new_policy
unsigned int cpu, bool new_policy)
if (!new_policy && cpufreq_driver->online) {
if (new_policy) {
if (new_policy) {
bool new_policy;
new_policy = false;
new_policy = true;
ret = cpufreq_policy_online(policy, cpu, new_policy);
if (new_policy && cpufreq_thermal_control_enabled(cpufreq_driver))
(new_policy || !cpufreq_boost_enabled())) {
struct dml_mode_eval_policy_st *new_policy;
struct dml_mode_eval_policy_st new_policy;
struct dml_mode_eval_policy_st new_policy;
if (p->new_policy != p->cur_policy)
*p->new_policy = *p->cur_policy;
p->new_policy->ODMUse[0] = dml_odm_use_policy_combine_2to1;
p->new_policy->ODMUse[0] = dml_odm_use_policy_combine_4to1;
s->optimize_configuration_params.new_policy = &s->new_policy;
dml2->v20.dml_core_ctx.policy = s->new_policy;
s->cur_policy = s->new_policy;
int new_policy = -1;
t.data = &new_policy;
if (ret || new_policy == -1)
mm_compute_batch(new_policy);
if (new_policy == OVERCOMMIT_NEVER)
sysctl_overcommit_memory = new_policy;
struct nf_conntrack_expect_policy *new_policy,
new_policy->max_expected =
if (new_policy->max_expected > NF_CT_EXPECT_MAX_CNT)
new_policy->timeout =
struct nf_conntrack_expect_policy *new_policy;
new_policy = kmalloc_objs(*new_policy, helper->expect_class_max + 1);
if (!new_policy)
&new_policy[i],
policy->max_expected = new_policy->max_expected;
policy->timeout = new_policy->timeout;
kfree(new_policy);
.write = new_policy,