arch/arm/kernel/smp.c
798
struct cpumask *cpus = freq->policy->cpus;
arch/arm/mm/mmu.c
101
.policy = "writealloc",
arch/arm/mm/mmu.c
146
int len = strlen(cache_policies[i].policy);
arch/arm/mm/mmu.c
148
if (memcmp(p, cache_policies[i].policy, len) == 0) {
arch/arm/mm/mmu.c
166
cache_policies[cachepolicy].policy);
arch/arm/mm/mmu.c
701
ecc_mask ? "ECC enabled, " : "", cp->policy);
arch/arm/mm/mmu.c
73
const char policy[16];
arch/arm/mm/mmu.c
81
.policy = "uncached",
arch/arm/mm/mmu.c
86
.policy = "buffered",
arch/arm/mm/mmu.c
91
.policy = "writethrough",
arch/arm/mm/mmu.c
96
.policy = "writeback",
arch/arm64/kernel/proton-pack.c
383
enum spectre_v4_policy policy;
arch/arm64/kernel/proton-pack.c
402
__spectre_v4_policy = param->policy;
arch/arm64/kernel/topology.c
209
struct cpufreq_policy *policy = cpufreq_cpu_get(cpu);
arch/arm64/kernel/topology.c
212
if (!policy)
arch/arm64/kernel/topology.c
215
if (!cpumask_intersects(policy->related_cpus,
arch/arm64/kernel/topology.c
217
cpufreq_cpu_put(policy);
arch/arm64/kernel/topology.c
221
for_each_cpu_wrap(ref_cpu, policy->cpus, cpu + 1) {
arch/arm64/kernel/topology.c
231
cpufreq_cpu_put(policy);
arch/arm64/kernel/topology.c
284
struct cpufreq_policy *policy = data;
arch/arm64/kernel/topology.c
287
amu_fie_setup(policy->cpus);
arch/arm64/kernel/topology.c
308
struct cpufreq_policy *policy = cpufreq_cpu_policy(cpu);
arch/arm64/kernel/topology.c
318
if (unlikely(!policy) || !cpumask_available(amu_fie_cpus) ||
arch/arm64/kernel/topology.c
326
if (unlikely(!cpumask_subset(policy->cpus, amu_fie_cpus)))
arch/arm64/kernel/topology.c
336
policy->related_cpus);
arch/arm64/kernel/topology.c
337
cpumask_andnot(amu_fie_cpus, amu_fie_cpus, policy->related_cpus);
arch/arm64/kernel/watchdog_hld.c
56
struct cpufreq_policy *policy = data;
arch/arm64/kernel/watchdog_hld.c
79
for_each_cpu(cpu, policy->cpus)
arch/mips/kernel/cevt-r4k.c
258
for_each_cpu(cpu, freq->policy->cpus) {
arch/mips/kernel/time.c
40
struct cpumask *cpus = freq->policy->cpus;
arch/powerpc/include/asm/plpks.h
56
u32 policy;
arch/powerpc/platforms/cell/spufs/sched.c
121
ctx->policy = current->policy;
arch/powerpc/platforms/cell/spufs/sched.c
913
if (ctx->policy == SCHED_FIFO)
arch/powerpc/platforms/cell/spufs/spufs.h
112
int policy;
arch/powerpc/platforms/pseries/plpks-secvar.c
124
var.policy = get_policy(key);
arch/powerpc/platforms/pseries/plpks.c
643
if (!(var->policy & PLPKS_SIGNEDUPDATE))
arch/powerpc/platforms/pseries/plpks.c
646
if (var->policy & PLPKS_WRAPPINGKEY)
arch/powerpc/platforms/pseries/plpks.c
666
label->size, var->policy, flags,
arch/powerpc/platforms/pseries/plpks.c
728
if (var.policy & PLPKS_SIGNEDUPDATE)
arch/powerpc/platforms/pseries/plpks.c
731
if (var.policy & PLPKS_WRAPPINGKEY)
arch/powerpc/platforms/pseries/plpks.c
745
virt_to_phys(label), label->size, var.policy,
arch/powerpc/platforms/pseries/plpks.c
828
if (var->policy & PLPKS_WRAPPINGKEY)
arch/powerpc/platforms/pseries/plpks.c
868
var->policy = retbuf[1];
arch/powerpc/platforms/pseries/plpks.c
938
.policy = PLPKS_WRAPPINGKEY,
arch/powerpc/platforms/pseries/plpks.c
955
label->size, var.policy,
arch/powerpc/platforms/pseries/plpks_sed_ops.c
60
var->policy = PLPKS_WORLDREADABLE;
arch/sparc/kernel/time_64.c
661
for_each_cpu(cpu, freq->policy->cpus) {
arch/x86/include/asm/efi.h
283
#define __efi64_argmap_load_file(protocol, path, policy, bufsize, buf) \
arch/x86/include/asm/efi.h
284
((protocol), (path), (policy), efi64_zero_upper(bufsize), (buf))
arch/x86/include/uapi/asm/kvm.h
771
__u32 policy;
arch/x86/include/uapi/asm/kvm.h
807
__u32 policy;
arch/x86/include/uapi/asm/kvm.h
826
__u32 policy;
arch/x86/include/uapi/asm/kvm.h
856
__u32 policy;
arch/x86/include/uapi/asm/kvm.h
878
__u64 policy;
arch/x86/kernel/tsc.c
1057
set_cyc2ns_scale(tsc_khz, freq->policy->cpu, rdtsc());
arch/x86/kvm/svm/sev.c
1133
params.policy = data.policy;
arch/x86/kvm/svm/sev.c
1598
params.policy = data.policy;
arch/x86/kvm/svm/sev.c
1779
start.policy = params.policy;
arch/x86/kvm/svm/sev.c
2228
if (params.policy & ~snp_supported_policy_bits)
arch/x86/kvm/svm/sev.c
2232
if (!(params.policy & SNP_POLICY_MASK_RSVD_MBO))
arch/x86/kvm/svm/sev.c
2247
start.policy = params.policy;
arch/x86/kvm/svm/sev.c
2257
sev->policy = params.policy;
arch/x86/kvm/svm/sev.c
5134
if (!(sev->policy & SNP_POLICY_MASK_DEBUG))
arch/x86/kvm/svm/sev.c
5137
if (sev->policy & SEV_POLICY_MASK_NODBG)
arch/x86/kvm/svm/sev.c
643
start.policy = params.policy;
arch/x86/kvm/svm/sev.c
665
sev->policy = params.policy;
arch/x86/kvm/svm/svm.h
102
unsigned long policy;
arch/x86/kvm/x86.c
9886
for_each_cpu(cpu, freq->policy->cpus)
arch/x86/kvm/x86.c
9908
struct cpufreq_policy *policy;
arch/x86/kvm/x86.c
9912
policy = cpufreq_cpu_get(cpu);
arch/x86/kvm/x86.c
9913
if (policy) {
arch/x86/kvm/x86.c
9914
if (policy->cpuinfo.max_freq)
arch/x86/kvm/x86.c
9915
max_tsc_khz = policy->cpuinfo.max_freq;
arch/x86/kvm/x86.c
9916
cpufreq_cpu_put(policy);
block/blk-flush.c
389
unsigned int policy = 0;
block/blk-flush.c
395
policy |= REQ_FSEQ_DATA;
block/blk-flush.c
402
policy |= REQ_FSEQ_PREFLUSH;
block/blk-flush.c
404
policy |= REQ_FSEQ_POSTFLUSH;
block/blk-flush.c
422
switch (policy) {
block/blk-flush.c
457
blk_flush_complete_seq(rq, fq, REQ_FSEQ_ACTIONS & ~policy, 0);
drivers/acpi/processor_driver.c
222
struct cpufreq_policy *policy = data;
drivers/acpi/processor_driver.c
225
acpi_thermal_cpufreq_init(policy);
drivers/acpi/processor_driver.c
226
acpi_processor_ppc_init(policy);
drivers/acpi/processor_driver.c
228
acpi_processor_ppc_exit(policy);
drivers/acpi/processor_driver.c
229
acpi_thermal_cpufreq_exit(policy);
drivers/acpi/processor_perflib.c
172
void acpi_processor_ppc_init(struct cpufreq_policy *policy)
drivers/acpi/processor_perflib.c
179
for_each_cpu(cpu, policy->related_cpus) {
drivers/acpi/processor_perflib.c
193
ret = freq_qos_add_request(&policy->constraints,
drivers/acpi/processor_perflib.c
210
void acpi_processor_ppc_exit(struct cpufreq_policy *policy)
drivers/acpi/processor_perflib.c
214
for_each_cpu(cpu, policy->related_cpus) {
drivers/acpi/processor_thermal.c
100
max_freq = (policy->cpuinfo.max_freq *
drivers/acpi/processor_thermal.c
160
void acpi_thermal_cpufreq_init(struct cpufreq_policy *policy)
drivers/acpi/processor_thermal.c
166
for_each_cpu(cpu, policy->related_cpus) {
drivers/acpi/processor_thermal.c
173
ret = freq_qos_add_request(&policy->constraints,
drivers/acpi/processor_thermal.c
186
void acpi_thermal_cpufreq_exit(struct cpufreq_policy *policy)
drivers/acpi/processor_thermal.c
190
for_each_cpu(cpu, policy->related_cpus) {
drivers/acpi/processor_thermal.c
70
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/acpi/processor_thermal.c
72
return policy != NULL;
drivers/acpi/processor_thermal.c
96
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/acpi/processor_thermal.c
97
if (!policy)
drivers/ata/ahci.c
1807
int policy = CONFIG_SATA_MOBILE_LPM_POLICY;
drivers/ata/ahci.c
1841
policy = mobile_lpm_policy;
drivers/ata/ahci.c
1845
if (policy > ATA_LPM_MED_POWER && pm_suspend_default_s2idle()) {
drivers/ata/ahci.c
1847
policy = ATA_LPM_MIN_POWER_WITH_PARTIAL;
drivers/ata/ahci.c
1849
policy = ATA_LPM_MIN_POWER;
drivers/ata/ahci.c
1853
if (policy >= ATA_LPM_UNKNOWN && policy <= ATA_LPM_MIN_POWER)
drivers/ata/ahci.c
1854
ap->target_lpm_policy = policy;
drivers/ata/ata_piix.c
812
static int piix_sidpr_set_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/ata_piix.c
815
return sata_link_scr_lpm(link, policy, false);
drivers/ata/libahci.c
46
static int ahci_set_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/libahci.c
809
static int ahci_set_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/libahci.c
817
if (policy != ATA_LPM_MAX_POWER) {
drivers/ata/libahci.c
829
sata_link_scr_lpm(link, policy, false);
drivers/ata/libahci.c
835
if (policy == ATA_LPM_MAX_POWER || !(hints & ATA_LPM_HIPM)) {
drivers/ata/libahci.c
850
if (policy == ATA_LPM_MIN_POWER)
drivers/ata/libahci.c
852
else if (policy == ATA_LPM_MIN_POWER_WITH_PARTIAL)
drivers/ata/libahci.c
864
if (policy == ATA_LPM_MIN_POWER ||
drivers/ata/libahci.c
865
policy == ATA_LPM_MIN_POWER_WITH_PARTIAL)
drivers/ata/libahci.c
871
if (policy == ATA_LPM_MAX_POWER) {
drivers/ata/libahci.c
872
sata_link_scr_lpm(link, policy, false);
drivers/ata/libata-eh.c
2148
enum ata_lpm_policy policy,
drivers/ata/libata-eh.c
2169
if (WARN_ON_ONCE(policy == ATA_LPM_UNKNOWN))
drivers/ata/libata-eh.c
2172
ata_link_dbg(link, "Set LPM policy: %d -> %d\n", old_policy, policy);
drivers/ata/libata-eh.c
2210
rc = ap->ops->set_lpm(link, policy, hints);
drivers/ata/libata-eh.c
2212
rc = ap->ops->set_lpm(ap->slave_link, policy, hints);
drivers/ata/libata-eh.c
2214
rc = sata_pmp_set_lpm(link, policy, hints);
drivers/ata/libata-eh.c
2233
link->lpm_policy = policy;
drivers/ata/libata-eh.c
2235
ap->slave_link->lpm_policy = policy;
drivers/ata/libata-eh.c
2245
if (policy >= ATA_LPM_MED_POWER_WITH_DIPM && host_has_dipm &&
drivers/ata/libata-pmp.c
203
int sata_pmp_set_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/libata-pmp.c
206
return sata_link_scr_lpm(link, policy, true);
drivers/ata/libata-sata.c
369
int sata_link_scr_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/libata-sata.c
381
switch (policy) {
drivers/ata/libata-sata.c
949
enum ata_lpm_policy policy;
drivers/ata/libata-sata.c
953
for (policy = ATA_LPM_MAX_POWER;
drivers/ata/libata-sata.c
954
policy < ARRAY_SIZE(ata_lpm_policy_names); policy++) {
drivers/ata/libata-sata.c
955
const char *name = ata_lpm_policy_names[policy];
drivers/ata/libata-sata.c
960
if (policy == ARRAY_SIZE(ata_lpm_policy_names))
drivers/ata/libata-sata.c
970
ap->target_lpm_policy = policy;
drivers/ata/libata-scsi.c
2505
u32 policy;
drivers/ata/libata-scsi.c
2535
policy = get_unaligned_le32(&cdl[0]);
drivers/ata/libata-scsi.c
2542
b[6] = ((policy >> 8) & 0x0f) << 4;
drivers/ata/libata-scsi.c
2546
b[6] |= (policy >> 4) & 0x0f;
drivers/ata/libata-scsi.c
2550
b[14] = policy & 0x0f;
drivers/ata/libata.h
208
extern int sata_pmp_set_lpm(struct ata_link *link, enum ata_lpm_policy policy,
drivers/ata/libata.h
223
enum ata_lpm_policy policy, unsigned hints)
drivers/base/arch_topology.c
394
struct cpufreq_policy *policy = data;
drivers/base/arch_topology.c
401
cpumask_pr_args(policy->related_cpus),
drivers/base/arch_topology.c
404
cpumask_andnot(cpus_to_visit, cpus_to_visit, policy->related_cpus);
drivers/base/arch_topology.c
406
for_each_cpu(cpu, policy->related_cpus) {
drivers/base/arch_topology.c
407
per_cpu(capacity_freq_ref, cpu) = policy->cpuinfo.max_freq;
drivers/block/drbd/drbd_nla.c
32
const struct nla_policy *policy)
drivers/block/drbd/drbd_nla.c
38
err = nla_parse_nested_deprecated(tb, maxtype, nla, policy,
drivers/block/drbd/drbd_nla.h
6
const struct nla_policy *policy);
drivers/block/nbd.c
2486
.policy = nbd_attr_policy,
drivers/clk/bcm/clk-bcm21664.c
127
.policy = {
drivers/clk/bcm/clk-bcm21664.c
230
.policy = {
drivers/clk/bcm/clk-bcm21664.c
44
.policy = {
drivers/clk/bcm/clk-iproc-armpll.c
69
unsigned int policy, fid, active_fid;
drivers/clk/bcm/clk-iproc-armpll.c
73
policy = val & IPROC_CLK_ARM_DIV_ARM_PLL_SELECT_MASK;
drivers/clk/bcm/clk-iproc-armpll.c
75
policy = 0;
drivers/clk/bcm/clk-iproc-armpll.c
78
BUG_ON(policy > IPROC_CLK_MAX_FREQ_POLICY);
drivers/clk/bcm/clk-iproc-armpll.c
81
fid = (val >> (IPROC_CLK_POLICY_FREQ_POLICY_FREQ_SHIFT * policy)) &
drivers/clk/bcm/clk-kona-setup.c
20
struct ccu_policy *ccu_policy = &ccu->policy;
drivers/clk/bcm/clk-kona-setup.c
231
static bool policy_valid(struct bcm_clk_policy *policy, const char *clock_name)
drivers/clk/bcm/clk-kona-setup.c
233
if (!bit_posn_valid(policy->bit, "policy", clock_name))
drivers/clk/bcm/clk-kona-setup.c
395
struct bcm_clk_policy *policy;
drivers/clk/bcm/clk-kona-setup.c
417
policy = &peri->policy;
drivers/clk/bcm/clk-kona-setup.c
418
if (policy_exists(policy) && !policy_valid(policy, name))
drivers/clk/bcm/clk-kona-setup.c
72
struct bcm_clk_policy *policy;
drivers/clk/bcm/clk-kona-setup.c
767
ccu_policy = &ccu->policy;
drivers/clk/bcm/clk-kona-setup.c
90
policy = &peri->policy;
drivers/clk/bcm/clk-kona-setup.c
91
if (policy_exists(policy)) {
drivers/clk/bcm/clk-kona-setup.c
92
if (policy->offset > limit) {
drivers/clk/bcm/clk-kona-setup.c
94
__func__, name, policy->offset, limit);
drivers/clk/bcm/clk-kona.c
1178
if (!policy_init(ccu, &peri->policy)) {
drivers/clk/bcm/clk-kona.c
191
struct bcm_policy_ctl *control = &ccu->policy.control;
drivers/clk/bcm/clk-kona.c
245
struct bcm_lvm_en *enable = &ccu->policy.enable;
drivers/clk/bcm/clk-kona.c
285
static bool policy_init(struct ccu_data *ccu, struct bcm_clk_policy *policy)
drivers/clk/bcm/clk-kona.c
292
if (!policy_exists(policy))
drivers/clk/bcm/clk-kona.c
309
offset = policy->offset;
drivers/clk/bcm/clk-kona.c
310
mask = (u32)1 << policy->bit;
drivers/clk/bcm/clk-kona.h
383
struct bcm_clk_policy policy;
drivers/clk/bcm/clk-kona.h
44
#define policy_exists(policy) ((policy)->offset != 0)
drivers/clk/bcm/clk-kona.h
474
struct ccu_policy policy;
drivers/clk/qcom/clk-rcg2.c
1370
enum freq_policy policy)
drivers/clk/qcom/clk-rcg2.c
1375
switch (policy) {
drivers/clk/qcom/clk-rcg2.c
231
enum freq_policy policy)
drivers/clk/qcom/clk-rcg2.c
238
switch (policy) {
drivers/clk/qcom/clk-rcg2.c
621
enum freq_policy policy)
drivers/clk/qcom/clk-rcg2.c
626
switch (policy) {
drivers/clk/tegra/clk-tegra20.c
958
unsigned int reg, policy;
drivers/clk/tegra/clk-tegra20.c
963
policy = (reg >> CCLK_BURST_POLICY_SHIFT) & 0xF;
drivers/clk/tegra/clk-tegra20.c
965
if (policy == CCLK_IDLE_POLICY)
drivers/clk/tegra/clk-tegra20.c
967
else if (policy == CCLK_RUN_POLICY)
drivers/clk/tegra/clk-tegra30.c
1156
unsigned int reg, policy;
drivers/clk/tegra/clk-tegra30.c
1161
policy = (reg >> CLK_RESET_CCLK_BURST_POLICY_SHIFT) & 0xF;
drivers/clk/tegra/clk-tegra30.c
1163
if (policy == CLK_RESET_CCLK_IDLE_POLICY)
drivers/clk/tegra/clk-tegra30.c
1165
else if (policy == CLK_RESET_CCLK_RUN_POLICY)
drivers/cpufreq/acpi-cpufreq.c
131
static int set_boost(struct cpufreq_policy *policy, int val)
drivers/cpufreq/acpi-cpufreq.c
133
on_each_cpu_mask(policy->cpus, boost_set_msr_each,
drivers/cpufreq/acpi-cpufreq.c
136
cpumask_pr_args(policy->cpus), str_enabled_disabled(val));
drivers/cpufreq/acpi-cpufreq.c
141
static ssize_t show_freqdomain_cpus(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/acpi-cpufreq.c
143
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
154
static ssize_t store_cpb(struct cpufreq_policy *policy, const char *buf,
drivers/cpufreq/acpi-cpufreq.c
168
set_boost(policy, val);
drivers/cpufreq/acpi-cpufreq.c
174
static ssize_t show_cpb(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/acpi-cpufreq.c
196
static unsigned extract_io(struct cpufreq_policy *policy, u32 value)
drivers/cpufreq/acpi-cpufreq.c
198
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
206
return policy->freq_table[i].frequency;
drivers/cpufreq/acpi-cpufreq.c
211
static unsigned extract_msr(struct cpufreq_policy *policy, u32 msr)
drivers/cpufreq/acpi-cpufreq.c
213
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
226
cpufreq_for_each_entry(pos, policy->freq_table)
drivers/cpufreq/acpi-cpufreq.c
229
return policy->freq_table[0].frequency;
drivers/cpufreq/acpi-cpufreq.c
232
static unsigned extract_freq(struct cpufreq_policy *policy, u32 val)
drivers/cpufreq/acpi-cpufreq.c
234
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
239
return extract_msr(policy, val);
drivers/cpufreq/acpi-cpufreq.c
241
return extract_io(policy, val);
drivers/cpufreq/acpi-cpufreq.c
358
struct cpufreq_policy *policy;
drivers/cpufreq/acpi-cpufreq.c
364
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/acpi-cpufreq.c
365
if (unlikely(!policy))
drivers/cpufreq/acpi-cpufreq.c
368
data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
369
if (unlikely(!data || !policy->freq_table))
drivers/cpufreq/acpi-cpufreq.c
372
cached_freq = policy->freq_table[to_perf_data(data)->state].frequency;
drivers/cpufreq/acpi-cpufreq.c
373
freq = extract_freq(policy, get_cur_val(cpumask_of(cpu), data));
drivers/cpufreq/acpi-cpufreq.c
387
static unsigned int check_freqs(struct cpufreq_policy *policy,
drivers/cpufreq/acpi-cpufreq.c
390
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
395
cur_freq = extract_freq(policy, get_cur_val(mask, data));
drivers/cpufreq/acpi-cpufreq.c
403
static int acpi_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/acpi-cpufreq.c
406
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
417
next_perf_state = policy->freq_table[index].driver_data;
drivers/cpufreq/acpi-cpufreq.c
434
mask = policy->shared_type == CPUFREQ_SHARED_TYPE_ANY ?
drivers/cpufreq/acpi-cpufreq.c
435
cpumask_of(policy->cpu) : policy->cpus;
drivers/cpufreq/acpi-cpufreq.c
440
if (!check_freqs(policy, mask,
drivers/cpufreq/acpi-cpufreq.c
441
policy->freq_table[index].frequency)) {
drivers/cpufreq/acpi-cpufreq.c
442
pr_debug("%s (%d)\n", __func__, policy->cpu);
drivers/cpufreq/acpi-cpufreq.c
453
static unsigned int acpi_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/acpi-cpufreq.c
456
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
464
if (policy->cached_target_freq == target_freq)
drivers/cpufreq/acpi-cpufreq.c
465
index = policy->cached_resolved_idx;
drivers/cpufreq/acpi-cpufreq.c
467
index = cpufreq_table_find_index_dl(policy, target_freq,
drivers/cpufreq/acpi-cpufreq.c
470
entry = &policy->freq_table[index];
drivers/cpufreq/acpi-cpufreq.c
678
static int acpi_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/acpi-cpufreq.c
683
unsigned int cpu = policy->cpu;
drivers/cpufreq/acpi-cpufreq.c
714
policy->driver_data = data;
drivers/cpufreq/acpi-cpufreq.c
723
policy->shared_type = perf->shared_type;
drivers/cpufreq/acpi-cpufreq.c
729
if (policy->shared_type == CPUFREQ_SHARED_TYPE_ALL ||
drivers/cpufreq/acpi-cpufreq.c
730
policy->shared_type == CPUFREQ_SHARED_TYPE_ANY) {
drivers/cpufreq/acpi-cpufreq.c
731
cpumask_copy(policy->cpus, perf->shared_cpu_map);
drivers/cpufreq/acpi-cpufreq.c
737
if (bios_with_sw_any_bug && !policy_is_shared(policy)) {
drivers/cpufreq/acpi-cpufreq.c
738
policy->shared_type = CPUFREQ_SHARED_TYPE_ALL;
drivers/cpufreq/acpi-cpufreq.c
739
cpumask_copy(policy->cpus, topology_core_cpumask(cpu));
drivers/cpufreq/acpi-cpufreq.c
744
cpumask_clear(policy->cpus);
drivers/cpufreq/acpi-cpufreq.c
745
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/acpi-cpufreq.c
748
policy->shared_type = CPUFREQ_SHARED_TYPE_HW;
drivers/cpufreq/acpi-cpufreq.c
808
policy->cpuinfo.transition_latency = 0;
drivers/cpufreq/acpi-cpufreq.c
811
policy->cpuinfo.transition_latency)
drivers/cpufreq/acpi-cpufreq.c
812
policy->cpuinfo.transition_latency =
drivers/cpufreq/acpi-cpufreq.c
818
policy->cpuinfo.transition_latency > 20 * 1000) {
drivers/cpufreq/acpi-cpufreq.c
819
policy->cpuinfo.transition_latency = 20 * 1000;
drivers/cpufreq/acpi-cpufreq.c
850
policy->cpuinfo.max_freq = freq * max_boost_ratio >> SCHED_CAPACITY_SHIFT;
drivers/cpufreq/acpi-cpufreq.c
861
policy->freq_table = freq_table;
drivers/cpufreq/acpi-cpufreq.c
872
policy->cur = acpi_cpufreq_guess_freq(data, policy->cpu);
drivers/cpufreq/acpi-cpufreq.c
898
policy->fast_switch_possible = !acpi_pstate_strict &&
drivers/cpufreq/acpi-cpufreq.c
899
!(policy_is_shared(policy) && policy->shared_type != CPUFREQ_SHARED_TYPE_ANY);
drivers/cpufreq/acpi-cpufreq.c
905
if (policy->boost_supported) {
drivers/cpufreq/acpi-cpufreq.c
911
if (policy->boost_enabled != boost_state(cpu))
drivers/cpufreq/acpi-cpufreq.c
912
set_boost(policy, policy->boost_enabled);
drivers/cpufreq/acpi-cpufreq.c
914
policy->boost_supported = true;
drivers/cpufreq/acpi-cpufreq.c
926
policy->driver_data = NULL;
drivers/cpufreq/acpi-cpufreq.c
931
static void acpi_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/acpi-cpufreq.c
933
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/acpi-cpufreq.c
937
cpufreq_boost_down_prep(policy->cpu);
drivers/cpufreq/acpi-cpufreq.c
938
policy->fast_switch_possible = false;
drivers/cpufreq/acpi-cpufreq.c
939
policy->driver_data = NULL;
drivers/cpufreq/acpi-cpufreq.c
942
kfree(policy->freq_table);
drivers/cpufreq/acpi-cpufreq.c
946
static int acpi_cpufreq_resume(struct cpufreq_policy *policy)
drivers/cpufreq/acpi-cpufreq.c
948
struct acpi_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/amd-pstate-ut.c
122
struct cpufreq_policy *policy __free(put_cpufreq_policy) = NULL;
drivers/cpufreq/amd-pstate-ut.c
125
policy = cpufreq_cpu_get(cpu);
drivers/cpufreq/amd-pstate-ut.c
126
if (!policy)
drivers/cpufreq/amd-pstate-ut.c
128
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate-ut.c
194
struct cpufreq_policy *policy __free(put_cpufreq_policy) = NULL;
drivers/cpufreq/amd-pstate-ut.c
197
policy = cpufreq_cpu_get(cpu);
drivers/cpufreq/amd-pstate-ut.c
198
if (!policy)
drivers/cpufreq/amd-pstate-ut.c
200
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate-ut.c
202
if (!((policy->cpuinfo.max_freq >= cpudata->nominal_freq) &&
drivers/cpufreq/amd-pstate-ut.c
204
(cpudata->lowest_nonlinear_freq >= policy->cpuinfo.min_freq) &&
drivers/cpufreq/amd-pstate-ut.c
205
(policy->cpuinfo.min_freq > 0))) {
drivers/cpufreq/amd-pstate-ut.c
207
__func__, cpu, policy->cpuinfo.max_freq, cpudata->nominal_freq,
drivers/cpufreq/amd-pstate-ut.c
208
cpudata->lowest_nonlinear_freq, policy->cpuinfo.min_freq);
drivers/cpufreq/amd-pstate-ut.c
212
if (cpudata->lowest_nonlinear_freq != policy->min) {
drivers/cpufreq/amd-pstate-ut.c
214
__func__, cpu, cpudata->lowest_nonlinear_freq, policy->min);
drivers/cpufreq/amd-pstate-ut.c
219
if ((policy->max != policy->cpuinfo.max_freq) &&
drivers/cpufreq/amd-pstate-ut.c
220
(policy->max != cpudata->nominal_freq)) {
drivers/cpufreq/amd-pstate-ut.c
222
__func__, cpu, policy->max, policy->cpuinfo.max_freq,
drivers/cpufreq/amd-pstate.c
1000
cpudata->cpu = policy->cpu;
drivers/cpufreq/amd-pstate.c
1016
policy->cpuinfo.transition_latency = amd_pstate_get_transition_latency(policy->cpu);
drivers/cpufreq/amd-pstate.c
1017
policy->transition_delay_us = amd_pstate_get_transition_delay_us(policy->cpu);
drivers/cpufreq/amd-pstate.c
1021
policy->cpuinfo.min_freq = policy->min = perf_to_freq(perf,
drivers/cpufreq/amd-pstate.c
1024
policy->cpuinfo.max_freq = policy->max = perf_to_freq(perf,
drivers/cpufreq/amd-pstate.c
1028
ret = amd_pstate_cppc_enable(policy);
drivers/cpufreq/amd-pstate.c
1032
policy->boost_supported = READ_ONCE(cpudata->boost_supported);
drivers/cpufreq/amd-pstate.c
1035
policy->cur = policy->cpuinfo.min_freq;
drivers/cpufreq/amd-pstate.c
1038
policy->fast_switch_possible = true;
drivers/cpufreq/amd-pstate.c
1040
ret = freq_qos_add_request(&policy->constraints, &cpudata->req[0],
drivers/cpufreq/amd-pstate.c
1047
ret = freq_qos_add_request(&policy->constraints, &cpudata->req[1],
drivers/cpufreq/amd-pstate.c
1048
FREQ_QOS_MAX, policy->cpuinfo.max_freq);
drivers/cpufreq/amd-pstate.c
1054
policy->driver_data = cpudata;
drivers/cpufreq/amd-pstate.c
1064
pr_warn("Failed to initialize CPU %d: %d\n", policy->cpu, ret);
drivers/cpufreq/amd-pstate.c
1069
static void amd_pstate_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1071
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1075
amd_pstate_update_perf(policy, perf.bios_min_perf, 0U, 0U, 0U, false);
drivers/cpufreq/amd-pstate.c
1079
policy->fast_switch_possible = false;
drivers/cpufreq/amd-pstate.c
1090
static ssize_t show_amd_pstate_max_freq(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
1096
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1103
static ssize_t show_amd_pstate_lowest_nonlinear_freq(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
1109
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1120
static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
1125
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1130
static ssize_t show_amd_pstate_prefcore_ranking(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
1134
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1141
static ssize_t show_amd_pstate_hw_prefcore(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
1145
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1153
struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/amd-pstate.c
1156
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1158
if (cpudata->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/amd-pstate.c
1171
struct cpufreq_policy *policy, const char *buf, size_t count)
drivers/cpufreq/amd-pstate.c
1173
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1186
if (epp > 0 && policy->policy == CPUFREQ_POLICY_PERFORMANCE) {
drivers/cpufreq/amd-pstate.c
1191
ret = amd_pstate_set_epp(policy, epp);
drivers/cpufreq/amd-pstate.c
1197
struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/amd-pstate.c
1199
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1465
static int amd_pstate_epp_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1476
amd_perf_ctl_reset(policy->cpu);
drivers/cpufreq/amd-pstate.c
1477
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/amd-pstate.c
1485
cpudata->cpu = policy->cpu;
drivers/cpufreq/amd-pstate.c
1503
policy->cpuinfo.min_freq = policy->min = perf_to_freq(perf,
drivers/cpufreq/amd-pstate.c
1506
policy->cpuinfo.max_freq = policy->max = perf_to_freq(perf,
drivers/cpufreq/amd-pstate.c
1509
policy->driver_data = cpudata;
drivers/cpufreq/amd-pstate.c
1511
ret = amd_pstate_cppc_enable(policy);
drivers/cpufreq/amd-pstate.c
1516
policy->cur = policy->cpuinfo.min_freq;
drivers/cpufreq/amd-pstate.c
1519
policy->boost_supported = READ_ONCE(cpudata->boost_supported);
drivers/cpufreq/amd-pstate.c
1527
policy->policy = CPUFREQ_POLICY_PERFORMANCE;
drivers/cpufreq/amd-pstate.c
1530
policy->policy = CPUFREQ_POLICY_POWERSAVE;
drivers/cpufreq/amd-pstate.c
1534
ret = amd_pstate_set_epp(policy, cpudata->epp_default);
drivers/cpufreq/amd-pstate.c
1543
pr_warn("Failed to initialize CPU %d: %d\n", policy->cpu, ret);
drivers/cpufreq/amd-pstate.c
1548
static void amd_pstate_epp_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1550
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1556
amd_pstate_update_perf(policy, perf.bios_min_perf, 0U, 0U, 0U, false);
drivers/cpufreq/amd-pstate.c
1559
policy->driver_data = NULL;
drivers/cpufreq/amd-pstate.c
1562
pr_debug("CPU %d exiting\n", policy->cpu);
drivers/cpufreq/amd-pstate.c
1565
static int amd_pstate_epp_update_limit(struct cpufreq_policy *policy, bool policy_change)
drivers/cpufreq/amd-pstate.c
1567
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1572
policy->min != cpudata->min_limit_freq ||
drivers/cpufreq/amd-pstate.c
1573
policy->max != cpudata->max_limit_freq)
drivers/cpufreq/amd-pstate.c
1574
amd_pstate_update_min_max_limit(policy);
drivers/cpufreq/amd-pstate.c
1576
if (cpudata->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/amd-pstate.c
1583
return amd_pstate_update_perf(policy, perf.min_limit_perf, 0U,
drivers/cpufreq/amd-pstate.c
1587
static int amd_pstate_epp_set_policy(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1589
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1592
if (!policy->cpuinfo.max_freq)
drivers/cpufreq/amd-pstate.c
1595
cpudata->policy = policy->policy;
drivers/cpufreq/amd-pstate.c
1597
ret = amd_pstate_epp_update_limit(policy, true);
drivers/cpufreq/amd-pstate.c
1605
policy->cur = policy->min;
drivers/cpufreq/amd-pstate.c
1610
static int amd_pstate_cpu_online(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1612
return amd_pstate_cppc_enable(policy);
drivers/cpufreq/amd-pstate.c
1615
static int amd_pstate_cpu_offline(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1617
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1625
return amd_pstate_update_perf(policy, perf.bios_min_perf,
drivers/cpufreq/amd-pstate.c
1632
static int amd_pstate_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1634
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1643
ret = amd_pstate_update_perf(policy, perf.bios_min_perf,
drivers/cpufreq/amd-pstate.c
1657
static int amd_pstate_resume(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1659
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1661
int cur_perf = freq_to_perf(perf, cpudata->nominal_freq, policy->cur);
drivers/cpufreq/amd-pstate.c
1664
return amd_pstate_update_perf(policy, perf.min_limit_perf, cur_perf, perf.max_limit_perf,
drivers/cpufreq/amd-pstate.c
1668
static int amd_pstate_epp_resume(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
1670
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
1676
ret = amd_pstate_epp_update_limit(policy, false);
drivers/cpufreq/amd-pstate.c
232
static int msr_update_perf(struct cpufreq_policy *policy, u8 min_perf,
drivers/cpufreq/amd-pstate.c
235
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
255
policy->boost_enabled,
drivers/cpufreq/amd-pstate.c
279
static inline int amd_pstate_update_perf(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
284
return static_call(amd_pstate_update_perf)(policy, min_perf, des_perf,
drivers/cpufreq/amd-pstate.c
288
static int msr_set_epp(struct cpufreq_policy *policy, u8 epp)
drivers/cpufreq/amd-pstate.c
290
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
307
policy->boost_enabled,
drivers/cpufreq/amd-pstate.c
328
static inline int amd_pstate_set_epp(struct cpufreq_policy *policy, u8 epp)
drivers/cpufreq/amd-pstate.c
330
return static_call(amd_pstate_set_epp)(policy, epp);
drivers/cpufreq/amd-pstate.c
333
static int shmem_set_epp(struct cpufreq_policy *policy, u8 epp)
drivers/cpufreq/amd-pstate.c
335
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
352
policy->boost_enabled,
drivers/cpufreq/amd-pstate.c
374
static inline int msr_cppc_enable(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
376
return wrmsrq_safe_on_cpu(policy->cpu, MSR_AMD_CPPC_ENABLE, 1);
drivers/cpufreq/amd-pstate.c
379
static int shmem_cppc_enable(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
381
return cppc_set_enable(policy->cpu, 1);
drivers/cpufreq/amd-pstate.c
386
static inline int amd_pstate_cppc_enable(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
388
return static_call(amd_pstate_cppc_enable)(policy);
drivers/cpufreq/amd-pstate.c
483
static int shmem_update_perf(struct cpufreq_policy *policy, u8 min_perf,
drivers/cpufreq/amd-pstate.c
486
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
492
int ret = shmem_set_epp(policy, epp);
drivers/cpufreq/amd-pstate.c
515
policy->boost_enabled,
drivers/cpufreq/amd-pstate.c
571
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpudata->cpu);
drivers/cpufreq/amd-pstate.c
574
if (!policy)
drivers/cpufreq/amd-pstate.c
583
policy->cur = perf_to_freq(perf, cpudata->nominal_freq, des_perf);
drivers/cpufreq/amd-pstate.c
596
amd_pstate_update_perf(policy, min_perf, des_perf, max_perf, 0, fast_switch);
drivers/cpufreq/amd-pstate.c
607
struct cpufreq_policy *policy __free(put_cpufreq_policy) =
drivers/cpufreq/amd-pstate.c
612
if (!policy)
drivers/cpufreq/amd-pstate.c
615
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
630
static void amd_pstate_update_min_max_limit(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
632
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
635
perf.max_limit_perf = freq_to_perf(perf, cpudata->nominal_freq, policy->max);
drivers/cpufreq/amd-pstate.c
636
WRITE_ONCE(cpudata->max_limit_freq, policy->max);
drivers/cpufreq/amd-pstate.c
638
if (cpudata->policy == CPUFREQ_POLICY_PERFORMANCE) {
drivers/cpufreq/amd-pstate.c
655
perf.min_limit_perf = freq_to_perf(perf, cpudata->nominal_freq, policy->min);
drivers/cpufreq/amd-pstate.c
656
WRITE_ONCE(cpudata->min_limit_freq, policy->min);
drivers/cpufreq/amd-pstate.c
662
static int amd_pstate_update_freq(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
670
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
672
if (policy->min != cpudata->min_limit_freq || policy->max != cpudata->max_limit_freq)
drivers/cpufreq/amd-pstate.c
673
amd_pstate_update_min_max_limit(policy);
drivers/cpufreq/amd-pstate.c
677
freqs.old = policy->cur;
drivers/cpufreq/amd-pstate.c
682
WARN_ON(fast_switch && !policy->fast_switch_enabled);
drivers/cpufreq/amd-pstate.c
689
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/amd-pstate.c
693
policy->governor->flags);
drivers/cpufreq/amd-pstate.c
696
cpufreq_freq_transition_end(policy, &freqs, false);
drivers/cpufreq/amd-pstate.c
701
static int amd_pstate_target(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
705
return amd_pstate_update_freq(policy, target_freq, false);
drivers/cpufreq/amd-pstate.c
708
static unsigned int amd_pstate_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/amd-pstate.c
711
if (!amd_pstate_update_freq(policy, target_freq, true))
drivers/cpufreq/amd-pstate.c
713
return policy->cur;
drivers/cpufreq/amd-pstate.c
722
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/amd-pstate.c
726
if (!policy)
drivers/cpufreq/amd-pstate.c
729
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
731
if (policy->min != cpudata->min_limit_freq || policy->max != cpudata->max_limit_freq)
drivers/cpufreq/amd-pstate.c
732
amd_pstate_update_min_max_limit(policy);
drivers/cpufreq/amd-pstate.c
754
policy->governor->flags);
drivers/cpufreq/amd-pstate.c
757
static int amd_pstate_cpu_boost_update(struct cpufreq_policy *policy, bool on)
drivers/cpufreq/amd-pstate.c
759
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
768
policy->cpuinfo.max_freq = max_freq;
drivers/cpufreq/amd-pstate.c
769
else if (policy->cpuinfo.max_freq > nominal_freq)
drivers/cpufreq/amd-pstate.c
770
policy->cpuinfo.max_freq = nominal_freq;
drivers/cpufreq/amd-pstate.c
772
policy->max = policy->cpuinfo.max_freq;
drivers/cpufreq/amd-pstate.c
775
ret = freq_qos_update_request(&cpudata->req[1], policy->cpuinfo.max_freq);
drivers/cpufreq/amd-pstate.c
783
static int amd_pstate_set_boost(struct cpufreq_policy *policy, int state)
drivers/cpufreq/amd-pstate.c
785
struct amd_cpudata *cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
793
ret = amd_pstate_cpu_boost_update(policy, state);
drivers/cpufreq/amd-pstate.c
794
refresh_frequency_limits(policy);
drivers/cpufreq/amd-pstate.c
857
static void amd_pstate_update_limits(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
862
unsigned int cpu = policy->cpu;
drivers/cpufreq/amd-pstate.c
870
cpudata = policy->driver_data;
drivers/cpufreq/amd-pstate.c
980
static int amd_pstate_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/amd-pstate.c
991
amd_perf_ctl_reset(policy->cpu);
drivers/cpufreq/amd-pstate.c
992
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/amd-pstate.h
107
u32 policy;
drivers/cpufreq/amd_freq_sensitivity.c
39
static unsigned int amd_powersave_bias_target(struct cpufreq_policy *policy,
drivers/cpufreq/amd_freq_sensitivity.c
46
struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu);
drivers/cpufreq/amd_freq_sensitivity.c
47
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/amd_freq_sensitivity.c
51
if (!policy->freq_table)
drivers/cpufreq/amd_freq_sensitivity.c
54
rdmsr_on_cpu(policy->cpu, MSR_AMD64_FREQ_SENSITIVITY_ACTUAL,
drivers/cpufreq/amd_freq_sensitivity.c
56
rdmsr_on_cpu(policy->cpu, MSR_AMD64_FREQ_SENSITIVITY_REFERENCE,
drivers/cpufreq/amd_freq_sensitivity.c
63
freq_next = policy->cur;
drivers/cpufreq/amd_freq_sensitivity.c
72
freq_next = policy->cur;
drivers/cpufreq/amd_freq_sensitivity.c
83
if (data->freq_prev == policy->cur)
drivers/cpufreq/amd_freq_sensitivity.c
84
freq_next = policy->cur;
drivers/cpufreq/amd_freq_sensitivity.c
86
if (freq_next > policy->cur)
drivers/cpufreq/amd_freq_sensitivity.c
87
freq_next = policy->cur;
drivers/cpufreq/amd_freq_sensitivity.c
88
else if (freq_next < policy->cur)
drivers/cpufreq/amd_freq_sensitivity.c
89
freq_next = policy->min;
drivers/cpufreq/amd_freq_sensitivity.c
93
index = cpufreq_table_find_index_h(policy,
drivers/cpufreq/amd_freq_sensitivity.c
94
policy->cur - 1,
drivers/cpufreq/amd_freq_sensitivity.c
96
freq_next = policy->freq_table[index].frequency;
drivers/cpufreq/apple-soc-cpufreq.c
137
struct cpufreq_policy *policy;
drivers/cpufreq/apple-soc-cpufreq.c
142
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/apple-soc-cpufreq.c
143
if (unlikely(!policy))
drivers/cpufreq/apple-soc-cpufreq.c
146
priv = policy->driver_data;
drivers/cpufreq/apple-soc-cpufreq.c
162
cpufreq_for_each_valid_entry(p, policy->freq_table)
drivers/cpufreq/apple-soc-cpufreq.c
171
static int apple_soc_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/apple-soc-cpufreq.c
174
struct apple_cpu_priv *priv = policy->driver_data;
drivers/cpufreq/apple-soc-cpufreq.c
175
unsigned int pstate = policy->freq_table[index].driver_data;
drivers/cpufreq/apple-soc-cpufreq.c
201
static unsigned int apple_soc_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/apple-soc-cpufreq.c
204
if (apple_soc_cpufreq_set_target(policy, policy->cached_resolved_idx) < 0)
drivers/cpufreq/apple-soc-cpufreq.c
207
return policy->freq_table[policy->cached_resolved_idx].frequency;
drivers/cpufreq/apple-soc-cpufreq.c
210
static int apple_soc_cpufreq_find_cluster(struct cpufreq_policy *policy,
drivers/cpufreq/apple-soc-cpufreq.c
218
ret = of_perf_domain_get_sharing_cpumask(policy->cpu, "performance-domains",
drivers/cpufreq/apple-soc-cpufreq.c
220
policy->cpus, &args);
drivers/cpufreq/apple-soc-cpufreq.c
238
static int apple_soc_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/apple-soc-cpufreq.c
248
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/apple-soc-cpufreq.c
250
pr_err("failed to get cpu%d device\n", policy->cpu);
drivers/cpufreq/apple-soc-cpufreq.c
260
ret = apple_soc_cpufreq_find_cluster(policy, ®_base, &info);
drivers/cpufreq/apple-soc-cpufreq.c
266
ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/apple-soc-cpufreq.c
307
policy->driver_data = priv;
drivers/cpufreq/apple-soc-cpufreq.c
308
policy->freq_table = freq_table;
drivers/cpufreq/apple-soc-cpufreq.c
314
policy->cpuinfo.transition_latency = transition_latency;
drivers/cpufreq/apple-soc-cpufreq.c
315
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/apple-soc-cpufreq.c
316
policy->fast_switch_possible = true;
drivers/cpufreq/apple-soc-cpufreq.c
317
policy->suspend_freq = freq_table[0].frequency;
drivers/cpufreq/apple-soc-cpufreq.c
332
static void apple_soc_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/apple-soc-cpufreq.c
334
struct apple_cpu_priv *priv = policy->driver_data;
drivers/cpufreq/apple-soc-cpufreq.c
336
dev_pm_opp_free_cpufreq_table(priv->cpu_dev, &policy->freq_table);
drivers/cpufreq/armada-37xx-cpufreq.c
362
static int armada37xx_cpufreq_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/armada-37xx-cpufreq.c
375
static int armada37xx_cpufreq_resume(struct cpufreq_policy *policy)
drivers/cpufreq/bmips-cpufreq.c
105
static int bmips_cpufreq_target_index(struct cpufreq_policy *policy,
drivers/cpufreq/bmips-cpufreq.c
108
unsigned int div = policy->freq_table[index].driver_data;
drivers/cpufreq/bmips-cpufreq.c
124
static void bmips_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/bmips-cpufreq.c
126
kfree(policy->freq_table);
drivers/cpufreq/bmips-cpufreq.c
129
static int bmips_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/bmips-cpufreq.c
133
freq_table = bmips_cpufreq_get_freq_table(policy);
drivers/cpufreq/bmips-cpufreq.c
140
cpufreq_generic_init(policy, freq_table, TRANSITION_LATENCY);
drivers/cpufreq/bmips-cpufreq.c
66
bmips_cpufreq_get_freq_table(const struct cpufreq_policy *policy)
drivers/cpufreq/brcmstb-avs-cpufreq.c
483
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/brcmstb-avs-cpufreq.c
486
if (!policy)
drivers/cpufreq/brcmstb-avs-cpufreq.c
489
priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
494
static int brcm_avs_target_index(struct cpufreq_policy *policy,
drivers/cpufreq/brcmstb-avs-cpufreq.c
497
return brcm_avs_set_pstate(policy->driver_data,
drivers/cpufreq/brcmstb-avs-cpufreq.c
498
policy->freq_table[index].driver_data);
drivers/cpufreq/brcmstb-avs-cpufreq.c
501
static int brcm_avs_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/brcmstb-avs-cpufreq.c
503
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
526
static int brcm_avs_resume(struct cpufreq_policy *policy)
drivers/cpufreq/brcmstb-avs-cpufreq.c
528
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
617
static int brcm_avs_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/brcmstb-avs-cpufreq.c
627
policy->driver_data = priv;
drivers/cpufreq/brcmstb-avs-cpufreq.c
637
policy->freq_table = freq_table;
drivers/cpufreq/brcmstb-avs-cpufreq.c
640
cpumask_setall(policy->cpus);
drivers/cpufreq/brcmstb-avs-cpufreq.c
648
policy->cur = freq_table[pstate].frequency;
drivers/cpufreq/brcmstb-avs-cpufreq.c
659
static ssize_t show_brcm_avs_pstate(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/brcmstb-avs-cpufreq.c
661
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
670
static ssize_t show_brcm_avs_mode(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/brcmstb-avs-cpufreq.c
672
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
682
static ssize_t show_brcm_avs_pmap(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/brcmstb-avs-cpufreq.c
685
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
700
static ssize_t show_brcm_avs_voltage(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/brcmstb-avs-cpufreq.c
702
struct private_data *priv = policy->driver_data;
drivers/cpufreq/brcmstb-avs-cpufreq.c
707
static ssize_t show_brcm_avs_frequency(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/brcmstb-avs-cpufreq.c
709
struct private_data *priv = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
147
static void cppc_cpufreq_cpu_fie_init(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
156
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/cppc_cpufreq.c
159
cppc_fi->cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
180
topology_set_scale_freq_source(sftd, policy->cpus);
drivers/cpufreq/cppc_cpufreq.c
191
static void cppc_cpufreq_cpu_fie_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
200
topology_clear_scale_freq_source(SCALE_FREQ_SOURCE_CPPC, policy->related_cpus);
drivers/cpufreq/cppc_cpufreq.c
202
for_each_cpu(cpu, policy->related_cpus) {
drivers/cpufreq/cppc_cpufreq.c
273
static inline void cppc_cpufreq_cpu_fie_init(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
277
static inline void cppc_cpufreq_cpu_fie_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
290
static int cppc_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/cppc_cpufreq.c
294
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
295
unsigned int cpu = policy->cpu;
drivers/cpufreq/cppc_cpufreq.c
301
freqs.old = policy->cur;
drivers/cpufreq/cppc_cpufreq.c
304
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/cppc_cpufreq.c
306
cpufreq_freq_transition_end(policy, &freqs, ret != 0);
drivers/cpufreq/cppc_cpufreq.c
315
static unsigned int cppc_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/cppc_cpufreq.c
318
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
319
unsigned int cpu = policy->cpu;
drivers/cpufreq/cppc_cpufreq.c
336
static int cppc_verify_policy(struct cpufreq_policy_data *policy)
drivers/cpufreq/cppc_cpufreq.c
338
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/cppc_cpufreq.c
395
static unsigned int get_perf_level_count(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
400
int cpu = policy->cpu;
drivers/cpufreq/cppc_cpufreq.c
402
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
429
struct cpufreq_policy *policy;
drivers/cpufreq/cppc_cpufreq.c
434
policy = cpufreq_cpu_get_raw(cpu_dev->id);
drivers/cpufreq/cppc_cpufreq.c
435
if (!policy)
drivers/cpufreq/cppc_cpufreq.c
438
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
499
struct cpufreq_policy *policy;
drivers/cpufreq/cppc_cpufreq.c
504
policy = cpufreq_cpu_get_raw(cpu_dev->id);
drivers/cpufreq/cppc_cpufreq.c
505
if (!policy)
drivers/cpufreq/cppc_cpufreq.c
508
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
521
static void cppc_cpufreq_register_em(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
527
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
528
em_dev_register_perf_domain(get_cpu_device(policy->cpu),
drivers/cpufreq/cppc_cpufreq.c
529
get_perf_level_count(policy), &em_cb,
drivers/cpufreq/cppc_cpufreq.c
607
static void cppc_cpufreq_put_cpu_data(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
609
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
613
policy->driver_data = NULL;
drivers/cpufreq/cppc_cpufreq.c
616
static int cppc_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
618
unsigned int cpu = policy->cpu;
drivers/cpufreq/cppc_cpufreq.c
629
policy->driver_data = cpu_data;
drivers/cpufreq/cppc_cpufreq.c
635
policy->min = cppc_perf_to_khz(caps, caps->lowest_nonlinear_perf);
drivers/cpufreq/cppc_cpufreq.c
636
policy->max = cppc_perf_to_khz(caps, policy->boost_enabled ?
drivers/cpufreq/cppc_cpufreq.c
644
policy->cpuinfo.min_freq = cppc_perf_to_khz(caps, caps->lowest_perf);
drivers/cpufreq/cppc_cpufreq.c
645
policy->cpuinfo.max_freq = policy->max;
drivers/cpufreq/cppc_cpufreq.c
647
policy->transition_delay_us = cppc_cpufreq_get_transition_delay_us(cpu);
drivers/cpufreq/cppc_cpufreq.c
648
policy->shared_type = cpu_data->shared_type;
drivers/cpufreq/cppc_cpufreq.c
650
switch (policy->shared_type) {
drivers/cpufreq/cppc_cpufreq.c
661
cpumask_copy(policy->cpus, cpu_data->shared_cpu_map);
drivers/cpufreq/cppc_cpufreq.c
665
policy->shared_type);
drivers/cpufreq/cppc_cpufreq.c
670
policy->fast_switch_possible = cppc_allow_fast_switch();
drivers/cpufreq/cppc_cpufreq.c
671
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/cppc_cpufreq.c
678
policy->boost_supported = true;
drivers/cpufreq/cppc_cpufreq.c
681
policy->cur = cppc_perf_to_khz(caps, caps->highest_perf);
drivers/cpufreq/cppc_cpufreq.c
691
cppc_cpufreq_cpu_fie_init(policy);
drivers/cpufreq/cppc_cpufreq.c
695
cppc_cpufreq_put_cpu_data(policy);
drivers/cpufreq/cppc_cpufreq.c
699
static void cppc_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cppc_cpufreq.c
701
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
703
unsigned int cpu = policy->cpu;
drivers/cpufreq/cppc_cpufreq.c
706
cppc_cpufreq_cpu_fie_exit(policy);
drivers/cpufreq/cppc_cpufreq.c
715
cppc_cpufreq_put_cpu_data(policy);
drivers/cpufreq/cppc_cpufreq.c
766
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cppc_cpufreq.c
772
if (!policy)
drivers/cpufreq/cppc_cpufreq.c
775
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
806
static int cppc_cpufreq_set_boost(struct cpufreq_policy *policy, int state)
drivers/cpufreq/cppc_cpufreq.c
808
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
813
policy->max = cppc_perf_to_khz(caps, caps->highest_perf);
drivers/cpufreq/cppc_cpufreq.c
815
policy->max = cppc_perf_to_khz(caps, caps->nominal_perf);
drivers/cpufreq/cppc_cpufreq.c
816
policy->cpuinfo.max_freq = policy->max;
drivers/cpufreq/cppc_cpufreq.c
818
ret = freq_qos_update_request(policy->max_freq_req, policy->max);
drivers/cpufreq/cppc_cpufreq.c
825
static ssize_t show_freqdomain_cpus(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cppc_cpufreq.c
827
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
832
static ssize_t show_auto_select(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cppc_cpufreq.c
837
ret = cppc_get_auto_sel(policy->cpu, &val);
drivers/cpufreq/cppc_cpufreq.c
849
static ssize_t store_auto_select(struct cpufreq_policy *policy,
drivers/cpufreq/cppc_cpufreq.c
859
ret = cppc_set_auto_sel(policy->cpu, val);
drivers/cpufreq/cppc_cpufreq.c
899
static ssize_t show_##_name(struct cpufreq_policy *policy, char *buf) \
drivers/cpufreq/cppc_cpufreq.c
901
return cppc_cpufreq_sysfs_show_u64(policy->cpu, _get_func, buf);\
drivers/cpufreq/cppc_cpufreq.c
903
static ssize_t store_##_name(struct cpufreq_policy *policy, \
drivers/cpufreq/cppc_cpufreq.c
906
return cppc_cpufreq_sysfs_store_u64(policy->cpu, _set_func, \
drivers/cpufreq/cpufreq-dt.c
109
cpumask_copy(policy->cpus, priv->cpus);
drivers/cpufreq/cpufreq-dt.c
110
policy->driver_data = priv;
drivers/cpufreq/cpufreq-dt.c
111
policy->clk = cpu_clk;
drivers/cpufreq/cpufreq-dt.c
112
policy->freq_table = priv->freq_table;
drivers/cpufreq/cpufreq-dt.c
113
policy->suspend_freq = dev_pm_opp_get_suspend_opp_freq(cpu_dev) / 1000;
drivers/cpufreq/cpufreq-dt.c
114
policy->cpuinfo.transition_latency = transition_latency;
drivers/cpufreq/cpufreq-dt.c
115
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/cpufreq-dt.c
120
static int cpufreq_online(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq-dt.c
126
static int cpufreq_offline(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq-dt.c
135
static void cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq-dt.c
137
clk_put(policy->clk);
drivers/cpufreq/cpufreq-dt.c
51
static int set_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/cpufreq-dt.c
53
struct private_data *priv = policy->driver_data;
drivers/cpufreq/cpufreq-dt.c
54
unsigned long freq = policy->freq_table[index].frequency;
drivers/cpufreq/cpufreq-dt.c
83
static int cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq-dt.c
91
priv = cpufreq_dt_find_data(policy->cpu);
drivers/cpufreq/cpufreq-dt.c
93
pr_err("failed to find data for cpu%d\n", policy->cpu);
drivers/cpufreq/cpufreq-dt.h
17
unsigned int (*get_intermediate)(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq-dt.h
19
int (*target_intermediate)(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq-dt.h
21
int (*suspend)(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq-dt.h
22
int (*resume)(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq-nforce2.c
252
static int nforce2_target(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq-nforce2.c
259
if ((target_freq > policy->max) || (target_freq < policy->min))
drivers/cpufreq/cpufreq-nforce2.c
264
freqs.old = nforce2_get(policy->cpu);
drivers/cpufreq/cpufreq-nforce2.c
273
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/cpufreq-nforce2.c
287
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/cpufreq-nforce2.c
296
static int nforce2_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/cpufreq-nforce2.c
300
fsb_pol_max = policy->max / (fid * 100);
drivers/cpufreq/cpufreq-nforce2.c
302
if (policy->min < (fsb_pol_max * fid * 100))
drivers/cpufreq/cpufreq-nforce2.c
303
policy->max = (fsb_pol_max + 1) * fid * 100;
drivers/cpufreq/cpufreq-nforce2.c
305
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/cpufreq-nforce2.c
309
static int nforce2_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq-nforce2.c
315
if (policy->cpu != 0)
drivers/cpufreq/cpufreq-nforce2.c
358
policy->min = policy->cpuinfo.min_freq = min_fsb * fid * 100;
drivers/cpufreq/cpufreq-nforce2.c
359
policy->max = policy->cpuinfo.max_freq = max_fsb * fid * 100;
drivers/cpufreq/cpufreq.c
1002
guard(cpufreq_policy_read)(policy);
drivers/cpufreq/cpufreq.c
1004
if (likely(!policy_is_inactive(policy)))
drivers/cpufreq/cpufreq.c
1005
return fattr->show(policy, buf);
drivers/cpufreq/cpufreq.c
1013
struct cpufreq_policy *policy = to_policy(kobj);
drivers/cpufreq/cpufreq.c
1019
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
1021
if (likely(!policy_is_inactive(policy)))
drivers/cpufreq/cpufreq.c
1022
return fattr->store(policy, buf, count);
drivers/cpufreq/cpufreq.c
1029
struct cpufreq_policy *policy = to_policy(kobj);
drivers/cpufreq/cpufreq.c
1031
complete(&policy->kobj_unregister);
drivers/cpufreq/cpufreq.c
1045
static void add_cpu_dev_symlink(struct cpufreq_policy *policy, unsigned int cpu,
drivers/cpufreq/cpufreq.c
1051
if (cpumask_test_and_set_cpu(cpu, policy->real_cpus))
drivers/cpufreq/cpufreq.c
1055
if (sysfs_create_link(&dev->kobj, &policy->kobj, "cpufreq"))
drivers/cpufreq/cpufreq.c
1059
static void remove_cpu_dev_symlink(struct cpufreq_policy *policy, int cpu,
drivers/cpufreq/cpufreq.c
1064
cpumask_clear_cpu(cpu, policy->real_cpus);
drivers/cpufreq/cpufreq.c
1067
static int cpufreq_add_dev_interface(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1073
if (policy->freq_table) {
drivers/cpufreq/cpufreq.c
1074
ret = sysfs_create_file(&policy->kobj,
drivers/cpufreq/cpufreq.c
1080
ret = sysfs_create_file(&policy->kobj,
drivers/cpufreq/cpufreq.c
1090
ret = sysfs_create_file(&policy->kobj, &((*drv_attr)->attr));
drivers/cpufreq/cpufreq.c
1096
ret = sysfs_create_file(&policy->kobj, &cpuinfo_cur_freq.attr);
drivers/cpufreq/cpufreq.c
1101
if (cpufreq_avg_freq_supported(policy)) {
drivers/cpufreq/cpufreq.c
1102
ret = sysfs_create_file(&policy->kobj, &cpuinfo_avg_freq.attr);
drivers/cpufreq/cpufreq.c
1108
ret = sysfs_create_file(&policy->kobj, &bios_limit.attr);
drivers/cpufreq/cpufreq.c
1114
ret = sysfs_create_file(&policy->kobj, &local_boost.attr);
drivers/cpufreq/cpufreq.c
1122
static int cpufreq_init_policy(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1130
if (policy->last_governor[0] != '\0')
drivers/cpufreq/cpufreq.c
1131
gov = get_governor(policy->last_governor);
drivers/cpufreq/cpufreq.c
1134
gov->name, policy->cpu);
drivers/cpufreq/cpufreq.c
1147
if (policy->last_policy) {
drivers/cpufreq/cpufreq.c
1148
pol = policy->last_policy;
drivers/cpufreq/cpufreq.c
1157
pol = policy->policy;
drivers/cpufreq/cpufreq.c
1164
ret = cpufreq_set_policy(policy, gov, pol);
drivers/cpufreq/cpufreq.c
1171
static int cpufreq_add_policy_cpu(struct cpufreq_policy *policy, unsigned int cpu)
drivers/cpufreq/cpufreq.c
1176
if (cpumask_test_cpu(cpu, policy->cpus))
drivers/cpufreq/cpufreq.c
1179
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
1182
cpufreq_stop_governor(policy);
drivers/cpufreq/cpufreq.c
1184
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/cpufreq.c
1187
ret = cpufreq_start_governor(policy);
drivers/cpufreq/cpufreq.c
1195
void refresh_frequency_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1197
if (!policy_is_inactive(policy)) {
drivers/cpufreq/cpufreq.c
1198
pr_debug("updating policy for CPU %u\n", policy->cpu);
drivers/cpufreq/cpufreq.c
1200
cpufreq_set_policy(policy, policy->governor, policy->policy);
drivers/cpufreq/cpufreq.c
1207
struct cpufreq_policy *policy =
drivers/cpufreq/cpufreq.c
1210
pr_debug("handle_update for cpu %u called\n", policy->cpu);
drivers/cpufreq/cpufreq.c
1212
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
1214
refresh_frequency_limits(policy);
drivers/cpufreq/cpufreq.c
1220
struct cpufreq_policy *policy = container_of(nb, struct cpufreq_policy, nb_min);
drivers/cpufreq/cpufreq.c
1222
schedule_work(&policy->update);
drivers/cpufreq/cpufreq.c
1229
struct cpufreq_policy *policy = container_of(nb, struct cpufreq_policy, nb_max);
drivers/cpufreq/cpufreq.c
1231
schedule_work(&policy->update);
drivers/cpufreq/cpufreq.c
1235
static void cpufreq_policy_put_kobj(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
124
struct kobject *get_governor_parent_kobj(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1240
scoped_guard(cpufreq_policy_write, policy) {
drivers/cpufreq/cpufreq.c
1241
cpufreq_stats_free_table(policy);
drivers/cpufreq/cpufreq.c
1242
kobj = &policy->kobj;
drivers/cpufreq/cpufreq.c
1243
cmp = &policy->kobj_unregister;
drivers/cpufreq/cpufreq.c
1259
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
1266
policy = kzalloc_obj(*policy);
drivers/cpufreq/cpufreq.c
1267
if (!policy)
drivers/cpufreq/cpufreq.c
127
return &policy->kobj;
drivers/cpufreq/cpufreq.c
1270
if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL))
drivers/cpufreq/cpufreq.c
1273
if (!zalloc_cpumask_var(&policy->related_cpus, GFP_KERNEL))
drivers/cpufreq/cpufreq.c
1276
if (!zalloc_cpumask_var(&policy->real_cpus, GFP_KERNEL))
drivers/cpufreq/cpufreq.c
1279
init_completion(&policy->kobj_unregister);
drivers/cpufreq/cpufreq.c
1280
ret = kobject_init_and_add(&policy->kobj, &ktype_cpufreq,
drivers/cpufreq/cpufreq.c
1289
kobject_put(&policy->kobj);
drivers/cpufreq/cpufreq.c
1293
init_rwsem(&policy->rwsem);
drivers/cpufreq/cpufreq.c
1295
freq_constraints_init(&policy->constraints);
drivers/cpufreq/cpufreq.c
1297
policy->nb_min.notifier_call = cpufreq_notifier_min;
drivers/cpufreq/cpufreq.c
1298
policy->nb_max.notifier_call = cpufreq_notifier_max;
drivers/cpufreq/cpufreq.c
1300
ret = freq_qos_add_notifier(&policy->constraints, FREQ_QOS_MIN,
drivers/cpufreq/cpufreq.c
1301
&policy->nb_min);
drivers/cpufreq/cpufreq.c
1308
ret = freq_qos_add_notifier(&policy->constraints, FREQ_QOS_MAX,
drivers/cpufreq/cpufreq.c
1309
&policy->nb_max);
drivers/cpufreq/cpufreq.c
1316
INIT_LIST_HEAD(&policy->policy_list);
drivers/cpufreq/cpufreq.c
1317
spin_lock_init(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
1318
init_waitqueue_head(&policy->transition_wait);
drivers/cpufreq/cpufreq.c
1319
INIT_WORK(&policy->update, handle_update);
drivers/cpufreq/cpufreq.c
1321
return policy;
drivers/cpufreq/cpufreq.c
1324
freq_qos_remove_notifier(&policy->constraints, FREQ_QOS_MIN,
drivers/cpufreq/cpufreq.c
1325
&policy->nb_min);
drivers/cpufreq/cpufreq.c
1327
cpufreq_policy_put_kobj(policy);
drivers/cpufreq/cpufreq.c
1329
free_cpumask_var(policy->real_cpus);
drivers/cpufreq/cpufreq.c
1331
free_cpumask_var(policy->related_cpus);
drivers/cpufreq/cpufreq.c
1333
free_cpumask_var(policy->cpus);
drivers/cpufreq/cpufreq.c
1335
kfree(policy);
drivers/cpufreq/cpufreq.c
1340
static void cpufreq_policy_free(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1349
if (unlikely(!policy_is_inactive(policy)))
drivers/cpufreq/cpufreq.c
1354
list_del(&policy->policy_list);
drivers/cpufreq/cpufreq.c
1356
for_each_cpu(cpu, policy->related_cpus)
drivers/cpufreq/cpufreq.c
1360
freq_qos_remove_notifier(&policy->constraints, FREQ_QOS_MAX,
drivers/cpufreq/cpufreq.c
1361
&policy->nb_max);
drivers/cpufreq/cpufreq.c
1362
freq_qos_remove_notifier(&policy->constraints, FREQ_QOS_MIN,
drivers/cpufreq/cpufreq.c
1363
&policy->nb_min);
drivers/cpufreq/cpufreq.c
1366
cancel_work_sync(&policy->update);
drivers/cpufreq/cpufreq.c
1368
if (policy->max_freq_req) {
drivers/cpufreq/cpufreq.c
1375
CPUFREQ_REMOVE_POLICY, policy);
drivers/cpufreq/cpufreq.c
1376
freq_qos_remove_request(policy->max_freq_req);
drivers/cpufreq/cpufreq.c
1379
freq_qos_remove_request(policy->min_freq_req);
drivers/cpufreq/cpufreq.c
1380
kfree(policy->min_freq_req);
drivers/cpufreq/cpufreq.c
1382
cpufreq_policy_put_kobj(policy);
drivers/cpufreq/cpufreq.c
1383
free_cpumask_var(policy->real_cpus);
drivers/cpufreq/cpufreq.c
1384
free_cpumask_var(policy->related_cpus);
drivers/cpufreq/cpufreq.c
1385
free_cpumask_var(policy->cpus);
drivers/cpufreq/cpufreq.c
1386
kfree(policy);
drivers/cpufreq/cpufreq.c
1389
static int cpufreq_policy_online(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
1396
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
1398
policy->cpu = cpu;
drivers/cpufreq/cpufreq.c
1399
policy->governor = NULL;
drivers/cpufreq/cpufreq.c
1403
cpumask_copy(policy->cpus, policy->related_cpus);
drivers/cpufreq/cpufreq.c
1405
ret = cpufreq_driver->online(policy);
drivers/cpufreq/cpufreq.c
1412
cpumask_copy(policy->cpus, cpumask_of(cpu));
drivers/cpufreq/cpufreq.c
1418
ret = cpufreq_driver->init(policy);
drivers/cpufreq/cpufreq.c
1430
ret = cpufreq_table_validate_and_sort(policy);
drivers/cpufreq/cpufreq.c
1435
cpumask_copy(policy->related_cpus, policy->cpus);
drivers/cpufreq/cpufreq.c
1442
cpumask_and(policy->cpus, policy->cpus, cpu_online_mask);
drivers/cpufreq/cpufreq.c
1445
for_each_cpu(j, policy->related_cpus) {
drivers/cpufreq/cpufreq.c
1446
per_cpu(cpufreq_cpu_data, j) = policy;
drivers/cpufreq/cpufreq.c
1447
add_cpu_dev_symlink(policy, j, get_cpu_device(j));
drivers/cpufreq/cpufreq.c
1450
policy->min_freq_req = kzalloc(2 * sizeof(*policy->min_freq_req),
drivers/cpufreq/cpufreq.c
1452
if (!policy->min_freq_req) {
drivers/cpufreq/cpufreq.c
1457
ret = freq_qos_add_request(&policy->constraints,
drivers/cpufreq/cpufreq.c
1458
policy->min_freq_req, FREQ_QOS_MIN,
drivers/cpufreq/cpufreq.c
1465
kfree(policy->min_freq_req);
drivers/cpufreq/cpufreq.c
1466
policy->min_freq_req = NULL;
drivers/cpufreq/cpufreq.c
1475
policy->max_freq_req = policy->min_freq_req + 1;
drivers/cpufreq/cpufreq.c
1477
ret = freq_qos_add_request(&policy->constraints,
drivers/cpufreq/cpufreq.c
1478
policy->max_freq_req, FREQ_QOS_MAX,
drivers/cpufreq/cpufreq.c
1481
policy->max_freq_req = NULL;
drivers/cpufreq/cpufreq.c
1486
CPUFREQ_CREATE_POLICY, policy);
drivers/cpufreq/cpufreq.c
1488
ret = freq_qos_update_request(policy->max_freq_req, policy->max);
drivers/cpufreq/cpufreq.c
1494
policy->cur = cpufreq_driver->get(policy->cpu);
drivers/cpufreq/cpufreq.c
1495
if (!policy->cur) {
drivers/cpufreq/cpufreq.c
1522
unsigned int old_freq = policy->cur;
drivers/cpufreq/cpufreq.c
1525
ret = cpufreq_frequency_table_get_index(policy, old_freq);
drivers/cpufreq/cpufreq.c
1527
ret = __cpufreq_driver_target(policy, old_freq - 1,
drivers/cpufreq/cpufreq.c
1537
__func__, policy->cpu, old_freq, policy->cur);
drivers/cpufreq/cpufreq.c
1542
ret = cpufreq_add_dev_interface(policy);
drivers/cpufreq/cpufreq.c
1546
cpufreq_stats_create_table(policy);
drivers/cpufreq/cpufreq.c
1549
list_add(&policy->policy_list, &cpufreq_policy_list);
drivers/cpufreq/cpufreq.c
1563
cpufreq_driver->register_em(policy);
drivers/cpufreq/cpufreq.c
1566
ret = cpufreq_init_policy(policy);
drivers/cpufreq/cpufreq.c
1576
for_each_cpu(j, policy->real_cpus)
drivers/cpufreq/cpufreq.c
1577
remove_cpu_dev_symlink(policy, j, get_cpu_device(j));
drivers/cpufreq/cpufreq.c
1581
cpufreq_driver->offline(policy);
drivers/cpufreq/cpufreq.c
1585
cpufreq_driver->exit(policy);
drivers/cpufreq/cpufreq.c
1588
cpumask_clear(policy->cpus);
drivers/cpufreq/cpufreq.c
1595
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
1602
policy = per_cpu(cpufreq_cpu_data, cpu);
drivers/cpufreq/cpufreq.c
1603
if (policy) {
drivers/cpufreq/cpufreq.c
1604
WARN_ON(!cpumask_test_cpu(cpu, policy->related_cpus));
drivers/cpufreq/cpufreq.c
1605
if (!policy_is_inactive(policy))
drivers/cpufreq/cpufreq.c
1606
return cpufreq_add_policy_cpu(policy, cpu);
drivers/cpufreq/cpufreq.c
1612
policy = cpufreq_policy_alloc(cpu);
drivers/cpufreq/cpufreq.c
1613
if (!policy)
drivers/cpufreq/cpufreq.c
1617
ret = cpufreq_policy_online(policy, cpu, new_policy);
drivers/cpufreq/cpufreq.c
1619
cpufreq_policy_free(policy);
drivers/cpufreq/cpufreq.c
1623
kobject_uevent(&policy->kobj, KOBJ_ADD);
drivers/cpufreq/cpufreq.c
1627
cpufreq_driver->ready(policy);
drivers/cpufreq/cpufreq.c
1631
policy->cdev = of_cpufreq_cooling_register(policy);
drivers/cpufreq/cpufreq.c
1638
if (cpufreq_driver->set_boost && policy->boost_supported &&
drivers/cpufreq/cpufreq.c
1640
ret = policy_set_boost(policy, cpufreq_boost_enabled());
drivers/cpufreq/cpufreq.c
1643
pr_info("%s: CPU%d: Cannot %s BOOST\n", __func__, policy->cpu,
drivers/cpufreq/cpufreq.c
1660
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
1673
policy = per_cpu(cpufreq_cpu_data, cpu);
drivers/cpufreq/cpufreq.c
1674
if (policy)
drivers/cpufreq/cpufreq.c
1675
add_cpu_dev_symlink(policy, cpu, dev);
drivers/cpufreq/cpufreq.c
1680
static void __cpufreq_offline(unsigned int cpu, struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1685
cpufreq_stop_governor(policy);
drivers/cpufreq/cpufreq.c
1687
cpumask_clear_cpu(cpu, policy->cpus);
drivers/cpufreq/cpufreq.c
1689
if (!policy_is_inactive(policy)) {
drivers/cpufreq/cpufreq.c
1691
if (cpu == policy->cpu)
drivers/cpufreq/cpufreq.c
1692
policy->cpu = cpumask_any(policy->cpus);
drivers/cpufreq/cpufreq.c
1696
ret = cpufreq_start_governor(policy);
drivers/cpufreq/cpufreq.c
1705
strscpy(policy->last_governor, policy->governor->name,
drivers/cpufreq/cpufreq.c
1707
cpufreq_exit_governor(policy);
drivers/cpufreq/cpufreq.c
1709
policy->last_policy = policy->policy;
drivers/cpufreq/cpufreq.c
1717
cpufreq_driver->offline(policy);
drivers/cpufreq/cpufreq.c
1722
cpufreq_driver->exit(policy);
drivers/cpufreq/cpufreq.c
1724
policy->freq_table = NULL;
drivers/cpufreq/cpufreq.c
1729
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
1733
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/cpufreq.c
1734
if (!policy) {
drivers/cpufreq/cpufreq.c
1739
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
1741
__cpufreq_offline(cpu, policy);
drivers/cpufreq/cpufreq.c
1754
struct cpufreq_policy *policy = per_cpu(cpufreq_cpu_data, cpu);
drivers/cpufreq/cpufreq.c
1756
if (!policy)
drivers/cpufreq/cpufreq.c
1759
scoped_guard(cpufreq_policy_write, policy) {
drivers/cpufreq/cpufreq.c
1761
__cpufreq_offline(cpu, policy);
drivers/cpufreq/cpufreq.c
1763
remove_cpu_dev_symlink(policy, cpu, dev);
drivers/cpufreq/cpufreq.c
1765
if (!cpumask_empty(policy->real_cpus))
drivers/cpufreq/cpufreq.c
1773
cpufreq_cooling_unregister(policy->cdev);
drivers/cpufreq/cpufreq.c
1774
policy->cdev = NULL;
drivers/cpufreq/cpufreq.c
1779
cpufreq_driver->exit(policy);
drivers/cpufreq/cpufreq.c
178
void cpufreq_generic_init(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
1782
cpufreq_policy_free(policy);
drivers/cpufreq/cpufreq.c
1793
static void cpufreq_out_of_sync(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
1799
policy->cur, new_freq);
drivers/cpufreq/cpufreq.c
1801
freqs.old = policy->cur;
drivers/cpufreq/cpufreq.c
1804
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/cpufreq.c
1805
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/cpufreq.c
1808
static unsigned int cpufreq_verify_current_freq(struct cpufreq_policy *policy, bool update)
drivers/cpufreq/cpufreq.c
1815
new_freq = cpufreq_driver->get(policy->cpu);
drivers/cpufreq/cpufreq.c
182
policy->freq_table = table;
drivers/cpufreq/cpufreq.c
1823
if (policy->fast_switch_enabled || !has_target())
drivers/cpufreq/cpufreq.c
1826
if (policy->cur != new_freq) {
drivers/cpufreq/cpufreq.c
183
policy->cpuinfo.transition_latency = transition_latency;
drivers/cpufreq/cpufreq.c
1834
if (abs(policy->cur - new_freq) < KHZ_PER_MHZ)
drivers/cpufreq/cpufreq.c
1835
return policy->cur;
drivers/cpufreq/cpufreq.c
1837
cpufreq_out_of_sync(policy, new_freq);
drivers/cpufreq/cpufreq.c
1839
schedule_work(&policy->update);
drivers/cpufreq/cpufreq.c
1868
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
1869
if (policy)
drivers/cpufreq/cpufreq.c
1870
return policy->cur;
drivers/cpufreq/cpufreq.c
1884
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
1885
if (policy)
drivers/cpufreq/cpufreq.c
1886
return policy->max;
drivers/cpufreq/cpufreq.c
189
cpumask_setall(policy->cpus);
drivers/cpufreq/cpufreq.c
1900
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
1901
if (policy)
drivers/cpufreq/cpufreq.c
1902
return policy->cpuinfo.max_freq;
drivers/cpufreq/cpufreq.c
1908
static unsigned int __cpufreq_get(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1910
if (unlikely(policy_is_inactive(policy)))
drivers/cpufreq/cpufreq.c
1913
return cpufreq_verify_current_freq(policy, true);
drivers/cpufreq/cpufreq.c
1924
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
1925
if (!policy)
drivers/cpufreq/cpufreq.c
1928
guard(cpufreq_policy_read)(policy);
drivers/cpufreq/cpufreq.c
1930
return __cpufreq_get(policy);
drivers/cpufreq/cpufreq.c
1945
int cpufreq_generic_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
1949
if (!policy->suspend_freq) {
drivers/cpufreq/cpufreq.c
195
struct cpufreq_policy *policy = per_cpu(cpufreq_cpu_data, cpu);
drivers/cpufreq/cpufreq.c
1955
policy->suspend_freq);
drivers/cpufreq/cpufreq.c
1957
ret = __cpufreq_driver_target(policy, policy->suspend_freq,
drivers/cpufreq/cpufreq.c
1961
__func__, policy->suspend_freq, ret);
drivers/cpufreq/cpufreq.c
197
return policy && cpumask_test_cpu(cpu, policy->cpus) ? policy : NULL;
drivers/cpufreq/cpufreq.c
1977
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
1987
for_each_active_policy(policy) {
drivers/cpufreq/cpufreq.c
1989
scoped_guard(cpufreq_policy_write, policy) {
drivers/cpufreq/cpufreq.c
1990
cpufreq_stop_governor(policy);
drivers/cpufreq/cpufreq.c
1994
if (cpufreq_driver->suspend && cpufreq_driver->suspend(policy))
drivers/cpufreq/cpufreq.c
2011
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
2027
for_each_active_policy(policy) {
drivers/cpufreq/cpufreq.c
2028
if (cpufreq_driver->resume && cpufreq_driver->resume(policy)) {
drivers/cpufreq/cpufreq.c
2032
scoped_guard(cpufreq_policy_write, policy) {
drivers/cpufreq/cpufreq.c
2033
ret = cpufreq_start_governor(policy);
drivers/cpufreq/cpufreq.c
2038
__func__, policy->cpu);
drivers/cpufreq/cpufreq.c
209
struct cpufreq_policy *policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/cpufreq.c
211
if (!policy || IS_ERR(policy->clk)) {
drivers/cpufreq/cpufreq.c
213
__func__, policy ? "clk" : "policy", cpu);
drivers/cpufreq/cpufreq.c
217
return clk_get_rate(policy->clk) / 1000;
drivers/cpufreq/cpufreq.c
2203
unsigned int cpufreq_driver_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
2209
target_freq = clamp_val(target_freq, policy->min, policy->max);
drivers/cpufreq/cpufreq.c
2210
freq = cpufreq_driver->fast_switch(policy, target_freq);
drivers/cpufreq/cpufreq.c
2215
policy->cur = freq;
drivers/cpufreq/cpufreq.c
2216
arch_set_freq_scale(policy->related_cpus, freq,
drivers/cpufreq/cpufreq.c
2217
arch_scale_freq_ref(policy->cpu));
drivers/cpufreq/cpufreq.c
2218
cpufreq_stats_record_transition(policy, freq);
drivers/cpufreq/cpufreq.c
2221
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/cpufreq.c
2270
static int __target_intermediate(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
2275
freqs->new = cpufreq_driver->get_intermediate(policy, index);
drivers/cpufreq/cpufreq.c
2282
__func__, policy->cpu, freqs->old, freqs->new);
drivers/cpufreq/cpufreq.c
2284
cpufreq_freq_transition_begin(policy, freqs);
drivers/cpufreq/cpufreq.c
2285
ret = cpufreq_driver->target_intermediate(policy, index);
drivers/cpufreq/cpufreq.c
2286
cpufreq_freq_transition_end(policy, freqs, ret);
drivers/cpufreq/cpufreq.c
2295
static int __target_index(struct cpufreq_policy *policy, int index)
drivers/cpufreq/cpufreq.c
2297
struct cpufreq_freqs freqs = {.old = policy->cur, .flags = 0};
drivers/cpufreq/cpufreq.c
2299
unsigned int newfreq = policy->freq_table[index].frequency;
drivers/cpufreq/cpufreq.c
2303
if (newfreq == policy->cur)
drivers/cpufreq/cpufreq.c
2307
restore_freq = policy->cur;
drivers/cpufreq/cpufreq.c
2313
retval = __target_intermediate(policy, &freqs, index);
drivers/cpufreq/cpufreq.c
2325
__func__, policy->cpu, freqs.old, freqs.new);
drivers/cpufreq/cpufreq.c
2327
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/cpufreq.c
2330
retval = cpufreq_driver->target_index(policy, index);
drivers/cpufreq/cpufreq.c
2336
cpufreq_freq_transition_end(policy, &freqs, retval);
drivers/cpufreq/cpufreq.c
234
struct cpufreq_policy *policy = NULL;
drivers/cpufreq/cpufreq.c
2347
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/cpufreq.c
2348
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/cpufreq.c
2355
int __cpufreq_driver_target(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
2364
target_freq = __resolve_freq(policy, target_freq, policy->min,
drivers/cpufreq/cpufreq.c
2365
policy->max, relation);
drivers/cpufreq/cpufreq.c
2368
policy->cpu, target_freq, relation, old_target_freq);
drivers/cpufreq/cpufreq.c
2376
if (target_freq == policy->cur &&
drivers/cpufreq/cpufreq.c
2385
if (!policy->efficiencies_available)
drivers/cpufreq/cpufreq.c
2388
return cpufreq_driver->target(policy, target_freq, relation);
drivers/cpufreq/cpufreq.c
2394
return __target_index(policy, policy->cached_resolved_idx);
drivers/cpufreq/cpufreq.c
2398
int cpufreq_driver_target(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
2402
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
2404
return __cpufreq_driver_target(policy, target_freq, relation);
drivers/cpufreq/cpufreq.c
2413
static int cpufreq_init_governor(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2424
if (!policy->governor)
drivers/cpufreq/cpufreq.c
2428
if (policy->governor->flags & CPUFREQ_GOV_DYNAMIC_SWITCHING &&
drivers/cpufreq/cpufreq.c
2434
policy->governor->name, gov->name);
drivers/cpufreq/cpufreq.c
2435
policy->governor = gov;
drivers/cpufreq/cpufreq.c
2441
if (!try_module_get(policy->governor->owner))
drivers/cpufreq/cpufreq.c
2444
pr_debug("%s: for CPU %u\n", __func__, policy->cpu);
drivers/cpufreq/cpufreq.c
2446
if (policy->governor->init) {
drivers/cpufreq/cpufreq.c
2447
ret = policy->governor->init(policy);
drivers/cpufreq/cpufreq.c
2449
module_put(policy->governor->owner);
drivers/cpufreq/cpufreq.c
245
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/cpufreq.c
2454
policy->strict_target = !!(policy->governor->flags & CPUFREQ_GOV_STRICT_TARGET);
drivers/cpufreq/cpufreq.c
2459
static void cpufreq_exit_governor(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
246
if (policy)
drivers/cpufreq/cpufreq.c
2461
if (cpufreq_suspended || !policy->governor)
drivers/cpufreq/cpufreq.c
2464
pr_debug("%s: for CPU %u\n", __func__, policy->cpu);
drivers/cpufreq/cpufreq.c
2466
if (policy->governor->exit)
drivers/cpufreq/cpufreq.c
2467
policy->governor->exit(policy);
drivers/cpufreq/cpufreq.c
2469
module_put(policy->governor->owner);
drivers/cpufreq/cpufreq.c
247
kobject_get(&policy->kobj);
drivers/cpufreq/cpufreq.c
2472
int cpufreq_start_governor(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2479
if (!policy->governor)
drivers/cpufreq/cpufreq.c
2482
pr_debug("%s: for CPU %u\n", __func__, policy->cpu);
drivers/cpufreq/cpufreq.c
2484
cpufreq_verify_current_freq(policy, false);
drivers/cpufreq/cpufreq.c
2486
if (policy->governor->start) {
drivers/cpufreq/cpufreq.c
2487
ret = policy->governor->start(policy);
drivers/cpufreq/cpufreq.c
2492
if (policy->governor->limits)
drivers/cpufreq/cpufreq.c
2493
policy->governor->limits(policy);
drivers/cpufreq/cpufreq.c
2498
void cpufreq_stop_governor(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2500
if (cpufreq_suspended || !policy->governor)
drivers/cpufreq/cpufreq.c
2503
pr_debug("%s: for CPU %u\n", __func__, policy->cpu);
drivers/cpufreq/cpufreq.c
2505
if (policy->governor->stop)
drivers/cpufreq/cpufreq.c
2506
policy->governor->stop(policy);
drivers/cpufreq/cpufreq.c
2509
static void cpufreq_governor_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2511
if (cpufreq_suspended || !policy->governor)
drivers/cpufreq/cpufreq.c
2514
pr_debug("%s: for CPU %u\n", __func__, policy->cpu);
drivers/cpufreq/cpufreq.c
2516
if (policy->governor->limits)
drivers/cpufreq/cpufreq.c
2517
policy->governor->limits(policy);
drivers/cpufreq/cpufreq.c
252
return policy;
drivers/cpufreq/cpufreq.c
2545
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
2556
for_each_inactive_policy(policy) {
drivers/cpufreq/cpufreq.c
2557
if (!strcmp(policy->last_governor, governor->name)) {
drivers/cpufreq/cpufreq.c
2558
policy->governor = NULL;
drivers/cpufreq/cpufreq.c
2559
policy->last_governor[0] = '\0';
drivers/cpufreq/cpufreq.c
2583
static void cpufreq_update_pressure(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2589
cpu = cpumask_first(policy->related_cpus);
drivers/cpufreq/cpufreq.c
2591
capped_freq = policy->max;
drivers/cpufreq/cpufreq.c
260
void cpufreq_cpu_put(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2605
for_each_cpu(cpu, policy->related_cpus)
drivers/cpufreq/cpufreq.c
262
kobject_put(&policy->kobj);
drivers/cpufreq/cpufreq.c
2624
static int cpufreq_set_policy(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
2632
memcpy(&new_data.cpuinfo, &policy->cpuinfo, sizeof(policy->cpuinfo));
drivers/cpufreq/cpufreq.c
2633
new_data.freq_table = policy->freq_table;
drivers/cpufreq/cpufreq.c
2634
new_data.cpu = policy->cpu;
drivers/cpufreq/cpufreq.c
2639
new_data.min = freq_qos_read_value(&policy->constraints, FREQ_QOS_MIN);
drivers/cpufreq/cpufreq.c
2640
new_data.max = freq_qos_read_value(&policy->constraints, FREQ_QOS_MAX);
drivers/cpufreq/cpufreq.c
2662
WRITE_ONCE(policy->max, __resolve_freq(policy, new_data.max,
drivers/cpufreq/cpufreq.c
2665
new_data.min = __resolve_freq(policy, new_data.min, new_data.min,
drivers/cpufreq/cpufreq.c
2667
WRITE_ONCE(policy->min, new_data.min > policy->max ? policy->max : new_data.min);
drivers/cpufreq/cpufreq.c
2669
trace_cpu_frequency_limits(policy);
drivers/cpufreq/cpufreq.c
2671
cpufreq_update_pressure(policy);
drivers/cpufreq/cpufreq.c
2673
policy->cached_target_freq = UINT_MAX;
drivers/cpufreq/cpufreq.c
2676
policy->min, policy->max);
drivers/cpufreq/cpufreq.c
2679
policy->policy = new_pol;
drivers/cpufreq/cpufreq.c
2681
return cpufreq_driver->setpolicy(policy);
drivers/cpufreq/cpufreq.c
2684
if (new_gov == policy->governor) {
drivers/cpufreq/cpufreq.c
2686
cpufreq_governor_limits(policy);
drivers/cpufreq/cpufreq.c
2693
old_gov = policy->governor;
drivers/cpufreq/cpufreq.c
2696
cpufreq_stop_governor(policy);
drivers/cpufreq/cpufreq.c
2697
cpufreq_exit_governor(policy);
drivers/cpufreq/cpufreq.c
2701
policy->governor = new_gov;
drivers/cpufreq/cpufreq.c
2702
ret = cpufreq_init_governor(policy);
drivers/cpufreq/cpufreq.c
2704
ret = cpufreq_start_governor(policy);
drivers/cpufreq/cpufreq.c
2709
cpufreq_exit_governor(policy);
drivers/cpufreq/cpufreq.c
2713
pr_debug("starting governor %s failed\n", policy->governor->name);
drivers/cpufreq/cpufreq.c
2715
policy->governor = old_gov;
drivers/cpufreq/cpufreq.c
2716
if (cpufreq_init_governor(policy)) {
drivers/cpufreq/cpufreq.c
2717
policy->governor = NULL;
drivers/cpufreq/cpufreq.c
2718
} else if (cpufreq_start_governor(policy)) {
drivers/cpufreq/cpufreq.c
2719
cpufreq_exit_governor(policy);
drivers/cpufreq/cpufreq.c
2720
policy->governor = NULL;
drivers/cpufreq/cpufreq.c
2727
static void cpufreq_policy_refresh(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
2729
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/cpufreq.c
2736
(cpufreq_suspended || WARN_ON(!cpufreq_verify_current_freq(policy, false))))
drivers/cpufreq/cpufreq.c
2739
refresh_frequency_limits(policy);
drivers/cpufreq/cpufreq.c
2753
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
2754
if (!policy)
drivers/cpufreq/cpufreq.c
2757
cpufreq_policy_refresh(policy);
drivers/cpufreq/cpufreq.c
2770
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
2771
if (!policy)
drivers/cpufreq/cpufreq.c
2775
cpufreq_driver->update_limits(policy);
drivers/cpufreq/cpufreq.c
2777
cpufreq_policy_refresh(policy);
drivers/cpufreq/cpufreq.c
2784
int cpufreq_boost_set_sw(struct cpufreq_policy *policy, int state)
drivers/cpufreq/cpufreq.c
2788
if (!policy->freq_table)
drivers/cpufreq/cpufreq.c
2791
ret = cpufreq_frequency_table_cpuinfo(policy);
drivers/cpufreq/cpufreq.c
2797
ret = freq_qos_update_request(policy->max_freq_req, policy->max);
drivers/cpufreq/cpufreq.c
2807
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq.c
2821
for_each_active_policy(policy) {
drivers/cpufreq/cpufreq.c
2822
if (!policy->boost_supported)
drivers/cpufreq/cpufreq.c
2825
ret = policy_set_boost(policy, state);
drivers/cpufreq/cpufreq.c
3054
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/cpufreq.c
3055
if (!policy) {
drivers/cpufreq/cpufreq.c
3060
return sugov_is_governor(policy);
drivers/cpufreq/cpufreq.c
314
static void cpufreq_notify_transition(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
325
freqs->policy = policy;
drivers/cpufreq/cpufreq.c
337
if (policy->cur && policy->cur != freqs->old) {
drivers/cpufreq/cpufreq.c
339
freqs->old, policy->cur);
drivers/cpufreq/cpufreq.c
340
freqs->old = policy->cur;
drivers/cpufreq/cpufreq.c
352
cpumask_pr_args(policy->cpus));
drivers/cpufreq/cpufreq.c
354
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/cpufreq.c
360
cpufreq_stats_record_transition(policy, freqs->new);
drivers/cpufreq/cpufreq.c
361
policy->cur = freqs->new;
drivers/cpufreq/cpufreq.c
366
static void cpufreq_notify_post_transition(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
369
cpufreq_notify_transition(policy, freqs, CPUFREQ_POSTCHANGE);
drivers/cpufreq/cpufreq.c
374
cpufreq_notify_transition(policy, freqs, CPUFREQ_PRECHANGE);
drivers/cpufreq/cpufreq.c
375
cpufreq_notify_transition(policy, freqs, CPUFREQ_POSTCHANGE);
drivers/cpufreq/cpufreq.c
378
void cpufreq_freq_transition_begin(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
391
&& current == policy->transition_task);
drivers/cpufreq/cpufreq.c
394
wait_event(policy->transition_wait, !policy->transition_ongoing);
drivers/cpufreq/cpufreq.c
396
spin_lock(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
398
if (unlikely(policy->transition_ongoing)) {
drivers/cpufreq/cpufreq.c
399
spin_unlock(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
403
policy->transition_ongoing = true;
drivers/cpufreq/cpufreq.c
404
policy->transition_task = current;
drivers/cpufreq/cpufreq.c
406
spin_unlock(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
408
cpufreq_notify_transition(policy, freqs, CPUFREQ_PRECHANGE);
drivers/cpufreq/cpufreq.c
412
void cpufreq_freq_transition_end(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
415
if (WARN_ON(!policy->transition_ongoing))
drivers/cpufreq/cpufreq.c
418
cpufreq_notify_post_transition(policy, freqs, transition_failed);
drivers/cpufreq/cpufreq.c
420
arch_set_freq_scale(policy->related_cpus,
drivers/cpufreq/cpufreq.c
421
policy->cur,
drivers/cpufreq/cpufreq.c
422
arch_scale_freq_ref(policy->cpu));
drivers/cpufreq/cpufreq.c
424
spin_lock(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
425
policy->transition_ongoing = false;
drivers/cpufreq/cpufreq.c
426
policy->transition_task = NULL;
drivers/cpufreq/cpufreq.c
427
spin_unlock(&policy->transition_lock);
drivers/cpufreq/cpufreq.c
429
wake_up(&policy->transition_wait);
drivers/cpufreq/cpufreq.c
465
void cpufreq_enable_fast_switch(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
467
lockdep_assert_held(&policy->rwsem);
drivers/cpufreq/cpufreq.c
469
if (!policy->fast_switch_possible)
drivers/cpufreq/cpufreq.c
475
policy->fast_switch_enabled = true;
drivers/cpufreq/cpufreq.c
478
policy->cpu);
drivers/cpufreq/cpufreq.c
489
void cpufreq_disable_fast_switch(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
492
if (policy->fast_switch_enabled) {
drivers/cpufreq/cpufreq.c
493
policy->fast_switch_enabled = false;
drivers/cpufreq/cpufreq.c
501
static unsigned int __resolve_freq(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
510
if (!policy->freq_table)
drivers/cpufreq/cpufreq.c
513
idx = cpufreq_frequency_table_target(policy, target_freq, min, max, relation);
drivers/cpufreq/cpufreq.c
514
policy->cached_resolved_idx = idx;
drivers/cpufreq/cpufreq.c
515
policy->cached_target_freq = target_freq;
drivers/cpufreq/cpufreq.c
516
return policy->freq_table[idx].frequency;
drivers/cpufreq/cpufreq.c
530
unsigned int cpufreq_driver_resolve_freq(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
533
unsigned int min = READ_ONCE(policy->min);
drivers/cpufreq/cpufreq.c
534
unsigned int max = READ_ONCE(policy->max);
drivers/cpufreq/cpufreq.c
547
return __resolve_freq(policy, target_freq, min, max, CPUFREQ_RELATION_LE);
drivers/cpufreq/cpufreq.c
551
unsigned int cpufreq_policy_transition_delay_us(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
555
if (policy->transition_delay_us)
drivers/cpufreq/cpufreq.c
556
return policy->transition_delay_us;
drivers/cpufreq/cpufreq.c
558
latency = policy->cpuinfo.transition_latency / NSEC_PER_USEC;
drivers/cpufreq/cpufreq.c
597
static ssize_t show_local_boost(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
599
return sysfs_emit(buf, "%d\n", policy->boost_enabled);
drivers/cpufreq/cpufreq.c
602
static int policy_set_boost(struct cpufreq_policy *policy, bool enable)
drivers/cpufreq/cpufreq.c
606
if (policy->boost_enabled == enable)
drivers/cpufreq/cpufreq.c
609
policy->boost_enabled = enable;
drivers/cpufreq/cpufreq.c
611
ret = cpufreq_driver->set_boost(policy, enable);
drivers/cpufreq/cpufreq.c
613
policy->boost_enabled = !policy->boost_enabled;
drivers/cpufreq/cpufreq.c
618
static ssize_t store_local_boost(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
630
if (!policy->boost_supported)
drivers/cpufreq/cpufreq.c
633
ret = policy_set_boost(policy, enable);
drivers/cpufreq/cpufreq.c
710
(struct cpufreq_policy *policy, char *buf) \
drivers/cpufreq/cpufreq.c
712
return sysfs_emit(buf, "%u\n", policy->object); \
drivers/cpufreq/cpufreq.c
726
static inline bool cpufreq_avg_freq_supported(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq.c
728
return arch_freq_get_on_cpu(policy->cpu) != -EOPNOTSUPP;
drivers/cpufreq/cpufreq.c
731
static ssize_t show_scaling_cur_freq(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
737
? arch_freq_get_on_cpu(policy->cpu)
drivers/cpufreq/cpufreq.c
743
ret = sysfs_emit(buf, "%u\n", cpufreq_driver->get(policy->cpu));
drivers/cpufreq/cpufreq.c
745
ret = sysfs_emit(buf, "%u\n", policy->cur);
drivers/cpufreq/cpufreq.c
754
(struct cpufreq_policy *policy, const char *buf, size_t count) \
drivers/cpufreq/cpufreq.c
763
ret = freq_qos_update_request(policy->object##_freq_req, val);\
drivers/cpufreq/cpufreq.c
773
static ssize_t show_cpuinfo_cur_freq(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
776
unsigned int cur_freq = __cpufreq_get(policy);
drivers/cpufreq/cpufreq.c
787
static ssize_t show_cpuinfo_avg_freq(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
790
int avg_freq = arch_freq_get_on_cpu(policy->cpu);
drivers/cpufreq/cpufreq.c
800
static ssize_t show_scaling_governor(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
802
if (policy->policy == CPUFREQ_POLICY_POWERSAVE)
drivers/cpufreq/cpufreq.c
804
else if (policy->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/cpufreq.c
806
else if (policy->governor)
drivers/cpufreq/cpufreq.c
807
return sysfs_emit(buf, "%s\n", policy->governor->name);
drivers/cpufreq/cpufreq.c
814
static ssize_t store_scaling_governor(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
83
static unsigned int __cpufreq_get(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq.c
831
ret = cpufreq_set_policy(policy, NULL, new_pol);
drivers/cpufreq/cpufreq.c
839
ret = cpufreq_set_policy(policy, new_gov,
drivers/cpufreq/cpufreq.c
84
static int cpufreq_init_governor(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq.c
85
static void cpufreq_exit_governor(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq.c
851
static ssize_t show_scaling_driver(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
859
static ssize_t show_scaling_available_governors(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
86
static void cpufreq_governor_limits(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq.c
87
static int cpufreq_set_policy(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
906
static ssize_t show_related_cpus(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
908
return cpufreq_show_cpus(policy->related_cpus, buf);
drivers/cpufreq/cpufreq.c
914
static ssize_t show_affected_cpus(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
916
return cpufreq_show_cpus(policy->cpus, buf);
drivers/cpufreq/cpufreq.c
919
static ssize_t store_scaling_setspeed(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq.c
925
if (!policy->governor || !policy->governor->store_setspeed)
drivers/cpufreq/cpufreq.c
932
policy->governor->store_setspeed(policy, freq);
drivers/cpufreq/cpufreq.c
937
static ssize_t show_scaling_setspeed(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
939
if (!policy->governor || !policy->governor->show_setspeed)
drivers/cpufreq/cpufreq.c
942
return policy->governor->show_setspeed(policy, buf);
drivers/cpufreq/cpufreq.c
948
static ssize_t show_bios_limit(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq.c
952
ret = cpufreq_driver->bios_limit(policy->cpu, &limit);
drivers/cpufreq/cpufreq.c
955
return sysfs_emit(buf, "%u\n", policy->cpuinfo.max_freq);
drivers/cpufreq/cpufreq.c
996
struct cpufreq_policy *policy = to_policy(kobj);
drivers/cpufreq/cpufreq_conservative.c
107
if (requested_freq == policy->max)
drivers/cpufreq/cpufreq_conservative.c
111
if (requested_freq > policy->max)
drivers/cpufreq/cpufreq_conservative.c
112
requested_freq = policy->max;
drivers/cpufreq/cpufreq_conservative.c
114
__cpufreq_driver_target(policy, requested_freq,
drivers/cpufreq/cpufreq_conservative.c
130
if (requested_freq == policy->min)
drivers/cpufreq/cpufreq_conservative.c
136
requested_freq = policy->min;
drivers/cpufreq/cpufreq_conservative.c
138
__cpufreq_driver_target(policy, requested_freq,
drivers/cpufreq/cpufreq_conservative.c
308
static void cs_start(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_conservative.c
310
struct cs_policy_dbs_info *dbs_info = to_dbs_info(policy->governor_data);
drivers/cpufreq/cpufreq_conservative.c
313
dbs_info->requested_freq = policy->cur;
drivers/cpufreq/cpufreq_conservative.c
316
static void cs_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_conservative.c
318
struct cs_policy_dbs_info *dbs_info = to_dbs_info(policy->governor_data);
drivers/cpufreq/cpufreq_conservative.c
324
dbs_info->requested_freq = policy->cur;
drivers/cpufreq/cpufreq_conservative.c
38
struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_conservative.c
40
unsigned int freq_step = (cs_tuners->freq_step * policy->max) / 100;
drivers/cpufreq/cpufreq_conservative.c
58
static unsigned int cs_dbs_update(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_conservative.c
60
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_conservative.c
65
unsigned int load = dbs_update(policy);
drivers/cpufreq/cpufreq_conservative.c
80
if (requested_freq > policy->max || requested_freq < policy->min) {
drivers/cpufreq/cpufreq_conservative.c
81
requested_freq = policy->cur;
drivers/cpufreq/cpufreq_conservative.c
85
freq_step = get_freq_step(cs_tuners, policy);
drivers/cpufreq/cpufreq_conservative.c
94
if (requested_freq > policy->min + freq_steps)
drivers/cpufreq/cpufreq_conservative.c
97
requested_freq = policy->min;
drivers/cpufreq/cpufreq_governor.c
102
for_each_cpu(j, policy_dbs->policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
114
unsigned int dbs_update(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
116
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_governor.c
137
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
236
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq_governor.c
240
policy = policy_dbs->policy;
drivers/cpufreq/cpufreq_governor.c
241
gov = dbs_governor_of(policy);
drivers/cpufreq/cpufreq_governor.c
248
gov_update_sample_delay(policy_dbs, gov->gov_dbs_update(policy));
drivers/cpufreq/cpufreq_governor.c
277
if (!cpufreq_this_cpu_can_update(policy_dbs->policy))
drivers/cpufreq/cpufreq_governor.c
326
struct cpufreq_policy *policy = policy_dbs->policy;
drivers/cpufreq/cpufreq_governor.c
332
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
340
static inline void gov_clear_update_util(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
344
for_each_cpu(i, policy->cpus)
drivers/cpufreq/cpufreq_governor.c
350
static struct policy_dbs_info *alloc_policy_dbs_info(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq_governor.c
361
policy_dbs->policy = policy;
drivers/cpufreq/cpufreq_governor.c
368
for_each_cpu(j, policy->related_cpus) {
drivers/cpufreq/cpufreq_governor.c
383
for_each_cpu(j, policy_dbs->policy->related_cpus) {
drivers/cpufreq/cpufreq_governor.c
401
int cpufreq_dbs_governor_init(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
403
struct dbs_governor *gov = dbs_governor_of(policy);
drivers/cpufreq/cpufreq_governor.c
409
if (policy->governor_data)
drivers/cpufreq/cpufreq_governor.c
412
policy_dbs = alloc_policy_dbs_info(policy, gov);
drivers/cpufreq/cpufreq_governor.c
426
policy->governor_data = policy_dbs;
drivers/cpufreq/cpufreq_governor.c
452
cpufreq_policy_transition_delay_us(policy));
drivers/cpufreq/cpufreq_governor.c
458
policy->governor_data = policy_dbs;
drivers/cpufreq/cpufreq_governor.c
463
get_governor_parent_kobj(policy),
drivers/cpufreq/cpufreq_governor.c
471
policy->governor_data = NULL;
drivers/cpufreq/cpufreq_governor.c
491
void cpufreq_dbs_governor_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
493
struct dbs_governor *gov = dbs_governor_of(policy);
drivers/cpufreq/cpufreq_governor.c
494
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_governor.c
503
policy->governor_data = NULL;
drivers/cpufreq/cpufreq_governor.c
514
int cpufreq_dbs_governor_start(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
516
struct dbs_governor *gov = dbs_governor_of(policy);
drivers/cpufreq/cpufreq_governor.c
517
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_governor.c
522
if (!policy->cur)
drivers/cpufreq/cpufreq_governor.c
525
policy_dbs->is_shared = policy_is_shared(policy);
drivers/cpufreq/cpufreq_governor.c
532
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
545
gov->start(policy);
drivers/cpufreq/cpufreq_governor.c
552
void cpufreq_dbs_governor_stop(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
554
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_governor.c
556
gov_clear_update_util(policy_dbs->policy);
drivers/cpufreq/cpufreq_governor.c
564
void cpufreq_dbs_governor_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.c
566
struct dbs_governor *gov = dbs_governor_of(policy);
drivers/cpufreq/cpufreq_governor.c
571
policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_governor.c
576
cpufreq_policy_apply_limits(policy);
drivers/cpufreq/cpufreq_governor.c
579
gov->limits(policy);
drivers/cpufreq/cpufreq_governor.h
135
unsigned int (*gov_dbs_update)(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
140
void (*start)(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
141
void (*limits)(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
144
static inline struct dbs_governor *dbs_governor_of(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_governor.h
146
return container_of(policy->governor, struct dbs_governor, gov);
drivers/cpufreq/cpufreq_governor.h
150
int cpufreq_dbs_governor_init(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
151
void cpufreq_dbs_governor_exit(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
152
int cpufreq_dbs_governor_start(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
153
void cpufreq_dbs_governor_stop(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
154
void cpufreq_dbs_governor_limits(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
170
unsigned int (*powersave_bias_target)(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq_governor.h
174
unsigned int dbs_update(struct cpufreq_policy *policy);
drivers/cpufreq/cpufreq_governor.h
79
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq_ondemand.c
101
else if (policy->cur == policy->max)
drivers/cpufreq/cpufreq_ondemand.c
104
__cpufreq_driver_target(policy, freq, od_tuners->powersave_bias ?
drivers/cpufreq/cpufreq_ondemand.c
113
static void od_update(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_ondemand.c
115
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_ondemand.c
119
unsigned int load = dbs_update(policy);
drivers/cpufreq/cpufreq_ondemand.c
126
if (policy->cur < policy->max)
drivers/cpufreq/cpufreq_ondemand.c
128
dbs_freq_increase(policy, policy->max);
drivers/cpufreq/cpufreq_ondemand.c
133
min_f = policy->cpuinfo.min_freq;
drivers/cpufreq/cpufreq_ondemand.c
134
max_f = policy->cpuinfo.max_freq;
drivers/cpufreq/cpufreq_ondemand.c
141
freq_next = od_ops.powersave_bias_target(policy,
drivers/cpufreq/cpufreq_ondemand.c
145
__cpufreq_driver_target(policy, freq_next, CPUFREQ_RELATION_CE);
drivers/cpufreq/cpufreq_ondemand.c
149
static unsigned int od_dbs_update(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_ondemand.c
151
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_ondemand.c
163
__cpufreq_driver_target(policy, dbs_info->freq_lo,
drivers/cpufreq/cpufreq_ondemand.c
168
od_update(policy);
drivers/cpufreq/cpufreq_ondemand.c
289
ondemand_powersave_bias_init(policy_dbs->policy);
drivers/cpufreq/cpufreq_ondemand.c
363
static void od_start(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_ondemand.c
365
struct od_policy_dbs_info *dbs_info = to_dbs_info(policy->governor_data);
drivers/cpufreq/cpufreq_ondemand.c
368
ondemand_powersave_bias_init(policy);
drivers/cpufreq/cpufreq_ondemand.c
37
static unsigned int generic_powersave_bias_target(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq_ondemand.c
401
struct cpufreq_policy *policy;
drivers/cpufreq/cpufreq_ondemand.c
409
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/cpufreq_ondemand.c
410
if (!policy || policy->governor != &CPU_FREQ_GOV_ONDEMAND)
drivers/cpufreq/cpufreq_ondemand.c
413
policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_ondemand.c
417
cpumask_or(done, done, policy->cpus);
drivers/cpufreq/cpufreq_ondemand.c
44
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_ondemand.c
48
struct cpufreq_frequency_table *freq_table = policy->freq_table;
drivers/cpufreq/cpufreq_ondemand.c
56
index = cpufreq_frequency_table_target(policy, freq_next, policy->min,
drivers/cpufreq/cpufreq_ondemand.c
57
policy->max, relation);
drivers/cpufreq/cpufreq_ondemand.c
63
index = cpufreq_table_find_index_h(policy, freq_avg,
drivers/cpufreq/cpufreq_ondemand.c
66
index = cpufreq_table_find_index_l(policy, freq_avg,
drivers/cpufreq/cpufreq_ondemand.c
85
static void ondemand_powersave_bias_init(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_ondemand.c
87
struct od_policy_dbs_info *dbs_info = to_dbs_info(policy->governor_data);
drivers/cpufreq/cpufreq_ondemand.c
92
static void dbs_freq_increase(struct cpufreq_policy *policy, unsigned int freq)
drivers/cpufreq/cpufreq_ondemand.c
94
struct policy_dbs_info *policy_dbs = policy->governor_data;
drivers/cpufreq/cpufreq_ondemand.c
99
freq = od_ops.powersave_bias_target(policy, freq,
drivers/cpufreq/cpufreq_performance.c
14
static void cpufreq_gov_performance_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_performance.c
16
pr_debug("setting to %u kHz\n", policy->max);
drivers/cpufreq/cpufreq_performance.c
17
__cpufreq_driver_target(policy, policy->max, CPUFREQ_RELATION_H);
drivers/cpufreq/cpufreq_powersave.c
14
static void cpufreq_gov_powersave_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_powersave.c
16
pr_debug("setting to %u kHz\n", policy->min);
drivers/cpufreq/cpufreq_powersave.c
17
__cpufreq_driver_target(policy, policy->min, CPUFREQ_RELATION_L);
drivers/cpufreq/cpufreq_stats.c
103
static ssize_t store_reset(struct cpufreq_policy *policy, const char *buf,
drivers/cpufreq/cpufreq_stats.c
106
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_stats.c
124
static ssize_t show_trans_table(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq_stats.c
126
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_stats.c
194
void cpufreq_stats_free_table(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_stats.c
196
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_stats.c
204
sysfs_remove_group(&policy->kobj, &stats_attr_group);
drivers/cpufreq/cpufreq_stats.c
207
policy->stats = NULL;
drivers/cpufreq/cpufreq_stats.c
210
void cpufreq_stats_create_table(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_stats.c
217
count = cpufreq_table_count_valid_entries(policy);
drivers/cpufreq/cpufreq_stats.c
222
if (policy->stats)
drivers/cpufreq/cpufreq_stats.c
245
cpufreq_for_each_valid_entry(pos, policy->freq_table)
drivers/cpufreq/cpufreq_stats.c
246
if (policy->freq_table_sorted != CPUFREQ_TABLE_UNSORTED ||
drivers/cpufreq/cpufreq_stats.c
252
stats->last_index = freq_table_get_index(stats, policy->cur);
drivers/cpufreq/cpufreq_stats.c
254
policy->stats = stats;
drivers/cpufreq/cpufreq_stats.c
255
if (!sysfs_create_group(&policy->kobj, &stats_attr_group))
drivers/cpufreq/cpufreq_stats.c
259
policy->stats = NULL;
drivers/cpufreq/cpufreq_stats.c
265
void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
drivers/cpufreq/cpufreq_stats.c
268
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_stats.c
58
static ssize_t show_total_trans(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq_stats.c
60
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_stats.c
69
static ssize_t show_time_in_state(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq_stats.c
71
struct cpufreq_stats *stats = policy->stats;
drivers/cpufreq/cpufreq_userspace.c
107
static void cpufreq_userspace_policy_limits(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_userspace.c
109
struct userspace_policy *userspace = policy->governor_data;
drivers/cpufreq/cpufreq_userspace.c
114
policy->cpu, policy->min, policy->max, policy->cur, userspace->setspeed);
drivers/cpufreq/cpufreq_userspace.c
116
if (policy->max < userspace->setspeed)
drivers/cpufreq/cpufreq_userspace.c
117
__cpufreq_driver_target(policy, policy->max,
drivers/cpufreq/cpufreq_userspace.c
119
else if (policy->min > userspace->setspeed)
drivers/cpufreq/cpufreq_userspace.c
120
__cpufreq_driver_target(policy, policy->min,
drivers/cpufreq/cpufreq_userspace.c
123
__cpufreq_driver_target(policy, userspace->setspeed,
drivers/cpufreq/cpufreq_userspace.c
31
static int cpufreq_set(struct cpufreq_policy *policy, unsigned int freq)
drivers/cpufreq/cpufreq_userspace.c
34
struct userspace_policy *userspace = policy->governor_data;
drivers/cpufreq/cpufreq_userspace.c
36
pr_debug("cpufreq_set for cpu %u, freq %u kHz\n", policy->cpu, freq);
drivers/cpufreq/cpufreq_userspace.c
44
ret = __cpufreq_driver_target(policy, freq, CPUFREQ_RELATION_L);
drivers/cpufreq/cpufreq_userspace.c
50
static ssize_t show_speed(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/cpufreq_userspace.c
52
struct userspace_policy *userspace = policy->governor_data;
drivers/cpufreq/cpufreq_userspace.c
57
static int cpufreq_userspace_policy_init(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_userspace.c
67
policy->governor_data = userspace;
drivers/cpufreq/cpufreq_userspace.c
75
static void cpufreq_userspace_policy_exit(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_userspace.c
77
kfree(policy->governor_data);
drivers/cpufreq/cpufreq_userspace.c
78
policy->governor_data = NULL;
drivers/cpufreq/cpufreq_userspace.c
81
static int cpufreq_userspace_policy_start(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_userspace.c
83
struct userspace_policy *userspace = policy->governor_data;
drivers/cpufreq/cpufreq_userspace.c
85
BUG_ON(!policy->cur);
drivers/cpufreq/cpufreq_userspace.c
86
pr_debug("started managing cpu %u\n", policy->cpu);
drivers/cpufreq/cpufreq_userspace.c
90
userspace->setspeed = policy->cur;
drivers/cpufreq/cpufreq_userspace.c
95
static void cpufreq_userspace_policy_stop(struct cpufreq_policy *policy)
drivers/cpufreq/cpufreq_userspace.c
97
struct userspace_policy *userspace = policy->governor_data;
drivers/cpufreq/cpufreq_userspace.c
99
pr_debug("managing cpu %u stopped\n", policy->cpu);
drivers/cpufreq/davinci-cpufreq.c
35
static int davinci_target(struct cpufreq_policy *policy, unsigned int idx)
drivers/cpufreq/davinci-cpufreq.c
42
old_freq = policy->cur;
drivers/cpufreq/davinci-cpufreq.c
69
static int davinci_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/davinci-cpufreq.c
75
if (policy->cpu != 0)
drivers/cpufreq/davinci-cpufreq.c
85
policy->clk = cpufreq.armclk;
drivers/cpufreq/davinci-cpufreq.c
93
cpufreq_generic_init(policy, freq_table, 2000 * 1000);
drivers/cpufreq/e_powersaver.c
107
struct cpufreq_policy *policy,
drivers/cpufreq/e_powersaver.c
152
static int eps_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/e_powersaver.c
155
unsigned int cpu = policy->cpu;
drivers/cpufreq/e_powersaver.c
165
ret = eps_set_state(centaur, policy, dest_state);
drivers/cpufreq/e_powersaver.c
171
static int eps_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/e_powersaver.c
190
if (policy->cpu != 0)
drivers/cpufreq/e_powersaver.c
288
if (!acpi_processor_get_bios_limit(policy->cpu, &limit)) {
drivers/cpufreq/e_powersaver.c
292
eps_acpi_exit(policy);
drivers/cpufreq/e_powersaver.c
355
policy->cpuinfo.transition_latency = 140000; /* 844mV -> 700mV in ns */
drivers/cpufreq/e_powersaver.c
356
policy->freq_table = ¢aur->freq_table[0];
drivers/cpufreq/e_powersaver.c
361
static void eps_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/e_powersaver.c
363
unsigned int cpu = policy->cpu;
drivers/cpufreq/e_powersaver.c
78
static int eps_acpi_exit(struct cpufreq_policy *policy)
drivers/cpufreq/elanfreq.c
104
static int elanfreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/elanfreq.c
143
static int elanfreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/elanfreq.c
162
policy->freq_table = elanfreq_table;
drivers/cpufreq/freq_table.c
106
int cpufreq_generic_frequency_table_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/freq_table.c
108
if (!policy->freq_table)
drivers/cpufreq/freq_table.c
111
return cpufreq_frequency_table_verify(policy);
drivers/cpufreq/freq_table.c
115
int cpufreq_table_index_unsorted(struct cpufreq_policy *policy,
drivers/cpufreq/freq_table.c
128
struct cpufreq_frequency_table *table = policy->freq_table;
drivers/cpufreq/freq_table.c
133
target_freq, relation, policy->cpu);
drivers/cpufreq/freq_table.c
17
static bool policy_has_boost_freq(struct cpufreq_policy *policy)
drivers/cpufreq/freq_table.c
19
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
194
WARN(1, "Invalid frequency table: %u\n", policy->cpu);
drivers/cpufreq/freq_table.c
208
int cpufreq_frequency_table_get_index(struct cpufreq_policy *policy,
drivers/cpufreq/freq_table.c
211
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
230
static ssize_t show_available_freqs(struct cpufreq_policy *policy, char *buf,
drivers/cpufreq/freq_table.c
234
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
270
static ssize_t scaling_available_frequencies_show(struct cpufreq_policy *policy,
drivers/cpufreq/freq_table.c
273
return show_available_freqs(policy, buf, false);
drivers/cpufreq/freq_table.c
281
static ssize_t scaling_boost_frequencies_show(struct cpufreq_policy *policy,
drivers/cpufreq/freq_table.c
284
return show_available_freqs(policy, buf, true);
drivers/cpufreq/freq_table.c
288
static int set_freq_table_sorted(struct cpufreq_policy *policy)
drivers/cpufreq/freq_table.c
290
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
294
policy->freq_table_sorted = CPUFREQ_TABLE_UNSORTED;
drivers/cpufreq/freq_table.c
31
int cpufreq_frequency_table_cpuinfo(struct cpufreq_policy *policy)
drivers/cpufreq/freq_table.c
33
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
333
policy->freq_table_sorted = CPUFREQ_TABLE_SORTED_ASCENDING;
drivers/cpufreq/freq_table.c
335
policy->freq_table_sorted = CPUFREQ_TABLE_SORTED_DESCENDING;
drivers/cpufreq/freq_table.c
343
int cpufreq_table_validate_and_sort(struct cpufreq_policy *policy)
drivers/cpufreq/freq_table.c
347
if (!policy->freq_table) {
drivers/cpufreq/freq_table.c
355
ret = cpufreq_frequency_table_cpuinfo(policy);
drivers/cpufreq/freq_table.c
360
if (policy_has_boost_freq(policy))
drivers/cpufreq/freq_table.c
361
policy->boost_supported = true;
drivers/cpufreq/freq_table.c
363
if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING ||
drivers/cpufreq/freq_table.c
364
policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_DESCENDING)
drivers/cpufreq/freq_table.c
367
return set_freq_table_sorted(policy);
drivers/cpufreq/freq_table.c
41
if ((!cpufreq_boost_enabled() || !policy->boost_enabled)
drivers/cpufreq/freq_table.c
52
policy->min = policy->cpuinfo.min_freq = min_freq;
drivers/cpufreq/freq_table.c
53
policy->max = max_freq;
drivers/cpufreq/freq_table.c
58
if (policy->cpuinfo.max_freq < max_freq)
drivers/cpufreq/freq_table.c
59
policy->max = policy->cpuinfo.max_freq = max_freq;
drivers/cpufreq/freq_table.c
61
if (policy->min == ~0)
drivers/cpufreq/freq_table.c
67
int cpufreq_frequency_table_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/freq_table.c
69
struct cpufreq_frequency_table *pos, *table = policy->freq_table;
drivers/cpufreq/freq_table.c
74
policy->min, policy->max, policy->cpu);
drivers/cpufreq/freq_table.c
76
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/freq_table.c
81
if ((freq >= policy->min) && (freq <= policy->max)) {
drivers/cpufreq/freq_table.c
86
if ((prev_smaller < freq) && (freq <= policy->max))
drivers/cpufreq/freq_table.c
91
policy->max = prev_smaller;
drivers/cpufreq/freq_table.c
92
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/freq_table.c
96
policy->min, policy->max, policy->cpu);
drivers/cpufreq/gx-suspmod.c
250
static void gx_set_cpuspeed(struct cpufreq_policy *policy, unsigned int khz)
drivers/cpufreq/gx-suspmod.c
264
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/gx-suspmod.c
313
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/gx-suspmod.c
331
static int cpufreq_gx_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/gx-suspmod.c
336
if (!stock_freq || !policy)
drivers/cpufreq/gx-suspmod.c
339
policy->cpu = 0;
drivers/cpufreq/gx-suspmod.c
340
cpufreq_verify_within_limits(policy, (stock_freq / max_duration),
drivers/cpufreq/gx-suspmod.c
349
tmp_freq = gx_validate_speed(policy->min, &tmp1, &tmp2);
drivers/cpufreq/gx-suspmod.c
350
if (tmp_freq < policy->min)
drivers/cpufreq/gx-suspmod.c
352
policy->min = tmp_freq;
drivers/cpufreq/gx-suspmod.c
353
if (policy->min > policy->max)
drivers/cpufreq/gx-suspmod.c
354
policy->max = tmp_freq;
drivers/cpufreq/gx-suspmod.c
355
tmp_freq = gx_validate_speed(policy->max, &tmp1, &tmp2);
drivers/cpufreq/gx-suspmod.c
356
if (tmp_freq > policy->max)
drivers/cpufreq/gx-suspmod.c
358
policy->max = tmp_freq;
drivers/cpufreq/gx-suspmod.c
359
if (policy->max < policy->min)
drivers/cpufreq/gx-suspmod.c
360
policy->max = policy->min;
drivers/cpufreq/gx-suspmod.c
361
cpufreq_verify_within_limits(policy, (stock_freq / max_duration),
drivers/cpufreq/gx-suspmod.c
371
static int cpufreq_gx_target(struct cpufreq_policy *policy,
drivers/cpufreq/gx-suspmod.c
378
if (!stock_freq || !policy)
drivers/cpufreq/gx-suspmod.c
381
policy->cpu = 0;
drivers/cpufreq/gx-suspmod.c
384
while (tmp_freq < policy->min) {
drivers/cpufreq/gx-suspmod.c
388
while (tmp_freq > policy->max) {
drivers/cpufreq/gx-suspmod.c
393
gx_set_cpuspeed(policy, tmp_freq);
drivers/cpufreq/gx-suspmod.c
398
static int cpufreq_gx_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/gx-suspmod.c
402
if (!policy || policy->cpu != 0)
drivers/cpufreq/gx-suspmod.c
418
policy->cpu = 0;
drivers/cpufreq/gx-suspmod.c
421
policy->min = maxfreq / max_duration;
drivers/cpufreq/gx-suspmod.c
423
policy->min = maxfreq / POLICY_MIN_DIV;
drivers/cpufreq/gx-suspmod.c
424
policy->max = maxfreq;
drivers/cpufreq/gx-suspmod.c
425
policy->cpuinfo.min_freq = maxfreq / max_duration;
drivers/cpufreq/gx-suspmod.c
426
policy->cpuinfo.max_freq = maxfreq;
drivers/cpufreq/imx-cpufreq-dt.c
55
static unsigned int imx7ulp_get_intermediate(struct cpufreq_policy *policy,
drivers/cpufreq/imx-cpufreq-dt.c
61
static int imx7ulp_target_intermediate(struct cpufreq_policy *policy,
drivers/cpufreq/imx-cpufreq-dt.c
64
unsigned int newfreq = policy->freq_table[index].frequency;
drivers/cpufreq/imx6q-cpufreq.c
192
static int imx6q_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/imx6q-cpufreq.c
194
policy->clk = clks[ARM].clk;
drivers/cpufreq/imx6q-cpufreq.c
195
cpufreq_generic_init(policy, freq_table, transition_latency);
drivers/cpufreq/imx6q-cpufreq.c
196
policy->suspend_freq = max_freq;
drivers/cpufreq/imx6q-cpufreq.c
60
static int imx6q_set_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/intel_pstate.c
1269
if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/intel_pstate.c
1280
if (cpu_data->epp_policy == cpu_data->policy)
drivers/cpufreq/intel_pstate.c
1283
cpu_data->epp_policy = cpu_data->policy;
drivers/cpufreq/intel_pstate.c
1285
if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) {
drivers/cpufreq/intel_pstate.c
1417
static int intel_pstate_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
1419
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
1431
static int intel_pstate_resume(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
1433
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
1465
static void __intel_pstate_update_max_freq(struct cpufreq_policy *policy,
drivers/cpufreq/intel_pstate.c
1468
guard(cpufreq_policy_write)(policy);
drivers/cpufreq/intel_pstate.c
1473
policy->cpuinfo.max_freq = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
1476
refresh_frequency_limits(policy);
drivers/cpufreq/intel_pstate.c
1481
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/intel_pstate.c
1482
if (!policy)
drivers/cpufreq/intel_pstate.c
1485
__intel_pstate_update_max_freq(policy, all_cpu_data[cpu]);
drivers/cpufreq/intel_pstate.c
1490
static void intel_pstate_update_limits(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
1492
struct cpudata *cpudata = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
1494
__intel_pstate_update_max_freq(policy, cpudata);
drivers/cpufreq/intel_pstate.c
1653
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/intel_pstate.c
1654
if (!policy)
drivers/cpufreq/intel_pstate.c
1657
req = policy->driver_data;
drivers/cpufreq/intel_pstate.c
232
unsigned int policy;
drivers/cpufreq/intel_pstate.c
2684
#define X86_MATCH(vfm, policy) \
drivers/cpufreq/intel_pstate.c
2685
X86_MATCH_VFM_FEATURE(vfm, X86_FEATURE_APERFMPERF, &policy)
drivers/cpufreq/intel_pstate.c
2885
static int intel_pstate_set_policy(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
2889
if (!policy->cpuinfo.max_freq)
drivers/cpufreq/intel_pstate.c
2893
policy->cpuinfo.max_freq, policy->max);
drivers/cpufreq/intel_pstate.c
2895
cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
2896
cpu->policy = policy->policy;
drivers/cpufreq/intel_pstate.c
2900
intel_pstate_update_perf_limits(cpu, policy->min, policy->max);
drivers/cpufreq/intel_pstate.c
2902
if (cpu->policy == CPUFREQ_POLICY_PERFORMANCE) {
drivers/cpufreq/intel_pstate.c
2909
intel_pstate_clear_update_util_hook(policy->cpu);
drivers/cpufreq/intel_pstate.c
2912
intel_pstate_set_update_util_hook(policy->cpu);
drivers/cpufreq/intel_pstate.c
2922
intel_pstate_clear_update_util_hook(policy->cpu);
drivers/cpufreq/intel_pstate.c
2923
intel_pstate_hwp_set(policy->cpu);
drivers/cpufreq/intel_pstate.c
2929
policy->cur = policy->min;
drivers/cpufreq/intel_pstate.c
2937
struct cpufreq_policy_data *policy)
drivers/cpufreq/intel_pstate.c
2941
policy->max < policy->cpuinfo.max_freq &&
drivers/cpufreq/intel_pstate.c
2942
policy->max > cpu->pstate.max_freq) {
drivers/cpufreq/intel_pstate.c
2944
policy->max = policy->cpuinfo.max_freq;
drivers/cpufreq/intel_pstate.c
2949
struct cpufreq_policy_data *policy)
drivers/cpufreq/intel_pstate.c
2960
cpufreq_verify_within_limits(policy, policy->cpuinfo.min_freq, max_freq);
drivers/cpufreq/intel_pstate.c
2962
intel_pstate_adjust_policy_max(cpu, policy);
drivers/cpufreq/intel_pstate.c
2965
static int intel_pstate_verify_policy(struct cpufreq_policy_data *policy)
drivers/cpufreq/intel_pstate.c
2967
intel_pstate_verify_cpu_policy(all_cpu_data[policy->cpu], policy);
drivers/cpufreq/intel_pstate.c
2972
static int intel_cpufreq_cpu_offline(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
2974
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
2992
intel_pstate_exit_perf_limits(policy);
drivers/cpufreq/intel_pstate.c
2997
static int intel_pstate_cpu_online(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
2999
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3003
intel_pstate_init_acpi_perf_limits(policy);
drivers/cpufreq/intel_pstate.c
3019
static int intel_pstate_cpu_offline(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3021
intel_pstate_clear_update_util_hook(policy->cpu);
drivers/cpufreq/intel_pstate.c
3023
return intel_cpufreq_cpu_offline(policy);
drivers/cpufreq/intel_pstate.c
3026
static void intel_pstate_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3028
pr_debug("CPU %d exiting\n", policy->cpu);
drivers/cpufreq/intel_pstate.c
3030
policy->fast_switch_possible = false;
drivers/cpufreq/intel_pstate.c
3033
static int __intel_pstate_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3038
rc = intel_pstate_init_cpu(policy->cpu);
drivers/cpufreq/intel_pstate.c
3042
cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3048
policy->cpuinfo.min_freq = cpu->pstate.min_freq;
drivers/cpufreq/intel_pstate.c
3049
policy->cpuinfo.max_freq = READ_ONCE(global.no_turbo) ?
drivers/cpufreq/intel_pstate.c
3052
policy->min = policy->cpuinfo.min_freq;
drivers/cpufreq/intel_pstate.c
3053
policy->max = policy->cpuinfo.max_freq;
drivers/cpufreq/intel_pstate.c
3055
intel_pstate_init_acpi_perf_limits(policy);
drivers/cpufreq/intel_pstate.c
3057
policy->fast_switch_possible = true;
drivers/cpufreq/intel_pstate.c
3062
static int intel_pstate_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3064
int ret = __intel_pstate_cpu_init(policy);
drivers/cpufreq/intel_pstate.c
3073
policy->policy = CPUFREQ_POLICY_POWERSAVE;
drivers/cpufreq/intel_pstate.c
3076
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3098
static int intel_cpufreq_verify_policy(struct cpufreq_policy_data *policy)
drivers/cpufreq/intel_pstate.c
3100
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3102
intel_pstate_verify_cpu_policy(cpu, policy);
drivers/cpufreq/intel_pstate.c
3103
intel_pstate_update_perf_limits(cpu, policy->min, policy->max);
drivers/cpufreq/intel_pstate.c
3181
static int intel_cpufreq_update_pstate(struct cpufreq_policy *policy,
drivers/cpufreq/intel_pstate.c
3184
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3189
int max_pstate = policy->strict_target ?
drivers/cpufreq/intel_pstate.c
3206
static int intel_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/intel_pstate.c
3210
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3214
freqs.old = policy->cur;
drivers/cpufreq/intel_pstate.c
3217
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/intel_pstate.c
3220
target_pstate = intel_cpufreq_update_pstate(policy, target_pstate, false);
drivers/cpufreq/intel_pstate.c
3224
cpufreq_freq_transition_end(policy, &freqs, false);
drivers/cpufreq/intel_pstate.c
3229
static unsigned int intel_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/intel_pstate.c
3232
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3237
target_pstate = intel_cpufreq_update_pstate(policy, target_pstate, true);
drivers/cpufreq/intel_pstate.c
3287
static int intel_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3294
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/intel_pstate.c
3298
ret = __intel_pstate_cpu_init(policy);
drivers/cpufreq/intel_pstate.c
3302
policy->cpuinfo.transition_latency = INTEL_CPUFREQ_TRANSITION_LATENCY;
drivers/cpufreq/intel_pstate.c
3304
policy->cur = policy->cpuinfo.min_freq;
drivers/cpufreq/intel_pstate.c
3312
cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
3317
policy->transition_delay_us = INTEL_CPUFREQ_TRANSITION_DELAY_HWP;
drivers/cpufreq/intel_pstate.c
3326
policy->transition_delay_us = INTEL_CPUFREQ_TRANSITION_DELAY;
drivers/cpufreq/intel_pstate.c
3331
ret = freq_qos_add_request(&policy->constraints, req, FREQ_QOS_MIN,
drivers/cpufreq/intel_pstate.c
3340
ret = freq_qos_add_request(&policy->constraints, req + 1, FREQ_QOS_MAX,
drivers/cpufreq/intel_pstate.c
3347
policy->driver_data = req;
drivers/cpufreq/intel_pstate.c
3356
intel_pstate_exit_perf_limits(policy);
drivers/cpufreq/intel_pstate.c
3361
static void intel_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3365
req = policy->driver_data;
drivers/cpufreq/intel_pstate.c
3371
intel_pstate_cpu_exit(policy);
drivers/cpufreq/intel_pstate.c
3374
static int intel_cpufreq_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
3376
intel_pstate_suspend(policy);
drivers/cpufreq/intel_pstate.c
3379
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
442
static void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
449
intel_pstate_set_itmt_prio(policy->cpu);
drivers/cpufreq/intel_pstate.c
456
cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
459
policy->cpu);
drivers/cpufreq/intel_pstate.c
479
pr_debug("CPU%u - ACPI _PSS perf data\n", policy->cpu);
drivers/cpufreq/intel_pstate.c
495
acpi_processor_unregister_performance(policy->cpu);
drivers/cpufreq/intel_pstate.c
498
static void intel_pstate_exit_perf_limits(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
502
cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
506
acpi_processor_unregister_performance(policy->cpu);
drivers/cpufreq/intel_pstate.c
509
static inline void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
513
static inline void intel_pstate_exit_perf_limits(struct cpufreq_policy *policy)
drivers/cpufreq/intel_pstate.c
767
if (epp > 0 && cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/intel_pstate.c
777
struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/intel_pstate.c
795
struct cpufreq_policy *policy, const char *buf, size_t count)
drivers/cpufreq/intel_pstate.c
797
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
847
cpufreq_stop_governor(policy);
drivers/cpufreq/intel_pstate.c
849
err = cpufreq_start_governor(policy);
drivers/cpufreq/intel_pstate.c
863
struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/intel_pstate.c
865
struct cpudata *cpu_data = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
880
static ssize_t show_base_frequency(struct cpufreq_policy *policy, char *buf)
drivers/cpufreq/intel_pstate.c
882
struct cpudata *cpu = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
885
ratio = intel_pstate_get_cppc_guaranteed(policy->cpu);
drivers/cpufreq/intel_pstate.c
889
rdmsrq_on_cpu(policy->cpu, MSR_HWP_CAPABILITIES, &cap);
drivers/cpufreq/kirkwood-cpufreq.c
50
static int kirkwood_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/kirkwood-cpufreq.c
86
static int kirkwood_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/kirkwood-cpufreq.c
88
cpufreq_generic_init(policy, kirkwood_freq_table, 5000);
drivers/cpufreq/longhaul.c
246
static int longhaul_setstate(struct cpufreq_policy *policy,
drivers/cpufreq/longhaul.c
619
static int longhaul_target(struct cpufreq_policy *policy,
drivers/cpufreq/longhaul.c
628
retval = longhaul_setstate(policy, table_index);
drivers/cpufreq/longhaul.c
643
retval = longhaul_setstate(policy, i);
drivers/cpufreq/longhaul.c
652
retval = longhaul_setstate(policy, table_index);
drivers/cpufreq/longhaul.c
759
static int longhaul_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/longhaul.c
896
policy->transition_delay_us = 200000; /* usec */
drivers/cpufreq/longhaul.c
897
policy->freq_table = longhaul_table;
drivers/cpufreq/longhaul.c
952
struct cpufreq_policy *policy = cpufreq_cpu_get(0);
drivers/cpufreq/longhaul.c
955
if (unlikely(!policy))
drivers/cpufreq/longhaul.c
962
freqs.old = policy->cur;
drivers/cpufreq/longhaul.c
966
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/longhaul.c
967
longhaul_setstate(policy, i);
drivers/cpufreq/longhaul.c
968
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/longhaul.c
973
cpufreq_cpu_put(policy);
drivers/cpufreq/longrun.c
125
static int longrun_verify_policy(struct cpufreq_policy_data *policy)
drivers/cpufreq/longrun.c
127
if (!policy)
drivers/cpufreq/longrun.c
130
policy->cpu = 0;
drivers/cpufreq/longrun.c
131
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/longrun.c
252
static int longrun_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/longrun.c
257
if (policy->cpu != 0)
drivers/cpufreq/longrun.c
266
policy->cpuinfo.min_freq = longrun_low_freq;
drivers/cpufreq/longrun.c
267
policy->cpuinfo.max_freq = longrun_high_freq;
drivers/cpufreq/longrun.c
268
longrun_get_policy(policy);
drivers/cpufreq/longrun.c
35
static void longrun_get_policy(struct cpufreq_policy *policy)
drivers/cpufreq/longrun.c
42
policy->policy = CPUFREQ_POLICY_PERFORMANCE;
drivers/cpufreq/longrun.c
44
policy->policy = CPUFREQ_POLICY_POWERSAVE;
drivers/cpufreq/longrun.c
53
policy->min = policy->max = longrun_high_freq;
drivers/cpufreq/longrun.c
55
policy->min = longrun_low_freq + msr_lo *
drivers/cpufreq/longrun.c
57
policy->max = longrun_low_freq + msr_hi *
drivers/cpufreq/longrun.c
60
policy->cpu = 0;
drivers/cpufreq/longrun.c
71
static int longrun_set_policy(struct cpufreq_policy *policy)
drivers/cpufreq/longrun.c
76
if (!policy)
drivers/cpufreq/longrun.c
83
pctg_lo = (policy->min - longrun_low_freq) /
drivers/cpufreq/longrun.c
85
pctg_hi = (policy->max - longrun_low_freq) /
drivers/cpufreq/longrun.c
97
switch (policy->policy) {
drivers/cpufreq/loongson2_cpufreq.c
49
static int loongson2_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/loongson2_cpufreq.c
64
static int loongson2_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/loongson2_cpufreq.c
84
cpufreq_generic_init(policy, &loongson2_clockmod_table[0], 0);
drivers/cpufreq/loongson3_cpufreq.c
229
static int loongson3_cpufreq_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/loongson3_cpufreq.c
233
ret = do_service_request(cpu_data[policy->cpu].core,
drivers/cpufreq/loongson3_cpufreq.c
284
static int loongson3_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/loongson3_cpufreq.c
286
int i, ret, cpu = policy->cpu;
drivers/cpufreq/loongson3_cpufreq.c
292
policy->cpuinfo.transition_latency = 10000;
drivers/cpufreq/loongson3_cpufreq.c
293
policy->freq_table = per_cpu(freq_data, cpu)->table;
drivers/cpufreq/loongson3_cpufreq.c
294
policy->suspend_freq = policy->freq_table[per_cpu(freq_data, cpu)->def_freq_level].frequency;
drivers/cpufreq/loongson3_cpufreq.c
295
cpumask_copy(policy->cpus, topology_sibling_cpumask(cpu));
drivers/cpufreq/loongson3_cpufreq.c
297
for_each_cpu(i, policy->cpus) {
drivers/cpufreq/loongson3_cpufreq.c
305
static void loongson3_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/loongson3_cpufreq.c
307
int cpu = policy->cpu;
drivers/cpufreq/loongson3_cpufreq.c
309
loongson3_cpufreq_target(policy, per_cpu(freq_data, cpu)->def_freq_level);
drivers/cpufreq/loongson3_cpufreq.c
312
static int loongson3_cpufreq_cpu_online(struct cpufreq_policy *policy)
drivers/cpufreq/loongson3_cpufreq.c
317
static int loongson3_cpufreq_cpu_offline(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
100
struct cpufreq_policy *policy;
drivers/cpufreq/mediatek-cpufreq-hw.c
103
policy = cpufreq_cpu_get_raw(cpu_dev->id);
drivers/cpufreq/mediatek-cpufreq-hw.c
104
if (!policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
107
data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
124
struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
126
struct mtk_cpufreq_domain *data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
131
for_each_cpu(cpu, policy->real_cpus) {
drivers/cpufreq/mediatek-cpufreq-hw.c
136
static int mtk_cpufreq_hw_target_index(struct cpufreq_policy *policy,
drivers/cpufreq/mediatek-cpufreq-hw.c
139
struct mtk_cpufreq_domain *data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
143
target_freq = policy->freq_table[index].frequency;
drivers/cpufreq/mediatek-cpufreq-hw.c
144
mtk_cpufreq_hw_fdvfs_switch(target_freq, policy);
drivers/cpufreq/mediatek-cpufreq-hw.c
155
struct cpufreq_policy *policy;
drivers/cpufreq/mediatek-cpufreq-hw.c
158
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/mediatek-cpufreq-hw.c
159
if (!policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
162
data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
170
static unsigned int mtk_cpufreq_hw_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/mediatek-cpufreq-hw.c
173
struct mtk_cpufreq_domain *data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
176
index = cpufreq_table_find_index_dl(policy, target_freq, false);
drivers/cpufreq/mediatek-cpufreq-hw.c
179
mtk_cpufreq_hw_fdvfs_switch(target_freq, policy);
drivers/cpufreq/mediatek-cpufreq-hw.c
183
return policy->freq_table[index].frequency;
drivers/cpufreq/mediatek-cpufreq-hw.c
221
struct cpufreq_policy *policy,
drivers/cpufreq/mediatek-cpufreq-hw.c
236
ret = of_perf_domain_get_sharing_cpumask(policy->cpu, "performance-domains",
drivers/cpufreq/mediatek-cpufreq-hw.c
238
policy->cpus, &args);
drivers/cpufreq/mediatek-cpufreq-hw.c
284
policy->freq_table = data->table;
drivers/cpufreq/mediatek-cpufreq-hw.c
285
policy->driver_data = data;
drivers/cpufreq/mediatek-cpufreq-hw.c
293
static int mtk_cpufreq_hw_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
302
ret = mtk_cpu_resources_init(pdev, policy, platform_get_drvdata(pdev));
drivers/cpufreq/mediatek-cpufreq-hw.c
308
data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
314
policy->cpuinfo.transition_latency = latency;
drivers/cpufreq/mediatek-cpufreq-hw.c
315
policy->fast_switch_possible = true;
drivers/cpufreq/mediatek-cpufreq-hw.c
324
policy->cpu);
drivers/cpufreq/mediatek-cpufreq-hw.c
328
pr_info("SVS of CPU%d is not enabled\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq-hw.c
334
static void mtk_cpufreq_hw_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
336
struct mtk_cpufreq_domain *data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
346
static void mtk_cpufreq_register_em(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq-hw.c
349
struct mtk_cpufreq_domain *data = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq-hw.c
351
em_dev_register_perf_domain(get_cpu_device(policy->cpu), data->nr_opp,
drivers/cpufreq/mediatek-cpufreq-hw.c
352
&em_cb, policy->cpus, true);
drivers/cpufreq/mediatek-cpufreq.c
200
static int mtk_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/mediatek-cpufreq.c
203
struct cpufreq_frequency_table *freq_table = policy->freq_table;
drivers/cpufreq/mediatek-cpufreq.c
204
struct clk *cpu_clk = policy->clk;
drivers/cpufreq/mediatek-cpufreq.c
206
struct mtk_cpu_dvfs_info *info = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq.c
234
policy->cpu, freq_hz);
drivers/cpufreq/mediatek-cpufreq.c
258
"cpu%d: failed to scale up voltage!\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
268
"cpu%d: failed to re-parent cpu clock!\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
277
"cpu%d: failed to scale cpu clock rate!\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
287
"cpu%d: failed to re-parent cpu clock!\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
300
"cpu%d: failed to scale down voltage!\n", policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
356
struct cpufreq_policy *policy __free(put_cpufreq_policy)
drivers/cpufreq/mediatek-cpufreq.c
358
if (policy)
drivers/cpufreq/mediatek-cpufreq.c
359
cpufreq_driver_target(policy, freq / 1000,
drivers/cpufreq/mediatek-cpufreq.c
581
static int mtk_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq.c
587
info = mtk_cpu_dvfs_info_lookup(policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
590
policy->cpu);
drivers/cpufreq/mediatek-cpufreq.c
598
policy->cpu, ret);
drivers/cpufreq/mediatek-cpufreq.c
602
cpumask_copy(policy->cpus, &info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
603
policy->freq_table = freq_table;
drivers/cpufreq/mediatek-cpufreq.c
604
policy->driver_data = info;
drivers/cpufreq/mediatek-cpufreq.c
605
policy->clk = info->cpu_clk;
drivers/cpufreq/mediatek-cpufreq.c
610
static void mtk_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/mediatek-cpufreq.c
612
struct mtk_cpu_dvfs_info *info = policy->driver_data;
drivers/cpufreq/mediatek-cpufreq.c
614
dev_pm_opp_free_cpufreq_table(info->cpu_dev, &policy->freq_table);
drivers/cpufreq/p4-clockmod.c
103
static int cpufreq_p4_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/p4-clockmod.c
111
for_each_cpu(i, policy->cpus)
drivers/cpufreq/p4-clockmod.c
155
static int cpufreq_p4_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/p4-clockmod.c
157
struct cpuinfo_x86 *c = &cpu_data(policy->cpu);
drivers/cpufreq/p4-clockmod.c
162
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
drivers/cpufreq/p4-clockmod.c
172
has_N44_O17_errata[policy->cpu] = 1;
drivers/cpufreq/p4-clockmod.c
179
cpufreq_p4_setdc(policy->cpu, DC_DISABLE);
drivers/cpufreq/p4-clockmod.c
189
if ((i < 2) && (has_N44_O17_errata[policy->cpu]))
drivers/cpufreq/p4-clockmod.c
199
policy->cpuinfo.transition_latency = 10000001;
drivers/cpufreq/p4-clockmod.c
200
policy->freq_table = &p4clockmod_table[0];
drivers/cpufreq/pasemi-cpufreq.c
124
static int pas_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/pasemi-cpufreq.c
134
cpu = of_get_cpu_node(policy->cpu, NULL);
drivers/cpufreq/pasemi-cpufreq.c
182
pr_debug("init cpufreq on CPU %d\n", policy->cpu);
drivers/cpufreq/pasemi-cpufreq.c
192
cur_astate = get_cur_astate(policy->cpu);
drivers/cpufreq/pasemi-cpufreq.c
195
policy->cur = pas_freqs[cur_astate].frequency;
drivers/cpufreq/pasemi-cpufreq.c
196
ppc_proc_freq = policy->cur * 1000ul;
drivers/cpufreq/pasemi-cpufreq.c
198
cpufreq_generic_init(policy, pas_freqs, get_gizmo_latency());
drivers/cpufreq/pasemi-cpufreq.c
207
static void pas_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/pasemi-cpufreq.c
222
static int pas_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/pasemi-cpufreq.c
228
policy->cpu,
drivers/cpufreq/pcc-cpufreq.c
113
static int pcc_cpufreq_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/pcc-cpufreq.c
115
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/pcc-cpufreq.c
198
static int pcc_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/pcc-cpufreq.c
208
cpu = policy->cpu;
drivers/cpufreq/pcc-cpufreq.c
216
freqs.old = policy->cur;
drivers/cpufreq/pcc-cpufreq.c
218
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/pcc-cpufreq.c
236
cpufreq_freq_transition_end(policy, &freqs, status != CMD_COMPLETE);
drivers/cpufreq/pcc-cpufreq.c
538
static int pcc_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/pcc-cpufreq.c
540
unsigned int cpu = policy->cpu;
drivers/cpufreq/pcc-cpufreq.c
554
policy->max = policy->cpuinfo.max_freq =
drivers/cpufreq/pcc-cpufreq.c
556
policy->min = policy->cpuinfo.min_freq =
drivers/cpufreq/pcc-cpufreq.c
560
policy->max, policy->min);
drivers/cpufreq/pmac32-cpufreq.c
333
static int do_set_cpu_speed(struct cpufreq_policy *policy, int speed_mode)
drivers/cpufreq/pmac32-cpufreq.c
363
static int pmac_cpufreq_target( struct cpufreq_policy *policy,
drivers/cpufreq/pmac32-cpufreq.c
368
rc = do_set_cpu_speed(policy, index);
drivers/cpufreq/pmac32-cpufreq.c
374
static int pmac_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/pmac32-cpufreq.c
376
cpufreq_generic_init(policy, pmac_cpu_freqs, transition_latency);
drivers/cpufreq/pmac32-cpufreq.c
397
static int pmac_cpufreq_suspend(struct cpufreq_policy *policy)
drivers/cpufreq/pmac32-cpufreq.c
409
do_set_cpu_speed(policy, CPUFREQ_HIGH);
drivers/cpufreq/pmac32-cpufreq.c
413
static int pmac_cpufreq_resume(struct cpufreq_policy *policy)
drivers/cpufreq/pmac32-cpufreq.c
425
do_set_cpu_speed(policy, sleep_freq == low_freq ?
drivers/cpufreq/pmac64-cpufreq.c
312
static int g5_cpufreq_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/pmac64-cpufreq.c
322
static int g5_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/pmac64-cpufreq.c
324
cpufreq_generic_init(policy, g5_cpu_freqs, transition_latency);
drivers/cpufreq/powernow-k6.c
139
static int powernow_k6_target(struct cpufreq_policy *policy,
drivers/cpufreq/powernow-k6.c
153
static int powernow_k6_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/powernow-k6.c
159
if (policy->cpu != 0)
drivers/cpufreq/powernow-k6.c
215
policy->cpuinfo.transition_latency = 500000;
drivers/cpufreq/powernow-k6.c
216
policy->freq_table = clock_ratio;
drivers/cpufreq/powernow-k6.c
222
static void powernow_k6_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/powernow-k6.c
230
freqs.old = policy->cur;
drivers/cpufreq/powernow-k6.c
234
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/powernow-k6.c
235
powernow_k6_target(policy, i);
drivers/cpufreq/powernow-k6.c
236
cpufreq_freq_transition_end(policy, &freqs, 0);
drivers/cpufreq/powernow-k7.c
248
static int powernow_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/powernow-k7.c
593
static int powernow_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/powernow-k7.c
598
if (policy->cpu != 0)
drivers/cpufreq/powernow-k7.c
640
policy->cpuinfo.transition_latency =
drivers/cpufreq/powernow-k7.c
642
policy->freq_table = powernow_table;
drivers/cpufreq/powernow-k7.c
647
static void powernow_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/powernow-k8.c
882
struct cpufreq_policy *policy)
drivers/cpufreq/powernow-k8.c
915
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/powernow-k8.c
917
cpufreq_freq_transition_end(policy, &freqs, res);
drivers/cpufreq/powernv-cpufreq.c
1017
static unsigned int powernv_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/powernv-cpufreq.c
1023
index = cpufreq_table_find_index_dl(policy, target_freq, false);
drivers/cpufreq/powernv-cpufreq.c
219
static inline void reset_gpstates(struct cpufreq_policy *policy)
drivers/cpufreq/powernv-cpufreq.c
221
struct global_pstate_info *gpstates = policy->driver_data;
drivers/cpufreq/powernv-cpufreq.c
381
static ssize_t cpuinfo_nominal_freq_show(struct cpufreq_policy *policy,
drivers/cpufreq/powernv-cpufreq.c
397
static ssize_t name##_show(struct cpufreq_policy *policy, char *buf) \
drivers/cpufreq/powernv-cpufreq.c
399
struct chip *chip = per_cpu(chip_info, policy->cpu); \
drivers/cpufreq/powernv-cpufreq.c
674
struct cpufreq_policy *policy = gpstates->policy;
drivers/cpufreq/powernv-cpufreq.c
687
if (!cpumask_test_cpu(raw_smp_processor_id(), policy->cpus)) {
drivers/cpufreq/powernv-cpufreq.c
689
add_timer_on(&gpstates->timer, cpumask_first(policy->cpus));
drivers/cpufreq/powernv-cpufreq.c
703
reset_gpstates(policy);
drivers/cpufreq/powernv-cpufreq.c
714
reset_gpstates(policy);
drivers/cpufreq/powernv-cpufreq.c
741
static int powernv_cpufreq_target_index(struct cpufreq_policy *policy,
drivers/cpufreq/powernv-cpufreq.c
746
struct global_pstate_info *gpstates = policy->driver_data;
drivers/cpufreq/powernv-cpufreq.c
786
reset_gpstates(policy);
drivers/cpufreq/powernv-cpufreq.c
796
reset_gpstates(policy);
drivers/cpufreq/powernv-cpufreq.c
824
smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1);
drivers/cpufreq/powernv-cpufreq.c
828
static int powernv_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/powernv-cpufreq.c
834
base = cpu_first_thread_sibling(policy->cpu);
drivers/cpufreq/powernv-cpufreq.c
837
cpumask_set_cpu(base + i, policy->cpus);
drivers/cpufreq/powernv-cpufreq.c
839
kn = kernfs_find_and_get(policy->kobj.sd, throttle_attr_grp.name);
drivers/cpufreq/powernv-cpufreq.c
843
ret = sysfs_create_group(&policy->kobj, &throttle_attr_grp);
drivers/cpufreq/powernv-cpufreq.c
846
policy->cpu);
drivers/cpufreq/powernv-cpufreq.c
853
policy->freq_table = powernv_freqs;
drivers/cpufreq/powernv-cpufreq.c
854
policy->fast_switch_possible = true;
drivers/cpufreq/powernv-cpufreq.c
864
policy->driver_data = gpstates;
drivers/cpufreq/powernv-cpufreq.c
867
gpstates->policy = policy;
drivers/cpufreq/powernv-cpufreq.c
877
static void powernv_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/powernv-cpufreq.c
88
struct cpufreq_policy *policy;
drivers/cpufreq/powernv-cpufreq.c
880
struct global_pstate_info *gpstates = policy->driver_data;
drivers/cpufreq/powernv-cpufreq.c
884
smp_call_function_single(policy->cpu, set_pstate, &freq_data, 1);
drivers/cpufreq/powernv-cpufreq.c
888
kfree(policy->driver_data);
drivers/cpufreq/powernv-cpufreq.c
916
struct cpufreq_policy *policy;
drivers/cpufreq/powernv-cpufreq.c
932
policy = cpufreq_cpu_get(cpu);
drivers/cpufreq/powernv-cpufreq.c
933
if (!policy)
drivers/cpufreq/powernv-cpufreq.c
935
index = cpufreq_table_find_index_c(policy, policy->cur, false);
drivers/cpufreq/powernv-cpufreq.c
936
powernv_cpufreq_target_index(policy, index);
drivers/cpufreq/powernv-cpufreq.c
937
cpumask_andnot(&mask, &mask, policy->cpus);
drivers/cpufreq/powernv-cpufreq.c
938
cpufreq_cpu_put(policy);
drivers/cpufreq/pxa2xx-cpufreq.c
184
static int pxa_set_target(struct cpufreq_policy *policy, unsigned int idx)
drivers/cpufreq/pxa2xx-cpufreq.c
199
policy->cur / 1000, new_freq_cpu / 1000);
drivers/cpufreq/pxa2xx-cpufreq.c
201
if (vcc_core && new_freq_cpu > policy->cur) {
drivers/cpufreq/pxa2xx-cpufreq.c
218
if (vcc_core && new_freq_cpu < policy->cur)
drivers/cpufreq/pxa2xx-cpufreq.c
224
static int pxa_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/pxa2xx-cpufreq.c
238
policy->cpuinfo.transition_latency = 1000; /* FIXME: 1 ms, assumed */
drivers/cpufreq/pxa2xx-cpufreq.c
277
policy->freq_table = pxa255_freq_table;
drivers/cpufreq/pxa2xx-cpufreq.c
280
policy->freq_table = pxa27x_freq_table;
drivers/cpufreq/pxa3xx-cpufreq.c
107
static int setup_freqs_table(struct cpufreq_policy *policy,
drivers/cpufreq/pxa3xx-cpufreq.c
128
policy->freq_table = table;
drivers/cpufreq/pxa3xx-cpufreq.c
165
static int pxa3xx_cpufreq_set(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/pxa3xx-cpufreq.c
170
if (policy->cpu != 0)
drivers/cpufreq/pxa3xx-cpufreq.c
183
static int pxa3xx_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/pxa3xx-cpufreq.c
188
policy->min = policy->cpuinfo.min_freq = 104000;
drivers/cpufreq/pxa3xx-cpufreq.c
189
policy->max = policy->cpuinfo.max_freq =
drivers/cpufreq/pxa3xx-cpufreq.c
191
policy->cpuinfo.transition_latency = 1000; /* FIXME: 1 ms, assumed */
drivers/cpufreq/pxa3xx-cpufreq.c
194
ret = setup_freqs_table(policy, pxa300_freqs,
drivers/cpufreq/pxa3xx-cpufreq.c
198
ret = setup_freqs_table(policy, pxa320_freqs,
drivers/cpufreq/qcom-cpufreq-hw.c
113
static int qcom_cpufreq_hw_target_index(struct cpufreq_policy *policy,
drivers/cpufreq/qcom-cpufreq-hw.c
116
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
118
unsigned long freq = policy->freq_table[index].frequency;
drivers/cpufreq/qcom-cpufreq-hw.c
124
for (i = 1; i < cpumask_weight(policy->related_cpus); i++)
drivers/cpufreq/qcom-cpufreq-hw.c
128
qcom_cpufreq_set_bw(policy, freq);
drivers/cpufreq/qcom-cpufreq-hw.c
146
static unsigned int qcom_cpufreq_get_freq(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
152
if (!policy)
drivers/cpufreq/qcom-cpufreq-hw.c
155
data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
161
return policy->freq_table[index].frequency;
drivers/cpufreq/qcom-cpufreq-hw.c
164
static unsigned int __qcom_cpufreq_hw_get(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
168
if (!policy)
drivers/cpufreq/qcom-cpufreq-hw.c
171
data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
176
return qcom_cpufreq_get_freq(policy);
drivers/cpufreq/qcom-cpufreq-hw.c
184
static unsigned int qcom_cpufreq_hw_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/qcom-cpufreq-hw.c
187
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
192
index = policy->cached_resolved_idx;
drivers/cpufreq/qcom-cpufreq-hw.c
196
for (i = 1; i < cpumask_weight(policy->related_cpus); i++)
drivers/cpufreq/qcom-cpufreq-hw.c
199
return policy->freq_table[index].frequency;
drivers/cpufreq/qcom-cpufreq-hw.c
203
struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
211
struct qcom_cpufreq_data *drv_data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
235
policy->fast_switch_possible = true;
drivers/cpufreq/qcom-cpufreq-hw.c
297
policy->freq_table = table;
drivers/cpufreq/qcom-cpufreq-hw.c
298
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
328
struct cpufreq_policy *policy = data->policy;
drivers/cpufreq/qcom-cpufreq-hw.c
329
int cpu = cpumask_first(policy->related_cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
353
arch_update_hw_pressure(policy->related_cpus, throttled_freq);
drivers/cpufreq/qcom-cpufreq-hw.c
428
static int qcom_cpufreq_hw_lmh_init(struct cpufreq_policy *policy, int index)
drivers/cpufreq/qcom-cpufreq-hw.c
430
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
449
snprintf(data->irq_name, sizeof(data->irq_name), "dcvsh-irq-%u", policy->cpu);
drivers/cpufreq/qcom-cpufreq-hw.c
457
ret = irq_set_affinity_and_hint(data->throttle_irq, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
465
static int qcom_cpufreq_hw_cpu_online(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
467
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
478
ret = irq_set_affinity_and_hint(data->throttle_irq, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
486
static int qcom_cpufreq_hw_cpu_offline(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
488
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
512
static int qcom_cpufreq_hw_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
522
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/qcom-cpufreq-hw.c
525
policy->cpu);
drivers/cpufreq/qcom-cpufreq-hw.c
529
cpu_np = of_cpu_device_node_get(policy->cpu);
drivers/cpufreq/qcom-cpufreq-hw.c
551
qcom_get_related_cpus(index, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
553
policy->driver_data = data;
drivers/cpufreq/qcom-cpufreq-hw.c
554
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/qcom-cpufreq-hw.c
555
data->policy = policy;
drivers/cpufreq/qcom-cpufreq-hw.c
557
ret = qcom_cpufreq_hw_read_lut(cpu_dev, policy);
drivers/cpufreq/qcom-cpufreq-hw.c
569
return qcom_cpufreq_hw_lmh_init(policy, index);
drivers/cpufreq/qcom-cpufreq-hw.c
572
static void qcom_cpufreq_hw_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
574
struct device *cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/qcom-cpufreq-hw.c
575
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
578
dev_pm_opp_of_cpumask_remove_table(policy->related_cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
58
struct cpufreq_policy *policy;
drivers/cpufreq/qcom-cpufreq-hw.c
580
kfree(policy->freq_table);
drivers/cpufreq/qcom-cpufreq-hw.c
584
static void qcom_cpufreq_ready(struct cpufreq_policy *policy)
drivers/cpufreq/qcom-cpufreq-hw.c
586
struct qcom_cpufreq_data *data = policy->driver_data;
drivers/cpufreq/qcom-cpufreq-hw.c
614
return __qcom_cpufreq_hw_get(data->policy) * HZ_PER_KHZ;
drivers/cpufreq/qcom-cpufreq-hw.c
72
static int qcom_cpufreq_set_bw(struct cpufreq_policy *policy,
drivers/cpufreq/qcom-cpufreq-hw.c
80
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/qoriq-cpufreq.c
155
static int qoriq_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/qoriq-cpufreq.c
164
unsigned int cpu = policy->cpu;
drivers/cpufreq/qoriq-cpufreq.c
175
policy->clk = of_clk_get(np, 0);
drivers/cpufreq/qoriq-cpufreq.c
176
if (IS_ERR(policy->clk)) {
drivers/cpufreq/qoriq-cpufreq.c
181
hwclk = __clk_get_hw(policy->clk);
drivers/cpufreq/qoriq-cpufreq.c
202
policy->freq_table = table;
drivers/cpufreq/qoriq-cpufreq.c
206
set_affected_cpus(policy);
drivers/cpufreq/qoriq-cpufreq.c
207
policy->driver_data = data;
drivers/cpufreq/qoriq-cpufreq.c
212
policy->cpuinfo.transition_latency = u64temp + 1;
drivers/cpufreq/qoriq-cpufreq.c
228
static void qoriq_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/qoriq-cpufreq.c
230
struct cpu_data *data = policy->driver_data;
drivers/cpufreq/qoriq-cpufreq.c
235
policy->driver_data = NULL;
drivers/cpufreq/qoriq-cpufreq.c
238
static int qoriq_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/qoriq-cpufreq.c
242
struct cpu_data *data = policy->driver_data;
drivers/cpufreq/qoriq-cpufreq.c
245
return clk_set_parent(policy->clk, parent);
drivers/cpufreq/qoriq-cpufreq.c
86
static void set_affected_cpus(struct cpufreq_policy *policy)
drivers/cpufreq/qoriq-cpufreq.c
88
struct cpumask *dstp = policy->cpus;
drivers/cpufreq/qoriq-cpufreq.c
99
if (clk_is_match(policy->clk, clk))
drivers/cpufreq/s3c64xx-cpufreq.c
102
clk_get_rate(policy->clk) / 1000);
drivers/cpufreq/s3c64xx-cpufreq.c
146
static int s3c64xx_cpufreq_driver_init(struct cpufreq_policy *policy)
drivers/cpufreq/s3c64xx-cpufreq.c
150
if (policy->cpu != 0)
drivers/cpufreq/s3c64xx-cpufreq.c
153
policy->clk = clk_get(NULL, "armclk");
drivers/cpufreq/s3c64xx-cpufreq.c
154
if (IS_ERR(policy->clk)) {
drivers/cpufreq/s3c64xx-cpufreq.c
156
PTR_ERR(policy->clk));
drivers/cpufreq/s3c64xx-cpufreq.c
157
return PTR_ERR(policy->clk);
drivers/cpufreq/s3c64xx-cpufreq.c
175
r = clk_round_rate(policy->clk, freq->frequency * 1000);
drivers/cpufreq/s3c64xx-cpufreq.c
185
if (!vddarm && freq->frequency > clk_get_rate(policy->clk) / 1000)
drivers/cpufreq/s3c64xx-cpufreq.c
193
cpufreq_generic_init(policy, s3c64xx_freq_table,
drivers/cpufreq/s3c64xx-cpufreq.c
53
static int s3c64xx_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/s3c64xx-cpufreq.c
63
old_freq = clk_get_rate(policy->clk) / 1000;
drivers/cpufreq/s3c64xx-cpufreq.c
78
ret = clk_set_rate(policy->clk, new_freq * 1000);
drivers/cpufreq/s3c64xx-cpufreq.c
93
if (clk_set_rate(policy->clk, old_freq * 1000) < 0)
drivers/cpufreq/s5pv210-cpufreq.c
223
static int s5pv210_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/s5pv210-cpufreq.c
242
old_freq = policy->cur;
drivers/cpufreq/s5pv210-cpufreq.c
246
priv_index = cpufreq_table_find_index_h(policy, old_freq, false);
drivers/cpufreq/s5pv210-cpufreq.c
498
static int s5pv210_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/s5pv210-cpufreq.c
503
policy->clk = clk_get(NULL, "armclk");
drivers/cpufreq/s5pv210-cpufreq.c
504
if (IS_ERR(policy->clk))
drivers/cpufreq/s5pv210-cpufreq.c
505
return PTR_ERR(policy->clk);
drivers/cpufreq/s5pv210-cpufreq.c
519
if (policy->cpu != 0) {
drivers/cpufreq/s5pv210-cpufreq.c
543
policy->suspend_freq = SLEEP_FREQ;
drivers/cpufreq/s5pv210-cpufreq.c
544
cpufreq_generic_init(policy, s5pv210_freq_table, 40000);
drivers/cpufreq/s5pv210-cpufreq.c
552
clk_put(policy->clk);
drivers/cpufreq/s5pv210-cpufreq.c
559
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(0);
drivers/cpufreq/s5pv210-cpufreq.c
562
if (!policy) {
drivers/cpufreq/s5pv210-cpufreq.c
567
ret = cpufreq_driver_target(policy, SLEEP_FREQ, 0);
drivers/cpufreq/sa1110-cpufreq.c
229
static int sa1110_target(struct cpufreq_policy *policy, unsigned int ppcr)
drivers/cpufreq/sa1110-cpufreq.c
244
if (policy->max < 147500) {
drivers/cpufreq/sa1110-cpufreq.c
304
static int __init sa1110_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/sa1110-cpufreq.c
306
cpufreq_generic_init(policy, sa11x0_freq_table, 0);
drivers/cpufreq/sc520_freq.c
52
static int sc520_freq_target(struct cpufreq_policy *policy, unsigned int state)
drivers/cpufreq/sc520_freq.c
71
static int sc520_freq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/sc520_freq.c
81
policy->cpuinfo.transition_latency = 1000000; /* 1ms */
drivers/cpufreq/sc520_freq.c
82
policy->freq_table = sc520_freq_table;
drivers/cpufreq/scmi-cpufreq.c
198
static int scmi_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/scmi-cpufreq.c
207
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/scmi-cpufreq.c
209
pr_err("failed to get cpu%d device\n", policy->cpu);
drivers/cpufreq/scmi-cpufreq.c
227
ret = scmi_get_sharing_cpus(cpu_dev, domain, policy->cpus);
drivers/cpufreq/scmi-cpufreq.c
245
cpumask_copy(priv->opp_shared_cpus, policy->cpus);
drivers/cpufreq/scmi-cpufreq.c
290
policy->driver_data = priv;
drivers/cpufreq/scmi-cpufreq.c
291
policy->freq_table = freq_table;
drivers/cpufreq/scmi-cpufreq.c
294
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/scmi-cpufreq.c
300
policy->cpuinfo.transition_latency = latency;
drivers/cpufreq/scmi-cpufreq.c
302
policy->fast_switch_possible =
drivers/cpufreq/scmi-cpufreq.c
305
policy->transition_delay_us =
drivers/cpufreq/scmi-cpufreq.c
306
scmi_get_rate_limit(domain, policy->fast_switch_possible);
drivers/cpufreq/scmi-cpufreq.c
308
ret = freq_qos_add_request(&policy->constraints, &priv->limits_freq_req, FREQ_QOS_MAX,
drivers/cpufreq/scmi-cpufreq.c
341
static void scmi_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/scmi-cpufreq.c
343
struct scmi_data *priv = policy->driver_data;
drivers/cpufreq/scmi-cpufreq.c
351
dev_pm_opp_free_cpufreq_table(priv->cpu_dev, &policy->freq_table);
drivers/cpufreq/scmi-cpufreq.c
357
static void scmi_cpufreq_register_em(struct cpufreq_policy *policy)
drivers/cpufreq/scmi-cpufreq.c
361
struct scmi_data *priv = policy->driver_data;
drivers/cpufreq/scmi-cpufreq.c
378
em_dev_register_perf_domain(get_cpu_device(policy->cpu), priv->nr_opp,
drivers/cpufreq/scmi-cpufreq.c
41
struct cpufreq_policy *policy;
drivers/cpufreq/scmi-cpufreq.c
46
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/scmi-cpufreq.c
47
if (unlikely(!policy))
drivers/cpufreq/scmi-cpufreq.c
50
priv = policy->driver_data;
drivers/cpufreq/scmi-cpufreq.c
64
scmi_cpufreq_set_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/scmi-cpufreq.c
66
struct scmi_data *priv = policy->driver_data;
drivers/cpufreq/scmi-cpufreq.c
67
u64 freq = policy->freq_table[index].frequency;
drivers/cpufreq/scmi-cpufreq.c
72
static unsigned int scmi_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/scmi-cpufreq.c
75
struct scmi_data *priv = policy->driver_data;
drivers/cpufreq/scpi-cpufreq.c
101
pr_err("failed to get cpu%d device\n", policy->cpu);
drivers/cpufreq/scpi-cpufreq.c
111
ret = scpi_get_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/scpi-cpufreq.c
117
ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/scpi-cpufreq.c
152
policy->driver_data = priv;
drivers/cpufreq/scpi-cpufreq.c
153
policy->freq_table = freq_table;
drivers/cpufreq/scpi-cpufreq.c
156
policy->dvfs_possible_from_any_cpu = true;
drivers/cpufreq/scpi-cpufreq.c
162
policy->cpuinfo.transition_latency = latency;
drivers/cpufreq/scpi-cpufreq.c
164
policy->fast_switch_possible = false;
drivers/cpufreq/scpi-cpufreq.c
178
static void scpi_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/scpi-cpufreq.c
180
struct scpi_data *priv = policy->driver_data;
drivers/cpufreq/scpi-cpufreq.c
183
dev_pm_opp_free_cpufreq_table(priv->cpu_dev, &policy->freq_table);
drivers/cpufreq/scpi-cpufreq.c
32
struct cpufreq_policy *policy;
drivers/cpufreq/scpi-cpufreq.c
36
policy = cpufreq_cpu_get_raw(cpu);
drivers/cpufreq/scpi-cpufreq.c
37
if (unlikely(!policy))
drivers/cpufreq/scpi-cpufreq.c
40
priv = policy->driver_data;
drivers/cpufreq/scpi-cpufreq.c
47
scpi_cpufreq_set_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/scpi-cpufreq.c
49
unsigned long freq_khz = policy->freq_table[index].frequency;
drivers/cpufreq/scpi-cpufreq.c
50
struct scpi_data *priv = policy->driver_data;
drivers/cpufreq/scpi-cpufreq.c
91
static int scpi_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/scpi-cpufreq.c
99
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/sh-cpufreq.c
101
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/sh-cpufreq.c
105
static int sh_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/sh-cpufreq.c
107
unsigned int cpu = policy->cpu;
drivers/cpufreq/sh-cpufreq.c
122
policy->freq_table = freq_table;
drivers/cpufreq/sh-cpufreq.c
127
policy->min = policy->cpuinfo.min_freq =
drivers/cpufreq/sh-cpufreq.c
129
policy->max = policy->cpuinfo.max_freq =
drivers/cpufreq/sh-cpufreq.c
136
static void sh_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/sh-cpufreq.c
138
unsigned int cpu = policy->cpu;
drivers/cpufreq/sh-cpufreq.c
33
struct cpufreq_policy *policy;
drivers/cpufreq/sh-cpufreq.c
45
struct cpufreq_policy *policy = target->policy;
drivers/cpufreq/sh-cpufreq.c
46
int cpu = policy->cpu;
drivers/cpufreq/sh-cpufreq.c
60
if (freq < (policy->min * 1000) || freq > (policy->max * 1000))
drivers/cpufreq/sh-cpufreq.c
69
cpufreq_freq_transition_begin(target->policy, &freqs);
drivers/cpufreq/sh-cpufreq.c
71
cpufreq_freq_transition_end(target->policy, &freqs, 0);
drivers/cpufreq/sh-cpufreq.c
80
static int sh_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/sh-cpufreq.c
84
struct cpufreq_target data = { .policy = policy, .freq = target_freq };
drivers/cpufreq/sh-cpufreq.c
86
return work_on_cpu(policy->cpu, __sh_cpufreq_target, &data);
drivers/cpufreq/sh-cpufreq.c
89
static int sh_cpufreq_verify(struct cpufreq_policy_data *policy)
drivers/cpufreq/sh-cpufreq.c
91
struct clk *cpuclk = &per_cpu(sh_cpuclk, policy->cpu);
drivers/cpufreq/sh-cpufreq.c
93
if (policy->freq_table)
drivers/cpufreq/sh-cpufreq.c
94
return cpufreq_frequency_table_verify(policy);
drivers/cpufreq/sh-cpufreq.c
96
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/sh-cpufreq.c
98
policy->min = (clk_round_rate(cpuclk, 1) + 500) / 1000;
drivers/cpufreq/sh-cpufreq.c
99
policy->max = (clk_round_rate(cpuclk, ~0UL) + 500) / 1000;
drivers/cpufreq/sparc-us2e-cpufreq.c
265
static int us2e_freq_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/sparc-us2e-cpufreq.c
267
unsigned int cpu = policy->cpu;
drivers/cpufreq/sparc-us2e-cpufreq.c
272
static int us2e_freq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/sparc-us2e-cpufreq.c
274
unsigned int cpu = policy->cpu;
drivers/cpufreq/sparc-us2e-cpufreq.c
292
policy->cpuinfo.transition_latency = 0;
drivers/cpufreq/sparc-us2e-cpufreq.c
293
policy->cur = clock_tick;
drivers/cpufreq/sparc-us2e-cpufreq.c
294
policy->freq_table = table;
drivers/cpufreq/sparc-us2e-cpufreq.c
299
static void us2e_freq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/sparc-us2e-cpufreq.c
301
us2e_freq_target(policy, 0);
drivers/cpufreq/sparc-us3-cpufreq.c
120
static int us3_freq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/sparc-us3-cpufreq.c
122
unsigned int cpu = policy->cpu;
drivers/cpufreq/sparc-us3-cpufreq.c
136
policy->cpuinfo.transition_latency = 0;
drivers/cpufreq/sparc-us3-cpufreq.c
137
policy->cur = clock_tick;
drivers/cpufreq/sparc-us3-cpufreq.c
138
policy->freq_table = table;
drivers/cpufreq/sparc-us3-cpufreq.c
143
static void us3_freq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/sparc-us3-cpufreq.c
145
us3_freq_target(policy, 0);
drivers/cpufreq/sparc-us3-cpufreq.c
93
static int us3_freq_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/sparc-us3-cpufreq.c
95
unsigned int cpu = policy->cpu;
drivers/cpufreq/spear-cpufreq.c
103
static int spear_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/spear-cpufreq.c
153
static int spear_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/spear-cpufreq.c
155
policy->clk = spear_cpufreq.clk;
drivers/cpufreq/spear-cpufreq.c
156
cpufreq_generic_init(policy, spear_cpufreq.freq_tbl,
drivers/cpufreq/speedstep-centrino.c
234
static int centrino_cpu_init_table(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-centrino.c
236
struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
drivers/cpufreq/speedstep-centrino.c
261
per_cpu(centrino_model, policy->cpu) = model;
drivers/cpufreq/speedstep-centrino.c
270
static inline int centrino_cpu_init_table(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-centrino.c
345
static int centrino_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-centrino.c
347
struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
drivers/cpufreq/speedstep-centrino.c
359
if (policy->cpu != 0)
drivers/cpufreq/speedstep-centrino.c
367
per_cpu(centrino_cpu, policy->cpu) = &cpu_ids[i];
drivers/cpufreq/speedstep-centrino.c
369
if (!per_cpu(centrino_cpu, policy->cpu)) {
drivers/cpufreq/speedstep-centrino.c
376
if (centrino_cpu_init_table(policy))
drivers/cpufreq/speedstep-centrino.c
396
policy->cpuinfo.transition_latency = 10000;
drivers/cpufreq/speedstep-centrino.c
398
policy->freq_table = per_cpu(centrino_model, policy->cpu)->op_points;
drivers/cpufreq/speedstep-centrino.c
403
static void centrino_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-centrino.c
405
unsigned int cpu = policy->cpu;
drivers/cpufreq/speedstep-centrino.c
418
static int centrino_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/speedstep-centrino.c
420
unsigned int msr, oldmsr = 0, h = 0, cpu = policy->cpu;
drivers/cpufreq/speedstep-centrino.c
436
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/speedstep-centrino.c
443
if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY)
drivers/cpufreq/speedstep-centrino.c
444
good_cpu = cpumask_any_and(policy->cpus,
drivers/cpufreq/speedstep-centrino.c
478
if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY)
drivers/cpufreq/speedstep-ich.c
259
static int speedstep_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/speedstep-ich.c
263
policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
drivers/cpufreq/speedstep-ich.c
273
struct cpufreq_policy *policy;
drivers/cpufreq/speedstep-ich.c
285
&get_freqs->policy->cpuinfo.transition_latency,
drivers/cpufreq/speedstep-ich.c
289
static int speedstep_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-ich.c
296
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
drivers/cpufreq/speedstep-ich.c
298
policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
drivers/cpufreq/speedstep-ich.c
301
gf.policy = policy;
drivers/cpufreq/speedstep-ich.c
306
policy->freq_table = speedstep_freqs;
drivers/cpufreq/speedstep-smi.c
221
static int speedstep_target(struct cpufreq_policy *policy, unsigned int index)
drivers/cpufreq/speedstep-smi.c
229
static int speedstep_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/speedstep-smi.c
235
if (policy->cpu != 0)
drivers/cpufreq/speedstep-smi.c
267
policy->freq_table = speedstep_freqs;
drivers/cpufreq/speedstep-smi.c
280
static int speedstep_resume(struct cpufreq_policy *policy)
drivers/cpufreq/tegra186-cpufreq.c
108
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/tegra186-cpufreq.c
110
pr_err("%s: failed to get cpu%d device\n", __func__, policy->cpu);
drivers/cpufreq/tegra186-cpufreq.c
166
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/tegra186-cpufreq.c
169
tegra_cpufreq_set_bw(policy, freq_table[j - 1].frequency);
drivers/cpufreq/tegra186-cpufreq.c
174
static int tegra186_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/tegra186-cpufreq.c
177
unsigned int cluster = data->cpus[policy->cpu].bpmp_cluster_id;
drivers/cpufreq/tegra186-cpufreq.c
183
policy->cpuinfo.transition_latency = 300 * 1000;
drivers/cpufreq/tegra186-cpufreq.c
184
policy->driver_data = NULL;
drivers/cpufreq/tegra186-cpufreq.c
189
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/tegra186-cpufreq.c
195
ret = tegra_cpufreq_init_cpufreq_table(policy, bpmp_lut, &freq_table);
drivers/cpufreq/tegra186-cpufreq.c
197
policy->freq_table = freq_table;
drivers/cpufreq/tegra186-cpufreq.c
203
policy->freq_table = bpmp_lut;
drivers/cpufreq/tegra186-cpufreq.c
209
static int tegra186_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/tegra186-cpufreq.c
213
struct cpufreq_frequency_table *tbl = policy->freq_table + index;
drivers/cpufreq/tegra186-cpufreq.c
218
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/tegra186-cpufreq.c
224
tegra_cpufreq_set_bw(policy, tbl->frequency);
drivers/cpufreq/tegra186-cpufreq.c
232
struct cpufreq_policy *policy __free(put_cpufreq_policy) = cpufreq_cpu_get(cpu);
drivers/cpufreq/tegra186-cpufreq.c
238
if (!policy)
drivers/cpufreq/tegra186-cpufreq.c
241
edvd_offset = data->cpus[policy->cpu].edvd_offset;
drivers/cpufreq/tegra186-cpufreq.c
243
cluster_id = data->cpus[policy->cpu].bpmp_cluster_id;
drivers/cpufreq/tegra186-cpufreq.c
74
static int tegra_cpufreq_set_bw(struct cpufreq_policy *policy, unsigned long freq_khz)
drivers/cpufreq/tegra186-cpufreq.c
80
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/tegra186-cpufreq.c
96
static int tegra_cpufreq_init_cpufreq_table(struct cpufreq_policy *policy,
drivers/cpufreq/tegra194-cpufreq.c
133
static void tegra234_set_cpu_ndiv(struct cpufreq_policy *policy, u64 ndiv)
drivers/cpufreq/tegra194-cpufreq.c
138
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/tegra194-cpufreq.c
380
static void tegra194_set_cpu_ndiv(struct cpufreq_policy *policy, u64 ndiv)
drivers/cpufreq/tegra194-cpufreq.c
382
on_each_cpu_mask(policy->cpus, tegra194_set_cpu_ndiv_sysreg, &ndiv, true);
drivers/cpufreq/tegra194-cpufreq.c
423
static int tegra_cpufreq_init_cpufreq_table(struct cpufreq_policy *policy,
drivers/cpufreq/tegra194-cpufreq.c
436
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/tegra194-cpufreq.c
438
pr_err("%s: failed to get cpu%d device\n", __func__, policy->cpu);
drivers/cpufreq/tegra194-cpufreq.c
495
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/tegra194-cpufreq.c
500
static int tegra194_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/tegra194-cpufreq.c
504
u32 clusterid = data->cpu_data[policy->cpu].clusterid;
drivers/cpufreq/tegra194-cpufreq.c
513
start_cpu = rounddown(policy->cpu, maxcpus_per_cluster);
drivers/cpufreq/tegra194-cpufreq.c
517
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/tegra194-cpufreq.c
519
policy->cpuinfo.transition_latency = TEGRA_CPUFREQ_TRANSITION_LATENCY;
drivers/cpufreq/tegra194-cpufreq.c
524
ret = tegra_cpufreq_init_cpufreq_table(policy, bpmp_lut, &freq_table);
drivers/cpufreq/tegra194-cpufreq.c
526
policy->freq_table = freq_table;
drivers/cpufreq/tegra194-cpufreq.c
532
policy->freq_table = bpmp_lut;
drivers/cpufreq/tegra194-cpufreq.c
538
static int tegra194_cpufreq_online(struct cpufreq_policy *policy)
drivers/cpufreq/tegra194-cpufreq.c
544
static int tegra194_cpufreq_offline(struct cpufreq_policy *policy)
drivers/cpufreq/tegra194-cpufreq.c
554
static void tegra194_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/tegra194-cpufreq.c
556
struct device *cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/tegra194-cpufreq.c
559
dev_pm_opp_of_cpumask_remove_table(policy->related_cpus);
drivers/cpufreq/tegra194-cpufreq.c
562
static int tegra194_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/tegra194-cpufreq.c
565
struct cpufreq_frequency_table *tbl = policy->freq_table + index;
drivers/cpufreq/tegra194-cpufreq.c
573
data->soc->ops->set_cpu_ndiv(policy, (u64)tbl->driver_data);
drivers/cpufreq/tegra194-cpufreq.c
576
tegra_cpufreq_set_bw(policy, tbl->frequency);
drivers/cpufreq/tegra194-cpufreq.c
61
void (*set_cpu_ndiv)(struct cpufreq_policy *policy, u64 ndiv);
drivers/cpufreq/tegra194-cpufreq.c
84
static int tegra_cpufreq_set_bw(struct cpufreq_policy *policy, unsigned long freq_khz)
drivers/cpufreq/tegra194-cpufreq.c
91
dev = get_cpu_device(policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
177
static int ve_spc_cpufreq_set_target(struct cpufreq_policy *policy,
drivers/cpufreq/vexpress-spc-cpufreq.c
180
u32 cpu = policy->cpu, cur_cluster, new_cluster, actual_cluster;
drivers/cpufreq/vexpress-spc-cpufreq.c
410
static int ve_spc_cpufreq_init(struct cpufreq_policy *policy)
drivers/cpufreq/vexpress-spc-cpufreq.c
412
u32 cur_cluster = cpu_to_cluster(policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
416
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
419
policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
426
dev_pm_opp_get_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/vexpress-spc-cpufreq.c
428
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/vexpress-spc-cpufreq.c
432
per_cpu(physical_cluster, policy->cpu) = A15_CLUSTER;
drivers/cpufreq/vexpress-spc-cpufreq.c
435
ret = get_cluster_clk_and_freq_table(cpu_dev, policy->cpus);
drivers/cpufreq/vexpress-spc-cpufreq.c
439
policy->freq_table = freq_table[cur_cluster];
drivers/cpufreq/vexpress-spc-cpufreq.c
440
policy->cpuinfo.transition_latency = 1000000; /* 1 ms */
drivers/cpufreq/vexpress-spc-cpufreq.c
443
per_cpu(cpu_last_req_freq, policy->cpu) =
drivers/cpufreq/vexpress-spc-cpufreq.c
444
clk_get_cpu_rate(policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
446
dev_info(cpu_dev, "%s: CPU %d initialized\n", __func__, policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
450
static void ve_spc_cpufreq_exit(struct cpufreq_policy *policy)
drivers/cpufreq/vexpress-spc-cpufreq.c
454
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
457
policy->cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
461
put_cluster_clk_and_freq_table(cpu_dev, policy->related_cpus);
drivers/cpufreq/virtual-cpufreq.c
100
static unsigned int virt_cpufreq_fast_switch(struct cpufreq_policy *policy,
drivers/cpufreq/virtual-cpufreq.c
103
virt_cpufreq_set_perf(policy, target_freq);
drivers/cpufreq/virtual-cpufreq.c
115
static int virt_cpufreq_target(struct cpufreq_policy *policy,
drivers/cpufreq/virtual-cpufreq.c
122
freqs.old = policy->cur;
drivers/cpufreq/virtual-cpufreq.c
125
cpufreq_freq_transition_begin(policy, &freqs);
drivers/cpufreq/virtual-cpufreq.c
126
ret = virt_cpufreq_set_perf(policy, target_freq);
drivers/cpufreq/virtual-cpufreq.c
127
cpufreq_freq_transition_end(policy, &freqs, ret != 0);
drivers/cpufreq/virtual-cpufreq.c
132
static int virt_cpufreq_get_sharing_cpus(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
138
cur_perf_domain = readl_relaxed(base + policy->cpu *
drivers/cpufreq/virtual-cpufreq.c
150
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/virtual-cpufreq.c
156
static int virt_cpufreq_get_freq_info(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
161
num_perftbl_entries = per_cpu(perftbl_num_entries, policy->cpu);
drivers/cpufreq/virtual-cpufreq.c
164
policy->cpuinfo.min_freq = 1;
drivers/cpufreq/virtual-cpufreq.c
165
policy->cpuinfo.max_freq = virt_cpufreq_get_perftbl_entry(policy->cpu, 0);
drivers/cpufreq/virtual-cpufreq.c
167
policy->min = policy->cpuinfo.min_freq;
drivers/cpufreq/virtual-cpufreq.c
168
policy->max = policy->cpuinfo.max_freq;
drivers/cpufreq/virtual-cpufreq.c
170
policy->cur = policy->max;
drivers/cpufreq/virtual-cpufreq.c
179
table[idx].frequency = virt_cpufreq_get_perftbl_entry(policy->cpu, idx);
drivers/cpufreq/virtual-cpufreq.c
182
policy->freq_table = table;
drivers/cpufreq/virtual-cpufreq.c
187
static int virt_cpufreq_cpu_init(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
192
cpu_dev = get_cpu_device(policy->cpu);
drivers/cpufreq/virtual-cpufreq.c
196
ret = virt_cpufreq_get_freq_info(policy);
drivers/cpufreq/virtual-cpufreq.c
202
ret = virt_cpufreq_get_sharing_cpus(policy);
drivers/cpufreq/virtual-cpufreq.c
216
policy->dvfs_possible_from_any_cpu = false;
drivers/cpufreq/virtual-cpufreq.c
217
policy->fast_switch_possible = true;
drivers/cpufreq/virtual-cpufreq.c
227
topology_set_scale_freq_source(&virt_sfd, policy->cpus);
drivers/cpufreq/virtual-cpufreq.c
232
static void virt_cpufreq_cpu_exit(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
234
topology_clear_scale_freq_source(SCALE_FREQ_SOURCE_VIRT, policy->related_cpus);
drivers/cpufreq/virtual-cpufreq.c
235
kfree(policy->freq_table);
drivers/cpufreq/virtual-cpufreq.c
238
static int virt_cpufreq_online(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
244
static int virt_cpufreq_offline(struct cpufreq_policy *policy)
drivers/cpufreq/virtual-cpufreq.c
250
static int virt_cpufreq_verify_policy(struct cpufreq_policy_data *policy)
drivers/cpufreq/virtual-cpufreq.c
252
if (policy->freq_table)
drivers/cpufreq/virtual-cpufreq.c
253
return cpufreq_frequency_table_verify(policy);
drivers/cpufreq/virtual-cpufreq.c
255
cpufreq_verify_within_cpu_limits(policy);
drivers/cpufreq/virtual-cpufreq.c
92
static unsigned int virt_cpufreq_set_perf(struct cpufreq_policy *policy,
drivers/cpufreq/virtual-cpufreq.c
96
base + policy->cpu * PER_CPU_OFFSET + REG_SET_PERF_STATE_OFFSET);
drivers/cxl/pci.c
623
struct cxl_event_interrupt_policy *policy)
drivers/cxl/pci.c
628
.payload_out = policy,
drivers/cxl/pci.c
629
.size_out = sizeof(*policy),
drivers/cxl/pci.c
642
struct cxl_event_interrupt_policy *policy)
drivers/cxl/pci.c
648
*policy = (struct cxl_event_interrupt_policy) {
drivers/cxl/pci.c
657
.payload_in = policy,
drivers/cxl/pci.c
658
.size_in = sizeof(*policy),
drivers/cxl/pci.c
669
return cxl_event_get_int_policy(mds, policy);
drivers/cxl/pci.c
675
struct cxl_event_interrupt_policy policy;
drivers/cxl/pci.c
678
rc = cxl_event_config_msgnums(mds, &policy);
drivers/cxl/pci.c
682
rc = cxl_event_req_irq(cxlds, policy.info_settings);
drivers/cxl/pci.c
688
rc = cxl_event_req_irq(cxlds, policy.warn_settings);
drivers/cxl/pci.c
694
rc = cxl_event_req_irq(cxlds, policy.failure_settings);
drivers/cxl/pci.c
700
rc = cxl_event_req_irq(cxlds, policy.fatal_settings);
drivers/cxl/pci.c
719
struct cxl_event_interrupt_policy policy;
drivers/cxl/pci.c
734
rc = cxl_event_get_int_policy(mds, &policy);
drivers/cxl/pci.c
738
if (cxl_event_int_is_fw(policy.info_settings) ||
drivers/cxl/pci.c
739
cxl_event_int_is_fw(policy.warn_settings) ||
drivers/cxl/pci.c
740
cxl_event_int_is_fw(policy.failure_settings) ||
drivers/cxl/pci.c
741
cxl_event_int_is_fw(policy.fatal_settings)) {
drivers/devfreq/governor_passive.c
108
struct cpufreq_policy *policy;
drivers/devfreq/governor_passive.c
115
policy = cpufreq_cpu_get(cpu);
drivers/devfreq/governor_passive.c
116
if (!policy) {
drivers/devfreq/governor_passive.c
121
parent_cpu_data = get_parent_cpu_data(p_data, policy);
drivers/devfreq/governor_passive.c
123
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
134
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
149
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
240
parent_cpu_data = get_parent_cpu_data(p_data, freqs->policy);
drivers/devfreq/governor_passive.c
284
struct cpufreq_policy *policy;
drivers/devfreq/governor_passive.c
301
policy = cpufreq_cpu_get(cpu);
drivers/devfreq/governor_passive.c
302
if (!policy) {
drivers/devfreq/governor_passive.c
307
parent_cpu_data = get_parent_cpu_data(p_data, policy);
drivers/devfreq/governor_passive.c
309
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
335
parent_cpu_data->first_cpu = cpumask_first(policy->related_cpus);
drivers/devfreq/governor_passive.c
336
parent_cpu_data->cur_freq = policy->cur;
drivers/devfreq/governor_passive.c
337
parent_cpu_data->min_freq = policy->cpuinfo.min_freq;
drivers/devfreq/governor_passive.c
338
parent_cpu_data->max_freq = policy->cpuinfo.max_freq;
drivers/devfreq/governor_passive.c
341
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
355
cpufreq_cpu_put(policy);
drivers/devfreq/governor_passive.c
47
struct cpufreq_policy *policy)
drivers/devfreq/governor_passive.c
51
if (!p_data || !policy)
drivers/devfreq/governor_passive.c
55
if (parent_cpu_data->first_cpu == cpumask_first(policy->related_cpus))
drivers/dpll/dpll_nl.c
103
.policy = dpll_device_get_nl_policy,
drivers/dpll/dpll_nl.c
117
.policy = dpll_device_set_nl_policy,
drivers/dpll/dpll_nl.c
126
.policy = dpll_pin_id_get_nl_policy,
drivers/dpll/dpll_nl.c
135
.policy = dpll_pin_get_do_nl_policy,
drivers/dpll/dpll_nl.c
142
.policy = dpll_pin_get_dump_nl_policy,
drivers/dpll/dpll_nl.c
151
.policy = dpll_pin_set_nl_policy,
drivers/dpll/dpll_nl.c
94
.policy = dpll_device_id_get_nl_policy,
drivers/firmware/qcom/qcom_qseecom_uefisecapp.c
819
pool_config.policy = QCOM_TZMEM_POLICY_MULTIPLIER;
drivers/firmware/qcom/qcom_scm.c
2741
pool_config.policy = QCOM_TZMEM_POLICY_ON_DEMAND;
drivers/firmware/qcom/qcom_tzmem.c
250
switch (config->policy) {
drivers/firmware/qcom/qcom_tzmem.c
275
pool->policy = config->policy;
drivers/firmware/qcom/qcom_tzmem.c
37
enum qcom_tzmem_policy policy;
drivers/firmware/qcom/qcom_tzmem.c
382
switch (pool->policy) {
drivers/gpu/drm/amd/display/dc/core/dc.c
3463
struct pipe_split_policy_backup *policy)
drivers/gpu/drm/amd/display/dc/core/dc.c
3468
policy->mpc_policy = dc->debug.pipe_split_policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3471
policy->dynamic_odm_policy = dc->debug.enable_single_display_2to1_odm_policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3473
policy->subvp_policy = dc->debug.force_disable_subvp;
drivers/gpu/drm/amd/display/dc/core/dc.c
3476
policy->force_odm[i] = context->streams[i]->debug.force_odm_combine_segments;
drivers/gpu/drm/amd/display/dc/core/dc.c
3484
struct pipe_split_policy_backup *policy)
drivers/gpu/drm/amd/display/dc/core/dc.c
3489
dc->debug.pipe_split_policy = policy->mpc_policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3491
policy->dynamic_odm_policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3492
dc->debug.force_disable_subvp = policy->subvp_policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3494
context->streams[i]->debug.force_odm_combine_segments = policy->force_odm[i];
drivers/gpu/drm/amd/display/dc/core/dc.c
3647
struct pipe_split_policy_backup policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
3660
backup_and_set_minimal_pipe_split_policy(dc, context, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
3664
restore_minimal_pipe_split_policy(dc, context, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
3670
restore_minimal_pipe_split_policy(dc, context, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4713
struct pipe_split_policy_backup *policy)
drivers/gpu/drm/amd/display/dc/core/dc.c
4715
restore_minimal_pipe_split_policy(dc, base_context, policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4734
struct dc_state *base_context, struct pipe_split_policy_backup *policy)
drivers/gpu/drm/amd/display/dc/core/dc.c
4742
backup_and_set_minimal_pipe_split_policy(dc, base_context, policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4753
release_minimal_transition_state(dc, minimal_transition_context, base_context, policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4823
struct pipe_split_policy_backup policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
4826
&policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4844
dc, intermediate_context, new_context, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4853
struct pipe_split_policy_backup policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
4873
old_current_state, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4893
old_current_state, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
4962
struct pipe_split_policy_backup policy;
drivers/gpu/drm/amd/display/dc/core/dc.c
5021
transition_base_context, &policy);
drivers/gpu/drm/amd/display/dc/core/dc.c
5024
release_minimal_transition_state(dc, transition_context, transition_base_context, &policy);
drivers/gpu/drm/amd/display/dc/dc_dsc.h
109
struct dc_dsc_policy *policy,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10067
mode_lib->ms.policy = mode_lib->policy;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6224
CalculatePrefetchSchedule_params->EnhancedPrefetchScheduleAccelerationFinal = mode_lib->ms.policy.EnhancedPrefetchScheduleAccelerationFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6344
CalculatePrefetchMode(mode_lib->ms.policy.AllowForPStateChangeOrStutterInVBlank[k], &s->MinPrefetchMode[k], &s->MaxPrefetchMode[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6357
mode_lib->ms.policy.SynchronizeDRRDisplaysForUCLKPStateChangeFinal,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6393
dml_print("DML::%s: AllowForPStateChangeOrStutterInVBlank = %u\n", __func__, mode_lib->ms.policy.AllowForPStateChangeOrStutterInVBlank[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6498
((s->MaxVStartup < s->MaximumVStartup[j][k] || mode_lib->ms.policy.EnhancedPrefetchScheduleAccelerationFinal == 0) &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6531
if (!(mode_lib->ms.policy.ImmediateFlipRequirement[k] == dml_immediate_flip_not_required)) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6579
mode_lib->ms.policy.ImmediateFlipRequirement,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6606
if (!(mode_lib->ms.policy.ImmediateFlipRequirement[k] == dml_immediate_flip_not_required) && (mode_lib->ms.ImmediateFlipSupportedForPipe[k] == false))
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6648
CalculateWatermarks_params->USRRetrainingRequiredFinal = mode_lib->ms.policy.USRRetrainingRequiredFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6657
CalculateWatermarks_params->SynchronizeTimingsFinal = mode_lib->ms.policy.SynchronizeTimingsFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6658
CalculateWatermarks_params->SynchronizeDRRDisplaysForUCLKPStateChangeFinal = mode_lib->ms.policy.SynchronizeDRRDisplaysForUCLKPStateChangeFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6802
mode_lib->ms.policy.NomDETInKByteOverrideEnable, // VBA_DELTA
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6803
mode_lib->ms.policy.NomDETInKByteOverrideValue, // VBA_DELTA
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7048
CalculateSwathAndDETConfiguration_params->UseUnboundedRequestingFinal = mode_lib->ms.policy.UseUnboundedRequesting;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7106
if (mode_lib->ms.policy.MPCCombineUse[k] == dml_mpc_as_needed_for_pstate_and_voltage)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7108
if (mode_lib->ms.policy.MPCCombineUse[k] == dml_mpc_as_possible)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7123
mode_lib->ms.policy.ODMUse[k],
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7146
mode_lib->ms.policy.ODMUse[k],
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7216
} else if (mode_lib->ms.policy.MPCCombineUse[k] == dml_mpc_disabled) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7249
if (j == 1 && !UnboundedRequest(mode_lib->ms.policy.UseUnboundedRequesting,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7256
if (mode_lib->ms.policy.MPCCombineUse[k] != dml_mpc_disabled && mode_lib->ms.policy.MPCCombineUse[k] != dml_mpc_as_needed_for_voltage &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7425
if (mode_lib->ms.cache_display_cfg.output.OutputEncoder[k] != dml_dp && (mode_lib->ms.policy.ODMUse[k] == dml_odm_use_policy_split_1to2 ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7426
mode_lib->ms.policy.ODMUse[k] == dml_odm_use_policy_mso_1to2 || mode_lib->ms.policy.ODMUse[k] == dml_odm_use_policy_mso_1to4))
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7429
if ((mode_lib->ms.policy.ODMUse[k] == dml_odm_use_policy_mso_1to2 && mode_lib->ms.cache_display_cfg.output.OutputLinkDPLanes[k] < 2) ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7430
(mode_lib->ms.policy.ODMUse[k] == dml_odm_use_policy_mso_1to4 && mode_lib->ms.cache_display_cfg.output.OutputLinkDPLanes[k] < 4))
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7587
CalculateSwathAndDETConfiguration_params->UseUnboundedRequestingFinal = mode_lib->ms.policy.UseUnboundedRequesting;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7919
s->ImmediateFlipRequiredFinal = s->ImmediateFlipRequiredFinal || (mode_lib->ms.policy.ImmediateFlipRequirement[k] == dml_immediate_flip_required);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7925
((mode_lib->ms.policy.ImmediateFlipRequirement[k] != dml_immediate_flip_required) &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7926
(mode_lib->ms.policy.ImmediateFlipRequirement[k] != dml_immediate_flip_not_required));
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7933
mode_lib->ms.support.ImmediateFlipOrHostVMAndPStateWithMALLFullFrameOrPhantomPipe || ((mode_lib->ms.cache_display_cfg.plane.HostVMEnable == true || mode_lib->ms.policy.ImmediateFlipRequirement[k] != dml_immediate_flip_not_required) &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7962
if (mode_lib->ms.policy.UseMinimumRequiredDCFCLK == true) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7965
UseMinimumDCFCLK_params->SynchronizeDRRDisplaysForUCLKPStateChangeFinal = mode_lib->ms.policy.SynchronizeDRRDisplaysForUCLKPStateChangeFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8133
s->dram_clock_change_support = (!mode_lib->ms.policy.DRAMClockChangeRequirementFinal ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8134
(s->is_max_dram_pwr_state && mode_lib->policy.AssumeModeSupportAtMaxPwrStateEvenDRAMClockChangeNotSupported) ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8136
s->f_clock_change_support = (!mode_lib->ms.policy.FCLKChangeRequirementFinal ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8137
(s->is_max_pwr_state && mode_lib->policy.AssumeModeSupportAtMaxPwrStateEvenFClockChangeNotSupported) ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8190
&& (!mode_lib->ms.policy.USRRetrainingRequiredFinal || mode_lib->ms.support.USRRetrainingSupport[j])) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8207
(s->MPCCombineMethodAsNeededForPStateChangeAndVoltage && mode_lib->ms.policy.DRAMClockChangeRequirementFinal &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8215
|| (s->MPCCombineMethodAsNeededForPStateChangeAndVoltage && mode_lib->ms.policy.FCLKChangeRequirementFinal &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8278
mode_lib->ms.policy.NomDETInKByteOverrideEnable,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8279
mode_lib->ms.policy.NomDETInKByteOverrideValue,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8334
dml_print_dml_policy(&mode_lib->ms.policy);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8510
CalculateSwathAndDETConfiguration_params->UseUnboundedRequestingFinal = mode_lib->ms.policy.UseUnboundedRequesting;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8932
s->ImmediateFlipRequirementFinal = s->ImmediateFlipRequirementFinal || (mode_lib->ms.policy.ImmediateFlipRequirement[k] == dml_immediate_flip_required);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8945
CalculatePrefetchMode(mode_lib->ms.policy.AllowForPStateChangeOrStutterInVBlank[k], &s->MinPrefetchMode[k], &s->MaxPrefetchMode[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8963
mode_lib->ms.policy.SynchronizeDRRDisplaysForUCLKPStateChangeFinal,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8996
dml_print("DML::%s: AllowForPStateChangeOrStutterInVBlank = %u\n", __func__, mode_lib->ms.policy.AllowForPStateChangeOrStutterInVBlank[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9000
CalculatePrefetchSchedule_params->EnhancedPrefetchScheduleAccelerationFinal = mode_lib->ms.policy.EnhancedPrefetchScheduleAccelerationFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9122
((s->VStartupLines < s->MaxVStartupLines[k] || mode_lib->ms.policy.EnhancedPrefetchScheduleAccelerationFinal == 0) &&
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9240
if (mode_lib->ms.policy.ImmediateFlipRequirement[k] != dml_immediate_flip_not_required) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9294
mode_lib->ms.policy.ImmediateFlipRequirement,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9322
mode_lib->ms.policy.ImmediateFlipRequirement,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9348
if (mode_lib->ms.policy.ImmediateFlipRequirement[k] != dml_immediate_flip_not_required && locals->ImmediateFlipSupportedForPipe[k] == false) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9371
dml_print("DML::%s: ImmediateFlipRequirement[%u] = %u\n", __func__, k, mode_lib->ms.policy.ImmediateFlipRequirement[k] == dml_immediate_flip_required);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9421
CalculateWatermarks_params->USRRetrainingRequiredFinal = mode_lib->ms.policy.USRRetrainingRequiredFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9430
CalculateWatermarks_params->SynchronizeTimingsFinal = mode_lib->ms.policy.SynchronizeTimingsFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9431
CalculateWatermarks_params->SynchronizeDRRDisplaysForUCLKPStateChangeFinal = mode_lib->ms.policy.SynchronizeDRRDisplaysForUCLKPStateChangeFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9651
mode_lib->ms.policy.DCCProgrammingAssumesScanDirectionUnknownFinal,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9791
CalculateStutterEfficiency_params->SynchronizeTimingsFinal = mode_lib->ms.policy.SynchronizeTimingsFinal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9863
mode_lib->ms.policy.SynchronizeTimingsFinal,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
1901
struct dml_mode_eval_policy_st policy;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
814
struct dml_mode_eval_policy_st policy;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
327
void dml_print_dml_policy(const struct dml_mode_eval_policy_st *policy)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
331
dml_print("DML: Policy: UseUnboundedRequesting = 0x%x\n", policy->UseUnboundedRequesting);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
332
dml_print("DML: Policy: UseMinimumRequiredDCFCLK = 0x%x\n", policy->UseMinimumRequiredDCFCLK);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
333
dml_print("DML: Policy: DRAMClockChangeRequirementFinal = 0x%x\n", policy->DRAMClockChangeRequirementFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
334
dml_print("DML: Policy: FCLKChangeRequirementFinal = 0x%x\n", policy->FCLKChangeRequirementFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
335
dml_print("DML: Policy: USRRetrainingRequiredFinal = 0x%x\n", policy->USRRetrainingRequiredFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
336
dml_print("DML: Policy: EnhancedPrefetchScheduleAccelerationFinal = 0x%x\n", policy->EnhancedPrefetchScheduleAccelerationFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
337
dml_print("DML: Policy: NomDETInKByteOverrideEnable = 0x%x\n", policy->NomDETInKByteOverrideEnable);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
338
dml_print("DML: Policy: NomDETInKByteOverrideValue = 0x%x\n", policy->NomDETInKByteOverrideValue);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
339
dml_print("DML: Policy: DCCProgrammingAssumesScanDirectionUnknownFinal = 0x%x\n", policy->DCCProgrammingAssumesScanDirectionUnknownFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
340
dml_print("DML: Policy: SynchronizeTimingsFinal = 0x%x\n", policy->SynchronizeTimingsFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
341
dml_print("DML: Policy: SynchronizeDRRDisplaysForUCLKPStateChangeFinal = 0x%x\n", policy->SynchronizeDRRDisplaysForUCLKPStateChangeFinal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
342
dml_print("DML: Policy: AssumeModeSupportAtMaxPwrStateEvenDRAMClockChangeNotSupported = 0x%x\n", policy->AssumeModeSupportAtMaxPwrStateEvenDRAMClockChangeNotSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
343
dml_print("DML: Policy: AssumeModeSupportAtMaxPwrStateEvenFClockChangeNotSupported = 0x%x\n", policy->AssumeModeSupportAtMaxPwrStateEvenFClockChangeNotSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
346
dml_print("DML: i=%0d, Policy: MPCCombineUse = 0x%x\n", i, policy->MPCCombineUse[i]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
347
dml_print("DML: i=%0d, Policy: ODMUse = 0x%x\n", i, policy->ODMUse[i]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
348
dml_print("DML: i=%0d, Policy: ImmediateFlipRequirement = 0x%x\n", i, policy->ImmediateFlipRequirement[i]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
349
dml_print("DML: i=%0d, Policy: AllowForPStateChangeOrStutterInVBlank = 0x%x\n", i, policy->AllowForPStateChangeOrStutterInVBlank[i]);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
407
dml_print("DML: MODE SUPPORT: USR Retraining Support : %s\n", (!mode_lib->ms.policy.USRRetrainingRequiredFinal || &mode_lib->ms.support.USRRetrainingSupport[j]) ? "Supported" : "NOT Supported");
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.h
57
__DML_DLL_EXPORT__ void dml_print_dml_policy(const struct dml_mode_eval_policy_st *policy);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
278
void build_unoptimized_policy_settings(enum dml_project_id project, struct dml_mode_eval_policy_st *policy)
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
281
policy->MPCCombineUse[i] = dml_mpc_as_needed_for_voltage; // TOREVIEW: Is this still needed? When is MPCC useful for pstate given CRB?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
282
policy->ODMUse[i] = dml_odm_use_policy_combine_as_needed;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
283
policy->ImmediateFlipRequirement[i] = dml_immediate_flip_required;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
284
policy->AllowForPStateChangeOrStutterInVBlank[i] = dml_prefetch_support_uclk_fclk_and_stutter_if_possible;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
290
policy->UseUnboundedRequesting = dml_unbounded_requesting_enable;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
291
policy->UseMinimumRequiredDCFCLK = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
292
policy->DRAMClockChangeRequirementFinal = true; // TOREVIEW: What does this mean?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
293
policy->FCLKChangeRequirementFinal = true; // TOREVIEW: What does this mean?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
294
policy->USRRetrainingRequiredFinal = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
295
policy->EnhancedPrefetchScheduleAccelerationFinal = true; // TOREVIEW: What does this mean?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
296
policy->NomDETInKByteOverrideEnable = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
297
policy->NomDETInKByteOverrideValue = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
298
policy->DCCProgrammingAssumesScanDirectionUnknownFinal = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
299
policy->SynchronizeTimingsFinal = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
300
policy->SynchronizeDRRDisplaysForUCLKPStateChangeFinal = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
301
policy->AssumeModeSupportAtMaxPwrStateEvenDRAMClockChangeNotSupported = true; // TOREVIEW: What does this mean?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
302
policy->AssumeModeSupportAtMaxPwrStateEvenFClockChangeNotSupported = true; // TOREVIEW: What does this mean?
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
306
policy->DCCProgrammingAssumesScanDirectionUnknownFinal = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
307
policy->EnhancedPrefetchScheduleAccelerationFinal = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
308
policy->AllowForPStateChangeOrStutterInVBlankFinal = dml_prefetch_support_uclk_fclk_and_stutter_if_possible; /*new*/
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.c
309
policy->UseOnlyMaxPrefetchModes = 1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_policy.h
45
void build_unoptimized_policy_settings(enum dml_project_id project, struct dml_mode_eval_policy_st *policy);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1313
dml2->v20.dml_core_ctx.policy.AllowForPStateChangeOrStutterInVBlankFinal = dml_prefetch_support_uclk_fclk_and_stutter;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1344
dml2->v20.dml_core_ctx.policy.ODMUse[disp_cfg_stream_location] = dml_odm_use_policy_combine_2to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1347
dml2->v20.dml_core_ctx.policy.ODMUse[disp_cfg_stream_location] = dml_odm_use_policy_combine_4to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1388
dml2->v20.dml_core_ctx.policy.ImmediateFlipRequirement[disp_cfg_plane_location] = dml_immediate_flip_not_required;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1405
dml2->v20.dml_core_ctx.policy.ODMUse[disp_cfg_plane_location] = dml_odm_use_policy_combine_2to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1408
dml2->v20.dml_core_ctx.policy.ODMUse[disp_cfg_plane_location] = dml_odm_use_policy_combine_4to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
167
build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
274
build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
312
s->cur_policy = dml2->v20.dml_core_ctx.policy;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
323
dml2->v20.dml_core_ctx.policy = s->new_policy;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
338
dml2->v20.dml_core_ctx.policy = s->cur_policy;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
427
memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st));
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
521
memset(&dml2->v20.dml_core_ctx.policy, 0, sizeof(struct dml_mode_eval_policy_st));
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
525
build_unoptimized_policy_settings(dml2->v20.dml_core_ctx.project, &dml2->v20.dml_core_ctx.policy);
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1066
struct dc_dsc_policy policy;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1070
dc_dsc_get_policy_for_timing(timing, options->max_target_bpp_limit_override_x16, &policy, link_encoding);
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1185
if (policy.use_min_slices_h) {
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1189
if (policy.max_slices_h)
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1190
num_slices_h = min(policy.max_slices_h, max_slices_h);
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1197
if (policy.max_slices_h)
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1198
num_slices_h = min(policy.max_slices_h, max_slices_h);
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1243
slice_height = min(policy.min_slice_height, pic_height);
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1267
&policy,
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1364
struct dc_dsc_policy *policy,
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1369
policy->min_target_bpp = 0;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1370
policy->max_target_bpp = 0;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1373
policy->use_min_slices_h = true;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1378
policy->max_slices_h = 0;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1383
policy->min_slice_height = 108;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1406
policy->min_target_bpp = 8;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1408
policy->max_target_bpp = 3 * bpc;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1412
policy->min_target_bpp = 6;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1414
policy->max_target_bpp = bpc * 3 / 2;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1421
if (policy->max_target_bpp > dsc_policy_max_target_bpp_limit)
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1422
policy->max_target_bpp = dsc_policy_max_target_bpp_limit;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1425
if (max_target_bpp_limit_override_x16 && policy->max_target_bpp > max_target_bpp_limit_override_x16 / 16)
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1426
policy->max_target_bpp = max_target_bpp_limit_override_x16 / 16;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
1429
policy->enable_dsc_when_not_needed = dsc_policy_enable_dsc_when_not_needed;
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
854
const struct dc_dsc_policy *policy,
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
867
if (decide_dsc_bandwidth_range(policy->min_target_bpp * 16, policy->max_target_bpp * 16,
drivers/gpu/drm/amd/display/dc/dsc/dc_dsc.c
870
if (policy->enable_dsc_when_not_needed || options->force_dsc_when_not_needed)
drivers/gpu/drm/amd/display/dc/link/protocols/link_dp_capability.c
843
unsigned int policy = 0;
drivers/gpu/drm/amd/display/dc/link/protocols/link_dp_capability.c
845
policy = link->panel_config.dsc.force_dsc_edp_policy;
drivers/gpu/drm/amd/display/dc/link/protocols/link_dp_capability.c
877
if (policy) {
drivers/gpu/drm/amd/display/dc/link/protocols/link_dp_capability.c
930
if (policy) {
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1294
struct smu_dpm_policy *policy;
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1296
policy = smu_get_pm_policy(smu, PP_PM_POLICY_XGMI_PLPD);
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1298
if (policy)
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1299
policy->current_level = XGMI_PLPD_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1306
if (policy)
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
1307
policy->current_level = XGMI_PLPD_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
3768
static void smu_print_dpm_policy(struct smu_dpm_policy *policy, char *sysbuf,
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
3774
for_each_set_bit(level, &policy->level_mask, PP_POLICY_MAX_LEVELS) {
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
3775
if (level == policy->current_level)
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
3778
policy->desc->get_desc(policy, level));
drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c
3782
policy->desc->get_desc(policy, level));
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
308
struct smu_dpm_policy *policy;
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
321
policy = &(smu_dpm->dpm_policies->policies[0]);
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
322
policy->policy_type = PP_PM_POLICY_XGMI_PLPD;
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
323
policy->level_mask = BIT(XGMI_PLPD_DISALLOW) | BIT(XGMI_PLPD_DEFAULT);
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
324
policy->current_level = XGMI_PLPD_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
325
policy->set_policy = arcturus_select_plpd_policy;
drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c
326
smu_cmn_generic_plpd_policy_desc(policy);
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
295
struct smu_dpm_policy *policy;
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
308
policy = &(smu_dpm->dpm_policies->policies[0]);
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
309
policy->policy_type = PP_PM_POLICY_XGMI_PLPD;
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
310
policy->level_mask = BIT(XGMI_PLPD_DISALLOW) | BIT(XGMI_PLPD_DEFAULT);
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
311
policy->current_level = XGMI_PLPD_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
312
policy->set_policy = aldebaran_select_plpd_policy;
drivers/gpu/drm/amd/pm/swsmu/smu13/aldebaran_ppt.c
313
smu_cmn_generic_plpd_policy_desc(policy);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
1077
struct smu_dpm_policy *policy;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
1079
policy = smu_get_pm_policy(smu, PP_PM_POLICY_SOC_PSTATE);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
1080
if (policy)
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
1081
policy->current_level = SOC_PSTATE_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
620
int policy)
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
625
switch (policy) {
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
647
policy);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
690
struct smu_dpm_policy *policy;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
706
policy = &(smu_dpm->dpm_policies->policies[0]);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
708
policy->policy_type = PP_PM_POLICY_SOC_PSTATE;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
709
policy->level_mask = BIT(SOC_PSTATE_DEFAULT) |
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
712
policy->current_level = SOC_PSTATE_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
713
policy->set_policy = smu_v13_0_6_select_policy_soc_pstate;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
714
smu_cmn_generic_soc_policy_desc(policy);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
718
policy = &(smu_dpm->dpm_policies->policies[1]);
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
720
policy->policy_type = PP_PM_POLICY_XGMI_PLPD;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
721
policy->level_mask = BIT(XGMI_PLPD_DISALLOW) | BIT(XGMI_PLPD_DEFAULT) |
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
723
policy->current_level = XGMI_PLPD_DEFAULT;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
724
policy->set_policy = smu_v13_0_6_select_plpd_policy;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
725
smu_cmn_generic_plpd_policy_desc(policy);
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1197
static char *smu_soc_policy_get_desc(struct smu_dpm_policy *policy, int level)
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1199
if (level < 0 || !(policy->level_mask & BIT(level)))
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1221
void smu_cmn_generic_soc_policy_desc(struct smu_dpm_policy *policy)
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1223
policy->desc = &pstate_policy_desc;
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1226
static char *smu_xgmi_plpd_policy_get_desc(struct smu_dpm_policy *policy,
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1229
if (level < 0 || !(policy->level_mask & BIT(level)))
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1249
void smu_cmn_generic_plpd_policy_desc(struct smu_dpm_policy *policy)
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c
1251
policy->desc = &xgmi_plpd_policy_desc;
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h
191
void smu_cmn_generic_soc_policy_desc(struct smu_dpm_policy *policy);
drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h
192
void smu_cmn_generic_plpd_policy_desc(struct smu_dpm_policy *policy);
drivers/gpu/drm/i915/gt/intel_llc.c
32
struct cpufreq_policy *policy;
drivers/gpu/drm/i915/gt/intel_llc.c
35
policy = cpufreq_cpu_get(0);
drivers/gpu/drm/i915/gt/intel_llc.c
36
if (policy) {
drivers/gpu/drm/i915/gt/intel_llc.c
37
max_khz = policy->cpuinfo.max_freq;
drivers/gpu/drm/i915/gt/intel_llc.c
38
cpufreq_cpu_put(policy);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2671
static u32 __guc_context_policy_action_size(struct context_policy *policy)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2673
size_t bytes = sizeof(policy->h2g.header) +
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2674
(sizeof(policy->h2g.klv[0]) * policy->count);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2679
static void __guc_context_policy_start_klv(struct context_policy *policy, u16 guc_id)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2681
policy->h2g.header.action = INTEL_GUC_ACTION_HOST2GUC_UPDATE_CONTEXT_POLICIES;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2682
policy->h2g.header.ctx_id = guc_id;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2683
policy->count = 0;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2687
static void __guc_context_policy_add_##func(struct context_policy *policy, u32 data) \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2689
GEM_BUG_ON(policy->count >= GUC_CONTEXT_POLICIES_KLV_NUM_IDS); \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2690
policy->h2g.klv[policy->count].kl = \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2693
policy->h2g.klv[policy->count].value = data; \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2694
policy->count++; \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2706
struct context_policy *policy,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2709
return guc_submission_send_busy_loop(guc, (u32 *)&policy->h2g,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2710
__guc_context_policy_action_size(policy),
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2718
struct context_policy policy;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2736
__guc_context_policy_start_klv(&policy, ce->guc_id.id);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2738
__guc_context_policy_add_priority(&policy, ce->guc_state.prio);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2739
__guc_context_policy_add_execution_quantum(&policy, execution_quantum);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2740
__guc_context_policy_add_preemption_timeout(&policy, preemption_timeout);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2741
__guc_context_policy_add_slpc_ctx_freq_req(&policy, slpc_ctx_freq_req);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2744
__guc_context_policy_add_preempt_to_idle(&policy, 1);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2746
ret = __guc_context_set_context_policies(guc, &policy, loop);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3238
struct context_policy policy;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3240
__guc_context_policy_start_klv(&policy, guc_id);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3241
__guc_context_policy_add_preemption_timeout(&policy, preemption_timeout);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3242
__guc_context_set_context_policies(guc, &policy, true);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3626
struct context_policy policy;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3628
__guc_context_policy_start_klv(&policy, ce->guc_id.id);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3629
__guc_context_policy_add_priority(&policy, ce->guc_state.prio);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
3630
__guc_context_set_context_policies(guc, &policy, true);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4682
static u32 __guc_scheduling_policy_action_size(struct scheduling_policy *policy)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4684
u32 *start = (void *)&policy->h2g;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4685
u32 *end = policy->h2g.data + policy->num_words;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4691
static struct scheduling_policy *__guc_scheduling_policy_start_klv(struct scheduling_policy *policy)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4693
policy->h2g.header.action = INTEL_GUC_ACTION_UPDATE_SCHEDULING_POLICIES_KLV;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4694
policy->max_words = ARRAY_SIZE(policy->h2g.data);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4695
policy->num_words = 0;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4696
policy->count = 0;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4698
return policy;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4701
static void __guc_scheduling_policy_add_klv(struct scheduling_policy *policy,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4704
u32 *klv_ptr = policy->h2g.data + policy->num_words;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4706
GEM_BUG_ON((policy->num_words + 1 + len) > policy->max_words);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4710
policy->num_words += 1 + len;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4711
policy->count++;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4715
struct scheduling_policy *policy)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4719
ret = intel_guc_send(guc, (u32 *)&policy->h2g,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4720
__guc_scheduling_policy_action_size(policy));
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4727
if (ret != policy->count) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4729
ret, policy->count);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4730
if (ret > policy->count)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4739
struct scheduling_policy policy;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4747
__guc_scheduling_policy_start_klv(&policy);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4755
__guc_scheduling_policy_add_klv(&policy,
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4759
ret = __guc_action_set_scheduling_policies(guc, &policy);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
286
enum nvkm_gsp_rpc_reply_policy policy, u32 repc)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
288
return gsp->rm->api->rpc->push(gsp, argv, policy, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
304
enum nvkm_gsp_rpc_reply_policy policy)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
306
void *repv = nvkm_gsp_rpc_push(gsp, argv, policy, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
552
enum nvkm_gsp_rpc_reply_policy policy,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
558
switch (policy) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
579
enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
592
if (policy == NVKM_GSP_RPC_REPLY_NOSEQ)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
601
return r535_gsp_rpc_handle_reply(gsp, fn, policy, gsp_rpc_len);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
633
enum nvkm_gsp_rpc_reply_policy policy, u32 gsp_rpc_len)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
686
repv = r535_gsp_rpc_handle_reply(gsp, fn, policy, payload_size +
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/rpc.c
691
repv = r535_gsp_rpc_send(gsp, payload, policy, gsp_rpc_len);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/rm.h
47
enum nvkm_gsp_rpc_reply_policy policy, u32 repc);
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
199
gt->sriov.pf.policy.guc.sched_groups.current_mode;
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
372
struct xe_gt_sriov_scheduler_groups *info = >->sriov.pf.policy.guc.sched_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
117
static int pf_update_policy_bool(struct xe_gt *gt, u16 key, bool *policy, bool value)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
133
*policy = value;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
137
static int pf_update_policy_u32(struct xe_gt *gt, u16 key, u32 *policy, u32 value)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
152
*policy = value;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
176
>->sriov.pf.policy.guc.sched_if_idle,
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
190
return pf_provision_sched_if_idle(gt, gt->sriov.pf.policy.guc.sched_if_idle);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
198
gt->sriov.pf.policy.guc.sched_if_idle = false;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
236
enable = gt->sriov.pf.policy.guc.sched_if_idle;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
248
>->sriov.pf.policy.guc.reset_engine, enable);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
256
return pf_provision_reset_engine(gt, gt->sriov.pf.policy.guc.reset_engine);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
264
gt->sriov.pf.policy.guc.reset_engine = false;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
302
enable = gt->sriov.pf.policy.guc.reset_engine;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
314
>->sriov.pf.policy.guc.sample_period, value);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
322
return pf_provision_sample_period(gt, gt->sriov.pf.policy.guc.sample_period);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
330
gt->sriov.pf.policy.guc.sample_period = 0;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
368
value = gt->sriov.pf.policy.guc.sample_period;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
422
if (group > gt->sriov.pf.policy.guc.sched_groups.max_groups) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
424
group, gt->sriov.pf.policy.guc.sched_groups.max_groups);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
497
gt->sriov.pf.policy.guc.sched_groups.max_groups = 2;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
500
u32 *num_groups = >->sriov.pf.policy.guc.sched_groups.modes[m].num_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
502
>->sriov.pf.policy.guc.sched_groups.modes[m].groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
526
gt->sriov.pf.policy.guc.sched_groups.supported_modes |= BIT(m);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
541
return gt->sriov.pf.policy.guc.sched_groups.supported_modes;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
560
return gt->sriov.pf.policy.guc.sched_groups.supported_modes & BIT(mode);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
565
struct guc_sched_group *groups = gt->sriov.pf.policy.guc.sched_groups.modes[mode].groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
566
u32 num_groups = gt->sriov.pf.policy.guc.sched_groups.modes[mode].num_groups;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
583
if (gt->sriov.pf.policy.guc.sched_groups.current_mode == mode)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
610
gt->sriov.pf.policy.guc.sched_groups.current_mode = mode;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
624
return __pf_provision_sched_groups(gt, gt->sriov.pf.policy.guc.sched_groups.current_mode);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
632
gt->sriov.pf.policy.guc.sched_groups.current_mode = XE_SRIOV_SCHED_GROUPS_DISABLED;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
665
return gt->sriov.pf.policy.guc.sched_groups.current_mode != XE_SRIOV_SCHED_GROUPS_DISABLED;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
732
static void print_guc_policies(struct drm_printer *p, struct xe_gt_sriov_guc_policies *policy)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
736
str_enabled_disabled(policy->sched_if_idle));
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
739
str_enabled_disabled(policy->reset_engine));
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
742
policy->sample_period, policy->sample_period ? "ms" : "(disabled)");
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
759
print_guc_policies(p, >->sriov.pf.policy.guc);
drivers/gpu/drm/xe/xe_gt_sriov_pf_types.h
60
struct xe_gt_sriov_pf_policy policy;
drivers/gpu/drm/xe/xe_guc_submit.c
480
static u32 __guc_exec_queue_policy_action_size(struct exec_queue_policy *policy)
drivers/gpu/drm/xe/xe_guc_submit.c
482
size_t bytes = sizeof(policy->h2g.header) +
drivers/gpu/drm/xe/xe_guc_submit.c
483
(sizeof(policy->h2g.klv[0]) * policy->count);
drivers/gpu/drm/xe/xe_guc_submit.c
488
static void __guc_exec_queue_policy_start_klv(struct exec_queue_policy *policy,
drivers/gpu/drm/xe/xe_guc_submit.c
491
policy->h2g.header.action =
drivers/gpu/drm/xe/xe_guc_submit.c
493
policy->h2g.header.guc_id = guc_id;
drivers/gpu/drm/xe/xe_guc_submit.c
494
policy->count = 0;
drivers/gpu/drm/xe/xe_guc_submit.c
498
static void __guc_exec_queue_policy_add_##func(struct exec_queue_policy *policy, \
drivers/gpu/drm/xe/xe_guc_submit.c
501
XE_WARN_ON(policy->count >= GUC_CONTEXT_POLICIES_KLV_NUM_IDS); \
drivers/gpu/drm/xe/xe_guc_submit.c
503
policy->h2g.klv[policy->count].kl = \
drivers/gpu/drm/xe/xe_guc_submit.c
507
policy->h2g.klv[policy->count].value = data; \
drivers/gpu/drm/xe/xe_guc_submit.c
508
policy->count++; \
drivers/gpu/drm/xe/xe_guc_submit.c
526
struct exec_queue_policy policy;
drivers/gpu/drm/xe/xe_guc_submit.c
538
__guc_exec_queue_policy_start_klv(&policy, q->guc->id);
drivers/gpu/drm/xe/xe_guc_submit.c
539
__guc_exec_queue_policy_add_priority(&policy, xe_exec_queue_prio_to_guc[prio]);
drivers/gpu/drm/xe/xe_guc_submit.c
540
__guc_exec_queue_policy_add_execution_quantum(&policy, timeslice_us);
drivers/gpu/drm/xe/xe_guc_submit.c
541
__guc_exec_queue_policy_add_preemption_timeout(&policy, preempt_timeout_us);
drivers/gpu/drm/xe/xe_guc_submit.c
542
__guc_exec_queue_policy_add_slpc_exec_queue_freq_req(&policy,
drivers/gpu/drm/xe/xe_guc_submit.c
545
xe_guc_ct_send(&guc->ct, (u32 *)&policy.h2g,
drivers/gpu/drm/xe/xe_guc_submit.c
546
__guc_exec_queue_policy_action_size(&policy), 0, 0);
drivers/gpu/drm/xe/xe_guc_submit.c
551
struct exec_queue_policy policy;
drivers/gpu/drm/xe/xe_guc_submit.c
555
__guc_exec_queue_policy_start_klv(&policy, q->guc->id);
drivers/gpu/drm/xe/xe_guc_submit.c
556
__guc_exec_queue_policy_add_preemption_timeout(&policy, 1);
drivers/gpu/drm/xe/xe_guc_submit.c
558
xe_guc_ct_send(&guc->ct, (u32 *)&policy.h2g,
drivers/gpu/drm/xe/xe_guc_submit.c
559
__guc_exec_queue_policy_action_size(&policy), 0, 0);
drivers/hwtracing/stm/core.c
956
if (stm->policy)
drivers/hwtracing/stm/core.c
957
stp_policy_unbind(stm->policy);
drivers/hwtracing/stm/policy.c
110
stm = policy_node->policy->stm;
drivers/hwtracing/stm/policy.c
157
stm = policy_node->policy->stm;
drivers/hwtracing/stm/policy.c
228
struct stp_policy *policy;
drivers/hwtracing/stm/policy.c
231
policy = container_of(group, struct stp_policy, group);
drivers/hwtracing/stm/policy.c
235
policy = parent_node->policy;
drivers/hwtracing/stm/policy.c
238
if (!policy->stm)
drivers/hwtracing/stm/policy.c
241
pdrv = policy->stm->pdrv;
drivers/hwtracing/stm/policy.c
251
if (policy->stm->pdrv_node_type)
drivers/hwtracing/stm/policy.c
252
type = policy->stm->pdrv_node_type;
drivers/hwtracing/stm/policy.c
256
policy_node->policy = policy;
drivers/hwtracing/stm/policy.c
259
policy_node->first_master = policy->stm->data->sw_start;
drivers/hwtracing/stm/policy.c
260
policy_node->last_master = policy->stm->data->sw_end;
drivers/hwtracing/stm/policy.c
262
policy_node->last_channel = policy->stm->data->sw_nchannels - 1;
drivers/hwtracing/stm/policy.c
291
struct stp_policy *policy = to_stp_policy(item);
drivers/hwtracing/stm/policy.c
295
(policy && policy->stm) ?
drivers/hwtracing/stm/policy.c
296
policy->stm->data->name :
drivers/hwtracing/stm/policy.c
307
struct stp_policy *policy = to_stp_policy(item);
drivers/hwtracing/stm/policy.c
31
struct stp_policy *policy;
drivers/hwtracing/stm/policy.c
311
(policy && policy->stm) ?
drivers/hwtracing/stm/policy.c
312
policy->stm->pdrv->name :
drivers/hwtracing/stm/policy.c
326
void stp_policy_unbind(struct stp_policy *policy)
drivers/hwtracing/stm/policy.c
328
struct stm_device *stm = policy->stm;
drivers/hwtracing/stm/policy.c
335
if (WARN_ON_ONCE(!policy->stm))
drivers/hwtracing/stm/policy.c
340
stm->policy = NULL;
drivers/hwtracing/stm/policy.c
341
policy->stm = NULL;
drivers/hwtracing/stm/policy.c
353
struct stp_policy *policy = to_stp_policy(item);
drivers/hwtracing/stm/policy.c
354
struct stm_device *stm = policy->stm;
drivers/hwtracing/stm/policy.c
361
stp_policy_unbind(policy);
drivers/hwtracing/stm/policy.c
364
kfree(policy);
drivers/hwtracing/stm/policy.c
435
if (stm->policy) {
drivers/hwtracing/stm/policy.c
440
stm->policy = kzalloc_obj(*stm->policy);
drivers/hwtracing/stm/policy.c
441
if (!stm->policy) {
drivers/hwtracing/stm/policy.c
446
config_group_init_type_name(&stm->policy->group, name,
drivers/hwtracing/stm/policy.c
451
stm->policy->stm = stm;
drivers/hwtracing/stm/policy.c
452
ret = &stm->policy->group;
drivers/hwtracing/stm/policy.c
491
__stp_policy_node_lookup(struct stp_policy *policy, char *s)
drivers/hwtracing/stm/policy.c
494
struct list_head *head = &policy->group.cg_children;
drivers/hwtracing/stm/policy.c
540
if (stm->policy)
drivers/hwtracing/stm/policy.c
541
policy_node = __stp_policy_node_lookup(stm->policy, s);
drivers/hwtracing/stm/stm.h
32
struct stp_policy *policy;
drivers/hwtracing/stm/stm.h
54
void stp_policy_unbind(struct stp_policy *policy);
drivers/infiniband/hw/mlx5/ib_virt.c
105
in->policy = net_to_mlx_policy(state);
drivers/infiniband/hw/mlx5/ib_virt.c
106
if (in->policy == MLX5_POLICY_INVALID) {
drivers/infiniband/hw/mlx5/ib_virt.c
113
vfs_ctx[vf].policy = in->policy;
drivers/infiniband/hw/mlx5/ib_virt.c
69
info->linkstate = mlx_to_net_policy(rep->policy);
drivers/infiniband/hw/mlx5/ib_virt.c
78
static inline enum port_state_policy net_to_mlx_policy(int policy)
drivers/infiniband/hw/mlx5/ib_virt.c
80
switch (policy) {
drivers/infiniband/ulp/ipoib/ipoib_netlink.c
175
.policy = ipoib_policy,
drivers/macintosh/windfarm_cpufreq_clamp.c
65
struct cpufreq_policy *policy;
drivers/macintosh/windfarm_cpufreq_clamp.c
70
policy = cpufreq_cpu_get(0);
drivers/macintosh/windfarm_cpufreq_clamp.c
71
if (!policy) {
drivers/macintosh/windfarm_cpufreq_clamp.c
76
min_freq = policy->cpuinfo.min_freq;
drivers/macintosh/windfarm_cpufreq_clamp.c
77
max_freq = policy->cpuinfo.max_freq;
drivers/macintosh/windfarm_cpufreq_clamp.c
79
ret = freq_qos_add_request(&policy->constraints, &qos_req, FREQ_QOS_MAX,
drivers/macintosh/windfarm_cpufreq_clamp.c
82
cpufreq_cpu_put(policy);
drivers/md/dm-cache-metadata.c
1277
struct dm_cache_policy *policy)
drivers/md/dm-cache-metadata.c
1279
const char *policy_name = dm_cache_policy_get_name(policy);
drivers/md/dm-cache-metadata.c
1280
const unsigned int *policy_version = dm_cache_policy_get_version(policy);
drivers/md/dm-cache-metadata.c
1281
size_t policy_hint_size = dm_cache_policy_get_hint_size(policy);
drivers/md/dm-cache-metadata.c
1310
struct dm_cache_policy *policy)
drivers/md/dm-cache-metadata.c
1312
return cmd->clean_when_opened && policy_unchanged(cmd, policy) &&
drivers/md/dm-cache-metadata.c
1400
struct dm_cache_policy *policy,
drivers/md/dm-cache-metadata.c
1406
bool hints_valid = hints_array_available(cmd, policy);
drivers/md/dm-cache-metadata.c
1489
struct dm_cache_policy *policy,
drivers/md/dm-cache-metadata.c
1495
r = __load_mappings(cmd, policy, fn, context);
drivers/md/dm-cache-metadata.c
1664
struct dm_cache_policy *policy = context;
drivers/md/dm-cache-metadata.c
1666
value = policy_get_hint(policy, to_cblock(index));
drivers/md/dm-cache-metadata.c
1676
static int write_hints(struct dm_cache_metadata *cmd, struct dm_cache_policy *policy)
drivers/md/dm-cache-metadata.c
1680
const char *policy_name = dm_cache_policy_get_name(policy);
drivers/md/dm-cache-metadata.c
1681
const unsigned int *policy_version = dm_cache_policy_get_version(policy);
drivers/md/dm-cache-metadata.c
1690
hint_size = dm_cache_policy_get_hint_size(policy);
drivers/md/dm-cache-metadata.c
1703
get_hint, policy);
drivers/md/dm-cache-metadata.c
1706
int dm_cache_write_hints(struct dm_cache_metadata *cmd, struct dm_cache_policy *policy)
drivers/md/dm-cache-metadata.c
1711
r = write_hints(cmd, policy);
drivers/md/dm-cache-metadata.h
94
struct dm_cache_policy *policy,
drivers/md/dm-cache-policy-smq.c
1358
return container_of(p, struct smq_policy, policy);
drivers/md/dm-cache-policy-smq.c
1690
mq->policy.destroy = smq_destroy;
drivers/md/dm-cache-policy-smq.c
1691
mq->policy.lookup = smq_lookup;
drivers/md/dm-cache-policy-smq.c
1692
mq->policy.lookup_with_work = smq_lookup_with_work;
drivers/md/dm-cache-policy-smq.c
1693
mq->policy.get_background_work = smq_get_background_work;
drivers/md/dm-cache-policy-smq.c
1694
mq->policy.complete_background_work = smq_complete_background_work;
drivers/md/dm-cache-policy-smq.c
1695
mq->policy.set_dirty = smq_set_dirty;
drivers/md/dm-cache-policy-smq.c
1696
mq->policy.clear_dirty = smq_clear_dirty;
drivers/md/dm-cache-policy-smq.c
1697
mq->policy.load_mapping = smq_load_mapping;
drivers/md/dm-cache-policy-smq.c
1698
mq->policy.invalidate_mapping = smq_invalidate_mapping;
drivers/md/dm-cache-policy-smq.c
1699
mq->policy.get_hint = smq_get_hint;
drivers/md/dm-cache-policy-smq.c
1700
mq->policy.residency = smq_residency;
drivers/md/dm-cache-policy-smq.c
1701
mq->policy.tick = smq_tick;
drivers/md/dm-cache-policy-smq.c
1702
mq->policy.allow_migrations = smq_allow_migrations;
drivers/md/dm-cache-policy-smq.c
1705
mq->policy.set_config_value = mq_set_config_value;
drivers/md/dm-cache-policy-smq.c
1706
mq->policy.emit_config_values = mq_emit_config_values;
drivers/md/dm-cache-policy-smq.c
1822
return &mq->policy;
drivers/md/dm-cache-policy-smq.c
798
struct dm_cache_policy policy;
drivers/md/dm-cache-target.c
1187
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1210
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1217
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1438
policy_complete_background_work(cache->policy, op, false);
drivers/md/dm-cache-target.c
1488
r = policy_invalidate_mapping(cache->policy, cblock);
drivers/md/dm-cache-target.c
1644
r = policy_lookup_with_work(cache->policy, block, &cblock, data_dir, true, &op);
drivers/md/dm-cache-target.c
1659
r = policy_lookup(cache->policy, block, &cblock, data_dir, false, &background_queued);
drivers/md/dm-cache-target.c
1871
policy_tick(cache->policy, true);
drivers/md/dm-cache-target.c
1887
r = policy_get_background_work(cache->policy, b == IDLE, &op);
drivers/md/dm-cache-target.c
1949
if (cache->policy)
drivers/md/dm-cache-target.c
1950
dm_cache_policy_destroy(cache->policy);
drivers/md/dm-cache-target.c
2293
r = policy_set_config_value(cache->policy, key, value);
drivers/md/dm-cache-target.c
2333
cache->policy = p;
drivers/md/dm-cache-target.c
2334
BUG_ON(!cache->policy);
drivers/md/dm-cache-target.c
2455
dm_cache_policy_get_hint_size(cache->policy),
drivers/md/dm-cache-target.c
2485
policy_allow_migrations(cache->policy, false);
drivers/md/dm-cache-target.c
2683
policy_tick(cache->policy, false);
drivers/md/dm-cache-target.c
2744
r = dm_cache_write_hints(cache->cmd, cache->policy);
drivers/md/dm-cache-target.c
2818
return policy_load_mapping(cache->policy, oblock, cblock, dirty, hint, hint_valid);
drivers/md/dm-cache-target.c
3042
r = dm_cache_load_mappings(cache->cmd, cache->policy,
drivers/md/dm-cache-target.c
3176
residency = policy_residency(cache->policy);
drivers/md/dm-cache-target.c
3197
DMEMIT("%s ", dm_cache_policy_get_name(cache->policy));
drivers/md/dm-cache-target.c
3199
r = policy_emit_config_values(cache->policy, result, maxlen, &sz);
drivers/md/dm-cache-target.c
383
struct dm_cache_policy *policy;
drivers/md/dm-cache-target.c
622
policy_set_dirty(cache->policy, cblock);
drivers/md/dm-cache-target.c
634
policy_set_dirty(cache->policy, cblock);
drivers/md/dm-cache-target.c
644
policy_clear_dirty(cache->policy, cblock);
drivers/net/amt.c
3340
.policy = amt_policy,
drivers/net/bareudp.c
763
.policy = bareudp_policy,
drivers/net/bonding/bond_netlink.c
948
.policy = bond_policy,
drivers/net/can/dev/netlink.c
1094
.policy = can_policy,
drivers/net/can/vxcan.c
322
.policy = vxcan_policy,
drivers/net/dsa/mv88e6xxx/chip.c
2307
const struct mv88e6xxx_policy *policy)
drivers/net/dsa/mv88e6xxx/chip.c
2309
enum mv88e6xxx_policy_mapping mapping = policy->mapping;
drivers/net/dsa/mv88e6xxx/chip.c
2310
enum mv88e6xxx_policy_action action = policy->action;
drivers/net/dsa/mv88e6xxx/chip.c
2311
const u8 *addr = policy->addr;
drivers/net/dsa/mv88e6xxx/chip.c
2312
u16 vid = policy->vid;
drivers/net/dsa/mv88e6xxx/chip.c
2345
idr_for_each_entry(&chip->policies, policy, id)
drivers/net/dsa/mv88e6xxx/chip.c
2346
if (policy->port == port &&
drivers/net/dsa/mv88e6xxx/chip.c
2347
policy->mapping == mapping &&
drivers/net/dsa/mv88e6xxx/chip.c
2348
policy->action != action)
drivers/net/dsa/mv88e6xxx/chip.c
2361
struct mv88e6xxx_policy *policy;
drivers/net/dsa/mv88e6xxx/chip.c
2400
idr_for_each_entry(&chip->policies, policy, id) {
drivers/net/dsa/mv88e6xxx/chip.c
2401
if (policy->port == port && policy->mapping == mapping &&
drivers/net/dsa/mv88e6xxx/chip.c
2402
policy->action == action && policy->vid == vid &&
drivers/net/dsa/mv88e6xxx/chip.c
2403
ether_addr_equal(policy->addr, addr))
drivers/net/dsa/mv88e6xxx/chip.c
2407
policy = devm_kzalloc(chip->dev, sizeof(*policy), GFP_KERNEL);
drivers/net/dsa/mv88e6xxx/chip.c
2408
if (!policy)
drivers/net/dsa/mv88e6xxx/chip.c
2412
err = idr_alloc_u32(&chip->policies, policy, &fs->location, 0xffffffff,
drivers/net/dsa/mv88e6xxx/chip.c
2415
devm_kfree(chip->dev, policy);
drivers/net/dsa/mv88e6xxx/chip.c
2419
memcpy(&policy->fs, fs, sizeof(*fs));
drivers/net/dsa/mv88e6xxx/chip.c
2420
ether_addr_copy(policy->addr, addr);
drivers/net/dsa/mv88e6xxx/chip.c
2421
policy->mapping = mapping;
drivers/net/dsa/mv88e6xxx/chip.c
2422
policy->action = action;
drivers/net/dsa/mv88e6xxx/chip.c
2423
policy->port = port;
drivers/net/dsa/mv88e6xxx/chip.c
2424
policy->vid = vid;
drivers/net/dsa/mv88e6xxx/chip.c
2426
err = mv88e6xxx_policy_apply(chip, port, policy);
drivers/net/dsa/mv88e6xxx/chip.c
2429
devm_kfree(chip->dev, policy);
drivers/net/dsa/mv88e6xxx/chip.c
2441
struct mv88e6xxx_policy *policy;
drivers/net/dsa/mv88e6xxx/chip.c
2452
idr_for_each_entry(&chip->policies, policy, id)
drivers/net/dsa/mv88e6xxx/chip.c
2453
if (policy->port == port)
drivers/net/dsa/mv88e6xxx/chip.c
2459
policy = idr_find(&chip->policies, fs->location);
drivers/net/dsa/mv88e6xxx/chip.c
2460
if (policy) {
drivers/net/dsa/mv88e6xxx/chip.c
2461
memcpy(fs, &policy->fs, sizeof(*fs));
drivers/net/dsa/mv88e6xxx/chip.c
2468
idr_for_each_entry(&chip->policies, policy, id)
drivers/net/dsa/mv88e6xxx/chip.c
2469
if (policy->port == port)
drivers/net/dsa/mv88e6xxx/chip.c
2488
struct mv88e6xxx_policy *policy;
drivers/net/dsa/mv88e6xxx/chip.c
2499
policy = idr_remove(&chip->policies, fs->location);
drivers/net/dsa/mv88e6xxx/chip.c
2500
if (policy) {
drivers/net/dsa/mv88e6xxx/chip.c
2501
policy->action = MV88E6XXX_POLICY_ACTION_NORMAL;
drivers/net/dsa/mv88e6xxx/chip.c
2502
err = mv88e6xxx_policy_apply(chip, port, policy);
drivers/net/dsa/mv88e6xxx/chip.c
2503
devm_kfree(chip->dev, policy);
drivers/net/dsa/mv88e6xxx/chip.c
2616
vlan.policy = true;
drivers/net/dsa/mv88e6xxx/chip.h
195
bool policy;
drivers/net/dsa/mv88e6xxx/global1_vtu.c
31
entry->policy = !!(val & MV88E6352_G1_VTU_FID_VID_POLICY);
drivers/net/dsa/mv88e6xxx/global1_vtu.c
40
if (entry->policy)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
262
u64 policy;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
274
policy = FIELD_PREP(MCS_RX_SECY_PLCY_RW_MASK, secy->replay_window);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
276
policy |= MCS_RX_SECY_PLCY_RP;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
278
policy |= MCS_RX_SECY_PLCY_AUTH_ENA;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
293
policy |= FIELD_PREP(MCS_RX_SECY_PLCY_CIP, cipher);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
294
policy |= FIELD_PREP(MCS_RX_SECY_PLCY_VAL, secy->validate_frames);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
296
policy |= MCS_RX_SECY_PLCY_ENA;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
298
req->plcy = policy;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
504
u64 policy;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
534
policy = FIELD_PREP(MCS_TX_SECY_PLCY_MTU,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
537
policy |= FIELD_PREP(MCS_TX_SECY_PLCY_ST_TCI, sectag_tci >> 2);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
538
policy |= FIELD_PREP(MCS_TX_SECY_PLCY_ST_OFFSET, tag_offset);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
539
policy |= MCS_TX_SECY_PLCY_INS_MODE;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
540
policy |= MCS_TX_SECY_PLCY_AUTH_ENA;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
555
policy |= FIELD_PREP(MCS_TX_SECY_PLCY_CIP, cipher);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
558
policy |= MCS_TX_SECY_PLCY_PROTECT;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
565
policy |= MCS_TX_SECY_PLCY_ENA;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_macsec.c
567
req->plcy = policy;
drivers/net/ethernet/mellanox/mlx5/core/en/port.c
315
#define MLX5E_FEC_OVERRIDE_ADMIN_POLICY(buf, policy, write, link) \
drivers/net/ethernet/mellanox/mlx5/core/en/port.c
317
u16 *_policy = &(policy); \
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
49
sriov->vfs_ctx[vf].policy != MLX5_POLICY_INVALID) {
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
56
in->policy = sriov->vfs_ctx[vf].policy;
drivers/net/ethernet/mellanox/mlx5/core/sriov.c
60
!!(in->policy) * MLX5_HCA_VPORT_SEL_STATE_POLICY;
drivers/net/ethernet/mellanox/mlx5/core/vport.c
1162
req->policy);
drivers/net/ethernet/mellanox/mlx5/core/vport.c
792
rep->policy = MLX5_GET_PR(hca_vport_context, ctx, vport_state_policy);
drivers/net/ethernet/mellanox/mlxsw/core_env.c
1332
enum ethtool_module_power_mode_policy policy;
drivers/net/ethernet/mellanox/mlxsw/core_env.c
1338
policy = module_info->power_mode_policy;
drivers/net/ethernet/mellanox/mlxsw/core_env.c
1341
policy, &extack);
drivers/net/ethernet/mellanox/mlxsw/core_env.c
659
params->policy = module_info->power_mode_policy;
drivers/net/ethernet/mellanox/mlxsw/core_env.c
768
enum ethtool_module_power_mode_policy policy,
drivers/net/ethernet/mellanox/mlxsw/core_env.c
783
if (module_info->power_mode_policy == policy)
drivers/net/ethernet/mellanox/mlxsw/core_env.c
790
low_power = policy == ETHTOOL_MODULE_POWER_MODE_POLICY_AUTO;
drivers/net/ethernet/mellanox/mlxsw/core_env.c
797
module_info->power_mode_policy = policy;
drivers/net/ethernet/mellanox/mlxsw/core_env.c
805
enum ethtool_module_power_mode_policy policy,
drivers/net/ethernet/mellanox/mlxsw/core_env.c
811
if (policy != ETHTOOL_MODULE_POWER_MODE_POLICY_HIGH &&
drivers/net/ethernet/mellanox/mlxsw/core_env.c
812
policy != ETHTOOL_MODULE_POWER_MODE_POLICY_AUTO) {
drivers/net/ethernet/mellanox/mlxsw/core_env.c
819
module, policy, extack);
drivers/net/ethernet/mellanox/mlxsw/core_env.h
50
enum ethtool_module_power_mode_policy policy,
drivers/net/ethernet/mellanox/mlxsw/minimal.c
190
params->policy, extack);
drivers/net/ethernet/mellanox/mlxsw/reg.h
373
enum mlxsw_reg_sfd_rec_policy policy,
drivers/net/ethernet/mellanox/mlxsw/reg.h
380
mlxsw_reg_sfd_rec_policy_set(payload, rec_index, policy);
drivers/net/ethernet/mellanox/mlxsw/reg.h
439
enum mlxsw_reg_sfd_rec_policy policy,
drivers/net/ethernet/mellanox/mlxsw/reg.h
447
mlxsw_reg_sfd_rec_policy_set(payload, rec_index, policy);
drivers/net/ethernet/mellanox/mlxsw/reg.h
534
enum mlxsw_reg_sfd_rec_policy policy,
drivers/net/ethernet/mellanox/mlxsw/reg.h
542
mlxsw_reg_sfd_rec_policy_set(payload, rec_index, policy);
drivers/net/ethernet/mellanox/mlxsw/reg.h
549
enum mlxsw_reg_sfd_rec_policy policy,
drivers/net/ethernet/mellanox/mlxsw/reg.h
555
mlxsw_reg_sfd_uc_tunnel_pack(payload, rec_index, policy, mac, fid,
drivers/net/ethernet/mellanox/mlxsw/spectrum_ethtool.c
1256
module, params->policy, extack);
drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c
1782
enum mlxsw_reg_sfd_rec_policy policy)
drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c
1793
mlxsw_reg_sfd_uc_pack(sfd_pl, 0, policy, mac, fid, vid, action,
drivers/net/ethernet/netronome/nfp/nfp_main.c
485
int err, reset, policy, ifcs = 0;
drivers/net/ethernet/netronome/nfp/nfp_main.c
527
NFP_NSP_APP_FW_LOAD_PREF, &policy);
drivers/net/ethernet/netronome/nfp/nfp_main.c
545
if (fw && policy != NFP_NSP_APP_FW_LOAD_FLASH) {
drivers/net/ethernet/netronome/nfp/nfp_main.c
557
} else if (policy != NFP_NSP_APP_FW_LOAD_DISK &&
drivers/net/ethernet/qualcomm/rmnet/rmnet_config.c
410
.policy = rmnet_policy,
drivers/net/fddi/skfp/h/smt.h
743
struct smt_p_policy policy ; /* station policy */
drivers/net/fddi/skfp/pcmplc.c
1415
int policy ;
drivers/net/fddi/skfp/pcmplc.c
1419
policy = smc->mib.fddiSMTConnectionPolicy ;
drivers/net/fddi/skfp/pcmplc.c
1425
((policy & POLICY_AA) && ne == TA) ||
drivers/net/fddi/skfp/pcmplc.c
1426
((policy & POLICY_AB) && ne == TB) ||
drivers/net/fddi/skfp/pcmplc.c
1427
((policy & POLICY_AS) && ne == TS) ||
drivers/net/fddi/skfp/pcmplc.c
1428
((policy & POLICY_AM) && ne == TM) )
drivers/net/fddi/skfp/pcmplc.c
1433
((policy & POLICY_BA) && ne == TA) ||
drivers/net/fddi/skfp/pcmplc.c
1434
((policy & POLICY_BB) && ne == TB) ||
drivers/net/fddi/skfp/pcmplc.c
1435
((policy & POLICY_BS) && ne == TS) ||
drivers/net/fddi/skfp/pcmplc.c
1436
((policy & POLICY_BM) && ne == TM) )
drivers/net/fddi/skfp/pcmplc.c
1441
((policy & POLICY_SA) && ne == TA) ||
drivers/net/fddi/skfp/pcmplc.c
1442
((policy & POLICY_SB) && ne == TB) ||
drivers/net/fddi/skfp/pcmplc.c
1443
((policy & POLICY_SS) && ne == TS) ||
drivers/net/fddi/skfp/pcmplc.c
1444
((policy & POLICY_SM) && ne == TM) )
drivers/net/fddi/skfp/pcmplc.c
1449
((policy & POLICY_MA) && ne == TA) ||
drivers/net/fddi/skfp/pcmplc.c
1450
((policy & POLICY_MB) && ne == TB) ||
drivers/net/fddi/skfp/pcmplc.c
1451
((policy & POLICY_MS) && ne == TS) ||
drivers/net/fddi/skfp/pcmplc.c
1452
((policy & POLICY_MM) && ne == TM) )
drivers/net/fddi/skfp/smt.c
1033
smt_fill_policy(smc,&sif->policy) ; /* set station policy */
drivers/net/fddi/skfp/smt.c
1254
static void smt_fill_policy(struct s_smc *smc, struct smt_p_policy *policy)
drivers/net/fddi/skfp/smt.c
1268
SMTSETPARA(policy,SMT_P_POLICY) ;
drivers/net/fddi/skfp/smt.c
1278
policy->pl_config = smc->mib.fddiSMTConfigPolicy ;
drivers/net/fddi/skfp/smt.c
1279
policy->pl_connect = out ;
drivers/net/fddi/skfp/smt.c
77
static void smt_fill_policy(struct s_smc *smc, struct smt_p_policy *policy);
drivers/net/geneve.c
2357
.policy = geneve_policy,
drivers/net/gtp.c
1612
.policy = gtp_policy,
drivers/net/gtp.c
2463
.policy = gtp_genl_policy,
drivers/net/ieee802154/mac802154_hwsim.c
794
.policy = hwsim_genl_policy,
drivers/net/ipvlan/ipvlan_main.c
716
ops->policy = ipvlan_nl_policy;
drivers/net/macsec.c
3384
.policy = macsec_genl_policy,
drivers/net/macsec.c
4375
.policy = macsec_rtnl_policy,
drivers/net/macvlan.c
1758
ops->policy = macvlan_policy;
drivers/net/netkit.c
22
enum netkit_action policy;
drivers/net/netkit.c
300
static int netkit_check_policy(int policy, struct nlattr *tb,
drivers/net/netkit.c
303
switch (policy) {
drivers/net/netkit.c
418
nk->policy = policy_peer;
drivers/net/netkit.c
444
nk->policy = policy_prim;
drivers/net/netkit.c
83
enum netkit_action ret = READ_ONCE(nk->policy);
drivers/net/netkit.c
869
enum netkit_action policy;
drivers/net/netkit.c
902
policy = nla_get_u32(attr);
drivers/net/netkit.c
903
err = netkit_check_policy(policy, attr, extack);
drivers/net/netkit.c
906
WRITE_ONCE(nk->policy, policy);
drivers/net/netkit.c
912
policy = nla_get_u32(attr);
drivers/net/netkit.c
914
err = netkit_check_policy(policy, attr, extack);
drivers/net/netkit.c
918
WRITE_ONCE(nk->policy, policy);
drivers/net/netkit.c
944
if (nla_put_u32(skb, IFLA_NETKIT_POLICY, nk->policy))
drivers/net/netkit.c
957
if (nla_put_u32(skb, IFLA_NETKIT_PEER_POLICY, nk->policy))
drivers/net/netkit.c
989
.policy = netkit_policy,
drivers/net/ovpn/main.c
235
.policy = ovpn_policy,
drivers/net/ovpn/netlink-gen.c
173
.policy = ovpn_peer_new_nl_policy,
drivers/net/ovpn/netlink-gen.c
182
.policy = ovpn_peer_set_nl_policy,
drivers/net/ovpn/netlink-gen.c
191
.policy = ovpn_peer_get_do_nl_policy,
drivers/net/ovpn/netlink-gen.c
198
.policy = ovpn_peer_get_dump_nl_policy,
drivers/net/ovpn/netlink-gen.c
207
.policy = ovpn_peer_del_nl_policy,
drivers/net/ovpn/netlink-gen.c
216
.policy = ovpn_key_new_nl_policy,
drivers/net/ovpn/netlink-gen.c
225
.policy = ovpn_key_get_nl_policy,
drivers/net/ovpn/netlink-gen.c
234
.policy = ovpn_key_swap_nl_policy,
drivers/net/ovpn/netlink-gen.c
243
.policy = ovpn_key_del_nl_policy,
drivers/net/ppp/ppp_generic.c
1389
.policy = ppp_nl_policy,
drivers/net/team/team_core.c
2835
.policy = team_nl_policy,
drivers/net/veth.c
1982
.policy = veth_policy,
drivers/net/vrf.c
1778
.policy = vrf_nl_policy,
drivers/net/vxlan/vxlan_core.c
4679
.policy = vxlan_policy,
drivers/net/wireguard/generated/netlink.c
62
.policy = wireguard_get_device_nl_policy,
drivers/net/wireguard/generated/netlink.c
69
.policy = wireguard_set_device_nl_policy,
drivers/net/wireless/ath/wcn36xx/hal.h
2940
u8 policy;
drivers/net/wireless/ath/wcn36xx/smd.c
2448
msg_body.policy = 1;
drivers/net/wireless/ath/wil6210/cfg80211.c
225
.policy = wil_rf_sector_policy,
drivers/net/wireless/ath/wil6210/cfg80211.c
233
.policy = wil_rf_sector_policy,
drivers/net/wireless/ath/wil6210/cfg80211.c
242
.policy = wil_rf_sector_policy,
drivers/net/wireless/ath/wil6210/cfg80211.c
251
.policy = wil_rf_sector_policy,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/vendor.c
115
.policy = VENDOR_CMD_RAW_DATA,
drivers/net/wireless/intel/iwlwifi/fw/api/time-event.h
227
__le16 policy;
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
3170
u32 policy = le32_to_cpu(dump_data->trig->apply_policy);
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
3173
if (policy & IWL_FW_INI_APPLY_POLICY_DUMP_COMPLETE_CMD) {
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1059
u32 policy = le32_to_cpu(trig->apply_policy);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1064
if (!(policy & IWL_FW_INI_APPLY_POLICY_OVERRIDE_DATA)) {
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1105
if (policy & IWL_FW_INI_APPLY_POLICY_OVERRIDE_CFG) {
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1114
if (policy & IWL_FW_INI_APPLY_POLICY_OVERRIDE_REGIONS) {
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1138
u32 policy = le32_to_cpu(trig->apply_policy);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1141
if (!(policy & IWL_FW_INI_APPLY_POLICY_MATCH_TIME_POINT))
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
1144
if (!(policy & IWL_FW_INI_APPLY_POLICY_MATCH_DATA) ||
drivers/net/wireless/intel/iwlwifi/mvm/time-event.c
1204
time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
drivers/net/wireless/intel/iwlwifi/mvm/time-event.c
1376
time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
drivers/net/wireless/intel/iwlwifi/mvm/time-event.c
1379
time_cmd.policy |= cpu_to_le16(TE_V2_START_IMMEDIATELY);
drivers/net/wireless/intel/iwlwifi/mvm/time-event.c
693
time_cmd.policy = cpu_to_le16(TE_V2_NOTIF_HOST_EVENT_START |
drivers/net/wireless/intel/iwlwifi/mvm/vendor-cmd.c
101
.policy = iwl_mvm_vendor_attr_policy,
drivers/net/wireless/intel/iwlwifi/mvm/vendor-cmd.c
91
.policy = iwl_mvm_vendor_attr_policy,
drivers/net/wireless/marvell/mwifiex/fw.h
2309
__le16 policy;
drivers/net/wireless/marvell/mwifiex/sta_cmd.c
1650
mc_pol->policy = cpu_to_le16(*drcs_info);
drivers/net/wireless/quantenna/qtnfmac/qlink.h
808
__le32 policy;
drivers/net/wireless/quantenna/qtnfmac/qlink_util.c
175
qacl->policy =
drivers/net/wireless/quantenna/qtnfmac/qlink_util.c
179
qacl->policy = cpu_to_le32(QLINK_ACL_POLICY_DENY_UNLESS_LISTED);
drivers/net/wireless/realtek/rtw89/acpi.c
206
struct rtw89_acpi_policy_6ghz_sp **policy)
drivers/net/wireless/realtek/rtw89/acpi.c
230
*policy = kmemdup(ptr, sizeof(*ptr), GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/acpi.c
231
if (!*policy)
drivers/net/wireless/realtek/rtw89/acpi.c
234
rtw89_hex_dump(rtwdev, RTW89_DBG_ACPI, "policy_6ghz_sp: ", *policy,
drivers/net/wireless/realtek/rtw89/acpi.c
250
struct rtw89_acpi_policy_6ghz_vlp **policy)
drivers/net/wireless/realtek/rtw89/acpi.c
274
*policy = kmemdup(ptr, sizeof(*ptr), GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/acpi.c
275
if (!*policy)
drivers/net/wireless/realtek/rtw89/acpi.c
278
rtw89_hex_dump(rtwdev, RTW89_DBG_ACPI, "policy_6ghz_vlp: ", *policy,
drivers/net/wireless/realtek/rtw89/acpi.c
293
struct rtw89_acpi_policy_tas **policy)
drivers/net/wireless/realtek/rtw89/acpi.c
317
*policy = kmemdup(ptr, sizeof(*ptr), GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/acpi.c
318
if (!*policy)
drivers/net/wireless/realtek/rtw89/acpi.c
321
rtw89_hex_dump(rtwdev, RTW89_DBG_ACPI, "policy_tas: ", *policy,
drivers/net/wireless/realtek/rtw89/acpi.c
338
struct rtw89_acpi_policy_reg_rules **policy)
drivers/net/wireless/realtek/rtw89/acpi.c
362
*policy = kmemdup(ptr, sizeof(*ptr), GFP_KERNEL);
drivers/net/wireless/realtek/rtw89/acpi.c
363
if (!*policy)
drivers/net/wireless/realtek/rtw89/acpi.c
366
rtw89_hex_dump(rtwdev, RTW89_DBG_ACPI, "policy_reg_rules: ", *policy,
drivers/net/wireless/realtek/rtw89/chan.c
1844
struct rtw89_mcc_policy *policy = &role->policy;
drivers/net/wireless/realtek/rtw89/chan.c
1857
req.c2h_rpt = policy->c2h_rpt;
drivers/net/wireless/realtek/rtw89/chan.c
1858
req.tx_null_early = policy->tx_null_early;
drivers/net/wireless/realtek/rtw89/chan.c
1859
req.dis_tx_null = policy->dis_tx_null;
drivers/net/wireless/realtek/rtw89/chan.c
1860
req.in_curr_ch = policy->in_curr_ch;
drivers/net/wireless/realtek/rtw89/chan.c
1861
req.sw_retry_count = policy->sw_retry_count;
drivers/net/wireless/realtek/rtw89/chan.c
1862
req.dis_sw_retry = policy->dis_sw_retry;
drivers/net/wireless/realtek/rtw89/chan.c
1897
struct rtw89_mcc_policy *policy = &role->policy;
drivers/net/wireless/realtek/rtw89/chan.c
1915
slot_arg->roles[0].en_tx_null = !policy->dis_tx_null;
drivers/net/wireless/realtek/rtw89/chan.c
1916
slot_arg->roles[0].null_early = policy->tx_null_early;
drivers/net/wireless/realtek/rtw89/chan.c
867
struct rtw89_mcc_policy *policy = &mcc_role->policy;
drivers/net/wireless/realtek/rtw89/chan.c
869
policy->c2h_rpt = RTW89_FW_MCC_C2H_RPT_ALL;
drivers/net/wireless/realtek/rtw89/chan.c
870
policy->tx_null_early = RTW89_MCC_DFLT_TX_NULL_EARLY;
drivers/net/wireless/realtek/rtw89/chan.c
871
policy->in_curr_ch = false;
drivers/net/wireless/realtek/rtw89/chan.c
872
policy->dis_sw_retry = true;
drivers/net/wireless/realtek/rtw89/chan.c
873
policy->sw_retry_count = false;
drivers/net/wireless/realtek/rtw89/chan.c
876
policy->dis_tx_null = true;
drivers/net/wireless/realtek/rtw89/chan.c
878
policy->dis_tx_null = false;
drivers/net/wireless/realtek/rtw89/coex.c
2238
tlv = (struct rtw89_btc_btf_tlv *)&btc->policy[len];
drivers/net/wireless/realtek/rtw89/coex.c
2246
tlv_v7 = (struct rtw89_btc_btf_tlv_v7 *)&btc->policy[len];
drivers/net/wireless/realtek/rtw89/coex.c
2288
tlv = (struct rtw89_btc_btf_tlv *)&btc->policy[len];
drivers/net/wireless/realtek/rtw89/coex.c
2335
tlv = (struct rtw89_btc_btf_tlv_v7 *)&btc->policy[len];
drivers/net/wireless/realtek/rtw89/coex.c
2348
btc->policy[len] = i; /* slot-id */
drivers/net/wireless/realtek/rtw89/coex.c
2349
memcpy(&btc->policy[len + 1], &dm->slot.v7[i],
drivers/net/wireless/realtek/rtw89/coex.c
2756
btc->policy, btc->policy_len);
drivers/net/wireless/realtek/rtw89/core.h
3326
u8 policy[RTW89_BTC_POLICY_MAXLEN];
drivers/net/wireless/realtek/rtw89/core.h
6014
struct rtw89_mcc_policy policy;
drivers/net/wireless/realtek/rtw89/regd.c
884
goto policy;
drivers/net/wireless/realtek/rtw89/regd.c
891
policy:
drivers/net/wireless/silabs/wfx/data_tx.c
46
static void wfx_tx_policy_build(struct wfx_vif *wvif, struct wfx_tx_policy *policy,
drivers/net/wireless/silabs/wfx/data_tx.c
54
memset(policy, 0, sizeof(*policy));
drivers/net/wireless/silabs/wfx/data_tx.c
64
policy->rates[rateid / 2] |= count;
drivers/net/wireless/st/cw1200/txrx.c
170
policy->defined = cw1200_get_tx_rate(priv, &rates[0])->hw_value + 1;
drivers/net/wireless/st/cw1200/txrx.c
184
policy->tbl[off] |= __cpu_to_le32(retries << shift);
drivers/net/wireless/st/cw1200/txrx.c
185
policy->retry_count += retries;
drivers/net/wireless/st/cw1200/txrx.c
223
if (tx_policy_is_equal(wanted, &it->policy))
drivers/net/wireless/st/cw1200/txrx.c
228
if (tx_policy_is_equal(wanted, &it->policy))
drivers/net/wireless/st/cw1200/txrx.c
237
++entry->policy.usage_count;
drivers/net/wireless/st/cw1200/txrx.c
244
int ret = --entry->policy.usage_count;
drivers/net/wireless/st/cw1200/txrx.c
265
if (WARN_ON(entry->policy.usage_count)) {
drivers/net/wireless/st/cw1200/txrx.c
266
entry->policy.usage_count = 0;
drivers/net/wireless/st/cw1200/txrx.c
269
memset(&entry->policy, 0, sizeof(entry->policy));
drivers/net/wireless/st/cw1200/txrx.c
323
entry->policy = wanted;
drivers/net/wireless/st/cw1200/txrx.c
326
tx_policy_dump(&entry->policy);
drivers/net/wireless/st/cw1200/txrx.c
363
struct tx_policy *src = &cache->cache[i].policy;
drivers/net/wireless/st/cw1200/txrx.c
47
static void tx_policy_dump(struct tx_policy *policy)
drivers/net/wireless/st/cw1200/txrx.c
50
policy->raw[0] & 0x0F, policy->raw[0] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
51
policy->raw[1] & 0x0F, policy->raw[1] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
52
policy->raw[2] & 0x0F, policy->raw[2] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
53
policy->raw[3] & 0x0F, policy->raw[3] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
54
policy->raw[4] & 0x0F, policy->raw[4] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
55
policy->raw[5] & 0x0F, policy->raw[5] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
56
policy->raw[6] & 0x0F, policy->raw[6] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
57
policy->raw[7] & 0x0F, policy->raw[7] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
58
policy->raw[8] & 0x0F, policy->raw[8] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
59
policy->raw[9] & 0x0F, policy->raw[9] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
60
policy->raw[10] & 0x0F, policy->raw[10] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
61
policy->raw[11] & 0x0F, policy->raw[11] >> 4,
drivers/net/wireless/st/cw1200/txrx.c
62
policy->defined);
drivers/net/wireless/st/cw1200/txrx.c
66
/* [out] */ struct tx_policy *policy,
drivers/net/wireless/st/cw1200/txrx.c
73
memset(policy, 0, sizeof(*policy));
drivers/net/wireless/st/cw1200/txrx.h
33
struct tx_policy policy;
drivers/net/wireless/ti/wlcore/main.c
2204
u8 policy = find_first_zero_bit(wl->rate_policies_map,
drivers/net/wireless/ti/wlcore/main.c
2206
if (policy >= WL12XX_MAX_RATE_POLICIES)
drivers/net/wireless/ti/wlcore/main.c
2209
__set_bit(policy, wl->rate_policies_map);
drivers/net/wireless/ti/wlcore/main.c
2210
*idx = policy;
drivers/net/wireless/ti/wlcore/main.c
2225
u8 policy = find_first_zero_bit(wl->klv_templates_map,
drivers/net/wireless/ti/wlcore/main.c
2227
if (policy >= WLCORE_MAX_KLV_TEMPLATES)
drivers/net/wireless/ti/wlcore/main.c
2230
__set_bit(policy, wl->klv_templates_map);
drivers/net/wireless/ti/wlcore/main.c
2231
*idx = policy;
drivers/net/wireless/ti/wlcore/vendor_cmd.c
157
.policy = wlcore_vendor_attr_policy,
drivers/net/wireless/ti/wlcore/vendor_cmd.c
167
.policy = wlcore_vendor_attr_policy,
drivers/net/wireless/ti/wlcore/vendor_cmd.c
177
.policy = wlcore_vendor_attr_policy,
drivers/net/wireless/virtual/mac80211_hwsim.c
633
.policy = hwsim_vendor_test_policy,
drivers/net/wireless/virtual/mac80211_hwsim.c
6697
.policy = hwsim_genl_policy,
drivers/net/wwan/wwan_core.c
1229
.policy = wwan_rtnl_policy,
drivers/pci/pcie/aspm.c
1580
module_param_call(policy, pcie_aspm_set_policy, pcie_aspm_get_policy,
drivers/power/supply/wm8350_power.c
103
eoc_mA = WM8350_CHG_EOC_mA(policy->eoc_mA);
drivers/power/supply/wm8350_power.c
110
reg | eoc_mA | policy->trickle_start_mV |
drivers/power/supply/wm8350_power.c
117
WM8350_CHG_FAST_LIMIT_mA(policy->fast_limit_USB_mA);
drivers/power/supply/wm8350_power.c
119
policy->charge_mV | policy->trickle_charge_USB_mA |
drivers/power/supply/wm8350_power.c
121
policy->charge_timeout));
drivers/power/supply/wm8350_power.c
125
WM8350_CHG_FAST_LIMIT_mA(policy->fast_limit_mA);
drivers/power/supply/wm8350_power.c
127
policy->charge_mV | policy->trickle_charge_mA |
drivers/power/supply/wm8350_power.c
129
policy->charge_timeout));
drivers/power/supply/wm8350_power.c
188
struct wm8350_charger_policy *policy = power->policy;
drivers/power/supply/wm8350_power.c
208
wm8350_charger_config(wm8350, policy);
drivers/power/supply/wm8350_power.c
229
wm8350_charger_config(wm8350, policy);
drivers/power/supply/wm8350_power.c
540
struct wm8350_charger_policy *policy = power->policy;
drivers/power/supply/wm8350_power.c
561
if (wm8350_charger_config(wm8350, policy) == 0) {
drivers/power/supply/wm8350_power.c
87
struct wm8350_charger_policy *policy)
drivers/power/supply/wm8350_power.c
91
if (!policy) {
drivers/power/supply/wm8350_power.c
98
if (policy->fast_limit_USB_mA > 500) {
drivers/powercap/dtpm_cpu.c
147
struct cpufreq_policy *policy;
drivers/powercap/dtpm_cpu.c
152
policy = cpufreq_cpu_get(dtpm_cpu->cpu);
drivers/powercap/dtpm_cpu.c
153
if (policy) {
drivers/powercap/dtpm_cpu.c
154
for_each_cpu(dtpm_cpu->cpu, policy->related_cpus)
drivers/powercap/dtpm_cpu.c
157
cpufreq_cpu_put(policy);
drivers/powercap/dtpm_cpu.c
195
struct cpufreq_policy *policy;
drivers/powercap/dtpm_cpu.c
205
policy = cpufreq_cpu_get(cpu);
drivers/powercap/dtpm_cpu.c
206
if (!policy)
drivers/powercap/dtpm_cpu.c
224
for_each_cpu(cpu, policy->related_cpus)
drivers/powercap/dtpm_cpu.c
235
ret = freq_qos_add_request(&policy->constraints,
drivers/powercap/dtpm_cpu.c
242
cpufreq_cpu_put(policy);
drivers/powercap/dtpm_cpu.c
250
for_each_cpu(cpu, policy->related_cpus)
drivers/powercap/dtpm_cpu.c
255
cpufreq_cpu_put(policy);
drivers/scsi/megaraid/megaraid_sas.h
2476
u8 policy[];
drivers/scsi/megaraid/megaraid_sas.h
2503
u8 policy[MAX_VIRTUAL_FUNCTIONS];
drivers/scsi/megaraid/megaraid_sas_base.c
2446
if (instance->vf_affiliation_111->map[ld].policy[thisVf] !=
drivers/scsi/megaraid/megaraid_sas_base.c
2447
new_affiliation_111->map[ld].policy[thisVf]) {
drivers/scsi/megaraid/megaraid_sas_base.c
2569
if (newmap->policy[thisVf] !=
drivers/scsi/megaraid/megaraid_sas_base.c
2570
savedmap->policy[thisVf]) {
drivers/scsi/megaraid/megaraid_sas_base.c
2579
if (!found && newmap->policy[thisVf] !=
drivers/scsi/megaraid/megaraid_sas_base.c
2597
if (savedmap->policy[thisVf] !=
drivers/scsi/megaraid/megaraid_sas_base.c
2598
newmap->policy[thisVf]) {
drivers/scsi/megaraid/megaraid_sas_base.c
2607
if (!found && savedmap->policy[thisVf] !=
drivers/target/target_core_user.c
483
.policy = tcmu_attr_policy,
drivers/thermal/cpufreq_cooling.c
237
struct cpufreq_policy *policy = cpufreq_cdev->policy;
drivers/thermal/cpufreq_cooling.c
239
freq = cpufreq_quick_get(policy->cpu);
drivers/thermal/cpufreq_cooling.c
241
for_each_cpu(cpu, policy->related_cpus) {
drivers/thermal/cpufreq_cooling.c
256
trace_thermal_power_cpu_get_power_simple(policy->cpu, *power);
drivers/thermal/cpufreq_cooling.c
285
num_cpus = cpumask_weight(cpufreq_cdev->policy->cpus);
drivers/thermal/cpufreq_cooling.c
320
struct cpufreq_policy *policy = cpufreq_cdev->policy;
drivers/thermal/cpufreq_cooling.c
327
trace_thermal_power_cpu_limit(policy->related_cpus, target_freq, *state,
drivers/thermal/cpufreq_cooling.c
334
struct cpufreq_policy *policy;
drivers/thermal/cpufreq_cooling.c
340
policy = cpufreq_cdev->policy;
drivers/thermal/cpufreq_cooling.c
341
if (!cpumask_equal(policy->related_cpus, em_span_cpus(em))) {
drivers/thermal/cpufreq_cooling.c
344
cpumask_pr_args(policy->related_cpus));
drivers/thermal/cpufreq_cooling.c
372
unsigned int num_cpus = cpumask_weight(cpufreq_cdev->policy->related_cpus);
drivers/thermal/cpufreq_cooling.c
392
struct cpufreq_policy *policy;
drivers/thermal/cpufreq_cooling.c
413
policy = cpufreq_cdev->policy;
drivers/thermal/cpufreq_cooling.c
414
if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
drivers/thermal/cpufreq_cooling.c
419
return policy->freq_table[idx].frequency;
drivers/thermal/cpufreq_cooling.c
516
struct cpufreq_policy *policy,
drivers/thermal/cpufreq_cooling.c
527
if (IS_ERR_OR_NULL(policy)) {
drivers/thermal/cpufreq_cooling.c
528
pr_err("%s: cpufreq policy isn't valid: %p\n", __func__, policy);
drivers/thermal/cpufreq_cooling.c
532
dev = get_cpu_device(policy->cpu);
drivers/thermal/cpufreq_cooling.c
534
pr_warn("No cpu device for cpu %d\n", policy->cpu);
drivers/thermal/cpufreq_cooling.c
538
i = cpufreq_table_count_valid_entries(policy);
drivers/thermal/cpufreq_cooling.c
549
cpufreq_cdev->policy = policy;
drivers/thermal/cpufreq_cooling.c
573
if (policy->freq_table_sorted == CPUFREQ_TABLE_UNSORTED) {
drivers/thermal/cpufreq_cooling.c
580
ret = freq_qos_add_request(&policy->constraints,
drivers/thermal/cpufreq_cooling.c
625
cpufreq_cooling_register(struct cpufreq_policy *policy)
drivers/thermal/cpufreq_cooling.c
627
return __cpufreq_cooling_register(NULL, policy, NULL);
drivers/thermal/cpufreq_cooling.c
648
of_cpufreq_cooling_register(struct cpufreq_policy *policy)
drivers/thermal/cpufreq_cooling.c
650
struct device_node *np = of_get_cpu_node(policy->cpu, NULL);
drivers/thermal/cpufreq_cooling.c
655
policy->cpu);
drivers/thermal/cpufreq_cooling.c
660
struct em_perf_domain *em = em_cpu_get(policy->cpu);
drivers/thermal/cpufreq_cooling.c
662
cdev = __cpufreq_cooling_register(np, policy, em);
drivers/thermal/cpufreq_cooling.c
665
policy->cpu, PTR_ERR(cdev));
drivers/thermal/cpufreq_cooling.c
73
struct cpufreq_policy *policy;
drivers/thermal/imx_thermal.c
204
struct cpufreq_policy *policy;
drivers/thermal/imx_thermal.c
546
data->policy = cpufreq_cpu_get(0);
drivers/thermal/imx_thermal.c
547
if (!data->policy) {
drivers/thermal/imx_thermal.c
552
np = of_get_cpu_node(data->policy->cpu, NULL);
drivers/thermal/imx_thermal.c
555
data->cdev = cpufreq_cooling_register(data->policy);
drivers/thermal/imx_thermal.c
558
cpufreq_cpu_put(data->policy);
drivers/thermal/imx_thermal.c
570
cpufreq_cpu_put(data->policy);
drivers/thermal/thermal_core.c
194
char *policy)
drivers/thermal/thermal_core.c
202
gov = __find_governor(strim(policy));
drivers/thermal/thermal_core.c
206
thermal_notify_tz_gov_change(tz, policy);
drivers/thermal/thermal_netlink.c
909
.policy = thermal_genl_policy,
drivers/thermal/thermal_sysfs.c
334
static DEVICE_ATTR_RW(policy);
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
237
data->policy = cpufreq_cpu_get(0);
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
238
if (!data->policy) {
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
244
data->cool_dev = cpufreq_cooling_register(data->policy);
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
249
cpufreq_cpu_put(data->policy);
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
266
if (data->policy)
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
267
cpufreq_cpu_put(data->policy);
drivers/thermal/ti-soc-thermal/ti-thermal-common.c
30
struct cpufreq_policy *policy;
drivers/vdpa/vdpa.c
1588
.policy = vdpa_nl_policy,
fs/btrfs/extent-tree.c
3834
if (ffe_ctl->policy == BTRFS_EXTENT_ALLOC_ZONED)
fs/btrfs/extent-tree.c
4202
switch (ffe_ctl->policy) {
fs/btrfs/extent-tree.c
4216
switch (ffe_ctl->policy) {
fs/btrfs/extent-tree.c
4247
switch (ffe_ctl->policy) {
fs/btrfs/extent-tree.c
4315
switch (ffe_ctl->policy) {
fs/btrfs/extent-tree.c
4405
if (ffe_ctl->policy != BTRFS_EXTENT_ALLOC_CLUSTERED)
fs/btrfs/extent-tree.c
4524
switch (ffe_ctl->policy) {
fs/btrfs/extent-tree.c
4587
ffe_ctl->policy = BTRFS_EXTENT_ALLOC_CLUSTERED;
fs/btrfs/extent-tree.c
4591
ffe_ctl->policy = BTRFS_EXTENT_ALLOC_ZONED;
fs/btrfs/extent-tree.h
86
enum btrfs_extent_allocation_policy policy;
fs/btrfs/sysfs.c
1395
const enum btrfs_read_policy policy = READ_ONCE(fs_devices->read_policy);
fs/btrfs/sysfs.c
1403
if (i == policy)
fs/btrfs/sysfs.c
1417
if (i == policy)
fs/btrfs/volumes.c
6280
const enum btrfs_read_policy policy = READ_ONCE(fs_info->fs_devices->read_policy);
fs/btrfs/volumes.c
6295
switch (policy) {
fs/btrfs/volumes.c
6299
policy);
fs/ceph/crypto.c
79
return ceph_sb_to_fs_client(sb)->fsc_dummy_enc_policy.policy;
fs/crypto/fname.c
166
bool __fscrypt_fname_encrypted_size(const union fscrypt_policy *policy,
fs/crypto/fname.c
170
int padding = 4 << (fscrypt_policy_flags(policy) &
fs/crypto/fscrypt_private.h
155
static inline int fscrypt_policy_size(const union fscrypt_policy *policy)
fs/crypto/fscrypt_private.h
157
switch (policy->version) {
fs/crypto/fscrypt_private.h
159
return sizeof(policy->v1);
fs/crypto/fscrypt_private.h
161
return sizeof(policy->v2);
fs/crypto/fscrypt_private.h
168
fscrypt_policy_contents_mode(const union fscrypt_policy *policy)
fs/crypto/fscrypt_private.h
170
switch (policy->version) {
fs/crypto/fscrypt_private.h
172
return policy->v1.contents_encryption_mode;
fs/crypto/fscrypt_private.h
174
return policy->v2.contents_encryption_mode;
fs/crypto/fscrypt_private.h
181
fscrypt_policy_fnames_mode(const union fscrypt_policy *policy)
fs/crypto/fscrypt_private.h
183
switch (policy->version) {
fs/crypto/fscrypt_private.h
185
return policy->v1.filenames_encryption_mode;
fs/crypto/fscrypt_private.h
187
return policy->v2.filenames_encryption_mode;
fs/crypto/fscrypt_private.h
194
fscrypt_policy_flags(const union fscrypt_policy *policy)
fs/crypto/fscrypt_private.h
196
switch (policy->version) {
fs/crypto/fscrypt_private.h
198
return policy->v1.flags;
fs/crypto/fscrypt_private.h
200
return policy->v2.flags;
fs/crypto/fscrypt_private.h
206
fscrypt_policy_v2_du_bits(const struct fscrypt_policy_v2 *policy,
fs/crypto/fscrypt_private.h
209
return policy->log2_data_unit_size ?: inode->i_blkbits;
fs/crypto/fscrypt_private.h
213
fscrypt_policy_du_bits(const union fscrypt_policy *policy,
fs/crypto/fscrypt_private.h
216
switch (policy->version) {
fs/crypto/fscrypt_private.h
220
return fscrypt_policy_v2_du_bits(&policy->v2, inode);
fs/crypto/fscrypt_private.h
379
bool __fscrypt_fname_encrypted_size(const union fscrypt_policy *policy,
fs/crypto/fscrypt_private.h
777
int fscrypt_policy_to_key_spec(const union fscrypt_policy *policy,
fs/crypto/hooks.c
244
const union fscrypt_policy *policy;
fs/crypto/hooks.c
251
policy = fscrypt_policy_to_inherit(dir);
fs/crypto/hooks.c
252
if (policy == NULL) {
fs/crypto/hooks.c
260
if (IS_ERR(policy))
fs/crypto/hooks.c
261
return PTR_ERR(policy);
fs/crypto/hooks.c
274
if (!__fscrypt_fname_encrypted_size(policy, len,
fs/crypto/keysetup.c
581
const union fscrypt_policy *policy,
fs/crypto/keysetup.c
599
crypt_info->ci_policy = *policy;
fs/crypto/keysetup.c
673
union fscrypt_policy policy;
fs/crypto/keysetup.c
686
res = fscrypt_policy_from_context(&policy, &ctx, res);
fs/crypto/keysetup.c
695
if (!fscrypt_supported_policy(&policy, inode)) {
fs/crypto/keysetup.c
701
res = fscrypt_setup_encryption_info(inode, &policy,
fs/crypto/keysetup.c
737
const union fscrypt_policy *policy;
fs/crypto/keysetup.c
740
policy = fscrypt_policy_to_inherit(dir);
fs/crypto/keysetup.c
741
if (policy == NULL)
fs/crypto/keysetup.c
743
if (IS_ERR(policy))
fs/crypto/keysetup.c
744
return PTR_ERR(policy);
fs/crypto/keysetup.c
764
return fscrypt_setup_encryption_info(inode, policy, nonce,
fs/crypto/keysetup.c
83
select_encryption_mode(const union fscrypt_policy *policy,
fs/crypto/keysetup.c
89
return &fscrypt_modes[fscrypt_policy_contents_mode(policy)];
fs/crypto/keysetup.c
92
return &fscrypt_modes[fscrypt_policy_fnames_mode(policy)];
fs/crypto/policy.c
122
static bool supported_iv_ino_lblk_policy(const struct fscrypt_policy_v2 *policy,
fs/crypto/policy.c
125
const char *type = (policy->flags & FSCRYPT_POLICY_FLAG_IV_INO_LBLK_64)
fs/crypto/policy.c
136
if (policy->contents_encryption_mode != FSCRYPT_MODE_AES_256_XTS) {
fs/crypto/policy.c
174
fscrypt_policy_v2_du_bits(policy, inode)) > 32) {
fs/crypto/policy.c
183
static bool fscrypt_supported_v1_policy(const struct fscrypt_policy_v1 *policy,
fs/crypto/policy.c
186
if (!fscrypt_valid_enc_modes_v1(policy->contents_encryption_mode,
fs/crypto/policy.c
187
policy->filenames_encryption_mode)) {
fs/crypto/policy.c
190
policy->contents_encryption_mode,
fs/crypto/policy.c
191
policy->filenames_encryption_mode);
fs/crypto/policy.c
195
if (policy->flags & ~(FSCRYPT_POLICY_FLAGS_PAD_MASK |
fs/crypto/policy.c
198
policy->flags);
fs/crypto/policy.c
202
if ((policy->flags & FSCRYPT_POLICY_FLAG_DIRECT_KEY) &&
fs/crypto/policy.c
203
!supported_direct_key_modes(inode, policy->contents_encryption_mode,
fs/crypto/policy.c
204
policy->filenames_encryption_mode))
fs/crypto/policy.c
217
static bool fscrypt_supported_v2_policy(const struct fscrypt_policy_v2 *policy,
fs/crypto/policy.c
222
if (!fscrypt_valid_enc_modes_v2(policy->contents_encryption_mode,
fs/crypto/policy.c
223
policy->filenames_encryption_mode)) {
fs/crypto/policy.c
226
policy->contents_encryption_mode,
fs/crypto/policy.c
227
policy->filenames_encryption_mode);
fs/crypto/policy.c
231
if (policy->flags & ~(FSCRYPT_POLICY_FLAGS_PAD_MASK |
fs/crypto/policy.c
236
policy->flags);
fs/crypto/policy.c
240
count += !!(policy->flags & FSCRYPT_POLICY_FLAG_DIRECT_KEY);
fs/crypto/policy.c
241
count += !!(policy->flags & FSCRYPT_POLICY_FLAG_IV_INO_LBLK_64);
fs/crypto/policy.c
242
count += !!(policy->flags & FSCRYPT_POLICY_FLAG_IV_INO_LBLK_32);
fs/crypto/policy.c
245
policy->flags);
fs/crypto/policy.c
249
if (policy->log2_data_unit_size) {
fs/crypto/policy.c
255
if (policy->log2_data_unit_size > inode->i_blkbits ||
fs/crypto/policy.c
256
policy->log2_data_unit_size < SECTOR_SHIFT /* 9 */) {
fs/crypto/policy.c
259
policy->log2_data_unit_size);
fs/crypto/policy.c
262
if (policy->log2_data_unit_size != inode->i_blkbits &&
fs/crypto/policy.c
263
(policy->flags & FSCRYPT_POLICY_FLAG_IV_INO_LBLK_32)) {
fs/crypto/policy.c
274
if ((policy->flags & FSCRYPT_POLICY_FLAG_DIRECT_KEY) &&
fs/crypto/policy.c
275
!supported_direct_key_modes(inode, policy->contents_encryption_mode,
fs/crypto/policy.c
276
policy->filenames_encryption_mode))
fs/crypto/policy.c
279
if ((policy->flags & (FSCRYPT_POLICY_FLAG_IV_INO_LBLK_64 |
fs/crypto/policy.c
281
!supported_iv_ino_lblk_policy(policy, inode))
fs/crypto/policy.c
284
if (memchr_inv(policy->__reserved, 0, sizeof(policy->__reserved))) {
fs/crypto/policy.c
335
const struct fscrypt_policy_v1 *policy = &policy_u->v1;
fs/crypto/policy.c
340
policy->contents_encryption_mode;
fs/crypto/policy.c
342
policy->filenames_encryption_mode;
fs/crypto/policy.c
343
ctx->flags = policy->flags;
fs/crypto/policy.c
345
policy->master_key_descriptor,
fs/crypto/policy.c
351
const struct fscrypt_policy_v2 *policy = &policy_u->v2;
fs/crypto/policy.c
356
policy->contents_encryption_mode;
fs/crypto/policy.c
358
policy->filenames_encryption_mode;
fs/crypto/policy.c
359
ctx->flags = policy->flags;
fs/crypto/policy.c
360
ctx->log2_data_unit_size = policy->log2_data_unit_size;
fs/crypto/policy.c
362
policy->master_key_identifier,
fs/crypto/policy.c
38
int fscrypt_policy_to_key_spec(const union fscrypt_policy *policy,
fs/crypto/policy.c
398
struct fscrypt_policy_v1 *policy = &policy_u->v1;
fs/crypto/policy.c
400
policy->version = FSCRYPT_POLICY_V1;
fs/crypto/policy.c
401
policy->contents_encryption_mode =
fs/crypto/policy.c
403
policy->filenames_encryption_mode =
fs/crypto/policy.c
405
policy->flags = ctx->flags;
fs/crypto/policy.c
406
memcpy(policy->master_key_descriptor,
fs/crypto/policy.c
408
sizeof(policy->master_key_descriptor));
fs/crypto/policy.c
41
switch (policy->version) {
fs/crypto/policy.c
413
struct fscrypt_policy_v2 *policy = &policy_u->v2;
fs/crypto/policy.c
415
policy->version = FSCRYPT_POLICY_V2;
fs/crypto/policy.c
416
policy->contents_encryption_mode =
fs/crypto/policy.c
418
policy->filenames_encryption_mode =
fs/crypto/policy.c
420
policy->flags = ctx->flags;
fs/crypto/policy.c
421
policy->log2_data_unit_size = ctx->log2_data_unit_size;
fs/crypto/policy.c
422
memcpy(policy->__reserved, ctx->__reserved,
fs/crypto/policy.c
423
sizeof(policy->__reserved));
fs/crypto/policy.c
424
memcpy(policy->master_key_identifier,
fs/crypto/policy.c
426
sizeof(policy->master_key_identifier));
fs/crypto/policy.c
435
static int fscrypt_get_policy(struct inode *inode, union fscrypt_policy *policy)
fs/crypto/policy.c
44
memcpy(key_spec->u.descriptor, policy->v1.master_key_descriptor,
fs/crypto/policy.c
444
*policy = ci->ci_policy;
fs/crypto/policy.c
455
return fscrypt_policy_from_context(policy, &ctx, ret);
fs/crypto/policy.c
459
const union fscrypt_policy *policy)
fs/crypto/policy.c
466
if (!fscrypt_supported_policy(policy, inode))
fs/crypto/policy.c
469
switch (policy->version) {
fs/crypto/policy.c
487
policy->v2.master_key_identifier);
fs/crypto/policy.c
49
memcpy(key_spec->u.identifier, policy->v2.master_key_identifier,
fs/crypto/policy.c
490
if (policy->v2.flags & FSCRYPT_POLICY_FLAG_IV_INO_LBLK_32)
fs/crypto/policy.c
500
ctxsize = fscrypt_new_context(&ctx, policy, nonce);
fs/crypto/policy.c
507
union fscrypt_policy policy;
fs/crypto/policy.c
514
if (get_user(policy.version, (const u8 __user *)arg))
fs/crypto/policy.c
517
size = fscrypt_policy_size(&policy);
fs/crypto/policy.c
532
version = policy.version;
fs/crypto/policy.c
533
if (copy_from_user(&policy, arg, size))
fs/crypto/policy.c
535
policy.version = version;
fs/crypto/policy.c
555
ret = set_encryption_policy(inode, &policy);
fs/crypto/policy.c
557
(ret == 0 && !fscrypt_policies_equal(&policy,
fs/crypto/policy.c
573
union fscrypt_policy policy;
fs/crypto/policy.c
576
err = fscrypt_get_policy(file_inode(filp), &policy);
fs/crypto/policy.c
580
if (policy.version != FSCRYPT_POLICY_V1)
fs/crypto/policy.c
583
if (copy_to_user(arg, &policy, sizeof(policy.v1)))
fs/crypto/policy.c
593
union fscrypt_policy *policy = (union fscrypt_policy *)&arg.policy;
fs/crypto/policy.c
600
offsetof(typeof(arg), policy));
fs/crypto/policy.c
601
BUILD_BUG_ON(sizeof(arg.policy) != sizeof(*policy));
fs/crypto/policy.c
603
err = fscrypt_get_policy(file_inode(filp), policy);
fs/crypto/policy.c
606
policy_size = fscrypt_policy_size(policy);
fs/crypto/policy.c
810
union fscrypt_policy *policy;
fs/crypto/policy.c
816
policy = kzalloc_obj(*policy);
fs/crypto/policy.c
817
if (!policy)
fs/crypto/policy.c
821
policy->version = FSCRYPT_POLICY_V1;
fs/crypto/policy.c
822
policy->v1.contents_encryption_mode = FSCRYPT_MODE_AES_256_XTS;
fs/crypto/policy.c
823
policy->v1.filenames_encryption_mode = FSCRYPT_MODE_AES_256_CTS;
fs/crypto/policy.c
824
memset(policy->v1.master_key_descriptor, 0x42,
fs/crypto/policy.c
827
policy->version = FSCRYPT_POLICY_V2;
fs/crypto/policy.c
828
policy->v2.contents_encryption_mode = FSCRYPT_MODE_AES_256_XTS;
fs/crypto/policy.c
829
policy->v2.filenames_encryption_mode = FSCRYPT_MODE_AES_256_CTS;
fs/crypto/policy.c
831
policy->v2.master_key_identifier);
fs/crypto/policy.c
837
if (dummy_policy->policy) {
fs/crypto/policy.c
838
if (fscrypt_policies_equal(policy, dummy_policy->policy))
fs/crypto/policy.c
844
dummy_policy->policy = policy;
fs/crypto/policy.c
845
policy = NULL;
fs/crypto/policy.c
848
kfree(policy);
fs/crypto/policy.c
863
if (!p1->policy && !p2->policy)
fs/crypto/policy.c
865
if (!p1->policy || !p2->policy)
fs/crypto/policy.c
867
return fscrypt_policies_equal(p1->policy, p2->policy);
fs/crypto/policy.c
883
const union fscrypt_policy *policy = fscrypt_get_dummy_policy(sb);
fs/crypto/policy.c
886
if (!policy)
fs/crypto/policy.c
889
vers = policy->version;
fs/ext4/crypto.c
228
return EXT4_SB(sb)->s_dummy_enc_policy.policy;
fs/f2fs/debug.c
477
unsigned long policy = SM_I(sbi)->ipu_policy;
fs/f2fs/debug.c
479
for_each_set_bit(j, &policy, F2FS_IPU_MAX)
fs/f2fs/super.c
3710
return F2FS_OPTION(F2FS_SB(sb)).dummy_enc_policy.policy;
fs/lockd/netlink.c
26
.policy = lockd_server_set_nl_policy,
fs/nfs/client.c
339
if (clp->cl_xprtsec.policy != data->xprtsec.policy)
fs/nfs/client.c
341
if (clp->cl_xprtsec.policy == RPC_XPRTSEC_TLS_X509) {
fs/nfs/fs_context.c
1752
ctx->xprtsec.policy = RPC_XPRTSEC_NONE;
fs/nfs/fs_context.c
375
if (ctx->xprtsec.policy != RPC_XPRTSEC_NONE)
fs/nfs/fs_context.c
505
ctx->xprtsec.policy = RPC_XPRTSEC_NONE;
fs/nfs/fs_context.c
508
ctx->xprtsec.policy = RPC_XPRTSEC_TLS_ANON;
fs/nfs/fs_context.c
511
ctx->xprtsec.policy = RPC_XPRTSEC_TLS_X509;
fs/nfs/nfs3client.c
103
.policy = RPC_XPRTSEC_NONE,
fs/nfs/nfs3client.c
120
if (mds_clp->cl_xprtsec.policy != RPC_XPRTSEC_NONE)
fs/nfs/nfs4client.c
1100
if (parent_client->cl_xprtsec.policy != RPC_XPRTSEC_NONE)
fs/nfs/nfs4client.c
810
.policy = RPC_XPRTSEC_NONE,
fs/nfs/nfs4client.c
823
if (mds_srv->nfs_client->cl_xprtsec.policy != RPC_XPRTSEC_NONE)
fs/nfs/pnfs_nfs.c
813
enum xprtsec_policies xprtsec_policy = mds_clp->cl_xprtsec.policy;
fs/nfs/pnfs_nfs.c
888
enum xprtsec_policies xprtsec_policy = mds_clp->cl_xprtsec.policy;
fs/nfs/super.c
502
switch (clp->cl_xprtsec.policy) {
fs/nfsd/netlink.c
60
.policy = nfsd_threads_set_nl_policy,
fs/nfsd/netlink.c
72
.policy = nfsd_version_set_nl_policy,
fs/nfsd/netlink.c
84
.policy = nfsd_listener_set_nl_policy,
fs/nfsd/netlink.c
96
.policy = nfsd_pool_mode_set_nl_policy,
fs/proc/array.c
638
seq_put_decimal_ull(m, " ", task->policy);
fs/smb/client/netlink.c
51
.policy = cifs_genl_policy,
fs/smb/server/transport_ipc.c
226
ksmbd_genl_family.policy = ksmbd_nl_policy;
include/acpi/processor.h
323
void acpi_processor_ppc_init(struct cpufreq_policy *policy);
include/acpi/processor.h
324
void acpi_processor_ppc_exit(struct cpufreq_policy *policy);
include/acpi/processor.h
332
static inline void acpi_processor_ppc_init(struct cpufreq_policy *policy)
include/acpi/processor.h
336
static inline void acpi_processor_ppc_exit(struct cpufreq_policy *policy)
include/acpi/processor.h
438
void acpi_thermal_cpufreq_init(struct cpufreq_policy *policy);
include/acpi/processor.h
439
void acpi_thermal_cpufreq_exit(struct cpufreq_policy *policy);
include/acpi/processor.h
441
static inline void acpi_thermal_cpufreq_init(struct cpufreq_policy *policy)
include/acpi/processor.h
445
static inline void acpi_thermal_cpufreq_exit(struct cpufreq_policy *policy)
include/linux/cpu_cooling.h
27
cpufreq_cooling_register(struct cpufreq_policy *policy);
include/linux/cpu_cooling.h
40
of_cpufreq_cooling_register(struct cpufreq_policy *policy);
include/linux/cpu_cooling.h
44
cpufreq_cooling_register(struct cpufreq_policy *policy)
include/linux/cpu_cooling.h
56
of_cpufreq_cooling_register(struct cpufreq_policy *policy)
include/linux/cpufreq.h
1026
static inline int find_index_c(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1033
if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
include/linux/cpufreq.h
1034
return cpufreq_table_find_index_ac(policy, target_freq,
include/linux/cpufreq.h
1037
return cpufreq_table_find_index_dc(policy, target_freq,
include/linux/cpufreq.h
1042
static inline int cpufreq_table_find_index_c(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1046
return find_index_c(policy, target_freq, policy->min, policy->max, efficiencies);
include/linux/cpufreq.h
1049
static inline bool cpufreq_is_in_limits(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1058
freq = policy->freq_table[idx].frequency;
include/linux/cpufreq.h
1063
static inline int cpufreq_frequency_table_target(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1069
bool efficiencies = policy->efficiencies_available &&
include/linux/cpufreq.h
1076
if (unlikely(policy->freq_table_sorted == CPUFREQ_TABLE_UNSORTED))
include/linux/cpufreq.h
1077
return cpufreq_table_index_unsorted(policy, target_freq, min,
include/linux/cpufreq.h
1082
idx = find_index_l(policy, target_freq, min, max, efficiencies);
include/linux/cpufreq.h
1085
idx = find_index_h(policy, target_freq, min, max, efficiencies);
include/linux/cpufreq.h
1088
idx = find_index_c(policy, target_freq, min, max, efficiencies);
include/linux/cpufreq.h
1096
if (!cpufreq_is_in_limits(policy, min, max, idx) && efficiencies) {
include/linux/cpufreq.h
1104
static inline int cpufreq_table_count_valid_entries(const struct cpufreq_policy *policy)
include/linux/cpufreq.h
1109
if (unlikely(!policy->freq_table))
include/linux/cpufreq.h
1112
cpufreq_for_each_valid_entry(pos, policy->freq_table)
include/linux/cpufreq.h
1129
cpufreq_table_set_inefficient(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1135
if (policy->freq_table_sorted == CPUFREQ_TABLE_UNSORTED)
include/linux/cpufreq.h
1138
cpufreq_for_each_valid_entry(pos, policy->freq_table) {
include/linux/cpufreq.h
1141
policy->efficiencies_available = true;
include/linux/cpufreq.h
1201
static inline int cpufreq_boost_set_sw(struct cpufreq_policy *policy, int state)
include/linux/cpufreq.h
1207
cpufreq_table_set_inefficient(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1235
int cpufreq_table_validate_and_sort(struct cpufreq_policy *policy);
include/linux/cpufreq.h
1238
void cpufreq_generic_init(struct cpufreq_policy *policy,
include/linux/cpufreq.h
1244
static inline void cpufreq_register_em_with_opp(struct cpufreq_policy *policy)
include/linux/cpufreq.h
1246
dev_pm_opp_of_register_em(get_cpu_device(policy->cpu),
include/linux/cpufreq.h
1247
policy->related_cpus);
include/linux/cpufreq.h
192
struct cpufreq_policy *policy;
include/linux/cpufreq.h
208
void cpufreq_cpu_put(struct cpufreq_policy *policy);
include/linux/cpufreq.h
222
static inline void cpufreq_cpu_put(struct cpufreq_policy *policy) { }
include/linux/cpufreq.h
228
static inline bool policy_is_inactive(struct cpufreq_policy *policy)
include/linux/cpufreq.h
230
return cpumask_empty(policy->cpus);
include/linux/cpufreq.h
233
static inline bool policy_is_shared(struct cpufreq_policy *policy)
include/linux/cpufreq.h
235
return cpumask_weight(policy->cpus) > 1;
include/linux/cpufreq.h
247
void refresh_frequency_limits(struct cpufreq_policy *policy);
include/linux/cpufreq.h
252
struct kobject *get_governor_parent_kobj(struct cpufreq_policy *policy);
include/linux/cpufreq.h
253
void cpufreq_enable_fast_switch(struct cpufreq_policy *policy);
include/linux/cpufreq.h
254
void cpufreq_disable_fast_switch(struct cpufreq_policy *policy);
include/linux/cpufreq.h
292
void cpufreq_stats_create_table(struct cpufreq_policy *policy);
include/linux/cpufreq.h
293
void cpufreq_stats_free_table(struct cpufreq_policy *policy);
include/linux/cpufreq.h
294
void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
include/linux/cpufreq.h
297
static inline void cpufreq_stats_create_table(struct cpufreq_policy *policy) { }
include/linux/cpufreq.h
298
static inline void cpufreq_stats_free_table(struct cpufreq_policy *policy) { }
include/linux/cpufreq.h
299
static inline void cpufreq_stats_record_transition(struct cpufreq_policy *policy,
include/linux/cpufreq.h
354
int (*init)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
355
int (*verify)(struct cpufreq_policy_data *policy);
include/linux/cpufreq.h
358
int (*setpolicy)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
360
int (*target)(struct cpufreq_policy *policy,
include/linux/cpufreq.h
363
int (*target_index)(struct cpufreq_policy *policy,
include/linux/cpufreq.h
365
unsigned int (*fast_switch)(struct cpufreq_policy *policy,
include/linux/cpufreq.h
395
unsigned int (*get_intermediate)(struct cpufreq_policy *policy,
include/linux/cpufreq.h
397
int (*target_intermediate)(struct cpufreq_policy *policy,
include/linux/cpufreq.h
404
void (*update_limits)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
409
int (*online)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
410
int (*offline)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
411
void (*exit)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
412
int (*suspend)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
413
int (*resume)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
416
void (*ready)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
422
int (*set_boost)(struct cpufreq_policy *policy, int state);
include/linux/cpufreq.h
428
void (*register_em)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
493
static inline void cpufreq_verify_within_limits(struct cpufreq_policy_data *policy,
include/linux/cpufreq.h
497
policy->max = clamp(policy->max, min, max);
include/linux/cpufreq.h
498
policy->min = clamp(policy->min, min, policy->max);
include/linux/cpufreq.h
502
cpufreq_verify_within_cpu_limits(struct cpufreq_policy_data *policy)
include/linux/cpufreq.h
504
cpufreq_verify_within_limits(policy, policy->cpuinfo.min_freq,
include/linux/cpufreq.h
505
policy->cpuinfo.max_freq);
include/linux/cpufreq.h
511
int cpufreq_generic_suspend(struct cpufreq_policy *policy);
include/linux/cpufreq.h
536
void cpufreq_freq_transition_begin(struct cpufreq_policy *policy,
include/linux/cpufreq.h
538
void cpufreq_freq_transition_end(struct cpufreq_policy *policy,
include/linux/cpufreq.h
594
int (*init)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
595
void (*exit)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
596
int (*start)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
597
void (*stop)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
598
void (*limits)(struct cpufreq_policy *policy);
include/linux/cpufreq.h
599
ssize_t (*show_setspeed) (struct cpufreq_policy *policy,
include/linux/cpufreq.h
601
int (*store_setspeed) (struct cpufreq_policy *policy,
include/linux/cpufreq.h
618
unsigned int cpufreq_driver_fast_switch(struct cpufreq_policy *policy,
include/linux/cpufreq.h
625
int cpufreq_driver_target(struct cpufreq_policy *policy,
include/linux/cpufreq.h
628
int __cpufreq_driver_target(struct cpufreq_policy *policy,
include/linux/cpufreq.h
631
unsigned int cpufreq_driver_resolve_freq(struct cpufreq_policy *policy,
include/linux/cpufreq.h
633
unsigned int cpufreq_policy_transition_delay_us(struct cpufreq_policy *policy);
include/linux/cpufreq.h
636
int cpufreq_start_governor(struct cpufreq_policy *policy);
include/linux/cpufreq.h
637
void cpufreq_stop_governor(struct cpufreq_policy *policy);
include/linux/cpufreq.h
657
bool sugov_is_governor(struct cpufreq_policy *policy);
include/linux/cpufreq.h
659
static inline bool sugov_is_governor(struct cpufreq_policy *policy)
include/linux/cpufreq.h
665
static inline void cpufreq_policy_apply_limits(struct cpufreq_policy *policy)
include/linux/cpufreq.h
667
if (policy->max < policy->cur)
include/linux/cpufreq.h
668
__cpufreq_driver_target(policy, policy->max,
include/linux/cpufreq.h
670
else if (policy->min > policy->cur)
include/linux/cpufreq.h
671
__cpufreq_driver_target(policy, policy->min,
include/linux/cpufreq.h
72
unsigned int policy; /* see above */
include/linux/cpufreq.h
786
int cpufreq_frequency_table_cpuinfo(struct cpufreq_policy *policy);
include/linux/cpufreq.h
788
int cpufreq_frequency_table_verify(struct cpufreq_policy_data *policy);
include/linux/cpufreq.h
790
int cpufreq_generic_frequency_table_verify(struct cpufreq_policy_data *policy);
include/linux/cpufreq.h
792
int cpufreq_table_index_unsorted(struct cpufreq_policy *policy,
include/linux/cpufreq.h
795
int cpufreq_frequency_table_get_index(struct cpufreq_policy *policy,
include/linux/cpufreq.h
802
int cpufreq_boost_set_sw(struct cpufreq_policy *policy, int state);
include/linux/cpufreq.h
805
static inline int cpufreq_table_find_index_al(struct cpufreq_policy *policy,
include/linux/cpufreq.h
809
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/cpufreq.h
827
static inline int cpufreq_table_find_index_dl(struct cpufreq_policy *policy,
include/linux/cpufreq.h
831
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/cpufreq.h
857
static inline int find_index_l(struct cpufreq_policy *policy,
include/linux/cpufreq.h
864
if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
include/linux/cpufreq.h
865
return cpufreq_table_find_index_al(policy, target_freq,
include/linux/cpufreq.h
868
return cpufreq_table_find_index_dl(policy, target_freq,
include/linux/cpufreq.h
873
static inline int cpufreq_table_find_index_l(struct cpufreq_policy *policy,
include/linux/cpufreq.h
877
return find_index_l(policy, target_freq, policy->min, policy->max, efficiencies);
include/linux/cpufreq.h
881
static inline int cpufreq_table_find_index_ah(struct cpufreq_policy *policy,
include/linux/cpufreq.h
885
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/cpufreq.h
912
static inline int cpufreq_table_find_index_dh(struct cpufreq_policy *policy,
include/linux/cpufreq.h
916
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/cpufreq.h
933
static inline int find_index_h(struct cpufreq_policy *policy,
include/linux/cpufreq.h
940
if (policy->freq_table_sorted == CPUFREQ_TABLE_SORTED_ASCENDING)
include/linux/cpufreq.h
941
return cpufreq_table_find_index_ah(policy, target_freq,
include/linux/cpufreq.h
944
return cpufreq_table_find_index_dh(policy, target_freq,
include/linux/cpufreq.h
949
static inline int cpufreq_table_find_index_h(struct cpufreq_policy *policy,
include/linux/cpufreq.h
953
return find_index_h(policy, target_freq, policy->min, policy->max, efficiencies);
include/linux/cpufreq.h
957
static inline int cpufreq_table_find_index_ac(struct cpufreq_policy *policy,
include/linux/cpufreq.h
961
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/cpufreq.h
992
static inline int cpufreq_table_find_index_dc(struct cpufreq_policy *policy,
include/linux/cpufreq.h
996
struct cpufreq_frequency_table *table = policy->freq_table;
include/linux/ethtool.h
682
enum ethtool_module_power_mode_policy policy;
include/linux/firmware/qcom/qcom_tzmem.h
46
enum qcom_tzmem_policy policy;
include/linux/fscrypt.h
392
const union fscrypt_policy *policy;
include/linux/fscrypt.h
404
return dummy_policy->policy != NULL;
include/linux/fscrypt.h
409
kfree(dummy_policy->policy);
include/linux/fscrypt.h
410
dummy_policy->policy = NULL;
include/linux/genl_magic_func.h
294
.policy = CONCATENATE(GENL_MAGIC_FAMILY, _tla_nl_policy),
include/linux/ioprio.h
41
if (task->policy == SCHED_IDLE)
include/linux/kernel_read_file.h
15
id(POLICY, security-policy) \
include/linux/libata.h
1306
extern int sata_link_scr_lpm(struct ata_link *link, enum ata_lpm_policy policy,
include/linux/libata.h
999
int (*set_lpm)(struct ata_link *link, enum ata_lpm_policy policy,
include/linux/mempolicy.h
120
struct mempolicy *policy;
include/linux/mempolicy.h
181
extern bool apply_policy_zone(struct mempolicy *policy, enum zone_type zone);
include/linux/mfd/wm8350/supply.h
124
struct wm8350_charger_policy *policy;
include/linux/mlx5/driver.h
472
enum port_state_policy policy;
include/linux/mlx5/driver.h
868
enum port_state_policy policy;
include/linux/netfilter/nfnetlink.h
28
const struct nla_policy *policy;
include/linux/netlink.h
146
(extack)->policy = (pol); \
include/linux/netlink.h
161
__extack->policy = (pol); \
include/linux/netlink.h
181
__extack->policy = (pol); \
include/linux/netlink.h
85
const struct nla_policy *policy;
include/linux/page-flags.h
449
#define TESTPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
450
FOLIO_TEST_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
452
{ return test_bit(PG_##lname, &policy(page, 0)->flags.f); }
include/linux/page-flags.h
454
#define SETPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
455
FOLIO_SET_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
457
{ set_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
459
#define CLEARPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
460
FOLIO_CLEAR_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
462
{ clear_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
464
#define __SETPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
465
__FOLIO_SET_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
467
{ __set_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
469
#define __CLEARPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
470
__FOLIO_CLEAR_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
472
{ __clear_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
474
#define TESTSETFLAG(uname, lname, policy) \
include/linux/page-flags.h
475
FOLIO_TEST_SET_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
477
{ return test_and_set_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
479
#define TESTCLEARFLAG(uname, lname, policy) \
include/linux/page-flags.h
480
FOLIO_TEST_CLEAR_FLAG(lname, FOLIO_##policy) \
include/linux/page-flags.h
482
{ return test_and_clear_bit(PG_##lname, &policy(page, 1)->flags.f); }
include/linux/page-flags.h
484
#define PAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
485
TESTPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
486
SETPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
487
CLEARPAGEFLAG(uname, lname, policy)
include/linux/page-flags.h
489
#define __PAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
490
TESTPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
491
__SETPAGEFLAG(uname, lname, policy) \
include/linux/page-flags.h
492
__CLEARPAGEFLAG(uname, lname, policy)
include/linux/page-flags.h
494
#define TESTSCFLAG(uname, lname, policy) \
include/linux/page-flags.h
495
TESTSETFLAG(uname, lname, policy) \
include/linux/page-flags.h
496
TESTCLEARFLAG(uname, lname, policy)
include/linux/pagemap.h
655
struct mempolicy *policy);
include/linux/pagemap.h
658
struct mempolicy *policy)
include/linux/pagemap.h
756
pgoff_t index, fgf_t fgf_flags, gfp_t gfp, struct mempolicy *policy);
include/linux/psp-sev.h
294
u32 policy; /* Out */
include/linux/psp-sev.h
311
u32 policy; /* In */
include/linux/psp-sev.h
410
u32 policy; /* Out */
include/linux/psp-sev.h
500
u32 policy; /* In */
include/linux/psp-sev.h
643
u64 policy; /* In */
include/linux/sched.h
920
unsigned int policy;
include/linux/sched/cpufreq.h
24
bool cpufreq_this_cpu_can_update(struct cpufreq_policy *policy);
include/linux/sched/rt.h
44
int policy = tsk->policy;
include/linux/sched/rt.h
46
if (policy == SCHED_FIFO || policy == SCHED_RR)
include/linux/sched/rt.h
48
if (policy == SCHED_DEADLINE)
include/linux/shmem_fs.h
50
struct shared_policy policy; /* NUMA memory alloc policy */
include/linux/sunrpc/xprt.h
145
enum xprtsec_policies policy;
include/linux/syscalls.h
647
asmlinkage long sys_sched_setscheduler(pid_t pid, int policy,
include/linux/syscalls.h
657
asmlinkage long sys_sched_get_priority_max(int policy);
include/linux/syscalls.h
658
asmlinkage long sys_sched_get_priority_min(int policy);
include/linux/syscalls.h
853
asmlinkage long sys_get_mempolicy(int __user *policy,
include/net/bluetooth/hci.h
1097
__le16 policy;
include/net/bluetooth/hci.h
1103
__le16 policy;
include/net/bluetooth/hci.h
1113
__le16 policy;
include/net/bluetooth/hci.h
1118
__le16 policy;
include/net/cfg80211.h
5646
const struct nla_policy *policy;
include/net/genetlink.h
220
const struct nla_policy *policy;
include/net/genetlink.h
272
const struct nla_policy *policy;
include/net/genetlink.h
390
const struct nla_policy *policy,
include/net/genetlink.h
394
policy, NL_VALIDATE_LIBERAL, extack);
include/net/genetlink.h
409
const struct nla_policy *policy,
include/net/genetlink.h
413
policy, NL_VALIDATE_STRICT, extack);
include/net/genetlink.h
90
const struct nla_policy *policy;
include/net/netfilter/nf_tables.h
1248
u8 policy;
include/net/netfilter/nf_tables.h
1434
const struct nla_policy *policy;
include/net/netfilter/nf_tables.h
1747
u8 policy;
include/net/netfilter/nf_tables.h
1766
nft_trans_container_chain(trans)->policy
include/net/netfilter/nf_tables.h
364
u32 policy;
include/net/netfilter/nf_tables.h
603
u16 policy;
include/net/netfilter/nf_tables.h
911
const struct nla_policy *policy;
include/net/netlink.h
1352
const struct nla_policy *policy,
include/net/netlink.h
1360
return __nla_parse(tb, maxtype, nla_data(nla), nla_len(nla), policy,
include/net/netlink.h
1376
const struct nla_policy *policy,
include/net/netlink.h
1379
return __nla_parse(tb, maxtype, nla_data(nla), nla_len(nla), policy,
include/net/netlink.h
2309
const struct nla_policy *policy,
include/net/netlink.h
2313
return __nla_validate(nla_data(start), nla_len(start), maxtype, policy,
include/net/netlink.h
2319
const struct nla_policy *policy,
include/net/netlink.h
2322
return __nla_validate_nested(start, maxtype, policy,
include/net/netlink.h
2328
const struct nla_policy *policy,
include/net/netlink.h
2331
return __nla_validate_nested(start, maxtype, policy,
include/net/netlink.h
2452
const struct nla_policy *policy,
include/net/netlink.h
2455
const struct nla_policy *policy,
include/net/netlink.h
392
#define _NLA_POLICY_NESTED(maxattr, policy) \
include/net/netlink.h
393
{ .type = NLA_NESTED, .nested_policy = policy, .len = maxattr }
include/net/netlink.h
394
#define _NLA_POLICY_NESTED_ARRAY(maxattr, policy) \
include/net/netlink.h
395
{ .type = NLA_NESTED_ARRAY, .nested_policy = policy, .len = maxattr }
include/net/netlink.h
396
#define NLA_POLICY_NESTED(policy) \
include/net/netlink.h
397
_NLA_POLICY_NESTED(ARRAY_SIZE(policy) - 1, policy)
include/net/netlink.h
398
#define NLA_POLICY_NESTED_ARRAY(policy) \
include/net/netlink.h
399
_NLA_POLICY_NESTED_ARRAY(ARRAY_SIZE(policy) - 1, policy)
include/net/netlink.h
542
const struct nla_policy *policy, unsigned int validate,
include/net/netlink.h
545
int len, const struct nla_policy *policy, unsigned int validate,
include/net/netlink.h
708
const struct nla_policy *policy,
include/net/netlink.h
711
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
733
const struct nla_policy *policy,
include/net/netlink.h
736
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
759
const struct nla_policy *policy,
include/net/netlink.h
762
return __nla_parse(tb, maxtype, head, len, policy,
include/net/netlink.h
780
const struct nla_policy *policy,
include/net/netlink.h
790
nlmsg_attrlen(nlh, hdrlen), policy, validate,
include/net/netlink.h
807
const struct nla_policy *policy,
include/net/netlink.h
810
return __nlmsg_parse(nlh, hdrlen, tb, maxtype, policy,
include/net/netlink.h
827
const struct nla_policy *policy,
include/net/netlink.h
830
return __nlmsg_parse(nlh, hdrlen, tb, maxtype, policy,
include/net/netlink.h
848
const struct nla_policy *policy,
include/net/netlink.h
851
return __nlmsg_parse(nlh, hdrlen, tb, maxtype, policy,
include/net/netlink.h
886
const struct nla_policy *policy,
include/net/netlink.h
889
return __nla_validate(head, len, maxtype, policy, NL_VALIDATE_LIBERAL,
include/net/netlink.h
908
const struct nla_policy *policy,
include/net/netlink.h
911
return __nla_validate(head, len, maxtype, policy, NL_VALIDATE_STRICT,
include/net/netlink.h
925
const struct nla_policy *policy,
include/net/netlink.h
933
policy, NL_VALIDATE_LIBERAL, extack);
include/net/rtnetlink.h
157
const struct nla_policy *policy;
include/net/xfrm.h
1867
int xfrm_policy_insert(int dir, struct xfrm_policy *policy, int excl);
include/net/xfrm.h
2250
int maxtype, const struct nla_policy *policy,
include/net/xfrm.h
898
static inline void xfrm_pol_hold(struct xfrm_policy *policy)
include/net/xfrm.h
900
if (likely(policy != NULL))
include/net/xfrm.h
901
refcount_inc(&policy->refcnt);
include/net/xfrm.h
904
void xfrm_policy_destroy(struct xfrm_policy *policy);
include/net/xfrm.h
906
static inline void xfrm_pol_put(struct xfrm_policy *policy)
include/net/xfrm.h
908
if (refcount_dec_and_test(&policy->refcnt))
include/net/xfrm.h
909
xfrm_policy_destroy(policy);
include/trace/events/power.h
194
TP_PROTO(struct cpufreq_policy *policy),
include/trace/events/power.h
196
TP_ARGS(policy),
include/trace/events/power.h
205
__entry->min_freq = policy->min;
include/trace/events/power.h
206
__entry->max_freq = policy->max;
include/trace/events/power.h
207
__entry->cpu_id = policy->cpu;
include/trace/events/sunrpc.h
143
#define rpc_show_xprtsec_policy(policy) \
include/trace/events/sunrpc.h
144
__print_symbolic(policy, \
include/trace/events/sunrpc.h
1545
__entry->requested_policy = clnt->cl_xprtsec.policy;
include/trace/events/sunrpc.h
187
__entry->xprtsec = args->xprtsec.policy;
include/uapi/linux/fscrypt.h
86
} policy; /* output */
include/uapi/linux/netfilter_bridge/ebtables.h
86
int policy;
include/uapi/linux/xfrm.h
462
struct xfrm_userpolicy_info policy;
init/init_task.c
108
.policy = SCHED_NORMAL,
kernel/power/em_netlink_autogen.c
29
.policy = dev_energymodel_get_perf_domains_nl_policy,
kernel/power/em_netlink_autogen.c
41
.policy = dev_energymodel_get_perf_table_nl_policy,
kernel/power/energy_model.c
490
struct cpufreq_policy *policy;
kernel/power/energy_model.c
504
policy = cpufreq_cpu_get(cpu);
kernel/power/energy_model.c
505
if (!policy) {
kernel/power/energy_model.c
514
if (!cpufreq_table_set_inefficient(policy, table[i].frequency))
kernel/power/energy_model.c
518
cpufreq_cpu_put(policy);
kernel/power/energy_model.c
863
struct cpufreq_policy *policy;
kernel/power/energy_model.c
870
policy = cpufreq_cpu_get(cpu);
kernel/power/energy_model.c
871
if (!policy) {
kernel/power/energy_model.c
875
cpufreq_cpu_put(policy);
kernel/sched/core.c
4649
p->policy = SCHED_NORMAL;
kernel/sched/core.c
4675
} else if (task_should_scx(p->policy)) {
kernel/sched/core.c
7254
const struct sched_class *__setscheduler_class(int policy, int prio)
kernel/sched/core.c
7263
if (task_should_scx(policy))
kernel/sched/core.c
7371
next_class = __setscheduler_class(p->policy, prio);
kernel/sched/cpufreq.c
70
bool cpufreq_this_cpu_can_update(struct cpufreq_policy *policy)
kernel/sched/cpufreq.c
72
return cpumask_test_cpu(smp_processor_id(), policy->cpus) ||
kernel/sched/cpufreq.c
73
(policy->dvfs_possible_from_any_cpu &&
kernel/sched/cpufreq_schedutil.c
154
unsigned long get_capacity_ref_freq(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
156
unsigned int freq = arch_scale_freq_ref(policy->cpu);
kernel/sched/cpufreq_schedutil.c
162
return policy->cpuinfo.max_freq;
kernel/sched/cpufreq_schedutil.c
168
return policy->cur + (policy->cur >> 2);
kernel/sched/cpufreq_schedutil.c
19
struct cpufreq_policy *policy;
kernel/sched/cpufreq_schedutil.c
196
struct cpufreq_policy *policy = sg_policy->policy;
kernel/sched/cpufreq_schedutil.c
199
freq = get_capacity_ref_freq(policy);
kernel/sched/cpufreq_schedutil.c
206
return cpufreq_driver_resolve_freq(policy, freq);
kernel/sched/cpufreq_schedutil.c
451
if (sg_policy->policy->fast_switch_enabled) {
kernel/sched/cpufreq_schedutil.c
452
cpufreq_driver_fast_switch(sg_policy->policy, next_f);
kernel/sched/cpufreq_schedutil.c
494
struct cpufreq_policy *policy = sg_policy->policy;
kernel/sched/cpufreq_schedutil.c
500
for_each_cpu(j, policy->cpus) {
kernel/sched/cpufreq_schedutil.c
533
if (sg_policy->policy->fast_switch_enabled)
kernel/sched/cpufreq_schedutil.c
534
cpufreq_driver_fast_switch(sg_policy->policy, next_f);
kernel/sched/cpufreq_schedutil.c
564
__cpufreq_driver_target(sg_policy->policy, freq, CPUFREQ_RELATION_L);
kernel/sched/cpufreq_schedutil.c
637
static struct sugov_policy *sugov_policy_alloc(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
645
sg_policy->policy = policy;
kernel/sched/cpufreq_schedutil.c
672
struct cpufreq_policy *policy = sg_policy->policy;
kernel/sched/cpufreq_schedutil.c
676
if (policy->fast_switch_enabled)
kernel/sched/cpufreq_schedutil.c
683
cpumask_first(policy->related_cpus));
kernel/sched/cpufreq_schedutil.c
697
if (policy->dvfs_possible_from_any_cpu)
kernel/sched/cpufreq_schedutil.c
698
set_cpus_allowed_ptr(thread, policy->related_cpus);
kernel/sched/cpufreq_schedutil.c
700
kthread_bind_mask(thread, policy->related_cpus);
kernel/sched/cpufreq_schedutil.c
713
if (sg_policy->policy->fast_switch_enabled)
kernel/sched/cpufreq_schedutil.c
740
static int sugov_init(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
747
if (policy->governor_data)
kernel/sched/cpufreq_schedutil.c
750
cpufreq_enable_fast_switch(policy);
kernel/sched/cpufreq_schedutil.c
752
sg_policy = sugov_policy_alloc(policy);
kernel/sched/cpufreq_schedutil.c
769
policy->governor_data = sg_policy;
kernel/sched/cpufreq_schedutil.c
782
tunables->rate_limit_us = cpufreq_policy_transition_delay_us(policy);
kernel/sched/cpufreq_schedutil.c
784
policy->governor_data = sg_policy;
kernel/sched/cpufreq_schedutil.c
788
get_governor_parent_kobj(policy), "%s",
kernel/sched/cpufreq_schedutil.c
804
policy->governor_data = NULL;
kernel/sched/cpufreq_schedutil.c
815
cpufreq_disable_fast_switch(policy);
kernel/sched/cpufreq_schedutil.c
821
static void sugov_exit(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
823
struct sugov_policy *sg_policy = policy->governor_data;
kernel/sched/cpufreq_schedutil.c
83
if (!cpufreq_this_cpu_can_update(sg_policy->policy))
kernel/sched/cpufreq_schedutil.c
830
policy->governor_data = NULL;
kernel/sched/cpufreq_schedutil.c
838
cpufreq_disable_fast_switch(policy);
kernel/sched/cpufreq_schedutil.c
843
static int sugov_start(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
845
struct sugov_policy *sg_policy = policy->governor_data;
kernel/sched/cpufreq_schedutil.c
858
if (policy_is_shared(policy))
kernel/sched/cpufreq_schedutil.c
860
else if (policy->fast_switch_enabled && cpufreq_driver_has_adjust_perf())
kernel/sched/cpufreq_schedutil.c
865
for_each_cpu(cpu, policy->cpus) {
kernel/sched/cpufreq_schedutil.c
876
static void sugov_stop(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
878
struct sugov_policy *sg_policy = policy->governor_data;
kernel/sched/cpufreq_schedutil.c
881
for_each_cpu(cpu, policy->cpus)
kernel/sched/cpufreq_schedutil.c
886
if (!policy->fast_switch_enabled) {
kernel/sched/cpufreq_schedutil.c
892
static void sugov_limits(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
894
struct sugov_policy *sg_policy = policy->governor_data;
kernel/sched/cpufreq_schedutil.c
896
if (!policy->fast_switch_enabled) {
kernel/sched/cpufreq_schedutil.c
898
cpufreq_policy_apply_limits(policy);
kernel/sched/cpufreq_schedutil.c
932
bool sugov_is_governor(struct cpufreq_policy *policy)
kernel/sched/cpufreq_schedutil.c
934
return policy->governor == &schedutil_gov;
kernel/sched/deadline.c
3538
int sched_dl_overflow(struct task_struct *p, int policy,
kernel/sched/deadline.c
3543
u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0;
kernel/sched/deadline.c
3564
if (dl_policy(policy) && !task_has_dl_policy(p) &&
kernel/sched/deadline.c
3570
} else if (dl_policy(policy) && task_has_dl_policy(p) &&
kernel/sched/deadline.c
3583
} else if (!dl_policy(policy) && task_has_dl_policy(p)) {
kernel/sched/debug.c
1348
P(policy);
kernel/sched/debug.c
1353
} else if (fair_policy(p->policy)) {
kernel/sched/ext.c
264
return __setscheduler_class(p->policy, p->prio);
kernel/sched/ext.c
2985
if (p->policy == SCHED_EXT) {
kernel/sched/ext.c
2986
p->policy = SCHED_NORMAL;
kernel/sched/ext.c
2991
} else if (p->policy == SCHED_EXT) {
kernel/sched/ext.c
3258
int scx_check_setscheduler(struct task_struct *p, int policy)
kernel/sched/ext.c
3264
p->policy != policy && policy == SCHED_EXT)
kernel/sched/ext.c
3846
bool task_should_scx(int policy)
kernel/sched/ext.c
3852
return policy == SCHED_EXT;
kernel/sched/ext.h
20
int scx_check_setscheduler(struct task_struct *p, int policy);
kernel/sched/ext.h
21
bool task_should_scx(int policy);
kernel/sched/ext.h
54
static inline int scx_check_setscheduler(struct task_struct *p, int policy) { return 0; }
kernel/sched/fair.c
8875
if (unlikely(!normal_policy(p->policy)))
kernel/sched/idle.c
416
WARN_ON_ONCE(current->policy != SCHED_FIFO);
kernel/sched/rt.c
1171
return (tsk->policy == SCHED_RR) ? 1 : 0;
kernel/sched/rt.c
2533
if (p->policy != SCHED_RR)
kernel/sched/rt.c
2559
if (task->policy == SCHED_RR)
kernel/sched/sched.h
189
static inline int idle_policy(int policy)
kernel/sched/sched.h
191
return policy == SCHED_IDLE;
kernel/sched/sched.h
194
static inline int normal_policy(int policy)
kernel/sched/sched.h
197
if (policy == SCHED_EXT)
kernel/sched/sched.h
200
return policy == SCHED_NORMAL;
kernel/sched/sched.h
203
static inline int fair_policy(int policy)
kernel/sched/sched.h
205
return normal_policy(policy) || policy == SCHED_BATCH;
kernel/sched/sched.h
208
static inline int rt_policy(int policy)
kernel/sched/sched.h
210
return policy == SCHED_FIFO || policy == SCHED_RR;
kernel/sched/sched.h
213
static inline int dl_policy(int policy)
kernel/sched/sched.h
215
return policy == SCHED_DEADLINE;
kernel/sched/sched.h
218
static inline bool valid_policy(int policy)
kernel/sched/sched.h
220
return idle_policy(policy) || fair_policy(policy) ||
kernel/sched/sched.h
221
rt_policy(policy) || dl_policy(policy);
kernel/sched/sched.h
226
return idle_policy(p->policy);
kernel/sched/sched.h
231
return rt_policy(p->policy);
kernel/sched/sched.h
236
return dl_policy(p->policy);
kernel/sched/sched.h
357
extern int sched_dl_overflow(struct task_struct *p, int policy, const struct sched_attr *attr);
kernel/sched/sched.h
4065
extern const struct sched_class *__setscheduler_class(int policy, int prio);
kernel/sched/syscalls.c
1010
retval = p->policy;
kernel/sched/syscalls.c
1080
kattr.sched_policy = p->policy;
kernel/sched/syscalls.c
1463
SYSCALL_DEFINE1(sched_get_priority_max, int, policy)
kernel/sched/syscalls.c
1467
switch (policy) {
kernel/sched/syscalls.c
1491
SYSCALL_DEFINE1(sched_get_priority_min, int, policy)
kernel/sched/syscalls.c
1495
switch (policy) {
kernel/sched/syscalls.c
19
static inline int __normal_prio(int policy, int rt_prio, int nice)
kernel/sched/syscalls.c
23
if (dl_policy(policy))
kernel/sched/syscalls.c
244
int policy = attr->sched_policy;
kernel/sched/syscalls.c
246
if (policy == SETPARAM_POLICY)
kernel/sched/syscalls.c
247
policy = p->policy;
kernel/sched/syscalls.c
249
p->policy = policy;
kernel/sched/syscalls.c
25
else if (rt_policy(policy))
kernel/sched/syscalls.c
251
if (dl_policy(policy))
kernel/sched/syscalls.c
253
else if (fair_policy(policy))
kernel/sched/syscalls.c
288
static inline void __setscheduler_dl_pi(int newprio, int policy,
kernel/sched/syscalls.c
299
if (dl_prio(newprio) && !dl_policy(policy)) {
kernel/sched/syscalls.c
309
static inline void __setscheduler_dl_pi(int newprio, int policy,
kernel/sched/syscalls.c
42
return __normal_prio(p->policy, p->rt_priority, PRIO_TO_NICE(p->static_prio));
kernel/sched/syscalls.c
437
int policy, int reset_on_fork)
kernel/sched/syscalls.c
439
if (fair_policy(policy)) {
kernel/sched/syscalls.c
445
if (rt_policy(policy)) {
kernel/sched/syscalls.c
449
if (policy != p->policy && !rlim_rtprio)
kernel/sched/syscalls.c
464
if (dl_policy(policy))
kernel/sched/syscalls.c
471
if (task_has_idle_policy(p) && !idle_policy(policy)) {
kernel/sched/syscalls.c
497
int oldpolicy = -1, policy = attr->sched_policy;
kernel/sched/syscalls.c
511
if (policy < 0) {
kernel/sched/syscalls.c
513
policy = oldpolicy = p->policy;
kernel/sched/syscalls.c
517
if (!valid_policy(policy))
kernel/sched/syscalls.c
531
if ((dl_policy(policy) && !__checkparam_dl(attr)) ||
kernel/sched/syscalls.c
532
(rt_policy(policy) != (attr->sched_priority != 0)))
kernel/sched/syscalls.c
536
retval = user_check_sched_setscheduler(p, attr, policy, reset_on_fork);
kernel/sched/syscalls.c
559
if (dl_policy(policy) || dl_policy(p->policy)) {
kernel/sched/syscalls.c
582
retval = scx_check_setscheduler(p, policy);
kernel/sched/syscalls.c
590
if (unlikely(policy == p->policy)) {
kernel/sched/syscalls.c
591
if (fair_policy(policy) &&
kernel/sched/syscalls.c
595
if (rt_policy(policy) && attr->sched_priority != p->rt_priority)
kernel/sched/syscalls.c
597
if (dl_policy(policy) && dl_param_changed(p, attr))
kernel/sched/syscalls.c
615
rt_bandwidth_enabled() && rt_policy(policy) &&
kernel/sched/syscalls.c
622
if (dl_bandwidth_enabled() && dl_policy(policy) &&
kernel/sched/syscalls.c
640
if (unlikely(oldpolicy != -1 && oldpolicy != p->policy)) {
kernel/sched/syscalls.c
641
policy = oldpolicy = -1;
kernel/sched/syscalls.c
653
if ((dl_policy(policy) || dl_task(p)) && sched_dl_overflow(p, policy, attr)) {
kernel/sched/syscalls.c
661
newprio = __normal_prio(policy, attr->sched_priority, attr->sched_nice);
kernel/sched/syscalls.c
676
next_class = __setscheduler_class(policy, newprio);
kernel/sched/syscalls.c
687
__setscheduler_dl_pi(newprio, policy, p, scope);
kernel/sched/syscalls.c
725
static int _sched_setscheduler(struct task_struct *p, int policy,
kernel/sched/syscalls.c
729
.sched_policy = policy,
kernel/sched/syscalls.c
738
if ((policy != SETPARAM_POLICY) && (policy & SCHED_RESET_ON_FORK)) {
kernel/sched/syscalls.c
740
policy &= ~SCHED_RESET_ON_FORK;
kernel/sched/syscalls.c
741
attr.sched_policy = policy;
kernel/sched/syscalls.c
758
int sched_setscheduler(struct task_struct *p, int policy,
kernel/sched/syscalls.c
761
return _sched_setscheduler(p, policy, param, true);
kernel/sched/syscalls.c
788
int sched_setscheduler_nocheck(struct task_struct *p, int policy,
kernel/sched/syscalls.c
791
return _sched_setscheduler(p, policy, param, false);
kernel/sched/syscalls.c
853
do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param)
kernel/sched/syscalls.c
866
return sched_setscheduler(p, policy, &lparam);
kernel/sched/syscalls.c
934
SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy, struct sched_param __user *, param)
kernel/sched/syscalls.c
936
if (policy < 0)
kernel/sched/syscalls.c
939
return do_sched_setscheduler(pid, policy, param);
kernel/taskstats.c
665
.policy = taskstats_cmd_get_policy,
kernel/taskstats.c
673
.policy = cgroupstats_cmd_get_policy,
kernel/trace/trace.c
1581
max_data->policy = tsk->policy;
kernel/trace/trace.c
3445
data->policy, data->rt_priority);
kernel/trace/trace.h
200
unsigned long policy;
kernel/tsacct.c
57
stats->ac_sched = tsk->policy;
lib/nlattr.c
104
NL_SET_ERR_MSG_ATTR_POL(extack, entry, policy,
lib/nlattr.c
110
maxtype, policy, validate, extack,
lib/nlattr.c
394
const struct nla_policy *policy, unsigned int validate,
lib/nlattr.c
397
u16 strict_start_type = policy[0].strict_start_type;
lib/nlattr.c
409
pt = &policy[type];
lib/nlattr.c
605
const struct nla_policy *policy,
lib/nlattr.c
61
const struct nla_policy *policy,
lib/nlattr.c
634
if (policy) {
lib/nlattr.c
635
int err = validate_nla(nla, maxtype, policy,
lib/nlattr.c
674
const struct nla_policy *policy, unsigned int validate,
lib/nlattr.c
677
return __nla_validate_parse(head, len, maxtype, policy, validate,
lib/nlattr.c
729
const struct nla_policy *policy, unsigned int validate,
lib/nlattr.c
732
return __nla_validate_parse(head, len, maxtype, policy, validate,
lib/nlattr.c
90
const struct nla_policy *policy,
mm/debug.c
260
vmg->file, vmg->anon_vma, vmg->policy,
mm/filemap.c
1000
return folio_alloc_mpol_noprof(gfp, order, policy,
mm/filemap.c
1941
pgoff_t index, fgf_t fgp_flags, gfp_t gfp, struct mempolicy *policy)
mm/filemap.c
2011
folio = filemap_alloc_folio(alloc_gfp, order, policy);
mm/filemap.c
994
struct mempolicy *policy)
mm/filemap.c
999
if (policy)
mm/huge_memory.c
1009
policy = token;
mm/huge_memory.c
1011
if (!policy)
mm/huge_memory.c
1041
if (!strcmp(policy, "always")) {
mm/huge_memory.c
1045
} else if (!strcmp(policy, "madvise")) {
mm/huge_memory.c
1049
} else if (!strcmp(policy, "inherit")) {
mm/huge_memory.c
1053
} else if (!strcmp(policy, "never")) {
mm/huge_memory.c
1058
pr_err("invalid policy %s in thp_anon boot parameter\n", policy);
mm/huge_memory.c
993
char *token, *range, *policy, *subtoken;
mm/kfence/kfence_test.c
245
static void *test_alloc(struct kunit *test, size_t size, gfp_t gfp, enum allocation_policy policy)
mm/kfence/kfence_test.c
251
switch (policy) {
mm/kfence/kfence_test.c
300
if (policy == ALLOCATE_ANY)
mm/kfence/kfence_test.c
302
if (policy == ALLOCATE_LEFT && PAGE_ALIGNED(alloc))
mm/kfence/kfence_test.c
304
if (policy == ALLOCATE_RIGHT && !PAGE_ALIGNED(alloc))
mm/kfence/kfence_test.c
306
} else if (policy == ALLOCATE_NONE)
mm/mempolicy.c
1147
static long do_get_mempolicy(int *policy, nodemask_t *nmask,
mm/mempolicy.c
1162
*policy = 0; /* just so it's initialized */
mm/mempolicy.c
1203
*policy = err;
mm/mempolicy.c
1206
*policy = next_node_in(current->il_prev, pol->nodes);
mm/mempolicy.c
1210
*policy = current->il_prev;
mm/mempolicy.c
1212
*policy = next_node_in(current->il_prev,
mm/mempolicy.c
1219
*policy = pol == &default_policy ? MPOL_DEFAULT :
mm/mempolicy.c
1225
*policy |= (pol->flags & MPOL_MODE_FLAGS);
mm/mempolicy.c
1954
static int kernel_get_mempolicy(int __user *policy,
mm/mempolicy.c
1974
if (policy && put_user(pval, policy))
mm/mempolicy.c
1983
SYSCALL_DEFINE5(get_mempolicy, int __user *, policy,
mm/mempolicy.c
1987
return kernel_get_mempolicy(policy, nmask, maxnode, addr, flags);
mm/mempolicy.c
2080
bool apply_policy_zone(struct mempolicy *policy, enum zone_type zone)
mm/mempolicy.c
2094
if (!nodes_intersects(policy->nodes, node_states[N_HIGH_MEMORY]))
mm/mempolicy.c
2100
static unsigned int weighted_interleave_nodes(struct mempolicy *policy)
mm/mempolicy.c
2109
if (!current->il_weight || !node_isset(node, policy->nodes)) {
mm/mempolicy.c
2110
node = next_node_in(node, policy->nodes);
mm/mempolicy.c
2123
static unsigned int interleave_nodes(struct mempolicy *policy)
mm/mempolicy.c
2131
nid = next_node_in(current->il_prev, policy->nodes);
mm/mempolicy.c
2145
struct mempolicy *policy;
mm/mempolicy.c
2151
policy = current->mempolicy;
mm/mempolicy.c
2152
if (!policy)
mm/mempolicy.c
2155
switch (policy->mode) {
mm/mempolicy.c
2157
return first_node(policy->nodes);
mm/mempolicy.c
2160
return interleave_nodes(policy);
mm/mempolicy.c
2163
return weighted_interleave_nodes(policy);
mm/mempolicy.c
2178
&policy->nodes);
mm/mempolicy.c
2956
mpol_get(sn->policy);
mm/mempolicy.c
2957
pol = sn->policy;
mm/mempolicy.c
2966
mpol_put(n->policy);
mm/mempolicy.c
3109
node->policy = pol;
mm/mempolicy.c
3159
*mpol_new = *n->policy;
mm/mempolicy.c
444
struct mempolicy *policy;
mm/mempolicy.c
474
policy = kmem_cache_alloc(policy_cache, GFP_KERNEL);
mm/mempolicy.c
475
if (!policy)
mm/mempolicy.c
477
atomic_set(&policy->refcnt, 1);
mm/mempolicy.c
478
policy->mode = mode;
mm/mempolicy.c
479
policy->flags = flags;
mm/mempolicy.c
480
policy->home_node = NUMA_NO_NODE;
mm/mempolicy.c
482
return policy;
mm/shmem.c
2891
return mpol_set_shared_policy(&SHMEM_I(inode)->policy, vma, mpol);
mm/shmem.c
2908
return mpol_shared_policy_lookup(&SHMEM_I(inode)->policy, index);
mm/shmem.c
2919
mpol = mpol_shared_policy_lookup(&info->policy, index);
mm/shmem.c
3122
mpol_shared_policy_init(&info->policy,
mm/shmem.c
3138
mpol_shared_policy_init(&info->policy, NULL);
mm/shmem.c
5183
mpol_free_shared_policy(&SHMEM_I(inode)->policy);
mm/shmem.c
5650
char *token, *range, *policy, *subtoken;
mm/shmem.c
5667
policy = token;
mm/shmem.c
5669
if (!policy)
mm/shmem.c
5703
if (!strcmp(policy, "always")) {
mm/shmem.c
5708
} else if (!strcmp(policy, "advise")) {
mm/shmem.c
5713
} else if (!strcmp(policy, "inherit")) {
mm/shmem.c
5718
} else if (!strcmp(policy, "within_size")) {
mm/shmem.c
5723
} else if (!strcmp(policy, "never")) {
mm/shmem.c
5729
pr_err("invalid policy %s in thp_shmem boot parameter\n", policy);
mm/vma.c
1733
vmg.policy = new_pol;
mm/vma.c
90
if (!mpol_equal(vmg->policy, vma_policy(vma)))
mm/vma.h
104
struct mempolicy *policy;
mm/vma.h
260
.policy = vma_policy(vma_), \
net/8021q/vlan_netlink.c
293
.policy = vlan_policy,
net/batman-adv/mesh-interface.c
1129
.policy = batadv_ifla_policy,
net/batman-adv/netlink.c
1529
.policy = batadv_netlink_policy,
net/bluetooth/hci_core.c
684
__le16 policy;
net/bluetooth/hci_core.c
747
policy = cpu_to_le16(dr.dev_opt);
net/bluetooth/hci_core.c
750
2, &policy, HCI_CMD_TIMEOUT);
net/bluetooth/hci_event.c
202
conn->link_policy = __le16_to_cpu(rp->policy);
net/bluetooth/hci_event.c
246
hdev->link_policy = __le16_to_cpu(rp->policy);
net/bluetooth/hci_sync.c
4231
cp.policy = cpu_to_le16(link_policy);
net/bridge/br_netlink.c
1904
.policy = br_policy,
net/bridge/netfilter/ebtable_broute.c
26
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_filter.c
22
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_filter.c
26
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_filter.c
30
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_nat.c
22
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_nat.c
26
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtable_nat.c
30
.policy = EBT_ACCEPT,
net/bridge/netfilter/ebtables.c
309
if (chaininfo->policy == EBT_RETURN)
net/bridge/netfilter/ebtables.c
311
if (chaininfo->policy == EBT_ACCEPT) {
net/bridge/netfilter/ebtables.c
554
if (((struct ebt_entries *)e)->policy != EBT_DROP &&
net/bridge/netfilter/ebtables.c
555
((struct ebt_entries *)e)->policy != EBT_ACCEPT) {
net/bridge/netfilter/ebtables.c
558
((struct ebt_entries *)e)->policy != EBT_RETURN)
net/bridge/netfilter/nft_meta_bridge.c
232
.policy = nft_meta_policy,
net/bridge/netfilter/nft_reject_bridge.c
194
.policy = nft_reject_policy,
net/caif/chnl_net.c
502
.policy = ipcaif_policy,
net/core/drop_monitor.c
1661
.policy = net_dm_nl_policy,
net/core/netdev-genl-gen.c
115
.policy = netdev_dev_get_nl_policy,
net/core/netdev-genl-gen.c
128
.policy = netdev_page_pool_get_nl_policy,
net/core/netdev-genl-gen.c
142
.policy = netdev_page_pool_stats_get_nl_policy,
net/core/netdev-genl-gen.c
155
.policy = netdev_queue_get_do_nl_policy,
net/core/netdev-genl-gen.c
162
.policy = netdev_queue_get_dump_nl_policy,
net/core/netdev-genl-gen.c
169
.policy = netdev_napi_get_do_nl_policy,
net/core/netdev-genl-gen.c
176
.policy = netdev_napi_get_dump_nl_policy,
net/core/netdev-genl-gen.c
183
.policy = netdev_qstats_get_nl_policy,
net/core/netdev-genl-gen.c
190
.policy = netdev_bind_rx_nl_policy,
net/core/netdev-genl-gen.c
197
.policy = netdev_napi_set_nl_policy,
net/core/netdev-genl-gen.c
204
.policy = netdev_bind_tx_nl_policy,
net/core/rtnetlink.c
4043
ops->policy, extack);
net/devlink/netlink_gen.c
1004
.policy = devlink_info_get_nl_policy,
net/devlink/netlink_gen.c
1020
.policy = devlink_health_reporter_get_do_nl_policy,
net/devlink/netlink_gen.c
1027
.policy = devlink_health_reporter_get_dump_nl_policy,
net/devlink/netlink_gen.c
1037
.policy = devlink_health_reporter_set_nl_policy,
net/devlink/netlink_gen.c
1047
.policy = devlink_health_reporter_recover_nl_policy,
net/devlink/netlink_gen.c
1057
.policy = devlink_health_reporter_diagnose_nl_policy,
net/devlink/netlink_gen.c
1065
.policy = devlink_health_reporter_dump_get_nl_policy,
net/devlink/netlink_gen.c
1075
.policy = devlink_health_reporter_dump_clear_nl_policy,
net/devlink/netlink_gen.c
1085
.policy = devlink_flash_update_nl_policy,
net/devlink/netlink_gen.c
1095
.policy = devlink_trap_get_do_nl_policy,
net/devlink/netlink_gen.c
1102
.policy = devlink_trap_get_dump_nl_policy,
net/devlink/netlink_gen.c
1112
.policy = devlink_trap_set_nl_policy,
net/devlink/netlink_gen.c
1122
.policy = devlink_trap_group_get_do_nl_policy,
net/devlink/netlink_gen.c
1129
.policy = devlink_trap_group_get_dump_nl_policy,
net/devlink/netlink_gen.c
1139
.policy = devlink_trap_group_set_nl_policy,
net/devlink/netlink_gen.c
1149
.policy = devlink_trap_policer_get_do_nl_policy,
net/devlink/netlink_gen.c
1156
.policy = devlink_trap_policer_get_dump_nl_policy,
net/devlink/netlink_gen.c
1166
.policy = devlink_trap_policer_set_nl_policy,
net/devlink/netlink_gen.c
1176
.policy = devlink_health_reporter_test_nl_policy,
net/devlink/netlink_gen.c
1186
.policy = devlink_rate_get_do_nl_policy,
net/devlink/netlink_gen.c
1193
.policy = devlink_rate_get_dump_nl_policy,
net/devlink/netlink_gen.c
1203
.policy = devlink_rate_set_nl_policy,
net/devlink/netlink_gen.c
1213
.policy = devlink_rate_new_nl_policy,
net/devlink/netlink_gen.c
1223
.policy = devlink_rate_del_nl_policy,
net/devlink/netlink_gen.c
1233
.policy = devlink_linecard_get_do_nl_policy,
net/devlink/netlink_gen.c
1240
.policy = devlink_linecard_get_dump_nl_policy,
net/devlink/netlink_gen.c
1250
.policy = devlink_linecard_set_nl_policy,
net/devlink/netlink_gen.c
1260
.policy = devlink_selftests_get_nl_policy,
net/devlink/netlink_gen.c
1276
.policy = devlink_selftests_run_nl_policy,
net/devlink/netlink_gen.c
1283
.policy = devlink_notify_filter_set_nl_policy,
net/devlink/netlink_gen.c
615
.policy = devlink_get_nl_policy,
net/devlink/netlink_gen.c
631
.policy = devlink_port_get_do_nl_policy,
net/devlink/netlink_gen.c
638
.policy = devlink_port_get_dump_nl_policy,
net/devlink/netlink_gen.c
648
.policy = devlink_port_set_nl_policy,
net/devlink/netlink_gen.c
658
.policy = devlink_port_new_nl_policy,
net/devlink/netlink_gen.c
668
.policy = devlink_port_del_nl_policy,
net/devlink/netlink_gen.c
678
.policy = devlink_port_split_nl_policy,
net/devlink/netlink_gen.c
688
.policy = devlink_port_unsplit_nl_policy,
net/devlink/netlink_gen.c
698
.policy = devlink_sb_get_do_nl_policy,
net/devlink/netlink_gen.c
705
.policy = devlink_sb_get_dump_nl_policy,
net/devlink/netlink_gen.c
715
.policy = devlink_sb_pool_get_do_nl_policy,
net/devlink/netlink_gen.c
722
.policy = devlink_sb_pool_get_dump_nl_policy,
net/devlink/netlink_gen.c
732
.policy = devlink_sb_pool_set_nl_policy,
net/devlink/netlink_gen.c
742
.policy = devlink_sb_port_pool_get_do_nl_policy,
net/devlink/netlink_gen.c
749
.policy = devlink_sb_port_pool_get_dump_nl_policy,
net/devlink/netlink_gen.c
759
.policy = devlink_sb_port_pool_set_nl_policy,
net/devlink/netlink_gen.c
769
.policy = devlink_sb_tc_pool_bind_get_do_nl_policy,
net/devlink/netlink_gen.c
776
.policy = devlink_sb_tc_pool_bind_get_dump_nl_policy,
net/devlink/netlink_gen.c
786
.policy = devlink_sb_tc_pool_bind_set_nl_policy,
net/devlink/netlink_gen.c
796
.policy = devlink_sb_occ_snapshot_nl_policy,
net/devlink/netlink_gen.c
806
.policy = devlink_sb_occ_max_clear_nl_policy,
net/devlink/netlink_gen.c
816
.policy = devlink_eswitch_get_nl_policy,
net/devlink/netlink_gen.c
826
.policy = devlink_eswitch_set_nl_policy,
net/devlink/netlink_gen.c
836
.policy = devlink_dpipe_table_get_nl_policy,
net/devlink/netlink_gen.c
846
.policy = devlink_dpipe_entries_get_nl_policy,
net/devlink/netlink_gen.c
856
.policy = devlink_dpipe_headers_get_nl_policy,
net/devlink/netlink_gen.c
866
.policy = devlink_dpipe_table_counters_set_nl_policy,
net/devlink/netlink_gen.c
876
.policy = devlink_resource_set_nl_policy,
net/devlink/netlink_gen.c
886
.policy = devlink_resource_dump_nl_policy,
net/devlink/netlink_gen.c
896
.policy = devlink_reload_nl_policy,
net/devlink/netlink_gen.c
906
.policy = devlink_param_get_do_nl_policy,
net/devlink/netlink_gen.c
913
.policy = devlink_param_get_dump_nl_policy,
net/devlink/netlink_gen.c
923
.policy = devlink_param_set_nl_policy,
net/devlink/netlink_gen.c
933
.policy = devlink_region_get_do_nl_policy,
net/devlink/netlink_gen.c
940
.policy = devlink_region_get_dump_nl_policy,
net/devlink/netlink_gen.c
950
.policy = devlink_region_new_nl_policy,
net/devlink/netlink_gen.c
960
.policy = devlink_region_del_nl_policy,
net/devlink/netlink_gen.c
968
.policy = devlink_region_read_nl_policy,
net/devlink/netlink_gen.c
978
.policy = devlink_port_param_get_nl_policy,
net/devlink/netlink_gen.c
994
.policy = devlink_port_param_set_nl_policy,
net/dsa/netlink.c
59
.policy = dsa_policy,
net/ethtool/module.c
151
power_new.policy = nla_get_u8(tb[ETHTOOL_A_MODULE_POWER_MODE_POLICY]);
net/ethtool/module.c
156
if (power_new.policy == power.policy)
net/ethtool/module.c
77
if (data->power.policy)
net/ethtool/module.c
92
if (data->power.policy &&
net/ethtool/module.c
94
data->power.policy))
net/ethtool/netlink.c
1140
.policy = ethnl_strset_get_policy,
net/ethtool/netlink.c
1149
.policy = ethnl_linkinfo_get_policy,
net/ethtool/netlink.c
1156
.policy = ethnl_linkinfo_set_policy,
net/ethtool/netlink.c
1165
.policy = ethnl_linkmodes_get_policy,
net/ethtool/netlink.c
1172
.policy = ethnl_linkmodes_set_policy,
net/ethtool/netlink.c
1181
.policy = ethnl_linkstate_get_policy,
net/ethtool/netlink.c
1190
.policy = ethnl_debug_get_policy,
net/ethtool/netlink.c
1197
.policy = ethnl_debug_set_policy,
net/ethtool/netlink.c
1207
.policy = ethnl_wol_get_policy,
net/ethtool/netlink.c
1214
.policy = ethnl_wol_set_policy,
net/ethtool/netlink.c
1223
.policy = ethnl_features_get_policy,
net/ethtool/netlink.c
1230
.policy = ethnl_features_set_policy,
net/ethtool/netlink.c
1239
.policy = ethnl_privflags_get_policy,
net/ethtool/netlink.c
1246
.policy = ethnl_privflags_set_policy,
net/ethtool/netlink.c
1255
.policy = ethnl_rings_get_policy,
net/ethtool/netlink.c
1262
.policy = ethnl_rings_set_policy,
net/ethtool/netlink.c
1271
.policy = ethnl_channels_get_policy,
net/ethtool/netlink.c
1278
.policy = ethnl_channels_set_policy,
net/ethtool/netlink.c
1287
.policy = ethnl_coalesce_get_policy,
net/ethtool/netlink.c
1294
.policy = ethnl_coalesce_set_policy,
net/ethtool/netlink.c
1303
.policy = ethnl_pause_get_policy,
net/ethtool/netlink.c
1310
.policy = ethnl_pause_set_policy,
net/ethtool/netlink.c
1319
.policy = ethnl_eee_get_policy,
net/ethtool/netlink.c
1326
.policy = ethnl_eee_set_policy,
net/ethtool/netlink.c
1335
.policy = ethnl_tsinfo_get_policy,
net/ethtool/netlink.c
1342
.policy = ethnl_cable_test_act_policy,
net/ethtool/netlink.c
1349
.policy = ethnl_cable_test_tdr_act_policy,
net/ethtool/netlink.c
1357
.policy = ethnl_tunnel_info_get_policy,
net/ethtool/netlink.c
1366
.policy = ethnl_fec_get_policy,
net/ethtool/netlink.c
1373
.policy = ethnl_fec_set_policy,
net/ethtool/netlink.c
1383
.policy = ethnl_module_eeprom_get_policy,
net/ethtool/netlink.c
1392
.policy = ethnl_stats_get_policy,
net/ethtool/netlink.c
1401
.policy = ethnl_phc_vclocks_get_policy,
net/ethtool/netlink.c
1410
.policy = ethnl_module_get_policy,
net/ethtool/netlink.c
1417
.policy = ethnl_module_set_policy,
net/ethtool/netlink.c
1426
.policy = ethnl_pse_get_policy,
net/ethtool/netlink.c
1433
.policy = ethnl_pse_set_policy,
net/ethtool/netlink.c
1441
.policy = ethnl_rss_get_policy,
net/ethtool/netlink.c
1450
.policy = ethnl_plca_get_cfg_policy,
net/ethtool/netlink.c
1457
.policy = ethnl_plca_set_cfg_policy,
net/ethtool/netlink.c
1466
.policy = ethnl_plca_get_status_policy,
net/ethtool/netlink.c
1475
.policy = ethnl_mm_get_policy,
net/ethtool/netlink.c
1482
.policy = ethnl_mm_set_policy,
net/ethtool/netlink.c
1489
.policy = ethnl_module_fw_flash_act_policy,
net/ethtool/netlink.c
1498
.policy = ethnl_phy_get_policy,
net/ethtool/netlink.c
1507
.policy = ethnl_tsconfig_get_policy,
net/ethtool/netlink.c
1514
.policy = ethnl_tsconfig_set_policy,
net/ethtool/netlink.c
1521
.policy = ethnl_rss_set_policy,
net/ethtool/netlink.c
1528
.policy = ethnl_rss_create_policy,
net/ethtool/netlink.c
1535
.policy = ethnl_rss_delete_policy,
net/ethtool/netlink.c
1544
.policy = ethnl_mse_get_policy,
net/handshake/genl.c
31
.policy = handshake_accept_nl_policy,
net/handshake/genl.c
38
.policy = handshake_done_nl_policy,
net/hsr/hsr_netlink.c
195
.policy = hsr_policy,
net/hsr/hsr_netlink.c
556
.policy = hsr_genl_policy,
net/ieee802154/netlink.c
131
.policy = ieee802154_policy,
net/ieee802154/nl802154.c
3087
.policy = nl802154_policy,
net/ipv4/fou_core.c
911
.policy = fou_nl_policy,
net/ipv4/ip_gre.c
1670
.policy = ipgre_policy,
net/ipv4/ip_gre.c
1685
.policy = ipgre_policy,
net/ipv4/ip_gre.c
1700
.policy = ipgre_policy,
net/ipv4/ip_vti.c
649
.policy = vti_policy,
net/ipv4/ipip.c
601
.policy = ipip_policy,
net/ipv4/netfilter/nft_dup_ipv4.c
91
.policy = nft_dup_ipv4_policy,
net/ipv4/netfilter/nft_fib_ipv4.c
205
.policy = nft_fib_policy,
net/ipv4/netfilter/nft_reject_ipv4.c
55
.policy = nft_reject_policy,
net/ipv4/tcp_metrics.c
992
.policy = tcp_metrics_nl_policy,
net/ipv6/ila/ila_main.c
51
.policy = ila_nl_policy,
net/ipv6/ioam6.c
558
.policy = ioam6_genl_policy_addns,
net/ipv6/ioam6.c
566
.policy = ioam6_genl_policy_delns,
net/ipv6/ioam6.c
582
.policy = ioam6_genl_policy_addsc,
net/ipv6/ioam6.c
590
.policy = ioam6_genl_policy_delsc,
net/ipv6/ioam6.c
606
.policy = ioam6_genl_policy_ns_sc,
net/ipv6/ip6_gre.c
2284
.policy = ip6gre_policy,
net/ipv6/ip6_gre.c
2299
.policy = ip6gre_policy,
net/ipv6/ip6_gre.c
2313
.policy = ip6gre_policy,
net/ipv6/ip6_tunnel.c
2245
.policy = ip6_tnl_policy,
net/ipv6/ip6_vti.c
1103
.policy = vti6_policy,
net/ipv6/netfilter/nft_dup_ipv6.c
89
.policy = nft_dup_ipv6_policy,
net/ipv6/netfilter/nft_fib_ipv6.c
267
.policy = nft_fib_policy,
net/ipv6/netfilter/nft_reject_ipv6.c
56
.policy = nft_reject_policy,
net/ipv6/seg6.c
496
.policy = seg6_genl_policy,
net/ipv6/sit.c
1767
.policy = ipip6_policy,
net/key/af_key.c
2727
return xfrm_policy_walk(net, &pfk->dump.u.policy, dump_sp, (void *) pfk);
net/key/af_key.c
2734
xfrm_policy_walk_done(&pfk->dump.u.policy, net);
net/key/af_key.c
2751
xfrm_policy_walk_init(&pfk->dump.u.policy, XFRM_POLICY_TYPE_MAIN);
net/key/af_key.c
57
struct xfrm_policy_walk policy;
net/l2tp/l2tp_netlink.c
999
.policy = l2tp_nl_policy,
net/mac80211/agg-rx.c
234
u8 dialog_token, u16 status, u16 policy,
net/mac80211/agg-rx.c
260
capab |= u16_encode_bits(policy, IEEE80211_ADDBA_PARAM_POLICY_MASK);
net/mptcp/mptcp_pm_gen.c
105
.policy = mptcp_pm_del_addr_nl_policy,
net/mptcp/mptcp_pm_gen.c
114
.policy = mptcp_pm_get_addr_nl_policy,
net/mptcp/mptcp_pm_gen.c
121
.policy = mptcp_pm_flush_addrs_nl_policy,
net/mptcp/mptcp_pm_gen.c
129
.policy = mptcp_pm_set_limits_nl_policy,
net/mptcp/mptcp_pm_gen.c
137
.policy = mptcp_pm_get_limits_nl_policy,
net/mptcp/mptcp_pm_gen.c
144
.policy = mptcp_pm_set_flags_nl_policy,
net/mptcp/mptcp_pm_gen.c
152
.policy = mptcp_pm_announce_nl_policy,
net/mptcp/mptcp_pm_gen.c
160
.policy = mptcp_pm_remove_nl_policy,
net/mptcp/mptcp_pm_gen.c
168
.policy = mptcp_pm_subflow_create_nl_policy,
net/mptcp/mptcp_pm_gen.c
176
.policy = mptcp_pm_subflow_destroy_nl_policy,
net/mptcp/mptcp_pm_gen.c
97
.policy = mptcp_pm_add_addr_nl_policy,
net/ncsi/ncsi-netlink.c
767
.policy = ncsi_genl_policy,
net/netfilter/ipset/ip_set_core.c
2144
.policy = ip_set_create_policy,
net/netfilter/ipset/ip_set_core.c
2150
.policy = ip_set_setname_policy,
net/netfilter/ipset/ip_set_core.c
2156
.policy = ip_set_setname_policy,
net/netfilter/ipset/ip_set_core.c
2162
.policy = ip_set_setname2_policy,
net/netfilter/ipset/ip_set_core.c
2168
.policy = ip_set_setname2_policy,
net/netfilter/ipset/ip_set_core.c
2174
.policy = ip_set_dump_policy,
net/netfilter/ipset/ip_set_core.c
2180
.policy = ip_set_setname_policy,
net/netfilter/ipset/ip_set_core.c
2186
.policy = ip_set_adt_policy,
net/netfilter/ipset/ip_set_core.c
2192
.policy = ip_set_adt_policy,
net/netfilter/ipset/ip_set_core.c
2198
.policy = ip_set_adt_policy,
net/netfilter/ipset/ip_set_core.c
2204
.policy = ip_set_setname_policy,
net/netfilter/ipset/ip_set_core.c
2210
.policy = ip_set_type_policy,
net/netfilter/ipset/ip_set_core.c
2216
.policy = ip_set_protocol_policy,
net/netfilter/ipset/ip_set_core.c
2222
.policy = ip_set_setname_policy,
net/netfilter/ipset/ip_set_core.c
2228
.policy = ip_set_index_policy,
net/netfilter/ipvs/ip_vs_ctl.c
4240
.policy = ip_vs_cmd_policy,
net/netfilter/nf_conntrack_netlink.c
3768
.policy = ct_nla_policy
net/netfilter/nf_conntrack_netlink.c
3774
.policy = ct_nla_policy
net/netfilter/nf_conntrack_netlink.c
3780
.policy = ct_nla_policy
net/netfilter/nf_conntrack_netlink.c
3786
.policy = ct_nla_policy
net/netfilter/nf_conntrack_netlink.c
3811
.policy = exp_nla_policy
net/netfilter/nf_conntrack_netlink.c
3817
.policy = exp_nla_policy
net/netfilter/nf_conntrack_netlink.c
3823
.policy = exp_nla_policy
net/netfilter/nf_tables_api.c
10038
if (trans->policy != NF_DROP)
net/netfilter/nf_tables_api.c
10045
basechain->policy = NF_DROP;
net/netfilter/nf_tables_api.c
10071
switch (trans->policy) {
net/netfilter/nf_tables_api.c
10074
basechain->policy = trans->policy;
net/netfilter/nf_tables_api.c
2081
htonl(basechain->policy)))
net/netfilter/nf_tables_api.c
2667
basechain->policy = NF_ACCEPT;
net/netfilter/nf_tables_api.c
2695
static int nf_tables_addchain(struct nft_ctx *ctx, u8 family, u8 policy,
net/netfilter/nf_tables_api.c
2815
nft_trans_chain_policy(trans) = policy;
net/netfilter/nf_tables_api.c
2841
static int nf_tables_updchain(struct nft_ctx *ctx, u8 genmask, u8 policy,
net/netfilter/nf_tables_api.c
2978
nft_trans_chain_policy(trans) = policy;
net/netfilter/nf_tables_api.c
3067
u8 policy = NF_ACCEPT;
net/netfilter/nf_tables_api.c
3118
policy = ntohl(nla_get_be32(nla[NFTA_CHAIN_POLICY]));
net/netfilter/nf_tables_api.c
3119
switch (policy) {
net/netfilter/nf_tables_api.c
3150
return nf_tables_updchain(&ctx, genmask, policy, flags, attr,
net/netfilter/nf_tables_api.c
3154
return nf_tables_addchain(&ctx, family, policy, flags, extack);
net/netfilter/nf_tables_api.c
3465
type->policy, NULL);
net/netfilter/nf_tables_api.c
3528
type->policy, NULL);
net/netfilter/nf_tables_api.c
4615
switch (desc->policy) {
net/netfilter/nf_tables_api.c
4945
if (set->policy != NFT_SET_POL_PERFORMANCE) {
net/netfilter/nf_tables_api.c
4946
if (nla_put_be32(skb, NFTA_SET_POLICY, htonl(set->policy)))
net/netfilter/nf_tables_api.c
5459
desc.policy = NFT_SET_POL_PERFORMANCE;
net/netfilter/nf_tables_api.c
5461
desc.policy = ntohl(nla_get_be32(nla[NFTA_SET_POLICY]));
net/netfilter/nf_tables_api.c
5462
switch (desc.policy) {
net/netfilter/nf_tables_api.c
5599
set->policy = desc.policy;
net/netfilter/nf_tables_api.c
8079
type->policy, NULL);
net/netfilter/nf_tables_api.c
9815
.policy = nft_table_policy,
net/netfilter/nf_tables_api.c
9821
.policy = nft_table_policy,
net/netfilter/nf_tables_api.c
9827
.policy = nft_table_policy,
net/netfilter/nf_tables_api.c
9833
.policy = nft_table_policy,
net/netfilter/nf_tables_api.c
9839
.policy = nft_chain_policy,
net/netfilter/nf_tables_api.c
9845
.policy = nft_chain_policy,
net/netfilter/nf_tables_api.c
9851
.policy = nft_chain_policy,
net/netfilter/nf_tables_api.c
9857
.policy = nft_chain_policy,
net/netfilter/nf_tables_api.c
9863
.policy = nft_rule_policy,
net/netfilter/nf_tables_api.c
9869
.policy = nft_rule_policy,
net/netfilter/nf_tables_api.c
9875
.policy = nft_rule_policy,
net/netfilter/nf_tables_api.c
9881
.policy = nft_rule_policy,
net/netfilter/nf_tables_api.c
9887
.policy = nft_rule_policy,
net/netfilter/nf_tables_api.c
9893
.policy = nft_set_policy,
net/netfilter/nf_tables_api.c
9899
.policy = nft_set_policy,
net/netfilter/nf_tables_api.c
9905
.policy = nft_set_policy,
net/netfilter/nf_tables_api.c
9911
.policy = nft_set_policy,
net/netfilter/nf_tables_api.c
9917
.policy = nft_set_elem_list_policy,
net/netfilter/nf_tables_api.c
9923
.policy = nft_set_elem_list_policy,
net/netfilter/nf_tables_api.c
9929
.policy = nft_set_elem_list_policy,
net/netfilter/nf_tables_api.c
9935
.policy = nft_set_elem_list_policy,
net/netfilter/nf_tables_api.c
9941
.policy = nft_set_elem_list_policy,
net/netfilter/nf_tables_api.c
9951
.policy = nft_obj_policy,
net/netfilter/nf_tables_api.c
9957
.policy = nft_obj_policy,
net/netfilter/nf_tables_api.c
9963
.policy = nft_obj_policy,
net/netfilter/nf_tables_api.c
9969
.policy = nft_obj_policy,
net/netfilter/nf_tables_api.c
9975
.policy = nft_obj_policy,
net/netfilter/nf_tables_api.c
9981
.policy = nft_flowtable_policy,
net/netfilter/nf_tables_api.c
9987
.policy = nft_flowtable_policy,
net/netfilter/nf_tables_api.c
9993
.policy = nft_flowtable_policy,
net/netfilter/nf_tables_api.c
9999
.policy = nft_flowtable_policy,
net/netfilter/nf_tables_core.c
343
if (nft_base_chain(basechain)->policy == NF_DROP)
net/netfilter/nf_tables_core.c
346
return nft_base_chain(basechain)->policy;
net/netfilter/nf_tables_offload.c
501
u8 policy;
net/netfilter/nf_tables_offload.c
507
policy = ppolicy ? *ppolicy : basechain->policy;
net/netfilter/nf_tables_offload.c
510
if (cmd == FLOW_BLOCK_BIND && policy == NF_DROP)
net/netfilter/nf_tables_offload.c
571
u8 policy;
net/netfilter/nf_tables_offload.c
583
policy = nft_trans_chain_policy(trans);
net/netfilter/nf_tables_offload.c
584
err = nft_flow_offload_chain(nft_trans_chain(trans), &policy,
net/netfilter/nf_tables_offload.c
591
policy = nft_trans_chain_policy(trans);
net/netfilter/nf_tables_offload.c
592
err = nft_flow_offload_chain(nft_trans_chain(trans), &policy,
net/netfilter/nf_tables_trace.c
331
htonl(info->basechain->policy)))
net/netfilter/nfnetlink.c
277
ss->cb[cb_id].policy, extack);
net/netfilter/nfnetlink.c
522
ss->cb[cb_id].policy, &extack);
net/netfilter/nfnetlink_acct.c
386
.policy = nfnl_acct_policy
net/netfilter/nfnetlink_acct.c
392
.policy = nfnl_acct_policy
net/netfilter/nfnetlink_acct.c
398
.policy = nfnl_acct_policy
net/netfilter/nfnetlink_acct.c
404
.policy = nfnl_acct_policy
net/netfilter/nfnetlink_cthelper.c
287
nfnl_cthelper_update_policy_one(const struct nf_conntrack_expect_policy *policy,
net/netfilter/nfnetlink_cthelper.c
304
if (nla_strcmp(tb[NFCTH_POLICY_NAME], policy->name))
net/netfilter/nfnetlink_cthelper.c
322
struct nf_conntrack_expect_policy *policy;
net/netfilter/nfnetlink_cthelper.c
347
policy = (struct nf_conntrack_expect_policy *)
net/netfilter/nfnetlink_cthelper.c
349
policy->max_expected = new_policy->max_expected;
net/netfilter/nfnetlink_cthelper.c
350
policy->timeout = new_policy->timeout;
net/netfilter/nfnetlink_cthelper.c
746
.policy = nfnl_cthelper_policy
net/netfilter/nfnetlink_cthelper.c
752
.policy = nfnl_cthelper_policy
net/netfilter/nfnetlink_cthelper.c
758
.policy = nfnl_cthelper_policy
net/netfilter/nfnetlink_cttimeout.c
540
.policy = cttimeout_nla_policy
net/netfilter/nfnetlink_cttimeout.c
546
.policy = cttimeout_nla_policy
net/netfilter/nfnetlink_cttimeout.c
552
.policy = cttimeout_nla_policy
net/netfilter/nfnetlink_cttimeout.c
558
.policy = cttimeout_nla_policy
net/netfilter/nfnetlink_cttimeout.c
564
.policy = cttimeout_nla_policy
net/netfilter/nfnetlink_hook.c
458
.policy = nfnl_hook_nla_policy
net/netfilter/nfnetlink_log.c
1008
.policy = nfula_cfg_policy
net/netfilter/nfnetlink_osf.c
404
.policy = nfnl_osf_policy,
net/netfilter/nfnetlink_osf.c
410
.policy = nfnl_osf_policy,
net/netfilter/nfnetlink_queue.c
1726
.policy = nfqa_verdict_policy
net/netfilter/nfnetlink_queue.c
1732
.policy = nfqa_cfg_policy
net/netfilter/nfnetlink_queue.c
1738
.policy = nfqa_verdict_batch_policy
net/netfilter/nft_bitwise.c
625
.policy = nft_bitwise_policy,
net/netfilter/nft_byteorder.c
195
.policy = nft_byteorder_policy,
net/netfilter/nft_cmp.c
433
.policy = nft_cmp_policy,
net/netfilter/nft_compat.c
768
.policy = nfnl_compat_policy_get
net/netfilter/nft_compat.c
863
.policy = nft_match_policy,
net/netfilter/nft_compat.c
946
.policy = nft_target_policy,
net/netfilter/nft_connlimit.c
180
.policy = nft_connlimit_policy,
net/netfilter/nft_connlimit.c
267
.policy = nft_connlimit_policy,
net/netfilter/nft_counter.c
213
.policy = nft_counter_policy,
net/netfilter/nft_counter.c
324
.policy = nft_counter_policy,
net/netfilter/nft_ct.c
1070
.policy = nft_ct_timeout_policy,
net/netfilter/nft_ct.c
1257
.policy = nft_ct_helper_policy,
net/netfilter/nft_ct.c
1409
.policy = nft_ct_expect_policy,
net/netfilter/nft_ct.c
855
.policy = nft_ct_policy,
net/netfilter/nft_dup_netdev.c
92
.policy = nft_dup_netdev_policy,
net/netfilter/nft_dynset.c
438
.policy = nft_dynset_policy,
net/netfilter/nft_exthdr.c
845
.policy = nft_exthdr_policy,
net/netfilter/nft_fib_inet.c
59
.policy = nft_fib_policy,
net/netfilter/nft_fib_netdev.c
68
.policy = nft_fib_policy,
net/netfilter/nft_flow_offload.c
234
.policy = nft_flow_offload_policy,
net/netfilter/nft_fwd_netdev.c
252
.policy = nft_fwd_netdev_policy,
net/netfilter/nft_hash.c
266
.policy = nft_hash_policy,
net/netfilter/nft_immediate.c
352
.policy = nft_immediate_policy,
net/netfilter/nft_inner.c
428
.policy = nft_inner_policy,
net/netfilter/nft_last.c
134
.policy = nft_last_policy,
net/netfilter/nft_limit.c
324
.policy = nft_limit_policy,
net/netfilter/nft_limit.c
449
.policy = nft_limit_policy,
net/netfilter/nft_log.c
244
.policy = nft_log_policy,
net/netfilter/nft_lookup.c
296
.policy = nft_lookup_policy,
net/netfilter/nft_masq.c
153
.policy = nft_masq_policy,
net/netfilter/nft_masq.c
181
.policy = nft_masq_policy,
net/netfilter/nft_masq.c
223
.policy = nft_masq_policy,
net/netfilter/nft_meta.c
1009
.policy = nft_secmark_policy,
net/netfilter/nft_meta.c
901
.policy = nft_meta_policy,
net/netfilter/nft_nat.c
329
.policy = nft_nat_policy,
net/netfilter/nft_nat.c
361
.policy = nft_nat_policy,
net/netfilter/nft_numgen.c
236
.policy = nft_ng_policy,
net/netfilter/nft_objref.c
279
.policy = nft_objref_policy,
net/netfilter/nft_osf.c
169
.policy = nft_osf_policy,
net/netfilter/nft_payload.c
1095
.policy = nft_payload_policy,
net/netfilter/nft_queue.c
228
.policy = nft_queue_policy,
net/netfilter/nft_quota.c
205
.policy = nft_quota_policy,
net/netfilter/nft_quota.c
275
.policy = nft_quota_policy,
net/netfilter/nft_range.c
147
.policy = nft_range_policy,
net/netfilter/nft_redir.c
156
.policy = nft_redir_policy,
net/netfilter/nft_redir.c
184
.policy = nft_redir_policy,
net/netfilter/nft_redir.c
213
.policy = nft_redir_policy,
net/netfilter/nft_reject_inet.c
89
.policy = nft_reject_policy,
net/netfilter/nft_reject_netdev.c
168
.policy = nft_reject_policy,
net/netfilter/nft_rt.c
204
.policy = nft_rt_policy,
net/netfilter/nft_socket.c
305
.policy = nft_socket_policy,
net/netfilter/nft_synproxy.c
302
.policy = nft_synproxy_policy,
net/netfilter/nft_synproxy.c
362
.policy = nft_synproxy_policy,
net/netfilter/nft_tproxy.c
341
.policy = nft_tproxy_policy,
net/netfilter/nft_tunnel.c
166
.policy = nft_tunnel_policy,
net/netfilter/nft_tunnel.c
723
.policy = nft_tunnel_key_policy,
net/netfilter/nft_xfrm.c
302
.policy = nft_xfrm_policy,
net/netlabel/netlabel_calipso.c
368
.policy = calipso_genl_policy,
net/netlabel/netlabel_cipso_v4.c
766
.policy = netlbl_cipsov4_genl_policy,
net/netlabel/netlabel_mgmt.c
820
.policy = netlbl_mgmt_genl_policy,
net/netlabel/netlabel_unlabeled.c
1365
.policy = netlbl_unlabel_genl_policy,
net/netlink/af_netlink.c
2177
if (extack->policy)
net/netlink/af_netlink.c
2178
tlvlen += netlink_policy_dump_attr_size_estimate(extack->policy);
net/netlink/af_netlink.c
2209
if (extack->policy)
net/netlink/af_netlink.c
2210
netlink_policy_dump_write_attr(skb, extack->policy,
net/netlink/genetlink.c
111
if (op->policy || op->cmd < family->resv_start_op)
net/netlink/genetlink.c
114
op->policy = genl_policy_reject_all;
net/netlink/genetlink.c
122
if (op->policy)
net/netlink/genetlink.c
125
op->policy = genl_policy_reject_all;
net/netlink/genetlink.c
1257
if (i.doit.policy || i.dumpit.policy)
net/netlink/genetlink.c
1571
if (doit.policy) {
net/netlink/genetlink.c
1573
doit.policy,
net/netlink/genetlink.c
1578
if (dump.policy) {
net/netlink/genetlink.c
1580
dump.policy,
net/netlink/genetlink.c
1601
if (i.doit.policy) {
net/netlink/genetlink.c
1603
i.doit.policy,
net/netlink/genetlink.c
1608
if (i.dumpit.policy) {
net/netlink/genetlink.c
1610
i.dumpit.policy,
net/netlink/genetlink.c
163
if (!op->policy)
net/netlink/genetlink.c
164
op->policy = family->policy;
net/netlink/genetlink.c
1659
if (!doit->policy && !dumpit->policy)
net/netlink/genetlink.c
1674
if (doit->policy) {
net/netlink/genetlink.c
1676
doit->policy,
net/netlink/genetlink.c
1682
if (dumpit->policy) {
net/netlink/genetlink.c
1684
dumpit->policy,
net/netlink/genetlink.c
1770
.policy = ctrl_policy_family,
net/netlink/genetlink.c
1778
.policy = ctrl_policy_family,
net/netlink/genetlink.c
1785
.policy = ctrl_policy_policy,
net/netlink/genetlink.c
195
op->policy = family->policy;
net/netlink/genetlink.c
282
op->policy = NULL;
net/netlink/genetlink.c
285
op->policy = full->policy;
net/netlink/genetlink.c
944
err = __nlmsg_parse(nlh, hdrlen, attrbuf, ops->maxattr, ops->policy,
net/netlink/policy.c
127
const struct nla_policy *policy,
net/netlink/policy.c
145
err = add_policy(&state, policy, maxtype);
net/netlink/policy.c
150
policy_idx < state->n_alloc && state->policies[policy_idx].policy;
net/netlink/policy.c
152
const struct nla_policy *policy;
net/netlink/policy.c
155
policy = state->policies[policy_idx].policy;
net/netlink/policy.c
160
switch (policy[type].type) {
net/netlink/policy.c
164
policy[type].nested_policy,
net/netlink/policy.c
165
policy[type].len);
net/netlink/policy.c
193
!state->policies[state->policy_idx].policy;
net/netlink/policy.c
22
const struct nla_policy *policy;
net/netlink/policy.c
28
const struct nla_policy *policy,
net/netlink/policy.c
34
if (!policy || !maxtype)
net/netlink/policy.c
38
if (state->policies[i].policy == policy &&
net/netlink/policy.c
42
if (!state->policies[i].policy) {
net/netlink/policy.c
43
state->policies[i].policy = policy;
net/netlink/policy.c
438
struct nlattr *policy;
net/netlink/policy.c
445
pt = &state->policies[state->policy_idx].policy[state->attr_idx];
net/netlink/policy.c
447
policy = nla_nest_start(skb, state->policy_idx);
net/netlink/policy.c
448
if (!policy)
net/netlink/policy.c
453
nla_nest_cancel(skb, policy);
net/netlink/policy.c
461
nla_nest_end(skb, policy);
net/netlink/policy.c
479
nla_nest_cancel(skb, policy);
net/netlink/policy.c
60
state->policies[old_n_alloc].policy = policy;
net/netlink/policy.c
83
const struct nla_policy *policy,
net/netlink/policy.c
88
if (WARN_ON(!policy || !maxtype))
net/netlink/policy.c
92
if (state->policies[i].policy == policy &&
net/nfc/netlink.c
1819
.policy = nfc_genl_policy,
net/openvswitch/conntrack.c
1987
.policy = ct_limit_policy,
net/openvswitch/datapath.c
1568
.policy = flow_policy,
net/openvswitch/datapath.c
2114
.policy = datapath_policy,
net/openvswitch/datapath.c
2611
.policy = vport_policy,
net/openvswitch/datapath.c
743
.policy = packet_policy,
net/openvswitch/flow_netlink.c
3113
static const struct nla_policy policy[OVS_PSAMPLE_ATTR_MAX + 1] = {
net/openvswitch/flow_netlink.c
3126
err = nla_parse_nested(a, OVS_PSAMPLE_ATTR_MAX, attr, policy, NULL);
net/openvswitch/meter.c
718
.policy = meter_policy,
net/psp/psp-nl-gen.c
104
.policy = psp_tx_assoc_nl_policy,
net/psp/psp-nl-gen.c
113
.policy = psp_get_stats_nl_policy,
net/psp/psp-nl-gen.c
63
.policy = psp_dev_get_nl_policy,
net/psp/psp-nl-gen.c
77
.policy = psp_dev_set_nl_policy,
net/psp/psp-nl-gen.c
86
.policy = psp_key_rotate_nl_policy,
net/psp/psp-nl-gen.c
95
.policy = psp_rx_assoc_nl_policy,
net/sched/sch_netem.c
960
const struct nla_policy *policy, int len)
net/sched/sch_netem.c
972
nested_len, policy, NULL);
net/sctp/socket.c
7460
int policy;
net/sctp/socket.c
7472
policy = params.sprstat_policy;
net/sctp/socket.c
7473
if (!policy || (policy & ~(SCTP_PR_SCTP_MASK | SCTP_PR_SCTP_ALL)) ||
net/sctp/socket.c
7474
((policy & SCTP_PR_SCTP_ALL) && (policy & SCTP_PR_SCTP_MASK)))
net/sctp/socket.c
7481
if (policy == SCTP_PR_SCTP_ALL) {
net/sctp/socket.c
7484
for (policy = 0; policy <= SCTP_PR_INDEX(MAX); policy++) {
net/sctp/socket.c
7486
asoc->abandoned_unsent[policy];
net/sctp/socket.c
7488
asoc->abandoned_sent[policy];
net/sctp/socket.c
7492
asoc->abandoned_unsent[__SCTP_PR_INDEX(policy)];
net/sctp/socket.c
7494
asoc->abandoned_sent[__SCTP_PR_INDEX(policy)];
net/sctp/socket.c
7521
int policy;
net/sctp/socket.c
7532
policy = params.sprstat_policy;
net/sctp/socket.c
7533
if (!policy || (policy & ~(SCTP_PR_SCTP_MASK | SCTP_PR_SCTP_ALL)) ||
net/sctp/socket.c
7534
((policy & SCTP_PR_SCTP_ALL) && (policy & SCTP_PR_SCTP_MASK)))
net/sctp/socket.c
7550
if (policy == SCTP_PR_SCTP_ALL) {
net/sctp/socket.c
7553
for (policy = 0; policy <= SCTP_PR_INDEX(MAX); policy++) {
net/sctp/socket.c
7555
streamoute->abandoned_unsent[policy];
net/sctp/socket.c
7557
streamoute->abandoned_sent[policy];
net/sctp/socket.c
7561
streamoute->abandoned_unsent[__SCTP_PR_INDEX(policy)];
net/sctp/socket.c
7563
streamoute->abandoned_sent[__SCTP_PR_INDEX(policy)];
net/shaper/shaper_nl_gen.c
105
.policy = net_shaper_set_nl_policy,
net/shaper/shaper_nl_gen.c
114
.policy = net_shaper_delete_nl_policy,
net/shaper/shaper_nl_gen.c
123
.policy = net_shaper_group_nl_policy,
net/shaper/shaper_nl_gen.c
132
.policy = net_shaper_cap_get_do_nl_policy,
net/shaper/shaper_nl_gen.c
141
.policy = net_shaper_cap_get_dump_nl_policy,
net/shaper/shaper_nl_gen.c
87
.policy = net_shaper_get_do_nl_policy,
net/shaper/shaper_nl_gen.c
96
.policy = net_shaper_get_dump_nl_policy,
net/smc/smc_netlink.c
141
.policy = smc_gen_nl_policy,
net/smc/smc_netlink.c
86
.policy = smc_gen_ueid_policy,
net/smc/smc_netlink.c
92
.policy = smc_gen_ueid_policy,
net/smc/smc_pnet.c
718
.policy = smc_pnet_policy,
net/sunrpc/sysfs.c
205
ret = sprintf(buf, "%s\n", xprtsec_strings[xprt->xprtsec.policy]);
net/sunrpc/xprtsock.c
2630
switch (xprtsec->policy) {
net/sunrpc/xprtsock.c
2705
.policy = RPC_XPRTSEC_NONE,
net/sunrpc/xprtsock.c
3495
switch (args->xprtsec.policy) {
net/tipc/netlink.c
292
.policy = tipc_nl_policy,
net/tipc/netlink_compat.c
216
tipc_genl_family.policy, NULL);
net/tipc/netlink_compat.c
229
tipc_genl_family.policy,
net/tls/tls_sw.c
853
bool enospc, policy, redir_ingress;
net/tls/tls_sw.c
857
policy = !(flags & MSG_SENDPAGE_NOPOLICY);
net/tls/tls_sw.c
859
if (!psock || !policy) {
net/wireless/core.c
978
if (WARN_ON(!rdev->wiphy.vendor_commands[i].policy))
net/wireless/nl80211.c
16467
if (vcmd->policy == VENDOR_CMD_RAW_DATA) {
net/wireless/nl80211.c
16482
return nla_validate_nested(attr, vcmd->maxattr, vcmd->policy, extack);
net/wireless/nl80211.c
19170
.policy = nl80211_policy,
net/xfrm/xfrm_compat.c
49
struct compat_xfrm_userpolicy_info policy;
net/xfrm/xfrm_compat.c
597
int maxtype, const struct nla_policy *policy,
net/xfrm/xfrm_compat.c
619
maxtype ? : XFRMA_MAX, policy ? : compat_policy, extack);
net/xfrm/xfrm_interface_core.c
938
.policy = xfrmi_policy,
net/xfrm/xfrm_policy.c
1180
struct xfrm_policy *policy, u8 dir)
net/xfrm/xfrm_policy.c
1185
net = xp_net(policy);
net/xfrm/xfrm_policy.c
1188
if (xfrm_policy_inexact_insert_use_any_list(policy))
net/xfrm/xfrm_policy.c
1191
if (xfrm_pol_inexact_addr_use_any_list(&policy->selector.daddr,
net/xfrm/xfrm_policy.c
1192
policy->family,
net/xfrm/xfrm_policy.c
1193
policy->selector.prefixlen_d)) {
net/xfrm/xfrm_policy.c
1197
&policy->selector.saddr,
net/xfrm/xfrm_policy.c
1198
policy->family,
net/xfrm/xfrm_policy.c
1199
policy->selector.prefixlen_s,
net/xfrm/xfrm_policy.c
1212
&policy->selector.daddr,
net/xfrm/xfrm_policy.c
1213
policy->family,
net/xfrm/xfrm_policy.c
1214
policy->selector.prefixlen_d, dir);
net/xfrm/xfrm_policy.c
1220
if (xfrm_pol_inexact_addr_use_any_list(&policy->selector.saddr,
net/xfrm/xfrm_policy.c
1221
policy->family,
net/xfrm/xfrm_policy.c
1222
policy->selector.prefixlen_s))
net/xfrm/xfrm_policy.c
1228
&policy->selector.saddr,
net/xfrm/xfrm_policy.c
1229
policy->family,
net/xfrm/xfrm_policy.c
1230
policy->selector.prefixlen_s, dir);
net/xfrm/xfrm_policy.c
1239
xfrm_policy_inexact_insert(struct xfrm_policy *policy, u8 dir, int excl)
net/xfrm/xfrm_policy.c
1246
bin = xfrm_policy_inexact_alloc_bin(policy, dir);
net/xfrm/xfrm_policy.c
1250
net = xp_net(policy);
net/xfrm/xfrm_policy.c
1253
chain = xfrm_policy_inexact_alloc_chain(bin, policy, dir);
net/xfrm/xfrm_policy.c
1259
delpol = xfrm_policy_insert_list(chain, policy, excl);
net/xfrm/xfrm_policy.c
1271
static bool xfrm_policy_is_dead_or_sk(const struct xfrm_policy *policy)
net/xfrm/xfrm_policy.c
1275
if (policy->walk.dead)
net/xfrm/xfrm_policy.c
1278
dir = xfrm_policy_id2dir(policy->index);
net/xfrm/xfrm_policy.c
1287
struct xfrm_policy *policy;
net/xfrm/xfrm_policy.c
1312
list_for_each_entry(policy, &net->xfrm.policy_all, walk.all) {
net/xfrm/xfrm_policy.c
1316
if (xfrm_policy_is_dead_or_sk(policy))
net/xfrm/xfrm_policy.c
1319
dir = xfrm_policy_id2dir(policy->index);
net/xfrm/xfrm_policy.c
1321
if (policy->family == AF_INET) {
net/xfrm/xfrm_policy.c
1329
if (policy->family == AF_INET) {
net/xfrm/xfrm_policy.c
1338
if (policy->selector.prefixlen_d < dbits ||
net/xfrm/xfrm_policy.c
1339
policy->selector.prefixlen_s < sbits)
net/xfrm/xfrm_policy.c
1342
bin = xfrm_policy_inexact_alloc_bin(policy, dir);
net/xfrm/xfrm_policy.c
1346
if (!xfrm_policy_inexact_alloc_chain(bin, policy, dir))
net/xfrm/xfrm_policy.c
1367
list_for_each_entry_reverse(policy, &net->xfrm.policy_all, walk.all) {
net/xfrm/xfrm_policy.c
1368
if (xfrm_policy_is_dead_or_sk(policy))
net/xfrm/xfrm_policy.c
1371
hlist_del_rcu(&policy->bydst);
net/xfrm/xfrm_policy.c
1374
dir = xfrm_policy_id2dir(policy->index);
net/xfrm/xfrm_policy.c
1375
chain = policy_hash_bysel(net, &policy->selector,
net/xfrm/xfrm_policy.c
1376
policy->family, dir);
net/xfrm/xfrm_policy.c
1379
void *p = xfrm_policy_inexact_insert(policy, dir, 0);
net/xfrm/xfrm_policy.c
1386
if (policy->priority >= pol->priority)
net/xfrm/xfrm_policy.c
1391
if (newpos && policy->xdo.type != XFRM_DEV_OFFLOAD_PACKET)
net/xfrm/xfrm_policy.c
1392
hlist_add_behind_rcu(&policy->bydst, newpos);
net/xfrm/xfrm_policy.c
1394
hlist_add_head_rcu(&policy->bydst, chain);
net/xfrm/xfrm_policy.c
1542
struct xfrm_policy *policy,
net/xfrm/xfrm_policy.c
1548
if (pol->type == policy->type &&
net/xfrm/xfrm_policy.c
1549
pol->if_id == policy->if_id &&
net/xfrm/xfrm_policy.c
1550
!selector_cmp(&pol->selector, &policy->selector) &&
net/xfrm/xfrm_policy.c
1551
xfrm_policy_mark_match(&policy->mark, pol) &&
net/xfrm/xfrm_policy.c
1552
xfrm_sec_ctx_match(pol->security, policy->security) &&
net/xfrm/xfrm_policy.c
1557
if (policy->priority > pol->priority)
net/xfrm/xfrm_policy.c
1559
} else if (policy->priority >= pol->priority) {
net/xfrm/xfrm_policy.c
1567
if (newpos && policy->xdo.type != XFRM_DEV_OFFLOAD_PACKET)
net/xfrm/xfrm_policy.c
1568
hlist_add_behind_rcu(&policy->bydst, &newpos->bydst);
net/xfrm/xfrm_policy.c
1573
hlist_add_head_rcu(&policy->bydst, chain);
net/xfrm/xfrm_policy.c
1578
int xfrm_policy_insert(int dir, struct xfrm_policy *policy, int excl)
net/xfrm/xfrm_policy.c
1580
struct net *net = xp_net(policy);
net/xfrm/xfrm_policy.c
1585
policy->mark.v &= policy->mark.m;
net/xfrm/xfrm_policy.c
1588
chain = policy_hash_bysel(net, &policy->selector, policy->family, dir);
net/xfrm/xfrm_policy.c
1590
delpol = xfrm_policy_insert_list(chain, policy, excl);
net/xfrm/xfrm_policy.c
1592
delpol = xfrm_policy_inexact_insert(policy, dir, excl);
net/xfrm/xfrm_policy.c
1599
__xfrm_policy_link(policy, dir);
net/xfrm/xfrm_policy.c
1602
if (policy->family == AF_INET)
net/xfrm/xfrm_policy.c
1608
xfrm_policy_requeue(delpol, policy);
net/xfrm/xfrm_policy.c
1611
policy->index = delpol ? delpol->index : xfrm_gen_index(net, dir, policy->index);
net/xfrm/xfrm_policy.c
1612
hlist_add_head(&policy->byidx, net->xfrm.policy_byidx+idx_hash(net, policy->index));
net/xfrm/xfrm_policy.c
1613
policy->curlft.add_time = ktime_get_real_seconds();
net/xfrm/xfrm_policy.c
1614
policy->curlft.use_time = 0;
net/xfrm/xfrm_policy.c
1615
if (!mod_timer(&policy->timer, jiffies + HZ))
net/xfrm/xfrm_policy.c
1616
xfrm_pol_hold(policy);
net/xfrm/xfrm_policy.c
202
xfrm_policy_insert_list(struct hlist_head *chain, struct xfrm_policy *policy,
net/xfrm/xfrm_policy.c
211
static inline bool xfrm_pol_hold_rcu(struct xfrm_policy *policy)
net/xfrm/xfrm_policy.c
213
return refcount_inc_not_zero(&policy->refcnt);
net/xfrm/xfrm_policy.c
2285
struct xfrm_policy *policy;
net/xfrm/xfrm_policy.c
2289
list_for_each_entry_reverse(policy, &net->xfrm.policy_all, walk.all) {
net/xfrm/xfrm_policy.c
2290
if (!xfrm_policy_is_dead_or_sk(policy))
net/xfrm/xfrm_policy.c
2291
policy->pos = ++i;
net/xfrm/xfrm_policy.c
2299
const struct xfrm_policy *policy;
net/xfrm/xfrm_policy.c
2303
list_for_each_entry(policy, &net->xfrm.policy_all, walk.all) {
net/xfrm/xfrm_policy.c
2304
if (xfrm_policy_is_dead_or_sk(policy))
net/xfrm/xfrm_policy.c
2307
if (policy->pos == UINT_MAX)
net/xfrm/xfrm_policy.c
2310
i = policy->pos + 1;
net/xfrm/xfrm_policy.c
2486
xfrm_tmpl_resolve_one(struct xfrm_policy *policy, const struct flowi *fl,
net/xfrm/xfrm_policy.c
2489
struct net *net = xp_net(policy);
net/xfrm/xfrm_policy.c
2496
for (nx = 0, i = 0; i < policy->xfrm_nr; i++) {
net/xfrm/xfrm_policy.c
2500
struct xfrm_tmpl *tmpl = &policy->xfrm_vec[i];
net/xfrm/xfrm_policy.c
2522
x = xfrm_state_find(remote, local, fl, tmpl, policy, &error,
net/xfrm/xfrm_policy.c
2523
family, policy->if_id);
net/xfrm/xfrm_policy.c
2666
static struct dst_entry *xfrm_bundle_create(struct xfrm_policy *policy,
net/xfrm/xfrm_policy.c
2675
struct net *net = xp_net(policy);
net/xfrm/xfrm_policy.c
2685
int family = policy->selector.family;
net/xfrm/xfrm_policy.c
430
struct xfrm_policy *policy;
net/xfrm/xfrm_policy.c
432
policy = kzalloc_obj(struct xfrm_policy, gfp);
net/xfrm/xfrm_policy.c
434
if (policy) {
net/xfrm/xfrm_policy.c
435
write_pnet(&policy->xp_net, net);
net/xfrm/xfrm_policy.c
436
INIT_LIST_HEAD(&policy->walk.all);
net/xfrm/xfrm_policy.c
437
INIT_HLIST_HEAD(&policy->state_cache_list);
net/xfrm/xfrm_policy.c
438
INIT_HLIST_NODE(&policy->bydst);
net/xfrm/xfrm_policy.c
439
INIT_HLIST_NODE(&policy->byidx);
net/xfrm/xfrm_policy.c
440
rwlock_init(&policy->lock);
net/xfrm/xfrm_policy.c
441
refcount_set(&policy->refcnt, 1);
net/xfrm/xfrm_policy.c
442
skb_queue_head_init(&policy->polq.hold_queue);
net/xfrm/xfrm_policy.c
443
timer_setup(&policy->timer, xfrm_policy_timer, 0);
net/xfrm/xfrm_policy.c
444
timer_setup(&policy->polq.hold_timer,
net/xfrm/xfrm_policy.c
447
return policy;
net/xfrm/xfrm_policy.c
453
struct xfrm_policy *policy = container_of(head, struct xfrm_policy, rcu);
net/xfrm/xfrm_policy.c
455
security_xfrm_policy_free(policy->security);
net/xfrm/xfrm_policy.c
456
kfree(policy);
net/xfrm/xfrm_policy.c
461
void xfrm_policy_destroy(struct xfrm_policy *policy)
net/xfrm/xfrm_policy.c
463
BUG_ON(!policy->walk.dead);
net/xfrm/xfrm_policy.c
465
if (timer_delete(&policy->timer) || timer_delete(&policy->polq.hold_timer))
net/xfrm/xfrm_policy.c
468
xfrm_dev_policy_free(policy);
net/xfrm/xfrm_policy.c
469
call_rcu(&policy->rcu, xfrm_policy_destroy_rcu);
net/xfrm/xfrm_policy.c
477
static void xfrm_policy_kill(struct xfrm_policy *policy)
net/xfrm/xfrm_policy.c
479
struct net *net = xp_net(policy);
net/xfrm/xfrm_policy.c
482
xfrm_dev_policy_delete(policy);
net/xfrm/xfrm_policy.c
484
write_lock_bh(&policy->lock);
net/xfrm/xfrm_policy.c
485
policy->walk.dead = 1;
net/xfrm/xfrm_policy.c
486
write_unlock_bh(&policy->lock);
net/xfrm/xfrm_policy.c
488
atomic_inc(&policy->genid);
net/xfrm/xfrm_policy.c
490
if (timer_delete(&policy->polq.hold_timer))
net/xfrm/xfrm_policy.c
491
xfrm_pol_put(policy);
net/xfrm/xfrm_policy.c
492
skb_queue_purge(&policy->polq.hold_queue);
net/xfrm/xfrm_policy.c
494
if (timer_delete(&policy->timer))
net/xfrm/xfrm_policy.c
495
xfrm_pol_put(policy);
net/xfrm/xfrm_policy.c
499
hlist_for_each_entry_rcu(x, &policy->state_cache_list, state_cache) {
net/xfrm/xfrm_policy.c
504
xfrm_pol_put(policy);
net/xfrm/xfrm_policy.c
807
xfrm_policy_inexact_insert_use_any_list(const struct xfrm_policy *policy)
net/xfrm/xfrm_policy.c
813
addr = &policy->selector.saddr;
net/xfrm/xfrm_policy.c
814
prefixlen = policy->selector.prefixlen_s;
net/xfrm/xfrm_policy.c
817
policy->family,
net/xfrm/xfrm_policy.c
819
addr = &policy->selector.daddr;
net/xfrm/xfrm_policy.c
820
prefixlen = policy->selector.prefixlen_d;
net/xfrm/xfrm_policy.c
822
policy->family,
net/xfrm/xfrm_policy.c
897
struct xfrm_policy *policy, *p;
net/xfrm/xfrm_policy.c
902
list_for_each_entry_reverse(policy, &net->xfrm.policy_all, walk.all) {
net/xfrm/xfrm_policy.c
906
if (policy->walk.dead || !policy->bydst_reinsert)
net/xfrm/xfrm_policy.c
909
WARN_ON_ONCE(policy->family != family);
net/xfrm/xfrm_policy.c
911
policy->bydst_reinsert = false;
net/xfrm/xfrm_policy.c
913
if (policy->priority > p->priority)
net/xfrm/xfrm_policy.c
915
else if (policy->priority == p->priority &&
net/xfrm/xfrm_policy.c
916
policy->pos > p->pos)
net/xfrm/xfrm_policy.c
922
if (newpos && policy->xdo.type != XFRM_DEV_OFFLOAD_PACKET)
net/xfrm/xfrm_policy.c
923
hlist_add_behind_rcu(&policy->bydst, newpos);
net/xfrm/xfrm_policy.c
925
hlist_add_head_rcu(&policy->bydst, &n->hhead);
net/xfrm/xfrm_policy.c
935
matches_s = xfrm_policy_addr_delta(&policy->selector.saddr,
net/xfrm/xfrm_policy.c
939
matches_d = xfrm_policy_addr_delta(&policy->selector.daddr,
net/xfrm/xfrm_user.c
2501
static bool xfrm_userpolicy_is_valid(__u8 policy)
net/xfrm/xfrm_user.c
2503
return policy == XFRM_USERPOLICY_BLOCK ||
net/xfrm/xfrm_user.c
2504
policy == XFRM_USERPOLICY_ACCEPT;
net/xfrm/xfrm_user.c
3020
err = verify_newpolicy_info(&ua->policy, extack);
net/xfrm/xfrm_user.c
3028
xp = xfrm_policy_construct(net, &ua->policy, attrs, &err, extack);
net/xfrm/xfrm_user.c
3819
copy_to_user_policy(xp, &ua->policy, XFRM_POLICY_OUT);
rust/helpers/cpufreq.c
7
rust_helper_cpufreq_register_em_with_opp(struct cpufreq_policy *policy)
rust/helpers/cpufreq.c
9
cpufreq_register_em_with_opp(policy);
scripts/ipe/polgen/polgen.c
117
char *policy = NULL;
scripts/ipe/polgen/polgen.c
125
rc = policy_to_buffer(argv[2], &policy, &len);
scripts/ipe/polgen/polgen.c
130
rc = write_boot_policy(argv[1], policy, len);
scripts/ipe/polgen/polgen.c
132
if (policy)
scripts/ipe/polgen/polgen.c
133
free(policy);
security/apparmor/af_unix.c
122
static aa_state_t match_to_sk(struct aa_policydb *policy,
security/apparmor/af_unix.c
130
return match_to_local(policy, state, request, u->sk.sk_type,
security/apparmor/af_unix.c
138
static aa_state_t match_to_cmd(struct aa_policydb *policy, aa_state_t state,
security/apparmor/af_unix.c
145
state = match_to_sk(policy, state, request, u, p, info);
security/apparmor/af_unix.c
147
state = aa_dfa_match_len(policy->dfa, state, &cmd, 1);
security/apparmor/af_unix.c
155
static aa_state_t match_to_peer(struct aa_policydb *policy, aa_state_t state,
security/apparmor/af_unix.c
162
state = match_to_cmd(policy, state, request, u, CMD_ADDR, p, info);
security/apparmor/af_unix.c
164
state = match_addr(policy->dfa, state, peer_addr, peer_addrlen);
security/apparmor/af_unix.c
184
state = aa_dfa_match(rule->policy->dfa, state,
security/apparmor/af_unix.c
191
return aa_do_perms(profile, rule->policy, state, request, p, ad);
security/apparmor/af_unix.c
213
state = aa_match_to_prot(rules->policy, state, AA_MAY_CREATE,
security/apparmor/af_unix.c
217
return aa_do_perms(profile, rules->policy, state, AA_MAY_CREATE,
security/apparmor/af_unix.c
244
state = match_to_sk(rules->policy, state, request, unix_sk(sk),
security/apparmor/af_unix.c
247
return aa_do_perms(profile, rules->policy, state, request, p,
security/apparmor/af_unix.c
272
state = match_to_local(rules->policy, state, AA_MAY_BIND,
security/apparmor/af_unix.c
278
return aa_do_perms(profile, rules->policy, state, AA_MAY_BIND,
security/apparmor/af_unix.c
306
state = match_to_cmd(rules->policy, state, AA_MAY_LISTEN,
security/apparmor/af_unix.c
309
state = aa_dfa_match_len(rules->policy->dfa, state,
security/apparmor/af_unix.c
314
return aa_do_perms(profile, rules->policy, state, AA_MAY_LISTEN,
security/apparmor/af_unix.c
341
state = match_to_sk(rules->policy, state, AA_MAY_ACCEPT,
security/apparmor/af_unix.c
344
return aa_do_perms(profile, rules->policy, state, AA_MAY_ACCEPT,
security/apparmor/af_unix.c
372
state = match_to_cmd(rules->policy, state, request, unix_sk(sk),
security/apparmor/af_unix.c
375
state = aa_dfa_match_len(rules->policy->dfa, state,
security/apparmor/af_unix.c
380
return aa_do_perms(profile, rules->policy, state, request, p,
security/apparmor/af_unix.c
415
state = match_to_peer(rules->policy, state, request,
security/apparmor/af_unix.c
84
static aa_state_t match_to_local(struct aa_policydb *policy,
security/apparmor/af_unix.c
91
state = aa_match_to_prot(policy, state, request, PF_UNIX, type,
security/apparmor/af_unix.c
94
state = match_addr(policy->dfa, state, addr, addrlen);
security/apparmor/af_unix.c
97
state = aa_dfa_null_transition(policy->dfa, state);
security/apparmor/apparmorfs.c
715
} else if (rules->policy->dfa) {
security/apparmor/apparmorfs.c
726
state = aa_dfa_match_len(rules->policy->dfa,
security/apparmor/apparmorfs.c
727
rules->policy->start[0],
security/apparmor/apparmorfs.c
730
tmp = *aa_lookup_perms(rules->policy, state);
security/apparmor/capability.c
134
state = aa_dfa_next(rules->policy->dfa, state, cap >> 5);
security/apparmor/capability.c
136
perms = *aa_lookup_perms(rules->policy, state);
security/apparmor/capability.c
209
tmp = aa_dfa_next(rules->policy->dfa, state, i);
security/apparmor/capability.c
210
perms = *aa_lookup_perms(rules->policy, tmp);
security/apparmor/include/lib.h
236
struct aa_policy *policy;
security/apparmor/include/lib.h
238
list_for_each_entry_rcu(policy, head, list) {
security/apparmor/include/lib.h
239
if (!strcmp(policy->name, name))
security/apparmor/include/lib.h
240
return policy;
security/apparmor/include/lib.h
261
struct aa_policy *policy;
security/apparmor/include/lib.h
263
list_for_each_entry_rcu(policy, head, list) {
security/apparmor/include/lib.h
264
if (aa_strneq(policy->name, str, len))
security/apparmor/include/lib.h
265
return policy;
security/apparmor/include/lib.h
271
bool aa_policy_init(struct aa_policy *policy, const char *prefix,
security/apparmor/include/lib.h
273
void aa_policy_destroy(struct aa_policy *policy);
security/apparmor/include/net.h
88
int aa_do_perms(struct aa_profile *profile, struct aa_policydb *policy,
security/apparmor/include/net.h
92
aa_state_t aa_match_to_prot(struct aa_policydb *policy, aa_state_t state,
security/apparmor/include/policy.h
157
static inline struct aa_perms *aa_lookup_perms(struct aa_policydb *policy,
security/apparmor/include/policy.h
160
unsigned int index = ACCEPT_TABLE(policy->dfa)[state];
security/apparmor/include/policy.h
162
if (!(policy->perms))
security/apparmor/include/policy.h
165
return &(policy->perms[index]);
security/apparmor/include/policy.h
195
struct aa_policydb *policy;
security/apparmor/include/policy.h
332
return rules->policy->start[class];
security/apparmor/include/policy.h
334
return aa_dfa_match_len(rules->policy->dfa,
security/apparmor/include/policy.h
335
rules->policy->start[0], &class, 1);
security/apparmor/include/policy_compat.h
29
int aa_compat_map_xmatch(struct aa_policydb *policy);
security/apparmor/include/policy_compat.h
30
int aa_compat_map_policy(struct aa_policydb *policy, u32 version);
security/apparmor/include/policy_compat.h
31
int aa_compat_map_file(struct aa_policydb *policy);
security/apparmor/ipc.c
96
state = aa_dfa_next(rules->policy->dfa, state, ad->signal);
security/apparmor/label.c
1265
return aa_dfa_match(rules->policy->dfa, state, tp->base.hname);
security/apparmor/label.c
1269
state = aa_dfa_match_len(rules->policy->dfa, state, ":", 1);
security/apparmor/label.c
1270
state = aa_dfa_match(rules->policy->dfa, state, ns_name);
security/apparmor/label.c
1271
state = aa_dfa_match_len(rules->policy->dfa, state, ":", 1);
security/apparmor/label.c
1272
return aa_dfa_match(rules->policy->dfa, state, tp->base.hname);
security/apparmor/label.c
1318
state = aa_dfa_match(rules->policy->dfa, state, "//&");
security/apparmor/label.c
1323
*perms = *aa_lookup_perms(rules->policy, state);
security/apparmor/label.c
1372
tmp = *aa_lookup_perms(rules->policy, state);
security/apparmor/label.c
1380
tmp = *aa_lookup_perms(rules->policy, state);
security/apparmor/lib.c
396
state = aa_dfa_next(rules->policy->dfa,
security/apparmor/lib.c
397
rules->policy->start[AA_CLASS_LABEL],
security/apparmor/lib.c
479
bool aa_policy_init(struct aa_policy *policy, const char *prefix,
security/apparmor/lib.c
494
policy->hname = hname;
security/apparmor/lib.c
496
policy->name = basename(policy->hname);
security/apparmor/lib.c
497
INIT_LIST_HEAD(&policy->list);
security/apparmor/lib.c
498
INIT_LIST_HEAD(&policy->profiles);
security/apparmor/lib.c
507
void aa_policy_destroy(struct aa_policy *policy)
security/apparmor/lib.c
509
AA_BUG(on_list_rcu(&policy->profiles));
security/apparmor/lib.c
510
AA_BUG(on_list_rcu(&policy->list));
security/apparmor/lib.c
513
aa_put_str(policy->hname);
security/apparmor/lsm.c
639
perms = *aa_lookup_perms(rules->policy, state);
security/apparmor/mount.c
226
static int do_match_mnt(struct aa_policydb *policy, aa_state_t start,
security/apparmor/mount.c
233
AA_BUG(!policy);
security/apparmor/mount.c
234
AA_BUG(!policy->dfa);
security/apparmor/mount.c
235
AA_BUG(!policy->perms);
security/apparmor/mount.c
238
state = aa_dfa_match(policy->dfa, start, mntpnt);
security/apparmor/mount.c
239
state = aa_dfa_null_transition(policy->dfa, state);
security/apparmor/mount.c
244
state = aa_dfa_match(policy->dfa, state, devname);
security/apparmor/mount.c
245
state = aa_dfa_null_transition(policy->dfa, state);
security/apparmor/mount.c
250
state = aa_dfa_match(policy->dfa, state, type);
security/apparmor/mount.c
251
state = aa_dfa_null_transition(policy->dfa, state);
security/apparmor/mount.c
255
state = match_mnt_flags(policy->dfa, state, flags);
security/apparmor/mount.c
258
*perms = *aa_lookup_perms(policy, state);
security/apparmor/mount.c
264
state = aa_dfa_null_transition(policy->dfa, state);
security/apparmor/mount.c
268
state = aa_dfa_match(policy->dfa, state, data);
security/apparmor/mount.c
271
*perms = *aa_lookup_perms(policy, state);
security/apparmor/mount.c
336
pos = do_match_mnt(rules->policy,
security/apparmor/mount.c
337
rules->policy->start[AA_CLASS_MOUNT],
security/apparmor/mount.c
622
state = aa_dfa_match(rules->policy->dfa,
security/apparmor/mount.c
623
rules->policy->start[AA_CLASS_MOUNT],
security/apparmor/mount.c
625
perms = *aa_lookup_perms(rules->policy, state);
security/apparmor/mount.c
695
state = aa_dfa_match(rules->policy->dfa,
security/apparmor/mount.c
696
rules->policy->start[AA_CLASS_MOUNT],
security/apparmor/mount.c
698
state = aa_dfa_null_transition(rules->policy->dfa, state);
security/apparmor/mount.c
699
state = aa_dfa_match(rules->policy->dfa, state, old_name);
security/apparmor/mount.c
700
perms = *aa_lookup_perms(rules->policy, state);
security/apparmor/net.c
165
int aa_do_perms(struct aa_profile *profile, struct aa_policydb *policy,
security/apparmor/net.c
172
AA_BUG(!policy);
security/apparmor/net.c
176
p = aa_lookup_perms(policy, state);
security/apparmor/net.c
188
static struct aa_perms *early_match(struct aa_policydb *policy,
security/apparmor/net.c
193
p = aa_lookup_perms(policy, state);
security/apparmor/net.c
224
aa_state_t aa_match_to_prot(struct aa_policydb *policy, aa_state_t state,
security/apparmor/net.c
228
state = aa_dfa_match_be16(policy->dfa, state, (u16)af);
security/apparmor/net.c
233
state = aa_dfa_match_be16(policy->dfa, state, (u16)type);
security/apparmor/net.c
236
*p = early_match(policy, state, request);
security/apparmor/net.c
238
state = aa_dfa_match_be16(policy->dfa, state, (u16)protocol);
security/apparmor/net.c
267
state = aa_match_to_prot(rules->policy, state, request, family, type,
security/apparmor/net.c
269
return aa_do_perms(profile, rules->policy, state, request, p, ad);
security/apparmor/policy.c
1238
struct aa_policy *policy;
security/apparmor/policy.c
1265
policy = __lookup_parent(ns, ent->new->base.hname);
security/apparmor/policy.c
1266
if (!policy) {
security/apparmor/policy.c
1282
policy = __create_missing_ancestors(ns,
security/apparmor/policy.c
1285
if (!policy) {
security/apparmor/policy.c
1292
if (!p && policy != &ns->base)
security/apparmor/policy.c
1294
p = (struct aa_profile *) policy;
security/apparmor/policy.c
288
aa_put_pdb(rules->policy);
security/apparmor/policy.c
518
struct aa_policy *policy;
security/apparmor/policy.c
522
policy = &ns->base;
security/apparmor/policy.c
525
profile = __strn_find_child(&policy->profiles, hname,
security/apparmor/policy.c
529
policy = &profile->base;
security/apparmor/policy.c
553
struct aa_policy *policy;
security/apparmor/policy.c
560
policy = &ns->base;
security/apparmor/policy.c
564
profile = __strn_find_child(&policy->profiles, hname,
security/apparmor/policy.c
578
policy = &profile->base;
security/apparmor/policy.c
699
rules->policy = aa_get_pdb(nullpdb);
security/apparmor/policy_compat.c
302
int aa_compat_map_xmatch(struct aa_policydb *policy)
security/apparmor/policy_compat.c
304
policy->perms = compute_xmatch_perms(policy->dfa, &policy->size);
security/apparmor/policy_compat.c
305
if (!policy->perms)
security/apparmor/policy_compat.c
308
remap_dfa_accept(policy->dfa, 1);
security/apparmor/policy_compat.c
313
int aa_compat_map_policy(struct aa_policydb *policy, u32 version)
security/apparmor/policy_compat.c
315
policy->perms = compute_perms(policy->dfa, version, &policy->size);
security/apparmor/policy_compat.c
316
if (!policy->perms)
security/apparmor/policy_compat.c
319
remap_dfa_accept(policy->dfa, 1);
security/apparmor/policy_compat.c
324
int aa_compat_map_file(struct aa_policydb *policy)
security/apparmor/policy_compat.c
326
policy->perms = compute_fperms(policy->dfa, &policy->size);
security/apparmor/policy_compat.c
327
if (!policy->perms)
security/apparmor/policy_compat.c
330
remap_dfa_accept(policy->dfa, 2);
security/apparmor/policy_unpack.c
1082
*policy = pdb;
security/apparmor/policy_unpack.c
1300
error = unpack_pdb(e, &rules->policy, true, false,
security/apparmor/policy_unpack.c
1305
if (aa_dfa_next(rules->policy->dfa, rules->policy->start[0],
security/apparmor/policy_unpack.c
1307
rules->policy->start[AA_CLASS_FILE] =
security/apparmor/policy_unpack.c
1308
aa_dfa_next(rules->policy->dfa,
security/apparmor/policy_unpack.c
1309
rules->policy->start[0],
security/apparmor/policy_unpack.c
1313
if (!rules->policy->perms) {
security/apparmor/policy_unpack.c
1314
error = aa_compat_map_policy(rules->policy,
security/apparmor/policy_unpack.c
1322
rules->policy = aa_get_pdb(nullpdb);
security/apparmor/policy_unpack.c
1337
} else if (rules->policy->dfa &&
security/apparmor/policy_unpack.c
1338
rules->policy->start[AA_CLASS_FILE]) {
security/apparmor/policy_unpack.c
1340
rules->file = aa_get_pdb(rules->policy);
security/apparmor/policy_unpack.c
1570
if (rules->policy->dfa &&
security/apparmor/policy_unpack.c
1571
!verify_dfa_accept_index(rules->policy->dfa, rules->policy->size)) {
security/apparmor/policy_unpack.c
1583
if (!verify_perms(rules->policy)) {
security/apparmor/policy_unpack.c
965
static int unpack_pdb(struct aa_ext *e, struct aa_policydb **policy,
security/apparmor/task.c
365
perms = *aa_lookup_perms(rules->policy, state);
security/integrity/ima/ima.h
325
hook(POLICY_CHECK, policy) \
security/integrity/ima/ima_policy.c
1051
struct list_head *policy = &ima_policy_rules;
security/integrity/ima/ima_policy.c
1053
list_splice_tail_init_rcu(&ima_temp_rules, policy, synchronize_rcu);
security/integrity/ima/ima_policy.c
1055
if (ima_rules != (struct list_head __rcu *)policy) {
security/integrity/ima/ima_policy.c
1058
rcu_assign_pointer(ima_rules, policy);
security/ipe/policy_parser.c
508
char *policy = NULL, *dup = NULL;
security/ipe/policy_parser.c
517
policy = kmemdup_nul(p->text, p->textlen, GFP_KERNEL);
security/ipe/policy_parser.c
518
if (!policy)
security/ipe/policy_parser.c
520
dup = policy;
security/ipe/policy_parser.c
528
while ((line = strsep(&policy, IPE_LINE_DELIM)) != NULL) {
security/ipe/policy_tests.c
12
const char *const policy;
security/ipe/policy_tests.c
252
pol = ipe_new_policy(p->policy, strlen(p->policy), NULL, 0);
security/ipe/policy_tests.c
261
KUNIT_EXPECT_STREQ(test, pol->text, p->policy);
security/ipe/policy_tests.c
276
const unsigned short policy[] = L"policy_name=Test policy_version=0.0.0\n"
security/ipe/policy_tests.c
280
pol = ipe_new_policy((const char *)policy, (ARRAY_SIZE(policy) - 1) * 2, NULL, 0);
security/safesetid/lsm.c
33
enum sid_policy_type _setid_policy_lookup(struct setid_ruleset *policy,
security/safesetid/lsm.c
39
if (policy->type == UID) {
security/safesetid/lsm.c
40
hash_for_each_possible(policy->rules, rule, next, __kuid_val(src.uid)) {
security/safesetid/lsm.c
47
} else if (policy->type == GID) {
security/safesetid/lsm.c
48
hash_for_each_possible(policy->rules, rule, next, __kgid_val(src.gid)) {
security/safesetid/lsm.h
67
enum sid_policy_type _setid_policy_lookup(struct setid_ruleset *policy,
security/selinux/ima.c
104
policy, policy_len, true,
security/selinux/ima.c
107
vfree(policy);
security/selinux/ima.c
73
void *policy = NULL;
security/selinux/ima.c
97
rc = security_read_state_kernel(&policy, &policy_len);
security/selinux/include/conditional.h
16
int security_get_bools(struct selinux_policy *policy, u32 *len, char ***names,
security/selinux/include/security.h
104
struct selinux_policy __rcu *policy;
security/selinux/include/security.h
226
struct selinux_policy *policy;
security/selinux/include/security.h
340
int security_get_classes(struct selinux_policy *policy, char ***classes,
security/selinux/include/security.h
342
int security_get_permissions(struct selinux_policy *policy, const char *class,
security/selinux/include/security.h
361
int selinux_policy_genfs_sid(struct selinux_policy *policy, const char *fstype,
security/selinux/selinuxfs.c
620
length = sel_make_policy_nodes(fsi, load_state.policy);
security/selinux/ss/services.c
1033
struct selinux_policy *policy;
security/selinux/ss/services.c
1055
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1056
policydb = &policy->policydb;
security/selinux/ss/services.c
1057
sidtab = policy->sidtab;
security/selinux/ss/services.c
1073
tclass = unmap_class(&policy->map, orig_tclass);
security/selinux/ss/services.c
1129
struct selinux_policy *policy;
security/selinux/ss/services.c
1136
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1137
avd_init(policy, avd);
security/selinux/ss/services.c
1142
policydb = &policy->policydb;
security/selinux/ss/services.c
1143
sidtab = policy->sidtab;
security/selinux/ss/services.c
1171
tclass = unmap_class(&policy->map, orig_tclass);
security/selinux/ss/services.c
1179
map_decision(&policy->map, orig_tclass, avd,
security/selinux/ss/services.c
1196
struct selinux_policy *policy;
security/selinux/ss/services.c
1202
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1203
avd_init(policy, avd);
security/selinux/ss/services.c
1207
policydb = &policy->policydb;
security/selinux/ss/services.c
1208
sidtab = policy->sidtab;
security/selinux/ss/services.c
1332
struct selinux_policy *policy;
security/selinux/ss/services.c
1342
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1343
rc = sidtab_hash_stats(policy->sidtab, page);
security/selinux/ss/services.c
1360
struct selinux_policy *policy;
security/selinux/ss/services.c
1402
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1403
policydb = &policy->policydb;
security/selinux/ss/services.c
1404
sidtab = policy->sidtab;
security/selinux/ss/services.c
1556
struct selinux_policy *policy;
security/selinux/ss/services.c
1597
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1598
policydb = &policy->policydb;
security/selinux/ss/services.c
1599
sidtab = policy->sidtab;
security/selinux/ss/services.c
1686
struct selinux_policy *policy,
security/selinux/ss/services.c
1692
struct policydb *policydb = &policy->policydb;
security/selinux/ss/services.c
1693
struct sidtab *sidtab = policy->sidtab;
security/selinux/ss/services.c
1761
struct selinux_policy *policy;
security/selinux/ss/services.c
1791
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
1794
tclass = unmap_class(&policy->map, orig_tclass);
security/selinux/ss/services.c
1798
sock = security_is_socket_class(map_class(&policy->map,
security/selinux/ss/services.c
1802
policydb = &policy->policydb;
security/selinux/ss/services.c
1803
sidtab = policy->sidtab;
security/selinux/ss/services.c
1925
rc = compute_sid_handle_invalid_context(policy, sentry,
security/selinux/ss/services.c
2170
static void security_load_policycaps(struct selinux_policy *policy)
security/selinux/ss/services.c
2176
p = &policy->policydb;
security/selinux/ss/services.c
2197
static void selinux_policy_free(struct selinux_policy *policy)
security/selinux/ss/services.c
2199
if (!policy)
security/selinux/ss/services.c
2202
sidtab_destroy(policy->sidtab);
security/selinux/ss/services.c
2203
kfree(policy->map.mapping);
security/selinux/ss/services.c
2204
policydb_destroy(&policy->policydb);
security/selinux/ss/services.c
2205
kfree(policy->sidtab);
security/selinux/ss/services.c
2206
kfree(policy);
security/selinux/ss/services.c
2209
static void selinux_policy_cond_free(struct selinux_policy *policy)
security/selinux/ss/services.c
2211
cond_policydb_destroy_dup(&policy->policydb);
security/selinux/ss/services.c
2212
kfree(policy);
security/selinux/ss/services.c
2220
oldpolicy = rcu_dereference_protected(state->policy,
security/selinux/ss/services.c
2224
selinux_policy_free(load_state->policy);
security/selinux/ss/services.c
2242
struct selinux_policy *oldpolicy, *newpolicy = load_state->policy;
security/selinux/ss/services.c
2246
oldpolicy = rcu_dereference_protected(state->policy,
security/selinux/ss/services.c
2267
rcu_assign_pointer(state->policy, newpolicy);
security/selinux/ss/services.c
2270
rcu_assign_pointer(state->policy, newpolicy);
security/selinux/ss/services.c
2343
load_state->policy = newpolicy;
security/selinux/ss/services.c
2348
oldpolicy = rcu_dereference_protected(state->policy,
security/selinux/ss/services.c
2383
load_state->policy = newpolicy;
security/selinux/ss/services.c
242
struct selinux_policy *policy;
security/selinux/ss/services.c
2448
struct selinux_policy *policy;
security/selinux/ss/services.c
2462
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2463
policydb = &policy->policydb;
security/selinux/ss/services.c
2464
sidtab = policy->sidtab;
security/selinux/ss/services.c
248
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
249
mls_enabled = policy->policydb.mls_enabled;
security/selinux/ss/services.c
2500
struct selinux_policy *policy;
security/selinux/ss/services.c
2514
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2515
policydb = &policy->policydb;
security/selinux/ss/services.c
2516
sidtab = policy->sidtab;
security/selinux/ss/services.c
2552
struct selinux_policy *policy;
security/selinux/ss/services.c
2566
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2567
policydb = &policy->policydb;
security/selinux/ss/services.c
2568
sidtab = policy->sidtab;
security/selinux/ss/services.c
2604
struct selinux_policy *policy;
security/selinux/ss/services.c
2619
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2620
policydb = &policy->policydb;
security/selinux/ss/services.c
2621
sidtab = policy->sidtab;
security/selinux/ss/services.c
2676
struct selinux_policy *policy;
security/selinux/ss/services.c
2689
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2690
policydb = &policy->policydb;
security/selinux/ss/services.c
2691
sidtab = policy->sidtab;
security/selinux/ss/services.c
2770
struct selinux_policy *policy;
security/selinux/ss/services.c
2794
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2795
policydb = &policy->policydb;
security/selinux/ss/services.c
2796
sidtab = policy->sidtab;
security/selinux/ss/services.c
2889
static inline int __security_genfs_sid(struct selinux_policy *policy,
security/selinux/ss/services.c
2895
struct policydb *policydb = &policy->policydb;
security/selinux/ss/services.c
2896
struct sidtab *sidtab = policy->sidtab;
security/selinux/ss/services.c
2906
sclass = unmap_class(&policy->map, orig_sclass);
security/selinux/ss/services.c
2918
wildcard = ebitmap_get_bit(&policy->policydb.policycaps,
security/selinux/ss/services.c
2955
struct selinux_policy *policy;
security/selinux/ss/services.c
2965
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
2966
retval = __security_genfs_sid(policy, fstype, path,
security/selinux/ss/services.c
2973
int selinux_policy_genfs_sid(struct selinux_policy *policy,
security/selinux/ss/services.c
2980
return __security_genfs_sid(policy, fstype, path, orig_sclass, sid);
security/selinux/ss/services.c
2989
struct selinux_policy *policy;
security/selinux/ss/services.c
3005
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3006
policydb = &policy->policydb;
security/selinux/ss/services.c
3007
sidtab = policy->sidtab;
security/selinux/ss/services.c
3026
rc = __security_genfs_sid(policy, fstype, "/",
security/selinux/ss/services.c
3045
int security_get_bools(struct selinux_policy *policy,
security/selinux/ss/services.c
3052
policydb = &policy->policydb;
security/selinux/ss/services.c
3108
oldpolicy = rcu_dereference_protected(state->policy,
security/selinux/ss/services.c
3155
rcu_assign_pointer(state->policy, newpolicy);
security/selinux/ss/services.c
3172
struct selinux_policy *policy;
security/selinux/ss/services.c
3181
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3182
policydb = &policy->policydb;
security/selinux/ss/services.c
3230
struct selinux_policy *policy;
security/selinux/ss/services.c
3250
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3251
policydb = &policy->policydb;
security/selinux/ss/services.c
3252
sidtab = policy->sidtab;
security/selinux/ss/services.c
3341
struct selinux_policy *policy;
security/selinux/ss/services.c
3369
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3370
policydb = &policy->policydb;
security/selinux/ss/services.c
3371
sidtab = policy->sidtab;
security/selinux/ss/services.c
3425
int security_get_classes(struct selinux_policy *policy,
security/selinux/ss/services.c
3431
policydb = &policy->policydb;
security/selinux/ss/services.c
3466
int security_get_permissions(struct selinux_policy *policy,
security/selinux/ss/services.c
3474
policydb = &policy->policydb;
security/selinux/ss/services.c
3514
struct selinux_policy *policy;
security/selinux/ss/services.c
3521
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3522
value = policy->policydb.reject_unknown;
security/selinux/ss/services.c
3529
struct selinux_policy *policy;
security/selinux/ss/services.c
3536
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3537
value = policy->policydb.allow_unknown;
security/selinux/ss/services.c
3554
struct selinux_policy *policy;
security/selinux/ss/services.c
3561
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3562
rc = ebitmap_get_bit(&policy->policydb.policycaps, req_cap);
security/selinux/ss/services.c
3594
struct selinux_policy *policy;
security/selinux/ss/services.c
3638
policy = rcu_dereference(state->policy);
security/selinux/ss/services.c
3639
policydb = &policy->policydb;
security/selinux/ss/services.c
3640
tmprule->au_seqno = policy->latest_granting;
security/selinux/ss/services.c
3719
struct selinux_policy *policy;
security/selinux/ss/services.c
3735
policy = rcu_dereference(state->policy);
security/selinux/ss/services.c
3737
if (rule->au_seqno < policy->latest_granting) {
security/selinux/ss/services.c
3742
ctxt = sidtab_search(policy->sidtab, prop->selinux.secid);
security/selinux/ss/services.c
3880
struct selinux_policy *policy;
security/selinux/ss/services.c
3895
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3896
policydb = &policy->policydb;
security/selinux/ss/services.c
3897
sidtab = policy->sidtab;
security/selinux/ss/services.c
3955
struct selinux_policy *policy;
security/selinux/ss/services.c
3964
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
3965
policydb = &policy->policydb;
security/selinux/ss/services.c
3968
ctx = sidtab_search(policy->sidtab, sid);
security/selinux/ss/services.c
3995
static int __security_read_policy(struct selinux_policy *policy,
security/selinux/ss/services.c
4004
rc = policydb_write(&policy->policydb, &fp);
security/selinux/ss/services.c
4021
struct selinux_policy *policy;
security/selinux/ss/services.c
4023
policy = rcu_dereference_protected(
security/selinux/ss/services.c
4024
state->policy, lockdep_is_held(&state->policy_mutex));
security/selinux/ss/services.c
4025
if (!policy)
security/selinux/ss/services.c
4028
*len = policy->policydb.len;
security/selinux/ss/services.c
4033
return __security_read_policy(policy, *data, len);
security/selinux/ss/services.c
4051
struct selinux_policy *policy;
security/selinux/ss/services.c
4053
policy = rcu_dereference_protected(
security/selinux/ss/services.c
4054
state->policy, lockdep_is_held(&state->policy_mutex));
security/selinux/ss/services.c
4055
if (!policy)
security/selinux/ss/services.c
4058
*len = policy->policydb.len;
security/selinux/ss/services.c
4063
err = __security_read_policy(policy, *data, len);
security/selinux/ss/services.c
725
static int security_validtrans_handle_fail(struct selinux_policy *policy,
security/selinux/ss/services.c
731
struct policydb *p = &policy->policydb;
security/selinux/ss/services.c
732
struct sidtab *sidtab = policy->sidtab;
security/selinux/ss/services.c
759
struct selinux_policy *policy;
security/selinux/ss/services.c
776
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
777
policydb = &policy->policydb;
security/selinux/ss/services.c
778
sidtab = policy->sidtab;
security/selinux/ss/services.c
781
tclass = unmap_class(&policy->map, orig_tclass);
security/selinux/ss/services.c
823
rc = security_validtrans_handle_fail(policy,
security/selinux/ss/services.c
863
struct selinux_policy *policy;
security/selinux/ss/services.c
875
policy = rcu_dereference(selinux_state.policy);
security/selinux/ss/services.c
876
policydb = &policy->policydb;
security/selinux/ss/services.c
877
sidtab = policy->sidtab;
security/selinux/ss/services.c
943
static void avd_init(struct selinux_policy *policy, struct av_decision *avd)
security/selinux/ss/services.c
948
if (policy)
security/selinux/ss/services.c
949
avd->seqno = policy->latest_granting;
sound/soc/qcom/qdsp6/audioreach.c
1470
eos->policy = WR_SH_MEM_EP_EOS_POLICY_LAST;
sound/soc/qcom/qdsp6/audioreach.h
284
uint32_t policy;
tools/arch/x86/include/uapi/asm/kvm.h
771
__u32 policy;
tools/arch/x86/include/uapi/asm/kvm.h
807
__u32 policy;
tools/arch/x86/include/uapi/asm/kvm.h
826
__u32 policy;
tools/arch/x86/include/uapi/asm/kvm.h
856
__u32 policy;
tools/arch/x86/include/uapi/asm/kvm.h
878
__u64 policy;
tools/include/uapi/linux/fscrypt.h
86
} policy; /* output */
tools/lib/bpf/nlattr.c
105
int len, struct libbpf_nla_policy *policy)
tools/lib/bpf/nlattr.c
118
if (policy) {
tools/lib/bpf/nlattr.c
119
err = validate_nla(nla, maxtype, policy);
tools/lib/bpf/nlattr.c
150
struct libbpf_nla_policy *policy)
tools/lib/bpf/nlattr.c
153
libbpf_nla_len(nla), policy);
tools/lib/bpf/nlattr.c
46
struct libbpf_nla_policy *policy)
tools/lib/bpf/nlattr.c
55
pt = &policy[type];
tools/lib/bpf/nlattr.h
124
int len, struct libbpf_nla_policy *policy);
tools/lib/bpf/nlattr.h
127
struct libbpf_nla_policy *policy);
tools/net/ynl/lib/ynl-priv.h
130
const struct ynl_policy_nest *policy;
tools/net/ynl/lib/ynl.c
105
".%s", policy->table[type].name);
tools/net/ynl/lib/ynl.c
110
if (policy->table[type].is_submsg) {
tools/net/ynl/lib/ynl.c
118
policy->table[type].selector_type) {
tools/net/ynl/lib/ynl.c
134
const struct ynl_policy_nest *policy, char *str, int str_sz,
tools/net/ynl/lib/ynl.c
146
if (!policy) {
tools/net/ynl/lib/ynl.c
158
if (ynl_err_walk_is_sel(policy, attr))
tools/net/ynl/lib/ynl.c
174
if (ynl_err_walk_report_one(policy, selector, type, str, str_sz, &n))
tools/net/ynl/lib/ynl.c
177
next_pol = ynl_err_walk_sel_policy(&policy->table[type], selector);
tools/net/ynl/lib/ynl.c
366
const struct ynl_policy_attr *policy;
tools/net/ynl/lib/ynl.c
378
policy = &yarg->rsp_policy->table[type];
tools/net/ynl/lib/ynl.c
380
switch (policy->type) {
tools/net/ynl/lib/ynl.c
383
"Rejected attribute (%s)", policy->name);
tools/net/ynl/lib/ynl.c
391
"Invalid attribute (u8 %s)", policy->name);
tools/net/ynl/lib/ynl.c
397
"Invalid attribute (u16 %s)", policy->name);
tools/net/ynl/lib/ynl.c
403
"Invalid attribute (u32 %s)", policy->name);
tools/net/ynl/lib/ynl.c
409
"Invalid attribute (u64 %s)", policy->name);
tools/net/ynl/lib/ynl.c
415
"Invalid attribute (uint %s)", policy->name);
tools/net/ynl/lib/ynl.c
424
"Invalid attribute (nest %s)", policy->name);
tools/net/ynl/lib/ynl.c
427
if (!policy->len || len == policy->len)
tools/net/ynl/lib/ynl.c
430
"Invalid attribute (binary %s)", policy->name);
tools/net/ynl/lib/ynl.c
433
if (len && (!policy->len || len <= policy->len) && !data[len - 1])
tools/net/ynl/lib/ynl.c
436
"Invalid attribute (string %s)", policy->name);
tools/net/ynl/lib/ynl.c
442
"Invalid attribute (bitfield32 %s)", policy->name);
tools/net/ynl/lib/ynl.c
446
"Invalid attribute (unknown %s)", policy->name);
tools/net/ynl/lib/ynl.c
49
ynl_err_walk_is_sel(const struct ynl_policy_nest *policy,
tools/net/ynl/lib/ynl.c
54
return policy && type <= policy->max_attr &&
tools/net/ynl/lib/ynl.c
55
policy->table[type].is_selector;
tools/net/ynl/lib/ynl.c
62
const struct ynl_policy_nest *policy = policy_attr->nest;
tools/net/ynl/lib/ynl.c
67
return policy;
tools/net/ynl/lib/ynl.c
70
for (i = 0; i <= policy->max_attr; i++) {
tools/net/ynl/lib/ynl.c
71
if (!strcmp(sel, policy->table[i].name))
tools/net/ynl/lib/ynl.c
72
return policy->table[i].nest;
tools/net/ynl/lib/ynl.c
79
ynl_err_walk_report_one(const struct ynl_policy_nest *policy,
tools/net/ynl/lib/ynl.c
83
if (!policy) {
tools/net/ynl/lib/ynl.c
89
if (type > policy->max_attr) {
tools/net/ynl/lib/ynl.c
894
yarg.rsp_policy = info->policy;
tools/net/ynl/lib/ynl.c
95
if (!policy->table[type].name) {
tools/net/ynl/samples/rt-link.c
44
if (netkit->_present.policy)
tools/net/ynl/samples/rt-link.c
45
name = rt_link_netkit_policy_str(netkit->policy);
tools/perf/trace/beauty/sched_policy.c
26
int policy = arg->val,
tools/perf/trace/beauty/sched_policy.c
27
flags = policy & ~SCHED_POLICY_MASK;
tools/perf/trace/beauty/sched_policy.c
29
policy &= SCHED_POLICY_MASK;
tools/perf/trace/beauty/sched_policy.c
30
if (policy <= SCHED_DEADLINE)
tools/perf/trace/beauty/sched_policy.c
31
printed = scnprintf(bf, size, "%s%s", show_prefix ? prefix : "", policies[policy]);
tools/perf/trace/beauty/sched_policy.c
33
printed = scnprintf(bf, size, "%#x", policy);
tools/power/cpupower/lib/cpufreq.c
256
struct cpufreq_policy *policy;
tools/power/cpupower/lib/cpufreq.c
258
policy = malloc(sizeof(struct cpufreq_policy));
tools/power/cpupower/lib/cpufreq.c
259
if (!policy)
tools/power/cpupower/lib/cpufreq.c
262
policy->governor = sysfs_cpufreq_get_one_string(cpu, SCALING_GOVERNOR);
tools/power/cpupower/lib/cpufreq.c
263
if (!policy->governor) {
tools/power/cpupower/lib/cpufreq.c
264
free(policy);
tools/power/cpupower/lib/cpufreq.c
267
policy->min = sysfs_cpufreq_get_one_value(cpu, SCALING_MIN_FREQ);
tools/power/cpupower/lib/cpufreq.c
268
policy->max = sysfs_cpufreq_get_one_value(cpu, SCALING_MAX_FREQ);
tools/power/cpupower/lib/cpufreq.c
269
if ((!policy->min) || (!policy->max)) {
tools/power/cpupower/lib/cpufreq.c
270
free(policy->governor);
tools/power/cpupower/lib/cpufreq.c
271
free(policy);
tools/power/cpupower/lib/cpufreq.c
275
return policy;
tools/power/cpupower/lib/cpufreq.c
278
void cpufreq_put_policy(struct cpufreq_policy *policy)
tools/power/cpupower/lib/cpufreq.c
280
if ((!policy) || (!policy->governor))
tools/power/cpupower/lib/cpufreq.c
283
free(policy->governor);
tools/power/cpupower/lib/cpufreq.c
284
policy->governor = NULL;
tools/power/cpupower/lib/cpufreq.c
285
free(policy);
tools/power/cpupower/lib/cpufreq.c
616
int cpufreq_set_policy(unsigned int cpu, struct cpufreq_policy *policy)
tools/power/cpupower/lib/cpufreq.c
625
if (!policy || !(policy->governor))
tools/power/cpupower/lib/cpufreq.c
628
if (policy->max < policy->min)
tools/power/cpupower/lib/cpufreq.c
631
if (verify_gov(gov, policy->governor))
tools/power/cpupower/lib/cpufreq.c
634
snprintf(min, SYSFS_PATH_MAX, "%lu", policy->min);
tools/power/cpupower/lib/cpufreq.c
635
snprintf(max, SYSFS_PATH_MAX, "%lu", policy->max);
tools/power/cpupower/lib/cpufreq.c
638
write_max_first = (old_min && (policy->max < old_min) ? 0 : 1);
tools/power/cpupower/lib/cpufreq.h
110
void cpufreq_put_policy(struct cpufreq_policy *policy);
tools/power/cpupower/lib/cpufreq.h
191
int cpufreq_set_policy(unsigned int cpu, struct cpufreq_policy *policy);
tools/power/cpupower/utils/cpufreq-info.c
331
struct cpufreq_policy *policy = cpufreq_get_policy(cpu);
tools/power/cpupower/utils/cpufreq-info.c
332
if (!policy) {
tools/power/cpupower/utils/cpufreq-info.c
337
print_speed(policy->min, no_rounding);
tools/power/cpupower/utils/cpufreq-info.c
339
print_speed(policy->max, no_rounding);
tools/power/cpupower/utils/cpufreq-info.c
344
policy->governor);
tools/power/cpupower/utils/cpufreq-info.c
345
cpufreq_put_policy(policy);
tools/power/cpupower/utils/cpufreq-info.c
59
struct cpufreq_policy *policy;
tools/power/cpupower/utils/cpufreq-info.c
68
policy = cpufreq_get_policy(cpu);
tools/power/cpupower/utils/cpufreq-info.c
69
if (!policy)
tools/power/cpupower/utils/cpufreq-info.c
75
min_pctg = (policy->min * 100) / max;
tools/power/cpupower/utils/cpufreq-info.c
76
max_pctg = (policy->max * 100) / max;
tools/power/cpupower/utils/cpufreq-info.c
79
cpu , policy->min, max ? min_pctg : 0, policy->max,
tools/power/cpupower/utils/cpufreq-info.c
80
max ? max_pctg : 0, policy->governor);
tools/power/cpupower/utils/cpufreq-info.c
82
cpufreq_put_policy(policy);
tools/sched_ext/scx_qmap.bpf.c
549
switch (next->policy) {
tools/testing/selftests/bpf/prog_tests/tc_netkit.c
43
static int create_netkit(int mode, int policy, int peer_policy, int *ifindex,
tools/testing/selftests/bpf/prog_tests/tc_netkit.c
67
addattr32(&req.n, sizeof(req), IFLA_NETKIT_POLICY, policy);
tools/testing/selftests/futex/functional/futex_requeue_pi.c
122
int policy, int prio)
tools/testing/selftests/futex/functional/futex_requeue_pi.c
137
ret = pthread_attr_setschedpolicy(&attr, policy);
tools/testing/selftests/futex/functional/futex_requeue_pi_signal_restart.c
41
int policy, int prio)
tools/testing/selftests/futex/functional/futex_requeue_pi_signal_restart.c
54
ret = pthread_attr_setschedpolicy(&attr, policy);
tools/testing/selftests/kvm/guest_memfd_test.c
102
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem + page_size * 2, MPOL_F_ADDR);
tools/testing/selftests/kvm/guest_memfd_test.c
103
TEST_ASSERT(policy == MPOL_BIND && nodemask == nodemask_0,
tools/testing/selftests/kvm/guest_memfd_test.c
105
MPOL_BIND, nodemask_0, policy, nodemask);
tools/testing/selftests/kvm/guest_memfd_test.c
109
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR);
tools/testing/selftests/kvm/guest_memfd_test.c
110
TEST_ASSERT(policy == MPOL_DEFAULT && !nodemask,
tools/testing/selftests/kvm/guest_memfd_test.c
112
MPOL_DEFAULT, policy, nodemask);
tools/testing/selftests/kvm/guest_memfd_test.c
84
int policy;
tools/testing/selftests/kvm/guest_memfd_test.c
95
kvm_get_mempolicy(&policy, &nodemask, maxnode, mem, MPOL_F_ADDR);
tools/testing/selftests/kvm/guest_memfd_test.c
96
TEST_ASSERT(policy == MPOL_INTERLEAVE && nodemask == nodemask_0,
tools/testing/selftests/kvm/guest_memfd_test.c
98
MPOL_INTERLEAVE, nodemask_0, policy, nodemask);
tools/testing/selftests/kvm/include/numaif.h
13
KVM_SYSCALL_DEFINE(get_mempolicy, 5, int *, policy, const unsigned long *, nmask,
tools/testing/selftests/kvm/include/x86/sev.h
49
void sev_vm_launch(struct kvm_vm *vm, uint32_t policy);
tools/testing/selftests/kvm/include/x86/sev.h
52
void snp_vm_launch_start(struct kvm_vm *vm, uint64_t policy);
tools/testing/selftests/kvm/include/x86/sev.h
58
void vm_sev_launch(struct kvm_vm *vm, uint64_t policy, uint8_t *measurement);
tools/testing/selftests/kvm/lib/x86/sev.c
100
if (policy & SEV_POLICY_ES)
tools/testing/selftests/kvm/lib/x86/sev.c
134
void snp_vm_launch_start(struct kvm_vm *vm, uint64_t policy)
tools/testing/selftests/kvm/lib/x86/sev.c
137
.policy = policy,
tools/testing/selftests/kvm/lib/x86/sev.c
177
void vm_sev_launch(struct kvm_vm *vm, uint64_t policy, uint8_t *measurement)
tools/testing/selftests/kvm/lib/x86/sev.c
182
snp_vm_launch_start(vm, policy);
tools/testing/selftests/kvm/lib/x86/sev.c
191
sev_vm_launch(vm, policy);
tools/testing/selftests/kvm/lib/x86/sev.c
82
void sev_vm_launch(struct kvm_vm *vm, uint32_t policy)
tools/testing/selftests/kvm/lib/x86/sev.c
85
.policy = policy,
tools/testing/selftests/kvm/lib/x86/sev.c
94
TEST_ASSERT_EQ(status.policy, policy);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
107
static void test_sync_vmsa(uint32_t type, uint64_t policy)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
134
vm_sev_launch(vm, policy, NULL);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
153
static void test_sev(void *guest_code, uint32_t type, uint64_t policy)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
162
vm_sev_launch(vm, policy, NULL);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
204
static void test_sev_shutdown(uint32_t type, uint64_t policy)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
211
vm_sev_launch(vm, policy, NULL);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
221
static void test_sev_smoke(void *guest, uint32_t type, uint64_t policy)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
226
test_sev(guest, type, policy | SNP_POLICY_DBG);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
228
test_sev(guest, type, policy | SEV_POLICY_NO_DBG);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
229
test_sev(guest, type, policy);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
234
test_sev_shutdown(type, policy);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
238
test_sync_vmsa(type, policy);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
240
test_sync_vmsa(type, policy | SNP_POLICY_DBG);
tools/testing/selftests/kvm/x86/sev_smoke_test.c
242
test_sync_vmsa(type, policy | SEV_POLICY_NO_DBG);
tools/testing/selftests/net/ipsec.c
1395
req.acq.policy.sel.family = AF_INET;
tools/testing/selftests/net/nettest.c
1454
struct xfrm_userpolicy_info policy = {};
tools/testing/selftests/net/nettest.c
1464
policy.action = XFRM_POLICY_ALLOW;
tools/testing/selftests/net/nettest.c
1465
policy.sel.family = args->version;
tools/testing/selftests/net/nettest.c
1471
policy.dir = XFRM_POLICY_OUT;
tools/testing/selftests/net/nettest.c
1472
if (setsockopt(sd, level, xfrm_af, &policy, sizeof(policy)) < 0)
tools/testing/selftests/net/nettest.c
1475
policy.dir = XFRM_POLICY_IN;
tools/testing/selftests/net/nettest.c
1476
if (setsockopt(sd, level, xfrm_af, &policy, sizeof(policy)) < 0)
tools/testing/selftests/sched_ext/rt_stall.c
73
static void set_sched(int policy, int priority)
tools/testing/selftests/sched_ext/rt_stall.c
78
if (sched_setscheduler(0, policy, ¶m) != 0) {
tools/tracing/latency/latency-collector.c
1055
int policy;
tools/tracing/latency/latency-collector.c
1060
r = pthread_getschedparam(pthread_self(), &policy, ¶m);
tools/tracing/latency/latency-collector.c
1067
tid, policy_name(policy), param.sched_priority);
tools/tracing/latency/latency-collector.c
138
int policy;
tools/tracing/latency/latency-collector.c
142
static const struct policy policies[] = {
tools/tracing/latency/latency-collector.c
1851
const struct policy *p;
tools/tracing/latency/latency-collector.c
1902
sched_policy = p->policy;
tools/tracing/latency/latency-collector.c
880
static const struct policy *policy_from_name(const char *name)
tools/tracing/latency/latency-collector.c
882
const struct policy *p = &policies[0];
tools/tracing/latency/latency-collector.c
892
static const char *policy_name(int policy)
tools/tracing/latency/latency-collector.c
894
const struct policy *p = &policies[0];
tools/tracing/latency/latency-collector.c
898
if (p->policy == policy)
virt/kvm/guest_memfd.c
122
struct mempolicy *policy;
virt/kvm/guest_memfd.c
134
policy = mpol_shared_policy_lookup(&GMEM_I(inode)->policy, index);
virt/kvm/guest_memfd.c
137
mapping_gfp_mask(inode->i_mapping), policy);
virt/kvm/guest_memfd.c
138
mpol_cond_put(policy);
virt/kvm/guest_memfd.c
31
struct shared_policy policy;
virt/kvm/guest_memfd.c
437
return mpol_set_shared_policy(&GMEM_I(inode)->policy, vma, mpol);
virt/kvm/guest_memfd.c
456
return mpol_shared_policy_lookup(&GMEM_I(inode)->policy, *pgoff);
virt/kvm/guest_memfd.c
945
mpol_shared_policy_init(&gi->policy, NULL);
virt/kvm/guest_memfd.c
953
mpol_free_shared_policy(&GMEM_I(inode)->policy);