arch/arm/common/mcpm_entry.c
103
sync_cache_r(&c->cpus);
arch/arm/common/mcpm_entry.c
111
cpustate = c->cpus[i].cpu;
arch/arm/common/mcpm_entry.c
116
sync_cache_r(&c->cpus[i].cpu);
arch/arm/common/mcpm_entry.c
36
mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN;
arch/arm/common/mcpm_entry.c
37
sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
arch/arm/common/mcpm_entry.c
439
mcpm_sync.clusters[i].cpus[j].cpu = CPU_DOWN;
arch/arm/common/mcpm_entry.c
445
mcpm_sync.clusters[this_cluster].cpus[i].cpu = CPU_UP;
arch/arm/common/mcpm_entry.c
50
mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN;
arch/arm/common/mcpm_entry.c
51
sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
arch/arm/include/asm/mcpm.h
288
} cpus[MAX_CPUS_PER_CLUSTER];
arch/arm/kernel/devtree.c
148
set_smp_ops_by_method(cpus);
arch/arm/kernel/devtree.c
73
struct device_node *cpu, *cpus;
arch/arm/kernel/devtree.c
80
cpus = of_find_node_by_path("/cpus");
arch/arm/kernel/devtree.c
82
if (!cpus)
arch/arm/kernel/smp.c
798
struct cpumask *cpus = freq->policy->cpus;
arch/arm/kernel/smp.c
799
int cpu, first = cpumask_first(cpus);
arch/arm/kernel/smp.c
806
for_each_cpu(cpu, cpus) {
arch/arm/kernel/smp.c
826
for_each_cpu(cpu, cpus)
arch/arm/mach-s3c/cpu.h
69
struct cpu_table *cpus, unsigned int cputab_size);
arch/arm64/include/asm/cpufeature.h
404
const struct cpumask *cpus;
arch/arm64/kernel/cpufeature.c
1984
static const struct midr_range cpus[] = {
arch/arm64/kernel/cpufeature.c
1996
return is_midr_in_range_list(cpus);
arch/arm64/kernel/cpufeature.c
2707
.cpus = &amu_cpus,
arch/arm64/kernel/cpufeature.c
2752
.cpus = &dbm_cpus,
arch/arm64/kernel/cpufeature.c
3538
if (!match_all && caps->desc && !caps->cpus)
arch/arm64/kernel/cpufeature.c
3954
if (caps->cpus && cpumask_any(caps->cpus) < nr_cpu_ids)
arch/arm64/kernel/cpufeature.c
3956
caps->desc, cpumask_pr_args(caps->cpus));
arch/arm64/kernel/topology.c
221
for_each_cpu_wrap(ref_cpu, policy->cpus, cpu + 1) {
arch/arm64/kernel/topology.c
253
static void amu_fie_setup(const struct cpumask *cpus)
arch/arm64/kernel/topology.c
259
unlikely(cpumask_subset(cpus, amu_fie_cpus)))
arch/arm64/kernel/topology.c
262
for_each_cpu(cpu, cpus)
arch/arm64/kernel/topology.c
269
cpumask_pr_args(cpus));
arch/arm64/kernel/topology.c
273
cpumask_or(amu_fie_cpus, amu_fie_cpus, cpus);
arch/arm64/kernel/topology.c
275
topology_set_scale_freq_source(&amu_sfd, cpus);
arch/arm64/kernel/topology.c
278
cpumask_pr_args(cpus));
arch/arm64/kernel/topology.c
287
amu_fie_setup(policy->cpus);
arch/arm64/kernel/topology.c
326
if (unlikely(!cpumask_subset(policy->cpus, amu_fie_cpus)))
arch/arm64/kernel/watchdog_hld.c
79
for_each_cpu(cpu, policy->cpus)
arch/mips/cavium-octeon/smp.c
143
int cpus;
arch/mips/cavium-octeon/smp.c
162
cpus = 1;
arch/mips/cavium-octeon/smp.c
165
set_cpu_possible(cpus, true);
arch/mips/cavium-octeon/smp.c
166
set_cpu_present(cpus, true);
arch/mips/cavium-octeon/smp.c
167
__cpu_number_map[id] = cpus;
arch/mips/cavium-octeon/smp.c
168
__cpu_logical_map[cpus] = id;
arch/mips/cavium-octeon/smp.c
169
cpus++;
arch/mips/cavium-octeon/smp.c
182
set_cpu_possible(cpus, true);
arch/mips/cavium-octeon/smp.c
183
__cpu_number_map[id] = cpus;
arch/mips/cavium-octeon/smp.c
184
__cpu_logical_map[cpus] = id;
arch/mips/cavium-octeon/smp.c
185
cpus++;
arch/mips/kernel/cevt-r4k.c
258
for_each_cpu(cpu, freq->policy->cpus) {
arch/mips/kernel/sync-r4k.c
111
int cpus = 2;
arch/mips/kernel/sync-r4k.c
116
while (atomic_read(&start_count) != cpus - 1)
arch/mips/kernel/sync-r4k.c
126
while (atomic_read(&stop_count) != cpus-1)
arch/mips/kernel/sync-r4k.c
177
int cpus = 2;
arch/mips/kernel/sync-r4k.c
192
while (atomic_read(&start_count) != cpus)
arch/mips/kernel/sync-r4k.c
210
while (atomic_read(&stop_count) != cpus)
arch/mips/kernel/time.c
40
struct cpumask *cpus = freq->policy->cpus;
arch/mips/kernel/time.c
73
for_each_cpu(cpu, cpus) {
arch/powerpc/mm/book3s64/radix_tlb.c
371
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/book3s64/radix_tlb.c
374
on_each_cpu_mask(cpus, do_tlbiel_pid, &t, 1);
arch/powerpc/mm/book3s64/radix_tlb.c
510
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/book3s64/radix_tlb.c
512
on_each_cpu_mask(cpus, do_tlbiel_va, &t, 1);
arch/powerpc/mm/book3s64/radix_tlb.c
561
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/book3s64/radix_tlb.c
566
on_each_cpu_mask(cpus, do_tlbiel_va_range, &t, 1);
arch/powerpc/platforms/powermac/smp.c
630
struct device_node *cpus;
arch/powerpc/platforms/powermac/smp.c
633
cpus = of_find_node_by_path("/cpus");
arch/powerpc/platforms/powermac/smp.c
634
BUG_ON(cpus == NULL);
arch/powerpc/platforms/powermac/smp.c
637
pmf_call_function(cpus, "cpu-timebase", &args);
arch/powerpc/platforms/powermac/smp.c
638
of_node_put(cpus);
arch/powerpc/platforms/powermac/smp.c
708
struct device_node *cpus =
arch/powerpc/platforms/powermac/smp.c
710
if (cpus &&
arch/powerpc/platforms/powermac/smp.c
711
of_property_read_bool(cpus, "platform-cpu-timebase")) {
arch/powerpc/platforms/powermac/smp.c
716
of_node_put(cpus);
arch/powerpc/platforms/powermac/smp.c
765
struct device_node *cpus;
arch/powerpc/platforms/powermac/smp.c
771
for_each_node_by_type(cpus, "cpu")
arch/riscv/include/asm/vendor_extensions/mips_hwprobe.h
13
void hwprobe_isa_vendor_ext_mips_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/mips_hwprobe.h
16
const struct cpumask *cpus)
arch/riscv/include/asm/vendor_extensions/sifive_hwprobe.h
10
void hwprobe_isa_vendor_ext_sifive_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/sifive_hwprobe.h
13
const struct cpumask *cpus)
arch/riscv/include/asm/vendor_extensions/thead_hwprobe.h
10
void hwprobe_isa_vendor_ext_thead_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/thead_hwprobe.h
13
const struct cpumask *cpus)
arch/riscv/include/asm/vendor_extensions/vendor_hwprobe.h
26
#define VENDOR_EXTENSION_SUPPORTED(pair, cpus, per_hart_vendor_bitmap, _extension_checks) \
arch/riscv/include/asm/vendor_extensions/vendor_hwprobe.h
30
for_each_cpu(cpu, (cpus)) { \
arch/riscv/kernel/sys_hwprobe.c
101
for_each_cpu(cpu, cpus) {
arch/riscv/kernel/sys_hwprobe.c
187
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
198
for_each_cpu(cpu, cpus) {
arch/riscv/kernel/sys_hwprobe.c
214
static bool hwprobe_ext0_has(const struct cpumask *cpus, u64 ext)
arch/riscv/kernel/sys_hwprobe.c
218
hwprobe_isa_ext0(&pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
223
static u64 hwprobe_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
228
for_each_cpu(cpu, cpus) {
arch/riscv/kernel/sys_hwprobe.c
246
static u64 hwprobe_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
259
static u64 hwprobe_vec_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
265
for_each_cpu(cpu, cpus) {
arch/riscv/kernel/sys_hwprobe.c
283
static u64 hwprobe_vec_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
296
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
302
hwprobe_arch_id(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
315
hwprobe_isa_ext0(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
319
hwprobe_isa_ext1(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
324
pair->value = hwprobe_misaligned(cpus);
arch/riscv/kernel/sys_hwprobe.c
328
pair->value = hwprobe_vec_misaligned(cpus);
arch/riscv/kernel/sys_hwprobe.c
333
if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOZ))
arch/riscv/kernel/sys_hwprobe.c
338
if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOM))
arch/riscv/kernel/sys_hwprobe.c
343
if (hwprobe_ext0_has(cpus, RISCV_HWPROBE_EXT_ZICBOP))
arch/riscv/kernel/sys_hwprobe.c
355
hwprobe_isa_vendor_ext_sifive_0(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
359
hwprobe_isa_vendor_ext_thead_0(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
36
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
362
hwprobe_isa_vendor_ext_mips_0(pair, cpus);
arch/riscv/kernel/sys_hwprobe.c
384
cpumask_t cpus;
arch/riscv/kernel/sys_hwprobe.c
395
cpumask_clear(&cpus);
arch/riscv/kernel/sys_hwprobe.c
397
cpumask_copy(&cpus, cpu_online_mask);
arch/riscv/kernel/sys_hwprobe.c
402
ret = copy_from_user(&cpus, cpus_user, cpusetsize);
arch/riscv/kernel/sys_hwprobe.c
410
cpumask_and(&cpus, &cpus, cpu_online_mask);
arch/riscv/kernel/sys_hwprobe.c
411
if (cpumask_empty(&cpus))
arch/riscv/kernel/sys_hwprobe.c
422
hwprobe_one_pair(&pair, &cpus);
arch/riscv/kernel/sys_hwprobe.c
439
cpumask_t cpus, one_cpu;
arch/riscv/kernel/sys_hwprobe.c
453
ret = copy_from_user(&cpus, cpus_user, cpusetsize);
arch/riscv/kernel/sys_hwprobe.c
457
if (cpumask_empty(&cpus))
arch/riscv/kernel/sys_hwprobe.c
458
cpumask_copy(&cpus, cpu_online_mask);
arch/riscv/kernel/sys_hwprobe.c
460
cpumask_and(&cpus, &cpus, cpu_online_mask);
arch/riscv/kernel/sys_hwprobe.c
47
for_each_cpu(cpu, cpus) {
arch/riscv/kernel/sys_hwprobe.c
485
for_each_cpu(cpu, &cpus) {
arch/riscv/kernel/sys_hwprobe.c
491
cpumask_clear_cpu(cpu, &cpus);
arch/riscv/kernel/sys_hwprobe.c
498
cpumask_clear(&cpus);
arch/riscv/kernel/sys_hwprobe.c
500
ret = copy_to_user(cpus_user, &cpus, cpusetsize);
arch/riscv/kernel/sys_hwprobe.c
608
cpus, unsigned int, flags)
arch/riscv/kernel/sys_hwprobe.c
611
cpus, flags);
arch/riscv/kernel/sys_hwprobe.c
82
const struct cpumask *cpus)
arch/riscv/kernel/vdso/hwprobe.c
101
size_t cpusetsize, unsigned long *cpus,
arch/riscv/kernel/vdso/hwprobe.c
105
size_t cpusetsize, unsigned long *cpus,
arch/riscv/kernel/vdso/hwprobe.c
110
cpus, flags);
arch/riscv/kernel/vdso/hwprobe.c
113
cpus, flags);
arch/riscv/kernel/vdso/hwprobe.c
12
size_t cpusetsize, unsigned long *cpus,
arch/riscv/kernel/vdso/hwprobe.c
16
size_t cpusetsize, unsigned long *cpus,
arch/riscv/kernel/vdso/hwprobe.c
20
bool all_cpus = !cpusetsize && !cpus;
arch/riscv/kernel/vdso/hwprobe.c
31
return riscv_hwprobe(pairs, pair_count, cpusetsize, cpus, flags);
arch/riscv/kernel/vdso/hwprobe.c
50
size_t cpusetsize, unsigned long *cpus,
arch/riscv/kernel/vdso/hwprobe.c
56
unsigned char *c = (unsigned char *)cpus;
arch/riscv/kernel/vdso/hwprobe.c
61
if (!cpusetsize || !cpus)
arch/riscv/kernel/vdso/hwprobe.c
72
return riscv_hwprobe(pairs, pair_count, cpusetsize, cpus, flags);
arch/riscv/kernel/vendor_extensions/mips_hwprobe.c
18
const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/mips_hwprobe.c
20
VENDOR_EXTENSION_SUPPORTED(pair, cpus,
arch/riscv/kernel/vendor_extensions/sifive_hwprobe.c
13
void hwprobe_isa_vendor_ext_sifive_0(struct riscv_hwprobe *pair, const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/sifive_hwprobe.c
15
VENDOR_EXTENSION_SUPPORTED(pair, cpus,
arch/riscv/kernel/vendor_extensions/thead_hwprobe.c
13
void hwprobe_isa_vendor_ext_thead_0(struct riscv_hwprobe *pair, const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/thead_hwprobe.c
15
VENDOR_EXTENSION_SUPPORTED(pair, cpus,
arch/s390/hypfs/hypfs_diag_fs.c
158
return ((struct diag204_phys_hdr *)hdr)->cpus;
arch/s390/hypfs/hypfs_diag_fs.c
160
return ((struct diag204_x_phys_hdr *)hdr)->cpus;
arch/s390/hypfs/hypfs_diag_fs.c
76
return ((struct diag204_part_hdr *)hdr)->cpus;
arch/s390/include/asm/diag.h
160
__u8 cpus;
arch/s390/include/asm/diag.h
167
__u8 cpus;
arch/s390/include/asm/diag.h
219
__u8 cpus;
arch/s390/include/asm/diag.h
226
__u8 cpus;
arch/s390/include/asm/diag.h
253
struct diag204_x_cpu_info cpus[];
arch/s390/include/asm/diag.h
258
struct diag204_x_phys_cpu cpus[];
arch/s390/kernel/crash_dump.c
473
int cpus = 0;
arch/s390/kernel/crash_dump.c
477
cpus++;
arch/s390/kernel/crash_dump.c
478
return cpus;
arch/s390/kernel/hiperdispatch.c
210
int cpus, cpu;
arch/s390/kernel/hiperdispatch.c
213
cpus = 0;
arch/s390/kernel/hiperdispatch.c
218
cpus++;
arch/s390/kernel/hiperdispatch.c
224
if (cpus == 0)
arch/s390/kernel/hiperdispatch.c
230
percentage = steal_delta / cpus;
arch/s390/kernel/perf_cpum_cf.c
1477
unsigned int cpu, cpus, rc = 0;
arch/s390/kernel/perf_cpum_cf.c
1498
cpus = cpumask_weight(mask);
arch/s390/kernel/perf_cpum_cf.c
1499
if (put_user(cpus, &ctrset_read->no_cpus))
arch/s390/kernel/sthyi.c
237
for (i = 0; i < block->hdr.cpus; i++) {
arch/s390/kernel/sthyi.c
238
switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
arch/s390/kernel/sthyi.c
240
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
246
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
266
if (!(block->cpus[i].cflag & DIAG204_CPU_ONLINE))
arch/s390/kernel/sthyi.c
269
switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
arch/s390/kernel/sthyi.c
272
if (block->cpus[i].cur_weight < DED_WEIGHT)
arch/s390/kernel/sthyi.c
273
weight_cp |= block->cpus[i].cur_weight;
arch/s390/kernel/sthyi.c
277
if (block->cpus[i].cur_weight < DED_WEIGHT)
arch/s390/kernel/sthyi.c
278
weight_ifl |= block->cpus[i].cur_weight;
arch/s390/kernel/sthyi.c
287
capped |= block->cpus[i].cflag & DIAG204_CPU_CAPPED;
arch/s390/kernel/sthyi.c
288
cpu_inf->lpar_cap |= block->cpus[i].cpu_type_cap;
arch/s390/kernel/sthyi.c
289
cpu_inf->lpar_grp_cap |= block->cpus[i].group_cpu_type_cap;
arch/s390/kernel/sthyi.c
291
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
303
return (struct diag204_x_part_block *)&block->cpus[i];
arch/s390/kernel/time.c
391
atomic_t cpus;
arch/s390/kernel/time.c
525
while (atomic_read(&sync->cpus) != 0)
arch/s390/kernel/time.c
545
atomic_dec(&sync->cpus);
arch/s390/kernel/time.c
589
atomic_set(&stp_sync.cpus, num_online_cpus() - 1);
arch/s390/kvm/priv.c
824
int cpus = 0;
arch/s390/kvm/priv.c
827
cpus = atomic_read(&vcpu->kvm->online_vcpus);
arch/s390/kvm/priv.c
838
mem->vm[0].cpus_total = cpus;
arch/s390/kvm/priv.c
839
mem->vm[0].cpus_configured = cpus;
arch/sparc/kernel/time_64.c
661
for_each_cpu(cpu, freq->policy->cpus) {
arch/x86/events/intel/core.c
7284
static DEVICE_ATTR(cpus, S_IRUGO, intel_hybrid_get_attr_cpus, NULL);
arch/x86/hyperv/mmu.c
110
cpu = cpumask_last(cpus);
arch/x86/hyperv/mmu.c
115
for_each_cpu(cpu, cpus) {
arch/x86/hyperv/mmu.c
161
status = hyperv_flush_tlb_others_ex(cpus, info);
arch/x86/hyperv/mmu.c
169
native_flush_tlb_multi(cpus, info);
arch/x86/hyperv/mmu.c
172
static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
arch/x86/hyperv/mmu.c
19
static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
arch/x86/hyperv/mmu.c
200
nr_bank = cpumask_to_vpset_skip(&flush->hv_vp_set, cpus,
arch/x86/hyperv/mmu.c
59
static void hyperv_flush_tlb_multi(const struct cpumask *cpus,
arch/x86/hyperv/mmu.c
68
trace_hyperv_mmu_flush_tlb_multi(cpus, info);
arch/x86/hyperv/mmu.c
96
if (cpumask_equal(cpus, cpu_present_mask)) {
arch/x86/include/asm/paravirt_types.h
112
void (*flush_tlb_multi)(const struct cpumask *cpus,
arch/x86/include/asm/smp.h
116
void wbinvd_on_cpus_mask(struct cpumask *cpus);
arch/x86/include/asm/smp.h
118
void wbnoinvd_on_cpus_mask(struct cpumask *cpus);
arch/x86/include/asm/smp.h
159
static inline void wbinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/include/asm/smp.h
169
static inline void wbnoinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/include/asm/trace/hyperv.h
12
TP_PROTO(const struct cpumask *cpus,
arch/x86/include/asm/trace/hyperv.h
14
TP_ARGS(cpus, info),
arch/x86/include/asm/trace/hyperv.h
21
TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
arch/x86/include/asm/trace/hyperv.h
60
TP_PROTO(const struct cpumask *cpus,
arch/x86/include/asm/trace/hyperv.h
62
TP_ARGS(cpus, vector),
arch/x86/include/asm/trace/hyperv.h
67
TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
arch/x86/kernel/smpboot.c
538
const struct cpumask *cpus = cpumask_of_node(N + n);
arch/x86/kernel/smpboot.c
541
for_each_cpu(cpu, cpus) {
arch/x86/kernel/tsc_sync.c
358
int cpus = 2;
arch/x86/kernel/tsc_sync.c
371
while (atomic_read(&start_count) != cpus - 1)
arch/x86/kernel/tsc_sync.c
381
while (atomic_read(&stop_count) != cpus-1)
arch/x86/kernel/tsc_sync.c
437
int cpus = 2;
arch/x86/kernel/tsc_sync.c
464
while (atomic_read(&start_count) != cpus)
arch/x86/kernel/tsc_sync.c
482
while (atomic_read(&stop_count) != cpus)
arch/x86/kvm/x86.c
9886
for_each_cpu(cpu, freq->policy->cpus)
arch/x86/lib/cache-smp.c
23
void wbinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/lib/cache-smp.c
25
on_each_cpu_mask(cpus, __wbinvd, NULL, 1);
arch/x86/lib/cache-smp.c
40
void wbnoinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/lib/cache-smp.c
42
on_each_cpu_mask(cpus, __wbnoinvd, NULL, 1);
arch/x86/xen/mmu_pv.c
1329
static void xen_flush_tlb_multi(const struct cpumask *cpus,
arch/x86/xen/mmu_pv.c
1340
trace_xen_mmu_flush_tlb_multi(cpus, info->mm, info->start, info->end);
arch/x86/xen/mmu_pv.c
1342
if (cpumask_empty(cpus))
arch/x86/xen/mmu_pv.c
1350
cpumask_and(to_cpumask(args->mask), cpus, cpu_online_mask);
arch/x86/xen/smp_pv.c
410
unsigned int cpus;
arch/x86/xen/smp_pv.c
412
for (cpus = 0; cpus < nr_cpu_ids; cpus++) {
arch/x86/xen/smp_pv.c
413
if (HYPERVISOR_vcpu_op(VCPUOP_is_up, cpus, NULL) < 0)
arch/x86/xen/smp_pv.c
417
pr_info("Xen PV: Detected %u vCPUS\n", cpus);
arch/x86/xen/smp_pv.c
418
if (cpus < nr_cpu_ids)
arch/x86/xen/smp_pv.c
419
set_nr_cpu_ids(cpus);
drivers/acpi/pptt.c
1016
int acpi_pptt_get_cpumask_from_cache_id(u32 cache_id, cpumask_t *cpus)
drivers/acpi/pptt.c
1021
cpumask_clear(cpus);
drivers/acpi/pptt.c
1058
cpumask_set_cpu(cpu, cpus);
drivers/acpi/pptt.c
866
cpumask_t *cpus)
drivers/acpi/pptt.c
872
cpumask_clear(cpus);
drivers/acpi/pptt.c
880
cpumask_set_cpu(cpu, cpus);
drivers/acpi/pptt.c
902
void acpi_pptt_get_cpus_from_container(u32 acpi_cpu_id, cpumask_t *cpus)
drivers/acpi/pptt.c
909
cpumask_clear(cpus);
drivers/acpi/pptt.c
927
acpi_pptt_get_child_cpus(table_hdr, cpu_node, cpus);
drivers/base/arch_topology.c
102
const struct cpumask *cpus)
drivers/base/arch_topology.c
109
for_each_cpu(cpu, cpus) {
drivers/base/arch_topology.c
141
void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq,
drivers/base/arch_topology.c
155
if (supports_scale_freq_counters(cpus))
drivers/base/arch_topology.c
160
for_each_cpu(i, cpus)
drivers/base/arch_topology.c
180
void topology_update_hw_pressure(const struct cpumask *cpus,
drivers/base/arch_topology.c
187
cpu = cpumask_first(cpus);
drivers/base/arch_topology.c
204
for_each_cpu(cpu, cpus)
drivers/base/arch_topology.c
35
static bool supports_scale_freq_counters(const struct cpumask *cpus)
drivers/base/arch_topology.c
39
for_each_cpu(i, cpus) {
drivers/base/arch_topology.c
71
const struct cpumask *cpus)
drivers/base/arch_topology.c
85
for_each_cpu(cpu, cpus) {
drivers/clk/sunxi/clk-sun9i-cpus.c
152
struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw);
drivers/clk/sunxi/clk-sun9i-cpus.c
159
reg = readl(cpus->reg);
drivers/clk/sunxi/clk-sun9i-cpus.c
167
writel(reg, cpus->reg);
drivers/clk/sunxi/clk-sun9i-cpus.c
189
struct sun9i_a80_cpus_clk *cpus;
drivers/clk/sunxi/clk-sun9i-cpus.c
194
cpus = kzalloc_obj(*cpus);
drivers/clk/sunxi/clk-sun9i-cpus.c
195
if (!cpus)
drivers/clk/sunxi/clk-sun9i-cpus.c
198
cpus->reg = of_io_request_and_map(node, 0, of_node_full_name(node));
drivers/clk/sunxi/clk-sun9i-cpus.c
199
if (IS_ERR(cpus->reg))
drivers/clk/sunxi/clk-sun9i-cpus.c
212
mux->reg = cpus->reg;
drivers/clk/sunxi/clk-sun9i-cpus.c
220
&cpus->hw, &sun9i_a80_cpus_clk_ops,
drivers/clk/sunxi/clk-sun9i-cpus.c
236
iounmap(cpus->reg);
drivers/clk/sunxi/clk-sun9i-cpus.c
240
kfree(cpus);
drivers/clk/sunxi/clk-sun9i-cpus.c
52
struct sun9i_a80_cpus_clk *cpus = to_sun9i_a80_cpus_clk(hw);
drivers/clk/sunxi/clk-sun9i-cpus.c
57
reg = readl(cpus->reg);
drivers/cpufreq/acpi-cpufreq.c
133
on_each_cpu_mask(policy->cpus, boost_set_msr_each,
drivers/cpufreq/acpi-cpufreq.c
136
cpumask_pr_args(policy->cpus), str_enabled_disabled(val));
drivers/cpufreq/acpi-cpufreq.c
435
cpumask_of(policy->cpu) : policy->cpus;
drivers/cpufreq/acpi-cpufreq.c
731
cpumask_copy(policy->cpus, perf->shared_cpu_map);
drivers/cpufreq/acpi-cpufreq.c
739
cpumask_copy(policy->cpus, topology_core_cpumask(cpu));
drivers/cpufreq/acpi-cpufreq.c
744
cpumask_clear(policy->cpus);
drivers/cpufreq/acpi-cpufreq.c
745
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/apple-soc-cpufreq.c
220
policy->cpus, &args);
drivers/cpufreq/apple-soc-cpufreq.c
266
ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/armada-8k-cpufreq.c
135
static struct cpumask cpus, shared_cpus;
drivers/cpufreq/armada-8k-cpufreq.c
149
cpumask_copy(&cpus, cpu_possible_mask);
drivers/cpufreq/armada-8k-cpufreq.c
156
for_each_cpu(cpu, &cpus) {
drivers/cpufreq/armada-8k-cpufreq.c
185
cpumask_andnot(&cpus, &cpus, &shared_cpus);
drivers/cpufreq/brcmstb-avs-cpufreq.c
640
cpumask_setall(policy->cpus);
drivers/cpufreq/cppc_cpufreq.c
156
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/cppc_cpufreq.c
180
topology_set_scale_freq_source(sftd, policy->cpus);
drivers/cpufreq/cppc_cpufreq.c
661
cpumask_copy(policy->cpus, cpu_data->shared_cpu_map);
drivers/cpufreq/cpufreq-dt.c
109
cpumask_copy(policy->cpus, priv->cpus);
drivers/cpufreq/cpufreq-dt.c
176
if (!zalloc_cpumask_var(&priv->cpus, GFP_KERNEL))
drivers/cpufreq/cpufreq-dt.c
179
cpumask_set_cpu(cpu, priv->cpus);
drivers/cpufreq/cpufreq-dt.c
197
ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, priv->cpus);
drivers/cpufreq/cpufreq-dt.c
207
if (dev_pm_opp_get_sharing_cpus(cpu_dev, priv->cpus))
drivers/cpufreq/cpufreq-dt.c
222
ret = dev_pm_opp_of_cpumask_add_table(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
241
cpumask_setall(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
242
ret = dev_pm_opp_set_sharing_cpus(cpu_dev, priv->cpus);
drivers/cpufreq/cpufreq-dt.c
259
dev_pm_opp_of_cpumask_remove_table(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
262
free_cpumask_var(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
273
dev_pm_opp_of_cpumask_remove_table(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
275
free_cpumask_var(priv->cpus);
drivers/cpufreq/cpufreq-dt.c
30
cpumask_var_t cpus;
drivers/cpufreq/cpufreq-dt.c
44
if (cpumask_test_cpu(cpu, priv->cpus))
drivers/cpufreq/cpufreq.c
1176
if (cpumask_test_cpu(cpu, policy->cpus))
drivers/cpufreq/cpufreq.c
1184
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/cpufreq.c
1270
if (!alloc_cpumask_var(&policy->cpus, GFP_KERNEL))
drivers/cpufreq/cpufreq.c
1333
free_cpumask_var(policy->cpus);
drivers/cpufreq/cpufreq.c
1385
free_cpumask_var(policy->cpus);
drivers/cpufreq/cpufreq.c
1403
cpumask_copy(policy->cpus, policy->related_cpus);
drivers/cpufreq/cpufreq.c
1412
cpumask_copy(policy->cpus, cpumask_of(cpu));
drivers/cpufreq/cpufreq.c
1435
cpumask_copy(policy->related_cpus, policy->cpus);
drivers/cpufreq/cpufreq.c
1442
cpumask_and(policy->cpus, policy->cpus, cpu_online_mask);
drivers/cpufreq/cpufreq.c
1588
cpumask_clear(policy->cpus);
drivers/cpufreq/cpufreq.c
1687
cpumask_clear_cpu(cpu, policy->cpus);
drivers/cpufreq/cpufreq.c
1692
policy->cpu = cpumask_any(policy->cpus);
drivers/cpufreq/cpufreq.c
189
cpumask_setall(policy->cpus);
drivers/cpufreq/cpufreq.c
197
return policy && cpumask_test_cpu(cpu, policy->cpus) ? policy : NULL;
drivers/cpufreq/cpufreq.c
2221
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/cpufreq.c
352
cpumask_pr_args(policy->cpus));
drivers/cpufreq/cpufreq.c
354
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/cpufreq.c
916
return cpufreq_show_cpus(policy->cpus, buf);
drivers/cpufreq/cpufreq_governor.c
102
for_each_cpu(j, policy_dbs->policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
137
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
332
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/cpufreq_governor.c
344
for_each_cpu(i, policy->cpus)
drivers/cpufreq/cpufreq_governor.c
532
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/cpufreq_ondemand.c
417
cpumask_or(done, done, policy->cpus);
drivers/cpufreq/loongson3_cpufreq.c
295
cpumask_copy(policy->cpus, topology_sibling_cpumask(cpu));
drivers/cpufreq/loongson3_cpufreq.c
297
for_each_cpu(i, policy->cpus) {
drivers/cpufreq/mediatek-cpufreq-hw.c
238
policy->cpus, &args);
drivers/cpufreq/mediatek-cpufreq-hw.c
352
&em_cb, policy->cpus, true);
drivers/cpufreq/mediatek-cpufreq.c
40
struct cpumask cpus;
drivers/cpufreq/mediatek-cpufreq.c
454
ret = dev_pm_opp_of_get_sharing_cpus(cpu_dev, &info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
461
ret = dev_pm_opp_of_cpumask_add_table(&info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
534
dev_pm_opp_of_cpumask_remove_table(&info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
575
dev_pm_opp_of_cpumask_remove_table(&info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
602
cpumask_copy(policy->cpus, &info->cpus);
drivers/cpufreq/mediatek-cpufreq.c
71
if (cpumask_test_cpu(cpu, &info->cpus))
drivers/cpufreq/p4-clockmod.c
111
for_each_cpu(i, policy->cpus)
drivers/cpufreq/p4-clockmod.c
162
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
drivers/cpufreq/powernow-k8.c
1071
cpumask_copy(pol->cpus, topology_core_cpumask(pol->cpu));
drivers/cpufreq/powernow-k8.c
1072
data->available_cores = pol->cpus;
drivers/cpufreq/powernow-k8.c
1079
for_each_cpu(cpu, pol->cpus)
drivers/cpufreq/powernv-cpufreq.c
687
if (!cpumask_test_cpu(raw_smp_processor_id(), policy->cpus)) {
drivers/cpufreq/powernv-cpufreq.c
689
add_timer_on(&gpstates->timer, cpumask_first(policy->cpus));
drivers/cpufreq/powernv-cpufreq.c
824
smp_call_function_any(policy->cpus, set_pstate, &freq_data, 1);
drivers/cpufreq/powernv-cpufreq.c
837
cpumask_set_cpu(base + i, policy->cpus);
drivers/cpufreq/powernv-cpufreq.c
937
cpumask_andnot(&mask, &mask, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
298
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
457
ret = irq_set_affinity_and_hint(data->throttle_irq, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
478
ret = irq_set_affinity_and_hint(data->throttle_irq, policy->cpus);
drivers/cpufreq/qcom-cpufreq-hw.c
551
qcom_get_related_cpus(index, policy->cpus);
drivers/cpufreq/qcom-cpufreq-nvmem.c
468
struct dev_pm_domain_list *pd_list = drv->cpus[cpu].pd_list;
drivers/cpufreq/qcom-cpufreq-nvmem.c
503
drv = devm_kzalloc(&pdev->dev, struct_size(drv, cpus, num_possible_cpus()),
drivers/cpufreq/qcom-cpufreq-nvmem.c
548
drv->cpus[cpu].opp_token = dev_pm_opp_set_config(cpu_dev, &config);
drivers/cpufreq/qcom-cpufreq-nvmem.c
549
if (drv->cpus[cpu].opp_token < 0) {
drivers/cpufreq/qcom-cpufreq-nvmem.c
550
ret = drv->cpus[cpu].opp_token;
drivers/cpufreq/qcom-cpufreq-nvmem.c
565
&drv->cpus[cpu].pd_list);
drivers/cpufreq/qcom-cpufreq-nvmem.c
583
dev_pm_domain_detach_list(drv->cpus[cpu].pd_list);
drivers/cpufreq/qcom-cpufreq-nvmem.c
584
dev_pm_opp_clear_config(drv->cpus[cpu].opp_token);
drivers/cpufreq/qcom-cpufreq-nvmem.c
597
dev_pm_domain_detach_list(drv->cpus[cpu].pd_list);
drivers/cpufreq/qcom-cpufreq-nvmem.c
598
dev_pm_opp_clear_config(drv->cpus[cpu].opp_token);
drivers/cpufreq/qcom-cpufreq-nvmem.c
67
struct qcom_cpufreq_drv_cpu cpus[];
drivers/cpufreq/qoriq-cpufreq.c
88
struct cpumask *dstp = policy->cpus;
drivers/cpufreq/scmi-cpufreq.c
227
ret = scmi_get_sharing_cpus(cpu_dev, domain, policy->cpus);
drivers/cpufreq/scmi-cpufreq.c
245
cpumask_copy(priv->opp_shared_cpus, policy->cpus);
drivers/cpufreq/scpi-cpufreq.c
111
ret = scpi_get_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/scpi-cpufreq.c
117
ret = dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/speedstep-centrino.c
436
for_each_cpu(j, policy->cpus) {
drivers/cpufreq/speedstep-centrino.c
444
good_cpu = cpumask_any_and(policy->cpus,
drivers/cpufreq/speedstep-ich.c
263
policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
drivers/cpufreq/speedstep-ich.c
296
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
drivers/cpufreq/speedstep-ich.c
298
policy_cpu = cpumask_any_and(policy->cpus, cpu_online_mask);
drivers/cpufreq/tegra186-cpufreq.c
166
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/tegra186-cpufreq.c
177
unsigned int cluster = data->cpus[policy->cpu].bpmp_cluster_id;
drivers/cpufreq/tegra186-cpufreq.c
188
if (data->cpus[cpu].bpmp_cluster_id == cluster)
drivers/cpufreq/tegra186-cpufreq.c
189
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/tegra186-cpufreq.c
218
for_each_cpu(cpu, policy->cpus) {
drivers/cpufreq/tegra186-cpufreq.c
219
edvd_offset = data->cpus[cpu].edvd_offset;
drivers/cpufreq/tegra186-cpufreq.c
241
edvd_offset = data->cpus[policy->cpu].edvd_offset;
drivers/cpufreq/tegra186-cpufreq.c
243
cluster_id = data->cpus[policy->cpu].bpmp_cluster_id;
drivers/cpufreq/tegra186-cpufreq.c
365
data->cpus = tegra186_cpus;
drivers/cpufreq/tegra186-cpufreq.c
390
if (data->cpus[cpu].bpmp_cluster_id == i) {
drivers/cpufreq/tegra186-cpufreq.c
392
edvd_offset = data->cpus[cpu].edvd_offset;
drivers/cpufreq/tegra186-cpufreq.c
69
const struct tegra186_cpufreq_cpu *cpus;
drivers/cpufreq/tegra194-cpufreq.c
138
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/tegra194-cpufreq.c
382
on_each_cpu_mask(policy->cpus, tegra194_set_cpu_ndiv_sysreg, &ndiv, true);
drivers/cpufreq/tegra194-cpufreq.c
495
dev_pm_opp_set_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/tegra194-cpufreq.c
517
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/vexpress-spc-cpufreq.c
426
dev_pm_opp_get_sharing_cpus(cpu_dev, policy->cpus);
drivers/cpufreq/vexpress-spc-cpufreq.c
428
for_each_cpu(cpu, policy->cpus)
drivers/cpufreq/vexpress-spc-cpufreq.c
435
ret = get_cluster_clk_and_freq_table(cpu_dev, policy->cpus);
drivers/cpufreq/virtual-cpufreq.c
150
cpumask_set_cpu(cpu, policy->cpus);
drivers/cpufreq/virtual-cpufreq.c
227
topology_set_scale_freq_source(&virt_sfd, policy->cpus);
drivers/crypto/caam/caamalg_qi2.c
5002
static void free_dpaa2_pcpu_netdev(struct dpaa2_caam_priv *priv, const cpumask_t *cpus)
drivers/crypto/caam/caamalg_qi2.c
5007
for_each_cpu(i, cpus) {
drivers/crypto/caam/qi.c
410
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
442
if (!cpumask_test_cpu(*cpu, cpus)) {
drivers/crypto/caam/qi.c
445
*pcpu = cpumask_next_wrap(*pcpu, cpus);
drivers/crypto/caam/qi.c
519
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
521
for_each_cpu(i, cpus) {
drivers/crypto/caam/qi.c
696
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
699
for_each_cpu(i, cpus) {
drivers/crypto/caam/qi.c
713
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
715
for_each_cpu(i, cpus)
drivers/crypto/caam/qi.c
719
static void free_caam_qi_pcpu_netdev(const cpumask_t *cpus)
drivers/crypto/caam/qi.c
724
for_each_cpu(i, cpus) {
drivers/crypto/caam/qi.c
735
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
762
for_each_cpu(i, cpus) {
drivers/crypto/intel/qat/qat_common/adf_gen2_config.c
116
int cpus = num_online_cpus();
drivers/crypto/intel/qat/qat_common/adf_gen2_config.c
123
instances = min(cpus, banks);
drivers/crypto/intel/qat/qat_common/adf_gen2_config.c
17
int cpus = num_online_cpus();
drivers/crypto/intel/qat/qat_common/adf_gen2_config.c
24
instances = min(cpus, banks);
drivers/crypto/intel/qat/qat_common/adf_gen4_config.c
124
int cpus = num_online_cpus();
drivers/crypto/intel/qat/qat_common/adf_gen4_config.c
131
instances = min(cpus, banks);
drivers/crypto/intel/qat/qat_common/adf_gen4_config.c
18
int cpus = num_online_cpus();
drivers/crypto/intel/qat/qat_common/adf_gen4_config.c
25
instances = min(cpus, banks / 2);
drivers/crypto/intel/qat/qat_common/adf_isr.c
216
unsigned int cpu, cpus = num_online_cpus();
drivers/crypto/intel/qat/qat_common/adf_isr.c
239
i) % cpus;
drivers/firmware/psci/psci_checker.c
77
static unsigned int down_and_up_cpus(const struct cpumask *cpus,
drivers/firmware/psci/psci_checker.c
86
for_each_cpu(cpu, cpus) {
drivers/gpu/drm/vc4/vc4_hdmi.c
2394
dai_link->cpus = &vc4_hdmi->audio.cpu;
drivers/gpu/drm/vc4/vc4_hdmi.c
2405
dai_link->cpus->dai_name = dev_name(dev);
drivers/gpu/drm/xlnx/zynqmp_dp_audio.c
401
link->cpus = &audio->components[i].cpu;
drivers/gpu/drm/xlnx/zynqmp_dp_audio.c
403
link->cpus[0].dai_name = audio->dai_name;
drivers/irqchip/irq-bcm6345-l1.c
149
intc->cpus[cpu_idx]->enable_cache[word] |= mask;
drivers/irqchip/irq-bcm6345-l1.c
150
__raw_writel(intc->cpus[cpu_idx]->enable_cache[word],
drivers/irqchip/irq-bcm6345-l1.c
151
intc->cpus[cpu_idx]->map_base + reg_enable(intc, word));
drivers/irqchip/irq-bcm6345-l1.c
161
intc->cpus[cpu_idx]->enable_cache[word] &= ~mask;
drivers/irqchip/irq-bcm6345-l1.c
162
__raw_writel(intc->cpus[cpu_idx]->enable_cache[word],
drivers/irqchip/irq-bcm6345-l1.c
163
intc->cpus[cpu_idx]->map_base + reg_enable(intc, word));
drivers/irqchip/irq-bcm6345-l1.c
206
enabled = intc->cpus[old_cpu]->enable_cache[word] & mask;
drivers/irqchip/irq-bcm6345-l1.c
241
cpu = intc->cpus[idx] = kzalloc_flex(*cpu, enable_cache, n_words);
drivers/irqchip/irq-bcm6345-l1.c
329
struct bcm6345_l1_cpu *cpu = intc->cpus[idx];
drivers/irqchip/irq-bcm6345-l1.c
338
struct bcm6345_l1_cpu *cpu = intc->cpus[idx];
drivers/irqchip/irq-bcm6345-l1.c
80
struct bcm6345_l1_cpu *cpus[NR_CPUS];
drivers/irqchip/irq-bcm7038-l1.c
121
cpu = intc->cpus[cpu_logical_map(smp_processor_id())];
drivers/irqchip/irq-bcm7038-l1.c
123
cpu = intc->cpus[0];
drivers/irqchip/irq-bcm7038-l1.c
151
intc->cpus[cpu_idx]->mask_cache[word] &= ~mask;
drivers/irqchip/irq-bcm7038-l1.c
152
l1_writel(mask, intc->cpus[cpu_idx]->map_base +
drivers/irqchip/irq-bcm7038-l1.c
162
intc->cpus[cpu_idx]->mask_cache[word] |= mask;
drivers/irqchip/irq-bcm7038-l1.c
163
l1_writel(mask, intc->cpus[cpu_idx]->map_base +
drivers/irqchip/irq-bcm7038-l1.c
202
was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] &
drivers/irqchip/irq-bcm7038-l1.c
245
cpu = intc->cpus[idx] = kzalloc_flex(*cpu, mask_cache, n_words);
drivers/irqchip/irq-bcm7038-l1.c
304
intc->cpus[boot_cpu]->map_base + reg_mask_set(intc, word));
drivers/irqchip/irq-bcm7038-l1.c
306
intc->cpus[boot_cpu]->map_base + reg_mask_clr(intc, word));
drivers/irqchip/irq-bcm7038-l1.c
326
l1_writel(intc->cpus[boot_cpu]->mask_cache[word],
drivers/irqchip/irq-bcm7038-l1.c
327
intc->cpus[boot_cpu]->map_base + reg_mask_set(intc, word));
drivers/irqchip/irq-bcm7038-l1.c
328
l1_writel(~intc->cpus[boot_cpu]->mask_cache[word],
drivers/irqchip/irq-bcm7038-l1.c
329
intc->cpus[boot_cpu]->map_base + reg_mask_clr(intc, word));
drivers/irqchip/irq-bcm7038-l1.c
41
struct bcm7038_l1_cpu *cpus[NR_CPUS];
drivers/irqchip/irq-bcm7038-l1.c
440
struct bcm7038_l1_cpu *cpu = intc->cpus[idx];
drivers/leds/trigger/ledtrig-activity.c
135
target = (cpus > 1) ? (100 / cpus) : 50;
drivers/leds/trigger/ledtrig-activity.c
45
int cpus;
drivers/leds/trigger/ledtrig-activity.c
57
cpus = 0;
drivers/leds/trigger/ledtrig-activity.c
70
cpus++;
drivers/leds/trigger/ledtrig-activity.c
78
curr_boot = ktime_get_boottime_ns() * cpus;
drivers/misc/sgi-gru/grufile.c
169
info.cpus = num_online_cpus();
drivers/misc/sgi-gru/grulib.h
146
int cpus;
drivers/misc/sgi-gru/grutlbpurge.c
295
int cpus, shift = 0, n;
drivers/misc/sgi-gru/grutlbpurge.c
297
cpus = uv_blade_nr_possible_cpus(gru->gs_blade_id);
drivers/misc/sgi-gru/grutlbpurge.c
300
if (cpus) {
drivers/misc/sgi-gru/grutlbpurge.c
301
n = 1 << fls(cpus - 1);
drivers/misc/sgi-gru/grutlbpurge.c
314
gru->gs_tgh_first_remote = (cpus + (1 << shift) - 1) >> shift;
drivers/net/ethernet/broadcom/bnx2.c
6296
int cpus = netif_get_num_default_rss_queues();
drivers/net/ethernet/broadcom/bnx2.c
6300
msix_vecs = max(cpus + 1, bp->num_req_tx_rings);
drivers/net/ethernet/broadcom/bnx2.c
6302
msix_vecs = max(cpus, bp->num_req_rx_rings);
drivers/net/ethernet/chelsio/cxgb3/common.h
698
const u8 * cpus, const u16 *rspq);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
572
u8 cpus[SGE_QSETS + 1];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
576
cpus[i] = i;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
577
cpus[SGE_QSETS] = 0xff; /* terminator */
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
587
V_RRCPLCPUSIZE(6) | F_HASHTOEPLITZ, cpus, rspq_map);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2471
const u8 * cpus, const u16 *rspq)
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2475
if (cpus)
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2480
val |= (cpus[cpu_idx++] & 0x3f) << (8 * j);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2481
if (cpus[cpu_idx] == 0xff)
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
815
const cpumask_t *cpus = qman_affine_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
819
for_each_cpu_and(cpu, cpus, cpu_online_mask) {
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
417
const cpumask_t *cpus = qman_affine_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
437
for_each_cpu_and(cpu, cpus, cpu_online_mask) {
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
456
for_each_cpu_and(cpu, cpus, cpu_online_mask) {
drivers/net/ethernet/freescale/enetc/enetc.c
2544
int cpus = num_online_cpus();
drivers/net/ethernet/freescale/enetc/enetc.c
2553
priv->num_rx_rings = min_t(int, cpus, si->num_rx_rings);
drivers/net/ethernet/intel/i40e/i40e_main.c
11809
int cpus, extra_vectors;
drivers/net/ethernet/intel/i40e/i40e_main.c
11849
cpus = num_online_cpus();
drivers/net/ethernet/intel/i40e/i40e_main.c
11850
pf->num_lan_msix = min_t(int, cpus, vectors_left / 2);
drivers/net/ethernet/intel/i40e/i40e_main.c
11916
extra_vectors = min_t(int, cpus - pf->num_lan_msix, vectors_left);
drivers/net/ethernet/intel/ice/ice_main.c
2940
u16 cpus = num_possible_cpus();
drivers/net/ethernet/intel/ice/ice_main.c
2942
if (avail < cpus / 2)
drivers/net/ethernet/intel/ice/ice_main.c
2948
vsi->num_xdp_txq = min_t(u16, avail, cpus);
drivers/net/ethernet/intel/ice/ice_main.c
2950
if (vsi->num_xdp_txq < cpus)
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1458
int nrxqs, cpu, cpus = num_possible_cpus();
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1461
nrxqs = port->nrxqs / cpus;
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1472
return port->first_rxq + ((rxq * nrxqs + rxq / cpus) % port->nrxqs);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1588
cpumask_var_t cpus __free(free_cpumask_var);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1591
if (!alloc_cpumask_var(&cpus, GFP_KERNEL))
drivers/net/ethernet/microsoft/mana/gdma_main.c
1598
cpumask_andnot(cpus, next, prev);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1599
for_each_cpu(cpu, cpus) {
drivers/net/ethernet/microsoft/mana/gdma_main.c
1600
cpumask_andnot(cpus, cpus, topology_sibling_cpumask(cpu));
drivers/opp/of.c
1542
int dev_pm_opp_of_register_em(struct device *dev, struct cpumask *cpus)
drivers/opp/of.c
1587
ret = em_dev_register_perf_domain(dev, nr_opp, &em_cb, cpus, true);
drivers/perf/arm_pmu.c
576
static DEVICE_ATTR_RO(cpus);
drivers/pmdomain/core.c
1891
cpumask_set_cpu(cpu, genpd->cpus);
drivers/pmdomain/core.c
1893
cpumask_clear_cpu(cpu, genpd->cpus);
drivers/pmdomain/core.c
2294
!zalloc_cpumask_var(&genpd->cpus, GFP_KERNEL))
drivers/pmdomain/core.c
2343
free_cpumask_var(genpd->cpus);
drivers/pmdomain/core.c
2354
free_cpumask_var(genpd->cpus);
drivers/pmdomain/governor.c
376
for_each_cpu_and(cpu, genpd->cpus, cpu_online_mask) {
drivers/pmdomain/governor.c
419
if (cpus_peek_for_pending_ipi(genpd->cpus))
drivers/powercap/dtpm_cpu.c
128
nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(em->cpus));
drivers/powercap/dtpm_cpu.c
50
nr_cpus = cpumask_weight_and(cpu_online_mask, to_cpumask(pd->cpus));
drivers/scsi/elx/efct/efct_hw.c
232
u32 i, max_sgl, cpus;
drivers/scsi/elx/efct/efct_hw.c
306
cpus = num_possible_cpus();
drivers/scsi/elx/efct/efct_hw.c
307
hw->config.n_eq = cpus > EFCT_HW_MAX_NUM_EQ ? EFCT_HW_MAX_NUM_EQ : cpus;
drivers/thermal/cpufreq_cooling.c
285
num_cpus = cpumask_weight(cpufreq_cdev->policy->cpus);
drivers/thermal/intel/intel_hfi.c
125
cpumask_var_t cpus;
drivers/thermal/intel/intel_hfi.c
178
for_each_cpu(cpu, hfi_instance->cpus) {
drivers/thermal/intel/intel_hfi.c
209
cpu_count = cpumask_weight(hfi_instance->cpus);
drivers/thermal/intel/intel_hfi.c
479
cpumask_set_cpu(cpu, hfi_instance->cpus);
drivers/thermal/intel/intel_hfi.c
485
if (cpumask_weight(hfi_instance->cpus) == 1 && hfi_clients_nr > 0) {
drivers/thermal/intel/intel_hfi.c
527
cpumask_clear_cpu(cpu, hfi_instance->cpus);
drivers/thermal/intel/intel_hfi.c
529
if (!cpumask_weight(hfi_instance->cpus))
drivers/thermal/intel/intel_hfi.c
656
if (cpumask_empty(hfi_instance->cpus))
drivers/thermal/intel/intel_hfi.c
659
cpu = cpumask_any(hfi_instance->cpus);
drivers/thermal/intel/intel_hfi.c
699
if (!zalloc_cpumask_var(&hfi_instance->cpus, GFP_KERNEL))
drivers/thermal/intel/intel_hfi.c
726
free_cpumask_var(hfi_instance->cpus);
drivers/thermal/thermal_trace.h
116
TP_PROTO(const struct cpumask *cpus, unsigned int freq,
drivers/thermal/thermal_trace.h
119
TP_ARGS(cpus, freq, cdev_state, power),
drivers/thermal/thermal_trace.h
129
__assign_bitmask(cpumask, cpumask_bits(cpus),
fs/btrfs/compression.c
799
int cpus = num_online_cpus();
fs/btrfs/compression.c
824
if (atomic_read(total_ws) > cpus) {
fs/btrfs/compression.c
829
if (atomic_read(total_ws) > cpus && !*free_ws)
fs/resctrl/rdtgroup.c
4109
cpumask_var_t cpus)
fs/resctrl/rdtgroup.c
4128
rdt_move_group_tasks(rdtgrp, rdtgrp, cpus);
fs/resctrl/rdtgroup.c
4130
update_closid_rmid(cpus, NULL);
include/asm-generic/mshyperv.h
232
const struct cpumask *cpus,
include/asm-generic/mshyperv.h
253
for_each_cpu(cpu, cpus) {
include/asm-generic/mshyperv.h
278
const struct cpumask *cpus)
include/asm-generic/mshyperv.h
280
return __cpumask_to_vpset(vpset, cpus, NULL);
include/asm-generic/mshyperv.h
284
const struct cpumask *cpus,
include/asm-generic/mshyperv.h
287
return __cpumask_to_vpset(vpset, cpus, func);
include/linux/acpi.h
1548
void acpi_pptt_get_cpus_from_container(u32 acpi_cpu_id, cpumask_t *cpus);
include/linux/acpi.h
1550
int acpi_pptt_get_cpumask_from_cache_id(u32 cache_id, cpumask_t *cpus);
include/linux/acpi.h
1573
cpumask_t *cpus) { }
include/linux/acpi.h
1579
cpumask_t *cpus)
include/linux/arch_topology.h
32
void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq,
include/linux/arch_topology.h
49
void topology_set_scale_freq_source(struct scale_freq_data *data, const struct cpumask *cpus);
include/linux/arch_topology.h
50
void topology_clear_scale_freq_source(enum scale_freq_source source, const struct cpumask *cpus);
include/linux/arch_topology.h
59
void topology_update_hw_pressure(const struct cpumask *cpus,
include/linux/cpufreq.h
1225
void arch_set_freq_scale(const struct cpumask *cpus,
include/linux/cpufreq.h
230
return cpumask_empty(policy->cpus);
include/linux/cpufreq.h
235
return cpumask_weight(policy->cpus) > 1;
include/linux/cpufreq.h
55
cpumask_var_t cpus; /* Online CPUs only */
include/linux/energy_model.h
101
#define em_span_cpus(em) (to_cpumask((em)->cpus))
include/linux/energy_model.h
177
const cpumask_t *cpus, bool microwatts);
include/linux/energy_model.h
180
const cpumask_t *cpus, bool microwatts);
include/linux/energy_model.h
356
const cpumask_t *cpus, bool microwatts)
include/linux/energy_model.h
363
const cpumask_t *cpus, bool microwatts)
include/linux/energy_model.h
82
unsigned long cpus[];
include/linux/pm_domain.h
214
cpumask_var_t cpus; /* A cpumask of the attached CPUs */
include/linux/pm_opp.h
526
int dev_pm_opp_of_register_em(struct device *dev, struct cpumask *cpus);
include/linux/pm_opp.h
583
struct cpumask *cpus)
include/linux/sched/topology.h
229
void arch_update_hw_pressure(const struct cpumask *cpus,
include/linux/stop_machine.h
130
int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
include/linux/stop_machine.h
142
int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
include/linux/stop_machine.h
162
const struct cpumask *cpus);
include/linux/stop_machine.h
166
const struct cpumask *cpus)
include/linux/stop_machine.h
177
stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus)
include/linux/stop_machine.h
179
return stop_machine_cpuslocked(fn, data, cpus);
include/linux/stop_machine.h
184
const struct cpumask *cpus)
include/linux/stop_machine.h
186
return stop_machine(fn, data, cpus);
include/linux/topology.h
269
int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node);
include/linux/topology.h
272
static __always_inline int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node)
include/linux/topology.h
274
return cpumask_nth_and(cpu, cpus, cpu_online_mask);
include/net/rps.h
22
u16 cpus[];
include/sound/simple_card_utils.h
101
((i) < (props)->num.cpus) && \
include/sound/simple_card_utils.h
123
((i) < (props)->num.cpus) && \
include/sound/simple_card_utils.h
174
struct snd_soc_dai_link_component *cpus);
include/sound/simple_card_utils.h
175
void simple_util_canonicalize_cpu(struct snd_soc_dai_link_component *cpus,
include/sound/simple_card_utils.h
52
int cpus;
include/sound/simple_card_utils.h
84
#define simple_props_to_dlc_cpu(props, i) ((props)->cpus + i)
include/sound/soc.h
1401
struct snd_soc_dai_link_component *cpus);
include/sound/soc.h
719
struct snd_soc_dai_link_component *cpus;
include/sound/soc.h
822
return &(link)->cpus[n];
include/sound/soc.h
915
.cpus = cpu, \
include/sound/soc_sdw_utils.h
161
struct snd_soc_dai_link_component *cpus, int cpus_num,
include/trace/events/xen.h
350
TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm,
include/trace/events/xen.h
352
TP_ARGS(cpus, mm, addr, end),
include/trace/events/xen.h
359
TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
kernel/power/em_netlink.c
33
nr_cpus = cpumask_weight(to_cpumask(pd->cpus));
kernel/power/em_netlink.c
63
cpumask = to_cpumask(pd->cpus);
kernel/power/energy_model.c
151
debugfs_create_file("cpus", 0444, d, dev->em_pd->cpus,
kernel/power/energy_model.c
419
const cpumask_t *cpus,
kernel/power/energy_model.c
428
num_cpus = cpumask_weight(cpus);
kernel/power/energy_model.c
440
cpumask_copy(em_span_cpus(pd), cpus);
kernel/power/energy_model.c
469
for_each_cpu(cpu, cpus) {
kernel/power/energy_model.c
590
const cpumask_t *cpus, bool microwatts)
kernel/power/energy_model.c
592
int ret = em_dev_register_pd_no_update(dev, nr_states, cb, cpus, microwatts);
kernel/power/energy_model.c
614
const cpumask_t *cpus, bool microwatts)
kernel/power/energy_model.c
636
if (!cpus) {
kernel/power/energy_model.c
642
for_each_cpu(cpu, cpus) {
kernel/power/energy_model.c
656
cpumask_pr_args(cpus));
kernel/power/energy_model.c
681
ret = em_create_pd(dev, nr_states, cb, cpus, flags);
kernel/power/power.h
15
int cpus;
kernel/sched/cpufreq.c
72
return cpumask_test_cpu(smp_processor_id(), policy->cpus) ||
kernel/sched/cpufreq_schedutil.c
500
for_each_cpu(j, policy->cpus) {
kernel/sched/cpufreq_schedutil.c
865
for_each_cpu(cpu, policy->cpus) {
kernel/sched/cpufreq_schedutil.c
881
for_each_cpu(cpu, policy->cpus)
kernel/sched/deadline.c
190
void __dl_sub(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
kernel/sched/deadline.c
1901
int cpus;
kernel/sched/deadline.c
1906
cpus = dl_bw_cpus(cpu);
kernel/sched/deadline.c
1914
__dl_add(dl_b, new_bw, cpus);
kernel/sched/deadline.c
1916
__dl_sub(dl_b, dl_se->dl_bw, cpus);
kernel/sched/deadline.c
1917
__dl_add(dl_b, new_bw, cpus);
kernel/sched/deadline.c
193
__dl_update(dl_b, (s32)tsk_bw / cpus);
kernel/sched/deadline.c
197
void __dl_add(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
kernel/sched/deadline.c
200
__dl_update(dl_b, -((s32)tsk_bw / cpus));
kernel/sched/deadline.c
3160
static void dl_get_task_effective_cpus(struct task_struct *p, struct cpumask *cpus)
kernel/sched/deadline.c
3171
cpumask_andnot(cpus, cpu_active_mask, hk_msk);
kernel/sched/deadline.c
3181
cpuset_cpus_allowed_locked(p, cpus);
kernel/sched/deadline.c
3453
int cpu, cpus, ret = 0;
kernel/sched/deadline.c
3468
cpus = dl_bw_cpus(cpu);
kernel/sched/deadline.c
3471
if (new_bw * cpus < dl_b->total_bw)
kernel/sched/deadline.c
3544
int cpus, err = -1, cpu = task_cpu(p);
kernel/sched/deadline.c
3561
cpus = dl_bw_cpus(cpu);
kernel/sched/deadline.c
3567
__dl_sub(dl_b, p->dl.dl_bw, cpus);
kernel/sched/deadline.c
3568
__dl_add(dl_b, new_bw, cpus);
kernel/sched/deadline.c
3579
__dl_sub(dl_b, p->dl.dl_bw, cpus);
kernel/sched/deadline.c
3580
__dl_add(dl_b, new_bw, cpus);
kernel/sched/ext_idle.c
498
const struct cpumask *cpus = numa_span(prev_cpu);
kernel/sched/ext_idle.c
501
numa_cpus = cpus;
kernel/sched/ext_idle.c
502
else if (cpus && cpumask_and(local_cpus, allowed, cpus))
kernel/sched/ext_idle.c
508
const struct cpumask *cpus = llc_span(prev_cpu);
kernel/sched/ext_idle.c
511
llc_cpus = cpus;
kernel/sched/ext_idle.c
512
else if (cpus && cpumask_and(local_cpus, allowed, cpus))
kernel/sched/fair.c
10477
for_each_cpu_and(i, sched_group_span(group), env->cpus) {
kernel/sched/fair.c
11555
for_each_cpu_and(i, sched_group_span(group), env->cpus) {
kernel/sched/fair.c
11768
if (!cpumask_test_cpu(env->dst_cpu, env->cpus))
kernel/sched/fair.c
11786
for_each_cpu_and(cpu, swb_cpus, env->cpus) {
kernel/sched/fair.c
11874
struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask);
kernel/sched/fair.c
11882
.cpus = cpus,
kernel/sched/fair.c
11888
cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
kernel/sched/fair.c
11991
__cpumask_clear_cpu(env.dst_cpu, env.cpus);
kernel/sched/fair.c
12018
__cpumask_clear_cpu(cpu_of(busiest), cpus);
kernel/sched/fair.c
12027
if (!cpumask_subset(cpus, env.dst_grpmask)) {
kernel/sched/fair.c
194
unsigned int cpus = min_t(unsigned int, num_online_cpus(), 8);
kernel/sched/fair.c
202
factor = cpus;
kernel/sched/fair.c
206
factor = 1 + ilog2(cpus);
kernel/sched/fair.c
7622
static int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle_cpu)
kernel/sched/fair.c
7631
if (sched_idle_cpu(cpu) && cpumask_test_cpu(cpu, cpus)) {
kernel/sched/fair.c
7639
if (*idle_cpu == -1 && cpumask_test_cpu(cpu, cpus))
kernel/sched/fair.c
7646
cpumask_andnot(cpus, cpus, cpu_smt_mask(core));
kernel/sched/fair.c
7684
static inline int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle_cpu)
kernel/sched/fair.c
7703
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
kernel/sched/fair.c
7707
cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
kernel/sched/fair.c
7725
if (!cpumask_test_cpu(cpu, cpus))
kernel/sched/fair.c
7729
i = select_idle_core(p, cpu, cpus, &idle_cpu);
kernel/sched/fair.c
7740
cpumask_andnot(cpus, cpus, sched_group_span(sg));
kernel/sched/fair.c
7744
for_each_cpu_wrap(cpu, cpus, target + 1) {
kernel/sched/fair.c
7746
i = select_idle_core(p, cpu, cpus, &idle_cpu);
kernel/sched/fair.c
7776
struct cpumask *cpus;
kernel/sched/fair.c
7778
cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
kernel/sched/fair.c
7779
cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
kernel/sched/fair.c
7785
for_each_cpu_wrap(cpu, cpus, target) {
kernel/sched/fair.c
8386
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
kernel/sched/fair.c
8431
if (!cpumask_and(cpus, perf_domain_span(pd), cpu_online_mask))
kernel/sched/fair.c
8435
cpu = cpumask_first(cpus);
kernel/sched/fair.c
8441
for_each_cpu(cpu, cpus) {
kernel/sched/fair.c
8503
eenv_pd_busy_time(&eenv, cpus, p);
kernel/sched/fair.c
8505
base_energy = compute_energy(&eenv, pd, cpus, p, -1);
kernel/sched/fair.c
8509
prev_delta = compute_energy(&eenv, pd, cpus, p,
kernel/sched/fair.c
8533
cur_delta = compute_energy(&eenv, pd, cpus, p,
kernel/sched/fair.c
9339
struct cpumask *cpus;
kernel/sched/fair.c
9546
cpu = cpumask_first_and_and(env->dst_grpmask, env->cpus, p->cpus_ptr);
kernel/sched/sched.h
123
unsigned long cpus[];
kernel/sched/sched.h
128
#define cpu_capacity_span(asym_data) to_cpumask((asym_data)->cpus)
kernel/sched/sched.h
2025
extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu);
kernel/sched/sched.h
2034
static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
kernel/sched/sched.h
3701
#define perf_domain_span(pd) (to_cpumask(((pd)->em_pd->cpus)))
kernel/sched/topology.c
2233
int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
kernel/sched/topology.c
2245
cpu = cpumask_any_and_distribute(cpus, masks[i][j]);
kernel/sched/topology.c
2258
const struct cpumask *cpus;
kernel/sched/topology.c
2270
if (cpumask_weight_and(k->cpus, cur_hop[k->node]) <= k->cpu)
kernel/sched/topology.c
2279
k->w = cpumask_weight_and(k->cpus, prev_hop[k->node]);
kernel/sched/topology.c
2296
int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node)
kernel/sched/topology.c
2298
struct __cmp_key k = { .cpus = cpus, .cpu = cpu };
kernel/sched/topology.c
2303
return cpumask_nth_and(cpu, cpus, cpu_online_mask);
kernel/sched/topology.c
2321
cpumask_nth_and_andnot(cpu - k.w, cpus, k.masks[hop][node], k.masks[hop-1][node]) :
kernel/sched/topology.c
2322
cpumask_nth_and(cpu, cpus, k.masks[0][node]);
kernel/stop_machine.c
588
const struct cpumask *cpus)
kernel/stop_machine.c
594
.active_cpus = cpus,
kernel/stop_machine.c
623
int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus)
kernel/stop_machine.c
629
ret = stop_machine_cpuslocked(fn, data, cpus);
kernel/stop_machine.c
679
const struct cpumask *cpus)
kernel/stop_machine.c
682
.active_cpus = cpus };
kernel/trace/ring_buffer.c
2510
buffer->cpus = nr_cpu_ids;
kernel/trace/ring_buffer.c
573
int cpus;
mm/slub.c
8640
DECLARE_BITMAP(cpus, NR_CPUS);
mm/slub.c
8728
to_cpumask(l->cpus));
mm/slub.c
8765
cpumask_clear(to_cpumask(l->cpus));
mm/slub.c
8766
cpumask_set_cpu(track->cpu, to_cpumask(l->cpus));
mm/slub.c
9641
if (num_online_cpus() > 1 && !cpumask_empty(to_cpumask(l->cpus)))
mm/slub.c
9643
cpumask_pr_args(to_cpumask(l->cpus)));
net/ceph/osdmap.c
1046
int cpus = num_online_cpus();
net/ceph/osdmap.c
1059
if (atomic_read(&wsm->total_ws) > cpus) {
net/ceph/osdmap.c
1064
if (atomic_read(&wsm->total_ws) > cpus && !wsm->free_ws)
net/core/dev.c
3363
cpumask_var_t cpus;
net/core/dev.c
3366
if (unlikely(is_kdump_kernel() || !zalloc_cpumask_var(&cpus, GFP_KERNEL)))
net/core/dev.c
3369
cpumask_copy(cpus, cpu_online_mask);
net/core/dev.c
3370
for_each_cpu(cpu, cpus) {
net/core/dev.c
3372
cpumask_andnot(cpus, cpus, topology_sibling_cpumask(cpu));
net/core/dev.c
3374
free_cpumask_var(cpus);
net/core/dev.c
5175
tcpu = map->cpus[reciprocal_scale(hash, map->len)];
net/core/net-sysfs.c
972
cpumask_set_cpu(map->cpus[i], mask);
net/core/net-sysfs.c
996
map->cpus[i++] = cpu;
samples/trace_events/trace-events-sample.h
304
__bitmask( cpus, num_possible_cpus() )
samples/trace_events/trace-events-sample.h
318
__assign_bitmask(cpus, cpumask_bits(mask), num_possible_cpus());
samples/trace_events/trace-events-sample.h
373
__get_bitmask(cpus), __get_cpumask(cpum),
samples/trace_events/trace-events-sample.h
375
__get_dynamic_array_len(cpus),
samples/trace_events/trace-events-sample.h
376
__get_dynamic_array_len(cpus),
samples/trace_events/trace-events-sample.h
377
__get_dynamic_array(cpus))
samples/trace_events/trace-events-sample.h
574
TP_PROTO(const char *foo, int bar, unsigned long *mask, const cpumask_t *cpus),
samples/trace_events/trace-events-sample.h
576
TP_ARGS(foo, bar, mask, cpus),
samples/trace_events/trace-events-sample.h
591
__assign_rel_cpumask(cpumask, cpus);
scripts/dtc/livetree.c
713
struct node *cpus, *bootcpu;
scripts/dtc/livetree.c
716
cpus = get_node_by_path(tree, "/cpus");
scripts/dtc/livetree.c
717
if (!cpus)
scripts/dtc/livetree.c
721
bootcpu = cpus->children;
sound/soc/amd/acp/acp-mach-common.c
1422
links[i].cpus = sof_sp;
sound/soc/amd/acp/acp-mach-common.c
1457
links[i].cpus = sof_hs;
sound/soc/amd/acp/acp-mach-common.c
1487
links[i].cpus = sof_sp;
sound/soc/amd/acp/acp-mach-common.c
1490
links[i].cpus = sof_sp_virtual;
sound/soc/amd/acp/acp-mach-common.c
1523
links[i].cpus = sof_hs_virtual;
sound/soc/amd/acp/acp-mach-common.c
1564
links[i].cpus = sof_bt;
sound/soc/amd/acp/acp-mach-common.c
1583
links[i].cpus = sof_dmic;
sound/soc/amd/acp/acp-mach-common.c
1622
links[i].cpus = i2s_sp;
sound/soc/amd/acp/acp-mach-common.c
1656
links[i].cpus = i2s_hs;
sound/soc/amd/acp/acp-mach-common.c
1696
links[i].cpus = i2s_sp;
sound/soc/amd/acp/acp-mach-common.c
1726
links[i].cpus = i2s_hs;
sound/soc/amd/acp/acp-mach-common.c
1778
links[i].cpus = pdm_dmic;
sound/soc/amd/acp/acp-sdw-legacy-mach.c
189
struct snd_soc_dai_link_component *cpus;
sound/soc/amd/acp/acp-sdw-legacy-mach.c
242
cpus = devm_kcalloc(dev, num_cpus, sizeof(*cpus), GFP_KERNEL);
sound/soc/amd/acp/acp-sdw-legacy-mach.c
243
if (!cpus)
sound/soc/amd/acp/acp-sdw-legacy-mach.c
260
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/amd/acp/acp-sdw-legacy-mach.c
263
dev_dbg(dev, "cpu->dai_name:%s\n", cpus->dai_name);
sound/soc/amd/acp/acp-sdw-legacy-mach.c
264
if (!cpus->dai_name)
sound/soc/amd/acp/acp-sdw-legacy-mach.c
281
cpus, num_cpus, sdw_platform_component,
sound/soc/amd/acp/acp-sdw-sof-mach.c
105
struct snd_soc_dai_link_component *cpus;
sound/soc/amd/acp/acp-sdw-sof-mach.c
158
cpus = devm_kcalloc(dev, num_cpus, sizeof(*cpus), GFP_KERNEL);
sound/soc/amd/acp/acp-sdw-sof-mach.c
159
if (!cpus)
sound/soc/amd/acp/acp-sdw-sof-mach.c
176
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/amd/acp/acp-sdw-sof-mach.c
179
if (!cpus->dai_name)
sound/soc/amd/acp/acp-sdw-sof-mach.c
181
dev_dbg(dev, "cpu->dai_name:%s\n", cpus->dai_name);
sound/soc/amd/acp/acp-sdw-sof-mach.c
197
cpus, num_cpus, platform_component,
sound/soc/amd/acp3x-rt5682-max9836.c
335
.cpus = acp3x_bt,
sound/soc/atmel/atmel-classd.c
481
dai_link->cpus = &comp[0];
sound/soc/atmel/atmel-classd.c
491
dai_link->cpus->dai_name = dev_name(dev);
sound/soc/atmel/atmel-pdmic.c
503
dai_link->cpus = comp;
sound/soc/atmel/atmel-pdmic.c
511
dai_link->cpus->dai_name = dev_name(dev);
sound/soc/atmel/atmel_wm8904.c
114
dailink->cpus->of_node = cpu_np;
sound/soc/atmel/atmel_wm8904.c
143
id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc");
sound/soc/atmel/atmel_wm8904.c
169
id = of_alias_get_id((struct device_node *)dailink->cpus->of_node, "ssc");
sound/soc/atmel/mikroe-proto.c
121
dai->cpus->of_node = cpu_np;
sound/soc/atmel/mikroe-proto.c
96
dai->cpus = &comp[0];
sound/soc/atmel/sam9g20_wm8731.c
159
at91sam9g20ek_dai.cpus->dai_name = NULL;
sound/soc/atmel/sam9g20_wm8731.c
168
at91sam9g20ek_dai.cpus->of_node = cpu_np;
sound/soc/atmel/sam9x5_wm8731.c
106
dai->cpus = &comp[0];
sound/soc/atmel/sam9x5_wm8731.c
147
dai->cpus->of_node = cpu_np;
sound/soc/atmel/tse850-pcm5142.c
330
dailink->cpus->of_node = cpu_np;
sound/soc/fsl/eukrea-tlv320.c
136
eukrea_tlv320_dai.cpus->of_node = ssi_np;
sound/soc/fsl/eukrea-tlv320.c
139
eukrea_tlv320_dai.cpus->dai_name = "imx-ssi.0";
sound/soc/fsl/fsl-asoc-card.c
719
priv->dai_link[0].cpus = &dlc[0];
sound/soc/fsl/fsl-asoc-card.c
726
priv->dai_link[1].cpus = &dlc[4];
sound/soc/fsl/fsl-asoc-card.c
733
priv->dai_link[2].cpus = &dlc[7];
sound/soc/fsl/fsl-asoc-card.c
953
priv->dai_link[0].cpus->of_node = cpu_np;
sound/soc/fsl/fsl-asoc-card.c
992
priv->dai_link[1].cpus->of_node = asrc_np;
sound/soc/fsl/fsl-asoc-card.c
999
priv->dai_link[2].cpus->of_node = cpu_np;
sound/soc/fsl/imx-audmix.c
230
priv->dai[i].cpus =
sound/soc/fsl/imx-audmix.c
239
priv->dai[i].cpus->of_node = args.np;
sound/soc/fsl/imx-audmix.c
240
priv->dai[i].cpus->dai_name = name[1][i];
sound/soc/fsl/imx-audmix.c
256
priv->dai[num_dai + i].cpus = &dlc[1];
sound/soc/fsl/imx-audmix.c
263
priv->dai[num_dai + i].cpus->of_node = audmix_np;
sound/soc/fsl/imx-audmix.c
264
priv->dai[num_dai + i].cpus->dai_name = be_name;
sound/soc/fsl/imx-card.c
587
link->cpus = &dlc[0];
sound/soc/fsl/imx-card.c
606
ret = snd_soc_of_get_dlc(cpu, &args, link->cpus, 0);
sound/soc/fsl/imx-card.c
637
link->platforms->of_node = link->cpus->of_node;
sound/soc/fsl/imx-es8328.c
185
data->dai.cpus =
sound/soc/fsl/imx-es8328.c
197
data->dai.cpus->of_node = ssi_np;
sound/soc/fsl/imx-hdmi.c
125
data->dai.cpus = &dlc[0];
sound/soc/fsl/imx-hdmi.c
134
data->dai.cpus->of_node = cpu_np;
sound/soc/fsl/imx-rpmsg.c
126
data->dai.cpus = &dlc[0];
sound/soc/fsl/imx-rpmsg.c
147
data->dai.cpus->dai_name = pdev->dev.platform_data;
sound/soc/fsl/imx-rpmsg.c
148
cpu_dai = snd_soc_find_dai_with_mutex(data->dai.cpus);
sound/soc/fsl/imx-sgtl5000.c
140
data->dai.cpus = &comp[0];
sound/soc/fsl/imx-sgtl5000.c
152
data->dai.cpus->of_node = ssi_np;
sound/soc/fsl/p1022_ds.c
226
mdata->dai[0].cpus = &comp[0];
sound/soc/fsl/p1022_ds.c
234
mdata->dai[1].cpus = &comp[3];
sound/soc/fsl/p1022_ds.c
243
mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev);
sound/soc/fsl/p1022_rdk.c
241
mdata->dai[0].cpus = &comp[0];
sound/soc/fsl/p1022_rdk.c
249
mdata->dai[1].cpus = &comp[3];
sound/soc/fsl/p1022_rdk.c
257
mdata->dai[0].cpus->dai_name = dev_name(&ssi_pdev->dev);
sound/soc/generic/audio-graph-card.c
213
struct snd_soc_dai_link_component *cpus = snd_soc_link_to_cpu(dai_link, 0);
sound/soc/generic/audio-graph-card.c
228
"fe.%pOFP.%s", cpus->of_node, cpus->dai_name);
sound/soc/generic/audio-graph-card.c
239
if (card->component_chaining && !soc_component_is_pcm(cpus)) {
sound/soc/generic/audio-graph-card.c
244
simple_util_canonicalize_cpu(cpus, is_single_links);
sound/soc/generic/audio-graph-card.c
245
simple_util_canonicalize_platform(platforms, cpus);
sound/soc/generic/audio-graph-card.c
288
struct snd_soc_dai_link_component *cpus = snd_soc_link_to_cpu(dai_link, 0);
sound/soc/generic/audio-graph-card.c
305
"%s-%s", cpus->dai_name, codecs->dai_name);
sound/soc/generic/audio-graph-card.c
307
simple_util_canonicalize_cpu(cpus, is_single_links);
sound/soc/generic/audio-graph-card.c
308
simple_util_canonicalize_platform(platforms, cpus);
sound/soc/generic/audio-graph-card.c
451
li->num[li->link].cpus = 1;
sound/soc/generic/audio-graph-card.c
481
li->num[li->link].cpus = 1;
sound/soc/generic/audio-graph-card2.c
1151
li->num[li->link].cpus =
sound/soc/generic/audio-graph-card2.c
1187
li->num[li->link].cpus = graph_counter(rport); /* FE */
sound/soc/generic/audio-graph-card2.c
1221
li->num[li->link].cpus =
sound/soc/generic/audio-graph-card2.c
439
struct snd_soc_dai_link_component *cpus = dlc;
sound/soc/generic/audio-graph-card2.c
454
cpus->dai_name, cpu_multi,
sound/soc/generic/audio-graph-card2.c
460
cpus->of_node, cpus->dai_name, cpu_multi);
sound/soc/generic/audio-graph-card2.c
469
cpus->dai_name, cpu_multi,
sound/soc/generic/audio-graph-card2.c
492
struct snd_soc_dai_link_component *cpus = dlc;
sound/soc/generic/audio-graph-card2.c
495
simple_util_canonicalize_cpu(cpus, is_single_links);
sound/soc/generic/audio-graph-card2.c
496
simple_util_canonicalize_platform(platforms, cpus);
sound/soc/generic/simple-card-utils.c
681
struct snd_soc_dai_link_component *cpus)
sound/soc/generic/simple-card-utils.c
694
snd_soc_dlc_use_cpu_as_platform(platforms, cpus);
sound/soc/generic/simple-card-utils.c
698
void simple_util_canonicalize_cpu(struct snd_soc_dai_link_component *cpus,
sound/soc/generic/simple-card-utils.c
711
cpus->dai_name = NULL;
sound/soc/generic/simple-card-utils.c
913
int cc = li->num[i].cpus + li->num[i].codecs;
sound/soc/generic/simple-card-utils.c
918
if (!li->num[i].cpus)
sound/soc/generic/simple-card-utils.c
948
if (li->num[i].cpus) {
sound/soc/generic/simple-card-utils.c
950
dai_link[i].cpus = dlcs;
sound/soc/generic/simple-card-utils.c
951
dai_props[i].num.cpus =
sound/soc/generic/simple-card-utils.c
952
dai_link[i].num_cpus = li->num[i].cpus;
sound/soc/generic/simple-card-utils.c
955
dlcs += li->num[i].cpus;
sound/soc/generic/simple-card-utils.c
956
dais += li->num[i].cpus;
sound/soc/generic/simple-card-utils.c
959
dai_link[i].cpus = &snd_soc_dummy_dlc;
sound/soc/generic/simple-card-utils.c
960
dai_props[i].num.cpus =
sound/soc/generic/simple-card-utils.c
975
if (!li->num[i].cpus) {
sound/soc/generic/simple-card.c
253
struct snd_soc_dai_link_component *cpus = snd_soc_link_to_cpu(dai_link, 0);
sound/soc/generic/simple-card.c
267
snprintf(dai_name, sizeof(dai_name), "fe.%s", cpus->dai_name);
sound/soc/generic/simple-card.c
269
simple_util_canonicalize_cpu(cpus, is_single_links);
sound/soc/generic/simple-card.c
270
simple_util_canonicalize_platform(platforms, cpus);
sound/soc/generic/simple-card.c
316
struct snd_soc_dai_link_component *cpus = snd_soc_link_to_cpu(dai_link, 0);
sound/soc/generic/simple-card.c
350
"%s-%s", cpus->dai_name, codecs->dai_name);
sound/soc/generic/simple-card.c
352
simple_util_canonicalize_cpu(cpus, single_cpu);
sound/soc/generic/simple-card.c
353
simple_util_canonicalize_platform(platforms, cpus);
sound/soc/generic/simple-card.c
584
li->num[li->link].cpus = 1;
sound/soc/generic/simple-card.c
611
li->num[li->link].cpus = 1;
sound/soc/generic/simple-card.c
678
li->num[0].cpus = 1;
sound/soc/generic/simple-card.c
755
struct snd_soc_dai_link_component *cpus;
sound/soc/generic/simple-card.c
778
cpus = dai_link->cpus;
sound/soc/generic/simple-card.c
779
cpus->dai_name = cinfo->cpu_dai.name;
sound/soc/intel/avs/boards/da7219.c
181
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/da7219.c
183
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/da7219.c
186
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/da7219.c
190
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/dmic.c
58
dl[0].cpus = dmic_pin;
sound/soc/intel/avs/boards/dmic.c
61
dl[1].cpus = dmic_wov_pin;
sound/soc/intel/avs/boards/es8336.c
213
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/es8336.c
215
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/es8336.c
218
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/es8336.c
222
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/hdaudio.c
165
.cpus = &snd_soc_dummy_dlc,
sound/soc/intel/avs/boards/hdaudio.c
48
dl[i].cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/hdaudio.c
49
if (!dl[i].codecs || !dl[i].cpus)
sound/soc/intel/avs/boards/hdaudio.c
52
dl[i].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "%s-cpu%d", cname, i);
sound/soc/intel/avs/boards/hdaudio.c
53
if (!dl[i].cpus->dai_name)
sound/soc/intel/avs/boards/i2s_test.c
30
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/i2s_test.c
31
if (!dl->name || !dl->cpus)
sound/soc/intel/avs/boards/i2s_test.c
34
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/i2s_test.c
37
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/max98357a.c
62
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/max98357a.c
64
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/max98357a.c
67
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/max98357a.c
71
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/max98373.c
111
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/max98373.c
113
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/max98373.c
116
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/max98373.c
122
if (!dl->cpus->dai_name || !dl->codecs[0].name || !dl->codecs[0].dai_name ||
sound/soc/intel/avs/boards/max98927.c
108
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/max98927.c
110
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/max98927.c
113
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/max98927.c
119
if (!dl->cpus->dai_name || !dl->codecs[0].name || !dl->codecs[0].dai_name ||
sound/soc/intel/avs/boards/nau8825.c
187
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/nau8825.c
189
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/nau8825.c
192
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/nau8825.c
196
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/pcm3168a.c
87
dl[0].cpus = cpu_ssp0;
sound/soc/intel/avs/boards/pcm3168a.c
90
dl[1].cpus = cpu_ssp2;
sound/soc/intel/avs/boards/probe.c
21
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/probe.c
23
if (!dl->cpus || !dl->platforms)
sound/soc/intel/avs/boards/probe.c
27
dl->cpus->dai_name = "Probe Extraction CPU DAI";
sound/soc/intel/avs/boards/rt274.c
163
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt274.c
165
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt274.c
168
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt274.c
172
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt286.c
131
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt286.c
133
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt286.c
136
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt286.c
140
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt298.c
148
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt298.c
150
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt298.c
153
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt298.c
157
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt5514.c
101
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt5514.c
103
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt5514.c
106
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt5514.c
110
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt5640.c
154
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt5640.c
156
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt5640.c
159
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt5640.c
163
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt5663.c
150
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt5663.c
152
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt5663.c
155
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt5663.c
159
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/rt5682.c
220
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/rt5682.c
222
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/boards/rt5682.c
225
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/rt5682.c
229
if (!dl->cpus->dai_name || !dl->codecs->name || !dl->codecs->dai_name)
sound/soc/intel/avs/boards/ssm4567.c
102
dl->cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/avs/boards/ssm4567.c
108
if (!dl->cpus->dai_name || !dl->codecs[0].name || !dl->codecs[0].dai_name ||
sound/soc/intel/avs/boards/ssm4567.c
97
dl->cpus = devm_kzalloc(dev, sizeof(*dl->cpus), GFP_KERNEL);
sound/soc/intel/avs/boards/ssm4567.c
99
if (!dl->name || !dl->cpus || !dl->codecs)
sound/soc/intel/avs/pcm.c
1481
struct snd_soc_dai_driver *cpus, *dai;
sound/soc/intel/avs/pcm.c
1500
cpus = devm_kcalloc(adev->dev, cpu_count, sizeof(*cpus), GFP_KERNEL);
sound/soc/intel/avs/pcm.c
1501
if (!cpus)
sound/soc/intel/avs/pcm.c
1504
dai = cpus;
sound/soc/intel/avs/pcm.c
1545
return avs_register_component(adev->dev, name, &avs_component_driver, cpus, cpu_count);
sound/soc/intel/boards/bytcht_es8316.c
622
byt_cht_es8316_dais[dai_index].cpus->dai_name = "ssp0-port";
sound/soc/intel/boards/bytcr_rt5640.c
1880
byt_rt5640_dais[dai_index].cpus->dai_name = "ssp0-port";
sound/soc/intel/boards/bytcr_rt5651.c
1070
byt_rt5651_dais[dai_index].cpus->dai_name = "ssp0-port";
sound/soc/intel/boards/bytcr_wm5102.c
621
byt_wm5102_dais[dai_index].cpus->dai_name = "ssp2-port";
sound/soc/intel/boards/cht_bsw_rt5645.c
661
cht_dailink[dai_index].cpus->dai_name = "ssp0-port";
sound/soc/intel/boards/cht_bsw_rt5672.c
488
cht_dailink[dai_index].cpus->dai_name = "ssp0-port";
sound/soc/intel/boards/sof_board_helpers.c
184
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_board_helpers.c
195
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
197
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
202
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "ssp%d-port",
sound/soc/intel/boards/sof_board_helpers.c
205
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "SSP%d Pin",
sound/soc/intel/boards/sof_board_helpers.c
208
if (!cpus->dai_name)
sound/soc/intel/boards/sof_board_helpers.c
211
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
229
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_board_helpers.c
232
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
234
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
242
cpus->dai_name = "DMIC01 Pin";
sound/soc/intel/boards/sof_board_helpers.c
248
cpus->dai_name = "DMIC16k Pin";
sound/soc/intel/boards/sof_board_helpers.c
255
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
279
struct snd_soc_dai_link_component *cpus, *codecs;
sound/soc/intel/boards/sof_board_helpers.c
290
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
292
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
295
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "iDisp%d Pin", hdmi_id);
sound/soc/intel/boards/sof_board_helpers.c
296
if (!cpus->dai_name)
sound/soc/intel/boards/sof_board_helpers.c
299
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
338
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_board_helpers.c
349
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
351
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
354
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "SSP%d Pin", ssp_amp);
sound/soc/intel/boards/sof_board_helpers.c
355
if (!cpus->dai_name)
sound/soc/intel/boards/sof_board_helpers.c
358
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
377
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_board_helpers.c
387
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
389
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
392
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "SSP%d Pin", ssp_bt);
sound/soc/intel/boards/sof_board_helpers.c
393
if (!cpus->dai_name)
sound/soc/intel/boards/sof_board_helpers.c
396
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
416
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_board_helpers.c
426
cpus = devm_kzalloc(dev, sizeof(struct snd_soc_dai_link_component),
sound/soc/intel/boards/sof_board_helpers.c
428
if (!cpus)
sound/soc/intel/boards/sof_board_helpers.c
431
cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL, "SSP%d Pin", ssp_hdmi);
sound/soc/intel/boards/sof_board_helpers.c
432
if (!cpus->dai_name)
sound/soc/intel/boards/sof_board_helpers.c
435
link->cpus = cpus;
sound/soc/intel/boards/sof_board_helpers.c
463
link->cpus = hda_analog_cpus;
sound/soc/intel/boards/sof_board_helpers.c
476
link->cpus = hda_digital_cpus;
sound/soc/intel/boards/sof_es8336.c
440
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_es8336.c
449
cpus = devm_kcalloc(dev, sof_es8336_card.num_links,
sound/soc/intel/boards/sof_es8336.c
451
if (!links || !cpus)
sound/soc/intel/boards/sof_es8336.c
470
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_es8336.c
473
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_es8336.c
476
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_es8336.c
485
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_es8336.c
486
links[id].cpus->dai_name = "DMIC01 Pin";
sound/soc/intel/boards/sof_es8336.c
491
links[id + 1].cpus = &cpus[id + 1];
sound/soc/intel/boards/sof_es8336.c
492
links[id + 1].cpus->dai_name = "DMIC16k Pin";
sound/soc/intel/boards/sof_es8336.c
531
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_es8336.c
533
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_es8336.c
535
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_es8336.c
568
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_es8336.c
569
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_es8336.c
571
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_pcm512x.c
226
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_pcm512x.c
232
cpus = devm_kcalloc(dev, sof_audio_card_pcm512x.num_links,
sound/soc/intel/boards/sof_pcm512x.c
234
if (!links || !cpus)
sound/soc/intel/boards/sof_pcm512x.c
256
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_pcm512x.c
259
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_pcm512x.c
262
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_pcm512x.c
265
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_pcm512x.c
268
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_pcm512x.c
277
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_pcm512x.c
278
links[id].cpus->dai_name = "DMIC01 Pin";
sound/soc/intel/boards/sof_pcm512x.c
283
links[id + 1].cpus = &cpus[id + 1];
sound/soc/intel/boards/sof_pcm512x.c
284
links[id + 1].cpus->dai_name = "DMIC16k Pin";
sound/soc/intel/boards/sof_pcm512x.c
317
links[id].cpus = &cpus[id];
sound/soc/intel/boards/sof_pcm512x.c
319
links[id].cpus->dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_pcm512x.c
321
if (!links[id].cpus->dai_name)
sound/soc/intel/boards/sof_sdw.c
1000
cpus, num_cpus, platform, 1, codecs, num_codecs,
sound/soc/intel/boards/sof_sdw.c
913
struct snd_soc_dai_link_component *cpus;
sound/soc/intel/boards/sof_sdw.c
947
cpus = devm_kcalloc(dev, num_cpus, sizeof(*cpus), GFP_KERNEL);
sound/soc/intel/boards/sof_sdw.c
948
if (!cpus)
sound/soc/intel/boards/sof_sdw.c
973
cpus[i].dai_name = devm_kasprintf(dev, GFP_KERNEL,
sound/soc/intel/boards/sof_sdw.c
976
if (!cpus[i].dai_name)
sound/soc/kirkwood/armada-370-db.c
111
a370db_dai[0].cpus->of_node =
sound/soc/kirkwood/armada-370-db.c
114
a370db_dai[0].platforms->of_node = a370db_dai[0].cpus->of_node;
sound/soc/kirkwood/armada-370-db.c
120
a370db_dai[1].cpus->of_node = a370db_dai[0].cpus->of_node;
sound/soc/kirkwood/armada-370-db.c
121
a370db_dai[1].platforms->of_node = a370db_dai[0].cpus->of_node;
sound/soc/kirkwood/armada-370-db.c
127
a370db_dai[2].cpus->of_node = a370db_dai[0].cpus->of_node;
sound/soc/kirkwood/armada-370-db.c
128
a370db_dai[2].platforms->of_node = a370db_dai[0].cpus->of_node;
sound/soc/loongson/loongson_card.c
142
ret = snd_soc_of_get_dlc(cpu, NULL, loongson_dai_links[i].cpus, 0);
sound/soc/loongson/loongson_card.c
147
loongson_dai_links[i].platforms->of_node = loongson_dai_links[i].cpus->of_node;
sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c
768
dai_link->cpus = i2s3_max98357a_cpus;
sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c
781
dai_link->cpus = i2s3_rt1015_cpus;
sound/soc/mediatek/mt8183/mt8183-da7219-max98357.c
794
dai_link->cpus = i2s3_rt1015p_cpus;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
727
dai_link->cpus[0].name = NULL;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
728
dai_link->cpus[0].of_node = ec_codec;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
729
dai_link->cpus[0].dai_name = NULL;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
743
dai_link->cpus = i2s3_max98357a_cpus;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
756
dai_link->cpus = i2s3_rt1015_cpus;
sound/soc/mediatek/mt8183/mt8183-mt6358-ts3a227-max98357.c
769
dai_link->cpus = i2s3_rt1015p_cpus;
sound/soc/meson/axg-card.c
128
lb->cpus = dlc;
sound/soc/meson/axg-card.c
134
lb->cpus->of_node = pad->cpus->of_node;
sound/soc/meson/axg-card.c
135
lb->cpus->dai_name = "TDM Loopback";
sound/soc/meson/axg-card.c
148
of_node_get(lb->cpus->of_node);
sound/soc/meson/axg-card.c
263
link->dai_fmt = meson_card_parse_daifmt(node, link->cpus->of_node);
sound/soc/meson/axg-card.c
319
dai_link->cpus = cpu;
sound/soc/meson/axg-card.c
323
ret = meson_card_parse_dai(card, np, dai_link->cpus);
sound/soc/meson/axg-card.c
327
if (axg_card_cpu_is_playback_fe(dai_link->cpus->of_node))
sound/soc/meson/axg-card.c
329
else if (axg_card_cpu_is_capture_fe(dai_link->cpus->of_node))
sound/soc/meson/axg-card.c
337
if (axg_card_cpu_is_codec(dai_link->cpus->of_node)) {
sound/soc/meson/axg-card.c
342
if (axg_card_cpu_is_tdm_iface(dai_link->cpus->of_node))
sound/soc/meson/gx-card.c
105
if (gx_card_cpu_identify(dai_link->cpus, "CODEC CTRL")) {
sound/soc/meson/gx-card.c
111
if (gx_card_cpu_identify(dai_link->cpus, "I2S Encoder"))
sound/soc/meson/gx-card.c
60
link->dai_fmt = meson_card_parse_daifmt(node, link->cpus->of_node);
sound/soc/meson/gx-card.c
90
dai_link->cpus = cpu;
sound/soc/meson/gx-card.c
93
ret = meson_card_parse_dai(card, np, dai_link->cpus);
sound/soc/meson/gx-card.c
97
if (gx_card_cpu_identify(dai_link->cpus, "FIFO"))
sound/soc/meson/meson-card-utils.c
245
if (link->cpus)
sound/soc/meson/meson-card-utils.c
246
of_node_put(link->cpus->of_node);
sound/soc/mxs/mxs-sgtl5000.c
129
mxs_sgtl5000_dai[i].cpus->dai_name = NULL;
sound/soc/mxs/mxs-sgtl5000.c
130
mxs_sgtl5000_dai[i].cpus->of_node = saif_np[i];
sound/soc/qcom/common.c
114
ret = snd_soc_of_get_dlc(cpu, &args, link->cpus, 0);
sound/soc/qcom/common.c
133
link->platforms->of_node = link->cpus->of_node;
sound/soc/qcom/common.c
92
link->cpus = &dlc[0];
sound/soc/qcom/storm.c
73
dai_link->cpus->of_node = of_parse_phandle(np, "cpu", 0);
sound/soc/qcom/storm.c
74
if (!dai_link->cpus->of_node) {
sound/soc/qcom/storm.c
78
dai_link->platforms->of_node = dai_link->cpus->of_node;
sound/soc/rockchip/rk3288_hdmi_analog.c
215
rk_dailink.cpus->of_node = of_parse_phandle(np, "rockchip,i2s-controller",
sound/soc/rockchip/rk3288_hdmi_analog.c
217
if (!rk_dailink.cpus->of_node) {
sound/soc/rockchip/rk3288_hdmi_analog.c
223
rk_dailink.platforms->of_node = rk_dailink.cpus->of_node;
sound/soc/rockchip/rk3399_gru_sound.c
564
dai->cpus->of_node = np_cpu;
sound/soc/rockchip/rockchip_max98090.c
401
card->dai_link[DAILINK_MAX98090].cpus->of_node = np_cpu;
sound/soc/rockchip/rockchip_max98090.c
403
card->dai_link[DAILINK_HDMI].cpus->of_node = np_cpu;
sound/soc/rockchip/rockchip_max98090.c
408
card->dai_link[0].cpus->of_node = np_cpu;
sound/soc/rockchip/rockchip_max98090.c
413
card->dai_link[0].cpus->of_node = np_cpu;
sound/soc/rockchip/rockchip_rt5645.c
182
rk_dailink.cpus->of_node = of_parse_phandle(np,
sound/soc/rockchip/rockchip_rt5645.c
184
if (!rk_dailink.cpus->of_node) {
sound/soc/rockchip/rockchip_rt5645.c
191
rk_dailink.platforms->of_node = rk_dailink.cpus->of_node;
sound/soc/rockchip/rockchip_rt5645.c
210
of_node_put(rk_dailink.cpus->of_node);
sound/soc/rockchip/rockchip_rt5645.c
211
rk_dailink.cpus->of_node = NULL;
sound/soc/rockchip/rockchip_rt5645.c
221
of_node_put(rk_dailink.cpus->of_node);
sound/soc/rockchip/rockchip_rt5645.c
222
rk_dailink.cpus->of_node = NULL;
sound/soc/samsung/aries_wm8994.c
654
aries_dai[0].cpus->of_node = of_parse_phandle(cpu,
sound/soc/samsung/aries_wm8994.c
656
if (!aries_dai[0].cpus->of_node) {
sound/soc/samsung/aries_wm8994.c
661
aries_dai[0].platforms->of_node = aries_dai[0].cpus->of_node;
sound/soc/samsung/aries_wm8994.c
664
aries_dai[2].cpus->of_node = of_parse_phandle(cpu,
sound/soc/samsung/aries_wm8994.c
666
if (!aries_dai[2].cpus->of_node) {
sound/soc/samsung/arndale.c
141
of_node_put(dai_link->cpus->of_node);
sound/soc/samsung/arndale.c
157
dai_link->cpus->of_node = of_parse_phandle(np, "samsung,audio-cpu", 0);
sound/soc/samsung/arndale.c
158
if (!dai_link->cpus->of_node) {
sound/soc/samsung/arndale.c
165
dai_link->platforms->of_node = dai_link->cpus->of_node;
sound/soc/samsung/midas_wm1811.c
731
dai_link->cpus->of_node = cpu_dai_node;
sound/soc/samsung/odroid.c
260
ret = snd_soc_of_get_dai_name(cpu, &link->cpus->dai_name, i);
sound/soc/samsung/smdk_wm8994.c
144
smdk_dai[0].cpus->dai_name = NULL;
sound/soc/samsung/smdk_wm8994.c
145
smdk_dai[0].cpus->of_node = of_parse_phandle(np,
sound/soc/samsung/smdk_wm8994.c
147
if (!smdk_dai[0].cpus->of_node) {
sound/soc/samsung/smdk_wm8994.c
155
smdk_dai[0].platforms->of_node = smdk_dai[0].cpus->of_node;
sound/soc/samsung/snow.c
145
link->cpus = links_cpus;
sound/soc/samsung/snow.c
162
link->cpus->of_node = of_parse_phandle(cpu, "sound-dai", 0);
sound/soc/samsung/snow.c
165
if (!link->cpus->of_node) {
sound/soc/samsung/snow.c
175
of_node_put(link->cpus->of_node);
sound/soc/samsung/snow.c
180
priv->clk_i2s_bus = of_clk_get_by_name(link->cpus->of_node,
sound/soc/samsung/snow.c
184
of_node_put(link->cpus->of_node);
sound/soc/samsung/snow.c
190
link->cpus->of_node = of_parse_phandle(dev->of_node,
sound/soc/samsung/snow.c
192
if (!link->cpus->of_node) {
sound/soc/samsung/snow.c
200
of_node_put(link->cpus->of_node);
sound/soc/samsung/snow.c
206
link->platforms->of_node = link->cpus->of_node;
sound/soc/samsung/snow.c
226
of_node_put(link->cpus->of_node);
sound/soc/samsung/tm2_wm5110.c
579
dai_link->cpus->name = NULL;
sound/soc/samsung/tm2_wm5110.c
586
dai_link->cpus->of_node = cpu_dai_node[dai_index];
sound/soc/sdw_utils/soc_sdw_utils.c
1295
struct snd_soc_dai_link_component *cpus, int cpus_num,
sound/soc/sdw_utils/soc_sdw_utils.c
1309
dai_links->cpus = cpus;
sound/soc/soc-core.c
1502
ext_fmt = rtd->dai_link->cpus[i].ext_fmt;
sound/soc/soc-core.c
3198
struct snd_soc_dai_link_component *cpus)
sound/soc/soc-core.c
3200
platforms->of_node = cpus->of_node;
sound/soc/soc-core.c
3201
platforms->dai_args = cpus->dai_args;
sound/soc/soc-core.c
3795
&dai_link->cpus, &dai_link->num_cpus);
sound/soc/soc-core.c
3808
dai_link->cpus = NULL;
sound/soc/soc-core.c
599
if (dai_link->num_cpus == 0 && dai_link->cpus) {
sound/soc/soc-core.c
601
dai_link->cpus = &snd_soc_dummy_dlc;
sound/soc/soc-topology.c
1476
link->cpus = &dlc[0];
sound/soc/soc-topology.c
1495
link->cpus->dai_name = devm_kstrdup(tplg->dev, pcm->dai_name, GFP_KERNEL);
sound/soc/soc-topology.c
1496
if (!link->cpus->dai_name) {
sound/soc/sof/nocodec.c
46
links[i].cpus = &dlc[0];
sound/soc/sof/nocodec.c
56
links[i].cpus->dai_name = drv[i].name;
sound/soc/sof/sof-client-probes.c
444
struct snd_soc_dai_link_component *cpus;
sound/soc/sof/sof-client-probes.c
514
cpus = devm_kcalloc(dev, SOF_PROBES_NUM_DAI_LINKS, sizeof(*cpus), GFP_KERNEL);
sound/soc/sof/sof-client-probes.c
515
if (!links || !cpus) {
sound/soc/sof/sof-client-probes.c
524
links[0].cpus = &cpus[0];
sound/soc/sof/sof-client-probes.c
526
links[0].cpus->dai_name = "Probe Extraction CPU DAI";
sound/soc/sunxi/sun4i-codec.c
1641
link->cpus = &dlc[0];
sound/soc/sunxi/sun4i-codec.c
1652
link->cpus->dai_name = dev_name(dev);
sound/soc/tegra/tegra_asoc_machine.c
498
card->dai_link->cpus->of_node = np_ac97;
sound/soc/tegra/tegra_asoc_machine.c
509
card->dai_link->cpus->of_node = np_i2s;
sound/soc/ti/davinci-evm.c
196
dai->cpus->of_node = of_parse_phandle(np, "ti,mcasp-controller", 0);
sound/soc/ti/davinci-evm.c
197
if (!dai->cpus->of_node) {
sound/soc/ti/davinci-evm.c
202
dai->platforms->of_node = dai->cpus->of_node;
sound/soc/ti/davinci-evm.c
259
if (dai->cpus->of_node) {
sound/soc/ti/davinci-evm.c
260
of_node_put(dai->cpus->of_node);
sound/soc/ti/davinci-evm.c
261
dai->cpus->of_node = NULL;
sound/soc/ti/j721e-evm.c
662
priv->dai_links[*link_idx].cpus = &compnent[comp_idx++];
sound/soc/ti/j721e-evm.c
671
priv->dai_links[*link_idx].cpus->of_node = dai_node;
sound/soc/ti/j721e-evm.c
682
priv->dai_links[*link_idx].cpus = &compnent[comp_idx++];
sound/soc/ti/j721e-evm.c
691
priv->dai_links[*link_idx].cpus->of_node = dai_node;
sound/soc/ti/j721e-evm.c
776
priv->dai_links[*link_idx].cpus = &compnent[comp_idx++];
sound/soc/ti/j721e-evm.c
786
priv->dai_links[*link_idx].cpus->of_node = dai_node;
sound/soc/ti/j721e-evm.c
799
priv->dai_links[*link_idx].cpus = &compnent[comp_idx++];
sound/soc/ti/j721e-evm.c
808
priv->dai_links[*link_idx].cpus->of_node = dai_node;
sound/soc/ti/omap-abe-twl6040.c
256
priv->dai_links[0].cpus = link0_cpus;
sound/soc/ti/omap-abe-twl6040.c
258
priv->dai_links[0].cpus->of_node = dai_node;
sound/soc/ti/omap-abe-twl6040.c
272
priv->dai_links[1].cpus = link1_cpus;
sound/soc/ti/omap-abe-twl6040.c
274
priv->dai_links[1].cpus->of_node = dai_node;
sound/soc/ti/omap-hdmi.c
367
card->dai_link->cpus = &compnent[0];
sound/soc/ti/omap-hdmi.c
376
card->dai_link->cpus->dai_name = dev_name(ad->dssdev);
sound/soc/ti/omap-twl4030.c
266
omap_twl4030_dai_links[0].cpus->dai_name = NULL;
sound/soc/ti/omap-twl4030.c
267
omap_twl4030_dai_links[0].cpus->of_node = dai_node;
sound/soc/ti/omap-twl4030.c
276
omap_twl4030_dai_links[1].cpus->dai_name = NULL;
sound/soc/ti/omap-twl4030.c
277
omap_twl4030_dai_links[1].cpus->of_node = dai_node;
sound/soc/ti/rx51.c
380
rx51_dai[0].cpus->dai_name = NULL;
sound/soc/ti/rx51.c
382
rx51_dai[0].cpus->of_node = dai_node;
sound/soc/ux500/mop500.c
65
of_node_put(mop500_dai_links[i].cpus->of_node);
sound/soc/ux500/mop500.c
90
mop500_dai_links[i].cpus->of_node = msp_np[i];
sound/soc/ux500/mop500.c
91
mop500_dai_links[i].cpus->dai_name = NULL;
tools/bpf/bpftool/common.c
649
int cpus = libbpf_num_possible_cpus();
tools/bpf/bpftool/common.c
651
if (cpus < 0) {
tools/bpf/bpftool/common.c
652
p_err("Can't get # of possible cpus: %s", strerror(-cpus));
tools/bpf/bpftool/common.c
655
return cpus;
tools/bpf/bpftool/map_perf_ring.c
188
opts.cpus = &ctx.cpu;
tools/lib/bpf/libbpf.c
13800
p.cpus = OPTS_GET(opts, cpus, NULL);
tools/lib/bpf/libbpf.c
13902
cpu = p->cpu_cnt > 0 ? p->cpus[i] : i;
tools/lib/bpf/libbpf.c
14280
static int cpus;
tools/lib/bpf/libbpf.c
14284
tmp_cpus = READ_ONCE(cpus);
tools/lib/bpf/libbpf.c
14299
WRITE_ONCE(cpus, tmp_cpus);
tools/lib/perf/cpumap.c
101
return cpus;
tools/lib/perf/cpumap.c
106
struct perf_cpu_map *cpus = NULL;
tools/lib/perf/cpumap.c
111
cpus = perf_cpu_map__new(buf);
tools/lib/perf/cpumap.c
114
return cpus;
tools/lib/perf/cpumap.c
119
struct perf_cpu_map *cpus = cpu_map__new_sysfs_online();
tools/lib/perf/cpumap.c
121
if (cpus)
tools/lib/perf/cpumap.c
122
return cpus;
tools/lib/perf/cpumap.c
135
static struct perf_cpu __perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx)
tools/lib/perf/cpumap.c
137
return RC_CHK_ACCESS(cpus)->map[idx];
tools/lib/perf/cpumap.c
143
struct perf_cpu_map *cpus = perf_cpu_map__alloc(nr_cpus);
tools/lib/perf/cpumap.c
146
if (cpus != NULL) {
tools/lib/perf/cpumap.c
147
memcpy(RC_CHK_ACCESS(cpus)->map, tmp_cpus, payload_size);
tools/lib/perf/cpumap.c
148
qsort(RC_CHK_ACCESS(cpus)->map, nr_cpus, sizeof(struct perf_cpu), cmp_cpu);
tools/lib/perf/cpumap.c
153
__perf_cpu_map__cpu(cpus, i).cpu !=
tools/lib/perf/cpumap.c
154
__perf_cpu_map__cpu(cpus, i - 1).cpu) {
tools/lib/perf/cpumap.c
155
RC_CHK_ACCESS(cpus)->map[j++].cpu =
tools/lib/perf/cpumap.c
156
__perf_cpu_map__cpu(cpus, i).cpu;
tools/lib/perf/cpumap.c
159
perf_cpu_map__set_nr(cpus, j);
tools/lib/perf/cpumap.c
162
return cpus;
tools/lib/perf/cpumap.c
167
struct perf_cpu_map *cpus = NULL;
tools/lib/perf/cpumap.c
231
cpus = cpu_map__trim_new(nr_cpus, tmp_cpus);
tools/lib/perf/cpumap.c
235
cpus = perf_cpu_map__new_online_cpus();
tools/lib/perf/cpumap.c
237
cpus = perf_cpu_map__new_any_cpu();
tools/lib/perf/cpumap.c
242
return cpus;
tools/lib/perf/cpumap.c
247
struct perf_cpu_map *cpus = perf_cpu_map__alloc(1);
tools/lib/perf/cpumap.c
249
if (cpus)
tools/lib/perf/cpumap.c
25
RC_STRUCT(perf_cpu_map) *cpus;
tools/lib/perf/cpumap.c
250
RC_CHK_ACCESS(cpus)->map[0].cpu = cpu;
tools/lib/perf/cpumap.c
252
return cpus;
tools/lib/perf/cpumap.c
255
static int __perf_cpu_map__nr(const struct perf_cpu_map *cpus)
tools/lib/perf/cpumap.c
257
return RC_CHK_ACCESS(cpus)->nr;
tools/lib/perf/cpumap.c
260
struct perf_cpu perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx)
tools/lib/perf/cpumap.c
266
if (cpus && idx < __perf_cpu_map__nr(cpus))
tools/lib/perf/cpumap.c
267
return __perf_cpu_map__cpu(cpus, idx);
tools/lib/perf/cpumap.c
272
int perf_cpu_map__nr(const struct perf_cpu_map *cpus)
tools/lib/perf/cpumap.c
274
return cpus ? __perf_cpu_map__nr(cpus) : 1;
tools/lib/perf/cpumap.c
295
int perf_cpu_map__idx(const struct perf_cpu_map *cpus, struct perf_cpu cpu)
tools/lib/perf/cpumap.c
299
if (!cpus)
tools/lib/perf/cpumap.c
303
high = __perf_cpu_map__nr(cpus);
tools/lib/perf/cpumap.c
306
struct perf_cpu cpu_at_idx = __perf_cpu_map__cpu(cpus, idx);
tools/lib/perf/cpumap.c
31
cpus = malloc(sizeof(*cpus) + sizeof(struct perf_cpu) * nr_cpus);
tools/lib/perf/cpumap.c
32
if (ADD_RC_CHK(result, cpus)) {
tools/lib/perf/cpumap.c
320
bool perf_cpu_map__has(const struct perf_cpu_map *cpus, struct perf_cpu cpu)
tools/lib/perf/cpumap.c
322
return perf_cpu_map__idx(cpus, cpu) != -1;
tools/lib/perf/cpumap.c
33
cpus->nr = nr_cpus;
tools/lib/perf/cpumap.c
34
refcount_set(&cpus->refcnt, 1);
tools/lib/perf/cpumap.c
41
struct perf_cpu_map *cpus = perf_cpu_map__alloc(1);
tools/lib/perf/cpumap.c
43
if (cpus)
tools/lib/perf/cpumap.c
44
RC_CHK_ACCESS(cpus)->map[0].cpu = -1;
tools/lib/perf/cpumap.c
46
return cpus;
tools/lib/perf/cpumap.c
80
struct perf_cpu_map *cpus;
tools/lib/perf/cpumap.c
93
cpus = perf_cpu_map__alloc(nr_cpus);
tools/lib/perf/cpumap.c
94
if (cpus != NULL) {
tools/lib/perf/cpumap.c
98
RC_CHK_ACCESS(cpus)->map[i].cpu = i;
tools/lib/perf/evlist.c
100
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evlist.c
101
evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus);
tools/lib/perf/evlist.c
110
if (evsel->reads_only_on_cpu_idx0 && perf_cpu_map__nr(evsel->cpus) > 0) {
tools/lib/perf/evlist.c
120
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evlist.c
121
evsel->cpus = perf_cpu_map__new_int(perf_cpu_map__cpu(srcs[i], 0).cpu);
tools/lib/perf/evlist.c
127
assert(!evsel->requires_cpu || !perf_cpu_map__has_any_cpu(evsel->cpus));
tools/lib/perf/evlist.c
133
if (perf_cpu_map__is_empty(evsel->cpus)) {
tools/lib/perf/evlist.c
153
perf_cpu_map__merge(&evlist->all_cpus, evsel->cpus);
tools/lib/perf/evlist.c
260
struct perf_cpu_map *cpus,
tools/lib/perf/evlist.c
270
if (cpus != evlist->user_requested_cpus) {
tools/lib/perf/evlist.c
272
evlist->user_requested_cpus = perf_cpu_map__get(cpus);
tools/lib/perf/evlist.c
289
err = perf_evsel__open(evsel, evsel->cpus, evsel->threads);
tools/lib/perf/evlist.c
39
if (perf_cpu_map__is_empty(evsel->cpus)) {
tools/lib/perf/evlist.c
496
sid->cpu = perf_cpu_map__cpu(evsel->cpus, cpu);
tools/lib/perf/evlist.c
52
evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus);
tools/lib/perf/evlist.c
557
cpu = perf_cpu_map__idx(evsel->cpus, evlist_cpu);
tools/lib/perf/evlist.c
58
evsel->cpus = perf_cpu_map__get(evsel->pmu_cpus);
tools/lib/perf/evlist.c
65
if (evsel->requires_cpu && perf_cpu_map__has_any_cpu(evsel->cpus)) {
tools/lib/perf/evlist.c
66
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evlist.c
67
evsel->cpus = perf_cpu_map__get(evsel->pmu_cpus);
tools/lib/perf/evlist.c
727
const struct perf_cpu_map *cpus = evlist->all_cpus;
tools/lib/perf/evlist.c
747
if (perf_cpu_map__has_any_cpu_or_is_empty(cpus))
tools/lib/perf/evlist.c
76
if (!perf_cpu_map__equal(evsel->cpus, evlist->user_requested_cpus)) {
tools/lib/perf/evlist.c
77
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evlist.c
78
evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus);
tools/lib/perf/evlist.c
83
if (!perf_cpu_map__has_any_cpu(evsel->cpus) &&
tools/lib/perf/evlist.c
84
!perf_cpu_map__is_subset(evsel->pmu_cpus, evsel->cpus)) {
tools/lib/perf/evlist.c
85
struct perf_cpu_map *tmp = perf_cpu_map__intersect(evsel->pmu_cpus, evsel->cpus);
tools/lib/perf/evlist.c
87
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evlist.c
88
evsel->cpus = tmp;
tools/lib/perf/evlist.c
98
perf_cpu_map__equal(evsel->cpus, evsel->pmu_cpus) &&
tools/lib/perf/evsel.c
126
int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
tools/lib/perf/evsel.c
132
if (cpus == NULL) {
tools/lib/perf/evsel.c
141
cpus = empty_cpu_map;
tools/lib/perf/evsel.c
157
perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0)
tools/lib/perf/evsel.c
160
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/lib/perf/evsel.c
276
struct perf_cpu cpu = perf_cpu_map__cpu(evsel->cpus, idx);
tools/lib/perf/evsel.c
466
perf_cpu_map__for_each_cpu(cpu, idx, evsel->cpus) {
tools/lib/perf/evsel.c
48
perf_cpu_map__put(evsel->cpus);
tools/lib/perf/evsel.c
504
for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++)
tools/lib/perf/evsel.c
513
return evsel->cpus;
tools/lib/perf/include/internal/cpumap.h
25
int perf_cpu_map__idx(const struct perf_cpu_map *cpus, struct perf_cpu cpu);
tools/lib/perf/include/internal/evsel.h
95
struct perf_cpu_map *cpus;
tools/lib/perf/include/perf/cpumap.h
101
#define perf_cpu_map__for_each_idx(idx, cpus) \
tools/lib/perf/include/perf/cpumap.h
102
for ((idx) = 0; (idx) < perf_cpu_map__nr(cpus); (idx)++)
tools/lib/perf/include/perf/cpumap.h
52
LIBPERF_API struct perf_cpu perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx);
tools/lib/perf/include/perf/cpumap.h
60
LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus);
tools/lib/perf/include/perf/cpumap.h
90
#define perf_cpu_map__for_each_cpu(cpu, idx, cpus) \
tools/lib/perf/include/perf/cpumap.h
91
for ((idx) = 0, (cpu) = perf_cpu_map__cpu(cpus, idx); \
tools/lib/perf/include/perf/cpumap.h
92
(idx) < perf_cpu_map__nr(cpus); \
tools/lib/perf/include/perf/cpumap.h
93
(idx)++, (cpu) = perf_cpu_map__cpu(cpus, idx))
tools/lib/perf/include/perf/cpumap.h
95
#define perf_cpu_map__for_each_cpu_skip_any(_cpu, idx, cpus) \
tools/lib/perf/include/perf/cpumap.h
96
for ((idx) = 0, (_cpu) = perf_cpu_map__cpu(cpus, idx); \
tools/lib/perf/include/perf/cpumap.h
97
(idx) < perf_cpu_map__nr(cpus); \
tools/lib/perf/include/perf/cpumap.h
98
(idx)++, (_cpu) = perf_cpu_map__cpu(cpus, idx)) \
tools/lib/perf/include/perf/event.h
268
struct perf_record_cpu_map_data cpus;
tools/lib/perf/include/perf/event.h
287
struct perf_record_event_update_cpus cpus;
tools/lib/perf/include/perf/evlist.h
32
struct perf_cpu_map *cpus,
tools/lib/perf/include/perf/evsel.h
30
LIBPERF_API int perf_evsel__open(struct perf_evsel *evsel, struct perf_cpu_map *cpus,
tools/lib/perf/tests/test-cpumap.c
16
struct perf_cpu_map *cpus;
tools/lib/perf/tests/test-cpumap.c
24
cpus = perf_cpu_map__new_any_cpu();
tools/lib/perf/tests/test-cpumap.c
25
if (!cpus)
tools/lib/perf/tests/test-cpumap.c
28
perf_cpu_map__get(cpus);
tools/lib/perf/tests/test-cpumap.c
29
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-cpumap.c
30
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-cpumap.c
32
cpus = perf_cpu_map__new_online_cpus();
tools/lib/perf/tests/test-cpumap.c
33
if (!cpus)
tools/lib/perf/tests/test-cpumap.c
36
perf_cpu_map__for_each_cpu(cpu, idx, cpus)
tools/lib/perf/tests/test-cpumap.c
39
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-evlist.c
218
struct perf_cpu_map *cpus;
tools/lib/perf/tests/test-evlist.c
264
cpus = perf_cpu_map__new_any_cpu();
tools/lib/perf/tests/test-evlist.c
265
__T("failed to create cpus", cpus);
tools/lib/perf/tests/test-evlist.c
278
perf_evlist__set_maps(evlist, cpus, threads);
tools/lib/perf/tests/test-evlist.c
313
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-evlist.c
329
struct perf_cpu_map *cpus;
tools/lib/perf/tests/test-evlist.c
353
cpus = perf_cpu_map__new_online_cpus();
tools/lib/perf/tests/test-evlist.c
354
__T("failed to create cpus", cpus);
tools/lib/perf/tests/test-evlist.c
36
struct perf_cpu_map *cpus;
tools/lib/perf/tests/test-evlist.c
365
perf_evlist__set_maps(evlist, cpus, NULL);
tools/lib/perf/tests/test-evlist.c
378
perf_cpu_map__for_each_cpu(cpu, tmp, cpus) {
tools/lib/perf/tests/test-evlist.c
414
__T("failed count", count >= perf_cpu_map__nr(cpus));
tools/lib/perf/tests/test-evlist.c
416
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-evlist.c
49
cpus = perf_cpu_map__new_online_cpus();
tools/lib/perf/tests/test-evlist.c
50
__T("failed to create cpus", cpus);
tools/lib/perf/tests/test-evlist.c
69
perf_evlist__set_maps(evlist, cpus, NULL);
tools/lib/perf/tests/test-evlist.c
75
cpus = perf_evsel__cpus(evsel);
tools/lib/perf/tests/test-evlist.c
77
for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) {
tools/lib/perf/tests/test-evlist.c
88
perf_cpu_map__put(cpus);
tools/lib/perf/tests/test-evsel.c
22
struct perf_cpu_map *cpus;
tools/lib/perf/tests/test-evsel.c
30
cpus = perf_cpu_map__new_online_cpus();
tools/lib/perf/tests/test-evsel.c
31
__T("failed to create cpus", cpus);
tools/lib/perf/tests/test-evsel.c
36
err = perf_evsel__open(evsel, cpus, NULL);
tools/lib/perf/tests/test-evsel.c
39
for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) {
tools/lib/perf/tests/test-evsel.c
49
perf_cpu_map__put(cpus);
tools/perf/arch/arm/util/cs-etm.c
325
struct perf_cpu_map *cpus = evlist->core.user_requested_cpus;
tools/perf/arch/arm/util/cs-etm.c
454
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus)) {
tools/perf/arch/arm/util/cs-etm.c
477
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus))
tools/perf/arch/arm/util/pmu.c
42
intersect = perf_cpu_map__intersect(online, pmu->cpus);
tools/perf/arch/arm/util/pmu.c
44
perf_cpu_map__put(pmu->cpus);
tools/perf/arch/arm/util/pmu.c
45
pmu->cpus = intersect;
tools/perf/arch/arm64/util/arm-spe.c
118
if (perf_cpu_map__has(sper->arm_spe_pmu->cpus, cpu))
tools/perf/arch/arm64/util/arm-spe.c
257
static void arm_spe_setup_evsel(struct evsel *evsel, struct perf_cpu_map *cpus)
tools/perf/arch/arm64/util/arm-spe.c
275
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus)) {
tools/perf/arch/arm64/util/arm-spe.c
365
struct perf_cpu_map *cpus = evlist->core.user_requested_cpus;
tools/perf/arch/arm64/util/arm-spe.c
379
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus)) {
tools/perf/arch/arm64/util/arm-spe.c
398
struct perf_cpu_map *cpus = evlist->core.user_requested_cpus;
tools/perf/arch/arm64/util/arm-spe.c
429
arm_spe_setup_evsel(evsel, cpus);
tools/perf/arch/arm64/util/header.c
45
struct perf_cpu_map *cpus;
tools/perf/arch/arm64/util/header.c
51
cpus = perf_cpu_map__new_online_cpus();
tools/perf/arch/arm64/util/header.c
52
if (!cpus)
tools/perf/arch/arm64/util/header.c
55
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/perf/arch/x86/util/intel-bts.c
113
const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus;
tools/perf/arch/x86/util/intel-bts.c
146
if (opts->full_auxtrace && !perf_cpu_map__is_any_cpu_or_is_empty(cpus)) {
tools/perf/arch/x86/util/intel-bts.c
227
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus))
tools/perf/arch/x86/util/intel-pt.c
621
const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus;
tools/perf/arch/x86/util/intel-pt.c
792
if (have_timing_info && !perf_cpu_map__is_any_cpu_or_is_empty(cpus) &&
tools/perf/arch/x86/util/intel-pt.c
850
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus))
tools/perf/arch/x86/util/intel-pt.c
876
if (!perf_cpu_map__is_any_cpu_or_is_empty(cpus)) {
tools/perf/arch/x86/util/intel-pt.c
888
if (!ptr->have_sched_switch && !perf_cpu_map__is_any_cpu_or_is_empty(cpus) &&
tools/perf/arch/x86/util/pmu.c
233
if (perf_cpu_map__cpu(pmu->cpus, 0).cpu != 0) {
tools/perf/arch/x86/util/pmu.c
248
adjusted[pmu_snc] = perf_cpu_map__empty_new(perf_cpu_map__nr(pmu->cpus));
tools/perf/arch/x86/util/pmu.c
253
perf_cpu_map__for_each_cpu(cpu, idx, pmu->cpus) {
tools/perf/arch/x86/util/pmu.c
268
perf_cpu_map__put(pmu->cpus);
tools/perf/arch/x86/util/pmu.c
269
pmu->cpus = perf_cpu_map__get(adjusted[pmu_snc]);
tools/perf/arch/x86/util/pmu.c
43
struct perf_cpu_map *cpus;
tools/perf/arch/x86/util/pmu.c
50
cpus = perf_cpu_map__new(buf);
tools/perf/arch/x86/util/pmu.c
52
return cpus;
tools/perf/arch/x86/util/topdown.c
95
evsel->core.cpus = perf_cpu_map__get(metric_event->core.cpus);
tools/perf/bench/evlist-open-close.c
56
cnt += evsel->core.threads->nr * perf_cpu_map__nr(evsel->core.cpus);
tools/perf/builtin-record.c
1064
struct perf_cpu_map *cpus = evlist->core.all_cpus;
tools/perf/builtin-record.c
1089
test_bit(perf_cpu_map__cpu(cpus, m).cpu, thread_data->mask->maps.bits)) {
tools/perf/builtin-record.c
1093
thread_data, perf_cpu_map__cpu(cpus, m).cpu, tm, m);
tools/perf/builtin-record.c
1098
thread_data, perf_cpu_map__cpu(cpus, m).cpu, tm, m);
tools/perf/builtin-record.c
1371
if (evsel__open(pos, pos->core.cpus, pos->core.threads) < 0) {
tools/perf/builtin-record.c
3696
static int record__mmap_cpu_mask_init(struct mmap_cpu_mask *mask, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
3701
if (cpu_map__is_dummy(cpus))
tools/perf/builtin-record.c
3704
perf_cpu_map__for_each_cpu_skip_any(cpu, idx, cpus) {
tools/perf/builtin-record.c
3716
struct perf_cpu_map *cpus;
tools/perf/builtin-record.c
3718
cpus = perf_cpu_map__new(mask_spec);
tools/perf/builtin-record.c
3719
if (!cpus)
tools/perf/builtin-record.c
3723
if (record__mmap_cpu_mask_init(mask, cpus))
tools/perf/builtin-record.c
3726
perf_cpu_map__put(cpus);
tools/perf/builtin-record.c
3768
static int record__init_thread_cpu_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
3770
int t, ret, nr_cpus = perf_cpu_map__nr(cpus);
tools/perf/builtin-record.c
3780
__set_bit(perf_cpu_map__cpu(cpus, t).cpu, rec->thread_masks[t].maps.bits);
tools/perf/builtin-record.c
3781
__set_bit(perf_cpu_map__cpu(cpus, t).cpu, rec->thread_masks[t].affinity.bits);
tools/perf/builtin-record.c
3793
static int record__init_thread_masks_spec(struct record *rec, struct perf_cpu_map *cpus,
tools/perf/builtin-record.c
3808
ret = record__mmap_cpu_mask_init(&cpus_mask, cpus);
tools/perf/builtin-record.c
3907
static int record__init_thread_core_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
3918
ret = record__init_thread_masks_spec(rec, cpus, topo->core_cpus_list,
tools/perf/builtin-record.c
3925
static int record__init_thread_package_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
3936
ret = record__init_thread_masks_spec(rec, cpus, topo->package_cpus_list,
tools/perf/builtin-record.c
3943
static int record__init_thread_numa_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
3963
spec[s] = topo->nodes[s].cpus;
tools/perf/builtin-record.c
3965
ret = record__init_thread_masks_spec(rec, cpus, spec, spec, topo->nr);
tools/perf/builtin-record.c
3975
static int record__init_thread_user_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
4028
ret = record__init_thread_masks_spec(rec, cpus, (const char **)maps_spec,
tools/perf/builtin-record.c
4045
static int record__init_thread_default_masks(struct record *rec, struct perf_cpu_map *cpus)
tools/perf/builtin-record.c
4053
if (record__mmap_cpu_mask_init(&rec->thread_masks->maps, cpus))
tools/perf/builtin-record.c
4064
struct perf_cpu_map *cpus = rec->evlist->core.all_cpus;
tools/perf/builtin-record.c
4067
return record__init_thread_default_masks(rec, cpus);
tools/perf/builtin-record.c
4076
ret = record__init_thread_cpu_masks(rec, cpus);
tools/perf/builtin-record.c
4079
ret = record__init_thread_core_masks(rec, cpus);
tools/perf/builtin-record.c
4082
ret = record__init_thread_package_masks(rec, cpus);
tools/perf/builtin-record.c
4085
ret = record__init_thread_numa_masks(rec, cpus);
tools/perf/builtin-record.c
4088
ret = record__init_thread_user_masks(rec, cpus);
tools/perf/builtin-sched.c
165
struct perf_cpu_map *cpus;
tools/perf/builtin-sched.c
1723
if (sched->map.cpus && !perf_cpu_map__has(sched->map.cpus, this_cpu))
tools/perf/builtin-sched.c
3568
sched->map.cpus = perf_cpu_map__new(sched->map.cpus_str);
tools/perf/builtin-sched.c
3569
if (!sched->map.cpus) {
tools/perf/builtin-sched.c
3652
perf_cpu_map__put(sched->map.cpus);
tools/perf/builtin-script.c
160
struct perf_cpu_map *cpus;
tools/perf/builtin-script.c
2331
cpu_map_idx = perf_cpu_map__idx(evsel->core.cpus, (struct perf_cpu){sample->cpu});
tools/perf/builtin-script.c
2334
if (perf_cpu_map__cpu(evsel->core.cpus, /*idx=*/0).cpu == -1 ||
tools/perf/builtin-script.c
3184
perf_cpu_map__put(script->cpus);
tools/perf/builtin-script.c
3883
if (!script->cpus || !script->threads)
tools/perf/builtin-script.c
3889
perf_evlist__set_maps(&evlist->core, script->cpus, script->threads);
tools/perf/builtin-script.c
3930
if (script->cpus) {
tools/perf/builtin-script.c
3935
script->cpus = cpu_map__new_data(&event->cpu_map.data);
tools/perf/builtin-script.c
3936
if (!script->cpus)
tools/perf/builtin-stat.c
150
struct perf_cpu_map *cpus;
tools/perf/builtin-stat.c
200
if (perf_cpu_map__equal(leader->core.cpus, evsel->core.cpus))
tools/perf/builtin-stat.c
217
cpu_map__snprint(leader->core.cpus, buf, sizeof(buf));
tools/perf/builtin-stat.c
219
cpu_map__snprint(evsel->core.cpus, buf, sizeof(buf));
tools/perf/builtin-stat.c
2215
if (perf_cpu_map__is_empty(st->cpus)) {
tools/perf/builtin-stat.c
2240
if (!st->cpus || !st->threads)
tools/perf/builtin-stat.c
2246
perf_evlist__set_maps(&evsel_list->core, st->cpus, st->threads);
tools/perf/builtin-stat.c
2280
struct perf_cpu_map *cpus;
tools/perf/builtin-stat.c
2282
if (st->cpus) {
tools/perf/builtin-stat.c
2287
cpus = cpu_map__new_data(&event->cpu_map.data);
tools/perf/builtin-stat.c
2288
if (!cpus)
tools/perf/builtin-stat.c
2291
st->cpus = cpus;
tools/perf/builtin-stat.c
418
perf_cpu_map__for_each_idx(idx, counter->core.cpus) {
tools/perf/builtin-top.c
1031
if (evsel__open(counter, counter->core.cpus,
tools/perf/pmu-events/empty-pmu-events.c
3217
cpu = perf_cpu_map__min(pmu->cpus);
tools/perf/tests/code-reading.c
699
struct perf_cpu_map *cpus = NULL;
tools/perf/tests/code-reading.c
765
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/code-reading.c
766
if (!cpus) {
tools/perf/tests/code-reading.c
780
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/code-reading.c
847
perf_cpu_map__put(cpus);
tools/perf/tests/cpumap.c
105
struct perf_cpu_map *cpus;
tools/perf/tests/cpumap.c
108
cpus = perf_cpu_map__new("0,2-20");
tools/perf/tests/cpumap.c
111
!perf_event__synthesize_cpu_map(NULL, cpus, process_event_mask, NULL));
tools/perf/tests/cpumap.c
113
perf_cpu_map__put(cpus);
tools/perf/tests/cpumap.c
116
cpus = perf_cpu_map__new("1,256");
tools/perf/tests/cpumap.c
119
!perf_event__synthesize_cpu_map(NULL, cpus, process_event_cpus, NULL));
tools/perf/tests/cpumap.c
121
perf_cpu_map__put(cpus);
tools/perf/tests/cpumap.c
124
cpus = perf_cpu_map__new("1-256");
tools/perf/tests/cpumap.c
127
!perf_event__synthesize_cpu_map(NULL, cpus, process_event_range_cpus, NULL));
tools/perf/tests/cpumap.c
129
perf_cpu_map__put(cpus);
tools/perf/tests/event-times.c
113
struct perf_cpu_map *cpus;
tools/perf/tests/event-times.c
118
cpus = perf_cpu_map__new("0");
tools/perf/tests/event-times.c
119
if (cpus == NULL) {
tools/perf/tests/event-times.c
126
err = evsel__open_per_cpu(evsel, cpus, -1);
tools/perf/tests/event-times.c
127
perf_cpu_map__put(cpus);
tools/perf/tests/event-times.c
142
struct perf_cpu_map *cpus;
tools/perf/tests/event-times.c
147
cpus = perf_cpu_map__new("0");
tools/perf/tests/event-times.c
148
if (cpus == NULL) {
tools/perf/tests/event-times.c
153
err = evsel__open_per_cpu(evsel, cpus, -1);
tools/perf/tests/event-times.c
154
perf_cpu_map__put(cpus);
tools/perf/tests/event_update.c
68
map = cpu_map__new_data(&ev->cpus.cpus);
tools/perf/tests/keep-tracking.c
158
perf_cpu_map__put(cpus);
tools/perf/tests/keep-tracking.c
75
struct perf_cpu_map *cpus = NULL;
tools/perf/tests/keep-tracking.c
84
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/keep-tracking.c
85
CHECK_NOT_NULL__(cpus);
tools/perf/tests/keep-tracking.c
90
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/mmap-basic.c
101
if (evsel__open(evsels[i], cpus, threads) < 0) {
tools/perf/tests/mmap-basic.c
172
perf_cpu_map__put(cpus);
tools/perf/tests/mmap-basic.c
261
if (pmu->cpus != NULL)
tools/perf/tests/mmap-basic.c
262
cpu_map__set_affinity(pmu->cpus);
tools/perf/tests/mmap-basic.c
363
if (pmu->cpus != NULL) {
tools/perf/tests/mmap-basic.c
364
struct perf_cpu_map *cpus = cpu_map__online();
tools/perf/tests/mmap-basic.c
366
cpu_map__set_affinity(cpus);
tools/perf/tests/mmap-basic.c
367
perf_cpu_map__put(cpus);
tools/perf/tests/mmap-basic.c
40
struct perf_cpu_map *cpus;
tools/perf/tests/mmap-basic.c
58
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/mmap-basic.c
59
if (cpus == NULL) {
tools/perf/tests/mmap-basic.c
65
CPU_SET(perf_cpu_map__cpu(cpus, 0).cpu, &cpu_set);
tools/perf/tests/mmap-basic.c
69
perf_cpu_map__cpu(cpus, 0).cpu,
tools/perf/tests/mmap-basic.c
80
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/openat-syscall-all-cpus.c
121
perf_cpu_map__put(cpus);
tools/perf/tests/openat-syscall-all-cpus.c
27
struct perf_cpu_map *cpus;
tools/perf/tests/openat-syscall-all-cpus.c
40
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/openat-syscall-all-cpus.c
41
if (cpus == NULL) {
tools/perf/tests/openat-syscall-all-cpus.c
56
if (evsel__open(evsel, cpus, threads) < 0) {
tools/perf/tests/openat-syscall-all-cpus.c
64
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/perf/tests/openat-syscall-all-cpus.c
91
evsel->core.cpus = perf_cpu_map__get(cpus);
tools/perf/tests/openat-syscall-all-cpus.c
95
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/perf/tests/parse-metric.c
115
perf_cpu_map__put(cpus);
tools/perf/tests/parse-metric.c
73
struct perf_cpu_map *cpus;
tools/perf/tests/parse-metric.c
85
cpus = perf_cpu_map__new("0");
tools/perf/tests/parse-metric.c
86
if (!cpus) {
tools/perf/tests/parse-metric.c
91
perf_evlist__set_maps(&evlist->core, cpus, NULL);
tools/perf/tests/perf-time-to-tsc.c
102
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/perf-time-to-tsc.c
205
perf_cpu_map__put(cpus);
tools/perf/tests/perf-time-to-tsc.c
80
struct perf_cpu_map *cpus = NULL;
tools/perf/tests/perf-time-to-tsc.c
96
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/perf-time-to-tsc.c
97
CHECK_NOT_NULL__(cpus);
tools/perf/tests/pmu-events.c
827
struct perf_cpu_map *cpus;
tools/perf/tests/pmu-events.c
845
cpus = perf_cpu_map__new("0");
tools/perf/tests/pmu-events.c
846
if (!cpus) {
tools/perf/tests/pmu-events.c
851
perf_evlist__set_maps(&evlist->core, cpus, NULL);
tools/perf/tests/pmu-events.c
901
perf_cpu_map__put(cpus);
tools/perf/tests/sw-clock.c
135
perf_cpu_map__put(cpus);
tools/perf/tests/sw-clock.c
47
struct perf_cpu_map *cpus = NULL;
tools/perf/tests/sw-clock.c
66
cpus = perf_cpu_map__new_any_cpu();
tools/perf/tests/sw-clock.c
68
if (!cpus || !threads) {
tools/perf/tests/sw-clock.c
74
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/switch-tracking.c
347
struct perf_cpu_map *cpus = NULL;
tools/perf/tests/switch-tracking.c
360
cpus = perf_cpu_map__new_online_cpus();
tools/perf/tests/switch-tracking.c
361
if (!cpus) {
tools/perf/tests/switch-tracking.c
372
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/tests/switch-tracking.c
582
perf_cpu_map__put(cpus);
tools/perf/tests/task-exit.c
148
perf_cpu_map__put(cpus);
tools/perf/tests/task-exit.c
53
struct perf_cpu_map *cpus;
tools/perf/tests/task-exit.c
72
cpus = perf_cpu_map__new_any_cpu();
tools/perf/tests/task-exit.c
74
if (!cpus || !threads) {
tools/perf/tests/task-exit.c
80
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/util/auxtrace.c
674
int cpu_map_idx = perf_cpu_map__idx(evsel->core.cpus, evlist_cpu);
tools/perf/util/bpf_counter.c
339
int cpu = perf_cpu_map__cpu(evsel->core.cpus, cpu_map_idx).cpu;
tools/perf/util/bpf_counter.c
499
assert(!perf_cpu_map__has_any_cpu_or_is_empty(evsel->core.cpus));
tools/perf/util/bpf_counter.c
502
evsel__open(evsel, evsel->core.cpus, threads);
tools/perf/util/bpf_counter.c
627
key = perf_cpu_map__cpu(evsel->core.cpus, i).cpu;
tools/perf/util/bpf_counter.c
658
int cpu = perf_cpu_map__cpu(evsel->core.cpus, cpu_map_idx).cpu;
tools/perf/util/bpf_counter.c
673
perf_cpu_map__for_each_cpu(cpu, idx, evsel->core.cpus)
tools/perf/util/bpf_counter_cgroup.c
153
err = evsel__open_per_cpu(evsel, evsel->core.cpus, -1);
tools/perf/util/bpf_counter_cgroup.c
158
perf_cpu_map__for_each_cpu(cpu, j, evsel->core.cpus) {
tools/perf/util/bpf_counter_cgroup.c
299
perf_cpu_map__for_each_cpu(cpu, i, evsel->core.cpus) {
tools/perf/util/bpf_off_cpu.c
91
perf_cpu_map__for_each_cpu(pcpu, i, evsel->core.cpus) {
tools/perf/util/bpf_trace_augment.c
72
perf_cpu_map__for_each_cpu(cpu, i, bpf_output->core.cpus) {
tools/perf/util/counts.c
60
struct perf_cpu_map *cpus = evsel__cpus(evsel);
tools/perf/util/counts.c
63
evsel->counts = perf_counts__new(perf_cpu_map__nr(cpus), nthreads);
tools/perf/util/cpumap.c
181
struct perf_cpu_map *cpus = perf_cpu_map__alloc(nr);
tools/perf/util/cpumap.c
183
if (cpus != NULL) {
tools/perf/util/cpumap.c
185
RC_CHK_ACCESS(cpus)->map[i].cpu = -1;
tools/perf/util/cpumap.c
188
return cpus;
tools/perf/util/cpumap.c
193
struct cpu_aggr_map *cpus = malloc(sizeof(*cpus) + sizeof(struct aggr_cpu_id) * nr);
tools/perf/util/cpumap.c
195
if (cpus != NULL) {
tools/perf/util/cpumap.c
198
cpus->nr = nr;
tools/perf/util/cpumap.c
200
cpus->map[i] = aggr_cpu_id__empty();
tools/perf/util/cpumap.c
203
return cpus;
tools/perf/util/cpumap.c
253
struct cpu_aggr_map *cpu_aggr_map__new(const struct perf_cpu_map *cpus,
tools/perf/util/cpumap.c
259
struct cpu_aggr_map *c = cpu_aggr_map__empty_new(perf_cpu_map__nr(cpus));
tools/perf/util/cpumap.c
267
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/perf/util/cpumap.c
283
if (c->nr != perf_cpu_map__nr(cpus)) {
tools/perf/util/cpumap.h
113
struct cpu_aggr_map *cpu_aggr_map__new(const struct perf_cpu_map *cpus,
tools/perf/util/cpumap.h
69
static inline bool cpu_map__is_dummy(const struct perf_cpu_map *cpus)
tools/perf/util/cpumap.h
71
return perf_cpu_map__nr(cpus) == 1 && perf_cpu_map__cpu(cpus, 0).cpu == -1;
tools/perf/util/cputopo.c
364
node->cpus = buf;
tools/perf/util/cputopo.c
432
zfree(&tp->nodes[i].cpus);
tools/perf/util/cputopo.c
462
node->cpus = buf;
tools/perf/util/cputopo.c
502
zfree(&tp->nodes[i].cpus);
tools/perf/util/cputopo.h
38
char *cpus;
tools/perf/util/cputopo.h
51
char *cpus;
tools/perf/util/drm_pmu.c
123
drm->pmu.cpus = perf_cpu_map__new_int(0);
tools/perf/util/drm_pmu.c
124
if (!drm->pmu.cpus) {
tools/perf/util/env.c
285
zfree(&env->hybrid_nodes[i].cpus);
tools/perf/util/env.h
44
char *cpus;
tools/perf/util/event.c
384
struct perf_cpu_map *cpus = cpu_map__new_data(&event->cpu_map.data);
tools/perf/util/event.c
389
if (cpus)
tools/perf/util/event.c
390
ret += cpu_map__fprintf(cpus, fp);
tools/perf/util/event.c
394
perf_cpu_map__put(cpus);
tools/perf/util/evlist.c
1050
struct perf_cpu_map *cpus;
tools/perf/util/evlist.c
1077
cpus = perf_cpu_map__new_any_cpu();
tools/perf/util/evlist.c
1079
cpus = perf_cpu_map__new(target->cpu_list);
tools/perf/util/evlist.c
1081
if (!cpus)
tools/perf/util/evlist.c
1086
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/util/evlist.c
1089
perf_cpu_map__put(cpus);
tools/perf/util/evlist.c
1409
struct perf_cpu_map *cpus;
tools/perf/util/evlist.c
1421
cpus = perf_cpu_map__new_online_cpus();
tools/perf/util/evlist.c
1422
if (!cpus)
tools/perf/util/evlist.c
1427
perf_cpu_map__put(cpus);
tools/perf/util/evlist.c
1431
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/util/evlist.c
1433
perf_cpu_map__put(cpus);
tools/perf/util/evlist.c
1455
err = evsel__open(evsel, evsel->core.cpus, evsel->core.threads);
tools/perf/util/evlist.c
399
used_cpus = perf_cpu_map__get(pos->core.cpus);
tools/perf/util/evlist.c
407
intersect = perf_cpu_map__intersect(used_cpus, pos->core.cpus);
tools/perf/util/evlist.c
415
perf_cpu_map__merge(&used_cpus, pos->core.cpus);
tools/perf/util/evlist.c
447
itr->cpu_map_idx = perf_cpu_map__idx(itr->evsel->core.cpus, itr->cpu);
tools/perf/util/evlist.c
470
perf_cpu_map__idx(evlist_cpu_itr->evsel->core.cpus,
tools/perf/util/evlist.c
484
perf_cpu_map__idx(evlist_cpu_itr->evsel->core.cpus,
tools/perf/util/evlist.c
76
void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus,
tools/perf/util/evlist.c
80
perf_evlist__set_maps(&evlist->core, cpus, threads);
tools/perf/util/evlist.h
109
void evlist__init(struct evlist *evlist, struct perf_cpu_map *cpus,
tools/perf/util/evsel.c
2123
cpu = perf_cpu_map__cpu(evsel->core.cpus, cpu_map_idx);
tools/perf/util/evsel.c
2124
return perf_cpu_map__idx(other->core.cpus, cpu);
tools/perf/util/evsel.c
2278
static int __evsel__prepare_open(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/evsel.c
2288
if (cpus == NULL) {
tools/perf/util/evsel.c
2295
cpus = empty_cpu_map;
tools/perf/util/evsel.c
2309
perf_evsel__alloc_fd(&evsel->core, perf_cpu_map__nr(cpus), nthreads) < 0)
tools/perf/util/evsel.c
2313
ret = evsel__tool_pmu_prepare_open(evsel, cpus, nthreads);
tools/perf/util/evsel.c
2366
int evsel__prepare_open(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/evsel.c
2371
err = __evsel__prepare_open(evsel, cpus, threads);
tools/perf/util/evsel.c
2782
static int evsel__open_cpu(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/evsel.c
2796
err = __evsel__prepare_open(evsel, cpus, threads);
tools/perf/util/evsel.c
2800
if (cpus == NULL)
tools/perf/util/evsel.c
2801
cpus = empty_cpu_map;
tools/perf/util/evsel.c
2837
cpu = perf_cpu_map__cpu(cpus, idx);
tools/perf/util/evsel.c
2916
if (evsel__ignore_missing_thread(evsel, perf_cpu_map__nr(cpus),
tools/perf/util/evsel.c
2958
int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/evsel.c
2961
return evsel__open_cpu(evsel, cpus, threads, 0, perf_cpu_map__nr(cpus));
tools/perf/util/evsel.c
2973
struct perf_cpu_map *cpus, int cpu_map_idx,
tools/perf/util/evsel.c
2977
return evsel__open_cpu(evsel, cpus, threads, 0, perf_cpu_map__nr(cpus));
tools/perf/util/evsel.c
2979
return evsel__open_cpu(evsel, cpus, threads, cpu_map_idx, cpu_map_idx + 1);
tools/perf/util/evsel.c
2982
int evsel__open_per_cpu(struct evsel *evsel, struct perf_cpu_map *cpus, int cpu_map_idx)
tools/perf/util/evsel.c
2985
int ret = evsel__open_per_cpu_and_thread(evsel, cpus, cpu_map_idx, threads);
tools/perf/util/evsel.c
2993
struct perf_cpu_map *cpus = perf_cpu_map__new_any_cpu();
tools/perf/util/evsel.c
2994
int ret = evsel__open_per_cpu_and_thread(evsel, cpus, -1, threads);
tools/perf/util/evsel.c
2996
perf_cpu_map__put(cpus);
tools/perf/util/evsel.c
4160
struct perf_cpu_map *cpus = evsel->core.cpus;
tools/perf/util/evsel.c
4163
if (perf_evsel__alloc_id(&evsel->core, perf_cpu_map__nr(cpus), threads->nr))
tools/perf/util/evsel.c
4390
intersect = perf_cpu_map__intersect(pmu->cpus, user_requested_cpus);
tools/perf/util/evsel.c
4400
cpu_map__snprint(online ?: pmu->cpus, buf2, sizeof(buf2));
tools/perf/util/evsel.c
493
evsel->core.cpus = perf_cpu_map__get(orig->core.cpus);
tools/perf/util/evsel.h
357
struct perf_cpu_map *cpus, int cpu_map_idx,
tools/perf/util/evsel.h
359
int evsel__open_per_cpu(struct evsel *evsel, struct perf_cpu_map *cpus, int cpu_map_idx);
tools/perf/util/evsel.h
361
int evsel__open(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/evsel.h
364
int evsel__prepare_open(struct evsel *evsel, struct perf_cpu_map *cpus,
tools/perf/util/header.c
1002
ret = do_write_string(ff, n->cpus);
tools/perf/util/header.c
1968
fprintf(fp, "# %s cpu list : %s\n", n->pmu_name, n->cpus);
tools/perf/util/header.c
3299
n->cpus = do_read_string(ff);
tools/perf/util/header.c
3300
if (!n->cpus)
tools/perf/util/header.c
3311
free(nodes[i].cpus);
tools/perf/util/header.c
4737
map = cpu_map__new_data(&ev->cpus.cpus);
tools/perf/util/header.c
4833
map = cpu_map__new_data(&ev->cpus.cpus);
tools/perf/util/header.c
734
ret = do_write_string(ff, n->cpus);
tools/perf/util/hwmon_pmu.c
379
hwm->pmu.cpus = perf_cpu_map__new_int(0);
tools/perf/util/hwmon_pmu.c
380
if (!hwm->pmu.cpus) {
tools/perf/util/mem-events.c
297
ret = perf_cpu_map__merge(&cpu_map, pmu->cpus);
tools/perf/util/parse-events.c
108
term_cpus = pmu->is_core && perf_cpu_map__is_empty(pmu->cpus)
tools/perf/util/parse-events.c
110
: perf_cpu_map__get(pmu->cpus);
tools/perf/util/parse-events.c
122
perf_cpu_map__merge(&cpus, term_cpus);
tools/perf/util/parse-events.c
126
return cpus;
tools/perf/util/parse-events.c
1409
struct perf_cpu_map *cpus;
tools/perf/util/parse-events.c
1430
cpus = get_config_cpu(head_config, parse_state->fake_pmu);
tools/perf/util/parse-events.c
1433
cpus, /*alternate_hw_config=*/PERF_COUNT_HW_MAX) ? 0 : -ENOMEM;
tools/perf/util/parse-events.c
1434
perf_cpu_map__put(cpus);
tools/perf/util/parse-events.c
183
struct perf_cpu_map *cpus, *pmu_cpus;
tools/perf/util/parse-events.c
231
pmu_cpus = perf_cpu_map__get(pmu->cpus);
tools/perf/util/parse-events.c
241
cpus = perf_cpu_map__get(user_cpus);
tools/perf/util/parse-events.c
243
cpus = perf_cpu_map__get(pmu_cpus);
tools/perf/util/parse-events.c
250
perf_cpu_map__put(cpus);
tools/perf/util/parse-events.c
268
evsel->core.cpus = cpus;
tools/perf/util/parse-events.c
288
perf_cpu_map__put(cpus);
tools/perf/util/parse-events.c
91
struct perf_cpu_map *cpus = NULL;
tools/perf/util/perf_api_probe.c
150
struct perf_cpu_map *cpus;
tools/perf/util/perf_api_probe.c
154
cpus = perf_cpu_map__new_online_cpus();
tools/perf/util/perf_api_probe.c
155
if (!cpus)
tools/perf/util/perf_api_probe.c
158
cpu = perf_cpu_map__cpu(cpus, 0);
tools/perf/util/perf_api_probe.c
159
perf_cpu_map__put(cpus);
tools/perf/util/perf_api_probe.c
63
struct perf_cpu_map *cpus;
tools/perf/util/perf_api_probe.c
67
cpus = perf_cpu_map__new_online_cpus();
tools/perf/util/perf_api_probe.c
68
if (!cpus)
tools/perf/util/perf_api_probe.c
70
cpu = perf_cpu_map__cpu(cpus, 0);
tools/perf/util/perf_api_probe.c
71
perf_cpu_map__put(cpus);
tools/perf/util/pfm.c
124
static bool is_libpfm_event_supported(const char *name, struct perf_cpu_map *cpus,
tools/perf/util/pfm.c
145
ret = evsel__open(evsel, cpus, threads);
tools/perf/util/pfm.c
155
ret = evsel__open(evsel, cpus, threads);
tools/perf/util/pfm.c
190
struct perf_cpu_map *cpus = perf_cpu_map__empty_new(1);
tools/perf/util/pfm.c
232
if (is_libpfm_event_supported(name, cpus, threads)) {
tools/perf/util/pfm.c
264
if (!is_libpfm_event_supported(name, cpus, threads))
tools/perf/util/pfm.c
279
perf_cpu_map__put(cpus);
tools/perf/util/pmu.c
1263
pmu->cpus = pmu_cpumask(dirfd, name, pmu->is_core);
tools/perf/util/pmu.c
1304
pmu->cpus = cpu_map__online();
tools/perf/util/pmu.c
2707
perf_cpu_map__put(pmu->cpus);
tools/perf/util/pmu.c
824
struct perf_cpu_map *cpus;
tools/perf/util/pmu.c
836
cpus = perf_cpu_map__new(cpumask);
tools/perf/util/pmu.c
838
if (cpus)
tools/perf/util/pmu.c
839
return cpus;
tools/perf/util/pmu.h
133
struct perf_cpu_map *cpus;
tools/perf/util/python.c
1067
struct perf_cpu_map *cpus = NULL;
tools/perf/util/python.c
1081
cpus = ((struct pyrf_cpu_map *)pcpus)->cpus;
tools/perf/util/python.c
1088
if (evsel__open(evsel, cpus, threads) < 0) {
tools/perf/util/python.c
1102
pcpu_map->cpus = perf_cpu_map__get(pevsel->evsel.core.cpus);
tools/perf/util/python.c
1131
ncpus = perf_cpu_map__nr(evsel->core.cpus);
tools/perf/util/python.c
1159
cpu_idx = perf_cpu_map__idx(evsel->core.cpus, (struct perf_cpu){.cpu = cpu});
tools/perf/util/python.c
1279
struct perf_cpu_map *cpus;
tools/perf/util/python.c
1286
cpus = ((struct pyrf_cpu_map *)pcpus)->cpus;
tools/perf/util/python.c
1287
evlist__init(&pevlist->evlist, cpus, threads);
tools/perf/util/python.c
1302
pcpu_map->cpus = perf_cpu_map__get(pevlist->evlist.core.all_cpus);
tools/perf/util/python.c
1440
cpu_idx = perf_cpu_map__idx(pos2->core.cpus,
tools/perf/util/python.c
2069
struct perf_cpu_map *cpus;
tools/perf/util/python.c
2076
cpus = pcpus ? ((struct pyrf_cpu_map *)pcpus)->cpus : NULL;
tools/perf/util/python.c
2079
evlist__init(&evlist, cpus, threads);
tools/perf/util/python.c
2096
struct perf_cpu_map *cpus;
tools/perf/util/python.c
2104
cpus = pcpus ? ((struct pyrf_cpu_map *)pcpus)->cpus : NULL;
tools/perf/util/python.c
2106
evlist__init(&evlist, cpus, threads);
tools/perf/util/python.c
514
struct perf_cpu_map *cpus;
tools/perf/util/python.c
527
pcpus->cpus = perf_cpu_map__new(cpustr);
tools/perf/util/python.c
528
if (pcpus->cpus == NULL)
tools/perf/util/python.c
535
perf_cpu_map__put(pcpus->cpus);
tools/perf/util/python.c
543
return perf_cpu_map__nr(pcpus->cpus);
tools/perf/util/python.c
550
if (i >= perf_cpu_map__nr(pcpus->cpus)) {
tools/perf/util/python.c
555
return Py_BuildValue("i", perf_cpu_map__cpu(pcpus->cpus, i).cpu);
tools/perf/util/record.c
241
struct perf_cpu_map *cpus = perf_cpu_map__new_online_cpus();
tools/perf/util/record.c
243
if (cpus)
tools/perf/util/record.c
244
cpu = perf_cpu_map__cpu(cpus, 0);
tools/perf/util/record.c
246
perf_cpu_map__put(cpus);
tools/perf/util/scripting-engines/trace-event-python.c
1701
struct perf_cpu_map *cpus = counter->core.cpus;
tools/perf/util/scripting-engines/trace-event-python.c
1707
perf_cpu_map__for_each_cpu(cpu, idx, cpus) {
tools/perf/util/stat-display.c
942
perf_cpu_map__for_each_cpu(cpu, idx, counter->core.cpus) {
tools/perf/util/stat.c
308
struct perf_cpu_map *cpus = evsel__cpus(counter);
tools/perf/util/stat.c
309
struct perf_cpu cpu = perf_cpu_map__cpu(cpus, cpu_map_idx);
tools/perf/util/stat.c
318
if (perf_cpu_map__is_any_cpu_or_is_empty(cpus))
tools/perf/util/stat.c
425
struct perf_cpu cpu = perf_cpu_map__cpu(evsel->core.cpus, cpu_map_idx);
tools/perf/util/stat.c
586
perf_cpu_map__for_each_cpu(cpu, idx, evsel->core.cpus) {
tools/perf/util/stat.c
599
perf_cpu_map__for_each_cpu(cpu, idx, evsel->core.cpus) {
tools/perf/util/stat.c
625
perf_cpu_map__for_each_cpu(cpu, idx, evsel->core.cpus) {
tools/perf/util/svghelper.c
108
total_height = (1 + rows + cpu2slot(cpus)) * SLOT_MULT;
tools/perf/util/svghelper.c
86
void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end)
tools/perf/util/svghelper.h
9
void open_svg(const char *filename, int cpus, int rows, u64 start, u64 end);
tools/perf/util/synthetic-events.c
2073
syn_data.data = &ev->cpus.cpus;
tools/perf/util/synthetic-events.h
68
int perf_event__synthesize_cpu_map(const struct perf_tool *tool, const struct perf_cpu_map *cpus, perf_event__handler_t process, struct machine *machine);
tools/perf/util/tool_pmu.c
205
struct perf_cpu_map *cpus,
tools/perf/util/tool_pmu.c
211
evsel->start_times = xyarray__new(perf_cpu_map__nr(cpus),
tools/perf/util/tool_pmu.c
279
cpu = perf_cpu_map__cpu(evsel->core.cpus, idx);
tools/perf/util/tool_pmu.c
354
if (!evsel || perf_cpu_map__is_empty(evsel->core.cpus)) {
tools/perf/util/tool_pmu.c
357
} else if (!perf_cpu_map__has_any_cpu(evsel->core.cpus)) {
tools/perf/util/tool_pmu.c
359
*result = perf_cpu_map__nr(evsel->core.cpus);
tools/perf/util/tool_pmu.c
379
if (!evsel || perf_cpu_map__is_empty(evsel->core.cpus)) {
tools/perf/util/tool_pmu.c
382
} else if (!perf_cpu_map__has_any_cpu(evsel->core.cpus)) {
tools/perf/util/tool_pmu.c
385
perf_cpu_map__intersect(online, evsel->core.cpus);
tools/perf/util/tool_pmu.c
537
struct perf_cpu cpu = perf_cpu_map__cpu(evsel->core.cpus,
tools/perf/util/tool_pmu.h
54
struct perf_cpu_map *cpus,
tools/perf/util/util.c
266
char cpus[MAX_NR_CPUS];
tools/perf/util/util.c
296
bitmap_scnprintf(bm, nbits, cpus, sizeof(cpus));
tools/perf/util/util.c
297
strcpy(cpulist, cpus);
tools/power/cpupower/lib/cpupower.c
170
int cpu, last_pkg, cpus = sysconf(_SC_NPROCESSORS_CONF);
tools/power/cpupower/lib/cpupower.c
174
cpu_top->core_info = malloc(sizeof(struct cpuid_core_info) * cpus);
tools/power/cpupower/lib/cpupower.c
178
for (cpu = 0; cpu < cpus; cpu++) {
tools/power/cpupower/lib/cpupower.c
214
qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info),
tools/power/cpupower/lib/cpupower.c
219
for (cpu = 1; cpu < cpus; cpu++) {
tools/power/cpupower/lib/cpupower.c
227
qsort(cpu_top->core_info, cpus, sizeof(struct cpuid_core_info),
tools/power/cpupower/lib/cpupower.c
234
for(cpu = 1; cpu < cpus; cpu++) {
tools/power/cpupower/lib/cpupower.c
245
return cpus;
tools/power/cpupower/utils/cpufreq-info.c
376
struct cpufreq_affected_cpus *cpus = cpufreq_get_affected_cpus(cpu);
tools/power/cpupower/utils/cpufreq-info.c
379
if (!cpus) {
tools/power/cpupower/utils/cpufreq-info.c
384
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-info.c
385
printf("%d ", cpus->cpu);
tools/power/cpupower/utils/cpufreq-info.c
386
cpus = cpus->next;
tools/power/cpupower/utils/cpufreq-info.c
388
printf("%d\n", cpus->cpu);
tools/power/cpupower/utils/cpufreq-info.c
389
cpufreq_put_affected_cpus(cpus);
tools/power/cpupower/utils/cpufreq-info.c
397
struct cpufreq_affected_cpus *cpus = cpufreq_get_related_cpus(cpu);
tools/power/cpupower/utils/cpufreq-info.c
400
if (!cpus) {
tools/power/cpupower/utils/cpufreq-info.c
405
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-info.c
406
printf("%d ", cpus->cpu);
tools/power/cpupower/utils/cpufreq-info.c
407
cpus = cpus->next;
tools/power/cpupower/utils/cpufreq-info.c
409
printf("%d\n", cpus->cpu);
tools/power/cpupower/utils/cpufreq-info.c
410
cpufreq_put_related_cpus(cpus);
tools/power/cpupower/utils/cpufreq-set.c
299
struct cpufreq_affected_cpus *cpus;
tools/power/cpupower/utils/cpufreq-set.c
305
cpus = cpufreq_get_related_cpus(cpu);
tools/power/cpupower/utils/cpufreq-set.c
306
if (!cpus)
tools/power/cpupower/utils/cpufreq-set.c
308
while (cpus->next) {
tools/power/cpupower/utils/cpufreq-set.c
309
bitmask_setbit(cpus_chosen, cpus->cpu);
tools/power/cpupower/utils/cpufreq-set.c
310
cpus = cpus->next;
tools/power/cpupower/utils/cpufreq-set.c
313
bitmask_setbit(cpus_chosen, cpus->cpu);
tools/power/cpupower/utils/cpufreq-set.c
314
cpufreq_put_related_cpus(cpus);
tools/power/x86/intel-speed-select/isst-config.c
558
int cpus[MAX_PACKAGE_COUNT][MAX_DIE_PER_PACKAGE][MAX_PUNIT_PER_DIE];
tools/power/x86/intel-speed-select/isst-config.c
562
memset(cpus, -1, sizeof(cpus));
tools/power/x86/intel-speed-select/isst-config.c
587
if (cpus[id.pkg][id.die][id.punit] == -1)
tools/power/x86/intel-speed-select/isst-config.c
588
cpus[id.pkg][id.die][id.punit] = i;
tools/power/x86/intel-speed-select/isst-config.c
594
id.cpu = cpus[i][k][k];
tools/power/x86/intel-speed-select/isst-config.c
612
id.cpu = cpus[i][j][k];
tools/power/x86/turbostat/turbostat.c
2395
} *cpus;
tools/power/x86/turbostat/turbostat.c
2458
int pkg_id = cpus[cpu].package_id;
tools/power/x86/turbostat/turbostat.c
2463
if (cpus[cpu].ht_id > 0) /* skip HT sibling */
tools/power/x86/turbostat/turbostat.c
2467
c = &core_base[GLOBAL_CORE_ID(cpus[cpu].core_id, pkg_id)];
tools/power/x86/turbostat/turbostat.c
2476
if (cpus[cpu].ht_sibling_cpu_id[i] <= 0)
tools/power/x86/turbostat/turbostat.c
2478
t = &thread_base[cpus[cpu].ht_sibling_cpu_id[i]];
tools/power/x86/turbostat/turbostat.c
2661
return cpus[cpu].core_id;
tools/power/x86/turbostat/turbostat.c
2664
return cpus[cpu].package_id;
tools/power/x86/turbostat/turbostat.c
3180
outp += sprintf(outp, "core: %d\n", cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
3365
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
3371
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].die_id);
tools/power/x86/turbostat/turbostat.c
3377
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].l3_id);
tools/power/x86/turbostat/turbostat.c
3383
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].physical_node_id);
tools/power/x86/turbostat/turbostat.c
3389
outp += sprintf(outp, "%s%d", (printed++ ? delim : ""), cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
5165
return cpus[cpu].package_id;
tools/power/x86/turbostat/turbostat.c
5167
return GLOBAL_CORE_ID(cpus[cpu].core_id, cpus[cpu].package_id);
tools/power/x86/turbostat/turbostat.c
5268
c->pmt_counter[i] = pmt_read_counter(pp, cpus[t->cpu_id].core_id);
tools/power/x86/turbostat/turbostat.c
5309
p->uncore_mhz = get_legacy_uncore_mhz(cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5333
path = find_sysfs_path_by_id(mp->sp, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5335
warnx("%s: package_id %d not found", __func__, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5347
p->pmt_counter[i] = pmt_read_counter(pp, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5991
if (cpus[i].put_ids)
tools/power/x86/turbostat/turbostat.c
5992
CPU_FREE(cpus[i].put_ids);
tools/power/x86/turbostat/turbostat.c
5994
free(cpus);
tools/power/x86/turbostat/turbostat.c
6057
cpus[cpu].logical_node_id = -1;
tools/power/x86/turbostat/turbostat.c
6063
if (cpus[cpu].package_id != pkg)
tools/power/x86/turbostat/turbostat.c
6066
if (cpus[cpu].logical_node_id != -1)
tools/power/x86/turbostat/turbostat.c
6068
cpus[cpu].logical_node_id = lnode;
tools/power/x86/turbostat/turbostat.c
6069
node = cpus[cpu].physical_node_id;
tools/power/x86/turbostat/turbostat.c
6076
if ((cpus[cpux].package_id == pkg) && (cpus[cpux].physical_node_id == node)) {
tools/power/x86/turbostat/turbostat.c
6077
cpus[cpux].logical_node_id = lnode;
tools/power/x86/turbostat/turbostat.c
6201
if ((so != cpu) && (cpus[so].ht_id < 0)) {
tools/power/x86/turbostat/turbostat.c
6202
cpus[so].ht_id = thread_id;
tools/power/x86/turbostat/turbostat.c
6203
cpus[cpu].ht_sibling_cpu_id[thread_id] = so;
tools/power/x86/turbostat/turbostat.c
6240
if (cpus[cpu].ht_id > 0) /* skip HT sibling */
tools/power/x86/turbostat/turbostat.c
6245
c = &core_base[GLOBAL_CORE_ID(cpus[cpu].core_id, cpus[cpu].package_id)];
tools/power/x86/turbostat/turbostat.c
6246
c2 = &core_base2[GLOBAL_CORE_ID(cpus[cpu].core_id, cpus[cpu].package_id)];
tools/power/x86/turbostat/turbostat.c
6247
p = &pkg_base[cpus[cpu].package_id];
tools/power/x86/turbostat/turbostat.c
6248
p2 = &pkg_base2[cpus[cpu].package_id];
tools/power/x86/turbostat/turbostat.c
6256
if (cpus[cpu].ht_sibling_cpu_id[i] <= 0)
tools/power/x86/turbostat/turbostat.c
6258
t = &thread_base[cpus[cpu].ht_sibling_cpu_id[i]];
tools/power/x86/turbostat/turbostat.c
6259
t2 = &thread_base2[cpus[cpu].ht_sibling_cpu_id[i]];
tools/power/x86/turbostat/turbostat.c
6404
cpus[cpu].ht_id = -1;
tools/power/x86/turbostat/turbostat.c
6406
cpus[cpu].ht_sibling_cpu_id[i] = -1;
tools/power/x86/turbostat/turbostat.c
6432
cpus[cpu].type = type;
tools/power/x86/turbostat/turbostat.c
8944
const int core_id = cpus[cpu].core_id;
tools/power/x86/turbostat/turbostat.c
8945
const int pkg_id = cpus[cpu].package_id;
tools/power/x86/turbostat/turbostat.c
9458
err(-1, "%s: cpu%d: type %d", __func__, cpu, cpus[cpu].type);
tools/power/x86/turbostat/turbostat.c
9493
cpus = calloc(1, (topo.max_cpu_num + 1) * sizeof(struct cpu_topology));
tools/power/x86/turbostat/turbostat.c
9494
if (cpus == NULL)
tools/power/x86/turbostat/turbostat.c
9600
cpus[i].cpu_id = i;
tools/power/x86/turbostat/turbostat.c
9603
cpus[i].package_id = get_package_id(i);
tools/power/x86/turbostat/turbostat.c
9604
if (cpus[i].package_id > max_package_id)
tools/power/x86/turbostat/turbostat.c
9605
max_package_id = cpus[i].package_id;
tools/power/x86/turbostat/turbostat.c
9608
cpus[i].die_id = get_die_id(i);
tools/power/x86/turbostat/turbostat.c
9609
if (cpus[i].die_id > topo.max_die_id)
tools/power/x86/turbostat/turbostat.c
9610
topo.max_die_id = cpus[i].die_id;
tools/power/x86/turbostat/turbostat.c
9613
cpus[i].l3_id = get_l3_id(i);
tools/power/x86/turbostat/turbostat.c
9614
if (cpus[i].l3_id > topo.max_l3_id)
tools/power/x86/turbostat/turbostat.c
9615
topo.max_l3_id = cpus[i].l3_id;
tools/power/x86/turbostat/turbostat.c
9618
cpus[i].physical_node_id = get_physical_node_id(&cpus[i]);
tools/power/x86/turbostat/turbostat.c
9619
if (cpus[i].physical_node_id > topo.max_node_num)
tools/power/x86/turbostat/turbostat.c
9620
topo.max_node_num = cpus[i].physical_node_id;
tools/power/x86/turbostat/turbostat.c
9623
cpus[i].core_id = get_core_id(i);
tools/power/x86/turbostat/turbostat.c
9624
if (cpus[i].core_id > max_core_id)
tools/power/x86/turbostat/turbostat.c
9625
max_core_id = cpus[i].core_id;
tools/power/x86/turbostat/turbostat.c
9628
siblings = set_thread_siblings(&cpus[i]);
tools/power/x86/turbostat/turbostat.c
9631
if (cpus[i].ht_id == 0)
tools/power/x86/turbostat/turbostat.c
9676
i, cpus[i].package_id, cpus[i].die_id, cpus[i].l3_id,
tools/power/x86/turbostat/turbostat.c
9677
cpus[i].physical_node_id, cpus[i].logical_node_id, cpus[i].core_id, cpus[i].ht_id);
tools/power/x86/turbostat/turbostat.c
9740
int pkg_id = cpus[cpu_id].package_id;
tools/power/x86/turbostat/turbostat.c
9741
int node_id = cpus[cpu_id].logical_node_id;
tools/power/x86/turbostat/turbostat.c
9742
int core_id = cpus[cpu_id].core_id;
tools/power/x86/turbostat/turbostat.c
9980
switch (cpus[cpu].type) {
tools/testing/selftests/arm64/fp/fp-stress.c
456
int cpus, i, j, c;
tools/testing/selftests/arm64/fp/fp-stress.c
477
cpus = num_processors();
tools/testing/selftests/arm64/fp/fp-stress.c
482
tests += sve_vl_count * cpus;
tools/testing/selftests/arm64/fp/fp-stress.c
489
tests += sme_vl_count * cpus * 2;
tools/testing/selftests/arm64/fp/fp-stress.c
495
tests += cpus;
tools/testing/selftests/arm64/fp/fp-stress.c
501
tests += cpus * 2;
tools/testing/selftests/arm64/fp/fp-stress.c
507
cpus, sve_vl_count, sme_vl_count,
tools/testing/selftests/arm64/fp/fp-stress.c
555
for (i = 0; i < cpus; i++) {
tools/testing/selftests/arm64/gcs/gcs-stress.c
358
int ret, cpus, i, c;
tools/testing/selftests/arm64/gcs/gcs-stress.c
374
cpus = num_processors();
tools/testing/selftests/arm64/gcs/gcs-stress.c
379
gcs_threads = cpus + 1;
tools/testing/selftests/arm64/gcs/gcs-stress.c
389
cpus, gcs_threads);
tools/testing/selftests/bpf/bench.c
359
if (parse_num_list(arg, &env.prod_cpus.cpus,
tools/testing/selftests/bpf/bench.c
367
if (parse_num_list(arg, &env.cons_cpus.cpus,
tools/testing/selftests/bpf/bench.c
473
if (cpu_set->cpus) {
tools/testing/selftests/bpf/bench.c
478
if (cpu_set->cpus[i]) {
tools/testing/selftests/bpf/bench.c
713
if (!env.prod_cpus.cpus)
tools/testing/selftests/bpf/bench.h
16
bool *cpus;
tools/testing/selftests/bpf/prog_tests/htab_update.c
71
cpu_set_t cpus;
tools/testing/selftests/bpf/prog_tests/htab_update.c
75
CPU_ZERO(&cpus);
tools/testing/selftests/bpf/prog_tests/htab_update.c
76
CPU_SET(0, &cpus);
tools/testing/selftests/bpf/prog_tests/htab_update.c
77
pthread_setaffinity_np(pthread_self(), sizeof(cpus), &cpus);
tools/testing/selftests/kvm/lib/x86/sev.c
169
struct kvm_vcpu *cpus[1];
tools/testing/selftests/kvm/lib/x86/sev.c
171
vm = __vm_create_with_vcpus(shape, 1, 0, guest_code, cpus);
tools/testing/selftests/kvm/lib/x86/sev.c
172
*cpu = cpus[0];
tools/testing/selftests/net/reuseport_bpf_cpu.c
233
int *rcv_fd, cpus;
tools/testing/selftests/net/reuseport_bpf_cpu.c
235
cpus = sysconf(_SC_NPROCESSORS_ONLN);
tools/testing/selftests/net/reuseport_bpf_cpu.c
236
if (cpus <= 0)
tools/testing/selftests/net/reuseport_bpf_cpu.c
239
rcv_fd = calloc(cpus, sizeof(int));
tools/testing/selftests/net/reuseport_bpf_cpu.c
244
test(rcv_fd, cpus, AF_INET, SOCK_DGRAM);
tools/testing/selftests/net/reuseport_bpf_cpu.c
247
test(rcv_fd, cpus, AF_INET6, SOCK_DGRAM);
tools/testing/selftests/net/reuseport_bpf_cpu.c
250
test(rcv_fd, cpus, AF_INET, SOCK_STREAM);
tools/testing/selftests/net/reuseport_bpf_cpu.c
253
test(rcv_fd, cpus, AF_INET6, SOCK_STREAM);
tools/testing/selftests/riscv/hwprobe/cbo.c
119
cpu_set_t *cpus = (cpu_set_t *)arg;
tools/testing/selftests/riscv/hwprobe/cbo.c
128
rc = riscv_hwprobe(&pair, 1, sizeof(cpu_set_t), (unsigned long *)cpus, 0);
tools/testing/selftests/riscv/hwprobe/cbo.c
157
cpu_set_t *cpus = (cpu_set_t *)arg;
tools/testing/selftests/riscv/hwprobe/cbo.c
161
rc = riscv_hwprobe(&pair, 1, sizeof(cpu_set_t), (unsigned long *)cpus, 0);
tools/testing/selftests/riscv/hwprobe/cbo.c
181
cpu_set_t *cpus = (cpu_set_t *)arg;
tools/testing/selftests/riscv/hwprobe/cbo.c
186
rc = riscv_hwprobe(&pair, 1, sizeof(cpu_set_t), (unsigned long *)cpus, 0);
tools/testing/selftests/riscv/hwprobe/cbo.c
224
static void check_no_zicbo_cpus(cpu_set_t *cpus, __u64 cbo)
tools/testing/selftests/riscv/hwprobe/cbo.c
234
while (i++ < CPU_COUNT(cpus)) {
tools/testing/selftests/riscv/hwprobe/cbo.c
235
while (!CPU_ISSET(c, cpus))
tools/testing/selftests/riscv/hwprobe/cbo.c
303
cpu_set_t cpus;
tools/testing/selftests/riscv/hwprobe/cbo.c
333
rc = sched_getaffinity(0, sizeof(cpu_set_t), &cpus);
tools/testing/selftests/riscv/hwprobe/cbo.c
339
rc = riscv_hwprobe(&pair, 1, sizeof(cpu_set_t), (unsigned long *)&cpus, 0);
tools/testing/selftests/riscv/hwprobe/cbo.c
348
check_no_zicbo_cpus(&cpus, RISCV_HWPROBE_EXT_ZICBOZ);
tools/testing/selftests/riscv/hwprobe/cbo.c
355
check_no_zicbo_cpus(&cpus, RISCV_HWPROBE_EXT_ZICBOM);
tools/testing/selftests/riscv/hwprobe/cbo.c
361
check_no_zicbo_cpus(&cpus, RISCV_HWPROBE_EXT_ZICBOP);
tools/testing/selftests/riscv/hwprobe/cbo.c
373
tests[i].test_fn(&cpus);
tools/testing/selftests/riscv/hwprobe/hwprobe.c
15
cpus = -1;
tools/testing/selftests/riscv/hwprobe/hwprobe.c
24
out = riscv_hwprobe(pairs, 8, 1, &cpus, 0);
tools/testing/selftests/riscv/hwprobe/hwprobe.c
46
out = riscv_hwprobe(pairs, 8, 0, &cpus, 0);
tools/testing/selftests/riscv/hwprobe/hwprobe.c
53
out = riscv_hwprobe(pairs, 1, 1, &cpus, 0);
tools/testing/selftests/riscv/hwprobe/hwprobe.c
8
unsigned long cpus;
tools/testing/selftests/riscv/hwprobe/hwprobe.h
23
size_t cpusetsize, unsigned long *cpus, unsigned int flags);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
121
CPU_ZERO(&cpus);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
122
rc = riscv_hwprobe(pairs, 1, 0, (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
130
CPU_ZERO(&cpus);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
131
rc = riscv_hwprobe(pairs, 1, sizeof(cpu_set_t), (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
132
ksft_test_result(rc == 0 && CPU_COUNT(&cpus) == 0, "unknown key\n");
tools/testing/selftests/riscv/hwprobe/which-cpus.c
136
CPU_ZERO(&cpus);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
137
rc = riscv_hwprobe(pairs, 2, sizeof(cpu_set_t), (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
143
CPU_ZERO(&cpus);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
144
rc = riscv_hwprobe(pairs, 3, sizeof(cpu_set_t), (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
145
ksft_test_result(rc == 0 && CPU_COUNT(&cpus) == sysconf(_SC_NPROCESSORS_ONLN), "set all cpus\n");
tools/testing/selftests/riscv/hwprobe/which-cpus.c
150
memcpy(&cpus, &cpus_aff, sizeof(cpu_set_t));
tools/testing/selftests/riscv/hwprobe/which-cpus.c
151
rc = riscv_hwprobe(pairs, 3, sizeof(cpu_set_t), (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
152
ksft_test_result(rc == 0 && CPU_EQUAL(&cpus, &cpus_aff), "set all affinity cpus\n");
tools/testing/selftests/riscv/hwprobe/which-cpus.c
157
memcpy(&cpus, &cpus_aff, sizeof(cpu_set_t));
tools/testing/selftests/riscv/hwprobe/which-cpus.c
158
rc = riscv_hwprobe(pairs, 3, sizeof(cpu_set_t), (unsigned long *)&cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
159
ksft_test_result(rc == 0 && CPU_COUNT(&cpus) == 0, "clear all cpus\n");
tools/testing/selftests/riscv/hwprobe/which-cpus.c
29
static void print_cpulist(cpu_set_t *cpus)
tools/testing/selftests/riscv/hwprobe/which-cpus.c
33
if (!CPU_COUNT(cpus)) {
tools/testing/selftests/riscv/hwprobe/which-cpus.c
39
for (int i = 0, c = 0; i < CPU_COUNT(cpus); i++, c++) {
tools/testing/selftests/riscv/hwprobe/which-cpus.c
40
if (start != end && !CPU_ISSET(c, cpus))
tools/testing/selftests/riscv/hwprobe/which-cpus.c
43
while (!CPU_ISSET(c, cpus))
tools/testing/selftests/riscv/hwprobe/which-cpus.c
59
static void do_which_cpus(int argc, char **argv, cpu_set_t *cpus)
tools/testing/selftests/riscv/hwprobe/which-cpus.c
78
rc = riscv_hwprobe(pairs, nr_pairs, sizeof(cpu_set_t), (unsigned long *)cpus, RISCV_HWPROBE_WHICH_CPUS);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
80
print_cpulist(cpus);
tools/testing/selftests/riscv/hwprobe/which-cpus.c
87
cpu_set_t cpus_aff, cpus;
tools/tracing/rtla/src/common.c
131
retval = osnoise_set_cpus(tool->context, params->cpus ? params->cpus : "all");
tools/tracing/rtla/src/common.c
137
if (!params->cpus) {
tools/tracing/rtla/src/common.c
149
} else if (params->cpus) {
tools/tracing/rtla/src/common.c
79
common->cpus = optarg;
tools/tracing/rtla/src/common.h
112
if (!(common)->cpus || CPU_ISSET(cpu, &(common)->monitored_cpus))
tools/tracing/rtla/src/common.h
146
int osnoise_set_cpus(struct osnoise_context *context, char *cpus);
tools/tracing/rtla/src/common.h
75
char *cpus;
tools/tracing/rtla/src/osnoise.c
52
int osnoise_set_cpus(struct osnoise_context *context, char *cpus)
tools/tracing/rtla/src/osnoise.c
61
context->curr_cpus = strdup(cpus);
tools/tracing/rtla/src/osnoise.c
65
snprintf(buffer, 1024, "%s\n", cpus);
tools/tracing/rtla/src/osnoise.c
67
debug_msg("setting cpus to %s from %s", cpus, context->orig_cpus);
tools/tracing/rtla/src/timerlat_bpf.c
151
int cpus)
tools/tracing/rtla/src/timerlat_bpf.c
157
sizeof(long long) * cpus, 0);
tools/tracing/rtla/src/timerlat_bpf.c
162
sizeof(long long) * cpus, 0);
tools/tracing/rtla/src/timerlat_bpf.c
167
sizeof(long long) * cpus, 0);
tools/tracing/rtla/src/timerlat_bpf.c
180
int cpus)
tools/tracing/rtla/src/timerlat_bpf.c
185
key, value_irq, value_thread, value_user, cpus);
tools/tracing/rtla/src/timerlat_bpf.c
195
int cpus)
tools/tracing/rtla/src/timerlat_bpf.c
200
key, value_irq, value_thread, value_user, cpus);
tools/tracing/rtla/src/timerlat_bpf.h
26
int cpus);
tools/tracing/rtla/src/timerlat_bpf.h
31
int cpus);
tools/tracing/rtla/src/timerlat_bpf.h
48
int cpus)
tools/tracing/rtla/src/timerlat_bpf.h
56
int cpus)
tools/virtio/virtio-trace/trace-agent.c
161
for (cpu = 0; cpu < s->cpus; cpu++) {
tools/virtio/virtio-trace/trace-agent.c
228
for (cpu = 0; cpu < s->cpus; cpu++)
tools/virtio/virtio-trace/trace-agent.c
234
for (cpu = 0; cpu < s->cpus; cpu++) {
tools/virtio/virtio-trace/trace-agent.c
250
for (i = 0; i < s->cpus; i++) {
tools/virtio/virtio-trace/trace-agent.c
61
s->cpus = get_total_cpus();
tools/virtio/virtio-trace/trace-agent.c
65
for (i = 0; i < s->cpus; i++)
tools/virtio/virtio-trace/trace-agent.h
21
int cpus;
virt/kvm/kvm_main.c
208
static inline bool kvm_kick_many_cpus(struct cpumask *cpus, bool wait)
virt/kvm/kvm_main.c
210
if (cpumask_empty(cpus))
virt/kvm/kvm_main.c
213
smp_call_function_many(cpus, ack_kick, NULL, wait);
virt/kvm/kvm_main.c
249
struct cpumask *cpus;
virt/kvm/kvm_main.c
255
cpus = this_cpu_cpumask_var_ptr(cpu_kick_mask);
virt/kvm/kvm_main.c
256
cpumask_clear(cpus);
virt/kvm/kvm_main.c
262
kvm_make_vcpu_request(vcpu, req, cpus, me);
virt/kvm/kvm_main.c
265
called = kvm_kick_many_cpus(cpus, !!(req & KVM_REQUEST_WAIT));
virt/kvm/kvm_main.c
274
struct cpumask *cpus;
virt/kvm/kvm_main.c
281
cpus = this_cpu_cpumask_var_ptr(cpu_kick_mask);
virt/kvm/kvm_main.c
282
cpumask_clear(cpus);
virt/kvm/kvm_main.c
285
kvm_make_vcpu_request(vcpu, req, cpus, me);
virt/kvm/kvm_main.c
287
called = kvm_kick_many_cpus(cpus, !!(req & KVM_REQUEST_WAIT));