Symbol: cpumask
arch/alpha/include/asm/smp.h
49
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/alpha/kernel/smp.c
175
wait_for_txrdy (unsigned long cpumask)
arch/alpha/kernel/smp.c
179
if (!(hwrpb->txrdy & cpumask))
arch/alpha/kernel/smp.c
184
if (!(hwrpb->txrdy & cpumask))
arch/alpha/kernel/smp.c
202
unsigned long cpumask;
arch/alpha/kernel/smp.c
210
cpumask = (1UL << cpuid);
arch/alpha/kernel/smp.c
211
if (wait_for_txrdy(cpumask))
arch/alpha/kernel/smp.c
224
if (wait_for_txrdy(cpumask))
arch/alpha/kernel/smp.c
497
send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation)
arch/alpha/kernel/smp.c
584
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/alpha/kernel/sys_dp264.c
151
dp264_set_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/sys_dp264.c
163
clipper_set_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/sys_titan.c
149
titan_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/time.c
125
.cpumask = cpumask_of(cpu),
arch/alpha/kernel/time.c
192
.cpumask = cpumask_of(cpu),
arch/arc/include/asm/smp.h
18
struct cpumask;
arch/arc/include/asm/smp.h
24
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/arc/kernel/mcip.c
264
idu_irq_set_affinity(struct irq_data *data, const struct cpumask *cpumask,
arch/arc/kernel/mcip.c
273
if (!cpumask_and(&online, cpumask, cpu_online_mask))
arch/arc/kernel/smp.c
283
static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg)
arch/arc/kernel/smp.c
298
struct cpumask targets;
arch/arc/kernel/smp.c
309
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/arc/kernel/smp.c
42
static int __init arc_get_cpu_map(const char *name, struct cpumask *cpumask)
arch/arc/kernel/smp.c
51
if (cpulist_parse(buf, cpumask))
arch/arc/kernel/smp.c
63
struct cpumask cpumask;
arch/arc/kernel/smp.c
65
if (arc_get_cpu_map("possible-cpus", &cpumask)) {
arch/arc/kernel/smp.c
69
cpumask_setall(&cpumask);
arch/arc/kernel/smp.c
72
if (!cpumask_test_cpu(0, &cpumask))
arch/arc/kernel/smp.c
75
init_cpu_possible(&cpumask);
arch/arm/include/asm/smp.h
70
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/arm/include/asm/smp.h
71
extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask);
arch/arm/kernel/smp.c
544
static void smp_cross_call(const struct cpumask *target, unsigned int ipinr);
arch/arm/kernel/smp.c
564
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/arm/kernel/smp.c
569
void arch_send_wakeup_ipi_mask(const struct cpumask *mask)
arch/arm/kernel/smp.c
588
void tick_broadcast(const struct cpumask *mask)
arch/arm/kernel/smp.c
707
static void smp_cross_call(const struct cpumask *target, unsigned int ipinr)
arch/arm/kernel/smp.c
756
struct cpumask mask;
arch/arm/kernel/smp.c
798
struct cpumask *cpus = freq->policy->cpus;
arch/arm/kernel/smp_twd.c
252
clk->cpumask = cpumask_of(cpu);
arch/arm/mach-footbridge/dc21285-timer.c
118
ce->cpumask = cpumask_of(smp_processor_id());
arch/arm/mach-imx/mmdc.c
134
__ATTR(cpumask, S_IRUGO, mmdc_pmu_cpumask_show, NULL);
arch/arm/mach-mmp/time.c
181
ckevt.cpumask = cpumask_of(0);
arch/arm/mach-omap1/time.c
167
clockevent_mpu_timer1.cpumask = cpumask_of(0);
arch/arm/mach-omap1/timer32k.c
155
clockevent_32k_timer.cpumask = cpumask_of(0);
arch/arm/mach-omap2/board-generic.c
44
void tick_broadcast(const struct cpumask *mask)
arch/arm/mach-spear/time.c
191
clkevt.cpumask = cpumask_of(0);
arch/arm/mm/cache-l2x0-pmu.c
397
__ATTR(cpumask, S_IRUGO, l2x0_pmu_cpumask_show, NULL);
arch/arm/plat-orion/time.c
236
orion_clkevt.cpumask = cpumask_of(0);
arch/arm64/include/asm/cpufeature.h
404
const struct cpumask *cpus;
arch/arm64/include/asm/cpufeature.h
673
const struct cpumask *system_32bit_el0_cpumask(void);
arch/arm64/include/asm/cpufeature.h
674
const struct cpumask *fallback_32bit_el0_cpumask(void);
arch/arm64/include/asm/mmu_context.h
265
static inline const struct cpumask *
arch/arm64/include/asm/mmu_context.h
266
__task_cpu_possible_mask(struct task_struct *p, const struct cpumask *mask)
arch/arm64/include/asm/mmu_context.h
277
static inline const struct cpumask *
arch/arm64/include/asm/mmu_context.h
284
const struct cpumask *task_cpu_fallback_mask(struct task_struct *p);
arch/arm64/include/asm/smp.h
100
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/arm64/kernel/cpufeature.c
1660
const struct cpumask *system_32bit_el0_cpumask(void)
arch/arm64/kernel/cpufeature.c
1671
const struct cpumask *task_cpu_fallback_mask(struct task_struct *p)
arch/arm64/kernel/cpufeature.c
1686
const struct cpumask *mask = system_32bit_el0_cpumask();
arch/arm64/kernel/cpufeature.c
1970
static struct cpumask dbm_cpus __read_mostly;
arch/arm64/kernel/cpufeature.c
2040
static struct cpumask amu_cpus __read_mostly;
arch/arm64/kernel/smp.c
1025
static void smp_cross_call(const struct cpumask *target, unsigned int ipinr)
arch/arm64/kernel/smp.c
1168
void tick_broadcast(const struct cpumask *mask)
arch/arm64/kernel/smp.c
827
static void smp_cross_call(const struct cpumask *target, unsigned int ipinr);
arch/arm64/kernel/smp.c
847
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/arm64/kernel/topology.c
253
static void amu_fie_setup(const struct cpumask *cpus)
arch/csky/include/asm/smp.h
16
void arch_send_call_function_ipi_mask(struct cpumask *mask);
arch/csky/include/asm/smp.h
20
void __init set_send_ipi(void (*func)(const struct cpumask *mask), int irq);
arch/csky/kernel/smp.c
123
void arch_send_call_function_ipi_mask(struct cpumask *mask)
arch/csky/kernel/smp.c
76
static void (*send_arch_ipi)(const struct cpumask *mask);
arch/csky/kernel/smp.c
79
void __init set_send_ipi(void (*func)(const struct cpumask *mask), int irq)
arch/csky/kernel/smp.c
89
send_ipi_message(const struct cpumask *to_whom, enum ipi_message_type operation)
arch/hexagon/include/asm/smp.h
23
extern void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg);
arch/hexagon/include/asm/smp.h
26
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/hexagon/kernel/smp.c
101
for_each_cpu(cpu, cpumask) {
arch/hexagon/kernel/smp.c
223
struct cpumask targets;
arch/hexagon/kernel/smp.c
234
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/hexagon/kernel/smp.c
93
void send_ipi(const struct cpumask *cpumask, enum ipi_message_type msg)
arch/hexagon/kernel/time.c
131
dummy_clock_dev->cpumask = cpumask_of(cpu);
arch/hexagon/kernel/time.c
171
ce_dev->cpumask = cpu_all_mask;
arch/hexagon/kernel/time.c
99
static void broadcast(const struct cpumask *mask)
arch/loongarch/include/asm/irq.h
61
void arch_trigger_cpumask_backtrace(const struct cpumask *mask, int exclude_cpu);
arch/loongarch/include/asm/smp.h
104
static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/loongarch/include/asm/smp.h
20
void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action);
arch/loongarch/include/asm/topology.h
33
static inline const struct cpumask *cpu_coregroup_mask(int cpu)
arch/loongarch/kernel/paravirt.c
68
static void pv_send_ipi_mask(const struct cpumask *mask, unsigned int action)
arch/loongarch/kernel/process.c
348
static struct cpumask backtrace_csd_busy;
arch/loongarch/kernel/smp.c
240
static void loongson_send_ipi_mask(const struct cpumask *mask, unsigned int action)
arch/loongarch/kernel/time.c
169
cd->cpumask = cpumask_of(cpu);
arch/m68k/coldfire/pit.c
143
cf_pit_clockevent.cpumask = cpumask_of(smp_processor_id());
arch/microblaze/kernel/timer.c
175
clockevent_xilinx_timer.cpumask = cpumask_of(0);
arch/mips/alchemy/common/time.c
72
.cpumask = cpu_possible_mask,
arch/mips/bcm63xx/irq.c
134
const struct cpumask *m) \
arch/mips/bcm63xx/irq.c
34
static void (*internal_irq_unmask)(struct irq_data *d, const struct cpumask *m);
arch/mips/bcm63xx/irq.c
376
const struct cpumask *dest,
arch/mips/bcm63xx/irq.c
54
const struct cpumask *m)
arch/mips/cavium-octeon/octeon-irq.c
1806
const struct cpumask *dest, bool force)
arch/mips/cavium-octeon/octeon-irq.c
2514
const struct cpumask *dest, bool force)
arch/mips/cavium-octeon/octeon-irq.c
276
const struct cpumask *mask = irq_data_get_affinity_mask(data);
arch/mips/cavium-octeon/octeon-irq.c
771
const struct cpumask *mask = irq_data_get_affinity_mask(data);
arch/mips/cavium-octeon/octeon-irq.c
792
const struct cpumask *dest, bool force)
arch/mips/cavium-octeon/octeon-irq.c
853
const struct cpumask *dest,
arch/mips/cavium-octeon/octeon-irq.c
898
const struct cpumask *dest,
arch/mips/cavium-octeon/smp.c
110
static inline void octeon_send_ipi_mask(const struct cpumask *mask,
arch/mips/cavium-octeon/smp.c
484
static void octeon_78xx_send_ipi_mask(const struct cpumask *mask,
arch/mips/include/asm/irq.h
79
void arch_trigger_cpumask_backtrace(const struct cpumask *mask,
arch/mips/include/asm/smp-cps.h
27
struct cpumask cpumask;
arch/mips/include/asm/smp-ops.h
24
void (*send_ipi_mask)(const struct cpumask *mask, unsigned int action);
arch/mips/include/asm/smp-ops.h
51
extern void mips_smp_send_ipi_mask(const struct cpumask *mask,
arch/mips/include/asm/smp.h
116
int mips_smp_ipi_allocate(const struct cpumask *mask);
arch/mips/include/asm/smp.h
123
int mips_smp_ipi_free(const struct cpumask *mask);
arch/mips/include/asm/smp.h
132
static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/mips/include/asm/topology.h
20
extern struct cpumask __cpu_primary_thread_mask;
arch/mips/include/asm/topology.h
21
#define cpu_primary_thread_mask ((const struct cpumask *)&__cpu_primary_thread_mask)
arch/mips/jazz/irq.c
135
cd->cpumask = cpumask_of(cpu);
arch/mips/kernel/cevt-bcm1480.c
117
cd->cpumask = cpumask_of(cpu);
arch/mips/kernel/cevt-ds1287.c
117
cd->cpumask = cpumask_of(0);
arch/mips/kernel/cevt-gt641xx.c
139
cd->cpumask = cpumask_of(0);
arch/mips/kernel/cevt-r4k.c
305
cd->cpumask = cpumask_of(cpu);
arch/mips/kernel/cevt-sb1250.c
117
cd->cpumask = cpumask_of(cpu);
arch/mips/kernel/cevt-txx9.c
196
cd->cpumask = cpumask_of(0);
arch/mips/kernel/mips-cm.c
541
struct cpumask *local_cl_mask;
arch/mips/kernel/mips-cm.c
549
local_cl_mask = &mips_cps_cluster_bootcfg[local_cl].cpumask;
arch/mips/kernel/process.c
722
static struct cpumask backtrace_csd_busy;
arch/mips/kernel/process.c
789
struct cpumask process_cpus;
arch/mips/kernel/smp-bmips.c
308
static void bmips5000_send_ipi_mask(const struct cpumask *mask,
arch/mips/kernel/smp-bmips.c
360
static void bmips43xx_send_ipi_mask(const struct cpumask *mask,
arch/mips/kernel/smp-cps.c
373
cpumask_set_cpu(nvpe++, &mips_cps_cluster_bootcfg[cl].cpumask);
arch/mips/kernel/smp-up.c
21
static inline void up_send_ipi_mask(const struct cpumask *mask,
arch/mips/kernel/smp.c
177
void mips_smp_send_ipi_mask(const struct cpumask *mask, unsigned int action)
arch/mips/kernel/smp.c
245
int mips_smp_ipi_allocate(const struct cpumask *mask)
arch/mips/kernel/smp.c
306
int mips_smp_ipi_free(const struct cpumask *mask)
arch/mips/kernel/smp.c
748
void tick_broadcast(const struct cpumask *mask)
arch/mips/kernel/smp.c
79
struct cpumask __cpu_primary_thread_mask __read_mostly;
arch/mips/kernel/time.c
40
struct cpumask *cpus = freq->policy->cpus;
arch/mips/lantiq/irq.c
245
const struct cpumask *cpumask, bool force)
arch/mips/lantiq/irq.c
247
struct cpumask tmask;
arch/mips/lantiq/irq.c
249
if (!cpumask_and(&tmask, cpumask, cpu_online_mask))
arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
113
cd->cpumask = cpumask_of(cpu);
arch/mips/loongson64/hpet.c
237
cd->cpumask = cpumask_of(cpu);
arch/mips/loongson64/smp.c
376
loongson3_send_ipi_mask(const struct cpumask *mask, unsigned int action)
arch/mips/pci/pci-xtalk-bridge.c
303
static int bridge_set_affinity(struct irq_data *d, const struct cpumask *mask,
arch/mips/sgi-ip27/ip27-irq.c
72
static void setup_hub_mask(struct hub_irq_data *hd, const struct cpumask *mask)
arch/mips/sgi-ip27/ip27-irq.c
92
static int set_affinity_hub_irq(struct irq_data *d, const struct cpumask *mask,
arch/mips/sgi-ip27/ip27-smp.c
123
static void ip27_send_ipi_mask(const struct cpumask *mask, unsigned int action)
arch/mips/sgi-ip27/ip27-timer.c
88
cd->cpumask = cpumask_of(cpu);
arch/mips/sgi-ip30/ip30-irq.c
173
const struct cpumask *mask, bool force)
arch/mips/sgi-ip30/ip30-smp.c
67
static void ip30_smp_send_ipi_mask(const struct cpumask *mask, u32 action)
arch/mips/sibyte/bcm1480/irq.c
77
static int bcm1480_set_affinity(struct irq_data *d, const struct cpumask *mask,
arch/mips/sibyte/bcm1480/smp.c
72
static void bcm1480_send_ipi_mask(const struct cpumask *mask,
arch/mips/sibyte/sb1250/irq.c
70
static int sb1250_set_affinity(struct irq_data *d, const struct cpumask *mask,
arch/mips/sibyte/sb1250/smp.c
61
static inline void sb1250_send_ipi_mask(const struct cpumask *mask,
arch/mips/sni/time.c
67
cd->cpumask = cpumask_of(cpu);
arch/nios2/kernel/time.c
262
nios2_ce.ced.cpumask = cpumask_of(0);
arch/openrisc/include/asm/smp.h
21
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/openrisc/include/asm/smp.h
23
extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int),
arch/openrisc/kernel/smp.c
209
void __init set_smp_cross_call(void (*fn)(const struct cpumask *, unsigned int),
arch/openrisc/kernel/smp.c
228
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/openrisc/kernel/smp.c
246
static void smp_flush_tlb_mm(struct cpumask *cmask, struct mm_struct *mm)
arch/openrisc/kernel/smp.c
283
static void smp_flush_tlb_range(const struct cpumask *cmask, unsigned long start,
arch/openrisc/kernel/smp.c
30
static void (*smp_cross_call)(const struct cpumask *, unsigned int);
arch/openrisc/kernel/smp.c
331
const struct cpumask *cmask = vma ? mm_cpumask(vma->vm_mm)
arch/openrisc/kernel/time.c
86
evt->cpumask = cpumask_of(cpu);
arch/parisc/include/asm/irq.h
48
extern int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest);
arch/parisc/include/asm/smp.h
33
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/parisc/kernel/irq.c
97
int cpu_check_affinity(struct irq_data *d, const struct cpumask *dest)
arch/parisc/kernel/smp.c
210
send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op)
arch/parisc/kernel/smp.c
258
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/parisc/kernel/time.c
92
cd->cpumask = cpumask_of(cpu);
arch/powerpc/include/asm/fadump-internal.h
78
struct cpumask cpu_mask;
arch/powerpc/include/asm/irq.h
54
int irq_choose_cpu(const struct cpumask *mask);
arch/powerpc/include/asm/smp.h
114
static inline struct cpumask *cpu_sibling_mask(int cpu)
arch/powerpc/include/asm/smp.h
119
static inline struct cpumask *cpu_core_mask(int cpu)
arch/powerpc/include/asm/smp.h
124
static inline struct cpumask *cpu_l2_cache_mask(int cpu)
arch/powerpc/include/asm/smp.h
129
static inline struct cpumask *cpu_smallcore_mask(int cpu)
arch/powerpc/include/asm/smp.h
142
static inline const struct cpumask *cpu_smt_mask(int cpu)
arch/powerpc/include/asm/smp.h
194
static inline const struct cpumask *cpu_sibling_mask(int cpu)
arch/powerpc/include/asm/smp.h
199
static inline const struct cpumask *cpu_smallcore_mask(int cpu)
arch/powerpc/include/asm/smp.h
204
static inline const struct cpumask *cpu_l2_cache_mask(int cpu)
arch/powerpc/include/asm/smp.h
256
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/powerpc/include/asm/topology.h
134
struct cpumask *cpu_coregroup_mask(int cpu);
arch/powerpc/include/asm/topology.h
135
const struct cpumask *cpu_die_mask(int cpu);
arch/powerpc/include/asm/xics.h
170
extern int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask,
arch/powerpc/include/asm/xics.h
173
#define xics_get_irq_server(virq, cpumask, strict_check) (xics_default_server)
arch/powerpc/kernel/cacheinfo.c
119
struct cpumask shared_cpu_map; /* online CPUs using this cache */
arch/powerpc/kernel/cacheinfo.c
685
const struct cpumask *mask;
arch/powerpc/kernel/fadump.c
1135
fdh->cpu_mask_sz = sizeof(struct cpumask);
arch/powerpc/kernel/fadump.c
1697
fdh->cpu_mask_sz != sizeof(struct cpumask)) {
arch/powerpc/kernel/irq.c
360
int irq_choose_cpu(const struct cpumask *mask)
arch/powerpc/kernel/irq.c
389
int irq_choose_cpu(const struct cpumask *mask)
arch/powerpc/kernel/smp.c
1033
static const struct cpumask *tl_cache_mask(struct sched_domain_topology_level *tl, int cpu)
arch/powerpc/kernel/smp.c
1039
static const struct cpumask *tl_smallcore_smt_mask(struct sched_domain_topology_level *tl, int cpu)
arch/powerpc/kernel/smp.c
1045
struct cpumask *cpu_coregroup_mask(int cpu)
arch/powerpc/kernel/smp.c
1095
const struct cpumask *cpu_die_mask(int cpu)
arch/powerpc/kernel/smp.c
1433
struct cpumask *(*submask_fn)(int) = cpu_sibling_mask;
arch/powerpc/kernel/smp.c
1504
struct cpumask *(*mask_fn)(int) = cpu_sibling_mask;
arch/powerpc/kernel/smp.c
1546
struct cpumask *(*submask_fn)(int) = cpu_sibling_mask;
arch/powerpc/kernel/smp.c
1582
struct cpumask *(*submask_fn)(int) = cpu_sibling_mask;
arch/powerpc/kernel/smp.c
1685
struct cpumask *(*sibling_mask)(int) = cpu_sibling_mask;
arch/powerpc/kernel/smp.c
1686
struct cpumask *mask = cpu_l2_cache_mask(cpu);
arch/powerpc/kernel/smp.c
374
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/powerpc/kernel/smp.c
405
static struct cpumask nmi_ipi_pending_mask;
arch/powerpc/kernel/smp.c
571
void tick_broadcast(const struct cpumask *mask)
arch/powerpc/kernel/smp.c
717
static void set_cpus_related(int i, int j, struct cpumask *(*get_cpumask)(int))
arch/powerpc/kernel/smp.c
725
struct cpumask *(*get_cpumask)(int))
arch/powerpc/kernel/smp.c
736
static void or_cpumasks_related(int i, int j, struct cpumask *(*srcmask)(int),
arch/powerpc/kernel/smp.c
737
struct cpumask *(*dstmask)(int))
arch/powerpc/kernel/smp.c
739
struct cpumask *mask;
arch/powerpc/kernel/time.c
835
dec->cpumask = cpumask_of(cpu);
arch/powerpc/mm/book3s64/radix_tlb.c
371
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/book3s64/radix_tlb.c
510
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/book3s64/radix_tlb.c
561
struct cpumask *cpus = mm_cpumask(mm);
arch/powerpc/mm/nohash/tlb.c
236
struct cpumask *cpu_mask;
arch/powerpc/perf/hv-24x7.c
1114
static DEVICE_ATTR_RO(cpumask);
arch/powerpc/perf/hv-gpci.c
614
static DEVICE_ATTR_RO(cpumask);
arch/powerpc/perf/imc-pmu.c
123
static DEVICE_ATTR(cpumask, S_IRUGO, imc_pmu_cpumask_get_attr, NULL);
arch/powerpc/perf/imc-pmu.c
360
const struct cpumask *l_cpumask;
arch/powerpc/perf/imc-pmu.c
423
const struct cpumask *l_cpumask;
arch/powerpc/perf/imc-pmu.c
424
static struct cpumask tmp_mask;
arch/powerpc/perf/imc-pmu.c
649
const struct cpumask *l_cpumask;
arch/powerpc/perf/imc-pmu.c
650
static struct cpumask tmp_mask;
arch/powerpc/platforms/cell/spufs/sched.c
158
const struct cpumask *mask = cpumask_of_node(node);
arch/powerpc/platforms/powernv/opal-imc.c
185
const struct cpumask *l_cpumask;
arch/powerpc/sysdev/ehv_pic.c
71
static int ehv_pic_set_affinity(struct irq_data *d, const struct cpumask *dest,
arch/powerpc/sysdev/mpic.c
627
static inline u32 mpic_physmask(u32 cpumask)
arch/powerpc/sysdev/mpic.c
632
for (i = 0; i < min(32, NR_CPUS) && cpu_possible(i); ++i, cpumask >>= 1)
arch/powerpc/sysdev/mpic.c
633
mask |= (cpumask & 1) << get_hard_smp_processor_id(i);
arch/powerpc/sysdev/mpic.c
827
int mpic_set_affinity(struct irq_data *d, const struct cpumask *cpumask,
arch/powerpc/sysdev/mpic.c
834
int cpuid = irq_choose_cpu(cpumask);
arch/powerpc/sysdev/mpic.c
838
u32 mask = cpumask_bits(cpumask)[0];
arch/powerpc/sysdev/mpic.h
35
const struct cpumask *cpumask, bool force);
arch/powerpc/sysdev/xics/ics-native.c
109
server = xics_get_irq_server(d->irq, cpumask, 1);
arch/powerpc/sysdev/xics/ics-native.c
112
__func__, cpumask_pr_args(cpumask), d->irq);
arch/powerpc/sysdev/xics/ics-native.c
97
const struct cpumask *cpumask,
arch/powerpc/sysdev/xics/ics-opal.c
115
wanted_server = xics_get_irq_server(d->irq, cpumask, 1);
arch/powerpc/sysdev/xics/ics-opal.c
118
__func__, cpumask_pr_args(cpumask), d->irq);
arch/powerpc/sysdev/xics/ics-opal.c
95
const struct cpumask *cpumask,
arch/powerpc/sysdev/xics/ics-rtas.c
100
const struct cpumask *cpumask,
arch/powerpc/sysdev/xics/ics-rtas.c
119
irq_server = xics_get_irq_server(d->irq, cpumask, 1);
arch/powerpc/sysdev/xics/ics-rtas.c
122
__func__, cpumask_pr_args(cpumask), d->irq);
arch/powerpc/sysdev/xics/xics-common.c
278
int xics_get_irq_server(unsigned int virq, const struct cpumask *cpumask,
arch/powerpc/sysdev/xics/xics-common.c
285
if (!cpumask_subset(cpu_possible_mask, cpumask)) {
arch/powerpc/sysdev/xics/xics-common.c
286
int server = cpumask_first_and(cpu_online_mask, cpumask);
arch/powerpc/sysdev/xive/common.c
548
static int xive_find_target_in_mask(const struct cpumask *mask,
arch/powerpc/sysdev/xive/common.c
595
const struct cpumask *affinity)
arch/powerpc/sysdev/xive/common.c
718
const struct cpumask *cpumask,
arch/powerpc/sysdev/xive/common.c
729
if (!cpumask_intersects(cpumask, cpu_online_mask))
arch/powerpc/sysdev/xive/common.c
738
cpumask_test_cpu(xd->target, cpumask))
arch/powerpc/sysdev/xive/common.c
742
target = xive_pick_irq_target(d, cpumask);
arch/riscv/include/asm/sbi.h
588
int sbi_remote_fence_i(const struct cpumask *cpu_mask);
arch/riscv/include/asm/sbi.h
590
int sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask,
arch/riscv/include/asm/sbi.h
594
int sbi_remote_hfence_gvma(const struct cpumask *cpu_mask,
arch/riscv/include/asm/sbi.h
597
int sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask,
arch/riscv/include/asm/sbi.h
601
int sbi_remote_hfence_vvma(const struct cpumask *cpu_mask,
arch/riscv/include/asm/sbi.h
604
int sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask,
arch/riscv/include/asm/sbi.h
687
static inline int sbi_remote_fence_i(const struct cpumask *cpu_mask) { return -1; }
arch/riscv/include/asm/smp.h
35
void arch_send_call_function_ipi_mask(struct cpumask *mask);
arch/riscv/include/asm/tlbbatch.h
12
struct cpumask cpumask;
arch/riscv/include/asm/vendor_extensions/mips_hwprobe.h
13
void hwprobe_isa_vendor_ext_mips_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/mips_hwprobe.h
16
const struct cpumask *cpus)
arch/riscv/include/asm/vendor_extensions/sifive_hwprobe.h
10
void hwprobe_isa_vendor_ext_sifive_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/sifive_hwprobe.h
13
const struct cpumask *cpus)
arch/riscv/include/asm/vendor_extensions/thead_hwprobe.h
10
void hwprobe_isa_vendor_ext_thead_0(struct riscv_hwprobe *pair, const struct cpumask *cpus);
arch/riscv/include/asm/vendor_extensions/thead_hwprobe.h
13
const struct cpumask *cpus)
arch/riscv/kernel/sbi.c
114
static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
166
static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
23
static int (*__sbi_rfence)(int fid, const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
257
static int __sbi_rfence_v02(int fid, const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
28
static unsigned long __sbi_v01_cpumask_to_hartmask(const struct cpumask *cpu_mask)
arch/riscv/kernel/sbi.c
399
int sbi_remote_fence_i(const struct cpumask *cpu_mask)
arch/riscv/kernel/sbi.c
418
int sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
441
int sbi_remote_hfence_gvma(const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
461
int sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
480
int sbi_remote_hfence_vvma(const struct cpumask *cpu_mask,
arch/riscv/kernel/sbi.c
501
int sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask,
arch/riscv/kernel/smp.c
115
static void send_ipi_mask(const struct cpumask *mask, enum ipi_message_type op)
arch/riscv/kernel/smp.c
237
void arch_send_call_function_ipi_mask(struct cpumask *mask)
arch/riscv/kernel/smp.c
248
void tick_broadcast(const struct cpumask *mask)
arch/riscv/kernel/sys_hwprobe.c
187
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
214
static bool hwprobe_ext0_has(const struct cpumask *cpus, u64 ext)
arch/riscv/kernel/sys_hwprobe.c
223
static u64 hwprobe_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
246
static u64 hwprobe_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
259
static u64 hwprobe_vec_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
283
static u64 hwprobe_vec_misaligned(const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
296
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
36
const struct cpumask *cpus)
arch/riscv/kernel/sys_hwprobe.c
82
const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/mips_hwprobe.c
18
const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/sifive_hwprobe.c
13
void hwprobe_isa_vendor_ext_sifive_0(struct riscv_hwprobe *pair, const struct cpumask *cpus)
arch/riscv/kernel/vendor_extensions/thead_hwprobe.c
13
void hwprobe_isa_vendor_ext_thead_0(struct riscv_hwprobe *pair, const struct cpumask *cpus)
arch/riscv/mm/tlbflush.c
119
const struct cpumask *cmask,
arch/riscv/mm/tlbflush.c
233
cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm));
arch/riscv/mm/tlbflush.c
239
__flush_tlb_range(NULL, &batch->cpumask,
arch/riscv/mm/tlbflush.c
241
cpumask_clear(&batch->cpumask);
arch/s390/include/asm/pci.h
342
static inline const struct cpumask *
arch/s390/include/asm/smp.h
44
extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/s390/include/asm/topology.h
50
const struct cpumask *cpu_coregroup_mask(int cpu);
arch/s390/include/asm/topology.h
91
static inline const struct cpumask *cpumask_of_node(int node)
arch/s390/include/uapi/asm/hwctrset.h
26
__u64 *cpumask; /* Pointer to CPU mask */
arch/s390/kernel/hiperdispatch.c
108
const struct cpumask *siblings;
arch/s390/kernel/perf_cpum_cf.c
1624
umask = (void __user *)start.cpumask;
arch/s390/kernel/perf_pai.c
166
struct cpumask *mask = PAI_CPU_MASK(event);
arch/s390/kernel/perf_pai.c
314
struct cpumask *maskptr;
arch/s390/kernel/processor.c
69
void notrace stop_machine_yield(const struct cpumask *cpumask)
arch/s390/kernel/processor.c
76
cpu = cpumask_next_wrap(this_cpu, cpumask);
arch/s390/kernel/smp.c
165
static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address)
arch/s390/kernel/smp.c
427
static cpumask_t cpumask;
arch/s390/kernel/smp.c
432
cpumask_copy(&cpumask, cpu_online_mask);
arch/s390/kernel/smp.c
433
cpumask_clear_cpu(smp_processor_id(), &cpumask);
arch/s390/kernel/smp.c
436
for_each_cpu(cpu, &cpumask) {
arch/s390/kernel/smp.c
445
for_each_cpu(cpu, &cpumask)
arch/s390/kernel/smp.c
447
cpumask_clear_cpu(cpu, &cpumask);
arch/s390/kernel/smp.c
448
if (cpumask_empty(&cpumask))
arch/s390/kernel/smp.c
513
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/s390/kernel/time.c
162
cd->cpumask = cpumask_of(cpu);
arch/s390/kernel/topology.c
511
const struct cpumask *cpu_coregroup_mask(int cpu)
arch/s390/kernel/topology.c
516
static const struct cpumask *tl_book_mask(struct sched_domain_topology_level *tl, int cpu)
arch/s390/kernel/topology.c
521
static const struct cpumask *tl_drawer_mask(struct sched_domain_topology_level *tl, int cpu)
arch/s390/pci/pci_irq.c
126
static int zpci_set_irq_affinity(struct irq_data *data, const struct cpumask *dest,
arch/sh/include/asm/smp.h
39
void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/sh/include/asm/topology.h
20
const struct cpumask *cpu_coregroup_mask(int cpu);
arch/sh/kernel/irq.c
235
const struct cpumask *mask = irq_data_get_affinity_mask(data);
arch/sh/kernel/smp.c
271
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/sh/kernel/smp.c
285
void tick_broadcast(const struct cpumask *mask)
arch/sh/kernel/topology.c
30
const struct cpumask *cpu_coregroup_mask(int cpu)
arch/sparc/include/asm/irq_64.h
89
void arch_trigger_cpumask_backtrace(const struct cpumask *mask,
arch/sparc/include/asm/smp_32.h
95
void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/sparc/include/asm/smp_64.h
41
void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/sparc/include/asm/topology_64.h
60
static inline const struct cpumask *cpu_coregroup_mask(int cpu)
arch/sparc/kernel/irq_64.c
349
static int irq_choose_cpu(unsigned int irq, const struct cpumask *affinity)
arch/sparc/kernel/irq_64.c
392
const struct cpumask *mask, bool force)
arch/sparc/kernel/irq_64.c
469
const struct cpumask *mask, bool force)
arch/sparc/kernel/irq_64.c
534
const struct cpumask *mask, bool force)
arch/sparc/kernel/leon_kernel.c
107
static int irq_choose_cpu(const struct cpumask *affinity)
arch/sparc/kernel/leon_kernel.c
120
static int leon_set_affinity(struct irq_data *data, const struct cpumask *dest,
arch/sparc/kernel/smp_32.c
143
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/sparc/kernel/smp_64.c
855
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/sparc/kernel/time_32.c
130
ce->cpumask = cpu_possible_mask;
arch/sparc/kernel/time_32.c
187
int cpu = cpumask_first(evt->cpumask);
arch/sparc/kernel/time_32.c
195
int cpu = cpumask_first(evt->cpumask);
arch/sparc/kernel/time_32.c
204
int cpu = cpumask_first(evt->cpumask);
arch/sparc/kernel/time_32.c
226
ce->cpumask = cpumask_of(cpu);
arch/sparc/kernel/time_64.c
766
sevt->cpumask = cpumask_of(smp_processor_id());
arch/um/include/asm/smp.h
16
void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/um/kernel/smp.c
42
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/um/kernel/time.c
926
evt->cpumask = cpumask_of(cpu);
arch/x86/events/amd/iommu.c
142
static DEVICE_ATTR(cpumask, S_IRUGO, _iommu_cpumask_show, NULL);
arch/x86/events/amd/power.c
153
static DEVICE_ATTR(cpumask, S_IRUGO, get_attr_cpumask, NULL);
arch/x86/events/amd/uncore.c
326
static DEVICE_ATTR(cpumask, S_IRUGO, amd_uncore_attr_show_cpumask, NULL);
arch/x86/events/intel/uncore.c
847
static DEVICE_ATTR(cpumask, S_IRUGO, uncore_get_attr_cpumask, NULL);
arch/x86/hyperv/hv_apic.c
112
static bool __send_ipi_mask_ex(const struct cpumask *mask, int vector,
arch/x86/hyperv/hv_apic.c
167
static bool __send_ipi_mask(const struct cpumask *mask, int vector,
arch/x86/hyperv/hv_apic.c
270
static void hv_send_ipi_mask(const struct cpumask *mask, int vector)
arch/x86/hyperv/hv_apic.c
276
static void hv_send_ipi_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/hyperv/mmu.c
172
static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
arch/x86/hyperv/mmu.c
19
static u64 hyperv_flush_tlb_others_ex(const struct cpumask *cpus,
arch/x86/hyperv/mmu.c
59
static void hyperv_flush_tlb_multi(const struct cpumask *cpus,
arch/x86/include/asm/apic.h
282
void (*send_IPI_mask)(const struct cpumask *mask, int vector);
arch/x86/include/asm/apic.h
283
void (*send_IPI_mask_allbutself)(const struct cpumask *msk, int vec);
arch/x86/include/asm/apic.h
329
void (*send_IPI_mask)(const struct cpumask *mask, int vector);
arch/x86/include/asm/apic.h
330
void (*send_IPI_mask_allbutself)(const struct cpumask *msk, int vec);
arch/x86/include/asm/apic.h
433
static __always_inline void __apic_send_IPI_mask(const struct cpumask *mask, int vector)
arch/x86/include/asm/apic.h
438
static __always_inline void __apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/include/asm/cpu.h
68
extern struct cpumask cpus_stop_mask;
arch/x86/include/asm/cpumask.h
21
static __always_inline void arch_cpumask_clear_cpu(int cpu, struct cpumask *dstp)
arch/x86/include/asm/cpumask.h
31
static __always_inline void arch_cpumask_clear_cpu(int cpu, struct cpumask *dstp)
arch/x86/include/asm/hw_irq.h
78
const struct cpumask *mask;
arch/x86/include/asm/irq.h
44
void arch_trigger_cpumask_backtrace(const struct cpumask *mask,
arch/x86/include/asm/irqdomain.h
19
const struct cpumask *mask);
arch/x86/include/asm/msr.h
263
void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs);
arch/x86/include/asm/msr.h
264
void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs);
arch/x86/include/asm/msr.h
292
static inline void rdmsr_on_cpus(const struct cpumask *m, u32 msr_no,
arch/x86/include/asm/msr.h
297
static inline void wrmsr_on_cpus(const struct cpumask *m, u32 msr_no,
arch/x86/include/asm/paravirt.h
36
void native_flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/include/asm/paravirt.h
54
static inline void __flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/include/asm/paravirt.h
57
PVOP_VCALL2(pv_ops, mmu.flush_tlb_multi, cpumask, info);
arch/x86/include/asm/paravirt_types.h
112
void (*flush_tlb_multi)(const struct cpumask *cpus,
arch/x86/include/asm/paravirt_types.h
18
struct cpumask;
arch/x86/include/asm/pci.h
115
static inline const struct cpumask *
arch/x86/include/asm/smp.h
116
void wbinvd_on_cpus_mask(struct cpumask *cpus);
arch/x86/include/asm/smp.h
118
void wbnoinvd_on_cpus_mask(struct cpumask *cpus);
arch/x86/include/asm/smp.h
124
void native_send_call_func_ipi(const struct cpumask *mask);
arch/x86/include/asm/smp.h
142
static inline struct cpumask *cpu_llc_shared_mask(int cpu)
arch/x86/include/asm/smp.h
147
static inline struct cpumask *cpu_l2c_shared_mask(int cpu)
arch/x86/include/asm/smp.h
159
static inline void wbinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/include/asm/smp.h
169
static inline void wbnoinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/include/asm/smp.h
174
static inline struct cpumask *cpu_llc_shared_mask(int cpu)
arch/x86/include/asm/smp.h
176
return (struct cpumask *)cpumask_of(0);
arch/x86/include/asm/smp.h
41
void (*send_call_func_ipi)(const struct cpumask *mask);
arch/x86/include/asm/smp.h
98
static inline void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/x86/include/asm/tlbbatch.h
12
struct cpumask cpumask;
arch/x86/include/asm/tlbflush.h
235
void flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/include/asm/tlbflush.h
353
cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm));
arch/x86/include/asm/topology.h
140
extern const struct cpumask *cpu_coregroup_mask(int cpu);
arch/x86/include/asm/topology.h
141
extern const struct cpumask *cpu_clustergroup_mask(int cpu);
arch/x86/include/asm/topology.h
233
extern struct cpumask __cpu_primary_thread_mask;
arch/x86/include/asm/topology.h
234
#define cpu_primary_thread_mask ((const struct cpumask *)&__cpu_primary_thread_mask)
arch/x86/include/asm/topology.h
67
extern const struct cpumask *cpumask_of_node(int node);
arch/x86/include/asm/topology.h
70
static inline const struct cpumask *cpumask_of_node(int node)
arch/x86/include/asm/trace/hyperv.h
12
TP_PROTO(const struct cpumask *cpus,
arch/x86/include/asm/trace/hyperv.h
60
TP_PROTO(const struct cpumask *cpus,
arch/x86/kernel/apic/apic.c
486
static void lapic_timer_broadcast(const struct cpumask *mask)
arch/x86/kernel/apic/apic.c
585
levt->cpumask = cpumask_of(smp_processor_id());
arch/x86/kernel/apic/apic_noop.c
24
static void noop_send_IPI_mask(const struct cpumask *cpumask, int vector) { }
arch/x86/kernel/apic/apic_noop.c
25
static void noop_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) { }
arch/x86/kernel/apic/apic_numachip.c
102
static void numachip_send_IPI_mask_allbutself(const struct cpumask *mask,
arch/x86/kernel/apic/apic_numachip.c
94
static void numachip_send_IPI_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/io_apic.c
1797
static int ioapic_set_affinity(struct irq_data *irq_data, const struct cpumask *mask, bool force)
arch/x86/kernel/apic/ipi.c
199
void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/ipi.c
212
void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask,
arch/x86/kernel/apic/ipi.c
252
void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/ipi.c
263
void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask,
arch/x86/kernel/apic/ipi.c
278
void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector)
arch/x86/kernel/apic/ipi.c
280
unsigned long mask = cpumask_bits(cpumask)[0];
arch/x86/kernel/apic/ipi.c
82
void native_send_call_func_ipi(const struct cpumask *mask)
arch/x86/kernel/apic/local.h
66
void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector);
arch/x86/kernel/apic/local.h
67
void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, int vector);
arch/x86/kernel/apic/local.h
73
void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector);
arch/x86/kernel/apic/local.h
74
void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, int vector);
arch/x86/kernel/apic/local.h
75
void default_send_IPI_mask_logical(const struct cpumask *mask, int vector);
arch/x86/kernel/apic/msi.c
35
msi_set_affinity(struct irq_data *irqd, const struct cpumask *mask, bool force)
arch/x86/kernel/apic/vector.c
204
const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd);
arch/x86/kernel/apic/vector.c
240
assign_vector_locked(struct irq_data *irqd, const struct cpumask *dest)
arch/x86/kernel/apic/vector.c
275
static int assign_irq_vector(struct irq_data *irqd, const struct cpumask *dest)
arch/x86/kernel/apic/vector.c
290
const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd);
arch/x86/kernel/apic/vector.c
330
assign_managed_vector(struct irq_data *irqd, const struct cpumask *dest)
arch/x86/kernel/apic/vector.c
332
const struct cpumask *affmsk = irq_data_get_affinity_mask(irqd);
arch/x86/kernel/apic/vector.c
439
const struct cpumask *dest = irq_data_get_affinity_mask(irqd);
arch/x86/kernel/apic/vector.c
486
const struct cpumask *dest = irq_data_get_affinity_mask(irqd);
arch/x86/kernel/apic/vector.c
75
const struct cpumask *mask)
arch/x86/kernel/apic/vector.c
880
const struct cpumask *dest, bool force)
arch/x86/kernel/apic/x2apic_cluster.c
105
static void prefill_clustermask(struct cpumask *cmsk, unsigned int cpu, u32 cluster)
arch/x86/kernel/apic/x2apic_cluster.c
110
struct cpumask **cpu_cmsk = &per_cpu(cluster_masks, cpu_i);
arch/x86/kernel/apic/x2apic_cluster.c
126
struct cpumask *cmsk = NULL;
arch/x86/kernel/apic/x2apic_cluster.c
196
struct cpumask *cmsk = per_cpu(cluster_masks, dead_cpu);
arch/x86/kernel/apic/x2apic_cluster.c
22
static DEFINE_PER_CPU_READ_MOSTLY(struct cpumask *, cluster_masks);
arch/x86/kernel/apic/x2apic_cluster.c
39
__x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest)
arch/x86/kernel/apic/x2apic_cluster.c
42
struct cpumask *tmpmsk;
arch/x86/kernel/apic/x2apic_cluster.c
58
struct cpumask *cmsk = per_cpu(cluster_masks, cpu);
arch/x86/kernel/apic/x2apic_cluster.c
75
static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_cluster.c
81
x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_cluster.c
93
struct cpumask *cmsk = this_cpu_read(cluster_masks);
arch/x86/kernel/apic/x2apic_phys.c
54
__x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest)
arch/x86/kernel/apic/x2apic_phys.c
75
static void x2apic_send_IPI_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_phys.c
81
x2apic_send_IPI_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_savic.c
259
static void send_ipi_mask(const struct cpumask *mask, unsigned int vector, bool excl_self)
arch/x86/kernel/apic/x2apic_savic.c
274
static void savic_send_ipi_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_savic.c
279
static void savic_send_ipi_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_uv_x.c
713
static void uv_send_IPI_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/apic/x2apic_uv_x.c
721
static void uv_send_IPI_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/kernel/cpu/microcode/core.c
441
struct cpumask *secondaries = topology_sibling_cpumask(cpu);
arch/x86/kernel/cpu/sgx/encl.c
928
cpumask_t *cpumask = &encl->cpumask;
arch/x86/kernel/cpu/sgx/encl.c
932
cpumask_clear(cpumask);
arch/x86/kernel/cpu/sgx/encl.c
940
cpumask_or(cpumask, cpumask, mm_cpumask(encl_mm->mm));
arch/x86/kernel/cpu/sgx/encl.c
947
return cpumask;
arch/x86/kernel/cpu/sgx/encl.h
63
cpumask_t cpumask;
arch/x86/kernel/cpu/topology_common.c
20
struct cpumask __cpu_primary_thread_mask __read_mostly;
arch/x86/kernel/hpet.c
407
evt->cpumask = cpumask_of(hc->cpu);
arch/x86/kernel/kvm.c
518
static void __send_ipi_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/kvm.c
574
static void kvm_send_ipi_mask(const struct cpumask *mask, int vector)
arch/x86/kernel/kvm.c
579
static void kvm_send_ipi_mask_allbutself(const struct cpumask *mask, int vector)
arch/x86/kernel/kvm.c
582
struct cpumask *new_mask = this_cpu_cpumask_var_ptr(__pv_cpu_mask);
arch/x86/kernel/kvm.c
583
const struct cpumask *local_mask;
arch/x86/kernel/kvm.c
650
static void kvm_smp_send_call_func_ipi(const struct cpumask *mask)
arch/x86/kernel/kvm.c
665
static void kvm_flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/kernel/kvm.c
671
struct cpumask *flushmask = this_cpu_cpumask_var_ptr(__pv_cpu_mask);
arch/x86/kernel/kvm.c
673
cpumask_copy(flushmask, cpumask);
arch/x86/kernel/nmi.c
640
void nmi_backtrace_stall_snap(const struct cpumask *btp)
arch/x86/kernel/nmi.c
654
void nmi_backtrace_stall_check(const struct cpumask *btp)
arch/x86/kernel/nmi_selftest.c
62
static void __init test_nmi_ipi(struct cpumask *mask)
arch/x86/kernel/process.c
819
struct cpumask cpus_stop_mask;
arch/x86/kernel/smpboot.c
538
const struct cpumask *cpus = cpumask_of_node(N + n);
arch/x86/kernel/smpboot.c
754
const struct cpumask *cpu_coregroup_mask(int cpu)
arch/x86/kernel/smpboot.c
759
const struct cpumask *cpu_clustergroup_mask(int cpu)
arch/x86/kernel/tsc.c
1587
const struct cpumask *mask = topology_core_cpumask(cpu);
arch/x86/kernel/tsc_sync.c
185
struct cpumask *mask;
arch/x86/lib/cache-smp.c
23
void wbinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/lib/cache-smp.c
40
void wbnoinvd_on_cpus_mask(struct cpumask *cpus)
arch/x86/lib/msr-smp.c
125
void rdmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs)
arch/x86/lib/msr-smp.c
139
void wrmsr_on_cpus(const struct cpumask *mask, u32 msr_no, struct msr __percpu *msrs)
arch/x86/lib/msr-smp.c
97
static void __rwmsr_on_cpus(const struct cpumask *mask, u32 msr_no,
arch/x86/mm/numa.c
373
struct cpumask *mask;
arch/x86/mm/numa.c
417
const struct cpumask *cpumask_of_node(int node)
arch/x86/mm/tlb.c
1347
STATIC_NOPV void native_flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/mm/tlb.c
1373
on_each_cpu_mask(cpumask, flush_tlb_func, (void *)info, true);
arch/x86/mm/tlb.c
1376
(void *)info, 1, cpumask);
arch/x86/mm/tlb.c
1379
void flush_tlb_multi(const struct cpumask *cpumask,
arch/x86/mm/tlb.c
1382
__flush_tlb_multi(cpumask, info);
arch/x86/mm/tlb.c
1745
} else if (cpumask_any_but(&batch->cpumask, cpu) < nr_cpu_ids) {
arch/x86/mm/tlb.c
1746
flush_tlb_multi(&batch->cpumask, info);
arch/x86/mm/tlb.c
1747
} else if (cpumask_test_cpu(cpu, &batch->cpumask)) {
arch/x86/mm/tlb.c
1754
cpumask_clear(&batch->cpumask);
arch/x86/platform/uv/uv_irq.c
51
uv_set_irq_affinity(struct irq_data *data, const struct cpumask *mask,
arch/x86/platform/uv/uv_time.c
294
int ced_cpu = cpumask_first(ced->cpumask);
arch/x86/platform/uv/uv_time.c
304
int ced_cpu = cpumask_first(evt->cpumask);
arch/x86/platform/uv/uv_time.c
337
ced->cpumask = cpumask_of(smp_processor_id());
arch/x86/xen/mmu_pv.c
1329
static void xen_flush_tlb_multi(const struct cpumask *cpus,
arch/x86/xen/smp.c
144
static void __xen_send_IPI_mask(const struct cpumask *mask,
arch/x86/xen/smp.c
153
void xen_smp_send_call_function_ipi(const struct cpumask *mask)
arch/x86/xen/smp.c
206
void xen_send_IPI_mask(const struct cpumask *mask,
arch/x86/xen/smp.c
231
void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
arch/x86/xen/time.c
368
evt->cpumask = cpumask_of(cpu);
arch/x86/xen/xen-ops.h
282
extern void xen_send_IPI_mask(const struct cpumask *mask,
arch/x86/xen/xen-ops.h
284
extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
arch/x86/xen/xen-ops.h
299
void xen_smp_send_call_function_ipi(const struct cpumask *mask);
arch/xtensa/include/asm/smp.h
22
struct cpumask;
arch/xtensa/include/asm/smp.h
23
void arch_send_call_function_ipi_mask(const struct cpumask *mask);
arch/xtensa/kernel/irq.c
173
const struct cpumask *mask;
arch/xtensa/kernel/smp.c
373
static void send_ipi_message(const struct cpumask *callmask,
arch/xtensa/kernel/smp.c
385
void arch_send_call_function_ipi_mask(const struct cpumask *mask)
arch/xtensa/kernel/smp.c
402
struct cpumask targets;
arch/xtensa/kernel/time.c
135
clockevent->cpumask = cpumask_of(cpu);
block/blk-mq-cpumap.c
112
const struct cpumask *mask;
block/blk-mq-cpumap.c
20
static unsigned int blk_mq_num_queues(const struct cpumask *mask,
block/blk-mq-cpumap.c
61
const struct cpumask *masks;
block/blk-mq-sysfs.c
38
free_cpumask_var(hctx->cpumask);
block/blk-mq-sysfs.c
87
for_each_cpu(i, hctx->cpumask) {
block/blk-mq.c
2246
int cpu = cpumask_first_and(hctx->cpumask, cpu_online_mask);
block/blk-mq.c
2249
cpu = cpumask_first(hctx->cpumask);
block/blk-mq.c
2279
next_cpu = cpumask_next_and(next_cpu, hctx->cpumask,
block/blk-mq.c
2381
if (async || !cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)) {
block/blk-mq.c
4037
if (!zalloc_cpumask_var_node(&hctx->cpumask, gfp, node))
block/blk-mq.c
4080
free_cpumask_var(hctx->cpumask);
block/blk-mq.c
4180
cpumask_clear(hctx->cpumask);
block/blk-mq.c
4219
if (cpumask_test_cpu(i, hctx->cpumask))
block/blk-mq.c
4222
cpumask_set_cpu(i, hctx->cpumask);
block/blk-mq.c
4275
for_each_cpu(cpu, hctx->cpumask) {
block/blk-mq.c
4277
cpumask_clear_cpu(cpu, hctx->cpumask);
block/blk-mq.c
749
cpu = cpumask_first_and(data.hctx->cpumask, cpu_online_mask);
drivers/accel/habanalabs/common/device.c
2875
struct cpumask *available_mask = &hdev->irq_affinity_mask;
drivers/accel/habanalabs/common/device.c
2877
static struct cpumask cpu_mask;
drivers/acpi/acpi_pad.c
136
struct cpumask *pad_busy_cpus = to_cpumask(pad_busy_cpus_bits);
drivers/acpi/acpi_pad.c
93
struct cpumask *pad_busy_cpus = to_cpumask(pad_busy_cpus_bits);
drivers/acpi/irq.c
303
const struct cpumask *acpi_irq_get_affinity(acpi_handle handle,
drivers/base/arch_numa.c
358
struct cpumask *mask;
drivers/base/arch_numa.c
44
const struct cpumask *cpumask_of_node(int node)
drivers/base/arch_topology.c
102
const struct cpumask *cpus)
drivers/base/arch_topology.c
141
void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq,
drivers/base/arch_topology.c
180
void topology_update_hw_pressure(const struct cpumask *cpus,
drivers/base/arch_topology.c
30
static struct cpumask scale_freq_counters_mask;
drivers/base/arch_topology.c
35
static bool supports_scale_freq_counters(const struct cpumask *cpus)
drivers/base/arch_topology.c
705
const struct cpumask *cpu_coregroup_mask(int cpu)
drivers/base/arch_topology.c
71
const struct cpumask *cpus)
drivers/base/arch_topology.c
732
const struct cpumask *cpu_clustergroup_mask(int cpu)
drivers/base/cacheinfo.c
675
const struct cpumask *mask = &this_leaf->shared_cpu_map;
drivers/base/cacheinfo.c
684
const struct cpumask *mask = &this_leaf->shared_cpu_map;
drivers/base/cacheinfo.c
780
const struct cpumask *mask = &this_leaf->shared_cpu_map;
drivers/base/cpu.c
212
const struct cpumask *const map;
drivers/base/cpu.c
306
const struct cpumask *hk_mask;
drivers/base/platform.c
153
static const struct cpumask *get_irq_affinity(struct platform_device *dev,
drivers/base/platform.c
156
const struct cpumask *mask = NULL;
drivers/base/platform.c
183
const struct cpumask **affinity)
drivers/block/mtip32xx/mtip32xx.c
3591
const struct cpumask *node_mask;
drivers/block/mtip32xx/mtip32xx.c
3694
const struct cpumask *node_mask;
drivers/block/ublk_drv.c
4471
cpumask_var_t cpumask;
drivers/block/ublk_drv.c
4488
if (!zalloc_cpumask_var(&cpumask, GFP_KERNEL))
drivers/block/ublk_drv.c
4493
cpumask_set_cpu(i, cpumask);
drivers/block/ublk_drv.c
4498
if (copy_to_user(argp, cpumask, retlen))
drivers/block/ublk_drv.c
4506
free_cpumask_var(cpumask);
drivers/clocksource/arc_timer.c
309
evt->cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/arm_arch_timer.c
687
clk->cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/arm_arch_timer_mmio.c
255
.cpumask = cpu_possible_mask,
drivers/clocksource/arm_global_timer.c
185
clk->cpumask = cpumask_of(cpu);
drivers/clocksource/asm9260_timer.c
237
event_dev.cpumask = cpumask_of(0);
drivers/clocksource/bcm2835_timer.c
114
timer->evt.cpumask = cpumask_of(0);
drivers/clocksource/bcm_kona_timer.c
139
kona_clockevent_timer.cpumask = cpumask_of(0);
drivers/clocksource/clps711x-timer.c
69
clkevt->cpumask = cpumask_of(0);
drivers/clocksource/dummy_timer.c
25
evt->cpumask = cpumask_of(cpu);
drivers/clocksource/dw_apb_timer.c
108
cpumask_first(evt->cpumask));
drivers/clocksource/dw_apb_timer.c
122
cpumask_first(evt->cpumask));
drivers/clocksource/dw_apb_timer.c
155
cpumask_first(evt->cpumask));
drivers/clocksource/dw_apb_timer.c
179
cpumask_first(evt->cpumask));
drivers/clocksource/dw_apb_timer.c
241
dw_ced->ced.cpumask = cpu < 0 ? cpu_possible_mask : cpumask_of(cpu);
drivers/clocksource/em_sti.c
269
ced->cpumask = cpu_possible_mask;
drivers/clocksource/exynos_mct.c
347
mct_comp_device.cpumask = cpumask_of(0);
drivers/clocksource/exynos_mct.c
464
evt->cpumask = cpumask_of(cpu);
drivers/clocksource/hyperv_timer.c
141
ce->cpumask = cpumask_of(cpu);
drivers/clocksource/i8253.c
208
i8253_clockevent.cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/ingenic-sysost.c
352
ost->cevt.cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/ingenic-timer.c
183
timer->cevt.cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/jcore-pit.c
240
pit->ced.cpumask = cpumask_of(cpu);
drivers/clocksource/mips-gic-timer.c
55
int cpu = cpumask_first(evt->cpumask);
drivers/clocksource/mips-gic-timer.c
89
cd->cpumask = cpumask_of(cpu);
drivers/clocksource/mps2-timer.c
151
ce->clkevt.cpumask = cpu_possible_mask;
drivers/clocksource/mxs_timer.c
166
mxs_clockevent_device.cpumask = cpumask_of(0);
drivers/clocksource/nomadik-mtu.c
233
nmdk_clkevt.cpumask = cpumask_of(0);
drivers/clocksource/numachip.c
68
ced->cpumask = cpumask_of(smp_processor_id());
drivers/clocksource/renesas-ostm.c
155
ced->cpumask = cpumask_of(0);
drivers/clocksource/samsung_pwm_timer.c
274
time_event_device.cpumask = cpumask_of(0);
drivers/clocksource/sh_cmt.c
852
ced->cpumask = cpu_possible_mask;
drivers/clocksource/sh_mtu2.c
316
ced->cpumask = cpu_possible_mask;
drivers/clocksource/sh_tmu.c
400
ced->cpumask = cpu_possible_mask;
drivers/clocksource/timer-armada-370-xp.c
192
evt->cpumask = cpumask_of(cpu);
drivers/clocksource/timer-atmel-pit.c
249
data->clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-atmel-st.c
242
clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-atmel-tcb.c
287
clkevt.clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-cadence-ttc.c
453
ttcce->ce.cpumask = cpu_possible_mask;
drivers/clocksource/timer-clint.c
133
ce->cpumask = cpumask_of(cpu);
drivers/clocksource/timer-davinci.c
282
clockevent->dev.cpumask = cpumask_of(0);
drivers/clocksource/timer-digicolor.c
196
dc_timer_dev.ce.cpumask = cpu_possible_mask;
drivers/clocksource/timer-econet-en751221.c
155
cd->cpumask = cpumask_of(i);
drivers/clocksource/timer-econet-en751221.c
68
int cpu = cpumask_first(dev->cpumask);
drivers/clocksource/timer-econet-en751221.c
86
cpu = cpumask_first(dev->cpumask);
drivers/clocksource/timer-fsl-ftm.c
195
ftm_clockevent.cpumask = cpumask_of(0);
drivers/clocksource/timer-fttmr010.c
409
fttmr010->clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-gx6605s.c
86
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-gxp.c
115
gxp_timer->evt.cpumask = cpumask_of(0);
drivers/clocksource/timer-imx-gpt.c
282
ced->cpumask = cpumask_of(0);
drivers/clocksource/timer-imx-sysctr.c
157
to_sysctr.clkevt.cpumask = cpu_possible_mask;
drivers/clocksource/timer-imx-tpm.c
144
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-ixp4xx.c
207
tmr->clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-keystone.c
210
event_dev->cpumask = cpu_possible_mask;
drivers/clocksource/timer-mediatek-cpux.c
93
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-mediatek.c
276
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-meson6.c
198
meson6_clockevent.cpumask = cpu_possible_mask;
drivers/clocksource/timer-microchip-pit64b.c
417
ce->clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-milbeaut.c
149
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-mp-csky.c
80
to->clkevt.cpumask = cpumask_of(cpu);
drivers/clocksource/timer-msc313e.c
190
msc313e_clkevt.cpumask = cpu_possible_mask;
drivers/clocksource/timer-npcm7xx.c
163
npcm7xx_to.clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-nxp-pit.c
221
pit->ced.cpumask = cpumask_of(cpu);
drivers/clocksource/timer-nxp-stm.c
313
stm_timer->ced.cpumask = cpumask_of(cpu);
drivers/clocksource/timer-orion.c
175
orion_clkevt.cpumask = cpumask_of(0);
drivers/clocksource/timer-owl.c
166
owl_clockevent.cpumask = cpumask_of(0);
drivers/clocksource/timer-pxa.c
155
ckevt_pxa_osmr0.cpumask = cpumask_of(0);
drivers/clocksource/timer-qcom.c
112
evt->cpumask = cpumask_of(cpu);
drivers/clocksource/timer-realtek.c
121
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-riscv.c
115
ce->cpumask = cpumask_of(cpu);
drivers/clocksource/timer-rockchip.c
228
ce->cpumask = cpu_possible_mask;
drivers/clocksource/timer-rtl-otto.c
251
to->clkevt.cpumask = cpumask_of(cpu);
drivers/clocksource/timer-rtl-otto.c
252
irq_force_affinity(to->of_irq.irq, to->clkevt.cpumask);
drivers/clocksource/timer-sp804.c
245
evt->cpumask = cpu_possible_mask;
drivers/clocksource/timer-sprd.c
136
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-stm32-lp.c
204
priv->clkevt.cpumask = cpu_possible_mask;
drivers/clocksource/timer-sun4i.c
154
.cpumask = cpu_possible_mask,
drivers/clocksource/timer-sun5i.c
217
ce->clkevt.cpumask = cpu_possible_mask;
drivers/clocksource/timer-tegra.c
316
cpu_to->clkevt.cpumask = cpumask_of(cpu);
drivers/clocksource/timer-ti-dm-systimer.c
536
const struct cpumask *cpumask,
drivers/clocksource/timer-ti-dm-systimer.c
559
dev->cpumask = cpumask;
drivers/clocksource/timer-vt8500.c
133
clockevent.cpumask = cpumask_of(0);
drivers/clocksource/timer-zevio.c
160
timer->clkevt.cpumask = cpu_possible_mask;
drivers/cpufreq/acpi-cpufreq.c
307
static u32 drv_read(struct acpi_cpufreq_data *data, const struct cpumask *mask)
drivers/cpufreq/acpi-cpufreq.c
329
const struct cpumask *mask, u32 val)
drivers/cpufreq/acpi-cpufreq.c
341
static u32 get_cur_val(const struct cpumask *mask, struct acpi_cpufreq_data *data)
drivers/cpufreq/acpi-cpufreq.c
388
const struct cpumask *mask, unsigned int freq)
drivers/cpufreq/acpi-cpufreq.c
408
const struct cpumask *mask;
drivers/cpufreq/armada-8k-cpufreq.c
135
static struct cpumask cpus, shared_cpus;
drivers/cpufreq/armada-8k-cpufreq.c
46
struct cpumask *cpumask)
drivers/cpufreq/armada-8k-cpufreq.c
65
cpumask_set_cpu(cpu, cpumask);
drivers/cpufreq/cpufreq.c
3063
bool cpufreq_ready_for_eas(const struct cpumask *cpu_mask)
drivers/cpufreq/cpufreq.c
883
ssize_t cpufreq_show_cpus(const struct cpumask *mask, char *buf)
drivers/cpufreq/mediatek-cpufreq.c
40
struct cpumask cpus;
drivers/cpufreq/powernow-k8.h
38
struct cpumask *available_cores;
drivers/cpufreq/qcom-cpufreq-hw.c
303
static void qcom_get_related_cpus(int index, struct cpumask *m)
drivers/cpufreq/qoriq-cpufreq.c
88
struct cpumask *dstp = policy->cpus;
drivers/cpufreq/scmi-cpufreq.c
110
struct cpumask *cpumask)
drivers/cpufreq/scmi-cpufreq.c
125
cpumask_set_cpu(cpu, cpumask);
drivers/cpufreq/scpi-cpufreq.c
66
scpi_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask)
drivers/cpufreq/scpi-cpufreq.c
85
cpumask_set_cpu(cpu, cpumask);
drivers/cpufreq/vexpress-spc-cpufreq.c
280
const struct cpumask *cpumask)
drivers/cpufreq/vexpress-spc-cpufreq.c
292
const struct cpumask *cpumask)
drivers/cpufreq/vexpress-spc-cpufreq.c
301
return _put_cluster_clk_and_freq_table(cpu_dev, cpumask);
drivers/cpufreq/vexpress-spc-cpufreq.c
309
_put_cluster_clk_and_freq_table(cdev, cpumask);
drivers/cpufreq/vexpress-spc-cpufreq.c
317
const struct cpumask *cpumask)
drivers/cpufreq/vexpress-spc-cpufreq.c
353
const struct cpumask *cpumask)
drivers/cpufreq/vexpress-spc-cpufreq.c
362
ret = _get_cluster_clk_and_freq_table(cpu_dev, cpumask);
drivers/cpufreq/vexpress-spc-cpufreq.c
378
ret = _get_cluster_clk_and_freq_table(cdev, cpumask);
drivers/cpufreq/vexpress-spc-cpufreq.c
401
_put_cluster_clk_and_freq_table(cdev, cpumask);
drivers/cpuidle/cpuidle-arm.c
87
drv->cpumask = (struct cpumask *)cpumask_of(cpu);
drivers/cpuidle/cpuidle-big_little.c
144
struct cpumask *cpumask;
drivers/cpuidle/cpuidle-big_little.c
147
cpumask = kzalloc(cpumask_size(), GFP_KERNEL);
drivers/cpuidle/cpuidle-big_little.c
148
if (!cpumask)
drivers/cpuidle/cpuidle-big_little.c
153
cpumask_set_cpu(cpu, cpumask);
drivers/cpuidle/cpuidle-big_little.c
155
drv->cpumask = cpumask;
drivers/cpuidle/cpuidle-big_little.c
218
kfree(bl_idle_big_driver.cpumask);
drivers/cpuidle/cpuidle-big_little.c
220
kfree(bl_idle_little_driver.cpumask);
drivers/cpuidle/cpuidle-powernv.c
225
drv->cpumask = (struct cpumask *)cpu_present_mask;
drivers/cpuidle/cpuidle-psci.c
379
drv->cpumask = (struct cpumask *)cpumask_of(cpu);
drivers/cpuidle/cpuidle-qcom-spm.c
120
data->cpuidle_driver.cpumask = (struct cpumask *)cpumask_of(cpu);
drivers/cpuidle/cpuidle-riscv-sbi.c
300
drv->cpumask = (struct cpumask *)cpumask_of(cpu);
drivers/cpuidle/cpuidle.c
753
for_each_cpu(cpu, drv->cpumask) {
drivers/cpuidle/cpuidle.c
774
const struct cpumask *const coupled_cpus)
drivers/cpuidle/cpuidle.c
785
for_each_cpu(cpu, drv->cpumask) {
drivers/cpuidle/driver.c
166
if (!drv->cpumask)
drivers/cpuidle/driver.c
167
drv->cpumask = (struct cpumask *)cpu_possible_mask;
drivers/cpuidle/driver.c
243
on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer,
drivers/cpuidle/driver.c
262
on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer,
drivers/cpuidle/driver.c
381
if (!drv->cpumask) {
drivers/cpuidle/driver.c
386
for_each_cpu(cpu, drv->cpumask) {
drivers/cpuidle/driver.c
55
for_each_cpu(cpu, drv->cpumask) {
drivers/cpuidle/driver.c
75
for_each_cpu(cpu, drv->cpumask) {
drivers/cpuidle/driver.c
83
for_each_cpu(cpu, drv->cpumask)
drivers/cpuidle/dt_idle_states.c
109
cpu = cpumask_first(cpumask) + 1;
drivers/cpuidle/dt_idle_states.c
110
for_each_cpu_from(cpu, cpumask) {
drivers/cpuidle/dt_idle_states.c
152
const cpumask_t *cpumask;
drivers/cpuidle/dt_idle_states.c
163
cpumask = drv->cpumask ? : cpu_possible_mask;
drivers/cpuidle/dt_idle_states.c
164
cpu_node = of_cpu_device_node_get(cpumask_first(cpumask));
drivers/cpuidle/dt_idle_states.c
182
if (!idle_state_valid(state_node, i, cpumask)) {
drivers/cpuidle/dt_idle_states.c
97
const cpumask_t *cpumask)
drivers/crypto/intel/iaa/iaa_crypto_main.c
889
const struct cpumask *node_cpus;
drivers/devfreq/event/rockchip-dfi.c
361
__ATTR(cpumask, 0444, ddr_perf_cpumask_show, NULL);
drivers/devfreq/hisi_uncore_freq.c
465
const struct cpumask *(*get_cpumask)(int cpu))
drivers/devfreq/hisi_uncore_freq.c
505
static const struct cpumask *get_package_cpumask(int cpu)
drivers/devfreq/hisi_uncore_freq.c
515
static const struct cpumask *get_cluster_cpumask(int cpu)
drivers/devfreq/hisi_uncore_freq.c
84
struct cpumask related_cpus;
drivers/edac/a72_edac.c
46
static struct cpumask compat_mask;
drivers/edac/amd64_edac.c
3195
static void get_cpus_on_this_dct_cpumask(struct cpumask *mask, u16 nid)
drivers/edac/cpc925_edac.c
618
u32 cpumask;
drivers/edac/cpc925_edac.c
622
cpumask = cpc925_cpu_mask_disabled();
drivers/edac/cpc925_edac.c
623
if (apimask & cpumask) {
drivers/edac/cpc925_edac.c
626
apimask &= ~cpumask;
drivers/firmware/psci/psci_checker.c
167
const struct cpumask *cpu_group =
drivers/firmware/psci/psci_checker.c
77
static unsigned int down_and_up_cpus(const struct cpumask *cpus,
drivers/firmware/psci/psci_checker.c
78
struct cpumask *offlined_cpus)
drivers/fpga/dfl-fme-perf.c
188
static DEVICE_ATTR_RO(cpumask);
drivers/gpio/gpio-realtek-otto.c
274
const struct cpumask *dest, bool force)
drivers/gpio/gpio-realtek-otto.c
71
struct cpumask cpu_irq_maskable;
drivers/gpio/gpio-sifive.c
133
const struct cpumask *dest,
drivers/gpio/gpio-tegra.c
573
const struct cpumask *dest,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1752
const struct cpumask *cpumask;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1765
cpumask = cpumask_of_node(numa_node_id);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1774
sub_type_hdr->num_cpu_cores = cpumask_weight(cpumask);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2331
static int kfd_cpumask_to_apic_id(const struct cpumask *cpumask)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2335
if (!cpumask || cpumask == cpu_none_mask)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2337
first_cpu_of_numa_node = cpumask_first(cpumask);
drivers/hv/channel_mgmt.c
746
struct cpumask *allocated_mask;
drivers/hv/channel_mgmt.c
747
const struct cpumask *hk_mask = housekeeping_cpumask(HK_TYPE_MANAGED_IRQ);
drivers/hv/hv.c
186
hv_context.hv_numa_map = kzalloc_objs(struct cpumask, nr_node_ids);
drivers/hv/hyperv_vmbus.h
177
struct cpumask *hv_numa_map;
drivers/hwmon/coretemp.c
719
if (!cpumask_intersects(&pdata->cpumask, topology_sibling_cpumask(cpu)))
drivers/hwmon/coretemp.c
722
cpumask_set_cpu(cpu, &pdata->cpumask);
drivers/hwmon/coretemp.c
744
cpumask_clear_cpu(cpu, &pd->cpumask);
drivers/hwmon/coretemp.c
751
target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu));
drivers/hwmon/coretemp.c
764
if (cpumask_empty(&pd->cpumask)) {
drivers/hwmon/coretemp.c
777
target = cpumask_first(&pd->cpumask);
drivers/hwmon/coretemp.c
96
struct cpumask cpumask;
drivers/hwtracing/coresight/coresight-trbe.c
1477
const struct cpumask *affinity;
drivers/hwtracing/ptt/hisi_ptt.c
781
const cpumask_t *cpumask = cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev));
drivers/hwtracing/ptt/hisi_ptt.c
783
return cpumap_print_to_pagebuf(true, buf, cpumask);
drivers/hwtracing/ptt/hisi_ptt.c
785
static DEVICE_ATTR_RO(cpumask);
drivers/infiniband/hw/hfi1/affinity.c
490
struct cpumask *dev_comp_vect_mask = &dd->comp_vect->mask;
drivers/infiniband/hw/hfi1/affinity.c
585
const struct cpumask *local_mask;
drivers/infiniband/hw/hfi1/affinity.c
981
const struct cpumask *node_mask,
drivers/infiniband/hw/hfi1/affinity.h
28
struct cpumask mask;
drivers/infiniband/hw/hfi1/affinity.h
29
struct cpumask used;
drivers/infiniband/hw/hfi1/affinity.h
64
struct cpumask general_intr_mask;
drivers/infiniband/hw/hfi1/affinity.h
65
struct cpumask comp_vect_mask;
drivers/infiniband/hw/hfi1/affinity.h
71
struct cpumask real_cpu_mask;
drivers/infiniband/hw/hfi1/netdev.h
89
struct cpumask *cpu_mask);
drivers/infiniband/hw/hfi1/netdev_rx.c
157
struct cpumask *cpu_mask)
drivers/infiniband/hw/hfi1/sdma.h
366
struct cpumask cpu_mask;
drivers/infiniband/sw/siw/siw_main.c
107
struct cpumask **tx_valid_cpus;
drivers/infiniband/sw/siw/siw_main.c
131
kzalloc_objs(struct cpumask *, num_nodes);
drivers/infiniband/sw/siw/siw_main.c
138
kzalloc_obj(struct cpumask);
drivers/infiniband/sw/siw/siw_main.c
162
const struct cpumask *tx_cpumask;
drivers/iommu/amd/init.c
2458
const struct cpumask *mask, bool force)
drivers/iommu/amd/iommu.c
4110
const struct cpumask *mask, bool force)
drivers/iommu/hyperv-iommu.c
198
const struct cpumask *affinity;
drivers/iommu/hyperv-iommu.c
241
const struct cpumask *mask, bool force)
drivers/iommu/hyperv-iommu.c
40
const struct cpumask *mask, bool force)
drivers/iommu/intel/irq_remapping.c
1212
intel_ir_set_affinity(struct irq_data *data, const struct cpumask *mask,
drivers/irqchip/exynos-combiner.c
94
const struct cpumask *mask_val, bool force)
drivers/irqchip/irq-apple-aic.c
429
const struct cpumask *mask_val, bool force)
drivers/irqchip/irq-apple-aic.c
641
const struct cpumask *mask;
drivers/irqchip/irq-armada-370-xp.c
246
static int mpic_msi_set_affinity(struct irq_data *d, const struct cpumask *mask, bool force)
drivers/irqchip/irq-armada-370-xp.c
421
static void mpic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
drivers/irqchip/irq-armada-370-xp.c
512
static int mpic_set_affinity(struct irq_data *d, const struct cpumask *mask_val, bool force)
drivers/irqchip/irq-aspeed-scu-ic.c
173
const struct cpumask *dest,
drivers/irqchip/irq-bcm2836.c
182
const struct cpumask *mask)
drivers/irqchip/irq-bcm6345-l1.c
113
return cpumask_first_and(&intc->cpumask, irq_data_get_affinity_mask(d));
drivers/irqchip/irq-bcm6345-l1.c
187
const struct cpumask *dest,
drivers/irqchip/irq-bcm6345-l1.c
198
new_cpu = cpumask_first_and_and(&intc->cpumask, dest, cpu_online_mask);
drivers/irqchip/irq-bcm6345-l1.c
308
cpumask_set_cpu(idx, &intc->cpumask);
drivers/irqchip/irq-bcm6345-l1.c
311
if (cpumask_empty(&intc->cpumask)) {
drivers/irqchip/irq-bcm6345-l1.c
328
for_each_cpu(idx, &intc->cpumask) {
drivers/irqchip/irq-bcm6345-l1.c
79
struct cpumask cpumask;
drivers/irqchip/irq-bcm7038-l1.c
189
const struct cpumask *dest,
drivers/irqchip/irq-csky-mpintc.c
128
const struct cpumask *mask_val,
drivers/irqchip/irq-csky-mpintc.c
213
static void csky_mpintc_send_ipi(const struct cpumask *mask)
drivers/irqchip/irq-gic-v3-its.c
1675
const struct cpumask *cpu_mask)
drivers/irqchip/irq-gic-v3-its.c
1696
const struct cpumask *aff_mask)
drivers/irqchip/irq-gic-v3-its.c
1700
static struct cpumask __tmpmask;
drivers/irqchip/irq-gic-v3-its.c
1701
struct cpumask *tmpmask;
drivers/irqchip/irq-gic-v3-its.c
1767
static int its_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v3-its.c
3910
const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v3-its.c
3915
struct cpumask *table_mask;
drivers/irqchip/irq-gic-v3-its.c
4373
const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v3.c
1331
static u16 gic_compute_target_list(int *base_cpu, const struct cpumask *mask,
drivers/irqchip/irq-gic-v3.c
1379
static void gic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
drivers/irqchip/irq-gic-v3.c
1428
static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v3.c
1715
const struct cpumask *mask = NULL;
drivers/irqchip/irq-gic-v3.c
48
static struct cpumask broken_rdists __read_mostly __maybe_unused;
drivers/irqchip/irq-gic-v5.c
247
const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v5.c
274
const struct cpumask *mask_val,
drivers/irqchip/irq-gic-v5.c
282
const struct cpumask *mask_val,
drivers/irqchip/irq-gic.c
467
u32 cpumask;
drivers/irqchip/irq-gic.c
476
cpumask = gic_get_cpumask(gic);
drivers/irqchip/irq-gic.c
477
cpumask |= cpumask << 8;
drivers/irqchip/irq-gic.c
478
cpumask |= cpumask << 16;
drivers/irqchip/irq-gic.c
480
writel_relaxed(cpumask, base + GIC_DIST_TARGET + i * 4 / 4);
drivers/irqchip/irq-gic.c
793
static int gic_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
drivers/irqchip/irq-gic.c
820
static void gic_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
drivers/irqchip/irq-hip04.c
147
const struct cpumask *mask_val,
drivers/irqchip/irq-hip04.c
175
static void hip04_ipi_send_mask(struct irq_data *d, const struct cpumask *mask)
drivers/irqchip/irq-hip04.c
249
u32 cpumask;
drivers/irqchip/irq-hip04.c
258
cpumask = hip04_get_cpumask(intc);
drivers/irqchip/irq-hip04.c
259
cpumask |= cpumask << 16;
drivers/irqchip/irq-hip04.c
261
writel_relaxed(cpumask, base + GIC_DIST_TARGET + ((i * 2) & ~3));
drivers/irqchip/irq-imx-mu-msi.c
131
const struct cpumask *mask, bool force)
drivers/irqchip/irq-loongarch-avec.c
102
static int avecintc_set_affinity(struct irq_data *data, const struct cpumask *dest, bool force)
drivers/irqchip/irq-loongarch-avec.c
34
static struct cpumask intersect_mask;
drivers/irqchip/irq-loongson-eiointc.c
145
static int eiointc_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, bool force)
drivers/irqchip/irq-ls-scfg-msi.c
105
const struct cpumask *mask, bool force)
drivers/irqchip/irq-ls-scfg-msi.c
94
const struct cpumask *mask;
drivers/irqchip/irq-mips-gic.c
368
static int gic_set_affinity(struct irq_data *d, const struct cpumask *cpumask,
drivers/irqchip/irq-mips-gic.c
381
cpu = cpumask_first(cpumask);
drivers/irqchip/irq-mips-gic.c
383
cpu = cpumask_first_and(cpumask, cpu_online_mask);
drivers/irqchip/irq-mips-gic.c
755
struct cpumask *ipimask = arg;
drivers/irqchip/irq-mvebu-sei.c
96
const struct cpumask *mask_val,
drivers/irqchip/irq-ompic.c
99
static void ompic_raise_softirq(const struct cpumask *mask,
drivers/irqchip/irq-riscv-aplic-direct.c
31
struct cpumask lmask;
drivers/irqchip/irq-riscv-aplic-direct.c
52
static int aplic_direct_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
drivers/irqchip/irq-riscv-imsic-platform.c
101
static int imsic_irq_set_affinity(struct irq_data *d, const struct cpumask *mask_val,
drivers/irqchip/irq-riscv-imsic-state.c
437
struct imsic_vector *imsic_vector_alloc(unsigned int irq, const struct cpumask *mask)
drivers/irqchip/irq-riscv-imsic-state.h
98
struct imsic_vector *imsic_vector_alloc(unsigned int irq, const struct cpumask *mask);
drivers/irqchip/irq-sifive-plic.c
135
static inline void plic_irq_toggle(const struct cpumask *mask,
drivers/irqchip/irq-sifive-plic.c
192
const struct cpumask *mask_val, bool force)
drivers/irqchip/irq-sifive-plic.c
72
struct cpumask lmask;
drivers/irqchip/irq-stm32mp-exti.c
420
static int stm32mp_exti_set_affinity(struct irq_data *d, const struct cpumask *dest, bool force)
drivers/irqchip/irq-ti-sci-inta.c
503
const struct cpumask *mask_val, bool force)
drivers/irqchip/irq-xtensa-mx.c
132
const struct cpumask *dest, bool force)
drivers/md/dm-ps-io-affinity.c
105
free_cpumask_var(pi->cpumask);
drivers/md/dm-ps-io-affinity.c
16
cpumask_var_t cpumask;
drivers/md/dm-ps-io-affinity.c
170
DMEMIT("%*pb ", cpumask_pr_args(pi->cpumask));
drivers/md/dm-ps-io-affinity.c
200
const struct cpumask *cpumask;
drivers/md/dm-ps-io-affinity.c
218
cpumask = cpumask_of_node(node);
drivers/md/dm-ps-io-affinity.c
219
for_each_cpu(i, cpumask) {
drivers/md/dm-ps-io-affinity.c
36
free_cpumask_var(pi->cpumask);
drivers/md/dm-ps-io-affinity.c
66
if (!zalloc_cpumask_var(&pi->cpumask, GFP_KERNEL)) {
drivers/md/dm-ps-io-affinity.c
72
ret = cpumask_parse(argv[0], pi->cpumask);
drivers/md/dm-ps-io-affinity.c
79
for_each_cpu(cpu, pi->cpumask) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12950
const struct cpumask *cpu_mask_ptr;
drivers/net/ethernet/cavium/liquidio/octeon_device.h
413
struct cpumask affinity_mask;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h
587
struct cpumask dpio_cpumask;
drivers/net/ethernet/google/gve/gve_main.c
343
const struct cpumask *aff_mask;
drivers/net/ethernet/google/gve/gve_main.c
471
static const struct cpumask *gve_get_node_mask(struct gve_priv *priv)
drivers/net/ethernet/google/gve/gve_main.c
482
const struct cpumask *node_mask;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h
105
struct cpumask affinity_mask;
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1154
struct cpumask *mask;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
373
const struct cpumask *aff_mask;
drivers/net/ethernet/mellanox/mlx5/core/en.h
785
const struct cpumask *aff_mask;
drivers/net/ethernet/mellanox/mlx5/core/en.h
890
cpumask_var_t cpumask;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3159
cpumask_clear(priv->scratchpad.cpumask);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3165
cpumask_set_cpu(cpu, priv->scratchpad.cpumask);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3168
netif_set_xps_queue(priv->netdev, priv->scratchpad.cpumask, ix);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6226
if (!alloc_cpumask_var(&priv->scratchpad.cpumask, GFP_KERNEL))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6280
free_cpumask_var(priv->scratchpad.cpumask);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6302
free_cpumask_var(priv->scratchpad.cpumask);
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1106
static struct cpumask *
drivers/net/ethernet/mellanox/mlx5/core/eq.c
1121
struct cpumask *mask;
drivers/net/ethernet/mellanox/mlx5/core/eq.c
66
struct cpumask used_cpus;
drivers/net/ethernet/mellanox/mlx5/core/irq_affinity.c
108
struct cpumask *iter_mask = mlx5_irq_get_affinity_mask(iter);
drivers/net/ethernet/mellanox/mlx5/core/irq_affinity.c
20
const struct cpumask *req_mask)
drivers/net/ethernet/mellanox/mlx5/core/irq_affinity.c
97
irq_pool_find_least_loaded(struct mlx5_irq_pool *pool, const struct cpumask *req_mask)
drivers/net/ethernet/mellanox/mlx5/core/mlx5_irq.h
40
struct cpumask *mlx5_irq_get_affinity_mask(struct mlx5_irq *irq);
drivers/net/ethernet/mellanox/mlx5/core/mlx5_irq.h
46
struct cpumask *used_cpus, u16 vecidx);
drivers/net/ethernet/mellanox/mlx5/core/mlx5_irq.h
54
struct cpumask *used_cpus, u16 vecidx)
drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
364
struct cpumask *mlx5_irq_get_affinity_mask(struct mlx5_irq *irq)
drivers/net/ethernet/microsoft/mana/gdma_main.c
1587
const struct cpumask *next, *prev = cpu_none_mask;
drivers/net/ethernet/sfc/efx_channels.c
365
const struct cpumask *numa_mask = cpumask_of_pcibus(efx->pci_dev->bus);
drivers/net/ethernet/sfc/siena/efx_channels.c
366
const struct cpumask *numa_mask = cpumask_of_pcibus(efx->pci_dev->bus);
drivers/net/wireless/ath/ath11k/pci.c
937
const struct cpumask *m)
drivers/net/wireless/ath/ath12k/pci.c
615
const struct cpumask *m)
drivers/nvdimm/nd_perf.c
134
const struct cpumask *cpumask;
drivers/nvdimm/nd_perf.c
157
cpumask = cpumask_of_node(nodeid);
drivers/nvdimm/nd_perf.c
158
target = cpumask_any_but(cpumask, cpu);
drivers/nvdimm/nd_perf.c
220
const struct cpumask *cpumask;
drivers/nvdimm/nd_perf.c
233
cpumask = cpumask_of_node(nodeid);
drivers/nvdimm/nd_perf.c
234
nd_pmu->cpu = cpumask_any(cpumask);
drivers/of/irq.c
552
const struct cpumask *of_irq_get_affinity(struct device_node *dev, int index)
drivers/opp/cpu.c
107
void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask,
drivers/opp/cpu.c
113
WARN_ON(cpumask_empty(cpumask));
drivers/opp/cpu.c
115
for_each_cpu(cpu, cpumask) {
drivers/opp/cpu.c
138
void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask)
drivers/opp/cpu.c
140
_dev_pm_opp_cpumask_remove_table(cpumask, -1);
drivers/opp/cpu.c
155
const struct cpumask *cpumask)
drivers/opp/cpu.c
167
for_each_cpu(cpu, cpumask) {
drivers/opp/cpu.c
203
int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask)
drivers/opp/cpu.c
216
cpumask_clear(cpumask);
drivers/opp/cpu.c
221
cpumask_set_cpu(opp_dev->dev->id, cpumask);
drivers/opp/cpu.c
223
cpumask_set_cpu(cpu_dev->id, cpumask);
drivers/opp/of.c
1202
void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask)
drivers/opp/of.c
1204
_dev_pm_opp_cpumask_remove_table(cpumask, -1);
drivers/opp/of.c
1214
int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask)
drivers/opp/of.c
1219
if (WARN_ON(cpumask_empty(cpumask)))
drivers/opp/of.c
1222
for_each_cpu(cpu, cpumask) {
drivers/opp/of.c
1248
_dev_pm_opp_cpumask_remove_table(cpumask, cpu);
drivers/opp/of.c
1272
struct cpumask *cpumask)
drivers/opp/of.c
1285
cpumask_set_cpu(cpu_dev->id, cpumask);
drivers/opp/of.c
1315
cpumask_set_cpu(cpu, cpumask);
drivers/opp/of.c
1542
int dev_pm_opp_of_register_em(struct device *dev, struct cpumask *cpus)
drivers/opp/opp.h
262
void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, int last_cpu);
drivers/parisc/dino.c
346
static int dino_set_affinity_irq(struct irq_data *d, const struct cpumask *dest,
drivers/parisc/dino.c
350
struct cpumask tmask;
drivers/parisc/gsc.c
139
static int gsc_set_affinity_irq(struct irq_data *d, const struct cpumask *dest,
drivers/parisc/gsc.c
143
struct cpumask tmask;
drivers/parisc/iosapic.c
660
const struct cpumask *dest, bool force)
drivers/pci/controller/pci-hyperv.c
1803
static int hv_compose_msi_req_get_cpu(const struct cpumask *affinity)
drivers/pci/controller/pci-hyperv.c
1882
const struct cpumask *dest;
drivers/pci/controller/pci-hyperv.c
616
const struct cpumask *dest;
drivers/pci/controller/pci-xgene-msi.c
150
const struct cpumask *mask, bool force)
drivers/pci/controller/pcie-iproc-msi.c
208
const struct cpumask *mask, bool force)
drivers/pci/msi/api.c
340
const struct cpumask *pci_irq_get_affinity(struct pci_dev *dev, int nr)
drivers/pci/pci-driver.c
1717
static const struct cpumask *pci_device_irq_get_affinity(struct device *dev,
drivers/pci/pci-sysfs.c
107
const struct cpumask *mask;
drivers/pci/pci-sysfs.c
140
const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev));
drivers/pci/pci-sysfs.c
142
return cpumap_print_to_pagebuf(false, buf, cpumask);
drivers/pci/pci-sysfs.c
149
const struct cpumask *cpumask = cpumask_of_pcibus(to_pci_bus(dev));
drivers/pci/pci-sysfs.c
151
return cpumap_print_to_pagebuf(true, buf, cpumask);
drivers/perf/alibaba_uncore_drw_pmu.c
228
__ATTR(cpumask, 0444, ali_drw_pmu_cpumask_show, NULL);
drivers/perf/amlogic/meson_ddr_pmu_core.c
198
__ATTR(cpumask, 0444, meson_ddr_perf_cpumask_show, NULL);
drivers/perf/arm-cci.c
1358
__ATTR(cpumask, S_IRUGO, pmu_cpumask_attr_show, NULL);
drivers/perf/arm-ccn.c
545
__ATTR(cpumask, S_IRUGO, arm_ccn_pmu_cpumask_show, NULL);
drivers/perf/arm-cmn.c
1333
__ATTR(cpumask, 0444, arm_cmn_cpumask_show, NULL);
drivers/perf/arm-ni.c
246
__ATTR(cpumask, 0444, arm_ni_cpumask_show, NULL);
drivers/perf/arm_cspmu/arm_cspmu.c
296
const cpumask_t *cpumask;
drivers/perf/arm_cspmu/arm_cspmu.c
300
cpumask = &cspmu->active_cpu;
drivers/perf/arm_cspmu/arm_cspmu.c
303
cpumask = &cspmu->associated_cpus;
drivers/perf/arm_cspmu/arm_cspmu.c
308
return cpumap_print_to_pagebuf(true, buf, cpumask);
drivers/perf/arm_cspmu/arm_cspmu.c
312
ARM_CSPMU_CPUMASK_ATTR(cpumask, ARM_CSPMU_ACTIVE_CPU_MASK),
drivers/perf/arm_dmc620_pmu.c
245
__ATTR(cpumask, 0444, dmc620_pmu_cpumask_show, NULL);
drivers/perf/arm_dsu_pmu.c
148
const cpumask_t *cpumask;
drivers/perf/arm_dsu_pmu.c
152
cpumask = &dsu_pmu->active_cpu;
drivers/perf/arm_dsu_pmu.c
155
cpumask = &dsu_pmu->associated_cpus;
drivers/perf/arm_dsu_pmu.c
160
return cpumap_print_to_pagebuf(true, buf, cpumask);
drivers/perf/arm_dsu_pmu.c
205
DSU_CPUMASK_ATTR(cpumask, DSU_ACTIVE_CPU_MASK),
drivers/perf/arm_pmu.c
29
static int armpmu_count_irq_users(const struct cpumask *affinity,
drivers/perf/arm_pmu.c
587
static int armpmu_count_irq_users(const struct cpumask *affinity, const int irq)
drivers/perf/arm_pmu.c
600
armpmu_find_irq_ops(const struct cpumask *affinity, int irq)
drivers/perf/arm_pmu.c
634
const struct cpumask *affinity = *armpmu ? &(*armpmu)->supported_cpus :
drivers/perf/arm_pmu_platform.c
117
const struct cpumask *affinity;
drivers/perf/arm_pmu_platform.c
46
const struct cpumask *affinity)
drivers/perf/arm_smmuv3_pmu.c
544
__ATTR(cpumask, 0444, smmu_pmu_cpumask_show, NULL);
drivers/perf/arm_spe_pmu.c
1333
const struct cpumask *affinity;
drivers/perf/arm_spe_pmu.c
348
static DEVICE_ATTR_RO(cpumask);
drivers/perf/cxl_pmu.c
498
static DEVICE_ATTR_RO(cpumask);
drivers/perf/dwc_pcie_pmu.c
122
static DEVICE_ATTR_RO(cpumask);
drivers/perf/fsl_imx8_ddr_perf.c
244
__ATTR(cpumask, 0444, ddr_perf_cpumask_show, NULL);
drivers/perf/fsl_imx9_ddr_perf.c
166
__ATTR(cpumask, 0444, ddr_perf_cpumask_show, NULL);
drivers/perf/fujitsu_uncore_pmu.c
379
static DEVICE_ATTR_RO(cpumask);
drivers/perf/hisilicon/hisi_pcie_pmu.c
126
static DEVICE_ATTR_RO(cpumask);
drivers/perf/hisilicon/hisi_uncore_pmu.c
52
static DEVICE_ATTR(cpumask, 0444, hisi_cpumask_sysfs_show, NULL);
drivers/perf/hisilicon/hns3_pmu.c
474
static DEVICE_ATTR_RO(cpumask);
drivers/perf/marvell_cn10k_ddr_pmu.c
371
__ATTR(cpumask, 0444, cn10k_ddr_perf_cpumask_show, NULL);
drivers/perf/marvell_cn10k_tad_pmu.c
264
static DEVICE_ATTR(cpumask, 0444, tad_pmu_cpumask_show, NULL);
drivers/perf/marvell_pem_pmu.c
171
__ATTR(cpumask, 0444, pem_perf_cpumask_show, NULL);
drivers/perf/qcom_l2_pmu.c
114
cpumask_t cpumask;
drivers/perf/qcom_l2_pmu.c
641
return cpumap_print_to_pagebuf(true, buf, &l2cache_pmu->cpumask);
drivers/perf/qcom_l2_pmu.c
645
__ATTR(cpumask, S_IRUGO, l2_cache_pmu_cpumask_show, NULL);
drivers/perf/qcom_l2_pmu.c
793
cpumask_set_cpu(cpu, &l2cache_pmu->cpumask);
drivers/perf/qcom_l2_pmu.c
818
cpumask_clear_cpu(cpu, &l2cache_pmu->cpumask);
drivers/perf/qcom_l2_pmu.c
831
cpumask_set_cpu(target, &l2cache_pmu->cpumask);
drivers/perf/qcom_l2_pmu.c
930
cpumask_clear(&l2cache_pmu->cpumask);
drivers/perf/qcom_l3_pmu.c
160
cpumask_t cpumask;
drivers/perf/qcom_l3_pmu.c
519
event->cpu = cpumask_first(&l3pmu->cpumask);
drivers/perf/qcom_l3_pmu.c
666
return cpumap_print_to_pagebuf(true, buf, &l3pmu->cpumask);
drivers/perf/qcom_l3_pmu.c
669
static DEVICE_ATTR_RO(cpumask);
drivers/perf/qcom_l3_pmu.c
699
if (cpumask_empty(&l3pmu->cpumask))
drivers/perf/qcom_l3_pmu.c
700
cpumask_set_cpu(cpu, &l3pmu->cpumask);
drivers/perf/qcom_l3_pmu.c
710
if (!cpumask_test_and_clear_cpu(cpu, &l3pmu->cpumask))
drivers/perf/qcom_l3_pmu.c
716
cpumask_set_cpu(target, &l3pmu->cpumask);
drivers/perf/starfive_starlink_pmu.c
134
return cpumap_print_to_pagebuf(true, buf, &starlink_pmu->cpumask);
drivers/perf/starfive_starlink_pmu.c
137
static DEVICE_ATTR_RO(cpumask);
drivers/perf/starfive_starlink_pmu.c
392
event->cpu = cpumask_first(&starlink_pmu->cpumask);
drivers/perf/starfive_starlink_pmu.c
594
if (cpumask_empty(&starlink_pmu->cpumask))
drivers/perf/starfive_starlink_pmu.c
595
cpumask_set_cpu(cpu, &starlink_pmu->cpumask);
drivers/perf/starfive_starlink_pmu.c
610
if (!cpumask_test_and_clear_cpu(cpu, &starlink_pmu->cpumask))
drivers/perf/starfive_starlink_pmu.c
619
cpumask_set_cpu(target, &starlink_pmu->cpumask);
drivers/perf/starfive_starlink_pmu.c
75
cpumask_t cpumask;
drivers/perf/thunderx2_pmu.c
259
static DEVICE_ATTR_RO(cpumask);
drivers/perf/xgene_pmu.c
601
static DEVICE_ATTR_RO(cpumask);
drivers/pinctrl/pinctrl-rp1.c
993
static int rp1_gpio_irq_set_affinity(struct irq_data *data, const struct cpumask *dest, bool force)
drivers/pinctrl/qcom/pinctrl-msm.c
1281
const struct cpumask *dest, bool force)
drivers/pinctrl/samsung/pinctrl-exynos.c
213
const struct cpumask *dest, bool force)
drivers/platform/x86/intel/ifs/runtest.c
165
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1253
static struct isst_perf_level_cpu_mask cpumask;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1257
if (copy_from_user(&cpumask, argp, sizeof(cpumask)))
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1260
power_domain_info = get_instance(cpumask.socket_id, cpumask.power_domain_id);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1264
_read_pp_level_info("mask", mask, cpumask.level, SST_PP_INFO_2_OFFSET,
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1268
cpumask.mask = mask;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1270
if (!cpumask.punit_cpu_map)
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1273
if (copy_to_user(argp, &cpumask, sizeof(cpumask)))
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1334
static struct isst_perf_level_cpu_mask cpumask;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1338
if (copy_from_user(&cpumask, argp, sizeof(cpumask)))
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1341
power_domain_info = get_instance(cpumask.socket_id, cpumask.power_domain_id);
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1345
_read_bf_level_info("BF-cpumask", mask, cpumask.level, SST_BF_INFO_1_OFFSET,
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1349
cpumask.mask = mask;
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1351
if (!cpumask.punit_cpu_map)
drivers/platform/x86/intel/speed_select_if/isst_tpmi_core.c
1354
if (copy_to_user(argp, &cpumask, sizeof(cpumask)))
drivers/powercap/dtpm_cpu.c
71
static u64 scale_pd_power_uw(struct cpumask *pd_mask, u64 power)
drivers/powercap/dtpm_cpu.c
93
struct cpumask *pd_mask;
drivers/powercap/idle_inject.c
107
for_each_cpu_and(cpu, to_cpumask(ii_dev->cpumask), cpu_online_mask) {
drivers/powercap/idle_inject.c
230
cpumask_pr_args(to_cpumask(ii_dev->cpumask)));
drivers/powercap/idle_inject.c
259
cpumask_pr_args(to_cpumask(ii_dev->cpumask)));
drivers/powercap/idle_inject.c
278
for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) {
drivers/powercap/idle_inject.c
330
struct idle_inject_device *idle_inject_register_full(struct cpumask *cpumask,
drivers/powercap/idle_inject.c
340
cpumask_copy(to_cpumask(ii_dev->cpumask), cpumask);
drivers/powercap/idle_inject.c
345
for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) {
drivers/powercap/idle_inject.c
358
for_each_cpu(cpu_rb, to_cpumask(ii_dev->cpumask)) {
drivers/powercap/idle_inject.c
381
struct idle_inject_device *idle_inject_register(struct cpumask *cpumask)
drivers/powercap/idle_inject.c
383
return idle_inject_register_full(cpumask, NULL);
drivers/powercap/idle_inject.c
401
for_each_cpu(cpu, to_cpumask(ii_dev->cpumask))
drivers/powercap/idle_inject.c
89
unsigned long cpumask[];
drivers/powercap/intel_rapl_common.c
1905
static DEVICE_ATTR_RO(cpumask);
drivers/powercap/intel_rapl_msr.c
100
cpumask_clear_cpu(cpu, &rp->cpumask);
drivers/powercap/intel_rapl_msr.c
101
lead_cpu = cpumask_first(&rp->cpumask);
drivers/powercap/intel_rapl_msr.c
87
cpumask_set_cpu(cpu, &rp->cpumask);
drivers/scsi/elx/efct/efct_hw_queues.c
108
const struct cpumask *maskp;
drivers/scsi/hisi_sas/hisi_sas.h
216
const struct cpumask *irq_mask;
drivers/scsi/hisi_sas/hisi_sas_main.c
1203
const struct cpumask *mask = cq->irq_mask;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3558
const struct cpumask *mask;
drivers/scsi/hpsa.c
7473
const struct cpumask *mask;
drivers/scsi/lpfc/lpfc.h
1674
lpfc_next_online_cpu(const struct cpumask *mask, unsigned int start)
drivers/scsi/lpfc/lpfc_attr.c
5637
const struct cpumask *sibling_mask;
drivers/scsi/lpfc/lpfc_attr.c
5638
struct cpumask *aff_mask = &phba->sli4_hba.irq_aff_mask;
drivers/scsi/lpfc/lpfc_attr.c
5727
const struct cpumask *aff_mask;
drivers/scsi/lpfc/lpfc_init.c
12689
const struct cpumask *maskp;
drivers/scsi/lpfc/lpfc_init.c
12691
struct cpumask *tmp;
drivers/scsi/lpfc/lpfc_init.c
12835
struct cpumask *aff_mask;
drivers/scsi/lpfc/lpfc_init.c
12837
const struct cpumask *orig_mask;
drivers/scsi/lpfc/lpfc_init.c
12970
const struct cpumask *aff_mask = NULL;
drivers/scsi/lpfc/lpfc_init.c
12974
const struct cpumask *maskp;
drivers/scsi/lpfc/lpfc_sli4.h
495
struct cpumask aff_mask;
drivers/scsi/lpfc/lpfc_sli4.h
959
struct cpumask irq_aff_mask;
drivers/scsi/megaraid/megaraid_sas_base.c
5840
const struct cpumask *mask;
drivers/scsi/megaraid/megaraid_sas_base.c
5908
const struct cpumask *mask;
drivers/scsi/mpt3sas/mpt3sas_base.c
3218
const struct cpumask *mask;
drivers/scsi/qla2xxx/qla_inline.h
594
const struct cpumask *mask;
drivers/scsi/storvsc_drv.c
1401
const struct cpumask *node_mask;
drivers/scsi/storvsc_drv.c
1461
const struct cpumask *node_mask;
drivers/scsi/storvsc_drv.c
441
struct cpumask alloced_cpus;
drivers/sh/intc/chip.c
69
const struct cpumask *cpumask,
drivers/sh/intc/chip.c
72
if (!cpumask_intersects(cpumask, cpu_online_mask))
drivers/sh/intc/chip.c
75
irq_data_update_affinity(data, cpumask);
drivers/soc/fsl/qbman/bman_portal.c
34
static struct cpumask portal_cpus;
drivers/soc/fsl/qbman/qman_portal.c
40
static struct cpumask portal_cpus;
drivers/soc/tegra/pmc.c
2706
const struct cpumask *dest,
drivers/soc/ti/knav_qmss.h
325
struct cpumask *cpu_mask;
drivers/soc/ti/knav_qmss_acc.c
200
struct cpumask *cpu_mask;
drivers/thermal/cpuidle_cooling.c
188
ii_dev = idle_inject_register(drv->cpumask);
drivers/thermal/cpuidle_cooling.c
202
dev = get_cpu_device(cpumask_first(drv->cpumask));
drivers/thermal/cpuidle_cooling.c
247
for_each_cpu(cpu, drv->cpumask) {
drivers/thermal/intel/intel_powerclamp.c
130
static int allocate_copy_idle_injection_mask(const struct cpumask *copy_mask)
drivers/thermal/intel/intel_powerclamp.c
212
module_param_cb(cpumask, &cpumask_ops, NULL, 0644);
drivers/thermal/intel/intel_powerclamp.c
213
MODULE_PARM_DESC(cpumask, "Mask of CPUs to use for idle injection.");
drivers/thermal/intel/x86_pkg_temp_thermal.c
363
cpumask_set_cpu(cpu, &zonedev->cpumask);
drivers/thermal/intel/x86_pkg_temp_thermal.c
386
target = cpumask_any_but(&zonedev->cpumask, cpu);
drivers/thermal/intel/x86_pkg_temp_thermal.c
387
cpumask_clear_cpu(cpu, &zonedev->cpumask);
drivers/thermal/intel/x86_pkg_temp_thermal.c
475
cpumask_set_cpu(cpu, &zonedev->cpumask);
drivers/thermal/intel/x86_pkg_temp_thermal.c
57
struct cpumask cpumask;
drivers/thermal/thermal_trace.h
116
TP_PROTO(const struct cpumask *cpus, unsigned int freq,
drivers/thermal/thermal_trace.h
122
__bitmask(cpumask, num_possible_cpus())
drivers/thermal/thermal_trace.h
129
__assign_bitmask(cpumask, cpumask_bits(cpus),
drivers/thermal/thermal_trace.h
137
__get_bitmask(cpumask), __entry->freq, __entry->cdev_state,
drivers/vdpa/vdpa_user/vduse_dev.c
82
struct cpumask irq_affinity;
drivers/vdpa/vdpa_user/vduse_dev.c
832
const struct cpumask *cpu_mask)
drivers/vdpa/vdpa_user/vduse_dev.c
844
static const struct cpumask *
drivers/virtio/virtio.c
393
static const struct cpumask *virtio_irq_get_affinity(struct device *_d,
drivers/virtio/virtio_pci_common.c
557
int vp_set_vq_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask)
drivers/virtio/virtio_pci_common.c
562
struct cpumask *mask;
drivers/virtio/virtio_pci_common.c
581
const struct cpumask *vp_get_vq_affinity(struct virtio_device *vdev, int index)
drivers/virtio/virtio_pci_common.h
145
int vp_set_vq_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask);
drivers/virtio/virtio_pci_common.h
147
const struct cpumask *vp_get_vq_affinity(struct virtio_device *vdev, int index);
drivers/virtio/virtio_vdpa.c
273
static struct cpumask *
drivers/virtio/virtio_vdpa.c
277
struct cpumask *masks = NULL;
drivers/virtio/virtio_vdpa.c
302
struct cpumask *result = group_cpus_evenly(this_vecs, &nr_masks);
drivers/virtio/virtio_vdpa.c
336
struct cpumask *masks;
drivers/virtio/virtio_vdpa.c
408
const struct cpumask *cpu_mask)
drivers/virtio/virtio_vdpa.c
421
static const struct cpumask *
drivers/xen/events/events_base.c
1824
static unsigned int select_target_cpu(const struct cpumask *dest)
drivers/xen/events/events_base.c
1847
static int set_affinity_irq(struct irq_data *data, const struct cpumask *dest,
fs/fuse/virtio_fs.c
861
const struct cpumask *mask, *masks;
fs/resctrl/ctrlmondata.c
556
cpumask_t *cpumask, struct mon_evt *evt, int first)
fs/resctrl/ctrlmondata.c
587
cpu = cpumask_any_housekeeping(cpumask, RESCTRL_PICK_ANY_CPU);
fs/resctrl/ctrlmondata.c
596
smp_call_function_any(cpumask, mon_event_count, rr, 1);
fs/resctrl/internal.h
225
struct cpumask cpu_mask;
fs/resctrl/internal.h
25
cpumask_any_housekeeping(const struct cpumask *mask, int exclude_cpu)
fs/resctrl/internal.h
390
cpumask_t *cpumask, struct mon_evt *evt, int first);
fs/resctrl/rdtgroup.c
2991
struct cpumask *mask)
fs/resctrl/rdtgroup.c
354
struct cpumask *mask;
fs/resctrl/rdtgroup.c
390
update_closid_rmid(const struct cpumask *cpu_mask, struct rdtgroup *r)
fs/resctrl/rdtgroup.c
446
static void cpumask_rdtgrp_clear(struct rdtgroup *r, struct cpumask *m)
fs/xfs/xfs_log_priv.h
255
struct cpumask cil_pcpmask;
fs/xfs/xfs_mount.h
342
struct cpumask m_inodegc_cpumask;
include/asm-generic/mshyperv.h
232
const struct cpumask *cpus,
include/asm-generic/mshyperv.h
278
const struct cpumask *cpus)
include/asm-generic/mshyperv.h
284
const struct cpumask *cpus,
include/asm-generic/numa.h
21
const struct cpumask *cpumask_of_node(int node);
include/asm-generic/numa.h
24
static inline const struct cpumask *cpumask_of_node(int node)
include/linux/acpi.h
1509
const struct cpumask *acpi_irq_get_affinity(acpi_handle handle,
include/linux/acpi.h
1517
static inline const struct cpumask *acpi_irq_get_affinity(acpi_handle handle,
include/linux/arch_topology.h
32
void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq,
include/linux/arch_topology.h
49
void topology_set_scale_freq_source(struct scale_freq_data *data, const struct cpumask *cpus);
include/linux/arch_topology.h
50
void topology_clear_scale_freq_source(enum scale_freq_source source, const struct cpumask *cpus);
include/linux/arch_topology.h
59
void topology_update_hw_pressure(const struct cpumask *cpus,
include/linux/arch_topology.h
90
const struct cpumask *cpu_coregroup_mask(int cpu);
include/linux/arch_topology.h
91
const struct cpumask *cpu_clustergroup_mask(int cpu);
include/linux/blk-mq.h
346
cpumask_var_t cpumask;
include/linux/clockchips.h
119
void (*broadcast)(const struct cpumask *mask);
include/linux/clockchips.h
129
const struct cpumask *cpumask;
include/linux/clockchips.h
203
extern void tick_broadcast(const struct cpumask *mask);
include/linux/cpu_rmap.h
40
const struct cpumask *affinity);
include/linux/cpufreq.h
1167
const char *cell_name, struct cpumask *cpumask,
include/linux/cpufreq.h
1177
cpumask_set_cpu(pcpu, cpumask);
include/linux/cpufreq.h
1188
cpumask_set_cpu(cpu, cpumask);
include/linux/cpufreq.h
1214
const char *cell_name, struct cpumask *cpumask,
include/linux/cpufreq.h
1225
void arch_set_freq_scale(const struct cpumask *cpus,
include/linux/cpufreq.h
1242
bool cpufreq_ready_for_eas(const struct cpumask *cpu_mask);
include/linux/cpufreq.h
798
ssize_t cpufreq_show_cpus(const struct cpumask *mask, char *buf);
include/linux/cpuidle.h
164
struct cpumask *cpumask;
include/linux/cpuidle.h
192
const struct cpumask *const coupled_cpus);
include/linux/cpuidle.h
230
const struct cpumask *const coupled_cpus)
include/linux/cpumask.h
1008
static __always_inline int cpulist_parse(const char *buf, struct cpumask *dstp)
include/linux/cpumask.h
1118
DEFINE_FREE(free_cpumask_var, struct cpumask *, if (_T) free_cpumask_var(_T));
include/linux/cpumask.h
1151
void init_cpu_present(const struct cpumask *src);
include/linux/cpumask.h
1152
void init_cpu_possible(const struct cpumask *src);
include/linux/cpumask.h
117
extern struct cpumask __cpu_possible_mask;
include/linux/cpumask.h
1179
((struct cpumask *)(1 ? (bitmap) \
include/linux/cpumask.h
118
extern struct cpumask __cpu_online_mask;
include/linux/cpumask.h
119
extern struct cpumask __cpu_enabled_mask;
include/linux/cpumask.h
1197
static __always_inline const struct cpumask *get_cpu_mask(unsigned int cpu)
include/linux/cpumask.h
120
extern struct cpumask __cpu_present_mask;
include/linux/cpumask.h
121
extern struct cpumask __cpu_active_mask;
include/linux/cpumask.h
122
extern struct cpumask __cpu_dying_mask;
include/linux/cpumask.h
123
#define cpu_possible_mask ((const struct cpumask *)&__cpu_possible_mask)
include/linux/cpumask.h
124
#define cpu_online_mask ((const struct cpumask *)&__cpu_online_mask)
include/linux/cpumask.h
125
#define cpu_enabled_mask ((const struct cpumask *)&__cpu_enabled_mask)
include/linux/cpumask.h
126
#define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask)
include/linux/cpumask.h
127
#define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask)
include/linux/cpumask.h
128
#define cpu_dying_mask ((const struct cpumask *)&__cpu_dying_mask)
include/linux/cpumask.h
1327
cpumap_print_to_pagebuf(bool list, char *buf, const struct cpumask *mask)
include/linux/cpumask.h
1350
ssize_t cpumap_print_bitmask_to_buf(char *buf, const struct cpumask *mask,
include/linux/cpumask.h
1372
ssize_t cpumap_print_list_to_buf(char *buf, const struct cpumask *mask,
include/linux/cpumask.h
155
static __always_inline unsigned int cpumask_first(const struct cpumask *srcp)
include/linux/cpumask.h
166
static __always_inline unsigned int cpumask_first_zero(const struct cpumask *srcp)
include/linux/cpumask.h
179
unsigned int cpumask_first_and(const struct cpumask *srcp1, const struct cpumask *srcp2)
include/linux/cpumask.h
192
unsigned int cpumask_first_andnot(const struct cpumask *srcp1, const struct cpumask *srcp2)
include/linux/cpumask.h
206
unsigned int cpumask_first_and_and(const struct cpumask *srcp1,
include/linux/cpumask.h
207
const struct cpumask *srcp2,
include/linux/cpumask.h
208
const struct cpumask *srcp3)
include/linux/cpumask.h
220
static __always_inline unsigned int cpumask_last(const struct cpumask *srcp)
include/linux/cpumask.h
233
unsigned int cpumask_next(int n, const struct cpumask *srcp)
include/linux/cpumask.h
249
unsigned int cpumask_next_zero(int n, const struct cpumask *srcp)
include/linux/cpumask.h
266
unsigned int cpumask_any_and_distribute(const struct cpumask *src1p,
include/linux/cpumask.h
267
const struct cpumask *src2p)
include/linux/cpumask.h
273
unsigned int cpumask_any_distribute(const struct cpumask *srcp)
include/linux/cpumask.h
279
unsigned int cpumask_any_and_distribute(const struct cpumask *src1p,
include/linux/cpumask.h
280
const struct cpumask *src2p);
include/linux/cpumask.h
281
unsigned int cpumask_any_distribute(const struct cpumask *srcp);
include/linux/cpumask.h
293
unsigned int cpumask_next_and(int n, const struct cpumask *src1p,
include/linux/cpumask.h
294
const struct cpumask *src2p)
include/linux/cpumask.h
312
unsigned int cpumask_next_andnot(int n, const struct cpumask *src1p,
include/linux/cpumask.h
313
const struct cpumask *src2p)
include/linux/cpumask.h
333
unsigned int cpumask_next_and_wrap(int n, const struct cpumask *src1p,
include/linux/cpumask.h
334
const struct cpumask *src2p)
include/linux/cpumask.h
352
unsigned int cpumask_next_wrap(int n, const struct cpumask *src)
include/linux/cpumask.h
367
unsigned int cpumask_random(const struct cpumask *src)
include/linux/cpumask.h
467
unsigned int cpumask_any_but(const struct cpumask *mask, int cpu)
include/linux/cpumask.h
491
unsigned int cpumask_any_and_but(const struct cpumask *mask1,
include/linux/cpumask.h
492
const struct cpumask *mask2,
include/linux/cpumask.h
518
unsigned int cpumask_any_andnot_but(const struct cpumask *mask1,
include/linux/cpumask.h
519
const struct cpumask *mask2,
include/linux/cpumask.h
543
unsigned int cpumask_nth(unsigned int cpu, const struct cpumask *srcp)
include/linux/cpumask.h
557
unsigned int cpumask_nth_and(unsigned int cpu, const struct cpumask *srcp1,
include/linux/cpumask.h
558
const struct cpumask *srcp2)
include/linux/cpumask.h
574
unsigned int cpumask_nth_and_andnot(unsigned int cpu, const struct cpumask *srcp1,
include/linux/cpumask.h
575
const struct cpumask *srcp2,
include/linux/cpumask.h
576
const struct cpumask *srcp3)
include/linux/cpumask.h
600
void cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp)
include/linux/cpumask.h
606
void __cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp)
include/linux/cpumask.h
617
static __always_inline void cpumask_clear_cpus(struct cpumask *dstp,
include/linux/cpumask.h
629
static __always_inline void cpumask_clear_cpu(int cpu, struct cpumask *dstp)
include/linux/cpumask.h
634
static __always_inline void __cpumask_clear_cpu(int cpu, struct cpumask *dstp)
include/linux/cpumask.h
647
bool cpumask_test_cpu(int cpu, const struct cpumask *cpumask)
include/linux/cpumask.h
649
return test_bit(cpumask_check(cpu), cpumask_bits((cpumask)));
include/linux/cpumask.h
662
bool cpumask_test_and_set_cpu(int cpu, struct cpumask *cpumask)
include/linux/cpumask.h
664
return test_and_set_bit(cpumask_check(cpu), cpumask_bits(cpumask));
include/linux/cpumask.h
677
bool cpumask_test_and_clear_cpu(int cpu, struct cpumask *cpumask)
include/linux/cpumask.h
679
return test_and_clear_bit(cpumask_check(cpu), cpumask_bits(cpumask));
include/linux/cpumask.h
686
static __always_inline void cpumask_setall(struct cpumask *dstp)
include/linux/cpumask.h
699
static __always_inline void cpumask_clear(struct cpumask *dstp)
include/linux/cpumask.h
713
bool cpumask_and(struct cpumask *dstp, const struct cpumask *src1p,
include/linux/cpumask.h
714
const struct cpumask *src2p)
include/linux/cpumask.h
727
void cpumask_or(struct cpumask *dstp, const struct cpumask *src1p,
include/linux/cpumask.h
728
const struct cpumask *src2p)
include/linux/cpumask.h
743
unsigned int cpumask_weighted_or(struct cpumask *dstp, const struct cpumask *src1p,
include/linux/cpumask.h
744
const struct cpumask *src2p)
include/linux/cpumask.h
757
void cpumask_xor(struct cpumask *dstp, const struct cpumask *src1p,
include/linux/cpumask.h
758
const struct cpumask *src2p)
include/linux/cpumask.h
773
bool cpumask_andnot(struct cpumask *dstp, const struct cpumask *src1p,
include/linux/cpumask.h
774
const struct cpumask *src2p)
include/linux/cpumask.h
788
bool cpumask_equal(const struct cpumask *src1p, const struct cpumask *src2p)
include/linux/cpumask.h
804
bool cpumask_or_equal(const struct cpumask *src1p, const struct cpumask *src2p,
include/linux/cpumask.h
805
const struct cpumask *src3p)
include/linux/cpumask.h
820
bool cpumask_intersects(const struct cpumask *src1p, const struct cpumask *src2p)
include/linux/cpumask.h
834
bool cpumask_subset(const struct cpumask *src1p, const struct cpumask *src2p)
include/linux/cpumask.h
846
static __always_inline bool cpumask_empty(const struct cpumask *srcp)
include/linux/cpumask.h
857
static __always_inline bool cpumask_full(const struct cpumask *srcp)
include/linux/cpumask.h
868
static __always_inline unsigned int cpumask_weight(const struct cpumask *srcp)
include/linux/cpumask.h
881
unsigned int cpumask_weight_and(const struct cpumask *srcp1, const struct cpumask *srcp2)
include/linux/cpumask.h
894
unsigned int cpumask_weight_andnot(const struct cpumask *srcp1,
include/linux/cpumask.h
895
const struct cpumask *srcp2)
include/linux/cpumask.h
907
void cpumask_shift_right(struct cpumask *dstp, const struct cpumask *srcp, int n)
include/linux/cpumask.h
920
void cpumask_shift_left(struct cpumask *dstp, const struct cpumask *srcp, int n)
include/linux/cpumask.h
932
void cpumask_copy(struct cpumask *dstp, const struct cpumask *srcp)
include/linux/cpumask.h
969
int cpumask_parse_user(const char __user *buf, int len, struct cpumask *dstp)
include/linux/cpumask.h
983
int cpumask_parselist_user(const char __user *buf, int len, struct cpumask *dstp)
include/linux/cpumask.h
996
static __always_inline int cpumask_parse(const char *buf, struct cpumask *dstp)
include/linux/cpumask_types.h
61
typedef struct cpumask *cpumask_var_t;
include/linux/cpumask_types.h
63
typedef struct cpumask cpumask_var_t[1];
include/linux/cpuset.h
203
struct cpumask *mask)
include/linux/cpuset.h
209
struct cpumask *mask)
include/linux/cpuset.h
80
extern void cpuset_cpus_allowed_locked(struct task_struct *p, struct cpumask *mask);
include/linux/cpuset.h
81
extern void cpuset_cpus_allowed(struct task_struct *p, struct cpumask *mask);
include/linux/device/bus.h
96
const struct cpumask *(*irq_get_affinity)(struct device *dev,
include/linux/group_cpus.h
12
struct cpumask *group_cpus_evenly(unsigned int numgrps, unsigned int *nummasks);
include/linux/idle_inject.h
14
struct idle_inject_device *idle_inject_register(struct cpumask *cpumask);
include/linux/idle_inject.h
16
struct idle_inject_device *idle_inject_register_full(struct cpumask *cpumask,
include/linux/intel_rapl.h
196
struct cpumask cpumask;
include/linux/interrupt.h
129
const struct cpumask *affinity;
include/linux/interrupt.h
201
const struct cpumask *affinity, void __percpu *dev_id);
include/linux/interrupt.h
310
struct cpumask mask;
include/linux/interrupt.h
318
extern int irq_set_affinity(unsigned int irq, const struct cpumask *cpumask);
include/linux/interrupt.h
319
extern int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask);
include/linux/interrupt.h
324
extern int __irq_apply_affinity_hint(unsigned int irq, const struct cpumask *m,
include/linux/interrupt.h
335
irq_update_affinity_hint(unsigned int irq, const struct cpumask *m)
include/linux/interrupt.h
350
irq_set_affinity_and_hint(unsigned int irq, const struct cpumask *m)
include/linux/interrupt.h
359
static inline int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m)
include/linux/interrupt.h
378
static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m)
include/linux/interrupt.h
383
static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask)
include/linux/interrupt.h
396
const struct cpumask *m)
include/linux/interrupt.h
402
const struct cpumask *m)
include/linux/interrupt.h
408
const struct cpumask *m)
include/linux/irq.h
1249
int irq_matrix_reserve_managed(struct irq_matrix *m, const struct cpumask *msk);
include/linux/irq.h
1250
void irq_matrix_remove_managed(struct irq_matrix *m, const struct cpumask *msk);
include/linux/irq.h
1251
int irq_matrix_alloc_managed(struct irq_matrix *m, const struct cpumask *msk,
include/linux/irq.h
1255
int irq_matrix_alloc(struct irq_matrix *m, const struct cpumask *msk,
include/linux/irq.h
1269
int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest);
include/linux/irq.h
1271
int ipi_send_mask(unsigned int virq, const struct cpumask *dest);
include/linux/irq.h
507
int (*irq_set_affinity)(struct irq_data *data, const struct cpumask *dest, bool force);
include/linux/irq.h
539
void (*ipi_send_mask)(struct irq_data *data, const struct cpumask *dest);
include/linux/irq.h
606
const struct cpumask *cpumask, bool force);
include/linux/irq.h
682
const struct cpumask *dest,
include/linux/irq.h
696
int irq_chip_redirect_set_affinity(struct irq_data *data, const struct cpumask *dest, bool force);
include/linux/irq.h
881
const struct cpumask *irq_data_get_affinity_mask(struct irq_data *d)
include/linux/irq.h
891
const struct cpumask *m)
include/linux/irq.h
898
static inline const struct cpumask *irq_get_affinity_mask(int irq)
include/linux/irq.h
907
const struct cpumask *irq_data_get_effective_affinity_mask(struct irq_data *d)
include/linux/irq.h
912
const struct cpumask *m)
include/linux/irq.h
918
const struct cpumask *m)
include/linux/irq.h
922
const struct cpumask *irq_data_get_effective_affinity_mask(struct irq_data *d)
include/linux/irq.h
929
const struct cpumask *irq_get_effective_affinity_mask(unsigned int irq)
include/linux/irqdesc.h
100
const struct cpumask *affinity_hint;
include/linux/irqdesc.h
97
struct cpumask *percpu_enabled;
include/linux/irqdomain.h
24
struct cpumask;
include/linux/irqdomain.h
562
int irq_reserve_ipi(struct irq_domain *domain, const struct cpumask *dest);
include/linux/irqdomain.h
563
int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest);
include/linux/irqdomain.h
59
const struct cpumask *affinity;
include/linux/kthread.h
105
void kthread_bind_mask(struct task_struct *k, const struct cpumask *mask);
include/linux/kthread.h
106
int kthread_affine_preferred(struct task_struct *p, const struct cpumask *mask);
include/linux/mm_types.h
1428
cpumask_clear((struct cpumask *)cpu_bitmap);
include/linux/mm_types.h
1434
return (struct cpumask *)&mm->flexible_array;
include/linux/mm_types.h
1528
return (struct cpumask *)bitmap;
include/linux/msi.h
633
int msi_domain_set_affinity(struct irq_data *data, const struct cpumask *mask,
include/linux/nd.h
56
struct cpumask arch_cpumask;
include/linux/netdevice.h
4072
int netif_set_xps_queue(struct net_device *dev, const struct cpumask *mask,
include/linux/netdevice.h
4164
const struct cpumask *mask,
include/linux/nmi.h
171
static inline bool trigger_cpumask_backtrace(struct cpumask *mask)
include/linux/nmi.h
198
static inline bool trigger_cpumask_backtrace(struct cpumask *mask)
include/linux/nmi.h
225
void nmi_backtrace_stall_snap(const struct cpumask *btp);
include/linux/nmi.h
226
void nmi_backtrace_stall_check(const struct cpumask *btp);
include/linux/nmi.h
228
static inline void nmi_backtrace_stall_snap(const struct cpumask *btp) {}
include/linux/nmi.h
229
static inline void nmi_backtrace_stall_check(const struct cpumask *btp) {}
include/linux/nmi.h
25
extern struct cpumask watchdog_cpumask;
include/linux/of_irq.h
110
static inline const struct cpumask *of_irq_get_affinity(struct device_node *dev,
include/linux/of_irq.h
70
extern const struct cpumask *of_irq_get_affinity(struct device_node *dev,
include/linux/padata.h
102
struct padata_cpumask cpumask;
include/linux/padata.h
168
struct padata_cpumask cpumask;
include/linux/padata.h
188
cpumask_var_t cpumask);
include/linux/pci.h
1774
const struct cpumask *pci_irq_get_affinity(struct pci_dev *pdev, int vec);
include/linux/pci.h
1833
static inline const struct cpumask *pci_irq_get_affinity(struct pci_dev *pdev,
include/linux/platform_device.h
101
const struct cpumask **);
include/linux/pm_opp.h
218
int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask);
include/linux/pm_opp.h
219
int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask);
include/linux/pm_opp.h
221
void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask);
include/linux/pm_opp.h
472
static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask)
include/linux/pm_opp.h
477
static inline int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask)
include/linux/pm_opp.h
486
static inline void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask)
include/linux/pm_opp.h
518
int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask);
include/linux/pm_opp.h
519
void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask);
include/linux/pm_opp.h
520
int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask);
include/linux/pm_opp.h
526
int dev_pm_opp_of_register_em(struct device *dev, struct cpumask *cpus);
include/linux/pm_opp.h
558
static inline int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask)
include/linux/pm_opp.h
563
static inline void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask)
include/linux/pm_opp.h
567
static inline int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask)
include/linux/pm_opp.h
583
struct cpumask *cpus)
include/linux/resctrl.h
143
struct cpumask cpu_mask;
include/linux/sched.h
1884
extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
include/linux/sched.h
1890
extern void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask);
include/linux/sched.h
1899
extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask);
include/linux/sched.h
1902
extern int dl_task_check_affinity(struct task_struct *p, const struct cpumask *mask);
include/linux/sched.h
2301
extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
include/linux/sched.h
2302
extern long sched_getaffinity(pid_t pid, struct cpumask *mask);
include/linux/sched/isolation.h
38
extern const struct cpumask *housekeeping_cpumask(enum hk_type type);
include/linux/sched/isolation.h
42
extern int housekeeping_update(struct cpumask *isol_mask);
include/linux/sched/isolation.h
52
static inline const struct cpumask *housekeeping_cpumask(enum hk_type type)
include/linux/sched/isolation.h
70
static inline int housekeeping_update(struct cpumask *isol_mask) { return 0; }
include/linux/sched/topology.h
153
static inline struct cpumask *sched_domain_span(struct sched_domain *sd)
include/linux/sched/topology.h
169
typedef const struct cpumask *(*sched_domain_mask_f)(struct sched_domain_topology_level *tl, int cpu);
include/linux/sched/topology.h
229
void arch_update_hw_pressure(const struct cpumask *cpus,
include/linux/sched/topology.h
37
extern const struct cpumask *tl_smt_mask(struct sched_domain_topology_level *tl, int cpu);
include/linux/sched/topology.h
42
extern const struct cpumask *tl_cls_mask(struct sched_domain_topology_level *tl, int cpu);
include/linux/sched/topology.h
47
extern const struct cpumask *tl_mc_mask(struct sched_domain_topology_level *tl, int cpu);
include/linux/sched/topology.h
50
extern const struct cpumask *tl_pkg_mask(struct sched_domain_topology_level *tl, int cpu);
include/linux/smp.h
164
void smp_call_function_many(const struct cpumask *mask,
include/linux/smp.h
167
int smp_call_function_any(const struct cpumask *mask,
include/linux/smp.h
172
bool cpus_peek_for_pending_ipi(const struct cpumask *mask);
include/linux/smp.h
213
smp_call_function_any(const struct cpumask *mask, smp_call_func_t func,
include/linux/smp.h
221
static inline bool cpus_peek_for_pending_ipi(const struct cpumask *mask)
include/linux/smp.h
54
void *info, bool wait, const struct cpumask *mask);
include/linux/smp.h
91
static inline void on_each_cpu_mask(const struct cpumask *mask,
include/linux/stop_machine.h
130
int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
include/linux/stop_machine.h
142
int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus);
include/linux/stop_machine.h
162
const struct cpumask *cpus);
include/linux/stop_machine.h
166
const struct cpumask *cpus)
include/linux/stop_machine.h
177
stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus)
include/linux/stop_machine.h
184
const struct cpumask *cpus)
include/linux/stop_machine.h
38
void stop_machine_yield(const struct cpumask *cpumask);
include/linux/tick.h
276
extern void __init tick_nohz_full_setup(cpumask_var_t cpumask);
include/linux/tick.h
301
static inline void tick_nohz_full_setup(cpumask_var_t cpumask) { }
include/linux/timer.h
192
extern int tmigr_isolated_exclude_cpumask(struct cpumask *exclude_cpumask);
include/linux/timer.h
194
static inline int tmigr_isolated_exclude_cpumask(struct cpumask *exclude_cpumask)
include/linux/topology.h
234
static inline const struct cpumask *cpu_smt_mask(int cpu)
include/linux/topology.h
263
static inline const struct cpumask *cpu_node_mask(int cpu)
include/linux/topology.h
269
int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node);
include/linux/topology.h
270
extern const struct cpumask *sched_numa_hop_mask(unsigned int node, unsigned int hops);
include/linux/topology.h
272
static __always_inline int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node)
include/linux/topology.h
277
static inline const struct cpumask *
include/linux/torture.h
134
long torture_sched_setaffinity(pid_t pid, const struct cpumask *in_mask, bool dowarn);
include/linux/vdpa.h
430
const struct cpumask *cpu_mask);
include/linux/vdpa.h
431
const struct cpumask *(*get_vq_affinity)(struct vdpa_device *vdev,
include/linux/virtio_config.h
133
const struct cpumask *cpu_mask);
include/linux/virtio_config.h
134
const struct cpumask *(*get_vq_affinity)(struct virtio_device *vdev,
include/linux/virtio_config.h
392
int virtqueue_set_affinity(struct virtqueue *vq, const struct cpumask *cpu_mask)
include/linux/workqueue.h
160
cpumask_var_t cpumask;
include/linux/workqueue.h
591
extern int workqueue_unbound_housekeeping_update(const struct cpumask *hk);
include/net/ip_vs.h
1158
static inline const struct cpumask *sysctl_est_cpulist(struct netns_ipvs *ipvs)
include/net/ip_vs.h
1166
static inline const struct cpumask *sysctl_est_preferred_cpulist(struct netns_ipvs *ipvs)
include/net/ip_vs.h
1276
static inline const struct cpumask *sysctl_est_cpulist(struct netns_ipvs *ipvs)
include/net/ip_vs.h
1281
static inline const struct cpumask *sysctl_est_preferred_cpulist(struct netns_ipvs *ipvs)
include/net/netns/core.h
27
struct cpumask *rps_default_mask;
include/net/netns/core.h
9
struct cpumask;
include/rdma/ib_verbs.h
2422
const struct cpumask *(*get_vector_affinity)(struct ib_device *ibdev,
include/rdma/ib_verbs.h
4887
static inline const struct cpumask *
include/trace/events/ipi.h
34
TP_PROTO(const struct cpumask *cpumask, unsigned long callsite, void *callback),
include/trace/events/ipi.h
36
TP_ARGS(cpumask, callsite, callback),
include/trace/events/ipi.h
39
__cpumask(cpumask)
include/trace/events/ipi.h
45
__assign_cpumask(cpumask, cpumask_bits(cpumask));
include/trace/events/ipi.h
51
__get_cpumask(cpumask), __entry->callsite, __entry->callback)
include/trace/events/ipi.h
66
TP_PROTO(const struct cpumask *mask, const char *reason),
include/trace/events/xen.h
350
TP_PROTO(const struct cpumask *cpus, struct mm_struct *mm,
kernel/bpf/cpumask.c
102
bpf_mem_cache_free_rcu(&bpf_cpumask_ma, cpumask);
kernel/bpf/cpumask.c
105
__bpf_kfunc void bpf_cpumask_release_dtor(void *cpumask)
kernel/bpf/cpumask.c
107
bpf_cpumask_release(cpumask);
kernel/bpf/cpumask.c
121
__bpf_kfunc u32 bpf_cpumask_first(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
123
return cpumask_first(cpumask);
kernel/bpf/cpumask.c
137
__bpf_kfunc u32 bpf_cpumask_first_zero(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
139
return cpumask_first_zero(cpumask);
kernel/bpf/cpumask.c
154
__bpf_kfunc u32 bpf_cpumask_first_and(const struct cpumask *src1,
kernel/bpf/cpumask.c
155
const struct cpumask *src2)
kernel/bpf/cpumask.c
165
__bpf_kfunc void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
170
cpumask_set_cpu(cpu, (struct cpumask *)cpumask);
kernel/bpf/cpumask.c
178
__bpf_kfunc void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
183
cpumask_clear_cpu(cpu, (struct cpumask *)cpumask);
kernel/bpf/cpumask.c
195
__bpf_kfunc bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask)
kernel/bpf/cpumask.c
200
return cpumask_test_cpu(cpu, (struct cpumask *)cpumask);
kernel/bpf/cpumask.c
212
__bpf_kfunc bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
217
return cpumask_test_and_set_cpu(cpu, (struct cpumask *)cpumask);
kernel/bpf/cpumask.c
230
__bpf_kfunc bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
235
return cpumask_test_and_clear_cpu(cpu, (struct cpumask *)cpumask);
kernel/bpf/cpumask.c
242
__bpf_kfunc void bpf_cpumask_setall(struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
244
cpumask_setall((struct cpumask *)cpumask);
kernel/bpf/cpumask.c
251
__bpf_kfunc void bpf_cpumask_clear(struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
253
cpumask_clear((struct cpumask *)cpumask);
kernel/bpf/cpumask.c
26
cpumask_t cpumask;
kernel/bpf/cpumask.c
269
const struct cpumask *src1,
kernel/bpf/cpumask.c
270
const struct cpumask *src2)
kernel/bpf/cpumask.c
272
return cpumask_and((struct cpumask *)dst, src1, src2);
kernel/bpf/cpumask.c
284
const struct cpumask *src1,
kernel/bpf/cpumask.c
285
const struct cpumask *src2)
kernel/bpf/cpumask.c
287
cpumask_or((struct cpumask *)dst, src1, src2);
kernel/bpf/cpumask.c
299
const struct cpumask *src1,
kernel/bpf/cpumask.c
300
const struct cpumask *src2)
kernel/bpf/cpumask.c
302
cpumask_xor((struct cpumask *)dst, src1, src2);
kernel/bpf/cpumask.c
316
__bpf_kfunc bool bpf_cpumask_equal(const struct cpumask *src1, const struct cpumask *src2)
kernel/bpf/cpumask.c
332
__bpf_kfunc bool bpf_cpumask_intersects(const struct cpumask *src1, const struct cpumask *src2)
kernel/bpf/cpumask.c
348
__bpf_kfunc bool bpf_cpumask_subset(const struct cpumask *src1, const struct cpumask *src2)
kernel/bpf/cpumask.c
363
__bpf_kfunc bool bpf_cpumask_empty(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
365
return cpumask_empty(cpumask);
kernel/bpf/cpumask.c
378
__bpf_kfunc bool bpf_cpumask_full(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
380
return cpumask_full(cpumask);
kernel/bpf/cpumask.c
390
__bpf_kfunc void bpf_cpumask_copy(struct bpf_cpumask *dst, const struct cpumask *src)
kernel/bpf/cpumask.c
392
cpumask_copy((struct cpumask *)dst, src);
kernel/bpf/cpumask.c
405
__bpf_kfunc u32 bpf_cpumask_any_distribute(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
407
return cpumask_any_distribute(cpumask);
kernel/bpf/cpumask.c
423
__bpf_kfunc u32 bpf_cpumask_any_and_distribute(const struct cpumask *src1,
kernel/bpf/cpumask.c
424
const struct cpumask *src2)
kernel/bpf/cpumask.c
438
__bpf_kfunc u32 bpf_cpumask_weight(const struct cpumask *cpumask)
kernel/bpf/cpumask.c
440
return cpumask_weight(cpumask);
kernel/bpf/cpumask.c
457
__bpf_kfunc int bpf_cpumask_populate(struct cpumask *cpumask, void *src, size_t src__sz)
kernel/bpf/cpumask.c
470
bitmap_copy(cpumask_bits(cpumask), src, nr_cpu_ids);
kernel/bpf/cpumask.c
55
struct bpf_cpumask *cpumask;
kernel/bpf/cpumask.c
58
BUILD_BUG_ON(offsetof(struct bpf_cpumask, cpumask) != 0);
kernel/bpf/cpumask.c
60
cpumask = bpf_mem_cache_alloc(&bpf_cpumask_ma);
kernel/bpf/cpumask.c
61
if (!cpumask)
kernel/bpf/cpumask.c
64
memset(cpumask, 0, sizeof(*cpumask));
kernel/bpf/cpumask.c
65
refcount_set(&cpumask->usage, 1);
kernel/bpf/cpumask.c
67
return cpumask;
kernel/bpf/cpumask.c
83
__bpf_kfunc struct bpf_cpumask *bpf_cpumask_acquire(struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
85
refcount_inc(&cpumask->usage);
kernel/bpf/cpumask.c
86
return cpumask;
kernel/bpf/cpumask.c
97
__bpf_kfunc void bpf_cpumask_release(struct bpf_cpumask *cpumask)
kernel/bpf/cpumask.c
99
if (!refcount_dec_and_test(&cpumask->usage))
kernel/cgroup/cpuset-internal.h
303
void cpuset_update_tasks_cpumask(struct cpuset *cs, struct cpumask *new_cpus);
kernel/cgroup/cpuset-internal.h
321
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset-internal.h
335
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset-v1.c
286
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset-v1.c
736
struct cpumask *dp = doms[nslot];
kernel/cgroup/cpuset.c
1054
void cpuset_update_tasks_cpumask(struct cpuset *cs, struct cpumask *new_cpus)
kernel/cgroup/cpuset.c
1062
const struct cpumask *possible_mask = task_cpu_possible_mask(task);
kernel/cgroup/cpuset.c
1088
static void compute_effective_cpumask(struct cpumask *new_cpus,
kernel/cgroup/cpuset.c
1165
struct cpumask *xcpus)
kernel/cgroup/cpuset.c
1200
static void isolated_cpus_update(int old_prs, int new_prs, struct cpumask *xcpus)
kernel/cgroup/cpuset.c
1226
struct cpumask *xcpus)
kernel/cgroup/cpuset.c
1253
struct cpumask *xcpus)
kernel/cgroup/cpuset.c
1283
static bool isolated_cpus_can_update(struct cpumask *add_cpus,
kernel/cgroup/cpuset.c
1284
struct cpumask *del_cpus)
kernel/cgroup/cpuset.c
1319
static bool prstate_housekeeping_conflict(int prstate, struct cpumask *new_cpus)
kernel/cgroup/cpuset.c
1384
struct cpumask *excpus)
kernel/cgroup/cpuset.c
1398
struct cpumask *sibling_xcpus;
kernel/cgroup/cpuset.c
1431
static int compute_excpus(struct cpuset *cs, struct cpumask *excpus)
kernel/cgroup/cpuset.c
1455
struct cpumask *excpus = trialcs->effective_xcpus;
kernel/cgroup/cpuset.c
1584
static void remote_cpus_update(struct cpuset *cs, struct cpumask *xcpus,
kernel/cgroup/cpuset.c
1585
struct cpumask *excpus, struct tmpmasks *tmp)
kernel/cgroup/cpuset.c
1687
struct cpumask *newmask,
kernel/cgroup/cpuset.c
1695
struct cpumask *xcpus = user_xcpus(cs);
kernel/cgroup/cpuset.c
2016
struct cpumask *new_ecpus)
kernel/cgroup/cpuset.c
2301
static int parse_cpuset_cpulist(const char *buf, struct cpumask *out_mask)
kernel/cgroup/cpuset.c
3704
struct cpumask *new_cpus, nodemask_t *new_mems,
kernel/cgroup/cpuset.c
4007
static void __cpuset_cpus_allowed_locked(struct task_struct *tsk, struct cpumask *pmask)
kernel/cgroup/cpuset.c
4020
const struct cpumask *possible_mask = task_cpu_possible_mask(tsk);
kernel/cgroup/cpuset.c
4040
void cpuset_cpus_allowed_locked(struct task_struct *tsk, struct cpumask *pmask)
kernel/cgroup/cpuset.c
4057
void cpuset_cpus_allowed(struct task_struct *tsk, struct cpumask *pmask)
kernel/cgroup/cpuset.c
4082
const struct cpumask *possible_mask = task_cpu_possible_mask(tsk);
kernel/cgroup/cpuset.c
4083
const struct cpumask *cs_mask;
kernel/cgroup/cpuset.c
469
struct cpumask *pmask)
kernel/cgroup/cpuset.c
471
const struct cpumask *possible_mask = task_cpu_possible_mask(tsk);
kernel/cgroup/cpuset.c
620
static inline struct cpumask *user_xcpus(struct cpuset *cs)
kernel/cgroup/cpuset.c
639
struct cpumask *xcpus1 = user_xcpus(cs1);
kernel/cgroup/cpuset.c
640
struct cpumask *xcpus2 = user_xcpus(cs2);
kernel/compat.c
111
unsigned len, struct cpumask *new_mask)
kernel/cpu.c
1764
static void __init cpuhp_bringup_mask(const struct cpumask *mask, unsigned int ncpus,
kernel/cpu.c
1801
static inline const struct cpumask *cpuhp_get_primary_thread_mask(void)
kernel/cpu.c
1810
static inline const struct cpumask *cpuhp_get_primary_thread_mask(void)
kernel/cpu.c
1833
const struct cpumask *mask = cpu_present_mask;
kernel/cpu.c
1841
const struct cpumask *pmask = cpuhp_get_primary_thread_mask();
kernel/cpu.c
1842
static struct cpumask tmp_mask __initdata;
kernel/cpu.c
3082
struct cpumask __cpu_possible_mask __ro_after_init
kernel/cpu.c
3086
struct cpumask __cpu_possible_mask __ro_after_init;
kernel/cpu.c
3092
struct cpumask __cpu_online_mask __read_mostly;
kernel/cpu.c
3095
struct cpumask __cpu_enabled_mask __read_mostly;
kernel/cpu.c
3098
struct cpumask __cpu_present_mask __read_mostly;
kernel/cpu.c
3101
struct cpumask __cpu_active_mask __read_mostly;
kernel/cpu.c
3104
struct cpumask __cpu_dying_mask __read_mostly;
kernel/cpu.c
3110
void init_cpu_present(const struct cpumask *src)
kernel/cpu.c
3115
void init_cpu_possible(const struct cpumask *src)
kernel/events/core.c
12578
static inline const struct cpumask *perf_scope_cpu_topology_cpumask(unsigned int scope, int cpu)
kernel/events/core.c
12596
static inline struct cpumask *perf_scope_cpumask(unsigned int scope)
kernel/events/core.c
12618
struct cpumask *mask = perf_scope_cpumask(pmu->scope);
kernel/events/core.c
12625
static DEVICE_ATTR_RO(cpumask);
kernel/events/core.c
13050
const struct cpumask *cpumask;
kernel/events/core.c
13051
struct cpumask *pmu_cpumask;
kernel/events/core.c
13054
cpumask = perf_scope_cpu_topology_cpumask(pmu->scope, event->cpu);
kernel/events/core.c
13058
if (!pmu_cpumask || !cpumask)
kernel/events/core.c
13061
cpu = cpumask_any_and(pmu_cpumask, cpumask);
kernel/events/core.c
15097
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(scope, cpu);
kernel/events/core.c
15098
struct cpumask *pmu_cpumask = perf_scope_cpumask(scope);
kernel/events/core.c
15101
if (WARN_ON_ONCE(!pmu_cpumask || !cpumask))
kernel/events/core.c
15106
target[scope] = cpumask_any_but(cpumask, cpu);
kernel/events/core.c
15152
struct cpumask *pmu_cpumask;
kernel/events/core.c
15171
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(scope, cpu);
kernel/events/core.c
15175
if (WARN_ON_ONCE(!pmu_cpumask || !cpumask))
kernel/events/core.c
15178
if (!cpumask_empty(cpumask) &&
kernel/events/core.c
15179
cpumask_any_and(pmu_cpumask, cpumask) >= nr_cpu_ids)
kernel/events/core.c
4779
static inline const struct cpumask *perf_scope_cpu_topology_cpumask(unsigned int scope, int cpu);
kernel/events/core.c
4790
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(event->pmu->scope, event_cpu);
kernel/events/core.c
4792
if (cpumask && cpumask_test_cpu(local_cpu, cpumask))
kernel/events/hw_breakpoint.c
359
static const struct cpumask *cpumask_of_bp(struct perf_event *bp)
kernel/events/hw_breakpoint.c
373
const struct cpumask *cpumask = cpumask_of_bp(bp);
kernel/events/hw_breakpoint.c
390
for_each_cpu(cpu, cpumask) {
kernel/events/hw_breakpoint.c
505
const struct cpumask *cpumask = cpumask_of_bp(bp);
kernel/events/hw_breakpoint.c
507
for_each_cpu(cpu, cpumask) {
kernel/irq/affinity.c
73
struct cpumask *result = group_cpus_evenly(this_vecs, &nr_masks);
kernel/irq/chip.c
1381
const struct cpumask *dest, bool force)
kernel/irq/chip.c
1491
int irq_chip_redirect_set_affinity(struct irq_data *data, const struct cpumask *dest, bool force)
kernel/irq/chip.c
170
__irq_startup_managed(struct irq_desc *desc, const struct cpumask *aff,
kernel/irq/chip.c
228
__irq_startup_managed(struct irq_desc *desc, const struct cpumask *aff,
kernel/irq/chip.c
272
const struct cpumask *aff = irq_data_get_affinity_mask(d);
kernel/irq/chip.c
480
const struct cpumask *aff;
kernel/irq/cpuhotplug.c
194
const struct cpumask *hk_mask;
kernel/irq/cpuhotplug.c
209
const struct cpumask *affinity = irq_data_get_affinity_mask(data);
kernel/irq/cpuhotplug.c
22
const struct cpumask *m = irq_data_get_effective_affinity_mask(d);
kernel/irq/cpuhotplug.c
58
const struct cpumask *affinity;
kernel/irq/debugfs.c
27
const struct cpumask *msk;
kernel/irq/internals.h
137
const struct cpumask *dest, bool force);
kernel/irq/internals.h
312
irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask)
kernel/irq/internals.h
317
irq_get_pending(struct cpumask *mask, struct irq_desc *desc)
kernel/irq/internals.h
321
static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc)
kernel/irq/internals.h
337
irq_copy_pending(struct irq_desc *desc, const struct cpumask *mask)
kernel/irq/internals.h
341
irq_get_pending(struct cpumask *mask, struct irq_desc *desc)
kernel/irq/internals.h
344
static inline struct cpumask *irq_desc_get_pending_mask(struct irq_desc *desc)
kernel/irq/ipi-mux.c
54
static void ipi_mux_send_mask(struct irq_data *d, const struct cpumask *mask)
kernel/irq/ipi.c
115
int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest)
kernel/irq/ipi.c
118
const struct cpumask *ipimask;
kernel/irq/ipi.c
166
const struct cpumask *ipimask;
kernel/irq/ipi.c
189
const struct cpumask *dest, unsigned int cpu)
kernel/irq/ipi.c
191
const struct cpumask *ipimask;
kernel/irq/ipi.c
24
const struct cpumask *dest)
kernel/irq/ipi.c
269
int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest)
kernel/irq/ipi.c
334
int ipi_send_mask(unsigned int virq, const struct cpumask *dest)
kernel/irq/irq_test.c
23
static int noop_affinity(struct irq_data *data, const struct cpumask *dest,
kernel/irq/irqdesc.c
1002
static unsigned int kstat_irqs_desc(struct irq_desc *desc, const struct cpumask *cpumask)
kernel/irq/irqdesc.c
1012
for_each_cpu(cpu, cpumask)
kernel/irq/irqdesc.c
116
desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { }
kernel/irq/irqdesc.c
121
const struct cpumask *affinity, struct module *owner)
kernel/irq/irqdesc.c
213
const struct cpumask *affinity,
kernel/irq/irqdesc.c
434
const struct cpumask *affinity,
kernel/irq/irqdesc.c
513
const struct cpumask *mask = NULL;
kernel/irq/irqdesc.c
780
const struct cpumask *m = irq_data_get_effective_affinity_mask(&desc->irq_data);
kernel/irq/irqdesc.c
86
static void desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity)
kernel/irq/manage.c
1041
const struct cpumask *m;
kernel/irq/manage.c
218
const struct cpumask *m = irq_data_get_effective_affinity_mask(data);
kernel/irq/manage.c
230
static DEFINE_PER_CPU(struct cpumask, __tmp_mask);
kernel/irq/manage.c
232
int irq_do_set_affinity(struct irq_data *data, const struct cpumask *mask, bool force)
kernel/irq/manage.c
234
struct cpumask *tmp_mask = this_cpu_ptr(&__tmp_mask);
kernel/irq/manage.c
237
const struct cpumask *prog_mask;
kernel/irq/manage.c
2585
const struct cpumask *affinity, void __percpu *dev_id)
kernel/irq/manage.c
264
const struct cpumask *hk_mask;
kernel/irq/manage.c
306
const struct cpumask *dest)
kernel/irq/manage.c
316
const struct cpumask *dest)
kernel/irq/manage.c
323
const struct cpumask *dest, bool force)
kernel/irq/manage.c
338
const struct cpumask *mask)
kernel/irq/manage.c
376
int irq_set_affinity_locked(struct irq_data *data, const struct cpumask *mask,
kernel/irq/manage.c
462
static int __irq_set_affinity(unsigned int irq, const struct cpumask *mask,
kernel/irq/manage.c
481
int irq_set_affinity(unsigned int irq, const struct cpumask *cpumask)
kernel/irq/manage.c
483
return __irq_set_affinity(irq, cpumask, false);
kernel/irq/manage.c
498
int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask)
kernel/irq/manage.c
500
return __irq_set_affinity(irq, cpumask, true);
kernel/irq/manage.c
504
int __irq_apply_affinity_hint(unsigned int irq, const struct cpumask *m, bool setaffinity)
kernel/irq/manage.c
523
cpumask_var_t cpumask;
kernel/irq/manage.c
525
if (!desc || !alloc_cpumask_var(&cpumask, GFP_KERNEL))
kernel/irq/manage.c
530
irq_get_pending(cpumask, desc);
kernel/irq/manage.c
532
cpumask_copy(cpumask, desc->irq_common_data.affinity);
kernel/irq/manage.c
535
notify->notify(notify, cpumask);
kernel/irq/manage.c
537
free_cpumask_var(cpumask);
kernel/irq/manage.c
594
struct cpumask *set = irq_default_affinity;
kernel/irq/manage.c
598
static struct cpumask mask;
kernel/irq/manage.c
623
const struct cpumask *nodemask = cpumask_of_node(node);
kernel/irq/matrix.c
136
const struct cpumask *msk)
kernel/irq/matrix.c
157
const struct cpumask *msk)
kernel/irq/matrix.c
216
int irq_matrix_reserve_managed(struct irq_matrix *m, const struct cpumask *msk)
kernel/irq/matrix.c
257
void irq_matrix_remove_managed(struct irq_matrix *m, const struct cpumask *msk)
kernel/irq/matrix.c
292
int irq_matrix_alloc_managed(struct irq_matrix *m, const struct cpumask *msk,
kernel/irq/matrix.c
383
int irq_matrix_alloc(struct irq_matrix *m, const struct cpumask *msk,
kernel/irq/msi.c
665
const struct cpumask *mask, bool force)
kernel/irq/proc.c
50
const struct cpumask *mask;
kernel/kthread.c
330
static void kthread_fetch_affinity(struct kthread *kthread, struct cpumask *cpumask)
kernel/kthread.c
332
const struct cpumask *pref;
kernel/kthread.c
345
cpumask_and(cpumask, pref, housekeeping_cpumask(HK_TYPE_DOMAIN));
kernel/kthread.c
346
if (cpumask_empty(cpumask))
kernel/kthread.c
347
cpumask_copy(cpumask, housekeeping_cpumask(HK_TYPE_DOMAIN));
kernel/kthread.c
566
static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, unsigned int state)
kernel/kthread.c
585
void kthread_bind_mask(struct task_struct *p, const struct cpumask *mask)
kernel/kthread.c
73
struct cpumask *preferred_affinity;
kernel/kthread.c
839
int kthread_affine_preferred(struct task_struct *p, const struct cpumask *mask)
kernel/kthread.c
855
kthread->preferred_affinity = kzalloc(sizeof(struct cpumask), GFP_KERNEL);
kernel/locking/locktorture.c
109
long torture_sched_setaffinity(pid_t pid, const struct cpumask *in_mask, bool dowarn);
kernel/padata.c
1004
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
1005
free_cpumask_var(pinst->cpumask.cbcpu);
kernel/padata.c
196
if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) {
kernel/padata.c
197
if (cpumask_empty(pd->cpumask.cbcpu))
kernel/padata.c
201
cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu);
kernel/padata.c
202
*cb_cpu = cpumask_nth(cpu_index, pd->cpumask.cbcpu);
kernel/padata.c
297
cpu = cpumask_first(pd->cpumask.pcpu);
kernel/padata.c
299
cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu);
kernel/padata.c
397
cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu);
kernel/padata.c
515
for_each_cpu(cpu, pd->cpumask.cbcpu) {
kernel/padata.c
530
for_each_cpu(cpu, pd->cpumask.pcpu) {
kernel/padata.c
557
if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL))
kernel/padata.c
559
if (!alloc_cpumask_var(&pd->cpumask.cbcpu, GFP_KERNEL))
kernel/padata.c
562
cpumask_and(pd->cpumask.pcpu, pinst->cpumask.pcpu, cpu_online_mask);
kernel/padata.c
563
cpumask_and(pd->cpumask.cbcpu, pinst->cpumask.cbcpu, cpu_online_mask);
kernel/padata.c
569
pd->cpu = cpumask_first(pd->cpumask.pcpu);
kernel/padata.c
574
free_cpumask_var(pd->cpumask.pcpu);
kernel/padata.c
587
free_cpumask_var(pd->cpumask.pcpu);
kernel/padata.c
588
free_cpumask_var(pd->cpumask.cbcpu);
kernel/padata.c
649
const struct cpumask *cpumask)
kernel/padata.c
651
if (!cpumask_intersects(cpumask, cpu_online_mask)) {
kernel/padata.c
678
cpumask_copy(pinst->cpumask.pcpu, pcpumask);
kernel/padata.c
679
cpumask_copy(pinst->cpumask.cbcpu, cbcpumask);
kernel/padata.c
700
cpumask_var_t cpumask)
kernel/padata.c
702
struct cpumask *serial_mask, *parallel_mask;
kernel/padata.c
710
serial_mask = pinst->cpumask.cbcpu;
kernel/padata.c
711
parallel_mask = cpumask;
kernel/padata.c
714
parallel_mask = pinst->cpumask.pcpu;
kernel/padata.c
715
serial_mask = cpumask;
kernel/padata.c
72
int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu);
kernel/padata.c
74
return cpumask_nth(cpu_index, pd->cpumask.pcpu);
kernel/padata.c
740
if (padata_validate_cpumask(pinst, pinst->cpumask.pcpu) &&
kernel/padata.c
741
padata_validate_cpumask(pinst, pinst->cpumask.cbcpu))
kernel/padata.c
753
if (!padata_validate_cpumask(pinst, pinst->cpumask.pcpu) ||
kernel/padata.c
754
!padata_validate_cpumask(pinst, pinst->cpumask.cbcpu))
kernel/padata.c
765
return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) ||
kernel/padata.c
766
cpumask_test_cpu(cpu, pinst->cpumask.cbcpu);
kernel/padata.c
812
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
813
free_cpumask_var(pinst->cpumask.cbcpu);
kernel/padata.c
840
struct cpumask *cpumask;
kernel/padata.c
845
cpumask = pinst->cpumask.cbcpu;
kernel/padata.c
847
cpumask = pinst->cpumask.pcpu;
kernel/padata.c
850
nr_cpu_ids, cpumask_bits(cpumask));
kernel/padata.c
972
if (!alloc_cpumask_var(&pinst->cpumask.pcpu, GFP_KERNEL))
kernel/padata.c
974
if (!alloc_cpumask_var(&pinst->cpumask.cbcpu, GFP_KERNEL)) {
kernel/padata.c
975
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
981
cpumask_copy(pinst->cpumask.pcpu, cpu_possible_mask);
kernel/padata.c
982
cpumask_copy(pinst->cpumask.cbcpu, cpu_possible_mask);
kernel/power/em_netlink.c
52
struct cpumask *cpumask;
kernel/power/em_netlink.c
63
cpumask = to_cpumask(pd->cpus);
kernel/power/em_netlink.c
64
for_each_cpu(cpu, cpumask) {
kernel/printk/printk_ringbuffer_kunit_test.c
226
KUNIT_DEFINE_ACTION_WRAPPER(prbtest_cpumask_cleanup, free_cpumask_var, struct cpumask *);
kernel/rcu/tree_nocb.h
1266
const struct cpumask *cpumask = NULL;
kernel/rcu/tree_nocb.h
1271
cpumask = tick_nohz_full_mask;
kernel/rcu/tree_nocb.h
1275
!rcu_state.nocb_is_setup && !cpumask)
kernel/rcu/tree_nocb.h
1276
cpumask = cpu_possible_mask;
kernel/rcu/tree_nocb.h
1278
if (cpumask) {
kernel/rcu/tree_nocb.h
1286
cpumask_or(rcu_nocb_mask, rcu_nocb_mask, cpumask);
kernel/rcu/update.c
530
long torture_sched_setaffinity(pid_t pid, const struct cpumask *in_mask, bool dowarn)
kernel/sched/core.c
10520
static inline void mm_update_cpus_allowed(struct mm_struct *mm, const struct cpumask *affmsk)
kernel/sched/core.c
10522
struct cpumask *mm_allowed;
kernel/sched/core.c
10863
static inline void mm_update_cpus_allowed(struct mm_struct *mm, const struct cpumask *affmsk) { }
kernel/sched/core.c
1209
const struct cpumask *hk_mask;
kernel/sched/core.c
2705
void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask)
kernel/sched/core.c
2713
cpumask_t cpumask;
kernel/sched/core.c
2771
static inline struct cpumask *clear_user_cpus_ptr(struct task_struct *p)
kernel/sched/core.c
2773
struct cpumask *user_mask = NULL;
kernel/sched/core.c
3020
const struct cpumask *cpu_allowed_mask = task_cpu_possible_mask(p);
kernel/sched/core.c
3021
const struct cpumask *cpu_valid_mask = cpu_active_mask;
kernel/sched/core.c
3117
int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask)
kernel/sched/core.c
3138
struct cpumask *new_mask,
kernel/sched/core.c
3139
const struct cpumask *subset_mask)
kernel/sched/core.c
3182
const struct cpumask *override_mask = task_cpu_possible_mask(p);
kernel/sched/core.c
3450
const struct cpumask *nodemask = NULL;
kernel/sched/core.c
399
static struct cpumask sched_core_mask;
kernel/sched/core.c
405
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/sched/core.c
417
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/sched/core.c
437
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/sched/core.c
6014
const struct cpumask *smt_mask;
kernel/sched/core.c
6362
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/sched/core.c
6401
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/sched/core.c
7992
int cpuset_cpumask_can_shrink(const struct cpumask *cur,
kernel/sched/core.c
7993
const struct cpumask *trial)
kernel/sched/core_sched.c
244
const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq));
kernel/sched/cpudeadline.c
118
struct cpumask *later_mask)
kernel/sched/cpudeadline.h
20
int cpudl_find(struct cpudl *cp, struct task_struct *p, struct cpumask *later_mask);
kernel/sched/cpupri.c
123
struct cpumask *lowest_mask)
kernel/sched/cpupri.c
146
struct cpumask *lowest_mask,
kernel/sched/cpupri.c
69
struct cpumask *lowest_mask, int idx)
kernel/sched/cpupri.h
24
struct cpumask *lowest_mask);
kernel/sched/cpupri.h
26
struct cpumask *lowest_mask,
kernel/sched/deadline.c
135
static inline unsigned long __dl_bw_capacity(const struct cpumask *mask)
kernel/sched/deadline.c
2718
struct cpumask *later_mask = this_cpu_cpumask_var_ptr(local_cpu_mask_dl);
kernel/sched/deadline.c
3160
static void dl_get_task_effective_cpus(struct task_struct *p, struct cpumask *cpus)
kernel/sched/deadline.c
3162
const struct cpumask *hk_msk;
kernel/sched/deadline.c
3191
struct cpumask *msk;
kernel/sched/deadline.c
3730
int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur,
kernel/sched/deadline.c
3731
const struct cpumask *trial)
kernel/sched/ext.c
2744
p, (struct cpumask *)p->cpus_ptr);
kernel/sched/ext.c
3243
p, (struct cpumask *)p->cpus_ptr);
kernel/sched/ext.c
3969
struct cpumask *donee_mask, struct cpumask *resched_mask,
kernel/sched/ext.c
4066
const struct cpumask *node_mask = cpumask_of_node(node);
kernel/sched/ext.c
4067
struct cpumask *donee_mask = scx_bypass_lb_donee_cpumask;
kernel/sched/ext.c
4068
struct cpumask *resched_mask = scx_bypass_lb_resched_cpumask;
kernel/sched/ext.c
5534
static void sched_ext_ops__set_cpumask(struct task_struct *p, const struct cpumask *mask) {}
kernel/sched/ext.c
7160
__bpf_kfunc const struct cpumask *scx_bpf_get_possible_cpumask(void)
kernel/sched/ext.c
7168
__bpf_kfunc const struct cpumask *scx_bpf_get_online_cpumask(void)
kernel/sched/ext.c
7177
__bpf_kfunc void scx_bpf_put_cpumask(const struct cpumask *cpumask)
kernel/sched/ext_idle.c
1036
__scx_bpf_select_cpu_and(struct task_struct *p, const struct cpumask *cpus_allowed,
kernel/sched/ext_idle.c
1055
const struct cpumask *cpus_allowed, u64 flags)
kernel/sched/ext_idle.c
1078
__bpf_kfunc const struct cpumask *scx_bpf_get_idle_cpumask_node(int node)
kernel/sched/ext_idle.c
1102
__bpf_kfunc const struct cpumask *scx_bpf_get_idle_cpumask(void)
kernel/sched/ext_idle.c
1133
__bpf_kfunc const struct cpumask *scx_bpf_get_idle_smtmask_node(int node)
kernel/sched/ext_idle.c
115
static s32 pick_idle_cpu_in_node(const struct cpumask *cpus_allowed, int node, u64 flags)
kernel/sched/ext_idle.c
1161
__bpf_kfunc const struct cpumask *scx_bpf_get_idle_smtmask(void)
kernel/sched/ext_idle.c
1190
__bpf_kfunc void scx_bpf_put_idle_cpumask(const struct cpumask *idle_mask)
kernel/sched/ext_idle.c
1248
__bpf_kfunc s32 scx_bpf_pick_idle_cpu_node(const struct cpumask *cpus_allowed,
kernel/sched/ext_idle.c
1288
__bpf_kfunc s32 scx_bpf_pick_idle_cpu(const struct cpumask *cpus_allowed,
kernel/sched/ext_idle.c
1331
__bpf_kfunc s32 scx_bpf_pick_any_cpu_node(const struct cpumask *cpus_allowed,
kernel/sched/ext_idle.c
1378
__bpf_kfunc s32 scx_bpf_pick_any_cpu(const struct cpumask *cpus_allowed,
kernel/sched/ext_idle.c
150
static s32 pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags)
kernel/sched/ext_idle.c
191
pick_idle_cpu_from_online_nodes(const struct cpumask *cpus_allowed, int node, u64 flags)
kernel/sched/ext_idle.c
200
static s32 scx_pick_idle_cpu(const struct cpumask *cpus_allowed, int node, u64 flags)
kernel/sched/ext_idle.c
246
static struct cpumask *llc_span(s32 cpu)
kernel/sched/ext_idle.c
280
static struct cpumask *numa_span(s32 cpu)
kernel/sched/ext_idle.c
452
const struct cpumask *cpus_allowed, u64 flags)
kernel/sched/ext_idle.c
454
const struct cpumask *llc_cpus = NULL, *numa_cpus = NULL;
kernel/sched/ext_idle.c
455
const struct cpumask *allowed = cpus_allowed ?: p->cpus_ptr;
kernel/sched/ext_idle.c
472
struct cpumask *local_cpus = this_cpu_cpumask_var_ptr(local_idle_cpumask);
kernel/sched/ext_idle.c
497
struct cpumask *local_cpus = this_cpu_cpumask_var_ptr(local_numa_idle_cpumask);
kernel/sched/ext_idle.c
498
const struct cpumask *cpus = numa_span(prev_cpu);
kernel/sched/ext_idle.c
507
struct cpumask *local_cpus = this_cpu_cpumask_var_ptr(local_llc_idle_cpumask);
kernel/sched/ext_idle.c
508
const struct cpumask *cpus = llc_span(prev_cpu);
kernel/sched/ext_idle.c
693
struct cpumask *idle_cpus = idle_cpumask(node)->cpu;
kernel/sched/ext_idle.c
699
const struct cpumask *smt = cpu_smt_mask(cpu);
kernel/sched/ext_idle.c
700
struct cpumask *idle_smts = idle_cpumask(node)->smt;
kernel/sched/ext_idle.c
789
const struct cpumask *node_mask = cpumask_of_node(node);
kernel/sched/ext_idle.c
80
struct cpumask *idle_cpus = idle_cpumask(node)->cpu;
kernel/sched/ext_idle.c
89
const struct cpumask *smt = cpu_smt_mask(cpu);
kernel/sched/ext_idle.c
893
const struct cpumask *allowed, u64 flags)
kernel/sched/ext_idle.c
90
struct cpumask *idle_smts = idle_cpumask(node)->smt;
kernel/sched/ext_idle.h
19
const struct cpumask *cpus_allowed, u64 flags);
kernel/sched/ext_internal.h
498
const struct cpumask *cpumask);
kernel/sched/fair.c
11760
struct cpumask *swb_cpus = this_cpu_cpumask_var_ptr(should_we_balance_tmpmask);
kernel/sched/fair.c
11874
struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask);
kernel/sched/fair.c
12418
const struct cpumask *hk_mask;
kernel/sched/fair.c
7622
static int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle_cpu)
kernel/sched/fair.c
7684
static inline int select_idle_core(struct task_struct *p, int core, struct cpumask *cpus, int *idle_cpu)
kernel/sched/fair.c
7703
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
kernel/sched/fair.c
7776
struct cpumask *cpus;
kernel/sched/fair.c
8258
struct cpumask *pd_cpus,
kernel/sched/fair.c
8281
eenv_pd_max_util(struct energy_env *eenv, struct cpumask *pd_cpus,
kernel/sched/fair.c
8329
struct cpumask *pd_cpus, struct task_struct *p, int dst_cpu)
kernel/sched/fair.c
8386
struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask);
kernel/sched/fair.c
8711
cpumask_t *cpumask;
kernel/sched/fair.c
8713
cpumask = cpu_capacity_span(entry);
kernel/sched/fair.c
8714
if (!cpumask_intersects(p->cpus_ptr, cpumask))
kernel/sched/fair.c
9334
struct cpumask *dst_grpmask;
kernel/sched/fair.c
9339
struct cpumask *cpus;
kernel/sched/isolation.c
121
int housekeeping_update(struct cpumask *isol_mask)
kernel/sched/isolation.c
123
struct cpumask *trial, *old = NULL;
kernel/sched/isolation.c
181
struct cpumask *omask, *nmask = kmalloc(cpumask_size(), GFP_KERNEL);
kernel/sched/isolation.c
199
struct cpumask *mask = memblock_alloc_or_panic(cpumask_size(), SMP_CACHE_BYTES);
kernel/sched/isolation.c
25
struct cpumask __rcu *cpumasks[HK_TYPE_MAX];
kernel/sched/isolation.c
58
static inline struct cpumask *housekeeping_cpumask_dereference(enum hk_type type)
kernel/sched/isolation.c
64
const struct cpumask *housekeeping_cpumask(enum hk_type type)
kernel/sched/isolation.c
66
const struct cpumask *mask = NULL;
kernel/sched/rt.c
1766
struct cpumask *lowest_mask = this_cpu_cpumask_var_ptr(local_cpu_mask);
kernel/sched/rt.c
581
static inline const struct cpumask *sched_rt_period_mask(void)
kernel/sched/rt.c
781
const struct cpumask *span;
kernel/sched/rt.c
942
static inline const struct cpumask *sched_rt_period_mask(void)
kernel/sched/sched.h
1053
extern int sched_init_domains(const struct cpumask *cpu_map);
kernel/sched/sched.h
1431
static inline struct cpumask *sched_group_span(struct sched_group *sg);
kernel/sched/sched.h
2025
extern int sched_numa_find_closest(const struct cpumask *cpus, int cpu);
kernel/sched/sched.h
2034
static inline int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
kernel/sched/sched.h
2181
unsigned long cpumask[]; /* Balance mask */
kernel/sched/sched.h
2201
unsigned long cpumask[];
kernel/sched/sched.h
2204
static inline struct cpumask *sched_group_span(struct sched_group *sg)
kernel/sched/sched.h
2206
return to_cpumask(sg->cpumask);
kernel/sched/sched.h
2212
static inline struct cpumask *group_balance_mask(struct sched_group *sg)
kernel/sched/sched.h
2214
return to_cpumask(sg->sgc->cpumask);
kernel/sched/sched.h
2224
static inline const struct cpumask *task_user_cpus(struct task_struct *p)
kernel/sched/sched.h
2493
const struct cpumask *new_mask;
kernel/sched/sched.h
2494
struct cpumask *user_mask;
kernel/sched/sched.h
362
extern int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
kernel/sched/syscalls.c
1101
int dl_task_check_affinity(struct task_struct *p, const struct cpumask *mask)
kernel/sched/syscalls.c
1191
long sched_setaffinity(pid_t pid, const struct cpumask *in_mask)
kernel/sched/syscalls.c
1194
struct cpumask *user_mask;
kernel/sched/syscalls.c
1238
struct cpumask *new_mask)
kernel/sched/syscalls.c
1272
long sched_getaffinity(pid_t pid, struct cpumask *mask)
kernel/sched/topology.c
1033
const struct cpumask *span = sched_domain_span(sd);
kernel/sched/topology.c
1034
struct cpumask *covered = sched_domains_tmpmask;
kernel/sched/topology.c
1042
struct cpumask *sg_span;
kernel/sched/topology.c
1245
const struct cpumask *span = sched_domain_span(sd);
kernel/sched/topology.c
1246
struct cpumask *covered;
kernel/sched/topology.c
1289
struct cpumask *mask = sched_domains_tmpmask2;
kernel/sched/topology.c
1395
asym_cpu_capacity_classify(const struct cpumask *sd_span,
kernel/sched/topology.c
1396
const struct cpumask *cpu_map)
kernel/sched/topology.c
1534
static void __sdt_free(const struct cpumask *cpu_map);
kernel/sched/topology.c
1535
static int __sdt_alloc(const struct cpumask *cpu_map);
kernel/sched/topology.c
1538
const struct cpumask *cpu_map)
kernel/sched/topology.c
1557
__visit_domain_allocation_hell(struct s_data *d, const struct cpumask *cpu_map)
kernel/sched/topology.c
1609
static struct cpumask ***sched_domains_numa_masks;
kernel/sched/topology.c
1638
const struct cpumask *cpu_map,
kernel/sched/topology.c
1644
struct cpumask *sd_span;
kernel/sched/topology.c
1751
const struct cpumask *tl_smt_mask(struct sched_domain_topology_level *tl, int cpu)
kernel/sched/topology.c
1763
const struct cpumask *tl_cls_mask(struct sched_domain_topology_level *tl, int cpu)
kernel/sched/topology.c
1775
const struct cpumask *tl_mc_mask(struct sched_domain_topology_level *tl, int cpu)
kernel/sched/topology.c
1781
const struct cpumask *tl_pkg_mask(struct sched_domain_topology_level *tl, int cpu)
kernel/sched/topology.c
1827
static const struct cpumask *sd_numa_mask(struct sched_domain_topology_level *tl, int cpu)
kernel/sched/topology.c
2021
struct cpumask ***masks;
kernel/sched/topology.c
2077
struct cpumask *mask = kzalloc(cpumask_size(), GFP_KERNEL);
kernel/sched/topology.c
214
static bool sched_is_eas_possible(const struct cpumask *cpu_mask)
kernel/sched/topology.c
2140
struct cpumask ***masks;
kernel/sched/topology.c
2233
int sched_numa_find_closest(const struct cpumask *cpus, int cpu)
kernel/sched/topology.c
2236
struct cpumask ***masks;
kernel/sched/topology.c
2258
const struct cpumask *cpus;
kernel/sched/topology.c
2259
struct cpumask ***masks;
kernel/sched/topology.c
2267
struct cpumask **prev_hop, **cur_hop = *(struct cpumask ***)b;
kernel/sched/topology.c
2278
prev_hop = *((struct cpumask ***)b - 1);
kernel/sched/topology.c
2296
int sched_numa_find_nth_cpu(const struct cpumask *cpus, int cpu, int node)
kernel/sched/topology.c
2299
struct cpumask ***hop_masks;
kernel/sched/topology.c
2346
const struct cpumask *sched_numa_hop_mask(unsigned int node, unsigned int hops)
kernel/sched/topology.c
2348
struct cpumask ***masks;
kernel/sched/topology.c
2363
static int __sdt_alloc(const struct cpumask *cpu_map)
kernel/sched/topology.c
2430
static void __sdt_free(const struct cpumask *cpu_map)
kernel/sched/topology.c
2467
const struct cpumask *cpu_map, struct sched_domain_attr *attr,
kernel/sched/topology.c
2498
static bool topology_span_sane(const struct cpumask *cpu_map)
kernel/sched/topology.c
2501
struct cpumask *covered, *id_seen;
kernel/sched/topology.c
2528
const struct cpumask *tl_cpu_mask = tl->mask(tl, cpu);
kernel/sched/topology.c
2556
build_sched_domains(const struct cpumask *cpu_map, struct sched_domain_attr *attr)
kernel/sched/topology.c
2761
int __init sched_init_domains(const struct cpumask *cpu_map)
kernel/sched/topology.c
2785
static void detach_destroy_domains(const struct cpumask *cpu_map)
kernel/sched/topology.c
361
static void perf_domain_debug(const struct cpumask *cpu_map,
kernel/sched/topology.c
409
static bool build_perf_domains(const struct cpumask *cpu_map)
kernel/sched/topology.c
44
struct cpumask *groupmask)
kernel/sched/topology.c
916
build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask)
kernel/sched/topology.c
918
const struct cpumask *sg_span = sched_group_span(sg);
kernel/sched/topology.c
956
struct cpumask *sg_span;
kernel/sched/topology.c
979
struct cpumask *mask = sched_domains_tmpmask2;
kernel/sched/topology.c
981
struct cpumask *sg_span;
kernel/smp.c
1035
void *info, bool wait, const struct cpumask *mask)
kernel/smp.c
1100
bool cpus_peek_for_pending_ipi(const struct cpumask *mask)
kernel/smp.c
125
send_call_function_ipi_mask(struct cpumask *mask)
kernel/smp.c
42
cpumask_var_t cpumask;
kernel/smp.c
58
if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL,
kernel/smp.c
63
free_cpumask_var(cfd->cpumask);
kernel/smp.c
68
free_cpumask_var(cfd->cpumask);
kernel/smp.c
754
int smp_call_function_any(const struct cpumask *mask,
kernel/smp.c
780
static void smp_call_function_many_cond(const struct cpumask *mask,
kernel/smp.c
80
free_cpumask_var(cfd->cpumask);
kernel/smp.c
814
cpumask_and(cfd->cpumask, mask, cpu_online_mask);
kernel/smp.c
815
__cpumask_clear_cpu(this_cpu, cfd->cpumask);
kernel/smp.c
818
for_each_cpu(cpu, cfd->cpumask) {
kernel/smp.c
822
__cpumask_clear_cpu(cpu, cfd->cpumask);
kernel/smp.c
873
for_each_cpu(cpu, cfd->cpumask) {
kernel/smp.c
897
void smp_call_function_many(const struct cpumask *mask,
kernel/stop_machine.c
173
const struct cpumask *active_cpus;
kernel/stop_machine.c
195
notrace void __weak stop_machine_yield(const struct cpumask *cpumask)
kernel/stop_machine.c
206
const struct cpumask *cpumask;
kernel/stop_machine.c
217
cpumask = cpu_online_mask;
kernel/stop_machine.c
218
is_active = cpu == cpumask_first(cpumask);
kernel/stop_machine.c
220
cpumask = msdata->active_cpus;
kernel/stop_machine.c
221
is_active = cpumask_test_cpu(cpu, cpumask);
kernel/stop_machine.c
227
stop_machine_yield(cpumask);
kernel/stop_machine.c
392
static bool queue_stop_cpus_work(const struct cpumask *cpumask,
kernel/stop_machine.c
408
for_each_cpu(cpu, cpumask) {
kernel/stop_machine.c
424
static int __stop_cpus(const struct cpumask *cpumask,
kernel/stop_machine.c
429
cpu_stop_init_done(&done, cpumask_weight(cpumask));
kernel/stop_machine.c
430
if (!queue_stop_cpus_work(cpumask, fn, arg, &done))
kernel/stop_machine.c
464
static int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg)
kernel/stop_machine.c
470
ret = __stop_cpus(cpumask, fn, arg);
kernel/stop_machine.c
588
const struct cpumask *cpus)
kernel/stop_machine.c
623
int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus)
kernel/stop_machine.c
638
const struct cpumask *smt_mask = cpu_smt_mask(cpu);
kernel/stop_machine.c
679
const struct cpumask *cpus)
kernel/taskstats.c
295
static int add_del_listener(pid_t pid, const struct cpumask *mask, int isadd)
kernel/taskstats.c
354
static int parse(struct nlattr *na, struct cpumask *mask)
kernel/time/clockevents.c
458
if (!dev->cpumask) {
kernel/time/clockevents.c
460
dev->cpumask = cpumask_of(smp_processor_id());
kernel/time/clockevents.c
463
if (dev->cpumask == cpu_all_mask) {
kernel/time/clockevents.c
466
dev->cpumask = cpu_possible_mask;
kernel/time/clockevents.c
649
if (cpumask_test_cpu(cpu, dev->cpumask) &&
kernel/time/clockevents.c
650
cpumask_weight(dev->cpumask) == 1 &&
kernel/time/tick-broadcast-hrtimer.c
91
.cpumask = cpu_possible_mask,
kernel/time/tick-broadcast.c
132
if (!cpumask_equal(newdev->cpumask, cpumask_of(cpu)))
kernel/time/tick-broadcast.c
227
static void err_broadcast(const struct cpumask *mask)
kernel/time/tick-broadcast.c
346
static bool tick_do_broadcast(struct cpumask *mask)
kernel/time/tick-broadcast.c
614
struct cpumask *tick_get_broadcast_oneshot_mask(void)
kernel/time/tick-broadcast.c
62
struct cpumask *tick_get_broadcast_mask(void)
kernel/time/tick-broadcast.c
639
const struct cpumask *cpumask)
kernel/time/tick-broadcast.c
644
if (cpumask_equal(bc->cpumask, cpumask))
kernel/time/tick-broadcast.c
647
bc->cpumask = cpumask;
kernel/time/tick-broadcast.c
648
irq_set_affinity(bc->irq, bc->cpumask);
kernel/time/tick-broadcast.c
992
static void tick_broadcast_init_next_event(struct cpumask *mask,
kernel/time/tick-common.c
186
const struct cpumask *cpumask)
kernel/time/tick-common.c
242
if (!cpumask_equal(newdev->cpumask, cpumask))
kernel/time/tick-common.c
243
irq_set_affinity(newdev->irq, cpumask);
kernel/time/tick-common.c
275
if (!cpumask_test_cpu(cpu, newdev->cpumask))
kernel/time/tick-common.c
277
if (cpumask_equal(newdev->cpumask, cpumask_of(cpu)))
kernel/time/tick-common.c
283
if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu)))
kernel/time/tick-common.c
305
!cpumask_equal(curdev->cpumask, newdev->cpumask);
kernel/time/tick-internal.h
137
extern struct cpumask *tick_get_broadcast_oneshot_mask(void);
kernel/time/tick-internal.h
78
extern struct cpumask *tick_get_broadcast_mask(void);
kernel/time/tick-sched.c
625
void __init tick_nohz_full_setup(cpumask_var_t cpumask)
kernel/time/tick-sched.c
628
cpumask_copy(tick_nohz_full_mask, cpumask);
kernel/time/timer_migration.c
1555
int tmigr_isolated_exclude_cpumask(struct cpumask *exclude_cpumask)
kernel/time/timer_migration.c
1559
cpumask_var_t cpumask __free(free_cpumask_var) = CPUMASK_VAR_NULL;
kernel/time/timer_migration.c
1564
if (!alloc_cpumask_var(&cpumask, GFP_KERNEL))
kernel/time/timer_migration.c
1572
cpumask_andnot(cpumask, cpu_online_mask, exclude_cpumask);
kernel/time/timer_migration.c
1573
cpumask_andnot(cpumask, cpumask, tmigr_available_cpumask);
kernel/time/timer_migration.c
1575
for_each_cpu(cpu, cpumask) {
kernel/time/timer_migration.c
1581
for_each_cpu(cpu, cpumask)
kernel/time/timer_migration.c
1589
cpumask_and(cpumask, exclude_cpumask, tmigr_available_cpumask);
kernel/time/timer_migration.c
1590
cpumask_and(cpumask, cpumask, housekeeping_cpumask(HK_TYPE_KERNEL_NOISE));
kernel/time/timer_migration.c
1595
for_each_cpu(cpu, cpumask) {
kernel/time/timer_migration.c
1597
cpumask_clear_cpu(cpu, cpumask);
kernel/time/timer_migration.c
1602
for_each_cpu(cpu, cpumask) {
kernel/time/timer_migration.c
1608
for_each_cpu(cpu, cpumask)
kernel/time/timer_migration.c
1616
cpumask_var_t cpumask __free(free_cpumask_var) = CPUMASK_VAR_NULL;
kernel/time/timer_migration.c
1622
if (!alloc_cpumask_var(&cpumask, GFP_KERNEL))
kernel/time/timer_migration.c
1625
cpumask_andnot(cpumask, cpu_possible_mask, housekeeping_cpumask(HK_TYPE_DOMAIN));
kernel/time/timer_migration.c
1628
return tmigr_isolated_exclude_cpumask(cpumask);
kernel/trace/ring_buffer.c
1004
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
1054
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
2493
if (!zalloc_cpumask_var(&buffer->cpumask, GFP_KERNEL))
kernel/trace/ring_buffer.c
2582
cpumask_set_cpu(cpu, buffer->cpumask);
kernel/trace/ring_buffer.c
2603
free_cpumask_var(buffer->cpumask);
kernel/trace/ring_buffer.c
2688
free_cpumask_var(buffer->cpumask);
kernel/trace/ring_buffer.c
2959
!cpumask_test_cpu(cpu_id, buffer->cpumask))
kernel/trace/ring_buffer.c
314
for_each_cpu(cpu, buffer->cpumask)
kernel/trace/ring_buffer.c
317
for_each_cpu_and(cpu, buffer->cpumask, cpu_online_mask)
kernel/trace/ring_buffer.c
4716
if (unlikely(!cpumask_test_cpu(cpu, buffer->cpumask)))
kernel/trace/ring_buffer.c
4863
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5048
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5068
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5088
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5119
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5138
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5158
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5181
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5203
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5223
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
576
cpumask_var_t cpumask;
kernel/trace/ring_buffer.c
5874
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5956
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6001
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6078
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6242
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6383
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6416
if (!cpumask_test_cpu(cpu, buffer_a->cpumask) ||
kernel/trace/ring_buffer.c
6417
!cpumask_test_cpu(cpu, buffer_b->cpumask))
kernel/trace/ring_buffer.c
6509
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6636
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6894
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6932
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7081
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7257
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7321
if (WARN_ON(!cpumask_test_cpu(cpu, buffer->cpumask)))
kernel/trace/ring_buffer.c
7339
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7484
if (cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7510
cpumask_set_cpu(cpu, buffer->cpumask);
kernel/trace/trace_events_filter.c
653
do_filter_cpumask(int op, const struct cpumask *mask, const struct cpumask *cmp)
kernel/trace/trace_events_filter.c
669
do_filter_scalar_cpumask(int op, unsigned int cpu, const struct cpumask *mask)
kernel/trace/trace_events_filter.c
688
do_filter_cpumask_scalar(int op, const struct cpumask *mask, unsigned int cpu)
kernel/trace/trace_events_filter.c
81
struct cpumask *mask;
kernel/trace/trace_events_filter.c
974
const struct cpumask *mask = (event + loc);
kernel/trace/trace_events_filter.c
975
const struct cpumask *cmp = pred->mask;
kernel/trace/trace_events_filter.c
985
const struct cpumask *mask = (event + loc);
kernel/trace/trace_hwlat.c
309
static struct cpumask save_cpumask;
kernel/trace/trace_hwlat.c
313
struct cpumask *current_mask = &save_cpumask;
kernel/trace/trace_hwlat.c
418
struct cpumask *current_mask = &save_cpumask;
kernel/trace/trace_hwlat.c
577
struct cpumask *current_mask = &save_cpumask;
kernel/trace/trace_osnoise.c
1600
static struct cpumask osnoise_cpumask;
kernel/trace/trace_osnoise.c
1601
static struct cpumask save_cpumask;
kernel/trace/trace_osnoise.c
1602
static struct cpumask kthread_cpumask;
kernel/trace/trace_osnoise.c
2027
struct cpumask *current_mask = &save_cpumask;
kernel/up.c
44
void *info, bool wait, const struct cpumask *mask)
kernel/watchdog.c
362
static struct cpumask watchdog_allowed_mask __read_mostly;
kernel/watchdog.c
54
struct cpumask watchdog_cpumask __read_mostly;
kernel/workqueue.c
1594
struct cpumask *effective = unbound_effective_cpumask(wq);
kernel/workqueue.c
2676
return pool->attrs->cpumask;
kernel/workqueue.c
467
static struct cpumask wq_cmdline_cpumask __initdata;
kernel/workqueue.c
4700
free_cpumask_var(attrs->cpumask);
kernel/workqueue.c
4721
if (!alloc_cpumask_var(&attrs->cpumask, GFP_KERNEL))
kernel/workqueue.c
4726
cpumask_copy(attrs->cpumask, cpu_possible_mask);
kernel/workqueue.c
4738
cpumask_copy(to->cpumask, from->cpumask);
kernel/workqueue.c
4760
cpumask_copy(attrs->cpumask, cpu_possible_mask);
kernel/workqueue.c
4773
hash = jhash(cpumask_bits(attrs->cpumask),
kernel/workqueue.c
4788
if (!a->affn_strict && !cpumask_equal(a->cpumask, b->cpumask))
kernel/workqueue.c
4802
cpumask_and(attrs->cpumask, attrs->cpumask, unbound_cpumask);
kernel/workqueue.c
4803
if (unlikely(cpumask_empty(attrs->cpumask)))
kernel/workqueue.c
4804
cpumask_copy(attrs->cpumask, unbound_cpumask);
kernel/workqueue.c
5307
cpumask_and(attrs->__pod_cpumask, pt->pod_cpus[pod], attrs->cpumask);
kernel/workqueue.c
5310
cpumask_copy(attrs->__pod_cpumask, attrs->cpumask);
kernel/workqueue.c
5387
cpumask_copy(new_attrs->__pod_cpumask, new_attrs->cpumask);
kernel/workqueue.c
5406
cpumask_and(new_attrs->cpumask, new_attrs->cpumask, cpu_possible_mask);
kernel/workqueue.c
5407
cpumask_copy(new_attrs->__pod_cpumask, new_attrs->cpumask);
kernel/workqueue.c
6263
pr_cont(" cpus=%*pbl", nr_cpumask_bits, pool->attrs->cpumask);
kernel/workqueue.c
6717
static cpumask_t cpumask;
kernel/workqueue.c
6723
if (!cpumask_test_cpu(cpu, pool->attrs->cpumask))
kernel/workqueue.c
6726
cpumask_and(&cpumask, pool->attrs->cpumask, cpu_online_mask);
kernel/workqueue.c
6730
WARN_ON_ONCE(set_cpus_allowed_ptr(worker->task, &cpumask) < 0);
kernel/workqueue.c
7037
int workqueue_unbound_housekeeping_update(const struct cpumask *hk)
kernel/workqueue.c
7039
cpumask_var_t cpumask;
kernel/workqueue.c
7042
if (!zalloc_cpumask_var(&cpumask, GFP_KERNEL))
kernel/workqueue.c
7053
if (!cpumask_and(cpumask, wq_requested_unbound_cpumask, hk))
kernel/workqueue.c
7054
cpumask_copy(cpumask, wq_requested_unbound_cpumask);
kernel/workqueue.c
7055
if (!cpumask_equal(cpumask, wq_unbound_cpumask))
kernel/workqueue.c
7056
ret = workqueue_apply_unbound_cpumask(cpumask);
kernel/workqueue.c
7063
free_cpumask_var(cpumask);
kernel/workqueue.c
7245
cpumask_pr_args(wq->unbound_attrs->cpumask));
kernel/workqueue.c
7264
ret = cpumask_parse(buf, attrs->cpumask);
kernel/workqueue.c
7349
__ATTR(cpumask, 0644, wq_cpumask_show, wq_cpumask_store),
kernel/workqueue.c
7372
static int workqueue_set_unbound_cpumask(cpumask_var_t cpumask)
kernel/workqueue.c
7380
cpumask_and(cpumask, cpumask, cpu_possible_mask);
kernel/workqueue.c
7381
if (!cpumask_empty(cpumask)) {
kernel/workqueue.c
7384
if (!cpumask_equal(cpumask, wq_unbound_cpumask))
kernel/workqueue.c
7385
ret = workqueue_apply_unbound_cpumask(cpumask);
kernel/workqueue.c
7387
cpumask_copy(wq_requested_unbound_cpumask, cpumask);
kernel/workqueue.c
7429
cpumask_var_t cpumask;
kernel/workqueue.c
7432
if (!zalloc_cpumask_var(&cpumask, GFP_KERNEL))
kernel/workqueue.c
7435
ret = cpumask_parse(buf, cpumask);
kernel/workqueue.c
7437
ret = workqueue_set_unbound_cpumask(cpumask);
kernel/workqueue.c
7439
free_cpumask_var(cpumask);
kernel/workqueue.c
7442
static DEVICE_ATTR_RW(cpumask);
kernel/workqueue.c
749
static struct cpumask *unbound_effective_cpumask(struct workqueue_struct *wq)
kernel/workqueue.c
7827
static void __init restrict_unbound_cpumask(const char *name, const struct cpumask *mask)
kernel/workqueue.c
7842
cpumask_copy(pool->attrs->cpumask, cpumask_of(cpu));
lib/cpu_rmap.c
168
const struct cpumask *affinity)
lib/cpu_rmap.c
95
const struct cpumask *mask, u16 dist)
lib/cpumask.c
134
unsigned int cpumask_any_and_distribute(const struct cpumask *src1p,
lib/cpumask.c
135
const struct cpumask *src2p)
lib/cpumask.c
156
unsigned int cpumask_any_distribute(const struct cpumask *srcp)
lib/group_cpus.c
14
static void grp_spread_init_one(struct cpumask *irqmsk, struct cpumask *nmsk,
lib/group_cpus.c
17
const struct cpumask *siblmsk;
lib/group_cpus.c
230
const struct cpumask *cpu_mask,
lib/group_cpus.c
232
struct cpumask *nmsk,
lib/group_cpus.c
259
struct cpumask *nmsk,
lib/group_cpus.c
261
struct cpumask *masks,
lib/group_cpus.c
291
struct cpumask *node_cpumask,
lib/group_cpus.c
293
const struct cpumask ***clusters_ptr,
lib/group_cpus.c
298
const struct cpumask *cluster_mask;
lib/group_cpus.c
299
const struct cpumask **clusters;
lib/group_cpus.c
361
struct cpumask *node_cpumask,
lib/group_cpus.c
362
struct cpumask *masks,
lib/group_cpus.c
367
const struct cpumask **clusters;
lib/group_cpus.c
406
const struct cpumask *cpu_mask,
lib/group_cpus.c
407
struct cpumask *nmsk, struct cpumask *masks)
lib/group_cpus.c
488
struct cpumask *group_cpus_evenly(unsigned int numgrps, unsigned int *nummasks)
lib/group_cpus.c
494
struct cpumask *masks = NULL;
lib/group_cpus.c
568
struct cpumask *group_cpus_evenly(unsigned int numgrps, unsigned int *nummasks)
lib/group_cpus.c
570
struct cpumask *masks;
lib/group_cpus.c
86
const struct cpumask *mask, nodemask_t *nodemsk)
mm/mm_init.c
2106
deferred_page_init_max_threads(const struct cpumask *node_cpumask)
mm/mm_init.c
2115
const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
mm/mm_init.c
2116
int max_threads = deferred_page_init_max_threads(cpumask);
mm/mm_init.c
2122
if (!cpumask_empty(cpumask))
mm/mm_init.c
2123
set_cpus_allowed_ptr(current, cpumask);
mm/percpu.c
2812
static struct cpumask mask __initdata;
mm/swap.c
810
static struct cpumask has_work;
mm/vmstat.c
2206
const struct cpumask *node_cpus;
net/core/dev.c
3053
int netif_set_xps_queue(struct net_device *dev, const struct cpumask *mask,
net/core/dev.h
14
struct cpumask;
net/core/dev.h
347
int rps_cpumask_housekeeping(struct cpumask *mask);
net/core/net-sysfs.c
1022
int rps_cpumask_housekeeping(struct cpumask *mask)
net/core/net-sysfs.c
1215
struct cpumask *rps_default_mask;
net/core/sysctl_net_core.c
107
struct cpumask *mask;
net/core/sysctl_net_core.c
58
struct cpumask *mask)
net/iucv/iucv.c
550
static cpumask_t cpumask;
net/iucv/iucv.c
554
cpumask_copy(&cpumask, &iucv_irq_cpumask);
net/iucv/iucv.c
555
cpumask_clear_cpu(cpumask_first(&iucv_irq_cpumask), &cpumask);
net/iucv/iucv.c
556
for_each_cpu(cpu, &cpumask)
net/iucv/iucv.c
656
cpumask_var_t cpumask;
net/iucv/iucv.c
662
if (!alloc_cpumask_var(&cpumask, GFP_KERNEL))
net/iucv/iucv.c
665
cpumask_copy(cpumask, &iucv_buffer_cpumask);
net/iucv/iucv.c
666
cpumask_clear_cpu(cpu, cpumask);
net/iucv/iucv.c
667
if (cpumask_empty(cpumask)) {
net/iucv/iucv.c
681
free_cpumask_var(cpumask);
net/netfilter/ipvs/ip_vs_ctl.c
1971
struct cpumask *mask;
net/netfilter/ipvs/ip_vs_ctl.c
1979
mask = (struct cpumask *)housekeeping_cpumask(HK_TYPE_KTHREAD);
net/openvswitch/flow.h
232
struct cpumask *cpu_used_mask;
net/openvswitch/flow_table.c
82
flow->cpu_used_mask = (struct cpumask *)&flow->stats[nr_cpu_ids];
rust/helpers/cpumask.c
12
void rust_helper___cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp)
rust/helpers/cpumask.c
18
void rust_helper_cpumask_clear_cpu(int cpu, struct cpumask *dstp)
rust/helpers/cpumask.c
24
void rust_helper___cpumask_clear_cpu(int cpu, struct cpumask *dstp)
rust/helpers/cpumask.c
30
bool rust_helper_cpumask_test_cpu(int cpu, struct cpumask *srcp)
rust/helpers/cpumask.c
36
void rust_helper_cpumask_setall(struct cpumask *dstp)
rust/helpers/cpumask.c
42
bool rust_helper_cpumask_empty(struct cpumask *srcp)
rust/helpers/cpumask.c
48
bool rust_helper_cpumask_full(struct cpumask *srcp)
rust/helpers/cpumask.c
54
unsigned int rust_helper_cpumask_weight(struct cpumask *srcp)
rust/helpers/cpumask.c
6
void rust_helper_cpumask_set_cpu(unsigned int cpu, struct cpumask *dstp)
rust/helpers/cpumask.c
60
void rust_helper_cpumask_copy(struct cpumask *dstp, const struct cpumask *srcp)
samples/bpf/cpustat_user.c
109
cpu_set_t cpumask;
samples/bpf/cpustat_user.c
133
CPU_ZERO(&cpumask);
samples/bpf/cpustat_user.c
134
CPU_SET(i, &cpumask);
samples/bpf/cpustat_user.c
136
sched_setaffinity(0, sizeof(cpumask), &cpumask);
samples/trace_events/trace-events-sample.h
294
const char *string, const struct cpumask *mask,
samples/trace_events/trace-events-sample.h
583
__rel_cpumask( cpumask )
samples/trace_events/trace-events-sample.h
591
__assign_rel_cpumask(cpumask, cpus);
samples/trace_events/trace-events-sample.h
596
__get_rel_cpumask(cpumask))
tools/accounting/getdelays.c
507
strncpy(cpumask, optarg, sizeof(cpumask));
tools/accounting/getdelays.c
508
cpumask[sizeof(cpumask) - 1] = '\0';
tools/accounting/getdelays.c
510
printf("cpumask %s maskset %d\n", cpumask, maskset);
tools/accounting/getdelays.c
585
&cpumask, strlen(cpumask) + 1);
tools/accounting/getdelays.c
727
&cpumask, strlen(cpumask) + 1);
tools/accounting/getdelays.c
73
char cpumask[100+6*MAX_CPUS];
tools/accounting/procacct.c
298
strncpy(cpumask, optarg, sizeof(cpumask));
tools/accounting/procacct.c
299
cpumask[sizeof(cpumask) - 1] = '\0';
tools/accounting/procacct.c
313
strncpy(cpumask, "1", sizeof(cpumask));
tools/accounting/procacct.c
314
cpumask[sizeof(cpumask) - 1] = '\0';
tools/accounting/procacct.c
316
printf("cpumask %s maskset %d\n", cpumask, maskset);
tools/accounting/procacct.c
341
&cpumask, strlen(cpumask) + 1);
tools/accounting/procacct.c
399
&cpumask, strlen(cpumask) + 1);
tools/accounting/procacct.c
84
char cpumask[100+6*MAX_CPUS];
tools/perf/bench/numa.c
254
struct bitmask *cpumask = numa_allocate_cpumask();
tools/perf/bench/numa.c
258
BUG_ON(!cpumask);
tools/perf/bench/numa.c
259
if (!numa_node_to_cpus(node, cpumask)) {
tools/perf/bench/numa.c
260
for (cpu = 0; cpu < (int)cpumask->size; cpu++) {
tools/perf/bench/numa.c
261
if (numa_bitmask_isbitset(cpumask, cpu)) {
tools/perf/bench/numa.c
267
numa_free_cpumask(cpumask);
tools/perf/bench/numa.c
344
struct bitmask *cpumask = numa_allocate_cpumask();
tools/perf/bench/numa.c
346
if (!cpumask)
tools/perf/bench/numa.c
349
if (!numa_node_to_cpus(target_node, cpumask)) {
tools/perf/bench/numa.c
350
for (cpu = 0; cpu < (int)cpumask->size; cpu++) {
tools/perf/bench/numa.c
351
if (numa_bitmask_isbitset(cpumask, cpu))
tools/perf/bench/numa.c
355
numa_free_cpumask(cpumask);
tools/perf/builtin-ftrace.c
390
char *cpumask;
tools/perf/builtin-ftrace.c
399
cpumask = malloc(mask_size);
tools/perf/builtin-ftrace.c
400
if (cpumask == NULL) {
tools/perf/builtin-ftrace.c
405
cpu_map__snprint_mask(cpumap, cpumask, mask_size);
tools/perf/builtin-ftrace.c
407
ret = write_tracing_file("tracing_cpumask", cpumask);
tools/perf/builtin-ftrace.c
409
free(cpumask);
tools/perf/util/env.c
237
zfree(&d_info->cpumask);
tools/perf/util/env.h
60
char *cpumask;
tools/perf/util/header.c
1635
char dname[DNAME_LEN], cpumask[MAX_NR_CPUS];
tools/perf/util/header.c
1692
cpumask);
tools/perf/util/header.c
1700
retval = sscanf(line, "domain%u %s %*s", &domain, cpumask);
tools/perf/util/header.c
1709
domain_info->cpumask = strdup(cpumask);
tools/perf/util/header.c
1710
if (!domain_info->cpumask)
tools/perf/util/header.c
1713
cpumask_to_cpulist(cpumask, cpulist);
tools/perf/util/header.c
1773
ret = do_write_string(ff, cd_map[i]->domains[j]->cpumask);
tools/perf/util/header.c
2454
fprintf(fp, "# Domain cpu map : %s\n", d_info->cpumask);
tools/perf/util/header.c
3616
char *dname, *cpumask, *cpulist;
tools/perf/util/header.c
3682
cpumask = do_read_string(ff);
tools/perf/util/header.c
3683
if (!cpumask)
tools/perf/util/header.c
3686
d_info->cpumask = cpumask;
tools/perf/util/pmu.c
821
char *cpumask = NULL;
tools/perf/util/pmu.c
831
ret = io__getline(&io, &cpumask, &cpumask_len);
tools/perf/util/pmu.c
836
cpus = perf_cpu_map__new(cpumask);
tools/perf/util/pmu.c
837
free(cpumask);
tools/perf/util/util.c
261
void cpumask_to_cpulist(char *cpumask, char *cpulist)
tools/perf/util/util.c
264
int len = strlen(cpumask);
tools/perf/util/util.c
269
if (cpumask[i] == ',') {
tools/perf/util/util.c
271
cpumask[j] = cpumask[j + 1];
tools/perf/util/util.c
275
len = strlen(cpumask);
tools/perf/util/util.c
289
strncpy(blk, cpumask + len - blklen, blklen);
tools/perf/util/util.c
292
cpumask[len - blklen] = '\0';
tools/perf/util/util.c
293
len = strlen(cpumask);
tools/perf/util/util.h
52
void cpumask_to_cpulist(char *cpumask, char *cpulist);
tools/sched_ext/include/scx/common.bpf.h
354
struct bpf_cpumask *bpf_cpumask_acquire(struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
355
void bpf_cpumask_release(struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
356
u32 bpf_cpumask_first(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
357
u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
358
void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
359
void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
360
bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
361
bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
362
bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
363
void bpf_cpumask_setall(struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
364
void bpf_cpumask_clear(struct bpf_cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
365
bool bpf_cpumask_and(struct bpf_cpumask *dst, const struct cpumask *src1,
tools/sched_ext/include/scx/common.bpf.h
366
const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
367
void bpf_cpumask_or(struct bpf_cpumask *dst, const struct cpumask *src1,
tools/sched_ext/include/scx/common.bpf.h
368
const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
369
void bpf_cpumask_xor(struct bpf_cpumask *dst, const struct cpumask *src1,
tools/sched_ext/include/scx/common.bpf.h
370
const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
371
bool bpf_cpumask_equal(const struct cpumask *src1, const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
372
bool bpf_cpumask_intersects(const struct cpumask *src1, const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
373
bool bpf_cpumask_subset(const struct cpumask *src1, const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
374
bool bpf_cpumask_empty(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
375
bool bpf_cpumask_full(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
376
void bpf_cpumask_copy(struct bpf_cpumask *dst, const struct cpumask *src) __ksym;
tools/sched_ext/include/scx/common.bpf.h
377
u32 bpf_cpumask_any_distribute(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
378
u32 bpf_cpumask_any_and_distribute(const struct cpumask *src1,
tools/sched_ext/include/scx/common.bpf.h
379
const struct cpumask *src2) __ksym;
tools/sched_ext/include/scx/common.bpf.h
380
u32 bpf_cpumask_weight(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/common.bpf.h
389
const struct cpumask *bitmap; \
tools/sched_ext/include/scx/common.bpf.h
398
sizeof(struct cpumask) / 8); \
tools/sched_ext/include/scx/common.bpf.h
445
static __always_inline const struct cpumask *cast_mask(struct bpf_cpumask *mask)
tools/sched_ext/include/scx/common.bpf.h
447
return (const struct cpumask *)mask;
tools/sched_ext/include/scx/common.bpf.h
63
s32 __scx_bpf_select_cpu_and(struct task_struct *p, const struct cpumask *cpus_allowed,
tools/sched_ext/include/scx/common.bpf.h
84
const struct cpumask *scx_bpf_get_possible_cpumask(void) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
85
const struct cpumask *scx_bpf_get_online_cpumask(void) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
86
void scx_bpf_put_cpumask(const struct cpumask *cpumask) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
87
const struct cpumask *scx_bpf_get_idle_cpumask_node(int node) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
88
const struct cpumask *scx_bpf_get_idle_cpumask(void) __ksym;
tools/sched_ext/include/scx/common.bpf.h
89
const struct cpumask *scx_bpf_get_idle_smtmask_node(int node) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
90
const struct cpumask *scx_bpf_get_idle_smtmask(void) __ksym;
tools/sched_ext/include/scx/common.bpf.h
91
void scx_bpf_put_idle_cpumask(const struct cpumask *cpumask) __ksym;
tools/sched_ext/include/scx/compat.bpf.h
236
const struct cpumask *cpus_allowed, u64 flags) __ksym __weak;
tools/sched_ext/include/scx/compat.bpf.h
253
const struct cpumask *cpus_allowed, u64 flags)
tools/sched_ext/include/scx/compat.bpf.h
82
int bpf_cpumask_populate(struct cpumask *dst, void *src, size_t src__sz) __ksym __weak;
tools/sched_ext/include/scx/compat.bpf.h
84
#define __COMPAT_bpf_cpumask_populate(cpumask, src, size__sz) \
tools/sched_ext/include/scx/compat.bpf.h
86
(bpf_cpumask_populate(cpumask, src, size__sz)) : -EOPNOTSUPP)
tools/sched_ext/scx_qmap.bpf.c
678
const struct cpumask *possible, *online;
tools/sched_ext/scx_qmap.bpf.c
749
const struct cpumask *online;
tools/testing/selftests/bpf/bpf_arena_alloc.h
14
#define NR_CPUS (sizeof(struct cpumask) * 8)
tools/testing/selftests/bpf/progs/cpumask_common.h
107
local.cpumask = NULL;
tools/testing/selftests/bpf/progs/cpumask_common.h
120
old = bpf_kptr_xchg(&v->cpumask, mask);
tools/testing/selftests/bpf/progs/cpumask_common.h
21
struct bpf_cpumask __kptr * cpumask;
tools/testing/selftests/bpf/progs/cpumask_common.h
32
void bpf_cpumask_release(struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
33
struct bpf_cpumask *bpf_cpumask_acquire(struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
34
u32 bpf_cpumask_first(const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
35
u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
36
u32 bpf_cpumask_first_and(const struct cpumask *src1,
tools/testing/selftests/bpf/progs/cpumask_common.h
37
const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
38
void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
39
void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
40
bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
41
bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
42
bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
43
void bpf_cpumask_setall(struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
44
void bpf_cpumask_clear(struct bpf_cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
45
bool bpf_cpumask_and(struct bpf_cpumask *cpumask,
tools/testing/selftests/bpf/progs/cpumask_common.h
46
const struct cpumask *src1,
tools/testing/selftests/bpf/progs/cpumask_common.h
47
const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
48
void bpf_cpumask_or(struct bpf_cpumask *cpumask,
tools/testing/selftests/bpf/progs/cpumask_common.h
49
const struct cpumask *src1,
tools/testing/selftests/bpf/progs/cpumask_common.h
50
const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
51
void bpf_cpumask_xor(struct bpf_cpumask *cpumask,
tools/testing/selftests/bpf/progs/cpumask_common.h
52
const struct cpumask *src1,
tools/testing/selftests/bpf/progs/cpumask_common.h
53
const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
54
bool bpf_cpumask_equal(const struct cpumask *src1, const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
55
bool bpf_cpumask_intersects(const struct cpumask *src1, const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
56
bool bpf_cpumask_subset(const struct cpumask *src1, const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
57
bool bpf_cpumask_empty(const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
58
bool bpf_cpumask_full(const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
59
void bpf_cpumask_copy(struct bpf_cpumask *dst, const struct cpumask *src) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
60
u32 bpf_cpumask_any_distribute(const struct cpumask *src) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
61
u32 bpf_cpumask_any_and_distribute(const struct cpumask *src1,
tools/testing/selftests/bpf/progs/cpumask_common.h
62
const struct cpumask *src2) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
63
u32 bpf_cpumask_weight(const struct cpumask *cpumask) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
64
int bpf_cpumask_populate(struct cpumask *cpumask, void *src, size_t src__sz) __ksym __weak;
tools/testing/selftests/bpf/progs/cpumask_common.h
69
static inline const struct cpumask *cast(struct bpf_cpumask *cpumask)
tools/testing/selftests/bpf/progs/cpumask_common.h
71
return (const struct cpumask *)cpumask;
tools/testing/selftests/bpf/progs/cpumask_common.h
76
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_common.h
78
cpumask = bpf_cpumask_create();
tools/testing/selftests/bpf/progs/cpumask_common.h
79
if (!cpumask) {
tools/testing/selftests/bpf/progs/cpumask_common.h
84
if (!bpf_cpumask_empty(cast(cpumask))) {
tools/testing/selftests/bpf/progs/cpumask_common.h
86
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_common.h
90
return cpumask;
tools/testing/selftests/bpf/progs/cpumask_failure.c
103
cpumask = bpf_kptr_xchg(&v->cpumask, NULL);
tools/testing/selftests/bpf/progs/cpumask_failure.c
148
bpf_cpumask_test_cpu(0, (const struct cpumask *)local);
tools/testing/selftests/bpf/progs/cpumask_failure.c
174
bpf_cpumask_test_cpu(0, (const struct cpumask *)local);
tools/testing/selftests/bpf/progs/cpumask_failure.c
234
ret = bpf_cpumask_populate((struct cpumask *)invalid, &bits, sizeof(bits));
tools/testing/selftests/bpf/progs/cpumask_failure.c
255
ret = bpf_cpumask_populate((struct cpumask *)local, garbage, 8);
tools/testing/selftests/bpf/progs/cpumask_failure.c
38
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_failure.c
40
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_failure.c
41
__sink(cpumask);
tools/testing/selftests/bpf/progs/cpumask_failure.c
51
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_failure.c
53
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_failure.c
56
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_failure.c
57
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_failure.c
66
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_failure.c
69
cpumask = bpf_cpumask_acquire((struct bpf_cpumask *)task->cpus_ptr);
tools/testing/selftests/bpf/progs/cpumask_failure.c
70
__sink(cpumask);
tools/testing/selftests/bpf/progs/cpumask_failure.c
89
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_failure.c
92
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_failure.c
93
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_failure.c
96
if (cpumask_map_insert(cpumask))
tools/testing/selftests/bpf/progs/cpumask_success.c
141
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
146
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
147
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
150
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
157
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
162
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
163
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
166
bpf_cpumask_set_cpu(0, cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
167
if (!bpf_cpumask_test_cpu(0, cast(cpumask))) {
tools/testing/selftests/bpf/progs/cpumask_success.c
172
bpf_cpumask_clear_cpu(0, cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
173
if (bpf_cpumask_test_cpu(0, cast(cpumask))) {
tools/testing/selftests/bpf/progs/cpumask_success.c
179
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
186
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
191
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
192
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
195
bpf_cpumask_setall(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
196
if (!bpf_cpumask_full(cast(cpumask))) {
tools/testing/selftests/bpf/progs/cpumask_success.c
201
bpf_cpumask_clear(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
202
if (!bpf_cpumask_empty(cast(cpumask))) {
tools/testing/selftests/bpf/progs/cpumask_success.c
208
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
215
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
220
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
221
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
224
if (bpf_cpumask_first(cast(cpumask)) < nr_cpus) {
tools/testing/selftests/bpf/progs/cpumask_success.c
229
if (bpf_cpumask_first_zero(cast(cpumask)) != 0) {
tools/testing/selftests/bpf/progs/cpumask_success.c
230
bpf_printk("first zero: %d", bpf_cpumask_first_zero(cast(cpumask)));
tools/testing/selftests/bpf/progs/cpumask_success.c
235
bpf_cpumask_set_cpu(0, cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
236
if (bpf_cpumask_first(cast(cpumask)) != 0) {
tools/testing/selftests/bpf/progs/cpumask_success.c
241
if (bpf_cpumask_first_zero(cast(cpumask)) != 1) {
tools/testing/selftests/bpf/progs/cpumask_success.c
247
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
286
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
291
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
292
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
295
if (bpf_cpumask_test_and_set_cpu(0, cpumask)) {
tools/testing/selftests/bpf/progs/cpumask_success.c
300
if (!bpf_cpumask_test_and_set_cpu(0, cpumask)) {
tools/testing/selftests/bpf/progs/cpumask_success.c
305
if (!bpf_cpumask_test_and_clear_cpu(0, cpumask)) {
tools/testing/selftests/bpf/progs/cpumask_success.c
311
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
461
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
463
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
464
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
467
if (cpumask_map_insert(cpumask))
tools/testing/selftests/bpf/progs/cpumask_success.c
476
struct bpf_cpumask *cpumask;
tools/testing/selftests/bpf/progs/cpumask_success.c
479
cpumask = create_cpumask();
tools/testing/selftests/bpf/progs/cpumask_success.c
480
if (!cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
483
if (cpumask_map_insert(cpumask)) {
tools/testing/selftests/bpf/progs/cpumask_success.c
494
cpumask = bpf_kptr_xchg(&v->cpumask, NULL);
tools/testing/selftests/bpf/progs/cpumask_success.c
495
if (cpumask)
tools/testing/selftests/bpf/progs/cpumask_success.c
496
bpf_cpumask_release(cpumask);
tools/testing/selftests/bpf/progs/cpumask_success.c
530
bpf_cpumask_test_cpu(0, (const struct cpumask *)local);
tools/testing/selftests/bpf/progs/cpumask_success.c
762
bpf_cpumask_test_cpu(0, (const struct cpumask *)mask1);
tools/testing/selftests/bpf/progs/cpumask_success.c
763
bpf_cpumask_test_cpu(0, (const struct cpumask *)mask2);
tools/testing/selftests/bpf/progs/cpumask_success.c
788
ret = bpf_cpumask_populate((struct cpumask *)local, &toofewbits, sizeof(toofewbits));
tools/testing/selftests/bpf/progs/cpumask_success.c
827
ret = bpf_cpumask_populate((struct cpumask *)mask, src, CPUMASK_TEST_MASKLEN);
tools/testing/selftests/bpf/progs/cpumask_success.c
858
ret = bpf_cpumask_populate((struct cpumask *)mask, bits, CPUMASK_TEST_MASKLEN);
tools/testing/selftests/bpf/progs/cpumask_success.c
876
bit = bpf_cpumask_test_cpu(i, (const struct cpumask *)mask);
tools/testing/selftests/bpf/progs/nested_trust_common.h
10
__u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) __ksym;
tools/testing/selftests/bpf/progs/nested_trust_common.h
9
bool bpf_cpumask_test_cpu(unsigned int cpu, const struct cpumask *cpumask) __ksym;
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
134
bpf_cpumask_test_cpu(0, (const struct cpumask *)ref);
tools/testing/selftests/net/bench/page_pool/time_bench.c
293
const struct cpumask *mask)
tools/testing/selftests/net/bench/page_pool/time_bench.c
330
const struct cpumask *mask, /* Support masking outsome CPUs*/
tools/testing/selftests/net/bench/page_pool/time_bench.h
206
const struct cpumask *mask, /* Support masking outsome CPUs*/
tools/testing/selftests/net/bench/page_pool/time_bench.h
211
const struct cpumask *mask);
tools/testing/selftests/sched_ext/allowed_cpus.bpf.c
18
validate_idle_cpu(const struct task_struct *p, const struct cpumask *allowed, s32 cpu)
tools/testing/selftests/sched_ext/allowed_cpus.bpf.c
32
const struct cpumask *allowed;
tools/testing/selftests/sched_ext/allowed_cpus.bpf.c
57
const struct cpumask *allowed;
tools/testing/selftests/sched_ext/allowed_cpus.bpf.c
98
const struct cpumask *online = scx_bpf_get_online_cpumask();
tools/testing/selftests/sched_ext/maximal.bpf.c
64
const struct cpumask *cpumask)
tools/testing/selftests/sched_ext/numa.bpf.c
24
const struct cpumask *idle_cpumask;
tools/testing/selftests/sched_ext/select_cpu_dfl.bpf.c
25
const struct cpumask *idle_mask = scx_bpf_get_idle_cpumask();
virt/kvm/kvm_main.c
208
static inline bool kvm_kick_many_cpus(struct cpumask *cpus, bool wait)
virt/kvm/kvm_main.c
218
struct cpumask *tmp, int current_cpu)
virt/kvm/kvm_main.c
249
struct cpumask *cpus;
virt/kvm/kvm_main.c
274
struct cpumask *cpus;