Symbol: cpumask_t
arch/alpha/kernel/smp.c
574
cpumask_t to_whom;
arch/alpha/kernel/sys_dp264.c
136
cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity)
arch/alpha/kernel/sys_titan.c
135
titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity)
arch/alpha/kernel/sys_titan.c
67
cpumask_t cpm;
arch/arc/kernel/mcip.c
268
cpumask_t online;
arch/arm/common/bL_switcher.c
380
static cpumask_t bL_switcher_removed_logical_cpus;
arch/arm/common/bL_switcher.c
421
cpumask_t available_cpus;
arch/arm/include/asm/irq.h
34
extern void arch_trigger_cpumask_backtrace(const cpumask_t *mask,
arch/arm/include/asm/mmu_context.h
50
cpumask_t *mask);
arch/arm/include/asm/mmu_context.h
53
cpumask_t *mask)
arch/arm/kernel/hw_breakpoint.c
999
static cpumask_t debug_err_mask;
arch/arm/kernel/smp.c
845
static void raise_nmi(cpumask_t *mask)
arch/arm/kernel/smp.c
850
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/arm/kernel/smp_tlb.c
167
cpumask_t mask = { CPU_BITS_NONE };
arch/arm/mach-bcm/platsmp.c
130
const cpumask_t only_cpu_0 = { CPU_BITS_CPU0 };
arch/arm/mach-imx/mmdc.c
100
cpumask_t cpu;
arch/arm/mach-tegra/platsmp.c
34
static cpumask_t tegra_cpu_init_mask;
arch/arm/mm/cache-l2x0-pmu.c
22
static cpumask_t pmu_cpu;
arch/arm/mm/context.c
48
static cpumask_t tlb_flush_pending;
arch/arm/mm/context.c
52
cpumask_t *mask)
arch/arm64/include/asm/irq.h
11
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu);
arch/arm64/kernel/smp.c
1188
static cpumask_t mask;
arch/arm64/kernel/smp.c
912
static void arm64_send_ipi(const cpumask_t *mask, unsigned int nr)
arch/arm64/kernel/smp.c
923
static void arm64_backtrace_ipi(cpumask_t *mask)
arch/arm64/kernel/smp.c
928
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/arm64/mm/context.c
28
static cpumask_t tlb_flush_pending;
arch/csky/abiv2/cacheflush.c
42
cpumask_t *mask = &mm->context.icache_stale_mask;
arch/csky/abiv2/cacheflush.c
59
cpumask_t others, *mask;
arch/csky/include/asm/asid.h
21
cpumask_t flush_pending;
arch/csky/include/asm/mmu.h
9
cpumask_t icache_stale_mask;
arch/loongarch/include/asm/smp.h
27
extern cpumask_t cpu_sibling_map[];
arch/loongarch/include/asm/smp.h
28
extern cpumask_t cpu_llc_shared_map[];
arch/loongarch/include/asm/smp.h
29
extern cpumask_t cpu_core_map[];
arch/loongarch/include/asm/smp.h
30
extern cpumask_t cpu_foreign_map[];
arch/loongarch/include/asm/topology.h
13
extern cpumask_t cpus_on_node[];
arch/loongarch/kernel/machine_kexec.c
29
static cpumask_t cpus_in_crash = CPU_MASK_NONE;
arch/loongarch/kernel/numa.c
31
cpumask_t cpus_on_node[MAX_NUMNODES];
arch/loongarch/kernel/numa.c
32
cpumask_t phys_cpus_on_node[MAX_NUMNODES];
arch/loongarch/kernel/process.c
356
static void raise_backtrace(cpumask_t *mask)
arch/loongarch/kernel/process.c
380
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/loongarch/kernel/smp.c
175
cpumask_t temp_foreign_map;
arch/loongarch/kernel/smp.c
46
cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
50
cpumask_t cpu_llc_shared_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
54
cpumask_t cpu_core_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
64
cpumask_t cpu_foreign_map[NR_CPUS] __read_mostly;
arch/loongarch/kernel/smp.c
68
static cpumask_t cpu_sibling_setup_map;
arch/loongarch/kernel/smp.c
71
static cpumask_t cpu_llc_shared_setup_map;
arch/loongarch/kernel/smp.c
74
static cpumask_t cpu_core_setup_map;
arch/mips/cavium-octeon/octeon-irq.c
770
cpumask_t new_affinity;
arch/mips/include/asm/bmips.h
87
extern cpumask_t bmips_booted_mask;
arch/mips/include/asm/kvm_host.h
179
cpumask_t asid_flush_mask;
arch/mips/include/asm/mach-ip27/mmzone.h
15
cpumask_t h_cpus;
arch/mips/include/asm/mach-loongson64/topology.h
9
extern cpumask_t __node_cpumask[];
arch/mips/include/asm/mips_mt.h
17
extern cpumask_t mt_fpu_cpumask;
arch/mips/include/asm/processor.h
252
cpumask_t user_cpus_allowed;
arch/mips/include/asm/smp.h
22
extern cpumask_t cpu_sibling_map[];
arch/mips/include/asm/smp.h
23
extern cpumask_t cpu_core_map[];
arch/mips/include/asm/smp.h
24
extern cpumask_t cpu_foreign_map[];
arch/mips/include/asm/smp.h
55
extern cpumask_t cpu_coherent_mask;
arch/mips/kernel/cacheinfo.c
58
static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map)
arch/mips/kernel/cacheinfo.c
67
static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map)
arch/mips/kernel/crash.c
16
static cpumask_t cpus_in_crash = CPU_MASK_NONE;
arch/mips/kernel/mips-mt-fpaff.c
162
cpumask_t allowed, mask;
arch/mips/kernel/mips-mt-fpaff.c
23
cpumask_t mt_fpu_cpumask;
arch/mips/kernel/pm-cps.c
107
cpumask_t *coupled_mask = this_cpu_ptr(&online_coupled);
arch/mips/kernel/pm-cps.c
58
static DEFINE_PER_CPU_ALIGNED(cpumask_t, online_coupled);
arch/mips/kernel/process.c
733
static void raise_backtrace(cpumask_t *mask)
arch/mips/kernel/process.c
756
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/mips/kernel/smp-bmips.c
47
cpumask_t bmips_booted_mask;
arch/mips/kernel/smp.c
142
cpumask_t temp_foreign_map;
arch/mips/kernel/smp.c
52
cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly;
arch/mips/kernel/smp.c
56
cpumask_t cpu_core_map[NR_CPUS] __read_mostly;
arch/mips/kernel/smp.c
68
cpumask_t cpu_foreign_map[NR_CPUS] __read_mostly;
arch/mips/kernel/smp.c
72
static cpumask_t cpu_sibling_setup_map;
arch/mips/kernel/smp.c
75
static cpumask_t cpu_core_setup_map;
arch/mips/kernel/smp.c
77
cpumask_t cpu_coherent_mask;
arch/mips/kernel/traps.c
946
cpumask_t tmask;
arch/mips/loongson64/numa.c
33
cpumask_t __node_cpumask[MAX_NUMNODES];
arch/mips/mm/c-octeon.c
68
cpumask_t mask;
arch/mips/mm/c-r4k.c
435
const cpumask_t *mask = cpu_present_mask;
arch/mips/mm/context.c
14
static cpumask_t tlb_flush_pending;
arch/parisc/kernel/irq.c
496
cpumask_t dest;
arch/powerpc/include/asm/cputhreads.h
26
extern cpumask_t threads_core_mask;
arch/powerpc/include/asm/irq.h
57
extern void arch_trigger_cpumask_backtrace(const cpumask_t *mask,
arch/powerpc/include/asm/kvm_book3s_64.h
33
cpumask_t need_tlb_flush;
arch/powerpc/include/asm/kvm_host.h
295
cpumask_t need_tlb_flush;
arch/powerpc/kernel/setup-common.c
385
cpumask_t threads_core_mask __read_mostly;
arch/powerpc/kernel/stacktrace.c
159
static void raise_backtrace_ipi(cpumask_t *mask)
arch/powerpc/kernel/stacktrace.c
209
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/powerpc/kernel/watchdog.c
203
static cpumask_t wd_smp_cpus_ipi; // protected by reporting
arch/powerpc/kernel/watchdog.c
77
static cpumask_t wd_cpus_enabled __read_mostly;
arch/powerpc/kernel/watchdog.c
91
static cpumask_t wd_smp_cpus_pending;
arch/powerpc/kernel/watchdog.c
92
static cpumask_t wd_smp_cpus_stuck;
arch/powerpc/kexec/crash.c
83
static cpumask_t cpus_state_saved = CPU_MASK_NONE;
arch/powerpc/kvm/book3s_hv.c
3243
cpumask_t *need_tlb_flush;
arch/powerpc/kvm/book3s_hv_p9_entry.c
456
cpumask_t *need_tlb_flush;
arch/powerpc/perf/hv-24x7.c
35
static cpumask_t hv_24x7_cpumask;
arch/powerpc/perf/hv-gpci.c
51
static cpumask_t hv_gpci_cpumask;
arch/powerpc/perf/imc-pmu.c
107
cpumask_t *active_mask;
arch/powerpc/perf/imc-pmu.c
28
static cpumask_t nest_imc_cpumask;
arch/powerpc/perf/imc-pmu.c
34
static cpumask_t core_imc_cpumask;
arch/powerpc/platforms/cell/spufs/spufs.h
111
cpumask_t cpus_allowed;
arch/powerpc/xmon/xmon.c
67
static cpumask_t cpus_in_xmon = CPU_MASK_NONE;
arch/powerpc/xmon/xmon.c
73
static cpumask_t xmon_batch_cpus = CPU_MASK_NONE;
arch/riscv/include/asm/irq.h
18
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu);
arch/riscv/include/asm/mmu.h
21
cpumask_t icache_stale_mask;
arch/riscv/include/asm/sbi.h
611
int sbi_fwft_set_cpumask(const cpumask_t *mask, u32 feature,
arch/riscv/kernel/sbi.c
351
int sbi_fwft_set_cpumask(const cpumask_t *mask, u32 feature,
arch/riscv/kernel/smp.c
259
cpumask_t mask;
arch/riscv/kernel/smp.c
294
cpumask_t mask;
arch/riscv/kernel/smp.c
343
static void riscv_backtrace_ipi(cpumask_t *mask)
arch/riscv/kernel/smp.c
348
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/riscv/kernel/sys_hwprobe.c
384
cpumask_t cpus;
arch/riscv/kernel/sys_hwprobe.c
439
cpumask_t cpus, one_cpu;
arch/riscv/kernel/unaligned_access_speed.c
183
static void modify_unaligned_access_branches(cpumask_t *mask, int weight)
arch/riscv/kernel/unaligned_access_speed.c
200
cpumask_t fast_except_me;
arch/riscv/kernel/unaligned_access_speed.c
219
cpumask_t fast_and_online;
arch/riscv/kernel/unaligned_access_speed.c
30
static cpumask_t fast_misaligned_access;
arch/riscv/mm/cacheflush.c
183
cpumask_t *mask;
arch/riscv/mm/cacheflush.c
60
cpumask_t others, *mask;
arch/riscv/mm/context.c
29
static cpumask_t context_tlb_flush_pending;
arch/s390/include/asm/mmu.h
11
cpumask_t cpu_attach_mask;
arch/s390/include/asm/smp.h
39
extern cpumask_t cpu_setup_mask;
arch/s390/include/asm/topology.h
21
cpumask_t thread_mask;
arch/s390/include/asm/topology.h
22
cpumask_t core_mask;
arch/s390/include/asm/topology.h
23
cpumask_t book_mask;
arch/s390/include/asm/topology.h
24
cpumask_t drawer_mask;
arch/s390/kernel/hiperdispatch.c
70
static cpumask_t hd_vl_coremask; /* Mask containing all vertical low COREs */
arch/s390/kernel/hiperdispatch.c
71
static cpumask_t hd_vmvl_cpumask; /* Mask containing vertical medium and low CPUs */
arch/s390/kernel/perf_cpum_cf.c
1250
cpumask_t mask; /* CPU mask to read from */
arch/s390/kernel/perf_cpum_cf.c
1474
static int cfset_all_copy(unsigned long arg, cpumask_t *mask)
arch/s390/kernel/smp.c
427
static cpumask_t cpumask;
arch/s390/kernel/smp.c
709
static int smp_add_core(struct sclp_core_entry *core, cpumask_t *avail,
arch/s390/kernel/smp.c
746
static cpumask_t avail;
arch/s390/kernel/smp.c
99
cpumask_t cpu_setup_mask;
arch/s390/kernel/topology.c
100
static cpumask_t mask;
arch/s390/kernel/topology.c
45
cpumask_t mask;
arch/s390/kernel/topology.c
67
static void cpu_group_map(cpumask_t *dst, struct mask_info *info, unsigned int cpu)
arch/s390/kernel/topology.c
69
static cpumask_t mask;
arch/s390/kernel/topology.c
98
static void cpu_thread_map(cpumask_t *dst, unsigned int cpu)
arch/sh/include/asm/topology.h
22
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sh/kernel/topology.c
18
cpumask_t cpu_core_map[NR_CPUS];
arch/sh/kernel/topology.c
21
static cpumask_t cpu_coregroup_map(int cpu)
arch/sparc/include/asm/mdesc.h
93
void mdesc_fill_in_cpu_data(cpumask_t *mask);
arch/sparc/include/asm/mdesc.h
94
void mdesc_populate_present_mask(cpumask_t *mask);
arch/sparc/include/asm/mdesc.h
95
void mdesc_get_page_sizes(cpumask_t *mask, unsigned long *pgsz_mask);
arch/sparc/include/asm/mmzone.h
10
extern cpumask_t numa_cpumask_lookup_table[];
arch/sparc/include/asm/smp_32.h
33
extern cpumask_t smp_commenced_mask;
arch/sparc/include/asm/smp_32.h
57
void (*cross_call)(void *func, cpumask_t mask, unsigned long arg1,
arch/sparc/include/asm/smp_64.h
34
DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
arch/sparc/include/asm/smp_64.h
35
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sparc/include/asm/topology_64.h
53
extern cpumask_t cpu_core_map[NR_CPUS];
arch/sparc/include/asm/topology_64.h
54
extern cpumask_t cpu_core_sib_map[NR_CPUS];
arch/sparc/include/asm/topology_64.h
55
extern cpumask_t cpu_core_sib_cache_map[NR_CPUS];
arch/sparc/kernel/ds.c
488
cpumask_t *mask, u32 default_stat)
arch/sparc/kernel/ds.c
534
u64 req_num, cpumask_t *mask)
arch/sparc/kernel/ds.c
593
cpumask_t *mask)
arch/sparc/kernel/ds.c
637
cpumask_t mask;
arch/sparc/kernel/leon_smp.c
375
static void leon_cross_call(void *func, cpumask_t mask, unsigned long arg1,
arch/sparc/kernel/leon_smp.c
55
extern cpumask_t smp_commenced_mask;
arch/sparc/kernel/mdesc.c
1107
static void *mdesc_iterate_over_cpus(void *(*func)(struct mdesc_handle *, u64, int, void *), void *arg, cpumask_t *mask)
arch/sparc/kernel/mdesc.c
1147
void mdesc_populate_present_mask(cpumask_t *mask)
arch/sparc/kernel/mdesc.c
1174
void __init mdesc_get_page_sizes(cpumask_t *mask, unsigned long *pgsz_mask)
arch/sparc/kernel/mdesc.c
1230
void mdesc_fill_in_cpu_data(cpumask_t *mask)
arch/sparc/kernel/process_64.c
239
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/sparc/kernel/smp_32.c
44
cpumask_t smp_commenced_mask = CPU_MASK_NONE;
arch/sparc/kernel/smp_64.c
61
DEFINE_PER_CPU(cpumask_t, cpu_sibling_map) = CPU_MASK_NONE;
arch/sparc/kernel/smp_64.c
62
cpumask_t cpu_core_map[NR_CPUS] __read_mostly =
arch/sparc/kernel/smp_64.c
65
cpumask_t cpu_core_sib_map[NR_CPUS] __read_mostly = {
arch/sparc/kernel/smp_64.c
68
cpumask_t cpu_core_sib_cache_map[NR_CPUS] __read_mostly = {
arch/sparc/kernel/smp_64.c
76
static cpumask_t smp_commenced_mask;
arch/sparc/kernel/smp_64.c
783
static void xcall_deliver(u64 data0, u64 data1, u64 data2, const cpumask_t *mask)
arch/sparc/kernel/smp_64.c
832
static void smp_cross_call_masked(unsigned long *func, u32 ctx, u64 data1, u64 data2, const cpumask_t *mask)
arch/sparc/kernel/sun4d_smp.c
284
static void sun4d_cross_call(void *func, cpumask_t mask, unsigned long arg1,
arch/sparc/kernel/sun4m_smp.c
173
static void sun4m_cross_call(void *func, cpumask_t mask, unsigned long arg1,
arch/sparc/mm/init_64.c
1313
u64 grp, cpumask_t *mask)
arch/sparc/mm/init_64.c
1428
cpumask_t mask;
arch/sparc/mm/init_64.c
929
cpumask_t numa_cpumask_lookup_table[MAX_NUMNODES];
arch/x86/events/amd/iommu.c
134
static cpumask_t iommu_cpumask;
arch/x86/events/amd/power.c
42
static cpumask_t cpu_mask;
arch/x86/events/amd/uncore.c
54
cpumask_t active_mask;
arch/x86/events/intel/uncore.c
30
static cpumask_t uncore_cpu_mask;
arch/x86/events/intel/uncore.h
151
cpumask_t cpu_mask;
arch/x86/events/perf_event.h
735
cpumask_t supported_cpus;
arch/x86/kernel/apic/hw_nmi.c
34
static void nmi_raise_cpu_backtrace(cpumask_t *mask)
arch/x86/kernel/apic/hw_nmi.c
39
void arch_trigger_cpumask_backtrace(const cpumask_t *mask, int exclude_cpu)
arch/x86/kernel/cpu/mce/core.c
1034
static cpumask_t mce_missing_cpus = CPU_MASK_ALL;
arch/x86/kernel/cpu/microcode/core.c
331
static cpumask_t cpu_offline_mask;
arch/x86/kernel/cpu/sgx/encl.c
926
const cpumask_t *sgx_encl_cpumask(struct sgx_encl *encl)
arch/x86/kernel/cpu/sgx/encl.c
928
cpumask_t *cpumask = &encl->cpumask;
arch/x86/kernel/cpu/sgx/encl.h
109
const cpumask_t *sgx_encl_cpumask(struct sgx_encl *encl);
arch/x86/kernel/cpu/sgx/encl.h
63
cpumask_t cpumask;
crypto/scompress.c
44
static cpumask_t scomp_scratch_want;
drivers/accel/habanalabs/common/habanalabs.h
3543
cpumask_t irq_affinity_mask;
drivers/acpi/pptt.c
1016
int acpi_pptt_get_cpumask_from_cache_id(u32 cache_id, cpumask_t *cpus)
drivers/acpi/pptt.c
866
cpumask_t *cpus)
drivers/acpi/pptt.c
902
void acpi_pptt_get_cpus_from_container(u32 acpi_cpu_id, cpumask_t *cpus)
drivers/base/arch_topology.c
707
const cpumask_t *core_mask = cpumask_of_node(cpu_to_node(cpu));
drivers/base/cacheinfo.c
1012
cpumask_t *cpu_map;
drivers/base/cacheinfo.c
1029
cpumask_t *cpu_map;
drivers/base/cacheinfo.c
641
static cpumask_t cache_dev_map;
drivers/base/cacheinfo.c
935
cpumask_t **map)
drivers/base/cacheinfo.c
997
cpumask_t *cpu_map)
drivers/clocksource/arm_arch_timer.c
65
static cpumask_t evtstrm_available = CPU_MASK_NONE;
drivers/cpufreq/intel_pstate.c
1926
static cpumask_t hwp_intr_enable_mask;
drivers/cpufreq/powernv-cpufreq.c
1048
cpumask_t *chip_cpu_mask;
drivers/cpufreq/powernv-cpufreq.c
1056
chip_cpu_mask = kzalloc_objs(cpumask_t, MAX_NR_CHIPS);
drivers/cpufreq/powernv-cpufreq.c
139
cpumask_t mask;
drivers/cpufreq/powernv-cpufreq.c
918
cpumask_t mask;
drivers/cpuidle/coupled.c
122
static cpumask_t cpuidle_coupled_poke_pending;
drivers/cpuidle/coupled.c
129
static cpumask_t cpuidle_coupled_poked;
drivers/cpuidle/coupled.c
98
cpumask_t coupled_cpus;
drivers/cpuidle/dt_idle_states.c
152
const cpumask_t *cpumask;
drivers/cpuidle/dt_idle_states.c
97
const cpumask_t *cpumask)
drivers/crypto/caam/caamalg_qi2.c
5002
static void free_dpaa2_pcpu_netdev(struct dpaa2_caam_priv *priv, const cpumask_t *cpus)
drivers/crypto/caam/qi.c
410
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
519
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
696
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
713
const cpumask_t *cpus = qman_affine_cpus();
drivers/crypto/caam/qi.c
719
static void free_caam_qi_pcpu_netdev(const cpumask_t *cpus)
drivers/crypto/caam/qi.c
735
const cpumask_t *cpus = qman_affine_cpus();
drivers/hwtracing/coresight/coresight-cpu-debug.c
419
cpumask_t mask;
drivers/hwtracing/coresight/coresight-etm-perf.c
201
cpumask_t *mask = &event_data->mask;
drivers/hwtracing/coresight/coresight-etm-perf.c
218
cpumask_t *mask;
drivers/hwtracing/coresight/coresight-etm-perf.c
259
cpumask_t *mask;
drivers/hwtracing/coresight/coresight-etm-perf.c
323
cpumask_t *mask;
drivers/hwtracing/coresight/coresight-etm-perf.h
96
cpumask_t mask;
drivers/hwtracing/coresight/coresight-etm-perf.h
97
cpumask_t aux_hwid_done;
drivers/hwtracing/coresight/coresight-trbe.c
145
cpumask_t supported_cpus;
drivers/hwtracing/ptt/hisi_ptt.c
781
const cpumask_t *cpumask = cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev));
drivers/infiniband/hw/bng_re/bng_re.h
35
cpumask_t mask;
drivers/infiniband/hw/bnxt_re/qplib_fp.h
528
cpumask_t mask;
drivers/infiniband/hw/efa/efa.h
31
cpumask_t affinity_hint_mask;
drivers/infiniband/hw/erdma/erdma.h
159
cpumask_t affinity_hint_mask;
drivers/infiniband/hw/hfi1/affinity.c
779
const cpumask_t *mask)
drivers/infiniband/hw/hfi1/hfi.h
684
cpumask_t mask;
drivers/infiniband/hw/irdma/main.h
224
cpumask_t mask;
drivers/infiniband/ulp/rtrs/rtrs-srv.c
57
static cpumask_t cq_affinity_mask = { CPU_BITS_ALL };
drivers/iommu/hyperv-iommu.c
36
static cpumask_t ioapic_max_cpumask = { CPU_BITS_NONE };
drivers/irqchip/irq-apple-aic.c
321
cpumask_t aff;
drivers/irqchip/irq-gic-v3-its.c
2791
static u64 inherit_vpe_l1_table_from_rd(cpumask_t **mask)
drivers/irqchip/irq-gic-v3-its.c
2930
gic_data_rdist()->vpe_table_mask = kzalloc_obj(cpumask_t, GFP_ATOMIC);
drivers/irqchip/irq-gic-v3.c
73
cpumask_t mask;
drivers/irqchip/irq-loongson-eiointc.c
76
cpumask_t cpuspan_map;
drivers/mailbox/bcm-flexrm-mailbox.c
262
cpumask_t irq_aff_hint;
drivers/net/ethernet/amazon/ena/ena_netdev.h
121
cpumask_t affinity_hint_mask;
drivers/net/ethernet/amd/xgbe/xgbe.h
445
cpumask_t affinity_mask;
drivers/net/ethernet/aquantia/atlantic/aq_pci_func.c
149
void *irq_arg, cpumask_t *affinity_mask)
drivers/net/ethernet/aquantia/atlantic/aq_pci_func.h
25
void *irq_arg, cpumask_t *affinity_mask);
drivers/net/ethernet/aquantia/atlantic/aq_ring.h
160
cpumask_t affinity_mask;
drivers/net/ethernet/aquantia/atlantic/aq_vec.c
358
cpumask_t *aq_vec_get_affinity_mask(struct aq_vec_s *self)
drivers/net/ethernet/aquantia/atlantic/aq_vec.h
44
cpumask_t *aq_vec_get_affinity_mask(struct aq_vec_s *self);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11626
const cpumask_t *mask)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
12952
cpumask_t *q_map;
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
283
cpumask_t affinity_mask;
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
815
const cpumask_t *cpus = qman_affine_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
957
const cpumask_t *affine_cpus = qman_affine_cpus();
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
417
const cpumask_t *cpus = qman_affine_cpus();
drivers/net/ethernet/fungible/funeth/funeth_main.c
235
const cpumask_t *mask)
drivers/net/ethernet/fungible/funeth/funeth_txrx.h
223
cpumask_t affinity_mask; /* IRQ affinity */
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1257
struct hnae_ring *ring, cpumask_t *mask)
drivers/net/ethernet/hisilicon/hns/hns_enet.h
36
cpumask_t mask; /* affinity mask */
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
566
cpumask_t affinity_mask;
drivers/net/ethernet/huawei/hinic3/hinic3_nic_dev.h
73
cpumask_t affinity_mask;
drivers/net/ethernet/intel/fm10k/fm10k.h
194
cpumask_t affinity_mask;
drivers/net/ethernet/intel/i40e/i40e.h
961
cpumask_t affinity_mask;
drivers/net/ethernet/intel/i40e/i40e_main.c
4101
const cpumask_t *mask)
drivers/net/ethernet/intel/ixgbe/ixgbe.h
516
cpumask_t affinity_mask;
drivers/net/ethernet/marvell/mvneta.c
705
cpumask_t affinity_mask;
drivers/net/ethernet/marvell/octeon_ep/octep_main.h
152
cpumask_t affinity_mask;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.h
128
cpumask_t affinity_mask;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
310
cpumask_t sp_affinity_mask;
drivers/net/ethernet/netronome/nfp/nfp_net.h
442
cpumask_t affinity_mask;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
279
const cpumask_t *mask)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
493
cpumask_t affinity_mask[IWL_MAX_RX_HW_QUEUES];
drivers/perf/arm_cspmu/arm_cspmu.c
296
const cpumask_t *cpumask;
drivers/perf/arm_cspmu/arm_cspmu.h
230
cpumask_t associated_cpus;
drivers/perf/arm_cspmu/arm_cspmu.h
231
cpumask_t active_cpu;
drivers/perf/arm_dsu_pmu.c
113
cpumask_t associated_cpus;
drivers/perf/arm_dsu_pmu.c
114
cpumask_t active_cpu;
drivers/perf/arm_dsu_pmu.c
148
const cpumask_t *cpumask;
drivers/perf/arm_dsu_pmu.c
592
static int dsu_pmu_dt_get_cpus(struct device *dev, cpumask_t *mask)
drivers/perf/arm_dsu_pmu.c
622
static int dsu_pmu_acpi_get_cpus(struct device *dev, cpumask_t *mask)
drivers/perf/arm_spe_pmu.c
1297
cpumask_t *mask = &spe_pmu->supported_cpus;
drivers/perf/arm_spe_pmu.c
73
cpumask_t supported_cpus;
drivers/perf/hisilicon/hisi_uncore_pmu.h
129
cpumask_t associated_cpus;
drivers/perf/qcom_l2_pmu.c
114
cpumask_t cpumask;
drivers/perf/qcom_l2_pmu.c
146
cpumask_t cluster_cpus;
drivers/perf/qcom_l3_pmu.c
160
cpumask_t cpumask;
drivers/perf/starfive_starlink_pmu.c
75
cpumask_t cpumask;
drivers/perf/xgene_pmu.c
126
cpumask_t cpu;
drivers/platform/x86/intel/tpmi_power_domains.c
136
cpumask_t *tpmi_get_power_domain_mask(int cpu_no)
drivers/platform/x86/intel/tpmi_power_domains.c
139
cpumask_t *mask;
drivers/platform/x86/intel/tpmi_power_domains.c
75
static cpumask_t *tpmi_power_domain_mask;
drivers/platform/x86/intel/tpmi_power_domains.h
16
cpumask_t *tpmi_get_power_domain_mask(int cpu_no);
drivers/platform/x86/intel/uncore-frequency/uncore-frequency.c
33
static cpumask_t uncore_cpu_mask;
drivers/resctrl/mpam_devices.c
449
cpumask_t *affinity)
drivers/resctrl/mpam_devices.c
458
static void get_cpumask_from_node_id(u32 node_id, cpumask_t *affinity)
drivers/resctrl/mpam_devices.c
468
static int mpam_ris_get_affinity(struct mpam_msc *msc, cpumask_t *affinity,
drivers/resctrl/mpam_internal.h
245
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
282
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
323
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
392
cpumask_t *affinity);
drivers/resctrl/mpam_internal.h
61
cpumask_t accessibility;
drivers/scsi/mpt3sas/mpt3sas_base.c
3250
const cpumask_t *mask;
drivers/soc/fsl/qbman/bman.c
226
static cpumask_t affine_mask;
drivers/soc/fsl/qbman/qman.c
1000
static cpumask_t affine_mask;
drivers/soc/fsl/qbman/qman.c
1725
const cpumask_t *qman_affine_cpus(void)
drivers/watchdog/octeon-wdt-main.c
86
static cpumask_t irq_enabled_cpus;
fs/resctrl/ctrlmondata.c
556
cpumask_t *cpumask, struct mon_evt *evt, int first)
fs/resctrl/internal.h
390
cpumask_t *cpumask, struct mon_evt *evt, int first);
include/crypto/internal/acompress.h
70
cpumask_t stream_want;
include/linux/acpi.h
1548
void acpi_pptt_get_cpus_from_container(u32 acpi_cpu_id, cpumask_t *cpus);
include/linux/acpi.h
1550
int acpi_pptt_get_cpumask_from_cache_id(u32 cache_id, cpumask_t *cpus);
include/linux/acpi.h
1573
cpumask_t *cpus) { }
include/linux/acpi.h
1579
cpumask_t *cpus)
include/linux/arch_topology.h
67
cpumask_t thread_sibling;
include/linux/arch_topology.h
68
cpumask_t core_sibling;
include/linux/arch_topology.h
69
cpumask_t cluster_sibling;
include/linux/arch_topology.h
70
cpumask_t llc_sibling;
include/linux/cacheinfo.h
59
cpumask_t shared_cpu_map;
include/linux/cpuidle.h
111
cpumask_t coupled_cpus;
include/linux/cpumask.h
133
extern cpumask_t cpus_booted_once_mask;
include/linux/cpumask.h
1381
(cpumask_t) { { \
include/linux/cpumask.h
1386
(cpumask_t) { { \
include/linux/cpumask.h
1393
(cpumask_t) { { \
include/linux/cpumask.h
1398
(cpumask_t) { { \
include/linux/energy_model.h
177
const cpumask_t *cpus, bool microwatts);
include/linux/energy_model.h
180
const cpumask_t *cpus, bool microwatts);
include/linux/energy_model.h
356
const cpumask_t *cpus, bool microwatts)
include/linux/energy_model.h
363
const cpumask_t *cpus, bool microwatts)
include/linux/interrupt.h
185
const cpumask_t *affinity, void __percpu *percpu_dev_id);
include/linux/interrupt.h
275
void (*notify)(struct irq_affinity_notify *, const cpumask_t *mask);
include/linux/irqchip/arm-gic-v3.h
622
cpumask_t *vpe_table_mask;
include/linux/mm_types.h
1419
[0 ... sizeof(cpumask_t) + MM_CID_STATIC_SIZE - 1] = 0 \
include/linux/mm_types.h
1432
static inline cpumask_t *mm_cpumask(struct mm_struct *mm)
include/linux/mm_types.h
1521
static inline cpumask_t *mm_cpus_allowed(struct mm_struct *mm)
include/linux/mm_types.h
1566
# define MM_CID_STATIC_SIZE (2 * sizeof(cpumask_t))
include/linux/netdevice.h
373
cpumask_t affinity_mask;
include/linux/nmi.h
184
void nmi_trigger_cpumask_backtrace(const cpumask_t *mask,
include/linux/nmi.h
186
void (*raise)(cpumask_t *mask));
include/linux/perf/arm_pmu.h
90
cpumask_t supported_cpus;
include/linux/sched.h
923
const cpumask_t *cpus_ptr;
include/linux/sched.h
924
cpumask_t *user_cpus_ptr;
include/linux/sched.h
925
cpumask_t cpus_mask;
include/soc/fsl/qman.h
901
const cpumask_t *qman_affine_cpus(void);
kernel/bpf/cpumask.c
26
cpumask_t cpumask;
kernel/bpf/verifier.c
7263
const cpumask_t *cpus_ptr;
kernel/cgroup/cpuset.c
3740
static cpumask_t new_cpus;
kernel/cgroup/cpuset.c
3855
static cpumask_t new_cpus;
kernel/cpu.c
91
cpumask_t cpus_booted_once_mask;
kernel/irq/manage.c
2481
const char *devname, const cpumask_t *affinity,
kernel/irq/manage.c
2527
const cpumask_t *affinity, void __percpu *dev_id)
kernel/locking/locktorture.c
1095
static cpumask_t cpumask_all;
kernel/locking/locktorture.c
1096
cpumask_t *rcmp = cpumask_nonempty(bind_readers) ? bind_readers : &cpumask_all;
kernel/locking/locktorture.c
1097
cpumask_t *wcmp = cpumask_nonempty(bind_writers) ? bind_writers : &cpumask_all;
kernel/power/energy_model.c
419
const cpumask_t *cpus,
kernel/power/energy_model.c
590
const cpumask_t *cpus, bool microwatts)
kernel/power/energy_model.c
614
const cpumask_t *cpus, bool microwatts)
kernel/sched/core.c
2670
static inline void mm_update_cpus_allowed(struct mm_struct *mm, const cpumask_t *affmask);
kernel/sched/core.c
2713
cpumask_t cpumask;
kernel/sched/core.c
2731
cpumask_t *user_mask;
kernel/sched/fair.c
8711
cpumask_t *cpumask;
kernel/sched/sched.h
2814
static inline cpumask_t *alloc_user_cpus_ptr(int node)
kernel/sched/syscalls.c
624
cpumask_t *span = rq->rd->span;
kernel/time/clocksource.c
305
static cpumask_t cpus_ahead;
kernel/time/clocksource.c
306
static cpumask_t cpus_behind;
kernel/time/clocksource.c
307
static cpumask_t cpus_chosen;
kernel/watchdog_buddy.c
9
static cpumask_t __read_mostly watchdog_cpus;
kernel/workqueue.c
2671
static cpumask_t *pool_allowed_cpus(struct worker_pool *pool)
kernel/workqueue.c
4795
const cpumask_t *unbound_cpumask)
kernel/workqueue.c
6717
static cpumask_t cpumask;
lib/cpu_rmap.c
263
irq_cpu_rmap_notify(struct irq_affinity_notify *notify, const cpumask_t *mask)
lib/nmi_backtrace.c
36
void nmi_trigger_cpumask_backtrace(const cpumask_t *mask,
lib/nmi_backtrace.c
38
void (*raise)(cpumask_t *mask))
lib/tests/cpumask_kunit.c
18
const cpumask_t *m = (mask); \
lib/tests/cpumask_kunit.c
28
const cpumask_t *m1 = (mask1); \
lib/tests/cpumask_kunit.c
29
const cpumask_t *m2 = (mask2); \
lib/tests/cpumask_kunit.c
41
const cpumask_t *m = (mask); \
lib/tests/cpumask_kunit.c
58
static cpumask_t mask_empty;
lib/tests/cpumask_kunit.c
59
static cpumask_t mask_all;
lib/tests/cpumask_kunit.c
60
static cpumask_t mask_tmp;
mm/page_alloc.c
2709
static cpumask_t cpus_with_pcps;
mm/vmalloc.c
2346
static cpumask_t purge_nodes;
net/core/dev.c
6532
cpumask_t flush_cpus;
net/core/dev.c
7299
const cpumask_t *mask)
net/iucv/iucv.c
132
static cpumask_t iucv_buffer_cpumask = { CPU_BITS_NONE };
net/iucv/iucv.c
133
static cpumask_t iucv_irq_cpumask = { CPU_BITS_NONE };
net/iucv/iucv.c
550
static cpumask_t cpumask;
samples/trace_events/trace-events-sample.h
574
TP_PROTO(const char *foo, int bar, unsigned long *mask, const cpumask_t *cpus),
tools/perf/util/svghelper.c
694
cpumask_t *sib_core;
tools/perf/util/svghelper.c
696
cpumask_t *sib_thr;
tools/perf/util/svghelper.c
727
static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus)
tools/perf/util/svghelper.c
762
t.sib_core = calloc(env->nr_sibling_cores, sizeof(cpumask_t));
tools/perf/util/svghelper.c
763
t.sib_thr = calloc(env->nr_sibling_threads, sizeof(cpumask_t));
tools/sched_ext/include/scx/common.bpf.h
93
s32 scx_bpf_pick_idle_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
94
s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
tools/sched_ext/include/scx/common.bpf.h
95
s32 scx_bpf_pick_any_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
tools/sched_ext/include/scx/common.bpf.h
96
s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
tools/testing/selftests/bpf/progs/cpumask_success.c
798
#define CPUMASK_TEST_MASKLEN (sizeof(cpumask_t))
tools/testing/selftests/net/bench/page_pool/time_bench.c
258
cpumask_t newmask = CPU_MASK_NONE;