Symbol: cpumask_var_t
arch/arm64/include/asm/kvm_host.h
371
cpumask_var_t supported_cpus;
arch/arm64/kernel/cpufeature.c
147
static cpumask_var_t cpu_32bit_el0_mask __cpumask_var_read_mostly;
arch/arm64/kernel/topology.c
45
static cpumask_var_t amu_fie_cpus;
arch/mips/kernel/mips-mt-fpaff.c
69
cpumask_var_t cpus_allowed, new_mask, effective_mask;
arch/powerpc/include/asm/mmzone.h
28
extern cpumask_var_t node_to_cpumask_map[];
arch/powerpc/include/asm/smp.h
109
DECLARE_PER_CPU(cpumask_var_t, cpu_sibling_map);
arch/powerpc/include/asm/smp.h
110
DECLARE_PER_CPU(cpumask_var_t, cpu_l2_cache_map);
arch/powerpc/include/asm/smp.h
111
DECLARE_PER_CPU(cpumask_var_t, cpu_core_map);
arch/powerpc/include/asm/smp.h
112
DECLARE_PER_CPU(cpumask_var_t, cpu_smallcore_map);
arch/powerpc/include/asm/smp.h
38
DECLARE_PER_CPU(cpumask_var_t, thread_group_l1_cache_map);
arch/powerpc/include/asm/smp.h
39
DECLARE_PER_CPU(cpumask_var_t, thread_group_l2_cache_map);
arch/powerpc/include/asm/smp.h
40
DECLARE_PER_CPU(cpumask_var_t, thread_group_l3_cache_map);
arch/powerpc/kernel/smp.c
120
DEFINE_PER_CPU(cpumask_var_t, thread_group_l1_cache_map);
arch/powerpc/kernel/smp.c
127
DEFINE_PER_CPU(cpumask_var_t, thread_group_l2_cache_map);
arch/powerpc/kernel/smp.c
133
DEFINE_PER_CPU(cpumask_var_t, thread_group_l3_cache_map);
arch/powerpc/kernel/smp.c
1431
static bool update_mask_by_l2(int cpu, cpumask_var_t *mask)
arch/powerpc/kernel/smp.c
1544
static void update_coregroup_mask(int cpu, cpumask_var_t *mask)
arch/powerpc/kernel/smp.c
1584
cpumask_var_t mask;
arch/powerpc/kernel/smp.c
86
DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map);
arch/powerpc/kernel/smp.c
87
DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map);
arch/powerpc/kernel/smp.c
88
DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map);
arch/powerpc/kernel/smp.c
89
DEFINE_PER_CPU(cpumask_var_t, cpu_core_map);
arch/powerpc/kernel/smp.c
90
static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map);
arch/powerpc/kernel/smp.c
923
static int __init update_mask_from_threadgroup(cpumask_var_t *mask, struct thread_groups *tg,
arch/powerpc/kernel/smp.c
951
cpumask_var_t *mask = NULL;
arch/powerpc/mm/numa.c
45
cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
arch/powerpc/platforms/powernv/subcore.c
141
static cpumask_var_t cpu_offline_mask;
arch/powerpc/platforms/pseries/hotplug-cpu.c
157
cpumask_var_t *cpu_mask)
arch/powerpc/platforms/pseries/hotplug-cpu.c
159
cpumask_var_t candidate_mask;
arch/powerpc/platforms/pseries/hotplug-cpu.c
218
cpumask_var_t cpu_mask;
arch/powerpc/platforms/pseries/hotplug-cpu.c
47
static cpumask_var_t node_recorded_ids_map[MAX_NUMNODES];
arch/powerpc/platforms/pseries/smp.c
52
static cpumask_var_t of_spin_mask;
arch/powerpc/sysdev/xive/common.c
599
cpumask_var_t mask;
arch/s390/kernel/perf_cpum_cf.c
1436
cpumask_var_t mask;
arch/s390/kernel/perf_cpum_cf.c
1555
cpumask_var_t mask;
arch/s390/kernel/perf_cpum_cf.c
289
cpumask_var_t mask;
arch/x86/include/asm/kvm_host.h
1026
cpumask_var_t wbinvd_dirty_mask;
arch/x86/include/asm/smp.h
12
DECLARE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_sibling_map);
arch/x86/include/asm/smp.h
13
DECLARE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_core_map);
arch/x86/include/asm/smp.h
14
DECLARE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_die_map);
arch/x86/include/asm/smp.h
16
DECLARE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_llc_shared_map);
arch/x86/include/asm/smp.h
17
DECLARE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_l2c_shared_map);
arch/x86/include/asm/topology.h
64
extern cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
arch/x86/kernel/apic/vector.c
43
static cpumask_var_t vector_searchmask;
arch/x86/kernel/apic/x2apic_cluster.c
21
static DEFINE_PER_CPU(cpumask_var_t, ipi_mask);
arch/x86/kernel/cpu/cacheinfo.c
27
DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_llc_shared_map);
arch/x86/kernel/cpu/cacheinfo.c
30
DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_l2c_shared_map);
arch/x86/kernel/cpu/cacheinfo.c
32
static cpumask_var_t cpu_cacheinfo_mask;
arch/x86/kernel/cpu/mce/core.c
2677
static cpumask_var_t mce_device_initialized;
arch/x86/kernel/cpu/mce/inject.c
175
static cpumask_var_t mce_inject_cpumask;
arch/x86/kernel/cpu/resctrl/rdtgroup.c
137
cpumask_var_t cpu_mask;
arch/x86/kernel/kvm.c
490
static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask);
arch/x86/kernel/smpboot.c
103
DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_die_map);
arch/x86/kernel/smpboot.c
107
static cpumask_var_t cpu_sibling_setup_mask;
arch/x86/kernel/smpboot.c
95
DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_sibling_map);
arch/x86/kernel/smpboot.c
99
DEFINE_PER_CPU_READ_MOSTLY(cpumask_var_t, cpu_core_map);
arch/x86/kvm/svm/svm.h
117
cpumask_var_t have_run_cpus; /* CPUs that have done VMRUN for this VM. */
arch/x86/kvm/vmx/tdx.c
2382
cpumask_var_t packages;
arch/x86/kvm/vmx/tdx.c
495
cpumask_var_t packages, targets;
arch/x86/mm/mmio-mod.c
366
static cpumask_var_t downed_cpus;
arch/x86/mm/numa.c
67
cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
arch/x86/platform/uv/uv_nmi.c
94
static cpumask_var_t uv_nmi_cpu_mask;
arch/x86/virt/vmx/tdx/tdx.c
991
cpumask_var_t packages;
arch/x86/xen/mmu_pv.c
988
cpumask_var_t mask;
arch/x86/xen/smp_pv.c
50
cpumask_var_t xen_cpu_initialized_map;
arch/x86/xen/xen-ops.h
92
extern cpumask_var_t xen_cpu_initialized_map;
drivers/acpi/acpi_pad.c
94
cpumask_var_t tmp;
drivers/acpi/processor_perflib.c
597
cpumask_var_t covered_cpus;
drivers/acpi/processor_throttling.c
59
cpumask_var_t covered_cpus;
drivers/base/arch_numa.c
36
cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
drivers/base/arch_topology.c
385
static cpumask_var_t cpus_to_visit;
drivers/base/cpu.c
251
cpumask_var_t offline;
drivers/base/cpu.c
288
cpumask_var_t isolated;
drivers/base/node.c
37
cpumask_var_t mask;
drivers/base/node.c
58
cpumask_var_t mask;
drivers/base/topology.c
30
cpumask_var_t mask; \
drivers/base/topology.c
48
cpumask_var_t mask; \
drivers/block/drbd/drbd_int.h
607
cpumask_var_t cpu_mask;
drivers/block/drbd/drbd_main.c
2464
cpumask_var_t new_cpu_mask;
drivers/block/drbd/drbd_main.c
2476
cpumask_var_t tmp_cpu_mask;
drivers/block/drbd/drbd_main.c
472
static void drbd_calc_cpu_mask(cpumask_var_t *cpu_mask)
drivers/block/ublk_drv.c
4471
cpumask_var_t cpumask;
drivers/char/random.c
1288
cpumask_var_t timer_cpus;
drivers/cpufreq/acpi-cpufreq.c
57
cpumask_var_t freqdomain_cpus;
drivers/cpufreq/cpufreq-dt.c
30
cpumask_var_t cpus;
drivers/cpufreq/cpufreq_ondemand.c
391
cpumask_var_t done;
drivers/cpufreq/scmi-cpufreq.c
30
cpumask_var_t opp_shared_cpus;
drivers/cpufreq/speedstep-centrino.c
424
cpumask_var_t covered_cpus;
drivers/crypto/caam/caamalg_qi2.h
69
cpumask_var_t clean_mask;
drivers/crypto/caam/qi.c
736
cpumask_var_t clean_mask;
drivers/crypto/cavium/cpt/cptvf.h
107
cpumask_var_t affinity_mask[CPT_VF_MSIX_VECTORS];
drivers/crypto/marvell/octeontx/otx_cptvf.h
84
cpumask_var_t affinity_mask[OTX_CPT_VF_MSIX_VECTORS];
drivers/crypto/marvell/octeontx2/otx2_cptlf.h
88
cpumask_var_t affinity_mask; /* IRQs affinity mask */
drivers/edac/amd64_edac.c
3207
cpumask_var_t mask;
drivers/edac/amd64_edac.c
3239
cpumask_var_t cmask;
drivers/firmware/psci/psci_checker.c
140
static void free_cpu_groups(int num, cpumask_var_t **pcpu_groups)
drivers/firmware/psci/psci_checker.c
143
cpumask_var_t *cpu_groups = *pcpu_groups;
drivers/firmware/psci/psci_checker.c
150
static int alloc_init_cpu_groups(cpumask_var_t **pcpu_groups)
drivers/firmware/psci/psci_checker.c
153
cpumask_var_t tmp, *cpu_groups;
drivers/firmware/psci/psci_checker.c
188
cpumask_var_t offlined_cpus, *cpu_groups;
drivers/hv/channel_mgmt.c
745
cpumask_var_t available_mask;
drivers/hwmon/fam15h_power.c
159
cpumask_var_t mask;
drivers/infiniband/hw/hfi1/affinity.c
234
static int per_cpu_affinity_get(cpumask_var_t possible_cpumask,
drivers/infiniband/hw/hfi1/affinity.c
274
static int per_cpu_affinity_put_max(cpumask_var_t possible_cpumask,
drivers/infiniband/hw/hfi1/affinity.c
313
cpumask_var_t non_intr_cpus,
drivers/infiniband/hw/hfi1/affinity.c
314
cpumask_var_t available_cpus)
drivers/infiniband/hw/hfi1/affinity.c
396
cpumask_var_t non_intr_cpus;
drivers/infiniband/hw/hfi1/affinity.c
397
cpumask_var_t available_cpus;
drivers/infiniband/hw/hfi1/affinity.c
62
static int cpu_mask_set_get_first(struct cpu_mask_set *set, cpumask_var_t diff)
drivers/infiniband/hw/hfi1/affinity.c
827
cpumask_var_t diff;
drivers/infiniband/hw/hfi1/affinity.c
954
static void find_hw_thread_mask(uint hw_thread_no, cpumask_var_t hw_thread_mask,
drivers/infiniband/hw/hfi1/affinity.c
980
cpumask_var_t diff, hw_thread_mask, available_mask, intrs_mask;
drivers/infiniband/hw/hfi1/netdev_rx.c
159
cpumask_var_t node_cpu_mask;
drivers/infiniband/hw/hfi1/sdma.c
897
cpumask_var_t mask, new_mask;
drivers/md/dm-ps-io-affinity.c
16
cpumask_var_t cpumask;
drivers/md/dm-ps-io-affinity.c
23
cpumask_var_t path_mask;
drivers/net/ethernet/broadcom/bnge/bnge_resc.h
65
cpumask_var_t cpu_mask;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1250
cpumask_var_t cpu_mask;
drivers/net/ethernet/cavium/thunder/nic.h
375
cpumask_var_t affinity_mask[NIC_VF_MSIX_VECTORS];
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1054
cpumask_var_t aff_mask;
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2129
cpumask_var_t *aff_mask, int idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
2130
void cxgb4_clear_msix_aff(unsigned short vec, cpumask_var_t aff_mask);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
690
cpumask_var_t *aff_mask, int idx)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
711
void cxgb4_clear_msix_aff(unsigned short vec, cpumask_var_t aff_mask)
drivers/net/ethernet/cisco/enic/enic.h
47
cpumask_var_t affinity_mask;
drivers/net/ethernet/emulex/benet/be.h
189
cpumask_var_t affinity_mask;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
2909
cpumask_var_t xps_mask;
drivers/net/ethernet/ibm/ibmvnic.c
220
cpumask_var_t mask;
drivers/net/ethernet/ibm/ibmvnic.h
834
cpumask_var_t affinity_mask;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
253
cpumask_var_t *affinity_mask;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
3048
sizeof(cpumask_var_t), GFP_KERNEL);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_vf.c
635
sizeof(cpumask_var_t), GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx4/mlx4.h
406
cpumask_var_t affinity_mask;
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
351
cpumask_var_t affinity_mask;
drivers/net/ethernet/mellanox/mlx5/core/en.h
890
cpumask_var_t cpumask;
drivers/net/ethernet/mellanox/mlx5/core/pci_irq.c
32
cpumask_var_t mask;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1588
cpumask_var_t cpus __free(free_cpumask_var);
drivers/net/ethernet/pensando/ionic/ionic.h
58
cpumask_var_t *affinity_masks;
drivers/net/ethernet/pensando/ionic/ionic_api.h
54
cpumask_var_t *affinity_mask;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
3181
cpumask_var_t *affinity_masks;
drivers/net/ethernet/pensando/ionic/ionic_lif.c
3185
affinity_masks = kzalloc_objs(cpumask_var_t, nintrs);
drivers/net/ethernet/pensando/ionic/ionic_lif.c
506
cpumask_var_t *affinity_mask;
drivers/net/ethernet/sfc/efx_channels.c
57
cpumask_var_t filter_mask;
drivers/net/ethernet/sfc/falcon/efx.c
1316
cpumask_var_t thread_mask;
drivers/net/ethernet/sfc/siena/efx_channels.c
57
cpumask_var_t filter_mask;
drivers/net/virtio_net.c
3940
cpumask_var_t mask;
drivers/pci/controller/pci-hyperv.c
617
cpumask_var_t tmp;
drivers/pci/controller/pcie-iproc-msi.c
483
cpumask_var_t mask;
drivers/powercap/intel_rapl_common.c
1592
static void set_pmu_cpumask(struct rapl_package *rp, cpumask_var_t mask)
drivers/powercap/intel_rapl_common.c
1882
cpumask_var_t cpu_mask;
drivers/soc/fsl/dpio/dpio-driver.c
33
static cpumask_var_t cpus_unused_mask;
drivers/target/iscsi/iscsi_target.c
3571
cpumask_var_t conn_allowed_cpumask;
drivers/target/iscsi/iscsi_target_configfs.c
1156
cpumask_var_t new_allowed_cpumask;
drivers/thermal/intel/intel_hfi.c
125
cpumask_var_t cpus;
drivers/thermal/intel/intel_powerclamp.c
128
static cpumask_var_t idle_injection_cpu_mask;
drivers/thermal/intel/intel_powerclamp.c
146
static bool check_invalid(cpumask_var_t mask, u8 idle)
drivers/thermal/intel/intel_powerclamp.c
156
cpumask_var_t new_mask;
drivers/vdpa/vdpa_user/vduse_dev.c
1693
cpumask_var_t new_value;
drivers/virt/nitro_enclaves/ne_misc_dev.c
120
cpumask_var_t *avail_threads_per_core;
drivers/virt/nitro_enclaves/ne_misc_dev.c
183
cpumask_var_t cpu_pool;
drivers/virt/nitro_enclaves/ne_misc_dev.h
80
cpumask_var_t *threads_per_core;
drivers/virt/nitro_enclaves/ne_misc_dev.h
81
cpumask_var_t vcpu_ids;
drivers/virtio/virtio_pci_common.h
89
cpumask_var_t *msix_affinity_masks;
fs/resctrl/pseudo_lock.c
644
cpumask_var_t cpu_with_psl;
fs/resctrl/rdtgroup.c
3952
static int rdtgroup_rmdir_mon(struct rdtgroup *rdtgrp, cpumask_var_t tmpmask)
fs/resctrl/rdtgroup.c
4003
static int rdtgroup_rmdir_ctrl(struct rdtgroup *rdtgrp, cpumask_var_t tmpmask)
fs/resctrl/rdtgroup.c
403
static int cpus_mon_write(struct rdtgroup *rdtgrp, cpumask_var_t newmask,
fs/resctrl/rdtgroup.c
404
cpumask_var_t tmpmask)
fs/resctrl/rdtgroup.c
4056
cpumask_var_t tmpmask;
fs/resctrl/rdtgroup.c
4109
cpumask_var_t cpus)
fs/resctrl/rdtgroup.c
4139
cpumask_var_t tmpmask;
fs/resctrl/rdtgroup.c
456
static int cpus_ctrl_write(struct rdtgroup *rdtgrp, cpumask_var_t newmask,
fs/resctrl/rdtgroup.c
457
cpumask_var_t tmpmask, cpumask_var_t tmpmask1)
fs/resctrl/rdtgroup.c
514
cpumask_var_t tmpmask, newmask, tmpmask1;
include/acpi/cppc_acpi.h
146
cpumask_var_t shared_cpu_map;
include/acpi/processor.h
138
cpumask_var_t shared_cpu_map;
include/acpi/processor.h
183
cpumask_var_t shared_cpu_map;
include/asm-generic/numa.h
17
extern cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
include/linux/blk-mq.h
346
cpumask_var_t cpumask;
include/linux/cpufreq.h
55
cpumask_var_t cpus; /* Online CPUs only */
include/linux/cpufreq.h
56
cpumask_var_t related_cpus; /* Online + Offline CPUs */
include/linux/cpufreq.h
57
cpumask_var_t real_cpus; /* Related and present */
include/linux/cpumask.h
1029
bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node);
include/linux/cpumask.h
1032
bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node)
include/linux/cpumask.h
1050
bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
include/linux/cpumask.h
1056
bool zalloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
include/linux/cpumask.h
1061
void alloc_bootmem_cpumask_var(cpumask_var_t *mask);
include/linux/cpumask.h
1062
void free_cpumask_var(cpumask_var_t mask);
include/linux/cpumask.h
1063
void free_bootmem_cpumask_var(cpumask_var_t mask);
include/linux/cpumask.h
1065
static __always_inline bool cpumask_available(cpumask_var_t mask)
include/linux/cpumask.h
1076
static __always_inline bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
include/linux/cpumask.h
1081
static __always_inline bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags,
include/linux/cpumask.h
1087
static __always_inline bool zalloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
include/linux/cpumask.h
1093
static __always_inline bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags,
include/linux/cpumask.h
1100
static __always_inline void alloc_bootmem_cpumask_var(cpumask_var_t *mask)
include/linux/cpumask.h
1104
static __always_inline void free_cpumask_var(cpumask_var_t mask)
include/linux/cpumask.h
1108
static __always_inline void free_bootmem_cpumask_var(cpumask_var_t mask)
include/linux/cpumask.h
1112
static __always_inline bool cpumask_available(cpumask_var_t mask)
include/linux/interrupt.h
316
extern cpumask_var_t irq_default_affinity;
include/linux/irq.h
153
cpumask_var_t affinity;
include/linux/irq.h
156
cpumask_var_t effective_affinity;
include/linux/irqdesc.h
103
cpumask_var_t pending_mask;
include/linux/padata.h
188
cpumask_var_t cpumask);
include/linux/padata.h
77
cpumask_var_t pcpu;
include/linux/padata.h
78
cpumask_var_t cbcpu;
include/linux/pm_domain.h
214
cpumask_var_t cpus; /* A cpumask of the attached CPUs */
include/linux/sched/topology.h
158
extern void partition_sched_domains(int ndoms_new, cpumask_var_t doms_new[],
include/linux/sched/topology.h
162
cpumask_var_t *alloc_sched_domains(unsigned int ndoms);
include/linux/sched/topology.h
163
void free_sched_domains(cpumask_var_t doms[], unsigned int ndoms);
include/linux/tick.h
175
extern cpumask_var_t tick_nohz_full_mask;
include/linux/tick.h
276
extern void __init tick_nohz_full_setup(cpumask_var_t cpumask);
include/linux/tick.h
301
static inline void tick_nohz_full_setup(cpumask_var_t cpumask) { }
include/linux/trace_events.h
114
cpumask_var_t started;
include/linux/workqueue.h
160
cpumask_var_t cpumask;
include/linux/workqueue.h
172
cpumask_var_t __pod_cpumask;
include/net/ip_vs.h
1000
cpumask_var_t sysctl_est_cpulist; /* kthread cpumask */
include/target/iscsi/iscsi_target_core.h
580
cpumask_var_t conn_cpumask;
include/target/iscsi/iscsi_target_core.h
581
cpumask_var_t allowed_cpumask;
include/target/iscsi/iscsi_target_core.h
891
cpumask_var_t allowed_cpumask;
io_uring/io-wq.c
134
cpumask_var_t cpu_mask;
io_uring/io-wq.c
1445
int io_wq_cpu_affinity(struct io_uring_task *tctx, cpumask_var_t mask)
io_uring/io-wq.c
1447
cpumask_var_t allowed_mask;
io_uring/io-wq.h
50
int io_wq_cpu_affinity(struct io_uring_task *tctx, cpumask_var_t mask);
io_uring/register.c
300
cpumask_var_t new_mask)
io_uring/register.c
318
cpumask_var_t new_mask;
io_uring/sqpoll.c
494
cpumask_var_t allowed_mask;
io_uring/sqpoll.c
546
cpumask_var_t mask)
io_uring/sqpoll.h
31
int io_sqpoll_wq_cpu_affinity(struct io_ring_ctx *ctx, cpumask_var_t mask);
kernel/cgroup/cpuset-internal.h
101
cpumask_var_t cpus_allowed;
kernel/cgroup/cpuset-internal.h
105
cpumask_var_t effective_cpus;
kernel/cgroup/cpuset-internal.h
120
cpumask_var_t effective_xcpus;
kernel/cgroup/cpuset-internal.h
134
cpumask_var_t exclusive_cpus;
kernel/cgroup/cpuset-internal.h
327
int cpuset1_generate_sched_domains(cpumask_var_t **domains,
kernel/cgroup/cpuset-internal.h
343
static inline int cpuset1_generate_sched_domains(cpumask_var_t **domains,
kernel/cgroup/cpuset-v1.c
629
int cpuset1_generate_sched_domains(cpumask_var_t **domains,
kernel/cgroup/cpuset-v1.c
636
cpumask_var_t *doms; /* resulting partition; i.e. sched domains */
kernel/cgroup/cpuset.c
1286
cpumask_var_t full_hk_cpus;
kernel/cgroup/cpuset.c
142
static cpumask_var_t subpartitions_cpus; /* RWCS */
kernel/cgroup/cpuset.c
147
static cpumask_var_t isolated_cpus; /* CSCB */
kernel/cgroup/cpuset.c
157
static cpumask_var_t isolated_hk_cpus; /* T */
kernel/cgroup/cpuset.c
217
cpumask_var_t addmask, delmask; /* For partition root */
kernel/cgroup/cpuset.c
218
cpumask_var_t new_cpus; /* For update_cpumasks_hier() */
kernel/cgroup/cpuset.c
3098
static cpumask_var_t cpus_attach;
kernel/cgroup/cpuset.c
514
static inline int alloc_cpumasks(cpumask_var_t *pmasks[], u32 size)
kernel/cgroup/cpuset.c
539
cpumask_var_t *pmask[3] = {
kernel/cgroup/cpuset.c
583
cpumask_var_t *pmask[4] = {
kernel/cgroup/cpuset.c
808
static int generate_sched_domains(cpumask_var_t **domains,
kernel/cgroup/cpuset.c
814
cpumask_var_t *doms; /* resulting partition; i.e. sched domains */
kernel/cgroup/cpuset.c
982
cpumask_var_t *doms;
kernel/compat.c
128
cpumask_var_t new_mask;
kernel/compat.c
148
cpumask_var_t mask;
kernel/cpu.c
1884
static cpumask_var_t frozen_cpus;
kernel/events/core.c
466
static cpumask_var_t perf_online_mask;
kernel/events/core.c
467
static cpumask_var_t perf_online_core_mask;
kernel/events/core.c
468
static cpumask_var_t perf_online_die_mask;
kernel/events/core.c
469
static cpumask_var_t perf_online_cluster_mask;
kernel/events/core.c
470
static cpumask_var_t perf_online_pkg_mask;
kernel/events/core.c
471
static cpumask_var_t perf_online_sys_mask;
kernel/irq/manage.c
1024
cpumask_var_t mask;
kernel/irq/manage.c
155
cpumask_var_t irq_default_affinity;
kernel/irq/manage.c
523
cpumask_var_t cpumask;
kernel/irq/proc.c
140
cpumask_var_t new_value;
kernel/irq/proc.c
237
cpumask_var_t new_value;
kernel/irq/proc.c
87
cpumask_var_t mask;
kernel/kthread.c
353
cpumask_var_t affinity;
kernel/kthread.c
842
cpumask_var_t affinity;
kernel/kthread.c
880
cpumask_var_t affinity;
kernel/locking/locktorture.c
62
static cpumask_var_t bind_readers; // Bind the readers to the specified set of CPUs.
kernel/locking/locktorture.c
63
static cpumask_var_t bind_writers; // Bind the writers to the specified set of CPUs.
kernel/locking/locktorture.c
69
cpumask_var_t *cm_bind = kp->arg;
kernel/locking/locktorture.c
91
cpumask_var_t *cm_bind = kp->arg;
kernel/locking/locktorture.c
96
static bool cpumask_nonempty(cpumask_var_t mask)
kernel/padata.c
661
cpumask_var_t pcpumask,
kernel/padata.c
662
cpumask_var_t cbcpumask)
kernel/padata.c
700
cpumask_var_t cpumask)
kernel/padata.c
859
cpumask_var_t new_cpumask;
kernel/power/energy_model.c
853
cpumask_var_t cpu_done_mask;
kernel/printk/printk_ringbuffer_kunit_test.c
229
static void prbtest_add_cpumask_cleanup(struct kunit *test, cpumask_var_t mask)
kernel/printk/printk_ringbuffer_kunit_test.c
259
cpumask_var_t test_cpus;
kernel/rcu/tasks.h
755
cpumask_var_t cm;
kernel/rcu/tree.c
4163
cpumask_var_t affinity;
kernel/rcu/tree_nocb.h
17
static cpumask_var_t rcu_nocb_mask; /* CPUs to have callbacks offloaded. */
kernel/sched/core.c
3181
cpumask_var_t new_mask;
kernel/sched/cpudeadline.h
16
cpumask_var_t free_cpus;
kernel/sched/cpupri.h
15
cpumask_var_t mask;
kernel/sched/deadline.c
2713
static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask_dl);
kernel/sched/debug.c
639
static cpumask_var_t sd_sysctl_cpus;
kernel/sched/ext.c
37
static cpumask_var_t scx_bypass_lb_donee_cpumask;
kernel/sched/ext.c
38
static cpumask_var_t scx_bypass_lb_resched_cpumask;
kernel/sched/ext_idle.c
33
cpumask_var_t cpu;
kernel/sched/ext_idle.c
34
cpumask_var_t smt;
kernel/sched/ext_idle.c
51
static DEFINE_PER_CPU(cpumask_var_t, local_idle_cpumask);
kernel/sched/ext_idle.c
52
static DEFINE_PER_CPU(cpumask_var_t, local_llc_idle_cpumask);
kernel/sched/ext_idle.c
53
static DEFINE_PER_CPU(cpumask_var_t, local_numa_idle_cpumask);
kernel/sched/fair.c
7206
static DEFINE_PER_CPU(cpumask_var_t, load_balance_mask);
kernel/sched/fair.c
7207
static DEFINE_PER_CPU(cpumask_var_t, select_rq_mask);
kernel/sched/fair.c
7208
static DEFINE_PER_CPU(cpumask_var_t, should_we_balance_tmpmask);
kernel/sched/fair.c
7213
cpumask_var_t idle_cpus_mask;
kernel/sched/isolation.c
197
cpumask_var_t housekeeping_staging)
kernel/sched/isolation.c
207
cpumask_var_t non_housekeeping_mask, housekeeping_staging;
kernel/sched/membarrier.c
253
cpumask_var_t tmpmask;
kernel/sched/membarrier.c
318
cpumask_var_t tmpmask;
kernel/sched/membarrier.c
441
cpumask_var_t tmpmask;
kernel/sched/rt.c
1761
static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask);
kernel/sched/sched.h
1011
cpumask_var_t dlo_mask;
kernel/sched/sched.h
1042
cpumask_var_t rto_mask;
kernel/sched/sched.h
1339
cpumask_var_t scratch_mask;
kernel/sched/sched.h
804
cpumask_var_t cpus_to_kick;
kernel/sched/sched.h
805
cpumask_var_t cpus_to_kick_if_idle;
kernel/sched/sched.h
806
cpumask_var_t cpus_to_preempt;
kernel/sched/sched.h
807
cpumask_var_t cpus_to_wait;
kernel/sched/sched.h
808
cpumask_var_t cpus_to_sync;
kernel/sched/sched.h
994
cpumask_var_t span;
kernel/sched/sched.h
995
cpumask_var_t online;
kernel/sched/syscalls.c
1133
cpumask_var_t cpus_allowed, new_mask;
kernel/sched/syscalls.c
1259
cpumask_var_t new_mask;
kernel/sched/syscalls.c
1305
cpumask_var_t mask;
kernel/sched/topology.c
21
static cpumask_var_t sched_domains_tmpmask;
kernel/sched/topology.c
22
static cpumask_var_t sched_domains_tmpmask2;
kernel/sched/topology.c
2707
static cpumask_var_t *doms_cur;
kernel/sched/topology.c
2720
static cpumask_var_t fallback_doms;
kernel/sched/topology.c
2732
cpumask_var_t *alloc_sched_domains(unsigned int ndoms)
kernel/sched/topology.c
2735
cpumask_var_t *doms;
kernel/sched/topology.c
2749
void free_sched_domains(cpumask_var_t doms[], unsigned int ndoms)
kernel/sched/topology.c
2845
static void partition_sched_domains_locked(int ndoms_new, cpumask_var_t doms_new[],
kernel/sched/topology.c
2941
void partition_sched_domains(int ndoms_new, cpumask_var_t doms_new[],
kernel/smp.c
42
cpumask_var_t cpumask;
kernel/smp.c
43
cpumask_var_t cpumask_ipi;
kernel/taskstats.c
452
cpumask_var_t mask;
kernel/taskstats.c
468
cpumask_var_t mask;
kernel/time/hrtimer.c
951
cpumask_var_t mask;
kernel/time/tick-broadcast.c
28
static cpumask_var_t tick_broadcast_mask __cpumask_var_read_mostly;
kernel/time/tick-broadcast.c
29
static cpumask_var_t tick_broadcast_on __cpumask_var_read_mostly;
kernel/time/tick-broadcast.c
30
static cpumask_var_t tmpmask __cpumask_var_read_mostly;
kernel/time/tick-broadcast.c
607
static cpumask_var_t tick_broadcast_oneshot_mask __cpumask_var_read_mostly;
kernel/time/tick-broadcast.c
608
static cpumask_var_t tick_broadcast_pending_mask __cpumask_var_read_mostly;
kernel/time/tick-broadcast.c
609
static cpumask_var_t tick_broadcast_force_mask __cpumask_var_read_mostly;
kernel/time/tick-sched.c
337
cpumask_var_t tick_nohz_full_mask;
kernel/time/tick-sched.c
625
void __init tick_nohz_full_setup(cpumask_var_t cpumask)
kernel/time/timer_migration.c
1559
cpumask_var_t cpumask __free(free_cpumask_var) = CPUMASK_VAR_NULL;
kernel/time/timer_migration.c
1616
cpumask_var_t cpumask __free(free_cpumask_var) = CPUMASK_VAR_NULL;
kernel/time/timer_migration.c
433
static cpumask_var_t tmigr_available_cpumask;
kernel/torture.c
482
static cpumask_var_t shuffle_tmp_mask;
kernel/trace/preemptirq_delay_test.c
120
cpumask_var_t cpu_mask;
kernel/trace/ring_buffer.c
576
cpumask_var_t cpumask;
kernel/trace/trace.c
118
cpumask_var_t __read_mostly tracing_buffer_mask;
kernel/trace/trace.c
4343
cpumask_var_t tracing_cpumask_new)
kernel/trace/trace.c
4385
cpumask_var_t tracing_cpumask_new;
kernel/trace/trace.h
2163
cpumask_var_t tracing_cpumask_new);
kernel/trace/trace.h
420
cpumask_var_t tracing_cpumask; /* only trace on set CPUs */
kernel/trace/trace.h
422
cpumask_var_t pipe_cpumask;
kernel/trace/trace.h
773
extern cpumask_var_t __read_mostly tracing_buffer_mask;
kernel/trace/trace_boot.c
70
cpumask_var_t new_mask;
kernel/trace/trace_osnoise.c
2322
cpumask_var_t osnoise_cpumask_new;
kernel/workqueue.c
396
cpumask_var_t *pod_cpus; /* pod -> cpus */
kernel/workqueue.c
455
static cpumask_var_t wq_online_cpumask;
kernel/workqueue.c
458
static cpumask_var_t wq_unbound_cpumask;
kernel/workqueue.c
461
static cpumask_var_t wq_requested_unbound_cpumask;
kernel/workqueue.c
464
static cpumask_var_t wq_isolated_cpumask;
kernel/workqueue.c
5362
const cpumask_var_t unbound_cpumask)
kernel/workqueue.c
6973
static int workqueue_apply_unbound_cpumask(const cpumask_var_t unbound_cpumask)
kernel/workqueue.c
7039
cpumask_var_t cpumask;
kernel/workqueue.c
7372
static int workqueue_set_unbound_cpumask(cpumask_var_t cpumask)
kernel/workqueue.c
7395
struct device_attribute *attr, char *buf, cpumask_var_t mask)
kernel/workqueue.c
7429
cpumask_var_t cpumask;
lib/cpu_rmap.c
170
cpumask_var_t update_mask;
lib/cpumask.c
28
bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node)
lib/cpumask.c
52
void __init alloc_bootmem_cpumask_var(cpumask_var_t *mask)
lib/cpumask.c
63
void free_cpumask_var(cpumask_var_t mask)
lib/cpumask.c
73
void __init free_bootmem_cpumask_var(cpumask_var_t mask)
lib/group_cpus.c
229
cpumask_var_t *node_to_cpumask,
lib/group_cpus.c
292
cpumask_var_t msk,
lib/group_cpus.c
370
cpumask_var_t nmsk;
lib/group_cpus.c
405
cpumask_var_t *node_to_cpumask,
lib/group_cpus.c
45
static cpumask_var_t *alloc_node_to_cpumask(void)
lib/group_cpus.c
47
cpumask_var_t *masks;
lib/group_cpus.c
491
cpumask_var_t *node_to_cpumask;
lib/group_cpus.c
492
cpumask_var_t nmsk, npresmsk;
lib/group_cpus.c
50
masks = kzalloc_objs(cpumask_var_t, nr_node_ids);
lib/group_cpus.c
68
static void free_node_to_cpumask(cpumask_var_t *masks)
lib/group_cpus.c
77
static void build_node_to_cpumask(cpumask_var_t *masks)
lib/group_cpus.c
85
static int get_nodes_in_cpumask(cpumask_var_t *node_to_cpumask,
net/core/dev.c
3363
cpumask_var_t cpus;
net/core/net-sysfs.c
1036
cpumask_var_t mask;
net/core/net-sysfs.c
1806
cpumask_var_t mask;
net/core/net-sysfs.c
962
cpumask_var_t mask;
net/core/net-sysfs.c
982
cpumask_var_t mask)
net/core/sysctl_net_core.c
215
cpumask_var_t mask;
net/iucv/iucv.c
656
cpumask_var_t cpumask;
net/netfilter/ipvs/ip_vs_ctl.c
1931
cpumask_var_t *valp = table->data;
net/netfilter/ipvs/ip_vs_ctl.c
1932
cpumask_var_t newmask;
net/netfilter/ipvs/ip_vs_ctl.c
1970
cpumask_var_t *valp = table->data;
rust/helpers/cpumask.c
66
bool rust_helper_alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
rust/helpers/cpumask.c
72
bool rust_helper_zalloc_cpumask_var(cpumask_var_t *mask, gfp_t flags)
rust/helpers/cpumask.c
79
void rust_helper_free_cpumask_var(cpumask_var_t mask)
virt/kvm/kvm_main.c
156
static DEFINE_PER_CPU(cpumask_var_t, cpu_kick_mask);