arch/alpha/kernel/irq.c
60
!cpumask_test_cpu(cpu, irq_default_affinity))
arch/alpha/kernel/sys_dp264.c
142
if (cpumask_test_cpu(cpu, &affinity))
arch/alpha/kernel/sys_titan.c
140
if (cpumask_test_cpu(cpu, &affinity))
arch/alpha/kernel/sys_titan.c
87
if (!cpumask_test_cpu(0, &cpm)) dim0 = &dummy;
arch/alpha/kernel/sys_titan.c
88
if (!cpumask_test_cpu(1, &cpm)) dim1 = &dummy;
arch/alpha/kernel/sys_titan.c
89
if (!cpumask_test_cpu(2, &cpm)) dim2 = &dummy;
arch/alpha/kernel/sys_titan.c
90
if (!cpumask_test_cpu(3, &cpm)) dim3 = &dummy;
arch/arc/kernel/smp.c
72
if (!cpumask_test_cpu(0, &cpumask))
arch/arm/include/asm/cacheflush.h
220
if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
arch/arm/include/asm/cacheflush.h
229
if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm)))
arch/arm/include/asm/cacheflush.h
239
if (!mm || cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) {
arch/arm/include/asm/mmu_context.h
130
!cpumask_test_cpu(cpu, mm_cpumask(next)))
arch/arm/include/asm/tlbflush.h
367
if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm))) {
arch/arm/include/asm/tlbflush.h
421
cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm))) {
arch/arm/mach-omap2/cpuidle44xx.c
133
if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
arch/arm/mach-omap2/cpuidle44xx.c
191
if (dev->cpu == 0 && cpumask_test_cpu(1, cpu_online_mask)) {
arch/arm/mach-tegra/platsmp.c
135
if (cpumask_test_cpu(cpu, &tegra_cpu_init_mask)) {
arch/arm/mach-tegra/platsmp.c
93
if (cpumask_test_cpu(cpu, &tegra_cpu_init_mask)) {
arch/arm/mm/flush.c
163
if (cpumask_test_cpu(smp_processor_id(), mm_cpumask(vma->vm_mm)))
arch/arm64/kernel/cpufeature.c
2044
return cpumask_test_cpu(cpu, &amu_cpus);
arch/arm64/kernel/cpufeature.c
4032
if (cpumask_test_cpu(0, cpu_32bit_el0_mask) == cpu_32bit)
arch/arm64/kernel/topology.c
166
cpumask_test_cpu(cpu, amu_fie_cpus);
arch/arm64/kernel/topology.c
319
cpumask_test_cpu(cpu, amu_fie_cpus))
arch/arm64/kernel/topology.c
444
if ((cpu >= nr_cpu_ids) || !cpumask_test_cpu(cpu, cpu_present_mask))
arch/arm64/kernel/topology.c
67
if ((cpu >= nr_cpu_ids) || !cpumask_test_cpu(cpu, cpu_present_mask))
arch/arm64/kvm/arm.c
706
if (!cpumask_test_cpu(cpu, vcpu->kvm->arch.supported_cpus))
arch/arm64/kvm/pmu-emul.c
837
if (cpumask_test_cpu(cpu, &pmu->supported_cpus))
arch/csky/abiv2/cacheflush.c
44
if (cpumask_test_cpu(cpu, mask)) {
arch/loongarch/kernel/machine_kexec.c
170
if (!cpumask_test_cpu(cpu, &cpus_in_crash))
arch/mips/bcm63xx/irq.c
60
enable &= cpumask_test_cpu(cpu, m);
arch/mips/bcm63xx/irq.c
62
enable &= cpumask_test_cpu(cpu, irq_data_get_affinity_mask(d));
arch/mips/cavium-octeon/octeon-irq.c
1821
if (cpumask_test_cpu(cpu, dest) && enable_one) {
arch/mips/cavium-octeon/octeon-irq.c
287
} else if (cpumask_test_cpu(cpu, cpu_online_mask)) {
arch/mips/cavium-octeon/octeon-irq.c
773
if (!cpumask_test_cpu(cpu, mask))
arch/mips/cavium-octeon/octeon-irq.c
826
if (cpumask_test_cpu(cpu, dest) && enable_one) {
arch/mips/cavium-octeon/octeon-irq.c
871
if (cpumask_test_cpu(cpu, dest) && enable_one) {
arch/mips/cavium-octeon/octeon-irq.c
884
if (cpumask_test_cpu(cpu, dest) && enable_one) {
arch/mips/cavium-octeon/octeon-irq.c
915
if (cpumask_test_cpu(cpu, dest) && enable_one) {
arch/mips/include/asm/mmu_context.h
218
} else if (cpumask_test_cpu(cpu, mm_cpumask(mm))) {
arch/mips/kernel/crash.c
43
if (!cpumask_test_cpu(cpu, &cpus_in_crash))
arch/mips/kernel/smp-bmips.c
214
if (cpumask_test_cpu(cpu, &bmips_booted_mask)) {
arch/mips/kernel/smp.c
205
while (!cpumask_test_cpu(cpu, &cpu_coherent_mask)) {
arch/mips/mm/context.c
229
if (cpumask_test_cpu(cpu, &tlb_flush_pending)) {
arch/parisc/kernel/irq.c
517
!cpumask_test_cpu(smp_processor_id(), &dest)) {
arch/powerpc/include/asm/book3s/64/mmu.h
260
if (cpumask_test_cpu(cpu, mm_cpumask(mm))) { \
arch/powerpc/include/asm/dbell.h
143
if (cpumask_test_cpu(cpu, cpu_sibling_mask(this_cpu))) {
arch/powerpc/include/asm/tlb.h
64
return cpumask_test_cpu(smp_processor_id(), mm_cpumask(mm));
arch/powerpc/kernel/cacheinfo.c
199
WARN_ONCE(cpumask_test_cpu(cpu, &next->shared_cpu_map),
arch/powerpc/kernel/cacheinfo.c
895
WARN_ONCE(!cpumask_test_cpu(cpu, &cache->shared_cpu_map),
arch/powerpc/kernel/setup-common.c
983
VM_WARN_ON(cpumask_test_cpu(smp_processor_id(), mm_cpumask(&init_mm)));
arch/powerpc/kernel/smp.c
1650
VM_WARN_ON(cpumask_test_cpu(smp_processor_id(), mm_cpumask(&init_mm)));
arch/powerpc/kernel/smp.c
1802
VM_WARN_ON_ONCE(!cpumask_test_cpu(cpu, mm_cpumask(&init_mm)));
arch/powerpc/kernel/smp.c
456
if (cpumask_test_cpu(me, &nmi_ipi_pending_mask)) {
arch/powerpc/kernel/stacktrace.c
175
while (cpumask_test_cpu(cpu, mask) && delay_us) {
arch/powerpc/kernel/watchdog.c
214
if (cpumask_test_cpu(cpu, &wd_smp_cpus_pending))
arch/powerpc/kernel/watchdog.c
219
if (!cpumask_test_cpu(c, &wd_smp_cpus_pending))
arch/powerpc/kernel/watchdog.c
268
if (!cpumask_test_cpu(cpu, &wd_smp_cpus_pending)) {
arch/powerpc/kernel/watchdog.c
269
if (unlikely(cpumask_test_cpu(cpu, &wd_smp_cpus_stuck))) {
arch/powerpc/kernel/watchdog.c
381
if (!cpumask_test_cpu(cpu, &wd_cpus_enabled))
arch/powerpc/kernel/watchdog.c
395
if (cpumask_test_cpu(cpu, &wd_smp_cpus_stuck)) {
arch/powerpc/kernel/watchdog.c
449
if (!cpumask_test_cpu(cpu, &watchdog_cpumask))
arch/powerpc/kernel/watchdog.c
465
if (!cpumask_test_cpu(cpu, &watchdog_cpumask))
arch/powerpc/kernel/watchdog.c
482
if (cpumask_test_cpu(cpu, &wd_cpus_enabled)) {
arch/powerpc/kernel/watchdog.c
490
if (!cpumask_test_cpu(cpu, &watchdog_cpumask))
arch/powerpc/kernel/watchdog.c
519
if (!cpumask_test_cpu(cpu, &wd_cpus_enabled))
arch/powerpc/kexec/crash.c
88
if (!cpumask_test_cpu(cpu, &cpus_state_saved)) {
arch/powerpc/kvm/book3s_hv_builtin.c
622
if (cpumask_test_cpu(pcpu, &kvm->arch.need_tlb_flush)) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
465
if (likely(!cpumask_test_cpu(pcpu, need_tlb_flush)))
arch/powerpc/kvm/book3s_hv_p9_entry.c
480
if (!cpumask_test_cpu(i, need_tlb_flush)) {
arch/powerpc/mm/book3s64/radix_tlb.c
703
if (cpumask_test_cpu(cpu, mm_cpumask(mm))) {
arch/powerpc/mm/book3s64/radix_tlb.c
772
if (active_cpus == 1 && cpumask_test_cpu(cpu, mm_cpumask(mm))) {
arch/powerpc/mm/book3s64/radix_tlb.c
825
if (cpumask_test_cpu(cpu, mm_cpumask(mm)))
arch/powerpc/mm/mmu_context.c
106
VM_WARN_ON_ONCE(!cpumask_test_cpu(cpu, mm_cpumask(prev)));
arch/powerpc/mm/mmu_context.c
50
if (!cpumask_test_cpu(cpu, mm_cpumask(next))) {
arch/powerpc/mm/numa.c
1074
if (cpumask_test_cpu(cpu,
arch/powerpc/mm/numa.c
148
if (!(cpumask_test_cpu(cpu, node_to_cpumask_map[node]))) {
arch/powerpc/mm/numa.c
159
if (cpumask_test_cpu(cpu, node_to_cpumask_map[node])) {
arch/powerpc/platforms/pseries/rtas-fadump.c
409
if (fdh && !cpumask_test_cpu(cpu, &fdh->cpu_mask)) {
arch/powerpc/platforms/pseries/smp.c
95
if (cpumask_test_cpu(lcpu, of_spin_mask))
arch/powerpc/sysdev/xive/common.c
738
cpumask_test_cpu(xd->target, cpumask))
arch/powerpc/xmon/xmon.c
1341
if (cpumask_test_cpu(cpu, &cpus_in_xmon)) {
arch/powerpc/xmon/xmon.c
1358
if (!cpumask_test_cpu(cpu, &cpus_in_xmon)) {
arch/powerpc/xmon/xmon.c
512
if (cpumask_test_cpu(cpu, &cpus_in_xmon)) {
arch/powerpc/xmon/xmon.c
812
if (in_xmon && !cpumask_test_cpu(smp_processor_id(), &cpus_in_xmon))
arch/riscv/mm/cacheflush.c
193
stale_cpu = cpumask_test_cpu(cpu, mask);
arch/s390/kernel/topology.c
104
if (!cpumask_test_cpu(cpu, &cpu_setup_mask))
arch/s390/kernel/topology.c
112
if (cpumask_test_cpu(cpu, &cpu_setup_mask))
arch/s390/kernel/topology.c
72
if (!cpumask_test_cpu(cpu, &cpu_setup_mask))
arch/s390/kernel/topology.c
78
if (cpumask_test_cpu(cpu, &info->mask)) {
arch/sparc/include/asm/mmu_context_64.h
130
if (!ctx_valid || !cpumask_test_cpu(cpu, mm_cpumask(mm))) {
arch/sparc/kernel/leon_smp.c
409
if (cpumask_test_cpu(i, &mask)) {
arch/sparc/kernel/leon_smp.c
423
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/leon_smp.c
432
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/leon_smp.c
98
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/mdesc.c
1124
if (!cpumask_test_cpu(cpuid, mask))
arch/sparc/kernel/process_64.c
255
if (cpumask_test_cpu(this_cpu, mask) && this_cpu != exclude_cpu)
arch/sparc/kernel/smp_64.c
136
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/smp_64.c
1364
if (!cpumask_test_cpu(cpu, &smp_commenced_mask))
arch/sparc/kernel/smp_64.c
1368
if (cpumask_test_cpu(cpu, &smp_commenced_mask)) {
arch/sparc/kernel/sun4d_smp.c
103
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/sun4d_smp.c
321
if (cpumask_test_cpu(i, &mask)) {
arch/sparc/kernel/sun4d_smp.c
334
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/sun4d_smp.c
342
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/sun4m_smp.c
197
if (cpumask_test_cpu(i, &mask)) {
arch/sparc/kernel/sun4m_smp.c
213
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/sun4m_smp.c
221
if (!cpumask_test_cpu(i, &mask))
arch/sparc/kernel/sun4m_smp.c
66
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/x86/events/core.c
2521
if (!cpumask_test_cpu(event->cpu, &pmu->supported_cpus))
arch/x86/events/intel/core.c
6232
*ret = !cpumask_test_cpu(cpu, &hpmu->supported_cpus);
arch/x86/hyperv/hv_apic.c
185
(exclude_self && weight == 1 && cpumask_test_cpu(this_cpu, mask)))
arch/x86/include/asm/topology.h
242
return cpumask_test_cpu(cpu, cpu_primary_thread_mask);
arch/x86/kernel/apic/ipi.c
105
if (WARN_ON_ONCE(!cpumask_test_cpu(cpu, &cpus_booted_once_mask)))
arch/x86/kernel/apic/ipi.c
90
if (cpumask_test_cpu(cpu, mask))
arch/x86/kernel/apic/vector.c
254
if (vector && cpu_online(cpu) && cpumask_test_cpu(cpu, dest))
arch/x86/kernel/apic/vector.c
339
if (apicd->vector && cpumask_test_cpu(apicd->cpu, vector_searchmask))
arch/x86/kernel/cpu/mce/core.c
2739
if (!cpumask_test_cpu(cpu, mce_device_initialized))
arch/x86/kernel/cpu/mce/inject.c
182
if (!cpumask_test_cpu(cpu, mce_inject_cpumask))
arch/x86/kernel/cpu/mce/inject.c
197
if (cpumask_test_cpu(cpu, mce_inject_cpumask) &&
arch/x86/kernel/cpu/resctrl/ctrlmondata.c
30
if (!cpumask_test_cpu(smp_processor_id(), &d->hdr.cpu_mask))
arch/x86/kvm/svm/sev.c
3579
if (!cpumask_test_cpu(cpu, to_kvm_sev_info(kvm)->have_run_cpus))
arch/x86/lib/msr-smp.c
111
if (cpumask_test_cpu(this_cpu, mask))
arch/x86/mm/tlb.c
1747
} else if (cpumask_test_cpu(cpu, &batch->cpumask)) {
arch/x86/mm/tlb.c
859
!cpumask_test_cpu(cpu, mm_cpumask(next))))
arch/x86/mm/tlb.c
935
if (next != &init_mm && !cpumask_test_cpu(cpu, mm_cpumask(next)))
arch/x86/platform/uv/uv_nmi.c
661
cpumask_test_cpu(0, uv_nmi_cpu_mask))
arch/x86/platform/uv/uv_nmi.c
810
if (cpumask_test_cpu(tcpu, uv_nmi_cpu_mask))
arch/xtensa/kernel/irq.c
180
if (!cpumask_test_cpu(cpu, mask))
block/blk-mq.c
2381
if (async || !cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)) {
block/blk-mq.c
4219
if (cpumask_test_cpu(i, hctx->cpumask))
drivers/acpi/processor_perflib.c
655
if (cpumask_test_cpu(i, covered_cpus))
drivers/acpi/processor_throttling.c
97
if (cpumask_test_cpu(i, covered_cpus))
drivers/base/arch_topology.c
40
if (cpumask_test_cpu(i, &scale_freq_counters_mask))
drivers/clocksource/arm_arch_timer.c
897
return cpumask_test_cpu(raw_smp_processor_id(), &evtstrm_available);
drivers/cpufreq/cpufreq-dt.c
44
if (cpumask_test_cpu(cpu, priv->cpus))
drivers/cpufreq/cpufreq.c
1176
if (cpumask_test_cpu(cpu, policy->cpus))
drivers/cpufreq/cpufreq.c
1604
WARN_ON(!cpumask_test_cpu(cpu, policy->related_cpus));
drivers/cpufreq/cpufreq.c
197
return policy && cpumask_test_cpu(cpu, policy->cpus) ? policy : NULL;
drivers/cpufreq/cpufreq_ondemand.c
406
if (cpumask_test_cpu(cpu, done))
drivers/cpufreq/intel_pstate.c
1950
if (!cpumask_test_cpu(this_cpu, &hwp_intr_enable_mask))
drivers/cpufreq/mediatek-cpufreq.c
71
if (cpumask_test_cpu(cpu, &info->cpus))
drivers/cpufreq/powernv-cpufreq.c
687
if (!cpumask_test_cpu(raw_smp_processor_id(), policy->cpus)) {
drivers/cpuidle/coupled.c
429
if (!cpumask_test_cpu(cpu, &cpuidle_coupled_poke_pending))
drivers/cpuidle/coupled.c
433
while (cpumask_test_cpu(cpu, &cpuidle_coupled_poke_pending))
drivers/cpuidle/coupled.c
515
!cpumask_test_cpu(dev->cpu, &cpuidle_coupled_poked)) {
drivers/crypto/caam/qi.c
442
if (!cpumask_test_cpu(*cpu, cpus)) {
drivers/dma/dmaengine.c
308
cpumask_test_cpu(cpu, cpumask_of_node(node));
drivers/gpio/gpio-realtek-otto.c
292
if (cpumask_test_cpu(cpu, dest))
drivers/hv/vmbus_drv.c
1774
if (!cpumask_test_cpu(target_cpu, housekeeping_cpumask(HK_TYPE_MANAGED_IRQ)))
drivers/hwtracing/coresight/coresight-etm-perf.c
521
if (!cpumask_test_cpu(cpu, &event_data->mask))
drivers/hwtracing/coresight/coresight-etm-perf.c
543
if (!cpumask_test_cpu(cpu, &event_data->aux_hwid_done)) {
drivers/hwtracing/coresight/coresight-etm-perf.c
668
!cpumask_test_cpu(cpu, &event_data->mask)) {
drivers/hwtracing/coresight/coresight-trbe.c
1112
if (!cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/coresight/coresight-trbe.c
1390
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/coresight/coresight-trbe.c
1392
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/coresight/coresight-trbe.c
1421
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus)) {
drivers/hwtracing/coresight/coresight-trbe.c
1429
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/coresight/coresight-trbe.c
1431
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/coresight/coresight-trbe.c
1444
if (cpumask_test_cpu(cpu, &drvdata->supported_cpus))
drivers/hwtracing/ptt/hisi_ptt.c
1168
if (!cpumask_test_cpu(cpu, cpumask_of_node(dev_to_node(&hisi_ptt->pdev->dev))))
drivers/infiniband/core/uverbs_std_types_dmah.c
50
if (!cpumask_test_cpu(dmah->cpu_id, current->cpus_ptr)) {
drivers/infiniband/hw/hfi1/affinity.c
769
if (cpumask_test_cpu(old_cpu, &other_msix->mask))
drivers/infiniband/hw/hfi1/sdma.c
932
if (cpumask_test_cpu(cpu, &sde->cpu_mask)) {
drivers/infiniband/hw/hfi1/sdma.c
997
if (cpumask_test_cpu(cpu, mask))
drivers/irqchip/irq-gic-v3-its.c
3966
if (cpumask_test_cpu(from, mask_val) &&
drivers/irqchip/irq-gic-v3-its.c
3967
cpumask_test_cpu(from, table_mask))
drivers/irqchip/irq-gic-v3.c
1314
if (cpumask_test_cpu(cpu, &broken_rdists))
drivers/irqchip/irq-gic-v3.c
666
!cpumask_test_cpu(raw_smp_processor_id(),
drivers/irqchip/irq-loongarch-avec.c
113
if (cpu_online(adata->cpu) && cpumask_test_cpu(adata->cpu, dest))
drivers/irqchip/irq-riscv-imsic-platform.c
126
if (cpumask_test_cpu(old_vec->cpu, mask_val))
drivers/net/ethernet/google/gve/gve_main.c
349
return cpumask_test_cpu(cpu_curr, aff_mask);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2820
if (!cpumask_test_cpu(cpu_id, &q_vector->affinity_mask)) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1653
if (!cpumask_test_cpu(cpu_id,
drivers/net/ethernet/mellanox/mlx4/en_rx.c
1003
if (likely(cpumask_test_cpu(cpu_curr, cq->aff_mask)))
drivers/net/ethernet/mellanox/mlx5/core/en_txrx.c
46
return cpumask_test_cpu(current_cpu, c->aff_mask);
drivers/net/ethernet/sfc/falcon/efx.c
1331
if (!cpumask_test_cpu(cpu, thread_mask)) {
drivers/perf/arm_cspmu/arm_cspmu.c
1296
if (!cpumask_test_cpu(cpu, &cspmu->associated_cpus))
drivers/perf/arm_cspmu/arm_cspmu.c
668
if (!cpumask_test_cpu(event->cpu, &cspmu->associated_cpus)) {
drivers/perf/arm_cspmu/arm_cspmu.c
878
if (WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(),
drivers/perf/arm_dsu_pmu.c
234
if (WARN_ON(!cpumask_test_cpu(smp_processor_id(),
drivers/perf/arm_dsu_pmu.c
260
if (WARN_ON(!cpumask_test_cpu(smp_processor_id(),
drivers/perf/arm_dsu_pmu.c
429
if (WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(),
drivers/perf/arm_dsu_pmu.c
549
if (!cpumask_test_cpu(event->cpu, &dsu_pmu->associated_cpus)) {
drivers/perf/arm_dsu_pmu.c
807
if (!cpumask_test_cpu(cpu, &dsu_pmu->associated_cpus))
drivers/perf/arm_pmu.c
347
if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus))
drivers/perf/arm_pmu.c
524
!cpumask_test_cpu(event->cpu, &armpmu->supported_cpus))
drivers/perf/arm_pmu.c
540
if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus))
drivers/perf/arm_pmu.c
552
if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus))
drivers/perf/arm_pmu.c
566
return !cpumask_test_cpu(cpu, &armpmu->supported_cpus);
drivers/perf/arm_pmu.c
722
if (!cpumask_test_cpu(cpu, &pmu->supported_cpus))
drivers/perf/arm_pmu.c
739
if (!cpumask_test_cpu(cpu, &pmu->supported_cpus))
drivers/perf/arm_pmu.c
788
if (!cpumask_test_cpu(smp_processor_id(), &armpmu->supported_cpus))
drivers/perf/arm_spe_pmu.c
1275
if (!cpumask_test_cpu(cpu, &spe_pmu->supported_cpus))
drivers/perf/arm_spe_pmu.c
1287
if (!cpumask_test_cpu(cpu, &spe_pmu->supported_cpus))
drivers/perf/arm_spe_pmu.c
815
!cpumask_test_cpu(event->cpu, &spe_pmu->supported_cpus))
drivers/perf/arm_spe_pmu.c
973
if (!cpumask_test_cpu(cpu, &spe_pmu->supported_cpus))
drivers/perf/hisilicon/hisi_uncore_pmu.c
524
cpumask_test_cpu(hisi_pmu->on_cpu, &hisi_pmu->associated_cpus))
drivers/resctrl/mpam_devices.c
1539
if (cpumask_test_cpu(cpu, &msc->accessibility))
drivers/resctrl/mpam_devices.c
160
WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(), &msc->accessibility));
drivers/resctrl/mpam_devices.c
1625
if (!cpumask_test_cpu(cpu, &msc->accessibility))
drivers/resctrl/mpam_devices.c
1651
if (!cpumask_test_cpu(cpu, &msc->accessibility))
drivers/resctrl/mpam_devices.c
1681
if (!cpumask_test_cpu(cpu, &msc->accessibility))
drivers/resctrl/mpam_devices.c
176
WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(), &msc->accessibility));
drivers/resctrl/mpam_devices.c
2245
WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(),
drivers/resctrl/mpam_devices.c
923
WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(), &msc->accessibility));
drivers/resctrl/mpam_devices.c
946
WARN_ON_ONCE(!cpumask_test_cpu(smp_processor_id(), &msc->accessibility));
drivers/scsi/lpfc/lpfc_init.c
12844
if (!cpumask_test_cpu(cpu, orig_mask))
drivers/scsi/lpfc/lpfc_init.c
12866
if (cpumask_test_cpu(cpu, aff_mask))
drivers/scsi/storvsc_drv.c
1420
if (cpumask_test_cpu(q_num, &stor_device->alloced_cpus))
drivers/scsi/storvsc_drv.c
1427
if (cpumask_test_cpu(tgt_cpu, node_mask))
drivers/scsi/storvsc_drv.c
1440
if (!cpumask_test_cpu(tgt_cpu, node_mask))
drivers/scsi/storvsc_drv.c
1488
if (!cpumask_test_cpu(tgt_cpu, node_mask))
drivers/scsi/storvsc_drv.c
653
if (stor_device->stor_chns[cpu] && !cpumask_test_cpu(
drivers/sh/intc/chip.c
25
if (!cpumask_test_cpu(cpu, irq_data_get_affinity_mask(data)))
drivers/sh/intc/chip.c
53
if (!cpumask_test_cpu(cpu, irq_data_get_affinity_mask(data)))
drivers/soc/fsl/qbman/qman.c
1739
WARN_ON(!cpumask_test_cpu(cpu, &affine_mask));
drivers/virt/nitro_enclaves/ne_misc_dev.c
1103
if (!cpumask_test_cpu(cpu, ne_enclave->vcpu_ids)) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
253
if (cpumask_test_cpu(0, cpu_pool)) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
262
if (cpumask_test_cpu(cpu_sibling, cpu_pool)) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
279
if (!cpumask_test_cpu(cpu_sibling, cpu_pool)) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
481
if (cpumask_test_cpu(cpu, ne_enclave->vcpu_ids))
drivers/virt/nitro_enclaves/ne_misc_dev.c
631
if (cpumask_test_cpu(vcpu_id, ne_cpu_pool.avail_threads_per_core[i])) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
669
if (cpumask_test_cpu(vcpu_id, ne_enclave->threads_per_core[i]))
drivers/watchdog/octeon-wdt-main.c
416
!cpumask_test_cpu(cpu, &irq_enabled_cpus)) {
fs/resctrl/monitor.c
454
if (!cpumask_test_cpu(cpu, &d->hdr.cpu_mask))
fs/resctrl/monitor.c
492
if (!cpumask_test_cpu(cpu, &rr->ci->shared_cpu_map))
fs/resctrl/monitor.c
633
if (cpumask_test_cpu(cpu, &d->hdr.cpu_mask))
fs/resctrl/rdtgroup.c
4495
if (cpumask_test_cpu(cpu, &d->hdr.cpu_mask))
fs/xfs/xfs_icache.c
2203
if (!cpumask_test_cpu(cpu_nr, &mp->m_inodegc_cpumask))
fs/xfs/xfs_log_cil.c
673
if (!cpumask_test_cpu(cpu_nr, &ctx->cil_pcpmask))
include/linux/cpumask.h
1231
return cpumask_test_cpu(cpu, cpu_online_mask);
include/linux/cpumask.h
1236
return cpumask_test_cpu(cpu, cpu_enabled_mask);
include/linux/cpumask.h
1241
return cpumask_test_cpu(cpu, cpu_possible_mask);
include/linux/cpumask.h
1246
return cpumask_test_cpu(cpu, cpu_present_mask);
include/linux/cpumask.h
1251
return cpumask_test_cpu(cpu, cpu_active_mask);
include/linux/cpumask.h
1256
return cpumask_test_cpu(cpu, cpu_dying_mask);
include/linux/mmu_context.h
29
# define task_cpu_possible(cpu, p) cpumask_test_cpu((cpu), task_cpu_possible_mask(p))
include/linux/tick.h
196
__ret = cpumask_test_cpu((_cpu), tick_nohz_full_mask); \
io_uring/io-wq.c
738
exit_mask = !cpumask_test_cpu(raw_smp_processor_id(),
io_uring/sqpoll.c
505
if (!cpumask_test_cpu(cpu, allowed_mask)) {
kernel/bpf/cpumask.c
200
return cpumask_test_cpu(cpu, (struct cpumask *)cpumask);
kernel/cpu.c
699
return !cpumask_test_cpu(cpu, &cpus_booted_once_mask);
kernel/events/core.c
11139
cpumask_test_cpu(cpu, perf_online_mask)) {
kernel/events/core.c
15053
cpuctx->online = cpumask_test_cpu(cpu, perf_online_mask);
kernel/events/core.c
4792
if (cpumask && cpumask_test_cpu(local_cpu, cpumask))
kernel/irq/chip.c
915
if (cpumask_test_cpu(cpu, action->affinity))
kernel/irq/chip.c
923
bool enabled = cpumask_test_cpu(cpu, desc->percpu_enabled);
kernel/irq/cpuhotplug.c
203
return cpumask_test_cpu(cpu, hk_mask);
kernel/irq/cpuhotplug.c
212
!irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity))
kernel/irq/cpuhotplug.c
50
return cpumask_test_cpu(cpu, m);
kernel/irq/ipi.c
172
if (!ipimask || !cpumask_test_cpu(cpu, ipimask))
kernel/irq/ipi.c
210
if (!cpumask_test_cpu(cpu, ipimask))
kernel/irq/irqdesc.c
790
if (cpumask_test_cpu(smp_processor_id(), m))
kernel/irq/manage.c
2375
return cpumask_test_cpu(smp_processor_id(), scoped_irqdesc->percpu_enabled);
kernel/padata.c
196
if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) {
kernel/padata.c
737
if (cpumask_test_cpu(cpu, cpu_online_mask)) {
kernel/padata.c
752
if (!cpumask_test_cpu(cpu, cpu_online_mask)) {
kernel/padata.c
765
return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) ||
kernel/padata.c
766
cpumask_test_cpu(cpu, pinst->cpumask.cbcpu);
kernel/power/energy_model.c
867
if (cpumask_test_cpu(cpu, cpu_done_mask))
kernel/rcu/tree_nocb.h
1466
if (cpumask_test_cpu(cpu, rcu_nocb_mask))
kernel/sched/core.c
2562
if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask))
kernel/sched/core.c
2595
if (cpumask_test_cpu(task_cpu(p), p->cpus_ptr)) {
kernel/sched/core.c
2875
if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask) ||
kernel/sched/core.c
3063
!cpumask_test_cpu(task_cpu(p), ctx->new_mask))) {
kernel/sched/core.c
3350
if (!cpumask_test_cpu(arg->dst_cpu, arg->src_task->cpus_ptr))
kernel/sched/core.c
3353
if (!cpumask_test_cpu(arg->src_cpu, arg->dst_task->cpus_ptr))
kernel/sched/core.c
3388
if (!cpumask_test_cpu(arg.dst_cpu, arg.src_task->cpus_ptr))
kernel/sched/core.c
3391
if (!cpumask_test_cpu(arg.src_cpu, arg.dst_task->cpus_ptr))
kernel/sched/core.c
3608
if (cpumask_test_cpu(cpu, sched_domain_span(sd))) {
kernel/sched/core.c
3867
if (!cpumask_test_cpu(cpu, p->cpus_ptr))
kernel/sched/core.c
8036
if (!cpumask_test_cpu(target_cpu, p->cpus_ptr))
kernel/sched/core.c
8267
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
kernel/sched/core.c
8279
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
kernel/sched/cpudeadline.c
154
if (cpumask_test_cpu(best_cpu, &p->cpus_mask) &&
kernel/sched/cpufreq.c
72
return cpumask_test_cpu(smp_processor_id(), policy->cpus) ||
kernel/sched/deadline.c
2748
if (cpumask_test_cpu(cpu, later_mask))
kernel/sched/deadline.c
2754
if (!cpumask_test_cpu(this_cpu, later_mask))
kernel/sched/deadline.c
2767
cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
kernel/sched/deadline.c
2870
!cpumask_test_cpu(later_rq->cpu, &task->cpus_mask) ||
kernel/sched/ext.c
1706
WARN_ON_ONCE(!cpumask_test_cpu(cpu_of(dst_rq), p->cpus_ptr));
kernel/sched/ext.c
4111
if (cpumask_test_cpu(cpu, donee_mask))
kernel/sched/ext.c
5682
if (cpumask_test_cpu(cpu, this_scx->cpus_to_preempt)) {
kernel/sched/ext.c
5688
if (cpumask_test_cpu(cpu, this_scx->cpus_to_wait)) {
kernel/sched/ext_idle.c
104
else if (cpumask_test_cpu(cpu, idle_smts))
kernel/sched/ext_idle.c
466
is_prev_allowed = cpumask_test_cpu(prev_cpu, allowed);
kernel/sched/ext_idle.c
551
if (cpumask_test_cpu(cpu, allowed))
kernel/sched/ext_idle.c
565
cpumask_test_cpu(prev_cpu, idle_cpumask(node)->smt) &&
kernel/sched/ext_idle.c
935
if (cpumask_test_cpu(prev_cpu, allowed ?: p->cpus_ptr) &&
kernel/sched/fair.c
10921
local_group = cpumask_test_cpu(this_cpu,
kernel/sched/fair.c
11151
local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(sg));
kernel/sched/fair.c
11768
if (!cpumask_test_cpu(env->dst_cpu, env->cpus))
kernel/sched/fair.c
12061
if (!cpumask_test_cpu(this_cpu, busiest->curr->cpus_ptr)) {
kernel/sched/fair.c
12233
if (cpumask_test_cpu(busiest_cpu, sched_domain_span(sd)))
kernel/sched/fair.c
12717
if (!cpumask_test_cpu(cpu, nohz.idle_cpus_mask))
kernel/sched/fair.c
2211
!cpumask_test_cpu(cpu, env->p->cpus_ptr))
kernel/sched/fair.c
2243
!cpumask_test_cpu(cpu, env->p->cpus_ptr)) {
kernel/sched/fair.c
2357
if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr))
kernel/sched/fair.c
6665
if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask))
kernel/sched/fair.c
7631
if (sched_idle_cpu(cpu) && cpumask_test_cpu(cpu, cpus)) {
kernel/sched/fair.c
7639
if (*idle_cpu == -1 && cpumask_test_cpu(cpu, cpus))
kernel/sched/fair.c
7664
if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
kernel/sched/fair.c
7725
if (!cpumask_test_cpu(cpu, cpus))
kernel/sched/fair.c
7900
cpumask_test_cpu(recent_used_cpu, p->cpus_ptr) &&
kernel/sched/fair.c
8409
while (sd && !cpumask_test_cpu(prev_cpu, sched_domain_span(sd)))
kernel/sched/fair.c
8446
if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
kernel/sched/fair.c
8449
if (!cpumask_test_cpu(cpu, p->cpus_ptr))
kernel/sched/fair.c
8598
cpumask_test_cpu(cpu, p->cpus_ptr))
kernel/sched/fair.c
8608
want_affine = !wake_wide(p) && cpumask_test_cpu(cpu, p->cpus_ptr);
kernel/sched/fair.c
8618
cpumask_test_cpu(prev_cpu, sched_domain_span(tmp))) {
kernel/sched/fair.c
9524
if (!cpumask_test_cpu(env->dst_cpu, p->cpus_ptr)) {
kernel/sched/isolation.c
116
return cpumask_test_cpu(cpu, housekeeping_cpumask(type));
kernel/sched/rt.c
1804
if (cpumask_test_cpu(cpu, lowest_mask))
kernel/sched/rt.c
1811
if (!cpumask_test_cpu(this_cpu, lowest_mask))
kernel/sched/rt.c
1824
cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
kernel/sched/rt.c
1913
!cpumask_test_cpu(lowest_rq->cpu, &task->cpus_mask) ||
kernel/sched/rt.c
2259
cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask))
kernel/sched/sched.h
2804
if (!cpumask_test_cpu(cpu, p->cpus_ptr))
kernel/sched/sched.h
4031
cpumask_test_cpu(cpu, &p->cpus_mask))
kernel/sched/topology.c
1044
if (cpumask_test_cpu(i, covered))
kernel/sched/topology.c
1059
if (!cpumask_test_cpu(i, sched_domain_span(sibling)))
kernel/sched/topology.c
1257
if (cpumask_test_cpu(i, covered))
kernel/sched/topology.c
2534
if (cpumask_test_cpu(id, id_seen)) {
kernel/sched/topology.c
2668
if (!cpumask_test_cpu(i, cpu_map))
kernel/sched/topology.c
334
if (cpumask_test_cpu(cpu, perf_domain_span(pd)))
kernel/sched/topology.c
482
if (cpumask_test_cpu(rq->cpu, old_rd->online))
kernel/sched/topology.c
500
if (cpumask_test_cpu(rq->cpu, cpu_active_mask))
kernel/sched/topology.c
56
if (!cpumask_test_cpu(cpu, sched_domain_span(sd))) {
kernel/sched/topology.c
59
if (group && !cpumask_test_cpu(cpu, sched_group_span(group))) {
kernel/smp.c
762
if (!cpumask_test_cpu(cpu, mask))
kernel/smp.c
863
if ((scf_flags & SCF_RUN_LOCAL) && cpumask_test_cpu(this_cpu, mask) &&
kernel/stop_machine.c
221
is_active = cpumask_test_cpu(cpu, cpumask);
kernel/time/clockevents.c
649
if (cpumask_test_cpu(cpu, dev->cpumask) &&
kernel/time/tick-broadcast.c
284
if (!cpumask_test_cpu(cpu, tick_broadcast_on))
kernel/time/tick-broadcast.c
318
ret = cpumask_test_cpu(cpu, tick_broadcast_mask);
kernel/time/tick-broadcast.c
355
if (cpumask_test_cpu(cpu, mask)) {
kernel/time/tick-broadcast.c
576
return cpumask_test_cpu(smp_processor_id(), tick_broadcast_mask);
kernel/time/tick-broadcast.c
631
return cpumask_test_cpu(smp_processor_id(), tick_broadcast_force_mask);
kernel/time/tick-broadcast.c
672
if (cpumask_test_cpu(smp_processor_id(), tick_broadcast_oneshot_mask)) {
kernel/time/tick-broadcast.c
831
WARN_ON_ONCE(cpumask_test_cpu(cpu, tick_broadcast_pending_mask));
kernel/time/tick-broadcast.c
846
if (cpumask_test_cpu(cpu, tick_broadcast_force_mask)) {
kernel/time/tick-common.c
275
if (!cpumask_test_cpu(cpu, newdev->cpumask))
kernel/time/tick-sched.c
672
if (cpumask_test_cpu(cpu, tick_nohz_full_mask)) {
kernel/trace/ring_buffer.c
1004
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
1054
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
2959
!cpumask_test_cpu(cpu_id, buffer->cpumask))
kernel/trace/ring_buffer.c
4716
if (unlikely(!cpumask_test_cpu(cpu, buffer->cpumask)))
kernel/trace/ring_buffer.c
4863
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5048
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5068
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5088
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5119
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5138
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5158
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5181
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5203
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5223
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5874
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
5956
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6001
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6078
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6242
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6383
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6416
if (!cpumask_test_cpu(cpu, buffer_a->cpumask) ||
kernel/trace/ring_buffer.c
6417
!cpumask_test_cpu(cpu, buffer_b->cpumask))
kernel/trace/ring_buffer.c
6509
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6636
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6894
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
6932
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7081
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7257
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7321
if (WARN_ON(!cpumask_test_cpu(cpu, buffer->cpumask)))
kernel/trace/ring_buffer.c
7339
if (!cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/ring_buffer.c
7484
if (cpumask_test_cpu(cpu, buffer->cpumask))
kernel/trace/trace.c
3473
cpumask_test_cpu(iter->cpu, iter->started))
kernel/trace/trace.c
4357
if (cpumask_test_cpu(cpu, tr->tracing_cpumask) &&
kernel/trace/trace.c
4358
!cpumask_test_cpu(cpu, tracing_cpumask_new)) {
kernel/trace/trace.c
4364
if (!cpumask_test_cpu(cpu, tr->tracing_cpumask) &&
kernel/trace/trace.c
4365
cpumask_test_cpu(cpu, tracing_cpumask_new)) {
kernel/trace/trace.c
5340
if (!cpumask_test_cpu(cpu_id, tracing_buffer_mask))
kernel/trace/trace.c
5775
} else if (!cpumask_test_cpu(cpu, tr->pipe_cpumask)) {
kernel/trace/trace.c
5788
WARN_ON(!cpumask_test_cpu(cpu, tr->pipe_cpumask));
kernel/trace/trace_events_filter.c
681
return cpumask_test_cpu(cpu, mask);
kernel/trace/trace_events_filter.c
692
return cpumask_test_cpu(cpu, mask) &&
kernel/trace/trace_events_filter.c
695
return !cpumask_test_cpu(cpu, mask) ||
kernel/trace/trace_events_filter.c
698
return cpumask_test_cpu(cpu, mask);
kernel/trace/trace_hwlat.c
519
if (!cpumask_test_cpu(cpu, tr->tracing_cpumask))
kernel/trace/trace_osnoise.c
2082
if (!cpumask_test_cpu(cpu, &osnoise_cpumask))
kernel/up.c
49
if ((!cond_func || cond_func(0, info)) && cpumask_test_cpu(0, mask)) {
kernel/watchdog.c
968
if (cpumask_test_cpu(cpu, &watchdog_allowed_mask))
kernel/watchdog.c
975
if (cpumask_test_cpu(cpu, &watchdog_allowed_mask))
kernel/workqueue.c
1287
!cpumask_test_cpu(p->wake_cpu, pool->attrs->__pod_cpumask)) {
kernel/workqueue.c
1604
if (off_cpu >= 0 && !cpumask_test_cpu(off_cpu, effective))
kernel/workqueue.c
2233
if (cpumask_test_cpu(cpu, wq_unbound_cpumask))
kernel/workqueue.c
6723
if (!cpumask_test_cpu(cpu, pool->attrs->cpumask))
kernel/workqueue.c
8124
return cpumask_test_cpu(cpu0, cpu_smt_mask(cpu1));
lib/nmi_backtrace.c
61
if (cpumask_test_cpu(this_cpu, to_cpumask(backtrace_mask)))
lib/nmi_backtrace.c
99
if (cpumask_test_cpu(cpu, to_cpumask(backtrace_mask))) {
mm/vmalloc.c
2394
if (cpumask_test_cpu(i, cpu_online_mask))
net/core/sysctl_net_core.c
232
if (cur && !cpumask_test_cpu(i, mask)) {
net/core/sysctl_net_core.c
235
} else if (!cur && cpumask_test_cpu(i, mask)) {
net/iucv/iucv.c
461
if (cpumask_test_cpu(cpu, &iucv_buffer_cpumask))
net/iucv/iucv.c
514
if (!cpumask_test_cpu(cpu, &iucv_buffer_cpumask))
net/iucv/iucv.c
538
if (cpumask_test_cpu(cpu, &iucv_buffer_cpumask) &&
net/iucv/iucv.c
539
!cpumask_test_cpu(cpu, &iucv_irq_cpumask))
rust/helpers/cpumask.c
32
return cpumask_test_cpu(cpu, srcp);