Symbol: cpu_active_mask
arch/arm64/kernel/cpufeature.c
4044
cpu_active_mask);
arch/mips/kernel/mips-mt-fpaff.c
182
cpumask_and(&mask, &allowed, cpu_active_mask);
arch/powerpc/perf/hv-24x7.c
1670
target = cpumask_last(cpu_active_mask);
arch/powerpc/perf/hv-gpci.c
862
target = cpumask_last(cpu_active_mask);
include/linux/cpumask.h
1227
#define num_active_cpus() cpumask_weight(cpu_active_mask)
include/linux/cpumask.h
1251
return cpumask_test_cpu(cpu, cpu_active_mask);
kernel/cgroup/cpuset.c
1004
if (WARN_ON_ONCE(!cpumask_subset(doms[i], cpu_active_mask)))
kernel/cgroup/cpuset.c
1172
(!cpumask_intersects(xcpus, cpu_active_mask) &&
kernel/cgroup/cpuset.c
1268
cpumask_and(parent->effective_cpus, parent->effective_cpus, cpu_active_mask);
kernel/cgroup/cpuset.c
1302
cpumask_and(full_hk_cpus, full_hk_cpus, cpu_active_mask);
kernel/cgroup/cpuset.c
1508
if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) ||
kernel/cgroup/cpuset.c
1776
cpumask_and(tmp->new_cpus, xcpus, cpu_active_mask);
kernel/cgroup/cpuset.c
1854
cpumask_and(tmp->new_cpus, tmp->delmask, cpu_active_mask);
kernel/cgroup/cpuset.c
1863
!cpumask_intersects(tmp->addmask, cpu_active_mask))) {
kernel/cgroup/cpuset.c
2031
cpumask_and(new_ecpus, new_ecpus, cpu_active_mask);
kernel/cgroup/cpuset.c
3046
int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus);
kernel/cgroup/cpuset.c
3868
cpumask_copy(&new_cpus, cpu_active_mask);
kernel/cgroup/cpuset.c
3995
cpumask_copy(top_cpuset.effective_cpus, cpu_active_mask);
kernel/cgroup/cpuset.c
4027
if (!cpumask_intersects(pmask, cpu_active_mask))
kernel/cgroup/cpuset.c
474
if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_active_mask)))
kernel/cgroup/cpuset.c
475
cpumask_copy(pmask, cpu_active_mask);
kernel/power/energy_model.c
498
cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask);
kernel/sched/core.c
3021
const struct cpumask *cpu_valid_mask = cpu_active_mask;
kernel/sched/core.c
8556
sched_init_domains(cpu_active_mask);
kernel/sched/cpupri.c
105
cpumask_and(lowest_mask, lowest_mask, cpu_active_mask);
kernel/sched/deadline.c
132
return cpumask_weight_and(rd->span, cpu_active_mask);
kernel/sched/deadline.c
140
for_each_cpu_and(i, mask, cpu_active_mask)
kernel/sched/deadline.c
182
for_each_cpu_and(i, rd->span, cpu_active_mask) {
kernel/sched/deadline.c
3171
cpumask_andnot(cpus, cpu_active_mask, hk_msk);
kernel/sched/deadline.c
3201
cpu = cpumask_first_and(cpu_active_mask, msk);
kernel/sched/deadline.c
658
cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr);
kernel/sched/deadline.c
671
cpu = cpumask_any(cpu_active_mask);
kernel/sched/fair.c
11888
cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
kernel/sched/fair.c
6665
if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask))
kernel/sched/syscalls.c
1287
cpumask_and(mask, &p->cpus_mask, cpu_active_mask);
kernel/sched/topology.c
280
if (!sched_is_eas_possible(cpu_active_mask)) {
kernel/sched/topology.c
2866
cpumask_and(doms_new[0], cpu_active_mask,
kernel/sched/topology.c
2890
cpumask_and(doms_new[0], cpu_active_mask,
kernel/sched/topology.c
500
if (cpumask_test_cpu(rq->cpu, cpu_active_mask))
kernel/stop_machine.c
697
queue_stop_cpus_work(cpu_active_mask, multi_cpu_stop, &msdata,
kernel/time/hrtimer.c
2304
int i, ncpu = cpumask_any_and(cpu_active_mask, housekeeping_cpumask(HK_TYPE_TIMER));
kernel/workqueue.c
2719
if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask))
kernel/workqueue.c
5682
if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask))