cpu_active_mask
cpu_active_mask);
cpumask_and(&mask, &allowed, cpu_active_mask);
target = cpumask_last(cpu_active_mask);
target = cpumask_last(cpu_active_mask);
#define num_active_cpus() cpumask_weight(cpu_active_mask)
return cpumask_test_cpu(cpu, cpu_active_mask);
if (WARN_ON_ONCE(!cpumask_subset(doms[i], cpu_active_mask)))
(!cpumask_intersects(xcpus, cpu_active_mask) &&
cpumask_and(parent->effective_cpus, parent->effective_cpus, cpu_active_mask);
cpumask_and(full_hk_cpus, full_hk_cpus, cpu_active_mask);
if (!cpumask_intersects(tmp->new_cpus, cpu_active_mask) ||
cpumask_and(tmp->new_cpus, xcpus, cpu_active_mask);
cpumask_and(tmp->new_cpus, tmp->delmask, cpu_active_mask);
!cpumask_intersects(tmp->addmask, cpu_active_mask))) {
cpumask_and(new_ecpus, new_ecpus, cpu_active_mask);
int cpu = cpumask_any_and(cpu_active_mask, cs->effective_cpus);
cpumask_copy(&new_cpus, cpu_active_mask);
cpumask_copy(top_cpuset.effective_cpus, cpu_active_mask);
if (!cpumask_intersects(pmask, cpu_active_mask))
if (WARN_ON(!cpumask_and(pmask, possible_mask, cpu_active_mask)))
cpumask_copy(pmask, cpu_active_mask);
cpu = cpumask_first_and(em_span_cpus(pd), cpu_active_mask);
const struct cpumask *cpu_valid_mask = cpu_active_mask;
sched_init_domains(cpu_active_mask);
cpumask_and(lowest_mask, lowest_mask, cpu_active_mask);
return cpumask_weight_and(rd->span, cpu_active_mask);
for_each_cpu_and(i, mask, cpu_active_mask)
for_each_cpu_and(i, rd->span, cpu_active_mask) {
cpumask_andnot(cpus, cpu_active_mask, hk_msk);
cpu = cpumask_first_and(cpu_active_mask, msk);
cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr);
cpu = cpumask_any(cpu_active_mask);
cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask))
cpumask_and(mask, &p->cpus_mask, cpu_active_mask);
if (!sched_is_eas_possible(cpu_active_mask)) {
cpumask_and(doms_new[0], cpu_active_mask,
cpumask_and(doms_new[0], cpu_active_mask,
if (cpumask_test_cpu(rq->cpu, cpu_active_mask))
queue_stop_cpus_work(cpu_active_mask, multi_cpu_stop, &msdata,
int i, ncpu = cpumask_any_and(cpu_active_mask, housekeeping_cpumask(HK_TYPE_TIMER));
if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask))
if (cpumask_intersects(wq_unbound_cpumask, cpu_active_mask))