cpumask_equal
return cpumask_equal(mm_cpumask(mm),
if (cpumask_equal(mask, cpu_online_mask)) {
if (!cpumask_equal(submask_fn(cpu), cpu_l2_cache_mask(cpu)) &&
if (cpumask_equal(cpu_online_mask, cpu_present_mask))
if (cpumask_equal(&mm->context.cpu_attach_mask,
if (cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id())))
cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id())))
if (cpumask_equal(&mm->context.cpu_attach_mask,
cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id())))
cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id())))
if (cpumask_equal(affinity, cpu_online_mask)) {
if (!cpumask_equal(mask, cpu_present_mask) || exclude_self) {
if (cpumask_equal(cpus, cpu_present_mask)) {
!cpumask_equal(cpu_present_mask, &cpus_booted_once_mask)) {
if (!cpumask_equal(resource->cpu_mask, new_cpu_mask)) {
if (WARN_ON(!cpumask_equal(&dev->coupled_cpus, &coupled->coupled_cpus)))
if (cpumask_equal(&set->mask, &set->used)) {
if (cpumask_equal(mask_val, cpu_present_mask))
cpumask_equal(&cpus_booted_once_mask, cpu_possible_mask))
if (!cpumask_equal(&msc->accessibility, cpu_possible_mask)) {
if (cpumask_equal(&portal_cpus, cpu_possible_mask)) {
if (!cpumask_equal(iscsit_global->allowed_cpumask,
if (!cpumask_equal(policy->related_cpus, em_span_cpus(em))) {
if (cpumask_equal(cpu_present_mask, mask) && idle > MAX_ALL_CPU_IDLE)
if (cpumask_equal(cpu_present_mask, idle_injection_cpu_mask)) {
return cpumask_equal(src1, src2);
if (cpumask_equal(cp->effective_xcpus, tmp->new_cpus)) {
cpumask_equal(tmp->new_cpus, cp->effective_cpus) &&
!cpumask_equal(cp->cpus_allowed, cp->effective_cpus));
if (cpumask_equal(tmp->new_cpus, sibling->effective_cpus))
if (cpumask_equal(cs->cpus_allowed, trialcs->cpus_allowed))
force = !cpumask_equal(cs->effective_xcpus, trialcs->effective_xcpus);
if (cpumask_equal(cs->exclusive_cpus, trialcs->exclusive_cpus))
force = !cpumask_equal(cs->effective_xcpus, trialcs->effective_xcpus);
!cpumask_equal(cs->effective_cpus, oldcs->effective_cpus) ||
cpus_updated = !cpumask_equal(cs->effective_cpus,
cpus_updated = !cpumask_equal(&new_cpus, cs->effective_cpus);
cpus_updated = !cpumask_equal(top_cpuset.effective_cpus, &new_cpus) ||
xcpus_changed = !cpumask_equal(cur->exclusive_cpus, trial->exclusive_cpus);
if (!cpumask_equal(affinity, cpu_possible_mask))
(!affinity || cpumask_equal(affinity, cpu_possible_mask)))
if (cpumask_equal(&p->cpus_mask, ctx->new_mask)) {
if (!cpumask_equal(housekeeping_cpumask(HK_TYPE_DOMAIN),
if (!cpumask_equal(housekeeping_staging,
cpumask_equal(sched_domain_span(sibling->child),
!cpumask_equal(group_balance_mask(group), sched_group_span(group))) {
!cpumask_equal(sched_domain_span(sd->child),
if (!cpumask_equal(sched_domain_span(sd), groupmask))
if (!cpumask_equal(sched_domain_span(sd), sched_domain_span(parent)))
if (!cpumask_equal(tl->mask(tl, id), tl_cpu_mask))
if (cpumask_equal(cpu_map, sched_domain_span(sd)))
if (cpumask_equal(doms_cur[i], doms_new[j]) &&
if (cpumask_equal(doms_new[i], doms_cur[j]) &&
if (cpumask_equal(doms_new[i], doms_cur[j]) &&
if (!cpumask_equal(sg_span, sched_domain_span(sibling->child)))
WARN_ON_ONCE(!cpumask_equal(group_balance_mask(sg), mask));
if (!cpumask_equal(newdev->cpumask, cpumask_of(cpu)))
if (cpumask_equal(bc->cpumask, cpumask))
if (!cpumask_equal(newdev->cpumask, cpumask))
if (cpumask_equal(newdev->cpumask, cpumask_of(cpu)))
if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu)))
!cpumask_equal(curdev->cpumask, newdev->cpumask);
return cpumask_equal(mask, cmp);
return !cpumask_equal(mask, cmp);
if (!cpumask_equal(current_mask, current->cpus_ptr))
if (!cpumask_equal(a->__pod_cpumask, b->__pod_cpumask))
if (!a->affn_strict && !cpumask_equal(a->cpumask, b->cpumask))
if (!cpumask_equal(cpumask, wq_unbound_cpumask))
if (!cpumask_equal(cpumask, wq_unbound_cpumask))