cpumask_any
return cpumask_any(&amu_cpus);
if (caps->cpus && cpumask_any(caps->cpus) < nr_cpu_ids)
cpu = cpumask_any(cpu_online_mask);
boot_cpuid = cpumask_any(cpu_online_mask);
boot_cpuid = cpumask_any(cpu_online_mask);
this_cpu_ci = get_cpu_cacheinfo(cpumask_any(cpu_online_mask));
int cpu = cpumask_any(&d->hdr.cpu_mask);
cpu = cpumask_any(&hdr->cpu_mask);
cpumask_set_cpu(cpumask_any(&d->hdr.cpu_mask), cpu_mask);
policy->cpu = cpumask_any(policy->cpus);
topology_core_cpumask(cpumask_any(tmp));
cpumask_set_cpu(cpumask_any(topology_sibling_cpumask(cpu)), mask);
cpu = cpumask_any(cpu_online_mask);
target = cpumask_any(&nd_pmu->arch_cpumask);
nd_pmu->cpu = cpumask_any(&nd_pmu->arch_cpumask);
nd_pmu->cpu = cpumask_any(cpumask);
cpu = cpumask_any(cpu_online_mask);
cpu = cpumask_any(hfi_instance->cpus);
cpu = cpumask_any(cpu_pool);
cpu = cpumask_any(cpu_pool);
*vcpu_id = cpumask_any(ne_enclave->threads_per_core[core_id]);
cpu = cpumask_any(&d->hdr.cpu_mask);
ci = get_cpu_cacheinfo_level(cpumask_any(&d->hdr.cpu_mask), r->ctrl_scope);
int cpu = cpumask_any(&d->hdr.cpu_mask);
int cpu = cpumask_any(cs->effective_cpus);
cpu = cpumask_any(p->cpus_ptr);
cur_dl_b = dl_bw_of(cpumask_any(cur));
cpu = cpumask_any(cpu_active_mask);
unsigned int cpu = cpumask_any(cpu_map);
migrator = cpumask_any(tmigr_available_cpumask);