topology_sibling_cpumask
topology_sibling_cpumask(smp_processor_id()));
siblings = topology_sibling_cpumask(cpu);
target = cpumask_any_but(topology_sibling_cpumask(cpu), cpu);
target = cpumask_any_but(topology_sibling_cpumask(cpu), cpu);
for_each_cpu(i, topology_sibling_cpumask(cpu)) {
for_each_cpu(i, topology_sibling_cpumask(cpu)) {
struct cpumask *secondaries = topology_sibling_cpumask(cpu);
ctrl.ctrl_cpu = cpumask_first(topology_sibling_cpumask(cpu));
for_each_cpu(cpu, topology_sibling_cpumask(this_cpu)) {
cpumask_set_cpu(0, topology_sibling_cpumask(0));
int threads = cpumask_weight(topology_sibling_cpumask(cpu));
if (cpumask_weight(topology_sibling_cpumask(cpu)) == 1)
for_each_cpu(sibling, topology_sibling_cpumask(cpu)) {
cpumask_clear_cpu(cpu, topology_sibling_cpumask(sibling));
if (cpumask_weight(topology_sibling_cpumask(sibling)) == 1)
cpumask_clear(topology_sibling_cpumask(cpu));
cpumask_set_cpu(cpu, topology_sibling_cpumask(cpu));
link_mask(topology_sibling_cpumask, cpu, i);
threads = cpumask_weight(topology_sibling_cpumask(cpu));
for_each_cpu(i, topology_sibling_cpumask(cpu))
topology_sibling_cpumask(i)) == i)
cpumask_set_cpu(cpumask_first(topology_sibling_cpumask(i)), available_mask);
cpumask_or(tmp, tmp, topology_sibling_cpumask(cpu));
return topology_sibling_cpumask(cpu);
for_each_cpu(sibling, topology_sibling_cpumask(cpu))
cpumask_clear_cpu(cpu, topology_sibling_cpumask(sibling));
topology_sibling_cpumask(cpu));
cpumask_copy(policy->cpus, topology_sibling_cpumask(cpu));
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
cpumask_copy(policy->cpus, topology_sibling_cpumask(policy->cpu));
for_each_cpu(i, topology_sibling_cpumask(cpu))
if (!cpumask_intersects(&pdata->cpumask, topology_sibling_cpumask(cpu)))
target = cpumask_any_and(&pd->cpumask, topology_sibling_cpumask(cpu));
cpumask_set_cpu(cpumask_any(topology_sibling_cpumask(cpu)), mask);
ht = cpumask_weight(topology_sibling_cpumask(
cpumask_weight(topology_sibling_cpumask(
if (cpu % cpumask_weight(topology_sibling_cpumask(cpu)))
cpumask_andnot(cpus, cpus, topology_sibling_cpumask(cpu));
irq_set_affinity_and_hint(*irqs++, topology_sibling_cpumask(cpu));
cpumask_andnot(filter_mask, filter_mask, topology_sibling_cpumask(cpu));
topology_sibling_cpumask(cpu));
cpumask_andnot(filter_mask, filter_mask, topology_sibling_cpumask(cpu));
sibling_mask = topology_sibling_cpumask(cpu);
for_each_cpu(cpu_sibling, topology_sibling_cpumask(0)) {
for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu)) {
for_each_cpu(cpu_sibling, topology_sibling_cpumask(cpu))
#ifndef topology_sibling_cpumask
return topology_sibling_cpumask(cpu);
return cpu == cpumask_first(topology_sibling_cpumask(cpu));
return topology_sibling_cpumask(cpu);
topology_sibling_cpumask(cpu), 1))
siblmsk = topology_sibling_cpumask(cpu);
cpumask_andnot(cpus, cpus, topology_sibling_cpumask(cpu));