Symbol: sched_domain_span
kernel/sched/core.c
1222
for_each_cpu_and(i, sched_domain_span(sd), hk_mask) {
kernel/sched/core.c
3608
if (cpumask_test_cpu(cpu, sched_domain_span(sd))) {
kernel/sched/core.c
6305
for_each_cpu_wrap(i, sched_domain_span(sd), cpu + 1) {
kernel/sched/deadline.c
2767
cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
kernel/sched/deadline.c
2773
sched_domain_span(sd));
kernel/sched/ext_idle.c
254
return sched_domain_span(sd);
kernel/sched/fair.c
11888
cpumask_and(cpus, sched_domain_span(sd), cpu_active_mask);
kernel/sched/fair.c
12233
if (cpumask_test_cpu(busiest_cpu, sched_domain_span(sd)))
kernel/sched/fair.c
12553
for_each_cpu_and(i, sched_domain_span(sd), nohz.idle_cpus_mask) {
kernel/sched/fair.c
7507
if (!cpumask_intersects(sched_domain_span(sd), p->cpus_ptr))
kernel/sched/fair.c
7664
if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
kernel/sched/fair.c
7707
cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
kernel/sched/fair.c
7779
cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr);
kernel/sched/fair.c
8409
while (sd && !cpumask_test_cpu(prev_cpu, sched_domain_span(sd)))
kernel/sched/fair.c
8446
if (!cpumask_test_cpu(cpu, sched_domain_span(sd)))
kernel/sched/fair.c
8618
cpumask_test_cpu(prev_cpu, sched_domain_span(tmp))) {
kernel/sched/rt.c
1824
cpumask_test_cpu(this_cpu, sched_domain_span(sd))) {
kernel/sched/rt.c
1830
sched_domain_span(sd));
kernel/sched/stats.c
140
cpumask_pr_args(sched_domain_span(sd)));
kernel/sched/topology.c
1012
!cpumask_subset(sched_domain_span(sibling->child),
kernel/sched/topology.c
1013
sched_domain_span(sd)))
kernel/sched/topology.c
1022
cpumask_equal(sched_domain_span(sibling->child),
kernel/sched/topology.c
1023
sched_domain_span(sibling)))
kernel/sched/topology.c
1033
const struct cpumask *span = sched_domain_span(sd);
kernel/sched/topology.c
1059
if (!cpumask_test_cpu(i, sched_domain_span(sibling)))
kernel/sched/topology.c
1094
!cpumask_subset(sched_domain_span(sibling->child), span))
kernel/sched/topology.c
115
!cpumask_equal(sched_domain_span(sd->child),
kernel/sched/topology.c
1203
cpu = cpumask_first(sched_domain_span(child));
kernel/sched/topology.c
1218
cpumask_copy(sched_group_span(sg), sched_domain_span(child));
kernel/sched/topology.c
1245
const struct cpumask *span = sched_domain_span(sd);
kernel/sched/topology.c
130
if (!cpumask_equal(sched_domain_span(sd), groupmask))
kernel/sched/topology.c
134
!cpumask_subset(groupmask, sched_domain_span(sd->parent)))
kernel/sched/topology.c
1689
sd_span = sched_domain_span(sd);
kernel/sched/topology.c
172
if (cpumask_weight(sched_domain_span(sd)) == 1)
kernel/sched/topology.c
195
if (!cpumask_equal(sched_domain_span(sd), sched_domain_span(parent)))
kernel/sched/topology.c
2477
if (!cpumask_subset(sched_domain_span(child),
kernel/sched/topology.c
2478
sched_domain_span(sd))) {
kernel/sched/topology.c
2483
cpumask_or(sched_domain_span(sd),
kernel/sched/topology.c
2484
sched_domain_span(sd),
kernel/sched/topology.c
2485
sched_domain_span(child));
kernel/sched/topology.c
2586
if (cpumask_equal(cpu_map, sched_domain_span(sd)))
kernel/sched/topology.c
2597
sd->span_weight = cpumask_weight(sched_domain_span(sd));
kernel/sched/topology.c
54
cpumask_pr_args(sched_domain_span(sd)), sd->name);
kernel/sched/topology.c
56
if (!cpumask_test_cpu(cpu, sched_domain_span(sd))) {
kernel/sched/topology.c
685
id = cpumask_first(sched_domain_span(sd));
kernel/sched/topology.c
686
size = cpumask_weight(sched_domain_span(sd));
kernel/sched/topology.c
697
id = cpumask_first(sched_domain_span(sd));
kernel/sched/topology.c
937
if (!cpumask_equal(sg_span, sched_domain_span(sibling->child)))
kernel/sched/topology.c
966
cpumask_copy(sg_span, sched_domain_span(sd->child));
kernel/sched/topology.c
969
cpumask_copy(sg_span, sched_domain_span(sd));