Symbol: cpu_of
kernel/sched/core.c
1130
cpu = cpu_of(rq);
kernel/sched/core.c
1316
int cpu = cpu_of(rq);
kernel/sched/core.c
2970
stop_one_cpu_nowait(cpu_of(rq), migration_cpu_stop,
kernel/sched/core.c
3751
if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq)))
kernel/sched/core.c
3752
set_task_cpu(p, cpu_of(rq));
kernel/sched/core.c
5567
hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
kernel/sched/core.c
6025
cpu = cpu_of(rq);
kernel/sched/core.c
6323
int cpu = cpu_of(rq);
kernel/sched/core.c
6598
int this_cpu = cpu_of(rq);
kernel/sched/core.c
796
irq_delta = irq_time_read(cpu_of(rq)) - rq->prev_irq_time;
kernel/sched/core.c
825
steal = prev_steal = paravirt_steal_clock(cpu_of(rq));
kernel/sched/core.c
8500
int cpu = cpu_of(rq);
kernel/sched/core.c
859
clock = sched_clock_cpu(cpu_of(rq));
kernel/sched/core.c
890
WARN_ON_ONCE(cpu_of(rq) != smp_processor_id());
kernel/sched/core.c
941
smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd);
kernel/sched/core_sched.c
244
const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq));
kernel/sched/deadline.c
1411
int cpu = cpu_of(rq);
kernel/sched/deadline.c
1810
if (WARN_ON_ONCE(!cpu_online(cpu_of(rq))))
kernel/sched/deadline.c
1881
int cpu = cpu_of(rq);
kernel/sched/deadline.c
1884
dl_b = dl_bw_of(cpu_of(rq));
kernel/sched/deadline.c
1898
int cpu = cpu_of(rq);
kernel/sched/deadline.c
3111
src_dl_b = dl_bw_of(cpu_of(rq));
kernel/sched/debug.c
388
cpu_of(rq),
kernel/sched/ext.c
1360
return likely((rq->scx.flags & SCX_RQ_ONLINE) && cpu_active(cpu_of(rq)));
kernel/sched/ext.c
1374
if (sticky_cpu == cpu_of(rq))
kernel/sched/ext.c
1508
sticky_cpu = cpu_of(rq);
kernel/sched/ext.c
1535
unlikely(cpu_of(rq) != p->scx.selected_cpu))
kernel/sched/ext.c
1695
set_task_cpu(p, cpu_of(dst_rq));
kernel/sched/ext.c
1696
p->scx.sticky_cpu = cpu_of(dst_rq);
kernel/sched/ext.c
1706
WARN_ON_ONCE(!cpumask_test_cpu(cpu_of(dst_rq), p->cpus_ptr));
kernel/sched/ext.c
1736
int cpu = cpu_of(rq);
kernel/sched/ext.c
1963
int node = cpu_to_node(cpu_of(rq));
kernel/sched/ext.c
2196
cpu_of(rq), NULL);
kernel/sched/ext.c
2249
cpu_of(rq), prev_on_scx ? prev : NULL);
kernel/sched/ext.c
2272
scx_kick_cpu(sch, cpu_of(rq), 0);
kernel/sched/ext.c
2415
cpu_of(rq), &args);
kernel/sched/ext.c
2492
if (cpu == cpu_of(rq) ||
kernel/sched/ext.c
2750
int cpu = cpu_of(rq);
kernel/sched/ext.c
5680
if ((cpu_online(cpu) || cpu == cpu_of(this_rq)) &&
kernel/sched/ext.c
5716
(cpu_online(cpu) || cpu == cpu_of(this_rq)))
kernel/sched/ext.c
7369
clock = sched_clock_cpu(cpu_of(rq));
kernel/sched/ext.c
924
p->scx.core_sched_at = sched_clock_cpu(cpu_of(rq));
kernel/sched/ext_idle.c
736
int cpu = cpu_of(rq);
kernel/sched/ext_idle.c
771
SCX_CALL_OP(sch, SCX_KF_REST, update_idle, rq, cpu_of(rq), idle);
kernel/sched/fair.c
10195
(arch_scale_cpu_capacity(cpu_of(rq)) * 100));
kernel/sched/fair.c
1199
long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq)));
kernel/sched/fair.c
12018
__cpumask_clear_cpu(cpu_of(busiest), cpus);
kernel/sched/fair.c
12083
stop_one_cpu_nowait(cpu_of(busiest),
kernel/sched/fair.c
12198
int busiest_cpu = cpu_of(busiest_rq);
kernel/sched/fair.c
13103
if (unlikely(on_null_domain(rq) || !cpu_active(cpu_of(rq))))
kernel/sched/fair.c
313
int cpu = cpu_of(rq);
kernel/sched/fair.c
4173
if (!cpu_active(cpu_of(rq_of(cfs_rq))))
kernel/sched/fair.c
4180
now = sched_clock_cpu(cpu_of(rq_of(cfs_rq)));
kernel/sched/fair.c
4203
now = sched_clock_cpu(cpu_of(rq_of(cfs_rq)));
kernel/sched/fair.c
4226
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
kernel/sched/fair.c
4603
now += sched_clock_cpu(cpu_of(rq)) - u64_u32_load(rq->clock_idle);
kernel/sched/fair.c
5124
int cpu = cpu_of(rq);
kernel/sched/fair.c
5910
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
kernel/sched/fair.c
5981
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
kernel/sched/fair.c
6046
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
kernel/sched/fair.c
6151
smp_call_function_single_async(cpu_of(rq), &rq->cfsb_csd);
kernel/sched/fair.c
6208
if (cpu_of(rq) != this_cpu) {
kernel/sched/fair.c
6648
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
kernel/sched/fair.c
6665
if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask))
kernel/sched/fair.c
6677
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
kernel/sched/fair.c
6721
int cpu = cpu_of(rq);
kernel/sched/fair.c
7246
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
7269
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
7597
int core = cpu_of(rq);
kernel/sched/fair.c
9900
int cpu = cpu_of(rq);
kernel/sched/fair.c
9947
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
kernel/sched/pelt.c
440
running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq)));
kernel/sched/pelt.c
441
running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq)));
kernel/sched/pelt.c
481
unsigned long hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
kernel/sched/pelt.h
123
delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq)));
kernel/sched/pelt.h
124
delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq)));
kernel/sched/rt.c
2441
if (p->prio < rq->donor->prio && cpu_online(cpu_of(rq)))
kernel/sched/rt.c
513
iter && (rt_rq = iter->rt_rq[cpu_of(rq)]); \
kernel/sched/rt.c
533
int cpu = cpu_of(rq);
kernel/sched/rt.c
551
int cpu = cpu_of(rq_of_rt_rq(rt_rq));
kernel/sched/sched.h
1498
for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) {
kernel/sched/sched.h
2918
int cpu = cpu_of(rq);
kernel/sched/sched.h
3038
if (!cpu_active(cpu_of(rq)))
kernel/sched/sched.h
3475
cpu_of(rq)));
kernel/sched/sched.h
3589
rq_util = cpu_util_cfs(cpu_of(rq)) + cpu_util_rt(rq);