cpu_of
cpu = cpu_of(rq);
int cpu = cpu_of(rq);
stop_one_cpu_nowait(cpu_of(rq), migration_cpu_stop,
if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq)))
set_task_cpu(p, cpu_of(rq));
hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
cpu = cpu_of(rq);
int cpu = cpu_of(rq);
int this_cpu = cpu_of(rq);
irq_delta = irq_time_read(cpu_of(rq)) - rq->prev_irq_time;
steal = prev_steal = paravirt_steal_clock(cpu_of(rq));
int cpu = cpu_of(rq);
clock = sched_clock_cpu(cpu_of(rq));
WARN_ON_ONCE(cpu_of(rq) != smp_processor_id());
smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd);
const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq));
int cpu = cpu_of(rq);
if (WARN_ON_ONCE(!cpu_online(cpu_of(rq))))
int cpu = cpu_of(rq);
dl_b = dl_bw_of(cpu_of(rq));
int cpu = cpu_of(rq);
src_dl_b = dl_bw_of(cpu_of(rq));
cpu_of(rq),
return likely((rq->scx.flags & SCX_RQ_ONLINE) && cpu_active(cpu_of(rq)));
if (sticky_cpu == cpu_of(rq))
sticky_cpu = cpu_of(rq);
unlikely(cpu_of(rq) != p->scx.selected_cpu))
set_task_cpu(p, cpu_of(dst_rq));
p->scx.sticky_cpu = cpu_of(dst_rq);
WARN_ON_ONCE(!cpumask_test_cpu(cpu_of(dst_rq), p->cpus_ptr));
int cpu = cpu_of(rq);
int node = cpu_to_node(cpu_of(rq));
cpu_of(rq), NULL);
cpu_of(rq), prev_on_scx ? prev : NULL);
scx_kick_cpu(sch, cpu_of(rq), 0);
cpu_of(rq), &args);
if (cpu == cpu_of(rq) ||
int cpu = cpu_of(rq);
if ((cpu_online(cpu) || cpu == cpu_of(this_rq)) &&
(cpu_online(cpu) || cpu == cpu_of(this_rq)))
clock = sched_clock_cpu(cpu_of(rq));
p->scx.core_sched_at = sched_clock_cpu(cpu_of(rq));
int cpu = cpu_of(rq);
SCX_CALL_OP(sch, SCX_KF_REST, update_idle, rq, cpu_of(rq), idle);
(arch_scale_cpu_capacity(cpu_of(rq)) * 100));
long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq)));
__cpumask_clear_cpu(cpu_of(busiest), cpus);
stop_one_cpu_nowait(cpu_of(busiest),
int busiest_cpu = cpu_of(busiest_rq);
if (unlikely(on_null_domain(rq) || !cpu_active(cpu_of(rq))))
int cpu = cpu_of(rq);
if (!cpu_active(cpu_of(rq_of(cfs_rq))))
now = sched_clock_cpu(cpu_of(rq_of(cfs_rq)));
now = sched_clock_cpu(cpu_of(rq_of(cfs_rq)));
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
now += sched_clock_cpu(cpu_of(rq)) - u64_u32_load(rq->clock_idle);
int cpu = cpu_of(rq);
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
smp_call_function_single_async(cpu_of(rq), &rq->cfsb_csd);
if (cpu_of(rq) != this_cpu) {
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
if (cpumask_test_cpu(cpu_of(rq), cpu_active_mask))
struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)];
int cpu = cpu_of(rq);
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
int core = cpu_of(rq);
int cpu = cpu_of(rq);
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq)));
running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq)));
unsigned long hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq)));
delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq)));
if (p->prio < rq->donor->prio && cpu_online(cpu_of(rq)))
iter && (rt_rq = iter->rt_rq[cpu_of(rq)]); \
int cpu = cpu_of(rq);
int cpu = cpu_of(rq_of_rt_rq(rt_rq));
for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) {
int cpu = cpu_of(rq);
if (!cpu_active(cpu_of(rq)))
cpu_of(rq)));
rq_util = cpu_util_cfs(cpu_of(rq)) + cpu_util_rt(rq);