task_rq
if (task_rq(t)->curr == t)
struct rq *rq = task_rq(p);
struct rq *rq = task_rq(p);
p->sched_class->reweight_task(task_rq(p), p, &lw);
rq = task_rq(p);
if (task_rq(p) == rq) {
if (task_rq(p) != rq)
if (task_rq(p) == rq) {
if (prio_less(b, a, !!task_rq(a)->core->core_forceidle_count))
lockdep_is_held(__rq_lockp(task_rq(p)))));
src_rq = task_rq(p);
atomic_dec(&task_rq(p)->nr_iowait);
atomic_dec(&task_rq(p)->nr_iowait);
struct sched_entity *curr = task_rq(p)->cfs.curr;
return (task_rq(t)->idle == t);
rq = task_rq(p);
if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) {
rq = task_rq(p);
if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) {
if (task_rq(p) == rq && task_on_rq_queued(p))
rq = task_rq(p);
if (!cpudl_find(&task_rq(task)->rd->cpudl, task, later_mask))
(task_rq(task) != rq ||
rq = task_rq(p);
dl_rq_change_utilization(task_rq(p), &p->dl, new_bw);
rq = task_rq(dl_task_of(dl_se));
lockdep_assert_rq_held(task_rq(p));
struct rq *rq = task_rq(p);
dsq = &task_rq(p)->scx.bypass_dsq;
!WARN_ON_ONCE(src_rq != task_rq(p));
struct rq *src_rq = task_rq(p), *dst_rq;
struct rq *task_rq = task_rq(p);
if (rq == task_rq) {
if (likely(consume_remote_task(rq, p, dsq, task_rq)))
struct rq *src_rq = task_rq(p);
!WARN_ON_ONCE(src_rq != task_rq(p))) {
!scx_rq_bypassing(task_rq(a)))
rq_bypass = scx_rq_bypassing(task_rq(p));
struct rq *rq = task_rq(p);
struct rq *rq = task_rq(p);
lockdep_assert_rq_held(task_rq(p));
SCX_CALL_OP_TASK(sch, SCX_KF_REST, exit_task, task_rq(p),
struct rq *rq = task_rq(p);
lockdep_assert_rq_held(task_rq(p));
lockdep_assert_rq_held(task_rq(p));
update_rq_clock(task_rq(p));
src_rq = task_rq(p);
WARN_ON_ONCE(src_rq != task_rq(p))) {
return task_rq(p)->curr == p;
struct rq *rq = task_rq(a);
WARN_ON_ONCE(task_rq(b)->core != rq->core);
cfs_rqa = &task_rq(a)->cfs;
cfs_rqb = &task_rq(b)->cfs;
check_update_overutilized_status(task_rq(curr));
(lockdep_is_held(__rq_lockp(task_rq(p))) && !READ_ONCE(p->on_cpu)));
WARN_ON_ONCE(task_rq(p) != rq);
WARN_ON_ONCE(task_rq(p) != rq);
ret = cpupri_find_fitness(&task_rq(task)->rd->cpupri,
ret = cpupri_find(&task_rq(task)->rd->cpupri,
return task_rq(p);
return &task_rq(p)->cfs;
struct rq *rq = task_rq(p);
if (task_on_cpu(task_rq(p), p))
if (!cpumask_subset(task_rq(p)->rd->span, mask))
p_rq = task_rq(p);
if (task_rq(p) != p_rq)