Symbol: sched_class
include/linux/sched.h
878
const struct sched_class *sched_class;
kernel/sched/core.c
10887
if ((flags & DEQUEUE_CLASS) && p->sched_class->switching_from)
kernel/sched/core.c
10888
p->sched_class->switching_from(rq, p);
kernel/sched/core.c
10892
.class = p->sched_class,
kernel/sched/core.c
10899
if (p->sched_class->get_prio)
kernel/sched/core.c
10900
ctx->prio = p->sched_class->get_prio(rq, p);
kernel/sched/core.c
10910
if ((flags & DEQUEUE_CLASS) && p->sched_class->switched_from)
kernel/sched/core.c
10911
p->sched_class->switched_from(rq, p);
kernel/sched/core.c
10926
WARN_ON_ONCE(p->sched_class != ctx->class && !(ctx->flags & ENQUEUE_CLASS));
kernel/sched/core.c
10928
if ((ctx->flags & ENQUEUE_CLASS) && p->sched_class->switching_to)
kernel/sched/core.c
10929
p->sched_class->switching_to(rq, p);
kernel/sched/core.c
10937
if (p->sched_class->switched_to)
kernel/sched/core.c
10938
p->sched_class->switched_to(rq, p);
kernel/sched/core.c
10947
if (sched_class_above(p->sched_class, ctx->class)) {
kernel/sched/core.c
10949
rq->next_class = p->sched_class;
kernel/sched/core.c
10955
if (sched_class_above(ctx->class, p->sched_class))
kernel/sched/core.c
10959
p->sched_class->prio_changed(rq, p, ctx->prio);
kernel/sched/core.c
1340
if (p->sched_class != &fair_sched_class)
kernel/sched/core.c
1464
if (update_load && p->sched_class->reweight_task)
kernel/sched/core.c
1465
p->sched_class->reweight_task(task_rq(p), p, &lw);
kernel/sched/core.c
1792
if (unlikely(!p->sched_class->uclamp_enabled))
kernel/sched/core.c
1820
if (unlikely(!p->sched_class->uclamp_enabled))
kernel/sched/core.c
197
if (p->sched_class == &stop_sched_class) /* trumps deadline */
kernel/sched/core.c
206
if (p->sched_class == &idle_sched_class)
kernel/sched/core.c
2107
p->sched_class->enqueue_task(rq, p, flags);
kernel/sched/core.c
2139
return p->sched_class->dequeue_task(rq, p, flags);
kernel/sched/core.c
2189
if (p->sched_class == rq->next_class) {
kernel/sched/core.c
2192
} else if (sched_class_above(p->sched_class, rq->next_class)) {
kernel/sched/core.c
2195
rq->next_class = p->sched_class;
kernel/sched/core.c
2645
if (p->sched_class->find_lock_rq)
kernel/sched/core.c
2646
lowest_rq = p->sched_class->find_lock_rq(p, rq);
kernel/sched/core.c
2698
p->sched_class->set_cpus_allowed(p, ctx);
kernel/sched/core.c
3259
p->sched_class == &fair_sched_class &&
kernel/sched/core.c
3286
if (p->sched_class->migrate_task_rq)
kernel/sched/core.c
3287
p->sched_class->migrate_task_rq(p, new_cpu);
kernel/sched/core.c
339
if (p->sched_class->task_is_throttled)
kernel/sched/core.c
340
return p->sched_class->task_is_throttled(p, cpu);
kernel/sched/core.c
3521
cpu = p->sched_class->select_task_rq(p, cpu, *wake_flags);
kernel/sched/core.c
3560
stop->sched_class = &stop_sched_class;
kernel/sched/core.c
3584
old_stop->sched_class = &rt_sched_class;
kernel/sched/core.c
3672
if (p->sched_class->task_woken) {
kernel/sched/core.c
3678
p->sched_class->task_woken(rq, p);
kernel/sched/core.c
3855
if (p->sched_class == &stop_sched_class)
kernel/sched/core.c
4673
p->sched_class = &rt_sched_class;
kernel/sched/core.c
4676
p->sched_class = &ext_sched_class;
kernel/sched/core.c
4679
p->sched_class = &fair_sched_class;
kernel/sched/core.c
4720
if (p->sched_class->task_fork)
kernel/sched/core.c
4721
p->sched_class->task_fork(p);
kernel/sched/core.c
4789
if (p->sched_class->task_woken) {
kernel/sched/core.c
4795
p->sched_class->task_woken(rq, p);
kernel/sched/core.c
5184
if (prev->sched_class->task_dead)
kernel/sched/core.c
5185
prev->sched_class->task_dead(prev);
kernel/sched/core.c
5418
dest_cpu = p->sched_class->select_task_rq(p, task_cpu(p), WF_EXEC);
kernel/sched/core.c
5489
p->sched_class->update_curr(rq);
kernel/sched/core.c
5573
donor->sched_class->task_tick(rq, donor, 0);
kernel/sched/core.c
5669
curr->sched_class->task_tick(rq, curr, 0);
kernel/sched/core.c
5888
const struct sched_class *start_class = prev->sched_class;
kernel/sched/core.c
5889
const struct sched_class *class;
kernel/sched/core.c
5912
const struct sched_class *class;
kernel/sched/core.c
5926
if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) &&
kernel/sched/core.c
5991
const struct sched_class *class;
kernel/sched/core.c
6854
rq->next_class = next->sched_class;
kernel/sched/core.c
7254
const struct sched_class *__setscheduler_class(int policy, int prio)
kernel/sched/core.c
7313
const struct sched_class *prev_class, *next_class;
kernel/sched/core.c
7370
prev_class = p->sched_class;
kernel/sched/core.c
7407
p->sched_class = next_class;
kernel/sched/core.c
7986
idle->sched_class = &idle_sched_class;
kernel/sched/core.c
8233
const struct sched_class *class;
kernel/sched/core.c
8248
const struct sched_class *class;
kernel/sched/core.c
894
rq->donor->sched_class->task_tick(rq, rq->donor, 1);
kernel/sched/core.c
9155
if (tsk->sched_class->task_change_group)
kernel/sched/core.c
9156
tsk->sched_class->task_change_group(tsk);
kernel/sched/deadline.c
1169
rq->donor->sched_class->update_curr(rq);
kernel/sched/deadline.c
1808
rq->donor->sched_class->update_curr(rq);
kernel/sched/deadline.c
2536
if (p->sched_class != &dl_sched_class)
kernel/sched/deadline.c
2579
if (rq->donor->sched_class != &dl_sched_class)
kernel/sched/ext.c
1008
rq->curr->sched_class == &ext_sched_class) {
kernel/sched/ext.c
1013
if (preempt || sched_class_above(&ext_sched_class, rq->curr->sched_class))
kernel/sched/ext.c
2053
if (sched_class_above(p->sched_class, dst_rq->curr->sched_class))
kernel/sched/ext.c
2178
bool prev_on_scx = prev->sched_class == &ext_sched_class;
kernel/sched/ext.c
2370
preempt_reason_from_class(const struct sched_class *class)
kernel/sched/ext.c
2384
const struct sched_class *next_class = next->sched_class;
kernel/sched/ext.c
2456
if (next && sched_class_above(&ext_sched_class, next->sched_class)) {
kernel/sched/ext.c
2465
if (next && next->sched_class != &ext_sched_class)
kernel/sched/ext.c
2558
prev->sched_class != &ext_sched_class)) {
kernel/sched/ext.c
259
static const struct sched_class *scx_setscheduler_class(struct task_struct *p)
kernel/sched/ext.c
261
if (p->sched_class == &stop_sched_class)
kernel/sched/ext.c
3119
if (p->sched_class == &ext_sched_class) {
kernel/sched/ext.c
3278
if (p->sched_class != &ext_sched_class)
kernel/sched/ext.c
3869
if (unlikely(p->sched_class != &ext_sched_class))
kernel/sched/ext.c
4423
const struct sched_class *old_class = p->sched_class;
kernel/sched/ext.c
4424
const struct sched_class *new_class = scx_setscheduler_class(p);
kernel/sched/ext.c
4432
p->sched_class = new_class;
kernel/sched/ext.c
4756
rq->curr->sched_class == &idle_sched_class;
kernel/sched/ext.c
4777
rq->curr->sched_class);
kernel/sched/ext.c
4819
if (rq->curr->sched_class == &ext_sched_class)
kernel/sched/ext.c
5258
const struct sched_class *old_class = p->sched_class;
kernel/sched/ext.c
5259
const struct sched_class *new_class = scx_setscheduler_class(p);
kernel/sched/ext.c
5269
p->sched_class = new_class;
kernel/sched/ext.c
5667
const struct sched_class *cur_class;
kernel/sched/ext.c
5672
cur_class = rq->curr->sched_class;
kernel/sched/ext.c
5779
struct sched_class *class;
kernel/sched/ext.c
5789
if (copy_from_kernel_nofault(&class, &p->sched_class, sizeof(class)) ||
kernel/sched/ext.c
669
if (p->sched_class != &idle_sched_class)
kernel/sched/ext.h
35
return scx_enabled() && p->sched_class == &ext_sched_class;
kernel/sched/fair.c
1202
if (p->sched_class != &fair_sched_class) {
kernel/sched/fair.c
13351
if (p->sched_class != &fair_sched_class)
kernel/sched/fair.c
298
const struct sched_class fair_sched_class;
kernel/sched/fair.c
5795
if (p->sched_class != &fair_sched_class)
kernel/sched/fair.c
6834
if (!hrtick_enabled_fair(rq) || donor->sched_class != &fair_sched_class)
kernel/sched/fair.c
8819
if (p->sched_class != &fair_sched_class)
kernel/sched/fair.c
8992
if (prev->sched_class != &fair_sched_class)
kernel/sched/fair.c
9361
if (p->sched_class != &fair_sched_class)
kernel/sched/pelt.c
480
const struct sched_class *curr_class = rq->donor->sched_class;
kernel/sched/rt.c
1621
if (p->sched_class != &rt_sched_class)
kernel/sched/rt.c
1665
if (rq->donor->sched_class != &rt_sched_class)
kernel/sched/rt.c
1979
if (rq->donor->sched_class != &rt_sched_class)
kernel/sched/rt.c
979
if (donor->sched_class != &rt_sched_class)
kernel/sched/sched.h
1205
const struct sched_class *next_class;
kernel/sched/sched.h
1828
extern const struct sched_class ext_sched_class;
kernel/sched/sched.h
2666
prev->sched_class->put_prev_task(rq, prev, NULL);
kernel/sched/sched.h
2671
next->sched_class->set_next_task(rq, next, false);
kernel/sched/sched.h
2695
prev->sched_class->put_prev_task(rq, prev, next);
kernel/sched/sched.h
2696
next->sched_class->set_next_task(rq, next, true);
kernel/sched/sched.h
2710
const struct sched_class name##_sched_class \
kernel/sched/sched.h
2711
__aligned(__alignof__(struct sched_class)) \
kernel/sched/sched.h
2715
extern struct sched_class __sched_class_highest[];
kernel/sched/sched.h
2716
extern struct sched_class __sched_class_lowest[];
kernel/sched/sched.h
2718
extern const struct sched_class stop_sched_class;
kernel/sched/sched.h
2719
extern const struct sched_class dl_sched_class;
kernel/sched/sched.h
2720
extern const struct sched_class rt_sched_class;
kernel/sched/sched.h
2721
extern const struct sched_class fair_sched_class;
kernel/sched/sched.h
2722
extern const struct sched_class idle_sched_class;
kernel/sched/sched.h
2728
static inline const struct sched_class *next_active_class(const struct sched_class *class)
kernel/sched/sched.h
2754
static inline void rq_modified_begin(struct rq *rq, const struct sched_class *class)
kernel/sched/sched.h
2760
static inline bool rq_modified_above(struct rq *rq, const struct sched_class *class)
kernel/sched/sched.h
4065
extern const struct sched_class *__setscheduler_class(int policy, int prio);
kernel/sched/sched.h
4094
const struct sched_class *class;
kernel/sched/syscalls.c
1337
rq->donor->sched_class->yield_task(rq);
kernel/sched/syscalls.c
1427
if (!curr->sched_class->yield_to_task)
kernel/sched/syscalls.c
1430
if (curr->sched_class != p->sched_class)
kernel/sched/syscalls.c
1436
yielded = curr->sched_class->yield_to_task(rq, p);
kernel/sched/syscalls.c
1529
if (p->sched_class->get_rr_interval)
kernel/sched/syscalls.c
1530
time_slice = p->sched_class->get_rr_interval(rq, p);
kernel/sched/syscalls.c
499
const struct sched_class *prev_class, *next_class;
kernel/sched/syscalls.c
675
prev_class = p->sched_class;
kernel/sched/syscalls.c
685
p->sched_class = next_class;
tools/perf/builtin-kwork.c
915
struct kwork_class *sched_class;
tools/perf/builtin-kwork.c
925
sched_class = get_kwork_class(kwork, KWORK_CLASS_SCHED);
tools/perf/builtin-kwork.c
926
if (sched_class) {
tools/perf/builtin-kwork.c
927
sched_work = find_work_by_id(&sched_class->work_root,