Symbol: sched_entity
drivers/accel/ethosu/ethosu_drv.h
12
struct drm_sched_entity sched_entity;
drivers/accel/ethosu/ethosu_job.c
355
ret = drm_sched_entity_init(&ethosu_priv->sched_entity,
drivers/accel/ethosu/ethosu_job.c
363
struct drm_sched_entity *entity = &ethosu_priv->sched_entity;
drivers/accel/ethosu/ethosu_job.c
400
&file_priv->sched_entity,
drivers/accel/rocket/rocket_drv.h
26
struct drm_sched_entity sched_entity;
drivers/accel/rocket/rocket_job.c
507
ret = drm_sched_entity_init(&rocket_priv->sched_entity,
drivers/accel/rocket/rocket_job.c
519
struct drm_sched_entity *entity = &rocket_priv->sched_entity;
drivers/accel/rocket/rocket_job.c
554
&file_priv->sched_entity,
drivers/gpu/drm/etnaviv/etnaviv_drv.c
116
drm_sched_entity_destroy(&ctx->sched_entity[i]);
drivers/gpu/drm/etnaviv/etnaviv_drv.c
91
drm_sched_entity_init(&ctx->sched_entity[i],
drivers/gpu/drm/etnaviv/etnaviv_drv.h
34
struct drm_sched_entity sched_entity[ETNA_MAX_PIPES];
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
536
&ctx->sched_entity[args->pipe],
drivers/gpu/drm/panfrost/panfrost_job.h
44
struct drm_sched_entity sched_entity;
drivers/gpu/drm/scheduler/tests/sched_tests.h
120
drm_sched_entity_to_mock_entity(struct drm_sched_entity *sched_entity)
drivers/gpu/drm/scheduler/tests/sched_tests.h
122
return container_of(sched_entity, struct drm_mock_sched_entity, base);
drivers/gpu/drm/v3d/v3d_drv.c
144
drm_sched_entity_init(&v3d_priv->sched_entity[i],
drivers/gpu/drm/v3d/v3d_drv.c
170
drm_sched_entity_destroy(&v3d_priv->sched_entity[q]);
drivers/gpu/drm/v3d/v3d_drv.c
172
if (job && job->base.entity == &v3d_priv->sched_entity[q]) {
drivers/gpu/drm/v3d/v3d_drv.h
228
struct drm_sched_entity sched_entity[V3D_MAX_QUEUES];
drivers/gpu/drm/v3d/v3d_submit.c
172
ret = drm_sched_job_init(&job->base, &v3d_priv->sched_entity[queue],
include/linux/sched.h
605
struct sched_entity *parent;
include/linux/sched.h
871
struct sched_entity se;
include/trace/events/sched.h
854
TP_PROTO(struct sched_entity *se),
include/trace/events/sched.h
870
TP_PROTO(struct sched_entity *se),
kernel/sched/core.c
5445
struct sched_entity *curr = p->se.cfs_rq->curr;
kernel/sched/core.c
5447
struct sched_entity *curr = task_rq(p)->cfs.curr;
kernel/sched/core.c
8626
root_task_group.se = (struct sched_entity **)ptr;
kernel/sched/debug.c
752
struct sched_entity *se = tg->se[cpu];
kernel/sched/debug.c
906
struct sched_entity *last, *first, *root;
kernel/sched/fair.c
1010
static struct sched_entity *__pick_eevdf(struct cfs_rq *cfs_rq, bool protect)
kernel/sched/fair.c
1013
struct sched_entity *se = __pick_first_entity(cfs_rq);
kernel/sched/fair.c
1014
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
1015
struct sched_entity *best = NULL;
kernel/sched/fair.c
1081
static struct sched_entity *pick_eevdf(struct cfs_rq *cfs_rq)
kernel/sched/fair.c
1086
struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq)
kernel/sched/fair.c
1111
static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se);
kernel/sched/fair.c
1117
static bool update_deadline(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
1149
void init_entity_runnable_average(struct sched_entity *se)
kernel/sched/fair.c
1196
struct sched_entity *se = &p->se;
kernel/sched/fair.c
1232
static s64 update_se(struct rq *rq, struct sched_entity *se)
kernel/sched/fair.c
1273
static void set_next_buddy(struct sched_entity *se);
kernel/sched/fair.c
1294
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
13132
__entity_slice_used(struct sched_entity *se, int min_nr_tasks)
kernel/sched/fair.c
13331
static void se_fi_update(const struct sched_entity *se, unsigned int fi_seq,
kernel/sched/fair.c
13349
struct sched_entity *se = &p->se;
kernel/sched/fair.c
13361
const struct sched_entity *sea = &a->se;
kernel/sched/fair.c
13362
const struct sched_entity *seb = &b->se;
kernel/sched/fair.c
1340
update_stats_wait_start_fair(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
13431
struct sched_entity *se = &curr->se;
kernel/sched/fair.c
13497
static void propagate_entity_cfs_rq(struct sched_entity *se)
kernel/sched/fair.c
13525
static void propagate_entity_cfs_rq(struct sched_entity *se) { }
kernel/sched/fair.c
13528
static void detach_entity_cfs_rq(struct sched_entity *se)
kernel/sched/fair.c
13548
static void attach_entity_cfs_rq(struct sched_entity *se)
kernel/sched/fair.c
13561
struct sched_entity *se = &p->se;
kernel/sched/fair.c
13568
struct sched_entity *se = &p->se;
kernel/sched/fair.c
1357
update_stats_wait_end_fair(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
13607
struct sched_entity *se = &p->se;
kernel/sched/fair.c
13636
struct sched_entity *se = &p->se;
kernel/sched/fair.c
13691
struct sched_entity *se;
kernel/sched/fair.c
13732
struct sched_entity *se;
kernel/sched/fair.c
13756
struct sched_entity *se = tg->se[cpu];
kernel/sched/fair.c
13783
struct sched_entity *se, int cpu,
kernel/sched/fair.c
13784
struct sched_entity *parent)
kernel/sched/fair.c
1383
update_stats_enqueue_sleeper_fair(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
13835
struct sched_entity *se = tg->se[i];
kernel/sched/fair.c
13886
struct sched_entity *se = tg->se[i];
kernel/sched/fair.c
13935
struct sched_entity *se = &task->se;
kernel/sched/fair.c
1403
update_stats_enqueue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
1420
update_stats_dequeue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
1452
update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
290
static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se)
kernel/sched/fair.c
3755
account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
3768
account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
3831
enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
3838
dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
3844
static void place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags);
kernel/sched/fair.c
3846
static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se,
kernel/sched/fair.c
3905
struct sched_entity *se = &p->se;
kernel/sched/fair.c
4027
static void update_cfs_group(struct sched_entity *se)
kernel/sched/fair.c
4045
static inline void update_cfs_group(struct sched_entity *se)
kernel/sched/fair.c
410
is_same_group(struct sched_entity *se, struct sched_entity *pse)
kernel/sched/fair.c
418
static inline struct sched_entity *parent_entity(const struct sched_entity *se)
kernel/sched/fair.c
424
find_matching_se(struct sched_entity **se, struct sched_entity **pse)
kernel/sched/fair.c
4240
void set_task_rq_fair(struct sched_entity *se,
kernel/sched/fair.c
4334
update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
kernel/sched/fair.c
4360
update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
kernel/sched/fair.c
4386
update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq)
kernel/sched/fair.c
4456
static inline int propagate_entity_load_avg(struct sched_entity *se)
kernel/sched/fair.c
4487
static inline bool skip_blocked_update(struct sched_entity *se)
kernel/sched/fair.c
4519
static inline int propagate_entity_load_avg(struct sched_entity *se)
kernel/sched/fair.c
4529
static inline void migrate_se_pelt_lag(struct sched_entity *se)
kernel/sched/fair.c
4608
static void migrate_se_pelt_lag(struct sched_entity *se) {}
kernel/sched/fair.c
465
static int se_is_idle(struct sched_entity *se)
kernel/sched/fair.c
4678
static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
4733
static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
4755
static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
4801
static void sync_entity_load_avg(struct sched_entity *se)
kernel/sched/fair.c
4814
static void remove_entity_load_avg(struct sched_entity *se)
kernel/sched/fair.c
493
static inline struct sched_entity *parent_entity(struct sched_entity *se)
kernel/sched/fair.c
499
find_matching_se(struct sched_entity **se, struct sched_entity **pse)
kernel/sched/fair.c
513
static int se_is_idle(struct sched_entity *se)
kernel/sched/fair.c
5150
struct sched_entity *se = &p->se;
kernel/sched/fair.c
5165
place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
5183
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
5276
requeue_delayed_entity(struct sched_entity *se);
kernel/sched/fair.c
5279
enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
5344
static void __clear_buddies_next(struct sched_entity *se)
kernel/sched/fair.c
5355
static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
5363
static void set_delayed(struct sched_entity *se)
kernel/sched/fair.c
5382
static void clear_delayed(struct sched_entity *se)
kernel/sched/fair.c
5402
static inline void finish_delayed_dequeue_entity(struct sched_entity *se)
kernel/sched/fair.c
5410
dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
5492
set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, bool first)
kernel/sched/fair.c
5533
static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags);
kernel/sched/fair.c
5542
static struct sched_entity *
kernel/sched/fair.c
5545
struct sched_entity *se;
kernel/sched/fair.c
5560
static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev)
kernel/sched/fair.c
5584
entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued)
kernel/sched/fair.c
5776
struct sched_entity *se;
kernel/sched/fair.c
582
static inline bool entity_before(const struct sched_entity *a,
kernel/sched/fair.c
583
const struct sched_entity *b)
kernel/sched/fair.c
6046
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
kernel/sched/fair.c
607
static inline s64 entity_key(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
613
rb_entry((node), struct sched_entity, run_node)
kernel/sched/fair.c
674
sum_w_vruntime_add(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
6807
struct sched_entity *se = &p->se;
kernel/sched/fair.c
684
sum_w_vruntime_sub(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
6905
requeue_delayed_entity(struct sched_entity *se)
kernel/sched/fair.c
6944
struct sched_entity *se = &p->se;
kernel/sched/fair.c
7066
static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags)
kernel/sched/fair.c
717
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
767
static void update_entity_lag(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
799
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
813
int entity_eligible(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
820
struct sched_entity *root = __pick_root_entity(cfs_rq);
kernel/sched/fair.c
821
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
835
struct sched_entity *root = __pick_root_entity(cfs_rq);
kernel/sched/fair.c
836
struct sched_entity *curr = cfs_rq->curr;
kernel/sched/fair.c
853
static inline void __min_vruntime_update(struct sched_entity *se, struct rb_node *node)
kernel/sched/fair.c
856
struct sched_entity *rse = __node_2_se(node);
kernel/sched/fair.c
863
static inline void __min_slice_update(struct sched_entity *se, struct rb_node *node)
kernel/sched/fair.c
8656
struct sched_entity *se = &p->se;
kernel/sched/fair.c
866
struct sched_entity *rse = __node_2_se(node);
kernel/sched/fair.c
8682
struct sched_entity *se = &p->se;
kernel/sched/fair.c
872
static inline void __max_slice_update(struct sched_entity *se, struct rb_node *node)
kernel/sched/fair.c
8729
static void set_next_buddy(struct sched_entity *se)
kernel/sched/fair.c
8749
struct sched_entity *pse, struct sched_entity *se)
kernel/sched/fair.c
875
struct sched_entity *rse = __node_2_se(node);
kernel/sched/fair.c
8771
struct sched_entity *pse, struct sched_entity *se)
kernel/sched/fair.c
8812
struct sched_entity *se = &donor->se, *pse = &p->se;
kernel/sched/fair.c
884
static inline bool min_vruntime_update(struct sched_entity *se, bool exit)
kernel/sched/fair.c
8943
struct sched_entity *se;
kernel/sched/fair.c
8981
struct sched_entity *se;
kernel/sched/fair.c
9009
struct sched_entity *pse = &prev->se;
kernel/sched/fair.c
9079
struct sched_entity *se = &prev->se;
kernel/sched/fair.c
908
RB_DECLARE_CALLBACKS(static, min_vruntime_cb, struct sched_entity,
kernel/sched/fair.c
9095
struct sched_entity *se = &curr->se;
kernel/sched/fair.c
9133
struct sched_entity *se = &p->se;
kernel/sched/fair.c
914
static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
923
static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
930
struct sched_entity *__pick_root_entity(struct cfs_rq *cfs_rq)
kernel/sched/fair.c
940
struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq)
kernel/sched/fair.c
958
static inline void set_protect_slice(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
973
static inline void update_protect_slice(struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/fair.c
980
static inline bool protect_slice(struct sched_entity *se)
kernel/sched/fair.c
985
static inline void cancel_protect_slice(struct sched_entity *se)
kernel/sched/fair.c
9907
struct sched_entity *se;
kernel/sched/fair.c
9947
struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)];
kernel/sched/idle.c
550
struct sched_entity *se = &rq->idle->se;
kernel/sched/pelt.c
296
int __update_load_avg_blocked_se(u64 now, struct sched_entity *se)
kernel/sched/pelt.c
307
int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se)
kernel/sched/pelt.h
8
int __update_load_avg_blocked_se(u64 now, struct sched_entity *se);
kernel/sched/pelt.h
9
int __update_load_avg_se(u64 now, struct cfs_rq *cfs_rq, struct sched_entity *se);
kernel/sched/sched.h
1673
static inline struct task_struct *task_of(struct sched_entity *se)
kernel/sched/sched.h
1685
static inline struct cfs_rq *cfs_rq_of(const struct sched_entity *se)
kernel/sched/sched.h
1691
static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
kernel/sched/sched.h
1705
static inline struct cfs_rq *cfs_rq_of(const struct sched_entity *se)
kernel/sched/sched.h
1714
static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
kernel/sched/sched.h
2904
extern void init_entity_runnable_average(struct sched_entity *se);
kernel/sched/sched.h
3311
extern struct sched_entity *__pick_root_entity(struct cfs_rq *cfs_rq);
kernel/sched/sched.h
3312
extern struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq);
kernel/sched/sched.h
3313
extern struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq);
kernel/sched/sched.h
4014
extern int entity_eligible(struct cfs_rq *cfs_rq, struct sched_entity *se);
kernel/sched/sched.h
484
struct sched_entity **se;
kernel/sched/sched.h
583
struct sched_entity *se, int cpu,
kernel/sched/sched.h
584
struct sched_entity *parent);
kernel/sched/sched.h
614
extern void set_task_rq_fair(struct sched_entity *se,
kernel/sched/sched.h
700
struct sched_entity *curr;
kernel/sched/sched.h
701
struct sched_entity *next;
kernel/sched/sched.h
732
struct sched_entity *h_load_next;
kernel/sched/sched.h
929
static inline void se_update_runnable(struct sched_entity *se)
kernel/sched/sched.h
935
static inline long se_runnable(struct sched_entity *se)
kernel/sched/sched.h
950
static inline void se_update_runnable(struct sched_entity *se) { }
kernel/sched/sched.h
952
static inline long se_runnable(struct sched_entity *se)
kernel/sched/sched.h
965
static inline long se_weight(struct sched_entity *se)
kernel/sched/stats.h
100
__schedstats_from_se(struct sched_entity *se)
kernel/sched/stats.h
94
struct sched_entity se;