Symbol: dl_bw
include/linux/sched.h
1886
extern int dl_bw_alloc(int cpu, u64 dl_bw);
include/linux/sched.h
1887
extern void dl_bw_free(int cpu, u64 dl_bw);
include/linux/sched.h
655
u64 dl_bw; /* dl_runtime / dl_period */
kernel/cgroup/cpuset.c
3038
cs->sum_migrate_dl_bw += task->dl.dl_bw;
kernel/sched/deadline.c
118
static inline struct dl_bw *dl_bw_of(int i)
kernel/sched/deadline.c
122
return &cpu_rq(i)->rd->dl_bw;
kernel/sched/deadline.c
1388
if (u_inact + rq->dl.extra_bw > rq->dl.max_bw - dl_se->dl_bw)
kernel/sched/deadline.c
1389
u_act = dl_se->dl_bw;
kernel/sched/deadline.c
175
void __dl_update(struct dl_bw *dl_b, s64 bw)
kernel/sched/deadline.c
177
struct root_domain *rd = container_of(dl_b, struct root_domain, dl_bw);
kernel/sched/deadline.c
1880
u64 new_bw = dl_se->dl_bw;
kernel/sched/deadline.c
1882
struct dl_bw *dl_b;
kernel/sched/deadline.c
1899
struct dl_bw *dl_b;
kernel/sched/deadline.c
190
void __dl_sub(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
kernel/sched/deadline.c
1916
__dl_sub(dl_b, dl_se->dl_bw, cpus);
kernel/sched/deadline.c
1929
dl_se->dl_bw = to_ratio(dl_se->dl_period, dl_se->dl_runtime);
kernel/sched/deadline.c
197
void __dl_add(struct dl_bw *dl_b, u64 tsk_bw, int cpus)
kernel/sched/deadline.c
1984
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
kernel/sched/deadline.c
1993
__dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/deadline.c
204
__dl_overflow(struct dl_bw *dl_b, unsigned long cap, u64 old_bw, u64 new_bw)
kernel/sched/deadline.c
211
void __add_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
kernel/sched/deadline.c
216
dl_rq->running_bw += dl_bw;
kernel/sched/deadline.c
224
void __sub_running_bw(u64 dl_bw, struct dl_rq *dl_rq)
kernel/sched/deadline.c
229
dl_rq->running_bw -= dl_bw;
kernel/sched/deadline.c
238
void __add_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
kernel/sched/deadline.c
243
dl_rq->this_bw += dl_bw;
kernel/sched/deadline.c
248
void __sub_rq_bw(u64 dl_bw, struct dl_rq *dl_rq)
kernel/sched/deadline.c
253
dl_rq->this_bw -= dl_bw;
kernel/sched/deadline.c
264
__add_rq_bw(dl_se->dl_bw, dl_rq);
kernel/sched/deadline.c
271
__sub_rq_bw(dl_se->dl_bw, dl_rq);
kernel/sched/deadline.c
278
__add_running_bw(dl_se->dl_bw, dl_rq);
kernel/sched/deadline.c
285
__sub_running_bw(dl_se->dl_bw, dl_rq);
kernel/sched/deadline.c
306
__sub_rq_bw(dl_se->dl_bw, &rq->dl);
kernel/sched/deadline.c
3109
struct dl_bw *src_dl_b;
kernel/sched/deadline.c
3118
__dl_sub(src_dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/deadline.c
3189
struct dl_bw *dl_b;
kernel/sched/deadline.c
3204
dl_b = &rq->rd->dl_bw;
kernel/sched/deadline.c
3207
__dl_add(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span));
kernel/sched/deadline.c
3218
__dl_add(&rd->dl_bw, dl_se->dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3223
__dl_add(&rd->dl_bw, dl_se->dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3229
u64 dl_bw = 0;
kernel/sched/deadline.c
3232
dl_bw += cpu_rq(cpu)->fair_server.dl_bw;
kernel/sched/deadline.c
3236
dl_bw += cpu_rq(cpu)->ext_server.dl_bw;
kernel/sched/deadline.c
3239
return dl_bw;
kernel/sched/deadline.c
3246
guard(raw_spinlock_irqsave)(&rd->dl_bw.lock);
kernel/sched/deadline.c
3252
rd->dl_bw.total_bw = 0;
kernel/sched/deadline.c
3452
struct dl_bw *dl_b;
kernel/sched/deadline.c
3502
struct dl_bw *dl_b;
kernel/sched/deadline.c
3545
struct dl_bw *dl_b = dl_bw_of(cpu);
kernel/sched/deadline.c
3552
if (new_bw == p->dl.dl_bw && task_has_dl_policy(p))
kernel/sched/deadline.c
3567
__dl_sub(dl_b, p->dl.dl_bw, cpus);
kernel/sched/deadline.c
3571
!__dl_overflow(dl_b, cap, p->dl.dl_bw, new_bw)) {
kernel/sched/deadline.c
3579
__dl_sub(dl_b, p->dl.dl_bw, cpus);
kernel/sched/deadline.c
3612
dl_se->dl_bw = to_ratio(dl_se->dl_period, dl_se->dl_runtime);
kernel/sched/deadline.c
3692
dl_se->dl_bw = 0;
kernel/sched/deadline.c
3734
struct dl_bw *cur_dl_b;
kernel/sched/deadline.c
3755
static int dl_bw_manage(enum dl_bw_request req, int cpu, u64 dl_bw)
kernel/sched/deadline.c
3758
struct dl_bw *dl_b;
kernel/sched/deadline.c
3769
__dl_sub(dl_b, dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3772
overflow = __dl_overflow(dl_b, cap, 0, dl_bw);
kernel/sched/deadline.c
3781
__dl_add(dl_b, dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3831
int dl_bw_alloc(int cpu, u64 dl_bw)
kernel/sched/deadline.c
3833
return dl_bw_manage(dl_bw_req_alloc, cpu, dl_bw);
kernel/sched/deadline.c
3836
void dl_bw_free(int cpu, u64 dl_bw)
kernel/sched/deadline.c
3838
dl_bw_manage(dl_bw_req_free, cpu, dl_bw);
kernel/sched/deadline.c
442
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
kernel/sched/deadline.c
447
__dl_sub(dl_b, dl_se->dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/deadline.c
506
void init_dl_bw(struct dl_bw *dl_b)
kernel/sched/deadline.c
648
struct dl_bw *dl_b;
kernel/sched/deadline.c
699
dl_b = &rq->rd->dl_bw;
kernel/sched/deadline.c
701
__dl_sub(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span));
kernel/sched/deadline.c
704
dl_b = &later_rq->rd->dl_bw;
kernel/sched/deadline.c
706
__dl_add(dl_b, p->dl.dl_bw, cpumask_weight(later_rq->rd->span));
kernel/sched/debug.c
1013
struct dl_bw *dl_bw;
kernel/sched/debug.c
1022
dl_bw = &cpu_rq(cpu)->rd->dl_bw;
kernel/sched/debug.c
1023
SEQ_printf(m, " .%-30s: %lld\n", "dl_bw->bw", dl_bw->bw);
kernel/sched/debug.c
1024
SEQ_printf(m, " .%-30s: %lld\n", "dl_bw->total_bw", dl_bw->total_bw);
kernel/sched/sched.h
1013
struct dl_bw dl_bw;
kernel/sched/sched.h
354
extern void init_dl_bw(struct dl_bw *dl_b);
kernel/sched/syscalls.c
632
rq->rd->dl_bw.bw == 0) {
kernel/sched/topology.c
553
init_dl_bw(&rd->dl_bw);