Symbol: rt_bandwidth
kernel/sched/core.c
8651
init_rt_bandwidth(&root_task_group.rt_bandwidth,
kernel/sched/rt.c
103
struct rt_bandwidth *rt_b =
kernel/sched/rt.c
104
container_of(timer, struct rt_bandwidth, rt_period_timer);
kernel/sched/rt.c
1125
start_rt_bandwidth(&rt_rq->tg->rt_bandwidth);
kernel/sched/rt.c
125
void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime)
kernel/sched/rt.c
136
static inline void do_start_rt_bandwidth(struct rt_bandwidth *rt_b)
kernel/sched/rt.c
156
static void start_rt_bandwidth(struct rt_bandwidth *rt_b)
kernel/sched/rt.c
164
static void destroy_rt_bandwidth(struct rt_bandwidth *rt_b)
kernel/sched/rt.c
205
destroy_rt_bandwidth(&tg->rt_bandwidth);
kernel/sched/rt.c
2658
period = ktime_to_ns(tg->rt_bandwidth.rt_period);
kernel/sched/rt.c
2659
runtime = tg->rt_bandwidth.rt_runtime;
kernel/sched/rt.c
2676
tg->rt_bandwidth.rt_runtime && tg_has_rt_tasks(tg))
kernel/sched/rt.c
269
init_rt_bandwidth(&tg->rt_bandwidth, ktime_to_ns(global_rt_period()), 0);
kernel/sched/rt.c
2694
period = ktime_to_ns(child->rt_bandwidth.rt_period);
kernel/sched/rt.c
2695
runtime = child->rt_bandwidth.rt_runtime;
kernel/sched/rt.c
2755
raw_spin_lock_irq(&tg->rt_bandwidth.rt_runtime_lock);
kernel/sched/rt.c
2756
tg->rt_bandwidth.rt_period = ns_to_ktime(rt_period);
kernel/sched/rt.c
2757
tg->rt_bandwidth.rt_runtime = rt_runtime;
kernel/sched/rt.c
2766
raw_spin_unlock_irq(&tg->rt_bandwidth.rt_runtime_lock);
kernel/sched/rt.c
2777
rt_period = ktime_to_ns(tg->rt_bandwidth.rt_period);
kernel/sched/rt.c
2791
if (tg->rt_bandwidth.rt_runtime == RUNTIME_INF)
kernel/sched/rt.c
2794
rt_runtime_us = tg->rt_bandwidth.rt_runtime;
kernel/sched/rt.c
2807
rt_runtime = tg->rt_bandwidth.rt_runtime;
kernel/sched/rt.c
2816
rt_period_us = ktime_to_ns(tg->rt_bandwidth.rt_period);
kernel/sched/rt.c
283
rt_rq->rt_runtime = tg->rt_bandwidth.rt_runtime;
kernel/sched/rt.c
2837
if (rt_group_sched_enabled() && rt_task(tsk) && tg->rt_bandwidth.rt_runtime == 0)
kernel/sched/rt.c
488
return ktime_to_ns(rt_rq->tg->rt_bandwidth.rt_period);
kernel/sched/rt.c
587
struct rt_rq *sched_rt_period_rt_rq(struct rt_bandwidth *rt_b, int cpu)
kernel/sched/rt.c
589
return container_of(rt_b, struct task_group, rt_bandwidth)->rt_rq[cpu];
kernel/sched/rt.c
592
static inline struct rt_bandwidth *sched_rt_bandwidth(struct rt_rq *rt_rq)
kernel/sched/rt.c
594
return &rt_rq->tg->rt_bandwidth;
kernel/sched/rt.c
599
struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
kernel/sched/rt.c
610
struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
kernel/sched/rt.c
670
struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
kernel/sched/rt.c
754
struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
kernel/sched/rt.c
778
static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun)
kernel/sched/rt.c
794
if (rt_b == &root_task_group.rt_bandwidth)
kernel/sched/rt.c
879
struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq);
kernel/sched/rt.c
948
struct rt_rq *sched_rt_period_rt_rq(struct rt_bandwidth *rt_b, int cpu)
kernel/sched/rt.c
99
static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun);
kernel/sched/sched.h
2889
extern void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime);
kernel/sched/sched.h
500
struct rt_bandwidth rt_bandwidth;
kernel/sched/syscalls.c
616
task_group(p)->rt_bandwidth.rt_runtime == 0 &&