arch/arm64/kernel/process.c
705
__notrace_funcgraph __sched
drivers/tty/tty_ldsem.c
155
static struct ld_semaphore __sched *
drivers/tty/tty_ldsem.c
229
static struct ld_semaphore __sched *
drivers/tty/tty_ldsem.c
337
int __sched ldsem_down_read(struct ld_semaphore *sem, long timeout)
drivers/tty/tty_ldsem.c
363
int __sched ldsem_down_write(struct ld_semaphore *sem, long timeout)
include/drm/gpu_scheduler.h
732
__drm_sched_pending_job_iter_begin(__sched),
include/drm/gpu_scheduler.h
733
struct drm_gpu_scheduler *__sched);
include/drm/gpu_scheduler.h
748
#define drm_sched_for_each_pending_job(__job, __sched, __entity) \
include/drm/gpu_scheduler.h
749
scoped_guard(drm_sched_pending_job_iter, (__sched)) \
include/drm/gpu_scheduler.h
750
list_for_each_entry((__job), &(__sched)->pending_list, list) \
kernel/latencytop.c
110
static void __sched
kernel/latencytop.c
175
void __sched
kernel/locking/mutex.c
1007
int __sched mutex_lock_interruptible(struct mutex *lock)
kernel/locking/mutex.c
1031
int __sched mutex_lock_killable(struct mutex *lock)
kernel/locking/mutex.c
1052
void __sched mutex_lock_io(struct mutex *lock)
kernel/locking/mutex.c
1062
static noinline void __sched
kernel/locking/mutex.c
1068
static noinline int __sched
kernel/locking/mutex.c
1074
static noinline int __sched
kernel/locking/mutex.c
1080
static noinline int __sched
kernel/locking/mutex.c
1087
static noinline int __sched
kernel/locking/mutex.c
1112
int __sched mutex_trylock(struct mutex *lock)
kernel/locking/mutex.c
1119
int __sched _mutex_trylock_nest_lock(struct mutex *lock, struct lockdep_map *nest_lock)
kernel/locking/mutex.c
1134
int __sched
kernel/locking/mutex.c
1149
int __sched
kernel/locking/mutex.c
262
static void __sched __mutex_lock_slowpath(struct mutex *lock);
kernel/locking/mutex.c
285
void __sched mutex_lock(struct mutex *lock)
kernel/locking/mutex.c
528
static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip);
kernel/locking/mutex.c
546
void __sched mutex_unlock(struct mutex *lock)
kernel/locking/mutex.c
567
void __sched ww_mutex_unlock(struct ww_mutex *lock)
kernel/locking/mutex.c
577
static __always_inline int __sched
kernel/locking/mutex.c
772
static int __sched
kernel/locking/mutex.c
779
static int __sched
kernel/locking/mutex.c
825
void __sched
kernel/locking/mutex.c
833
void __sched
kernel/locking/mutex.c
840
int __sched
kernel/locking/mutex.c
848
int __sched
kernel/locking/mutex.c
855
void __sched
kernel/locking/mutex.c
895
int __sched
kernel/locking/mutex.c
910
int __sched
kernel/locking/mutex.c
931
static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip)
kernel/locking/mutex.c
989
static noinline int __sched
kernel/locking/mutex.c
992
static noinline int __sched
kernel/locking/percpu-rwsem.c
169
bool __sched __percpu_down_read(struct percpu_rw_semaphore *sem, bool try,
kernel/locking/percpu-rwsem.c
227
void __sched percpu_down_write(struct percpu_rw_semaphore *sem)
kernel/locking/rtmutex.c
1086
static int __sched
kernel/locking/rtmutex.c
1203
static int __sched task_blocks_on_rt_mutex(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1312
static void __sched mark_wakeup_next_waiter(struct rt_wake_q_head *wqh,
kernel/locking/rtmutex.c
1358
static int __sched __rt_mutex_slowtrylock(struct rt_mutex_base *lock)
kernel/locking/rtmutex.c
1374
static int __sched rt_mutex_slowtrylock(struct rt_mutex_base *lock)
kernel/locking/rtmutex.c
1411
static void __sched rt_mutex_slowunlock(struct rt_mutex_base *lock)
kernel/locking/rtmutex.c
1539
static void __sched remove_waiter(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1604
static int __sched rt_mutex_slowlock_block(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1658
static void __sched rt_mutex_handle_deadlock(int res, int detect_deadlock,
kernel/locking/rtmutex.c
1691
static int __sched __rt_mutex_slowlock(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1774
static int __sched rt_mutex_slowlock(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1829
static void __sched rtlock_slowlock_locked(struct rt_mutex_base *lock,
kernel/locking/rtmutex.c
1889
static __always_inline void __sched rtlock_slowlock(struct rt_mutex_base *lock)
kernel/locking/rtmutex.c
309
static int __sched rt_mutex_slowtrylock(struct rt_mutex_base *lock);
kernel/locking/rtmutex.c
678
static int __sched rt_mutex_adjust_prio_chain(struct task_struct *task,
kernel/locking/rtmutex_api.c
103
int __sched rt_mutex_lock_interruptible(struct rt_mutex *lock)
kernel/locking/rtmutex_api.c
118
int __sched rt_mutex_lock_killable(struct rt_mutex *lock)
kernel/locking/rtmutex_api.c
136
int __sched rt_mutex_trylock(struct rt_mutex *lock)
kernel/locking/rtmutex_api.c
156
void __sched rt_mutex_unlock(struct rt_mutex *lock)
kernel/locking/rtmutex_api.c
166
int __sched rt_mutex_futex_trylock(struct rt_mutex_base *lock)
kernel/locking/rtmutex_api.c
171
int __sched __rt_mutex_futex_trylock(struct rt_mutex_base *lock)
kernel/locking/rtmutex_api.c
183
bool __sched __rt_mutex_futex_unlock(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
206
void __sched rt_mutex_futex_unlock(struct rt_mutex_base *lock)
kernel/locking/rtmutex_api.c
231
void __sched __rt_mutex_init(struct rt_mutex *lock, const char *name,
kernel/locking/rtmutex_api.c
254
void __sched rt_mutex_init_proxy_locked(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
285
void __sched rt_mutex_proxy_unlock(struct rt_mutex_base *lock)
kernel/locking/rtmutex_api.c
311
int __sched __rt_mutex_start_proxy_lock(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
359
int __sched rt_mutex_start_proxy_lock(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
395
int __sched rt_mutex_wait_proxy_lock(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
435
bool __sched rt_mutex_cleanup_proxy_lock(struct rt_mutex_base *lock,
kernel/locking/rtmutex_api.c
477
void __sched rt_mutex_adjust_pi(struct task_struct *task)
kernel/locking/rtmutex_api.c
503
void __sched rt_mutex_postunlock(struct rt_wake_q_head *wqh)
kernel/locking/rtmutex_api.c
550
void __sched mutex_lock_nested(struct mutex *lock, unsigned int subclass)
kernel/locking/rtmutex_api.c
556
void __sched _mutex_lock_nest_lock(struct mutex *lock,
kernel/locking/rtmutex_api.c
563
int __sched mutex_lock_interruptible_nested(struct mutex *lock,
kernel/locking/rtmutex_api.c
570
int __sched _mutex_lock_killable(struct mutex *lock, unsigned int subclass,
kernel/locking/rtmutex_api.c
577
void __sched mutex_lock_io_nested(struct mutex *lock, unsigned int subclass)
kernel/locking/rtmutex_api.c
589
int __sched _mutex_trylock_nest_lock(struct mutex *lock,
kernel/locking/rtmutex_api.c
612
void __sched mutex_lock(struct mutex *lock)
kernel/locking/rtmutex_api.c
618
int __sched mutex_lock_interruptible(struct mutex *lock)
kernel/locking/rtmutex_api.c
624
int __sched mutex_lock_killable(struct mutex *lock)
kernel/locking/rtmutex_api.c
630
void __sched mutex_lock_io(struct mutex *lock)
kernel/locking/rtmutex_api.c
639
int __sched mutex_trylock(struct mutex *lock)
kernel/locking/rtmutex_api.c
649
void __sched mutex_unlock(struct mutex *lock)
kernel/locking/rtmutex_api.c
68
void __sched rt_mutex_lock_nested(struct rt_mutex *lock, unsigned int subclass)
kernel/locking/rtmutex_api.c
74
void __sched _rt_mutex_lock_nest_lock(struct rt_mutex *lock, struct lockdep_map *nest_lock)
kernel/locking/rtmutex_api.c
87
void __sched rt_mutex_lock(struct rt_mutex *lock)
kernel/locking/rwbase_rt.c
150
static void __sched __rwbase_read_unlock(struct rwbase_rt *rwb,
kernel/locking/rwbase_rt.c
237
static int __sched rwbase_write_lock(struct rwbase_rt *rwb,
kernel/locking/rwbase_rt.c
68
static int __sched __rwbase_read_lock(struct rwbase_rt *rwb,
kernel/locking/rwsem.c
1110
static struct rw_semaphore __sched *
kernel/locking/rwsem.c
1489
static inline void __sched __down_write(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1494
static inline int __sched __down_write_killable(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1534
void __sched down_read(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1543
int __sched down_read_interruptible(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1557
int __sched down_read_killable(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1587
void __sched down_write(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1598
int __sched down_write_killable(struct rw_semaphore *sem)
kernel/locking/rwsem.c
1711
int __sched down_write_killable_nested(struct rw_semaphore *sem, int subclass)
kernel/locking/rwsem.c
992
static struct rw_semaphore __sched *
kernel/locking/semaphore.c
114
int __sched down_interruptible(struct semaphore *sem)
kernel/locking/semaphore.c
141
int __sched down_killable(struct semaphore *sem)
kernel/locking/semaphore.c
171
int __sched down_trylock(struct semaphore *sem)
kernel/locking/semaphore.c
196
int __sched down_timeout(struct semaphore *sem, long timeout)
kernel/locking/semaphore.c
220
void __sched up(struct semaphore *sem)
kernel/locking/semaphore.c
252
static inline int __sched ___down_common(struct semaphore *sem, long state,
kernel/locking/semaphore.c
285
static inline int __sched __down_common(struct semaphore *sem, long state,
kernel/locking/semaphore.c
301
static noinline void __sched __down(struct semaphore *sem)
kernel/locking/semaphore.c
306
static noinline int __sched __down_interruptible(struct semaphore *sem)
kernel/locking/semaphore.c
311
static noinline int __sched __down_killable(struct semaphore *sem)
kernel/locking/semaphore.c
316
static noinline int __sched __down_timeout(struct semaphore *sem, long timeout)
kernel/locking/semaphore.c
321
static noinline void __sched __up(struct semaphore *sem,
kernel/locking/semaphore.c
91
void __sched down(struct semaphore *sem)
kernel/locking/spinlock_rt.c
116
int __sched rt_spin_trylock(spinlock_t *lock)
kernel/locking/spinlock_rt.c
122
int __sched rt_spin_trylock_bh(spinlock_t *lock)
kernel/locking/spinlock_rt.c
201
int __sched rt_read_trylock(rwlock_t *rwlock)
kernel/locking/spinlock_rt.c
215
int __sched rt_write_trylock(rwlock_t *rwlock)
kernel/locking/spinlock_rt.c
229
void __sched rt_read_lock(rwlock_t *rwlock) __acquires(RCU)
kernel/locking/spinlock_rt.c
239
void __sched rt_write_lock(rwlock_t *rwlock) __acquires(RCU)
kernel/locking/spinlock_rt.c
250
void __sched rt_write_lock_nested(rwlock_t *rwlock, int subclass) __acquires(RCU)
kernel/locking/spinlock_rt.c
261
void __sched rt_read_unlock(rwlock_t *rwlock) __releases(RCU)
kernel/locking/spinlock_rt.c
270
void __sched rt_write_unlock(rwlock_t *rwlock) __releases(RCU)
kernel/locking/spinlock_rt.c
54
void __sched rt_spin_lock(spinlock_t *lock) __acquires(RCU)
kernel/locking/spinlock_rt.c
62
void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass)
kernel/locking/spinlock_rt.c
69
void __sched rt_spin_lock_nest_lock(spinlock_t *lock,
kernel/locking/spinlock_rt.c
78
void __sched rt_spin_unlock(spinlock_t *lock) __releases(RCU)
kernel/locking/spinlock_rt.c
94
void __sched rt_spin_lock_unlock(spinlock_t *lock)
kernel/locking/ww_rt_mutex.c
37
static int __sched
kernel/locking/ww_rt_mutex.c
78
int __sched
kernel/locking/ww_rt_mutex.c
85
int __sched
kernel/locking/ww_rt_mutex.c
92
void __sched ww_mutex_unlock(struct ww_mutex *lock)
kernel/sched/completion.c
112
static inline long __sched
kernel/sched/completion.c
129
static long __sched
kernel/sched/completion.c
135
static long __sched
kernel/sched/completion.c
151
void __sched wait_for_completion(struct completion *x)
kernel/sched/completion.c
169
unsigned long __sched
kernel/sched/completion.c
184
void __sched wait_for_completion_io(struct completion *x)
kernel/sched/completion.c
203
unsigned long __sched
kernel/sched/completion.c
219
int __sched wait_for_completion_interruptible(struct completion *x)
kernel/sched/completion.c
240
long __sched
kernel/sched/completion.c
257
int __sched wait_for_completion_killable(struct completion *x)
kernel/sched/completion.c
267
int __sched wait_for_completion_state(struct completion *x, unsigned int state)
kernel/sched/completion.c
289
long __sched
kernel/sched/completion.c
85
static inline long __sched
kernel/sched/core.c
6764
static void __sched notrace __schedule(int sched_mode)
kernel/sched/core.c
6998
asmlinkage __visible void __sched schedule(void)
kernel/sched/core.c
7023
void __sched schedule_idle(void)
kernel/sched/core.c
7039
asmlinkage __visible void __sched schedule_user(void)
kernel/sched/core.c
7062
void __sched schedule_preempt_disabled(void)
kernel/sched/core.c
7070
void __sched notrace schedule_rtlock(void)
kernel/sched/core.c
7077
static void __sched notrace preempt_schedule_common(void)
kernel/sched/core.c
7111
asmlinkage __visible void __sched notrace preempt_schedule(void)
kernel/sched/core.c
7134
void __sched notrace dynamic_preempt_schedule(void)
kernel/sched/core.c
7159
asmlinkage __visible void __sched notrace preempt_schedule_notrace(void)
kernel/sched/core.c
7207
void __sched notrace dynamic_preempt_schedule_notrace(void)
kernel/sched/core.c
7226
asmlinkage __visible void __sched preempt_schedule_irq(void)
kernel/sched/core.c
7419
int __sched __cond_resched(void)
kernel/sched/core.c
7459
int __sched dynamic_cond_resched(void)
kernel/sched/core.c
7468
int __sched dynamic_might_resched(void)
kernel/sched/core.c
7817
long __sched io_schedule_timeout(long timeout)
kernel/sched/core.c
7830
void __sched io_schedule(void)
kernel/sched/syscalls.c
1382
void __sched yield(void)
kernel/sched/syscalls.c
1404
int __sched yield_to(struct task_struct *p, bool preempt)
kernel/sched/wait_bit.c
114
int __sched out_of_line_wait_on_bit_lock(unsigned long *word, int bit,
kernel/sched/wait_bit.c
238
__sched int bit_wait(struct wait_bit_key *word, int mode)
kernel/sched/wait_bit.c
248
__sched int bit_wait_io(struct wait_bit_key *word, int mode)
kernel/sched/wait_bit.c
258
__sched int bit_wait_timeout(struct wait_bit_key *word, int mode)
kernel/sched/wait_bit.c
43
int __sched
kernel/sched/wait_bit.c
61
int __sched out_of_line_wait_on_bit(unsigned long *word, int bit,
kernel/sched/wait_bit.c
71
int __sched out_of_line_wait_on_bit_timeout(
kernel/sched/wait_bit.c
84
int __sched
kernel/time/alarmtimer.c
752
static long __sched alarm_timer_nsleep_restart(struct restart_block *restart)
kernel/time/hrtimer.c
2115
static int __sched do_nanosleep(struct hrtimer_sleeper *t, enum hrtimer_mode mode)
kernel/time/hrtimer.c
2150
static long __sched hrtimer_nanosleep_restart(struct restart_block *restart)
kernel/time/sleep_timeout.c
125
signed long __sched schedule_timeout_interruptible(signed long timeout)
kernel/time/sleep_timeout.c
140
signed long __sched schedule_timeout_killable(signed long timeout)
kernel/time/sleep_timeout.c
155
signed long __sched schedule_timeout_uninterruptible(signed long timeout)
kernel/time/sleep_timeout.c
172
signed long __sched schedule_timeout_idle(signed long timeout)
kernel/time/sleep_timeout.c
189
int __sched schedule_hrtimeout_range_clock(ktime_t *expires, u64 delta,
kernel/time/sleep_timeout.c
260
int __sched schedule_hrtimeout_range(ktime_t *expires, u64 delta,
kernel/time/sleep_timeout.c
276
int __sched schedule_hrtimeout(ktime_t *expires, const enum hrtimer_mode mode)
kernel/time/sleep_timeout.c
362
void __sched usleep_range_state(unsigned long min, unsigned long max, unsigned int state)
kernel/time/sleep_timeout.c
61
signed long __sched schedule_timeout(signed long timeout)
lib/closure.c
135
void __sched __closure_sync(struct closure *cl)
lib/closure.c
160
void __sched closure_return_sync(struct closure *cl)
lib/closure.c
188
int __sched __closure_sync_timeout(struct closure *cl, unsigned long timeout)