Symbol: __lock
drivers/block/zram/zram_drv.c
105
unsigned long *lock = &zram->table[index].__lock;
drivers/block/zram/zram_drv.c
83
unsigned long *lock = &zram->table[index].__lock;
drivers/block/zram/zram_drv.c
96
unsigned long *lock = &zram->table[index].__lock;
drivers/block/zram/zram_drv.h
69
unsigned long __lock;
drivers/gpu/drm/drm_gpuvm.c
913
#define drm_gpuvm_bo_list_add(__vm_bo, __list_name, __lock) \
drivers/gpu/drm/drm_gpuvm.c
915
__lock ? &(__vm_bo)->vm->__list_name.lock : \
drivers/gpu/drm/drm_gpuvm.c
942
#define drm_gpuvm_bo_list_del_init(__vm_bo, __list_name, __lock) \
drivers/gpu/drm/drm_gpuvm.c
944
__lock ? &(__vm_bo)->vm->__list_name.lock : \
drivers/gpu/drm/drm_gpuvm.c
957
#define drm_gpuvm_bo_list_del(__vm_bo, __list_name, __lock) \
drivers/gpu/drm/drm_gpuvm.c
959
__lock ? &(__vm_bo)->vm->__list_name.lock : \
drivers/md/dm-bio-prison-v2.c
266
r = __lock(prison, key, lock_level, cell_prealloc, cell_result);
include/linux/local_lock_internal.h
254
#define __local_lock(__lock) \
include/linux/local_lock_internal.h
257
spin_lock((__lock)); \
include/linux/local_lock_internal.h
269
#define __local_unlock(__lock) \
include/linux/local_lock_internal.h
271
spin_unlock((__lock)); \
include/linux/local_lock_internal.h
313
#define __local_lock_is_locked(__lock) \
include/linux/local_lock_internal.h
314
(rt_mutex_owner(&this_cpu_ptr(__lock)->lock) == current)
include/linux/maple_tree.h
252
#define MTREE_INIT_EXT(name, __flags, __lock) { \
include/linux/maple_tree.h
253
.ma_external_lock = &(__lock).dep_map, \
include/linux/maple_tree.h
258
#define MTREE_INIT_EXT(name, __flags, __lock) MTREE_INIT(name, __flags)
kernel/locking/lockdep.c
138
static arch_spinlock_t __lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/locking/lockdep.c
146
arch_spin_lock(&__lock);
kernel/locking/lockdep.c
158
arch_spin_unlock(&__lock);
kernel/sched/core.c
403
__acquires(&runqueues.__lock) /* overapproximation */
kernel/sched/core.c
410
raw_spin_lock_nested(&cpu_rq(t)->__lock, i++);
kernel/sched/core.c
415
__releases(&runqueues.__lock) /* overapproximation */
kernel/sched/core.c
421
raw_spin_unlock(&cpu_rq(t)->__lock);
kernel/sched/core.c
647
raw_spin_lock_nested(&rq->__lock, subclass);
kernel/sched/core.c
674
ret = raw_spin_trylock(&rq->__lock);
kernel/sched/core.c
8668
raw_spin_lock_init(&rq->__lock);
kernel/sched/sched.h
1159
raw_spinlock_t __lock;
kernel/sched/sched.h
1452
return &rq->core->__lock;
kernel/sched/sched.h
1454
return &rq->__lock;
kernel/sched/sched.h
1461
return &rq->core->__lock;
kernel/sched/sched.h
1463
return &rq->__lock;
kernel/sched/sched.h
1554
return &rq->__lock;
kernel/sched/sched.h
1560
return &rq->__lock;
tools/perf/util/bpf_skel/lock_contention.bpf.c
860
raw_spinlock_t __lock;
tools/perf/util/bpf_skel/lock_contention.bpf.c
922
if (bpf_core_field_exists(struct rq___new, __lock))
tools/perf/util/bpf_skel/lock_contention.bpf.c
923
lock_off = offsetof(struct rq___new, __lock);