Symbol: arch_spinlock_t
arch/alpha/include/asm/spinlock.h
19
static inline int arch_spin_value_unlocked(arch_spinlock_t lock)
arch/alpha/include/asm/spinlock.h
24
static inline void arch_spin_unlock(arch_spinlock_t * lock)
arch/alpha/include/asm/spinlock.h
30
static inline void arch_spin_lock(arch_spinlock_t * lock)
arch/alpha/include/asm/spinlock.h
50
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/arc/include/asm/smp.h
109
extern arch_spinlock_t smp_atomic_ops_lock;
arch/arc/include/asm/spinlock.h
17
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock.h
222
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock.h
244
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock.h
261
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock.h
44
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock.h
67
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/arc/include/asm/spinlock_types.h
27
arch_spinlock_t lock_mutex;
arch/arc/kernel/smp.c
32
arch_spinlock_t smp_atomic_ops_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/arm/common/mcpm_entry.c
183
static arch_spinlock_t mcpm_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/arm/include/asm/spinlock.h
107
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/arm/include/asm/spinlock.h
114
static inline int arch_spin_value_unlocked(arch_spinlock_t lock)
arch/arm/include/asm/spinlock.h
119
static inline int arch_spin_is_locked(arch_spinlock_t *lock)
arch/arm/include/asm/spinlock.h
124
static inline int arch_spin_is_contended(arch_spinlock_t *lock)
arch/arm/include/asm/spinlock.h
56
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/arm/include/asm/spinlock.h
60
arch_spinlock_t lockval;
arch/arm/include/asm/spinlock.h
81
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/arm/kernel/traps.c
303
static arch_spinlock_t die_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/hexagon/include/asm/spinlock.h
115
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/hexagon/include/asm/spinlock.h
130
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/hexagon/include/asm/spinlock.h
136
static inline unsigned int arch_spin_trylock(arch_spinlock_t *lock)
arch/mips/kernel/sync-r4k.c
38
static arch_spinlock_t sync_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/mips/sgi-ip27/ip27-nmi.c
27
static arch_spinlock_t nmi_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/parisc/include/asm/atomic.h
32
extern arch_spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] __lock_aligned;
arch/parisc/include/asm/atomic.h
37
arch_spinlock_t *s = ATOMIC_HASH(l); \
arch/parisc/include/asm/atomic.h
43
arch_spinlock_t *s = ATOMIC_HASH(l); \
arch/parisc/include/asm/futex.h
109
s = (arch_spinlock_t *)&lws_lock_start[_futex_hash_index(ua)];
arch/parisc/include/asm/futex.h
21
_futex_spin_lock_irqsave(arch_spinlock_t *s, unsigned long *flags)
arch/parisc/include/asm/futex.h
28
_futex_spin_unlock_irqrestore(arch_spinlock_t *s, unsigned long *flags)
arch/parisc/include/asm/futex.h
39
arch_spinlock_t *s;
arch/parisc/include/asm/futex.h
44
s = (arch_spinlock_t *)&lws_lock_start[_futex_hash_index(ua)];
arch/parisc/include/asm/futex.h
95
arch_spinlock_t *s;
arch/parisc/include/asm/spinlock.h
19
static inline int arch_spin_is_locked(arch_spinlock_t *x)
arch/parisc/include/asm/spinlock.h
30
static inline void arch_spin_lock(arch_spinlock_t *x)
arch/parisc/include/asm/spinlock.h
49
static inline void arch_spin_unlock(arch_spinlock_t *x)
arch/parisc/include/asm/spinlock.h
59
static inline int arch_spin_trylock(arch_spinlock_t *x)
arch/parisc/include/asm/spinlock_types.h
25
arch_spinlock_t lock_mutex;
arch/parisc/lib/bitops.c
15
arch_spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] __lock_aligned = {
arch/powerpc/include/asm/simple_spinlock.h
114
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/powerpc/include/asm/simple_spinlock.h
128
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/powerpc/include/asm/simple_spinlock.h
35
static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock)
arch/powerpc/include/asm/simple_spinlock.h
40
static inline int arch_spin_is_locked(arch_spinlock_t *lock)
arch/powerpc/include/asm/simple_spinlock.h
49
static inline unsigned long __arch_spin_trylock(arch_spinlock_t *lock)
arch/powerpc/include/asm/simple_spinlock.h
70
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/powerpc/include/asm/simple_spinlock.h
91
void splpar_spin_yield(arch_spinlock_t *lock);
arch/powerpc/include/asm/simple_spinlock.h
94
static inline void splpar_spin_yield(arch_spinlock_t *lock) {}
arch/powerpc/include/asm/simple_spinlock.h
98
static inline void spin_yield(arch_spinlock_t *lock)
arch/powerpc/kernel/traps.c
155
static arch_spinlock_t die_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/powerpc/kvm/book3s_xics.h
98
arch_spinlock_t lock;
arch/powerpc/kvm/book3s_xive.h
89
arch_spinlock_t lock;
arch/powerpc/lib/locks.c
21
void splpar_spin_yield(arch_spinlock_t *lock)
arch/powerpc/platforms/pasemi/setup.c
101
static arch_spinlock_t timebase_lock;
arch/riscv/include/asm/spinlock.h
31
SPINLOCK_BASE_DECLARE(lock, void, arch_spinlock_t *)
arch/riscv/include/asm/spinlock.h
32
SPINLOCK_BASE_DECLARE(unlock, void, arch_spinlock_t *)
arch/riscv/include/asm/spinlock.h
33
SPINLOCK_BASE_DECLARE(is_locked, int, arch_spinlock_t *)
arch/riscv/include/asm/spinlock.h
34
SPINLOCK_BASE_DECLARE(is_contended, int, arch_spinlock_t *)
arch/riscv/include/asm/spinlock.h
35
SPINLOCK_BASE_DECLARE(trylock, bool, arch_spinlock_t *)
arch/riscv/include/asm/spinlock.h
36
SPINLOCK_BASE_DECLARE(value_unlocked, int, arch_spinlock_t)
arch/s390/include/asm/spinlock.h
52
void arch_spin_relax(arch_spinlock_t *lock);
arch/s390/include/asm/spinlock.h
55
void arch_spin_lock_wait(arch_spinlock_t *);
arch/s390/include/asm/spinlock.h
56
int arch_spin_trylock_retry(arch_spinlock_t *);
arch/s390/include/asm/spinlock.h
64
static inline int arch_spin_value_unlocked(arch_spinlock_t lock)
arch/s390/include/asm/spinlock.h
69
static inline int arch_spin_is_locked(arch_spinlock_t *lp)
arch/s390/include/asm/spinlock.h
74
static inline int arch_spin_trylock_once(arch_spinlock_t *lp)
arch/s390/include/asm/spinlock.h
82
static inline void arch_spin_lock(arch_spinlock_t *lp)
arch/s390/include/asm/spinlock.h
88
static inline int arch_spin_trylock(arch_spinlock_t *lp)
arch/s390/include/asm/spinlock.h
95
static inline void arch_spin_unlock(arch_spinlock_t *lp)
arch/s390/include/asm/spinlock_types.h
17
arch_spinlock_t wait;
arch/s390/kernel/smp.c
426
static arch_spinlock_t lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/s390/lib/spinlock.c
156
static inline void arch_spin_lock_queued(arch_spinlock_t *lp)
arch/s390/lib/spinlock.c
250
static inline void arch_spin_lock_classic(arch_spinlock_t *lp)
arch/s390/lib/spinlock.c
282
void arch_spin_lock_wait(arch_spinlock_t *lp)
arch/s390/lib/spinlock.c
291
int arch_spin_trylock_retry(arch_spinlock_t *lp)
arch/s390/lib/spinlock.c
352
void arch_spin_relax(arch_spinlock_t *lp)
arch/sh/include/asm/spinlock-cas.h
28
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/sh/include/asm/spinlock-cas.h
33
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/sh/include/asm/spinlock-cas.h
38
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/sh/include/asm/spinlock-llsc.h
26
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/sh/include/asm/spinlock-llsc.h
46
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/sh/include/asm/spinlock-llsc.h
61
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/sparc/include/asm/spinlock_32.h
18
static inline void arch_spin_lock(arch_spinlock_t *lock)
arch/sparc/include/asm/spinlock_32.h
38
static inline int arch_spin_trylock(arch_spinlock_t *lock)
arch/sparc/include/asm/spinlock_32.h
48
static inline void arch_spin_unlock(arch_spinlock_t *lock)
arch/x86/kernel/dumpstack.c
345
static arch_spinlock_t die_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/x86/kernel/hpet.c
778
arch_spinlock_t lock;
arch/x86/kernel/tsc_sync.c
255
static arch_spinlock_t sync_lock = __ARCH_SPIN_LOCK_UNLOCKED;
arch/x86/mm/kmmio.c
71
static arch_spinlock_t kmmio_lock = __ARCH_SPIN_LOCK_UNLOCKED;
include/asm-generic/qrwlock_types.h
26
arch_spinlock_t wait_lock;
include/asm-generic/ticket_spinlock.h
33
static __always_inline void ticket_spin_lock(arch_spinlock_t *lock)
include/asm-generic/ticket_spinlock.h
53
static __always_inline bool ticket_spin_trylock(arch_spinlock_t *lock)
include/asm-generic/ticket_spinlock.h
63
static __always_inline void ticket_spin_unlock(arch_spinlock_t *lock)
include/asm-generic/ticket_spinlock.h
71
static __always_inline int ticket_spin_value_unlocked(arch_spinlock_t lock)
include/asm-generic/ticket_spinlock.h
78
static __always_inline int ticket_spin_is_locked(arch_spinlock_t *lock)
include/asm-generic/ticket_spinlock.h
80
arch_spinlock_t val = READ_ONCE(*lock);
include/asm-generic/ticket_spinlock.h
85
static __always_inline int ticket_spin_is_contended(arch_spinlock_t *lock)
include/linux/spinlock_types_raw.h
15
arch_spinlock_t raw_lock;
include/linux/spinlock_up.h
29
static inline void arch_spin_lock(arch_spinlock_t *lock)
include/linux/spinlock_up.h
35
static inline int arch_spin_trylock(arch_spinlock_t *lock)
include/linux/spinlock_up.h
45
static inline void arch_spin_unlock(arch_spinlock_t *lock)
kernel/bpf/helpers.c
288
arch_spinlock_t *l = (void *)lock;
kernel/bpf/helpers.c
291
arch_spinlock_t lock;
kernel/bpf/helpers.c
303
arch_spinlock_t *l = (void *)lock;
kernel/kcsan/kcsan_test.c
535
arch_spinlock_t arch_spinlock = __ARCH_SPIN_LOCK_UNLOCKED;
kernel/kcsan/selftest.c
118
arch_spinlock_t arch_spinlock = __ARCH_SPIN_LOCK_UNLOCKED;
kernel/locking/lockdep.c
138
static arch_spinlock_t __lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/locking/spinlock_debug.c
27
lock->raw_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/rcu/tree.h
422
arch_spinlock_t ofl_lock ____cacheline_internodealigned_in_smp;
kernel/trace/ring_buffer.c
2368
cpu_buffer->lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/trace/ring_buffer.c
519
arch_spinlock_t lock;
kernel/trace/trace.c
10873
global_trace.max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/trace/trace.c
9601
tr->max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/trace/trace.h
389
arch_spinlock_t max_lock;
kernel/trace/trace_clock.c
88
arch_spinlock_t lock;
kernel/trace/trace_clock.c
91
.lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED,
kernel/trace/trace_sched_switch.c
169
static arch_spinlock_t trace_cmdline_lock = __ARCH_SPIN_LOCK_UNLOCKED;
kernel/trace/trace_sched_wakeup.c
33
static arch_spinlock_t wakeup_lock =
kernel/trace/trace_sched_wakeup.c
34
(arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
kernel/trace/trace_stack.c
29
static arch_spinlock_t stack_trace_max_lock =
kernel/trace/trace_stack.c
30
(arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
lib/atomic64.c
107
arch_spinlock_t *lock = lock_addr(v); \
lib/atomic64.c
144
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
161
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
178
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
194
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
28
arch_spinlock_t lock;
lib/atomic64.c
36
static inline arch_spinlock_t *lock_addr(const atomic64_t *v)
lib/atomic64.c
48
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
63
arch_spinlock_t *lock = lock_addr(v);
lib/atomic64.c
77
arch_spinlock_t *lock = lock_addr(v); \
lib/atomic64.c
91
arch_spinlock_t *lock = lock_addr(v); \
tools/include/linux/spinlock.h
26
static inline void arch_spin_lock(arch_spinlock_t *mutex)
tools/include/linux/spinlock.h
31
static inline void arch_spin_unlock(arch_spinlock_t *mutex)
tools/include/linux/spinlock.h
36
static inline bool arch_spin_is_locked(arch_spinlock_t *mutex)