Symbol: preempt_count
arch/alpha/include/asm/thread_info.h
23
int preempt_count; /* 0 => preemptable, <0 => BUG */
arch/alpha/include/asm/thread_info.h
38
.preempt_count = INIT_PREEMPT_COUNT, \
arch/arc/include/asm/thread_info.h
41
int preempt_count; /* 0 => preemptible, <0 => BUG */
arch/arc/include/asm/thread_info.h
56
.preempt_count = INIT_PREEMPT_COUNT, \
arch/arc/kernel/asm-offsets.c
33
offsetof(struct thread_info, preempt_count));
arch/arm/include/asm/thread_info.h
65
int preempt_count; /* 0 => preemptable, <0 => bug */
arch/arm/include/asm/thread_info.h
81
.preempt_count = INIT_PREEMPT_COUNT, \
arch/arm/kernel/asm-offsets.c
49
DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/arm64/include/asm/preempt.h
22
task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \
arch/arm64/include/asm/preempt.h
26
task_thread_info(p)->preempt_count = PREEMPT_DISABLED; \
arch/arm64/include/asm/preempt.h
61
u64 pc = READ_ONCE(ti->preempt_count);
arch/arm64/include/asm/preempt.h
73
return !pc || !READ_ONCE(ti->preempt_count);
arch/arm64/include/asm/preempt.h
78
u64 pc = READ_ONCE(current_thread_info()->preempt_count);
arch/arm64/include/asm/thread_info.h
123
.preempt_count = INIT_PREEMPT_COUNT, \
arch/arm64/include/asm/thread_info.h
30
u64 preempt_count; /* 0 => preemptible, <0 => bug */
arch/arm64/kernel/asm-offsets.c
31
DEFINE(TSK_TI_PREEMPT, offsetof(struct task_struct, thread_info.preempt_count));
arch/csky/include/asm/thread_info.h
17
int preempt_count;
arch/csky/include/asm/thread_info.h
27
.preempt_count = INIT_PREEMPT_COUNT, \
arch/csky/kernel/asm-offsets.c
28
DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/hexagon/include/asm/thread_info.h
35
int preempt_count; /* 0=>preemptible,<0=>BUG */
arch/hexagon/include/asm/thread_info.h
63
.preempt_count = 1, \
arch/loongarch/include/asm/thread_info.h
29
int preempt_count; /* 0 => preemptible, <0 => BUG */
arch/loongarch/include/asm/thread_info.h
43
.preempt_count = INIT_PREEMPT_COUNT, \
arch/loongarch/kernel/asm-offsets.c
91
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
arch/m68k/include/asm/thread_info.h
30
int preempt_count; /* 0 => preemptable, <0 => BUG */
arch/m68k/include/asm/thread_info.h
39
.preempt_count = INIT_PREEMPT_COUNT, \
arch/m68k/kernel/asm-offsets.c
43
DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/microblaze/include/asm/thread_info.h
64
__s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/
arch/microblaze/include/asm/thread_info.h
77
.preempt_count = INIT_PREEMPT_COUNT, \
arch/microblaze/kernel/asm-offsets.c
91
DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count));
arch/mips/include/asm/thread_info.h
30
int preempt_count; /* 0 => preemptible, <0 => BUG */
arch/mips/include/asm/thread_info.h
43
.preempt_count = INIT_PREEMPT_COUNT, \
arch/mips/kernel/asm-offsets.c
110
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
arch/nios2/include/asm/thread_info.h
40
int preempt_count; /* 0 => preemptable,<0 => BUG */
arch/nios2/include/asm/thread_info.h
54
.preempt_count = INIT_PREEMPT_COUNT, \
arch/nios2/kernel/asm-offsets.c
71
OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count);
arch/openrisc/include/asm/thread_info.h
47
__s32 preempt_count; /* 0 => preemptable, <0 => BUG */
arch/openrisc/include/asm/thread_info.h
67
.preempt_count = INIT_PREEMPT_COUNT, \
arch/openrisc/kernel/asm-offsets.c
50
DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/parisc/include/asm/thread_info.h
11
int preempt_count; /* 0=premptable, <0=BUG; will also serve as bh-counter */
arch/parisc/include/asm/thread_info.h
20
.preempt_count = INIT_PREEMPT_COUNT, \
arch/parisc/kernel/asm-offsets.c
224
DEFINE(TI_PRE_COUNT, offsetof(struct task_struct, thread_info.preempt_count));
arch/powerpc/include/asm/thread_info.h
55
int preempt_count; /* 0 => preemptable,
arch/powerpc/include/asm/thread_info.h
80
.preempt_count = INIT_PREEMPT_COUNT, \
arch/powerpc/kernel/interrupt.c
406
if (preempt_count() == 0)
arch/powerpc/kexec/core_64.c
357
current_thread_info()->preempt_count = HARDIRQ_OFFSET;
arch/riscv/include/asm/thread_info.h
55
int preempt_count; /* 0=>preemptible, <0=>BUG */
arch/riscv/include/asm/thread_info.h
97
.preempt_count = INIT_PREEMPT_COUNT, \
arch/riscv/kernel/asm-offsets.c
41
OFFSET(TASK_TI_PREEMPT_COUNT, task_struct, thread_info.preempt_count);
arch/s390/include/asm/lowcore.h
163
__s32 preempt_count; /* 0x03a8 */
arch/s390/include/asm/preempt.h
101
__atomic_add(val, &get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
120
lc_preempt = offsetof(struct lowcore, preempt_count);
arch/s390/include/asm/preempt.h
125
: "=@cc" (cc), "+m" (((struct lowcore *)0)->preempt_count)
arch/s390/include/asm/preempt.h
130
return __atomic_add_const_and_test(-1, &get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
139
return unlikely(READ_ONCE(get_lowcore()->preempt_count) == preempt_offset);
arch/s390/include/asm/preempt.h
31
BUILD_BUG_ON(sizeof_field(struct lowcore, preempt_count) != sizeof(int));
arch/s390/include/asm/preempt.h
32
lc_preempt = offsetof(struct lowcore, preempt_count);
arch/s390/include/asm/preempt.h
41
"m" (((struct lowcore *)0)->preempt_count));
arch/s390/include/asm/preempt.h
49
old = READ_ONCE(get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
52
} while (!arch_try_cmpxchg(&get_lowcore()->preempt_count, &old, new));
arch/s390/include/asm/preempt.h
66
__atomic_and(~PREEMPT_NEED_RESCHED, &get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
71
__atomic_or(PREEMPT_NEED_RESCHED, &get_lowcore()->preempt_count);
arch/s390/include/asm/preempt.h
76
return !(READ_ONCE(get_lowcore()->preempt_count) & PREEMPT_NEED_RESCHED);
arch/s390/include/asm/preempt.h
89
lc_preempt = offsetof(struct lowcore, preempt_count);
arch/s390/include/asm/preempt.h
94
: "+m" (((struct lowcore *)0)->preempt_count)
arch/s390/kernel/early.c
190
lc->preempt_count = INIT_PREEMPT_COUNT;
arch/s390/kernel/setup.c
415
lc->preempt_count = get_lowcore()->preempt_count;
arch/s390/kernel/setup.c
447
lc->preempt_count = PREEMPT_DISABLED;
arch/s390/kernel/smp.c
204
lc->preempt_count = PREEMPT_DISABLED;
arch/sh/include/asm/thread_info.h
32
int preempt_count; /* 0 => preemptable, <0 => BUG */
arch/sh/include/asm/thread_info.h
59
.preempt_count = INIT_PREEMPT_COUNT, \
arch/sh/kernel/asm-offsets.c
28
DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count));
arch/sh/kernel/irq.c
130
irqctx->tinfo.preempt_count = HARDIRQ_OFFSET;
arch/sh/kernel/irq.c
138
irqctx->tinfo.preempt_count = 0;
arch/sh/kernel/irq.c
96
irqctx->tinfo.preempt_count =
arch/sh/kernel/irq.c
97
(irqctx->tinfo.preempt_count & ~SOFTIRQ_MASK) |
arch/sh/kernel/irq.c
98
(curctx->tinfo.preempt_count & SOFTIRQ_MASK);
arch/sparc/include/asm/thread_info_32.h
33
int preempt_count; /* 0 => preemptable,
arch/sparc/include/asm/thread_info_32.h
63
.preempt_count = INIT_PREEMPT_COUNT, \
arch/sparc/include/asm/thread_info_64.h
118
.preempt_count = INIT_PREEMPT_COUNT, \
arch/sparc/include/asm/thread_info_64.h
47
int preempt_count; /* 0 => preemptable, <0 => BUG */
arch/sparc/kernel/traps_32.c
376
TI_PREEMPT != offsetof(struct thread_info, preempt_count) ||
arch/sparc/kernel/traps_64.c
2859
preempt_count) ||
arch/um/include/asm/thread_info.h
22
int preempt_count; /* 0 => preemptable,
arch/um/include/asm/thread_info.h
30
.preempt_count = INIT_PREEMPT_COUNT, \
arch/xtensa/include/asm/thread_info.h
53
__s32 preempt_count; /* 0 => preemptable,< 0 => BUG*/
arch/xtensa/include/asm/thread_info.h
90
.preempt_count = INIT_PREEMPT_COUNT, \
arch/xtensa/kernel/asm-offsets.c
91
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
include/asm-generic/preempt.h
11
return READ_ONCE(current_thread_info()->preempt_count);
include/asm-generic/preempt.h
16
return &current_thread_info()->preempt_count;
include/asm-generic/preempt.h
28
task_thread_info(p)->preempt_count = FORK_PREEMPT_COUNT; \
include/asm-generic/preempt.h
32
task_thread_info(p)->preempt_count = PREEMPT_DISABLED; \
include/asm-generic/preempt.h
77
return unlikely(preempt_count() == preempt_offset &&
include/linux/bit_spinlock.h
110
return preempt_count();
include/linux/ftrace.h
665
WARN_ON_ONCE(!preempt_count() || !irqs_disabled());
include/linux/ftrace.h
678
WARN_ON_ONCE(!preempt_count() || !irqs_disabled());
include/linux/lockdep.h
600
(preempt_count() != 0 || \
include/linux/lockdep.h
608
(preempt_count() == 0 && \
include/linux/preempt.h
108
#define nmi_count() (preempt_count() & NMI_MASK)
include/linux/preempt.h
109
#define hardirq_count() (preempt_count() & HARDIRQ_MASK)
include/linux/preempt.h
112
# define irq_count() ((preempt_count() & (NMI_MASK | HARDIRQ_MASK)) | softirq_count())
include/linux/preempt.h
114
# define softirq_count() (preempt_count() & SOFTIRQ_MASK)
include/linux/preempt.h
115
# define irq_count() (preempt_count() & (NMI_MASK | HARDIRQ_MASK | SOFTIRQ_MASK))
include/linux/preempt.h
130
# define in_task() (!((preempt_count() & (NMI_MASK | HARDIRQ_MASK)) | in_serving_softirq()))
include/linux/preempt.h
132
# define in_task() (!(preempt_count() & (NMI_MASK | HARDIRQ_MASK | SOFTIRQ_OFFSET)))
include/linux/preempt.h
184
#define in_atomic() (preempt_count() != 0)
include/linux/preempt.h
190
#define in_atomic_preempt_off() (preempt_count() != PREEMPT_DISABLE_OFFSET)
include/linux/preempt.h
225
#define preemptible() (preempt_count() == 0 && !irqs_disabled())
include/linux/preempt.h
92
unsigned long pc = preempt_count();
include/linux/trace_events.h
178
entry->preempt_count = trace_ctx & 0xff;
include/linux/trace_events.h
85
unsigned char preempt_count;
init/main.c
1374
int count = preempt_count();
init/main.c
1387
if (preempt_count() != count) {
kernel/bpf/syscall.c
601
return preempt_count() == 0 && !irqs_disabled() &&
kernel/entry/common.c
190
if (!preempt_count()) {
kernel/exit.c
1050
preempt_count());
kernel/rcu/rcutorture.c
2043
#define ROEC_ARGS "%s %s: Current %#x To add %#x To remove %#x preempt_count() %#x\n", __func__, s, curstate, new, old, preempt_count()
kernel/rcu/rcutorture.c
2061
!(preempt_count() & PREEMPT_MASK), ROEC_ARGS);
kernel/rcu/rcutorture.c
2084
(preempt_count() & PREEMPT_MASK), ROEC_ARGS);
kernel/rcu/rcutorture.c
475
if ((preempt_count() & HARDIRQ_MASK) || softirq_count())
kernel/rcu/rcutorture.c
487
if (!preempt_count() &&
kernel/rcu/rcutorture.c
502
return (preempt_count() & PREEMPT_MASK);
kernel/rcu/tree.c
2846
if (IS_ENABLED(CONFIG_PREEMPT_COUNT) && (!(preempt_count() & PREEMPT_MASK))) {
kernel/rcu/tree_exp.h
772
if (!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK)) ||
kernel/rcu/tree_exp.h
869
bool preempt_bh_enabled = !(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK));
kernel/rcu/tree_plugin.h
1070
(preempt_count() == HARDIRQ_OFFSET))) {
kernel/rcu/tree_plugin.h
730
!!(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK));
kernel/rcu/tree_plugin.h
816
(preempt_count() & (PREEMPT_MASK | SOFTIRQ_MASK))) {
kernel/sched/core.c
5130
if (WARN_ONCE(preempt_count() != 2*PREEMPT_DISABLE_OFFSET,
kernel/sched/core.c
5132
current->comm, current->pid, preempt_count()))
kernel/sched/core.c
5746
if (preempt_count() == val) {
kernel/sched/core.c
5761
if (DEBUG_LOCKS_WARN_ON((preempt_count() < 0)))
kernel/sched/core.c
5769
DEBUG_LOCKS_WARN_ON((preempt_count() & PREEMPT_MASK) >=
kernel/sched/core.c
5783
if (preempt_count() == val)
kernel/sched/core.c
5793
if (DEBUG_LOCKS_WARN_ON(val > preempt_count()))
kernel/sched/core.c
5799
!(preempt_count() & PREEMPT_MASK)))
kernel/sched/core.c
5835
prev->comm, prev->pid, preempt_count());
kernel/sched/core.c
7231
BUG_ON(preempt_count() || !irqs_disabled());
kernel/sched/core.c
8826
if (preempt_count() == preempt_offset)
kernel/sched/core.c
8835
unsigned int nested = preempt_count();
kernel/sched/core.c
8870
pr_err("preempt_count: %x, expected: %x\n", preempt_count(),
kernel/sched/core.c
8903
if (preempt_count() > preempt_offset)
kernel/sched/core.c
8935
if (preempt_count() > 0)
kernel/softirq.c
393
if (preempt_count() == cnt) {
kernel/softirq.c
407
if (preempt_count() == cnt)
kernel/softirq.c
617
prev_count = preempt_count();
kernel/softirq.c
624
if (unlikely(prev_count != preempt_count())) {
kernel/softirq.c
627
prev_count, preempt_count());
kernel/stop_machine.c
521
WARN_ONCE(preempt_count(),
kernel/time/timer.c
1726
int count = preempt_count();
kernel/time/timer.c
1753
if (count != preempt_count()) {
kernel/time/timer.c
1755
fn, count, preempt_count());
kernel/trace/bpf_trace.c
850
if (preempt_count() != 0 || irqs_disabled()) {
kernel/trace/ring_buffer.c
4312
unsigned long pc = preempt_count();
kernel/trace/trace.c
2186
pc = preempt_count();
kernel/trace/trace_events.c
214
__common_field(unsigned char, preempt_count);
kernel/trace/trace_irqsoff.c
122
if (!irqs_disabled_flags(*flags) && !preempt_count())
kernel/trace/trace_irqsoff.c
467
if (preempt_trace(preempt_count()) || irq_trace())
kernel/trace/trace_irqsoff.c
475
if (preempt_trace(preempt_count()) || irq_trace())
kernel/trace/trace_irqsoff.c
638
if (!preempt_trace(preempt_count()) && irq_trace())
kernel/trace/trace_irqsoff.c
645
if (!preempt_trace(preempt_count()) && irq_trace())
kernel/trace/trace_irqsoff.c
686
if (preempt_trace(preempt_count()) && !irq_trace())
kernel/trace/trace_irqsoff.c
692
if (preempt_trace(preempt_count()) && !irq_trace())
kernel/trace/trace_output.c
537
if (entry->preempt_count & 0xf)
kernel/trace/trace_output.c
538
trace_seq_printf(s, "%x", entry->preempt_count & 0xf);
kernel/trace/trace_output.c
542
if (entry->preempt_count & 0xf0)
kernel/trace/trace_output.c
543
trace_seq_printf(s, "%x", entry->preempt_count >> 4);
kernel/trace/trace_output.c
706
entry->preempt_count & 0xf, iter->idx);
kernel/trace/trace_selftest.c
1259
if (preempt_count()) {
kernel/trace/trace_selftest.c
1321
if (preempt_count()) {
kernel/workqueue.c
3292
current->comm, task_pid_nr(current), preempt_count(),
lib/locking-selftest.c
1432
int saved_preempt_count = preempt_count();
lib/smp_processor_id.c
16
if (likely(preempt_count()))
lib/smp_processor_id.c
44
what1, what2, preempt_count() - 1, current->comm, current->pid);
tools/perf/util/bpf_skel/vmlinux/vmlinux.h
113
unsigned char preempt_count;
tools/testing/radix-tree/main.c
244
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
248
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
252
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
256
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
260
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
264
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
269
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
273
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
281
nr_allocated, preempt_count);
tools/testing/radix-tree/main.c
324
nr_allocated, preempt_count);
tools/testing/selftests/bpf/bpf_experimental.h
620
int preempt_count;
tools/testing/selftests/bpf/bpf_experimental.h
643
if (bpf_core_field_exists(pcpu_hot.preempt_count))
tools/testing/selftests/bpf/bpf_experimental.h
645
bpf_this_cpu_ptr(&pcpu_hot))->preempt_count;
tools/testing/selftests/bpf/prog_tests/timer.c
346
if (skel->bss->preempt_count)
tools/testing/selftests/bpf/progs/timer_interrupt.c
11
int preempt_count;
tools/testing/selftests/bpf/progs/timer_interrupt.c
28
preempt_count = get_preempt_count();
tools/testing/shared/linux.c
16
int preempt_count;
tools/testing/shared/linux/preempt.h
5
extern int preempt_count;
tools/testing/shared/linux/preempt.h
7
#define preempt_disable() uatomic_inc(&preempt_count)
tools/testing/shared/linux/preempt.h
8
#define preempt_enable() uatomic_dec(&preempt_count)