Symbol: stack
arch/alpha/kernel/process.c
235
unsigned long usp = args->stack;
arch/alpha/kernel/process.c
243
struct switch_stack *childstack, *stack;
arch/alpha/kernel/process.c
276
stack = ((struct switch_stack *) regs) - 1;
arch/alpha/kernel/process.c
277
*childstack = *stack;
arch/alpha/kernel/traps.c
115
unsigned long *stack;
arch/alpha/kernel/traps.c
125
stack = sp;
arch/alpha/kernel/traps.c
127
if (((long) stack & (THREAD_SIZE-1)) == 0)
arch/alpha/kernel/traps.c
136
pr_cont("%016lx", *stack++);
arch/arc/include/asm/entry-arcv2.h
158
; Auto stack switch works
arch/arc/include/asm/entry-arcv2.h
196
; - for K mode, it will be implicitly restored as stack is unwound
arch/arc/include/asm/entry-arcv2.h
52
; 1. SP auto-switched to kernel mode stack
arch/arc/include/asm/entry-arcv2.h
54
; 3. Auto save: (mandatory) Push PC and STAT32 on stack
arch/arc/include/asm/entry-arcv2.h
65
; carve pt_regs on stack (case #3), PC/STAT32 already on stack
arch/arc/include/asm/entry-arcv2.h
70
; carve pt_regs on stack (case #4), which grew partially already
arch/arc/include/asm/entry-arcv2.h
81
; 1. SP auto-switched to kernel mode stack
arch/arc/kernel/asm-offsets.c
20
DEFINE(TASK_THREAD_INFO, offsetof(struct task_struct, stack));
arch/arc/kernel/process.c
170
unsigned long usp = args->stack;
arch/arc/kernel/unwind.c
151
const u8 *stack[MAX_STACK_DEPTH];
arch/arc/kernel/unwind.c
793
state->stack[state->stackDepth++] = ptr.p8;
arch/arc/kernel/unwind.c
802
state->stack[state->stackDepth - 1];
arch/arm/common/bL_switcher.c
126
void *stack = current_thread_info() + 1;
arch/arm/common/bL_switcher.c
127
stack = PTR_ALIGN(stack, L1_CACHE_BYTES);
arch/arm/common/bL_switcher.c
128
stack += clusterid * STACK_SIZE + STACK_SIZE;
arch/arm/common/bL_switcher.c
129
call_with_stack(bL_do_switch, (void *)_arg, stack);
arch/arm/include/asm/smp.h
58
void *stack;
arch/arm/include/asm/stacktrace.h
32
unsigned long delta = current_stack_pointer ^ (unsigned long)current->stack;
arch/arm/include/asm/traps.h
39
asmlinkage void dump_backtrace_stm(u32 *stack, u32 instruction, const char *loglvl);
arch/arm/kernel/irq.c
56
u8 *stack;
arch/arm/kernel/irq.c
61
stack = (u8 *)__get_free_pages(GFP_KERNEL,
arch/arm/kernel/irq.c
64
stack = __vmalloc_node(THREAD_SIZE, THREAD_ALIGN,
arch/arm/kernel/irq.c
68
if (WARN_ON(!stack))
arch/arm/kernel/irq.c
70
per_cpu(irq_stack_ptr, cpu) = &stack[THREAD_SIZE];
arch/arm/kernel/process.c
238
unsigned long stack_start = args->stack;
arch/arm/kernel/reboot.c
62
u64 *stack = soft_restart_stack + ARRAY_SIZE(soft_restart_stack);
arch/arm/kernel/reboot.c
73
call_with_stack(__soft_restart, (void *)addr, (void *)stack);
arch/arm/kernel/setup.c
144
static struct stack stacks[NR_CPUS];
arch/arm/kernel/setup.c
535
struct stack *stk = &stacks[cpu];
arch/arm/kernel/setup.c
582
"I" (offsetof(struct stack, irq[0])),
arch/arm/kernel/setup.c
584
"I" (offsetof(struct stack, abt[0])),
arch/arm/kernel/setup.c
586
"I" (offsetof(struct stack, und[0])),
arch/arm/kernel/setup.c
588
"I" (offsetof(struct stack, fiq[0])),
arch/arm/kernel/smp.c
146
secondary_data.stack = task_stack_page(idle) + THREAD_START_SP;
arch/arm/kernel/traps.c
106
p += sprintf(p, " r%d:%08x", reg, *stack--);
arch/arm/kernel/traps.c
908
u8 *stack;
arch/arm/kernel/traps.c
912
stack = (u8 *)__get_free_page(GFP_KERNEL);
arch/arm/kernel/traps.c
913
if (WARN_ON(!stack))
arch/arm/kernel/traps.c
915
per_cpu(overflow_stack_ptr, cpu) = &stack[OVERFLOW_STACK_SIZE];
arch/arm/kernel/traps.c
923
unsigned long tsk_stk = (unsigned long)current->stack;
arch/arm/kernel/traps.c
98
void dump_backtrace_stm(u32 *stack, u32 instruction, const char *loglvl)
arch/arm/probes/kprobes/test-core.c
1315
static uintptr_t __used kprobes_test_case_start(const char **title, void *stack)
arch/arm/probes/kprobes/test-core.c
1324
current_stack = stack;
arch/arm64/include/asm/stacktrace/common.h
129
state->stack = tmp;
arch/arm64/include/asm/stacktrace/common.h
135
state->stack.low = sp + size;
arch/arm64/include/asm/stacktrace/common.h
33
struct stack_info stack;
arch/arm64/include/asm/stacktrace/common.h
60
state->stack = stackinfo_get_unknown();
arch/arm64/include/asm/stacktrace/common.h
77
struct stack_info *info = &state->stack;
arch/arm64/kernel/asm-offsets.c
39
DEFINE(TSK_STACK, offsetof(struct task_struct, stack));
arch/arm64/kernel/probes/uprobes.c
150
return regs->sp <= ret->stack;
arch/arm64/kernel/probes/uprobes.c
152
return regs->sp < ret->stack;
arch/arm64/kernel/process.c
414
unsigned long stack_start = args->stack;
arch/arm64/kernel/traps.c
905
unsigned long tsk_stk = (unsigned long)current->stack;
arch/csky/kernel/asm-offsets.c
13
DEFINE(TASK_THREAD_INFO, offsetof(struct task_struct, stack));
arch/csky/kernel/probes/uprobes.c
120
return regs->usp <= ret->stack;
arch/csky/kernel/probes/uprobes.c
122
return regs->usp < ret->stack;
arch/csky/kernel/process.c
36
unsigned long usp = args->stack;
arch/hexagon/kernel/asm-offsets.c
85
OFFSET(_TASK_THREAD_INFO, task_struct, stack);
arch/hexagon/kernel/process.c
56
unsigned long usp = args->stack;
arch/hexagon/kernel/smp.c
176
struct thread_info *thread = (struct thread_info *)idle->stack;
arch/loongarch/include/asm/smp.h
84
unsigned long stack;
arch/loongarch/include/asm/stacktrace.h
30
bool in_irq_stack(unsigned long stack, struct stack_info *info);
arch/loongarch/include/asm/stacktrace.h
31
bool in_task_stack(unsigned long stack, struct task_struct *task, struct stack_info *info);
arch/loongarch/include/asm/stacktrace.h
32
int get_stack_info(unsigned long stack, struct task_struct *task, struct stack_info *info);
arch/loongarch/include/asm/stacktrace.h
36
return !(((unsigned long)(current->stack) ^ current_stack_pointer) & ~(THREAD_SIZE - 1));
arch/loongarch/kernel/asm-offsets.c
269
OFFSET(CPU_BOOT_STACK, secondary_data, stack);
arch/loongarch/kernel/asm-offsets.c
73
OFFSET(TASK_THREAD_INFO, task_struct, stack);
arch/loongarch/kernel/process.c
174
unsigned long usp = args->stack;
arch/loongarch/kernel/process.c
258
bool in_irq_stack(unsigned long stack, struct stack_info *info)
arch/loongarch/kernel/process.c
264
if (stack < begin || stack >= end)
arch/loongarch/kernel/process.c
279
bool in_task_stack(unsigned long stack, struct task_struct *task,
arch/loongarch/kernel/process.c
285
if (stack < begin || stack >= end)
arch/loongarch/kernel/process.c
296
int get_stack_info(unsigned long stack, struct task_struct *task,
arch/loongarch/kernel/process.c
301
if (!stack || stack & (SZREG - 1))
arch/loongarch/kernel/process.c
304
if (in_task_stack(stack, task, info))
arch/loongarch/kernel/process.c
310
if (in_irq_stack(stack, info))
arch/loongarch/kernel/smp.c
403
cpuboot_data.stack = (unsigned long)__KSTK_TOS(idle);
arch/loongarch/kernel/traps.c
1102
register unsigned long stack;
arch/loongarch/kernel/traps.c
1110
stack = per_cpu(irq_stack, cpu) + IRQ_STACK_START;
arch/loongarch/kernel/traps.c
1113
*(unsigned long *)stack = sp;
arch/loongarch/kernel/traps.c
1122
: [stk] "r" (stack), [regs] "r" (regs)
arch/loongarch/kernel/uprobes.c
103
return regs->regs[3] <= ret->stack;
arch/loongarch/kernel/uprobes.c
105
return regs->regs[3] < ret->stack;
arch/loongarch/net/bpf_jit.c
21
#define BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack) (round_up(stack, 16) - 80)
arch/m68k/fpsp040/fpsp.h
11
| fpsp.h --- stack frame offsets during FPSP exception handling
arch/m68k/fpsp040/fpsp.h
24
| After initialization, the stack looks like this:
arch/m68k/fpsp040/fpsp.h
66
| a second fsave frame can be pushed onto the stack and the
arch/m68k/fpsp040/fpsp.h
77
| Local Variables on the stack
arch/m68k/kernel/asm-offsets.c
29
DEFINE(TASK_STACK, offsetof(struct task_struct, stack));
arch/m68k/kernel/process.c
125
.stack = regs->d2,
arch/m68k/kernel/process.c
145
unsigned long usp = args->stack;
arch/m68k/kernel/traps.c
810
static void show_trace(unsigned long *stack, const char *loglvl)
arch/m68k/kernel/traps.c
817
addr = (unsigned long)stack + THREAD_SIZE - 1;
arch/m68k/kernel/traps.c
820
while (stack + 1 <= endstack) {
arch/m68k/kernel/traps.c
821
addr = *stack++;
arch/m68k/kernel/traps.c
931
void show_stack(struct task_struct *task, unsigned long *stack,
arch/m68k/kernel/traps.c
938
if (!stack) {
arch/m68k/kernel/traps.c
940
stack = (unsigned long *)task->thread.esp0;
arch/m68k/kernel/traps.c
942
stack = (unsigned long *)&stack;
arch/m68k/kernel/traps.c
944
endstack = (unsigned long *)(((unsigned long)stack + THREAD_SIZE - 1) & -THREAD_SIZE);
arch/m68k/kernel/traps.c
946
printk("%sStack from %08lx:", loglvl, (unsigned long)stack);
arch/m68k/kernel/traps.c
947
p = stack;
arch/m68k/kernel/traps.c
956
show_trace(stack, loglvl);
arch/m68k/math-emu/fp_decode.h
227
| adjust stack for byte moves from/to stack
arch/microblaze/kernel/asm-offsets.c
73
DEFINE(TS_THREAD_INFO, offsetof(struct task_struct, stack));
arch/microblaze/kernel/process.c
58
unsigned long usp = args->stack;
arch/microblaze/kernel/traps.c
43
(task->stack))->cpu_context.r1;
arch/microblaze/kernel/unwind.c
283
(struct thread_info *)(task->stack);
arch/mips/include/asm/sgiarcs.h
373
extern long call_o32(long vec, void *stack, ...);
arch/mips/kernel/asm-offsets.c
91
OFFSET(TASK_THREAD_INFO, task_struct, stack);
arch/mips/kernel/process.c
111
unsigned long usp = args->stack;
arch/mips/net/bpf_jit_comp32.c
1377
int stack, saved, locals, reserved;
arch/mips/net/bpf_jit_comp32.c
1429
stack = ALIGN(saved + locals + reserved, MIPS_STACK_ALIGNMENT);
arch/mips/net/bpf_jit_comp32.c
1430
emit(ctx, addiu, MIPS_R_SP, MIPS_R_SP, -stack);
arch/mips/net/bpf_jit_comp32.c
1433
push_regs(ctx, ctx->clobbered & JIT_CALLEE_REGS, 0, stack - saved);
arch/mips/net/bpf_jit_comp32.c
1437
emit(ctx, addiu, lo(fp), MIPS_R_SP, stack - saved);
arch/mips/net/bpf_jit_comp32.c
1440
ctx->stack_size = stack;
arch/mips/net/bpf_jit_comp64.c
551
int stack, saved, locals, reserved;
arch/mips/net/bpf_jit_comp64.c
597
stack = ALIGN(saved + locals + reserved, MIPS_STACK_ALIGNMENT);
arch/mips/net/bpf_jit_comp64.c
598
if (stack)
arch/mips/net/bpf_jit_comp64.c
599
emit(ctx, daddiu, MIPS_R_SP, MIPS_R_SP, -stack);
arch/mips/net/bpf_jit_comp64.c
602
push_regs(ctx, ctx->clobbered & JIT_CALLEE_REGS, 0, stack - saved);
arch/mips/net/bpf_jit_comp64.c
606
emit(ctx, daddiu, fp, MIPS_R_SP, stack - saved);
arch/mips/net/bpf_jit_comp64.c
615
ctx->stack_size = stack;
arch/nios2/kernel/process.c
105
unsigned long usp = args->stack;
arch/nios2/kernel/process.c
109
struct switch_stack *stack;
arch/nios2/kernel/process.c
133
stack = ((struct switch_stack *) regs) - 1;
arch/nios2/kernel/process.c
134
*childstack = *stack;
arch/nios2/kernel/process.c
263
.stack = newsp,
arch/nios2/kernel/traps.c
60
void show_stack(struct task_struct *task, unsigned long *stack,
arch/nios2/kernel/traps.c
66
if (!stack) {
arch/nios2/kernel/traps.c
68
stack = (unsigned long *)task->thread.ksp;
arch/nios2/kernel/traps.c
70
stack = (unsigned long *)&stack;
arch/nios2/kernel/traps.c
73
addr = (unsigned long) stack;
arch/nios2/kernel/traps.c
76
printk("%sStack from %08lx:", loglvl, (unsigned long)stack);
arch/nios2/kernel/traps.c
78
if (stack + 1 > endstack)
arch/nios2/kernel/traps.c
82
printk("%s %08lx", loglvl, *stack++);
arch/nios2/kernel/traps.c
87
while (stack + 1 <= endstack) {
arch/nios2/kernel/traps.c
88
addr = *stack++;
arch/openrisc/include/asm/unwinder.h
16
void unwind_stack(void *data, unsigned long *stack,
arch/openrisc/kernel/process.c
169
unsigned long usp = args->stack;
arch/openrisc/kernel/process.c
254
new_ti = new->stack;
arch/openrisc/kernel/process.c
255
old_ti = old->stack;
arch/openrisc/kernel/unwinder.c
60
void unwind_stack(void *data, unsigned long *stack,
arch/openrisc/kernel/unwinder.c
67
while (!kstack_end(stack)) {
arch/openrisc/kernel/unwinder.c
68
frameinfo = container_of(stack,
arch/openrisc/kernel/unwinder.c
83
stack++;
arch/openrisc/kernel/unwinder.c
93
void unwind_stack(void *data, unsigned long *stack,
arch/openrisc/kernel/unwinder.c
98
while (!kstack_end(stack)) {
arch/openrisc/kernel/unwinder.c
99
addr = *stack++;
arch/parisc/include/asm/processor.h
150
void show_trace(struct task_struct *task, unsigned long *stack);
arch/parisc/kernel/asm-offsets.c
47
DEFINE(TASK_STACK, offsetof(struct task_struct, stack));
arch/parisc/kernel/irq.c
363
unsigned long stack[IRQ_STACK_SIZE/sizeof(unsigned long)];
arch/parisc/kernel/irq.c
403
stack_start = (unsigned long) &per_cpu(irq_stack_union, cpu).stack;
arch/parisc/kernel/irq.c
455
irq_stack = (unsigned long) &union_ptr->stack;
arch/parisc/kernel/process.c
208
unsigned long usp = args->stack;
arch/parisc/kernel/process.c
211
void *stack = task_stack_page(p);
arch/parisc/kernel/process.c
227
cregs->ksp = (unsigned long) stack + FRAME_SIZE + PT_SZ_ALGN;
arch/parisc/kernel/process.c
250
cregs->ksp = (unsigned long) stack + FRAME_SIZE;
arch/powerpc/include/asm/thread_info.h
201
static inline int arch_within_stack_frames(const void * const stack,
arch/powerpc/include/asm/thread_info.h
220
while (stack <= frame && frame < stackend) {
arch/powerpc/include/asm/xics.h
106
unsigned char stack[MAX_NUM_PRIORITIES];
arch/powerpc/include/asm/xics.h
120
os_cppr->stack[++os_cppr->index] = IPI_PRIORITY;
arch/powerpc/include/asm/xics.h
122
os_cppr->stack[++os_cppr->index] = DEFAULT_PRIORITY;
arch/powerpc/include/asm/xics.h
132
return os_cppr->stack[--os_cppr->index];
arch/powerpc/include/asm/xics.h
144
os_cppr->stack[0] = cppr;
arch/powerpc/include/asm/xics.h
151
return os_cppr->stack[os_cppr->index];
arch/powerpc/kernel/asm-offsets.c
93
OFFSET(TASK_STACK, task_struct, stack);
arch/powerpc/kernel/process.c
1784
unsigned long usp = args->stack;
arch/powerpc/kernel/process.c
2281
unsigned long *stack,
arch/powerpc/kernel/process.c
2296
sp = (unsigned long) stack;
arch/powerpc/kernel/process.c
2310
stack = (unsigned long *) sp;
arch/powerpc/kernel/process.c
2311
newsp = stack[0];
arch/powerpc/kernel/process.c
2312
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/process.c
2317
&ftrace_idx, ip, stack);
arch/powerpc/kernel/process.c
2335
&& stack[STACK_INT_FRAME_MARKER_LONGS] == STACK_FRAME_REGS_MARKER) {
arch/powerpc/kernel/stacktrace.c
103
newsp = stack[0];
arch/powerpc/kernel/stacktrace.c
123
stack[STACK_INT_FRAME_MARKER_LONGS] == STACK_FRAME_REGS_MARKER) {
arch/powerpc/kernel/stacktrace.c
128
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/stacktrace.c
136
ip = ftrace_graph_ret_addr(task, &graph_idx, ip, stack);
arch/powerpc/kernel/stacktrace.c
44
unsigned long *stack = (unsigned long *) sp;
arch/powerpc/kernel/stacktrace.c
50
newsp = stack[0];
arch/powerpc/kernel/stacktrace.c
51
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/stacktrace.c
96
unsigned long *stack = (unsigned long *) sp;
arch/powerpc/kernel/uprobes.c
214
return regs->gpr[1] <= ret->stack;
arch/powerpc/kernel/uprobes.c
216
return regs->gpr[1] < ret->stack;
arch/powerpc/net/bpf_jit_comp64.c
382
unsigned long *stack = (unsigned long *) sp;
arch/powerpc/net/bpf_jit_comp64.c
388
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/net/bpf_jit_comp64.c
400
sp = stack[0];
arch/riscv/include/asm/stacktrace.h
21
return !(((unsigned long)(current->stack) ^ current_stack_pointer) & ~(THREAD_SIZE - 1));
arch/riscv/kernel/probes/uprobes.c
124
return regs->sp <= ret->stack;
arch/riscv/kernel/probes/uprobes.c
126
return regs->sp < ret->stack;
arch/riscv/kernel/process.c
241
unsigned long usp = args->stack;
arch/riscv/kernel/traps.c
483
unsigned long tsk_stk = (unsigned long)current->stack;
arch/s390/include/asm/stacktrace.h
188
#define call_on_stack(nr, stack, rettype, fn, ...) \
arch/s390/include/asm/stacktrace.h
192
unsigned long __stack = stack; \
arch/s390/kernel/asm-offsets.c
21
OFFSET(__TASK_stack, task_struct, stack);
arch/s390/kernel/dumpstack.c
130
void show_stack(struct task_struct *task, unsigned long *stack,
arch/s390/kernel/dumpstack.c
136
unwind_for_each_frame(&state, task, NULL, (unsigned long) stack)
arch/s390/kernel/dumpstack.c
45
enum stack_type type, unsigned long stack)
arch/s390/kernel/dumpstack.c
47
if (sp < stack || sp >= stack + THREAD_SIZE)
arch/s390/kernel/dumpstack.c
50
info->begin = stack;
arch/s390/kernel/dumpstack.c
51
info->end = stack + THREAD_SIZE;
arch/s390/kernel/dumpstack.c
58
unsigned long stack = (unsigned long)task_stack_page(task);
arch/s390/kernel/dumpstack.c
60
return in_stack(sp, info, STACK_TYPE_TASK, stack);
arch/s390/kernel/dumpstack.c
65
unsigned long stack = get_lowcore()->async_stack - STACK_INIT_OFFSET;
arch/s390/kernel/dumpstack.c
67
return in_stack(sp, info, STACK_TYPE_IRQ, stack);
arch/s390/kernel/dumpstack.c
72
unsigned long stack = get_lowcore()->nodat_stack - STACK_INIT_OFFSET;
arch/s390/kernel/dumpstack.c
74
return in_stack(sp, info, STACK_TYPE_NODAT, stack);
arch/s390/kernel/dumpstack.c
79
unsigned long stack = get_lowcore()->mcck_stack - STACK_INIT_OFFSET;
arch/s390/kernel/dumpstack.c
81
return in_stack(sp, info, STACK_TYPE_MCCK, stack);
arch/s390/kernel/dumpstack.c
86
unsigned long stack = get_lowcore()->restart_stack - STACK_INIT_OFFSET;
arch/s390/kernel/dumpstack.c
88
return in_stack(sp, info, STACK_TYPE_RESTART, stack);
arch/s390/kernel/entry.h
60
void stack_free(unsigned long stack);
arch/s390/kernel/perf_event.c
28
struct stack_frame *stack = (struct stack_frame *) regs->gprs[15];
arch/s390/kernel/perf_event.c
30
if (!stack)
arch/s390/kernel/perf_event.c
33
return (struct kvm_s390_sie_block *)stack->sie_control_block;
arch/s390/kernel/process.c
109
unsigned long new_stackp = args->stack;
arch/s390/kernel/setup.c
365
void *stack;
arch/s390/kernel/setup.c
367
stack = __vmalloc_node(THREAD_SIZE, THREAD_SIZE, THREADINFO_GFP,
arch/s390/kernel/setup.c
369
kmemleak_not_leak(stack);
arch/s390/kernel/setup.c
370
return (unsigned long)stack;
arch/s390/kernel/setup.c
373
void stack_free(unsigned long stack)
arch/s390/kernel/setup.c
375
vfree((void *)stack);
arch/s390/kernel/setup.c
380
unsigned long stack;
arch/s390/kernel/setup.c
382
stack = (unsigned long)memblock_alloc_or_panic(THREAD_SIZE, THREAD_SIZE);
arch/s390/kernel/setup.c
383
return stack;
arch/s390/kernel/smp.c
312
void *data, unsigned long stack)
arch/s390/kernel/smp.c
321
call_on_stack(2, stack, void, __pcpu_delegate,
arch/s390/kernel/smp.c
329
lc->restart_stack = stack;
arch/s390/kernel/smp.c
335
abs_lc->restart_stack = stack;
arch/s390/kernel/uprobes.c
156
return user_stack_pointer(regs) <= ret->stack;
arch/s390/kernel/uprobes.c
158
return user_stack_pointer(regs) < ret->stack;
arch/sh/include/asm/stacktrace.h
18
unsigned long *stack,
arch/sh/kernel/dumpstack.c
143
unsigned long stack;
arch/sh/kernel/dumpstack.c
152
stack = (unsigned long)sp;
arch/sh/kernel/dumpstack.c
153
dump_mem("Stack: ", loglvl, stack, THREAD_SIZE +
arch/sh/kernel/irq.c
63
u32 stack[THREAD_SIZE/sizeof(u32)];
arch/sh/kernel/process_32.c
93
unsigned long usp = args->stack;
arch/sh/kernel/ptrace_32.c
40
unsigned char *stack;
arch/sh/kernel/ptrace_32.c
42
stack = (unsigned char *)task_pt_regs(task);
arch/sh/kernel/ptrace_32.c
43
stack += offset;
arch/sh/kernel/ptrace_32.c
44
return (*((int *)stack));
arch/sh/kernel/ptrace_32.c
53
unsigned char *stack;
arch/sh/kernel/ptrace_32.c
55
stack = (unsigned char *)task_pt_regs(task);
arch/sh/kernel/ptrace_32.c
56
stack += offset;
arch/sh/kernel/ptrace_32.c
57
*(unsigned long *) stack = data;
arch/sh/kernel/smp.c
222
stack_start.thread_info = tsk->stack;
arch/sparc/kernel/process.c
106
args.stack = regs->u_regs[UREG_I1];
arch/sparc/kernel/process.c
108
args.stack = regs->u_regs[UREG_FP];
arch/sparc/kernel/process.c
31
args.stack = regs->u_regs[UREG_FP];
arch/sparc/kernel/process.c
59
args.stack = regs->u_regs[UREG_FP];
arch/sparc/kernel/process_32.c
270
unsigned long sp = args->stack ? args->stack : regs->u_regs[UREG_FP];
arch/sparc/kernel/process_64.c
579
unsigned long sp = args->stack ? args->stack : regs->u_regs[UREG_FP];
arch/sparc/kernel/signal32.c
244
err |= compat_restore_altstack(&sf->stack);
arch/sparc/kernel/signal32.c
566
err |= __compat_save_altstack(&sf->stack, regs->u_regs[UREG_FP]);
arch/sparc/kernel/signal32.c
64
compat_stack_t stack;
arch/sparc/kernel/signal_32.c
180
err |= restore_altstack(&sf->stack);
arch/sparc/kernel/signal_32.c
373
err |= __save_altstack(&sf->stack, regs->u_regs[UREG_FP]);
arch/sparc/kernel/signal_32.c
53
stack_t stack;
arch/sparc/kernel/signal_64.c
250
stack_t stack;
arch/sparc/kernel/signal_64.c
303
err |= restore_altstack(&sf->stack);
arch/sparc/kernel/signal_64.c
404
err |= __save_altstack(&sf->stack, regs->u_regs[UREG_FP]);
arch/um/drivers/chan_user.c
298
unsigned long stack;
arch/um/drivers/chan_user.c
312
thread = winch_tramp(fd, port, &thread_fd, &stack);
arch/um/drivers/chan_user.c
316
register_winch_irq(thread_fd, fd, thread, port, stack);
arch/um/drivers/chan_user.h
45
struct tty_port *port, unsigned long stack);
arch/um/drivers/line.c
597
unsigned long stack;
arch/um/drivers/line.c
608
if (winch->stack != 0)
arch/um/drivers/line.c
609
free_stack(winch->stack, 0);
arch/um/drivers/line.c
671
unsigned long stack)
arch/um/drivers/line.c
685
.stack = stack });
arch/um/drivers/line.c
708
if (stack != 0)
arch/um/drivers/line.c
709
free_stack(stack, 0);
arch/um/include/asm/processor-generic.h
70
unsigned long stack);
arch/um/include/shared/kern_util.h
19
extern void free_stack(unsigned long stack, int order);
arch/um/include/shared/os.h
296
extern void new_thread(void *stack, jmp_buf *buf, void (*handler)(void));
arch/um/include/shared/os.h
298
extern int start_idle_thread(void *stack, jmp_buf *switch_buf);
arch/um/include/shared/skas/mm_id.h
15
unsigned long stack;
arch/um/kernel/process.c
149
unsigned long sp = args->stack;
arch/um/kernel/process.c
51
void free_stack(unsigned long stack, int order)
arch/um/kernel/process.c
53
free_pages(stack, order);
arch/um/kernel/skas/mmu.c
109
free_pages(mmu->id.stack, ilog2(STUB_DATA_PAGES));
arch/um/kernel/skas/mmu.c
134
stub_data = (void *)mm_context->id.stack;
arch/um/kernel/skas/mmu.c
46
unsigned long stack = 0;
arch/um/kernel/skas/mmu.c
52
stack = __get_free_pages(GFP_KERNEL | __GFP_ZERO, ilog2(STUB_DATA_PAGES));
arch/um/kernel/skas/mmu.c
53
if (stack == 0)
arch/um/kernel/skas/mmu.c
56
new_id->stack = stack;
arch/um/kernel/skas/mmu.c
75
free_pages(new_id->stack, ilog2(STUB_DATA_PAGES));
arch/um/kernel/skas/process.c
50
return current->mm->context.id.stack;
arch/um/kernel/skas/stub_exe.c
21
} stack = {
arch/um/kernel/skas/stub_exe.c
79
stack.ss_sp = (void *)init_data.stub_start + UM_KERN_PAGE_SIZE;
arch/um/kernel/skas/stub_exe.c
80
stub_syscall2(__NR_sigaltstack, (unsigned long)&stack, 0);
arch/um/kernel/sysrq.c
29
void show_stack(struct task_struct *task, unsigned long *stack,
arch/um/kernel/sysrq.c
35
if (!stack)
arch/um/kernel/sysrq.c
36
stack = get_stack_pointer(task, segv_regs);
arch/um/kernel/sysrq.c
40
if (kstack_end(stack))
arch/um/kernel/sysrq.c
44
pr_cont(" %08lx", READ_ONCE_NOCHECK(*stack));
arch/um/kernel/sysrq.c
45
stack++;
arch/um/kernel/um_arch.c
311
unsigned long stack;
arch/um/kernel/um_arch.c
384
stack = (unsigned long) argv;
arch/um/kernel/um_arch.c
385
stack &= ~(1024 * 1024 - 1);
arch/um/kernel/um_arch.c
386
avail = stack - start_vm;
arch/um/kernel/um_arch.c
411
stack_protections((unsigned long) init_task.stack);
arch/um/os-Linux/helper.c
115
free_stack(stack, 0);
arch/um/os-Linux/helper.c
122
unsigned long stack, sp;
arch/um/os-Linux/helper.c
129
stack = alloc_stack(0, __uml_cant_sleep());
arch/um/os-Linux/helper.c
130
if (stack == 0)
arch/um/os-Linux/helper.c
133
sp = stack + UM_KERN_PAGE_SIZE;
arch/um/os-Linux/helper.c
152
free_stack(stack, 0);
arch/um/os-Linux/helper.c
154
*stack_out = stack;
arch/um/os-Linux/helper.c
47
unsigned long stack, sp;
arch/um/os-Linux/helper.c
50
stack = alloc_stack(0, __uml_cant_sleep());
arch/um/os-Linux/helper.c
51
if (stack == 0)
arch/um/os-Linux/helper.c
69
sp = stack + UM_KERN_PAGE_SIZE;
arch/um/os-Linux/signal.c
168
stack_t stack = {
arch/um/os-Linux/signal.c
174
if (sigaltstack(&stack, NULL) != 0)
arch/um/os-Linux/skas/mem.c
160
struct stub_data *proc_data = (struct stub_data *) mm_idp->stack;
arch/um/os-Linux/skas/mem.c
183
struct stub_data *proc_data = (void *) mm_idp->stack;
arch/um/os-Linux/skas/mem.c
27
struct stub_data *proc_data = (void *)mm_idp->stack;
arch/um/os-Linux/skas/mem.c
59
unsigned long *stack)
arch/um/os-Linux/skas/mem.c
61
if (stack == NULL) {
arch/um/os-Linux/skas/mem.c
62
stack = (unsigned long *) mm_idp->stack + 2;
arch/um/os-Linux/skas/mem.c
63
*stack = 0;
arch/um/os-Linux/skas/mem.c
65
return stack;
arch/um/os-Linux/skas/mem.c
89
struct stub_data *proc_data = (void *)mm_idp->stack;
arch/um/os-Linux/skas/process.c
152
struct stub_data *data = (void *)mm_idp->stack;
arch/um/os-Linux/skas/process.c
444
struct stub_data *proc_data = (void *)mm_id->stack;
arch/um/os-Linux/skas/process.c
448
void *stack;
arch/um/os-Linux/skas/process.c
453
stack = mmap(NULL, UM_KERN_PAGE_SIZE,
arch/um/os-Linux/skas/process.c
456
if (stack == MAP_FAILED) {
arch/um/os-Linux/skas/process.c
464
sp = (unsigned long)stack + UM_KERN_PAGE_SIZE;
arch/um/os-Linux/skas/process.c
517
if (munmap(stack, UM_KERN_PAGE_SIZE) < 0) {
arch/um/os-Linux/skas/process.c
594
struct stub_data *proc_data = (void *) mm_id->stack;
arch/um/os-Linux/skas/process.c
793
void new_thread(void *stack, jmp_buf *buf, void (*handler)(void))
arch/um/os-Linux/skas/process.c
796
(*buf)[0].JB_SP = (unsigned long) stack + UM_THREAD_SIZE -
arch/um/os-Linux/skas/process.c
819
int start_idle_thread(void *stack, jmp_buf *switch_buf)
arch/um/os-Linux/skas/process.c
837
(*switch_buf)[0].JB_SP = (unsigned long) stack +
arch/x86/include/asm/cpu_entry_area.h
144
return &get_cpu_entry_area(cpu)->entry_stack_page.stack;
arch/x86/include/asm/cpu_entry_area.h
77
unsigned long stack[(PAGE_SIZE - sizeof(struct x86_hw_tss)) / sizeof(unsigned long)];
arch/x86/include/asm/irq_stack.h
81
#define call_on_stack(stack, func, asm_call, argconstr...) \
arch/x86/include/asm/irq_stack.h
85
tos = ((void *)(stack)); \
arch/x86/include/asm/processor.h
362
char stack[PAGE_SIZE];
arch/x86/include/asm/processor.h
366
struct entry_stack stack;
arch/x86/include/asm/processor.h
415
char stack[IRQ_STACK_SIZE];
arch/x86/include/asm/stacktrace.h
31
bool in_task_stack(unsigned long *stack, struct task_struct *task,
arch/x86/include/asm/stacktrace.h
34
bool in_entry_stack(unsigned long *stack, struct stack_info *info);
arch/x86/include/asm/stacktrace.h
36
int get_stack_info(unsigned long *stack, struct task_struct *task,
arch/x86/include/asm/stacktrace.h
38
bool get_stack_info_noinstr(unsigned long *stack, struct task_struct *task,
arch/x86/include/asm/stacktrace.h
42
bool get_stack_guard_info(unsigned long *stack, struct stack_info *info)
arch/x86/include/asm/stacktrace.h
45
if (get_stack_info_noinstr(stack, current, info))
arch/x86/include/asm/stacktrace.h
48
return get_stack_info_noinstr((void *)stack + PAGE_SIZE, current, info);
arch/x86/include/asm/thread_info.h
164
static inline int arch_within_stack_frames(const void * const stack,
arch/x86/include/asm/thread_info.h
181
while (stack <= frame && frame < stackend) {
arch/x86/kernel/asm-offsets_32.c
45
offsetofend(struct cpu_entry_area, entry_stack_page.stack));
arch/x86/kernel/doublefault_32.c
125
(unsigned long)&cea->doublefault_stack.stack +
arch/x86/kernel/doublefault_32.c
126
sizeof(doublefault_stack.stack));
arch/x86/kernel/dumpstack.c
185
unsigned long *stack, const char *log_lvl)
arch/x86/kernel/dumpstack.c
195
unwind_start(&state, task, regs, stack);
arch/x86/kernel/dumpstack.c
196
stack = stack ?: get_stack_pointer(task, regs);
arch/x86/kernel/dumpstack.c
215
for (; stack; stack = stack_info.next_sp) {
arch/x86/kernel/dumpstack.c
218
stack = PTR_ALIGN(stack, sizeof(long));
arch/x86/kernel/dumpstack.c
220
if (get_stack_info(stack, task, &stack_info, &visit_mask)) {
arch/x86/kernel/dumpstack.c
227
stack = (unsigned long *)PAGE_ALIGN((unsigned long)stack);
arch/x86/kernel/dumpstack.c
228
if (get_stack_info(stack, task, &stack_info, &visit_mask))
arch/x86/kernel/dumpstack.c
249
for (; stack < stack_info.end; stack++) {
arch/x86/kernel/dumpstack.c
252
unsigned long addr = READ_ONCE_NOCHECK(*stack);
arch/x86/kernel/dumpstack.c
263
if (regs && stack == &regs->ip)
arch/x86/kernel/dumpstack.c
266
if (stack == ret_addr_p)
arch/x86/kernel/dumpstack.c
279
addr, stack);
arch/x86/kernel/dumpstack.c
30
bool noinstr in_task_stack(unsigned long *stack, struct task_struct *task,
arch/x86/kernel/dumpstack.c
307
unsigned long *stack, const char *log_lvl)
arch/x86/kernel/dumpstack.c
319
__show_trace_log_lvl(task, regs, stack, log_lvl);
arch/x86/kernel/dumpstack.c
36
if (stack < begin || stack >= end)
arch/x86/kernel/dumpstack.c
48
bool noinstr in_entry_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack.c
55
if ((void *)stack < begin || (void *)stack >= end)
arch/x86/kernel/dumpstack_32.c
108
int get_stack_info(unsigned long *stack, struct task_struct *task,
arch/x86/kernel/dumpstack_32.c
111
if (!stack)
arch/x86/kernel/dumpstack_32.c
116
if (in_task_stack(stack, task, info))
arch/x86/kernel/dumpstack_32.c
122
if (in_entry_stack(stack, info))
arch/x86/kernel/dumpstack_32.c
125
if (in_hardirq_stack(stack, info))
arch/x86/kernel/dumpstack_32.c
128
if (in_softirq_stack(stack, info))
arch/x86/kernel/dumpstack_32.c
131
if (in_doublefault_stack(stack, info))
arch/x86/kernel/dumpstack_32.c
38
static bool in_hardirq_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack_32.c
47
if (stack < begin || stack > end)
arch/x86/kernel/dumpstack_32.c
63
static bool in_softirq_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack_32.c
72
if (stack < begin || stack > end)
arch/x86/kernel/dumpstack_32.c
88
static bool in_doublefault_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack_32.c
93
void *begin = ss->stack;
arch/x86/kernel/dumpstack_32.c
94
void *end = begin + sizeof(ss->stack);
arch/x86/kernel/dumpstack_32.c
96
if ((void *)stack < begin || (void *)stack >= end)
arch/x86/kernel/dumpstack_64.c
135
static __always_inline bool in_irq_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack_64.c
153
if (stack < begin || stack >= end)
arch/x86/kernel/dumpstack_64.c
170
bool noinstr get_stack_info_noinstr(unsigned long *stack, struct task_struct *task,
arch/x86/kernel/dumpstack_64.c
173
if (in_task_stack(stack, task, info))
arch/x86/kernel/dumpstack_64.c
179
if (in_exception_stack(stack, info))
arch/x86/kernel/dumpstack_64.c
182
if (in_irq_stack(stack, info))
arch/x86/kernel/dumpstack_64.c
185
if (in_entry_stack(stack, info))
arch/x86/kernel/dumpstack_64.c
191
int get_stack_info(unsigned long *stack, struct task_struct *task,
arch/x86/kernel/dumpstack_64.c
196
if (!stack)
arch/x86/kernel/dumpstack_64.c
199
if (!get_stack_info_noinstr(stack, task, info))
arch/x86/kernel/dumpstack_64.c
94
static __always_inline bool in_exception_stack(unsigned long *stack, struct stack_info *info)
arch/x86/kernel/dumpstack_64.c
96
unsigned long begin, end, stk = (unsigned long)stack;
arch/x86/kernel/ftrace.c
656
unsigned long *stack = (unsigned long *)kernel_stack_pointer(regs);
arch/x86/kernel/ftrace.c
658
unsigned long *parent = (unsigned long *)stack;
arch/x86/kernel/irq_32.c
54
static void call_on_stack(void *func, void *stack)
arch/x86/kernel/irq_32.c
59
: [sp] "+b" (stack)
arch/x86/kernel/irq_64.c
38
char *stack = (char *)per_cpu_ptr(&irq_stack_backing_store, cpu);
arch/x86/kernel/irq_64.c
44
phys_addr_t pa = per_cpu_ptr_to_phys(stack + (i << PAGE_SHIFT));
arch/x86/kernel/kexec-bzimage64.c
488
void *stack;
arch/x86/kernel/kexec-bzimage64.c
665
stack = kexec_purgatory_get_symbol_addr(image, "stack_end");
arch/x86/kernel/kexec-bzimage64.c
666
if (IS_ERR(stack)) {
arch/x86/kernel/kexec-bzimage64.c
672
regs64.rsp = (unsigned long)stack;
arch/x86/kernel/process.c
173
unsigned long sp = args->stack;
arch/x86/kernel/rethook.c
119
unsigned long *stack = (unsigned long *)regs->sp;
arch/x86/kernel/rethook.c
121
rh->ret_addr = stack[0];
arch/x86/kernel/rethook.c
125
stack[0] = (unsigned long) arch_rethook_trampoline;
arch/x86/kernel/sys_ia32.c
250
.stack = newsp,
arch/x86/kernel/traps.c
1053
unsigned long sp, *stack;
arch/x86/kernel/traps.c
1072
stack = (unsigned long *)sp;
arch/x86/kernel/traps.c
1074
if (!get_stack_info_noinstr(stack, current, &info) || info.type == STACK_TYPE_ENTRY ||
arch/x86/kernel/uprobes.c
1790
return regs->sp < ret->stack;
arch/x86/kernel/uprobes.c
1792
return regs->sp <= ret->stack;
arch/x86/net/bpf_jit_comp.c
1603
#define LOAD_TAIL_CALL_CNT_PTR(stack) \
arch/x86/net/bpf_jit_comp.c
1604
__LOAD_TCC_PTR(BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack))
arch/x86/net/bpf_jit_comp.c
3148
#define LOAD_TRAMP_TAIL_CALL_CNT_PTR(stack) \
arch/x86/net/bpf_jit_comp.c
3149
__LOAD_TCC_PTR(-round_up(stack, 8) - 8)
arch/x86/net/bpf_jit_comp.c
710
#define BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack) (-16 - round_up(stack, 8))
arch/x86/um/tls_32.c
35
struct stub_data *data = (void *)task->mm->context.id.stack;
arch/x86/xen/xen-ops.h
302
void __noreturn xen_cpu_bringup_again(unsigned long stack);
arch/xtensa/include/asm/smp.h
18
unsigned long stack;
arch/xtensa/kernel/asm-offsets.c
80
DEFINE(TASK_THREAD_INFO, offsetof (struct task_struct, stack));
arch/xtensa/kernel/process.c
271
unsigned long usp_thread_fn = args->stack;
arch/xtensa/kernel/smp.c
241
start_info.stack = (unsigned long)task_pt_regs(idle);
arch/xtensa/kernel/smp.c
245
__func__, cpu, idle, start_info.stack);
crypto/algapi.c
114
struct list_head *stack,
crypto/algapi.c
120
spawn = list_first_entry_or_null(stack, struct crypto_spawn, list);
crypto/algapi.c
127
if (list_is_last(&n->list, stack))
crypto/algapi.c
172
LIST_HEAD(stack);
crypto/algapi.c
197
list_move(&spawn->list, &stack);
crypto/algapi.c
228
} while ((spawns = crypto_more_spawns(alg, &stack, &top,
drivers/char/random.c
1285
struct entropy_timer_state *stack = PTR_ALIGN((void *)stack_bytes, SMP_CACHE_BYTES);
drivers/char/random.c
1292
stack->entropy = random_get_entropy();
drivers/char/random.c
1293
if (stack->entropy != last)
drivers/char/random.c
1295
last = stack->entropy;
drivers/char/random.c
1297
stack->samples_per_bit = DIV_ROUND_UP(NUM_TRIAL_SAMPLES, num_different + 1);
drivers/char/random.c
1298
if (stack->samples_per_bit > MAX_SAMPLES_PER_BIT)
drivers/char/random.c
1301
atomic_set(&stack->samples, 0);
drivers/char/random.c
1302
timer_setup_on_stack(&stack->timer, entropy_timer, 0);
drivers/char/random.c
1311
if (!timer_pending(&stack->timer) && timer_delete_sync_try(&stack->timer) >= 0) {
drivers/char/random.c
1337
stack->timer.expires = jiffies;
drivers/char/random.c
1339
add_timer_on(&stack->timer, cpu);
drivers/char/random.c
1343
mix_pool_bytes(&stack->entropy, sizeof(stack->entropy));
drivers/char/random.c
1345
stack->entropy = random_get_entropy();
drivers/char/random.c
1347
mix_pool_bytes(&stack->entropy, sizeof(stack->entropy));
drivers/char/random.c
1351
timer_delete_sync(&stack->timer);
drivers/char/random.c
1352
timer_destroy_on_stack(&stack->timer);
drivers/firewire/core-device.c
657
u32 *stack, *rom __free(kfree) = NULL;
drivers/firewire/core-device.c
663
sizeof(*stack) * MAX_CONFIG_ROM_SIZE, GFP_KERNEL);
drivers/firewire/core-device.c
667
stack = &rom[MAX_CONFIG_ROM_SIZE];
drivers/firewire/core-device.c
703
stack[sp++] = 0xc0000005;
drivers/firewire/core-device.c
711
key = stack[--sp];
drivers/firewire/core-device.c
762
stack[sp++] = i + rom[i];
drivers/firewire/core-topology.c
106
struct list_head stack;
drivers/firewire/core-topology.c
113
INIT_LIST_HEAD(&stack);
drivers/firewire/core-topology.c
177
for (i = 0, h = &stack; i < child_port_count; i++)
drivers/firewire/core-topology.c
234
__list_del(h->prev, &stack);
drivers/firewire/core-topology.c
235
list_add_tail(&node->link, &stack);
drivers/gpu/drm/drm_mm.c
116
node->stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/drm_mm.c
129
if (!node->stack) {
drivers/gpu/drm/drm_mm.c
135
stack_depot_snprint(node->stack, buf, BUFSZ, 0);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1516
struct drm_i915_gem_relocation_entry stack[N_RELOC(512)];
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1534
struct drm_i915_gem_relocation_entry *r = stack;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1536
min_t(unsigned long, remain, ARRAY_SIZE(stack));
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1588
__put_user(offset, &urelocs[r - stack].presumed_offset);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1590
urelocs += ARRAY_SIZE(stack);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
279
struct page *stack[32], **pages = stack, *page;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
315
if (n_pages > ARRAY_SIZE(stack)) {
drivers/gpu/drm/i915/gem/i915_gem_pages.c
326
if (pages != stack)
drivers/gpu/drm/i915/gem/i915_gem_pages.c
338
unsigned long stack[32], *pfns = stack, i;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
345
if (n_pfn > ARRAY_SIZE(stack)) {
drivers/gpu/drm/i915/gem/i915_gem_pages.c
356
if (pfns != stack)
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
1016
if (buf && stack_depot_snprint(ct->requests.lost_and_found[n].stack,
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.c
415
ct->requests.lost_and_found[lost].stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/i915/gt/uc/intel_guc_ct.h
91
depot_stack_handle_t stack;
drivers/gpu/drm/i915/i915_scheduler.c
161
struct i915_dependency stack;
drivers/gpu/drm/i915/i915_scheduler.c
172
stack.signaler = node;
drivers/gpu/drm/i915/i915_scheduler.c
173
list_add(&stack.dfs_link, &dfs);
drivers/gpu/drm/i915/i915_scheduler.c
227
if (stack.dfs_link.next == stack.dfs_link.prev)
drivers/gpu/drm/i915/i915_scheduler.c
230
__list_del_entry(&stack.dfs_link);
drivers/gpu/drm/i915/i915_vma.c
81
if (!vma->node.stack) {
drivers/gpu/drm/i915/i915_vma.c
88
stack_depot_snprint(vma->node.stack, buf, sizeof(buf), 0);
drivers/gpu/drm/i915/selftests/i915_random.c
48
char stack[128];
drivers/gpu/drm/i915/selftests/i915_random.c
50
if (WARN_ON(elsz > sizeof(stack) || count > U32_MAX))
drivers/gpu/drm/i915/selftests/i915_random.c
64
memcpy(stack, arr + count * elsz, elsz);
drivers/gpu/drm/i915/selftests/i915_random.c
66
memcpy(arr + swp * elsz, stack, elsz);
drivers/gpu/drm/nouveau/nvif/mem.c
53
u8 stack[128];
drivers/gpu/drm/nouveau/nvif/mem.c
60
if (sizeof(*args) + argc > sizeof(stack)) {
drivers/gpu/drm/nouveau/nvif/mem.c
64
args = (void *)stack;
drivers/gpu/drm/nouveau/nvif/mem.c
81
if (args != (void *)stack)
drivers/gpu/drm/nouveau/nvif/object.c
108
u8 stack[128];
drivers/gpu/drm/nouveau/nvif/object.c
114
if (args_size > sizeof(stack)) {
drivers/gpu/drm/nouveau/nvif/object.c
119
args = (void *)stack;
drivers/gpu/drm/nouveau/nvif/object.c
129
if (args != (void *)stack)
drivers/gpu/drm/nouveau/nvif/vmm.c
40
u8 stack[48];
drivers/gpu/drm/nouveau/nvif/vmm.c
43
if (sizeof(*args) + argc > sizeof(stack)) {
drivers/gpu/drm/nouveau/nvif/vmm.c
47
args = (void *)stack;
drivers/gpu/drm/nouveau/nvif/vmm.c
59
if (args != (void *)stack)
drivers/gpu/drm/xe/xe_guc_ct.c
109
ct->fast_req[slot].stack = stack_depot_save(entries, n, GFP_NOWAIT);
drivers/gpu/drm/xe/xe_guc_ct.c
119
if (buf && stack_depot_snprint(ct->fast_req[slot].stack, buf, SZ_4K, 0))
drivers/gpu/drm/xe/xe_guc_ct_types.h
117
depot_stack_handle_t stack;
drivers/gpu/drm/xe/xe_tlb_inval.c
44
bool stack = test_bit(FENCE_STACK_BIT, &fence->base.flags);
drivers/gpu/drm/xe/xe_tlb_inval.c
441
bool stack)
drivers/gpu/drm/xe/xe_tlb_inval.c
450
if (stack)
drivers/gpu/drm/xe/xe_tlb_inval.c
52
if (!stack)
drivers/gpu/drm/xe/xe_tlb_inval.h
30
bool stack);
drivers/iommu/iommu.c
864
LIST_HEAD(stack);
drivers/iommu/iommu.c
885
list_move_tail(&iter->list, &stack);
drivers/iommu/iommu.c
890
list_for_each_entry_reverse(top, &stack, list)
drivers/iommu/iommu.c
894
list_move_tail(&iter->list, &stack);
drivers/iommu/iommu.c
901
list_move_tail(&iter->list, &stack);
drivers/iommu/iommu.c
908
list_splice(&stack, regions);
drivers/md/dm-vdo/indexer/radix-sort.c
172
static inline int push_bins(struct task **stack, struct task *end_of_stack,
drivers/md/dm-vdo/indexer/radix-sort.c
190
if (*stack >= end_of_stack)
drivers/md/dm-vdo/indexer/radix-sort.c
193
push_task(stack, pile_start, size, offset, length);
drivers/md/dm-vdo/indexer/radix-sort.c
220
radix_sorter->end_of_stack = radix_sorter->stack + stack_size;
drivers/md/dm-vdo/indexer/radix-sort.c
240
struct task *task_stack = sorter->stack;
drivers/md/dm-vdo/indexer/radix-sort.c
268
for (*task_stack = start; task_stack >= sorter->stack; task_stack--) {
drivers/md/dm-vdo/indexer/radix-sort.c
62
struct task stack[];
drivers/media/mc/mc-entity.c
300
graph->stack[graph->top].link = entity->links.next;
drivers/media/mc/mc-entity.c
301
graph->stack[graph->top].entity = entity;
drivers/media/mc/mc-entity.c
308
entity = graph->stack[graph->top].entity;
drivers/media/mc/mc-entity.c
314
#define link_top(en) ((en)->stack[(en)->top].link)
drivers/media/mc/mc-entity.c
315
#define stack_top(en) ((en)->stack[(en)->top].entity)
drivers/media/mc/mc-entity.c
352
graph->stack[graph->top].entity = NULL;
drivers/media/mc/mc-entity.c
471
} stack;
drivers/media/mc/mc-entity.c
479
return &walk->stack.entries[walk->stack.top];
drivers/media/mc/mc-entity.c
484
return walk->stack.top == -1;
drivers/media/mc/mc-entity.c
494
if (walk->stack.size >= 256)
drivers/media/mc/mc-entity.c
497
new_size = walk->stack.size + MEDIA_PIPELINE_STACK_GROW_STEP;
drivers/media/mc/mc-entity.c
499
entries = krealloc(walk->stack.entries,
drivers/media/mc/mc-entity.c
500
new_size * sizeof(*walk->stack.entries),
drivers/media/mc/mc-entity.c
505
walk->stack.entries = entries;
drivers/media/mc/mc-entity.c
506
walk->stack.size = new_size;
drivers/media/mc/mc-entity.c
518
if (walk->stack.top + 1 >= walk->stack.size) {
drivers/media/mc/mc-entity.c
524
walk->stack.top++;
drivers/media/mc/mc-entity.c
531
walk->stack.top, pad->entity->name, pad->index);
drivers/media/mc/mc-entity.c
545
if (WARN_ON(walk->stack.top < 0))
drivers/media/mc/mc-entity.c
553
walk->stack.top);
drivers/media/mc/mc-entity.c
555
walk->stack.top--;
drivers/media/mc/mc-entity.c
563
walk->stack.top);
drivers/media/mc/mc-entity.c
571
kfree(walk->stack.entries);
drivers/media/mc/mc-entity.c
733
walk.stack.top = -1;
drivers/media/mc/mc-entity.c
758
WARN_ON(walk.stack.top != -1);
drivers/misc/altera-stapl/altera-jtag.h
67
long stack[ALTERA_STACK_SIZE];
drivers/misc/altera-stapl/altera.c
1058
swap(stack[stack_ptr - index], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
1067
stack[stack_ptr] = stack[stack_ptr - index];
drivers/misc/altera-stapl/altera.c
1078
vars[args[0]] = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1135
index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1138
longptr_tmp[index] = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1217
long_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1220
long_idx = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1247
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1273
if (stack[--stack_ptr] == 0) {
drivers/misc/altera-stapl/altera.c
1292
long_idx = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1293
long_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1302
long_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1361
index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1362
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1384
index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1385
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1407
index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1408
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1431
index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1432
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1454
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1467
index = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
1473
stack[stack_ptr - 1] = longptr_tmp[index];
drivers/misc/altera-stapl/altera.c
1478
stack[stack_ptr - 1] =
drivers/misc/altera-stapl/altera.c
1504
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1507
index = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
1528
stack[stack_ptr - 1] = long_tmp;
drivers/misc/altera-stapl/altera.c
1540
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1606
variable_id = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1607
long_idx = stack[--stack_ptr];/* right indx */
drivers/misc/altera-stapl/altera.c
1608
long_idx2 = stack[--stack_ptr];/* left indx */
drivers/misc/altera-stapl/altera.c
1676
copy_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1677
copy_index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1678
copy_index2 = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1688
destleft = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1833
capture_index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1834
scan_index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1844
scan_right = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1845
scan_left = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1851
long_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1951
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1957
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1995
index1 = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1996
index2 = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1997
mask_index = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
1998
long_count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
2009
s32 mask_right = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
2010
s32 mask_left = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
2052
stack[stack_ptr++] = long_tmp;
drivers/misc/altera-stapl/altera.c
213
long *stack = astate->stack;
drivers/misc/altera-stapl/altera.c
528
stack[stack_ptr] = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
534
swap(stack[stack_ptr - 2], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
539
stack[stack_ptr - 1] += stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
545
stack[stack_ptr - 1] -= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
551
stack[stack_ptr - 1] *= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
557
stack[stack_ptr - 1] /= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
563
stack[stack_ptr - 1] %= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
569
stack[stack_ptr - 1] <<= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
575
stack[stack_ptr - 1] >>= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
580
stack[stack_ptr - 1] ^= (-1L);
drivers/misc/altera-stapl/altera.c
586
stack[stack_ptr - 1] &= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
592
stack[stack_ptr - 1] |= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
598
stack[stack_ptr - 1] ^= stack[stack_ptr];
drivers/misc/altera-stapl/altera.c
604
stack[stack_ptr - 1] = stack[stack_ptr - 1] ? 0L : 1L;
drivers/misc/altera-stapl/altera.c
610
stack[stack_ptr - 1] =
drivers/misc/altera-stapl/altera.c
611
(stack[stack_ptr - 1] > stack[stack_ptr]) ?
drivers/misc/altera-stapl/altera.c
619
stack[stack_ptr - 1] =
drivers/misc/altera-stapl/altera.c
620
(stack[stack_ptr - 1] < stack[stack_ptr]) ?
drivers/misc/altera-stapl/altera.c
657
pc = stack[--stack_ptr] + code_sect;
drivers/misc/altera-stapl/altera.c
674
s32 a = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
675
s32 b = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
676
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
677
count = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
684
stack[stack_ptr - 1] =
drivers/misc/altera-stapl/altera.c
698
"%ld", stack[--stack_ptr]);
drivers/misc/altera-stapl/altera.c
715
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
716
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
728
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
729
count = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
733
stack[stack_ptr - 1] = get_unaligned_le32(&charbuf[0]);
drivers/misc/altera-stapl/altera.c
743
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
744
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
756
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
757
count = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
761
stack[stack_ptr - 1] = get_unaligned_le32(&charbuf[0]);
drivers/misc/altera-stapl/altera.c
766
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
777
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
778
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
789
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
802
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
803
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
810
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
822
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
823
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
835
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
848
count = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
849
long_tmp = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
858
ch = (char) stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
873
*exit_code = stack[--stack_ptr];
drivers/misc/altera-stapl/altera.c
881
stack[stack_ptr - 1] =
drivers/misc/altera-stapl/altera.c
882
(stack[stack_ptr - 1] == stack[stack_ptr]) ?
drivers/misc/altera-stapl/altera.c
893
if (stack[stack_ptr - 1] < 0)
drivers/misc/altera-stapl/altera.c
894
stack[stack_ptr - 1] = 0 - stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
913
swap(stack[stack_ptr - 2], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
918
swap(stack[stack_ptr - index], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
922
swap(stack[stack_ptr - 2], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
927
swap(stack[stack_ptr - index], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
932
stack[stack_ptr] = stack[stack_ptr - index];
drivers/misc/altera-stapl/altera.c
939
swap(stack[stack_ptr - index], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
943
swap(stack[stack_ptr - 2], stack[stack_ptr - 1]);
drivers/misc/altera-stapl/altera.c
948
stack[stack_ptr] = stack[stack_ptr - index];
drivers/misc/altera-stapl/altera.c
955
stack[stack_ptr] = stack[stack_ptr - index];
drivers/misc/altera-stapl/altera.c
960
stack[stack_ptr++] = 0;
drivers/misc/altera-stapl/altera.c
963
stack[stack_ptr++] = (s32) args[0];
drivers/misc/altera-stapl/altera.c
966
stack[stack_ptr++] = vars[args[0]];
drivers/misc/altera-stapl/altera.c
974
stack[stack_ptr++] = pc;
drivers/misc/altera-stapl/altera.c
988
s32 step = stack[stack_ptr - 1];
drivers/misc/altera-stapl/altera.c
989
s32 end = stack[stack_ptr - 2];
drivers/misc/altera-stapl/altera.c
990
s32 top = stack[stack_ptr - 3];
drivers/misc/lkdtm/bugs.c
190
static noinline void __lkdtm_CORRUPT_STACK(void *stack)
drivers/misc/lkdtm/bugs.c
192
memset(stack, '\xff', 64);
drivers/misc/lkdtm/bugs.c
238
static noinline void __lkdtm_REPORT_STACK_CANARY(void *stack)
drivers/misc/lkdtm/bugs.c
242
unsigned long *canary = (unsigned long *)stack;
drivers/misc/lkdtm/bugs.c
247
canary = (unsigned long *)stack + i;
drivers/misc/lkdtm/bugs.c
275
canary = (unsigned long *)stack + current_offset;
drivers/misc/lkdtm/bugs.c
651
const unsigned char *stack = task_stack_page(current);
drivers/misc/lkdtm/bugs.c
652
const unsigned char *ptr = stack - 1;
drivers/misc/lkdtm/bugs.c
665
const unsigned char *stack = task_stack_page(current);
drivers/misc/lkdtm/bugs.c
666
const unsigned char *ptr = stack + THREAD_SIZE;
drivers/misc/lkdtm/usercopy.c
32
static noinline unsigned char *trick_compiler(unsigned char *stack)
drivers/misc/lkdtm/usercopy.c
34
return stack + unconst;
drivers/net/ethernet/airoha/airoha_ppe.c
223
struct net_device_path_stack stack;
drivers/net/ethernet/airoha/airoha_ppe.c
231
err = dev_fill_forward_path(dev, addr, &stack);
drivers/net/ethernet/airoha/airoha_ppe.c
236
path = &stack.path[stack.num_paths - 1];
drivers/net/ethernet/intel/idpf/idpf_lib.c
170
struct idpf_vector_lifo *stack = &adapter->vector_stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
174
if (stack->top == stack->base) {
drivers/net/ethernet/intel/idpf/idpf_lib.c
176
stack->top);
drivers/net/ethernet/intel/idpf/idpf_lib.c
180
stack->vec_idx[--stack->top] = vec_idx;
drivers/net/ethernet/intel/idpf/idpf_lib.c
191
struct idpf_vector_lifo *stack = &adapter->vector_stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
195
if (stack->top == stack->size) {
drivers/net/ethernet/intel/idpf/idpf_lib.c
20
struct idpf_vector_lifo *stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
201
return stack->vec_idx[stack->top++];
drivers/net/ethernet/intel/idpf/idpf_lib.c
259
struct idpf_vector_lifo *stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
26
stack = &adapter->vector_stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
263
stack = &adapter->vector_stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
27
stack->size = adapter->num_msix_entries;
drivers/net/ethernet/intel/idpf/idpf_lib.c
278
q_vector_idxs[num_alloc_vecs++] = stack->vec_idx[j++];
drivers/net/ethernet/intel/idpf/idpf_lib.c
31
stack->base = min_vec;
drivers/net/ethernet/intel/idpf/idpf_lib.c
32
stack->top = min_vec;
drivers/net/ethernet/intel/idpf/idpf_lib.c
34
stack->vec_idx = kcalloc(stack->size, sizeof(u16), GFP_KERNEL);
drivers/net/ethernet/intel/idpf/idpf_lib.c
35
if (!stack->vec_idx) {
drivers/net/ethernet/intel/idpf/idpf_lib.c
41
for (i = 0; i < stack->size; i++)
drivers/net/ethernet/intel/idpf/idpf_lib.c
42
stack->vec_idx[i] = i;
drivers/net/ethernet/intel/idpf/idpf_lib.c
55
struct idpf_vector_lifo *stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
58
stack = &adapter->vector_stack;
drivers/net/ethernet/intel/idpf/idpf_lib.c
59
kfree(stack->vec_idx);
drivers/net/ethernet/intel/idpf/idpf_lib.c
60
stack->vec_idx = NULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
342
err = qmem_alloc(pfvf->dev, &pool->stack,
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
355
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
360
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
366
aq->pool.stack_base = pool->stack->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1359
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1471
err = qmem_alloc(pfvf->dev, &pool->stack,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1484
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1489
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1495
aq->pool.stack_base = pool->stack->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.h
131
struct qmem *stack;
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
109
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
27
qmem_free(pfvf->dev, pool->stack);
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
29
pool->stack = NULL;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
105
err = dev_fill_forward_path(dev, addr, &stack);
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
110
path = &stack.path[stack.num_paths - 1];
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
94
struct net_device_path_stack stack;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
108
stack_entry = &state->stack[soff / BPF_REG_SIZE];
drivers/net/ethernet/tehuti/tehuti.c
931
db->stack = (int *)(db + 1);
drivers/net/ethernet/tehuti/tehuti.c
932
db->elems = (void *)(db->stack + nelem);
drivers/net/ethernet/tehuti/tehuti.c
936
db->stack[i] = nelem - i - 1; /* to make first allocs
drivers/net/ethernet/tehuti/tehuti.c
946
return db->stack[--(db->top)];
drivers/net/ethernet/tehuti/tehuti.c
963
db->stack[(db->top)++] = n;
drivers/net/ethernet/tehuti/tehuti.h
177
int *stack;
drivers/net/ethernet/tehuti/tn40.c
111
db->stack[db->top++] = n;
drivers/net/ethernet/tehuti/tn40.c
78
db->stack = (int *)(db + 1);
drivers/net/ethernet/tehuti/tn40.c
79
db->elems = (void *)(db->stack + nelem);
drivers/net/ethernet/tehuti/tn40.c
84
db->stack[i] = nelem - i - 1;
drivers/net/ethernet/tehuti/tn40.c
96
return db->stack[--db->top];
drivers/net/ethernet/tehuti/tn40.h
79
unsigned int *stack;
drivers/net/wireguard/allowedips.c
41
static void push_rcu(struct allowedips_node **stack,
drivers/net/wireguard/allowedips.c
47
stack[(*len)++] = rcu_dereference_raw(p);
drivers/net/wireguard/allowedips.c
58
struct allowedips_node *node, *stack[MAX_ALLOWEDIPS_DEPTH] = {
drivers/net/wireguard/allowedips.c
62
while (len > 0 && (node = stack[--len])) {
drivers/net/wireguard/allowedips.c
63
push_rcu(stack, node->bit[0], &len);
drivers/net/wireguard/allowedips.c
64
push_rcu(stack, node->bit[1], &len);
drivers/net/wireguard/allowedips.c
71
struct allowedips_node *node, *stack[MAX_ALLOWEDIPS_DEPTH] = { root };
drivers/net/wireguard/allowedips.c
74
while (len > 0 && (node = stack[--len])) {
drivers/net/wireguard/allowedips.c
75
push_rcu(stack, node->bit[0], &len);
drivers/net/wireguard/allowedips.c
76
push_rcu(stack, node->bit[1], &len);
drivers/net/wireless/broadcom/b43/phy_g.c
793
_stack_save(stack, &stackidx, 0x1, (offset), \
drivers/net/wireless/broadcom/b43/phy_g.c
799
_stack_restore(stack, 0x1, \
drivers/net/wireless/broadcom/b43/phy_g.c
804
_stack_save(stack, &stackidx, 0x2, (offset), \
drivers/net/wireless/broadcom/b43/phy_g.c
810
_stack_restore(stack, 0x2, \
drivers/net/wireless/broadcom/b43/phy_g.c
815
_stack_save(stack, &stackidx, 0x3, (offset)|(table), \
drivers/net/wireless/broadcom/b43/phy_g.c
821
_stack_restore(stack, 0x3, \
drivers/net/wireless/broadcom/b43/phy_g.c
832
u32 *stack = gphy->interfstack;
drivers/net/wireless/broadcom/b43/phy_g.c
996
u32 *stack = gphy->interfstack;
drivers/net/wireless/broadcom/b43legacy/radio.c
1223
u32 *stack = phy->interfstack;
drivers/net/wireless/broadcom/b43legacy/radio.c
939
_stack_save(stack, &stackidx, 0x1, (offset), \
drivers/net/wireless/broadcom/b43legacy/radio.c
945
_stack_restore(stack, 0x1, \
drivers/net/wireless/broadcom/b43legacy/radio.c
950
_stack_save(stack, &stackidx, 0x2, (offset), \
drivers/net/wireless/broadcom/b43legacy/radio.c
956
_stack_restore(stack, 0x2, \
drivers/net/wireless/broadcom/b43legacy/radio.c
961
_stack_save(stack, &stackidx, 0x3, (offset), \
drivers/net/wireless/broadcom/b43legacy/radio.c
967
_stack_restore(stack, 0x3, \
drivers/net/wireless/broadcom/b43legacy/radio.c
980
u32 *stack = phy->interfstack;
drivers/scsi/mvumi.c
1019
mhba->tag_pool.stack = virmem;
drivers/scsi/mvumi.c
50
st->stack[i] = size - 1 - i;
drivers/scsi/mvumi.c
56
return st->stack[--st->top];
drivers/scsi/mvumi.c
63
st->stack[st->top++] = tag;
drivers/scsi/mvumi.h
470
unsigned short *stack;
fs/btrfs/send.c
3622
struct list_head *stack)
fs/btrfs/send.c
3625
list_add_tail(&moves->list, stack);
fs/btrfs/send.c
3629
list_add_tail(&moves->list, stack);
fs/btrfs/send.c
3630
list_splice_tail(&list, stack);
fs/btrfs/send.c
3641
LIST_HEAD(stack);
fs/btrfs/send.c
3649
tail_append_pending_moves(sctx, pm, &stack);
fs/btrfs/send.c
3651
while (!list_empty(&stack)) {
fs/btrfs/send.c
3652
pm = list_first_entry(&stack, struct pending_dir_move, list);
fs/btrfs/send.c
3660
tail_append_pending_moves(sctx, pm, &stack);
fs/btrfs/send.c
3665
while (!list_empty(&stack)) {
fs/btrfs/send.c
3666
pm = list_first_entry(&stack, struct pending_dir_move, list);
fs/jfs/jfs_btree.h
108
struct btframe stack[MAXTREEHEIGHT];
fs/jfs/jfs_btree.h
112
(btstack)->top = (btstack)->stack
fs/jfs/jfs_btree.h
115
( (btstack)->top == &((btstack)->stack[MAXTREEHEIGHT-1]))
fs/jfs/jfs_btree.h
126
( (btstack)->top == (btstack)->stack ? NULL : --(btstack)->top )
fs/jfs/jfs_btree.h
129
( (btstack)->top == (btstack)->stack ? NULL : (btstack)->top )
fs/jfs/jfs_btree.h
137
(long long)btstack->stack[i].bn,
fs/jfs/jfs_btree.h
138
btstack->stack[i].index);
fs/namei.c
1193
struct saved *last = nd->stack + --nd->depth;
fs/namei.c
1294
audit_inode(nd->name, nd->stack[0].link.dentry, 0);
fs/namei.c
1983
if (likely(nd->stack != nd->internal))
fs/namei.c
2026
last = nd->stack + nd->depth++;
fs/namei.c
2649
name = nd->stack[--depth].name;
fs/namei.c
2659
nd->stack[depth++].name = name;
fs/namei.c
738
} *stack, internal[EMBEDDED_LEVELS];
fs/namei.c
755
p->stack = p->internal;
fs/namei.c
785
if (now->stack != now->internal)
fs/namei.c
786
kfree(now->stack);
fs/namei.c
798
nd->stack = p;
fs/namei.c
825
struct saved *last = nd->stack + i;
fs/namei.c
846
path_put(&nd->stack[i].link);
fs/namei.c
889
struct saved *last = nd->stack + i;
fs/overlayfs/export.c
696
struct ovl_path *stack = &origin;
fs/overlayfs/export.c
703
err = ovl_check_origin_fh(ofs, fh, false, NULL, &stack);
fs/overlayfs/export.c
748
err = ovl_check_origin_fh(ofs, fh, true, NULL, &stack);
fs/overlayfs/namei.c
1082
struct ovl_path *stack;
fs/overlayfs/namei.c
1149
ctx->stack = ovl_stack_alloc(ofs->numlayer - 1);
fs/overlayfs/namei.c
1150
if (!ctx->stack)
fs/overlayfs/namei.c
1222
ctx->stack[ctx->ctr].dentry = this;
fs/overlayfs/namei.c
1223
ctx->stack[ctx->ctr].layer = lower.layer;
fs/overlayfs/namei.c
1263
if (WARN_ON(ctx->stack != NULL)) {
fs/overlayfs/namei.c
1267
ctx->stack = ctx->origin_path;
fs/overlayfs/namei.c
1293
origin = ctx->stack[0].dentry;
fs/overlayfs/namei.c
1311
ovl_stack_cpy(ovl_lowerstack(ctx->oe), ctx->stack, ctx->ctr);
fs/overlayfs/namei.c
1360
if (ctx->ctr > 1 && !d->is_dir && !ctx->stack[ctx->ctr - 1].dentry) {
fs/overlayfs/namei.c
1410
ovl_stack_free(ctx.stack, ctx.ctr);
fs/overlayfs/namei.c
641
struct ovl_path *stack = &origin;
fs/overlayfs/namei.c
711
err = ovl_check_origin_fh(ofs, fh, false, index, &stack);
fs/overlayfs/overlayfs.h
481
void ovl_stack_put(struct ovl_path *stack, unsigned int n);
fs/overlayfs/overlayfs.h
482
void ovl_stack_free(struct ovl_path *stack, unsigned int n);
fs/overlayfs/util.c
131
void ovl_stack_put(struct ovl_path *stack, unsigned int n)
fs/overlayfs/util.c
135
for (i = 0; stack && i < n; i++)
fs/overlayfs/util.c
136
dput(stack[i].dentry);
fs/overlayfs/util.c
139
void ovl_stack_free(struct ovl_path *stack, unsigned int n)
fs/overlayfs/util.c
141
ovl_stack_put(stack, n);
fs/overlayfs/util.c
142
kfree(stack);
include/asm-generic/vmlinux.lds.h
933
*(.note.GNU-stack) \
include/drm/drm_mm.h
178
depot_stack_handle_t stack;
include/linux/bpf_verifier.h
317
struct bpf_stack_state *stack;
include/linux/bpf_verifier.h
450
((1 << frame->stack[slot].slot_type[BPF_REG_SIZE - 1]) & (mask))) \
include/linux/bpf_verifier.h
451
? &frame->stack[slot].spilled_ptr : NULL)
include/linux/netdevice.h
3363
struct net_device_path_stack *stack);
include/linux/sched.h
1955
unsigned long stack[THREAD_SIZE/sizeof(long)];
include/linux/sched.h
1967
# define task_thread_info(task) ((struct thread_info *)(task)->stack)
include/linux/sched.h
839
void *stack;
include/linux/sched/task.h
34
unsigned long stack;
include/linux/sched/task_stack.h
23
return task->stack;
include/linux/sched/task_stack.h
31
return (unsigned long *)((unsigned long)task->stack + THREAD_SIZE) - 1;
include/linux/sched/task_stack.h
33
return task->stack;
include/linux/sched/task_stack.h
39
#define task_stack_page(task) ((void *)(task)->stack)
include/linux/sched/task_stack.h
91
void *stack = task_stack_page(current);
include/linux/sched/task_stack.h
94
return (obj >= stack) && (obj < (stack + THREAD_SIZE));
include/linux/stackdepot.h
207
void stack_depot_print(depot_stack_handle_t stack);
include/linux/thread_info.h
219
static inline int arch_within_stack_frames(const void * const stack,
include/linux/uprobes.h
162
unsigned long stack; /* stack pointer */
include/media/media-entity.h
94
} stack[MEDIA_ENTITY_ENUM_MAX_DEPTH];
include/uapi/linux/sched.h
98
__aligned_u64 stack;
init/init_task.c
102
.stack = init_stack,
io_uring/slist.h
60
struct io_wq_work_node *stack)
io_uring/slist.h
62
node->next = stack->next;
io_uring/slist.h
63
stack->next = node;
io_uring/slist.h
74
struct io_wq_work_node *wq_stack_extract(struct io_wq_work_node *stack)
io_uring/slist.h
76
struct io_wq_work_node *node = stack->next;
io_uring/slist.h
78
stack->next = node->next;
kernel/bpf/btf.c
2012
v = &env->stack[env->top_stack++];
kernel/bpf/btf.c
2030
env->stack[env->top_stack - 1].next_member = next_member;
kernel/bpf/btf.c
2037
u32 type_id = env->stack[--(env->top_stack)].type_id;
kernel/bpf/btf.c
2048
return env->top_stack ? &env->stack[env->top_stack - 1] : NULL;
kernel/bpf/btf.c
318
struct resolve_vertex stack[MAX_RESOLVE_DEPTH];
kernel/bpf/core.c
2327
u64 stack[stack_size / sizeof(u64)]; \
kernel/bpf/core.c
2330
kmsan_unpoison_memory(stack, sizeof(stack)); \
kernel/bpf/core.c
2331
FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \
kernel/bpf/core.c
2341
u64 stack[stack_size / sizeof(u64)]; \
kernel/bpf/core.c
2344
kmsan_unpoison_memory(stack, sizeof(stack)); \
kernel/bpf/core.c
2345
FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \
kernel/bpf/log.c
779
slot_type = state->stack[i].slot_type[j];
kernel/bpf/log.c
788
reg = &state->stack[i].spilled_ptr;
kernel/bpf/log.c
789
switch (state->stack[i].slot_type[BPF_REG_SIZE - 1]) {
kernel/bpf/log.c
793
if (state->stack[i].slot_type[j] == STACK_SPILL)
kernel/bpf/log.c
803
reg = &state->stack[i].spilled_ptr;
kernel/bpf/verifier.c
1009
if (!state->stack[spi].spilled_ptr.dynptr.first_slot)
kernel/bpf/verifier.c
1013
if (state->stack[spi].slot_type[i] != STACK_DYNPTR ||
kernel/bpf/verifier.c
1014
state->stack[spi - 1].slot_type[i] != STACK_DYNPTR)
kernel/bpf/verifier.c
1039
return state->stack[spi].spilled_ptr.dynptr.type == dynptr_type;
kernel/bpf/verifier.c
1066
struct bpf_stack_state *slot = &state->stack[spi - i];
kernel/bpf/verifier.c
1104
struct bpf_stack_state *slot = &state->stack[spi - i];
kernel/bpf/verifier.c
1139
struct bpf_stack_state *slot = &state->stack[spi - i];
kernel/bpf/verifier.c
1160
struct bpf_stack_state *slot = &state->stack[spi - i];
kernel/bpf/verifier.c
1202
slot = &state->stack[spi];
kernel/bpf/verifier.c
1230
slot = &state->stack[spi];
kernel/bpf/verifier.c
1286
slot = &state->stack[spi];
kernel/bpf/verifier.c
1305
slot = &state->stack[spi];
kernel/bpf/verifier.c
1323
static bool is_stack_slot_special(const struct bpf_stack_state *stack)
kernel/bpf/verifier.c
1325
enum bpf_stack_slot_type type = stack->slot_type[BPF_REG_SIZE - 1];
kernel/bpf/verifier.c
1346
static bool is_spilled_reg(const struct bpf_stack_state *stack)
kernel/bpf/verifier.c
1348
return stack->slot_type[BPF_REG_SIZE - 1] == STACK_SPILL;
kernel/bpf/verifier.c
1351
static bool is_spilled_scalar_reg(const struct bpf_stack_state *stack)
kernel/bpf/verifier.c
1353
return stack->slot_type[BPF_REG_SIZE - 1] == STACK_SPILL &&
kernel/bpf/verifier.c
1354
stack->spilled_ptr.type == SCALAR_VALUE;
kernel/bpf/verifier.c
1357
static bool is_spilled_scalar_reg64(const struct bpf_stack_state *stack)
kernel/bpf/verifier.c
1359
return stack->slot_type[0] == STACK_SPILL &&
kernel/bpf/verifier.c
1360
stack->spilled_ptr.type == SCALAR_VALUE;
kernel/bpf/verifier.c
1467
dst->stack = copy_array(dst->stack, src->stack, n, sizeof(struct bpf_stack_state),
kernel/bpf/verifier.c
1469
if (!dst->stack)
kernel/bpf/verifier.c
1501
state->stack = realloc_array(state->stack, old_n, n, sizeof(struct bpf_stack_state));
kernel/bpf/verifier.c
1502
if (!state->stack)
kernel/bpf/verifier.c
1684
kfree(state->stack);
kernel/bpf/verifier.c
17413
if (!is_spilled_reg(&func->stack[j]))
kernel/bpf/verifier.c
17415
reg = &func->stack[j].spilled_ptr;
kernel/bpf/verifier.c
17435
: &vstate->frame[e->frameno]->stack[e->spi].spilled_ptr;
kernel/bpf/verifier.c
1747
memcpy(dst, src, offsetof(struct bpf_func_state, stack));
kernel/bpf/verifier.c
19049
int *stack = NULL, *postorder = NULL, *state = NULL;
kernel/bpf/verifier.c
19054
stack = kvzalloc_objs(int, env->prog->len, GFP_KERNEL_ACCOUNT);
kernel/bpf/verifier.c
19055
if (!postorder || !state || !stack) {
kernel/bpf/verifier.c
19058
kvfree(stack);
kernel/bpf/verifier.c
19064
stack[0] = env->subprog_info[i].start;
kernel/bpf/verifier.c
19067
top = stack[stack_sz - 1];
kernel/bpf/verifier.c
19077
stack[stack_sz++] = succ->items[s];
kernel/bpf/verifier.c
19087
kvfree(stack);
kernel/bpf/verifier.c
19661
__mark_reg_not_init(env, &st->stack[i].spilled_ptr);
kernel/bpf/verifier.c
19663
st->stack[i].slot_type[j] = STACK_INVALID;
kernel/bpf/verifier.c
19984
struct bpf_stack_state *stack)
kernel/bpf/verifier.c
19988
for (i = 0; i < ARRAY_SIZE(stack->slot_type); ++i) {
kernel/bpf/verifier.c
19989
if ((stack->slot_type[i] == STACK_MISC) ||
kernel/bpf/verifier.c
19990
(stack->slot_type[i] == STACK_INVALID && env->allow_uninit_stack))
kernel/bpf/verifier.c
19999
struct bpf_stack_state *stack)
kernel/bpf/verifier.c
20001
if (is_spilled_scalar_reg64(stack))
kernel/bpf/verifier.c
20002
return &stack->spilled_ptr;
kernel/bpf/verifier.c
20004
if (is_stack_all_misc(env, stack))
kernel/bpf/verifier.c
20027
old->stack[spi].slot_type[i % BPF_REG_SIZE] !=
kernel/bpf/verifier.c
20028
cur->stack[spi].slot_type[i % BPF_REG_SIZE]))
kernel/bpf/verifier.c
20031
if (old->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_INVALID)
kernel/bpf/verifier.c
20035
old->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_MISC)
kernel/bpf/verifier.c
20049
old_reg = scalar_reg_for_stack(env, &old->stack[spi]);
kernel/bpf/verifier.c
20050
cur_reg = scalar_reg_for_stack(env, &cur->stack[spi]);
kernel/bpf/verifier.c
20062
if (old->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_MISC &&
kernel/bpf/verifier.c
20063
cur->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_ZERO)
kernel/bpf/verifier.c
20065
if (old->stack[spi].slot_type[i % BPF_REG_SIZE] !=
kernel/bpf/verifier.c
20066
cur->stack[spi].slot_type[i % BPF_REG_SIZE])
kernel/bpf/verifier.c
20076
switch (old->stack[spi].slot_type[BPF_REG_SIZE - 1]) {
kernel/bpf/verifier.c
20088
if (!regsafe(env, &old->stack[spi].spilled_ptr,
kernel/bpf/verifier.c
20089
&cur->stack[spi].spilled_ptr, idmap, exact))
kernel/bpf/verifier.c
20093
old_reg = &old->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20094
cur_reg = &cur->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20101
old_reg = &old->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20102
cur_reg = &cur->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20117
old_reg = &old->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20118
cur_reg = &cur->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
20307
if (!is_spilled_reg(&state->stack[i]))
kernel/bpf/verifier.c
20309
state_reg = &state->stack[i].spilled_ptr;
kernel/bpf/verifier.c
20461
if (state->stack[i].slot_type[0] != STACK_ITER)
kernel/bpf/verifier.c
20464
slot = &state->stack[i].spilled_ptr;
kernel/bpf/verifier.c
20468
cur_slot = &cur->frame[fr]->stack[i].spilled_ptr;
kernel/bpf/verifier.c
20582
iter_state = &func(env, iter_reg)->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
25784
u32 *stack, *pre, *low, *dfs;
kernel/bpf/verifier.c
25799
stack = kvcalloc(insn_cnt, sizeof(int), GFP_KERNEL_ACCOUNT);
kernel/bpf/verifier.c
25803
if (!stack || !pre || !low || !dfs) {
kernel/bpf/verifier.c
25893
stack[stack_sz++] = w;
kernel/bpf/verifier.c
25918
assign_scc = stack[stack_sz - 1] != w; /* two or more elements? */
kernel/bpf/verifier.c
25929
t = stack[--stack_sz];
kernel/bpf/verifier.c
25947
kvfree(stack);
kernel/bpf/verifier.c
4781
if (!is_spilled_reg(&func->stack[j]))
kernel/bpf/verifier.c
4783
reg = &func->stack[j].spilled_ptr;
kernel/bpf/verifier.c
4811
if (!is_spilled_reg(&func->stack[j]))
kernel/bpf/verifier.c
4813
reg = &func->stack[j].spilled_ptr;
kernel/bpf/verifier.c
5049
if (!is_spilled_scalar_reg(&func->stack[i])) {
kernel/bpf/verifier.c
5053
reg = &func->stack[i].spilled_ptr;
kernel/bpf/verifier.c
5199
copy_register_state(&state->stack[spi].spilled_ptr, reg);
kernel/bpf/verifier.c
5202
state->stack[spi].slot_type[i - 1] = STACK_SPILL;
kernel/bpf/verifier.c
5206
mark_stack_slot_misc(env, &state->stack[spi].slot_type[i - 1]);
kernel/bpf/verifier.c
5263
is_spilled_reg(&state->stack[spi]) &&
kernel/bpf/verifier.c
5264
!is_spilled_scalar_reg(&state->stack[spi]) &&
kernel/bpf/verifier.c
5277
u8 type = state->stack[spi].slot_type[i];
kernel/bpf/verifier.c
5317
state->stack[spi].spilled_ptr.id = 0;
kernel/bpf/verifier.c
5342
state->stack[spi].spilled_ptr.type = NOT_INIT;
kernel/bpf/verifier.c
5344
if (is_stack_slot_special(&state->stack[spi]))
kernel/bpf/verifier.c
5346
scrub_spilled_slot(&state->stack[spi].slot_type[i]);
kernel/bpf/verifier.c
5365
state->stack[spi].slot_type[(slot - i) % BPF_REG_SIZE] = type;
kernel/bpf/verifier.c
5437
stype = &state->stack[spi].slot_type[slot % BPF_REG_SIZE];
kernel/bpf/verifier.c
5461
is_spilled_scalar_reg(&state->stack[spi])) {
kernel/bpf/verifier.c
5462
struct bpf_reg_state *spill_reg = &state->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
5471
state->stack[spi].spilled_ptr.type = NOT_INIT;
kernel/bpf/verifier.c
5526
stype = ptr_state->stack[spi].slot_type;
kernel/bpf/verifier.c
5564
stype = reg_state->stack[spi].slot_type;
kernel/bpf/verifier.c
5565
reg = &reg_state->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
5573
if (is_spilled_reg(&reg_state->stack[spi])) {
kernel/bpf/verifier.c
797
state->stack[spi].slot_type[i] = STACK_DYNPTR;
kernel/bpf/verifier.c
798
state->stack[spi - 1].slot_type[i] = STACK_DYNPTR;
kernel/bpf/verifier.c
805
mark_dynptr_stack_regs(env, &state->stack[spi].spilled_ptr,
kernel/bpf/verifier.c
806
&state->stack[spi - 1].spilled_ptr, type);
kernel/bpf/verifier.c
820
state->stack[spi].spilled_ptr.ref_obj_id = id;
kernel/bpf/verifier.c
821
state->stack[spi - 1].spilled_ptr.ref_obj_id = id;
kernel/bpf/verifier.c
8282
if (state->stack[spi].slot_type[stack_off % BPF_REG_SIZE] == STACK_DYNPTR) {
kernel/bpf/verifier.c
8302
stype = &state->stack[spi].slot_type[slot % BPF_REG_SIZE];
kernel/bpf/verifier.c
8314
if (is_spilled_reg(&state->stack[spi]) &&
kernel/bpf/verifier.c
8315
(state->stack[spi].spilled_ptr.type == SCALAR_VALUE ||
kernel/bpf/verifier.c
8318
__mark_reg_unknown(env, &state->stack[spi].spilled_ptr);
kernel/bpf/verifier.c
8320
scrub_spilled_slot(&state->stack[spi].slot_type[j]);
kernel/bpf/verifier.c
834
state->stack[spi].slot_type[i] = STACK_INVALID;
kernel/bpf/verifier.c
835
state->stack[spi - 1].slot_type[i] = STACK_INVALID;
kernel/bpf/verifier.c
838
__mark_reg_not_init(env, &state->stack[spi].spilled_ptr);
kernel/bpf/verifier.c
839
__mark_reg_not_init(env, &state->stack[spi - 1].spilled_ptr);
kernel/bpf/verifier.c
862
if (!dynptr_type_refcounted(state->stack[spi].spilled_ptr.dynptr.type)) {
kernel/bpf/verifier.c
867
ref_obj_id = state->stack[spi].spilled_ptr.ref_obj_id;
kernel/bpf/verifier.c
881
if (state->stack[i].spilled_ptr.ref_obj_id != ref_obj_id)
kernel/bpf/verifier.c
888
if (state->stack[i].slot_type[0] != STACK_DYNPTR) {
kernel/bpf/verifier.c
892
if (state->stack[i].spilled_ptr.dynptr.first_slot)
kernel/bpf/verifier.c
8932
return state->stack[spi].spilled_ptr.ref_obj_id;
kernel/bpf/verifier.c
9127
if (!is_spilled_reg(&fold->stack[i]) ||
kernel/bpf/verifier.c
9128
!is_spilled_reg(&fcur->stack[i]))
kernel/bpf/verifier.c
9132
&fold->stack[i].spilled_ptr,
kernel/bpf/verifier.c
9133
&fcur->stack[i].spilled_ptr);
kernel/bpf/verifier.c
9145
return &cur_st->frame[iter_frameno]->stack[iter_spi].spilled_ptr;
kernel/bpf/verifier.c
922
if (state->stack[spi].slot_type[0] != STACK_DYNPTR)
kernel/bpf/verifier.c
926
if (!state->stack[spi].spilled_ptr.dynptr.first_slot)
kernel/bpf/verifier.c
929
if (dynptr_type_refcounted(state->stack[spi].spilled_ptr.dynptr.type)) {
kernel/bpf/verifier.c
939
state->stack[spi].slot_type[i] = STACK_INVALID;
kernel/bpf/verifier.c
940
state->stack[spi - 1].slot_type[i] = STACK_INVALID;
kernel/bpf/verifier.c
943
dynptr_id = state->stack[spi].spilled_ptr.id;
kernel/bpf/verifier.c
956
__mark_reg_not_init(env, &state->stack[spi].spilled_ptr);
kernel/bpf/verifier.c
957
__mark_reg_not_init(env, &state->stack[spi - 1].spilled_ptr);
kernel/bpf/verifier.c
9711
return state->stack[spi].spilled_ptr.id;
kernel/bpf/verifier.c
9724
return state->stack[spi].spilled_ptr.ref_obj_id;
kernel/bpf/verifier.c
9742
return state->stack[spi].spilled_ptr.dynptr.type;
kernel/bpf/verifier.c
9824
stype = state->stack[spi].slot_type;
kernel/bpf/verifier.c
9835
if (!is_spilled_scalar_reg(&state->stack[spi]))
kernel/bpf/verifier.c
9842
reg = &state->stack[spi].spilled_ptr;
kernel/bpf/verifier.c
9934
if (spi < 0 || !state->stack[spi].spilled_ptr.ref_obj_id) {
kernel/dma/debug.c
1100
addr = (u8 *)current->stack + i * PAGE_SIZE +
kernel/events/uprobes.c
2300
ri->stack = user_stack_pointer(regs);
kernel/fork.c
2790
.stack = newsp,
kernel/fork.c
285
struct vm_stack *vm_stack = tsk->stack;
kernel/fork.c
2850
.stack = args.stack,
kernel/fork.c
2877
if (kargs->stack == 0) {
kernel/fork.c
2884
if (!access_ok((void __user *)kargs->stack, kargs->stack_size))
kernel/fork.c
2888
kargs->stack += kargs->stack_size;
kernel/fork.c
333
void *stack;
kernel/fork.c
345
stack = kasan_reset_tag(vm_area->addr);
kernel/fork.c
348
memset(stack, 0, THREAD_SIZE);
kernel/fork.c
351
tsk->stack = stack;
kernel/fork.c
355
stack = __vmalloc_node(THREAD_SIZE, THREAD_ALIGN,
kernel/fork.c
358
if (!stack)
kernel/fork.c
361
vm_area = find_vm_area(stack);
kernel/fork.c
363
vfree(stack);
kernel/fork.c
372
stack = kasan_reset_tag(stack);
kernel/fork.c
373
tsk->stack = stack;
kernel/fork.c
382
tsk->stack = NULL;
kernel/fork.c
401
struct rcu_head *rh = tsk->stack;
kernel/fork.c
412
tsk->stack = kasan_reset_tag(page_address(page));
kernel/fork.c
421
tsk->stack = NULL;
kernel/fork.c
435
struct rcu_head *rh = tsk->stack;
kernel/fork.c
442
unsigned long *stack;
kernel/fork.c
443
stack = kmem_cache_alloc_node(thread_stack_cache, THREADINFO_GFP, node);
kernel/fork.c
444
stack = kasan_reset_tag(stack);
kernel/fork.c
445
tsk->stack = stack;
kernel/fork.c
446
return stack ? 0 : -ENOMEM;
kernel/fork.c
452
tsk->stack = NULL;
kernel/fork.c
491
void *stack = task_stack_page(tsk);
kernel/fork.c
494
mod_lruvec_kmem_state(stack, NR_KERNEL_STACK_KB,
kernel/time/timer_migration.c
1763
struct tmigr_group *group, *child, **stack;
kernel/time/timer_migration.c
1767
stack = kzalloc_objs(*stack, tmigr_hierarchy_levels);
kernel/time/timer_migration.c
1768
if (!stack)
kernel/time/timer_migration.c
1772
stack[start->level] = start;
kernel/time/timer_migration.c
1788
stack[i] = group;
kernel/time/timer_migration.c
1811
group = stack[i];
kernel/time/timer_migration.c
1837
child = stack[i - 1];
kernel/time/timer_migration.c
1890
kfree(stack);
kernel/trace/fprobe.c
165
static inline bool write_fprobe_header(unsigned long *stack,
kernel/trace/fprobe.c
172
*stack = arch_encode_fprobe_header(fp, size_words);
kernel/trace/fprobe.c
176
static inline void read_fprobe_header(unsigned long *stack,
kernel/trace/fprobe.c
179
*fp = arch_decode_fprobe_header_fp(*stack);
kernel/trace/fprobe.c
180
*size_words = arch_decode_fprobe_header_size(*stack);
kernel/trace/fprobe.c
193
static inline bool write_fprobe_header(unsigned long *stack,
kernel/trace/fprobe.c
196
struct __fprobe_header *fph = (struct __fprobe_header *)stack;
kernel/trace/fprobe.c
206
static inline void read_fprobe_header(unsigned long *stack,
kernel/trace/fprobe.c
209
struct __fprobe_header *fph = (struct __fprobe_header *)stack;
kernel/trace/trace_events_hist.c
5404
unsigned long *stack, n_entries;
kernel/trace/trace_events_hist.c
5407
stack = (unsigned long *)(long)field_contents;
kernel/trace/trace_events_hist.c
5408
n_entries = *stack;
kernel/trace/trace_events_hist.c
5409
memcpy(entries, ++stack, n_entries * sizeof(unsigned long));
kernel/trace/trace_events_synth.c
469
long *stack,
kernel/trace/trace_events_synth.c
482
if (!stack[len])
kernel/trace/trace_events_synth.c
490
memcpy(data_loc, stack, len);
kernel/trace/trace_events_synth.c
561
long *stack = (long *)(long)var_ref_vals[val_idx];
kernel/trace/trace_events_synth.c
563
len = trace_stack(entry, event, stack,
kernel/trace/trace_stack.c
155
static void check_stack(unsigned long ip, unsigned long *stack)
kernel/trace/trace_stack.c
162
this_size = ((unsigned long)stack) & (THREAD_SIZE-1);
kernel/trace/trace_stack.c
171
if (!object_is_on_stack(stack))
kernel/trace/trace_stack.c
212
start = stack;
kernel/trace/trace_stack.c
249
tracer_frame = (p - stack) *
kernel/trace/trace_stack.c
295
unsigned long stack;
kernel/trace/trace_stack.c
310
check_stack(ip, &stack);
lib/ref_tracker.c
146
depot_stack_handle_t stack;
lib/ref_tracker.c
164
stack = stats->stacks[i].stack_handle;
lib/ref_tracker.c
165
if (sbuf && !stack_depot_snprint(stack, sbuf, STACK_BUF_SIZE, 4))
lib/ref_tracker.c
84
depot_stack_handle_t stack = tracker->alloc_stack_handle;
lib/ref_tracker.c
89
if (stats->stacks[i].stack_handle == stack)
lib/ref_tracker.c
94
stats->stacks[i].stack_handle = stack;
lib/stackdepot.c
364
struct stack_record *stack;
lib/stackdepot.c
382
stack = current_pool + pool_offset;
lib/stackdepot.c
385
stack->handle.pool_index_plus_1 = pool_index + 1;
lib/stackdepot.c
386
stack->handle.offset = pool_offset >> DEPOT_STACK_ALIGN;
lib/stackdepot.c
387
stack->handle.extra = 0;
lib/stackdepot.c
388
INIT_LIST_HEAD(&stack->hash_list);
lib/stackdepot.c
392
return stack;
lib/stackdepot.c
399
struct stack_record *stack;
lib/stackdepot.c
412
stack = list_first_entry(&free_stacks, struct stack_record, free_list);
lib/stackdepot.c
413
if (!poll_state_synchronize_rcu(stack->rcu_state))
lib/stackdepot.c
416
list_del(&stack->free_list);
lib/stackdepot.c
419
return stack;
lib/stackdepot.c
437
struct stack_record *stack = NULL;
lib/stackdepot.c
455
record_size = depot_stack_record_size(stack, CONFIG_STACKDEPOT_MAX_FRAMES);
lib/stackdepot.c
456
stack = depot_pop_free();
lib/stackdepot.c
458
record_size = depot_stack_record_size(stack, nr_entries);
lib/stackdepot.c
461
if (!stack) {
lib/stackdepot.c
462
stack = depot_pop_free_pool(prealloc, record_size);
lib/stackdepot.c
463
if (!stack)
lib/stackdepot.c
468
stack->hash = hash;
lib/stackdepot.c
469
stack->size = nr_entries;
lib/stackdepot.c
471
memcpy(stack->entries, entries, flex_array_size(stack, entries, nr_entries));
lib/stackdepot.c
474
refcount_set(&stack->count, 1);
lib/stackdepot.c
479
refcount_set(&stack->count, REFCOUNT_SATURATED);
lib/stackdepot.c
488
kmsan_unpoison_memory(stack, record_size);
lib/stackdepot.c
490
return stack;
lib/stackdepot.c
501
struct stack_record *stack;
lib/stackdepot.c
516
stack = pool + offset;
lib/stackdepot.c
517
if (WARN_ON(!refcount_read(&stack->count)))
lib/stackdepot.c
520
return stack;
lib/stackdepot.c
524
static void depot_free_stack(struct stack_record *stack)
lib/stackdepot.c
539
list_del_rcu(&stack->hash_list);
lib/stackdepot.c
551
stack->rcu_state = get_state_synchronize_rcu();
lib/stackdepot.c
558
list_add_tail(&stack->free_list, &free_stacks);
lib/stackdepot.c
596
struct stack_record *stack, *ret = NULL;
lib/stackdepot.c
609
list_for_each_entry_rcu(stack, bucket, hash_list) {
lib/stackdepot.c
610
if (stack->hash != hash || stack->size != size)
lib/stackdepot.c
618
if (data_race(stackdepot_memcmp(entries, stack->entries, size)))
lib/stackdepot.c
629
if ((flags & STACK_DEPOT_FLAG_GET) && !refcount_inc_not_zero(&stack->count))
lib/stackdepot.c
632
ret = stack;
lib/stackdepot.c
763
struct stack_record *stack;
lib/stackdepot.c
775
stack = depot_fetch_stack(handle);
lib/stackdepot.c
780
if (WARN(!stack, "corrupt handle or use after stack_depot_put()"))
lib/stackdepot.c
783
*entries = stack->entries;
lib/stackdepot.c
784
return stack->size;
lib/stackdepot.c
790
struct stack_record *stack;
lib/stackdepot.c
795
stack = depot_fetch_stack(handle);
lib/stackdepot.c
800
if (WARN(!stack, "corrupt handle or unbalanced stack_depot_put()"))
lib/stackdepot.c
803
if (refcount_dec_and_test(&stack->count))
lib/stackdepot.c
804
depot_free_stack(stack);
lib/stackdepot.c
808
void stack_depot_print(depot_stack_handle_t stack)
lib/stackdepot.c
813
nr_entries = stack_depot_fetch(stack, &entries);
lib/tests/fortify_kunit.c
87
char stack[80] = "Test!";
lib/tests/fortify_kunit.c
89
KUNIT_EXPECT_FALSE(test, __is_constexpr(__builtin_strlen(stack)));
lib/tests/fortify_kunit.c
90
KUNIT_EXPECT_EQ(test, __compiletime_strlen(stack), 5);
mm/kasan/common.c
61
void kasan_set_track(struct kasan_track *track, depot_stack_handle_t stack)
mm/kasan/common.c
71
track->stack = stack;
mm/kasan/common.c
76
depot_stack_handle_t stack;
mm/kasan/common.c
78
stack = kasan_save_stack(flags, STACK_DEPOT_FLAG_CAN_ALLOC);
mm/kasan/common.c
79
kasan_set_track(track, stack);
mm/kasan/kasan.h
195
depot_stack_handle_t stack;
mm/kasan/kasan.h
398
void kasan_set_track(struct kasan_track *track, depot_stack_handle_t stack);
mm/kasan/report.c
280
if (track->stack)
mm/kasan/report.c
281
stack_depot_print(track->stack);
mm/kasan/report.c
337
if (info->alloc_track.stack) {
mm/kasan/report.c
342
if (info->free_track.stack) {
mm/kasan/report.c
368
return addr >= (void *)&init_thread_union.stack &&
mm/kasan/report.c
369
(addr <= (void *)&init_thread_union.stack +
mm/kasan/report.c
370
sizeof(init_thread_union.stack));
mm/kasan/tags.c
101
depot_stack_handle_t stack, old_stack;
mm/kasan/tags.c
106
stack = kasan_save_stack(gfp_flags,
mm/kasan/tags.c
126
old_stack = entry->track.stack;
mm/kasan/tags.c
129
kasan_set_track(&entry->track, stack);
mm/kmemleak.c
1785
void *stack = try_get_task_stack(p);
mm/kmemleak.c
1786
if (stack) {
mm/kmemleak.c
1787
scan_block(stack, stack + THREAD_SIZE, NULL);
mm/kmsan/init.c
149
static void smallstack_push(struct smallstack *stack, struct page *pages)
mm/kmsan/init.c
151
KMSAN_WARN_ON(stack->index == MAX_BLOCKS);
mm/kmsan/init.c
152
stack->items[stack->index] = pages;
mm/kmsan/init.c
153
stack->index++;
mm/kmsan/init.c
157
static struct page *smallstack_pop(struct smallstack *stack)
mm/kmsan/init.c
161
KMSAN_WARN_ON(stack->index == 0);
mm/kmsan/init.c
162
stack->index--;
mm/kmsan/init.c
163
ret = stack->items[stack->index];
mm/kmsan/init.c
164
stack->items[stack->index] = NULL;
mm/mm_init.c
2608
const char *stack;
mm/mm_init.c
2611
stack = "all(pattern)";
mm/mm_init.c
2613
stack = "all(zero)";
mm/mm_init.c
2615
stack = "off";
mm/mm_init.c
2618
stack, str_on_off(want_init_on_alloc(GFP_KERNEL)),
mm/page_owner.c
178
struct stack *stack;
mm/page_owner.c
184
stack = kmalloc_obj(*stack, gfp_nested_mask(gfp_mask));
mm/page_owner.c
185
if (!stack) {
mm/page_owner.c
191
stack->stack_record = stack_record;
mm/page_owner.c
192
stack->next = NULL;
mm/page_owner.c
195
stack->next = stack_list;
mm/page_owner.c
202
smp_store_release(&stack_list, stack);
mm/page_owner.c
41
struct stack *next;
mm/page_owner.c
43
static struct stack dummy_stack;
mm/page_owner.c
44
static struct stack failure_stack;
mm/page_owner.c
45
static struct stack *stack_list;
mm/page_owner.c
53
struct stack *stack;
mm/page_owner.c
857
struct stack *stack;
mm/page_owner.c
869
stack = smp_load_acquire(&stack_list);
mm/page_owner.c
870
ctx->stack = stack;
mm/page_owner.c
872
stack = ctx->stack;
mm/page_owner.c
875
return stack;
mm/page_owner.c
880
struct stack *stack = v;
mm/page_owner.c
883
stack = stack->next;
mm/page_owner.c
884
*ppos = stack ? *ppos + 1 : -1UL;
mm/page_owner.c
885
ctx->stack = stack;
mm/page_owner.c
887
return stack;
mm/page_owner.c
895
struct stack *stack = v;
mm/page_owner.c
898
struct stack_record *stack_record = stack->stack_record;
mm/page_owner.c
901
if (!stack->stack_record)
mm/usercopy.c
39
const void * const stack = task_stack_page(current);
mm/usercopy.c
40
const void * const stackend = stack + THREAD_SIZE;
mm/usercopy.c
44
if (obj + len <= stack || stackend <= obj)
mm/usercopy.c
52
if (obj < stack || stackend < obj + len)
mm/usercopy.c
56
ret = arch_within_stack_frames(stack, stackend, obj, len);
net/core/dev.c
743
static struct net_device_path *dev_fwd_path(struct net_device_path_stack *stack)
net/core/dev.c
745
int k = stack->num_paths++;
net/core/dev.c
750
return &stack->path[k];
net/core/dev.c
754
struct net_device_path_stack *stack)
net/core/dev.c
764
stack->num_paths = 0;
net/core/dev.c
767
path = dev_fwd_path(stack);
net/core/dev.c
783
path = dev_fwd_path(stack);
net/llc/llc_sap.c
367
struct sock **stack, int count)
net/llc/llc_sap.c
375
sock_put(stack[i]);
net/llc/llc_sap.c
379
llc_sap_rcv(sap, skb1, stack[i]);
net/llc/llc_sap.c
380
sock_put(stack[i]);
net/llc/llc_sap.c
399
struct sock *stack[256 / sizeof(struct sock *)];
net/llc/llc_sap.c
412
if (i < ARRAY_SIZE(stack))
net/llc/llc_sap.c
413
stack[i++] = sk;
net/llc/llc_sap.c
415
llc_do_mcast(sap, skb, stack, i);
net/llc/llc_sap.c
421
llc_do_mcast(sap, skb, stack, i);
net/netfilter/nf_flow_table_path.c
101
for (i = 0; i < stack->num_paths; i++) {
net/netfilter/nf_flow_table_path.c
102
path = &stack->path[i];
net/netfilter/nf_flow_table_path.c
116
i = stack->num_paths;
net/netfilter/nf_flow_table_path.c
246
struct net_device_path_stack stack;
net/netfilter/nf_flow_table_path.c
251
if (nft_dev_fill_forward_path(route, dst, ct, dir, ha, &stack) >= 0)
net/netfilter/nf_flow_table_path.c
252
nft_dev_path_info(&stack, &info, ha, &ft->data);
net/netfilter/nf_flow_table_path.c
49
struct net_device_path_stack *stack)
net/netfilter/nf_flow_table_path.c
73
return dev_fill_forward_path(dev, ha, stack);
net/netfilter/nf_flow_table_path.c
92
static void nft_dev_path_info(const struct net_device_path_stack *stack,
net/openvswitch/actions.c
165
struct mpls_shim_hdr *stack;
net/openvswitch/actions.c
172
stack = mpls_hdr(skb);
net/openvswitch/actions.c
173
lse = OVS_MASKED(stack->label_stack_entry, *mpls_lse, *mask);
net/sched/cls_u32.c
121
} stack[TC_U32_MAXDEPTH];
net/sched/cls_u32.c
206
stack[sdepth].knode = n;
net/sched/cls_u32.c
207
stack[sdepth].off = off;
net/sched/cls_u32.c
252
n = stack[sdepth].knode;
net/sched/cls_u32.c
254
off = stack[sdepth].off;
net/sched/ematch.c
505
int stack[CONFIG_NET_EMATCH_STACK];
net/sched/ematch.c
515
stack[stackp++] = match_idx;
net/sched/ematch.c
530
match_idx = stack[--stackp];
samples/pidfd/pidfd-metadata.c
37
char *stack[1024] = { 0 };
samples/pidfd/pidfd-metadata.c
40
return __clone2(do_child, stack, stack_size, flags | SIGCHLD, NULL, pidfd);
samples/pidfd/pidfd-metadata.c
42
return clone(do_child, stack + stack_size, flags | SIGCHLD, NULL, pidfd);
scripts/kconfig/symbol.c
1089
static void dep_stack_insert(struct dep_stack *stack, struct symbol *sym)
scripts/kconfig/symbol.c
1091
memset(stack, 0, sizeof(*stack));
scripts/kconfig/symbol.c
1093
check_top->next = stack;
scripts/kconfig/symbol.c
1094
stack->prev = check_top;
scripts/kconfig/symbol.c
1095
stack->sym = sym;
scripts/kconfig/symbol.c
1096
check_top = stack;
scripts/kconfig/symbol.c
1113
struct dep_stack *stack;
scripts/kconfig/symbol.c
1125
for (stack = check_top; stack != NULL; stack = stack->prev)
scripts/kconfig/symbol.c
1126
if (stack->sym == last_sym)
scripts/kconfig/symbol.c
1128
if (!stack) {
scripts/kconfig/symbol.c
1133
for (; stack; stack = stack->next) {
scripts/kconfig/symbol.c
1134
sym = stack->sym;
scripts/kconfig/symbol.c
1135
next_sym = stack->next ? stack->next->sym : last_sym;
scripts/kconfig/symbol.c
1136
type = stack->prop ? stack->prop->type : P_UNKNOWN;
scripts/kconfig/symbol.c
1138
if (stack->sym == last_sym)
scripts/kconfig/symbol.c
1147
} else if (stack->expr == &sym->dir_dep.expr) {
scripts/kconfig/symbol.c
1151
} else if (stack->expr == &sym->rev_dep.expr) {
scripts/kconfig/symbol.c
1154
} else if (stack->expr == &sym->implied.expr) {
scripts/kconfig/symbol.c
1157
} else if (stack->expr) {
scripts/kconfig/symbol.c
1218
struct dep_stack stack;
scripts/kconfig/symbol.c
1220
dep_stack_insert(&stack, sym);
scripts/kconfig/symbol.c
1222
stack.expr = &sym->dir_dep.expr;
scripts/kconfig/symbol.c
1227
stack.expr = &sym->rev_dep.expr;
scripts/kconfig/symbol.c
1232
stack.expr = &sym->implied.expr;
scripts/kconfig/symbol.c
1237
stack.expr = NULL;
scripts/kconfig/symbol.c
1242
stack.prop = prop;
scripts/kconfig/symbol.c
1248
stack.expr = &prop->expr;
scripts/kconfig/symbol.c
1252
stack.expr = NULL;
scripts/kconfig/symbol.c
1265
struct dep_stack stack;
scripts/kconfig/symbol.c
1267
dep_stack_insert(&stack, choice);
scripts/kconfig/symbol.c
1321
struct dep_stack stack;
scripts/kconfig/symbol.c
1324
dep_stack_insert(&stack, sym);
security/apparmor/domain.c
129
struct aa_label *label, bool stack,
security/apparmor/domain.c
1355
struct aa_label *target, bool stack,
security/apparmor/domain.c
1363
error = change_profile_perms(profile, target, stack, request,
security/apparmor/domain.c
1397
bool stack = flags & AA_CHANGE_STACK;
security/apparmor/domain.c
142
state = match_component(profile, tp, stack, state);
security/apparmor/domain.c
1424
if (stack)
security/apparmor/domain.c
1430
if (stack)
security/apparmor/domain.c
1439
if (!stack && unconfined(label) &&
security/apparmor/domain.c
1449
stack = true;
security/apparmor/domain.c
1459
stack = true;
security/apparmor/domain.c
1500
profile, target, stack,
security/apparmor/domain.c
1526
if (!stack) {
security/apparmor/domain.c
1546
if (stack)
security/apparmor/domain.c
1566
aa_set_current_onexec(target, stack);
security/apparmor/domain.c
191
struct aa_label *label, bool stack,
security/apparmor/domain.c
206
state = match_component(profile, tp, stack, start);
security/apparmor/domain.c
223
state = match_component(profile, tp, stack, start);
security/apparmor/domain.c
255
bool stack, aa_state_t state, bool inview, u32 request,
security/apparmor/domain.c
261
error = label_compound_match(profile, label, stack, state, inview,
security/apparmor/domain.c
267
return label_components_match(profile, label, stack, state, inview,
security/apparmor/domain.c
289
struct aa_label *target, bool stack,
security/apparmor/domain.c
300
return label_match(profile, target, stack, start, true, request, perms);
security/apparmor/domain.c
575
struct aa_label *stack = NULL;
security/apparmor/domain.c
594
stack = new;
security/apparmor/domain.c
645
if (new && stack) {
security/apparmor/domain.c
649
new = aa_label_merge(base, stack, GFP_KERNEL);
security/apparmor/domain.c
654
aa_put_label(stack);
security/apparmor/domain.c
794
bool stack, const struct linux_binprm *bprm,
security/apparmor/domain.c
842
error = change_profile_perms(profile, onexec, stack, AA_MAY_ONEXEC,
security/apparmor/domain.c
869
struct aa_label *onexec, bool stack,
security/apparmor/domain.c
887
profile_onexec(subj_cred, profile, onexec, stack,
security/apparmor/domain.c
893
stack ? aa_label_merge(&profile->label, onexec,
security/apparmor/domain.c
94
bool stack, aa_state_t state)
security/apparmor/domain.c
99
if (stack)
security/apparmor/include/task.h
33
void aa_set_current_onexec(struct aa_label *label, bool stack);
security/apparmor/label.c
1864
int i, len, stack = 0, error;
security/apparmor/label.c
1879
stack = base->size;
security/apparmor/label.c
1880
len += stack;
security/apparmor/label.c
1889
for (i = 0; i < stack; i++)
security/apparmor/label.c
1892
for (split = aa_label_strn_split(str, end - str), i = stack;
security/apparmor/task.c
105
ctx->token = stack;
security/apparmor/task.c
98
void aa_set_current_onexec(struct aa_label *label, bool stack)
security/keys/keyring.c
670
} stack[KEYRING_SEARCH_MAX_DEPTH];
security/keys/keyring.c
801
stack[sp].keyring = keyring;
security/keys/keyring.c
802
stack[sp].node = node;
security/keys/keyring.c
803
stack[sp].slot = slot;
security/keys/keyring.c
848
keyring = stack[sp].keyring;
security/keys/keyring.c
849
node = stack[sp].node;
security/keys/keyring.c
850
slot = stack[sp].slot + 1;
security/keys/keyring.c
862
stack[--sp].keyring->last_used_at = ctx->now;
sound/soc/qcom/qdsp6/audioreach.c
333
cfg->stack.stack_size = cont->stack_size;
sound/soc/qcom/qdsp6/audioreach.c
51
struct apm_cont_prop_id_stack_size stack;
sound/soc/sof/amd/acp-common.c
114
amd_get_registers(sdev, &xoops, &panic_info, stack, AMD_STACK_DUMP_SIZE);
sound/soc/sof/amd/acp-common.c
118
&panic_info, stack, AMD_STACK_DUMP_SIZE);
sound/soc/sof/amd/acp-common.c
62
u32 *stack, size_t stack_words)
sound/soc/sof/amd/acp-common.c
81
acp_mailbox_read(sdev, offset, stack, stack_words * sizeof(u32));
sound/soc/sof/amd/acp-common.c
94
u32 stack[AMD_STACK_DUMP_SIZE];
sound/soc/sof/core.c
126
void *stack, size_t stack_words)
sound/soc/sof/core.c
158
sof_stack(sdev, level, oops, stack, stack_words);
sound/soc/sof/imx/imx-common.c
31
u32 *stack, size_t stack_words)
sound/soc/sof/imx/imx-common.c
49
sof_mailbox_read(sdev, offset, stack, stack_words * sizeof(u32));
sound/soc/sof/imx/imx-common.c
62
u32 stack[IMX8_STACK_DUMP_SIZE];
sound/soc/sof/imx/imx-common.c
73
imx8_get_registers(sdev, &xoops, &panic_info, stack,
sound/soc/sof/imx/imx-common.c
78
&panic_info, stack, IMX8_STACK_DUMP_SIZE);
sound/soc/sof/imx/imx-common.h
163
u32 *stack, size_t stack_words);
sound/soc/sof/intel/atom.c
38
u32 *stack, size_t stack_words)
sound/soc/sof/intel/atom.c
58
sof_mailbox_read(sdev, offset, stack, stack_words * sizeof(u32));
sound/soc/sof/intel/atom.c
65
u32 stack[STACK_DUMP_SIZE];
sound/soc/sof/intel/atom.c
71
atom_get_registers(sdev, &xoops, &panic_info, stack,
sound/soc/sof/intel/atom.c
74
&panic_info, stack, STACK_DUMP_SIZE);
sound/soc/sof/intel/bdw.c
226
u32 *stack, size_t stack_words)
sound/soc/sof/intel/bdw.c
246
sof_mailbox_read(sdev, offset, stack, stack_words * sizeof(u32));
sound/soc/sof/intel/bdw.c
253
u32 stack[BDW_STACK_DUMP_SIZE];
sound/soc/sof/intel/bdw.c
259
bdw_get_registers(sdev, &xoops, &panic_info, stack,
sound/soc/sof/intel/bdw.c
262
&panic_info, stack, BDW_STACK_DUMP_SIZE);
sound/soc/sof/intel/hda-dsp.c
1567
u32 *stack, size_t stack_words)
sound/soc/sof/intel/hda-dsp.c
1588
sof_block_read(sdev, sdev->mmio_bar, offset, stack,
sound/soc/sof/intel/hda-dsp.c
1617
u32 stack[HDA_DSP_STACK_DUMP_SIZE];
sound/soc/sof/intel/hda-dsp.c
1627
hda_dsp_get_registers(sdev, &xoops, &panic_info, stack,
sound/soc/sof/intel/hda-dsp.c
1630
&panic_info, stack, HDA_DSP_STACK_DUMP_SIZE);
sound/soc/sof/mediatek/mtk-adsp-common.c
35
u32 *stack, size_t stack_words)
sound/soc/sof/mediatek/mtk-adsp-common.c
53
sof_mailbox_read(sdev, offset, stack, stack_words * sizeof(u32));
sound/soc/sof/mediatek/mtk-adsp-common.c
67
u32 stack[MTK_ADSP_STACK_DUMP_SIZE];
sound/soc/sof/mediatek/mtk-adsp-common.c
78
mtk_adsp_get_registers(sdev, &xoops, &panic_info, stack,
sound/soc/sof/mediatek/mtk-adsp-common.c
83
stack, MTK_ADSP_STACK_DUMP_SIZE);
sound/soc/sof/sof-priv.h
365
u32 *stack, u32 stack_words);
sound/soc/sof/sof-priv.h
779
void *stack, size_t stack_words);
sound/soc/sof/sof-priv.h
796
void *oops, u32 *stack, u32 stack_words)
sound/soc/sof/sof-priv.h
798
sof_dsp_arch_ops(sdev)->dsp_stack(sdev, level, oops, stack,
sound/soc/sof/xtensa/core.c
116
u32 *stack, u32 stack_words)
sound/soc/sof/xtensa/core.c
131
hex_dump_to_buffer(stack + i, 16, 16, 4,
tools/perf/arch/arm/tests/dwarf-unwind.c
16
struct stack_dump *stack = &sample->user_stack;
tools/perf/arch/arm/tests/dwarf-unwind.c
40
stack->data = (char *) buf;
tools/perf/arch/arm/tests/dwarf-unwind.c
41
stack->size = stack_size;
tools/perf/arch/arm64/tests/dwarf-unwind.c
16
struct stack_dump *stack = &sample->user_stack;
tools/perf/arch/arm64/tests/dwarf-unwind.c
40
stack->data = (char *) buf;
tools/perf/arch/arm64/tests/dwarf-unwind.c
41
stack->size = stack_size;
tools/perf/arch/powerpc/tests/dwarf-unwind.c
16
struct stack_dump *stack = &sample->user_stack;
tools/perf/arch/powerpc/tests/dwarf-unwind.c
40
stack->data = (char *) buf;
tools/perf/arch/powerpc/tests/dwarf-unwind.c
41
stack->size = stack_size;
tools/perf/arch/x86/tests/dwarf-unwind.c
16
struct stack_dump *stack = &sample->user_stack;
tools/perf/arch/x86/tests/dwarf-unwind.c
48
stack->data = (char *) buf;
tools/perf/arch/x86/tests/dwarf-unwind.c
49
stack->size = stack_size;
tools/perf/builtin-lock.c
1742
int broken = fails->task + fails->stack + fails->time + fails->data;
tools/perf/builtin-lock.c
1757
fprintf(lock_output, " %10s: %d\n", "stack", fails->stack);
tools/perf/builtin-lock.c
1767
bad = fails->task + fails->stack + fails->time + fails->data;
tools/perf/builtin-lock.c
1777
fprintf(lock_output, "%s bad_%s=%d", sep, "stack", fails->stack);
tools/perf/trace/beauty/include/uapi/linux/sched.h
98
__aligned_u64 stack;
tools/perf/util/annotate-arch/annotate-x86.c
372
struct type_state_stack *stack;
tools/perf/util/annotate-arch/annotate-x86.c
375
stack = find_stack_state(state, offset);
tools/perf/util/annotate-arch/annotate-x86.c
376
if (!stack)
tools/perf/util/annotate-arch/annotate-x86.c
379
tsr->type = stack->type;
tools/perf/util/annotate-arch/annotate-x86.c
381
tsr->offset = offset - stack->offset;
tools/perf/util/annotate-arch/annotate-x86.c
558
struct type_state_stack *stack;
tools/perf/util/annotate-arch/annotate-x86.c
561
stack = find_stack_state(state, offset);
tools/perf/util/annotate-arch/annotate-x86.c
562
if (stack == NULL) {
tools/perf/util/annotate-arch/annotate-x86.c
565
} else if (!stack->compound) {
tools/perf/util/annotate-arch/annotate-x86.c
566
tsr->type = stack->type;
tools/perf/util/annotate-arch/annotate-x86.c
567
tsr->kind = stack->kind;
tools/perf/util/annotate-arch/annotate-x86.c
568
tsr->offset = stack->ptr_offset;
tools/perf/util/annotate-arch/annotate-x86.c
570
} else if (die_get_member_type(&stack->type,
tools/perf/util/annotate-arch/annotate-x86.c
571
offset - stack->offset,
tools/perf/util/annotate-arch/annotate-x86.c
735
struct type_state_stack *stack;
tools/perf/util/annotate-arch/annotate-x86.c
740
stack = find_stack_state(state, offset);
tools/perf/util/annotate-arch/annotate-x86.c
741
if (stack) {
tools/perf/util/annotate-arch/annotate-x86.c
749
if (!stack->compound)
tools/perf/util/annotate-arch/annotate-x86.c
750
set_stack_state(stack, offset, tsr->kind,
tools/perf/util/annotate-data.c
1235
struct type_state_stack *stack;
tools/perf/util/annotate-data.c
1239
stack = find_stack_state(state, dloc->type_offset);
tools/perf/util/annotate-data.c
1240
if (stack == NULL) {
tools/perf/util/annotate-data.c
1252
if (stack->kind == TSR_KIND_CANARY) {
tools/perf/util/annotate-data.c
1257
if (stack->kind != TSR_KIND_TYPE)
tools/perf/util/annotate-data.c
1260
*type_die = stack->type;
tools/perf/util/annotate-data.c
1262
dloc->type_offset -= stack->offset;
tools/perf/util/annotate-data.c
1268
struct type_state_stack *stack;
tools/perf/util/annotate-data.c
1280
stack = find_stack_state(state, dloc->type_offset - fboff);
tools/perf/util/annotate-data.c
1281
if (stack == NULL) {
tools/perf/util/annotate-data.c
1293
if (stack->kind == TSR_KIND_CANARY) {
tools/perf/util/annotate-data.c
1298
if (stack->kind != TSR_KIND_TYPE)
tools/perf/util/annotate-data.c
1301
*type_die = stack->type;
tools/perf/util/annotate-data.c
1303
dloc->type_offset -= fboff + stack->offset;
tools/perf/util/annotate-data.c
185
struct type_state_stack *stack, *tmp;
tools/perf/util/annotate-data.c
187
list_for_each_entry_safe(stack, tmp, &state->stack_vars, list) {
tools/perf/util/annotate-data.c
188
list_del(&stack->list);
tools/perf/util/annotate-data.c
189
free(stack);
tools/perf/util/annotate-data.c
566
struct type_state_stack *stack;
tools/perf/util/annotate-data.c
568
list_for_each_entry(stack, &state->stack_vars, list) {
tools/perf/util/annotate-data.c
569
if (offset == stack->offset)
tools/perf/util/annotate-data.c
570
return stack;
tools/perf/util/annotate-data.c
572
if (stack->compound && stack->offset < offset &&
tools/perf/util/annotate-data.c
573
offset < stack->offset + stack->size)
tools/perf/util/annotate-data.c
574
return stack;
tools/perf/util/annotate-data.c
579
void set_stack_state(struct type_state_stack *stack, int offset, u8 kind,
tools/perf/util/annotate-data.c
592
stack->type = *type_die;
tools/perf/util/annotate-data.c
593
stack->size = size;
tools/perf/util/annotate-data.c
594
stack->offset = offset;
tools/perf/util/annotate-data.c
595
stack->ptr_offset = ptr_offset;
tools/perf/util/annotate-data.c
596
stack->kind = kind;
tools/perf/util/annotate-data.c
599
stack->compound = false;
tools/perf/util/annotate-data.c
608
stack->compound = (kind != TSR_KIND_PERCPU_POINTER);
tools/perf/util/annotate-data.c
611
stack->compound = false;
tools/perf/util/annotate-data.c
621
struct type_state_stack *stack = find_stack_state(state, offset);
tools/perf/util/annotate-data.c
623
if (stack) {
tools/perf/util/annotate-data.c
624
set_stack_state(stack, offset, kind, type_die, ptr_offset);
tools/perf/util/annotate-data.c
625
return stack;
tools/perf/util/annotate-data.c
628
stack = malloc(sizeof(*stack));
tools/perf/util/annotate-data.c
629
if (stack) {
tools/perf/util/annotate-data.c
630
set_stack_state(stack, offset, kind, type_die, ptr_offset);
tools/perf/util/annotate-data.c
631
list_add(&stack->list, &state->stack_vars);
tools/perf/util/annotate-data.c
633
return stack;
tools/perf/util/annotate-data.c
884
struct type_state_stack *stack;
tools/perf/util/annotate-data.c
894
stack = find_stack_state(state, offset);
tools/perf/util/annotate-data.c
895
if (stack && stack->kind == TSR_KIND_TYPE &&
tools/perf/util/annotate-data.c
896
!is_better_type(&stack->type, &mem_die))
tools/perf/util/annotate-data.h
254
void set_stack_state(struct type_state_stack *stack, int offset, u8 kind,
tools/perf/util/arm64-frame-pointer-unwind-support.c
13
u64 stack[2];
tools/perf/util/arm64-frame-pointer-unwind-support.c
34
entries->stack[entries->length++] = entry->ip;
tools/perf/util/arm64-frame-pointer-unwind-support.c
70
return callchain_param.order == ORDER_CALLER ? entries.stack[0] : entries.stack[1];
tools/perf/util/bpf_lock_contention.c
720
int fd, stack, err = 0;
tools/perf/util/bpf_lock_contention.c
729
stack = bpf_map__fd(skel->maps.stacks);
tools/perf/util/bpf_lock_contention.c
732
con->fails.stack = skel->bss->stack_fail;
tools/perf/util/bpf_lock_contention.c
768
bpf_map_lookup_elem(stack, &key.stack_id, stack_trace);
tools/perf/util/bpf_off_cpu.c
321
int fd, stack;
tools/perf/util/bpf_off_cpu.c
357
stack = bpf_map__fd(skel->maps.stacks);
tools/perf/util/bpf_off_cpu.c
396
bpf_map_lookup_elem(stack, &key.stack_id, &off_cpu_raw[i + 2]);
tools/perf/util/bpf_skel/off_cpu.bpf.c
235
struct __stack *stack, __u64 delta)
tools/perf/util/bpf_skel/off_cpu.bpf.c
245
len = copy_stack(stack, data, n);
tools/perf/util/bpf_skel/off_cpu.bpf.c
284
if (stack_id > 0 && bpf_get_stack(ctx, &pelem->stack, MAX_STACKS * sizeof(u64), BPF_F_USER_STACK)) {
tools/perf/util/bpf_skel/off_cpu.bpf.c
309
off_cpu_dump(ctx, data, &key, &pelem->stack, delta);
tools/perf/util/bpf_skel/off_cpu.bpf.c
33
struct __stack stack;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1208
err = intel_pt_push(&decoder->stack, decoder->ip +
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1214
decoder->ret_addr = intel_pt_pop(&decoder->stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
177
struct intel_pt_stack stack;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3149
intel_pt_clear_stack(&decoder->stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3754
intel_pt_clear_stack(&decoder->stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
387
static void intel_pt_pop_blk(struct intel_pt_stack *stack)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
389
struct intel_pt_blk *blk = stack->blk;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
391
stack->blk = blk->prev;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3916
intel_pt_clear_stack(&decoder->stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
392
if (!stack->spare)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
393
stack->spare = blk;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
398
static uint64_t intel_pt_pop(struct intel_pt_stack *stack)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
400
if (!stack->pos) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
401
if (!stack->blk)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
403
intel_pt_pop_blk(stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
404
if (!stack->blk)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
406
stack->pos = INTEL_PT_BLK_SIZE;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
408
return stack->blk->ip[--stack->pos];
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
411
static int intel_pt_alloc_blk(struct intel_pt_stack *stack)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
415
if (stack->spare) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
416
blk = stack->spare;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
417
stack->spare = NULL;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
424
blk->prev = stack->blk;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
425
stack->blk = blk;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
426
stack->pos = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
430
static int intel_pt_push(struct intel_pt_stack *stack, uint64_t ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
434
if (!stack->blk || stack->pos == INTEL_PT_BLK_SIZE) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
435
err = intel_pt_alloc_blk(stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
440
stack->blk->ip[stack->pos++] = ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
444
static void intel_pt_clear_stack(struct intel_pt_stack *stack)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
446
while (stack->blk)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
447
intel_pt_pop_blk(stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
448
stack->pos = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
451
static void intel_pt_free_stack(struct intel_pt_stack *stack)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
453
intel_pt_clear_stack(stack);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
454
zfree(&stack->blk);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
455
zfree(&stack->spare);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
460
intel_pt_free_stack(&decoder->stack);
tools/perf/util/lock-contention.h
138
int stack;
tools/perf/util/thread-stack.c
1008
tse = &ts->stack[ts->cnt - 1];
tools/perf/util/thread-stack.c
1027
if (!ts->cnt || (ts->cnt == 1 && ts->stack[0].ref == ref))
tools/perf/util/thread-stack.c
1030
cp = call_path__findnew(cpr, ts->stack[ts->cnt - 1].cp, NULL, 0,
tools/perf/util/thread-stack.c
1053
struct thread_stack_entry *tse = &ts->stack[ts->cnt - 1];
tools/perf/util/thread-stack.c
1071
sym = ts->stack[ts->cnt - 2].cp->sym;
tools/perf/util/thread-stack.c
1090
cp = call_path__findnew(cpr, ts->stack[ts->cnt - 2].cp, tsym,
tools/perf/util/thread-stack.c
1096
ts->stack[ts->cnt - 1].cp = cp;
tools/perf/util/thread-stack.c
1161
cp = call_path__findnew(cpr, ts->stack[ts->cnt - 1].cp,
tools/perf/util/thread-stack.c
1193
ts->stack[ts->cnt - 1].ret_addr != sample->addr)
tools/perf/util/thread-stack.c
1220
cp = call_path__findnew(cpr, ts->stack[ts->cnt - 1].cp,
tools/perf/util/thread-stack.c
1226
ts->stack[ts->cnt - 1].non_call = true;
tools/perf/util/thread-stack.c
126
new_stack = realloc(ts->stack, sz);
tools/perf/util/thread-stack.c
130
ts->stack = new_stack;
tools/perf/util/thread-stack.c
201
if (!ts->stack &&
tools/perf/util/thread-stack.c
220
if (!ts->stack)
tools/perf/util/thread-stack.c
251
ts->stack[ts->cnt].trace_end = trace_end;
tools/perf/util/thread-stack.c
252
ts->stack[ts->cnt++].ret_addr = ret_addr;
tools/perf/util/thread-stack.c
271
if (ts->stack[--i].ret_addr == ret_addr) {
tools/perf/util/thread-stack.c
283
if (ts->stack[--i].trace_end)
tools/perf/util/thread-stack.c
295
return ts->stack[ts->cnt - 1].cp->in_kernel;
tools/perf/util/thread-stack.c
311
tse = &ts->stack[idx];
tools/perf/util/thread-stack.c
490
zfree(&ts->stack);
tools/perf/util/thread-stack.c
546
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
587
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
597
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
798
tse = &ts->stack[ts->cnt++];
tools/perf/util/thread-stack.c
824
struct thread_stack_entry *tse = &ts->stack[0];
tools/perf/util/thread-stack.c
831
if (ts->stack[ts->cnt - 1].ret_addr == ret_addr &&
tools/perf/util/thread-stack.c
832
!ts->stack[ts->cnt - 1].non_call) {
tools/perf/util/thread-stack.c
839
if (ts->stack[i].ret_addr != ret_addr ||
tools/perf/util/thread-stack.c
840
ts->stack[i].non_call)
tools/perf/util/thread-stack.c
89
struct thread_stack_entry *stack;
tools/perf/util/thread-stack.c
940
parent = ts->stack[ts->cnt - 1].cp;
tools/perf/util/thread-stack.c
972
ts->stack[ts->cnt - 1].non_call = true;
tools/perf/util/unwind-libdw.c
216
struct stack_dump *stack = &ui->sample->user_stack;
tools/perf/util/unwind-libdw.c
229
end = start + stack->size;
tools/perf/util/unwind-libdw.c
247
*result = *(Dwarf_Word *)&stack->data[offset];
tools/perf/util/unwind-libunwind-local.c
575
struct stack_dump *stack = &ui->sample->user_stack;
tools/perf/util/unwind-libunwind-local.c
581
if (__write || !stack || !ui->sample->user_regs || !ui->sample->user_regs->regs) {
tools/perf/util/unwind-libunwind-local.c
593
end = start + stack->size;
tools/perf/util/unwind-libunwind-local.c
612
*valp = *(unw_word_t *)&stack->data[offset];
tools/testing/selftests/arm64/abi/tpidr2.c
146
void *stack;
tools/testing/selftests/arm64/abi/tpidr2.c
151
stack = malloc(__STACK_SIZE);
tools/testing/selftests/arm64/abi/tpidr2.c
152
if (!stack) {
tools/testing/selftests/arm64/abi/tpidr2.c
157
ret = sys_clone(CLONE_VM, (unsigned long)stack + __STACK_SIZE,
tools/testing/selftests/arm64/gcs/libc-gcs.c
303
unsigned long *stack;
tools/testing/selftests/arm64/gcs/libc-gcs.c
434
self->stack = (void *)syscall(__NR_map_shadow_stack, 0,
tools/testing/selftests/arm64/gcs/libc-gcs.c
437
ASSERT_FALSE(self->stack == MAP_FAILED);
tools/testing/selftests/arm64/gcs/libc-gcs.c
438
ksft_print_msg("Allocated stack from %p-%p\n", self->stack,
tools/testing/selftests/arm64/gcs/libc-gcs.c
439
self->stack + variant->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
446
if (self->stack != MAP_FAILED) {
tools/testing/selftests/arm64/gcs/libc-gcs.c
447
ret = munmap(self->stack, variant->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
455
unsigned long *stack = self->stack;
tools/testing/selftests/arm64/gcs/libc-gcs.c
473
ASSERT_EQ(stack[cap_index], GCS_CAP(&stack[cap_index]));
tools/testing/selftests/arm64/gcs/libc-gcs.c
479
unsigned long *stack = self->stack;
tools/testing/selftests/arm64/gcs/libc-gcs.c
487
ASSERT_EQ(stack[term_index], 0);
tools/testing/selftests/arm64/gcs/libc-gcs.c
493
self->stack[0] = 0;
tools/testing/selftests/arm64/gcs/libc-gcs.c
516
pivot_gcspr_el0 = &self->stack[cap_index];
tools/testing/selftests/arm64/gcs/libc-gcs.c
531
ASSERT_TRUE((unsigned long)get_gcspr() > (unsigned long)self->stack);
tools/testing/selftests/arm64/gcs/libc-gcs.c
533
(unsigned long)self->stack + variant->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
569
pivot_gcspr_el0 = &self->stack[cap_index];
tools/testing/selftests/arm64/gcs/libc-gcs.c
584
ASSERT_TRUE((unsigned long)get_gcspr() > (unsigned long)self->stack);
tools/testing/selftests/arm64/gcs/libc-gcs.c
586
(unsigned long)self->stack + variant->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
633
void *stack;
tools/testing/selftests/arm64/gcs/libc-gcs.c
635
stack = (void *)syscall(__NR_map_shadow_stack, 0,
tools/testing/selftests/arm64/gcs/libc-gcs.c
637
ASSERT_TRUE(stack == MAP_FAILED);
tools/testing/selftests/arm64/gcs/libc-gcs.c
638
if (stack != MAP_FAILED)
tools/testing/selftests/arm64/gcs/libc-gcs.c
639
munmap(stack, variant->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
644
unsigned long *stack;
tools/testing/selftests/arm64/gcs/libc-gcs.c
656
self->stack = (void *)syscall(__NR_map_shadow_stack, 0,
tools/testing/selftests/arm64/gcs/libc-gcs.c
658
ASSERT_FALSE(self->stack == MAP_FAILED);
tools/testing/selftests/arm64/gcs/libc-gcs.c
659
ksft_print_msg("Allocated stack from %p-%p\n", self->stack,
tools/testing/selftests/arm64/gcs/libc-gcs.c
660
self->stack + self->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
667
if (self->stack != MAP_FAILED) {
tools/testing/selftests/arm64/gcs/libc-gcs.c
668
ret = munmap(self->stack, self->stack_size);
tools/testing/selftests/arm64/gcs/libc-gcs.c
682
ret = mprotect(self->stack, self->stack_size, variant->flags);
tools/testing/selftests/arm64/gcs/libc-gcs.c
690
ret = mprotect(self->stack, self->stack_size,
tools/testing/selftests/bpf/prog_tests/build_id.c
10
static void print_stack(struct bpf_stack_build_id *stack, int frame_cnt)
tools/testing/selftests/bpf/prog_tests/build_id.c
100
ASSERT_EQ(stack[0].status, BPF_STACK_BUILD_ID_VALID, "build_id_status");
tools/testing/selftests/bpf/prog_tests/build_id.c
101
ASSERT_EQ(memcmp(stack[0].build_id, build_id, build_id_sz), 0, "build_id_match");
tools/testing/selftests/bpf/prog_tests/build_id.c
16
switch (stack[i].status) {
tools/testing/selftests/bpf/prog_tests/build_id.c
23
printf("%02hhx", (unsigned)stack[i].build_id[j]);
tools/testing/selftests/bpf/prog_tests/build_id.c
24
printf(" OFFSET = %llx", (unsigned long long)stack[i].offset);
tools/testing/selftests/bpf/prog_tests/build_id.c
27
printf("IP = %llx", (unsigned long long)stack[i].ip);
tools/testing/selftests/bpf/prog_tests/build_id.c
30
printf("UNEXPECTED STATUS %d ", stack[i].status);
tools/testing/selftests/bpf/prog_tests/build_id.c
40
struct bpf_stack_build_id *stack;
tools/testing/selftests/bpf/prog_tests/build_id.c
59
stack = skel->bss->stack_nofault;
tools/testing/selftests/bpf/prog_tests/build_id.c
62
print_stack(stack, frame_cnt);
tools/testing/selftests/bpf/prog_tests/build_id.c
65
ASSERT_EQ(stack[0].status, BPF_STACK_BUILD_ID_VALID, "build_id_status");
tools/testing/selftests/bpf/prog_tests/build_id.c
66
ASSERT_EQ(memcmp(stack[0].build_id, build_id, build_id_sz), 0, "build_id_match");
tools/testing/selftests/bpf/prog_tests/build_id.c
68
ASSERT_EQ(stack[0].status, BPF_STACK_BUILD_ID_IP, "build_id_status");
tools/testing/selftests/bpf/prog_tests/build_id.c
78
struct bpf_stack_build_id *stack;
tools/testing/selftests/bpf/prog_tests/build_id.c
95
stack = skel->bss->stack_sleepable;
tools/testing/selftests/bpf/prog_tests/build_id.c
98
print_stack(stack, frame_cnt);
tools/testing/selftests/bpf/prog_tests/stacktrace_map.c
11
__u64 stack[PERF_MAX_STACK_DEPTH];
tools/testing/selftests/bpf/prog_tests/stacktrace_map.c
53
err = bpf_map_lookup_and_delete_elem(stackmap_fd, &stack_id, stack);
tools/testing/selftests/bpf/prog_tests/stacktrace_map.c
57
err = bpf_map_lookup_elem(stackmap_fd, &stack_id, stack);
tools/testing/selftests/bpf/prog_tests/task_kfunc.c
117
char *stack;
tools/testing/selftests/bpf/prog_tests/task_kfunc.c
119
stack = (char *)malloc(stack_size);
tools/testing/selftests/bpf/prog_tests/task_kfunc.c
120
if (!ASSERT_OK_PTR(stack, "clone_stack"))
tools/testing/selftests/bpf/prog_tests/task_kfunc.c
123
child_pid = clone(run_vpid_test, stack + stack_size,
tools/testing/selftests/bpf/prog_tests/task_kfunc.c
136
free(stack);
tools/testing/selftests/bpf/prog_tests/uprobe_multi_test.c
113
child->pid = child->tid = clone(child_func, child->stack + sizeof(child->stack)/2,
tools/testing/selftests/bpf/prog_tests/uprobe_multi_test.c
54
char stack[65536];
tools/testing/selftests/bpf/progs/map_ptr_kern.c
603
struct bpf_queue_stack *stack = (struct bpf_queue_stack *)&m_stack;
tools/testing/selftests/bpf/progs/map_ptr_kern.c
606
VERIFY(check(&stack->map, map, 0, sizeof(__u32), MAX_ENTRIES));
tools/testing/selftests/bpf/progs/pyperf.h
206
event->stack[i] = *symbol_id;
tools/testing/selftests/bpf/progs/pyperf.h
326
event->stack[i] = *symbol_id;
tools/testing/selftests/bpf/progs/pyperf.h
55
int32_t stack[STACK_MAX_LEN];
tools/testing/selftests/bpf/progs/test_get_stack_rawtp_err.c
11
__u64 stack[MAX_STACK_RAWTP];
tools/testing/selftests/bpf/progs/test_get_stack_rawtp_err.c
15
error = bpf_get_stack(ctx, stack, 0, -1);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
11
static long stack[256];
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
115
return bpf_get_stack(ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
130
return bpf_get_stack(ctx, stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
142
return bpf_get_stack(ctx1, stack, sizeof(stack), 0) +
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
144
bpf_get_stack(ctx2, stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
19
return bpf_get_stack(ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
50
return bpf_get_stack((void *)ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
67
return bpf_get_stack(ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
83
return bpf_get_stack(ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/test_global_func_ctx_args.c
99
return bpf_get_stack(ctx, &stack, sizeof(stack), 0);
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
109
long stack[128];
tools/testing/selftests/bpf/progs/verifier_global_subprogs.c
176
return bpf_get_stack(ctx, stack, sizeof(stack), 0);
tools/testing/selftests/bpf/veristat.c
1000
char *state = NULL, *token, stack[512];
tools/testing/selftests/bpf/veristat.c
1026
if (1 == sscanf(cur, "stack depth %511s", stack))
tools/testing/selftests/bpf/veristat.c
1029
while ((token = strtok_r(cnt++ ? NULL : stack, "+", &state))) {
tools/testing/selftests/cgroup/test_core.c
797
static char stack[65536];
tools/testing/selftests/cgroup/test_core.c
833
pid = clone(lesser_ns_open_thread_fn, stack + sizeof(stack),
tools/testing/selftests/cgroup/test_kmem.c
251
long stack;
tools/testing/selftests/cgroup/test_kmem.c
266
stack = cg_read_key_long(cgroup, "memory.stat", "kernel_stack ");
tools/testing/selftests/cgroup/test_kmem.c
267
if (stack >= 4096 * 1000)
tools/testing/selftests/clone3/clone3_selftests.h
28
__aligned_u64 stack;
tools/testing/selftests/coredump/stackdump_test.c
107
unsigned long long stack;
tools/testing/selftests/coredump/stackdump_test.c
159
stack = strtoull(line, NULL, 10);
tools/testing/selftests/coredump/stackdump_test.c
160
ASSERT_NE(stack, 0);
tools/testing/selftests/filesystems/nsfs/pidns.c
29
char stack[128] __stack_aligned__;
tools/testing/selftests/filesystems/utils.c
141
void *stack;
tools/testing/selftests/filesystems/utils.c
143
stack = malloc(__STACK_SIZE);
tools/testing/selftests/filesystems/utils.c
144
if (!stack)
tools/testing/selftests/filesystems/utils.c
148
return __clone2(fn, stack, __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/filesystems/utils.c
150
return clone(fn, stack + __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/futex/functional/robust_list.c
71
char *stack;
tools/testing/selftests/futex/functional/robust_list.c
74
stack = mmap(NULL, STACK_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/futex/functional/robust_list.c
76
if (stack == MAP_FAILED)
tools/testing/selftests/futex/functional/robust_list.c
79
stack += STACK_SIZE;
tools/testing/selftests/futex/functional/robust_list.c
81
pid = clone(fn, stack, CLONE_VM | SIGCHLD, arg);
tools/testing/selftests/kvm/lib/assert.c
28
void *stack[n];
tools/testing/selftests/kvm/lib/assert.c
38
n = backtrace(stack, n);
tools/testing/selftests/kvm/lib/assert.c
52
c += sprintf(c, " %lx", ((unsigned long) stack[i]) - 1);
tools/testing/selftests/memfd/fuse_test.c
202
uint8_t *stack;
tools/testing/selftests/memfd/fuse_test.c
205
stack = malloc(STACK_SIZE);
tools/testing/selftests/memfd/fuse_test.c
206
if (!stack) {
tools/testing/selftests/memfd/fuse_test.c
212
stack + STACK_SIZE,
tools/testing/selftests/memfd/memfd_test.c
798
uint8_t *stack;
tools/testing/selftests/memfd/memfd_test.c
801
stack = malloc(STACK_SIZE);
tools/testing/selftests/memfd/memfd_test.c
802
if (!stack) {
tools/testing/selftests/memfd/memfd_test.c
807
pid = clone(fn, stack + STACK_SIZE, SIGCHLD | flags, arg);
tools/testing/selftests/mm/pkey_sighandler_tests.c
179
stack_t *stack = ptr;
tools/testing/selftests/mm/pkey_sighandler_tests.c
187
syscall_raw(SYS_sigaltstack, (long)stack, 0, 0, 0, 0, 0);
tools/testing/selftests/mm/pkey_sighandler_tests.c
286
void *stack;
tools/testing/selftests/mm/pkey_sighandler_tests.c
302
stack = mmap(0, STACK_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/pkey_sighandler_tests.c
305
assert(stack != MAP_FAILED);
tools/testing/selftests/mm/pkey_sighandler_tests.c
315
sys_mprotect_pkey(stack, STACK_SIZE, PROT_READ | PROT_WRITE, pkey);
tools/testing/selftests/mm/pkey_sighandler_tests.c
330
stack + STACK_SIZE,
tools/testing/selftests/mm/pkey_sighandler_tests.c
405
stack_t *stack = ptr;
tools/testing/selftests/mm/pkey_sighandler_tests.c
413
syscall(SYS_sigaltstack, (long)stack, 0, 0, 0, 0, 0);
tools/testing/selftests/mm/pkey_sighandler_tests.c
442
void *stack;
tools/testing/selftests/mm/pkey_sighandler_tests.c
470
stack = mmap(0, STACK_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/pkey_sighandler_tests.c
473
assert(stack != MAP_FAILED);
tools/testing/selftests/mm/pkey_sighandler_tests.c
488
sys_mprotect_pkey(stack, STACK_SIZE, PROT_READ | PROT_WRITE, pkey);
tools/testing/selftests/mm/pkey_sighandler_tests.c
501
stack + STACK_SIZE,
tools/testing/selftests/mm/pkey_sighandler_tests.c
80
static inline long clone_raw(unsigned long flags, void *stack,
tools/testing/selftests/mm/pkey_sighandler_tests.c
84
long a2 = (long)stack;
tools/testing/selftests/mount_setattr/mount_setattr_test.c
1168
void *stack;
tools/testing/selftests/mount_setattr/mount_setattr_test.c
1170
stack = malloc(__STACK_SIZE);
tools/testing/selftests/mount_setattr/mount_setattr_test.c
1171
if (!stack)
tools/testing/selftests/mount_setattr/mount_setattr_test.c
1175
return __clone2(fn, stack, __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/mount_setattr/mount_setattr_test.c
1177
return clone(fn, stack + __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/move_mount_set_group/move_mount_set_group_test.c
270
void *stack;
tools/testing/selftests/move_mount_set_group/move_mount_set_group_test.c
272
stack = malloc(__STACK_SIZE);
tools/testing/selftests/move_mount_set_group/move_mount_set_group_test.c
273
if (!stack)
tools/testing/selftests/move_mount_set_group/move_mount_set_group_test.c
277
return __clone2(fn, stack, __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/move_mount_set_group/move_mount_set_group_test.c
279
return clone(fn, stack + __STACK_SIZE, flags | SIGCHLD, arg, NULL);
tools/testing/selftests/mseal_system_mappings/sysmap_is_sealed.c
103
FIXTURE_VARIANT_ADD(basic, stack) {
tools/testing/selftests/pid_namespace/pid_max.c
22
char *stack;
tools/testing/selftests/pid_namespace/pid_max.c
25
stack = malloc(__STACK_SIZE);
tools/testing/selftests/pid_namespace/pid_max.c
26
if (!stack)
tools/testing/selftests/pid_namespace/pid_max.c
30
ret = __clone2(fn, stack, __STACK_SIZE, flags | SIGCHLD, arg);
tools/testing/selftests/pid_namespace/pid_max.c
32
ret = clone(fn, stack + __STACK_SIZE, flags | SIGCHLD, arg);
tools/testing/selftests/pid_namespace/pid_max.c
34
free(stack);
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
102
ret.stack = mmap(NULL, CHILD_STACK_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
104
if (ret.stack == MAP_FAILED) {
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
110
ret.pid = __clone2(fn, ret.stack, CHILD_STACK_SIZE, flags, args, &ret.fd);
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
112
ret.pid = clone(fn, ret.stack + CHILD_STACK_SIZE, flags, args, &ret.fd);
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
142
if (munmap(child->stack, CHILD_STACK_SIZE)) {
tools/testing/selftests/pidfd/pidfd_fdinfo_test.c
88
char *stack;
tools/testing/selftests/pidfd/pidfd_test.c
36
char *stack[1024] = { 0 };
tools/testing/selftests/pidfd/pidfd_test.c
39
return __clone2(fn, stack, stack_size, flags | SIGCHLD, NULL, pidfd);
tools/testing/selftests/pidfd/pidfd_test.c
41
return clone(fn, stack + stack_size, flags | SIGCHLD, NULL, pidfd);
tools/testing/selftests/proc/thread-self.c
51
void *stack;
tools/testing/selftests/proc/thread-self.c
56
stack = mmap(NULL, 2 * PAGE_SIZE, PROT_READ|PROT_WRITE, MAP_PRIVATE|MAP_ANONYMOUS, -1, 0);
tools/testing/selftests/proc/thread-self.c
57
assert(stack != MAP_FAILED);
tools/testing/selftests/proc/thread-self.c
59
pid = clone(f, stack + PAGE_SIZE, CLONE_THREAD|CLONE_SIGHAND|CLONE_VM, (void *)1);
tools/testing/selftests/user_events/abi_test.c
401
void *stack = mmap(NULL, stack_size, PROT_READ | PROT_WRITE,
tools/testing/selftests/user_events/abi_test.c
405
ASSERT_NE(MAP_FAILED, stack);
tools/testing/selftests/user_events/abi_test.c
410
ASSERT_NE(-1, clone(&clone_check, stack + stack_size,
tools/testing/selftests/user_events/abi_test.c
416
munmap(stack, stack_size);
tools/testing/selftests/x86/lam.c
856
char stack[STACK_SIZE];
tools/testing/selftests/x86/lam.c
868
pid = clone(thread_fn_get_lam, stack + STACK_SIZE,
tools/testing/selftests/x86/lam.c
891
char stack[STACK_SIZE];
tools/testing/selftests/x86/lam.c
896
pid = clone(thread_fn_set_lam, stack + STACK_SIZE,
tools/testing/selftests/x86/mov_ss_trap.c
239
stack_t stack = {
tools/testing/selftests/x86/mov_ss_trap.c
243
if (sigaltstack(&stack, NULL) != 0)
tools/testing/selftests/x86/mov_ss_trap.c
247
free(stack.ss_sp);
tools/testing/selftests/x86/sigreturn.c
750
stack_t stack = {
tools/testing/selftests/x86/sigreturn.c
755
if (sigaltstack(&stack, NULL) != 0)
tools/testing/selftests/x86/sigreturn.c
853
free(stack.ss_sp);
tools/testing/selftests/x86/single_step_syscall.c
189
stack_t stack = {
tools/testing/selftests/x86/single_step_syscall.c
193
if (sigaltstack(&stack, NULL) != 0)
tools/testing/selftests/x86/single_step_syscall.c
199
free(stack.ss_sp);
tools/testing/selftests/x86/syscall_arg_fault.c
221
free(stack.ss_sp);
tools/testing/selftests/x86/syscall_arg_fault.c
90
stack_t stack = {
tools/testing/selftests/x86/syscall_arg_fault.c
95
if (sigaltstack(&stack, NULL) != 0)
tools/testing/selftests/x86/sysret_ss_attrs.c
36
extern unsigned long call32_from_64(void *stack, void (*function)(void));