arch/alpha/include/asm/elf.h
110
struct thread_info;
arch/alpha/include/asm/elf.h
113
struct thread_info *ti);
arch/alpha/include/asm/thread_info.h
42
register struct thread_info *__current_thread_info __asm__("$8");
arch/alpha/kernel/asm-offsets.c
18
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/alpha/kernel/asm-offsets.c
19
DEFINE(TI_FP, offsetof(struct thread_info, fp));
arch/alpha/kernel/asm-offsets.c
20
DEFINE(TI_STATUS, offsetof(struct thread_info, status));
arch/alpha/kernel/process.c
240
struct thread_info *childti = task_thread_info(p);
arch/alpha/kernel/process.c
286
dump_elf_thread(elf_greg_t *dest, struct pt_regs *pt, struct thread_info *ti)
arch/alpha/kernel/ptrace.c
81
#define FP_REG(reg) (offsetof(struct thread_info, reg))
arch/arc/include/asm/thread_info.h
59
static inline __attribute_const__ struct thread_info *current_thread_info(void)
arch/arc/include/asm/thread_info.h
62
return (struct thread_info *)(sp & ~(THREAD_SIZE - 1));
arch/arc/kernel/asm-offsets.c
30
DEFINE(THREAD_INFO_KSP, offsetof(struct thread_info, ksp));
arch/arc/kernel/asm-offsets.c
31
DEFINE(THREAD_INFO_FLAGS, offsetof(struct thread_info, flags));
arch/arc/kernel/asm-offsets.c
33
offsetof(struct thread_info, preempt_count));
arch/arm/include/asm/switch_to.h
25
extern struct task_struct *__switch_to(struct task_struct *, struct thread_info *, struct thread_info *);
arch/arm/include/asm/thread_info.h
102
extern void iwmmxt_task_disable(struct thread_info *);
arch/arm/include/asm/thread_info.h
103
extern void iwmmxt_task_copy(struct thread_info *, void *);
arch/arm/include/asm/thread_info.h
104
extern void iwmmxt_task_restore(struct thread_info *, void *);
arch/arm/include/asm/thread_info.h
105
extern void iwmmxt_task_release(struct thread_info *);
arch/arm/include/asm/thread_info.h
106
extern void iwmmxt_task_switch(struct thread_info *);
arch/arm/include/asm/thread_info.h
123
extern void vfp_sync_hwstate(struct thread_info *);
arch/arm/include/asm/thread_info.h
124
extern void vfp_flush_hwstate(struct thread_info *);
arch/arm/include/asm/thread_info.h
84
static inline struct task_struct *thread_task(struct thread_info* ti)
arch/arm/include/asm/thread_notify.h
29
static inline void thread_notify(unsigned long rc, struct thread_info *thread)
arch/arm/include/asm/tls.h
75
struct thread_info *thread;
arch/arm/kernel/asm-offsets.c
48
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/arm/kernel/asm-offsets.c
49
DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/arm/kernel/asm-offsets.c
50
DEFINE(TI_CPU, offsetof(struct thread_info, cpu));
arch/arm/kernel/asm-offsets.c
51
DEFINE(TI_CPU_DOMAIN, offsetof(struct thread_info, cpu_domain));
arch/arm/kernel/asm-offsets.c
52
DEFINE(TI_CPU_SAVE, offsetof(struct thread_info, cpu_context));
arch/arm/kernel/asm-offsets.c
53
DEFINE(TI_ABI_SYSCALL, offsetof(struct thread_info, abi_syscall));
arch/arm/kernel/asm-offsets.c
54
DEFINE(TI_TP_VALUE, offsetof(struct thread_info, tp_value));
arch/arm/kernel/asm-offsets.c
55
DEFINE(TI_FPSTATE, offsetof(struct thread_info, fpstate));
arch/arm/kernel/asm-offsets.c
57
DEFINE(TI_VFPSTATE, offsetof(struct thread_info, vfpstate));
arch/arm/kernel/asm-offsets.c
64
DEFINE(TI_THUMBEE_STATE, offsetof(struct thread_info, thumbee_state));
arch/arm/kernel/asm-offsets.c
67
DEFINE(TI_IWMMXT_STATE, offsetof(struct thread_info, fpstate.iwmmxt));
arch/arm/kernel/kgdb.c
78
struct thread_info *ti;
arch/arm/kernel/process.c
220
struct thread_info *thread = current_thread_info();
arch/arm/kernel/process.c
240
struct thread_info *thread = task_thread_info(p);
arch/arm/kernel/ptrace.c
296
struct thread_info *thread = task_thread_info(tsk);
arch/arm/kernel/ptrace.c
310
struct thread_info *thread = task_thread_info(tsk);
arch/arm/kernel/ptrace.c
585
struct thread_info *thread = task_thread_info(target);
arch/arm/kernel/ptrace.c
615
struct thread_info *thread = task_thread_info(target);
arch/arm/kernel/ptrace.c
637
struct thread_info *thread = task_thread_info(target);
arch/arm/kernel/thumbee.c
32
struct thread_info *thread = t;
arch/arm/kernel/xscale-cp0.c
35
struct thread_info *thread = t;
arch/arm/kernel/xscale-cp0.c
60
struct thread_info *thread = t;
arch/arm/mm/context.c
107
struct thread_info *thread = t;
arch/arm/nwfpe/fpmodule.c
50
struct thread_info *thread = v;
arch/arm/vfp/vfpmodule.c
104
static void vfp_force_reload(unsigned int cpu, struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
118
static void vfp_thread_flush(struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
146
static void vfp_thread_exit(struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
157
static void vfp_thread_copy(struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
159
struct thread_info *parent = current_thread_info();
arch/arm/vfp/vfpmodule.c
189
struct thread_info *thread = v;
arch/arm/vfp/vfpmodule.c
482
struct thread_info *ti = current_thread_info();
arch/arm/vfp/vfpmodule.c
547
void vfp_sync_hwstate(struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
566
void vfp_flush_hwstate(struct thread_info *thread)
arch/arm/vfp/vfpmodule.c
582
struct thread_info *thread = current_thread_info();
arch/arm/vfp/vfpmodule.c
620
struct thread_info *thread = current_thread_info();
arch/arm/vfp/vfpmodule.c
708
struct thread_info *ti = current_thread_info();
arch/arm/vfp/vfpmodule.c
868
struct thread_info *thread = current_thread_info();
arch/arm/vfp/vfpmodule.c
90
static bool vfp_state_in_hw(unsigned int cpu, struct thread_info *thread)
arch/arm64/include/asm/compat.h
103
static inline int is_compat_thread(struct thread_info *thread)
arch/arm64/include/asm/compat.h
94
static inline int is_compat_thread(struct thread_info *thread)
arch/arm64/include/asm/preempt.h
60
struct thread_info *ti = current_thread_info();
arch/arm64/kernel/asm-offsets.c
29
DEFINE(TSK_TI_CPU, offsetof(struct task_struct, thread_info.cpu));
arch/arm64/kernel/asm-offsets.c
30
DEFINE(TSK_TI_FLAGS, offsetof(struct task_struct, thread_info.flags));
arch/arm64/kernel/asm-offsets.c
31
DEFINE(TSK_TI_PREEMPT, offsetof(struct task_struct, thread_info.preempt_count));
arch/arm64/kernel/asm-offsets.c
33
DEFINE(TSK_TI_TTBR0, offsetof(struct task_struct, thread_info.ttbr0));
arch/arm64/kernel/asm-offsets.c
36
DEFINE(TSK_TI_SCS_BASE, offsetof(struct task_struct, thread_info.scs_base));
arch/arm64/kernel/asm-offsets.c
37
DEFINE(TSK_TI_SCS_SP, offsetof(struct task_struct, thread_info.scs_sp));
arch/arm64/kernel/debug-monitors.c
384
struct thread_info *ti = task_thread_info(task);
arch/arm64/kernel/process.c
627
struct thread_info *ti = task_thread_info(next);
arch/arm64/kernel/process.c
854
struct thread_info *ti = task_thread_info(task);
arch/arm64/kernel/process.c
888
struct thread_info *ti = task_thread_info(task);
arch/arm64/kernel/setup.c
369
init_task.thread_info.ttbr0 = phys_to_ttbr(__pa_symbol(reserved_pg_dir));
arch/arm64/net/bpf_jit_comp.c
1578
cpu_offset = offsetof(struct thread_info, cpu);
arch/csky/include/asm/thread_info.h
45
static inline struct thread_info *current_thread_info(void)
arch/csky/include/asm/thread_info.h
51
return (struct thread_info *)(sp & ~(THREAD_SIZE - 1));
arch/csky/kernel/asm-offsets.c
27
DEFINE(TINFO_FLAGS, offsetof(struct thread_info, flags));
arch/csky/kernel/asm-offsets.c
28
DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/csky/kernel/asm-offsets.c
29
DEFINE(TINFO_TP_VALUE, offsetof(struct thread_info, tp_value));
arch/csky/kernel/asm-offsets.c
30
DEFINE(TINFO_TASK, offsetof(struct thread_info, task));
arch/csky/kernel/syscall.c
8
struct thread_info *ti = task_thread_info(current);
arch/hexagon/include/asm/thread_info.h
73
register struct thread_info *__current_thread_info asm(QUOTED_THREADINFO_REG);
arch/hexagon/kernel/asm-offsets.c
65
OFFSET(_THREAD_INFO_FLAGS, thread_info, flags);
arch/hexagon/kernel/asm-offsets.c
66
OFFSET(_THREAD_INFO_PT_REGS, thread_info, regs);
arch/hexagon/kernel/asm-offsets.c
67
OFFSET(_THREAD_INFO_SP, thread_info, sp);
arch/hexagon/kernel/process.c
136
if (fp < (stack_page + sizeof(struct thread_info)) ||
arch/hexagon/kernel/process.c
58
struct thread_info *ti = task_thread_info(p);
arch/hexagon/kernel/smp.c
176
struct thread_info *thread = (struct thread_info *)idle->stack;
arch/hexagon/kernel/traps.c
117
low += sizeof(struct thread_info);
arch/loongarch/include/asm/smp.h
85
unsigned long thread_info;
arch/loongarch/include/asm/switch_to.h
26
struct task_struct *next, struct thread_info *next_ti,
arch/loongarch/include/asm/thread_info.h
47
register struct thread_info *__current_thread_info __asm__("$tp");
arch/loongarch/include/asm/thread_info.h
49
static inline struct thread_info *current_thread_info(void)
arch/loongarch/kernel/asm-offsets.c
270
OFFSET(CPU_BOOT_TINFO, secondary_data, thread_info);
arch/loongarch/kernel/asm-offsets.c
87
OFFSET(TI_TASK, thread_info, task);
arch/loongarch/kernel/asm-offsets.c
88
OFFSET(TI_FLAGS, thread_info, flags);
arch/loongarch/kernel/asm-offsets.c
89
OFFSET(TI_TP_VALUE, thread_info, tp_value);
arch/loongarch/kernel/asm-offsets.c
90
OFFSET(TI_CPU, thread_info, cpu);
arch/loongarch/kernel/asm-offsets.c
91
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
arch/loongarch/kernel/asm-offsets.c
92
OFFSET(TI_REGS, thread_info, regs);
arch/loongarch/kernel/ptrace.c
1084
struct thread_info *ti = task_thread_info(task);
arch/loongarch/kernel/ptrace.c
592
struct thread_info *ti = task_thread_info(tsk);
arch/loongarch/kernel/smp.c
404
cpuboot_data.thread_info = (unsigned long)task_thread_info(idle);
arch/m68k/include/asm/thread_info.h
44
static inline struct thread_info *current_thread_info(void)
arch/m68k/include/asm/thread_info.h
46
struct thread_info *ti;
arch/m68k/kernel/asm-offsets.c
43
DEFINE(TINFO_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/m68k/kernel/asm-offsets.c
44
DEFINE(TINFO_FLAGS, offsetof(struct thread_info, flags));
arch/m68k/kernel/process.c
277
if (fp < stack_page+sizeof(struct thread_info) ||
arch/microblaze/include/asm/switch_to.h
10
struct thread_info;
arch/microblaze/include/asm/switch_to.h
12
extern struct task_struct *_switch_to(struct thread_info *prev,
arch/microblaze/include/asm/switch_to.h
13
struct thread_info *next);
arch/microblaze/include/asm/thread_info.h
81
static inline struct thread_info *current_thread_info(void)
arch/microblaze/include/asm/thread_info.h
85
return (struct thread_info *)(sp & ~(THREAD_SIZE-1));
arch/microblaze/kernel/asm-offsets.c
88
DEFINE(TI_TASK, offsetof(struct thread_info, task));
arch/microblaze/kernel/asm-offsets.c
89
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/microblaze/kernel/asm-offsets.c
90
DEFINE(TI_CPU_CONTEXT, offsetof(struct thread_info, cpu_context));
arch/microblaze/kernel/asm-offsets.c
91
DEFINE(TI_PREEMPT_COUNT, offsetof(struct thread_info, preempt_count));
arch/microblaze/kernel/process.c
61
struct thread_info *ti = task_thread_info(p);
arch/microblaze/kernel/traps.c
42
fp = ((struct thread_info *)
arch/microblaze/kernel/unwind.c
282
struct thread_info *thread_info =
arch/microblaze/kernel/unwind.c
283
(struct thread_info *)(task->stack);
arch/microblaze/kernel/unwind.c
285
&thread_info->cpu_context;
arch/mips/include/asm/switch_to.h
33
struct task_struct *next, struct thread_info *next_ti);
arch/mips/include/asm/switch_to.h
54
struct thread_info *__prev_ti = task_thread_info(prev); \
arch/mips/include/asm/thread_info.h
62
extern struct thread_info *__current_thread_info;
arch/mips/include/asm/thread_info.h
64
register struct thread_info *__current_thread_info __asm__("$28");
arch/mips/include/asm/thread_info.h
67
static inline struct thread_info *current_thread_info(void)
arch/mips/kernel/asm-offsets.c
106
OFFSET(TI_TASK, thread_info, task);
arch/mips/kernel/asm-offsets.c
107
OFFSET(TI_FLAGS, thread_info, flags);
arch/mips/kernel/asm-offsets.c
108
OFFSET(TI_TP_VALUE, thread_info, tp_value);
arch/mips/kernel/asm-offsets.c
109
OFFSET(TI_CPU, thread_info, cpu);
arch/mips/kernel/asm-offsets.c
110
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
arch/mips/kernel/asm-offsets.c
111
OFFSET(TI_REGS, thread_info, regs);
arch/mips/kernel/asm-offsets.c
112
OFFSET(TI_SYSCALL, thread_info, syscall);
arch/mips/kernel/irq.c
87
if (unlikely(sp < (sizeof(struct thread_info) + STACK_WARN))) {
arch/mips/kernel/irq.c
89
sp - sizeof(struct thread_info));
arch/mips/kernel/mips-mt-fpaff.c
70
struct thread_info *ti;
arch/mips/kernel/process.c
113
struct thread_info *ti = task_thread_info(p);
arch/mips/kernel/smp-mt.c
145
struct thread_info *gp = task_thread_info(idle);
arch/mips/kernel/smp-mt.c
169
(unsigned long)(gp + sizeof(struct thread_info)));
arch/mips/kernel/syscall.c
89
struct thread_info *ti = task_thread_info(current);
arch/mips/kernel/traps.c
639
struct thread_info *ti = task_thread_info(current);
arch/mips/kvm/entry.c
313
uasm_i_lw(&p, GPR_T2, offsetof(struct thread_info, cpu), GPR_GP);
arch/nios2/include/asm/thread_info.h
58
static inline struct thread_info *current_thread_info(void)
arch/nios2/include/asm/thread_info.h
62
return (struct thread_info *)(sp & ~(THREAD_SIZE - 1));
arch/nios2/kernel/asm-offsets.c
70
OFFSET(TI_FLAGS, thread_info, flags);
arch/nios2/kernel/asm-offsets.c
71
OFFSET(TI_PREEMPT_COUNT, thread_info, preempt_count);
arch/openrisc/include/asm/processor.h
58
#define user_regs(thread_info) (((struct pt_regs *)((unsigned long)(thread_info) + THREAD_SIZE - STACK_FRAME_OVERHEAD)) - 1)
arch/openrisc/include/asm/thread_info.h
72
register struct thread_info *current_thread_info_reg asm("r10");
arch/openrisc/kernel/asm-offsets.c
48
DEFINE(TI_TASK, offsetof(struct thread_info, task));
arch/openrisc/kernel/asm-offsets.c
49
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/openrisc/kernel/asm-offsets.c
50
DEFINE(TI_PREEMPT, offsetof(struct thread_info, preempt_count));
arch/openrisc/kernel/asm-offsets.c
51
DEFINE(TI_KSP, offsetof(struct thread_info, ksp));
arch/openrisc/kernel/process.c
235
extern struct thread_info *_switch(struct thread_info *old_ti,
arch/openrisc/kernel/process.c
236
struct thread_info *new_ti);
arch/openrisc/kernel/process.c
243
struct thread_info *new_ti, *old_ti;
arch/openrisc/kernel/process.c
52
struct thread_info *current_thread_info_set[NR_CPUS] = { &init_thread_info, };
arch/openrisc/kernel/smp.c
33
struct thread_info *secondary_thread_info;
arch/parisc/kernel/asm-offsets.c
223
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/parisc/kernel/asm-offsets.c
224
DEFINE(TI_PRE_COUNT, offsetof(struct task_struct, thread_info.preempt_count));
arch/parisc/kernel/asm-offsets.c
43
DEFINE(TASK_TI_FLAGS, offsetof(struct task_struct, thread_info.flags));
arch/parisc/kernel/asm-offsets.c
45
DEFINE(TASK_TI_CPU, offsetof(struct task_struct, thread_info.cpu));
arch/powerpc/include/asm/thread_info.h
167
struct thread_info *ti = current_thread_info();
arch/powerpc/include/asm/thread_info.h
173
struct thread_info *ti = current_thread_info();
arch/powerpc/kernel/asm-offsets.c
175
OFFSET(TI_LOCAL_FLAGS, thread_info, local_flags);
arch/powerpc/kernel/asm-offsets.c
186
offsetof(struct task_struct, thread_info));
arch/powerpc/kernel/asm-offsets.c
95
OFFSET(TASK_CPU, task_struct, thread_info.cpu);
arch/powerpc/kernel/asm-offsets.c
99
OFFSET(TI_livepatch_sp, thread_info, livepatch_sp);
arch/powerpc/kernel/process.c
2363
struct thread_info *ti = current_thread_info();
arch/powerpc/kernel/process.c
2390
struct thread_info *ti = current_thread_info();
arch/powerpc/kernel/syscalls.c
113
struct thread_info *ti;
arch/powerpc/mm/book3s64/slb.c
292
static bool preload_hit(struct thread_info *ti, unsigned long esid)
arch/powerpc/mm/book3s64/slb.c
306
static void preload_add(struct thread_info *ti, unsigned long ea)
arch/powerpc/mm/book3s64/slb.c
333
static void preload_age(struct thread_info *ti)
arch/powerpc/mm/book3s64/slb.c
369
struct thread_info *ti = task_thread_info(tsk);
arch/riscv/include/asm/compat.h
22
static inline int is_compat_thread(struct thread_info *thread)
arch/riscv/kernel/asm-offsets.c
40
OFFSET(TASK_TI_CPU, task_struct, thread_info.cpu);
arch/riscv/kernel/asm-offsets.c
41
OFFSET(TASK_TI_PREEMPT_COUNT, task_struct, thread_info.preempt_count);
arch/riscv/kernel/asm-offsets.c
42
OFFSET(TASK_TI_KERNEL_SP, task_struct, thread_info.kernel_sp);
arch/riscv/kernel/asm-offsets.c
43
OFFSET(TASK_TI_USER_SP, task_struct, thread_info.user_sp);
arch/riscv/kernel/asm-offsets.c
45
OFFSET(TASK_TI_SCS_SP, task_struct, thread_info.scs_sp);
arch/riscv/kernel/asm-offsets.c
48
OFFSET(TASK_TI_A0, task_struct, thread_info.a0);
arch/riscv/kernel/asm-offsets.c
49
OFFSET(TASK_TI_A1, task_struct, thread_info.a1);
arch/riscv/kernel/asm-offsets.c
50
OFFSET(TASK_TI_A2, task_struct, thread_info.a2);
arch/riscv/kernel/asm-offsets.c
53
OFFSET(TASK_TI_CPU_NUM, task_struct, thread_info.cpu);
arch/riscv/kernel/asm-offsets.c
55
OFFSET(TASK_TI_CFI_STATE, task_struct, thread_info.user_cfi_state);
arch/riscv/kernel/asm-offsets.c
56
OFFSET(TASK_TI_USER_SSP, task_struct, thread_info.user_cfi_state.user_shdw_stk);
arch/riscv/kernel/process.c
311
struct thread_info *ti = task_thread_info(task);
arch/riscv/kernel/process.c
373
struct thread_info *ti = task_thread_info(task);
arch/riscv/kernel/ptrace.c
633
if (is_compat_thread(&task->thread_info))
arch/riscv/kernel/smpboot.c
193
tidle->thread_info.cpu = cpu;
arch/riscv/kernel/usercfi.c
109
task->thread_info.user_cfi_state.ufcfi_locked = 1;
arch/riscv/kernel/usercfi.c
26
return task->thread_info.user_cfi_state.ubcfi_en;
arch/riscv/kernel/usercfi.c
31
return task->thread_info.user_cfi_state.shdw_stk_base;
arch/riscv/kernel/usercfi.c
36
return task->thread_info.user_cfi_state.ubcfi_locked;
arch/riscv/kernel/usercfi.c
41
task->thread_info.user_cfi_state.shdw_stk_base = shstk_addr;
arch/riscv/kernel/usercfi.c
42
task->thread_info.user_cfi_state.shdw_stk_size = size;
arch/riscv/kernel/usercfi.c
48
*size = task->thread_info.user_cfi_state.shdw_stk_size;
arch/riscv/kernel/usercfi.c
49
return task->thread_info.user_cfi_state.shdw_stk_base;
arch/riscv/kernel/usercfi.c
54
task->thread_info.user_cfi_state.user_shdw_stk = shstk_addr;
arch/riscv/kernel/usercfi.c
59
return task->thread_info.user_cfi_state.user_shdw_stk;
arch/riscv/kernel/usercfi.c
67
task->thread_info.user_cfi_state.ubcfi_en = enable ? 1 : 0;
arch/riscv/kernel/usercfi.c
79
task->thread_info.user_cfi_state.ubcfi_locked = 1;
arch/riscv/kernel/usercfi.c
84
return task->thread_info.user_cfi_state.ufcfi_en;
arch/riscv/kernel/usercfi.c
89
return task->thread_info.user_cfi_state.ufcfi_locked;
arch/riscv/kernel/usercfi.c
97
task->thread_info.user_cfi_state.ufcfi_en = enable ? 1 : 0;
arch/riscv/net/bpf_jit_comp64.c
1399
emit_lw(RV_REG_T1, offsetof(struct thread_info, cpu),
arch/riscv/net/bpf_jit_comp64.c
1806
emit_lw(bpf_to_rv_reg(BPF_REG_0, ctx), offsetof(struct thread_info, cpu),
arch/s390/kernel/asm-offsets.c
32
OFFSET(__TI_flags, task_struct, thread_info.flags);
arch/s390/kernel/asm-offsets.c
33
OFFSET(__TI_sie, task_struct, thread_info.sie);
arch/sh/include/asm/thread_info.h
156
struct thread_info *ti = current_thread_info();
arch/sh/include/asm/thread_info.h
163
struct thread_info *ti = current_thread_info();
arch/sh/include/asm/thread_info.h
66
static inline struct thread_info *current_thread_info(void)
arch/sh/include/asm/thread_info.h
68
struct thread_info *ti;
arch/sh/kernel/asm-offsets.c
25
DEFINE(TI_TASK, offsetof(struct thread_info, task));
arch/sh/kernel/asm-offsets.c
26
DEFINE(TI_FLAGS, offsetof(struct thread_info, flags));
arch/sh/kernel/asm-offsets.c
27
DEFINE(TI_CPU, offsetof(struct thread_info, cpu));
arch/sh/kernel/asm-offsets.c
28
DEFINE(TI_PRE_COUNT, offsetof(struct thread_info, preempt_count));
arch/sh/kernel/asm-offsets.c
29
DEFINE(TI_SIZE, sizeof(struct thread_info));
arch/sh/kernel/dumpstack.c
57
struct thread_info *tinfo, int *graph)
arch/sh/kernel/dumpstack.c
83
struct thread_info *tinfo, int *graph)
arch/sh/kernel/dumpstack.c
92
struct thread_info *context;
arch/sh/kernel/dumpstack.c
95
context = (struct thread_info *)
arch/sh/kernel/irq.c
155
struct thread_info *curctx;
arch/sh/kernel/irq.c
62
struct thread_info tinfo;
arch/sh/kernel/process_32.c
95
struct thread_info *ti = task_thread_info(p);
arch/sh/kernel/smp.c
211
void *thread_info;
arch/sh/kernel/smp.c
222
stack_start.thread_info = tsk->stack;
arch/sparc/include/asm/ptrace.h
47
struct thread_info *thread;
arch/sparc/include/asm/switch_to_32.h
7
extern struct thread_info *current_set[NR_CPUS];
arch/sparc/include/asm/thread_info_32.h
67
register struct thread_info *current_thread_info_reg asm("g6");
arch/sparc/include/asm/thread_info_64.h
124
register struct thread_info *current_thread_info_reg asm("g6");
arch/sparc/include/asm/thread_info_64.h
127
extern struct thread_info *current_thread_info(void);
arch/sparc/include/asm/trap_block.h
25
struct thread_info;
arch/sparc/include/asm/trap_block.h
28
struct thread_info *thread;
arch/sparc/include/asm/trap_block.h
57
void init_cur_cpu_trap(struct thread_info *);
arch/sparc/kernel/kgdb_32.c
49
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/kgdb_64.c
47
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/kstack.h
11
static inline bool kstack_valid(struct thread_info *tp, unsigned long sp)
arch/sparc/kernel/kstack.h
19
if (sp >= (base + sizeof(struct thread_info)) &&
arch/sparc/kernel/kstack.h
37
static inline bool kstack_is_trap_frame(struct thread_info *tp, struct pt_regs *regs)
arch/sparc/kernel/process_32.c
162
if (fp < (task_base + sizeof(struct thread_info)) ||
arch/sparc/kernel/process_32.c
267
struct thread_info *ti = task_thread_info(p);
arch/sparc/kernel/process_32.c
389
if (fp < (task_base + sizeof(struct thread_info)) ||
arch/sparc/kernel/process_32.c
67
struct thread_info *current_set[NR_CPUS];
arch/sparc/kernel/process_64.c
191
static void __global_reg_self(struct thread_info *tp, struct pt_regs *regs,
arch/sparc/kernel/process_64.c
241
struct thread_info *tp = current_thread_info();
arch/sparc/kernel/process_64.c
400
struct thread_info *t = task_thread_info(tsk);
arch/sparc/kernel/process_64.c
412
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
468
struct thread_info *t)
arch/sparc/kernel/process_64.c
481
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
519
struct thread_info *t = current_thread_info();
arch/sparc/kernel/process_64.c
574
struct thread_info *t = task_thread_info(p);
arch/sparc/kernel/process_64.c
675
struct thread_info *tp;
arch/sparc/kernel/ptrace_64.c
345
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/ptrace_64.c
680
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/ptrace_64.c
833
struct thread_info *t = task_thread_info(target);
arch/sparc/kernel/sigutil_32.c
102
struct thread_info *t = current_thread_info();
arch/sparc/kernel/sigutil_64.c
76
struct thread_info *t = current_thread_info();
arch/sparc/kernel/smp_64.c
345
static struct thread_info *cpu_new_thread = NULL;
arch/sparc/kernel/stacktrace.c
13
static void __save_stack_trace(struct thread_info *tp,
arch/sparc/kernel/stacktrace.c
85
struct thread_info *tp = task_thread_info(tsk);
arch/sparc/kernel/traps_32.c
372
if (TI_UWINMASK != offsetof(struct thread_info, uwinmask) ||
arch/sparc/kernel/traps_32.c
373
TI_TASK != offsetof(struct thread_info, task) ||
arch/sparc/kernel/traps_32.c
374
TI_FLAGS != offsetof(struct thread_info, flags) ||
arch/sparc/kernel/traps_32.c
375
TI_CPU != offsetof(struct thread_info, cpu) ||
arch/sparc/kernel/traps_32.c
376
TI_PREEMPT != offsetof(struct thread_info, preempt_count) ||
arch/sparc/kernel/traps_32.c
377
TI_SOFTIRQ != offsetof(struct thread_info, softirq_count) ||
arch/sparc/kernel/traps_32.c
378
TI_HARDIRQ != offsetof(struct thread_info, hardirq_count) ||
arch/sparc/kernel/traps_32.c
379
TI_KSP != offsetof(struct thread_info, ksp) ||
arch/sparc/kernel/traps_32.c
380
TI_KPC != offsetof(struct thread_info, kpc) ||
arch/sparc/kernel/traps_32.c
381
TI_KPSR != offsetof(struct thread_info, kpsr) ||
arch/sparc/kernel/traps_32.c
382
TI_KWIM != offsetof(struct thread_info, kwim) ||
arch/sparc/kernel/traps_32.c
383
TI_REG_WINDOW != offsetof(struct thread_info, reg_window) ||
arch/sparc/kernel/traps_32.c
384
TI_RWIN_SPTRS != offsetof(struct thread_info, rwbuf_stkptrs) ||
arch/sparc/kernel/traps_32.c
385
TI_W_SAVED != offsetof(struct thread_info, w_saved))
arch/sparc/kernel/traps_64.c
2453
struct thread_info *tp;
arch/sparc/kernel/traps_64.c
2538
struct thread_info *tp = current_thread_info();
arch/sparc/kernel/traps_64.c
2826
void notrace init_cur_cpu_trap(struct thread_info *t)
arch/sparc/kernel/traps_64.c
2843
BUILD_BUG_ON(TI_TASK != offsetof(struct thread_info, task) ||
arch/sparc/kernel/traps_64.c
2844
TI_FLAGS != offsetof(struct thread_info, flags) ||
arch/sparc/kernel/traps_64.c
2845
TI_CPU != offsetof(struct thread_info, cpu) ||
arch/sparc/kernel/traps_64.c
2846
TI_FPSAVED != offsetof(struct thread_info, fpsaved) ||
arch/sparc/kernel/traps_64.c
2847
TI_KSP != offsetof(struct thread_info, ksp) ||
arch/sparc/kernel/traps_64.c
2848
TI_FAULT_ADDR != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2850
TI_KREGS != offsetof(struct thread_info, kregs) ||
arch/sparc/kernel/traps_64.c
2851
TI_UTRAPS != offsetof(struct thread_info, utraps) ||
arch/sparc/kernel/traps_64.c
2852
TI_REG_WINDOW != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2854
TI_RWIN_SPTRS != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2856
TI_GSR != offsetof(struct thread_info, gsr) ||
arch/sparc/kernel/traps_64.c
2857
TI_XFSR != offsetof(struct thread_info, xfsr) ||
arch/sparc/kernel/traps_64.c
2858
TI_PRE_COUNT != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2860
TI_NEW_CHILD != offsetof(struct thread_info, new_child) ||
arch/sparc/kernel/traps_64.c
2861
TI_KUNA_REGS != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2863
TI_KUNA_INSN != offsetof(struct thread_info,
arch/sparc/kernel/traps_64.c
2865
TI_FPREGS != offsetof(struct thread_info, fpregs) ||
arch/sparc/kernel/windows.c
115
struct thread_info *tp = current_thread_info();
arch/sparc/kernel/windows.c
42
static inline void shift_window_buffer(int first_win, int last_win, struct thread_info *tp)
arch/sparc/kernel/windows.c
63
struct thread_info *tp = current_thread_info();
arch/sparc/net/bpf_jit_comp_32.c
221
emit_load32(G6, struct thread_info, cpu, REG)
arch/um/kernel/smp.c
152
idle->thread_info.cpu = cpu;
arch/x86/include/asm/syscall.h
124
if (task->thread_info.status & TS_COMPAT) {
arch/x86/include/asm/syscall.h
148
if (task->thread_info.status & TS_COMPAT) {
arch/x86/include/asm/syscall.h
171
task->thread_info.status & TS_COMPAT)
arch/x86/include/asm/syscall.h
63
if (task->thread_info.status & (TS_COMPAT|TS_I386_REGS_POKED))
arch/x86/kernel/cpu/bugs.c
188
struct thread_info *ti = current_thread_info();
arch/x86/kernel/cpu/bugs.c
2498
set_ti_thread_flag(&task->thread_info, TIF_SPEC_L1D_FLUSH);
arch/x86/kernel/cpu/bugs.c
2501
clear_ti_thread_flag(&task->thread_info, TIF_SPEC_L1D_FLUSH);
arch/x86/kernel/cpu/bugs.c
2642
if (test_ti_thread_flag(&task->thread_info, TIF_SPEC_L1D_FLUSH))
arch/x86/kernel/irq_32.c
36
return sp < (sizeof(struct thread_info) + STACK_WARN);
arch/x86/kernel/ptrace.c
934
child->thread_info.status |= TS_I386_REGS_POKED;
arch/x86/kernel/uprobes.c
1830
struct thread_info *ti = current_thread_info();
arch/x86/mm/tlb.c
646
clear_ti_thread_flag(&next->thread_info, TIF_SPEC_L1D_FLUSH);
arch/xtensa/include/asm/coprocessor.h
145
struct thread_info;
arch/xtensa/include/asm/coprocessor.h
146
void coprocessor_flush(struct thread_info *ti, int cp_index);
arch/xtensa/include/asm/coprocessor.h
147
void coprocessor_release_all(struct thread_info *ti);
arch/xtensa/include/asm/coprocessor.h
148
void coprocessor_flush_all(struct thread_info *ti);
arch/xtensa/include/asm/coprocessor.h
149
void coprocessor_flush_release_all(struct thread_info *ti);
arch/xtensa/include/asm/thread_info.h
94
static __always_inline struct thread_info *current_thread_info(void)
arch/xtensa/include/asm/thread_info.h
96
struct thread_info *ti;
arch/xtensa/include/asm/traps.h
32
struct thread_info *coprocessor_owner[XCHAL_CP_MAX];
arch/xtensa/kernel/asm-offsets.c
100
DEFINE(THREAD_ATOMCTL8, offsetof (struct thread_info, atomctl8));
arch/xtensa/kernel/asm-offsets.c
102
DEFINE(THREAD_CPENABLE, offsetof(struct thread_info, cpenable));
arch/xtensa/kernel/asm-offsets.c
103
DEFINE(THREAD_CPU, offsetof(struct thread_info, cpu));
arch/xtensa/kernel/asm-offsets.c
104
DEFINE(THREAD_CP_OWNER_CPU, offsetof(struct thread_info, cp_owner_cpu));
arch/xtensa/kernel/asm-offsets.c
106
DEFINE(THREAD_XTREGS_CP0, offsetof(struct thread_info, xtregs_cp.cp0));
arch/xtensa/kernel/asm-offsets.c
107
DEFINE(THREAD_XTREGS_CP1, offsetof(struct thread_info, xtregs_cp.cp1));
arch/xtensa/kernel/asm-offsets.c
108
DEFINE(THREAD_XTREGS_CP2, offsetof(struct thread_info, xtregs_cp.cp2));
arch/xtensa/kernel/asm-offsets.c
109
DEFINE(THREAD_XTREGS_CP3, offsetof(struct thread_info, xtregs_cp.cp3));
arch/xtensa/kernel/asm-offsets.c
110
DEFINE(THREAD_XTREGS_CP4, offsetof(struct thread_info, xtregs_cp.cp4));
arch/xtensa/kernel/asm-offsets.c
111
DEFINE(THREAD_XTREGS_CP5, offsetof(struct thread_info, xtregs_cp.cp5));
arch/xtensa/kernel/asm-offsets.c
112
DEFINE(THREAD_XTREGS_CP6, offsetof(struct thread_info, xtregs_cp.cp6));
arch/xtensa/kernel/asm-offsets.c
113
DEFINE(THREAD_XTREGS_CP7, offsetof(struct thread_info, xtregs_cp.cp7));
arch/xtensa/kernel/asm-offsets.c
115
DEFINE(THREAD_XTREGS_USER, offsetof (struct thread_info, xtregs_user));
arch/xtensa/kernel/asm-offsets.c
87
OFFSET(TI_TASK, thread_info, task);
arch/xtensa/kernel/asm-offsets.c
88
OFFSET(TI_FLAGS, thread_info, flags);
arch/xtensa/kernel/asm-offsets.c
89
OFFSET(TI_STSTUS, thread_info, status);
arch/xtensa/kernel/asm-offsets.c
90
OFFSET(TI_CPU, thread_info, cpu);
arch/xtensa/kernel/asm-offsets.c
91
OFFSET(TI_PRE_COUNT, thread_info, preempt_count);
arch/xtensa/kernel/asm-offsets.c
93
OFFSET(TI_PS_WOE_FIX_ADDR, thread_info, ps_woe_fix_addr);
arch/xtensa/kernel/irq.c
44
if (unlikely(sp < (sizeof(thread_info) + 1024)))
arch/xtensa/kernel/irq.c
46
sp - sizeof(struct thread_info));
arch/xtensa/kernel/process.c
103
struct thread_info *ti = info;
arch/xtensa/kernel/process.c
104
struct thread_info **coprocessor_owner;
arch/xtensa/kernel/process.c
122
void coprocessor_release_all(struct thread_info *ti)
arch/xtensa/kernel/process.c
135
struct thread_info *ti = info;
arch/xtensa/kernel/process.c
136
struct thread_info **coprocessor_owner;
arch/xtensa/kernel/process.c
150
void coprocessor_flush_all(struct thread_info *ti)
arch/xtensa/kernel/process.c
167
void coprocessor_flush_release_all(struct thread_info *ti)
arch/xtensa/kernel/process.c
207
struct thread_info *ti = current_thread_info();
arch/xtensa/kernel/process.c
276
struct thread_info *ti;
arch/xtensa/kernel/process.c
70
struct thread_info **coprocessor_owner;
arch/xtensa/kernel/process.c
71
struct thread_info *unique_owner[XCHAL_CP_MAX];
arch/xtensa/kernel/process.c
79
struct thread_info *ti = coprocessor_owner[i];
arch/xtensa/kernel/ptrace.c
125
struct thread_info *ti = task_thread_info(target);
arch/xtensa/kernel/ptrace.c
158
struct thread_info *ti = task_thread_info(target);
arch/xtensa/kernel/signal.c
147
struct thread_info *ti = current_thread_info();
arch/xtensa/kernel/signal.c
185
struct thread_info *ti = current_thread_info();
include/linux/sched.h
1953
struct thread_info thread_info;
include/linux/sched.h
1959
extern struct thread_info init_thread_info;
include/linux/sched.h
1965
# define task_thread_info(task) (&(task)->thread_info)
include/linux/sched.h
1967
# define task_thread_info(task) ((struct thread_info *)(task)->stack)
include/linux/sched.h
826
struct thread_info thread_info;
include/linux/thread_info.h
102
static inline void set_ti_thread_flag(struct thread_info *ti, int flag)
include/linux/thread_info.h
107
static inline void clear_ti_thread_flag(struct thread_info *ti, int flag)
include/linux/thread_info.h
112
static inline void update_ti_thread_flag(struct thread_info *ti, int flag,
include/linux/thread_info.h
121
static inline int test_and_set_ti_thread_flag(struct thread_info *ti, int flag)
include/linux/thread_info.h
126
static inline int test_and_clear_ti_thread_flag(struct thread_info *ti, int flag)
include/linux/thread_info.h
131
static inline int test_ti_thread_flag(struct thread_info *ti, int flag)
include/linux/thread_info.h
140
static __always_inline unsigned long read_ti_thread_flags(struct thread_info *ti)
include/linux/thread_info.h
24
#define current_thread_info() ((struct thread_info *)current)
init/init_task.c
260
struct thread_info init_thread_info __init_thread_info = INIT_THREAD_INFO(init_task);
init/init_task.c
98
.thread_info = INIT_THREAD_INFO(init_task),
kernel/sched/core.c
1006
static inline bool set_nr_and_not_polling(struct thread_info *ti, int tif)
kernel/sched/core.c
1115
struct thread_info *cti = task_thread_info(curr);
kernel/sched/core.c
979
static inline bool set_nr_and_not_polling(struct thread_info *ti, int tif)
kernel/sched/core.c
992
struct thread_info *ti = task_thread_info(p);
mm/kmsan/core.c
43
struct thread_info *info = current_thread_info();
tools/testing/selftests/bpf/bpf_experimental.h
647
return bpf_get_current_task_btf()->thread_info.preempt.count;
tools/virtio/virtio-trace/trace-agent-rw.c
104
static void *rw_thread_main(void *thread_info)
tools/virtio/virtio-trace/trace-agent-rw.c
108
struct rw_thread_info *ts = (struct rw_thread_info *)thread_info;