Symbol: current
arch/alpha/include/asm/cacheflush.h
43
if (current->active_mm == mm)
arch/alpha/include/asm/mmu_context.h
174
struct mm_struct * mm = current->active_mm; \
arch/alpha/include/asm/mmu_context.h
201
if (tsk != current)
arch/alpha/include/asm/processor.h
40
((tsk) == current ? rdusp() : task_thread_info(tsk)->pcb.usp)
arch/alpha/include/asm/tlbflush.h
51
flush_tlb_current(current->active_mm);
arch/alpha/include/asm/tlbflush.h
78
if (mm == current->active_mm)
arch/alpha/include/asm/tlbflush.h
90
if (mm == current->active_mm)
arch/alpha/kernel/osf_sys.c
1076
task_cputime(current, &utime, &stime);
arch/alpha/kernel/osf_sys.c
1081
r.ru_minflt = current->min_flt;
arch/alpha/kernel/osf_sys.c
1082
r.ru_majflt = current->maj_flt;
arch/alpha/kernel/osf_sys.c
1085
utime_jiffies = nsecs_to_jiffies(current->signal->cutime);
arch/alpha/kernel/osf_sys.c
1086
stime_jiffies = nsecs_to_jiffies(current->signal->cstime);
arch/alpha/kernel/osf_sys.c
1089
r.ru_minflt = current->signal->cmin_flt;
arch/alpha/kernel/osf_sys.c
1090
r.ru_majflt = current->signal->cmaj_flt;
arch/alpha/kernel/osf_sys.c
186
current->comm, flags);
arch/alpha/kernel/osf_sys.c
676
unsigned long oss_sp = current->sas_ss_sp + current->sas_ss_size;
arch/alpha/kernel/osf_sys.c
690
if (current->sas_ss_sp && on_sig_stack(usp))
arch/alpha/kernel/osf_sys.c
696
current->sas_ss_sp = (unsigned long)ss_sp - SIGSTKSZ;
arch/alpha/kernel/osf_sys.c
697
current->sas_ss_size = SIGSTKSZ;
arch/alpha/kernel/osf_sys.c
76
mm = current->mm;
arch/alpha/kernel/osf_sys.c
869
0, current);
arch/alpha/kernel/pci-sysfs.c
270
current->comm, sparse ? " sparse" : "", start, start + nr,
arch/alpha/kernel/pci-sysfs.c
50
current->comm, sparse ? " sparse" : "", start, start + nr,
arch/alpha/kernel/signal.c
155
current->restart_block.fn = do_no_restart_syscall;
arch/alpha/kernel/signal.c
222
if (ptrace_cancel_bpt (current)) {
arch/alpha/kernel/signal.c
224
current);
arch/alpha/kernel/signal.c
250
if (ptrace_cancel_bpt (current)) {
arch/alpha/kernel/signal.c
252
current);
arch/alpha/kernel/signal.c
368
current->comm, current->pid, frame, regs->pc, regs->r26);
arch/alpha/kernel/signal.c
422
current->comm, current->pid, frame, regs->pc, regs->r26);
arch/alpha/kernel/signal.c
486
unsigned long single_stepping = ptrace_cancel_bpt(current);
arch/alpha/kernel/signal.c
492
single_stepping |= ptrace_cancel_bpt(current);
arch/alpha/kernel/signal.c
498
single_stepping |= ptrace_cancel_bpt(current);
arch/alpha/kernel/signal.c
519
ptrace_set_bpt(current); /* re-set breakpoint */
arch/alpha/kernel/smp.c
148
current->active_mm = &init_mm;
arch/alpha/kernel/smp.c
168
cpuid, current, current->active_mm));
arch/alpha/kernel/smp.c
628
if (mm == current->active_mm && !asn_locked())
arch/alpha/kernel/smp.c
639
if (mm == current->active_mm) {
arch/alpha/kernel/smp.c
672
if (mm == current->active_mm && !asn_locked())
arch/alpha/kernel/smp.c
686
if (mm == current->active_mm) {
arch/alpha/kernel/smp.c
723
if (mm == current->active_mm && !asn_locked())
arch/alpha/kernel/smp.c
740
if (mm == current->active_mm) {
arch/alpha/kernel/traps.c
150
printk("%s(%d): %s %ld\n", current->comm, task_pid_nr(current), str, err);
arch/alpha/kernel/traps.c
196
send_sig_fault_trapno(SIGFPE, si_code, (void __user *) regs->pc, 0, current);
arch/alpha/kernel/traps.c
247
if (ptrace_cancel_bpt(current)) {
arch/alpha/kernel/traps.c
252
current);
arch/alpha/kernel/traps.c
257
(void __user *) regs->pc, 0, current);
arch/alpha/kernel/traps.c
319
regs->r16, current);
arch/alpha/kernel/traps.c
330
send_sig_fault(SIGILL, ILL_ILLOPC, (void __user *)regs->pc, current);
arch/alpha/kernel/traps.c
542
current->comm, task_pid_nr(current));
arch/alpha/kernel/traps.c
678
current->comm, task_pid_nr(current),
arch/alpha/kernel/traps.c
899
struct mm_struct *mm = current->mm;
arch/alpha/kernel/traps.c
907
send_sig_fault(SIGSEGV, si_code, va, current);
arch/alpha/kernel/traps.c
912
send_sig_fault(SIGBUS, BUS_ADRALN, va, current);
arch/alpha/mm/fault.c
239
pgd = current->active_mm->pgd + index;
arch/alpha/mm/fault.c
89
struct mm_struct *mm = current->mm;
arch/alpha/mm/tlbflush.c
45
if (mm == current->active_mm)
arch/alpha/mm/tlbflush.c
69
if (d->mm == current->active_mm && !asn_locked())
arch/arc/include/asm/entry-arcv2.h
160
; - K mode: add the offset from current SP where H/w starts auto push
arch/arc/kernel/kgdb.c
114
current->thread.callee_reg,
arch/arc/kernel/kgdb.c
60
current->thread.callee_reg);
arch/arc/kernel/kgdb.c
66
current->thread.callee_reg);
arch/arc/kernel/kprobes.c
171
(struct callee_regs *) current->thread.callee_reg,
arch/arc/kernel/process.c
232
task_thread_info(current)->thr_ptr;
arch/arc/kernel/process.c
27
task_thread_info(current)->thr_ptr = (unsigned int)user_tls_data_ptr;
arch/arc/kernel/process.c
43
return task_thread_info(current)->thr_ptr;
arch/arc/kernel/process.c
93
mmap_read_lock(current->mm);
arch/arc/kernel/process.c
94
ret = fixup_user_fault(current->mm, (unsigned long) uaddr,
arch/arc/kernel/process.c
96
mmap_read_unlock(current->mm);
arch/arc/kernel/ptrace.c
350
trace_sys_enter(regs, syscall_get_nr(current, regs));
arch/arc/kernel/signal.c
203
current->restart_block.fn = do_no_restart_syscall;
arch/arc/kernel/smp.c
170
current->active_mm = mm;
arch/arc/kernel/stacktrace.c
58
} else if (tsk == NULL || tsk == current) {
arch/arc/kernel/stacktrace.c
64
frame_info->task = current;
arch/arc/kernel/traps.c
43
struct task_struct *tsk = current;
arch/arc/kernel/troubleshoot.c
118
address = current->thread.fault_address;
arch/arc/kernel/troubleshoot.c
172
struct task_struct *tsk = current;
arch/arc/kernel/troubleshoot.c
190
regs->ecr.full, current->thread.fault_address, regs->ret);
arch/arc/kernel/troubleshoot.c
219
current->thread.fault_address = address;
arch/arc/kernel/troubleshoot.c
229
show_stacktrace(current, regs, KERN_DEFAULT);
arch/arc/kernel/troubleshoot.c
80
struct mm_struct *active_mm = current->active_mm;
arch/arc/kernel/unaligned.c
214
current->comm, task_pid_nr(current));
arch/arc/kernel/unaligned.c
218
current->comm, task_pid_nr(current),
arch/arc/mm/fault.c
37
pgd = pgd_offset(current->active_mm, address);
arch/arc/mm/fault.c
76
struct task_struct *tsk = current;
arch/arc/mm/mmap.c
29
struct mm_struct *mm = current->mm;
arch/arc/mm/tlb.c
196
if (current->mm == mm)
arch/arc/mm/tlb.c
425
if (current->active_mm != vma->vm_mm)
arch/arm/common/bL_switcher.c
277
sched_set_fifo_low(current);
arch/arm/common/bL_switcher.c
281
if (signal_pending(current))
arch/arm/common/bL_switcher.c
282
flush_signals(current);
arch/arm/include/asm/elf.h
147
(elf_addr_t)current->mm->context.vdso); \
arch/arm/include/asm/mmu_context.h
85
struct mm_struct *mm = current->mm;
arch/arm/include/asm/processor.h
20
#define STACK_TOP ((current->personality & ADDR_LIMIT_32BIT) ? \
arch/arm/include/asm/processor.h
63
current->personality & FDPIC_FUNCPTRS) { \
arch/arm/include/asm/processor.h
67
regs->ARM_r10 = current->mm->start_data; \
arch/arm/include/asm/processor.h
69
regs->ARM_r10 = current->mm->start_data; \
arch/arm/include/asm/processor.h
70
if (current->personality & ADDR_LIMIT_32BIT) \
arch/arm/include/asm/stackprotector.h
32
current->stack_canary = canary;
arch/arm/include/asm/stackprotector.h
34
__stack_chk_guard = current->stack_canary;
arch/arm/include/asm/stacktrace.h
32
unsigned long delta = current_stack_pointer ^ (unsigned long)current->stack;
arch/arm/include/asm/stacktrace.h
46
frame->tsk = current;
arch/arm/include/asm/syscall.h
42
return __in_oabi_syscall(current);
arch/arm/kernel/elf.c
46
unsigned int personality = current->personality & ~PER_MASK;
arch/arm/kernel/ftrace.c
234
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/arm/kernel/perf_callchain.c
69
if (!current->mm)
arch/arm/kernel/perf_regs.c
37
regs_user->regs = task_pt_regs(current);
arch/arm/kernel/perf_regs.c
38
regs_user->abi = perf_reg_abi(current);
arch/arm/kernel/process.c
221
struct task_struct *tsk = current;
arch/arm/kernel/process.c
385
current->mm->context.sigpage = new_vma->vm_start;
arch/arm/kernel/process.c
397
struct mm_struct *mm = current->mm;
arch/arm/kernel/ptrace.c
359
if (current->thread.debug.hbp[i] == bp)
arch/arm/kernel/ptrace.c
862
secure_computing_strict(syscall_get_nr(current, regs));
arch/arm/kernel/ptrace.c
866
scno = syscall_get_nr(current, regs);
arch/arm/kernel/return_address.c
48
frame.tsk = current;
arch/arm/kernel/signal.c
192
current->restart_block.fn = do_no_restart_syscall;
arch/arm/kernel/signal.c
222
current->restart_block.fn = do_no_restart_syscall;
arch/arm/kernel/signal.c
276
.trap_no = current->thread.trap_no,
arch/arm/kernel/signal.c
277
.error_code = current->thread.error_code,
arch/arm/kernel/signal.c
278
.fault_address = current->thread.address,
arch/arm/kernel/signal.c
330
(current->personality & FDPIC_FUNCPTRS);
arch/arm/kernel/signal.c
413
struct mm_struct *mm = current->mm;
arch/arm/kernel/smp.c
377
: "r" (task_stack_page(current) + THREAD_SIZE - 8),
arch/arm/kernel/smp.c
378
"r" (current)
arch/arm/kernel/smp.c
425
enter_lazy_tlb(mm, current);
arch/arm/kernel/smp.c
434
current->active_mm = mm;
arch/arm/kernel/stacktrace.c
170
} else if (task != current) {
arch/arm/kernel/suspend.c
23
struct mm_struct *mm = current->active_mm;
arch/arm/kernel/swp_emulate.c
101
mmap_read_lock(current->mm);
arch/arm/kernel/swp_emulate.c
102
if (find_vma(current->mm, addr) == NULL)
arch/arm/kernel/swp_emulate.c
106
mmap_read_unlock(current->mm);
arch/arm/kernel/swp_emulate.c
139
if (likely(res != -EAGAIN) || signal_pending(current))
arch/arm/kernel/swp_emulate.c
182
if (current->pid != previous_pid) {
arch/arm/kernel/swp_emulate.c
184
current->comm, (unsigned long)current->pid);
arch/arm/kernel/swp_emulate.c
185
previous_pid = current->pid;
arch/arm/kernel/sys_oabi-compat.c
347
ns = current->nsproxy->ipc_ns;
arch/arm/kernel/traps.c
227
tsk = current;
arch/arm/kernel/traps.c
232
} else if (tsk != current) {
arch/arm/kernel/traps.c
274
struct task_struct *tsk = current;
arch/arm/kernel/traps.c
332
if (regs && kexec_should_crash(current))
arch/arm/kernel/traps.c
378
current->thread.error_code = err;
arch/arm/kernel/traps.c
379
current->thread.trap_no = trap;
arch/arm/kernel/traps.c
488
current->comm, task_pid_nr(current), pc);
arch/arm/kernel/traps.c
542
if ((current->personality & PER_MASK) != PER_LINUX) {
arch/arm/kernel/traps.c
543
send_sig(SIGSEGV, current, 1);
arch/arm/kernel/traps.c
550
task_pid_nr(current), current->comm, n);
arch/arm/kernel/traps.c
572
if (fatal_signal_pending(current))
arch/arm/kernel/traps.c
673
task_pid_nr(current), current->comm, no);
arch/arm/kernel/traps.c
739
task_pid_nr(current), current->comm, code, instr);
arch/arm/kernel/traps.c
741
show_pte(KERN_ERR, current->mm, addr);
arch/arm/kernel/traps.c
923
unsigned long tsk_stk = (unsigned long)current->stack;
arch/arm/kernel/unwind.c
532
tsk = current;
arch/arm/kernel/unwind.c
539
} else if (tsk == current) {
arch/arm/kernel/vdso.c
40
current->mm->context.vdso = new_vma->vm_start;
arch/arm/lib/uaccess_with_memcpy.c
102
mmap_read_lock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
110
mmap_read_unlock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
114
mmap_read_lock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
134
mmap_read_unlock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
166
mmap_read_lock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
173
mmap_read_unlock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
176
mmap_read_lock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
194
mmap_read_unlock(current->mm);
arch/arm/lib/uaccess_with_memcpy.c
33
pgd = pgd_offset(current->mm, addr);
arch/arm/lib/uaccess_with_memcpy.c
60
ptl = &current->mm->page_table_lock;
arch/arm/lib/uaccess_with_memcpy.c
76
pte = pte_offset_map_lock(current->mm, pmd, addr, &ptl);
arch/arm/mach-rpc/ecard.c
248
struct mm_struct *active_mm = current->active_mm;
arch/arm/mach-rpc/ecard.c
253
current->mm = mm;
arch/arm/mach-rpc/ecard.c
254
current->active_mm = mm;
arch/arm/mm/alignment.c
975
"Address=0x%08lx FSR 0x%03x\n", current->comm,
arch/arm/mm/alignment.c
976
task_pid_nr(current), instrptr,
arch/arm/mm/fault.c
199
struct task_struct *tsk = current;
arch/arm/mm/fault.c
226
struct task_struct *tsk = current;
arch/arm/mm/fault.c
295
struct mm_struct *mm = current->mm;
arch/arm/mm/fault.c
566
do_kernel_address_page_fault(current->mm, addr, fsr, regs);
arch/arm/mm/fault.c
651
show_pte(KERN_ALERT, current->mm, addr);
arch/arm/mm/flush.c
240
struct mm_struct *mm = current->active_mm;
arch/arm/mm/init.c
382
set_section_perms(perms, n, true, current->active_mm);
arch/arm/mm/ioremap.c
207
check_vmalloc_seq(current->active_mm);
arch/arm/mm/mmap.c
34
struct mm_struct *mm = current->mm;
arch/arm/mm/mmap.c
86
struct mm_struct *mm = current->mm;
arch/arm/nwfpe/fpmodule.c
139
current->comm, current->pid, flags,
arch/arm/nwfpe/fpmodule.c
166
fp_send_sig(SIGFPE, current, 1);
arch/arm/probes/uprobes/core.c
135
struct uprobe_task *utask = current->utask;
arch/arm/probes/uprobes/core.c
140
utask->autask.saved_trap_no = current->thread.trap_no;
arch/arm/probes/uprobes/core.c
141
current->thread.trap_no = UPROBE_TRAP_NR;
arch/arm/probes/uprobes/core.c
149
struct uprobe_task *utask = current->utask;
arch/arm/probes/uprobes/core.c
151
WARN_ON_ONCE(current->thread.trap_no != UPROBE_TRAP_NR);
arch/arm/probes/uprobes/core.c
153
current->thread.trap_no = utask->autask.saved_trap_no;
arch/arm/probes/uprobes/core.c
172
struct uprobe_task *utask = current->utask;
arch/arm/probes/uprobes/core.c
174
current->thread.trap_no = utask->autask.saved_trap_no;
arch/arm/vfp/vfpmodule.c
248
current->thread.error_code = 0;
arch/arm/vfp/vfpmodule.c
249
current->thread.trap_no = 6;
arch/arm/vfp/vfpmodule.c
253
current);
arch/arm64/include/asm/compat.h
86
#define compat_user_stack_pointer() (user_stack_pointer(task_pt_regs(current)))
arch/arm64/include/asm/efi.h
126
if (mm != current->active_mm) {
arch/arm64/include/asm/efi.h
131
update_saved_ttbr0(current, mm);
arch/arm64/include/asm/efi.h
137
update_saved_ttbr0(current, current->active_mm);
arch/arm64/include/asm/efi.h
49
on_task_stack(current, READ_ONCE(efi_rt_stack_top[-1]), 1))
arch/arm64/include/asm/elf.h
164
current->personality &= ~READ_IMPLIES_EXEC; \
arch/arm64/include/asm/elf.h
171
(elf_addr_t)current->mm->context.vdso); \
arch/arm64/include/asm/elf.h
238
(Elf64_Off)current->mm->context.vdso); \
arch/arm64/include/asm/entry-common.h
21
send_sig_fault(SIGSEGV, SEGV_MTEAERR, (void __user *)NULL, current);
arch/arm64/include/asm/mmu_context.h
94
struct mm_struct *mm = current->active_mm;
arch/arm64/include/asm/pgtable.h
421
if (mm != current->active_mm && atomic_read(&mm->mm_users) <= 1)
arch/arm64/include/asm/pointer_auth.h
117
ptrauth_keys_install_user(&current->thread.keys_user)
arch/arm64/include/asm/pointer_auth.h
121
ptrauth_keys_init_user(&current->thread.keys_user); \
arch/arm64/include/asm/pointer_auth.h
125
ptrauth_set_enabled_keys(current, \
arch/arm64/include/asm/processor.h
340
spectre_v4_enable_task_mitigation(current);
arch/arm64/include/asm/processor.h
355
spectre_v4_enable_task_mitigation(current);
arch/arm64/include/asm/processor.h
436
#define SET_TAGGED_ADDR_CTRL(arg) set_tagged_addr_ctrl(current, arg)
arch/arm64/include/asm/processor.h
437
#define GET_TAGGED_ADDR_CTRL() get_tagged_addr_ctrl(current)
arch/arm64/include/asm/stackprotector.h
31
current->stack_canary = canary;
arch/arm64/include/asm/stackprotector.h
33
__stack_chk_guard = current->stack_canary;
arch/arm64/include/asm/stackprotector.h
35
ptrauth_thread_init_kernel(current);
arch/arm64/include/asm/stackprotector.h
36
ptrauth_thread_switch_kernel(current);
arch/arm64/include/asm/stacktrace.h
60
#define on_thread_stack() (on_task_stack(current, current_stack_pointer, 1))
arch/arm64/include/asm/uaccess.h
46
(current->flags & PF_KTHREAD || test_thread_flag(TIF_TAGGED_ADDR)))
arch/arm64/kernel/armv8_deprecated.c
150
if (likely(res != -EAGAIN) || signal_pending(current))
arch/arm64/kernel/armv8_deprecated.c
219
current->comm, (unsigned long)current->pid, regs->pc);
arch/arm64/kernel/armv8_deprecated.c
299
current->comm, (unsigned long)current->pid, regs->pc);
arch/arm64/kernel/armv8_deprecated.c
365
current->comm, (unsigned long)current->pid, regs->pc);
arch/arm64/kernel/cpuinfo.c
218
bool compat = personality(current->personality) == PER_LINUX32;
arch/arm64/kernel/debug-monitors.c
193
user_rewind_single_step(current);
arch/arm64/kernel/efi.c
173
if (preemptible() && (current->flags & PF_KTHREAD)) {
arch/arm64/kernel/efi.c
210
if (preemptible() && (current->flags & PF_KTHREAD)) {
arch/arm64/kernel/entry-common.c
237
sve_vq_minus_one = sve_vq_from_vl(task_get_sve_vl(current)) - 1;
arch/arm64/kernel/entry-common.c
44
mte_disable_tco_entry(current);
arch/arm64/kernel/entry-common.c
69
mte_disable_tco_entry(current);
arch/arm64/kernel/fpsimd.c
1308
sve_vq_from_vl(task_get_sve_vl(current)) - 1;
arch/arm64/kernel/fpsimd.c
1313
fpsimd_to_sve(current);
arch/arm64/kernel/fpsimd.c
1314
current->thread.fp_type = FP_STATE_SVE;
arch/arm64/kernel/fpsimd.c
1315
fpsimd_flush_task_state(current);
arch/arm64/kernel/fpsimd.c
1338
sve_alloc(current, true);
arch/arm64/kernel/fpsimd.c
1339
if (!current->thread.sve_state) {
arch/arm64/kernel/fpsimd.c
1389
sve_alloc(current, false);
arch/arm64/kernel/fpsimd.c
1390
sme_alloc(current, true);
arch/arm64/kernel/fpsimd.c
1391
if (!current->thread.sve_state || !current->thread.sme_state) {
arch/arm64/kernel/fpsimd.c
1404
sve_vq_from_vl(task_get_sme_vl(current)) - 1;
arch/arm64/kernel/fpsimd.c
1409
fpsimd_flush_task_state(current);
arch/arm64/kernel/fpsimd.c
1455
current);
arch/arm64/kernel/fpsimd.c
1520
fpsimd_save_kernel_state(current);
arch/arm64/kernel/fpsimd.c
1557
vl = task_get_vl_onexec(current, type);
arch/arm64/kernel/fpsimd.c
1568
task_set_vl(current, type, vl);
arch/arm64/kernel/fpsimd.c
1575
task_set_vl_onexec(current, type, 0);
arch/arm64/kernel/fpsimd.c
1588
fpsimd_flush_task_state(current);
arch/arm64/kernel/fpsimd.c
1589
memset(&current->thread.uw.fpsimd_state, 0,
arch/arm64/kernel/fpsimd.c
1590
sizeof(current->thread.uw.fpsimd_state));
arch/arm64/kernel/fpsimd.c
1596
sve_state = current->thread.sve_state;
arch/arm64/kernel/fpsimd.c
1597
current->thread.sve_state = NULL;
arch/arm64/kernel/fpsimd.c
1606
sme_state = current->thread.sme_state;
arch/arm64/kernel/fpsimd.c
1607
current->thread.sme_state = NULL;
arch/arm64/kernel/fpsimd.c
1610
current->thread.svcr = 0;
arch/arm64/kernel/fpsimd.c
1614
current->thread.uw.fpmr = 0;
arch/arm64/kernel/fpsimd.c
1616
current->thread.fp_type = FP_STATE_FPSIMD;
arch/arm64/kernel/fpsimd.c
1647
last->st = &current->thread.uw.fpsimd_state;
arch/arm64/kernel/fpsimd.c
1648
last->sve_state = current->thread.sve_state;
arch/arm64/kernel/fpsimd.c
1649
last->sme_state = current->thread.sme_state;
arch/arm64/kernel/fpsimd.c
1650
last->sve_vl = task_get_sve_vl(current);
arch/arm64/kernel/fpsimd.c
1651
last->sme_vl = task_get_sme_vl(current);
arch/arm64/kernel/fpsimd.c
1652
last->svcr = &current->thread.svcr;
arch/arm64/kernel/fpsimd.c
1653
last->fpmr = &current->thread.uw.fpmr;
arch/arm64/kernel/fpsimd.c
1654
last->fp_type = &current->thread.fp_type;
arch/arm64/kernel/fpsimd.c
1656
current->thread.fpsimd_cpu = smp_processor_id();
arch/arm64/kernel/fpsimd.c
1734
current->thread.uw.fpsimd_state = *state;
arch/arm64/kernel/fpsimd.c
1735
if (current->thread.fp_type == FP_STATE_SVE)
arch/arm64/kernel/fpsimd.c
1736
fpsimd_to_sve(current);
arch/arm64/kernel/fpsimd.c
1774
fpsimd_flush_task_state(current);
arch/arm64/kernel/fpsimd.c
1860
WARN_ON(current->thread.kernel_fpsimd_state != NULL);
arch/arm64/kernel/fpsimd.c
1861
current->thread.kernel_fpsimd_state = state;
arch/arm64/kernel/fpsimd.c
1902
WARN_ON(current->thread.kernel_fpsimd_state != state);
arch/arm64/kernel/fpsimd.c
1903
current->thread.kernel_fpsimd_state = NULL;
arch/arm64/kernel/fpsimd.c
369
switch (current->thread.fp_type) {
arch/arm64/kernel/fpsimd.c
375
if (!thread_sm_enabled(&current->thread))
arch/arm64/kernel/fpsimd.c
379
sve_set_vq(sve_vq_from_vl(task_get_sve_vl(current)) - 1);
arch/arm64/kernel/fpsimd.c
401
unsigned long sme_vl = task_get_sme_vl(current);
arch/arm64/kernel/fpsimd.c
407
write_sysreg_s(current->thread.svcr, SYS_SVCR);
arch/arm64/kernel/fpsimd.c
409
if (thread_za_enabled(&current->thread))
arch/arm64/kernel/fpsimd.c
410
sme_load_state(current->thread.sme_state,
arch/arm64/kernel/fpsimd.c
413
if (thread_sm_enabled(&current->thread))
arch/arm64/kernel/fpsimd.c
418
write_sysreg_s(current->thread.uw.fpmr, SYS_FPMR);
arch/arm64/kernel/fpsimd.c
421
WARN_ON_ONCE(current->thread.fp_type != FP_STATE_SVE);
arch/arm64/kernel/fpsimd.c
422
sve_load_state(sve_pffr(&current->thread),
arch/arm64/kernel/fpsimd.c
423
&current->thread.uw.fpsimd_state.fpsr,
arch/arm64/kernel/fpsimd.c
426
WARN_ON_ONCE(current->thread.fp_type != FP_STATE_FPSIMD);
arch/arm64/kernel/fpsimd.c
427
fpsimd_load_state(&current->thread.uw.fpsimd_state);
arch/arm64/kernel/fpsimd.c
832
if (task == current)
arch/arm64/kernel/fpsimd.c
914
ret = task_get_vl_onexec(current, type);
arch/arm64/kernel/fpsimd.c
916
ret = task_get_vl(current, type);
arch/arm64/kernel/fpsimd.c
936
ret = vec_set_vector_length(current, ARM64_VEC_SVE, vl, flags);
arch/arm64/kernel/fpsimd.c
965
ret = vec_set_vector_length(current, ARM64_VEC_SME, vl, flags);
arch/arm64/kernel/ftrace.c
533
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/arm64/kernel/ftrace.c
558
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/arm64/kernel/hw_breakpoint.c
229
struct debug_info *debug_info = &current->thread.debug;
arch/arm64/kernel/hw_breakpoint.c
633
debug_info = &current->thread.debug;
arch/arm64/kernel/hw_breakpoint.c
678
user_enable_single_step(current);
arch/arm64/kernel/hw_breakpoint.c
764
debug_info = &current->thread.debug;
arch/arm64/kernel/hw_breakpoint.c
826
user_enable_single_step(current);
arch/arm64/kernel/hw_breakpoint.c
849
struct debug_info *debug_info = &current->thread.debug;
arch/arm64/kernel/hw_breakpoint.c
877
user_disable_single_step(current);
arch/arm64/kernel/hw_breakpoint.c
917
current_debug_info = &current->thread.debug;
arch/arm64/kernel/mte.c
280
set_mte_ctrl(current, 0);
arch/arm64/kernel/mte.c
400
if (task == current) {
arch/arm64/kernel/mte.c
534
if (!tsk->ptrace || (current != tsk->parent) ||
arch/arm64/kernel/perf_callchain.c
39
arch_stack_walk(callchain_trace, entry, current, regs);
arch/arm64/kernel/perf_regs.c
104
regs_user->regs = task_pt_regs(current);
arch/arm64/kernel/perf_regs.c
105
regs_user->abi = perf_reg_abi(current);
arch/arm64/kernel/perf_regs.c
23
return (task_get_sve_vl(current) * 8) / 64;
arch/arm64/kernel/pointer_auth.c
86
if (tsk == current)
arch/arm64/kernel/probes/simulate-insn.c
165
if (user_mode(regs) && task_gcs_el0_enabled(current)) {
arch/arm64/kernel/probes/simulate-insn.c
57
if (user_mode(regs) && task_gcs_el0_enabled(current)) {
arch/arm64/kernel/probes/uprobes.c
128
struct uprobe_task *utask = current->utask;
arch/arm64/kernel/probes/uprobes.c
136
user_disable_single_step(current);
arch/arm64/kernel/probes/uprobes.c
166
if (task_gcs_el0_enabled(current)) {
arch/arm64/kernel/probes/uprobes.c
218
struct uprobe_task *utask = current->utask;
arch/arm64/kernel/probes/uprobes.c
73
struct uprobe_task *utask = current->utask;
arch/arm64/kernel/probes/uprobes.c
76
current->thread.fault_code = UPROBE_INV_FAULT_CODE;
arch/arm64/kernel/probes/uprobes.c
81
user_enable_single_step(current);
arch/arm64/kernel/probes/uprobes.c
88
struct uprobe_task *utask = current->utask;
arch/arm64/kernel/probes/uprobes.c
90
WARN_ON_ONCE(current->thread.fault_code != UPROBE_INV_FAULT_CODE);
arch/arm64/kernel/probes/uprobes.c
95
user_disable_single_step(current);
arch/arm64/kernel/process.c
258
current->thread.uw.tp_value = 0;
arch/arm64/kernel/process.c
291
current->thread.gcspr_el0 = 0;
arch/arm64/kernel/process.c
292
current->thread.gcs_base = 0;
arch/arm64/kernel/process.c
293
current->thread.gcs_size = 0;
arch/arm64/kernel/process.c
294
current->thread.gcs_el0_mode = 0;
arch/arm64/kernel/process.c
295
current->thread.gcs_el0_locked = 0;
arch/arm64/kernel/process.c
311
p->thread.gcs_el0_mode = current->thread.gcs_el0_mode;
arch/arm64/kernel/process.c
312
p->thread.gcs_el0_locked = current->thread.gcs_el0_locked;
arch/arm64/kernel/process.c
336
flush_ptrace_hw_breakpoint(current);
arch/arm64/kernel/process.c
471
ret = copy_thread_za(p, current);
arch/arm64/kernel/process.c
523
*task_user_tls(current) = read_sysreg(tpidr_el0);
arch/arm64/kernel/process.c
525
current->thread.tpidr2_el0 = read_sysreg_s(SYS_TPIDR2_EL0);
arch/arm64/kernel/process.c
583
current->thread.gcspr_el0 = read_sysreg_s(SYS_GCSPR_EL0);
arch/arm64/kernel/process.c
595
if (current->thread.gcs_el0_mode != next->thread.gcs_el0_mode)
arch/arm64/kernel/process.c
607
if (task_gcs_el0_enabled(current) || task_gcs_el0_enabled(next))
arch/arm64/kernel/process.c
662
update_cntkctl_el1(current);
arch/arm64/kernel/process.c
673
current->thread.por_el0 = read_sysreg_s(SYS_POR_EL0);
arch/arm64/kernel/process.c
674
if (current->thread.por_el0 != next->thread.por_el0) {
arch/arm64/kernel/process.c
782
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/arm64/kernel/process.c
805
!dl_task_check_affinity(current, system_32bit_el0_cpumask());
arch/arm64/kernel/process.c
829
force_compatible_cpus_allowed_ptr(current);
arch/arm64/kernel/process.c
831
relax_compatible_cpus_allowed_ptr(current);
arch/arm64/kernel/process.c
834
current->mm->context.flags = mmflags;
arch/arm64/kernel/process.c
839
if (task_spec_ssb_noexec(current)) {
arch/arm64/kernel/process.c
840
arch_prctl_spec_ctrl_set(current, PR_SPEC_STORE_BYPASS,
arch/arm64/kernel/ptrace.c
1109
if (target == current)
arch/arm64/kernel/ptrace.c
1487
if (target == current)
arch/arm64/kernel/ptrace.c
1488
current->thread.por_el0 = read_sysreg_s(SYS_POR_EL0);
arch/arm64/kernel/ptrace.c
1543
if (target == current)
arch/arm64/kernel/ptrace.c
183
if (current->thread.debug.hbp_break[i] == bp) {
arch/arm64/kernel/ptrace.c
190
if (current->thread.debug.hbp_watch[i] == bp) {
arch/arm64/kernel/ptrace.c
1904
if (target == current)
arch/arm64/kernel/ptrace.c
2442
trace_sys_exit(regs, syscall_get_return_value(current, regs));
arch/arm64/kernel/ptrace.c
610
if (target == current)
arch/arm64/kernel/ptrace.c
667
if (target == current)
arch/arm64/kernel/ptrace.c
707
if (target == current)
arch/arm64/kernel/ptrace.c
823
if (target == current)
arch/arm64/kernel/ptrace.c
986
memset(&current->thread.uw.fpsimd_state, 0,
arch/arm64/kernel/ptrace.c
987
sizeof(current->thread.uw.fpsimd_state));
arch/arm64/kernel/return_address.c
42
arch_stack_walk(save_return_addr, &data, current, NULL);
arch/arm64/kernel/signal.c
1007
err |= !valid_user_regs(&regs->user_regs, current);
arch/arm64/kernel/signal.c
1051
if (!(current->thread.gcs_el0_mode & PR_SHADOW_STACK_ENABLE))
arch/arm64/kernel/signal.c
1101
current->restart_block.fn = do_no_restart_syscall;
arch/arm64/kernel/signal.c
1153
if (add_all || current->thread.fault_code) {
arch/arm64/kernel/signal.c
1161
if (system_supports_gcs() && (add_all || current->thread.gcspr_el0)) {
arch/arm64/kernel/signal.c
1172
if (add_all || current->thread.fp_type == FP_STATE_SVE ||
arch/arm64/kernel/signal.c
1173
thread_sm_enabled(&current->thread)) {
arch/arm64/kernel/signal.c
1177
vl = thread_get_cur_vl(&current->thread);
arch/arm64/kernel/signal.c
1202
vl = task_get_sme_vl(current);
arch/arm64/kernel/signal.c
1204
if (thread_za_enabled(&current->thread))
arch/arm64/kernel/signal.c
1214
if (add_all || thread_za_enabled(&current->thread)) {
arch/arm64/kernel/signal.c
1257
__put_user_error(current->thread.fault_address, &sf->uc.uc_mcontext.fault_address, err);
arch/arm64/kernel/signal.c
1274
__put_user_error(current->thread.fault_code, &esr_ctx->esr, err);
arch/arm64/kernel/signal.c
1410
if (!task_gcs_el0_enabled(current))
arch/arm64/kernel/signal.c
1452
sigtramp = VDSO_SYMBOL(current->mm->context.vdso, sigtramp);
arch/arm64/kernel/signal.c
1497
task_smstop_sm(current);
arch/arm64/kernel/signal.c
1498
current->thread.svcr &= ~SVCR_ZA_MASK;
arch/arm64/kernel/signal.c
1579
ret |= !valid_user_regs(&regs->user_regs, current);
arch/arm64/kernel/signal.c
1644
syscall_set_return_value(current, regs, -EINTR, 0);
arch/arm64/kernel/signal.c
1659
user_rewind_single_step(current);
arch/arm64/kernel/signal.c
254
&current->thread.uw.fpsimd_state;
arch/arm64/kernel/signal.c
257
fpsimd_sync_from_effective_state(current);
arch/arm64/kernel/signal.c
299
current->thread.svcr &= ~SVCR_SM_MASK;
arch/arm64/kernel/signal.c
300
current->thread.fp_type = FP_STATE_FPSIMD;
arch/arm64/kernel/signal.c
313
__put_user_error(current->thread.uw.fpmr, &ctx->fpmr, err);
arch/arm64/kernel/signal.c
328
current->thread.uw.fpmr = fpmr;
arch/arm64/kernel/signal.c
368
unsigned int vl = task_get_sve_vl(current);
arch/arm64/kernel/signal.c
371
if (thread_sm_enabled(&current->thread)) {
arch/arm64/kernel/signal.c
372
vl = task_get_sme_vl(current);
arch/arm64/kernel/signal.c
375
} else if (current->thread.fp_type == FP_STATE_SVE) {
arch/arm64/kernel/signal.c
391
current->thread.sve_state,
arch/arm64/kernel/signal.c
419
vl = task_get_sme_vl(current);
arch/arm64/kernel/signal.c
429
vl = task_get_sve_vl(current);
arch/arm64/kernel/signal.c
453
sme_alloc(current, false);
arch/arm64/kernel/signal.c
454
if (!current->thread.sme_state)
arch/arm64/kernel/signal.c
458
sve_alloc(current, true);
arch/arm64/kernel/signal.c
459
if (!current->thread.sve_state) {
arch/arm64/kernel/signal.c
465
current->thread.svcr |= SVCR_SM_MASK;
arch/arm64/kernel/signal.c
468
current->thread.svcr &= ~SVCR_SM_MASK;
arch/arm64/kernel/signal.c
472
current->thread.fp_type = FP_STATE_SVE;
arch/arm64/kernel/signal.c
474
err = __copy_from_user(current->thread.sve_state,
arch/arm64/kernel/signal.c
537
unsigned int vl = task_get_sme_vl(current);
arch/arm64/kernel/signal.c
540
if (thread_za_enabled(&current->thread))
arch/arm64/kernel/signal.c
556
current->thread.sme_state,
arch/arm64/kernel/signal.c
576
if (user_vl != task_get_sme_vl(current))
arch/arm64/kernel/signal.c
580
current->thread.svcr &= ~SVCR_ZA_MASK;
arch/arm64/kernel/signal.c
589
sve_alloc(current, false);
arch/arm64/kernel/signal.c
590
if (!current->thread.sve_state)
arch/arm64/kernel/signal.c
593
sme_alloc(current, true);
arch/arm64/kernel/signal.c
594
if (!current->thread.sme_state) {
arch/arm64/kernel/signal.c
595
current->thread.svcr &= ~SVCR_ZA_MASK;
arch/arm64/kernel/signal.c
600
err = __copy_from_user(current->thread.sme_state,
arch/arm64/kernel/signal.c
608
current->thread.svcr |= SVCR_ZA_MASK;
arch/arm64/kernel/signal.c
618
if (WARN_ON(!thread_za_enabled(&current->thread)))
arch/arm64/kernel/signal.c
631
thread_zt_state(&current->thread),
arch/arm64/kernel/signal.c
643
if (!thread_za_enabled(&current->thread))
arch/arm64/kernel/signal.c
655
err = __copy_from_user(thread_zt_state(&current->thread),
arch/arm64/kernel/signal.c
691
if (task_gcs_el0_enabled(current))
arch/arm64/kernel/signal.c
698
__put_user_error(current->thread.gcs_el0_mode,
arch/arm64/kernel/signal.c
721
err = gcs_check_locked(current, enabled);
arch/arm64/kernel/signal.c
726
if (!task_gcs_el0_enabled(current) &&
arch/arm64/kernel/signal.c
734
current->thread.gcs_el0_mode = enabled;
arch/arm64/kernel/signal32.c
180
current->thread.uw.fpsimd_state = fpsimd;
arch/arm64/kernel/signal32.c
222
err |= !valid_user_regs(&regs->user_regs, current);
arch/arm64/kernel/signal32.c
237
current->restart_block.fn = do_no_restart_syscall;
arch/arm64/kernel/signal32.c
268
current->restart_block.fn = do_no_restart_syscall;
arch/arm64/kernel/signal32.c
349
retcode = (unsigned long)current->mm->context.sigpage +
arch/arm64/kernel/signal32.c
387
__put_user_error(!!(current->thread.fault_code & ESR_ELx_WNR) <<
arch/arm64/kernel/signal32.c
389
__put_user_error(current->thread.fault_address, &sf->uc.uc_mcontext.fault_address, err);
arch/arm64/kernel/signal32.c
95
&current->thread.uw.fpsimd_state;
arch/arm64/kernel/smp.c
207
current->active_mm = mm;
arch/arm64/kernel/stacktrace.c
102
kunwind_init(state, current);
arch/arm64/kernel/stacktrace.c
141
WARN_ON_ONCE(state->task == current);
arch/arm64/kernel/stacktrace.c
206
WARN_ON_ONCE(tsk == current);
arch/arm64/kernel/stacktrace.c
211
WARN_ON_ONCE(tsk == current);
arch/arm64/kernel/stacktrace.c
309
((task == current) && !preemptible()) \
arch/arm64/kernel/stacktrace.c
320
((task == current) && in_nmi()) \
arch/arm64/kernel/stacktrace.c
327
((task == current) && current_in_efi()) \
arch/arm64/kernel/stacktrace.c
357
if (task != current)
arch/arm64/kernel/stacktrace.c
360
} else if (task == current) {
arch/arm64/kernel/stacktrace.c
445
kunwind_stack_walk(arch_bpf_unwind_consume_entry, &data, current, NULL);
arch/arm64/kernel/stacktrace.c
486
tsk = current;
arch/arm64/kernel/stacktrace.c
83
kunwind_init(state, current);
arch/arm64/kernel/sys_compat.c
31
if (fatal_signal_pending(current))
arch/arm64/kernel/sys_compat.c
91
current->thread.uw.tp_value = regs->regs[0];
arch/arm64/kernel/syscall.c
105
syscall_set_return_value(current, regs, -ERESTARTNOINTR, 0);
arch/arm64/kernel/syscall.c
126
syscall_set_return_value(current, regs, -ENOSYS, 0);
arch/arm64/kernel/syscall.c
54
syscall_set_return_value(current, regs, 0, ret);
arch/arm64/kernel/time.c
49
arch_stack_walk(profile_pc_cb, &prof_pc, current, regs);
arch/arm64/kernel/traps.c
219
if (regs && kexec_should_crash(current))
arch/arm64/kernel/traps.c
241
struct task_struct *tsk = current;
arch/arm64/kernel/traps.c
297
current->thread.fault_address = 0;
arch/arm64/kernel/traps.c
298
current->thread.fault_code = err;
arch/arm64/kernel/traps.c
370
user_fastforward_single_step(current);
arch/arm64/kernel/traps.c
445
mmap_read_lock(current->mm);
arch/arm64/kernel/traps.c
446
if (find_vma(current->mm, untagged_addr(addr)) == NULL)
arch/arm64/kernel/traps.c
450
mmap_read_unlock(current->mm);
arch/arm64/kernel/traps.c
536
user_fastforward_single_step(current);
arch/arm64/kernel/traps.c
893
current->thread.fault_address = 0;
arch/arm64/kernel/traps.c
894
current->thread.fault_code = esr;
arch/arm64/kernel/traps.c
905
unsigned long tsk_stk = (unsigned long)current->stack;
arch/arm64/kernel/vdso.c
160
current->mm->context.sigpage = (void *)new_vma->vm_start;
arch/arm64/kernel/vdso.c
296
struct mm_struct *mm = current->mm;
arch/arm64/kernel/vdso.c
335
struct mm_struct *mm = current->mm;
arch/arm64/kernel/vdso.c
63
current->mm->context.vdso = (void *)new_vma->vm_start;
arch/arm64/kvm/fpsimd.c
26
BUG_ON(!current->mm);
arch/arm64/kvm/mmu.c
1051
vma = find_vma_intersection(current->mm, hva, reg_end);
arch/arm64/kvm/mmu.c
1083
mmap_read_lock(current->mm);
arch/arm64/kvm/mmu.c
1093
mmap_read_unlock(current->mm);
arch/arm64/kvm/mmu.c
1325
send_sig_mceerr(BUS_MCEERR_AR, (void __user *)address, lsb, current);
arch/arm64/kvm/mmu.c
1690
mmap_read_lock(current->mm);
arch/arm64/kvm/mmu.c
1691
vma = vma_lookup(current->mm, hva);
arch/arm64/kvm/mmu.c
1694
mmap_read_unlock(current->mm);
arch/arm64/kvm/mmu.c
1786
mmap_read_unlock(current->mm);
arch/arm64/kvm/mmu.c
2435
mmap_read_lock(current->mm);
arch/arm64/kvm/mmu.c
2450
vma = find_vma_intersection(current->mm, hva, reg_end);
arch/arm64/kvm/mmu.c
2478
mmap_read_unlock(current->mm);
arch/arm64/kvm/mmu.c
895
current->comm);
arch/arm64/kvm/pmu-emul.c
753
event = perf_event_create_kernel_counter(&attr, -1, current,
arch/arm64/kvm/pvtime.c
28
vcpu->arch.steal.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/arm64/kvm/pvtime.c
64
vcpu->arch.steal.last_steal = current->sched_info.run_delay;
arch/arm64/mm/fault.c
136
mm = current->active_mm;
arch/arm64/mm/fault.c
414
current->thread.fault_address = address;
arch/arm64/mm/fault.c
428
if (!is_ttbr0_addr(current->thread.fault_address)) {
arch/arm64/mm/fault.c
466
current->thread.fault_code = esr;
arch/arm64/mm/fault.c
560
struct mm_struct *mm = current->mm;
arch/arm64/mm/gcs.c
16
struct mm_struct *mm = current->mm;
arch/arm64/mm/gcs.c
160
if (!task->mm || task->mm != current->mm)
arch/arm64/mm/gcs.c
197
if (task != current)
arch/arm64/mm/gcs.c
208
if (task == current)
arch/arm64/mm/gcs.c
214
if (task == current)
arch/csky/abiv1/mmap.c
29
struct mm_struct *mm = current->mm;
arch/csky/abiv2/cacheflush.c
64
if (mm == current->mm) {
arch/csky/abiv2/cacheflush.c
86
if (mm != current->active_mm || !cpumask_empty(&others)) {
arch/csky/abiv2/inc/abi/cacheflush.h
53
flush_icache_mm_range(current->mm, \
arch/csky/include/asm/stackprotector.h
17
current->stack_canary = canary;
arch/csky/include/asm/stackprotector.h
18
__stack_chk_guard = current->stack_canary;
arch/csky/kernel/ftrace.c
153
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/csky/kernel/perf_callchain.c
15
unsigned long low = (unsigned long)task_stack_page(current);
arch/csky/kernel/perf_regs.c
37
regs_user->regs = task_pt_regs(current);
arch/csky/kernel/perf_regs.c
38
regs_user->abi = perf_reg_abi(current);
arch/csky/kernel/probes/uprobes.c
103
struct uprobe_task *utask = current->utask;
arch/csky/kernel/probes/uprobes.c
105
current->thread.trap_no = utask->autask.saved_trap_no;
arch/csky/kernel/probes/uprobes.c
113
user_disable_single_step(current);
arch/csky/kernel/probes/uprobes.c
50
struct uprobe_task *utask = current->utask;
arch/csky/kernel/probes/uprobes.c
52
utask->autask.saved_trap_no = current->thread.trap_no;
arch/csky/kernel/probes/uprobes.c
53
current->thread.trap_no = UPROBE_TRAP_NR;
arch/csky/kernel/probes/uprobes.c
57
user_enable_single_step(current);
arch/csky/kernel/probes/uprobes.c
64
struct uprobe_task *utask = current->utask;
arch/csky/kernel/probes/uprobes.c
66
WARN_ON_ONCE(current->thread.trap_no != UPROBE_TRAP_NR);
arch/csky/kernel/probes/uprobes.c
67
current->thread.trap_no = utask->autask.saved_trap_no;
arch/csky/kernel/probes/uprobes.c
71
user_disable_single_step(current);
arch/csky/kernel/process.c
75
memcpy(fpu, &current->thread.user_fp, sizeof(*fpu));
arch/csky/kernel/ptrace.c
330
trace_sys_enter(regs, syscall_get_nr(current, regs));
arch/csky/kernel/ptrace.c
344
trace_sys_exit(regs, syscall_get_return_value(current, regs));
arch/csky/kernel/ptrace.c
468
pr_info("COMM=%s PID=%d\n", current->comm, current->pid);
arch/csky/kernel/ptrace.c
470
if (current->mm) {
arch/csky/kernel/ptrace.c
472
(int) current->mm->start_code,
arch/csky/kernel/ptrace.c
473
(int) current->mm->end_code,
arch/csky/kernel/ptrace.c
474
(int) current->mm->start_data,
arch/csky/kernel/ptrace.c
475
(int) current->mm->end_data,
arch/csky/kernel/ptrace.c
476
(int) current->mm->end_data,
arch/csky/kernel/ptrace.c
477
(int) current->mm->brk);
arch/csky/kernel/ptrace.c
479
(int) current->mm->start_stack,
arch/csky/kernel/ptrace.c
480
(int) (((unsigned long) current) + 2 * PAGE_SIZE));
arch/csky/kernel/signal.c
159
current->mm->context.vdso, rt_sigreturn);
arch/csky/kernel/signal.c
76
current->restart_block.fn = do_no_restart_syscall;
arch/csky/kernel/smp.c
264
current->active_mm = mm;
arch/csky/kernel/stacktrace.c
25
} else if (task == NULL || task == current) {
arch/csky/kernel/stacktrace.c
53
pc = ftrace_graph_ret_addr(current, NULL, frame->ra,
arch/csky/kernel/stacktrace.c
69
} else if (task == NULL || task == current) {
arch/csky/kernel/syscall.c
8
struct thread_info *ti = task_thread_info(current);
arch/csky/kernel/traps.c
118
struct task_struct *tsk = current;
arch/csky/kernel/traps.c
135
current->thread.trap_no = trap_no(regs);
arch/csky/kernel/traps.c
163
current->thread.trap_no = trap_no(regs);
arch/csky/kernel/traps.c
180
send_sig(SIGTRAP, current, 0);
arch/csky/kernel/traps.c
190
current->thread.trap_no = trap_no(regs);
arch/csky/kernel/traps.c
202
send_sig(SIGTRAP, current, 0);
arch/csky/kernel/traps.c
99
show_stack(current, (unsigned long *)regs->regs[4], KERN_INFO);
arch/csky/kernel/vdso.c
44
struct mm_struct *mm = current->mm;
arch/csky/mm/fault.c
197
tsk = current;
arch/csky/mm/fault.c
56
current->thread.trap_no = trap_no(regs);
arch/csky/mm/fault.c
75
current->thread.trap_no = trap_no(regs);
arch/csky/mm/syscache.c
23
flush_icache_mm_range(current->mm,
arch/hexagon/include/asm/pgalloc.h
74
current->active_mm->context.generation = kmap_generation;
arch/hexagon/include/asm/pgalloc.h
84
ppmd = (pmd_t *)current->active_mm->pgd + pmdindex;
arch/hexagon/kernel/signal.c
100
struct hexagon_vdso *vdso = current->mm->context.vdso;
arch/hexagon/kernel/signal.c
230
current->restart_block.fn = do_no_restart_syscall;
arch/hexagon/kernel/smp.c
144
current->active_mm = &init_mm;
arch/hexagon/kernel/stacktrace.c
29
low = (unsigned long)task_stack_page(current);
arch/hexagon/kernel/traps.c
205
do_show_stack(current, &regs->r30, pt_elr(regs), KERN_EMERG);
arch/hexagon/kernel/traps.c
90
task = current;
arch/hexagon/kernel/traps.c
96
if (task == current) {
arch/hexagon/kernel/vdso.c
55
struct mm_struct *mm = current->mm;
arch/hexagon/mm/vm_fault.c
40
struct mm_struct *mm = current->mm;
arch/hexagon/mm/vm_tlb.c
31
if (mm->context.ptbase == current->active_mm->context.ptbase)
arch/hexagon/mm/vm_tlb.c
60
if (current->active_mm->context.ptbase == mm->context.ptbase)
arch/hexagon/mm/vm_tlb.c
71
if (mm->context.ptbase == current->active_mm->context.ptbase)
arch/loongarch/include/asm/elf.h
273
current->thread.vdso = &vdso_info; \
arch/loongarch/include/asm/elf.h
275
if (personality(current->personality) != PER_LINUX) \
arch/loongarch/include/asm/elf.h
290
current->thread.vdso = &vdso_info; \
arch/loongarch/include/asm/elf.h
292
p = personality(current->personality); \
arch/loongarch/include/asm/elf.h
345
(unsigned long)current->mm->context.vdso); \
arch/loongarch/include/asm/fpu.h
113
KSTK_EUEN(current) |= CSR_EUEN_FPEN;
arch/loongarch/include/asm/fpu.h
121
_restore_fp(&current->thread.fpu);
arch/loongarch/include/asm/fpu.h
159
lose_fpu_inatomic(save, current);
arch/loongarch/include/asm/fpu.h
165
unsigned int fcsr = current->thread.fpu.fcsr;
arch/loongarch/include/asm/fpu.h
188
if (tsk == current) {
arch/loongarch/include/asm/fpu.h
195
_save_lasx(&current->thread.fpu);
arch/loongarch/include/asm/fpu.h
200
_save_lsx(&current->thread.fpu);
arch/loongarch/include/asm/fpu.h
204
_save_fp(&current->thread.fpu);
arch/loongarch/include/asm/lbt.h
55
KSTK_EUEN(current) |= CSR_EUEN_LBTEN;
arch/loongarch/include/asm/lbt.h
63
_restore_lbt(&current->thread.lbt);
arch/loongarch/include/asm/lbt.h
89
lose_lbt_inatomic(save, current);
arch/loongarch/include/asm/mmu_context.h
133
#define activate_mm(prev, next) switch_mm(prev, next, current)
arch/loongarch/include/asm/mmu_context.h
153
if (!current->mm || (current->mm == mm)) {
arch/loongarch/include/asm/stackprotector.h
34
current->stack_canary = canary;
arch/loongarch/include/asm/stackprotector.h
35
__stack_chk_guard = current->stack_canary;
arch/loongarch/include/asm/stacktrace.h
36
return !(((unsigned long)(current->stack) ^ current_stack_pointer) & ~(THREAD_SIZE - 1));
arch/loongarch/include/asm/unwind.h
66
} else if (task && task != current) {
arch/loongarch/kernel/ftrace.c
58
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/loongarch/kernel/ftrace_dyn.c
230
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/loongarch/kernel/ftrace_dyn.c
248
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/loongarch/kernel/kfpu.c
54
_save_lasx(&current->thread.fpu);
arch/loongarch/kernel/kfpu.c
59
_save_lsx(&current->thread.fpu);
arch/loongarch/kernel/kfpu.c
63
_save_fp(&current->thread.fpu);
arch/loongarch/kernel/kfpu.c
79
_restore_lasx(&current->thread.fpu);
arch/loongarch/kernel/kfpu.c
84
_restore_lsx(&current->thread.fpu);
arch/loongarch/kernel/kfpu.c
88
_restore_fp(&current->thread.fpu);
arch/loongarch/kernel/kgdb.c
130
save_fp(current);
arch/loongarch/kernel/kgdb.c
135
memcpy(mem, (void *)&current->thread.fpu.fcsr, reg_size);
arch/loongarch/kernel/kgdb.c
138
memcpy(mem, (void *)&current->thread.fpu.fcc + reg_offset, reg_size);
arch/loongarch/kernel/kgdb.c
141
memcpy(mem, (void *)&current->thread.fpu.fpr[reg_offset], reg_size);
arch/loongarch/kernel/kgdb.c
176
memcpy((void *)&current->thread.fpu.fcsr, mem, reg_size);
arch/loongarch/kernel/kgdb.c
179
memcpy((void *)&current->thread.fpu.fcc + reg_offset, mem, reg_size);
arch/loongarch/kernel/kgdb.c
182
memcpy((void *)&current->thread.fpu.fpr[reg_offset], mem, reg_size);
arch/loongarch/kernel/kgdb.c
188
restore_fp(current);
arch/loongarch/kernel/machine_kexec.c
161
regs = task_pt_regs(current);
arch/loongarch/kernel/perf_event.c
79
for (unwind_start(&state, current, regs);
arch/loongarch/kernel/perf_regs.c
51
regs_user->regs = task_pt_regs(current);
arch/loongarch/kernel/perf_regs.c
52
regs_user->abi = perf_reg_abi(current);
arch/loongarch/kernel/process.c
104
flush_ptrace_hw_breakpoint(current);
arch/loongarch/kernel/process.c
124
save_lasx(current);
arch/loongarch/kernel/process.c
126
save_lsx(current);
arch/loongarch/kernel/process.c
128
save_fp(current);
arch/loongarch/kernel/process.c
299
task = task ? : current;
arch/loongarch/kernel/process.c
307
if (task != current)
arch/loongarch/kernel/process.c
322
if (current->thread.vdso) {
arch/loongarch/kernel/process.c
324
top -= PAGE_ALIGN(current->thread.vdso->size);
arch/loongarch/kernel/process.c
328
if (current->flags & PF_RANDOMIZE)
arch/loongarch/kernel/process.c
341
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/loongarch/kernel/process.c
92
current->thread.fpu.fcsr = boot_cpu_data.fpu_csr0;
arch/loongarch/kernel/ptrace.c
395
if (current->thread.hbp_break[i] == bp)
arch/loongarch/kernel/ptrace.c
399
if (current->thread.hbp_watch[i] == bp)
arch/loongarch/kernel/signal.c
1012
ret = setup_rt_frame(vdso + current->thread.vdso->offset_sigreturn, ksig, regs, oldset);
arch/loongarch/kernel/signal.c
109
set_fpr64(&current->thread.fpu.fpr[i], 0, fpr_val);
arch/loongarch/kernel/signal.c
111
err |= __get_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
112
err |= __get_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
126
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 0),
arch/loongarch/kernel/signal.c
128
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 1),
arch/loongarch/kernel/signal.c
131
err |= __put_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
132
err |= __put_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
148
set_fpr64(&current->thread.fpu.fpr[i], 0, fpr_val);
arch/loongarch/kernel/signal.c
150
set_fpr64(&current->thread.fpu.fpr[i], 1, fpr_val);
arch/loongarch/kernel/signal.c
152
err |= __get_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
153
err |= __get_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
167
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 0),
arch/loongarch/kernel/signal.c
169
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 1),
arch/loongarch/kernel/signal.c
171
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 2),
arch/loongarch/kernel/signal.c
173
err |= __put_user(get_fpr64(&current->thread.fpu.fpr[i], 3),
arch/loongarch/kernel/signal.c
176
err |= __put_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
177
err |= __put_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
193
set_fpr64(&current->thread.fpu.fpr[i], 0, fpr_val);
arch/loongarch/kernel/signal.c
195
set_fpr64(&current->thread.fpu.fpr[i], 1, fpr_val);
arch/loongarch/kernel/signal.c
197
set_fpr64(&current->thread.fpu.fpr[i], 2, fpr_val);
arch/loongarch/kernel/signal.c
199
set_fpr64(&current->thread.fpu.fpr[i], 3, fpr_val);
arch/loongarch/kernel/signal.c
201
err |= __get_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
202
err |= __get_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
214
err |= __put_user(current->thread.lbt.scr0, &regs[0]);
arch/loongarch/kernel/signal.c
215
err |= __put_user(current->thread.lbt.scr1, &regs[1]);
arch/loongarch/kernel/signal.c
216
err |= __put_user(current->thread.lbt.scr2, &regs[2]);
arch/loongarch/kernel/signal.c
217
err |= __put_user(current->thread.lbt.scr3, &regs[3]);
arch/loongarch/kernel/signal.c
218
err |= __put_user(current->thread.lbt.eflags, eflags);
arch/loongarch/kernel/signal.c
229
err |= __get_user(current->thread.lbt.scr0, &regs[0]);
arch/loongarch/kernel/signal.c
230
err |= __get_user(current->thread.lbt.scr1, &regs[1]);
arch/loongarch/kernel/signal.c
231
err |= __get_user(current->thread.lbt.scr2, &regs[2]);
arch/loongarch/kernel/signal.c
232
err |= __get_user(current->thread.lbt.scr3, &regs[3]);
arch/loongarch/kernel/signal.c
233
err |= __get_user(current->thread.lbt.eflags, eflags);
arch/loongarch/kernel/signal.c
242
return __put_user(current->thread.fpu.ftop, ftop);
arch/loongarch/kernel/signal.c
249
return __get_user(current->thread.fpu.ftop, ftop);
arch/loongarch/kernel/signal.c
451
save_fp(current);
arch/loongarch/kernel/signal.c
493
restore_fp(current);
arch/loongarch/kernel/signal.c
527
save_lsx(current);
arch/loongarch/kernel/signal.c
529
save_fp(current);
arch/loongarch/kernel/signal.c
572
restore_lsx(current);
arch/loongarch/kernel/signal.c
574
restore_fp(current);
arch/loongarch/kernel/signal.c
779
current->restart_block.fn = do_no_restart_syscall;
arch/loongarch/kernel/signal.c
807
switch (current->thread.error_code) {
arch/loongarch/kernel/signal.c
89
__put_user(get_fpr64(&current->thread.fpu.fpr[i], 0),
arch/loongarch/kernel/signal.c
92
err |= __put_user(current->thread.fpu.fcc, fcc);
arch/loongarch/kernel/signal.c
93
err |= __put_user(current->thread.fpu.fcsr, fcsr);
arch/loongarch/kernel/signal.c
977
current->comm, current->pid,
arch/loongarch/kernel/signal.c
987
void *vdso = current->mm->context.vdso;
arch/loongarch/kernel/smp.c
624
init_new_context(current, &init_mm);
arch/loongarch/kernel/smp.c
745
if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/loongarch/kernel/smp.c
778
if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/loongarch/kernel/smp.c
825
if ((atomic_read(&vma->vm_mm->mm_users) != 1) || (current->mm != vma->vm_mm)) {
arch/loongarch/kernel/stacktrace.c
24
if (task == current) {
arch/loongarch/kernel/stacktrace.c
51
if (task == current) {
arch/loongarch/kernel/traps.c
1067
read_csr_excode(), current->pid, current->comm);
arch/loongarch/kernel/traps.c
1158
current->active_mm = &init_mm;
arch/loongarch/kernel/traps.c
1159
BUG_ON(current->mm);
arch/loongarch/kernel/traps.c
1160
enter_lazy_tlb(&init_mm, current);
arch/loongarch/kernel/traps.c
144
if (!task || task == current)
arch/loongarch/kernel/traps.c
391
current->comm, current->pid, current_thread_info(), current);
arch/loongarch/kernel/traps.c
393
show_stacktrace(current, regs, KERN_DEFAULT, user_mode(regs));
arch/loongarch/kernel/traps.c
408
current->thread.trap_nr, SIGSEGV);
arch/loongarch/kernel/traps.c
424
if (regs && kexec_should_crash(current))
arch/loongarch/kernel/traps.c
482
force_fcsr_sig(fcsr, fault_addr, current);
arch/loongarch/kernel/traps.c
490
mmap_read_lock(current->mm);
arch/loongarch/kernel/traps.c
491
if (vma_lookup(current->mm, (unsigned long)fault_addr))
arch/loongarch/kernel/traps.c
495
mmap_read_unlock(current->mm);
arch/loongarch/kernel/traps.c
514
if (notify_die(DIE_FP, "FP exception", regs, 0, current->thread.trap_nr,
arch/loongarch/kernel/traps.c
640
current->thread.trap_nr = read_csr_excode();
arch/loongarch/kernel/traps.c
765
current->thread.trap_nr = read_csr_excode();
arch/loongarch/kernel/traps.c
767
current->thread.trap_nr, SIGTRAP) == NOTIFY_STOP)
arch/loongarch/kernel/traps.c
813
if (test_tsk_thread_flag(current, TIF_SINGLESTEP)) {
arch/loongarch/kernel/traps.c
829
if (pc == current->thread.single_step) {
arch/loongarch/kernel/traps.c
864
current->thread.trap_nr = read_csr_excode();
arch/loongarch/kernel/traps.c
866
if (notify_die(DIE_RI, "RI Fault", regs, 0, current->thread.trap_nr,
arch/loongarch/kernel/traps.c
874
current->thread.error_code = 1;
arch/loongarch/kernel/traps.c
91
task = current;
arch/loongarch/kernel/traps.c
910
restore_lsx_upper(current);
arch/loongarch/kernel/traps.c
913
restore_lsx(current);
arch/loongarch/kernel/traps.c
936
restore_lasx_upper(current);
arch/loongarch/kernel/traps.c
940
restore_lasx(current);
arch/loongarch/kernel/unaligned.c
454
set_fpr64(&current->thread.fpu.fpr[insn.reg3_format.rd], 0, value);
arch/loongarch/kernel/unaligned.c
464
value = get_fpr64(&current->thread.fpu.fpr[insn.reg3_format.rd], 0);
arch/loongarch/kernel/uprobes.c
40
struct uprobe_task *utask = current->utask;
arch/loongarch/kernel/uprobes.c
42
utask->autask.saved_trap_nr = current->thread.trap_nr;
arch/loongarch/kernel/uprobes.c
43
current->thread.trap_nr = UPROBE_TRAP_NR;
arch/loongarch/kernel/uprobes.c
51
struct uprobe_task *utask = current->utask;
arch/loongarch/kernel/uprobes.c
53
WARN_ON_ONCE(current->thread.trap_nr != UPROBE_TRAP_NR);
arch/loongarch/kernel/uprobes.c
54
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/loongarch/kernel/uprobes.c
62
struct uprobe_task *utask = current->utask;
arch/loongarch/kernel/uprobes.c
64
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/loongarch/kernel/vdso.c
30
current->mm->context.vdso = (void *)(new_vma->vm_start);
arch/loongarch/kernel/vdso.c
72
if (current->flags & PF_RANDOMIZE) {
arch/loongarch/kernel/vdso.c
84
struct mm_struct *mm = current->mm;
arch/loongarch/kernel/vdso.c
86
struct loongarch_vdso_info *info = current->thread.vdso;
arch/loongarch/kvm/exit.c
778
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/kvm/vcpu.c
1210
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/kvm/vcpu.c
1822
if (unlikely(current->mm != vcpu->kvm->mm))
arch/loongarch/kvm/vcpu.c
198
steal += current->sched_info.run_delay - vcpu->arch.st.last_steal;
arch/loongarch/kvm/vcpu.c
199
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/loongarch/mm/fault.c
127
current->thread.csr_badvaddr = address;
arch/loongarch/mm/fault.c
128
current->thread.trap_nr = read_csr_excode();
arch/loongarch/mm/fault.c
145
current->thread.csr_badvaddr = address;
arch/loongarch/mm/fault.c
147
current->thread.error_code = 1;
arch/loongarch/mm/fault.c
149
current->thread.error_code = 2;
arch/loongarch/mm/fault.c
150
current->thread.trap_nr = read_csr_excode();
arch/loongarch/mm/fault.c
153
unhandled_signal(current, SIGSEGV) && __ratelimit(&ratelimit_state)) {
arch/loongarch/mm/fault.c
155
current->comm,
arch/loongarch/mm/fault.c
180
struct task_struct *tsk = current;
arch/loongarch/mm/fault.c
185
if (kprobe_page_fault(regs, current->thread.trap_nr))
arch/loongarch/mm/mmap.c
25
struct mm_struct *mm = current->mm;
arch/loongarch/mm/tlb.c
177
if (current->active_mm != vma->vm_mm)
arch/loongarch/power/hibernate.c
26
save_fp(current);
arch/loongarch/power/hibernate.c
39
restore_fp(current);
arch/m68k/coldfire/sltimers.c
48
if (current->pid)
arch/m68k/coldfire/timers.c
162
if (current->pid)
arch/m68k/include/asm/cacheflush_mm.h
199
if (mm == current->mm)
arch/m68k/include/asm/cacheflush_mm.h
211
if (vma->vm_mm == current->mm)
arch/m68k/include/asm/cacheflush_mm.h
217
if (vma->vm_mm == current->mm)
arch/m68k/include/asm/current.h
7
register struct task_struct *current __asm__("%a2");
arch/m68k/include/asm/flat.h
13
if (current->mm) \
arch/m68k/include/asm/flat.h
14
(regs)->d5 = current->mm->start_data; \
arch/m68k/include/asm/math-emu.h
128
#define FPDATA ((struct fp_data *)current->thread.fp)
arch/m68k/include/asm/processor.h
173
#define KSTK_ESP(tsk) ((tsk) == current ? rdusp() : (tsk)->thread.usp)
arch/m68k/include/asm/tlbflush.h
80
if (mm == current->active_mm)
arch/m68k/include/asm/tlbflush.h
86
if (vma->vm_mm == current->active_mm)
arch/m68k/include/asm/tlbflush.h
93
if (vma->vm_mm == current->active_mm)
arch/m68k/kernel/process.c
223
memcpy(fpu->fpcntl, current->thread.fpcntl, 12);
arch/m68k/kernel/process.c
224
memcpy(fpu->fpregs, current->thread.fp, 96);
arch/m68k/kernel/process.c
96
current->thread.fc = USER_DATA;
arch/m68k/kernel/ptrace.c
282
ret = ptrace_report_syscall_entry(task_pt_regs(current));
arch/m68k/kernel/ptrace.c
293
ptrace_report_syscall_exit(task_pt_regs(current), 0);
arch/m68k/kernel/signal.c
1084
send_sig(SIGTRAP, current, 1);
arch/m68k/kernel/signal.c
1097
current->thread.esp0 = (unsigned long) regs;
arch/m68k/kernel/signal.c
263
memcpy(current->thread.fpcntl, sc->sc_fpcntl, 12);
arch/m68k/kernel/signal.c
264
memcpy(current->thread.fp, sc->sc_fpregs, 24);
arch/m68k/kernel/signal.c
342
if (__copy_from_user(current->thread.fpcntl,
arch/m68k/kernel/signal.c
346
if (__copy_from_user(current->thread.fp,
arch/m68k/kernel/signal.c
435
memcpy(sc->sc_fpcntl, current->thread.fpcntl, 12);
arch/m68k/kernel/signal.c
436
memcpy(sc->sc_fpregs, current->thread.fp, 24);
arch/m68k/kernel/signal.c
493
current->thread.fpcntl, 12);
arch/m68k/kernel/signal.c
496
current->thread.fp, 96);
arch/m68k/kernel/signal.c
668
current->thread.esp0 = (unsigned long)&new->ptregs;
arch/m68k/kernel/signal.c
687
current->restart_block.fn = do_no_restart_syscall;
arch/m68k/kernel/signal.c
720
current->restart_block.fn = do_no_restart_syscall;
arch/m68k/kernel/sys_m68k.c
392
mmap_read_lock(current->mm);
arch/m68k/kernel/sys_m68k.c
404
mmap_read_lock(current->mm);
arch/m68k/kernel/sys_m68k.c
405
vma = vma_lookup(current->mm, addr);
arch/m68k/kernel/sys_m68k.c
455
mmap_read_unlock(current->mm);
arch/m68k/kernel/sys_m68k.c
468
struct mm_struct *mm = current->mm;
arch/m68k/kernel/sys_m68k.c
549
struct mm_struct *mm = current->mm;
arch/m68k/kernel/traps.c
1141
current->thread.esp0 = ssp;
arch/m68k/kernel/traps.c
234
if (wba != current->thread.faddr)
arch/m68k/kernel/traps.c
352
current->thread.signo = SIGBUS;
arch/m68k/kernel/traps.c
353
current->thread.faddr = fp->un.fmt7.faddr;
arch/m68k/kernel/traps.c
506
pr_debug("pid = %x ", current->pid);
arch/m68k/kernel/traps.c
534
temp, addr, current);
arch/m68k/kernel/traps.c
643
temp, addr, current);
arch/m68k/kernel/traps.c
757
current->thread.esp0 = (unsigned long) fp;
arch/m68k/kernel/traps.c
858
current->comm, task_pid_nr(current), current);
arch/m68k/kernel/traps.c
998
pr_err("Current process id is %d\n", task_pid_nr(current));
arch/m68k/mm/fault.c
191
current->thread.signo = SIGBUS;
arch/m68k/mm/fault.c
192
current->thread.faddr = address;
arch/m68k/mm/fault.c
196
current->thread.signo = SIGBUS;
arch/m68k/mm/fault.c
197
current->thread.code = BUS_ADRERR;
arch/m68k/mm/fault.c
198
current->thread.faddr = address;
arch/m68k/mm/fault.c
204
current->thread.signo = SIGSEGV;
arch/m68k/mm/fault.c
205
current->thread.code = SEGV_MAPERR;
arch/m68k/mm/fault.c
206
current->thread.faddr = address;
arch/m68k/mm/fault.c
210
current->thread.signo = SIGSEGV;
arch/m68k/mm/fault.c
211
current->thread.code = SEGV_ACCERR;
arch/m68k/mm/fault.c
212
current->thread.faddr = address;
arch/m68k/mm/fault.c
29
signo = current->thread.signo;
arch/m68k/mm/fault.c
30
si_code = current->thread.code;
arch/m68k/mm/fault.c
31
addr = (void __user *)current->thread.faddr;
arch/m68k/mm/fault.c
73
struct mm_struct *mm = current->mm;
arch/m68k/mm/mcfmmu.c
74
current->mm = NULL;
arch/m68k/mm/mcfmmu.c
94
mm = (!user_mode(regs) && KMAPAREA(mmuar)) ? &init_mm : current->mm;
arch/m68k/mm/sun3mmu.c
81
current->mm = NULL;
arch/m68k/sun3/mmu_emu.c
362
if(current->mm == NULL) {
arch/m68k/sun3/mmu_emu.c
366
context = current->mm->context;
arch/m68k/sun3/mmu_emu.c
370
crp = current->mm->pgd;
arch/microblaze/include/asm/current.h
21
register struct task_struct *current asm("r31");
arch/microblaze/include/asm/current.h
23
# define get_current() current
arch/microblaze/include/asm/elf.h
25
set_personality(PER_LINUX_32BIT | (current->personality & (~PER_MASK)))
arch/microblaze/include/asm/mmu_context_mm.h
130
current->thread.pgdir = mm->pgd;
arch/microblaze/kernel/ftrace.c
32
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/microblaze/kernel/setup.c
173
per_cpu(CURRENT_SAVE, 0) = (unsigned long)current;
arch/microblaze/kernel/signal.c
193
pmdp = pmd_off(current->mm, address);
arch/microblaze/kernel/signal.c
224
current->comm, current->pid, frame, regs->pc);
arch/microblaze/kernel/signal.c
90
current->restart_block.fn = do_no_restart_syscall;
arch/microblaze/kernel/traps.c
76
task = current;
arch/microblaze/kernel/unwind.c
277
if (task == current) {
arch/microblaze/kernel/unwind.c
306
microblaze_unwind_inner(current, pc, fp, 0, trace, loglvl);
arch/microblaze/mm/fault.c
175
struct pt_regs *uregs = current->thread.regs;
arch/microblaze/mm/fault.c
90
struct mm_struct *mm = current->mm;
arch/microblaze/mm/pgtable.c
234
mm = current->mm;
arch/mips/cavium-octeon/cpu.c
31
prefetch(&current->thread.cp2);
arch/mips/cavium-octeon/cpu.c
33
KSTK_STATUS(current) |= ST0_CU2;
arch/mips/cavium-octeon/cpu.c
36
octeon_cop2_restore(&(current->thread.cp2));
arch/mips/cavium-octeon/octeon-crypto.c
35
if (KSTK_STATUS(current) & ST0_CU2) {
arch/mips/cavium-octeon/octeon-crypto.c
36
octeon_cop2_save(&(current->thread.cp2));
arch/mips/cavium-octeon/octeon-crypto.c
37
KSTK_STATUS(current) &= ~ST0_CU2;
arch/mips/include/asm/dsp.h
75
if (tsk == current) \
arch/mips/include/asm/dsp.h
76
__save_dsp(current); \
arch/mips/include/asm/elf.h
336
current->thread.abi = &mips_abi; \
arch/mips/include/asm/elf.h
341
if (personality(current->personality) != PER_LINUX) \
arch/mips/include/asm/elf.h
354
current->thread.abi = &mips_abi_n32; \
arch/mips/include/asm/elf.h
369
current->thread.abi = &mips_abi_32; \
arch/mips/include/asm/elf.h
399
current->personality &= ~READ_IMPLIES_EXEC; \
arch/mips/include/asm/elf.h
404
current->thread.abi = &mips_abi; \
arch/mips/include/asm/elf.h
408
p = personality(current->personality); \
arch/mips/include/asm/elf.h
464
(unsigned long)current->mm->context.vdso); \
arch/mips/include/asm/fpu.h
132
if (current->thread.fpu.fcr31 & FPU_CSR_NAN2008) {
arch/mips/include/asm/fpu.h
144
KSTK_STATUS(current) |= ST0_CU1;
arch/mips/include/asm/fpu.h
146
KSTK_STATUS(current) |= ST0_FR;
arch/mips/include/asm/fpu.h
148
KSTK_STATUS(current) &= ~ST0_FR;
arch/mips/include/asm/fpu.h
164
_restore_fp(current);
arch/mips/include/asm/fpu.h
206
lose_fpu_inatomic(save, current);
arch/mips/include/asm/fpu.h
256
if (tsk == current) {
arch/mips/include/asm/fpu.h
259
_save_fp(current);
arch/mips/include/asm/stackprotector.h
30
current->stack_canary = canary;
arch/mips/include/asm/stackprotector.h
31
__stack_chk_guard = current->stack_canary;
arch/mips/kernel/branch.c
156
fcr31 = current->thread.fpu.fcr31;
arch/mips/kernel/branch.c
693
if (!init_fp_ctx(current))
arch/mips/kernel/branch.c
696
bit = get_fpr32(&current->thread.fpu.fpr[reg], 0) & 0x1;
arch/mips/kernel/branch.c
714
fcr31 = current->thread.fpu.fcr31;
arch/mips/kernel/branch.c
833
current->comm);
arch/mips/kernel/branch.c
838
current->comm);
arch/mips/kernel/branch.c
843
current->comm);
arch/mips/kernel/branch.c
871
printk("%s: unaligned epc - sending SIGBUS.\n", current->comm);
arch/mips/kernel/crash.c
34
regs = task_pt_regs(current);
arch/mips/kernel/elf.c
311
struct task_struct *t = current;
arch/mips/kernel/ftrace.c
333
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/mips/kernel/kgdb.c
140
} else if (current && dbg_reg_def[regno].offset != -1 && regno < 72) {
arch/mips/kernel/kgdb.c
146
memcpy((void *)&current->thread.fpu.fcr31, mem,
arch/mips/kernel/kgdb.c
154
memcpy((void *)&current->thread.fpu.fpr[fp_reg], mem,
arch/mips/kernel/kgdb.c
157
restore_fp(current);
arch/mips/kernel/kgdb.c
174
} else if (current && dbg_reg_def[regno].offset != -1 && regno < 72) {
arch/mips/kernel/kgdb.c
178
save_fp(current);
arch/mips/kernel/kgdb.c
181
memcpy(mem, (void *)&current->thread.fpu.fcr31,
arch/mips/kernel/kgdb.c
190
memcpy(mem, (void *)&current->thread.fpu.fpr[fp_reg],
arch/mips/kernel/kprobes.c
229
pr_notice("Failed to emulate branch instruction because of unaligned epc - sending SIGBUS to %s.\n", current->comm);
arch/mips/kernel/linux32.c
94
if (personality(current->personality) == PER_LINUX32 &&
arch/mips/kernel/mips-mt-fpaff.c
44
return pid ? find_task_by_vpid(pid) : current;
arch/mips/kernel/mips-r2-to-r6-emul.c
1178
err = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 0,
arch/mips/kernel/mips-r2-to-r6-emul.c
1185
*fcr31 = res = mask_fcr31_x(current->thread.fpu.fcr31);
arch/mips/kernel/mips-r2-to-r6-emul.c
1186
current->thread.fpu.fcr31 &= ~res;
arch/mips/kernel/mips-r2-to-r6-emul.c
1199
current->thread.cp0_baduaddr = (unsigned long)fault_addr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1209
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1282
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1356
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1426
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1501
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1620
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1739
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1857
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1969
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
1974
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
2025
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
2030
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
204
csr = current->thread.fpu.fcr31;
arch/mips/kernel/mips-r2-to-r6-emul.c
2088
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
2093
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
2149
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
2154
current->thread.cp0_baduaddr = vaddr;
arch/mips/kernel/mips-r2-to-r6-emul.c
227
csr = current->thread.fpu.fcr31;
arch/mips/kernel/perf_event.c
52
(unsigned long)task_stack_page(current);
arch/mips/kernel/perf_event.c
62
pc = unwind_stack(current, &sp, pc, &ra);
arch/mips/kernel/perf_regs.c
66
regs_user->regs = task_pt_regs(current);
arch/mips/kernel/perf_regs.c
67
regs_user->abi = perf_reg_abi(current);
arch/mips/kernel/pm.c
30
save_dsp(current);
arch/mips/kernel/pm.c
44
if (current->mm)
arch/mips/kernel/pm.c
45
write_c0_entryhi(cpu_asid(cpu, current->mm));
arch/mips/kernel/pm.c
48
restore_dsp(current);
arch/mips/kernel/pm.c
55
__restore_watch(current);
arch/mips/kernel/process.c
64
atomic_set(&current->thread.bd_emu_frame, BD_EMUFRAME_NONE);
arch/mips/kernel/process.c
693
if (current->thread.abi) {
arch/mips/kernel/process.c
694
top -= PAGE_ALIGN(current->thread.abi->vdso->size);
arch/mips/kernel/process.c
699
if (current->flags & PF_RANDOMIZE)
arch/mips/kernel/process.c
716
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/mips/kernel/process.c
77
if (!(current->flags & PF_KTHREAD))
arch/mips/kernel/process.c
93
save_msa(current);
arch/mips/kernel/process.c
95
_save_fp(current);
arch/mips/kernel/process.c
97
save_dsp(current);
arch/mips/kernel/ptrace.c
1346
syscall_set_return_value(current, regs, -ENOSYS, 0);
arch/mips/kernel/rtlx.c
137
if (!signal_pending(current)) {
arch/mips/kernel/signal.c
102
set_fpr64(&current->thread.fpu.fpr[i], 0, fpr_val);
arch/mips/kernel/signal.c
104
err |= __get_user(current->thread.fpu.fcr31, csr);
arch/mips/kernel/signal.c
128
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
137
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
195
err = __put_user(current->thread.fpu.msacsr, &msa->csr);
arch/mips/kernel/signal.c
198
val = get_fpr64(&current->thread.fpu.fpr[i], 1);
arch/mips/kernel/signal.c
241
current->thread.fpu.msacsr = csr;
arch/mips/kernel/signal.c
245
set_fpr64(&current->thread.fpu.fpr[i], 1, val);
arch/mips/kernel/signal.c
327
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
380
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
523
current->restart_block.fn = do_no_restart_syscall;
arch/mips/kernel/signal.c
73
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
747
current->comm, current->pid,
arch/mips/kernel/signal.c
796
current->comm, current->pid,
arch/mips/kernel/signal.c
82
__put_user(get_fpr64(&current->thread.fpu.fpr[i], 0),
arch/mips/kernel/signal.c
820
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal.c
821
void *vdso = current->mm->context.vdso;
arch/mips/kernel/signal.c
85
err |= __put_user(current->thread.fpu.fcr31, csr);
arch/mips/kernel/signal.c
885
regs->regs[2] = current->thread.abi->restart;
arch/mips/kernel/signal.c
92
struct mips_abi *abi = current->thread.abi;
arch/mips/kernel/signal_n32.c
131
current->comm, current->pid,
arch/mips/kernel/signal_o32.c
149
current->comm, current->pid,
arch/mips/kernel/signal_o32.c
235
current->comm, current->pid,
arch/mips/kernel/signal_o32.c
94
current->restart_block.fn = do_no_restart_syscall;
arch/mips/kernel/smp.c
440
init_new_context(current, &init_mm);
arch/mips/kernel/smp.c
571
} else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/mips/kernel/smp.c
621
} else if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/mips/kernel/smp.c
689
(current->mm != vma->vm_mm)) {
arch/mips/kernel/stacktrace.c
74
save_stack_trace_tsk(current, trace);
arch/mips/kernel/stacktrace.c
85
if (tsk != current) {
arch/mips/kernel/stacktrace.c
91
save_context_stack(trace, tsk, regs, tsk == current);
arch/mips/kernel/syscall.c
171
ll_task = current;
arch/mips/kernel/syscall.c
89
struct thread_info *ti = task_thread_info(current);
arch/mips/kernel/traps.c
1037
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/kernel/traps.c
1079
current->thread.trap_nr, SIGTRAP) == NOTIFY_STOP)
arch/mips/kernel/traps.c
1085
current->thread.trap_nr, SIGTRAP) == NOTIFY_STOP)
arch/mips/kernel/traps.c
1091
current->thread.trap_nr, SIGTRAP) == NOTIFY_STOP)
arch/mips/kernel/traps.c
1097
current->thread.trap_nr, SIGTRAP) == NOTIFY_STOP)
arch/mips/kernel/traps.c
1125
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/kernel/traps.c
1180
&current->thread.cp0_baduaddr,
arch/mips/kernel/traps.c
1189
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/kernel/traps.c
1191
if (notify_die(DIE_RI, "RI Fault", regs, 0, current->thread.trap_nr,
arch/mips/kernel/traps.c
1282
first_fp = init_fp_ctx(current);
arch/mips/kernel/traps.c
1295
write_msa_csr(current->thread.fpu.msacsr);
arch/mips/kernel/traps.c
1349
write_msa_csr(current->thread.fpu.msacsr);
arch/mips/kernel/traps.c
1371
_restore_fp(current);
arch/mips/kernel/traps.c
1381
restore_msa(current);
arch/mips/kernel/traps.c
1386
current->thread.fpu.fcr31);
arch/mips/kernel/traps.c
1478
sig = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 0,
arch/mips/kernel/traps.c
1485
fcr31 = mask_fcr31_x(current->thread.fpu.fcr31);
arch/mips/kernel/traps.c
1486
current->thread.fpu.fcr31 &= ~fcr31;
arch/mips/kernel/traps.c
1514
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/kernel/traps.c
1516
current->thread.trap_nr, SIGFPE) == NOTIFY_STOP)
arch/mips/kernel/traps.c
1578
if (test_tsk_thread_flag(current, TIF_LOAD_WATCH)) {
arch/mips/kernel/traps.c
160
task = current;
arch/mips/kernel/traps.c
220
if (task && task != current) {
arch/mips/kernel/traps.c
2270
current->active_mm = &init_mm;
arch/mips/kernel/traps.c
2271
BUG_ON(current->mm);
arch/mips/kernel/traps.c
2272
enter_lazy_tlb(&init_mm, current);
arch/mips/kernel/traps.c
379
current->comm, current->pid, current_thread_info(), current,
arch/mips/kernel/traps.c
389
show_stacktrace(current, regs, KERN_DEFAULT, user_mode(regs));
arch/mips/kernel/traps.c
403
if (notify_die(DIE_OOPS, str, regs, 0, current->thread.trap_nr,
arch/mips/kernel/traps.c
424
if (regs && kexec_should_crash(current))
arch/mips/kernel/traps.c
489
if (notify_die(DIE_OOPS, "bus error", regs, 0, current->thread.trap_nr,
arch/mips/kernel/traps.c
555
if (ll_task == NULL || ll_task == current) {
arch/mips/kernel/traps.c
560
ll_task = current;
arch/mips/kernel/traps.c
594
if (ll_bit == 0 || ll_task != current) {
arch/mips/kernel/traps.c
639
struct thread_info *ti = task_thread_info(current);
arch/mips/kernel/traps.c
801
force_fcr31_sig(fcr31, fault_addr, current);
arch/mips/kernel/traps.c
809
mmap_read_lock(current->mm);
arch/mips/kernel/traps.c
810
if (vma_lookup(current->mm, (unsigned long)fault_addr))
arch/mips/kernel/traps.c
814
mmap_read_unlock(current->mm);
arch/mips/kernel/traps.c
854
sig = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 1,
arch/mips/kernel/traps.c
861
fcr31 = mask_fcr31_x(current->thread.fpu.fcr31);
arch/mips/kernel/traps.c
862
current->thread.fpu.fcr31 &= ~fcr31;
arch/mips/kernel/traps.c
883
if (notify_die(DIE_FP, "FP exception", regs, 0, current->thread.trap_nr,
arch/mips/kernel/traps.c
906
sig = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 1,
arch/mips/kernel/traps.c
913
fcr31 = mask_fcr31_x(current->thread.fpu.fcr31);
arch/mips/kernel/traps.c
914
current->thread.fpu.fcr31 &= ~fcr31;
arch/mips/kernel/traps.c
939
((current->thread.emulated_fp++ > mt_fpemul_threshold))) {
arch/mips/kernel/traps.c
945
if (cpumask_intersects(&current->cpus_mask, &mt_fpu_cpumask)) {
arch/mips/kernel/traps.c
948
current->thread.user_cpus_allowed
arch/mips/kernel/traps.c
949
= current->cpus_mask;
arch/mips/kernel/traps.c
950
cpumask_and(&tmask, &current->cpus_mask,
arch/mips/kernel/traps.c
952
set_cpus_allowed_ptr(current, &tmask);
arch/mips/kernel/traps.c
975
if (kgdb_ll_trap(DIE_TRAP, str, regs, code, current->thread.trap_nr,
arch/mips/kernel/traps.c
980
if (notify_die(DIE_TRAP, str, regs, code, current->thread.trap_nr,
arch/mips/kernel/unaligned.c
1535
current->thread.cp0_baduaddr = regs->cp0_badvaddr;
arch/mips/kernel/unaligned.c
447
res = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 1,
arch/mips/kernel/unaligned.c
479
fpr = &current->thread.fpu.fpr[wd];
arch/mips/kernel/unaligned.c
968
res = fpu_emulator_cop1Handler(regs, &current->thread.fpu, 1,
arch/mips/kernel/uprobes.c
107
struct uprobe_task *utask = current->utask;
arch/mips/kernel/uprobes.c
119
utask->autask.saved_trap_nr = current->thread.trap_nr;
arch/mips/kernel/uprobes.c
120
current->thread.trap_nr = UPROBE_TRAP_NR;
arch/mips/kernel/uprobes.c
121
regs->cp0_epc = current->utask->xol_vaddr;
arch/mips/kernel/uprobes.c
128
struct uprobe_task *utask = current->utask;
arch/mips/kernel/uprobes.c
130
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/mips/kernel/uprobes.c
192
struct uprobe_task *utask = current->utask;
arch/mips/kernel/uprobes.c
194
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/mips/kernel/vdso.c
69
if (current->flags & PF_RANDOMIZE) {
arch/mips/kernel/vdso.c
79
struct mips_vdso_image *image = current->thread.abi->vdso;
arch/mips/kernel/vdso.c
80
struct mm_struct *mm = current->mm;
arch/mips/kernel/watch.c
53
&current->thread.watch.mips3264;
arch/mips/kvm/mips.c
1315
if (signal_pending(current)) {
arch/mips/kvm/mips.c
1541
if (!(current->flags & PF_VCPU))
arch/mips/kvm/vz.c
2586
if (current->flags & PF_VCPU) {
arch/mips/kvm/vz.c
2709
if (current->flags & PF_VCPU)
arch/mips/loongson64/cop2-ex.c
131
value_next = get_fpr64(&current->thread.fpu.fpr[insn.loongson3_lswc2_format.rq], 0);
arch/mips/loongson64/cop2-ex.c
137
value = get_fpr64(&current->thread.fpu.fpr[insn.loongson3_lswc2_format.rt], 0);
arch/mips/loongson64/cop2-ex.c
203
set_fpr64(&current->thread.fpu.fpr[insn.loongson3_lsdc2_format.rt], 0, value);
arch/mips/loongson64/cop2-ex.c
219
set_fpr64(&current->thread.fpu.fpr[insn.loongson3_lsdc2_format.rt], 0, value);
arch/mips/loongson64/cop2-ex.c
283
value = get_fpr64(&current->thread.fpu.fpr[insn.loongson3_lsdc2_format.rt], 0);
arch/mips/loongson64/cop2-ex.c
301
value = get_fpr64(&current->thread.fpu.fpr[insn.loongson3_lsdc2_format.rt], 0);
arch/mips/loongson64/cop2-ex.c
51
KSTK_STATUS(current) |= (ST0_CU1 | ST0_CU2);
arch/mips/loongson64/cop2-ex.c
53
KSTK_STATUS(current) |= ST0_FR;
arch/mips/loongson64/cop2-ex.c
55
KSTK_STATUS(current) &= ~ST0_FR;
arch/mips/loongson64/cop2-ex.c
59
init_fp_ctx(current);
arch/mips/loongson64/cop2-ex.c
60
_restore_fp(current);
arch/mips/loongson64/cop2-ex.c
98
set_fpr64(&current->thread.fpu.fpr[insn.loongson3_lswc2_format.rt], 0, value);
arch/mips/loongson64/cop2-ex.c
99
set_fpr64(&current->thread.fpu.fpr[insn.loongson3_lswc2_format.rq], 0, value_next);
arch/mips/math-emu/cp1emu.c
1191
fpr = &current->thread.fpu.fpr[MIPSInst_RT(ir)];
arch/mips/math-emu/cp1emu.c
2850
if (!init_fp_ctx(current))
arch/mips/math-emu/cp1emu.c
706
fpr = &current->thread.fpu.fpr[insn.i_format.rt];
arch/mips/math-emu/cp1emu.c
736
fcr31 = current->thread.fpu.fcr31;
arch/mips/math-emu/dsemul.c
116
pr_debug("allocate emuframe %d to %d\n", idx, current->pid);
arch/mips/math-emu/dsemul.c
128
pr_debug("free emuframe %d from %d\n", idx, current->pid);
arch/mips/math-emu/dsemul.c
180
fr_idx = atomic_read(&current->thread.bd_emu_frame);
arch/mips/math-emu/dsemul.c
193
regs->cp0_epc = current->thread.bd_emu_branch_pc;
arch/mips/math-emu/dsemul.c
195
regs->cp0_epc = current->thread.bd_emu_cont_pc;
arch/mips/math-emu/dsemul.c
197
atomic_set(&current->thread.bd_emu_frame, BD_EMUFRAME_NONE);
arch/mips/math-emu/dsemul.c
198
free_emuframe(fr_idx, current->mm);
arch/mips/math-emu/dsemul.c
246
fr_idx = atomic_read(&current->thread.bd_emu_frame);
arch/mips/math-emu/dsemul.c
273
ret = access_process_vm(current, fr_uaddr, &fr, sizeof(fr),
arch/mips/math-emu/dsemul.c
277
free_emuframe(fr_idx, current->mm);
arch/mips/math-emu/dsemul.c
282
current->thread.bd_emu_branch_pc = branch_pc;
arch/mips/math-emu/dsemul.c
283
current->thread.bd_emu_cont_pc = cont_pc;
arch/mips/math-emu/dsemul.c
284
atomic_set(&current->thread.bd_emu_frame, fr_idx);
arch/mips/math-emu/dsemul.c
295
if (!dsemul_thread_cleanup(current)) {
arch/mips/math-emu/dsemul.c
301
xcp->cp0_epc = current->thread.bd_emu_cont_pc;
arch/mips/math-emu/dsemul.c
77
mm_context_t *mm_ctx = &current->mm->context;
arch/mips/math-emu/ieee754.h
170
#define ieee754_csr (*(struct _ieee754_csr *)(&current->thread.fpu.fcr31))
arch/mips/mm/c-r4k.c
568
if ((mm == current->active_mm) && (pte_val(*ptep) & _PAGE_VALID))
arch/mips/mm/c-r4k.c
593
if (vaddr && cpu_has_vtag_icache && mm == current->active_mm) {
arch/mips/mm/fault.c
122
current->comm, current->pid,
arch/mips/mm/fault.c
133
current->comm, current->pid,
arch/mips/mm/fault.c
215
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/mm/fault.c
223
current->thread.cp0_baduaddr = address;
arch/mips/mm/fault.c
270
current->thread.trap_nr = (regs->cp0_cause >> 2) & 0x1f;
arch/mips/mm/fault.c
43
struct task_struct *tsk = current;
arch/mips/mm/fault.c
54
current->comm, current->pid, field, address, write,
arch/mips/mm/fault.c
63
current->thread.trap_nr, SIGSEGV) == NOTIFY_STOP)
arch/mips/mm/mmap.c
33
struct mm_struct *mm = current->mm;
arch/mips/mm/tlb-r3k.c
188
if (current->active_mm != vma->vm_mm)
arch/mips/mm/tlb-r4k.c
310
if (current->active_mm != vma->vm_mm)
arch/mips/power/cpu.c
22
save_fp(current);
arch/mips/power/cpu.c
24
save_dsp(current);
arch/mips/power/cpu.c
32
restore_fp(current);
arch/mips/power/cpu.c
34
restore_dsp(current);
arch/nios2/kernel/misaligned.c
162
current->pid);
arch/nios2/kernel/process.c
159
pr_emerg("COMM=%s PID=%d\n", current->comm, current->pid);
arch/nios2/kernel/process.c
161
if (current->mm) {
arch/nios2/kernel/process.c
163
(int) current->mm->start_code,
arch/nios2/kernel/process.c
164
(int) current->mm->end_code,
arch/nios2/kernel/process.c
165
(int) current->mm->start_data,
arch/nios2/kernel/process.c
166
(int) current->mm->end_data,
arch/nios2/kernel/process.c
167
(int) current->mm->end_data,
arch/nios2/kernel/process.c
168
(int) current->mm->brk);
arch/nios2/kernel/process.c
170
(int) current->mm->start_stack,
arch/nios2/kernel/process.c
171
(int)(((unsigned long) current) + THREAD_SIZE));
arch/nios2/kernel/ptrace.c
136
ret = ptrace_report_syscall_entry(task_pt_regs(current));
arch/nios2/kernel/ptrace.c
144
ptrace_report_syscall_exit(task_pt_regs(current), 0);
arch/nios2/kernel/signal.c
240
current->thread.kregs = regs;
arch/nios2/kernel/signal.c
46
current->restart_block.fn = do_no_restart_syscall;
arch/nios2/kernel/sys_nios2.c
25
struct mm_struct *mm = current->mm;
arch/nios2/mm/cacheflush.c
76
struct mm_struct *mm = current->active_mm;
arch/nios2/mm/fault.c
171
if (unhandled_signal(current, SIGSEGV) && printk_ratelimit()) {
arch/nios2/mm/fault.c
173
"cause %ld\n", current->comm, SIGSEGV, address, cause);
arch/nios2/mm/fault.c
47
struct task_struct *tsk = current;
arch/nios2/mm/init.c
96
struct mm_struct *mm = current->mm;
arch/nios2/mm/tlb.c
259
if (current->mm == mm) {
arch/openrisc/include/asm/tlbflush.h
53
flush_tlb_mm(current->mm);
arch/openrisc/kernel/process.c
248
save_fpu(current);
arch/openrisc/kernel/process.c
262
restore_fpu(current);
arch/openrisc/kernel/signal.c
48
err = __copy_from_user(&current->thread.fpcsr, &sc->fpcsr, sizeof(unsigned long));
arch/openrisc/kernel/signal.c
53
restore_fpu(current);
arch/openrisc/kernel/signal.c
63
save_fpu(current);
arch/openrisc/kernel/signal.c
65
err = __copy_to_user(&sc->fpcsr, &current->thread.fpcsr, sizeof(unsigned long));
arch/openrisc/kernel/signal.c
80
current->restart_block.fn = do_no_restart_syscall;
arch/openrisc/kernel/smp.c
133
current->active_mm = mm;
arch/openrisc/kernel/stacktrace.c
75
if (tsk == current)
arch/openrisc/kernel/traps.c
110
current->comm, current->pid, (unsigned long)current);
arch/openrisc/kernel/traps.c
188
save_fpu(current);
arch/openrisc/kernel/traps.c
189
fpcsr = current->thread.fpcsr;
arch/openrisc/kernel/traps.c
203
current->thread.fpcsr &= ~SPR_FPCSR_ALLF;
arch/openrisc/kernel/traps.c
204
restore_fpu(current);
arch/openrisc/mm/fault.c
56
tsk = current;
arch/parisc/include/asm/compat.h
138
return __is_compat_task(current);
arch/parisc/include/asm/elf.h
236
set_personality((current->personality & ~PER_MASK) | PER_LINUX); \
arch/parisc/include/asm/elf.h
238
current->thread.map_base = DEFAULT_MAP_BASE; \
arch/parisc/include/asm/elf.h
239
current->thread.task_size = DEFAULT_TASK_SIZE; \
arch/parisc/include/asm/elf.h
248
current->thread.map_base = DEFAULT_MAP_BASE32; \
arch/parisc/include/asm/elf.h
249
current->thread.task_size = DEFAULT_TASK_SIZE32; \
arch/parisc/include/asm/elf.h
359
#define VDSO_CURRENT_BASE current->mm->context.vdso_base
arch/parisc/include/asm/mmu_context.h
95
switch_mm(prev,next,current);
arch/parisc/include/asm/processor.h
243
__u32 spaceid = (__u32)current->mm->context.space_id; \
arch/parisc/include/asm/processor.h
28
#define TASK_SIZE TASK_SIZE_OF(current)
arch/parisc/include/asm/processor.h
29
#define TASK_UNMAPPED_BASE (current->thread.map_base)
arch/parisc/include/asm/tlbflush.h
58
if (mm == current->active_mm)
arch/parisc/kernel/ftrace.c
43
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/parisc/kernel/irq.c
381
unsigned long stack_start = (unsigned long) task_stack_page(current);
arch/parisc/kernel/irq.c
415
current->comm, sp, stack_start, stack_start + IRQ_STACK_SIZE);
arch/parisc/kernel/irq.c
432
current->comm, sp, stack_start, stack_start + THREAD_SIZE);
arch/parisc/kernel/perf_event.c
18
unwind_frame_init_task(&info, current, NULL);
arch/parisc/kernel/perf_regs.c
59
regs_user->regs = task_pt_regs(current);
arch/parisc/kernel/perf_regs.c
60
regs_user->abi = perf_reg_abi(current);
arch/parisc/kernel/ptrace.c
188
task_user_regset_view(current),
arch/parisc/kernel/ptrace.c
195
task_user_regset_view(current),
arch/parisc/kernel/ptrace.c
202
task_user_regset_view(current),
arch/parisc/kernel/ptrace.c
209
task_user_regset_view(current),
arch/parisc/kernel/signal.c
172
sp = (current->sas_ss_sp + 0x7f) & ~0x3f; /* Stacks grow up! */
arch/parisc/kernel/signal.c
276
rp = VDSO64_SYMBOL(current, sigtramp_rt);
arch/parisc/kernel/signal.c
279
rp = VDSO32_SYMBOL(current, sigtramp_rt);
arch/parisc/kernel/signal.c
339
if (pa_psw(current)->r) {
arch/parisc/kernel/signal.c
340
pa_psw(current)->r = 0;
arch/parisc/kernel/signal.c
372
__func__, current->comm, current->pid, frame, regs->gr[30],
arch/parisc/kernel/signal.c
447
current->comm, task_pid_nr(current), opcode);
arch/parisc/kernel/signal.c
458
__func__, regs->orig_r28, task_pid_nr(current), regs->gr[20]);
arch/parisc/kernel/signal.c
470
__func__, task_pid_nr(current));
arch/parisc/kernel/signal.c
490
__func__, regs->gr[28], task_pid_nr(current));
arch/parisc/kernel/signal.c
514
rp = VDSO64_SYMBOL(current, restart_syscall);
arch/parisc/kernel/signal.c
519
rp = VDSO32_SYMBOL(current, restart_syscall);
arch/parisc/kernel/signal.c
567
__func__, task_pid_nr(current));
arch/parisc/kernel/signal.c
576
__func__, regs->gr[28], regs->orig_r28, task_pid_nr(current));
arch/parisc/kernel/signal.c
90
current->restart_block.fn = do_no_restart_syscall;
arch/parisc/kernel/signal.c
95
DBG(2, "%s: frame is %p pid %d\n", __func__, frame, task_pid_nr(current));
arch/parisc/kernel/smp.c
295
current->active_mm = &init_mm;
arch/parisc/kernel/smp.c
296
BUG_ON(current->mm);
arch/parisc/kernel/smp.c
297
enter_lazy_tlb(&init_mm, current);
arch/parisc/kernel/smp.c
94
switch_to_idle_task(current);
arch/parisc/kernel/sys_parisc.c
103
struct mm_struct *mm = current->mm;
arch/parisc/kernel/sys_parisc.c
296
if (personality(current->personality) == PER_LINUX32
arch/parisc/kernel/sys_parisc.c
327
current->comm, current->pid);
arch/parisc/kernel/sys_parisc.c
91
if (current->flags & PF_RANDOMIZE)
arch/parisc/kernel/sys_parisc32.c
23
current->comm, current->pid, r20);
arch/parisc/kernel/traps.c
154
level, task_cpu(current), cr30, cr31);
arch/parisc/kernel/traps.c
166
parisc_show_stack(current, regs, KERN_DEFAULT);
arch/parisc/kernel/traps.c
227
current->comm, task_pid_nr(current), str, err, regs->iaoq[0]);
arch/parisc/kernel/traps.c
251
current->comm, task_pid_nr(current), str, err);
arch/parisc/kernel/traps.c
254
if (current->thread.flags & PARISC_KERNEL_DEATH) {
arch/parisc/kernel/traps.c
259
current->thread.flags |= PARISC_KERNEL_DEATH;
arch/parisc/kernel/traps.c
329
task_pid_nr(current), current->comm);
arch/parisc/kernel/traps.c
450
unwind_frame_init(&info, current, regs);
arch/parisc/kernel/traps.c
718
mmap_read_lock(current->mm);
arch/parisc/kernel/traps.c
719
vma = find_vma(current->mm,regs->iaoq[0]);
arch/parisc/kernel/traps.c
726
mmap_read_unlock(current->mm);
arch/parisc/kernel/traps.c
729
mmap_read_unlock(current->mm);
arch/parisc/kernel/traps.c
756
task_pid_nr(current), current->comm);
arch/parisc/kernel/traps.c
773
task_pid_nr(current), current->comm);
arch/parisc/kernel/unaligned.c
386
if (current->thread.flags & PARISC_UAC_SIGBUS) {
arch/parisc/kernel/unaligned.c
390
if (!(current->thread.flags & PARISC_UAC_NOPRINT) &&
arch/parisc/kernel/unaligned.c
394
current->comm, task_pid_nr(current), regs->ior,
arch/parisc/kernel/unaligned.c
405
if (!(current->thread.flags & PARISC_UAC_NOPRINT) &&
arch/parisc/kernel/unwind.c
430
task = task ? task : current;
arch/parisc/kernel/unwind.c
432
if (task == current) {
arch/parisc/kernel/unwind.c
484
unwind_frame_init_task(&info, current, NULL);
arch/parisc/kernel/vdso.c
32
current->mm->context.vdso_base = vma->vm_start;
arch/parisc/kernel/vdso.c
58
struct mm_struct *mm = current->mm;
arch/parisc/kernel/vdso.c
77
if (current->flags & PF_RANDOMIZE)
arch/parisc/kernel/vdso.c
91
current->mm->context.vdso_base = vdso_text_start;
arch/parisc/mm/fault.c
278
tsk = current;
arch/parisc/mm/fault.c
491
tsk = current;
arch/powerpc/include/asm/book3s/32/kup.h
107
current->thread.kuap = (__force u32)to;
arch/powerpc/include/asm/book3s/32/kup.h
113
u32 kuap = current->thread.kuap;
arch/powerpc/include/asm/book3s/32/kup.h
120
current->thread.kuap = KUAP_NONE;
arch/powerpc/include/asm/book3s/32/kup.h
126
unsigned long flags = current->thread.kuap;
arch/powerpc/include/asm/book3s/32/kup.h
129
current->thread.kuap = KUAP_NONE;
arch/powerpc/include/asm/book3s/32/kup.h
139
current->thread.kuap = flags;
arch/powerpc/include/asm/book3s/32/kup.h
60
unsigned long kuap = current->thread.kuap;
arch/powerpc/include/asm/book3s/32/kup.h
66
current->thread.kuap = KUAP_NONE;
arch/powerpc/include/asm/book3s/32/kup.h
78
current->thread.kuap = KUAP_NONE;
arch/powerpc/include/asm/book3s/32/kup.h
85
current->thread.kuap = regs->kuap;
arch/powerpc/include/asm/book3s/32/kup.h
92
unsigned long kuap = current->thread.kuap;
arch/powerpc/include/asm/book3s/64/kup.h
218
if (current->thread.regs)
arch/powerpc/include/asm/book3s/64/kup.h
219
return current->thread.regs->amr;
arch/powerpc/include/asm/book3s/64/kup.h
225
if (current->thread.regs)
arch/powerpc/include/asm/book3s/64/kup.h
226
return current->thread.regs->iamr;
arch/powerpc/include/asm/cputime.h
48
struct cpu_accounting_data *acct = raw_get_accounting(current);
arch/powerpc/include/asm/cputime.h
57
struct cpu_accounting_data *acct = raw_get_accounting(current);
arch/powerpc/include/asm/current.h
33
register struct task_struct *current asm ("r2");
arch/powerpc/include/asm/elf.h
170
VDSO_AUX_ENT(AT_SYSINFO_EHDR, (unsigned long)current->mm->context.vdso);\
arch/powerpc/include/asm/elf.h
93
if (personality(current->personality) != PER_LINUX32) \
arch/powerpc/include/asm/elf.h
95
(current->personality & (~PER_MASK))); \
arch/powerpc/include/asm/interrupt.h
144
unsigned long dbcr0 = current->thread.debug.dbcr0;
arch/powerpc/include/asm/mmu_context.h
244
switch_mm_irqs_off(prev, next, current);
arch/powerpc/include/asm/nohash/kup-booke.h
41
mtspr(SPRN_PID, current->thread.pid);
arch/powerpc/include/asm/nohash/kup-booke.h
49
mtspr(SPRN_PID, current->thread.pid);
arch/powerpc/include/asm/nohash/kup-booke.h
78
uaccess_begin_booke(current->thread.pid);
arch/powerpc/include/asm/nohash/kup-booke.h
98
uaccess_begin_booke(current->thread.pid);
arch/powerpc/include/asm/ptrace.h
234
((struct pt_regs *)((unsigned long)task_stack_page(current) + THREAD_SIZE) - 1)
arch/powerpc/include/asm/sfp-machine.h
108
#define __FPU_FPSCR (current->thread.spefscr)
arch/powerpc/include/asm/sfp-machine.h
128
#define __FPU_FPSCR (current->thread.fp_state.fpscr)
arch/powerpc/include/asm/stackprotector.h
24
current->stack_canary = canary;
arch/powerpc/kernel/align.c
117
unsigned long *evr = &current->thread.evr[reg];
arch/powerpc/kernel/align.c
133
flush_spe_to_thread(current);
arch/powerpc/kernel/dexcr.c
19
current->thread.dexcr_onexec = mfspr(SPRN_DEXCR);
arch/powerpc/kernel/interrupt.c
84
unsigned long dbcr0 = current->thread.debug.dbcr0;
arch/powerpc/kernel/kgdb.c
333
if (current)
arch/powerpc/kernel/kgdb.c
334
memcpy(mem, &current->thread.evr[regno-32],
arch/powerpc/kernel/kgdb.c
359
memcpy(&current->thread.evr[regno-32], mem,
arch/powerpc/kernel/mce.c
583
level, evt->cpu, current->pid, current->comm,
arch/powerpc/kernel/mce_power.c
37
mm = current->mm;
arch/powerpc/kernel/pci-common.c
680
current->comm, current->pid, pci_domain_nr(bus), bus->number);
arch/powerpc/kernel/process.c
1133
msr_diff = current->thread.ckpt_regs.msr & ~regs->msr;
arch/powerpc/kernel/process.c
1138
current->thread.load_fp = 1;
arch/powerpc/kernel/process.c
1141
current->thread.load_vec = 1;
arch/powerpc/kernel/process.c
1197
if (!current->thread.regs)
arch/powerpc/kernel/process.c
1200
usermsr = current->thread.regs->msr;
arch/powerpc/kernel/process.c
1204
__giveup_fpu(current);
arch/powerpc/kernel/process.c
1206
__giveup_altivec(current);
arch/powerpc/kernel/process.c
1210
current->thread.tm_tfhar = mfspr(SPRN_TFHAR);
arch/powerpc/kernel/process.c
1211
current->thread.tm_tfiar = mfspr(SPRN_TFIAR);
arch/powerpc/kernel/process.c
1212
current->thread.tm_texasr = mfspr(SPRN_TEXASR);
arch/powerpc/kernel/process.c
1213
current->thread.regs->msr &= ~MSR_TM;
arch/powerpc/kernel/process.c
1221
save_sprs(&current->thread);
arch/powerpc/kernel/process.c
1286
old_thread = &current->thread;
arch/powerpc/kernel/process.c
1366
if (current->thread.regs)
arch/powerpc/kernel/process.c
1367
restore_math(current->thread.regs);
arch/powerpc/kernel/process.c
1438
pr_info("%s[%d]: code: %s\n", current->comm,
arch/powerpc/kernel/process.c
1439
current->pid, s.buffer);
arch/powerpc/kernel/process.c
1583
show_stack(current, (unsigned long *) regs->gpr[1], KERN_DEFAULT);
arch/powerpc/kernel/process.c
1591
flush_ptrace_hw_breakpoint(current);
arch/powerpc/kernel/process.c
1593
set_debug_reg_defaults(&current->thread);
arch/powerpc/kernel/process.c
1608
if (!current->thread.regs) {
arch/powerpc/kernel/process.c
1609
struct pt_regs *regs = task_stack_page(current) + THREAD_SIZE;
arch/powerpc/kernel/process.c
1610
current->thread.regs = regs - 1;
arch/powerpc/kernel/process.c
1614
current->thread.regs->amr = default_amr;
arch/powerpc/kernel/process.c
1615
current->thread.regs->iamr = default_iamr;
arch/powerpc/kernel/process.c
1620
current->thread.dexcr = current->thread.dexcr_onexec;
arch/powerpc/kernel/process.c
1621
mtspr(SPRN_DEXCR, current->thread.dexcr);
arch/powerpc/kernel/process.c
1665
if (t != current)
arch/powerpc/kernel/process.c
1859
p->thread.dscr_inherit = current->thread.dscr_inherit;
arch/powerpc/kernel/process.c
1867
p->thread.hashkeyr = current->thread.hashkeyr;
arch/powerpc/kernel/process.c
1955
current->thread.used_vsr = 0;
arch/powerpc/kernel/process.c
1957
current->thread.load_slb = 0;
arch/powerpc/kernel/process.c
1958
current->thread.load_fp = 0;
arch/powerpc/kernel/process.c
1960
memset(&current->thread.fp_state, 0, sizeof(current->thread.fp_state));
arch/powerpc/kernel/process.c
1961
current->thread.fp_save_area = NULL;
arch/powerpc/kernel/process.c
1964
memset(&current->thread.vr_state, 0, sizeof(current->thread.vr_state));
arch/powerpc/kernel/process.c
1965
current->thread.vr_state.vscr.u[3] = 0x00010000; /* Java mode disabled */
arch/powerpc/kernel/process.c
1966
current->thread.vr_save_area = NULL;
arch/powerpc/kernel/process.c
1967
current->thread.vrsave = 0;
arch/powerpc/kernel/process.c
1968
current->thread.used_vr = 0;
arch/powerpc/kernel/process.c
1969
current->thread.load_vec = 0;
arch/powerpc/kernel/process.c
197
BUG_ON(tsk != current);
arch/powerpc/kernel/process.c
1972
memset(current->thread.evr, 0, sizeof(current->thread.evr));
arch/powerpc/kernel/process.c
1973
current->thread.acc = 0;
arch/powerpc/kernel/process.c
1974
current->thread.spefscr = 0;
arch/powerpc/kernel/process.c
1975
current->thread.used_spe = 0;
arch/powerpc/kernel/process.c
1978
current->thread.tm_tfhar = 0;
arch/powerpc/kernel/process.c
1979
current->thread.tm_texasr = 0;
arch/powerpc/kernel/process.c
1980
current->thread.tm_tfiar = 0;
arch/powerpc/kernel/process.c
1981
current->thread.load_tm = 0;
arch/powerpc/kernel/process.c
1985
current->thread.hashkeyr = get_random_long();
arch/powerpc/kernel/process.c
1986
mtspr(SPRN_HASHKEYR, current->thread.hashkeyr);
arch/powerpc/kernel/process.c
213
if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) {
arch/powerpc/kernel/process.c
214
check_if_tm_restore_required(current);
arch/powerpc/kernel/process.c
223
MSR_TM_ACTIVE(current->thread.regs->msr))
arch/powerpc/kernel/process.c
225
__giveup_fpu(current);
arch/powerpc/kernel/process.c
2291
tsk = current;
arch/powerpc/kernel/process.c
2298
if (tsk == current)
arch/powerpc/kernel/process.c
2316
ret_addr = ftrace_graph_ret_addr(current,
arch/powerpc/kernel/process.c
2408
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/powerpc/kernel/process.c
264
if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) {
arch/powerpc/kernel/process.c
265
check_if_tm_restore_required(current);
arch/powerpc/kernel/process.c
274
MSR_TM_ACTIVE(current->thread.regs->msr))
arch/powerpc/kernel/process.c
276
__giveup_altivec(current);
arch/powerpc/kernel/process.c
290
BUG_ON(tsk != current);
arch/powerpc/kernel/process.c
334
if (current->thread.regs &&
arch/powerpc/kernel/process.c
335
(current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) {
arch/powerpc/kernel/process.c
336
check_if_tm_restore_required(current);
arch/powerpc/kernel/process.c
345
MSR_TM_ACTIVE(current->thread.regs->msr))
arch/powerpc/kernel/process.c
347
__giveup_vsx(current);
arch/powerpc/kernel/process.c
357
BUG_ON(tsk != current);
arch/powerpc/kernel/process.c
383
if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) {
arch/powerpc/kernel/process.c
384
check_if_tm_restore_required(current);
arch/powerpc/kernel/process.c
385
__giveup_spe(current);
arch/powerpc/kernel/process.c
395
BUG_ON(tsk != current);
arch/powerpc/kernel/process.c
454
if (current->thread.load_fp) {
arch/powerpc/kernel/process.c
455
current->thread.load_fp++;
arch/powerpc/kernel/process.c
463
load_fp_state(&current->thread.fp_state);
arch/powerpc/kernel/process.c
473
if (cpu_has_feature(CPU_FTR_ALTIVEC) && (current->thread.load_vec)) {
arch/powerpc/kernel/process.c
474
current->thread.load_vec++;
arch/powerpc/kernel/process.c
482
load_vr_state(&current->thread.vr_state);
arch/powerpc/kernel/process.c
483
current->thread.used_vr = 1;
arch/powerpc/kernel/process.c
499
current->thread.used_vsr = 1;
arch/powerpc/kernel/process.c
547
fpexc_mode = current->thread.fpexc_mode;
arch/powerpc/kernel/process.c
595
BUG_ON(tsk != current);
arch/powerpc/kernel/process.c
611
current->thread.trap_nr = TRAP_HWBKPT;
arch/powerpc/kernel/process.c
638
current->thread.hw_brk[0] = null_brk;
arch/powerpc/kernel/process.c
639
current->thread.hw_brk[0].flags |= HW_BRK_FLAG_DISABLED;
arch/powerpc/kernel/process.c
647
info = &current->thread.hw_brk[i];
arch/powerpc/kernel/process.c
653
current->thread.hw_brk[i] = null_brk;
arch/powerpc/kernel/process.c
654
current->thread.hw_brk[i].flags |= HW_BRK_FLAG_DISABLED;
arch/powerpc/kernel/process.c
661
current->thread.trap_nr = TRAP_HWBKPT;
arch/powerpc/kernel/process.c
751
if ((current->thread.debug.dbcr0 & DBCR0_IDM)
arch/powerpc/kernel/process.c
91
if (tsk == current && tsk->thread.regs &&
arch/powerpc/kernel/process.c
978
tm_reclaim_thread(&current->thread, cause);
arch/powerpc/kernel/ptrace/ptrace-tm.c
21
if (!cpu_has_feature(CPU_FTR_TM) || tsk != current)
arch/powerpc/kernel/ptrace/ptrace32.c
269
child, task_user_regset_view(current), 0,
arch/powerpc/kernel/ptrace/ptrace32.c
275
child, task_user_regset_view(current), 0,
arch/powerpc/kernel/rtas.c
1845
func->name, nargs, current->comm);
arch/powerpc/kernel/setup_64.c
398
task_thread_info(current)->cpu = boot_cpuid; // fix task_cpu(current)
arch/powerpc/kernel/signal.c
249
BUG_ON(tsk != current);
arch/powerpc/kernel/signal.c
301
klp_update_patch_state(current);
arch/powerpc/kernel/signal.c
304
BUG_ON(regs != current->thread.regs);
arch/powerpc/kernel/signal.c
305
do_signal(current);
arch/powerpc/kernel/signal.c
339
BUG_ON(tsk != current);
arch/powerpc/kernel/signal_32.c
1057
unsafe_put_sigset_t(&old_ctx->uc_sigmask, &current->blocked, failed);
arch/powerpc/kernel/signal_32.c
1106
current->restart_block.fn = do_no_restart_syscall;
arch/powerpc/kernel/signal_32.c
1186
signal_fault(current, regs, "sys_rt_sigreturn", rt_sf);
arch/powerpc/kernel/signal_32.c
1201
unsigned long new_dbcr0 = current->thread.debug.dbcr0;
arch/powerpc/kernel/signal_32.c
1216
current->thread.debug.dbcr1)) {
arch/powerpc/kernel/signal_32.c
1251
current->thread.debug.dbcr0 = new_dbcr0;
arch/powerpc/kernel/signal_32.c
1270
signal_fault(current, regs, "sys_debug_setcontext", ctx);
arch/powerpc/kernel/signal_32.c
1311
current->restart_block.fn = do_no_restart_syscall;
arch/powerpc/kernel/signal_32.c
1344
signal_fault(current, regs, "sys_sigreturn", sr);
arch/powerpc/kernel/signal_32.c
1355
signal_fault(current, regs, "sys_sigreturn", sc);
arch/powerpc/kernel/signal_32.c
250
flush_fp_to_thread(current);
arch/powerpc/kernel/signal_32.c
252
if (current->thread.used_vr)
arch/powerpc/kernel/signal_32.c
253
flush_altivec_to_thread(current);
arch/powerpc/kernel/signal_32.c
255
current->thread.vrsave = mfspr(SPRN_VRSAVE);
arch/powerpc/kernel/signal_32.c
258
if (current->thread.used_vsr && ctx_has_vsx_region)
arch/powerpc/kernel/signal_32.c
259
flush_vsx_to_thread(current);
arch/powerpc/kernel/signal_32.c
262
if (current->thread.used_spe)
arch/powerpc/kernel/signal_32.c
263
flush_spe_to_thread(current);
arch/powerpc/kernel/signal_32.c
278
if (current->thread.used_vr) {
arch/powerpc/kernel/signal_32.c
279
unsafe_copy_to_user(&frame->mc_vregs, &current->thread.vr_state,
arch/powerpc/kernel/signal_32.c
293
unsafe_put_user(current->thread.vrsave, (u32 __user *)&frame->mc_vregs[32],
arch/powerpc/kernel/signal_32.c
296
unsafe_copy_fpr_to_user(&frame->mc_fregs, current, failed);
arch/powerpc/kernel/signal_32.c
310
if (current->thread.used_vsr && ctx_has_vsx_region) {
arch/powerpc/kernel/signal_32.c
311
unsafe_copy_vsx_to_user(&frame->mc_vsregs, current, failed);
arch/powerpc/kernel/signal_32.c
317
if (current->thread.used_spe) {
arch/powerpc/kernel/signal_32.c
318
unsafe_copy_to_user(&frame->mc_vregs, current->thread.evr,
arch/powerpc/kernel/signal_32.c
327
unsafe_put_user(current->thread.spefscr,
arch/powerpc/kernel/signal_32.c
365
current->thread.ckvrsave = mfspr(SPRN_VRSAVE);
arch/powerpc/kernel/signal_32.c
373
unsafe_save_general_regs(&current->thread.ckpt_regs, frame, failed);
arch/powerpc/kernel/signal_32.c
385
if (current->thread.used_vr) {
arch/powerpc/kernel/signal_32.c
386
unsafe_copy_to_user(&frame->mc_vregs, &current->thread.ckvr_state,
arch/powerpc/kernel/signal_32.c
390
&current->thread.vr_state,
arch/powerpc/kernel/signal_32.c
394
&current->thread.ckvr_state,
arch/powerpc/kernel/signal_32.c
408
unsafe_put_user(current->thread.ckvrsave,
arch/powerpc/kernel/signal_32.c
411
unsafe_put_user(current->thread.vrsave,
arch/powerpc/kernel/signal_32.c
414
unsafe_put_user(current->thread.ckvrsave,
arch/powerpc/kernel/signal_32.c
417
unsafe_copy_ckfpr_to_user(&frame->mc_fregs, current, failed);
arch/powerpc/kernel/signal_32.c
419
unsafe_copy_fpr_to_user(&tm_frame->mc_fregs, current, failed);
arch/powerpc/kernel/signal_32.c
421
unsafe_copy_ckfpr_to_user(&tm_frame->mc_fregs, current, failed);
arch/powerpc/kernel/signal_32.c
429
if (current->thread.used_vsr) {
arch/powerpc/kernel/signal_32.c
430
unsafe_copy_ckvsx_to_user(&frame->mc_vsregs, current, failed);
arch/powerpc/kernel/signal_32.c
432
unsafe_copy_vsx_to_user(&tm_frame->mc_vsregs, current, failed);
arch/powerpc/kernel/signal_32.c
434
unsafe_copy_ckvsx_to_user(&tm_frame->mc_vsregs, current, failed);
arch/powerpc/kernel/signal_32.c
501
unsafe_copy_from_user(&current->thread.vr_state, &sr->mc_vregs,
arch/powerpc/kernel/signal_32.c
503
current->thread.used_vr = true;
arch/powerpc/kernel/signal_32.c
504
} else if (current->thread.used_vr)
arch/powerpc/kernel/signal_32.c
505
memset(&current->thread.vr_state, 0,
arch/powerpc/kernel/signal_32.c
509
unsafe_get_user(current->thread.vrsave, (u32 __user *)&sr->mc_vregs[32], failed);
arch/powerpc/kernel/signal_32.c
511
mtspr(SPRN_VRSAVE, current->thread.vrsave);
arch/powerpc/kernel/signal_32.c
513
unsafe_copy_fpr_from_user(current, &sr->mc_fregs, failed);
arch/powerpc/kernel/signal_32.c
526
unsafe_copy_vsx_from_user(current, &sr->mc_vsregs, failed);
arch/powerpc/kernel/signal_32.c
527
current->thread.used_vsr = true;
arch/powerpc/kernel/signal_32.c
528
} else if (current->thread.used_vsr)
arch/powerpc/kernel/signal_32.c
530
current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0;
arch/powerpc/kernel/signal_32.c
544
BUILD_BUG_ON(sizeof(current->thread.spe) != ELF_NEVRREG * sizeof(u32));
arch/powerpc/kernel/signal_32.c
548
unsafe_copy_from_user(&current->thread.spe, &sr->mc_vregs,
arch/powerpc/kernel/signal_32.c
549
sizeof(current->thread.spe), failed);
arch/powerpc/kernel/signal_32.c
550
current->thread.used_spe = true;
arch/powerpc/kernel/signal_32.c
551
} else if (current->thread.used_spe)
arch/powerpc/kernel/signal_32.c
552
memset(&current->thread.spe, 0, sizeof(current->thread.spe));
arch/powerpc/kernel/signal_32.c
555
unsafe_get_user(current->thread.spefscr, (u32 __user *)&sr->mc_vregs + ELF_NEVRREG, failed);
arch/powerpc/kernel/signal_32.c
591
unsafe_restore_general_regs(&current->thread.ckpt_regs, sr, failed);
arch/powerpc/kernel/signal_32.c
592
unsafe_get_user(current->thread.tm_tfhar, &sr->mc_gregs[PT_NIP], failed);
arch/powerpc/kernel/signal_32.c
601
unsafe_copy_from_user(&current->thread.ckvr_state, &sr->mc_vregs,
arch/powerpc/kernel/signal_32.c
603
current->thread.used_vr = true;
arch/powerpc/kernel/signal_32.c
604
} else if (current->thread.used_vr) {
arch/powerpc/kernel/signal_32.c
605
memset(&current->thread.vr_state, 0,
arch/powerpc/kernel/signal_32.c
607
memset(&current->thread.ckvr_state, 0,
arch/powerpc/kernel/signal_32.c
612
unsafe_get_user(current->thread.ckvrsave,
arch/powerpc/kernel/signal_32.c
615
mtspr(SPRN_VRSAVE, current->thread.ckvrsave);
arch/powerpc/kernel/signal_32.c
619
unsafe_copy_fpr_from_user(current, &sr->mc_fregs, failed);
arch/powerpc/kernel/signal_32.c
627
unsafe_copy_ckvsx_from_user(current, &sr->mc_vsregs, failed);
arch/powerpc/kernel/signal_32.c
628
current->thread.used_vsr = true;
arch/powerpc/kernel/signal_32.c
629
} else if (current->thread.used_vsr)
arch/powerpc/kernel/signal_32.c
631
current->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = 0;
arch/powerpc/kernel/signal_32.c
632
current->thread.ckfp_state.fpr[i][TS_VSRLOWOFFSET] = 0;
arch/powerpc/kernel/signal_32.c
644
unsafe_copy_from_user(&current->thread.vr_state, &tm_sr->mc_vregs,
arch/powerpc/kernel/signal_32.c
648
unsafe_get_user(current->thread.vrsave,
arch/powerpc/kernel/signal_32.c
651
unsafe_copy_ckfpr_from_user(current, &tm_sr->mc_fregs, failed);
arch/powerpc/kernel/signal_32.c
658
unsafe_copy_vsx_from_user(current, &tm_sr->mc_vsregs, failed);
arch/powerpc/kernel/signal_32.c
659
current->thread.used_vsr = true;
arch/powerpc/kernel/signal_32.c
695
current->thread.tm_texasr |= TEXASR_FS;
arch/powerpc/kernel/signal_32.c
697
tm_recheckpoint(&current->thread);
arch/powerpc/kernel/signal_32.c
702
load_fp_state(&current->thread.fp_state);
arch/powerpc/kernel/signal_32.c
703
regs_set_return_msr(regs, regs->msr | (MSR_FP | current->thread.fpexc_mode));
arch/powerpc/kernel/signal_32.c
706
load_vr_state(&current->thread.vr_state);
arch/powerpc/kernel/signal_64.c
134
BUG_ON(tsk != current);
arch/powerpc/kernel/signal_64.c
227
BUG_ON(tsk != current);
arch/powerpc/kernel/signal_64.c
350
BUG_ON(tsk != current);
arch/powerpc/kernel/signal_64.c
451
BUG_ON(tsk != current);
arch/powerpc/kernel/signal_64.c
523
current->thread.used_vr = true;
arch/powerpc/kernel/signal_64.c
685
prepare_setup_sigcontext(current);
arch/powerpc/kernel/signal_64.c
689
unsafe_setup_sigcontext(&old_ctx->uc_mcontext, current, 0, NULL,
arch/powerpc/kernel/signal_64.c
691
unsafe_copy_to_user(&old_ctx->uc_sigmask, &current->blocked,
arch/powerpc/kernel/signal_64.c
722
if (__unsafe_restore_sigcontext(current, NULL, 0, &new_ctx->uc_mcontext)) {
arch/powerpc/kernel/signal_64.c
752
current->restart_block.fn = do_no_restart_syscall;
arch/powerpc/kernel/signal_64.c
814
if (restore_tm_sigcontexts(current, &uc->uc_mcontext,
arch/powerpc/kernel/signal_64.c
830
regs_set_return_msr(current->thread.regs,
arch/powerpc/kernel/signal_64.c
831
current->thread.regs->msr & ~MSR_TS_MASK);
arch/powerpc/kernel/signal_64.c
835
unsafe_restore_sigcontext(current, NULL, 1, &uc->uc_mcontext,
arch/powerpc/kernel/signal_64.c
851
signal_fault(current, regs, "rt_sigreturn", uc);
arch/powerpc/kernel/signal_64.c
974
signal_fault(current, regs, "handle_rt_signal64", frame);
arch/powerpc/kernel/smp.c
1194
paca_ptrs[boot_cpuid]->__current = current;
arch/powerpc/kernel/smp.c
1197
current_set[boot_cpuid] = current;
arch/powerpc/kernel/smp.c
1649
current->active_mm = &init_mm;
arch/powerpc/kernel/stacktrace.c
38
else if (task == current)
arch/powerpc/kernel/stacktrace.c
84
if (task == current)
arch/powerpc/kernel/swsusp.c
20
flush_all_to_thread(current);
arch/powerpc/kernel/swsusp.c
31
switch_mmu_context(current->active_mm, current->active_mm, NULL);
arch/powerpc/kernel/syscalls.c
115
regs_set_return_msr(current->thread.regs,
arch/powerpc/kernel/syscalls.c
116
current->thread.regs->msr ^ MSR_LE);
arch/powerpc/kernel/syscalls.c
81
if (personality(current->personality) == PER_LINUX32
arch/powerpc/kernel/sysfs.c
161
if (!current->thread.dscr_inherit) {
arch/powerpc/kernel/sysfs.c
162
current->thread.dscr = *(unsigned long *)val;
arch/powerpc/kernel/sysfs.c
56
current->comm, current->pid);
arch/powerpc/kernel/sysfs.c
65
current->comm, current->pid);
arch/powerpc/kernel/time.c
373
struct cpu_accounting_data *acct = get_accounting(current);
arch/powerpc/kernel/trace/ftrace.c
687
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/powerpc/kernel/trace/ftrace_64_pg.c
796
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/powerpc/kernel/traps.c
1199
flush_fp_to_thread(current);
arch/powerpc/kernel/traps.c
1202
code = __parse_fpscr(current->thread.fp_state.fpscr);
arch/powerpc/kernel/traps.c
1431
current->thread.dscr = regs->gpr[rd];
arch/powerpc/kernel/traps.c
1432
current->thread.dscr_inherit = 1;
arch/powerpc/kernel/traps.c
1433
mtspr(SPRN_DSCR, current->thread.dscr);
arch/powerpc/kernel/traps.c
146
if (kexec_should_crash(current))
arch/powerpc/kernel/traps.c
1463
code = __parse_fpscr(current->thread.fp_state.fpscr);
arch/powerpc/kernel/traps.c
149
!current->pid || is_global_init(current))
arch/powerpc/kernel/traps.c
1658
if (!(current->thread.align_ctl & PR_UNALIGN_SIGBUS))
arch/powerpc/kernel/traps.c
1728
current->thread.load_tm++;
arch/powerpc/kernel/traps.c
1731
tm_restore_sprs(&current->thread);
arch/powerpc/kernel/traps.c
1808
current->thread.dscr = regs->gpr[rd];
arch/powerpc/kernel/traps.c
1809
current->thread.dscr_inherit = 1;
arch/powerpc/kernel/traps.c
1810
current->thread.fscr |= FSCR_DSCR;
arch/powerpc/kernel/traps.c
1811
mtspr(SPRN_FSCR, current->thread.fscr);
arch/powerpc/kernel/traps.c
1883
current->thread.load_fp = 1;
arch/powerpc/kernel/traps.c
1888
tm_recheckpoint(&current->thread);
arch/powerpc/kernel/traps.c
1901
current->thread.load_vec = 1;
arch/powerpc/kernel/traps.c
1902
tm_recheckpoint(&current->thread);
arch/powerpc/kernel/traps.c
1903
current->thread.used_vr = 1;
arch/powerpc/kernel/traps.c
1919
current->thread.used_vsr = 1;
arch/powerpc/kernel/traps.c
1924
current->thread.load_vec = 1;
arch/powerpc/kernel/traps.c
1925
current->thread.load_fp = 1;
arch/powerpc/kernel/traps.c
1927
tm_recheckpoint(&current->thread);
arch/powerpc/kernel/traps.c
1976
dbcr_dac(current) &= ~(DBCR_DAC1R | DBCR_DAC1W);
arch/powerpc/kernel/traps.c
1978
current->thread.debug.dbcr2 &= ~DBCR2_DAC12MODE;
arch/powerpc/kernel/traps.c
1984
dbcr_dac(current) &= ~(DBCR_DAC2R | DBCR_DAC2W);
arch/powerpc/kernel/traps.c
1989
current->thread.debug.dbcr0 &= ~DBCR0_IAC1;
arch/powerpc/kernel/traps.c
1990
dbcr_iac_range(current) &= ~DBCR_IAC12MODE;
arch/powerpc/kernel/traps.c
1995
current->thread.debug.dbcr0 &= ~DBCR0_IAC2;
arch/powerpc/kernel/traps.c
2000
current->thread.debug.dbcr0 &= ~DBCR0_IAC3;
arch/powerpc/kernel/traps.c
2001
dbcr_iac_range(current) &= ~DBCR_IAC34MODE;
arch/powerpc/kernel/traps.c
2006
current->thread.debug.dbcr0 &= ~DBCR0_IAC4;
arch/powerpc/kernel/traps.c
2016
if (DBCR_ACTIVE_EVENTS(current->thread.debug.dbcr0,
arch/powerpc/kernel/traps.c
2017
current->thread.debug.dbcr1))
arch/powerpc/kernel/traps.c
2021
current->thread.debug.dbcr0 &= ~DBCR0_IDM;
arch/powerpc/kernel/traps.c
2024
mtspr(SPRN_DBCR0, current->thread.debug.dbcr0);
arch/powerpc/kernel/traps.c
2031
current->thread.debug.dbsr = debug_status;
arch/powerpc/kernel/traps.c
2048
current->thread.debug.dbcr0 &= ~DBCR0_BT;
arch/powerpc/kernel/traps.c
2049
current->thread.debug.dbcr0 |= DBCR0_IDM | DBCR0_IC;
arch/powerpc/kernel/traps.c
2083
current->thread.debug.dbcr0 &= ~DBCR0_IC;
arch/powerpc/kernel/traps.c
2084
if (DBCR_ACTIVE_EVENTS(current->thread.debug.dbcr0,
arch/powerpc/kernel/traps.c
2085
current->thread.debug.dbcr1))
arch/powerpc/kernel/traps.c
2089
current->thread.debug.dbcr0 &= ~DBCR0_IDM;
arch/powerpc/kernel/traps.c
2109
flush_altivec_to_thread(current);
arch/powerpc/kernel/traps.c
2126
"in %s at %lx\n", current->comm, regs->nip);
arch/powerpc/kernel/traps.c
2127
current->thread.vr_state.vscr.u[3] |= 0x10000;
arch/powerpc/kernel/traps.c
2157
flush_spe_to_thread(current);
arch/powerpc/kernel/traps.c
2159
spefscr = current->thread.spefscr;
arch/powerpc/kernel/traps.c
2160
fpexc_mode = current->thread.fpexc_mode;
arch/powerpc/kernel/traps.c
2189
"in %s at %lx\n", current->comm, regs->nip);
arch/powerpc/kernel/traps.c
2205
giveup_spe(current);
arch/powerpc/kernel/traps.c
2222
"in %s at %lx\n", current->comm, regs->nip);
arch/powerpc/kernel/traps.c
230
if (kexec_should_crash(current))
arch/powerpc/kernel/traps.c
2307
pr_warn_ratelimited("%s used emulated %s instruction\n", current->comm,
arch/powerpc/kernel/traps.c
242
if (in_interrupt() || panic_on_oops || !current->pid ||
arch/powerpc/kernel/traps.c
243
is_global_init(current)) {
arch/powerpc/kernel/traps.c
318
if (!unhandled_signal(current, signr))
arch/powerpc/kernel/traps.c
325
current->comm, current->pid, signame(signr), signr,
arch/powerpc/kernel/traps.c
354
current->thread.trap_nr = code;
arch/powerpc/kernel/traps.c
581
#define single_stepping(regs) (current->thread.debug.dbcr0 & DBCR0_IC)
arch/powerpc/kernel/traps.c
582
#define clear_single_step(regs) (current->thread.debug.dbcr0 &= ~DBCR0_IC)
arch/powerpc/kernel/traps.c
904
smp_processor_id(), current->comm, current->pid,
arch/powerpc/kernel/traps.c
911
flush_vsx_to_thread(current);
arch/powerpc/kernel/traps.c
925
vdst = (u8 *)&current->thread.vr_state.vr[t];
arch/powerpc/kernel/traps.c
927
vdst = (u8 *)&current->thread.fp_state.fpr[t][0];
arch/powerpc/kernel/traps.c
939
smp_processor_id(), current->comm, current->pid,
arch/powerpc/kernel/traps.c
954
smp_processor_id(), current->comm, current->pid,
arch/powerpc/kernel/traps.c
961
smp_processor_id(), current->comm, current->pid, regs->nip,
arch/powerpc/kernel/traps.c
980
smp_processor_id(), current->comm, current->pid,
arch/powerpc/kernel/uprobes.c
114
struct uprobe_task *utask = current->utask;
arch/powerpc/kernel/uprobes.c
116
WARN_ON_ONCE(current->thread.trap_nr != UPROBE_TRAP_NR);
arch/powerpc/kernel/uprobes.c
118
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/powerpc/kernel/uprobes.c
129
user_disable_single_step(current);
arch/powerpc/kernel/uprobes.c
170
struct uprobe_task *utask = current->utask;
arch/powerpc/kernel/uprobes.c
172
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/powerpc/kernel/uprobes.c
175
user_disable_single_step(current);
arch/powerpc/kernel/uprobes.c
66
struct arch_uprobe_task *autask = &current->utask->autask;
arch/powerpc/kernel/uprobes.c
68
autask->saved_trap_nr = current->thread.trap_nr;
arch/powerpc/kernel/uprobes.c
69
current->thread.trap_nr = UPROBE_TRAP_NR;
arch/powerpc/kernel/uprobes.c
70
regs_set_return_ip(regs, current->utask->xol_vaddr);
arch/powerpc/kernel/uprobes.c
72
user_enable_single_step(current);
arch/powerpc/kernel/vdso.c
103
struct mm_struct *mm = current->mm;
arch/powerpc/kernel/vdso.c
158
struct mm_struct *mm = current->mm;
arch/powerpc/kernel/vdso.c
52
current->mm->context.vdso = (void __user *)new_vma->vm_start;
arch/powerpc/kernel/vecemu.c
280
vrs = current->thread.vr_state.vr;
arch/powerpc/kernel/vecemu.c
329
&current->thread.vr_state.vscr.u[3]);
arch/powerpc/kernel/vecemu.c
334
&current->thread.vr_state.vscr.u[3]);
arch/powerpc/kernel/watchdog.c
157
print_irqtrace_events(current);
arch/powerpc/kernel/watchdog.c
275
print_irqtrace_events(current);
arch/powerpc/kernel/watchdog.c
416
print_irqtrace_events(current);
arch/powerpc/kvm/book3s_hv.c
369
if (task_is_running(current) &&
arch/powerpc/kvm/book3s_hv.c
4758
vcpu->arch.run_task = current;
arch/powerpc/kvm/book3s_hv.c
4770
if (!signal_pending(current)) {
arch/powerpc/kvm/book3s_hv.c
4784
!signal_pending(current)) {
arch/powerpc/kvm/book3s_hv.c
4889
vcpu->arch.run_task = current;
arch/powerpc/kvm/book3s_hv.c
4918
if (signal_pending(current))
arch/powerpc/kvm/book3s_hv.c
5037
r = kvmppc_handle_exit_hv(vcpu, current);
arch/powerpc/kvm/book3s_hv.c
5049
if (signal_pending(current)) {
arch/powerpc/kvm/book3s_hv.c
5101
if (signal_pending(current)) {
arch/powerpc/kvm/book3s_hv.c
5111
if (cpu_has_feature(CPU_FTR_TM) && current->thread.regs &&
arch/powerpc/kvm/book3s_hv.c
5112
(current->thread.regs->msr & MSR_TM)) {
arch/powerpc/kvm/book3s_hv.c
5113
if (MSR_TM_ACTIVE(current->thread.regs->msr)) {
arch/powerpc/kvm/book3s_hv_p9_entry.c
169
current->thread.tidr != vcpu->arch.tid)
arch/powerpc/kvm/book3s_hv_p9_entry.c
170
mtspr(SPRN_TIDR, current->thread.tidr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
177
if (current->thread.fscr != vcpu->arch.fscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
178
mtspr(SPRN_FSCR, current->thread.fscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
179
if (current->thread.dscr != vcpu->arch.dscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
18
current->thread.vrsave != vcpu->arch.vrsave)
arch/powerpc/kvm/book3s_hv_p9_entry.c
180
mtspr(SPRN_DSCR, current->thread.dscr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
190
vcpu->arch.vrsave != current->thread.vrsave)
arch/powerpc/kvm/book3s_hv_p9_entry.c
191
mtspr(SPRN_VRSAVE, current->thread.vrsave);
arch/powerpc/kvm/book3s_hv_p9_entry.c
194
if (vcpu->arch.bescr != current->thread.bescr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
195
mtspr(SPRN_BESCR, current->thread.bescr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
196
if (vcpu->arch.ebbhr != current->thread.ebbhr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
197
mtspr(SPRN_EBBHR, current->thread.ebbhr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
198
if (vcpu->arch.ebbrr != current->thread.ebbrr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
199
mtspr(SPRN_EBBRR, current->thread.ebbrr);
arch/powerpc/kvm/book3s_hv_p9_entry.c
214
if (vcpu->arch.tar != current->thread.tar)
arch/powerpc/kvm/book3s_hv_p9_entry.c
215
mtspr(SPRN_TAR, current->thread.tar);
arch/powerpc/kvm/book3s_hv_p9_entry.c
23
if (current->thread.ebbhr != vcpu->arch.ebbhr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
25
if (current->thread.ebbrr != vcpu->arch.ebbrr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
27
if (current->thread.bescr != vcpu->arch.bescr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
32
current->thread.tidr != vcpu->arch.tid)
arch/powerpc/kvm/book3s_hv_p9_entry.c
40
if (current->thread.fscr != vcpu->arch.fscr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
42
if (current->thread.dscr != vcpu->arch.dscr)
arch/powerpc/kvm/book3s_pr.c
1012
current->thread.tar = mfspr(SPRN_TAR);
arch/powerpc/kvm/book3s_pr.c
145
if (cpu_has_feature(CPU_FTR_ARCH_300) && (current->thread.fscr & FSCR_SCV))
arch/powerpc/kvm/book3s_pr.c
152
current->thread.kvm_shadow_vcpu = vcpu->arch.shadow_vcpu;
arch/powerpc/kvm/book3s_pr.c
176
if (cpu_has_feature(CPU_FTR_ARCH_300) && (current->thread.fscr & FSCR_SCV))
arch/powerpc/kvm/book3s_pr.c
1832
giveup_all(current);
arch/powerpc/kvm/book3s_pr.c
799
struct thread_struct *t = &current->thread;
arch/powerpc/kvm/book3s_pr.c
823
giveup_fpu(current);
arch/powerpc/kvm/book3s_pr.c
829
if (current->thread.regs->msr & MSR_VEC)
arch/powerpc/kvm/book3s_pr.c
830
giveup_altivec(current);
arch/powerpc/kvm/book3s_pr.c
851
mtspr(SPRN_TAR, current->thread.tar);
arch/powerpc/kvm/book3s_pr.c
862
struct thread_struct *t = &current->thread;
arch/powerpc/kvm/book3s_pr.c
934
lost_ext = vcpu->arch.guest_owned_ext & ~current->thread.regs->msr;
arch/powerpc/kvm/book3s_pr.c
954
current->thread.regs->msr |= lost_ext;
arch/powerpc/kvm/booke.c
160
if (!(current->thread.regs->msr & MSR_FP)) {
arch/powerpc/kvm/booke.c
164
current->thread.fp_save_area = &vcpu->arch.fp;
arch/powerpc/kvm/booke.c
165
current->thread.regs->msr |= MSR_FP;
arch/powerpc/kvm/booke.c
177
if (current->thread.regs->msr & MSR_FP)
arch/powerpc/kvm/booke.c
178
giveup_fpu(current);
arch/powerpc/kvm/booke.c
179
current->thread.fp_save_area = NULL;
arch/powerpc/kvm/booke.c
202
if (!(current->thread.regs->msr & MSR_VEC)) {
arch/powerpc/kvm/booke.c
206
current->thread.vr_save_area = &vcpu->arch.vr;
arch/powerpc/kvm/booke.c
207
current->thread.regs->msr |= MSR_VEC;
arch/powerpc/kvm/booke.c
2112
current->thread.kvm_vcpu = vcpu;
arch/powerpc/kvm/booke.c
2117
current->thread.kvm_vcpu = NULL;
arch/powerpc/kvm/booke.c
221
if (current->thread.regs->msr & MSR_VEC)
arch/powerpc/kvm/booke.c
222
giveup_altivec(current);
arch/powerpc/kvm/booke.c
223
current->thread.vr_save_area = NULL;
arch/powerpc/kvm/booke.c
802
debug = current->thread.debug;
arch/powerpc/kvm/booke.c
803
current->thread.debug = vcpu->arch.dbg_reg;
arch/powerpc/kvm/booke.c
815
current->thread.debug = debug;
arch/powerpc/kvm/booke_emulate.c
373
current->thread.debug = vcpu->arch.dbg_reg;
arch/powerpc/kvm/mpic.c
116
struct kvm_vcpu *vcpu = current->thread.kvm_vcpu;
arch/powerpc/kvm/powerpc.c
94
if (signal_pending(current)) {
arch/powerpc/kvm/trace_hv.h
422
__entry->tgid = current->tgid;
arch/powerpc/kvm/trace_hv.h
446
__entry->tgid = current->tgid;
arch/powerpc/kvm/trace_hv.h
468
__entry->tgid = current->tgid;
arch/powerpc/kvm/trace_hv.h
488
__entry->tgid = current->tgid;
arch/powerpc/lib/code-patching.c
101
switch_mm_irqs_off(temp_mm, orig_mm, current);
arch/powerpc/lib/code-patching.c
86
struct mm_struct *orig_mm = current->active_mm;
arch/powerpc/lib/code-patching.c
89
switch_mm_irqs_off(orig_mm, temp_mm, current);
arch/powerpc/lib/sstep.c
1029
buf[j].d[0] = current->thread.fp_state.fpr[reg + i][0];
arch/powerpc/lib/sstep.c
1030
buf[j].d[1] = current->thread.fp_state.fpr[reg + i][1];
arch/powerpc/lib/sstep.c
1042
buf[j].v = current->thread.vr_state.vr[reg - 32 + i];
arch/powerpc/lib/sstep.c
614
current->thread.TS_FPR(rn) = u.l[0];
arch/powerpc/lib/sstep.c
621
current->thread.TS_FPR(rn) = u.l[1];
arch/powerpc/lib/sstep.c
650
u.l[0] = current->thread.TS_FPR(rn);
arch/powerpc/lib/sstep.c
662
u.l[1] = current->thread.TS_FPR(rn);
arch/powerpc/lib/sstep.c
703
current->thread.vr_state.vr[rn] = u.v;
arch/powerpc/lib/sstep.c
729
u.v = current->thread.vr_state.vr[rn];
arch/powerpc/lib/sstep.c
983
current->thread.fp_state.fpr[reg + i][0] = buf[j].d[0];
arch/powerpc/lib/sstep.c
984
current->thread.fp_state.fpr[reg + i][1] = buf[j].d[1];
arch/powerpc/lib/sstep.c
996
current->thread.vr_state.vr[reg - 32 + i] = buf[j].v;
arch/powerpc/math-emu/math.c
332
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
333
op1 = (void *)&current->thread.TS_FPR((insn >> 16) & 0x1f);
arch/powerpc/math-emu/math.c
334
op2 = (void *)&current->thread.TS_FPR((insn >> 11) & 0x1f);
arch/powerpc/math-emu/math.c
338
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
339
op1 = (void *)&current->thread.TS_FPR((insn >> 16) & 0x1f);
arch/powerpc/math-emu/math.c
340
op2 = (void *)&current->thread.TS_FPR((insn >> 6) & 0x1f);
arch/powerpc/math-emu/math.c
344
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
345
op1 = (void *)&current->thread.TS_FPR((insn >> 16) & 0x1f);
arch/powerpc/math-emu/math.c
346
op2 = (void *)&current->thread.TS_FPR((insn >> 11) & 0x1f);
arch/powerpc/math-emu/math.c
347
op3 = (void *)&current->thread.TS_FPR((insn >> 6) & 0x1f);
arch/powerpc/math-emu/math.c
353
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
363
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
368
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
372
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
373
op1 = (void *)&current->thread.TS_FPR((insn >> 16) & 0x1f);
arch/powerpc/math-emu/math.c
377
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
378
op1 = (void *)&current->thread.TS_FPR((insn >> 11) & 0x1f);
arch/powerpc/math-emu/math.c
383
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
392
op0 = (void *)&current->thread.TS_FPR((insn >> 21) & 0x1f);
arch/powerpc/math-emu/math.c
400
op2 = (void *)&current->thread.TS_FPR((insn >> 16) & 0x1f);
arch/powerpc/math-emu/math.c
401
op3 = (void *)&current->thread.TS_FPR((insn >> 11) & 0x1f);
arch/powerpc/math-emu/math.c
421
op1 = (void *)&current->thread.TS_FPR((insn >> 11) & 0x1f);
arch/powerpc/math-emu/math.c
433
flush_fp_to_thread(current);
arch/powerpc/math-emu/math_efp.c
200
vc.wp[0] = current->thread.evr[fc];
arch/powerpc/math-emu/math_efp.c
202
va.wp[0] = current->thread.evr[fa];
arch/powerpc/math-emu/math_efp.c
204
vb.wp[0] = current->thread.evr[fb];
arch/powerpc/math-emu/math_efp.c
680
&= ~(FP_EX_INVALID | FP_EX_UNDERFLOW) | current->thread.spefscr_last;
arch/powerpc/math-emu/math_efp.c
683
current->thread.spefscr_last = __FPU_FPSCR;
arch/powerpc/math-emu/math_efp.c
685
current->thread.evr[fc] = vc.wp[0];
arch/powerpc/math-emu/math_efp.c
695
if (current->thread.fpexc_mode & PR_FP_EXC_SW_ENABLE) {
arch/powerpc/math-emu/math_efp.c
697
&& (current->thread.fpexc_mode & PR_FP_EXC_DIV))
arch/powerpc/math-emu/math_efp.c
700
&& (current->thread.fpexc_mode & PR_FP_EXC_OVF))
arch/powerpc/math-emu/math_efp.c
703
&& (current->thread.fpexc_mode & PR_FP_EXC_UND))
arch/powerpc/math-emu/math_efp.c
706
&& (current->thread.fpexc_mode & PR_FP_EXC_RES))
arch/powerpc/math-emu/math_efp.c
709
&& (current->thread.fpexc_mode & PR_FP_EXC_INV))
arch/powerpc/math-emu/math_efp.c
756
s_hi = current->thread.evr[fc] & SIGN_BIT_S;
arch/powerpc/math-emu/math_efp.c
757
fgpr.wp[0] = current->thread.evr[fc];
arch/powerpc/math-emu/math_efp.c
802
s_hi = current->thread.evr[fb] & SIGN_BIT_S;
arch/powerpc/math-emu/math_efp.c
811
s_hi = current->thread.evr[fb] & SIGN_BIT_S;
arch/powerpc/math-emu/math_efp.c
883
current->thread.evr[fc] = fgpr.wp[0];
arch/powerpc/math-emu/math_efp.c
888
if (current->thread.fpexc_mode & PR_FP_EXC_SW_ENABLE)
arch/powerpc/math-emu/math_efp.c
889
return (current->thread.fpexc_mode & PR_FP_EXC_RES) ? 1 : 0;
arch/powerpc/mm/book3s32/kuap.c
12
current->thread.sr0 |= SR_KS;
arch/powerpc/mm/book3s32/mmu.c
333
if (!current->thread.regs)
arch/powerpc/mm/book3s32/mmu.c
337
if (TRAP(current->thread.regs) != 0x300 && TRAP(current->thread.regs) != 0x400)
arch/powerpc/mm/book3s64/hash_utils.c
1625
if ((get_paca_psize(addr) != MMU_PAGE_4K) && (current->mm == mm)) {
arch/powerpc/mm/book3s64/hash_utils.c
1692
ea, access, current->comm);
arch/powerpc/mm/book3s64/hash_utils.c
1855
if (current->mm == mm)
arch/powerpc/mm/book3s64/hash_utils.c
1901
if (current->mm == mm)
arch/powerpc/mm/book3s64/hash_utils.c
1943
struct mm_struct *mm = current->mm;
arch/powerpc/mm/book3s64/hash_utils.c
1975
mm = current->mm;
arch/powerpc/mm/book3s64/hash_utils.c
2144
trap = current->thread.regs ? TRAP(current->thread.regs) : 0UL;
arch/powerpc/mm/book3s64/hash_utils.c
2170
if (local && cpu_has_feature(CPU_FTR_TM) && current->thread.regs &&
arch/powerpc/mm/book3s64/hash_utils.c
2171
MSR_TM_ACTIVE(current->thread.regs->msr)) {
arch/powerpc/mm/book3s64/mmu_context.c
122
memcpy(mm->context.hash_context, current->mm->context.hash_context, sizeof(struct hash_mm_context));
arch/powerpc/mm/book3s64/mmu_context.c
125
if (current->mm->context.hash_context->spt) {
arch/powerpc/mm/book3s64/pkeys.c
320
current->thread.regs->amr = old_amr | new_amr_bits;
arch/powerpc/mm/book3s64/pkeys.c
331
current->thread.regs->iamr = old_iamr | new_iamr_bits;
arch/powerpc/mm/book3s64/radix_tlb.c
675
if (current->mm == mm)
arch/powerpc/mm/book3s64/radix_tlb.c
678
if (current->active_mm == mm) {
arch/powerpc/mm/book3s64/radix_tlb.c
681
WARN_ON_ONCE(current->mm != NULL);
arch/powerpc/mm/book3s64/radix_tlb.c
689
current->active_mm = &init_mm;
arch/powerpc/mm/book3s64/radix_tlb.c
690
switch_mm_irqs_off(mm, &init_mm, current);
arch/powerpc/mm/book3s64/radix_tlb.c
773
if (current->mm != mm) {
arch/powerpc/mm/book3s64/radix_tlb.c
805
if (atomic_read(&mm->mm_users) <= 1 && current->mm == mm) {
arch/powerpc/mm/book3s64/radix_tlb.c
823
if (current->mm == mm)
arch/powerpc/mm/book3s64/slb.c
783
struct mm_struct *mm = current->mm;
arch/powerpc/mm/book3s64/slice.c
176
if (mm != current->active_mm)
arch/powerpc/mm/book3s64/slice.c
179
copy_mm_to_paca(current->active_mm);
arch/powerpc/mm/book3s64/slice.c
436
struct mm_struct *mm = current->mm;
arch/powerpc/mm/book3s64/slice.c
667
psize = mm_ctx_user_psize(&current->mm->context);
arch/powerpc/mm/book3s64/slice.c
687
psize = mm_ctx_user_psize(&current->mm->context);
arch/powerpc/mm/book3s64/slice.c
747
struct mm_struct *mm = current->mm;
arch/powerpc/mm/book3s64/subpage_prot.c
187
struct mm_struct *mm = current->mm;
arch/powerpc/mm/book3s64/subpage_prot.c
92
struct mm_struct *mm = current->mm;
arch/powerpc/mm/fault.c
146
current->thread.trap_nr = BUS_ADRERR;
arch/powerpc/mm/fault.c
152
current->comm, current->pid, address);
arch/powerpc/mm/fault.c
175
if (fatal_signal_pending(current) && !user_mode(regs))
arch/powerpc/mm/fault.c
325
current->comm, current->pid, address,
arch/powerpc/mm/fault.c
418
struct mm_struct *mm = current->mm;
arch/powerpc/mm/fault.c
568
mmap_read_unlock(current->mm);
arch/powerpc/mm/fault.c
651
if (task_stack_end_corrupted(current))
arch/powerpc/mm/pgtable.c
42
return current->thread.regs && TRAP(current->thread.regs) == 0x400;
arch/powerpc/net/bpf_jit_comp64.c
385
if (!validate_sp(sp, current))
arch/powerpc/perf/callchain.c
108
if (!current->mm)
arch/powerpc/perf/callchain.c
30
if (!validate_sp(sp, current))
arch/powerpc/perf/callchain.c
56
if (!validate_sp(sp, current))
arch/powerpc/perf/callchain.c
64
validate_sp_size(sp, current, STACK_INT_FRAME_SIZE) &&
arch/powerpc/perf/callchain_32.c
62
if (current->mm->context.vdso &&
arch/powerpc/perf/callchain_32.c
63
nip == VDSO32_SYMBOL(current->mm->context.vdso, sigtramp32))
arch/powerpc/perf/callchain_32.c
73
if (current->mm->context.vdso &&
arch/powerpc/perf/callchain_32.c
74
nip == VDSO32_SYMBOL(current->mm->context.vdso, sigtramp_rt32))
arch/powerpc/perf/callchain_64.c
44
if (current->mm->context.vdso &&
arch/powerpc/perf/callchain_64.c
45
nip == VDSO64_SYMBOL(current->mm->context.vdso, sigtramp_rt64))
arch/powerpc/perf/core-book3s.c
614
if (!is_ebb_event(event) || current->thread.used_ebb)
arch/powerpc/perf/core-book3s.c
623
current->thread.used_ebb = 1;
arch/powerpc/perf/core-book3s.c
624
current->thread.mmcr0 |= MMCR0_PMXE;
arch/powerpc/perf/core-book3s.c
632
current->thread.siar = mfspr(SPRN_SIAR);
arch/powerpc/perf/core-book3s.c
633
current->thread.sier = mfspr(SPRN_SIER);
arch/powerpc/perf/core-book3s.c
634
current->thread.sdar = mfspr(SPRN_SDAR);
arch/powerpc/perf/core-book3s.c
635
current->thread.mmcr0 = mmcr0 & MMCR0_USER_MASK;
arch/powerpc/perf/core-book3s.c
636
current->thread.mmcr2 = mfspr(SPRN_MMCR2) & MMCR2_USER_MASK;
arch/powerpc/perf/core-book3s.c
638
current->thread.mmcr3 = mfspr(SPRN_MMCR3);
arch/powerpc/perf/core-book3s.c
639
current->thread.sier2 = mfspr(SPRN_SIER2);
arch/powerpc/perf/core-book3s.c
640
current->thread.sier3 = mfspr(SPRN_SIER3);
arch/powerpc/perf/core-book3s.c
659
mmcr0 |= current->thread.mmcr0;
arch/powerpc/perf/core-book3s.c
666
if (!(current->thread.mmcr0 & MMCR0_PMXE))
arch/powerpc/perf/core-book3s.c
669
mtspr(SPRN_SIAR, current->thread.siar);
arch/powerpc/perf/core-book3s.c
670
mtspr(SPRN_SIER, current->thread.sier);
arch/powerpc/perf/core-book3s.c
671
mtspr(SPRN_SDAR, current->thread.sdar);
arch/powerpc/perf/core-book3s.c
680
mtspr(SPRN_MMCR2, cpuhw->mmcr.mmcr2 | current->thread.mmcr2);
arch/powerpc/perf/core-book3s.c
683
mtspr(SPRN_MMCR3, current->thread.mmcr3);
arch/powerpc/perf/core-book3s.c
684
mtspr(SPRN_SIER2, current->thread.sier2);
arch/powerpc/perf/core-book3s.c
685
mtspr(SPRN_SIER3, current->thread.sier3);
arch/powerpc/perf/core-book3s.c
731
if ((current->thread.mmcr0 & (MMCR0_PMAO | MMCR0_PMAO_SYNC)) != MMCR0_PMAO)
arch/powerpc/perf/core-book3s.c
735
if (ebb && !(current->thread.bescr & BESCR_GE))
arch/powerpc/perf/perf_regs.c
146
regs_user->regs = task_pt_regs(current);
arch/powerpc/perf/perf_regs.c
147
regs_user->abi = (regs_user->regs) ? perf_reg_abi(current) :
arch/powerpc/platforms/83xx/suspend.c
270
if (signal_pending(current) || pci_pm_state < 2)
arch/powerpc/platforms/book3s/vas-api.c
361
struct pt_regs *regs = current->thread.regs;
arch/powerpc/platforms/book3s/vas-api.c
76
task_ref->pid = get_task_pid(current, PIDTYPE_PID);
arch/powerpc/platforms/book3s/vas-api.c
80
task_ref->mm = get_task_mm(current);
arch/powerpc/platforms/book3s/vas-api.c
84
current->pid);
arch/powerpc/platforms/book3s/vas-api.c
97
task_ref->tgid = find_get_pid(task_tgid_vnr(current));
arch/powerpc/platforms/cell/spufs/context.c
49
ctx->owner = get_task_mm(current);
arch/powerpc/platforms/cell/spufs/coredump.c
71
int n = iterate_fd(current->files, *fd, match_context, NULL);
arch/powerpc/platforms/cell/spufs/fault.c
128
ret = copro_handle_mm_fault(current->mm, ea, dsisr, &flt);
arch/powerpc/platforms/cell/spufs/file.c
1430
if (ctx->owner != current->mm)
arch/powerpc/platforms/cell/spufs/file.c
324
if (fatal_signal_pending(current))
arch/powerpc/platforms/cell/spufs/file.c
346
mmap_read_unlock(current->mm);
arch/powerpc/platforms/cell/spufs/file.c
350
mmap_read_lock(current->mm);
arch/powerpc/platforms/cell/spufs/run.c
259
if (signal_pending(current))
arch/powerpc/platforms/cell/spufs/run.c
411
if (signal_pending(current))
arch/powerpc/platforms/cell/spufs/sched.c
1075
idr_get_cursor(&task_active_pid_ns(current)->idr) - 1);
arch/powerpc/platforms/cell/spufs/sched.c
109
ctx->tid = current->pid;
arch/powerpc/platforms/cell/spufs/sched.c
117
if (rt_prio(current->prio))
arch/powerpc/platforms/cell/spufs/sched.c
118
ctx->prio = current->prio;
arch/powerpc/platforms/cell/spufs/sched.c
120
ctx->prio = current->static_prio;
arch/powerpc/platforms/cell/spufs/sched.c
121
ctx->policy = current->policy;
arch/powerpc/platforms/cell/spufs/sched.c
131
cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr);
arch/powerpc/platforms/cell/spufs/sched.c
226
spu->pid = current->pid;
arch/powerpc/platforms/cell/spufs/sched.c
227
spu->tgid = current->tgid;
arch/powerpc/platforms/cell/spufs/sched.c
539
if (!signal_pending(current)) {
arch/powerpc/platforms/cell/spufs/sched.c
782
if (signal_pending(current))
arch/powerpc/platforms/cell/spufs/spufs.h
315
if (signal_pending(current)) { \
arch/powerpc/platforms/chrp/nvram.c
30
current->comm, addr, nvram_size);
arch/powerpc/platforms/chrp/nvram.c
51
current->comm, addr, nvram_size);
arch/powerpc/platforms/powernv/vas-window.c
1004
trace_vas_tx_win_open(current, vasid, cop, attr);
arch/powerpc/platforms/powernv/vas-window.c
1106
trace_vas_paste_crb(current, txwin);
arch/powerpc/platforms/powernv/vas-window.c
860
trace_vas_rx_win_open(current, vasid, cop, rxattr);
arch/powerpc/platforms/pseries/papr-hvpipe.c
503
src_info->tsk = current;
arch/powerpc/platforms/pseries/papr-platform-dump.c
349
current->comm, current->pid, dump_tag);
arch/powerpc/platforms/pseries/papr-rtas-common.c
188
} while (!fatal_signal_pending(current));
arch/powerpc/xmon/xmon.c
1852
printf(" current = 0x%px\n", current);
arch/powerpc/xmon/xmon.c
1857
if (current) {
arch/powerpc/xmon/xmon.c
1859
current->pid, current->comm);
arch/riscv/include/asm/elf.h
139
if (personality(current->personality) != PER_LINUX32) \
arch/riscv/include/asm/elf.h
141
(current->personality & (~PER_MASK))); \
arch/riscv/include/asm/elf.h
94
(elf_addr_t)(ulong)current->mm->context.vdso); \
arch/riscv/include/asm/entry-common.h
19
riscv_v_vstate_restore(&current->thread.vstate, regs);
arch/riscv/include/asm/processor.h
215
#define SET_TAGGED_ADDR_CTRL(arg) set_tagged_addr_ctrl(current, arg)
arch/riscv/include/asm/processor.h
216
#define GET_TAGGED_ADDR_CTRL() get_tagged_addr_ctrl(current)
arch/riscv/include/asm/simd.h
43
if (IS_ENABLED(CONFIG_RISCV_ISA_V_PREEMPTIVE) && current->thread.kernel_vstate.datap) {
arch/riscv/include/asm/simd.h
44
if (!riscv_preempt_v_started(current))
arch/riscv/include/asm/stackprotector.h
18
current->stack_canary = canary;
arch/riscv/include/asm/stackprotector.h
20
__stack_chk_guard = current->stack_canary;
arch/riscv/include/asm/stacktrace.h
21
return !(((unsigned long)(current->stack) ^ current_stack_pointer) & ~(THREAD_SIZE - 1));
arch/riscv/include/asm/switch_to.h
80
if (task == current)
arch/riscv/include/asm/uaccess.h
33
(__force __typeof__(addr))__untagged_addr_remote(current->mm, __addr); \
arch/riscv/include/asm/vector.h
58
return READ_ONCE(current->thread.riscv_v_flags);
arch/riscv/kernel/compat_signal.c
127
current->restart_block.fn = do_no_restart_syscall;
arch/riscv/kernel/compat_signal.c
148
task = current;
arch/riscv/kernel/compat_signal.c
221
current->mm->context.vdso, rt_sigreturn);
arch/riscv/kernel/compat_signal.c
238
current->comm, task_pid_nr(current), ksig->sig,
arch/riscv/kernel/compat_signal.c
56
err = __copy_from_user(&current->thread.fstate, state, sizeof(*state));
arch/riscv/kernel/compat_signal.c
60
fstate_restore(current, regs);
arch/riscv/kernel/compat_signal.c
83
fstate_save(current, regs);
arch/riscv/kernel/compat_signal.c
84
err = __copy_to_user(state, &current->thread.fstate, sizeof(*state));
arch/riscv/kernel/cpufeature.c
1174
current->thread.envcfg |= ENVCFG_CBZE;
arch/riscv/kernel/cpufeature.c
1179
current->thread.envcfg |= ENVCFG_CBCFE;
arch/riscv/kernel/ftrace.c
236
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/riscv/kernel/ftrace.c
258
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/riscv/kernel/kernel_mode_fpu.c
17
fstate_save(current, task_pt_regs(current));
arch/riscv/kernel/kernel_mode_fpu.c
25
fstate_restore(current, task_pt_regs(current));
arch/riscv/kernel/kernel_mode_vector.c
116
if (riscv_v_ctx_get_depth() != 0 || !riscv_preempt_v_started(current))
arch/riscv/kernel/kernel_mode_vector.c
119
riscv_preempt_v_clear_dirty(current);
arch/riscv/kernel/kernel_mode_vector.c
128
kvstate = &current->thread.kernel_vstate;
arch/riscv/kernel/kernel_mode_vector.c
132
if (riscv_preempt_v_started(current)) {
arch/riscv/kernel/kernel_mode_vector.c
136
if (riscv_preempt_v_dirty(current)) {
arch/riscv/kernel/kernel_mode_vector.c
138
riscv_preempt_v_clear_dirty(current);
arch/riscv/kernel/kernel_mode_vector.c
140
riscv_preempt_v_set_restore(current);
arch/riscv/kernel/kernel_mode_vector.c
146
if (__riscv_v_vstate_check(task_pt_regs(current)->status, DIRTY)) {
arch/riscv/kernel/kernel_mode_vector.c
147
uvstate = &current->thread.vstate;
arch/riscv/kernel/kernel_mode_vector.c
150
riscv_preempt_v_clear_dirty(current);
arch/riscv/kernel/kernel_mode_vector.c
159
if (!riscv_preempt_v_started(current))
arch/riscv/kernel/kernel_mode_vector.c
171
struct __riscv_v_ext_state *vstate = &current->thread.kernel_vstate;
arch/riscv/kernel/kernel_mode_vector.c
176
if (!riscv_preempt_v_started(current))
arch/riscv/kernel/kernel_mode_vector.c
182
if (riscv_preempt_v_restore(current)) {
arch/riscv/kernel/kernel_mode_vector.c
218
riscv_v_vstate_save(&current->thread.vstate, task_pt_regs(current));
arch/riscv/kernel/kernel_mode_vector.c
222
riscv_v_vstate_set_restore(current, task_pt_regs(current));
arch/riscv/kernel/kernel_mode_vector.c
23
WRITE_ONCE(current->thread.riscv_v_flags, flags);
arch/riscv/kernel/kernel_mode_vector.c
86
return &current->thread.riscv_v_flags;
arch/riscv/kernel/perf_regs.c
41
regs_user->regs = task_pt_regs(current);
arch/riscv/kernel/perf_regs.c
42
regs_user->abi = perf_reg_abi(current);
arch/riscv/kernel/probes/uprobes.c
110
struct uprobe_task *utask = current->utask;
arch/riscv/kernel/probes/uprobes.c
112
current->thread.bad_cause = utask->autask.saved_cause;
arch/riscv/kernel/probes/uprobes.c
61
struct uprobe_task *utask = current->utask;
arch/riscv/kernel/probes/uprobes.c
63
utask->autask.saved_cause = current->thread.bad_cause;
arch/riscv/kernel/probes/uprobes.c
64
current->thread.bad_cause = UPROBE_TRAP_NR;
arch/riscv/kernel/probes/uprobes.c
73
struct uprobe_task *utask = current->utask;
arch/riscv/kernel/probes/uprobes.c
75
WARN_ON_ONCE(current->thread.bad_cause != UPROBE_TRAP_NR);
arch/riscv/kernel/probes/uprobes.c
76
current->thread.bad_cause = utask->autask.saved_cause;
arch/riscv/kernel/process.c
111
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/riscv/kernel/process.c
154
fstate_restore(current, regs);
arch/riscv/kernel/process.c
163
set_shstk_status(current, false);
arch/riscv/kernel/process.c
164
set_shstk_base(current, 0, 0);
arch/riscv/kernel/process.c
165
set_active_shstk(current, 0);
arch/riscv/kernel/process.c
170
set_indir_lp_status(current, false);
arch/riscv/kernel/process.c
190
fstate_off(current, task_pt_regs(current));
arch/riscv/kernel/process.c
191
memset(&current->thread.fstate, 0, sizeof(current->thread.fstate));
arch/riscv/kernel/process.c
195
riscv_v_vstate_ctrl_init(current);
arch/riscv/kernel/process.c
196
riscv_v_vstate_off(task_pt_regs(current));
arch/riscv/kernel/process.c
197
kfree(current->thread.vstate.datap);
arch/riscv/kernel/process.c
198
memset(&current->thread.vstate, 0, sizeof(struct __riscv_v_ext_state));
arch/riscv/kernel/process.c
199
clear_tsk_thread_flag(current, TIF_RISCV_V_DEFER_RESTORE);
arch/riscv/kernel/process.c
203
envcfg_update_bits(current, ENVCFG_PMM, ENVCFG_PMM_PMLEN_0);
arch/riscv/kernel/process.c
97
regs->t5, regs->t6, get_active_shstk(current));
arch/riscv/kernel/ptrace.c
112
if (target == current) {
arch/riscv/kernel/ptrace.c
114
riscv_v_vstate_save(&current->thread.vstate, task_pt_regs(current));
arch/riscv/kernel/ptrace.c
66
if (target == current)
arch/riscv/kernel/ptrace.c
67
fstate_save(current, task_pt_regs(current));
arch/riscv/kernel/return_address.c
39
arch_stack_walk(save_return_addr, &data, current, NULL);
arch/riscv/kernel/signal.c
102
err |= __copy_to_user(datap, current->thread.vstate.datap, riscv_v_vsize);
arch/riscv/kernel/signal.c
126
riscv_v_vstate_set_restore(current, regs);
arch/riscv/kernel/signal.c
129
err = __copy_from_user(&current->thread.vstate, &state->v_state,
arch/riscv/kernel/signal.c
142
return copy_from_user(current->thread.vstate.datap, datap, riscv_v_vsize);
arch/riscv/kernel/signal.c
151
if (!is_shstk_enabled(current))
arch/riscv/kernel/signal.c
166
err |= save_user_shstk(current, &ss_ptr);
arch/riscv/kernel/signal.c
198
return restore_user_shstk(current, ss_ptr);
arch/riscv/kernel/signal.c
268
if (!is_shstk_enabled(current) || size != riscv_zicfiss_sc_size)
arch/riscv/kernel/signal.c
290
if (cal_all || riscv_v_vstate_query(task_pt_regs(current)))
arch/riscv/kernel/signal.c
294
if (is_shstk_enabled(current))
arch/riscv/kernel/signal.c
319
current->restart_block.fn = do_no_restart_syscall;
arch/riscv/kernel/signal.c
342
task = current;
arch/riscv/kernel/signal.c
439
current->mm->context.vdso, rt_sigreturn);
arch/riscv/kernel/signal.c
442
if (is_shstk_enabled(current))
arch/riscv/kernel/signal.c
476
current->comm, task_pid_nr(current), ksig->sig,
arch/riscv/kernel/signal.c
50
err = __copy_from_user(&current->thread.fstate, state, sizeof(*state));
arch/riscv/kernel/signal.c
54
fstate_restore(current, regs);
arch/riscv/kernel/signal.c
64
fstate_save(current, regs);
arch/riscv/kernel/signal.c
65
err = __copy_to_user(state, &current->thread.fstate, sizeof(*state));
arch/riscv/kernel/signal.c
93
riscv_v_vstate_save(&current->thread.vstate, regs);
arch/riscv/kernel/signal.c
97
err = __copy_to_user(&state->v_state, &current->thread.vstate,
arch/riscv/kernel/smpboot.c
235
current->active_mm = mm;
arch/riscv/kernel/stacktrace.c
116
} else if (task == NULL || task == current) {
arch/riscv/kernel/stacktrace.c
28
if ((task) == current) \
arch/riscv/kernel/stacktrace.c
59
} else if (task == NULL || task == current) {
arch/riscv/kernel/stacktrace.c
90
pc = ftrace_graph_ret_addr(current, &graph_idx, pc,
arch/riscv/kernel/sys_riscv.c
76
flush_icache_mm(current->mm, flags & SYS_RISCV_FLUSH_ICACHE_LOCAL);
arch/riscv/kernel/traps.c
117
struct task_struct *tsk = current;
arch/riscv/kernel/traps.c
135
current->thread.bad_cause = regs->cause;
arch/riscv/kernel/traps.c
291
current->thread.bad_cause = regs->cause;
arch/riscv/kernel/traps.c
483
unsigned long tsk_stk = (unsigned long)current->stack;
arch/riscv/kernel/traps.c
49
if (regs != task_pt_regs(current))
arch/riscv/kernel/traps.c
99
if (kexec_should_crash(current))
arch/riscv/kernel/traps_misaligned.c
235
if (user_mode(regs) && (current->thread.align_ctl & PR_UNALIGN_SIGBUS))
arch/riscv/kernel/traps_misaligned.c
346
if (user_mode(regs) && (current->thread.align_ctl & PR_UNALIGN_SIGBUS))
arch/riscv/kernel/usercfi.c
234
struct mm_struct *mm = current->mm;
arch/riscv/kernel/usercfi.c
364
if (!tsk->mm || tsk->mm != current->mm)
arch/riscv/kernel/vdso.c
161
struct mm_struct *mm = current->mm;
arch/riscv/kernel/vdso.c
177
struct mm_struct *mm = current->mm;
arch/riscv/kernel/vdso.c
39
current->mm->context.vdso = (void *)new_vma->vm_start;
arch/riscv/kernel/vector.c
181
return riscv_v_ctrl_get_cur(current) == PR_RISCV_V_VSTATE_CTRL_ON;
arch/riscv/kernel/vector.c
212
WARN_ON(current->thread.vstate.datap);
arch/riscv/kernel/vector.c
219
if (riscv_v_thread_ctx_alloc(riscv_v_user_cachep, &current->thread.vstate)) {
arch/riscv/kernel/vector.c
225
riscv_v_vstate_set_restore(current, regs);
arch/riscv/kernel/vector.c
260
return current->thread.vstate_ctrl & PR_RISCV_V_VSTATE_CTRL_MASK;
arch/riscv/kernel/vector.c
278
if (riscv_v_ctrl_get_cur(current) != PR_RISCV_V_VSTATE_CTRL_OFF)
arch/riscv/kernel/vector.c
285
cur = riscv_v_ctrl_get_cur(current);
arch/riscv/kernel/vector.c
297
riscv_v_ctrl_set(current, cur, next, inherit);
arch/riscv/kvm/mmu.c
197
mmap_read_lock(current->mm);
arch/riscv/kvm/mmu.c
214
vma = find_vma_intersection(current->mm, hva, reg_end);
arch/riscv/kvm/mmu.c
241
mmap_read_unlock(current->mm);
arch/riscv/kvm/mmu.c
481
mmap_read_lock(current->mm);
arch/riscv/kvm/mmu.c
483
vma = vma_lookup(current->mm, hva);
arch/riscv/kvm/mmu.c
486
mmap_read_unlock(current->mm);
arch/riscv/kvm/mmu.c
510
mmap_read_unlock(current->mm);
arch/riscv/kvm/mmu.c
523
vma_pageshift, current);
arch/riscv/kvm/vcpu_pmu.c
336
event = perf_event_create_kernel_counter(attr, -1, current, kvm_riscv_pmu_overflow, pmc);
arch/riscv/kvm/vcpu_sbi_sta.c
118
vcpu->arch.sta.last_steal = current->sched_info.run_delay;
arch/riscv/kvm/vcpu_sbi_sta.c
70
vcpu->arch.sta.last_steal = READ_ONCE(current->sched_info.run_delay);
arch/riscv/mm/cacheflush.c
192
mask = &current->mm->context.icache_stale_mask;
arch/riscv/mm/cacheflush.c
253
current->mm->context.force_icache_flush = true;
arch/riscv/mm/cacheflush.c
256
current->thread.force_icache_flush = true;
arch/riscv/mm/cacheflush.c
266
current->mm->context.force_icache_flush = false;
arch/riscv/mm/cacheflush.c
270
current->thread.force_icache_flush = false;
arch/riscv/mm/cacheflush.c
78
if (mm == current->active_mm && local) {
arch/riscv/mm/fault.c
291
tsk = current;
arch/riscv/mm/fault.c
35
struct mm_struct *mm = current->mm;
arch/riscv/mm/fault.c
41
current->comm, PAGE_SIZE / SZ_1K, VA_BITS,
arch/s390/crypto/prng.c
522
if (signal_pending(current)) {
arch/s390/crypto/prng.c
605
if (signal_pending(current)) {
arch/s390/hypfs/hypfs_diag.c
151
if (signal_pending(current))
arch/s390/include/asm/elf.h
219
(current->personality & (~PER_MASK))); \
arch/s390/include/asm/elf.h
237
(unsigned long)current->mm->context.vdso_base); \
arch/s390/include/asm/entry-common.h
65
return current->flags & PF_VCPU;
arch/s390/include/asm/fpu.h
160
struct thread_struct *thread = &current->thread;
arch/s390/include/asm/fpu.h
176
struct thread_struct *thread = &current->thread;
arch/s390/include/asm/fpu.h
189
struct thread_struct *thread = &current->thread;
arch/s390/include/asm/fpu.h
208
WRITE_ONCE(current->thread.kfpu_flags, mask);
arch/s390/include/asm/kprobes.h
48
(((unsigned long)task_stack_page(current)) + THREAD_SIZE - (ADDR))) \
arch/s390/include/asm/kprobes.h
50
: (((unsigned long)task_stack_page(current)) + THREAD_SIZE - (ADDR)))
arch/s390/include/asm/mmu_context.h
123
switch_mm_irqs_off(prev, next, current);
arch/s390/include/asm/mmu_context.h
98
struct task_struct *tsk = current;
arch/s390/include/asm/pgtable.h
1464
return end <= current->mm->context.asce_limit;
arch/s390/include/asm/stackprotector.h
12
current->stack_canary = get_random_canary();
arch/s390/include/asm/stackprotector.h
13
get_lowcore()->stack_canary = current->stack_canary;
arch/s390/include/asm/stacktrace.h
92
if (task == current)
arch/s390/include/asm/unwind.h
78
task = task ?: current;
arch/s390/kernel/dumpstack.c
105
if (task != current)
arch/s390/kernel/dumpstack.c
140
debug_show_held_locks(task ? : current);
arch/s390/kernel/ftrace.c
274
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/s390/kernel/guarded_storage.c
100
for_each_thread(current, sibling) {
arch/s390/kernel/guarded_storage.c
26
if (!current->thread.gs_cb) {
arch/s390/kernel/guarded_storage.c
34
current->thread.gs_cb = gs_cb;
arch/s390/kernel/guarded_storage.c
42
if (current->thread.gs_cb) {
arch/s390/kernel/guarded_storage.c
44
kfree(current->thread.gs_cb);
arch/s390/kernel/guarded_storage.c
45
current->thread.gs_cb = NULL;
arch/s390/kernel/guarded_storage.c
56
gs_cb = current->thread.gs_bc_cb;
arch/s390/kernel/guarded_storage.c
61
current->thread.gs_bc_cb = gs_cb;
arch/s390/kernel/guarded_storage.c
72
gs_cb = current->thread.gs_bc_cb;
arch/s390/kernel/guarded_storage.c
73
current->thread.gs_bc_cb = NULL;
arch/s390/kernel/guarded_storage.c
84
gs_cb = current->thread.gs_bc_cb;
arch/s390/kernel/guarded_storage.c
86
kfree(current->thread.gs_cb);
arch/s390/kernel/guarded_storage.c
87
current->thread.gs_bc_cb = NULL;
arch/s390/kernel/guarded_storage.c
90
current->thread.gs_cb = gs_cb;
arch/s390/kernel/irq.c
158
current->thread.last_break = regs->last_break;
arch/s390/kernel/irq.c
197
current->thread.last_break = regs->last_break;
arch/s390/kernel/nmi.c
264
current->comm, current->pid);
arch/s390/kernel/nmi.c
265
if (is_global_init(current))
arch/s390/kernel/nmi.c
267
do_send_sig_info(SIGKILL, SEND_SIG_PRIV, current, PIDTYPE_PID);
arch/s390/kernel/perf_event.c
208
unwind_for_each_frame(&state, current, regs, 0) {
arch/s390/kernel/perf_pai.c
618
data.tid_entry.pid = task_tgid_nr(current);
arch/s390/kernel/perf_pai.c
619
data.tid_entry.tid = task_pid_nr(current);
arch/s390/kernel/perf_regs.c
22
fp = *(freg_t *)(current->thread.ufpu.vxrs + idx);
arch/s390/kernel/perf_regs.c
59
regs_user->regs = task_pt_regs(current);
arch/s390/kernel/perf_regs.c
62
regs_user->abi = perf_reg_abi(current);
arch/s390/kernel/process.c
182
current->thread.ufpu.fpc = 0;
arch/s390/kernel/process.c
232
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/s390/kernel/process.c
72
if (get_lowcore()->current_pid != current->pid) {
arch/s390/kernel/process.c
73
get_lowcore()->current_pid = current->pid;
arch/s390/kernel/processor.c
112
current->active_mm = &init_mm;
arch/s390/kernel/processor.c
113
BUG_ON(current->mm);
arch/s390/kernel/processor.c
114
enter_lazy_tlb(&init_mm, current);
arch/s390/kernel/ptrace.c
514
if (target == current)
arch/s390/kernel/ptrace.c
529
if (target == current)
arch/s390/kernel/ptrace.c
552
if (rc == 0 && target == current)
arch/s390/kernel/ptrace.c
564
if (target == current)
arch/s390/kernel/ptrace.c
581
if (target == current)
arch/s390/kernel/ptrace.c
649
if (target == current)
arch/s390/kernel/ptrace.c
666
if (target == current)
arch/s390/kernel/ptrace.c
686
if (target == current)
arch/s390/kernel/ptrace.c
701
if (target == current)
arch/s390/kernel/ptrace.c
736
if (target == current)
arch/s390/kernel/ptrace.c
758
else if (target == current)
arch/s390/kernel/ptrace.c
772
if (target == current) {
arch/s390/kernel/ptrace.c
870
if (target == current)
arch/s390/kernel/ptrace.c
896
if (target == current)
arch/s390/kernel/runtime_instr.c
33
struct task_struct *task = current;
arch/s390/kernel/runtime_instr.c
85
if (!current->thread.ri_cb) {
arch/s390/kernel/runtime_instr.c
90
cb = current->thread.ri_cb;
arch/s390/kernel/runtime_instr.c
98
current->thread.ri_cb = cb;
arch/s390/kernel/signal.c
110
save_access_regs(current->thread.acrs);
arch/s390/kernel/signal.c
117
restore_access_regs(current->thread.acrs);
arch/s390/kernel/signal.c
131
memcpy(&user_sregs.regs.acrs, current->thread.acrs,
arch/s390/kernel/signal.c
133
fpregs_store(&user_sregs.fpregs, &current->thread.ufpu);
arch/s390/kernel/signal.c
144
current->restart_block.fn = do_no_restart_syscall;
arch/s390/kernel/signal.c
149
if (!is_ri_task(current) && (user_sregs.regs.psw.mask & PSW_MASK_RI))
arch/s390/kernel/signal.c
164
memcpy(&current->thread.acrs, &user_sregs.regs.acrs,
arch/s390/kernel/signal.c
165
sizeof(current->thread.acrs));
arch/s390/kernel/signal.c
167
fpregs_load(&user_sregs.fpregs, &current->thread.ufpu);
arch/s390/kernel/signal.c
183
vxrs[i] = current->thread.ufpu.vxrs[i].low;
arch/s390/kernel/signal.c
187
current->thread.ufpu.vxrs + __NUM_VXRS_LOW,
arch/s390/kernel/signal.c
204
__copy_from_user(current->thread.ufpu.vxrs + __NUM_VXRS_LOW,
arch/s390/kernel/signal.c
209
current->thread.ufpu.vxrs[i].low = vxrs[i];
arch/s390/kernel/signal.c
216
struct pt_regs *regs = task_pt_regs(current);
arch/s390/kernel/signal.c
238
struct pt_regs *regs = task_pt_regs(current);
arch/s390/kernel/signal.c
278
sp = current->sas_ss_sp + current->sas_ss_size;
arch/s390/kernel/signal.c
329
restorer = VDSO_SYMBOL(current, sigreturn);
arch/s390/kernel/signal.c
350
regs->gprs[6] = current->thread.last_break;
arch/s390/kernel/signal.c
381
restorer = VDSO_SYMBOL(current, rt_sigreturn);
arch/s390/kernel/signal.c
411
regs->gprs[5] = current->thread.last_break;
arch/s390/kernel/signal.c
449
current->thread.system_call =
arch/s390/kernel/signal.c
454
if (current->thread.system_call) {
arch/s390/kernel/signal.c
455
regs->int_code = current->thread.system_call;
arch/s390/kernel/signal.c
486
if (current->thread.system_call) {
arch/s390/kernel/signal.c
487
regs->int_code = current->thread.system_call;
arch/s390/kernel/signal.c
492
current->restart_block.arch_data = regs->psw.addr;
arch/s390/kernel/signal.c
493
regs->psw.addr = VDSO_SYMBOL(current, restart_syscall);
arch/s390/kernel/stacktrace.c
108
if (!current->mm)
arch/s390/kernel/stacktrace.c
90
if (ip >= current->mm->context.asce_limit)
arch/s390/kernel/stacktrace.c
97
return in_range(ip, current->mm->context.vdso_base, vdso_text_size());
arch/s390/kernel/syscall.c
106
current->thread.last_break = regs->last_break;
arch/s390/kernel/syscall.c
120
if (nr == __NR_restart_syscall && !(current->restart_block.arch_data & 1)) {
arch/s390/kernel/syscall.c
121
regs->psw.addr = current->restart_block.arch_data;
arch/s390/kernel/syscall.c
122
current->restart_block.arch_data = 1;
arch/s390/kernel/syscall.c
76
unsigned int ret = current->personality;
arch/s390/kernel/syscall.c
78
if (personality(current->personality) == PER_LINUX32 &&
arch/s390/kernel/traps.c
151
if (current->ptrace)
arch/s390/kernel/traps.c
183
vic = (current->thread.ufpu.fpc & 0xf00) >> 8;
arch/s390/kernel/traps.c
209
if (current->thread.ufpu.fpc & FPC_DXC_MASK)
arch/s390/kernel/traps.c
210
do_fp_trap(regs, current->thread.ufpu.fpc);
arch/s390/kernel/traps.c
348
current->thread.gmap_teid.val = regs->int_parm_long;
arch/s390/kernel/traps.c
349
current->thread.gmap_int_code = regs->int_code & 0xffff;
arch/s390/kernel/traps.c
359
current->thread.last_break = regs->last_break;
arch/s390/kernel/traps.c
363
current->thread.trap_tdb = lc->pgm_tdb;
arch/s390/kernel/traps.c
367
struct per_event *ev = &current->thread.per_event;
arch/s390/kernel/traps.c
40
address = current->thread.trap_tdb.data[3];
arch/s390/kernel/traps.c
76
if (!current->ptrace)
arch/s390/kernel/traps.c
78
force_sig_fault(SIGTRAP, TRAP_HWBKPT, (void __force __user *)current->thread.per_event.address);
arch/s390/kernel/unwind_bc.c
148
} else if (task == current) {
arch/s390/kernel/uprobes.c
100
current->thread.per_user.control, regs)) {
arch/s390/kernel/uprobes.c
102
current->thread.per_event.address = utask->vaddr;
arch/s390/kernel/uprobes.c
138
regs->psw.addr = current->utask->vaddr;
arch/s390/kernel/uprobes.c
139
current->thread.per_event.address = current->utask->vaddr;
arch/s390/kernel/uprobes.c
242
if (!(current->thread.per_user.control & PER_EVENT_STORE))
arch/s390/kernel/uprobes.c
244
if ((void *)current->thread.per_user.start > (addr + len))
arch/s390/kernel/uprobes.c
246
if ((void *)current->thread.per_user.end < addr)
arch/s390/kernel/uprobes.c
248
current->thread.per_event.address = regs->psw.addr;
arch/s390/kernel/uprobes.c
249
current->thread.per_event.cause = PER_EVENT_STORE >> 16;
arch/s390/kernel/uprobes.c
37
regs->psw.addr = current->utask->xol_vaddr;
arch/s390/kernel/uprobes.c
38
set_tsk_thread_flag(current, TIF_UPROBE_SINGLESTEP);
arch/s390/kernel/uprobes.c
39
update_cr_regs(current);
arch/s390/kernel/uprobes.c
69
regs->psw.addr >= current->thread.per_user.start &&
arch/s390/kernel/uprobes.c
70
regs->psw.addr <= current->thread.per_user.end)
arch/s390/kernel/uprobes.c
79
struct uprobe_task *utask = current->utask;
arch/s390/kernel/uprobes.c
81
clear_tsk_thread_flag(current, TIF_UPROBE_SINGLESTEP);
arch/s390/kernel/uprobes.c
82
update_cr_regs(current);
arch/s390/kernel/uprobes.c
99
if (check_per_event(current->thread.per_event.cause,
arch/s390/kernel/vdso.c
124
if (current->flags & PF_RANDOMIZE)
arch/s390/kernel/vdso.c
125
addr = vdso_addr(current->mm->start_stack + PAGE_SIZE, size);
arch/s390/kernel/vdso.c
30
current->mm->context.vdso_base = vma->vm_start;
arch/s390/kernel/vdso.c
49
struct mm_struct *mm = current->mm;
arch/s390/kernel/vdso.c
76
current->mm->context.vdso_base = vdso_text_start;
arch/s390/kernel/vtime.c
181
lc->user_timer = current->thread.user_timer;
arch/s390/kernel/vtime.c
182
lc->guest_timer = current->thread.guest_timer;
arch/s390/kernel/vtime.c
183
lc->system_timer = current->thread.system_timer;
arch/s390/kernel/vtime.c
184
lc->hardirq_timer = current->thread.hardirq_timer;
arch/s390/kernel/vtime.c
185
lc->softirq_timer = current->thread.softirq_timer;
arch/s390/kernel/wti.c
84
st->dbg.pid = current->pid;
arch/s390/kvm/gmap.c
482
if (fatal_signal_pending(current))
arch/s390/kvm/gmap.c
963
if (interruptible && fatal_signal_pending(current))
arch/s390/kvm/intercept.c
219
if (current->thread.per_flags & PER_FLAG_NO_TE)
arch/s390/kvm/intercept.c
99
current->pid, vcpu->kvm);
arch/s390/kvm/interrupt.c
2834
pr_info("guest machine check pid %d: %lx", current->pid, mci.val);
arch/s390/kvm/interrupt.c
536
save_gs_cb(current->thread.gs_cb);
arch/s390/kvm/kvm-s390.c
3199
snprintf(debug_name, sizeof(debug_name), "kvm-%u", current->pid);
arch/s390/kvm/kvm-s390.c
3295
KVM_EVENT(3, "vm 0x%p created by pid %u", kvm, current->pid);
arch/s390/kvm/kvm-s390.c
4439
hva = gfn_to_hva(vcpu->kvm, current->thread.gmap_teid.addr);
arch/s390/kvm/kvm-s390.c
4443
return kvm_setup_async_pf(vcpu, current->thread.gmap_teid.addr * PAGE_SIZE, hva, &arch);
arch/s390/kvm/kvm-s390.c
4478
current->thread.gmap_int_code = 0;
arch/s390/kvm/kvm-s390.c
4524
KVM_BUG(current->thread.gmap_teid.as != PSW_BITS_AS_PRIMARY, vcpu->kvm,
arch/s390/kvm/kvm-s390.c
4526
current->thread.gmap_int_code, current->thread.gmap_teid.val);
arch/s390/kvm/kvm-s390.c
4556
gaddr = current->thread.gmap_teid.addr * PAGE_SIZE;
arch/s390/kvm/kvm-s390.c
4560
switch (current->thread.gmap_int_code & PGM_INT_CODE_MASK) {
arch/s390/kvm/kvm-s390.c
4583
current->thread.gmap_int_code, current->comm,
arch/s390/kvm/kvm-s390.c
4584
current->pid);
arch/s390/kvm/kvm-s390.c
4585
send_sig(SIGSEGV, current, 0);
arch/s390/kvm/kvm-s390.c
4597
send_sig(SIGSEGV, current, 0);
arch/s390/kvm/kvm-s390.c
4613
current->thread.gmap_int_code, current->thread.gmap_teid.val);
arch/s390/kvm/kvm-s390.c
4614
send_sig(SIGSEGV, current, 0);
arch/s390/kvm/kvm-s390.c
4816
if (current->thread.gs_cb) {
arch/s390/kvm/kvm-s390.c
4817
vcpu->arch.host_gscb = current->thread.gs_cb;
arch/s390/kvm/kvm-s390.c
4821
current->thread.gs_cb = (struct gs_cb *)
arch/s390/kvm/kvm-s390.c
4823
restore_gs_cb(current->thread.gs_cb);
arch/s390/kvm/kvm-s390.c
4883
save_gs_cb(current->thread.gs_cb);
arch/s390/kvm/kvm-s390.c
4884
current->thread.gs_cb = vcpu->arch.host_gscb;
arch/s390/kvm/kvm-s390.c
4967
if (signal_pending(current) && !rc) {
arch/s390/kvm/kvm-s390.h
586
if (current->thread.gmap_int_code == PGM_PROTECTION)
arch/s390/kvm/kvm-s390.h
588
return test_facility(75) && (current->thread.gmap_teid.fsi == TEID_FSI_STORE);
arch/s390/kvm/pci.c
199
if (current->mm)
arch/s390/kvm/pci.c
200
atomic64_sub(nr_pages, &current->mm->pinned_vm);
arch/s390/kvm/pci.c
217
atomic64_add(nr_pages, &current->mm->pinned_vm);
arch/s390/kvm/priv.c
62
current->thread.gs_cb = (struct gs_cb *)&vcpu->run->s.regs.gscb;
arch/s390/kvm/priv.c
63
restore_gs_cb(current->thread.gs_cb);
arch/s390/kvm/pv.c
814
if (fatal_signal_pending(current))
arch/s390/kvm/vsie.c
1153
current->thread.gmap_int_code = 0;
arch/s390/kvm/vsie.c
1191
else if (current->thread.gmap_int_code)
arch/s390/kvm/vsie.c
933
if ((current->thread.gmap_int_code & PGM_INT_CODE_MASK) == PGM_PROTECTION)
arch/s390/kvm/vsie.c
936
current->thread.gmap_teid.addr * PAGE_SIZE, 1);
arch/s390/kvm/vsie.c
938
rc = gaccess_shadow_fault(vcpu, sg, current->thread.gmap_teid.addr * PAGE_SIZE, NULL, wr);
arch/s390/kvm/vsie.c
941
current->thread.gmap_teid.addr * PAGE_SIZE, wr);
arch/s390/kvm/vsie.c
943
vsie_page->fault_addr = current->thread.gmap_teid.addr * PAGE_SIZE;
arch/s390/mm/fault.c
196
if ((task_pid_nr(current) > 1) && !show_unhandled_signals)
arch/s390/mm/fault.c
198
if (!unhandled_signal(current, signr))
arch/s390/mm/fault.c
244
struct mm_struct *mm = current->mm;
arch/s390/mm/fault.c
282
mm = current->mm;
arch/s390/mm/fault.c
431
send_sig(SIGSEGV, current, 0);
arch/s390/mm/fault.c
458
mm = current->mm;
arch/s390/mm/fault.c
474
send_sig(SIGSEGV, current, 0);
arch/s390/mm/gmap_helpers.c
281
struct mm_struct *mm = current->mm;
arch/s390/mm/mmap.c
119
struct mm_struct *mm = current->mm;
arch/s390/mm/mmap.c
175
if (current->flags & PF_RANDOMIZE)
arch/s390/mm/mmap.c
24
if (!(current->flags & PF_RANDOMIZE))
arch/s390/mm/mmap.c
31
if (current->personality & ADDR_COMPAT_LAYOUT)
arch/s390/mm/mmap.c
71
if (!(current->flags & PF_RANDOMIZE))
arch/s390/mm/mmap.c
82
struct mm_struct *mm = current->mm;
arch/s390/mm/pfault.c
172
if (WARN_ON_ONCE(tsk != current))
arch/s390/mm/pgalloc.c
46
if (current->active_mm == mm) {
arch/s390/pci/pci_mmio.c
170
mmap_read_lock(current->mm);
arch/s390/pci/pci_mmio.c
172
vma = vma_lookup(current->mm, mmio_addr);
arch/s390/pci/pci_mmio.c
185
fixup_user_fault(current->mm, mmio_addr, FAULT_FLAG_WRITE, NULL);
arch/s390/pci/pci_mmio.c
201
mmap_read_unlock(current->mm);
arch/s390/pci/pci_mmio.c
322
mmap_read_lock(current->mm);
arch/s390/pci/pci_mmio.c
324
vma = vma_lookup(current->mm, mmio_addr);
arch/s390/pci/pci_mmio.c
337
fixup_user_fault(current->mm, mmio_addr, 0, NULL);
arch/s390/pci/pci_mmio.c
355
mmap_read_unlock(current->mm);
arch/sh/include/asm/elf.h
164
set_personality(PER_LINUX_32BIT | (current->personality & (~PER_MASK)))
arch/sh/include/asm/elf.h
176
#define VDSO_BASE ((unsigned long)current->mm->context.vdso)
arch/sh/include/asm/stackprotector.h
17
current->stack_canary = canary;
arch/sh/include/asm/stackprotector.h
18
__stack_chk_guard = current->stack_canary;
arch/sh/kernel/cpu/fpu.c
14
if ((boot_cpu_data.flags & CPU_HAS_FPU) && tsk == current)
arch/sh/kernel/cpu/fpu.c
46
struct task_struct *tsk = current;
arch/sh/kernel/cpu/fpu.c
56
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh2a/fpu.c
454
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh2a/fpu.c
466
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh2a/fpu.c
509
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh2a/fpu.c
558
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
228
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
241
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
279
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
323
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
363
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
396
struct task_struct *tsk = current;
arch/sh/kernel/cpu/sh4/fpu.c
403
struct task_struct *tsk = current;
arch/sh/kernel/dumpstack.c
136
tsk = current;
arch/sh/kernel/dumpstack.c
146
tsk = current;
arch/sh/kernel/dumpstack.c
147
if (tsk == current)
arch/sh/kernel/dwarf.c
610
ret_stack = ftrace_graph_get_ret_stack(current, 0);
arch/sh/kernel/dwarf.c
619
WARN_ON(ftrace_graph_get_ret_stack(current, 1));
arch/sh/kernel/ftrace.c
325
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/sh/kernel/process_32.c
70
free_thread_xstate(current);
arch/sh/kernel/process_32.c
76
struct task_struct *tsk = current;
arch/sh/kernel/process_32.c
99
struct task_struct *tsk = current;
arch/sh/kernel/signal_32.c
139
struct task_struct *tsk = current;
arch/sh/kernel/signal_32.c
163
current->restart_block.fn = do_no_restart_syscall;
arch/sh/kernel/signal_32.c
193
current->restart_block.fn = do_no_restart_syscall;
arch/sh/kernel/signal_32.c
258
sp = current->sas_ss_sp + current->sas_ss_size;
arch/sh/kernel/signal_32.c
291
} else if (likely(current->mm->context.vdso)) {
arch/sh/kernel/signal_32.c
317
if (current->personality & FDPIC_FUNCPTRS) {
arch/sh/kernel/signal_32.c
330
current->comm, task_pid_nr(current), frame, regs->pc, regs->pr);
arch/sh/kernel/signal_32.c
361
} else if (likely(current->mm->context.vdso)) {
arch/sh/kernel/signal_32.c
387
if (current->personality & FDPIC_FUNCPTRS) {
arch/sh/kernel/signal_32.c
400
current->comm, task_pid_nr(current), frame, regs->pc, regs->pr);
arch/sh/kernel/signal_32.c
79
struct task_struct *tsk = current;
arch/sh/kernel/signal_32.c
92
struct task_struct *tsk = current;
arch/sh/kernel/smp.c
183
current->active_mm = mm;
arch/sh/kernel/smp.c
185
enter_lazy_tlb(mm, current);
arch/sh/kernel/smp.c
366
if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/sh/kernel/smp.c
398
if ((atomic_read(&mm->mm_users) != 1) || (current->mm != mm)) {
arch/sh/kernel/smp.c
442
(current->mm != vma->vm_mm)) {
arch/sh/kernel/smp.c
62
init_new_context(current, &init_mm);
arch/sh/kernel/stacktrace.c
45
unwind_stack(current, NULL, sp, &save_stack_ops, trace);
arch/sh/kernel/stacktrace.c
77
unwind_stack(current, NULL, sp, &save_stack_ops_nosched, trace);
arch/sh/kernel/swsusp.c
29
init_fpu(current);
arch/sh/kernel/sys_sh.c
72
mmap_read_lock(current->mm);
arch/sh/kernel/sys_sh.c
73
vma = find_vma (current->mm, addr);
arch/sh/kernel/sys_sh.c
75
mmap_read_unlock(current->mm);
arch/sh/kernel/sys_sh.c
94
mmap_read_unlock(current->mm);
arch/sh/kernel/traps.c
39
printk("Process: %s (pid: %d, stack limit = %p)\n", current->comm,
arch/sh/kernel/traps.c
40
task_pid_nr(current), task_stack_page(current) + 1);
arch/sh/kernel/traps.c
44
THREAD_SIZE + (unsigned long)task_stack_page(current));
arch/sh/kernel/traps.c
53
if (kexec_should_crash(current))
arch/sh/kernel/traps_32.c
344
unaligned_fixups_notify(current, instruction, regs);
arch/sh/kernel/traps_32.c
515
unaligned_fixups_notify(current, instruction, regs);
arch/sh/kernel/traps_32.c
543
"access (PC %lx PR %lx)\n", current->comm, regs->pc,
arch/sh/kernel/traps_32.c
561
unaligned_fixups_notify(current, instruction, regs);
arch/sh/kernel/traps_32.c
642
current->thread.dsp_status.status |= SR_DSP;
arch/sh/kernel/vsyscall/vsyscall.c
108
current->mm->context.vdso = (void *)addr;
arch/sh/kernel/vsyscall/vsyscall.c
85
struct mm_struct *mm = current->mm;
arch/sh/math-emu/math.c
496
struct task_struct *tsk = current;
arch/sh/mm/alignment.c
72
if (current->thread.flags & SH_THREAD_UAC_SIGBUS) {
arch/sh/mm/alignment.c
77
if (current->thread.flags & SH_THREAD_UAC_NOPRINT)
arch/sh/mm/cache-sh4.c
241
if ((vma->vm_mm == current->active_mm))
arch/sh/mm/fault.c
273
struct mm_struct *mm = current->mm;
arch/sh/mm/fault.c
300
struct task_struct *tsk = current;
arch/sh/mm/fault.c
328
mmap_read_unlock(current->mm);
arch/sh/mm/fault.c
400
tsk = current;
arch/sh/mm/mmap.c
106
struct mm_struct *mm = current->mm;
arch/sh/mm/mmap.c
58
struct mm_struct *mm = current->mm;
arch/sh/mm/tlb-pteaex.c
25
if (vma && current->active_mm != vma->vm_mm)
arch/sh/mm/tlb-sh3.c
34
if (vma && current->active_mm != vma->vm_mm)
arch/sh/mm/tlb-sh4.c
23
if (vma && current->active_mm != vma->vm_mm)
arch/sh/mm/tlbex_32.c
41
if (unlikely(address >= TASK_SIZE || !current->mm))
arch/sh/mm/tlbex_32.c
44
pgd = pgd_offset(current->mm, address);
arch/sh/mm/tlbflush_32.c
117
if (mm == current->mm)
arch/sh/mm/tlbflush_32.c
28
if (vma->vm_mm != current->mm) {
arch/sh/mm/tlbflush_32.c
53
if (mm == current->mm)
arch/sh/mm/tlbflush_32.c
63
if (mm != current->mm) {
arch/sparc/include/asm/cacheflush_64.h
22
do { if ((__mm) == current->mm) flushw_user(); } while(0)
arch/sparc/include/asm/current.h
18
register struct task_struct *current asm("g4");
arch/sparc/include/asm/elf_64.h
210
if (personality(current->personality) != PER_LINUX32) \
arch/sparc/include/asm/elf_64.h
212
(current->personality & (~PER_MASK))); \
arch/sparc/include/asm/elf_64.h
222
(unsigned long)current->mm->context.vdso); \
arch/sparc/include/asm/mman.h
22
if (current->mm == mm) {
arch/sparc/include/asm/mman.h
25
regs = task_pt_regs(current);
arch/sparc/include/asm/mman.h
36
if (!current->mm->context.adi) {
arch/sparc/include/asm/mman.h
37
regs = task_pt_regs(current);
arch/sparc/include/asm/mman.h
39
current->mm->context.adi = true;
arch/sparc/include/asm/mman.h
40
on_each_cpu_mask(mm_cpumask(current->mm),
arch/sparc/include/asm/mman.h
41
ipi_set_tstate_mcde, current->mm, 0);
arch/sparc/include/asm/mmu_context_64.h
179
if (current && current->mm && current->mm->context.adi) {
arch/sparc/include/asm/mmu_context_64.h
182
regs = task_pt_regs(current);
arch/sparc/include/asm/pgtable_64.h
965
flush_dcache_folio_all(current->mm, \
arch/sparc/include/asm/sfp-machine_32.h
185
#define FP_ROUNDMODE ((current->thread.fsr >> 30) & 0x3)
arch/sparc/include/asm/sfp-machine_32.h
201
#define FP_INHIBIT_RESULTS ((current->thread.fsr >> 23) & _fex)
arch/sparc/include/asm/sfp-machine_32.h
207
#define FP_TRAPPING_EXCEPTIONS ((current->thread.fsr >> 23) & 0x1f)
arch/sparc/include/asm/switch_to_64.h
55
: "=&r" (last), "=r" (current), "=r" (current_thread_info_reg), \
arch/sparc/kernel/ftrace.c
125
if (unlikely(atomic_read(&current->tracing_graph_pause)))
arch/sparc/kernel/leon_smp.c
96
current->active_mm = &init_mm;
arch/sparc/kernel/perf_event.c
1774
ret_stack = ftrace_graph_get_ret_stack(current,
arch/sparc/kernel/perf_event.c
1860
if (!current->mm)
arch/sparc/kernel/process_32.c
153
tsk = current;
arch/sparc/kernel/process_32.c
155
if (tsk == current && !_ksp)
arch/sparc/kernel/process_32.c
201
if(last_task_used_math == current) {
arch/sparc/kernel/process_32.c
207
fpsave(&current->thread.float_regs[0], &current->thread.fsr,
arch/sparc/kernel/process_32.c
208
&current->thread.fpqueue[0], &current->thread.fpqdepth);
arch/sparc/kernel/process_32.c
273
if(last_task_used_math == current) {
arch/sparc/kernel/process_32.c
317
ti->kpsr = current->thread.fork_kpsr | PSR_PIL;
arch/sparc/kernel/process_32.c
318
ti->kwim = current->thread.fork_kwim;
arch/sparc/kernel/process_32.c
365
childregs->u_regs[UREG_I0] = current->pid;
arch/sparc/kernel/process_64.c
185
show_stack(current, (unsigned long *)regs->u_regs[UREG_FP], KERN_DEFAULT);
arch/sparc/kernel/process_64.c
547
current->comm, current->pid,
arch/sparc/kernel/process_64.c
634
t->kregs->u_regs[UREG_I0] = current->pid;
arch/sparc/kernel/ptrace_32.c
116
if (target == current)
arch/sparc/kernel/ptrace_32.c
171
if (target == current)
arch/sparc/kernel/ptrace_32.c
191
if (target == current)
arch/sparc/kernel/ptrace_32.c
250
if (target == current)
arch/sparc/kernel/ptrace_32.c
269
if (target == current)
arch/sparc/kernel/ptrace_32.c
292
if (target == current)
arch/sparc/kernel/ptrace_32.c
309
if (target == current)
arch/sparc/kernel/ptrace_32.c
363
unsigned long addr2 = current->thread.kregs->u_regs[UREG_I4];
arch/sparc/kernel/ptrace_32.c
54
if (target == current) {
arch/sparc/kernel/ptrace_32.c
72
if (target == current) {
arch/sparc/kernel/ptrace_32.c
90
if (target == current)
arch/sparc/kernel/ptrace_64.c
1013
const struct user_regset_view *view = task_user_regset_view(current);
arch/sparc/kernel/ptrace_64.c
1014
unsigned long addr2 = task_pt_regs(current)->u_regs[UREG_I4];
arch/sparc/kernel/ptrace_64.c
163
if (target == current) {
arch/sparc/kernel/ptrace_64.c
178
if (target == current) {
arch/sparc/kernel/ptrace_64.c
253
if (target == current)
arch/sparc/kernel/ptrace_64.c
275
if (target == current)
arch/sparc/kernel/ptrace_64.c
348
if (target == current)
arch/sparc/kernel/ptrace_64.c
380
if (target == current)
arch/sparc/kernel/ptrace_64.c
448
if (target == current)
arch/sparc/kernel/ptrace_64.c
467
if (target == current)
arch/sparc/kernel/ptrace_64.c
540
if (target == current)
arch/sparc/kernel/ptrace_64.c
569
if (target == current)
arch/sparc/kernel/ptrace_64.c
581
if (target == current) {
arch/sparc/kernel/ptrace_64.c
609
if (target == current) {
arch/sparc/kernel/ptrace_64.c
683
if (target == current)
arch/sparc/kernel/ptrace_64.c
707
if (target == current)
arch/sparc/kernel/ptrace_64.c
783
if (target == current)
arch/sparc/kernel/ptrace_64.c
805
if (target == current)
arch/sparc/kernel/ptrace_64.c
835
if (target == current)
arch/sparc/kernel/ptrace_64.c
855
if (target == current)
arch/sparc/kernel/ptrace_64.c
930
compat_ulong_t caddr2 = task_pt_regs(current)->u_regs[UREG_I4];
arch/sparc/kernel/setup_32.c
74
if (!is_idle_task(current)) {
arch/sparc/kernel/signal32.c
184
current->restart_block.fn = do_no_restart_syscall;
arch/sparc/kernel/signal32.c
317
pgdp = pgd_offset(current->mm, address);
arch/sparc/kernel/signal32.c
380
current->comm, current->pid, (unsigned long)sf,
arch/sparc/kernel/signal32.c
514
current->comm, current->pid, (unsigned long)sf,
arch/sparc/kernel/signal32.c
717
if (put_user(current->sas_ss_sp + current->sas_ss_size,
arch/sparc/kernel/signal32.c
734
if (current->sas_ss_sp && on_sig_stack(sp))
arch/sparc/kernel/signal32.c
741
current->sas_ss_sp = (unsigned long)ss_sp - SIGSTKSZ;
arch/sparc/kernel/signal32.c
742
current->sas_ss_size = SIGSTKSZ;
arch/sparc/kernel/signal32.c
95
current->restart_block.fn = do_no_restart_syscall;
arch/sparc/kernel/signal_32.c
315
flush_sig_insns(current->mm, (unsigned long) &(sf->insns[0]));
arch/sparc/kernel/signal_32.c
412
flush_sig_insns(current->mm, (unsigned long) &(sf->insns[0]));
arch/sparc/kernel/signal_32.c
538
if (put_user(current->sas_ss_sp + current->sas_ss_size,
arch/sparc/kernel/signal_32.c
553
if (current->sas_ss_sp && on_sig_stack(sp))
arch/sparc/kernel/signal_32.c
559
current->sas_ss_sp = (unsigned long)ss_sp - SIGSTKSZ;
arch/sparc/kernel/signal_32.c
560
current->sas_ss_size = SIGSTKSZ;
arch/sparc/kernel/signal_32.c
84
current->restart_block.fn = do_no_restart_syscall;
arch/sparc/kernel/signal_64.c
174
err |= __put_user(current->blocked.sig[0],
arch/sparc/kernel/signal_64.c
177
err |= __copy_to_user(&ucp->uc_sigmask, &current->blocked,
arch/sparc/kernel/signal_64.c
265
current->restart_block.fn = do_no_restart_syscall;
arch/sparc/kernel/signal_64.c
374
current->comm, current->pid, (unsigned long)sf,
arch/sparc/kernel/sigutil_32.c
19
if (test_tsk_thread_flag(current, TIF_USEDFPU)) {
arch/sparc/kernel/sigutil_32.c
21
fpsave(&current->thread.float_regs[0], &current->thread.fsr,
arch/sparc/kernel/sigutil_32.c
22
&current->thread.fpqueue[0], &current->thread.fpqdepth);
arch/sparc/kernel/sigutil_32.c
24
clear_tsk_thread_flag(current, TIF_USEDFPU);
arch/sparc/kernel/sigutil_32.c
27
if (current == last_task_used_math) {
arch/sparc/kernel/sigutil_32.c
29
fpsave(&current->thread.float_regs[0], &current->thread.fsr,
arch/sparc/kernel/sigutil_32.c
30
&current->thread.fpqueue[0], &current->thread.fpqdepth);
arch/sparc/kernel/sigutil_32.c
36
&current->thread.float_regs[0],
arch/sparc/kernel/sigutil_32.c
38
err |= __put_user(current->thread.fsr, &fpu->si_fsr);
arch/sparc/kernel/sigutil_32.c
39
err |= __put_user(current->thread.fpqdepth, &fpu->si_fpqdepth);
arch/sparc/kernel/sigutil_32.c
40
if (current->thread.fpqdepth != 0)
arch/sparc/kernel/sigutil_32.c
42
&current->thread.fpqueue[0],
arch/sparc/kernel/sigutil_32.c
57
if (test_tsk_thread_flag(current, TIF_USEDFPU))
arch/sparc/kernel/sigutil_32.c
60
if (current == last_task_used_math) {
arch/sparc/kernel/sigutil_32.c
66
clear_tsk_thread_flag(current, TIF_USEDFPU);
arch/sparc/kernel/sigutil_32.c
71
err = __copy_from_user(&current->thread.float_regs[0], &fpu->si_float_regs[0],
arch/sparc/kernel/sigutil_32.c
73
err |= __get_user(current->thread.fsr, &fpu->si_fsr);
arch/sparc/kernel/sigutil_32.c
74
err |= __get_user(current->thread.fpqdepth, &fpu->si_fpqdepth);
arch/sparc/kernel/sigutil_32.c
75
if (current->thread.fpqdepth != 0)
arch/sparc/kernel/sigutil_32.c
76
err |= __copy_from_user(&current->thread.fpqueue[0],
arch/sparc/kernel/smp_64.c
131
current->active_mm = &init_mm;
arch/sparc/kernel/sun4d_smp.c
98
current->active_mm = &init_mm;
arch/sparc/kernel/sun4m_smp.c
64
current->active_mm = &init_mm;
arch/sparc/kernel/sys_sparc_32.c
146
current->comm, task_pid_nr(current), (int)regs->u_regs[1]);
arch/sparc/kernel/sys_sparc_64.c
103
struct mm_struct *mm = current->mm;
arch/sparc/kernel/sys_sparc_64.c
168
struct mm_struct *mm = current->mm;
arch/sparc/kernel/sys_sparc_64.c
287
if (current->flags & PF_RANDOMIZE) {
arch/sparc/kernel/sys_sparc_64.c
308
(current->personality & ADDR_COMPAT_LAYOUT) ||
arch/sparc/kernel/sys_sparc_64.c
442
if (personality(current->personality) == PER_LINUX32 &&
arch/sparc/kernel/sys_sparc_64.c
602
current->pid, current->comm);
arch/sparc/kernel/traps_32.c
118
send_sig_fault(SIGILL, ILL_ILLOPC, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
126
send_sig_fault(SIGILL, ILL_PRVOPC, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
147
current);
arch/sparc/kernel/traps_32.c
167
if(last_task_used_math == current)
arch/sparc/kernel/traps_32.c
175
last_task_used_math = current;
arch/sparc/kernel/traps_32.c
177
fpload(&current->thread.float_regs[0], &current->thread.fsr);
arch/sparc/kernel/traps_32.c
188
fpload(&current->thread.float_regs[0], &current->thread.fsr);
arch/sparc/kernel/traps_32.c
209
struct task_struct *fpt = current;
arch/sparc/kernel/traps_32.c
256
fpload(&current->thread.float_regs[0], &current->thread.fsr);
arch/sparc/kernel/traps_32.c
307
send_sig_fault(SIGEMT, EMT_TAGOVF, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
335
send_sig_fault(SIGILL, ILL_COPROC, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
345
send_sig_fault(SIGILL, ILL_COPROC, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
351
send_sig_fault(SIGFPE, FPE_INTDIV, (void __user *)pc, current);
arch/sparc/kernel/traps_32.c
390
current->active_mm = &init_mm;
arch/sparc/kernel/traps_32.c
63
printk("%s(%d): %s [#%d]\n", current->comm, task_pid_nr(current), str, ++die_counter);
arch/sparc/kernel/traps_64.c
2461
tsk = current;
arch/sparc/kernel/traps_64.c
2464
if (tsk == current)
arch/sparc/kernel/traps_64.c
2532
printk("%s(%d): %s [#%d]\n", current->comm, task_pid_nr(current), str, ++die_counter);
arch/sparc/kernel/traps_64.c
2924
current->active_mm = &init_mm;
arch/sparc/kernel/unaligned_32.c
228
(current->mm ? current->mm->context :
arch/sparc/kernel/unaligned_32.c
229
current->active_mm->context));
arch/sparc/kernel/unaligned_32.c
231
(current->mm ? (unsigned long) current->mm->pgd :
arch/sparc/kernel/unaligned_32.c
232
(unsigned long) current->active_mm->pgd));
arch/sparc/kernel/unaligned_32.c
281
current);
arch/sparc/kernel/unaligned_32.c
61
die_if_kernel("Byte sized unaligned access?!?!", current->thread.kregs);
arch/sparc/kernel/unaligned_64.c
279
(current->mm ? CTX_HWBITS(current->mm->context) :
arch/sparc/kernel/unaligned_64.c
280
CTX_HWBITS(current->active_mm->context)));
arch/sparc/kernel/unaligned_64.c
282
(current->mm ? (unsigned long) current->mm->pgd :
arch/sparc/kernel/unaligned_64.c
283
(unsigned long) current->active_mm->pgd));
arch/sparc/kernel/uprobes.c
194
struct uprobe_task *utask = current->utask;
arch/sparc/kernel/uprobes.c
195
struct arch_uprobe_task *autask = &current->utask->autask;
arch/sparc/kernel/uprobes.c
220
struct uprobe_task *utask = current->utask;
arch/sparc/kernel/uprobes.c
297
struct uprobe_task *utask = current->utask;
arch/sparc/mm/fault_32.c
116
struct task_struct *tsk = current;
arch/sparc/mm/fault_32.c
308
struct task_struct *tsk = current;
arch/sparc/mm/fault_32.c
84
addr, current);
arch/sparc/mm/fault_64.c
180
show_signal_msg(regs, sig, code, addr, current);
arch/sparc/mm/fault_64.c
252
unhandled_fault (address, current, regs);
arch/sparc/mm/fault_64.c
262
current->comm, current->pid,
arch/sparc/mm/fault_64.c
270
struct mm_struct *mm = current->mm;
arch/sparc/mm/fault_64.c
69
unhandled_fault(regs->tpc, current, regs);
arch/sparc/mm/fault_64.c
81
pgd_t *pgdp = pgd_offset(current->mm, tpc);
arch/sparc/mm/init_64.c
2970
if (mm == current->mm)
arch/sparc/mm/init_64.c
2976
struct mm_struct *mm = current->mm;
arch/sparc/mm/srmmu.c
1664
if (atomic_read(&mm->mm_users) == 1 && current->active_mm == mm)
arch/sparc/mm/tsb.c
268
current->comm, current->pid, tsb_bytes);
arch/sparc/power/hibernate.c
38
struct mm_struct *mm = current->active_mm;
arch/sparc/vdso/vma.c
365
struct mm_struct *mm = current->mm;
arch/sparc/vdso/vma.c
376
if (current->flags & PF_RANDOMIZE) {
arch/sparc/vdso/vma.c
394
current->mm->context.vdso = (void __user *)text_start;
arch/sparc/vdso/vma.c
424
current->mm->context.vdso = NULL;
arch/um/drivers/mconsole_kern.h
27
#define CONFIG_CHUNK(str, size, current, chunk, end) \
arch/um/drivers/mconsole_kern.h
29
current += strlen(chunk); \
arch/um/drivers/mconsole_kern.h
30
if(current >= size) \
arch/um/drivers/mconsole_kern.h
37
current++; \
arch/um/include/asm/stacktrace.h
21
if (!task || task == current)
arch/um/include/asm/stacktrace.h
36
if (!task || task == current)
arch/um/include/asm/uaccess.h
52
current->thread.segv_continue = NULL; \
arch/um/include/asm/uaccess.h
64
current->thread.segv_continue = NULL; \
arch/um/kernel/exec.c
25
arch_flush_thread(&current->thread.arch);
arch/um/kernel/process.c
103
return task_pid_nr(current);
arch/um/kernel/process.c
115
if (current->thread.prev_sched != NULL)
arch/um/kernel/process.c
116
schedule_tail(current->thread.prev_sched);
arch/um/kernel/process.c
117
current->thread.prev_sched = NULL;
arch/um/kernel/process.c
119
fn = current->thread.request.thread.proc;
arch/um/kernel/process.c
120
arg = current->thread.request.thread.arg;
arch/um/kernel/process.c
126
userspace(&current->thread.regs.regs);
arch/um/kernel/process.c
132
schedule_tail(current->thread.prev_sched);
arch/um/kernel/process.c
139
arch_switch_to(current);
arch/um/kernel/process.c
141
current->thread.prev_sched = NULL;
arch/um/kernel/process.c
143
userspace(&current->thread.regs.regs);
arch/um/kernel/process.c
165
arch_copy_thread(&current->thread.arch, &p->thread.arch);
arch/um/kernel/process.c
273
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/um/kernel/process.c
79
arch_switch_to(current);
arch/um/kernel/process.c
81
return current->thread.prev_sched;
arch/um/kernel/process.c
86
struct pt_regs *regs = &current->thread.regs;
arch/um/kernel/ptrace.c
143
int ptraced = current->ptrace;
arch/um/kernel/signal.c
56
if (test_thread_flag(TIF_SINGLESTEP) && (current->ptrace & PT_PTRACED))
arch/um/kernel/signal.c
83
sp = current->sas_ss_sp + current->sas_ss_size;
arch/um/kernel/skas/process.c
47
if (current->mm == NULL)
arch/um/kernel/skas/process.c
50
return current->mm->context.id.stack;
arch/um/kernel/skas/process.c
55
if (current->mm == NULL)
arch/um/kernel/skas/process.c
58
return &current->mm->context.id;
arch/um/kernel/skas/process.c
63
if (current->mm == NULL)
arch/um/kernel/skas/process.c
66
um_tlb_sync(current->mm);
arch/um/kernel/skas/uaccess.c
48
pte_t *pte = virt_to_pte(current->mm, virt);
arch/um/kernel/skas/uaccess.c
56
pte = virt_to_pte(current->mm, virt);
arch/um/kernel/smp.c
68
if (current->mm)
arch/um/kernel/smp.c
69
os_alarm_process(current->mm->context.id.pid);
arch/um/kernel/stacktrace.c
67
__save_stack_trace(current, trace);
arch/um/kernel/sysrq.c
32
struct pt_regs *segv_regs = current->thread.segv_regs;
arch/um/kernel/sysrq.c
49
dump_trace(task ?: current, &stackops, (void *)loglvl);
arch/um/kernel/tlb.c
212
if (atomic_read(&current->mm->mm_users) == 0)
arch/um/kernel/tlb.c
215
flush_tlb_mm(current->mm);
arch/um/kernel/trap.c
143
struct mm_struct *mm = current->mm;
arch/um/kernel/trap.c
182
if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current))
arch/um/kernel/trap.c
241
struct task_struct *tsk = current;
arch/um/kernel/trap.c
262
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
269
do_signal(&current->thread.regs);
arch/um/kernel/trap.c
317
current->thread.segv_regs = container_of(regs, struct pt_regs, regs);
arch/um/kernel/trap.c
332
else if (current->pagefault_disabled) {
arch/um/kernel/trap.c
337
if (!current->thread.segv_continue) {
arch/um/kernel/trap.c
341
mc_set_rip(mc, current->thread.segv_continue);
arch/um/kernel/trap.c
342
current->thread.segv_continue = NULL;
arch/um/kernel/trap.c
345
else if (current->mm == NULL) {
arch/um/kernel/trap.c
382
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
386
current->thread.arch.faultinfo = fi;
arch/um/kernel/trap.c
392
current->thread.segv_regs = NULL;
arch/um/kernel/trap.c
417
current->thread.arch.faultinfo = *fi;
arch/x86/coco/sev/vc-handle.c
191
if (user_mode(ctxt->regs) || mm_is_efi(current->active_mm))
arch/x86/coco/sev/vc-handle.c
79
struct thread_struct *t = &current->thread;
arch/x86/entry/syscall_32.c
321
unsigned long landing_pad = (unsigned long)current->mm->context.vdso +
arch/x86/entry/vdso/extable.c
15
const struct vdso_image *image = current->mm->context.vdso_image;
arch/x86/entry/vdso/extable.c
28
if (!current->mm->context.vdso)
arch/x86/entry/vdso/extable.c
31
base = (unsigned long)current->mm->context.vdso + image->extable_base;
arch/x86/entry/vdso/vma.c
135
struct mm_struct *mm = current->mm;
arch/x86/entry/vdso/vma.c
189
current->mm->context.vdso = (void __user *)text_start;
arch/x86/entry/vdso/vma.c
190
current->mm->context.vdso_image = image;
arch/x86/entry/vdso/vma.c
199
struct mm_struct *mm = current->mm;
arch/x86/entry/vdso/vma.c
263
const struct vdso_image *image = current->mm->context.vdso_image;
arch/x86/entry/vdso/vma.c
264
unsigned long vdso = (unsigned long) current->mm->context.vdso;
arch/x86/entry/vdso/vma.c
70
(unsigned long)current->mm->context.vdso;
arch/x86/entry/vdso/vma.c
79
const struct vdso_image *image = current->mm->context.vdso_image;
arch/x86/entry/vdso/vma.c
82
current->mm->context.vdso = (void __user *)new_vma->vm_start;
arch/x86/entry/vsyscall/vsyscall_64.c
101
struct thread_struct *thread = &current->thread;
arch/x86/entry/vsyscall/vsyscall_64.c
79
level, current->comm, task_pid_nr(current),
arch/x86/events/core.c
2866
unwind_start(&state, current, regs, NULL);
arch/x86/events/core.c
2868
unwind_start(&state, current, NULL, (void *)regs->sp);
arch/x86/events/core.c
2901
ldt = smp_load_acquire(&current->mm->context.ldt);
arch/x86/events/utils.c
121
if (!current->mm)
arch/x86/include/asm/compat.h
80
(!!(task_pt_regs(current)->orig_ax & __X32_SYSCALL_BIT))
arch/x86/include/asm/compat.h
86
if (task_pt_regs(current)->orig_ax & __X32_SYSCALL_BIT)
arch/x86/include/asm/elf.h
167
elf_common_init(&current->thread, _r, 0)
arch/x86/include/asm/elf.h
170
elf_common_init(&current->thread, regs, __USER_DS)
arch/x86/include/asm/elf.h
335
(unsigned long __force)current->mm->context.vdso); \
arch/x86/include/asm/elf.h
344
(unsigned long __force)current->mm->context.vdso); \
arch/x86/include/asm/elf.h
360
#define VDSO_CURRENT_BASE ((unsigned long)current->mm->context.vdso)
arch/x86/include/asm/elf.h
363
((unsigned long)current->mm->context.vdso + \
arch/x86/include/asm/elfcore-compat.h
23
(user_64bit_mode(task_pt_regs(current)) \
arch/x86/include/asm/elfcore-compat.h
27
(*(user_64bit_mode(task_pt_regs(current)) \
arch/x86/include/asm/entry-common.h
43
WARN_ON_ONCE(regs != task_pt_regs(current));
arch/x86/include/asm/fred.h
101
unsigned long rsp0 = (unsigned long) task_stack_page(current) + THREAD_SIZE;
arch/x86/include/asm/page_64_types.h
59
#define IA32_PAGE_OFFSET ((current->personality & ADDR_LIMIT_3GB) ? \
arch/x86/include/asm/stackprotector.h
40
current->stack_canary = canary;
arch/x86/include/asm/stacktrace.h
45
if (get_stack_info_noinstr(stack, current, info))
arch/x86/include/asm/stacktrace.h
48
return get_stack_info_noinstr((void *)stack + PAGE_SIZE, current, info);
arch/x86/include/asm/stacktrace.h
76
if (task == current)
arch/x86/include/asm/stacktrace.h
95
if (task == current)
arch/x86/include/asm/unwind.h
138
if (task == current) \
arch/x86/include/asm/unwind.h
148
return task != current && task->on_cpu;
arch/x86/kernel/acpi/sleep.c
131
current->thread.sp = (unsigned long)temp_stack + sizeof(temp_stack);
arch/x86/kernel/apic/x2apic_uv_x.c
1584
current->comm);
arch/x86/kernel/apic/x2apic_uv_x.c
1592
current->comm);
arch/x86/kernel/apic/x2apic_uv_x.c
1600
current->comm);
arch/x86/kernel/apm_32.c
1406
DECLARE_WAITQUEUE(wait, current);
arch/x86/kernel/apm_32.c
1471
if (signal_pending(current))
arch/x86/kernel/apm_32.c
1702
set_cpus_allowed_ptr(current, cpumask_of(0));
arch/x86/kernel/apm_32.c
913
task_cputime(current, &utime, &stime);
arch/x86/kernel/apm_32.c
984
set_cpus_allowed_ptr(current, cpumask_of(0));
arch/x86/kernel/cet.c
63
tsk = current;
arch/x86/kernel/cpu/bugs.c
2486
if (tsk == current)
arch/x86/kernel/cpu/bugs.c
2602
if (task == current)
arch/x86/kernel/cpu/bus_lock.c
248
if (!current->reported_split_lock)
arch/x86/kernel/cpu/bus_lock.c
250
current->comm, current->pid, ip);
arch/x86/kernel/cpu/bus_lock.c
251
current->reported_split_lock = 1;
arch/x86/kernel/cpu/bus_lock.c
285
current->comm, current->pid,
arch/x86/kernel/cpu/bus_lock.c
288
current->thread.error_code = 0;
arch/x86/kernel/cpu/bus_lock.c
289
current->thread.trap_nr = X86_TRAP_AC;
arch/x86/kernel/cpu/bus_lock.c
340
current->comm, current->pid, regs->ip);
arch/x86/kernel/cpu/common.c
2465
struct task_struct *cur = current;
arch/x86/kernel/cpu/mce/core.c
1491
int count = ++current->mce_count;
arch/x86/kernel/cpu/mce/core.c
1496
current->mce_addr = m->addr;
arch/x86/kernel/cpu/mce/core.c
1497
current->mce_kflags = m->kflags;
arch/x86/kernel/cpu/mce/core.c
1498
current->mce_ripv = !!(m->mcgstatus & MCG_STATUS_RIPV);
arch/x86/kernel/cpu/mce/core.c
1499
current->mce_whole_page = whole_page(m);
arch/x86/kernel/cpu/mce/core.c
1500
current->mce_kill_me.func = func;
arch/x86/kernel/cpu/mce/core.c
1509
if (count > 1 && (current->mce_addr >> PAGE_SHIFT) != (m->addr >> PAGE_SHIFT))
arch/x86/kernel/cpu/mce/core.c
1516
task_work_add(current, &current->mce_kill_me, TWA_RESUME);
arch/x86/kernel/cpu/mce/severity.c
270
current->mce_vaddr = (void __user *)addr;
arch/x86/kernel/cpu/resctrl/rdtgroup.c
61
resctrl_arch_sched_in(current);
arch/x86/kernel/cpu/sgx/encl.c
1013
return get_mem_cgroup_from_mm(current->mm);
arch/x86/kernel/cpu/sgx/encl.c
541
if (current->personality & READ_IMPLIES_EXEC)
arch/x86/kernel/cpu/sgx/ioctl.c
218
vma = find_vma(current->mm, src);
arch/x86/kernel/cpu/sgx/ioctl.c
292
mmap_read_lock(current->mm);
arch/x86/kernel/cpu/sgx/ioctl.c
335
mmap_read_unlock(current->mm);
arch/x86/kernel/cpu/sgx/ioctl.c
344
mmap_read_unlock(current->mm);
arch/x86/kernel/cpu/sgx/ioctl.c
443
if (signal_pending(current)) {
arch/x86/kernel/cpu/sgx/ioctl.c
535
if (signal_pending(current)) {
arch/x86/kernel/cpu/sgx/main.c
442
return current == ksgxd_tsk;
arch/x86/kernel/cpu/sgx/main.c
583
if (signal_pending(current)) {
arch/x86/kernel/dumpstack.c
314
bool disable_kasan = task && task != current;
arch/x86/kernel/dumpstack.c
328
task = task ? : current;
arch/x86/kernel/dumpstack.c
334
if (!sp && task == current)
arch/x86/kernel/dumpstack.c
335
sp = get_stack_pointer(current, NULL);
arch/x86/kernel/dumpstack.c
342
show_trace_log_lvl(current, regs, NULL, KERN_DEFAULT);
arch/x86/kernel/dumpstack.c
377
if (regs && kexec_should_crash(current))
arch/x86/kernel/dumpstack.c
407
kasan_unpoison_task_stack(current);
arch/x86/kernel/dumpstack.c
435
current->thread.trap_nr, SIGSEGV) == NOTIFY_STOP)
arch/x86/kernel/dumpstack.c
489
show_trace_log_lvl(current, regs, NULL, KERN_DEFAULT);
arch/x86/kernel/dumpstack.c
80
if (regs != task_pt_regs(current))
arch/x86/kernel/dumpstack_32.c
114
task = task ? : current;
arch/x86/kernel/dumpstack_32.c
119
if (task != current)
arch/x86/kernel/dumpstack_64.c
176
if (task != current)
arch/x86/kernel/dumpstack_64.c
194
task = task ? : current;
arch/x86/kernel/dumpstack_64.c
209
if (task == current)
arch/x86/kernel/fpu/context.h
56
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/context.h
59
if (WARN_ON_ONCE(current->flags & (PF_KTHREAD | PF_USER_WORKER)))
arch/x86/kernel/fpu/core.c
232
spin_lock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/core.c
233
fpuperm = &x86_task_fpu(current->group_leader)->guest_perm;
arch/x86/kernel/fpu/core.c
239
spin_unlock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/core.c
363
struct fpstate *fpstate = x86_task_fpu(current)->fpstate;
arch/x86/kernel/fpu/core.c
377
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/core.c
483
if (!(current->flags & (PF_KTHREAD | PF_USER_WORKER)) &&
arch/x86/kernel/fpu/core.c
486
save_fpregs_to_fpstate(x86_task_fpu(current));
arch/x86/kernel/fpu/core.c
516
WARN_ON_FPU(fpu != x86_task_fpu(current));
arch/x86/kernel/fpu/core.c
623
struct fpu *src_fpu = x86_task_fpu(current->group_leader);
arch/x86/kernel/fpu/core.c
625
spin_lock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/core.c
629
spin_unlock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/core.c
775
if (fpu == x86_task_fpu(current)) {
arch/x86/kernel/fpu/core.c
809
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/core.c
838
WARN_ON_FPU(fpu != x86_task_fpu(current));
arch/x86/kernel/fpu/core.c
874
fpstate_reset(x86_task_fpu(current));
arch/x86/kernel/fpu/core.c
897
WARN_ON_ONCE(current->flags & PF_KTHREAD);
arch/x86/kernel/fpu/core.c
915
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/core.c
927
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/regset.c
351
if (tsk == current) {
arch/x86/kernel/fpu/regset.c
48
if (fpu == x86_task_fpu(current))
arch/x86/kernel/fpu/regset.c
66
WARN_ON_FPU(fpu == x86_task_fpu(current));
arch/x86/kernel/fpu/signal.c
177
struct task_struct *tsk = current;
arch/x86/kernel/fpu/signal.c
188
fpregs_soft_get(current, NULL, (struct membuf){.p = &fp,
arch/x86/kernel/fpu/signal.c
266
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/signal.c
325
struct task_struct *tsk = current;
arch/x86/kernel/fpu/signal.c
446
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/signal.c
46
if (__get_user(magic2, (__u32 __user *)(fpstate + x86_task_fpu(current)->fpstate->user_size)))
arch/x86/kernel/fpu/signal.c
476
success = !fpregs_soft_set(current, NULL, 0,
arch/x86/kernel/fpu/signal.c
493
unsigned long frame_size = xstate_sigframe_size(x86_task_fpu(current)->fpstate);
arch/x86/kernel/fpu/signal.c
52
trace_x86_fpu_xstate_check_failed(x86_task_fpu(current));
arch/x86/kernel/fpu/xstate.c
1507
if (fpstate->xfd == x86_task_fpu(current)->fpstate->xfd)
arch/x86/kernel/fpu/xstate.c
1584
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/fpu/xstate.c
1653
struct task_struct *thread, *leader = current->group_leader;
arch/x86/kernel/fpu/xstate.c
1656
lockdep_assert_held(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
1677
struct fpu *fpu = x86_task_fpu(current->group_leader);
arch/x86/kernel/fpu/xstate.c
1750
spin_lock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
1758
spin_unlock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
1776
spin_lock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
1780
spin_unlock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
1784
fpu = x86_task_fpu(current->group_leader);
arch/x86/kernel/fpu/xstate.c
1795
spin_unlock_irq(&current->sighand->siglock);
arch/x86/kernel/fpu/xstate.c
781
fpstate_reset(x86_task_fpu(current));
arch/x86/kernel/fpu/xstate.c
990
wrmsrq(MSR_IA32_XFD, x86_task_fpu(current)->fpstate->xfd);
arch/x86/kernel/fpu/xstate.h
26
struct fpu *fpu = x86_task_fpu(current->group_leader);
arch/x86/kernel/fpu/xstate.h
302
struct fpstate *fpstate = x86_task_fpu(current)->fpstate;
arch/x86/kernel/fpu/xstate.h
336
xfd_validate_state(x86_task_fpu(current)->fpstate, mask, true);
arch/x86/kernel/ftrace.c
630
if (atomic_read(&current->tracing_graph_pause))
arch/x86/kernel/hw_breakpoint.c
571
if ((current->thread.virtual_dr6 & DR_TRAP_BITS) ||
arch/x86/kernel/ioport.c
111
io_bitmap_exit(current);
arch/x86/kernel/ioport.c
143
io_bitmap_exit(current);
arch/x86/kernel/ioport.c
181
struct thread_struct *t = &current->thread;
arch/x86/kernel/ioport.c
25
if (current->thread.io_bitmap) {
arch/x86/kernel/ioport.c
30
refcount_inc(&current->thread.io_bitmap->refcnt);
arch/x86/kernel/ioport.c
31
tsk->thread.io_bitmap = current->thread.io_bitmap;
arch/x86/kernel/ioport.c
38
struct task_struct *tsk = current;
arch/x86/kernel/ioport.c
62
if (tsk == current)
arch/x86/kernel/ioport.c
73
struct thread_struct *t = &current->thread;
arch/x86/kernel/kgdb.c
629
struct task_struct *tsk = current;
arch/x86/kernel/ldt.c
504
struct mm_struct *mm = current->mm;
arch/x86/kernel/ldt.c
580
struct mm_struct *mm = current->mm;
arch/x86/kernel/msr.c
102
reg, current->comm, current->pid);
arch/x86/kernel/perf_regs.c
106
regs_user->regs = task_pt_regs(current);
arch/x86/kernel/perf_regs.c
107
regs_user->abi = perf_reg_abi(current);
arch/x86/kernel/perf_regs.c
137
struct pt_regs *user_regs = task_pt_regs(current);
arch/x86/kernel/perf_regs.c
141
regs_user->abi = perf_reg_abi(current);
arch/x86/kernel/process.c
1021
if (!(current->personality & ADDR_NO_RANDOMIZE) && randomize_va_space)
arch/x86/kernel/process.c
1079
return do_arch_prctl_64(current, option, arg2);
arch/x86/kernel/process.c
195
p->thread.fsindex = current->thread.fsindex;
arch/x86/kernel/process.c
196
p->thread.fsbase = current->thread.fsbase;
arch/x86/kernel/process.c
197
p->thread.gsindex = current->thread.gsindex;
arch/x86/kernel/process.c
198
p->thread.gsbase = current->thread.gsbase;
arch/x86/kernel/process.c
269
if (!ret && unlikely(test_tsk_thread_flag(current, TIF_IO_BITMAP)))
arch/x86/kernel/process.c
286
struct task_struct *tsk = current;
arch/x86/kernel/process.c
422
task_spec_ssb_noexec(current)) {
arch/x86/kernel/process.c
424
task_clear_spec_ssb_disable(current);
arch/x86/kernel/process.c
425
task_clear_spec_ssb_noexec(current);
arch/x86/kernel/process.c
429
mm_reset_untag_mask(current->mm);
arch/x86/kernel/process.c
473
struct thread_struct *t = &current->thread;
arch/x86/kernel/process.c
702
speculation_ctrl_update(speculation_ctrl_update_tif(current));
arch/x86/kernel/process_64.c
304
save_fsgs(current);
arch/x86/kernel/process_64.c
489
if (task == current)
arch/x86/kernel/process_64.c
504
if (task == current)
arch/x86/kernel/process_64.c
517
WARN_ON_ONCE(task == current);
arch/x86/kernel/process_64.c
524
WARN_ON_ONCE(task == current);
arch/x86/kernel/process_64.c
723
task_pt_regs(current)->orig_ax = __NR_execve;
arch/x86/kernel/process_64.c
725
if (current->mm)
arch/x86/kernel/process_64.c
726
__set_bit(MM_CONTEXT_HAS_VSYSCALL, &current->mm->context.flags);
arch/x86/kernel/process_64.c
732
current->personality &= ~READ_IMPLIES_EXEC;
arch/x86/kernel/process_64.c
738
if (current->mm)
arch/x86/kernel/process_64.c
739
current->mm->context.flags = 0;
arch/x86/kernel/process_64.c
741
current->personality &= ~READ_IMPLIES_EXEC;
arch/x86/kernel/process_64.c
750
task_pt_regs(current)->orig_ax = __NR_x32_execve | __X32_SYSCALL_BIT;
arch/x86/kernel/process_64.c
758
if (current->mm) {
arch/x86/kernel/process_64.c
763
__set_bit(MM_CONTEXT_UPROBE_IA32, &current->mm->context.flags);
arch/x86/kernel/process_64.c
766
current->personality |= force_personality32;
arch/x86/kernel/process_64.c
768
task_pt_regs(current)->orig_ax = __NR_ia32_execve;
arch/x86/kernel/process_64.c
834
if (current->mm != mm)
arch/x86/kernel/process_64.c
882
if (task == current) {
arch/x86/kernel/process_64.c
912
if (task == current) {
arch/x86/kernel/process_64.c
960
if (current != task)
arch/x86/kernel/ptrace.c
1410
struct task_struct *tsk = current;
arch/x86/kernel/ptrace.c
192
if (task == current)
arch/x86/kernel/ptrace.c
203
if (WARN_ON_ONCE(task == current))
arch/x86/kernel/ptrace.c
258
if (task == current) {
arch/x86/kernel/ptrace.c
265
if (task == current) {
arch/x86/kernel/ptrace.c
271
if (task == current) {
arch/x86/kernel/ptrace.c
277
if (task == current) {
arch/x86/kernel/ptrace.c
293
if (WARN_ON_ONCE(task == current))
arch/x86/kernel/ptrace.c
476
struct thread_struct *thread = &(current->thread);
arch/x86/kernel/shstk.c
104
struct mm_struct *mm = current->mm;
arch/x86/kernel/shstk.c
159
struct thread_shstk *shstk = &current->thread.shstk;
arch/x86/kernel/shstk.c
189
memset(&current->thread.shstk, 0, sizeof(struct thread_shstk));
arch/x86/kernel/shstk.c
190
current->thread.features = 0;
arch/x86/kernel/shstk.c
191
current->thread.features_locked = 0;
arch/x86/kernel/shstk.c
33
return current->thread.features & features;
arch/x86/kernel/shstk.c
354
if (mmap_read_lock_killable(current->mm))
arch/x86/kernel/shstk.c
362
vma = find_vma(current->mm, *ssp);
arch/x86/kernel/shstk.c
368
mmap_read_unlock(current->mm);
arch/x86/kernel/shstk.c
38
current->thread.features |= features;
arch/x86/kernel/shstk.c
384
mmap_read_unlock(current->mm);
arch/x86/kernel/shstk.c
43
current->thread.features &= ~features;
arch/x86/kernel/shstk.c
460
if (!tsk->mm || tsk->mm != current->mm)
arch/x86/kernel/shstk.c
540
shstk_free(current);
arch/x86/kernel/shstk.c
590
if (task != current) {
arch/x86/kernel/signal.c
119
sp = current->sas_ss_sp + current->sas_ss_size;
arch/x86/kernel/signal.c
155
current->comm, task_pid_nr(current));
arch/x86/kernel/signal.c
258
struct fpu *fpu = x86_task_fpu(current);
arch/x86/kernel/signal.c
264
if (syscall_get_nr(current, regs) != -1) {
arch/x86/kernel/signal.c
266
switch (syscall_get_error(current, regs)) {
arch/x86/kernel/signal.c
292
user_disable_single_step(current);
arch/x86/kernel/signal.c
318
if (current->restart_block.arch_data & TS_COMPAT)
arch/x86/kernel/signal.c
344
if (syscall_get_nr(current, regs) != -1) {
arch/x86/kernel/signal.c
346
switch (syscall_get_error(current, regs)) {
arch/x86/kernel/signal.c
370
struct task_struct *me = current;
arch/x86/kernel/signal.c
375
task_pid_nr(current) > 1 ? KERN_INFO : KERN_EMERG,
arch/x86/kernel/signal.c
421
lockdep_assert_held(&current->sighand->siglock);
arch/x86/kernel/signal.c
426
fsize += x86_task_fpu(current->group_leader)->perm.__user_state_size;
arch/x86/kernel/signal.c
433
mask = x86_task_fpu(current->group_leader)->perm.__state_perm;
arch/x86/kernel/signal_32.c
113
current->restart_block.fn = do_no_restart_syscall;
arch/x86/kernel/signal_32.c
234
unsafe_put_user(current->thread.trap_nr, &sc->trapno, Efault);
arch/x86/kernel/signal_32.c
235
unsafe_put_user(current->thread.error_code, &sc->err, Efault);
arch/x86/kernel/signal_32.c
246
unsafe_put_user(current->thread.cr2, &sc->cr2, Efault);
arch/x86/kernel/signal_32.c
283
if (current->mm->context.vdso)
arch/x86/kernel/signal_32.c
284
restorer = current->mm->context.vdso +
arch/x86/kernel/signal_32.c
370
restorer = current->mm->context.vdso +
arch/x86/kernel/signal_64.c
119
unsafe_put_user(current->thread.trap_nr, &sc->trapno, Efault);
arch/x86/kernel/signal_64.c
120
unsafe_put_user(current->thread.error_code, &sc->err, Efault);
arch/x86/kernel/signal_64.c
132
unsafe_put_user(current->thread.cr2, &sc->cr2, Efault);
arch/x86/kernel/signal_64.c
57
current->restart_block.fn = do_no_restart_syscall;
arch/x86/kernel/step.c
196
if (task == current)
arch/x86/kernel/sys_x86_64.c
105
if (current->flags & PF_RANDOMIZE) {
arch/x86/kernel/sys_x86_64.c
130
struct mm_struct *mm = current->mm;
arch/x86/kernel/sys_x86_64.c
172
struct mm_struct *mm = current->mm;
arch/x86/kernel/sys_x86_64.c
37
if (!(current->flags & PF_RANDOMIZE))
arch/x86/kernel/tls.c
106
if (t == &current->thread)
arch/x86/kernel/tls.c
155
if (p == current) {
arch/x86/kernel/tls.c
188
return do_set_thread_area(current, -1, u_info, 1);
arch/x86/kernel/tls.c
24
struct thread_struct *t = &current->thread;
arch/x86/kernel/tls.c
240
return do_get_thread_area(current, -1, u_info);
arch/x86/kernel/traps.c
1074
if (!get_stack_info_noinstr(stack, current, &info) || info.type == STACK_TYPE_ENTRY ||
arch/x86/kernel/traps.c
1322
current->thread.virtual_dr6 = (dr6 & DR_STEP);
arch/x86/kernel/traps.c
1354
dr6 |= current->thread.virtual_dr6;
arch/x86/kernel/traps.c
1425
struct task_struct *task = current;
arch/x86/kernel/traps.c
337
struct task_struct *tsk = current;
arch/x86/kernel/traps.c
599
struct task_struct *tsk = current;
arch/x86/kernel/traps.c
804
struct thread_struct *t = &current->thread;
arch/x86/kernel/traps.c
822
current->comm, task_pid_nr(current), ip);
arch/x86/kernel/traps.c
862
if (!mm_valid_pasid(current->mm))
arch/x86/kernel/traps.c
865
pasid = mm_get_enqcmd_pasid(current->mm);
arch/x86/kernel/traps.c
871
if (current->pasid_activated)
arch/x86/kernel/traps.c
875
current->pasid_activated = 1;
arch/x86/kernel/traps.c
890
current->thread.error_code = error_code;
arch/x86/kernel/traps.c
891
current->thread.trap_nr = trapnr;
arch/x86/kernel/traps.c
907
current->thread.error_code = error_code;
arch/x86/kernel/traps.c
908
current->thread.trap_nr = trapnr;
arch/x86/kernel/traps.c
909
show_signal(current, SIGSEGV, "", str, regs, error_code);
arch/x86/kernel/umip.c
121
struct task_struct *tsk = current;
arch/x86/kernel/umip.c
269
down_read(&current->mm->context.ldt_usr_sem);
arch/x86/kernel/umip.c
270
if (current->mm->context.ldt)
arch/x86/kernel/umip.c
274
up_read(&current->mm->context.ldt_usr_sem);
arch/x86/kernel/umip.c
314
struct task_struct *tsk = current;
arch/x86/kernel/unwind_frame.c
330
if (state->task != current)
arch/x86/kernel/unwind_frame.c
351
if (state->task != current)
arch/x86/kernel/unwind_orc.c
21
if (state->task == current && !state->error) { \
arch/x86/kernel/unwind_orc.c
726
} else if (task == current) {
arch/x86/kernel/uprobes.c
1133
struct mm_struct *mm = current->mm;
arch/x86/kernel/uprobes.c
1242
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
1607
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
1616
utask->autask.saved_trap_nr = current->thread.trap_nr;
arch/x86/kernel/uprobes.c
1617
current->thread.trap_nr = UPROBE_TRAP_NR;
arch/x86/kernel/uprobes.c
1621
if (test_tsk_thread_flag(current, TIF_BLOCKSTEP))
arch/x86/kernel/uprobes.c
1622
set_task_blockstep(current, false);
arch/x86/kernel/uprobes.c
1654
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
1658
WARN_ON_ONCE(current->thread.trap_nr != UPROBE_TRAP_NR);
arch/x86/kernel/uprobes.c
1659
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/x86/kernel/uprobes.c
1681
send_sig(SIGTRAP, current, 0);
arch/x86/kernel/uprobes.c
1727
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
1732
current->thread.trap_nr = utask->autask.saved_trap_nr;
arch/x86/kernel/uprobes.c
1750
send_sig(SIGTRAP, current, 0);
arch/x86/kernel/uprobes.c
1778
current->pid, regs->sp, regs->ip);
arch/x86/kernel/uprobes.c
1809
if (!current->utask)
arch/x86/kernel/uprobes.c
1812
auprobe = current->utask->auprobe;
arch/x86/kernel/uprobes.c
352
struct pt_regs *regs = task_pt_regs(current);
arch/x86/kernel/uprobes.c
375
struct pt_regs *regs = task_pt_regs(current);
arch/x86/kernel/uprobes.c
603
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
614
struct uprobe_task *utask = current->utask;
arch/x86/kernel/uprobes.c
687
struct pt_regs *regs = task_pt_regs(current);
arch/x86/kernel/uprobes.c
688
struct mm_struct *mm = current->mm;
arch/x86/kernel/uprobes.c
716
struct uprobes_state *state = &current->mm->uprobes_state;
arch/x86/kernel/uprobes.c
766
struct vm_area_struct *vma = vma_lookup(current->mm, ip);
arch/x86/kernel/uprobes.c
773
struct mm_struct *mm = current->mm;
arch/x86/kernel/uprobes.c
806
struct pt_regs *regs = task_pt_regs(current);
arch/x86/kernel/vm86_32.c
100
struct vm86 *vm86 = current->thread.vm86;
arch/x86/kernel/vm86_32.c
201
struct task_struct *tsk = current;
arch/x86/kernel/vm86_32.c
229
current->comm, task_pid_nr(current),
arch/x86/kernel/vm86_32.c
250
current->comm);
arch/x86/kernel/vm86_32.c
379
set_flags(VEFLAGS, flags, current->thread.vm86->veflags_mask);
arch/x86/kernel/vm86_32.c
389
set_flags(VFLAGS, flags, current->thread.vm86->veflags_mask);
arch/x86/kernel/vm86_32.c
404
return flags | (VEFLAGS & current->thread.vm86->veflags_mask);
arch/x86/kernel/vm86_32.c
499
struct vm86 *vm86 = current->thread.vm86;
arch/x86/kernel/vm86_32.c
529
struct vm86 *vm86 = current->thread.vm86;
arch/x86/kernel/vm86_32.c
541
current->thread.trap_nr = trapno;
arch/x86/kernel/vm86_32.c
542
current->thread.error_code = error_code;
arch/x86/kernel/vm86_32.c
554
struct vm86plus_info_struct *vmpi = &current->thread.vm86->vm86plus;
arch/x86/kernel/vm86_32.c
784
if (vm86_irqs[irqnumber].tsk != current) return 0;
arch/x86/kernel/vm86_32.c
818
vm86_irqs[irq].tsk = current;
arch/x86/kernel/vm86_32.c
824
if (vm86_irqs[irqnumber].tsk != current) return -EPERM;
arch/x86/kernel/vm86_32.c
87
#define VFLAGS (*(unsigned short *)&(current->thread.vm86->veflags))
arch/x86/kernel/vm86_32.c
88
#define VEFLAGS (current->thread.vm86->veflags)
arch/x86/kernel/vm86_32.c
98
struct task_struct *tsk = current;
arch/x86/kvm/hyperv.c
1514
task_cputime_adjusted(current, &utime, &stime);
arch/x86/kvm/i8254.c
749
pid = get_pid(task_tgid(current));
arch/x86/kvm/mmu/mmu.c
3484
send_sig_mceerr(BUS_MCEERR_AR, (void __user *)hva, PAGE_SHIFT, current);
arch/x86/kvm/mmu/mmu.c
4946
if (signal_pending(current))
arch/x86/kvm/mmu/mmu.c
5082
if (signal_pending(current))
arch/x86/kvm/mmu/page_track.c
226
if (!kvm || kvm->mm != current->mm)
arch/x86/kvm/pmu.c
289
event = perf_event_create_kernel_counter(&attr, -1, current,
arch/x86/kvm/vmx/pmu_intel.c
274
current, NULL, NULL);
arch/x86/kvm/vmx/tdx.c
1942
if (kvm_vcpu_has_events(vcpu) || signal_pending(current))
arch/x86/kvm/vmx/tdx.c
3189
if (signal_pending(current)) {
arch/x86/kvm/vmx/tdx.c
801
if (likely(is_64bit_mm(current->mm)))
arch/x86/kvm/vmx/tdx.c
802
vt->msr_host_kernel_gs_base = current->thread.gsbase;
arch/x86/kvm/vmx/vmx.c
1397
if (likely(is_64bit_mm(current->mm))) {
arch/x86/kvm/vmx/vmx.c
1399
fs_sel = current->thread.fsindex;
arch/x86/kvm/vmx/vmx.c
1400
gs_sel = current->thread.gsindex;
arch/x86/kvm/vmx/vmx.c
1401
fs_base = current->thread.fsbase;
arch/x86/kvm/vmx/vmx.c
1402
vt->msr_host_kernel_gs_base = current->thread.gsbase;
arch/x86/kvm/x86.c
11961
if (signal_pending(current)) {
arch/x86/kvm/x86.c
13385
if (current->mm == kvm->mm) {
arch/x86/kvm/x86.c
3763
if (WARN_ON_ONCE(current->mm != vcpu->kvm->mm))
arch/x86/kvm/x86.c
3830
steal += current->sched_info.run_delay -
arch/x86/kvm/x86.c
3832
vcpu->arch.st.last_steal = current->sched_info.run_delay;
arch/x86/kvm/x86.c
5266
if (unlikely(current->mm != vcpu->kvm->mm))
arch/x86/kvm/xen.c
1056
vcpu->arch.xen.last_steal = current->sched_info.run_delay;
arch/x86/kvm/xen.c
1907
if (current->mm != kvm->mm) {
arch/x86/kvm/xen.c
1912
if (WARN_ON_ONCE(current->mm))
arch/x86/kvm/xen.c
592
u64 run_delay = current->sched_info.run_delay;
arch/x86/kvm/xen.c
728
if (in_atomic() || !task_is_running(current))
arch/x86/lib/insn-eval.c
641
mutex_lock(&current->mm->context.lock);
arch/x86/lib/insn-eval.c
642
ldt = current->mm->context.ldt;
arch/x86/lib/insn-eval.c
648
mutex_unlock(&current->mm->context.lock);
arch/x86/math-emu/fpu_aux.c
56
fpstate_init_soft(&x86_task_fpu(current)->fpstate->regs.soft);
arch/x86/math-emu/fpu_entry.c
232
current->thread.trap_nr = X86_TRAP_MF;
arch/x86/math-emu/fpu_entry.c
233
current->thread.error_code = 0;
arch/x86/math-emu/fpu_entry.c
234
send_sig(SIGFPE, current, 1);
arch/x86/math-emu/fpu_entry.c
625
current->thread.trap_nr = X86_TRAP_MF;
arch/x86/math-emu/fpu_entry.c
626
current->thread.error_code = 0;
arch/x86/math-emu/fpu_entry.c
627
send_sig(signal, current, 1);
arch/x86/math-emu/fpu_system.h
30
mutex_lock(&current->mm->context.lock);
arch/x86/math-emu/fpu_system.h
31
if (current->mm->context.ldt && seg < current->mm->context.ldt->nr_entries)
arch/x86/math-emu/fpu_system.h
32
ret = current->mm->context.ldt->entries[seg];
arch/x86/math-emu/fpu_system.h
33
mutex_unlock(&current->mm->context.lock);
arch/x86/math-emu/fpu_system.h
76
#define I387 (&x86_task_fpu(current)->fpstate->regs)
arch/x86/mm/debug_pagetables.c
18
if (current->mm->pgd)
arch/x86/mm/debug_pagetables.c
19
ptdump_walk_pgd_level_debugfs(m, current->mm, false);
arch/x86/mm/debug_pagetables.c
28
if (current->mm->pgd)
arch/x86/mm/debug_pagetables.c
29
ptdump_walk_pgd_level_debugfs(m, current->mm, true);
arch/x86/mm/fault.c
1217
tsk = current;
arch/x86/mm/fault.c
140
instr = (void *)convert_ip_to_linear(current, regs);
arch/x86/mm/fault.c
1418
if (fatal_signal_pending(current) && !user_mode(regs)) {
arch/x86/mm/fault.c
599
tsk = current;
arch/x86/mm/fault.c
631
struct task_struct *tsk = current;
arch/x86/mm/fault.c
705
if (task_stack_end_corrupted(current))
arch/x86/mm/fault.c
780
struct task_struct *tsk = current;
arch/x86/mm/fault.c
930
struct task_struct *tsk = current;
arch/x86/mm/mmap.c
148
struct mm_struct *mm = current->mm;
arch/x86/mm/mmap.c
44
if (current->flags & PF_RANDOMIZE) {
arch/x86/mm/mmap.c
64
if (current->personality & ADDR_COMPAT_LAYOUT)
arch/x86/mm/mmap.c
72
if (!(current->flags & PF_RANDOMIZE))
arch/x86/mm/pat/memtype.c
634
current->comm, current->pid, start, end - 1);
arch/x86/mm/pat/memtype.c
826
current->comm, current->pid,
arch/x86/mm/pat/memtype.c
860
current->comm, current->pid,
arch/x86/mm/pat/memtype.c
878
current->comm, current->pid,
arch/x86/mm/pat/memtype_interval.c
84
current->comm, current->pid, start, end,
arch/x86/mm/pkeys.c
45
ret = arch_set_user_pkey_access(current, execute_only_pkey,
arch/x86/mm/tlb.c
1029
switch_mm_irqs_off(NULL, temp_mm, current);
arch/x86/mm/tlb.c
1056
switch_mm_irqs_off(NULL, prev_mm, current);
arch/x86/mm/tlb.c
1770
struct mm_struct *current_mm = current->mm;
arch/x86/mm/tlb.c
452
if ((current->pid & 0x1f) != (jiffies & 0x1f))
arch/x86/net/bpf_jit_comp.c
4013
for (unwind_start(&state, current, NULL, NULL); !unwind_done(&state);
arch/x86/platform/uv/uv_nmi.c
716
cpu, current->pid, current->comm, (void *)regs->ip);
arch/x86/platform/uv/uv_nmi.c
733
if (current->pid != 0 || uv_nmi_action != nmi_act_ips)
arch/x86/power/cpu.c
180
load_mm_ldt(current->active_mm); /* This does lldt */
arch/x86/um/asm/processor.h
20
(address + 65536 + 32 * sizeof(unsigned long) >= UPT_SP(&current->thread.regs.regs))
arch/x86/um/shared/sysdep/faultinfo_32.h
41
"=m" (current->thread.segv_continue) :: \
arch/x86/um/shared/sysdep/faultinfo_64.h
41
"=m" (current->thread.segv_continue) :: \
arch/x86/um/signal.c
115
struct faultinfo * fi = &current->thread.arch.faultinfo;
arch/x86/um/signal.c
165
err = copy_regset_to_user(current,
arch/x86/um/signal.c
166
task_user_regset_view(current),
arch/x86/um/signal.c
222
err |= copy_sc_to_user(&uc->uc_mcontext, fp, &current->thread.regs, 0);
arch/x86/um/signal.c
341
unsigned long sp = PT_REGS_SP(&current->thread.regs);
arch/x86/um/signal.c
353
if (copy_sc_from_user(&current->thread.regs, sc))
arch/x86/um/signal.c
357
PT_REGS_SYSCALL_NR(&current->thread.regs) = -1;
arch/x86/um/signal.c
358
return PT_REGS_SYSCALL_RET(&current->thread.regs);
arch/x86/um/signal.c
41
current->restart_block.fn = do_no_restart_syscall;
arch/x86/um/signal.c
449
unsigned long sp = PT_REGS_SP(&current->thread.regs);
arch/x86/um/signal.c
460
if (copy_sc_from_user(&current->thread.regs, &uc->uc_mcontext))
arch/x86/um/signal.c
464
PT_REGS_SYSCALL_NR(&current->thread.regs) = -1;
arch/x86/um/signal.c
465
return PT_REGS_SYSCALL_RET(&current->thread.regs);
arch/x86/um/signal.c
97
err = copy_regset_from_user(current,
arch/x86/um/signal.c
98
task_user_regset_view(current),
arch/x86/um/syscalls_64.c
23
current->thread.regs.regs.gp[FS_BASE / sizeof(unsigned long)] =
arch/x86/um/syscalls_64.c
28
current->thread.regs.regs.gp[GS_BASE / sizeof(unsigned long)] =
arch/x86/um/syscalls_64.c
33
ret = put_user(current->thread.regs.regs.gp[FS_BASE / sizeof(unsigned long)], arg2);
arch/x86/um/syscalls_64.c
36
ret = put_user(current->thread.regs.regs.gp[GS_BASE / sizeof(unsigned long)], arg2);
arch/x86/um/syscalls_64.c
45
return arch_prctl(current, option, (unsigned long __user *) arg2);
arch/x86/um/sysrq_64.c
20
printk(KERN_INFO "Pid: %d, comm: %.20s %s %s\n", task_pid_nr(current),
arch/x86/um/sysrq_64.c
21
current->comm, print_tainted(), init_utsname()->release);
arch/x86/um/tls_32.c
112
ret = do_set_thread_area(current, &curr->tls);
arch/x86/um/tls_32.c
249
if (unlikely(task == current &&
arch/x86/um/tls_32.c
252
"without flushed TLS.", current->pid);
arch/x86/um/tls_32.c
281
idx = get_free_idx(current);
arch/x86/um/tls_32.c
290
ret = do_set_thread_area(current, &info);
arch/x86/um/tls_32.c
293
return set_tls_entry(current, &info, idx, 1);
arch/x86/um/tls_32.c
326
ret = get_tls_entry(current, &info, idx);
arch/x86/um/vdso/vma.c
38
struct mm_struct *mm = current->mm;
arch/x86/xen/smp_pv.c
355
xen_cpu_bringup_again((unsigned long)task_pt_regs(current));
arch/xtensa/include/asm/elf.h
194
set_personality(PER_LINUX_32BIT | (current->personality & (~PER_MASK)))
arch/xtensa/include/asm/stackprotector.h
29
current->stack_canary = canary;
arch/xtensa/include/asm/stackprotector.h
30
__stack_chk_guard = current->stack_canary;
arch/xtensa/include/asm/stacktrace.h
24
if (!task || task == current)
arch/xtensa/kernel/process.c
210
flush_ptrace_hw_breakpoint(current);
arch/xtensa/kernel/ptrace.c
373
if (current->thread.ptrace_bp[i] == bp)
arch/xtensa/kernel/ptrace.c
378
if (current->thread.ptrace_wp[i] == bp)
arch/xtensa/kernel/ptrace.c
563
trace_sys_enter(regs, syscall_get_nr(current, regs));
arch/xtensa/kernel/signal.c
256
current->restart_block.fn = do_no_restart_syscall;
arch/xtensa/kernel/signal.c
352
(current->personality & FDPIC_FUNCPTRS);
arch/xtensa/kernel/signal.c
365
sp = current->sas_ss_sp + current->sas_ss_size;
arch/xtensa/kernel/signal.c
439
current->comm, current->pid, sig, frame, regs->pc);
arch/xtensa/kernel/signal.c
457
task_pt_regs(current)->icountlevel = 0;
arch/xtensa/kernel/signal.c
497
task_pt_regs(current)->icountlevel = 1;
arch/xtensa/kernel/signal.c
523
task_pt_regs(current)->icountlevel = 1;
arch/xtensa/kernel/smp.c
146
current->active_mm = mm;
arch/xtensa/kernel/smp.c
148
enter_lazy_tlb(mm, current);
arch/xtensa/kernel/stacktrace.c
236
save_stack_trace_tsk(current, trace);
arch/xtensa/kernel/syscall.c
84
vma_iter_init(&vmi, current->mm, addr);
arch/xtensa/kernel/traps.c
210
current->comm, task_pid_nr(current), regs->pc,
arch/xtensa/kernel/traps.c
366
current->comm, task_pid_nr(current), regs->pc);
arch/xtensa/kernel/traps.c
383
regs->excvaddr, current->comm,
arch/xtensa/kernel/traps.c
384
task_pid_nr(current), regs->pc);
arch/xtensa/kernel/traps.c
403
regs->excvaddr, current->comm,
arch/xtensa/kernel/traps.c
404
task_pid_nr(current), regs->pc);
arch/xtensa/mm/fault.c
124
current->comm, current->pid,
arch/xtensa/mm/fault.c
244
current->comm, regs->pc, entry->fixup);
arch/xtensa/mm/fault.c
33
struct mm_struct *act_mm = current->active_mm;
arch/xtensa/mm/fault.c
91
struct mm_struct *mm = current->mm;
arch/xtensa/mm/kasan_init.c
96
current->kasan_depth = 0;
arch/xtensa/mm/tlb.c
68
if (mm == current->active_mm) {
block/bfq-iosched.c
457
if (!current->io_context)
block/bfq-iosched.c
5508
struct task_struct *tsk = current;
block/bfq-iosched.c
5854
bfq_init_bfqq(bfqd, bfqq, bic, current->pid,
block/bfq-iosched.c
6722
bfqq->pid = current->pid;
block/bio.c
454
while ((bio = bio_list_pop(&current->bio_list[0])))
block/bio.c
456
current->bio_list[0] = nopunt;
block/bio.c
459
while ((bio = bio_list_pop(&current->bio_list[1])))
block/bio.c
461
current->bio_list[1] = nopunt;
block/bio.c
592
if (current->bio_list &&
block/bio.c
593
(!bio_list_empty(&current->bio_list[0]) ||
block/bio.c
594
!bio_list_empty(&current->bio_list[1])) &&
block/blk-cgroup.c
110
return task_css(current, io_cgrp_id);
block/blk-cgroup.c
1978
} while (!fatal_signal_pending(current));
block/blk-cgroup.c
1997
struct gendisk *disk = current->throttle_disk;
block/blk-cgroup.c
2000
bool use_memdelay = current->use_memdelay;
block/blk-cgroup.c
2005
current->throttle_disk = NULL;
block/blk-cgroup.c
2006
current->use_memdelay = false;
block/blk-cgroup.c
2046
if (unlikely(current->flags & PF_KTHREAD))
block/blk-cgroup.c
2049
if (current->throttle_disk != disk) {
block/blk-cgroup.c
2054
if (current->throttle_disk)
block/blk-cgroup.c
2055
put_disk(current->throttle_disk);
block/blk-cgroup.c
2056
current->throttle_disk = disk;
block/blk-cgroup.c
2060
current->use_memdelay = use_memdelay;
block/blk-cgroup.c
2061
set_notify_resume(current);
block/blk-core.c
1131
struct task_struct *tsk = current;
block/blk-core.c
1204
struct blk_plug *plug = current->plug;
block/blk-core.c
1241
current->flags &= ~PF_BLOCK_TS;
block/blk-core.c
1256
if (plug == current->plug) {
block/blk-core.c
1258
current->plug = NULL;
block/blk-core.c
290
bool freeze = __blk_freeze_queue_start(q, current);
block/blk-core.c
568
current->comm, bio->bi_bdev, bio->bi_opf,
block/blk-core.c
678
current->bio_list = bio_list_on_stack;
block/blk-core.c
712
current->bio_list = NULL;
block/blk-core.c
719
current->bio_list = bio_list;
block/blk-core.c
725
current->bio_list = NULL;
block/blk-core.c
747
if (current->bio_list) {
block/blk-core.c
749
bio_list_add_head(&current->bio_list[0], bio);
block/blk-core.c
751
bio_list_add(&current->bio_list[0], bio);
block/blk-core.c
957
blk_flush_plug(current->plug, false);
block/blk-ioc.c
291
struct io_context *ioc = current->io_context;
block/blk-ioc.c
320
struct io_context *ioc = current->io_context;
block/blk-ioc.c
357
struct io_context *ioc = current->io_context;
block/blk-ioc.c
401
struct io_context *ioc = current->io_context;
block/blk-ioc.c
409
task_lock(current);
block/blk-ioc.c
410
if (current->io_context) {
block/blk-ioc.c
412
ioc = current->io_context;
block/blk-ioc.c
414
current->io_context = ioc;
block/blk-ioc.c
418
task_unlock(current);
block/blk-iocost.c
2651
use_debt = bio_issue_as_root_blkg(bio) || fatal_signal_pending(current);
block/blk-iolatency.c
304
if (issue_as_root || fatal_signal_pending(current)) {
block/blk-lib.c
129
fatal_signal_pending(current))
block/blk-lib.c
159
fatal_signal_pending(current)) {
block/blk-lib.c
204
fatal_signal_pending(current))
block/blk-lib.c
240
fatal_signal_pending(current)) {
block/blk-map.c
139
if (!current->mm)
block/blk-merge.c
1088
struct blk_plug *plug = current->plug;
block/blk-mq.c
140
if (q->mq_freeze_owner != current)
block/blk-mq.c
1455
if (current->plug && !at_head) {
block/blk-mq.c
1456
blk_add_rq_to_plug(current->plug, rq);
block/blk-mq.c
184
if (__blk_freeze_queue_start(q, current))
block/blk-mq.c
3144
struct blk_plug *plug = current->plug;
block/blk-mq.c
5235
if (task_sigpending(current))
block/blk-mq.c
630
struct blk_plug *plug = current->plug;
block/blk.h
675
struct blk_plug *plug = current->plug;
block/blk.h
687
current->flags |= PF_BLOCK_TS;
block/bsg.c
154
current->comm);
block/fops.c
199
dio->waiter = current;
block/ioctl.c
157
if (fatal_signal_pending(current)) {
block/ioprio.c
193
p = current;
block/ioprio.c
201
pgrp = task_pgrp(current);
block/ioprio.c
82
p = current;
block/ioprio.c
90
pgrp = task_pgrp(current);
crypto/af_alg.c
793
if (signal_pending(current))
crypto/af_alg.c
852
if (signal_pending(current))
crypto/af_alg.c
960
current->comm);
crypto/algboss.c
62
} while (err == -EAGAIN && !signal_pending(current));
crypto/api.c
493
if (fatal_signal_pending(current)) {
crypto/api.c
653
if (fatal_signal_pending(current)) {
crypto/arc4.c
44
current->comm, (unsigned long)current->pid);
crypto/jitterentropy-testing.c
150
if (signal_pending(current)) {
crypto/jitterentropy-testing.c
208
if (signal_pending(current)) {
drivers/accel/amdxdna/amdxdna_ctx.c
504
job->mm = current->mm;
drivers/accel/amdxdna/amdxdna_gem.c
218
current->mm,
drivers/accel/amdxdna/amdxdna_pci_drv.c
76
client->sva = iommu_sva_bind_device(xdna->ddev.dev, current->mm);
drivers/accel/amdxdna/amdxdna_pci_drv.c
88
client->mm = current->mm;
drivers/accel/amdxdna/amdxdna_ubuf.c
156
ubuf->mm = current->mm;
drivers/accel/habanalabs/common/context.c
273
current->comm, ctx->asid);
drivers/accel/habanalabs/common/habanalabs_drv.c
197
hpriv->taskpid = get_task_pid(current, PIDTYPE_PID);
drivers/accel/habanalabs/common/habanalabs_drv.c
306
hpriv->taskpid = get_task_pid(current, PIDTYPE_PID);
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1289
task_pid_nr(current), current->comm, cmd, nr);
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1314
task_pid_nr(current), current->comm, cmd, nr);
drivers/accel/habanalabs/common/memory.c
2397
userptr->pid = current->pid;
drivers/accel/ivpu/ivpu_drv.c
278
ctx_id, current->comm, task_pid_nr(current));
drivers/accel/ivpu/ivpu_drv.c
298
file_priv->ctx.id, current->comm, task_pid_nr(current));
drivers/accel/ivpu/ivpu_job.c
203
task_pid_nr(current), engine,
drivers/accel/qaic/qaic_control.c
1046
if (signal_pending(current))
drivers/accessibility/speakup/speakup_soft.c
248
if (signal_pending(current)) {
drivers/acpi/acpi_dbg.c
86
current != acpi_aml_io.thread)
drivers/acpi/acpi_pad.c
129
set_cpus_allowed_ptr(current, cpumask_of(preferred_cpu));
drivers/acpi/acpi_pad.c
152
sched_set_fifo_low(current);
drivers/acpi/apei/ghes.c
501
twcb->pfn, current->comm, task_pid_nr(current));
drivers/acpi/apei/ghes.c
515
if (flags == MF_ACTION_REQUIRED && current->mm) {
drivers/acpi/apei/ghes.c
523
task_work_add(current, &twcb->twork, TWA_RESUME);
drivers/acpi/apei/ghes.c
960
current->comm, task_pid_nr(current));
drivers/android/binder.c
2020
filp_close(twcb->file, current->files);
drivers/android/binder.c
2021
task_work_add(current, &twcb->twork, TWA_RESUME);
drivers/android/binder.c
3123
t->priority = task_nice(current);
drivers/android/binder.c
4662
if (signal_pending(current)) {
drivers/android/binder.c
493
thread->pid == current->pid && !thread->process_todo)
drivers/android/binder.c
5038
t->saved_priority = task_nice(current);
drivers/android/binder.c
5061
task_active_pid_ns(current));
drivers/android/binder.c
5259
if (current->pid < thread->pid)
drivers/android/binder.c
5261
else if (current->pid > thread->pid)
drivers/android/binder.c
5271
thread->pid = current->pid;
drivers/android/binder.c
5985
pr_info("%d:%d ioctl %x %lx returned %d\n", proc->pid, current->pid, cmd, arg, ret);
drivers/android/binder.c
6030
if (!same_thread_group(proc->tsk, current))
drivers/android/binder.c
6061
current->tgid, current->pid);
drivers/android/binder.c
6070
proc->tsk = get_task_struct(current->group_leader);
drivers/android/binder.c
6071
proc->pid = current->tgid;
drivers/android/binder.c
6075
proc->default_priority = task_nice(current);
drivers/android/binder.c
684
if (can_nice(current, nice)) {
drivers/android/binder.c
685
set_user_nice(current, nice);
drivers/android/binder.c
691
current->pid, nice, min_nice);
drivers/android/binder.c
692
set_user_nice(current, min_nice);
drivers/android/binder.c
695
binder_user_error("%d RLIMIT_NICE not set\n", current->pid);
drivers/android/binder.c
788
proc->pid, current->pid, node->debug_id,
drivers/android/binder/rust_binderfs.c
650
info->ipc_ns = get_ipc_ns(current->nsproxy->ipc_ns);
drivers/android/binder_alloc.c
1235
alloc->pid = current->tgid;
drivers/android/binder_alloc.c
1236
alloc->mm = current->mm;
drivers/android/binder_alloc.c
473
int pid = current->tgid;
drivers/android/binder_alloc.c
689
buffer->pid = current->tgid;
drivers/android/binderfs.c
646
info->ipc_ns = get_ipc_ns(current->nsproxy->ipc_ns);
drivers/ata/libata-core.c
6825
bool owns_eh = ap && ap->host->eh_owner == current;
drivers/ata/libata-eh.c
475
ap->host->eh_owner = current;
drivers/ata/libata-eh.c
490
WARN_ON_ONCE(ap->host->eh_owner != current);
drivers/atm/atmtcp.c
46
DECLARE_WAITQUEUE(wait,current);
drivers/atm/eni.c
1392
DECLARE_WAITQUEUE(wait,current);
drivers/atm/eni.c
837
DECLARE_WAITQUEUE(wait,current);
drivers/atm/he.c
2329
DECLARE_WAITQUEUE(wait, current);
drivers/atm/solos-pci.c
202
"L%05d\n%s\n", current->pid, attr->attr.name);
drivers/atm/solos-pci.c
210
prm.pid = current->pid;
drivers/atm/solos-pci.c
258
"L%05d\n%s\n%s\n", current->pid, attr->attr.name, buf);
drivers/atm/solos-pci.c
266
prm.pid = current->pid;
drivers/base/power/main.c
226
task_pid_nr(current),
drivers/base/power/main.c
573
wd->tsk = current;
drivers/base/power/wakeup.c
973
if (inpr == 0 || signal_pending(current))
drivers/block/aoe/aoecmd.c
1232
DECLARE_WAITQUEUE(wait, current);
drivers/block/aoe/aoecmd.c
1236
current->flags |= PF_NOFREEZE;
drivers/block/aoe/aoecmd.c
1237
set_user_nice(current, -10);
drivers/block/drbd/drbd_actlog.c
186
current->comm, current->pid, __func__,
drivers/block/drbd/drbd_actlog.c
193
current->comm, current->pid, __func__,
drivers/block/drbd/drbd_bitmap.c
119
current->comm, task_pid_nr(current),
drivers/block/drbd/drbd_bitmap.c
138
current->comm, task_pid_nr(current),
drivers/block/drbd/drbd_bitmap.c
148
b->bm_task = current;
drivers/block/drbd/drbd_main.c
1467
current->comm, current->pid, connection->ko_count);
drivers/block/drbd/drbd_main.c
1833
flush_signals(current);
drivers/block/drbd/drbd_main.c
307
snprintf(current->comm, sizeof(current->comm), "drbd_%c_%s",
drivers/block/drbd/drbd_main.c
341
drbd_info(resource, "Terminating %s\n", current->comm);
drivers/block/drbd/drbd_main.c
3482
D_ASSERT(device, current == peer_device->connection->worker.task);
drivers/block/drbd/drbd_main.c
3529
D_ASSERT(device, current != first_peer_device(device)->connection->worker.task);
drivers/block/drbd/drbd_main.c
3675
if (signal_pending(current))
drivers/block/drbd/drbd_main.c
377
thi->name, current->comm, current->pid);
drivers/block/drbd/drbd_main.c
394
flush_signals(current); /* otherw. may get -ERESTARTNOINTR */
drivers/block/drbd/drbd_main.c
417
thi->name, current->comm, current->pid);
drivers/block/drbd/drbd_main.c
455
if (thi->task != current)
drivers/block/drbd/drbd_main.c
513
struct task_struct *p = current;
drivers/block/drbd/drbd_nl.c
368
if (current == connection->worker.task)
drivers/block/drbd/drbd_nl.c
397
if (current == connection->worker.task)
drivers/block/drbd/drbd_nl.c
606
flush_signals(current);
drivers/block/drbd/drbd_receiver.c
1008
if (current->plug == &connection->receiver_plug) {
drivers/block/drbd/drbd_receiver.c
2181
if (signal_pending(current)) {
drivers/block/drbd/drbd_receiver.c
376
if (err && !signal_pending(current))
drivers/block/drbd/drbd_receiver.c
5780
sched_set_fifo_low(current);
drivers/block/drbd/drbd_receiver.c
5840
flush_signals(current);
drivers/block/drbd/drbd_receiver.c
825
if (signal_pending(current)) {
drivers/block/drbd/drbd_receiver.c
826
flush_signals(current);
drivers/block/drbd/drbd_state.c
1529
D_ASSERT(device, current == first_peer_device(device)->connection->worker.task);
drivers/block/drbd/drbd_state.c
625
D_ASSERT(device, current != first_peer_device(device)->connection->worker.task);
drivers/block/drbd/drbd_worker.c
1768
if (current == connection->worker.task) {
drivers/block/drbd/drbd_worker.c
2102
if (!list_empty(work_list) || signal_pending(current)) {
drivers/block/drbd/drbd_worker.c
2173
if (signal_pending(current)) {
drivers/block/drbd/drbd_worker.c
2174
flush_signals(current);
drivers/block/loop.c
1949
int orig_flags = current->flags;
drivers/block/loop.c
1952
current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
drivers/block/loop.c
1977
current->flags = orig_flags;
drivers/block/nbd.c
1266
nbd->task_setup = current;
drivers/block/nbd.c
1269
(nbd->task_setup != current ||
drivers/block/nbd.c
1498
nbd->pid = task_pid_nr(current);
drivers/block/swim3.c
1053
if (signal_pending(current))
drivers/block/swim3.c
867
if (signal_pending(current)) {
drivers/block/swim3.c
948
if (signal_pending(current)) {
drivers/block/ublk_drv.c
1710
if (unlikely(current != io->task || current->flags & PF_EXITING)) {
drivers/block/ublk_drv.c
2317
ub->ublksrv_tgid = current->tgid;
drivers/block/ublk_drv.c
2554
ub->mm = current->mm;
drivers/block/ublk_drv.c
2555
if (current->mm != ub->mm)
drivers/block/ublk_drv.c
2571
__func__, q_id, current->pid, vma->vm_start,
drivers/block/ublk_drv.c
3072
if (current == io->task && io->task_registered_buffers)
drivers/block/ublk_drv.c
3179
WRITE_ONCE(io->task, get_task_struct(current));
drivers/block/ublk_drv.c
3312
if (READ_ONCE(io->task) != current) {
drivers/block/ublk_drv.c
3972
on_daemon = current == READ_ONCE(io->task);
drivers/block/ublk_drv.c
5068
while (elapsed < timeout_ms && !signal_pending(current)) {
drivers/block/ublk_drv.c
5089
if (signal_pending(current))
drivers/block/xen-blkback/blkback.c
543
current->comm, ring->st_oo_req,
drivers/block/zloop.c
610
int orig_flags = current->flags;
drivers/block/zloop.c
612
current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
drivers/block/zloop.c
614
current->flags = orig_flags;
drivers/bluetooth/btmrvl_main.c
592
init_waitqueue_entry(&wait, current);
drivers/bus/vexpress-config.c
195
if (signal_pending(current))
drivers/cdrom/cdrom.c
1094
(unsigned int)task_pid_nr(current));
drivers/char/agp/generic.c
518
printk(KERN_INFO PFX "%s tried to set rate=x0. Setting to x1 mode.\n", current->comm);
drivers/char/agp/generic.c
596
printk(KERN_INFO PFX "%s tried to set rate=x0. Setting to AGP3 x4 mode.\n", current->comm);
drivers/char/agp/generic.c
600
printk(KERN_INFO PFX "%s tried to set rate=x%d. Setting to AGP3 x8 mode.\n", current->comm, tmp * 4);
drivers/char/agp/generic.c
630
current->comm, *requested_mode);
drivers/char/agp/generic.c
641
current->comm, *requested_mode);
drivers/char/agp/generic.c
650
printk(KERN_INFO PFX "%s requested AGPx8 but bridge not capable.\n", current->comm);
drivers/char/agp/generic.c
656
printk(KERN_INFO PFX "%s requested AGPx8 but graphic card not capable.\n", current->comm);
drivers/char/applicom.c
366
DECLARE_WAITQUEUE(wait, current);
drivers/char/applicom.c
445
if (signal_pending(current)) {
drivers/char/applicom.c
543
DECLARE_WAITQUEUE(wait, current);
drivers/char/applicom.c
610
if (signal_pending(current))
drivers/char/hpet.c
264
DECLARE_WAITQUEUE(wait, current);
drivers/char/hpet.c
296
} else if (signal_pending(current)) {
drivers/char/hw_random/core.c
305
if (signal_pending(current)) {
drivers/char/hw_random/s390-trng.c
77
if (signal_pending(current)) {
drivers/char/ipmi/ipmi_si_intf.c
1011
set_user_nice(current, MAX_NICE);
drivers/char/ipmi/ipmi_watchdog.c
765
init_waitqueue_entry(&wait, current);
drivers/char/ipmi/ipmi_watchdog.c
767
while (!data_to_read && !signal_pending(current)) {
drivers/char/ipmi/ipmi_watchdog.c
775
if (signal_pending(current)) {
drivers/char/lp.c
293
if (signal_pending(current)) {
drivers/char/lp.c
354
if (signal_pending(current)) {
drivers/char/lp.c
477
if (signal_pending(current)) {
drivers/char/mem.c
469
if (signal_pending(current))
drivers/char/mem.c
499
if (signal_pending(current))
drivers/char/mem.c
75
return signal_pending(current);
drivers/char/ppdev.c
175
if (signal_pending(current)) {
drivers/char/ppdev.c
257
if (signal_pending(current))
drivers/char/random.c
1306
while (!crng_ready() && !signal_pending(current)) {
drivers/char/random.c
1437
if (signal_pending(current))
drivers/char/random.c
1469
current->comm, iov_iter_count(iter));
drivers/char/random.c
472
if (signal_pending(current))
drivers/char/tpm/st33zp24/st33zp24.c
229
if (ret == -ERESTARTSYS && freezing(current))
drivers/char/tpm/st33zp24/st33zp24.c
247
} while (ret == -ERESTARTSYS && freezing(current));
drivers/char/tpm/tpm-dev-common.c
97
task_tgid_nr(current));
drivers/char/tpm/tpm_tis_core.c
103
if (rc == -ERESTARTSYS && freezing(current)) {
drivers/char/tpm/tpm_tis_core.c
216
if (rc == -ERESTARTSYS && freezing(current)) {
drivers/char/tpm/xen-tpmfront.c
85
if (rc == -ERESTARTSYS && freezing(current)) {
drivers/clk/clk.c
224
if (prepare_owner == current) {
drivers/clk/clk.c
232
prepare_owner = current;
drivers/clk/clk.c
238
WARN_ON_ONCE(prepare_owner != current);
drivers/clk/clk.c
259
if (enable_owner == current) {
drivers/clk/clk.c
270
enable_owner = current;
drivers/clk/clk.c
278
WARN_ON_ONCE(enable_owner != current);
drivers/comedi/comedi_fops.c
2714
DECLARE_WAITQUEUE(wait, current);
drivers/comedi/comedi_fops.c
2772
if (signal_pending(current)) {
drivers/comedi/comedi_fops.c
2842
DECLARE_WAITQUEUE(wait, current);
drivers/comedi/comedi_fops.c
2897
if (signal_pending(current)) {
drivers/connector/cn_proc.c
250
ev->event_data.ptrace.tracer_pid = current->pid;
drivers/connector/cn_proc.c
251
ev->event_data.ptrace.tracer_tgid = current->tgid;
drivers/connector/cn_proc.c
410
!task_is_in_init_pid_ns(current))
drivers/cpufreq/cpufreq.c
391
&& current == policy->transition_task);
drivers/cpufreq/cpufreq.c
404
policy->transition_task = current;
drivers/cpufreq/pmac32-cpufreq.c
300
switch_mmu_context(NULL, current->active_mm, NULL);
drivers/crypto/amcc/crypto4xx_core.c
625
static u32 get_next_gd(u32 current)
drivers/crypto/amcc/crypto4xx_core.c
627
if (current != PPC4XX_LAST_GD)
drivers/crypto/amcc/crypto4xx_core.c
628
return current + 1;
drivers/crypto/amcc/crypto4xx_core.c
633
static u32 get_next_sd(u32 current)
drivers/crypto/amcc/crypto4xx_core.c
635
if (current != PPC4XX_LAST_SD)
drivers/crypto/amcc/crypto4xx_core.c
636
return current + 1;
drivers/crypto/nx/nx.c
58
atomic_set(&(nx_ctx->stats->last_error_pid), current->pid);
drivers/dax/device.c
247
dev_dbg(&dev_dax->dev, "%s: op=%s addr=%#lx order=%d\n", current->comm,
drivers/dax/device.c
30
current->comm, func);
drivers/dax/device.c
38
current->comm, func, start, end,
drivers/dax/device.c
46
current->comm, func);
drivers/dma-buf/dma-fence.c
811
if (intr && signal_pending(current)) {
drivers/dma-buf/dma-fence.c
822
cb.task = current;
drivers/dma-buf/dma-fence.c
835
if (ret > 0 && intr && signal_pending(current))
drivers/dma-buf/dma-fence.c
917
cb[i].task = current;
drivers/dma-buf/dma-fence.c
938
if (ret > 0 && intr && signal_pending(current))
drivers/dma-buf/heaps/cma_heap.c
340
if (fatal_signal_pending(current))
drivers/dma-buf/heaps/system_heap.c
373
if (fatal_signal_pending(current)) {
drivers/dma-buf/selftest.c
97
if (signal_pending(current))
drivers/dma-buf/sw_sync.c
315
get_task_comm(task_comm, current);
drivers/dma/dmatest.c
382
const char *thread_name = current->comm;
drivers/dma/dmatest.c
428
current->comm, error_count - MAX_ERROR_COUNT);
drivers/dma/dmatest.c
466
current->comm, n, err, src_off, dst_off, len, data);
drivers/dma/dmatest.c
469
current->comm, n, err, src_off, dst_off, len, data);
drivers/dma/dmatest.c
478
current->comm, n, err, src_off, dst_off, len, data);
drivers/dma/dmatest.c
678
set_user_nice(current, 10);
drivers/dma/dmatest.c
876
pr_debug("%s: verifying source buffer...\n", current->comm);
drivers/dma/dmatest.c
886
pr_debug("%s: verifying dest buffer...\n", current->comm);
drivers/dma/dmatest.c
934
current->comm, total_tests, failed_tests,
drivers/dma/idxd/cdev.c
244
ctx->pid = current->pid;
drivers/dma/idxd/cdev.c
247
sva = iommu_sva_bind_device(dev, current->mm);
drivers/dma/idxd/cdev.c
262
ctx->mm = current->mm;
drivers/dma/idxd/cdev.c
377
current->comm, func,
drivers/dma/idxd/cdev.c
408
if (current->mm != ctx->mm)
drivers/dma/idxd/cdev.c
479
if (current->mm != ctx->mm)
drivers/dma/idxd/cdev.c
502
if (current->mm != ctx->mm)
drivers/dma/idxd/submit.c
50
if (signal_pending_state(TASK_INTERRUPTIBLE, current))
drivers/dpll/zl3073x/flash.c
149
if (signal_pending(current))
drivers/dpll/zl3073x/flash.c
67
if (signal_pending(current)) {
drivers/firmware/efi/arm-runtime.c
163
efi_set_pgd(current->active_mm);
drivers/firmware/efi/riscv-runtime.c
128
switch_mm(current->active_mm, &efi_mm, NULL);
drivers/firmware/efi/riscv-runtime.c
133
switch_mm(&efi_mm, current->active_mm, NULL);
drivers/firmware/efi/runtime-wrappers.c
224
efi_runtime_lock_owner = current;
drivers/firmware/efi/runtime-wrappers.c
452
efi_runtime_lock_owner = current;
drivers/firmware/efi/runtime-wrappers.c
491
efi_runtime_lock_owner = current;
drivers/firmware/efi/runtime-wrappers.c
521
efi_runtime_lock_owner = current;
drivers/firmware/efi/runtime-wrappers.c
603
WARN_ON(efi_runtime_lock_owner != current);
drivers/firmware/psci/psci_checker.c
285
sched_set_fifo(current);
drivers/fpga/dfl-afu-dma-region.c
41
ret = account_locked_vm(current->mm, npages, true);
drivers/fpga/dfl-afu-dma-region.c
70
account_locked_vm(current->mm, npages, false);
drivers/fpga/dfl-afu-dma-region.c
90
account_locked_vm(current->mm, npages, false);
drivers/fwctl/main.c
109
current->comm, task_pid_nr(current));
drivers/fwctl/mlx5/main.c
15
dev_dbg(&mcdev->fwctl.dev, "PID %u: " format, current->pid, \
drivers/gpib/common/gpib_os.c
1714
board->locking_pid = current->pid;
drivers/gpib/common/gpib_os.c
1722
if (current->pid != board->locking_pid) {
drivers/gpib/common/gpib_os.c
1724
current->pid, board->locking_pid);
drivers/gpib/common/gpib_os.c
731
if (current->pid != board->locking_pid) {
drivers/gpib/lpvo_usb_gpib/lpvo_usb_gpib.c
667
if (item->private == current) {
drivers/gpio/gpio-sloppy-logic-analyzer.c
62
if (ret == 0 && fatal_signal_pending(current))
drivers/gpu/drm/adp/adp_drv.c
72
if (current->comm[0] == 'X')
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
289
if ((mmptr) == current->mm) { \
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
291
} else if (current->flags & PF_KTHREAD) { \
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_fence.c
77
get_task_comm(fence->timeline_name, current);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1425
current->mm,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gpuvm.c
1433
info->pid = get_task_pid(current, PIDTYPE_TGID);
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
106
if (usermm != current->mm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
941
if (usermm && usermm != current->mm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
169
get_task_comm(ev_fence->timeline_name, current);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
239
if (mm && mm != current->mm)
drivers/gpu/drm/amd/amdgpu/amdgpu_hmm.c
135
r = mmu_interval_notifier_insert(&bo->notifier, current->mm,
drivers/gpu/drm/amd/amdgpu/amdgpu_hmm.c
139
r = mmu_interval_notifier_insert(&bo->notifier, current->mm, addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
260
((blk_info.task.pid != task_pid_nr(current)) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
261
strncmp(blk_info.task.comm, current->comm, TASK_COMM_LEN)))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4396
con->init_task_pid = task_pid_nr(current);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4397
get_task_comm(con->init_task_comm, current);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1321
gtt->usertask = current->group_leader;
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
1482
if (amdkfd_fence_check_mm(f, current->mm) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
105
get_task_comm(fence_drv->timeline_name, current);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2557
if (vm->task_info->task.pid == current->pid)
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2560
vm->task_info->task.pid = current->pid;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2561
get_task_comm(vm->task_info->task.comm, current);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2563
vm->task_info->tgid = current->tgid;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2564
get_task_comm(vm->task_info->process_name, current->group_leader);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
564
vres->task.pid = task_pid_nr(current);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
565
get_task_comm(vres->task.comm, current);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1083
svm_range_list_lock_and_flush_work(&p->svms, current->mm);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1085
mmap_write_unlock(current->mm);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
141
current->pid);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
145
process = kfd_create_process(current);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2980
create_process = thread && thread != current && ptrace_parent(thread) == current;
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3005
&& ptrace_parent(target->lead_thread) != current) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3186
process = create_process(current, false);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3375
ptrace_parent(process->lead_thread) == current)
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3379
if (process->lead_thread != current->group_leader
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3440
task_pid_nr(current), cmd, nr);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3497
if (process->lead_thread != current->group_leader)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1040
if (WARN(q->process->mm != current->mm,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1046
&q->properties, current->mm);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
701
if (WARN(q->process->mm != current->mm,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
706
q->queue, &q->properties, current->mm);
drivers/gpu/drm/amd/amdkfd/kfd_events.c
1003
if (signal_pending(current)) {
drivers/gpu/drm/amd/amdkfd/kfd_events.c
998
if (fatal_signal_pending(current)) {
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
971
if (READ_ONCE(p->svms.faulting_task) == current) {
drivers/gpu/drm/amd/amdkfd/kfd_process.c
1026
task = current;
drivers/gpu/drm/amd/amdkfd/kfd_smi_events.c
388
client->pid = current->tgid;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1790
WRITE_ONCE(p->svms.faulting_task, current);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4315
struct mm_struct *mm = current->mm;
drivers/gpu/drm/amd/pm/swsmu/smu_internal.h
85
#define smu_get_asic_power_limits(smu, current, default, max, min) smu_ppt_funcs(get_power_limit, 0, smu, current, default, max, min)
drivers/gpu/drm/display/drm_dp_aux_dev.c
161
if (signal_pending(current)) {
drivers/gpu/drm/display/drm_dp_aux_dev.c
204
if (signal_pending(current)) {
drivers/gpu/drm/drm_auth.c
236
rcu_access_pointer(file_priv->pid) == task_tgid(current))
drivers/gpu/drm/drm_file.c
145
rcu_assign_pointer(file->pid, get_pid(task_tgid(current)));
drivers/gpu/drm/drm_file.c
243
current->comm, task_pid_nr(current),
drivers/gpu/drm/drm_file.c
333
current->comm, task_pid_nr(current), minor->index);
drivers/gpu/drm/drm_file.c
465
pid = task_tgid(current);
drivers/gpu/drm/drm_framebuffer.c
886
strscpy(fb->comm, current->comm);
drivers/gpu/drm/drm_ioc32.c
384
current->comm, task_pid_nr(current),
drivers/gpu/drm/drm_ioctl.c
199
client->pid = task_pid_vnr(current);
drivers/gpu/drm/drm_ioctl.c
342
if (current->comm[0] == 'X' && req->value == 1) {
drivers/gpu/drm/drm_ioctl.c
870
current->comm, task_pid_nr(current),
drivers/gpu/drm/drm_ioctl.c
909
current->comm, task_pid_nr(current),
drivers/gpu/drm/drm_ioctl.c
917
current->comm, task_pid_nr(current), retcode);
drivers/gpu/drm/drm_syncobj.c
1079
entries[i].task = current;
drivers/gpu/drm/drm_syncobj.c
1167
if (signal_pending(current)) {
drivers/gpu/drm/drm_syncobj.c
484
wait.task = current;
drivers/gpu/drm/drm_syncobj.c
499
if (signal_pending(current)) {
drivers/gpu/drm/etnaviv/etnaviv_gem.c
673
might_lock_read(&current->mm->mmap_lock);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
675
if (userptr->mm != current->mm)
drivers/gpu/drm/etnaviv/etnaviv_gem.c
748
etnaviv_obj->userptr.mm = current->mm;
drivers/gpu/drm/etnaviv/etnaviv_gem_submit.c
425
struct pid *pid = get_pid(task_pid(current));
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1330
runqueue_node->pid = current->pid;
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
163
current->comm, fb->width, fb->height, fb->pitches[0], hv->fb_size);
drivers/gpu/drm/i915/display/intel_display_driver.c
373
set_display_access(display, false, current);
drivers/gpu/drm/i915/display/intel_display_driver.c
409
set_display_access(display, false, current);
drivers/gpu/drm/i915/display/intel_display_driver.c
429
display->access.allowed_task == current)
drivers/gpu/drm/i915/display/intel_display_driver.c
433
current->comm, task_pid_vnr(current));
drivers/gpu/drm/i915/gem/i915_gem_context.c
1723
ctx->pid = get_task_pid(current, PIDTYPE_PID);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1727
current->comm, pid_nr(ctx->pid));
drivers/gpu/drm/i915/gem/i915_gem_context.c
2387
current->comm, task_pid_nr(current));
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1776
if (signal_pending(current)) {
drivers/gpu/drm/i915/gem/i915_gem_mman.c
105
struct mm_struct *mm = current->mm;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
242
if (obj->userptr.notifier.mm != current->mm)
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
514
ret = probe_range(current->mm, args->user_ptr, args->user_size);
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
75
return mmu_interval_notifier_insert(&obj->userptr.notifier, current->mm,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1193
mmap_read_lock(current->mm);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1194
area = vma_lookup(current->mm, addr);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1195
mmap_read_unlock(current->mm);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1464
err = access_process_vm(current, addr, &x, sizeof(x), 0);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1471
err = access_process_vm(current, addr, &B, sizeof(B), FOLL_WRITE);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1708
return apply_to_page_range(current->mm, addr, len,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1714
return apply_to_page_range(current->mm, addr, len,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1849
if (!current->mm) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1850
kthread_use_mm(current->active_mm);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1857
kthread_unuse_mm(current->active_mm);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
925
mmap_read_lock(current->mm);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
926
area = vma_lookup(current->mm, addr);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
927
mmap_read_unlock(current->mm);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2649
current->comm);
drivers/gpu/drm/i915/gt/intel_gt.c
672
if (signal_pending(current))
drivers/gpu/drm/i915/gt/intel_ring_submission.c
671
current->comm);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
666
if (signal_pending_state(state, current)) {
drivers/gpu/drm/i915/i915_hwmon.c
460
if (signal_pending(current)) {
drivers/gpu/drm/i915/i915_request.c
1938
if (signal_pending_state(state, current))
drivers/gpu/drm/i915/i915_request.c
2050
wait.tsk = current;
drivers/gpu/drm/i915/i915_request.c
2078
if (signal_pending_state(state, current)) {
drivers/gpu/drm/i915/selftests/i915_selftest.c
210
if (signal_pending(current))
drivers/gpu/drm/i915/selftests/i915_selftest.c
218
if (err == -EINTR && !signal_pending(current))
drivers/gpu/drm/i915/selftests/i915_selftest.c
432
if (signal_pending(current))
drivers/gpu/drm/i915/selftests/i915_selftest.c
463
if (!signal_pending(current)) {
drivers/gpu/drm/imagination/pvr_queue.c
1047
cctx_fw->pid = task_tgid_nr(current);
drivers/gpu/drm/lima/lima_ctx.c
31
ctx->pid = task_pid_nr(current);
drivers/gpu/drm/lima/lima_ctx.c
32
get_task_comm(ctx->pname, current);
drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h
1000
#define DPU_ATRACE_END(name) trace_tracing_mark_write(current->tgid, name, 0)
drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h
1001
#define DPU_ATRACE_BEGIN(name) trace_tracing_mark_write(current->tgid, name, 1)
drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h
1005
trace_dpu_trace_counter(current->tgid, name, value)
drivers/gpu/drm/msm/disp/dpu1/dpu_trace.h
132
__entry->pid = current->tgid;
drivers/gpu/drm/msm/msm_drv.c
237
priv->gpu, current, !ctx->userspace_managed_vm);
drivers/gpu/drm/msm/msm_gem_submit.c
73
submit->pid = get_pid(task_pid(current));
drivers/gpu/drm/nouveau/nouveau_chan.c
337
current->comm, task_pid_nr(current));
drivers/gpu/drm/nouveau/nouveau_drm.c
1203
current->comm, pid_nr(rcu_dereference(fpriv->pid)));
drivers/gpu/drm/nouveau/nouveau_fence.c
288
if (intr && signal_pending(current))
drivers/gpu/drm/nouveau/nouveau_fence.c
312
if (intr && signal_pending(current)) {
drivers/gpu/drm/nouveau/nouveau_sched.c
418
current->pid);
drivers/gpu/drm/nouveau/nouveau_svm.c
163
mm = get_task_mm(current);
drivers/gpu/drm/nouveau/nouveau_svm.c
358
mmap_write_lock(current->mm);
drivers/gpu/drm/nouveau/nouveau_svm.c
360
ret = __mmu_notifier_register(&svmm->notifier, current->mm);
drivers/gpu/drm/nouveau/nouveau_svm.c
367
mmap_write_unlock(current->mm);
drivers/gpu/drm/nouveau/nouveau_svm.c
372
mmap_write_unlock(current->mm);
drivers/gpu/drm/panfrost/panfrost_gem.c
38
bo->debugfs.creator.tgid = current->tgid;
drivers/gpu/drm/panfrost/panfrost_gem.c
39
get_task_comm(bo->debugfs.creator.process_name, current->group_leader);
drivers/gpu/drm/panthor/panthor_drv.c
1487
if (test_tsk_thread_flag(current, TIF_32BIT))
drivers/gpu/drm/panthor/panthor_gem.c
48
bo->debugfs.creator.tgid = current->tgid;
drivers/gpu/drm/panthor/panthor_gem.c
49
get_task_comm(bo->debugfs.creator.process_name, current->group_leader);
drivers/gpu/drm/panthor/panthor_mmu.c
1394
} else if (TASK_SIZE_OF(current) < full_va_range) {
drivers/gpu/drm/panthor/panthor_mmu.c
1399
user_va_range = TASK_SIZE_OF(current);
drivers/gpu/drm/panthor/panthor_sched.c
3603
struct task_struct *task = current->group_leader;
drivers/gpu/drm/radeon/radeon_cs.c
194
mmap_read_lock(current->mm);
drivers/gpu/drm/radeon/radeon_cs.c
199
mmap_read_unlock(current->mm);
drivers/gpu/drm/radeon/radeon_fence.c
1010
cb.task = current;
drivers/gpu/drm/radeon/radeon_fence.c
1035
if (t > 0 && intr && signal_pending(current))
drivers/gpu/drm/radeon/radeon_gem.c
133
robj->pid = task_pid_nr(current);
drivers/gpu/drm/radeon/radeon_gem.c
401
mmap_read_lock(current->mm);
drivers/gpu/drm/radeon/radeon_gem.c
404
mmap_read_unlock(current->mm);
drivers/gpu/drm/radeon/radeon_gem.c
411
mmap_read_unlock(current->mm);
drivers/gpu/drm/radeon/radeon_mn.c
100
ret = mmu_interval_notifier_insert(&bo->notifier, current->mm, addr,
drivers/gpu/drm/radeon/radeon_ttm.c
334
if (current->mm != gtt->usermm)
drivers/gpu/drm/radeon/radeon_ttm.c
583
gtt->usermm = current->mm;
drivers/gpu/drm/scheduler/sched_entity.c
297
if (current->flags & PF_EXITING) {
drivers/gpu/drm/scheduler/sched_entity.c
309
last_user = cmpxchg(&entity->last_user, current->group_leader, NULL);
drivers/gpu/drm/scheduler/sched_entity.c
310
if (last_user == current->group_leader &&
drivers/gpu/drm/scheduler/sched_entity.c
311
(current->flags & PF_EXITING) && (current->exit_code == SIGKILL))
drivers/gpu/drm/scheduler/sched_entity.c
592
WRITE_ONCE(entity->last_user, current->group_leader);
drivers/gpu/drm/scheduler/sched_entity.c
73
entity->last_user = current->group_leader;
drivers/gpu/drm/tegra/submit.c
29
current->comm, ##__VA_ARGS__)
drivers/gpu/drm/tegra/submit.c
528
current->comm, args->context);
drivers/gpu/drm/tegra/uapi.c
118
struct pid *pid = get_task_pid(current, PIDTYPE_TGID);
drivers/gpu/drm/tegra/uapi.c
312
sp = host1x_syncpt_alloc(host1x, HOST1X_SYNCPT_CLIENT_MANAGED, current->comm);
drivers/gpu/drm/tests/drm_framebuffer_test.c
595
KUNIT_EXPECT_STREQ(test, fb1.comm, current->comm);
drivers/gpu/drm/vc4/vc4_gem.c
411
if (interruptible && signal_pending(current)) {
drivers/gpu/drm/virtio/virtgpu_ioctl.c
53
get_task_comm(dbgname, current);
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
219
if (interruptible && signal_pending(current)) {
drivers/gpu/drm/vmwgfx/vmwgfx_irq.c
216
if (interruptible && signal_pending(current)) {
drivers/gpu/drm/vmwgfx/vmwgfx_mksstat.h
102
.slot = vmw_mksstat_get_kern_slot(current->pid, dev_priv) \
drivers/gpu/drm/vmwgfx/vmwgfx_mksstat.h
116
const pid_t pid = atomic_cmpxchg(&dev_priv->mksstat_kern_pids[_##kern_cntr.slot], current->pid, MKSSTAT_PID_RESERVED); \
drivers/gpu/drm/vmwgfx/vmwgfx_mksstat.h
119
if (pid == current->pid) { \
drivers/gpu/drm/vmwgfx/vmwgfx_mksstat.h
134
atomic_set(&dev_priv->mksstat_kern_pids[_##kern_cntr.slot], current->pid); \
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
1048
atomic_set(&dev_priv->mksstat_user_pids[slot], task_pgrp_vnr(current));
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
1052
DRM_DEV_INFO(dev->dev, "pid=%d arg.description='%.*s' id=%zu\n", current->pid, (int)desc_len, pdesc->description, slot);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
1105
DRM_DEV_INFO(dev->dev, "pid=%d arg.id=%zu\n", current->pid, slot);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
1107
pgid = task_pgrp_vnr(current);
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
727
MKSSTAT_KERNEL_DESCRIPTION, current->pid);
drivers/gpu/drm/xe/tests/xe_bo.c
546
if (signal_pending(current))
drivers/gpu/drm/xe/tests/xe_bo.c
562
if (!signal_pending(current) && !failed) {
drivers/gpu/drm/xe/tests/xe_bo.c
574
} while (ret == -EINTR && !signal_pending(current));
drivers/gpu/drm/xe/xe_pm.c
562
xe_pm_write_callback_task(xe, current);
drivers/gpu/drm/xe/xe_pm.c
648
xe_pm_write_callback_task(xe, current);
drivers/gpu/drm/xe/xe_pm.c
767
if (xe_pm_read_callback_task(xe) == current)
drivers/gpu/drm/xe/xe_pm.c
781
if (xe_pm_read_callback_task(xe) == current) {
drivers/gpu/drm/xe/xe_pm.c
806
if (WARN_ON(xe_pm_read_callback_task(xe) == current))
drivers/gpu/drm/xe/xe_pm.c
834
if (xe_pm_read_callback_task(xe) == current) {
drivers/gpu/drm/xe/xe_pm.c
898
if (xe_pm_read_callback_task(xe) == current) {
drivers/gpu/drm/xe/xe_pt.c
2358
!current->mm)
drivers/gpu/drm/xe/xe_svm.c
877
current->mm, 0, vm->size,
drivers/gpu/drm/xe/xe_sync.c
70
ufence->mm = current->mm;
drivers/gpu/drm/xe/xe_userptr.c
287
err = mmu_interval_notifier_insert(&userptr->notifier, current->mm,
drivers/gpu/drm/xe/xe_validation.c
41
xe_assert(xe, current->kunit_test);
drivers/gpu/drm/xe/xe_vm.h
327
WRITE_ONCE(vm->validation.validating, current);
drivers/gpu/drm/xe/xe_vm.h
363
if (READ_ONCE(vm->validation.validating) == current) {
drivers/gpu/drm/xe/xe_wait_user_fence.c
143
if (signal_pending(current)) {
drivers/hid/hid-core.c
1421
__func__, n, current->comm);
drivers/hid/hid-core.c
1467
__func__, n, current->comm);
drivers/hid/hid-core.c
1475
__func__, value, n, current->comm);
drivers/hid/hid-debug.c
3718
DECLARE_WAITQUEUE(wait, current);
drivers/hid/hid-debug.c
3726
if (signal_pending(current)) {
drivers/hid/hid-roccat.c
80
DECLARE_WAITQUEUE(wait, current);
drivers/hid/hid-roccat.c
95
if (signal_pending(current)) {
drivers/hid/hidraw.c
131
task_pid_nr(current));
drivers/hid/hidraw.c
138
task_pid_nr(current));
drivers/hid/hidraw.c
215
task_pid_nr(current));
drivers/hid/hidraw.c
222
task_pid_nr(current));
drivers/hid/hidraw.c
50
DECLARE_WAITQUEUE(wait, current);
drivers/hid/hidraw.c
63
if (signal_pending(current)) {
drivers/hid/uhid.c
751
task_tgid_vnr(current), current->comm);
drivers/hid/usbhid/hiddev.c
342
if (signal_pending(current)) {
drivers/hsi/clients/cmt_speech.c
1158
} else if (signal_pending(current)) {
drivers/hsi/clients/cmt_speech.c
849
if (signal_pending(current)) {
drivers/hv/mshv_regions.c
580
ret = mmu_interval_notifier_insert(&region->mreg_mni, current->mm,
drivers/hv/mshv_root_main.c
1312
mmap_read_lock(current->mm);
drivers/hv/mshv_root_main.c
1313
vma = vma_lookup(current->mm, mem->userspace_addr);
drivers/hv/mshv_root_main.c
1316
mmap_read_unlock(current->mm);
drivers/hv/mshv_vtl_main.c
1082
current->pid);
drivers/hv/mshv_vtl_main.c
1100
current->pid);
drivers/hv/mshv_vtl_main.c
1198
pid_t pid = task_pid_vnr(current);
drivers/hv/vmbus_drv.c
1381
sched_set_fifo(current);
drivers/hwtracing/coresight/coresight-etm3x-sysfs.c
1034
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm3x-sysfs.c
1062
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm3x-sysfs.c
1088
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm3x-sysfs.c
1108
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm4x-core.c
820
task_is_in_init_pid_ns(current))
drivers/hwtracing/coresight/coresight-etm4x-core.c
835
if (task_is_in_init_pid_ns(current))
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
1908
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
1936
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
1969
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
1993
if (task_active_pid_ns(current) != &init_pid_ns)
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
2133
if (!task_is_in_init_pid_ns(current))
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
2154
if (!task_is_in_init_pid_ns(current))
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
2184
if (!task_is_in_init_pid_ns(current))
drivers/hwtracing/coresight/coresight-etm4x-sysfs.c
2208
if (!task_is_in_init_pid_ns(current))
drivers/hwtracing/stm/core.c
634
char comm[sizeof(current->comm)];
drivers/hwtracing/stm/core.c
637
get_task_comm(comm, current);
drivers/i2c/busses/i2c-ibm_iic.c
436
if (signal_pending(current)){
drivers/iio/industrialio-buffer.c
182
if (signal_pending(current)) {
drivers/iio/industrialio-buffer.c
237
if (signal_pending(current)) {
drivers/infiniband/core/cma.c
1074
ret = __rdma_create_id(current->nsproxy->net_ns, event_handler, context,
drivers/infiniband/core/device.c
1762
ret = rdma_dev_change_netns(dev, current->nsproxy->net_ns, net);
drivers/infiniband/core/restrack.c
142
rdma_restrack_attach_task(res, current);
drivers/infiniband/core/ucma.c
1836
__func__, task_tgid_vnr(current), current->comm);
drivers/infiniband/core/umem.c
205
umem->owning_mm = mm = current->mm;
drivers/infiniband/core/umem_odp.c
149
umem->owning_mm = current->mm;
drivers/infiniband/core/umem_odp.c
152
umem_odp->tgid = get_task_pid(current, PIDTYPE_TGID);
drivers/infiniband/core/umem_odp.c
252
umem_odp->umem.owning_mm = current->mm;
drivers/infiniband/core/umem_odp.c
261
umem_odp->tgid = get_task_pid(current, PIDTYPE_TGID);
drivers/infiniband/core/user_mad.c
1016
if (!rdma_dev_access_netns(port->ib_dev, current->nsproxy->net_ns)) {
drivers/infiniband/core/user_mad.c
1113
if (!rdma_dev_access_netns(port->ib_dev, current->nsproxy->net_ns)) {
drivers/infiniband/core/user_mad.c
778
current->comm);
drivers/infiniband/core/uverbs_main.c
585
task_tgid_vnr(current), current->comm);
drivers/infiniband/core/uverbs_main.c
952
if (!rdma_dev_access_netns(ib_dev, current->nsproxy->net_ns)) {
drivers/infiniband/core/uverbs_std_types_dmah.c
50
if (!cpumask_test_cpu(dmah->cpu_id, current->cpus_ptr)) {
drivers/infiniband/hw/erdma/erdma_cm.c
158
if (signal_pending(current))
drivers/infiniband/hw/erdma/erdma_cm.c
159
flush_signals(current);
drivers/infiniband/hw/hfi1/affinity.c
1001
} else if (current->nr_cpus_allowed < cpumask_weight(&set->mask)) {
drivers/infiniband/hw/hfi1/affinity.c
1003
current->pid, current->comm,
drivers/infiniband/hw/hfi1/affinity.c
982
*proc_mask = current->cpus_ptr;
drivers/infiniband/hw/hfi1/affinity.c
990
if (current->nr_cpus_allowed == 1) {
drivers/infiniband/hw/hfi1/affinity.c
992
current->pid, current->comm,
drivers/infiniband/hw/hfi1/file_ops.c
967
uctxt->ctxt, fd->subctxt, current->pid, fd->rec_cpu_num,
drivers/infiniband/hw/hfi1/file_ops.c
999
strscpy(uctxt->comm, current->comm, sizeof(uctxt->comm));
drivers/infiniband/hw/hfi1/mmu_rb.c
129
if (current->mm != handler->mn.mm)
drivers/infiniband/hw/hfi1/mmu_rb.c
233
if (current->mm != handler->mn.mm)
drivers/infiniband/hw/hfi1/mmu_rb.c
68
ret = mmu_notifier_register(&h->mn, current->mm);
drivers/infiniband/hw/hfi1/pin_system.c
127
if (!hfi1_can_pin_pages(pq->dd, current->mm, atomic_read(&pq->n_locked),
drivers/infiniband/hw/hfi1/pin_system.c
138
pinned = hfi1_acquire_user_pages(current->mm, start_address, npages, 0,
drivers/infiniband/hw/hfi1/pin_system.c
147
unpin_vector_pages(current->mm, pages, node->npages, pinned);
drivers/infiniband/hw/hfi1/sdma.c
839
if (current->nr_cpus_allowed != 1)
drivers/infiniband/hw/hfi1/user_exp_rcv.c
147
mm = current->mm;
drivers/infiniband/hw/hfi1/user_exp_rcv.c
179
if (!hfi1_can_pin_pages(dd, current->mm, fd->tid_n_pinned, npages)) {
drivers/infiniband/hw/hfi1/user_exp_rcv.c
184
pinned = hfi1_acquire_user_pages(current->mm, vaddr, npages, true, pages);
drivers/infiniband/hw/hfi1/user_exp_rcv.c
277
&tidbuf->notifier, current->mm,
drivers/infiniband/hw/hfi1/user_exp_rcv.c
761
&node->notifier, current->mm,
drivers/infiniband/hw/mlx4/mr.c
119
mmap_read_lock(current->mm);
drivers/infiniband/hw/mlx4/mr.c
125
vma = find_vma(current->mm, untagged_start);
drivers/infiniband/hw/mlx4/mr.c
134
mmap_read_unlock(current->mm);
drivers/infiniband/hw/mlx5/doorbell.c
57
if ((current->mm == page->mm) &&
drivers/infiniband/hw/mlx5/doorbell.c
76
mmgrab(current->mm);
drivers/infiniband/hw/mlx5/doorbell.c
77
page->mm = current->mm;
drivers/infiniband/hw/mlx5/mlx5_ib.h
33
__LINE__, current->pid, ##arg)
drivers/infiniband/hw/mlx5/mlx5_ib.h
37
__LINE__, current->pid, ##arg)
drivers/infiniband/hw/mlx5/mlx5_ib.h
41
__LINE__, current->pid, ##arg)
drivers/infiniband/hw/mlx5/mlx5_ib.h
45
__func__, __LINE__, current->pid, ##arg)
drivers/infiniband/hw/mlx5/qpc.c
178
qp->pid = current->pid;
drivers/infiniband/hw/mthca/mthca_provider.c
849
current->comm);
drivers/infiniband/hw/usnic/usnic_ib_qp_grp.c
703
qp_grp->owner_pid = current->pid;
drivers/infiniband/hw/usnic/usnic_uiom.c
123
uiomr->owning_mm = mm = current->mm;
drivers/infiniband/hw/usnic/usnic_uiom.c
126
locked = atomic64_add_return(npages, &current->mm->pinned_vm);
drivers/infiniband/hw/usnic/usnic_uiom.c
180
atomic64_sub(npages, &current->mm->pinned_vm);
drivers/infiniband/sw/rdmavt/qp.c
1216
qp->pid = current->pid;
drivers/infiniband/sw/siw/siw_cm.c
275
if (signal_pending(current))
drivers/infiniband/sw/siw/siw_cm.c
276
flush_signals(current);
drivers/infiniband/ulp/ipoib/ipoib_main.c
2568
current->comm);
drivers/infiniband/ulp/isert/ib_isert.c
537
isert_info("conn %p final kref %s/%d\n", isert_conn, current->comm,
drivers/infiniband/ulp/isert/ib_isert.c
538
current->pid);
drivers/iommu/iommufd/ioas.c
446
return pages->source_mm != current->mm;
drivers/iommu/iommufd/ioas.c
453
(pages->source_mm != current->mm);
drivers/iommu/iommufd/ioas.c
461
.source_mm = current->mm,
drivers/iommu/iommufd/ioas.c
462
.source_task = current->group_leader,
drivers/iommu/iommufd/ioas.c
499
pages->source_mm = current->mm;
drivers/iommu/iommufd/ioas.c
503
pages->source_task = current->group_leader;
drivers/iommu/iommufd/pages.c
1382
pages->source_mm = current->mm;
drivers/iommu/iommufd/pages.c
1392
pages->source_task = current->group_leader;
drivers/iommu/iommufd/pages.c
1393
get_task_struct(current->group_leader);
drivers/iommu/iommufd/pages.c
2346
bool change_mm = current->mm != pages->source_mm;
drivers/iommu/iommufd/pages.c
813
if (!user->file && pages->source_mm != current->mm)
drivers/iommu/iommufd/pages.c
881
bool remote_mm = pages->source_mm != current->mm;
drivers/iommu/iommufd/pages.c
995
pages->source_mm != current->mm) {
drivers/irqchip/irq-gic-v4.c
127
name = kasprintf(GFP_KERNEL, "GICv4-sgi-%d", task_pid_nr(current));
drivers/irqchip/irq-gic-v4.c
163
task_pid_nr(current));
drivers/isdn/mISDN/l1oip_core.c
709
while (!signal_pending(current)) {
drivers/isdn/mISDN/stack.c
202
sigfillset(&current->blocked);
drivers/isdn/mISDN/timerdev.c
114
if (signal_pending(current))
drivers/macintosh/adb.c
722
DECLARE_WAITQUEUE(wait, current);
drivers/macintosh/adb.c
748
if (signal_pending(current)) {
drivers/macintosh/ams/ams-core.c
56
static DEVICE_ATTR(current, S_IRUGO, ams_show_current, NULL);
drivers/macintosh/smu.c
1171
DECLARE_WAITQUEUE(wait, current);
drivers/macintosh/smu.c
1193
if (signal_pending(current))
drivers/macintosh/smu.c
1287
DECLARE_WAITQUEUE(wait, current);
drivers/macintosh/via-pmu.c
1922
switch_mmu_context(NULL, current->active_mm, NULL);
drivers/macintosh/via-pmu.c
2011
switch_mmu_context(NULL, current->active_mm, NULL);
drivers/macintosh/via-pmu.c
2173
DECLARE_WAITQUEUE(wait, current);
drivers/macintosh/via-pmu.c
2205
if (signal_pending(current))
drivers/mailbox/mailbox-test.c
180
DECLARE_WAITQUEUE(wait, current);
drivers/mailbox/mailbox-test.c
206
if (signal_pending(current)) {
drivers/md/bcache/btree.c
1017
if (current->bio_list)
drivers/md/bcache/btree.c
2409
if (current->bio_list) {
drivers/md/bcache/btree.c
2491
BUG_ON(current->bio_list);
drivers/md/bcache/btree.c
411
BUG_ON(current->bio_list);
drivers/md/bcache/btree.c
510
!current->bio_list)
drivers/md/bcache/btree.c
871
c->btree_cache_alloc_lock = current;
drivers/md/bcache/btree.c
872
} else if (c->btree_cache_alloc_lock != current) {
drivers/md/bcache/btree.c
915
if (c->btree_cache_alloc_lock == current) {
drivers/md/bcache/btree.c
927
BUG_ON(current->bio_list);
drivers/md/bcache/request.c
368
struct task_struct *task = current;
drivers/md/bcache/request.c
754
s->iop.write_point = hash_long((unsigned long) current, 16);
drivers/md/dm-bufio.c
1017
#define dm_bufio_in_request() (!!current->bio_list)
drivers/md/dm-bufio.c
1571
DECLARE_WAITQUEUE(wait, current);
drivers/md/dm-integrity.c
1282
new_range->task = current;
drivers/md/dm-integrity.c
1541
DECLARE_WAITQUEUE(wait, current);
drivers/md/dm-snap.c
1543
DECLARE_WAITQUEUE(wait, current);
drivers/md/dm-vdo/funnel-workqueue.c
609
if (kthread_func(current) != work_queue_runner)
drivers/md/dm-vdo/funnel-workqueue.c
613
return kthread_data(current);
drivers/md/dm-vdo/logger.c
131
device_instance, current->comm, prefix, vaf1,
drivers/md/dm-vdo/logger.c
140
if (((current->flags & PF_KTHREAD) != 0) &&
drivers/md/dm-vdo/logger.c
141
(strncmp(module, current->comm, strlen(module)) == 0)) {
drivers/md/dm-vdo/logger.c
142
emit_log_message_to_kernel(priority, "%s: %s%pV%pV\n", current->comm,
drivers/md/dm-vdo/logger.c
148
emit_log_message_to_kernel(priority, "%s: %s: %s%pV%pV\n", module, current->comm,
drivers/md/dm-vdo/thread-registry.c
33
new_thread->task = current;
drivers/md/dm-vdo/thread-registry.c
37
if (thread->task == current) {
drivers/md/dm-vdo/thread-registry.c
62
if (thread->task == current) {
drivers/md/dm-vdo/thread-registry.c
85
if (thread->task == current) {
drivers/md/dm-vdo/thread-utils.c
39
thread->thread_task = current;
drivers/md/dm-vdo/thread-utils.c
54
char *my_name_colon = strchr(current->comm, ':');
drivers/md/dm-vdo/thread-utils.c
84
(int) (my_name_colon - current->comm), current->comm,
drivers/md/dm-zone.c
198
md->revalidate_map_task = current;
drivers/md/dm-zone.c
57
if (!zone_revalidate_map || md->revalidate_map_task != current) {
drivers/md/dm.c
2789
if (signal_pending_state(task_state, current)) {
drivers/md/dm.c
2814
if (signal_pending_state(task_state, current)) {
drivers/md/dm.c
439
if (r == -ENOTCONN && *forward && !fatal_signal_pending(current)) {
drivers/md/dm.c
472
current->comm, cmd);
drivers/md/md.c
10194
if (signal_pending(current)) {
drivers/md/md.c
10200
flush_signals(current);
drivers/md/md.c
8495
if (signal_pending(current))
drivers/md/md.c
8496
flush_signals(current);
drivers/md/md.c
9598
if (signal_pending(current))
drivers/md/md.c
9599
flush_signals(current);
drivers/md/md.c
9683
flush_signals(current); /* just in case */
drivers/md/persistent-data/dm-block-manager.c
105
if (lock->holders[i] == current) {
drivers/md/persistent-data/dm-block-manager.c
207
__add_holder(lock, current);
drivers/md/persistent-data/dm-block-manager.c
212
get_task_struct(current);
drivers/md/persistent-data/dm-block-manager.c
214
w.task = current;
drivers/md/persistent-data/dm-block-manager.c
220
put_task_struct(current);
drivers/md/persistent-data/dm-block-manager.c
235
__add_holder(lock, current);
drivers/md/persistent-data/dm-block-manager.c
249
__del_holder(lock, current);
drivers/md/persistent-data/dm-block-manager.c
270
__add_holder(lock, current);
drivers/md/persistent-data/dm-block-manager.c
275
get_task_struct(current);
drivers/md/persistent-data/dm-block-manager.c
276
w.task = current;
drivers/md/persistent-data/dm-block-manager.c
287
put_task_struct(current);
drivers/md/persistent-data/dm-block-manager.c
295
__del_holder(lock, current);
drivers/md/raid1-10.c
166
mddev->bitmap_ops->unplug(mddev, current->bio_list == NULL);
drivers/md/raid10.c
956
struct bio_list *bio_list = current->bio_list;
drivers/md/raid10.c
980
if (thread->tsk == current) {
drivers/media/common/saa7146/saa7146_vbi.c
118
if(signal_pending(current)) {
drivers/media/common/saa7146/saa7146_vbi.c
16
DECLARE_WAITQUEUE(wait, current);
drivers/media/dvb-core/dvb_ca_en50221.c
1923
if (signal_pending(current)) {
drivers/media/dvb-core/dvb_demux.c
1164
if (signal_pending(current))
drivers/media/dvb-core/dvb_frontend.c
2804
if (signal_pending(current))
drivers/media/dvb-core/dvb_frontend.c
302
init_waitqueue_entry(&wait, current);
drivers/media/dvb-core/dvb_frontend.c
306
if (signal_pending(current)) {
drivers/media/dvb-core/dvb_frontend.c
871
if (signal_pending(current))
drivers/media/pci/cx18/cx18-driver.c
276
sig = intr ? signal_pending(current) : 0;
drivers/media/pci/cx18/cx18-fileops.c
235
if (signal_pending(current)) {
drivers/media/pci/ivtv/ivtv-driver.c
340
return signal_pending(current) ? -EINTR : 0;
drivers/media/pci/ivtv/ivtv-driver.c
352
int ret = signal_pending(current);
drivers/media/pci/ivtv/ivtv-fileops.c
278
if (signal_pending(current)) {
drivers/media/pci/ivtv/ivtv-fileops.c
556
while (!(got_sig = signal_pending(current)) &&
drivers/media/pci/ivtv/ivtv-fileops.c
650
if (signal_pending(current)) {
drivers/media/pci/ivtv/ivtv-ioctl.c
168
got_sig = signal_pending(current);
drivers/media/pci/ivtv/ivtv-streams.c
780
DECLARE_WAITQUEUE(wait, current);
drivers/media/pci/ivtv/ivtv-yuv.c
1094
got_sig = signal_pending(current);
drivers/media/pci/ivtv/ivtvfb.c
303
got_sig = signal_pending(current);
drivers/media/pci/pt1/pt1.c
1115
if (signal_pending(current))
drivers/media/pci/pt3/pt3.c
499
if (signal_pending(current))
drivers/media/pci/solo6x10/solo6x10-i2c.c
246
if (signal_pending(current))
drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c
636
DECLARE_WAITQUEUE(wait, current);
drivers/media/pci/solo6x10/solo6x10-v4l2.c
262
DECLARE_WAITQUEUE(wait, current);
drivers/media/platform/amphion/vpu_v4l2.c
745
inst->pid = current->pid;
drivers/media/platform/amphion/vpu_v4l2.c
746
inst->tgid = current->tgid;
drivers/media/platform/samsung/exynos-gsc/gsc-m2m.c
610
pr_debug("pid: %d, state: 0x%lx", task_pid_nr(current), gsc->state);
drivers/media/platform/samsung/exynos-gsc/gsc-m2m.c
672
task_pid_nr(current), gsc->state, gsc->m2m.refcnt);
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
471
dbg("pid: %d, state: 0x%lx", task_pid_nr(current), fimc->state);
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
520
dbg("pid: %d, state: 0x%lx", task_pid_nr(current), fimc->state);
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
608
pr_debug("pid: %d, state: %#lx\n", task_pid_nr(current), fimc->state);
drivers/media/platform/samsung/exynos4-is/fimc-m2m.c
681
task_pid_nr(current), fimc->state, fimc->m2m.refcnt);
drivers/media/platform/samsung/s3c-camif/camif-capture.c
539
vp->state, vp->owner, task_pid_nr(current));
drivers/media/platform/samsung/s3c-camif/camif-capture.c
571
vp->state, vp->owner, task_pid_nr(current));
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
443
&state->pending, &state->current, &state->complete,
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
512
state->complete = state->current;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
513
state->current = state->pending;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.h
43
struct sun6i_csi_buffer *current;
drivers/media/test-drivers/visl/visl.h
90
#define frame_dprintk(dev, current, fmt, arg...) \
drivers/media/test-drivers/visl/visl.h
93
(current) >= visl_dprintk_frame_start && \
drivers/media/test-drivers/visl/visl.h
94
(current) < visl_dprintk_frame_start + visl_dprintk_nframes) \
drivers/media/test-drivers/vivid/vivid-radio-rx.c
80
if (msleep_interruptible(20) && signal_pending(current))
drivers/media/test-drivers/vivid/vivid-radio-tx.c
65
if (msleep_interruptible(20) && signal_pending(current))
drivers/media/usb/gspca/cpia1.c
742
if (signal_pending(current))
drivers/mfd/ucb1x00-ts.c
207
DECLARE_WAITQUEUE(wait, current);
drivers/misc/bcm-vk/bcm_vk_dev.c
1171
vk->reset_pid = task_pid_nr(current);
drivers/misc/bcm-vk/bcm_vk_msg.c
1215
pid = task_tgid_nr(current);
drivers/misc/bcm-vk/bcm_vk_msg.c
980
ctx = bcm_vk_get_ctx(vk, task_tgid_nr(current));
drivers/misc/bcm-vk/bcm_vk_tty.c
144
vktty->pid = task_pid_nr(current);
drivers/misc/fastrpc.c
1043
mmap_read_lock(current->mm);
drivers/misc/fastrpc.c
1044
vma = find_vma(current->mm, ctx->args[i].ptr);
drivers/misc/fastrpc.c
1048
mmap_read_unlock(current->mm);
drivers/misc/fastrpc.c
1174
msg->tid = current->pid;
drivers/misc/fastrpc.c
1457
inbuf.namelen = strlen(current->comm) + 1;
drivers/misc/fastrpc.c
1482
args[1].ptr = (u64)(uintptr_t)current->comm;
drivers/misc/fastrpc.c
634
ctx->pid = current->pid;
drivers/misc/genwqe/card_dev.c
47
cfile->opener = get_pid(task_tgid(current));
drivers/misc/ibmasm/r_heartbeat.c
71
if (signal_pending(current) || rhb->stopped) {
drivers/misc/ibmvmc.c
1812
set_user_nice(current, -20);
drivers/misc/ibmvmc.c
967
if (signal_pending(current)) {
drivers/misc/lis3lv02d/lis3lv02d.c
605
DECLARE_WAITQUEUE(wait, current);
drivers/misc/lis3lv02d/lis3lv02d.c
625
if (signal_pending(current)) {
drivers/misc/lkdtm/bugs.c
223
pid_t pid = task_pid_nr(current);
drivers/misc/lkdtm/bugs.c
241
pid_t pid = task_pid_nr(current);
drivers/misc/lkdtm/bugs.c
249
if (*canary == current->stack_canary)
drivers/misc/lkdtm/bugs.c
651
const unsigned char *stack = task_stack_page(current);
drivers/misc/lkdtm/bugs.c
665
const unsigned char *stack = task_stack_page(current);
drivers/misc/lkdtm/bugs.c
785
ptrauth_thread_init_kernel(current);
drivers/misc/lkdtm/bugs.c
786
ptrauth_thread_switch_kernel(current);
drivers/misc/lkdtm/kstack_erase.c
30
const unsigned long task_stack_base = (unsigned long)task_stack_page(current);
drivers/misc/lkdtm/kstack_erase.c
31
const unsigned long task_stack_low = stackleak_task_low_bound(current);
drivers/misc/lkdtm/kstack_erase.c
32
const unsigned long task_stack_high = stackleak_task_high_bound(current);
drivers/misc/lkdtm/kstack_erase.c
34
const unsigned long lowest_sp = current->lowest_stack;
drivers/misc/lkdtm/perms.c
107
copied = access_process_vm(current, (unsigned long)dst, do_nothing_text,
drivers/misc/lkdtm/usercopy.c
71
bad_stack = task_stack_page(current) + THREAD_SIZE;
drivers/misc/mei/bus.c
104
if (signal_pending(current))
drivers/misc/mei/bus.c
203
if (signal_pending(current))
drivers/misc/mei/bus.c
212
if (signal_pending(current))
drivers/misc/mei/client.c
2077
if (signal_pending(current))
drivers/misc/mei/main.c
227
if (signal_pending(current))
drivers/misc/mei/main.c
368
if (signal_pending(current))
drivers/misc/mei/main.c
468
if (signal_pending(current))
drivers/misc/mei/main.c
913
if (signal_pending(current))
drivers/misc/ntsync.c
841
if (signal_pending(current)) {
drivers/misc/ntsync.c
890
q->task = current;
drivers/misc/ocxl/file.c
125
if (set_thread_tidr(current)) {
drivers/misc/ocxl/file.c
130
ctx->tidr = current->thread.tidr;
drivers/misc/ocxl/file.c
428
if (signal_pending(current)) {
drivers/misc/ocxl/file.c
87
return ocxl_context_attach(ctx, amr, current->mm);
drivers/misc/ocxl/link.c
532
if (!test_tsk_thread_flag(current, TIF_32BIT))
drivers/misc/ocxl/link.c
624
trace_ocxl_context_add(current->pid, spa->spa_mem, pasid, pidr, tidr);
drivers/misc/ocxl/link.c
706
trace_ocxl_context_remove(current->pid, spa->spa_mem, pasid,
drivers/misc/sgi-gru/grufault.c
112
mmap_read_unlock(current->mm);
drivers/misc/sgi-gru/grufault.c
52
vma = vma_lookup(current->mm, vaddr);
drivers/misc/sgi-gru/grufault.c
68
struct mm_struct *mm = current->mm;
drivers/misc/sgi-gru/grufault.c
85
struct mm_struct *mm = current->mm;
drivers/misc/sgi-gru/grufault.c
891
gts->ts_tgid_owner = current->tgid;
drivers/misc/sgi-gru/grufile.c
139
mmap_write_lock(current->mm);
drivers/misc/sgi-gru/grufile.c
150
mmap_write_unlock(current->mm);
drivers/misc/sgi-gru/grumain.c
336
gts->ts_mm = current->mm;
drivers/misc/sgi-gru/grumain.c
732
if (!gru || gts->ts_tgid_owner != current->tgid)
drivers/misc/sgi-gru/grutlbpurge.c
267
mn = mmu_notifier_get_locked(&gru_mmuops, current->mm);
drivers/misc/sgi-xp/xpc_main.c
253
set_cpus_allowed_ptr(current, cpumask_of(XPC_HB_CHECK_CPU));
drivers/misc/uacce/uacce.c
128
handle = iommu_sva_bind_device(uacce->parent, current->mm);
drivers/mmc/core/core.c
793
struct task_struct *task = ctx ? NULL : current;
drivers/mmc/core/core.c
794
DECLARE_WAITQUEUE(wait, current);
drivers/mmc/core/sdio_irq.c
145
sched_set_fifo_low(current);
drivers/mmc/core/sdio_uart.c
169
if (likely(port->in_sdio_uart_irq != current))
drivers/mmc/core/sdio_uart.c
177
if (likely(port->in_sdio_uart_irq != current))
drivers/mmc/core/sdio_uart.c
508
if (unlikely(port->in_sdio_uart_irq == current))
drivers/mmc/core/sdio_uart.c
515
port->in_sdio_uart_irq = current;
drivers/mmc/host/tmio_mmc_core.c
1036
current->comm, task_pid_nr(current),
drivers/mmc/host/tmio_mmc_core.c
987
current->comm, task_pid_nr(current),
drivers/mmc/host/tmio_mmc_core.c
993
current->comm, task_pid_nr(current),
drivers/mtd/chips/cfi_cmdset_0001.c
1233
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0001.c
1309
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0001.c
825
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0001.c
943
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
1117
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
1311
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
1676
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
1820
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
1954
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
2413
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
2513
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
2914
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0002.c
875
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
1036
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
1182
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
252
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
426
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
738
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/chips/cfi_cmdset_0020.c
979
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/lpddr/lpddr_cmds.c
170
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/lpddr/lpddr_cmds.c
200
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/lpddr/lpddr_cmds.c
297
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/nand/onenand/onenand_base.c
997
DECLARE_WAITQUEUE(wait, current);
drivers/mtd/tests/mtd_test.h
8
if (signal_pending(current)) {
drivers/mtd/ubi/debug.h
20
__func__, __LINE__, current->pid); \
drivers/mtd/ubi/debug.h
29
pr_debug("UBI DBG " type " (pid %d): " fmt "\n", current->pid, \
drivers/mtd/ubi/gluebi.c
32
current->pid, __func__, ##__VA_ARGS__)
drivers/mtd/ubi/wl.c
1673
ubi->bgt_name, task_pid_nr(current));
drivers/net/can/softing/softing_fw.c
330
if (signal_pending(current))
drivers/net/can/softing/softing_fw.c
87
} while (!signal_pending(current));
drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.c
255
struct mm_struct *mm = current->mm;
drivers/net/ethernet/broadcom/bnxt/bnxt_coredump.c
274
strscpy(record->commandline, current->comm, TASK_COMM_LEN);
drivers/net/ethernet/broadcom/tg3.c
12162
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
404
DECLARE_WAITQUEUE(wait, current);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
428
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1365
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1386
signal_pending(current))
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1408
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1550
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1582
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1603
current->comm, current->pid);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1702
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1725
signal_pending(current))
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1744
if (signal_pending(current)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
931
if (signal_pending(current))
drivers/net/ethernet/intel/igbvf/ethtool.c
22
#define IGBVF_STAT(current, base) \
drivers/net/ethernet/intel/igbvf/ethtool.c
23
sizeof(((struct igbvf_adapter *)0)->current), \
drivers/net/ethernet/intel/igbvf/ethtool.c
24
offsetof(struct igbvf_adapter, current), \
drivers/net/ethernet/mellanox/mlx5/core/cq.c
159
cq->pid = current->pid;
drivers/net/ethernet/mellanox/mlx5/core/en/rss.c
8
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/fpga/core.h
64
__func__, __LINE__, current->pid, ##__VA_ARGS__)
drivers/net/ethernet/mellanox/mlx5/core/fpga/core.h
68
__func__, __LINE__, current->pid, ##__VA_ARGS__)
drivers/net/ethernet/mellanox/mlx5/core/fpga/core.h
72
__func__, __LINE__, current->pid, ##__VA_ARGS__)
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
158
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
50
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
56
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
67
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
73
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
78
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
83
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
89
__func__, __LINE__, current->pid, \
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
98
__func__, __LINE__, current->pid, \
drivers/net/ethernet/sfc/falcon/falcon.c
635
if (signal_pending(current)) {
drivers/net/ethernet/sfc/falcon/falcon.c
728
if (signal_pending(current)) {
drivers/net/ethernet/sfc/falcon/falcon.c
758
if (signal_pending(current))
drivers/net/ethernet/sfc/falcon/falcon.c
844
if (signal_pending(current))
drivers/net/mctp/mctp-i2c.c
886
dev_net_set(ndev, current->nsproxy->net_ns);
drivers/net/netdevsim/bus.c
466
nsim_bus_dev->initial_net = current->nsproxy->net_ns;
drivers/net/netkit.c
537
dev = netkit_dev_fetch(current->nsproxy->net_ns, attr->target_ifindex,
drivers/net/netkit.c
577
dev = netkit_dev_fetch(current->nsproxy->net_ns, attr->target_ifindex,
drivers/net/netkit.c
608
dev = netkit_dev_fetch(current->nsproxy->net_ns,
drivers/net/netkit.c
792
dev = netkit_dev_fetch(current->nsproxy->net_ns,
drivers/net/ppp/ppp_generic.c
1684
if (xmit_recursion->owner == current)
drivers/net/ppp/ppp_generic.c
1687
xmit_recursion->owner = current;
drivers/net/ppp/ppp_generic.c
2217
xmit_recursion->owner = current;
drivers/net/ppp/ppp_generic.c
2920
return ppp_register_net_channel(current->nsproxy->net_ns, chan);
drivers/net/ppp/ppp_generic.c
436
DECLARE_WAITQUEUE(wait, current);
drivers/net/ppp/ppp_generic.c
475
if (signal_pending(current))
drivers/net/ppp/ppp_generic.c
738
err = ppp_unattached_ioctl(current->nsproxy->net_ns,
drivers/net/ppp/ppp_generic.c
750
current->comm, current->pid);
drivers/net/ppp/ppp_generic.c
777
pn = ppp_pernet(current->nsproxy->net_ns);
drivers/net/slip/slip.c
1249
if (sl->tty != current->signal->tty &&
drivers/net/slip/slip.c
1250
sl->pid != current->pid) {
drivers/net/slip/slip.c
823
sl->pid = current->pid;
drivers/net/tap.c
448
struct net *net = current->nsproxy->net_ns;
drivers/net/tap.c
784
if (signal_pending(current)) {
drivers/net/tun.c
1602
struct page_frag *alloc_frag = &current->task_frag;
drivers/net/tun.c
2118
DECLARE_WAITQUEUE(wait, current);
drivers/net/tun.c
2137
if (signal_pending(current)) {
drivers/net/tun.c
3384
__f_setown(file, task_pid(current), PIDTYPE_TGID, 0);
drivers/net/tun.c
3395
struct net *net = current->nsproxy->net_ns;
drivers/net/usb/hso.c
1536
DECLARE_WAITQUEUE(wait, current);
drivers/net/usb/hso.c
1564
if (signal_pending(current)) {
drivers/net/usb/lan78xx.c
3470
DECLARE_WAITQUEUE(wait, current);
drivers/net/usb/smsc95xx.c
123
if (current != pdata->pm_task)
drivers/net/usb/smsc95xx.c
1551
pdata->pm_task = current;
drivers/net/usb/smsc95xx.c
1816
pdata->pm_task = current;
drivers/net/usb/smsc95xx.c
1860
pdata->pm_task = current;
drivers/net/usb/smsc95xx.c
92
if (current != pdata->pm_task)
drivers/net/usb/usbnet.c
808
DECLARE_WAITQUEUE(wait, current);
drivers/net/wireless/ath/ath6kl/cfg80211.c
1471
if (signal_pending(current)) {
drivers/net/wireless/ath/ath6kl/cfg80211.c
500
if (signal_pending(current)) {
drivers/net/wireless/broadcom/b43legacy/main.c
1729
if (signal_pending(current)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/core.c
913
current->pid, name, ifp->mac_addr);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1731
DECLARE_WAITQUEUE(wait, current);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1738
while (!(*condition) && (!signal_pending(current) && timeout))
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1741
if (signal_pending(current))
drivers/net/wireless/marvell/libertas/main.c
419
init_waitqueue_entry(&wait, current);
drivers/net/wireless/marvell/mwl8k.c
1568
if (priv->hw_restart_owner == current)
drivers/net/wireless/marvell/mwl8k.c
2166
if (priv->fw_mutex_owner != current) {
drivers/net/wireless/marvell/mwl8k.c
2182
priv->fw_mutex_owner = current;
drivers/net/wireless/marvell/mwl8k.c
4991
priv->hw_restart_owner = current;
drivers/net/wireless/realtek/rtlwifi/debug.c
38
current->comm, current->pid, titlestring);
drivers/net/wireless/realtek/rtw88/sdio.c
1081
rtwsdio->irq_thread = current;
drivers/net/wireless/realtek/rtw88/sdio.c
32
rtwsdio->irq_thread != current;
drivers/net/wireless/rsi/rsi_91x_sdio.c
410
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
418
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
442
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
450
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
495
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
500
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
538
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/rsi/rsi_91x_sdio.c
543
if (likely(dev->sdio_irq_task != current))
drivers/net/wireless/st/cw1200/cw1200_spi.c
200
DECLARE_WAITQUEUE(wait, current);
drivers/nvdimm/region_devs.c
1138
idx = this_cpu_add_return(flush_idx, hash_32(current->pid + idx, 8));
drivers/nvme/host/core.c
2633
if (fatal_signal_pending(current))
drivers/nvme/host/ioctl.c
274
current->comm, nsid, ns->head->ns_id);
drivers/nvme/host/tcp.c
1786
ret = sock_create_kern(current->nsproxy->net_ns,
drivers/parport/daisy.c
147
if (signal_pending(current))
drivers/parport/ieee1284.c
120
if (signal_pending (current))
drivers/parport/ieee1284.c
189
if (signal_pending (current))
drivers/parport/ieee1284_ops.c
102
if (signal_pending (current))
drivers/parport/ieee1284_ops.c
109
if (signal_pending (current))
drivers/parport/ieee1284_ops.c
428
if (signal_pending (current)) {
drivers/parport/ieee1284_ops.c
531
if (signal_pending (current))
drivers/parport/ieee1284_ops.c
659
if (signal_pending (current)) {
drivers/parport/parport_ip32.c
1246
if (signal_pending(current)) {
drivers/parport/parport_ip32.c
1504
if (signal_pending(current))
drivers/parport/parport_ip32.c
1515
if (signal_pending(current))
drivers/parport/parport_pc.c
170
if (signal_pending(current))
drivers/parport/parport_pc.c
182
if (signal_pending(current))
drivers/parport/parport_pc.c
551
if (signal_pending(current))
drivers/parport/share.c
1103
if (signal_pending(current))
drivers/pci/hotplug/cpci_hotplug_core.c
459
if (kthread_should_stop() || signal_pending(current))
drivers/pci/hotplug/cpqphp.h
706
DECLARE_WAITQUEUE(wait, current);
drivers/pci/hotplug/cpqphp.h
714
if (signal_pending(current))
drivers/pci/p2pdma.c
107
current->comm);
drivers/pci/p2pdma.c
114
current->comm);
drivers/pci/pci-sysfs.c
831
current->comm, off);
drivers/pci/quirks.c
150
pci_info(dev, "calling %pS @ %i\n", fn, task_pid_nr(current));
drivers/pci/vgaarb.c
468
init_waitqueue_entry(&wait, current);
drivers/pci/vgaarb.c
473
if (interruptible && signal_pending(current)) {
drivers/pci/vpd.c
187
if (fatal_signal_pending(current)) {
drivers/pcmcia/cs.c
580
skt->thread = current;
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
495
ret = vchiq_open_service_internal(service, current->pid);
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
595
if (iter->pid == current->pid) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
629
if ((ret != -EAGAIN) || fatal_signal_pending(current) || !waiter->bulk_waiter.bulk) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
640
waiter->pid = current->pid;
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
645
waiter, current->pid);
drivers/platform/raspberrypi/vchiq-interface/vchiq_arm.c
981
flush_signals(current);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2660
flush_signals(current);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2685
flush_signals(current);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3388
if (current == service->state->slot_handler_thread) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3445
(current == service->state->slot_handler_thread)) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
1193
instance->pid = current->tgid;
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
319
if (iter->pid == current->pid) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
328
"arm: no bulk_waiter found for pid %d\n", current->pid);
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
333
waiter, current->pid);
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
353
if ((status != -EAGAIN) || fatal_signal_pending(current) ||
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
366
waiter->pid = current->pid;
drivers/platform/raspberrypi/vchiq-interface/vchiq_dev.c
371
waiter, current->pid);
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
551
if (signal_pending(current)) {
drivers/platform/x86/lenovo/thinkpad_acpi.c
464
what, task_tgid_vnr(current));
drivers/platform/x86/lenovo/thinkpad_acpi.c
472
what, task_tgid_vnr(current), ## arg); \
drivers/power/supply/cw2015_battery.c
250
#define HYSTERESIS(current, previous, up, down) \
drivers/power/supply/cw2015_battery.c
251
(((current) < (previous) + (up)) && ((current) > (previous) - (down)))
drivers/power/supply/rk817_charger.c
115
#define CURRENT_TO_ADC(current, samp_res) \
drivers/power/supply/rk817_charger.c
116
(current * samp_res / 172)
drivers/powercap/idle_inject.c
298
sched_set_fifo(current);
drivers/pwm/core.c
2184
label = kasprintf(GFP_KERNEL, "pwm-cdev (pid=%d)", current->pid);
drivers/rapidio/devices/rio_mport_cdev.c
1029
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
1038
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
1047
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
1938
filp, current->comm, task_pid_nr(current));
drivers/rapidio/devices/rio_mport_cdev.c
1950
filp, current->comm, task_pid_nr(current));
drivers/rapidio/devices/rio_mport_cdev.c
1971
current->comm, task_pid_nr(current), wret);
drivers/rapidio/devices/rio_mport_cdev.c
1976
filp, current->comm, task_pid_nr(current));
drivers/rapidio/devices/rio_mport_cdev.c
694
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
725
rmcd_debug(DMA, "pid=%d DMA_%s tx_cookie = %d", task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
751
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
759
current->comm, task_pid_nr(current),
drivers/rapidio/devices/rio_mport_cdev.c
767
current->comm, task_pid_nr(current),
drivers/rapidio/devices/tsi721_dma.c
317
bdma_chan->id, task_pid_nr(current));
drivers/rapidio/devices/tsi721_dma.c
323
task_pid_nr(current));
drivers/rapidio/rio_cm.c
1441
ch->id, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
1457
current->comm, task_pid_nr(current), ch->id);
drivers/rapidio/rio_cm.c
1462
current->comm, task_pid_nr(current), ch->id);
drivers/rapidio/rio_cm.c
1482
current->comm, task_pid_nr(current), filp);
drivers/rapidio/rio_cm.c
1500
current->comm, task_pid_nr(current), filp);
drivers/rapidio/rio_cm.c
1507
ch->id, current->comm,
drivers/rapidio/rio_cm.c
1508
task_pid_nr(current));
drivers/rapidio/rio_cm.c
1665
ch_num, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
1672
ch_num, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
1691
ch_num, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
1755
param.ch_num, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
1766
ch->id, current->comm, task_pid_nr(current));
drivers/rapidio/rio_cm.c
791
riocm_error("%s(%d) ch_%d not found", current->comm,
drivers/rapidio/rio_cm.c
792
task_pid_nr(current), ch_id);
drivers/regulator/core.c
170
if (rdev->mutex_owner == current)
drivers/regulator/core.c
186
rdev->mutex_owner = current;
drivers/regulator/core.c
263
contended->mutex_owner = current;
drivers/regulator/core.c
422
old_contended_rdev->mutex_owner = current;
drivers/regulator/core.c
6691
old_contended_rdev->mutex_owner = current;
drivers/rtc/dev.c
146
DECLARE_WAITQUEUE(wait, current);
drivers/rtc/dev.c
170
if (signal_pending(current)) {
drivers/s390/char/con3215.c
647
DECLARE_WAITQUEUE(wait, current);
drivers/s390/char/fs3270.c
455
fp->fs_pid = get_pid(task_pid(current));
drivers/s390/char/keyboard.c
493
perm = current->signal->tty == tty || capable(CAP_SYS_TTY_CONFIG);
drivers/s390/cio/device.c
1347
if (signal_pending(current))
drivers/s390/crypto/ap_bus.c
1138
if (ap_scan_bus_task == current)
drivers/s390/crypto/ap_bus.c
1145
ap_scan_bus_task = current;
drivers/s390/crypto/ap_bus.c
2460
ap_scan_bus_task = current;
drivers/s390/crypto/ap_bus.c
530
DECLARE_WAITQUEUE(wait, current);
drivers/s390/crypto/ap_bus.c
532
set_user_nice(current, MAX_NICE);
drivers/s390/crypto/zcrypt_msgtype50.c
465
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype50.c
514
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype6.c
1017
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype6.c
1060
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype6.c
1150
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype6.c
1264
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/s390/crypto/zcrypt_msgtype6.c
972
ap_msg->psmid = (((unsigned long)current->pid) << 32) +
drivers/sbus/char/bbc_i2c.c
126
DECLARE_WAITQUEUE(wait, current);
drivers/sbus/char/envctrl.c
1013
current->comm, whichcpu,
drivers/sbus/char/envctrl.c
1020
printk(KERN_INFO PFX "%s exiting...\n", current->comm);
drivers/sbus/char/envctrl.c
998
printk(KERN_INFO PFX "%s starting...\n", current->comm);
drivers/sbus/char/oradax.c
378
if (ctx->owner != current) {
drivers/sbus/char/oradax.c
549
if (ctx->client != current)
drivers/sbus/char/oradax.c
582
if (ctx->owner != current)
drivers/sbus/char/oradax.c
608
ctx->client = current;
drivers/sbus/char/oradax.c
625
ctx->client = current;
drivers/sbus/char/oradax.c
669
ctx->owner = current;
drivers/sbus/char/oradax.c
854
ctx->client = current;
drivers/sbus/char/oradax.c
858
if (ctx->owner != current) {
drivers/scsi/aacraid/commsup.c
1489
if (aac->thread && aac->thread->pid != current->pid) {
drivers/scsi/aacraid/commsup.c
2303
DECLARE_WAITQUEUE(wait, current);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
1937
if (signal_pending(current))
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
1938
flush_signals(current);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
485
set_user_nice(current, MIN_NICE);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
648
set_user_nice(current, MIN_NICE);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
977
if (signal_pending(current))
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
978
flush_signals(current);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
284
if (signal_pending(current))
drivers/scsi/bnx2fc/bnx2fc_tgt.c
285
flush_signals(current);
drivers/scsi/bnx2fc/bnx2fc_tgt.c
75
if (signal_pending(current))
drivers/scsi/bnx2fc/bnx2fc_tgt.c
76
flush_signals(current);
drivers/scsi/bnx2i/bnx2i_hwi.c
1858
set_user_nice(current, MIN_NICE);
drivers/scsi/bnx2i/bnx2i_iscsi.c
1627
if (signal_pending(current))
drivers/scsi/bnx2i/bnx2i_iscsi.c
1628
flush_signals(current);
drivers/scsi/bnx2i/bnx2i_iscsi.c
1750
if (signal_pending(current))
drivers/scsi/bnx2i/bnx2i_iscsi.c
1751
flush_signals(current);
drivers/scsi/bnx2i/bnx2i_iscsi.c
1862
if (signal_pending(current))
drivers/scsi/bnx2i/bnx2i_iscsi.c
1863
flush_signals(current);
drivers/scsi/bnx2i/bnx2i_iscsi.c
2101
if (signal_pending(current))
drivers/scsi/bnx2i/bnx2i_iscsi.c
2102
flush_signals(current);
drivers/scsi/ibmvscsi/ibmvfc.c
5771
set_user_nice(current, MIN_NICE);
drivers/scsi/ibmvscsi/ibmvscsi.c
2206
set_user_nice(current, MIN_NICE);
drivers/scsi/libfc/fc_exch.c
602
while (ep->resp_active && ep->resp_task != current) {
drivers/scsi/libfc/fc_exch.c
738
if (ep->resp_task != current)
drivers/scsi/libfc/fc_exch.c
739
ep->resp_task = !ep->resp_task ? current : NULL;
drivers/scsi/libiscsi.c
1948
if (signal_pending(current))
drivers/scsi/libiscsi.c
1949
flush_signals(current);
drivers/scsi/libiscsi.c
2641
if (signal_pending(current))
drivers/scsi/libiscsi.c
2642
flush_signals(current);
drivers/scsi/libiscsi.c
2960
if (signal_pending(current))
drivers/scsi/libiscsi.c
2961
flush_signals(current);
drivers/scsi/libsas/sas_discover.c
475
task_pid_nr(current));
drivers/scsi/libsas/sas_discover.c
509
task_pid_nr(current), error);
drivers/scsi/libsas/sas_discover.c
524
port->id, task_pid_nr(current));
drivers/scsi/libsas/sas_discover.c
531
task_pid_nr(current));
drivers/scsi/libsas/sas_discover.c
537
port->id, task_pid_nr(current), res);
drivers/scsi/libsas/sas_scsi_host.c
467
if (current != host->ehandler)
drivers/scsi/libsas/sas_scsi_host.c
490
if (current != host->ehandler)
drivers/scsi/lpfc/lpfc_bsg.c
2618
evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid,
drivers/scsi/lpfc/lpfc_bsg.c
3147
evt = lpfc_bsg_event_new(FC_REG_CT_EVENT, current->pid,
drivers/scsi/lpfc/lpfc_hbadisc.c
1101
set_user_nice(current, MIN_NICE);
drivers/scsi/lpfc/lpfc_hbadisc.c
1102
current->flags |= PF_NOFREEZE;
drivers/scsi/lpfc/lpfc_vport.c
149
if (signal_pending(current)) {
drivers/scsi/qedi/qedi_main.c
1923
set_user_nice(current, -20);
drivers/scsi/qedi/qedi_main.c
808
set_user_nice(current, -20);
drivers/scsi/qla2xxx/qla_mbx.c
525
} else if (current == ha->dpc_thread) {
drivers/scsi/qla2xxx/qla_os.c
6892
set_user_nice(current, MIN_NICE);
drivers/scsi/scsi_error.c
1381
current->comm));
drivers/scsi/scsi_error.c
1395
current->comm));
drivers/scsi/scsi_error.c
1497
current->comm));
drivers/scsi/scsi_error.c
1568
current->comm));
drivers/scsi/scsi_error.c
1586
current->comm));
drivers/scsi/scsi_error.c
1602
current->comm));
drivers/scsi/scsi_error.c
1635
current->comm));
drivers/scsi/scsi_error.c
1651
"%s: Sending BDR\n", current->comm));
drivers/scsi/scsi_error.c
1668
"%s: BDR failed\n", current->comm));
drivers/scsi/scsi_error.c
1705
current->comm));
drivers/scsi/scsi_error.c
1715
current->comm, id));
drivers/scsi/scsi_error.c
1722
current->comm, id));
drivers/scsi/scsi_error.c
1768
current->comm));
drivers/scsi/scsi_error.c
1789
current->comm, channel));
drivers/scsi/scsi_error.c
1806
current->comm, channel));
drivers/scsi/scsi_error.c
1833
current->comm));
drivers/scsi/scsi_error.c
1846
current->comm));
drivers/scsi/scsi_error.c
2267
current->comm));
drivers/scsi/scsi_error.c
2282
current->comm));
drivers/scsi/scsi_ioctl.c
897
"ioctl, please convert it to SG_IO\n", current->comm);
drivers/scsi/scsi_lib.c
2827
WARN_ON_ONCE(sdev->quiesced_by && sdev->quiesced_by != current);
drivers/scsi/scsi_lib.c
2829
if (sdev->quiesced_by == current)
drivers/scsi/scsi_lib.c
2847
sdev->quiesced_by = current;
drivers/scsi/sg.c
229
caller, task_tgid_vnr(current), current->comm);
drivers/scsi/sg.c
702
current->comm);
drivers/scsi/st.c
471
if (signal_pending(current))
drivers/scsi/st.c
619
if (signal_pending(current))
drivers/scsi/xen-scsifront.c
1060
info->curr = current;
drivers/scsi/xen-scsifront.c
747
if (current == info->curr) {
drivers/scsi/xen-scsifront.c
765
if (current == info->curr) {
drivers/soc/qcom/spm.c
26
#define FIELD_SET(current, mask, val) \
drivers/soc/qcom/spm.c
27
(((current) & ~(mask)) | FIELD_PREP((mask), (val)))
drivers/staging/greybus/uart.c
639
DECLARE_WAITQUEUE(wait, current);
drivers/staging/greybus/uart.c
668
} else if (signal_pending(current)) {
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
154
&state->pending, &state->current, &state->complete,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
227
state->complete = state->current;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
228
state->current = state->pending;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.h
32
struct sun6i_isp_buffer *current;
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
413
if (signal_pending(current)) {
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
414
flush_signals(current);
drivers/staging/rtl8723bs/include/osdep_service.h
67
if (signal_pending(current))
drivers/staging/rtl8723bs/include/osdep_service.h
68
flush_signals(current);
drivers/staging/rtl8723bs/os_dep/sdio_intf.c
56
rtw_sdio_set_irq_thd(psdpriv, current);
drivers/staging/rtl8723bs/os_dep/sdio_ops_linux.c
15
if (sdio_data->sys_sdio_irq_thd && sdio_data->sys_sdio_irq_thd == current)
drivers/target/iscsi/cxgbit/cxgbit_target.c
1615
if (signal_pending(current))
drivers/target/iscsi/cxgbit/cxgbit_target.c
1650
iscsit_thread_check_cpumask(conn, current, 0);
drivers/target/iscsi/cxgbit/cxgbit_target.c
320
signal_pending(current))) {
drivers/target/iscsi/iscsi_target.c
3879
iscsit_thread_check_cpumask(conn, current, 1);
drivers/target/iscsi/iscsi_target.c
3884
if (signal_pending(current))
drivers/target/iscsi/iscsi_target.c
4025
iscsit_thread_check_cpumask(conn, current, 0);
drivers/target/iscsi/iscsi_target.c
4132
if (!signal_pending(current))
drivers/target/iscsi/iscsi_target.c
4246
if (!strcmp(current->comm, ISCSI_RX_THREAD_NAME)) {
drivers/target/iscsi/iscsi_target.c
4252
} else if (!strcmp(current->comm, ISCSI_TX_THREAD_NAME)) {
drivers/target/iscsi/iscsi_target_login.c
1158
flush_signals(current);
drivers/target/iscsi/iscsi_target_login.c
1207
iscsit_start_login_timer(conn, current);
drivers/target/iscsi/iscsi_target_nego.c
550
conn, current->comm, current->pid);
drivers/target/iscsi/iscsi_target_nego.c
585
rc = iscsit_set_login_timer_kworker(conn, current);
drivers/target/iscsi/iscsi_target_nego.c
593
flush_signals(current);
drivers/target/iscsi/iscsi_target_nego.c
599
conn, current->comm, current->pid);
drivers/target/iscsi/iscsi_target_util.c
139
if (signal_pending_state(state, current))
drivers/tee/optee/call.c
627
struct mm_struct *mm = current->mm;
drivers/tty/amiserial.c
1155
if (signal_pending(current)) {
drivers/tty/amiserial.c
1308
if (signal_pending(current))
drivers/tty/mxser.c
1392
if (signal_pending(current))
drivers/tty/n_gsm.c
2594
if (signal_pending(current))
drivers/tty/n_hdlc.c
431
DECLARE_WAITQUEUE(wait, current);
drivers/tty/n_hdlc.c
462
if (signal_pending(current)) {
drivers/tty/n_hdlc.c
521
DECLARE_WAITQUEUE(wait, current);
drivers/tty/n_hdlc.c
549
if (signal_pending(current)) {
drivers/tty/n_tty.c
2154
if (signal_pending(current))
drivers/tty/n_tty.c
2356
if (signal_pending(current)) {
drivers/tty/serial/max3100.c
295
!freezing(current) &&
drivers/tty/serial/max3100.c
306
if (!s->force_end_work && !freezing(current) && !s->suspending)
drivers/tty/serial/serial_core.c
1018
current->comm,
drivers/tty/serial/serial_core.c
1209
DECLARE_WAITQUEUE(wait, current);
drivers/tty/serial/serial_core.c
1242
if (signal_pending(current)) {
drivers/tty/serial/serial_core.c
1806
if (signal_pending(current))
drivers/tty/synclink_gt.c
2656
DECLARE_WAITQUEUE(wait, current);
drivers/tty/synclink_gt.c
2697
if (signal_pending(current)) {
drivers/tty/synclink_gt.c
2921
init_waitqueue_entry(&w->wait, current);
drivers/tty/synclink_gt.c
3006
if (signal_pending(current))
drivers/tty/synclink_gt.c
3029
DECLARE_WAITQUEUE(wait, current);
drivers/tty/synclink_gt.c
3040
if (signal_pending(current)) {
drivers/tty/synclink_gt.c
3161
DECLARE_WAITQUEUE(wait, current);
drivers/tty/synclink_gt.c
3210
if (signal_pending(current)) {
drivers/tty/synclink_gt.c
854
if (signal_pending(current))
drivers/tty/tty_audit.c
118
buf = xchg(&current->signal->tty_audit_buf, ERR_PTR(-ESRCH));
drivers/tty/tty_audit.c
133
sig->audit_tty = current->signal->audit_tty;
drivers/tty/tty_audit.c
160
if (~current->signal->audit_tty & AUDIT_TTY_ENABLE)
drivers/tty/tty_audit.c
194
if (cmpxchg(&current->signal->tty_audit_buf, NULL, buf) != NULL)
drivers/tty/tty_audit.c
212
audit_tty = READ_ONCE(current->signal->audit_tty);
drivers/tty/tty_audit.c
29
buf = current->signal->tty_audit_buf;
drivers/tty/tty_audit.c
67
pid_t pid = task_pid_nr(current);
drivers/tty/tty_audit.c
68
uid_t uid = from_kuid(&init_user_ns, task_uid(current));
drivers/tty/tty_audit.c
69
uid_t loginuid = from_kuid(&init_user_ns, audit_get_loginuid(current));
drivers/tty/tty_audit.c
70
unsigned int sessionid = audit_get_sessionid(current);
drivers/tty/tty_audit.c
80
get_task_comm(name, current);
drivers/tty/tty_io.c
1022
if (signal_pending(current))
drivers/tty/tty_io.c
2125
if (retval != -EAGAIN || signal_pending(current))
drivers/tty/tty_io.c
2150
if (signal_pending(current))
drivers/tty/tty_io.c
2232
pid = task_pid(current);
drivers/tty/tty_io.c
2280
if ((current->signal->tty != tty) && !capable(CAP_SYS_ADMIN))
drivers/tty/tty_io.c
2492
if (signal_pending(current))
drivers/tty/tty_io.c
2621
__func__, current->comm, flags);
drivers/tty/tty_io.c
2695
if (signal_pending(current))
drivers/tty/tty_ioctl.c
506
if (signal_pending(current)) {
drivers/tty/tty_jobctrl.c
109
tty->ctrl.pgrp = get_pid(task_pgrp(current));
drivers/tty/tty_jobctrl.c
110
tty->ctrl.session = get_pid(task_session(current));
drivers/tty/tty_jobctrl.c
112
if (current->signal->tty) {
drivers/tty/tty_jobctrl.c
114
current->signal->tty->name);
drivers/tty/tty_jobctrl.c
115
tty_kref_put(current->signal->tty);
drivers/tty/tty_jobctrl.c
117
put_pid(current->signal->tty_old_pgrp);
drivers/tty/tty_jobctrl.c
118
current->signal->tty = tty_kref_get(tty);
drivers/tty/tty_jobctrl.c
119
current->signal->tty_old_pgrp = NULL;
drivers/tty/tty_jobctrl.c
124
spin_lock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
126
spin_unlock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
135
spin_lock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
136
if (current->signal->leader &&
drivers/tty/tty_jobctrl.c
137
!current->signal->tty &&
drivers/tty/tty_jobctrl.c
156
spin_unlock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
165
spin_lock_irqsave(&current->sighand->siglock, flags);
drivers/tty/tty_jobctrl.c
166
tty = tty_kref_get(current->signal->tty);
drivers/tty/tty_jobctrl.c
167
spin_unlock_irqrestore(&current->sighand->siglock, flags);
drivers/tty/tty_jobctrl.c
18
return (sigismember(&current->blocked, sig) ||
drivers/tty/tty_jobctrl.c
19
current->sighand->action[sig-1].sa.sa_handler == SIG_IGN);
drivers/tty/tty_jobctrl.c
269
if (!current->signal->leader)
drivers/tty/tty_jobctrl.c
291
spin_lock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
292
old_pgrp = current->signal->tty_old_pgrp;
drivers/tty/tty_jobctrl.c
293
current->signal->tty_old_pgrp = NULL;
drivers/tty/tty_jobctrl.c
294
spin_unlock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
323
spin_lock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
324
put_pid(current->signal->tty_old_pgrp);
drivers/tty/tty_jobctrl.c
325
current->signal->tty_old_pgrp = NULL;
drivers/tty/tty_jobctrl.c
326
spin_unlock_irq(&current->sighand->siglock);
drivers/tty/tty_jobctrl.c
330
session_clear_tty(task_session(current));
drivers/tty/tty_jobctrl.c
345
struct task_struct *tsk = current;
drivers/tty/tty_jobctrl.c
372
if (current->signal->leader &&
drivers/tty/tty_jobctrl.c
373
task_session(current) == tty->ctrl.session)
drivers/tty/tty_jobctrl.c
380
if (!current->signal->leader || current->signal->tty) {
drivers/tty/tty_jobctrl.c
39
if (current->signal->tty != tty)
drivers/tty/tty_jobctrl.c
43
pgrp = task_pgrp(current);
drivers/tty/tty_jobctrl.c
474
if (tty == real_tty && current->signal->tty != real_tty)
drivers/tty/tty_jobctrl.c
510
if (!current->signal->tty ||
drivers/tty/tty_jobctrl.c
511
(current->signal->tty != real_tty) ||
drivers/tty/tty_jobctrl.c
512
(real_tty->ctrl.session != task_session(current))) {
drivers/tty/tty_jobctrl.c
522
if (session_of_pgrp(pgrp) != task_session(current))
drivers/tty/tty_jobctrl.c
552
if (tty == real_tty && current->signal->tty != real_tty)
drivers/tty/tty_jobctrl.c
579
if (current->signal->tty != tty)
drivers/tty/tty_ldsem.c
183
waiter.task = current;
drivers/tty/tty_ldsem.c
184
get_task_struct(current);
drivers/tty/tty_ldsem.c
255
waiter.task = current;
drivers/tty/tty_port.c
540
if (signal_pending(current)) {
drivers/tty/vt/selection.c
397
DECLARE_WAITQUEUE(wait, current);
drivers/tty/vt/selection.c
418
if (signal_pending(current)) {
drivers/tty/vt/vt.c
3493
if (current->signal->tty != tty && !capable(CAP_SYS_ADMIN))
drivers/tty/vt/vt_ioctl.c
1069
if (current->signal->tty == tty || capable(CAP_SYS_TTY_CONFIG))
drivers/tty/vt/vt_ioctl.c
460
vt_spawn_con.pid = get_pid(task_pid(current));
drivers/tty/vt/vt_ioctl.c
615
nvc->vt_pid = get_pid(task_pid(current));
drivers/tty/vt/vt_ioctl.c
740
if (current->signal->tty == tty || capable(CAP_SYS_TTY_CONFIG))
drivers/tty/vt/vt_ioctl.c
770
vc->vt_pid = get_pid(task_pid(current));
drivers/uio/uio.c
585
DECLARE_WAITQUEUE(wait, current);
drivers/uio/uio.c
622
if (signal_pending(current)) {
drivers/uio/uio_pci_generic_sva.c
39
handle = iommu_sva_bind_device(&udev->pdev->dev, current->mm);
drivers/usb/class/cdc-acm.c
1014
DECLARE_WAITQUEUE(wait, current);
drivers/usb/class/cdc-acm.c
1044
if (signal_pending(current))
drivers/usb/class/usblp.c
1005
if (signal_pending(current))
drivers/usb/class/usblp.c
910
DECLARE_WAITQUEUE(waita, current);
drivers/usb/class/usblp.c
952
if (signal_pending(current))
drivers/usb/class/usblp.c
974
DECLARE_WAITQUEUE(waita, current);
drivers/usb/core/devio.c
1072
ps->disc_pid = get_pid(task_pid(current));
drivers/usb/core/devio.c
1080
snoop(&dev->dev, "opened by process %d: %s\n", task_pid_nr(current),
drivers/usb/core/devio.c
1081
current->comm);
drivers/usb/core/devio.c
1260
current->comm, ctrl->bRequestType, ctrl->bRequest,
drivers/usb/core/devio.c
1392
task_pid_nr(current), current->comm,
drivers/usb/core/devio.c
1520
number, interface->dev.driver->name, current->comm);
drivers/usb/core/devio.c
1574
current->comm, u);
drivers/usb/core/devio.c
1932
as->pid = get_pid(task_pid(current));
drivers/usb/core/devio.c
2082
DECLARE_WAITQUEUE(wait, current);
drivers/usb/core/devio.c
2092
if (signal_pending(current))
drivers/usb/core/devio.c
2115
if (signal_pending(current))
drivers/usb/core/devio.c
2260
if (signal_pending(current))
drivers/usb/core/devio.c
2430
portnum, task_pid_nr(current), current->comm);
drivers/usb/core/devio.c
851
"interface %u before use\n", task_pid_nr(current),
drivers/usb/core/devio.c
852
current->comm, ifnum);
drivers/usb/core/devio.c
926
__func__, task_pid_nr(current),
drivers/usb/core/devio.c
927
current->comm, index, index ^ 0x80);
drivers/usb/core/message.c
84
current->comm,
drivers/usb/gadget/function/f_fs.c
1260
p->mm = current->mm;
drivers/usb/gadget/function/f_fs.c
1304
p->mm = current->mm;
drivers/usb/gadget/function/f_mass_storage.c
1027
if (signal_pending(current))
drivers/usb/gadget/function/f_mass_storage.c
2171
if (reply == -EINTR || signal_pending(current))
drivers/usb/gadget/function/f_mass_storage.c
2588
if (exception_in_progress(common) || signal_pending(current)) {
drivers/usb/gadget/function/f_mass_storage.c
689
if (signal_pending(current))
drivers/usb/gadget/function/f_mass_storage.c
888
if (signal_pending(current))
drivers/usb/gadget/function/f_mass_storage.c
993
if (signal_pending(current))
drivers/usb/gadget/function/f_mass_storage.c
997
if (signal_pending(current))
drivers/usb/gadget/legacy/inode.c
543
priv->mm = current->mm; /* mm teardown waits for iocbs in exit_aio() */
drivers/usb/image/mdc800.c
707
if (signal_pending (current))
drivers/usb/image/mdc800.c
801
if (signal_pending (current))
drivers/usb/misc/adutux.c
353
DECLARE_WAITQUEUE(wait, current);
drivers/usb/misc/adutux.c
464
if (signal_pending(current)) {
drivers/usb/misc/adutux.c
506
DECLARE_WAITQUEUE(waita, current);
drivers/usb/misc/adutux.c
542
if (signal_pending(current)) {
drivers/usb/misc/idmouse.c
173
if (signal_pending(current)) {
drivers/usb/misc/uss720.c
291
if (signal_pending (current))
drivers/usb/mon/mon_bin.c
1306
DECLARE_WAITQUEUE(waita, current);
drivers/usb/mon/mon_bin.c
1322
if (signal_pending(current)) {
drivers/usb/mon/mon_text.c
484
DECLARE_WAITQUEUE(waita, current);
drivers/usb/mon/mon_text.c
500
if (signal_pending(current)) {
drivers/usb/serial/digi_acceleport.c
386
if (interruptible && signal_pending(current))
drivers/usb/serial/digi_acceleport.c
449
if (signal_pending(current))
drivers/usb/serial/digi_acceleport.c
527
if (interruptible && signal_pending(current))
drivers/usb/serial/digi_acceleport.c
599
if (signal_pending(current))
drivers/usb/serial/generic.c
291
if (signal_pending(current))
drivers/usb/usbip/stub_dev.c
478
if (usbip_in_eh(current))
drivers/usb/usbip/usbip_event.c
67
worker_context = current;
drivers/vdpa/mlx5/core/mlx5_vdpa.h
172
current->pid, ##__VA_ARGS__)
drivers/vdpa/mlx5/core/mlx5_vdpa.h
177
current->pid, ##__VA_ARGS__)
drivers/vdpa/mlx5/core/mlx5_vdpa.h
181
current->pid, ##__VA_ARGS__)
drivers/vdpa/mlx5/core/mlx5_vdpa.h
185
current->pid, ##__VA_ARGS__)
drivers/vdpa/vdpa_user/vduse_dev.c
1239
mmap_read_lock(current->mm);
drivers/vdpa/vdpa_user/vduse_dev.c
1242
if (npages + atomic64_read(&current->mm->pinned_vm) > lock_limit)
drivers/vdpa/vdpa_user/vduse_dev.c
1257
atomic64_add(npages, &current->mm->pinned_vm);
drivers/vdpa/vdpa_user/vduse_dev.c
1262
umem->mm = current->mm;
drivers/vdpa/vdpa_user/vduse_dev.c
1263
mmgrab(current->mm);
drivers/vdpa/vdpa_user/vduse_dev.c
1270
mmap_read_unlock(current->mm);
drivers/vfio/group.c
283
"(%s:%d)\n", current->comm, task_pid_nr(current));
drivers/vfio/vfio_iommu_spapr_tce.c
78
if (container->mm == current->mm)
drivers/vfio/vfio_iommu_spapr_tce.c
82
BUG_ON(!current->mm);
drivers/vfio/vfio_iommu_spapr_tce.c
83
container->mm = current->mm;
drivers/vfio/vfio_iommu_spapr_tce.c
853
if (container->mm && container->mm != current->mm)
drivers/vfio/vfio_iommu_type1.c
1651
struct task_struct *task = current->group_leader;
drivers/vfio/vfio_iommu_type1.c
1652
struct mm_struct *mm = current->mm;
drivers/vfio/vfio_iommu_type1.c
1776
get_task_struct(current->group_leader);
drivers/vfio/vfio_iommu_type1.c
1777
dma->task = current->group_leader;
drivers/vfio/vfio_iommu_type1.c
1779
dma->mm = current->mm;
drivers/vfio/vfio_iommu_type1.c
3147
bool kthread = current->mm == NULL;
drivers/vfio/vfio_iommu_type1.c
3166
else if (current->mm != mm)
drivers/vfio/vfio_iommu_type1.c
681
struct mm_struct *mm = current->mm;
drivers/vfio/vfio_main.c
1757
if (!current->mm)
drivers/vfio/vfio_main.c
424
current->comm, task_pid_nr(current));
drivers/vhost/net.c
430
!signal_pending(current));
drivers/vhost/vhost.c
1774
if (signal_pending(current)) {
drivers/vhost/vhost.c
629
return dev->mm == current->mm ? 0 : -EPERM;
drivers/vhost/vhost.c
644
s->ret = cgroup_attach_task_all(s->owner, current);
drivers/vhost/vhost.c
652
attach.owner = current;
drivers/vhost/vhost.c
684
dev->mm = get_task_mm(current);
drivers/vhost/vhost.c
692
dev->mm = current->mm;
drivers/vhost/vhost.c
841
snprintf(name, sizeof(name), "vhost-%d", current->pid);
drivers/vhost/vsock.c
689
net = current->nsproxy->net_ns;
drivers/video/fbdev/cobalt_lcdfb.c
163
if (retval < 0 && signal_pending(current))
drivers/video/fbdev/cobalt_lcdfb.c
215
if (retval < 0 && signal_pending(current))
drivers/video/fbdev/omap2/omapfb/displays/panel-dsi-cm.c
1051
if (signal_pending(current)) {
drivers/video/fbdev/sa1100fb.c
815
DECLARE_WAITQUEUE(wait, current);
drivers/virt/acrn/mm.c
176
mmap_read_lock(current->mm);
drivers/virt/acrn/mm.c
177
vma = vma_lookup(current->mm, memmap->vma_base);
drivers/virt/acrn/mm.c
183
mmap_read_unlock(current->mm);
drivers/virt/acrn/mm.c
223
mmap_read_unlock(current->mm);
drivers/virt/acrn/mm.c
235
mmap_read_unlock(current->mm);
drivers/virt/nitro_enclaves/ne_misc_dev.c
1685
ne_enclave->mm = current->mm;
drivers/virt/nitro_enclaves/ne_misc_dev.c
712
if (ne_enclave->mm != current->mm)
drivers/virt/nitro_enclaves/ne_misc_dev.c
752
if (ne_enclave->mm != current->mm)
drivers/w1/w1_family.c
78
flush_signals(current);
drivers/w1/w1_int.c
198
flush_signals(current);
drivers/xen/evtchn.c
649
u->name = kasprintf(GFP_KERNEL, "evtchn:%s", current->comm);
drivers/xen/gntdev.c
716
mmap_read_lock(current->mm);
drivers/xen/gntdev.c
717
vma = find_vma(current->mm, op.vaddr);
drivers/xen/gntdev.c
730
mmap_read_unlock(current->mm);
drivers/xen/grant-table.c
1166
pr_err("%s: %s eagain grant\n", func, current->comm);
drivers/xen/privcmd.c
1238
struct mm_struct *mm = current->mm;
drivers/xen/privcmd.c
282
struct mm_struct *mm = current->mm;
drivers/xen/privcmd.c
476
struct mm_struct *mm = current->mm;
drivers/xen/privcmd.c
753
struct mm_struct *mm = current->mm;
drivers/xen/xenbus/xenbus_xs.c
786
if (current->pid != xenwatch_pid)
drivers/xen/xenbus/xenbus_xs.c
802
if (current->pid != xenwatch_pid)
drivers/xen/xenbus/xenbus_xs.c
847
xenwatch_pid = current->pid;
fs/afs/fs_operation.c
112
struct afs_io_locker myself = { .task = current, };
fs/afs/fs_operation.c
128
signal_pending(current))
fs/afs/fs_operation.c
137
if (unlikely(signal_pending(current))) {
fs/afs/fs_operation.c
85
struct afs_io_locker myself = { .task = current, };
fs/afs/fs_probe.c
356
if (!still_probing || signal_pending(current))
fs/afs/fs_probe.c
367
if (!ret && signal_pending(current))
fs/afs/fs_probe.c
513
(is_intr && signal_pending(current)))
fs/afs/fs_probe.c
525
if (is_intr && signal_pending(current))
fs/afs/internal.h
1835
printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__)
fs/afs/rotate.c
137
if (signal_pending(current)) {
fs/afs/rotate.c
68
init_waitqueue_entry(&s->probe_waiter, current);
fs/afs/rxrpc.c
630
DECLARE_WAITQUEUE(myself, current);
fs/afs/vl_probe.c
259
init_waitqueue_entry(&waits[i], current);
fs/afs/vl_probe.c
278
if (!still_probing || signal_pending(current))
fs/afs/vl_probe.c
302
if (pref == -1 && signal_pending(current))
fs/afs/vl_rotate.c
28
if (signal_pending(current)) {
fs/aio.c
1077
struct mm_struct *mm = current->mm;
fs/aio.c
1339
hrtimer_set_expires_range_ns(&t.timer, until, current->timer_slack_ns);
fs/aio.c
1403
kill_ioctx(current->mm, ioctx, NULL);
fs/aio.c
1435
kill_ioctx(current->mm, ioctx, NULL);
fs/aio.c
1464
ret = kill_ioctx(current->mm, ioctx, &wait);
fs/aio.c
2263
if (!ret && signal_pending(current))
fs/aio.c
2300
interrupted = signal_pending(current);
fs/aio.c
2336
interrupted = signal_pending(current);
fs/aio.c
2361
if (!ret && signal_pending(current))
fs/aio.c
2402
interrupted = signal_pending(current);
fs/aio.c
2437
interrupted = signal_pending(current);
fs/aio.c
342
pr_debug("pid(%d) [%d] folio->count=%d\n", current->pid, i,
fs/aio.c
485
struct mm_struct *mm = current->mm;
fs/aio.c
529
pr_debug("pid(%d) [%d] folio->count=%d\n", current->pid, i,
fs/aio.c
734
struct mm_struct *mm = current->mm;
fs/attr.c
278
send_sig(SIGXFSZ, current, 0);
fs/autofs/autofs_i.h
157
task_pgrp(current) == sbi->oz_pgrp);
fs/autofs/autofs_i.h
46
#define pr_fmt(fmt) KBUILD_MODNAME ":pid:%d:%s: " fmt, current->pid, __func__
fs/autofs/dev-ioctl.c
345
new_pid = get_task_pid(current, PIDTYPE_PGID);
fs/autofs/dev-ioctl.c
366
sbi->mnt_ns_id = to_ns_common(current->nsproxy->mnt_ns)->ns_id;
fs/autofs/inode.c
254
sbi->mnt_ns_id = to_ns_common(current->nsproxy->mnt_ns)->ns_id;
fs/autofs/inode.c
347
sbi->oz_pgrp = get_task_pid(current, PIDTYPE_PGID);
fs/autofs/root.c
347
if (sbi->mnt_ns_id != to_ns_common(current->nsproxy->mnt_ns)->ns_id) {
fs/autofs/root.c
517
current->pid, task_pgrp_nr(current),
fs/autofs/root.c
866
cmd, arg, sbi, task_pgrp_nr(current));
fs/autofs/waitq.c
324
pid = task_pid_nr_ns(current, ns_of_pid(sbi->oz_pgrp));
fs/autofs/waitq.c
325
tgid = task_tgid_nr_ns(current, ns_of_pid(sbi->oz_pgrp));
fs/autofs/waitq.c
53
sigpipe = sigismember(&current->pending.signal, SIGPIPE);
fs/autofs/waitq.c
69
spin_lock_irqsave(&current->sighand->siglock, flags);
fs/autofs/waitq.c
70
sigdelset(&current->pending.signal, SIGPIPE);
fs/autofs/waitq.c
72
spin_unlock_irqrestore(&current->sighand->siglock, flags);
fs/binfmt_elf.c
1019
current->personality |= READ_IMPLIES_EXEC;
fs/binfmt_elf.c
1022
if (!(current->personality & ADDR_NO_RANDOMIZE) && snapshot_randomize_va_space)
fs/binfmt_elf.c
1023
current->flags |= PF_RANDOMIZE;
fs/binfmt_elf.c
1142
if (current->flags & PF_RANDOMIZE)
fs/binfmt_elf.c
1302
mm = current->mm;
fs/binfmt_elf.c
1333
if ((current->flags & PF_RANDOMIZE) && snapshot_randomize_va_space > 1) {
fs/binfmt_elf.c
1347
current->brk_randomized = 1;
fs/binfmt_elf.c
1350
if (current->personality & MMAP_PAGE_ZERO) {
fs/binfmt_elf.c
1361
task_pid_nr(current), retval);
fs/binfmt_elf.c
169
struct mm_struct *mm = current->mm;
fs/binfmt_elf.c
1820
struct task_struct *dump_task = current;
fs/binfmt_elf.c
1908
fill_auxv_note(&info->auxv, current->mm);
fs/binfmt_elf.c
405
task_pid_nr(current), current->comm, (void *)addr);
fs/binfmt_elf_fdpic.c
1502
for (ct = current->signal->core_state->dumper.next;
fs/binfmt_elf_fdpic.c
1515
current, &thread_status_size);
fs/binfmt_elf_fdpic.c
1540
fill_psinfo(psinfo, current->group_leader, current->mm);
fs/binfmt_elf_fdpic.c
1544
auxv = (elf_addr_t *) current->mm->saved_auxv;
fs/binfmt_elf_fdpic.c
200
kdebug("____ LOAD %d ____", current->pid);
fs/binfmt_elf_fdpic.c
350
current->personality |= PER_LINUX_FDPIC;
fs/binfmt_elf_fdpic.c
352
current->personality |= READ_IMPLIES_EXEC;
fs/binfmt_elf_fdpic.c
358
current->mm->start_code = 0;
fs/binfmt_elf_fdpic.c
359
current->mm->end_code = 0;
fs/binfmt_elf_fdpic.c
360
current->mm->start_stack = 0;
fs/binfmt_elf_fdpic.c
361
current->mm->start_data = 0;
fs/binfmt_elf_fdpic.c
362
current->mm->end_data = 0;
fs/binfmt_elf_fdpic.c
363
current->mm->context.exec_fdpic_loadmap = 0;
fs/binfmt_elf_fdpic.c
364
current->mm->context.interp_fdpic_loadmap = 0;
fs/binfmt_elf_fdpic.c
369
&current->mm->start_stack,
fs/binfmt_elf_fdpic.c
370
&current->mm->start_brk);
fs/binfmt_elf_fdpic.c
372
retval = setup_arg_pages(bprm, current->mm->start_stack,
fs/binfmt_elf_fdpic.c
384
retval = elf_fdpic_map_file(&exec_params, bprm->file, current->mm,
fs/binfmt_elf_fdpic.c
391
current->mm, "interpreter");
fs/binfmt_elf_fdpic.c
403
if (!current->mm->start_brk)
fs/binfmt_elf_fdpic.c
404
current->mm->start_brk = current->mm->end_data;
fs/binfmt_elf_fdpic.c
406
current->mm->brk = current->mm->start_brk =
fs/binfmt_elf_fdpic.c
407
PAGE_ALIGN(current->mm->start_brk);
fs/binfmt_elf_fdpic.c
420
current->mm->start_brk = vm_mmap(NULL, 0, stack_size, stack_prot,
fs/binfmt_elf_fdpic.c
425
if (IS_ERR_VALUE(current->mm->start_brk)) {
fs/binfmt_elf_fdpic.c
426
retval = current->mm->start_brk;
fs/binfmt_elf_fdpic.c
427
current->mm->start_brk = 0;
fs/binfmt_elf_fdpic.c
431
current->mm->brk = current->mm->start_brk;
fs/binfmt_elf_fdpic.c
432
current->mm->context.end_brk = current->mm->start_brk;
fs/binfmt_elf_fdpic.c
433
current->mm->start_stack = current->mm->start_brk + stack_size;
fs/binfmt_elf_fdpic.c
436
retval = create_elf_fdpic_tables(bprm, current->mm, &exec_params,
fs/binfmt_elf_fdpic.c
441
kdebug("- start_code %lx", current->mm->start_code);
fs/binfmt_elf_fdpic.c
442
kdebug("- end_code %lx", current->mm->end_code);
fs/binfmt_elf_fdpic.c
443
kdebug("- start_data %lx", current->mm->start_data);
fs/binfmt_elf_fdpic.c
444
kdebug("- end_data %lx", current->mm->end_data);
fs/binfmt_elf_fdpic.c
445
kdebug("- start_brk %lx", current->mm->start_brk);
fs/binfmt_elf_fdpic.c
446
kdebug("- brk %lx", current->mm->brk);
fs/binfmt_elf_fdpic.c
447
kdebug("- start_stack %lx", current->mm->start_stack);
fs/binfmt_elf_fdpic.c
464
start_thread(regs, entryaddr, current->mm->start_stack);
fs/binfmt_elf_fdpic.c
571
current->mm->context.exec_fdpic_loadmap = (unsigned long) sp;
fs/binfmt_elf_fdpic.c
584
current->mm->context.interp_fdpic_loadmap = (unsigned long) sp;
fs/binfmt_elf_fdpic.c
697
current->mm->arg_start = bprm->p;
fs/binfmt_elf_fdpic.c
699
current->mm->arg_start = current->mm->start_stack -
fs/binfmt_elf_fdpic.c
703
p = (char __user *) current->mm->arg_start;
fs/binfmt_elf_fdpic.c
714
current->mm->arg_end = (unsigned long) p;
fs/binfmt_elf_fdpic.c
717
current->mm->env_start = (unsigned long) p;
fs/binfmt_elf_fdpic.c
728
current->mm->env_end = (unsigned long) p;
fs/binfmt_flat.c
115
sp = (unsigned long __user *)current->mm->start_stack;
fs/binfmt_flat.c
123
current->mm->start_stack = (unsigned long)sp & -FLAT_STACK_ALIGN;
fs/binfmt_flat.c
124
sp = (unsigned long __user *)current->mm->start_stack;
fs/binfmt_flat.c
136
current->mm->arg_start = (unsigned long)p;
fs/binfmt_flat.c
147
current->mm->arg_end = (unsigned long)p;
fs/binfmt_flat.c
149
current->mm->env_start = (unsigned long) p;
fs/binfmt_flat.c
160
current->mm->env_end = (unsigned long)p;
fs/binfmt_flat.c
332
pr_cont(", killing %s!\n", current->comm);
fs/binfmt_flat.c
333
send_sig(SIGSEGV, current, 0);
fs/binfmt_flat.c
350
ptr = (unsigned long __user *)(current->mm->start_code + r.reloc.offset);
fs/binfmt_flat.c
352
ptr = (unsigned long __user *)(current->mm->start_data + r.reloc.offset);
fs/binfmt_flat.c
362
val += current->mm->start_code;
fs/binfmt_flat.c
365
val += current->mm->start_data;
fs/binfmt_flat.c
368
val += current->mm->end_data;
fs/binfmt_flat.c
695
current->mm->start_code = start_code;
fs/binfmt_flat.c
696
current->mm->end_code = end_code;
fs/binfmt_flat.c
697
current->mm->start_data = datapos;
fs/binfmt_flat.c
698
current->mm->end_data = datapos + data_len;
fs/binfmt_flat.c
706
current->mm->start_brk = datapos + data_len + bss_len;
fs/binfmt_flat.c
707
current->mm->brk = (current->mm->start_brk + 3) & ~3;
fs/binfmt_flat.c
709
current->mm->context.end_brk = memp + memp_size - stack_len;
fs/binfmt_flat.c
907
current->mm->start_stack =
fs/binfmt_flat.c
908
((current->mm->context.end_brk + stack_len + 3) & ~3) - 4;
fs/binfmt_flat.c
909
pr_debug("sp=%lx\n", current->mm->start_stack);
fs/binfmt_flat.c
912
res = transfer_args_to_stack(bprm, &current->mm->start_stack);
fs/binfmt_flat.c
914
res = create_flat_tables(bprm, current->mm->start_stack);
fs/binfmt_flat.c
931
regs, start_addr, current->mm->start_stack);
fs/binfmt_flat.c
932
start_thread(regs, start_addr, current->mm->start_stack);
fs/btrfs/backref.c
1249
if (!current->journal_info)
fs/btrfs/backref.c
1320
if (!current->journal_info)
fs/btrfs/defrag.h
28
return signal_pending(current);
fs/btrfs/delalloc-space.c
338
if (current->journal_info)
fs/btrfs/dev-replace.c
630
dev_replace->replace_task = current;
fs/btrfs/direct-io.c
920
ASSERT(current->journal_info == NULL);
fs/btrfs/direct-io.c
921
current->journal_info = BTRFS_TRANS_DIO_WRITE_STUB;
fs/btrfs/direct-io.c
923
current->journal_info = NULL;
fs/btrfs/disk-io.c
4524
if (current->flags & PF_MEMALLOC)
fs/btrfs/extent-tree.c
4377
trans = current->journal_info;
fs/btrfs/extent-tree.c
5239
if (eb->lock_owner == current->pid) {
fs/btrfs/extent-tree.c
5242
eb->start, btrfs_header_owner(eb), current->pid);
fs/btrfs/fiemap.c
796
if (fatal_signal_pending(current)) {
fs/btrfs/file.c
1570
if (current->journal_info == BTRFS_TRANS_DIO_WRITE_STUB) {
fs/btrfs/file.c
1572
current->journal_info = NULL;
fs/btrfs/file.c
3566
if (private && private->owner_task != current) {
fs/btrfs/file.c
3585
private->owner_task = current;
fs/btrfs/file.c
3758
if (fatal_signal_pending(current)) {
fs/btrfs/free-space-cache.h
62
return fatal_signal_pending(current) || freezing(current);
fs/btrfs/inode.c
10545
if (fatal_signal_pending(current)) {
fs/btrfs/ioctl.c
4398
add_rchar(current, ret);
fs/btrfs/ioctl.c
4399
inc_syscr(current);
fs/btrfs/ioctl.c
4498
add_wchar(current, ret);
fs/btrfs/ioctl.c
4499
inc_syscw(current);
fs/btrfs/ioctl.c
4583
add_rchar(current, ret);
fs/btrfs/ioctl.c
4820
add_rchar(current, ret);
fs/btrfs/ioctl.c
4821
inc_syscr(current);
fs/btrfs/ioctl.c
4948
add_wchar(current, ret);
fs/btrfs/ioctl.c
4949
inc_syscw(current);
fs/btrfs/locking.c
190
btrfs_set_eb_lock_owner(eb, current->pid);
fs/btrfs/print-tree.c
234
refcount_read(&eb->refs), eb->lock_owner, current->pid);
fs/btrfs/qgroup.c
4201
ASSERT(current->journal_info == NULL);
fs/btrfs/qgroup.c
4202
if (WARN_ON(current->journal_info))
fs/btrfs/reflink.c
568
if (fatal_signal_pending(current)) {
fs/btrfs/relocation.c
2763
fatal_signal_pending(current);
fs/btrfs/scrub.c
2111
freezing(current) || signal_pending(current))
fs/btrfs/space-info.c
1774
if (current->journal_info) {
fs/btrfs/space-info.c
1940
ASSERT(!current->journal_info || flush != BTRFS_RESERVE_FLUSH_DATA,
fs/btrfs/space-info.c
1942
(unsigned long)current->journal_info, flush);
fs/btrfs/space-info.c
745
trans = current->journal_info;
fs/btrfs/space-info.c
915
ASSERT(current->journal_info == NULL);
fs/btrfs/transaction.c
1099
if (current->journal_info == trans)
fs/btrfs/transaction.c
1100
current->journal_info = NULL;
fs/btrfs/transaction.c
2113
if (current->journal_info == trans)
fs/btrfs/transaction.c
2114
current->journal_info = NULL;
fs/btrfs/transaction.c
2624
if (current->journal_info == trans)
fs/btrfs/transaction.c
2625
current->journal_info = NULL;
fs/btrfs/transaction.c
2644
if (current->journal_info == trans)
fs/btrfs/transaction.c
2645
current->journal_info = NULL;
fs/btrfs/transaction.c
617
if (current->journal_info) {
fs/btrfs/transaction.c
619
h = current->journal_info;
fs/btrfs/transaction.c
734
current->journal_info = h;
fs/btrfs/transaction.c
757
if (!current->journal_info)
fs/btrfs/transaction.c
758
current->journal_info = h;
fs/btrfs/tree-log.c
333
root->log_start_pid = current->pid;
fs/btrfs/tree-log.c
334
} else if (root->log_start_pid != current->pid) {
fs/btrfs/tree-log.c
355
root->log_start_pid = current->pid;
fs/btrfs/volumes.c
6303
preferred_mirror = first + (current->pid % num_stripes);
fs/btrfs/volumes.c
6944
if (dev_replace->replace_task != current)
fs/btrfs/volumes.c
6952
if (!dev_replace_is_ongoing && dev_replace->replace_task != current)
fs/btrfs/volumes.c
7092
if (dev_replace_is_ongoing && dev_replace->replace_task != current) {
fs/btrfs/volumes.c
8175
current->comm, task_pid_nr(current));
fs/btrfs/volumes.c
860
fs_devices->fsid, current->comm,
fs/btrfs/volumes.c
861
task_pid_nr(current));
fs/btrfs/volumes.c
889
current->comm, task_pid_nr(current));
fs/btrfs/volumes.c
895
current->comm, task_pid_nr(current));
fs/btrfs/volumes.c
956
current->comm,
fs/btrfs/volumes.c
957
task_pid_nr(current));
fs/btrfs/volumes.c
963
path, current->comm,
fs/btrfs/volumes.c
964
task_pid_nr(current));
fs/btrfs/xattr.c
219
const bool start_trans = (current->journal_info == NULL);
fs/btrfs/xattr.c
244
trans = current->journal_info;
fs/buffer.c
2533
if (fatal_signal_pending(current)) {
fs/cachefiles/daemon.c
653
get_fs_pwd(current->fs, &path);
fs/cachefiles/daemon.c
724
get_fs_pwd(current->fs, &path);
fs/cachefiles/internal.h
456
printk(KERN_DEBUG "[%-6.6s] "FMT"\n", current->comm, ##__VA_ARGS__)
fs/cachefiles/security.c
23
new = prepare_kernel_cred(current);
fs/ceph/addr.c
754
WARN_ON(!(current->flags & PF_MEMALLOC));
fs/ceph/caps.c
3074
cw.tgid = current->tgid;
fs/ceph/caps.c
3090
if (signal_pending(current)) {
fs/ceph/mds_client.c
4004
current->journal_info = req;
fs/ceph/mds_client.c
4011
current->journal_info = NULL;
fs/ceph/super.h
898
.thread = current, \
fs/ceph/super.h
924
if (ctx->thread == current) {
fs/ceph/xattr.c
1021
if (current->journal_info) {
fs/ceph/xattr.c
1054
if (current->journal_info &&
fs/ceph/xattr.c
1302
if (current->journal_info) {
fs/ceph/xattr.c
953
struct ceph_mds_request *req = current->journal_info;
fs/coda/inode.c
361
if (task_active_pid_ns(current) != &init_pid_ns)
fs/coda/psdev.c
207
DECLARE_WAITQUEUE(wait, current);
fs/coda/psdev.c
225
if (signal_pending(current)) {
fs/coda/psdev.c
273
if (task_active_pid_ns(current) != &init_pid_ns)
fs/coda/upcall.c
54
inp->ih.pid = task_pid_nr_ns(current, &init_pid_ns);
fs/coda/upcall.c
55
inp->ih.pgid = task_pgrp_nr_ns(current, &init_pid_ns);
fs/coda/upcall.c
618
spin_lock_irq(&current->sighand->siglock);
fs/coda/upcall.c
619
*old = current->blocked;
fs/coda/upcall.c
621
sigfillset(&current->blocked);
fs/coda/upcall.c
622
sigdelset(&current->blocked, SIGKILL);
fs/coda/upcall.c
623
sigdelset(&current->blocked, SIGSTOP);
fs/coda/upcall.c
624
sigdelset(&current->blocked, SIGINT);
fs/coda/upcall.c
627
spin_unlock_irq(&current->sighand->siglock);
fs/coda/upcall.c
632
spin_lock_irq(&current->sighand->siglock);
fs/coda/upcall.c
633
current->blocked = *old;
fs/coda/upcall.c
635
spin_unlock_irq(&current->sighand->siglock);
fs/coda/upcall.c
654
DECLARE_WAITQUEUE(wait, current);
fs/coda/upcall.c
680
if (signal_pending(current)) {
fs/coda/upcall.c
775
if ((req->uc_flags & CODA_REQ_ABORT) || !signal_pending(current)) {
fs/coredump.c
1162
const struct mm_struct *mm = current->mm;
fs/coredump.c
1355
if (mmap_read_lock_killable(current->mm))
fs/coredump.c
1370
mmap_read_unlock(current->mm);
fs/coredump.c
1387
mmap_read_unlock(current->mm);
fs/coredump.c
1395
mmap_read_unlock(current->mm);
fs/coredump.c
1438
if (current->nsproxy->mnt_ns != init_task.nsproxy->mnt_ns)
fs/coredump.c
1723
struct mm_struct *mm = current->mm;
fs/coredump.c
201
exe_file = get_mm_exe_file(current->mm);
fs/coredump.c
203
return cn_esc_printf(cn, "%s (path unknown)", current->comm);
fs/coredump.c
367
task_tgid_vnr(current));
fs/coredump.c
372
task_tgid_nr(current));
fs/coredump.c
376
task_pid_vnr(current));
fs/coredump.c
380
task_pid_nr(current));
fs/coredump.c
420
err = cn_esc_printf(cn, "%s", current->comm);
fs/coredump.c
455
cprm->pid = task_tgid(current);
fs/coredump.c
476
return cn_printf(cn, ".%d", task_tgid_vnr(current)) == 0;
fs/coredump.c
492
if (t != current && !(t->flags & PF_POSTCOREDUMP)) {
fs/coredump.c
523
struct task_struct *tsk = current;
fs/coredump.c
556
spin_lock_irq(&current->sighand->siglock);
fs/coredump.c
557
if (core_dumped && !__fatal_signal_pending(current))
fs/coredump.c
558
current->signal->group_exit_code |= 0x80;
fs/coredump.c
559
next = current->signal->core_state->dumper.next;
fs/coredump.c
560
current->signal->core_state = NULL;
fs/coredump.c
561
spin_unlock_irq(&current->sighand->siglock);
fs/coredump.c
584
return fatal_signal_pending(current) || freezing(current);
fs/coredump.c
660
current->signal->rlim[RLIMIT_CORE] = (struct rlimit){1, 1};
fs/coredump.c
700
cprm->pid = task_tgid(current);
fs/crypto/policy.c
483
current->comm, current->pid);
fs/crypto/policy.c
492
current->comm, current->pid);
fs/d_path.c
286
get_fs_root_rcu(current->fs, &root);
fs/d_path.c
422
get_fs_root_and_pwd_rcu(current->fs, &root, &pwd);
fs/dax.c
1642
if (fatal_signal_pending(current)) {
fs/dcache.c
2586
DECLARE_WAITQUEUE(wait, current);
fs/devpts/inode.c
435
current->nsproxy->mnt_ns == init_task.nsproxy->mnt_ns)
fs/direct-io.c
481
dio->waiter = current;
fs/dlm/lock.c
2863
lkb->lkb_ownpid = (int) current->pid;
fs/dlm/lock.c
5902
lkb->lkb_ownpid = (int) current->pid;
fs/dlm/lock.c
6269
if (pid == current->pid)
fs/dlm/user.c
476
current->comm,
fs/dlm/user.c
477
task_pid_nr(current),
fs/dlm/user.c
789
DECLARE_WAITQUEUE(wait, current);
fs/dlm/user.c
825
if (list_empty(&proc->asts) && !signal_pending(current)) {
fs/dlm/user.c
834
if (signal_pending(current)) {
fs/drop_caches.c
73
current->comm, task_pid_nr(current),
fs/ecryptfs/messaging.c
55
(*msg_ctx)->task = current;
fs/ecryptfs/read_write.c
118
if (fatal_signal_pending(current)) {
fs/erofs/dir.c
67
if (fatal_signal_pending(current)) {
fs/eventfd.c
236
current->in_eventfd = 1;
fs/eventfd.c
239
current->in_eventfd = 0;
fs/eventfd.c
272
current->in_eventfd = 1;
fs/eventfd.c
275
current->in_eventfd = 0;
fs/eventfd.c
68
if (WARN_ON_ONCE(current->in_eventfd))
fs/eventfd.c
72
current->in_eventfd = 1;
fs/eventfd.c
77
current->in_eventfd = 0;
fs/eventpoll.c
1776
if (fatal_signal_pending(current))
fs/eventpoll.c
1982
if (signal_pending(current))
fs/exec.c
1093
struct task_struct *me = current;
fs/exec.c
1107
trace_sched_prepare_exec(current, bprm);
fs/exec.c
1119
current->fs->in_exec = 0;
fs/exec.c
1213
set_dumpable(current->mm, suid_dumpable);
fs/exec.c
1215
set_dumpable(current->mm, SUID_DUMP_USER);
fs/exec.c
1318
struct task_struct *me = current;
fs/exec.c
133
struct mm_struct *mm = current->mm;
fs/exec.c
1338
task_lock(current->group_leader);
fs/exec.c
1339
current->signal->rlim[RLIMIT_STACK] = bprm->rlim_stack;
fs/exec.c
1340
task_unlock(current->group_leader);
fs/exec.c
1352
if (mutex_lock_interruptible(&current->signal->cred_guard_mutex))
fs/exec.c
1359
mutex_unlock(&current->signal->cred_guard_mutex);
fs/exec.c
1381
current->fs->in_exec = 0;
fs/exec.c
1382
mutex_unlock(&current->signal->cred_guard_mutex);
fs/exec.c
1488
struct task_struct *p = current, *t;
fs/exec.c
1498
if (task_no_new_privs(current))
fs/exec.c
1541
if (task_no_new_privs(current))
fs/exec.c
1685
old_pid = current->pid;
fs/exec.c
1687
old_vpid = task_pid_nr_ns(current, task_active_pid_ns(current->parent));
fs/exec.c
1718
trace_sched_process_exec(current, old_pid, bprm);
fs/exec.c
1720
proc_exec_connector(current);
fs/exec.c
1738
current->in_execve = 1;
fs/exec.c
1739
sched_mm_cid_before_execve(current);
fs/exec.c
1752
sched_mm_cid_after_execve(current);
fs/exec.c
1753
rseq_execve(current);
fs/exec.c
1755
current->in_execve = 0;
fs/exec.c
1756
user_events_execve(current);
fs/exec.c
1757
acct_update_integrals(current);
fs/exec.c
1758
task_numa_free(current, false);
fs/exec.c
1768
if (bprm->point_of_no_return && !fatal_signal_pending(current))
fs/exec.c
1771
sched_mm_cid_after_execve(current);
fs/exec.c
1773
current->in_execve = 0;
fs/exec.c
1791
if ((current->flags & PF_NPROC_EXCEEDED) &&
fs/exec.c
1797
current->flags &= ~PF_NPROC_EXCEEDED;
fs/exec.c
1843
current->comm, bprm->filename);
fs/exec.c
1855
if (WARN_ON_ONCE(current->flags & PF_KTHREAD))
fs/exec.c
1897
struct mm_struct *mm = current->mm;
fs/exec.c
267
task_lock(current->group_leader);
fs/exec.c
268
bprm->rlim_stack = current->signal->rlim[RLIMIT_STACK];
fs/exec.c
269
task_unlock(current->group_leader);
fs/exec.c
344
if (fatal_signal_pending(current))
fs/exec.c
362
if (fatal_signal_pending(current))
fs/exec.c
484
if (fatal_signal_pending(current)) {
fs/exec.c
585
if (fatal_signal_pending(current))
fs/exec.c
604
struct mm_struct *mm = current->mm;
fs/exec.c
622
if (current->flags & PF_RANDOMIZE)
fs/exec.c
713
current->mm->start_stack = bprm->p;
fs/exec.c
844
tsk = current;
fs/exec.c
845
old_mm = current->mm;
fs/exfat/balloc.c
380
if (fatal_signal_pending(current)) {
fs/exfat/super.c
838
sbi->options.fs_fmask = current->fs->umask;
fs/exfat/super.c
839
sbi->options.fs_dmask = current->fs->umask;
fs/ext2/inode.c
316
colour = (current->pid % 16) *
fs/ext4/balloc.c
996
colour = (task_pid_nr(current) % 16) *
fs/ext4/balloc.c
999
colour = (task_pid_nr(current) % 16) *
fs/ext4/dir.c
178
if (fatal_signal_pending(current)) {
fs/ext4/dir.c
540
inode->i_ino, current->comm);
fs/ext4/ext4.h
96
current->comm, task_pid_nr(current), \
fs/ext4/ext4_jbd2.c
35
handle_t *handle = current->journal_info;
fs/ext4/ext4_jbd2.c
43
current->journal_info = handle;
fs/ext4/ext4_jbd2.c
58
current->journal_info = handle;
fs/ext4/fast_commit.c
1256
set_task_ioprio(current, journal_ioprio);
fs/ext4/fast_commit.c
1271
set_task_ioprio(current, old_ioprio);
fs/ext4/fast_commit.c
1281
set_task_ioprio(current, old_ioprio);
fs/ext4/fsmap.c
96
if (fatal_signal_pending(current))
fs/ext4/inode.c
5157
ino, current->comm);
fs/ext4/inode.c
5676
if (WARN_ON_ONCE(current->flags & PF_MEMALLOC))
fs/ext4/mballoc.c
6921
return fatal_signal_pending(current) || freezing(current);
fs/ext4/mballoc.h
32
current->comm, task_pid_nr(current), sb->s_id, \
fs/ext4/namei.c
1223
if (fatal_signal_pending(current)) {
fs/ext4/namei.c
150
current->comm, PTR_ERR(bh));
fs/ext4/super.c
1039
function, line, inode->i_ino, current->comm, &vaf);
fs/ext4/super.c
825
sb->s_id, function, line, current->comm, &vaf);
fs/ext4/super.c
853
block, current->comm, &vaf);
fs/ext4/super.c
858
current->comm, &vaf);
fs/ext4/super.c
894
block, current->comm, path, &vaf);
fs/ext4/super.c
900
current->comm, path, &vaf);
fs/ext4/xattr.c
1541
!(current->flags & PF_MEMALLOC_NOFS));
fs/ext4/xattr.c
2394
WARN_ON_ONCE(!(current->flags & PF_MEMALLOC_NOFS));
fs/f2fs/checkpoint.c
123
if (rt_task(current))
fs/f2fs/checkpoint.c
127
lc->orig_nice = task_nice(current);
fs/f2fs/checkpoint.c
131
set_user_nice(current, lc->new_nice);
fs/f2fs/checkpoint.c
1312
F2FS_I(inode)->cp_task = current;
fs/f2fs/checkpoint.c
1313
F2FS_I(inode)->wb_task = current;
fs/f2fs/checkpoint.c
134
trace_f2fs_priority_uplift(sem->sbi, sem->name, is_write, current,
fs/f2fs/checkpoint.c
1430
bool need_lock = sbi->umount_lock_holder != current;
fs/f2fs/checkpoint.c
144
if (task_nice(current) != lc->new_nice)
fs/f2fs/checkpoint.c
146
set_user_nice(current, lc->orig_nice);
fs/f2fs/checkpoint.c
148
trace_f2fs_priority_restore(sem->sbi, sem->name, is_write, current,
fs/f2fs/checkpoint.c
2100
sbi->umount_lock_holder == current) {
fs/f2fs/checkpoint.c
31
ts->running_time = current->se.sum_exec_runtime;
fs/f2fs/checkpoint.c
34
ts->runnable_time = current->sched_info.run_delay;
fs/f2fs/checkpoint.c
37
if (current->delays)
fs/f2fs/checkpoint.c
38
ts->io_sleep_time = current->delays->blkio_delay;
fs/f2fs/checkpoint.c
88
trace_f2fs_lock_elapsed_time(sem->sbi, sem->name, is_write, current,
fs/f2fs/data.c
2133
if (fatal_signal_pending(current))
fs/f2fs/data.c
3582
if (current->plug)
fs/f2fs/data.c
3583
blk_finish_plug(current->plug);
fs/f2fs/data.c
3634
F2FS_I(inode)->cp_task == current ?
fs/f2fs/dir.c
1089
if (fatal_signal_pending(current)) {
fs/f2fs/dir.c
393
F2FS_I(dir)->task = current;
fs/f2fs/dir.c
816
if (current != F2FS_I(dir)->task) {
fs/f2fs/f2fs.h
5053
if (fatal_signal_pending(current))
fs/f2fs/file.c
2094
if (F2FS_I(inode)->atomic_write_task == current &&
fs/f2fs/file.c
2095
(current->flags & PF_EXITING)) {
fs/f2fs/file.c
2365
fi->atomic_write_task = current;
fs/f2fs/file.c
4316
if (fatal_signal_pending(current)) {
fs/f2fs/file.c
4533
if (fatal_signal_pending(current)) {
fs/f2fs/file.c
4613
if (fatal_signal_pending(current)) {
fs/f2fs/file.c
4861
current->pid, path, current->comm);
fs/f2fs/file.c
4864
current->pid, path, current->comm);
fs/f2fs/gc.c
2141
if (fatal_signal_pending(current))
fs/f2fs/node.c
2263
if (current->plug)
fs/f2fs/node.c
2264
blk_finish_plug(current->plug);
fs/f2fs/segment.c
3470
if (fatal_signal_pending(current))
fs/f2fs/super.c
121
if (fatal_signal_pending(current))
fs/f2fs/super.c
134
if (fatal_signal_pending(current))
fs/f2fs/super.c
2131
sbi->umount_lock_holder = current;
fs/f2fs/super.c
2792
sbi->umount_lock_holder = current;
fs/f2fs/super.c
3421
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
3442
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
3503
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
5212
sbi->umount_lock_holder = current;
fs/f2fs/super.c
5472
sbi->umount_lock_holder = current;
fs/fat/fatent.c
817
if (fatal_signal_pending(current)) {
fs/fhandle.c
182
get_fs_pwd(current->fs, root);
fs/file.c
1108
return __fget_files(current->files, fd, mask);
fs/file.c
1184
struct files_struct *files = current->files;
fs/file.c
1279
struct files_struct *files = current->files;
fs/file.c
1289
res = close_on_exec(fd, current->files);
fs/file.c
1350
struct files_struct *files = current->files;
fs/file.c
1429
struct files_struct *files = current->files;
fs/file.c
1467
struct files_struct *files = current->files;
fs/file.c
572
struct files_struct *files = current->files;
fs/file.c
636
struct files_struct *files = current->files;
fs/file.c
662
struct files_struct *files = current->files;
fs/file.c
682
struct files_struct *files = current->files;
fs/file.c
736
struct files_struct *files = current->files;
fs/file.c
822
struct task_struct *me = current;
fs/file.c
881
struct files_struct *files = current->files;
fs/file_table.c
521
struct task_struct *task = current;
fs/fs-writeback.c
2076
blk_flush_plug(current->plug, false);
fs/fs-writeback.c
208
current->comm, current->pid, waited_secs);
fs/fs-writeback.c
2479
blk_flush_plug(current->plug, true);
fs/fs-writeback.c
288
memcg_css = task_get_css(current, memory_cgrp_id);
fs/fs_context.c
276
fc->net_ns = get_net(current->nsproxy->net_ns);
fs/fs_struct.c
129
struct fs_struct *fs = current->fs;
fs/fs_struct.c
136
task_lock(current);
fs/fs_struct.c
139
current->fs = new_fs;
fs/fs_struct.c
141
task_unlock(current);
fs/fuse/dev.c
233
req->in.h.pid = pid_nr_ns(task_pid(current), fc->pid_ns);
fs/fuse/dev.c
649
req->in.h.pid = pid_nr_ns(task_pid(current), fc->pid_ns);
fs/fuse/dev_uring.c
1244
qid = task_cpu(current);
fs/fuse/dir.c
2228
inarg.lock_owner = fuse_lock_owner_id(fc, current->files);
fs/fuse/file.c
1693
fl_owner_t owner = current->files;
fs/fuse/file.c
2512
struct pid *pid = fl->c.flc_type != F_UNLCK ? task_tgid(current) : NULL;
fs/fuse/inode.c
997
fc->pid_ns = get_pid_ns(task_active_pid_ns(current));
fs/gfs2/aops.c
134
BUG_ON(current->journal_info);
fs/gfs2/aops.c
548
if (current->journal_info)
fs/gfs2/bmap.c
1079
tr = current->journal_info;
fs/gfs2/bmap.c
1316
BUG_ON(current->journal_info);
fs/gfs2/bmap.c
1359
tr = current->journal_info;
fs/gfs2/bmap.c
1420
if (current->journal_info)
fs/gfs2/bmap.c
1528
if (current->journal_info == NULL) {
fs/gfs2/bmap.c
1550
tr = current->journal_info;
fs/gfs2/bmap.c
1598
if (current->journal_info) {
fs/gfs2/bmap.c
1964
if (current->journal_info == NULL) {
fs/gfs2/bmap.c
1984
if (current->journal_info) {
fs/gfs2/bmap.c
2401
tr = current->journal_info;
fs/gfs2/bmap.c
2466
BUG_ON(!current->journal_info);
fs/gfs2/bmap.c
2474
if (current->journal_info)
fs/gfs2/bmap.c
2481
if (current->journal_info)
fs/gfs2/bmap.c
990
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/file.c
801
nr_dirtied = max(current->nr_dirtied_pause -
fs/gfs2/file.c
802
current->nr_dirtied, 8);
fs/gfs2/glock.c
1187
gh->gh_owner_pid = get_pid(task_pid(current));
fs/gfs2/glock.c
1210
gh->gh_owner_pid = get_pid(task_pid(current));
fs/gfs2/glock.c
1315
if (signal_pending(current))
fs/gfs2/glock.c
1325
if (signal_pending(current))
fs/gfs2/glock.c
2660
struct pid_namespace *ns = task_active_pid_ns(current);
fs/gfs2/glock.h
160
pid = task_pid(current);
fs/gfs2/inode.c
2011
if (current->journal_info)
fs/gfs2/lock_dlm.c
864
if (retries++ && signal_pending(current)) {
fs/gfs2/log.c
1074
if (current == sdp->sd_logd_process)
fs/gfs2/log.c
211
__func__, current->journal_info ? 1 : 0);
fs/gfs2/log.c
529
if (current != sdp->sd_logd_process)
fs/gfs2/log.c
973
__func__, current->journal_info ? 1 : 0);
fs/gfs2/lops.c
46
BUG_ON(!current->journal_info);
fs/gfs2/meta_io.c
303
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/meta_io.c
330
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/meta_io.c
346
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/super.c
1023
unlikely(current->flags & PF_MEMALLOC) &&
fs/gfs2/super.c
1169
if (current->flags & PF_MEMALLOC)
fs/gfs2/super.c
1249
if (WARN_ON_ONCE(current->flags & PF_MEMALLOC))
fs/gfs2/super.c
509
if (current->journal_info == NULL) {
fs/gfs2/super.c
904
if (signal_pending(current))
fs/gfs2/trans.c
105
current->journal_info = tr;
fs/gfs2/trans.c
131
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/trans.c
134
current->journal_info = NULL;
fs/gfs2/trans.c
199
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/trans.c
262
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/trans.c
319
struct gfs2_trans *tr = current->journal_info;
fs/gfs2/trans.c
46
if (current->journal_info) {
fs/gfs2/trans.c
47
gfs2_print_trans(sdp, current->journal_info);
fs/hpfs/alloc.c
507
if (fatal_signal_pending(current))
fs/hugetlbfs/inode.c
1551
current->comm, current->pid);
fs/hugetlbfs/inode.c
740
struct mm_struct *mm = current->mm;
fs/hugetlbfs/inode.c
797
if (signal_pending(current)) {
fs/init.c
71
set_fs_pwd(current->fs, &path);
fs/init.c
93
set_fs_root(current->fs, &path);
fs/ioctl.c
81
current->comm, task_pid_nr(current),
fs/iomap/buffered-io.c
1958
if (WARN_ON_ONCE((current->flags & (PF_MEMALLOC | PF_KSWAPD)) ==
fs/iomap/buffered-io.c
957
if (fatal_signal_pending(current))
fs/iomap/direct-io.c
641
dio->iocb->ki_filp, current->comm);
fs/iomap/direct-io.c
707
dio->submit.waiter = current;
fs/jbd2/journal.c
174
journal->j_task = current;
fs/jbd2/journal.c
207
if (freezing(current)) {
fs/jbd2/journal.c
506
if (journal->j_running_transaction && !current->journal_info) {
fs/jbd2/journal.c
558
J_ASSERT(!current->journal_info);
fs/jbd2/transaction.c
1902
pid = current->pid;
fs/jbd2/transaction.c
1952
if (handle->h_sync && !(current->flags & PF_MEMALLOC))
fs/jbd2/transaction.c
332
current->comm, blocks, rsv_blocks,
fs/jbd2/transaction.c
442
current->journal_info = handle;
fs/jbd2/transaction.c
595
if (WARN_ON(current->journal_info)) {
fs/jbd2/transaction.c
705
current->journal_info = NULL;
fs/jffs2/background.c
122
while (signal_pending(current) || freezing(current)) {
fs/jffs2/background.c
85
c->gc_task = current;
fs/jffs2/background.c
88
set_user_nice(current, 10);
fs/jffs2/debug.h
81
task_pid_nr(current), __func__, ##__VA_ARGS__)
fs/jffs2/debug.h
85
task_pid_nr(current), __func__, ##__VA_ARGS__)
fs/jffs2/debug.h
89
task_pid_nr(current), __func__, ##__VA_ARGS__)
fs/jffs2/debug.h
93
task_pid_nr(current), __func__, ##__VA_ARGS__)
fs/jffs2/nodemgmt.c
178
DECLARE_WAITQUEUE(wait, current);
fs/jffs2/nodemgmt.c
194
if (signal_pending(current))
fs/jffs2/os-linux.h
43
DECLARE_WAITQUEUE(__wait, current); \
fs/jfs/jfs_lock.h
24
DECLARE_WAITQUEUE(__wait, current); \
fs/jfs/jfs_logmgr.c
1558
DECLARE_WAITQUEUE(__wait, current);
fs/jfs/jfs_logmgr.c
2325
if (freezing(current)) {
fs/jfs/jfs_metapage.c
45
DECLARE_WAITQUEUE(wait, current);
fs/jfs/jfs_txnmgr.c
120
DECLARE_WAITQUEUE(wait, current);
fs/jfs/jfs_txnmgr.c
2753
if (freezing(current)) {
fs/jfs/jfs_txnmgr.c
2757
DECLARE_WAITQUEUE(wq, current);
fs/jfs/jfs_txnmgr.c
2943
if (freezing(current)) {
fs/lockd/procfs.c
25
struct lockd_net *ln = net_generic(current->nsproxy->net_ns,
fs/lockd/procfs.c
52
struct lockd_net *ln = net_generic(current->nsproxy->net_ns,
fs/lockd/svc4proc.c
58
lock->fl.c.flc_pid = current->tgid;
fs/lockd/svcproc.c
82
lock->fl.c.flc_pid = current->tgid;
fs/locks.c
2212
pr_warn_once("%s(%d): Attempt to set a LOCK_MAND lock via flock(2). This support has been removed and the request ignored.\n", current->comm, current->pid);
fs/locks.c
2321
flock->l_pid = locks_translate_pid(&fl->c, task_active_pid_ns(current));
fs/locks.c
2343
flock->l_pid = locks_translate_pid(&fl->c, task_active_pid_ns(current));
fs/locks.c
2539
struct files_struct *files = current->files;
fs/locks.c
2661
struct files_struct *files = current->files;
fs/locks.c
2708
lock.c.flc_pid = current->tgid;
fs/locks.c
490
fl->c.flc_pid = current->tgid;
fs/locks.c
547
fl->c.flc_owner = current->files;
fs/locks.c
548
fl->c.flc_pid = current->tgid;
fs/locks.c
595
__f_setown(filp, task_pid(current), PIDTYPE_TGID, 0);
fs/locks.c
653
fl->c.flc_pid = current->tgid;
fs/minix/inode.c
41
current->comm, &vaf);
fs/namei.c
1102
struct fs_struct *fs = current->fs;
fs/namei.c
1166
struct nameidata *nd = current->nameidata;
fs/namei.c
2726
struct fs_struct *fs = current->fs;
fs/namei.c
2736
get_fs_pwd(current->fs, &nd->path);
fs/namei.c
754
struct nameidata *old = current->nameidata;
fs/namei.c
764
current->nameidata = p;
fs/namei.c
780
struct nameidata *now = current->nameidata, *old = now->saved;
fs/namei.c
782
current->nameidata = old;
fs/namespace.c
1376
struct task_struct *task = current;
fs/namespace.c
1881
if (&mnt->mnt == current->fs->root.mnt ||
fs/namespace.c
1923
if (&mnt->mnt == current->fs->root.mnt && !(flags & MNT_DETACH)) {
fs/namespace.c
2004
return ns_capable(current->nsproxy->mnt_ns->user_ns, CAP_SYS_ADMIN);
fs/namespace.c
2154
return current->nsproxy->mnt_ns->ns.ns_id >= mnt_ns->ns.ns_id;
fs/namespace.c
2562
struct user_namespace *user_ns = current->nsproxy->mnt_ns->user_ns;
fs/namespace.c
3035
struct mnt_namespace *ns, *mnt_ns = current->nsproxy->mnt_ns, *src_mnt_ns;
fs/namespace.c
3092
struct mnt_namespace *ns = current->nsproxy->mnt_ns;
fs/namespace.c
3546
if (&mnt_to->mnt == current->fs->root.mnt)
fs/namespace.c
3548
if (parent_mnt_to == current->nsproxy->mnt_ns->root)
fs/namespace.c
4488
ns = alloc_mnt_ns(current->nsproxy->mnt_ns->user_ns, true);
fs/namespace.c
4643
get_fs_root(current->fs, &root);
fs/namespace.c
4691
touch_mnt_namespace(current->nsproxy->mnt_ns);
fs/namespace.c
5295
s->sm.propagate_from = get_dominating_id(m, &current->fs->root);
fs/namespace.c
5640
if (ns == current->nsproxy->mnt_ns) {
fs/namespace.c
5641
get_fs_root(current->fs, root);
fs/namespace.c
5913
mnt_ns = current->nsproxy->mnt_ns;
fs/namespace.c
5949
if (kreq.mnt_ns_id && (ns != current->nsproxy->mnt_ns) &&
fs/namespace.c
6112
if (kreq.mnt_ns_id && (kls.ns != current->nsproxy->mnt_ns) &&
fs/namespace.c
6197
set_fs_pwd(current->fs, &root);
fs/namespace.c
6198
set_fs_root(current->fs, &root);
fs/namespace.c
6297
get_fs_root(current->fs, &fs_root);
fs/namespace.c
6303
root = topmost_overmount(current->nsproxy->mnt_ns->root);
fs/namespace.c
6371
struct mnt_namespace *ns = current->nsproxy->mnt_ns;
fs/namespace.c
840
struct mnt_namespace *ns = current->nsproxy->mnt_ns;
fs/namespace.c
948
return mnt->mnt_ns == current->nsproxy->mnt_ns;
fs/namespace.c
959
return !seq || (seq == current->nsproxy->mnt_ns->ns.ns_id);
fs/netfs/buffered_write.c
200
if (signal_pending(current)) {
fs/netfs/direct_write.c
155
if (!wreq->iocb && signal_pending(current)) {
fs/netfs/internal.h
449
printk("[%-6.6s] "FMT"\n", current->comm, ##__VA_ARGS__)
fs/nfs/blocklayout/rpc_pipefs.c
61
DECLARE_WAITQUEUE(wq, current);
fs/nfs/dir.c
3090
u64 login_time = nfs_access_login_time(current, cred);
fs/nfs/dir.c
3139
u64 login_time = nfs_access_login_time(current, cred);
fs/nfs/inode.c
1107
l_ctx->lockowner = current->files;
fs/nfs/inode.c
1117
if (pos->lockowner != current->files)
fs/nfs/inode.c
80
if (signal_pending_state(mode, current))
fs/nfs/internal.h
693
if ((current->flags & PF_WQ_WORKER) && ret == GFP_KERNEL)
fs/nfs/internal.h
925
return (current->flags & PF_EXITING) != 0;
fs/nfs/localio.c
1050
unsigned long old_flags = current->flags;
fs/nfs/localio.c
1056
current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
fs/nfs/localio.c
1065
current->flags = old_flags;
fs/nfs/localio.c
859
unsigned long old_flags = current->flags;
fs/nfs/localio.c
863
current->flags |= PF_LOCAL_THROTTLE | PF_MEMALLOC_NOIO;
fs/nfs/localio.c
892
current->flags = old_flags;
fs/nfs/nfs3acl.c
22
struct posix_acl *sentinel = uncached_acl_sentinel(current);
fs/nfs/nfs3acl.c
30
struct posix_acl *sentinel = uncached_acl_sentinel(current);
fs/nfs/nfs3acl.c
40
struct posix_acl *sentinel = uncached_acl_sentinel(current);
fs/nfs/nfs3proc.c
361
data->arg.create.verifier[1] = cpu_to_be32(current->pid);
fs/nfs/nfs3proc.c
42
} while (!fatal_signal_pending(current) && !nfs_current_task_exiting());
fs/nfs/nfs4proc.c
1743
if (!fatal_signal_pending(current) &&
fs/nfs/nfs4proc.c
3251
[1] = (__u32)current->pid,
fs/nfs/nfs4proc.c
3513
if (fatal_signal_pending(current) || nfs_current_task_exiting())
fs/nfs/nfs4proc.c
459
if (!__fatal_signal_pending(current))
fs/nfs/nfs4proc.c
472
if (!signal_pending(current))
fs/nfs/nfs4proc.c
474
return __fatal_signal_pending(current) ? -EINTR :-ERESTARTSYS;
fs/nfs/nfs4super.c
106
new->task = current;
fs/nfs/nfs4super.c
90
if (p->task == current)
fs/nfs/write.c
1137
do_flush |= l_ctx->lockowner != current->files;
fs/nfsd/export.c
262
nfsd_file_cache_purge(current->nsproxy->net_ns);
fs/nfsd/nfs4layouts.c
213
fl->c.flc_pid = current->tgid;
fs/nfsd/nfs4proc.c
476
current->fs->umask = open->op_umask;
fs/nfsd/nfs4proc.c
478
current->fs->umask = 0;
fs/nfsd/nfs4proc.c
847
current->fs->umask = create->cr_umask;
fs/nfsd/nfs4proc.c
916
current->fs->umask = 0;
fs/nfsd/nfs4state.c
5930
fl->c.flc_pid = current->tgid;
fs/nfsd/nfs4state.c
8426
file_lock->c.flc_pid = current->tgid;
fs/nfsd/nfs4state.c
8608
file_lock->c.flc_pid = current->tgid;
fs/nfsd/nfs4state.c
8683
file_lock->c.flc_pid = current->tgid;
fs/nfsd/nfsctl.c
1288
struct nfsd_net *nn = net_generic(current->nsproxy->net_ns,
fs/nfsd/nfsctl.c
1382
return exports_net_open(current->nsproxy->net_ns, file);
fs/nfsd/nfssvc.c
591
if (kthread_func(current) == nfsd)
fs/nfsd/nfssvc.c
592
return kthread_data(current);
fs/nfsd/nfssvc.c
898
current->fs->umask = 0;
fs/nfsd/vfs.c
1264
dprintk("nfsd: write defer %d\n", task_pid_nr(current));
fs/nfsd/vfs.c
1266
dprintk("nfsd: write resume %d\n", task_pid_nr(current));
fs/nfsd/vfs.c
1270
dprintk("nfsd: write sync %d\n", task_pid_nr(current));
fs/nfsd/vfs.c
1440
unsigned int pflags = current->flags;
fs/nfsd/vfs.c
1458
current->flags |= PF_LOCAL_THROTTLE;
fs/nilfs2/nilfs.h
190
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/nilfs.h
197
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/segment.c
144
struct nilfs_transaction_info *cur_ti = current->journal_info;
fs/nilfs2/segment.c
157
save = current->journal_info;
fs/nilfs2/segment.c
170
current->journal_info = ti;
fs/nilfs2/segment.c
210
trace_ti = current->journal_info;
fs/nilfs2/segment.c
2239
init_waitqueue_entry(&wait_req.wq, current);
fs/nilfs2/segment.c
2270
if (!signal_pending(current)) {
fs/nilfs2/segment.c
228
trace_ti = current->journal_info;
fs/nilfs2/segment.c
2324
BUG_ON((ti = current->journal_info) && ti->ti_magic == NILFS_TI_MAGIC);
fs/nilfs2/segment.c
235
ti = current->journal_info;
fs/nilfs2/segment.c
236
current->journal_info = ti->ti_save;
fs/nilfs2/segment.c
258
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/segment.c
2668
if (freezing(current)) {
fs/nilfs2/segment.c
282
current->journal_info = ti->ti_save;
fs/nilfs2/segment.c
294
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/segment.c
309
current->journal_info = ti->ti_save;
fs/nilfs2/segment.c
329
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/segment.c
342
struct nilfs_transaction_info *cur_ti = current->journal_info;
fs/nilfs2/segment.c
351
current->journal_info = ti;
fs/nilfs2/segment.c
375
struct nilfs_transaction_info *ti = current->journal_info;
fs/nilfs2/segment.c
382
current->journal_info = ti->ti_save;
fs/notify/dnotify/dnotify.c
265
fl_owner_t id = current->files;
fs/notify/dnotify/dnotify.c
369
__f_setown(filp, task_pid(current), PIDTYPE_TGID, 0);
fs/notify/fanotify/fanotify.c
867
pid = get_pid(task_pid(current));
fs/notify/fanotify/fanotify.c
869
pid = get_pid(task_tgid(current));
fs/notify/fanotify/fanotify_user.c
1022
if (signal_pending(current))
fs/notify/fanotify/fanotify_user.c
1050
FANOTIFY_PERM(event)->recv_pid = current->pid;
fs/notify/fanotify/fanotify_user.c
1710
group->memcg = get_mem_cgroup_from_mm(current->mm);
fs/notify/fanotify/fanotify_user.c
859
task_tgid(current) != event->pid)
fs/notify/inotify/inotify_user.c
286
if (signal_pending(current))
fs/notify/inotify/inotify_user.c
676
group->memcg = get_mem_cgroup_from_mm(current->mm);
fs/ocfs2/alloc.c
7538
if (fatal_signal_pending(current)) {
fs/ocfs2/aops.c
2088
dwc->dw_writer_pid = task_pid_nr(current);
fs/ocfs2/aops.c
2301
BUG_ON(dwc->dw_writer_pid != task_pid_nr(current));
fs/ocfs2/cluster/heartbeat.c
1206
set_user_nice(current, MIN_NICE);
fs/ocfs2/cluster/masklog.c
78
level, current->comm, task_pid_nr(current),
fs/ocfs2/cluster/tcp.c
1030
o2net_init_nst(&nst, msg_type, key, current, target_node);
fs/ocfs2/dlm/dlmdomain.c
1895
if (signal_pending(current)) {
fs/ocfs2/dlm/dlmdomain.c
2115
if (signal_pending(current)) {
fs/ocfs2/dlm/dlmthread.c
44
DECLARE_WAITQUEUE(wait, current);
fs/ocfs2/dlmfs/userdlm.c
430
if (signal_pending(current)) {
fs/ocfs2/dlmglue.c
1501
if (catch_signals && signal_pending(current)) {
fs/ocfs2/dlmglue.c
2680
struct pid *pid = task_pid(current);
fs/ocfs2/dlmglue.c
3515
if (lockres->l_flags & OCFS2_LOCK_QUEUED && current == osb->dc_task) {
fs/ocfs2/dlmglue.c
804
oh->oh_owner_pid = get_pid(task_pid(current));
fs/ocfs2/inode.c
923
trace_ocfs2_inode_is_valid_to_delete(current, osb->dc_task,
fs/ocfs2/inode.c
941
if (current == osb->dc_task)
fs/ocfs2/quota_global.c
746
if (current == osb->dc_task) {
fs/ocfs2/refcounttree.c
4517
if (fatal_signal_pending(current)) {
fs/open.c
1503
retval = filp_flush(file, current->files);
fs/open.c
563
set_fs_pwd(current->fs, &path);
fs/open.c
586
set_fs_pwd(current->fs, &fd_file(f)->f_path);
fs/open.c
610
set_fs_root(current->fs, &path);
fs/orangefs/devorangefs-req.c
317
current->comm);
fs/orangefs/devorangefs-req.c
340
current->comm);
fs/orangefs/devorangefs-req.c
499
current->comm);
fs/orangefs/orangefs-bufmap.c
108
if (signal_pending(current))
fs/orangefs/orangefs-mod.c
225
current->comm);
fs/orangefs/waitqueue.c
120
current->comm);
fs/orangefs/waitqueue.c
235
current->comm);
fs/orangefs/waitqueue.c
50
current->comm);
fs/orangefs/waitqueue.c
74
op->upcall.tgid = current->tgid;
fs/orangefs/waitqueue.c
75
op->upcall.pid = current->pid;
fs/orangefs/waitqueue.c
84
current->comm,
fs/orangefs/waitqueue.c
85
current->pid);
fs/overlayfs/copy_up.c
306
if (signal_pending_state(TASK_KILLABLE, current)) {
fs/overlayfs/dir.c
589
if (WARN_ON_ONCE(current->cred != ovl_creds(dentry->d_sb)))
fs/pidfs.c
303
const struct pid_namespace *ns = task_active_pid_ns(current);
fs/pipe.c
1110
if (signal_pending(current))
fs/pipe.c
462
send_sig(SIGPIPE, current, 0);
fs/pipe.c
503
send_sig(SIGPIPE, current, 0);
fs/pipe.c
557
if (signal_pending(current)) {
fs/posix_acl.c
135
sentinel = uncached_acl_sentinel(current);
fs/proc/base.c
1147
current->comm, task_pid_nr(current), task_pid_nr(task),
fs/proc/base.c
1175
if (!legacy && has_capability_noaudit(current, CAP_SYS_RESOURCE))
fs/proc/base.c
1194
if (!legacy && has_capability_noaudit(current, CAP_SYS_RESOURCE))
fs/proc/base.c
1339
if (current->flags & PF_KTHREAD)
fs/proc/base.c
1343
if (current != pid_task(proc_pid(inode), PIDTYPE_PID)) {
fs/proc/base.c
1738
if (same_thread_group(current, p)) {
fs/proc/base.c
2613
if (p != current) {
fs/proc/base.c
2653
if (p != current) {
fs/proc/base.c
2805
if (file->private_data != current->mm)
fs/proc/base.c
2815
if (current != task) {
fs/proc/base.c
2840
rv = mutex_lock_interruptible(&current->signal->cred_guard_mutex);
fs/proc/base.c
2847
mutex_unlock(&current->signal->cred_guard_mutex);
fs/proc/base.c
3638
is_same_tgroup = same_thread_group(current, task);
fs/proc/base.c
890
READ_ONCE(task->parent) == current;
fs/proc/fd.c
343
if (p && same_thread_group(p, current))
fs/proc/kcore.c
450
append_kcore_note(notes, &i, NN_TASKSTRUCT, NT_TASKSTRUCT, current,
fs/proc/loadavg.c
25
idr_get_cursor(&task_active_pid_ns(current)->idr) - 1);
fs/proc/proc_net.c
99
p->net = get_net_track(current->nsproxy->net_ns, &p->ns_tracker,
fs/proc/root.c
120
struct pid_namespace *target, *active = task_active_pid_ns(current);
fs/proc/root.c
338
ctx->pid_ns = get_pid_ns(task_active_pid_ns(current));
fs/proc/self.c
16
pid_t tgid = task_tgid_nr_ns(current, ns);
fs/proc/task_mmu.c
3043
if (fatal_signal_pending(current)) {
fs/proc/task_nommu.c
55
if (current->fs && current->fs->users > 1)
fs/proc/task_nommu.c
56
sbytes += kobjsize(current->fs);
fs/proc/task_nommu.c
58
bytes += kobjsize(current->fs);
fs/proc/task_nommu.c
60
if (current->files && atomic_read(&current->files->count) > 1)
fs/proc/task_nommu.c
61
sbytes += kobjsize(current->files);
fs/proc/task_nommu.c
63
bytes += kobjsize(current->files);
fs/proc/task_nommu.c
65
if (current->sighand && refcount_read(&current->sighand->count) > 1)
fs/proc/task_nommu.c
66
sbytes += kobjsize(current->sighand);
fs/proc/task_nommu.c
68
bytes += kobjsize(current->sighand);
fs/proc/task_nommu.c
70
bytes += kobjsize(current); /* includes kernel stack */
fs/proc/thread_self.c
16
pid_t tgid = task_tgid_nr_ns(current, ns);
fs/proc/thread_self.c
17
pid_t pid = task_pid_nr_ns(current, ns);
fs/read_write.c
1088
add_rchar(current, ret);
fs/read_write.c
1089
inc_syscr(current);
fs/read_write.c
1111
add_wchar(current, ret);
fs/read_write.c
1112
inc_syscw(current);
fs/read_write.c
1138
add_rchar(current, ret);
fs/read_write.c
1139
inc_syscr(current);
fs/read_write.c
1159
add_wchar(current, ret);
fs/read_write.c
1160
inc_syscw(current);
fs/read_write.c
1382
add_rchar(current, retval);
fs/read_write.c
1383
add_wchar(current, retval);
fs/read_write.c
1393
inc_syscr(current);
fs/read_write.c
1394
inc_syscw(current);
fs/read_write.c
1638
add_rchar(current, ret);
fs/read_write.c
1640
add_wchar(current, ret);
fs/read_write.c
1643
inc_syscr(current);
fs/read_write.c
1644
inc_syscw(current);
fs/read_write.c
1719
send_sig(SIGXFSZ, current, 0);
fs/read_write.c
504
op, file, current->pid, current->comm);
fs/read_write.c
537
add_rchar(current, ret);
fs/read_write.c
539
inc_syscr(current);
fs/read_write.c
579
add_rchar(current, ret);
fs/read_write.c
581
inc_syscr(current);
fs/read_write.c
626
add_wchar(current, ret);
fs/read_write.c
628
inc_syscw(current);
fs/read_write.c
693
add_wchar(current, ret);
fs/read_write.c
695
inc_syscw(current);
fs/readdir.c
286
if (!(flags & FILLDIR_FLAG_NOINTR) && prev_reclen && signal_pending(current))
fs/readdir.c
370
if (!(flags & FILLDIR_FLAG_NOINTR) && prev_reclen && signal_pending(current))
fs/readdir.c
542
if (!(flags & FILLDIR_FLAG_NOINTR) && prev_reclen && signal_pending(current))
fs/remap_range.c
562
if (fatal_signal_pending(current))
fs/resctrl/pseudo_lock.c
1030
if (!cpumask_subset(current->cpus_ptr, &plr->d->hdr.cpu_mask)) {
fs/resctrl/rdtgroup.c
606
if (task == current)
fs/resctrl/rdtgroup.c
755
tsk = current;
fs/select.c
1071
restart_block = &current->restart_block;
fs/select.c
1194
fdt = files_fdtable(current->files);
fs/select.c
122
pwq->polling_task = current;
fs/select.c
1234
if (signal_pending(current))
fs/select.c
306
if (current->personality & STICKY_TIMEOUTS)
fs/select.c
427
fdt = files_fdtable(current->files);
fs/select.c
575
if (retval || timed_out || signal_pending(current))
fs/select.c
61
if (task_nice(current) > 0)
fs/select.c
638
fdt = files_fdtable(current->files);
fs/select.c
683
if (signal_pending(current))
fs/select.c
80
u64 slack = current->timer_slack_ns;
fs/select.c
927
if (signal_pending(current))
fs/signalfd.c
159
DECLARE_WAITQUEUE(wait, current);
fs/signalfd.c
161
spin_lock_irq(&current->sighand->siglock);
fs/signalfd.c
170
spin_unlock_irq(&current->sighand->siglock);
fs/signalfd.c
174
add_wait_queue(&current->sighand->signalfd_wqh, &wait);
fs/signalfd.c
180
if (signal_pending(current)) {
fs/signalfd.c
184
spin_unlock_irq(&current->sighand->siglock);
fs/signalfd.c
186
spin_lock_irq(&current->sighand->siglock);
fs/signalfd.c
188
spin_unlock_irq(&current->sighand->siglock);
fs/signalfd.c
190
remove_wait_queue(&current->sighand->signalfd_wqh, &wait);
fs/signalfd.c
289
spin_lock_irq(&current->sighand->siglock);
fs/signalfd.c
291
spin_unlock_irq(&current->sighand->siglock);
fs/signalfd.c
293
wake_up(&current->sighand->signalfd_wqh);
fs/signalfd.c
56
poll_wait(file, &current->sighand->signalfd_wqh, wait);
fs/signalfd.c
58
spin_lock_irq(&current->sighand->siglock);
fs/signalfd.c
59
if (next_signal(&current->pending, &ctx->sigmask) ||
fs/signalfd.c
60
next_signal(&current->signal->shared_pending,
fs/signalfd.c
63
spin_unlock_irq(&current->sighand->siglock);
fs/smb/client/cifs_spnego.c
152
dp += sprintf(dp, ";pid=0x%x", current->pid);
fs/smb/client/cifssmb.c
5856
rc = CIFSSMBSetFileInfo(xid, tcon, data, fid.netfid, current->tgid);
fs/smb/client/connect.c
1269
cifs_dbg(FYI, "Demultiplex PID: %d\n", task_pid_nr(current));
fs/smb/client/connect.c
1444
while (!signal_pending(current)) {
fs/smb/client/connect.c
1634
if (!net_eq(cifs_net_ns(server), current->nsproxy->net_ns))
fs/smb/client/connect.c
1786
cifs_set_net_ns(tcp_ses, get_net(current->nsproxy->net_ns));
fs/smb/client/dir.c
391
current->tgid);
fs/smb/client/dns_resolve.c
28
rc = dns_query(current->nsproxy->net_ns, NULL, name,
fs/smb/client/file.c
1571
lock->pid = current->tgid;
fs/smb/client/file.c
1607
if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
fs/smb/client/file.c
1616
current->tgid == li->pid) || type == li->type))
fs/smb/client/file.c
2225
if (current->tgid != li->pid)
fs/smb/client/file.c
2614
if (!any_available && open_file->pid != current->tgid)
fs/smb/client/file.c
278
req->pid = current->tgid; // Ummm... This may be a workqueue
fs/smb/client/file.c
445
current->comm, current->tgid);
fs/smb/client/file.c
693
cfile->pid = current->tgid;
fs/smb/client/inode.c
1849
current->tgid);
fs/smb/client/inode.c
1870
current->tgid);
fs/smb/client/inode.c
1907
current->tgid))
fs/smb/client/inode.c
2747
if (signal_pending_state(mode, current))
fs/smb/client/inode.c
578
io_parms.pid = current->tgid;
fs/smb/client/link.c
260
io_parms.pid = current->tgid;
fs/smb/client/link.c
297
io_parms.pid = current->tgid;
fs/smb/client/link.c
359
io_parms.pid = current->tgid;
fs/smb/client/link.c
412
io_parms.pid = current->tgid;
fs/smb/client/smb1misc.c
36
buffer->Pid = cpu_to_le16((__u16)current->tgid);
fs/smb/client/smb1misc.c
37
buffer->PidHigh = cpu_to_le16((__u16)(current->tgid >> 16));
fs/smb/client/smb1ops.c
1081
netpid = current->tgid;
fs/smb/client/smb1ops.c
1159
return CIFSSMBLock(0, tcon, net_fid, current->tgid, 0, 0, 0, 0,
fs/smb/client/smb1ops.c
1200
current->tgid, length, offset, unlock, lock,
fs/smb/client/smb1transport.c
51
temp->pid = current->pid;
fs/smb/client/smb1transport.c
62
get_task_struct(current);
fs/smb/client/smb1transport.c
63
temp->creator = current;
fs/smb/client/smb1transport.c
65
temp->callback_data = current;
fs/smb/client/smb2file.c
296
if (current->tgid != li->pid)
fs/smb/client/smb2file.c
325
current->tgid, num, buf);
fs/smb/client/smb2file.c
348
cfile->fid.volatile_fid, current->tgid,
fs/smb/client/smb2file.c
382
current->tgid, num, buf);
fs/smb/client/smb2file.c
394
current->tgid, num, buf);
fs/smb/client/smb2inode.c
371
current->tgid,
fs/smb/client/smb2inode.c
380
current->tgid,
fs/smb/client/smb2inode.c
406
current->tgid,
fs/smb/client/smb2inode.c
415
current->tgid,
fs/smb/client/smb2inode.c
440
cfile->fid.volatile_fid, current->tgid,
fs/smb/client/smb2inode.c
447
COMPOUND_FID, current->tgid,
fs/smb/client/smb2inode.c
482
current->tgid, FILE_RENAME_INFORMATION,
fs/smb/client/smb2inode.c
488
current->tgid, FILE_RENAME_INFORMATION,
fs/smb/client/smb2inode.c
518
COMPOUND_FID, current->tgid,
fs/smb/client/smb2ops.c
1306
COMPOUND_FID, current->tgid,
fs/smb/client/smb2ops.c
1707
current->tgid, FILE_END_OF_FILE_INFORMATION,
fs/smb/client/smb2ops.c
3062
current->tgid, length, offset, type, wait);
fs/smb/client/smb2ops.c
3153
if (fatal_signal_pending(current)) {
fs/smb/client/smb2ops.c
3531
io_parms.pid = current->tgid;
fs/smb/client/smb2ops.c
5272
io_parms.pid = current->tgid;
fs/smb/client/smb2pdu.c
132
shdr->Id.SyncId.ProcessId = cpu_to_le32((__u16)current->tgid);
fs/smb/client/smb2pdu.c
5898
current->tgid, 0, SMB2_O_INFO_SECURITY, aclflag,
fs/smb/client/smb2pdu.c
5908
current->tgid, FILE_FULL_EA_INFORMATION, SMB2_O_INFO_FILE,
fs/smb/client/smb2transport.c
664
temp->pid = current->pid;
fs/smb/client/smb2transport.c
672
get_task_struct(current);
fs/smb/client/smb2transport.c
673
temp->creator = current;
fs/smb/client/smb2transport.c
675
temp->callback_data = current;
fs/smb/client/trace.h
1715
__entry->pid = current->pid;
fs/smb/client/transport.c
183
if (rc == -EAGAIN || unlikely(rc == -EINTR && task_work_pending(current))) {
fs/smb/client/transport.c
266
if (fatal_signal_pending(current)) {
fs/smb/client/transport.c
347
if (signal_pending(current) && (total_len != send_length)) {
fs/smb/server/smb2pdu.c
7359
fl->c.flc_pid = current->tgid;
fs/smb/server/transport_tcp.c
462
ret = sock_create_kern(current->nsproxy->net_ns, PF_INET6, SOCK_STREAM,
fs/smb/server/transport_tcp.c
467
ret = sock_create_kern(current->nsproxy->net_ns, PF_INET,
fs/splice.c
1045
pipe = current->splice_pipe;
fs/splice.c
1058
current->splice_pipe = pipe;
fs/splice.c
1263
send_sig(SIGPIPE, current, 0);
fs/splice.c
1270
if (signal_pending(current))
fs/splice.c
1657
if (signal_pending(current)) {
fs/splice.c
1694
send_sig(SIGPIPE, current, 0);
fs/splice.c
1702
if (signal_pending(current)) {
fs/splice.c
1750
send_sig(SIGPIPE, current, 0);
fs/splice.c
1871
send_sig(SIGPIPE, current, 0);
fs/splice.c
210
send_sig(SIGPIPE, current, 0);
fs/splice.c
252
send_sig(SIGPIPE, current, 0);
fs/splice.c
521
if (signal_pending(current))
fs/splice.c
535
if (signal_pending(current))
fs/splice.c
816
if (signal_pending(current))
fs/splice.c
832
if (signal_pending(current))
fs/splice.c
924
op, file, current->pid, current->comm);
fs/stat.c
391
current->comm);
fs/ubifs/commit.c
292
c->bgt_name, current->pid);
fs/ubifs/commit.c
388
dbg_cmt("pid %d goes sleep", current->pid);
fs/ubifs/commit.c
399
dbg_cmt("commit finished, pid %d woke up", current->pid);
fs/ubifs/compress.c
310
current->pid, compr->name, PTR_ERR(compr->cc));
fs/ubifs/debug.c
604
current->pid, lst->empty_lebs, lst->idx_lebs);
fs/ubifs/debug.c
623
current->pid, bi->data_growth + bi->dd_growth,
fs/ubifs/debug.c
777
pr_err("(pid %d) start dumping LEB properties\n", current->pid);
fs/ubifs/debug.c
790
pr_err("(pid %d) finish dumping LEB properties\n", current->pid);
fs/ubifs/debug.c
798
pr_err("(pid %d) dumping LPT information\n", current->pid);
fs/ubifs/debug.c
837
pr_err("(pid %d) start dumping LEB %d\n", current->pid, lnum);
fs/ubifs/debug.c
861
pr_err("(pid %d) finish dumping LEB %d\n", current->pid, lnum);
fs/ubifs/debug.c
912
current->pid, cat, heap->cnt);
fs/ubifs/debug.c
920
pr_err("(pid %d) finish dumping heap\n", current->pid);
fs/ubifs/debug.c
928
pr_err("(pid %d) dumping pnode:\n", current->pid);
fs/ubifs/debug.c
947
pr_err("(pid %d) start dumping TNC tree\n", current->pid);
fs/ubifs/debug.c
963
pr_err("(pid %d) finish dumping TNC tree\n", current->pid);
fs/ubifs/debug.h
159
pr_debug("UBIFS DBG " type " (pid %d): " fmt "\n", current->pid, \
fs/ubifs/debug.h
165
pr_debug("UBIFS DBG " type " (pid %d): " fmt "%s\n", current->pid, \
fs/ubifs/lpt_commit.c
1846
pr_err("(pid %d) start dumping LEB %d\n", current->pid, lnum);
fs/ubifs/lpt_commit.c
1932
pr_err("(pid %d) finish dumping LEB %d\n", current->pid, lnum);
fs/ubifs/lpt_commit.c
1948
pr_err("(pid %d) start dumping all LPT LEBs\n", current->pid);
fs/ubifs/lpt_commit.c
1951
pr_err("(pid %d) finish dumping all LPT LEBs\n", current->pid);
fs/ubifs/misc.c
34
c->vi.ubi_num, c->vi.vol_id, current->pid,
fs/ubifs/misc.c
53
c->vi.ubi_num, c->vi.vol_id, current->pid,
fs/ubifs/super.c
2254
current->pid, fc->source, err);
fs/ubifs/super.c
2440
current->pid, (unsigned int)PAGE_SIZE);
fs/ubifs/super.c
2473
current->pid, err);
fs/userfaultfd.c
1087
if (signal_pending(current)) {
fs/userfaultfd.c
1926
if (mm != current->mm)
fs/userfaultfd.c
2135
VM_WARN_ON_ONCE(!current->mm);
fs/userfaultfd.c
2153
ctx->mm = current->mm;
fs/userfaultfd.c
221
msg.arg.pagefault.feat.ptid = task_pid_vnr(current);
fs/userfaultfd.c
395
if (current->flags & (PF_EXITING|PF_DUMPCORE))
fs/userfaultfd.c
480
uwq.wq.private = current;
fs/userfaultfd.c
565
if (WARN_ON_ONCE(current->flags & PF_EXITING))
fs/userfaultfd.c
569
init_waitqueue_entry(&ewq->wq, current);
fs/userfaultfd.c
583
fatal_signal_pending(current)) {
fs/userfaultfd.c
996
DECLARE_WAITQUEUE(wait, current);
fs/verity/enable.c
147
if (fatal_signal_pending(current)) {
fs/verity/read_metadata.c
80
if (fatal_signal_pending(current)) {
fs/xfs/scrub/scrub.c
995
if (fatal_signal_pending(current)) {
fs/xfs/scrub/scrub.h
48
if (widget->interruptible && fatal_signal_pending(current))
fs/xfs/xfs_buf.c
1012
DECLARE_WAITQUEUE (wait, current);
fs/xfs/xfs_discard.c
324
return fatal_signal_pending(current) || freezing(current);
fs/xfs/xfs_fsmap.c
1178
has_capability_noaudit(current, CAP_SYS_ADMIN);
fs/xfs/xfs_fsmap.c
292
if (fatal_signal_pending(current))
fs/xfs/xfs_icache.c
2157
if (current->flags & PF_MEMALLOC_NOFS)
fs/xfs/xfs_ioctl.c
1205
current->comm);
fs/xfs/xfs_ioctl.c
624
has_capability_noaudit(current, CAP_FOWNER), &tp);
fs/xfs/xfs_iomap.c
1715
ip->i_ino, current->comm);
fs/xfs/xfs_iops.c
838
has_capability_noaudit(current, CAP_FOWNER), &tp);
fs/xfs/xfs_log.c
3137
tic->t_task = current;
fs/xfs/xfs_log_priv.h
599
DECLARE_WAITQUEUE(wait, current);
fs/xfs/xfs_platform.h
119
(*(sp) = current->flags, current->flags |= (f))
fs/xfs/xfs_platform.h
121
(current->flags = ((current->flags & ~(f)) | (*(sp) & (f))))
fs/xfs/xfs_pwork.c
72
trace_xfs_pwork_init(mp, nr_threads, current->pid);
fs/xfs/xfs_pwork.c
76
current->pid);
fs/xfs/xfs_reflink.c
1583
if (fatal_signal_pending(current)) {
fs/xfs/xfs_verify_media.c
381
if (fatal_signal_pending(current)) {
fs/xfs/xfs_zone_alloc.c
178
ASSERT(current == zi->zi_gc_thread);
fs/xfs/xfs_zone_space_resv.c
112
if (fatal_signal_pending(current))
fs/xfs/xfs_zone_space_resv.c
125
.task = current,
fs/zonefs/file.c
411
send_sig(SIGXFSZ, current, 0);
include/acpi/platform/aclinuxex.h
69
return (acpi_thread_id) (unsigned long)current;
include/asm-generic/mmu_context.h
60
switch_mm(prev_mm, next_mm, current);
include/kunit/test-bug.h
46
return current->kunit_test;
include/linux/alloc_tag.h
234
#define alloc_tag_record(p) ((p) = current->alloc_tag)
include/linux/audit.h
339
return current->audit_context;
include/linux/backing-dev.h
183
memcg_css = task_css(current, memory_cgrp_id);
include/linux/backing-dev.h
193
if (likely(wb && wb->blkcg_css == task_css(current, io_cgrp_id)))
include/linux/backing-dev.h
221
memcg_css = task_get_css(current, memory_cgrp_id);
include/linux/blkdev.h
1217
current->flags &= ~PF_BLOCK_TS;
include/linux/blkdev.h
1698
if (waiter == current)
include/linux/bpf.h
2433
old_ctx = current->bpf_ctx;
include/linux/bpf.h
2434
current->bpf_ctx = new_ctx;
include/linux/bpf.h
2442
current->bpf_ctx = old_ctx;
include/linux/bpf.h
3074
return !!current->bpf_ctx;
include/linux/cgroup.h
647
current->no_cgroup_migration = 1;
include/linux/cgroup.h
656
current->no_cgroup_migration = 0;
include/linux/compaction.h
85
return current->flags & PF_KCOMPACTD;
include/linux/compat.h
548
struct task_struct *t = current; \
include/linux/coredump.h
57
memcpy(comm, current->comm, sizeof(comm)); \
include/linux/coredump.h
59
task_tgid_vnr(current), (int)strlen(comm), comm, ##__VA_ARGS__); \
include/linux/cpuset.h
126
return task_spread_page(current);
include/linux/cpuset.h
149
return read_seqcount_begin(&current->mems_allowed_seq);
include/linux/cpuset.h
163
return read_seqcount_retry(&current->mems_allowed_seq, seq);
include/linux/cpuset.h
170
task_lock(current);
include/linux/cpuset.h
172
write_seqcount_begin(&current->mems_allowed_seq);
include/linux/cpuset.h
173
current->mems_allowed = nodemask;
include/linux/cpuset.h
174
write_seqcount_end(&current->mems_allowed_seq);
include/linux/cpuset.h
176
task_unlock(current);
include/linux/cpuset.h
84
#define cpuset_current_mems_allowed (current->mems_allowed)
include/linux/cred.h
181
return rcu_replace_pointer(current->cred, override_cred, 1);
include/linux/cred.h
186
return rcu_replace_pointer(current->cred, revert_cred, 1);
include/linux/cred.h
296
rcu_dereference_protected(current->cred, 1)
include/linux/cred.h
305
rcu_dereference_protected(current->real_cred, 1)
include/linux/delayacct.h
134
if (current->delays)
include/linux/delayacct.h
159
if (current->delays)
include/linux/delayacct.h
168
if (current->delays)
include/linux/delayacct.h
177
if (current->delays)
include/linux/delayacct.h
186
if (current->delays)
include/linux/delayacct.h
195
if (current->delays)
include/linux/delayacct.h
204
if (current->delays)
include/linux/delayacct.h
213
if (current->delays)
include/linux/delayacct.h
222
if (current->delays)
include/linux/delayacct.h
231
if (current->delays)
include/linux/delayacct.h
240
if (current->delays)
include/linux/elf.h
17
set_personality(PER_LINUX | (current->personality & (~PER_MASK)))
include/linux/entry-common.h
101
rseq_syscall_enter_work(syscall_get_nr(current, regs));
include/linux/entry-common.h
118
syscall = syscall_get_nr(current, regs);
include/linux/entry-common.h
242
if (unlikely(current->syscall_dispatch.on_dispatch)) {
include/linux/entry-common.h
243
current->syscall_dispatch.on_dispatch = false;
include/linux/entry-common.h
251
trace_syscall_exit(regs, syscall_get_return_value(current, regs));
include/linux/entry-common.h
269
unsigned long nr = syscall_get_nr(current, regs);
include/linux/entry-common.h
76
syscall_get_arguments(current, regs, args);
include/linux/eventfd.h
45
return !current->in_eventfd;
include/linux/filter.h
1320
proglen, pass, image, current->comm, task_pid_nr(current));
include/linux/filter.h
795
struct task_struct *tsk = current;
include/linux/filter.h
808
current->bpf_net_context = NULL;
include/linux/filter.h
813
return current->bpf_net_context;
include/linux/freezer.h
59
if (likely(!freezing(current)))
include/linux/freezer.h
61
if (!(current->flags & PF_NOFREEZE))
include/linux/fs_struct.h
47
return current->fs->umask;
include/linux/fsnotify_backend.h
302
WARN_ON_ONCE(!(current->flags & PF_MEMALLOC_NOFS));
include/linux/ftrace.h
1344
atomic_inc(&current->tracing_graph_pause);
include/linux/ftrace.h
1349
atomic_dec(&current->tracing_graph_pause);
include/linux/hardirq.h
38
account_hardirq_enter(current); \
include/linux/hardirq.h
66
account_hardirq_exit(current); \
include/linux/hfs_common.h
18
current->pid, __FILE__, __LINE__, __func__, ##__VA_ARGS__) \
include/linux/highmem-internal.h
17
DEBUG_LOCKS_WARN_ON(current->kmap_ctrl.idx);
include/linux/hung_task.h
46
WARN_ON_ONCE(READ_ONCE(current->blocker));
include/linux/hung_task.h
55
WRITE_ONCE(current->blocker, lock_ptr | type);
include/linux/hung_task.h
60
WRITE_ONCE(current->blocker, 0UL);
include/linux/io_uring.h
19
if (current->io_uring)
include/linux/io_uring.h
24
if (current->io_uring)
include/linux/iocontext.h
124
if (!current->io_context)
include/linux/ioprio.h
66
if (p != current)
include/linux/ioprio.h
84
return __get_task_ioprio(current);
include/linux/irqflags.h
100
current->irq_config = 1; \
include/linux/irqflags.h
105
current->irq_config = 0; \
include/linux/irqflags.h
133
current->softirq_context++; \
include/linux/irqflags.h
137
current->softirq_context--; \
include/linux/irqflags.h
59
current->hardirq_threaded = 0; \
include/linux/irqflags.h
63
current->hardirq_threaded = 1; \
include/linux/irqflags.h
75
current->irq_config = 1; \
include/linux/irqflags.h
84
current->irq_config = 0; \
include/linux/irqflags.h
89
current->irq_config = 1; \
include/linux/irqflags.h
94
current->irq_config = 0; \
include/linux/jbd2.h
1494
return current->journal_info;
include/linux/kernel.h
122
# define non_block_start() (current->non_block_count++)
include/linux/kernel.h
128
# define non_block_end() WARN_ON(current->non_block_count-- == 0)
include/linux/kernel.h
99
# define sched_annotate_sleep() (current->task_state_change = 0)
include/linux/kvm_host.h
881
pr_err("kvm [%i]: " fmt, task_pid_nr(current), ## __VA_ARGS__)
include/linux/kvm_host.h
883
pr_info("kvm [%i]: " fmt, task_pid_nr(current), ## __VA_ARGS__)
include/linux/kvm_host.h
885
pr_debug("kvm [%i]: " fmt, task_pid_nr(current), ## __VA_ARGS__)
include/linux/kvm_host.h
887
pr_debug_ratelimited("kvm [%i]: " fmt, task_pid_nr(current), \
include/linux/kvm_host.h
891
task_tgid_nr(current), ## __VA_ARGS__)
include/linux/local_lock_internal.h
314
(rt_mutex_owner(&this_cpu_ptr(__lock)->lock) == current)
include/linux/local_lock_internal.h
48
l->owner = current;
include/linux/local_lock_internal.h
55
l->owner = current;
include/linux/local_lock_internal.h
60
DEBUG_LOCKS_WARN_ON(l->owner != current);
include/linux/lockdep.h
111
current->lockdep_recursion += LOCKDEP_OFF; \
include/linux/lockdep.h
116
current->lockdep_recursion -= LOCKDEP_OFF; \
include/linux/lockdep.h
300
lockdep_assert_once(!current->lockdep_depth)
include/linux/lockdep.h
643
WARN_ONCE(debug_locks && !current->lockdep_recursion && \
include/linux/lockdep.h
645
!(current->hardirq_threaded || current->irq_config), \
include/linux/memcontrol.h
1889
WARN_ON(current->in_user_fault);
include/linux/memcontrol.h
1890
current->in_user_fault = 1;
include/linux/memcontrol.h
1895
WARN_ON(!current->in_user_fault);
include/linux/memcontrol.h
1896
current->in_user_fault = 0;
include/linux/memcontrol.h
915
if (unlikely(current->memcg_nr_pages_over_high))
include/linux/misc_cgroup.h
95
return css_misc(task_get_css(current, misc_cgrp_id));
include/linux/mm.h
1281
if (!current->mm)
include/linux/mm.h
1284
if (current->mm != vma->vm_mm)
include/linux/mm.h
2090
pid_bit = hash_32(current->pid, ilog2(BITS_PER_LONG));
include/linux/mm.h
516
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
include/linux/mm_inline.h
123
return current->in_lru_fault;
include/linux/mman.h
204
if (!mm_flags_test(MMF_HAS_MDWE, current->mm))
include/linux/netdevice.h
3606
return current->net_xmit.recursion;
include/linux/netdevice.h
3611
return unlikely(current->net_xmit.recursion > XMIT_RECURSION_LIMIT);
include/linux/netdevice.h
3616
current->net_xmit.recursion++;
include/linux/netdevice.h
3621
current->net_xmit.recursion--;
include/linux/netdevice.h
4894
return rt_mutex_owner(&txq->_xmit_lock.lock) == current;
include/linux/netdevice.h
5311
current->net_xmit.more = more;
include/linux/netdevice.h
5316
return current->net_xmit.more;
include/linux/nsfs.h
29
struct cgroup_namespace *: current->nsproxy->cgroup_ns, \
include/linux/nsfs.h
30
struct ipc_namespace *: current->nsproxy->ipc_ns, \
include/linux/nsfs.h
31
struct net *: current->nsproxy->net_ns, \
include/linux/nsfs.h
32
struct pid_namespace *: task_active_pid_ns(current), \
include/linux/nsfs.h
33
struct mnt_namespace *: current->nsproxy->mnt_ns, \
include/linux/nsfs.h
34
struct time_namespace *: current->nsproxy->time_ns, \
include/linux/nsfs.h
36
struct uts_namespace *: current->nsproxy->uts_ns)
include/linux/oom.h
61
current->signal->oom_flag_origin = true;
include/linux/oom.h
66
current->signal->oom_flag_origin = false;
include/linux/perf_regs.h
46
regs_user->regs = task_pt_regs(current);
include/linux/perf_regs.h
47
regs_user->abi = perf_reg_abi(current);
include/linux/personality.h
15
#define set_personality(pers) (current->personality = (pers))
include/linux/pgtable.h
263
struct lazy_mmu_state *state = &current->lazy_mmu_state;
include/linux/pgtable.h
289
struct lazy_mmu_state *state = &current->lazy_mmu_state;
include/linux/pgtable.h
318
struct lazy_mmu_state *state = &current->lazy_mmu_state;
include/linux/pgtable.h
344
struct lazy_mmu_state *state = &current->lazy_mmu_state;
include/linux/preempt.h
111
# define softirq_count() (current->softirq_disable_cnt & SOFTIRQ_MASK)
include/linux/ptrace.h
150
if (unlikely(ptrace_event_enabled(current, event))) {
include/linux/ptrace.h
154
if ((current->ptrace & (PT_PTRACED|PT_SEIZED)) == PT_PTRACED)
include/linux/ptrace.h
155
send_sig(SIGTRAP, current, 0);
include/linux/ptrace.h
182
ns = task_active_pid_ns(rcu_dereference(current->parent));
include/linux/ptrace.h
208
if (unlikely(ptrace) && current->ptrace) {
include/linux/ptrace.h
209
child->ptrace = current->ptrace;
include/linux/ptrace.h
210
__ptrace_link(child, current->parent, current->ptracer_cred);
include/linux/ptrace.h
389
#define current_pt_regs() task_pt_regs(current)
include/linux/ptrace.h
409
int ptrace = current->ptrace;
include/linux/ptrace.h
424
send_sig(signr, current, 1);
include/linux/ptrace.h
426
return fatal_signal_pending(current);
include/linux/ratelimit.h
61
pr_warn("%s: %d output lines suppressed due to ratelimiting\n", current->comm, m);
include/linux/rcupdate.h
229
rcu_tasks_qs(current, false); \
include/linux/rcupdate.h
91
#define rcu_preempt_depth() READ_ONCE(current->rcu_read_lock_nesting)
include/linux/rcupdate_trace.h
123
struct task_struct *t = current;
include/linux/rcupdate_trace.h
98
struct task_struct *t = current;
include/linux/rcupdate_wait.h
76
return !list_empty(&current->rcu_node_entry);
include/linux/rcutiny.h
103
rcu_tasks_qs(current, (preempt)); \
include/linux/rcuwait.h
36
rcu_assign_pointer(w->task, current);
include/linux/rcuwait.h
54
if (signal_pending_state(state, current)) { \
include/linux/resume_user_mode.h
49
if (unlikely(task_work_pending(current)))
include/linux/resume_user_mode.h
53
if (unlikely(current->cached_requested_key)) {
include/linux/resume_user_mode.h
54
key_put(current->cached_requested_key);
include/linux/resume_user_mode.h
55
current->cached_requested_key = NULL;
include/linux/rseq.h
116
current->rseq.event.sched_switch)
include/linux/rseq.h
117
rseq_raise_notify_resume(current);
include/linux/rseq.h
146
t->rseq = current->rseq;
include/linux/rseq.h
16
if (current->rseq.event.slowpath)
include/linux/rseq.h
20
if (current->rseq.event.sched_switch & current->rseq.event.has_rseq)
include/linux/rseq.h
35
if (current->rseq.event.has_rseq & current->rseq.event.user_irq)
include/linux/rseq.h
38
if (current->rseq.event.has_rseq)
include/linux/rseq.h
90
if (current->rseq.event.has_rseq) {
include/linux/rseq.h
91
current->rseq.event.ids_changed = true;
include/linux/rseq.h
92
current->rseq.event.sched_switch = true;
include/linux/rseq.h
93
rseq_raise_notify_resume(current);
include/linux/rseq_entry.h
100
if (likely(!current->rseq.slice.state.granted))
include/linux/rseq_entry.h
115
struct task_struct *curr = current;
include/linux/rseq_entry.h
231
current->rseq.event.user_irq = true;
include/linux/rseq_entry.h
645
struct task_struct *t = current;
include/linux/rseq_entry.h
697
current->rseq.event.slowpath = true;
include/linux/rseq_entry.h
698
set_tsk_thread_flag(current, TIF_NOTIFY_RESUME);
include/linux/rseq_entry.h
719
struct rseq_event *ev = &current->rseq.event;
include/linux/rseq_entry.h
732
struct rseq_event *ev = &current->rseq.event;
include/linux/rseq_entry.h
749
struct rseq_event *ev = &current->rseq.event;
include/linux/sbitmap.h
597
.private = current, \
include/linux/sched.h
170
current->task_state_change = _THIS_IP_; \
include/linux/sched.h
1743
return __task_lazy_mmu_mode_active(current);
include/linux/sched.h
176
current->task_state_change = _THIS_IP_; \
include/linux/sched.h
1800
#define clear_used_math() clear_stopped_child_used_math(current)
include/linux/sched.h
1801
#define set_used_math() set_stopped_child_used_math(current)
include/linux/sched.h
1806
#define conditional_used_math(condition) conditional_stopped_child_used_math(condition, current)
include/linux/sched.h
1809
do { (child)->flags &= ~PF_USED_MATH, (child)->flags |= current->flags & PF_USED_MATH; } while (0)
include/linux/sched.h
181
current->saved_state_change = current->task_state_change;\
include/linux/sched.h
1813
#define used_math() tsk_used_math(current)
include/linux/sched.h
1817
return (current->flags & PF_NO_SETAFFINITY) &&
include/linux/sched.h
1818
(current->nr_cpus_allowed == 1);
include/linux/sched.h
182
current->task_state_change = _THIS_IP_; \
include/linux/sched.h
187
current->task_state_change = current->saved_state_change;\
include/linux/sched.h
1880
current->flags &= ~flags;
include/linux/sched.h
1881
current->flags |= orig_flags & flags;
include/linux/sched.h
2089
set_tsk_need_resched(current);
include/linux/sched.h
2197
WARN_ON_ONCE(p != current);
include/linux/sched.h
2337
swap(current->alloc_tag, tag);
include/linux/sched.h
2344
WARN(current->alloc_tag != tag, "current->alloc_tag was changed:\n");
include/linux/sched.h
2346
current->alloc_tag = old;
include/linux/sched.h
2402
struct task_struct *p = current;
include/linux/sched.h
2437
struct task_struct *p = current;
include/linux/sched.h
244
WRITE_ONCE(current->__state, (state_value)); \
include/linux/sched.h
251
smp_store_mb(current->__state, (state_value)); \
include/linux/sched.h
264
raw_spin_lock_irqsave(&current->pi_lock, flags); \
include/linux/sched.h
267
WRITE_ONCE(current->__state, (state_value)); \
include/linux/sched.h
268
raw_spin_unlock_irqrestore(&current->pi_lock, flags); \
include/linux/sched.h
299
raw_spin_lock(&current->pi_lock); \
include/linux/sched.h
300
current->saved_state = current->__state; \
include/linux/sched.h
303
WRITE_ONCE(current->__state, TASK_RTLOCK_WAIT); \
include/linux/sched.h
304
raw_spin_unlock(&current->pi_lock); \
include/linux/sched.h
310
raw_spin_lock(&current->pi_lock); \
include/linux/sched.h
312
trace_set_current_state(current->saved_state); \
include/linux/sched.h
313
WRITE_ONCE(current->__state, current->saved_state); \
include/linux/sched.h
314
current->saved_state = TASK_RUNNING; \
include/linux/sched.h
315
raw_spin_unlock(&current->pi_lock); \
include/linux/sched.h
318
#define get_current_state() READ_ONCE(current->__state)
include/linux/sched/mm.h
251
unsigned int pflags = READ_ONCE(current->flags);
include/linux/sched/mm.h
320
if (current->flags & PF_MEMALLOC)
include/linux/sched/mm.h
335
unsigned oldflags = ~current->flags & flags;
include/linux/sched/mm.h
336
current->flags |= flags;
include/linux/sched/mm.h
342
current->flags &= ~flags;
include/linux/sched/mm.h
500
old = current->active_memcg;
include/linux/sched/mm.h
501
current->active_memcg = memcg;
include/linux/sched/mm.h
544
if (current->mm != mm)
include/linux/sched/signal.h
289
struct task_struct *task = current;
include/linux/sched/signal.h
303
spin_lock_irq(&current->sighand->siglock);
include/linux/sched/signal.h
304
if (current->jobctl & JOBCTL_STOP_DEQUEUED) {
include/linux/sched/signal.h
305
current->jobctl |= JOBCTL_STOPPED;
include/linux/sched/signal.h
308
spin_unlock_irq(&current->sighand->siglock);
include/linux/sched/signal.h
377
set_tsk_thread_flag(current, TIF_SIGPENDING);
include/linux/sched/signal.h
428
(fatal_signal_pending(current) ||
include/linux/sched/signal.h
429
(user_mode(regs) && signal_pending(current))));
include/linux/sched/signal.h
513
current->restore_sigmask = true;
include/linux/sched/signal.h
521
current->restore_sigmask = false;
include/linux/sched/signal.h
525
return current->restore_sigmask;
include/linux/sched/signal.h
533
if (!current->restore_sigmask)
include/linux/sched/signal.h
535
current->restore_sigmask = false;
include/linux/sched/signal.h
543
__set_current_blocked(&current->saved_sigmask);
include/linux/sched/signal.h
551
WARN_ON(!signal_pending(current));
include/linux/sched/signal.h
558
sigset_t *res = &current->blocked;
include/linux/sched/signal.h
560
res = &current->saved_sigmask;
include/linux/sched/signal.h
576
return sp >= current->sas_ss_sp &&
include/linux/sched/signal.h
577
sp - current->sas_ss_sp < current->sas_ss_size;
include/linux/sched/signal.h
579
return sp > current->sas_ss_sp &&
include/linux/sched/signal.h
580
sp - current->sas_ss_sp <= current->sas_ss_size;
include/linux/sched/signal.h
598
if (current->sas_ss_flags & SS_AUTODISARM)
include/linux/sched/signal.h
606
if (!current->sas_ss_size)
include/linux/sched/signal.h
623
return current->sas_ss_sp;
include/linux/sched/signal.h
625
return current->sas_ss_sp + current->sas_ss_size;
include/linux/sched/signal.h
771
return task_rlimit(current, limit);
include/linux/sched/signal.h
776
return task_rlimit_max(current, limit);
include/linux/sched/task_stack.h
91
void *stack = task_stack_page(current);
include/linux/signal.h
462
struct task_struct *t = current; \
include/linux/sunrpc/types.h
22
#define signalled() (signal_pending(current))
include/linux/swait.h
54
.task = current, \
include/linux/swap.h
172
if (current->reclaim_state)
include/linux/swap.h
173
current->reclaim_state->reclaimed += pages;
include/linux/swap.h
38
return current->flags & PF_KSWAPD;
include/linux/syscalls.h
1302
unsigned int old = current->personality;
include/linux/task_io_accounting_ops.h
13
current->ioac.read_bytes += bytes;
include/linux/task_io_accounting_ops.h
27
current->ioac.write_bytes += bytes;
include/linux/task_io_accounting_ops.h
41
current->ioac.cancelled_write_bytes += bytes;
include/linux/thread_info.h
24
#define current_thread_info() ((struct thread_info *)current)
include/linux/time_namespace.h
107
struct time_namespace *ns = current->nsproxy->time_ns;
include/linux/time_namespace.h
76
struct timens_offsets *ns_offsets = &current->nsproxy->time_ns->offsets;
include/linux/time_namespace.h
83
struct timens_offsets *ns_offsets = &current->nsproxy->time_ns->offsets;
include/linux/time_namespace.h
90
struct timens_offsets *ns_offsets = &current->nsproxy->time_ns->offsets;
include/linux/time_namespace.h
97
struct timens_offsets *ns_offsets = &current->nsproxy->time_ns->offsets;
include/linux/trace_events.h
179
entry->pid = current->pid;
include/linux/trace_recursion.h
129
unsigned int val = READ_ONCE(current->trace_recursion);
include/linux/trace_recursion.h
153
current->trace_recursion = val;
include/linux/trace_recursion.h
58
#define trace_recursion_set(bit) do { (current)->trace_recursion |= (1<<(bit)); } while (0)
include/linux/trace_recursion.h
59
#define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(1<<(bit)); } while (0)
include/linux/trace_recursion.h
60
#define trace_recursion_test(bit) ((current)->trace_recursion & (1<<(bit)))
include/linux/uaccess.h
255
current->pagefault_disabled++;
include/linux/uaccess.h
260
current->pagefault_disabled--;
include/linux/uaccess.h
295
return current->pagefault_disabled != 0;
include/linux/unwind_deferred.h
36
struct unwind_task_info *info = &current->unwind_info;
include/linux/unwind_deferred.h
48
current->unwind_info.id.id = 0;
include/linux/user_events.h
40
if (!t || !current->user_event_mm)
include/linux/user_events.h
43
old_mm = current->user_event_mm;
include/linux/utsname.h
31
return &current->nsproxy->uts_ns->name;
include/linux/vtime.h
104
vtime_guest_exit(current);
include/linux/vtime.h
106
current->flags &= ~PF_VCPU;
include/linux/vtime.h
116
current->flags |= PF_VCPU;
include/linux/vtime.h
121
current->flags &= ~PF_VCPU;
include/linux/vtime.h
52
vtime_account_kernel(current);
include/linux/vtime.h
53
current->flags |= PF_VCPU;
include/linux/vtime.h
58
vtime_account_kernel(current);
include/linux/vtime.h
59
current->flags &= ~PF_VCPU;
include/linux/vtime.h
96
vtime_guest_enter(current);
include/linux/vtime.h
98
current->flags |= PF_VCPU;
include/linux/wait.h
1235
.private = current, \
include/linux/wait.h
1244
(wait)->private = current; \
include/linux/wait.h
554
current->timer_slack_ns); \
include/linux/wait_bit.h
42
.private = current, \
include/linux/ww_mutex.h
147
ctx->task = current;
include/net/busy_poll.h
47
return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current);
include/net/cls_cgroup.h
44
classid = task_cls_classid(current);
include/net/cls_cgroup.h
55
u32 classid = __task_get_classid(current);
include/net/netprio_cgroup.h
39
sock_cgroup_set_prioidx(skcd, task_netprioidx(current));
include/net/scm.h
98
scm_set_cred(scm, task_tgid(current), current_uid(), current_gid());
include/net/sock.h
2624
return &current->task_frag;
include/net/xfrm.h
827
audit_get_loginuid(current) :
include/net/xfrm.h
829
const unsigned int ses = task_valid ? audit_get_sessionid(current) :
include/trace/events/block.h
216
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
include/trace/events/block.h
354
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
include/trace/events/block.h
437
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
include/trace/events/block.h
456
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
include/trace/events/block.h
507
memcpy(__entry->comm, current->comm, TASK_COMM_LEN);
include/trace/events/kmem.h
449
__entry->curr = current->mm == mm && !(current->flags & PF_KTHREAD);
include/trace/events/mmap.h
29
__entry->total_vm = current->mm->total_vm;
include/trace/events/power_cpu_migrate.h
46
__define_cpu_migrate_event(current);
include/trace/events/sched.h
196
BUG_ON(p != current);
include/trace/events/sched.h
378
__string( comm, current->comm )
include/trace/events/sched.h
386
__entry->prio = current->prio; /* XXX SCHED_DEADLINE */
include/trace/events/syscalls.h
31
syscall_get_arguments(current, regs, __entry->args);
include/trace/events/syscalls.h
56
__entry->id = syscall_get_nr(current, regs);
include/trace/events/writeback.h
699
__entry->dirtied_pause = current->nr_dirtied_pause;
include/trace/events/writeback.h
700
__entry->think = current->dirty_paused_when == 0 ? 0 :
include/trace/events/writeback.h
701
(long)(jiffies - current->dirty_paused_when) * 1000/HZ;
init/do_mounts.c
164
s = current->fs->pwd.dentry->d_sb;
init/main.c
1314
printk(KERN_DEBUG "calling %pS @ %i\n", fn, task_pid_nr(current));
init/main.c
1673
cad_pid = get_pid(task_pid(current));
io_uring/advise.c
59
ret = do_madvise(current->mm, ma->addr, ma->len, ma->advice);
io_uring/cancel.c
111
WARN_ON_ONCE(!io_wq_current_is_worker() && tctx != current->io_uring);
io_uring/cancel.c
298
ret = __io_sync_cancel(current->io_uring, &cd, sc.fd);
io_uring/cancel.c
322
ret = __io_sync_cancel(current->io_uring, &cd, sc.fd);
io_uring/cancel.c
579
struct io_uring_task *tctx = current->io_uring;
io_uring/cancel.c
586
WARN_ON_ONCE(sqd && sqpoll_task_locked(sqd) != current);
io_uring/cancel.c
588
if (!current->io_uring)
io_uring/cancel.c
597
io_uring_drop_tctx_refs(current);
io_uring/cancel.c
612
current->io_uring,
io_uring/cancel.c
619
current->io_uring,
io_uring/cancel.c
631
io_uring_drop_tctx_refs(current);
io_uring/cancel.c
635
node->ctx->submitter_task != current);
io_uring/cancel.c
658
__io_uring_free(current);
io_uring/io-wq.c
1272
task_rlimit(current, RLIMIT_NPROC);
io_uring/io-wq.c
1487
if (new_count[i] > task_rlimit(current, RLIMIT_NPROC))
io_uring/io-wq.c
1488
new_count[i] = task_rlimit(current, RLIMIT_NPROC);
io_uring/io-wq.c
200
struct io_worker *worker = current->worker_private;
io_uring/io-wq.c
260
current->worker_private = NULL;
io_uring/io-wq.c
694
set_task_comm(current, buf);
io_uring/io-wq.c
729
if (signal_pending(current)) {
io_uring/io-wq.c
812
if (fatal_signal_pending(current))
io_uring/io-wq.h
83
return in_task() && (current->flags & PF_IO_WORKER) &&
io_uring/io-wq.h
84
current->worker_private;
io_uring/io_uring.c
1237
if (task_work_pending(current) || list_empty(&ctx->iopoll_list)) {
io_uring/io_uring.c
1250
if (task_sigpending(current))
io_uring/io_uring.c
1731
req->tctx = current->io_uring;
io_uring/io_uring.c
1755
current->io_uring->cached_refs++;
io_uring/io_uring.c
2055
current->io_uring->cached_refs += left;
io_uring/io_uring.c
2287
struct io_uring_task *tctx = current->io_uring;
io_uring/io_uring.c
2558
struct io_uring_task *tctx = current->io_uring;
io_uring/io_uring.c
2997
if (current->io_uring_restrict)
io_uring/io_uring.c
2998
io_ctx_restriction_clone(ctx, current->io_uring_restrict);
io_uring/io_uring.c
3006
mmgrab(current->mm);
io_uring/io_uring.c
3007
ctx->mm_account = current->mm;
io_uring/io_uring.c
3026
ctx->submitter_task = get_task_struct(current);
io_uring/io_uring.c
3037
tctx = current->io_uring;
io_uring/io_uring.c
412
if ((current->flags & PF_KTHREAD) || !tctx->io_wq) {
io_uring/io_uring.c
427
if (WARN_ON_ONCE(!same_thread_group(tctx->task, current)))
io_uring/io_uring.c
597
if (likely(tctx->task == current)) {
io_uring/io_uring.c
612
refcount_add(refill, &current->usage);
io_uring/io_uring.h
236
lockdep_assert(current == ctx->submitter_task);
io_uring/io_uring.h
506
struct io_uring_task *tctx = current->io_uring;
io_uring/msg_ring.c
214
if (!(current->flags & PF_EXITING))
io_uring/napi.c
149
if (signal_pending(current))
io_uring/notif.c
126
notif->tctx = current->io_uring;
io_uring/openclose.c
238
struct files_struct *files = current->files;
io_uring/openclose.c
267
ret = filp_close(file, current->files);
io_uring/query.c
131
if (fatal_signal_pending(current))
io_uring/register.c
209
if (current->io_uring_restrict)
io_uring/register.c
215
if (!task_no_new_privs(current) &&
io_uring/register.c
238
current->io_uring_restrict = res;
io_uring/register.c
251
if (!task_no_new_privs(current) &&
io_uring/register.c
259
res = current->io_uring_restrict;
io_uring/register.c
268
if (res != current->io_uring_restrict)
io_uring/register.c
272
if (!current->io_uring_restrict)
io_uring/register.c
273
current->io_uring_restrict = res;
io_uring/register.c
283
ctx->submitter_task = get_task_struct(current);
io_uring/register.c
305
ret = io_wq_cpu_affinity(current->io_uring, new_mask);
io_uring/register.c
387
tctx = current->io_uring;
io_uring/register.c
733
if (ctx->submitter_task && ctx->submitter_task != current)
io_uring/register.c
963
struct io_uring_task *tctx = current->io_uring;
io_uring/rsrc.c
1286
src_ctx->submitter_task != current) {
io_uring/sqpoll.c
132
if (sqd->task_tgid != current->tgid)
io_uring/sqpoll.c
193
sqd->work_time += io_sq_cpu_usec(current) - ist->usec;
io_uring/sqpoll.c
201
ist->usec = io_sq_cpu_usec(current);
io_uring/sqpoll.c
251
signal_pending(current)) {
io_uring/sqpoll.c
253
if (signal_pending(current))
io_uring/sqpoll.c
270
struct io_uring_task *tctx = current->io_uring;
io_uring/sqpoll.c
281
if (task_work_pending(current))
io_uring/sqpoll.c
288
struct io_uring_task *tctx = current->io_uring;
io_uring/sqpoll.c
303
if (!current->io_uring) {
io_uring/sqpoll.c
306
put_task_struct(current);
io_uring/sqpoll.c
312
set_task_comm(current, buf);
io_uring/sqpoll.c
315
sqd->task_pid = current->pid;
io_uring/sqpoll.c
318
set_cpus_allowed_ptr(current, cpumask_of(sqd->sq_cpu));
io_uring/sqpoll.c
320
set_cpus_allowed_ptr(current, cpu_online_mask);
io_uring/sqpoll.c
337
if (io_sqd_events_pending(sqd) || signal_pending(current)) {
io_uring/sqpoll.c
36
WARN_ON_ONCE(sqpoll_task_locked(sqd) == current);
io_uring/sqpoll.c
419
put_task_struct(current);
io_uring/sqpoll.c
441
} while (!signal_pending(current));
io_uring/sqpoll.c
504
cpuset_cpus_allowed(current, allowed_mask);
io_uring/sqpoll.c
515
sqd->task_pid = current->pid;
io_uring/sqpoll.c
516
sqd->task_tgid = current->tgid;
io_uring/sqpoll.c
60
WARN_ON_ONCE(tsk == current);
io_uring/sqpoll.c
75
WARN_ON_ONCE(tsk == current);
io_uring/tctx.c
114
struct io_uring_task *tctx = current->io_uring;
io_uring/tctx.c
119
ret = io_uring_alloc_task_context(current, ctx);
io_uring/tctx.c
123
tctx = current->io_uring;
io_uring/tctx.c
146
node->task = current;
io_uring/tctx.c
167
&& ctx->submitter_task != current)
io_uring/tctx.c
174
current->io_uring->last = ctx;
io_uring/tctx.c
183
struct io_uring_task *tctx = current->io_uring;
io_uring/tctx.c
192
WARN_ON_ONCE(current != node->task);
io_uring/tctx.c
229
struct io_uring_task *tctx = current->io_uring;
io_uring/tctx.c
299
tctx = current->io_uring;
io_uring/tctx.c
345
struct io_uring_task *tctx = current->io_uring;
io_uring/tctx.h
27
struct io_uring_task *tctx = current->io_uring;
io_uring/tw.c
137
io_uring_drop_tctx_refs(current);
io_uring/tw.c
323
if (WARN_ON_ONCE(ctx->submitter_task != current))
io_uring/tw.h
107
return likely(ctx->submitter_task == current);
io_uring/tw.h
113
ctx->submitter_task == current);
io_uring/tw.h
22
return (current->flags & (PF_EXITING | PF_KTHREAD)) || percpu_ref_is_dying(&ctx->refs);
io_uring/tw.h
67
if (current->flags & PF_IO_WORKER) {
io_uring/tw.h
72
if (current->io_uring) {
io_uring/tw.h
76
tctx_task_work_run(current->io_uring, UINT_MAX, &count);
io_uring/tw.h
81
if (task_work_pending(current)) {
io_uring/tw.h
97
return task_work_pending(current) || io_local_work_pending(ctx);
io_uring/wait.c
155
current->in_iowait = 1;
io_uring/wait.c
160
current->in_iowait = 0;
io_uring/wait.c
174
if (unlikely(task_work_pending(current)))
io_uring/wait.c
176
if (unlikely(task_sigpending(current)))
io_uring/wait.c
216
iowq.wq.private = current;
io_uring/wait.c
40
if (task_sigpending(current))
io_uring/wait.c
47
struct io_uring_task *tctx = current->io_uring;
io_uring/waitid.c
220
iw->head = &current->signal->wait_chldexit;
io_uring/waitid.c
312
iw->head = &current->signal->wait_chldexit;
io_uring/zcrx.c
1159
if (fatal_signal_pending(current))
io_uring/zcrx.c
822
ifq->netdev = netdev_get_by_index_lock(current->nsproxy->net_ns, reg.if_idx);
ipc/ipc_sysctl.c
185
return &current->nsproxy->ipc_ns->ipc_set;
ipc/ipc_sysctl.c
190
return &current->nsproxy->ipc_ns->ipc_set == set;
ipc/mq_sysctl.c
71
return &current->nsproxy->ipc_ns->mq_set;
ipc/mq_sysctl.c
76
return &current->nsproxy->ipc_ns->mq_set == set;
ipc/mqueue.c
1108
wait.task = current;
ipc/mqueue.c
1201
wait.task = current;
ipc/mqueue.c
1337
if (info->notify_owner == task_tgid(current)) {
ipc/mqueue.c
1360
info->notify_self_exec_id = current->self_exec_id;
ipc/mqueue.c
1364
info->notify_owner = get_pid(task_tgid(current));
ipc/mqueue.c
456
ctx->ipc_ns = get_ipc_ns(current->nsproxy->ipc_ns);
ipc/mqueue.c
663
if (task_tgid(current) == info->notify_owner)
ipc/mqueue.c
695
if (walk->task->prio <= current->prio) {
ipc/mqueue.c
738
if (signal_pending(current)) {
ipc/mqueue.c
804
sig_i.si_pid = task_tgid_nr_ns(current,
ipc/mqueue.c
914
struct vfsmount *mnt = current->nsproxy->ipc_ns->mq_mnt;
ipc/mqueue.c
945
struct ipc_namespace *ipc_ns = current->nsproxy->ipc_ns;
ipc/msg.c
1081
!security_msg_queue_msgrcv(&msq->q_perm, msg, current,
ipc/msg.c
1107
ns = current->nsproxy->ipc_ns;
ipc/msg.c
1166
ipc_update_pid(&msq->q_lrpid, task_tgid(current));
ipc/msg.c
1182
msr_d.r_tsk = current;
ipc/msg.c
1240
if (signal_pending(current)) {
ipc/msg.c
196
mss->tsk = current;
ipc/msg.c
307
ns = current->nsproxy->ipc_ns;
ipc/msg.c
606
ns = current->nsproxy->ipc_ns;
ipc/msg.c
738
ns = current->nsproxy->ipc_ns;
ipc/msg.c
857
ns = current->nsproxy->ipc_ns;
ipc/msg.c
929
if (signal_pending(current)) {
ipc/msg.c
936
ipc_update_pid(&msq->q_lspid, task_tgid(current));
ipc/sem.c
1086
current->comm, task_pid_nr(current));
ipc/sem.c
1395
ipc_update_pid(&curr->sempid, task_tgid(current));
ipc/sem.c
1515
ipc_update_pid(&sma->sems[i].sempid, task_tgid(current));
ipc/sem.c
1661
ns = current->nsproxy->ipc_ns;
ipc/sem.c
1781
ns = current->nsproxy->ipc_ns;
ipc/sem.c
1854
undo_list = current->sysvsem.undo_list;
ipc/sem.c
1863
current->sysvsem.undo_list = undo_list;
ipc/sem.c
2097
queue.pid = task_tgid(current);
ipc/sem.c
2159
queue.sleeper = current;
ipc/sem.c
2167
current->timer_slack_ns, HRTIMER_MODE_ABS);
ipc/sem.c
2211
} while (error == -EINTR && !signal_pending(current)); /* spurious */
ipc/sem.c
2230
ns = current->nsproxy->ipc_ns;
ipc/sem.c
2434
ipc_update_pid(&semaphore->sempid, task_tgid(current));
ipc/sem.c
612
ns = current->nsproxy->ipc_ns;
ipc/shm.c
1251
ns = current->nsproxy->ipc_ns;
ipc/shm.c
1443
ns = current->nsproxy->ipc_ns;
ipc/shm.c
1580
ns = current->nsproxy->ipc_ns;
ipc/shm.c
1648
if (mmap_write_lock_killable(current->mm)) {
ipc/shm.c
1658
if (find_vma_intersection(current->mm, addr, addr + size))
ipc/shm.c
1668
mmap_write_unlock(current->mm);
ipc/shm.c
1730
struct mm_struct *mm = current->mm;
ipc/shm.c
297
ipc_update_pid(&shp->shm_lprid, task_tgid(current));
ipc/shm.c
387
ipc_update_pid(&shp->shm_lprid, task_tgid(current));
ipc/shm.c
770
shp->shm_cprid = get_pid(task_tgid(current));
ipc/shm.c
777
shp->shm_creator = current;
ipc/shm.c
786
task_lock(current);
ipc/shm.c
787
list_add(&shp->shm_clist, &current->sysvshm.shm_clist);
ipc/shm.c
788
task_unlock(current);
ipc/shm.c
838
ns = current->nsproxy->ipc_ns;
ipc/util.c
906
iter->ns = get_ipc_ns(current->nsproxy->ipc_ns);
ipc/util.c
907
iter->pid_ns = get_pid_ns(task_active_pid_ns(current));
kernel/acct.c
220
struct pid_namespace *ns = task_active_pid_ns(current);
kernel/acct.c
306
pin_kill(task_active_pid_ns(current)->bacct);
kernel/acct.c
432
struct pacct_struct *pacct = &current->signal->pacct;
kernel/acct.c
458
strscpy(ac->ac_comm, current->comm, sizeof(ac->ac_comm));
kernel/acct.c
462
run_time -= current->group_leader->start_time;
kernel/acct.c
487
spin_lock_irq(&current->sighand->siglock);
kernel/acct.c
488
tty = current->signal->tty; /* Safe as we hold the siglock */
kernel/acct.c
497
spin_unlock_irq(&current->sighand->siglock);
kernel/acct.c
510
ac->ac_pid = task_tgid_nr_ns(current, ns);
kernel/acct.c
512
ac->ac_ppid = task_tgid_nr_ns(rcu_dereference(current->real_parent), ns);
kernel/acct.c
546
current->signal->rlim[RLIMIT_FSIZE].rlim_cur = RLIM_INFINITY;
kernel/acct.c
549
current->signal->rlim[RLIMIT_FSIZE].rlim_cur = flim;
kernel/acct.c
559
struct pacct_struct *pacct = &current->signal->pacct;
kernel/acct.c
563
if (group_dead && current->mm) {
kernel/acct.c
564
struct mm_struct *mm = current->mm;
kernel/acct.c
574
spin_lock_irq(&current->sighand->siglock);
kernel/acct.c
577
if (thread_group_leader(current)) {
kernel/acct.c
579
if (current->flags & PF_FORKNOEXEC)
kernel/acct.c
582
if (current->flags & PF_SUPERPRIV)
kernel/acct.c
584
if (current->flags & PF_DUMPCORE)
kernel/acct.c
586
if (current->flags & PF_SIGNALED)
kernel/acct.c
589
task_cputime(current, &utime, &stime);
kernel/acct.c
592
pacct->ac_minflt += current->min_flt;
kernel/acct.c
593
pacct->ac_majflt += current->maj_flt;
kernel/acct.c
594
spin_unlock_irq(&current->sighand->siglock);
kernel/acct.c
621
for (ns = task_active_pid_ns(current); ns != NULL; ns = ns->parent) {
kernel/async.c
126
entry->func, task_pid_nr(current));
kernel/async.c
314
pr_debug("async_waiting @ %i\n", task_pid_nr(current));
kernel/async.c
319
pr_debug("async_continuing @ %i after %lli usec\n", task_pid_nr(current),
kernel/audit.c
1108
if (task_active_pid_ns(current) != &init_pid_ns)
kernel/audit.c
1131
pid_t pid = task_tgid_nr(current);
kernel/audit.c
1317
struct pid *req_pid = task_tgid(current);
kernel/audit.c
1541
t = READ_ONCE(current->signal->audit_tty);
kernel/audit.c
1562
t = READ_ONCE(current->signal->audit_tty);
kernel/audit.c
1565
t = xchg(&current->signal->audit_tty, t);
kernel/audit.c
1625
DECLARE_WAITQUEUE(wait, current);
kernel/audit.c
1642
char comm[sizeof(current->comm)];
kernel/audit.c
1655
task_tgid_nr(current),
kernel/audit.c
1657
from_kuid(&init_user_ns, audit_get_loginuid(current)),
kernel/audit.c
1659
audit_get_sessionid(current));
kernel/audit.c
1663
audit_log_untrustedstring(ab, get_task_comm(comm, current));
kernel/audit.c
1664
audit_log_d_path_exe(ab, current->mm); /* exe= */
kernel/audit.c
1926
if (!(auditd_test_task(current) || audit_ctl_owner_current())) {
kernel/audit.c
1939
DECLARE_WAITQUEUE(wait, current);
kernel/audit.c
2212
unsigned int sessionid = audit_get_sessionid(current);
kernel/audit.c
2213
uid_t auid = from_kuid(&init_user_ns, audit_get_loginuid(current));
kernel/audit.c
2424
spin_lock_irqsave(&current->sighand->siglock, flags);
kernel/audit.c
2425
if (current->signal)
kernel/audit.c
2426
tty = tty_kref_get(current->signal->tty);
kernel/audit.c
2427
spin_unlock_irqrestore(&current->sighand->siglock, flags);
kernel/audit.c
2439
char comm[sizeof(current->comm)];
kernel/audit.c
2451
task_ppid_nr(current),
kernel/audit.c
2452
task_tgid_nr(current),
kernel/audit.c
2453
from_kuid(&init_user_ns, audit_get_loginuid(current)),
kernel/audit.c
2463
audit_get_sessionid(current));
kernel/audit.c
2466
audit_log_untrustedstring(ab, get_task_comm(comm, current));
kernel/audit.c
2467
audit_log_d_path_exe(ab, current->mm);
kernel/audit.c
249
audit_cmd_mutex.owner = current;
kernel/audit.c
2656
if (!audit_loginuid_set(current))
kernel/audit.c
2686
uid = from_kuid(&init_user_ns, task_uid(current));
kernel/audit.c
2691
audit_log_format(ab, "pid=%d uid=%u", task_tgid_nr(current), uid);
kernel/audit.c
270
return (current == audit_cmd_mutex.owner);
kernel/audit.c
2714
oldloginuid = audit_get_loginuid(current);
kernel/audit.c
2715
oldsessionid = audit_get_sessionid(current);
kernel/audit.c
2728
current->sessionid = sessionid;
kernel/audit.c
2729
current->loginuid = loginuid;
kernel/audit.c
2750
audit_sig_pid = task_tgid_nr(current);
kernel/audit.c
2751
auid = audit_get_loginuid(current);
kernel/audit_watch.c
264
audit_filter_inodes(current, audit_context());
kernel/audit_watch.c
534
if (tsk != current)
kernel/audit_watch.c
537
if (!current->mm)
kernel/audit_watch.c
539
exe_file = get_mm_exe_file(current->mm);
kernel/auditfilter.c
1353
pid = task_tgid_nr(current);
kernel/auditfilter.c
1363
result = audit_uid_comparator(audit_get_loginuid(current),
kernel/auditfilter.c
1367
result = audit_comparator(audit_loginuid_set(current),
kernel/auditfilter.c
1386
result = audit_exe_compare(current, e->rule.exe);
kernel/auditsc.c
1135
const char __user *p = (const char __user *)current->mm->arg_start;
kernel/auditsc.c
1190
send_sig(SIGKILL, current, 0);
kernel/auditsc.c
1594
res = get_cmdline(current, buf, MAX_PROCTITLE_AUDIT_LEN);
kernel/auditsc.c
1638
task_ppid_nr(current), task_tgid_nr(current),
kernel/auditsc.c
1660
context->personality = current->personality;
kernel/auditsc.c
1818
if (tsk == current && !context->dummy) {
kernel/auditsc.c
1944
audit_filter_syscall(current, ctx);
kernel/auditsc.c
1946
audit_filter_uring(current, ctx);
kernel/auditsc.c
1947
audit_filter_inodes(current, ctx);
kernel/auditsc.c
1960
audit_filter_uring(current, ctx);
kernel/auditsc.c
1961
audit_filter_inodes(current, ctx);
kernel/auditsc.c
2009
if (auditd_test_task(current))
kernel/auditsc.c
2013
context->arch = syscall_get_arch(current);
kernel/auditsc.c
2049
audit_filter_syscall(current, context);
kernel/auditsc.c
2050
audit_filter_inodes(current, context);
kernel/auditsc.c
2168
get_fs_pwd(current->fs, &context->pwd);
kernel/auditsc.c
2787
context->capset.pid = task_tgid_nr(current);
kernel/auditsc.c
2869
char comm[sizeof(current->comm)];
kernel/auditsc.c
2877
audit_log_format(ab, " pid=%u", task_tgid_nr(current));
kernel/auditsc.c
2880
audit_log_untrustedstring(ab, get_task_comm(comm, current));
kernel/auditsc.c
2890
char comm[sizeof(current->comm)];
kernel/auditsc.c
2892
auid = audit_get_loginuid(current);
kernel/auditsc.c
2893
sessionid = audit_get_sessionid(current);
kernel/auditsc.c
2902
audit_log_format(ab, " pid=%d comm=", task_tgid_nr(current));
kernel/auditsc.c
2903
audit_log_untrustedstring(ab, get_task_comm(comm, current));
kernel/auditsc.c
2904
audit_log_d_path_exe(ab, current->mm);
kernel/auditsc.c
2953
signr, syscall_get_arch(current), syscall,
kernel/auditsc.c
2954
in_compat_syscall(), KSTK_EIP(current), code);
kernel/auditsc.c
479
cred = rcu_dereference_check(tsk->cred, tsk == current || task_creation);
kernel/bpf/bpf_lsm.c
213
run_ctx = container_of(current->bpf_ctx, struct bpf_trace_run_ctx, run_ctx);
kernel/bpf/cgroup.c
150
cgrp = task_dfl_cgroup(current);
kernel/bpf/cgroup.c
1744
cgrp = task_dfl_cgroup(current);
kernel/bpf/cgroup.c
1764
ctx = container_of(current->bpf_ctx, struct bpf_cg_run_ctx, run_ctx);
kernel/bpf/cgroup.c
1786
container_of(current->bpf_ctx, struct bpf_cg_run_ctx, run_ctx);
kernel/bpf/cgroup.c
1800
container_of(current->bpf_ctx, struct bpf_cg_run_ctx, run_ctx);
kernel/bpf/cgroup.c
1931
cgrp = task_dfl_cgroup(current);
kernel/bpf/cgroup.c
2097
.current_task = current,
kernel/bpf/cgroup.c
2183
.current_task = current,
kernel/bpf/cgroup_iter.c
259
current->nsproxy->cgroup_ns);
kernel/bpf/devmap.c
1036
return __dev_map_hash_update_elem(current->nsproxy->net_ns,
kernel/bpf/devmap.c
976
return __dev_map_update_elem(current->nsproxy->net_ns,
kernel/bpf/helpers.c
227
struct task_struct *task = current;
kernel/bpf/helpers.c
243
struct task_struct *task = current;
kernel/bpf/helpers.c
263
struct task_struct *task = current;
kernel/bpf/helpers.c
2710
return task_under_cgroup_hierarchy(current, cgrp);
kernel/bpf/helpers.c
412
cgrp = task_dfl_cgroup(current);
kernel/bpf/helpers.c
432
cgrp = task_dfl_cgroup(current);
kernel/bpf/helpers.c
593
struct task_struct *task = current;
kernel/bpf/memalloc.c
412
rcu_request_urgent_qs_task(current);
kernel/bpf/net_namespace.c
311
net = current->nsproxy->net_ns;
kernel/bpf/net_namespace.c
400
ret = __netns_bpf_prog_detach(current->nsproxy->net_ns, type, prog);
kernel/bpf/offload.c
242
netdev = dev_get_by_index(current->nsproxy->net_ns, attr->prog_ifindex);
kernel/bpf/offload.c
506
struct net *net = current->nsproxy->net_ns;
kernel/bpf/stackmap.c
178
if (!user || !current || !current->mm || irq_work_busy ||
kernel/bpf/stackmap.c
179
!mmap_read_trylock(current->mm)) {
kernel/bpf/stackmap.c
196
vma = find_vma(current->mm, ip);
kernel/bpf/stackmap.c
209
bpf_mmap_unlock_mm(work, current->mm);
kernel/bpf/stackmap.c
431
bool crosstask = task && task != current;
kernel/bpf/stream.c
392
current->pid, current->comm);
kernel/bpf/task_iter.c
232
tgid = pid_nr_ns(pid, task_active_pid_ns(current));
kernel/bpf/task_iter.c
385
common->ns = get_pid_ns(task_active_pid_ns(current));
kernel/bpf/tcx.c
125
struct net *net = current->nsproxy->net_ns;
kernel/bpf/tcx.c
13
struct net *net = current->nsproxy->net_ns;
kernel/bpf/tcx.c
312
struct net *net = current->nsproxy->net_ns;
kernel/bpf/tcx.c
62
struct net *net = current->nsproxy->net_ns;
kernel/bpf/verifier.c
21306
if (signal_pending(current))
kernel/capability.c
110
if (pid && (pid != task_pid_vnr(current))) {
kernel/capability.c
123
ret = security_capget(current, pEp, pIp, pPp);
kernel/capability.c
233
if (pid != 0 && pid != task_pid_vnr(current))
kernel/capability.c
344
current->flags |= PF_SUPERPRIV;
kernel/capability.c
42
current->comm);
kernel/capability.c
64
current->comm);
kernel/cgroup/cgroup-v1.c
1109
task_tgid_nr(current), current->comm);
kernel/cgroup/cgroup-v1.c
292
struct pid_namespace *ns = task_active_pid_ns(current);
kernel/cgroup/cgroup-v1.c
327
l->key.ns = get_pid_ns(task_active_pid_ns(current));
kernel/cgroup/cgroup.c
1488
cset = current->nsproxy->cgroup_ns->root_cset;
kernel/cgroup/cgroup.c
1516
if (current->nsproxy) {
kernel/cgroup/cgroup.c
1517
cset = current->nsproxy->cgroup_ns->root_cset;
kernel/cgroup/cgroup.c
2047
if (current->nsproxy->cgroup_ns == &init_cgroup_ns) {
kernel/cgroup/cgroup.c
2358
ctx->ns = current->nsproxy->cgroup_ns;
kernel/cgroup/cgroup.c
3068
tsk = current;
kernel/cgroup/cgroup.c
4260
ctx->ns = current->nsproxy->cgroup_ns;
kernel/cgroup/cgroup.c
6675
current->nsproxy->cgroup_ns);
kernel/cgroup/cgroup.c
6780
cgroup_threadgroup_change_begin(current);
kernel/cgroup/cgroup.c
6783
cset = task_css_set(current);
kernel/cgroup/cgroup.c
6840
current->nsproxy->cgroup_ns);
kernel/cgroup/cgroup.c
6855
cgroup_threadgroup_change_end(current);
kernel/cgroup/cgroup.c
6878
cgroup_threadgroup_change_end(current);
kernel/cgroup/cgroup.c
7423
cset = task_css_set(current);
kernel/cgroup/cpuset-v1.c
144
fmeter_markevent(&task_cs(current)->fmeter);
kernel/cgroup/cpuset-v1.c
431
current->nsproxy->cgroup_ns);
kernel/cgroup/cpuset.c
2573
if (task_work_add(current, flush_cb, TWA_RESUME))
kernel/cgroup/cpuset.c
2775
ret = task_cs(current) == cpuset_being_rebound;
kernel/cgroup/cpuset.c
3567
same_cs = (cs == task_cs(current));
kernel/cgroup/cpuset.c
3605
same_cs = (cs == task_cs(current));
kernel/cgroup/cpuset.c
3626
same_cs = (cs == task_cs(current));
kernel/cgroup/cpuset.c
3633
set_cpus_allowed_ptr(task, current->cpus_ptr);
kernel/cgroup/cpuset.c
3634
task->mems_allowed = current->mems_allowed;
kernel/cgroup/cpuset.c
4116
nodes_setall(current->mems_allowed);
kernel/cgroup/cpuset.c
4149
return nodes_intersects(*nodemask, current->mems_allowed);
kernel/cgroup/cpuset.c
4213
if (node_isset(node, current->mems_allowed))
kernel/cgroup/cpuset.c
4219
if (unlikely(tsk_is_oom_victim(current)))
kernel/cgroup/cpuset.c
4224
if (current->flags & PF_EXITING) /* Let dying task have memory */
kernel/cgroup/cpuset.c
4230
cs = nearest_hardwall_ancestor(task_cs(current));
kernel/cgroup/cpuset.c
4319
return *rotor = next_node_in(*rotor, current->mems_allowed);
kernel/cgroup/cpuset.c
4327
if (current->cpuset_mem_spread_rotor == NUMA_NO_NODE)
kernel/cgroup/cpuset.c
4328
current->cpuset_mem_spread_rotor =
kernel/cgroup/cpuset.c
4329
node_random(&current->mems_allowed);
kernel/cgroup/cpuset.c
4331
return cpuset_spread_node(&current->cpuset_mem_spread_rotor);
kernel/cgroup/cpuset.c
4363
cgrp = task_cs(current)->css.cgroup;
kernel/cgroup/cpuset.c
4367
nodemask_pr_args(&current->mems_allowed));
kernel/cgroup/debug.c
52
cset = task_css_set(current);
kernel/cgroup/debug.c
80
count = refcount_read(&task_css_set(current)->refcount);
kernel/cgroup/debug.c
96
cset = task_css_set(current);
kernel/cgroup/dmem.c
105
return css_to_dmemcs(task_get_css(current, dmem_cgrp_id));
kernel/cgroup/freezer.c
108
if (current->frozen)
kernel/cgroup/freezer.c
112
current->frozen = true;
kernel/cgroup/freezer.c
113
cgrp = task_dfl_cgroup(current);
kernel/cgroup/freezer.c
133
cgrp = task_dfl_cgroup(current);
kernel/cgroup/freezer.c
137
WARN_ON_ONCE(!current->frozen);
kernel/cgroup/freezer.c
138
current->frozen = false;
kernel/cgroup/freezer.c
139
} else if (!(current->jobctl & JOBCTL_TRAP_FREEZE)) {
kernel/cgroup/freezer.c
140
spin_lock(&current->sighand->siglock);
kernel/cgroup/freezer.c
141
current->jobctl |= JOBCTL_TRAP_FREEZE;
kernel/cgroup/freezer.c
143
spin_unlock(&current->sighand->siglock);
kernel/cgroup/namespace.c
73
cset = task_css_set(current);
kernel/cgroup/rdma.c
79
return css_rdmacg(task_get_css(current, rdma_cgrp_id));
kernel/compat.c
47
old_set = current->blocked.sig[0];
kernel/compat.c
54
new_blocked = current->blocked;
kernel/context_tracking.c
120
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
kernel/context_tracking.c
121
rcu_preempt_deferred_qs(current);
kernel/context_tracking.c
165
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
kernel/context_tracking.c
45
WRITE_ONCE(current->rcu_tasks_idle_cpu, smp_processor_id());
kernel/context_tracking.c
453
WARN_ON_ONCE(!current->mm);
kernel/context_tracking.c
470
vtime_user_enter(current);
kernel/context_tracking.c
53
WRITE_ONCE(current->rcu_tasks_idle_cpu, -1);
kernel/context_tracking.c
605
vtime_user_exit(current);
kernel/crash_core.c
467
prstatus.common.pr_pid = current->pid;
kernel/cred.c
181
struct task_struct *task = current;
kernel/cred.c
26
current->comm, current->pid, ##__VA_ARGS__)
kernel/cred.c
32
current->comm, current->pid, ##__VA_ARGS__); \
kernel/cred.c
370
struct task_struct *task = current;
kernel/cred.c
77
BUG_ON(cred == current->cred);
kernel/cred.c
78
BUG_ON(cred == current->real_cred);
kernel/debug/debug_core.c
603
kgdb_info[cpu].task = current;
kernel/debug/debug_core.c
729
kgdb_contthread = current;
kernel/debug/gdbstub.c
1058
if (kgdb_contthread && kgdb_contthread != current) {
kernel/debug/gdbstub.c
548
if (kgdb_usethread && kgdb_usethread != current) {
kernel/debug/gdbstub.c
621
!(!kgdb_usethread || kgdb_usethread == current) ||
kernel/debug/gdbstub.c
745
ks->threadid = shadow_pid(current->pid);
kernel/debug/gdbstub.c
976
int_to_threadref(thref, shadow_pid(current->pid));
kernel/delayacct.c
131
current->delays->blkio_start = local_clock();
kernel/delayacct.c
218
current->delays->freepages_start = local_clock();
kernel/delayacct.c
223
delayacct_end(&current->delays->lock,
kernel/delayacct.c
224
&current->delays->freepages_start,
kernel/delayacct.c
225
&current->delays->freepages_delay,
kernel/delayacct.c
226
&current->delays->freepages_count,
kernel/delayacct.c
227
&current->delays->freepages_delay_max,
kernel/delayacct.c
228
&current->delays->freepages_delay_min,
kernel/delayacct.c
229
&current->delays->freepages_delay_max_ts);
kernel/delayacct.c
234
*in_thrashing = !!current->in_thrashing;
kernel/delayacct.c
238
current->in_thrashing = 1;
kernel/delayacct.c
239
current->delays->thrashing_start = local_clock();
kernel/delayacct.c
247
current->in_thrashing = 0;
kernel/delayacct.c
248
delayacct_end(&current->delays->lock,
kernel/delayacct.c
249
&current->delays->thrashing_start,
kernel/delayacct.c
250
&current->delays->thrashing_delay,
kernel/delayacct.c
251
&current->delays->thrashing_count,
kernel/delayacct.c
252
&current->delays->thrashing_delay_max,
kernel/delayacct.c
253
&current->delays->thrashing_delay_min,
kernel/delayacct.c
254
&current->delays->thrashing_delay_max_ts);
kernel/delayacct.c
259
current->delays->swapin_start = local_clock();
kernel/delayacct.c
264
delayacct_end(&current->delays->lock,
kernel/delayacct.c
265
&current->delays->swapin_start,
kernel/delayacct.c
266
&current->delays->swapin_delay,
kernel/delayacct.c
267
&current->delays->swapin_count,
kernel/delayacct.c
268
&current->delays->swapin_delay_max,
kernel/delayacct.c
269
&current->delays->swapin_delay_min,
kernel/delayacct.c
270
&current->delays->swapin_delay_max_ts);
kernel/delayacct.c
275
current->delays->compact_start = local_clock();
kernel/delayacct.c
280
delayacct_end(&current->delays->lock,
kernel/delayacct.c
281
&current->delays->compact_start,
kernel/delayacct.c
282
&current->delays->compact_delay,
kernel/delayacct.c
283
&current->delays->compact_count,
kernel/delayacct.c
284
&current->delays->compact_delay_max,
kernel/delayacct.c
285
&current->delays->compact_delay_min,
kernel/delayacct.c
286
&current->delays->compact_delay_max_ts);
kernel/delayacct.c
291
current->delays->wpcopy_start = local_clock();
kernel/delayacct.c
296
delayacct_end(&current->delays->lock,
kernel/delayacct.c
297
&current->delays->wpcopy_start,
kernel/delayacct.c
298
&current->delays->wpcopy_delay,
kernel/delayacct.c
299
&current->delays->wpcopy_count,
kernel/delayacct.c
300
&current->delays->wpcopy_delay_max,
kernel/delayacct.c
301
&current->delays->wpcopy_delay_min,
kernel/delayacct.c
302
&current->delays->wpcopy_delay_max_ts);
kernel/dma/debug.c
1082
struct vm_struct *stack_vm_area = task_stack_vm_area(current);
kernel/dma/debug.c
1100
addr = (u8 *)current->stack + i * PAGE_SIZE +
kernel/entry/common.c
141
(is_idle_task(current) || arch_in_rcu_eqs())) {
kernel/entry/common.c
61
klp_update_patch_state(current);
kernel/entry/syscall-common.c
17
return syscall_get_nr(current, regs);
kernel/entry/syscall_user_dispatch.c
134
return task_set_syscall_user_dispatch(current, mode, offset, len, selector);
kernel/entry/syscall_user_dispatch.c
27
info.si_call_addr = (void __user *)KSTK_EIP(current);
kernel/entry/syscall_user_dispatch.c
29
info.si_arch = syscall_get_arch(current);
kernel/entry/syscall_user_dispatch.c
30
info.si_syscall = syscall_get_nr(current, regs);
kernel/entry/syscall_user_dispatch.c
37
struct syscall_user_dispatch *sd = &current->syscall_dispatch;
kernel/entry/syscall_user_dispatch.c
66
syscall_rollback(current, regs);
kernel/events/callchain.c
185
struct uprobe_task *utask = current->utask;
kernel/events/callchain.c
249
if (!is_user_task(current))
kernel/events/callchain.c
251
regs = task_pt_regs(current);
kernel/events/core.c
10484
rec.pid = perf_event_pid(event, current);
kernel/events/core.c
10485
rec.tid = perf_event_tid(event, current);
kernel/events/core.c
10734
!task_work_add(current, &event->pending_task, notify_mode)) {
kernel/events/core.c
11015
return get_recursion_context(current->perf_recursion);
kernel/events/core.c
11021
put_recursion_context(current->perf_recursion, rctx);
kernel/events/core.c
11423
if (task && task != current) {
kernel/events/core.c
1182
cpuctx->cgrp = perf_cgroup_from_task(current, ctx);
kernel/events/core.c
12194
if (!(event->attr.exclude_idle && is_idle_task(current)))
kernel/events/core.c
13332
event->ns = get_pid_ns(task_active_pid_ns(current));
kernel/events/core.c
14161
event->owner = current;
kernel/events/core.c
14173
mutex_lock(&current->perf_event_mutex);
kernel/events/core.c
14174
list_add_tail(&event->owner_entry, &current->perf_event_list);
kernel/events/core.c
14175
mutex_unlock(&current->perf_event_mutex);
kernel/events/core.c
14603
WARN_ON_ONCE(task != current);
kernel/events/core.c
14914
struct task_struct *parent = current;
kernel/events/core.c
297
if (ctx->task != current) {
kernel/events/core.c
3065
reprogram = (ctx->task == current);
kernel/events/core.c
3090
struct perf_cgroup *cgrp = perf_cgroup_from_task(current, ctx);
kernel/events/core.c
412
if (WARN_ON_ONCE(task != current))
kernel/events/core.c
4661
ctx = rcu_dereference(current->perf_event_ctxp);
kernel/events/core.c
4696
if (WARN_ON_ONCE(current->perf_event_ctxp != ctx))
kernel/events/core.c
4746
if (WARN_ON_ONCE(ctx->task != current))
kernel/events/core.c
4907
event->hw.target != current) {
kernel/events/core.c
5077
task = current;
kernel/events/core.c
6756
mutex_lock(&current->perf_event_mutex);
kernel/events/core.c
6757
list_for_each_entry(event, &current->perf_event_list, owner_entry) {
kernel/events/core.c
6762
mutex_unlock(&current->perf_event_mutex);
kernel/events/core.c
6772
mutex_lock(&current->perf_event_mutex);
kernel/events/core.c
6773
list_for_each_entry(event, &current->perf_event_list, owner_entry) {
kernel/events/core.c
6778
mutex_unlock(&current->perf_event_mutex);
kernel/events/core.c
7543
if (current->flags & PF_EXITING)
kernel/events/core.c
7551
if (WARN_ON_ONCE(event->ctx->task != current))
kernel/events/core.c
7753
regs_user->abi = perf_reg_abi(current);
kernel/events/core.c
7755
} else if (is_user_task(current)) {
kernel/events/core.c
7767
regs_intr->abi = perf_reg_abi(current);
kernel/events/core.c
7799
if (!current->mm)
kernel/events/core.c
8003
data->tid_entry.pid = perf_event_pid(event, current);
kernel/events/core.c
8004
data->tid_entry.tid = perf_event_tid(event, current);
kernel/events/core.c
8393
if (is_user_task(current)) {
kernel/events/core.c
8483
mm = current->mm;
kernel/events/core.c
8508
is_user_task(current);
kernel/events/core.c
8510
bool crosstask = event->ctx->task && event->ctx->task != current;
kernel/events/core.c
8517
if (!current->mm)
kernel/events/core.c
8680
cgrp = task_css_check(current, perf_event_cgrp_id, 1)->cgroup;
kernel/events/core.c
8953
ctx = rcu_dereference(current->perf_event_ctxp);
kernel/events/core.c
8998
ctx = perf_pin_task_context(current);
kernel/events/core.c
91
if (p != current)
kernel/events/core.c
9149
task_event->event_id.ppid = perf_event_pid(event, current);
kernel/events/core.c
9150
task_event->event_id.ptid = perf_event_tid(event, current);
kernel/events/core.c
9248
perf_event_alloc_task_data(task, current);
kernel/events/core.c
9647
mmap_event->event_id.pid = perf_event_pid(event, current);
kernel/events/core.c
9648
mmap_event->event_id.tid = perf_event_tid(event, current);
kernel/events/core.c
9883
ctx = rcu_dereference(current->perf_event_ctxp);
kernel/events/hw_breakpoint.c
808
if (irqs_disabled() && bp->ctx && bp->ctx->task == current)
kernel/events/hw_breakpoint_test.c
127
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
150
fill_bp_slots(test, &idx, -1, current, 0);
kernel/events/hw_breakpoint_test.c
151
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
152
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
164
fill_bp_slots(test, &idx, -1, current, 0);
kernel/events/hw_breakpoint_test.c
167
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
169
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
181
fill_bp_slots(test, &idx, get_test_cpu(0), current, 0);
kernel/events/hw_breakpoint_test.c
182
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
183
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
199
fill_one_bp_slot(test, &idx, get_test_cpu(0), current);
kernel/events/hw_breakpoint_test.c
200
fill_bp_slots(test, &idx, -1, current, 1);
kernel/events/hw_breakpoint_test.c
201
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
202
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
217
fill_bp_slots(test, &idx, get_test_cpu(0), current, 0);
kernel/events/hw_breakpoint_test.c
220
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
222
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
233
fill_bp_slots(test, &idx, get_test_cpu(0), current, 0);
kernel/events/hw_breakpoint_test.c
236
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
238
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
252
fill_bp_slots(test, &idx, -1, current, 2);
kernel/events/hw_breakpoint_test.c
255
fill_one_bp_slot(test, &idx, get_test_cpu(0), current);
kernel/events/hw_breakpoint_test.c
256
fill_one_bp_slot(test, &idx, -1, current);
kernel/events/hw_breakpoint_test.c
258
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
259
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/hw_breakpoint_test.c
270
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
273
fill_one_bp_slot(test, &idx, -1, current);
kernel/events/hw_breakpoint_test.c
275
TEST_EXPECT_NOSPC(register_test_bp(-1, current, idx));
kernel/events/hw_breakpoint_test.c
276
TEST_EXPECT_NOSPC(register_test_bp(get_test_cpu(0), current, idx));
kernel/events/uprobes.c
1348
uprobe_warn(current, "unregister, leaking uprobe");
kernel/events/uprobes.c
1626
if (!fatal_signal_pending(current) &&
kernel/events/uprobes.c
1753
struct mm_struct *mm = current->mm;
kernel/events/uprobes.c
1798
struct mm_struct *mm = current->mm;
kernel/events/uprobes.c
1894
struct xol_area *area = current->mm->uprobes_state.xol_area;
kernel/events/uprobes.c
1938
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2092
if (!current->utask)
kernel/events/uprobes.c
2093
current->utask = alloc_utask();
kernel/events/uprobes.c
2094
return current->utask;
kernel/events/uprobes.c
2176
if (current->flags & PF_EXITING)
kernel/events/uprobes.c
2179
if (!__create_xol_area(current->utask->dup_xol_addr) &&
kernel/events/uprobes.c
2180
!fatal_signal_pending(current))
kernel/events/uprobes.c
2181
uprobe_warn(current, "dup xol area");
kernel/events/uprobes.c
2189
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2190
struct mm_struct *mm = current->mm;
kernel/events/uprobes.c
2229
area = READ_ONCE(current->mm->uprobes_state.xol_area); /* ^^^ */
kernel/events/uprobes.c
2255
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2266
current->pid, current->tgid);
kernel/events/uprobes.c
2290
uprobe_warn(current, "handle tail call");
kernel/events/uprobes.c
2321
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2358
struct task_struct *t = current;
kernel/events/uprobes.c
2429
struct mm_struct *mm = current->mm;
kernel/events/uprobes.c
2471
struct mm_struct *mm = current->mm;
kernel/events/uprobes.c
2554
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2590
if (!filter_chain(uprobe, current->mm)) {
kernel/events/uprobes.c
2592
unapply_uprobe(uprobe, current->mm);
kernel/events/uprobes.c
2643
utask = current->utask;
kernel/events/uprobes.c
2689
uprobe_warn(current, "handle uretprobe, sending SIGILL.");
kernel/events/uprobes.c
2838
uprobe_warn(current, "execute the probed insn, sending SIGILL.");
kernel/events/uprobes.c
2860
utask = current->utask;
kernel/events/uprobes.c
2873
if (!current->mm)
kernel/events/uprobes.c
2876
if (!mm_flags_test(MMF_HAS_UPROBES, current->mm) &&
kernel/events/uprobes.c
2877
(!current->utask || !current->utask->return_instances))
kernel/events/uprobes.c
2890
struct uprobe_task *utask = current->utask;
kernel/events/uprobes.c
2892
if (!current->mm || !utask || !utask->active_uprobe)
kernel/exec_domain.c
40
unsigned int old = current->personality;
kernel/exit.c
1034
struct task_struct *tsk = current;
kernel/exit.c
1044
current->comm, task_pid_nr(current));
kernel/exit.c
1049
current->comm, task_pid_nr(current),
kernel/exit.c
1095
struct signal_struct *sig = current->signal;
kernel/exit.c
1102
struct sighand_struct *const sighand = current->sighand;
kernel/exit.c
1113
zap_other_threads(current);
kernel/exit.c
1212
struct signal_struct *psig = current->signal;
kernel/exit.c
1629
return current == parent || (!(wo->wo_flags & __WNOTHREAD) &&
kernel/exit.c
1630
same_thread_group(current, parent));
kernel/exit.c
1685
struct task_struct *tsk = current;
kernel/exit.c
1698
} while_each_thread(current, tsk);
kernel/exit.c
1717
wo->child_wait.private = current;
kernel/exit.c
1718
add_wait_queue(&current->signal->wait_chldexit, &wo->child_wait);
kernel/exit.c
1725
if (signal_pending(current))
kernel/exit.c
1731
remove_wait_queue(&current->signal->wait_chldexit, &wo->child_wait);
kernel/exit.c
1768
pid = get_task_pid(current, PIDTYPE_PGID);
kernel/exit.c
1869
pid = get_task_pid(current, PIDTYPE_PGID);
kernel/exit.c
375
retval = will_become_orphaned_pgrp(task_pgrp(current), NULL);
kernel/exit.c
489
struct task_struct *g, *p = current;
kernel/exit.c
552
struct mm_struct *mm = current->mm;
kernel/exit.c
554
exit_mm_release(current, mm);
kernel/exit.c
559
BUG_ON(mm != current->active_mm);
kernel/exit.c
561
task_lock(current);
kernel/exit.c
574
current->mm = NULL;
kernel/exit.c
576
enter_lazy_tlb(mm, current);
kernel/exit.c
578
task_unlock(current);
kernel/exit.c
849
free = stack_not_used(current);
kernel/exit.c
858
current->comm, task_pid_nr(current), free);
kernel/exit.c
898
struct task_struct *tsk = current;
kernel/exit.c
992
if (unlikely(current->pi_state_cache))
kernel/exit.c
993
kfree(current->pi_state_cache);
kernel/fork.c
1103
if (current->mm) {
kernel/fork.c
1104
unsigned long flags = __mm_flags_get_word(current->mm);
kernel/fork.c
1107
mm->def_flags = current->mm->def_flags & VM_INIT_DEF_MASK;
kernel/fork.c
1163
return mm_init(mm, current, current_user_ns());
kernel/fork.c
1384
if (mm == current->mm)
kernel/fork.c
1575
oldmm = current->mm;
kernel/fork.c
1583
mm = dup_mm(tsk, current->mm);
kernel/fork.c
1595
struct fs_struct *fs = current->fs;
kernel/fork.c
1622
oldf = current->files;
kernel/fork.c
1649
refcount_inc(&current->sighand->count);
kernel/fork.c
1658
spin_lock_irq(&current->sighand->siglock);
kernel/fork.c
1659
memcpy(sig->action, current->sighand->action, sizeof(sig->action));
kernel/fork.c
1660
spin_unlock_irq(&current->sighand->siglock);
kernel/fork.c
1727
task_lock(current->group_leader);
kernel/fork.c
1728
memcpy(sig->rlim, current->signal->rlim, sizeof sig->rlim);
kernel/fork.c
1729
task_unlock(current->group_leader);
kernel/fork.c
1740
sig->oom_score_adj = current->signal->oom_score_adj;
kernel/fork.c
1741
sig->oom_score_adj_min = current->signal->oom_score_adj_min;
kernel/fork.c
1758
assert_spin_locked(&current->sighand->siglock);
kernel/fork.c
1761
get_seccomp_filter(current);
kernel/fork.c
1762
p->seccomp = current->seccomp;
kernel/fork.c
1769
if (task_no_new_privs(current))
kernel/fork.c
1784
current->clear_child_tid = tidptr;
kernel/fork.c
1786
return task_pid_vnr(current);
kernel/fork.c
1935
tsk->signal->oom_score_adj = current->signal->oom_score_adj;
kernel/fork.c
1936
tsk->signal->oom_score_adj_min = current->signal->oom_score_adj_min;
kernel/fork.c
1975
struct nsproxy *nsp = current->nsproxy;
kernel/fork.c
2009
current->signal->flags & SIGNAL_UNKILLABLE)
kernel/fork.c
2018
(task_active_pid_ns(current) != nsp->pid_ns_for_children))
kernel/fork.c
2040
spin_lock_irq(&current->sighand->siglock);
kernel/fork.c
2042
hlist_add_head(&delayed.node, &current->signal->multiprocess);
kernel/fork.c
2044
spin_unlock_irq(&current->sighand->siglock);
kernel/fork.c
2046
if (task_sigpending(current))
kernel/fork.c
2050
p = dup_task_struct(current, node);
kernel/fork.c
2094
current->flags &= ~PF_NPROC_EXCEEDED;
kernel/fork.c
2136
p->default_timer_slack_ns = current->timer_slack_ns;
kernel/fork.c
2294
p->group_leader = current->group_leader;
kernel/fork.c
2295
p->tgid = current->tgid;
kernel/fork.c
2373
p->real_parent = current->real_parent;
kernel/fork.c
2374
p->parent_exec_id = current->parent_exec_id;
kernel/fork.c
2378
p->exit_signal = current->group_leader->exit_signal;
kernel/fork.c
2380
p->real_parent = current;
kernel/fork.c
2381
p->parent_exec_id = current->self_exec_id;
kernel/fork.c
2389
spin_lock(&current->sighand->siglock);
kernel/fork.c
2402
if (fatal_signal_pending(current)) {
kernel/fork.c
2422
init_task_pid(p, PIDTYPE_PGID, task_pgrp(current));
kernel/fork.c
2423
init_task_pid(p, PIDTYPE_SID, task_session(current));
kernel/fork.c
2430
p->signal->tty = tty_kref_get(current->signal->tty);
kernel/fork.c
2445
current->signal->nr_threads++;
kernel/fork.c
2446
current->signal->quick_threads++;
kernel/fork.c
2447
atomic_inc(&current->signal->live);
kernel/fork.c
2448
refcount_inc(&current->signal->sigcnt);
kernel/fork.c
2458
spin_unlock(&current->sighand->siglock);
kernel/fork.c
2480
spin_unlock(&current->sighand->siglock);
kernel/fork.c
2541
spin_lock_irq(&current->sighand->siglock);
kernel/fork.c
2543
spin_unlock_irq(&current->sighand->siglock);
kernel/fork.c
2649
if (likely(!ptrace_event_enabled(current, trace)))
kernel/fork.c
2663
trace_sched_process_fork(current, p);
kernel/fork.c
3061
if (!thread_group_empty(current))
kernel/fork.c
3065
if (refcount_read(&current->sighand->count) > 1)
kernel/fork.c
3081
struct fs_struct *fs = current->fs;
kernel/fork.c
3102
struct files_struct *fd = current->files;
kernel/fork.c
3189
exit_sem(current);
kernel/fork.c
3193
exit_shm(current);
kernel/fork.c
3194
shm_init_task(current);
kernel/fork.c
3198
switch_task_namespaces(current, new_nsproxy);
kernel/fork.c
3200
task_lock(current);
kernel/fork.c
3203
fs = current->fs;
kernel/fork.c
3205
current->fs = new_fs;
kernel/fork.c
3214
swap(current->files, new_fd);
kernel/fork.c
3216
task_unlock(current);
kernel/fork.c
3225
perf_event_namespaces(current);
kernel/fork.c
3255
struct task_struct *task = current;
kernel/fork.c
635
current->comm,
kernel/fork.c
636
task_pid_nr(current));
kernel/fork.c
656
WARN_ON_ONCE(current->active_mm == mm);
kernel/fork.c
663
if (current->active_mm == mm) {
kernel/fork.c
664
WARN_ON_ONCE(current->mm);
kernel/fork.c
665
current->active_mm = &init_mm;
kernel/fork.c
666
switch_mm(mm, &init_mm, current);
kernel/fork.c
721
WARN_ON_ONCE(mm == current->mm);
kernel/fork.c
726
WARN_ON_ONCE(mm == current->active_mm);
kernel/fork.c
780
WARN_ON(tsk == current);
kernel/freezer.c
121
if (p != current && task_curr(p))
kernel/freezer.c
242
current->flags &= ~PF_NOFREEZE;
kernel/freezer.c
68
pr_debug("%s entered refrigerator\n", current->comm);
kernel/freezer.c
75
raw_spin_lock_irq(&current->pi_lock);
kernel/freezer.c
76
WRITE_ONCE(current->__state, TASK_FROZEN);
kernel/freezer.c
78
current->saved_state = TASK_RUNNING;
kernel/freezer.c
79
raw_spin_unlock_irq(&current->pi_lock);
kernel/freezer.c
82
freeze = freezing(current) && !(check_kthr_stop && kthread_should_stop());
kernel/freezer.c
93
pr_debug("%s left refrigerator\n", current->comm);
kernel/futex/core.c
1339
WARN_ON(curr != current);
kernel/futex/core.c
1781
struct mm_struct *mm = current->mm;
kernel/futex/core.c
1894
if (!current->mm)
kernel/futex/core.c
1899
get_nr_threads(current),
kernel/futex/core.c
1902
fph = rcu_dereference(current->mm->futex_phash);
kernel/futex/core.c
1929
fph = rcu_dereference(current->mm->futex_phash);
kernel/futex/core.c
276
struct mm_struct *mm = current->mm;
kernel/futex/core.c
552
struct mm_struct *mm = current->mm;
kernel/futex/core.c
786
struct mm_struct *mm = current->mm;
kernel/futex/core.c
904
prio = min(current->normal_prio, MAX_RT_PRIO);
kernel/futex/pi.c
1035
ret = __rt_mutex_start_proxy_lock(&q.pi_state->pi_mutex, &rt_waiter, current, &wake_q);
kernel/futex/pi.c
109
current->pi_state_cache = pi_state;
kernel/futex/pi.c
1135
u32 curval, uval, vpid = task_pid_vnr(current);
kernel/futex/pi.c
1178
if (pi_state->owner != current)
kernel/futex/pi.c
17
if (likely(current->pi_state_cache))
kernel/futex/pi.c
31
current->pi_state_cache = pi_state;
kernel/futex/pi.c
38
struct futex_pi_state *pi_state = current->pi_state_cache;
kernel/futex/pi.c
41
current->pi_state_cache = NULL;
kernel/futex/pi.c
706
if (oldowner != current) {
kernel/futex/pi.c
737
WARN_ON_ONCE(argowner != current);
kernel/futex/pi.c
738
if (oldowner == current) {
kernel/futex/pi.c
775
return argowner == current;
kernel/futex/pi.c
816
return argowner == current;
kernel/futex/pi.c
883
if (q->pi_state->owner != current)
kernel/futex/pi.c
884
return fixup_pi_state_owner(uaddr, q, current);
kernel/futex/pi.c
896
if (q->pi_state->owner == current)
kernel/futex/pi.c
903
if (WARN_ON_ONCE(rt_mutex_owner(&q->pi_state->pi_mutex) == current))
kernel/futex/pi.c
904
return fixup_pi_state_owner(uaddr, q, current);
kernel/futex/pi.c
947
ret = futex_lock_pi_atomic(uaddr, hb, &q.key, &q.pi_state, current,
kernel/futex/pi.c
988
__futex_queue(&q, hb, current);
kernel/futex/pi.c
99
if (current->pi_state_cache) {
kernel/futex/requeue.c
734
else if (signal_pending(current))
kernel/futex/requeue.c
800
current->timer_slack_ns);
kernel/futex/requeue.c
820
ret = futex_wait_setup(uaddr, val, flags, &q, &key2, current);
kernel/futex/requeue.c
840
if (q.pi_state && (q.pi_state->owner != current)) {
kernel/futex/syscalls.c
37
current->robust_list = head;
kernel/futex/syscalls.c
485
current->compat_robust_list = head;
kernel/futex/syscalls.c
53
struct task_struct *p = current;
kernel/futex/waitwake.c
217
current->comm, oparg);
kernel/futex/waitwake.c
460
futex_queue(q, hb, current);
kernel/futex/waitwake.c
564
else if (signal_pending(current))
kernel/futex/waitwake.c
658
if (task == current)
kernel/futex/waitwake.c
682
ret = futex_wait_setup(uaddr, val, flags, &q, NULL, current);
kernel/futex/waitwake.c
700
if (!signal_pending(current))
kernel/futex/waitwake.c
713
current->timer_slack_ns);
kernel/futex/waitwake.c
725
restart = &current->restart_block;
kernel/hung_task.c
501
set_user_nice(current, 0);
kernel/irq/manage.c
1047
set_cpus_allowed_ptr(current, mask);
kernel/irq/manage.c
1180
struct task_struct *tsk = current;
kernel/irq/manage.c
1184
if (WARN_ON_ONCE(!(current->flags & PF_EXITING)))
kernel/irq/manage.c
1255
sched_set_fifo_secondary(current);
kernel/irq/manage.c
1257
sched_set_fifo(current);
kernel/irq/manage.c
1266
task_work_add(current, &on_exit_work, TWA_NONE);
kernel/irq/manage.c
1284
task_work_cancel_func(current, irq_thread_dtor);
kernel/irq_work.c
310
sched_set_fifo_low(current);
kernel/kcov.c
1019
struct task_struct *t = current;
kernel/kcov.c
1090
return current->kcov_handle;
kernel/kcov.c
1112
current->kcov_mode = KCOV_MODE_TRACE_PC;
kernel/kcov.c
1115
current->kcov_mode = 0;
kernel/kcov.c
216
t = current;
kernel/kcov.c
245
t = current;
kernel/kcov.c
390
t->kcov_handle = current->kcov_handle;
kernel/kcov.c
615
t = current;
kernel/kcov.c
632
if (unused != 0 || current->kcov != kcov)
kernel/kcov.c
634
t = current;
kernel/kcov.c
643
t = current;
kernel/kcov.c
863
struct task_struct *t = current;
kernel/kcsan/core.c
206
return in_task() ? &current->kcsan_ctx : raw_cpu_ptr(&kcsan_cpu_ctx);
kernel/kcsan/core.c
375
return current->kcsan_stack_depth;
kernel/kcsan/core.c
385
current->kcsan_stack_depth += val;
kernel/kcsan/core.c
506
kcsan_save_irqtrace(current);
kernel/kcsan/core.c
508
kcsan_restore_irqtrace(current);
kernel/kcsan/core.c
587
kcsan_save_irqtrace(current);
kernel/kcsan/core.c
704
kcsan_restore_irqtrace(current);
kernel/kcsan/debugfs.c
59
const struct kcsan_ctx ctx_save = current->kcsan_ctx;
kernel/kcsan/debugfs.c
64
memset(&current->kcsan_ctx, 0, sizeof(current->kcsan_ctx));
kernel/kcsan/debugfs.c
86
current->kcsan_ctx = ctx_save;
kernel/kcsan/kcsan_test.c
531
struct kcsan_scoped_access *reorder_access = &current->kcsan_ctx.reorder_access;
kernel/kcsan/permissive.h
31
return ptr == &current->flags;
kernel/kcsan/report.c
473
print_verbose_info(current);
kernel/kcsan/report.c
525
const bool is_running = task_is_running(current);
kernel/kcsan/report.c
536
other_info->task = current;
kernel/kcsan/report.c
569
other_info->task == current);
kernel/kcsan/report.c
649
.task_pid = in_task() ? task_pid_nr(current) : -1,
kernel/kcsan/selftest.c
113
struct kcsan_scoped_access *reorder_access = &current->kcsan_ctx.reorder_access;
kernel/kexec_core.c
283
if (fatal_signal_pending(current))
kernel/kprobes.c
2089
correct_ret_addr = __kretprobe_find_ret_addr(current, &node);
kernel/kprobes.c
2103
first = current->kretprobe_instances.first;
kernel/kprobes.c
2128
first = current->kretprobe_instances.first;
kernel/kprobes.c
2129
current->kretprobe_instances.first = node->next;
kernel/kprobes.c
2167
__llist_add(&ri->llist, &current->kretprobe_instances);
kernel/kstack_erase.c
119
current->lowest_stack = task_stack_high;
kernel/kstack_erase.c
172
if (sp < current->lowest_stack &&
kernel/kstack_erase.c
173
sp >= stackleak_task_low_bound(current)) {
kernel/kstack_erase.c
174
current->lowest_stack = sp;
kernel/kstack_erase.c
87
const unsigned long task_stack_low = stackleak_task_low_bound(current);
kernel/kstack_erase.c
88
const unsigned long task_stack_high = stackleak_task_high_bound(current);
kernel/kstack_erase.c
92
current->lowest_stack);
kernel/kstack_erase.c
95
current->prev_lowest_stack = erase_low;
kernel/kthread.c
1031
} else if (!freezing(current)) {
kernel/kthread.c
117
kthread->node = tsk_fork_get_node(current);
kernel/kthread.c
150
return test_bit(KTHREAD_SHOULD_STOP, &to_kthread(current)->flags);
kernel/kthread.c
1618
struct task_struct *tsk = current;
kernel/kthread.c
1664
struct task_struct *tsk = current;
kernel/kthread.c
1706
if (!(current->flags & PF_KTHREAD))
kernel/kthread.c
1708
kthread = to_kthread(current);
kernel/kthread.c
172
return __kthread_should_park(current);
kernel/kthread.c
1732
if (current->flags & PF_KTHREAD) {
kernel/kthread.c
1733
kthread = to_kthread(current);
kernel/kthread.c
178
struct kthread *kthread = tsk_is_kthread(current);
kernel/kthread.c
201
if (unlikely(freezing(current)))
kernel/kthread.c
290
__kthread_parkme(to_kthread(current));
kernel/kthread.c
352
struct kthread *kthread = to_kthread(current);
kernel/kthread.c
355
if (WARN_ON_ONCE(kthread_is_per_cpu(current)))
kernel/kthread.c
374
set_cpus_allowed_ptr(current, affinity);
kernel/kthread.c
391
self = to_kthread(current);
kernel/kthread.c
409
sched_setscheduler_nocheck(current, SCHED_NORMAL, &param);
kernel/kthread.c
413
create->result = current;
kernel/kthread.c
429
if (!(current->flags & PF_NO_SETAFFINITY) && !self->preferred_affinity)
kernel/kthread.c
456
current->pref_node_fork = create->node;
kernel/kthread.c
714
if (k != current) {
kernel/kthread.c
790
struct task_struct *tsk = current;
kernel/kthread.c
797
current->flags |= PF_NOFREEZE;
kernel/kthread.c
994
WARN_ON(worker->task && worker->task != current);
kernel/kthread.c
995
worker->task = current;
kernel/livepatch/patch.c
96
patch_state = current->patch_state;
kernel/livepatch/transition.c
288
if (task_curr(task) && task != current)
kernel/livepatch/transition.c
326
if (task == current)
kernel/livepatch/transition.c
327
ret = klp_check_and_switch_task(current, &old_name);
kernel/livepatch/transition.c
369
if (likely(!klp_patch_pending(current)))
kernel/livepatch/transition.c
380
klp_try_switch_task(current);
kernel/livepatch/transition.c
689
if (test_tsk_thread_flag(current, TIF_PATCH_PENDING))
kernel/livepatch/transition.c
694
child->patch_state = current->patch_state;
kernel/locking/lockdep.c
124
if (current->lockdep_recursion)
kernel/locking/lockdep.c
147
__owner = current;
kernel/locking/lockdep.c
154
if (debug_locks && DEBUG_LOCKS_WARN_ON(__owner != current))
kernel/locking/lockdep.c
165
return DEBUG_LOCKS_WARN_ON(__owner != current);
kernel/locking/lockdep.c
1955
struct task_struct *curr = current;
kernel/locking/lockdep.c
2010
struct task_struct *curr = current;
kernel/locking/lockdep.c
2173
print_deadlock_bug(current, src, target);
kernel/locking/lockdep.c
3648
pr_warn("%s/%d: ", current->comm, task_pid_nr(current));
kernel/locking/lockdep.c
4337
struct task_struct *curr = current;
kernel/locking/lockdep.c
4407
current->hardirq_chain_key = current->curr_chain_key;
kernel/locking/lockdep.c
4417
struct irqtrace_events *trace = &current->irqtrace;
kernel/locking/lockdep.c
4467
DEBUG_LOCKS_WARN_ON(current->hardirq_chain_key !=
kernel/locking/lockdep.c
4468
current->curr_chain_key);
kernel/locking/lockdep.c
4506
struct irqtrace_events *trace = &current->irqtrace;
kernel/locking/lockdep.c
4526
struct irqtrace_events *trace = &current->irqtrace;
kernel/locking/lockdep.c
4538
if (current->softirqs_enabled) {
kernel/locking/lockdep.c
4547
current->softirqs_enabled = 1;
kernel/locking/lockdep.c
4557
mark_held_locks(current, LOCK_ENABLED_SOFTIRQ);
kernel/locking/lockdep.c
4575
if (current->softirqs_enabled) {
kernel/locking/lockdep.c
4576
struct irqtrace_events *trace = &current->irqtrace;
kernel/locking/lockdep.c
4581
current->softirqs_enabled = 0;
kernel/locking/lockdep.c
5082
struct task_struct *curr = current;
kernel/locking/lockdep.c
5258
lockdep_print_held_locks(current);
kernel/locking/lockdep.c
5412
struct task_struct *curr = current;
kernel/locking/lockdep.c
5459
struct task_struct *curr = current;
kernel/locking/lockdep.c
5513
struct task_struct *curr = current;
kernel/locking/lockdep.c
5595
struct task_struct *curr = current;
kernel/locking/lockdep.c
5615
struct task_struct *curr = current;
kernel/locking/lockdep.c
5642
struct task_struct *curr = current;
kernel/locking/lockdep.c
5662
struct task_struct *curr = current;
kernel/locking/lockdep.c
5718
DEBUG_LOCKS_WARN_ON(current->softirqs_enabled);
kernel/locking/lockdep.c
5721
DEBUG_LOCKS_WARN_ON(!current->softirqs_enabled);
kernel/locking/lockdep.c
5727
print_irqtrace_events(current);
kernel/locking/lockdep.c
5746
check_chain_key(current);
kernel/locking/lockdep.c
5763
check_chain_key(current);
kernel/locking/lockdep.c
5792
print_usage_bug(current, hlock, LOCK_USED, LOCK_USAGE_STATES);
kernel/locking/lockdep.c
5890
check_chain_key(current);
kernel/locking/lockdep.c
5918
check_chain_key(current);
kernel/locking/lockdep.c
6038
struct task_struct *curr = current;
kernel/locking/lockdep.c
6082
struct task_struct *curr = current;
kernel/locking/lockdep.c
6178
lockdep_init_task(current);
kernel/locking/lockdep.c
6179
memset(current->held_locks, 0, MAX_LOCK_DEPTH*sizeof(struct held_lock));
kernel/locking/lockdep.c
6303
return current == lockdep_selftest_task_struct;
kernel/locking/lockdep.c
6712
struct task_struct *curr = current;
kernel/locking/lockdep.c
6747
current->comm, task_pid_nr(current));
kernel/locking/lockdep.c
6750
lockdep_print_held_locks(current);
kernel/locking/lockdep.c
6759
if (unlikely(current->lockdep_depth > 0))
kernel/locking/lockdep.c
6807
struct task_struct *curr = current;
kernel/locking/lockdep.c
6833
struct task_struct *curr = current;
kernel/locking/lockdep.c
799
if (p != current && task_is_running(p))
kernel/locking/locktorture.c
193
if (!rt_task(current)) {
kernel/locking/locktorture.c
201
sched_set_fifo(current);
kernel/locking/locktorture.c
215
sched_set_normal(current, 0);
kernel/locking/locktorture.c
907
if (!rt_task(current))
kernel/locking/locktorture.c
908
set_user_nice(current, MAX_NICE);
kernel/locking/locktorture.c
975
set_user_nice(current, MAX_NICE);
kernel/locking/mutex.c
154
unsigned long curr = (unsigned long)current;
kernel/locking/mutex.c
167
unsigned long curr = (unsigned long)current;
kernel/locking/mutex.c
211
debug_mutex_add_waiter(lock, waiter, current);
kernel/locking/mutex.c
225
debug_mutex_remove_waiter(lock, waiter, current);
kernel/locking/mutex.c
242
MUTEX_WARN_ON(__owner_task(owner) != current);
kernel/locking/mutex.c
640
waiter.task = current;
kernel/locking/mutex.c
659
__set_task_blocked_on(current, lock);
kernel/locking/mutex.c
679
if (signal_pending_state(state, current)) {
kernel/locking/mutex.c
701
set_task_blocked_on(current, lock);
kernel/locking/mutex.c
718
clear_task_blocked_on(current, lock);
kernel/locking/mutex.c
721
set_task_blocked_on(current, lock);
kernel/locking/mutex.c
729
__clear_task_blocked_on(current, lock);
kernel/locking/mutex.c
759
__clear_task_blocked_on(current, lock);
kernel/locking/mutex.c
763
WARN_ON(__get_task_blocked_on(current));
kernel/locking/mutex.c
86
unsigned long owner, curr = (unsigned long)current;
kernel/locking/mutex.c
949
MUTEX_WARN_ON(__owner_task(owner) != current);
kernel/locking/rtmutex.c
1319
raw_spin_lock(&current->pi_lock);
kernel/locking/rtmutex.c
1330
rt_mutex_dequeue_pi(current, waiter);
kernel/locking/rtmutex.c
1331
rt_mutex_adjust_prio(lock, current);
kernel/locking/rtmutex.c
1355
raw_spin_unlock(&current->pi_lock);
kernel/locking/rtmutex.c
1360
int ret = try_to_take_rt_mutex(lock, current, NULL);
kernel/locking/rtmutex.c
1402
if (likely(rt_mutex_cmpxchg_acquire(lock, NULL, current)))
kernel/locking/rtmutex.c
1474
if (likely(rt_mutex_cmpxchg_release(lock, current, NULL)))
kernel/locking/rtmutex.c
1548
raw_spin_lock(&current->pi_lock);
kernel/locking/rtmutex.c
1550
current->pi_blocked_on = NULL;
kernel/locking/rtmutex.c
1551
raw_spin_unlock(&current->pi_lock);
kernel/locking/rtmutex.c
1587
next_lock, NULL, current);
kernel/locking/rtmutex.c
1619
if (try_to_take_rt_mutex(lock, current, waiter)) {
kernel/locking/rtmutex.c
1628
if (signal_pending_state(state, current)) {
kernel/locking/rtmutex.c
1706
if (try_to_take_rt_mutex(lock, current, NULL)) {
kernel/locking/rtmutex.c
1719
ret = task_blocks_on_rt_mutex(lock, waiter, current, ww_ctx, chwalk, wake_q);
kernel/locking/rtmutex.c
1810
lockdep_assert(!current->pi_blocked_on);
kernel/locking/rtmutex.c
1839
if (try_to_take_rt_mutex(lock, current, NULL)) {
kernel/locking/rtmutex.c
1851
task_blocks_on_rt_mutex(lock, &waiter, current, NULL, RT_MUTEX_MIN_CHAINWALK, wake_q);
kernel/locking/rtmutex.c
1855
if (try_to_take_rt_mutex(lock, current, &waiter)) {
kernel/locking/rtmutex.c
226
return rt_mutex_cmpxchg_acquire(lock, NULL, current);
kernel/locking/rtmutex_api.c
452
try_to_take_rt_mutex(lock, current, waiter);
kernel/locking/rtmutex_api.c
457
if (rt_mutex_owner(lock) != current) {
kernel/locking/rtmutex_common.h
182
DEBUG_LOCKS_WARN_ON(rt_mutex_owner(lock) != current);
kernel/locking/rwbase_rt.c
142
lockdep_assert(!current->pi_blocked_on);
kernel/locking/rwbase_rt.c
260
if (rwbase_signal_pending_state(state, current)) {
kernel/locking/rwsem.c
1032
waiter.task = current;
kernel/locking/rwsem.c
1078
if (signal_pending_state(state, current)) {
kernel/locking/rwsem.c
1126
waiter.task = current;
kernel/locking/rwsem.c
1166
if (signal_pending_state(state, current))
kernel/locking/rwsem.c
1380
DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
kernel/locking/rwsem.c
1405
DEBUG_RWSEMS_WARN_ON(rwsem_owner(sem) != current, sem);
kernel/locking/rwsem.c
1438
#define rwbase_signal_pending_state(state, current) \
kernel/locking/rwsem.c
1439
signal_pending_state(state, current)
kernel/locking/rwsem.c
144
atomic_long_set(&sem->owner, (long)current);
kernel/locking/rwsem.c
182
__rwsem_set_reader_owned(sem, current);
kernel/locking/rwsem.c
220
while ((val & ~RWSEM_OWNER_FLAGS_MASK) == (unsigned long)current) {
kernel/locking/rwsem.c
74
atomic_long_read(&(sem)->owner), (long)current, \
kernel/locking/rwsem.c
914
if (rt_or_dl_task(current) &&
kernel/locking/semaphore.c
258
waiter.task = current;
kernel/locking/semaphore.c
262
if (signal_pending_state(state, current))
kernel/locking/semaphore.c
48
WRITE_ONCE((sem)->last_holder, (unsigned long)current);
kernel/locking/semaphore.c
53
if (READ_ONCE((sem)->last_holder) == (unsigned long)current)
kernel/locking/spinlock_debug.c
102
SPIN_BUG_ON(lock->owner != current, lock, "wrong owner");
kernel/locking/spinlock_debug.c
152
msg, raw_smp_processor_id(), current->comm,
kernel/locking/spinlock_debug.c
153
task_pid_nr(current), lock);
kernel/locking/spinlock_debug.c
187
RWLOCK_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion");
kernel/locking/spinlock_debug.c
195
WRITE_ONCE(lock->owner, current);
kernel/locking/spinlock_debug.c
201
RWLOCK_BUG_ON(lock->owner != current, lock, "wrong owner");
kernel/locking/spinlock_debug.c
63
current->comm, task_pid_nr(current));
kernel/locking/spinlock_debug.c
87
SPIN_BUG_ON(READ_ONCE(lock->owner) == current, lock, "recursion");
kernel/locking/spinlock_debug.c
95
WRITE_ONCE(lock->owner, current);
kernel/locking/spinlock_rt.c
105
if (unlikely(!rt_mutex_cmpxchg_acquire(&lock->lock, NULL, current)))
kernel/locking/spinlock_rt.c
159
if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
kernel/locking/spinlock_rt.c
174
if (likely(rt_mutex_cmpxchg_acquire(rtm, current, NULL)))
kernel/locking/spinlock_rt.c
182
if (likely(rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
kernel/locking/spinlock_rt.c
188
#define rwbase_signal_pending_state(state, current) (0)
kernel/locking/spinlock_rt.c
40
lockdep_assert(!current->pi_blocked_on);
kernel/locking/spinlock_rt.c
42
if (unlikely(!rt_mutex_cmpxchg_acquire(rtm, NULL, current)))
kernel/locking/spinlock_rt.c
84
if (unlikely(!rt_mutex_cmpxchg_release(&lock->lock, current, NULL)))
kernel/locking/ww_mutex.h
340
if (owner != current) {
kernel/nscommon.c
315
return (task_active_pid_ns(current) == &init_pid_ns) &&
kernel/nsproxy.c
226
*new_nsp = create_new_namespaces(unshare_flags, current, user_ns,
kernel/nsproxy.c
227
new_fs ? new_fs : current->fs);
kernel/nsproxy.c
278
struct task_struct *tsk = current;
kernel/nsproxy.c
350
struct task_struct *me = current;
kernel/nsproxy.c
532
struct task_struct *me = current;
kernel/nsproxy.c
596
perf_event_namespaces(current);
kernel/panic.c
1042
raw_smp_processor_id(), current->comm, current->pid);
kernel/panic.c
1046
raw_smp_processor_id(), current->comm, current->pid);
kernel/panic.c
1067
print_irqtrace_events(current);
kernel/pid.c
369
return find_pid_ns(nr, task_active_pid_ns(current));
kernel/pid.c
483
return find_task_by_pid_ns(vnr, task_active_pid_ns(current));
kernel/pid.c
549
return pid_nr_ns(pid, task_active_pid_ns(current));
kernel/pid.c
560
ns = task_active_pid_ns(current);
kernel/pid.c
623
pid = get_task_pid(current, type);
kernel/pid.c
627
pid = get_task_pid(current, type);
kernel/pid.c
712
return &task_active_pid_ns(current)->set;
kernel/pid.c
717
return &task_active_pid_ns(current)->set == set;
kernel/pid_namespace.c
180
if (task_active_pid_ns(current) != old_ns)
kernel/pid_namespace.c
196
struct task_struct *task, *me = current;
kernel/pid_namespace.c
279
current->signal->group_exit_code = pid_ns->reboot;
kernel/pid_namespace.c
289
struct pid_namespace *pid_ns = task_active_pid_ns(current);
kernel/pid_namespace.c
404
struct pid_namespace *active = task_active_pid_ns(current);
kernel/pid_namespace.c
429
struct pid_namespace *active = task_active_pid_ns(current);
kernel/pid_sysctl.h
11
struct pid_namespace *ns = task_active_pid_ns(current);
kernel/power/main.c
69
unsigned int flags = current->flags;
kernel/power/main.c
70
current->flags |= PF_NOFREEZE;
kernel/power/main.c
79
current->flags &= ~PF_NOFREEZE;
kernel/power/process.c
101
if (p != current && freezing(p) && !frozen(p))
kernel/power/process.c
130
current->flags |= PF_SUSPEND_TASK;
kernel/power/process.c
182
struct task_struct *curr = current;
kernel/power/process.c
54
if (p == current || !freeze_task(p))
kernel/printk/printk.c
1897
console_owner = current;
kernel/printk/printk.c
2005
if (!waiter && owner && owner != current) {
kernel/printk/printk.c
2140
return in_task() ? task_pid_nr(current) :
kernel/printk/printk.c
2148
return !in_task() ? task_pid_nr(current) :
kernel/printk/printk.c
2250
get_task_comm(info->comm, current);
kernel/printk/printk.c
758
if (!___ratelimit(&user->rs, current->comm))
kernel/printk/printk_ringbuffer_kunit_test.c
193
wakeup.task = current;
kernel/printk/printk_ringbuffer_kunit_test.c
307
set_cpus_allowed_ptr(current, cpumask_of(reader_cpu));
kernel/ptrace.c
251
if (child->ptrace && child->parent == current) {
kernel/ptrace.c
298
if (same_thread_group(task, current))
kernel/ptrace.c
371
if (seccomp_mode(&current->seccomp) != SECCOMP_MODE_DISABLED ||
kernel/ptrace.c
372
current->ptrace & PT_SUSPEND_SECCOMP)
kernel/ptrace.c
439
if (same_thread_group(task, current))
kernel/ptrace.c
463
ptrace_link(task, current);
kernel/ptrace.c
493
if (!current->ptrace) {
kernel/ptrace.c
494
ret = security_ptrace_traceme(current->parent);
kernel/ptrace.c
500
if (!ret && !(current->real_parent->flags & PF_EXITING)) {
kernel/ptrace.c
501
current->ptrace = PT_PTRACED;
kernel/ptrace.c
502
ptrace_link(current, current->real_parent);
kernel/ptrace.c
55
(current != tsk->parent) ||
kernel/ptrace.c
582
__ptrace_detach(current, child);
kernel/ptrace.c
779
if (signal_pending(current))
kernel/rcu/rcuscale.c
458
set_cpus_allowed_ptr(current, cpumask_of(me % nr_cpu_ids));
kernel/rcu/rcuscale.c
459
set_user_nice(current, MAX_NICE);
kernel/rcu/rcuscale.c
543
set_cpus_allowed_ptr(current, cpumask_of(me % nr_cpu_ids));
kernel/rcu/rcuscale.c
544
current->flags |= PF_NO_SETAFFINITY;
kernel/rcu/rcuscale.c
545
sched_set_fifo_low(current);
kernel/rcu/rcuscale.c
609
sched_set_normal(current, 0);
kernel/rcu/rcuscale.c
737
set_cpus_allowed_ptr(current, cpumask_of(me % nr_cpu_ids));
kernel/rcu/rcuscale.c
738
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
1024
torture_sched_setaffinity(current->pid, cpumask_of(cpu), true);
kernel/rcu/rcutorture.c
1272
current->rt_priority, gp_state, end - *start);
kernel/rcu/rcutorture.c
1310
sched_set_fifo_low(current);
kernel/rcu/rcutorture.c
1325
sched_set_fifo_low(current);
kernel/rcu/rcutorture.c
1343
sched_set_fifo_low(current);
kernel/rcu/rcutorture.c
1380
sched_set_fifo_low(current);
kernel/rcu/rcutorture.c
1403
int oldnice = task_nice(current);
kernel/rcu/rcutorture.c
1420
sched_set_normal(current, oldnice);
kernel/rcu/rcutorture.c
1592
int oldnice = task_nice(current);
kernel/rcu/rcutorture.c
1656
current == writer_task);
kernel/rcu/rcutorture.c
1853
sched_set_normal(current, oldnice);
kernel/rcu/rcutorture.c
1885
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
2192
raw_spin_lock_irqsave(&current->pi_lock, flags);
kernel/rcu/rcutorture.c
2197
raw_spin_unlock_irqrestore(&current->pi_lock, flags);
kernel/rcu/rcutorture.c
2488
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
2491
tick_dep_set_task(current, TICK_DEP_BIT_RCU); // CPU bound, so need tick.
kernel/rcu/rcutorture.c
2512
tick_dep_clear_task(current, TICK_DEP_BIT_RCU);
kernel/rcu/rcutorture.c
2671
int oldnice = task_nice(current);
kernel/rcu/rcutorture.c
2704
sched_set_normal(current, oldnice);
kernel/rcu/rcutorture.c
3386
tick_dep_set_task(current, TICK_DEP_BIT_RCU); // CPU bound, so need tick.
kernel/rcu/rcutorture.c
3447
tick_dep_clear_task(current, TICK_DEP_BIT_RCU);
kernel/rcu/rcutorture.c
3507
int oldnice = task_nice(current);
kernel/rcu/rcutorture.c
3518
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
3545
sched_set_normal(current, oldnice);
kernel/rcu/rcutorture.c
3662
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
3827
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
3844
set_user_nice(current, MAX_NICE);
kernel/rcu/rcutorture.c
3943
if (torture_sched_setaffinity(current->pid, cpumask_of(cpu), false))
kernel/rcu/rcutorture.c
3946
sched_set_fifo(current);
kernel/rcu/rcutorture.c
3947
torture_sched_setaffinity(current->pid, cpu_present_mask, true);
kernel/rcu/rcutorture.c
3949
sched_set_normal(current, 0);
kernel/rcu/refscale.c
1245
WARN_ON_ONCE(set_cpus_allowed_ptr(current, cpumask_of(me % nr_cpu_ids)));
kernel/rcu/refscale.c
1246
set_user_nice(current, MAX_NICE);
kernel/rcu/refscale.c
1373
set_cpus_allowed_ptr(current, cpumask_of(nreaders % nr_cpu_ids));
kernel/rcu/refscale.c
1374
set_user_nice(current, MAX_NICE);
kernel/rcu/tasks.h
1011
if (t != current && rcu_tasks_is_holdout(t)) {
kernel/rcu/tasks.h
1290
struct task_struct *t = current;
kernel/rcu/tasks.h
1311
struct task_struct *t = current;
kernel/rcu/tasks.h
642
housekeeping_affine(current, HK_TYPE_RCU);
kernel/rcu/tasks.h
643
smp_store_release(&rtp->kthread_ptr, current); // Let GPs start!
kernel/rcu/tasks.h
876
WARN_ON(signal_pending(current));
kernel/rcu/tiny.c
179
if (unlikely(is_idle_task(current))) {
kernel/rcu/tiny.c
214
if (unlikely(is_idle_task(current))) {
kernel/rcu/tree.c
1123
if ((current == t && !in_hardirq() && !in_serving_softirq()) ||
kernel/rcu/tree.c
2133
WARN_ON(signal_pending(current));
kernel/rcu/tree.c
2291
WARN_ON(signal_pending(current));
kernel/rcu/tree.c
2560
need_resched(), is_idle_task(current),
kernel/rcu/tree.c
2602
tick_dep_set_task(current, TICK_DEP_BIT_RCU);
kernel/rcu/tree.c
2625
if (count >= bl && (need_resched() || !is_idle_task(current)))
kernel/rcu/tree.c
2654
is_idle_task(current), rcu_is_callbacks_kthread(rdp));
kernel/rcu/tree.c
2685
tick_dep_clear_task(current, TICK_DEP_BIT_RCU);
kernel/rcu/tree.c
269
rcu_preempt_deferred_qs(current);
kernel/rcu/tree.c
270
rcu_tasks_qs(current, false);
kernel/rcu/tree.c
2719
rcu_note_voluntary_context_switch(current);
kernel/rcu/tree.c
2847
rcu_preempt_deferred_qs(current);
kernel/rcu/tree.c
2848
} else if (rcu_preempt_need_deferred_qs(current)) {
kernel/rcu/tree.c
2895
if (t && (status != RCU_KTHREAD_YIELDING || is_idle_task(current)))
kernel/rcu/tree.c
2907
if (t != NULL && t != current)
kernel/rcu/tree.c
366
rcu_preempt_deferred_qs(current);
kernel/rcu/tree.c
406
RCU_LOCKDEP_WARN(!in_task() || !is_idle_task(current),
kernel/rcu/tree.c
4413
rcu_preempt_deferred_qs(current);
kernel/rcu/tree.c
603
if (IS_ENABLED(CONFIG_GENERIC_ENTRY) && !(current->flags & PF_VCPU))
kernel/rcu/tree.c
606
if (IS_ENABLED(CONFIG_VIRT_XFER_TO_GUEST_WORK) && (current->flags & PF_VCPU))
kernel/rcu/tree_exp.h
755
struct task_struct *t = current;
kernel/rcu/tree_nocb.h
1480
WARN_ON(sched_setaffinity(current->pid, rcu_nocb_mask));
kernel/rcu/tree_nocb.h
26
if (current == rdp->nocb_cb_kthread || current == rdp->nocb_gp_kthread)
kernel/rcu/tree_nocb.h
822
WARN_ON(signal_pending(current));
kernel/rcu/tree_nocb.h
878
WARN_ON(signal_pending(current));
kernel/rcu/tree_plugin.h
1006
rcu_tasks_qs(current, preempt);
kernel/rcu/tree_plugin.h
1119
sched_setscheduler_nocheck(current, SCHED_FIFO, &sp);
kernel/rcu/tree_plugin.h
1128
return rdp->nocb_cb_kthread == current;
kernel/rcu/tree_plugin.h
1140
return rdp->rcu_cpu_kthread_task == current ||
kernel/rcu/tree_plugin.h
1368
housekeeping_affine(current, HK_TYPE_RCU);
kernel/rcu/tree_plugin.h
169
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
307
WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, false);
kernel/rcu/tree_plugin.h
326
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
371
rcu_tasks_qs(current, preempt);
kernel/rcu/tree_plugin.h
391
WRITE_ONCE(current->rcu_read_lock_nesting, READ_ONCE(current->rcu_read_lock_nesting) + 1);
kernel/rcu/tree_plugin.h
396
int ret = READ_ONCE(current->rcu_read_lock_nesting) - 1;
kernel/rcu/tree_plugin.h
398
WRITE_ONCE(current->rcu_read_lock_nesting, ret);
kernel/rcu/tree_plugin.h
404
WRITE_ONCE(current->rcu_read_lock_nesting, val);
kernel/rcu/tree_plugin.h
418
WRITE_ONCE(current->rcu_read_unlock_special.b.need_qs, true);
kernel/rcu/tree_plugin.h
432
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
812
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
847
struct task_struct *t = current;
kernel/rcu/tree_plugin.h
849
if (unlikely(!list_empty(&current->rcu_node_entry))) {
kernel/rcu/tree_plugin.h
859
rcu_preempt_deferred_qs(current);
kernel/rcu/update.c
321
current->lockdep_recursion == 0;
kernel/reboot.c
262
current->flags |= PF_NO_SETAFFINITY;
kernel/reboot.c
265
set_cpus_allowed_ptr(current, cpumask_of(cpu));
kernel/reboot.c
731
struct pid_namespace *pid_ns = task_active_pid_ns(current);
kernel/resource.c
1318
DECLARE_WAITQUEUE(wait, current);
kernel/rseq.c
266
struct task_struct *t = current;
kernel/rseq.c
344
if (unlikely(!rseq_handle_cs(current, regs))) {
kernel/rseq.c
349
current->rseq.event.error = 0;
kernel/rseq.c
360
struct task_struct *t = current;
kernel/rseq.c
397
if (rseq_set_ids(current, &ids, 0))
kernel/rseq.c
418
if (current->rseq.usrptr != rseq || !current->rseq.usrptr)
kernel/rseq.c
420
if (rseq_len != current->rseq.len)
kernel/rseq.c
422
if (current->rseq.sig != sig)
kernel/rseq.c
426
rseq_reset(current);
kernel/rseq.c
433
if (current->rseq.usrptr) {
kernel/rseq.c
439
if (current->rseq.usrptr != rseq || rseq_len != current->rseq.len)
kernel/rseq.c
441
if (current->rseq.sig != sig)
kernel/rseq.c
495
current->rseq.usrptr = rseq;
kernel/rseq.c
496
current->rseq.len = rseq_len;
kernel/rseq.c
497
current->rseq.sig = sig;
kernel/rseq.c
500
current->rseq.slice.state.enabled = !!(rseqfl & RSEQ_CS_FLAG_SLICE_EXT_ENABLED);
kernel/rseq.c
508
current->rseq.event.has_rseq = true;
kernel/rseq.c
545
if (st->cookie == current && current->rseq.slice.state.granted) {
kernel/rseq.c
555
struct task_struct *curr = current;
kernel/rseq.c
600
if (st->cookie == current)
kernel/rseq.c
630
u32 __user *sctrl = &current->rseq.usrptr->slice_ctrl.all;
kernel/rseq.c
649
struct task_struct *curr = current;
kernel/rseq.c
704
return current->rseq.slice.state.enabled ? PR_RSEQ_SLICE_EXT_ENABLE : 0;
kernel/rseq.c
714
if (!current->rseq.usrptr)
kernel/rseq.c
718
if (enable == !!current->rseq.slice.state.enabled)
kernel/rseq.c
721
if (get_user(rflags, &current->rseq.usrptr->flags))
kernel/rseq.c
724
if (current->rseq.slice.state.enabled)
kernel/rseq.c
735
if (put_user(rflags, &current->rseq.usrptr->flags))
kernel/rseq.c
738
current->rseq.slice.state.enabled = enable;
kernel/rseq.c
764
int yielded = !!current->rseq.slice.yielded;
kernel/rseq.c
766
current->rseq.slice.yielded = 0;
kernel/scftorture.c
486
WARN_ON_ONCE(set_cpus_allowed_ptr(current, cpumask_of(cpu)));
kernel/scftorture.c
487
set_user_nice(current, MAX_NICE);
kernel/scftorture.c
519
set_cpus_allowed_ptr(current, cpumask_of(cpu));
kernel/sched/autogroup.c
215
sig->autogroup = autogroup_task_get(current);
kernel/sched/autogroup.c
243
err = security_task_setnice(current, nice);
kernel/sched/autogroup.c
247
if (nice < 0 && !can_nice(current, nice))
kernel/sched/completion.c
93
if (signal_pending_state(state, current)) {
kernel/sched/core.c
10638
struct mm_struct *mm = current->mm;
kernel/sched/core.c
10645
if (t != current)
kernel/sched/core.c
10694
mm_cid_transit_to_task(current, pcp);
kernel/sched/core.c
10696
mm_cid_transit_to_cpu(current, pcp);
kernel/sched/core.c
10739
if (WARN_ON_ONCE(current->mm != mm))
kernel/sched/core.c
10779
if (t == current)
kernel/sched/core.c
2081
if (!p || p == current)
kernel/sched/core.c
2372
struct task_struct *p = current;
kernel/sched/core.c
3061
if (WARN_ON_ONCE(p == current &&
kernel/sched/core.c
4099
if (p == current) {
kernel/sched/core.c
4640
p->prio = current->normal_prio;
kernel/sched/core.c
4826
hlist_add_head(&notifier->link, &current->preempt_notifiers);
kernel/sched/core.c
5053
if (unlikely(current->kmap_ctrl.idx))
kernel/sched/core.c
5061
if (unlikely(current->kmap_ctrl.idx))
kernel/sched/core.c
5132
current->comm, current->pid, preempt_count()))
kernel/sched/core.c
5150
perf_event_task_sched_in(prev, current);
kernel/sched/core.c
5155
kcov_finish_switch(current);
kernel/sched/core.c
5165
fire_sched_in_preempt_notifiers(current);
kernel/sched/core.c
5229
if (current->set_child_tid)
kernel/sched/core.c
5230
put_user(task_pid_vnr(current), current->set_child_tid);
kernel/sched/core.c
538
trace_sched_set_state_tp(current, state_value);
kernel/sched/core.c
5413
struct task_struct *p = current;
kernel/sched/core.c
5749
current->preempt_disable_ip = ip;
kernel/sched/core.c
5829
unsigned long preempt_disable_ip = get_preempt_disable_ip(current);
kernel/sched/core.c
6930
current->flags |= PF_NOFREEZE;
kernel/sched/core.c
6966
WARN_ON_ONCE(current->__state & TASK_RTLOCK_WAIT);
kernel/sched/core.c
7000
struct task_struct *tsk = current;
kernel/sched/core.c
7032
WARN_ON_ONCE(current->__state);
kernel/sched/core.c
7282
lockdep_assert(!fetch_and_set(current->sched_rt_mutex, 1));
kernel/sched/core.c
7283
sched_submit_work(current);
kernel/sched/core.c
7288
lockdep_assert(current->sched_rt_mutex);
kernel/sched/core.c
7294
sched_update_worker(current);
kernel/sched/core.c
7295
lockdep_assert(fetch_and_set(current->sched_rt_mutex, 0));
kernel/sched/core.c
7801
int old_iowait = current->in_iowait;
kernel/sched/core.c
7803
current->in_iowait = 1;
kernel/sched/core.c
7804
blk_flush_plug(current->plug, true);
kernel/sched/core.c
7810
current->in_iowait = token;
kernel/sched/core.c
8069
struct mm_struct *mm = current->active_mm;
kernel/sched/core.c
8074
current->active_mm = &init_mm;
kernel/sched/core.c
8075
switch_mm_irqs_off(mm, &init_mm, current);
kernel/sched/core.c
8560
if (set_cpus_allowed_ptr(current, housekeeping_cpumask(HK_TYPE_DOMAIN)) < 0)
kernel/sched/core.c
8562
current->flags &= ~PF_NO_SETAFFINITY;
kernel/sched/core.c
8765
enter_lazy_tlb(&init_mm, current);
kernel/sched/core.c
8773
WARN_ON(!set_kthread_struct(current));
kernel/sched/core.c
8781
__sched_fork(0, current);
kernel/sched/core.c
8782
init_idle(current, smp_processor_id());
kernel/sched/core.c
8811
WARN_ONCE(state != TASK_RUNNING && current->task_state_change,
kernel/sched/core.c
8814
(void *)current->task_state_change,
kernel/sched/core.c
8815
(void *)current->task_state_change);
kernel/sched/core.c
8853
!is_idle_task(current) && !current->non_block_count) ||
kernel/sched/core.c
8863
preempt_disable_ip = get_preempt_disable_ip(current);
kernel/sched/core.c
8868
in_atomic(), irqs_disabled(), current->non_block_count,
kernel/sched/core.c
8869
current->pid, current->comm);
kernel/sched/core.c
8878
if (task_stack_end_corrupted(current))
kernel/sched/core.c
8881
debug_show_held_locks(current);
kernel/sched/core.c
8883
print_irqtrace_events(current);
kernel/sched/core.c
8913
current->pid, current->comm);
kernel/sched/core.c
8915
debug_show_held_locks(current);
kernel/sched/core.c
8929
if (is_migration_disabled(current))
kernel/sched/core.c
8944
in_atomic(), irqs_disabled(), is_migration_disabled(current),
kernel/sched/core.c
8945
current->pid, current->comm);
kernel/sched/core.c
8947
debug_show_held_locks(current);
kernel/sched/core_sched.c
115
p->core_cookie = sched_core_clone_cookie(current);
kernel/sched/core_sched.c
152
task = current;
kernel/sched/core_sched.c
195
cookie = sched_core_clone_cookie(current);
kernel/sched/core_sched.c
204
__sched_core_set(current, cookie);
kernel/sched/cputime.c
341
if (same_thread_group(current, tsk))
kernel/sched/cputime.c
342
(void) task_sched_runtime(current);
kernel/sched/cputime.c
419
irqtime_account_process_tick(current, 0, ticks);
kernel/sched/cputime.c
794
vtime = &current->vtime;
kernel/sched/cputime.c
797
if (is_idle_task(current))
kernel/sched/cputime.c
799
else if (current->flags & PF_VCPU)
kernel/sched/ext.c
278
WARN_ONCE((mask | higher_bits(mask)) & current->scx.kf_mask,
kernel/sched/ext.c
280
current->scx.kf_mask, mask);
kernel/sched/ext.c
281
current->scx.kf_mask |= mask;
kernel/sched/ext.c
288
current->scx.kf_mask &= ~mask;
kernel/sched/ext.c
358
current->scx.kf_tasks[0] = task; \
kernel/sched/ext.c
360
current->scx.kf_tasks[0] = NULL; \
kernel/sched/ext.c
367
current->scx.kf_tasks[0] = task; \
kernel/sched/ext.c
369
current->scx.kf_tasks[0] = NULL; \
kernel/sched/ext.c
377
current->scx.kf_tasks[0] = task0; \
kernel/sched/ext.c
378
current->scx.kf_tasks[1] = task1; \
kernel/sched/ext.c
380
current->scx.kf_tasks[0] = NULL; \
kernel/sched/ext.c
381
current->scx.kf_tasks[1] = NULL; \
kernel/sched/ext.c
388
if (unlikely(!(current->scx.kf_mask & mask))) {
kernel/sched/ext.c
390
mask, current->scx.kf_mask);
kernel/sched/ext.c
402
(current->scx.kf_mask & higher_bits(SCX_KF_CPU_RELEASE)))) {
kernel/sched/ext.c
408
(current->scx.kf_mask & higher_bits(SCX_KF_DISPATCH)))) {
kernel/sched/ext.c
424
if (unlikely((p != current->scx.kf_tasks[0] &&
kernel/sched/ext.c
425
p != current->scx.kf_tasks[1]))) {
kernel/sched/ext.c
4728
current->comm, current->pid, ei->kind);
kernel/sched/ext_idle.c
547
if (!(current->flags & PF_EXITING) &&
kernel/sched/ext_idle.c
885
return p != current;
kernel/sched/ext_internal.h
1177
return !current->scx.kf_mask;
kernel/sched/fair.c
1553
return rcu_dereference_check(p->numa_group, p == current ||
kernel/sched/fair.c
1559
return rcu_dereference_protected(p->numa_group, p == current);
kernel/sched/fair.c
2025
this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid);
kernel/sched/fair.c
3138
if (tsk->mm == current->mm)
kernel/sched/fair.c
3227
struct task_struct *p = current;
kernel/sched/fair.c
3229
int cpu_node = task_node(current);
kernel/sched/fair.c
3329
if ((READ_ONCE(current->mm->numa_scan_seq) - vma->numab_state->start_scan_seq) < 2)
kernel/sched/fair.c
3333
if (test_bit(hash_32(current->pid, ilog2(BITS_PER_LONG)), &pids))
kernel/sched/fair.c
3351
(vma->numab_state->prev_scan_seq + get_nr_threads(current)))
kernel/sched/fair.c
3366
struct task_struct *p = current;
kernel/sched/fair.c
3396
trace_sched_skip_cpuset_numa(current, &cpuset_current_mems_allowed);
kernel/sched/fair.c
3660
delay = min_t(unsigned int, task_scan_max(current),
kernel/sched/fair.c
3661
current->numa_scan_period * mm_users * NSEC_PER_MSEC);
kernel/sched/fair.c
5780
WARN_ON_ONCE(p != current);
kernel/sched/fair.c
7292
if (time_after(jiffies, current->wakee_flip_decay_ts + HZ)) {
kernel/sched/fair.c
7293
current->wakee_flips >>= 1;
kernel/sched/fair.c
7294
current->wakee_flip_decay_ts = jiffies;
kernel/sched/fair.c
7297
if (current->last_wakee != p) {
kernel/sched/fair.c
7298
current->last_wakee = p;
kernel/sched/fair.c
7299
current->wakee_flips++;
kernel/sched/fair.c
7322
unsigned int master = current->wakee_flips;
kernel/sched/fair.c
7386
unsigned long current_load = task_h_load(current);
kernel/sched/fair.c
7885
if (is_per_cpu_kthread(current) &&
kernel/sched/fair.c
8060
else if (p && unlikely(task_on_rq_queued(p) || current == p))
kernel/sched/fair.c
8582
int sync = (wake_flags & WF_SYNC) && !(current->flags & PF_EXITING);
kernel/sched/idle.c
383
if (unlikely(klp_patch_pending(current)))
kernel/sched/idle.c
384
klp_update_patch_state(current);
kernel/sched/idle.c
403
set_tsk_need_resched(current);
kernel/sched/idle.c
416
WARN_ON_ONCE(current->policy != SCHED_FIFO);
kernel/sched/idle.c
417
WARN_ON_ONCE(current->nr_cpus_allowed != 1);
kernel/sched/idle.c
418
WARN_ON_ONCE(!(current->flags & PF_KTHREAD));
kernel/sched/idle.c
419
WARN_ON_ONCE(!(current->flags & PF_NO_SETAFFINITY));
kernel/sched/idle.c
421
WARN_ON_ONCE(current->mm);
kernel/sched/idle.c
425
current->flags |= PF_IDLE;
kernel/sched/idle.c
438
current->flags &= ~PF_IDLE;
kernel/sched/idle.c
447
current->flags |= PF_IDLE;
kernel/sched/membarrier.c
202
rseq_sched_switch_event(current);
kernel/sched/membarrier.c
209
if (current->mm != mm)
kernel/sched/membarrier.c
319
struct mm_struct *mm = current->mm;
kernel/sched/membarrier.c
498
struct task_struct *p = current;
kernel/sched/membarrier.c
517
struct task_struct *p = current;
kernel/sched/membarrier.c
559
struct task_struct *p = current;
kernel/sched/psi.c
1064
*flags = current->in_memstall;
kernel/sched/psi.c
1074
current->in_memstall = 1;
kernel/sched/psi.c
1075
psi_task_change(current, 0, TSK_MEMSTALL | TSK_MEMSTALL_RUNNING);
kernel/sched/psi.c
1104
current->in_memstall = 0;
kernel/sched/psi.c
1105
psi_task_change(current, TSK_MEMSTALL | TSK_MEMSTALL_RUNNING, 0);
kernel/sched/psi.c
742
sched_set_fifo_low(current);
kernel/sched/swait.c
109
if (signal_pending_state(state, current)) {
kernel/sched/swait.c
87
wait->task = current;
kernel/sched/syscalls.c
142
nice = task_nice(current) + increment;
kernel/sched/syscalls.c
145
if (increment < 0 && !can_nice(current, nice))
kernel/sched/syscalls.c
148
retval = security_task_setnice(current, nice);
kernel/sched/syscalls.c
152
set_user_nice(current, nice);
kernel/sched/syscalls.c
217
return pid ? find_task_by_vpid(pid) : current;
kernel/sched/wait.c
283
wq_entry->private = current;
kernel/sched/wait.c
295
if (signal_pending_state(state, current)) {
kernel/sched/wait.c
338
if (signal_pending(current))
kernel/sched/wait.c
355
if (signal_pending(current))
kernel/sched/wait_bit.c
194
.private = current,
kernel/sched/wait_bit.c
241
if (signal_pending_state(mode, current))
kernel/sched/wait_bit.c
251
if (signal_pending_state(mode, current))
kernel/sched/wait_bit.c
265
if (signal_pending_state(mode, current))
kernel/seccomp.c
1082
current->seccomp.mode = SECCOMP_MODE_DEAD;
kernel/seccomp.c
1090
int mode = current->seccomp.mode;
kernel/seccomp.c
1093
unlikely(current->ptrace & PT_SUSPEND_SECCOMP))
kernel/seccomp.c
1105
int this_syscall = syscall_get_nr(current, current_pt_regs());
kernel/seccomp.c
1178
n.task = current;
kernel/seccomp.c
1254
syscall_set_return_value(current, current_pt_regs(),
kernel/seccomp.c
1283
syscall_set_return_value(current, current_pt_regs(),
kernel/seccomp.c
1289
syscall_rollback(current, current_pt_regs());
kernel/seccomp.c
1300
if (!ptrace_event_enabled(current, PTRACE_EVENT_SECCOMP)) {
kernel/seccomp.c
1301
syscall_set_return_value(current,
kernel/seccomp.c
1319
if (fatal_signal_pending(current))
kernel/seccomp.c
1322
this_syscall = syscall_get_nr(current, current_pt_regs());
kernel/seccomp.c
1358
current->seccomp.mode = SECCOMP_MODE_DEAD;
kernel/seccomp.c
1362
(atomic_read(&current->signal->live) == 1)) {
kernel/seccomp.c
1364
syscall_rollback(current, current_pt_regs());
kernel/seccomp.c
1390
int mode = current->seccomp.mode;
kernel/seccomp.c
1394
unlikely(current->ptrace & PT_SUSPEND_SECCOMP))
kernel/seccomp.c
1397
this_syscall = syscall_get_nr(current, current_pt_regs());
kernel/seccomp.c
1418
return current->seccomp.mode;
kernel/seccomp.c
1433
spin_lock_irq(&current->sighand->siglock);
kernel/seccomp.c
1441
seccomp_assign_mode(current, seccomp_mode, 0);
kernel/seccomp.c
1445
spin_unlock_irq(&current->sighand->siglock);
kernel/seccomp.c
1931
lockdep_assert_held(&current->sighand->siglock);
kernel/seccomp.c
1935
for (cur = current->seccomp.filter; cur; cur = cur->prev) {
kernel/seccomp.c
2014
mutex_lock_killable(&current->signal->cred_guard_mutex))
kernel/seccomp.c
2017
spin_lock_irq(&current->sighand->siglock);
kernel/seccomp.c
2033
seccomp_assign_mode(current, seccomp_mode, flags);
kernel/seccomp.c
2035
spin_unlock_irq(&current->sighand->siglock);
kernel/seccomp.c
2037
mutex_unlock(&current->signal->cred_guard_mutex);
kernel/seccomp.c
2221
current->seccomp.mode != SECCOMP_MODE_DISABLED) {
kernel/seccomp.c
2259
current->seccomp.mode != SECCOMP_MODE_DISABLED) {
kernel/seccomp.c
250
struct task_struct *task = current;
kernel/seccomp.c
410
READ_ONCE(current->seccomp.filter);
kernel/seccomp.c
437
assert_spin_locked(&current->sighand->siglock);
kernel/seccomp.c
439
if (current->seccomp.mode && current->seccomp.mode != seccomp_mode)
kernel/seccomp.c
492
BUG_ON(!mutex_is_locked(&current->signal->cred_guard_mutex));
kernel/seccomp.c
493
assert_spin_locked(&current->sighand->siglock);
kernel/seccomp.c
496
caller = current;
kernel/seccomp.c
602
BUG_ON(!mutex_is_locked(&current->signal->cred_guard_mutex));
kernel/seccomp.c
603
assert_spin_locked(&current->sighand->siglock);
kernel/seccomp.c
609
if (current->signal->flags & SIGNAL_GROUP_EXIT)
kernel/seccomp.c
613
caller = current;
kernel/seccomp.c
691
if (!task_no_new_privs(current) &&
kernel/seccomp.c
927
assert_spin_locked(&current->sighand->siglock);
kernel/seccomp.c
931
for (walker = current->seccomp.filter; walker; walker = walker->prev)
kernel/seccomp.c
961
filter->prev = current->seccomp.filter;
kernel/seccomp.c
963
current->seccomp.filter = filter;
kernel/seccomp.c
964
atomic_inc(&current->seccomp.filter_count);
kernel/signal.c
1097
q->info.si_pid = task_tgid_nr_ns(current,
kernel/signal.c
1191
force = !task_pid_nr_ns(current, task_active_pid_ns(t));
kernel/signal.c
1211
if (!task_pid_nr_ns(current, task_active_pid_ns(t))) {
kernel/signal.c
1221
struct pt_regs *regs = task_pt_regs(current);
kernel/signal.c
1224
exe_file = get_task_exe_file(current);
kernel/signal.c
1227
exe_file, current->comm, signr);
kernel/signal.c
1231
current->comm, signr);
kernel/signal.c
1331
return force_sig_info_to_task(info, current, HANDLER_CURRENT);
kernel/signal.c
1586
pid ? find_vpid(-pid) : task_pgrp(current));
kernel/signal.c
1593
!same_thread_group(p, current)) {
kernel/signal.c
1659
force_sig_info_to_task(&info, current, HANDLER_SIG_DFL);
kernel/signal.c
1672
force_sig_info_to_task(&info, current, HANDLER_EXIT);
kernel/signal.c
1704
return force_sig_fault_to_task(sig, code, addr, current);
kernel/signal.c
179
if (!recalc_sigpending_tsk(current) && !freezing(current)) {
kernel/signal.c
1796
info.si_perf_flags = sigismember(&current->blocked, info.si_signo) ?
kernel/signal.c
1800
return send_sig_info(info.si_signo, &info, current);
kernel/signal.c
1818
info.si_call_addr = (void __user *)KSTK_EIP(current);
kernel/signal.c
1820
info.si_arch = syscall_get_arch(current);
kernel/signal.c
1822
return force_sig_info_to_task(&info, current,
kernel/signal.c
191
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
192
set_tsk_thread_flag(current, TIF_SIGPENDING);
kernel/signal.c
1925
struct task_struct *tsk = current;
kernel/signal.c
1934
struct ucounts *ucounts = sig_get_ucounts(current, -1, 0);
kernel/signal.c
194
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
1971
same_thread_group(t, current) && !current->exit_state)
kernel/signal.c
1972
t = current;
kernel/signal.c
2351
__releases(&current->sighand->siglock)
kernel/signal.c
2352
__acquires(&current->sighand->siglock)
kernel/signal.c
2365
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2367
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
2376
if (!current->ptrace || __fatal_signal_pending(current))
kernel/signal.c
2380
current->jobctl |= JOBCTL_TRACED;
kernel/signal.c
2402
current->ptrace_message = message;
kernel/signal.c
2403
current->last_siginfo = info;
kernel/signal.c
2404
current->exit_code = exit_code;
kernel/signal.c
2413
if (why == CLD_STOPPED && (current->jobctl & JOBCTL_STOP_PENDING))
kernel/signal.c
2414
gstop_done = task_participate_group_stop(current);
kernel/signal.c
2417
task_clear_jobctl_pending(current, JOBCTL_TRAP_STOP);
kernel/signal.c
2419
task_clear_jobctl_pending(current, JOBCTL_TRAP_NOTIFY);
kernel/signal.c
2422
task_clear_jobctl_trapping(current);
kernel/signal.c
2424
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2436
if (current->ptrace)
kernel/signal.c
2437
do_notify_parent_cldstop(current, true, why);
kernel/signal.c
2438
if (gstop_done && (!current->ptrace || ptrace_reparented(current)))
kernel/signal.c
2439
do_notify_parent_cldstop(current, false, why);
kernel/signal.c
2482
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
2483
exit_code = current->exit_code;
kernel/signal.c
2484
current->last_siginfo = NULL;
kernel/signal.c
2485
current->ptrace_message = 0;
kernel/signal.c
2486
current->exit_code = 0;
kernel/signal.c
2489
current->jobctl &= ~(JOBCTL_LISTENING | JOBCTL_PTRACE_FROZEN);
kernel/signal.c
2496
recalc_sigpending_tsk(current);
kernel/signal.c
2507
info.si_pid = task_pid_vnr(current);
kernel/signal.c
2519
if (unlikely(task_work_pending(current)))
kernel/signal.c
2522
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
2524
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2551
__releases(&current->sighand->siglock)
kernel/signal.c
2553
struct signal_struct *sig = current->signal;
kernel/signal.c
2555
if (!(current->jobctl & JOBCTL_STOP_PENDING)) {
kernel/signal.c
2562
if (!likely(current->jobctl & JOBCTL_STOP_DEQUEUED) ||
kernel/signal.c
2589
if (task_set_jobctl_pending(current, signr | gstop))
kernel/signal.c
2592
for_other_threads(current, t) {
kernel/signal.c
260
current->comm, current->pid, sig);
kernel/signal.c
2609
if (likely(!current->ptrace)) {
kernel/signal.c
2617
if (task_participate_group_stop(current))
kernel/signal.c
2620
current->jobctl |= JOBCTL_STOPPED;
kernel/signal.c
2622
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2635
do_notify_parent_cldstop(current, false, notify);
kernel/signal.c
2648
task_set_jobctl_pending(current, JOBCTL_TRAP_STOP);
kernel/signal.c
2670
struct signal_struct *signal = current->signal;
kernel/signal.c
2671
int signr = current->jobctl & JOBCTL_STOP_SIGMASK;
kernel/signal.c
2673
if (current->ptrace & PT_SEIZED) {
kernel/signal.c
2697
__releases(&current->sighand->siglock)
kernel/signal.c
2704
if ((current->jobctl & (JOBCTL_PENDING_MASK | JOBCTL_TRAP_FREEZE)) !=
kernel/signal.c
2706
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2718
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
2727
if (unlikely(task_work_pending(current)))
kernel/signal.c
2742
current->jobctl |= JOBCTL_STOP_DEQUEUED;
kernel/signal.c
2761
info->si_pid = task_pid_vnr(current->parent);
kernel/signal.c
2763
task_uid(current->parent));
kernel/signal.c
2768
if (sigismember(&current->blocked, signr) ||
kernel/signal.c
2769
fatal_signal_pending(current)) {
kernel/signal.c
2770
send_signal_locked(signr, info, current, type);
kernel/signal.c
2801
struct sighand_struct *sighand = current->sighand;
kernel/signal.c
2802
struct signal_struct *signal = current->signal;
kernel/signal.c
2806
if (unlikely(task_work_pending(current)))
kernel/signal.c
2809
if (!task_sigpending(current))
kernel/signal.c
2851
do_notify_parent_cldstop(current, false, why);
kernel/signal.c
2853
if (ptrace_reparented(current->group_leader))
kernel/signal.c
2854
do_notify_parent_cldstop(current->group_leader,
kernel/signal.c
2869
sigdelset(&current->pending.signal, SIGKILL);
kernel/signal.c
2880
if (unlikely(current->jobctl & JOBCTL_STOP_PENDING) &&
kernel/signal.c
2884
if (unlikely(current->jobctl &
kernel/signal.c
2886
if (current->jobctl & JOBCTL_TRAP_MASK) {
kernel/signal.c
2889
} else if (current->jobctl & JOBCTL_TRAP_FREEZE)
kernel/signal.c
2899
if (unlikely(cgroup_task_frozen(current))) {
kernel/signal.c
2914
signr = dequeue_signal(&current->blocked, &ksig->info, &type);
kernel/signal.c
2919
if (unlikely(current->ptrace) && (signr != SIGKILL) &&
kernel/signal.c
2999
if (unlikely(cgroup_task_frozen(current)))
kernel/signal.c
3005
current->flags |= PF_SIGNALED;
kernel/signal.c
3010
proc_coredump_connector(current);
kernel/signal.c
3028
if (current->flags & PF_USER_WORKER)
kernel/signal.c
3067
sigorsets(&blocked, &current->blocked, &ksig->ka.sa.sa_mask);
kernel/signal.c
3071
if (current->sas_ss_flags & SS_AUTODISARM)
kernel/signal.c
3072
sas_ss_reset(current);
kernel/signal.c
3175
struct restart_block *restart = &current->restart_block;
kernel/signal.c
3189
sigandnsets(&newblocked, newset, &current->blocked);
kernel/signal.c
3211
struct task_struct *tsk = current;
kernel/signal.c
3235
struct task_struct *tsk = current;
kernel/signal.c
3282
current->saved_sigmask = current->blocked;
kernel/signal.c
3302
current->saved_sigmask = current->blocked;
kernel/signal.c
3326
old_set = current->blocked;
kernel/signal.c
3350
sigset_t old_set = current->blocked;
kernel/signal.c
3373
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
3374
sigorsets(set, &current->pending.signal,
kernel/signal.c
3375
&current->signal->shared_pending.signal);
kernel/signal.c
3376
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
3379
sigandsets(set, &current->blocked, set);
kernel/signal.c
3747
struct task_struct *tsk = current;
kernel/signal.c
389
unsigned long mask = current->jobctl & JOBCTL_STOP_SIGMASK;
kernel/signal.c
390
struct signal_struct *sig = current->signal;
kernel/signal.c
3938
info->si_pid = task_tgid_vnr(current);
kernel/signal.c
3963
struct pid_namespace *active = task_active_pid_ns(current);
kernel/signal.c
4036
if ((task_pid(current) != pid || type > PIDTYPE_TGID) &&
kernel/signal.c
4080
pid = get_task_pid(current, PIDTYPE_PID);
kernel/signal.c
4084
pid = get_task_pid(current, PIDTYPE_TGID);
kernel/signal.c
4196
(task_pid_vnr(current) != pid))
kernel/signal.c
4243
(task_pid_vnr(current) != pid))
kernel/signal.c
4279
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
4280
current->sighand->action[sig - 1].sa.sa_handler = action;
kernel/signal.c
4287
flush_sigqueue_mask(current, &mask, &current->signal->shared_pending);
kernel/signal.c
4288
flush_sigqueue_mask(current, &mask, &current->pending);
kernel/signal.c
4291
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
4302
struct task_struct *p = current, *t;
kernel/signal.c
4371
__acquires(&current->sighand->siglock)
kernel/signal.c
4373
spin_lock_irq(&current->sighand->siglock);
kernel/signal.c
4377
__releases(&current->sighand->siglock)
kernel/signal.c
4379
spin_unlock_irq(&current->sighand->siglock);
kernel/signal.c
4390
struct task_struct *t = current;
kernel/signal.c
4398
(current->sas_ss_flags & SS_FLAG_BITS);
kernel/signal.c
4471
struct task_struct *t = current;
kernel/signal.c
4525
struct task_struct *t = current;
kernel/signal.c
4585
old_set = current->blocked.sig[0];
kernel/signal.c
4591
new_blocked = current->blocked;
kernel/signal.c
4784
return current->blocked.sig[0];
kernel/signal.c
4789
int old = current->blocked.sig[0];
kernel/signal.c
4822
while (!signal_pending(current)) {
kernel/signal.c
4833
current->saved_sigmask = current->blocked;
kernel/signal.c
4836
while (!signal_pending(current)) {
kernel/signal.c
620
struct task_struct *tsk = current;
kernel/signal.c
656
current->jobctl |= JOBCTL_STOP_DEQUEUED;
kernel/signal.c
670
struct task_struct *tsk = current;
kernel/signal.c
815
if (!same_thread_group(current, t) &&
kernel/signal.c
824
if (!sid || sid == task_session(current))
kernel/smpboot.c
39
per_cpu(idle_threads, smp_processor_id()) = current;
kernel/softirq.c
1114
sched_set_fifo_low(current);
kernel/softirq.c
166
if (!current->softirq_disable_cnt) {
kernel/softirq.c
190
current->softirq_disable_cnt = newcnt;
kernel/softirq.c
200
if (!current->softirq_disable_cnt)
kernel/softirq.c
204
current->softirq_disable_cnt += cnt;
kernel/softirq.c
205
WARN_ON_ONCE(current->softirq_disable_cnt < 0);
kernel/softirq.c
223
DEBUG_LOCKS_WARN_ON(current->softirq_disable_cnt !=
kernel/softirq.c
228
if (current->softirq_disable_cnt == cnt)
kernel/softirq.c
240
current->softirq_disable_cnt = newcnt;
kernel/softirq.c
247
current->softirq_disable_cnt -= cnt;
kernel/softirq.c
249
if (unlock && !current->softirq_disable_cnt) {
kernel/softirq.c
253
WARN_ON_ONCE(current->softirq_disable_cnt < 0);
kernel/softirq.c
274
curcnt = current->softirq_disable_cnt;
kernel/softirq.c
395
current->preempt_disable_ip = get_lock_parent_ip();
kernel/softirq.c
582
unsigned long old_flags = current->flags;
kernel/softirq.c
594
current->flags &= ~PF_MEMALLOC;
kernel/softirq.c
600
account_softirq_enter(current);
kernel/softirq.c
648
account_softirq_exit(current);
kernel/softirq.c
667
(is_idle_task(current) && (irq_count() == HARDIRQ_OFFSET)))
kernel/softirq.c
670
account_hardirq_enter(current);
kernel/softirq.c
720
account_hardirq_exit(current);
kernel/stacktrace.c
122
arch_stack_walk(consume_entry, &c, current, NULL);
kernel/stacktrace.c
144
.skip = skipnr + (current == tsk),
kernel/stacktrace.c
175
arch_stack_walk(consume_entry, &c, current, regs);
kernel/stacktrace.c
232
if (current->flags & PF_KTHREAD)
kernel/stacktrace.c
235
arch_stack_walk_user(consume_entry, &c, task_pt_regs(current));
kernel/stacktrace.c
299
.skip = skipnr + (current == task),
kernel/sys.c
1001
return task_tgid_vnr(current);
kernel/sys.c
1007
return task_pid_vnr(current);
kernel/sys.c
1021
pid = task_tgid_vnr(rcu_dereference(current->real_parent));
kernel/sys.c
1055
thread_group_cputime_adjusted(current, &tgutime, &tgstime);
kernel/sys.c
1056
cutime = current->signal->cutime;
kernel/sys.c
1057
cstime = current->signal->cstime;
kernel/sys.c
1117
struct task_struct *group_leader = current->group_leader;
kernel/sys.c
1195
grp = task_pgrp(current);
kernel/sys.c
1237
sid = task_session(current);
kernel/sys.c
1259
struct task_struct *curr = current->group_leader;
kernel/sys.c
1270
struct task_struct *group_leader = current->group_leader;
kernel/sys.c
1312
(personality(current->personality) == PER_LINUX32 && \
kernel/sys.c
1329
if (current->personality & UNAME26) {
kernel/sys.c
1424
if (!ns_capable(current->nsproxy->uts_ns->user_ns, CAP_SYS_ADMIN))
kernel/sys.c
1478
if (!ns_capable(current->nsproxy->uts_ns->user_ns, CAP_SYS_ADMIN))
kernel/sys.c
1568
ret = do_prlimit(current, resource, NULL, &value);
kernel/sys.c
1594
return do_prlimit(current, resource, &r, NULL);
kernel/sys.c
1603
ret = do_prlimit(current, resource, NULL, &r);
kernel/sys.c
1636
task_lock(current->group_leader);
kernel/sys.c
1637
x = current->signal->rlim[resource];
kernel/sys.c
1638
task_unlock(current->group_leader);
kernel/sys.c
1656
task_lock(current->group_leader);
kernel/sys.c
1657
r = current->signal->rlim[resource];
kernel/sys.c
1658
task_unlock(current->group_leader);
kernel/sys.c
1713
if (current == task)
kernel/sys.c
1751
tsk = pid ? find_task_by_vpid(pid) : current;
kernel/sys.c
1764
need_tasklist = !same_thread_group(tsk, current);
kernel/sys.c
1800
return do_prlimit(current, resource, &new_rlim, NULL);
kernel/sys.c
1862
task_cputime_adjusted(current, &utime, &stime);
kernel/sys.c
1942
getrusage(current, who, &r);
kernel/sys.c
1955
getrusage(current, who, &r);
kernel/sys.c
1962
mask = xchg(&current->fs->umask, mask & S_IRWXUGO);
kernel/sys.c
2064
struct mm_struct *mm = current->mm;
kernel/sys.c
2191
task_lock(current);
kernel/sys.c
2193
task_unlock(current);
kernel/sys.c
2201
struct mm_struct *mm = current->mm;
kernel/sys.c
2428
if (mm_flags_test(MMF_HAS_MDWE, current->mm))
kernel/sys.c
2430
if (mm_flags_test(MMF_HAS_MDWE_NO_INHERIT, current->mm))
kernel/sys.c
2463
mm_flags_set(MMF_HAS_MDWE_NO_INHERIT, current->mm);
kernel/sys.c
2465
mm_flags_set(MMF_HAS_MDWE, current->mm);
kernel/sys.c
2480
struct mm_struct *mm = current->mm;
kernel/sys.c
2491
struct mm_struct *mm = current->mm;
kernel/sys.c
2507
struct mm_struct *mm = current->mm;
kernel/sys.c
2515
if (mmap_write_lock_killable(current->mm))
kernel/sys.c
2529
mmap_write_unlock(current->mm);
kernel/sys.c
2536
struct task_struct *me = current;
kernel/sys.c
2640
if (current->timer_slack_ns > ULONG_MAX)
kernel/sys.c
2643
error = current->timer_slack_ns;
kernel/sys.c
2646
if (rt_or_dl_task_policy(current))
kernel/sys.c
2649
current->timer_slack_ns =
kernel/sys.c
2650
current->default_timer_slack_ns;
kernel/sys.c
2652
current->timer_slack_ns = arg2;
kernel/sys.c
2661
current->flags &= ~PF_MCE_PROCESS;
kernel/sys.c
2664
current->flags |= PF_MCE_PROCESS;
kernel/sys.c
2666
current->flags |= PF_MCE_EARLY;
kernel/sys.c
2668
current->flags &= ~PF_MCE_EARLY;
kernel/sys.c
2670
current->flags &=
kernel/sys.c
2682
if (current->flags & PF_MCE_PROCESS)
kernel/sys.c
2683
error = (current->flags & PF_MCE_EARLY) ?
kernel/sys.c
2709
task_set_no_new_privs(current);
kernel/sys.c
2714
return task_no_new_privs(current) ? 1 : 0;
kernel/sys.c
2786
current->flags |= PR_IO_FLUSHER;
kernel/sys.c
2788
current->flags &= ~PR_IO_FLUSHER;
kernel/sys.c
2799
error = (current->flags & PR_IO_FLUSHER) == PR_IO_FLUSHER;
kernel/sys.c
284
p = current;
kernel/sys.c
292
pgrp = task_pgrp(current);
kernel/sys.c
347
p = current;
kernel/sys.c
358
pgrp = task_pgrp(current);
kernel/sys.c
550
current->flags |= PF_NPROC_EXCEEDED;
kernel/sys.c
552
current->flags &= ~PF_NPROC_EXCEEDED;
kernel/sysctl.c
143
current->comm, table->procname);
kernel/task_work.c
202
struct task_struct *task = current;
kernel/task_work.c
65
if (WARN_ON_ONCE(task != current))
kernel/task_work.c
94
set_tsk_thread_flag(current, TIF_NOTIFY_RESUME);
kernel/taskstats.c
208
fill_stats(current_user_ns(), task_active_pid_ns(current), tsk, stats);
kernel/taskstats.c
308
if (task_active_pid_ns(current) != &init_pid_ns)
kernel/time/alarmtimer.c
702
alarm->data = (void *)current;
kernel/time/alarmtimer.c
710
} while (alarm->data && !signal_pending(current));
kernel/time/alarmtimer.c
719
if (freezing(current))
kernel/time/alarmtimer.c
721
restart = &current->restart_block;
kernel/time/alarmtimer.c
775
struct restart_block *restart = &current->restart_block;
kernel/time/hrtimer.c
2074
if (rt_or_dl_task_policy(current) && !(mode & HRTIMER_MODE_SOFT))
kernel/time/hrtimer.c
2079
sl->task = current;
kernel/time/hrtimer.c
2129
} while (t->task && !signal_pending(current));
kernel/time/hrtimer.c
2136
restart = &current->restart_block;
kernel/time/hrtimer.c
2170
hrtimer_set_expires_range_ns(&t.timer, rqtp, current->timer_slack_ns);
kernel/time/hrtimer.c
2181
restart = &current->restart_block;
kernel/time/hrtimer.c
2203
current->restart_block.fn = do_no_restart_syscall;
kernel/time/hrtimer.c
2204
current->restart_block.nanosleep.type = rmtp ? TT_NATIVE : TT_NONE;
kernel/time/hrtimer.c
2205
current->restart_block.nanosleep.rmtp = rmtp;
kernel/time/hrtimer.c
2225
current->restart_block.fn = do_no_restart_syscall;
kernel/time/hrtimer.c
2226
current->restart_block.nanosleep.type = rmtp ? TT_COMPAT : TT_NONE;
kernel/time/hrtimer.c
2227
current->restart_block.nanosleep.compat_rmtp = rmtp;
kernel/time/itimer.c
228
struct task_struct *tsk = current;
kernel/time/itimer.c
365
current->comm);
kernel/time/itimer.c
412
current->comm);
kernel/time/itimer.c
78
struct task_struct *tsk = current;
kernel/time/namespace.c
196
if (likely(vma->vm_mm == current->mm))
kernel/time/namespace.c
197
return current->nsproxy->time_ns->vvar_page;
kernel/time/posix-cpu-timers.c
1137
handle_posix_cpu_timers(current);
kernel/time/posix-cpu-timers.c
1202
clear_posix_cputimers_work(current);
kernel/time/posix-cpu-timers.c
1404
struct task_struct *tsk = current;
kernel/time/posix-cpu-timers.c
1494
timer.it_process = current;
kernel/time/posix-cpu-timers.c
1511
while (!signal_pending(current)) {
kernel/time/posix-cpu-timers.c
1559
restart = &current->restart_block;
kernel/time/posix-cpu-timers.c
1573
struct restart_block *restart_block = &current->restart_block;
kernel/time/posix-cpu-timers.c
1581
CPUCLOCK_PID(which_clock) == task_pid_vnr(current)))
kernel/time/posix-cpu-timers.c
71
return thread ? task_pid(current) : task_tgid(current);
kernel/time/posix-cpu-timers.c
79
return (tsk && same_thread_group(tsk, current)) ? pid : NULL;
kernel/time/posix-cpu-timers.c
818
rcu_assign_pointer(ctmr->handling, current);
kernel/time/posix-cpu-timers.c
854
current->comm, task_pid_nr(current));
kernel/time/posix-cpu-timers.c
856
send_signal_locked(signo, SEND_SIG_PRIV, current, PIDTYPE_TGID);
kernel/time/posix-cpu-timers.c
88
if (gettime && (pid == task_pid(current)))
kernel/time/posix-cpu-timers.c
89
return task_tgid(current);
kernel/time/posix-stubs.c
116
current->restart_block.fn = do_no_restart_syscall;
kernel/time/posix-stubs.c
117
current->restart_block.nanosleep.type = rmtp ? TT_NATIVE : TT_NONE;
kernel/time/posix-stubs.c
118
current->restart_block.nanosleep.rmtp = rmtp;
kernel/time/posix-stubs.c
199
current->restart_block.fn = do_no_restart_syscall;
kernel/time/posix-stubs.c
200
current->restart_block.nanosleep.type = rmtp ? TT_COMPAT : TT_NONE;
kernel/time/posix-stubs.c
201
current->restart_block.nanosleep.compat_rmtp = rmtp;
kernel/time/posix-timers.c
1035
scoped_guard (spinlock, &current->sighand->siglock) {
kernel/time/posix-timers.c
1387
current->restart_block.fn = do_no_restart_syscall;
kernel/time/posix-timers.c
1388
current->restart_block.nanosleep.type = rmtp ? TT_NATIVE : TT_NONE;
kernel/time/posix-timers.c
1389
current->restart_block.nanosleep.rmtp = rmtp;
kernel/time/posix-timers.c
1415
current->restart_block.fn = do_no_restart_syscall;
kernel/time/posix-timers.c
1416
current->restart_block.nanosleep.type = rmtp ? TT_COMPAT : TT_NONE;
kernel/time/posix-timers.c
1417
current->restart_block.nanosleep.compat_rmtp = rmtp;
kernel/time/posix-timers.c
159
struct signal_struct *sig = current->signal;
kernel/time/posix-timers.c
338
spin_unlock(&current->sighand->siglock);
kernel/time/posix-timers.c
345
spin_lock(&current->sighand->siglock);
kernel/time/posix-timers.c
380
current->signal->timer_create_restore_ids = 0;
kernel/time/posix-timers.c
383
current->signal->timer_create_restore_ids = 1;
kernel/time/posix-timers.c
386
return current->signal->timer_create_restore_ids;
kernel/time/posix-timers.c
393
struct pid *pid = task_tgid(current);
kernel/time/posix-timers.c
400
if (!rtn || !same_thread_group(rtn, current))
kernel/time/posix-timers.c
472
if (unlikely(current->signal->timer_create_restore_ids)) {
kernel/time/posix-timers.c
514
new_timer->it_pid = get_pid(task_tgid(current));
kernel/time/posix-timers.c
547
guard(spinlock)(&current->sighand->siglock);
kernel/time/posix-timers.c
553
WRITE_ONCE(new_timer->it_signal, current->signal);
kernel/time/posix-timers.c
554
hlist_add_head_rcu(&new_timer->list, &current->signal->posix_timers);
kernel/time/posix-timers.c
643
if (timr->it_signal == current->signal)
kernel/time/posix-timers.c
91
struct signal_struct *sig = current->signal;
kernel/time/sleep_timeout.c
338
while (timeout && !signal_pending(current))
kernel/time/sleep_timeout.c
95
timer.task = current;
kernel/time/tick-sched.c
288
if (is_idle_task(current))
kernel/time/tick-sched.c
396
if (check_tick_dependency(&current->tick_dep_mask))
kernel/time/tick-sched.c
399
if (check_tick_dependency(&current->signal->tick_dep_mask))
kernel/time/tick-sched.c
618
if (atomic_read(&current->tick_dep_mask) ||
kernel/time/tick-sched.c
619
atomic_read(&current->signal->tick_dep_mask))
kernel/time/timer.c
2469
struct task_struct *p = current;
kernel/trace/blktrace.c
275
trace_note(bt, current->pid, BLK_TN_MESSAGE, buf, n, cgid);
kernel/trace/blktrace.c
314
struct task_struct *tsk = current;
kernel/trace/blktrace.c
420
tracing_record_cmdline(current);
kernel/trace/bpf_trace.c
1060
run_ctx = container_of(current->bpf_ctx, struct bpf_trace_run_ctx, run_ctx);
kernel/trace/bpf_trace.c
1062
return ((struct uprobe_dispatch_data *)current->utask->vaddr)->bp_addr;
kernel/trace/bpf_trace.c
1082
return bpf_kprobe_multi_entry_ip(current->bpf_ctx);
kernel/trace/bpf_trace.c
1094
return bpf_kprobe_multi_cookie(current->bpf_ctx);
kernel/trace/bpf_trace.c
1106
return bpf_uprobe_multi_entry_ip(current->bpf_ctx);
kernel/trace/bpf_trace.c
1118
return bpf_uprobe_multi_cookie(current->bpf_ctx);
kernel/trace/bpf_trace.c
1132
run_ctx = container_of(current->bpf_ctx, struct bpf_trace_run_ctx, run_ctx);
kernel/trace/bpf_trace.c
1159
run_ctx = container_of(current->bpf_ctx, struct bpf_trace_run_ctx, run_ctx);
kernel/trace/bpf_trace.c
2524
run_ctx = container_of(current->bpf_ctx, struct bpf_kprobe_multi_run_ctx,
kernel/trace/bpf_trace.c
2542
run_ctx = container_of(current->bpf_ctx, struct bpf_kprobe_multi_run_ctx,
kernel/trace/bpf_trace.c
2985
task_pid_nr_ns(umulti_link->task, task_active_pid_ns(current)) : 0;
kernel/trace/bpf_trace.c
3049
task_pid_nr_ns(umulti_link->task, task_active_pid_ns(current)) : 0;
kernel/trace/bpf_trace.c
3097
if (link->task && !same_thread_group(current, link->task))
kernel/trace/bpf_trace.c
3158
run_ctx = container_of(current->bpf_ctx, struct bpf_uprobe_multi_run_ctx,
kernel/trace/bpf_trace.c
3167
run_ctx = container_of(current->bpf_ctx, struct bpf_uprobe_multi_run_ctx,
kernel/trace/bpf_trace.c
3343
session_ctx = container_of(current->bpf_ctx, struct bpf_session_run_ctx, run_ctx);
kernel/trace/bpf_trace.c
3351
session_ctx = container_of(current->bpf_ctx, struct bpf_session_run_ctx, run_ctx);
kernel/trace/bpf_trace.c
342
current->flags & (PF_KTHREAD | PF_EXITING)))
kernel/trace/bpf_trace.c
757
return (long) current;
kernel/trace/bpf_trace.c
768
return (unsigned long) current;
kernel/trace/bpf_trace.c
824
task = current;
kernel/trace/fgraph.c
1208
pid != current->pid)
kernel/trace/fgraph.c
350
int curr_ret_stack = current->curr_ret_stack;
kernel/trace/fgraph.c
359
val = get_fgraph_entry(current, curr_ret_stack - 1);
kernel/trace/fgraph.c
360
data = &current->ret_stack[curr_ret_stack];
kernel/trace/fgraph.c
369
current->ret_stack[curr_ret_stack - 1] = val;
kernel/trace/fgraph.c
373
current->curr_ret_stack = curr_ret_stack;
kernel/trace/fgraph.c
375
current->ret_stack[curr_ret_stack - 1] = val;
kernel/trace/fgraph.c
413
return ret_stack_get_task_var(current, gops->idx);
kernel/trace/fgraph.c
467
int offset = current->curr_ret_stack;
kernel/trace/fgraph.c
476
ret_stack = get_ret_stack(current, offset, &next_offset);
kernel/trace/fgraph.c
487
val = get_fgraph_entry(current, offset);
kernel/trace/fgraph.c
492
val = get_fgraph_entry(current, offset);
kernel/trace/fgraph.c
498
return get_data_type_data(current, offset);
kernel/trace/fgraph.c
574
if (!current->ret_stack)
kernel/trace/fgraph.c
592
if (current->curr_ret_stack + FGRAPH_FRAME_OFFSET + 1 >= SHADOW_STACK_MAX_OFFSET) {
kernel/trace/fgraph.c
593
atomic_inc(&current->trace_overrun);
kernel/trace/fgraph.c
597
offset = READ_ONCE(current->curr_ret_stack);
kernel/trace/fgraph.c
598
ret_stack = RET_STACK(current, offset);
kernel/trace/fgraph.c
602
current->ret_stack[offset] = val;
kernel/trace/fgraph.c
618
WRITE_ONCE(current->curr_ret_stack, offset + 1);
kernel/trace/fgraph.c
626
current->ret_stack[offset] = val;
kernel/trace/fgraph.c
667
trace.depth = ++current->curr_ret_depth;
kernel/trace/fgraph.c
675
int save_curr_ret_stack = current->curr_ret_stack;
kernel/trace/fgraph.c
681
current->curr_ret_stack = save_curr_ret_stack;
kernel/trace/fgraph.c
693
save_curr_ret_stack = current->curr_ret_stack;
kernel/trace/fgraph.c
699
current->curr_ret_stack = save_curr_ret_stack;
kernel/trace/fgraph.c
710
set_bitmap(current, offset, bitmap | BIT(0));
kernel/trace/fgraph.c
714
current->curr_ret_stack -= FGRAPH_FRAME_OFFSET + 1;
kernel/trace/fgraph.c
716
current->curr_ret_depth--;
kernel/trace/fgraph.c
728
ret_stack = get_ret_stack(current, current->curr_ret_stack, offset);
kernel/trace/fgraph.c
733
current->curr_ret_stack);
kernel/trace/fgraph.c
770
trace->overrun = atomic_read(&current->trace_overrun);
kernel/trace/fgraph.c
771
trace->depth = current->curr_ret_depth;
kernel/trace/fgraph.c
847
bitmap = get_bitmap_bits(current, offset);
kernel/trace/fgraph.c
874
current->curr_ret_stack = offset - FGRAPH_FRAME_OFFSET;
kernel/trace/fgraph.c
876
current->curr_ret_depth--;
kernel/trace/ftrace.c
183
pid != current->pid)
kernel/trace/ftrace.c
850
if (!current->ret_stack)
kernel/trace/ftrace.c
858
profile_data->sleeptime = current->ftrace_sleeptime;
kernel/trace/ftrace.c
8870
if (trace_ignore_this_task(pid_list, no_pid_list, current))
kernel/trace/ftrace.c
8875
current->pid);
kernel/trace/ftrace.c
892
if (current->ftrace_sleeptime)
kernel/trace/ftrace.c
893
calltime -= current->ftrace_sleeptime - profile_data->sleeptime;
kernel/trace/preemptirq_delay_test.c
128
if (set_cpus_allowed_ptr(current, cpu_mask))
kernel/trace/rethook.c
200
__llist_add(&node->llist, &current->rethooks);
kernel/trace/rethook.c
253
if (tsk != current && task_is_running(tsk))
kernel/trace/rethook.c
287
correct_ret_addr = __rethook_find_ret_addr(current, &node);
kernel/trace/rethook.c
305
first = current->rethooks.first;
kernel/trace/rethook.c
324
first = current->rethooks.first;
kernel/trace/rethook.c
325
current->rethooks.first = node->next;
kernel/trace/ring_buffer.c
2247
bool user_thread = current->mm != NULL;
kernel/trace/ring_buffer.c
2308
if (user_thread && fatal_signal_pending(current))
kernel/trace/rv/monitors/nrp/nrp.c
26
da_handle_event(current, irq_entry_nrp);
kernel/trace/rv/monitors/nrp/nrp.c
61
da_handle_event(current, irq_entry_nrp);
kernel/trace/rv/monitors/nrp/nrp.c
79
da_handle_event(current, schedule_entry_preempt_nrp);
kernel/trace/rv/monitors/nrp/nrp.c
81
da_handle_event(current, schedule_entry_nrp);
kernel/trace/rv/monitors/pagefault/pagefault.c
41
ltl_atom_pulse(current, LTL_PAGEFAULT, true);
kernel/trace/rv/monitors/sleep/sleep.c
103
if (current->prio <= task->prio)
kernel/trace/rv/monitors/sleep/sleep.c
113
ltl_atom_update(current, LTL_BLOCK_ON_RT_MUTEX, true);
kernel/trace/rv/monitors/sleep/sleep.c
118
ltl_atom_update(current, LTL_BLOCK_ON_RT_MUTEX, false);
kernel/trace/rv/monitors/sleep/sleep.c
127
mon = ltl_get_monitor(current);
kernel/trace/rv/monitors/sleep/sleep.c
136
syscall_get_arguments(current, regs, args);
kernel/trace/rv/monitors/sleep/sleep.c
140
ltl_atom_update(current, LTL_CLOCK_NANOSLEEP, true);
kernel/trace/rv/monitors/sleep/sleep.c
149
syscall_get_arguments(current, regs, args);
kernel/trace/rv/monitors/sleep/sleep.c
156
ltl_atom_update(current, LTL_FUTEX_LOCK_PI, true);
kernel/trace/rv/monitors/sleep/sleep.c
161
ltl_atom_update(current, LTL_FUTEX_WAIT, true);
kernel/trace/rv/monitors/sleep/sleep.c
170
struct ltl_monitor *mon = ltl_get_monitor(current);
kernel/trace/rv/monitors/sleep/sleep.c
177
ltl_atom_update(current, LTL_CLOCK_NANOSLEEP, false);
kernel/trace/rv/monitors/sssw/sssw.c
59
da_handle_event(current, signal_deliver_sssw);
kernel/trace/trace.c
1575
if (tsk == current)
kernel/trace/trace.c
2175
return current->migration_disabled;
kernel/trace/trace.c
2756
entry->tgid = current->tgid;
kernel/trace/trace.c
7197
update_max_tr_single((struct trace_array *)tr, current, smp_processor_id());
kernel/trace/trace.c
7256
update_max_tr(tr, current, smp_processor_id(), NULL);
kernel/trace/trace.c
842
update_max_tr(tr, current, smp_processor_id(), cond_data);
kernel/trace/trace_branch.c
41
if (current->trace_recursion & TRACE_BRANCH_BIT)
kernel/trace/trace_branch.c
55
current->trace_recursion |= TRACE_BRANCH_BIT;
kernel/trace/trace_branch.c
83
current->trace_recursion &= ~TRACE_BRANCH_BIT;
kernel/trace/trace_events.c
1158
trace_ignore_this_task(pid_list, no_pid_list, current));
kernel/trace/trace_events.c
2476
trace_ignore_this_task(pid_list, no_pid_list, current));
kernel/trace/trace_events_filter.c
996
cmp = pred->regex->match(current->comm, pred->regex,
kernel/trace/trace_events_hist.c
1702
save_comm(elt_data->comm, current);
kernel/trace/trace_events_hist.c
902
return (u64)(unsigned long)current->comm;
kernel/trace/trace_events_user.c
2613
struct user_event_mm *mm = current->user_event_mm;
kernel/trace/trace_events_user.c
745
struct user_event_mm *user_mm = current->user_event_mm;
kernel/trace/trace_events_user.c
750
user_mm = user_event_mm_alloc(current);
kernel/trace/trace_events_user.c
755
user_event_mm_attach(user_mm, current);
kernel/trace/trace_functions.c
195
true_parent_ip = ftrace_graph_ret_addr(current, &idx, parent_ip,
kernel/trace/trace_functions_graph.c
264
ftimes->sleeptime = current->ftrace_sleeptime;
kernel/trace/trace_functions_graph.c
284
unsigned long retaddr = ftrace_graph_top_ret_addr(current);
kernel/trace/trace_functions_graph.c
365
ftimes->calltime += current->ftrace_sleeptime - ftimes->sleeptime;
kernel/trace/trace_hwlat.c
322
if (!cpumask_equal(current_mask, current->cpus_ptr))
kernel/trace/trace_hwlat.c
336
set_cpus_allowed_ptr(current, current_mask);
kernel/trace/trace_irqsoff.c
377
update_max_tr_single(tr, current, cpu);
kernel/trace/trace_osnoise.c
1428
osn_var->pid = current->pid;
kernel/trace/trace_osnoise.c
1649
if (!current->migration_pending)
kernel/trace/trace_osnoise.c
1690
raw_spin_lock_irqsave(&current->pi_lock, flags);
kernel/trace/trace_osnoise.c
1691
current->flags &= ~(PF_NO_SETAFFINITY);
kernel/trace/trace_osnoise.c
1692
raw_spin_unlock_irqrestore(&current->pi_lock, flags);
kernel/trace/trace_osnoise.c
1863
sched_setscheduler_nocheck(current, SCHED_FIFO, &sp);
kernel/trace/trace_osnoise.c
1873
raw_spin_lock_irqsave(&current->pi_lock, flags);
kernel/trace/trace_osnoise.c
1874
current->flags &= ~(PF_NO_SETAFFINITY);
kernel/trace/trace_osnoise.c
1875
raw_spin_unlock_irqrestore(&current->pi_lock, flags);
kernel/trace/trace_osnoise.c
1881
tlat->kthread = current;
kernel/trace/trace_osnoise.c
1882
osn_var->pid = current->pid;
kernel/trace/trace_osnoise.c
2410
if (current->nr_cpus_allowed > 1 || cpu != smp_processor_id()) {
kernel/trace/trace_osnoise.c
2421
get_task_struct(current);
kernel/trace/trace_osnoise.c
2423
osn_var->kthread = current;
kernel/trace/trace_osnoise.c
2424
osn_var->pid = current->pid;
kernel/trace/trace_osnoise.c
2516
tlat->kthread = current;
kernel/trace/trace_preemptirq.c
38
is_idle_task(current)) { \
kernel/trace/trace_probe_tmpl.h
110
*val = (unsigned long)current->comm;
kernel/trace/trace_sched_switch.c
49
tracing_record_taskinfo_sched_switch(current, wakee, flags);
kernel/trace/trace_sched_wakeup.c
547
tracing_record_cmdline(current);
kernel/trace/trace_sched_wakeup.c
559
(!dl_task(p) && (p->prio >= wakeup_prio || p->prio >= current->prio)))
kernel/trace/trace_sched_wakeup.c
596
tracing_sched_wakeup_trace(wakeup_trace, p, current, trace_ctx);
kernel/trace/trace_selftest.c
1433
sched_setattr(current, &attr);
kernel/trace/trace_stack.c
276
if (task_stack_end_corrupted(current)) {
kernel/trace/trace_syscalls.c
1421
syscall_nr = trace_get_syscall_nr(current, regs);
kernel/trace/trace_syscalls.c
1431
syscall_get_arguments(current, regs, args);
kernel/trace/trace_syscalls.c
1553
syscall_nr = trace_get_syscall_nr(current, regs);
kernel/trace/trace_syscalls.c
1577
rec->ret = syscall_get_return_value(current, regs);
kernel/trace/trace_syscalls.c
882
syscall_nr = trace_get_syscall_nr(current, regs);
kernel/trace/trace_syscalls.c
902
syscall_get_arguments(current, regs, args);
kernel/trace/trace_syscalls.c
943
syscall_nr = trace_get_syscall_nr(current, regs);
kernel/trace/trace_syscalls.c
964
entry->ret = syscall_get_return_value(current, regs);
kernel/trace/trace_uprobe.c
1453
if (!uprobe_perf_filter(&tu->consumer, current->mm))
kernel/trace/trace_uprobe.c
1542
current->utask->vaddr = (unsigned long) &udd;
kernel/trace/trace_uprobe.c
1573
current->utask->vaddr = (unsigned long) &udd;
kernel/trace/trace_uprobe.c
159
ret = strscpy(dst, current->comm, maxlen);
kernel/trace/trace_uprobe.c
193
len = strlen(current->comm) + 1;
kernel/trace/trace_uprobe.c
211
udd = (void *) current->utask->vaddr;
kernel/umh.c
69
spin_lock_irq(&current->sighand->siglock);
kernel/umh.c
70
flush_signal_handlers(current, 1);
kernel/umh.c
71
spin_unlock_irq(&current->sighand->siglock);
kernel/umh.c
79
current->fs->umask = 0022;
kernel/umh.c
85
set_user_nice(current, 0);
kernel/umh.c
88
new = prepare_kernel_cred(current);
kernel/unwind/deferred.c
113
struct unwind_task_info *info = &current->unwind_info;
kernel/unwind/deferred.c
119
if (!current->mm)
kernel/unwind/deferred.c
189
process_unwind_deferred(current);
kernel/unwind/deferred.c
194
struct unwind_task_info *info = &current->unwind_info;
kernel/unwind/deferred.c
231
struct unwind_task_info *info = &current->unwind_info;
kernel/unwind/deferred.c
239
if ((current->flags & (PF_KTHREAD | PF_EXITING)) ||
kernel/unwind/deferred.c
240
!user_mode(task_pt_regs(current)))
kernel/unwind/deferred.c
290
ret = task_work_add(current, &info->work, twa_mode);
kernel/unwind/user.c
117
struct pt_regs *regs = task_pt_regs(current);
kernel/unwind/user.c
121
if ((current->flags & PF_KTHREAD) || !user_mode(regs)) {
kernel/unwind/user.c
151
if (current->flags & PF_KTHREAD)
kernel/unwind/user.c
70
struct pt_regs *regs = task_pt_regs(current);
kernel/user-return-notifier.c
17
set_tsk_thread_flag(current, TIF_USER_RETURN_NOTIFY);
kernel/user-return-notifier.c
30
clear_tsk_thread_flag(current, TIF_USER_RETURN_NOTIFY);
kernel/user_namespace.c
1355
if (!thread_group_empty(current))
kernel/user_namespace.c
1358
if (current->fs->users != 1)
kernel/utsname_sysctl.c
23
uts_ns = current->nsproxy->uts_ns;
kernel/vhost_task.c
34
if (signal_pending(current)) {
kernel/watch_queue.c
474
if (atomic_inc_return(&cred->user->nr_watches) > task_rlimit(current, RLIMIT_NOFILE)) {
kernel/watchdog.c
246
print_irqtrace_events(current);
kernel/watchdog.c
864
current->comm, task_pid_nr(current));
kernel/watchdog.c
867
print_irqtrace_events(current);
kernel/workqueue.c
3248
lockdep_start_depth = lockdep_depth(current);
kernel/workqueue.c
3288
lockdep_depth(current) != lockdep_start_depth ||
kernel/workqueue.c
3292
current->comm, task_pid_nr(current), preempt_count(),
kernel/workqueue.c
3293
lockdep_start_depth, lockdep_depth(current),
kernel/workqueue.c
3296
debug_show_held_locks(current);
kernel/workqueue.c
3367
current->flags |= PF_WQ_WORKER;
kernel/workqueue.c
3369
current->flags &= ~PF_WQ_WORKER;
kernel/workqueue.c
3542
set_user_nice(current, RESCUER_NICE_LEVEL);
kernel/workqueue.c
3798
WARN_ONCE(current->flags & PF_MEMALLOC,
kernel/workqueue.c
3800
current->pid, current->comm, target_wq->name, target_func);
kernel/workqueue.c
3869
barr->task = current;
kernel/workqueue_internal.h
70
if (in_task() && (current->flags & PF_WQ_WORKER))
kernel/workqueue_internal.h
71
return kthread_data(current);
lib/alloc_tag.c
786
if (!current->mm)
lib/closure.c
137
struct closure_syncer s = { .task = current };
lib/closure.c
162
struct closure_syncer s = { .task = current };
lib/closure.c
190
struct closure_syncer s = { .task = current };
lib/debugobjects.c
667
task_stack_page(current));
lib/debugobjects.c
670
task_stack_page(current));
lib/dump_stack.c
60
current->pid, current->comm,
lib/dump_stack.c
74
print_worker_info(log_lvl, current);
lib/dump_stack.c
75
print_stop_info(log_lvl, current);
lib/dump_stack.c
76
print_scx_info(log_lvl, current);
lib/dynamic_debug.c
835
task_pid_vnr(current));
lib/fault-inject.c
129
unsigned int fail_nth = READ_ONCE(current->fail_nth);
lib/fault-inject.c
137
WRITE_ONCE(current->fail_nth, fail_nth);
lib/fault-inject.c
149
if (attr->task_filter && !fail_task(attr, current))
lib/is_single_threaded.c
17
struct task_struct *task = current;
lib/klist.c
243
waiter.process = current;
lib/kunit/kunit-test.c
512
kunit_log(KERN_INFO, current->kunit_test, "action_order_1");
lib/kunit/kunit-test.c
519
kunit_log(KERN_INFO, current->kunit_test, "action_order_2");
lib/kunit/kunit-test.c
711
KUNIT_EXPECT_PTR_EQ(test, test, current->kunit_test);
lib/kunit/kunit-test.c
723
current->kunit_test = &fake;
lib/kunit/kunit-test.c
730
current->kunit_test = test;
lib/kunit/string-stream-test.c
148
current->kunit_test = fake_test;
lib/kunit/string-stream-test.c
154
current->kunit_test = test;
lib/kunit/test.c
1054
current->kunit_test = NULL;
lib/kunit/test.c
39
if (!current->kunit_test)
lib/kunit/test.c
42
kunit_set_failure(current->kunit_test);
lib/kunit/test.c
486
current->kunit_test = test;
lib/kunit/test.c
49
buffer = kunit_kmalloc(current->kunit_test, len, GFP_KERNEL);
lib/kunit/test.c
502
current->kunit_test = test;
lib/kunit/test.c
57
kunit_err(current->kunit_test, "%s:%d: %s", file, line, buffer);
lib/kunit/test.c
58
kunit_kfree(current->kunit_test, buffer);
lib/kunit/user_alloc.c
29
if (current->mm)
lib/kunit/user_alloc.c
44
arch_pick_mmap_layout(mm, &current->signal->rlim[RLIMIT_STACK]);
lib/locking-selftest.c
1435
int saved_mgd_count = current->migration_disabled;
lib/locking-selftest.c
1437
int saved_rcu_count = current->rcu_read_lock_nesting;
lib/locking-selftest.c
1475
while (current->migration_disabled > saved_mgd_count)
lib/locking-selftest.c
1479
while (current->rcu_read_lock_nesting > saved_rcu_count)
lib/locking-selftest.c
1481
WARN_ON_ONCE(current->rcu_read_lock_nesting < saved_rcu_count);
lib/locking-selftest.c
1486
current->softirqs_enabled = 0;
lib/locking-selftest.c
1488
current->softirqs_enabled = 1;
lib/locking-selftest.c
2856
lockdep_set_selftest_task(current);
lib/nlattr.c
215
current->comm, pt->type);
lib/nlattr.c
415
current->comm, type);
lib/nlattr.c
648
rem, current->comm);
lib/smp_processor_id.c
25
if (current->migration_disabled)
lib/smp_processor_id.c
44
what1, what2, preempt_count() - 1, current->comm, current->pid);
lib/syscall.c
74
if (target == current)
lib/test_hmm.c
177
ret = mmu_interval_notifier_insert(&dmirror->notifier, current->mm,
lib/test_lockup.c
364
current->in_iowait = 1;
lib/test_lockup.c
369
current->in_iowait = 0;
lib/test_lockup.c
458
main_task = current;
lib/tests/test_ratelimit.c
86
set_user_nice(current, MAX_NICE);
lib/ubsan.c
133
return current->in_ubsan || was_reported(loc);
lib/ubsan.c
221
current->in_ubsan++;
lib/ubsan.c
236
current->in_ubsan--;
lib/vsprintf.c
892
if (!has_capability_noaudit(current, CAP_SYSLOG) ||
lib/zstd/compress/zstd_double_fast.c
160
U32 const current = (U32)(ip - base);
lib/zstd/compress/zstd_double_fast.c
161
U32 const windowLow = ZSTD_getLowestPrefixIndex(ms, current, cParams->windowLog);
lib/zstd/compress/zstd_double_fast.c
162
U32 const maxRep = current - windowLow;
mm/compaction.c
2339
if (cc->contended || fatal_signal_pending(current))
mm/compaction.c
2780
WRITE_ONCE(current->capture_control, &capc);
mm/compaction.c
2789
WRITE_ONCE(current->capture_control, NULL);
mm/compaction.c
2875
|| fatal_signal_pending(current))
mm/compaction.c
2911
if (fatal_signal_pending(current))
mm/compaction.c
3171
current->flags |= PF_KCOMPACTD;
mm/compaction.c
3228
current->flags &= ~PF_KCOMPACTD;
mm/compaction.c
3281
table->procname, current->comm,
mm/compaction.c
3282
task_pid_nr(current));
mm/compaction.c
541
if (fatal_signal_pending(current)) {
mm/compaction.c
872
if (fatal_signal_pending(current))
mm/compaction.c
920
if (fatal_signal_pending(current)) {
mm/damon/core.c
2752
pr_debug("kdamond (%d) starts\n", current->pid);
mm/damon/core.c
2861
pr_debug("kdamond (%d) finishes\n", current->pid);
mm/filemap.c
1320
if (signal_pending_state(state, current))
mm/filemap.c
1447
if (signal_pending_state(TASK_UNINTERRUPTIBLE, current))
mm/filemap.c
158
current->comm, folio_pfn(folio));
mm/filemap.c
2682
if (fatal_signal_pending(current))
mm/filemap.c
4224
pr_crit("File: %s PID: %d Comm: %.20s\n", path, current->pid,
mm/filemap.c
4225
current->comm);
mm/filemap.c
4319
if (fatal_signal_pending(current)) {
mm/gup.c
1293
current->comm, task_pid_nr(current),
mm/gup.c
1420
if (fatal_signal_pending(current)) {
mm/gup.c
1585
fatal_signal_pending(current))
mm/gup.c
1627
if (fatal_signal_pending(current))
mm/gup.c
1633
return signal_pending(current);
mm/gup.c
1927
struct mm_struct *mm = current->mm;
mm/gup.c
2120
struct mm_struct *mm = current->mm;
mm/gup.c
2192
ret = __get_user_pages_locked(current->mm, addr, 1, &page, locked,
mm/gup.c
2652
return __get_user_pages_locked(current->mm, start, nr_pages, pages,
mm/gup.c
2681
return __get_user_pages_locked(current->mm, start, nr_pages, pages,
mm/gup.c
3098
pgdp = pgd_offset(current->mm, addr);
mm/gup.c
3141
if (!raw_seqcount_try_begin(&current->mm->write_protect_seq, seq))
mm/gup.c
3165
if (read_seqcount_retry(&current->mm->write_protect_seq, seq)) {
mm/gup.c
3190
mm_set_has_pinned_flag(current->mm);
mm/gup.c
3193
might_lock_read(&current->mm->mmap_lock);
mm/gup.c
3209
ret = __gup_longterm_locked(current->mm, start, nr_pages - nr_pinned,
mm/gup.c
3383
return __gup_longterm_locked(current->mm, start, nr_pages,
mm/gup.c
3405
return __gup_longterm_locked(current->mm, start, nr_pages, pages,
mm/gup_test.c
120
if (needs_mmap_lock && mmap_read_lock_killable(current->mm)) {
mm/gup_test.c
201
mmap_read_unlock(current->mm);
mm/gup_test.c
257
if (!fast && mmap_read_lock_killable(current->mm)) {
mm/gup_test.c
285
mmap_read_unlock(current->mm);
mm/highmem.c
168
struct kmap_ctrl *kctrl = &current->kmap_ctrl;
mm/highmem.c
271
DECLARE_WAITQUEUE(wait, current);
mm/highmem.c
479
current->kmap_ctrl.idx += KM_INCR;
mm/highmem.c
480
BUG_ON(current->kmap_ctrl.idx >= KM_MAX_IDX);
mm/highmem.c
481
return current->kmap_ctrl.idx - 1;
mm/highmem.c
486
return current->kmap_ctrl.idx - 1;
mm/highmem.c
491
current->kmap_ctrl.idx -= KM_INCR;
mm/highmem.c
492
BUG_ON(current->kmap_ctrl.idx < 0);
mm/highmem.c
569
current->kmap_ctrl.pteval[kmap_local_idx()] = pteval;
mm/highmem.c
629
current->kmap_ctrl.pteval[kmap_local_idx()] = __pte(0);
mm/highmem.c
648
struct task_struct *tsk = current;
mm/highmem.c
685
struct task_struct *tsk = current;
mm/huge_memory.c
1227
if (mm_flags_test(MMF_TOPDOWN, current->mm) && !off_sub)
mm/hugetlb.c
1304
bool pin = !!(current->flags & PF_MEMALLOC_PIN);
mm/hugetlb.c
2280
struct mempolicy *mpol = get_task_policy(current);
mm/hugetlb.c
3888
if (signal_pending(current)) {
mm/hugetlb.c
5744
current->pid);
mm/hugetlb_cgroup.c
265
h_cg = hugetlb_cgroup_from_task(current);
mm/kasan/common.c
70
track->pid = current->pid;
mm/kasan/common.c
85
current->kasan_depth++;
mm/kasan/common.c
91
current->kasan_depth--;
mm/kasan/report.c
193
test = current->kunit_test;
mm/kasan/report.c
251
info->access_addr, current->comm, task_pid_nr(current));
mm/kasan/report.c
258
info->access_addr, current->comm, task_pid_nr(current));
mm/kasan/report.c
262
info->access_addr, current->comm, task_pid_nr(current));
mm/kasan/report.c
94
if (current->kasan_depth)
mm/kasan/report_generic.c
309
shadow_bottom = kasan_mem_to_shadow(end_of_stack(current));
mm/kasan/report_generic.c
348
current->comm, task_pid_nr(current));
mm/kasan/report_sw_tags.c
93
current->comm, task_pid_nr(current));
mm/kfence/core.c
316
track->pid = task_pid_nr(current);
mm/khugepaged.c
2617
set_user_nice(current, MAX_NICE);
mm/khugepaged.c
2722
return kthread_func(current) == khugepaged;
mm/kmemleak.c
1508
if (current->mm)
mm/kmemleak.c
1509
return signal_pending(current);
mm/kmemleak.c
1885
set_user_nice(current, 10);
mm/kmemleak.c
690
object->pid = current->pid;
mm/kmemleak.c
697
strscpy(object->comm, current->comm);
mm/kmsan/hooks.c
432
KMSAN_WARN_ON(current->kmsan_ctx.depth == 0);
mm/kmsan/hooks.c
433
current->kmsan_ctx.depth--;
mm/kmsan/hooks.c
439
current->kmsan_ctx.depth++;
mm/kmsan/hooks.c
440
KMSAN_WARN_ON(current->kmsan_ctx.depth == 0);
mm/kmsan/init.c
232
kmsan_internal_task_create(current);
mm/kmsan/kmsan.h
83
return in_task() ? &current->kmsan_ctx : raw_cpu_ptr(&kmsan_percpu_ctx);
mm/kmsan/report.c
162
if (current->kmsan_ctx.depth)
mm/ksm.c
2785
while (scan_npages-- && likely(!freezing(current))) {
mm/ksm.c
2806
set_user_nice(current, 5);
mm/ksm.c
421
cpu_time = task_sched_runtime(current);
mm/ksm.c
621
if (signal_pending(current))
mm/madvise.c
1604
if (current->mm != vma->vm_mm)
mm/madvise.c
1911
return current->mm == mm ? untagged_addr(start) :
mm/madvise.c
2037
return do_madvise(current->mm, start, len_in, behavior);
mm/madvise.c
2080
if (fatal_signal_pending(current)) {
mm/madvise.c
2144
if (mm != current->mm && !process_madvise_remote_valid(behavior)) {
mm/madvise.c
2154
if (mm != current->mm && !capable(CAP_SYS_NICE)) {
mm/madvise.c
2226
struct mm_struct *mm = current->mm;
mm/madvise.c
369
if (fatal_signal_pending(current))
mm/memcontrol-v1.c
1375
struct mem_cgroup *memcg = current->memcg_in_oom;
mm/memcontrol-v1.c
1389
owait.wait.private = current;
mm/memcontrol-v1.c
1407
current->memcg_in_oom = NULL;
mm/memcontrol-v1.c
1434
if (current->in_user_fault) {
mm/memcontrol-v1.c
1436
current->memcg_in_oom = memcg;
mm/memcontrol-v1.c
1471
if (signal_pending(current)) {
mm/memcontrol-v1.c
1531
if (signal_pending(current))
mm/memcontrol.c
110
return tsk_is_oom_victim(current) || fatal_signal_pending(current) ||
mm/memcontrol.c
111
(current->flags & PF_EXITING);
mm/memcontrol.c
2270
unsigned int nr_pages = current->memcg_nr_pages_over_high;
mm/memcontrol.c
2275
memcg = get_mem_cgroup_from_mm(current->mm);
mm/memcontrol.c
2276
current->memcg_nr_pages_over_high = 0;
mm/memcontrol.c
2401
if (unlikely(current->flags & PF_MEMALLOC))
mm/memcontrol.c
2404
if (unlikely(task_in_memcg_oom(current)))
mm/memcontrol.c
2530
current->memcg_nr_pages_over_high += batch;
mm/memcontrol.c
2531
set_notify_resume(current);
mm/memcontrol.c
2543
if (current->memcg_nr_pages_over_high > MEMCG_CHARGE_BATCH &&
mm/memcontrol.c
2544
!(current->flags & PF_MEMALLOC) &&
mm/memcontrol.c
2689
old = xchg(&current->objcg, NULL);
mm/memcontrol.c
2699
if (!current->mm || (current->flags & PF_KTHREAD))
mm/memcontrol.c
2719
memcg = mem_cgroup_from_task(current);
mm/memcontrol.c
2728
} while (!try_cmpxchg(&current->objcg, &old, objcg));
mm/memcontrol.c
2742
memcg = current->active_memcg;
mm/memcontrol.c
2746
objcg = READ_ONCE(current->objcg);
mm/memcontrol.c
4416
if (signal_pending(current))
mm/memcontrol.c
4467
if (signal_pending(current))
mm/memcontrol.c
5052
memcg = mem_cgroup_from_task(current);
mm/memcontrol.c
5506
VM_WARN_ON_ONCE(!(current->flags & PF_MEMALLOC));
mm/memcontrol.c
912
return current->active_memcg;
mm/memcontrol.c
949
mm = current->mm;
mm/memcontrol.c
977
memcg = mem_cgroup_from_task(current);
mm/memfd.c
347
struct pid_namespace *ns = task_active_pid_ns(current);
mm/memfd.c
360
current->comm, task_pid_nr(current), sysctl);
mm/memory-failure.c
2060
rv = kill_accessing_process(current, pfn, flags);
mm/memory-failure.c
2396
res = kill_accessing_process(current, pfn, flags);
mm/memory-failure.c
295
if ((flags & MF_ACTION_REQUIRED) && (t == current))
mm/memory-failure.c
536
if (force_early && tsk->mm == current->mm)
mm/memory-failure.c
537
return current;
mm/memory.c
612
pr_alert("BUG: Bad page map in process %s %s:%08llx", current->comm,
mm/memory.c
6508
current->maj_flt++;
mm/memory.c
6510
current->min_flt++;
mm/memory.c
6530
current->in_lru_fault = vma_has_recency(vma);
mm/memory.c
6535
current->in_lru_fault = false;
mm/memory.c
6646
if (task_in_memcg_oom(current) && !(ret & VM_FAULT_OOM))
mm/memory.c
7202
struct mm_struct *mm = current->mm;
mm/memory.c
7229
if (current->mm)
mm/memory.c
7230
might_lock_read(&current->mm->mmap_lock);
mm/memory_hotplug.c
2005
if (signal_pending(current)) {
mm/mempolicy.c
1083
task_lock(current);
mm/mempolicy.c
1086
task_unlock(current);
mm/mempolicy.c
1091
old = current->mempolicy;
mm/mempolicy.c
1092
current->mempolicy = new;
mm/mempolicy.c
1095
current->il_prev = MAX_NUMNODES-1;
mm/mempolicy.c
1096
current->il_weight = 0;
mm/mempolicy.c
1098
task_unlock(current);
mm/mempolicy.c
1151
struct mm_struct *mm = current->mm;
mm/mempolicy.c
1153
struct mempolicy *pol = current->mempolicy, *pol_refcount = NULL;
mm/mempolicy.c
1163
task_lock(current);
mm/mempolicy.c
1165
task_unlock(current);
mm/mempolicy.c
1204
} else if (pol == current->mempolicy &&
mm/mempolicy.c
1206
*policy = next_node_in(current->il_prev, pol->nodes);
mm/mempolicy.c
1207
} else if (pol == current->mempolicy &&
mm/mempolicy.c
1209
if (current->il_weight)
mm/mempolicy.c
1210
*policy = current->il_prev;
mm/mempolicy.c
1212
*policy = next_node_in(current->il_prev,
mm/mempolicy.c
1233
task_lock(current);
mm/mempolicy.c
1235
task_unlock(current);
mm/mempolicy.c
1490
struct mm_struct *mm = current->mm;
mm/mempolicy.c
1569
new = get_task_policy(current);
mm/mempolicy.c
1763
struct mm_struct *mm = current->mm;
mm/mempolicy.c
1888
task = pid ? find_task_by_vpid(pid) : current;
mm/mempolicy.c
1916
task_nodes = cpuset_mems_allowed(current);
mm/mempolicy.c
2048
pol = get_task_policy(current);
mm/mempolicy.c
2075
pol = get_task_policy(current);
mm/mempolicy.c
2108
node = current->il_prev;
mm/mempolicy.c
2109
if (!current->il_weight || !node_isset(node, policy->nodes)) {
mm/mempolicy.c
2115
current->il_prev = node;
mm/mempolicy.c
2116
current->il_weight = get_il_weight(node);
mm/mempolicy.c
2118
current->il_weight--;
mm/mempolicy.c
2131
nid = next_node_in(current->il_prev, policy->nodes);
mm/mempolicy.c
2135
current->il_prev = nid;
mm/mempolicy.c
2151
policy = current->mempolicy;
mm/mempolicy.c
2359
if (!(mask && current->mempolicy))
mm/mempolicy.c
2362
task_lock(current);
mm/mempolicy.c
2363
mempolicy = current->mempolicy;
mm/mempolicy.c
2380
task_unlock(current);
mm/mempolicy.c
2559
pol = get_task_policy(current);
mm/mempolicy.c
2635
struct task_struct *me = current;
mm/mempolicy.c
2782
pol = get_task_policy(current);
mm/mempolicy.c
2832
if (old == current->mempolicy) {
mm/mempolicy.c
2833
task_lock(current);
mm/mempolicy.c
2835
task_unlock(current);
mm/mempolicy.c
2840
nodemask_t mems = cpuset_mems_allowed(current);
mm/mempolicy.c
3068
if (!should_numa_migrate_memory(current, folio, curnid,
mm/mempolicy.c
3230
task_lock(current);
mm/mempolicy.c
3232
task_unlock(current);
mm/migrate.c
1239
if (current->flags & PF_MEMALLOC)
mm/migrate.c
2536
mmget(current->mm);
mm/migrate.c
2537
*mem_nodes = cpuset_mems_allowed(current);
mm/migrate.c
2538
return current->mm;
mm/mincore.c
253
vma = vma_lookup(current->mm, addr);
mm/mincore.c
326
mmap_read_lock(current->mm);
mm/mincore.c
328
mmap_read_unlock(current->mm);
mm/mlock.c
476
is_vm_hugetlb_page(vma) || vma == get_gate_vma(current->mm) ||
mm/mlock.c
519
VMA_ITERATOR(vmi, current->mm, start);
mm/mlock.c
630
if (mmap_write_lock_killable(current->mm))
mm/mlock.c
633
locked += current->mm->locked_vm;
mm/mlock.c
641
locked -= count_mm_mlocked_page_nr(current->mm,
mm/mlock.c
649
mmap_write_unlock(current->mm);
mm/mlock.c
686
if (mmap_write_lock_killable(current->mm))
mm/mlock.c
689
mmap_write_unlock(current->mm);
mm/mlock.c
706
VMA_ITERATOR(vmi, current->mm, 0);
mm/mlock.c
710
current->mm->def_flags &= ~VM_LOCKED_MASK;
mm/mlock.c
712
current->mm->def_flags |= VM_LOCKED;
mm/mlock.c
715
current->mm->def_flags |= VM_LOCKONFAULT;
mm/mlock.c
760
if (mmap_write_lock_killable(current->mm))
mm/mlock.c
764
if (!(flags & MCL_CURRENT) || (current->mm->total_vm <= lock_limit) ||
mm/mlock.c
767
mmap_write_unlock(current->mm);
mm/mlock.c
778
if (mmap_write_lock_killable(current->mm))
mm/mlock.c
781
mmap_write_unlock(current->mm);
mm/mm_init.c
2123
set_cpus_allowed_ptr(current, cpumask);
mm/mmap.c
1089
struct mm_struct *mm = current->mm;
mm/mmap.c
1097
current->comm, current->pid);
mm/mmap.c
111
return mlock_future_ok(current->mm,
mm/mmap.c
112
current->mm->def_flags & VM_LOCKED, len)
mm/mmap.c
119
struct mm_struct *mm = current->mm;
mm/mmap.c
1206
struct mm_struct *mm = current->mm;
mm/mmap.c
1348
current->comm, current->pid,
mm/mmap.c
138
if (!current->brk_randomized)
mm/mmap.c
1396
if (WARN_ON_ONCE(current->mm != new_vma->vm_mm))
mm/mmap.c
341
struct mm_struct *mm = current->mm;
mm/mmap.c
357
if ((prot & PROT_READ) && (current->personality & READ_IMPLIES_EXEC))
mm/mmap.c
693
struct mm_struct *mm = current->mm;
mm/mmap.c
743
struct mm_struct *mm = current->mm;
mm/mmap.c
805
if (mm_flags_test(MMF_TOPDOWN, current->mm))
mm/mprotect.c
724
error = walk_page_range(current->mm, start, end,
mm/mprotect.c
808
const bool rier = (current->personality & READ_IMPLIES_EXEC) &&
mm/mprotect.c
832
if (mmap_write_lock_killable(current->mm))
mm/mprotect.c
840
if ((pkey != -1) && !mm_pkey_is_allocated(current->mm, pkey))
mm/mprotect.c
843
vma_iter_init(&vmi, current->mm, start);
mm/mprotect.c
871
tlb_gather_mmu(&tlb, current->mm);
mm/mprotect.c
944
mmap_write_unlock(current->mm);
mm/mprotect.c
974
mmap_write_lock(current->mm);
mm/mprotect.c
975
pkey = mm_pkey_alloc(current->mm);
mm/mprotect.c
981
ret = arch_set_user_pkey_access(current, pkey, init_val);
mm/mprotect.c
983
mm_pkey_free(current->mm, pkey);
mm/mprotect.c
988
mmap_write_unlock(current->mm);
mm/mprotect.c
996
mmap_write_lock(current->mm);
mm/mprotect.c
997
ret = mm_pkey_free(current->mm, pkey);
mm/mprotect.c
998
mmap_write_unlock(current->mm);
mm/mremap.c
1023
struct mm_struct *mm = current->mm;
mm/mremap.c
1047
if (current->mm->map_count >= sysctl_max_map_count - 3)
mm/mremap.c
1084
struct mm_struct *mm = current->mm;
mm/mremap.c
1272
struct mm_struct *mm = current->mm;
mm/mremap.c
1333
struct mm_struct *mm = current->mm;
mm/mremap.c
1369
struct mm_struct *mm = current->mm;
mm/mremap.c
1458
struct mm_struct *mm = current->mm;
mm/mremap.c
1615
struct mm_struct *mm = current->mm;
mm/mremap.c
1658
struct mm_struct *mm = current->mm;
mm/mremap.c
1696
current->comm, current->pid);
mm/mremap.c
1820
if ((current->mm->map_count + 2) >= sysctl_max_map_count - 3)
mm/mremap.c
1837
VMA_ITERATOR(vmi, current->mm, start);
mm/mremap.c
1917
struct mm_struct *mm = current->mm;
mm/mremap.c
1935
vrm->vma = vma_lookup(current->mm, vrm->addr);
mm/mremap.c
994
if (security_vm_enough_memory_mm(current->mm, charged))
mm/mseal.c
144
struct mm_struct *mm = current->mm;
mm/mseal.c
43
VMA_ITERATOR(vmi, current->mm, start);
mm/msync.c
35
struct mm_struct *mm = current->mm;
mm/nommu.c
1004
len, current->pid, current->comm);
mm/nommu.c
1027
VMA_ITERATOR(vmi, current->mm, 0);
mm/nommu.c
1052
vma = vm_area_alloc(current->mm);
mm/nommu.c
1193
current->mm->total_vm += len >> PAGE_SHIFT;
mm/nommu.c
1201
setup_vma_to_mm(vma, current->mm);
mm/nommu.c
1202
current->mm->map_count++;
mm/nommu.c
1238
len, current->pid);
mm/nommu.c
1244
len, current->pid);
mm/nommu.c
1347
current->pid);
mm/nommu.c
1445
current->pid, current->comm,
mm/nommu.c
1490
struct mm_struct *mm = current->mm;
mm/nommu.c
150
mmap_write_lock(current->mm);
mm/nommu.c
151
vma = find_vma(current->mm, (unsigned long)ret);
mm/nommu.c
154
mmap_write_unlock(current->mm);
mm/nommu.c
1560
vma = find_vma_exact(current->mm, addr, old_len);
mm/nommu.c
1584
mmap_write_lock(current->mm);
mm/nommu.c
1586
mmap_write_unlock(current->mm);
mm/nommu.c
382
struct mm_struct *mm = current->mm;
mm/nommu.c
604
current->pid);
mm/nommu.c
811
if (current->personality & READ_IMPLIES_EXEC) {
mm/nommu.c
830
(current->personality & READ_IMPLIES_EXEC))
mm/nommu.c
870
if (!(prot & PROT_WRITE) && !current->ptrace)
mm/nommu.c
94
vma = find_vma(current->mm, (unsigned long)objp);
mm/oom_kill.c
112
ret = cpuset_mems_allowed_intersects(current, tsk);
mm/oom_kill.c
1138
if (task_will_free_mem(current)) {
mm/oom_kill.c
1139
mark_oom_victim(current);
mm/oom_kill.c
1140
queue_oom_reaper(current);
mm/oom_kill.c
1162
current->mm && !oom_unkillable_task(current) &&
mm/oom_kill.c
1163
oom_cpuset_eligible(current, oc) &&
mm/oom_kill.c
1164
current->signal->oom_score_adj != OOM_SCORE_ADJ_MIN) {
mm/oom_kill.c
1165
get_task_struct(current);
mm/oom_kill.c
1166
oc->chosen = current;
mm/oom_kill.c
1204
if (fatal_signal_pending(current))
mm/oom_kill.c
462
current->comm, oc->gfp_mask, &oc->gfp_mask, oc->order,
mm/oom_kill.c
463
current->signal->oom_score_adj);
mm/page-writeback.c
1766
if (!(current->flags & PF_LOCAL_THROTTLE))
mm/page-writeback.c
1853
current->dirty_paused_when = now;
mm/page-writeback.c
1854
current->nr_dirtied = 0;
mm/page-writeback.c
1857
current->nr_dirtied_pause = min(intv, m_intv);
mm/page-writeback.c
1929
if (current->dirty_paused_when)
mm/page-writeback.c
1930
pause -= now - current->dirty_paused_when;
mm/page-writeback.c
1948
current->dirty_paused_when = now;
mm/page-writeback.c
1949
current->nr_dirtied = 0;
mm/page-writeback.c
1951
current->dirty_paused_when += period;
mm/page-writeback.c
1952
current->nr_dirtied = 0;
mm/page-writeback.c
1953
} else if (current->nr_dirtied_pause <= pages_dirtied)
mm/page-writeback.c
1954
current->nr_dirtied_pause += pages_dirtied;
mm/page-writeback.c
1980
current->dirty_paused_when = now + pause;
mm/page-writeback.c
1981
current->nr_dirtied = 0;
mm/page-writeback.c
1982
current->nr_dirtied_pause = nr_dirtied_pause;
mm/page-writeback.c
2004
if (fatal_signal_pending(current))
mm/page-writeback.c
2063
ratelimit = current->nr_dirtied_pause;
mm/page-writeback.c
2075
if (unlikely(current->nr_dirtied >= ratelimit))
mm/page-writeback.c
2087
if (*p > 0 && current->nr_dirtied < ratelimit) {
mm/page-writeback.c
2089
nr_pages_dirtied = min(*p, ratelimit - current->nr_dirtied);
mm/page-writeback.c
2091
current->nr_dirtied += nr_pages_dirtied;
mm/page-writeback.c
2095
if (unlikely(current->nr_dirtied >= ratelimit))
mm/page-writeback.c
2096
ret = balance_dirty_pages(wb, current->nr_dirtied, flags);
mm/page-writeback.c
2238
current->comm);
mm/page-writeback.c
2638
current->nr_dirtied += nr;
mm/page-writeback.c
2756
current->nr_dirtied -= nr;
mm/page-writeback.c
390
tsk = current;
mm/page-writeback.c
441
struct task_struct *tsk = current;
mm/page_alloc.c
1891
pgalloc_tag_add(page, current, 1 << order);
mm/page_alloc.c
4015
if (tsk_is_oom_victim(current) ||
mm/page_alloc.c
4016
(current->flags & (PF_MEMALLOC | PF_EXITING)))
mm/page_alloc.c
4040
current->comm, &vaf, gfp_mask, &gfp_mask,
mm/page_alloc.c
4108
if (current->flags & PF_DUMPCORE)
mm/page_alloc.c
4238
if (fatal_signal_pending(current))
mm/page_alloc.c
4337
if (current->flags & PF_MEMALLOC)
mm/page_alloc.c
4536
} else if (unlikely(rt_or_dl_task(current)) && in_task())
mm/page_alloc.c
4572
if (in_serving_softirq() && (current->flags & PF_MEMALLOC))
mm/page_alloc.c
4575
if (current->flags & PF_MEMALLOC)
mm/page_alloc.c
4577
else if (oom_reserves_allowed(current))
mm/page_alloc.c
4664
if (current->flags & PF_WQ_WORKER)
mm/page_alloc.c
4741
WARN_ON_ONCE(current->flags & PF_MEMALLOC);
mm/page_alloc.c
4841
if (current->flags & PF_MEMALLOC)
mm/page_alloc.c
4942
if (tsk_is_oom_victim(current) &&
mm/page_alloc.c
672
current->comm, page_to_pfn(page));
mm/page_alloc.c
6860
if (fatal_signal_pending(current)) {
mm/page_alloc.c
761
struct capture_control *capc = current->capture_control;
mm/page_alloc.c
764
!(current->flags & PF_KTHREAD) &&
mm/page_io.c
585
get_task_struct(current);
mm/page_io.c
591
put_task_struct(current);
mm/page_owner.c
161
if (current->in_page_owner)
mm/page_owner.c
292
page_owner->free_pid = current->pid;
mm/page_owner.c
293
page_owner->free_tgid = current->tgid;
mm/page_owner.c
321
__update_page_owner_free_handle(page, handle, order, current->pid,
mm/page_owner.c
322
current->tgid, free_ts_nsec);
mm/page_owner.c
343
ts_nsec, current->pid, current->tgid,
mm/page_owner.c
344
current->comm);
mm/page_owner.c
74
current->in_page_owner = 1;
mm/page_owner.c
79
current->in_page_owner = 0;
mm/page_owner.c
829
-1, local_clock(), current->pid,
mm/page_owner.c
830
current->tgid, current->comm);
mm/percpu.c
1701
current->alloc_tag, size);
mm/rmap.c
713
struct tlbflush_unmap_batch *tlb_ubc = &current->tlb_ubc;
mm/rmap.c
726
struct tlbflush_unmap_batch *tlb_ubc = &current->tlb_ubc;
mm/rmap.c
745
struct tlbflush_unmap_batch *tlb_ubc = &current->tlb_ubc;
mm/shmem.c
178
0 : security_vm_enough_memory_mm(current->mm, VM_ACCT(size));
mm/shmem.c
192
return security_vm_enough_memory_mm(current->mm,
mm/shmem.c
211
return security_vm_enough_memory_mm(current->mm,
mm/shmem.c
2920
return mpol ? mpol : get_task_policy(current);
mm/shmem.c
3785
if (fatal_signal_pending(current))
mm/shrinker_debug.c
84
if (signal_pending(current)) {
mm/slub.c
1056
p->pid = current->pid;
mm/slub.c
2380
alloc_tag_add(&obj_ext->ref, current->alloc_tag, s->size);
mm/slub.c
2383
alloc_tag_set_inaccurate(current->alloc_tag);
mm/slub.c
4462
struct mempolicy *mpol = current->mempolicy;
mm/slub.c
4682
struct mempolicy *mpol = current->mempolicy;
mm/slub.c
928
resource = kunit_find_named_resource(current->kunit_test, "slab_errors");
mm/slub.c
944
resource = kunit_find_named_resource(current->kunit_test, "slab_errors");
mm/swap.c
508
lru_gen_in_fault() && !(current->flags & PF_MEMALLOC))
mm/swapfile.c
2424
!signal_pending(current) &&
mm/swapfile.c
2452
!signal_pending(current) &&
mm/swapfile.c
2487
if (!signal_pending(current))
mm/swapfile.c
2784
BUG_ON(!current->mm);
mm/swapfile.c
2806
if (!security_vm_enough_memory_mm(current->mm, p->pages))
mm/swapfile.c
3931
if (current->throttle_disk)
mm/usercopy.c
39
const void * const stack = task_stack_page(current);
mm/userfaultfd.c
1912
if (fatal_signal_pending(current)) {
mm/userfaultfd.c
631
if (fatal_signal_pending(current))
mm/userfaultfd.c
849
if (fatal_signal_pending(current))
mm/util.c
1297
do_munmap(current->mm, vma->vm_start, len, NULL);
mm/util.c
320
struct task_struct * __maybe_unused t = current;
mm/util.c
345
if (current->flags & PF_RANDOMIZE) {
mm/util.c
415
if (current->personality & ADDR_COMPAT_LAYOUT)
mm/util.c
449
if (current->flags & PF_RANDOMIZE)
mm/util.c
469
if (current->flags & PF_RANDOMIZE)
mm/util.c
557
ret = __account_locked_vm(mm, pages, inc, current,
mm/util.c
571
struct mm_struct *mm = current->mm;
mm/util.c
970
__func__, current->pid, current->comm, bytes_failed);
mm/vma.c
1317
mm = current->mm;
mm/vma.c
2214
if (signal_pending(current))
mm/vma.c
2221
if (signal_pending(current))
mm/vma.c
2230
if (signal_pending(current))
mm/vma.c
2239
if (signal_pending(current))
mm/vma.c
2724
struct mm_struct *mm = current->mm;
mm/vma.c
2825
mmap_assert_write_locked(current->mm);
mm/vma.c
2850
validate_mm(current->mm);
mm/vma.c
2869
struct mm_struct *mm = current->mm;
mm/vma.c
2952
VMA_ITERATOR(vmi, current->mm, 0);
mm/vma.c
3009
VMA_ITERATOR(vmi, current->mm, 0);
mm/vma.c
3254
struct mm_struct *mm = current->mm;
mm/vma.c
667
pr_emerg("issue in %s\n", current->comm);
mm/vmalloc.c
3727
if (!(gfp & __GFP_NOFAIL) && fatal_signal_pending(current))
mm/vmalloc.c
3901
if (!fatal_signal_pending(current) && page_order == 0)
mm/vmscan.c
1970
return !(current->flags & PF_LOCAL_THROTTLE);
mm/vmscan.c
2000
if (fatal_signal_pending(current))
mm/vmscan.c
337
if (current->reclaim_state && root_reclaim(sc)) {
mm/vmscan.c
338
sc->nr_reclaimed += current->reclaim_state->reclaimed;
mm/vmscan.c
339
current->reclaim_state->reclaimed = 0;
mm/vmscan.c
3815
struct lru_gen_mm_walk *walk = current->reclaim_state->mm_walk;
mm/vmscan.c
3828
current->reclaim_state->mm_walk = walk;
mm/vmscan.c
3835
struct lru_gen_mm_walk *walk = current->reclaim_state->mm_walk;
mm/vmscan.c
3840
current->reclaim_state->mm_walk = NULL;
mm/vmscan.c
4235
walk = current->reclaim_state ? current->reclaim_state->mm_walk : NULL;
mm/vmscan.c
450
if (fatal_signal_pending(current))
mm/vmscan.c
4750
walk = current->reclaim_state->mm_walk;
mm/vmscan.c
542
current->flags & (PF_USER_WORKER|PF_KTHREAD)) {
mm/vmscan.c
5493
while (!signal_pending(current)) {
mm/vmscan.c
5580
set_task_reclaim_state(current, &sc.reclaim_state);
mm/vmscan.c
5631
set_task_reclaim_state(current, NULL);
mm/vmscan.c
6500
if (current->flags & PF_KTHREAD)
mm/vmscan.c
6507
if (fatal_signal_pending(current))
mm/vmscan.c
6559
if (fatal_signal_pending(current))
mm/vmscan.c
6598
set_task_reclaim_state(current, &sc.reclaim_state);
mm/vmscan.c
6604
set_task_reclaim_state(current, NULL);
mm/vmscan.c
6627
WARN_ON_ONCE(!current->reclaim_state);
mm/vmscan.c
6679
set_task_reclaim_state(current, &sc.reclaim_state);
mm/vmscan.c
6687
set_task_reclaim_state(current, NULL);
mm/vmscan.c
6966
set_task_reclaim_state(current, &sc.reclaim_state);
mm/vmscan.c
7157
set_task_reclaim_state(current, NULL);
mm/vmscan.c
7189
if (freezing(current) || kthread_should_stop())
mm/vmscan.c
7285
struct task_struct *tsk = current;
mm/vmscan.c
7459
set_task_reclaim_state(current, &sc.reclaim_state);
mm/vmscan.c
7463
set_task_reclaim_state(current, NULL);
mm/vmscan.c
7626
struct task_struct *p = current;
mm/vmscan.c
7699
if (!gfpflags_allow_blocking(gfp_mask) || (current->flags & PF_MEMALLOC))
mm/vmscan.c
7793
if (signal_pending(current))
mm/zswap.c
1010
mpol = get_task_policy(current);
net/9p/client.c
576
if (signal_pending(current)) {
net/9p/client.c
626
spin_lock_irqsave(&current->sighand->siglock, flags);
net/9p/client.c
628
spin_unlock_irqrestore(&current->sighand->siglock, flags);
net/9p/client.c
678
if (signal_pending(current)) {
net/9p/client.c
711
spin_lock_irqsave(&current->sighand->siglock, flags);
net/9p/client.c
713
spin_unlock_irqrestore(&current->sighand->siglock, flags);
net/9p/mod.c
45
pr_notice("(%8.8d) %pV", task_pid_nr(current), &vaf);
net/9p/mod.c
47
pr_notice("-- %s (%d): %pV", func, task_pid_nr(current), &vaf);
net/9p/trans_fd.c
1054
p9_debug(P9_DEBUG_TRANS, "start %p\n", current);
net/9p/trans_fd.c
638
m, current, &req->tc, req->tc.id);
net/9p/trans_fd.c
777
__func__, task_pid_nr(current));
net/9p/trans_fd.c
890
err = inet_pton_with_scope(current->nsproxy->net_ns, AF_UNSPEC, addr,
net/9p/trans_fd.c
899
err = __sock_create(current->nsproxy->net_ns, stor.ss_family,
net/9p/trans_fd.c
903
__func__, task_pid_nr(current));
net/9p/trans_fd.c
911
__func__, task_pid_nr(current));
net/9p/trans_fd.c
922
__func__, task_pid_nr(current), addr);
net/9p/trans_fd.c
945
__func__, task_pid_nr(current), addr);
net/9p/trans_fd.c
951
err = __sock_create(current->nsproxy->net_ns, PF_UNIX,
net/9p/trans_fd.c
955
__func__, task_pid_nr(current));
net/9p/trans_fd.c
963
__func__, task_pid_nr(current), addr, err);
net/9p/trans_rdma.c
557
__func__, task_pid_nr(current), -err);
net/appletalk/ddp.c
1230
current->comm);
net/atm/common.c
590
send_sig(SIGPIPE, current, 0);
net/atm/common.c
611
if (signal_pending(current)) {
net/atm/common.c
619
send_sig(SIGPIPE, current, 0);
net/atm/ioctl.c
91
current->comm, task_pid_nr(current));
net/atm/svc.c
222
if (!signal_pending(current)) {
net/atm/svc.c
365
if (signal_pending(current)) {
net/ax25/af_ax25.c
1346
if (!signal_pending(current)) {
net/ax25/af_ax25.c
1419
if (!signal_pending(current)) {
net/ax25/af_ax25.c
1522
send_sig(SIGPIPE, current, 0);
net/ax25/af_ax25.c
1875
current->comm);
net/bluetooth/af_bluetooth.c
166
sk->sk_peer_pid = get_pid(task_tgid(current));
net/bluetooth/af_bluetooth.c
387
DECLARE_WAITQUEUE(wait, current);
net/bluetooth/af_bluetooth.c
399
if (signal_pending(current) || !timeo)
net/bluetooth/af_bluetooth.c
453
if (signal_pending(current)) {
net/bluetooth/af_bluetooth.c
704
DECLARE_WAITQUEUE(wait, current);
net/bluetooth/af_bluetooth.c
717
if (signal_pending(current)) {
net/bluetooth/af_bluetooth.c
740
DECLARE_WAITQUEUE(wait, current);
net/bluetooth/af_bluetooth.c
756
if (signal_pending(current)) {
net/bluetooth/bnep/core.c
489
set_user_nice(current, -15);
net/bluetooth/cmtp/capi.c
365
DECLARE_WAITQUEUE(wait, current);
net/bluetooth/cmtp/capi.c
416
if (signal_pending(current)) {
net/bluetooth/cmtp/core.c
287
set_user_nice(current, -15);
net/bluetooth/hci_sock.c
109
get_task_comm(hci_pi(sk)->comm, current);
net/bluetooth/hidp/core.c
1284
set_user_nice(current, -15);
net/bluetooth/iso.c
1375
if (signal_pending(current)) {
net/bluetooth/iso.c
1958
!(current->flags & PF_EXITING))
net/bluetooth/iso.c
1981
!(current->flags & PF_EXITING)) {
net/bluetooth/l2cap_sock.c
1295
DECLARE_WAITQUEUE(wait, current);
net/bluetooth/l2cap_sock.c
1311
if (signal_pending(current)) {
net/bluetooth/l2cap_sock.c
1420
!(current->flags & PF_EXITING))
net/bluetooth/l2cap_sock.c
360
if (signal_pending(current)) {
net/bluetooth/rfcomm/core.c
2119
set_user_nice(current, -10);
net/bluetooth/rfcomm/sock.c
508
if (signal_pending(current)) {
net/bluetooth/rfcomm/sock.c
905
!(current->flags & PF_EXITING))
net/bluetooth/sco.c
1330
!(current->flags & PF_EXITING))
net/bluetooth/sco.c
1354
!(current->flags & PF_EXITING)) {
net/bluetooth/sco.c
770
if (signal_pending(current)) {
net/bpf/test_run.c
1227
device = dev_get_by_index(current->nsproxy->net_ns,
net/bpf/test_run.c
1346
rxqueue = __netif_get_rx_queue(current->nsproxy->net_ns->loopback_dev, 0);
net/bpf/test_run.c
1714
struct net *net = current->nsproxy->net_ns;
net/bpf/test_run.c
64
if (signal_pending(current)) {
net/bpf/test_run.c
990
struct net *net = current->nsproxy->net_ns;
net/caif/caif_socket.c
318
signal_pending(current) ||
net/caif/caif_socket.c
409
if (signal_pending(current)) {
net/caif/caif_socket.c
474
if (signal_pending(current))
net/caif/caif_socket.c
659
send_sig(SIGPIPE, current, 0);
net/ceph/ceph_common.c
144
if (!net_eq(current->nsproxy->net_ns, read_pnet(&client->msgr.net)))
net/ceph/ceph_common.c
816
if (signal_pending(current)) {
net/ceph/messenger.c
1271
ip_len = dns_query(current->nsproxy->net_ns,
net/ceph/messenger.c
1695
write_pnet(&msgr->net, get_net(current->nsproxy->net_ns));
net/core/datagram.c
119
if (signal_pending(current))
net/core/dev.c
10677
struct net *net = current->nsproxy->net_ns;
net/core/dev.c
10833
return current == READ_ONCE(cleanup_net_task);
net/core/dev.c
12350
if (lockdep_depth(current) > limit)
net/core/dev.c
4368
return current->net_xmit.skip_txqueue;
net/core/dev.c
4373
current->net_xmit.skip_txqueue = skip;
net/core/dev.c
9602
from_kuid(&init_user_ns, audit_get_loginuid(current)),
net/core/dev.c
9605
audit_get_sessionid(current));
net/core/filter.c
3112
return __task_get_classid(current);
net/core/net-sysfs.c
2183
struct net *net = current->nsproxy->net_ns;
net/core/net-sysfs.c
2190
struct net *ns = current->nsproxy->net_ns;
net/core/net_namespace.c
670
WRITE_ONCE(cleanup_net_task, current);
net/core/netpoll.c
709
struct net *net = current->nsproxy->net_ns;
net/core/pktgen.c
2352
} while (t.task && pkt_dev->running && !signal_pending(current));
net/core/pktgen.c
3283
if (signal_pending(current))
net/core/pktgen.c
3500
if (signal_pending(current))
net/core/pktgen.c
3711
pr_debug("starting pktgen/%d: pid=%d\n", cpu, task_pid_nr(current));
net/core/pktgen.c
524
struct pktgen_net *pn = net_generic(current->nsproxy->net_ns, pg_net_id);
net/core/scm.c
58
if ((creds->pid == task_tgid_vnr(current) ||
net/core/scm.c
59
ns_capable(task_active_pid_ns(current)->user_ns, CAP_SYS_ADMIN)) &&
net/core/sock.c
2944
if (signal_pending(current))
net/core/sock.c
2991
if (signal_pending(current))
net/core/sock.c
446
__func__, current->comm, task_pid_nr(current));
net/core/stream.c
110
} while (!signal_pending(current) && timeout);
net/core/stream.c
141
if (signal_pending(current))
net/core/stream.c
195
send_sig(SIGPIPE, current, 0);
net/core/stream.c
63
struct task_struct *tsk = current;
net/dns_resolver/internal.h
47
current->comm, ##__VA_ARGS__); \
net/ethtool/ioctl.c
2484
} while (!signal_pending(current) && (!id.data || i < count));
net/ipv4/af_inet.c
441
!(current->flags & PF_EXITING))
net/ipv4/af_inet.c
620
if (signal_pending(current) || !timeo)
net/ipv4/af_inet.c
715
if (signal_pending(current))
net/ipv4/devinet.c
2784
memcpy(all, current->nsproxy->net_ns->ipv4.devconf_all,
net/ipv4/devinet.c
2787
current->nsproxy->net_ns->ipv4.devconf_dflt,
net/ipv4/inet_connection_sock.c
637
if (signal_pending(current))
net/ipv4/ip_output.c
1575
&current->task_frag, getfrag,
net/ipv4/netfilter/ip_tables.c
273
jumpstack += private->stacksize * current->in_nf_duplicate;
net/ipv4/netfilter/nf_dup_ipv4.c
57
if (current->in_nf_duplicate)
net/ipv4/netfilter/nf_dup_ipv4.c
89
current->in_nf_duplicate = true;
net/ipv4/netfilter/nf_dup_ipv4.c
91
current->in_nf_duplicate = false;
net/ipv4/raw.c
524
__func__, current->comm);
net/ipv4/tcp.c
2264
vma = find_tcp_vma(current->mm, address, &mmap_locked);
net/ipv4/tcp.c
2348
mmap_read_unlock(current->mm);
net/ipv4/tcp.c
2747
if (signal_pending(current)) {
net/ipv4/tcp.c
2791
signal_pending(current))
net/ipv4/tcp.c
2818
if (signal_pending(current)) {
net/ipv4/tcp.c
2839
current->comm,
net/ipv4/tcp.c
2840
task_pid_nr(current));
net/ipv4/tcp.c
884
if (signal_pending(current)) {
net/ipv4/tcp.c
900
signal_pending(current))
net/ipv4/tcp_bpf.c
309
if (signal_pending(current)) {
net/ipv4/tcp_ipv4.c
3485
old_net = current->nsproxy->net_ns;
net/ipv4/tcp_output.c
1400
if (refcount_read(&sk->sk_wmem_alloc) >= SKB_TRUESIZE(1) && this_cpu_ksoftirqd() == current)
net/ipv4/udp.c
3921
old_net = current->nsproxy->net_ns;
net/ipv6/addrconf.c
7422
memcpy(all, current->nsproxy->net_ns->ipv6.devconf_all,
net/ipv6/addrconf.c
7425
current->nsproxy->net_ns->ipv6.devconf_dflt,
net/ipv6/ip6_flowlabel.c
441
fl->owner.pid = get_task_pid(current, PIDTYPE_PID);
net/ipv6/ip6_output.c
2061
&current->task_frag, getfrag, from,
net/ipv6/ndisc.c
1918
if (strcmp(warncomm, current->comm) && warned < 5) {
net/ipv6/ndisc.c
1919
strscpy(warncomm, current->comm);
net/ipv6/netfilter/ip6_tables.c
295
jumpstack += private->stacksize * current->in_nf_duplicate;
net/ipv6/netfilter/nf_dup_ipv6.c
51
if (current->in_nf_duplicate)
net/ipv6/netfilter/nf_dup_ipv6.c
67
current->in_nf_duplicate = true;
net/ipv6/netfilter/nf_dup_ipv6.c
69
current->in_nf_duplicate = false;
net/iucv/af_iucv.c
67
if (signal_pending(current)) { \
net/iucv/af_iucv.c
801
DECLARE_WAITQUEUE(wait, current);
net/iucv/af_iucv.c
833
if (signal_pending(current)) {
net/l2tp/l2tp_debugfs.c
279
pd->net = get_net_ns_by_pid(current->pid);
net/l2tp/l2tp_ppp.c
523
ps->owner = current->pid;
net/llc/af_llc.c
527
if (signal_pending(current))
net/llc/af_llc.c
592
if (signal_pending(current))
net/llc/af_llc.c
612
if (signal_pending(current) || !timeout)
net/llc/af_llc.c
635
if (signal_pending(current))
net/llc/af_llc.c
663
if (signal_pending(current))
net/llc/af_llc.c
798
if (signal_pending(current)) {
net/llc/af_llc.c
860
current->comm,
net/llc/af_llc.c
861
task_pid_nr(current));
net/mctp/test/route-test.c
1294
struct kunit *test = current->kunit_test;
net/mptcp/protocol.c
2322
signal_pending(current))
net/mptcp/protocol.c
2343
if (signal_pending(current)) {
net/mptcp/protocol.c
4491
if (signal_pending(current)) {
net/mptcp/protocol.c
4507
signal_pending(current))
net/netfilter/ipvs/ip_vs_ctl.c
1951
cpumask_and(newmask, newmask, &current->cpus_mask);
net/netfilter/ipvs/ip_vs_sync.c
1750
IP_VS_DBG(7, "%s(): pid %d\n", __func__, task_pid_nr(current));
net/netfilter/ipvs/ip_vs_sync.c
1937
IP_VS_DBG(7, "%s(): pid %d\n", __func__, task_pid_nr(current));
net/netfilter/nf_bpf_link.c
212
struct net *net = current->nsproxy->net_ns;
net/netfilter/nf_conntrack_core.c
2617
if (current->nsproxy->net_ns != &init_net)
net/netfilter/nf_dup_netdev.c
27
return &current->net_xmit.nf_dup_skb_recursion;
net/netfilter/nf_tables_api.c
4111
if (fatal_signal_pending(current))
net/netfilter/nf_tables_api.c
9614
nla_put_be32(skb, NFTA_GEN_PROC_PID, htonl(task_pid_nr(current))) ||
net/netfilter/nf_tables_api.c
9615
nla_put_string(skb, NFTA_GEN_PROC_NAME, get_task_comm(buf, current)))
net/netfilter/nfnetlink.c
440
if (fatal_signal_pending(current)) {
net/netfilter/nft_set_rbtree.c
695
if (fatal_signal_pending(current))
net/netlabel/netlabel_user.h
36
audit_info->loginuid = audit_get_loginuid(current);
net/netlabel/netlabel_user.h
37
audit_info->sessionid = audit_get_sessionid(current);
net/netlink/af_netlink.c
1212
DECLARE_WAITQUEUE(wait, current);
net/netlink/af_netlink.c
1248
if (signal_pending(current)) {
net/netlink/af_netlink.c
421
DECLARE_WAITQUEUE(wait, current);
net/netlink/af_netlink.c
805
s32 portid = task_tgid_vnr(current);
net/netrom/af_netrom.c
1063
send_sig(SIGPIPE, current, 0);
net/netrom/af_netrom.c
747
if (!signal_pending(current)) {
net/netrom/af_netrom.c
812
if (!signal_pending(current)) {
net/nfc/llcp_sock.c
19
DECLARE_WAITQUEUE(wait, current);
net/nfc/llcp_sock.c
33
if (signal_pending(current)) {
net/nfc/llcp_sock.c
452
DECLARE_WAITQUEUE(wait, current);
net/nfc/llcp_sock.c
478
if (signal_pending(current)) {
net/openvswitch/datapath.c
303
if (IS_ENABLED(CONFIG_PREEMPT_RT) && ovs_pcpu->owner != current) {
net/openvswitch/datapath.c
305
ovs_pcpu->owner = current;
net/openvswitch/datapath.c
699
this_cpu_write(ovs_pcpu_storage->owner, current);
net/phonet/pep.c
1193
if (signal_pending(current)) {
net/phonet/socket.c
223
struct task_struct *tsk = current;
net/rose/af_rose.c
1134
send_sig(SIGPIPE, current, 0);
net/rose/af_rose.c
908
if (!signal_pending(current)) {
net/rose/af_rose.c
975
if (!signal_pending(current)) {
net/rxrpc/ar-internal.h
1598
printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__)
net/rxrpc/io_thread.c
467
set_user_nice(current, MIN_NICE);
net/rxrpc/recvmsg.c
416
if (signal_pending(current))
net/rxrpc/sendmsg.c
119
if (signal_pending(current))
net/rxrpc/sendmsg.c
155
tx_win == tx_start && signal_pending(current))
net/rxrpc/sendmsg.c
197
DECLARE_WAITQUEUE(myself, current);
net/rxrpc/sendmsg.c
48
DECLARE_WAITQUEUE(myself, current);
net/rxrpc/sendmsg.c
74
signal_pending(current)) {
net/sched/act_mirred.c
433
xmit = &current->net_xmit;
net/sctp/socket.c
1565
send_sig(SIGPIPE, current, 0);
net/sctp/socket.c
2842
current->comm, task_pid_nr(current));
net/sctp/socket.c
3265
current->comm, task_pid_nr(current));
net/sctp/socket.c
3531
current->comm, task_pid_nr(current));
net/sctp/socket.c
5685
if (!net_eq(current->nsproxy->net_ns, sock_net(sk)))
net/sctp/socket.c
6117
current->comm, task_pid_nr(current));
net/sctp/socket.c
6811
current->comm, task_pid_nr(current));
net/sctp/socket.c
6906
current->comm, task_pid_nr(current));
net/sctp/socket.c
9030
if (signal_pending(current))
net/sctp/socket.c
9272
if (signal_pending(current))
net/sctp/socket.c
9384
if (signal_pending(current))
net/sctp/socket.c
9453
if (signal_pending(current))
net/sctp/socket.c
9477
} while (!signal_pending(current) && timeout);
net/smc/af_smc.c
1635
else if (signal_pending(current))
net/smc/af_smc.c
2718
DECLARE_WAITQUEUE(wait, current);
net/smc/af_smc.c
2747
if (signal_pending(current)) {
net/smc/smc_clc.c
715
if (signal_pending(current)) {
net/smc/smc_close.c
204
timeout = current->flags & PF_EXITING ?
net/smc/smc_close.c
455
timeout = current->flags & PF_EXITING ?
net/smc/smc_close.c
65
while (!signal_pending(current) && timeout) {
net/smc/smc_rx.c
411
signal_pending(current))
net/smc/smc_rx.c
430
if (signal_pending(current)) {
net/smc/smc_tx.c
107
if (signal_pending(current)) {
net/socket.c
1556
current->comm);
net/socket.c
1663
return __sock_create(current->nsproxy->net_ns, family, type, protocol, res, 0);
net/sunrpc/auth_gss/auth_gss.c
686
if (fatal_signal_pending(current)) {
net/sunrpc/clnt.c
1918
if (RPC_IS_ASYNC(task) || !fatal_signal_pending(current)) {
net/sunrpc/sched.c
1111
task->tk_owner = current->tgid;
net/sunrpc/sched.c
280
if (signal_pending_state(mode, current))
net/sunrpc/sched.c
62
if (current->flags & PF_WQ_WORKER)
net/sunrpc/sched.c
914
unsigned long pflags = current->flags;
net/sunrpc/sched.c
944
current->flags |= PF_MEMALLOC;
net/sunrpc/svc.c
1283
sighand = rcu_dereference(current->sighand);
net/sunrpc/xprt.c
1519
current->flags |= PF_MEMALLOC;
net/sunrpc/xprtrdma/transport.c
228
unsigned int pflags = current->flags;
net/sunrpc/xprtrdma/transport.c
232
current->flags |= PF_MEMALLOC;
net/sunrpc/xprtsock.c
1282
if (!(current->flags & PF_WQ_WORKER)) {
net/sunrpc/xprtsock.c
2220
unsigned int pflags = current->flags;
net/sunrpc/xprtsock.c
2223
current->flags |= PF_MEMALLOC;
net/sunrpc/xprtsock.c
2425
unsigned int pflags = current->flags;
net/sunrpc/xprtsock.c
2428
current->flags |= PF_MEMALLOC;
net/sunrpc/xprtsock.c
2709
unsigned int pflags = current->flags;
net/sunrpc/xprtsock.c
2715
current->flags |= PF_MEMALLOC;
net/sysctl_net.c
33
return &current->nsproxy->net_ns->sysctls;
net/sysctl_net.c
38
return &current->nsproxy->net_ns->sysctls == set;
net/tipc/socket.c
1858
if (signal_pending(current))
net/tipc/socket.c
2538
if (signal_pending(current))
net/tipc/socket.c
2702
if (signal_pending(current))
net/tipc/socket.c
417
if (signal_pending(current))
net/tls/tls_main.c
153
if (signal_pending(current)) {
net/tls/tls_sw.c
1401
if (signal_pending(current))
net/tls/tls_sw.c
1989
if (signal_pending(current))
net/unix/af_unix.c
1704
if (signal_pending(current))
net/unix/af_unix.c
1938
if (too_many_unix_fds(current))
net/unix/af_unix.c
2026
pid = task_tgid(current);
net/unix/af_unix.c
2250
if (signal_pending(current))
net/unix/af_unix.c
2503
send_sig(SIGPIPE, current, 0);
net/unix/af_unix.c
2711
signal_pending(current) ||
net/unix/af_unix.c
2984
if (signal_pending(current)) {
net/unix/af_unix.c
749
pid = task_tgid(current);
net/vmw_vsock/af_vsock.c
1173
} while (!signal_pending(current) && timeout);
net/vmw_vsock/af_vsock.c
1787
if (signal_pending(current) || timeout == 0) {
net/vmw_vsock/af_vsock.c
1862
if (signal_pending(current)) {
net/vmw_vsock/af_vsock.c
2231
if (signal_pending(current)) {
net/vmw_vsock/af_vsock.c
2348
if (signal_pending(current)) {
net/vmw_vsock/af_vsock.c
2928
net->vsock.mode = vsock_net_child_mode(current->nsproxy->net_ns);
net/vmw_vsock/virtio_transport_common.c
1297
if (!(current->flags & PF_EXITING))
net/wireless/wext-core.c
644
current->comm);
net/x25/af_x25.c
1129
send_sig(SIGPIPE, current, 0);
net/x25/af_x25.c
714
DECLARE_WAITQUEUE(wait, current);
net/x25/af_x25.c
721
if (signal_pending(current))
net/x25/af_x25.c
847
DECLARE_WAITQUEUE(wait, current);
net/x25/af_x25.c
856
if (signal_pending(current))
net/xdp/xdp_umem.c
105
mmap_read_lock(current->mm);
net/xdp/xdp_umem.c
108
mmap_read_unlock(current->mm);
net/xfrm/xfrm_state.c
2616
if (signal_pending(current)) {
rust/helpers/task.c
13
return current;
samples/kprobes/kretprobe_example.c
43
if (!current->mm)
samples/trace_events/trace-events-sample.c
41
current->cpus_ptr, fmt, &va);
samples/trace_events/trace-events-sample.c
53
trace_foo_rel_loc("Hello __rel_loc", cnt, bitmask, current->cpus_ptr);
samples/workqueue/stall_detector/wq_stall.c
61
current->flags &= ~PF_WQ_WORKER;
samples/workqueue/stall_detector/wq_stall.c
68
current->flags |= PF_WQ_WORKER;
scripts/gendwarfksyms/dwarf.c
270
Dwarf_Die current;
scripts/gendwarfksyms/dwarf.c
277
res = checkp(dwarf_child(die, &current));
scripts/gendwarfksyms/dwarf.c
279
if (match(&current)) {
scripts/gendwarfksyms/dwarf.c
281
res = checkp(func(state, cache, &current));
scripts/gendwarfksyms/dwarf.c
286
res = checkp(dwarf_siblingof(&current, &current));
security/apparmor/audit.c
197
ad->common.u.tsk ? ad->common.u.tsk : current);
security/apparmor/domain.c
1226
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/domain.c
1245
if (task_no_new_privs(current) && !unconfined(label) && !ctx->nnp)
security/apparmor/domain.c
1287
if (task_no_new_privs(current) && !unconfined(label) &&
security/apparmor/domain.c
1309
if (task_no_new_privs(current) && !unconfined(label) &&
security/apparmor/domain.c
1398
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/domain.c
1413
if (task_no_new_privs(current) && !unconfined(label) && !ctx->nnp)
security/apparmor/domain.c
1534
if (task_no_new_privs(current) && !unconfined(label) &&
security/apparmor/domain.c
59
tracer = ptrace_parent(current);
security/apparmor/domain.c
937
ctx = task_ctx(current);
security/apparmor/lsm.c
120
aa_dup_task_ctx(new, task_ctx(current));
security/apparmor/lsm.c
2255
struct cred *cred = (__force struct cred *)current->real_cred;
security/apparmor/lsm.c
785
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/lsm.c
827
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/lsm.c
964
aa_inherit_files(bprm->cred, current->files);
security/apparmor/lsm.c
966
current->pdeath_signal = 0;
security/apparmor/lsm.c
979
aa_clear_task_ctx_trans(task_ctx(current));
security/apparmor/path.c
121
get_fs_root(current->fs, &root);
security/apparmor/procattr.c
130
, __func__, current->pid, token, count, hat);
security/apparmor/procattr.c
136
__func__, current->pid, token, count, "<NULL>");
security/apparmor/resource.c
174
rlim = current->signal->rlim + j;
security/apparmor/resource.c
194
rlim = current->signal->rlim + j;
security/apparmor/resource.c
202
(void) update_rlimit_cpu(current->group_leader,
security/apparmor/task.c
100
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
120
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
160
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
310
exe_file = get_task_exe_file(current);
security/apparmor/task.c
50
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
76
aa_clear_task_ctx_trans(task_ctx(current));
security/commoncap.c
1478
current->flags |= PF_SUPERPRIV;
security/commoncap.c
952
!ptracer_capable(current, new->user_ns))) {
security/device_cgroup.c
833
dev_cgroup = task_devcgroup(current);
security/integrity/ima/ima_policy.c
621
if (has_capability_noaudit(current, CAP_SETUID)) {
security/integrity/ima/ima_policy.c
632
if (has_capability_noaudit(current, CAP_SETGID)) {
security/integrity/integrity_audit.c
51
task_pid_nr(current),
security/integrity/integrity_audit.c
53
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/integrity/integrity_audit.c
54
audit_get_sessionid(current));
security/integrity/integrity_audit.c
57
audit_log_untrustedstring(ab, get_task_comm(name, current));
security/ipe/audit.c
133
char comm[sizeof(current->comm)];
security/ipe/audit.c
147
task_tgid_nr(current));
security/ipe/audit.c
148
audit_log_untrustedstring(ab, get_task_comm(comm, current));
security/ipe/audit.c
224
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/ipe/audit.c
225
audit_get_sessionid(current));
security/ipe/audit.c
252
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/ipe/audit.c
253
audit_get_sessionid(current), !err, err);
security/ipe/audit.c
275
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/ipe/audit.c
276
audit_get_sessionid(current));
security/ipe/hooks.c
222
ipe_sb(current->fs->root.mnt->mnt_sb)->initramfs = true;
security/keys/keyctl.c
1653
me = current;
security/keys/keyring.c
223
index_key->domain_tag = current->nsproxy->net_ns->key_domain;
security/keys/process_keys.c
916
if (unlikely(current->flags & PF_EXITING)) {
security/keys/request_key.c
25
struct key *key = current->cached_requested_key;
security/keys/request_key.c
39
struct task_struct *t = current;
security/keys/request_key_auth.c
207
rka->pid = current->pid;
security/landlock/cred.h
98
if (task == current)
security/landlock/domain.c
102
details->pid = get_pid(task_tgid(current));
security/landlock/domain.c
104
get_task_comm(details->comm, current);
security/landlock/domain.c
44
struct mm_struct *mm = current->mm;
security/landlock/fs.c
1786
return !same_thread_group(p, current);
security/landlock/syscalls.c
498
if (!task_no_new_privs(current) &&
security/landlock/task.c
161
.u.tsk = current,
security/landlock/task.c
355
if (same_thread_group(p, current))
security/landlock/tsync.c
143
task_set_no_new_privs(current);
security/landlock/tsync.c
337
caller = current;
security/landlock/tsync.c
378
caller = current;
security/landlock/tsync.c
486
shared_ctx.set_no_new_privs = task_no_new_privs(current);
security/landlock/tsync.c
495
if (!down_write_trylock(&current->signal->exec_update_lock))
security/landlock/tsync.c
613
up_write(&current->signal->exec_update_lock);
security/loadpin/loadpin.c
33
cmdline = kstrdup_quotable_cmdline(current, GFP_KERNEL);
security/loadpin/loadpin.c
40
task_pid_nr(current),
security/lockdown/lockdown.c
68
current->comm, lockdown_reasons[what]);
security/lsm_audit.c
417
char comm[sizeof(current->comm)];
security/lsm_audit.c
419
audit_log_format(ab, " pid=%d comm=", task_tgid_nr(current));
security/lsm_audit.c
420
audit_log_untrustedstring(ab, get_task_comm(comm, current));
security/lsm_init.c
470
if (lsm_cred_alloc((struct cred *)unrcu_pointer(current->cred),
security/lsm_init.c
473
if (lsm_task_alloc(current))
security/security.c
2466
if (!(current->personality & READ_IMPLIES_EXEC))
security/selinux/hooks.c
2132
return avc_has_perm(task_sid_obj(parent), task_sid_obj(current),
security/selinux/hooks.c
219
crsec = selinux_cred(unrcu_pointer(current->real_cred));
security/selinux/hooks.c
2251
tracer = ptrace_parent(current);
security/selinux/hooks.c
2505
flush_unauthorized_files(bprm->cred, current->files);
security/selinux/hooks.c
2508
current->pdeath_signal = 0;
security/selinux/hooks.c
2524
task_lock(current);
security/selinux/hooks.c
2526
rlim = current->signal->rlim + i;
security/selinux/hooks.c
2530
task_unlock(current);
security/selinux/hooks.c
2532
update_rlimit_cpu(current, rlimit(RLIMIT_CPU));
security/selinux/hooks.c
2563
spin_lock_irq(&unrcu_pointer(current->sighand)->siglock);
security/selinux/hooks.c
2564
if (!fatal_signal_pending(current)) {
security/selinux/hooks.c
2565
flush_sigqueue(&current->pending);
security/selinux/hooks.c
2566
flush_sigqueue(&current->signal->shared_pending);
security/selinux/hooks.c
2567
flush_signal_handlers(current, 1);
security/selinux/hooks.c
2568
sigemptyset(&current->blocked);
security/selinux/hooks.c
2571
spin_unlock_irq(&unrcu_pointer(current->sighand)->siglock);
security/selinux/hooks.c
2577
__wake_up_parent(current, unrcu_pointer(current->real_parent));
security/selinux/hooks.c
3239
tsec = selinux_task(current);
security/selinux/hooks.c
3311
tsec = selinux_task(current);
security/selinux/hooks.c
4177
struct task_security_struct *old_tsec = selinux_task(current);
security/selinux/hooks.c
6176
task_pid_nr(current), current->comm);
security/selinux/hooks.c
6563
if (p != current) {
security/selinux/hooks.c
6773
val_len = selinux_lsm_getattr(attr, current, &val);
security/selinux/selinuxfs.c
1085
" userspace.\n", current->comm, current->pid);
security/selinux/selinuxfs.c
172
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/selinux/selinuxfs.c
173
audit_get_sessionid(current));
security/selinux/selinuxfs.c
631
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/selinux/selinuxfs.c
632
audit_get_sessionid(current));
security/selinux/selinuxfs.c
720
char comm[sizeof(current->comm)];
security/selinux/selinuxfs.c
722
strscpy(comm, current->comm);
security/selinux/selinuxfs.c
724
comm, current->pid);
security/selinux/ss/services.c
3141
from_kuid(&init_user_ns, audit_get_loginuid(current)),
security/selinux/ss/services.c
3142
audit_get_sessionid(current));
security/smack/smack_access.c
752
if (unlikely(current->flags & PF_KTHREAD))
security/smack/smack_lsm.c
148
acc, current->comm, note);
security/smack/smack_lsm.c
170
current->comm, otp->comm);
security/smack/smack_lsm.c
186
inode->i_sb->s_id, inode->i_ino, current->comm);
security/smack/smack_lsm.c
200
inode->i_sb->s_id, inode->i_ino, current->comm);
security/smack/smack_lsm.c
218
inode->i_sb->s_id, inode->i_ino, current->comm);
security/smack/smack_lsm.c
229
current->comm);
security/smack/smack_lsm.c
2457
if (unlikely(current->flags & PF_KTHREAD)) {
security/smack/smack_lsm.c
248
inode->i_sb->s_id, inode->i_ino, current->comm);
security/smack/smack_lsm.c
259
current->comm);
security/smack/smack_lsm.c
2999
if (unlikely(current->flags & PF_KTHREAD)) {
security/smack/smack_lsm.c
481
return smk_ptrace_rule_check(current, skp, mode, __func__);
security/smack/smack_lsm.c
5328
struct cred *cred = (struct cred *) current->cred;
security/smack/smack_lsm.c
923
tracer = ptrace_parent(current);
security/smack/smackfs.c
188
nap->loginuid = audit_get_loginuid(current);
security/smack/smackfs.c
189
nap->sessionid = audit_get_sessionid(current);
security/tomoyo/audit.c
153
const pid_t gpid = task_pid_nr(current);
security/tomoyo/common.c
944
const struct task_struct *task = current;
security/tomoyo/common.c
968
const pid_t pid = current->pid;
security/tomoyo/common.h
1141
pid = task_tgid_vnr(rcu_dereference(current->real_parent));
security/tomoyo/common.h
1156
return task_tgid_vnr(current);
security/tomoyo/domain.c
880
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/memory.c
21
const pid_t pid = current->pid;
security/tomoyo/network.c
619
return current->flags & PF_KTHREAD;
security/tomoyo/realpath.c
168
task_tgid_nr_ns(current, proc_pidns)) {
security/tomoyo/securityfs_if.c
73
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/tomoyo.c
19
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/tomoyo.c
21
if (s->old_domain_info && !current->in_execve) {
security/tomoyo/tomoyo.c
41
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/tomoyo.c
43
if (s->old_domain_info && !current->in_execve) {
security/tomoyo/tomoyo.c
519
struct tomoyo_task *old = tomoyo_task(current);
security/tomoyo/tomoyo.c
59
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/tomoyo.c
600
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/tomoyo.c
94
struct tomoyo_task *s = tomoyo_task(current);
security/tomoyo/util.c
970
struct mm_struct *mm = current->mm;
security/yama/yama_lsm.c
101
if (task_work_add(current, &info->work, TWA_RESUME) == 0)
security/yama/yama_lsm.c
235
myself = current->group_leader;
security/yama/yama_lsm.c
364
if (!rc && !task_is_descendant(current, child) &&
security/yama/yama_lsm.c
365
!ptracer_exception_found(current, child) &&
security/yama/yama_lsm.c
384
report_access("attach", child, current);
security/yama/yama_lsm.c
411
task_lock(current);
security/yama/yama_lsm.c
412
report_access("traceme", current, parent);
security/yama/yama_lsm.c
413
task_unlock(current);
security/yama/yama_lsm.c
82
if (current->flags & PF_KTHREAD) {
sound/core/control.c
2001
init_waitqueue_entry(&wait, current);
sound/core/control.c
2009
if (signal_pending(current))
sound/core/control.c
2165
if (kctl->pid == task_pid(current)) {
sound/core/control.c
80
ctl->pid = get_pid(task_pid(current));
sound/core/hwdep.c
123
if (signal_pending(current)) {
sound/core/hwdep.c
93
init_waitqueue_entry(&wait, current);
sound/core/oss/pcm_oss.c
1460
if (signal_pending(current)) {
sound/core/oss/pcm_oss.c
1555
if (signal_pending(current)) {
sound/core/oss/pcm_oss.c
1613
init_waitqueue_entry(&wait, current);
sound/core/oss/pcm_oss.c
1636
if (signal_pending(current)) {
sound/core/oss/pcm_oss.c
2503
if (snd_task_name(current, task_name, sizeof(task_name)) < 0) {
sound/core/oss/pcm_oss.c
2519
init_waitqueue_entry(&wait, current);
sound/core/oss/pcm_oss.c
2542
if (signal_pending(current)) {
sound/core/pcm.c
971
substream->pid = get_pid(task_pid(current));
sound/core/pcm_lib.c
1959
init_waitqueue_entry(&wait, current);
sound/core/pcm_lib.c
1981
if (signal_pending(current)) {
sound/core/pcm_native.c
2151
if (signal_pending(current)) {
sound/core/pcm_native.c
2179
init_waitqueue_entry(&wait, current);
sound/core/pcm_native.c
2885
init_waitqueue_entry(&wait, current);
sound/core/pcm_native.c
2907
if (signal_pending(current)) {
sound/core/rawmidi.c
1295
init_waitqueue_entry(&wait, current);
sound/core/rawmidi.c
1303
if (signal_pending(current))
sound/core/rawmidi.c
1612
init_waitqueue_entry(&wait, current);
sound/core/rawmidi.c
1620
if (signal_pending(current))
sound/core/rawmidi.c
1644
init_waitqueue_entry(&wait, current);
sound/core/rawmidi.c
1650
if (signal_pending(current))
sound/core/rawmidi.c
268
if (signal_pending(current))
sound/core/rawmidi.c
358
substream->pid = get_pid(task_pid(current));
sound/core/rawmidi.c
481
init_waitqueue_entry(&wait, current);
sound/core/rawmidi.c
503
if (signal_pending(current)) {
sound/core/seq/oss/seq_oss_ioctl.c
88
if (signal_pending(current))
sound/core/seq/oss/seq_oss_rw.c
50
if (signal_pending(current))
sound/core/seq/oss/seq_oss_writeq.c
111
if (signal_pending(current))
sound/core/seq/seq_clientmgr.c
340
client->data.user.owner = get_pid(task_pid(current));
sound/core/seq/seq_fifo.c
168
init_waitqueue_entry(&wait, current);
sound/core/seq/seq_fifo.c
182
if (signal_pending(current)) {
sound/core/seq/seq_memory.c
282
init_waitqueue_entry(&wait, current);
sound/core/seq/seq_memory.c
302
if (signal_pending(current)) {
sound/core/timer.c
1729
sprintf(str, "application %i", current->pid);
sound/core/timer.c
1745
err = snd_timer_open(tu->timeri, &tselect.id, current->pid);
sound/core/timer.c
2347
init_waitqueue_entry(&wait, current);
sound/core/timer.c
2362
if (signal_pending(current)) {
sound/drivers/aloop.c
1210
err = snd_timer_open(timeri, &cable->snd_timer.id, current->pid);
sound/firewire/bebob/bebob_hwdep.c
33
if (signal_pending(current))
sound/firewire/dice/dice-hwdep.c
25
if (signal_pending(current))
sound/firewire/digi00x/digi00x-hwdep.c
33
if (signal_pending(current))
sound/firewire/fireface/ff-hwdep.c
39
if (signal_pending(current))
sound/firewire/fireworks/fireworks_hwdep.c
138
if (signal_pending(current))
sound/firewire/motu/motu-hwdep.c
41
if (signal_pending(current))
sound/firewire/oxfw/oxfw-hwdep.c
32
if (signal_pending(current))
sound/firewire/tascam/tascam-hwdep.c
111
if (signal_pending(current))
sound/isa/gus/gus_pcm.c
313
if (signal_pending(current))
sound/isa/msnd/msnd.c
198
if (!signal_pending(current))
sound/isa/sb/emu8000.c
116
if (signal_pending(current))
sound/isa/sb/emu8000.c
128
if (signal_pending(current))
sound/isa/sb/emu8000.c
427
if (signal_pending(current))
sound/isa/sb/emu8000_patch.c
102
if (signal_pending(current))
sound/isa/sb/emu8000_pcm.c
109
if (signal_pending(current))
sound/isa/sb/emu8000_pcm.c
406
if (signal_pending(current))\
sound/isa/wavefront/wavefront_synth.c
1798
init_waitqueue_entry(&wait, current);
sound/isa/wavefront/wavefront_synth.c
272
return signal_pending(current);
sound/oss/dmasound/dmasound_core.c
629
if (signal_pending(current)) {
sound/oss/dmasound/dmasound_core.c
852
if (signal_pending(current)) {
sound/pci/korg1212/korg1212.c
1331
korg1212->playback_pid = current->pid;
sound/pci/korg1212/korg1212.c
1359
korg1212->capture_pid = current->pid;
sound/pci/mixart/mixart_core.c
227
init_waitqueue_entry(&wait, current);
sound/pci/mixart/mixart_core.c
280
init_waitqueue_entry(&wait, current);
sound/pci/rme9652/hdsp.c
4429
hdsp->playback_pid = current->pid;
sound/pci/rme9652/hdsp.c
4502
hdsp->capture_pid = current->pid;
sound/pci/rme9652/hdspm.c
6003
hdspm->playback_pid = current->pid;
sound/pci/rme9652/hdspm.c
6009
hdspm->capture_pid = current->pid;
sound/pci/rme9652/rme9652.c
2204
rme9652->playback_pid = current->pid;
sound/pci/rme9652/rme9652.c
2259
rme9652->capture_pid = current->pid;
sound/pci/ymfpci/ymfpci_main.c
742
init_waitqueue_entry(&wait, current);
sound/soc/intel/avs/debugfs.c
219
if (!signal_pending(current))
sound/soc/sof/ipc3-dtrace.c
307
init_waitqueue_entry(&wait, current);
sound/soc/sof/ipc3-dtrace.c
311
if (!signal_pending(current)) {
sound/soc/sof/ipc4-mtrace.c
154
init_waitqueue_entry(&wait, current);
sound/soc/sof/ipc4-mtrace.c
158
if (!signal_pending(current)) {
tools/iio/iio_utils.c
31
char *current;
tools/iio/iio_utils.c
321
struct iio_channel_info *current;
tools/iio/iio_utils.c
385
current = &(*ci_array)[count++];
tools/iio/iio_utils.c
421
current->scale = 1.0;
tools/iio/iio_utils.c
422
current->offset = 0;
tools/iio/iio_utils.c
423
current->name = strndup(ent->d_name,
tools/iio/iio_utils.c
426
if (!current->name) {
tools/iio/iio_utils.c
43
current = strdup(full_name + strlen(prefix) + 1);
tools/iio/iio_utils.c
433
ret = iioutils_break_up_name(current->name,
tools/iio/iio_utils.c
434
&current->generic_name);
tools/iio/iio_utils.c
436
free(current->name);
tools/iio/iio_utils.c
44
if (!current)
tools/iio/iio_utils.c
444
current->name);
tools/iio/iio_utils.c
455
scan_el_dir, current->name);
tools/iio/iio_utils.c
460
if (fscanf(sysfsfp, "%u", &current->index) != 1) {
tools/iio/iio_utils.c
47
working = strtok(current, "_\0");
tools/iio/iio_utils.c
474
ret = iioutils_get_param_float(&current->scale,
tools/iio/iio_utils.c
477
current->name,
tools/iio/iio_utils.c
478
current->generic_name);
tools/iio/iio_utils.c
482
ret = iioutils_get_param_float(&current->offset,
tools/iio/iio_utils.c
485
current->name,
tools/iio/iio_utils.c
486
current->generic_name);
tools/iio/iio_utils.c
49
free(current);
tools/iio/iio_utils.c
490
ret = iioutils_get_type(&current->is_signed,
tools/iio/iio_utils.c
491
&current->bytes,
tools/iio/iio_utils.c
492
&current->bits_used,
tools/iio/iio_utils.c
493
&current->shift,
tools/iio/iio_utils.c
494
&current->mask,
tools/iio/iio_utils.c
495
&current->be,
tools/iio/iio_utils.c
498
current->name,
tools/iio/iio_utils.c
499
current->generic_name);
tools/iio/iio_utils.c
66
free(current);
tools/perf/builtin-timechart.c
105
struct per_pidcomm *current;
tools/perf/builtin-timechart.c
226
p->current = c;
tools/perf/builtin-timechart.c
240
p->current = c;
tools/perf/builtin-timechart.c
245
p->current = c;
tools/perf/builtin-timechart.c
261
if (pp->current && pp->current->comm && !p->current)
tools/perf/builtin-timechart.c
262
pid_set_comm(tchart, pid, pp->current->comm);
tools/perf/builtin-timechart.c
265
if (p->current && !p->current->start_time) {
tools/perf/builtin-timechart.c
266
p->current->start_time = timestamp;
tools/perf/builtin-timechart.c
267
p->current->state_since = timestamp;
tools/perf/builtin-timechart.c
276
if (p->current)
tools/perf/builtin-timechart.c
277
p->current->end_time = timestamp;
tools/perf/builtin-timechart.c
289
c = p->current;
tools/perf/builtin-timechart.c
446
if (p && p->current && p->current->state == TYPE_NONE) {
tools/perf/builtin-timechart.c
447
p->current->state_since = timestamp;
tools/perf/builtin-timechart.c
448
p->current->state = TYPE_WAITING;
tools/perf/builtin-timechart.c
450
if (p && p->current && p->current->state == TYPE_BLOCKED) {
tools/perf/builtin-timechart.c
451
pid_put_sample(tchart, p->pid, p->current->state, cpu,
tools/perf/builtin-timechart.c
452
p->current->state_since, timestamp, NULL);
tools/perf/builtin-timechart.c
453
p->current->state_since = timestamp;
tools/perf/builtin-timechart.c
454
p->current->state = TYPE_WAITING;
tools/perf/builtin-timechart.c
468
if (prev_p->current && prev_p->current->state != TYPE_NONE)
tools/perf/builtin-timechart.c
470
prev_p->current->state_since, timestamp,
tools/perf/builtin-timechart.c
472
if (p && p->current) {
tools/perf/builtin-timechart.c
473
if (p->current->state != TYPE_NONE)
tools/perf/builtin-timechart.c
474
pid_put_sample(tchart, next_pid, p->current->state, cpu,
tools/perf/builtin-timechart.c
475
p->current->state_since, timestamp,
tools/perf/builtin-timechart.c
478
p->current->state_since = timestamp;
tools/perf/builtin-timechart.c
479
p->current->state = TYPE_RUNNING;
tools/perf/builtin-timechart.c
482
if (prev_p->current) {
tools/perf/builtin-timechart.c
483
prev_p->current->state = TYPE_NONE;
tools/perf/builtin-timechart.c
484
prev_p->current->state_since = timestamp;
tools/perf/builtin-timechart.c
486
prev_p->current->state = TYPE_BLOCKED;
tools/perf/builtin-timechart.c
488
prev_p->current->state = TYPE_WAITING;
tools/perf/builtin-timechart.c
734
struct per_pidcomm *c = p->current;
tools/perf/builtin-timechart.c
776
struct per_pidcomm *c = p->current;
tools/perf/builtin-trace.c
162
struct thread *current;
tools/perf/builtin-trace.c
2699
if (trace->failure_only || trace->current == NULL)
tools/perf/builtin-trace.c
2702
ttrace = thread__priv(trace->current);
tools/perf/builtin-trace.c
2707
printed = trace__fprintf_entry_head(trace, trace->current, 0, false, ttrace->entry_time, trace->output);
tools/perf/builtin-trace.c
2847
if (trace->current != thread) {
tools/perf/builtin-trace.c
2848
thread__put(trace->current);
tools/perf/builtin-trace.c
2849
trace->current = thread__get(thread);
tools/perf/builtin-trace.c
4613
thread__zput(trace->current);
tools/perf/builtin-trace.c
5348
thread__zput(trace->current);
tools/perf/ui/browser.c
21
double percent, bool current)
tools/perf/ui/browser.c
23
if (current && (!browser->use_navkeypressed || browser->navkeypressed))
tools/perf/ui/browser.c
41
double percent, bool current)
tools/perf/ui/browser.c
43
int color = ui_browser__percent_color(browser, percent, current);
tools/perf/ui/browser.h
39
double percent, bool current);
tools/perf/ui/browsers/annotate-data.c
372
bool current = ui_browser__is_current_entry(uib, row);
tools/perf/ui/browsers/annotate-data.c
375
ui_browser__set_percent_color(uib, percent, current);
tools/perf/ui/browsers/annotate-data.c
384
ui_browser__set_percent_color(uib, 0, current);
tools/perf/ui/browsers/annotate-data.c
397
bool current = ui_browser__is_current_entry(uib, row);
tools/perf/ui/browsers/annotate-data.c
401
ui_browser__set_percent_color(uib, 0, current);
tools/perf/ui/browsers/annotate-data.c
410
ui_browser__set_percent_color(uib, 0, current);
tools/perf/ui/browsers/annotate.c
100
static void annotate_browser__set_percent_color(void *browser, double percent, bool current)
tools/perf/ui/browsers/annotate.c
102
ui_browser__set_percent_color(browser, percent, current);
tools/perf/ui/browsers/annotate.c
71
static int ui_browser__jumps_percent_color(struct ui_browser *browser, int nr, bool current)
tools/perf/ui/browsers/annotate.c
75
if (current && (!browser->use_navkeypressed || browser->navkeypressed))
tools/perf/ui/browsers/annotate.c
84
static int ui_browser__set_jumps_percent_color(void *browser, int nr, bool current)
tools/perf/ui/browsers/annotate.c
86
int color = ui_browser__jumps_percent_color(browser, nr, current);
tools/perf/util/annotate.c
1320
bool current __maybe_unused)
tools/perf/util/annotate.c
1325
int nr __maybe_unused, bool current __maybe_unused)
tools/perf/util/annotate.c
2025
void (*obj__set_percent_color)(void *obj, double percent, bool current) = wops->set_percent_color;
tools/perf/util/annotate.c
2026
int (*obj__set_jumps_percent_color)(void *obj, int nr, bool current) = wops->set_jumps_percent_color;
tools/perf/util/annotate.h
196
void (*set_percent_color)(void *obj, double percent, bool current);
tools/perf/util/annotate.h
197
int (*set_jumps_percent_color)(void *obj, int nr, bool current);
tools/perf/util/db-export.c
216
struct call_path *current = &dbe->cpr->call_path;
tools/perf/util/db-export.c
266
current = call_path__findnew(dbe->cpr, current,
tools/perf/util/db-export.c
277
if (current == &dbe->cpr->call_path) {
tools/perf/util/db-export.c
282
return current;
tools/power/cpupower/lib/cpufreq.c
292
struct cpufreq_available_governors *current = NULL;
tools/power/cpupower/lib/cpufreq.c
307
if (current) {
tools/power/cpupower/lib/cpufreq.c
308
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
309
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
311
current = current->next;
tools/power/cpupower/lib/cpufreq.c
316
current = first;
tools/power/cpupower/lib/cpufreq.c
318
current->first = first;
tools/power/cpupower/lib/cpufreq.c
319
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
321
current->governor = malloc(i - pos + 1);
tools/power/cpupower/lib/cpufreq.c
322
if (!current->governor)
tools/power/cpupower/lib/cpufreq.c
325
memcpy(current->governor, linebuf + pos, i - pos);
tools/power/cpupower/lib/cpufreq.c
326
current->governor[i - pos] = '\0';
tools/power/cpupower/lib/cpufreq.c
335
current = first->next;
tools/power/cpupower/lib/cpufreq.c
339
first = current;
tools/power/cpupower/lib/cpufreq.c
366
struct cpufreq_available_frequencies *current = NULL;
tools/power/cpupower/lib/cpufreq.c
384
if (current) {
tools/power/cpupower/lib/cpufreq.c
385
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
386
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
388
current = current->next;
tools/power/cpupower/lib/cpufreq.c
393
current = first;
tools/power/cpupower/lib/cpufreq.c
395
current->first = first;
tools/power/cpupower/lib/cpufreq.c
396
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
400
if (sscanf(one_value, "%lu", &current->frequency) != 1)
tools/power/cpupower/lib/cpufreq.c
411
current = first->next;
tools/power/cpupower/lib/cpufreq.c
413
first = current;
tools/power/cpupower/lib/cpufreq.c
422
struct cpufreq_available_frequencies *current = NULL;
tools/power/cpupower/lib/cpufreq.c
440
if (current) {
tools/power/cpupower/lib/cpufreq.c
441
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
442
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
444
current = current->next;
tools/power/cpupower/lib/cpufreq.c
449
current = first;
tools/power/cpupower/lib/cpufreq.c
451
current->first = first;
tools/power/cpupower/lib/cpufreq.c
452
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
456
if (sscanf(one_value, "%lu", &current->frequency) != 1)
tools/power/cpupower/lib/cpufreq.c
467
current = first->next;
tools/power/cpupower/lib/cpufreq.c
469
first = current;
tools/power/cpupower/lib/cpufreq.c
498
struct cpufreq_affected_cpus *current = NULL;
tools/power/cpupower/lib/cpufreq.c
515
if (current) {
tools/power/cpupower/lib/cpufreq.c
516
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
517
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
519
current = current->next;
tools/power/cpupower/lib/cpufreq.c
524
current = first;
tools/power/cpupower/lib/cpufreq.c
526
current->first = first;
tools/power/cpupower/lib/cpufreq.c
527
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
532
if (sscanf(one_value, "%u", &current->cpu) != 1)
tools/power/cpupower/lib/cpufreq.c
543
current = first->next;
tools/power/cpupower/lib/cpufreq.c
545
first = current;
tools/power/cpupower/lib/cpufreq.c
729
struct cpufreq_stats *current = NULL;
tools/power/cpupower/lib/cpufreq.c
748
if (current) {
tools/power/cpupower/lib/cpufreq.c
749
current->next = malloc(sizeof(*current));
tools/power/cpupower/lib/cpufreq.c
750
if (!current->next)
tools/power/cpupower/lib/cpufreq.c
752
current = current->next;
tools/power/cpupower/lib/cpufreq.c
757
current = first;
tools/power/cpupower/lib/cpufreq.c
759
current->first = first;
tools/power/cpupower/lib/cpufreq.c
760
current->next = NULL;
tools/power/cpupower/lib/cpufreq.c
765
&current->frequency,
tools/power/cpupower/lib/cpufreq.c
766
&current->time_in_state) != 2)
tools/power/cpupower/lib/cpufreq.c
769
*total_time = *total_time + current->time_in_state;
tools/power/cpupower/lib/cpufreq.c
778
current = first->next;
tools/power/cpupower/lib/cpufreq.c
780
first = current;
tools/spi/spidev_test.c
516
struct timespec current;
tools/spi/spidev_test.c
520
clock_gettime(CLOCK_MONOTONIC, &current);
tools/spi/spidev_test.c
521
if (current.tv_sec - last_stat.tv_sec > interval) {
tools/spi/spidev_test.c
523
last_stat = current;
tools/testing/selftests/arm64/signal/test_signals.c
19
struct tdescr *current = &tde;
tools/testing/selftests/arm64/signal/test_signals.c
33
ksft_print_msg("%s :: %s\n", current->name, current->descr);
tools/testing/selftests/arm64/signal/test_signals.c
34
if (test_setup(current) && test_init(current)) {
tools/testing/selftests/arm64/signal/test_signals.c
35
test_run(current);
tools/testing/selftests/arm64/signal/test_signals.c
36
test_cleanup(current);
tools/testing/selftests/arm64/signal/test_signals.c
38
test_result(current);
tools/testing/selftests/arm64/signal/test_signals.c
41
exit(current->result);
tools/testing/selftests/arm64/signal/test_signals_utils.c
107
default_result(current, 1);
tools/testing/selftests/arm64/signal/test_signals_utils.c
173
default_result(current, 1);
tools/testing/selftests/arm64/signal/test_signals_utils.c
23
extern struct tdescr *current;
tools/testing/selftests/arm64/signal/test_signals_utils.c
245
if (current->sig_unsupp && signum == current->sig_unsupp &&
tools/testing/selftests/arm64/signal/test_signals_utils.c
246
handle_signal_unsupported(current, si, uc)) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
248
} else if (current->sig_trig && signum == current->sig_trig &&
tools/testing/selftests/arm64/signal/test_signals_utils.c
249
handle_signal_trigger(current, si, uc)) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
251
} else if (current->sig_ok && signum == current->sig_ok &&
tools/testing/selftests/arm64/signal/test_signals_utils.c
252
handle_signal_ok(current, si, uc)) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
254
} else if (signum == sig_copyctx && current->live_uc &&
tools/testing/selftests/arm64/signal/test_signals_utils.c
255
handle_signal_copyctx(current, si, uc)) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
258
if (signum == SIGALRM && current->timeout) {
tools/testing/selftests/arm64/signal/test_signals_utils.c
265
default_result(current, 1);
tools/testing/selftests/arm64/signal/test_signals_utils.c
387
assert(current);
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
347
struct test_desc *current;
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
355
current = &tests[i];
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
356
if (!test__start_subtest(current->name))
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
358
if (ASSERT_OK(test_setup(skel, current), "subtest setup")) {
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
359
test_basic_access(current);
tools/testing/selftests/bpf/prog_tests/bpftool_maps_access.c
360
test_cleanup(current);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1987
struct task_struct *current = NULL;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1993
global_call_bpf_dynptr((const struct bpf_dynptr *)current);
tools/testing/selftests/bpf/progs/lru_bug.c
34
struct task_struct *current;
tools/testing/selftests/bpf/progs/lru_bug.c
41
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/lru_bug.c
42
v->ptr = current;
tools/testing/selftests/bpf/progs/lru_bug.c
43
pid = current->pid;
tools/testing/selftests/bpf/progs/map_kptr.c
511
struct task_struct *current;
tools/testing/selftests/bpf/progs/map_kptr.c
514
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/map_kptr.c
515
if (!current)
tools/testing/selftests/bpf/progs/map_kptr.c
517
v = bpf_task_storage_get(&task_ls_map, current, NULL, 0);
tools/testing/selftests/bpf/progs/map_kptr.c
520
v = bpf_task_storage_get(&task_ls_map, current, NULL, BPF_LOCAL_STORAGE_GET_F_CREATE);
tools/testing/selftests/bpf/progs/map_kptr.c
529
struct task_struct *current;
tools/testing/selftests/bpf/progs/map_kptr.c
532
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/map_kptr.c
533
if (!current)
tools/testing/selftests/bpf/progs/map_kptr.c
535
v = bpf_task_storage_get(&task_ls_map, current, NULL, 0);
tools/testing/selftests/bpf/progs/map_kptr.c
544
struct task_struct *current;
tools/testing/selftests/bpf/progs/map_kptr.c
547
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/map_kptr.c
548
if (!current)
tools/testing/selftests/bpf/progs/map_kptr.c
550
v = bpf_task_storage_get(&task_ls_map, current, NULL, 0);
tools/testing/selftests/bpf/progs/map_kptr.c
555
return bpf_task_storage_delete(&task_ls_map, current);
tools/testing/selftests/bpf/progs/task_kfunc_success.c
268
struct task_struct *current, *acquired;
tools/testing/selftests/bpf/progs/task_kfunc_success.c
273
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/task_kfunc_success.c
274
acquired = bpf_task_acquire(current);
tools/testing/selftests/bpf/progs/task_kfunc_success.c
373
struct task_struct *current, *v_task;
tools/testing/selftests/bpf/progs/task_kfunc_success.c
381
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/task_kfunc_success.c
384
if (current != v_task)
tools/testing/selftests/bpf/progs/test_task_pt_regs.c
20
struct task_struct *current;
tools/testing/selftests/bpf/progs/test_task_pt_regs.c
23
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/test_task_pt_regs.c
24
regs = (struct pt_regs *) bpf_task_pt_regs(current);
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
21
struct task_struct *current, *ref_1, *ref_2;
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
23
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_kfunc_prog_types.c
24
ref_1 = bpf_task_from_pid(current->pid);
tools/testing/selftests/bpf/progs/verifier_vfs_reject.c
119
struct task_struct *current;
tools/testing/selftests/bpf/progs/verifier_vfs_reject.c
121
current = bpf_get_current_task_btf();
tools/testing/selftests/bpf/progs/verifier_vfs_reject.c
125
pwd = &current->fs->pwd;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
528
trace_bpf_testmod_test_read(current, &ctx);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
571
trace_bpf_testmod_test_write_bare_tp(current, &ctx);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
942
err = sock_create_kern(current->nsproxy->net_ns, args->af, args->type,
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
104
current = cg_read_long(test_group, "memory.current");
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
105
if (current - old_current >= MB(2)) {
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
108
ksft_print_msg("before: %ld, after: %ld\n", old_current, current);
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
117
current = cg_read_long(test_group, "memory.current");
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
118
if (current - old_current >= MB(2)) {
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
120
ksft_print_msg("before: %ld, after: %ld\n", old_current, current);
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
123
old_current = current;
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
128
current = cg_read_long(test_group, "memory.current");
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
129
if (!values_close(expected_current, current, 5)) {
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
133
expected_current, current);
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
139
current = cg_read_long(test_group, "memory.current");
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
141
if (!values_close(expected_current, current, 5)) {
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
145
expected_current, current);
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
151
current = cg_read_long(test_group, "memory.current");
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
153
if (!values_close(expected_current, current, 5)) {
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
157
expected_current, current);
tools/testing/selftests/cgroup/test_hugetlb_memcg.c
99
long old_current, expected_current, current;
tools/testing/selftests/cgroup/test_kmem.c
170
long current, anon, file, kernel, sock, sum;
tools/testing/selftests/cgroup/test_kmem.c
187
current = cg_read_long(parent, "memory.current");
tools/testing/selftests/cgroup/test_kmem.c
192
if (current < 0 || anon < 0 || file < 0 || kernel < 0 || sock < 0)
tools/testing/selftests/cgroup/test_kmem.c
196
if (labs(sum - current) < MAX_VMSTAT_ERROR) {
tools/testing/selftests/cgroup/test_kmem.c
199
printf("memory.current = %ld\n", current);
tools/testing/selftests/cgroup/test_kmem.c
356
long current, percpu;
tools/testing/selftests/cgroup/test_kmem.c
380
current = cg_read_long(parent, "memory.current");
tools/testing/selftests/cgroup/test_kmem.c
383
if (current > 0 && percpu > 0 && labs(current - percpu) <
tools/testing/selftests/cgroup/test_kmem.c
388
current, percpu);
tools/testing/selftests/cgroup/test_kmem.c
57
long slab0, slab1, current;
tools/testing/selftests/cgroup/test_kmem.c
82
current = cg_read_long(cg, "memory.current");
tools/testing/selftests/cgroup/test_kmem.c
83
if (current < 0)
tools/testing/selftests/cgroup/test_kmem.c
86
if (slab1 < slab0 / 2 && current < slab0 / 2)
tools/testing/selftests/cgroup/test_memcontrol.c
1351
long current, sock;
tools/testing/selftests/cgroup/test_memcontrol.c
1356
current = cg_read_long(cgroup, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
1359
if (current < 0 || sock < 0)
tools/testing/selftests/cgroup/test_memcontrol.c
1363
if (values_close(current - allocated, sock, 10)) {
tools/testing/selftests/cgroup/test_memcontrol.c
177
long anon, current;
tools/testing/selftests/cgroup/test_memcontrol.c
189
current = cg_read_long(cgroup, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
190
if (current < size)
tools/testing/selftests/cgroup/test_memcontrol.c
193
if (!values_close(size, current, 3))
tools/testing/selftests/cgroup/test_memcontrol.c
200
if (!values_close(anon, current, 3))
tools/testing/selftests/cgroup/test_memcontrol.c
213
long current, file;
tools/testing/selftests/cgroup/test_memcontrol.c
223
current = cg_read_long(cgroup, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
224
if (current < size)
tools/testing/selftests/cgroup/test_memcontrol.c
231
if (!values_close(file, current, 10))
tools/testing/selftests/cgroup/test_memcontrol.c
249
long current, peak, peak_reset;
tools/testing/selftests/cgroup/test_memcontrol.c
262
current = cg_read_long(memcg, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
263
if (current != 0)
tools/testing/selftests/cgroup/test_memcontrol.c
488
long current;
tools/testing/selftests/cgroup/test_memcontrol.c
597
current = min ? MB(50) : MB(30);
tools/testing/selftests/cgroup/test_memcontrol.c
598
if (!values_close(cg_read_long(parent[1], "memory.current"), current, 3))
tools/testing/selftests/cgroup/test_memcontrol.c
670
long current, high, max;
tools/testing/selftests/cgroup/test_memcontrol.c
685
current = cg_read_long(cgroup, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
686
if (!values_close(current, MB(30), 5))
tools/testing/selftests/cgroup/test_memcontrol.c
831
long current, max;
tools/testing/selftests/cgroup/test_memcontrol.c
856
current = cg_read_long(memcg, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
857
if (current > MB(30) || !current)
tools/testing/selftests/cgroup/test_memcontrol.c
892
long current, to_reclaim;
tools/testing/selftests/cgroup/test_memcontrol.c
896
current = cg_read_long(memcg, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
898
if (current < goal || values_close(current, goal, 3))
tools/testing/selftests/cgroup/test_memcontrol.c
904
to_reclaim = current - goal;
tools/testing/selftests/cgroup/test_memcontrol.c
925
long current, expected_usage;
tools/testing/selftests/cgroup/test_memcontrol.c
934
current = cg_read_long(memcg, "memory.current");
tools/testing/selftests/cgroup/test_memcontrol.c
935
if (current != 0)
tools/testing/selftests/livepatch/test_modules/test_klp_syscall.c
50
if (current->pid == klp_pids[i]) {
tools/testing/selftests/livepatch/test_modules/test_klp_syscall.c
59
return task_tgid_vnr(current);
tools/testing/selftests/net/bench/page_pool/time_bench.c
263
set_cpus_allowed_ptr(current, &newmask);
tools/testing/selftests/powerpc/pmu/count_instructions.c
76
u64 current, overhead;
tools/testing/selftests/powerpc/pmu/count_instructions.c
84
current = events[0].result.value;
tools/testing/selftests/powerpc/pmu/count_instructions.c
85
if (current < overhead) {
tools/testing/selftests/powerpc/pmu/count_instructions.c
86
printf("Replacing overhead %llu with %llu\n", overhead, current);
tools/testing/selftests/powerpc/pmu/count_instructions.c
87
overhead = current;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
81
u64 current, overhead;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
89
current = events[0].result.value;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
90
if (current < overhead) {
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
91
printf("Replacing overhead %llu with %llu\n", overhead, current);
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
92
overhead = current;
tools/testing/selftests/powerpc/pmu/ebb/instruction_count_test.c
72
uint64_t current, overhead;
tools/testing/selftests/powerpc/pmu/ebb/instruction_count_test.c
80
current = event->result.value;
tools/testing/selftests/powerpc/pmu/ebb/instruction_count_test.c
81
if (current < overhead) {
tools/testing/selftests/powerpc/pmu/ebb/instruction_count_test.c
82
printf("Replacing overhead %lu with %lu\n", overhead, current);
tools/testing/selftests/powerpc/pmu/ebb/instruction_count_test.c
83
overhead = current;
tools/testing/selftests/powerpc/pmu/lib.c
179
long current;
tools/testing/selftests/powerpc/pmu/lib.c
181
err = read_long(PARANOID_PATH, &current, 10);
tools/testing/selftests/powerpc/pmu/lib.c
187
return current < level;
tools/testing/vma/include/dup.h
1262
if (mm_flags_test(MMF_HAS_MDWE, current->mm))
tools/testing/vma/include/dup.h
314
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
tools/testing/vma/include/dup.h
353
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
tools/testing/vma/tests/mmap.c
10
current->mm = &mm;
virt/kvm/dirty_ring.c
134
if (signal_pending(current))
virt/kvm/guest_memfd.c
271
if (signal_pending(current)) {
virt/kvm/guest_memfd.c
888
if (signal_pending(current)) {
virt/kvm/kvm_main.c
1021
snprintf(dir_name, sizeof(dir_name), "%d-%s", task_pid_nr(current), fdname);
virt/kvm/kvm_main.c
1115
mmgrab(current->mm);
virt/kvm/kvm_main.c
1116
kvm->mm = current->mm;
virt/kvm/kvm_main.c
1144
task_pid_nr(current));
virt/kvm/kvm_main.c
1220
mmu_notifier_unregister(&kvm->mmu_notifier, current->mm);
virt/kvm/kvm_main.c
1236
mmdrop(current->mm);
virt/kvm/kvm_main.c
2702
mmap_read_lock(current->mm);
virt/kvm/kvm_main.c
2703
vma = find_vma(current->mm, addr);
virt/kvm/kvm_main.c
2710
mmap_read_unlock(current->mm);
virt/kvm/kvm_main.c
2968
r = fixup_user_fault(current->mm, kfp->hva,
virt/kvm/kvm_main.c
3014
mmap_read_lock(current->mm);
virt/kvm/kvm_main.c
3016
vma = vma_lookup(current->mm, kfp->hva);
virt/kvm/kvm_main.c
3033
mmap_read_unlock(current->mm);
virt/kvm/kvm_main.c
3574
sigprocmask(SIG_SETMASK, &vcpu->sigset, &current->real_blocked);
virt/kvm/kvm_main.c
3582
sigprocmask(SIG_SETMASK, &current->real_blocked, NULL);
virt/kvm/kvm_main.c
3583
sigemptyset(&current->real_blocked);
virt/kvm/kvm_main.c
3633
if (signal_pending(current))
virt/kvm/kvm_main.c
4360
if (signal_pending(current)) {
virt/kvm/kvm_main.c
4421
if (vcpu->kvm->mm != current->mm || vcpu->kvm->vm_dead)
virt/kvm/kvm_main.c
4460
if (unlikely(oldpid != task_pid(current))) {
virt/kvm/kvm_main.c
4468
newpid = get_task_pid(current, PIDTYPE_PID);
virt/kvm/kvm_main.c
464
task_pid_nr(current), id);
virt/kvm/kvm_main.c
4670
if (vcpu->kvm->mm != current->mm || vcpu->kvm->vm_dead)
virt/kvm/kvm_main.c
4736
if (dev->kvm->mm != current->mm || dev->kvm->vm_dead)
virt/kvm/kvm_main.c
5161
if (kvm->mm != current->mm || kvm->vm_dead)
virt/kvm/kvm_main.c
5425
if (kvm->mm != current->mm || kvm->vm_dead)
virt/kvm/kvm_main.c
6323
kvm->userspace_pid = task_pid_nr(current);
virt/kvm/kvm_main.c
6402
if (task_is_runnable(current) && vcpu->wants_to_run) {
virt/kvm/kvm_main.c
900
return mmu_notifier_register(&kvm->mmu_notifier, current->mm);