Symbol: task_cpu
arch/mips/kernel/process.c
852
cpumask_set_cpu(task_cpu(t), &process_cpus);
arch/parisc/kernel/traps.c
154
level, task_cpu(current), cr30, cr31);
arch/powerpc/kernel/process.c
2132
unsigned long cpu = task_cpu(p);
arch/powerpc/kernel/process.c
2153
unsigned long cpu = task_cpu(p);
arch/powerpc/kernel/process.c
2187
unsigned long cpu = task_cpu(p);
arch/powerpc/xmon/xmon.c
3264
state, task_cpu(tsk),
fs/fuse/dev_uring.c
1244
qid = task_cpu(current);
fs/proc/array.c
636
seq_put_decimal_ll(m, " ", task_cpu(task));
fs/resctrl/rdtgroup.c
3018
cpumask_set_cpu(task_cpu(t), mask);
fs/resctrl/rdtgroup.c
613
smp_call_function_single(task_cpu(t), _update_task_closid_rmid, t, 1);
include/linux/kdb.h
188
unsigned int cpu = task_cpu(p);
include/linux/rseq_entry.h
633
.cpu_id = task_cpu(t),
include/linux/sched.h
2314
return READ_ONCE(owner->on_cpu) && !vcpu_is_preempted(task_cpu(owner));
include/linux/sched.h
2373
return task_cpu(t);
include/linux/sched/topology.h
244
return cpu_to_node(task_cpu(p));
include/trace/events/sched.h
158
__entry->target_cpu = task_cpu(p);
include/trace/events/sched.h
290
__entry->orig_cpu = task_cpu(p);
kernel/events/core.c
123
ret = smp_call_function_single(task_cpu(p), remote_function,
kernel/events/core.c
82
if (task_cpu(p) != smp_processor_id())
kernel/rcu/rcutorture.c
2835
wtp == NULL ? -1 : (int)task_cpu(wtp));
kernel/rcu/tasks.h
1109
cpu = task_cpu(t);
kernel/rcu/tasks.h
999
cpu = task_cpu(t);
kernel/rcu/tree.c
770
cpu = task_cpu(t);
kernel/rcu/tree_nocb.h
1520
rdp->nocb_gp_kthread ? (int)task_cpu(rdp->nocb_gp_kthread) : -1,
kernel/rcu/tree_nocb.h
1577
rdp->nocb_cb_kthread ? (int)task_cpu(rdp->nocb_cb_kthread) : -1,
kernel/rcu/tree_stall.h
467
cpu = task_cpu(rcuc);
kernel/rcu/tree_stall.h
576
cpu = gpk ? task_cpu(gpk) : -1;
kernel/rcu/tree_stall.h
619
cpu = task_cpu(gpk);
kernel/rseq.c
297
ids.cpu_id = task_cpu(t);
kernel/sched/core.c
10630
mm_cid_transit_to_cpu(t, per_cpu_ptr(mm->mm_cid.pcpu, task_cpu(t)));
kernel/sched/core.c
2182
return cpu_curr(task_cpu(p)) == p;
kernel/sched/core.c
2461
WARN_ON_ONCE(task_cpu(p) != new_cpu);
kernel/sched/core.c
2562
if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask))
kernel/sched/core.c
2595
if (cpumask_test_cpu(task_cpu(p), p->cpus_ptr)) {
kernel/sched/core.c
2610
stop_one_cpu_nowait(task_cpu(p), migration_cpu_stop,
kernel/sched/core.c
2875
if (cpumask_test_cpu(task_cpu(p), &p->cpus_mask) ||
kernel/sched/core.c
3063
!cpumask_test_cpu(task_cpu(p), ctx->new_mask))) {
kernel/sched/core.c
3285
if (task_cpu(p) != new_cpu) {
kernel/sched/core.c
3344
if (task_cpu(arg->dst_task) != arg->dst_cpu)
kernel/sched/core.c
3347
if (task_cpu(arg->src_task) != arg->src_cpu)
kernel/sched/core.c
3418
int cpu = task_cpu(p);
kernel/sched/core.c
348
int cpu = task_cpu(p);
kernel/sched/core.c
3538
cpu = select_fallback_rq(task_cpu(p), p);
kernel/sched/core.c
3751
if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq)))
kernel/sched/core.c
4216
ttwu_queue_wakelist(p, task_cpu(p), wake_flags))
kernel/sched/core.c
4231
if (task_cpu(p) != cpu) {
kernel/sched/core.c
4246
ttwu_stat(p, task_cpu(p), wake_flags);
kernel/sched/core.c
4780
p->recent_used_cpu = task_cpu(p);
kernel/sched/core.c
4781
__set_task_cpu(p, select_task_rq(p, task_cpu(p), &wake_flags));
kernel/sched/core.c
5418
dest_cpu = p->sched_class->select_task_rq(p, task_cpu(p), WF_EXEC);
kernel/sched/core.c
5427
stop_one_cpu(task_cpu(p), migration_cpu_stop, &arg);
kernel/sched/core.c
6636
if (task_cpu(owner) != this_cpu) {
kernel/sched/core.c
6661
if (!task_on_rq_queued(owner) || task_cpu(owner) != this_cpu)
kernel/sched/core.c
8031
int curr_cpu = task_cpu(p);
kernel/sched/core.c
8506
if (task_cpu(p) != cpu)
kernel/sched/core.c
9159
set_task_rq(tsk, task_cpu(tsk));
kernel/sched/cpuacct.c
338
unsigned int cpu = task_cpu(tsk);
kernel/sched/cpudeadline.c
138
(cpu == task_cpu(p) && cap == max_cap)) {
kernel/sched/deadline.c
1984
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
kernel/sched/deadline.c
1993
__dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/deadline.c
2720
int cpu = task_cpu(task);
kernel/sched/deadline.c
2811
WARN_ON_ONCE(rq->cpu != task_cpu(p));
kernel/sched/deadline.c
3118
__dl_sub(src_dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/deadline.c
3544
int cpus, err = -1, cpu = task_cpu(p);
kernel/sched/deadline.c
442
struct dl_bw *dl_b = dl_bw_of(task_cpu(p));
kernel/sched/deadline.c
447
__dl_sub(dl_b, dl_se->dl_bw, dl_bw_cpus(task_cpu(p)));
kernel/sched/debug.c
894
if (task_cpu(p) != rq_cpu)
kernel/sched/ext.c
1738
WARN_ON_ONCE(task_cpu(p) == cpu);
kernel/sched/ext.c
1755
p->comm, p->pid, task_cpu(p), cpu);
kernel/sched/ext.c
251
return sch->global_dsqs[cpu_to_node(task_cpu(p))];
kernel/sched/ext.c
7202
return task_cpu(p);
kernel/sched/fair.c
10741
if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
13670
set_task_rq(p, task_cpu(p));
kernel/sched/fair.c
2567
.src_cpu = task_cpu(p),
kernel/sched/fair.c
3702
int src_nid = cpu_to_node(task_cpu(p));
kernel/sched/fair.c
7246
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
7269
if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/fair.c
8022
if (p && task_cpu(p) == cpu && dst_cpu != cpu)
kernel/sched/fair.c
8024
else if (p && task_cpu(p) != cpu && dst_cpu == cpu)
kernel/sched/fair.c
8095
if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time))
kernel/sched/idle.c
461
return task_cpu(p); /* IDLE tasks as never migrated */
kernel/sched/psi.c
1008
int cpu = task_cpu(curr);
kernel/sched/psi.c
900
task->pid, task->comm, task_cpu(task),
kernel/sched/psi.c
911
int cpu = task_cpu(task);
kernel/sched/psi.c
930
int cpu = task_cpu(prev);
kernel/sched/rt.c
1768
int cpu = task_cpu(task);
kernel/sched/rt.c
1864
BUG_ON(rq->cpu != task_cpu(p));
kernel/sched/sched.h
1392
#define task_rq(p) cpu_rq(task_cpu(p))
kernel/sched/sched.h
2574
int (*select_task_rq)(struct task_struct *p, int task_cpu, int flags);
kernel/sched/stop_task.c
15
return task_cpu(p); /* stop tasks as never migrate */
kernel/stop_machine.c
58
struct cpu_stopper *stopper = per_cpu_ptr(&cpu_stopper, task_cpu(task));
kernel/time/tick-sched.c
477
cpu = task_cpu(tsk);
kernel/trace/trace_osnoise.c
1200
long cpu = task_cpu(p);
kernel/trace/trace_sched_wakeup.c
408
entry->next_cpu = task_cpu(next);
kernel/trace/trace_sched_wakeup.c
434
entry->next_cpu = task_cpu(wakee);
kernel/trace/trace_sched_wakeup.c
579
wakeup_cpu = task_cpu(p);