arch/arc/kernel/smp.c
252
unsigned long __percpu *ipi_data_ptr = per_cpu_ptr(&ipi_data, cpu);
arch/arc/kernel/smp.c
391
int *dev = per_cpu_ptr(&ipi_dev, cpu);
arch/arm/xen/enlighten.c
157
vcpup = per_cpu_ptr(xen_vcpu_info, cpu);
arch/arm64/include/asm/kvm_asm.h
172
? per_cpu_ptr(&sym, cpu) \
arch/arm64/kernel/paravirt.c
50
reg = per_cpu_ptr(&stolen_time_region, cpu);
arch/arm64/kernel/smp.c
76
#define get_ipi_desc(__cpu, __ipi) (per_cpu_ptr(&pcpu_ipi_desc, __cpu)->descs[__ipi])
arch/arm64/kernel/topology.c
197
amu_sample = per_cpu_ptr(&cpu_amu_samples, cpu);
arch/arm64/kernel/topology.c
65
struct amu_cntr_sample *amu_sample = per_cpu_ptr(&cpu_amu_samples, cpu);
arch/arm64/kvm/hyp/nvhe/mm.c
309
.arg = per_cpu_ptr(&fixmap_slots, cpu),
arch/arm64/kvm/hyp/nvhe/psci-relay.c
129
boot_args = per_cpu_ptr(&cpu_on_args, cpu_id);
arch/arm64/kvm/hyp/nvhe/psci-relay.c
130
init_params = per_cpu_ptr(&kvm_init_params, cpu_id);
arch/arm64/kvm/hyp/nvhe/setup.c
145
struct kvm_nvhe_init_params *params = per_cpu_ptr(&kvm_init_params, i);
arch/arm64/kvm/hyp/nvhe/setup.c
167
params = per_cpu_ptr(&kvm_init_params, i);
arch/arm64/kvm/hyp/nvhe/setup.c
84
struct kvm_host_data *host_data = per_cpu_ptr(&kvm_host_data, i);
arch/arm64/kvm/mmu.c
1001
*per_cpu_ptr(mmu->last_vcpu_ran, cpu) = -1;
arch/arm64/net/bpf_jit_comp.c
1975
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
arch/arm64/net/bpf_jit_comp.c
1990
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
arch/csky/include/asm/asid.h
31
#define active_asid(info, cpu) *per_cpu_ptr((info)->active, cpu)
arch/csky/kernel/smp.c
116
per_cpu_ptr(&ipi_data, cpu)->stats[i]);
arch/csky/kernel/smp.c
94
set_bit(operation, &per_cpu_ptr(&ipi_data, i)->bits);
arch/csky/mm/asid.c
16
#define reserved_asid(info, cpu) *per_cpu_ptr((info)->reserved, cpu)
arch/loongarch/kernel/kgdb.c
470
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/loongarch/kernel/kgdb.c
482
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/loongarch/kernel/kgdb.c
498
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/loongarch/kernel/kgdb.c
595
bp = *per_cpu_ptr(breakinfo[i].pev, cpu);
arch/loongarch/kernel/kgdb.c
617
bp = *per_cpu_ptr(breakinfo[i].pev, cpu);
arch/loongarch/kernel/kgdb.c
645
bp = *per_cpu_ptr(breakinfo[i].pev, cpu);
arch/loongarch/kernel/kgdb.c
707
pevent = per_cpu_ptr(breakinfo[i].pev, cpu);
arch/loongarch/kvm/main.c
210
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
239
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/main.c
399
context = per_cpu_ptr(vmcs, cpu);
arch/loongarch/kvm/vcpu.c
1608
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/loongarch/kvm/vcpu.c
1630
context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
arch/parisc/kernel/smp.c
347
*per_cpu_ptr(desc->kstat_irqs, cpuid) = (struct irqstat) { };
arch/powerpc/kvm/book3s_hv.c
3267
struct kvm *running = *per_cpu_ptr(&cpu_in_guest, i);
arch/powerpc/kvm/book3s_hv_rm_xics.c
810
raddr = per_cpu_ptr(addr, cpu);
arch/powerpc/lib/qspinlock.c
263
struct qnodes *qnodesp = per_cpu_ptr(&qnodes, prev_cpu);
arch/powerpc/net/bpf_jit_comp.c
138
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
arch/powerpc/net/bpf_jit_comp.c
153
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
arch/powerpc/platforms/pseries/lpar.c
536
disp = per_cpu_ptr(&vcpu_disp_data, cpu);
arch/powerpc/platforms/pseries/lpar.c
571
disp = per_cpu_ptr(&vcpu_disp_data, cpu);
arch/powerpc/xmon/xmon.c
3685
addr = (unsigned long)per_cpu_ptr(ptr, cpu);
arch/riscv/kernel/cpu.c
173
struct riscv_cpuinfo *ci = per_cpu_ptr(&riscv_cpuinfo, cpu_id);
arch/riscv/kernel/cpu.c
181
struct riscv_cpuinfo *ci = per_cpu_ptr(&riscv_cpuinfo, cpu_id);
arch/riscv/kernel/cpu.c
189
struct riscv_cpuinfo *ci = per_cpu_ptr(&riscv_cpuinfo, cpu_id);
arch/riscv/kernel/cpu.c
332
struct riscv_cpuinfo *ci = per_cpu_ptr(&riscv_cpuinfo, cpu_id);
arch/riscv/kernel/paravirt.c
84
struct sbi_sta_struct *st = per_cpu_ptr(&steal_time, cpu);
arch/riscv/kernel/traps_misaligned.c
525
long *mas_ptr = per_cpu_ptr(&misaligned_access_speed, cpu);
arch/riscv/kernel/traps_misaligned.c
537
long *mas_ptr = per_cpu_ptr(&misaligned_access_speed, cpu);
arch/riscv/kvm/aia.c
418
struct aia_hgei_control *hgctrl = per_cpu_ptr(&aia_hgei, cpu);
arch/riscv/kvm/aia.c
434
lc = (gc) ? per_cpu_ptr(gc->local, cpu) : NULL;
arch/riscv/kvm/aia.c
448
struct aia_hgei_control *hgctrl = per_cpu_ptr(&aia_hgei, cpu);
arch/riscv/kvm/aia.c
495
hgctrl = per_cpu_ptr(&aia_hgei, cpu);
arch/riscv/kvm/nacl.c
139
nacl = per_cpu_ptr(&kvm_riscv_nacl, cpu);
arch/riscv/kvm/nacl.c
91
nacl = per_cpu_ptr(&kvm_riscv_nacl, cpu);
arch/s390/kernel/perf_cpum_cf.c
150
struct cpu_cf_ptr *q = per_cpu_ptr(p, cpu);
arch/s390/kernel/perf_cpum_cf.c
222
p = per_cpu_ptr(cpu_cf_root.cfptr, cpu);
arch/s390/kernel/perf_cpum_cf.c
251
p = per_cpu_ptr(cpu_cf_root.cfptr, cpu);
arch/s390/kernel/perf_pai.c
146
struct pai_mapptr *mp = per_cpu_ptr(pai_root[idx].mapptr, cpu);
arch/s390/kernel/perf_pai.c
251
mp = per_cpu_ptr(pai_root[idx].mapptr, cpu);
arch/s390/kernel/processor.c
330
struct cpu_info *c = per_cpu_ptr(&cpu_info, n);
arch/s390/kernel/setup.c
399
lc->pcpu = (unsigned long)per_cpu_ptr(&pcpu_devices, 0);
arch/s390/kernel/smp.c
1028
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
1100
struct cpu *c = per_cpu_ptr(&cpu_devices, cpu);
arch/s390/kernel/smp.c
1107
struct cpu *c = per_cpu_ptr(&cpu_devices, cpu);
arch/s390/kernel/smp.c
1120
struct cpu *c = per_cpu_ptr(&cpu_devices, cpu);
arch/s390/kernel/smp.c
297
pcpu_sigp_retry(per_cpu_ptr(&pcpu_devices, cpu), SIGP_RESTART, 0);
arch/s390/kernel/smp.c
404
if (pcpu_running(per_cpu_ptr(&pcpu_devices, cpu)))
arch/s390/kernel/smp.c
437
struct pcpu *pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
446
if (pcpu_stopped(per_cpu_ptr(&pcpu_devices, cpu)))
arch/s390/kernel/smp.c
477
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
518
pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_call_function_single);
arch/s390/kernel/smp.c
523
pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_call_function_single);
arch/s390/kernel/smp.c
533
pcpu_ec_call(per_cpu_ptr(&pcpu_devices, cpu), ec_schedule);
arch/s390/kernel/smp.c
551
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
724
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
858
struct pcpu *pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
922
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
968
ipl_pcpu = per_cpu_ptr(&pcpu_devices, 0);
arch/s390/kernel/wti.c
104
struct wti_state *st = per_cpu_ptr(&wti_state, cpu);
arch/s390/kernel/wti.c
133
st = per_cpu_ptr(&wti_state, cpu);
arch/s390/kernel/wti.c
144
struct wti_state *st = per_cpu_ptr(&wti_state, cpu);
arch/s390/kernel/wti.c
179
st = per_cpu_ptr(&wti_state, cpu);
arch/s390/lib/spinlock.c
142
return per_cpu_ptr(&spin_wait[ix], cpu - 1);
arch/s390/lib/spinlock.c
85
node = per_cpu_ptr(&spin_wait[0], cpu);
arch/x86/events/amd/uncore.c
1004
*per_cpu_ptr(uncore->info, cpu) = info;
arch/x86/events/amd/uncore.c
1027
info = *per_cpu_ptr(uncore->info, i);
arch/x86/events/amd/uncore.c
166
struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/amd/uncore.c
184
struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/amd/uncore.c
205
struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/amd/uncore.c
252
struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/amd/uncore.c
280
ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/amd/uncore.c
454
union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu);
arch/x86/events/amd/uncore.c
461
union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu);
arch/x86/events/amd/uncore.c
468
union amd_uncore_info *info = per_cpu_ptr(uncore->info, cpu);
arch/x86/events/amd/uncore.c
483
ctx = *per_cpu_ptr(pmu->ctx, cpu);
arch/x86/events/amd/uncore.c
495
*per_cpu_ptr(pmu->ctx, cpu) = NULL;
arch/x86/events/amd/uncore.c
513
*per_cpu_ptr(pmu->ctx, cpu) = NULL;
arch/x86/events/amd/uncore.c
525
prev = *per_cpu_ptr(pmu->ctx, j);
arch/x86/events/amd/uncore.c
558
*per_cpu_ptr(pmu->ctx, cpu) = curr;
arch/x86/events/amd/uncore.c
580
curr = *per_cpu_ptr(pmu->ctx, cpu);
arch/x86/events/amd/uncore.c
586
next = *per_cpu_ptr(pmu->ctx, j);
arch/x86/events/amd/uncore.c
710
*per_cpu_ptr(uncore->info, cpu) = info;
arch/x86/events/amd/uncore.c
844
*per_cpu_ptr(uncore->info, cpu) = info;
arch/x86/events/amd/uncore.c
941
struct amd_uncore_ctx *ctx = *per_cpu_ptr(pmu->ctx, event->cpu);
arch/x86/events/intel/ds.c
1079
struct cpu_hw_events *cpuc = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/ds.c
849
struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/ds.c
892
struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/ds.c
913
struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/ds.c
942
struct cpu_hw_events *hwev = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/lbr.c
630
cpuc = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/events/intel/lbr.c
649
cpuc = per_cpu_ptr(&cpu_hw_events, cpu);
arch/x86/kernel/acpi/cstate.c
183
percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu);
arch/x86/kernel/acpi/cstate.c
214
percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu);
arch/x86/kernel/acpi/cstate.c
224
percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu);
arch/x86/kernel/alternative.c
2945
atomic_set_release(per_cpu_ptr(&text_poke_array_refs, i), 1);
arch/x86/kernel/alternative.c
3067
atomic_t *refs = per_cpu_ptr(&text_poke_array_refs, i);
arch/x86/kernel/apic/vector.c
1099
struct vector_cleanup *cl = per_cpu_ptr(&vector_cleanup, cpu);
arch/x86/kernel/apic/x2apic_savic.c
139
apic_set_reg(per_cpu_ptr(savic_page, cpu), SAVIC_NMI_REQ, 1);
arch/x86/kernel/apic/x2apic_savic.c
33
return &per_cpu_ptr(savic_page, cpu)->regs[offset];
arch/x86/kernel/cpu/aperfmperf.c
388
per_cpu_ptr(arch_cpu_scale, cpu)->capacity = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/aperfmperf.c
389
per_cpu_ptr(arch_cpu_scale, cpu)->freq_ratio = arch_max_freq_ratio;
arch/x86/kernel/cpu/aperfmperf.c
419
WRITE_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->capacity,
arch/x86/kernel/cpu/aperfmperf.c
421
WRITE_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->freq_ratio,
arch/x86/kernel/cpu/aperfmperf.c
431
return READ_ONCE(per_cpu_ptr(arch_cpu_scale, cpu)->capacity);
arch/x86/kernel/cpu/aperfmperf.c
508
struct aperfmperf *s = per_cpu_ptr(&cpu_samples, cpu);
arch/x86/kernel/cpu/bus_lock.c
222
struct delayed_work *work = per_cpu_ptr(&sl_reenable, cpu);
arch/x86/kernel/cpu/bus_lock.c
275
work = saved_sld_mitigate ? &sl_reenable_unlock : per_cpu_ptr(&sl_reenable, cpu);
arch/x86/kernel/cpu/debugfs.c
13
struct cpuinfo_x86 *c = per_cpu_ptr(&cpu_info, cpu);
arch/x86/kernel/irq_64.c
38
char *stack = (char *)per_cpu_ptr(&irq_stack_backing_store, cpu);
arch/x86/kernel/irq_64.c
64
void *va = per_cpu_ptr(&irq_stack_backing_store, cpu);
arch/x86/kernel/kgdb.c
208
bp = *per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/x86/kernel/kgdb.c
237
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/x86/kernel/kgdb.c
249
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/x86/kernel/kgdb.c
264
pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu);
arch/x86/kernel/kgdb.c
304
bp = *per_cpu_ptr(breakinfo[i].pev, cpu);
arch/x86/kernel/kgdb.c
397
bp = *per_cpu_ptr(breakinfo[i].pev, cpu);
arch/x86/kernel/kgdb.c
666
pevent = per_cpu_ptr(breakinfo[i].pev, cpu);
arch/x86/kernel/kvm.c
699
zalloc_cpumask_var_node(per_cpu_ptr(&__pv_cpu_mask, cpu),
arch/x86/kernel/nmi.c
646
nsp = per_cpu_ptr(&nmi_stats, cpu);
arch/x86/kernel/nmi.c
666
nsp = per_cpu_ptr(&nmi_stats, cpu);
arch/x86/kernel/smpboot.c
1460
md = per_cpu_ptr(&mwait_cpu_dead, cpu);
arch/x86/kernel/tsc.c
179
c2n = per_cpu_ptr(&cyc2ns, cpu);
arch/x86/kernel/tsc.c
227
c2n = per_cpu_ptr(&cyc2ns, cpu);
arch/x86/kernel/tsc_sync.c
217
ref = per_cpu_ptr(&tsc_adjust, refcpu);
arch/x86/kvm/svm/sev.c
283
sd = per_cpu_ptr(&svm_data, cpu);
arch/x86/kvm/svm/sev.c
3561
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, cpu);
arch/x86/kvm/svm/svm.c
1381
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, vcpu->cpu);
arch/x86/kvm/svm/svm.c
3643
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, vcpu->cpu);
arch/x86/kvm/svm/svm.c
4258
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, vcpu->cpu);
arch/x86/kvm/svm/svm.c
525
sd = per_cpu_ptr(&svm_data, me);
arch/x86/kvm/svm/svm.c
581
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, cpu);
arch/x86/kvm/svm/svm.c
594
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, cpu);
arch/x86/kvm/x86.c
12857
cmpxchg(per_cpu_ptr(&last_vcpu, cpu), vcpu, NULL);
arch/x86/mm/cpu_entry_area.c
141
struct exception_stacks *estacks = per_cpu_ptr(&exception_stacks, cpu);
arch/x86/mm/cpu_entry_area.c
202
per_cpu_ptr(&entry_stack_storage, cpu), 1,
arch/x86/net/bpf_jit_comp.c
3681
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
arch/x86/net/bpf_jit_comp.c
3694
stack_ptr = per_cpu_ptr(priv_stack_ptr, cpu);
block/bio.c
481
cache = per_cpu_ptr(bs->cache, get_cpu());
block/bio.c
766
struct bio_alloc_cache *cache = per_cpu_ptr(bs->cache, cpu);
block/bio.c
784
cache = per_cpu_ptr(bs->cache, cpu);
block/bio.c
795
cache = per_cpu_ptr(bio->bi_pool->cache, get_cpu());
block/blk-cgroup.c
1048
struct llist_head *lhead = per_cpu_ptr(blkcg->lhead, cpu);
block/blk-cgroup.c
1111
plhead = per_cpu_ptr(parent->blkcg->lhead, cpu);
block/blk-cgroup.c
1158
cpu_dkstats = per_cpu_ptr(bdev->bd_stats, cpu);
block/blk-cgroup.c
2219
bis = per_cpu_ptr(bio->bi_blkg->iostat_cpu, cpu);
block/blk-cgroup.c
329
u64_stats_init(&per_cpu_ptr(blkg->iostat_cpu, cpu)->sync);
block/blk-cgroup.c
330
per_cpu_ptr(blkg->iostat_cpu, cpu)->blkg = blkg;
block/blk-cgroup.c
645
struct blkg_iostat_set *s = per_cpu_ptr(blkg->iostat_cpu, cpu);
block/blk-cgroup.c
93
init_llist_head(per_cpu_ptr(blkcg->lhead, cpu));
block/blk-iocost.c
1608
struct ioc_pcpu_stat *stat = per_cpu_ptr(ioc->pcpu_stat, cpu);
block/blk-iocost.c
1718
per_cpu_ptr(&iocg->pcpu_stat->abs_vusage, cpu));
block/blk-iocost.c
2907
struct ioc_pcpu_stat *ccs = per_cpu_ptr(ioc->pcpu_stat, cpu);
block/blk-iolatency.c
528
s = per_cpu_ptr(iolat->stats, cpu);
block/blk-iolatency.c
927
s = per_cpu_ptr(iolat->stats, cpu);
block/blk-iolatency.c
995
stat = per_cpu_ptr(iolat->stats, cpu);
block/blk-mq-sysfs.c
199
ctx = per_cpu_ptr(q->queue_ctx, cpu);
block/blk-mq-sysfs.c
213
ctx = per_cpu_ptr(q->queue_ctx, cpu);
block/blk-mq.c
4115
struct blk_mq_ctx *__ctx = per_cpu_ptr(q->queue_ctx, i);
block/blk-mq.c
4213
ctx = per_cpu_ptr(q->queue_ctx, i);
block/blk-mq.c
4395
struct blk_mq_ctx *ctx = per_cpu_ptr(ctxs->queue_ctx, cpu);
block/blk-mq.h
146
return per_cpu_ptr(q->queue_ctx, cpu);
block/blk-stat.c
142
cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
block/blk-stat.c
70
stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket];
block/blk-stat.c
89
cpu_stat = per_cpu_ptr(cb->cpu_stat, cpu);
block/genhd.c
114
struct disk_stats *ptr = per_cpu_ptr(part->bd_stats, cpu);
block/kyber-iosched.c
281
cpu_latency = per_cpu_ptr(kqd->cpu_latency, cpu);
crypto/acompress.c
368
ps = per_cpu_ptr(streams, cpu);
crypto/acompress.c
398
struct crypto_acomp_stream *ps = per_cpu_ptr(streams, i);
crypto/acompress.c
431
ps = per_cpu_ptr(streams, i);
crypto/acompress.c
435
ps = per_cpu_ptr(streams, i);
crypto/acompress.c
453
ps = per_cpu_ptr(streams, cpu);
crypto/acompress.c
462
ps = per_cpu_ptr(streams, cpumask_first(cpu_possible_mask));
crypto/cryptd.c
111
cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
crypto/cryptd.c
126
cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
crypto/scompress.c
102
scratch = per_cpu_ptr(&scomp_scratch, cpu);
crypto/scompress.c
118
scratch = per_cpu_ptr(&scomp_scratch, i);
crypto/scompress.c
148
scratch = per_cpu_ptr(&scomp_scratch, cpu);
crypto/scompress.c
157
scratch = per_cpu_ptr(&scomp_scratch, cpumask_first(cpu_possible_mask));
crypto/scompress.c
74
scratch = per_cpu_ptr(&scomp_scratch, i);
drivers/acpi/processor_perflib.c
624
if (!performance || !per_cpu_ptr(performance, i)) {
drivers/acpi/processor_perflib.c
636
pr->performance = per_cpu_ptr(performance, i);
drivers/base/arch_topology.c
110
sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu));
drivers/base/arch_topology.c
86
sfd = rcu_dereference(*per_cpu_ptr(&sft_data, cpu));
drivers/base/cpu.c
167
addr = per_cpu_ptr_to_phys(per_cpu_ptr(crash_notes, cpunum));
drivers/block/rnbd/rnbd-clt.c
170
return per_cpu_ptr(sess->cpu_queues, bit);
drivers/block/rnbd/rnbd-clt.c
175
return per_cpu_ptr(sess->cpu_queues, bit);
drivers/block/rnbd/rnbd-clt.c
729
cpu_q = per_cpu_ptr(cpu_queues, cpu);
drivers/block/rnbd/rnbd-clt.c
814
* per_cpu_ptr(sess->cpu_rr, cpu) = cpu;
drivers/block/zram/zcomp.c
172
struct zcomp_strm *zstrm = per_cpu_ptr(comp->stream, cpu);
drivers/block/zram/zcomp.c
184
struct zcomp_strm *zstrm = per_cpu_ptr(comp->stream, cpu);
drivers/block/zram/zcomp.c
206
mutex_init(&per_cpu_ptr(comp->stream, cpu)->lock);
drivers/bus/mips_cdmm.c
304
bus_p = per_cpu_ptr(&mips_cdmm_buses, cpu);
drivers/char/random.c
1058
per_cpu_ptr(&irq_randomness, cpu)->count = 0;
drivers/char/random.c
591
per_cpu_ptr(&crngs, cpu)->generation = ULONG_MAX;
drivers/char/random.c
592
per_cpu_ptr(&batched_entropy_u8, cpu)->position = UINT_MAX;
drivers/char/random.c
593
per_cpu_ptr(&batched_entropy_u16, cpu)->position = UINT_MAX;
drivers/char/random.c
594
per_cpu_ptr(&batched_entropy_u32, cpu)->position = UINT_MAX;
drivers/char/random.c
595
per_cpu_ptr(&batched_entropy_u64, cpu)->position = UINT_MAX;
drivers/clocksource/dummy_timer.c
18
struct clock_event_device *evt = per_cpu_ptr(&dummy_timer_evt, cpu);
drivers/clocksource/exynos_mct.c
458
per_cpu_ptr(&percpu_mct_tick, cpu);
drivers/clocksource/exynos_mct.c
496
per_cpu_ptr(&percpu_mct_tick, cpu);
drivers/clocksource/exynos_mct.c
575
per_cpu_ptr(&percpu_mct_tick, cpu);
drivers/clocksource/exynos_mct.c
604
struct mct_clock_event_device *mevt = per_cpu_ptr(&percpu_mct_tick, cpu);
drivers/clocksource/exynos_mct.c
630
per_cpu_ptr(&percpu_mct_tick, cpu);
drivers/clocksource/hyperv_timer.c
138
ce = per_cpu_ptr(hv_clock_event, cpu);
drivers/clocksource/hyperv_timer.c
192
ce = per_cpu_ptr(hv_clock_event, cpu);
drivers/clocksource/jcore-pit.c
228
struct jcore_pit *pit = per_cpu_ptr(jcore_pit_percpu, cpu);
drivers/clocksource/timer-armada-370-xp.c
172
struct clock_event_device *evt = per_cpu_ptr(armada_370_xp_evt, cpu);
drivers/clocksource/timer-armada-370-xp.c
202
struct clock_event_device *evt = per_cpu_ptr(armada_370_xp_evt, cpu);
drivers/clocksource/timer-clint.c
131
struct clock_event_device *ce = per_cpu_ptr(&clint_clock_event, cpu);
drivers/clocksource/timer-mp-csky.c
145
to = per_cpu_ptr(&csky_to, cpu);
drivers/clocksource/timer-mp-csky.c
168
to = per_cpu_ptr(&csky_to, cpu_rollback);
drivers/clocksource/timer-mp-csky.c
78
struct timer_of *to = per_cpu_ptr(&csky_to, cpu);
drivers/clocksource/timer-qcom.c
101
struct clock_event_device *evt = per_cpu_ptr(msm_evt, cpu);
drivers/clocksource/timer-qcom.c
131
struct clock_event_device *evt = per_cpu_ptr(msm_evt, cpu);
drivers/clocksource/timer-riscv.c
110
struct clock_event_device *ce = per_cpu_ptr(&riscv_clock_event, cpu);
drivers/clocksource/timer-rtl-otto.c
248
struct timer_of *to = per_cpu_ptr(&rttm_to, cpu);
drivers/clocksource/timer-rtl-otto.c
268
to = per_cpu_ptr(&rttm_to, cpu);
drivers/clocksource/timer-rtl-otto.c
296
to = per_cpu_ptr(&rttm_to, cpu_rollback);
drivers/clocksource/timer-tegra.c
134
struct timer_of *to = per_cpu_ptr(&tegra_to, cpu);
drivers/clocksource/timer-tegra.c
159
struct timer_of *to = per_cpu_ptr(&tegra_to, cpu);
drivers/clocksource/timer-tegra.c
301
struct timer_of *cpu_to = per_cpu_ptr(&tegra_to, cpu);
drivers/clocksource/timer-tegra.c
358
cpu_to = per_cpu_ptr(&tegra_to, cpu);
drivers/clocksource/timer-ti-dm-systimer.c
648
clkevt = per_cpu_ptr(&dmtimer_percpu_timer, cpu);
drivers/clocksource/timer-ti-dm-systimer.c
662
struct dmtimer_clockevent *clkevt = per_cpu_ptr(&dmtimer_percpu_timer, cpu);
drivers/clocksource/timer-ti-dm-systimer.c
674
struct dmtimer_clockevent *clkevt = per_cpu_ptr(&dmtimer_percpu_timer, 0);
drivers/cpufreq/acpi-cpufreq.c
523
free_cpumask_var(per_cpu_ptr(acpi_perf_data, i)
drivers/cpufreq/acpi-cpufreq.c
557
&per_cpu_ptr(acpi_perf_data, i)->shared_cpu_map,
drivers/cpufreq/acpi-cpufreq.c
67
return per_cpu_ptr(acpi_perf_data, data->acpi_perf_cpu);
drivers/cpufreq/acpi-cpufreq.c
735
perf = per_cpu_ptr(acpi_perf_data, cpu);
drivers/cpufreq/pcc-cpufreq.c
153
pcc_cpu_data = per_cpu_ptr(pcc_cpu_info, cpu);
drivers/cpufreq/pcc-cpufreq.c
209
pcc_cpu_data = per_cpu_ptr(pcc_cpu_info, cpu);
drivers/cpufreq/pcc-cpufreq.c
259
pcc_cpu_data = per_cpu_ptr(pcc_cpu_info, cpu);
drivers/cpuidle/cpuidle-haltpoll.c
60
dev = per_cpu_ptr(haltpoll_cpuidle_devices, cpu);
drivers/cpuidle/cpuidle-haltpoll.c
77
dev = per_cpu_ptr(haltpoll_cpuidle_devices, cpu);
drivers/cpuidle/cpuidle-psci.c
160
dev = per_cpu_ptr(&psci_cpuidle_data, cpu)->dev;
drivers/cpuidle/cpuidle-psci.c
283
struct psci_cpuidle_data *data = per_cpu_ptr(&psci_cpuidle_data, cpu);
drivers/cpuidle/cpuidle-psci.c
344
struct psci_cpuidle_data *data = per_cpu_ptr(&psci_cpuidle_data, cpu);
drivers/cpuidle/cpuidle-riscv-sbi.c
239
struct sbi_cpuidle_data *data = per_cpu_ptr(&sbi_cpuidle_data, cpu);
drivers/cpuidle/cpuidle-riscv-sbi.c
282
struct sbi_cpuidle_data *data = per_cpu_ptr(&sbi_cpuidle_data, cpu);
drivers/cpuidle/governors/teo.c
557
struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
drivers/crypto/caam/caamalg_qi2.c
4692
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
4732
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
4739
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
4754
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
4774
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
4828
ppriv = per_cpu_ptr(priv->ppriv, i);
drivers/crypto/caam/caamalg_qi2.c
5008
ppriv = per_cpu_ptr(priv->ppriv, i);
drivers/crypto/caam/caamalg_qi2.c
5106
ppriv = per_cpu_ptr(priv->ppriv, cpu);
drivers/crypto/caam/caamalg_qi2.c
5158
ppriv = per_cpu_ptr(priv->ppriv, i);
drivers/crypto/caam/caamalg_qi2.c
5187
ppriv = per_cpu_ptr(priv->ppriv, i);
drivers/crypto/caam/qi.c
524
irqtask = &per_cpu_ptr(&pcpu_qipriv.caam_napi, i)->irqtask;
drivers/crypto/caam/qi.c
725
priv = per_cpu_ptr(&pcpu_qipriv, i);
drivers/crypto/caam/qi.c
763
struct caam_qi_pcpu_priv *priv = per_cpu_ptr(&pcpu_qipriv, i);
drivers/crypto/intel/iaa/iaa_crypto_main.c
42
struct wq_table_entry *entry = per_cpu_ptr(wq_table, cpu);
drivers/crypto/intel/iaa/iaa_crypto_main.c
59
struct wq_table_entry *entry = per_cpu_ptr(wq_table, cpu);
drivers/crypto/intel/iaa/iaa_crypto_main.c
720
entry = per_cpu_ptr(wq_table, cpu);
drivers/crypto/intel/iaa/iaa_crypto_main.c
73
struct wq_table_entry *entry = per_cpu_ptr(wq_table, cpu);
drivers/crypto/intel/iaa/iaa_crypto_main.c
81
struct wq_table_entry *entry = per_cpu_ptr(wq_table, cpu);
drivers/dma/dmaengine.c
176
count += per_cpu_ptr(chan->local, i)->memcpy_count;
drivers/dma/dmaengine.c
198
count += per_cpu_ptr(chan->local, i)->bytes_transferred;
drivers/dma/dmaengine.c
373
per_cpu_ptr(channel_table[cap], cpu)->chan = NULL;
drivers/dma/dmaengine.c
390
per_cpu_ptr(channel_table[cap], cpu)->chan = chan;
drivers/edac/amd64_edac.c
3221
struct msr *reg = per_cpu_ptr(msrs, cpu);
drivers/edac/amd64_edac.c
3253
struct msr *reg = per_cpu_ptr(msrs, cpu);
drivers/firmware/arm_ffa/driver.c
1932
per_cpu_ptr(irq_pcpu, cpu)->info = drv_info;
drivers/firmware/arm_sdei.c
253
reg = per_cpu_ptr(regs, cpu);
drivers/firmware/arm_sdei.c
562
reg = per_cpu_ptr(arg->event->private_registered, smp_processor_id());
drivers/hv/channel_mgmt.c
844
= per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/channel_mgmt.c
893
= per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
182
hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
193
hv_cpu = per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
250
per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
279
per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
350
= per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/hv.c
407
per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hv/mshv_vtl_main.c
198
return *per_cpu_ptr(&mshv_vtl_per_cpu.run, cpu);
drivers/hv/mshv_vtl_main.c
203
return *per_cpu_ptr(&mshv_vtl_per_cpu.reg_page, cpu);
drivers/hv/mshv_vtl_main.c
480
poll_file = per_cpu_ptr(&mshv_vtl_poll_file, READ_ONCE(input.cpu));
drivers/hv/vmbus_drv.c
1448
struct work_struct *work = per_cpu_ptr(works, cpu);
drivers/hv/vmbus_drv.c
1455
flush_work(per_cpu_ptr(works, cpu));
drivers/hv/vmbus_drv.c
2706
struct hv_per_cpu_context *hv_cpu = per_cpu_ptr(
drivers/hv/vmbus_drv.c
3036
= per_cpu_ptr(hv_context.cpu_context, cpu);
drivers/hwtracing/coresight/coresight-etm-perf.c
145
return per_cpu_ptr(data->path, cpu);
drivers/hwtracing/coresight/coresight-trace-id.c
137
atomic_set(per_cpu_ptr(id_map->cpu_map, cpu), 0);
drivers/hwtracing/coresight/coresight-trace-id.c
171
atomic_set(per_cpu_ptr(id_map->cpu_map, cpu), id);
drivers/hwtracing/coresight/coresight-trace-id.c
194
atomic_set(per_cpu_ptr(id_map->cpu_map, cpu), 0);
drivers/hwtracing/coresight/coresight-trace-id.c
51
return atomic_read(per_cpu_ptr(id_map->cpu_map, cpu));
drivers/hwtracing/coresight/coresight-trbe.c
1256
struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
drivers/hwtracing/coresight/coresight-trbe.c
1310
struct trbe_cpudata *cpudata = per_cpu_ptr(drvdata->cpudata, cpu);
drivers/idle/intel_idle.c
2320
dev = per_cpu_ptr(intel_idle_cpuidle_devices, cpu);
drivers/idle/intel_idle.c
2351
dev = per_cpu_ptr(intel_idle_cpuidle_devices, cpu);
drivers/idle/intel_idle.c
2366
cpuidle_unregister_device(per_cpu_ptr(intel_idle_cpuidle_devices, i));
drivers/infiniband/hw/hfi1/affinity.c
258
prev_cntr = *per_cpu_ptr(comp_vect_affinity, ret_cpu);
drivers/infiniband/hw/hfi1/affinity.c
260
cntr = *per_cpu_ptr(comp_vect_affinity, curr_cpu);
drivers/infiniband/hw/hfi1/affinity.c
268
*per_cpu_ptr(comp_vect_affinity, ret_cpu) += 1;
drivers/infiniband/hw/hfi1/affinity.c
292
prev_cntr = *per_cpu_ptr(comp_vect_affinity, max_cpu);
drivers/infiniband/hw/hfi1/affinity.c
294
cntr = *per_cpu_ptr(comp_vect_affinity, curr_cpu);
drivers/infiniband/hw/hfi1/affinity.c
302
*per_cpu_ptr(comp_vect_affinity, max_cpu) -= 1;
drivers/infiniband/hw/hfi1/chip.c
1603
counter += *per_cpu_ptr(cntr, cpu);
drivers/infiniband/hw/hfi1/debugfs.c
111
per_cpu_ptr(dd->tx_opstats, j);
drivers/infiniband/hw/hfi1/fault.c
62
per_cpu_ptr(dd->tx_opstats, j);
drivers/infiniband/hw/hfi1/pio.c
637
ret += *per_cpu_ptr(sc->buffers_allocated, cpu);
drivers/infiniband/hw/hfi1/pio.c
646
(*per_cpu_ptr(sc->buffers_allocated, cpu)) = 0;
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
119
s = per_cpu_ptr(stats->pcpu_stats, cpu);
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
135
s = per_cpu_ptr(stats->pcpu_stats, cpu);
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
27
s = per_cpu_ptr(stats->pcpu_stats, con->cpu);
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
47
s = per_cpu_ptr(stats->pcpu_stats, cpu);
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
66
s = per_cpu_ptr(stats->pcpu_stats, cpu);
drivers/infiniband/ulp/rtrs/rtrs-clt-stats.c
90
r = &per_cpu_ptr(stats->pcpu_stats, cpu)->rdma;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1589
INIT_LIST_HEAD(per_cpu_ptr(clt_path->mp_skip_entry, cpu));
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2299
ppcpu_path = per_cpu_ptr(clt->pcpu_path, cpu);
drivers/infiniband/ulp/rtrs/rtrs-srv-stats.c
21
r = per_cpu_ptr(stats->rdma_stats, cpu);
drivers/infiniband/ulp/rtrs/rtrs-srv-stats.c
40
r = per_cpu_ptr(stats->rdma_stats, cpu);
drivers/iommu/dma-iommu.c
193
fq_ring_free(cookie, per_cpu_ptr(cookie->percpu_fq, cpu));
drivers/iommu/dma-iommu.c
264
struct iova_fq *fq = per_cpu_ptr(percpu_fq, cpu);
drivers/iommu/dma-iommu.c
326
iommu_dma_init_one_fq(per_cpu_ptr(queue, cpu), fq_size);
drivers/iommu/intel/irq_remapping.c
1130
return __pa(per_cpu_ptr(&posted_msi_pi_desc, cpu));
drivers/iommu/iova.c
738
cpu_rcache = per_cpu_ptr(rcache->cpu_rcaches, cpu);
drivers/iommu/iova.c
887
cpu_rcache = per_cpu_ptr(rcache->cpu_rcaches, cpu);
drivers/iommu/iova.c
913
cpu_rcache = per_cpu_ptr(rcache->cpu_rcaches, cpu);
drivers/irqchip/irq-gic-v3-its.c
1559
rdbase = per_cpu_ptr(gic_rdists->rdist, cpu)->rd_base;
drivers/irqchip/irq-gic-v3-its.c
1653
return atomic_read(&per_cpu_ptr(&cpu_lpi_count, cpu)->managed);
drivers/irqchip/irq-gic-v3-its.c
1655
return atomic_read(&per_cpu_ptr(&cpu_lpi_count, cpu)->unmanaged);
drivers/irqchip/irq-gic-v3-its.c
1661
atomic_inc(&per_cpu_ptr(&cpu_lpi_count, cpu)->managed);
drivers/irqchip/irq-gic-v3-its.c
1663
atomic_inc(&per_cpu_ptr(&cpu_lpi_count, cpu)->unmanaged);
drivers/irqchip/irq-gic-v3-its.c
1669
atomic_dec(&per_cpu_ptr(&cpu_lpi_count, cpu)->managed);
drivers/irqchip/irq-gic-v3-its.c
1671
atomic_dec(&per_cpu_ptr(&cpu_lpi_count, cpu)->unmanaged);
drivers/irqchip/irq-gic-v3-its.c
207
#define gic_data_rdist_cpu(cpu) (per_cpu_ptr(gic_rdists->rdist, cpu))
drivers/irqchip/irq-gic-v3-its.c
2822
*this_cpu_ptr(&local_4_1_its) = *per_cpu_ptr(&local_4_1_its, cpu);
drivers/irqchip/irq-gic-v3-its.c
3877
rdbase = per_cpu_ptr(gic_rdists->rdist, from)->rd_base;
drivers/irqchip/irq-gic-v3-its.c
3904
rdbase = per_cpu_ptr(gic_rdists->rdist, cpu)->rd_base;
drivers/irqchip/irq-gic-v3-its.c
4160
rdbase = per_cpu_ptr(gic_rdists->rdist, vpe->col_idx)->rd_base;
drivers/irqchip/irq-gic.c
1184
*per_cpu_ptr(gic->dist_base.percpu_base, cpu) =
drivers/irqchip/irq-gic.c
1186
*per_cpu_ptr(gic->cpu_base.percpu_base, cpu) =
drivers/irqchip/irq-loongarch-avec.c
125
per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(data);
drivers/irqchip/irq-loongarch-avec.c
152
struct pending_list *plist = per_cpu_ptr(&pending_list, cpu);
drivers/irqchip/irq-loongarch-avec.c
269
per_cpu_ptr(irq_map, adata->cpu)[adata->vec] = irq_data_to_desc(irqd);
drivers/irqchip/irq-loongarch-avec.c
85
struct pending_list *plist = per_cpu_ptr(&pending_list, cpu);
drivers/irqchip/irq-loongarch-avec.c
95
plist = per_cpu_ptr(&pending_list, adata->prev_cpu);
drivers/irqchip/irq-mips-gic.c
156
clear_bit(intr, per_cpu_ptr(pcpu_masks, i));
drivers/irqchip/irq-mips-gic.c
293
set_bit(intr, per_cpu_ptr(pcpu_masks, cpu));
drivers/irqchip/irq-mips-gic.c
429
set_bit(irq, per_cpu_ptr(pcpu_masks, cpu));
drivers/irqchip/irq-mips-gic.c
439
set_bit(irq, per_cpu_ptr(pcpu_masks, cpu));
drivers/irqchip/irq-riscv-aplic-direct.c
181
aplic_idc_set_delivery(per_cpu_ptr(&aplic_idcs, cpu), true);
drivers/irqchip/irq-riscv-aplic-direct.c
277
idc = per_cpu_ptr(&aplic_idcs, cpu);
drivers/irqchip/irq-riscv-aplic-direct.c
69
idc = per_cpu_ptr(&aplic_idcs, cpu);
drivers/irqchip/irq-riscv-imsic-early.c
39
struct imsic_local_config *local = per_cpu_ptr(imsic->global.local, cpu);
drivers/irqchip/irq-riscv-imsic-platform.c
33
local = per_cpu_ptr(global->local, cpu);
drivers/irqchip/irq-riscv-imsic-platform.c
62
local = per_cpu_ptr(imsic->global.local, vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
186
tlocal = per_cpu_ptr(imsic->global.local, tvec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
190
mlocal = per_cpu_ptr(imsic->global.local, mvec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
195
mlocal = per_cpu_ptr(imsic->global.local, mvec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
296
lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
318
lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
342
lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
387
old_lpriv = per_cpu_ptr(imsic->lpriv, old_vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
391
new_lpriv = per_cpu_ptr(imsic->lpriv, new_vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
412
lpriv = per_cpu_ptr(imsic->lpriv, vec->cpu);
drivers/irqchip/irq-riscv-imsic-state.c
451
lpriv = per_cpu_ptr(imsic->lpriv, cpu);
drivers/irqchip/irq-riscv-imsic-state.c
477
lpriv = per_cpu_ptr(imsic->lpriv, cpu);
drivers/irqchip/irq-riscv-imsic-state.c
500
lpriv = per_cpu_ptr(imsic->lpriv, cpu);
drivers/irqchip/irq-riscv-imsic-state.c
918
local = per_cpu_ptr(global->local, cpu);
drivers/irqchip/irq-sifive-plic.c
141
struct plic_handler *handler = per_cpu_ptr(&plic_handlers, cpu);
drivers/irqchip/irq-sifive-plic.c
297
struct plic_handler *handler = per_cpu_ptr(&plic_handlers, cpu);
drivers/irqchip/irq-sifive-plic.c
720
handler = per_cpu_ptr(&plic_handlers, cpu);
drivers/irqchip/irq-sifive-plic.c
766
handler = per_cpu_ptr(&plic_handlers, cpu);
drivers/irqchip/irq-sifive-plic.c
810
handler = per_cpu_ptr(&plic_handlers, cpu);
drivers/mailbox/riscv-sbi-mpxy-mbox.c
345
mpxy = per_cpu_ptr(&mpxy_local, cpu);
drivers/md/dm-pcache/cache.c
322
per_cpu_ptr(cache->data_heads, cpu);
drivers/md/dm-stats.c
207
last = per_cpu_ptr(stats->last, cpu);
drivers/md/dm.c
2773
sum += *per_cpu_ptr(md->pending_io, cpu);
drivers/md/raid5.c
2491
percpu = per_cpu_ptr(conf->percpu, cpu);
drivers/md/raid5.c
7364
free_scratch_buffer(conf, per_cpu_ptr(conf->percpu, cpu));
drivers/md/raid5.c
7403
struct raid5_percpu *percpu = per_cpu_ptr(conf->percpu, cpu);
drivers/misc/vcpu_stall_detector.c
126
per_cpu_ptr(vcpu_stall_detectors, cpu);
drivers/net/ethernet/asix/ax88796c_main.c
662
s = per_cpu_ptr(ax_local->stats, cpu);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
339
tmp_stats += ((u64 *)per_cpu_ptr(nic->drv_stats, cpu))
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1522
memset(per_cpu_ptr(nic->drv_stats, cpu), 0,
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1692
drv_stats = per_cpu_ptr(nic->drv_stats, cpu);
drivers/net/ethernet/chelsio/cxgb/sge.c
984
struct sge_port_stats *st = per_cpu_ptr(sge->port_stats[port], cpu);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
129
pool = per_cpu_ptr(ppm->pool, cpu);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
202
pool = per_cpu_ptr(ppm->pool, cpu);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c
374
struct cxgbi_ppm_pool *ppool = per_cpu_ptr(pools, cpu);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
1610
int *count_ptr = per_cpu_ptr(dpaa_bp->percpu_count, i);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
217
percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2922
percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2935
percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
3171
percpu_priv = per_cpu_ptr(priv->percpu_priv, cpu);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
3186
percpu_priv = per_cpu_ptr(priv->percpu_priv, cpu);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
360
percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
197
percpu_priv = per_cpu_ptr(priv->percpu_priv, i);
drivers/net/ethernet/freescale/dpaa/dpaa_ethtool.c
201
bp_count = *(per_cpu_ptr(dpaa_bp->percpu_count, i));
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-debugfs.c
27
stats = per_cpu_ptr(priv->percpu_stats, i);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-debugfs.c
28
extras = per_cpu_ptr(priv->percpu_extras, i);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
1906
sgt_cache = per_cpu_ptr(priv->sgt_cache, k);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
2375
percpu_stats = per_cpu_ptr(priv->percpu_stats, i);
drivers/net/ethernet/freescale/dpaa2/dpaa2-ethtool.c
292
extras = per_cpu_ptr(priv->percpu_extras, k);
drivers/net/ethernet/intel/ice/ice_repr.c
182
repr_stats = per_cpu_ptr(repr->stats, i);
drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c
182
ddp_pool = per_cpu_ptr(fcoe->ddp_pool, get_cpu());
drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c
602
ddp_pool = per_cpu_ptr(fcoe->ddp_pool, cpu);
drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c
622
ddp_pool = per_cpu_ptr(fcoe->ddp_pool, cpu);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7842
ddp_pool = per_cpu_ptr(fcoe->ddp_pool, cpu);
drivers/net/ethernet/marvell/mvneta.c
3771
per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
3810
per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
4428
struct mvneta_pcpu_port *port = per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
4456
per_cpu_ptr(pp->ports, other_cpu);
drivers/net/ethernet/marvell/mvneta.c
4494
struct mvneta_pcpu_port *port = per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
4877
stats = per_cpu_ptr(pp->stats, cpu);
drivers/net/ethernet/marvell/mvneta.c
5033
per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
5061
per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
5713
per_cpu_ptr(pp->ports, cpu);
drivers/net/ethernet/marvell/mvneta.c
820
cpu_stats = per_cpu_ptr(pp->stats, cpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2033
cpu_stats = per_cpu_ptr(port->stats, cpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2550
txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2891
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3152
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3187
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3257
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3646
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3682
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4087
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4193
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4374
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4433
struct mvpp2_pcpu_stats *stats = per_cpu_ptr(port->stats, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4465
struct mvpp2_port_pcpu *port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4904
port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5156
cpu_stats = per_cpu_ptr(port->stats, cpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6045
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7009
port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
163
lmt_info = per_cpu_ptr(pfvf->hw.lmt_info, smp_processor_id());
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
79
lmt_info = per_cpu_ptr(pfvf->hw.lmt_info, cpu);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
387
lmt_info = per_cpu_ptr(pf->hw.lmt_info, smp_processor_id());
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
788
lmt_info = per_cpu_ptr(pfvf->hw.lmt_info, smp_processor_id());
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
728
p = per_cpu_ptr(mlxsw_sp_port->pcpu_stats, i);
drivers/net/ethernet/netronome/nfp/nfp_net_repr.c
135
repr_stats = per_cpu_ptr(repr->stats, i);
drivers/net/ethernet/nvidia/forcedeth.c
1717
struct nv_txrx_stats *src = per_cpu_ptr(np->txrx_stats, cpu);
drivers/net/ethernet/qualcomm/rmnet/rmnet_vnd.c
140
pcpu_ptr = per_cpu_ptr(priv->pcpu_stats, cpu);
drivers/net/ethernet/spacemit/k1_emac.c
1179
stats = per_cpu_ptr(priv->ndev->dstats, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
451
pcpu = per_cpu_ptr(priv->xstats.pcpu_stats, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
472
pcpu = per_cpu_ptr(priv->xstats.pcpu_stats, cpu);
drivers/net/hyperv/netvsc_drv.c
1234
= per_cpu_ptr(ndev_ctx->vf_stats, i);
drivers/net/hyperv/netvsc_drv.c
1264
per_cpu_ptr(ndev_ctx->vf_stats, i);
drivers/net/ipvlan/ipvlan_main.c
306
pcptr = per_cpu_ptr(ipvlan->pcpu_stats, idx);
drivers/net/loopback.c
108
lb_stats = per_cpu_ptr(dev->lstats, i);
drivers/net/macsec.c
2707
per_cpu_ptr(tx_sa->stats, cpu);
drivers/net/macsec.c
2752
per_cpu_ptr(rx_sa->stats, cpu);
drivers/net/macsec.c
2806
stats = per_cpu_ptr(rx_sc->stats, cpu);
drivers/net/macsec.c
2887
stats = per_cpu_ptr(macsec_priv(dev)->secy.tx_sc.stats, cpu);
drivers/net/macsec.c
2943
stats = per_cpu_ptr(macsec_priv(dev)->stats, cpu);
drivers/net/macvlan.c
985
p = per_cpu_ptr(vlan->pcpu_stats, i);
drivers/net/ppp/ppp_generic.c
1266
xmit_recursion = per_cpu_ptr(ppp->xmit_recursion, cpu);
drivers/net/ppp/ppp_generic.c
3336
struct pcpu_sw_netstats *p = per_cpu_ptr(ppp->dev->tstats, cpu);
drivers/net/team/team_core.c
1819
p = per_cpu_ptr(team->pcpu_stats, i);
drivers/net/team/team_mode_loadbalance.c
487
pcpu_stats = per_cpu_ptr(lb_priv->pcpu_stats, i);
drivers/net/team/team_mode_loadbalance.c
501
pcpu_stats = per_cpu_ptr(lb_priv->pcpu_stats, i);
drivers/net/team/team_mode_loadbalance.c
502
stats = per_cpu_ptr(lb_port_priv->pcpu_stats, i);
drivers/net/team/team_mode_loadbalance.c
610
team_lb_stats = per_cpu_ptr(lb_priv->pcpu_stats, i);
drivers/net/vxlan/vxlan_vnifilter.c
126
pstats = per_cpu_ptr(vninode->stats, i);
drivers/net/wireguard/queueing.c
19
per_cpu_ptr(worker, cpu)->ptr = ptr;
drivers/net/wireguard/queueing.c
20
INIT_WORK(&per_cpu_ptr(worker, cpu)->work, function);
drivers/net/wireguard/queueing.h
171
queue_work_on(cpu, wq, &per_cpu_ptr(device_queue->worker, cpu)->work);
drivers/net/wireguard/receive.c
571
&per_cpu_ptr(wg->handshake_queue.worker, cpu)->work);
drivers/net/wireless/ath/ath9k/common-spectral.c
457
if ((buf = *per_cpu_ptr(rc->buf, i))) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2016
per_cpu_ptr(trans_pcie->txqs.tso_hdr_page, i);
drivers/net/xen-netfront.c
1396
struct netfront_stats *rx_stats = per_cpu_ptr(np->rx_stats, cpu);
drivers/net/xen-netfront.c
1397
struct netfront_stats *tx_stats = per_cpu_ptr(np->tx_stats, cpu);
drivers/nvdimm/region_devs.c
1029
ndl = per_cpu_ptr(nd_region->lane, i);
drivers/nvdimm/region_devs.c
930
ndl_count = per_cpu_ptr(nd_region->lane, cpu);
drivers/nvdimm/region_devs.c
931
ndl_lock = per_cpu_ptr(nd_region->lane, lane);
drivers/nvdimm/region_devs.c
947
ndl_count = per_cpu_ptr(nd_region->lane, cpu);
drivers/nvdimm/region_devs.c
948
ndl_lock = per_cpu_ptr(nd_region->lane, lane);
drivers/perf/arm_pmu.c
40
free_irq(irq, per_cpu_ptr(devid, cpu));
drivers/perf/arm_pmu.c
51
free_nmi(irq, per_cpu_ptr(devid, cpu));
drivers/perf/arm_pmu.c
633
struct arm_pmu **armpmu = per_cpu_ptr(pcpu_armpmu, cpu);
drivers/perf/arm_pmu.c
68
struct arm_pmu *armpmu = *per_cpu_ptr((void * __percpu *)devid, cpu);
drivers/perf/arm_pmu.c
905
events = per_cpu_ptr(pmu->hw_events, cpu);
drivers/perf/arm_pmu.c
95
struct arm_pmu *armpmu = *per_cpu_ptr((void * __percpu *)devid, cpu);
drivers/perf/arm_pmuv3.c
1363
events_cpu = per_cpu_ptr(armpmu->hw_events, cpu);
drivers/perf/qcom_l2_pmu.c
166
return *per_cpu_ptr(l2cache_pmu->pmu_cluster, cpu);
drivers/perf/qcom_l2_pmu.c
760
*per_cpu_ptr(l2cache_pmu->pmu_cluster, cpu) = cluster;
drivers/perf/riscv_pmu.c
403
cpuc = per_cpu_ptr(pmu->hw_events, cpuid);
drivers/perf/riscv_pmu_sbi.c
661
struct cpu_hw_events *cpu_hw_evt = per_cpu_ptr(pmu->hw_events, cpu);
drivers/perf/riscv_pmu_sbi.c
678
struct cpu_hw_events *cpu_hw_evt = per_cpu_ptr(pmu->hw_events, cpu);
drivers/perf/riscv_pmu_sbi.c
711
cpu_hw_evt = per_cpu_ptr(pmu->hw_events, cpu);
drivers/perf/starfive_starlink_pmu.c
531
hw_events = per_cpu_ptr(starlink_pmu->hw_events, cpuid);
drivers/platform/x86/amd/hfi/hfi.c
184
info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu_index);
drivers/platform/x86/amd/hfi/hfi.c
188
info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu_index);
drivers/platform/x86/amd/hfi/hfi.c
232
hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, idx);
drivers/platform/x86/amd/hfi/hfi.c
278
struct amd_hfi_cpuinfo *hfi_info = per_cpu_ptr(&amd_hfi_cpuinfo, cpu);
drivers/platform/x86/amd/hfi/hfi.c
333
struct amd_hfi_cpuinfo *hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, cpu);
drivers/platform/x86/amd/hfi/hfi.c
408
struct amd_hfi_cpuinfo *hfi_cpuinfo = per_cpu_ptr(&amd_hfi_cpuinfo, cpu);
drivers/powercap/idle_inject.c
108
iit = per_cpu_ptr(&idle_inject_thread, cpu);
drivers/powercap/idle_inject.c
154
iit = per_cpu_ptr(&idle_inject_thread, cpu);
drivers/powercap/idle_inject.c
279
iit = per_cpu_ptr(&idle_inject_thread, cpu);
drivers/powercap/idle_inject.c
310
per_cpu_ptr(&idle_inject_thread, cpu);
drivers/resctrl/mpam_devices.c
1937
*per_cpu_ptr(msc->error_dev_id, cpu) = msc;
drivers/scsi/fcoe/fcoe.c
1290
p = per_cpu_ptr(&fcoe_percpu, cpu);
drivers/scsi/fcoe/fcoe.c
1662
stats = per_cpu_ptr(lport->stats, raw_smp_processor_id());
drivers/scsi/fcoe/fcoe.c
2456
p = per_cpu_ptr(&fcoe_percpu, cpu);
drivers/scsi/fcoe/fcoe_transport.c
185
stats = per_cpu_ptr(lport->stats, cpu);
drivers/scsi/hpsa.c
2872
lockup_detected = per_cpu_ptr(h->lockup_detected, cpu);
drivers/scsi/hpsa.c
8245
lockup_detected = per_cpu_ptr(h->lockup_detected, cpu);
drivers/scsi/libfc/fc_exch.c
1987
per_cpu_ptr(ema->mp->pool, cpu),
drivers/scsi/libfc/fc_exch.c
2523
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_exch.c
829
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_exch.c
943
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_lport.c
315
stats = per_cpu_ptr(lport->stats, cpu);
drivers/scsi/lpfc/lpfc_attr.c
260
cgs = per_cpu_ptr(phba->cmf_stat, cpu);
drivers/scsi/lpfc/lpfc_attr.c
394
lta = per_cpu_ptr(vmp->last_io_time, cpu);
drivers/scsi/lpfc/lpfc_attr.c
4961
eqi = per_cpu_ptr(phba->sli4_hba.eq_info, i);
drivers/scsi/lpfc/lpfc_debugfs.c
1736
c_stat = per_cpu_ptr(phba->sli4_hba.c_stat, j);
drivers/scsi/lpfc/lpfc_debugfs.c
3139
c_stat = per_cpu_ptr(phba->sli4_hba.c_stat, i);
drivers/scsi/lpfc/lpfc_hbadisc.c
337
lta = per_cpu_ptr(vmp->last_io_time, cpu);
drivers/scsi/lpfc/lpfc_init.c
10528
eqi = per_cpu_ptr(phba->sli4_hba.eq_info, qdesc->last_cpu);
drivers/scsi/lpfc/lpfc_init.c
12376
eqi = per_cpu_ptr(phba->sli4_hba.eq_info, cpu);
drivers/scsi/lpfc/lpfc_init.c
12656
c_stat = per_cpu_ptr(phba->sli4_hba.c_stat, cpu);
drivers/scsi/lpfc/lpfc_init.c
1380
eqi = per_cpu_ptr(phba->sli4_hba.eq_info, i);
drivers/scsi/lpfc/lpfc_init.c
1393
eqi_new = per_cpu_ptr(phba->sli4_hba.eq_info,
drivers/scsi/lpfc/lpfc_init.c
3205
cgs = per_cpu_ptr(phba->cmf_stat, cpu);
drivers/scsi/lpfc/lpfc_init.c
3270
cgs = per_cpu_ptr(phba->cmf_stat, cpu);
drivers/scsi/lpfc/lpfc_init.c
6010
cgs = per_cpu_ptr(phba->cmf_stat, cpu);
drivers/scsi/lpfc/lpfc_scsi.c
3766
cgs = per_cpu_ptr(phba->cmf_stat, raw_smp_processor_id());
drivers/scsi/lpfc/lpfc_scsi.c
3794
cgs = per_cpu_ptr(phba->cmf_stat, cpu);
drivers/scsi/lpfc/lpfc_scsi.c
3810
cgs = per_cpu_ptr(phba->cmf_stat, raw_smp_processor_id());
drivers/scsi/lpfc/lpfc_sli.c
16288
eqi = per_cpu_ptr(phba->sli4_hba.eq_info, raw_smp_processor_id());
drivers/scsi/lpfc/lpfc_vmid.c
125
lta = per_cpu_ptr(vmp->last_io_time, raw_smp_processor_id());
drivers/scsi/lpfc/lpfc_vmid.c
317
*per_cpu_ptr(vmp->last_io_time, cpu) = 0;
drivers/scsi/smartpqi/smartpqi_init.c
6030
per_cpu_ptr(device->raid_io_stats, raw_smp_processor_id())->write_stream_cnt++;
drivers/scsi/smartpqi/smartpqi_init.c
6110
per_cpu_ptr(device->raid_io_stats, raw_smp_processor_id())->raid_bypass_cnt++;
drivers/scsi/smartpqi/smartpqi_init.c
7444
raid_bypass_cnt += per_cpu_ptr(device->raid_io_stats, cpu)->raid_bypass_cnt;
drivers/scsi/smartpqi/smartpqi_init.c
7560
write_stream_cnt += per_cpu_ptr(device->raid_io_stats, cpu)->write_stream_cnt;
drivers/soc/ti/knav_qmss_queue.c
447
pushes += per_cpu_ptr(qh->stats, cpu)->pushes;
drivers/soc/ti/knav_qmss_queue.c
448
pops += per_cpu_ptr(qh->stats, cpu)->pops;
drivers/soc/ti/knav_qmss_queue.c
449
push_errors += per_cpu_ptr(qh->stats, cpu)->push_errors;
drivers/soc/ti/knav_qmss_queue.c
450
pop_errors += per_cpu_ptr(qh->stats, cpu)->pop_errors;
drivers/soc/ti/knav_qmss_queue.c
451
notifies += per_cpu_ptr(qh->stats, cpu)->notifies;
drivers/spi/spi.c
107
stat = per_cpu_ptr(pcpu_stats, cpu);
drivers/spi/spi.c
126
pcpu_stats = per_cpu_ptr(stat, i);
drivers/target/target_core_stat.c
289
stats = per_cpu_ptr(per_cpu_stats, cpu); \
drivers/tee/optee/smc_abi.c
1138
per_cpu_ptr(optee_pcpu, cpu)->optee = optee;
drivers/xen/time.c
63
state = per_cpu_ptr(&xen_runstate, cpu);
drivers/xen/xen-acpi-processor.c
452
free_cpumask_var(per_cpu_ptr(acpi_perf_data, i)
drivers/xen/xen-acpi-processor.c
536
&per_cpu_ptr(acpi_perf_data, i)->shared_cpu_map,
drivers/xen/xen-acpi-processor.c
551
perf = per_cpu_ptr(acpi_perf_data, i);
fs/buffer.c
1437
struct bh_lru *b = per_cpu_ptr(&bh_lrus, cpu);
fs/ext4/mballoc.c
3836
lg = per_cpu_ptr(sbi->s_locality_groups, i);
fs/gfs2/glock.c
2417
const struct gfs2_pcpu_lkstats *lkstats = per_cpu_ptr(sdp->sd_lkstats, i);
fs/gfs2/rgrp.c
1937
st = &per_cpu_ptr(sdp->sd_lkstats, cpu)->lkstats[LM_TYPE_RGRP];
fs/locks.c
3072
struct file_lock_list_struct *fll = per_cpu_ptr(&file_lock_list, i);
fs/locks.c
722
fll = per_cpu_ptr(&file_lock_list, flc->flc_link_cpu);
fs/namespace.c
276
count += per_cpu_ptr(mnt->mnt_pcp, cpu)->mnt_count;
fs/namespace.c
390
count += per_cpu_ptr(mnt->mnt_pcp, cpu)->mnt_writers;
fs/nfs/super.c
708
stats = per_cpu_ptr(nfss->io_stats, cpu);
fs/seq_file.c
1101
hlist_for_each(node, per_cpu_ptr(head, *cpu)) {
fs/seq_file.c
1132
struct hlist_head *bucket = per_cpu_ptr(head, *cpu);
fs/squashfs/decompressor_multi_percpu.c
39
stream = per_cpu_ptr(percpu, cpu);
fs/squashfs/decompressor_multi_percpu.c
53
stream = per_cpu_ptr(percpu, cpu);
fs/squashfs/decompressor_multi_percpu.c
70
stream = per_cpu_ptr(percpu, cpu);
fs/xfs/xfs_icache.c
2290
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
2314
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
465
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
487
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_log_cil.c
127
cilpcp = per_cpu_ptr(cil->xc_pcp, cpu);
fs/xfs/xfs_log_cil.c
173
struct xlog_cil_pcp *cilpcp = per_cpu_ptr(cil->xc_pcp, cpu);
fs/xfs/xfs_log_cil.c
2028
cilpcp = per_cpu_ptr(cil->xc_pcp, cpu);
fs/xfs/xfs_stats.c
112
xs_inodes_active = per_cpu_ptr(stats, c)->s.xs_inodes_active;
fs/xfs/xfs_stats.c
113
xs_inodes_meta = per_cpu_ptr(stats, c)->s.xs_inodes_meta;
fs/xfs/xfs_stats.c
114
memset(per_cpu_ptr(stats, c), 0, sizeof(*stats));
fs/xfs/xfs_stats.c
115
per_cpu_ptr(stats, c)->s.xs_inodes_active = xs_inodes_active;
fs/xfs/xfs_stats.c
116
per_cpu_ptr(stats, c)->s.xs_inodes_meta = xs_inodes_meta;
fs/xfs/xfs_stats.c
15
val += *(((__u32 *)per_cpu_ptr(stats, cpu) + idx));
fs/xfs/xfs_stats.c
79
xs_xstrat_bytes += per_cpu_ptr(stats, i)->s.xs_xstrat_bytes;
fs/xfs/xfs_stats.c
80
xs_write_bytes += per_cpu_ptr(stats, i)->s.xs_write_bytes;
fs/xfs/xfs_stats.c
81
xs_read_bytes += per_cpu_ptr(stats, i)->s.xs_read_bytes;
fs/xfs/xfs_stats.c
82
xs_defer_relog += per_cpu_ptr(stats, i)->s.xs_defer_relog;
fs/xfs/xfs_stats.c
83
xs_gc_bytes += per_cpu_ptr(stats, i)->s.xs_gc_bytes;
fs/xfs/xfs_stats.h
177
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v++; \
fs/xfs/xfs_stats.h
178
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
fs/xfs/xfs_stats.h
183
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v--; \
fs/xfs/xfs_stats.h
184
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
fs/xfs/xfs_stats.h
189
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
fs/xfs/xfs_stats.h
190
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
fs/xfs/xfs_stats.h
195
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]++; \
fs/xfs/xfs_stats.h
196
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
fs/xfs/xfs_stats.h
201
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off]; \
fs/xfs/xfs_stats.h
202
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]; \
fs/xfs/xfs_stats.h
207
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off] += (inc); \
fs/xfs/xfs_stats.h
208
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off] += (inc); \
fs/xfs/xfs_super.c
1197
gc = per_cpu_ptr(mp->m_inodegc, cpu);
include/linux/alloc_tag.h
144
counter = per_cpu_ptr(tag->counters, cpu);
include/linux/context_tracking_state.h
103
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
include/linux/context_tracking_state.h
115
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
include/linux/context_tracking_state.h
127
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
include/linux/context_tracking_state.h
96
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
include/linux/netdevice.h
3085
stat = per_cpu_ptr(pcpu_stats, __cpu); \
include/linux/netdevice.h
3102
stat = per_cpu_ptr(pcpu_stats, __cpu); \
include/linux/netfilter/x_tables.h
435
return per_cpu_ptr((void __percpu *) (unsigned long) cnt->pcnt, cpu);
include/linux/part_stat.h
29
(per_cpu_ptr((part)->bd_stats, (cpu))->field)
include/linux/part_stat.h
39
res += per_cpu_ptr((part)->bd_stats, _cpu)->field; \
include/linux/part_stat.h
48
memset(per_cpu_ptr(part->bd_stats, i), value,
include/linux/percpu-defs.h
268
#define raw_cpu_ptr(ptr) per_cpu_ptr(ptr, 0)
include/linux/percpu-defs.h
273
#define per_cpu(var, cpu) (*per_cpu_ptr(&(var), cpu))
include/linux/vmstat.h
229
x += per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_stat_diff[item];
include/net/ip.h
307
return *(((unsigned long *)per_cpu_ptr(mib, cpu)) + offt);
include/net/sch_generic.h
549
qlen += per_cpu_ptr(q->cpu_qstats, i)->qlen;
include/rv/da_monitor.h
251
da_mon = per_cpu_ptr(&DA_MON_NAME, cpu);
include/uapi/linux/bpf.h
6056
FN(per_cpu_ptr, 153, ##ctx) \
kernel/bpf/arraymap.c
307
return per_cpu_ptr(array->pptrs[index & array->index_mask], cpu);
kernel/bpf/arraymap.c
330
copy_map_value(map, value, per_cpu_ptr(pptr, cpu));
kernel/bpf/arraymap.c
335
copy_map_value_long(map, value + off, per_cpu_ptr(pptr, cpu));
kernel/bpf/arraymap.c
435
ptr = per_cpu_ptr(pptr, cpu);
kernel/bpf/arraymap.c
441
ptr = per_cpu_ptr(pptr, cpu);
kernel/bpf/arraymap.c
488
bpf_obj_free_fields(map->record, per_cpu_ptr(pptr, cpu));
kernel/bpf/arraymap.c
543
per_cpu_ptr(pptr, cpu), m);
kernel/bpf/arraymap.c
676
per_cpu_ptr(pptr, cpu));
kernel/bpf/bpf_lru_list.c
405
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
438
loc_l = per_cpu_ptr(clru->local_list, cpu);
kernel/bpf/bpf_lru_list.c
467
steal_loc_l = per_cpu_ptr(clru->local_list, steal);
kernel/bpf/bpf_lru_list.c
512
loc_l = per_cpu_ptr(lru->common_lru.local_list, node->cpu);
kernel/bpf/bpf_lru_list.c
539
l = per_cpu_ptr(lru->percpu_lru, node->cpu);
kernel/bpf/bpf_lru_list.c
592
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
659
l = per_cpu_ptr(lru->percpu_lru, cpu);
kernel/bpf/bpf_lru_list.c
673
loc_l = per_cpu_ptr(clru->local_list, cpu);
kernel/bpf/core.c
171
pstats = per_cpu_ptr(prog->stats, cpu);
kernel/bpf/cpumap.c
454
bq = per_cpu_ptr(rcpu->bulkq, i);
kernel/bpf/devmap.c
1145
bq = per_cpu_ptr(netdev->xdp_bulkq, cpu);
kernel/bpf/hashtab.c
1030
ptr = per_cpu_ptr(pptr, cpu);
kernel/bpf/hashtab.c
1037
ptr = per_cpu_ptr(pptr, cpu);
kernel/bpf/hashtab.c
1059
copy_map_value_long(&htab->map, per_cpu_ptr(pptr, cpu), value);
kernel/bpf/hashtab.c
1061
zero_map_value(&htab->map, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
1706
copy_map_value_long(&htab->map, value + off, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
1903
copy_map_value(&htab->map, dst_val, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
1908
per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
2187
per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
2436
return per_cpu_ptr(htab_elem_get_ptr(l, map->key_size), cpu);
kernel/bpf/hashtab.c
2463
return per_cpu_ptr(htab_elem_get_ptr(l, map->key_size), cpu);
kernel/bpf/hashtab.c
2493
copy_map_value(map, value, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
2498
copy_map_value_long(map, value + off, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
2546
per_cpu_ptr(pptr, cpu), m);
kernel/bpf/hashtab.c
256
bpf_obj_free_fields(htab->map.record, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
400
*per_cpu_ptr(pptr, cpu) = l_new;
kernel/bpf/hashtab.c
488
bpf_obj_free_fields(hrec->record, per_cpu_ptr(pptr, cpu));
kernel/bpf/hashtab.c
848
bpf_obj_free_fields(htab->map.record, per_cpu_ptr(pptr, cpu));
kernel/bpf/helpers.c
721
return (unsigned long)per_cpu_ptr((const void __percpu *)(const uintptr_t)ptr, cpu);
kernel/bpf/local_storage.c
203
copy_map_value(_map, value, per_cpu_ptr(storage->percpu_buf, cpu));
kernel/bpf/local_storage.c
208
copy_map_value_long(_map, value + off, per_cpu_ptr(storage->percpu_buf, cpu));
kernel/bpf/local_storage.c
243
copy_map_value(_map, per_cpu_ptr(storage->percpu_buf, cpu), value);
kernel/bpf/local_storage.c
249
copy_map_value(_map, per_cpu_ptr(storage->percpu_buf, cpu), val);
kernel/bpf/local_storage.c
444
per_cpu_ptr(storage->percpu_buf, cpu),
kernel/bpf/map_iter.c
208
pcount = per_cpu_ptr(map->elem_count, cpu);
kernel/bpf/memalloc.c
1037
c = per_cpu_ptr(ma->cache, cpu);
kernel/bpf/memalloc.c
1044
cc = per_cpu_ptr(ma->caches, cpu);
kernel/bpf/memalloc.c
544
c = per_cpu_ptr(pc, cpu);
kernel/bpf/memalloc.c
564
cc = per_cpu_ptr(pcc, cpu);
kernel/bpf/memalloc.c
614
cc = per_cpu_ptr(pcc, cpu);
kernel/bpf/memalloc.c
670
c = per_cpu_ptr(ma->cache, cpu);
kernel/bpf/memalloc.c
676
cc = per_cpu_ptr(ma->caches, cpu);
kernel/bpf/memalloc.c
758
c = per_cpu_ptr(ma->cache, cpu);
kernel/bpf/memalloc.c
771
cc = per_cpu_ptr(ma->caches, cpu);
kernel/bpf/percpu_freelist.c
107
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
15
struct pcpu_freelist_head *head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
58
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/percpu_freelist.c
89
head = per_cpu_ptr(s->freelist, cpu);
kernel/bpf/rqspinlock.c
135
struct rqspinlock_held *rqh_cpu = per_cpu_ptr(&rqspinlock_held_locks, cpu);
kernel/bpf/syscall.c
2470
st = per_cpu_ptr(prog->stats, cpu);
kernel/bpf/task_iter.c
1051
work = per_cpu_ptr(&mmap_unlock_work, cpu);
kernel/cgroup/cgroup.c
7134
init_llist_head(per_cpu_ptr(&cgrp_dead_tasks, cpu));
kernel/cgroup/rstat.c
31
return per_cpu_ptr(css->rstat_cpu, cpu);
kernel/cgroup/rstat.c
37
return per_cpu_ptr(cgrp->rstat_base_cpu, cpu);
kernel/cgroup/rstat.c
51
return per_cpu_ptr(ss->lhead, cpu);
kernel/cgroup/rstat.c
52
return per_cpu_ptr(&rstat_backlog_list, cpu);
kernel/cpu.c
1135
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1184
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1215
st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1304
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1406
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1578
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1618
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
173
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
1770
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2381
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2417
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2501
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2565
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2617
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
2729
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, dev->id);
kernel/cpu.c
2738
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, dev->id);
kernel/cpu.c
2779
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, dev->id);
kernel/cpu.c
2788
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, dev->id);
kernel/cpu.c
2838
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, dev->id);
kernel/cpu.c
311
atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu);
kernel/cpu.c
362
atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu);
kernel/cpu.c
404
atomic_t *st = per_cpu_ptr(&cpuhp_state.ap_sync_state, cpu);
kernel/cpu.c
800
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
833
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/cpu.c
865
struct cpuhp_cpu_state *st = per_cpu_ptr(&cpuhp_state, cpu);
kernel/crash_core.c
225
notes_addr = per_cpu_ptr_to_phys(per_cpu_ptr(crash_notes, cpu));
kernel/crash_core.c
463
buf = (u32 *)per_cpu_ptr(crash_notes, cpu);
kernel/events/core.c
1101
cpuctx = per_cpu_ptr(&perf_cpu_context, cpu);
kernel/events/core.c
12565
cpc = *per_cpu_ptr(pmu->cpu_pmu_context, cpu);
kernel/events/core.c
12745
cpc = *per_cpu_ptr(pmu->cpu_pmu_context, cpu);
kernel/events/core.c
12808
*per_cpu_ptr(pmu->cpu_pmu_context, cpu) = cpc;
kernel/events/core.c
13152
struct pmu_event_list *pel = per_cpu_ptr(&pmu_sb_events, event->cpu);
kernel/events/core.c
13991
struct perf_cpu_context *cpuctx = per_cpu_ptr(&perf_cpu_context, event->cpu);
kernel/events/core.c
14398
src_ctx = &per_cpu_ptr(&perf_cpu_context, src_cpu)->ctx;
kernel/events/core.c
14399
dst_ctx = &per_cpu_ptr(&perf_cpu_context, dst_cpu)->ctx;
kernel/events/core.c
15048
cpuctx = per_cpu_ptr(&perf_cpu_context, cpu);
kernel/events/core.c
15133
cpuctx = per_cpu_ptr(&perf_cpu_context, cpu);
kernel/events/core.c
15194
cpuctx = per_cpu_ptr(&perf_cpu_context, cpu);
kernel/events/core.c
5107
cpuctx = per_cpu_ptr(&perf_cpu_context, event->cpu);
kernel/events/core.c
5180
cpc = *per_cpu_ptr(pmu->cpu_pmu_context, event->cpu);
kernel/events/core.c
5314
struct pmu_event_list *pel = per_cpu_ptr(&pmu_sb_events, event->cpu);
kernel/events/core.c
944
t = per_cpu_ptr(event->cgrp->info, event->cpu);
kernel/events/core.c
952
t = per_cpu_ptr(event->cgrp->info, event->cpu);
kernel/events/hw_breakpoint.c
64
return per_cpu_ptr(bp_cpuinfo + type, cpu);
kernel/fork.c
295
struct vm_struct **cached_vm_stack_areas = per_cpu_ptr(cached_stacks, cpu);
kernel/futex/core.c
1587
unsigned int *ptr = per_cpu_ptr(mm->futex_ref, cpu);
kernel/irq/ipi-mux.c
62
icpu = per_cpu_ptr(ipi_mux_pcpu, cpu);
kernel/irq/irqdesc.c
631
*per_cpu_ptr(desc->kstat_irqs, cpu) = (struct irqstat) { };
kernel/irq/matrix.c
144
cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
165
cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
221
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
262
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
305
cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
400
cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
427
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
510
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
kernel/irq/matrix.c
72
struct cpumap *cm = per_cpu_ptr(m->maps, cpu);
kernel/kcov.c
1129
per_cpu_ptr(&kcov_percpu_data, cpu)->irq_area = area;
kernel/locking/lock_events.c
101
unsigned long *ptr = per_cpu_ptr(lockevents, cpu);
kernel/locking/osq_lock.c
41
return per_cpu_ptr(&osq_node, cpu_nr);
kernel/locking/qspinlock.h
68
return per_cpu_ptr(&qnodes[idx].mcs, cpu);
kernel/module/main.c
495
memcpy(per_cpu_ptr(mod->percpu, cpu), from, size);
kernel/module/main.c
510
void *start = per_cpu_ptr(mod->percpu, cpu);
kernel/module/main.c
517
per_cpu_ptr(mod->percpu,
kernel/padata.c
254
reorder = per_cpu_ptr(pd->reorder_list, cpu);
kernel/padata.c
302
squeue = per_cpu_ptr(pd->squeue, cb_cpu);
kernel/padata.c
363
struct padata_list *reorder = per_cpu_ptr(pd->reorder_list, hashed_cpu);
kernel/padata.c
516
squeue = per_cpu_ptr(pd->squeue, cpu);
kernel/padata.c
531
list = per_cpu_ptr(pd->reorder_list, cpu);
kernel/rcu/srcutree.c
1326
sdp = per_cpu_ptr(ssp->sda, get_boot_cpu_id());
kernel/rcu/srcutree.c
1717
srcu_barrier_one_cpu(ssp, per_cpu_ptr(ssp->sda, get_boot_cpu_id()));
kernel/rcu/srcutree.c
1720
srcu_barrier_one_cpu(ssp, per_cpu_ptr(ssp->sda, cpu));
kernel/rcu/srcutree.c
186
sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
2057
sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
475
struct srcu_data *sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
499
struct srcu_data *sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
616
struct srcu_data *sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
726
struct srcu_data *sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
903
srcu_schedule_cbs_sdp(per_cpu_ptr(ssp->sda, cpu), delay);
kernel/rcu/srcutree.c
953
srcu_schedule_cbs_sdp(per_cpu_ptr(ssp->sda, get_boot_cpu_id()),
kernel/rcu/srcutree.c
982
sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/srcutree.c
99
sdp = per_cpu_ptr(ssp->sda, cpu);
kernel/rcu/tasks.h
1053
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, cpu);
kernel/rcu/tasks.h
1314
rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, t->rcu_tasks_exit_cpu);
kernel/rcu/tasks.h
267
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
361
rtpcp = per_cpu_ptr(rtp->rtpcpu, chosen_cpu);
kernel/rcu/tasks.h
439
rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
470
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
525
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
624
rcu_tasks_invoke_cbs(rtp, per_cpu_ptr(rtp->rtpcpu, 0));
kernel/rcu/tasks.h
635
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
726
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tasks.h
773
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rtp->rtpcpu, cpu);
kernel/rcu/tree.c
2206
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
231
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
2765
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
3794
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
3861
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
3903
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4012
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4144
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4243
struct context_tracking *ct = per_cpu_ptr(&context_tracking, cpu);
kernel/rcu/tree.c
4244
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4293
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4308
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4343
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4452
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4520
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4539
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4728
per_cpu_ptr(&rcu_data, i)->mynode = rnp;
kernel/rcu/tree.c
4729
per_cpu_ptr(&rcu_data, i)->barrier_head.next =
kernel/rcu/tree.c
4730
&per_cpu_ptr(&rcu_data, i)->barrier_head;
kernel/rcu/tree.c
773
smp_store_release(per_cpu_ptr(&rcu_data.rcu_urgent_qs, cpu), true);
kernel/rcu/tree.c
86
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
1031
rdp = per_cpu_ptr(&rcu_data, raw_smp_processor_id());
kernel/rcu/tree_exp.h
255
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
303
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, raw_smp_processor_id());
kernel/rcu/tree_exp.h
372
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
407
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
581
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
646
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1154
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1215
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1250
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1342
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1372
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1466
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_plugin.h
1114
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_plugin.h
894
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
481
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
522
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
585
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
979
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
989
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/relay.c
202
*per_cpu_ptr(chan->buf, buf->cpu) = NULL;
kernel/relay.c
332
if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0))) {
kernel/relay.c
339
if ((buf = *per_cpu_ptr(chan->buf, i)))
kernel/relay.c
386
return *per_cpu_ptr(chan->buf, 0);
kernel/relay.c
410
*per_cpu_ptr(chan->buf, 0) = buf;
kernel/relay.c
444
if (*per_cpu_ptr(chan->buf, cpu))
kernel/relay.c
452
*per_cpu_ptr(chan->buf, cpu) = buf;
kernel/relay.c
520
*per_cpu_ptr(chan->buf, i) = buf;
kernel/relay.c
529
if ((buf = *per_cpu_ptr(chan->buf, i)))
kernel/relay.c
632
buf = *per_cpu_ptr(chan->buf, cpu);
kernel/relay.c
658
if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0)))
kernel/relay.c
662
if ((buf = *per_cpu_ptr(chan->buf, i)))
kernel/relay.c
685
if (chan->is_global && (buf = *per_cpu_ptr(chan->buf, 0))) {
kernel/relay.c
692
if ((buf = *per_cpu_ptr(chan->buf, i)))
kernel/relay.c
714
rbuf = *per_cpu_ptr(chan->buf, 0);
kernel/relay.c
721
rbuf = *per_cpu_ptr(chan->buf, i);
kernel/rseq.c
831
hrtimer_setup(per_cpu_ptr(&slice_timer.timer, cpu), rseq_slice_expired,
kernel/sched/core.c
10854
struct mm_cid_pcpu *pcp = per_cpu_ptr(mm->mm_cid.pcpu, cpu);
kernel/sched/core.c
10899
mm_cid_transit_to_cpu(t, per_cpu_ptr(mm->mm_cid.pcpu, task_cpu(t)));
kernel/sched/core.c
5781
twork = per_cpu_ptr(tick_work_cpu, cpu);
kernel/sched/core.c
5802
twork = per_cpu_ptr(tick_work_cpu, cpu);
kernel/sched/cpuacct.c
100
u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat;
kernel/sched/cpuacct.c
139
u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
kernel/sched/cpuacct.c
140
u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat;
kernel/sched/cpuacct.c
272
u64 *cpustat = per_cpu_ptr(ca->cpustat, cpu)->cpustat;
kernel/sched/cpuacct.c
280
cputime.sum_exec_runtime += *per_cpu_ptr(ca->cpuusage, cpu);
kernel/sched/cpuacct.c
344
*per_cpu_ptr(ca->cpuusage, cpu) += cputime;
kernel/sched/cpuacct.c
99
u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
kernel/sched/ext.c
4349
struct scx_kick_syncs **ksyncs = per_cpu_ptr(&scx_kick_syncs, cpu);
kernel/sched/ext.c
4899
struct scx_kick_syncs **ksyncs = per_cpu_ptr(&scx_kick_syncs, cpu);
kernel/sched/ext.c
7385
e_cpu = &per_cpu_ptr(sch->pcpu, cpu)->event_stats;
kernel/sched/psi.c
1037
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
183
write_seqcount_begin(per_cpu_ptr(&psi_seq, cpu));
kernel/sched/psi.c
188
write_seqcount_end(per_cpu_ptr(&psi_seq, cpu));
kernel/sched/psi.c
193
return read_seqcount_begin(per_cpu_ptr(&psi_seq, cpu));
kernel/sched/psi.c
198
return read_seqcount_retry(per_cpu_ptr(&psi_seq, cpu), seq);
kernel/sched/psi.c
275
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
805
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
944
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/sched.h
3495
data = rcu_dereference_sched(*per_cpu_ptr(&cpufreq_update_util_data,
kernel/sched/topology.c
1059
sibling = *per_cpu_ptr(sdd->sd, i);
kernel/sched/topology.c
1209
struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu);
kernel/sched/topology.c
1217
sg = *per_cpu_ptr(sdd->sg, cpu);
kernel/sched/topology.c
1218
sg->sgc = *per_cpu_ptr(sdd->sgc, cpu);
kernel/sched/topology.c
1602
if (atomic_read(&(*per_cpu_ptr(d->sds, cpu))->ref))
kernel/sched/topology.c
1603
*per_cpu_ptr(d->sds, cpu) = NULL;
kernel/sched/topology.c
1605
for (sd = *per_cpu_ptr(d->sd, cpu); sd; sd = sd->parent) {
kernel/sched/topology.c
1608
WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd);
kernel/sched/topology.c
1609
*per_cpu_ptr(sdd->sd, cpu) = NULL;
kernel/sched/topology.c
1611
if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref))
kernel/sched/topology.c
1612
*per_cpu_ptr(sdd->sg, cpu) = NULL;
kernel/sched/topology.c
1614
if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref))
kernel/sched/topology.c
1615
*per_cpu_ptr(sdd->sgc, cpu) = NULL;
kernel/sched/topology.c
1666
struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu);
kernel/sched/topology.c
2407
*per_cpu_ptr(sdd->sd, j) = sd;
kernel/sched/topology.c
2416
*per_cpu_ptr(sdd->sg, j) = sg;
kernel/sched/topology.c
2425
*per_cpu_ptr(sdd->sgc, j) = sgc;
kernel/sched/topology.c
2444
sd = *per_cpu_ptr(sdd->sd, j);
kernel/sched/topology.c
2447
kfree(*per_cpu_ptr(sdd->sd, j));
kernel/sched/topology.c
2451
kfree(*per_cpu_ptr(sdd->sg, j));
kernel/sched/topology.c
2453
kfree(*per_cpu_ptr(sdd->sgc, j));
kernel/sched/topology.c
2480
*per_cpu_ptr(d->sds, j) = sds;
kernel/sched/topology.c
2494
kfree(*per_cpu_ptr(d->sds, j));
kernel/sched/topology.c
2687
*per_cpu_ptr(d.sd, i) = sd;
kernel/sched/topology.c
2698
for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent) {
kernel/sched/topology.c
2711
sd = *per_cpu_ptr(d.sd, i);
kernel/sched/topology.c
2722
sd->shared = *per_cpu_ptr(d.sds, sd_id);
kernel/sched/topology.c
2742
for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent)
kernel/sched/topology.c
2750
sd = *per_cpu_ptr(d.sd, i);
kernel/sched/topology.c
938
sibling = *per_cpu_ptr(sdd->sd, i);
kernel/sched/topology.c
999
sg->sgc = *per_cpu_ptr(sdd->sgc, cpu);
kernel/scs.c
96
void **cache = per_cpu_ptr(scs_cache, cpu);
kernel/smp.c
1119
if (!llist_empty(per_cpu_ptr(&call_single_queue, cpu)))
kernel/smp.c
384
return per_cpu_ptr(&csd_data, cpu);
kernel/smp.c
834
call_single_data_t *csd = per_cpu_ptr(cfd->csd, cpu);
kernel/smp.c
891
csd = per_cpu_ptr(cfd->csd, cpu);
kernel/smpboot.c
168
struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
kernel/smpboot.c
193
*per_cpu_ptr(ht->store, cpu) = tsk;
kernel/smpboot.c
226
struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
kernel/smpboot.c
245
struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
kernel/smpboot.c
268
struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
kernel/smpboot.c
272
*per_cpu_ptr(ht->store, cpu) = NULL;
kernel/stop_machine.c
263
struct cpu_stopper *stopper1 = per_cpu_ptr(&cpu_stopper, cpu1);
kernel/stop_machine.c
264
struct cpu_stopper *stopper2 = per_cpu_ptr(&cpu_stopper, cpu2);
kernel/stop_machine.c
58
struct cpu_stopper *stopper = per_cpu_ptr(&cpu_stopper, task_cpu(task));
kernel/time/clocksource.c
394
wd = per_cpu_ptr(&watchdog_cpu_data, cpu);
kernel/time/tick-sched.c
535
ts = per_cpu_ptr(&tick_cpu_sched, cpu);
kernel/time/tick-sched.c
555
struct tick_sched *ts = per_cpu_ptr(&tick_cpu_sched, cpu);
kernel/time/tick-sched.c
725
struct tick_sched *ts = per_cpu_ptr(&tick_cpu_sched, cpu);
kernel/time/timer.c
2065
base_local = per_cpu_ptr(&timer_bases[BASE_LOCAL], cpu);
kernel/time/timer.c
2066
base_global = per_cpu_ptr(&timer_bases[BASE_GLOBAL], cpu);
kernel/time/timer.c
2086
base_local = per_cpu_ptr(&timer_bases[BASE_LOCAL], cpu);
kernel/time/timer.c
2087
base_global = per_cpu_ptr(&timer_bases[BASE_GLOBAL], cpu);
kernel/time/timer.c
2105
base_local = per_cpu_ptr(&timer_bases[BASE_LOCAL], cpu);
kernel/time/timer.c
2106
base_global = per_cpu_ptr(&timer_bases[BASE_GLOBAL], cpu);
kernel/time/timer.c
2134
struct timer_base *base = per_cpu_ptr(&timer_bases[BASE_GLOBAL], cpu);
kernel/time/timer.c
2505
base = per_cpu_ptr(&timer_bases[b], cpu);
kernel/time/timer.c
2522
old_base = per_cpu_ptr(&timer_bases[b], cpu);
kernel/time/timer.c
2558
base = per_cpu_ptr(&timer_bases[i], cpu);
kernel/time/timer.c
925
return per_cpu_ptr(&timer_bases[index], cpu);
kernel/time/timer_migration.c
1576
struct work_struct *work = per_cpu_ptr(works, cpu);
kernel/time/timer_migration.c
1582
flush_work(per_cpu_ptr(works, cpu));
kernel/time/timer_migration.c
1603
struct work_struct *work = per_cpu_ptr(works, cpu);
kernel/time/timer_migration.c
1609
flush_work(per_cpu_ptr(works, cpu));
kernel/time/timer_migration.c
1825
struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
kernel/time/timer_migration.c
1919
WARN_ON_ONCE(!per_cpu_ptr(&tmigr_cpu, raw_smp_processor_id())->available);
kernel/time/timer_migration.c
1928
struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
kernel/time/timer_migration.c
948
tmc = per_cpu_ptr(&tmigr_cpu, cpu);
kernel/trace/blktrace.c
462
sequence = per_cpu_ptr(bt->sequence, cpu);
kernel/trace/bpf_trace.c
2222
work = per_cpu_ptr(&send_signal_work, cpu);
kernel/trace/ftrace.c
8672
per_cpu_ptr(tr->array_buffer.data, cpu)->ftrace_ignore_pid = FTRACE_PID_TRACE;
kernel/trace/trace.c
1558
struct trace_array_cpu *data = per_cpu_ptr(trace_buf->data, cpu);
kernel/trace/trace.c
1560
struct trace_array_cpu *max_data = per_cpu_ptr(max_buf->data, cpu);
kernel/trace/trace.c
3195
per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = 0;
kernel/trace/trace.c
3217
per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = entries;
kernel/trace/trace.c
3304
if (per_cpu_ptr(buf->data, cpu)->skipped_entries) {
kernel/trace/trace.c
3305
count -= per_cpu_ptr(buf->data, cpu)->skipped_entries;
kernel/trace/trace.c
3414
struct trace_array_cpu *data = per_cpu_ptr(buf->data, buf->cpu);
kernel/trace/trace.c
3476
if (per_cpu_ptr(iter->array_buffer->data, iter->cpu)->skipped_entries)
kernel/trace/trace.c
5230
per_cpu_ptr(buf->data, cpu)->entries = val;
kernel/trace/trace.c
5238
per_cpu_ptr(buf->data, cpu)->entries = ring_buffer_size(buf->buffer, cpu);
kernel/trace/trace.c
5252
per_cpu_ptr(size_buf->data, cpu)->entries, cpu);
kernel/trace/trace.c
5255
per_cpu_ptr(trace_buf->data, cpu)->entries =
kernel/trace/trace.c
5256
per_cpu_ptr(size_buf->data, cpu)->entries;
kernel/trace/trace.c
5260
per_cpu_ptr(size_buf->data, cpu_id)->entries, cpu_id);
kernel/trace/trace.c
5262
per_cpu_ptr(trace_buf->data, cpu_id)->entries =
kernel/trace/trace.c
5263
per_cpu_ptr(size_buf->data, cpu_id)->entries;
kernel/trace/trace.c
6267
size = per_cpu_ptr(tr->array_buffer.data, cpu)->entries;
kernel/trace/trace.c
6268
if (size != per_cpu_ptr(tr->array_buffer.data, cpu)->entries) {
kernel/trace/trace.c
6284
r = sprintf(buf, "%lu\n", per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10);
kernel/trace/trace.c
6331
size += per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10;
kernel/trace/trace.c
6574
buf = per_cpu_ptr(tinfo->tbuf, cpu)->buf;
kernel/trace/trace.c
6596
per_cpu_ptr(tinfo->tbuf, cpu)->buf = NULL;
kernel/trace/trace.c
6604
per_cpu_ptr(tinfo->tbuf, cpu)->buf = buf;
kernel/trace/trace.c
6761
char *buffer = per_cpu_ptr(tinfo->tbuf, cpu)->buf;
kernel/trace/trace.c
6800
buffer = per_cpu_ptr(tinfo->tbuf, cpu)->buf;
kernel/trace/trace_event_perf.c
112
INIT_HLIST_HEAD(per_cpu_ptr(list, cpu));
kernel/trace/trace_events.c
1200
per_cpu_ptr(tr->array_buffer.data, cpu)->ignore_pid = false;
kernel/trace/trace_events_hist.c
5356
hist_pad = per_cpu_ptr(hist_pads, smp_processor_id());
kernel/trace/trace_functions.c
293
data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_functions.c
404
data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_functions.c
408
last_info = per_cpu_ptr(tr->last_func_repeats, cpu);
kernel/trace/trace_functions_graph.c
1035
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_functions_graph.c
1154
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
1200
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
1310
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_functions_graph.c
1382
depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth;
kernel/trace/trace_functions_graph.c
1448
if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) {
kernel/trace/trace_functions_graph.c
1449
per_cpu_ptr(data->cpu_data, cpu)->ignore = 0;
kernel/trace/trace_functions_graph.c
1462
per_cpu_ptr(data->cpu_data, iter->cpu)->ignore = 1;
kernel/trace/trace_functions_graph.c
1632
pid_t *pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
kernel/trace/trace_functions_graph.c
1633
int *depth = &(per_cpu_ptr(data->cpu_data, cpu)->depth);
kernel/trace/trace_functions_graph.c
1634
int *ignore = &(per_cpu_ptr(data->cpu_data, cpu)->ignore);
kernel/trace/trace_functions_graph.c
1635
int *depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
610
last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
kernel/trace/trace_functions_graph.c
971
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_irqsoff.c
125
*data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_irqsoff.c
407
data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_irqsoff.c
446
data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_kprobe.c
187
nhit += *per_cpu_ptr(tk->nhit, cpu);
kernel/trace/trace_osnoise.c
1202
osn_var = per_cpu_ptr(&per_cpu_osnoise_var, cpu);
kernel/trace/trace_osnoise.c
1204
per_cpu_ptr(&per_cpu_timerlat_var, cpu)->uthread_migrate = 1;
kernel/trace/trace_osnoise.c
2475
per_cpu_ptr(&per_cpu_timerlat_var, cpu)->uthread_migrate = 1;
kernel/trace/trace_osnoise.c
2565
osn_var = per_cpu_ptr(&per_cpu_osnoise_var, cpu);
kernel/trace/trace_osnoise.c
2566
tlat_var = per_cpu_ptr(&per_cpu_timerlat_var, cpu);
kernel/trace/trace_osnoise.c
276
tlat_var = per_cpu_ptr(&per_cpu_timerlat_var, cpu);
kernel/trace/trace_osnoise.c
300
osn_var = per_cpu_ptr(&per_cpu_osnoise_var, cpu);
kernel/trace/trace_sched_wakeup.c
470
disabled = local_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);
kernel/trace/trace_sched_wakeup.c
484
data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu);
kernel/trace/trace_sched_wakeup.c
507
local_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);
kernel/trace/trace_sched_wakeup.c
562
disabled = local_inc_return(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);
kernel/trace/trace_sched_wakeup.c
594
data = per_cpu_ptr(wakeup_trace->array_buffer.data, wakeup_cpu);
kernel/trace/trace_sched_wakeup.c
609
local_dec(&per_cpu_ptr(wakeup_trace->array_buffer.data, cpu)->disabled);
kernel/trace/trace_sched_wakeup.c
85
*data = per_cpu_ptr(tr->array_buffer.data, cpu);
kernel/trace/trace_uprobe.c
894
per_cpu_ptr(uprobe_cpu_buffer, cpu)->buf = page_address(p);
kernel/trace/trace_uprobe.c
895
mutex_init(&per_cpu_ptr(uprobe_cpu_buffer, cpu)->mutex);
kernel/trace/trace_uprobe.c
904
free_page((unsigned long)per_cpu_ptr(uprobe_cpu_buffer, cpu)->buf);
kernel/trace/trace_uprobe.c
934
free_page((unsigned long)per_cpu_ptr(uprobe_cpu_buffer,
kernel/trace/trace_uprobe.c
948
ucb = per_cpu_ptr(uprobe_cpu_buffer, cpu);
kernel/workqueue.c
2296
pwq = rcu_dereference(*per_cpu_ptr(wq->cpu_pwq, cpu));
kernel/workqueue.c
4663
struct work_struct *work = per_cpu_ptr(works, cpu);
kernel/workqueue.c
4670
flush_work(per_cpu_ptr(works, cpu));
kernel/workqueue.c
5601
pool = &(per_cpu_ptr(pools, cpu)[highpri]);
kernel/workqueue.c
5602
pwq_p = per_cpu_ptr(wq->cpu_pwq, cpu);
kernel/workqueue.c
5636
struct pool_workqueue *pwq = *per_cpu_ptr(wq->cpu_pwq, cpu);
kernel/workqueue.c
6158
pwq = *per_cpu_ptr(wq->cpu_pwq, cpu);
kernel/workqueue.c
728
return per_cpu_ptr(wq->cpu_pwq, cpu);
lib/debugobjects.c
575
struct obj_pool *pcp = per_cpu_ptr(&pool_pcpu, cpu);
lib/percpu-refcount.c
178
count += *per_cpu_ptr(percpu_count, cpu);
lib/percpu-refcount.c
259
*per_cpu_ptr(percpu_count, cpu) = 0;
lib/percpu_counter.c
180
s32 *pcount = per_cpu_ptr(fbc->counters, cpu);
lib/percpu_counter.c
278
pcount = per_cpu_ptr(fbc->counters, cpu);
lib/percpu_counter.c
372
pcount = per_cpu_ptr(fbc->counters, cpu);
lib/percpu_counter.c
67
s32 *pcount = per_cpu_ptr(fbc->counters, cpu);
lib/random32.c
115
struct rnd_state *state = per_cpu_ptr(pcpu_state, i);
lib/sbitmap.c
24
*per_cpu_ptr(sb->alloc_hint, i) = get_random_u32_below(depth);
lib/sbitmap.c
639
data_race(*per_cpu_ptr(sb->alloc_hint, cpu) = tag);
lib/sbitmap.c
738
seq_printf(m, "%u", *per_cpu_ptr(sbq->sb.alloc_hint, i));
lib/test_context-analysis.c
607
raw_spin_lock(per_cpu_ptr(&test_per_cpu_lock, cpu));
lib/test_context-analysis.c
608
raw_spin_unlock(per_cpu_ptr(&test_per_cpu_lock, cpu));
lib/test_lockup.c
408
work == per_cpu_ptr(&test_works, master_cpu));
lib/test_lockup.c
584
INIT_WORK(per_cpu_ptr(&test_works, cpu), test_work_fn);
lib/test_lockup.c
586
per_cpu_ptr(&test_works, cpu));
lib/test_lockup.c
591
flush_work(per_cpu_ptr(&test_works, cpu));
lib/test_objpool.c
250
struct ot_item *item = per_cpu_ptr(&ot_pcup_items, cpu);
lib/test_objpool.c
364
struct ot_item *item = per_cpu_ptr(&ot_pcup_items, cpu);
lib/test_objpool.c
551
struct ot_item *item = per_cpu_ptr(&ot_pcup_items, cpu);
mm/kasan/quarantine.c
359
sq = per_cpu_ptr(&shrink_qlist, cpu);
mm/kmemleak.c
1400
void *ptr = per_cpu_ptr((void __percpu *)object->pointer, cpu);
mm/kmemleak.c
1586
void *start = per_cpu_ptr((void __percpu *)object->pointer, cpu);
mm/memcontrol.c
3779
statc = per_cpu_ptr(memcg->vmstats_percpu, cpu);
mm/memcontrol.c
4115
statc = per_cpu_ptr(memcg->vmstats_percpu, cpu);
mm/memcontrol.c
4148
lstatc = per_cpu_ptr(pn->lruvec_stats_percpu, cpu);
mm/memcontrol.c
5182
INIT_WORK(&per_cpu_ptr(&memcg_stock, cpu)->work,
mm/memcontrol.c
5184
INIT_WORK(&per_cpu_ptr(&obj_stock, cpu)->work,
mm/mm_init.c
1575
p = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu);
mm/page_alloc.c
2648
struct per_cpu_pages *pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
mm/page_alloc.c
2740
pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
mm/page_alloc.c
2745
pcp = per_cpu_ptr(z->per_cpu_pageset, cpu);
mm/page_alloc.c
6073
pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
mm/page_alloc.c
6127
pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
mm/page_alloc.c
6128
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
mm/page_alloc.c
6152
pcp = per_cpu_ptr(zone->per_cpu_pageset, cpu);
mm/page_alloc.c
7408
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
mm/percpu.c
2298
void *start = per_cpu_ptr(base, cpu);
mm/percpu.c
2305
per_cpu_ptr(base, get_boot_cpu_id());
mm/percpu.c
2378
void *start = per_cpu_ptr(base, cpu);
mm/shmem.c
395
next_ino = per_cpu_ptr(sbinfo->ino_batch, get_cpu());
mm/show_mem.c
194
free_pcp += per_cpu_ptr(zone->per_cpu_pageset, cpu)->count;
mm/show_mem.c
295
free_pcp += per_cpu_ptr(zone->per_cpu_pageset, cpu)->count;
mm/slab_common.c
1922
struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu);
mm/slab_common.c
2032
krcp = per_cpu_ptr(&krc, cpu);
mm/slab_common.c
2071
krcp = per_cpu_ptr(&krc, cpu);
mm/slab_common.c
2133
struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu);
mm/slab_common.c
2150
struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu);
mm/slab_common.c
2189
struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu);
mm/slub.c
3023
pcs = per_cpu_ptr(s->cpu_sheaves, cpu);
mm/slub.c
3057
pcs = per_cpu_ptr(s->cpu_sheaves, cpu);
mm/slub.c
3952
pcs = per_cpu_ptr(s->cpu_sheaves, cpu);
mm/slub.c
6155
irq_work_sync(&per_cpu_ptr(&defer_free_objects, cpu)->work);
mm/slub.c
7469
pcs = per_cpu_ptr(s->cpu_sheaves, cpu);
mm/slub.c
8329
pcs = per_cpu_ptr(s->cpu_sheaves, cpu);
mm/slub.c
9189
unsigned int x = per_cpu_ptr(s->cpu_stats, cpu)->stat[si];
mm/slub.c
9215
per_cpu_ptr(s->cpu_stats, cpu)->stat[si] = 0;
mm/swapfile.c
2757
pcp_si = per_cpu_ptr(percpu_swap_cluster.si, cpu);
mm/vmalloc.c
2757
vbq = per_cpu_ptr(&vmap_block_queue, vb->cpu);
mm/vmstat.c
181
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
mm/vmstat.c
1818
pcp = per_cpu_ptr(zone->per_cpu_pageset, i);
mm/vmstat.c
1833
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, i);
mm/vmstat.c
2058
struct per_cpu_zonestat *pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
mm/vmstat.c
2070
n = per_cpu_ptr(zone->zone_pgdat->per_cpu_nodestats, cpu);
mm/vmstat.c
2158
INIT_DEFERRABLE_WORK(per_cpu_ptr(&vmstat_work, cpu),
mm/vmstat.c
285
per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0;
mm/vmstat.c
298
per_cpu_ptr(zone->per_cpu_zonestats, cpu)->stat_threshold
mm/vmstat.c
302
pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold;
mm/vmstat.c
303
per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold
mm/vmstat.c
335
per_cpu_ptr(zone->per_cpu_zonestats, cpu)->stat_threshold
mm/vmstat.c
47
per_cpu_ptr(zone->per_cpu_zonestats, cpu)->vm_numa_event[item]
mm/vmstat.c
900
pzstats = per_cpu_ptr(zone->per_cpu_zonestats, cpu);
mm/vmstat.c
928
p = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu);
mm/zswap.c
273
mutex_init(&per_cpu_ptr(pool->acomp_ctx, cpu)->mutex);
mm/zswap.c
738
struct crypto_acomp_ctx *acomp_ctx = per_cpu_ptr(pool->acomp_ctx, cpu);
mm/zswap.c
798
struct crypto_acomp_ctx *acomp_ctx = per_cpu_ptr(pool->acomp_ctx, cpu);
net/8021q/vlan_dev.c
675
p = per_cpu_ptr(vlan_dev_priv(dev)->vlan_pcpu_stats, i);
net/batman-adv/mesh-interface.c
92
counters = per_cpu_ptr(bat_priv->bat_counters, cpu);
net/bridge/br_multicast.c
5189
struct bridge_mcast_stats *cpu_stats = per_cpu_ptr(stats, i);
net/bridge/br_vlan.c
1394
cpu_stats = per_cpu_ptr(v->stats, i);
net/caif/caif_dev.c
88
refcnt += *per_cpu_ptr(e->pcpu_refcnt, i);
net/caif/cffrml.c
202
refcnt += *per_cpu_ptr(this->pcpu_refcnt, i);
net/core/dev.c
11571
refcnt += *per_cpu_ptr(dev->pcpu_refcnt, i);
net/core/dev.c
11770
stats = per_cpu_ptr(dstats, cpu);
net/core/dev.c
11905
core_stats = per_cpu_ptr(p, i);
net/core/dev.c
11933
stats = per_cpu_ptr(netstats, cpu);
net/core/dev.c
13187
struct softnet_data *sd = per_cpu_ptr(&softnet_data, cpu);
net/core/dev.c
13195
struct softnet_data *sd = per_cpu_ptr(&softnet_data, cpu);
net/core/dev.c
13202
struct softnet_data *sd = per_cpu_ptr(&softnet_data, cpu);
net/core/dst.c
333
__metadata_dst_init(per_cpu_ptr(md_dst, cpu), type, optslen);
net/core/dst.c
344
struct metadata_dst *one_md_dst = per_cpu_ptr(md_dst, cpu);
net/core/dst_cache.c
172
local_lock_init(&per_cpu_ptr(dst_cache->cache, i)->bh_lock);
net/core/dst_cache.c
187
dst_release(per_cpu_ptr(dst_cache->cache, i)->dst);
net/core/dst_cache.c
202
struct dst_cache_pcpu *idst = per_cpu_ptr(dst_cache->cache, i);
net/core/gen_stats.c
133
struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i);
net/core/gen_stats.c
185
struct gnet_stats_basic_sync *bcpu = per_cpu_ptr(cpu, i);
net/core/gen_stats.c
346
const struct gnet_stats_queue *qcpu = per_cpu_ptr(q, i);
net/core/gro_cells.c
123
struct gro_cell *cell = per_cpu_ptr(gcells->cells, i);
net/core/gro_cells.c
87
struct gro_cell *cell = per_cpu_ptr(gcells->cells, i);
net/core/neighbour.c
2301
st = per_cpu_ptr(tbl->stats, cpu);
net/core/neighbour.c
3496
return per_cpu_ptr(tbl->stats, cpu);
net/core/neighbour.c
3510
return per_cpu_ptr(tbl->stats, cpu);
net/core/page_pool.c
103
per_cpu_ptr(pool->recycle_stats, cpu);
net/core/skbuff.c
7299
sdn = per_cpu_ptr(net_hotdata.skb_defer_nodes, cpu) + numa_node_id();
net/core/sock.c
4036
res += per_cpu_ptr(net->core.prot_inuse, cpu)->val[idx];
net/core/sock.c
4047
res += per_cpu_ptr(net->core.prot_inuse, cpu)->all;
net/devlink/trap.c
167
cpu_stats = per_cpu_ptr(trap_stats, i);
net/dsa/user.c
1057
s = per_cpu_ptr(dev->tstats, i);
net/ipv4/af_inet.c
1681
bhptr = per_cpu_ptr(mib, cpu);
net/ipv4/af_inet.c
1730
af_inet_stats = per_cpu_ptr(net->mib.ip_statistics, i);
net/ipv4/fib_semantics.c
193
rt = rcu_dereference_protected(*per_cpu_ptr(rtp, cpu), 1);
net/ipv4/fib_trie.c
2578
const struct trie_use_stats *pcpu = per_cpu_ptr(stats, cpu);
net/ipv4/nexthop.c
695
cpu_stats = per_cpu_ptr(nhge->stats, i);
net/ipv4/route.c
323
src = (struct ip_rt_acct *)per_cpu_ptr(ip_rt_acct, i);
net/ipv4/route.c
731
prt = per_cpu_ptr(nhc->nhc_pcpu_rth_output, i);
net/ipv6/addrconf.c
353
addrconf_stats = per_cpu_ptr(idev->stats.ipv6, i);
net/ipv6/addrconf.c
6685
rtp = per_cpu_ptr(nh->rt6i_pcpu, cpu);
net/ipv6/af_inet6.c
908
af_inet6_stats = per_cpu_ptr(net->mib.ipv6_statistics, i);
net/ipv6/ip6_fib.c
991
ppcpu_rt = per_cpu_ptr(fib6_nh->rt6i_pcpu, cpu);
net/ipv6/route.c
165
struct uncached_list *ul = per_cpu_ptr(&rt6_uncached_list, cpu);
net/ipv6/route.c
3738
ppcpu_rt = per_cpu_ptr(fib6_nh->rt6i_pcpu, cpu);
net/ipv6/route.c
6933
struct uncached_list *ul = per_cpu_ptr(&rt6_uncached_list, cpu);
net/ipv6/seg6_local.c
2041
pcounters = per_cpu_ptr(slwt->pcpu_counters, i);
net/mac80211/sta_info.c
2505
cpustats = per_cpu_ptr(link_sta_info->pcpu_rx_stats, cpu);
net/mac80211/sta_info.c
2633
cpurxs = per_cpu_ptr(link_sta_info->pcpu_rx_stats,
net/mac80211/sta_info.c
2815
cpurxs = per_cpu_ptr(link_sta_info->pcpu_rx_stats,
net/mac80211/sta_info.c
2831
cpurxs = per_cpu_ptr(link_sta_info->pcpu_rx_stats,
net/mac80211/sta_info.c
2873
cpurxs = per_cpu_ptr(link_sta_info->pcpu_rx_stats,
net/mac80211/sta_info.c
3046
cpurxs = per_cpu_ptr(sta->deflink.pcpu_rx_stats,
net/mac80211/sta_info.c
3061
cpurxs = per_cpu_ptr(sta->deflink.pcpu_rx_stats,
net/mac80211/sta_info.c
3101
cpurxs = per_cpu_ptr(sta->deflink.pcpu_rx_stats, cpu);
net/mpls/af_mpls.c
1113
p = per_cpu_ptr(mdev->stats, i);
net/mpls/af_mpls.c
1500
mpls_stats = per_cpu_ptr(mdev->stats, i);
net/mptcp/protocol.c
4601
delegated = per_cpu_ptr(&mptcp_delegated_actions, cpu);
net/netfilter/ipvs/ip_vs_ctl.c
2528
struct ip_vs_cpu_stats *u = per_cpu_ptr(cpustats, i);
net/netfilter/ipvs/ip_vs_ctl.c
932
struct ip_vs_cpu_stats *cs = per_cpu_ptr(s->cpustats, i);
net/netfilter/ipvs/ip_vs_est.c
97
c = per_cpu_ptr(s->cpustats, i);
net/netfilter/nf_conntrack_netlink.c
2544
st = per_cpu_ptr(net->ct.stat, cpu);
net/netfilter/nf_conntrack_netlink.c
3731
st = per_cpu_ptr(net->ct.stat, cpu);
net/netfilter/nf_conntrack_standalone.c
414
return per_cpu_ptr(net->ct.stat, cpu);
net/netfilter/nf_conntrack_standalone.c
429
return per_cpu_ptr(net->ct.stat, cpu);
net/netfilter/nf_flow_table_procfs.c
18
return per_cpu_ptr(net->ft.stat, cpu);
net/netfilter/nf_flow_table_procfs.c
33
return per_cpu_ptr(net->ft.stat, cpu);
net/netfilter/nf_synproxy_core.c
253
return per_cpu_ptr(snet->stats, cpu);
net/netfilter/nf_synproxy_core.c
268
return per_cpu_ptr(snet->stats, cpu);
net/netfilter/nf_tables_api.c
1959
cpu_stats = per_cpu_ptr(stats, cpu);
net/netfilter/nft_counter.c
140
struct u64_stats_sync *nft_sync = per_cpu_ptr(&nft_counter_sync, cpu);
net/netfilter/nft_counter.c
142
this_cpu = per_cpu_ptr(priv->counter, cpu);
net/netfilter/nft_counter.c
303
u64_stats_init(per_cpu_ptr(&nft_counter_sync, cpu));
net/netfilter/nft_set_pipapo.c
1201
s = *per_cpu_ptr(m->scratch, cpu);
net/netfilter/nft_set_pipapo.c
1237
*per_cpu_ptr(clone->scratch, i) = scratch;
net/netfilter/nft_set_pipapo.c
1428
*per_cpu_ptr(new->scratch, i) = NULL;
net/netfilter/nft_set_pipapo.c
2292
*per_cpu_ptr(m->scratch, i) = NULL;
net/openvswitch/datapath.c
2762
ovs_pcpu = per_cpu_ptr(ovs_pcpu_storage, cpu);
net/openvswitch/datapath.c
769
percpu_stats = per_cpu_ptr(dp->stats_percpu, i);
net/openvswitch/flow_table.c
1132
stats = per_cpu_ptr(ma->masks_usage_stats, cpu);
net/openvswitch/flow_table.c
205
stats = per_cpu_ptr(ma->masks_usage_stats, cpu);
net/openvswitch/vport.c
321
stats = per_cpu_ptr(vport->upcall_stats, i);
net/packet/af_packet.c
1217
refcnt += *per_cpu_ptr(rb->pending_refcnt, cpu);
net/rds/ib_recv.c
112
head = per_cpu_ptr(cache->percpu, cpu);
net/rds/ib_recv.c
143
head = per_cpu_ptr(cache->percpu, cpu);
net/sched/cls_basic.c
291
struct tc_basic_pcnt *pf = per_cpu_ptr(f->pf, cpu);
net/sched/cls_matchall.c
356
struct tc_matchall_pcnt *pf = per_cpu_ptr(head->pf, cpu);
net/sched/cls_u32.c
1400
__u32 cnt = *per_cpu_ptr(n->pcpu_success, cpum);
net/sched/cls_u32.c
1426
struct tc_u32_pcnt *pf = per_cpu_ptr(n->pf, cpu);
net/sched/sch_generic.c
835
q = per_cpu_ptr(qdisc->cpu_qstats, i);
net/smc/smc_stats.c
293
src = (u64 *)per_cpu_ptr(net->smc.smc_stats, cpu);
net/tipc/crypto.c
2053
stat = per_cpu_ptr(tx->stats, cpu)->stat[i];
net/tipc/crypto.c
2067
stat = per_cpu_ptr(rx->stats, cpu)->stat[i];
net/tipc/crypto.c
576
*per_cpu_ptr(tmp->tfm_entry, cpu) = head;
net/tipc/crypto.c
651
*per_cpu_ptr(aead->tfm_entry, cpu) =
net/tipc/crypto.c
652
*per_cpu_ptr(src->tfm_entry, cpu);
net/xfrm/xfrm_nat_keepalive.c
227
per_cpu_ptr(socks, i)->sock = sk;
net/xfrm/xfrm_nat_keepalive.c
233
inet_ctl_sock_destroy(per_cpu_ptr(socks, i)->sock);
net/xfrm/xfrm_nat_keepalive.c
242
inet_ctl_sock_destroy(per_cpu_ptr(socks, i)->sock);
tools/include/uapi/linux/bpf.h
6056
FN(per_cpu_ptr, 153, ##ctx) \
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
312
struct rqsl_cpu_hist *hist = per_cpu_ptr(&rqsl_cpu_hists, cpu);
tools/testing/shared/linux/percpu.h
11
#define per_cpu(var, cpu) (*per_cpu_ptr(&(var), cpu))