Symbol: raw_cpu_ptr
arch/arm/kernel/smp_twd.c
107
clockevents_update_freq(raw_cpu_ptr(twd_evt), twd_timer_rate);
arch/arm/kernel/smp_twd.c
133
if (twd_evt && raw_cpu_ptr(twd_evt) && !IS_ERR(twd_clk))
arch/arm/kernel/smp_twd.c
220
struct clock_event_device *clk = raw_cpu_ptr(twd_evt);
arch/arm/kernel/smp_twd.c
94
struct clock_event_device *clk = raw_cpu_ptr(twd_evt);
arch/arm64/include/asm/percpu.h
159
op(raw_cpu_ptr(&(pcp)), __VA_ARGS__); \
arch/arm64/include/asm/percpu.h
167
__retval = (typeof(pcp))op(raw_cpu_ptr(&(pcp)), ##args); \
arch/arm64/include/asm/percpu.h
254
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/arm64/include/asm/percpu.h
271
#define this_cpu_ptr raw_cpu_ptr
arch/arm64/include/asm/stacktrace.h
66
unsigned long low = (unsigned long)raw_cpu_ptr(overflow_stack);
arch/loongarch/include/asm/percpu.h
111
typeof(*raw_cpu_ptr(&(pcp))) __ret; \
arch/loongarch/include/asm/percpu.h
113
__ret = cmpxchg_local(raw_cpu_ptr(&(pcp)), o, n); \
arch/loongarch/include/asm/percpu.h
122
__retval = (typeof(pcp))operation(raw_cpu_ptr(&(pcp)), \
arch/s390/include/asm/percpu.h
110
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
132
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
153
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
164
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
25
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
64
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/s390/include/asm/percpu.h
91
ptr__ = raw_cpu_ptr(&(pcp)); \
arch/x86/events/core.c
2913
desc = raw_cpu_ptr(gdt_page.gdt) + idx;
arch/x86/include/asm/msr.h
295
rdmsr_on_cpu(0, msr_no, raw_cpu_ptr(&msrs->l), raw_cpu_ptr(&msrs->h));
arch/x86/kernel/cpu/mce/core.c
2465
__mcheck_cpu_init_vendor(raw_cpu_ptr(&cpu_info));
arch/x86/kernel/cpu/mce/core.c
2486
if (!mce_available(raw_cpu_ptr(&cpu_info)))
arch/x86/kernel/cpu/mce/core.c
2505
if (!mce_available(raw_cpu_ptr(&cpu_info)))
arch/x86/kernel/cpu/mce/core.c
2512
if (!mce_available(raw_cpu_ptr(&cpu_info)))
arch/x86/kernel/cpu/mce/core.c
2756
if (!mce_available(raw_cpu_ptr(&cpu_info)))
arch/x86/kernel/cpu/mce/core.c
2770
if (!mce_available(raw_cpu_ptr(&cpu_info)))
arch/x86/kernel/cpu/mce/intel.c
312
if (!mce_available(raw_cpu_ptr(&cpu_info)) || !cmci_supported(&banks))
arch/x86/lib/delay.c
134
__monitorx(raw_cpu_ptr(&cpu_tss_rw), 0, 0);
drivers/block/zram/zcomp.c
113
struct zcomp_strm *zstrm = raw_cpu_ptr(comp->stream);
drivers/char/random.c
362
crng = raw_cpu_ptr(&crngs);
drivers/char/random.c
524
batch = raw_cpu_ptr(&batched_entropy_##type); \
drivers/crypto/caam/caamalg_qi2.c
5521
ppriv = raw_cpu_ptr(priv->ppriv);
drivers/crypto/caam/qi.c
147
struct device *qidev = &(raw_cpu_ptr(&pcpu_qipriv)->net_dev->dev);
drivers/crypto/caam/qi.c
571
struct caam_napi *caam_napi = raw_cpu_ptr(&pcpu_qipriv.caam_napi);
drivers/crypto/caam/qi.c
574
struct device *qidev = &(raw_cpu_ptr(&pcpu_qipriv)->net_dev->dev);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
832
if (!list_empty(raw_cpu_ptr(clt_path->mp_skip_entry)))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
848
list_add(raw_cpu_ptr(min_path->mp_skip_entry), &it->skip_list);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
883
if (!list_empty(raw_cpu_ptr(clt_path->mp_skip_entry)))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
899
list_add(raw_cpu_ptr(min_path->mp_skip_entry), &it->skip_list);
drivers/iommu/dma-iommu.c
217
fq = raw_cpu_ptr(cookie->percpu_fq);
drivers/iommu/iova.c
776
cpu_rcache = raw_cpu_ptr(rcache->cpu_rcaches);
drivers/iommu/iova.c
830
cpu_rcache = raw_cpu_ptr(rcache->cpu_rcaches);
drivers/irqchip/irq-gic-v3-its.c
206
#define gic_data_rdist() (raw_cpu_ptr(gic_rdists->rdist))
drivers/irqchip/irq-gic.c
651
ptr = raw_cpu_ptr(gic->saved_ppi_enable);
drivers/irqchip/irq-gic.c
655
ptr = raw_cpu_ptr(gic->saved_ppi_active);
drivers/irqchip/irq-gic.c
659
ptr = raw_cpu_ptr(gic->saved_ppi_conf);
drivers/irqchip/irq-gic.c
681
ptr = raw_cpu_ptr(gic->saved_ppi_enable);
drivers/irqchip/irq-gic.c
688
ptr = raw_cpu_ptr(gic->saved_ppi_active);
drivers/irqchip/irq-gic.c
695
ptr = raw_cpu_ptr(gic->saved_ppi_conf);
drivers/md/dm-stats.c
693
last = raw_cpu_ptr(stats->last);
fs/ext4/mballoc.c
5879
ac->ac_lg = raw_cpu_ptr(sbi->s_locality_groups);
include/asm-generic/mmiowb.h
30
#define __mmiowb_state() raw_cpu_ptr(&__mmiowb_state)
include/asm-generic/percpu.h
100
TYPEOF_UNQUAL(pcp) *__p = raw_cpu_ptr(&(pcp)); \
include/asm-generic/percpu.h
118
TYPEOF_UNQUAL(pcp) *__p = raw_cpu_ptr(&(pcp)); \
include/asm-generic/percpu.h
142
___ret = READ_ONCE(*raw_cpu_ptr(&(pcp))); \
include/asm-generic/percpu.h
82
*raw_cpu_ptr(&(pcp)); \
include/asm-generic/percpu.h
87
*raw_cpu_ptr(&(pcp)) op val; \
include/asm-generic/percpu.h
92
TYPEOF_UNQUAL(pcp) *__p = raw_cpu_ptr(&(pcp)); \
include/linux/percpu-defs.h
256
#define this_cpu_ptr(ptr) raw_cpu_ptr(ptr)
include/linux/percpu-defs.h
269
#define this_cpu_ptr(ptr) raw_cpu_ptr(ptr)
include/linux/sbitmap.h
333
*raw_cpu_ptr(sb->alloc_hint) = bitnr;
include/linux/srcutree.h
297
atomic_long_inc(raw_cpu_ptr(&scp->srcu_locks)); // Y, and implicit RCU reader.
include/linux/srcutree.h
321
atomic_long_inc(raw_cpu_ptr(&scp->srcu_unlocks)); // Z, and implicit RCU reader.
include/linux/srcutree.h
341
atomic_long_inc(raw_cpu_ptr(&scp->srcu_locks)); // Y, and implicit RCU reader.
include/linux/srcutree.h
365
atomic_long_inc(raw_cpu_ptr(&scp->srcu_unlocks)); // Z, and implicit RCU reader.
include/net/snmp.h
157
TYPEOF_UNQUAL(*mib) *ptr = raw_cpu_ptr(mib); \
include/net/snmp.h
174
TYPEOF_UNQUAL(*mib) *ptr = raw_cpu_ptr(mib); \
kernel/irq/chip.c
920
res = action->handler(irq, raw_cpu_ptr(action->percpu_dev_id));
kernel/kcsan/core.c
206
return in_task() ? &current->kcsan_ctx : raw_cpu_ptr(&kcsan_cpu_ctx);
kernel/printk/nbcon.c
1427
return raw_cpu_ptr(&nbcon_pcpu_emergency_nesting);
kernel/rcu/srcutree.c
1263
sdp = raw_cpu_ptr(ssp->sda);
kernel/rcu/srcutree.c
1328
sdp = raw_cpu_ptr(ssp->sda);
kernel/rcu/srcutree.c
769
sdp = raw_cpu_ptr(ssp->sda);
kernel/rcu/srcutree.c
822
struct srcu_ctr *scp = raw_cpu_ptr(scpp);
kernel/rcu/srcutree.c
838
atomic_long_inc(&raw_cpu_ptr(__srcu_ctr_to_ptr(ssp, idx))->srcu_unlocks);
kernel/rcu/tree.c
2837
struct rcu_data *rdp = raw_cpu_ptr(&rcu_data);
kernel/sched/sched.h
1394
#define raw_rq() raw_cpu_ptr(&runqueues)
kernel/taskstats.c
623
listeners = raw_cpu_ptr(&listener_array);
kernel/time/hrtimer.c
1634
cpu_base = raw_cpu_ptr(&hrtimer_bases);
kernel/time/hrtimer.c
950
struct hrtimer_cpu_base *cpu_base = raw_cpu_ptr(&hrtimer_bases);
mm/kmsan/kmsan.h
83
return in_task() ? &current->kmsan_ctx : raw_cpu_ptr(&kmsan_percpu_ctx);
mm/vmalloc.c
2871
vbq = raw_cpu_ptr(&vmap_block_queue);
mm/vmalloc.c
3410
struct vfree_deferred *p = raw_cpu_ptr(&vfree_deferred);
mm/zswap.c
833
acomp_ctx = raw_cpu_ptr(pool->acomp_ctx);
net/ipv4/route.c
1498
p = (struct rtable **)raw_cpu_ptr(nhc->nhc_pcpu_rth_output);
net/ipv4/route.c
1529
struct uncached_list *ul = raw_cpu_ptr(&rt_uncached_list);
net/ipv4/route.c
2647
prth = raw_cpu_ptr(nhc->nhc_pcpu_rth_output);
net/ipv6/route.c
140
struct uncached_list *ul = raw_cpu_ptr(&rt6_uncached_list);
net/mpls/internal.h
37
raw_cpu_ptr((mdev)->stats); \
net/mpls/internal.h
49
raw_cpu_ptr((mdev)->stats); \
net/netfilter/nft_counter.c
264
this_cpu = raw_cpu_ptr(cpu_stats);
net/netfilter/nft_counter.c
76
this_cpu = raw_cpu_ptr(cpu_stats);
net/netfilter/nft_set_pipapo.c
429
scratch = *raw_cpu_ptr(m->scratch);
net/netfilter/nft_set_pipapo_avx2.c
1162
scratch = *raw_cpu_ptr(m->scratch);
net/xfrm/xfrm_state.c
1210
state_cache_input = raw_cpu_ptr(net->xfrm.state_cache_input);