raw_cpu_read
unsigned long irq_stk = (unsigned long)raw_cpu_read(irq_stack_ptr);
unsigned long ovf_stk = (unsigned long)raw_cpu_read(overflow_stack_ptr);
#define __this_cpu_read raw_cpu_read
unsigned long low = (unsigned long)raw_cpu_read(irq_stack_ptr);
unsigned long low = (unsigned long)raw_cpu_read(sdei_stack_normal_ptr);
unsigned long low = (unsigned long)raw_cpu_read(sdei_stack_critical_ptr);
pool_nr = raw_cpu_read(iommu_pool_hash) & (tbl->nr_pools - 1);
rng = raw_cpu_read(pnv_rng);
wrmsr_on_cpu(0, msr_no, raw_cpu_read(msrs->l), raw_cpu_read(msrs->h));
TYPEOF_UNQUAL(_var) pxo_old__ = raw_cpu_read(_var); \
!!(BIT_MASK(_nr) & raw_cpu_read(*addr__)); \
if (raw_cpu_read(ucode_ctrl.ctrl) != SCTRL_WAIT)
unsigned int ctrl_cpu = raw_cpu_read(ucode_ctrl.ctrl_cpu);
if (!raw_cpu_read(ucode_ctrl.nmi_enabled))
if (raw_cpu_read(ucode_ctrl.ctrl_cpu) == cpu) {
if (!raw_cpu_read(ucode_ctrl.nmi_enabled))
return raw_cpu_read(*base->percpu_base);
# define lockdep_hardirq_context() (raw_cpu_read(hardirq_context))
raw_cpu_read(pcp); \
u32 offset = raw_cpu_read(kstack_offset); \
u32 offset = raw_cpu_read(kstack_offset); \
return raw_cpu_read(_numa_mem_);
return raw_cpu_read(numa_node);
if (raw_cpu_read(lockdep_recursion))
rnp = raw_cpu_read(rcu_data.mynode);
if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs)))
if (!raw_cpu_read(rcu_data.rcu_urgent_qs))
if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) {
if (use_backlog_threads() && thread == raw_cpu_read(backlog_napi))