arch/arm64/kernel/topology.c
148
scale *= this_cpu_read(arch_max_freq_scale);
arch/arm64/kernel/traps.c
906
unsigned long irq_stk = (unsigned long)this_cpu_read(irq_stack_ptr);
arch/loongarch/kernel/kfpu.c
45
WARN_ON(this_cpu_read(in_kernel_fpu));
arch/loongarch/kernel/kfpu.c
73
WARN_ON(!this_cpu_read(in_kernel_fpu));
arch/loongarch/kernel/process.c
261
unsigned long begin = (unsigned long)this_cpu_read(irq_stack);
arch/powerpc/lib/code-patching.c
135
free_vm_area(this_cpu_read(cpu_patching_context.area));
arch/powerpc/lib/code-patching.c
195
put_patching_mm(this_cpu_read(cpu_patching_context.mm),
arch/powerpc/lib/code-patching.c
196
this_cpu_read(cpu_patching_context.addr));
arch/powerpc/lib/qspinlock.c
64
u64 seen = this_cpu_read(sleepy_lock_seen_clock);
arch/powerpc/mm/nohash/e500_hugetlbpage.c
84
index = this_cpu_read(next_tlbcam_idx);
arch/x86/coco/sev/core.c
1083
data = this_cpu_read(runtime_data);
arch/x86/coco/sev/core.c
489
call.caa = this_cpu_read(svsm_caa);
arch/x86/coco/sev/internal.h
105
return this_cpu_read(svsm_caa_pa);
arch/x86/coco/sev/internal.h
97
return this_cpu_read(svsm_caa);
arch/x86/coco/sev/noinstr.c
127
data = this_cpu_read(runtime_data);
arch/x86/coco/sev/noinstr.c
173
data = this_cpu_read(runtime_data);
arch/x86/coco/sev/vc-handle.c
365
regs->ax = lower_32_bits(this_cpu_read(svsm_caa_pa));
arch/x86/coco/sev/vc-handle.c
366
regs->dx = upper_32_bits(this_cpu_read(svsm_caa_pa));
arch/x86/coco/sev/vc-handle.c
663
struct sev_es_runtime_data *data = this_cpu_read(runtime_data);
arch/x86/coco/sev/vc-handle.c
704
struct sev_es_runtime_data *data = this_cpu_read(runtime_data);
arch/x86/events/amd/core.c
884
if (time_after(jiffies, this_cpu_read(perf_nmi_tstamp)))
arch/x86/events/core.c
1803
if (this_cpu_read(guest_lvtpc_loaded))
arch/x86/events/intel/ds.c
2046
u8 *buf = this_cpu_read(insn_buffer);
arch/x86/events/intel/ds.c
2369
WARN_ON(this_cpu_read(cpu_hw_events.enabled));
arch/x86/events/intel/ds.c
2799
WARN_ON(this_cpu_read(cpu_hw_events.enabled));
arch/x86/events/intel/p4.c
1014
s64 left = this_cpu_read(pmc_prev_left[hwc->idx]);
arch/x86/include/asm/desc.h
330
if (unlikely(this_cpu_read(__tss_limit_invalid)))
arch/x86/include/asm/entry-common.h
99
this_cpu_read(x86_ibpb_exit_to_user)) {
arch/x86/include/asm/kvmclock.h
11
return &this_cpu_read(hv_clock_per_cpu)->pvti;
arch/x86/include/asm/kvmclock.h
16
return this_cpu_read(hv_clock_per_cpu);
arch/x86/include/asm/percpu.h
278
TYPEOF_UNQUAL(_var) pxo_old__ = this_cpu_read(_var); \
arch/x86/include/asm/percpu.h
56
#define __my_cpu_offset this_cpu_read(this_cpu_off)
arch/x86/include/asm/smp.h
139
#define raw_smp_processor_id() this_cpu_read(cpu_number)
arch/x86/include/asm/switch_to.h
60
if (unlikely(this_cpu_read(cpu_tss_rw.x86_tss.ss1) == thread->sysenter_cs))
arch/x86/include/asm/tlbflush.h
461
u64 lam = this_cpu_read(cpu_tlbstate.lam);
arch/x86/include/asm/uv/uv_hub.h
769
#define uv_hub_nmi this_cpu_read(uv_cpu_nmi.hub)
arch/x86/include/asm/xen/hypervisor.h
78
BUG_ON(this_cpu_read(xen_lazy_mode) != XEN_LAZY_NONE);
arch/x86/include/asm/xen/hypervisor.h
85
BUG_ON(this_cpu_read(xen_lazy_mode) != mode);
arch/x86/kernel/apic/msi.c
111
if (IS_ERR_OR_NULL(this_cpu_read(vector_irq[cfg->vector])))
arch/x86/kernel/apic/x2apic_cluster.c
93
struct cpumask *cmsk = this_cpu_read(cluster_masks);
arch/x86/kernel/cpu/bugs.c
114
return this_cpu_read(x86_spec_ctrl_current);
arch/x86/kernel/cpu/bugs.c
99
if (this_cpu_read(x86_spec_ctrl_current) == val)
arch/x86/kernel/cpu/common.c
2469
if (this_cpu_read(numa_node) == 0 &&
arch/x86/kernel/cpu/common.c
496
unsigned long newval, cr4 = this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/cpu/common.c
511
return this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/cpu/common.c
537
cr4_pinned_bits = this_cpu_read(cpu_tlbstate.cr4) & cr4_pinned_mask;
arch/x86/kernel/cpu/intel.c
511
msr = this_cpu_read(msr_misc_features_shadow);
arch/x86/kernel/cpu/intel_epb.c
94
u64 val = this_cpu_read(saved_epb);
arch/x86/kernel/cpu/mce/amd.c
1079
if ((bank >= this_cpu_read(mce_num_banks)) || (block >= NR_BLOCKS))
arch/x86/kernel/cpu/mce/amd.c
1147
struct device *dev = this_cpu_read(mce_device);
arch/x86/kernel/cpu/mce/amd.c
1205
unsigned int bank, numbanks = this_cpu_read(mce_num_banks);
arch/x86/kernel/cpu/mce/amd.c
1219
struct threshold_bank **bp = this_cpu_read(threshold_banks);
arch/x86/kernel/cpu/mce/amd.c
1253
bp = this_cpu_read(threshold_banks);
arch/x86/kernel/cpu/mce/amd.c
1257
numbanks = this_cpu_read(mce_num_banks);
arch/x86/kernel/cpu/mce/amd.c
1263
if (!(this_cpu_read(bank_map) & BIT_ULL(bank)))
arch/x86/kernel/cpu/mce/amd.c
436
if (!this_cpu_read(threshold_banks) && !tr->set_lvt_off)
arch/x86/kernel/cpu/mce/amd.c
481
struct threshold_bank **thr_banks = this_cpu_read(threshold_banks);
arch/x86/kernel/cpu/mce/amd.c
672
if (c->x86 == 15 && this_cpu_read(mce_num_banks) > 4) {
arch/x86/kernel/cpu/mce/amd.c
685
if (c->x86 == 6 && this_cpu_read(mce_num_banks))
arch/x86/kernel/cpu/mce/amd.c
728
for (bank = 0; bank < this_cpu_read(mce_num_banks); ++bank) {
arch/x86/kernel/cpu/mce/core.c
1315
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
1366
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
1849
u8 n_banks = this_cpu_read(mce_num_banks);
arch/x86/kernel/cpu/mce/core.c
1916
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
2045
if (this_cpu_read(mce_num_banks) > 8)
arch/x86/kernel/cpu/mce/core.c
2330
if (bank >= this_cpu_read(mce_num_banks)) {
arch/x86/kernel/cpu/mce/core.c
2419
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
2775
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
831
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/core.c
997
for (i = 0; i < this_cpu_read(mce_num_banks); i++) {
arch/x86/kernel/cpu/mce/intel.c
484
if (c->x86_vfm < INTEL_NEHALEM_EP && this_cpu_read(mce_num_banks))
arch/x86/kernel/cpu/microcode/core.c
429
if (this_cpu_read(ucode_ctrl.ctrl) == SCTRL_APPLY)
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
167
closid_p = this_cpu_read(pqr_state.cur_closid);
arch/x86/kernel/cpu/resctrl/pseudo_lock.c
168
rmid_p = this_cpu_read(pqr_state.cur_rmid);
arch/x86/kernel/doublefault_32.c
16
#define TSS(x) this_cpu_read(cpu_tss_rw.x86_tss.x)
arch/x86/kernel/dumpstack_32.c
102
info->next_sp = (unsigned long *)this_cpu_read(cpu_tss_rw.x86_tss.sp);
arch/x86/kernel/dumpstack_32.c
40
unsigned long *begin = (unsigned long *)this_cpu_read(hardirq_stack_ptr);
arch/x86/kernel/dumpstack_32.c
65
unsigned long *begin = (unsigned long *)this_cpu_read(softirq_stack_ptr);
arch/x86/kernel/dumpstack_64.c
137
unsigned long *end = (unsigned long *)this_cpu_read(hardirq_stack_ptr);
arch/x86/kernel/fpu/context.h
38
return fpu == this_cpu_read(fpu_fpregs_owner_ctx) && cpu == fpu->last_cpu;
arch/x86/kernel/fpu/core.c
480
WARN_ON_FPU(!this_cpu_read(kernel_fpu_allowed));
arch/x86/kernel/fpu/core.c
502
WARN_ON_FPU(this_cpu_read(kernel_fpu_allowed));
arch/x86/kernel/fpu/core.c
90
if (!this_cpu_read(kernel_fpu_allowed))
arch/x86/kernel/head64.c
274
__native_tlb_flush_global(this_cpu_read(cpu_tlbstate.cr4));
arch/x86/kernel/hw_breakpoint.c
165
dr7 = this_cpu_read(cpu_dr7);
arch/x86/kernel/hw_breakpoint.c
532
bp = this_cpu_read(bp_per_reg[i]);
arch/x86/kernel/irq.c
430
apic_id = this_cpu_read(x86_cpu_to_apicid);
arch/x86/kernel/ldt.c
140
if (this_cpu_read(cpu_tlbstate.loaded_mm) != mm)
arch/x86/kernel/machine_kexec_64.c
449
if (this_cpu_read(cache_state_incoherent))
arch/x86/kernel/nmi.c
555
if (this_cpu_read(nmi_state) != NMI_NOT_RUNNING) {
arch/x86/kernel/nmi.c
597
local_db_restore(this_cpu_read(nmi_dr7));
arch/x86/kernel/nmi.c
601
if (unlikely(this_cpu_read(nmi_cr2) != read_cr2()))
arch/x86/kernel/nmi.c
602
write_cr2(this_cpu_read(nmi_cr2));
arch/x86/kernel/nmi.c
733
if (unlikely(this_cpu_read(nmi_cr2) != read_cr2()))
arch/x86/kernel/nmi.c
734
write_cr2(this_cpu_read(nmi_cr2));
arch/x86/kernel/process.c
351
msrval = this_cpu_read(msr_misc_features_shadow);
arch/x86/kernel/process.c
708
unsigned long newval, cr4 = this_cpu_read(cpu_tlbstate.cr4);
arch/x86/kernel/process.c
840
if (this_cpu_read(cache_state_incoherent))
arch/x86/kernel/process_64.c
617
this_cpu_read(hardirq_stack_inuse));
arch/x86/kernel/process_64.c
807
if (this_cpu_read(cpu_tlbstate.loaded_mm) == mm) {
arch/x86/kernel/traps.c
627
struct pt_regs *gpregs = (struct pt_regs *)this_cpu_read(cpu_tss_rw.x86_tss.sp0) - 1;
arch/x86/kernel/tsc.c
82
seq = this_cpu_read(cyc2ns.seq.seqcount.sequence);
arch/x86/kernel/tsc.c
85
data->cyc2ns_offset = this_cpu_read(cyc2ns.data[idx].cyc2ns_offset);
arch/x86/kernel/tsc.c
86
data->cyc2ns_mul = this_cpu_read(cyc2ns.data[idx].cyc2ns_mul);
arch/x86/kernel/tsc.c
87
data->cyc2ns_shift = this_cpu_read(cyc2ns.data[idx].cyc2ns_shift);
arch/x86/kernel/tsc.c
89
} while (unlikely(seq != this_cpu_read(cyc2ns.seq.seqcount.sequence)));
arch/x86/kvm/vmx/vmx.c
7502
u64 hostval = this_cpu_read(x86_spec_ctrl_current);
arch/x86/kvm/vmx/vmx_onhyperv.h
15
#define current_evmcs ((struct hv_enlightened_vmcs *)this_cpu_read(current_vmcs))
arch/x86/lib/delay.c
209
unsigned long lpj = this_cpu_read(cpu_info.loops_per_jiffy) ? : loops_per_jiffy;
arch/x86/mm/tlb.c
1025
if (this_cpu_read(cpu_tlbstate_shared.is_lazy))
arch/x86/mm/tlb.c
1028
prev_mm = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
1054
cpumask_clear_cpu(smp_processor_id(), mm_cpumask(this_cpu_read(cpu_tlbstate.loaded_mm)));
arch/x86/mm/tlb.c
1082
struct mm_struct *mm = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
1136
struct mm_struct *loaded_mm = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
1137
u32 loaded_mm_asid = this_cpu_read(cpu_tlbstate.loaded_mm_asid);
arch/x86/mm/tlb.c
1164
loaded_mm_asid = this_cpu_read(cpu_tlbstate.loaded_mm_asid);
arch/x86/mm/tlb.c
1171
VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].ctx_id) !=
arch/x86/mm/tlb.c
1174
if (this_cpu_read(cpu_tlbstate_shared.is_lazy)) {
arch/x86/mm/tlb.c
1188
local_tlb_gen = this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].tlb_gen);
arch/x86/mm/tlb.c
1474
} else if (mm == this_cpu_read(cpu_tlbstate.loaded_mm)) {
arch/x86/mm/tlb.c
1576
build_cr3(this_cpu_read(cpu_tlbstate.loaded_mm)->pgd,
arch/x86/mm/tlb.c
1577
this_cpu_read(cpu_tlbstate.loaded_mm_asid),
arch/x86/mm/tlb.c
1635
loaded_mm_asid = this_cpu_read(cpu_tlbstate.loaded_mm_asid);
arch/x86/mm/tlb.c
1636
cpu_pcide = this_cpu_read(cpu_tlbstate.cr4) & X86_CR4_PCIDE;
arch/x86/mm/tlb.c
1679
__native_tlb_flush_global(this_cpu_read(cpu_tlbstate.cr4));
arch/x86/mm/tlb.c
1696
invalidate_user_asid(this_cpu_read(cpu_tlbstate.loaded_mm_asid));
arch/x86/mm/tlb.c
1769
struct mm_struct *loaded_mm = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
207
if (asid == this_cpu_read(cpu_tlbstate.loaded_mm_asid))
arch/x86/mm/tlb.c
250
if (this_cpu_read(cpu_tlbstate.invalidate_other))
arch/x86/mm/tlb.c
254
if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) !=
arch/x86/mm/tlb.c
259
ns.need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < next_tlb_gen);
arch/x86/mm/tlb.c
587
struct mm_struct *loaded_mm = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
601
WARN_ON(!this_cpu_read(cpu_tlbstate_shared.is_lazy));
arch/x86/mm/tlb.c
645
if (this_cpu_read(cpu_info.smt_active)) {
arch/x86/mm/tlb.c
674
prev_mm = this_cpu_read(cpu_tlbstate.last_user_mm_spec);
arch/x86/mm/tlb.c
770
cr4_update_pce_mm(this_cpu_read(cpu_tlbstate.loaded_mm));
arch/x86/mm/tlb.c
786
struct mm_struct *prev = this_cpu_read(cpu_tlbstate.loaded_mm);
arch/x86/mm/tlb.c
787
u16 prev_asid = this_cpu_read(cpu_tlbstate.loaded_mm_asid);
arch/x86/mm/tlb.c
788
bool was_lazy = this_cpu_read(cpu_tlbstate_shared.is_lazy);
arch/x86/mm/tlb.c
844
this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) !=
arch/x86/mm/tlb.c
892
if (this_cpu_read(cpu_tlbstate.ctxs[prev_asid].tlb_gen) ==
arch/x86/mm/tlb.c
989
if (this_cpu_read(cpu_tlbstate.loaded_mm) == &init_mm)
arch/x86/platform/uv/uv_nmi.c
1018
if (!this_cpu_read(uv_cpu_nmi.pinging)) {
arch/x86/platform/uv/uv_nmi.c
825
while (this_cpu_read(uv_cpu_nmi.state) != UV_NMI_STATE_DUMP)
arch/x86/platform/uv/uv_nmi.c
947
if (!this_cpu_read(uv_cpu_nmi.pinging) && !uv_check_nmi(hub_nmi)) {
arch/x86/virt/vmx/tdx/tdx.c
1877
if (!this_cpu_read(cache_state_incoherent))
arch/x86/xen/enlighten_pv.c
1056
unsigned long cr0 = this_cpu_read(xen_cr0_value);
arch/x86/xen/enlighten_pv.c
148
return this_cpu_read(xen_lazy_mode);
arch/x86/xen/enlighten_pv.c
427
if (this_cpu_read(xen_lazy_mode) == XEN_LAZY_MMU) {
arch/x86/xen/mmu_pv.c
1288
this_cpu_read(xen_vcpu)->arch.cr2 = cr2;
arch/x86/xen/mmu_pv.c
1366
return this_cpu_read(xen_cr3);
arch/x86/xen/mmu_pv.c
970
if (this_cpu_read(cpu_tlbstate.loaded_mm) == mm)
arch/x86/xen/mmu_pv.c
977
if (this_cpu_read(xen_current_cr3) == __pa(mm->pgd))
arch/x86/xen/suspend.c
43
wrmsrq(MSR_IA32_SPEC_CTRL, this_cpu_read(spec_ctrl));
arch/x86/xen/xen-ops.h
245
local_irq_restore(this_cpu_read(xen_mc_irq_flags));
drivers/accessibility/speakup/fakekey.c
86
return this_cpu_read(reporting_keystroke);
drivers/acpi/processor_idle.c
642
cx = this_cpu_read(acpi_cstate[index]);
drivers/acpi/processor_throttling.c
705
if ((this_cpu_read(cpu_info.x86_vendor) != X86_VENDOR_INTEL) ||
drivers/acpi/processor_throttling.c
725
if ((this_cpu_read(cpu_info.x86_vendor) != X86_VENDOR_INTEL) ||
drivers/cpufreq/powernv-cpufreq.c
563
chip = this_cpu_read(chip_info);
drivers/crypto/nx/nx-common-powernv.c
587
txwin = this_cpu_read(cpu_txwin);
drivers/dma/dmaengine.c
549
return this_cpu_read(channel_table[tx_type]->chan);
drivers/firmware/psci/psci_checker.c
287
dev = this_cpu_read(cpuidle_devices);
drivers/input/gameport/gameport.c
157
return (this_cpu_read(cpu_info.loops_per_jiffy) *
drivers/irqchip/irq-aclint-sswi.c
29
writel_relaxed(0x0, this_cpu_read(sswi_cpu_regs));
drivers/irqchip/irq-csky-mpintc.c
215
void __iomem *reg_base = this_cpu_read(intcl_reg);
drivers/irqchip/irq-csky-mpintc.c
56
(this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR)))
drivers/irqchip/irq-csky-mpintc.c
75
void __iomem *reg_base = this_cpu_read(intcl_reg);
drivers/irqchip/irq-csky-mpintc.c
83
void __iomem *reg_base = this_cpu_read(intcl_reg);
drivers/irqchip/irq-csky-mpintc.c
92
void __iomem *reg_base = this_cpu_read(intcl_reg);
drivers/irqchip/irq-csky-mpintc.c
99
void __iomem *reg_base = this_cpu_read(intcl_reg);
drivers/irqchip/irq-gic.c
223
hwirq = this_cpu_read(sgi_intid);
drivers/irqchip/irq-gic.c
237
hwirq = this_cpu_read(sgi_intid);
drivers/irqchip/irq-loongarch-avec.c
245
d = this_cpu_read(irq_map[vector]);
drivers/nvdimm/region_devs.c
1137
idx = this_cpu_read(flush_idx);
fs/ext4/mballoc.c
6306
seq = this_cpu_read(discard_pa_seq);
include/crypto/internal/simd.h
54
(may_use_simd() && !this_cpu_read(crypto_simd_disabled_for_test))
include/linux/cpumask.h
1025
#define this_cpu_cpumask_var_ptr(x) this_cpu_read(x)
include/linux/interrupt.h
665
return this_cpu_read(ksoftirqd);
include/linux/irqflags.h
54
# define lockdep_hardirqs_enabled() (this_cpu_read(hardirqs_enabled))
include/linux/lockdep.h
573
#define __lockdep_enabled (debug_locks && !this_cpu_read(lockdep_recursion))
include/linux/lockdep.h
577
WARN_ON_ONCE(__lockdep_enabled && !this_cpu_read(hardirqs_enabled)); \
include/linux/lockdep.h
582
WARN_ON_ONCE(__lockdep_enabled && this_cpu_read(hardirqs_enabled)); \
include/linux/lockdep.h
587
WARN_ON_ONCE(__lockdep_enabled && !this_cpu_read(hardirq_context)); \
include/linux/lockdep.h
592
WARN_ON_ONCE(__lockdep_enabled && (this_cpu_read(hardirq_context) || \
include/linux/lockdep.h
593
!this_cpu_read(hardirqs_enabled))); \
include/linux/lockdep.h
601
!this_cpu_read(hardirqs_enabled))); \
include/linux/lockdep.h
609
this_cpu_read(hardirqs_enabled))); \
include/linux/netdevice.h
3585
return this_cpu_read(softnet_data.xmit.recursion);
include/linux/perf_event.h
751
(this_cpu_read(hardirqs_enabled) && \
include/linux/sched/mm.h
497
old = this_cpu_read(int_active_memcg);
kernel/bpf/helpers.c
1454
if (this_cpu_read(async_cb_running) == cb) {
kernel/bpf/helpers.c
1563
cur_t = this_cpu_read(hrtimer_running);
kernel/bpf/helpers.c
791
if (WARN_ON_ONCE(this_cpu_read(bpf_bprintf_nest_level) == 0))
kernel/cpu.c
1225
kthread_unpark(this_cpu_read(cpuhp_state.thread));
kernel/irq/irqdesc.c
1036
this_cpu_write(desc->kstat_irqs->ref, this_cpu_read(desc->kstat_irqs->cnt));
kernel/irq/irqdesc.c
1046
return this_cpu_read(desc->kstat_irqs->cnt) - this_cpu_read(desc->kstat_irqs->ref);
kernel/kcsan/core.c
298
u32 state = this_cpu_read(kcsan_rand_state);
kernel/locking/lockdep.c
121
if (this_cpu_read(lockdep_recursion))
kernel/locking/lockdep.c
4373
if (unlikely(this_cpu_read(lockdep_recursion)))
kernel/locking/lockdep.c
4442
if (unlikely(this_cpu_read(lockdep_recursion)))
kernel/printk/printk_safe.c
70
this_cpu_read(printk_context) ||
kernel/rcu/srcutree.c
1260
if (this_cpu_read(ssp->sda->srcu_reader_flavor) & SRCU_READ_FLAVOR_SLOWGP)
kernel/softirq.c
176
DEBUG_LOCKS_WARN_ON(this_cpu_read(softirq_ctrl.cnt));
kernel/softirq.c
224
this_cpu_read(softirq_ctrl.cnt));
kernel/softirq.c
272
curcnt = this_cpu_read(softirq_ctrl.cnt);
kernel/softirq.c
335
return !this_cpu_read(softirq_ctrl.cnt);
kernel/trace/fgraph.c
1204
pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid);
kernel/trace/ftrace.c
179
pid = this_cpu_read(tr->array_buffer.data->ftrace_ignore_pid);
kernel/trace/rv/monitors/opid/opid.c
88
if (this_cpu_read(hardirq_context))
kernel/trace/rv/monitors/opid/opid.c
97
if (this_cpu_read(hardirq_context))
kernel/trace/rv/monitors/sleep/sleep.c
100
if (this_cpu_read(hardirq_context)) {
kernel/trace/trace.c
7139
if (rbe == this_cpu_read(trace_buffered_event))
kernel/trace/trace.h
1622
if (this_cpu_read(trace_buffered_event) == event) {
kernel/trace/trace.h
1637
if (this_cpu_read(trace_buffered_event) == event) {
kernel/trace/trace_events.c
1132
if (!this_cpu_read(tr->array_buffer.data->ignore_pid))
kernel/trace/trace_events.c
1150
if (this_cpu_read(tr->array_buffer.data->ignore_pid))
kernel/trace/trace_events.c
648
return this_cpu_read(tr->array_buffer.data->ignore_pid) != 0;
kernel/trace/trace_events_hist.c
5358
if (this_cpu_read(hist_pad_cnt) == MAX_HIST_CNT) {
kernel/trace/trace_preemptirq.c
106
if (!this_cpu_read(tracing_irq_cpu)) {
kernel/trace/trace_preemptirq.c
61
if (this_cpu_read(tracing_irq_cpu)) {
kernel/trace/trace_preemptirq.c
72
if (this_cpu_read(tracing_irq_cpu)) {
kernel/trace/trace_preemptirq.c
92
if (!this_cpu_read(tracing_irq_cpu)) {
kernel/watchdog_perf.c
163
WARN_ONCE(this_cpu_read(watchdog_ev), "unexpected watchdog_ev leak");
kernel/watchdog_perf.c
176
struct perf_event *event = this_cpu_read(watchdog_ev);
kernel/watchdog_perf.c
195
struct perf_event *event = this_cpu_read(watchdog_ev);
lib/percpu_counter.c
98
count = this_cpu_read(*fbc->counters);
lib/sbitmap.c
34
hint = this_cpu_read(*sb->alloc_hint);
mm/kasan/sw_tags.c
67
u32 state = this_cpu_read(prng_state);
mm/kmsan/kmsan_test.c
453
check = this_cpu_read(per_cpu_var);
mm/memcontrol.c
2756
memcg = this_cpu_read(int_active_memcg);
mm/memcontrol.c
910
return this_cpu_read(int_active_memcg);
mm/show_mem.c
346
K(this_cpu_read(zone->per_cpu_pageset->count)),
mm/swapfile.c
1329
si = this_cpu_read(percpu_swap_cluster.si[order]);
mm/swapfile.c
1330
offset = this_cpu_read(percpu_swap_cluster.offset[order]);
mm/vmalloc.c
1907
if (!this_cpu_read(ne_fit_preload_node))
mm/vmstat.c
568
o = this_cpu_read(*p);
mm/vmstat.c
582
t = this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
637
o = this_cpu_read(*p);
mm/vmstat.c
651
t = this_cpu_read(pcp->stat_threshold);
net/core/dev.c
13205
napi->thread = this_cpu_read(backlog_napi);
net/core/dev.c
5529
err = skb_cow_data_for_xdp(this_cpu_read(system_page_pool.pool), pskb, prog);
net/core/xdp.c
753
pp = this_cpu_read(system_page_pool.pool);
net/ipv4/icmp.c
209
sk = this_cpu_read(ipv4_icmp_sk);
net/ipv4/tcp_ipv4.c
1016
ctl_sk = this_cpu_read(ipv4_tcp_sk.sock);
net/ipv4/tcp_ipv4.c
893
ctl_sk = this_cpu_read(ipv4_tcp_sk.sock);
net/ipv6/icmp.c
107
sk = this_cpu_read(ipv6_icmp_sk);
net/ipv6/route.c
1444
pcpu_rt = this_cpu_read(*res->nh->rt6i_pcpu);
net/netfilter/nft_ct.c
254
ct = this_cpu_read(nft_ct_pcpu_template);
net/xfrm/xfrm_nat_keepalive.c
64
sk = this_cpu_read(nat_keepalive_sk_ipv4.sock);
net/xfrm/xfrm_nat_keepalive.c
99
sk = this_cpu_read(nat_keepalive_sk_ipv6.sock);