Symbol: __this_cpu_read
arch/alpha/kernel/time.c
61
#define test_irq_work_pending() __this_cpu_read(irq_work_pending)
arch/arm/kernel/irq.c
83
__this_cpu_read(irq_stack_ptr));
arch/arm64/include/asm/arch_timer.h
27
__wa = __this_cpu_read(timer_unstable_counter_workaround); \
arch/arm64/include/asm/arch_timer.h
34
__wa = __this_cpu_read(timer_unstable_counter_workaround); \
arch/arm64/kernel/cpufeature.c
1948
if (__this_cpu_read(this_cpu_vector) == vectors) {
arch/arm64/kernel/entry-common.c
196
if (!__this_cpu_read(__in_cortex_a76_erratum_1463225_wa))
arch/arm64/kernel/fpsimd.c
1534
wrong_task = __this_cpu_read(fpsimd_last_state.st) !=
arch/arm64/kernel/mte.c
215
pref = __this_cpu_read(mte_tcf_preferred);
arch/arm64/kernel/proton-pack.c
1083
if (__this_cpu_read(bp_hardening_data.fn) != cpu_cb)
arch/arm64/kvm/arm.c
2215
if (!__this_cpu_read(kvm_hyp_initialized)) {
arch/arm64/kvm/arm.c
2223
if (!is_protected_kvm_enabled() && __this_cpu_read(kvm_hyp_initialized)) {
arch/arm64/kvm/arm.c
2270
if (__this_cpu_read(kvm_hyp_initialized))
arch/arm64/kvm/arm.c
2281
if (__this_cpu_read(kvm_hyp_initialized))
arch/arm64/kvm/hyp/nvhe/pkvm.c
253
if (__this_cpu_read(loaded_hyp_vcpu))
arch/arm64/kvm/hyp/nvhe/pkvm.c
294
return __this_cpu_read(loaded_hyp_vcpu);
arch/arm64/kvm/hyp/nvhe/switch.c
60
write_sysreg(__this_cpu_read(kvm_hyp_vector), vbar_el2);
arch/arm64/kvm/hyp/vhe/switch.c
141
write_sysreg(__this_cpu_read(kvm_hyp_vector), vbar_el1);
arch/arm64/kvm/hyp/vhe/switch.c
187
host_vectors = __this_cpu_read(this_cpu_vector);
arch/mips/cavium-octeon/octeon-irq.c
1304
ciu_sum &= __this_cpu_read(octeon_irq_ciu0_en_mirror);
arch/mips/cavium-octeon/octeon-irq.c
1321
ciu_sum &= __this_cpu_read(octeon_irq_ciu1_en_mirror);
arch/mips/cavium-octeon/octeon-irq.c
2602
ciu3_info = __this_cpu_read(octeon_ciu3_info);
arch/mips/cavium-octeon/octeon-irq.c
2669
ciu3_info = __this_cpu_read(octeon_ciu3_info);
arch/mips/cavium-octeon/octeon-irq.c
2780
ciu3_info = __this_cpu_read(octeon_ciu3_info);
arch/mips/kernel/mips-r2-to-r6-emul.c
2243
(unsigned long)__this_cpu_read(mipsr2emustats.movs),
arch/mips/kernel/mips-r2-to-r6-emul.c
2244
(unsigned long)__this_cpu_read(mipsr2bdemustats.movs));
arch/mips/kernel/mips-r2-to-r6-emul.c
2246
(unsigned long)__this_cpu_read(mipsr2emustats.hilo),
arch/mips/kernel/mips-r2-to-r6-emul.c
2247
(unsigned long)__this_cpu_read(mipsr2bdemustats.hilo));
arch/mips/kernel/mips-r2-to-r6-emul.c
2249
(unsigned long)__this_cpu_read(mipsr2emustats.muls),
arch/mips/kernel/mips-r2-to-r6-emul.c
2250
(unsigned long)__this_cpu_read(mipsr2bdemustats.muls));
arch/mips/kernel/mips-r2-to-r6-emul.c
2252
(unsigned long)__this_cpu_read(mipsr2emustats.divs),
arch/mips/kernel/mips-r2-to-r6-emul.c
2253
(unsigned long)__this_cpu_read(mipsr2bdemustats.divs));
arch/mips/kernel/mips-r2-to-r6-emul.c
2255
(unsigned long)__this_cpu_read(mipsr2emustats.dsps),
arch/mips/kernel/mips-r2-to-r6-emul.c
2256
(unsigned long)__this_cpu_read(mipsr2bdemustats.dsps));
arch/mips/kernel/mips-r2-to-r6-emul.c
2258
(unsigned long)__this_cpu_read(mipsr2emustats.bops),
arch/mips/kernel/mips-r2-to-r6-emul.c
2259
(unsigned long)__this_cpu_read(mipsr2bdemustats.bops));
arch/mips/kernel/mips-r2-to-r6-emul.c
2261
(unsigned long)__this_cpu_read(mipsr2emustats.traps),
arch/mips/kernel/mips-r2-to-r6-emul.c
2262
(unsigned long)__this_cpu_read(mipsr2bdemustats.traps));
arch/mips/kernel/mips-r2-to-r6-emul.c
2264
(unsigned long)__this_cpu_read(mipsr2emustats.fpus),
arch/mips/kernel/mips-r2-to-r6-emul.c
2265
(unsigned long)__this_cpu_read(mipsr2bdemustats.fpus));
arch/mips/kernel/mips-r2-to-r6-emul.c
2267
(unsigned long)__this_cpu_read(mipsr2emustats.loads),
arch/mips/kernel/mips-r2-to-r6-emul.c
2268
(unsigned long)__this_cpu_read(mipsr2bdemustats.loads));
arch/mips/kernel/mips-r2-to-r6-emul.c
2270
(unsigned long)__this_cpu_read(mipsr2emustats.stores),
arch/mips/kernel/mips-r2-to-r6-emul.c
2271
(unsigned long)__this_cpu_read(mipsr2bdemustats.stores));
arch/mips/kernel/mips-r2-to-r6-emul.c
2273
(unsigned long)__this_cpu_read(mipsr2emustats.llsc),
arch/mips/kernel/mips-r2-to-r6-emul.c
2274
(unsigned long)__this_cpu_read(mipsr2bdemustats.llsc));
arch/mips/kernel/mips-r2-to-r6-emul.c
2276
(unsigned long)__this_cpu_read(mipsr2emustats.dsemul),
arch/mips/kernel/mips-r2-to-r6-emul.c
2277
(unsigned long)__this_cpu_read(mipsr2bdemustats.dsemul));
arch/mips/kernel/mips-r2-to-r6-emul.c
2279
(unsigned long)__this_cpu_read(mipsr2bremustats.jrs));
arch/mips/kernel/mips-r2-to-r6-emul.c
2281
(unsigned long)__this_cpu_read(mipsr2bremustats.bltzl));
arch/mips/kernel/mips-r2-to-r6-emul.c
2283
(unsigned long)__this_cpu_read(mipsr2bremustats.bgezl));
arch/mips/kernel/mips-r2-to-r6-emul.c
2285
(unsigned long)__this_cpu_read(mipsr2bremustats.bltzll));
arch/mips/kernel/mips-r2-to-r6-emul.c
2287
(unsigned long)__this_cpu_read(mipsr2bremustats.bgezll));
arch/mips/kernel/mips-r2-to-r6-emul.c
2289
(unsigned long)__this_cpu_read(mipsr2bremustats.bltzal));
arch/mips/kernel/mips-r2-to-r6-emul.c
2291
(unsigned long)__this_cpu_read(mipsr2bremustats.bgezal));
arch/mips/kernel/mips-r2-to-r6-emul.c
2293
(unsigned long)__this_cpu_read(mipsr2bremustats.beql));
arch/mips/kernel/mips-r2-to-r6-emul.c
2295
(unsigned long)__this_cpu_read(mipsr2bremustats.bnel));
arch/mips/kernel/mips-r2-to-r6-emul.c
2297
(unsigned long)__this_cpu_read(mipsr2bremustats.blezl));
arch/mips/kernel/mips-r2-to-r6-emul.c
2299
(unsigned long)__this_cpu_read(mipsr2bremustats.bgtzl));
arch/mips/kernel/smp-bmips.c
347
action = __this_cpu_read(ipi_action_mask);
arch/powerpc/include/asm/time.h
103
return __this_cpu_read(decrementers_next_tb);
arch/powerpc/kernel/hw_breakpoint.c
236
struct perf_event *bp = __this_cpu_read(bp_per_reg[i]);
arch/powerpc/kernel/hw_breakpoint.c
246
info = counter_arch_bp(__this_cpu_read(bp_per_reg[i]));
arch/powerpc/kernel/hw_breakpoint.c
408
bp[i] = __this_cpu_read(bp_per_reg[i]);
arch/powerpc/kernel/hw_breakpoint.c
519
bp = __this_cpu_read(bp_per_reg[i]);
arch/powerpc/kernel/sysfs.c
547
if (__this_cpu_read(pmcs_enabled))
arch/powerpc/kernel/time.c
456
#define test_irq_work_pending() __this_cpu_read(irq_work_pending)
arch/powerpc/kvm/e500.c
107
__this_cpu_read(pcpu_sids.entry[entry->val]) == entry &&
arch/powerpc/kvm/e500mc.c
148
__this_cpu_read(last_vcpu_of_lpid[get_lpid(vcpu)]) != vcpu) {
arch/powerpc/lib/code-patching.c
292
patching_mm = __this_cpu_read(cpu_patching_context.mm);
arch/powerpc/lib/code-patching.c
293
text_poke_addr = __this_cpu_read(cpu_patching_context.addr);
arch/powerpc/lib/code-patching.c
335
text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK;
arch/powerpc/lib/code-patching.c
338
pte = __this_cpu_read(cpu_patching_context.pte);
arch/powerpc/lib/code-patching.c
478
patching_mm = __this_cpu_read(cpu_patching_context.mm);
arch/powerpc/lib/code-patching.c
479
text_poke_addr = __this_cpu_read(cpu_patching_context.addr);
arch/powerpc/lib/code-patching.c
527
text_poke_addr = (unsigned long)__this_cpu_read(cpu_patching_context.addr) & PAGE_MASK;
arch/powerpc/lib/code-patching.c
530
pte = __this_cpu_read(cpu_patching_context.pte);
arch/powerpc/perf/hv-24x7.c
1465
txn_flags = __this_cpu_read(hv_24x7_txn_flags);
arch/powerpc/perf/hv-24x7.c
1478
if (__this_cpu_read(hv_24x7_txn_err))
arch/powerpc/perf/hv-24x7.c
1536
WARN_ON_ONCE(__this_cpu_read(hv_24x7_txn_flags));
arch/powerpc/perf/hv-24x7.c
1581
txn_flags = __this_cpu_read(hv_24x7_txn_flags);
arch/powerpc/perf/hv-24x7.c
1588
ret = __this_cpu_read(hv_24x7_txn_err);
arch/powerpc/perf/hv-24x7.c
1632
WARN_ON_ONCE(!__this_cpu_read(hv_24x7_txn_flags));
arch/powerpc/platforms/pseries/iommu.c
273
tcep = __this_cpu_read(tce_page);
arch/powerpc/platforms/pseries/iommu.c
502
tcep = __this_cpu_read(tce_page);
arch/powerpc/platforms/pseries/lpar.c
364
u64 i = __this_cpu_read(dtl_entry_ridx);
arch/powerpc/sysdev/xive/common.c
1077
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
1514
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
1606
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
1626
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
1639
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
346
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/powerpc/sysdev/xive/common.c
441
struct xive_cpu *xc = __this_cpu_read(xive_cpu);
arch/s390/kernel/kprobes.c
238
kcb->prev_kprobe.kp = __this_cpu_read(current_kprobe);
arch/s390/kernel/vtime.c
139
if (smp_cpu_mtid && time_after64(jiffies_64, __this_cpu_read(mt_scaling_jiffies)))
arch/s390/kernel/vtime.c
98
u64 mult = __this_cpu_read(mt_scaling_mult);
arch/s390/kernel/vtime.c
99
u64 div = __this_cpu_read(mt_scaling_div);
arch/sh/kernel/kprobes.c
327
addr = __this_cpu_read(saved_current_opcode.addr);
arch/sparc/kernel/iommu-common.c
104
unsigned int pool_hash = __this_cpu_read(iommu_hash_common);
arch/sparc/kernel/nmi.c
107
if (__this_cpu_read(nmi_touch)) {
arch/sparc/kernel/nmi.c
111
if (!touched && __this_cpu_read(last_irq_sum) == sum) {
arch/sparc/kernel/nmi.c
113
if (__this_cpu_read(alert_counter) == 30 * nmi_hz)
arch/sparc/kernel/nmi.c
120
if (__this_cpu_read(wd_enabled)) {
arch/sparc/kernel/nmi.c
160
if (!__this_cpu_read(wd_enabled))
arch/sparc/kernel/nmi.c
217
if (__this_cpu_read(wd_enabled))
arch/sparc/kernel/nmi.c
231
if (!__this_cpu_read(wd_enabled))
arch/sparc/kernel/smp_64.c
1406
if (!__this_cpu_read(poke))
arch/x86/coco/sev/noinstr.c
58
new_ist = old_ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]);
arch/x86/coco/sev/noinstr.c
84
ist = __this_cpu_read(cpu_tss_rw.x86_tss.ist[IST_INDEX_VC]);
arch/x86/events/core.c
1465
if (__this_cpu_read(cpu_hw_events.enabled))
arch/x86/events/core.c
2330
__this_cpu_sub(cpu_hw_events.n_added, __this_cpu_read(cpu_hw_events.n_txn));
arch/x86/events/core.c
2331
__this_cpu_sub(cpu_hw_events.n_events, __this_cpu_read(cpu_hw_events.n_txn));
arch/x86/events/core.c
2332
__this_cpu_sub(cpu_hw_events.n_pair, __this_cpu_read(cpu_hw_events.n_txn_pair));
arch/x86/events/core.c
2333
__this_cpu_sub(cpu_hw_events.n_metric, __this_cpu_read(cpu_hw_events.n_txn_metric));
arch/x86/events/intel/core.c
3319
if (!__this_cpu_read(cpu_hw_events.enabled))
arch/x86/events/intel/core.c
3419
struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds);
arch/x86/events/intel/ds.c
3447
struct debug_store *ds = __this_cpu_read(cpu_hw_events.ds);
arch/x86/events/perf_event.h
1271
u64 disable_mask = __this_cpu_read(cpu_hw_events.perf_ctr_virt_mask);
arch/x86/events/perf_event.h
1296
u64 disable_mask = __this_cpu_read(cpu_hw_events.perf_ctr_virt_mask);
arch/x86/include/asm/cpu_entry_area.h
148
CEA_ESTACK_TOP(__this_cpu_read(cea_exception_stacks), name)
arch/x86/include/asm/cpu_entry_area.h
151
CEA_ESTACK_BOT(__this_cpu_read(cea_exception_stacks), name)
arch/x86/include/asm/debugreg.h
123
return __this_cpu_read(cpu_dr7) & DR_GLOBAL_ENABLE_MASK;
arch/x86/include/asm/fred.h
103
if (cpu_feature_enabled(X86_FEATURE_FRED) && (__this_cpu_read(fred_rsp0) != rsp0)) {
arch/x86/include/asm/hardirq.h
91
return __this_cpu_read(irq_stat.kvm_cpu_l1tf_flush_l1d);
arch/x86/include/asm/irq_stack.h
119
call_on_stack(__this_cpu_read(hardirq_stack_ptr), \
arch/x86/include/asm/irq_stack.h
138
if (user_mode(regs) || __this_cpu_read(hardirq_stack_inuse)) { \
arch/x86/include/asm/smp.h
140
#define __smp_processor_id() __this_cpu_read(cpu_number)
arch/x86/kernel/acpi/sleep.c
95
if (__this_cpu_read(cpu_info.cpuid_level) >= 0) {
arch/x86/kernel/apic/apic_numachip.c
75
local_apicid = __this_cpu_read(x86_cpu_to_apicid);
arch/x86/kernel/cpu/cacheinfo.c
188
size_in_kb = __this_cpu_read(cpu_info.x86_cache_size);
arch/x86/kernel/cpu/mce/core.c
1798
iv = __this_cpu_read(mce_next_interval);
arch/x86/kernel/cpu/mce/core.c
357
unsigned bank = __this_cpu_read(injectm.bank);
arch/x86/kernel/cpu/mce/core.c
396
if (__this_cpu_read(injectm.finished)) {
arch/x86/kernel/cpu/mce/core.c
431
if (__this_cpu_read(injectm.finished)) {
arch/x86/kernel/cpu/mce/threshold.c
65
return __this_cpu_read(storm_desc.poll_mode);
arch/x86/kernel/dumpstack_64.c
103
begin = (unsigned long)__this_cpu_read(cea_exception_stacks);
arch/x86/kernel/fpu/xstate.c
1498
u64 xfd = __this_cpu_read(xfd_state);
arch/x86/kernel/fpu/xstate.h
194
if (__this_cpu_read(xfd_state) != xfd)
arch/x86/kernel/fred.c
59
wrmsrq(MSR_IA32_FRED_RSP0, __this_cpu_read(fred_rsp0));
arch/x86/kernel/hw_breakpoint.c
486
set_debugreg(__this_cpu_read(cpu_debugreg[0]), 0);
arch/x86/kernel/hw_breakpoint.c
487
set_debugreg(__this_cpu_read(cpu_debugreg[1]), 1);
arch/x86/kernel/hw_breakpoint.c
488
set_debugreg(__this_cpu_read(cpu_debugreg[2]), 2);
arch/x86/kernel/hw_breakpoint.c
489
set_debugreg(__this_cpu_read(cpu_debugreg[3]), 3);
arch/x86/kernel/hw_breakpoint.c
491
set_debugreg(__this_cpu_read(cpu_dr7), 7);
arch/x86/kernel/irq.c
269
struct irq_desc *desc = __this_cpu_read(vector_irq[vector]);
arch/x86/kernel/irq.c
283
struct irq_desc *desc = __this_cpu_read(vector_irq[vector]);
arch/x86/kernel/irq.c
450
if (unlikely(!__this_cpu_read(posted_msi_handler_active)))
arch/x86/kernel/irq.c
546
if (IS_ERR_OR_NULL(__this_cpu_read(vector_irq[vector])))
arch/x86/kernel/irq.c
550
desc = __this_cpu_read(vector_irq[vector]);
arch/x86/kernel/irq.c
561
if (__this_cpu_read(vector_irq[vector]) != VECTOR_RETRIGGERED)
arch/x86/kernel/irq_32.c
135
irqstk = __this_cpu_read(softirq_stack_ptr);
arch/x86/kernel/irq_32.c
75
irqstk = __this_cpu_read(hardirq_stack_ptr);
arch/x86/kernel/kvm.c
264
if (__this_cpu_read(async_pf_enabled)) {
arch/x86/kernel/kvm.c
265
flags = __this_cpu_read(apf_reason.flags);
arch/x86/kernel/kvm.c
315
if (__this_cpu_read(async_pf_enabled)) {
arch/x86/kernel/kvm.c
316
token = __this_cpu_read(apf_reason.token);
arch/x86/kernel/kvm.c
403
if (!__this_cpu_read(async_pf_enabled))
arch/x86/kernel/nmi.c
376
if (regs->ip == __this_cpu_read(last_nmi_rip))
arch/x86/kernel/nmi.c
473
if (b2b && __this_cpu_read(swallow_nmi))
arch/x86/kernel/smpboot.c
1480
if (__this_cpu_read(cpu_info.x86) >= 4)
arch/x86/kernel/traps.c
1105
new_stack = (struct pt_regs *)__this_cpu_read(cpu_tss_rw.x86_tss.sp0) - 1;
arch/x86/kvm/svm/svm.c
468
if (multiplier == __this_cpu_read(current_tsc_ratio))
arch/x86/kvm/x86.c
3206
return __this_cpu_read(cpu_tsc_khz);
arch/x86/kvm/x86.c
3220
(static_cpu_has(X86_FEATURE_CONSTANT_TSC) || __this_cpu_read(cpu_tsc_khz))) {
arch/x86/virt/vmx/tdx/tdx.c
165
if (__this_cpu_read(tdx_lp_initialized))
arch/x86/xen/enlighten_pv.c
119
bool inhcall = __this_cpu_read(xen_in_preemptible_hcall);
arch/x86/xen/enlighten_pv.c
884
start = __this_cpu_read(idt_desc.address);
arch/x86/xen/enlighten_pv.c
885
end = start + __this_cpu_read(idt_desc.size) + 1;
arch/x86/xen/spinlock.c
37
int irq = __this_cpu_read(lock_kicker_irq);
arch/x86/xen/time.c
549
pvti = &__this_cpu_read(xen_vcpu)->time;
arch/x86/xen/time.c
55
src = &__this_cpu_read(xen_vcpu)->time;
arch/x86/xen/time.c
630
if (!__this_cpu_read(xen_vcpu)) {
arch/x86/xen/time.c
71
src = &__this_cpu_read(xen_vcpu)->time;
drivers/acpi/processor_idle.c
1171
pr = __this_cpu_read(processors);
drivers/acpi/processor_idle.c
683
pr = __this_cpu_read(processors);
drivers/acpi/processor_idle.c
712
struct acpi_processor *pr = __this_cpu_read(processors);
drivers/clocksource/arm_arch_timer.c
536
__wa = __this_cpu_read(timer_unstable_counter_workaround);
drivers/clocksource/arm_arch_timer.c
979
arch_timer_set_cntkctl(__this_cpu_read(saved_cntkctl));
drivers/clocksource/numachip.c
59
unsigned local_apicid = __this_cpu_read(x86_cpu_to_apicid) & 0xff;
drivers/cpufreq/powernow-k8.c
1112
struct powernow_k8_data *data = __this_cpu_read(powernow_data);
drivers/cpuidle/cpuidle-psci.c
124
struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev);
drivers/cpuidle/cpuidle-psci.c
138
struct device *pd_dev = __this_cpu_read(psci_cpuidle_data.dev);
drivers/cpuidle/cpuidle-psci.c
221
u32 *state = __this_cpu_read(psci_cpuidle_data.psci_states);
drivers/cpuidle/cpuidle-riscv-sbi.c
149
struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev);
drivers/cpuidle/cpuidle-riscv-sbi.c
159
struct device *pd_dev = __this_cpu_read(sbi_cpuidle_data.dev);
drivers/cpuidle/cpuidle-riscv-sbi.c
81
u32 *states = __this_cpu_read(sbi_cpuidle_data.states);
drivers/cpuidle/cpuidle.c
65
struct cpuidle_device *dev = __this_cpu_read(cpuidle_devices);
drivers/dma/fsl-qdma.c
654
__this_cpu_read(pre.queue) &&
drivers/dma/fsl-qdma.c
656
__this_cpu_read(pre.addr))
drivers/dma/fsl-qdma.c
674
__this_cpu_read(pre.addr)) {
drivers/firmware/arm_sdei.c
1111
if (__this_cpu_read(sdei_active_critical_event)) {
drivers/firmware/arm_sdei.c
1116
if (__this_cpu_read(sdei_active_normal_event)) {
drivers/gpu/drm/amd/display/amdgpu_dm/dc_fpu.c
56
depth = __this_cpu_read(fpu_recursion_depth);
drivers/hv/vmbus_drv.c
1373
struct task_struct *tsk = __this_cpu_read(vmbus_irqd);
drivers/hv/vmbus_drv.c
1386
return __this_cpu_read(vmbus_irq_pending);
drivers/irqchip/irq-apple-aic.c
531
if (__this_cpu_read(aic_fiq_unmasked) & BIT(aic_fiq_get_idx(d)))
drivers/irqchip/irq-xtensa-mx.c
101
mask |= __this_cpu_read(cached_irq_mask);
drivers/irqchip/irq-xtensa-mx.c
83
mask = __this_cpu_read(cached_irq_mask) & ~mask;
drivers/xen/events/events_2l.c
123
struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu);
drivers/xen/events/events_2l.c
173
struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu);
drivers/xen/events/events_2l.c
192
start_word_idx = __this_cpu_read(current_word_idx);
drivers/xen/events/events_2l.c
193
start_bit_idx = __this_cpu_read(current_bit_idx);
drivers/xen/events/events_base.c
1695
info->irq_epoch = __this_cpu_read(irq_epoch);
drivers/xen/events/events_base.c
1704
struct vcpu_info *vcpu_info = __this_cpu_read(xen_vcpu);
fs/buffer.c
1370
struct buffer_head *bh = __this_cpu_read(bh_lrus.bhs[i]);
fs/buffer.c
1377
__this_cpu_read(bh_lrus.bhs[i - 1]));
include/asm-generic/irq_regs.h
21
return __this_cpu_read(__irq_regs);
include/asm-generic/irq_regs.h
28
old_regs = __this_cpu_read(__irq_regs);
include/linux/context_tracking_state.h
110
return __this_cpu_read(context_tracking.nesting);
include/linux/context_tracking_state.h
122
return __this_cpu_read(context_tracking.nmi_nesting);
include/linux/context_tracking_state.h
148
return context_tracking_enabled() && __this_cpu_read(context_tracking.active);
include/linux/cpuidle.h
204
{return __this_cpu_read(cpuidle_devices); }
include/linux/interrupt.h
528
#define local_softirq_pending() (__this_cpu_read(local_softirq_pending_ref))
include/linux/interrupt.h
641
return __this_cpu_read(pending_timer_softirq);
include/linux/kprobes.h
389
return __this_cpu_read(current_kprobe);
include/linux/netdevice.h
3590
return unlikely(__this_cpu_read(softnet_data.xmit.recursion) >
include/linux/netdevice.h
5306
return __this_cpu_read(softnet_data.xmit.more);
include/linux/netfilter/x_tables.h
362
addend = (__this_cpu_read(xt_recseq.sequence) + 1) & 1;
kernel/bpf/helpers.c
359
flags = __this_cpu_read(irqsave_flags);
kernel/events/core.c
10537
seq = __this_cpu_read(perf_throttled_seq);
kernel/events/core.c
12476
unsigned int flags = __this_cpu_read(nop_txn_flags);
kernel/events/core.c
12489
unsigned int flags = __this_cpu_read(nop_txn_flags);
kernel/events/core.c
3947
if (__this_cpu_read(perf_sched_cb_usages))
kernel/events/core.c
4324
if (__this_cpu_read(perf_sched_cb_usages))
kernel/events/core.c
479
return __this_cpu_read(guest_ctx_loaded);
kernel/events/core.c
6423
if (WARN_ON_ONCE(__this_cpu_read(guest_ctx_loaded)))
kernel/events/core.c
6449
if (WARN_ON_ONCE(!__this_cpu_read(guest_ctx_loaded)))
kernel/events/core.c
652
running_len = __this_cpu_read(running_sample_length);
kernel/irq_work.c
33
struct task_struct *tsk = __this_cpu_read(irq_workd);
kernel/locking/lockdep.c
4495
} else if (__this_cpu_read(lockdep_recursion))
kernel/rcu/tree.c
2702
WARN_ON_ONCE(time_before(j, __this_cpu_read(rcu_data.last_sched_clock)));
kernel/rcu/tree.c
2906
t = __this_cpu_read(rcu_data.rcu_cpu_kthread_task);
kernel/rcu/tree.c
2908
rcu_wake_cond(t, __this_cpu_read(rcu_data.rcu_cpu_kthread_status));
kernel/rcu/tree.c
2932
return __this_cpu_read(rcu_data.rcu_cpu_has_work);
kernel/rcu/tree_exp.h
873
__this_cpu_read(rcu_data.cpu_no_qs.b.exp))
kernel/rcu/tree_plugin.h
301
if (__this_cpu_read(rcu_data.cpu_no_qs.b.norm)) {
kernel/rcu/tree_plugin.h
303
__this_cpu_read(rcu_data.gp_seq),
kernel/rcu/tree_plugin.h
602
return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) ||
kernel/rcu/tree_plugin.h
830
__this_cpu_read(rcu_data.core_needs_qs) &&
kernel/rcu/tree_plugin.h
831
__this_cpu_read(rcu_data.cpu_no_qs.b.norm) &&
kernel/rcu/tree_plugin.h
953
if (!__this_cpu_read(rcu_data.cpu_no_qs.s))
kernel/rcu/tree_plugin.h
956
__this_cpu_read(rcu_data.gp_seq), TPS("cpuqs"));
kernel/rcu/tree_plugin.h
958
if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp))
kernel/sched/ext.c
2471
struct scx_kick_syncs __rcu *ks = __this_cpu_read(scx_kick_syncs);
kernel/sched/ext.c
5726
struct scx_kick_syncs __rcu *ksyncs_pcpu = __this_cpu_read(scx_kick_syncs);
kernel/sched/ext.c
5906
ddsp_task = __this_cpu_read(direct_dispatch_task);
kernel/sched/ext.c
6208
return scx_dsp_max_batch - __this_cpu_read(scx_dsp_ctx->cursor);
kernel/sched/ext_internal.h
1172
return __this_cpu_read(scx_locked_rq_state);
kernel/sched/fair.c
7324
int factor = __this_cpu_read(sd_llc_size);
kernel/sched/sched.h
3966
cpu_cid = __this_cpu_read(mm->mm_cid.pcpu->cid);
kernel/softirq.c
1081
*__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head;
kernel/softirq.c
1089
*__this_cpu_read(tasklet_hi_vec.tail) = per_cpu(tasklet_hi_vec, cpu).head;
kernel/softirq.c
153
return __this_cpu_read(softirq_ctrl.cnt) != 0;
kernel/softirq.c
489
if (!force_irqthreads() || !__this_cpu_read(ksoftirqd)) {
kernel/softirq.c
701
struct task_struct *tsk = __this_cpu_read(ktimerd);
kernel/softirq.c
78
struct task_struct *tsk = __this_cpu_read(ksoftirqd);
kernel/time/tick-common.c
74
struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev);
kernel/time/tick-oneshot.c
127
ret = __this_cpu_read(tick_cpu_device.mode) == TICKDEV_MODE_ONESHOT;
kernel/time/tick-oneshot.c
29
struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev);
kernel/time/tick-oneshot.c
56
struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev);
kernel/time/tick-sched.c
1002
struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev);
kernel/time/tick-sched.c
1355
return __this_cpu_read(tick_cpu_device.evtdev)->next_event;
kernel/time/tick-sched.c
1372
struct clock_event_device *dev = __this_cpu_read(tick_cpu_device.evtdev);
kernel/time/timer.c
2121
return __this_cpu_read(timer_bases[BASE_LOCAL].is_idle);
kernel/trace/trace.c
2261
__this_cpu_read(trace_buffered_event) !=
kernel/trace/trace.c
2358
if ((entry = __this_cpu_read(trace_buffered_event))) {
kernel/trace/trace.c
2745
if (__this_cpu_read(user_stack_count))
kernel/trace/trace_sched_switch.c
363
if (!__this_cpu_read(trace_taskinfo_save))
kernel/trace/trace_stack.c
301
if (__this_cpu_read(disable_stack_tracer) != 1)
kernel/unwind/deferred.c
88
cnt |= __this_cpu_read(unwind_ctx_ctr) + 2;
kernel/watchdog.c
471
u8 tail = __this_cpu_read(cpustat_tail);
kernel/watchdog.c
477
old_stat = __this_cpu_read(cpustat_old[i]);
kernel/watchdog.c
498
u8 tail = __this_cpu_read(cpustat_tail);
kernel/watchdog.c
515
__this_cpu_read(cpustat_util[group][STATS_SYSTEM]),
kernel/watchdog.c
516
__this_cpu_read(cpustat_util[group][STATS_SOFTIRQ]),
kernel/watchdog.c
517
__this_cpu_read(cpustat_util[group][STATS_HARDIRQ]),
kernel/watchdog.c
518
__this_cpu_read(cpustat_util[group][STATS_IDLE]));
kernel/watchdog.c
551
int tail = __this_cpu_read(cpustat_tail);
kernel/watchdog.c
554
util = __this_cpu_read(cpustat_util[tail][STATS_HARDIRQ]);
kernel/watchdog.c
560
if (!__this_cpu_read(snapshot_taken)) {
kernel/watchdog.c
578
if (__this_cpu_read(snapshot_taken)) {
kernel/watchdog.c
828
if (unlikely(__this_cpu_read(softlockup_touch_sync))) {
kernel/watchdog.c
842
touch_ts = __this_cpu_read(watchdog_touch_ts);
kernel/watchdog_perf.c
63
delta = now - __this_cpu_read(last_timestamp);
kernel/workqueue.c
2239
new_cpu = __this_cpu_read(wq_rr_cpu_last);
lib/cpumask.c
140
prev = __this_cpu_read(distribute_cpu_mask_prev);
lib/cpumask.c
161
prev = __this_cpu_read(distribute_cpu_mask_prev);
lib/percpu_counter.c
106
count = __this_cpu_read(*fbc->counters);
lib/percpu_counter.c
126
count = __this_cpu_read(*fbc->counters) + amount;
lib/percpu_counter.c
152
count = __this_cpu_read(*fbc->counters);
lib/percpu_counter.c
340
count = __this_cpu_read(*fbc->counters);
lib/percpu_counter.c
385
count = __this_cpu_read(*fbc->counters);
lib/percpu_test.c
12
WARN(__this_cpu_read(pcp) != (expected), \
lib/percpu_test.c
14
__this_cpu_read(pcp), __this_cpu_read(pcp), \
mm/memcontrol-v1.c
554
val = __this_cpu_read(memcg->events_percpu->nr_page_events);
mm/memcontrol-v1.c
555
next = __this_cpu_read(memcg->events_percpu->targets[target]);
mm/vmstat.c
362
x = delta + __this_cpu_read(*p);
mm/vmstat.c
364
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
398
x = delta + __this_cpu_read(*p);
mm/vmstat.c
400
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
445
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
468
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
501
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
524
t = __this_cpu_read(pcp->stat_threshold);
mm/vmstat.c
840
if (!__this_cpu_read(pcp->expire) ||
mm/vmstat.c
841
!__this_cpu_read(pcp->count))
mm/vmstat.c
857
if (__this_cpu_read(pcp->count)) {
net/core/dev.c
3457
skb->next = __this_cpu_read(softnet_data.completion_queue);
net/core/dev.c
4356
return __this_cpu_read(softnet_data.xmit.skip_txqueue);
net/ipv4/tcp_input.c
7661
isn = __this_cpu_read(tcp_tw_isn);
net/rds/ib_recv.c
487
chpfirst = __this_cpu_read(cache->percpu->first);
net/rds/ib_recv.c
496
if (__this_cpu_read(cache->percpu->count) < RDS_IB_RECYCLE_BATCH_COUNT)
virt/kvm/kvm_main.c
3830
if (vcpu == __this_cpu_read(kvm_running_vcpu)) {
virt/kvm/kvm_main.c
5600
if (__this_cpu_read(virtualization_enabled))
virt/kvm/kvm_main.c
5625
if (!__this_cpu_read(virtualization_enabled))
virt/kvm/kvm_main.c
6424
vcpu = __this_cpu_read(kvm_running_vcpu);