arch/alpha/kernel/irq.c
34
DEFINE_PER_CPU(unsigned long, irq_pmi_count);
arch/alpha/kernel/perf_event.c
53
DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
arch/alpha/kernel/time.c
58
DEFINE_PER_CPU(u8, irq_work_pending);
arch/alpha/kernel/time.c
88
static DEFINE_PER_CPU(struct clock_event_device, cpu_ce);
arch/arc/kernel/kprobes.c
20
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/arc/kernel/kprobes.c
21
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/arc/kernel/perf_event.c
259
static DEFINE_PER_CPU(struct arc_pmu_cpu, arc_pmu_cpu);
arch/arc/kernel/setup.c
642
static DEFINE_PER_CPU(struct cpu, cpu_topology);
arch/arc/kernel/smp.c
248
static DEFINE_PER_CPU(unsigned long, ipi_data);
arch/arc/kernel/smp.c
387
static DEFINE_PER_CPU(int, ipi_dev);
arch/arc/mm/tlb.c
19
DEFINE_PER_CPU(unsigned int, asid_cache) = MM_CTXT_FIRST_CYCLE;
arch/arm/kernel/hw_breakpoint.c
31
static DEFINE_PER_CPU(struct perf_event *, bp_on_reg[ARM_MAX_BRP]);
arch/arm/kernel/hw_breakpoint.c
34
static DEFINE_PER_CPU(struct perf_event *, wp_on_reg[ARM_MAX_WRP]);
arch/arm/kernel/machine_kexec.c
97
static DEFINE_PER_CPU(call_single_data_t, cpu_stop_csd) =
arch/arm/kernel/process.c
40
DEFINE_PER_CPU(struct task_struct *, __entry_task);
arch/arm/kernel/setup.c
158
DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data);
arch/arm/kernel/smp.c
618
static DEFINE_PER_CPU(struct completion *, cpu_completion);
arch/arm/kernel/smp.c
789
static DEFINE_PER_CPU(unsigned long, l_p_j_ref);
arch/arm/kernel/smp.c
790
static DEFINE_PER_CPU(unsigned long, l_p_j_ref_freq);
arch/arm/kernel/smp_twd.c
30
static DEFINE_PER_CPU(bool, percpu_setup_called);
arch/arm/kernel/traps.c
904
asmlinkage DEFINE_PER_CPU(u8 *, overflow_stack_ptr);
arch/arm/mach-omap2/omap-mpuss-lowpower.c
91
static DEFINE_PER_CPU(struct omap4_cpu_pm_info, omap4_pm_info);
arch/arm/mach-omap2/omap-wakeupgen.c
188
static DEFINE_PER_CPU(u32 [MAX_NR_REG_BANKS], irqmasks);
arch/arm/mach-qcom/platsmp.c
314
static DEFINE_PER_CPU(int, cold_boot_done);
arch/arm/mm/context.c
46
static DEFINE_PER_CPU(atomic64_t, active_asids);
arch/arm/mm/context.c
47
static DEFINE_PER_CPU(u64, reserved_asids);
arch/arm/mm/proc-v7-bugs.c
271
static DEFINE_PER_CPU(bool, spectre_warned);
arch/arm/mm/proc-v7-bugs.c
39
DEFINE_PER_CPU(harden_branch_predictor_fn_t, harden_branch_predictor_fn);
arch/arm/probes/kprobes/core.c
44
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/arm/probes/kprobes/core.c
45
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/arm/xen/enlighten.c
51
DEFINE_PER_CPU(struct vcpu_info *, xen_vcpu);
arch/arm/xen/enlighten.c
55
DEFINE_PER_CPU(uint32_t, xen_vcpu_id);
arch/arm64/kernel/cpuinfo.c
33
DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data);
arch/arm64/kernel/debug-monitors.c
81
static DEFINE_PER_CPU(int, mde_ref_count);
arch/arm64/kernel/debug-monitors.c
82
static DEFINE_PER_CPU(int, kde_ref_count);
arch/arm64/kernel/entry-common.c
168
static DEFINE_PER_CPU(int, __in_cortex_a76_erratum_1463225_wa);
arch/arm64/kernel/fpsimd.c
122
DEFINE_PER_CPU(struct cpu_fp_state, fpsimd_last_state);
arch/arm64/kernel/hw_breakpoint.c
32
static DEFINE_PER_CPU(struct perf_event *, bp_on_reg[ARM_MAX_BRP]);
arch/arm64/kernel/hw_breakpoint.c
35
static DEFINE_PER_CPU(struct perf_event *, wp_on_reg[ARM_MAX_WRP]);
arch/arm64/kernel/hw_breakpoint.c
38
static DEFINE_PER_CPU(int, stepping_kernel_bp);
arch/arm64/kernel/irq.c
31
DEFINE_PER_CPU(struct nmi_ctx, nmi_contexts);
arch/arm64/kernel/irq.c
33
DEFINE_PER_CPU(unsigned long *, irq_stack_ptr);
arch/arm64/kernel/irq.c
39
DEFINE_PER_CPU(unsigned long *, irq_shadow_call_stack_ptr);
arch/arm64/kernel/paravirt.c
32
static DEFINE_PER_CPU(struct pv_time_stolen_time_region, stolen_time_region);
arch/arm64/kernel/probes/kprobes.c
39
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/arm64/kernel/probes/kprobes.c
40
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/arm64/kernel/process.c
572
DEFINE_PER_CPU(struct task_struct *, __entry_task);
arch/arm64/kernel/sdei.c
37
DEFINE_PER_CPU(unsigned long *, sdei_stack_normal_ptr);
arch/arm64/kernel/sdei.c
38
DEFINE_PER_CPU(unsigned long *, sdei_stack_critical_ptr);
arch/arm64/kernel/sdei.c
44
DEFINE_PER_CPU(unsigned long *, sdei_shadow_call_stack_normal_ptr);
arch/arm64/kernel/sdei.c
45
DEFINE_PER_CPU(unsigned long *, sdei_shadow_call_stack_critical_ptr);
arch/arm64/kernel/sdei.c
48
DEFINE_PER_CPU(struct sdei_registered_event *, sdei_active_normal_event);
arch/arm64/kernel/sdei.c
49
DEFINE_PER_CPU(struct sdei_registered_event *, sdei_active_critical_event);
arch/arm64/kvm/arm.c
109
DEFINE_PER_CPU(unsigned long, kvm_arm_hyp_stack_base);
arch/arm64/kvm/arm.c
116
static DEFINE_PER_CPU(unsigned char, kvm_hyp_initialized);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
23
DEFINE_PER_CPU(struct kvm_nvhe_init_params, kvm_init_params);
arch/arm64/kvm/hyp/nvhe/mem_protect.c
28
static DEFINE_PER_CPU(struct pkvm_hyp_vm *, __current_vm);
arch/arm64/kvm/hyp/nvhe/mm.c
33
static DEFINE_PER_CPU(struct hyp_fixmap_slot, fixmap_slots);
arch/arm64/kvm/hyp/nvhe/pkvm.c
29
static DEFINE_PER_CPU(struct pkvm_hyp_vcpu *, loaded_hyp_vcpu);
arch/arm64/kvm/hyp/nvhe/psci-relay.c
41
static DEFINE_PER_CPU(struct psci_boot_args, cpu_on_args) = PSCI_BOOT_ARGS_INIT;
arch/arm64/kvm/hyp/nvhe/psci-relay.c
42
static DEFINE_PER_CPU(struct psci_boot_args, suspend_args) = PSCI_BOOT_ARGS_INIT;
arch/arm64/kvm/hyp/nvhe/stacktrace.c
15
DEFINE_PER_CPU(struct kvm_nvhe_stacktrace_info, kvm_stacktrace_info);
arch/arm64/kvm/hyp/nvhe/stacktrace.c
40
DEFINE_PER_CPU(unsigned long [NVHE_STACKTRACE_SIZE/sizeof(long)], pkvm_stacktrace);
arch/arm64/kvm/hyp/nvhe/switch.c
32
DEFINE_PER_CPU(struct kvm_host_data, kvm_host_data);
arch/arm64/kvm/hyp/nvhe/switch.c
33
DEFINE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/kvm/hyp/nvhe/switch.c
34
DEFINE_PER_CPU(unsigned long, kvm_hyp_vector);
arch/arm64/kvm/hyp/vhe/switch.c
32
DEFINE_PER_CPU(struct kvm_host_data, kvm_host_data);
arch/arm64/kvm/hyp/vhe/switch.c
33
DEFINE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/kvm/hyp/vhe/switch.c
34
DEFINE_PER_CPU(unsigned long, kvm_hyp_vector);
arch/arm64/kvm/pmu.c
11
static DEFINE_PER_CPU(struct kvm_pmu_events, kvm_pmu_events);
arch/arm64/kvm/vgic/vgic-v3-nested.c
37
static DEFINE_PER_CPU(struct shadow_if, shadow_if);
arch/arm64/kvm/vmid.c
25
static DEFINE_PER_CPU(atomic64_t, active_vmids);
arch/arm64/kvm/vmid.c
26
static DEFINE_PER_CPU(u64, reserved_vmids);
arch/arm64/mm/context.c
26
static DEFINE_PER_CPU(atomic64_t, active_asids);
arch/arm64/mm/context.c
27
static DEFINE_PER_CPU(u64, reserved_asids);
arch/csky/kernel/probes/kprobes.c
16
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/csky/kernel/probes/kprobes.c
17
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/csky/kernel/smp.c
42
static DEFINE_PER_CPU(struct ipi_data_struct, ipi_data);
arch/csky/mm/context.c
14
static DEFINE_PER_CPU(atomic64_t, active_asids);
arch/csky/mm/context.c
15
static DEFINE_PER_CPU(u64, reserved_asids);
arch/hexagon/kernel/smp.c
36
static DEFINE_PER_CPU(struct ipi_data, ipi_data);
arch/hexagon/kernel/time.c
118
static DEFINE_PER_CPU(struct clock_event_device, clock_events);
arch/loongarch/kernel/hw_breakpoint.c
14
static DEFINE_PER_CPU(struct perf_event *, bp_on_reg[LOONGARCH_MAX_BRP]);
arch/loongarch/kernel/hw_breakpoint.c
17
static DEFINE_PER_CPU(struct perf_event *, wp_on_reg[LOONGARCH_MAX_WRP]);
arch/loongarch/kernel/irq.c
24
DEFINE_PER_CPU(unsigned long, irq_stack);
arch/loongarch/kernel/kfpu.c
19
static DEFINE_PER_CPU(bool, in_kernel_fpu);
arch/loongarch/kernel/kfpu.c
20
static DEFINE_PER_CPU(unsigned int, euen_current);
arch/loongarch/kernel/kprobes.c
10
DEFINE_PER_CPU(struct kprobe *, current_kprobe);
arch/loongarch/kernel/kprobes.c
11
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/loongarch/kernel/paravirt.c
13
static DEFINE_PER_CPU(struct kvm_steal_time, steal_time) __aligned(64);
arch/loongarch/kernel/perf_event.c
104
static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = {
arch/loongarch/kernel/process.c
347
static DEFINE_PER_CPU(call_single_data_t, backtrace_csd);
arch/loongarch/kernel/setup.c
63
DEFINE_PER_CPU(unsigned long, kernelsp);
arch/loongarch/kernel/smp.c
77
static DEFINE_PER_CPU(int, cpu_state);
arch/loongarch/kernel/time.c
29
static DEFINE_PER_CPU(struct clock_event_device, constant_clockevent_device);
arch/microblaze/kernel/setup.c
38
DEFINE_PER_CPU(unsigned int, KSP); /* Saved kernel stack pointer */
arch/microblaze/kernel/setup.c
39
DEFINE_PER_CPU(unsigned int, KM); /* Kernel/user mode */
arch/microblaze/kernel/setup.c
40
DEFINE_PER_CPU(unsigned int, ENTRY_SP); /* Saved SP on kernel entry */
arch/microblaze/kernel/setup.c
41
DEFINE_PER_CPU(unsigned int, R11_SAVE); /* Temp variable for entry */
arch/microblaze/kernel/setup.c
42
DEFINE_PER_CPU(unsigned int, CURRENT_SAVE); /* Saved current pointer */
arch/mips/cavium-octeon/octeon-irq.c
24
static DEFINE_PER_CPU(unsigned long, octeon_irq_ciu0_en_mirror);
arch/mips/cavium-octeon/octeon-irq.c
25
static DEFINE_PER_CPU(unsigned long, octeon_irq_ciu1_en_mirror);
arch/mips/cavium-octeon/octeon-irq.c
26
static DEFINE_PER_CPU(raw_spinlock_t, octeon_irq_ciu_spinlock);
arch/mips/cavium-octeon/octeon-irq.c
27
static DEFINE_PER_CPU(unsigned int, octeon_irq_ciu3_idt_ip2);
arch/mips/cavium-octeon/octeon-irq.c
29
static DEFINE_PER_CPU(unsigned int, octeon_irq_ciu3_idt_ip3);
arch/mips/cavium-octeon/octeon-irq.c
30
static DEFINE_PER_CPU(struct octeon_ciu3_info *, octeon_ciu3_info);
arch/mips/cavium-octeon/smp.c
286
static DEFINE_PER_CPU(int, cpu_state);
arch/mips/kernel/cevt-bcm1480.c
93
static DEFINE_PER_CPU(struct clock_event_device, sibyte_hpt_clockevent);
arch/mips/kernel/cevt-bcm1480.c
94
static DEFINE_PER_CPU(char [18], sibyte_hpt_name);
arch/mips/kernel/cevt-r4k.c
109
DEFINE_PER_CPU(struct clock_event_device, mips_clockevent_device);
arch/mips/kernel/cevt-sb1250.c
92
static DEFINE_PER_CPU(struct clock_event_device, sibyte_hpt_clockevent);
arch/mips/kernel/cevt-sb1250.c
93
static DEFINE_PER_CPU(char [18], sibyte_hpt_name);
arch/mips/kernel/kprobes.c
44
DEFINE_PER_CPU(struct kprobe *, current_kprobe);
arch/mips/kernel/kprobes.c
45
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/mips/kernel/mips-r2-to-r6-emul.c
50
static DEFINE_PER_CPU(struct mips_r2_emulator_stats, mipsr2emustats);
arch/mips/kernel/mips-r2-to-r6-emul.c
51
static DEFINE_PER_CPU(struct mips_r2_emulator_stats, mipsr2bdemustats);
arch/mips/kernel/mips-r2-to-r6-emul.c
52
static DEFINE_PER_CPU(struct mips_r2br_emulator_stats, mipsr2bremustats);
arch/mips/kernel/perf_event_mipsxx.c
48
DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = {
arch/mips/kernel/process.c
730
static DEFINE_PER_CPU(call_single_data_t, backtrace_csd) =
arch/mips/kernel/smp-bmips.c
328
static DEFINE_PER_CPU(int, ipi_action_mask);
arch/mips/kernel/smp.c
745
static DEFINE_PER_CPU(call_single_data_t, tick_broadcast_csd) =
arch/mips/kernel/time.c
31
static DEFINE_PER_CPU(unsigned long, pcp_lpj_ref);
arch/mips/kernel/time.c
32
static DEFINE_PER_CPU(unsigned long, pcp_lpj_ref_freq);
arch/mips/kernel/topology.c
9
static DEFINE_PER_CPU(struct cpu, cpu_devices);
arch/mips/loongson64/hpet.c
21
DEFINE_PER_CPU(struct clock_event_device, hpet_clockevent_device);
arch/mips/loongson64/smp.c
27
DEFINE_PER_CPU(int, cpu_state);
arch/mips/math-emu/me-debugfs.c
12
DEFINE_PER_CPU(struct mips_fpu_emulator_stats, fpuemustats);
arch/mips/mm/context.c
13
static DEFINE_PER_CPU(u64, reserved_mmids);
arch/mips/sgi-ip27/ip27-irq.c
35
static DEFINE_PER_CPU(unsigned long [2], irq_enable_mask);
arch/mips/sgi-ip27/ip27-timer.c
43
static DEFINE_PER_CPU(struct clock_event_device, hub_rt_clockevent);
arch/mips/sgi-ip27/ip27-timer.c
44
static DEFINE_PER_CPU(char [11], hub_rt_name);
arch/mips/sgi-ip30/ip30-irq.c
27
static DEFINE_PER_CPU(unsigned long, irq_enable_mask);
arch/openrisc/kernel/time.c
67
static DEFINE_PER_CPU(struct clock_event_device, clockevent_openrisc_timer);
arch/parisc/kernel/irq.c
368
static DEFINE_PER_CPU(union irq_stack_union, irq_stack_union) = {
arch/parisc/kernel/irq.c
39
static DEFINE_PER_CPU(unsigned long, local_ack_eiem) = ~0UL;
arch/parisc/kernel/kprobes.c
17
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/parisc/kernel/kprobes.c
18
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/parisc/kernel/processor.c
41
DEFINE_PER_CPU(struct cpuinfo_parisc, cpu_data);
arch/parisc/kernel/smp.c
64
static DEFINE_PER_CPU(spinlock_t, ipi_lock);
arch/parisc/kernel/time.c
26
static DEFINE_PER_CPU(struct clock_event_device, parisc_clockevent_device);
arch/parisc/kernel/topology.c
21
static DEFINE_PER_CPU(struct cpu, cpu_devices);
arch/powerpc/kernel/cacheinfo.c
127
static DEFINE_PER_CPU(struct cache_dir *, cache_dir_pcpu);
arch/powerpc/kernel/hw_breakpoint.c
34
static DEFINE_PER_CPU(struct perf_event *, bp_per_reg[HBP_NUM_MAX]);
arch/powerpc/kernel/iommu.c
97
static DEFINE_PER_CPU(unsigned int, iommu_pool_hash);
arch/powerpc/kernel/kprobes.c
31
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/powerpc/kernel/kprobes.c
32
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/powerpc/kernel/process.c
684
static DEFINE_PER_CPU(struct arch_hw_breakpoint, current_brk[HBP_NUM_MAX]);
arch/powerpc/kernel/setup-common.c
186
DEFINE_PER_CPU(unsigned int, cpu_pvr);
arch/powerpc/kernel/smp.c
120
DEFINE_PER_CPU(cpumask_var_t, thread_group_l1_cache_map);
arch/powerpc/kernel/smp.c
127
DEFINE_PER_CPU(cpumask_var_t, thread_group_l2_cache_map);
arch/powerpc/kernel/smp.c
133
DEFINE_PER_CPU(cpumask_var_t, thread_group_l3_cache_map);
arch/powerpc/kernel/smp.c
77
static DEFINE_PER_CPU(int, cpu_state) = { 0 };
arch/powerpc/kernel/smp.c
86
DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map);
arch/powerpc/kernel/smp.c
87
DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map);
arch/powerpc/kernel/smp.c
88
DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map);
arch/powerpc/kernel/smp.c
89
DEFINE_PER_CPU(cpumask_var_t, cpu_core_map);
arch/powerpc/kernel/smp.c
90
static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map);
arch/powerpc/kernel/sysfs.c
34
static DEFINE_PER_CPU(struct cpu, cpu_devices);
arch/powerpc/kernel/sysfs.c
540
static DEFINE_PER_CPU(char, pmcs_enabled);
arch/powerpc/kernel/time.c
116
DEFINE_PER_CPU(u64, decrementers_next_tb) = DEC_CLOCKEVENT_STOPPED;
arch/powerpc/kernel/time.c
118
static DEFINE_PER_CPU(struct clock_event_device, decrementers);
arch/powerpc/kernel/time.c
453
DEFINE_PER_CPU(u8, irq_work_pending);
arch/powerpc/kernel/watchdog.c
84
static DEFINE_PER_CPU(struct hrtimer, wd_hrtimer);
arch/powerpc/kernel/watchdog.c
85
static DEFINE_PER_CPU(u64, wd_timer_tb);
arch/powerpc/kvm/book3s_hv.c
3238
static DEFINE_PER_CPU(struct kvm *, cpu_in_guest);
arch/powerpc/kvm/book3s_hv.c
3431
static DEFINE_PER_CPU(struct preempted_vcore_list, preempted_vcores);
arch/powerpc/kvm/book3s_hv.c
4113
static DEFINE_PER_CPU(u64, l1_to_l2_cs);
arch/powerpc/kvm/book3s_hv.c
4114
static DEFINE_PER_CPU(u64, l2_to_l1_cs);
arch/powerpc/kvm/book3s_hv.c
4115
static DEFINE_PER_CPU(u64, l2_runtime_agg);
arch/powerpc/kvm/e500.c
56
static DEFINE_PER_CPU(struct pcpu_id_table, pcpu_sids);
arch/powerpc/kvm/e500.c
60
static DEFINE_PER_CPU(unsigned long, pcpu_last_used_sid);
arch/powerpc/kvm/e500mc.c
115
static DEFINE_PER_CPU(struct kvm_vcpu *[KVMPPC_NR_LPIDS], last_vcpu_of_lpid);
arch/powerpc/lib/code-patching.c
61
static DEFINE_PER_CPU(struct patch_context, cpu_patching_context);
arch/powerpc/mm/book3s64/hash_tlb.c
32
DEFINE_PER_CPU(struct ppc64_tlb_batch, ppc64_tlb_batch);
arch/powerpc/mm/book3s64/radix_tlb.c
738
static DEFINE_PER_CPU(unsigned int, mm_cpumask_trim_clock);
arch/powerpc/mm/nohash/tlb.c
108
DEFINE_PER_CPU(int, next_tlbcam_idx);
arch/powerpc/perf/core-book3s.c
64
static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
arch/powerpc/perf/core-fsl-emb.c
25
static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
arch/powerpc/perf/hv-24x7.c
216
static DEFINE_PER_CPU(int, hv_24x7_txn_flags);
arch/powerpc/perf/hv-24x7.c
217
static DEFINE_PER_CPU(int, hv_24x7_txn_err);
arch/powerpc/perf/hv-24x7.c
223
static DEFINE_PER_CPU(struct hv_24x7_hw, hv_24x7_hw);
arch/powerpc/perf/hv-24x7.c
232
static DEFINE_PER_CPU(char, hv_24x7_resb[H24x7_DATA_BUFFER_SIZE]) __aligned(4096);
arch/powerpc/perf/imc-pmu.c
26
static DEFINE_PER_CPU(struct imc_pmu_ref *, local_nest_imc_refc);
arch/powerpc/perf/imc-pmu.c
40
static DEFINE_PER_CPU(u64 *, thread_imc_mem);
arch/powerpc/perf/imc-pmu.c
45
static DEFINE_PER_CPU(u64 *, trace_imc_mem);
arch/powerpc/perf/vpa-dtl.c
111
static DEFINE_PER_CPU(struct vpa_pmu_ctx, vpa_pmu_ctx);
arch/powerpc/perf/vpa-dtl.c
112
static DEFINE_PER_CPU(struct vpa_dtl, vpa_dtl_cpu);
arch/powerpc/platforms/chrp/setup.c
63
DEFINE_PER_CPU(struct timer_list, heartbeat_timer);
arch/powerpc/platforms/powernv/opal-call.c
14
static DEFINE_PER_CPU(unsigned int, opal_trace_depth);
arch/powerpc/platforms/powernv/opal-tracepoints.c
45
static DEFINE_PER_CPU(unsigned int, opal_trace_depth);
arch/powerpc/platforms/powernv/rng.c
30
static DEFINE_PER_CPU(struct pnv_rng *, pnv_rng);
arch/powerpc/platforms/powernv/subcore.c
148
static DEFINE_PER_CPU(struct split_state, split_state);
arch/powerpc/platforms/powernv/vas.c
27
static DEFINE_PER_CPU(int, cpu_vas_id);
arch/powerpc/platforms/ps3/interrupt.c
86
static DEFINE_PER_CPU(struct ps3_private, ps3_private);
arch/powerpc/platforms/ps3/smp.c
28
static DEFINE_PER_CPU(unsigned int [MSG_COUNT], ps3_ipi_virqs);
arch/powerpc/platforms/pseries/dtl.c
29
static DEFINE_PER_CPU(struct dtl, cpu_dtl);
arch/powerpc/platforms/pseries/dtl.c
57
static DEFINE_PER_CPU(struct dtl_ring, dtl_rings);
arch/powerpc/platforms/pseries/hvCall_inst.c
29
static DEFINE_PER_CPU(struct hcall_stats[HCALL_STAT_ARRAY_SIZE], hcall_stats);
arch/powerpc/platforms/pseries/iommu.c
248
static DEFINE_PER_CPU(__be64 *, tce_page);
arch/powerpc/platforms/pseries/lpar.c
174
static DEFINE_PER_CPU(struct vcpu_dispatch_data, vcpu_disp_data);
arch/powerpc/platforms/pseries/lpar.c
175
static DEFINE_PER_CPU(u64, dtl_entry_ridx);
arch/powerpc/platforms/pseries/lpar.c
176
static DEFINE_PER_CPU(struct dtl_worker, dtl_workers);
arch/powerpc/platforms/pseries/lpar.c
1840
static DEFINE_PER_CPU(unsigned int, hcall_trace_depth);
arch/powerpc/platforms/pseries/setup.c
356
DEFINE_PER_CPU(u64, idle_spurr_cycles);
arch/powerpc/platforms/pseries/setup.c
357
DEFINE_PER_CPU(u64, idle_entry_purr_snap);
arch/powerpc/platforms/pseries/setup.c
358
DEFINE_PER_CPU(u64, idle_entry_spurr_snap);
arch/powerpc/sysdev/xics/xics-common.c
37
DEFINE_PER_CPU(struct xics_cppr, xics_cppr);
arch/powerpc/sysdev/xive/common.c
82
static DEFINE_PER_CPU(struct xive_cpu *, xive_cpu);
arch/riscv/kernel/cpu.c
169
DEFINE_PER_CPU(struct riscv_cpuinfo, riscv_cpuinfo);
arch/riscv/kernel/irq.c
75
DEFINE_PER_CPU(ulong *, irq_shadow_call_stack_ptr);
arch/riscv/kernel/irq.c
90
DEFINE_PER_CPU(ulong *, irq_stack_ptr);
arch/riscv/kernel/probes/kprobes.c
19
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/riscv/kernel/probes/kprobes.c
20
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/riscv/kernel/unaligned_access_speed.c
24
DEFINE_PER_CPU(long, misaligned_access_speed) = RISCV_HWPROBE_MISALIGNED_SCALAR_UNKNOWN;
arch/riscv/kernel/unaligned_access_speed.c
25
DEFINE_PER_CPU(long, vector_misaligned_access) = RISCV_HWPROBE_MISALIGNED_VECTOR_UNSUPPORTED;
arch/riscv/kvm/aia.c
27
static DEFINE_PER_CPU(struct aia_hgei_control, aia_hgei);
arch/riscv/kvm/nacl.c
15
DEFINE_PER_CPU(struct kvm_riscv_nacl, kvm_riscv_nacl);
arch/riscv/mm/context.c
32
static DEFINE_PER_CPU(atomic_long_t, active_context);
arch/riscv/mm/context.c
33
static DEFINE_PER_CPU(unsigned long, reserved_context);
arch/s390/kernel/diag/diag.c
26
static DEFINE_PER_CPU(struct diag_stat, diag_stat);
arch/s390/kernel/idle.c
22
DEFINE_PER_CPU(struct s390_idle_data, s390_idle);
arch/s390/kernel/kprobes.c
31
DEFINE_PER_CPU(struct kprobe *, current_kprobe);
arch/s390/kernel/kprobes.c
32
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/s390/kernel/nmi.c
45
static DEFINE_PER_CPU(struct mcck_struct, cpu_mcck);
arch/s390/kernel/perf_cpum_sf.c
112
static DEFINE_PER_CPU(struct cpu_hw_sf, cpu_hw_sf);
arch/s390/kernel/processor.c
40
static DEFINE_PER_CPU(struct cpu_info, cpu_info);
arch/s390/kernel/processor.c
41
static DEFINE_PER_CPU(int, cpu_relax_retry);
arch/s390/kernel/smp.c
79
DEFINE_PER_CPU(struct pcpu, pcpu_devices);
arch/s390/kernel/time.c
271
static DEFINE_PER_CPU(atomic_t, clock_sync_word);
arch/s390/kernel/time.c
62
static DEFINE_PER_CPU(struct clock_event_device, comparators);
arch/s390/kernel/trace.c
15
static DEFINE_PER_CPU(unsigned int, diagnose_trace_depth);
arch/s390/kernel/vtime.c
31
DEFINE_PER_CPU(u64, mt_cycles[8]);
arch/s390/kernel/vtime.c
32
static DEFINE_PER_CPU(u64, mt_scaling_mult) = { 1 };
arch/s390/kernel/vtime.c
33
static DEFINE_PER_CPU(u64, mt_scaling_div) = { 1 };
arch/s390/kernel/vtime.c
34
static DEFINE_PER_CPU(u64, mt_scaling_jiffies);
arch/s390/kernel/wti.c
42
static DEFINE_PER_CPU(struct wti_state, wti_state);
arch/sh/kernel/cpu/sh2/smp-j2.c
15
DEFINE_PER_CPU(unsigned, j2_ipi_messages);
arch/sh/kernel/hw_breakpoint.c
29
static DEFINE_PER_CPU(struct perf_event *, bp_per_reg[HBP_NUM]);
arch/sh/kernel/kprobes.c
17
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/sh/kernel/kprobes.c
18
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/sh/kernel/kprobes.c
20
static DEFINE_PER_CPU(struct kprobe, saved_current_opcode);
arch/sh/kernel/kprobes.c
21
static DEFINE_PER_CPU(struct kprobe, saved_next_opcode);
arch/sh/kernel/kprobes.c
22
static DEFINE_PER_CPU(struct kprobe, saved_next_opcode2);
arch/sh/kernel/perf_event.c
34
DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events);
arch/sh/kernel/smp.c
39
DEFINE_PER_CPU(int, cpu_state) = { 0 };
arch/sh/kernel/topology.c
16
static DEFINE_PER_CPU(struct cpu, cpu_devices);
arch/sparc/kernel/cpu.c
28
DEFINE_PER_CPU(cpuinfo_sparc, __cpu_data) = { 0 };
arch/sparc/kernel/iommu-common.c
17
static DEFINE_PER_CPU(unsigned int, iommu_hash_common);
arch/sparc/kernel/kprobes.c
45
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/sparc/kernel/kprobes.c
46
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/sparc/kernel/nmi.c
48
static DEFINE_PER_CPU(short, wd_enabled);
arch/sparc/kernel/nmi.c
51
static DEFINE_PER_CPU(unsigned int, last_irq_sum);
arch/sparc/kernel/nmi.c
52
static DEFINE_PER_CPU(long, alert_counter);
arch/sparc/kernel/nmi.c
53
static DEFINE_PER_CPU(int, nmi_touch);
arch/sparc/kernel/pci_sun4v.c
64
static DEFINE_PER_CPU(struct iommu_batch, iommu_batch);
arch/sparc/kernel/perf_event.c
115
static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { .enabled = 1, };
arch/sparc/kernel/smp_64.c
61
DEFINE_PER_CPU(cpumask_t, cpu_sibling_map) = CPU_MASK_NONE;
arch/sparc/kernel/smp_64.c
78
static DEFINE_PER_CPU(bool, poke);
arch/sparc/kernel/sysfs.c
17
static DEFINE_PER_CPU(struct hv_mmu_statistics, mmu_stats) __attribute__((aligned(64)));
arch/sparc/kernel/sysfs.c
206
static DEFINE_PER_CPU(struct cpu, cpu_devices);
arch/sparc/kernel/time_32.c
61
DEFINE_PER_CPU(struct clock_event_device, sparc32_clockevent);
arch/sparc/kernel/time_64.c
640
static DEFINE_PER_CPU(struct freq_table, sparc64_freq_table) = { 0, 0 };
arch/sparc/kernel/time_64.c
716
static DEFINE_PER_CPU(struct clock_event_device, sparc64_events);
arch/sparc/mm/init_64.c
758
DEFINE_PER_CPU(struct mm_struct *, per_cpu_secondary_mm) = {0};
arch/sparc/mm/tlb.c
23
static DEFINE_PER_CPU(struct tlb_batch, tlb_batch);
arch/x86/coco/sev/core.c
104
DEFINE_PER_CPU(struct sev_es_runtime_data*, runtime_data);
arch/x86/coco/sev/core.c
105
DEFINE_PER_CPU(struct sev_es_save_area *, sev_vmsa);
arch/x86/coco/sev/svsm.c
27
DEFINE_PER_CPU(struct svsm_ca *, svsm_caa);
arch/x86/coco/sev/svsm.c
28
DEFINE_PER_CPU(u64, svsm_caa_pa);
arch/x86/events/amd/core.c
18
static DEFINE_PER_CPU(unsigned long, perf_nmi_tstamp);
arch/x86/events/core.c
1398
DEFINE_PER_CPU(u64 [X86_PMC_IDX_MAX], pmc_prev_left);
arch/x86/events/core.c
55
DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = {
arch/x86/events/core.c
60
static DEFINE_PER_CPU(bool, guest_lvtpc_loaded);
arch/x86/events/intel/ds.c
796
static DEFINE_PER_CPU(void *, insn_buffer);
arch/x86/events/intel/p4.c
988
static DEFINE_PER_CPU(unsigned long [BITS_TO_LONGS(X86_PMC_IDX_MAX)], p4_running);
arch/x86/events/intel/pt.c
33
static DEFINE_PER_CPU(struct pt, pt_ctx);
arch/x86/hyperv/ivm.c
252
static DEFINE_PER_CPU(struct sev_es_save_area *, hv_sev_vmsa);
arch/x86/include/asm/percpu.h
614
DEFINE_PER_CPU(_type, _name) = _initvalue; \
arch/x86/include/asm/percpu.h
648
DEFINE_PER_CPU(_type, _name) = _initvalue
arch/x86/kernel/alternative.c
2787
static DEFINE_PER_CPU(atomic_t, text_poke_array_refs);
arch/x86/kernel/apic/apic.c
512
static DEFINE_PER_CPU(struct clock_event_device, lapic_events);
arch/x86/kernel/apic/vector.c
55
static DEFINE_PER_CPU(struct vector_cleanup, vector_cleanup) = {
arch/x86/kernel/apic/x2apic_cluster.c
21
static DEFINE_PER_CPU(cpumask_var_t, ipi_mask);
arch/x86/kernel/apic/x2apic_uv_x.c
506
DEFINE_PER_CPU(struct uv_cpu_info_s, __uv_cpu_info);
arch/x86/kernel/callthunks.c
42
DEFINE_PER_CPU(u64, __x86_call_count);
arch/x86/kernel/callthunks.c
43
DEFINE_PER_CPU(u64, __x86_ret_count);
arch/x86/kernel/callthunks.c
44
DEFINE_PER_CPU(u64, __x86_stuffs_count);
arch/x86/kernel/callthunks.c
45
DEFINE_PER_CPU(u64, __x86_ctxsw_count);
arch/x86/kernel/cpu/aperfmperf.c
354
DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
arch/x86/kernel/cpu/bugs.c
60
DEFINE_PER_CPU(u64, x86_spec_ctrl_current);
arch/x86/kernel/cpu/bugs.c
68
DEFINE_PER_CPU(bool, x86_ibpb_exit_to_user);
arch/x86/kernel/cpu/bus_lock.c
203
static DEFINE_PER_CPU(struct delayed_work, sl_reenable);
arch/x86/kernel/cpu/common.c
756
DEFINE_PER_CPU(struct cpu_entry_area *, cpu_entry_area);
arch/x86/kernel/cpu/intel_epb.c
56
static DEFINE_PER_CPU(u8, saved_epb);
arch/x86/kernel/cpu/mce/amd.c
258
static DEFINE_PER_CPU(struct threshold_bank **, threshold_banks);
arch/x86/kernel/cpu/mce/amd.c
264
static DEFINE_PER_CPU(u64, bank_map);
arch/x86/kernel/cpu/mce/core.c
150
DEFINE_PER_CPU(struct mce, injectm);
arch/x86/kernel/cpu/mce/core.c
1763
static DEFINE_PER_CPU(unsigned long, mce_next_interval); /* in jiffies */
arch/x86/kernel/cpu/mce/core.c
1764
static DEFINE_PER_CPU(struct timer_list, mce_timer);
arch/x86/kernel/cpu/mce/core.c
2525
DEFINE_PER_CPU(struct device *, mce_device);
arch/x86/kernel/cpu/mce/core.c
70
DEFINE_PER_CPU(unsigned, mce_exception_count);
arch/x86/kernel/cpu/mce/core.c
719
DEFINE_PER_CPU(unsigned, mce_poll_count);
arch/x86/kernel/cpu/mce/core.c
92
static DEFINE_PER_CPU(struct mce_hw_err, hw_errs_seen);
arch/x86/kernel/cpu/mce/core.c
99
DEFINE_PER_CPU(mce_banks_t, mce_poll_banks) = {
arch/x86/kernel/cpu/mce/intel.c
42
static DEFINE_PER_CPU(mce_banks_t, mce_banks_owned);
arch/x86/kernel/cpu/mce/threshold.c
46
DEFINE_PER_CPU(struct mca_storm_desc, storm_desc);
arch/x86/kernel/cpu/microcode/core.c
328
static DEFINE_PER_CPU(struct microcode_ctrl, ucode_ctrl);
arch/x86/kernel/cpu/resctrl/core.c
45
DEFINE_PER_CPU(struct resctrl_pqr_state, pqr_state);
arch/x86/kernel/fpu/core.c
35
DEFINE_PER_CPU(u64, xfd_state);
arch/x86/kernel/fpu/core.c
53
DEFINE_PER_CPU(bool, kernel_fpu_allowed);
arch/x86/kernel/fpu/core.c
58
DEFINE_PER_CPU(struct fpu *, fpu_fpregs_owner_ctx);
arch/x86/kernel/fred.c
25
DEFINE_PER_CPU(unsigned long, fred_rsp0);
arch/x86/kernel/hpet.c
58
static DEFINE_PER_CPU(struct hpet_channel *, cpu_hpet_channel);
arch/x86/kernel/hw_breakpoint.c
40
DEFINE_PER_CPU(unsigned long, cpu_dr7);
arch/x86/kernel/hw_breakpoint.c
44
static DEFINE_PER_CPU(unsigned long, cpu_debugreg[HBP_NUM]);
arch/x86/kernel/hw_breakpoint.c
50
static DEFINE_PER_CPU(struct perf_event *, bp_per_reg[HBP_NUM]);
arch/x86/kernel/irqinit.c
50
DEFINE_PER_CPU(vector_irq_t, vector_irq) = {
arch/x86/kernel/kprobes/core.c
61
DEFINE_PER_CPU(struct kprobe *, current_kprobe) = NULL;
arch/x86/kernel/kprobes/core.c
62
DEFINE_PER_CPU(struct kprobe_ctlblk, kprobe_ctlblk);
arch/x86/kernel/kvm.c
490
static DEFINE_PER_CPU(cpumask_var_t, __pv_cpu_mask);
arch/x86/kernel/kvmclock.c
54
DEFINE_PER_CPU(struct pvclock_vsyscall_time_info *, hv_clock_per_cpu);
arch/x86/kernel/nmi.c
353
static DEFINE_PER_CPU(bool, swallow_nmi);
arch/x86/kernel/nmi.c
354
static DEFINE_PER_CPU(unsigned long, last_nmi_rip);
arch/x86/kernel/nmi.c
532
static DEFINE_PER_CPU(enum nmi_states, nmi_state);
arch/x86/kernel/nmi.c
533
static DEFINE_PER_CPU(unsigned long, nmi_cr2);
arch/x86/kernel/nmi.c
534
static DEFINE_PER_CPU(unsigned long, nmi_dr7);
arch/x86/kernel/nmi.c
83
static DEFINE_PER_CPU(struct nmi_stats, nmi_stats);
arch/x86/kernel/perf_regs.c
131
static DEFINE_PER_CPU(struct pt_regs, nmi_user_regs);
arch/x86/kernel/process.c
343
DEFINE_PER_CPU(u64, msr_misc_features_shadow);
arch/x86/kernel/process.c
525
static DEFINE_PER_CPU(struct ssb_state, ssb_state);
arch/x86/kernel/process.c
88
DEFINE_PER_CPU(bool, __tss_limit_invalid);
arch/x86/kernel/process.c
99
DEFINE_PER_CPU(bool, cache_state_incoherent);
arch/x86/kernel/tsc_sync.c
34
static DEFINE_PER_CPU(struct tsc_adjust, tsc_adjust);
arch/x86/kvm/svm/svm.c
181
DEFINE_PER_CPU(struct svm_cpu_data, svm_data);
arch/x86/kvm/svm/svm.c
81
static DEFINE_PER_CPU(u64, current_tsc_ratio);
arch/x86/kvm/vmx/posted_intr.c
26
static DEFINE_PER_CPU(struct list_head, wakeup_vcpus_on_cpu);
arch/x86/kvm/vmx/posted_intr.c
34
static DEFINE_PER_CPU(raw_spinlock_t, wakeup_vcpus_on_cpu_lock);
arch/x86/kvm/vmx/tdx.c
242
static DEFINE_PER_CPU(struct list_head, associated_tdvcpus);
arch/x86/kvm/vmx/vmx.c
582
static DEFINE_PER_CPU(struct vmcs *, vmxarea);
arch/x86/kvm/vmx/vmx.c
583
DEFINE_PER_CPU(struct vmcs *, current_vmcs);
arch/x86/kvm/vmx/vmx.c
588
static DEFINE_PER_CPU(struct list_head, loaded_vmcss_on_cpu);
arch/x86/kvm/x86.c
218
static DEFINE_PER_CPU(struct kvm_user_return_msrs, user_return_msrs);
arch/x86/kvm/x86.c
2513
static DEFINE_PER_CPU(unsigned long, cpu_tsc_khz);
arch/x86/kvm/x86.c
5161
static DEFINE_PER_CPU(struct kvm_vcpu *, last_vcpu);
arch/x86/mm/cpu_entry_area.c
19
DEFINE_PER_CPU(struct cea_exception_stacks*, cea_exception_stacks);
arch/x86/mm/init.c
1020
DEFINE_PER_CPU(u64, tlbstate_untag_mask);
arch/x86/mm/kmmio.c
93
static DEFINE_PER_CPU(struct kmmio_context, kmmio_ctx);
arch/x86/mm/kmsan_shadow.c
19
DEFINE_PER_CPU(char[CPU_ENTRY_AREA_SIZE], cpu_entry_area_shadow);
arch/x86/mm/kmsan_shadow.c
20
DEFINE_PER_CPU(char[CPU_ENTRY_AREA_SIZE], cpu_entry_area_origin);
arch/x86/mm/mmio-mod.c
42
static DEFINE_PER_CPU(struct trap_reason, pf_reason);
arch/x86/mm/mmio-mod.c
43
static DEFINE_PER_CPU(struct mmiotrace_rw, cpu_trace);
arch/x86/mm/tlb.c
1400
static DEFINE_PER_CPU(unsigned int, flush_tlb_info_idx);
arch/x86/platform/uv/uv_nmi.c
58
DEFINE_PER_CPU(struct uv_cpu_nmi_s, uv_cpu_nmi);
arch/x86/platform/uv/uv_time.c
44
static DEFINE_PER_CPU(struct clock_event_device, cpu_ced);
arch/x86/virt/vmx/tdx/tdx.c
50
static DEFINE_PER_CPU(bool, tdx_lp_initialized);
arch/x86/xen/enlighten.c
38
DEFINE_PER_CPU(struct vcpu_info *, xen_vcpu);
arch/x86/xen/enlighten.c
42
DEFINE_PER_CPU(uint32_t, xen_vcpu_id);
arch/x86/xen/enlighten_pv.c
1052
static DEFINE_PER_CPU(unsigned long, xen_cr0_value);
arch/x86/xen/enlighten_pv.c
110
DEFINE_PER_CPU(bool, xen_in_preemptible_hcall);
arch/x86/xen/enlighten_pv.c
141
DEFINE_PER_CPU(enum xen_lazy_mode, xen_lazy_mode) = XEN_LAZY_NONE;
arch/x86/xen/enlighten_pv.c
158
static DEFINE_PER_CPU(struct tls_descs, shadow_tls_desc);
arch/x86/xen/enlighten_pv.c
871
static DEFINE_PER_CPU(struct desc_ptr, idt_desc);
arch/x86/xen/mmu_pv.c
176
DEFINE_PER_CPU(unsigned long, xen_cr3); /* cr3 stored as physaddr */
arch/x86/xen/mmu_pv.c
177
static DEFINE_PER_CPU(unsigned long, xen_current_cr3); /* actual vcpu cr3 */
arch/x86/xen/multicalls.c
55
static DEFINE_PER_CPU(struct mc_buffer, mc_buffer);
arch/x86/xen/multicalls.c
58
DEFINE_PER_CPU(unsigned long, xen_mc_irq_flags);
arch/x86/xen/pmu.c
25
static DEFINE_PER_CPU(struct xenpmu, xenpmu_shared);
arch/x86/xen/smp.c
13
static DEFINE_PER_CPU(struct xen_common_irq, xen_resched_irq) = { .irq = -1 };
arch/x86/xen/smp.c
14
static DEFINE_PER_CPU(struct xen_common_irq, xen_callfunc_irq) = { .irq = -1 };
arch/x86/xen/smp.c
15
static DEFINE_PER_CPU(struct xen_common_irq, xen_callfuncsingle_irq) = { .irq = -1 };
arch/x86/xen/smp.c
16
static DEFINE_PER_CPU(struct xen_common_irq, xen_debug_irq) = { .irq = -1 };
arch/x86/xen/smp_pv.c
52
static DEFINE_PER_CPU(struct xen_common_irq, xen_irq_work) = { .irq = -1 };
arch/x86/xen/smp_pv.c
53
static DEFINE_PER_CPU(struct xen_common_irq, xen_pmu_irq) = { .irq = -1 };
arch/x86/xen/spinlock.c
17
static DEFINE_PER_CPU(int, lock_kicker_irq) = -1;
arch/x86/xen/spinlock.c
18
static DEFINE_PER_CPU(char *, irq_name);
arch/x86/xen/spinlock.c
19
static DEFINE_PER_CPU(atomic_t, xen_qlock_wait_nest);
arch/x86/xen/suspend.c
20
static DEFINE_PER_CPU(u64, spec_ctrl);
arch/x86/xen/time.c
319
static DEFINE_PER_CPU(struct xen_clock_event_device, xen_clock_events) = { .evt.irq = -1 };
arch/xtensa/kernel/hw_breakpoint.c
19
static DEFINE_PER_CPU(struct perf_event *, bp_on_reg[XCHAL_NUM_IBREAK]);
arch/xtensa/kernel/hw_breakpoint.c
22
static DEFINE_PER_CPU(struct perf_event *, wp_on_reg[XCHAL_NUM_DBREAK]);
arch/xtensa/kernel/perf_event.c
65
static DEFINE_PER_CPU(struct xtensa_pmu_events, xtensa_pmu_events);
arch/xtensa/kernel/setup.c
374
static DEFINE_PER_CPU(struct cpu, cpu_data);
arch/xtensa/kernel/smp.c
371
static DEFINE_PER_CPU(struct ipi_data, ipi_data);
arch/xtensa/kernel/time.c
107
static DEFINE_PER_CPU(struct ccount_timer, ccount_timer) = {
arch/xtensa/kernel/traps.c
168
DEFINE_PER_CPU(struct exc_table, exc_table);
arch/xtensa/kernel/traps.c
169
DEFINE_PER_CPU(struct debug_table, debug_table);
arch/xtensa/kernel/traps.c
259
DEFINE_PER_CPU(unsigned long, nmi_count);
arch/xtensa/mm/mmu.c
21
DEFINE_PER_CPU(unsigned long, asid_cache) = ASID_USER_FIRST;
block/blk-mq.c
45
static DEFINE_PER_CPU(struct llist_head, blk_cpu_done);
block/blk-mq.c
46
static DEFINE_PER_CPU(call_single_data_t, blk_cpu_csd);
crypto/scompress.c
36
static DEFINE_PER_CPU(struct scomp_scratch, scomp_scratch) = {
drivers/accessibility/speakup/fakekey.c
18
static DEFINE_PER_CPU(int, reporting_keystroke);
drivers/acpi/acpi_processor.c
200
static DEFINE_PER_CPU(void *, processor_device_array);
drivers/acpi/acpi_processor.c
31
DEFINE_PER_CPU(struct acpi_processor *, processors);
drivers/acpi/cppc_acpi.c
85
static DEFINE_PER_CPU(int, cpu_pcc_subspace_idx);
drivers/acpi/cppc_acpi.c
94
static DEFINE_PER_CPU(struct cpc_desc *, cpc_desc_ptr);
drivers/acpi/processor_idle.c
52
static DEFINE_PER_CPU(struct cpuidle_device *, acpi_cpuidle_device);
drivers/acpi/processor_idle.c
67
DEFINE_PER_CPU(struct acpi_processor_cx * [CPUIDLE_STATE_MAX], acpi_cstate);
drivers/acpi/processor_thermal.c
42
static DEFINE_PER_CPU(unsigned int, cpufreq_thermal_reduction_step);
drivers/base/arch_topology.c
138
DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE;
drivers/base/arch_topology.c
164
DEFINE_PER_CPU(unsigned long, hw_pressure);
drivers/base/arch_topology.c
29
static DEFINE_PER_CPU(struct scale_freq_data __rcu *, sft_data);
drivers/base/arch_topology.c
32
DEFINE_PER_CPU(unsigned long, capacity_freq_ref) = 0;
drivers/base/cacheinfo.c
25
static DEFINE_PER_CPU(struct cpu_cacheinfo, ci_cpu_cacheinfo);
drivers/base/cacheinfo.c
638
static DEFINE_PER_CPU(struct device *, ci_cache_dev);
drivers/base/cacheinfo.c
644
static DEFINE_PER_CPU(struct device **, ci_index_dev);
drivers/base/cpu.c
27
static DEFINE_PER_CPU(struct device *, cpu_sys_devices);
drivers/base/cpu.c
556
DEFINE_PER_CPU(struct cpu, cpu_devices);
drivers/base/topology.c
212
DEFINE_PER_CPU(unsigned long, cpu_scale) = SCHED_CAPACITY_SCALE;
drivers/bus/mips_cdmm.c
274
static DEFINE_PER_CPU(struct mips_cdmm_bus *, mips_cdmm_buses);
drivers/char/random.c
1017
static DEFINE_PER_CPU(struct fast_pool, irq_randomness) = {
drivers/char/random.c
218
static DEFINE_PER_CPU(struct crng, crngs) = {
drivers/char/random.c
506
static DEFINE_PER_CPU(struct batch_ ##type, batched_entropy_ ##type) = { \
drivers/clocksource/arc_timer.c
269
static DEFINE_PER_CPU(struct clock_event_device, arc_clockevent_device) = {
drivers/clocksource/arm_arch_timer.c
310
DEFINE_PER_CPU(const struct arch_timer_erratum_workaround *, timer_unstable_counter_workaround);
drivers/clocksource/arm_arch_timer.c
970
static DEFINE_PER_CPU(unsigned long, saved_cntkctl);
drivers/clocksource/dummy_timer.c
14
static DEFINE_PER_CPU(struct clock_event_device, dummy_timer_evt);
drivers/clocksource/exynos_mct.c
358
static DEFINE_PER_CPU(struct mct_clock_event_device, percpu_mct_tick);
drivers/clocksource/hyperv_timer.c
54
static __maybe_unused DEFINE_PER_CPU(long, stimer0_evt);
drivers/clocksource/ingenic-timer.c
25
static DEFINE_PER_CPU(call_single_data_t, ingenic_cevt_csd);
drivers/clocksource/mips-gic-timer.c
19
static DEFINE_PER_CPU(struct clock_event_device, gic_clockevent_device);
drivers/clocksource/numachip.c
13
static DEFINE_PER_CPU(struct clock_event_device, numachip2_ced);
drivers/clocksource/timer-clint.c
122
static DEFINE_PER_CPU(struct clock_event_device, clint_clock_event) = {
drivers/clocksource/timer-econet-en751221.c
30
static DEFINE_PER_CPU(struct clock_event_device, econet_timer_pcpu);
drivers/clocksource/timer-mp-csky.c
49
static DEFINE_PER_CPU(struct timer_of, csky_to) = {
drivers/clocksource/timer-nxp-pit.c
51
static DEFINE_PER_CPU(struct pit_timer *, pit_timers);
drivers/clocksource/timer-nxp-stm.c
65
static DEFINE_PER_CPU(struct stm_timer *, stm_timers);
drivers/clocksource/timer-riscv.c
72
static DEFINE_PER_CPU(struct clock_event_device, riscv_clock_event) = {
drivers/clocksource/timer-rtl-otto.c
200
static DEFINE_PER_CPU(struct timer_of, rttm_to) = {
drivers/clocksource/timer-tegra.c
116
static DEFINE_PER_CPU(struct timer_of, tegra_to) = {
drivers/clocksource/timer-ti-dm-systimer.c
634
static DEFINE_PER_CPU(struct dmtimer_clockevent, dmtimer_percpu_timer);
drivers/connector/cn_proc.c
47
static DEFINE_PER_CPU(struct local_event, local_event) = {
drivers/cpufreq/amd_freq_sensitivity.c
37
static DEFINE_PER_CPU(struct cpu_data_t, cpu_data);
drivers/cpufreq/cppc_cpufreq.c
385
static DEFINE_PER_CPU(unsigned int, efficiency_class);
drivers/cpufreq/cppc_cpufreq.c
50
static DEFINE_PER_CPU(struct cppc_freq_invariance, cppc_freq_inv);
drivers/cpufreq/cpufreq.c
2575
DEFINE_PER_CPU(unsigned long, cpufreq_pressure);
drivers/cpufreq/cpufreq.c
60
static DEFINE_PER_CPU(struct cpufreq_policy *, cpufreq_cpu_data);
drivers/cpufreq/cpufreq_governor.c
24
static DEFINE_PER_CPU(struct cpu_dbs_info, cpu_dbs);
drivers/cpufreq/loongson3_cpufreq.c
174
static DEFINE_PER_CPU(struct loongson3_freq_data *, freq_data);
drivers/cpufreq/powernow-k8.c
53
static DEFINE_PER_CPU(struct powernow_k8_data *, powernow_data);
drivers/cpufreq/powernv-cpufreq.c
147
static DEFINE_PER_CPU(struct chip *, chip_info);
drivers/cpufreq/sh-cpufreq.c
30
static DEFINE_PER_CPU(struct clk, sh_cpuclk);
drivers/cpufreq/speedstep-centrino.c
75
static DEFINE_PER_CPU(struct cpu_model *, centrino_model);
drivers/cpufreq/speedstep-centrino.c
76
static DEFINE_PER_CPU(const struct cpu_id *, centrino_cpu);
drivers/cpufreq/vexpress-spc-cpufreq.c
55
static DEFINE_PER_CPU(unsigned int, physical_cluster);
drivers/cpufreq/vexpress-spc-cpufreq.c
56
static DEFINE_PER_CPU(unsigned int, cpu_last_req_freq);
drivers/cpufreq/virtual-cpufreq.c
68
static DEFINE_PER_CPU(u32, perftbl_num_entries);
drivers/cpuidle/coupled.c
114
static DEFINE_PER_CPU(call_single_data_t, cpuidle_coupled_poke_cb);
drivers/cpuidle/cpuidle-psci.c
46
static DEFINE_PER_CPU(struct psci_cpuidle_domain_state, psci_domain_state);
drivers/cpuidle/cpuidle-riscv-sbi.c
45
static DEFINE_PER_CPU(struct sbi_domain_state, domain_state);
drivers/cpuidle/cpuidle.c
33
DEFINE_PER_CPU(struct cpuidle_device *, cpuidle_devices);
drivers/cpuidle/cpuidle.c
34
DEFINE_PER_CPU(struct cpuidle_device, cpuidle_dev);
drivers/cpuidle/driver.c
29
static DEFINE_PER_CPU(struct cpuidle_driver *, cpuidle_drivers);
drivers/cpuidle/governors/ladder.c
43
static DEFINE_PER_CPU(struct ladder_device, ladder_devices);
drivers/cpuidle/governors/menu.c
98
static DEFINE_PER_CPU(struct menu_device, menu_devices);
drivers/cpuidle/governors/teo.c
155
static DEFINE_PER_CPU(struct teo_cpu, teo_cpus);
drivers/crypto/caam/qi.c
64
static DEFINE_PER_CPU(struct caam_qi_pcpu_priv, pcpu_qipriv);
drivers/crypto/caam/qi.c
65
static DEFINE_PER_CPU(int, last_cpu);
drivers/crypto/nx/nx-common-powernv.c
59
static DEFINE_PER_CPU(struct vas_window *, cpu_txwin);
drivers/crypto/padlock-aes.c
72
static DEFINE_PER_CPU(struct cword *, paes_last_cword);
drivers/dma/fsl-qdma.c
181
static DEFINE_PER_CPU(struct fsl_pre_status, pre);
drivers/gpu/drm/amd/display/amdgpu_dm/dc_fpu.c
42
static DEFINE_PER_CPU(int, fpu_recursion_depth);
drivers/hv/mshv_synic.c
691
static DEFINE_PER_CPU(long, mshv_evt);
drivers/hv/mshv_vtl_main.c
102
static DEFINE_PER_CPU(struct mshv_vtl_poll_file, mshv_vtl_poll_file);
drivers/hv/mshv_vtl_main.c
103
static DEFINE_PER_CPU(unsigned long long, num_vtl0_transitions);
drivers/hv/mshv_vtl_main.c
104
static DEFINE_PER_CPU(struct mshv_vtl_per_cpu, mshv_vtl_per_cpu);
drivers/hv/vmbus_drv.c
1368
static DEFINE_PER_CPU(bool, vmbus_irq_pending);
drivers/hv/vmbus_drv.c
1369
static DEFINE_PER_CPU(struct task_struct *, vmbus_irqd);
drivers/hv/vmbus_drv.c
55
static DEFINE_PER_CPU(long, vmbus_evt);
drivers/hwtracing/coresight/coresight-core.c
36
static DEFINE_PER_CPU(struct coresight_device *, csdev_sink);
drivers/hwtracing/coresight/coresight-cpu-debug.c
107
static DEFINE_PER_CPU(struct debug_drvdata *, debug_drvdata);
drivers/hwtracing/coresight/coresight-etm-perf.c
51
static DEFINE_PER_CPU(struct etm_ctxt, etm_ctxt);
drivers/hwtracing/coresight/coresight-etm-perf.c
52
static DEFINE_PER_CPU(struct coresight_device *, csdev_src);
drivers/hwtracing/coresight/coresight-etm4x-core.c
77
static DEFINE_PER_CPU(struct etm4_init_arg *, delayed_probe);
drivers/hwtracing/coresight/coresight-sysfs.c
26
static DEFINE_PER_CPU(struct coresight_path *, tracer_path);
drivers/hwtracing/coresight/coresight-trace-id.c
22
static DEFINE_PER_CPU(atomic_t, id_map_default_cpu_ids) = ATOMIC_INIT(0);
drivers/infiniband/sw/siw/siw_main.c
104
static DEFINE_PER_CPU(atomic_t, siw_use_cnt);
drivers/infiniband/sw/siw/siw_qp_tx.c
1201
static DEFINE_PER_CPU(struct tx_task_t, siw_tx_task_g);
drivers/irqchip/irq-aclint-sswi.c
20
static DEFINE_PER_CPU(void __iomem *, sswi_cpu_regs);
drivers/irqchip/irq-apple-aic.c
332
static DEFINE_PER_CPU(uint32_t, aic_fiq_unmasked);
drivers/irqchip/irq-csky-mpintc.c
42
static DEFINE_PER_CPU(void __iomem *, intcl_reg);
drivers/irqchip/irq-gic-v3-its.c
129
static DEFINE_PER_CPU(struct its_node *, local_4_1_its);
drivers/irqchip/irq-gic-v3-its.c
193
static DEFINE_PER_CPU(struct cpu_lpi_count, cpu_lpi_count);
drivers/irqchip/irq-gic-v3.c
235
static DEFINE_PER_CPU(bool, has_rss);
drivers/irqchip/irq-gic-v5-irs.c
373
static DEFINE_PER_CPU(struct iaffid_entry, cpu_iaffid);
drivers/irqchip/irq-gic.c
127
static DEFINE_PER_CPU(u32, sgi_intid);
drivers/irqchip/irq-loongarch-avec.c
35
static DEFINE_PER_CPU(struct pending_list, pending_list);
drivers/irqchip/irq-loongarch-avec.c
38
static DEFINE_PER_CPU(struct irq_desc * [NR_VECTORS], irq_map);
drivers/irqchip/irq-ompic.c
83
DEFINE_PER_CPU(unsigned long, ops);
drivers/irqchip/irq-riscv-aplic-direct.c
41
static DEFINE_PER_CPU(struct aplic_idc, aplic_idcs);
drivers/irqchip/irq-sifive-plic.c
104
static DEFINE_PER_CPU(struct plic_handler, plic_handlers);
drivers/irqchip/irq-xtensa-mx.c
24
static DEFINE_PER_CPU(unsigned int, cached_irq_mask);
drivers/leds/trigger/ledtrig-cpu.c
39
static DEFINE_PER_CPU(struct led_trigger_cpu, cpu_trig);
drivers/mailbox/riscv-sbi-mpxy-mbox.c
124
static DEFINE_PER_CPU(struct mpxy_local, mpxy_local);
drivers/mailbox/zynqmp-ipi-mailbox.c
143
static DEFINE_PER_CPU(struct zynqmp_ipi_pdata *, per_cpu_pdata);
drivers/md/dm-vdo/funnel-workqueue.c
25
static DEFINE_PER_CPU(unsigned int, service_queue_rotor);
drivers/nvdimm/region_devs.c
24
static DEFINE_PER_CPU(int, flush_idx);
drivers/perf/arm_pmu.c
107
static DEFINE_PER_CPU(int, cpu_irq);
drivers/perf/arm_pmu.c
108
static DEFINE_PER_CPU(const struct pmu_irq_ops *, cpu_irq_ops);
drivers/perf/arm_pmu_acpi.c
19
static DEFINE_PER_CPU(struct arm_pmu *, probed_pmus);
drivers/perf/arm_pmu_acpi.c
20
static DEFINE_PER_CPU(int, pmu_irqs);
drivers/platform/x86/amd/hfi/hfi.c
113
static DEFINE_PER_CPU(struct amd_hfi_cpuinfo, amd_hfi_cpuinfo) = {.class_index = -1};
drivers/platform/x86/intel/pmc/cnp.c
224
static DEFINE_PER_CPU(u64, pkg_cst_config);
drivers/platform/x86/intel/tpmi_power_domains.c
68
static DEFINE_PER_CPU(struct tpmi_cpu_info, tpmi_cpu_info);
drivers/powercap/dtpm_cpu.c
34
static DEFINE_PER_CPU(struct dtpm_cpu *, dtpm_per_cpu);
drivers/powercap/idle_inject.c
92
static DEFINE_PER_CPU(struct idle_inject_thread, idle_inject_thread);
drivers/powercap/idle_inject.c
93
static DEFINE_PER_CPU(struct idle_inject_device *, idle_inject_device);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
25
DEFINE_PER_CPU(struct bnx2fc_percpu_s, bnx2fc_percpu);
drivers/scsi/bnx2i/bnx2i_init.c
71
DEFINE_PER_CPU(struct bnx2i_percpu_s, bnx2i_percpu);
drivers/scsi/fcoe/fcoe.c
69
static DEFINE_PER_CPU(struct fcoe_percpu_s, fcoe_percpu);
drivers/scsi/qedi/qedi_main.c
60
static DEFINE_PER_CPU(struct qedi_percpu_s, qedi_percpu);
drivers/soc/fsl/qbman/bman.c
228
static DEFINE_PER_CPU(struct bman_portal, bman_affine_portal);
drivers/soc/fsl/qbman/qman.c
1003
static DEFINE_PER_CPU(struct qman_portal, qman_affine_portal);
drivers/soc/fsl/qbman/qman_test_stash.c
172
static DEFINE_PER_CPU(struct hp_cpu, hp_cpus);
drivers/thermal/intel/intel_hfi.c
159
static DEFINE_PER_CPU(struct hfi_cpu_info, hfi_cpu_info) = { .index = -1 };
drivers/thermal/intel/therm_throt.c
124
static DEFINE_PER_CPU(struct thermal_state, thermal_state);
drivers/xen/events/events_2l.c
139
static DEFINE_PER_CPU(unsigned int, current_word_idx);
drivers/xen/events/events_2l.c
140
static DEFINE_PER_CPU(unsigned int, current_bit_idx);
drivers/xen/events/events_2l.c
43
static DEFINE_PER_CPU(xen_ulong_t [EVTCHN_MASK_SIZE], cpu_evtchn_mask);
drivers/xen/events/events_base.c
163
static DEFINE_PER_CPU(int [NR_VIRQS], virq_to_irq) = {[0 ... NR_VIRQS-1] = -1};
drivers/xen/events/events_base.c
166
static DEFINE_PER_CPU(int [XEN_NR_IPIS], ipi_to_irq) = {[0 ... XEN_NR_IPIS-1] = -1};
drivers/xen/events/events_base.c
168
static DEFINE_PER_CPU(evtchn_port_t [XEN_NR_IPIS], ipi_to_evtchn) = {[0 ... XEN_NR_IPIS-1] = 0};
drivers/xen/events/events_base.c
194
static DEFINE_PER_CPU(unsigned int, irq_epoch);
drivers/xen/events/events_base.c
553
static DEFINE_PER_CPU(struct lateeoi_work, lateeoi);
drivers/xen/events/events_fifo.c
64
static DEFINE_PER_CPU(struct evtchn_fifo_control_block *, cpu_control_block);
drivers/xen/events/events_fifo.c
65
static DEFINE_PER_CPU(struct evtchn_fifo_queue, cpu_queue);
drivers/xen/time.c
23
static DEFINE_PER_CPU(struct vcpu_runstate_info, xen_runstate);
drivers/xen/time.c
25
static DEFINE_PER_CPU(u64[4], old_runstate_time);
fs/buffer.c
1298
static DEFINE_PER_CPU(struct bh_lru, bh_lrus) = {{ NULL }};
fs/buffer.c
3009
static DEFINE_PER_CPU(struct bh_accounting, bh_accounting) = {0, 0};
fs/dcache.c
142
static DEFINE_PER_CPU(long, nr_dentry);
fs/dcache.c
143
static DEFINE_PER_CPU(long, nr_dentry_unused);
fs/dcache.c
144
static DEFINE_PER_CPU(long, nr_dentry_negative);
fs/ext4/mballoc.c
453
static DEFINE_PER_CPU(u64, discard_pa_seq);
fs/inode.c
108
static DEFINE_PER_CPU(long, mg_ctime_updates);
fs/inode.c
109
static DEFINE_PER_CPU(long, mg_fine_stamps);
fs/inode.c
110
static DEFINE_PER_CPU(long, mg_ctime_swaps);
fs/inode.c
1143
static DEFINE_PER_CPU(unsigned int, last_ino);
fs/inode.c
77
static DEFINE_PER_CPU(unsigned long, nr_inodes);
fs/inode.c
78
static DEFINE_PER_CPU(unsigned long, nr_unused);
fs/locks.c
139
static DEFINE_PER_CPU(struct file_lock_list_struct, file_lock_list);
fs/nfsd/filecache.c
58
static DEFINE_PER_CPU(unsigned long, nfsd_file_cache_hits);
fs/nfsd/filecache.c
59
static DEFINE_PER_CPU(unsigned long, nfsd_file_acquisitions);
fs/nfsd/filecache.c
60
static DEFINE_PER_CPU(unsigned long, nfsd_file_allocations);
fs/nfsd/filecache.c
61
static DEFINE_PER_CPU(unsigned long, nfsd_file_releases);
fs/nfsd/filecache.c
62
static DEFINE_PER_CPU(unsigned long, nfsd_file_total_age);
fs/nfsd/filecache.c
63
static DEFINE_PER_CPU(unsigned long, nfsd_file_evictions);
include/linux/alloc_tag.h
118
static DEFINE_PER_CPU(struct alloc_tag_counters, _alloc_tag_cntr); \
include/linux/cookie.h
23
static DEFINE_PER_CPU(struct pcpu_gen_cookie, __##name); \
include/linux/notifier.h
127
static DEFINE_PER_CPU(struct srcu_data, name##_head_srcu_data); \
include/linux/percpu-defs.h
187
#define DEFINE_PER_CPU_DECRYPTED(type, name) DEFINE_PER_CPU(type, name)
include/linux/percpu-rwsem.h
31
static DEFINE_PER_CPU(unsigned int, __percpu_rwsem_rc_##name); \
include/linux/srcutree.h
223
static DEFINE_PER_CPU(struct srcu_data, name##_srcu_data); \
include/linux/torture.h
76
DEFINE_PER_CPU(struct torture_random_state, name)
include/rv/da_monitor.h
232
static DEFINE_PER_CPU(struct da_monitor, DA_MON_NAME);
init/calibrate.c
254
static DEFINE_PER_CPU(unsigned long, cpu_loops_per_jiffy) = { 0 };
init/main.c
836
DEFINE_PER_CPU(u32, kstack_offset);
kernel/bpf/core.c
2985
static DEFINE_PER_CPU(struct rnd_state, bpf_user_rnd_state);
kernel/bpf/helpers.c
1164
static DEFINE_PER_CPU(struct bpf_hrtimer *, hrtimer_running);
kernel/bpf/helpers.c
1442
static DEFINE_PER_CPU(struct bpf_async_cb *, async_cb_running);
kernel/bpf/helpers.c
330
static DEFINE_PER_CPU(unsigned long, irqsave_flags);
kernel/bpf/helpers.c
770
static DEFINE_PER_CPU(struct bpf_bprintf_buffers[MAX_BPRINTF_NEST_LEVEL], bpf_bprintf_bufs);
kernel/bpf/helpers.c
771
static DEFINE_PER_CPU(int, bpf_bprintf_nest_level);
kernel/bpf/syscall.c
59
DEFINE_PER_CPU(int, bpf_prog_active);
kernel/bpf/task_iter.c
1032
DEFINE_PER_CPU(struct mmap_unlock_irq_work, mmap_unlock_work);
kernel/cgroup/cgroup.c
189
static DEFINE_PER_CPU(struct css_rstat_cpu, root_rstat_cpu);
kernel/cgroup/cgroup.c
190
static DEFINE_PER_CPU(struct cgroup_rstat_base_cpu, root_rstat_base_cpu);
kernel/cgroup/cgroup.c
7114
static DEFINE_PER_CPU(struct llist_head, cgrp_dead_tasks);
kernel/cgroup/cgroup.c
7115
static DEFINE_PER_CPU(struct irq_work, cgrp_dead_tasks_iwork);
kernel/cgroup/rstat.c
13
static DEFINE_PER_CPU(struct llist_head, rstat_backlog_list);
kernel/context_tracking.c
29
DEFINE_PER_CPU(struct context_tracking, context_tracking) = {
kernel/cpu.c
86
static DEFINE_PER_CPU(struct cpuhp_cpu_state, cpuhp_state) = {
kernel/debug/debug_core.c
238
static DEFINE_PER_CPU(call_single_data_t, kgdb_roundup_csd) =
kernel/events/callchain.c
34
static DEFINE_PER_CPU(u8, callchain_recursion[PERF_NR_CONTEXTS]);
kernel/events/core.c
10796
static DEFINE_PER_CPU(struct swevent_htable, swevent_htable);
kernel/events/core.c
11011
DEFINE_PER_CPU(struct pt_regs, __perf_regs[4]);
kernel/events/core.c
12462
static DEFINE_PER_CPU(unsigned int, nop_txn_flags);
kernel/events/core.c
244
static DEFINE_PER_CPU(struct perf_cpu_context, perf_cpu_context);
kernel/events/core.c
3860
static DEFINE_PER_CPU(struct list_head, sched_cb_list);
kernel/events/core.c
3861
static DEFINE_PER_CPU(int, perf_sched_cb_usages);
kernel/events/core.c
4401
static DEFINE_PER_CPU(int, perf_throttled_count);
kernel/events/core.c
4402
static DEFINE_PER_CPU(u64, perf_throttled_seq);
kernel/events/core.c
449
static DEFINE_PER_CPU(struct pmu_event_list, pmu_sb_events);
kernel/events/core.c
475
static DEFINE_PER_CPU(bool, guest_ctx_loaded);
kernel/events/core.c
625
static DEFINE_PER_CPU(u64, running_sample_length);
kernel/events/hw_breakpoint.c
60
static DEFINE_PER_CPU(struct bp_cpuinfo, bp_cpuinfo[TYPE_MAX]);
kernel/fork.c
154
DEFINE_PER_CPU(unsigned long, process_counts) = 0;
kernel/fork.c
199
static DEFINE_PER_CPU(struct vm_struct *, cached_stacks[NR_CACHED_STACKS]);
kernel/irq/manage.c
230
static DEFINE_PER_CPU(struct cpumask, __tmp_mask);
kernel/irq_work.c
27
static DEFINE_PER_CPU(struct llist_head, raised_list);
kernel/irq_work.c
28
static DEFINE_PER_CPU(struct llist_head, lazy_list);
kernel/irq_work.c
29
static DEFINE_PER_CPU(struct task_struct *, irq_workd);
kernel/irq_work.c
45
static DEFINE_PER_CPU(struct irq_work, irq_work_wakeup) =
kernel/kcov.c
102
static DEFINE_PER_CPU(struct kcov_percpu_data, kcov_percpu_data) = {
kernel/kcsan/core.c
109
static DEFINE_PER_CPU(long, kcsan_skip);
kernel/kcsan/core.c
112
static DEFINE_PER_CPU(u32, kcsan_rand_state);
kernel/kcsan/core.c
55
static DEFINE_PER_CPU(struct kcsan_ctx, kcsan_cpu_ctx) = {
kernel/kprobes.c
71
static DEFINE_PER_CPU(struct kprobe *, kprobe_instance);
kernel/livepatch/transition.c
18
static DEFINE_PER_CPU(unsigned long[MAX_STACK_ENTRIES], klp_stack_entries);
kernel/locking/lock_events.c
56
DEFINE_PER_CPU(unsigned long, lockevents[lockevent_num]);
kernel/locking/lockdep.c
113
DEFINE_PER_CPU(unsigned int, lockdep_recursion);
kernel/locking/lockdep.c
250
static DEFINE_PER_CPU(struct lock_class_stats[MAX_LOCKDEP_KEYS], cpu_lock_stats);
kernel/locking/lockdep.c
642
DEFINE_PER_CPU(struct lockdep_stats, lockdep_stats);
kernel/locking/qspinlock_stat.h
23
static DEFINE_PER_CPU(u64, pv_kick_time);
kernel/locking/spinlock.c
27
DEFINE_PER_CPU(struct mmiowb_state, __mmiowb_state);
kernel/printk/nbcon.c
1402
static DEFINE_PER_CPU(unsigned int, nbcon_pcpu_emergency_nesting);
kernel/printk/printk.c
2055
static DEFINE_PER_CPU(u8, printk_count);
kernel/printk/printk.c
2058
static DEFINE_PER_CPU(u8, printk_count_nmi);
kernel/printk/printk.c
4595
static DEFINE_PER_CPU(int, printk_pending);
kernel/printk/printk.c
4615
static DEFINE_PER_CPU(struct irq_work, wake_up_klogd_work) =
kernel/printk/printk_safe.c
33
static DEFINE_PER_CPU(int, printk_context);
kernel/rcu/rcutorture.c
199
static DEFINE_PER_CPU(long [RCU_TORTURE_PIPE_LEN + 1], rcu_torture_count);
kernel/rcu/rcutorture.c
200
static DEFINE_PER_CPU(long [RCU_TORTURE_PIPE_LEN + 1], rcu_torture_batch);
kernel/rcu/refscale.c
372
static DEFINE_PER_CPU(unsigned long, test_acqrel);
kernel/rcu/tasks.h
134
static DEFINE_PER_CPU(struct rcu_tasks_percpu, rt_name ## __percpu) = { \
kernel/rcu/tree.c
586
static DEFINE_PER_CPU(struct irq_work, late_wakeup_work) =
kernel/rseq.c
128
DEFINE_PER_CPU(struct rseq_stats, rseq_stats);
kernel/rseq.c
525
static DEFINE_PER_CPU(struct slice_timer, slice_timer);
kernel/scftorture.c
100
static DEFINE_PER_CPU(struct llist_head, scf_free_pool);
kernel/scftorture.c
99
static DEFINE_PER_CPU(long long, scf_invoked_count);
kernel/sched/core.c
10867
static DEFINE_PER_CPU(struct sched_change_ctx, sched_change_ctx);
kernel/sched/core.c
127
DEFINE_PER_CPU(struct rnd_state, sched_rnd_state);
kernel/sched/core.c
5430
DEFINE_PER_CPU(struct kernel_stat, kstat);
kernel/sched/core.c
5431
DEFINE_PER_CPU(struct kernel_cpustat, kernel_cpustat);
kernel/sched/core.c
6339
static DEFINE_PER_CPU(struct balance_callback, core_balance_head);
kernel/sched/core.c
8112
static DEFINE_PER_CPU(struct cpu_stop_work, push_work);
kernel/sched/cpuacct.c
49
static DEFINE_PER_CPU(u64, root_cpuacct_cpuusage);
kernel/sched/cpufreq.c
10
DEFINE_PER_CPU(struct update_util_data __rcu *, cpufreq_update_util_data);
kernel/sched/cpufreq_schedutil.c
60
static DEFINE_PER_CPU(struct sugov_cpu, sugov_cpu);
kernel/sched/cputime.c
28
DEFINE_PER_CPU(struct irqtime, cpu_irqtime);
kernel/sched/deadline.c
2713
static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask_dl);
kernel/sched/deadline.c
624
static DEFINE_PER_CPU(struct balance_callback, dl_push_head);
kernel/sched/deadline.c
625
static DEFINE_PER_CPU(struct balance_callback, dl_pull_head);
kernel/sched/ext.c
297
DEFINE_PER_CPU(struct rq *, scx_locked_rq_state);
kernel/sched/ext.c
90
static DEFINE_PER_CPU(struct scx_kick_syncs __rcu *, scx_kick_syncs);
kernel/sched/ext.c
99
static DEFINE_PER_CPU(struct task_struct *, direct_dispatch_task);
kernel/sched/ext_idle.c
145
static DEFINE_PER_CPU(nodemask_t, per_cpu_unvisited);
kernel/sched/ext_idle.c
51
static DEFINE_PER_CPU(cpumask_var_t, local_idle_cpumask);
kernel/sched/ext_idle.c
52
static DEFINE_PER_CPU(cpumask_var_t, local_llc_idle_cpumask);
kernel/sched/ext_idle.c
53
static DEFINE_PER_CPU(cpumask_var_t, local_numa_idle_cpumask);
kernel/sched/fair.c
7206
static DEFINE_PER_CPU(cpumask_var_t, load_balance_mask);
kernel/sched/fair.c
7207
static DEFINE_PER_CPU(cpumask_var_t, select_rq_mask);
kernel/sched/fair.c
7208
static DEFINE_PER_CPU(cpumask_var_t, should_we_balance_tmpmask);
kernel/sched/psi.c
174
static DEFINE_PER_CPU(struct psi_group_cpu, system_group_pcpu);
kernel/sched/psi.c
179
static DEFINE_PER_CPU(seqcount_t, psi_seq) = SEQCNT_ZERO(psi_seq);
kernel/sched/rt.c
1761
static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask);
kernel/sched/rt.c
378
static DEFINE_PER_CPU(struct balance_callback, rt_push_head);
kernel/sched/rt.c
379
static DEFINE_PER_CPU(struct balance_callback, rt_pull_head);
kernel/sched/topology.c
664
DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
kernel/sched/topology.c
665
DEFINE_PER_CPU(int, sd_llc_size);
kernel/sched/topology.c
666
DEFINE_PER_CPU(int, sd_llc_id);
kernel/sched/topology.c
667
DEFINE_PER_CPU(int, sd_share_id);
kernel/sched/topology.c
668
DEFINE_PER_CPU(struct sched_domain_shared __rcu *, sd_llc_shared);
kernel/sched/topology.c
669
DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
kernel/sched/topology.c
670
DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
kernel/sched/topology.c
671
DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
kernel/scs.c
29
static DEFINE_PER_CPU(void *, scs_cache[NR_CACHED_SCS]);
kernel/smp.c
170
static DEFINE_PER_CPU(call_single_data_t *, cur_csd);
kernel/smp.c
171
static DEFINE_PER_CPU(smp_call_func_t, cur_csd_func);
kernel/smp.c
172
static DEFINE_PER_CPU(void *, cur_csd_info);
kernel/smp.c
50
static DEFINE_PER_CPU(atomic_t, trigger_backtrace) = ATOMIC_INIT(1);
kernel/smpboot.c
26
static DEFINE_PER_CPU(struct task_struct *, idle_threads);
kernel/softirq.c
125
static DEFINE_PER_CPU(struct softirq_ctrl, softirq_ctrl) = {
kernel/softirq.c
62
DEFINE_PER_CPU(struct task_struct *, ksoftirqd);
kernel/softirq.c
696
DEFINE_PER_CPU(struct task_struct *, ktimerd);
kernel/softirq.c
697
DEFINE_PER_CPU(unsigned long, pending_timer_softirq);
kernel/softirq.c
806
static DEFINE_PER_CPU(struct tasklet_head, tasklet_vec);
kernel/softirq.c
807
static DEFINE_PER_CPU(struct tasklet_head, tasklet_hi_vec);
kernel/softirq.c
85
DEFINE_PER_CPU(int, hardirqs_enabled);
kernel/softirq.c
857
static DEFINE_PER_CPU(struct tasklet_sync_callback, tasklet_sync_callback) = {
kernel/softirq.c
86
DEFINE_PER_CPU(int, hardirq_context);
kernel/stop_machine.c
49
static DEFINE_PER_CPU(struct cpu_stopper, cpu_stopper);
kernel/task_work.c
19
static DEFINE_PER_CPU(struct irq_work, irq_work_NMI_resume) =
kernel/taskstats.c
32
static DEFINE_PER_CPU(__u32, taskstats_seqnum);
kernel/taskstats.c
58
static DEFINE_PER_CPU(struct listener_list, listener_array);
kernel/time/clockevents.c
667
static DEFINE_PER_CPU(struct device, tick_percpu_dev);
kernel/time/hrtimer.c
80
DEFINE_PER_CPU(struct hrtimer_cpu_base, hrtimer_bases) =
kernel/time/tick-broadcast.c
36
static DEFINE_PER_CPU(struct clock_event_device *, tick_oneshot_wakeup_device);
kernel/time/tick-common.c
29
DEFINE_PER_CPU(struct tick_device, tick_cpu_device);
kernel/time/tick-sched.c
40
static DEFINE_PER_CPU(struct tick_sched, tick_cpu_sched);
kernel/time/tick-sched.c
410
static DEFINE_PER_CPU(struct irq_work, nohz_full_kick_work) =
kernel/time/timekeeping_debug.c
21
DEFINE_PER_CPU(unsigned long, timekeeping_mg_floor_swaps);
kernel/time/timer.c
267
static DEFINE_PER_CPU(struct timer_base, timer_bases[NR_BASES]);
kernel/time/timer_migration.c
426
static DEFINE_PER_CPU(struct tmigr_cpu, tmigr_cpu);
kernel/trace/bpf_trace.c
1574
static DEFINE_PER_CPU(struct bpf_raw_tp_regs, bpf_raw_tp_regs);
kernel/trace/bpf_trace.c
1575
static DEFINE_PER_CPU(int, bpf_raw_tp_nest_level);
kernel/trace/bpf_trace.c
2313
static DEFINE_PER_CPU(struct pt_regs, bpf_kprobe_multi_pt_regs);
kernel/trace/bpf_trace.c
655
static DEFINE_PER_CPU(struct bpf_trace_sample_data, bpf_trace_sds);
kernel/trace/bpf_trace.c
656
static DEFINE_PER_CPU(int, bpf_trace_nest_level);
kernel/trace/bpf_trace.c
706
static DEFINE_PER_CPU(int, bpf_event_output_nest_level);
kernel/trace/bpf_trace.c
710
static DEFINE_PER_CPU(struct bpf_nested_pt_regs, bpf_pt_regs);
kernel/trace/bpf_trace.c
711
static DEFINE_PER_CPU(struct bpf_trace_sample_data, bpf_misc_sds);
kernel/trace/bpf_trace.c
803
static DEFINE_PER_CPU(struct send_signal_irq_work, send_signal_work);
kernel/trace/fgraph.c
1112
static DEFINE_PER_CPU(unsigned long *, idle_ret_stack);
kernel/trace/ftrace.c
450
static DEFINE_PER_CPU(struct ftrace_profile_stat, ftrace_profile_stats);
kernel/trace/ring_buffer.c
4378
static DEFINE_PER_CPU(atomic_t, checking);
kernel/trace/trace.c
108
DEFINE_PER_CPU(bool, trace_taskinfo_save);
kernel/trace/trace.c
2216
DEFINE_PER_CPU(struct ring_buffer_event *, trace_buffered_event);
kernel/trace/trace.c
2217
DEFINE_PER_CPU(int, trace_buffered_event_cnt);
kernel/trace/trace.c
2586
static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks);
kernel/trace/trace.c
2587
static DEFINE_PER_CPU(int, ftrace_stack_reserve);
kernel/trace/trace.c
2721
static DEFINE_PER_CPU(int, user_stack_count);
kernel/trace/trace.c
724
static DEFINE_PER_CPU(struct mutex, cpu_access_lock);
kernel/trace/trace_events.c
5005
static DEFINE_PER_CPU(atomic_t, ftrace_test_event_disable);
kernel/trace/trace_events_hist.c
5311
static DEFINE_PER_CPU(int, hist_pad_cnt);
kernel/trace/trace_hwlat.c
83
static DEFINE_PER_CPU(struct hwlat_kthread_data, hwlat_per_cpu_data);
kernel/trace/trace_irqsoff.c
27
static DEFINE_PER_CPU(int, tracing_cpu);
kernel/trace/trace_osnoise.c
222
static DEFINE_PER_CPU(struct osnoise_variables, per_cpu_osnoise_var);
kernel/trace/trace_osnoise.c
251
static DEFINE_PER_CPU(struct timerlat_variables, per_cpu_timerlat_var);
kernel/trace/trace_osnoise.c
602
static DEFINE_PER_CPU(struct trace_stack, trace_stack);
kernel/trace/trace_preemptirq.c
51
static DEFINE_PER_CPU(int, tracing_irq_cpu);
kernel/trace/trace_stack.c
32
DEFINE_PER_CPU(int, disable_stack_tracer);
kernel/unwind/deferred.c
68
static DEFINE_PER_CPU(u32, unwind_ctx_ctr);
kernel/user-return-notifier.c
8
static DEFINE_PER_CPU(struct hlist_head, return_notifier_list);
kernel/watchdog.c
138
static DEFINE_PER_CPU(atomic_t, hrtimer_interrupts);
kernel/watchdog.c
139
static DEFINE_PER_CPU(int, hrtimer_interrupts_saved);
kernel/watchdog.c
140
static DEFINE_PER_CPU(bool, watchdog_hardlockup_warned);
kernel/watchdog.c
141
static DEFINE_PER_CPU(bool, watchdog_hardlockup_touched);
kernel/watchdog.c
394
static DEFINE_PER_CPU(unsigned long, watchdog_touch_ts);
kernel/watchdog.c
396
static DEFINE_PER_CPU(unsigned long, watchdog_report_ts);
kernel/watchdog.c
397
static DEFINE_PER_CPU(struct hrtimer, watchdog_hrtimer);
kernel/watchdog.c
398
static DEFINE_PER_CPU(bool, softlockup_touch_sync);
kernel/watchdog.c
445
static DEFINE_PER_CPU(u16, cpustat_old[NUM_STATS_PER_GROUP]);
kernel/watchdog.c
446
static DEFINE_PER_CPU(u8, cpustat_util[NUM_SAMPLE_PERIODS][NUM_STATS_PER_GROUP]);
kernel/watchdog.c
447
static DEFINE_PER_CPU(u8, cpustat_tail);
kernel/watchdog.c
529
static DEFINE_PER_CPU(bool, snapshot_taken);
kernel/watchdog.c
752
static DEFINE_PER_CPU(struct completion, softlockup_completion);
kernel/watchdog.c
753
static DEFINE_PER_CPU(struct cpu_stop_work, softlockup_stop_work);
kernel/watchdog_perf.c
24
static DEFINE_PER_CPU(struct perf_event *, watchdog_ev);
kernel/watchdog_perf.c
29
static DEFINE_PER_CPU(ktime_t, last_timestamp);
kernel/watchdog_perf.c
30
static DEFINE_PER_CPU(unsigned int, nmi_rearmed);
kernel/workqueue.c
470
static DEFINE_PER_CPU(int, wq_rr_cpu_last);
kernel/workqueue.c
7575
static DEFINE_PER_CPU(unsigned long, wq_watchdog_touched_cpu) = INITIAL_JIFFIES;
lib/alloc_tag.c
30
DEFINE_PER_CPU(struct alloc_tag_counters, _shared_alloc_tag);
lib/cpumask.c
122
static DEFINE_PER_CPU(int, distribute_cpu_mask_prev);
lib/crypto/simd.c
10
DEFINE_PER_CPU(bool, crypto_simd_disabled_for_test);
lib/fault-inject.c
19
static DEFINE_PER_CPU(struct rnd_state, fault_rnd_state);
lib/irq_poll.c
17
static DEFINE_PER_CPU(struct list_head, blk_cpu_iopoll);
lib/irq_regs.c
12
DEFINE_PER_CPU(struct pt_regs *, __irq_regs);
lib/locking-selftest.c
148
static DEFINE_PER_CPU(local_lock_t, local_A);
lib/percpu_test.c
18
static DEFINE_PER_CPU(long, long_counter);
lib/percpu_test.c
19
static DEFINE_PER_CPU(unsigned long, ulong_counter);
lib/radix-tree.c
62
DEFINE_PER_CPU(struct radix_tree_preload, radix_tree_preloads) = {
lib/test_context-analysis.c
472
static DEFINE_PER_CPU(struct test_local_lock_data, test_local_lock_data) = {
lib/test_context-analysis.c
516
static DEFINE_PER_CPU(struct test_local_trylock_data, test_local_trylock_data) = {
lib/test_lockup.c
403
static DEFINE_PER_CPU(struct work_struct, test_works);
lib/test_objpool.c
134
static DEFINE_PER_CPU(struct ot_item, ot_pcup_items);
mm/huge_memory.c
598
DEFINE_PER_CPU(struct mthp_stat, mthp_stats) = {{{0}}};
mm/kasan/hw_tags.c
82
DEFINE_PER_CPU(long, kasan_page_alloc_skip);
mm/kasan/quarantine.c
109
static DEFINE_PER_CPU(struct cpu_shrink_qlist, shrink_qlist) = {
mm/kasan/quarantine.c
93
static DEFINE_PER_CPU(struct qlist_head, cpu_quarantine);
mm/kasan/sw_tags.c
37
static DEFINE_PER_CPU(u32, prng_state);
mm/kmsan/core.c
38
DEFINE_PER_CPU(struct kmsan_ctx, kmsan_percpu_ctx);
mm/kmsan/kmsan_test.c
28
static DEFINE_PER_CPU(int, per_cpu_var);
mm/memcontrol.c
87
DEFINE_PER_CPU(struct mem_cgroup *, int_active_memcg);
mm/memory-failure.c
2577
static DEFINE_PER_CPU(struct memory_failure_cpu, memory_failure_cpu);
mm/mlock.c
36
static DEFINE_PER_CPU(struct mlock_fbatch, mlock_fbatch) = {
mm/mm_init.c
250
static DEFINE_PER_CPU(struct per_cpu_nodestat, boot_nodestats);
mm/page-writeback.c
2010
static DEFINE_PER_CPU(int, bdp_ratelimits);
mm/page-writeback.c
2026
DEFINE_PER_CPU(int, dirty_throttle_leaks) = 0;
mm/page_alloc.c
197
DEFINE_PER_CPU(int, numa_node);
mm/page_alloc.c
210
DEFINE_PER_CPU(int, _numa_mem_); /* Kernel "local memory" node */
mm/page_alloc.c
5794
static DEFINE_PER_CPU(struct per_cpu_pages, boot_pageset);
mm/page_alloc.c
5795
static DEFINE_PER_CPU(struct per_cpu_zonestat, boot_zonestats);
mm/slab_common.c
1407
static DEFINE_PER_CPU(struct kfree_rcu_cpu, krc) = {
mm/slub.c
3368
static DEFINE_PER_CPU(struct rnd_state, slab_rnd_state);
mm/slub.c
489
static DEFINE_PER_CPU(struct slub_flush_work, slub_flush);
mm/slub.c
6094
static DEFINE_PER_CPU(struct defer_free, defer_free_objects) = {
mm/swap.c
68
static DEFINE_PER_CPU(struct cpu_fbatches, cpu_fbatches) = {
mm/swap.c
768
static DEFINE_PER_CPU(struct work_struct, lru_add_drain_work);
mm/swapfile.c
127
static DEFINE_PER_CPU(struct percpu_swap_cluster, percpu_swap_cluster) = {
mm/vmalloc.c
2618
static DEFINE_PER_CPU(struct vmap_block_queue, vmap_block_queue);
mm/vmalloc.c
905
static DEFINE_PER_CPU(struct vmap_area *, ne_fit_preload_node);
mm/vmalloc.c
91
static DEFINE_PER_CPU(struct vfree_deferred, vfree_deferred);
mm/vmstat.c
110
DEFINE_PER_CPU(struct vm_event_state, vm_event_states) = {{0}};
mm/vmstat.c
1965
static DEFINE_PER_CPU(struct delayed_work, vmstat_work);
net/bridge/br_netfilter_hooks.c
150
static DEFINE_PER_CPU(struct brnf_frag_data, brnf_frag_data_storage) = {
net/core/dev.c
471
DEFINE_PER_CPU(struct page_pool_bh, system_page_pool) = {
net/core/dev.c
4935
static DEFINE_PER_CPU(struct task_struct *, backlog_napi);
net/core/drop_monitor.c
98
static DEFINE_PER_CPU(struct per_cpu_dm_data, dm_cpu_data);
net/core/drop_monitor.c
99
static DEFINE_PER_CPU(struct per_cpu_dm_data, dm_hw_cpu_data);
net/core/page_pool.c
43
static DEFINE_PER_CPU(struct page_pool_recycle_stats, pp_system_recycle_stats);
net/core/skbuff.c
238
static DEFINE_PER_CPU(struct page_frag_cache, netdev_alloc_cache);
net/core/skbuff.c
239
static DEFINE_PER_CPU(struct napi_alloc_cache, napi_alloc_cache) = {
net/ipv4/icmp.c
202
static DEFINE_PER_CPU(struct sock *, ipv4_icmp_sk);
net/ipv4/route.c
192
static DEFINE_PER_CPU(struct rt_cache_stat, rt_cache_stat);
net/ipv4/tcp.c
299
DEFINE_PER_CPU(unsigned int, tcp_orphan_count);
net/ipv4/tcp.c
302
DEFINE_PER_CPU(u32, tcp_tw_isn);
net/ipv4/tcp.c
308
DEFINE_PER_CPU(int, tcp_memory_per_cpu_fw_alloc);
net/ipv4/tcp_ipv4.c
102
static DEFINE_PER_CPU(struct sock_bh_locked, ipv4_tcp_sk) = {
net/ipv4/tcp_output.c
1256
static DEFINE_PER_CPU(struct tsq_work, tsq_work);
net/ipv4/tcp_sigpool.c
18
static DEFINE_PER_CPU(struct sigpool_scratch, sigpool_scratch) = {
net/ipv4/udp.c
130
DEFINE_PER_CPU(int, udp_memory_per_cpu_fw_alloc);
net/ipv6/icmp.c
72
static DEFINE_PER_CPU(struct sock *, ipv6_icmp_sk);
net/ipv6/seg6_hmac.c
49
static DEFINE_PER_CPU(struct hmac_storage, hmac_storage) = {
net/ipv6/seg6_local.c
1386
DEFINE_PER_CPU(struct seg6_bpf_srh_state, seg6_bpf_srh_states) = {
net/mptcp/protocol.c
53
DEFINE_PER_CPU(struct mptcp_delegated_action, mptcp_delegated_actions) = {
net/netfilter/nft_counter.c
33
static DEFINE_PER_CPU(struct u64_stats_sync, nft_counter_sync);
net/netfilter/nft_ct.c
35
static DEFINE_PER_CPU(struct nf_conn *, nft_ct_pcpu_template);
net/netfilter/nft_inner.c
31
static DEFINE_PER_CPU(struct nft_inner_tun_ctx_locked, nft_pcpu_tun_ctx) = {
net/netfilter/x_tables.c
1347
DEFINE_PER_CPU(seqcount_t, xt_recseq);
net/rds/tcp_stats.c
40
DEFINE_PER_CPU(struct rds_tcp_statistics, rds_tcp_stats)
net/sched/sch_frag.c
22
static DEFINE_PER_CPU(struct sch_frag_data, sch_frag_data_storage) = {
net/sctp/socket.c
98
static DEFINE_PER_CPU(int, sctp_memory_per_cpu_fw_alloc);
net/xfrm/xfrm_input.c
53
static DEFINE_PER_CPU(struct xfrm_trans_tasklet, xfrm_trans_tasklet);
net/xfrm/xfrm_nat_keepalive.c
12
static DEFINE_PER_CPU(struct sock_bh_locked, nat_keepalive_sk_ipv4) = {
net/xfrm/xfrm_nat_keepalive.c
16
static DEFINE_PER_CPU(struct sock_bh_locked, nat_keepalive_sk_ipv6) = {
samples/kmemleak/kmemleak-test.c
29
static DEFINE_PER_CPU(void *, kmemleak_test_pointer);
security/apparmor/capability.c
40
static DEFINE_PER_CPU(struct audit_cache, audit_cache);
security/apparmor/lsm.c
67
static DEFINE_PER_CPU(struct aa_local_cache, aa_local_buffers);
security/selinux/avc.c
87
DEFINE_PER_CPU(struct avc_cache_stats, avc_cache_stats) = { 0 };
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
48
static DEFINE_PER_CPU(struct rqsl_cpu_hist, rqsl_cpu_hists);
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
36
DEFINE_PER_CPU(int, bpf_testmod_ksym_percpu) = 123;
virt/kvm/kvm_main.c
117
static DEFINE_PER_CPU(struct kvm_vcpu *, kvm_running_vcpu);
virt/kvm/kvm_main.c
156
static DEFINE_PER_CPU(cpumask_var_t, cpu_kick_mask);
virt/kvm/kvm_main.c
5584
static DEFINE_PER_CPU(bool, virtualization_enabled);