arch/alpha/include/asm/mmu_context.h
143
cpu_data[cpu].asn_lock = 1;
arch/alpha/include/asm/mmu_context.h
154
cpu_data[cpu].need_new_asn = 1;
arch/alpha/include/asm/mmu_context.h
171
cpu_data[cpu].asn_lock = 0; \
arch/alpha/include/asm/mmu_context.h
173
if (cpu_data[cpu].need_new_asn) { \
arch/alpha/include/asm/mmu_context.h
175
cpu_data[cpu].need_new_asn = 0; \
arch/alpha/include/asm/mmu_context.h
91
#define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn)
arch/alpha/include/asm/smp.h
41
extern struct cpuinfo_alpha cpu_data[NR_CPUS];
arch/alpha/kernel/irq.c
77
seq_printf(p, "%10lu ", cpu_data[j].ipi_count);
arch/alpha/kernel/proto.h
208
#define mcheck_expected(cpu) (cpu_data[cpu].mcheck_expected)
arch/alpha/kernel/proto.h
209
#define mcheck_taken(cpu) (cpu_data[cpu].mcheck_taken)
arch/alpha/kernel/proto.h
210
#define mcheck_extra(cpu) (cpu_data[cpu].mcheck_extra)
arch/alpha/kernel/smp.c
487
bogosum += cpu_data[cpu].loops_per_jiffy;
arch/alpha/kernel/smp.c
55
struct cpuinfo_alpha cpu_data[NR_CPUS];
arch/alpha/kernel/smp.c
554
cpu_data[this_cpu].ipi_count++;
arch/alpha/kernel/smp.c
56
EXPORT_SYMBOL(cpu_data);
arch/alpha/kernel/smp.c
622
#define asn_locked() (cpu_data[smp_processor_id()].asn_lock)
arch/alpha/kernel/smp.c
83
cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy;
arch/alpha/kernel/smp.c
84
cpu_data[cpuid].last_asn = ASN_FIRST_VERSION;
arch/alpha/kernel/smp.c
85
cpu_data[cpuid].need_new_asn = 0;
arch/alpha/kernel/smp.c
86
cpu_data[cpuid].asn_lock = 0;
arch/alpha/kernel/smp.c
95
cpu_data[cpuid].prof_counter = 1;
arch/alpha/kernel/smp.c
96
cpu_data[cpuid].prof_multiplier = 1;
arch/alpha/lib/udelay.c
37
#define LPJ cpu_data[smp_processor_id()].loops_per_jiffy
arch/alpha/mm/tlbflush.c
22
#define asn_locked() (cpu_data[smp_processor_id()].asn_lock)
arch/arm/include/asm/cpu.h
20
DECLARE_PER_CPU(struct cpuinfo_arm, cpu_data);
arch/arm/include/asm/smp_plat.h
38
struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpu);
arch/arm/kernel/setup.c
1271
cpuid = is_smp() ? per_cpu(cpu_data, i).cpuid : read_cpuid_id();
arch/arm/kernel/setup.c
1277
per_cpu(cpu_data, i).loops_per_jiffy / (500000UL/HZ),
arch/arm/kernel/setup.c
1278
(per_cpu(cpu_data, i).loops_per_jiffy / (5000UL/HZ)) % 100);
arch/arm/kernel/setup.c
158
DEFINE_PER_CPU(struct cpuinfo_arm, cpu_data);
arch/arm/kernel/smp.c
391
struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid);
arch/arm/kernel/smp.c
487
bogosum += per_cpu(cpu_data, cpu).loops_per_jiffy;
arch/arm/kernel/smp.c
808
per_cpu(cpu_data, cpu).loops_per_jiffy;
arch/arm/kernel/smp.c
827
per_cpu(cpu_data, cpu).loops_per_jiffy = lpj;
arch/arm64/include/asm/cpu.h
73
DECLARE_PER_CPU(struct cpuinfo_arm64, cpu_data);
arch/arm64/kernel/cpufeature.c
4017
struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
arch/arm64/kernel/cpuinfo.c
282
return *pos < nr_cpu_ids ? &per_cpu(cpu_data, *pos) : NULL;
arch/arm64/kernel/cpuinfo.c
33
DEFINE_PER_CPU(struct cpuinfo_arm64, cpu_data);
arch/arm64/kernel/cpuinfo.c
363
struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
arch/arm64/kernel/cpuinfo.c
385
struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
arch/arm64/kernel/cpuinfo.c
403
struct cpuinfo_arm64 *info = &per_cpu(cpu_data, cpu);
arch/arm64/kernel/cpuinfo.c
524
struct cpuinfo_arm64 *info = this_cpu_ptr(&cpu_data);
arch/arm64/kernel/cpuinfo.c
531
struct cpuinfo_arm64 *info = &per_cpu(cpu_data, 0);
arch/csky/include/asm/processor.h
21
extern struct cpuinfo_csky cpu_data[];
arch/csky/kernel/process.c
17
struct cpuinfo_csky cpu_data[NR_CPUS];
arch/loongarch/include/asm/cpu-features.h
15
#define cpu_opt(opt) (cpu_data[0].options & (opt))
arch/loongarch/include/asm/cpu-features.h
16
#define cpu_has(feat) (cpu_data[0].options & BIT_ULL(feat))
arch/loongarch/include/asm/cpu-features.h
19
#define cpu_has_loongarch32 (cpu_data[0].isa_level & LOONGARCH_CPU_ISA_32BIT)
arch/loongarch/include/asm/cpu-features.h
20
#define cpu_has_loongarch64 (cpu_data[0].isa_level & LOONGARCH_CPU_ISA_64BIT)
arch/loongarch/include/asm/cpu-features.h
28
# define cpu_vabits cpu_data[0].vabits
arch/loongarch/include/asm/cpu-features.h
29
# define cpu_pabits cpu_data[0].pabits
arch/loongarch/include/asm/cpu-info.h
67
extern struct cpuinfo_loongarch cpu_data[];
arch/loongarch/include/asm/cpu-info.h
68
#define boot_cpu_data cpu_data[0]
arch/loongarch/include/asm/cpu-info.h
69
#define current_cpu_data cpu_data[smp_processor_id()]
arch/loongarch/include/asm/cpu-info.h
70
#define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
arch/loongarch/include/asm/cpu-info.h
81
struct cpuinfo_loongarch *infoa = &cpu_data[cpua];
arch/loongarch/include/asm/cpu-info.h
82
struct cpuinfo_loongarch *infob = &cpu_data[cpub];
arch/loongarch/include/asm/mmu_context.h
26
return ~(u64)(cpu_asid_mask(&cpu_data[cpu]));
arch/loongarch/include/asm/mmu_context.h
31
return cpu_asid_mask(&cpu_data[cpu]) + 1;
arch/loongarch/include/asm/mmu_context.h
35
#define asid_cache(cpu) (cpu_data[cpu].asid_cache)
arch/loongarch/include/asm/mmu_context.h
36
#define cpu_asid(cpu, mm) (cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu]))
arch/loongarch/include/asm/mmu_context.h
56
if (!((++asid) & cpu_asid_mask(&cpu_data[cpu])))
arch/loongarch/include/asm/topology.h
38
#define topology_physical_package_id(cpu) (cpu_data[cpu].package)
arch/loongarch/include/asm/topology.h
39
#define topology_core_id(cpu) (cpu_data[cpu].core)
arch/loongarch/kernel/acpi.c
189
cpu_data[cpu].core = topology_id;
arch/loongarch/kernel/acpi.c
195
cpu_data[cpu].core = topology_id;
arch/loongarch/kernel/cpu-probe.c
70
struct cpuinfo_loongarch *c = &cpu_data[0];
arch/loongarch/kernel/hw_breakpoint.c
537
cpu_data[cpu].watch_ireg_count = boot_cpu_data.watch_ireg_count;
arch/loongarch/kernel/hw_breakpoint.c
538
cpu_data[cpu].watch_dreg_count = boot_cpu_data.watch_dreg_count;
arch/loongarch/kernel/proc.c
100
cpu_data[n].watch_ireg_count, cpu_data[n].watch_dreg_count);
arch/loongarch/kernel/proc.c
19
unsigned int isa = cpu_data[n].isa_level;
arch/loongarch/kernel/proc.c
20
unsigned int prid = cpu_data[n].processor_id;
arch/loongarch/kernel/proc.c
21
unsigned int version = cpu_data[n].processor_id & 0xff;
arch/loongarch/kernel/proc.c
22
unsigned int fp_version = cpu_data[n].fpu_vers;
arch/loongarch/kernel/proc.c
39
seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package);
arch/loongarch/kernel/proc.c
40
seq_printf(m, "core\t\t\t: %d\n", cpu_data[n].core);
arch/loongarch/kernel/proc.c
41
seq_printf(m, "global_id\t\t: %d\n", cpu_data[n].global_id);
arch/loongarch/kernel/proc.c
49
seq_printf(m, "TLB Entries\t\t: %d\n", cpu_data[n].tlbsize);
arch/loongarch/kernel/setup.c
64
struct cpuinfo_loongarch cpu_data[NR_CPUS] __read_mostly;
arch/loongarch/kernel/setup.c
66
EXPORT_SYMBOL(cpu_data);
arch/loongarch/kernel/smp.c
105
if (cpu_data[cpu].package == cpu_data[i].package) {
arch/loongarch/kernel/smp.c
357
cpu_data[0].core = cpu_logical_map(0) % loongson_sysconf.cores_per_package;
arch/loongarch/kernel/smp.c
358
cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package;
arch/loongarch/kernel/smp.c
371
cpu_data[0].global_id = cpu_logical_map(0);
arch/loongarch/kernel/smp.c
428
cpu_data[cpu].package =
arch/loongarch/kernel/smp.c
430
cpu_data[cpu].core = pptt_enabled ? cpu_data[cpu].core :
arch/loongarch/kernel/smp.c
432
cpu_data[cpu].global_id = cpu_logical_map(cpu);
arch/loongarch/kernel/traps.c
1154
if (!cpu_data[cpu].asid_cache)
arch/loongarch/kernel/traps.c
1155
cpu_data[cpu].asid_cache = asid_first_version(cpu);
arch/mips/include/asm/cpu-features.h
17
#define __ase(ase) (cpu_data[0].ases & (ase))
arch/mips/include/asm/cpu-features.h
18
#define __isa(isa) (cpu_data[0].isa_level & (isa))
arch/mips/include/asm/cpu-features.h
19
#define __opt(opt) (cpu_data[0].options & (opt))
arch/mips/include/asm/cpu-features.h
249
#define cpu_has_vtag_icache (cpu_data[0].icache.flags & MIPS_CACHE_VTAG)
arch/mips/include/asm/cpu-features.h
252
#define cpu_has_dc_aliases (cpu_data[0].dcache.flags & MIPS_CACHE_ALIASES)
arch/mips/include/asm/cpu-features.h
255
#define cpu_has_ic_fills_f_dc (cpu_data[0].icache.flags & MIPS_CACHE_IC_F_DC)
arch/mips/include/asm/cpu-features.h
258
#define cpu_has_pindexed_dcache (cpu_data[0].dcache.flags & MIPS_CACHE_PINDEX)
arch/mips/include/asm/cpu-features.h
274
#define cpu_icache_snoops_remote_store (cpu_data[0].icache.flags & MIPS_IC_SNOOPS_REMOTE)
arch/mips/include/asm/cpu-features.h
459
# define cpu_has_64bits (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
arch/mips/include/asm/cpu-features.h
462
# define cpu_has_64bit_zero_reg (cpu_data[0].isa_level & MIPS_CPU_ISA_64BIT)
arch/mips/include/asm/cpu-features.h
486
# define cpu_vmbits cpu_data[0].vmbits
arch/mips/include/asm/cpu-features.h
507
#define cpu_dcache_line_size() cpu_data[0].dcache.linesz
arch/mips/include/asm/cpu-features.h
510
#define cpu_icache_line_size() cpu_data[0].icache.linesz
arch/mips/include/asm/cpu-features.h
513
#define cpu_scache_line_size() cpu_data[0].scache.linesz
arch/mips/include/asm/cpu-features.h
516
#define cpu_tcache_line_size() cpu_data[0].tcache.linesz
arch/mips/include/asm/cpu-features.h
665
#define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1))
arch/mips/include/asm/cpu-features.h
668
#define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2))
arch/mips/include/asm/cpu-features.h
671
#define cpu_guest_has_conf3 (cpu_data[0].guest.conf & (1 << 3))
arch/mips/include/asm/cpu-features.h
674
#define cpu_guest_has_conf4 (cpu_data[0].guest.conf & (1 << 4))
arch/mips/include/asm/cpu-features.h
677
#define cpu_guest_has_conf5 (cpu_data[0].guest.conf & (1 << 5))
arch/mips/include/asm/cpu-features.h
680
#define cpu_guest_has_conf6 (cpu_data[0].guest.conf & (1 << 6))
arch/mips/include/asm/cpu-features.h
683
#define cpu_guest_has_conf7 (cpu_data[0].guest.conf & (1 << 7))
arch/mips/include/asm/cpu-features.h
686
#define cpu_guest_has_fpu (cpu_data[0].guest.options & MIPS_CPU_FPU)
arch/mips/include/asm/cpu-features.h
689
#define cpu_guest_has_watch (cpu_data[0].guest.options & MIPS_CPU_WATCH)
arch/mips/include/asm/cpu-features.h
692
#define cpu_guest_has_contextconfig (cpu_data[0].guest.options & MIPS_CPU_CTXTC)
arch/mips/include/asm/cpu-features.h
695
#define cpu_guest_has_segments (cpu_data[0].guest.options & MIPS_CPU_SEGMENTS)
arch/mips/include/asm/cpu-features.h
698
#define cpu_guest_has_badinstr (cpu_data[0].guest.options & MIPS_CPU_BADINSTR)
arch/mips/include/asm/cpu-features.h
701
#define cpu_guest_has_badinstrp (cpu_data[0].guest.options & MIPS_CPU_BADINSTRP)
arch/mips/include/asm/cpu-features.h
704
#define cpu_guest_has_htw (cpu_data[0].guest.options & MIPS_CPU_HTW)
arch/mips/include/asm/cpu-features.h
707
#define cpu_guest_has_ldpte (cpu_data[0].guest.options & MIPS_CPU_LDPTE)
arch/mips/include/asm/cpu-features.h
710
#define cpu_guest_has_mvh (cpu_data[0].guest.options & MIPS_CPU_MVH)
arch/mips/include/asm/cpu-features.h
713
#define cpu_guest_has_msa (cpu_data[0].guest.ases & MIPS_ASE_MSA)
arch/mips/include/asm/cpu-features.h
716
#define cpu_guest_has_kscr(n) (cpu_data[0].guest.kscratch_mask & (1u << (n)))
arch/mips/include/asm/cpu-features.h
719
#define cpu_guest_has_rw_llb (cpu_has_mips_r6 || (cpu_data[0].guest.options & MIPS_CPU_RW_LLB))
arch/mips/include/asm/cpu-features.h
722
#define cpu_guest_has_perf (cpu_data[0].guest.options & MIPS_CPU_PERF)
arch/mips/include/asm/cpu-features.h
725
#define cpu_guest_has_maar (cpu_data[0].guest.options & MIPS_CPU_MAAR)
arch/mips/include/asm/cpu-features.h
728
#define cpu_guest_has_userlocal (cpu_data[0].guest.options & MIPS_CPU_ULRI)
arch/mips/include/asm/cpu-features.h
735
#define cpu_guest_has_dyn_fpu (cpu_data[0].guest.options_dyn & MIPS_CPU_FPU)
arch/mips/include/asm/cpu-features.h
738
#define cpu_guest_has_dyn_watch (cpu_data[0].guest.options_dyn & MIPS_CPU_WATCH)
arch/mips/include/asm/cpu-features.h
741
#define cpu_guest_has_dyn_contextconfig (cpu_data[0].guest.options_dyn & MIPS_CPU_CTXTC)
arch/mips/include/asm/cpu-features.h
744
#define cpu_guest_has_dyn_perf (cpu_data[0].guest.options_dyn & MIPS_CPU_PERF)
arch/mips/include/asm/cpu-features.h
747
#define cpu_guest_has_dyn_msa (cpu_data[0].guest.ases_dyn & MIPS_ASE_MSA)
arch/mips/include/asm/cpu-features.h
750
#define cpu_guest_has_dyn_maar (cpu_data[0].guest.options_dyn & MIPS_CPU_MAAR)
arch/mips/include/asm/cpu-info.h
117
extern struct cpuinfo_mips cpu_data[];
arch/mips/include/asm/cpu-info.h
118
#define current_cpu_data cpu_data[smp_processor_id()]
arch/mips/include/asm/cpu-info.h
119
#define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
arch/mips/include/asm/cpu-info.h
120
#define boot_cpu_data cpu_data[0]
arch/mips/include/asm/cpu-info.h
182
struct cpuinfo_mips *infoa = &cpu_data[cpua];
arch/mips/include/asm/cpu-info.h
183
struct cpuinfo_mips *infob = &cpu_data[cpub];
arch/mips/include/asm/cpu-type.h
197
const int cpu_type = cpu_data[0].cputype;
arch/mips/include/asm/mach-cavium-octeon/cpu-feature-overrides.h
59
#define cpu_has_rixi (cpu_data[0].cputype != CPU_CAVIUM_OCTEON)
arch/mips/include/asm/mips-cm.h
457
unsigned int core = cpu_core(&cpu_data[cpu]);
arch/mips/include/asm/mips-cm.h
458
unsigned int vp = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/include/asm/mips-cm.h
513
struct cpuinfo_mips *d = &cpu_data[cpu];
arch/mips/include/asm/mips-cps.h
255
last_cl = cpu_cluster(&cpu_data[nr_cpu_ids - 1]);
arch/mips/include/asm/mmu_context.h
123
#define asid_cache(cpu) (cpu_data[cpu].asid_cache)
arch/mips/include/asm/mmu_context.h
125
(cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu]))
arch/mips/include/asm/mmu_context.h
211
write_c0_memorymapid(ctx & cpu_asid_mask(&cpu_data[cpu]));
arch/mips/include/asm/mmu_context.h
96
unsigned long asid_mask = cpu_asid_mask(&cpu_data[cpu]);
arch/mips/include/asm/pgtable.h
476
prot = (prot & ~_CACHE_MASK) | cpu_data[0].writecombine;
arch/mips/include/asm/processor.h
57
#define TASK_SIZE64 (0x1UL << ((cpu_data[0].vmbits>48)?48:cpu_data[0].vmbits))
arch/mips/include/asm/timex.h
64
asm volatile("" : "=m" (cpu_data[0].options));
arch/mips/include/asm/topology.h
15
#define topology_physical_package_id(cpu) (cpu_data[cpu].package)
arch/mips/include/asm/topology.h
16
#define topology_core_id(cpu) (cpu_core(&cpu_data[cpu]))
arch/mips/kernel/cacheinfo.c
70
int cluster = cpu_cluster(&cpu_data[cpu]);
arch/mips/kernel/cacheinfo.c
73
if (cpu_cluster(&cpu_data[cpu1]) == cluster)
arch/mips/kernel/cpu-probe.c
104
if (set_ftlb_enable(&cpu_data[0], 0)) {
arch/mips/kernel/cpu-probe.c
134
cpu_data[0].tlbsize -= cpu_data[0].tlbsizeftlbways *
arch/mips/kernel/cpu-probe.c
135
cpu_data[0].tlbsizeftlbsets;
arch/mips/kernel/cpu-probe.c
136
cpu_data[0].tlbsizeftlbsets = 0;
arch/mips/kernel/cpu-probe.c
137
cpu_data[0].tlbsizeftlbways = 0;
arch/mips/kernel/cpu-probe.c
2079
cpu_data[cpu].asid_cache = asid_first_version(cpu);
arch/mips/kernel/cpu-probe.c
60
cpu_data[0].ases &= ~(MIPS_ASE_DSP | MIPS_ASE_DSP2P);
arch/mips/kernel/cpu-probe.c
73
cpu_data[0].options &= ~MIPS_CPU_HTW;
arch/mips/kernel/perf_event_mipsxx.c
385
ctrl = M_PERFCTL_VPEID(cpu_vpe_id(&cpu_data[cpu]));
arch/mips/kernel/pm-cps.c
443
vpe_id = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/kernel/pm-cps.c
464
cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].icache,
arch/mips/kernel/pm-cps.c
468
cps_gen_cache_routine(&p, &l, &r, &cpu_data[cpu].dcache,
arch/mips/kernel/pm-cps.c
481
uasm_i_addiu(&p, GPR_T0, GPR_ZERO, 1 << cpu_core(&cpu_data[cpu]));
arch/mips/kernel/pm-cps.c
495
err = cps_gen_flush_fsb(&p, &l, &r, &cpu_data[cpu],
arch/mips/kernel/pm-cps.c
638
core = cpu_core(&cpu_data[cpu]);
arch/mips/kernel/proc.c
280
cpu_data[n].srsets);
arch/mips/kernel/proc.c
282
hweight8(cpu_data[n].kscratch_mask));
arch/mips/kernel/proc.c
283
seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package);
arch/mips/kernel/proc.c
284
seq_printf(m, "core\t\t\t: %d\n", cpu_core(&cpu_data[n]));
arch/mips/kernel/proc.c
288
seq_printf(m, "VPE\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n]));
arch/mips/kernel/proc.c
290
seq_printf(m, "VP\t\t\t: %d\n", cpu_vpe_id(&cpu_data[n]));
arch/mips/kernel/proc.c
40
unsigned int version = cpu_data[n].processor_id;
arch/mips/kernel/proc.c
41
unsigned int fp_vers = cpu_data[n].fpu_id;
arch/mips/kernel/proc.c
62
cpu_data[n].options & MIPS_CPU_FPU ? " FPU V%d.%d" : "");
arch/mips/kernel/proc.c
67
cpu_data[n].udelay_val / (500000/HZ),
arch/mips/kernel/proc.c
68
(cpu_data[n].udelay_val / (5000/HZ)) % 100);
arch/mips/kernel/proc.c
72
seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize);
arch/mips/kernel/proc.c
78
cpu_data[n].watch_reg_count);
arch/mips/kernel/proc.c
79
for (i = 0; i < cpu_data[n].watch_reg_count; i++)
arch/mips/kernel/proc.c
81
cpu_data[n].watch_reg_masks[i]);
arch/mips/kernel/setup.c
54
struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
arch/mips/kernel/setup.c
56
EXPORT_SYMBOL(cpu_data);
arch/mips/kernel/setup.c
824
cpu_data[cpu].udelay_val = loops_per_jiffy;
arch/mips/kernel/smp-cps.c
242
cpu_set_cluster(&cpu_data[nvpes + v], cl);
arch/mips/kernel/smp-cps.c
243
cpu_set_core(&cpu_data[nvpes + v], c);
arch/mips/kernel/smp-cps.c
244
cpu_set_vpe_id(&cpu_data[nvpes + v], v);
arch/mips/kernel/smp-cps.c
292
return cpu_data[first_cpu_cluster].udelay_val;
arch/mips/kernel/smp-cps.c
619
unsigned int cluster = cpu_cluster(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
620
unsigned core = cpu_core(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
621
unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
633
atomic_or(1 << cpu_vpe_id(&cpu_data[cpu]), &core_cfg->vpe_mask);
arch/mips/kernel/smp-cps.c
738
core = cpu_core(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
741
vpe_id = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
839
unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
855
unsigned int cluster = cpu_cluster(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
856
unsigned core = cpu_core(&cpu_data[cpu]);
arch/mips/kernel/smp-cps.c
857
unsigned int vpe_id = cpu_vpe_id(&cpu_data[cpu]);
arch/mips/kernel/smp-mt.c
76
cpu_set_vpe_id(&cpu_data[ncpu], tc);
arch/mips/kernel/smp.c
128
if (cpu_data[cpu].package == cpu_data[i].package) {
arch/mips/kernel/smp.c
203
core = cpu_core(&cpu_data[cpu]);
arch/mips/kernel/smp.c
378
cpu_data[cpu].udelay_val = loops_per_jiffy;
arch/mips/kernel/time.c
58
cpu_data[cpu].udelay_val;
arch/mips/kernel/time.c
77
cpu_data[cpu].udelay_val = (unsigned int)lpj;
arch/mips/kernel/traps.c
2265
cpu_data[cpu].asid_cache = 0;
arch/mips/kernel/traps.c
2266
else if (!cpu_data[cpu].asid_cache)
arch/mips/kernel/traps.c
2267
cpu_data[cpu].asid_cache = asid_first_version(cpu);
arch/mips/kvm/entry.c
326
UASM_i_LA_mostly(&p, GPR_AT, (long)&cpu_data[0].asid_mask);
arch/mips/kvm/entry.c
328
UASM_i_LW(&p, GPR_T2, uasm_rel_lo((long)&cpu_data[0].asid_mask), GPR_AT);
arch/mips/kvm/entry.c
71
unsigned int kscratch_mask = cpu_data[0].kscratch_mask;
arch/mips/kvm/vz.c
1801
ret += __arch_hweight8(cpu_data[0].guest.kscratch_mask);
arch/mips/kvm/vz.c
2411
#define guestid_cache(cpu) (cpu_data[cpu].guestid_cache)
arch/mips/loongson64/smp.c
417
cpu_set_core(&cpu_data[cpu],
arch/mips/loongson64/smp.c
419
cpu_data[cpu].package =
arch/mips/loongson64/smp.c
466
cpu_set_cluster(&cpu_data[num], i / 4);
arch/mips/loongson64/smp.c
486
cpu_set_core(&cpu_data[0],
arch/mips/loongson64/smp.c
488
cpu_data[0].package = cpu_logical_map(0) / loongson_sysconf.cores_per_package;
arch/mips/loongson64/smp.c
571
[sets] "r" (cpu_data[smp_processor_id()].dcache.sets));
arch/mips/loongson64/smp.c
633
[sets] "r" (cpu_data[smp_processor_id()].dcache.sets));
arch/mips/loongson64/smp.c
717
[sets] "r" (cpu_data[smp_processor_id()].dcache.sets),
arch/mips/loongson64/smp.c
718
[vsets] "r" (cpu_data[smp_processor_id()].vcache.sets));
arch/mips/loongson64/smp.c
800
uint64_t core_id = cpu_core(&cpu_data[cpu]);
arch/mips/loongson64/smp.c
801
uint64_t package_id = cpu_data[cpu].package;
arch/mips/loongson64/smp.c
817
uint64_t core_id = cpu_core(&cpu_data[cpu]);
arch/mips/loongson64/smp.c
818
uint64_t package_id = cpu_data[cpu].package;
arch/mips/mm/context.c
210
old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache);
arch/mips/mm/context.c
213
!cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) {
arch/mips/mm/context.c
220
WRITE_ONCE(cpu_data[cpu].asid_cache, ctx);
arch/mips/mm/context.c
36
if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) {
arch/mips/mm/context.c
76
mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0);
arch/mips/mm/context.c
88
__set_bit(mmid & cpu_asid_mask(&cpu_data[cpu]), mmid_map);
arch/mips/mm/tlbex.c
314
unsigned int a = cpu_data[0].kscratch_mask & ~kscratch_used_mask;
arch/parisc/include/asm/processor.h
100
DECLARE_PER_CPU(struct cpuinfo_parisc, cpu_data);
arch/parisc/kernel/irq.c
318
return per_cpu(cpu_data, cpu).txn_addr;
arch/parisc/kernel/irq.c
330
(!per_cpu(cpu_data, next_cpu).txn_addr ||
arch/parisc/kernel/irq.c
523
per_cpu(cpu_data, cpu).hpa);
arch/parisc/kernel/perf.c
523
cpu_device = per_cpu(cpu_data, 0).dev;
arch/parisc/kernel/perf.c
525
per_cpu(cpu_data, 0).dev->name);
arch/parisc/kernel/processor.c
161
p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/processor.c
341
per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision;
arch/parisc/kernel/processor.c
342
per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model;
arch/parisc/kernel/processor.c
385
strscpy(cpu_name, per_cpu(cpu_data, 0).dev->name, sizeof(cpu_name));
arch/parisc/kernel/processor.c
392
const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
arch/parisc/kernel/processor.c
41
DEFINE_PER_CPU(struct cpuinfo_parisc, cpu_data);
arch/parisc/kernel/setup.c
299
per_cpu(cpu_data, cpunum).fp_rev = coproc_cfg.revision;
arch/parisc/kernel/setup.c
300
per_cpu(cpu_data, cpunum).fp_model = coproc_cfg.model;
arch/parisc/kernel/smp.c
123
struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu);
arch/parisc/kernel/smp.c
199
struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu);
arch/parisc/kernel/smp.c
335
const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/topology.c
49
p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/topology.c
51
const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu);
arch/s390/pci/pci_irq.c
216
struct cpu_irq_data *cpu_data;
arch/s390/pci/pci_irq.c
233
cpu_data = &per_cpu(irq_data, cpu);
arch/s390/pci/pci_irq.c
234
if (atomic_inc_return(&cpu_data->scheduled) > 1)
arch/s390/pci/pci_irq.c
237
INIT_CSD(&cpu_data->csd, zpci_handle_remote_irq, &cpu_data->scheduled);
arch/s390/pci/pci_irq.c
238
smp_call_function_single_async(cpu, &cpu_data->csd);
arch/sh/include/asm/mmu_context.h
37
#define asid_cache(cpu) (cpu_data[cpu].asid_cache)
arch/sh/include/asm/processor.h
89
extern struct sh_cpuinfo cpu_data[];
arch/sh/include/asm/processor.h
90
#define boot_cpu_data cpu_data[0]
arch/sh/include/asm/processor.h
91
#define current_cpu_data cpu_data[smp_processor_id()]
arch/sh/include/asm/processor.h
92
#define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
arch/sh/kernel/cpu/proc.c
135
return *pos < nr_cpu_ids ? cpu_data + *pos : NULL;
arch/sh/kernel/cpu/proc.c
82
unsigned int cpu = c - cpu_data;
arch/sh/kernel/setup.c
54
struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
arch/sh/kernel/setup.c
62
EXPORT_SYMBOL(cpu_data);
arch/sh/kernel/smp.c
253
bogosum += cpu_data[cpu].loops_per_jiffy;
arch/sh/kernel/smp.c
51
struct sh_cpuinfo *c = cpu_data + cpu;
arch/sh/lib/delay.c
40
"r" (cpu_data[raw_smp_processor_id()].loops_per_jiffy * (HZ/4))
arch/sh/mm/cache-sh4.c
74
cpu_data->icache.entry_mask);
arch/sh/mm/cache-sh4.c
78
for (i = 0; i < cpu_data->icache.ways; i++) {
arch/sh/mm/cache-sh4.c
81
icacheaddr += cpu_data->icache.way_incr;
arch/sh/mm/tlb-sh4.c
44
if (cpu_data->flags & CPU_HAS_PTEA) {
arch/sparc/include/asm/delay_32.h
28
#define __udelay_val cpu_data(smp_processor_id()).udelay_val
arch/sparc/include/asm/topology_64.h
46
#define topology_physical_package_id(cpu) (cpu_data(cpu).proc_id)
arch/sparc/include/asm/topology_64.h
47
#define topology_core_id(cpu) (cpu_data(cpu).core_id)
arch/sparc/kernel/cpu.c
348
, cpu_data(0).udelay_val/(500000/HZ),
arch/sparc/kernel/cpu.c
349
(cpu_data(0).udelay_val/(5000/HZ)) % 100,
arch/sparc/kernel/cpu.c
350
cpu_data(0).clock_tick
arch/sparc/kernel/cpu.c
398
, cpu_data(0).clock_tick
arch/sparc/kernel/cpumap.c
108
id = cpu_data(cpu).core_id;
arch/sparc/kernel/cpumap.c
111
id = cpu_data(cpu).proc_id;
arch/sparc/kernel/devices.c
130
cpu_data(0).clock_tick = prom_getintdefault(cpu_node,
arch/sparc/kernel/irq_32.c
204
seq_put_decimal_ull_width(p, " ", cpu_data(j).irq_resched_count, 10);
arch/sparc/kernel/irq_32.c
208
seq_put_decimal_ull_width(p, " ", cpu_data(j).irq_call_count, 10);
arch/sparc/kernel/irq_32.c
213
seq_put_decimal_ull_width(p, " ", cpu_data(j).counter, 10);
arch/sparc/kernel/irq_64.c
308
seq_put_decimal_ull_width(p, " ", cpu_data(j).__nmi_count, 10);
arch/sparc/kernel/leon_smp.c
240
prev = &cpu_data(i).next;
arch/sparc/kernel/mdesc.c
1028
cpu_data(*id).proc_id = proc_id;
arch/sparc/kernel/mdesc.c
1198
c = &cpu_data(cpuid);
arch/sparc/kernel/mdesc.c
889
cpu_data(*id).core_id = core_id;
arch/sparc/kernel/mdesc.c
898
cpu_data(*id).max_cache_id = max_cache_id;
arch/sparc/kernel/mdesc.c
904
cpu_data(*id).sock_id = max_cache_id;
arch/sparc/kernel/mdesc.c
990
cpu_data(*id).sock_id = idx;
arch/sparc/kernel/nmi.c
132
return cpu_data(cpu).__nmi_count;
arch/sparc/kernel/prom_64.c
418
int proc_id = cpu_data(cpu).proc_id;
arch/sparc/kernel/prom_64.c
534
cpu_data(cpuid).clock_tick =
arch/sparc/kernel/prom_64.c
538
cpu_data(cpuid).dcache_size =
arch/sparc/kernel/prom_64.c
541
cpu_data(cpuid).dcache_line_size =
arch/sparc/kernel/prom_64.c
544
cpu_data(cpuid).icache_size =
arch/sparc/kernel/prom_64.c
547
cpu_data(cpuid).icache_line_size =
arch/sparc/kernel/prom_64.c
550
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
552
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
554
if (!cpu_data(cpuid).ecache_size ||
arch/sparc/kernel/prom_64.c
555
!cpu_data(cpuid).ecache_line_size) {
arch/sparc/kernel/prom_64.c
556
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
560
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
565
cpu_data(cpuid).core_id = portid + 1;
arch/sparc/kernel/prom_64.c
566
cpu_data(cpuid).proc_id = portid;
arch/sparc/kernel/prom_64.c
568
cpu_data(cpuid).dcache_size =
arch/sparc/kernel/prom_64.c
570
cpu_data(cpuid).dcache_line_size =
arch/sparc/kernel/prom_64.c
573
cpu_data(cpuid).icache_size =
arch/sparc/kernel/prom_64.c
575
cpu_data(cpuid).icache_line_size =
arch/sparc/kernel/prom_64.c
578
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
581
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
584
cpu_data(cpuid).core_id = 0;
arch/sparc/kernel/prom_64.c
585
cpu_data(cpuid).proc_id = -1;
arch/sparc/kernel/setup_32.c
406
cpu_data(0).udelay_val = loops_per_jiffy;
arch/sparc/kernel/smp_32.c
377
cpu_data(i).udelay_val/(500000/HZ),
arch/sparc/kernel/smp_32.c
378
(cpu_data(i).udelay_val/(5000/HZ))%100);
arch/sparc/kernel/smp_32.c
61
cpu_data(id).udelay_val = loops_per_jiffy;
arch/sparc/kernel/smp_32.c
64
cpu_data(id).clock_tick = prom_getintdefault(cpu_node,
arch/sparc/kernel/smp_32.c
66
cpu_data(id).prom_node = cpu_node;
arch/sparc/kernel/smp_32.c
73
cpu_data(id).mid = mid;
arch/sparc/kernel/smp_32.c
83
bogosum += cpu_data(cpu).udelay_val;
arch/sparc/kernel/smp_64.c
1224
if (cpu_data(i).core_id == 0) {
arch/sparc/kernel/smp_64.c
1230
if (cpu_data(i).core_id ==
arch/sparc/kernel/smp_64.c
1231
cpu_data(j).core_id)
arch/sparc/kernel/smp_64.c
1240
if (cpu_data(i).max_cache_id ==
arch/sparc/kernel/smp_64.c
1241
cpu_data(j).max_cache_id)
arch/sparc/kernel/smp_64.c
1244
if (cpu_data(i).sock_id == cpu_data(j).sock_id)
arch/sparc/kernel/smp_64.c
1253
if (cpu_data(i).proc_id == -1) {
arch/sparc/kernel/smp_64.c
1259
if (cpu_data(i).proc_id ==
arch/sparc/kernel/smp_64.c
1260
cpu_data(j).proc_id)
arch/sparc/kernel/smp_64.c
1337
c = &cpu_data(cpu);
arch/sparc/kernel/smp_64.c
97
i, cpu_data(i).clock_tick);
arch/sparc/kernel/sun4d_smp.c
173
prev = &cpu_data(i).next;
arch/sparc/kernel/sun4m_smp.c
131
prev = &cpu_data(i).next;
arch/sparc/kernel/sysfs.c
176
cpuinfo_sparc *c = &cpu_data(dev->id); \
arch/sparc/kernel/sysfs.c
184
cpuinfo_sparc *c = &cpu_data(dev->id); \
arch/sparc/kernel/time_64.c
648
return cpu_data(cpu).clock_tick;
arch/sparc/kernel/time_64.c
666
ft->clock_tick_ref = cpu_data(cpu).clock_tick;
arch/sparc/kernel/time_64.c
671
cpu_data(cpu).clock_tick =
arch/sparc/kernel/traps_64.c
868
val = cpu_data(i).ecache_size;
arch/sparc/kernel/traps_64.c
875
val = cpu_data(i).ecache_line_size;
arch/x86/events/intel/core.c
5983
struct cpuinfo_x86 *c = &cpu_data(smp_processor_id());
arch/x86/events/intel/core.c
7220
return (cpu >= 0) && is_attr_for_this_pmu(kobj, attr) && cpu_has(&cpu_data(cpu), X86_FEATURE_RTM) ? attr->mode : 0;
arch/x86/events/intel/uncore.c
76
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/include/asm/acpi.h
112
struct cpuinfo_x86 *c = &cpu_data(0);
arch/x86/include/asm/acpi.h
119
struct cpuinfo_x86 *c = &cpu_data(0);
arch/x86/include/asm/topology.h
143
#define topology_logical_package_id(cpu) (cpu_data(cpu).topo.logical_pkg_id)
arch/x86/include/asm/topology.h
144
#define topology_physical_package_id(cpu) (cpu_data(cpu).topo.pkg_id)
arch/x86/include/asm/topology.h
145
#define topology_logical_die_id(cpu) (cpu_data(cpu).topo.logical_die_id)
arch/x86/include/asm/topology.h
146
#define topology_logical_core_id(cpu) (cpu_data(cpu).topo.logical_core_id)
arch/x86/include/asm/topology.h
147
#define topology_die_id(cpu) (cpu_data(cpu).topo.die_id)
arch/x86/include/asm/topology.h
148
#define topology_core_id(cpu) (cpu_data(cpu).topo.core_id)
arch/x86/include/asm/topology.h
149
#define topology_ppin(cpu) (cpu_data(cpu).ppin)
arch/x86/include/asm/topology.h
151
#define topology_amd_node_id(cpu) (cpu_data(cpu).topo.amd_node_id)
arch/x86/include/asm/topology.h
198
#define topology_cluster_id(cpu) (cpu_data(cpu).topo.l2c_id)
arch/x86/kernel/acpi/cppc.c
244
enum x86_topology_cpu_type core_type = get_topology_cpu_type(&cpu_data(cpu));
arch/x86/kernel/acpi/cstate.c
174
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/acpi/cstate.c
35
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/alternative.c
2228
clear_cpu_cap(&cpu_data(0), X86_FEATURE_UP);
arch/x86/kernel/amd_nb.c
181
return (mask >> (4 * cpu_data(cpu).topo.core_id)) & 0xf;
arch/x86/kernel/amd_nb.c
207
cuid = cpu_data(cpu).topo.core_id;
arch/x86/kernel/cpu/amd.c
1301
struct cpuinfo_x86 *c = &cpu_data(smp_processor_id());
arch/x86/kernel/cpu/cacheinfo.c
516
apicid = cpu_data(cpu).topo.apicid;
arch/x86/kernel/cpu/cacheinfo.c
525
apicid = cpu_data(i).topo.apicid;
arch/x86/kernel/cpu/cacheinfo.c
532
apicid = cpu_data(sibling).topo.apicid;
arch/x86/kernel/cpu/cacheinfo.c
551
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/cacheinfo.c
571
if (cpu_data(i).topo.apicid >> index_msb == c->topo.apicid >> index_msb) {
arch/x86/kernel/cpu/cacheinfo.c
615
u32 apicid = cpu_data(cpu).topo.apicid;
arch/x86/kernel/cpu/common.c
2171
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/mce/apei.c
101
if (cpu_data(cpu).topo.initial_apicid == lapic_id) {
arch/x86/kernel/cpu/mce/core.c
134
m->apicid = cpu_data(cpu).topo.initial_apicid;
arch/x86/kernel/cpu/mce/core.c
135
m->microcode = cpu_data(cpu).microcode;
arch/x86/kernel/cpu/microcode/amd.c
1180
c = &cpu_data(cpu);
arch/x86/kernel/cpu/microcode/amd.c
1242
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/microcode/amd.c
972
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/microcode/core.c
855
cpu_data(cpu).microcode = uci->cpu_sig.rev;
arch/x86/kernel/cpu/microcode/intel.c
798
cpu_data(cpu).microcode = uci->cpu_sig.rev;
arch/x86/kernel/cpu/microcode/intel.c
910
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/microcode/intel.c
932
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/cpu/proc.c
163
return &cpu_data(*pos);
arch/x86/kernel/cpuid.c
109
c = &cpu_data(cpu);
arch/x86/kernel/msr.c
217
c = &cpu_data(cpu);
arch/x86/kernel/smpboot.c
1226
print_cpu_info(&cpu_data(0));
arch/x86/kernel/smpboot.c
1297
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/smpboot.c
1305
cpu_data(sibling).booted_cores--;
arch/x86/kernel/smpboot.c
1314
cpu_data(sibling).smt_active = false;
arch/x86/kernel/smpboot.c
223
cpu_data(smp_processor_id()).loops_per_jiffy = loops_per_jiffy;
arch/x86/kernel/smpboot.c
678
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kernel/smpboot.c
695
o = &cpu_data(i);
arch/x86/kernel/smpboot.c
718
cpu_data(i).smt_active = threads > 1;
arch/x86/kernel/smpboot.c
725
o = &cpu_data(i);
arch/x86/kernel/smpboot.c
746
cpu_data(i).booted_cores++;
arch/x86/kernel/smpboot.c
748
c->booted_cores = cpu_data(i).booted_cores;
arch/x86/kernel/smpboot.c
774
bogosum += cpu_data(cpu).loops_per_jiffy;
arch/x86/kernel/tsc.c
1586
int constant_tsc = cpu_has(&cpu_data(cpu), X86_FEATURE_CONSTANT_TSC);
arch/x86/kernel/tsc.c
1594
return cpu_data(0).loops_per_jiffy;
arch/x86/kernel/tsc.c
1609
return cpu_data(sibling).loops_per_jiffy;
arch/x86/kernel/tsc.c
961
cpu_data(0).loops_per_jiffy = cpufreq_scale(cpu_data(0).loops_per_jiffy,
arch/x86/kvm/svm/svm.c
426
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/kvm/x86.c
10022
struct cpuinfo_x86 *c = &cpu_data(cpu);
arch/x86/power/cpu.c
283
c = &cpu_data(smp_processor_id());
arch/x86/xen/apic.c
108
return cpu_data(cpu).topo.apicid;
arch/xtensa/kernel/setup.c
374
static DEFINE_PER_CPU(struct cpu, cpu_data);
arch/xtensa/kernel/setup.c
381
struct cpu *cpu = &per_cpu(cpu_data, i);
drivers/acpi/cppc_acpi.c
498
int acpi_get_psd_map(unsigned int cpu, struct cppc_cpudata *cpu_data)
drivers/acpi/cppc_acpi.c
514
cpumask_set_cpu(cpu, cpu_data->shared_cpu_map);
drivers/acpi/cppc_acpi.c
521
cpu_data->shared_type = CPUFREQ_SHARED_TYPE_ALL;
drivers/acpi/cppc_acpi.c
523
cpu_data->shared_type = CPUFREQ_SHARED_TYPE_HW;
drivers/acpi/cppc_acpi.c
525
cpu_data->shared_type = CPUFREQ_SHARED_TYPE_ANY;
drivers/acpi/cppc_acpi.c
546
cpumask_set_cpu(i, cpu_data->shared_cpu_map);
drivers/acpi/cppc_acpi.c
553
cpumask_clear(cpu_data->shared_cpu_map);
drivers/acpi/cppc_acpi.c
554
cpumask_set_cpu(cpu, cpu_data->shared_cpu_map);
drivers/acpi/cppc_acpi.c
555
cpu_data->shared_type = CPUFREQ_SHARED_TYPE_NONE;
drivers/acpi/processor_idle.c
138
if (cpu_has(&cpu_data(pr->id), X86_FEATURE_ARAT))
drivers/char/hw_random/via-rng.c
131
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/acpi-cpufreq.c
184
struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
drivers/cpufreq/acpi-cpufreq.c
191
struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
drivers/cpufreq/acpi-cpufreq.c
684
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/cpufreq/amd-pstate.c
1725
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/amd_freq_sensitivity.c
37
static DEFINE_PER_CPU(struct cpu_data_t, cpu_data);
drivers/cpufreq/amd_freq_sensitivity.c
46
struct cpu_data_t *data = &per_cpu(cpu_data, policy->cpu);
drivers/cpufreq/cppc_cpufreq.c
159
cppc_fi->cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
294
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
299
cpu_data->perf_ctrls.desired_perf =
drivers/cpufreq/cppc_cpufreq.c
300
cppc_khz_to_perf(&cpu_data->perf_caps, target_freq);
drivers/cpufreq/cppc_cpufreq.c
305
ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
drivers/cpufreq/cppc_cpufreq.c
318
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
323
desired_perf = cppc_khz_to_perf(&cpu_data->perf_caps, target_freq);
drivers/cpufreq/cppc_cpufreq.c
324
cpu_data->perf_ctrls.desired_perf = desired_perf;
drivers/cpufreq/cppc_cpufreq.c
325
ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
drivers/cpufreq/cppc_cpufreq.c
399
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
402
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
403
perf_caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
432
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
438
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
439
perf_caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
47
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
500
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
508
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
509
perf_caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
523
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
527
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
530
cpu_data->shared_cpu_map, 0);
drivers/cpufreq/cppc_cpufreq.c
575
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
578
cpu_data = kzalloc_obj(struct cppc_cpudata);
drivers/cpufreq/cppc_cpufreq.c
579
if (!cpu_data)
drivers/cpufreq/cppc_cpufreq.c
582
if (!zalloc_cpumask_var(&cpu_data->shared_cpu_map, GFP_KERNEL))
drivers/cpufreq/cppc_cpufreq.c
585
ret = acpi_get_psd_map(cpu, cpu_data);
drivers/cpufreq/cppc_cpufreq.c
591
ret = cppc_get_perf_caps(cpu, &cpu_data->perf_caps);
drivers/cpufreq/cppc_cpufreq.c
597
return cpu_data;
drivers/cpufreq/cppc_cpufreq.c
600
free_cpumask_var(cpu_data->shared_cpu_map);
drivers/cpufreq/cppc_cpufreq.c
602
kfree(cpu_data);
drivers/cpufreq/cppc_cpufreq.c
609
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
611
free_cpumask_var(cpu_data->shared_cpu_map);
drivers/cpufreq/cppc_cpufreq.c
612
kfree(cpu_data);
drivers/cpufreq/cppc_cpufreq.c
619
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
623
cpu_data = cppc_cpufreq_get_cpu_data(cpu);
drivers/cpufreq/cppc_cpufreq.c
624
if (!cpu_data) {
drivers/cpufreq/cppc_cpufreq.c
628
caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
629
policy->driver_data = cpu_data;
drivers/cpufreq/cppc_cpufreq.c
648
policy->shared_type = cpu_data->shared_type;
drivers/cpufreq/cppc_cpufreq.c
661
cpumask_copy(policy->cpus, cpu_data->shared_cpu_map);
drivers/cpufreq/cppc_cpufreq.c
682
cpu_data->perf_ctrls.desired_perf = caps->highest_perf;
drivers/cpufreq/cppc_cpufreq.c
684
ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
drivers/cpufreq/cppc_cpufreq.c
701
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
702
struct cppc_perf_caps *caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
708
cpu_data->perf_ctrls.desired_perf = caps->lowest_perf;
drivers/cpufreq/cppc_cpufreq.c
71
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
710
ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
drivers/cpufreq/cppc_cpufreq.c
75
cpu_data = cppc_fi->cpu_data;
drivers/cpufreq/cppc_cpufreq.c
768
struct cppc_cpudata *cpu_data;
drivers/cpufreq/cppc_cpufreq.c
775
cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
790
return cppc_perf_to_khz(&cpu_data->perf_caps, delivered_perf);
drivers/cpufreq/cppc_cpufreq.c
801
delivered_perf = cpu_data->perf_ctrls.desired_perf;
drivers/cpufreq/cppc_cpufreq.c
803
return cppc_perf_to_khz(&cpu_data->perf_caps, delivered_perf);
drivers/cpufreq/cppc_cpufreq.c
808
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
809
struct cppc_perf_caps *caps = &cpu_data->perf_caps;
drivers/cpufreq/cppc_cpufreq.c
827
struct cppc_cpudata *cpu_data = policy->driver_data;
drivers/cpufreq/cppc_cpufreq.c
829
return cpufreq_show_cpus(cpu_data->shared_cpu_map, buf);
drivers/cpufreq/cppc_cpufreq.c
89
local_freq_scale = div64_u64(perf, cpu_data->perf_caps.highest_perf);
drivers/cpufreq/e_powersaver.c
182
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/elanfreq.c
145
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/intel_pstate.c
1261
struct cpudata *cpu_data = all_cpu_data[cpu];
drivers/cpufreq/intel_pstate.c
1266
max = cpu_data->max_perf_ratio;
drivers/cpufreq/intel_pstate.c
1267
min = cpu_data->min_perf_ratio;
drivers/cpufreq/intel_pstate.c
1269
if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/intel_pstate.c
1280
if (cpu_data->epp_policy == cpu_data->policy)
drivers/cpufreq/intel_pstate.c
1283
cpu_data->epp_policy = cpu_data->policy;
drivers/cpufreq/intel_pstate.c
1285
if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) {
drivers/cpufreq/intel_pstate.c
1286
epp = intel_pstate_get_epp(cpu_data, value);
drivers/cpufreq/intel_pstate.c
1287
cpu_data->epp_powersave = epp;
drivers/cpufreq/intel_pstate.c
1295
if (cpu_data->epp_powersave < 0)
drivers/cpufreq/intel_pstate.c
1305
epp = intel_pstate_get_epp(cpu_data, value);
drivers/cpufreq/intel_pstate.c
1309
epp = cpu_data->epp_powersave;
drivers/cpufreq/intel_pstate.c
1317
WRITE_ONCE(cpu_data->hwp_req_cached, value);
drivers/cpufreq/intel_pstate.c
2806
struct cpudata *cpu_data = all_cpu_data[cpu];
drivers/cpufreq/intel_pstate.c
2808
if (!cpu_data->update_util_set)
drivers/cpufreq/intel_pstate.c
2812
cpu_data->update_util_set = false;
drivers/cpufreq/intel_pstate.c
625
static s16 intel_pstate_get_epp(struct cpudata *cpu_data, u64 hwp_req_data)
drivers/cpufreq/intel_pstate.c
635
epp = rdmsrq_on_cpu(cpu_data->cpu, MSR_HWP_REQUEST,
drivers/cpufreq/intel_pstate.c
682
static int intel_pstate_get_energy_pref_index(struct cpudata *cpu_data, int *raw_epp)
drivers/cpufreq/intel_pstate.c
688
epp = intel_pstate_get_epp(cpu_data, 0);
drivers/cpufreq/intel_pstate.c
746
static int intel_pstate_set_energy_pref_index(struct cpudata *cpu_data,
drivers/cpufreq/intel_pstate.c
754
epp = cpu_data->epp_default;
drivers/cpufreq/intel_pstate.c
767
if (epp > 0 && cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE)
drivers/cpufreq/intel_pstate.c
770
ret = intel_pstate_set_epp(cpu_data, epp);
drivers/cpufreq/intel_pstate.c
865
struct cpudata *cpu_data = all_cpu_data[policy->cpu];
drivers/cpufreq/intel_pstate.c
868
preference = intel_pstate_get_energy_pref_index(cpu_data, &raw_epp);
drivers/cpufreq/intel_pstate.c
919
return cpu_data(cpu).topo.intel_type;
drivers/cpufreq/longhaul.c
761
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/longhaul.c
918
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/longrun.c
166
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/loongson3_cpufreq.c
180
unsigned int package = cpu_data[cpu].package;
drivers/cpufreq/loongson3_cpufreq.c
233
ret = do_service_request(cpu_data[policy->cpu].core,
drivers/cpufreq/p4-clockmod.c
157
struct cpuinfo_x86 *c = &cpu_data(policy->cpu);
drivers/cpufreq/powernow-k7.c
119
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/qoriq-cpufreq.c
163
struct cpu_data *data;
drivers/cpufreq/qoriq-cpufreq.c
230
struct cpu_data *data = policy->driver_data;
drivers/cpufreq/qoriq-cpufreq.c
242
struct cpu_data *data = policy->driver_data;
drivers/cpufreq/sc520_freq.c
73
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/speedstep-centrino.c
236
struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
drivers/cpufreq/speedstep-centrino.c
347
struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu);
drivers/cpufreq/speedstep-lib.c
256
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/cpufreq/tegra194-cpufreq.c
128
*ndiv = readl(data->cpu_data[cpu].freq_core_reg) & NDIV_MASK;
drivers/cpufreq/tegra194-cpufreq.c
139
writel(ndiv, data->cpu_data[cpu].freq_core_reg);
drivers/cpufreq/tegra194-cpufreq.c
157
actmon_reg = CORE_ACTMON_CNTR_REG(data, data->cpu_data[c->cpu].clusterid,
drivers/cpufreq/tegra194-cpufreq.c
158
data->cpu_data[c->cpu].cpuid);
drivers/cpufreq/tegra194-cpufreq.c
388
u32 clusterid = data->cpu_data[cpu].clusterid;
drivers/cpufreq/tegra194-cpufreq.c
398
ret = data->soc->ops->get_cpu_ndiv(cpu, data->cpu_data[cpu].cpuid, clusterid, &ndiv);
drivers/cpufreq/tegra194-cpufreq.c
504
u32 clusterid = data->cpu_data[policy->cpu].clusterid;
drivers/cpufreq/tegra194-cpufreq.c
700
data->cpu_data[cpu].cpuid = cpuid;
drivers/cpufreq/tegra194-cpufreq.c
701
data->cpu_data[cpu].clusterid = clusterid;
drivers/cpufreq/tegra194-cpufreq.c
702
data->cpu_data[cpu].freq_core_reg = SCRATCH_FREQ_CORE_REG(data, mpidr_id);
drivers/cpufreq/tegra194-cpufreq.c
741
data->cpu_data = devm_kcalloc(&pdev->dev, data->soc->num_clusters *
drivers/cpufreq/tegra194-cpufreq.c
743
sizeof(*data->cpu_data), GFP_KERNEL);
drivers/cpufreq/tegra194-cpufreq.c
744
if (!data->cpu_data)
drivers/cpufreq/tegra194-cpufreq.c
79
struct tegra_cpu_data *cpu_data;
drivers/cpuidle/governors/teo.c
175
struct teo_cpu *cpu_data = this_cpu_ptr(&teo_cpus);
drivers/cpuidle/governors/teo.c
180
teo_decay(&cpu_data->short_idles);
drivers/cpuidle/governors/teo.c
200
cpu_data->short_idles += PULSE;
drivers/cpuidle/governors/teo.c
203
cpu_data->short_idles += PULSE;
drivers/cpuidle/governors/teo.c
213
struct teo_bin *bin = &cpu_data->state_bins[i];
drivers/cpuidle/governors/teo.c
222
if (target_residency_ns <= cpu_data->sleep_length_ns) {
drivers/cpuidle/governors/teo.c
229
cpu_data->total = total + PULSE;
drivers/cpuidle/governors/teo.c
231
teo_decay(&cpu_data->tick_intercepts);
drivers/cpuidle/governors/teo.c
233
teo_decay(&cpu_data->total_tick);
drivers/cpuidle/governors/teo.c
234
if (cpu_data->tick_wakeup) {
drivers/cpuidle/governors/teo.c
235
cpu_data->total_tick += PULSE;
drivers/cpuidle/governors/teo.c
241
if (3 * cpu_data->total_tick > 2 * cpu_data->total) {
drivers/cpuidle/governors/teo.c
242
cpu_data->state_bins[drv->state_count-1].hits += PULSE;
drivers/cpuidle/governors/teo.c
252
if (3 * cpu_data->tick_intercepts < 2 * total) {
drivers/cpuidle/governors/teo.c
253
cpu_data->state_bins[idx_timer].hits += PULSE;
drivers/cpuidle/governors/teo.c
269
cpu_data->sleep_length_ns - measured_ns < lat_ns / 2) {
drivers/cpuidle/governors/teo.c
270
cpu_data->state_bins[idx_timer].hits += PULSE;
drivers/cpuidle/governors/teo.c
272
cpu_data->state_bins[idx_duration].intercepts += PULSE;
drivers/cpuidle/governors/teo.c
274
cpu_data->tick_intercepts += PULSE;
drivers/cpuidle/governors/teo.c
311
struct teo_cpu *cpu_data = this_cpu_ptr(&teo_cpus);
drivers/cpuidle/governors/teo.c
339
cpu_data->sleep_length_ns = KTIME_MAX;
drivers/cpuidle/governors/teo.c
351
struct teo_bin *prev_bin = &cpu_data->state_bins[i-1];
drivers/cpuidle/governors/teo.c
409
if (2 * idx_intercept_sum > cpu_data->total - idx_hit_sum) {
drivers/cpuidle/governors/teo.c
454
intercept_sum += cpu_data->state_bins[i].intercepts;
drivers/cpuidle/governors/teo.c
498
(2 * cpu_data->short_idles >= cpu_data->total ||
drivers/cpuidle/governors/teo.c
503
cpu_data->sleep_length_ns = duration_ns;
drivers/cpuidle/governors/teo.c
521
3 * cpu_data->tick_intercepts >= 2 * cpu_data->total)
drivers/cpuidle/governors/teo.c
555
struct teo_cpu *cpu_data = this_cpu_ptr(&teo_cpus);
drivers/cpuidle/governors/teo.c
557
cpu_data->tick_wakeup = tick_nohz_idle_got_tick();
drivers/cpuidle/governors/teo.c
570
struct teo_cpu *cpu_data = per_cpu_ptr(&teo_cpus, dev->cpu);
drivers/cpuidle/governors/teo.c
572
memset(cpu_data, 0, sizeof(*cpu_data));
drivers/crypto/padlock-aes.c
486
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/crypto/padlock-sha.c
328
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/crypto/padlock-sha.c
374
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/edac/e752x_edac.c
1213
char *cpu_id = cpu_data(0).x86_model_id;
drivers/edac/skx_common.c
280
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
2008
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
2030
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1830
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2341
return cpu_data(first_cpu_of_numa_node).topo.apicid;
drivers/hwmon/coretemp.c
279
struct cpuinfo_x86 *c = &cpu_data(tdata->cpu);
drivers/hwmon/coretemp.c
459
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/hwmon/coretemp.c
562
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/hwmon/coretemp.c
669
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/hwmon/hwmon-vid.c
281
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/hwmon/via-cputemp.c
113
struct cpuinfo_x86 *c = &cpu_data(pdev->id);
drivers/irqchip/irq-mips-cpu.c
103
settc(cpu_vpe_id(&cpu_data[cpu]));
drivers/irqchip/irq-mips-gic.c
141
cl = cpu_cluster(&cpu_data[cpu]);
drivers/irqchip/irq-mips-gic.c
389
old_cl = cpu_cluster(&cpu_data[old_cpu]);
drivers/irqchip/irq-mips-gic.c
390
cl = cpu_cluster(&cpu_data[cpu]);
drivers/pci/p2pdma.c
521
struct cpuinfo_x86 *c = &cpu_data(0);
drivers/perf/arm_pmu_acpi.c
345
unsigned long cpu_cpuid = per_cpu(cpu_data, cpu).reg_midr;
drivers/perf/arm_pmu_acpi.c
397
cpuid = per_cpu(cpu_data, cpu).reg_midr;
drivers/platform/x86/amd/hfi/hfi.c
128
struct cpuinfo_x86 *info = &cpu_data(cpu_index);
drivers/thermal/intel/therm_throt.c
472
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/thermal/intel/x86_pkg_temp_thermal.c
467
struct cpuinfo_x86 *c = &cpu_data(cpu);
drivers/video/fbdev/geode/video_gx.c
126
if (cpu_data(0).x86_stepping == 1) {
drivers/virt/acrn/hsm.c
450
lapicid = cpu_data(cpu).topo.apicid;
include/acpi/cppc_acpi.h
163
extern int acpi_get_psd_map(unsigned int cpu, struct cppc_cpudata *cpu_data);
kernel/trace/trace_functions_graph.c
1032
struct fgraph_cpu_data *cpu_data;
kernel/trace/trace_functions_graph.c
1035
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_functions_graph.c
1036
cpu_data->depth = call->depth;
kernel/trace/trace_functions_graph.c
1041
cpu_data->enter_funcs[call->depth] = call->func;
kernel/trace/trace_functions_graph.c
1154
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
1200
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
1307
struct fgraph_cpu_data *cpu_data;
kernel/trace/trace_functions_graph.c
1310
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_functions_graph.c
1317
cpu_data->depth = trace->depth - 1;
kernel/trace/trace_functions_graph.c
1321
if (cpu_data->enter_funcs[trace->depth] != trace->func)
kernel/trace/trace_functions_graph.c
1323
cpu_data->enter_funcs[trace->depth] = 0;
kernel/trace/trace_functions_graph.c
1382
depth = per_cpu_ptr(data->cpu_data, iter->cpu)->depth;
kernel/trace/trace_functions_graph.c
1448
if (data && per_cpu_ptr(data->cpu_data, cpu)->ignore) {
kernel/trace/trace_functions_graph.c
1449
per_cpu_ptr(data->cpu_data, cpu)->ignore = 0;
kernel/trace/trace_functions_graph.c
1462
per_cpu_ptr(data->cpu_data, iter->cpu)->ignore = 1;
kernel/trace/trace_functions_graph.c
1627
data->cpu_data = alloc_percpu_gfp(struct fgraph_cpu_data, gfpflags);
kernel/trace/trace_functions_graph.c
1628
if (!data->cpu_data)
kernel/trace/trace_functions_graph.c
1632
pid_t *pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
kernel/trace/trace_functions_graph.c
1633
int *depth = &(per_cpu_ptr(data->cpu_data, cpu)->depth);
kernel/trace/trace_functions_graph.c
1634
int *ignore = &(per_cpu_ptr(data->cpu_data, cpu)->ignore);
kernel/trace/trace_functions_graph.c
1635
int *depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
kernel/trace/trace_functions_graph.c
1658
free_percpu(data->cpu_data);
kernel/trace/trace_functions_graph.c
46
struct fgraph_cpu_data __percpu *cpu_data;
kernel/trace/trace_functions_graph.c
610
last_pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
kernel/trace/trace_functions_graph.c
969
struct fgraph_cpu_data *cpu_data;
kernel/trace/trace_functions_graph.c
971
cpu_data = per_cpu_ptr(data->cpu_data, cpu);
kernel/trace/trace_functions_graph.c
978
cpu_data->depth = call->depth - 1;
kernel/trace/trace_functions_graph.c
983
cpu_data->enter_funcs[call->depth] = 0;
tools/perf/builtin-sched.c
3941
struct perf_record_schedstat_cpu *cpu_data;
tools/perf/builtin-sched.c
3952
struct perf_record_schedstat_cpu *before = cpu_second_pass->cpu_data;
tools/perf/builtin-sched.c
3953
struct perf_record_schedstat_cpu *after = after_workload->cpu_data;
tools/perf/builtin-sched.c
3954
__u16 version = after_workload->cpu_data->version;
tools/perf/builtin-sched.c
4124
struct perf_record_schedstat_cpu *summary_cs = summary_cpu->cpu_data,
tools/perf/builtin-sched.c
4125
*temp_cs = cptr->cpu_data;
tools/perf/builtin-sched.c
4134
if (cptr->cpu_data->version == 15) {
tools/perf/builtin-sched.c
4136
} else if (cptr->cpu_data->version == 16) {
tools/perf/builtin-sched.c
4138
} else if (cptr->cpu_data->version == 17) {
tools/perf/builtin-sched.c
4187
summary_head->cpu_data = zalloc(sizeof(*cs));
tools/perf/builtin-sched.c
4188
memcpy(summary_head->cpu_data, cptr->cpu_data, sizeof(*cs));
tools/perf/builtin-sched.c
4253
jiffies1 = cptr1->cpu_data->timestamp;
tools/perf/builtin-sched.c
4257
jiffies2 = cptr2->cpu_data->timestamp;
tools/perf/builtin-sched.c
4271
cs1 = cptr1->cpu_data;
tools/perf/builtin-sched.c
4274
cs2 = cptr2->cpu_data;
tools/perf/builtin-sched.c
4415
temp->cpu_data = zalloc(sizeof(*temp->cpu_data));
tools/perf/builtin-sched.c
4416
if (!temp->cpu_data)
tools/perf/builtin-sched.c
4419
memcpy(temp->cpu_data, &event->schedstat_cpu, sizeof(*temp->cpu_data));
tools/perf/builtin-sched.c
4421
if (!list_empty(&cpu_head) && temp->cpu_data->cpu == initial_cpu)
tools/perf/builtin-sched.c
4426
initial_cpu = temp->cpu_data->cpu;
tools/perf/builtin-sched.c
4431
if (temp->cpu_data->cpu == initial_cpu) {
tools/perf/builtin-sched.c
4434
cpu_second_pass->cpu_data->timestamp =
tools/perf/builtin-sched.c
4435
temp->cpu_data->timestamp - cpu_second_pass->cpu_data->timestamp;
tools/perf/util/bpf_lock_contention.c
497
struct tstamp_data ts_data, *cpu_data;
tools/perf/util/bpf_lock_contention.c
525
cpu_data = calloc(total_cpus, sizeof(*cpu_data));
tools/perf/util/bpf_lock_contention.c
526
if (cpu_data == NULL)
tools/perf/util/bpf_lock_contention.c
531
if (bpf_map_lookup_elem(ts_fd, &key, cpu_data) < 0)
tools/perf/util/bpf_lock_contention.c
535
if (cpu_data[i].lock == 0)
tools/perf/util/bpf_lock_contention.c
539
&cpu_data[i]);
tools/perf/util/bpf_lock_contention.c
545
free(cpu_data);
tools/perf/util/cs-etm.c
326
u64 *cpu_data;
tools/perf/util/cs-etm.c
329
cpu_data = get_cpu_data(etm, cpu);
tools/perf/util/cs-etm.c
330
if (cpu_data == NULL)
tools/perf/util/cs-etm.c
333
err = cs_etm__map_trace_id_v0(etm, trace_chan_id, cpu_data);
tools/perf/util/cs-etm.c
341
return cs_etm__metadata_set_trace_id(trace_chan_id, cpu_data);
tools/perf/util/cs-etm.c
349
u64 *cpu_data;
tools/perf/util/cs-etm.c
390
cpu_data = get_cpu_data(etm, cpu);
tools/perf/util/cs-etm.c
391
ret = cs_etm__insert_trace_id_node(etmq, trace_id, cpu_data);
tools/perf/util/cs-etm.c
395
ret = cs_etm__metadata_set_trace_id(trace_id, cpu_data);
tools/testing/selftests/bpf/prog_tests/perf_buffer.c
19
int cpu_data = *(int *)data, duration = 0;
tools/testing/selftests/bpf/prog_tests/perf_buffer.c
22
if (cpu_data != cpu)
tools/testing/selftests/bpf/prog_tests/perf_buffer.c
23
CHECK(cpu_data != cpu, "check_cpu_data",
tools/testing/selftests/bpf/prog_tests/perf_buffer.c
24
"cpu_data %d != cpu %d\n", cpu_data, cpu);
tools/tracing/rtla/src/osnoise_top.c
101
update_max(&cpu_data->max_noise, &val);
tools/tracing/rtla/src/osnoise_top.c
102
update_sum(&cpu_data->sum_noise, &val);
tools/tracing/rtla/src/osnoise_top.c
105
update_max(&cpu_data->max_sample, &val);
tools/tracing/rtla/src/osnoise_top.c
108
update_sum(&cpu_data->hw_count, &val);
tools/tracing/rtla/src/osnoise_top.c
111
update_sum(&cpu_data->nmi_count, &val);
tools/tracing/rtla/src/osnoise_top.c
114
update_sum(&cpu_data->irq_count, &val);
tools/tracing/rtla/src/osnoise_top.c
117
update_sum(&cpu_data->softirq_count, &val);
tools/tracing/rtla/src/osnoise_top.c
120
update_sum(&cpu_data->thread_count, &val);
tools/tracing/rtla/src/osnoise_top.c
193
struct osnoise_top_cpu *cpu_data;
tools/tracing/rtla/src/osnoise_top.c
199
cpu_data = &data->cpu_data[cpu];
tools/tracing/rtla/src/osnoise_top.c
201
if (!cpu_data->sum_runtime)
tools/tracing/rtla/src/osnoise_top.c
204
percentage = ((cpu_data->sum_runtime - cpu_data->sum_noise) * 10000000)
tools/tracing/rtla/src/osnoise_top.c
205
/ cpu_data->sum_runtime;
tools/tracing/rtla/src/osnoise_top.c
209
trace_seq_printf(s, "%3d #%-6d %12llu ", cpu, cpu_data->sum_cycles, cpu_data->sum_runtime);
tools/tracing/rtla/src/osnoise_top.c
210
trace_seq_printf(s, "%12llu ", cpu_data->sum_noise);
tools/tracing/rtla/src/osnoise_top.c
212
trace_seq_printf(s, "%12llu %12llu", cpu_data->max_noise, cpu_data->max_sample);
tools/tracing/rtla/src/osnoise_top.c
214
trace_seq_printf(s, "%12llu ", cpu_data->hw_count);
tools/tracing/rtla/src/osnoise_top.c
215
trace_seq_printf(s, "%12llu ", cpu_data->nmi_count);
tools/tracing/rtla/src/osnoise_top.c
222
trace_seq_printf(s, "%12llu ", cpu_data->irq_count);
tools/tracing/rtla/src/osnoise_top.c
223
trace_seq_printf(s, "%12llu ", cpu_data->softirq_count);
tools/tracing/rtla/src/osnoise_top.c
224
trace_seq_printf(s, "%12llu\n", cpu_data->thread_count);
tools/tracing/rtla/src/osnoise_top.c
33
struct osnoise_top_cpu *cpu_data;
tools/tracing/rtla/src/osnoise_top.c
42
free(data->cpu_data);
tools/tracing/rtla/src/osnoise_top.c
65
data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus);
tools/tracing/rtla/src/osnoise_top.c
66
if (!data->cpu_data)
tools/tracing/rtla/src/osnoise_top.c
86
struct osnoise_top_cpu *cpu_data;
tools/tracing/rtla/src/osnoise_top.c
93
cpu_data = &data->cpu_data[cpu];
tools/tracing/rtla/src/osnoise_top.c
95
cpu_data->sum_cycles++;
tools/tracing/rtla/src/osnoise_top.c
98
update_sum(&cpu_data->sum_runtime, &val);
tools/tracing/rtla/src/timerlat_hist.c
471
struct timerlat_hist_cpu *cpu_data;
tools/tracing/rtla/src/timerlat_hist.c
488
cpu_data = &data->hist[cpu];
tools/tracing/rtla/src/timerlat_hist.c
490
sum.irq_count += cpu_data->irq_count;
tools/tracing/rtla/src/timerlat_hist.c
491
update_min(&sum.min_irq, &cpu_data->min_irq);
tools/tracing/rtla/src/timerlat_hist.c
492
update_sum(&sum.sum_irq, &cpu_data->sum_irq);
tools/tracing/rtla/src/timerlat_hist.c
493
update_max(&sum.max_irq, &cpu_data->max_irq);
tools/tracing/rtla/src/timerlat_hist.c
495
sum.thread_count += cpu_data->thread_count;
tools/tracing/rtla/src/timerlat_hist.c
496
update_min(&sum.min_thread, &cpu_data->min_thread);
tools/tracing/rtla/src/timerlat_hist.c
497
update_sum(&sum.sum_thread, &cpu_data->sum_thread);
tools/tracing/rtla/src/timerlat_hist.c
498
update_max(&sum.max_thread, &cpu_data->max_thread);
tools/tracing/rtla/src/timerlat_hist.c
500
sum.user_count += cpu_data->user_count;
tools/tracing/rtla/src/timerlat_hist.c
501
update_min(&sum.min_user, &cpu_data->min_user);
tools/tracing/rtla/src/timerlat_hist.c
502
update_sum(&sum.sum_user, &cpu_data->sum_user);
tools/tracing/rtla/src/timerlat_hist.c
503
update_max(&sum.max_user, &cpu_data->max_user);
tools/tracing/rtla/src/timerlat_top.c
108
struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu];
tools/tracing/rtla/src/timerlat_top.c
110
sum->irq_count += cpu_data->irq_count;
tools/tracing/rtla/src/timerlat_top.c
111
update_min(&sum->min_irq, &cpu_data->min_irq);
tools/tracing/rtla/src/timerlat_top.c
112
update_sum(&sum->sum_irq, &cpu_data->sum_irq);
tools/tracing/rtla/src/timerlat_top.c
113
update_max(&sum->max_irq, &cpu_data->max_irq);
tools/tracing/rtla/src/timerlat_top.c
115
sum->thread_count += cpu_data->thread_count;
tools/tracing/rtla/src/timerlat_top.c
116
update_min(&sum->min_thread, &cpu_data->min_thread);
tools/tracing/rtla/src/timerlat_top.c
117
update_sum(&sum->sum_thread, &cpu_data->sum_thread);
tools/tracing/rtla/src/timerlat_top.c
118
update_max(&sum->max_thread, &cpu_data->max_thread);
tools/tracing/rtla/src/timerlat_top.c
120
sum->user_count += cpu_data->user_count;
tools/tracing/rtla/src/timerlat_top.c
121
update_min(&sum->min_user, &cpu_data->min_user);
tools/tracing/rtla/src/timerlat_top.c
122
update_sum(&sum->sum_user, &cpu_data->sum_user);
tools/tracing/rtla/src/timerlat_top.c
123
update_max(&sum->max_user, &cpu_data->max_user);
tools/tracing/rtla/src/timerlat_top.c
136
struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu];
tools/tracing/rtla/src/timerlat_top.c
142
cpu_data->irq_count++;
tools/tracing/rtla/src/timerlat_top.c
143
cpu_data->cur_irq = latency;
tools/tracing/rtla/src/timerlat_top.c
144
update_min(&cpu_data->min_irq, &latency);
tools/tracing/rtla/src/timerlat_top.c
145
update_sum(&cpu_data->sum_irq, &latency);
tools/tracing/rtla/src/timerlat_top.c
146
update_max(&cpu_data->max_irq, &latency);
tools/tracing/rtla/src/timerlat_top.c
148
cpu_data->thread_count++;
tools/tracing/rtla/src/timerlat_top.c
149
cpu_data->cur_thread = latency;
tools/tracing/rtla/src/timerlat_top.c
150
update_min(&cpu_data->min_thread, &latency);
tools/tracing/rtla/src/timerlat_top.c
151
update_sum(&cpu_data->sum_thread, &latency);
tools/tracing/rtla/src/timerlat_top.c
152
update_max(&cpu_data->max_thread, &latency);
tools/tracing/rtla/src/timerlat_top.c
154
cpu_data->user_count++;
tools/tracing/rtla/src/timerlat_top.c
155
cpu_data->cur_user = latency;
tools/tracing/rtla/src/timerlat_top.c
156
update_min(&cpu_data->min_user, &latency);
tools/tracing/rtla/src/timerlat_top.c
157
update_sum(&cpu_data->sum_user, &latency);
tools/tracing/rtla/src/timerlat_top.c
158
update_max(&cpu_data->max_user, &latency);
tools/tracing/rtla/src/timerlat_top.c
204
data->cpu_data[i].cur_irq = value_irq[i];
tools/tracing/rtla/src/timerlat_top.c
205
data->cpu_data[i].cur_thread = value_thread[i];
tools/tracing/rtla/src/timerlat_top.c
206
data->cpu_data[i].cur_user = value_user[i];
tools/tracing/rtla/src/timerlat_top.c
215
data->cpu_data[i].irq_count = value_irq[i];
tools/tracing/rtla/src/timerlat_top.c
216
data->cpu_data[i].thread_count = value_thread[i];
tools/tracing/rtla/src/timerlat_top.c
217
data->cpu_data[i].user_count = value_user[i];
tools/tracing/rtla/src/timerlat_top.c
226
data->cpu_data[i].min_irq = value_irq[i];
tools/tracing/rtla/src/timerlat_top.c
227
data->cpu_data[i].min_thread = value_thread[i];
tools/tracing/rtla/src/timerlat_top.c
228
data->cpu_data[i].min_user = value_user[i];
tools/tracing/rtla/src/timerlat_top.c
237
data->cpu_data[i].max_irq = value_irq[i];
tools/tracing/rtla/src/timerlat_top.c
238
data->cpu_data[i].max_thread = value_thread[i];
tools/tracing/rtla/src/timerlat_top.c
239
data->cpu_data[i].max_user = value_user[i];
tools/tracing/rtla/src/timerlat_top.c
248
data->cpu_data[i].sum_irq = value_irq[i];
tools/tracing/rtla/src/timerlat_top.c
249
data->cpu_data[i].sum_thread = value_thread[i];
tools/tracing/rtla/src/timerlat_top.c
250
data->cpu_data[i].sum_user = value_user[i];
tools/tracing/rtla/src/timerlat_top.c
309
struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu];
tools/tracing/rtla/src/timerlat_top.c
315
if (!cpu_data->irq_count && !cpu_data->thread_count)
tools/tracing/rtla/src/timerlat_top.c
321
trace_seq_printf(s, "%3d #%-9llu |", cpu, cpu_data->irq_count);
tools/tracing/rtla/src/timerlat_top.c
323
if (!cpu_data->irq_count) {
tools/tracing/rtla/src/timerlat_top.c
326
trace_seq_printf(s, "%9llu ", cpu_data->cur_irq);
tools/tracing/rtla/src/timerlat_top.c
327
trace_seq_printf(s, "%9llu ", cpu_data->min_irq);
tools/tracing/rtla/src/timerlat_top.c
328
trace_seq_printf(s, "%9llu ", cpu_data->sum_irq / cpu_data->irq_count);
tools/tracing/rtla/src/timerlat_top.c
329
trace_seq_printf(s, "%9llu |", cpu_data->max_irq);
tools/tracing/rtla/src/timerlat_top.c
332
if (!cpu_data->thread_count) {
tools/tracing/rtla/src/timerlat_top.c
335
trace_seq_printf(s, "%9llu ", cpu_data->cur_thread);
tools/tracing/rtla/src/timerlat_top.c
336
trace_seq_printf(s, "%9llu ", cpu_data->min_thread);
tools/tracing/rtla/src/timerlat_top.c
338
cpu_data->sum_thread / cpu_data->thread_count);
tools/tracing/rtla/src/timerlat_top.c
339
trace_seq_printf(s, "%9llu", cpu_data->max_thread);
tools/tracing/rtla/src/timerlat_top.c
349
if (!cpu_data->user_count) {
tools/tracing/rtla/src/timerlat_top.c
352
trace_seq_printf(s, "%9llu ", cpu_data->cur_user);
tools/tracing/rtla/src/timerlat_top.c
353
trace_seq_printf(s, "%9llu ", cpu_data->min_user);
tools/tracing/rtla/src/timerlat_top.c
355
cpu_data->sum_user / cpu_data->user_count);
tools/tracing/rtla/src/timerlat_top.c
356
trace_seq_printf(s, "%9llu\n", cpu_data->max_user);
tools/tracing/rtla/src/timerlat_top.c
43
struct timerlat_top_cpu *cpu_data;
tools/tracing/rtla/src/timerlat_top.c
52
free(data->cpu_data);
tools/tracing/rtla/src/timerlat_top.c
77
data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus);
tools/tracing/rtla/src/timerlat_top.c
78
if (!data->cpu_data)
tools/tracing/rtla/src/timerlat_top.c
83
data->cpu_data[cpu].min_irq = ~0;
tools/tracing/rtla/src/timerlat_top.c
84
data->cpu_data[cpu].min_thread = ~0;
tools/tracing/rtla/src/timerlat_top.c
85
data->cpu_data[cpu].min_user = ~0;