Symbol: cpuid
arch/alpha/include/asm/err_common.h
110
u64 cpuid;
arch/alpha/include/asm/hwrpb.h
169
unsigned long cpuid;
arch/alpha/include/asm/mmu_context.h
91
#define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn)
arch/alpha/include/asm/mmu_context.h
94
#define cpu_last_asn(cpuid) last_asn
arch/alpha/kernel/err_titan.c
552
(int)header->by_type.regatta_frame.cpuid);
arch/alpha/kernel/process.c
100
set_cpu_present(cpuid, false);
arch/alpha/kernel/process.c
101
set_cpu_possible(cpuid, false);
arch/alpha/kernel/process.c
81
int cpuid = smp_processor_id();
arch/alpha/kernel/process.c
88
+ hwrpb->processor_size * cpuid);
arch/alpha/kernel/process.c
97
if (cpuid != boot_cpuid) {
arch/alpha/kernel/smp.c
100
wait_boot_cpu_to_stop(int cpuid)
arch/alpha/kernel/smp.c
110
printk("wait_boot_cpu_to_stop: FAILED on CPU %d, hanging now\n", cpuid);
arch/alpha/kernel/smp.c
121
int cpuid = hard_smp_processor_id();
arch/alpha/kernel/smp.c
123
if (cpu_online(cpuid)) {
arch/alpha/kernel/smp.c
124
printk("??, cpu 0x%x already present??\n", cpuid);
arch/alpha/kernel/smp.c
127
set_cpu_online(cpuid, true);
arch/alpha/kernel/smp.c
139
smp_setup_percpu_timer(cpuid);
arch/alpha/kernel/smp.c
151
notify_cpu_starting(cpuid);
arch/alpha/kernel/smp.c
158
wait_boot_cpu_to_stop(cpuid);
arch/alpha/kernel/smp.c
162
smp_store_cpu_info(cpuid);
arch/alpha/kernel/smp.c
168
cpuid, current, current->active_mm));
arch/alpha/kernel/smp.c
198
send_secondary_console_msg(char *str, int cpuid)
arch/alpha/kernel/smp.c
208
+ cpuid * hwrpb->processor_size);
arch/alpha/kernel/smp.c
210
cpumask = (1UL << cpuid);
arch/alpha/kernel/smp.c
222
set_bit(cpuid, &hwrpb->rxrdy);
arch/alpha/kernel/smp.c
229
printk("Processor %x not ready\n", cpuid);
arch/alpha/kernel/smp.c
290
secondary_cpu_start(int cpuid, struct task_struct *idle)
arch/alpha/kernel/smp.c
299
+ cpuid * hwrpb->processor_size);
arch/alpha/kernel/smp.c
321
cpuid, idle->state, ipcb->flags));
arch/alpha/kernel/smp.c
339
send_secondary_console_msg("START\r\n", cpuid);
arch/alpha/kernel/smp.c
349
printk(KERN_ERR "SMP: Processor %d failed to start.\n", cpuid);
arch/alpha/kernel/smp.c
353
DBGS(("secondary_cpu_start: SUCCESS for CPU %d!!!\n", cpuid));
arch/alpha/kernel/smp.c
361
smp_boot_one_cpu(int cpuid, struct task_struct *idle)
arch/alpha/kernel/smp.c
369
if (secondary_cpu_start(cpuid, idle))
arch/alpha/kernel/smp.c
388
printk(KERN_ERR "SMP: Processor %d is stuck.\n", cpuid);
arch/alpha/kernel/smp.c
81
smp_store_cpu_info(int cpuid)
arch/alpha/kernel/smp.c
83
cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy;
arch/alpha/kernel/smp.c
84
cpu_data[cpuid].last_asn = ASN_FIRST_VERSION;
arch/alpha/kernel/smp.c
85
cpu_data[cpuid].need_new_asn = 0;
arch/alpha/kernel/smp.c
86
cpu_data[cpuid].asn_lock = 0;
arch/alpha/kernel/smp.c
93
smp_setup_percpu_timer(int cpuid)
arch/alpha/kernel/smp.c
95
cpu_data[cpuid].prof_counter = 1;
arch/alpha/kernel/smp.c
96
cpu_data[cpuid].prof_multiplier = 1;
arch/alpha/kernel/sys_marvel.c
405
int cpuid = hard_smp_processor_id();
arch/alpha/kernel/sys_marvel.c
406
struct io7 *io7 = marvel_find_io7(cpuid);
arch/alpha/kernel/sys_marvel.c
415
printk("Redirecting IO7 interrupts to local CPU at PE %u\n", cpuid);
arch/alpha/kernel/sys_marvel.c
418
io7_redirect_irq(io7, &io7->csrs->HLT_CTL.csr, cpuid);
arch/alpha/kernel/sys_marvel.c
419
io7_redirect_irq(io7, &io7->csrs->HPI_CTL.csr, cpuid);
arch/alpha/kernel/sys_marvel.c
420
io7_redirect_irq(io7, &io7->csrs->CRD_CTL.csr, cpuid);
arch/alpha/kernel/sys_marvel.c
421
io7_redirect_irq(io7, &io7->csrs->STV_CTL.csr, cpuid);
arch/alpha/kernel/sys_marvel.c
422
io7_redirect_irq(io7, &io7->csrs->HEI_CTL.csr, cpuid);
arch/alpha/kernel/sys_marvel.c
426
io7_redirect_one_lsi(io7, i, cpuid);
arch/alpha/kernel/sys_marvel.c
428
io7_redirect_one_lsi(io7, 0x74, cpuid);
arch/alpha/kernel/sys_marvel.c
429
io7_redirect_one_lsi(io7, 0x75, cpuid);
arch/alpha/kernel/sys_marvel.c
433
io7_redirect_one_msi(io7, i, cpuid);
arch/arm/include/asm/cacheflush.h
471
void check_cpu_icache_size(int cpuid);
arch/arm/include/asm/cacheflush.h
473
static inline void check_cpu_icache_size(int cpuid) { }
arch/arm/include/asm/cpu.h
14
u32 cpuid;
arch/arm/include/asm/smp_plat.h
40
return is_smp() ? cpu_info->cpuid & ARM_CPU_PART_MASK :
arch/arm/include/asm/topology.h
32
static inline void store_cpu_topology(unsigned int cpuid) { }
arch/arm/kernel/setup.c
1262
u32 cpuid;
arch/arm/kernel/setup.c
1271
cpuid = is_smp() ? per_cpu(cpu_data, i).cpuid : read_cpuid_id();
arch/arm/kernel/setup.c
1273
cpu_name, cpuid & 15, elf_platform);
arch/arm/kernel/setup.c
1295
seq_printf(m, "\nCPU implementer\t: 0x%02x\n", cpuid >> 24);
arch/arm/kernel/setup.c
1299
if ((cpuid & 0x0008f000) == 0x00000000) {
arch/arm/kernel/setup.c
1301
seq_printf(m, "CPU part\t: %07x\n", cpuid >> 4);
arch/arm/kernel/setup.c
1303
if ((cpuid & 0x0008f000) == 0x00007000) {
arch/arm/kernel/setup.c
1306
(cpuid >> 16) & 127);
arch/arm/kernel/setup.c
1310
(cpuid >> 20) & 15);
arch/arm/kernel/setup.c
1313
(cpuid >> 4) & 0xfff);
arch/arm/kernel/setup.c
1315
seq_printf(m, "CPU revision\t: %d\n\n", cpuid & 15);
arch/arm/kernel/smp.c
389
static void smp_store_cpu_info(unsigned int cpuid)
arch/arm/kernel/smp.c
391
struct cpuinfo_arm *cpu_info = &per_cpu(cpu_data, cpuid);
arch/arm/kernel/smp.c
394
cpu_info->cpuid = read_cpuid_id();
arch/arm/kernel/smp.c
396
store_cpu_topology(cpuid);
arch/arm/kernel/smp.c
397
check_cpu_icache_size(cpuid);
arch/arm/kernel/suspend.c
17
extern int __cpu_suspend(unsigned long, int (*)(unsigned long), u32 cpuid);
arch/arm/kernel/topology.c
177
static inline void update_cpu_capacity(unsigned int cpuid) {}
arch/arm/kernel/topology.c
185
void store_cpu_topology(unsigned int cpuid)
arch/arm/kernel/topology.c
187
struct cpu_topology *cpuid_topo = &cpu_topology[cpuid];
arch/arm/kernel/topology.c
224
update_cpu_capacity(cpuid);
arch/arm/kernel/topology.c
227
cpuid, cpu_topology[cpuid].thread_id,
arch/arm/kernel/topology.c
228
cpu_topology[cpuid].core_id,
arch/arm/kernel/topology.c
229
cpu_topology[cpuid].package_id, mpidr);
arch/arm/kernel/topology.c
232
update_siblings_masks(cpuid);
arch/arm/mach-exynos/pm.c
155
unsigned int cpuid = smp_processor_id();
arch/arm/mach-exynos/pm.c
160
exynos_set_boot_flag(cpuid, C2_STATE);
arch/arm/mach-exynos/pm.c
181
exynos_clear_boot_flag(cpuid, C2_STATE);
arch/arm/mach-exynos/suspend.c
413
u32 cpuid = read_cpuid_part();
arch/arm/mach-exynos/suspend.c
418
if (cpuid == ARM_CPU_PART_CORTEX_A9)
arch/arm/mach-exynos/suspend.c
422
&& cpuid == ARM_CPU_PART_CORTEX_A9)
arch/arm/mach-exynos/suspend.c
434
u32 cpuid = read_cpuid_part();
arch/arm/mach-exynos/suspend.c
442
&& cpuid == ARM_CPU_PART_CORTEX_A9)
arch/arm/mach-omap2/id.c
351
u32 cpuid, idcode;
arch/arm/mach-omap2/id.c
360
cpuid = read_cpuid_id();
arch/arm/mach-omap2/id.c
361
if ((((cpuid >> 4) & 0xfff) == 0xc08) && ((cpuid & 0xf) == 0x0)) {
arch/arm/mm/init.c
165
void check_cpu_icache_size(int cpuid)
arch/arm/mm/init.c
172
if (cpuid != 0 && icache_size != size)
arch/arm/mm/init.c
174
cpuid);
arch/arm64/kvm/vgic/vgic-kvm-device.c
347
int cpuid = FIELD_GET(KVM_DEV_ARM_VGIC_CPUID_MASK, attr->attr);
arch/arm64/kvm/vgic/vgic-kvm-device.c
350
reg_attr->vcpu = kvm_get_vcpu_by_id(dev->kvm, cpuid);
arch/arm64/kvm/vgic/vgic-v2.c
160
u8 cpuid;
arch/arm64/kvm/vgic/vgic-v2.c
163
cpuid = FIELD_GET(GENMASK_ULL(12, 10), val);
arch/arm64/kvm/vgic/vgic-v2.c
195
if (val < VGIC_NR_SGIS && irq->active_source != cpuid) {
arch/arm64/kvm/vgic/vgic-v2.c
58
u32 cpuid, intid = val & GICH_LR_VIRTUALID;
arch/arm64/kvm/vgic/vgic-v2.c
63
cpuid = FIELD_GET(GICH_LR_PHYSID_CPUID, val) & 7;
arch/arm64/kvm/vgic/vgic-v2.c
78
irq->active_source = cpuid;
arch/arm64/kvm/vgic/vgic-v2.c
86
irq->source |= (1 << cpuid);
arch/arm64/kvm/vgic/vgic-v3.c
111
u8 cpuid = FIELD_GET(GICH_LR_PHYSID_CPUID, val);
arch/arm64/kvm/vgic/vgic-v3.c
114
irq->active_source = cpuid;
arch/arm64/kvm/vgic/vgic-v3.c
117
irq->source |= BIT(cpuid);
arch/arm64/kvm/vgic/vgic-v3.c
205
u8 cpuid;
arch/arm64/kvm/vgic/vgic-v3.c
208
cpuid = FIELD_GET(GENMASK_ULL(12, 10), val);
arch/arm64/kvm/vgic/vgic-v3.c
269
if (is_v2_sgi && irq->active_source != cpuid) {
arch/loongarch/include/asm/kvm_vcpu.h
91
struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid);
arch/loongarch/include/asm/numa.h
25
extern void __init early_numa_add_cpu(int cpuid, s16 node);
arch/loongarch/include/asm/numa.h
33
static inline void set_cpuid_to_node(int cpuid, s16 node)
arch/loongarch/include/asm/numa.h
35
__cpuid_to_node[cpuid] = node;
arch/loongarch/include/asm/numa.h
42
static inline void early_numa_add_cpu(int cpuid, s16 node) { }
arch/loongarch/include/asm/numa.h
45
static inline void set_cpuid_to_node(int cpuid, s16 node) { }
arch/loongarch/kernel/acpi.c
65
int cpu = -1, cpuid = id;
arch/loongarch/kernel/acpi.c
69
" processor 0x%x ignored.\n", NR_CPUS, cpuid);
arch/loongarch/kernel/acpi.c
75
if (cpuid == loongson_sysconf.boot_cpu_id)
arch/loongarch/kernel/acpi.c
95
__cpu_number_map[cpuid] = cpu;
arch/loongarch/kernel/acpi.c
96
__cpu_logical_map[cpu] = cpuid;
arch/loongarch/kernel/numa.c
120
void __init early_numa_add_cpu(int cpuid, s16 node)
arch/loongarch/kernel/numa.c
122
int cpu = __cpu_number_map[cpuid];
arch/loongarch/kernel/numa.c
128
cpumask_set_cpu(cpuid, &phys_cpus_on_node[node]);
arch/loongarch/kernel/smp.c
319
unsigned int cpu, cpuid;
arch/loongarch/kernel/smp.c
326
cpuid = of_get_cpu_hwid(node, 0);
arch/loongarch/kernel/smp.c
327
if (cpuid >= nr_cpu_ids)
arch/loongarch/kernel/smp.c
330
if (cpuid == loongson_sysconf.boot_cpu_id)
arch/loongarch/kernel/smp.c
338
__cpu_number_map[cpuid] = cpu;
arch/loongarch/kernel/smp.c
339
__cpu_logical_map[cpu] = cpuid;
arch/loongarch/kernel/smp.c
341
early_numa_add_cpu(cpuid, 0);
arch/loongarch/kernel/smp.c
342
set_cpuid_to_node(cpuid, 0);
arch/loongarch/kvm/intc/eiointc.c
12
int ipnum, cpu, cpuid, irq;
arch/loongarch/kvm/intc/eiointc.c
22
cpuid = ((u8 *)s->coremap)[irq];
arch/loongarch/kvm/intc/eiointc.c
23
vcpu = kvm_get_vcpu_by_cpuid(s->kvm, cpuid);
arch/loongarch/kvm/intc/eiointc.c
77
int i, cpu, cpuid;
arch/loongarch/kvm/intc/eiointc.c
81
cpuid = val & 0xff;
arch/loongarch/kvm/intc/eiointc.c
85
cpuid = ffs(cpuid) - 1;
arch/loongarch/kvm/intc/eiointc.c
86
cpuid = ((cpuid < 0) || (cpuid >= 4)) ? 0 : cpuid;
arch/loongarch/kvm/intc/eiointc.c
89
vcpu = kvm_get_vcpu_by_cpuid(s->kvm, cpuid);
arch/loongarch/kvm/vcpu.c
517
int cpuid;
arch/loongarch/kvm/vcpu.c
525
cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
arch/loongarch/kvm/vcpu.c
528
if ((cpuid < KVM_MAX_PHYID) && map->phys_map[cpuid].enabled) {
arch/loongarch/kvm/vcpu.c
530
if (cpuid == val) {
arch/loongarch/kvm/vcpu.c
568
int cpuid;
arch/loongarch/kvm/vcpu.c
573
cpuid = kvm_read_sw_gcsr(csr, LOONGARCH_CSR_CPUID);
arch/loongarch/kvm/vcpu.c
575
if (cpuid >= KVM_MAX_PHYID)
arch/loongarch/kvm/vcpu.c
579
if (map->phys_map[cpuid].enabled) {
arch/loongarch/kvm/vcpu.c
580
map->phys_map[cpuid].vcpu = NULL;
arch/loongarch/kvm/vcpu.c
581
map->phys_map[cpuid].enabled = false;
arch/loongarch/kvm/vcpu.c
587
struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid)
arch/loongarch/kvm/vcpu.c
591
if (cpuid < 0)
arch/loongarch/kvm/vcpu.c
594
if (cpuid >= KVM_MAX_PHYID)
arch/loongarch/kvm/vcpu.c
598
if (!map->phys_map[cpuid].enabled)
arch/loongarch/kvm/vcpu.c
601
return map->phys_map[cpuid].vcpu;
arch/m68k/include/asm/macintosh.h
111
unsigned long cpuid;
arch/m68k/mac/config.c
111
mac_bi_data.cpuid = be32_to_cpup(data);
arch/m68k/mac/config.c
789
model = (mac_bi_data.cpuid >> 2) & 63;
arch/m68k/mac/config.c
846
mac_bi_data.id, mac_bi_data.cpuid, mac_bi_data.memsize);
arch/mips/include/asm/octeon/octeon-model.h
366
uint32_t cpuid = cvmx_get_proc_id();
arch/mips/include/asm/octeon/octeon-model.h
368
return __OCTEON_IS_MODEL_COMPILE__(model, cpuid);
arch/mips/include/asm/sgi/heart.h
127
u64 cpuid; /* + 0x50000 */
arch/mips/include/asm/sn/agent.h
37
SET_HUB_NIC(cpuid(), (_v))
arch/mips/include/asm/sn/agent.h
43
GET_HUB_NIC(cpuid())
arch/mips/loongson64/smp.c
546
register long cpuid, core, node, count;
arch/mips/loongson64/smp.c
599
[base] "=&r" (base), [cpuid] "=&r" (cpuid),
arch/mips/loongson64/smp.c
608
register long cpuid, core, node, count;
arch/mips/loongson64/smp.c
663
[base] "=&r" (base), [cpuid] "=&r" (cpuid),
arch/mips/loongson64/smp.c
672
register long cpuid, core, node, count;
arch/mips/loongson64/smp.c
747
[base] "=&r" (base), [cpuid] "=&r" (cpuid),
arch/mips/sgi-ip27/ip27-smp.c
34
cpuid_t cpuid;
arch/mips/sgi-ip27/ip27-smp.c
41
cpuid = acpu->cpu_info.virtid;
arch/mips/sgi-ip27/ip27-smp.c
45
if (cpuid > highest)
arch/mips/sgi-ip27/ip27-smp.c
46
highest = cpuid;
arch/mips/sgi-ip27/ip27-smp.c
47
set_cpu_possible(cpuid, true);
arch/openrisc/kernel/smp.c
248
unsigned int cpuid;
arch/openrisc/kernel/smp.c
253
cpuid = get_cpu();
arch/openrisc/kernel/smp.c
255
if (cpumask_any_but(cmask, cpuid) >= nr_cpu_ids) {
arch/openrisc/kernel/smp.c
286
unsigned int cpuid;
arch/openrisc/kernel/smp.c
291
cpuid = get_cpu();
arch/openrisc/kernel/smp.c
293
if (cpumask_any_but(cmask, cpuid) >= nr_cpu_ids) {
arch/parisc/include/asm/processor.h
70
unsigned long cpuid;
arch/parisc/include/asm/processor.h
84
unsigned long cpuid; /* aka slot_number or set to NO_PROC_ID */
arch/parisc/include/asm/smp.h
5
extern int init_per_cpu(int cpuid);
arch/parisc/include/asm/topology.h
12
static inline void store_cpu_topology(unsigned int cpuid) { }
arch/parisc/kernel/drivers.c
942
boot_cpu_data.pdc.cpuid);
arch/parisc/kernel/processor.c
103
cpuid = boot_cpu_data.cpu_count;
arch/parisc/kernel/processor.c
105
cpu_info.cpu_num = cpu_info.cpu_loc = cpuid;
arch/parisc/kernel/processor.c
136
cpuid, cpu_info.cpu_num, cpu_info.cpu_loc,
arch/parisc/kernel/processor.c
155
cpuid = cpu_info.cpu_num;
arch/parisc/kernel/processor.c
161
p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/processor.c
165
if (cpuid)
arch/parisc/kernel/processor.c
170
p->cpuid = cpuid; /* save CPU id */
arch/parisc/kernel/processor.c
175
store_cpu_topology(cpuid);
arch/parisc/kernel/processor.c
182
init_percpu_prof(cpuid);
arch/parisc/kernel/processor.c
193
if (cpuid) {
arch/parisc/kernel/processor.c
206
cpu_irq_actions[cpuid] = actions;
arch/parisc/kernel/processor.c
214
if (cpuid) {
arch/parisc/kernel/processor.c
215
set_cpu_present(cpuid, true);
arch/parisc/kernel/processor.c
216
add_cpu(cpuid);
arch/parisc/kernel/processor.c
261
if (pdc_model_cpuid(&boot_cpu_data.pdc.cpuid) == PDC_OK) {
arch/parisc/kernel/processor.c
263
(boot_cpu_data.pdc.cpuid >> 5) & 127,
arch/parisc/kernel/processor.c
264
boot_cpu_data.pdc.cpuid & 31,
arch/parisc/kernel/processor.c
265
boot_cpu_data.pdc.cpuid);
arch/parisc/kernel/processor.c
267
add_device_randomness(&boot_cpu_data.pdc.cpuid,
arch/parisc/kernel/processor.c
268
sizeof(boot_cpu_data.pdc.cpuid));
arch/parisc/kernel/processor.c
84
unsigned long cpuid;
arch/parisc/kernel/smp.c
333
static int smp_boot_one_cpu(int cpuid, struct task_struct *idle)
arch/parisc/kernel/smp.c
335
const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/smp.c
342
memset(&per_cpu(irq_stat, cpuid), 0, sizeof(irq_cpustat_t));
arch/parisc/kernel/smp.c
347
*per_cpu_ptr(desc->kstat_irqs, cpuid) = (struct irqstat) { };
arch/parisc/kernel/smp.c
358
cpu_now_booting = cpuid;
arch/parisc/kernel/smp.c
367
printk(KERN_INFO "Releasing cpu %d now, hpa=%lx\n", cpuid, p->hpa);
arch/parisc/kernel/smp.c
387
if(cpu_online(cpuid)) {
arch/parisc/kernel/smp.c
395
printk(KERN_CRIT "SMP: CPU:%d is stuck.\n", cpuid);
arch/parisc/kernel/smp.c
401
cpuid, timeout * 100);
arch/parisc/kernel/smp.c
88
ipi_init(int cpuid)
arch/parisc/kernel/smp.c
92
if(cpu_online(cpuid) )
arch/parisc/kernel/topology.c
28
void store_cpu_topology(unsigned int cpuid)
arch/parisc/kernel/topology.c
30
struct cpu_topology *cpuid_topo = &cpu_topology[cpuid];
arch/parisc/kernel/topology.c
40
per_cpu(cpu_devices, cpuid).hotpluggable = 1;
arch/parisc/kernel/topology.c
42
if (register_cpu(&per_cpu(cpu_devices, cpuid), cpuid))
arch/parisc/kernel/topology.c
43
pr_warn("Failed to register CPU%d device", cpuid);
arch/parisc/kernel/topology.c
49
p = &per_cpu(cpu_data, cpuid);
arch/parisc/kernel/topology.c
53
if (cpu == cpuid) /* ignore current cpu */
arch/parisc/kernel/topology.c
72
update_siblings_masks(cpuid);
arch/parisc/kernel/topology.c
75
cpuid,
arch/parisc/kernel/topology.c
76
cpu_topology[cpuid].core_id,
arch/parisc/kernel/topology.c
77
cpu_topology[cpuid].package_id);
arch/powerpc/kernel/irq.c
362
int cpuid;
arch/powerpc/kernel/irq.c
377
cpuid = irq_rover;
arch/powerpc/kernel/irq.c
381
cpuid = cpumask_first_and(mask, cpu_online_mask);
arch/powerpc/kernel/irq.c
382
if (cpuid >= nr_cpu_ids)
arch/powerpc/kernel/irq.c
386
return get_hard_smp_processor_id(cpuid);
arch/powerpc/sysdev/ehv_pic.c
76
int cpuid = irq_choose_cpu(dest);
arch/powerpc/sysdev/ehv_pic.c
81
ev_int_set_config(src, config, prio, cpuid);
arch/powerpc/sysdev/mpic.c
1907
int cpuid = get_hard_smp_processor_id(cpu);
arch/powerpc/sysdev/mpic.c
1912
pir |= (1 << cpuid);
arch/powerpc/sysdev/mpic.c
1917
pir &= ~(1 << cpuid);
arch/powerpc/sysdev/mpic.c
1925
_mpic_write(mpic->reg_type, &mpic->cpuregs[cpuid],
arch/powerpc/sysdev/mpic.c
834
int cpuid = irq_choose_cpu(cpumask);
arch/powerpc/sysdev/mpic.c
836
mpic_irq_write(src, MPIC_INFO(IRQ_DESTINATION), 1 << cpuid);
arch/powerpc/sysdev/mpic.c
948
static void mpic_set_destination(unsigned int virq, unsigned int cpuid)
arch/powerpc/sysdev/mpic.c
954
mpic, virq, src, cpuid);
arch/powerpc/sysdev/mpic.c
959
mpic_irq_write(src, MPIC_INFO(IRQ_DESTINATION), 1 << cpuid);
arch/riscv/kernel/acpi.c
172
int cpuid;
arch/riscv/kernel/acpi.c
177
cpuid = riscv_hartid_to_cpuid(rintc->hart_id);
arch/riscv/kernel/acpi.c
183
if (cpuid >= 0 && cpuid < num_possible_cpus())
arch/riscv/kernel/acpi.c
184
cpu_madt_rintc[cpuid] = *rintc;
arch/riscv/kernel/cpu_ops_sbi.c
65
static int sbi_cpu_start(unsigned int cpuid, struct task_struct *tidle)
arch/riscv/kernel/cpu_ops_sbi.c
68
unsigned long hartid = cpuid_to_hartid_map(cpuid);
arch/riscv/kernel/cpu_ops_sbi.c
70
struct sbi_hart_boot_data *bdata = &boot_data[cpuid];
arch/riscv/kernel/cpu_ops_sbi.c
91
static int sbi_cpu_is_stopped(unsigned int cpuid)
arch/riscv/kernel/cpu_ops_sbi.c
94
unsigned long hartid = cpuid_to_hartid_map(cpuid);
arch/riscv/kernel/cpu_ops_spinwait.c
20
static void cpu_update_secondary_bootdata(unsigned int cpuid,
arch/riscv/kernel/cpu_ops_spinwait.c
23
unsigned long hartid = cpuid_to_hartid_map(cpuid);
arch/riscv/kernel/cpu_ops_spinwait.c
41
static int spinwait_cpu_start(unsigned int cpuid, struct task_struct *tidle)
arch/riscv/kernel/cpu_ops_spinwait.c
51
cpu_update_secondary_bootdata(cpuid, tidle);
arch/riscv/kernel/sbi.c
261
unsigned long hartid, cpuid, hmask = 0, hbase = 0, htop = 0;
arch/riscv/kernel/sbi.c
267
for_each_cpu(cpuid, cpu_mask) {
arch/riscv/kernel/sbi.c
268
hartid = cpuid_to_hartid_map(cpuid);
arch/riscv/kernel/sbi.c
30
unsigned long cpuid, hartid;
arch/riscv/kernel/sbi.c
40
for_each_cpu(cpuid, cpu_mask) {
arch/riscv/kernel/sbi.c
41
hartid = cpuid_to_hartid_map(cpuid);
arch/riscv/kernel/smpboot.c
128
int cpuid = 1;
arch/riscv/kernel/smpboot.c
142
if (cpuid >= NR_CPUS) {
arch/riscv/kernel/smpboot.c
144
cpuid, hart);
arch/riscv/kernel/smpboot.c
148
cpuid_to_hartid_map(cpuid) = hart;
arch/riscv/kernel/smpboot.c
149
early_map_cpu_to_node(cpuid, of_node_to_nid(dn));
arch/riscv/kernel/smpboot.c
150
cpuid++;
arch/riscv/kernel/smpboot.c
155
if (cpuid > nr_cpu_ids)
arch/riscv/kernel/smpboot.c
157
cpuid, nr_cpu_ids);
arch/riscv/kernel/smpboot.c
162
int cpuid;
arch/riscv/kernel/smpboot.c
171
for (cpuid = 1; cpuid < nr_cpu_ids; cpuid++)
arch/riscv/kernel/smpboot.c
172
if (cpuid_to_hartid_map(cpuid) != INVALID_HARTID)
arch/riscv/kernel/smpboot.c
173
set_cpu_possible(cpuid, true);
arch/riscv/kernel/smpboot.c
48
int cpuid;
arch/riscv/kernel/smpboot.c
62
for_each_possible_cpu(cpuid) {
arch/riscv/kernel/smpboot.c
63
if (cpuid == curr_cpuid)
arch/riscv/kernel/smpboot.c
65
set_cpu_present(cpuid, true);
arch/riscv/kernel/smpboot.c
66
numa_store_cpu_info(cpuid);
arch/s390/boot/als.c
62
struct cpuid id;
arch/s390/include/asm/kvm_host.h
501
u64 cpuid;
arch/s390/include/asm/processor.h
101
static inline void get_cpu_id(struct cpuid *ptr)
arch/s390/include/uapi/asm/kvm.h
413
__u64 cpuid;
arch/s390/include/uapi/asm/kvm.h
422
__u64 cpuid;
arch/s390/kernel/perf_cpum_cf.c
393
struct cpuid cpuid;
arch/s390/kernel/perf_cpum_cf.c
398
get_cpu_id(&cpuid); /* Machine type */
arch/s390/kernel/perf_cpum_cf.c
399
te->mach_type = cpuid.machine;
arch/s390/kernel/perf_cpum_cf_events.c
999
struct cpuid cpu_id;
arch/s390/kernel/processor.c
106
struct cpuid *id = this_cpu_ptr(&cpu_info.cpu_id);
arch/s390/kernel/processor.c
171
struct cpuid *id = &per_cpu(cpu_info.cpu_id, cpu);
arch/s390/kernel/processor.c
264
struct cpuid cpu_id;
arch/s390/kernel/processor.c
321
struct cpuid *id = &per_cpu(cpu_info.cpu_id, n);
arch/s390/kernel/processor.c
37
struct cpuid cpu_id;
arch/s390/kvm/kvm-s390.c
1427
kvm->arch.model.cpuid = proc->cpuid;
arch/s390/kvm/kvm-s390.c
1442
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
1625
proc->cpuid = kvm->arch.model.cpuid;
arch/s390/kvm/kvm-s390.c
1631
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
1653
get_cpu_id((struct cpuid *) &mach->cpuid);
arch/s390/kvm/kvm-s390.c
1661
kvm->arch.model.cpuid);
arch/s390/kvm/kvm-s390.c
3131
struct cpuid cpuid;
arch/s390/kvm/kvm-s390.c
3133
get_cpu_id(&cpuid);
arch/s390/kvm/kvm-s390.c
3134
cpuid.version = 0xff;
arch/s390/kvm/kvm-s390.c
3135
return *((u64 *) &cpuid);
arch/s390/kvm/kvm-s390.c
3237
kvm->arch.model.cpuid = kvm_s390_get_initial_cpuid();
arch/s390/kvm/pci.h
68
struct cpuid cpu_id;
arch/s390/kvm/priv.c
799
u64 stidp_data = vcpu->kvm->arch.model.cpuid;
arch/sparc/include/asm/hypervisor.h
1727
unsigned long sun4v_intr_settarget(unsigned long sysino, unsigned long cpuid);
arch/sparc/include/asm/hypervisor.h
1831
unsigned long *cpuid);
arch/sparc/include/asm/hypervisor.h
1834
unsigned long cpuid);
arch/sparc/include/asm/hypervisor.h
258
unsigned long sun4v_cpu_start(unsigned long cpuid,
arch/sparc/include/asm/hypervisor.h
286
unsigned long sun4v_cpu_stop(unsigned long cpuid);
arch/sparc/include/asm/hypervisor.h
321
unsigned long sun4v_cpu_poke(unsigned long cpuid);
arch/sparc/include/asm/hypervisor.h
453
long sun4v_cpu_state(unsigned long cpuid);
arch/sparc/include/asm/obio.h
197
static inline unsigned int cc_get_imsk_other(int cpuid)
arch/sparc/include/asm/obio.h
203
"r" (ECSR_BASE(cpuid) | CC_IMSK),
arch/sparc/include/asm/obio.h
208
static inline void cc_set_imsk_other(int cpuid, unsigned int mask)
arch/sparc/include/asm/obio.h
212
"r" (ECSR_BASE(cpuid) | CC_IMSK),
arch/sparc/include/asm/oplib_64.h
116
void prom_startcpu_cpuid(int cpuid, unsigned long pc, unsigned long arg);
arch/sparc/include/asm/oplib_64.h
119
void prom_stopcpu_cpuid(int cpuid);
arch/sparc/include/asm/prom.h
35
struct device_node *of_find_node_by_cpuid(int cpuid);
arch/sparc/kernel/irq_64.c
313
static unsigned int sun4u_compute_tid(unsigned long imap, unsigned long cpuid)
arch/sparc/kernel/irq_64.c
318
tid = starfire_translate(imap, cpuid);
arch/sparc/kernel/irq_64.c
328
tid = cpuid << IMAP_TID_SHIFT;
arch/sparc/kernel/irq_64.c
331
unsigned int a = cpuid & 0x1f;
arch/sparc/kernel/irq_64.c
332
unsigned int n = (cpuid >> 5) & 0x1f;
arch/sparc/kernel/irq_64.c
340
tid = cpuid << IMAP_TID_SHIFT;
arch/sparc/kernel/irq_64.c
351
int cpuid;
arch/sparc/kernel/irq_64.c
354
cpuid = map_to_cpu(irq);
arch/sparc/kernel/irq_64.c
356
cpuid = cpumask_first_and(affinity, cpu_online_mask);
arch/sparc/kernel/irq_64.c
357
cpuid = cpuid < nr_cpu_ids ? cpuid : map_to_cpu(irq);
arch/sparc/kernel/irq_64.c
360
return cpuid;
arch/sparc/kernel/irq_64.c
373
unsigned long cpuid, imap, val;
arch/sparc/kernel/irq_64.c
376
cpuid = irq_choose_cpu(data->irq,
arch/sparc/kernel/irq_64.c
380
tid = sun4u_compute_tid(imap, cpuid);
arch/sparc/kernel/irq_64.c
398
unsigned long cpuid, imap, val;
arch/sparc/kernel/irq_64.c
401
cpuid = irq_choose_cpu(data->irq, mask);
arch/sparc/kernel/irq_64.c
404
tid = sun4u_compute_tid(imap, cpuid);
arch/sparc/kernel/irq_64.c
449
unsigned long cpuid = irq_choose_cpu(data->irq,
arch/sparc/kernel/irq_64.c
454
err = sun4v_intr_settarget(ino, cpuid);
arch/sparc/kernel/irq_64.c
457
"err(%d)\n", ino, cpuid, err);
arch/sparc/kernel/irq_64.c
471
unsigned long cpuid = irq_choose_cpu(data->irq, mask);
arch/sparc/kernel/irq_64.c
475
err = sun4v_intr_settarget(ino, cpuid);
arch/sparc/kernel/irq_64.c
478
"err(%d)\n", ino, cpuid, err);
arch/sparc/kernel/irq_64.c
509
unsigned long cpuid;
arch/sparc/kernel/irq_64.c
512
cpuid = irq_choose_cpu(data->irq, irq_data_get_affinity_mask(data));
arch/sparc/kernel/irq_64.c
514
err = sun4v_vintr_set_target(dev_handle, dev_ino, cpuid);
arch/sparc/kernel/irq_64.c
518
dev_handle, dev_ino, cpuid, err);
arch/sparc/kernel/irq_64.c
538
unsigned long cpuid;
arch/sparc/kernel/irq_64.c
541
cpuid = irq_choose_cpu(data->irq, mask);
arch/sparc/kernel/irq_64.c
543
err = sun4v_vintr_set_target(dev_handle, dev_ino, cpuid);
arch/sparc/kernel/irq_64.c
547
dev_handle, dev_ino, cpuid, err);
arch/sparc/kernel/leon_smp.c
78
int cpuid = hard_smp_processor_id();
arch/sparc/kernel/leon_smp.c
85
do_swap(&cpu_callin_map[cpuid], 1);
arch/sparc/kernel/leon_smp.c
91
__asm__ __volatile__("ld [%0], %%g6\n\t" : : "r"(&current_set[cpuid])
arch/sparc/kernel/leon_smp.c
98
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/mdesc.c
1115
int cpuid = *id;
arch/sparc/kernel/mdesc.c
1118
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/mdesc.c
1121
cpuid, NR_CPUS);
arch/sparc/kernel/mdesc.c
1124
if (!cpumask_test_cpu(cpuid, mask))
arch/sparc/kernel/mdesc.c
1128
ret = func(hp, mp, cpuid, arg);
arch/sparc/kernel/mdesc.c
1137
static void *record_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid,
arch/sparc/kernel/mdesc.c
1142
set_cpu_present(cpuid, true);
arch/sparc/kernel/mdesc.c
1156
static void * __init check_one_pgsz(struct mdesc_handle *hp, u64 mp, int cpuid, void *arg)
arch/sparc/kernel/mdesc.c
1180
static void *fill_in_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid,
arch/sparc/kernel/mdesc.c
1193
if (cpuid != real_hard_smp_processor_id())
arch/sparc/kernel/mdesc.c
1195
cpuid = 0;
arch/sparc/kernel/mdesc.c
1198
c = &cpu_data(cpuid);
arch/sparc/kernel/mdesc.c
1201
tb = &trap_block[cpuid];
arch/sparc/kernel/prom_64.c
443
int cpuid = of_getintprop_default(dp, mid_prop, -1);
arch/sparc/kernel/prom_64.c
447
if (cpuid < 0) {
arch/sparc/kernel/prom_64.c
449
cpuid = of_getintprop_default(dp, this_mid_prop, -1);
arch/sparc/kernel/prom_64.c
451
if (cpuid < 0) {
arch/sparc/kernel/prom_64.c
457
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/prom_64.c
460
cpuid, NR_CPUS);
arch/sparc/kernel/prom_64.c
464
ret = func(dp, cpuid, arg);
arch/sparc/kernel/prom_64.c
471
static void *check_cpu_node(struct device_node *dp, int cpuid, int id)
arch/sparc/kernel/prom_64.c
473
if (id == cpuid)
arch/sparc/kernel/prom_64.c
478
struct device_node *of_find_node_by_cpuid(int cpuid)
arch/sparc/kernel/prom_64.c
480
return of_iterate_over_cpus(check_cpu_node, cpuid);
arch/sparc/kernel/prom_64.c
483
static void *record_one_cpu(struct device_node *dp, int cpuid, int arg)
arch/sparc/kernel/prom_64.c
487
set_cpu_present(cpuid, true);
arch/sparc/kernel/prom_64.c
490
set_cpu_possible(cpuid, true);
arch/sparc/kernel/prom_64.c
504
static void *fill_in_one_cpu(struct device_node *dp, int cpuid, int arg)
arch/sparc/kernel/prom_64.c
529
if (cpuid != real_hard_smp_processor_id())
arch/sparc/kernel/prom_64.c
531
cpuid = 0;
arch/sparc/kernel/prom_64.c
534
cpu_data(cpuid).clock_tick =
arch/sparc/kernel/prom_64.c
538
cpu_data(cpuid).dcache_size =
arch/sparc/kernel/prom_64.c
541
cpu_data(cpuid).dcache_line_size =
arch/sparc/kernel/prom_64.c
544
cpu_data(cpuid).icache_size =
arch/sparc/kernel/prom_64.c
547
cpu_data(cpuid).icache_line_size =
arch/sparc/kernel/prom_64.c
550
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
552
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
554
if (!cpu_data(cpuid).ecache_size ||
arch/sparc/kernel/prom_64.c
555
!cpu_data(cpuid).ecache_line_size) {
arch/sparc/kernel/prom_64.c
556
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
560
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
565
cpu_data(cpuid).core_id = portid + 1;
arch/sparc/kernel/prom_64.c
566
cpu_data(cpuid).proc_id = portid;
arch/sparc/kernel/prom_64.c
568
cpu_data(cpuid).dcache_size =
arch/sparc/kernel/prom_64.c
570
cpu_data(cpuid).dcache_line_size =
arch/sparc/kernel/prom_64.c
573
cpu_data(cpuid).icache_size =
arch/sparc/kernel/prom_64.c
575
cpu_data(cpuid).icache_line_size =
arch/sparc/kernel/prom_64.c
578
cpu_data(cpuid).ecache_size =
arch/sparc/kernel/prom_64.c
581
cpu_data(cpuid).ecache_line_size =
arch/sparc/kernel/prom_64.c
584
cpu_data(cpuid).core_id = 0;
arch/sparc/kernel/prom_64.c
585
cpu_data(cpuid).proc_id = -1;
arch/sparc/kernel/smp_32.c
179
int i, cpuid, extra;
arch/sparc/kernel/smp_32.c
184
for (i = 0; !cpu_find_by_instance(i, NULL, &cpuid); i++) {
arch/sparc/kernel/smp_32.c
185
if (cpuid >= NR_CPUS)
arch/sparc/kernel/smp_32.c
239
int cpuid = hard_smp_processor_id();
arch/sparc/kernel/smp_32.c
241
if (cpuid >= NR_CPUS) {
arch/sparc/kernel/smp_32.c
245
if (cpuid != 0)
arch/sparc/kernel/smp_32.c
248
current_thread_info()->cpu = cpuid;
arch/sparc/kernel/smp_32.c
249
set_cpu_online(cpuid, true);
arch/sparc/kernel/smp_32.c
250
set_cpu_possible(cpuid, true);
arch/sparc/kernel/smp_32.c
311
unsigned int cpuid = hard_smp_processor_id();
arch/sparc/kernel/smp_32.c
313
register_percpu_ce(cpuid);
arch/sparc/kernel/smp_32.c
316
smp_store_cpu_info(cpuid);
arch/sparc/kernel/smp_64.c
106
int cpuid = hard_smp_processor_id();
arch/sparc/kernel/smp_64.c
108
__local_per_cpu_offset = __per_cpu_offset(cpuid);
arch/sparc/kernel/smp_64.c
134
notify_cpu_starting(cpuid);
arch/sparc/kernel/smp_64.c
136
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/smp_64.c
139
set_cpu_online(cpuid, true);
arch/sparc/kernel/smp_64.c
623
#define CPU_MONDO_COUNTER(cpuid) (cpu_mondo_counter[cpuid])
arch/sparc/kernel/sun4d_irq.c
195
int cpuid = handler_data->cpuid;
arch/sparc/kernel/sun4d_irq.c
201
cc_set_imsk_other(cpuid, cc_get_imsk_other(cpuid) | (1 << real_irq));
arch/sparc/kernel/sun4d_irq.c
213
int cpuid = handler_data->cpuid;
arch/sparc/kernel/sun4d_irq.c
220
cc_set_imsk_other(cpuid, cc_get_imsk_other(cpuid) & ~(1 << real_irq));
arch/sparc/kernel/sun4d_irq.c
254
int cpuid = cpu_logical_map(1);
arch/sparc/kernel/sun4d_irq.c
256
if (cpuid == -1)
arch/sparc/kernel/sun4d_irq.c
257
cpuid = cpu_logical_map(0);
arch/sparc/kernel/sun4d_irq.c
261
board_to_cpu[board] = cpuid;
arch/sparc/kernel/sun4d_irq.c
262
set_sbi_tid(devid, cpuid << 3);
arch/sparc/kernel/sun4d_irq.c
264
printk(KERN_ERR "All sbus IRQs directed to CPU%d\n", cpuid);
arch/sparc/kernel/sun4d_irq.c
312
handler_data->cpuid = board_to_cpu[board];
arch/sparc/kernel/sun4d_irq.c
33
unsigned int cpuid; /* target cpu */
arch/sparc/kernel/sun4d_smp.c
103
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/sun4d_smp.c
45
static inline void show_leds(int cpuid)
arch/sparc/kernel/sun4d_smp.c
47
cpuid &= 0x1e;
arch/sparc/kernel/sun4d_smp.c
49
"r" ((cpu_leds[cpuid] << 4) | cpu_leds[cpuid+1]),
arch/sparc/kernel/sun4d_smp.c
50
"r" (ECSR_BASE(cpuid) | BB_LEDS),
arch/sparc/kernel/sun4d_smp.c
56
int cpuid = hard_smp_processor_id();
arch/sparc/kernel/sun4d_smp.c
59
cpu_leds[cpuid] = 0x6;
arch/sparc/kernel/sun4d_smp.c
60
show_leds(cpuid);
arch/sparc/kernel/sun4d_smp.c
69
int cpuid;
arch/sparc/kernel/sun4d_smp.c
71
cpuid = hard_smp_processor_id();
arch/sparc/kernel/sun4d_smp.c
78
sun4d_swap((unsigned long *)&cpu_callin_map[cpuid], 1);
arch/sparc/kernel/sun4d_smp.c
82
while ((unsigned long)current_set[cpuid] < PAGE_OFFSET)
arch/sparc/kernel/sun4d_smp.c
85
while (current_set[cpuid]->cpu != cpuid)
arch/sparc/kernel/sun4d_smp.c
90
: : "r" (&current_set[cpuid])
arch/sparc/kernel/sun4d_smp.c
93
cpu_leds[cpuid] = 0x9;
arch/sparc/kernel/sun4d_smp.c
94
show_leds(cpuid);
arch/sparc/kernel/sun4m_smp.c
44
int cpuid = hard_smp_processor_id();
arch/sparc/kernel/sun4m_smp.c
51
swap_ulong(&cpu_callin_map[cpuid], 1);
arch/sparc/kernel/sun4m_smp.c
59
: : "r" (&current_set[cpuid])
arch/sparc/kernel/sun4m_smp.c
66
while (!cpumask_test_cpu(cpuid, &smp_commenced_mask))
arch/sparc/kernel/time_64.c
167
static unsigned long cpuid_to_freq(phandle node, int cpuid)
arch/sparc/kernel/time_64.c
180
if (is_cpu_node && (prom_getint(node, "upa-portid") == cpuid ||
arch/sparc/kernel/time_64.c
181
prom_getint(node, "cpuid") == cpuid))
arch/sparc/kernel/time_64.c
184
freq = cpuid_to_freq(prom_getchild(node), cpuid);
arch/sparc/kernel/time_64.c
186
freq = cpuid_to_freq(prom_getsibling(node), cpuid);
arch/sparc/prom/misc_64.c
393
void prom_startcpu_cpuid(int cpuid, unsigned long pc, unsigned long arg)
arch/sparc/prom/misc_64.c
400
args[3] = (unsigned int) cpuid;
arch/sparc/prom/misc_64.c
406
void prom_stopcpu_cpuid(int cpuid)
arch/sparc/prom/misc_64.c
413
args[3] = (unsigned int) cpuid;
arch/x86/boot/cpuflags.c
101
cpuid(0x80000000, &max_amd_level, &ignored, &ignored,
arch/x86/boot/cpuflags.c
106
cpuid(0x80000001, &ignored, &ignored, &cpu.flags[6],
arch/x86/boot/cpuflags.c
82
cpuid(0x0, &max_intel_level, &cpu_vendor[0], &cpu_vendor[2],
arch/x86/boot/cpuflags.c
87
cpuid(0x1, &tfms, &ignored, &cpu.flags[4],
arch/x86/events/amd/uncore.c
999
cpuid(EXT_PERFMON_DEBUG_FEATURES, &eax, &ebx.full, &ecx, &edx);
arch/x86/events/intel/core.c
5904
cpuid(ARCH_PERFMON_EXT_LEAF, &eax_0.full, &ebx_0.full, &ecx, &edx);
arch/x86/events/intel/core.c
7562
cpuid(10, &eax.full, &ebx.full, &fixed_mask, &edx.full);
arch/x86/events/intel/lbr.c
1599
cpuid(28, &eax.full, &ebx.full, &ecx.full, &unused_edx);
arch/x86/events/intel/pt.c
210
cpuid(CPUID_LEAF_TSC, &eax, &ebx, &ecx, &edx);
arch/x86/events/zhaoxin/core.c
520
cpuid(10, &eax.full, &ebx.full, &unused, &edx.full);
arch/x86/include/asm/alternative.h
74
u32 cpuid: 16; /* CPUID bit set for replacement */
arch/x86/include/asm/cpuid/api.h
104
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
113
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
122
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
199
cpuid(base, &eax, &signature[0], &signature[1], &signature[2]);
arch/x86/include/asm/cpuid/api.h
95
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/kvm_host.h
761
struct kvm_hypervisor_cpuid cpuid;
arch/x86/include/asm/paravirt.h
91
PVOP_VCALL4(pv_ops, cpu.cpuid, eax, ebx, ecx, edx);
arch/x86/include/asm/paravirt_types.h
69
void (*cpuid)(unsigned int *eax, unsigned int *ebx,
arch/x86/include/asm/xen/interface.h
388
#define XEN_CPUID XEN_EMULATE_PREFIX __ASM_FORM(cpuid)
arch/x86/include/uapi/asm/kvm.h
1017
struct kvm_cpuid2 cpuid;
arch/x86/include/uapi/asm/kvm.h
1040
struct kvm_cpuid2 cpuid;
arch/x86/include/uapi/asm/mce.h
26
__u32 cpuid; /* CPUID 1 EAX */
arch/x86/kernel/acpi/cstate.c
133
cpuid(CPUID_LEAF_MWAIT, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/alternative.c
614
BUG_ON(alt->cpuid >= (NCAPINTS + NBUGINTS) * 32);
arch/x86/kernel/alternative.c
621
if (!boot_cpu_has(alt->cpuid) != !(alt->flags & ALT_FLAG_NOT))
arch/x86/kernel/alternative.c
644
alt->cpuid >> 5, alt->cpuid & 0x1f,
arch/x86/kernel/cpu/amd.c
1192
cpuid(0x80000006, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/amd.c
1202
cpuid(0x80000005, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/amd.c
1221
cpuid(0x80000005, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/amd.c
437
u32 cpuid, assoc;
arch/x86/kernel/cpu/amd.c
439
cpuid = cpuid_edx(0x80000005);
arch/x86/kernel/cpu/amd.c
440
assoc = cpuid >> 16 & 0xff;
arch/x86/kernel/cpu/amd.c
441
upperbit = ((cpuid >> 24) << 10) / assoc;
arch/x86/kernel/cpu/cacheinfo.c
163
cpuid(0x80000005, &dummy, &dummy, &l1d.val, &l1i.val);
arch/x86/kernel/cpu/cacheinfo.c
164
cpuid(0x80000006, &dummy, &dummy, &l2.val, &l3.val);
arch/x86/kernel/cpu/centaur.c
191
cpuid(0x80000005, &aa, &bb, &cc, &dd);
arch/x86/kernel/cpu/common.c
1022
cpuid(0x00000001, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
1061
cpuid(0x80000001, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
1071
cpuid(0x80000008, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
1118
cpuid(0x80000008, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
833
cpuid(0x80000002, &v[0], &v[1], &v[2], &v[3]);
arch/x86/kernel/cpu/common.c
834
cpuid(0x80000003, &v[4], &v[5], &v[6], &v[7]);
arch/x86/kernel/cpu/common.c
835
cpuid(0x80000004, &v[8], &v[9], &v[10], &v[11]);
arch/x86/kernel/cpu/common.c
862
cpuid(0x80000005, &dummy, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
873
cpuid(0x80000006, &dummy, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/common.c
943
cpuid(0x00000000, (unsigned int *)&c->cpuid_level,
arch/x86/kernel/cpu/common.c
953
cpuid(0x00000001, &tfms, &misc, &junk, &cap0);
arch/x86/kernel/cpu/hygon.c
239
cpuid(0x80000006, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/hygon.c
255
cpuid(0x80000005, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/mce/core.c
123
m->cpuid = cpuid_eax(1);
arch/x86/kernel/cpu/mce/core.c
221
m->cpuvendor, m->cpuid, m->time, m->socketid, m->apicid,
arch/x86/kernel/cpu/mce/inject.c
122
m->cpuid = cpuid_eax(1);
arch/x86/kernel/cpu/mshyperv.c
347
cpuid(HYPERV_CPUID_VENDOR_AND_MAX_FUNCTIONS,
arch/x86/kernel/cpu/mshyperv.c
475
cpuid(HYPERV_CPUID_VERSION, &info->eax, &info->ebx, &info->ecx, &info->edx);
arch/x86/kernel/cpu/transmeta.c
37
cpuid(0x80860001, &dummy, &cpu_rev, &cpu_freq, &cpu_flags);
arch/x86/kernel/cpu/transmeta.c
48
cpuid(0x80860002, &new_cpu_rev, &cms_rev1, &cms_rev2, &dummy);
arch/x86/kernel/cpu/transmeta.c
61
cpuid(0x80860003,
arch/x86/kernel/cpu/transmeta.c
66
cpuid(0x80860004,
arch/x86/kernel/cpu/transmeta.c
71
cpuid(0x80860005,
arch/x86/kernel/cpu/transmeta.c
76
cpuid(0x80860006,
arch/x86/kernel/cpu/vmware.c
450
cpuid(CPUID_VMWARE_FEATURES_LEAF, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/cpu/vmware.c
468
cpuid(CPUID_VMWARE_INFO_LEAF, &eax, &hyper_vendor_id[0],
arch/x86/kernel/hpet.c
924
cpuid(CPUID_LEAF_MWAIT, &eax, &ebx, &ecx, &mwait_substates);
arch/x86/kernel/paravirt.c
107
.cpu.cpuid = native_cpuid,
arch/x86/kernel/process.c
893
cpuid(CPUID_LEAF_MWAIT, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/smpboot.c
172
int cpuid = smp_processor_id();
arch/x86/kernel/smpboot.c
190
identify_secondary_cpu(cpuid);
arch/x86/kernel/smpboot.c
196
set_cpu_sibling_map(cpuid);
arch/x86/kernel/smpboot.c
200
pr_debug("Stack at about %p\n", &cpuid);
arch/x86/kernel/smpboot.c
208
notify_cpu_starting(cpuid);
arch/x86/kernel/tsc.c
1104
cpuid(CPUID_LEAF_TSC, &art_base_clk.denominator,
arch/x86/kernel/tsc.c
686
cpuid(CPUID_LEAF_TSC, &eax_denominator, &ebx_numerator, &ecx_hz, &edx);
arch/x86/kernel/tsc.c
718
cpuid(CPUID_LEAF_FREQ, &eax_base_mhz, &ebx, &ecx, &edx);
arch/x86/kernel/tsc.c
758
cpuid(CPUID_LEAF_FREQ, &eax_base_mhz, &ebx_max_mhz, &ecx_bus_mhz, &edx);
arch/x86/kvm/cpuid.c
1976
int kvm_dev_ioctl_get_cpuid(struct kvm_cpuid2 *cpuid,
arch/x86/kvm/cpuid.c
1989
if (cpuid->nent < 1)
arch/x86/kvm/cpuid.c
1991
if (cpuid->nent > KVM_MAX_CPUID_ENTRIES)
arch/x86/kvm/cpuid.c
1992
cpuid->nent = KVM_MAX_CPUID_ENTRIES;
arch/x86/kvm/cpuid.c
1994
if (sanity_check_entries(entries, cpuid->nent, type))
arch/x86/kvm/cpuid.c
1997
array.entries = kvzalloc_objs(struct kvm_cpuid_entry2, cpuid->nent);
arch/x86/kvm/cpuid.c
2001
array.maxnent = cpuid->nent;
arch/x86/kvm/cpuid.c
2008
cpuid->nent = array.nent;
arch/x86/kvm/cpuid.c
210
struct kvm_hypervisor_cpuid cpuid = {};
arch/x86/kvm/cpuid.c
225
cpuid.base = base;
arch/x86/kvm/cpuid.c
226
cpuid.limit = entry->eax;
arch/x86/kvm/cpuid.c
232
return cpuid;
arch/x86/kvm/cpuid.c
392
const struct cpuid_reg cpuid = reverse_cpuid[i];
arch/x86/kvm/cpuid.c
395
if (!cpuid.function)
arch/x86/kvm/cpuid.c
398
entry = kvm_find_cpuid_entry_index(vcpu, cpuid.function, cpuid.index);
arch/x86/kvm/cpuid.c
402
cpuid_func_emulated(&emulated, cpuid.function, true);
arch/x86/kvm/cpuid.c
410
cpuid_get_reg_unsafe(&emulated, cpuid.reg);
arch/x86/kvm/cpuid.c
411
vcpu->arch.cpu_caps[i] &= cpuid_get_reg_unsafe(entry, cpuid.reg);
arch/x86/kvm/cpuid.c
573
vcpu->arch.xen.cpuid = kvm_get_hypervisor_cpuid(vcpu, XEN_SIGNATURE);
arch/x86/kvm/cpuid.c
590
struct kvm_cpuid *cpuid,
arch/x86/kvm/cpuid.c
597
if (cpuid->nent > KVM_MAX_CPUID_ENTRIES)
arch/x86/kvm/cpuid.c
600
if (cpuid->nent) {
arch/x86/kvm/cpuid.c
601
e = vmemdup_array_user(entries, cpuid->nent, sizeof(*e));
arch/x86/kvm/cpuid.c
605
e2 = kvmalloc_objs(*e2, cpuid->nent, GFP_KERNEL_ACCOUNT);
arch/x86/kvm/cpuid.c
611
for (i = 0; i < cpuid->nent; i++) {
arch/x86/kvm/cpuid.c
624
r = kvm_set_cpuid(vcpu, e2, cpuid->nent);
arch/x86/kvm/cpuid.c
635
struct kvm_cpuid2 *cpuid,
arch/x86/kvm/cpuid.c
641
if (cpuid->nent > KVM_MAX_CPUID_ENTRIES)
arch/x86/kvm/cpuid.c
644
if (cpuid->nent) {
arch/x86/kvm/cpuid.c
645
e2 = vmemdup_array_user(entries, cpuid->nent, sizeof(*e2));
arch/x86/kvm/cpuid.c
650
r = kvm_set_cpuid(vcpu, e2, cpuid->nent);
arch/x86/kvm/cpuid.c
658
struct kvm_cpuid2 *cpuid,
arch/x86/kvm/cpuid.c
661
if (cpuid->nent < vcpu->arch.cpuid_nent)
arch/x86/kvm/cpuid.c
671
cpuid->nent = vcpu->arch.cpuid_nent;
arch/x86/kvm/cpuid.c
675
static __always_inline u32 raw_cpuid_get(struct cpuid_reg cpuid)
arch/x86/kvm/cpuid.c
685
base = cpuid.function & 0xffff0000;
arch/x86/kvm/cpuid.c
689
if (cpuid_eax(base) < cpuid.function)
arch/x86/kvm/cpuid.c
692
cpuid_count(cpuid.function, cpuid.index,
arch/x86/kvm/cpuid.c
695
return *__cpuid_entry_get_reg(&entry, cpuid.reg);
arch/x86/kvm/cpuid.c
705
const struct cpuid_reg cpuid = x86_feature_cpuid(leaf * 32); \
arch/x86/kvm/cpuid.c
721
kvm_cpu_caps[leaf] &= (raw_cpuid_get(cpuid) | \
arch/x86/kvm/cpuid.h
105
const struct cpuid_reg cpuid = x86_feature_cpuid(x86_feature);
arch/x86/kvm/cpuid.h
125
entry = kvm_find_cpuid_entry_index(vcpu, cpuid.function, cpuid.index);
arch/x86/kvm/cpuid.h
129
reg = __cpuid_entry_get_reg(entry, cpuid.reg);
arch/x86/kvm/cpuid.h
50
int kvm_dev_ioctl_get_cpuid(struct kvm_cpuid2 *cpuid,
arch/x86/kvm/cpuid.h
54
struct kvm_cpuid *cpuid,
arch/x86/kvm/cpuid.h
57
struct kvm_cpuid2 *cpuid,
arch/x86/kvm/cpuid.h
60
struct kvm_cpuid2 *cpuid,
arch/x86/kvm/emulate.c
4407
II(ImplicitOps, em_cpuid, cpuid),
arch/x86/kvm/hyperv.c
2769
int kvm_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid,
arch/x86/kvm/hyperv.c
2790
if (cpuid->nent < nent)
arch/x86/kvm/hyperv.c
2793
if (cpuid->nent > nent)
arch/x86/kvm/hyperv.c
2794
cpuid->nent = nent;
arch/x86/kvm/hyperv.h
194
int kvm_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid,
arch/x86/kvm/reverse_cpuid.h
211
const struct cpuid_reg cpuid = x86_feature_cpuid(x86_feature);
arch/x86/kvm/reverse_cpuid.h
213
return __cpuid_entry_get_reg(entry, cpuid.reg);
arch/x86/kvm/svm/sev.c
3035
cpuid(0x8000001f, &eax, &ebx, &ecx, &edx);
arch/x86/kvm/vmx/tdx.c
204
caps->cpuid.nent = td_conf->num_cpuid_config;
arch/x86/kvm/vmx/tdx.c
210
td_init_cpuid_entry2(&caps->cpuid.entries[i], i);
arch/x86/kvm/vmx/tdx.c
2215
if (get_user(nr_user_entries, &user_caps->cpuid.nent))
arch/x86/kvm/vmx/tdx.c
2221
caps = kzalloc_flex(*caps, cpuid.entries, td_conf->num_cpuid_config);
arch/x86/kvm/vmx/tdx.c
2229
if (copy_to_user(user_caps, caps, struct_size(caps, cpuid.entries,
arch/x86/kvm/vmx/tdx.c
2230
caps->cpuid.nent))) {
arch/x86/kvm/vmx/tdx.c
2251
static int setup_tdparams_eptp_controls(struct kvm_cpuid2 *cpuid,
arch/x86/kvm/vmx/tdx.c
2257
entry = kvm_find_cpuid_entry2(cpuid->entries, cpuid->nent, 0x80000008, 0);
arch/x86/kvm/vmx/tdx.c
2280
static int setup_tdparams_cpuids(struct kvm_cpuid2 *cpuid,
arch/x86/kvm/vmx/tdx.c
2298
entry = kvm_find_cpuid_entry2(cpuid->entries, cpuid->nent,
arch/x86/kvm/vmx/tdx.c
2328
if (copy_cnt != cpuid->nent)
arch/x86/kvm/vmx/tdx.c
2338
struct kvm_cpuid2 *cpuid = &init_vm->cpuid;
arch/x86/kvm/vmx/tdx.c
2357
ret = setup_tdparams_eptp_controls(cpuid, td_params);
arch/x86/kvm/vmx/tdx.c
2361
ret = setup_tdparams_cpuids(cpuid, td_params);
arch/x86/kvm/vmx/tdx.c
2715
BUILD_BUG_ON(sizeof(*init_vm) != 256 + sizeof_field(struct kvm_tdx_init_vm, cpuid));
arch/x86/kvm/vmx/tdx.c
2724
if (get_user(nr_user_entries, &user_data->cpuid.nent))
arch/x86/kvm/vmx/tdx.c
2731
struct_size(user_data, cpuid.entries, nr_user_entries));
arch/x86/kvm/vmx/tdx.c
2740
if (init_vm->cpuid.padding) {
arch/x86/kvm/x86.c
4759
struct kvm_cpuid2 cpuid;
arch/x86/kvm/x86.c
4763
if (copy_from_user(&cpuid, cpuid_arg, sizeof(cpuid)))
arch/x86/kvm/x86.c
4766
r = kvm_get_hv_cpuid(vcpu, &cpuid, cpuid_arg->entries);
arch/x86/kvm/x86.c
4771
if (copy_to_user(cpuid_arg, &cpuid, sizeof(cpuid)))
arch/x86/kvm/x86.c
5078
struct kvm_cpuid2 cpuid;
arch/x86/kvm/x86.c
5081
if (copy_from_user(&cpuid, cpuid_arg, sizeof(cpuid)))
arch/x86/kvm/x86.c
5084
r = kvm_dev_ioctl_get_cpuid(&cpuid, cpuid_arg->entries,
arch/x86/kvm/x86.c
5090
if (copy_to_user(cpuid_arg, &cpuid, sizeof(cpuid)))
arch/x86/kvm/x86.c
6256
struct kvm_cpuid cpuid;
arch/x86/kvm/x86.c
6259
if (copy_from_user(&cpuid, cpuid_arg, sizeof(cpuid)))
arch/x86/kvm/x86.c
6261
r = kvm_vcpu_ioctl_set_cpuid(vcpu, &cpuid, cpuid_arg->entries);
arch/x86/kvm/x86.c
6266
struct kvm_cpuid2 cpuid;
arch/x86/kvm/x86.c
6269
if (copy_from_user(&cpuid, cpuid_arg, sizeof(cpuid)))
arch/x86/kvm/x86.c
6271
r = kvm_vcpu_ioctl_set_cpuid2(vcpu, &cpuid,
arch/x86/kvm/x86.c
6277
struct kvm_cpuid2 cpuid;
arch/x86/kvm/x86.c
6280
if (copy_from_user(&cpuid, cpuid_arg, sizeof(cpuid)))
arch/x86/kvm/x86.c
6282
r = kvm_vcpu_ioctl_get_cpuid2(vcpu, &cpuid,
arch/x86/kvm/x86.c
6287
if (copy_to_user(cpuid_arg, &cpuid, sizeof(cpuid)))
arch/x86/kvm/xen.h
56
vcpu->arch.xen.cpuid.base &&
arch/x86/kvm/xen.h
57
function <= vcpu->arch.xen.cpuid.limit &&
arch/x86/kvm/xen.h
58
function == (vcpu->arch.xen.cpuid.base | XEN_CPUID_LEAF(3));
arch/x86/xen/enlighten_hvm.c
114
cpuid(base + 4, &eax, &ebx, &ecx, &edx);
arch/x86/xen/enlighten_pv.c
1366
pv_ops.cpu.cpuid = xen_cpuid;
arch/x86/xen/pmu.c
103
cpuid(0xa, &eax, &ebx, &ecx, &edx);
drivers/acpi/acpi_pad.c
51
cpuid(CPUID_LEAF_MWAIT, &eax, &ebx, &ecx, &edx);
drivers/acpi/acpi_processor.c
528
int cpuid, type;
drivers/acpi/acpi_processor.c
568
cpuid = acpi_get_cpuid(handle, type, acpi_id);
drivers/acpi/acpi_processor.c
570
return !invalid_logical_cpuid(cpuid);
drivers/base/arch_topology.c
745
void update_siblings_masks(unsigned int cpuid)
drivers/base/arch_topology.c
747
struct cpu_topology *cpu_topo, *cpuid_topo = &cpu_topology[cpuid];
drivers/base/arch_topology.c
750
ret = detect_cache_attributes(cpuid);
drivers/base/arch_topology.c
758
if (last_level_cache_is_shared(cpu, cpuid)) {
drivers/base/arch_topology.c
760
cpumask_set_cpu(cpuid, &cpu_topo->llc_sibling);
drivers/base/arch_topology.c
766
cpumask_set_cpu(cpuid, &cpu_topo->core_sibling);
drivers/base/arch_topology.c
774
cpumask_set_cpu(cpuid, &cpu_topo->cluster_sibling);
drivers/base/arch_topology.c
780
cpumask_set_cpu(cpuid, &cpu_topo->thread_sibling);
drivers/base/arch_topology.c
961
void store_cpu_topology(unsigned int cpuid)
drivers/base/arch_topology.c
963
struct cpu_topology *cpuid_topo = &cpu_topology[cpuid];
drivers/base/arch_topology.c
969
cpuid_topo->core_id = cpuid;
drivers/base/arch_topology.c
970
cpuid_topo->package_id = cpu_to_node(cpuid);
drivers/base/arch_topology.c
973
cpuid, cpuid_topo->package_id, cpuid_topo->core_id,
drivers/base/arch_topology.c
977
update_siblings_masks(cpuid);
drivers/base/cpu.c
42
int cpuid = cpu->dev.id;
drivers/base/cpu.c
43
unregister_cpu_under_node(cpuid, from_nid);
drivers/base/cpu.c
44
register_cpu_under_node(cpuid, to_nid);
drivers/base/cpu.c
51
int cpuid = dev->id;
drivers/base/cpu.c
56
from_nid = cpu_to_node(cpuid);
drivers/base/cpu.c
82
to_nid = cpu_to_node(cpuid);
drivers/clocksource/timer-riscv.c
210
int cpuid, error;
drivers/clocksource/timer-riscv.c
221
cpuid = riscv_hartid_to_cpuid(hartid);
drivers/clocksource/timer-riscv.c
222
if (cpuid < 0) {
drivers/clocksource/timer-riscv.c
224
return cpuid;
drivers/clocksource/timer-riscv.c
227
if (cpuid != smp_processor_id())
drivers/cpufreq/acpi-cpufreq.c
182
static int check_est_cpu(unsigned int cpuid)
drivers/cpufreq/acpi-cpufreq.c
184
struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
drivers/cpufreq/acpi-cpufreq.c
189
static int check_amd_hwpstate_cpu(unsigned int cpuid)
drivers/cpufreq/acpi-cpufreq.c
191
struct cpuinfo_x86 *cpu = &cpu_data(cpuid);
drivers/cpufreq/longrun.c
143
cpuid(0x80860007, &eax, &ebx, &ecx, &edx);
drivers/cpufreq/longrun.c
211
cpuid(0x80860007, &eax, &ebx, &ecx, &edx);
drivers/cpufreq/longrun.c
222
cpuid(0x80860007, &eax, &ebx, &ecx, &edx);
drivers/cpufreq/p4-clockmod.c
158
int cpuid = 0;
drivers/cpufreq/p4-clockmod.c
166
cpuid = (c->x86 << 8) | (c->x86_model << 4) | c->x86_stepping;
drivers/cpufreq/p4-clockmod.c
167
switch (cpuid) {
drivers/cpufreq/pmac64-cpufreq.c
480
struct device_node *cpuid = NULL, *hwclock = NULL;
drivers/cpufreq/pmac64-cpufreq.c
490
cpuid = of_find_node_by_path("/u3@0,f8000000/i2c@f8001000/cpuid@a0");
drivers/cpufreq/pmac64-cpufreq.c
491
if (cpuid != NULL)
drivers/cpufreq/pmac64-cpufreq.c
492
eeprom = of_get_property(cpuid, "cpuid", NULL);
drivers/cpufreq/pmac64-cpufreq.c
640
of_node_put(cpuid);
drivers/cpufreq/powernow-k7.c
139
cpuid(0x80000007, &eax, &ebx, &ecx, &edx);
drivers/cpufreq/powernow-k7.c
444
pst->cpuid, pst->fsbspeed, pst->maxfid, pst->startvid);
drivers/cpufreq/powernow-k7.c
497
if ((etuple == pst->cpuid) &&
drivers/cpufreq/powernow-k7.c
50
u32 cpuid;
drivers/cpufreq/powernow-k8.c
482
cpuid(CPUID_FREQ_VOLT_CAPABILITIES, &eax, &ebx, &ecx, &edx);
drivers/cpufreq/powernow-k8.c
670
if ((psb->cpuid == 0x00000fc0) ||
drivers/cpufreq/powernow-k8.c
671
(psb->cpuid == 0x00000fe0)) {
drivers/cpufreq/powernow-k8.h
168
u32 cpuid;
drivers/cpufreq/speedstep-centrino.c
200
#define _BANIAS(cpuid, max, name) \
drivers/cpufreq/speedstep-centrino.c
201
{ .cpu_id = cpuid, \
drivers/cpufreq/tegra194-cpufreq.c
112
static void tegra234_get_cpu_cluster_id(u32 cpu, u32 *cpuid, u32 *clusterid)
drivers/cpufreq/tegra194-cpufreq.c
118
if (cpuid)
drivers/cpufreq/tegra194-cpufreq.c
119
*cpuid = MPIDR_AFFINITY_LEVEL(mpidr, 1);
drivers/cpufreq/tegra194-cpufreq.c
124
static int tegra234_get_cpu_ndiv(u32 cpu, u32 cpuid, u32 clusterid, u64 *ndiv)
drivers/cpufreq/tegra194-cpufreq.c
158
data->cpu_data[c->cpu].cpuid);
drivers/cpufreq/tegra194-cpufreq.c
207
static void tegra194_get_cpu_cluster_id(u32 cpu, u32 *cpuid, u32 *clusterid)
drivers/cpufreq/tegra194-cpufreq.c
213
if (cpuid)
drivers/cpufreq/tegra194-cpufreq.c
214
*cpuid = MPIDR_AFFINITY_LEVEL(mpidr, 0);
drivers/cpufreq/tegra194-cpufreq.c
368
static int tegra194_get_cpu_ndiv(u32 cpu, u32 cpuid, u32 clusterid, u64 *ndiv)
drivers/cpufreq/tegra194-cpufreq.c
398
ret = data->soc->ops->get_cpu_ndiv(cpu, data->cpu_data[cpu].cpuid, clusterid, &ndiv);
drivers/cpufreq/tegra194-cpufreq.c
43
u32 cpuid;
drivers/cpufreq/tegra194-cpufreq.c
62
void (*get_cpu_cluster_id)(u32 cpu, u32 *cpuid, u32 *clusterid);
drivers/cpufreq/tegra194-cpufreq.c
63
int (*get_cpu_ndiv)(u32 cpu, u32 cpuid, u32 clusterid, u64 *ndiv);
drivers/cpufreq/tegra194-cpufreq.c
688
u32 cpuid, clusterid;
drivers/cpufreq/tegra194-cpufreq.c
696
data->soc->ops->get_cpu_cluster_id(cpu, &cpuid, &clusterid);
drivers/cpufreq/tegra194-cpufreq.c
698
mpidr_id = (clusterid * data->soc->maxcpus_per_cluster) + cpuid;
drivers/cpufreq/tegra194-cpufreq.c
700
data->cpu_data[cpu].cpuid = cpuid;
drivers/crypto/hisilicon/sec/sec_drv.h
386
u32 cpuid;
drivers/dma/cv1800b-dmamux.c
104
unsigned int chid, devid, cpuid;
drivers/dma/cv1800b-dmamux.c
113
cpuid = dma_spec->args[1];
drivers/dma/cv1800b-dmamux.c
121
if (cpuid > MAX_DMA_CPU_ID) {
drivers/dma/cv1800b-dmamux.c
122
dev_err(&pdev->dev, "invalid cpu id: %u\n", cpuid);
drivers/dma/cv1800b-dmamux.c
136
if (map->peripheral == devid && map->cpu == cpuid)
drivers/dma/cv1800b-dmamux.c
155
map->cpu = cpuid;
drivers/dma/cv1800b-dmamux.c
162
DMAMUX_INT_CH_MASK(chid, cpuid),
drivers/dma/cv1800b-dmamux.c
163
DMAMUX_INT_CH_BIT(chid, cpuid));
drivers/dma/cv1800b-dmamux.c
170
chid, devid, cpuid);
drivers/dma/cv1800b-dmamux.c
47
#define DMAMUX_INT_BIT(chid, cpuid) \
drivers/dma/cv1800b-dmamux.c
48
BIT((cpuid) * DMAMUX_INT_BIT_PER_CPU + (chid))
drivers/dma/cv1800b-dmamux.c
49
#define DMAMUX_INTEN_BIT(cpuid) \
drivers/dma/cv1800b-dmamux.c
50
DMAMUX_INT_BIT(8, cpuid)
drivers/dma/cv1800b-dmamux.c
51
#define DMAMUX_INT_CH_BIT(chid, cpuid) \
drivers/dma/cv1800b-dmamux.c
52
(DMAMUX_INT_BIT(chid, cpuid) | DMAMUX_INTEN_BIT(cpuid))
drivers/dma/cv1800b-dmamux.c
57
#define DMAMUX_INT_CH_MASK(chid, cpuid) \
drivers/dma/cv1800b-dmamux.c
58
(DMAMUX_INT_MASK(chid) | DMAMUX_INTEN_BIT(cpuid))
drivers/edac/ie31200_edac.c
572
mce->cpuvendor, mce->cpuid, mce->time,
drivers/edac/igen6_edac.c
1055
mce->cpuvendor, mce->cpuid, mce->time,
drivers/edac/mce_amd.c
587
unsigned int fam = x86_family(m->cpuid);
drivers/edac/mce_amd.c
645
unsigned int fam = x86_family(m->cpuid);
drivers/edac/mce_amd.c
798
unsigned int fam = x86_family(m->cpuid);
drivers/edac/mce_amd.c
809
fam, x86_model(m->cpuid), x86_stepping(m->cpuid),
drivers/edac/pnd2_edac.c
1399
mce->cpuvendor, mce->cpuid, mce->time, mce->socketid, mce->apicid);
drivers/edac/sb_edac.c
3305
"%u APIC %x\n", mce->cpuvendor, mce->cpuid,
drivers/edac/skx_common.c
798
"%u APIC 0x%x\n", mce->cpuvendor, mce->cpuid,
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
105
u32 cpuid, u64 vector, bool start,
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
112
ret = scmi_imx_cpu_validate_cpuid(ph, cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
122
in->cpuid = cpu_to_le32(cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
139
static int scmi_imx_cpu_started(const struct scmi_protocol_handle *ph, u32 cpuid,
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
151
ret = scmi_imx_cpu_validate_cpuid(ph, cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
160
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
207
u32 cpuid)
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
218
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
225
dev_err(ph->dev, "i.MX cpu: Failed to get info of cpu(%u)\n", cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
45
__le32 cpuid;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
66
u32 cpuid)
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
70
if (cpuid >= info->nr_cpu)
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
77
u32 cpuid, bool start)
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
83
ret = scmi_imx_cpu_validate_cpuid(ph, cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
96
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
135
u32 lmid, u32 cpuid, u32 flags, u64 vector)
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
148
in->cpuid = cpu_to_le32(cpuid);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
54
__le32 cpuid;
drivers/firmware/efi/cper-x86.c
270
print_hex_dump(pfx, "", DUMP_PREFIX_OFFSET, 16, 4, proc->cpuid,
drivers/firmware/efi/cper-x86.c
271
sizeof(proc->cpuid), 0);
drivers/firmware/imx/sm-cpu.c
16
int scmi_imx_cpu_reset_vector_set(u32 cpuid, u64 vector, bool start, bool boot,
drivers/firmware/imx/sm-cpu.c
22
return imx_cpu_ops->cpu_reset_vector_set(ph, cpuid, vector, start,
drivers/firmware/imx/sm-cpu.c
27
int scmi_imx_cpu_start(u32 cpuid, bool start)
drivers/firmware/imx/sm-cpu.c
33
return imx_cpu_ops->cpu_start(ph, cpuid, true);
drivers/firmware/imx/sm-cpu.c
35
return imx_cpu_ops->cpu_start(ph, cpuid, false);
drivers/firmware/imx/sm-cpu.c
39
int scmi_imx_cpu_started(u32 cpuid, bool *started)
drivers/firmware/imx/sm-cpu.c
47
return imx_cpu_ops->cpu_started(ph, cpuid, started);
drivers/firmware/imx/sm-lmm.c
28
int scmi_imx_lmm_reset_vector_set(u32 lmid, u32 cpuid, u32 flags, u64 vector)
drivers/firmware/imx/sm-lmm.c
33
return imx_lmm_ops->lmm_reset_vector_set(ph, lmid, cpuid, flags, vector);
drivers/firmware/psci/psci.c
217
static int __psci_cpu_on(u32 fn, unsigned long cpuid, unsigned long entry_point)
drivers/firmware/psci/psci.c
221
err = invoke_psci_fn(fn, cpuid, entry_point, 0);
drivers/firmware/psci/psci.c
225
static int psci_0_1_cpu_on(unsigned long cpuid, unsigned long entry_point)
drivers/firmware/psci/psci.c
227
return __psci_cpu_on(psci_0_1_function_ids.cpu_on, cpuid, entry_point);
drivers/firmware/psci/psci.c
230
static int psci_0_2_cpu_on(unsigned long cpuid, unsigned long entry_point)
drivers/firmware/psci/psci.c
232
return __psci_cpu_on(PSCI_FN_NATIVE(0_2, CPU_ON), cpuid, entry_point);
drivers/firmware/psci/psci.c
235
static int __psci_migrate(u32 fn, unsigned long cpuid)
drivers/firmware/psci/psci.c
239
err = invoke_psci_fn(fn, cpuid, 0, 0);
drivers/firmware/psci/psci.c
243
static int psci_0_1_migrate(unsigned long cpuid)
drivers/firmware/psci/psci.c
245
return __psci_migrate(psci_0_1_function_ids.migrate, cpuid);
drivers/firmware/psci/psci.c
248
static int psci_0_2_migrate(unsigned long cpuid)
drivers/firmware/psci/psci.c
250
return __psci_migrate(PSCI_FN_NATIVE(0_2, MIGRATE), cpuid);
drivers/firmware/psci/psci.c
438
unsigned long cpuid;
drivers/firmware/psci/psci.c
442
cpuid = psci_migrate_info_up_cpu();
drivers/firmware/psci/psci.c
444
cpuid, resident_cpu);
drivers/firmware/psci/psci.c
599
unsigned long cpuid;
drivers/firmware/psci/psci.c
620
cpuid = psci_migrate_info_up_cpu();
drivers/firmware/psci/psci.c
621
if (cpuid & ~MPIDR_HWID_BITMASK) {
drivers/firmware/psci/psci.c
623
cpuid);
drivers/firmware/psci/psci.c
627
cpu = get_logical_index(cpuid);
drivers/firmware/psci/psci.c
630
pr_info("Trusted OS resident on physical CPU 0x%lx\n", cpuid);
drivers/gpu/drm/gma500/mmu.c
461
cpuid(0x00000001, &tfms, &misc, &cap0, &cap4);
drivers/hwmon/ibmpowernv.c
195
int cpuid = get_logical_cpu(id);
drivers/hwmon/ibmpowernv.c
197
if (cpuid >= 0)
drivers/hwmon/ibmpowernv.c
204
cpuid);
drivers/hwtracing/coresight/coresight-cti-platform.c
178
int cpuid = 0;
drivers/hwtracing/coresight/coresight-cti-platform.c
183
cpuid = cti_plat_get_cpu_at_node(dev_fwnode(dev));
drivers/hwtracing/coresight/coresight-cti-platform.c
184
if (cpuid < 0) {
drivers/hwtracing/coresight/coresight-cti-platform.c
189
cti_dev->cpu = cpuid;
drivers/hwtracing/coresight/coresight-cti-platform.c
204
scnprintf(cpu_name_str, sizeof(cpu_name_str), "cpu%d", cpuid);
drivers/hwtracing/coresight/coresight-cti-platform.c
345
int cpuid = -1, err = 0;
drivers/hwtracing/coresight/coresight-cti-platform.c
392
cpuid = cti_plat_get_cpu_at_node(fwnode);
drivers/hwtracing/coresight/coresight-cti-platform.c
393
if (cpuid >= 0) {
drivers/hwtracing/coresight/coresight-cti-platform.c
394
drvdata->ctidev.cpu = cpuid;
drivers/hwtracing/coresight/coresight-cti-platform.c
395
scnprintf(cpu_name_str, sizeof(cpu_name_str), "cpu%d", cpuid);
drivers/idle/intel_idle.c
2018
cpuid(7, &eax, &ebx, &ecx, &edx);
drivers/idle/intel_idle.c
2627
cpuid(CPUID_LEAF_MWAIT, &eax, &ebx, &ecx, &mwait_substates);
drivers/infiniband/hw/hfi1/sdma.c
1078
unsigned long cpuid)
drivers/infiniband/hw/hfi1/sdma.c
1083
rht_node = rhashtable_lookup_fast(dd->sdma_rht, &cpuid,
drivers/infiniband/hw/hfi1/sdma.c
1088
seq_printf(s, "cpu%3lu: ", cpuid);
drivers/infiniband/hw/hfi1/sdma.h
1035
unsigned long cpuid);
drivers/irqchip/irq-armada-370-xp.c
124
#define MPIC_INT_IRQ_FIQ_MASK(cpuid) ((BIT(0) | BIT(8)) << (cpuid))
drivers/irqchip/irq-armada-370-xp.c
385
u32 cpuid;
drivers/irqchip/irq-armada-370-xp.c
394
cpuid = cpu_logical_map(smp_processor_id());
drivers/irqchip/irq-armada-370-xp.c
397
writel(MPIC_INT_CAUSE_PERF(cpuid), mpic->per_cpu + MPIC_INT_FABRIC_MASK);
drivers/irqchip/irq-armada-370-xp.c
675
u32 irqsrc, cpuid;
drivers/irqchip/irq-armada-370-xp.c
681
cpuid = cpu_logical_map(smp_processor_id());
drivers/irqchip/irq-armada-370-xp.c
689
if (!(irqsrc & MPIC_INT_IRQ_FIQ_MASK(cpuid)))
drivers/irqchip/irq-gic-v5-irs.c
375
int gicv5_irs_cpu_to_iaffid(int cpuid, u16 *iaffid)
drivers/irqchip/irq-gic-v5-irs.c
377
if (!per_cpu(cpu_iaffid, cpuid).valid) {
drivers/irqchip/irq-gic-v5-irs.c
378
pr_err("IAFFID for CPU %d has not been initialised\n", cpuid);
drivers/irqchip/irq-gic-v5-irs.c
382
*iaffid = per_cpu(cpu_iaffid, cpuid).iaffid;
drivers/irqchip/irq-gic-v5-irs.c
507
int gicv5_irs_register_cpu(int cpuid)
drivers/irqchip/irq-gic-v5-irs.c
514
ret = gicv5_irs_cpu_to_iaffid(cpuid, &iaffid);
drivers/irqchip/irq-gic-v5-irs.c
516
pr_err("IAFFID for CPU %d has not been initialised\n", cpuid);
drivers/irqchip/irq-gic-v5-irs.c
520
irs_data = per_cpu(per_cpu_irs_data, cpuid);
drivers/irqchip/irq-gic-v5-irs.c
522
pr_err("No IRS associated with CPU %u\n", cpuid);
drivers/irqchip/irq-gic-v5-irs.c
542
pr_debug("CPU %d enabled PE IAFFID 0x%x\n", cpuid, iaffid);
drivers/irqchip/irq-gic-v5.c
250
int ret, cpuid;
drivers/irqchip/irq-gic-v5.c
255
cpuid = cpumask_first(mask_val);
drivers/irqchip/irq-gic-v5.c
257
cpuid = cpumask_any_and(mask_val, cpu_online_mask);
drivers/irqchip/irq-gic-v5.c
259
ret = gicv5_irs_cpu_to_iaffid(cpuid, &iaffid);
drivers/irqchip/irq-gic-v5.c
268
irq_data_update_effective_affinity(d, cpumask_of(cpuid));
drivers/irqchip/irq-vf610-mscm-ir.c
183
int ret, cpuid;
drivers/irqchip/irq-vf610-mscm-ir.c
209
regmap_read(mscm_cp_regmap, MSCM_CPxNUM, &cpuid);
drivers/irqchip/irq-vf610-mscm-ir.c
210
mscm_ir_data->cpu_mask = 0x1 << cpuid;
drivers/mailbox/mailbox-mchp-ipc-sbi.c
374
int cpuid, ret;
drivers/mailbox/mailbox-mchp-ipc-sbi.c
378
for_each_online_cpu(cpuid) {
drivers/mailbox/mailbox-mchp-ipc-sbi.c
379
hartid = cpuid_to_hartid_map(cpuid);
drivers/mailbox/mailbox-mchp-ipc-sbi.c
385
ipc->cluster_cfg[cpuid].irq = ret;
drivers/mailbox/mailbox-mchp-ipc-sbi.c
386
ret = devm_request_irq(ipc->dev, ipc->cluster_cfg[cpuid].irq,
drivers/mailbox/mailbox-mchp-ipc-sbi.c
392
ipc->cluster_cfg[cpuid].buf_base = devm_kmalloc(ipc->dev,
drivers/mailbox/mailbox-mchp-ipc-sbi.c
396
if (!ipc->cluster_cfg[cpuid].buf_base)
drivers/mailbox/mailbox-mchp-ipc-sbi.c
399
ipc->cluster_cfg[cpuid].buf_base_addr = __pa(ipc->cluster_cfg[cpuid].buf_base);
drivers/misc/sgi-xp/xp_main.c
55
int (*xp_cpu_to_nasid) (int cpuid);
drivers/misc/sgi-xp/xp_uv.c
82
xp_cpu_to_nasid_uv(int cpuid)
drivers/misc/sgi-xp/xp_uv.c
85
return UV_PNODE_TO_NASID(uv_cpu_to_pnode(cpuid));
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
1284
void hw_atl_tdm_cpu_id_set(struct aq_hw_s *aq_hw, u32 cpuid, u32 dca)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
1288
HW_ATL_TDM_DCADCPUID_SHIFT, cpuid);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
394
void hw_atl_rdm_cpu_id_set(struct aq_hw_s *aq_hw, u32 cpuid, u32 dca)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
398
HW_ATL_RDM_DCADCPUID_SHIFT, cpuid);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
186
void hw_atl_rdm_cpu_id_set(struct aq_hw_s *aq_hw, u32 cpuid, u32 dca);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
603
void hw_atl_tdm_cpu_id_set(struct aq_hw_s *aq_hw, u32 cpuid, u32 dca);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
678
{I40E_HMC_STORE(i40e_hmc_obj_txq, cpuid), 8, 96 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
698
{ I40E_HMC_STORE(i40e_hmc_obj_rxq, cpuid), 8, 13 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
23
u16 cpuid; /* bigger than needed, see above for reason */
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
63
u8 cpuid;
drivers/net/ethernet/intel/ice/ice_common.c
1387
ICE_CTX_STORE(ice_rlan_ctx, cpuid, 8, 13),
drivers/net/ethernet/intel/ice/ice_common.c
1501
ICE_CTX_STORE(ice_tlan_ctx, cpuid, 8, 93),
drivers/net/ethernet/intel/ice/ice_common.c
1709
ICE_CTX_STORE(ice_txtime_ctx, cpuid, 8, 82),
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
384
u8 cpuid;
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
556
u8 cpuid;
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
598
u8 cpuid;
drivers/parisc/sba_iommu.c
1150
if ( ((boot_cpu_data.pdc.cpuid >> 5) & 0x7f) != 0x13
drivers/perf/arm_pmu_acpi.c
251
unsigned long cpuid = read_cpuid_id();
drivers/perf/arm_pmu_acpi.c
257
if (!pmu || pmu->acpi_cpuid != cpuid)
drivers/perf/arm_pmu_acpi.c
340
unsigned long cpuid)
drivers/perf/arm_pmu_acpi.c
347
if (cpu_cpuid == cpuid)
drivers/perf/arm_pmu_acpi.c
383
unsigned long cpuid;
drivers/perf/arm_pmu_acpi.c
397
cpuid = per_cpu(cpu_data, cpu).reg_midr;
drivers/perf/arm_pmu_acpi.c
398
pmu->acpi_cpuid = cpuid;
drivers/perf/arm_pmu_acpi.c
400
arm_pmu_acpi_probe_matching_cpus(pmu, cpuid);
drivers/perf/arm_pmu_platform.c
29
unsigned int cpuid = read_cpuid_id();
drivers/perf/arm_pmu_platform.c
35
if ((cpuid & info->mask) != info->cpuid)
drivers/perf/riscv_pmu.c
389
int cpuid, i;
drivers/perf/riscv_pmu.c
402
for_each_possible_cpu(cpuid) {
drivers/perf/riscv_pmu.c
403
cpuc = per_cpu_ptr(pmu->hw_events, cpuid);
drivers/perf/starfive_starlink_pmu.c
512
int cpuid, i, ret;
drivers/perf/starfive_starlink_pmu.c
530
for_each_possible_cpu(cpuid) {
drivers/perf/starfive_starlink_pmu.c
531
hw_events = per_cpu_ptr(starlink_pmu->hw_events, cpuid);
drivers/platform/x86/intel/pmc/core.c
1043
cpuid(CPUID_LEAF_TSC, &eax_denominator, &ebx_numerator, &ecx_hz, &edx);
drivers/ras/amd/fmpm.c
567
u64 cpuid;
drivers/ras/amd/fmpm.c
586
cpuid = cpuid_eax(1);
drivers/ras/amd/fmpm.c
587
if (fmp->fru_arch != cpuid) {
drivers/ras/amd/fmpm.c
589
fmp->fru_arch, cpuid);
drivers/remoteproc/imx_rproc.c
1150
ret = scmi_imx_cpu_started(dcfg->cpuid, &started);
drivers/remoteproc/imx_rproc.c
1152
dev_err(dev, "Failed to detect cpu(%d) status: %d\n", dcfg->cpuid, ret);
drivers/remoteproc/imx_rproc.c
1459
.cpuid = 1, /* Use 1 as cpu id for M7 core */
drivers/remoteproc/imx_rproc.c
348
ret = scmi_imx_cpu_reset_vector_set(dcfg->cpuid, 0, true, false, false);
drivers/remoteproc/imx_rproc.c
350
dev_err(priv->dev, "Failed to set reset vector cpuid(%u): %d\n", dcfg->cpuid, ret);
drivers/remoteproc/imx_rproc.c
354
return scmi_imx_cpu_start(dcfg->cpuid, true);
drivers/remoteproc/imx_rproc.c
368
ret = scmi_imx_lmm_reset_vector_set(dcfg->lmid, dcfg->cpuid, 0, 0);
drivers/remoteproc/imx_rproc.c
371
dcfg->lmid, dcfg->cpuid, ret);
drivers/remoteproc/imx_rproc.c
445
return scmi_imx_cpu_start(dcfg->cpuid, false);
drivers/remoteproc/imx_rproc.h
42
u32 cpuid; /* ID of the remote core */
drivers/s390/cio/css.c
852
struct cpuid cpu_id;
drivers/scsi/qla2xxx/qla_def.h
3546
int cpuid;
drivers/scsi/qla2xxx/qla_def.h
3884
uint16_t cpuid;
drivers/scsi/qla2xxx/qla_inline.h
330
qla_cpu_update(struct qla_qpair *qpair, uint16_t cpuid)
drivers/scsi/qla2xxx/qla_inline.h
332
qpair->cpuid = cpuid;
drivers/scsi/qla2xxx/qla_inline.h
338
h->cpuid = qpair->cpuid;
drivers/scsi/qla2xxx/qla_inline.h
580
int cpuid = raw_smp_processor_id();
drivers/scsi/qla2xxx/qla_inline.h
582
if (qpair->cpuid != cpuid &&
drivers/scsi/qla2xxx/qla_inline.h
583
ha->qp_cpu_map[cpuid]) {
drivers/scsi/qla2xxx/qla_inline.h
584
qpair = ha->qp_cpu_map[cpuid];
drivers/scsi/qla2xxx/qla_inline.h
602
qpair->cpuid = cpumask_first(mask);
drivers/scsi/qla2xxx/qla_inline.h
606
msix->cpuid = qpair->cpuid;
drivers/scsi/qla2xxx/qla_isr.c
3984
if (rsp->qpair->cpuid != raw_smp_processor_id() || !rsp->qpair->rcv_intr) {
drivers/scsi/qla2xxx/qla_target.c
2077
mcmd->se_cmd.cpuid = h->cpuid;
drivers/scsi/qla2xxx/qla_target.c
2089
mcmd->se_cmd.cpuid = abort_cmd->se_cmd.cpuid;
drivers/scsi/qla2xxx/qla_target.c
2095
queue_work_on(mcmd->se_cmd.cpuid, qla_tgt_wq, &mcmd->work);
drivers/scsi/qla2xxx/qla_target.c
3986
queue_work_on(cmd->se_cmd.cpuid, qla_tgt_wq, &tgt->srr_work);
drivers/scsi/qla2xxx/qla_target.c
4589
cmd->se_cmd.cpuid = h->cpuid;
drivers/scsi/qla2xxx/qla_target.c
4736
queue_work_on(cmd->se_cmd.cpuid, qla_tgt_wq, &cmd->work);
drivers/scsi/qla2xxx/qla_target.c
4741
queue_work_on(cmd->se_cmd.cpuid, qla_tgt_wq,
drivers/scsi/qla2xxx/qla_target.c
4780
mcmd->se_cmd.cpuid = h->cpuid;
drivers/scsi/qla2xxx/qla_target.c
4792
mcmd->se_cmd.cpuid = h->cpuid;
drivers/scsi/qla2xxx/qla_target.c
4805
queue_work_on(mcmd->se_cmd.cpuid, qla_tgt_wq,
drivers/scsi/qla2xxx/qla_target.c
7489
h->cpuid = ha->base_qpair->cpuid;
drivers/scsi/qla2xxx/qla_target.c
7504
h->cpuid = qpair->cpuid;
drivers/scsi/qla2xxx/qla_target.h
789
u16 cpuid;
drivers/scsi/qla2xxx/tcm_qla2xxx.c
463
if (se_cmd->cpuid != WORK_CPU_UNBOUND)
drivers/soc/samsung/exynos-asv.c
110
cpuid);
drivers/soc/samsung/exynos-asv.c
85
int ret, cpuid;
drivers/soc/samsung/exynos-asv.c
87
for_each_possible_cpu(cpuid) {
drivers/soc/samsung/exynos-asv.c
90
cpu = get_cpu_device(cpuid);
drivers/soc/tegra/flowctrl.c
106
reg |= TEGRA20_FLOW_CTRL_CSR_WFE_CPU0 << cpuid;
drivers/soc/tegra/flowctrl.c
109
reg |= TEGRA30_FLOW_CTRL_CSR_WFI_CPU0 << cpuid;
drivers/soc/tegra/flowctrl.c
116
flowctrl_write_cpu_csr(cpuid, reg);
drivers/soc/tegra/flowctrl.c
119
if (i == cpuid)
drivers/soc/tegra/flowctrl.c
128
void flowctrl_cpu_suspend_exit(unsigned int cpuid)
drivers/soc/tegra/flowctrl.c
133
reg = flowctrl_read_cpu_csr(cpuid);
drivers/soc/tegra/flowctrl.c
153
flowctrl_write_cpu_csr(cpuid, reg);
drivers/soc/tegra/flowctrl.c
51
u32 flowctrl_read_cpu_csr(unsigned int cpuid)
drivers/soc/tegra/flowctrl.c
53
u8 offset = flowctrl_offset_cpu_csr[cpuid];
drivers/soc/tegra/flowctrl.c
62
void flowctrl_write_cpu_csr(unsigned int cpuid, u32 value)
drivers/soc/tegra/flowctrl.c
64
return flowctrl_update(flowctrl_offset_cpu_csr[cpuid], value);
drivers/soc/tegra/flowctrl.c
67
void flowctrl_write_cpu_halt(unsigned int cpuid, u32 value)
drivers/soc/tegra/flowctrl.c
69
return flowctrl_update(flowctrl_offset_halt_cpu[cpuid], value);
drivers/soc/tegra/flowctrl.c
72
void flowctrl_cpu_suspend_enter(unsigned int cpuid)
drivers/soc/tegra/flowctrl.c
77
reg = flowctrl_read_cpu_csr(cpuid);
drivers/soc/tegra/flowctrl.c
85
reg |= TEGRA20_FLOW_CTRL_CSR_WFE_CPU0 << cpuid;
drivers/soc/tegra/pmc.c
1186
unsigned int cpuid)
drivers/soc/tegra/pmc.c
1188
if (pmc->soc && cpuid < pmc->soc->num_cpu_powergates)
drivers/soc/tegra/pmc.c
1189
return pmc->soc->cpu_powergates[cpuid];
drivers/soc/tegra/pmc.c
1198
bool tegra_pmc_cpu_is_powered(unsigned int cpuid)
drivers/soc/tegra/pmc.c
1202
id = tegra_get_cpu_powergate_id(pmc, cpuid);
drivers/soc/tegra/pmc.c
1213
int tegra_pmc_cpu_power_on(unsigned int cpuid)
drivers/soc/tegra/pmc.c
1217
id = tegra_get_cpu_powergate_id(pmc, cpuid);
drivers/soc/tegra/pmc.c
1228
int tegra_pmc_cpu_remove_clamping(unsigned int cpuid)
drivers/soc/tegra/pmc.c
1232
id = tegra_get_cpu_powergate_id(pmc, cpuid);
drivers/soc/ux500/ux500-soc-id.c
85
unsigned int cpuid = read_cpuid_id();
drivers/soc/ux500/ux500-soc-id.c
89
switch (cpuid) {
drivers/target/target_core_transport.c
1476
cmd->cpuid = raw_smp_processor_id();
drivers/target/target_core_transport.c
1920
int cpu = se_cmd->cpuid;
drivers/target/target_core_transport.c
704
spin_lock_irqsave(&dev->queues[cmd->cpuid].lock, flags);
drivers/target/target_core_transport.c
709
spin_unlock_irqrestore(&dev->queues[cmd->cpuid].lock, flags);
drivers/target/target_core_transport.c
940
cpu = cmd->cpuid;
drivers/target/target_core_transport.c
987
spin_lock_irqsave(&dev->queues[cmd->cpuid].lock, flags);
drivers/target/target_core_transport.c
990
&dev->queues[cmd->cpuid].state_list);
drivers/target/target_core_transport.c
993
spin_unlock_irqrestore(&dev->queues[cmd->cpuid].lock, flags);
drivers/thermal/intel/x86_pkg_temp_thermal.c
326
cpuid(6, &eax, &ebx, &ecx, &edx);
drivers/usb/gadget/function/f_tcm.c
1358
cmd->se_cmd.cpuid = cpu;
drivers/xen/pcpu.c
82
.u.cpu_ol.cpuid = cpu_id,
drivers/xen/pcpu.c
93
.u.cpu_ol.cpuid = cpu_id,
include/linux/acpi.h
287
static inline bool invalid_logical_cpuid(u32 cpuid)
include/linux/acpi.h
289
return (int)cpuid < 0;
include/linux/arch_topology.h
89
void store_cpu_topology(unsigned int cpuid);
include/linux/arch_topology.h
93
void remove_cpu_topology(unsigned int cpuid);
include/linux/cper.h
409
u8 cpuid[48];
include/linux/firmware/imx/sm.h
47
int scmi_imx_cpu_start(u32 cpuid, bool start);
include/linux/firmware/imx/sm.h
48
int scmi_imx_cpu_started(u32 cpuid, bool *started);
include/linux/firmware/imx/sm.h
49
int scmi_imx_cpu_reset_vector_set(u32 cpuid, u64 vector, bool start, bool boot,
include/linux/firmware/imx/sm.h
52
static inline int scmi_imx_cpu_start(u32 cpuid, bool start)
include/linux/firmware/imx/sm.h
57
static inline int scmi_imx_cpu_started(u32 cpuid, bool *started)
include/linux/firmware/imx/sm.h
62
static inline int scmi_imx_cpu_reset_vector_set(u32 cpuid, u64 vector, bool start,
include/linux/firmware/imx/sm.h
82
int scmi_imx_lmm_reset_vector_set(u32 lmid, u32 cpuid, u32 flags, u64 vector);
include/linux/firmware/imx/sm.h
94
static inline int scmi_imx_lmm_reset_vector_set(u32 lmid, u32 cpuid, u32 flags, u64 vector)
include/linux/irqchip/arm-gic-v5.h
361
int gicv5_irs_register_cpu(int cpuid);
include/linux/perf/arm_pmu.h
150
unsigned int cpuid;
include/linux/perf/arm_pmu.h
157
.cpuid = (_cpuid), \
include/linux/psci.h
28
int (*cpu_on)(unsigned long cpuid, unsigned long entry_point);
include/linux/psci.h
29
int (*migrate)(unsigned long cpuid);
include/linux/scmi_imx_protocol.h
101
int (*cpu_started)(const struct scmi_protocol_handle *ph, u32 cpuid,
include/linux/scmi_imx_protocol.h
90
u32 lmid, u32 cpuid, u32 flags, u64 vector);
include/linux/scmi_imx_protocol.h
97
u32 cpuid, u64 vector, bool start,
include/linux/scmi_imx_protocol.h
99
int (*cpu_start)(const struct scmi_protocol_handle *ph, u32 cpuid,
include/linux/smp.h
50
int smp_call_function_single(int cpuid, smp_call_func_t func, void *info,
include/net/page_pool/types.h
170
int cpuid;
include/net/page_pool/types.h
264
int cpuid);
include/soc/tegra/flowctrl.h
44
u32 flowctrl_read_cpu_csr(unsigned int cpuid);
include/soc/tegra/flowctrl.h
45
void flowctrl_write_cpu_csr(unsigned int cpuid, u32 value);
include/soc/tegra/flowctrl.h
46
void flowctrl_write_cpu_halt(unsigned int cpuid, u32 value);
include/soc/tegra/flowctrl.h
48
void flowctrl_cpu_suspend_enter(unsigned int cpuid);
include/soc/tegra/flowctrl.h
49
void flowctrl_cpu_suspend_exit(unsigned int cpuid);
include/soc/tegra/flowctrl.h
51
static inline u32 flowctrl_read_cpu_csr(unsigned int cpuid)
include/soc/tegra/flowctrl.h
56
static inline void flowctrl_write_cpu_csr(unsigned int cpuid, u32 value)
include/soc/tegra/flowctrl.h
60
static inline void flowctrl_write_cpu_halt(unsigned int cpuid, u32 value) {}
include/soc/tegra/flowctrl.h
62
static inline void flowctrl_cpu_suspend_enter(unsigned int cpuid)
include/soc/tegra/flowctrl.h
66
static inline void flowctrl_cpu_suspend_exit(unsigned int cpuid)
include/soc/tegra/pmc.h
21
bool tegra_pmc_cpu_is_powered(unsigned int cpuid);
include/soc/tegra/pmc.h
22
int tegra_pmc_cpu_power_on(unsigned int cpuid);
include/soc/tegra/pmc.h
23
int tegra_pmc_cpu_remove_clamping(unsigned int cpuid);
include/target/target_core_base.h
571
int cpuid;
include/trace/events/mce.h
39
__field( u32, cpuid )
include/trace/events/mce.h
62
__entry->cpuid = err->m.cpuid;
include/trace/events/mce.h
84
__entry->cpuid,
include/xen/interface/platform.h
438
uint32_t cpuid;
include/xen/interface/xen-mca.h
351
__u32 cpuid; /* CPUID 1 EAX */
lib/zstd/common/cpu.h
83
ZSTD_cpuid_t cpuid;
lib/zstd/common/cpu.h
84
cpuid.f1c = f1c;
lib/zstd/common/cpu.h
85
cpuid.f1d = f1d;
lib/zstd/common/cpu.h
86
cpuid.f7b = f7b;
lib/zstd/common/cpu.h
87
cpuid.f7c = f7c;
lib/zstd/common/cpu.h
88
return cpuid;
lib/zstd/common/cpu.h
93
MEM_STATIC int ZSTD_cpuid_##name(ZSTD_cpuid_t const cpuid) { \
lib/zstd/common/cpu.h
94
return ((cpuid.r) & (1U << bit)) != 0; \
lib/zstd/common/zstd_internal.h
314
ZSTD_cpuid_t cpuid = ZSTD_cpuid();
lib/zstd/common/zstd_internal.h
315
return ZSTD_cpuid_bmi1(cpuid) && ZSTD_cpuid_bmi2(cpuid);
net/core/dev.c
13159
static int net_page_pool_create(int cpuid)
net/core/dev.c
13165
.nid = cpu_to_mem(cpuid),
net/core/dev.c
13170
pp_ptr = page_pool_create_percpu(&page_pool_params, cpuid);
net/core/dev.c
13180
per_cpu(system_page_pool.pool, cpuid) = pp_ptr;
net/core/page_pool.c
1276
WRITE_ONCE(pool->cpuid, -1);
net/core/page_pool.c
195
int cpuid)
net/core/page_pool.c
206
pool->cpuid = cpuid;
net/core/page_pool.c
338
page_pool_create_percpu(const struct page_pool_params *params, int cpuid)
net/core/page_pool.c
347
err = page_pool_init(pool, params, cpuid);
net/core/page_pool.c
875
u32 cpuid;
net/core/page_pool.c
890
cpuid = smp_processor_id();
net/core/page_pool.c
891
if (READ_ONCE(pool->cpuid) == cpuid)
net/core/page_pool.c
896
return napi && READ_ONCE(napi->list_owner) == cpuid;
net/smc/smc_ism.c
54
struct cpuid id;
tools/arch/s390/include/uapi/asm/kvm.h
413
__u64 cpuid;
tools/arch/s390/include/uapi/asm/kvm.h
422
__u64 cpuid;
tools/arch/x86/include/uapi/asm/kvm.h
1017
struct kvm_cpuid2 cpuid;
tools/arch/x86/include/uapi/asm/kvm.h
1040
struct kvm_cpuid2 cpuid;
tools/arch/x86/kcpuid/kcpuid.c
243
cpuid(range->index, max_func, ebx, ecx, edx);
tools/arch/x86/kcpuid/kcpuid.c
268
cpuid(f, eax, ebx, ecx, edx);
tools/perf/arch/arm64/util/arm-spe.c
103
char *cpuid = NULL;
tools/perf/arch/arm64/util/arm-spe.c
107
cpuid = get_cpuid_allow_env_override(cpu);
tools/perf/arch/arm64/util/arm-spe.c
108
if (!cpuid)
tools/perf/arch/arm64/util/arm-spe.c
110
val = strtol(cpuid, NULL, 16);
tools/perf/arch/arm64/util/arm-spe.c
138
free(cpuid);
tools/perf/arch/loongarch/util/header.c
46
char *line, *model, *cpuid;
tools/perf/arch/loongarch/util/header.c
53
line = model = cpuid = NULL;
tools/perf/arch/loongarch/util/header.c
64
if (model && (asprintf(&cpuid, "%s", model) < 0))
tools/perf/arch/loongarch/util/header.c
65
cpuid = NULL;
tools/perf/arch/loongarch/util/header.c
70
return cpuid;
tools/perf/arch/loongarch/util/header.c
76
char *cpuid = _get_cpuid();
tools/perf/arch/loongarch/util/header.c
78
if (!cpuid)
tools/perf/arch/loongarch/util/header.c
81
if (sz < strlen(cpuid)) {
tools/perf/arch/loongarch/util/header.c
86
scnprintf(buffer, sz, "%s", cpuid);
tools/perf/arch/loongarch/util/header.c
89
free(cpuid);
tools/perf/arch/riscv/util/header.c
42
char *cpuid = NULL;
tools/perf/arch/riscv/util/header.c
49
return cpuid;
tools/perf/arch/riscv/util/header.c
72
if (asprintf(&cpuid, "%s-%s-%s", mvendorid, marchid, mimpid) < 0)
tools/perf/arch/riscv/util/header.c
73
cpuid = NULL;
tools/perf/arch/riscv/util/header.c
81
return cpuid;
tools/perf/arch/riscv/util/header.c
86
char *cpuid = _get_cpuid();
tools/perf/arch/riscv/util/header.c
89
if (sz < strlen(cpuid)) {
tools/perf/arch/riscv/util/header.c
94
scnprintf(buffer, sz, "%s", cpuid);
tools/perf/arch/riscv/util/header.c
96
free(cpuid);
tools/perf/arch/x86/tests/intel-pt-test.c
366
cpuid(20, i, &r.eax, &r.ebx, &r.ecx, &r.edx);
tools/perf/arch/x86/tests/intel-pt-test.c
383
cpuid(7, 0, &eax, &ebx, &ecx, &edx);
tools/perf/arch/x86/util/header.c
18
cpuid(0, 0, lvl, &b, &c, &d);
tools/perf/arch/x86/util/header.c
36
cpuid(1, 0, &a, &b, &c, &d);
tools/perf/arch/x86/util/intel-pt.c
314
cpuid(0x15, 0, &eax, &ebx, &ecx, &edx);
tools/perf/arch/x86/util/pmu.c
32
char *cpuid = get_cpuid_str((struct perf_cpu){0});
tools/perf/arch/x86/util/pmu.c
34
is_graniterapids = cpuid && strcmp_cpuid_str(graniterapids_cpuid, cpuid) == 0;
tools/perf/arch/x86/util/pmu.c
35
free(cpuid);
tools/perf/arch/x86/util/tsc.c
84
cpuid(0x15, 0, &a, &b, &c, &d);
tools/perf/builtin-kvm.c
1168
char buf[128], *cpuid;
tools/perf/builtin-kvm.c
1181
cpuid = buf;
tools/perf/builtin-kvm.c
1183
cpuid = perf_session__env(kvm->session)->cpuid;
tools/perf/builtin-kvm.c
1185
if (!cpuid) {
tools/perf/builtin-kvm.c
1191
err = cpu_isa_init(kvm, e_machine, cpuid);
tools/perf/builtin-kvm.c
1193
pr_err("CPU %s is not supported.\n", cpuid);
tools/perf/pmu-events/empty-pmu-events.c
2795
const char *cpuid;
tools/perf/pmu-events/empty-pmu-events.c
2807
.cpuid = "common",
tools/perf/pmu-events/empty-pmu-events.c
2819
.cpuid = "testcpu",
tools/perf/pmu-events/empty-pmu-events.c
2831
.cpuid = 0,
tools/perf/pmu-events/empty-pmu-events.c
3154
char *cpuid;
tools/perf/pmu-events/empty-pmu-events.c
3158
char *cpuid = NULL;
tools/perf/pmu-events/empty-pmu-events.c
3164
cpuid = get_cpuid_allow_env_override(cpu);
tools/perf/pmu-events/empty-pmu-events.c
3170
if (!cpuid)
tools/perf/pmu-events/empty-pmu-events.c
3173
if (has_last_map_search && !strcmp(last_map_search.cpuid, cpuid)) {
tools/perf/pmu-events/empty-pmu-events.c
3175
free(cpuid);
tools/perf/pmu-events/empty-pmu-events.c
3186
if (!strcmp_cpuid_str(map->cpuid, cpuid))
tools/perf/pmu-events/empty-pmu-events.c
3189
free(last_map_search.cpuid);
tools/perf/pmu-events/empty-pmu-events.c
3190
last_map_search.cpuid = cpuid;
tools/perf/pmu-events/empty-pmu-events.c
3252
if (!strcmp(map->cpuid, "common"))
tools/perf/pmu-events/empty-pmu-events.c
3276
if (!strcmp(map->cpuid, "common"))
tools/perf/pmu-events/empty-pmu-events.c
3282
const struct pmu_events_table *find_core_events_table(const char *arch, const char *cpuid)
tools/perf/pmu-events/empty-pmu-events.c
3287
if (!strcmp(tables->arch, arch) && !strcmp_cpuid_str(tables->cpuid, cpuid))
tools/perf/pmu-events/empty-pmu-events.c
3293
const struct pmu_metrics_table *find_core_metrics_table(const char *arch, const char *cpuid)
tools/perf/pmu-events/empty-pmu-events.c
3298
if (!strcmp(tables->arch, arch) && !strcmp_cpuid_str(tables->cpuid, cpuid))
tools/perf/pmu-events/pmu-events.h
132
const struct pmu_events_table *find_core_events_table(const char *arch, const char *cpuid);
tools/perf/pmu-events/pmu-events.h
133
const struct pmu_metrics_table *find_core_metrics_table(const char *arch, const char *cpuid);
tools/perf/tests/expr.c
248
bool is_intel = strstr(cpuid, "Intel") != NULL;
tools/perf/tests/expr.c
278
escaped_cpuid1 = strreplace_chars('-', cpuid, "\\-");
tools/perf/tests/expr.c
279
free(cpuid);
tools/perf/tests/expr.c
80
char *cpuid = get_cpuid_allow_env_override(cpu);
tools/perf/tests/expr.c
83
TEST_ASSERT_VAL("get_cpuid", cpuid);
tools/perf/util/amd-sample-raw.c
338
const char *cpuid;
tools/perf/util/amd-sample-raw.c
341
cpuid = perf_env__cpuid(env);
tools/perf/util/amd-sample-raw.c
345
ret = sscanf(cpuid, "%*[^,],%u,%u", &cpu_family, &cpu_model);
tools/perf/util/annotate-arch/annotate-arc.c
7
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-arm.c
34
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-arm64.c
92
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-csky.c
44
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-loongarch.c
144
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-mips.c
41
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-powerpc.c
394
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-riscv64.c
29
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-s390.c
151
static int s390__cpuid_parse(struct arch *arch, const char *cpuid)
tools/perf/util/annotate-arch/annotate-s390.c
161
ret = sscanf(cpuid, "%*[^,],%u,%[^,],%[^,],%[^,],%s", &family, model_c,
tools/perf/util/annotate-arch/annotate-s390.c
172
const struct arch *arch__new_s390(const struct e_machine_and_e_flags *id, const char *cpuid)
tools/perf/util/annotate-arch/annotate-s390.c
182
if (cpuid) {
tools/perf/util/annotate-arch/annotate-s390.c
183
if (s390__cpuid_parse(arch, cpuid)) {
tools/perf/util/annotate-arch/annotate-sparc.c
165
const char *cpuid __maybe_unused)
tools/perf/util/annotate-arch/annotate-x86.c
185
static int x86__cpuid_parse(struct arch *arch, const char *cpuid)
tools/perf/util/annotate-arch/annotate-x86.c
193
ret = sscanf(cpuid, "%*[^,],%u,%u,%u", &family, &model, &stepping);
tools/perf/util/annotate-arch/annotate-x86.c
197
arch->ins_is_fused = strstarts(cpuid, "AuthenticAMD") ?
tools/perf/util/annotate-arch/annotate-x86.c
780
const struct arch *arch__new_x86(const struct e_machine_and_e_flags *id, const char *cpuid)
tools/perf/util/annotate-arch/annotate-x86.c
789
if (cpuid) {
tools/perf/util/annotate-arch/annotate-x86.c
790
if (x86__cpuid_parse(arch, cpuid)) {
tools/perf/util/annotate.c
1000
arch = arch__find(e_machine, e_flags, machine->env ? machine->env->cpuid : NULL);
tools/perf/util/arm-spe.c
972
const char *cpuid;
tools/perf/util/arm-spe.c
975
cpuid = perf_env__cpuid(perf_session__env(spe->session));
tools/perf/util/arm-spe.c
976
midr = strtol(cpuid, NULL, 16);
tools/perf/util/data-convert-json.c
306
output_json_key_string(out, true, 2, "cpuid", env->cpuid);
tools/perf/util/disasm.c
137
const struct arch *arch__find(uint16_t e_machine, uint32_t e_flags, const char *cpuid)
tools/perf/util/disasm.c
140
const char *cpuid) = {
tools/perf/util/disasm.c
182
result = arch_new_fn[e_machine](&key, cpuid);
tools/perf/util/disasm.h
111
const struct arch *arch__find(uint16_t e_machine, uint32_t e_flags, const char *cpuid);
tools/perf/util/disasm.h
125
const struct arch *arch__new_arc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
126
const struct arch *arch__new_arm(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
127
const struct arch *arch__new_arm64(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
128
const struct arch *arch__new_csky(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
129
const struct arch *arch__new_loongarch(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
130
const struct arch *arch__new_mips(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
131
const struct arch *arch__new_powerpc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
132
const struct arch *arch__new_riscv64(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
133
const struct arch *arch__new_s390(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
134
const struct arch *arch__new_sparc(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/disasm.h
135
const struct arch *arch__new_x86(const struct e_machine_and_e_flags *id, const char *cpuid);
tools/perf/util/env.c
258
zfree(&env->cpuid);
tools/perf/util/env.c
429
char cpuid[128];
tools/perf/util/env.c
431
int err = get_cpuid(cpuid, sizeof(cpuid), cpu);
tools/perf/util/env.c
436
free(env->cpuid);
tools/perf/util/env.c
437
env->cpuid = strdup(cpuid);
tools/perf/util/env.c
438
if (env->cpuid == NULL)
tools/perf/util/env.c
658
if (!env->cpuid) { /* Assume local operation */
tools/perf/util/env.c
664
return env->cpuid;
tools/perf/util/env.c
818
is_amd = env->cpuid && strstarts(env->cpuid, "AuthenticAMD") ? 1 : -1;
tools/perf/util/env.c
840
is_intel = env->cpuid && strstarts(env->cpuid, "GenuineIntel") ? 1 : -1;
tools/perf/util/env.h
83
char *cpuid;
tools/perf/util/expr.c
461
char *cpuid = get_cpuid_allow_env_override(cpu);
tools/perf/util/expr.c
463
if (!cpuid)
tools/perf/util/expr.c
466
ret = !strcmp_cpuid_str(test_id, cpuid);
tools/perf/util/expr.c
468
free(cpuid);
tools/perf/util/header.c
2235
fprintf(fp, "# cpuid : %s\n", ff->ph->env.cpuid);
tools/perf/util/header.c
2634
FEAT_PROCESS_STR_FUN(cpuid, cpuid);
tools/perf/util/header.c
3739
FEAT_OPR(CPUID, cpuid, false),
tools/perf/util/header.c
847
char *cpuid;
tools/perf/util/header.c
850
cpuid = getenv("PERF_CPUID");
tools/perf/util/header.c
851
if (cpuid)
tools/perf/util/header.c
852
cpuid = strdup(cpuid);
tools/perf/util/header.c
853
if (!cpuid)
tools/perf/util/header.c
854
cpuid = get_cpuid_str(cpu);
tools/perf/util/header.c
855
if (!cpuid)
tools/perf/util/header.c
859
pr_debug("Using CPUID %s\n", cpuid);
tools/perf/util/header.c
862
return cpuid;
tools/perf/util/header.c
869
int __weak strcmp_cpuid_str(const char *mapcpuid, const char *cpuid)
tools/perf/util/header.c
881
match = !regexec(&re, cpuid, 1, pmatch, 0);
tools/perf/util/header.c
887
if (match_len == strlen(cpuid))
tools/perf/util/intel-pt.c
1383
if (env->cpuid && !strncmp(env->cpuid, "GenuineIntel,6,92,", 18))
tools/perf/util/intel-pt.c
3573
if (!env->cpuid)
tools/perf/util/intel-pt.c
3584
if (!strncmp(env->cpuid, "GenuineIntel,6,190,", 19))
tools/perf/util/intel-pt.c
3587
if (!strncmp(env->cpuid, "GenuineIntel,6,175,", 19) ||
tools/perf/util/intel-pt.c
3588
!strncmp(env->cpuid, "GenuineIntel,6,182,", 19))
tools/perf/util/kvm-stat-arch/kvm-stat-s390.c
96
int __cpu_isa_init_s390(struct perf_kvm_stat *kvm, const char *cpuid)
tools/perf/util/kvm-stat-arch/kvm-stat-s390.c
98
if (strstr(cpuid, "IBM")) {
tools/perf/util/kvm-stat-arch/kvm-stat-x86.c
198
int __cpu_isa_init_x86(struct perf_kvm_stat *kvm, const char *cpuid)
tools/perf/util/kvm-stat-arch/kvm-stat-x86.c
200
if (strstr(cpuid, "Intel")) {
tools/perf/util/kvm-stat-arch/kvm-stat-x86.c
203
} else if (strstr(cpuid, "AMD") || strstr(cpuid, "Hygon")) {
tools/perf/util/kvm-stat.c
102
return __cpu_isa_init_x86(kvm, cpuid);
tools/perf/util/kvm-stat.c
86
int cpu_isa_init(struct perf_kvm_stat *kvm, uint16_t e_machine, const char *cpuid)
tools/perf/util/kvm-stat.c
99
return __cpu_isa_init_s390(kvm, cpuid);
tools/perf/util/kvm-stat.h
146
int cpu_isa_init(struct perf_kvm_stat *kvm, uint16_t e_machine, const char *cpuid);
tools/perf/util/kvm-stat.h
151
int __cpu_isa_init_s390(struct perf_kvm_stat *kvm, const char *cpuid);
tools/perf/util/kvm-stat.h
152
int __cpu_isa_init_x86(struct perf_kvm_stat *kvm, const char *cpuid);
tools/perf/util/s390-cpumsf.c
1056
static int s390_cpumsf_get_type(const char *cpuid)
tools/perf/util/s390-cpumsf.c
1060
ret = sscanf(cpuid, "%*[^,],%u", &family);
tools/perf/util/s390-cpumsf.c
1145
sf->machine_type = s390_cpumsf_get_type(perf_session__env(session)->cpuid);
tools/perf/util/sample-raw.c
18
const char *cpuid = perf_env__cpuid(env);
tools/perf/util/sample-raw.c
23
cpuid && strstarts(cpuid, "AuthenticAMD") &&
tools/power/cpupower/debug/i386/dump_psb.c
113
if (relevant!= pst->cpuid)
tools/power/cpupower/debug/i386/dump_psb.c
119
pst->cpuid,
tools/power/cpupower/debug/i386/dump_psb.c
48
u_int32_t cpuid;
tools/testing/selftests/kvm/arm64/vgic_irq.c
805
static void guest_code_asym_dir(struct test_args *args, int cpuid)
tools/testing/selftests/kvm/arm64/vgic_irq.c
812
if (cpuid == 0) {
tools/testing/selftests/kvm/arm64/vgic_irq.c
849
static void guest_code_group_en(struct test_args *args, int cpuid)
tools/testing/selftests/kvm/arm64/vgic_irq.c
897
static void guest_code_timer_spi(struct test_args *args, int cpuid)
tools/testing/selftests/kvm/arm64/vgic_irq.c
908
gic_set_priority(MIN_SPI + cpuid, IRQ_DEFAULT_PRIO);
tools/testing/selftests/kvm/arm64/vgic_irq.c
909
gic_irq_set_pending(MIN_SPI + cpuid);
tools/testing/selftests/kvm/arm64/vgic_irq.c
923
gic_irq_enable(MIN_SPI + cpuid);
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
101
u32 cpuid = guest_get_vcpuid();
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
105
test_data.lpi_pend_tables + (cpuid * SZ_64K));
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
109
if (cpuid > 0)
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
123
for (cpuid = 0; cpuid < test_data.nr_cpus; cpuid++)
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
124
its_send_sync_cmd(test_data.cmdq_base_va, cpuid);
tools/testing/selftests/kvm/include/kvm_util.h
66
struct kvm_cpuid2 *cpuid;
tools/testing/selftests/kvm/include/x86/processor.h
1040
struct kvm_cpuid2 *cpuid;
tools/testing/selftests/kvm/include/x86/processor.h
1042
cpuid = malloc(kvm_cpuid2_size(nr_entries));
tools/testing/selftests/kvm/include/x86/processor.h
1043
TEST_ASSERT(cpuid, "-ENOMEM when allocating kvm_cpuid2");
tools/testing/selftests/kvm/include/x86/processor.h
1045
cpuid->nent = nr_entries;
tools/testing/selftests/kvm/include/x86/processor.h
1047
return cpuid;
tools/testing/selftests/kvm/include/x86/processor.h
1050
void vcpu_init_cpuid(struct kvm_vcpu *vcpu, const struct kvm_cpuid2 *cpuid);
tools/testing/selftests/kvm/include/x86/processor.h
1054
vcpu_ioctl(vcpu, KVM_GET_CPUID2, vcpu->cpuid);
tools/testing/selftests/kvm/include/x86/processor.h
1061
TEST_ASSERT(vcpu->cpuid, "Must do vcpu_init_cpuid() first (or equivalent)");
tools/testing/selftests/kvm/include/x86/processor.h
1065
return (struct kvm_cpuid_entry2 *)get_cpuid_entry(vcpu->cpuid,
tools/testing/selftests/kvm/include/x86/processor.h
1079
TEST_ASSERT(vcpu->cpuid, "Must do vcpu_init_cpuid() first");
tools/testing/selftests/kvm/include/x86/processor.h
1080
r = __vcpu_ioctl(vcpu, KVM_SET_CPUID2, vcpu->cpuid);
tools/testing/selftests/kvm/include/x86/processor.h
1091
TEST_ASSERT(vcpu->cpuid, "Must do vcpu_init_cpuid() first");
tools/testing/selftests/kvm/include/x86/processor.h
1092
vcpu_ioctl(vcpu, KVM_SET_CPUID2, vcpu->cpuid);
tools/testing/selftests/kvm/include/x86/processor.h
684
cpuid(1, &eax, &ebx, &ecx, &edx);
tools/testing/selftests/kvm/include/x86/processor.h
703
cpuid(0, &eax, &ebx, &ecx, &edx);
tools/testing/selftests/kvm/include/x86/processor.h
949
const struct kvm_cpuid_entry2 *get_cpuid_entry(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/include/x86/processor.h
968
bool kvm_cpuid_has(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/include/x86/processor.h
976
uint32_t kvm_cpuid_property(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/lib/x86/hyperv.c
13
static struct kvm_cpuid2 *cpuid;
tools/testing/selftests/kvm/lib/x86/hyperv.c
16
if (cpuid)
tools/testing/selftests/kvm/lib/x86/hyperv.c
17
return cpuid;
tools/testing/selftests/kvm/lib/x86/hyperv.c
19
cpuid = allocate_kvm_cpuid2(MAX_NR_CPUID_ENTRIES);
tools/testing/selftests/kvm/lib/x86/hyperv.c
22
kvm_ioctl(kvm_fd, KVM_GET_SUPPORTED_HV_CPUID, cpuid);
tools/testing/selftests/kvm/lib/x86/hyperv.c
25
return cpuid;
tools/testing/selftests/kvm/lib/x86/hyperv.c
63
struct kvm_cpuid2 *cpuid = allocate_kvm_cpuid2(MAX_NR_CPUID_ENTRIES);
tools/testing/selftests/kvm/lib/x86/hyperv.c
65
vcpu_ioctl(vcpu, KVM_GET_SUPPORTED_HV_CPUID, cpuid);
tools/testing/selftests/kvm/lib/x86/hyperv.c
67
return cpuid;
tools/testing/selftests/kvm/lib/x86/processor.c
1001
TEST_ASSERT(cpuid != vcpu->cpuid, "@cpuid can't be the vCPU's CPUID");
tools/testing/selftests/kvm/lib/x86/processor.c
1004
if (vcpu->cpuid && vcpu->cpuid->nent < cpuid->nent) {
tools/testing/selftests/kvm/lib/x86/processor.c
1005
free(vcpu->cpuid);
tools/testing/selftests/kvm/lib/x86/processor.c
1006
vcpu->cpuid = NULL;
tools/testing/selftests/kvm/lib/x86/processor.c
1009
if (!vcpu->cpuid)
tools/testing/selftests/kvm/lib/x86/processor.c
1010
vcpu->cpuid = allocate_kvm_cpuid2(cpuid->nent);
tools/testing/selftests/kvm/lib/x86/processor.c
1012
memcpy(vcpu->cpuid, cpuid, kvm_cpuid2_size(cpuid->nent));
tools/testing/selftests/kvm/lib/x86/processor.c
1030
TEST_ASSERT_EQ(kvm_cpuid_property(vcpu->cpuid, property), value);
tools/testing/selftests/kvm/lib/x86/processor.c
1325
const struct kvm_cpuid_entry2 *get_cpuid_entry(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/lib/x86/processor.c
1330
for (i = 0; i < cpuid->nent; i++) {
tools/testing/selftests/kvm/lib/x86/processor.c
1331
if (cpuid->entries[i].function == function &&
tools/testing/selftests/kvm/lib/x86/processor.c
1332
cpuid->entries[i].index == index)
tools/testing/selftests/kvm/lib/x86/processor.c
1333
return &cpuid->entries[i];
tools/testing/selftests/kvm/lib/x86/processor.c
882
if (vcpu->cpuid)
tools/testing/selftests/kvm/lib/x86/processor.c
883
free(vcpu->cpuid);
tools/testing/selftests/kvm/lib/x86/processor.c
906
static uint32_t __kvm_cpu_has(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/lib/x86/processor.c
913
for (i = 0; i < cpuid->nent; i++) {
tools/testing/selftests/kvm/lib/x86/processor.c
914
entry = &cpuid->entries[i];
tools/testing/selftests/kvm/lib/x86/processor.c
928
bool kvm_cpuid_has(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/lib/x86/processor.c
931
return __kvm_cpu_has(cpuid, feature.function, feature.index,
tools/testing/selftests/kvm/lib/x86/processor.c
935
uint32_t kvm_cpuid_property(const struct kvm_cpuid2 *cpuid,
tools/testing/selftests/kvm/lib/x86/processor.c
938
return __kvm_cpu_has(cpuid, property.function, property.index,
tools/testing/selftests/kvm/lib/x86/processor.c
999
void vcpu_init_cpuid(struct kvm_vcpu *vcpu, const struct kvm_cpuid2 *cpuid)
tools/testing/selftests/kvm/s390/ucontrol_test.c
137
TH_LOG("create VM 0x%llx", info.cpuid);
tools/testing/selftests/kvm/x86/cpuid_test.c
143
struct kvm_cpuid2 *vcpu_alloc_cpuid(struct kvm_vm *vm, vm_vaddr_t *p_gva, struct kvm_cpuid2 *cpuid)
tools/testing/selftests/kvm/x86/cpuid_test.c
145
int size = sizeof(*cpuid) + cpuid->nent * sizeof(cpuid->entries[0]);
tools/testing/selftests/kvm/x86/cpuid_test.c
149
memcpy(guest_cpuids, cpuid, size);
tools/testing/selftests/kvm/x86/cpuid_test.c
200
struct kvm_cpuid2 *cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent + 1);
tools/testing/selftests/kvm/x86/cpuid_test.c
203
vcpu_ioctl(vcpu, KVM_GET_CPUID2, cpuid);
tools/testing/selftests/kvm/x86/cpuid_test.c
204
TEST_ASSERT(cpuid->nent == vcpu->cpuid->nent,
tools/testing/selftests/kvm/x86/cpuid_test.c
206
vcpu->cpuid->nent, cpuid->nent);
tools/testing/selftests/kvm/x86/cpuid_test.c
208
for (i = 0; i < vcpu->cpuid->nent; i++) {
tools/testing/selftests/kvm/x86/cpuid_test.c
209
cpuid->nent = i;
tools/testing/selftests/kvm/x86/cpuid_test.c
210
r = __vcpu_ioctl(vcpu, KVM_GET_CPUID2, cpuid);
tools/testing/selftests/kvm/x86/cpuid_test.c
212
TEST_ASSERT(cpuid->nent == i, "KVM modified nent on failure");
tools/testing/selftests/kvm/x86/cpuid_test.c
214
free(cpuid);
tools/testing/selftests/kvm/x86/cpuid_test.c
226
compare_cpuids(kvm_get_supported_cpuid(), vcpu->cpuid);
tools/testing/selftests/kvm/x86/cpuid_test.c
228
vcpu_alloc_cpuid(vm, &cpuid_gva, vcpu->cpuid);
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
117
static struct kvm_cpuid2 cpuid = {.nent = 0};
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
121
ret = __vcpu_ioctl(vcpu, KVM_GET_SUPPORTED_HV_CPUID, &cpuid);
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
123
ret = __kvm_ioctl(vm->kvm_fd, KVM_GET_SUPPORTED_HV_CPUID, &cpuid);
tools/testing/selftests/kvm/x86/hyperv_features.c
154
prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent);
tools/testing/selftests/kvm/x86/hyperv_features.c
494
memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent));
tools/testing/selftests/kvm/x86/hyperv_features.c
546
prev_cpuid = allocate_kvm_cpuid2(vcpu->cpuid->nent);
tools/testing/selftests/kvm/x86/hyperv_features.c
663
memcpy(prev_cpuid, vcpu->cpuid, kvm_cpuid2_size(vcpu->cpuid->nent));
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
382
cpuid(1, &eax, &ebx, &ecx, &edx);
tools/testing/selftests/kvm/x86/set_sregs_test.c
89
(vcpu->cpuid && vcpu_cpuid_has(vcpu, X86_FEATURE_OSXSAVE)),
tools/testing/selftests/kvm/x86/set_sregs_test.c
94
(vcpu->cpuid && vcpu_cpuid_has(vcpu, X86_FEATURE_OSPKE)),
tools/testing/selftests/net/psock_fanout.c
532
static int set_cpuaffinity(int cpuid)
tools/testing/selftests/net/psock_fanout.c
537
CPU_SET(cpuid, &mask);
tools/testing/selftests/net/psock_fanout.c
540
fprintf(stderr, "setaffinity %d\n", cpuid);