lib/libsys/_libsys.h
869
int __sys_membarrier(int cmd, unsigned flags, int cpu_id);
sys/amd64/amd64/initcpu.c
100
switch (CPUID_TO_FAMILY(cpu_id)) {
sys/amd64/amd64/initcpu.c
114
if (CPUID_TO_FAMILY(cpu_id) == 0x10) {
sys/amd64/amd64/initcpu.c
128
if (CPUID_TO_FAMILY(cpu_id) == 0x10) {
sys/amd64/amd64/initcpu.c
142
if (CPUID_TO_FAMILY(cpu_id) == 0x16 && CPUID_TO_MODEL(cpu_id) <= 0xf) {
sys/amd64/amd64/initcpu.c
151
if (CPUID_TO_FAMILY(cpu_id) == 0x17 && CPUID_TO_MODEL(cpu_id) == 0x1 &&
sys/amd64/amd64/initcpu.c
191
if (CPUID_TO_FAMILY(cpu_id) == 0x17 ||
sys/amd64/amd64/initcpu.c
192
CPUID_TO_FAMILY(cpu_id) == 0x18) {
sys/amd64/amd64/initcpu.c
82
if ((CPUID_TO_FAMILY(cpu_id) == 0xf ||
sys/amd64/amd64/initcpu.c
83
CPUID_TO_FAMILY(cpu_id) == 0x10) && (cpu_feature2 & CPUID2_HV) == 0)
sys/amd64/amd64/pmap.c
2357
(CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/amd64/amd64/pmap.c
2358
(CPUID_TO_MODEL(cpu_id) == 0x26 || /* Atoms */
sys/amd64/amd64/pmap.c
2359
CPUID_TO_MODEL(cpu_id) == 0x27 ||
sys/amd64/amd64/pmap.c
2360
CPUID_TO_MODEL(cpu_id) == 0x35 ||
sys/amd64/amd64/pmap.c
2361
CPUID_TO_MODEL(cpu_id) == 0x36 ||
sys/amd64/amd64/pmap.c
2362
CPUID_TO_MODEL(cpu_id) == 0x37 ||
sys/amd64/amd64/pmap.c
2363
CPUID_TO_MODEL(cpu_id) == 0x86 ||
sys/amd64/amd64/pmap.c
2364
CPUID_TO_MODEL(cpu_id) == 0x1c ||
sys/amd64/amd64/pmap.c
2365
CPUID_TO_MODEL(cpu_id) == 0x4a ||
sys/amd64/amd64/pmap.c
2366
CPUID_TO_MODEL(cpu_id) == 0x4c ||
sys/amd64/amd64/pmap.c
2367
CPUID_TO_MODEL(cpu_id) == 0x4d ||
sys/amd64/amd64/pmap.c
2368
CPUID_TO_MODEL(cpu_id) == 0x5a ||
sys/amd64/amd64/pmap.c
2369
CPUID_TO_MODEL(cpu_id) == 0x5c ||
sys/amd64/amd64/pmap.c
2370
CPUID_TO_MODEL(cpu_id) == 0x5d ||
sys/amd64/amd64/pmap.c
2371
CPUID_TO_MODEL(cpu_id) == 0x5f ||
sys/amd64/amd64/pmap.c
2372
CPUID_TO_MODEL(cpu_id) == 0x6e ||
sys/amd64/amd64/pmap.c
2373
CPUID_TO_MODEL(cpu_id) == 0x7a ||
sys/amd64/amd64/pmap.c
2374
CPUID_TO_MODEL(cpu_id) == 0x57 || /* Knights */
sys/amd64/amd64/pmap.c
2375
CPUID_TO_MODEL(cpu_id) == 0x85))))
sys/amd64/amd64/pmap.c
2514
CPUID_TO_FAMILY(cpu_id) == 0x6 && CPUID_TO_MODEL(cpu_id) == 0x55) {
sys/amd64/pt/pt.c
158
pt_cpu_get_state(int cpu_id)
sys/amd64/pt/pt.c
160
return (atomic_load_int(&pt_pcpu[cpu_id].state));
sys/amd64/pt/pt.c
164
pt_cpu_set_state(int cpu_id, enum pt_cpu_state state)
sys/amd64/pt/pt.c
166
atomic_store_int(&pt_pcpu[cpu_id].state, state);
sys/amd64/pt/pt.c
424
pt_backend_configure(struct hwt_context *ctx, int cpu_id, int thread_id)
sys/amd64/pt/pt.c
467
if (hwt_cpu->cpu_id != cpu_id)
sys/amd64/pt/pt.c
469
pt_ctx = &pt_pcpu_ctx[cpu_id];
sys/amd64/pt/pt.c
505
pt_pcpu[cpu_id].ctx = pt_ctx;
sys/amd64/pt/pt.c
514
pt_backend_enable(struct hwt_context *ctx, int cpu_id)
sys/amd64/pt/pt.c
519
KASSERT(curcpu == cpu_id,
sys/amd64/pt/pt.c
522
CPU_SET(cpu_id, &ctx->cpu_map);
sys/amd64/pt/pt.c
529
pt_backend_disable(struct hwt_context *ctx, int cpu_id)
sys/amd64/pt/pt.c
535
KASSERT(curcpu == cpu_id,
sys/amd64/pt/pt.c
538
cpu = &pt_pcpu[cpu_id];
sys/amd64/pt/pt.c
541
cpu_id);
sys/amd64/pt/pt.c
542
pt_cpu_set_state(cpu_id, PT_INACTIVE);
sys/amd64/pt/pt.c
547
CPU_CLR(cpu_id, &ctx->cpu_map);
sys/amd64/pt/pt.c
587
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/amd64/pt/pt.c
588
cpu = &pt_pcpu[cpu_id];
sys/amd64/pt/pt.c
590
__func__, cpu_id);
sys/amd64/pt/pt.c
591
pt_cpu_set_state(cpu_id, PT_INACTIVE);
sys/amd64/pt/pt.c
616
error = pt_init_ctx(&pt_pcpu_ctx[hwt_cpu->cpu_id], hwt_cpu->vm,
sys/amd64/pt/pt.c
617
hwt_cpu->cpu_id);
sys/amd64/pt/pt.c
636
int cpu_id;
sys/amd64/pt/pt.c
649
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/amd64/pt/pt.c
650
if (pt_pcpu[cpu_id].ctx == NULL)
sys/amd64/pt/pt.c
652
KASSERT(pt_pcpu[cpu_id].ctx == &pt_pcpu_ctx[cpu_id],
sys/amd64/pt/pt.c
656
pt_deinit_ctx(pt_pcpu[cpu_id].ctx);
sys/amd64/pt/pt.c
657
pt_pcpu[cpu_id].ctx = NULL;
sys/amd64/pt/pt.c
658
atomic_set_int(&pt_pcpu[cpu_id].in_pcint_handler, 0);
sys/amd64/pt/pt.c
678
buf = &pt_pcpu[vm->cpu->cpu_id].ctx->buf;
sys/amd64/pt/pt.c
723
pt_backend_dump(int cpu_id)
sys/amd64/vmm/intel/vmx_msr.c
184
family = CPUID_TO_FAMILY(cpu_id);
sys/amd64/vmm/intel/vmx_msr.c
185
model = CPUID_TO_MODEL(cpu_id);
sys/amd64/vmm/intel/vmx_msr.c
209
family = CPUID_TO_FAMILY(cpu_id);
sys/amd64/vmm/intel/vmx_msr.c
210
model = CPUID_TO_MODEL(cpu_id);
sys/amd64/vmm/x86.c
286
if (!vmm_is_svm() || CPUID_TO_FAMILY(cpu_id) < 0x16)
sys/arm/arm/machdep.c
228
cpu_est_clockrate(int cpu_id, uint64_t *rate)
sys/arm/arm/machdep.c
232
pc = pcpu_find(cpu_id);
sys/arm/nvidia/tegra124/tegra124_coretemp.c
159
sc->tsens_id = 0x100 + sc->cpu_id; //cells[0];
sys/arm/nvidia/tegra124/tegra124_coretemp.c
204
sc->cpu_id = device_get_unit(dev);
sys/arm/nvidia/tegra124/tegra124_coretemp.c
56
int cpu_id;
sys/arm64/arm64/machdep.c
353
cpu_est_clockrate(int cpu_id, uint64_t *rate)
sys/arm64/arm64/machdep.c
357
pc = pcpu_find(cpu_id);
sys/arm64/nvidia/tegra210/tegra210_coretemp.c
164
sc->tsens_id = 0x100 + sc->cpu_id;
sys/arm64/nvidia/tegra210/tegra210_coretemp.c
209
sc->cpu_id = device_get_unit(dev);
sys/arm64/nvidia/tegra210/tegra210_coretemp.c
58
int cpu_id;
sys/arm64/spe/arm_spe_backend.c
142
int cpu_id;
sys/arm64/spe/arm_spe_backend.c
148
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/arm64/spe/arm_spe_backend.c
149
info = &spe_info_cpu[cpu_id];
sys/arm64/spe/arm_spe_backend.c
151
info->ident = cpu_id;
sys/arm64/spe/arm_spe_backend.c
156
snprintf(lock_name, sizeof(lock_name), "%s%d", tmp, cpu_id);
sys/arm64/spe/arm_spe_backend.c
159
spe_info[cpu_id] = info;
sys/arm64/spe/arm_spe_backend.c
226
int cpu_id;
sys/arm64/spe/arm_spe_backend.c
229
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/arm64/spe/arm_spe_backend.c
230
info = &spe_info_cpu[cpu_id];
sys/arm64/spe/arm_spe_backend.c
231
printf("CPU %u:\n", cpu_id);
sys/arm64/spe/arm_spe_backend.c
297
spe_backend_configure(struct hwt_context *ctx, int cpu_id, int thread_id)
sys/arm64/spe/arm_spe_backend.c
305
info = &spe_info_cpu[cpu_id];
sys/arm64/spe/arm_spe_backend.c
319
info->ident = cpu_id;
sys/arm64/spe/arm_spe_backend.c
352
spe_info[cpu_id] = info;
sys/arm64/spe/arm_spe_backend.c
424
int cpu_id;
sys/arm64/spe/arm_spe_backend.c
429
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/arm64/spe/arm_spe_backend.c
430
vm = hwt_cpu_get(ctx, cpu_id)->vm;
sys/arm64/spe/arm_spe_backend.c
431
KASSERT(spe_info[cpu_id] == &spe_info_cpu[cpu_id], ("%s: spe_info mismatch for cpu_id=%u", __func__, cpu_id));
sys/arm64/spe/arm_spe_backend.c
432
info = &spe_info_cpu[cpu_id];
sys/arm64/spe/arm_spe_backend.c
441
cpu_id = CPU_FFS(&ctx->cpu_map) - 1;
sys/arm64/spe/arm_spe_backend.c
442
KASSERT(spe_info[cpu_id] == &spe_info_cpu[cpu_id], ("%s: spe_info mismatch for cpu_id=%u", __func__, cpu_id));
sys/arm64/spe/arm_spe_backend.c
443
info = spe_info[cpu_id];
sys/arm64/spe/arm_spe_backend.c
508
int cpu_id;
sys/arm64/spe/arm_spe_backend.c
516
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/arm64/spe/arm_spe_backend.c
517
info = spe_info[cpu_id];
sys/arm64/spe/arm_spe_backend.c
538
spe_backend_enable(struct hwt_context *ctx, int cpu_id)
sys/arm64/spe/arm_spe_backend.c
544
KASSERT(curcpu == cpu_id, ("%s: attempting to enable SPE on another cpu", __func__));
sys/arm64/spe/arm_spe_backend.c
546
info = spe_info[cpu_id];
sys/arm64/spe/arm_spe_backend.c
559
spe_backend_disable(struct hwt_context *ctx, int cpu_id)
sys/arm64/spe/arm_spe_backend.c
566
KASSERT(curcpu == cpu_id, ("%s: attempting to disable SPE on another cpu", __func__));
sys/arm64/spe/arm_spe_dev.c
129
int cpu_id = PCPU_GET(cpuid);
sys/arm64/spe/arm_spe_dev.c
134
struct arm_spe_info *info = sc->spe_info[cpu_id];
sys/arm64/spe/arm_spe_dev.c
181
__func__, cpu_id);
sys/arm64/spe/arm_spe_dev.c
196
device_printf(dev, "CPU:%d PMBSR_EL1:%#lx\n", cpu_id, pmbsr);
sys/arm64/spe/arm_spe_dev.c
253
cpu_id);
sys/compat/freebsd32/freebsd32_systrace_args.c
3307
iarg[a++] = p->cpu_id; /* int */
sys/compat/linprocfs/linprocfs.c
344
i, cpu_vendor, CPUID_TO_FAMILY(cpu_id),
sys/compat/linprocfs/linprocfs.c
345
CPUID_TO_MODEL(cpu_id), cpu_model, cpu_id & CPUID_STEPPING,
sys/compat/linprocfs/linprocfs.c
350
CPUID_TO_FAMILY(cpu_id), "yes");
sys/compat/linux/linux_rseq.c
74
uint32_t cpu_id;
sys/compat/linuxkpi/common/include/linux/interrupt.h
113
bind_irq_to_cpu(unsigned int irq, int cpu_id)
sys/compat/linuxkpi/common/include/linux/interrupt.h
115
return (lkpi_bind_irq_to_cpu(irq, cpu_id));
sys/compat/linuxkpi/common/src/linux_compat.c
2862
boot_cpu_data.x86 = CPUID_TO_FAMILY(cpu_id);
sys/compat/linuxkpi/common/src/linux_compat.c
2863
boot_cpu_data.x86_model = CPUID_TO_MODEL(cpu_id);
sys/compat/linuxkpi/common/src/linux_compat.c
2871
__cpu_data[i].x86 = CPUID_TO_FAMILY(cpu_id);
sys/compat/linuxkpi/common/src/linux_compat.c
2872
__cpu_data[i].x86_model = CPUID_TO_MODEL(cpu_id);
sys/compat/linuxkpi/common/src/linux_interrupt.c
201
lkpi_bind_irq_to_cpu(unsigned int irq, int cpu_id)
sys/compat/linuxkpi/common/src/linux_interrupt.c
214
return (-bus_bind_intr(dev->bsddev, irqe->res, cpu_id));
sys/crypto/openssl/ossl_x86.c
108
if ((cpu_id & 0x0fff0ff0) == 0x00050650)
sys/crypto/openssl/ossl_x86.c
75
if ((cpu_id & 0xf00) != 0xf00)
sys/dev/acpica/acpi_cpu.c
168
u_int *cpu_id);
sys/dev/acpica/acpi_cpu.c
235
int acpi_id, cpu_id;
sys/dev/acpica/acpi_cpu.c
289
if (acpi_pcpu_get_id(dev, acpi_id, &cpu_id) != 0) {
sys/dev/acpica/acpi_cpu.c
296
if (device_set_unit(dev, cpu_id) != 0)
sys/dev/acpica/acpi_cpu.c
320
int cpu_id, drv_count, i;
sys/dev/acpica/acpi_cpu.c
334
cpu_id = device_get_unit(dev);
sys/dev/acpica/acpi_cpu.c
335
cpu_softc[cpu_id] = sc;
sys/dev/acpica/acpi_cpu.c
336
pcpu_data = pcpu_find(cpu_id);
sys/dev/acpica/acpi_cpu.c
561
acpi_pcpu_get_id(device_t dev, uint32_t acpi_id, u_int *cpu_id)
sys/dev/acpica/acpi_cpu.c
569
*cpu_id = pc->pc_cpuid;
sys/dev/acpica/acpi_cpu.c
582
*cpu_id = 0;
sys/dev/amd_ecc_inject/ecc_inject.c
205
if ((cpu_vendor_id != CPU_VENDOR_AMD || CPUID_TO_FAMILY(cpu_id) < 0x10) &&
sys/dev/amdsmn/amdsmn.c
230
family = CPUID_TO_FAMILY(cpu_id);
sys/dev/amdtemp/amdtemp.c
316
family = CPUID_TO_FAMILY(cpu_id);
sys/dev/amdtemp/amdtemp.c
317
model = CPUID_TO_MODEL(cpu_id);
sys/dev/amdtemp/amdtemp.c
318
stepping = CPUID_TO_STEPPING(cpu_id);
sys/dev/amdtemp/amdtemp.c
371
cpuid = cpu_id;
sys/dev/amdtemp/amdtemp.c
767
minus49 = (CPUID_TO_FAMILY(cpu_id) >= 0x15 &&
sys/dev/bxe/ecore_hsi.h
12508
uint32_t cpu_id /* CPU id for sending completion for TSS (only 8 bits are used) */;
sys/dev/coretemp/coretemp.c
165
cpu_model = CPUID_TO_MODEL(cpu_id);
sys/dev/coretemp/coretemp.c
166
cpu_stepping = CPUID_TO_STEPPING(cpu_id);
sys/dev/cpuctl/cpuctl.c
223
if (cpu_id == 0)
sys/dev/cpuctl/cpuctl.c
530
if (cpu_id == 0)
sys/dev/hwpmc/hwpmc_amd.c
825
family = CPUID_TO_FAMILY(cpu_id);
sys/dev/hwpmc/hwpmc_amd.c
826
model = CPUID_TO_MODEL(cpu_id);
sys/dev/hwpmc/hwpmc_amd.c
827
stepping = CPUID_TO_STEPPING(cpu_id);
sys/dev/hwpmc/hwpmc_amd.c
836
switch (cpu_id & 0xF00) {
sys/dev/hwpmc/hwpmc_amd.c
841
printf("pmc: Unknown AMD CPU %x %d-%d.\n", cpu_id, family,
sys/dev/hwpmc/hwpmc_intel.c
85
PMCDBG1(MDP,INI,0, "intel-initialize cpuid=0x%x", cpu_id);
sys/dev/hwpmc/hwpmc_intel.c
91
family = CPUID_TO_FAMILY(cpu_id);
sys/dev/hwpmc/hwpmc_intel.c
92
model = CPUID_TO_MODEL(cpu_id);
sys/dev/hwpmc/hwpmc_intel.c
93
stepping = CPUID_TO_STEPPING(cpu_id);
sys/dev/hwpmc/hwpmc_intel.c
98
switch (cpu_id & 0xF00) {
sys/dev/hwt/hwt_backend.c
100
hwt_backend_enable(struct hwt_context *ctx, int cpu_id)
sys/dev/hwt/hwt_backend.c
105
ctx->hwt_backend->ops->hwt_backend_enable(ctx, cpu_id);
sys/dev/hwt/hwt_backend.c
109
hwt_backend_disable(struct hwt_context *ctx, int cpu_id)
sys/dev/hwt/hwt_backend.c
114
ctx->hwt_backend->ops->hwt_backend_disable(ctx, cpu_id);
sys/dev/hwt/hwt_backend.c
136
hwt_backend_dump(struct hwt_context *ctx, int cpu_id)
sys/dev/hwt/hwt_backend.c
141
ctx->hwt_backend->ops->hwt_backend_dump(cpu_id);
sys/dev/hwt/hwt_backend.c
87
hwt_backend_configure(struct hwt_context *ctx, int cpu_id, int thread_id)
sys/dev/hwt/hwt_backend.c
93
error = ctx->hwt_backend->ops->hwt_backend_configure(ctx, cpu_id,
sys/dev/hwt/hwt_backend.h
35
int (*hwt_backend_configure)(struct hwt_context *, int cpu_id,
sys/dev/hwt/hwt_backend.h
39
void (*hwt_backend_enable)(struct hwt_context *, int cpu_id);
sys/dev/hwt/hwt_backend.h
40
void (*hwt_backend_disable)(struct hwt_context *, int cpu_id);
sys/dev/hwt/hwt_backend.h
51
void (*hwt_backend_dump)(int cpu_id);
sys/dev/hwt/hwt_backend.h
63
int hwt_backend_configure(struct hwt_context *ctx, int cpu_id, int thread_id);
sys/dev/hwt/hwt_backend.h
64
void hwt_backend_enable(struct hwt_context *ctx, int cpu_id);
sys/dev/hwt/hwt_backend.h
65
void hwt_backend_disable(struct hwt_context *ctx, int cpu_id);
sys/dev/hwt/hwt_backend.h
68
void hwt_backend_dump(struct hwt_context *ctx, int cpu_id);
sys/dev/hwt/hwt_cpu.c
100
if (cpu->cpu_id == cpu_id) {
sys/dev/hwt/hwt_cpu.c
92
hwt_cpu_get(struct hwt_context *ctx, int cpu_id)
sys/dev/hwt/hwt_cpu.h
33
int cpu_id;
sys/dev/hwt/hwt_cpu.h
42
struct hwt_cpu * hwt_cpu_get(struct hwt_context *ctx, int cpu_id);
sys/dev/hwt/hwt_hook.c
101
int cpu_id;
sys/dev/hwt/hwt_hook.c
105
cpu_id = PCPU_GET(cpuid);
sys/dev/hwt/hwt_hook.c
122
thr->thread_id, td->td_tid, cpu_id);
sys/dev/hwt/hwt_hook.c
124
hwt_backend_disable(ctx, cpu_id);
sys/dev/hwt/hwt_hook.c
135
int cpu_id;
sys/dev/hwt/hwt_hook.c
139
cpu_id = PCPU_GET(cpuid);
sys/dev/hwt/hwt_hook.c
154
thr->thread_id, td->td_tid, cpu_id);
sys/dev/hwt/hwt_hook.c
157
hwt_backend_disable(ctx, cpu_id);
sys/dev/hwt/hwt_hook.c
65
int cpu_id;
sys/dev/hwt/hwt_hook.c
69
cpu_id = PCPU_GET(cpuid);
sys/dev/hwt/hwt_hook.c
87
thr->thread_id, td->td_tid, cpu_id);
sys/dev/hwt/hwt_hook.c
89
hwt_backend_configure(ctx, cpu_id, thr->thread_id);
sys/dev/hwt/hwt_hook.c
90
hwt_backend_enable(ctx, cpu_id);
sys/dev/hwt/hwt_ioctl.c
297
int cpu_id;
sys/dev/hwt/hwt_ioctl.c
306
CPU_FOREACH_ISSET(cpu_id, &cpu_map) {
sys/dev/hwt/hwt_ioctl.c
309
if (CPU_ISSET(cpu_id, &hlt_cpus_mask))
sys/dev/hwt/hwt_ioctl.c
343
CPU_FOREACH_ISSET(cpu_id, &cpu_map) {
sys/dev/hwt/hwt_ioctl.c
344
sprintf(path, "hwt_%d_%d", ctx->ident, cpu_id);
sys/dev/hwt/hwt_ioctl.c
354
cpu->cpu_id = cpu_id;
sys/dev/hwt/hwt_thread.h
45
int cpu_id; /* last cpu_id */
sys/dev/hwt/hwt_vm.c
211
int cpu_id;
sys/dev/hwt/hwt_vm.c
215
CPU_FOREACH_ISSET(cpu_id, &ctx->cpu_map) {
sys/dev/hwt/hwt_vm.c
218
if (CPU_ISSET(cpu_id, &hlt_cpus_mask))
sys/dev/hwt/hwt_vm.c
222
hwt_backend_configure(ctx, cpu_id, cpu_id);
sys/dev/hwt/hwt_vm.c
224
CPU_SET(cpu_id, &enable_cpus);
sys/dev/hwt/hwt_vm.c
228
CPU_FOREACH_ISSET(cpu_id, &enable_cpus)
sys/dev/hwt/hwt_vm.c
229
hwt_backend_enable(ctx, cpu_id);
sys/dev/liquidio/base/lio_droq.h
331
uint32_t cpu_id;
sys/dev/liquidio/lio_core.c
412
static int cpu_id, cpu_id_modulus;
sys/dev/liquidio/lio_core.c
424
cpu_id = 0;
sys/dev/liquidio/lio_core.c
447
droq->cpu_id = cpu_id;
sys/dev/liquidio/lio_core.c
448
cpu_id++;
sys/dev/liquidio/lio_core.c
449
if (cpu_id >= cpu_id_modulus)
sys/dev/liquidio/lio_core.c
450
cpu_id = 0;
sys/dev/liquidio/lio_core.c
552
int cpu_id, err, i;
sys/dev/liquidio/lio_core.c
612
cpu_id = rss_getcpu(i % rss_getnumbuckets());
sys/dev/liquidio/lio_core.c
614
cpu_id = i % mp_ncpus;
sys/dev/liquidio/lio_core.c
616
CPU_SETOF(cpu_id, &ioq_vector->affinity_mask);
sys/dev/liquidio/lio_core.c
619
err = bus_bind_intr(device, ioq_vector->msix_res, cpu_id);
sys/dev/liquidio/lio_core.c
623
lio_dev_dbg(oct, "Bound RSS bucket %d to CPU %d\n", i, cpu_id);
sys/dev/liquidio/lio_core.c
625
lio_dev_dbg(oct, "Bound Queue %d to CPU %d\n", i, cpu_id);
sys/dev/qat/qat/qat_ocf.c
561
u_int cpu_id = PCPU_GET(cpuid);
sys/dev/qat/qat/qat_ocf.c
568
->cyInstHandles[cpu_id % qat_softc->numCyInstances];
sys/dev/qcom_ess_edma/qcom_ess_edma.c
463
struct qcom_ess_edma_intr *intr, int rid, int cpu_id)
sys/dev/qcom_ess_edma/qcom_ess_edma.c
490
if (cpu_id != -1) {
sys/dev/qcom_ess_edma/qcom_ess_edma.c
491
if (intr_bind_irq(sc->sc_dev, intr->irq_res, cpu_id) != 0) {
sys/dev/qcom_ess_edma/qcom_ess_edma.c
494
rid, cpu_id);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
743
int cpu_id;
sys/dev/qcom_ess_edma/qcom_ess_edma.c
751
cpu_id = qcom_ess_edma_tx_queue_to_cpu(sc, i);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
753
i, cpu_id) != 0)
sys/dev/qcom_ess_edma/qcom_ess_edma.c
758
i, cpu_id);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
763
int cpu_id = qcom_ess_edma_rx_queue_to_cpu(sc, i);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
765
i + QCOM_ESS_EDMA_NUM_TX_IRQS, cpu_id) != 0)
sys/dev/qcom_ess_edma/qcom_ess_edma.c
770
i, cpu_id);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
828
int cpu_id;
sys/dev/qcom_ess_edma/qcom_ess_edma.c
843
cpu_id = qcom_ess_edma_tx_queue_to_cpu(sc, i);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
845
if (qcom_ess_edma_setup_tx_state(sc, i, cpu_id) != 0)
sys/dev/qcom_ess_edma/qcom_ess_edma.c
852
int cpu_id;
sys/dev/qcom_ess_edma/qcom_ess_edma.c
867
cpu_id = qcom_ess_edma_rx_queue_to_cpu(sc, i);
sys/dev/qcom_ess_edma/qcom_ess_edma.c
869
if (qcom_ess_edma_setup_rx_state(sc, i, cpu_id) != 0)
sys/i386/i386/initcpu.c
412
if (CPUID_TO_MODEL(cpu_id) >= 8)
sys/i386/i386/initcpu.c
574
if (CPUID_TO_MODEL(cpu_id) <= 9)
sys/i386/i386/initcpu.c
674
if (((cpu_id & 0x0f0) > 0) &&
sys/i386/i386/initcpu.c
675
((cpu_id & 0x0f0) < 0x60) &&
sys/i386/i386/initcpu.c
676
((cpu_id & 0x00f) > 3))
sys/i386/i386/initcpu.c
678
else if (((cpu_id & 0x0f0) > 0x80) ||
sys/i386/i386/initcpu.c
679
(((cpu_id & 0x0f0) == 0x80) &&
sys/i386/i386/initcpu.c
680
(cpu_id & 0x00f) > 0x07))
sys/i386/i386/initcpu.c
682
else if ((cpu_id & 0x0f0) > 0x50)
sys/i386/i386/initcpu.c
685
if ((cpu_id & 0xf0) == 0xa0)
sys/i386/i386/initcpu.c
712
switch (cpu_id & 0xff0) {
sys/i386/i386/initcpu.c
731
((cpu_id & ~0xf) == 0x660 ||
sys/i386/i386/initcpu.c
732
(cpu_id & ~0xf) == 0x670 ||
sys/i386/i386/initcpu.c
733
(cpu_id & ~0xf) == 0x680)) {
sys/i386/i386/initcpu.c
744
if ((CPUID_TO_FAMILY(cpu_id) == 0xf ||
sys/i386/i386/initcpu.c
745
CPUID_TO_FAMILY(cpu_id) == 0x10) &&
sys/i386/i386/initcpu.c
821
if (((cpu_id & 0xf0) > 0) && ((cpu_id & 0x0f) > 3)) {
sys/i386/i386/k6_mem.c
181
if ((cpu_id & 0xf00) != 0x500)
sys/i386/i386/k6_mem.c
183
if ((cpu_id & 0xf0) < 0x80 ||
sys/i386/i386/k6_mem.c
184
((cpu_id & 0xf0) == 0x80 && (cpu_id & 0xf) <= 0x7))
sys/i386/i386/machdep.c
699
cpu_vendor_id == CPU_VENDOR_AMD) && CPUID_TO_FAMILY(cpu_id) >= 6) ||
sys/i386/i386/pmap.c
817
!(CPUID_TO_FAMILY(cpu_id) == 6 && CPUID_TO_MODEL(cpu_id) >= 0xe))
sys/kern/kern_membarrier.c
117
kern_membarrier(struct thread *td, int cmd, unsigned flags, int cpu_id)
sys/kern/kern_membarrier.c
238
return (kern_membarrier(td, uap->cmd, uap->flags, uap->cpu_id));
sys/kern/systrace_args.c
3394
iarg[a++] = p->cpu_id; /* int */
sys/powerpc/powerpc/cpu.c
377
cpu_est_clockrate(int cpu_id, uint64_t *cps)
sys/powerpc/ps3/ps3-hvcall.h
56
int lv1_configure_irq_state_bitmap(uint64_t ppe_id, uint64_t cpu_id, uint64_t bitmap_addr);
sys/powerpc/ps3/ps3-hvcall.h
57
int lv1_connect_irq_plug_ext(uint64_t ppe_id, uint64_t cpu_id, uint64_t virq, uint64_t outlet, uint64_t zero);
sys/powerpc/ps3/ps3-hvcall.h
60
int lv1_disconnect_irq_plug_ext(uint64_t ppe_id, uint64_t cpu_id, uint64_t virq);
sys/powerpc/ps3/ps3-hvcall.h
67
int lv1_end_of_interrupt_ext(uint64_t ppe_id, uint64_t cpu_id, uint64_t virq);
sys/powerpc/ps3/ps3-hvcall.h
68
int lv1_did_update_interrupt_mask(uint64_t ppe_id, uint64_t cpu_id);
sys/riscv/riscv/machdep.c
218
cpu_est_clockrate(int cpu_id, uint64_t *rate)
sys/riscv/riscv/machdep.c
222
pc = pcpu_find(cpu_id);
sys/riscv/vmm/riscv.h
95
int cpu_id;
sys/riscv/vmm/vmm_aplic.c
208
aplic_handle_idc_claimi(struct hyp *hyp, struct aplic *aplic, int cpu_id,
sys/riscv/vmm/vmm_aplic.c
224
if (irq->target_hart != cpu_id)
sys/riscv/vmm/vmm_aplic.c
468
if (irq->target_hart != hypctx->cpu_id)
sys/riscv/vmm/vmm_riscv.c
231
hypctx->cpu_id = vcpuid;
sys/riscv/vmm/vmm_vtimer.c
95
vcpu_notify_event(vm_vcpu(hyp->vm, hypctx->cpu_id));
sys/sys/cpu.h
200
int cpu_est_clockrate(int cpu_id, uint64_t *rate);
sys/sys/syscallsubr.h
236
int cpu_id);
sys/sys/sysproto.h
1845
char cpu_id_l_[PADL_(int)]; int cpu_id; char cpu_id_r_[PADR_(int)];
sys/x86/acpica/madt.c
170
CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/acpica/madt.c
171
CPUID_TO_MODEL(cpu_id) == 0x2a) {
sys/x86/cpufreq/hwpstate_amd.c
1155
family = CPUID_TO_FAMILY(cpu_id);
sys/x86/cpufreq/hwpstate_amd.c
781
if ((cpu_vendor_id != CPU_VENDOR_AMD || CPUID_TO_FAMILY(cpu_id) < 0x10) &&
sys/x86/cpufreq/p4tcc.c
176
switch (cpu_id & 0xff) {
sys/x86/cpufreq/powernow.c
866
switch (cpu_id & 0xf00) {
sys/x86/cpufreq/powernow.c
899
switch (cpu_id & 0xf00) {
sys/x86/cpufreq/smist.c
289
switch (cpu_id & 0xff0) {
sys/x86/include/x86_var.h
56
extern u_int cpu_id;
sys/x86/pci/qpi.c
69
CPUID_TO_FAMILY(cpu_id) != 0x6)
sys/x86/x86/cpu_machdep.c
1324
} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/cpu_machdep.c
1325
(CPUID_TO_MODEL(cpu_id) == 0x2e || CPUID_TO_MODEL(cpu_id) == 0x1e ||
sys/x86/x86/cpu_machdep.c
1326
CPUID_TO_MODEL(cpu_id) == 0x1f || CPUID_TO_MODEL(cpu_id) == 0x1a ||
sys/x86/x86/cpu_machdep.c
1327
CPUID_TO_MODEL(cpu_id) == 0x2f || CPUID_TO_MODEL(cpu_id) == 0x25 ||
sys/x86/x86/cpu_machdep.c
1328
CPUID_TO_MODEL(cpu_id) == 0x2c || CPUID_TO_MODEL(cpu_id) == 0x2d ||
sys/x86/x86/cpu_machdep.c
1329
CPUID_TO_MODEL(cpu_id) == 0x2a || CPUID_TO_MODEL(cpu_id) == 0x3e ||
sys/x86/x86/cpu_machdep.c
1330
CPUID_TO_MODEL(cpu_id) == 0x3a) &&
sys/x86/x86/cpu_machdep.c
1344
} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/cpu_machdep.c
1345
(CPUID_TO_MODEL(cpu_id) == 0x3f || CPUID_TO_MODEL(cpu_id) == 0x3c ||
sys/x86/x86/cpu_machdep.c
1346
CPUID_TO_MODEL(cpu_id) == 0x45 || CPUID_TO_MODEL(cpu_id) == 0x46 ||
sys/x86/x86/cpu_machdep.c
1347
CPUID_TO_MODEL(cpu_id) == 0x56 || CPUID_TO_MODEL(cpu_id) == 0x4f ||
sys/x86/x86/cpu_machdep.c
1348
CPUID_TO_MODEL(cpu_id) == 0x47 || CPUID_TO_MODEL(cpu_id) == 0x3d) &&
sys/x86/x86/cpu_machdep.c
1362
} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/cpu_machdep.c
1363
((CPUID_TO_MODEL(cpu_id) == 0x55 && (cpu_id &
sys/x86/x86/cpu_machdep.c
1365
CPUID_TO_MODEL(cpu_id) == 0x4e || CPUID_TO_MODEL(cpu_id) == 0x5e ||
sys/x86/x86/cpu_machdep.c
1366
(CPUID_TO_MODEL(cpu_id) == 0x8e && (cpu_id &
sys/x86/x86/cpu_machdep.c
1368
(CPUID_TO_MODEL(cpu_id) == 0x9e && (cpu_id &
sys/x86/x86/cpu_machdep.c
1397
} else if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/cpu_machdep.c
1398
((CPUID_TO_MODEL(cpu_id) == 0x37 ||
sys/x86/x86/cpu_machdep.c
1399
CPUID_TO_MODEL(cpu_id) == 0x4a ||
sys/x86/x86/cpu_machdep.c
1400
CPUID_TO_MODEL(cpu_id) == 0x4c ||
sys/x86/x86/cpu_machdep.c
1401
CPUID_TO_MODEL(cpu_id) == 0x4d ||
sys/x86/x86/cpu_machdep.c
1402
CPUID_TO_MODEL(cpu_id) == 0x5a ||
sys/x86/x86/cpu_machdep.c
1403
CPUID_TO_MODEL(cpu_id) == 0x5d ||
sys/x86/x86/cpu_machdep.c
1404
CPUID_TO_MODEL(cpu_id) == 0x6e ||
sys/x86/x86/cpu_machdep.c
1405
CPUID_TO_MODEL(cpu_id) == 0x65 ||
sys/x86/x86/cpu_machdep.c
1406
CPUID_TO_MODEL(cpu_id) == 0x75 ||
sys/x86/x86/cpu_machdep.c
1407
CPUID_TO_MODEL(cpu_id) == 0x1c ||
sys/x86/x86/cpu_machdep.c
1408
CPUID_TO_MODEL(cpu_id) == 0x26 ||
sys/x86/x86/cpu_machdep.c
1409
CPUID_TO_MODEL(cpu_id) == 0x27 ||
sys/x86/x86/cpu_machdep.c
1410
CPUID_TO_MODEL(cpu_id) == 0x35 ||
sys/x86/x86/cpu_machdep.c
1411
CPUID_TO_MODEL(cpu_id) == 0x36 ||
sys/x86/x86/cpu_machdep.c
1412
CPUID_TO_MODEL(cpu_id) == 0x7a))) {
sys/x86/x86/cpu_machdep.c
1743
CPUID_TO_FAMILY(cpu_id) == 0x17 &&
sys/x86/x86/cpu_machdep.c
1744
CPUID_TO_MODEL(cpu_id) >= 0x30 &&
sys/x86/x86/cpu_machdep.c
423
cpu_est_clockrate(int cpu_id, uint64_t *rate)
sys/x86/x86/cpu_machdep.c
430
if (pcpu_find(cpu_id) == NULL || rate == NULL)
sys/x86/x86/cpu_machdep.c
448
sched_bind(curthread, cpu_id);
sys/x86/x86/cpu_machdep.c
939
CPUID_TO_FAMILY(cpu_id) == 0x17 && CPUID_TO_MODEL(cpu_id) == 0x1) {
sys/x86/x86/cpu_machdep.c
947
CPUID_TO_FAMILY(cpu_id) == 0x6 && CPUID_TO_MODEL(cpu_id) == 0x5c) {
sys/x86/x86/identcpu.c
107
u_int cpu_id; /* Stepping ID */
sys/x86/x86/identcpu.c
1174
if (*cpu_vendor || cpu_id)
sys/x86/x86/identcpu.c
1529
((CPUID_TO_FAMILY(cpu_id) == 0xf &&
sys/x86/x86/identcpu.c
1530
CPUID_TO_MODEL(cpu_id) >= 0x3) ||
sys/x86/x86/identcpu.c
1531
(CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/identcpu.c
1532
CPUID_TO_MODEL(cpu_id) >= 0xe))) {
sys/x86/x86/identcpu.c
1549
CPUID_TO_FAMILY(cpu_id) == 0x15) {
sys/x86/x86/identcpu.c
1573
cpu_id = regs[0];
sys/x86/x86/identcpu.c
1724
switch (cpu_id & 0xf00) {
sys/x86/x86/identcpu.c
1777
cpu_id = regs[0]; /* eax */
sys/x86/x86/identcpu.c
1914
if (((cpu_id & 0xf00) == 0x500)
sys/x86/x86/identcpu.c
1915
&& (((cpu_id & 0x0f0) > 0x80)
sys/x86/x86/identcpu.c
1916
|| (((cpu_id & 0x0f0) == 0x80)
sys/x86/x86/identcpu.c
1917
&& (cpu_id & 0x00f) > 0x07))) {
sys/x86/x86/identcpu.c
1928
} else if (((cpu_id & 0xf00) == 0x500)
sys/x86/x86/identcpu.c
1929
&& ((cpu_id & 0x0f0) > 0x50)) {
sys/x86/x86/identcpu.c
1954
if (CPUID_TO_FAMILY(cpu_id) == 0xf && CPUID_TO_MODEL(cpu_id) >= 0x20 &&
sys/x86/x86/identcpu.c
1955
CPUID_TO_MODEL(cpu_id) <= 0x3f)
sys/x86/x86/identcpu.c
2103
if (CPUID_TO_FAMILY(cpu_id) == 0xf &&
sys/x86/x86/identcpu.c
2104
CPUID_TO_MODEL(cpu_id) == 0x6)
sys/x86/x86/identcpu.c
2722
u_int family = CPUID_TO_FAMILY(cpu_id);
sys/x86/x86/identcpu.c
2723
u_int model = CPUID_TO_MODEL(cpu_id);
sys/x86/x86/identcpu.c
277
if ((cpu_id & 0xf00) > 0x300) {
sys/x86/x86/identcpu.c
282
switch (cpu_id & 0x3000) {
sys/x86/x86/identcpu.c
291
switch (cpu_id & 0xf00) {
sys/x86/x86/identcpu.c
295
switch (cpu_id & 0xf0) {
sys/x86/x86/identcpu.c
324
switch (cpu_id & 0xf0) {
sys/x86/x86/identcpu.c
361
switch (cpu_id & 0xf0) {
sys/x86/x86/identcpu.c
422
switch (cpu_id & 0xFF0) {
sys/x86/x86/identcpu.c
476
if ((cpu_id & 0xf00) == 0xf00)
sys/x86/x86/identcpu.c
485
switch (cpu_id & 0xff0) {
sys/x86/x86/identcpu.c
622
switch (cpu_id & 0xff0) {
sys/x86/x86/identcpu.c
634
switch (cpu_id & 0xff0) {
sys/x86/x86/identcpu.c
648
if (cpu_id & 0x8)
sys/x86/x86/identcpu.c
671
if ((cpu_id & 0xff0) == 0x6f0)
sys/x86/x86/identcpu.c
682
switch (cpu_id & 0xff0) {
sys/x86/x86/identcpu.c
698
if ((cpu_id & 0xf00) == 0xf00)
sys/x86/x86/identcpu.c
759
if (cpu_id)
sys/x86/x86/identcpu.c
760
printf(" Id=0x%x", cpu_id);
sys/x86/x86/identcpu.c
770
(cpu_vendor_id == CPU_VENDOR_CYRIX && ((cpu_id & 0xf00) > 0x500)) ||
sys/x86/x86/identcpu.c
773
printf(" Family=0x%x", CPUID_TO_FAMILY(cpu_id));
sys/x86/x86/identcpu.c
774
printf(" Model=0x%x", CPUID_TO_MODEL(cpu_id));
sys/x86/x86/identcpu.c
775
printf(" Stepping=%u", cpu_id & CPUID_STEPPING);
sys/x86/x86/mca.c
1332
CPUID_TO_FAMILY(cpu_id) == 0x10 && amd10h_L1TP)
sys/x86/x86/mca.c
1501
if (CPUID_TO_FAMILY(cpu_id) < 0x15 && i >= 5)
sys/x86/x86/mca.c
1578
family = CPUID_TO_FAMILY(cpu_id);
sys/x86/x86/mca.c
1640
&& CPUID_TO_MODEL(cpu_id) < 0x1a)
sys/x86/x86/mca.c
266
if (CPUID_TO_FAMILY(cpu_id) >= 0x10 && CPUID_TO_FAMILY(cpu_id) <= 0x16)
sys/x86/x86/mca.c
268
if (CPUID_TO_FAMILY(cpu_id) >= 0x17)
sys/x86/x86/mca.c
471
CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/mca.c
472
(CPUID_TO_MODEL(cpu_id) == 0x3c || /* HSD131, HSM142, HSW131 */
sys/x86/x86/mca.c
473
CPUID_TO_MODEL(cpu_id) == 0x3d || /* BDM48 */
sys/x86/x86/mca.c
474
CPUID_TO_MODEL(cpu_id) == 0x45 ||
sys/x86/x86/mca.c
475
CPUID_TO_MODEL(cpu_id) == 0x46) && /* HSM142 */
sys/x86/x86/mca.c
856
rec->mr_cpu_id = cpu_id;
sys/x86/x86/mp_x86.c
299
CPUID_TO_FAMILY(cpu_id) >= 0x16) {
sys/x86/x86/msi.c
361
if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/msi.c
362
CPUID_TO_MODEL(cpu_id) >= 0xf)
sys/x86/x86/tsc.c
292
switch (cpu_id & 0xFF0) {
sys/x86/x86/tsc.c
300
switch (cpu_id & 0xff0) {
sys/x86/x86/tsc.c
313
switch (cpu_id & 0xff0) {
sys/x86/x86/tsc.c
315
if ((cpu_id & CPUID_STEPPING) == 0) {
sys/x86/x86/tsc.c
330
CPUID_TO_FAMILY(cpu_id) >= 0x10))
sys/x86/x86/tsc.c
340
((CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/tsc.c
341
CPUID_TO_MODEL(cpu_id) >= 0xe) ||
sys/x86/x86/tsc.c
342
(CPUID_TO_FAMILY(cpu_id) == 0xf &&
sys/x86/x86/tsc.c
343
CPUID_TO_MODEL(cpu_id) >= 0x3))))
sys/x86/x86/tsc.c
352
CPUID_TO_FAMILY(cpu_id) == 0x6 &&
sys/x86/x86/tsc.c
353
CPUID_TO_MODEL(cpu_id) >= 0xf &&
sys/x86/x86/tsc.c
627
if (CPUID_TO_FAMILY(cpu_id) >= 0x17)
sys/x86/x86/tsc.c
635
if (CPUID_TO_FAMILY(cpu_id) < 0x15 ||
tools/tools/ncpus/acpi.c
75
apic->cpu_id,
tools/tools/ncpus/acpi.c
84
sapic->cpu_id,
tools/tools/ncpus/acpidump.h
112
u_char cpu_id;
tools/tools/ncpus/acpidump.h
125
u_char cpu_id;
tools/tools/ncpus/acpidump.h
60
u_char cpu_id;
tools/tools/ncpus/acpidump.h
94
u_char cpu_id;
usr.sbin/acpi/acpidump/acpi.c
336
acpi_print_cpu(u_char cpu_id)
usr.sbin/acpi/acpidump/acpi.c
340
if (cpu_id == 0xff)
usr.sbin/acpi/acpidump/acpi.c
343
printf("%d\n", (u_int)cpu_id);
usr.sbin/acpi/acpidump/acpi.c
60
static void acpi_print_cpu(u_char cpu_id);