Symbol: threads_per_core
arch/loongarch/kernel/smp.c
368
int threads_per_core = 0;
arch/loongarch/kernel/smp.c
374
threads_per_core = 1;
arch/loongarch/kernel/smp.c
380
threads_per_core++;
arch/loongarch/kernel/smp.c
390
cpu_smt_set_num_threads(threads_per_core, threads_per_core);
arch/powerpc/include/asm/cputhreads.h
23
extern int threads_per_core;
arch/powerpc/include/asm/cputhreads.h
50
return cpu & (threads_per_core - 1);
arch/powerpc/include/asm/cputhreads.h
60
return cpu & ~(threads_per_core - 1);
arch/powerpc/include/asm/cputhreads.h
65
return cpu | (threads_per_core - 1);
arch/powerpc/include/asm/cputhreads.h
76
if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8))
arch/powerpc/include/asm/cputhreads.h
84
if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8))
arch/powerpc/include/asm/cputhreads.h
92
if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8))
arch/powerpc/include/asm/paravirt.h
196
for (i = first_cpu; i < first_cpu + threads_per_core; i++) {
arch/powerpc/include/asm/topology.h
172
for (i = first_cpu; i < first_cpu + threads_per_core; ++i) {
arch/powerpc/kernel/prom.c
1018
idx = cpu / threads_per_core;
arch/powerpc/kernel/setup-common.c
384
int threads_per_core, threads_per_subcore, threads_shift __read_mostly;
arch/powerpc/kernel/setup-common.c
386
EXPORT_SYMBOL_GPL(threads_per_core);
arch/powerpc/kernel/setup-common.c
395
threads_per_core = tpc;
arch/powerpc/kernel/setup_64.c
134
smt_enabled_at_boot = threads_per_core;
arch/powerpc/kernel/setup_64.c
139
smt_enabled_at_boot = threads_per_core;
arch/powerpc/kernel/setup_64.c
146
min(threads_per_core, smt);
arch/powerpc/kernel/setup_64.c
156
smt_enabled_at_boot = threads_per_core;
arch/powerpc/kernel/smp.c
1167
int idx = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/kernel/smp.c
1187
cpu_smt_set_num_threads(num_threads, threads_per_core);
arch/powerpc/kernel/smp.c
1303
if (threads_per_core > 1 && secondaries_inhibited() &&
arch/powerpc/kernel/smp.c
1597
for (i = first_thread; i < first_thread + threads_per_core; i++)
arch/powerpc/kernel/smp.c
1770
return -cpu / threads_per_core;
arch/powerpc/kernel/smp.c
931
for (i = first_thread; i < first_thread + threads_per_core; i++) {
arch/powerpc/kexec/core_64.c
216
cpu_smt_num_threads = threads_per_core;
arch/powerpc/kexec/file_load_64.c
658
possible_cpu_nodes = num_possible_cpus() / threads_per_core;
arch/powerpc/kvm/book3s_hv.c
5654
for (cpu = 0; cpu < nr_cpu_ids; cpu += threads_per_core) {
arch/powerpc/kvm/book3s_hv.c
6565
int first_cpu = i * threads_per_core;
arch/powerpc/kvm/book3s_hv.c
6579
for (j = 0; j < threads_per_core; j++) {
arch/powerpc/kvm/e500.h
297
if (threads_per_core == 2)
arch/powerpc/kvm/e500mc.c
358
if (threads_per_core == 2)
arch/powerpc/kvm/e500mc.c
369
if (threads_per_core == 2)
arch/powerpc/kvm/e500mc.c
410
kvmppc_init_lpid(KVMPPC_NR_LPIDS/threads_per_core);
arch/powerpc/mm/numa.c
773
for (i = 0; i < threads_per_core; i++) {
arch/powerpc/perf/hv-24x7.c
1712
if (threads_per_core == 8)
arch/powerpc/perf/imc-pmu.c
1135
core_id = smp_processor_id() / threads_per_core;
arch/powerpc/perf/imc-pmu.c
1170
core_id = smp_processor_id() / threads_per_core;
arch/powerpc/perf/imc-pmu.c
1209
int core_id = (cpu_id / threads_per_core);
arch/powerpc/perf/imc-pmu.c
1360
int core_id = smp_processor_id() / threads_per_core;
arch/powerpc/perf/imc-pmu.c
1410
int core_id = smp_processor_id() / threads_per_core;
arch/powerpc/perf/imc-pmu.c
1570
int i, nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/perf/imc-pmu.c
1714
nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/perf/imc-pmu.c
1752
nr_cores = DIV_ROUND_UP(num_possible_cpus(), threads_per_core);
arch/powerpc/perf/imc-pmu.c
601
int nid, rc = 0, core_id = (cpu / threads_per_core);
arch/powerpc/perf/imc-pmu.c
638
int core_id = (cpu / threads_per_core);
arch/powerpc/perf/imc-pmu.c
716
core_id = cpu / threads_per_core;
arch/powerpc/perf/imc-pmu.c
774
core_id = event->cpu / threads_per_core;
arch/powerpc/perf/imc-pmu.c
842
core_id = event->cpu / threads_per_core;
arch/powerpc/platforms/85xx/smp.c
288
if (threads_per_core == 2) {
arch/powerpc/platforms/85xx/smp.c
319
} else if (threads_per_core == 1) {
arch/powerpc/platforms/85xx/smp.c
326
} else if (threads_per_core > 2) {
arch/powerpc/platforms/powernv/idle.c
1439
p->idle_state = (1 << threads_per_core) - 1;
arch/powerpc/platforms/powernv/idle.c
320
unsigned long core_thread_mask = (1UL << threads_per_core) - 1;
arch/powerpc/platforms/powernv/idle.c
371
== threads_per_core)
arch/powerpc/platforms/powernv/idle.c
613
unsigned long core_thread_mask = (1UL << threads_per_core) - 1;
arch/powerpc/platforms/powernv/idle.c
822
int need_awake = threads_per_core;
arch/powerpc/platforms/powernv/idle.c
825
cpu0 = cpu & ~(threads_per_core - 1);
arch/powerpc/platforms/powernv/idle.c
826
for (thr = 0; thr < threads_per_core; ++thr) {
arch/powerpc/platforms/powernv/idle.c
832
for (thr = 0; thr < threads_per_core; ++thr) {
arch/powerpc/platforms/powernv/idle.c
842
for (thr = 0; thr < threads_per_core; ++thr) {
arch/powerpc/platforms/powernv/idle.c
851
for (thr = 0; thr < threads_per_core; ++thr) {
arch/powerpc/platforms/powernv/idle.c
868
cpu0 = cpu & ~(threads_per_core - 1);
arch/powerpc/platforms/powernv/idle.c
871
for (thr = 0; thr < threads_per_core; ++thr) {
arch/powerpc/platforms/powernv/idle.c
898
unsigned long core_thread_mask = (1UL << threads_per_core) - 1;
arch/powerpc/platforms/powernv/subcore.c
154
for (i = cpu + 1; i < cpu + threads_per_core; i++)
arch/powerpc/platforms/powernv/subcore.c
209
for (i = cpu + 1; i < cpu + threads_per_core; i++)
arch/powerpc/platforms/powernv/subcore.c
328
threads_per_subcore = threads_per_core / subcores_per_core;
arch/powerpc/platforms/powernv/subcore.c
436
if (setup_max_cpus % threads_per_core)
arch/powerpc/platforms/pseries/lpar.c
210
vcpu_associativity = kcalloc(num_possible_cpus() / threads_per_core,
arch/powerpc/platforms/pseries/lpar.c
212
pcpu_associativity = kcalloc(NR_CPUS_H / threads_per_core,
arch/powerpc/platforms/pseries/lpar.c
235
assoc = &cpu_assoc[(int)(cpu / threads_per_core) * VPHN_ASSOC_BUFSIZE];
arch/powerpc/sysdev/fsl_rcpm.c
148
if (threads_per_core == 2) {
arch/powerpc/sysdev/fsl_rcpm.c
160
if (threads_per_core == 1)
drivers/cpufreq/powernv-cpufreq.c
836
for (i = 0; i < threads_per_core; i++)
drivers/virt/nitro_enclaves/ne_misc_dev.c
1102
for_each_cpu(cpu, ne_enclave->threads_per_core[i])
drivers/virt/nitro_enclaves/ne_misc_dev.c
1422
for_each_cpu(cpu, ne_enclave->threads_per_core[i])
drivers/virt/nitro_enclaves/ne_misc_dev.c
1426
free_cpumask_var(ne_enclave->threads_per_core[i]);
drivers/virt/nitro_enclaves/ne_misc_dev.c
1431
kfree(ne_enclave->threads_per_core);
drivers/virt/nitro_enclaves/ne_misc_dev.c
1629
ne_enclave->threads_per_core = kzalloc_objs(*ne_enclave->threads_per_core,
drivers/virt/nitro_enclaves/ne_misc_dev.c
1631
if (!ne_enclave->threads_per_core) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
1638
if (!zalloc_cpumask_var(&ne_enclave->threads_per_core[i], GFP_KERNEL)) {
drivers/virt/nitro_enclaves/ne_misc_dev.c
1715
free_cpumask_var(ne_enclave->threads_per_core[i]);
drivers/virt/nitro_enclaves/ne_misc_dev.c
1716
kfree(ne_enclave->threads_per_core);
drivers/virt/nitro_enclaves/ne_misc_dev.c
553
cpumask_set_cpu(cpu, ne_enclave->threads_per_core[core_id]);
drivers/virt/nitro_enclaves/ne_misc_dev.c
585
for_each_cpu(cpu, ne_enclave->threads_per_core[i])
drivers/virt/nitro_enclaves/ne_misc_dev.c
604
*vcpu_id = cpumask_any(ne_enclave->threads_per_core[core_id]);
drivers/virt/nitro_enclaves/ne_misc_dev.c
669
if (cpumask_test_cpu(vcpu_id, ne_enclave->threads_per_core[i]))
drivers/virt/nitro_enclaves/ne_misc_dev.h
80
cpumask_var_t *threads_per_core;
tools/power/x86/turbostat/turbostat.c
2413
int threads_per_core;
tools/power/x86/turbostat/turbostat.c
9664
topo.threads_per_core = max_siblings;
tools/power/x86/turbostat/turbostat.c
9705
int num_threads = topo.threads_per_core * num_cores;