Symbol: pcpu
arch/alpha/kernel/err_common.c
242
cdl_process_console_data_log(int cpu, struct percpu_struct *pcpu)
arch/alpha/kernel/err_common.c
245
(IDENT_ADDR | pcpu->console_data_log_pa);
arch/alpha/kernel/err_common.c
256
pcpu->console_data_log_pa = 0;
arch/alpha/kernel/err_common.c
266
struct percpu_struct *pcpu;
arch/alpha/kernel/err_common.c
270
pcpu = (struct percpu_struct *)
arch/alpha/kernel/err_common.c
273
if (pcpu->console_data_log_pa)
arch/alpha/kernel/err_common.c
274
cdl_process_console_data_log(cpu, pcpu);
arch/arm/common/mcpm_platsmp.c
20
unsigned int *pcpu, unsigned int *pcluster)
arch/arm/common/mcpm_platsmp.c
25
*pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0);
arch/arm/common/mcpm_platsmp.c
31
unsigned int pcpu, pcluster, ret;
arch/arm/common/mcpm_platsmp.c
34
cpu_to_pcpu(cpu, &pcpu, &pcluster);
arch/arm/common/mcpm_platsmp.c
37
__func__, cpu, pcpu, pcluster);
arch/arm/common/mcpm_platsmp.c
39
mcpm_set_entry_vector(pcpu, pcluster, NULL);
arch/arm/common/mcpm_platsmp.c
40
ret = mcpm_cpu_power_up(pcpu, pcluster);
arch/arm/common/mcpm_platsmp.c
43
mcpm_set_entry_vector(pcpu, pcluster, secondary_startup);
arch/arm/common/mcpm_platsmp.c
58
unsigned int pcpu, pcluster;
arch/arm/common/mcpm_platsmp.c
60
cpu_to_pcpu(cpu, &pcpu, &pcluster);
arch/arm/common/mcpm_platsmp.c
62
return !mcpm_wait_for_cpu_powerdown(pcpu, pcluster);
arch/arm/common/mcpm_platsmp.c
73
unsigned int mpidr, pcpu, pcluster;
arch/arm/common/mcpm_platsmp.c
75
pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0);
arch/arm/common/mcpm_platsmp.c
77
mcpm_set_entry_vector(pcpu, pcluster, NULL);
arch/arm64/kernel/acpi.c
443
int *pcpu)
arch/arm64/kernel/acpi.c
446
if (*pcpu < 0) {
arch/arm64/kernel/acpi.c
448
return *pcpu;
arch/loongarch/kernel/acpi.c
353
int acpi_map_cpu(acpi_handle handle, phys_cpuid_t physid, u32 acpi_id, int *pcpu)
arch/loongarch/kernel/acpi.c
367
*pcpu = cpu;
arch/powerpc/include/asm/kvm_book3s.h
75
u16 pcpu;
arch/powerpc/include/asm/kvm_ppc.h
573
extern void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu);
arch/powerpc/kernel/module_64.c
269
unsigned int symindex, unsigned int pcpu)
arch/powerpc/kernel/module_64.c
284
if (sym->st_shndx == pcpu &&
arch/powerpc/kvm/book3s_hv.c
1504
int thr, cpu, pcpu, nthreads;
arch/powerpc/kvm/book3s_hv.c
1520
pcpu = READ_ONCE(v->cpu);
arch/powerpc/kvm/book3s_hv.c
1521
if (pcpu >= 0)
arch/powerpc/kvm/book3s_hv.c
1522
smp_call_function_single(pcpu, do_nothing, NULL, 1);
arch/powerpc/kvm/book3s_hv.c
3293
static void kvmppc_prepare_radix_vcpu(struct kvm_vcpu *vcpu, int pcpu)
arch/powerpc/kvm/book3s_hv.c
3318
if (prev_cpu != pcpu) {
arch/powerpc/kvm/book3s_hv.c
3321
cpu_first_tlb_thread_sibling(pcpu))
arch/powerpc/kvm/book3s_hv.c
3328
nested->prev_cpu[vcpu->arch.nested_vcpu_id] = pcpu;
arch/powerpc/kvm/book3s_hv.c
3330
vcpu->arch.prev_cpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
3339
cpu = vc->pcpu;
arch/powerpc/kvm/book3s_hv.c
3346
vcpu->cpu = vc->pcpu;
arch/powerpc/kvm/book3s_hv.c
3351
tpaca->kvm_hstate.ptid = cpu - vc->pcpu;
arch/powerpc/kvm/book3s_hv.c
3451
vc->pcpu = smp_processor_id();
arch/powerpc/kvm/book3s_hv.c
3470
lp = &per_cpu(preempted_vcores, vc->pcpu);
arch/powerpc/kvm/book3s_hv.c
3789
int pcpu, thr;
arch/powerpc/kvm/book3s_hv.c
3841
pcpu = smp_processor_id();
arch/powerpc/kvm/book3s_hv.c
3869
kvmppc_release_hwthread(pcpu + i);
arch/powerpc/kvm/book3s_hv.c
3873
kvmppc_clear_host_core(pcpu);
arch/powerpc/kvm/book3s_hv.c
3911
struct paca_struct *paca = paca_ptrs[pcpu + thr];
arch/powerpc/kvm/book3s_hv.c
3959
pvc->pcpu = pcpu + thr;
arch/powerpc/kvm/book3s_hv.c
3999
kvmppc_ipi_thread(pcpu + thr);
arch/powerpc/kvm/book3s_hv.c
4051
kvmppc_set_host_core(pcpu);
arch/powerpc/kvm/book3s_hv.c
4071
kvmppc_release_hwthread(pcpu + i);
arch/powerpc/kvm/book3s_hv.c
4073
kvmppc_ipi_thread(pcpu + i);
arch/powerpc/kvm/book3s_hv.c
4873
int trap, r, pcpu;
arch/powerpc/kvm/book3s_hv.c
4909
pcpu = smp_processor_id();
arch/powerpc/kvm/book3s_hv.c
4911
kvmppc_prepare_radix_vcpu(vcpu, pcpu);
arch/powerpc/kvm/book3s_hv.c
4923
vcpu->cpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
4924
vcpu->arch.thread_cpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
4925
vc->pcpu = pcpu;
arch/powerpc/kvm/book3s_hv.c
794
unsigned int pcpu, u64 now,
arch/powerpc/kvm/book3s_hv.c
806
dt->processor_id = cpu_to_be16(pcpu + vcpu->arch.ptid);
arch/powerpc/kvm/book3s_hv.c
851
__kvmppc_create_dtl_entry(vcpu, vpa, vc->pcpu, now + kvmppc_get_tb_offset(vcpu), stolen);
arch/powerpc/kvm/book3s_hv.c
874
__kvmppc_create_dtl_entry(vcpu, vpa, vc->pcpu, now, stolen_delta);
arch/powerpc/kvm/book3s_hv_builtin.c
241
int cpu = vc->pcpu;
arch/powerpc/kvm/book3s_hv_builtin.c
620
void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu)
arch/powerpc/kvm/book3s_hv_builtin.c
622
if (cpumask_test_cpu(pcpu, &kvm->arch.need_tlb_flush)) {
arch/powerpc/kvm/book3s_hv_builtin.c
626
cpumask_clear_cpu(pcpu, &kvm->arch.need_tlb_flush);
arch/powerpc/kvm/book3s_hv_p9_entry.c
453
static void check_need_tlb_flush(struct kvm *kvm, int pcpu,
arch/powerpc/kvm/book3s_hv_p9_entry.c
465
if (likely(!cpumask_test_cpu(pcpu, need_tlb_flush)))
arch/powerpc/kvm/book3s_hv_p9_entry.c
477
for (i = cpu_first_tlb_thread_sibling(pcpu);
arch/powerpc/kvm/book3s_hv_p9_entry.c
478
i <= cpu_last_tlb_thread_sibling(pcpu);
arch/powerpc/kvm/book3s_hv_p9_entry.c
491
cpumask_clear_cpu(pcpu, need_tlb_flush);
arch/powerpc/kvm/book3s_hv_p9_entry.c
703
check_need_tlb_flush(kvm, vc->pcpu, nested);
arch/powerpc/kvm/book3s_hv_rm_mmu.c
57
cpu = local_paca->kvm_hstate.kvm_vcore->pcpu;
arch/powerpc/kvm/book3s_hv_rm_xics.c
718
int pcpu = raw_smp_processor_id();
arch/powerpc/kvm/book3s_hv_rm_xics.c
720
pcpu = cpu_first_thread_sibling(pcpu);
arch/powerpc/kvm/book3s_hv_rm_xics.c
722
if (state->intr_cpu != pcpu) {
arch/powerpc/kvm/book3s_hv_rm_xics.c
724
xics_opal_set_server(state->host_irq, pcpu);
arch/powerpc/platforms/powernv/smp.c
107
pr_devel("OPAL: Starting CPU %d (HW 0x%x)...\n", nr, pcpu);
arch/powerpc/platforms/powernv/smp.c
108
rc = opal_start_cpu(pcpu, start_here);
arch/powerpc/platforms/powernv/smp.c
121
" (status %d)...\n", nr, pcpu, status);
arch/powerpc/platforms/powernv/smp.c
67
unsigned int pcpu;
arch/powerpc/platforms/powernv/smp.c
76
pcpu = get_hard_smp_processor_id(nr);
arch/powerpc/platforms/powernv/smp.c
90
rc = opal_query_cpu_status(pcpu, &status);
arch/powerpc/platforms/pseries/hotplug-cpu.c
121
unsigned int pcpu = get_hard_smp_processor_id(cpu);
arch/powerpc/platforms/pseries/hotplug-cpu.c
125
cpu_status = smp_query_cpu_stopped(pcpu);
arch/powerpc/platforms/pseries/hotplug-cpu.c
132
cpu, pcpu);
arch/powerpc/platforms/pseries/hotplug-cpu.c
141
cpu, pcpu);
arch/powerpc/platforms/pseries/pseries.h
30
int smp_query_cpu_stopped(unsigned int pcpu);
arch/powerpc/platforms/pseries/smp.c
102
if (smp_query_cpu_stopped(pcpu) == QCSS_NOT_STOPPED){
arch/powerpc/platforms/pseries/smp.c
115
status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, pcpu);
arch/powerpc/platforms/pseries/smp.c
55
int smp_query_cpu_stopped(unsigned int pcpu)
arch/powerpc/platforms/pseries/smp.c
66
status = rtas_call(qcss_tok, 1, 2, &cpu_status, pcpu);
arch/powerpc/platforms/pseries/smp.c
92
unsigned int pcpu;
arch/powerpc/platforms/pseries/smp.c
99
pcpu = get_hard_smp_processor_id(lcpu);
arch/s390/include/asm/lowcore.h
107
__u64 pcpu; /* 0x0288 */
arch/s390/include/asm/processor.h
56
DECLARE_PER_CPU(struct pcpu, pcpu_devices);
arch/s390/include/asm/processor.h
60
static __always_inline struct pcpu *this_pcpu(void)
arch/s390/include/asm/processor.h
62
return (struct pcpu *)(get_lowcore()->pcpu);
arch/s390/include/asm/smp.h
65
u16 pcpu = stap();
arch/s390/include/asm/smp.h
68
__pcpu_sigp(pcpu, SIGP_STOP, 0, NULL);
arch/s390/kernel/asm-offsets.c
121
OFFSET(__LC_PCPU, lowcore, pcpu);
arch/s390/kernel/asm-offsets.c
186
OFFSET(__PCPU_FLAGS, pcpu, flags);
arch/s390/kernel/setup.c
399
lc->pcpu = (unsigned long)per_cpu_ptr(&pcpu_devices, 0);
arch/s390/kernel/smp.c
1011
struct pcpu *pcpu;
arch/s390/kernel/smp.c
1028
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
1032
if (pcpu->state != CPU_STATE_CONFIGURED)
arch/s390/kernel/smp.c
1034
rc = sclp_core_deconfigure(pcpu->address >> smp_cpu_mt_shift);
arch/s390/kernel/smp.c
1047
if (pcpu->state != CPU_STATE_STANDBY)
arch/s390/kernel/smp.c
1049
rc = sclp_core_configure(pcpu->address >> smp_cpu_mt_shift);
arch/s390/kernel/smp.c
129
static int pcpu_sigp_retry(struct pcpu *pcpu, u8 order, u32 parm)
arch/s390/kernel/smp.c
134
cc = __pcpu_sigp(pcpu->address, order, parm, NULL);
arch/s390/kernel/smp.c
143
static inline int pcpu_stopped(struct pcpu *pcpu)
arch/s390/kernel/smp.c
147
if (__pcpu_sigp(pcpu->address, SIGP_SENSE,
arch/s390/kernel/smp.c
153
static inline int pcpu_running(struct pcpu *pcpu)
arch/s390/kernel/smp.c
155
if (__pcpu_sigp(pcpu->address, SIGP_SENSE_RUNNING,
arch/s390/kernel/smp.c
165
static struct pcpu *pcpu_find_address(const struct cpumask *mask, u16 address)
arch/s390/kernel/smp.c
175
static void pcpu_ec_call(struct pcpu *pcpu, int ec_bit)
arch/s390/kernel/smp.c
177
if (test_and_set_bit(ec_bit, &pcpu->ec_mask))
arch/s390/kernel/smp.c
179
pcpu->ec_clk = get_tod_clock_fast();
arch/s390/kernel/smp.c
180
pcpu_sigp_retry(pcpu, SIGP_EXTERNAL_CALL, 0);
arch/s390/kernel/smp.c
183
static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
arch/s390/kernel/smp.c
210
pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, __pa(lc));
arch/s390/kernel/smp.c
223
static void pcpu_free_lowcore(struct pcpu *pcpu, int cpu)
arch/s390/kernel/smp.c
232
pcpu_sigp_retry(pcpu, SIGP_SET_PREFIX, 0);
arch/s390/kernel/smp.c
242
static void pcpu_prepare_secondary(struct pcpu *pcpu, int cpu)
arch/s390/kernel/smp.c
250
lc->pcpu = (unsigned long)pcpu;
arch/s390/kernel/smp.c
310
static void __noreturn pcpu_delegate(struct pcpu *pcpu, int cpu,
arch/s390/kernel/smp.c
320
if (pcpu->address == source_cpu) {
arch/s390/kernel/smp.c
325
pcpu_sigp_retry(pcpu, SIGP_STOP, 0);
arch/s390/kernel/smp.c
326
pcpu_sigp_retry(pcpu, SIGP_CPU_RESET, 0);
arch/s390/kernel/smp.c
346
: : "d" (pcpu->address), "d" (source_cpu),
arch/s390/kernel/smp.c
437
struct pcpu *pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
438
set_bit(ec_stop_cpu, &pcpu->ec_mask);
arch/s390/kernel/smp.c
439
while (__pcpu_sigp(pcpu->address, SIGP_EMERGENCY_SIGNAL,
arch/s390/kernel/smp.c
461
struct pcpu *pcpu;
arch/s390/kernel/smp.c
477
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
478
pcpu_sigp_retry(pcpu, SIGP_STOP, 0);
arch/s390/kernel/smp.c
479
while (!pcpu_stopped(pcpu))
arch/s390/kernel/smp.c
548
struct pcpu *pcpu;
arch/s390/kernel/smp.c
551
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
554
if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_STATUS_AT_ADDRESS,
arch/s390/kernel/smp.c
562
if (__pcpu_sigp_relax(pcpu->address, SIGP_STORE_ADDITIONAL_STATUS,
arch/s390/kernel/smp.c
712
struct pcpu *pcpu;
arch/s390/kernel/smp.c
724
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
725
pcpu->address = address + i;
arch/s390/kernel/smp.c
727
pcpu->state = CPU_STATE_CONFIGURED;
arch/s390/kernel/smp.c
729
pcpu->state = CPU_STATE_STANDBY;
arch/s390/kernel/smp.c
79
DEFINE_PER_CPU(struct pcpu, pcpu_devices);
arch/s390/kernel/smp.c
858
struct pcpu *pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
86
static struct pcpu *ipl_pcpu;
arch/s390/kernel/smp.c
861
if (pcpu->state != CPU_STATE_CONFIGURED)
arch/s390/kernel/smp.c
863
if (pcpu_sigp_retry(pcpu, SIGP_INITIAL_CPU_RESET, 0) !=
arch/s390/kernel/smp.c
867
rc = pcpu_alloc_lowcore(pcpu, cpu);
arch/s390/kernel/smp.c
875
pcpu_prepare_secondary(pcpu, cpu);
arch/s390/kernel/smp.c
919
struct pcpu *pcpu;
arch/s390/kernel/smp.c
922
pcpu = per_cpu_ptr(&pcpu_devices, cpu);
arch/s390/kernel/smp.c
923
while (!pcpu_stopped(pcpu))
arch/s390/kernel/smp.c
925
pcpu_free_lowcore(pcpu, cpu);
arch/s390/kernel/smp.c
928
pcpu->flags = 0;
arch/s390/kernel/smp.c
970
lc->pcpu = (unsigned long)ipl_pcpu;
arch/x86/events/amd/ibs.c
1228
struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
arch/x86/events/amd/ibs.c
1229
struct perf_event *event = pcpu->event;
arch/x86/events/amd/ibs.c
1240
if (!test_bit(IBS_STARTED, pcpu->state)) {
arch/x86/events/amd/ibs.c
1248
if (test_and_clear_bit(IBS_STOPPED, pcpu->state))
arch/x86/events/amd/ibs.c
1416
struct cpu_perf_ibs __percpu *pcpu;
arch/x86/events/amd/ibs.c
1419
pcpu = alloc_percpu(struct cpu_perf_ibs);
arch/x86/events/amd/ibs.c
1420
if (!pcpu)
arch/x86/events/amd/ibs.c
1423
perf_ibs->pcpu = pcpu;
arch/x86/events/amd/ibs.c
1427
perf_ibs->pcpu = NULL;
arch/x86/events/amd/ibs.c
1428
free_percpu(pcpu);
arch/x86/events/amd/ibs.c
1497
free_percpu(perf_ibs_op.pcpu);
arch/x86/events/amd/ibs.c
1498
perf_ibs_op.pcpu = NULL;
arch/x86/events/amd/ibs.c
1501
free_percpu(perf_ibs_fetch.pcpu);
arch/x86/events/amd/ibs.c
1502
perf_ibs_fetch.pcpu = NULL;
arch/x86/events/amd/ibs.c
471
struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
arch/x86/events/amd/ibs.c
494
set_bit(IBS_STARTED, pcpu->state);
arch/x86/events/amd/ibs.c
495
clear_bit(IBS_STOPPING, pcpu->state);
arch/x86/events/amd/ibs.c
505
struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
arch/x86/events/amd/ibs.c
509
if (test_and_set_bit(IBS_STOPPING, pcpu->state))
arch/x86/events/amd/ibs.c
512
stopping = test_bit(IBS_STARTED, pcpu->state);
arch/x86/events/amd/ibs.c
526
set_bit(IBS_STOPPED, pcpu->state);
arch/x86/events/amd/ibs.c
537
clear_bit(IBS_STARTED, pcpu->state);
arch/x86/events/amd/ibs.c
558
struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
arch/x86/events/amd/ibs.c
560
if (test_and_set_bit(IBS_ENABLED, pcpu->state))
arch/x86/events/amd/ibs.c
565
pcpu->event = event;
arch/x86/events/amd/ibs.c
576
struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu);
arch/x86/events/amd/ibs.c
578
if (!test_and_clear_bit(IBS_ENABLED, pcpu->state))
arch/x86/events/amd/ibs.c
583
pcpu->event = NULL;
arch/x86/events/amd/ibs.c
96
struct cpu_perf_ibs __percpu *pcpu;
arch/x86/include/asm/topology.h
250
int pcpu = topology_get_primary_thread(cpu);
arch/x86/include/asm/topology.h
252
return pcpu >= 0 ? cpu_online(pcpu) : false;
arch/x86/kernel/acpi/boot.c
785
int acpi_map_cpu(acpi_handle handle, phys_cpuid_t physid, u32 acpi_id, int *pcpu)
arch/x86/kernel/acpi/boot.c
797
*pcpu = cpu;
arch/x86/xen/enlighten.c
367
pin_override.pcpu = cpu;
drivers/base/devres.c
1241
void __percpu *pcpu;
drivers/base/devres.c
1243
pcpu = __alloc_percpu(size, align);
drivers/base/devres.c
1244
if (!pcpu)
drivers/base/devres.c
1249
free_percpu(pcpu);
drivers/base/devres.c
1253
*(void __percpu **)p = pcpu;
drivers/base/devres.c
1257
return pcpu;
drivers/crypto/caam/qi.c
443
int *pcpu = &get_cpu_var(last_cpu);
drivers/crypto/caam/qi.c
445
*pcpu = cpumask_next_wrap(*pcpu, cpus);
drivers/crypto/caam/qi.c
446
*cpu = *pcpu;
drivers/firmware/arm_ffa/driver.c
1800
struct ffa_pcpu_irq *pcpu = irq_data;
drivers/firmware/arm_ffa/driver.c
1801
struct ffa_drv_info *info = pcpu->info;
drivers/firmware/arm_ffa/driver.c
1810
struct ffa_pcpu_irq *pcpu = irq_data;
drivers/firmware/arm_ffa/driver.c
1811
struct ffa_drv_info *info = pcpu->info;
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1223
struct mvpp2_port_pcpu __percpu *pcpu;
drivers/net/ethernet/marvell/mvpp2/mvpp2.h
1452
struct mvpp2_txq_pcpu __percpu *pcpu;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2550
txq_pcpu_aux = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2891
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3152
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3187
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3257
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3646
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3682
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4087
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4193
struct mvpp2_txq_pcpu *txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4374
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4465
struct mvpp2_port_pcpu *port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4904
port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6035
txq->pcpu = alloc_percpu(struct mvpp2_txq_pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6036
if (!txq->pcpu) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6045
txq_pcpu = per_cpu_ptr(txq->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6119
free_percpu(port->txqs[queue]->pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7001
port->pcpu = alloc_percpu(struct mvpp2_port_pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7002
if (!port->pcpu) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7009
port_pcpu = per_cpu_ptr(port->pcpu, thread);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7172
free_percpu(port->pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7175
free_percpu(port->txqs[i]->pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7196
free_percpu(port->pcpu);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
7199
free_percpu(port->txqs[i]->pcpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
447
struct stmmac_pcpu_stats *pcpu;
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
451
pcpu = per_cpu_ptr(priv->xstats.pcpu_stats, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
453
start = u64_stats_fetch_begin(&pcpu->syncp);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
454
irq_n = u64_stats_read(&pcpu->rx_normal_irq_n[q]);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
455
} while (u64_stats_fetch_retry(&pcpu->syncp, start));
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
468
struct stmmac_pcpu_stats *pcpu;
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
472
pcpu = per_cpu_ptr(priv->xstats.pcpu_stats, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
474
start = u64_stats_fetch_begin(&pcpu->syncp);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
475
irq_n = u64_stats_read(&pcpu->tx_normal_irq_n[q]);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
476
} while (u64_stats_fetch_retry(&pcpu->syncp, start));
drivers/tee/optee/smc_abi.c
1109
struct optee_pcpu *pcpu = dev_id;
drivers/tee/optee/smc_abi.c
1110
struct optee *optee = pcpu->optee;
drivers/xen/pcpu.c
103
struct pcpu *cpu = container_of(dev, struct pcpu, dev);
drivers/xen/pcpu.c
112
struct pcpu *pcpu = container_of(dev, struct pcpu, dev);
drivers/xen/pcpu.c
124
ret = xen_pcpu_down(pcpu->cpu_id);
drivers/xen/pcpu.c
127
ret = xen_pcpu_up(pcpu->cpu_id);
drivers/xen/pcpu.c
172
struct pcpu *pcpu)
drivers/xen/pcpu.c
175
!xen_pcpu_online(pcpu->flags)) {
drivers/xen/pcpu.c
177
pcpu->flags |= XEN_PCPU_FLAGS_ONLINE;
drivers/xen/pcpu.c
178
kobject_uevent(&pcpu->dev.kobj, KOBJ_ONLINE);
drivers/xen/pcpu.c
180
xen_pcpu_online(pcpu->flags)) {
drivers/xen/pcpu.c
182
pcpu->flags &= ~XEN_PCPU_FLAGS_ONLINE;
drivers/xen/pcpu.c
183
kobject_uevent(&pcpu->dev.kobj, KOBJ_OFFLINE);
drivers/xen/pcpu.c
187
static struct pcpu *get_pcpu(uint32_t cpu_id)
drivers/xen/pcpu.c
189
struct pcpu *pcpu;
drivers/xen/pcpu.c
191
list_for_each_entry(pcpu, &xen_pcpus, list) {
drivers/xen/pcpu.c
192
if (pcpu->cpu_id == cpu_id)
drivers/xen/pcpu.c
193
return pcpu;
drivers/xen/pcpu.c
201
struct pcpu *pcpu = container_of(dev, struct pcpu, dev);
drivers/xen/pcpu.c
203
list_del(&pcpu->list);
drivers/xen/pcpu.c
204
kfree(pcpu);
drivers/xen/pcpu.c
207
static void unregister_and_remove_pcpu(struct pcpu *pcpu)
drivers/xen/pcpu.c
211
if (!pcpu)
drivers/xen/pcpu.c
214
dev = &pcpu->dev;
drivers/xen/pcpu.c
219
static int register_pcpu(struct pcpu *pcpu)
drivers/xen/pcpu.c
224
if (!pcpu)
drivers/xen/pcpu.c
227
dev = &pcpu->dev;
drivers/xen/pcpu.c
229
dev->id = pcpu->cpu_id;
drivers/xen/pcpu.c
242
static struct pcpu *create_and_register_pcpu(struct xenpf_pcpuinfo *info)
drivers/xen/pcpu.c
244
struct pcpu *pcpu;
drivers/xen/pcpu.c
250
pcpu = kzalloc_obj(struct pcpu);
drivers/xen/pcpu.c
251
if (!pcpu)
drivers/xen/pcpu.c
254
INIT_LIST_HEAD(&pcpu->list);
drivers/xen/pcpu.c
255
pcpu->cpu_id = info->xen_cpuid;
drivers/xen/pcpu.c
256
pcpu->acpi_id = info->acpi_id;
drivers/xen/pcpu.c
257
pcpu->flags = info->flags;
drivers/xen/pcpu.c
260
list_add_tail(&pcpu->list, &xen_pcpus);
drivers/xen/pcpu.c
262
err = register_pcpu(pcpu);
drivers/xen/pcpu.c
268
return pcpu;
drivers/xen/pcpu.c
277
struct pcpu *pcpu = NULL;
drivers/xen/pcpu.c
293
pcpu = get_pcpu(cpu);
drivers/xen/pcpu.c
299
unregister_and_remove_pcpu(pcpu);
drivers/xen/pcpu.c
303
if (!pcpu) {
drivers/xen/pcpu.c
304
pcpu = create_and_register_pcpu(info);
drivers/xen/pcpu.c
305
if (IS_ERR_OR_NULL(pcpu))
drivers/xen/pcpu.c
308
pcpu_online_status(info, pcpu);
drivers/xen/pcpu.c
323
struct pcpu *pcpu, *tmp;
drivers/xen/pcpu.c
333
list_for_each_entry_safe(pcpu, tmp, &xen_pcpus, list)
drivers/xen/pcpu.c
334
unregister_and_remove_pcpu(pcpu);
drivers/xen/pcpu.c
393
const struct pcpu *pcpu;
drivers/xen/pcpu.c
397
list_for_each_entry(pcpu, &xen_pcpus, list)
drivers/xen/pcpu.c
398
if (pcpu->acpi_id == acpi_id) {
drivers/xen/pcpu.c
399
online = pcpu->flags & XEN_PCPU_FLAGS_ONLINE;
include/linux/acpi.h
321
int *pcpu);
include/linux/cpufreq.h
1166
static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name,
include/linux/cpufreq.h
1173
ret = parse_perf_domain(pcpu, list_name, cell_name, pargs);
include/linux/cpufreq.h
1177
cpumask_set_cpu(pcpu, cpumask);
include/linux/cpufreq.h
1180
if (cpu == pcpu)
include/linux/cpufreq.h
1213
static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name,
include/linux/mm_types.h
1545
mm->mm_cid.pcpu = alloc_percpu_noprof(struct mm_cid_pcpu);
include/linux/mm_types.h
1546
if (!mm->mm_cid.pcpu)
include/linux/mm_types.h
1555
free_percpu(mm->mm_cid.pcpu);
include/linux/mm_types.h
1556
mm->mm_cid.pcpu = NULL;
include/linux/notifier.h
107
#define SRCU_NOTIFIER_INIT(name, pcpu) \
include/linux/notifier.h
112
.srcu = __SRCU_STRUCT_INIT(name.srcu, name.srcuu, pcpu, 0), \
include/linux/padata.h
77
cpumask_var_t pcpu;
include/linux/psi_types.h
168
struct psi_group_cpu __percpu *pcpu;
include/linux/rseq_types.h
173
struct mm_cid_pcpu __percpu *pcpu;
include/xen/interface/sched.h
141
int32_t pcpu;
kernel/module/internal.h
91
unsigned int pcpu;
kernel/module/kallsyms.c
135
info->index.pcpu)) {
kernel/module/kallsyms.c
203
info->index.pcpu)) {
kernel/module/kallsyms.c
50
if (sym->st_shndx == SHN_ABS || sym->st_shndx == info->index.pcpu)
kernel/module/main.c
1579
if (sym[i].st_shndx == info->index.pcpu)
kernel/module/main.c
1611
(!infosec || infosec != info->index.pcpu))
kernel/module/main.c
2289
info->index.pcpu = find_pcpusec(info);
kernel/module/main.c
2900
info->sechdrs[info->index.pcpu].sh_flags &= ~(unsigned long)SHF_ALLOC;
kernel/module/main.c
2953
percpu_modcopy(mod, (void *)info->sechdrs[info->index.pcpu].sh_addr,
kernel/module/main.c
2954
info->sechdrs[info->index.pcpu].sh_size);
kernel/module/main.c
459
Elf_Shdr *pcpusec = &info->sechdrs[info->index.pcpu];
kernel/module/main.c
551
if (info->sechdrs[info->index.pcpu].sh_size != 0)
kernel/padata.c
1004
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
297
cpu = cpumask_first(pd->cpumask.pcpu);
kernel/padata.c
299
cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu);
kernel/padata.c
397
cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu);
kernel/padata.c
530
for_each_cpu(cpu, pd->cpumask.pcpu) {
kernel/padata.c
557
if (!alloc_cpumask_var(&pd->cpumask.pcpu, GFP_KERNEL))
kernel/padata.c
562
cpumask_and(pd->cpumask.pcpu, pinst->cpumask.pcpu, cpu_online_mask);
kernel/padata.c
569
pd->cpu = cpumask_first(pd->cpumask.pcpu);
kernel/padata.c
574
free_cpumask_var(pd->cpumask.pcpu);
kernel/padata.c
587
free_cpumask_var(pd->cpumask.pcpu);
kernel/padata.c
678
cpumask_copy(pinst->cpumask.pcpu, pcpumask);
kernel/padata.c
714
parallel_mask = pinst->cpumask.pcpu;
kernel/padata.c
72
int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu);
kernel/padata.c
74
return cpumask_nth(cpu_index, pd->cpumask.pcpu);
kernel/padata.c
740
if (padata_validate_cpumask(pinst, pinst->cpumask.pcpu) &&
kernel/padata.c
753
if (!padata_validate_cpumask(pinst, pinst->cpumask.pcpu) ||
kernel/padata.c
765
return cpumask_test_cpu(cpu, pinst->cpumask.pcpu) ||
kernel/padata.c
812
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
847
cpumask = pinst->cpumask.pcpu;
kernel/padata.c
972
if (!alloc_cpumask_var(&pinst->cpumask.pcpu, GFP_KERNEL))
kernel/padata.c
975
free_cpumask_var(pinst->cpumask.pcpu);
kernel/padata.c
981
cpumask_copy(pinst->cpumask.pcpu, cpu_possible_mask);
kernel/sched/core.c
10585
struct mm_cid_pcpu *pcp = per_cpu_ptr(mm->mm_cid.pcpu, cpu);
kernel/sched/core.c
10630
mm_cid_transit_to_cpu(t, per_cpu_ptr(mm->mm_cid.pcpu, task_cpu(t)));
kernel/sched/core.c
10674
struct mm_cid_pcpu *pcp = this_cpu_ptr(mm->mm_cid.pcpu);
kernel/sched/core.c
10771
mm_drop_cid_on_cpu(mm, this_cpu_ptr(mm->mm_cid.pcpu));
kernel/sched/core.c
10780
mm_cid_transit_to_task(t, this_cpu_ptr(mm->mm_cid.pcpu));
kernel/sched/ext.c
3754
free_percpu(sch->pcpu);
kernel/sched/ext.c
4955
sch->pcpu = alloc_percpu(struct scx_sched_pcpu);
kernel/sched/ext.c
4956
if (!sch->pcpu) {
kernel/sched/ext.c
4985
free_percpu(sch->pcpu);
kernel/sched/ext.c
690
this_cpu_add((sch)->pcpu->event_stats.name, (cnt)); \
kernel/sched/ext.c
703
__this_cpu_add((sch)->pcpu->event_stats.name, (cnt)); \
kernel/sched/ext.c
7385
e_cpu = &per_cpu_ptr(sch->pcpu, cpu)->event_stats;
kernel/sched/ext_internal.h
901
struct scx_sched_pcpu __percpu *pcpu;
kernel/sched/psi.c
1037
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
1121
cgroup->psi->pcpu = alloc_percpu(struct psi_group_cpu);
kernel/sched/psi.c
1122
if (!cgroup->psi->pcpu) {
kernel/sched/psi.c
1137
free_percpu(cgroup->psi->pcpu);
kernel/sched/psi.c
176
.pcpu = &system_group_pcpu,
kernel/sched/psi.c
275
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
805
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
944
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/sched.h
3887
__this_cpu_write(mm->mm_cid.pcpu->cid, cid);
kernel/sched/sched.h
3944
mm_drop_cid_on_cpu(mm, this_cpu_ptr(mm->mm_cid.pcpu));
kernel/sched/sched.h
3966
cpu_cid = __this_cpu_read(mm->mm_cid.pcpu->cid);
lib/test_vmalloc.c
319
void __percpu **pcpu;
lib/test_vmalloc.c
323
pcpu = vmalloc(sizeof(void __percpu *) * nr_pcpu_objects);
lib/test_vmalloc.c
324
if (!pcpu)
lib/test_vmalloc.c
335
pcpu[i] = __alloc_percpu(size, align);
lib/test_vmalloc.c
336
if (!pcpu[i])
lib/test_vmalloc.c
341
free_percpu(pcpu[i]);
lib/test_vmalloc.c
343
vfree(pcpu);
mm/percpu.c
254
static void pcpu_set_page_chunk(struct page *page, struct pcpu_chunk *pcpu)
mm/percpu.c
256
page->private = (unsigned long)pcpu;
net/core/page_pool.c
102
const struct page_pool_recycle_stats *pcpu =
net/core/page_pool.c
105
stats->recycle_stats.cached += pcpu->cached;
net/core/page_pool.c
106
stats->recycle_stats.cache_full += pcpu->cache_full;
net/core/page_pool.c
107
stats->recycle_stats.ring += pcpu->ring;
net/core/page_pool.c
108
stats->recycle_stats.ring_full += pcpu->ring_full;
net/core/page_pool.c
109
stats->recycle_stats.released_refcnt += pcpu->released_refcnt;
net/ipv4/fib_trie.c
2578
const struct trie_use_stats *pcpu = per_cpu_ptr(stats, cpu);
net/ipv4/fib_trie.c
2580
s.gets += pcpu->gets;
net/ipv4/fib_trie.c
2581
s.backtrack += pcpu->backtrack;
net/ipv4/fib_trie.c
2582
s.semantic_match_passed += pcpu->semantic_match_passed;
net/ipv4/fib_trie.c
2583
s.semantic_match_miss += pcpu->semantic_match_miss;
net/ipv4/fib_trie.c
2584
s.null_node_hit += pcpu->null_node_hit;
net/ipv4/fib_trie.c
2585
s.resize_node_skipped += pcpu->resize_node_skipped;
tools/perf/util/bpf_off_cpu.c
69
struct perf_cpu pcpu;
tools/perf/util/bpf_off_cpu.c
91
perf_cpu_map__for_each_cpu(pcpu, i, evsel->core.cpus) {
tools/perf/util/bpf_off_cpu.c
93
int cpu_nr = pcpu.cpu;
tools/testing/selftests/kvm/arch_timer.c
83
uint32_t pcpu;
tools/testing/selftests/kvm/arch_timer.c
92
pcpu = rand() % nproc_conf;
tools/testing/selftests/kvm/arch_timer.c
93
} while (!CPU_ISSET(pcpu, &online_cpuset));
tools/testing/selftests/kvm/arch_timer.c
95
return pcpu;
tools/testing/selftests/kvm/lib/kvm_util.c
641
uint32_t pcpu = atoi_non_negative("CPU number", cpu_str);
tools/testing/selftests/kvm/lib/kvm_util.c
643
TEST_ASSERT(CPU_ISSET(pcpu, allowed_mask),
tools/testing/selftests/kvm/lib/kvm_util.c
644
"Not allowed to run on pCPU '%d', check cgroups?", pcpu);
tools/testing/selftests/kvm/lib/kvm_util.c
645
return pcpu;