Symbol: target_cpu
arch/powerpc/platforms/pseries/mobility.c
493
static void prod_single(unsigned int target_cpu)
arch/powerpc/platforms/pseries/mobility.c
498
hwid = get_hard_smp_processor_id(target_cpu);
arch/powerpc/platforms/pseries/mobility.c
503
target_cpu, hwid, hvrc);
arch/sparc/kernel/sun4d_irq.c
487
int target_cpu;
arch/sparc/kernel/sun4d_irq.c
489
target_cpu = boot_cpu_id;
arch/sparc/kernel/sun4d_irq.c
495
set_sbi_tid(devid, target_cpu << 3);
arch/sparc/kernel/sun4d_irq.c
496
board_to_cpu[board] = target_cpu;
drivers/hv/channel.c
179
page = alloc_pages_node(cpu_to_node(newchannel->target_cpu),
drivers/hv/channel.c
732
open_msg->target_vp = hv_cpu_number_to_vp_number(newchannel->target_cpu);
drivers/hv/channel_mgmt.c
474
hv_clear_allocated_cpu(channel->target_cpu);
drivers/hv/channel_mgmt.c
717
if (primary->target_cpu == cpu)
drivers/hv/channel_mgmt.c
721
if (sc != chn && sc->target_cpu == cpu)
drivers/hv/channel_mgmt.c
748
u32 target_cpu;
drivers/hv/channel_mgmt.c
762
channel->target_cpu = VMBUS_CONNECT_CPU;
drivers/hv/channel_mgmt.c
794
target_cpu = cpumask_first(available_mask);
drivers/hv/channel_mgmt.c
795
cpumask_set_cpu(target_cpu, allocated_mask);
drivers/hv/channel_mgmt.c
798
i > ncpu || !hv_cpuself_used(target_cpu, channel))
drivers/hv/channel_mgmt.c
802
channel->target_cpu = target_cpu;
drivers/hv/hv.c
559
if (channel->target_cpu == cpu ||
drivers/hv/hv.c
560
channel->target_cpu == VMBUS_CONNECT_CPU)
drivers/hv/hv.c
601
if (channel->target_cpu == cpu) {
drivers/hv/hv.c
609
if (sc->target_cpu == cpu) {
drivers/hv/hyperv_vmbus.h
491
if (channel->target_cpu == cpu)
drivers/hv/hyperv_vmbus.h
494
if (sc->target_cpu == cpu)
drivers/hv/vmbus_drv.c
1756
return sprintf(buf, "%u\n", channel->target_cpu);
drivers/hv/vmbus_drv.c
1759
int vmbus_channel_set_cpu(struct vmbus_channel *channel, u32 target_cpu)
drivers/hv/vmbus_drv.c
1771
if (target_cpu >= nr_cpumask_bits)
drivers/hv/vmbus_drv.c
1774
if (!cpumask_test_cpu(target_cpu, housekeeping_cpumask(HK_TYPE_MANAGED_IRQ)))
drivers/hv/vmbus_drv.c
1777
if (!cpu_online(target_cpu))
drivers/hv/vmbus_drv.c
1812
origin_cpu = channel->target_cpu;
drivers/hv/vmbus_drv.c
1813
if (target_cpu == origin_cpu)
drivers/hv/vmbus_drv.c
1817
hv_cpu_number_to_vp_number(target_cpu))) {
drivers/hv/vmbus_drv.c
1837
channel->target_cpu = target_cpu;
drivers/hv/vmbus_drv.c
1841
hv_update_allocated_cpus(origin_cpu, target_cpu);
drivers/hv/vmbus_drv.c
1846
origin_cpu, target_cpu);
drivers/hv/vmbus_drv.c
1856
u32 target_cpu;
drivers/hv/vmbus_drv.c
1859
if (sscanf(buf, "%u", &target_cpu) != 1)
drivers/hv/vmbus_drv.c
1864
ret = vmbus_channel_set_cpu(channel, target_cpu);
drivers/hv/vmbus_drv.c
223
return sysfs_emit(buf, "%d\n", cpu_to_node(hv_dev->channel->target_cpu));
drivers/hv/vmbus_drv.c
508
channel->target_cpu);
drivers/hv/vmbus_drv.c
515
cur_sc->target_cpu);
drivers/irqchip/irq-apple-aic.c
1062
if (irqc->info.target_cpu)
drivers/irqchip/irq-apple-aic.c
1064
aic_ic_write(irqc, irqc->info.target_cpu + off + i * 4, 1);
drivers/irqchip/irq-apple-aic.c
247
u32 target_cpu;
drivers/irqchip/irq-apple-aic.c
265
.target_cpu = AIC_TARGET_CPU,
drivers/irqchip/irq-apple-aic.c
272
.target_cpu = AIC_TARGET_CPU,
drivers/irqchip/irq-apple-aic.c
281
.target_cpu = AIC_TARGET_CPU,
drivers/irqchip/irq-apple-aic.c
435
BUG_ON(!ic->info.target_cpu);
drivers/irqchip/irq-apple-aic.c
442
aic_ic_write(ic, ic->info.target_cpu + AIC_HWIRQ_IRQ(hwirq) * 4, BIT(cpu));
drivers/irqchip/irq-apple-aic.c
987
off = start_off = irqc->info.target_cpu;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-debugfs.c
82
fq->target_cpu,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
2925
cpumask_set_cpu(fq->target_cpu, xps_mask);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3325
fq->target_cpu = rx_cpu;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3331
fq->target_cpu = txc_cpu;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3339
fq->channel = dpaa2_eth_get_affine_channel(priv, fq->target_cpu);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.h
463
int target_cpu;
drivers/net/hyperv/netvsc.c
324
int node = cpu_to_node(nvchan->channel->target_cpu);
drivers/net/hyperv/netvsc_drv.c
1287
&pcpu_tot[nvchan->channel->target_cpu];
drivers/pci/controller/pci-xgene-msi.c
152
int target_cpu = cpumask_first(mask);
drivers/pci/controller/pci-xgene-msi.c
154
irq_data_update_effective_affinity(irqdata, cpumask_of(target_cpu));
drivers/pci/controller/pcie-iproc-msi.c
211
int target_cpu = cpumask_first(mask);
drivers/pci/controller/pcie-iproc-msi.c
216
if (curr_cpu == target_cpu)
drivers/pci/controller/pcie-iproc-msi.c
220
data->hwirq = hwirq_to_canonical_hwirq(msi, data->hwirq) + target_cpu;
drivers/pci/controller/pcie-iproc-msi.c
224
irq_data_update_effective_affinity(data, cpumask_of(target_cpu));
drivers/scsi/storvsc_drv.c
631
if (device->channel != channel && device->channel->target_cpu == old) {
drivers/scsi/storvsc_drv.c
639
if (cur_chn->target_cpu == old) {
drivers/scsi/storvsc_drv.c
714
stor_device->stor_chns[new_sc->target_cpu] = new_sc;
drivers/scsi/storvsc_drv.c
715
cpumask_set_cpu(new_sc->target_cpu, &stor_device->alloced_cpus);
drivers/scsi/storvsc_drv.c
939
stor_device->stor_chns[device->channel->target_cpu] = device->channel;
drivers/scsi/storvsc_drv.c
940
cpumask_set_cpu(device->channel->target_cpu,
include/linux/hyperv.h
1635
int vmbus_channel_set_cpu(struct vmbus_channel *channel, u32 target_cpu);
include/linux/hyperv.h
877
u32 target_cpu;
include/linux/irqdesc.h
41
unsigned int target_cpu;
include/linux/smp.h
65
int panic_smp_redirect_cpu(int target_cpu, void *msg);
include/trace/events/sched.h
151
__field( int, target_cpu )
include/trace/events/sched.h
158
__entry->target_cpu = task_cpu(p);
include/trace/events/sched.h
163
__entry->target_cpu)
kernel/irq/chip.c
1495
WRITE_ONCE(redir->target_cpu, cpumask_first(dest));
kernel/irq/irqdesc.c
781
unsigned int target_cpu = READ_ONCE(desc->redirect.target_cpu);
kernel/irq/irqdesc.c
819
irq_work_queue_on(&desc->redirect.work, target_cpu);
kernel/panic.c
354
int __weak panic_smp_redirect_cpu(int target_cpu, void *msg)
kernel/panic.c
361
return smp_call_function_single_async(target_cpu, &panic_csd);
kernel/sched/core.c
3366
int target_cpu, int curr_cpu)
kernel/sched/core.c
3375
.dst_cpu = target_cpu,
kernel/sched/core.c
8028
int migrate_task_to(struct task_struct *p, int target_cpu)
kernel/sched/core.c
8030
struct migration_arg arg = { p, target_cpu };
kernel/sched/core.c
8033
if (curr_cpu == target_cpu)
kernel/sched/core.c
8036
if (!cpumask_test_cpu(target_cpu, p->cpus_ptr))
kernel/sched/core.c
8041
trace_sched_move_numa(p, curr_cpu, target_cpu);
kernel/sched/fair.c
12199
int target_cpu = busiest_rq->push_cpu;
kernel/sched/fair.c
12200
struct rq *target_rq = cpu_rq(target_cpu);
kernel/sched/fair.c
12211
if (!cpu_active(busiest_cpu) || !cpu_active(target_cpu))
kernel/sched/fair.c
12232
for_each_domain(target_cpu, sd) {
kernel/sched/fair.c
12240
.dst_cpu = target_cpu,
tools/perf/bench/numa.c
1331
int target_cpu;
tools/perf/bench/numa.c
1343
target_cpu = g->p.nr_cpus-1;
tools/perf/bench/numa.c
1345
target_cpu = 0;
tools/perf/bench/numa.c
1347
orig_mask = bind_to_cpu(target_cpu);
tools/perf/bench/numa.c
1351
printf(" (injecting perturbalance, moved to CPU#%d)\n", target_cpu);
tools/perf/bench/numa.c
272
static cpu_set_t *bind_to_cpu(int target_cpu)
tools/perf/bench/numa.c
292
if (target_cpu == -1) {
tools/perf/bench/numa.c
298
if (target_cpu < 0 || target_cpu >= g->p.nr_cpus)
tools/perf/bench/numa.c
301
CPU_SET_S(target_cpu, size, mask);
tools/testing/selftests/kvm/arm64/psci_test.c
126
static void guest_test_cpu_on(uint64_t target_cpu)
tools/testing/selftests/kvm/arm64/psci_test.c
130
GUEST_ASSERT(!psci_cpu_on(target_cpu, CPU_ON_ENTRY_ADDR, CPU_ON_CONTEXT_ID));
tools/testing/selftests/kvm/arm64/psci_test.c
133
target_state = psci_affinity_info(target_cpu, 0);
tools/testing/selftests/kvm/arm64/psci_test.c
25
static uint64_t psci_cpu_on(uint64_t target_cpu, uint64_t entry_addr,
tools/testing/selftests/kvm/arm64/psci_test.c
30
do_smccc(PSCI_0_2_FN64_CPU_ON, target_cpu, entry_addr, context_id,
tools/testing/selftests/sched_ext/cyclic_kick_wait.bpf.c
49
tgt = target_cpu(this_cpu);