target_cpu
static void prod_single(unsigned int target_cpu)
hwid = get_hard_smp_processor_id(target_cpu);
target_cpu, hwid, hvrc);
int target_cpu;
target_cpu = boot_cpu_id;
set_sbi_tid(devid, target_cpu << 3);
board_to_cpu[board] = target_cpu;
page = alloc_pages_node(cpu_to_node(newchannel->target_cpu),
open_msg->target_vp = hv_cpu_number_to_vp_number(newchannel->target_cpu);
hv_clear_allocated_cpu(channel->target_cpu);
if (primary->target_cpu == cpu)
if (sc != chn && sc->target_cpu == cpu)
u32 target_cpu;
channel->target_cpu = VMBUS_CONNECT_CPU;
target_cpu = cpumask_first(available_mask);
cpumask_set_cpu(target_cpu, allocated_mask);
i > ncpu || !hv_cpuself_used(target_cpu, channel))
channel->target_cpu = target_cpu;
if (channel->target_cpu == cpu ||
channel->target_cpu == VMBUS_CONNECT_CPU)
if (channel->target_cpu == cpu) {
if (sc->target_cpu == cpu) {
if (channel->target_cpu == cpu)
if (sc->target_cpu == cpu)
return sprintf(buf, "%u\n", channel->target_cpu);
int vmbus_channel_set_cpu(struct vmbus_channel *channel, u32 target_cpu)
if (target_cpu >= nr_cpumask_bits)
if (!cpumask_test_cpu(target_cpu, housekeeping_cpumask(HK_TYPE_MANAGED_IRQ)))
if (!cpu_online(target_cpu))
origin_cpu = channel->target_cpu;
if (target_cpu == origin_cpu)
hv_cpu_number_to_vp_number(target_cpu))) {
channel->target_cpu = target_cpu;
hv_update_allocated_cpus(origin_cpu, target_cpu);
origin_cpu, target_cpu);
u32 target_cpu;
if (sscanf(buf, "%u", &target_cpu) != 1)
ret = vmbus_channel_set_cpu(channel, target_cpu);
return sysfs_emit(buf, "%d\n", cpu_to_node(hv_dev->channel->target_cpu));
channel->target_cpu);
cur_sc->target_cpu);
if (irqc->info.target_cpu)
aic_ic_write(irqc, irqc->info.target_cpu + off + i * 4, 1);
u32 target_cpu;
.target_cpu = AIC_TARGET_CPU,
.target_cpu = AIC_TARGET_CPU,
.target_cpu = AIC_TARGET_CPU,
BUG_ON(!ic->info.target_cpu);
aic_ic_write(ic, ic->info.target_cpu + AIC_HWIRQ_IRQ(hwirq) * 4, BIT(cpu));
off = start_off = irqc->info.target_cpu;
fq->target_cpu,
cpumask_set_cpu(fq->target_cpu, xps_mask);
fq->target_cpu = rx_cpu;
fq->target_cpu = txc_cpu;
fq->channel = dpaa2_eth_get_affine_channel(priv, fq->target_cpu);
int target_cpu;
int node = cpu_to_node(nvchan->channel->target_cpu);
&pcpu_tot[nvchan->channel->target_cpu];
int target_cpu = cpumask_first(mask);
irq_data_update_effective_affinity(irqdata, cpumask_of(target_cpu));
int target_cpu = cpumask_first(mask);
if (curr_cpu == target_cpu)
data->hwirq = hwirq_to_canonical_hwirq(msi, data->hwirq) + target_cpu;
irq_data_update_effective_affinity(data, cpumask_of(target_cpu));
if (device->channel != channel && device->channel->target_cpu == old) {
if (cur_chn->target_cpu == old) {
stor_device->stor_chns[new_sc->target_cpu] = new_sc;
cpumask_set_cpu(new_sc->target_cpu, &stor_device->alloced_cpus);
stor_device->stor_chns[device->channel->target_cpu] = device->channel;
cpumask_set_cpu(device->channel->target_cpu,
int vmbus_channel_set_cpu(struct vmbus_channel *channel, u32 target_cpu);
u32 target_cpu;
unsigned int target_cpu;
int panic_smp_redirect_cpu(int target_cpu, void *msg);
__field( int, target_cpu )
__entry->target_cpu = task_cpu(p);
__entry->target_cpu)
WRITE_ONCE(redir->target_cpu, cpumask_first(dest));
unsigned int target_cpu = READ_ONCE(desc->redirect.target_cpu);
irq_work_queue_on(&desc->redirect.work, target_cpu);
int __weak panic_smp_redirect_cpu(int target_cpu, void *msg)
return smp_call_function_single_async(target_cpu, &panic_csd);
int target_cpu, int curr_cpu)
.dst_cpu = target_cpu,
int migrate_task_to(struct task_struct *p, int target_cpu)
struct migration_arg arg = { p, target_cpu };
if (curr_cpu == target_cpu)
if (!cpumask_test_cpu(target_cpu, p->cpus_ptr))
trace_sched_move_numa(p, curr_cpu, target_cpu);
int target_cpu = busiest_rq->push_cpu;
struct rq *target_rq = cpu_rq(target_cpu);
if (!cpu_active(busiest_cpu) || !cpu_active(target_cpu))
for_each_domain(target_cpu, sd) {
.dst_cpu = target_cpu,
int target_cpu;
target_cpu = g->p.nr_cpus-1;
target_cpu = 0;
orig_mask = bind_to_cpu(target_cpu);
printf(" (injecting perturbalance, moved to CPU#%d)\n", target_cpu);
static cpu_set_t *bind_to_cpu(int target_cpu)
if (target_cpu == -1) {
if (target_cpu < 0 || target_cpu >= g->p.nr_cpus)
CPU_SET_S(target_cpu, size, mask);
static void guest_test_cpu_on(uint64_t target_cpu)
GUEST_ASSERT(!psci_cpu_on(target_cpu, CPU_ON_ENTRY_ADDR, CPU_ON_CONTEXT_ID));
target_state = psci_affinity_info(target_cpu, 0);
static uint64_t psci_cpu_on(uint64_t target_cpu, uint64_t entry_addr,
do_smccc(PSCI_0_2_FN64_CPU_ON, target_cpu, entry_addr, context_id,
tgt = target_cpu(this_cpu);