next_cpu
static int next_cpu = -1;
next_cpu++; /* assign to "next" CPU we want this bugger on */
while ((next_cpu < nr_cpu_ids) &&
(!per_cpu(cpu_data, next_cpu).txn_addr ||
!cpu_online(next_cpu)))
next_cpu++;
if (next_cpu >= nr_cpu_ids)
next_cpu = 0; /* nothing else, assign monarch */
return txn_affinity_addr(virt_irq, next_cpu);
int next_cpu = next->cpu;
if (vcpu_is_preempted(next_cpu))
prod_cpu(next_cpu);
int next_cpu;
next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask);
if (next_cpu >= nr_cpu_ids)
next_cpu = cpumask_first(cpu_online_mask);
add_timer_on(&stress_hpt_timer, next_cpu);
next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask);
if (next_cpu >= nr_cpu_ids)
next_cpu = cpumask_first(cpu_online_mask);
add_timer_on(&tsc_sync_check_timer, next_cpu);
int next_cpu;
head->next_cpu = -1;
head->next_cpu = -1;
head->next_cpu = bcpu;
int next_cpu;
next_cpu = head->next_cpu;
if (next_cpu < 0 || bcpu == next_cpu ||
expires < head->cpu[next_cpu].expires) {
head->next_cpu = bcpu;
if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force)
if (head->next_cpu == bcpu)
int next_cpu;
return hctx->next_cpu >= nr_cpu_ids;
int next_cpu = hctx->next_cpu;
next_cpu = cpumask_next_and(next_cpu, hctx->cpumask,
if (next_cpu >= nr_cpu_ids)
next_cpu = blk_mq_first_mapped_cpu(hctx);
if (!cpu_online(next_cpu)) {
hctx->next_cpu = next_cpu;
hctx->next_cpu = next_cpu;
return next_cpu;
hctx->next_cpu = blk_mq_first_mapped_cpu(hctx);
int next_cpu, cpu = *base_cpu;
next_cpu = cpumask_next(cpu, mask);
if (next_cpu >= nr_cpu_ids)
cpu = next_cpu;
u32 next_cpu = desc->txd2;
cpu = next_cpu;
int next_cpu;
goto next_cpu;
next_cpu:
int cpu, next_cpu = 0;
next_cpu = cpu;
tick_broadcast_set_event(dev, next_cpu, next_event);
int next_cpu = -1;
next_cpu = cpu;
*ent_cpu = next_cpu;
__field( unsigned int, next_cpu ) \
__entry->next_cpu)
__entry->next_cpu)
int next_cpu;
next_cpu = cpumask_next_wrap(raw_smp_processor_id(), current_mask);
if (next_cpu >= nr_cpu_ids) /* Shouldn't happen! */
cpumask_set_cpu(next_cpu, current_mask);
int next_cpu;
next_cpu = cpumask_first(current_mask);
cpumask_set_cpu(next_cpu, current_mask);
field->next_cpu,
field->next_cpu,
SEQ_PUT_HEX_FIELD(s, field->next_cpu);
SEQ_PUT_FIELD(s, field->next_cpu);
entry->next_cpu = task_cpu(next);
entry->next_cpu = task_cpu(wakee);
watchdog_hardlockup_check(next_cpu, NULL);
unsigned int next_cpu;
next_cpu = cpumask_next_wrap(cpu, &watchdog_cpus);
if (next_cpu == cpu)
return next_cpu;
unsigned int next_cpu;
next_cpu = watchdog_next_cpu(cpu);
if (next_cpu < nr_cpu_ids)
watchdog_hardlockup_touch_cpu(next_cpu);
unsigned int next_cpu = watchdog_next_cpu(cpu);
if (next_cpu < nr_cpu_ids)
watchdog_hardlockup_touch_cpu(next_cpu);
unsigned int next_cpu;
next_cpu = watchdog_next_cpu(smp_processor_id());
if (next_cpu >= nr_cpu_ids)
struct rps_dev_flow *rflow, u16 next_cpu, u32 hash)
if (next_cpu < nr_cpu_ids) {
rxq_index = cpu_rmap_lookup_index(dev->rx_cpu_rmap, next_cpu);
next_cpu == tmp_cpu)
head = READ_ONCE(per_cpu(softnet_data, next_cpu).input_queue_head);
WRITE_ONCE(rflow->cpu, next_cpu);
u32 next_cpu;
next_cpu = ident & net_hotdata.rps_cpu_mask;
if (unlikely(tcpu != next_cpu) &&
tcpu = next_cpu;
rflow = set_rps_cpu(dev, skb, rflow, next_cpu, hash);
for (i = cpu_set->next_cpu; i < cpu_set->cpus_len; i++) {
cpu_set->next_cpu = i + 1;
return cpu_set->next_cpu++ % env.nr_cpus;
next_cpu(&env.cons_cpus));
env.prod_cpus.next_cpu = env.cons_cpus.next_cpu;
next_cpu(&env.prod_cpus));
int next_cpu;
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
while (sched_next_online(0, &next_cpu) != -1) {
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
int next_cpu = 0;
assert(sched_next_online(0, &next_cpu) != -1);
for (i = 0, cpu = min_cpu; i < NR_TASK_MIGRATIONS; i++, cpu = next_cpu(cpu)) {