Symbol: next_cpu
arch/parisc/kernel/irq.c
324
static int next_cpu = -1;
arch/parisc/kernel/irq.c
326
next_cpu++; /* assign to "next" CPU we want this bugger on */
arch/parisc/kernel/irq.c
329
while ((next_cpu < nr_cpu_ids) &&
arch/parisc/kernel/irq.c
330
(!per_cpu(cpu_data, next_cpu).txn_addr ||
arch/parisc/kernel/irq.c
331
!cpu_online(next_cpu)))
arch/parisc/kernel/irq.c
332
next_cpu++;
arch/parisc/kernel/irq.c
334
if (next_cpu >= nr_cpu_ids)
arch/parisc/kernel/irq.c
335
next_cpu = 0; /* nothing else, assign monarch */
arch/parisc/kernel/irq.c
337
return txn_affinity_addr(virt_irq, next_cpu);
arch/powerpc/lib/qspinlock.c
688
int next_cpu = next->cpu;
arch/powerpc/lib/qspinlock.c
692
if (vcpu_is_preempted(next_cpu))
arch/powerpc/lib/qspinlock.c
693
prod_cpu(next_cpu);
arch/powerpc/mm/book3s64/hash_utils.c
1296
int next_cpu;
arch/powerpc/mm/book3s64/hash_utils.c
1302
next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask);
arch/powerpc/mm/book3s64/hash_utils.c
1303
if (next_cpu >= nr_cpu_ids)
arch/powerpc/mm/book3s64/hash_utils.c
1304
next_cpu = cpumask_first(cpu_online_mask);
arch/powerpc/mm/book3s64/hash_utils.c
1306
add_timer_on(&stress_hpt_timer, next_cpu);
arch/x86/kernel/tsc_sync.c
102
next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask);
arch/x86/kernel/tsc_sync.c
103
if (next_cpu >= nr_cpu_ids)
arch/x86/kernel/tsc_sync.c
104
next_cpu = cpumask_first(cpu_online_mask);
arch/x86/kernel/tsc_sync.c
107
add_timer_on(&tsc_sync_check_timer, next_cpu);
arch/x86/kernel/tsc_sync.c
97
int next_cpu;
arch/x86/platform/uv/uv_time.c
159
head->next_cpu = -1;
arch/x86/platform/uv/uv_time.c
176
head->next_cpu = -1;
arch/x86/platform/uv/uv_time.c
185
head->next_cpu = bcpu;
arch/x86/platform/uv/uv_time.c
209
int next_cpu;
arch/x86/platform/uv/uv_time.c
213
next_cpu = head->next_cpu;
arch/x86/platform/uv/uv_time.c
217
if (next_cpu < 0 || bcpu == next_cpu ||
arch/x86/platform/uv/uv_time.c
218
expires < head->cpu[next_cpu].expires) {
arch/x86/platform/uv/uv_time.c
219
head->next_cpu = bcpu;
arch/x86/platform/uv/uv_time.c
249
if ((head->next_cpu == bcpu && uv_read_rtc(NULL) >= *t) || force)
arch/x86/platform/uv/uv_time.c
255
if (head->next_cpu == bcpu)
arch/x86/platform/uv/uv_time.c
50
int next_cpu;
block/blk-mq.c
2259
return hctx->next_cpu >= nr_cpu_ids;
block/blk-mq.c
2271
int next_cpu = hctx->next_cpu;
block/blk-mq.c
2279
next_cpu = cpumask_next_and(next_cpu, hctx->cpumask,
block/blk-mq.c
2281
if (next_cpu >= nr_cpu_ids)
block/blk-mq.c
2282
next_cpu = blk_mq_first_mapped_cpu(hctx);
block/blk-mq.c
2290
if (!cpu_online(next_cpu)) {
block/blk-mq.c
2300
hctx->next_cpu = next_cpu;
block/blk-mq.c
2305
hctx->next_cpu = next_cpu;
block/blk-mq.c
2306
return next_cpu;
block/blk-mq.c
4305
hctx->next_cpu = blk_mq_first_mapped_cpu(hctx);
drivers/irqchip/irq-gic-v3.c
1334
int next_cpu, cpu = *base_cpu;
drivers/irqchip/irq-gic-v3.c
1343
next_cpu = cpumask_next(cpu, mask);
drivers/irqchip/irq-gic-v3.c
1344
if (next_cpu >= nr_cpu_ids)
drivers/irqchip/irq-gic-v3.c
1346
cpu = next_cpu;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2486
u32 next_cpu = desc->txd2;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2509
cpu = next_cpu;
include/linux/blk-mq.h
351
int next_cpu;
kernel/sched/fair.c
14116
goto next_cpu;
kernel/sched/fair.c
14136
next_cpu:
kernel/time/tick-broadcast.c
697
int cpu, next_cpu = 0;
kernel/time/tick-broadcast.c
727
next_cpu = cpu;
kernel/time/tick-broadcast.c
764
tick_broadcast_set_event(dev, next_cpu, next_event);
kernel/trace/trace.c
2384
int next_cpu = -1;
kernel/trace/trace.c
2414
next_cpu = cpu;
kernel/trace/trace.c
2424
*ent_cpu = next_cpu;
kernel/trace/trace_entries.h
173
__field( unsigned int, next_cpu ) \
kernel/trace/trace_entries.h
190
__entry->next_cpu)
kernel/trace/trace_entries.h
208
__entry->next_cpu)
kernel/trace/trace_hwlat.c
315
int next_cpu;
kernel/trace/trace_hwlat.c
327
next_cpu = cpumask_next_wrap(raw_smp_processor_id(), current_mask);
kernel/trace/trace_hwlat.c
330
if (next_cpu >= nr_cpu_ids) /* Shouldn't happen! */
kernel/trace/trace_hwlat.c
334
cpumask_set_cpu(next_cpu, current_mask);
kernel/trace/trace_hwlat.c
420
int next_cpu;
kernel/trace/trace_hwlat.c
437
next_cpu = cpumask_first(current_mask);
kernel/trace/trace_hwlat.c
439
cpumask_set_cpu(next_cpu, current_mask);
kernel/trace/trace_output.c
1293
field->next_cpu,
kernel/trace/trace_output.c
1327
field->next_cpu,
kernel/trace/trace_output.c
1363
SEQ_PUT_HEX_FIELD(s, field->next_cpu);
kernel/trace/trace_output.c
1394
SEQ_PUT_FIELD(s, field->next_cpu);
kernel/trace/trace_sched_wakeup.c
408
entry->next_cpu = task_cpu(next);
kernel/trace/trace_sched_wakeup.c
434
entry->next_cpu = task_cpu(wakee);
kernel/watchdog_buddy.c
102
watchdog_hardlockup_check(next_cpu, NULL);
kernel/watchdog_buddy.c
13
unsigned int next_cpu;
kernel/watchdog_buddy.c
15
next_cpu = cpumask_next_wrap(cpu, &watchdog_cpus);
kernel/watchdog_buddy.c
16
if (next_cpu == cpu)
kernel/watchdog_buddy.c
19
return next_cpu;
kernel/watchdog_buddy.c
30
unsigned int next_cpu;
kernel/watchdog_buddy.c
48
next_cpu = watchdog_next_cpu(cpu);
kernel/watchdog_buddy.c
49
if (next_cpu < nr_cpu_ids)
kernel/watchdog_buddy.c
50
watchdog_hardlockup_touch_cpu(next_cpu);
kernel/watchdog_buddy.c
64
unsigned int next_cpu = watchdog_next_cpu(cpu);
kernel/watchdog_buddy.c
73
if (next_cpu < nr_cpu_ids)
kernel/watchdog_buddy.c
74
watchdog_hardlockup_touch_cpu(next_cpu);
kernel/watchdog_buddy.c
88
unsigned int next_cpu;
kernel/watchdog_buddy.c
91
next_cpu = watchdog_next_cpu(smp_processor_id());
kernel/watchdog_buddy.c
92
if (next_cpu >= nr_cpu_ids)
net/core/dev.c
5037
struct rps_dev_flow *rflow, u16 next_cpu, u32 hash)
net/core/dev.c
5039
if (next_cpu < nr_cpu_ids) {
net/core/dev.c
5056
rxq_index = cpu_rmap_lookup_index(dev->rx_cpu_rmap, next_cpu);
net/core/dev.c
5075
next_cpu == tmp_cpu)
net/core/dev.c
5094
head = READ_ONCE(per_cpu(softnet_data, next_cpu).input_queue_head);
net/core/dev.c
5098
WRITE_ONCE(rflow->cpu, next_cpu);
net/core/dev.c
5147
u32 next_cpu;
net/core/dev.c
5160
next_cpu = ident & net_hotdata.rps_cpu_mask;
net/core/dev.c
5181
if (unlikely(tcpu != next_cpu) &&
net/core/dev.c
5185
tcpu = next_cpu;
net/core/dev.c
5186
rflow = set_rps_cpu(dev, skb, rflow, next_cpu, hash);
tools/testing/selftests/bpf/bench.c
477
for (i = cpu_set->next_cpu; i < cpu_set->cpus_len; i++) {
tools/testing/selftests/bpf/bench.c
479
cpu_set->next_cpu = i + 1;
tools/testing/selftests/bpf/bench.c
487
return cpu_set->next_cpu++ % env.nr_cpus;
tools/testing/selftests/bpf/bench.c
711
next_cpu(&env.cons_cpus));
tools/testing/selftests/bpf/bench.c
718
env.prod_cpus.next_cpu = env.cons_cpus.next_cpu;
tools/testing/selftests/bpf/bench.c
734
next_cpu(&env.prod_cpus));
tools/testing/selftests/bpf/bench.h
18
int next_cpu;
tools/testing/selftests/bpf/test_lru_map.c
165
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
170
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
258
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
267
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
327
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
336
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
432
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
441
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
492
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
497
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
567
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
583
while (sched_next_online(0, &next_cpu) != -1) {
tools/testing/selftests/bpf/test_lru_map.c
616
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
624
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
682
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
687
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/bpf/test_lru_map.c
773
int next_cpu = 0;
tools/testing/selftests/bpf/test_lru_map.c
778
assert(sched_next_online(0, &next_cpu) != -1);
tools/testing/selftests/kvm/rseq_test.c
82
for (i = 0, cpu = min_cpu; i < NR_TASK_MIGRATIONS; i++, cpu = next_cpu(cpu)) {