arch/arm64/kernel/sdei.c
129
s = scs_alloc(cpu_to_node(cpu));
arch/arm64/kernel/sdei.c
76
p = arch_alloc_vmap_stack(SDEI_STACK_SIZE, cpu_to_node(cpu));
arch/loongarch/kernel/cacheinfo.c
50
(cpu_to_node(i) != cpu_to_node(cpu)))
arch/loongarch/kernel/irq.c
118
page = alloc_pages_node(cpu_to_node(i), GFP_KERNEL, order);
arch/loongarch/kernel/numa.c
133
int nid = cpu_to_node(cpu);
arch/loongarch/kernel/numa.c
139
int nid = cpu_to_node(cpu);
arch/loongarch/kernel/smp.c
119
if (cpu_to_node(cpu) == cpu_to_node(i)) {
arch/loongarch/kernel/smp.c
131
if (cpu_to_node(cpu) == cpu_to_node(i)) {
arch/loongarch/kernel/smp.c
377
if (cpu_to_node(i) != 0)
arch/loongarch/kernel/vdso.c
51
vdso_k_arch_data->pdata[cpu].node = cpu_to_node(cpu);
arch/loongarch/mm/tlb.c
299
page = alloc_pages_node(cpu_to_node(cpu), GFP_ATOMIC, get_order(vec_sz));
arch/mips/include/asm/sn/agent.h
29
REMOTE_HUB_ADDR(cpu_to_node(_cpuid), \
arch/mips/mm/init.c
512
return node_distance(cpu_to_node(from), cpu_to_node(to));
arch/mips/mm/init.c
517
return cpu_to_node(cpu);
arch/mips/pci/pci-xtalk-bridge.c
315
data->nasid = cpu_to_node(cpu);
arch/mips/sgi-ip27/ip27-irq.c
147
info->nasid = cpu_to_node(hd->cpu);
arch/mips/sgi-ip27/ip27-irq.c
81
nasid = cpu_to_node(cpu);
arch/mips/sgi-ip27/ip27-smp.c
120
REMOTE_HUB_SEND_INTR(cpu_to_node(destid), irq);
arch/powerpc/kernel/smp.c
1071
cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1131
GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1133
GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1135
GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1138
GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1593
map_cpu_to_node(cpu, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
1604
ret = alloc_cpumask_var_node(&mask, GFP_ATOMIC, cpu_to_node(cpu));
arch/powerpc/kernel/smp.c
929
zalloc_cpumask_var_node(mask, GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/kernel/vdso.c
227
node = cpu_to_node(cpu);
arch/powerpc/kvm/book3s_hv.c
6566
int node = cpu_to_node(first_cpu);
arch/powerpc/lib/qspinlock.c
451
if (numa_node_id() != cpu_to_node(cpu))
arch/powerpc/mm/numa.c
779
if (cpu_to_node(sibling) != node) {
arch/powerpc/perf/imc-pmu.c
1208
int phys_id = cpu_to_node(cpu_id), rc = 0;
arch/powerpc/perf/imc-pmu.c
1557
nid = cpu_to_node(cpu);
arch/powerpc/perf/imc-pmu.c
387
nid = cpu_to_node(cpu);
arch/powerpc/perf/imc-pmu.c
428
l_cpumask = cpumask_of_node(cpu_to_node(cpu));
arch/powerpc/perf/imc-pmu.c
467
node_id = cpu_to_node(event->cpu);
arch/powerpc/perf/imc-pmu.c
563
node_id = cpu_to_node(event->cpu);
arch/powerpc/perf/imc-pmu.c
609
nid = cpu_to_node(cpu);
arch/powerpc/perf/imc-pmu.c
923
int nid = cpu_to_node(cpu_id);
arch/powerpc/perf/vpa-dtl.c
341
buf = kmem_cache_alloc_node(dtl_cache, GFP_KERNEL | GFP_ATOMIC, cpu_to_node(cpu));
arch/powerpc/perf/vpa-dtl.c
522
buf = kzalloc_node(sizeof(*buf), GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/platforms/cell/spufs/sched.c
298
node = cpu_to_node(raw_smp_processor_id());
arch/powerpc/platforms/cell/spufs/sched.c
579
node = cpu_to_node(raw_smp_processor_id());
arch/powerpc/platforms/cell/spufs/sched.c
627
node = cpu_to_node(raw_smp_processor_id());
arch/powerpc/platforms/powernv/setup.c
257
cpu_to_node(i));
arch/powerpc/platforms/pseries/dtl.c
198
buf = kmem_cache_alloc_node(dtl_cache, GFP_KERNEL, cpu_to_node(dtl->cpu));
arch/powerpc/sysdev/xive/common.c
1497
GFP_KERNEL, cpu_to_node(cpu));
arch/powerpc/sysdev/xive/common.c
1699
pages = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, alloc_order);
arch/riscv/kernel/irq.c
87
scs_alloc(cpu_to_node(cpu));
arch/riscv/kernel/irq.c
99
p = arch_alloc_vmap_stack(IRQ_STACK_SIZE, cpu_to_node(cpu));
arch/s390/include/asm/topology.h
83
#define cpu_to_node cpu_to_node
arch/sparc/kernel/cpumap.c
105
id = cpu_to_node(cpu);
arch/sparc/kernel/setup_64.c
607
node = cpu_to_node(i);
arch/sparc/kernel/smp_64.c
1520
if (cpu_to_node(from) == cpu_to_node(to))
arch/sparc/kernel/smp_64.c
1528
return cpu_to_node(cpu);
arch/x86/coco/sev/core.c
1194
data = memblock_alloc_node(sizeof(*data), PAGE_SIZE, cpu_to_node(cpu));
arch/x86/coco/sev/core.c
749
p = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL_ACCOUNT | __GFP_ZERO, 1);
arch/x86/events/amd/core.c
540
nb = kzalloc_node(sizeof(struct amd_nb), GFP_KERNEL, cpu_to_node(cpu));
arch/x86/events/amd/core.c
583
cpu_to_node(cpu));
arch/x86/events/amd/uncore.c
537
node = cpu_to_node(cpu);
arch/x86/events/intel/bts.c
87
int node = (cpu == -1) ? cpu : cpu_to_node(cpu);
arch/x86/events/intel/core.c
5578
GFP_KERNEL, cpu_to_node(cpu));
arch/x86/events/intel/core.c
5596
GFP_KERNEL, cpu_to_node(cpu));
arch/x86/events/intel/core.c
5618
cpuc->constraint_list = kzalloc_node(sz, GFP_KERNEL, cpu_to_node(cpu));
arch/x86/events/intel/ds.c
834
int node = cpu_to_node(cpu);
arch/x86/events/intel/ds.c
852
int max, node = cpu_to_node(cpu);
arch/x86/events/intel/lbr.c
656
cpu_to_node(cpu));
arch/x86/events/intel/pt.c
1345
node = cpu_to_node(cpu);
arch/x86/events/intel/pt.c
1386
int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu);
arch/x86/events/intel/pt.c
682
int node = cpu_to_node(cpu);
arch/x86/events/intel/uncore.c
1560
box = uncore_alloc_box(type, cpu_to_node(cpu));
arch/x86/events/intel/uncore.c
78
if (c->initialized && cpu_to_node(cpu) == node)
arch/x86/kernel/apic/io_apic.c
2056
int node = cpu_to_node(0);
arch/x86/kernel/apic/x2apic_cluster.c
181
int node = cpu_to_node(cpu);
arch/x86/kernel/apic/x2apic_uv_x.c
1781
uv_cpu_hub_info(cpu)->memory_nid = cpu_to_node(cpu);
arch/x86/kernel/cpu/intel.c
481
node = cpu_to_node(cpu);
arch/x86/kernel/cpu/resctrl/core.c
461
return cpu_to_node(cpu);
arch/x86/kernel/cpu/resctrl/core.c
500
hw_dom = kzalloc_node(sizeof(*hw_dom), GFP_KERNEL, cpu_to_node(cpu));
arch/x86/kernel/cpu/resctrl/core.c
534
hw_dom = kzalloc_node(sizeof(*hw_dom), GFP_KERNEL, cpu_to_node(cpu));
arch/x86/kernel/cpu/resctrl/intel_aet.c
393
d = kzalloc_node(sizeof(*d), GFP_KERNEL, cpu_to_node(cpu));
arch/x86/kernel/cpu/resctrl/monitor.c
133
return lrmid + (cpu_to_node(cpu) % snc_nodes_per_l3_cache) * r->mon.num_rmid;
arch/x86/kernel/espfix_64.c
159
node = cpu_to_node(cpu);
arch/x86/kernel/irq_32.c
109
int node = cpu_to_node(cpu);
arch/x86/kernel/kvm.c
706
GFP_KERNEL, cpu_to_node(cpu));
arch/x86/kernel/smpboot.c
1166
node = cpu_to_node(cpu);
arch/x86/kernel/smpboot.c
321
return (cpu_to_node(cpu1) == cpu_to_node(cpu2));
arch/x86/kernel/smpboot.c
332
cpu1, name, cpu2, cpu_to_node(cpu1), cpu_to_node(cpu2));
arch/x86/kvm/svm/svm.c
599
save_area_page = snp_safe_alloc_page_node(cpu_to_node(cpu), GFP_KERNEL);
arch/x86/kvm/vmx/vmx.c
3058
int node = cpu_to_node(cpu);
arch/x86/platform/uv/uv_nmi.c
1062
int nid = cpu_to_node(cpu);
arch/x86/platform/uv/uv_time.c
144
int nid = cpu_to_node(cpu);
block/blk-mq-cpumap.c
93
return cpu_to_node(i);
block/blk-mq.c
4112
hctx->numa_node = cpu_to_node(i);
crypto/scompress.c
83
int node = cpu_to_node(cpu);
drivers/acpi/acpi_processor.c
517
try_offline_node(cpu_to_node(pr->id));
drivers/base/arch_numa.c
65
int nid = cpu_to_node(cpu);
drivers/base/arch_topology.c
707
const cpumask_t *core_mask = cpumask_of_node(cpu_to_node(cpu));
drivers/base/arch_topology.c
970
cpuid_topo->package_id = cpu_to_node(cpuid);
drivers/base/cpu.c
420
cpu->node_id = cpu_to_node(num);
drivers/base/cpu.c
438
register_cpu_under_node(num, cpu_to_node(num));
drivers/base/cpu.c
56
from_nid = cpu_to_node(cpuid);
drivers/base/cpu.c
82
to_nid = cpu_to_node(cpuid);
drivers/base/cpu.c
99
unregister_cpu_under_node(logical_cpu, cpu_to_node(logical_cpu));
drivers/base/node.c
902
if (cpu_to_node(cpu) == nid)
drivers/base/test/test_async_driver_probe.c
113
nid = cpu_to_node(cpu);
drivers/base/test/test_async_driver_probe.c
147
nid = cpu_to_node(cpu);
drivers/base/test/test_async_driver_probe.c
175
nid = cpu_to_node(cpu);
drivers/block/mtip32xx/mtip32xx.c
3709
cpu_to_node(raw_smp_processor_id()), raw_smp_processor_id());
drivers/block/mtip32xx/mtip32xx.c
3766
cpu_to_node(dd->isr_binding), dd->isr_binding);
drivers/block/ublk_drv.c
4086
return cpu_to_node(cpu);
drivers/cpufreq/acpi-cpufreq.c
558
GFP_KERNEL, cpu_to_node(i))) {
drivers/crypto/hisilicon/hpre/hpre_main.c
457
int node = cpu_to_node(raw_smp_processor_id());
drivers/crypto/hisilicon/sec2/sec_main.c
417
int node = cpu_to_node(raw_smp_processor_id());
drivers/crypto/hisilicon/zip/zip_main.c
452
node = cpu_to_node(raw_smp_processor_id());
drivers/crypto/virtio/virtio_crypto_common.h
138
node = cpu_to_node(cpu);
drivers/edac/skx_common.c
282
if (c->initialized && cpu_to_node(cpu) == node) {
drivers/firmware/stratix10-svc.c
887
cpu_to_node(cpu),
drivers/hv/channel.c
179
page = alloc_pages_node(cpu_to_node(newchannel->target_cpu),
drivers/hv/hyperv_vmbus.h
503
cpumask_set_cpu(cpu, &hv_context.hv_numa_map[cpu_to_node(cpu)]);
drivers/hv/hyperv_vmbus.h
510
cpumask_clear_cpu(cpu, &hv_context.hv_numa_map[cpu_to_node(cpu)]);
drivers/hv/vmbus_drv.c
223
return sysfs_emit(buf, "%d\n", cpu_to_node(hv_dev->channel->target_cpu));
drivers/hwtracing/coresight/coresight-etb10.c
382
node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
drivers/hwtracing/coresight/coresight-etm-perf.c
159
int node = event->cpu == -1 ? -1 : cpu_to_node(event->cpu);
drivers/hwtracing/coresight/coresight-tmc-etf.c
428
node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1180
0, cpu_to_node(0), NULL);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1379
node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
drivers/hwtracing/coresight/coresight-tmc-etr.c
1495
node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
drivers/hwtracing/coresight/coresight-trbe.c
209
return cpu_to_node(event->cpu);
drivers/hwtracing/coresight/ultrasoc-smb.c
313
node = (event->cpu == -1) ? NUMA_NO_NODE : cpu_to_node(event->cpu);
drivers/infiniband/hw/hfi1/file_ops.c
958
numa = cpu_to_node(fd->rec_cpu_num);
drivers/infiniband/sw/siw/siw_main.c
145
cpumask_set_cpu(i, siw_cpu_info.tx_valid_cpus[cpu_to_node(i)]);
drivers/iommu/intel/dmar.c
2051
if (iommu->irq || iommu->node != cpu_to_node(cpu))
drivers/md/dm-ps-io-affinity.c
217
node = cpu_to_node(cpu);
drivers/md/dm-stats.c
367
p = dm_kvzalloc(percpu_alloc_size, cpu_to_node(cpu));
drivers/md/dm-stats.c
376
hi = dm_kvzalloc(s->histogram_alloc_size, cpu_to_node(cpu));
drivers/md/raid5.c
65
#define cpu_to_group(cpu) cpu_to_node(cpu)
drivers/misc/sgi-xp/xpc_uv.c
172
nid = cpu_to_node(cpu);
drivers/net/ethernet/amazon/ena/ena_netdev.c
1415
numa_node = cpu_to_node(cpu);
drivers/net/ethernet/amazon/ena/ena_netdev.c
247
node = cpu_to_node(ena_irq->cpu);
drivers/net/ethernet/amazon/ena/ena_netdev.c
384
node = cpu_to_node(ena_irq->cpu);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
95
node = cpu_to_node(cpu);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
640
cpu_node = cpu_to_node(cpu);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3934
cpu_node = cpu_to_node(cpu);
drivers/net/ethernet/fungible/funeth/funeth_tx.c
635
numa_node = cpu_to_node(qidx); /* XDP Tx queue */
drivers/net/ethernet/huawei/hinic/hinic_main.c
397
node = cpu_to_node(i);
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
852
node = cpu_to_node(cpu);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2165
node = cpu_to_node(i % num_online_cpus());
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2179
node = cpu_to_node(i % num_online_cpus());
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
613
.node = cpu_to_node(c->cpu),
drivers/net/ethernet/mellanox/mlx5/core/en/trap.c
139
t = kvzalloc_node(sizeof(*t), GFP_KERNEL, cpu_to_node(cpu));
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/setup.c
105
err = mlx5e_open_rq(params, rq_params, xsk, cpu_to_node(c->cpu), q_counter, xskrq);
drivers/net/ethernet/mellanox/mlx5/core/en_dim.c
73
dim = kvzalloc_node(sizeof(*dim), GFP_KERNEL, cpu_to_node(cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1509
param->wq.db_numa_node = cpu_to_node(c->cpu);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1515
err = mlx5e_alloc_xdpsq_db(sq, cpu_to_node(c->cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1583
param->wq.db_numa_node = cpu_to_node(c->cpu);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1589
err = mlx5e_alloc_icosq_db(sq, cpu_to_node(c->cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1677
param->wq.db_numa_node = cpu_to_node(c->cpu);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1683
err = mlx5e_alloc_txqsq_db(sq, cpu_to_node(c->cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2104
xdpsq = kvzalloc_node(sizeof(*xdpsq), GFP_KERNEL, cpu_to_node(c->cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2519
return mlx5e_open_rq(params, rq_params, NULL, cpu_to_node(c->cpu), q_counter, &c->rq);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2533
cpu_to_node(c->cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2712
GFP_KERNEL, cpu_to_node(cpu));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2793
c = kvzalloc_node(sizeof(*c), GFP_KERNEL, cpu_to_node(cpu));
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2111
q_vector->numa_node = cpu_to_node(v_idx);
drivers/net/hyperv/netvsc.c
324
int node = cpu_to_node(nvchan->channel->target_cpu);
drivers/nvdimm/nd_perf.c
156
nodeid = cpu_to_node(cpu);
drivers/perf/alibaba_uncore_drw_pmu.c
747
target = cpumask_any_and_but(cpumask_of_node(cpu_to_node(cpu)),
drivers/perf/arm-cmn.c
2027
if (cpu_to_node(cmn->cpu) != node && cpu_to_node(cpu) == node)
drivers/perf/arm-ni.c
807
if (cpu_to_node(ni->cpu) != node && cpu_to_node(cpu) == node)
drivers/perf/arm_cspmu/nvidia_cspmu.c
376
const int socket = cpu_to_node(cpu);
drivers/perf/arm_spe_pmu.c
1019
buf = kzalloc_node(sizeof(*buf), GFP_KERNEL, cpu_to_node(cpu));
drivers/perf/fujitsu_uncore_pmu.c
418
if (cpu_to_node(uncorepmu->cpu) != node && cpu_to_node(cpu) == node)
drivers/perf/thunderx2_pmu.c
921
(tx2_pmu->node == cpu_to_node(cpu)))
drivers/platform/x86/intel/speed_select_if/isst_if_common.c
389
isst_cpu_info[cpu].numa_node = cpu_to_node(cpu);
drivers/platform/x86/intel_scu_wdt.c
38
ioapic_set_alloc_attr(&info, cpu_to_node(0), 1, 0);
drivers/resctrl/mpam_devices.c
463
if (node_id == cpu_to_node(cpu))
drivers/scsi/lpfc/lpfc_attr.c
5654
if (cpu_to_node(cpu) == numa_node)
drivers/scsi/lpfc/lpfc_init.c
10729
cpu_to_node(cpu));
drivers/scsi/lpfc/lpfc_sli.c
15899
GFP_KERNEL, cpu_to_node(cpu));
drivers/scsi/lpfc/lpfc_sli.c
22418
cpu_to_node(hdwq->io_wq->chann));
drivers/scsi/lpfc/lpfc_sli.c
22561
cpu_to_node(hdwq->io_wq->chann));
drivers/scsi/smartpqi/smartpqi_init.c
9320
node = cpu_to_node(0);
drivers/scsi/storvsc_drv.c
1423
node_mask = cpumask_of_node(cpu_to_node(q_num));
drivers/scsi/storvsc_drv.c
1485
node_mask = cpumask_of_node(cpu_to_node(q_num));
drivers/virt/nitro_enclaves/ne_misc_dev.c
229
numa_node = cpu_to_node(cpu);
drivers/virt/nitro_enclaves/ne_misc_dev.c
239
if (numa_node != cpu_to_node(cpu)) {
drivers/watchdog/octeon-wdt-main.c
134
int node = cpu_to_node(cpu);
drivers/watchdog/octeon-wdt-main.c
309
node = cpu_to_node(cpu);
drivers/watchdog/octeon-wdt-main.c
333
node = cpu_to_node(cpu);
drivers/watchdog/octeon-wdt-main.c
356
node = cpu_to_node(cpu);
drivers/watchdog/octeon-wdt-main.c
412
node = cpu_to_node(cpu);
drivers/watchdog/octeon-wdt-main.c
469
node = cpu_to_node(cpu);
drivers/xen/xen-acpi-processor.c
537
GFP_KERNEL, cpu_to_node(i))) {
fs/resctrl/rdtgroup.c
2493
GFP_KERNEL, cpu_to_node(cpu));
include/asm-generic/topology.h
34
#ifndef cpu_to_node
include/linux/mm.h
2030
return cpu_to_node(cpupid_to_cpu(cpupid));
include/linux/rseq_entry.h
360
cpu_to_node(t->rseq.ids.cpu_id) : 0;
include/linux/rseq_entry.h
636
u32 node_id = cpu_to_node(ids.cpu_id);
include/linux/sched/topology.h
244
return cpu_to_node(task_cpu(p));
include/linux/topology.h
117
return cpu_to_node(raw_smp_processor_id());
include/linux/topology.h
174
return cpu_to_node(cpu);
include/linux/topology.h
265
return cpumask_of_node(cpu_to_node(cpu));
include/linux/topology.h
90
#ifndef cpu_to_node
include/trace/events/rseq.h
25
__entry->node_id = cpu_to_node(__entry->cpu_id);
include/trace/events/sched.h
657
__entry->src_nid = cpu_to_node(src_cpu);
include/trace/events/sched.h
659
__entry->dst_nid = cpu_to_node(dst_cpu);
include/trace/events/sched.h
693
__entry->src_nid = cpu_to_node(src_cpu);
include/trace/events/sched.h
698
__entry->dst_nid = dst_cpu >= 0 ? cpu_to_node(dst_cpu) : -1;
init/main.c
902
#ifndef cpu_to_node
kernel/bpf/cpumap.c
441
numa = cpu_to_node(cpu);
kernel/bpf/memalloc.c
503
alloc_bulk(c, cnt, cpu_to_node(cpu), false);
kernel/cpu.c
1691
err = try_online_node(cpu_to_node(cpu));
kernel/cpu.c
2710
if (cpu_online(cpu) || !node_online(cpu_to_node(cpu)))
kernel/events/callchain.c
93
cpu_to_node(cpu));
kernel/events/core.c
1106
GFP_KERNEL, cpu_to_node(cpu));
kernel/events/core.c
11744
int node = cpu_to_node(event->cpu == -1 ? 0 : event->cpu);
kernel/events/core.c
12804
cpu_to_node(cpu));
kernel/events/core.c
13288
node = (cpu >= 0) ? cpu_to_node(cpu) : -1;
kernel/events/core.c
15067
hlist = kzalloc_node(sizeof(*hlist), GFP_KERNEL, cpu_to_node(cpu));
kernel/events/ring_buffer.c
681
int node = (event->cpu == -1) ? -1 : cpu_to_node(event->cpu);
kernel/events/ring_buffer.c
821
node = (cpu == -1) ? cpu : cpu_to_node(cpu);
kernel/events/ring_buffer.c
848
node = (cpu == -1) ? cpu : cpu_to_node(cpu);
kernel/events/ring_buffer.c
928
node = (cpu == -1) ? cpu : cpu_to_node(cpu);
kernel/fork.c
2574
task = copy_process(&init_struct_pid, 0, cpu_to_node(cpu), &args);
kernel/irq/irqdesc.c
523
node = cpu_to_node(cpumask_first(mask));
kernel/kcov.c
1126
sizeof(unsigned long), cpu_to_node(cpu));
kernel/kthread.c
1142
worker = kthread_create_worker_on_node(flags, cpu_to_node(cpu), namefmt, cpu);
kernel/kthread.c
625
p = kthread_create_on_node(threadfn, data, cpu_to_node(cpu), namefmt,
kernel/rseq.c
304
node_id = cpu_to_node(ids.cpu_id);
kernel/sched/core.c
3449
int nid = cpu_to_node(cpu);
kernel/sched/core.c
8755
zalloc_cpumask_var_node(&rq->scratch_mask, GFP_KERNEL, cpu_to_node(i));
kernel/sched/deadline.c
3152
GFP_KERNEL, cpu_to_node(i));
kernel/sched/ext.c
1963
int node = cpu_to_node(cpu_of(rq));
kernel/sched/ext.c
251
return sch->global_dsqs[cpu_to_node(task_cpu(p))];
kernel/sched/ext.c
4905
GFP_KERNEL, cpu_to_node(cpu));
kernel/sched/ext.c
5851
int n = cpu_to_node(cpu);
kernel/sched/ext_idle.c
682
GFP_KERNEL, cpu_to_node(i)));
kernel/sched/ext_idle.c
684
GFP_KERNEL, cpu_to_node(i)));
kernel/sched/ext_idle.c
686
GFP_KERNEL, cpu_to_node(i)));
kernel/sched/ext_idle.c
74
return cpu_to_node(cpu);
kernel/sched/ext_idle.c
965
return cpu_to_node(cpu);
kernel/sched/fair.c
11017
if (cpu_to_node(this_cpu) == p->numa_preferred_nid)
kernel/sched/fair.c
11021
if (cpu_to_node(idlest_cpu) == p->numa_preferred_nid)
kernel/sched/fair.c
13708
GFP_KERNEL, cpu_to_node(i));
kernel/sched/fair.c
13713
GFP_KERNEL, cpu_to_node(i));
kernel/sched/fair.c
14038
zalloc_cpumask_var_node(&per_cpu(load_balance_mask, i), GFP_KERNEL, cpu_to_node(i));
kernel/sched/fair.c
14039
zalloc_cpumask_var_node(&per_cpu(select_rq_mask, i), GFP_KERNEL, cpu_to_node(i));
kernel/sched/fair.c
14041
GFP_KERNEL, cpu_to_node(i));
kernel/sched/fair.c
1986
int dst_nid = cpu_to_node(dst_cpu);
kernel/sched/fair.c
2667
nid = cpu_to_node(env.best_cpu);
kernel/sched/fair.c
3702
int src_nid = cpu_to_node(task_cpu(p));
kernel/sched/fair.c
3703
int dst_nid = cpu_to_node(new_cpu);
kernel/sched/fair.c
9414
src_nid = cpu_to_node(env->src_cpu);
kernel/sched/fair.c
9415
dst_nid = cpu_to_node(env->dst_cpu);
kernel/sched/rt.c
2413
GFP_KERNEL, cpu_to_node(i));
kernel/sched/rt.c
273
GFP_KERNEL, cpu_to_node(i));
kernel/sched/rt.c
278
GFP_KERNEL, cpu_to_node(i));
kernel/sched/topology.c
1829
return sched_domains_numa_masks[tl->numa_level][cpu_to_node(cpu)];
kernel/sched/topology.c
2183
node = cpu_to_node(cpu);
kernel/sched/topology.c
2197
int node = cpu_to_node(cpu);
kernel/sched/topology.c
2235
int i, j = cpu_to_node(cpu), found = nr_cpu_ids;
kernel/sched/topology.c
2394
GFP_KERNEL, cpu_to_node(j));
kernel/sched/topology.c
2401
GFP_KERNEL, cpu_to_node(j));
kernel/sched/topology.c
2408
GFP_KERNEL, cpu_to_node(j));
kernel/sched/topology.c
2417
GFP_KERNEL, cpu_to_node(j));
kernel/sched/topology.c
959
GFP_KERNEL, cpu_to_node(cpu));
kernel/smp.c
59
cpu_to_node(cpu)))
kernel/smp.c
62
cpu_to_node(cpu))) {
kernel/smp.c
763
cpu = sched_numa_find_nth_cpu(mask, 0, cpu_to_node(cpu));
kernel/smpboot.c
174
td = kzalloc_node(sizeof(*td), GFP_KERNEL, cpu_to_node(cpu));
kernel/sys.c
2922
err |= put_user(cpu_to_node(cpu), nodep);
kernel/taskstats.c
314
GFP_KERNEL, cpu_to_node(cpu));
kernel/time/timer_migration.c
1898
int node = cpu_to_node(cpu);
kernel/trace/ring_buffer.c
411
cache_line_size()), GFP_KERNEL, cpu_to_node(cpu));
kernel/trace/ring_buffer.c
415
cache_line_size()), GFP_KERNEL, cpu_to_node(cpu));
kernel/trace/ring_buffer.c
430
page = alloc_pages_node(cpu_to_node(cpu), mflags, order);
kernel/trace/trace.c
2246
page = alloc_pages_node(cpu_to_node(cpu),
kernel/trace/trace.c
6601
cpu_to_node(cpu));
kernel/trace/trace_uprobe.c
888
struct page *p = alloc_pages_node(cpu_to_node(cpu),
kernel/workqueue.c
1624
if (off_cpu >= 0 && cpu_to_node(off_cpu) == node)
kernel/workqueue.c
2433
if (node == cpu_to_node(cpu))
kernel/workqueue.c
7846
pool->node = cpu_to_node(cpu);
kernel/workqueue.c
8036
pool->node = cpu_to_node(cpu);
kernel/workqueue.c
8038
pool->node = cpu_to_node(cpu);
kernel/workqueue.c
8112
pt->pod_node[pt->cpu_pod[cpu]] = cpu_to_node(cpu);
kernel/workqueue.c
8132
return cpu_to_node(cpu0) == cpu_to_node(cpu1);
lib/cpu_rmap.c
195
cpumask_of_node(cpu_to_node(cpu)));
lib/cpu_rmap.c
209
cpumask_of_node(cpu_to_node(cpu)), 3))
lib/group_cpus.c
82
cpumask_set_cpu(cpu, masks[cpu_to_node(cpu)]);
lib/objpool.c
85
cpu_to_node(i), __builtin_return_address(0));
lib/objpool.c
88
slot = kmalloc_node(size, pool->gfp, cpu_to_node(i));
mm/memory_hotplug.c
2152
if (cpu_to_node(cpu) == nid)
mm/page_alloc.c
5848
set_cpu_numa_mem(cpu, local_memory_node(cpu_to_node(cpu)));
mm/percpu-vm.c
95
*pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0);
mm/slub.c
4239
cpu, cpu_to_node(cpu), nid, gfpflags, &gfpflags);
mm/vmstat.c
2190
if (!node_state(cpu_to_node(cpu), N_CPU)) {
mm/vmstat.c
2191
node_set_state(cpu_to_node(cpu), N_CPU);
mm/vmstat.c
2209
node = cpu_to_node(cpu);
mm/zswap.c
744
buffer = kmalloc_node(PAGE_SIZE, GFP_KERNEL, cpu_to_node(cpu));
mm/zswap.c
750
acomp = crypto_alloc_acomp_node(pool->tfm_name, 0, 0, cpu_to_node(cpu));
net/bridge/netfilter/ebtables.c
931
cpu_to_node(i));
net/core/dev.c
2817
cpu_to_node(attr_index));
net/core/dev.c
2955
numa_node_id = cpu_to_node(j);
net/core/dev.c
2956
else if (numa_node_id != cpu_to_node(j))
net/core/pktgen.c
3831
int node = cpu_to_node(t->cpu);
net/core/pktgen.c
3922
cpu_to_node(cpu));
net/core/sysctl_net_core.c
237
cpu_to_node(i));
net/ipv4/tcp_sigpool.c
76
scratch = kmalloc_node(size, GFP_KERNEL, cpu_to_node(cpu));
net/iucv/iucv.c
624
GFP_KERNEL|GFP_DMA, cpu_to_node(cpu));
net/iucv/iucv.c
630
GFP_KERNEL|GFP_DMA, cpu_to_node(cpu));
net/iucv/iucv.c
635
GFP_KERNEL|GFP_DMA, cpu_to_node(cpu));
net/netfilter/nft_set_pipapo.c
1223
GFP_KERNEL_ACCOUNT, cpu_to_node(i));
net/netfilter/x_tables.c
1380
cpu_to_node(cpu));
net/sunrpc/svc.c
351
return cpu_to_node(m->pool_to[pidx]);
net/sunrpc/svc.c
414
pidx = m->to_pool[cpu_to_node(cpu)];
virt/kvm/kvm_main.c
6507
GFP_KERNEL, cpu_to_node(cpu))) {