Symbol: numa_node_id
arch/arm/mach-imx/avic.c
200
irq_base = irq_alloc_descs(-1, 0, AVIC_NUM_IRQS, numa_node_id());
arch/arm/mach-imx/tzic.c
175
irq_base = irq_alloc_descs(-1, 0, TZIC_NUM_IRQS, numa_node_id());
arch/mips/mti-malta/malta-int.c
180
16, numa_node_id()) < 0,
arch/powerpc/lib/qspinlock.c
451
if (numa_node_id() != cpu_to_node(cpu))
arch/s390/include/asm/topology.h
100
#define numa_node_id numa_node_id
arch/sh/boards/mach-se/7724/irq.c
125
SE7724_FPGA_IRQ_NR, numa_node_id());
arch/sh/kernel/cpu/irq/ipr.c
66
res = irq_alloc_desc_at(p->irq, numa_node_id());
arch/sh/kernel/topology.c
68
if (i != numa_node_id())
arch/sparc/kernel/irq_64.c
245
irq = __irq_alloc_descs(-1, 1, 1, numa_node_id(), NULL, NULL);
arch/sparc/mm/tsb.c
425
gfp_flags, numa_node_id());
arch/um/drivers/virt-pci.c
488
dev->irq = irq_alloc_desc(numa_node_id());
arch/x86/include/asm/topology.h
92
#define numa_node_id numa_node_id
arch/x86/kernel/cpu/sgx/main.c
480
int nid_of_current = numa_node_id();
arch/x86/kvm/svm/svm.h
884
return snp_safe_alloc_page_node(numa_node_id(), GFP_KERNEL_ACCOUNT);
arch/x86/kvm/svm/svm.h
910
return snp_safe_alloc_page_node(numa_node_id(), GFP_KERNEL_ACCOUNT);
drivers/base/test/test_async_driver_probe.c
48
dev_to_node(dev) != numa_node_id()) {
drivers/base/test/test_async_driver_probe.c
50
dev_to_node(dev), numa_node_id());
drivers/crypto/intel/qat/qat_common/qat_algs.c
532
int node = numa_node_id();
drivers/crypto/intel/qat/qat_common/qat_algs.c
817
int node = numa_node_id();
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1281
qat_crypto_get_instance_node(numa_node_id());
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
530
qat_crypto_get_instance_node(numa_node_id());
drivers/crypto/intel/qat/qat_common/qat_comp_algs.c
139
node = numa_node_id();
drivers/gpio/gpio-mxc.c
506
irq_base = devm_irq_alloc_descs(&pdev->dev, -1, 0, 32, numa_node_id());
drivers/gpio/gpio-mxs.c
302
irq_base = devm_irq_alloc_descs(&pdev->dev, -1, 0, 32, numa_node_id());
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1748
static int kfd_fill_cu_for_cpu(int numa_node_id, int *avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1765
cpumask = cpumask_of_node(numa_node_id);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1770
sub_type_hdr->processor_id_low = kfd_numa_node_to_apic_id(numa_node_id);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1787
static int kfd_fill_mem_info_for_cpu(int numa_node_id, int *avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1812
pgdat = NODE_DATA(numa_node_id);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1825
static int kfd_fill_iolink_info_for_cpu(int numa_node_id, int *avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1842
if (nid == numa_node_id) /* node itself */
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1857
sub_type_hdr->proximity_domain_from = numa_node_id;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1882
int numa_node_id;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1919
for_each_online_node(numa_node_id) {
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1920
if (kfd_numa_node_to_apic_id(numa_node_id) == -1)
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1924
ret = kfd_fill_cu_for_cpu(numa_node_id, &avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1936
ret = kfd_fill_mem_info_for_cpu(numa_node_id, &avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1949
ret = kfd_fill_iolink_info_for_cpu(numa_node_id, &avail_size,
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1194
int kfd_numa_node_to_apic_id(int numa_node_id);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2351
int kfd_numa_node_to_apic_id(int numa_node_id)
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2353
if (numa_node_id == -1) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
2357
return kfd_cpumask_to_apic_id(cpumask_of_node(numa_node_id));
drivers/gpu/drm/nouveau/nouveau_dmem.c
333
ptr = memremap_pages(&chunk->pagemap, numa_node_id());
drivers/infiniband/hw/hfi1/file_ops.c
960
numa = numa_node_id();
drivers/irqchip/irq-clps711x.c
180
err = irq_alloc_descs(-1, 0, ARRAY_SIZE(clps711x_irqs), numa_node_id());
drivers/irqchip/irq-hip04.c
383
irq_base = irq_alloc_descs(-1, 0, nr_irqs, numa_node_id());
drivers/md/dm-core.h
68
int numa_node_id;
drivers/md/dm-rq.c
550
md->tag_set = kzalloc_node(sizeof(struct blk_mq_tag_set), GFP_KERNEL, md->numa_node_id);
drivers/md/dm-rq.c
556
md->tag_set->numa_node = md->numa_node_id;
drivers/md/dm-table.c
1050
pools = kzalloc_node(sizeof(*pools), GFP_KERNEL, md->numa_node_id);
drivers/md/dm.c
2285
int r, numa_node_id = dm_get_numa_node();
drivers/md/dm.c
2290
md = kvzalloc_node(sizeof(*md), GFP_KERNEL, numa_node_id);
drivers/md/dm.c
2311
md->numa_node_id = numa_node_id;
drivers/md/dm.c
2331
md->disk = blk_alloc_disk(NULL, md->numa_node_id);
drivers/md/dm.c
733
td = kmalloc_node(sizeof(*td), GFP_KERNEL, md->numa_node_id);
drivers/net/ethernet/cavium/liquidio/octeon_network.h
342
unlikely(page_to_nid(pg_info->page) != numa_node_id())) {
drivers/net/ethernet/hisilicon/hns/hns_enet.c
459
if (unlikely(page_to_nid(desc_cb->priv) != numa_node_id()))
drivers/net/ethernet/hisilicon/hns/hns_enet.c
603
if (likely(page_to_nid(desc_cb->priv) == numa_node_id()))
drivers/nvme/host/multipath.c
365
int node = numa_node_id();
drivers/nvme/host/multipath.c
455
int node = numa_node_id();
drivers/sh/intc/virq.c
225
irq = irq_alloc_desc(numa_node_id());
drivers/soc/tegra/cbb/tegra234-cbb.c
442
local_socket_id = numa_node_id();
include/linux/gfp.h
331
return alloc_pages_node_noprof(numa_node_id(), gfp_mask, order);
include/linux/gfp.h
335
return __folio_alloc_node_noprof(gfp, order, numa_node_id());
include/linux/netdevice.h
3499
int n = numa_node_id() % 2;
include/linux/topology.h
114
#ifndef numa_node_id
include/linux/topology.h
167
return numa_node_id();
include/linux/topology.h
82
#ifndef numa_node_id
kernel/bpf/helpers.c
169
return numa_node_id();
kernel/fork.c
226
if (node != NUMA_NO_NODE && numa_node_id() != node)
kernel/fork.c
253
nid = numa_node_id();
kernel/futex/core.c
601
node = numa_node_id();
lib/test_bpf.c
15526
addr = (long)&numa_node_id;
lib/test_hmm.c
554
ptr = memremap_pages(&devmem->pagemap, numa_node_id());
mm/filemap.c
1001
NO_INTERLEAVE_INDEX, numa_node_id());
mm/hugetlb.c
1339
nid = numa_node_id();
mm/internal.h
907
return __alloc_frozen_pages_noprof(gfp, order, numa_node_id(), NULL);
mm/memfd.c
94
numa_node_id(),
mm/memory.c
5986
if (folio_nid(folio) == numa_node_id()) {
mm/mempolicy.c
1441
int nid = numa_node_id();
mm/mempolicy.c
2215
return numa_node_id();
mm/mempolicy.c
2257
return numa_node_id();
mm/mempolicy.c
2333
nid = numa_node_id();
mm/mempolicy.c
2374
init_nodemask_of_node(mask, numa_node_id());
mm/mempolicy.c
2544
folio = folio_alloc_mpol_noprof(gfp, order, pol, ilx, numa_node_id());
mm/mempolicy.c
2562
numa_node_id());
mm/mempolicy.c
2762
nr_allocated += alloc_pages_bulk_noprof(gfp, numa_node_id(), NULL,
mm/mempolicy.c
2794
numa_node_id(), pol, nr_pages, page_array);
mm/mempolicy.c
2796
nid = numa_node_id();
mm/mempolicy.c
2993
int thisnid = numa_node_id();
mm/mempolicy.c
3022
polnid = numa_node_id();
mm/mempolicy.c
348
node = numa_node_id();
mm/mempolicy.c
864
if (is_private_single_threaded && (nid == numa_node_id()))
mm/page-writeback.c
2588
reclaim_throttle(NODE_DATA(numa_node_id()),
mm/page_alloc.c
3225
if (zone_to_nid(z) != numa_node_id())
mm/page_alloc.c
5513
struct zonelist *zonelist = node_zonelist(numa_node_id(), GFP_KERNEL);
mm/page_alloc.c
7809
nid = numa_node_id();
mm/shmem.c
1930
folio = folio_alloc_mpol(gfp, order, mpol, ilx, numa_node_id());
mm/swap_state.c
570
folio = folio_alloc_mpol(gfp_mask, 0, mpol, ilx, numa_node_id());
mm/vmalloc.c
2713
node = numa_node_id();
mm/vmscan.c
6677
struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask);
mm/vmscan.c
7453
struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask);
mm/vmscan.c
7708
if (node_state(pgdat->node_id, N_CPU) && pgdat->node_id != numa_node_id())
mm/vmstat.c
847
if (zone_to_nid(zone) == numa_node_id()) {
net/core/dev.c
2855
int i, j, tci, numa_node_id = -2;
net/core/dev.c
2954
if (numa_node_id == -2)
net/core/dev.c
2955
numa_node_id = cpu_to_node(j);
net/core/dev.c
2956
else if (numa_node_id != cpu_to_node(j))
net/core/dev.c
2957
numa_node_id = -1;
net/core/dev.c
3000
(numa_node_id >= 0) ?
net/core/dev.c
3001
numa_node_id : NUMA_NO_NODE);
net/core/pktgen.c
2849
int node = numa_node_id();
net/core/pktgen.c
2909
int node = pkt_dev->node >= 0 ? pkt_dev->node : numa_node_id();
net/core/skbuff.c
7302
sdn = per_cpu_ptr(net_hotdata.skb_defer_nodes, cpu) + numa_node_id();
net/ipv4/udp.c
1717
udp_prod_queue = &udp_sk(sk)->udp_prod_queue[numa_node_id()];
net/openvswitch/flow.c
100
numa_node_id());