page_to_nid
page_to_nid(page),
page_to_nid(page),
page_to_nid(page),
__GFP_HIGHMEM | __GFP_MOVABLE, page_to_nid(page));
__GFP_HIGHMEM | __GFP_MOVABLE, page_to_nid(page));
page_to_nid(page));
unlikely(page_to_nid(pg_info->page) != numa_node_id())) {
rb->node = page_is_pfmemalloc(p) ? -1 : page_to_nid(p);
if (unlikely(page_to_nid(desc_cb->priv) != numa_node_id()))
if (likely(page_to_nid(desc_cb->priv) == numa_node_id()))
page_to_nid(page) != numa_mem_id();
if (ne_enclave->numa_node != page_to_nid(mem_region_page)) {
nid = page_to_nid(page);
nid = page_to_nid(page);
return &NODE_DATA(page_to_nid(page))->node_zones[page_zonenum(page)];
return NODE_DATA(page_to_nid(page));
page_to_nid(pfn_to_page(__pfn_to_nid_pfn)); \
return likely(page_to_nid(page) == numa_mem_id() &&
return page_to_nid(netmem_to_page(netmem)) == pref_nid;
if (cma_release(dma_contiguous_pernuma_area[page_to_nid(page)],
if (cma_release(dma_contiguous_numa_area[page_to_nid(page)],
if (page_to_nid(page) != nid)
int nid = page_to_nid((struct page *)reuse);
int nid = page_to_nid((struct page *)start);
page_to_nid(tree_page) != nid) {
int nid = page_to_nid(virt_to_page(item));
int nid = page_to_nid(virt_to_page(item));
ret = page_to_nid(p);
page_to_nid(page) == nid) {
VM_BUG_ON_PAGE(page_to_nid(page) != zone_to_nid(zone), page);
base = NODE_DATA(page_to_nid(page))->node_page_ext;
index = pfn - round_down(node_start_pfn(page_to_nid(page)),
if (page_to_nid(virt_to_page(object)) != node) {
nid != page_to_nid(virt_to_page(p)))
node_set(page_to_nid(virt_to_page(track)), l->nodes);
node_set(page_to_nid(virt_to_page(track)), l->nodes);
nid != page_to_nid(vmalloc_to_page(p)))
counters[page_to_nid(v->pages[nr])] += step;
entry = zswap_entry_cache_alloc(GFP_KERNEL, page_to_nid(page));
return page_to_nid(virt_to_page(entry));
handle = zs_malloc(pool->zs_pool, dlen, gfp, page_to_nid(page));
return page_to_nid(page);