alloc_pages_node
page = alloc_pages_node(cpu_to_node(i), GFP_KERNEL, order);
page = alloc_pages_node(cpu_to_node(cpu), GFP_ATOMIC, get_order(vec_sz));
page = alloc_pages_node(node, flag, order);
page = alloc_pages_node(phys_id,
page = alloc_pages_node(nid,
page = alloc_pages_node(nid,
tce_mem = alloc_pages_node(nid, GFP_ATOMIC | __GFP_NOWARN,
table_pages = alloc_pages_node(pe->phb->hose->node, GFP_KERNEL,
pages = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, alloc_order);
page = alloc_pages_node(numa_node, GFP_KERNEL, 0);
page = alloc_pages_node(numa_node, GFP_KERNEL, order);
page = alloc_pages_node(nid, gfp, order);
page = alloc_pages_node(nid, gfp, order);
p = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL_ACCOUNT | __GFP_ZERO, 1);
p = alloc_pages_node(node, gfp | __GFP_ZERO, 0);
struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0);
struct page *page = alloc_pages_node(node, PGALLOC_GFP, 0);
stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0));
ph = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER);
ps = alloc_pages_node(node, THREADINFO_GFP, THREAD_SIZE_ORDER);
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
p = alloc_pages_node(node, gfp | __GFP_ZERO, 1);
return alloc_pages_node(node, gfp | __GFP_ZERO, 0);
page = alloc_pages_node(node,
page = alloc_pages_node(node, GFP_KERNEL, 0);
page = alloc_pages_node(numa_node, gfp_flags, get_order(size));
lmc->mem = alloc_pages_node(lmc->node, GFP_KERNEL, 0);
p = alloc_pages_node(dev_to_node(pagetable->iommu_dev),
p = alloc_pages_node(dev_to_node(vm->ptdev->base.dev),
p = alloc_pages_node(pool->nid, gfp_flags, order);
page = alloc_pages_node(cpu_to_node(newchannel->target_cpu),
pages[i] = alloc_pages_node(node, GFP_KERNEL, order);
page = alloc_pages_node(node,
page = alloc_pages_node(node, gfp, get_order(alloc_size));
page = alloc_pages_node(nid, alloc_flags, order);
page = alloc_pages_node(node, gfp | gfp_flags_quirk, order);
page = alloc_pages_node(nid, GFP_KERNEL, order);
pages = alloc_pages_node(node, gfp, order);
newpage = alloc_pages_node(node, __GFP_NOWARN | GFP_KERNEL |
pg = alloc_pages_node(node, gfp | __GFP_COMP, s->fl_pg_order);
pg = alloc_pages_node(node, gfp, 0);
rqinfo->page = alloc_pages_node(node, GFP_KERNEL, 0);
*page = alloc_pages_node(priv->numa_node, gfp_flags, 0);
struct page *page = alloc_pages_node(priv->numa_node,
struct page *page = alloc_pages_node(rx->gve->numa_node, GFP_ATOMIC, 0);
page = alloc_pages_node(dev_to_node(ring_to_dev(ring)),
page = alloc_pages_node(node, gfp_mask, order);
struct page *page = alloc_pages_node(node, GFP_KERNEL, 0);
page = alloc_pages_node(nid, GFP_HIGHUSER, 0);
data = alloc_pages_node(node, GFP_KERNEL, get_order(R8169_RX_BUF_SIZE));
page = alloc_pages_node(numa_node, flags, 0);
return alloc_pages_node(nid,
while ((page = alloc_pages_node(node, gfp, get_order(size)))
page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size));
page = alloc_pages_node(node, PERF_AUX_GFP, order);
page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);
struct page *page = alloc_pages_node(node, THREADINFO_GFP,
page = alloc_pages_node(cpu_to_node(cpu), mflags, order);
page = alloc_pages_node(cpu_to_node(cpu),
struct page *p = alloc_pages_node(cpu_to_node(cpu),
walk.reuse_page = alloc_pages_node(nid, gfp_mask, 0);
page = alloc_pages_node(nid, gfp_mask, 0);
*pagep = alloc_pages_node(cpu_to_node(cpu), gfp, 0);
page = alloc_pages_node(node, gfp_mask, order);
struct page *page = alloc_pages_node(nid, gfp, 0);
page = alloc_pages_node(pool->p.nid, gfp, pool->p.order);
pkt_dev->page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0);