GFP_HIGHUSER_MOVABLE
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
gfp_t flags = GFP_HIGHUSER_MOVABLE | __GFP_ZERO;
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
vma_alloc_folio(GFP_HIGHUSER_MOVABLE | __GFP_ZERO, 0, vma, vaddr)
GFP_HIGHUSER_MOVABLE, order), 0);
mapping_set_gfp_mask(mapping, GFP_HIGHUSER_MOVABLE);
#define GFP_TRANSHUGE_LIGHT ((GFP_HIGHUSER_MOVABLE | __GFP_COMP | \
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, vaddr);
gfp |= hugepage_movable_supported(h) ? GFP_HIGHUSER_MOVABLE : GFP_HIGHUSER;
gfpflag_string(GFP_HIGHUSER_MOVABLE), \
dpage = alloc_page_vma(GFP_HIGHUSER_MOVABLE, args->vma, addr);
dpage = folio_page(folio_alloc(GFP_HIGHUSER_MOVABLE,
dpage = alloc_page(GFP_HIGHUSER_MOVABLE | __GFP_NOFAIL);
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER_MOVABLE,
dpage = alloc_page_vma(GFP_HIGHUSER_MOVABLE, args->vma, addr);
gfp_flags |= GFP_HIGHUSER_MOVABLE;
.gfp_mask = (GFP_HIGHUSER_MOVABLE & ~__GFP_RECLAIM) |
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr);
folio = read_swap_cache_async(entry, GFP_HIGHUSER_MOVABLE,
new_folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, addr);
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, vma, vmf->address);
folio = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE, vmf);
.gfp_mask = GFP_HIGHUSER_MOVABLE | __GFP_THISNODE,
gfp = GFP_HIGHUSER_MOVABLE | __GFP_RETRY_MAYFAIL | __GFP_COMP;
.gfp_mask = GFP_HIGHUSER_MOVABLE | __GFP_THISNODE,
gfp |= GFP_HIGHUSER_MOVABLE | __GFP_NOMEMALLOC | __GFP_NORETRY |
vm_total_pages = nr_free_zone_pages(gfp_zone(GFP_HIGHUSER_MOVABLE));
folio = swapin_readahead(entry, GFP_HIGHUSER_MOVABLE,
folio = vma_alloc_folio(GFP_HIGHUSER_MOVABLE, 0, dst_vma,
.gfp_mask = (GFP_HIGHUSER_MOVABLE & ~__GFP_RECLAIM) |
(GFP_HIGHUSER_MOVABLE & ~GFP_RECLAIM_MASK);
(GFP_HIGHUSER_MOVABLE & ~GFP_RECLAIM_MASK),
.gfp_mask = GFP_HIGHUSER_MOVABLE,
#define GFP_TRANSHUGE_LIGHT ((GFP_HIGHUSER_MOVABLE | __GFP_COMP | \