__GFP_NORETRY
pages[i] = alloc_pages(gfp | __GFP_NORETRY, order);
#define KEXEC_CONTROL_MEMORY_GFP (GFP_DMA | __GFP_NORETRY)
gfp_flags |= __GFP_NOWARN | __GFP_NORETRY;
return __get_dma_pages(GFP_KERNEL|__GFP_NORETRY, get_order(size));
__GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN);
folio = folio_alloc(gfp | __GFP_NORETRY, get_order(*size));
__GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
gfp_t gfp = GFP_NOIO | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY;
GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY,
GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY,
GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY | __GFP_ZERO,
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
(order ? __GFP_NORETRY : __GFP_RETRY_MAYFAIL),
buf = kvmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
buf = __vmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
__GFP_NORETRY | __GFP_ZERO,
| __GFP_NORETRY) & ~__GFP_RECLAIM) \
buf = kmalloc(len + 1, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
flags |= __GFP_NORETRY | __GFP_NOWARN;
#define QUIET (__GFP_NORETRY | __GFP_NOWARN)
noreclaim |= __GFP_NORETRY | __GFP_NOWARN;
i915_deps_init(&deps, GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
i915_deps_init(&deps, GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
#define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
#define GFP (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN))
GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
__GFP_NORETRY);
gfp_flags |= __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN |
pg = alloc_pages(GFP_HIGHUSER | __GFP_NORETRY |
gfp_mask |= __GFP_ZERO | __GFP_NORETRY;
alloc_flags |= __GFP_NORETRY;
res = kmalloc(*size, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
res = kmalloc(*size, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
gfp_mask & __GFP_NORETRY) {
b = alloc_buffer(c, GFP_NOWAIT | __GFP_NORETRY | __GFP_NOMEMALLOC);
b = alloc_buffer(c, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
| __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP,
likely(gfp_mask & __GFP_NORETRY))
io->integrity_metadata = kmalloc(tag_len, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
clone = bio_kmalloc(nr_iovecs, GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN);
pages = alloc_pages(GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP, order);
GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN);
checksums = kmalloc(PAGE_SIZE, GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN);
dio->integrity_payload = kmalloc(dio->payload_len, GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
pl = alloc_pl(__GFP_NOWARN | __GFP_NORETRY | __GFP_KSWAPD_RECLAIM);
GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
GFP_NOIO | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
__GFP_NORETRY;
gfp_t flags = __GFP_NOWARN | __GFP_DIRECT_RECLAIM | __GFP_NORETRY;
__GFP_ZERO | __GFP_NORETRY);
__GFP_NORETRY,
buf = __vmalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
__GFP_NORETRY, get_order(dmb->dmb_len));
buf = kvzalloc(bufsize, GFP_KERNEL | __GFP_NORETRY);
const gfp_t high_order_gfp = __GFP_NOWARN | __GFP_NORETRY;
#define VIRTIO_BALLOON_FREE_PAGE_ALLOC_FLAG (__GFP_NORETRY | __GFP_NOWARN | \
(GFP_HIGHUSER | __GFP_NOWARN | __GFP_NORETRY | __GFP_NOMEMALLOC)
__GFP_NOMEMALLOC | __GFP_NORETRY |
GFP_NOWAIT | __GFP_NORETRY
__GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
gfp |= __GFP_NORETRY | __GFP_NOWARN;
gfp |= __GFP_NORETRY | __GFP_NOWARN;
ret |= __GFP_NORETRY | __GFP_NOWARN;
const gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
gfp |= __GFP_NORETRY | __GFP_NOWARN;
gfp_mask |= __GFP_NORETRY;
gfp_mask |= __GFP_NORETRY;
flags |= __GFP_NOWARN | __GFP_NORETRY;
folio_alloc(GFP_KERNEL | __GFP_NORETRY, order);
(__GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN));
#define KEXEC_CONTROL_MEMORY_GFP (GFP_KERNEL | __GFP_NORETRY)
GFP_KERNEL | __GFP_NORETRY)
return mapping_gfp_mask(x) | __GFP_NORETRY | __GFP_NOWARN;
!(gfp_flags & __GFP_NORETRY))
area = kmalloc_node(size, gfp | GFP_USER | __GFP_NORETRY,
#define PERF_AUX_GFP (GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY)
__GFP_NORETRY);
gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
GFP_KERNEL | __GFP_NORETRY, 0);
folio = folio_alloc(GFP_KERNEL | __GFP_NORETRY, order);
gfp_t gfp_flags = __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
err = __execmem_cache_free(&mas, area, GFP_KERNEL | __GFP_NORETRY);
alloc_gfp |= __GFP_NORETRY | __GFP_NOWARN;
return GFP_TRANSHUGE | (vma_madvised ? 0 : __GFP_NORETRY);
gfp_t gfp_mask = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
__GFP_NORETRY|__GFP_MEMALLOC|__GFP_NOMEMALLOC|\
__GFP_NORETRY | __GFP_NOWARN);
if (gfp_mask & __GFP_NORETRY)
gfp | __GFP_THISNODE | __GFP_NORETRY, order,
*gfp_mask |= __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN;
gfp |= GFP_HIGHUSER_MOVABLE | __GFP_NOMEMALLOC | __GFP_NORETRY |
if (gfp_has_flags(gfp_mask, __GFP_NORETRY | __GFP_THISNODE))
if (!(gfp_mask & __GFP_NORETRY))
if (gfp_mask & __GFP_NORETRY)
__GFP_NOWARN | __GFP_NORETRY | __GFP_NOMEMALLOC;
pcpu_gfp = gfp & (GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
const gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
gfp_t denyflags = __GFP_NOWARN | __GFP_NORETRY;
__get_free_page(GFP_KERNEL | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
__get_free_page(GFP_KERNEL | __GFP_NORETRY | __GFP_NOMEMALLOC | __GFP_NOWARN);
alloc_gfp = (flags | __GFP_NOWARN | __GFP_NORETRY) & ~__GFP_NOFAIL;
__GFP_NOFAIL | __GFP_ZERO | __GFP_NORETRY |\
gfp = GFP_NOWAIT | __GFP_NORETRY | __GFP_HIGHMEM | __GFP_MOVABLE;
__GFP_NORETRY,
__GFP_NOWARN | __GFP_NORETRY);
__GFP_NOWARN | __GFP_NORETRY);
__GFP_ZERO | __GFP_NOWARN | __GFP_NORETRY;
gfp_flags &= ~__GFP_NORETRY;
__GFP_NORETRY | __GFP_ZERO,
__GFP_NOWARN | __GFP_NORETRY |
gfp_t gfp_flags = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
return GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;
#define XPRTRDMA_GFP_FLAGS (__GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN)
page = alloc_page(GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN);
gfp_t gfp_mask = __GFP_RECLAIM | __GFP_NOWARN | __GFP_NORETRY;
gfp_t gfp = GFP_KERNEL | __GFP_NORETRY | __GFP_NOWARN;