__GFP_COMP
o = (void *)__get_free_pages(GFP_KERNEL_ACCOUNT | __GFP_COMP, CRST_ALLOC_ORDER);
page = alloc_pages(GFP_KERNEL_ACCOUNT | __GFP_COMP, CRST_ALLOC_ORDER);
return (void *)__get_free_pages(GFP_KVM_S390_MMU_CACHE | __GFP_COMP, CRST_ALLOC_ORDER);
pages = __get_free_pages(GFP_KERNEL | __GFP_COMP, order);
pages = __get_free_pages(GFP_KERNEL | __GFP_COMP, order);
| __GFP_COMP)
gfp &= ~(__GFP_DMA | __GFP_DMA32 | __GFP_HIGHMEM | __GFP_COMP);
if (WARN_ON_ONCE(gfp & (__GFP_COMP | __GFP_DMA | __GFP_DMA32 |
t->data = (void *) __get_free_pages(__GFP_COMP|gfp, b->page_order);
__get_free_pages(GFP_KERNEL|__GFP_COMP,
!(j->w[0].data = (void *) __get_free_pages(GFP_KERNEL|__GFP_COMP, JSET_BITS)) ||
!(j->w[1].data = (void *) __get_free_pages(GFP_KERNEL|__GFP_COMP, JSET_BITS)))
((void *) __get_free_pages(__GFP_ZERO|__GFP_COMP|gfp, ilog2(meta_bucket_pages(sb))))
| __GFP_NOMEMALLOC | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP,
pages = alloc_pages(GFP_NOIO | __GFP_NORETRY | __GFP_NOWARN | __GFP_COMP, order);
gfp = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN;
page = alloc_pages(GFP_ATOMIC | __GFP_COMP,
page = alloc_pages(gfp | __GFP_COMP | __GFP_NOWARN, 0);
GFP_KERNEL | __GFP_COMP);
GFP_KERNEL | __GFP_COMP);
GFP_ATOMIC | __GFP_COMP);
GFP_ATOMIC | __GFP_COMP) == 0)
pg = alloc_pages_node(node, gfp | __GFP_COMP, s->fl_pg_order);
page = alloc_pages(gfp | __GFP_COMP |
page = alloc_pages(sk->sk_allocation | __GFP_COMP, 0);
gfp |= __GFP_COMP;
data = __get_free_pages(gfp_mask | __GFP_COMP | __GFP_NOWARN,
__GFP_ZERO | __GFP_COMP |
alloc_pages(GFP_ATOMIC | __GFP_COMP,
page = alloc_pages(__GFP_COMP |
page = alloc_pages(__GFP_COMP |
page = alloc_pages(__GFP_COMP |
gfp |= __GFP_COMP | __GFP_NOWARN;
rx_agg->page = alloc_pages(mflags | __GFP_COMP | __GFP_NOWARN, order);
gfp_mask |= __GFP_COMP;
gfp_mask |= __GFP_COMP;
gfp_mask |= __GFP_COMP;
udev->ll2_ring = (void *)get_zeroed_page(GFP_KERNEL | __GFP_COMP);
udev->ll2_buf = (void *)__get_free_pages(GFP_KERNEL | __GFP_COMP |
gfp_t gfp_mask = GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN | __GFP_ZERO;
#define GFP_TRANSHUGE_LIGHT ((GFP_HIGHUSER_MOVABLE | __GFP_COMP | \
gfp_t gfp = __GFP_COMP | __GFP_NOWARN;
struct page *page = alloc_pages_noprof(gfp | __GFP_COMP, order);
gfp_mask |= __GFP_COMP | __GFP_MEMALLOC;
gfp |= __GFP_COMP;
if (WARN_ON_ONCE(flag & __GFP_COMP))
if (WARN_ON_ONCE(gfp & __GFP_COMP))
if (WARN_ON_ONCE(gfp & __GFP_COMP))
mflags = GFP_KERNEL | __GFP_RETRY_MAYFAIL | __GFP_COMP | __GFP_ZERO;
test_free_pages(GFP_KERNEL | __GFP_COMP);
gfp_t gfp = GFP_KERNEL | __GFP_COMP | __GFP_NOWARN;
gfp = GFP_HIGHUSER_MOVABLE | __GFP_RETRY_MAYFAIL | __GFP_COMP;
struct page *page = alloc_pages_mpol(gfp | __GFP_COMP, order, pol,
return page_rmappable_folio(alloc_pages_noprof(gfp | __GFP_COMP, order));
return kmalloc_noprof(size, (gfp_mask | __GFP_COMP) & ~__GFP_HIGHMEM);
return krealloc_noprof(p, size, (flags | __GFP_COMP) & ~__GFP_HIGHMEM);
if (order && (gfp_flags & __GFP_COMP))
struct page *page = __alloc_pages_noprof(gfp | __GFP_COMP, order,
if (WARN_ON_ONCE(gfp_mask & (__GFP_COMP | __GFP_HIGHMEM)))
gfp_mask &= ~(__GFP_COMP | __GFP_HIGHMEM);
if (WARN_ON_ONCE(gfp_mask & (__GFP_COMP | __GFP_HIGHMEM)))
gfp_mask &= ~(__GFP_COMP | __GFP_HIGHMEM);
const gfp_t action_mask = __GFP_COMP | __GFP_RETRY_MAYFAIL | __GFP_NOWARN |
if (WARN_ON_ONCE((gfp_mask & __GFP_COMP) && order > MAX_FOLIO_ORDER))
if (!(gfp_mask & __GFP_COMP)) {
if (WARN_ON(gfp_mask & __GFP_COMP))
if (WARN_ON(gfp_mask & __GFP_COMP))
gfp_t alloc_gfp = __GFP_NOWARN | __GFP_ZERO | __GFP_NOMEMALLOC | __GFP_COMP
gfp_mask = (gfp_mask & ~__GFP_DIRECT_RECLAIM) | __GFP_COMP |
flags |= __GFP_COMP;
s->allocflags = __GFP_COMP;
page = alloc_pages(__GFP_NOWARN | GFP_ATOMIC | __GFP_COMP,
page = alloc_pages(__GFP_NOWARN | GFP_ATOMIC | __GFP_COMP,
__GFP_COMP | GFP_ATOMIC,
gfp |= __GFP_COMP;
page = alloc_pages(gfp_mask | __GFP_COMP, order);
__GFP_COMP |
__GFP_COMP | __GFP_NOWARN |
gfp_t gfp_flags = GFP_KERNEL | __GFP_COMP |
gfp |= __GFP_COMP;
__GFP_NOMEMALLOC | __GFP_COMP |
#define GFP_TRANSHUGE_LIGHT ((GFP_HIGHUSER_MOVABLE | __GFP_COMP | \