__GFP_ACCOUNT
return gfp | __GFP_ACCOUNT;
return gfp | __GFP_ACCOUNT;
gfp &= ~__GFP_ACCOUNT;
gfp |= __GFP_ACCOUNT;
.gfp_custom = (in_atomic) ? GFP_ATOMIC | __GFP_ACCOUNT : 0,
buf = __vmalloc_node(data_size, PAGE_SIZE, GFP_KERNEL | __GFP_ZERO | __GFP_ACCOUNT,
#define GFP_KVM_S390_MMU_CACHE (GFP_ATOMIC | __GFP_ACCOUNT | __GFP_NOWARN)
gfp &= ~__GFP_ACCOUNT;
gfp &= ~__GFP_ACCOUNT;
gfp &= ~__GFP_ACCOUNT;
flags |= __GFP_ACCOUNT;
GFP_KERNEL | __GFP_ACCOUNT |
GFP_KERNEL | __GFP_ACCOUNT | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
GFP_KERNEL | __GFP_ACCOUNT | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
GFP_KERNEL | __GFP_ACCOUNT | __GFP_RETRY_MAYFAIL | __GFP_NOWARN);
bh = folio_alloc_buffers(folio, size, gfp | __GFP_ACCOUNT);
gfp_t gfp = GFP_NOFS | __GFP_ACCOUNT | __GFP_NOFAIL;
gfp_t gfp = GFP_NOFS | __GFP_ACCOUNT;
return alloc_page(GFP_HIGHUSER | __GFP_ACCOUNT);
#define GFP_PGTABLE_USER (GFP_PGTABLE_KERNEL | __GFP_ACCOUNT)
return flags | __GFP_ACCOUNT;
#define GFP_KERNEL_ACCOUNT (GFP_KERNEL | __GFP_ACCOUNT)
(IS_ENABLED(CONFIG_MEMCG) ? __GFP_ACCOUNT : 0))
((gfp_flags & __GFP_ACCOUNT) ||
(__entry->gfp_flags & (__force unsigned long)__GFP_ACCOUNT)) ? "true" : "false")
ocqe = kzalloc(ocq_size, gfp | __GFP_ACCOUNT);
pages = kmalloc_nolock(alloc_pages * sizeof(struct page *), __GFP_ACCOUNT, NUMA_NO_NODE);
s = kmalloc_nolock(sizeof(struct arena_free_span), __GFP_ACCOUNT, -1);
ret = __alloc(c, NUMA_NO_NODE, GFP_KERNEL | __GFP_NOWARN | __GFP_ACCOUNT);
gfp = __GFP_NOWARN | __GFP_ACCOUNT;
new_rn = kmalloc_nolock(sizeof(struct range_node), __GFP_ACCOUNT,
left = kmalloc_nolock(sizeof(struct range_node), __GFP_ACCOUNT, NUMA_NO_NODE);
ptr = kmalloc_node(size, flags | __GFP_ACCOUNT, node);
ptr = kmalloc_nolock(size, flags | __GFP_ACCOUNT, node);
ptr = kzalloc(size, flags | __GFP_ACCOUNT);
ptr = kvcalloc(n, size, flags | __GFP_ACCOUNT);
ptr = __alloc_percpu_gfp(size, align, flags | __GFP_ACCOUNT);
return alloc_pages_nolock(__GFP_ACCOUNT, nid, 0);
GFP_KERNEL | __GFP_ZERO | __GFP_ACCOUNT
enabler = kzalloc_obj(*enabler, GFP_NOWAIT | __GFP_ACCOUNT);
gfp_mask &= ~__GFP_ACCOUNT;
gfp |= __GFP_ACCOUNT;
gfp |= __GFP_ACCOUNT;
gfp |= __GFP_ACCOUNT;
gfp |= __GFP_ACCOUNT;
if (memcg_kmem_online() && (gfp & __GFP_ACCOUNT))
if (memcg_kmem_online() && (gfp & __GFP_ACCOUNT) && page &&
VM_WARN_ON_ONCE(gfp_flags & ~__GFP_ACCOUNT);
if (memcg_kmem_online() && page && (gfp_flags & __GFP_ACCOUNT) &&
if (!memcg_kmem_online() || !(gfp & __GFP_ACCOUNT))
__GFP_ACCOUNT | __GFP_NOFAIL)
if (likely(!(flags & __GFP_ACCOUNT) && !(s->flags & SLAB_ACCOUNT)))
VM_WARN_ON_ONCE(gfp_flags & ~(__GFP_ACCOUNT | __GFP_ZERO |
if (gfp_mask & __GFP_ACCOUNT && area->nr_pages)
ifa = kzalloc_obj(*ifa, gfp_flags | __GFP_ACCOUNT);
GFP_ATOMIC | __GFP_ACCOUNT | __GFP_NOWARN);
#define GFP_KERNEL_ACCOUNT (GFP_KERNEL | __GFP_ACCOUNT)
p = mmu_memory_cache_alloc_obj(mc, GFP_ATOMIC | __GFP_ACCOUNT);