__GFP_NOFAIL
GFP_KERNEL | __GFP_NOFAIL);
ctx = kzalloc_obj(*ctx, GFP_KERNEL | __GFP_NOFAIL);
src_pfns = kvcalloc(npages, sizeof(*src_pfns), GFP_KERNEL | __GFP_NOFAIL);
dst_pfns = kvcalloc(npages, sizeof(*dst_pfns), GFP_KERNEL | __GFP_NOFAIL);
dma_info = kvzalloc_objs(*dma_info, npages, GFP_KERNEL | __GFP_NOFAIL);
dpage = alloc_page(GFP_HIGHUSER | __GFP_NOFAIL);
vbuf = kmem_cache_zalloc(vgdev->vbufs, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
skb = alloc_skb(wr_len, GFP_KERNEL | __GFP_NOFAIL);
skb = alloc_skb(wr_len, GFP_KERNEL | __GFP_NOFAIL);
nreg = kmalloc_obj(*nreg, GFP_NOIO | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
skb = alloc_skb(len, GFP_KERNEL | __GFP_NOFAIL);
skb = alloc_skb(len, GFP_KERNEL | __GFP_NOFAIL);
skb = alloc_skb(wrlen, GFP_KERNEL | __GFP_NOFAIL);
struct sk_buff *skb = alloc_skb(0, GFP_KERNEL | __GFP_NOFAIL);
ld = kmalloc_obj(struct tty_ldisc, GFP_KERNEL | __GFP_NOFAIL);
eb = kmem_cache_zalloc(extent_buffer_cache, GFP_NOFS|__GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
const gfp_t gfp = nofail ? (GFP_NOFS | __GFP_NOFAIL) : GFP_NOFS;
gfp |= __GFP_NOFAIL;
gfp_t gfp = GFP_NOFS | __GFP_ACCOUNT | __GFP_NOFAIL;
struct erofs_fileio_rq *rq = kzalloc_obj(*rq, GFP_KERNEL | __GFP_NOFAIL);
io = kmalloc_obj(*io, GFP_KERNEL | __GFP_NOFAIL);
item = kmalloc_obj(*item, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
gfp_flags |= __GFP_NOFAIL;
GFP_NOFS | __GFP_NOFAIL);
gfp_flags |= __GFP_NOFAIL;
gfp_flags |= __GFP_NOFAIL;
return kmem_cache_zalloc(ext4_pending_cachep, GFP_KERNEL | __GFP_NOFAIL);
return kmem_cache_zalloc(ext4_es_cachep, GFP_KERNEL | __GFP_NOFAIL);
GFP_NOFS|__GFP_NOFAIL);
GFP_NOFS|__GFP_NOFAIL);
GFP_NOFS|__GFP_NOFAIL);
GFP_NOFS|__GFP_NOFAIL);
new_gfp_flags |= __GFP_NOFAIL;
~__GFP_FS) | __GFP_MOVABLE | __GFP_NOFAIL;
ret = radix_tree_preload(GFP_NOFS | __GFP_NOFAIL);
gfp_flags |= __GFP_NOFAIL;
cc.log_cluster_size, GFP_NOFS | __GFP_NOFAIL);
entry = kmem_cache_alloc(cachep, flags | __GFP_NOFAIL);
err = radix_tree_preload(GFP_NOFS | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
req = fuse_request_alloc(fm, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL))
err = fuse_simple_background(fm, args, GFP_NOFS | __GFP_NOFAIL);
ia = kzalloc_obj(*ia, GFP_KERNEL | __GFP_NOFAIL);
bucket = kzalloc_obj(*bucket, GFP_KERNEL | __GFP_NOFAIL);
forget = kmalloc_obj(*forget, GFP_NOFS | __GFP_NOFAIL);
w = kzalloc_obj(*w, GFP_NOFS | __GFP_NOFAIL);
mapping_gfp_mask(mapping) | __GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
bd = kmem_cache_zalloc(gfs2_bufdata_cachep, GFP_NOFS | __GFP_NOFAIL);
gfp = GFP_NOFS | __GFP_NOFAIL;
GFP_NOFS | __GFP_NOFAIL);
new_bh = alloc_buffer_head(GFP_NOFS|__GFP_NOFAIL);
tmp = jbd2_alloc(bh_in->b_size, GFP_NOFS | __GFP_NOFAIL);
gfp_mask |= __GFP_NOFAIL;
GFP_NOFS | __GFP_NOFAIL);
GFP_NOFS|__GFP_NOFAIL);
gfp_mask |= __GFP_NOFAIL;
gfp |= __GFP_NOFAIL;
iter->namebuf = kmalloc(UDF_NAME_LEN_CS0, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
tmpbuffer = kmalloc(size, GFP_KERNEL | __GFP_NOFAIL);
tmpbuffer = kvmalloc(args->geo->blksize, GFP_KERNEL | __GFP_NOFAIL);
tmpbuffer = kvmalloc(args->geo->blksize, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
bi = kmem_cache_alloc(xfs_bmap_intent_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
ifp->if_broot = kzalloc(new_size, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
dfc = kzalloc_obj(*dfc, GFP_KERNEL | __GFP_NOFAIL);
args = kzalloc_obj(*args, GFP_KERNEL | __GFP_NOFAIL);
args = kzalloc_obj(*args, GFP_KERNEL | __GFP_NOFAIL);
args = kzalloc_obj(*args, GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
args = kzalloc_obj(*args, GFP_KERNEL | __GFP_NOFAIL);
args = kzalloc_obj(*args, GFP_KERNEL | __GFP_NOFAIL);
sfp = kmalloc(ifp->if_bytes, GFP_KERNEL | __GFP_NOFAIL);
buf = kmalloc(oldsize, GFP_KERNEL | __GFP_NOFAIL);
buf = kmalloc(oldsize, GFP_KERNEL | __GFP_NOFAIL);
sfp = kmalloc(mp->m_sb.sb_inodesize, GFP_KERNEL | __GFP_NOFAIL);
buf = kmalloc(old_isize, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
return kzalloc(size, GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
ifp->if_broot = kmalloc(new_size, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
ri = kmem_cache_alloc(xfs_rmap_intent_cache, GFP_KERNEL | __GFP_NOFAIL);
attrip = kmem_cache_zalloc(xfs_attri_cache, GFP_KERNEL | __GFP_NOFAIL);
sizeof(struct xfs_da_args), GFP_KERNEL | __GFP_NOFAIL);
attrdp = kmem_cache_zalloc(xfs_attrd_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_NOFS | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
buip = kmem_cache_zalloc(xfs_bui_cache, GFP_KERNEL | __GFP_NOFAIL);
budp = kmem_cache_zalloc(xfs_bud_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
bp->b_addr = kmalloc(size, gfp_mask | __GFP_NOFAIL);
btp = kzalloc_obj(*btp, GFP_KERNEL | __GFP_NOFAIL);
gfp_mask |= __GFP_NOFAIL;
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
bip = kmem_cache_zalloc(xfs_buf_item_cache, GFP_KERNEL | __GFP_NOFAIL);
bcp = kmalloc_obj(struct xfs_buf_cancel, GFP_KERNEL | __GFP_NOFAIL);
dqp = kmem_cache_zalloc(xfs_dquot_cache, GFP_KERNEL | __GFP_NOFAIL);
xmi_lip = kmem_cache_zalloc(xfs_xmi_cache, GFP_KERNEL | __GFP_NOFAIL);
xmd_lip = kmem_cache_zalloc(xfs_xmd_cache, GFP_KERNEL | __GFP_NOFAIL);
new = kzalloc_obj(struct xfs_extent_busy, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
ip = alloc_inode_sb(mp->m_super, xfs_inode_cache, GFP_KERNEL | __GFP_NOFAIL);
icp = kmem_cache_zalloc(xfs_icreate_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
iup = kmem_cache_zalloc(xfs_iunlink_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
ctx = kzalloc_obj(*ctx, GFP_KERNEL | __GFP_NOFAIL);
item = kzalloc_obj(struct xlog_recover_item, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
trans = kzalloc_obj(struct xlog_recover, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
mru = kzalloc_obj(*mru, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
cudp = kmem_cache_zalloc(xfs_cud_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
rudp = kmem_cache_zalloc(xfs_rud_cache, GFP_KERNEL | __GFP_NOFAIL);
ri = kmem_cache_alloc(xfs_rmap_intent_cache, GFP_KERNEL | __GFP_NOFAIL);
tp = kmem_cache_zalloc(xfs_trans_cache, GFP_KERNEL | __GFP_NOFAIL);
ntp = kmem_cache_zalloc(xfs_trans_cache, GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
oz = kzalloc_obj(*oz, GFP_NOFS | __GFP_NOFAIL);
gfp |= __GFP_NOFAIL;
gfp |= __GFP_MOVABLE | __GFP_NOFAIL;
new_res = alloc_resource(GFP_KERNEL | (alloc_nofail ? __GFP_NOFAIL : 0));
gfp_mask |= __GFP_NOFAIL;
tbl = bucket_table_alloc(ht, size, GFP_KERNEL | __GFP_NOFAIL);
src_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*src_pfns), GFP_KERNEL | __GFP_NOFAIL);
dst_pfns = kvcalloc(PTRS_PER_PTE, sizeof(*dst_pfns), GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
GFP_KERNEL | __GFP_NOFAIL);
src_pfns = kvcalloc(npages, sizeof(*src_pfns), GFP_KERNEL | __GFP_NOFAIL);
dst_pfns = kvcalloc(npages, sizeof(*dst_pfns), GFP_KERNEL | __GFP_NOFAIL);
dpage = alloc_page(GFP_HIGHUSER_MOVABLE | __GFP_NOFAIL);
list4 = kzalloc_obj(*list4, GFP_KERNEL | __GFP_NOFAIL);
list5 = kmalloc_obj(*list5, GFP_KERNEL | __GFP_NOFAIL);
list4 = kzalloc_obj(*list4, GFP_KERNEL | __GFP_NOFAIL);
list5 = kmalloc_obj(*list5, GFP_KERNEL | __GFP_NOFAIL);
if (gfp_mask & __GFP_NOFAIL)
if (gfpflags & __GFP_NOFAIL)
gfp_t gfp = gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
__GFP_NOWARN|__GFP_RETRY_MAYFAIL|__GFP_NOFAIL|\
if (gfp_mask & __GFP_NOFAIL) {
if (!(gfp_mask & (__GFP_NOFAIL | __GFP_HIGH)))
preferred_gfp &= ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
preferred_gfp &= ~(__GFP_DIRECT_RECLAIM | __GFP_NOFAIL);
WARN_ON_ONCE_GFP(gfp_mask & __GFP_NOFAIL, gfp_mask)) {
if (gfp_mask & __GFP_NOFAIL)
bool nofail = gfp_mask & __GFP_NOFAIL;
if (gfp & __GFP_NOFAIL) {
__GFP_ACCOUNT | __GFP_NOFAIL)
alloc_gfp = (flags | __GFP_NOWARN | __GFP_NORETRY) & ~__GFP_NOFAIL;
slab_post_alloc_hook(s, NULL, gfp | __GFP_NOFAIL, 1, &ret, init, s->object_size);
flags &= ~__GFP_NOFAIL;
mas_store_gfp(mas, NULL, GFP_KERNEL|__GFP_NOFAIL);
flags &= ~__GFP_NOFAIL;
if (!(gfp & __GFP_NOFAIL) && fatal_signal_pending(current))
bool nofail = gfp_mask & __GFP_NOFAIL;
__GFP_NOFAIL | __GFP_ZERO | __GFP_NORETRY |\
bool nofail = gfp_mask & __GFP_NOFAIL;
osd = kzalloc_obj(*osd, GFP_NOIO | __GFP_NOFAIL);
GFP_NOIO | __GFP_NOFAIL);
buf = kmalloc(total, GFP_NOIO | __GFP_NOFAIL);
gfp_memcg_charge() | __GFP_NOFAIL);
gfp |= __GFP_NOFAIL;
mem_cgroup_sk_charge(sk, amt, gfp_memcg_charge() | __GFP_NOFAIL);
gfp_t gfp = GFP_KERNEL | __GFP_NOFAIL;
GFP_KERNEL | __GFP_NOFAIL);