__GFP_IO
return (gfp & ~(__GFP_DIRECT_RECLAIM | __GFP_IO)) |
lo->old_gfp_mask & ~(__GFP_IO | __GFP_FS));
.writeback = !ctx.no_wait_gpu && (sc->gfp_mask & __GFP_IO),
if (sc->gfp_mask & __GFP_IO)
cfolio = filemap_alloc_folio(__GFP_NOWARN | __GFP_IO, 0, NULL);
#define NILFS_MDT_GFP (__GFP_RECLAIM | __GFP_IO | __GFP_HIGHMEM)
return gfp_has_flags(gfp, __GFP_IO | __GFP_FS);
return IS_ENABLED(CONFIG_COMPACTION) && (gfp_mask & __GFP_IO);
#define GFP_KERNEL (__GFP_RECLAIM | __GFP_IO | __GFP_FS)
#define GFP_NOFS (__GFP_RECLAIM | __GFP_IO)
#define GFP_USER (__GFP_RECLAIM | __GFP_IO | __GFP_FS | __GFP_HARDWALL)
flags &= ~(__GFP_IO | __GFP_FS);
WARN_ON((saved_gfp_mask & ~(__GFP_IO | __GFP_FS)) != gfp_allowed_mask);
gfp_allowed_mask &= ~(__GFP_IO | __GFP_FS);
#define GFP_RECLAIM_MASK (__GFP_RECLAIM|__GFP_HIGH|__GFP_IO|__GFP_FS|\
#define GFP_BOOT_MASK (__GFP_BITS_MASK & ~(__GFP_RECLAIM|__GFP_IO|__GFP_FS))
return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS | __GFP_IO;
return *gfp_mask & ~(__GFP_DIRECT_RECLAIM | __GFP_IO);
const gfp_t reclaim_mask = __GFP_IO | __GFP_FS | __GFP_RECLAIM;
gfp_t allowflags = __GFP_IO | __GFP_FS | __GFP_RECLAIM;
if (!(gfp & __GFP_IO))
else if ((gfp_mask & (__GFP_FS | __GFP_IO)) == __GFP_IO)
else if ((gfp_mask & (__GFP_FS | __GFP_IO)) == 0)
if (!(gfp & (__GFP_HIGHMEM | __GFP_MOVABLE | __GFP_IO | __GFP_FS)))
if (!folio_test_swapcache(folio) || !(gfp_mask & __GFP_IO))
if (!(sc->gfp_mask & __GFP_IO))
if (!(sc->gfp_mask & __GFP_IO) &&
#define GFP_KERNEL (__GFP_RECLAIM | __GFP_IO | __GFP_FS)
#define GFP_NOFS (__GFP_RECLAIM | __GFP_IO)
#define GFP_USER (__GFP_RECLAIM | __GFP_IO | __GFP_FS | __GFP_HARDWALL)