__GFP_FS
lo->old_gfp_mask & ~(__GFP_IO | __GFP_FS));
bool can_backup = !!(sc->gfp_mask & __GFP_FS);
bool can_backup = !!(sc->gfp_mask & __GFP_FS);
const gfp_t gfp = __GFP_NOWARN | __GFP_RECLAIM | __GFP_FS;
folio = filemap_alloc_folio(mapping_gfp_constraint(mapping, ~__GFP_FS),
~(__GFP_FS | __GFP_HIGHMEM)));
return mapping_gfp_constraint(mapping, ~__GFP_FS);
folio = filemap_alloc_folio(mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS),
gfp |= mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
mapping_gfp_constraint(mapping, ~__GFP_FS));
~__GFP_FS));
mapping_gfp_constraint(mapping, ~__GFP_FS));
~__GFP_FS) | __GFP_MOVABLE;
~__GFP_FS);
~__GFP_FS) | __GFP_MOVABLE | __GFP_NOFAIL;
if (!(sc->gfp_mask & __GFP_FS))
mapping_set_gfp_mask(inode->i_mapping, gfp_mask & ~__GFP_FS);
if (!(sc->gfp_mask & __GFP_FS))
if ((gfp_mask & __GFP_FS) == 0)
bool nofs = !mapping_gfp_constraint(mapping, __GFP_FS);
if (current_is_kswapd() || !(gfp & __GFP_FS))
if (gfp_flags & __GFP_FS)
if (current_is_kswapd() || !(gfp & __GFP_FS))
mapping_set_gfp_mask(m, (gfp_mask & ~(__GFP_FS)));
mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS));
mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS));
~__GFP_FS));
if (!(sc->gfp_mask & __GFP_FS))
gfp_t ra_gfp_mask = readahead_gfp_mask(mapping) & ~__GFP_FS;
mapping_set_gfp_mask(inode->i_mapping, (gfp_mask & ~(__GFP_FS)));
if ((sc->gfp_mask & (__GFP_FS|__GFP_DIRECT_RECLAIM)) != (__GFP_FS|__GFP_DIRECT_RECLAIM))
gfp = mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
gfp = mapping_gfp_constraint(bdev->bd_mapping, ~__GFP_FS);
return gfp_has_flags(gfp, __GFP_IO | __GFP_FS);
#define GFP_KERNEL (__GFP_RECLAIM | __GFP_IO | __GFP_FS)
#define GFP_USER (__GFP_RECLAIM | __GFP_IO | __GFP_FS | __GFP_HARDWALL)
flags &= ~(__GFP_IO | __GFP_FS);
flags &= ~__GFP_FS;
WARN_ON((saved_gfp_mask & ~(__GFP_IO | __GFP_FS)) != gfp_allowed_mask);
gfp_allowed_mask &= ~(__GFP_IO | __GFP_FS);
if (!(cc->gfp_mask & __GFP_FS) && mapping)
if (cc->gfp_mask & __GFP_FS) {
gfp &= ~__GFP_FS;
#define GFP_RECLAIM_MASK (__GFP_RECLAIM|__GFP_HIGH|__GFP_IO|__GFP_FS|\
#define GFP_BOOT_MASK (__GFP_BITS_MASK & ~(__GFP_RECLAIM|__GFP_IO|__GFP_FS))
return mapping_gfp_mask(vm_file->f_mapping) | __GFP_FS | __GFP_IO;
if (!(oc->gfp_mask & __GFP_FS) && !is_memcg_oom(oc))
if (gfp_mask & __GFP_FS)
if (gfp_mask & __GFP_FS)
const gfp_t reclaim_mask = __GFP_IO | __GFP_FS | __GFP_RECLAIM;
gfp_t allowflags = __GFP_IO | __GFP_FS | __GFP_RECLAIM;
else if ((gfp_mask & (__GFP_FS | __GFP_IO)) == __GFP_IO)
else if ((gfp_mask & (__GFP_FS | __GFP_IO)) == 0)
if (!(gfp & (__GFP_HIGHMEM | __GFP_MOVABLE | __GFP_IO | __GFP_FS)))
if (gfp_mask & __GFP_FS)
if (!(gfp_mask & __GFP_FS))
#define GFP_KERNEL (__GFP_RECLAIM | __GFP_IO | __GFP_FS)
#define GFP_USER (__GFP_RECLAIM | __GFP_IO | __GFP_FS | __GFP_HARDWALL)