Symbol: page_folio
arch/arc/mm/cache.c
714
return flush_dcache_folio(page_folio(page));
arch/arc/mm/cache.c
885
struct folio *src = page_folio(from);
arch/arc/mm/cache.c
886
struct folio *dst = page_folio(to);
arch/arc/mm/cache.c
901
struct folio *folio = page_folio(page);
arch/arc/mm/tlb.c
490
struct folio *folio = page_folio(page);
arch/arm/mm/copypage-v4mc.c
67
struct folio *src = page_folio(from);
arch/arm/mm/copypage-v6.c
72
struct folio *src = page_folio(from);
arch/arm/mm/copypage-xscale.c
87
struct folio *src = page_folio(from);
arch/arm/mm/fault-armv.c
204
folio = page_folio(pfn_to_page(pfn));
arch/arm/mm/flush.c
298
folio = page_folio(pfn_to_page(pfn));
arch/arm/mm/flush.c
369
flush_dcache_folio(page_folio(page));
arch/arm64/include/asm/mte.h
44
VM_WARN_ON_ONCE(folio_test_hugetlb(page_folio(page)));
arch/arm64/include/asm/mte.h
58
VM_WARN_ON_ONCE(folio_test_hugetlb(page_folio(page)));
arch/arm64/include/asm/mte.h
83
VM_WARN_ON_ONCE(folio_test_hugetlb(page_folio(page)));
arch/arm64/kernel/hibernate.c
273
folio = page_folio(page);
arch/arm64/kernel/mte.c
41
struct folio *folio = page_folio(page);
arch/arm64/kernel/mte.c
477
folio = page_folio(page);
arch/arm64/kvm/guest.c
1027
folio = page_folio(page);
arch/arm64/kvm/mmu.c
1475
struct folio *folio = page_folio(page);
arch/arm64/mm/contpte.c
269
folio = page_folio(page);
arch/arm64/mm/copypage.c
21
struct folio *src = page_folio(from);
arch/arm64/mm/copypage.c
22
struct folio *dst = page_folio(to);
arch/arm64/mm/flush.c
54
struct folio *folio = page_folio(pte_page(pte));
arch/arm64/mm/flush.c
79
flush_dcache_folio(page_folio(page));
arch/csky/abiv1/cacheflush.c
40
flush_dcache_folio(page_folio(page));
arch/csky/abiv1/cacheflush.c
58
folio = page_folio(pfn_to_page(pfn));
arch/csky/abiv2/cacheflush.c
22
folio = page_folio(pfn_to_page(pfn));
arch/csky/abiv2/inc/abi/cacheflush.h
31
flush_dcache_folio(page_folio(page));
arch/mips/include/asm/cacheflush.h
68
struct folio *folio = page_folio(page);
arch/mips/mm/c-r4k.c
571
struct folio *folio = page_folio(page);
arch/mips/mm/cache.c
130
struct folio *folio = page_folio(page);
arch/mips/mm/cache.c
157
folio = page_folio(pfn_to_page(pfn));
arch/mips/mm/init.c
166
struct folio *src = page_folio(from);
arch/mips/mm/init.c
192
struct folio *folio = page_folio(page);
arch/mips/mm/init.c
212
struct folio *folio = page_folio(page);
arch/mips/mm/init.c
85
BUG_ON(folio_test_dcache_dirty(page_folio(page)));
arch/nios2/mm/cacheflush.c
205
flush_dcache_folio(page_folio(page));
arch/nios2/mm/cacheflush.c
229
folio = page_folio(pfn_to_page(pfn));
arch/openrisc/include/asm/cacheflush.h
85
flush_dcache_folio(page_folio(page));
arch/openrisc/mm/cache.c
85
struct folio *folio = page_folio(pfn_to_page(pfn));
arch/parisc/include/asm/cacheflush.h
51
flush_dcache_folio(page_folio(page));
arch/parisc/kernel/cache.c
121
folio = page_folio(pfn_to_page(pfn));
arch/powerpc/include/asm/cacheflush.h
50
flush_dcache_folio(page_folio(page));
arch/powerpc/include/asm/kvm_ppc.h
941
folio = page_folio(pfn_to_page(pfn));
arch/powerpc/mm/book3s64/hash_utils.c
1581
folio = page_folio(pte_page(pte));
arch/powerpc/mm/pgtable.c
72
return page_folio(page);
arch/riscv/include/asm/cacheflush.h
34
flush_dcache_folio(page_folio(page));
arch/riscv/mm/cacheflush.c
102
struct folio *folio = page_folio(pte_page(pte));
arch/s390/include/asm/page.h
260
#define phys_to_folio(phys) page_folio(phys_to_page(phys))
arch/s390/include/asm/tlb.h
50
free_folio_and_swap_cache(page_folio(page));
arch/s390/include/asm/tlb.h
63
VM_WARN_ON_ONCE(page_folio(page) != page_folio(page + nr_pages - 1));
arch/s390/kvm/gmap.c
690
order = folio_order(page_folio(f->page));
arch/s390/kvm/kvm-s390.h
345
struct folio *folio = page_folio(page);
arch/sh/include/asm/cacheflush.h
49
flush_dcache_folio(page_folio(page));
arch/sh/mm/cache-sh4.c
244
struct folio *folio = page_folio(page);
arch/sh/mm/cache.c
102
struct folio *src = page_folio(from);
arch/sh/mm/cache.c
150
struct folio *folio = page_folio(pfn_to_page(pfn));
arch/sh/mm/cache.c
160
struct folio *folio = page_folio(page);
arch/sh/mm/cache.c
247
cacheop_on_each_cpu(local_flush_icache_folio, page_folio(page), 1);
arch/sh/mm/cache.c
64
struct folio *folio = page_folio(page);
arch/sh/mm/cache.c
85
struct folio *folio = page_folio(page);
arch/sh/mm/kmap.c
30
struct folio *folio = page_folio(page);
arch/sparc/include/asm/cacheflush_32.h
45
flush_dcache_folio(page_folio(page));
arch/sparc/include/asm/cacheflush_64.h
53
flush_dcache_folio(page_folio(page));
arch/sparc/include/asm/pgtable_64.h
966
page_folio(pfn_to_page(this_pfn))); \
arch/sparc/mm/init_64.c
292
struct folio *folio = page_folio(page);
arch/sparc/mm/tlb.c
140
folio = page_folio(page);
arch/xtensa/include/asm/cacheflush.h
133
flush_dcache_folio(page_folio(page));
arch/xtensa/mm/cache.c
105
struct folio *folio = page_folio(dst);
arch/xtensa/mm/cache.c
226
folio = page_folio(pfn_to_page(pfn));
arch/xtensa/mm/cache.c
90
struct folio *folio = page_folio(page);
arch/xtensa/mm/tlb.c
259
struct folio *f = page_folio(p);
block/bio-integrity.c
280
struct folio *folio = page_folio(pages[i]);
block/bio-integrity.c
286
if (page_folio(pages[j]) != folio ||
block/bio.c
1323
struct folio *folio = page_folio(bv->bv_page);
block/bio.c
1427
struct folio *folio = page_folio(bv->bv_page);
block/bio.c
1461
folio_put(page_folio(bio->bi_io_vec[0].bv_page));
drivers/dax/device.c
183
return vmf_insert_folio_pmd(vmf, page_folio(pfn_to_page(pfn)),
drivers/dax/device.c
229
return vmf_insert_folio_pud(vmf, page_folio(pfn_to_page(pfn)),
drivers/gpu/drm/drm_gem.c
720
struct folio *f = page_folio(pages[j]);
drivers/gpu/drm/drm_gem.c
764
folio = page_folio(pages[i]);
drivers/gpu/drm/drm_gem_shmem_helper.c
564
folio_test_pmd_mappable(page_folio(pfn_to_page(pfn)))) {
drivers/gpu/drm/drm_pagemap.c
243
folio = page_folio(page);
drivers/gpu/drm/drm_pagemap.c
561
unsigned long nr_pages = src_page ? NR_PAGES(folio_order(page_folio(src_page))) : 1;
drivers/gpu/drm/drm_pagemap.c
653
unsigned long nr_pages = page ? NR_PAGES(folio_order(page_folio(page))) : 1;
drivers/gpu/drm/drm_pagemap.c
723
order = folio_order(page_folio(src_page));
drivers/gpu/drm/drm_pagemap.c
753
WARN_ON_ONCE(!folio_trylock(page_folio(page)));
drivers/gpu/drm/drm_pagemap.c
755
order = folio_order(page_folio(page));
drivers/gpu/drm/drm_pagemap.c
775
order = folio_order(page_folio(page));
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
46
struct folio *folio = page_folio(page);
drivers/gpu/drm/nouveau/nouveau_dmem.c
205
sfolio = page_folio(vmf->page);
drivers/gpu/drm/nouveau/nouveau_dmem.c
255
dfolio = page_folio(dpage);
drivers/gpu/drm/nouveau/nouveau_dmem.c
360
drm->dmem->free_folios = page_folio(page);
drivers/gpu/drm/nouveau/nouveau_dmem.c
417
folio = page_folio(page);
drivers/gpu/drm/nouveau/nouveau_dmem.c
423
folio = page_folio(page);
drivers/gpu/drm/nouveau/nouveau_dmem.c
492
struct folio *folio = page_folio(
drivers/gpu/drm/nouveau/nouveau_dmem.c
511
page_folio(migrate_pfn_to_page(src_pfns[i])),
drivers/gpu/drm/nouveau/nouveau_dmem.c
512
page_folio(dpage),
drivers/gpu/drm/nouveau/nouveau_dmem.c
753
if (drm->dmem->migrate.copy_func(drm, folio_nr_pages(page_folio(spage)),
drivers/gpu/drm/nouveau/nouveau_dmem.c
770
if (folio_order(page_folio(dpage)))
drivers/gpu/drm/nouveau/nouveau_dmem.c
803
folio = page_folio(migrate_pfn_to_page(args->dst[i]));
drivers/gpu/drm/nouveau/nouveau_svm.c
619
folio = page_folio(page);
drivers/gpu/drm/panthor/panthor_mmu.c
2114
return folio_size(page_folio(pg)) >= SZ_2M;
drivers/hv/mshv_regions.c
49
page_order = folio_order(page_folio(page));
fs/btrfs/compression.c
1178
const struct folio *folio = page_folio(page);
fs/btrfs/extent_io.c
710
eb->folios[i] = page_folio(page_array[i]);
fs/btrfs/file.c
1854
struct folio *folio = page_folio(page);
fs/btrfs/free-space-cache.c
1457
ret = btrfs_dirty_folio(BTRFS_I(inode), page_folio(io_ctl->pages[i]),
fs/btrfs/free-space-cache.c
437
page_folio(io_ctl->pages[i]),
fs/btrfs/inode.c
3458
struct folio *folio = page_folio(phys_to_page(paddr));
fs/btrfs/inode.c
890
return page_folio(phys_to_page(paddr));
fs/btrfs/tests/extent-io-tests.c
182
found = find_lock_delalloc_range(inode, page_folio(locked_page), &start,
fs/btrfs/tests/extent-io-tests.c
213
found = find_lock_delalloc_range(inode, page_folio(locked_page), &start,
fs/btrfs/tests/extent-io-tests.c
247
found = find_lock_delalloc_range(inode, page_folio(locked_page), &start,
fs/btrfs/tests/extent-io-tests.c
268
found = find_lock_delalloc_range(inode, page_folio(locked_page), &start,
fs/btrfs/tests/extent-io-tests.c
309
found = find_lock_delalloc_range(inode, page_folio(locked_page), &start,
fs/buffer.c
2616
struct folio *folio = page_folio(vmf->page);
fs/buffer.c
968
return folio_alloc_buffers(page_folio(page), size, gfp);
fs/ceph/addr.c
2046
struct folio *folio = page_folio(vmf->page);
fs/ceph/addr.c
550
folio_start_private_2(page_folio(page)); /* [DEPRECATED] */
fs/ceph/addr.c
952
page_folio(page));
fs/dax.c
75
return page_folio(pfn_to_page(dax_to_pfn(entry)));
fs/direct-io.c
224
folio_add_pin(page_folio(page));
fs/erofs/data.c
25
folio_put(page_folio(buf->page));
fs/erofs/data.c
35
folio = page_folio(buf->page);
fs/erofs/zdata.c
1103
return !page_folio(page)->mapping && !z_erofs_is_shortlived_page(page);
fs/erofs/zdata.c
1187
erofs_onlinefolio_end(page_folio(bvi->bvec.page), err, true);
fs/erofs/zdata.c
1238
erofs_folio_is_managed(EROFS_SB(be->sb), page_folio(page))) {
fs/erofs/zdata.c
1330
folio_put(page_folio(pcl->compressed_bvecs[0].page));
fs/erofs/zdata.c
1338
if (erofs_folio_is_managed(sbi, page_folio(page))) {
fs/erofs/zdata.c
1359
erofs_onlinefolio_end(page_folio(page), err, true);
fs/erofs/zdata.c
1513
folio = page_folio(zbv.page);
fs/erofs/zdata.c
1577
folio = page_folio(page);
fs/erofs/zdata.c
603
folio = page_folio(pcl->compressed_bvecs[i].page);
fs/erofs/zdata.c
633
if (bvec->page && page_folio(bvec->page) == folio) {
fs/erofs/zdata.c
868
folio_get(page_folio((struct page *)ptr));
fs/ext4/inode.c
6646
struct folio *folio = page_folio(vmf->page);
fs/ext4/page-io.c
587
io_folio = page_folio(bounce_page);
fs/f2fs/compress.c
1070
if (folio_test_writeback(page_folio(cc->rpages[i])))
fs/f2fs/compress.c
1200
struct folio *folio = page_folio(cc.rpages[0]);
fs/f2fs/compress.c
1244
struct folio *folio = page_folio(rpages[i]);
fs/f2fs/compress.c
1330
folio = page_folio(cc->rpages[last_index]);
fs/f2fs/compress.c
1354
page_folio(cc->rpages[i + 1])->index, cic);
fs/f2fs/compress.c
140
return page_folio(ctx->rpages[0]);
fs/f2fs/compress.c
1542
folio = page_folio(cc->rpages[i]);
fs/f2fs/compress.c
1820
rfolio = page_folio(rpage);
fs/f2fs/compress.c
596
folio = page_folio(page);
fs/f2fs/compress.c
847
unsigned long pgidx = page_folio(pages[index])->index;
fs/f2fs/compress.c
861
struct folio *folio = page_folio(pages[index + i]);
fs/f2fs/compress.c
884
if (page_folio(page)->index >= nr_pages)
fs/f2fs/compress.c
89
struct folio *folio = page_folio(page);
fs/f2fs/data.c
1040
bio_folio = page_folio(fio->encrypted_page);
fs/f2fs/data.c
1042
bio_folio = page_folio(fio->compressed_page);
fs/f2fs/data.c
2298
folio = page_folio(page);
fs/f2fs/data.c
2364
struct folio *folio = page_folio(dic->cpages[i]);
fs/f2fs/data.c
2801
fio->encrypted_page = fscrypt_encrypt_pagecache_blocks(page_folio(page),
fs/f2fs/data.c
3325
struct folio *folio = page_folio(page);
fs/f2fs/data.c
3896
*foliop = page_folio(page);
fs/f2fs/data.c
752
page_folio(fio->encrypted_page) : fio_folio;
fs/f2fs/data.c
946
page_folio(fio->encrypted_page) : fio->folio;
fs/f2fs/f2fs.h
3084
f2fs_folio_put(page_folio(page), unlock);
fs/f2fs/f2fs.h
4041
f2fs_folio_wait_writeback(page_folio(page), type, ordered, locked)
fs/f2fs/file.c
77
struct folio *folio = page_folio(vmf->page);
fs/fuse/dev.c
1009
newfolio = page_folio(buf->page);
fs/fuse/file.c
1607
struct folio *folio = page_folio(pages[i]);
fs/fuse/file.c
2347
struct folio *folio = page_folio(vmf->page);
fs/gfs2/file.c
418
struct folio *folio = page_folio(vmf->page);
fs/gfs2/lops.c
217
struct folio *folio = page_folio(page);
fs/iomap/buffered-io.c
1719
struct folio *folio = page_folio(vmf->page);
fs/netfs/buffered_write.c
508
struct folio *folio = page_folio(vmf->page);
fs/netfs/fscache_io.c
185
folio_end_private_2(page_folio(page));
fs/nfs/file.c
648
struct folio *folio = page_folio(vmf->page);
fs/nfs/fscache.c
326
err = nfs_read_add_folio(&pgio, ctx, page_folio(page));
fs/nilfs2/file.c
49
struct folio *folio = page_folio(vmf->page);
fs/ntfs3/file.c
1052
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1063
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1074
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1118
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1136
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1159
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
956
folio = page_folio(pages[npages]);
fs/ntfs3/frecord.c
2483
struct folio *folio = page_folio(pages[0]);
fs/ocfs2/mmap.c
115
struct folio *folio = page_folio(vmf->page);
fs/orangefs/inode.c
607
struct folio *folio = page_folio(vmf->page);
fs/proc/task_mmu.c
1088
folio = page_folio(page);
fs/proc/task_mmu.c
1246
folio = page_folio(pte_page(ptent));
fs/proc/task_mmu.c
1985
folio = page_folio(page);
fs/proc/task_mmu.c
2050
folio = page_folio(page);
fs/proc/task_mmu.c
2140
struct folio *folio = page_folio(pte_page(pte));
fs/proc/task_mmu.c
3139
struct folio *folio = page_folio(page);
fs/proc/task_mmu.c
920
struct folio *folio = page_folio(page);
fs/splice.c
122
struct folio *folio = page_folio(buf->page);
fs/splice.c
67
struct folio *folio = page_folio(buf->page);
fs/squashfs/page_actor.c
65
return page_folio(actor->page[actor->next_page])->index;
fs/ubifs/file.c
1491
struct folio *folio = page_folio(vmf->page);
fs/udf/file.c
43
struct folio *folio = page_folio(vmf->page);
fs/xfs/xfs_file.c
1905
unsigned int len = folio_size(page_folio(vmf->page));
include/linux/bio.h
256
return page_folio(bio_first_page_all(bio));
include/linux/bio.h
286
fi->folio = page_folio(bvec->bv_page);
include/linux/huge_mm.h
805
return page_folio(ZERO_PAGE(0));
include/linux/leafops.h
395
softleaf_migration_sync(entry, page_folio(page));
include/linux/memcontrol.h
543
return folio_memcg_kmem(page_folio(page));
include/linux/memcontrol.h
943
memcg = folio_memcg(page_folio(page));
include/linux/memremap.h
176
folio_is_device_private(page_folio(page));
include/linux/memremap.h
201
folio_is_pci_p2pdma(page_folio(page));
include/linux/memremap.h
212
return folio_is_device_coherent(page_folio(page));
include/linux/memremap.h
223
return folio_is_fsdax(page_folio(page));
include/linux/mm.h
1625
return folio_mapped(page_folio(page));
include/linux/mm.h
1639
return page_folio(page);
include/linux/mm.h
1784
struct folio *folio = page_folio(page);
include/linux/mm.h
1881
struct folio *folio = page_folio(page);
include/linux/mm.h
2264
return page_folio(pfn_to_page(pfn));
include/linux/mm_inline.h
35
return folio_is_file_lru(page_folio(page));
include/linux/mmzone.h
1208
return page_folio(page)->pgmap;
include/linux/page-flags.h
1079
return folio_test_hugetlb(page_folio(page));
include/linux/page-flags.h
1093
folio = page_folio(page);
include/linux/page-flags.h
729
unsigned long flags = (unsigned long)page_folio(page)->mapping;
include/linux/page-flags.h
736
return folio_test_anon(page_folio(page));
include/linux/page-flags.h
804
return folio_test_uptodate(page_folio(page));
include/linux/page_ref.h
94
return folio_ref_count(page_folio(page));
include/linux/pagemap.h
1055
struct folio *folio = page_folio(page);
include/linux/pagemap.h
1130
return folio_trylock(page_folio(page));
include/linux/pagemap.h
1178
folio = page_folio(page);
include/linux/pagemap.h
645
folio_attach_private(page_folio(page), data);
include/linux/pagemap.h
650
return folio_detach_private(page_folio(page));
include/linux/pgtable.h
34
#define pmd_folio(pmd) page_folio(pmd_page(pmd))
include/linux/rmap.h
350
VM_WARN_ON_FOLIO(page_folio(page) != folio, folio);
include/linux/rmap.h
351
VM_WARN_ON_FOLIO(page_folio(page + nr_pages - 1) != folio, folio);
include/linux/swap.h
307
struct folio *folio = page_folio(page);
include/linux/vmstat.h
528
lruvec_stat_mod_folio(page_folio(page), idx, val);
io_uring/rsrc.c
105
struct folio *folio = page_folio(imu->bvec[i].bv_page);
io_uring/rsrc.c
692
struct folio *folio = page_folio(p);
io_uring/rsrc.c
718
struct folio *folio = page_folio(page_array[0]);
io_uring/rsrc.c
731
if (page_folio(page_array[i]) == folio &&
io_uring/rsrc.c
747
folio = page_folio(page_array[i]);
io_uring/rsrc.c
849
unpin_user_folio(page_folio(pages[i]), 1);
io_uring/zcrx.c
1256
if (folio_test_partial_kmap(page_folio(dst_page)) ||
io_uring/zcrx.c
1257
folio_test_partial_kmap(page_folio(src_page))) {
io_uring/zcrx.c
181
struct folio *folio = page_folio(pages[i]);
kernel/events/uprobes.c
532
folio = page_folio(page);
kernel/futex/core.c
669
folio = page_folio(page);
kernel/liveupdate/kexec_handover.c
290
return page ? page_folio(page) : NULL;
lib/iov_iter.c
1108
struct folio *folio = page_folio(page + k);
lib/iov_iter.c
1852
struct folio *folio = page_folio(pages[0]);
lib/iov_iter.c
1869
if (page_folio(pages[i]) != folio ||
lib/test_hmm.c
1405
order = folio_order(page_folio(spage));
lib/test_hmm.c
1422
folio_copy(page_folio(dpage), page_folio(spage));
lib/test_hmm.c
1590
struct folio *rfolio = page_folio(rpage);
lib/test_hmm.c
1609
mdevice->free_folios = page_folio(page);
lib/test_hmm.c
1631
rpage = folio_zone_device_data(page_folio(vmf->page));
lib/test_hmm.c
1635
order = folio_order(page_folio(vmf->page));
lib/test_hmm.c
1701
rfolio = page_folio(rpage);
lib/test_hmm.c
1720
folio_set_count(page_folio(rpage_tail), 1);
lib/test_hmm.c
583
mdevice->free_folios = page_folio(page);
lib/test_hmm.c
665
zone_device_folio_init(page_folio(dpage),
lib/test_hmm.c
666
page_pgmap(folio_page(page_folio(dpage), 0)),
lib/test_hmm.c
723
folio = page_folio(spage);
lib/test_hmm.c
763
struct folio *folio = page_folio(dpage);
lib/test_hmm.c
860
nr = folio_nr_pages(page_folio(dpage));
lib/test_hmm.c
867
VM_WARN_ON(folio_nr_pages(page_folio(rpage)) != nr);
lib/test_hmm.c
870
entry = folio_page(page_folio(rpage), i);
lib/test_hmm.c
969
order = folio_order(page_folio(spage));
lib/tests/kunit_iov_iter.c
396
folioq_append(p, page_folio(pages[i]));
mm/compaction.c
1061
folio = page_folio(page);
mm/compaction.c
971
folio = page_folio(page);
mm/damon/ops-common.c
34
folio = page_folio(page);
mm/damon/ops-common.c
37
if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) {
mm/filemap.c
3959
struct folio *folio = page_folio(vmf->page);
mm/folio-compat.c
16
return folio_unlock(page_folio(page));
mm/folio-compat.c
22
return folio_end_writeback(page_folio(page));
mm/folio-compat.c
28
return folio_wait_writeback(page_folio(page));
mm/folio-compat.c
34
folio_mark_accessed(page_folio(page));
mm/folio-compat.c
40
folio_start_writeback(page_folio(page));
mm/folio-compat.c
46
return folio_mark_dirty(page_folio(page));
mm/folio-compat.c
52
return folio_mark_dirty_lock(page_folio(page));
mm/folio-compat.c
58
return folio_clear_dirty_for_io(page_folio(page));
mm/folio-compat.c
65
return folio_redirty_for_writepage(wbc, page_folio(page));
mm/folio-compat.c
72
return filemap_add_folio(mapping, page_folio(page), index, gfp);
mm/gup.c
1072
ret = try_grab_folio(page_folio(*page), 1, gup_flags);
mm/gup.c
1478
struct folio *folio = page_folio(page);
mm/gup.c
188
gup_put_folio(page_folio(page), 1, FOLL_PIN);
mm/gup.c
2221
return page_folio(pofs->pages[i]);
mm/gup.c
236
struct folio *folio = page_folio(next);
mm/gup.c
250
struct folio *folio = page_folio(list[i]);
mm/gup.c
254
if (page_folio(list[nr]) != folio)
mm/gup.c
541
return page_folio(page);
mm/gup.c
56
folio = page_folio(page);
mm/gup.c
674
ret = try_grab_folio(page_folio(page), 1, flags);
mm/gup.c
731
ret = try_grab_folio(page_folio(page), 1, flags);
mm/gup.c
79
folio = page_folio(page);
mm/gup.c
847
folio = page_folio(page);
mm/gup.c
94
if (unlikely(page_folio(page) != folio)) {
mm/gup_test.c
51
folio = page_folio(pages[i]);
mm/huge_memory.c
1397
folio = page_folio(page);
mm/huge_memory.c
1923
src_folio = page_folio(src_page);
mm/huge_memory.c
2090
folio = page_folio(page);
mm/huge_memory.c
2380
folio = page_folio(page);
mm/huge_memory.c
2749
src_folio = page_folio(src_page);
mm/huge_memory.c
2879
folio = page_folio(page);
mm/huge_memory.c
2909
folio = page_folio(page);
mm/huge_memory.c
3028
folio = page_folio(page);
mm/huge_memory.c
3059
folio = page_folio(page);
mm/huge_memory.c
3075
folio = page_folio(page);
mm/huge_memory.c
3127
folio = page_folio(page);
mm/huge_memory.c
3685
folio = page_folio(split_at);
mm/huge_memory.c
3963
if (folio != page_folio(split_at) || folio != page_folio(lock_at)) {
mm/huge_memory.c
4086
if (new_folio == page_folio(lock_at))
mm/huge_memory.c
4207
struct folio *folio = page_folio(page);
mm/huge_memory.c
4546
folio = page_folio(page);
mm/huge_memory.c
4550
if (unlikely(page_folio(page) != folio))
mm/huge_memory.c
4865
struct folio *folio = page_folio(page);
mm/huge_memory.c
4917
struct folio *folio = page_folio(new);
mm/hugetlb.c
2837
struct folio *folio = page_folio(page);
mm/hugetlb.c
4981
pte_folio = page_folio(pte_page(entry));
mm/hugetlb.c
5248
if (folio != page_folio(pte_page(pte))) {
mm/hugetlb.c
5259
folio = page_folio(pte_page(pte));
mm/hugetlb.c
5480
old_folio = page_folio(pte_page(pte));
mm/hugetlb.c
6114
folio = page_folio(pte_page(vmf.orig_pte));
mm/hugetlb_cma.c
54
folio = page_folio(page);
mm/hwpoison-inject.c
116
folio = page_folio(p);
mm/hwpoison-inject.c
22
struct folio *folio = page_folio(p);
mm/internal.h
347
VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio);
mm/khugepaged.c
1315
folio = page_folio(page);
mm/khugepaged.c
571
folio = page_folio(page);
mm/khugepaged.c
700
src = page_folio(src_page);
mm/ksm.c
1375
struct folio *kfolio = page_folio(kpage);
mm/ksm.c
1377
struct folio *folio = page_folio(page);
mm/ksm.c
1478
struct folio *folio = page_folio(page);
mm/ksm.c
1500
folio = page_folio(page);
mm/ksm.c
1636
return err ? NULL : page_folio(page);
mm/ksm.c
1835
folio = page_folio(page);
mm/ksm.c
2250
struct folio *folio = page_folio(page);
mm/ksm.c
2382
folio = page_folio(page);
mm/ksm.c
2522
folio = page_folio(page);
mm/ksm.c
2546
folio = page_folio(page);
mm/memcontrol.c
284
memcg = folio_memcg_check(page_folio(page));
mm/memory-failure.c
1022
struct folio *folio = page_folio(p);
mm/memory-failure.c
1080
struct folio *folio = page_folio(p);
mm/memory-failure.c
1118
struct folio *folio = page_folio(p);
mm/memory-failure.c
1140
struct folio *folio = page_folio(p);
mm/memory-failure.c
1162
struct folio *folio = page_folio(p);
mm/memory-failure.c
1356
struct folio *folio = page_folio(page);
mm/memory-failure.c
1363
if (folio == page_folio(page))
mm/memory-failure.c
1367
folio = page_folio(page);
mm/memory-failure.c
1380
if (folio == page_folio(page))
mm/memory-failure.c
1455
struct folio *folio = page_folio(page);
mm/memory-failure.c
1462
if (folio == page_folio(page))
mm/memory-failure.c
1841
struct folio *folio = page_folio(page);
mm/memory-failure.c
187
ret = dissolve_free_hugetlb_folio(page_folio(page));
mm/memory-failure.c
1976
struct folio *folio = page_folio(page);
mm/memory-failure.c
2071
folio = page_folio(p);
mm/memory-failure.c
2439
folio = page_folio(p);
mm/memory-failure.c
2479
folio = page_folio(p);
mm/memory-failure.c
2487
folio = page_folio(p);
mm/memory-failure.c
2690
folio = page_folio(p);
mm/memory-failure.c
2794
struct folio *folio = page_folio(page);
mm/memory-failure.c
2820
folio = page_folio(page);
mm/memory-failure.c
334
shake_folio(page_folio(page));
mm/memory-failure.c
400
tk->size_shift = folio_shift(page_folio(p));
mm/memory-failure.c
984
count -= folio_nr_pages(page_folio(p));
mm/memory.c
1139
folio = page_folio(page);
mm/memory.c
1708
folio = page_folio(page);
mm/memory.c
1743
struct folio *folio = page_folio(page);
mm/memory.c
2326
struct folio *folio = page_folio(page);
mm/memory.c
2345
struct folio *folio = page_folio(page);
mm/memory.c
3615
struct folio *folio = page_folio(vmf->page);
mm/memory.c
3774
old_folio = page_folio(vmf->page);
mm/memory.c
4191
folio = page_folio(vmf->page);
mm/memory.c
4377
struct folio *folio = page_folio(vmf->page);
mm/memory.c
5369
folio = page_folio(vmf->page);
mm/memory.c
5577
folio = page_folio(page);
mm/memory.c
5804
folio = page_folio(vmf->page);
mm/memory.c
5867
folio = page_folio(vmf->page);
mm/memory.c
7010
folio = page_folio(page);
mm/memory.c
7120
folio = page_folio(page);
mm/memory.c
789
return page_folio(page);
mm/memory.c
831
return page_folio(page);
mm/memory.c
920
struct folio *folio = page_folio(page);
mm/memory.c
990
folio = page_folio(page);
mm/memory_hotplug.c
1774
folio = page_folio(page);
mm/memory_hotplug.c
1804
folio = page_folio(page);
mm/memory_hotplug.c
1809
if (unlikely(page_folio(page) != folio))
mm/migrate.c
199
struct folio *folio = page_folio(page);
mm/migrate_device.c
103
new_fault_folio = fault_page ? page_folio(fault_page) : NULL;
mm/migrate_device.c
1163
newfolio = page_folio(newpage);
mm/migrate_device.c
1164
folio = page_folio(page);
mm/migrate_device.c
1224
folio = page_folio(migrate_pfn_to_page(src_pfns[i+j]));
mm/migrate_device.c
1225
newfolio = page_folio(migrate_pfn_to_page(dst_pfns[i+j]));
mm/migrate_device.c
1277
page_folio(fault_page) : NULL;
mm/migrate_device.c
1286
dst = page_folio(newpage);
mm/migrate_device.c
1297
src = page_folio(page);
mm/migrate_device.c
1400
struct folio *folio = page_folio(page);
mm/migrate_device.c
1434
struct folio *folio = page_folio(page);
mm/migrate_device.c
258
page_folio(migrate->fault_page) : NULL;
mm/migrate_device.c
312
folio = page_folio(page);
mm/migrate_device.c
355
folio = page ? page_folio(page) : NULL;
mm/migrate_device.c
391
folio = page_folio(page);
mm/migrate_device.c
539
struct folio *folio = page_folio(page);
mm/migrate_device.c
571
page_folio(fault_page) : NULL;
mm/migrate_device.c
589
folio = page_folio(page);
mm/migrate_device.c
641
folio = page_folio(page);
mm/migrate_device.c
801
struct folio *folio = page_folio(page);
mm/migrate_device.c
86
struct folio *fault_folio = fault_page ? page_folio(fault_page) : NULL;
mm/migrate_device.c
974
struct folio *folio = page_folio(page);
mm/mm_init.c
1025
page_folio(page)->pgmap = pgmap;
mm/mmu_gather.c
178
VM_WARN_ON_ONCE(page_folio(page) != page_folio(page + nr_pages - 1));
mm/mmu_gather.c
68
folio_remove_rmap_ptes(page_folio(page), page, nr_pages,
mm/mprotect.c
253
folio = page_folio(page);
mm/page_alloc.c
1349
struct folio *folio = page_folio(page);
mm/page_alloc.c
3125
pgalloc_tag_split(page_folio(page), order, 0);
mm/page_idle.c
42
folio = page_folio(page);
mm/page_idle.c
45
if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) {
mm/page_io.c
489
struct folio *folio = page_folio(sio->bvec[p].bv_page);
mm/page_io.c
499
struct folio *folio = page_folio(sio->bvec[p].bv_page);
mm/page_isolation.c
45
struct folio *folio = page_folio(page);
mm/page_vma_mapped.c
351
const struct folio *folio = page_folio(page);
mm/pagewalk.c
1049
return page_folio(page);
mm/rmap.c
2838
folio = page_folio(page);
mm/swap.c
1028
struct folio *folio = page_folio(encoded_page_ptr(encoded[i]));
mm/swap_state.c
414
struct folio *folio = page_folio(encoded_page_ptr(pages[i]));
mm/swapfile.c
434
folio = page_folio(container_of(head, struct page, rcu_head));
mm/truncate.c
268
folio2 = page_folio(split_at2);
mm/userfaultfd.c
179
struct folio *folio = page_folio(page);
mm/util.c
1255
set_ps_flags(ps, page_folio(page), page);
mm/zswap.c
897
folio_memcg(page_folio(page)))) {
net/sunrpc/svc.c
960
page_folio(*rqstp->rq_next_page)))