Symbol: folio_page
arch/arm/mm/flush.c
220
void *addr = kmap_high_get(folio_page(folio, i));
arch/arm/mm/flush.c
223
kunmap_high(folio_page(folio, i));
arch/arm64/mm/copypage.c
35
from != folio_page(src, 0))
arch/arm64/mm/copypage.c
48
kfrom = page_address(folio_page(src, i));
arch/arm64/mm/copypage.c
49
kto = page_address(folio_page(dst, i));
arch/arm64/mm/mteswap.c
105
err = mte_save_tags(folio_page(folio, i));
arch/arm64/mm/mteswap.c
113
__mte_invalidate_tags(folio_page(folio, i));
arch/arm64/mm/mteswap.c
127
mte_restore_tags(entry, folio_page(folio, i));
arch/mips/include/asm/cacheflush.h
59
__flush_dcache_folio_pages(folio, folio_page(folio, 0),
arch/openrisc/mm/cache.c
97
sync_icache_dcache(folio_page(folio, nr));
block/bio.c
1095
__bio_add_page(bio, folio_page(folio, nr), len, off % PAGE_SIZE);
block/bio.c
1120
return bio_add_page(bio, folio_page(folio, nr), len, off % PAGE_SIZE) > 0;
drivers/dma-buf/udmabuf.c
123
pages[pg] = folio_page(ubuf->folios[pg],
drivers/gpu/drm/drm_pagemap.c
734
page = folio_page(folio, 0);
drivers/gpu/drm/nouveau/nouveau_dmem.c
160
struct page *dpage = folio_page(dfolio, 0);
drivers/gpu/drm/nouveau/nouveau_dmem.c
161
struct page *spage = folio_page(sfolio, 0);
drivers/gpu/drm/nouveau/nouveau_dmem.c
242
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER | __GFP_ZERO,
drivers/gpu/drm/nouveau/nouveau_dmem.c
428
zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order);
drivers/gpu/drm/nouveau/nouveau_dmem.c
497
dpage = folio_page(
drivers/gpu/drm/panfrost/panfrost_mmu.c
669
pages[pg++] = folio_page(folio, i);
drivers/iommu/iommufd/pages.c
673
unsigned long pfn = page_to_pfn(folio_page(folio, offset));
drivers/iommu/iommufd/pages.c
858
*upages++ = folio_page(folio, offset);
drivers/iommu/iommufd/pages.c
866
*upages++ = folio_page(folio, offset++);
fs/btrfs/extent_io.c
3368
__free_page(folio_page(eb->folios[i], 0));
fs/btrfs/extent_io.c
3383
__free_page(folio_page(eb->folios[i], 0));
fs/btrfs/extent_io.c
3520
if (i && folio_page(eb->folios[i - 1], 0) + 1 != folio_page(folio, 0))
fs/btrfs/inode.c
3477
paddrs[i] = page_to_phys(folio_page(folio, pindex)) + offset_in_page(paddr);
fs/btrfs/scrub.c
709
return page_to_phys(folio_page(folio, 0)) + offset_in_folio(folio, offset);
fs/btrfs/subpage.c
802
dump_page(folio_page(folio, 0), "btrfs folio state dump");
fs/btrfs/tests/extent-io-tests.c
673
struct page *page = folio_page(eb->folios[i >> PAGE_SHIFT], 0);
fs/btrfs/tree-checker.c
1070
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/btrfs/tree-checker.c
1343
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/btrfs/tree-checker.c
157
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/btrfs/tree-checker.c
673
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/btrfs/tree-checker.c
68
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/btrfs/tree-checker.c
96
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
fs/ceph/addr.c
2280
pages[0] = folio_page(folio, 0);
fs/crypto/crypto.c
274
struct page *page = folio_page(folio, i >> PAGE_SHIFT);
fs/dax.c
402
struct page *page = folio_page(folio, i);
fs/ecryptfs/crypto.c
376
folio_page(folio, 0), folio->index,
fs/ecryptfs/crypto.c
449
struct page *page = folio_page(folio, 0);
fs/erofs/zdata.c
1073
.page = folio_page(folio, pgs >> PAGE_SHIFT),
fs/erofs/zdata.c
575
folio_page(folio ?: newfolio, 0);
fs/f2fs/compress.c
171
cc->rpages[cluster_ofs] = folio_page(folio, 0);
fs/f2fs/data.c
3310
pages[nr_pages] = folio_page(folio, idx);
fs/f2fs/f2fs.h
2209
(page_address(folio_page(folio, idx_in_folio)) +
fs/freevxfs/vxfs_immed.c
38
memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE);
fs/gfs2/lops.c
361
gfs2_log_write(sdp, sdp->sd_jdesc, folio_page(bh->b_folio, 0),
fs/hugetlbfs/inode.c
218
struct page *page = folio_page(folio, offset / PAGE_SIZE);
fs/smb/client/smbdirect.c
3272
if (!smb_set_sge(rdma, folio_page(folio, 0), offset, part))
include/linux/bvec.h
62
bvec_set_page(bv, folio_page(folio, nr), len, offset % PAGE_SIZE);
include/linux/highmem-internal.h
86
const struct page *page = folio_page(folio, offset / PAGE_SIZE);
include/linux/mm.h
2542
return (struct folio *)folio_page(folio, folio_nr_pages(folio));
include/linux/nfs_page.h
195
return folio_page(folio, pgbase >> PAGE_SHIFT);
include/linux/pagemap.h
1477
array[i++] = folio_page(folio, 0);
include/linux/pagemap.h
972
return folio_page(folio, index & (folio_nr_pages(folio) - 1));
lib/alloc_tag.c
184
if (get_page_tag_ref(folio_page(folio, i), &ref, &handle)) {
lib/iov_iter.c
1536
p[nr++] = folio_page(folio, offset / PAGE_SIZE);
lib/iov_iter.c
946
*pages = folio_page(folio, offset / PAGE_SIZE);
lib/scatterlist.c
1302
sg_set_page(sg, folio_page(folio, 0), part, offset);
lib/scatterlist.c
1363
sg_set_page(sg, folio_page(folio, 0), len, offset);
lib/test_hmm.c
1408
dpage = folio_page(folio_alloc(GFP_HIGHUSER_MOVABLE,
lib/test_hmm.c
1439
if (dmirror_page_to_chunk(folio_page(folio, 0)) == devmem)
lib/test_hmm.c
1691
struct page *rpage = BACKING_PAGE(folio_page(head, 0));
lib/test_hmm.c
1712
rpage_tail = folio_page(rfolio, offset);
lib/test_hmm.c
1718
folio_page(tail, 0)->mapping = folio_page(head, 0)->mapping;
lib/test_hmm.c
600
*ppage = folio_page(mdevice->free_folios, 0);
lib/test_hmm.c
643
rpage = folio_page(folio_alloc(GFP_HIGHUSER, order), 0);
lib/test_hmm.c
650
dpage = folio_page(mdevice->free_folios, 0);
lib/test_hmm.c
666
page_pgmap(folio_page(page_folio(dpage), 0)),
lib/test_hmm.c
870
entry = folio_page(page_folio(rpage), i);
lib/test_hmm.c
980
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER_MOVABLE,
mm/filemap.c
3010
page = folio_page(folio, offset / PAGE_SIZE);
mm/filemap.c
3755
struct page *page = folio_page(folio, start);
mm/huge_memory.c
3496
page_range_has_hwpoisoned(folio_page(folio, 0), new_nr_pages))
mm/huge_memory.c
4392
if (pages_identical(folio_page(folio, i), ZERO_PAGE(0))) {
mm/huge_memory.c
4674
struct page *split_at = folio_page(folio,
mm/huge_memory.c
4750
struct page *split_at = folio_page(folio,
mm/hugetlb.c
3176
struct page *page = folio_page(folio, start_page_number);
mm/hugetlb.c
3239
init_cma_pageblock(folio_page(folio, i));
mm/hugetlb.c
3241
init_pageblock_migratetype(folio_page(folio, i),
mm/hugetlb.c
4005
struct page *page = folio_page(folio, i);
mm/hugetlb.c
5315
tlb_remove_page_size(tlb, folio_page(folio, 0),
mm/khugepaged.c
1558
if (folio_page(folio, i) != page)
mm/khugepaged.c
1614
if (folio_page(folio, i) != page)
mm/khugepaged.c
2104
dst = folio_page(new_folio, 0);
mm/khugepaged.c
2115
if (copy_mc_highpage(dst, folio_page(folio, i)) > 0) {
mm/khugepaged.c
786
struct page *page = folio_page(folio, i);
mm/ksm.c
1286
pvmw.address = page_address_in_vma(folio, folio_page(folio, 0), vma);
mm/ksm.c
3107
struct page *page = folio_page(folio, 0);
mm/ksm.c
3136
if (copy_mc_user_highpage(folio_page(new_folio, 0), page,
mm/memory.c
4864
page = folio_page(folio, 0);
mm/memory.c
7361
struct page *page = folio_page(folio, r[i].start);
mm/memory.c
7379
dst_page = folio_page(dst, i);
mm/memory.c
7380
src_page = folio_page(src, i);
mm/memory.c
7399
struct page *dst = folio_page(copy_arg->dst, idx);
mm/memory.c
7400
struct page *src = folio_page(copy_arg->src, idx);
mm/memory.c
7434
subpage = folio_page(dst_folio, i);
mm/mempolicy.c
1595
folio_page(folio, 0), vma);
mm/memremap.c
429
__ClearPageAnonExclusive(folio_page(folio, i));
mm/migrate.c
301
struct page *page = folio_page(folio, idx);
mm/migrate.c
363
new = folio_page(folio, idx);
mm/migrate_device.c
210
unsigned long pfn = page_to_pfn(folio_page(folio, 0));
mm/migrate_device.c
217
ret = set_pmd_migration_entry(&pvmw, folio_page(folio, 0));
mm/page_io.c
213
entry = page_swap_entry(folio_page(folio, i));
mm/page_io.c
231
entry = page_swap_entry(folio_page(folio, i));
mm/rmap.c
1663
struct page *page = folio_page(folio, i);
mm/rmap.c
2104
subpage = folio_page(folio, pfn - folio_pfn(folio));
mm/rmap.c
2479
subpage = folio_page(folio, pfn - folio_pfn(folio));
mm/rmap.c
2510
subpage = folio_page(folio, pfn - folio_pfn(folio));
mm/secretmem.c
154
set_direct_map_default_noflush(folio_page(folio, 0));
mm/secretmem.c
75
err = set_direct_map_invalid_noflush(folio_page(folio, 0));
mm/secretmem.c
90
set_direct_map_default_noflush(folio_page(folio, 0));
mm/shmem.c
2617
clear_highpage(folio_page(folio, i));
mm/shmem.c
4182
if (PageHWPoison(folio_page(folio, 0)) ||
mm/shmem.c
4193
if (PageHWPoison(folio_page(folio, 0))) {
mm/slab.h
153
#define slab_page(s) folio_page(slab_folio(s), 0)
mm/swapfile.c
445
call_rcu(&(folio_page(virt_to_folio(table), 0)->rcu_head),
mm/truncate.c
254
split_at = folio_page(folio, PAGE_ALIGN_DOWN(offset) / PAGE_SIZE);
mm/truncate.c
266
split_at2 = folio_page(folio,
mm/util.c
1132
flush_dcache_page(folio_page(folio, i));
mm/util.c
727
copy_highpage(folio_page(dst, i), folio_page(src, i));
mm/util.c
741
if (copy_mc_highpage(folio_page(dst, i), folio_page(src, i)))
mm/zswap.c
1531
struct page *page = folio_page(folio, index);
virt/kvm/guest_memfd.c
417
clear_highpage(folio_page(folio, 0));
virt/kvm/guest_memfd.c
528
struct page *page = folio_page(folio, 0);
virt/kvm/guest_memfd.c
805
clear_highpage(folio_page(folio, 0));