folio_page
void *addr = kmap_high_get(folio_page(folio, i));
kunmap_high(folio_page(folio, i));
from != folio_page(src, 0))
kfrom = page_address(folio_page(src, i));
kto = page_address(folio_page(dst, i));
err = mte_save_tags(folio_page(folio, i));
__mte_invalidate_tags(folio_page(folio, i));
mte_restore_tags(entry, folio_page(folio, i));
__flush_dcache_folio_pages(folio, folio_page(folio, 0),
sync_icache_dcache(folio_page(folio, nr));
__bio_add_page(bio, folio_page(folio, nr), len, off % PAGE_SIZE);
return bio_add_page(bio, folio_page(folio, nr), len, off % PAGE_SIZE) > 0;
pages[pg] = folio_page(ubuf->folios[pg],
page = folio_page(folio, 0);
struct page *dpage = folio_page(dfolio, 0);
struct page *spage = folio_page(sfolio, 0);
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER | __GFP_ZERO,
zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order);
dpage = folio_page(
pages[pg++] = folio_page(folio, i);
unsigned long pfn = page_to_pfn(folio_page(folio, offset));
*upages++ = folio_page(folio, offset);
*upages++ = folio_page(folio, offset++);
__free_page(folio_page(eb->folios[i], 0));
__free_page(folio_page(eb->folios[i], 0));
if (i && folio_page(eb->folios[i - 1], 0) + 1 != folio_page(folio, 0))
paddrs[i] = page_to_phys(folio_page(folio, pindex)) + offset_in_page(paddr);
return page_to_phys(folio_page(folio, 0)) + offset_in_folio(folio, offset);
dump_page(folio_page(folio, 0), "btrfs folio state dump");
struct page *page = folio_page(eb->folios[i >> PAGE_SHIFT], 0);
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
dump_page(folio_page(eb->folios[0], 0), "eb page dump");
pages[0] = folio_page(folio, 0);
struct page *page = folio_page(folio, i >> PAGE_SHIFT);
struct page *page = folio_page(folio, i);
folio_page(folio, 0), folio->index,
struct page *page = folio_page(folio, 0);
.page = folio_page(folio, pgs >> PAGE_SHIFT),
folio_page(folio ?: newfolio, 0);
cc->rpages[cluster_ofs] = folio_page(folio, 0);
pages[nr_pages] = folio_page(folio, idx);
(page_address(folio_page(folio, idx_in_folio)) +
memcpy_to_page(folio_page(folio, i), 0, src, PAGE_SIZE);
gfs2_log_write(sdp, sdp->sd_jdesc, folio_page(bh->b_folio, 0),
struct page *page = folio_page(folio, offset / PAGE_SIZE);
if (!smb_set_sge(rdma, folio_page(folio, 0), offset, part))
bvec_set_page(bv, folio_page(folio, nr), len, offset % PAGE_SIZE);
const struct page *page = folio_page(folio, offset / PAGE_SIZE);
return (struct folio *)folio_page(folio, folio_nr_pages(folio));
return folio_page(folio, pgbase >> PAGE_SHIFT);
array[i++] = folio_page(folio, 0);
return folio_page(folio, index & (folio_nr_pages(folio) - 1));
if (get_page_tag_ref(folio_page(folio, i), &ref, &handle)) {
p[nr++] = folio_page(folio, offset / PAGE_SIZE);
*pages = folio_page(folio, offset / PAGE_SIZE);
sg_set_page(sg, folio_page(folio, 0), part, offset);
sg_set_page(sg, folio_page(folio, 0), len, offset);
dpage = folio_page(folio_alloc(GFP_HIGHUSER_MOVABLE,
if (dmirror_page_to_chunk(folio_page(folio, 0)) == devmem)
struct page *rpage = BACKING_PAGE(folio_page(head, 0));
rpage_tail = folio_page(rfolio, offset);
folio_page(tail, 0)->mapping = folio_page(head, 0)->mapping;
*ppage = folio_page(mdevice->free_folios, 0);
rpage = folio_page(folio_alloc(GFP_HIGHUSER, order), 0);
dpage = folio_page(mdevice->free_folios, 0);
page_pgmap(folio_page(page_folio(dpage), 0)),
entry = folio_page(page_folio(rpage), i);
dpage = folio_page(vma_alloc_folio(GFP_HIGHUSER_MOVABLE,
page = folio_page(folio, offset / PAGE_SIZE);
struct page *page = folio_page(folio, start);
page_range_has_hwpoisoned(folio_page(folio, 0), new_nr_pages))
if (pages_identical(folio_page(folio, i), ZERO_PAGE(0))) {
struct page *split_at = folio_page(folio,
struct page *split_at = folio_page(folio,
struct page *page = folio_page(folio, start_page_number);
init_cma_pageblock(folio_page(folio, i));
init_pageblock_migratetype(folio_page(folio, i),
struct page *page = folio_page(folio, i);
tlb_remove_page_size(tlb, folio_page(folio, 0),
if (folio_page(folio, i) != page)
if (folio_page(folio, i) != page)
dst = folio_page(new_folio, 0);
if (copy_mc_highpage(dst, folio_page(folio, i)) > 0) {
struct page *page = folio_page(folio, i);
pvmw.address = page_address_in_vma(folio, folio_page(folio, 0), vma);
struct page *page = folio_page(folio, 0);
if (copy_mc_user_highpage(folio_page(new_folio, 0), page,
page = folio_page(folio, 0);
struct page *page = folio_page(folio, r[i].start);
dst_page = folio_page(dst, i);
src_page = folio_page(src, i);
struct page *dst = folio_page(copy_arg->dst, idx);
struct page *src = folio_page(copy_arg->src, idx);
subpage = folio_page(dst_folio, i);
folio_page(folio, 0), vma);
__ClearPageAnonExclusive(folio_page(folio, i));
struct page *page = folio_page(folio, idx);
new = folio_page(folio, idx);
unsigned long pfn = page_to_pfn(folio_page(folio, 0));
ret = set_pmd_migration_entry(&pvmw, folio_page(folio, 0));
entry = page_swap_entry(folio_page(folio, i));
entry = page_swap_entry(folio_page(folio, i));
struct page *page = folio_page(folio, i);
subpage = folio_page(folio, pfn - folio_pfn(folio));
subpage = folio_page(folio, pfn - folio_pfn(folio));
subpage = folio_page(folio, pfn - folio_pfn(folio));
set_direct_map_default_noflush(folio_page(folio, 0));
err = set_direct_map_invalid_noflush(folio_page(folio, 0));
set_direct_map_default_noflush(folio_page(folio, 0));
clear_highpage(folio_page(folio, i));
if (PageHWPoison(folio_page(folio, 0)) ||
if (PageHWPoison(folio_page(folio, 0))) {
#define slab_page(s) folio_page(slab_folio(s), 0)
call_rcu(&(folio_page(virt_to_folio(table), 0)->rcu_head),
split_at = folio_page(folio, PAGE_ALIGN_DOWN(offset) / PAGE_SIZE);
split_at2 = folio_page(folio,
flush_dcache_page(folio_page(folio, i));
copy_highpage(folio_page(dst, i), folio_page(src, i));
if (copy_mc_highpage(folio_page(dst, i), folio_page(src, i)))
struct page *page = folio_page(folio, index);
clear_highpage(folio_page(folio, 0));
struct page *page = folio_page(folio, 0);
clear_highpage(folio_page(folio, 0));