page_pgmap
zone_device_page_init(page, page_pgmap(page), 0);
if (pagemap != page_pgmap(page)) {
pagemap = page_pgmap(page);
cur = page_pgmap(page);
.pgmap_owner = page_pgmap(page)->owner,
zone_device_page_init(page, page_pgmap(page), 0);
if (page_pgmap(src_page) == pagemap)
if (page_pgmap(src_page) == pagemap &&
zone_device_folio_init(folio, page_pgmap(folio_page(folio, 0)), order);
return container_of(page_pgmap(page), struct nouveau_dmem_chunk,
return container_of(page_pgmap(page), struct xe_pagemap, pagemap);
struct pci_p2pdma_pagemap *p2p_pgmap = to_p2p_pgmap(page_pgmap(page));
struct pci_p2pdma_pagemap *pgmap = to_p2p_pgmap(page_pgmap(page));
struct dev_pagemap *pgmap = page_pgmap(&folio->page);
return page_pgmap(a) == page_pgmap(b);
return container_of(page_pgmap(page), struct dmirror_chunk,
page_pgmap(folio_page(page_folio(dpage), 0)),
page_pgmap(softleaf_to_page(entry))->owner ==
ret = page_pgmap(page)->ops->migrate_to_ram(vmf);
pgmap = page_pgmap(vmf->page);
WARN_ON_ONCE(!percpu_ref_tryget_many(&page_pgmap(page)->ref, 1 << order));
pgmap = page_pgmap(page);
pgmap = page_pgmap(page);