arch/alpha/include/asm/page.h
81
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/alpha/include/asm/pgtable.h
172
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/alpha/include/asm/pgtable.h
197
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> 32))
arch/alpha/include/asm/pgtable.h
198
#define pud_page(pud) (pfn_to_page(pud_val(pud) >> 32))
arch/arc/include/asm/page.h
130
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
arch/arc/include/asm/pgtable-levels.h
175
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/arc/mm/tlb.c
477
struct page *page = pfn_to_page(pte_pfn(*ptep));
arch/arm/include/asm/kexec.h
77
return pfn_to_page(boot_pfn - (arch_phys_to_idmap_offset >> PAGE_SHIFT));
arch/arm/include/asm/memory.h
384
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
arch/arm/include/asm/pgtable.h
165
#define pmd_page(pmd) pfn_to_page(__phys_to_pfn(pmd_val(pmd) & PHYS_MASK))
arch/arm/include/asm/pgtable.h
170
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/arm/mm/fault-armv.c
204
folio = page_folio(pfn_to_page(pfn));
arch/arm/mm/fault.c
91
if (PageHighMem(pfn_to_page(pmd_val(*pmd) >> PAGE_SHIFT)))
arch/arm/mm/flush.c
298
folio = page_folio(pfn_to_page(pfn));
arch/arm64/include/asm/memory.h
407
#define virt_to_page(x) pfn_to_page(virt_to_pfn(x))
arch/arm64/include/asm/pgtable.h
1054
#define p4d_page(p4d) pfn_to_page(__phys_to_pfn(__p4d_to_phys(p4d)))
arch/arm64/include/asm/pgtable.h
1176
#define pgd_page(pgd) pfn_to_page(__phys_to_pfn(__pgd_to_phys(pgd)))
arch/arm64/include/asm/pgtable.h
147
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/arm64/kernel/vdso.c
92
vdso_pagelist[i] = pfn_to_page(pfn + i);
arch/arm64/kvm/mmu.c
1474
struct page *page = pfn_to_page(pfn);
arch/csky/abiv1/cacheflush.c
58
folio = page_folio(pfn_to_page(pfn));
arch/csky/abiv2/cacheflush.c
22
folio = page_folio(pfn_to_page(pfn));
arch/csky/include/asm/pgtable.h
33
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/csky/include/asm/pgtable.h
45
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/hexagon/include/asm/page.h
90
#define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
arch/hexagon/include/asm/pgtable.h
222
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/hexagon/include/asm/pgtable.h
242
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/loongarch/include/asm/page.h
101
#define pfn_to_virt(pfn) page_to_virt(pfn_to_page(pfn))
arch/loongarch/include/asm/pgtable.h
184
#define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
230
#define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
268
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/loongarch/include/asm/pgtable.h
275
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/loongarch/include/asm/pgtable.h
579
return pfn_to_page(pmd_pfn(pmd));
arch/loongarch/include/asm/pgtable.h
581
return pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT);
arch/loongarch/kernel/vdso.c
62
vdso_info.code_mapping.pages[i] = pfn_to_page(pfn + i);
arch/loongarch/mm/init.c
96
struct page *page = pfn_to_page(start_pfn);
arch/m68k/include/asm/mcf_pgtable.h
290
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/m68k/include/asm/page_mm.h
136
pfn_to_page(virt_to_pfn(addr)); \
arch/microblaze/include/asm/page.h
98
# define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT))
arch/microblaze/include/asm/pgtable.h
377
#define pmd_page(pmd) (pfn_to_page(__pa(pmd_val(pmd)) >> PAGE_SHIFT))
arch/mips/include/asm/page.h
210
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
arch/mips/include/asm/pgtable-32.h
192
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/mips/include/asm/pgtable-64.h
213
#define p4d_page(p4d) (pfn_to_page(p4d_phys(p4d) >> PAGE_SHIFT))
arch/mips/include/asm/pgtable-64.h
299
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/mips/include/asm/pgtable-64.h
311
#define pud_page(pud) (pfn_to_page(pud_phys(pud) >> PAGE_SHIFT))
arch/mips/include/asm/pgtable.h
421
return pfn_to_page(pmd_pfn(pmd));
arch/mips/include/asm/pgtable.h
423
return pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT);
arch/mips/include/asm/pgtable.h
73
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/mips/kernel/vdso.c
41
image->mapping.pages[i] = pfn_to_page(data_pfn + i);
arch/mips/mm/c-r4k.c
542
struct page *page = pfn_to_page(fcp_args->pfn);
arch/mips/mm/cache.c
157
folio = page_folio(pfn_to_page(pfn));
arch/mips/mm/dma-noncoherent.c
98
struct page *page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/mips/mm/init.c
436
struct page *page = pfn_to_page(tmp);
arch/mips/mm/init.c
476
struct page *page = pfn_to_page(pfn);
arch/mips/mm/ioremap.c
32
!PageReserved(pfn_to_page(start_pfn + i)))
arch/nios2/include/asm/page.h
85
# define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr)))
arch/nios2/include/asm/pgtable.h
174
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/nios2/include/asm/pgtable.h
226
#define pmd_page(pmd) (pfn_to_page(pmd_phys(pmd) >> PAGE_SHIFT))
arch/nios2/mm/cacheflush.c
229
folio = page_folio(pfn_to_page(pfn));
arch/openrisc/include/asm/pgtable.h
331
#define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT))
arch/openrisc/mm/cache.c
85
struct folio *folio = page_folio(pfn_to_page(pfn));
arch/parisc/include/asm/page.h
168
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/parisc/include/asm/pgtable.h
364
#define pte_page(pte) (pfn_to_page(pte_pfn(pte)))
arch/parisc/kernel/cache.c
121
folio = page_folio(pfn_to_page(pfn));
arch/powerpc/include/asm/book3s/32/pgtable.h
346
#define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd))
arch/powerpc/include/asm/kvm_ppc.h
941
folio = page_folio(pfn_to_page(pfn));
arch/powerpc/include/asm/nohash/32/pgtable.h
175
#define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd))
arch/powerpc/include/asm/page.h
231
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
arch/powerpc/include/asm/pgtable.h
47
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/powerpc/kernel/crash_dump.c
131
free_reserved_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/powerpc/kernel/fadump.c
1189
free_reserved_page(pfn_to_page(pfn));
arch/powerpc/kernel/iommu.c
1109
SetPageDirty(pfn_to_page(*hpa >> PAGE_SHIFT));
arch/powerpc/kvm/book3s.c
440
*page = pfn_to_page(pfn);
arch/powerpc/kvm/book3s_hv_uvmem.c
639
uvmem_page = pfn_to_page(uvmem_pfn);
arch/powerpc/kvm/book3s_hv_uvmem.c
724
dpage = pfn_to_page(uvmem_pfn);
arch/powerpc/kvm/book3s_hv_uvmem.c
891
uvmem_page = pfn_to_page(uvmem_pfn);
arch/powerpc/kvm/book3s_hv_uvmem.c
909
uvmem_page = pfn_to_page(uvmem_pfn);
arch/powerpc/kvm/e500_mmu_host.c
170
get_page(pfn_to_page(pfn));
arch/powerpc/kvm/e500_mmu_host.c
666
page = pfn_to_page(pfn);
arch/powerpc/mm/book3s64/iommu_api.c
214
page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT);
arch/powerpc/mm/dma-noncoherent.c
96
struct page *page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/powerpc/mm/init_64.c
353
page = pfn_to_page(addr >> PAGE_SHIFT);
arch/powerpc/mm/init_64.c
82
return pfn_to_page(start_pfn);
arch/powerpc/mm/pgtable.c
69
page = pfn_to_page(pfn);
arch/powerpc/platforms/512x/mpc512x_shared.c
188
mpc512x_free_bootmem(pfn_to_page(start));
arch/powerpc/platforms/powernv/memtrace.c
124
__SetPageOffline(pfn_to_page(pfn));
arch/powerpc/platforms/powernv/memtrace.c
210
__ClearPageOffline(pfn_to_page(pfn));
arch/powerpc/platforms/pseries/papr_scm.c
1336
!is_zone_device_page(pfn_to_page(phys_addr >> PAGE_SHIFT)))
arch/powerpc/platforms/pseries/svm.c
87
struct page *page = pfn_to_page(pfn);
arch/riscv/include/asm/page.h
188
#define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr)))
arch/riscv/include/asm/pgtable-64.h
230
return pfn_to_page(__page_val_to_pfn(pud_val(pud)));
arch/riscv/include/asm/pgtable-64.h
335
return pfn_to_page(__page_val_to_pfn(p4d_val(p4d)));
arch/riscv/include/asm/pgtable-64.h
392
return pfn_to_page(__page_val_to_pfn(pgd_val(pgd)));
arch/riscv/include/asm/pgtable.h
271
return pfn_to_page(__page_val_to_pfn(pmd_val(pmd)));
arch/riscv/include/asm/pgtable.h
332
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/riscv/kernel/vdso.c
66
vdso_pagelist[i] = pfn_to_page(pfn + i);
arch/s390/include/asm/page.h
275
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
arch/s390/include/asm/pgtable.h
1470
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/s390/include/asm/pgtable.h
1472
#define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd))
arch/s390/include/asm/pgtable.h
1473
#define pud_page(pud) pfn_to_page(pud_pfn(pud))
arch/s390/include/asm/pgtable.h
1474
#define p4d_page(p4d) pfn_to_page(p4d_pfn(p4d))
arch/s390/include/asm/pgtable.h
1475
#define pgd_page(pgd) pfn_to_page(pgd_pfn(pgd))
arch/s390/kernel/machine_kexec.c
144
free_reserved_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/s390/kvm/gmap.h
189
SetPageDirty(pfn_to_page(newpte.h.pfra));
arch/s390/kvm/vsie.c
680
kvm_release_page_dirty(pfn_to_page(hpa >> PAGE_SHIFT));
arch/sh/include/asm/page.h
157
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sh/include/asm/pgtable_32.h
333
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/sh/mm/cache-sh4.c
229
page = pfn_to_page(pfn);
arch/sh/mm/cache.c
150
struct folio *folio = page_folio(pfn_to_page(pfn));
arch/sh/mm/fault.c
109
if (PageHighMem(pfn_to_page(pmd_val(*pmd) >> PAGE_SHIFT)))
arch/sparc/include/asm/page_32.h
129
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/sparc/include/asm/page_64.h
150
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr)>>PAGE_SHIFT)
arch/sparc/include/asm/pgtable_32.h
118
return pfn_to_page(pmd_pfn(pmd));
arch/sparc/include/asm/pgtable_32.h
277
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/sparc/include/asm/pgtable_64.h
258
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/sparc/include/asm/pgtable_64.h
966
page_folio(pfn_to_page(this_pfn))); \
arch/sparc/kernel/traps_64.c
1739
get_page(pfn_to_page(afar >> PAGE_SHIFT));
arch/sparc/kernel/traps_64.c
2176
get_page(pfn_to_page(addr >> PAGE_SHIFT));
arch/sparc/mm/init_64.c
290
page = pfn_to_page(pfn);
arch/sparc/mm/srmmu.c
350
page = pfn_to_page(__nocache_pa((unsigned long)ptep) >> PAGE_SHIFT);
arch/sparc/mm/srmmu.c
366
page = pfn_to_page(__nocache_pa((unsigned long)ptep) >> PAGE_SHIFT);
arch/sparc/mm/tlb.c
135
page = pfn_to_page(pfn);
arch/um/include/asm/pgtable.h
103
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/x86/include/asm/page.h
62
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/x86/include/asm/pgtable.h
1031
#define pmd_page(pmd) pfn_to_page(pmd_pfn(pmd))
arch/x86/include/asm/pgtable.h
1064
#define pud_page(pud) pfn_to_page(pud_pfn(pud))
arch/x86/include/asm/pgtable.h
1098
#define p4d_page(p4d) pfn_to_page(p4d_pfn(p4d))
arch/x86/include/asm/pgtable.h
1133
#define pgd_page(pgd) pfn_to_page(pgd_pfn(pgd))
arch/x86/include/asm/pgtable.h
296
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
arch/x86/kernel/crash.c
552
old_elfcorehdr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
arch/x86/kernel/irq_64.c
46
pages[i] = pfn_to_page(pa >> PAGE_SHIFT);
arch/x86/kernel/machine_kexec_64.c
642
page = pfn_to_page(start >> PAGE_SHIFT);
arch/x86/kvm/mmu/spte.c
110
return !is_zero_pfn(pfn) && PageReserved(pfn_to_page(pfn)) &&
arch/x86/kvm/mmu/spte.h
261
struct page *page = pfn_to_page((shadow_page) >> PAGE_SHIFT);
arch/x86/kvm/svm/svm.h
41
return pfn_to_page(__sme_clr(pa) >> PAGE_SHIFT);
arch/x86/kvm/vmx/tdx.c
1637
err = tdh_mem_page_add(&kvm_tdx->td, gpa, pfn_to_page(pfn),
arch/x86/kvm/vmx/tdx.c
1653
struct page *page = pfn_to_page(pfn);
arch/x86/kvm/vmx/tdx.c
1790
struct page *page = pfn_to_page(spte_to_pfn(mirror_spte));
arch/x86/mm/ioremap.c
83
if (!PageReserved(pfn_to_page(pfn)))
arch/x86/mm/pat/cpa-test.c
173
pages[k] = pfn_to_page(pfn + k);
arch/x86/mm/pat/memtype.c
458
page = pfn_to_page(pfn);
arch/x86/mm/pat/memtype.c
474
page = pfn_to_page(pfn);
arch/x86/mm/pat/memtype.c
486
page = pfn_to_page(pfn);
arch/x86/mm/pat/memtype.c
666
page = pfn_to_page(paddr >> PAGE_SHIFT);
arch/x86/platform/efi/memmap.c
42
struct page *p = pfn_to_page(PHYS_PFN(phys));
arch/x86/virt/svm/sev.c
1034
struct page *page = pfn_to_page(pfn);
arch/x86/xen/enlighten.c
462
pfn_to_page(xen_extra_mem[i].start_pfn + j);
arch/x86/xen/mmu_pv.c
1601
struct page *page = pfn_to_page(pfn);
arch/x86/xen/mmu_pv.c
1634
struct page *page = pfn_to_page(pfn);
arch/xtensa/include/asm/page.h
170
#define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
arch/xtensa/include/asm/pgtable.h
274
#define pte_page(x) pfn_to_page(pte_pfn(x))
arch/xtensa/kernel/pci-dma.c
29
struct page *page = pfn_to_page(pfn);
arch/xtensa/mm/cache.c
204
unsigned long phys = page_to_phys(pfn_to_page(pfn));
arch/xtensa/mm/cache.c
226
folio = page_folio(pfn_to_page(pfn));
arch/xtensa/mm/tlb.c
258
struct page *p = pfn_to_page(r1 >> PAGE_SHIFT);
drivers/acpi/osl.c
294
return (void __iomem __force *)kmap(pfn_to_page(pfn));
drivers/acpi/osl.c
305
kunmap(pfn_to_page(pfn));
drivers/base/memory.c
268
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
299
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
307
adjust_present_page_count(pfn_to_page(start_pfn),
drivers/dax/device.c
139
return vmf_insert_page_mkwrite(vmf, pfn_to_page(pfn),
drivers/dax/device.c
183
return vmf_insert_folio_pmd(vmf, page_folio(pfn_to_page(pfn)),
drivers/dax/device.c
229
return vmf_insert_folio_pud(vmf, page_folio(pfn_to_page(pfn)),
drivers/dma/loongson1-apb-dma.c
344
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(buf_addr)),
drivers/dma/sh/rcar-dmac.c
1137
sg_set_page(&sgl, pfn_to_page(PFN_DOWN(dma_src)), len,
drivers/dma/sh/rcar-dmac.c
1266
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
drivers/dma/sh/shdma-base.c
664
sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
drivers/dma/sh/shdma-base.c
752
sg_set_page(&sgl[i], pfn_to_page(PFN_DOWN(src)), period_len,
drivers/edac/edac_mc.c
735
pg = pfn_to_page(page);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2758
p = pfn_to_page(pfn);
drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
2809
p = pfn_to_page(pfn);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
223
page = pfn_to_page(pfn);
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
234
page = pfn_to_page(svm_migrate_addr_to_pfn(adev, addr));
drivers/gpu/drm/amd/amdkfd/kfd_migrate.c
264
page = pfn_to_page(addr >> PAGE_SHIFT);
drivers/gpu/drm/drm_gem_shmem_helper.c
564
folio_test_pmd_mappable(page_folio(pfn_to_page(pfn)))) {
drivers/gpu/drm/drm_pagemap.c
590
struct page *page = pfn_to_page(migrate.dst[i]);
drivers/gpu/drm/i915/i915_scatterlist.h
114
pfn_to_page((__iter).pfn + ((__iter).curr >> PAGE_SHIFT))); \
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
87
sg_set_page(sg, pfn_to_page(PFN_BIAS), len, 0);
drivers/gpu/drm/i915/selftests/scatterlist.c
122
if (page != pfn_to_page(pfn)) {
drivers/gpu/drm/i915/selftests/scatterlist.c
245
if (!page_contiguous(pfn_to_page(pfn),
drivers/gpu/drm/i915/selftests/scatterlist.c
246
pfn_to_page(pfn + npages),
drivers/gpu/drm/i915/selftests/scatterlist.c
254
sg_set_page(sg, pfn_to_page(pfn), npages * PAGE_SIZE, 0);
drivers/gpu/drm/i915/selftests/scatterlist.c
92
if (page != pfn_to_page(pfn)) {
drivers/gpu/drm/nouveau/nouveau_dmem.c
344
page = pfn_to_page(pfn_first);
drivers/gpu/drm/nouveau/nouveau_dmem.c
353
page = pfn_to_page(pfn);
drivers/gpu/drm/nouveau/nouveau_dmem.c
358
page = pfn_to_page(pfn);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
351
addr = dma_map_page(dev, pfn_to_page(addr), 0,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
94
addr = dma_map_page(dev, pfn_to_page(addr), 0,
drivers/gpu/drm/omapdrm/omap_gem.c
1023
sg_set_page(sg, pfn_to_page(__phys_to_pfn(addr)),
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
808
unpin_user_page(pfn_to_page(pdesc->statPPNs[i]));
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
811
unpin_user_page(pfn_to_page(pdesc->infoPPNs[i]));
drivers/gpu/drm/vmwgfx/vmwgfx_msg.c
814
unpin_user_page(pfn_to_page(pdesc->strsPPNs[i]));
drivers/hv/hv_balloon.c
1191
pg = pfn_to_page(i + start_frame);
drivers/hv/hv_balloon.c
703
hv_page_online_one(has, pfn_to_page(start_pfn + i));
drivers/hv/hv_proc.c
103
__free_page(pfn_to_page(base_pfn + j));
drivers/hv/mshv_root_hv_call.c
565
*state_page = pfn_to_page(output->map_location);
drivers/hv/mshv_root_hv_call.c
74
__free_page(pfn_to_page(output_page->gpa_page_list[i]));
drivers/hv/mshv_root_hv_call.c
931
*addr = page_address(pfn_to_page(pfn));
drivers/infiniband/hw/efa/efa_verbs.c
2065
err = vm_insert_page(vma, va, pfn_to_page(pfn));
drivers/iommu/iommufd/pages.c
2110
*(out_pages++) = pfn_to_page(xa_to_value(entry));
drivers/iommu/iommufd/pages.c
449
*(out_pages++) = pfn_to_page(PHYS_PFN(phys));
drivers/iommu/iommufd/pages.c
715
pfn_to_page(batch->pfns[cur] + first_page_off),
drivers/iommu/iommufd/pages.c
750
copy_data_page(pfn_to_page(batch->pfns[cur] + npage), data,
drivers/md/dm-pcache/cache_dev.c
44
pages[i++] = pfn_to_page(pfn);
drivers/md/dm-writecache.c
321
pages[i++] = pfn_to_page(pfn);
drivers/media/common/videobuf2/frame_vector.c
124
pages[i] = pfn_to_page(nums[i]);
drivers/misc/vmw_balloon.c
745
*p = pfn_to_page(b->batch_page[idx].pfn);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
676
if (page_shift(pfn_to_page(pfn)) != EHEA_HUGEPAGESHIFT)
drivers/nvdimm/pmem.c
76
struct page *page = pfn_to_page(pfn);
drivers/usb/musb/ux500_dma.c
94
sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_addr)), len,
drivers/vdpa/mlx5/core/mr.c
404
pg = pfn_to_page(__phys_to_pfn(pa));
drivers/vdpa/vdpa_user/iova_domain.c
171
page = pfn_to_page(pfn);
drivers/vdpa/vdpa_user/iova_domain.c
231
page = pfn_to_page((map->addr + iova - map->start) >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
189
page = pfn_to_page(hpa >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_spapr_tce.c
385
page = pfn_to_page(hpa >> PAGE_SHIFT);
drivers/vfio/vfio_iommu_type1.c
476
return PageReserved(pfn_to_page(pfn));
drivers/vfio/vfio_iommu_type1.c
484
struct page *page = pfn_to_page(pfn);
drivers/vfio/vfio_iommu_type1.c
814
unpin_user_page_range_dirty_lock(pfn_to_page(start_pfn), npage,
drivers/vfio/vfio_iommu_type1.c
960
pages[i] = pfn_to_page(vpfn->pfn);
drivers/vfio/vfio_iommu_type1.c
982
pages[i] = pfn_to_page(phys_pfn);
drivers/vhost/vdpa.c
1208
unpin_user_page(pfn_to_page(pfn));
drivers/vhost/vdpa.c
932
page = pfn_to_page(pfn);
drivers/vhost/vringh.c
1082
bvec_set_page(&bvec[ret], pfn_to_page(pfn), io_len,
drivers/video/fbdev/core/fb_defio.c
39
page = pfn_to_page((info->fix.smem_start + offs) >> PAGE_SHIFT);
drivers/virt/acrn/mm.c
212
!PageReserved(pfn_to_page(cur_pfn))) {
drivers/virtio/virtio_mem.c
1149
struct page *page = pfn_to_page(pfn);
drivers/virtio/virtio_mem.c
1173
struct page *page = pfn_to_page(pfn);
drivers/virtio/virtio_mem.c
1201
struct page *page = pfn_to_page(pfn + i);
drivers/virtio/virtio_mem.c
1227
const bool is_movable = is_zone_movable_page(pfn_to_page(pfn));
drivers/virtio/virtio_mem.c
1257
adjust_managed_page_count(pfn_to_page(pfn), -nr_pages);
drivers/virtio/virtio_mem.c
1276
page = pfn_to_page(pfn + i);
drivers/virtio/virtio_mem.c
1296
page_ref_inc(pfn_to_page(pfn + i));
drivers/virtio/virtio_mem.c
1350
generic_online_page(pfn_to_page(PFN_DOWN(addr)), order);
drivers/virtio/virtio_mem.c
906
const bool is_movable = is_zone_movable_page(pfn_to_page(start_pfn));
drivers/xen/balloon.c
353
p = pfn_to_page(start_pfn + i);
drivers/xen/balloon.c
703
balloon_append(pfn_to_page(pfn));
drivers/xen/grant-table.c
1086
struct page *page = pfn_to_page(pfn);
drivers/xen/grant-table.c
563
page = pfn_to_page(gfn_to_pfn(gfn));
fs/dax.c
1173
dax_flush(dax_dev, page_address(pfn_to_page(pfn)), count * PAGE_SIZE);
fs/dax.c
1368
ret = vmf_insert_page_mkwrite(vmf, pfn_to_page(pfn), false);
fs/dax.c
1856
ret = vmf_insert_page_mkwrite(vmf, pfn_to_page(pfn), write);
fs/dax.c
600
*page = pfn_to_page(dax_to_pfn(entry));
fs/dax.c
75
return page_folio(pfn_to_page(dax_to_pfn(entry)));
fs/proc/kcore.c
169
start = ((unsigned long)pfn_to_page(pfn)) & PAGE_MASK;
fs/proc/kcore.c
170
end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1;
fs/proc/kcore.c
211
p = pfn_to_page(pfn);
fs/pstore/ram_core.c
450
pages[i] = pfn_to_page(addr >> PAGE_SHIFT);
include/asm-generic/memory_model.h
87
#define phys_to_page(phys) pfn_to_page(PHYS_PFN(phys))
include/linux/highmem-internal.h
208
return kmap_local_page(pfn_to_page(pfn));
include/linux/highmem-internal.h
235
return kmap_atomic(pfn_to_page(pfn));
include/linux/hmm.h
69
return pfn_to_page(hmm_pfn & ~HMM_PFN_FLAGS);
include/linux/kexec.h
482
return pfn_to_page(boot_pfn);
include/linux/leafops.h
391
struct page *page = pfn_to_page(softleaf_to_pfn(entry));
include/linux/memory_hotplug.h
184
___page = pfn_to_page(pfn); \
include/linux/migrate.h
143
return pfn_to_page(mpfn >> MIGRATE_PFN_SHIFT);
include/linux/mm.h
2264
return page_folio(pfn_to_page(pfn));
include/linux/mmzone.h
2296
page_to_nid(pfn_to_page(__pfn_to_nid_pfn)); \
include/trace/events/kmem.h
156
pfn_to_page(__entry->pfn),
include/trace/events/kmem.h
176
pfn_to_page(__entry->pfn),
include/trace/events/kmem.h
202
__entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
include/trace/events/kmem.h
231
__entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL,
include/trace/events/kmem.h
265
pfn_to_page(__entry->pfn), __entry->pfn,
include/trace/events/kmem.h
299
pfn_to_page(__entry->pfn),
include/trace/events/vmscan.h
354
pfn_to_page(__entry->pfn),
include/xen/page.h
19
(pfn_to_page((unsigned long)(xen_pfn) >> (PAGE_SHIFT - XEN_PAGE_SHIFT)))
kernel/crash_core.c
615
ptr = kmap_local_page(pfn_to_page(mem >> PAGE_SHIFT));
kernel/crash_dump_dm_crypt.c
93
keys_header_loaded = kmap_local_page(pfn_to_page(
kernel/debug/kdb/kdb_support.c
354
page = pfn_to_page(pfn);
kernel/dma/direct.c
37
return pfn_to_page(PHYS_PFN(dma_to_phys(dev, dma_addr)));
kernel/dma/pool.c
271
return pfn_to_page(__phys_to_pfn(phys));
kernel/dma/swiotlb.c
1790
return pfn_to_page(PFN_DOWN(tlb_addr));
kernel/dma/swiotlb.c
1816
if (PageHighMem(pfn_to_page(PHYS_PFN(rmem->base)))) {
kernel/dma/swiotlb.c
894
if (PageHighMem(pfn_to_page(pfn))) {
kernel/dma/swiotlb.c
904
page = pfn_to_page(pfn);
kernel/iomem.c
34
if (pfn_valid(pfn) && !PageHighMem(pfn_to_page(pfn)) &&
kernel/kexec_core.c
450
pages = pfn_to_page(hole_start >> PAGE_SHIFT);
kernel/liveupdate/kexec_handover.c
1401
init_cma_reserved_pageblock(pfn_to_page(pfn));
kernel/liveupdate/kexec_handover.c
1443
init_pageblock_migratetype(pfn_to_page(pfn),
kernel/liveupdate/kexec_handover.c
320
return pfn_to_page(start_pfn);
kernel/power/snapshot.c
1207
clear_or_poison_free_page(pfn_to_page(pfn));
kernel/power/snapshot.c
1258
page = pfn_to_page(pfn);
kernel/power/snapshot.c
1283
swsusp_set_page_free(pfn_to_page(pfn + i));
kernel/power/snapshot.c
1479
s_page = pfn_to_page(src_pfn);
kernel/power/snapshot.c
1480
d_page = pfn_to_page(dst_pfn);
kernel/power/snapshot.c
1508
return safe_copy_page(page_address(pfn_to_page(dst_pfn)),
kernel/power/snapshot.c
1509
pfn_to_page(src_pfn));
kernel/power/snapshot.c
1616
struct page *page = pfn_to_page(fr_pfn);
kernel/power/snapshot.c
1752
struct page *page = pfn_to_page(pfn);
kernel/power/snapshot.c
2261
page = pfn_to_page(memory_bm_next_pfn(©_bm));
kernel/power/snapshot.c
2423
if (PageHighMem(pfn_to_page(pfn)))
kernel/power/snapshot.c
2533
tmp = pfn_to_page(memory_bm_next_pfn(safe_highmem_bm));
kernel/power/snapshot.c
2724
page = pfn_to_page(pfn);
lib/scatterlist.c
488
last_pg = pfn_to_page(next_pfn - 1);
lib/test_hmm.c
1006
dst_page = pfn_to_page(page_to_pfn(dpage));
lib/test_hmm.c
1009
dst_page = pfn_to_page(page_to_pfn(dpage) + i);
lib/test_hmm.c
1012
src_page = pfn_to_page(page_to_pfn(spage) + i);
lib/test_hmm.c
578
struct page *page = pfn_to_page(pfn);
lib/test_hmm.c
733
src_page = pfn_to_page(spfn + i);
lib/test_hmm.c
769
pfn_to_page(page_to_pfn(rpage) + i);
lib/test_hmm.c
771
pfn_to_page(page_to_pfn(spage) + i);
mm/cma.c
1025
ret += !put_page_testzero(pfn_to_page(pfn));
mm/cma.c
169
init_cma_reserved_pageblock(pfn_to_page(pfn));
mm/cma.c
194
free_reserved_page(pfn_to_page(pfn));
mm/cma.c
816
page = pfn_to_page(pfn);
mm/compaction.c
1591
page = pfn_to_page(high_pfn);
mm/compaction.c
1643
page = pfn_to_page(highest);
mm/compaction.c
574
page = pfn_to_page(blockpfn);
mm/compaction.c
932
page = pfn_to_page(low_pfn);
mm/debug_vm_pgtable.c
1008
__free_page(pfn_to_page(args->pte_pfn));
mm/debug_vm_pgtable.c
130
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
mm/debug_vm_pgtable.c
228
page = (args->pmd_pfn != ULONG_MAX) ? pfn_to_page(args->pmd_pfn) : NULL;
mm/debug_vm_pgtable.c
350
page = (args->pud_pfn != ULONG_MAX) ? pfn_to_page(args->pud_pfn) : NULL;
mm/debug_vm_pgtable.c
611
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
mm/debug_vm_pgtable.c
859
page = (args->pte_pfn != ULONG_MAX) ? pfn_to_page(args->pte_pfn) : NULL;
mm/debug_vm_pgtable.c
984
__free_pages(pfn_to_page(pfn), order);
mm/gup.c
665
!can_follow_write_pud(pud, pfn_to_page(pfn), vma, flags))
mm/gup.c
669
page = pfn_to_page(pfn);
mm/hugetlb.c
2833
page = pfn_to_page(start_pfn);
mm/hugetlb_vmemmap.c
810
map = pfn_to_page(pfn);
mm/hugetlb_vmemmap.c
861
map = pfn_to_page(pfn);
mm/hwpoison-inject.c
115
p = pfn_to_page(pfn);
mm/internal.h
347
VM_WARN_ON_FOLIO(page_folio(pfn_to_page(pte_pfn(pte))) != folio, folio);
mm/internal.h
590
set_page_refcounted(pfn_to_page(pfn));
mm/internal.h
797
return pfn_to_page(start_pfn);
mm/kasan/shadow.c
484
__free_page(pfn_to_page(pte_pfn(pte)));
mm/kfence/core.c
627
page = pfn_to_page(start_pfn + i);
mm/kfence/core.c
696
page = pfn_to_page(start_pfn + i);
mm/memblock.c
2128
start_pg = pfn_to_page(start_pfn - 1) + 1;
mm/memblock.c
2129
end_pg = pfn_to_page(end_pfn - 1) + 1;
mm/memory-failure.c
1975
struct page *page = pfn_to_page(pfn);
mm/memory-failure.c
2038
struct page *p = pfn_to_page(pfn);
mm/memory-failure.c
2127
put_page(pfn_to_page(pfn));
mm/memory.c
2810
page = pfn_to_page(pfn);
mm/memory.c
749
return pfn_to_page(pfn);
mm/memory_hotplug.c
1104
struct page *page = pfn_to_page(pfn + i);
mm/memory_hotplug.c
1137
remove_pfn_range_from_zone(page_zone(pfn_to_page(pfn)), pfn, nr_pages);
mm/memory_hotplug.c
1208
adjust_present_page_count(pfn_to_page(pfn), group, nr_pages);
mm/memory_hotplug.c
1759
page = pfn_to_page(pfn);
mm/memory_hotplug.c
1803
page = pfn_to_page(pfn);
mm/memory_hotplug.c
1953
if (WARN_ON_ONCE(page_zone(pfn_to_page(start_pfn)) != zone ||
mm/memory_hotplug.c
1954
page_zone(pfn_to_page(end_pfn - 1)) != zone)) {
mm/memory_hotplug.c
2061
adjust_managed_page_count(pfn_to_page(start_pfn), -managed_pages);
mm/memory_hotplug.c
2062
adjust_present_page_count(pfn_to_page(start_pfn), group, -nr_pages);
mm/memory_hotplug.c
370
return pfn_to_page(pfn);
mm/memory_hotplug.c
385
return pfn_to_page(pfn);
mm/memory_hotplug.c
445
if (zone != page_zone(pfn_to_page(start_pfn)))
mm/memory_hotplug.c
470
if (zone != page_zone(pfn_to_page(pfn)))
mm/memory_hotplug.c
562
page_init_poison(pfn_to_page(pfn),
mm/memory_hotplug.c
673
struct page *page = pfn_to_page(pfn);
mm/memremap.c
92
first_page = pfn_to_page(pfn_first(pgmap, range_id));
mm/migrate_device.c
1362
folio = folio_get_nontail_page(pfn_to_page(pfn));
mm/migrate_device.c
1399
struct page *page = pfn_to_page(pfn);
mm/migrate_device.c
1433
struct page *page = pfn_to_page(src_pfns[i]);
mm/mm_init.c
1099
struct page *page = pfn_to_page(pfn);
mm/mm_init.c
1135
struct page *page = pfn_to_page(pfn);
mm/mm_init.c
1995
page = pfn_to_page(pfn);
mm/mm_init.c
2038
struct page *page = pfn_to_page(pfn);
mm/mm_init.c
2483
struct page *page = pfn_to_page(pfn);
mm/mm_init.c
682
__init_single_page(pfn_to_page(pfn), pfn, zid, nid);
mm/mm_init.c
685
init_pageblock_migratetype(pfn_to_page(pfn), MIGRATE_MOVABLE,
mm/mm_init.c
787
struct page *page = pfn_to_page(pfn);
mm/mm_init.c
853
__init_single_page(pfn_to_page(pfn), pfn, zone, node);
mm/mm_init.c
854
__SetPageReserved(pfn_to_page(pfn));
mm/mm_init.c
916
page = pfn_to_page(pfn);
mm/page_alloc.c
1559
page = pfn_to_page(pfn);
mm/page_alloc.c
1709
end_page = pfn_to_page(end_pfn);
mm/page_alloc.c
1984
page = pfn_to_page(pfn);
mm/page_alloc.c
2033
page = pfn_to_page(pfn);
mm/page_alloc.c
2065
set_pageblock_migratetype(pfn_to_page(start_pfn), new_mt);
mm/page_alloc.c
2085
while (!PageBuddy(page = pfn_to_page(pfn))) {
mm/page_alloc.c
2151
buddy = pfn_to_page(buddy_pfn);
mm/page_alloc.c
2176
toggle_pageblock_isolate(pfn_to_page(start_pfn), isolate);
mm/page_alloc.c
2371
set_pageblock_migratetype(pfn_to_page(start_pfn), start_type);
mm/page_alloc.c
6965
free_frozen_pages(pfn_to_page(pfn), 0);
mm/page_alloc.c
7001
.zone = page_zone(pfn_to_page(start)),
mm/page_alloc.c
7116
struct page *head = pfn_to_page(start);
mm/page_alloc.c
7158
set_pages_refcounted(pfn_to_page(start), end - start);
mm/page_alloc.c
7281
return pfn_to_page(pfn);
mm/page_alloc.c
7345
struct page *first_page = pfn_to_page(pfn);
mm/page_alloc.c
7370
if (WARN_ON_ONCE(PageHead(pfn_to_page(pfn))))
mm/page_alloc.c
7374
__free_page(pfn_to_page(pfn));
mm/page_alloc.c
7439
zone = page_zone(pfn_to_page(pfn));
mm/page_alloc.c
7442
page = pfn_to_page(pfn);
mm/page_ext.c
513
return lookup_page_ext(pfn_to_page(pfn));
mm/page_isolation.c
130
struct page *page = pfn_to_page(start_pfn);
mm/page_isolation.c
151
page = pfn_to_page(start_pfn);
mm/page_isolation.c
347
zone = page_zone(pfn_to_page(isolate_pageblock));
mm/page_isolation.c
352
VM_BUG_ON(!get_pageblock_isolate(pfn_to_page(isolate_pageblock)));
mm/page_isolation.c
354
ret = set_migratetype_isolate(pfn_to_page(isolate_pageblock),
mm/page_isolation.c
446
unset_migratetype_isolate(pfn_to_page(isolate_pageblock));
mm/page_isolation.c
509
unset_migratetype_isolate(pfn_to_page(isolate_start));
mm/page_isolation.c
522
pfn_to_page(isolate_end - pageblock_nr_pages));
mm/page_isolation.c
566
page = pfn_to_page(pfn);
mm/page_owner.c
459
page = pfn_to_page(pfn);
mm/page_owner.c
699
page = pfn_to_page(pfn);
mm/page_owner.c
795
struct page *page = pfn_to_page(pfn);
mm/page_table_check.c
108
page = pfn_to_page(pfn);
mm/page_table_check.c
73
page = pfn_to_page(pfn);
mm/pagewalk.c
1025
page = pfn_to_page(pte_pfn(pte));
mm/pagewalk.c
991
page = pfn_to_page(pmd_pfn(pmd));
mm/sparse-vmemmap.c
179
get_page(pfn_to_page(ptpfn));
mm/sparse-vmemmap.c
565
unsigned long start = (unsigned long) pfn_to_page(pfn);
mm/sparse-vmemmap.c
581
return pfn_to_page(pfn);
mm/sparse.c
679
unsigned long start = (unsigned long) pfn_to_page(pfn);
mm/sparse.c
752
kvfree(pfn_to_page(pfn));
mm/sparse.c
902
return pfn_to_page(pfn);
mm/sparse.c
962
memmap = pfn_to_page(section_nr_to_pfn(section_nr));
mm/util.c
1465
if (unlikely(page + (pfn - start_pfn) != pfn_to_page(pfn)))
mm/vmalloc.c
116
page = pfn_to_page(pfn);
mm/zpdesc.h
152
return page_zpdesc(pfn_to_page(pfn));
sound/soc/renesas/siu_pcm.c
116
sg_set_page(&sg, pfn_to_page(PFN_DOWN(buff)),
sound/soc/renesas/siu_pcm.c
166
sg_set_page(&sg, pfn_to_page(PFN_DOWN(buff)),
virt/kvm/kvm_main.c
3139
map->page = pfn_to_page(map->pfn);
virt/kvm/pfncache.c
102
return kmap(pfn_to_page(pfn));
virt/kvm/pfncache.c
118
kunmap(pfn_to_page(pfn));