Symbol: PHYS_PFN
arch/arm64/mm/init.c
169
if (PHYS_PFN(addr) != pfn)
arch/arm64/mm/ioremap.c
55
unsigned long pfn = PHYS_PFN(offset);
arch/arm64/mm/mmu.c
2099
ms = __pfn_to_section(PHYS_PFN(addr));
arch/m68k/include/asm/page_no.h
37
#define ARCH_PFN_OFFSET PHYS_PFN(PAGE_OFFSET_RAW)
arch/mips/loongson64/numa.c
140
max_low_pfn = PHYS_PFN(memblock_end_of_DRAM());
arch/mips/sgi-ip27/ip27-memory.c
398
max_low_pfn = PHYS_PFN(memblock_end_of_DRAM());
arch/powerpc/include/asm/video.h
11
return __phys_mem_access_prot(PHYS_PFN(offset), vm_end - vm_start, prot);
arch/powerpc/kernel/fadump.c
1207
spfn = PHYS_PFN(start);
arch/powerpc/kernel/fadump.c
1208
epfn = PHYS_PFN(end);
arch/powerpc/kernel/paca.c
85
uv_share_page(PHYS_PFN(__pa(shared_lppaca)),
arch/powerpc/mm/init_64.c
355
base_pfn = PHYS_PFN(addr);
arch/powerpc/mm/kasan/8xx.c
30
pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL));
arch/powerpc/mm/kasan/8xx.c
68
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/mm/kasan/book3s_32.c
52
pte_t pte = pfn_pte(PHYS_PFN(phys + k_cur - k_nobat), PAGE_KERNEL);
arch/powerpc/mm/kasan/init_32.c
109
pfn_pte(PHYS_PFN(pa), prot));
arch/powerpc/mm/kasan/init_32.c
28
__set_pte_at(&init_mm, va, ptep, pfn_pte(PHYS_PFN(pa), prot), 1);
arch/powerpc/mm/kasan/init_32.c
75
pte_t pte = pfn_pte(PHYS_PFN(__pa(va)), PAGE_KERNEL);
arch/powerpc/platforms/powernv/memtrace.c
200
const unsigned long nr_pages = PHYS_PFN(size);
arch/powerpc/platforms/powernv/memtrace.c
201
const unsigned long start_pfn = PHYS_PFN(start);
arch/powerpc/platforms/powernv/memtrace.c
63
return remap_pfn_range(vma, vma->vm_start, PHYS_PFN(ent->start) + vma->vm_pgoff,
arch/powerpc/platforms/powernv/memtrace.c
97
const unsigned long nr_pages = PHYS_PFN(size);
arch/powerpc/platforms/pseries/svm.c
49
uv_unshare_page(PHYS_PFN(__pa(addr)), numpages);
arch/powerpc/platforms/pseries/svm.c
62
uv_share_page(PHYS_PFN(__pa(addr)), numpages);
arch/powerpc/platforms/pseries/svm.c
86
unsigned long pfn = PHYS_PFN(__pa(addr));
arch/powerpc/sysdev/xive/spapr.c
524
uv_share_page(PHYS_PFN(qpage_phys),
arch/powerpc/sysdev/xive/spapr.c
560
uv_unshare_page(PHYS_PFN(__pa(q->qpage)), 1 << alloc_order);
arch/s390/kvm/gmap.c
530
f->pfn = PHYS_PFN(large_crste_to_phys(oldcrste, f->gfn));
arch/x86/kvm/svm/sev.c
5159
if (snp_page_reclaim(vcpu->kvm, PHYS_PFN(__pa(vmsa))))
arch/x86/platform/efi/memmap.c
42
struct page *p = pfn_to_page(PHYS_PFN(phys));
arch/x86/virt/svm/sev.c
116
#define RMP_ENTRY_INDEX(x) ((u64)(PHYS_PFN((x) & rmp_segment_mask)))
arch/x86/virt/svm/sev.c
240
size = PHYS_PFN(mapped_size) << 4;
arch/x86/virt/svm/sev.c
287
rmp_segment_size_max = PHYS_PFN(rmp_segment_size) << 4;
arch/x86/virt/svm/sev.c
454
rmp_size = PHYS_PFN(mapped_size) << 4;
arch/x86/virt/svm/sev.c
856
dump_rmpentry(PHYS_PFN(paddr));
arch/x86/virt/vmx/tdx/tdx.c
241
start_pfn = max(start_pfn, PHYS_PFN(SZ_1M));
arch/x86/virt/vmx/tdx/tdx.c
491
if (tmb->end_pfn > PHYS_PFN(tdmr->base))
arch/x86/xen/enlighten_hvm.c
67
shared_info_pfn = PHYS_PFN(pa);
arch/xtensa/include/asm/page.h
26
#define MAX_LOW_PFN (PHYS_PFN(XCHAL_KSEG_PADDR) + \
arch/xtensa/include/asm/page.h
27
PHYS_PFN(XCHAL_KSEG_SIZE))
arch/xtensa/include/asm/page.h
31
#define MAX_LOW_PFN PHYS_PFN(0xfffffffful)
arch/xtensa/mm/kasan_init.c
58
set_pte(pte + j, pfn_pte(PHYS_PFN(phys), PAGE_KERNEL));
drivers/acpi/acpi_mrrm.c
47
if (zone_intersects(zone, PHYS_PFN(e->base), PHYS_PFN(e->length)))
drivers/acpi/apei/ghes.c
351
vaddr = ghes_map(PHYS_PFN(paddr), fixmap_idx);
drivers/acpi/apei/ghes.c
513
pfn = PHYS_PFN(physical_addr);
drivers/base/memory.c
1154
unit_pages < PHYS_PFN(memory_block_size_bytes()))
drivers/dax/bus.c
873
pgoff += PHYS_PFN(range_len(&ranges[i].range));
drivers/dax/device.c
135
pfn = PHYS_PFN(phys);
drivers/dax/device.c
179
pfn = PHYS_PFN(phys);
drivers/dax/device.c
225
pfn = PHYS_PFN(phys);
drivers/dax/device.c
72
pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1;
drivers/edac/ghes_edac.c
341
e->page_frame_number = PHYS_PFN(mem_err->physical_addr);
drivers/edac/versalnet_edac.c
462
pfn = PHYS_PFN(pa);
drivers/gpu/drm/xe/xe_device.c
285
pfn = PHYS_PFN(pci_resource_start(to_pci_dev(dev->dev), 0) +
drivers/gpu/drm/xe/xe_mmio_gem.c
219
ret = vmf_insert_pfn(vma, addr, PHYS_PFN(phys_addr));
drivers/gpu/drm/xe/xe_svm.c
747
return PHYS_PFN(offset + xpagemap->hpa_base);
drivers/infiniband/hw/ionic/ionic_controlpath.c
2270
PHYS_PFN(qp->sq_cmb_addr),
drivers/infiniband/hw/ionic/ionic_controlpath.c
2307
PHYS_PFN(qp->rq_cmb_addr),
drivers/infiniband/hw/ionic/ionic_controlpath.c
388
PHYS_PFN(db_phys), 0, NULL);
drivers/iommu/iommufd/pages.c
1110
batch_add_pfn_num(batch, PHYS_PFN(dmabuf->phys.paddr + start),
drivers/iommu/iommufd/pages.c
426
if (!batch_add_pfn(batch, PHYS_PFN(phys)))
drivers/iommu/iommufd/pages.c
449
*(out_pages++) = pfn_to_page(PHYS_PFN(phys));
drivers/media/pci/intel/ipu6/ipu6-dma.c
103
int count = PHYS_PFN(size);
drivers/media/pci/intel/ipu6/ipu6-dma.c
172
count = PHYS_PFN(size);
drivers/media/pci/intel/ipu6/ipu6-dma.c
175
PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0);
drivers/media/pci/intel/ipu6/ipu6-dma.c
249
struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle));
drivers/media/pci/intel/ipu6/ipu6-dma.c
276
for (i = 0; i < PHYS_PFN(size); i++) {
drivers/media/pci/intel/ipu6/ipu6-dma.c
337
PHYS_PFN(sg_dma_address(sglist)));
drivers/media/pci/intel/ipu6/ipu6-dma.c
401
PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0);
drivers/media/pci/intel/ipu6/ipu6-dma.c
60
int count = PHYS_PFN(size);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
425
PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
713
iova = find_iova(&dmap->iovad, PHYS_PFN(mmu->iova_trash_page));
drivers/misc/vmw_balloon.c
795
pfn = PHYS_PFN(virt_to_phys(b->batch_page));
drivers/nvdimm/pfn_devs.c
658
unsigned long base_pfn = PHYS_PFN(base);
drivers/nvdimm/pfn_devs.c
666
unsigned long base_pfn = PHYS_PFN(base);
drivers/nvdimm/pfn_devs.c
688
.end_pfn = PHYS_PFN(end),
drivers/nvdimm/pfn_devs.c
701
nd_pfn->npfns = PHYS_PFN((range_len(range) - offset));
drivers/nvdimm/pfn_devs.c
708
altmap->free = PHYS_PFN(offset - reserve);
drivers/nvdimm/pfn_devs.c
760
npfns = PHYS_PFN(size - SZ_8K);
drivers/nvdimm/pfn_devs.c
815
npfns = PHYS_PFN(size - offset - end_trunc);
drivers/nvdimm/pmem.c
256
*pfn = PHYS_PFN(pmem->phys_addr + offset);
drivers/nvdimm/pmem.c
270
actual_nr = PHYS_PFN(
drivers/nvdimm/pmem.c
285
return PHYS_PFN(pmem->size - pmem->pfn_pad - offset);
drivers/nvdimm/pmem.c
73
pfn_start = PHYS_PFN(phys);
drivers/nvdimm/pmem.c
74
pfn_end = pfn_start + PHYS_PFN(len);
drivers/nvdimm/region_devs.c
35
unsigned long pfn = PHYS_PFN(res->start);
drivers/nvdimm/region_devs.c
41
unsigned long pfn_j = PHYS_PFN(res_j->start);
drivers/platform/x86/intel/pmc/core.c
1300
if (page_is_ram(PHYS_PFN(pmc->base_addr)))
drivers/ras/amd/atl/umc.c
325
addr = PHYS_PFN(addr);
drivers/staging/media/ipu7/ipu7-dma.c
103
unsigned int count = PHYS_PFN(size);
drivers/staging/media/ipu7/ipu7-dma.c
172
count = PHYS_PFN(size);
drivers/staging/media/ipu7/ipu7-dma.c
175
PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0);
drivers/staging/media/ipu7/ipu7-dma.c
249
struct iova *iova = find_iova(&mmu->dmap->iovad, PHYS_PFN(dma_handle));
drivers/staging/media/ipu7/ipu7-dma.c
276
for (i = 0; i < PHYS_PFN(size); i++) {
drivers/staging/media/ipu7/ipu7-dma.c
337
PHYS_PFN(sg_dma_address(sglist)));
drivers/staging/media/ipu7/ipu7-dma.c
417
PHYS_PFN(mmu->dmap->mmu_info->aperture_end),
drivers/staging/media/ipu7/ipu7-dma.c
60
unsigned int count = PHYS_PFN(size);
drivers/staging/media/ipu7/ipu7-mmu.c
408
PHYS_PFN(mmu->dmap->mmu_info->aperture_end), 0);
drivers/staging/media/ipu7/ipu7-mmu.c
758
iova = find_iova(&dmap->iovad, PHYS_PFN(mmu->iova_trash_page));
drivers/tee/tee_heap.c
472
if (!pfn_valid(PHYS_PFN(paddr)))
drivers/vfio/pci/nvgrace-gpu/main.c
105
start_pfn = PHYS_PFN(region->memphys);
drivers/vfio/pci/nvgrace-gpu/main.c
156
pfn = PHYS_PFN(region->memphys);
drivers/vfio/pci/nvgrace-gpu/main.c
313
pfn = PHYS_PFN(memregion->memphys) + addr_to_pgoff(vma, addr);
drivers/virtio/virtio_mem.c
2620
unit_pages = PHYS_PFN(memory_block_size_bytes());
drivers/virtio/virtio_mem.c
2622
unit_pages = PHYS_PFN(vm->bbm.bb_size);
fs/cramfs/inode.c
425
PHYS_PFN(address + off));
fs/dax.c
1000
return PHYS_PFN(iomap->addr + (pos & PAGE_MASK) - iomap->offset);
fs/dax.c
1248
length = dax_direct_access(iomap->dax_dev, pgoff, PHYS_PFN(size),
fs/dax.c
1259
if (*pfnp & (PHYS_PFN(size)-1))
fs/dax.c
1647
map_len = dax_direct_access(dax_dev, pgoff, PHYS_PFN(size),
fs/dax.c
1651
PHYS_PFN(size), DAX_RECOVERY_WRITE,
fs/fuse/dax.c
1209
nr_pages = dax_direct_access(fcd->dev, 0, PHYS_PFN(dax_size),
fs/fuse/virtio_fs.c
1017
*pfn = PHYS_PFN(fs->window_phys_addr + offset);
include/asm-generic/memory_model.h
70
#define __phys_to_pfn(paddr) PHYS_PFN(paddr)
include/asm-generic/memory_model.h
87
#define phys_to_page(phys) pfn_to_page(PHYS_PFN(phys))
include/linux/io-mapping.h
77
return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot);
include/linux/io-mapping.h
98
return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot);
include/linux/mmzone.h
2191
if (PHYS_PFN(PFN_PHYS(pfn)) != pfn)
kernel/dma/contiguous.c
154
unsigned long total_pages = PHYS_PFN(memblock_phys_mem_size());
kernel/dma/direct.c
37
return pfn_to_page(PHYS_PFN(dma_to_phys(dev, dma_addr)));
kernel/dma/direct.c
523
unsigned long pfn = PHYS_PFN(dma_to_phys(dev, dma_addr));
kernel/dma/mapping.c
360
WARN_ON_ONCE(pfn_valid(PHYS_PFN(phys_addr))))
kernel/dma/swiotlb.c
1816
if (PageHighMem(pfn_to_page(PHYS_PFN(rmem->base)))) {
kernel/iomem.c
31
unsigned long pfn = PHYS_PFN(offset);
kernel/liveupdate/kexec_handover.c
1386
unsigned long base_pfn = PHYS_PFN(kho_scratch[i].addr);
kernel/liveupdate/kexec_handover.c
243
struct page *page = pfn_to_online_page(PHYS_PFN(phys));
kernel/liveupdate/kexec_handover.c
306
const unsigned long start_pfn = PHYS_PFN(phys);
kernel/liveupdate/kexec_handover.c
967
unsigned long pfn = PHYS_PFN(virt_to_phys(chunk));
kernel/liveupdate/kexec_handover.c
972
pfn = PHYS_PFN(chunk->phys[i]);
kernel/resource.c
1288
if (devmem_is_allowed(PHYS_PFN(res->start)) &&
kernel/resource.c
1289
devmem_is_allowed(PHYS_PFN(res->end))) {
mm/cma.c
678
cmrp->base_pfn = PHYS_PFN(mlp->base);
mm/damon/paddr.c
178
folio = damon_get_folio(PHYS_PFN(addr));
mm/damon/paddr.c
219
folio = damon_get_folio(PHYS_PFN(addr));
mm/damon/paddr.c
267
folio = damon_get_folio(PHYS_PFN(addr));
mm/damon/paddr.c
303
folio = damon_get_folio(PHYS_PFN(addr));
mm/damon/paddr.c
42
struct folio *folio = damon_get_folio(PHYS_PFN(paddr));
mm/damon/paddr.c
72
struct folio *folio = damon_get_folio(PHYS_PFN(paddr));
mm/hugetlb.c
3104
listnode = early_pfn_to_nid(PHYS_PFN(__pa(m)));
mm/hugetlb_vmemmap.c
809
pfn = PHYS_PFN(paddr);
mm/hugetlb_vmemmap.c
851
pfn = PHYS_PFN(phys);
mm/kfence/core.c
611
start_pfn = PHYS_PFN(virt_to_phys(__kfence_pool));
mm/kmemleak.c
1724
if (PHYS_PFN(phys) < min_low_pfn ||
mm/kmemleak.c
1725
PHYS_PFN(phys + object->size) > max_low_pfn)
mm/madvise.c
1191
PHYS_PFN(range->end - range->start);
mm/memblock.c
1830
return PHYS_PFN(memblock_phys_mem_size() -
mm/memory_hotplug.c
1405
if (memmap_pages == PHYS_PFN(memory_block_size_bytes()))
mm/memory_hotplug.c
1456
.base_pfn = PHYS_PFN(cur_start),
mm/memory_hotplug.c
1457
.end_pfn = PHYS_PFN(cur_start + memblock_size - 1),
mm/memory_hotplug.c
55
return PHYS_PFN(memory_block_size_bytes()) * sizeof(struct page);
mm/memremap.c
100
PHYS_PFN(range_len(range)), NULL);
mm/memremap.c
108
pfnmap_untrack(PHYS_PFN(range->start), range_len(range));
mm/memremap.c
156
conflict_pgmap = get_dev_pagemap(PHYS_PFN(range->start));
mm/memremap.c
163
conflict_pgmap = get_dev_pagemap(PHYS_PFN(range->end));
mm/memremap.c
180
error = xa_err(xa_store_range(&pgmap_array, PHYS_PFN(range->start),
mm/memremap.c
181
PHYS_PFN(range->end), pgmap, GFP_KERNEL));
mm/memremap.c
188
error = pfnmap_track(PHYS_PFN(range->start), range_len(range),
mm/memremap.c
212
error = add_pages(nid, PHYS_PFN(range->start),
mm/memremap.c
213
PHYS_PFN(range_len(range)), params);
mm/memremap.c
229
move_pfn_range_to_zone(zone, PHYS_PFN(range->start),
mm/memremap.c
230
PHYS_PFN(range_len(range)), params->altmap,
mm/memremap.c
243
PHYS_PFN(range->start),
mm/memremap.c
244
PHYS_PFN(range_len(range)), pgmap);
mm/memremap.c
254
pfnmap_untrack(PHYS_PFN(range->start), range_len(range));
mm/memremap.c
407
pgmap = xa_load(&pgmap_array, PHYS_PFN(phys));
mm/memremap.c
43
xa_store_range(&pgmap_array, PHYS_PFN(range->start), PHYS_PFN(range->end),
mm/memremap.c
51
unsigned long pfn = PHYS_PFN(range->start);
mm/memremap.c
65
if (pfn >= PHYS_PFN(range->start) &&
mm/memremap.c
66
pfn <= PHYS_PFN(range->end))
mm/memremap.c
96
remove_pfn_range_from_zone(page_zone(first_page), PHYS_PFN(range->start),
mm/memremap.c
97
PHYS_PFN(range_len(range)));
mm/memremap.c
99
__remove_pages(PHYS_PFN(range->start),
mm/mm_init.c
1830
start_pfn = PHYS_PFN(memblock_start_of_DRAM());
mm/mm_init.c
415
if (usable_startpfn < PHYS_PFN(SZ_4G)) {
mm/numa_emulation.c
206
unsigned long max_pfn = PHYS_PFN(max_addr);
mm/numa_emulation.c
207
unsigned long base_pfn = PHYS_PFN(base);
mm/numa_emulation.c
208
unsigned long hole_pfns = PHYS_PFN(hole);
mm/sparse-vmemmap.c
167
ptpfn = PHYS_PFN(__pa(p));
mm/sparse-vmemmap.c
484
PHYS_PFN(pgmap->ranges[pgmap->nr_range].start);
mm/vma.c
197
pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start);
mm/vma.c
2891
VMG_STATE(vmg, mm, vmi, addr, addr + len, vm_flags, PHYS_PFN(addr));
mm/vma.c
54
.pglen = PHYS_PFN(len_), \
mm/vma.c
709
pgoff = adjust->vm_pgoff + PHYS_PFN(vmg->end - adjust->vm_start);
mm/vma.c
712
pgoff = adjust->vm_pgoff - PHYS_PFN(adjust->vm_start - vmg->end);
mm/vma.c
956
pgoff_t pglen = PHYS_PFN(vmg->end - vmg->start);
mm/vma.h
233
return vma->vm_pgoff + PHYS_PFN(addr - vma->vm_start);
security/integrity/ima/ima_kexec.c
316
start_pfn = PHYS_PFN(phys);
security/integrity/ima/ima_kexec.c
317
end_pfn = PHYS_PFN(end_phys);
tools/testing/nvdimm/dax-dev.c
20
pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1;
tools/testing/nvdimm/pmem-dax.c
42
*pfn = PHYS_PFN(pmem->phys_addr + offset);
tools/testing/nvdimm/pmem-dax.c
50
return PHYS_PFN(pmem->size - pmem->pfn_pad - offset);