Symbol: PFN_PHYS
arch/alpha/kernel/setup.c
308
memblock_add(PFN_PHYS(cluster->start_pfn),
arch/alpha/kernel/setup.c
315
memblock_reserve(PFN_PHYS(cluster->start_pfn),
arch/alpha/kernel/setup.c
356
if ((void *)initrd_end > phys_to_virt(PFN_PHYS(max_low_pfn))) {
arch/alpha/kernel/setup.c
357
if (!move_initrd(PFN_PHYS(max_low_pfn)))
arch/alpha/kernel/setup.c
361
phys_to_virt(PFN_PHYS(max_low_pfn)));
arch/arm/mm/ioremap.c
318
if (WARN_ON(memblock_is_map_memory(PFN_PHYS(pfn)) &&
arch/arm64/include/asm/vmalloc.h
40
if (!IS_ALIGNED(PFN_PHYS(pfn), CONT_PTE_SIZE))
arch/arm64/kernel/setup.c
260
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/arm64/kernel/setup.c
261
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/arm64/mm/init.c
166
phys_addr_t addr = PFN_PHYS(pfn);
arch/arm64/mm/mmu.c
2034
unsigned long start = PFN_PHYS(pfn);
arch/csky/include/asm/page.h
82
#define pfn_to_kaddr(x) __va(PFN_PHYS(x))
arch/csky/kernel/setup.c
100
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
arch/csky/kernel/setup.c
25
if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
arch/hexagon/kernel/dma.c
41
return dma_init_global_coherent(PFN_PHYS(max_low_pfn),
arch/hexagon/mm/init.c
136
memblock_reserve(PFN_PHYS(max_low_pfn), DMA_RESERVED_BYTES);
arch/loongarch/kernel/mem.c
51
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
arch/loongarch/kernel/setup.c
409
dma_contiguous_reserve(PFN_PHYS(max_low_pfn));
arch/loongarch/kernel/setup.c
418
early_memtest(PFN_PHYS(ARCH_PFN_OFFSET), PFN_PHYS(max_low_pfn));
arch/loongarch/kernel/setup.c
555
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/loongarch/kernel/setup.c
556
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/loongarch/mm/init.c
58
unsigned long addr = PFN_PHYS(pfn);
arch/microblaze/include/asm/page.h
96
# define pfn_to_phys(pfn) (PFN_PHYS(pfn))
arch/mips/include/asm/mach-generic/spaces.h
26
# define PHYS_OFFSET ((unsigned long)PFN_PHYS(ARCH_PFN_OFFSET))
arch/mips/kernel/setup.c
222
if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
arch/mips/kernel/setup.c
661
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
arch/mips/kernel/setup.c
670
dma_contiguous_reserve(PFN_PHYS(max_low_pfn));
arch/mips/kernel/setup.c
676
early_memtest(PFN_PHYS(ARCH_PFN_OFFSET), PFN_PHYS(max_low_pfn));
arch/mips/mm/init.c
285
cfg->lower = ALIGN(PFN_PHYS(start_pfn), maar_align);
arch/mips/mm/init.c
286
cfg->upper = ALIGN_DOWN(PFN_PHYS(start_pfn + nr_pages), maar_align) - 1;
arch/mips/mm/init.c
431
memblock_remove(PFN_PHYS(highstart_pfn), -1);
arch/mips/mm/init.c
438
if (!memblock_is_memory(PFN_PHYS(tmp)))
arch/mips/mm/init.c
477
void *addr = phys_to_virt(PFN_PHYS(pfn));
arch/mips/sgi-ip27/ip27-memory.c
343
memblock_add_node(PFN_PHYS(slot_getbasepfn(node, slot)),
arch/mips/sgi-ip27/ip27-memory.c
344
PFN_PHYS(slot_psize), node,
arch/parisc/kernel/cache.c
107
#define pfn_va(pfn) __va(PFN_PHYS(pfn))
arch/parisc/kernel/cache.c
637
__flush_cache_page(vma, vmaddr, PFN_PHYS(pte_pfn(pte)));
arch/parisc/kernel/cache.c
654
__flush_cache_page(vma, vaddr, PFN_PHYS(page_to_pfn(from)));
arch/parisc/kernel/cache.c
663
__flush_cache_page(vma, user_vaddr, PFN_PHYS(page_to_pfn(page)));
arch/parisc/kernel/cache.c
671
__flush_cache_page(vma, user_vaddr, PFN_PHYS(page_to_pfn(page)));
arch/parisc/kernel/cache.c
773
__flush_cache_page(vma, vmaddr, PFN_PHYS(pfn));
arch/parisc/kernel/cache.c
781
__flush_cache_page(vma, vmaddr, PFN_PHYS(page_to_pfn(page)));
arch/parisc/kernel/cache.c
793
__flush_cache_page(vma, addr, PFN_PHYS(pte_pfn(pte)));
arch/parisc/kernel/cache.c
812
__flush_cache_page(vma, addr, PFN_PHYS(pfn));
arch/powerpc/include/asm/nohash/32/mmu-8xx.h
230
if (!IS_ALIGNED(PFN_PHYS(pfn), size))
arch/powerpc/include/asm/pte-walk.h
53
pa = PFN_PHYS(pte_pfn(*ptep));
arch/powerpc/kernel/fadump.c
1186
PFN_PHYS(start_pfn), PFN_PHYS(end_pfn));
arch/powerpc/kexec/crash.c
451
base_addr = PFN_PHYS(mn->start_pfn);
arch/powerpc/mm/mem.c
368
if (iomem_is_exclusive(PFN_PHYS(pfn)))
arch/powerpc/mm/numa.c
1050
memblock_set_node(PFN_PHYS(start_pfn),
arch/powerpc/mm/numa.c
1051
PFN_PHYS(end_pfn - start_pfn),
arch/powerpc/platforms/powernv/memtrace.c
126
arch_remove_linear_mapping(PFN_PHYS(start_pfn), size);
arch/powerpc/platforms/powernv/memtrace.c
128
return PFN_PHYS(start_pfn);
arch/riscv/include/asm/page.h
183
#define pfn_to_phys(pfn) (PFN_PHYS(pfn))
arch/riscv/kernel/setup.c
248
start = max(PFN_PHYS(PFN_DOWN(r_start)), mem->start);
arch/riscv/kernel/setup.c
249
end = min(PFN_PHYS(PFN_UP(r_end)) - 1, mem->end);
arch/riscv/mm/init.c
293
dma32_phys_limit = min(4UL * SZ_1G, (unsigned long)PFN_PHYS(max_low_pfn));
arch/riscv/mm/init.c
550
pte_phys = PFN_PHYS(_pmd_pfn(pmdp[pmd_idx]));
arch/riscv/mm/init.c
652
next_phys = PFN_PHYS(_pud_pfn(pudp[pud_index]));
arch/riscv/mm/init.c
678
next_phys = PFN_PHYS(_p4d_pfn(p4dp[p4d_index]));
arch/riscv/mm/init.c
735
next_phys = PFN_PHYS(_pgd_pfn(pgdp[pgd_idx]));
arch/riscv/mm/ptdump.c
282
u64 pa = PFN_PHYS(pte_pfn(__pte(val)));
arch/s390/kvm/dat.h
395
union pte res = { .val = PFN_PHYS(pfn) };
arch/s390/kvm/dat.h
409
union crste res = { .val = PFN_PHYS(pfn) };
arch/s390/kvm/dat.h
433
union crste res = { .val = PFN_PHYS(pfn) & _SEGMENT_MASK };
arch/s390/kvm/gaccess.c
1108
context->exception = __cmpxchg_with_key(__va(PFN_PHYS(f->pfn) | context->offset),
arch/s390/kvm/gaccess.c
872
ptr = __va(PFN_PHYS(f->pfn) | context->offset);
arch/sh/kernel/setup.c
235
memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn),
arch/sh/mm/init.c
250
memblock_reserve(start, (PFN_PHYS(start_pfn) + PAGE_SIZE - 1) - start);
arch/um/include/asm/page.h
89
#define pfn_to_phys(pfn) PFN_PHYS(pfn)
arch/x86/include/asm/xen/page.h
233
return XMADDR(PFN_PHYS(pfn_to_mfn(PFN_DOWN(phys.paddr))) | offset);
arch/x86/include/asm/xen/page.h
239
return XPADDR(PFN_PHYS(mfn_to_pfn(PFN_DOWN(machine.maddr))) | offset);
arch/x86/kernel/cpu/sgx/encl.c
125
shmem_truncate_range(inode, PFN_PHYS(page_index), PFN_PHYS(page_index) + PAGE_SIZE - 1);
arch/x86/kernel/cpu/sgx/encl.c
154
pcmd_first_page = PFN_PHYS(page_index & ~PCMD_FIRST_MASK) + encl->base;
arch/x86/kernel/e820.c
1370
memblock_remove(PFN_PHYS(max_pfn), -1);
arch/x86/kernel/kvm.c
956
u64 tolud = PFN_PHYS(e820__end_of_low_ram_pfn());
arch/x86/kernel/machine_kexec_64.c
657
control = PFN_PHYS(page_to_pfn(kexec_crash_image->control_code_page));
arch/x86/kernel/setup.c
328
PFN_PHYS(max_pfn_mapped));
arch/x86/mm/amdtopology.c
59
u64 prevbase, start = PFN_PHYS(0);
arch/x86/mm/amdtopology.c
60
u64 end = PFN_PHYS(max_pfn);
arch/x86/mm/init.c
579
u64 start = clamp_val(PFN_PHYS(start_pfn), r_start, r_end);
arch/x86/mm/init.c
580
u64 end = clamp_val(PFN_PHYS(end_pfn), r_start, r_end);
arch/x86/mm/init.c
869
if (region_intersects(PFN_PHYS(pagenr), PAGE_SIZE,
arch/x86/mm/numa.c
217
0LLU, PFN_PHYS(max_pfn) - 1);
arch/x86/mm/numa.c
221
numa_add_memblk(0, 0, PFN_PHYS(max_pfn));
arch/x86/mm/numa.c
462
return PFN_PHYS(MAX_DMA32_PFN);
arch/x86/mm/pat/memtype.c
695
enum page_cache_mode cm = lookup_memtype(PFN_PHYS(pfn));
arch/x86/mm/pat/set_memory.c
1263
if (PFN_PHYS(pfn) & ~PMD_MASK)
arch/x86/mm/pat/set_memory.c
1334
if (!pmd_leaf(first) || (PFN_PHYS(pfn) & ~PUD_MASK))
arch/x86/platform/efi/memmap.c
30
return PFN_PHYS(page_to_pfn(p));
arch/x86/virt/svm/sev.c
855
paddr = PFN_PHYS(pte_pfn(*pte)) | (hva & ~page_level_mask(level));
arch/x86/virt/vmx/tdx/tdx.c
391
start = TDMR_ALIGN_DOWN(PFN_PHYS(tmb->start_pfn));
arch/x86/virt/vmx/tdx/tdx.c
392
end = TDMR_ALIGN_UP(PFN_PHYS(tmb->end_pfn));
arch/x86/virt/vmx/tdx/tdx.c
747
start = PFN_PHYS(tmb->start_pfn);
arch/x86/virt/vmx/tdx/tdx.c
748
end = PFN_PHYS(tmb->end_pfn);
arch/x86/xen/enlighten.c
439
memblock_reserve(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns));
arch/x86/xen/enlighten_hvm.c
76
HYPERVISOR_shared_info = __va(PFN_PHYS(shared_info_pfn));
arch/x86/xen/enlighten_pv.c
1501
initrd_start = PFN_PHYS(xen_start_info->mod_start);
arch/x86/xen/enlighten_pv.c
496
va = __va(PFN_PHYS(pfn));
arch/x86/xen/enlighten_pv.c
575
virt = __va(PFN_PHYS(pfn));
arch/x86/xen/enlighten_pvh.c
113
next->addr = PAGE_ALIGN(e->addr) + PFN_PHYS(pages);
arch/x86/xen/mmu_pv.c
1542
make_lowmem_page_readonly(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1552
make_lowmem_page_readonly(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1560
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1565
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
1882
xen_pt_base = PFN_PHYS(pt_base);
arch/x86/xen/mmu_pv.c
1971
new_area = xen_find_free_area(PFN_PHYS(n_frames));
arch/x86/xen/mmu_pv.c
1986
pmd_phys = pud_phys + PFN_PHYS(n_pud);
arch/x86/xen/mmu_pv.c
1987
pt_phys = pmd_phys + PFN_PHYS(n_pmd);
arch/x86/xen/mmu_pv.c
2053
memblock_phys_free(PFN_PHYS(pfn), PAGE_SIZE * (pfn_end - pfn));
arch/x86/xen/mmu_pv.c
2059
make_lowmem_page_readwrite(__va(PFN_PHYS(pfn)));
arch/x86/xen/mmu_pv.c
2074
paddr = PFN_PHYS(mfn_to_pfn(xen_start_info->store_mfn));
arch/x86/xen/mmu_pv.c
2078
paddr = PFN_PHYS(mfn_to_pfn(xen_start_info->console.domU.mfn));
arch/x86/xen/setup.c
116
memblock_phys_free(PFN_PHYS(start_pfn), PFN_PHYS(n_pfns));
arch/x86/xen/setup.c
636
phys_addr_t mem_end = PFN_PHYS(ini_nr_pages);
arch/x86/xen/setup.c
774
start = PFN_PHYS(xen_start_info->first_p2m_pfn);
arch/x86/xen/setup.c
775
size = PFN_PHYS(xen_start_info->nr_p2m_frames);
arch/x86/xen/setup.c
804
mem_end = PFN_PHYS(ini_nr_pages);
arch/x86/xen/setup.c
909
chunk_size = min(size, PFN_PHYS(extra_pages));
arch/xtensa/mm/init.c
63
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
arch/xtensa/mm/init.c
64
dma_contiguous_reserve(PFN_PHYS(max_low_pfn));
drivers/acpi/apei/ghes.c
184
paddr = PFN_PHYS(pfn);
drivers/base/node.c
860
walk_memory_blocks(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn),
drivers/dax/device.c
75
phys = PFN_PHYS(pgoff - dax_range->pgoff) + range->start;
drivers/gpu/drm/i915/gt/intel_ring_submission.c
74
set_hws_pga(engine, PFN_PHYS(page_to_pfn(status_page(engine))));
drivers/hv/hv_balloon.c
728
nid = memory_add_physaddr_to_nid(PFN_PHYS(start_pfn));
drivers/hv/hv_balloon.c
729
ret = add_memory(nid, PFN_PHYS((start_pfn)),
drivers/hv/mshv_vtl_main.c
400
pgmap->ranges[0].start = PFN_PHYS(vtl0_mem.start_pfn);
drivers/hv/mshv_vtl_main.c
401
pgmap->ranges[0].end = PFN_PHYS(vtl0_mem.last_pfn) - 1;
drivers/iommu/iommufd/pages.c
537
PFN_PHYS(batch->pfns[cur]) + page_offset,
drivers/iommu/iommufd/pages.c
541
PFN_PHYS(batch->pfns[cur]) + page_offset,
drivers/media/pci/intel/ipu6/ipu6-dma.c
197
PFN_PHYS(iova->pfn_lo + i), pci_dma_addr,
drivers/media/pci/intel/ipu6/ipu6-dma.c
213
*dma_handle = PFN_PHYS(iova->pfn_lo);
drivers/media/pci/intel/ipu6/ipu6-dma.c
224
ipu6_iova = PFN_PHYS(iova->pfn_lo + i);
drivers/media/pci/intel/ipu6/ipu6-dma.c
277
ipu6_iova = PFN_PHYS(iova->pfn_lo + i);
drivers/media/pci/intel/ipu6/ipu6-dma.c
284
ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/media/pci/intel/ipu6/ipu6-dma.c
285
PFN_PHYS(iova_size(iova)));
drivers/media/pci/intel/ipu6/ipu6-dma.c
321
ret = vm_insert_page(vma, vma->vm_start + PFN_PHYS(i),
drivers/media/pci/intel/ipu6/ipu6-dma.c
366
ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/media/pci/intel/ipu6/ipu6-dma.c
367
PFN_PHYS(iova_size(iova)));
drivers/media/pci/intel/ipu6/ipu6-dma.c
413
iova_pa = PFN_PHYS(iova_addr);
drivers/media/pci/intel/ipu6/ipu6-dma.c
417
ret = ipu6_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr),
drivers/media/pci/intel/ipu6/ipu6-dma.c
423
sg_dma_address(sg) = PFN_PHYS(iova_addr);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
447
ret = ipu6_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr),
drivers/media/pci/intel/ipu6/ipu6-mmu.c
458
mmu->iova_trash_page = PFN_PHYS(iova->pfn_lo);
drivers/media/pci/intel/ipu6/ipu6-mmu.c
464
ipu6_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/media/pci/intel/ipu6/ipu6-mmu.c
465
PFN_PHYS(iova_size(iova)));
drivers/media/pci/intel/ipu6/ipu6-mmu.c
716
ipu6_mmu_unmap(mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/media/pci/intel/ipu6/ipu6-mmu.c
717
PFN_PHYS(iova_size(iova)));
drivers/nvdimm/pmem.c
246
resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset;
drivers/nvdimm/pmem.c
247
sector_t sector = PFN_PHYS(pgoff) >> SECTOR_SHIFT;
drivers/nvdimm/pmem.c
248
unsigned int num = PFN_PHYS(nr_pages) >> SECTOR_SHIFT;
drivers/nvdimm/pmem.c
299
PFN_PHYS(pgoff) >> SECTOR_SHIFT,
drivers/nvdimm/pmem.c
335
len = PFN_PHYS(PFN_UP(off + bytes));
drivers/nvdimm/pmem.c
336
if (!is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) >> SECTOR_SHIFT, len))
drivers/nvdimm/pmem.c
349
pmem_off = PFN_PHYS(pgoff) + pmem->data_offset;
drivers/nvdimm/pmem.c
438
u64 offset = PFN_PHYS(pfn) - pmem->phys_addr - pmem->data_offset;
drivers/nvdimm/region_devs.c
53
PFN_PHYS(pfn), PAGE_SIZE);
drivers/s390/char/sclp_mem.c
74
return PFN_PHYS(start_pfn) >> ilog2(sclp.rzm);
drivers/staging/media/ipu7/ipu7-dma.c
197
PFN_PHYS(iova->pfn_lo + i), pci_dma_addr,
drivers/staging/media/ipu7/ipu7-dma.c
213
*dma_handle = PFN_PHYS(iova->pfn_lo);
drivers/staging/media/ipu7/ipu7-dma.c
224
ipu7_iova = PFN_PHYS(iova->pfn_lo + i);
drivers/staging/media/ipu7/ipu7-dma.c
277
ipu7_iova = PFN_PHYS(iova->pfn_lo + i);
drivers/staging/media/ipu7/ipu7-dma.c
284
ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/staging/media/ipu7/ipu7-dma.c
285
PFN_PHYS(iova_size(iova)));
drivers/staging/media/ipu7/ipu7-dma.c
321
ret = vm_insert_page(vma, vma->vm_start + PFN_PHYS(i),
drivers/staging/media/ipu7/ipu7-dma.c
366
ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/staging/media/ipu7/ipu7-dma.c
367
PFN_PHYS(iova_size(iova)));
drivers/staging/media/ipu7/ipu7-dma.c
431
iova_pa = PFN_PHYS(iova_addr);
drivers/staging/media/ipu7/ipu7-dma.c
435
ret = ipu7_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr),
drivers/staging/media/ipu7/ipu7-dma.c
441
sg_dma_address(sg) = PFN_PHYS(iova_addr);
drivers/staging/media/ipu7/ipu7-mmu.c
430
ret = ipu7_mmu_map(mmu->dmap->mmu_info, PFN_PHYS(iova_addr),
drivers/staging/media/ipu7/ipu7-mmu.c
441
mmu->iova_trash_page = PFN_PHYS(iova->pfn_lo);
drivers/staging/media/ipu7/ipu7-mmu.c
447
ipu7_mmu_unmap(mmu->dmap->mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/staging/media/ipu7/ipu7-mmu.c
448
PFN_PHYS(iova_size(iova)));
drivers/staging/media/ipu7/ipu7-mmu.c
761
ipu7_mmu_unmap(mmu_info, PFN_PHYS(iova->pfn_lo),
drivers/staging/media/ipu7/ipu7-mmu.c
762
PFN_PHYS(iova_size(iova)));
drivers/vfio/pci/nvgrace-gpu/main.c
380
check_add_overflow(PFN_PHYS(pgoff), req_len, &end))
drivers/vhost/vdpa.c
1159
csize = PFN_PHYS(last_pfn - map_pfn + 1);
drivers/vhost/vdpa.c
1161
PFN_PHYS(map_pfn),
drivers/vhost/vdpa.c
1185
cur_base += PFN_PHYS(pinned);
drivers/vhost/vdpa.c
1190
ret = vhost_vdpa_map(v, iotlb, iova, PFN_PHYS(last_pfn - map_pfn + 1),
drivers/vhost/vdpa.c
1191
PFN_PHYS(map_pfn), perm, NULL);
drivers/virt/acrn/mm.c
232
PFN_PHYS(start_pfn), memmap->len,
drivers/virtio/virtio_mem.c
1303
const unsigned long end = start + PFN_PHYS(1 << order);
drivers/virtio/virtio_mem.c
1315
next = addr + PFN_PHYS(1 << order);
drivers/virtio/virtio_mem.c
1369
if (!virtio_mem_contains_range(vm, addr, PFN_PHYS(1 << order)))
drivers/virtio/virtio_mem.c
2700
uint64_t addr = PFN_PHYS(pfn);
drivers/virtio/virtio_mem.c
2760
chunk_size = max_t(uint64_t, PFN_PHYS(PAGES_PER_SECTION),
drivers/virtio/virtio_mem.c
996
const unsigned long start = PFN_PHYS(mhp->start_pfn);
drivers/virtio/virtio_mem.c
997
const unsigned long size = PFN_PHYS(mhp->nr_pages);
fs/dax.c
1257
if (PFN_PHYS(length) < size)
fs/dax.c
1668
map_len = PFN_PHYS(map_len);
fs/fuse/virtio_fs.c
1011
phys_addr_t offset = PFN_PHYS(pgoff);
include/asm-generic/memory_model.h
71
#define __pfn_to_phys(pfn) PFN_PHYS(pfn)
include/asm-generic/memory_model.h
82
PFN_PHYS(__pfn); \
include/asm-generic/memory_model.h
85
#define page_to_phys(page) PFN_PHYS(page_to_pfn(page))
include/linux/bootmem_info.h
89
kmemleak_free_part_phys(PFN_PHYS(page_to_pfn(page)), PAGE_SIZE);
include/linux/dma-direct.h
135
min(addr, end) < phys_to_dma(dev, PFN_PHYS(min_low_pfn)))
include/linux/mm.h
121
#define page_to_virt(x) __va(PFN_PHYS(page_to_pfn(x)))
include/linux/mmzone.h
2191
if (PHYS_PFN(PFN_PHYS(pfn)) != pfn)
kernel/dma/coherent.c
34
return phys_to_dma(dev, PFN_PHYS(mem->pfn_base));
kernel/liveupdate/kexec_handover.c
313
struct page *page = kho_restore_page(PFN_PHYS(pfn), false);
kernel/power/snapshot.c
2381
(unsigned long long)PFN_PHYS(decoded_pfn));
lib/devmem_is_allowed.c
23
if (iomem_is_exclusive(PFN_PHYS(pfn)))
mm/bootmem_info.c
38
kmemleak_free_part_phys(PFN_PHYS(page_to_pfn(page)), PAGE_SIZE);
mm/cma.c
1071
rstart = PFN_PHYS(cmr->base_pfn);
mm/cma.c
1072
rend = PFN_PHYS(cmr->base_pfn + cmr->count);
mm/cma.c
1135
ret = phys_to_virt(PFN_PHYS(cmr->early_pfn));
mm/cma.c
43
return PFN_PHYS(cma->ranges[0].base_pfn);
mm/memblock.c
1978
int mid = memblock_search(type, PFN_PHYS(pfn));
mm/memfd_luo.c
349
phys = PFN_PHYS(pfolio->pfn);
mm/memfd_luo.c
410
phys = PFN_PHYS(pfolio->pfn);
mm/memory-failure.c
2366
if (!pfn_valid(pfn) && !arch_is_platform_page(PFN_PHYS(pfn))) {
mm/memory_hotplug.c
1096
ret = kasan_add_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages));
mm/memory_hotplug.c
1138
kasan_remove_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages));
mm/memory_hotplug.c
2127
beginpa = PFN_PHYS(section_nr_to_pfn(mem->start_section_nr));
mm/memory_hotplug.c
400
VM_BUG_ON(!mhp_range_allowed(PFN_PHYS(pfn), nr_pages * PAGE_SIZE, false));
mm/memremap.c
404
resource_size_t phys = PFN_PHYS(pfn);
mm/mm_init.c
1801
highmem > PFN_PHYS(arch_zone_lowest_possible_pfn[ZONE_HIGHMEM]))
mm/mm_init.c
1802
highmem = PFN_PHYS(arch_zone_lowest_possible_pfn[ZONE_HIGHMEM]);
mm/mm_init.c
2007
accept_memory(PFN_PHYS(pfn), nr_pages * PAGE_SIZE);
mm/numa_emulation.c
109
size = PFN_PHYS((unsigned long)(size >> PAGE_SHIFT) / nr_nodes);
mm/numa_emulation.c
210
return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes);
mm/numa_emulation.c
378
const u64 max_addr = PFN_PHYS(max_pfn);
mm/numa_emulation.c
41
return PFN_PHYS(absent_pages_in_range(start_pfn, end_pfn));
mm/numa_memblks.c
432
unsigned long node_align_mb = PFN_PHYS(pfn_align) / SZ_1M;
mm/numa_memblks.c
434
unsigned long sect_align_mb = PFN_PHYS(PAGES_PER_SECTION) / SZ_1M;
mm/sparse-vmemmap.c
350
memblock_phys_free(PFN_PHYS(pfn), PAGE_SIZE);
samples/damon/mtier.c
67
range->start = PFN_PHYS(node_start_pfn(target_node));
samples/damon/mtier.c
68
range->end = PFN_PHYS(node_end_pfn(target_node));
tools/testing/nvdimm/dax-dev.c
23
addr = PFN_PHYS(pgoff - dax_range->pgoff) + range->start;
tools/testing/nvdimm/dax-dev.c
32
return PFN_PHYS(page_to_pfn(page));
tools/testing/nvdimm/pmem-dax.c
15
resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset;
tools/testing/nvdimm/pmem-dax.c
17
if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512,
tools/testing/nvdimm/pmem-dax.c
18
PFN_PHYS(nr_pages))))