virt_to_pfn
#define ARCH_PFN_OFFSET virt_to_pfn((void *)CONFIG_LINUX_RAM_BASE)
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
#define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr))
min_low_pfn = virt_to_pfn((void *)CONFIG_LINUX_RAM_BASE);
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
&& pfn_valid(virt_to_pfn(kaddr)))
unsigned long nosave_begin_pfn = virt_to_pfn(&__nosave_begin);
unsigned long nosave_end_pfn = virt_to_pfn(&__nosave_end - 1);
pfn_pte(virt_to_pfn(kasan_early_shadow_page),
entry = pfn_pte(virt_to_pfn(p),
entry = pfn_pte(virt_to_pfn(kasan_early_shadow_page),
#define virt_to_page(x) pfn_to_page(virt_to_pfn(x))
__is_lm_address(__addr) && pfn_is_map_memory(virt_to_pfn(__addr)); \
early_pfn_to_nid(virt_to_pfn(lm_alias(KERNEL_START))));
early_pfn_to_nid(virt_to_pfn(start)));
pfn = virt_to_pfn(levels[this_level]);
entry = pfn_pmd(virt_to_pfn(p), PAGE_KERNEL);
pfn_to_page(virt_to_pfn(addr)); \
pte_t pte = pfn_pte(virt_to_pfn((void *)address),
pte_t pte = pfn_pte(virt_to_pfn((void *)address), PAGE_INIT);
ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL);
set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr),
#define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr)))
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
#define virt_addr_valid(kaddr) (pfn_valid(virt_to_pfn(kaddr)))
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
pfn_valid(virt_to_pfn((void *)_addr)); \
pte_t zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL_RO);
pte_t zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL);
pte_t zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL);
zero_pte = pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL_RO);
#define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr)))
(unsigned long)(_addr) >= PAGE_OFFSET && pfn_valid(virt_to_pfn(_addr)); \
unsigned long pfn = virt_to_pfn(pte);
unsigned long pfn = virt_to_pfn(page_address(pte));
unsigned long pfn = virt_to_pfn(pmd);
unsigned long pfn = virt_to_pfn(pud);
unsigned long pfn = virt_to_pfn(pud);
unsigned long pfn = virt_to_pfn(p4d);
unsigned long pfn = virt_to_pfn(p4d);
csr_write(CSR_SATP, virt_to_pfn(mm->pgd) |
csr_write(CSR_SATP, virt_to_pfn(mm->pgd) | satp_mode);
pfn_pte(virt_to_pfn(kasan_early_shadow_page), PAGE_KERNEL));
#define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr))
asce.rsto = virt_to_pfn(table);
newcrste = _crste_fc0(virt_to_pfn(pt), TABLE_TYPE_SEGMENT);
diag10_range(virt_to_pfn((void *)addr), 1);
max_zone_pfns[ZONE_DMA] = virt_to_pfn(MAX_DMA_ADDRESS);
#define virt_to_mfn(v) (pfn_to_mfn(virt_to_pfn(v)))
#define virt_to_gfn(v) (pfn_to_gfn(virt_to_pfn(v)))
pfn = virt_to_pfn((void *)va);
__set_phys_to_machine(virt_to_pfn((void *)vaddr), INVALID_P2M_ENTRY);
out_frames[i] = virt_to_pfn((void *)vaddr);
set_phys_to_machine(virt_to_pfn((void *)vaddr), mfn);
out_frame = virt_to_pfn((void *)vstart);
pfn = virt_to_pfn(xenpmu_data);
pte = _io_pte_entry(virt_to_pfn(addr), _PAGE_TABLE);
const unsigned long pfn = virt_to_pfn(domain->pgd_root);
FIELD_PREP(RISCV_IOMMU_PC_FSC_PPN, virt_to_pfn(domain->pgd_root));
pfn = virt_to_pfn(ret);
pfn_to_gfn(virt_to_pfn(bytes) + i), 0);
pfn_to_gfn(virt_to_pfn((void *)map->active.ring)), 0);
virt_to_pfn(xen_store_interface),
virt_to_pfn(xen_store_interface),
pfn = virt_to_pfn(addr);