pfn_to_pa
*pa = pa_to_ma(pfn_to_pa(hat_getpfnum(as->a_hat, addr)));
*pa = pfn_to_pa(hat_getpfnum(as->a_hat, addr));
return (pfn_to_pa(xen_assign_pfn(btop(paddr))));
paddr = pfn_to_pa(hat_getpfnum(kas.a_hat, vaddr));
paddr = pfn_to_pa(page->p_pagenum) + offset;
paddr = pfn_to_pa(pparray[pcnt]->p_pagenum) + offset;
paddr = pfn_to_pa(hat_getpfnum(vas->a_hat,
paddr = pfn_to_pa(page->p_pagenum);
paddr = pfn_to_pa(pparray[pcnt]->p_pagenum);
paddr = pfn_to_pa(hat_getpfnum(vas->a_hat, vaddr));
pgtable->hwpg_paddr = pfn_to_pa(hat_getpfnum(kas.a_hat, vaddr));
intrmap->intrmap_paddr = pfn_to_pa(
qinv->qinv_table.qinv_mem_paddr = pfn_to_pa(
qinv->qinv_sync.qinv_mem_paddr = pfn_to_pa(
pbase = pfn_to_pa(xen_assign_pfn(mmu_btop(rbase)));
pbase = pfn_to_pa(xen_assign_pfn(mmu_btop(rbase))) |
paddr = pfn_to_pa(pp->p_pagenum) + offset;
paddr = pfn_to_pa(pplist[pcnt]->p_pagenum);
paddr = pfn_to_pa(hat_getpfnum(sglinfo->si_asp->a_hat, vaddr));
paddr = pfn_to_pa(pp->p_pagenum);
paddr = pfn_to_pa(pplist[pcnt]->p_pagenum);
paddr = pfn_to_pa(hat_getpfnum(sglinfo->si_asp->a_hat,
paddr = pfn_to_pa(pp->p_pagenum) + offset;
paddr = pfn_to_pa(pplist[pcnt]->p_pagenum);
paddr = pfn_to_pa(hat_getpfnum(sglinfo->si_asp->a_hat, vaddr));
paddr = pfn_to_pa(pp->p_pagenum);
paddr = pfn_to_pa(pplist[pcnt]->p_pagenum);
paddr = pfn_to_pa(hat_getpfnum(sglinfo->si_asp->a_hat,
paddr = pfn_to_pa(hat_getpfnum(kas.a_hat,
paddr = pfn_to_pa(hat_getpfnum(kas.a_hat, dma->dp_cbaddr)) +
paddr = pfn_to_pa(hat_getpfnum(kas.a_hat,
if (physmem != 0 && high_phys > pfn_to_pa(physmem))
high_phys = pfn_to_pa(physmem);
pa = pfn_to_pa(xen_assign_pfn(mmu_btop(pa))) | (pa & MMU_PAGEOFFSET);
base = pfn_to_pa(xen_assign_pfn(mmu_btop(addr))) |
phys_addr = pfn_to_pa(xen_assign_pfn(mmu_btop(phys_addr))) |
pfn_addr = pfn_to_pa(pfn);
pfn_addr = pfn_to_pa(pfn);
!IS_P2ALIGNED(pfn_to_pa(pfn), pgsize))
p = PT_INDEX_PHYSADDR(pfn_to_pa(ht->ht_pfn), entry);
return (pfn_to_pa(mfn) | (pa & MMU_PAGEOFFSET));
return (pfn_to_pa(pfn) | (ma & MMU_PAGEOFFSET));
*pap = pfn_to_pa(pfn) + (vaddr & MMU_PAGEOFFSET);
*pap = pfn_to_pa(CPU->cpu_current_hat->hat_htable->ht_pfn);
(pa_to_ma(pfn_to_pa(pfn)) | mmu.ptp_bits[(l) + 1])
((pfn_to_pa(pfn & ~PFN_IS_FOREIGN_MFN) | mmu.pte_bits[l]) | \
(pa_to_ma(pfn_to_pa(pfn)) | mmu.pte_bits[l]))
(pfn_to_pa(pfn) | mmu.ptp_bits[(l) + 1])
(pfn_to_pa(pfn) | mmu.pte_bits[l])
ptep = kbm_remap_window(pfn_to_pa(pfn), 0);
ptep = kbm_remap_window(pfn_to_pa(pfn), 0);
caddr_t va = kbm_remap_window(pfn_to_pa(pfn), 1);
ma = pa_to_ma(PT_INDEX_PHYSADDR(pfn_to_pa(ht->ht_pfn), entry));
ma = pa_to_ma(PT_INDEX_PHYSADDR(pfn_to_pa(
ma = pa_to_ma(PT_INDEX_PHYSADDR(pfn_to_pa(ht->ht_pfn), entry));
set_pteval(pfn_to_pa(dest->ht_pfn), entry,
pfn_to_pa(dest->ht_hat->hat_user_ptable),
block_zero_no_xmm(kpm_vbase + pfn_to_pa(hat->hat_user_ptable),
x86pte_t pte_val = pa_to_ma(pfn_to_pa(pfn)) | PT_WRITABLE |
return (address_in_memlist(phys_install, pfn_to_pa(pf), 1));
if (pa_to_ma(pfn_to_pa(pp->p_pagenum)) <
if (pa_to_ma(pfn_to_pa(pp->p_pagenum)) >=
pgaddr = pa_to_ma(pfn_to_pa(pp->p_pagenum));
pgaddr = pa_to_ma(pfn_to_pa(pp->p_pagenum));
metasz = pfn_to_pa(metapgs);
memlist_add(pfn_to_pa(meta_start), num_pages, &mem->memlist,
gdtpa = pfn_to_pa(va_to_pfn(bgdt));
mach_addr = pa_to_ma(pfn_to_pa(va_to_pfn(
pa_to_ma(pfn_to_pa(va_to_pfn((caddr_t)mfn_list + off)));
kbm_map_ma(pfn_to_pa(xen_info->console.domU.mfn), addr, 0);
return (pfn_to_pa(pfn) + (ma & MMU_PAGEOFFSET));
pte = pfn_to_pa(pfn) | PT_VALID;
const uint64_t paddr = pfn_to_pa(pfn) & EPT_PA_MASK;
const uint64_t paddr = pfn_to_pa(pfn) & EPT_PA_MASK;
const uint64_t paddr = pfn_to_pa(pfn);
const uint64_t paddr = pfn_to_pa(pfn);
gdtpa = pfn_to_pa(va_to_pfn(gdt0));
CPU->cpu_m.mcpu_gdtpa = pfn_to_pa(va_to_pfn(gdt));