mmu_ptob
physaddr_t pa = mmu_ptob(pte2mfn(pte, curlevel));
return (mmu_ptob((paddr_t)pfn) | (ma & (MMU_PAGESIZE - 1)));
paddr_t paddr = mmu_ptob((paddr_t)htable->ht_pfn);
paddr = mmu_ptob((physaddr_t)htable.ht_pfn);
paddr = mmu_ptob((physaddr_t)pfn);
len = mmu_ptob(descp->cpd_pages);
size = mmu_ptob(descp->cpd_pages);
dst_off = mmu_ptob(dst_ppn - sfile.low_ppn);
count, mmu_ptob(count));
cpr_pagedata_size = mmu_ptob(CPR_MAXCONTIG + 1);
mmu_ptob(cdump.cdd_dumppgsize + bitmap_pages);
kas_cnt, mmu_ptob(kas_cnt));
dcnt, mmu_ptob(dcnt));
nbytes = mmu_ptob(pages);
STAT->cs_dumped_statefsz += mmu_ptob(npg);
ksize = ndvram + mmu_ptob(npages);
nback = mmu_ptob(STAT->cs_nosw_pages);
paddr = mmu_ptob(pfn);
hat_devload(kas.a_hat, cvaddr, mmu_ptob(1),
pm_msgbuf = kmem_zalloc(mmu_ptob(PMLOGPGS), KM_SLEEP);
pm_bufend = pm_msgbuf + mmu_ptob(PMLOGPGS) - 1;
" page 0x%08x.%08x", mmu_ptob((uint64_t)pp->p_pagenum));
pa = mmu_ptob((uint64_t)pp->p_pagenum);
vpm_cache_size = mmu_ptob((physmem * vpm_cache_percent)/100);
pte = mmu_ptob((x86pte_t)mfn) | PT_VALID | PT_USER
if ((seg == NULL) || ((uaddr + mmu_ptob(req->nr_segments)) >
if ((seg == NULL) || ((uaddr + mmu_ptob(req->nr_segments)) >
pa = mfn_base + mmu_ptob((paddr_t)pfn) + pgoff;
iommu->aiomt_reg_size = mmu_ptob(iommu->aiomt_reg_pages);
va_sz = mmu_ptob(pfn_end - pfn_start + 1);
dhp->dh_len = mmu_ptob(mmu_btopr(length));
if (start == mmu_ptob(mmu_btop(getcr3_pa())))
if (start == mmu_ptob(mmu_btop(getcr3_pa())))
hat_devload(kas.a_hat, virt_base, mmu_ptob(*num_pages), pfn,
hat_devload(kas.a_hat, cvaddr, mmu_ptob(npages),
maxphysaddr = mmu_ptob((paddr_t)physmax) + MMU_PAGEOFFSET;
maxphysaddr = mmu_ptob((paddr_t)HYPERVISOR_memory_op(
if (attr->dma_attr_addr_lo > mmu_ptob((uint64_t)ddiphysmin))
fb->fb_pte_list_pa = mmu_ptob((uint64_t)hat_getpfnum(kas.a_hat,
paddr = mmu_ptob((uint64_t)hat_getpfnum(kas.a_hat,
mmu_ptob((uint64_t)hat_getpfnum(kas.a_hat,
start_addr_pa = mmu_ptob((uint64_t)hat_getpfnum(kas.a_hat,
if (address + mmu_ptob(pages) >= mlp_last_addr) {
mlp_next_addr += mmu_ptob(pages);
hat_devload(kas.a_hat, cvaddr, mmu_ptob(npages), mmu_btop(base),
if (mmu_ptob(kas.a_hat->hat_htable->ht_pfn) > 0xffffffffUL) {
mmu_ptob(kas.a_hat->hat_htable->ht_pfn));
uint64_t physmem_bytes = mmu_ptob(physmax + 1);
const size_t physmem_size = mmu_ptob(physmem);
kpm_size = ROUND_UP_LPAGE(mmu_ptob(physmax + 1));
size = mmu_ptob(segkpsize);
segkvmm_base = segkp_base + mmu_ptob(segkpsize);
size = segkvmmsize != 0 ? mmu_ptob(segkvmmsize) : (physmem_size * 4);
segzio_base = segkvmm_base + mmu_ptob(segkvmmsize);
size = (segziosize != 0) ? mmu_ptob(segziosize) :
ROUND_UP_LPAGE((uintptr_t)segzio_base + mmu_ptob(segziosize));
} else if (seg_attach(&kas, (caddr_t)segkp_base, mmu_ptob(segkpsize),
(void) seg_attach(&kas, segkvmm_base, mmu_ptob(segkvmmsize), &kvmmseg);
segkmem_kvmm_init(segkvmm_base, mmu_ptob(segkvmmsize));
(void) seg_attach(&kas, segzio_base, mmu_ptob(segziosize),
segkmem_zio_init(segzio_base, mmu_ptob(segziosize));
msrval = mmu_ptob(pfn);
uintptr_t vaddr = (uintptr_t)kpm_vbase + mmu_ptob(pfn);
hat_devload(kas.a_hat, kpm_vbase + mmu_ptob(base),
mmu_ptob(end - base), base, PROT_READ | PROT_WRITE,
pte = kdi_ptom(mmu_ptob(mmu_btop(pa))) | PT_VALID;
#define MAKECR3(pfn, pcid) (mmu_ptob(pfn) | pcid)
#define pfn_to_pa(pfn) (mmu_ptob((paddr_t)(pfn)))
paddr_t pa = mmu_ptob((paddr_t)pfn);
addr1 = addr = vmem_alloc(heap_arena, mmu_ptob(pgcnt), VM_SLEEP);
addr = vmem_alloc(heap_arena, mmu_ptob(1), VM_SLEEP);
addr = vmem_alloc(heap_arena, mmu_ptob(1), VM_SLEEP);
vmem_free(heap_arena, cpup->cpu_caddr2, mmu_ptob(1));
vmem_free(heap_arena, cpup->cpu_caddr1, mmu_ptob(1));
mmu_ptob((x86pte_t)xen_info->console.domU.mfn) | PTE_BITS,
addr + mmu_ptob(mme->npages) > seg->s_base + seg->s_size)
pa_to_ma(mmu_ptob(kas.a_hat->hat_htable->ht_pfn));
mmu_ptob(hat_getpfnum(kas.a_hat, (caddr_t)cp->cpu_gdt));
pte = mmu_ptob((x86pte_t)mfn) | PT_VALID | PT_USER | PT_FOREIGN;
dpa = mmu_ptob(hat_getpfnum(kas.a_hat, (caddr_t)ldp)) |
!(mmu_ptob(pfn) & MMU_PAGEOFFSET4M)) {
!(mmu_ptob(pfn) & MMU_PAGEOFFSET512K)) {
!(mmu_ptob(pfn) & MMU_PAGEOFFSET64K)) {
size_t physmem_b = mmu_ptob(physmem);
size = mmu_ptob(segziosize);
va += mmu_ptob(segziosize);
(void) seg_attach(&kas, segzio_base, mmu_ptob(segziosize),
segkmem_zio_init(segzio_base, mmu_ptob(segziosize));
i = mmu_ptob((freemem * segmap_percent) / 100);
if (i > MIN(SEGMAPSIZE, mmu_ptob(freemem)))
i = MIN(SEGMAPSIZE, mmu_ptob(freemem));
test_usum = checksum32(CPR->c_mapping_area, mmu_ptob(pages));
(size_t)mmu_ptob(i_cpr_storage_desc_pgcnt),
(size_t)mmu_ptob(i_cpr_storage_data_sz),
kmem_free((caddr_t)*basepp, mmu_ptob(*pgsp));
descp = *basepp = kmem_alloc(mmu_ptob(npages), KM_NOSLEEP);
len = mmu_ptob(npages);
size_t len = mmu_ptob(npages);
STAT->cs_dumped_statefsz += mmu_ptob(dirty_npages);
clean_sz, mmu_ptob(dirty_npages));
ASSERT(cpd.cpd_length == mmu_ptob(dirty_npages));
ppage_buf = kmem_alloc(mmu_ptob(ppage_count), KM_NOSLEEP);
kmem_free(ppage_buf, mmu_ptob(ppage_count));
mmu_ptob(NHMENTS), mmu_ptob(NHMENTS),
vmem_free(heap_arena, cpr_vaddr, mmu_ptob(NHMENTS));
mmu_ptob(i_cpr_storage_data_sz));
addr + mmu_ptob(i_cpr_storage_data_sz);
addr = kmem_alloc(mmu_ptob(*alloc_pages), KM_NOSLEEP);
mmu_ptob(i_cpr_storage_desc_pgcnt));
mmu_ptob(i_cpr_storage_data_sz));
color = addr_to_vcolor(mmu_ptob(pp->p_pagenum));
size_t allocsz = mmu_ptob(mmu_btopr(mblocksz + mstripesz));
hat_devload(kas.a_hat, kpm_vbase + mmu_ptob(base),
mmu_ptob(end - base), base, PROT_READ | PROT_WRITE,
hat_unload(kas.a_hat, kpm_vbase + mmu_ptob(base), mmu_ptob(end - base),