MMU_PTOB
baddr = MMU_PTOB(seg_pfn0);
MAKE_DMA_COOKIE(cookie_p, baddr, MMU_PTOB(pfn_no));
MMU_PTOB(seg_pfn0), pfn_no, cookie_no);
if (px_lib_iommu_getbypass(dip, MMU_PTOB(pfn),
if (px_lib_iommu_getbypass(dip, MMU_PTOB(pfn), attr,
(MMU_PTOB(pfn_no + 1) - 1 <= count_max))
xfer_sz = MMU_PTOB(MMU_BTOP(xfer_sz + pg_off)); /* page align */
ASSERT(MMU_PTOB(npages) == mp->dmai_winsize);
mp->dmai_mapping = mp->dmai_roffset | MMU_PTOB(dvma_pg);
MMU_PTOB(npages + PX_HAS_REDZONE(mp)),
mp->dmai_mapping = mp->dmai_roffset | MMU_PTOB(dvma_pg);
MMU_PTOB(npages + PX_HAS_REDZONE(mp)));
dvma_addr = MMU_PTOB(dvma_pg);
MMU_PTOB(npages));
win_p->win_size = MMU_PTOB(end_idx - start_idx + 1);
(MMU_PTOB(pfn_no + 1) - 1 <= count_max))
baddr = MMU_PTOB(seg_pfn0);
MAKE_DMA_COOKIE(cookie_p, baddr, MMU_PTOB(pfn_no));
MMU_PTOB(seg_pfn0), pfn_no);
px_dvma_addr_t dvma_pg = MMU_BTOP(mp->dmai_mapping + MMU_PTOB(index));
mp->dmai_mapping, MMU_PTOB(index), MMU_PTOB(npages));
mp->dmai_mapping, MMU_PTOB(index), MMU_PTOB(npg));
MMU_PTOB(npages), MMU_PAGE_SIZE, 0,
mp->dmai_mapping = MMU_PTOB(dvma_pg);
MMU_PTOB(npages));
dvma_addr = mp->dmai_mapping + MMU_PTOB(index);
cache_size = MMU_PTOB(px_dvma_page_cache_entries *
MMU_PTOB(tsb_entries) - cache_size, MMU_PAGE_SIZE,
va = (caddr_t)(MMU_PTOB(mmu_p->dvma_base_pg + tsb_i));
mp->dmai_mapping = MMU_PTOB(dvma_pg) | win_pg0_off;
pxu_p->tsb_vaddr[tsb_index] = MMU_PTOB(pfn) | attr;
pxu_p->tsb_vaddr[tsb_index] = MMU_PTOB(pfn) | attr;
va = (caddr_t)(MMU_PTOB(mmu_p->dvma_base_pg + obp_tsb_bias +
pfns[i] = MMU_PTOB(PX_ADDR2PFN(addr, pfn_index, flags, i));
ra = MMU_PTOB(MMU_BTOP(ra));
if (hvio_dma_sync(hdl, MMU_PTOB(PX_GET_MP_PFN(mp, idx)) +
MMU_PTOB(PX_GET_MP_PFN(mp, idx)) + pg_off, bytes_to_sync,