_ptob64
uint64_t endpa = _ptob64(physmax + 1);
endpa = _ptob64(physmax + 1);
uint64_t basepa = _ptob64(mp->sbm_basepfn);
endpa = _ptob64(physmax + 1);
_ptob64(s_mp->sbm_basepfn),
_ptob64(s_mp->sbm_npages),
_ptob64(t_mp->sbm_basepfn),
_ptob64(t_mp->sbm_npages),
_ptob64(t_mp->sbm_basepfn),
_ptob64(s_mp->sbm_npages),
_ptob64(s_mp->sbm_basepfn),
_ptob64(t_mp->sbm_npages),
_ptob64(s_mp->sbm_basepfn),
_ptob64(excess),
_ptob64(p),
_ptob64(excess),
ASSERT(x_ml->ml_address == _ptob64(s_mp->sbm_basepfn) +
_ptob64(t_mp->sbm_slice_offset));
_ptob64(s_mq.last_nonrelocatable + 1));
_ptob64(pfn),
_ptob64(t_mp->sbm_npages),
s_basepa = _ptob64(s_mp->sbm_basepfn);
s_nbytes = _ptob64(s_mp->sbm_npages);
t_basepa = _ptob64(t_mp->sbm_basepfn);
_ptob64(t_mp->sbm_slice_offset);
uint64_t basepa = _ptob64(mp->sbm_basepfn);
s_old_basepa = _ptob64(s_mp->sbm_basepfn);
t_old_basepa = _ptob64(t_mp->sbm_basepfn);
if (((t_new_basepa & sm) + _ptob64(t_mp->sbm_npages)) > sz) {
endpa = _ptob64(physmax + 1);
_ptob64(s_mp->sbm_basepfn),
_ptob64(excess));
_ptob64(p),
_ptob64(excess));
_ptob64(s_mq.last_nonrelocatable + 1));
s_del_pa = _ptob64(pfn + t_mp->sbm_npages);
d_ml = memlist_del_span(d_ml, _ptob64(pfn),
s_del_pa - _ptob64(pfn));
t_mp->sbm_cm.sbdev_id, _ptob64(t_mp->sbm_slice_offset),
uint64_t pa = _ptob64(base);
s_old_basepa = _ptob64(s_mp->sbm_basepfn);
t_old_basepa = _ptob64(t_mp->sbm_basepfn);
uint64_t basepa = _ptob64(mp->sbm_basepfn);
endpa = _ptob64(physmax + 1);
_ptob64(mq.first_nonrelocatable),
_ptob64(mq.last_nonrelocatable),
nr_base = _ptob64(mq.first_nonrelocatable);
nr_end = _ptob64(mq.last_nonrelocatable + 1);
endpa = _ptob64(physmax + 1);