pmap_l2_pindex
pages = pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
pmap_l2_pindex(seg->start);
pages = pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
pmap_l2_pindex(seg->start);
pvd += pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
pmap_l2_pindex(seg->start);
if (pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) - 1 ==
pmap_l2_pindex(next_seg->start)) {
pmap_l2_pindex(pa) - pmap_l2_pindex(seg->start));
pmap_l2_pindex(VM_PAGE_TO_PHYS(m)) - pmap_l2_pindex(seg->start));
l2pindex = pmap_l2_pindex(va) >> Ln_ENTRIES_SHIFT;
ptepindex = pmap_l2_pindex(va);
nkpg->pindex = pmap_l2_pindex(kernel_vm_end);
ml3 = vm_radix_lookup(&pmap->pm_root, pmap_l2_pindex(sva));
return (vm_radix_remove(&pmap->pm_root, pmap_l2_pindex(va)));
KASSERT(mpte->pindex == pmap_l2_pindex(va),
mp = _pmap_alloc_l3(pmap, pmap_l2_pindex(va), NULL);
mpte = _pmap_alloc_l3(pmap, pmap_l2_pindex(va),
uwptpg->pindex = pmap_l2_pindex(va);
l2pindex = pmap_l2_pindex(va);
l2pindex = pmap_l2_pindex(va);
KASSERT(dstmpte->pindex == pmap_l2_pindex(addr),
ml3->pindex = pmap_l2_pindex(va);
return (vm_radix_remove(&pmap->pm_root, pmap_l2_pindex(va)));
ptepindex = pmap_l2_pindex(va);
nkpg->pindex = pmap_l2_pindex(kernel_vm_end);
mpte->pindex = pmap_l2_pindex(va);
KASSERT(ml3->pindex == pmap_l2_pindex(va),
uwptpg->pindex = pmap_l2_pindex(va);
l2pindex = pmap_l2_pindex(va);