pmap_l2_index
l2_idx = pmap_l2_index(va);
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
for (k = pmap_l2_index(sva); k < Ln_ENTRIES; k++) {
l2_slot = pmap_l2_index(state->va);
l2_slot = pmap_l2_index(state->va);
l2 = &l2[pmap_l2_index(va)];
l2p = &l2p[pmap_l2_index(va)];
return (&l2p[pmap_l2_index(va)]);
slot = pmap_l2_index(va);
l2e = atomic_load_64(&l2[pmap_l2_index(va)]);
rv = atomic_cmpset_64(&l2[pmap_l2_index(va)], l2e, new_l2e);
l2e = atomic_load_64(&l2[pmap_l2_index(va)]);
l2 = &l2[pmap_l2_index(va)];
return (&l2[pmap_l2_index(va)]);
for (j = pmap_l2_index(sva); j < Ln_ENTRIES; j++) {
if (l3 == NULL || pmap_l2_index(va) != l2slot) {
l2slot = pmap_l2_index(va);
l2slot = pmap_l2_index(va);
l2slot = pmap_l2_index(va);
if (l3 == NULL || pmap_l2_index(va) != l2slot) {
l2slot = pmap_l2_index(va);
slot = pmap_l2_index(KERNBASE);
slot = pmap_l2_index(freemempos - kernstart + KERNBASE);
slot = pmap_l2_index(DEVMAP_MIN_VADDR);