pmap_l1_index
l1_idx = pmap_l1_index(va);
for (j = pmap_l1_index(sva); j < Ln_ENTRIES; j++) {
l1_slot = pmap_l1_index(state->va);
&bs_state.l1[pmap_l1_index(bs_state.va)],
return (&l1[pmap_l1_index(va)]);
l1e = atomic_load_64(&l1[pmap_l1_index(va)]);
rv = atomic_cmpset_64(&l1[pmap_l1_index(va)], l1e, new_l1e);
l1e = atomic_load_64(&l1[pmap_l1_index(va)]);
pmap_distribute_l1(pmap, pmap_l1_index(va), 0);
for (i = pmap_l1_index(VM_MIN_KERNEL_ADDRESS);
i < pmap_l1_index(VM_MAX_KERNEL_ADDRESS); i++)
for (i = pmap_l1_index(DMAP_MIN_ADDRESS);
i < pmap_l1_index(DMAP_MAX_ADDRESS); i++)
pmap_l1_index(kernel_vm_end), entry);
return (&l1[pmap_l1_index(va)]);
return (&pmap->pm_top[pmap_l1_index(va)]);
for (i = pmap_l1_index(sva); i < Ln_ENTRIES; i++) {
if (i == pmap_l1_index(DMAP_MIN_ADDRESS))
else if (i == pmap_l1_index(VM_MIN_KERNEL_ADDRESS))
if (l2 == NULL || pmap_l1_index(va) != l1slot) {
l1slot = pmap_l1_index(va);
if (l2 == NULL || pmap_l1_index(va) != l1slot) {
l1slot = pmap_l1_index(va);
l1slot = pmap_l1_index(va);
if (l2 == NULL || pmap_l1_index(va) != l1slot) {
l1slot = pmap_l1_index(va);
if (l2 == NULL || pmap_l1_index(va) != l1slot) {
l1slot = pmap_l1_index(va);
slot = pmap_l1_index(KERNBASE);
slot = pmap_l1_index(DEVMAP_MIN_VADDR);