Symbol: pmap_l2_index
sys/arm64/arm64/efirt_machdep.c
128
l2_idx = pmap_l2_index(va);
sys/arm64/arm64/kexec_support.c
180
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/kexec_support.c
182
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/kexec_support.c
185
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
sys/arm64/arm64/pmap.c
10146
for (k = pmap_l2_index(sva); k < Ln_ENTRIES; k++) {
sys/arm64/arm64/pmap.c
1092
l2_slot = pmap_l2_index(state->va);
sys/arm64/arm64/pmap.c
1165
l2_slot = pmap_l2_index(state->va);
sys/arm64/arm64/pmap.c
3008
l2 = &l2[pmap_l2_index(va)];
sys/arm64/arm64/pmap.c
5282
l2p = &l2p[pmap_l2_index(va)];
sys/arm64/arm64/pmap.c
606
return (&l2p[pmap_l2_index(va)]);
sys/arm64/arm64/pmap.c
9890
slot = pmap_l2_index(va);
sys/arm64/vmm/vmm_mmu.c
252
l2e = atomic_load_64(&l2[pmap_l2_index(va)]);
sys/arm64/vmm/vmm_mmu.c
264
rv = atomic_cmpset_64(&l2[pmap_l2_index(va)], l2e, new_l2e);
sys/arm64/vmm/vmm_mmu.c
371
l2e = atomic_load_64(&l2[pmap_l2_index(va)]);
sys/riscv/riscv/pmap.c
3588
l2 = &l2[pmap_l2_index(va)];
sys/riscv/riscv/pmap.c
455
return (&l2[pmap_l2_index(va)]);
sys/riscv/riscv/pmap.c
5559
for (j = pmap_l2_index(sva); j < Ln_ENTRIES; j++) {
sys/riscv/riscv/pmap.c
648
if (l3 == NULL || pmap_l2_index(va) != l2slot) {
sys/riscv/riscv/pmap.c
652
l2slot = pmap_l2_index(va);
sys/riscv/riscv/pmap.c
682
l2slot = pmap_l2_index(va);
sys/riscv/riscv/pmap.c
713
l2slot = pmap_l2_index(va);
sys/riscv/riscv/pmap.c
732
if (l3 == NULL || pmap_l2_index(va) != l2slot) {
sys/riscv/riscv/pmap.c
736
l2slot = pmap_l2_index(va);
sys/riscv/riscv/pmap.c
861
slot = pmap_l2_index(KERNBASE);
sys/riscv/riscv/pmap.c
871
slot = pmap_l2_index(freemempos - kernstart + KERNBASE);
sys/riscv/riscv/pmap.c
894
slot = pmap_l2_index(DEVMAP_MIN_VADDR);