L2_BLOCK
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
pte[pmap_l2_index(tmp)] = (tmp | L2_BLOCK | ATTR_AF | ATTR_S1_UXN);
} else if ((l2e & ATTR_DESCR_MASK) == L2_BLOCK) {
} else if ((l2e & ATTR_DESCR_MASK) == L2_BLOCK) {
if ((l2e & ATTR_DESCR_MASK) == L2_BLOCK) {
ATTR_S1_IDX(VM_MEMATTR_WRITE_BACK) | contig | L2_BLOCK);
pmap_store(l2, PHYS_TO_PTE(pa) | PMAP_SAN_PTE_BITS | L2_BLOCK);
KASSERT((tpte & ATTR_DESCR_MASK) == L2_BLOCK,
PHYS_TO_PTE(pa) | attr | L2_BLOCK, va,
KASSERT((old_l2 & ATTR_DESCR_MASK) == L2_BLOCK,
if ((l3_paddr & ATTR_DESCR_MASK) == L2_BLOCK) {
KASSERT((old_l2 & ATTR_DESCR_MASK) == L2_BLOCK,
if ((pmap_load(l2) & ATTR_DESCR_MASK) == L2_BLOCK) {
pmap_update_entry(pmap, l2, newl2 | L2_BLOCK, va & ~L2_OFFSET, L2_SIZE);
((origpte & ATTR_DESCR_MASK) == L2_BLOCK &&
new_l3 |= L2_BLOCK;
rv = pmap_enter_l2(pmap, va, (new_l3 & ~L3_PAGE) | L2_BLOCK,
if ((pmap_load(l2) & ATTR_DESCR_MASK) == L2_BLOCK &&
L2_BLOCK);
if ((old_l2 & ATTR_DESCR_MASK) == L2_BLOCK) {
if ((old_l2 & ATTR_DESCR_MASK) == L2_BLOCK) {
L2_BLOCK) {
KASSERT((pmap_load(l2p) & ATTR_DESCR_MASK) == L2_BLOCK,
L2_BLOCK)
if ((pmap_load(l2) & ATTR_DESCR_MASK) == L2_BLOCK) {
if ((srcptepaddr & ATTR_DESCR_MASK) == L2_BLOCK) {
if (desc == L2_BLOCK) {
L2_BLOCK,
value |= ATTR_AF | L2_BLOCK;
if (desc == L2_BLOCK && level == 2)
if ((oldl2 & ATTR_DESCR_MASK) == L2_BLOCK) {
ATTR_S1_IDX(VM_MEMATTR_WRITE_BACK) | L2_BLOCK);
if ((pmap_load(l2p) & ATTR_DESCR_MASK) == L2_BLOCK) {
L2_BLOCK), ("Invalid l2 page (%lx != %lx)", l2[0],
ATTR_CONTIGUOUS | (oldl1 & ~ATTR_DESCR_MASK) | L2_BLOCK));
CTASSERT(L1_BLOCK == L2_BLOCK);
KASSERT((oldl2 & ATTR_DESCR_MASK) == L2_BLOCK,
PMAP_SAN_PTE_BITS | L2_BLOCK);
PMAP_SAN_PTE_BITS | L2_BLOCK);
if ((pmap_load(l2) & ATTR_DESCR_MASK) == L2_BLOCK)