sys/arm64/arm64/minidump_machdep.c
176
for (va = VM_MIN_KERNEL_ADDRESS; va < kva_end; va += L2_SIZE) {
sys/arm64/arm64/minidump_machdep.c
191
va += L1_SIZE - L2_SIZE;
sys/arm64/arm64/minidump_machdep.c
288
for (va = VM_MIN_KERNEL_ADDRESS; va < kva_end; va += L2_SIZE) {
sys/arm64/arm64/minidump_machdep.c
312
tmpbuffer[j] = (pa + i * L2_SIZE +
sys/arm64/arm64/minidump_machdep.c
326
va += L1_SIZE - L2_SIZE;
sys/arm64/arm64/pmap.c
10150
sva += L2_SIZE;
sys/arm64/arm64/pmap.c
10162
sva += L2_SIZE;
sys/arm64/arm64/pmap.c
1132
if ((physmap[i + 1] - state->pa) < L2_SIZE)
sys/arm64/arm64/pmap.c
1141
(physmap[i + 1] - state->pa) >= L2_SIZE;
sys/arm64/arm64/pmap.c
1142
state->va += L2_SIZE, state->pa += L2_SIZE) {
sys/arm64/arm64/pmap.c
1289
(physmap[i + 1] - bs_state.pa) >= L2_SIZE) {
sys/arm64/arm64/pmap.c
1334
for (; bs_state.va < VM_MAX_KERNEL_ADDRESS; bs_state.va += L2_SIZE)
sys/arm64/arm64/pmap.c
1362
preinit_map_va = roundup2(bs_state.freemempos, L2_SIZE);
sys/arm64/arm64/pmap.c
1437
pa = rounddown2(end_pa - L2_SIZE, L2_SIZE);
sys/arm64/arm64/pmap.c
1438
for (; pa >= start_pa && va < eva; va += L2_SIZE, pa -= L2_SIZE) {
sys/arm64/arm64/pmap.c
1446
pa += L2_SIZE;
sys/arm64/arm64/pmap.c
1450
bzero_early((void *)PHYS_TO_DMAP(pa), L2_SIZE);
sys/arm64/arm64/pmap.c
1451
physmem_exclude_region(pa, L2_SIZE, EXFLAG_NOALLOC);
sys/arm64/arm64/pmap.c
1489
plow = roundup2(physmap[i], L2_SIZE);
sys/arm64/arm64/pmap.c
1495
if (phigh - plow >= L2_SIZE) {
sys/arm64/arm64/pmap.c
1598
pages = pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
sys/arm64/arm64/pmap.c
1613
pages = pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
sys/arm64/arm64/pmap.c
1644
pvd += pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) -
sys/arm64/arm64/pmap.c
1656
if (pmap_l2_pindex(roundup2(seg->end, L2_SIZE)) - 1 ==
sys/arm64/arm64/pmap.c
1798
pagesizes[2] = L2_SIZE;
sys/arm64/arm64/pmap.c
2225
sva = (sva & ~L2_OFFSET) + L2_SIZE;
sys/arm64/arm64/pmap.c
2367
if ((va & L2_OFFSET) == 0 && size >= L2_SIZE &&
sys/arm64/arm64/pmap.c
2390
va += L2_SIZE;
sys/arm64/arm64/pmap.c
2391
pa += L2_SIZE;
sys/arm64/arm64/pmap.c
2392
size -= L2_SIZE;
sys/arm64/arm64/pmap.c
2477
KASSERT(size >= L2_SIZE, ("Insufficient size"));
sys/arm64/arm64/pmap.c
2489
va += L2_SIZE;
sys/arm64/arm64/pmap.c
2491
size -= L2_SIZE;
sys/arm64/arm64/pmap.c
307
#define PMAP_PREINIT_MAPPING_SIZE (PMAP_PREINIT_MAPPING_COUNT * L2_SIZE)
sys/arm64/arm64/pmap.c
3186
addr = roundup2(addr, L2_SIZE);
sys/arm64/arm64/pmap.c
3213
kernel_vm_end = (kernel_vm_end + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
3230
kernel_vm_end = (kernel_vm_end + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
3778
va_last = va + L2_SIZE - PAGE_SIZE;
sys/arm64/arm64/pmap.c
3969
pmap->pm_stats.wired_count -= L2_SIZE / PAGE_SIZE;
sys/arm64/arm64/pmap.c
3970
pmap_resident_count_dec(pmap, L2_SIZE / PAGE_SIZE);
sys/arm64/arm64/pmap.c
3976
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++) {
sys/arm64/arm64/pmap.c
4156
KASSERT(rounddown2(sva, L2_SIZE) + L2_SIZE == roundup2(eva, L2_SIZE),
sys/arm64/arm64/pmap.c
4304
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
4312
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/arm64/arm64/pmap.c
4506
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++)
sys/arm64/arm64/pmap.c
4608
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
4617
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/arm64/arm64/pmap.c
4932
va_last = va + L2_SIZE - PAGE_SIZE;
sys/arm64/arm64/pmap.c
5018
+ L2_SIZE - PAGE_SIZE;
sys/arm64/arm64/pmap.c
5081
pmap_update_entry(pmap, l2, newl2 | L2_BLOCK, va & ~L2_OFFSET, L2_SIZE);
sys/arm64/arm64/pmap.c
5268
KASSERT(pagesizes[psind] == L2_SIZE,
sys/arm64/arm64/pmap.c
5801
if (!pmap_bti_same(pmap, va, va + L2_SIZE, &new_l2)) {
sys/arm64/arm64/pmap.c
5856
pmap_remove_l3_range(pmap, old_l2, va, va + L2_SIZE,
sys/arm64/arm64/pmap.c
5932
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++)
sys/arm64/arm64/pmap.c
5940
pmap->pm_stats.wired_count += L2_SIZE / PAGE_SIZE;
sys/arm64/arm64/pmap.c
5941
pmap->pm_stats.resident_count += L2_SIZE / PAGE_SIZE;
sys/arm64/arm64/pmap.c
5950
L2_SIZE);
sys/arm64/arm64/pmap.c
6226
if ((va & L2_OFFSET) == 0 && va + L2_SIZE <= end &&
sys/arm64/arm64/pmap.c
6230
m = vm_radix_iter_jump(&pages, L2_SIZE / PAGE_SIZE);
sys/arm64/arm64/pmap.c
6502
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
6519
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/arm64/arm64/pmap.c
6521
pmap->pm_stats.wired_count -= L2_SIZE /
sys/arm64/arm64/pmap.c
6695
va_next = (addr + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
6707
addr + L2_SIZE > end_addr)
sys/arm64/arm64/pmap.c
6729
pmap_resident_count_inc(dst_pmap, L2_SIZE /
sys/arm64/arm64/pmap.c
7194
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++)
sys/arm64/arm64/pmap.c
7208
L2_SIZE / PAGE_SIZE);
sys/arm64/arm64/pmap.c
7213
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++)
sys/arm64/arm64/pmap.c
7725
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/arm64/arm64/pmap.c
7978
l2_blocks = (roundup2(pa + size, L2_SIZE) -
sys/arm64/arm64/pmap.c
7979
rounddown2(pa, L2_SIZE)) >> L2_SHIFT;
sys/arm64/arm64/pmap.c
8012
va = preinit_map_va + (start_idx * L2_SIZE);
sys/arm64/arm64/pmap.c
8022
pa = rounddown2(pa, L2_SIZE);
sys/arm64/arm64/pmap.c
8039
va += L2_SIZE;
sys/arm64/arm64/pmap.c
8040
pa += L2_SIZE;
sys/arm64/arm64/pmap.c
8052
va = preinit_map_va + (start_idx * L2_SIZE);
sys/arm64/arm64/pmap.c
8102
(roundup2(va + size, L2_SIZE) - rounddown2(va, L2_SIZE)) >> L2_SHIFT;
sys/arm64/arm64/pmap.c
8117
offset = block * L2_SIZE;
sys/arm64/arm64/pmap.c
8118
va_trunc = rounddown2(va, L2_SIZE) + offset;
sys/arm64/arm64/pmap.c
8295
tmpva = (tmpva & ~L2_OFFSET) + L2_SIZE;
sys/arm64/arm64/pmap.c
8338
(base + size - tmpva) >= L2_SIZE) {
sys/arm64/arm64/pmap.c
8339
pte_size = L2_SIZE;
sys/arm64/arm64/pmap.c
8371
L2C_ENTRIES, pte, tmpva, L2_SIZE, L2C_SIZE);
sys/arm64/arm64/pmap.c
8465
phys += L2_SIZE;
sys/arm64/arm64/pmap.c
8569
if (va <= (vm_offset_t)l2 && va + L2_SIZE > (vm_offset_t)l2) {
sys/arm64/arm64/pmap.c
8789
L2C_SIZE) & ~L2C_OFFSET, L2_SIZE, true);
sys/arm64/arm64/pmap.c
8798
l2e += L2_SIZE;
sys/arm64/arm64/pmap.c
9527
if (size - ((L2_SIZE - superpage_offset) & L2_OFFSET) >= L2_SIZE) {
sys/arm64/arm64/pmap.c
9801
#define SAN_BOOTSTRAP_L2_SIZE (1 * L2_SIZE)
sys/arm64/arm64/pmap.c
9806
static uint8_t bootstrap_data[SAN_BOOTSTRAP_L2_SIZE] __aligned(L2_SIZE);
sys/arm64/arm64/pmap.c
9810
if (offset + L2_SIZE > sizeof(bootstrap_data)) {
sys/arm64/arm64/pmap.c
9816
offset += L2_SIZE;
sys/arm64/arm64/pmap.c
9869
Ln_ENTRIES, 0, ~0ul, L2_SIZE, 0, VM_MEMATTR_DEFAULT));
sys/arm64/include/pte.h
197
#define L2_OFFSET (L2_SIZE - 1)
sys/arm64/include/pte.h
214
#define PMAP_MAPDEV_EARLY_SIZE (L2_SIZE * 20)
sys/arm64/include/pte.h
247
#define L2C_SIZE (L2C_ENTRIES * L2_SIZE)
sys/arm64/vmm/vmm_arm64.c
310
next_hyp_va = roundup2(vmm_base + hyp_code_len, L2_SIZE);
sys/arm64/vmm/vmm_arm64.c
323
next_hyp_va += L2_SIZE;
sys/arm64/vmm/vmm_arm64.c
413
if (vmm_base > (L2_SIZE + PAGE_SIZE)) {
sys/arm64/vmm/vmm_arm64.c
419
vmm_base -= L2_SIZE + PAGE_SIZE;
sys/arm64/vmm/vmm_arm64.c
420
vmm_base = rounddown2(vmm_base, L2_SIZE);
sys/arm64/vmm/vmm_arm64.c
428
if (vmm_base > L2_SIZE)
sys/arm64/vmm/vmm_arm64.c
429
vmem_add(el2_mem_alloc, L2_SIZE, vmm_base - L2_SIZE,
sys/kern/subr_devmap.c
287
if (size >= L2_SIZE && (pa & L2_OFFSET) == 0)
sys/kern/subr_devmap.c
288
va = kva_alloc_aligned(size, L2_SIZE);
sys/riscv/include/pte.h
58
#define L2_OFFSET (L2_SIZE - 1)
sys/riscv/include/vmparam.h
250
#define PMAP_MAPDEV_EARLY_SIZE (4 * L2_SIZE)
sys/riscv/riscv/machdep.c
527
physmem_exclude_region(mr->mr_start, L2_SIZE,
sys/riscv/riscv/machdep.c
552
min(p->md_pages * EFI_PAGE_SIZE, L2_SIZE),
sys/riscv/riscv/minidump_machdep.c
183
for (va = VM_MIN_KERNEL_ADDRESS; va < kva_max; va += L2_SIZE) {
sys/riscv/riscv/minidump_machdep.c
283
for (va = VM_MIN_KERNEL_ADDRESS; va < kva_max; va += L2_SIZE) {
sys/riscv/riscv/pmap.c
1050
pv_npg = howmany(vm_phys_segs[vm_phys_nsegs - 1].end, L2_SIZE);
sys/riscv/riscv/pmap.c
1063
pagesizes[1] = L2_SIZE;
sys/riscv/riscv/pmap.c
1626
4, 0, ~0ul, L2_SIZE, 0, VM_MEMATTR_DEFAULT);
sys/riscv/riscv/pmap.c
1970
addr = roundup2(addr, L2_SIZE);
sys/riscv/riscv/pmap.c
1996
kernel_vm_end = (kernel_vm_end + L2_SIZE) & ~L2_OFFSET;
sys/riscv/riscv/pmap.c
2018
kernel_vm_end = (kernel_vm_end + L2_SIZE) & ~L2_OFFSET;
sys/riscv/riscv/pmap.c
2378
va_last = va + L2_SIZE - PAGE_SIZE;
sys/riscv/riscv/pmap.c
2434
va_last = va + L2_SIZE - PAGE_SIZE;
sys/riscv/riscv/pmap.c
2528
pmap_invalidate_range(pmap, sva, sva + L2_SIZE);
sys/riscv/riscv/pmap.c
2530
pmap->pm_stats.wired_count -= L2_SIZE / PAGE_SIZE;
sys/riscv/riscv/pmap.c
2531
pmap_resident_count_dec(pmap, L2_SIZE / PAGE_SIZE);
sys/riscv/riscv/pmap.c
2536
eva = sva + L2_SIZE;
sys/riscv/riscv/pmap.c
2656
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/riscv/riscv/pmap.c
2664
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/riscv/riscv/pmap.c
2839
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/riscv/riscv/pmap.c
2847
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/riscv/riscv/pmap.c
3179
pa += L2_SIZE - PAGE_SIZE;
sys/riscv/riscv/pmap.c
3613
for (sva = va; sva < va + L2_SIZE; sva += PAGE_SIZE) {
sys/riscv/riscv/pmap.c
3676
for (mt = m; mt < &m[L2_SIZE / PAGE_SIZE]; mt++)
sys/riscv/riscv/pmap.c
3684
pmap->pm_stats.wired_count += L2_SIZE / PAGE_SIZE;
sys/riscv/riscv/pmap.c
3685
pmap->pm_stats.resident_count += L2_SIZE / PAGE_SIZE;
sys/riscv/riscv/pmap.c
3732
if ((va & L2_OFFSET) == 0 && va + L2_SIZE <= end &&
sys/riscv/riscv/pmap.c
3736
m = vm_radix_iter_jump(&pages, L2_SIZE / PAGE_SIZE);
sys/riscv/riscv/pmap.c
3959
va_next = (sva + L2_SIZE) & ~L2_OFFSET;
sys/riscv/riscv/pmap.c
3967
if (sva + L2_SIZE == va_next && eva >= va_next) {
sys/riscv/riscv/pmap.c
5019
tmpva = (tmpva & ~L2_OFFSET) + L2_SIZE;
sys/riscv/riscv/pmap.c
5028
tmpva + L2_SIZE <= base + size) {
sys/riscv/riscv/pmap.c
5029
tmpva += L2_SIZE;
sys/riscv/riscv/pmap.c
5076
tmpva += L2_SIZE;
sys/riscv/riscv/pmap.c
5089
PHYS_TO_DMAP(phys), L2_SIZE, mode);
sys/riscv/riscv/pmap.c
5093
tmpva += L2_SIZE;
sys/riscv/riscv/pmap.c
5270
if (size < L2_SIZE)
sys/riscv/riscv/pmap.c
5275
if (size - ((L2_SIZE - superpage_offset) & L2_OFFSET) < L2_SIZE ||
sys/riscv/riscv/pmap.c
5563
sva += L2_SIZE;
sys/riscv/riscv/pmap.c
5569
sva += L2_SIZE;
sys/riscv/riscv/pmap.c
633
if (roundup(pa, L2_SIZE) + L2_SIZE > endpa)
sys/riscv/riscv/pmap.c
685
pa += L2_SIZE;
sys/riscv/riscv/pmap.c
686
va += L2_SIZE;
sys/riscv/riscv/pmap.c
701
while (pa + L2_SIZE - 1 < endpa) {
sys/riscv/riscv/pmap.c
716
pa += L2_SIZE;
sys/riscv/riscv/pmap.c
717
va += L2_SIZE;
sys/riscv/riscv/pmap.c
814
nkernl2 = howmany(howmany(kernlen, L2_SIZE), Ln_ENTRIES);
sys/riscv/riscv/pmap.c
845
if (freemempos < roundup2(kernend, L2_SIZE))
sys/riscv/riscv/pmap.c
846
freemempos = roundup2(kernend, L2_SIZE);
sys/riscv/riscv/pmap.c
862
for (pa = kernstart; pa < kernend; pa += L2_SIZE, slot++) {