MMU_PAGESIZE4M
#define MMU_PAGEOFFSET4M (MMU_PAGESIZE4M - 1)
ASSERT(tsb_size <= MMU_PAGESIZE4M || use_bigtsb_arena);
if (tsb_size > MMU_PAGESIZE4M)
ASSERT(tsbbytes <= MMU_PAGESIZE4M || use_bigtsb_arena);
if (tsbbytes > MMU_PAGESIZE4M)
if (tsbbytes > MMU_PAGESIZE4M) {
if ((len >= MMU_PAGESIZE4M) &&
len -= MMU_PAGESIZE4M;
addr += MMU_PAGESIZE4M;
pfn += MMU_PAGESIZE4M / MMU_PAGESIZE;
vpgsz = MMU_PAGESIZE4M;
size_t tsb_slab_size = MMU_PAGESIZE4M;
#define MAX_BIGKTSB_TTES (TSB_BYTES(MAX_BIGKTSB_SZCODE) / MMU_PAGESIZE4M)
#define DEFAULT_ISM_PAGESIZE MMU_PAGESIZE4M
MMU_PAGESIZE4M * tstat_num4m_mapping);
va += MMU_PAGESIZE4M;
tstat_va[i] = contig_mem_alloc(MMU_PAGESIZE4M);
contig_mem_free(tstat_va[j], MMU_PAGESIZE4M);
contig_mem_free(tstat_va[i], MMU_PAGESIZE4M);
contig_mem_free(tstat_va[i], MMU_PAGESIZE4M);
contig_mem_free(tstat_va[i], MMU_PAGESIZE4M);
va += MMU_PAGESIZE4M;
va += MMU_PAGESIZE4M;
bootaux[BA_LPAGESZ].ba_val = MMU_PAGESIZE4M;
modtext_sz = (caddr_t)roundup((uintptr_t)modtext, MMU_PAGESIZE4M) -
npages = physinstalled - btop(MMU_PAGESIZE4M);
memlist_new(va_to_pa(s_text), MMU_PAGESIZE4M, &memlist);
memlist_add(va_to_pa(s_data), MMU_PAGESIZE4M - ndata_remain_sz,
(void) seg_attach(&kas, (caddr_t)(KERNELBASE + MMU_PAGESIZE4M),
(size_t)(MMU_PAGESIZE4M), &ktexthole);
((uintptr_t)(addr) < KERNELBASE + 2 * MMU_PAGESIZE4M ? 0 : \
(((uintptr_t)(addr) >= KERNELBASE + 2 * MMU_PAGESIZE4M) && \
((uintptr_t)(addr) < KERNELBASE + 3 * MMU_PAGESIZE4M))
(void *)(KERNELBASE + MMU_PAGESIZE4M),
MMU_PAGESIZE4M, PAGESIZE, NULL, NULL, NULL,
nalloc_end = (caddr_t)roundup((uintptr_t)nalloc_base, MMU_PAGESIZE4M);
MMU_PAGESIZE4M) >> MMU_PAGESHIFT4M)
MMU_PAGESIZE4M) >> MMU_PAGESHIFT4M)
return (MMU_PAGESIZE4M);
case MMU_PAGESIZE4M:
return (MMU_PAGESIZE4M);
case MMU_PAGESIZE4M:
copy_sz = roundup(copy_sz, MMU_PAGESIZE4M);
copy_sz += MMU_PAGESIZE4M;
(((va) + MMU_PAGESIZE) <= ((base) + MMU_PAGESIZE4M)))
psz = MMU_PAGESIZE4M;
#define MEMSCRUB_BPP4M MMU_PAGESIZE4M >> MEMSCRUB_BLOCK_SIZE_SHIFT
#define MMU_PAGEOFFSET4M (MMU_PAGESIZE4M - 1)
#define FLUSH_ADDR (KERNELBASE + 2 * MMU_PAGESIZE4M)
ASSERT(va >= datava + MMU_PAGESIZE4M);
ASSERT(tsbsz >= MMU_PAGESIZE4M);
va += MMU_PAGESIZE4M;
tsbsz -= MMU_PAGESIZE4M;
va += MMU_PAGESIZE4M;
size_t max_uheap_lpsize = MMU_PAGESIZE4M;
size_t max_ustack_lpsize = MMU_PAGESIZE4M;
size_t max_privmap_lpsize = MMU_PAGESIZE4M;
size_t max_utext_lpsize = MMU_PAGESIZE4M;
size_t max_shm_lpsize = MMU_PAGESIZE4M;
if (max_uheap_lpsize == MMU_PAGESIZE4M) {
if (max_ustack_lpsize == MMU_PAGESIZE4M) {
if (max_privmap_lpsize == MMU_PAGESIZE4M) {
if (max_shm_lpsize == MMU_PAGESIZE4M) {
} else if (len >= MMU_PAGESIZE4M) { /* 4MB mappings */
align_amount = MMU_PAGESIZE4M;
{MMU_PAGESIZE4M, MMU_PAGESHIFT4M, 0, MMU_PAGESIZE4M >> MMU_PAGESHIFT},
max_uheap_lpsize = MMU_PAGESIZE4M;
max_ustack_lpsize = MMU_PAGESIZE4M;
max_privmap_lpsize = MMU_PAGESIZE4M;
#define MMU_PAGEOFFSET4M (MMU_PAGESIZE4M - 1)
#define FLUSH_ADDR (KERNELBASE + 2 * MMU_PAGESIZE4M)
ASSERT(va >= datava + MMU_PAGESIZE4M);
ASSERT(tsbsz >= MMU_PAGESIZE4M);
va += MMU_PAGESIZE4M;
tsbsz -= MMU_PAGESIZE4M;
va += MMU_PAGESIZE4M;
size_t max_utext_lpsize = MMU_PAGESIZE4M;
size_t max_shm_lpsize = MMU_PAGESIZE4M;
MMU_PAGESIZE4M,
static size_t contig_mem_import_size_max = MMU_PAGESIZE4M;
size_t contig_mem_slab_size = MMU_PAGESIZE4M;
} else if (len >= MMU_PAGESIZE4M) { /* 4MB mappings */
align_amount = MMU_PAGESIZE4M;
lpsize = MMU_PAGESIZE4M;
MMU_PAGESIZE4M);
alloc_base = (caddr_t)roundup((uintptr_t)alloc_base, MMU_PAGESIZE4M);
MMU_PAGESIZE4M) != alloc_base) {
MMU_PAGESIZE4M) != chunkp) {
{MMU_PAGESIZE4M, MMU_PAGESHIFT4M, 0, MMU_PAGESIZE4M >> MMU_PAGESHIFT},