root/sys/arm64/include/vmparam.h
/*-
 * Copyright (c) 1990 The Regents of the University of California.
 * All rights reserved.
 * Copyright (c) 1994 John S. Dyson
 * All rights reserved.
 *
 * This code is derived from software contributed to Berkeley by
 * William Jolitz.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. Neither the name of the University nor the names of its contributors
 *    may be used to endorse or promote products derived from this software
 *    without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 *      from: FreeBSD: src/sys/i386/include/vmparam.h,v 1.33 2000/03/30
 */

#ifdef __arm__
#include <arm/vmparam.h>
#else /* !__arm__ */

#ifndef _MACHINE_VMPARAM_H_
#define _MACHINE_VMPARAM_H_

/*
 * Virtual memory related constants, all in bytes
 */
#ifndef MAXTSIZ
#define MAXTSIZ         (1*1024*1024*1024)      /* max text size */
#endif
#ifndef DFLDSIZ
#define DFLDSIZ         (128*1024*1024)         /* initial data size limit */
#endif
#ifndef MAXDSIZ
#define MAXDSIZ         (1*1024*1024*1024)      /* max data size */
#endif
#ifndef DFLSSIZ
#define DFLSSIZ         (128*1024*1024)         /* initial stack size limit */
#endif
#ifndef MAXSSIZ
#define MAXSSIZ         (1*1024*1024*1024)      /* max stack size */
#endif
#ifndef SGROWSIZ
#define SGROWSIZ        (128*1024)              /* amount to grow stack */
#endif

/*
 * The physical address space is sparsely populated.
 */
#define VM_PHYSSEG_SPARSE

/*
 * The number of PHYSSEG entries.
 */
#define VM_PHYSSEG_MAX          64

/*
 * Create three free page pools: VM_FREEPOOL_DEFAULT is the default pool from
 * which physical pages are allocated and VM_FREEPOOL_DIRECT is the pool from
 * which physical pages for page tables and small UMA objects are allocated.
 * VM_FREEPOOL_LAZYINIT is a special-purpose pool that is populated only during
 * boot and is used to implement deferred initialization of page structures.
 */
#define VM_NFREEPOOL            3
#define VM_FREEPOOL_LAZYINIT    0
#define VM_FREEPOOL_DEFAULT     1
#define VM_FREEPOOL_DIRECT      2

/*
 * Create two free page lists: VM_FREELIST_DMA32 is for physical pages that have
 * physical addresses below 4G, and VM_FREELIST_DEFAULT is for all other
 * physical pages.
 */
#define VM_NFREELIST            2
#define VM_FREELIST_DEFAULT     0
#define VM_FREELIST_DMA32       1

/*
 * When PAGE_SIZE is 4KB, an allocation size of 16MB is supported in order
 * to optimize the use of the direct map by UMA.  Specifically, a 64-byte
 * cache line contains at most 8 L2 BLOCK entries, collectively mapping 16MB
 * of physical memory.  By reducing the number of distinct 16MB "pages" that
 * are used by UMA, the physical memory allocator reduces the likelihood of
 * both 2MB page TLB misses and cache misses during the page table walk when
 * a 2MB page TLB miss does occur.
 *
 * When PAGE_SIZE is 16KB, an allocation size of 32MB is supported.  This
 * size is used by level 0 reservations and L2 BLOCK mappings.
 */
#if PAGE_SIZE == PAGE_SIZE_4K
#define VM_NFREEORDER           13
#elif PAGE_SIZE == PAGE_SIZE_16K
#define VM_NFREEORDER           12
#else
#error Unsupported page size
#endif

/*
 * Enable superpage reservations: 2 levels.
 */
#ifndef VM_NRESERVLEVEL
#define VM_NRESERVLEVEL         2
#endif

/*
 * Level 0 reservations consist of 16 pages when PAGE_SIZE is 4KB, and 128
 * pages when PAGE_SIZE is 16KB.  Level 1 reservations consist of 32 64KB
 * pages when PAGE_SIZE is 4KB, and 16 2M pages when PAGE_SIZE is 16KB.
 */
#if PAGE_SIZE == PAGE_SIZE_4K
#ifndef VM_LEVEL_0_ORDER
#define VM_LEVEL_0_ORDER        4
#endif
#ifndef VM_LEVEL_1_ORDER
#define VM_LEVEL_1_ORDER        5
#endif
#elif PAGE_SIZE == PAGE_SIZE_16K
#ifndef VM_LEVEL_0_ORDER
#define VM_LEVEL_0_ORDER        7
#endif
#ifndef VM_LEVEL_1_ORDER
#define VM_LEVEL_1_ORDER        4
#endif
#else
#error Unsupported page size
#endif

/**
 * Address space layout.
 *
 * ARMv8 implements up to a 48 bit virtual address space. The address space is
 * split into 2 regions at each end of the 64 bit address space, with an
 * out of range "hole" in the middle.
 *
 * We use the full 48 bits for each region, however the kernel may only use
 * a limited range within this space.
 *
 * Upper region:    0xffffffffffffffff  Top of virtual memory
 *
 *                  0xfffffeffffffffff  End of DMAP
 *                  0xffffa00000000000  Start of DMAP
 *
 *                  0xffff027fffffffff  End of KMSAN origin map
 *                  0xffff020000000000  Start of KMSAN origin map
 *
 *                  0xffff017fffffffff  End of KMSAN shadow map
 *                  0xffff010000000000  Start of KMSAN shadow map
 *
 *                  0xffff009fffffffff  End of KASAN shadow map
 *                  0xffff008000000000  Start of KASAN shadow map
 *
 *                  0xffff007fffffffff  End of KVA
 *                  0xffff000000000000  Kernel base address & start of KVA
 *
 * Hole:            0xfffeffffffffffff
 *                  0x0001000000000000
 *
 * Lower region:    0x0000ffffffffffff End of user address space
 *                  0x0000000000000000 Start of user address space
 *
 * We use the upper region for the kernel, and the lower region for userland.
 *
 * We define some interesting address constants:
 *
 * VM_MIN_ADDRESS and VM_MAX_ADDRESS define the start and end of the entire
 * 64 bit address space, mostly just for convenience.
 *
 * VM_MIN_KERNEL_ADDRESS and VM_MAX_KERNEL_ADDRESS define the start and end of
 * mappable kernel virtual address space.
 *
 * VM_MIN_USER_ADDRESS and VM_MAX_USER_ADDRESS define the start and end of the
 * user address space.
 */
#define VM_MIN_ADDRESS          (0x0000000000000000UL)
#define VM_MAX_ADDRESS          (0xffffffffffffffffUL)

/* 512 GiB of kernel addresses */
#define VM_MIN_KERNEL_ADDRESS   (0xffff000000000000UL)
#define VM_MAX_KERNEL_ADDRESS   (0xffff008000000000UL)

/* 128 GiB KASAN shadow map */
#define KASAN_MIN_ADDRESS       (0xffff008000000000UL)
#define KASAN_MAX_ADDRESS       (0xffff00a000000000UL)

/* 512GiB KMSAN shadow map */
#define KMSAN_SHAD_MIN_ADDRESS  (0xffff010000000000UL)
#define KMSAN_SHAD_MAX_ADDRESS  (0xffff018000000000UL)

/* 512GiB KMSAN origin map */
#define KMSAN_ORIG_MIN_ADDRESS  (0xffff020000000000UL)
#define KMSAN_ORIG_MAX_ADDRESS  (0xffff028000000000UL)

/* The address bits that hold a pointer authentication code */
#define PAC_ADDR_MASK           (0x007f000000000000UL)
#define PAC_ADDR_MASK_14        (0xff7f000000000000UL)

/* The top-byte ignore address bits */
#define TBI_ADDR_MASK           0xff00000000000000UL

/* If true addr is in the kernel address space */
#define ADDR_IS_KERNEL(addr)    (((addr) & (1ul << 55)) == (1ul << 55))
/* If true addr is in the user address space */
#define ADDR_IS_USER(addr)      (((addr) & (1ul << 55)) == 0)
/* If true addr is in its canonical form (i.e. no TBI, PAC, etc.) */
#define ADDR_IS_CANONICAL(addr) \
    (((addr) & 0xffff000000000000UL) == 0 || \
     ((addr) & 0xffff000000000000UL) == 0xffff000000000000UL)
#define ADDR_MAKE_CANONICAL(addr) ({                    \
        __typeof(addr) _tmp_addr = (addr);              \
                                                        \
        _tmp_addr &= ~0xffff000000000000UL;             \
        if (ADDR_IS_KERNEL(addr))                       \
                _tmp_addr |= 0xffff000000000000UL;      \
                                                        \
        _tmp_addr;                                      \
})

/* 95 TiB maximum for the direct map region */
#define DMAP_MIN_ADDRESS        (0xffffa00000000000UL)
#define DMAP_MAX_ADDRESS        (0xffffff0000000000UL)

#define DMAP_MIN_PHYSADDR       (dmap_phys_base)
#define DMAP_MAX_PHYSADDR       (dmap_phys_max)

/*
 * Checks to see if a physical address is in the DMAP range.
 * - PHYS_IN_DMAP_RANGE will return true that may be within the DMAP range
 *   but not accessible through the DMAP, e.g. device memory between two
 *   DMAP physical address regions.
 * - PHYS_IN_DMAP will check if DMAP address is mapped before returning true.
 *
 * PHYS_IN_DMAP_RANGE should only be used when a check on the address is
 * performed, e.g. by checking the physical address is within phys_avail,
 * or checking the virtual address is mapped.
 */
#define PHYS_IN_DMAP_RANGE(pa)  ((pa) >= DMAP_MIN_PHYSADDR && \
    (pa) < DMAP_MAX_PHYSADDR)
#define PHYS_IN_DMAP(pa)        (PHYS_IN_DMAP_RANGE(pa) && \
    pmap_klookup(PHYS_TO_DMAP(pa), NULL))
/* True if va is in the dmap range */
#define VIRT_IN_DMAP(va)        ((va) >= DMAP_MIN_ADDRESS && \
    (va) < (dmap_max_addr))

#define PMAP_HAS_DMAP   1
#define PHYS_TO_DMAP(pa)                                                \
({                                                                      \
        KASSERT(PHYS_IN_DMAP_RANGE(pa),                                 \
            ("%s: PA out of range, PA: 0x%lx", __func__,                \
            (vm_paddr_t)(pa)));                                         \
        ((pa) - dmap_phys_base) + DMAP_MIN_ADDRESS;                     \
})

#define DMAP_TO_PHYS(va)                                                \
({                                                                      \
        KASSERT(VIRT_IN_DMAP(va),                                       \
            ("%s: VA out of range, VA: 0x%lx", __func__,                \
            (vm_offset_t)(va)));                                        \
        ((va) - DMAP_MIN_ADDRESS) + dmap_phys_base;                     \
})

#define VM_MIN_USER_ADDRESS     (0x0000000000000000UL)
#define VM_MAX_USER_ADDRESS     (0x0001000000000000UL)

#define VM_MINUSER_ADDRESS      (VM_MIN_USER_ADDRESS)
#define VM_MAXUSER_ADDRESS      (VM_MAX_USER_ADDRESS)

#define KERNBASE                (VM_MIN_KERNEL_ADDRESS)
#define SHAREDPAGE              (VM_MAXUSER_ADDRESS - PAGE_SIZE)
#define USRSTACK                SHAREDPAGE

/*
 * How many physical pages per kmem arena virtual page.
 */
#ifndef VM_KMEM_SIZE_SCALE
#define VM_KMEM_SIZE_SCALE      (1)
#endif

/*
 * Optional ceiling (in bytes) on the size of the kmem arena: 60% of the
 * kernel map.
 */
#ifndef VM_KMEM_SIZE_MAX
#define VM_KMEM_SIZE_MAX        ((VM_MAX_KERNEL_ADDRESS - \
    VM_MIN_KERNEL_ADDRESS + 1) * 3 / 5)
#endif

/*
 * Initial pagein size of beginning of executable file.
 */
#ifndef VM_INITIAL_PAGEIN
#define VM_INITIAL_PAGEIN       16
#endif

#if !defined(KASAN) && !defined(KMSAN)
#define UMA_USE_DMAP
#endif

#ifndef LOCORE

extern vm_paddr_t dmap_phys_base;
extern vm_paddr_t dmap_phys_max;
extern vm_offset_t dmap_max_addr;

#endif

#define ZERO_REGION_SIZE        (64 * 1024)     /* 64KB */

#define DEVMAP_MAX_VADDR        VM_MAX_KERNEL_ADDRESS

/*
 * The pmap can create non-transparent large page mappings.
 */
#define PMAP_HAS_LARGEPAGES     1

/*
 * Need a page dump array for minidump.
 */
#define MINIDUMP_PAGE_TRACKING  1
#define MINIDUMP_STARTUP_PAGE_TRACKING 1

#endif /* !_MACHINE_VMPARAM_H_ */

#endif /* !__arm__ */