KERNBASE
kernbase = KERNBASE;
kernbase = KERNBASE; /* for old kernels */
#define WARMBOOT_OFF (KERNBASE + 0x0467)
#define WARMBOOT_SEG (KERNBASE + 0x0469)
ASSYM(KERNBASE, KERNBASE);
preload_metadata = (caddr_t)(uintptr_t)(modulep + KERNBASE);
preload_bootstrap_relocate(KERNBASE);
envp += KERNBASE;
kva_end = MAX(KERNBASE + nkpt * NBPDR, kernel_vm_end);
#define WARMBOOT_OFF (KERNBASE + 0x0467)
#define WARMBOOT_SEG (KERNBASE + 0x0469)
kasankernbase = kasan_md_addr_to_shad(KERNBASE);
mpte->pindex = pmap_pde_pindex(KERNBASE) + i;
if (KERNBASE < addr) {
end = KERNBASE + nkpt * NBPDR;
if (end <= KERNBASE)
nkpt = howmany(end - KERNBASE, NBPDR);
*ptr = (orig & KMSAN_ORIG_PTR_MASK) | KERNBASE;
return (addr < VM_MIN_KERNEL_ADDRESS || addr >= KERNBASE);
#define BIOS_PADDRTOVADDR(x) ((x) + KERNBASE)
#define BIOS_VADDRTOPADDR(x) ((x) - KERNBASE)
#define KERNSTART (KERNBASE + NBPDR)
ASSYM(KERNBASE, KERNBASE);
if (kdb_thrctx->pcb_regs.sf_sp > KERNBASE) {
ASSYM(KERNBASE, KERNBASE);
#if defined(EARLY_PRINTF) && defined(SOCDEV_PA) && defined(SOCDEV_VA) && SOCDEV_VA < KERNBASE
#if defined(EARLY_PRINTF) && defined(SOCDEV_PA) && defined(SOCDEV_VA) && SOCDEV_VA < KERNBASE
for (va = KERNBASE; va < kva_end; va += PAGE_SIZE) {
mdhdr.kernbase = KERNBASE;
for (va = KERNBASE; va < kva_end; va += PAGE_SIZE) {
((vm_offset_t)(PT2MAP + (KERNBASE >> PT2MAP_SHIFT)))
pte1_idx = pte1_index(KERNBASE - PT2MAP_SIZE);
pte2p = kern_pt2tab_entry(KERNBASE - PT2MAP_SIZE);
unsigned long ksize = vm_max_kernel_address - KERNBASE;
pte1_copy_nosync(kern_pt1, pmap->pm_pt1, KERNBASE,
pt2tab_copy_nosync(kern_pt2tab, pmap->pm_pt2tab, KERNBASE,
start = pte1_index(KERNBASE) * sizeof(pt1_entry_t);
start = pt2tab_index(KERNBASE) * sizeof(pt2_entry_t);
vm_offset_t kernel_vm_end = KERNBASE + NKPT2PG * NPT2_IN_PG * PTE1_SIZE;
CTASSERT((KERNBASE & (NPT2_IN_PG * PTE1_SIZE - 1)) == 0);
CTASSERT((KERNBASE - VM_MAXUSER_ADDRESS) >= PT2MAP_SIZE);
pte2p = kern_pt2tab_entry(KERNBASE);
pte1p = kern_pte1(KERNBASE);
pte1p = kern_pte1(KERNBASE);
for (pa = KERNEL_V2P(KERNBASE); pa < last; pa += PTE1_SIZE)
PT2MAP = (pt2_entry_t *)(KERNBASE - PT2MAP_SIZE);
if (va >= KERNBASE) {
#define VM_MAXUSER_ADDRESS (KERNBASE - 0x00400000) /* !!! PT2MAP_SIZE */
#define VM_MIN_KERNEL_ADDRESS KERNBASE
#ifndef KERNBASE
pmap_bootstrap_dmap(lastaddr - KERNBASE);
stop_mmu(release_addr, pmap_kextract(KERNBASE));
start_pa = pmap_early_vtophys(KERNBASE);
bs_state.freemempos = KERNBASE + kernlen;
pmap_bootstrap_l2(KERNBASE + L1_SIZE);
kernstart = pmap_early_vtophys(KERNBASE);
*ptr = (orig & KMSAN_ORIG_PTR_MASK) | KERNBASE;
trampaddr = KERNBASE;
idt_paddr = (char *)idt - (char *)KERNBASE;
dg.buf = (struct dcons_buf *)(vm_offset_t)(KERNBASE + addr);
fb = KERNBASE + 0xb8000;
ASSYM(KERNBASE, KERNBASE);
addend = (vm_paddr_t)bootinfo.bi_modulep < KERNBASE ?
addend = (vm_paddr_t)bootinfo.bi_envp < KERNBASE ?
for (va = KERNBASE; va < kva_end; va += NBPDR) {
mdhdr.kernbase = KERNBASE;
for (va = KERNBASE; va < kva_end; va += NBPDR) {
_Static_assert(LOWPTDI * 2 * NBPDR == KERNBASE,
pmap_cold_mapident(KERNBASE, atop(KERNend - KERNBASE));
for (a = KERNBASE; a < KERNend; a += NBPDR)
if (va >= PMAP_MAP_LOW && va <= KERNBASE && va + size <= KERNBASE)
KERNBASE <= i << PDRSHIFT && i << PDRSHIFT < KERNend &&
ksize = VM_MAX_KERNEL_ADDRESS - KERNBASE;
mapbase = (caddr_t)KERNBASE;
ef->address = (caddr_t) (__startkernel - KERNBASE);
static unsigned long kern_relbase = KERNBASE;
SYSCTL_NULL_ULONG_PTR, KERNBASE, "Kernel base address");
ef->address = (caddr_t) (__startkernel - KERNBASE);
linker_kernel_file->address += KERNBASE;
mapbase = KERNBASE;
ksymtab_relbase = (__startkernel - KERNBASE);
frame->srr0, frame->srr0 - (register_t)(__startkernel - KERNBASE));
frame->lr, frame->lr - (register_t)(__startkernel - KERNBASE));
ASSYM(KERNBASE, KERNBASE);
PRELOAD_PUSH_VALUE(uint64_t, KERNBASE);
PRELOAD_PUSH_VALUE(uint64_t, (size_t)((vm_offset_t)&end - KERNBASE));
rvbp->dtbp_phys >= rvbp->kern_phys + (lastaddr - KERNBASE),
rvbp->kern_phys + (lastaddr - KERNBASE)));
rvbp->kern_phys, rvbp->kern_phys + (lastaddr - KERNBASE));
kernlen = (lastaddr - KERNBASE);
mdhdr.kernbase = KERNBASE;
slot = pmap_l2_index(KERNBASE);
slot = pmap_l2_index(freemempos - kernstart + KERNBASE);
slot = pmap_l1_index(KERNBASE);
slot = pmap_l0_index(KERNBASE);
freeva = freemempos - kernstart + KERNBASE;
KERNBASE,
phdr.p_vaddr = KERNBASE;
if (__predict_false((uintptr_t)addr < KERNBASE)) {
return ((uintptr_t)addr - KERNBASE
start_info = (struct hvm_start_info *)(start_info_paddr + KERNBASE);
physfree = MAX(roundup2((vm_paddr_t)_end - KERNBASE, PAGE_SIZE),
(start_info->modlist_paddr + KERNBASE);
(start_info->modlist_paddr + KERNBASE);
(const char *)(mod[0].cmdline_paddr + KERNBASE) : NULL;
header = (struct xen_header *)(mod[0].paddr + KERNBASE);
header->modulep_offset + KERNBASE);
preload_metadata = (caddr_t)(mod[0].paddr + KERNBASE);
off += mod[0].paddr + KERNBASE - metadata;
(char *)(start_info->cmdline_paddr + KERNBASE),
entries = (struct hvm_memmap_table_entry *)(start_info->memmap_paddr + KERNBASE);