Symbol: KERNSTART
sys/amd64/amd64/machdep.c
1286
pml4e = (pml4_entry_t *)cr3 + pmap_pml4e_index(KERNSTART);
sys/amd64/amd64/machdep.c
1287
pdpe = (pdp_entry_t *)(*pml4e & PG_FRAME) + pmap_pdpe_index(KERNSTART);
sys/amd64/amd64/machdep.c
1288
pde = (pd_entry_t *)(*pdpe & PG_FRAME) + pmap_pde_index(KERNSTART);
sys/amd64/amd64/machdep.c
1320
physfree += ucode_load_bsp(physfree - kernphys + KERNSTART);
sys/amd64/amd64/machdep.c
1368
thread0.td_kstack = physfree - kernphys + KERNSTART;
sys/amd64/amd64/machdep.c
1405
dpcpu_init((void *)(physfree - kernphys + KERNSTART), 0);
sys/amd64/amd64/pmap.c
11173
pmap_pti_add_kva_locked((vm_offset_t)KERNSTART, (vm_offset_t)etext,
sys/amd64/amd64/pmap.c
11716
pa = pabase + ((vm_offset_t)&data[offset] - KERNSTART);
sys/amd64/amd64/pmap.c
1706
if (pa < trunc_2mpage(kernphys + btext - KERNSTART) ||
sys/amd64/amd64/pmap.c
1707
pa >= trunc_2mpage(kernphys + _end - KERNSTART))
sys/amd64/amd64/pmap.c
1716
if (pa >= trunc_2mpage(kernphys + brwsection - KERNSTART))
sys/amd64/amd64/pmap.c
1728
if (pa < round_2mpage(kernphys + etext - KERNSTART))
sys/amd64/amd64/pmap.c
1758
KERNSTART + amd64_loadaddr());
sys/amd64/amd64/pmap.c
1835
nkdmpde = howmany((vm_offset_t)brwsection - KERNSTART +
sys/amd64/amd64/pmap.c
2102
virtual_avail = (vm_offset_t)KERNSTART + round_2mpage(KERNend -
sys/amd64/include/pmap.h
384
m->phys_addr >= kernphys + (vm_offset_t)&_end - KERNSTART, \
sys/x86/x86/sdt_machdep.c
26
if (patchpoint < KERNSTART || target < KERNSTART)
sys/x86/x86/sdt_machdep.c
74
KASSERT(patchpoint >= KERNSTART,