_stext
extern char _stext[], _etext[];
return addr >= (unsigned long)_stext && addr <= (unsigned long)_end;
#define KERNEL_START _stext
.end = (unsigned long)_stext,
.start = (unsigned long)_stext,
PI_EXPORT_SYM(_stext);
unmap_segment(init_pg_dir, va_offset, _stext, _etext,
map_segment(init_pg_dir, NULL, va_offset, _stext, _etext,
map_segment(init_pg_dir, &pgdp, va_offset, _text, _stext, data_prot,
map_segment(init_pg_dir, &pgdp, va_offset, _stext, _etext, prot,
map_range(&ptep, (u64)_stext, (u64)__initdata_begin,
(phys_addr_t)_stext, text_prot, IDMAP_ROOT_LEVEL,
(unsigned long)_stext - (unsigned long)_text,
unsigned long kstart = (unsigned long)lm_alias(_stext);
#define LOOKUP_START_IP (unsigned long)_stext
_stext, _etext, _sdata, _edata, __bss_start, __bss_stop);
setup_initial_init_mm(_stext, _etext, _edata, NULL);
begin = (unsigned long) &_stext;
begin = (unsigned long) &_stext;
source_start = (unsigned long) _stext;
setup_initial_init_mm(_stext, _etext, _edata, _end);
memblock_reserve(__pa_symbol(_stext), _end - _stext);
if (((addr >= (unsigned long) _stext) &&
setup_initial_init_mm(_stext, _etext, _edata, _end);
memblock_reserve(__pa(_stext), _end - _stext);
s = (uintptr_t) dereference_function_descriptor(&_stext);
real_start = __start_interrupts - _stext;
real_end = __end_interrupts - _stext;
(unsigned long)_stext < end;
(addr >= (unsigned long)_stext &&
start = (void*)_stext;
new_paca->kernelbase = (unsigned long) _stext;
int_vector_size = __end_interrupts - _stext;
if (IS_ENABLED(CONFIG_PPC64) && (unsigned long)_stext & 0xffff)
(unsigned long)_stext);
setup_initial_init_mm(_stext, _etext, _edata, _end);
unsigned long kbase = (unsigned long)_stext;
kernel_start = __pa(_stext);
kernel_size = _end - _stext;
end = (void *)src + (__end_interrupts - _stext);
unsigned long base = (unsigned long)_stext - PAGE_OFFSET;
start = (unsigned long)_stext;
(unsigned long)_stext % 0x1000000) {
start = (unsigned long)_stext;
for (start = PAGE_OFFSET; start < (unsigned long)_stext; start += PAGE_SIZE) {
stext_phys = __pa_symbol(_stext);
boundary = (unsigned long)(_sinittext - _stext);
if (regions_overlap(start, end, __pa(_stext), __pa(_end)))
kernel_sz = (unsigned long)_end - (unsigned long)_stext;
memcpy((void *)kernstart_virt_addr, (void *)_stext, kernel_sz);
if (v_block_mapped((unsigned long)_stext + 1))
PFN_DOWN((unsigned long)_stext);
return set_memory_ro((unsigned long)_stext, numpages);
for (ptr = ((unsigned long) &_stext) & PAGE_MASK;
setup_initial_init_mm(_stext, _etext, _edata, _end);
os_info_entry_add_val(OS_INFO_IMAGE_START, (unsigned long)_stext);
os_info_entry_add_val(OS_INFO_IMAGE_PHYS, __pa_symbol(_stext));
memblock_reserve(__pa(_stext), _end - _stext);
rc |= add_marker((unsigned long)_stext, (unsigned long)_end, "Kernel Image");
__set_memory_rox(_stext, _etext);
(unsigned long)(__end_rodata - _stext) >> 10);
mp_ops->start_cpu(cpu, (unsigned long)_stext);
#define LOOKUP_START_IP (unsigned long)_stext
return _stext + tpl->rel_addr;
tpl->rel_addr = addr - (void *)_stext;
if ((address >= (u64)_stext && address <= (u64)_etext) ||
kasan_populate_shadow((unsigned long)kasan_mem_to_shadow(_stext),
early_pfn_to_nid(__pa(_stext)));
mem_reserve(__pa(_stext), __pa(_end));
return addr >= (unsigned long)_stext &&
return ((addr >= (unsigned long)_stext &&
extern char _text[], _stext[], _etext[];
text_size = _etext - _stext; \
__entry->caller_offs = (s32)(ip - (unsigned long)_stext);
__entry->parent_offs = (s32)(parent_ip - (unsigned long)_stext);
(void *)((unsigned long)(_stext) + __entry->caller_offs),
(void *)((unsigned long)(_stext) + __entry->parent_offs))
if (memory_intersects(_stext, _etext, addr, len) ||
pc = ((unsigned long)__pc - (unsigned long)_stext) >> prof_shift;
prof_len = (_etext - _stext) >> prof_shift;
ip = addr + (unsigned long)_stext;
VMCOREINFO_SYMBOL(_stext);
codesize = _etext - _stext;
adj_init_size(_stext, _etext, codesize, _sinittext, init_code_size);
adj_init_size(_stext, _etext, codesize, __start_rodata, rosize);
unsigned long textlow = (unsigned long)_stext;