Symbol: _stext
arch/alpha/kernel/traps.c
97
extern char _stext[], _etext[];
arch/arc/kernel/setup.c
441
return addr >= (unsigned long)_stext && addr <= (unsigned long)_end;
arch/arm/include/asm/memory.h
137
#define KERNEL_START _stext
arch/arm/mm/init.c
267
.end = (unsigned long)_stext,
arch/arm/mm/init.c
293
.start = (unsigned long)_stext,
arch/arm64/kernel/image-vars.h
65
PI_EXPORT_SYM(_stext);
arch/arm64/kernel/pi/map_kernel.c
118
unmap_segment(init_pg_dir, va_offset, _stext, _etext,
arch/arm64/kernel/pi/map_kernel.c
129
map_segment(init_pg_dir, NULL, va_offset, _stext, _etext,
arch/arm64/kernel/pi/map_kernel.c
85
map_segment(init_pg_dir, &pgdp, va_offset, _text, _stext, data_prot,
arch/arm64/kernel/pi/map_kernel.c
87
map_segment(init_pg_dir, &pgdp, va_offset, _stext, _etext, prot,
arch/arm64/kernel/pi/map_range.c
101
map_range(&ptep, (u64)_stext, (u64)__initdata_begin,
arch/arm64/kernel/pi/map_range.c
102
(phys_addr_t)_stext, text_prot, IDMAP_ROOT_LEVEL,
arch/arm64/mm/mmu.c
1198
(unsigned long)_stext - (unsigned long)_text,
arch/arm64/mm/mmu.c
915
unsigned long kstart = (unsigned long)lm_alias(_stext);
arch/loongarch/include/asm/orc_lookup.h
26
#define LOOKUP_START_IP (unsigned long)_stext
arch/m68k/kernel/setup_no.c
139
_stext, _etext, _sdata, _edata, __bss_start, __bss_stop);
arch/m68k/kernel/setup_no.c
87
setup_initial_init_mm(_stext, _etext, _edata, NULL);
arch/mips/mm/c-r4k.c
1395
begin = (unsigned long) &_stext;
arch/mips/mm/sc-rm7k.c
191
begin = (unsigned long) &_stext;
arch/mips/sgi-ip27/ip27-klnuma.c
77
source_start = (unsigned long) _stext;
arch/nios2/kernel/setup.c
168
setup_initial_init_mm(_stext, _etext, _edata, _end);
arch/nios2/kernel/setup.c
177
memblock_reserve(__pa_symbol(_stext), _end - _stext);
arch/nios2/kernel/traps.c
97
if (((addr >= (unsigned long) _stext) &&
arch/openrisc/kernel/setup.c
239
setup_initial_init_mm(_stext, _etext, _edata, _end);
arch/openrisc/kernel/setup.c
76
memblock_reserve(__pa(_stext), _end - _stext);
arch/parisc/mm/init.c
762
s = (uintptr_t) dereference_function_descriptor(&_stext);
arch/powerpc/include/asm/sections.h
65
real_start = __start_interrupts - _stext;
arch/powerpc/include/asm/sections.h
66
real_end = __end_interrupts - _stext;
arch/powerpc/include/asm/sections.h
75
(unsigned long)_stext < end;
arch/powerpc/kernel/kprobes.c
40
(addr >= (unsigned long)_stext &&
arch/powerpc/kernel/kvm.c
679
start = (void*)_stext;
arch/powerpc/kernel/paca.c
197
new_paca->kernelbase = (unsigned long) _stext;
arch/powerpc/kernel/prom.c
847
int_vector_size = __end_interrupts - _stext;
arch/powerpc/kernel/setup-common.c
1011
if (IS_ENABLED(CONFIG_PPC64) && (unsigned long)_stext & 0xffff)
arch/powerpc/kernel/setup-common.c
1013
(unsigned long)_stext);
arch/powerpc/kernel/setup-common.c
981
setup_initial_init_mm(_stext, _etext, _edata, _end);
arch/powerpc/kernel/traps.c
399
unsigned long kbase = (unsigned long)_stext;
arch/powerpc/kexec/core.c
126
kernel_start = __pa(_stext);
arch/powerpc/kexec/core.c
127
kernel_size = _end - _stext;
arch/powerpc/lib/feature-fixups.c
620
end = (void *)src + (__end_interrupts - _stext);
arch/powerpc/mm/book3s32/mmu.c
200
unsigned long base = (unsigned long)_stext - PAGE_OFFSET;
arch/powerpc/mm/book3s64/hash_pgtable.c
548
start = (unsigned long)_stext;
arch/powerpc/mm/book3s64/hash_utils.c
1089
(unsigned long)_stext % 0x1000000) {
arch/powerpc/mm/book3s64/radix_pgtable.c
235
start = (unsigned long)_stext;
arch/powerpc/mm/book3s64/radix_pgtable.c
240
for (start = PAGE_OFFSET; start < (unsigned long)_stext; start += PAGE_SIZE) {
arch/powerpc/mm/book3s64/radix_pgtable.c
277
stext_phys = __pa_symbol(_stext);
arch/powerpc/mm/nohash/e500.c
172
boundary = (unsigned long)(_sinittext - _stext);
arch/powerpc/mm/nohash/kaslr_booke.c
156
if (regions_overlap(start, end, __pa(_stext), __pa(_end)))
arch/powerpc/mm/nohash/kaslr_booke.c
360
kernel_sz = (unsigned long)_end - (unsigned long)_stext;
arch/powerpc/mm/nohash/kaslr_booke.c
380
memcpy((void *)kernstart_virt_addr, (void *)_stext, kernel_sz);
arch/powerpc/mm/pgtable_32.c
161
if (v_block_mapped((unsigned long)_stext + 1))
arch/powerpc/mm/pgtable_32.c
170
PFN_DOWN((unsigned long)_stext);
arch/powerpc/mm/pgtable_32.c
172
return set_memory_ro((unsigned long)_stext, numpages);
arch/powerpc/platforms/powermac/bootx_init.c
571
for (ptr = ((unsigned long) &_stext) & PAGE_MASK;
arch/riscv/kernel/setup.c
319
setup_initial_init_mm(_stext, _etext, _edata, _end);
arch/s390/kernel/os_info.c
86
os_info_entry_add_val(OS_INFO_IMAGE_START, (unsigned long)_stext);
arch/s390/kernel/os_info.c
88
os_info_entry_add_val(OS_INFO_IMAGE_PHYS, __pa_symbol(_stext));
arch/s390/kernel/setup.c
751
memblock_reserve(__pa(_stext), _end - _stext);
arch/s390/mm/dump_pagetables.c
332
rc |= add_marker((unsigned long)_stext, (unsigned long)_end, "Kernel Image");
arch/s390/mm/vmem.c
667
__set_memory_rox(_stext, _etext);
arch/s390/mm/vmem.c
680
(unsigned long)(__end_rodata - _stext) >> 10);
arch/sh/kernel/smp.c
230
mp_ops->start_cpu(cpu, (unsigned long)_stext);
arch/x86/include/asm/orc_lookup.h
29
#define LOOKUP_START_IP (unsigned long)_stext
arch/x86/kernel/alternative.c
2814
return _stext + tpl->rel_addr;
arch/x86/kernel/alternative.c
3094
tpl->rel_addr = addr - (void *)_stext;
arch/x86/mm/fault.c
436
if ((address >= (u64)_stext && address <= (u64)_etext) ||
arch/x86/mm/kasan_init_64.c
424
kasan_populate_shadow((unsigned long)kasan_mem_to_shadow(_stext),
arch/x86/mm/kasan_init_64.c
426
early_pfn_to_nid(__pa(_stext)));
arch/xtensa/kernel/setup.c
300
mem_reserve(__pa(_stext), __pa(_end));
include/asm-generic/sections.h
213
return addr >= (unsigned long)_stext &&
include/asm-generic/sections.h
229
return ((addr >= (unsigned long)_stext &&
include/asm-generic/sections.h
35
extern char _text[], _stext[], _etext[];
include/asm-generic/vmlinux.lds.h
882
text_size = _etext - _stext; \
include/trace/events/preemptirq.h
26
__entry->caller_offs = (s32)(ip - (unsigned long)_stext);
include/trace/events/preemptirq.h
27
__entry->parent_offs = (s32)(parent_ip - (unsigned long)_stext);
include/trace/events/preemptirq.h
31
(void *)((unsigned long)(_stext) + __entry->caller_offs),
include/trace/events/preemptirq.h
32
(void *)((unsigned long)(_stext) + __entry->parent_offs))
kernel/dma/debug.c
1110
if (memory_intersects(_stext, _etext, addr, len) ||
kernel/profile.c
120
pc = ((unsigned long)__pc - (unsigned long)_stext) >> prof_shift;
kernel/profile.c
91
prof_len = (_etext - _stext) >> prof_shift;
kernel/trace/trace_output.c
1081
ip = addr + (unsigned long)_stext;
kernel/vmcore_info.c
171
VMCOREINFO_SYMBOL(_stext);
mm/mm_init.c
2630
codesize = _etext - _stext;
mm/mm_init.c
2652
adj_init_size(_stext, _etext, codesize, _sinittext, init_code_size);
mm/mm_init.c
2654
adj_init_size(_stext, _etext, codesize, __start_rodata, rosize);
mm/usercopy.c
123
unsigned long textlow = (unsigned long)_stext;