__start
#define SET_BANK(__nr,__start,__size) \
mi->bank[__nr].start = (__start), \
#define SET_BANK(__nr,__start,__size) \
mi->bank[__nr].start = (__start), \
#define for_each_mapping_in_range_safe(__pgt, __start, __end, __map) \
__start, __end - 1); \
__tmp = pkvm_mapping_iter_next(__map, __start, __end - 1); \
unsigned long __start;
v->__start = start;
write_tc_c0_tcrestart((unsigned long)v->__start);
v->__start = sym[i].st_value;
if ((v->__start == 0) || (v->shared_ptr == NULL))
if (v->__start == 0) {
v->__start = 0;
unsigned long __start = mftb(); \
(tb_ticks_since(__start) <= __loops)) \
(tb_ticks_since(__start) <= __loops)) \
extern void __start(unsigned long r3, unsigned long r4, unsigned long r5,
__start(hdr, kbase, 0, 0, 0, 0, 0);
void *__start = (void *)VDSO##bits##_SYMBOL(&vdso##bits##_start, sec##_start); \
do_##type##_fixups((value), __start, __end); \
extern void __start(unsigned long r3, unsigned long r4, unsigned long r5);
__start(hdr, KERNELBASE + offset, 0);
unsigned long mask, idx, tmp, sz = (size), __start = (start); \
if (unlikely(__start >= sz)) \
mask = MUNGE(BITMAP_FIRST_WORD_MASK(__start)); \
idx = __start / BITS_PER_LONG; \
#define __DPTUN_REG_RANGE(__start, __size) \
GENMASK_ULL((__start) + (__size) - 1, (__start))
#define queue_loop(__entry, __start, __end) \
for ((__entry) = (__start); \
#define bfa_msix_get_rme_range(__bfa, __start, __end) \
((__bfa)->iocfc.hwif.hw_msix_get_rme_range(__bfa, __start, __end))
loff_t __start; \
__start = (loff_t)(sec) << HFS_SECTOR_SIZE_BITS;\
__block = __start >> (sb)->s_blocksize_bits; \
__offset = __start & ((sb)->s_blocksize - 1); \
void *__start;
__start = kc_xlate_dev_mem_ptr(phys);
if (!__start) {
__start = (void *)start;
ret = copy_from_kernel_nofault(buf, __start, tsz);
kc_unxlate_dev_mem_ptr(phys, __start);
BOUNDED_SECTION_PRE_LABEL(_sec_, _label_, __start, __stop)
HEADERED_SECTION_PRE_LABEL(_sec_, _label_, __start, __stop)
BOUNDED_SECTION_PRE_LABEL(.lsm_info.init, _lsm_info, __start, __end)
BOUNDED_SECTION_PRE_LABEL(.early_lsm_info.init, _early_lsm_info, __start, __end)
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_early, _pci_fixups_early, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_header, _pci_fixups_header, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_final, _pci_fixups_final, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_enable, _pci_fixups_enable, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_resume, _pci_fixups_resume, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_suspend, _pci_fixups_suspend, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_resume_early, _pci_fixups_resume_early, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.pci_fixup_suspend_late, _pci_fixups_suspend_late, __start, __end) \
BOUNDED_SECTION_PRE_LABEL(.builtin_fw, _builtin_fw, __start, __end) \
for (int __start = (node), \
(node) = nearest_node_nodemask((__start), &(unvisited)); \
(node) = nearest_node_nodemask((__start), &(unvisited)))
({ struct nfattr *__start = (struct nfattr *)skb_tail_pointer(skb); \
__start; })
return __start(m, pos, TRACE_PIDS);
return __start(m, pos, TRACE_NO_PIDS);
unsigned long mask, idx, tmp, sz = (size), __start = (start); \
if (unlikely(__start >= sz)) \
mask = MUNGE(BITMAP_FIRST_WORD_MASK(__start)); \
idx = __start / BITS_PER_LONG; \
void __start(void);
unsigned long mask, idx, tmp, sz = (size), __start = (start); \
if (unlikely(__start >= sz)) \
mask = MUNGE(BITMAP_FIRST_WORD_MASK(__start)); \
idx = __start / BITS_PER_LONG; \