Symbol: vdso_base
arch/arm64/kernel/vdso.c
104
unsigned long vdso_base, vdso_text_len, vdso_mapping_len;
arch/arm64/kernel/vdso.c
114
vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0);
arch/arm64/kernel/vdso.c
115
if (IS_ERR_VALUE(vdso_base)) {
arch/arm64/kernel/vdso.c
116
ret = ERR_PTR(vdso_base);
arch/arm64/kernel/vdso.c
120
ret = vdso_install_vvar_mapping(mm, vdso_base);
arch/arm64/kernel/vdso.c
127
vdso_base += VDSO_NR_PAGES * PAGE_SIZE;
arch/arm64/kernel/vdso.c
128
mm->context.vdso = (void *)vdso_base;
arch/arm64/kernel/vdso.c
129
ret = _install_special_mapping(mm, vdso_base, vdso_text_len,
arch/csky/kernel/vdso.c
45
unsigned long vdso_base, vdso_len;
arch/csky/kernel/vdso.c
54
vdso_base = get_unmapped_area(NULL, 0, vdso_len, 0, 0);
arch/csky/kernel/vdso.c
55
if (IS_ERR_VALUE(vdso_base)) {
arch/csky/kernel/vdso.c
56
ret = vdso_base;
arch/csky/kernel/vdso.c
65
mm->context.vdso = (void *)vdso_base;
arch/csky/kernel/vdso.c
69
_install_special_mapping(mm, vdso_base, vdso_pages << PAGE_SHIFT,
arch/csky/kernel/vdso.c
79
vdso_base += (vdso_pages << PAGE_SHIFT);
arch/hexagon/kernel/vdso.c
53
unsigned long vdso_base;
arch/hexagon/kernel/vdso.c
64
vdso_base = STACK_TOP;
arch/hexagon/kernel/vdso.c
66
vdso_base = get_unmapped_area(NULL, vdso_base, PAGE_SIZE, 0, 0);
arch/hexagon/kernel/vdso.c
67
if (IS_ERR_VALUE(vdso_base)) {
arch/hexagon/kernel/vdso.c
68
ret = vdso_base;
arch/hexagon/kernel/vdso.c
74
vma = _install_special_mapping(mm, vdso_base, PAGE_SIZE,
arch/hexagon/kernel/vdso.c
83
mm->context.vdso = (void *)vdso_base;
arch/loongarch/kernel/vdso.c
97
data_addr = get_unmapped_area(NULL, vdso_base(), size, 0, 0);
arch/mips/kernel/vdso.c
119
base = get_unmapped_area(NULL, vdso_base(), size, 0, 0);
arch/parisc/include/asm/elf.h
359
#define VDSO_CURRENT_BASE current->mm->context.vdso_base
arch/parisc/include/asm/mmu.h
7
unsigned long vdso_base;
arch/parisc/include/asm/vdso.h
12
#define VDSO64_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso64_offset_##name))
arch/parisc/include/asm/vdso.h
13
#define VDSO32_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso32_offset_##name))
arch/parisc/kernel/vdso.c
32
current->mm->context.vdso_base = vma->vm_start;
arch/parisc/kernel/vdso.c
91
current->mm->context.vdso_base = vdso_text_start;
arch/powerpc/kernel/vdso.c
100
unsigned long vdso_size, vdso_base, mappings_size;
arch/powerpc/kernel/vdso.c
121
vdso_base = get_unmapped_area(NULL, 0, mappings_size, 0, 0);
arch/powerpc/kernel/vdso.c
122
if (IS_ERR_VALUE(vdso_base))
arch/powerpc/kernel/vdso.c
123
return vdso_base;
arch/powerpc/kernel/vdso.c
126
vdso_base = ALIGN(vdso_base, VDSO_ALIGNMENT);
arch/powerpc/kernel/vdso.c
128
vma = vdso_install_vvar_mapping(mm, vdso_base);
arch/powerpc/kernel/vdso.c
142
vma = _install_special_mapping(mm, vdso_base + vvar_size, vdso_size,
arch/powerpc/kernel/vdso.c
146
do_munmap(mm, vdso_base, vvar_size, NULL);
arch/powerpc/kernel/vdso.c
151
mm->context.vdso = (void __user *)vdso_base + vvar_size;
arch/riscv/kernel/vdso.c
120
unsigned long vdso_base, vdso_text_len, vdso_mapping_len;
arch/riscv/kernel/vdso.c
129
vdso_base = get_unmapped_area(NULL, 0, vdso_mapping_len, 0, 0);
arch/riscv/kernel/vdso.c
130
if (IS_ERR_VALUE(vdso_base)) {
arch/riscv/kernel/vdso.c
131
ret = ERR_PTR(vdso_base);
arch/riscv/kernel/vdso.c
135
ret = vdso_install_vvar_mapping(mm, vdso_base);
arch/riscv/kernel/vdso.c
139
vdso_base += VVAR_SIZE;
arch/riscv/kernel/vdso.c
140
mm->context.vdso = (void *)vdso_base;
arch/riscv/kernel/vdso.c
143
_install_special_mapping(mm, vdso_base, vdso_text_len,
arch/s390/include/asm/elf.h
237
(unsigned long)current->mm->context.vdso_base); \
arch/s390/include/asm/mmu.h
18
unsigned long vdso_base;
arch/s390/include/asm/vdso-symbols.h
7
#define VDSO_SYMBOL(tsk, name) ((tsk)->mm->context.vdso_base + (vdso_offset_##name))
arch/s390/kernel/stacktrace.c
97
return in_range(ip, current->mm->context.vdso_base, vdso_text_size());
arch/s390/kernel/vdso.c
30
current->mm->context.vdso_base = vma->vm_start;
arch/s390/kernel/vdso.c
76
current->mm->context.vdso_base = vdso_text_start;