arch/alpha/include/asm/pgtable.h
48
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/alpha/include/asm/processor.h
23
#define TASK_UNMAPPED_BASE (TASK_SIZE / 2)
arch/alpha/kernel/osf_sys.c
1220
unsigned long limit = TASK_SIZE;
arch/alpha/kernel/traps.c
691
if ((unsigned long)va >= TASK_SIZE)
arch/alpha/kernel/traps.c
896
if ((unsigned long)va >= TASK_SIZE)
arch/alpha/mm/fault.c
115
if (address >= TASK_SIZE)
arch/arc/include/asm/elf.h
50
#define ELF_ET_DYN_BASE (2UL * TASK_SIZE / 3)
arch/arc/include/asm/pgtable.h
20
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/arc/include/asm/processor.h
102
#define TASK_UNMAPPED_BASE (TASK_SIZE / 3)
arch/arc/include/asm/processor.h
94
#define USER_KERNEL_GUTTER (VMALLOC_START - TASK_SIZE)
arch/arc/include/asm/processor.h
96
#define STACK_TOP TASK_SIZE
arch/arc/mm/cache.c
791
WARN(kstart < TASK_SIZE, "%s() can't handle user vaddr", __func__);
arch/arc/mm/mmap.c
43
if (len > TASK_SIZE)
arch/arc/mm/mmap.c
50
if (TASK_SIZE - len >= addr &&
arch/arc/mm/mmap.c
57
info.high_limit = TASK_SIZE;
arch/arm/include/asm/memory.h
48
#define TASK_UNMAPPED_BASE ALIGN(TASK_SIZE / 3, SZ_16M)
arch/arm/include/asm/memory.h
66
#if TASK_SIZE > MODULES_VADDR
arch/arm/include/asm/pgtable-2level.h
106
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/arm/include/asm/pgtable-3level.h
233
BUG_ON(addr >= TASK_SIZE);
arch/arm/include/asm/pgtable.h
77
#define USER_PGTABLES_CEILING TASK_SIZE
arch/arm/include/asm/processor.h
21
TASK_SIZE : TASK_SIZE_26)
arch/arm/include/asm/processor.h
22
#define STACK_TOP_MAX TASK_SIZE
arch/arm/include/asm/uaccess.h
119
: "r" (size), "r" (TASK_SIZE)
arch/arm/include/asm/uaccess.h
183
unsigned long __limit = TASK_SIZE - 1; \
arch/arm/include/asm/uaccess.h
236
unsigned long __limit = TASK_SIZE - 1; \
arch/arm/kernel/elf.c
125
*start_stack = TASK_SIZE - SZ_16M;
arch/arm/kernel/hw_breakpoint.c
460
return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE);
arch/arm/kernel/process.c
360
last = TASK_SIZE - (npages << PAGE_SHIFT);
arch/arm/mm/alignment.c
813
if (addr >= TASK_SIZE && user_mode(regs))
arch/arm/mm/fault.c
309
if (addr >= TASK_SIZE)
arch/arm/mm/fault.c
34
return addr >= TASK_SIZE && ULONG_MAX - addr >= size;
arch/arm/mm/fault.c
513
if (addr < TASK_SIZE)
arch/arm/mm/fault.c
592
if (addr >= TASK_SIZE && user_mode(regs))
arch/arm/mm/init.c
239
BUILD_BUG_ON(TASK_SIZE > MODULES_VADDR);
arch/arm/mm/init.c
240
BUG_ON(TASK_SIZE > MODULES_VADDR);
arch/arm/mm/mmap.c
100
if (len > TASK_SIZE)
arch/arm/mm/mmap.c
117
if (TASK_SIZE - len >= addr &&
arch/arm/mm/mmap.c
140
info.high_limit = TASK_SIZE;
arch/arm/mm/mmap.c
57
if (len > TASK_SIZE)
arch/arm/mm/mmap.c
67
if (TASK_SIZE - len >= addr &&
arch/arm/mm/mmap.c
74
info.high_limit = TASK_SIZE;
arch/arm/mm/mmu.c
1796
if (addr < TASK_SIZE && pte_valid_user(pteval)) {
arch/arm/mm/mmu.c
979
if (md->virtual != vectors_base() && md->virtual < TASK_SIZE) {
arch/arm/mm/pgd.c
71
init_pgd = pgd_offset_k(TASK_SIZE);
arch/arm/mm/pgd.c
72
init_p4d = p4d_offset(init_pgd, TASK_SIZE);
arch/arm/mm/pgd.c
73
init_pud = pud_offset(init_p4d, TASK_SIZE);
arch/arm/mm/pgd.c
74
init_pmd = pmd_offset(init_pud, TASK_SIZE);
arch/arm/mm/pgd.c
75
new_pmd = pmd_offset(new_pud, TASK_SIZE);
arch/arm/mm/pgd.c
77
(pmd_index(MODULES_VADDR) - pmd_index(TASK_SIZE))
arch/arm64/include/asm/processor.h
101
base + TASK_SIZE - DEFAULT_MAP_WINDOW :\
arch/arm64/include/asm/processor.h
363
return addr < TASK_SIZE;
arch/arm64/include/asm/processor.h
82
#define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 4))
arch/arm64/include/asm/processor.h
98
(((addr) > DEFAULT_MAP_WINDOW) ? TASK_SIZE : DEFAULT_MAP_WINDOW)
arch/arm64/kernel/hw_breakpoint.c
345
return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE);
arch/csky/abiv1/mmap.c
35
.high_limit = TASK_SIZE,
arch/csky/abiv1/mmap.c
55
if (len > TASK_SIZE)
arch/csky/abiv1/mmap.c
65
if (TASK_SIZE - len >= addr &&
arch/csky/include/asm/processor.h
32
#define STACK_TOP TASK_SIZE
arch/csky/include/asm/processor.h
39
#define TASK_UNMAPPED_BASE (TASK_SIZE / 3)
arch/hexagon/include/asm/mem-layout.h
44
#define STACK_TOP TASK_SIZE
arch/hexagon/include/asm/mem-layout.h
45
#define STACK_TOP_MAX TASK_SIZE
arch/hexagon/include/asm/processor.h
54
#define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE/3))
arch/loongarch/include/asm/elf.h
339
#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
arch/loongarch/include/asm/pgtable.h
62
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/loongarch/include/asm/processor.h
23
#define TASK_SIZE_MIN TASK_SIZE
arch/loongarch/include/asm/processor.h
24
#define STACK_TOP_MAX TASK_SIZE
arch/loongarch/include/asm/processor.h
55
#define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
arch/loongarch/kernel/hw_breakpoint.c
306
return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE);
arch/loongarch/kernel/process.c
320
unsigned long top = TASK_SIZE & PAGE_MASK;
arch/loongarch/kernel/ptrace.c
1076
info->mask = TASK_SIZE - 1;
arch/loongarch/mm/mmap.c
31
if (unlikely(len > TASK_SIZE))
arch/loongarch/mm/mmap.c
36
if (TASK_SIZE - len < addr)
arch/loongarch/mm/mmap.c
61
if (TASK_SIZE - len >= addr &&
arch/loongarch/mm/mmap.c
91
info.high_limit = TASK_SIZE;
arch/m68k/include/asm/pgtable_mm.h
72
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/m68k/include/asm/processor.h
58
#define STACK_TOP TASK_SIZE
arch/m68k/include/asm/tlbflush.h
119
for (addr = 0x00000000; addr < TASK_SIZE; addr += SUN3_PMEG_SIZE) {
arch/m68k/include/asm/tlbflush.h
148
for (i = 0; i < TASK_SIZE; i += SUN3_PMEG_SIZE) {
arch/m68k/mm/mcfmmu.c
131
if ((mmuar < TASK_UNMAPPED_BASE) || (mmuar >= TASK_SIZE))
arch/microblaze/include/asm/pgtable.h
101
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/microblaze/include/asm/processor.h
44
# define TASK_UNMAPPED_BASE (TASK_SIZE / 8 * 3)
arch/microblaze/include/asm/processor.h
84
# define STACK_TOP TASK_SIZE
arch/microblaze/mm/fault.c
100
if (unlikely(kernel_mode(regs) && (address >= TASK_SIZE))) {
arch/microblaze/mm/pgtable.c
233
if (addr < TASK_SIZE)
arch/mips/include/asm/elf.h
458
#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
arch/mips/include/asm/processor.h
41
#define STACK_TOP_MAX TASK_SIZE
arch/mips/include/asm/processor.h
80
#define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
arch/mips/kernel/process.c
685
unsigned long top = TASK_SIZE & PAGE_MASK;
arch/mips/mm/mmap.c
39
if (unlikely(len > TASK_SIZE))
arch/mips/mm/mmap.c
44
if (TASK_SIZE - len < addr)
arch/mips/mm/mmap.c
69
if (TASK_SIZE - len >= addr &&
arch/mips/mm/mmap.c
96
info.high_limit = TASK_SIZE;
arch/nios2/include/asm/processor.h
31
#define STACK_TOP TASK_SIZE
arch/nios2/include/asm/processor.h
42
# define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3))
arch/nios2/mm/fault.c
74
if (unlikely(address >= TASK_SIZE))
arch/openrisc/include/asm/pgtable.h
75
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/openrisc/include/asm/processor.h
22
#define STACK_TOP TASK_SIZE
arch/openrisc/include/asm/processor.h
40
#define TASK_UNMAPPED_BASE (TASK_SIZE / 8 * 3)
arch/openrisc/kernel/asm-offsets.c
59
DEFINE(NUM_USER_SEGMENTS, TASK_SIZE >> 28);
arch/parisc/include/asm/processor.h
45
#define STACK_TOP TASK_SIZE
arch/parisc/kernel/sys_parisc.c
111
if (unlikely(len > TASK_SIZE))
arch/parisc/kernel/sys_parisc.c
121
if (TASK_SIZE - len < addr)
arch/parisc/kernel/sys_parisc.c
138
if (TASK_SIZE - len >= addr &&
arch/powerpc/include/asm/book3s/32/mmu-hash.h
203
if (n << 28 < ALIGN(TASK_SIZE, SZ_256M))
arch/powerpc/include/asm/book3s/32/pgtable.h
133
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/powerpc/include/asm/kexec.h
35
#define KEXEC_CONTROL_MEMORY_LIMIT TASK_SIZE
arch/powerpc/include/asm/nohash/32/pgtable.h
52
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/powerpc/include/asm/page.h
261
#define is_kernel_addr(x) ((x) >= TASK_SIZE)
arch/powerpc/include/asm/task_size_32.h
41
#define TASK_UNMAPPED_BASE (TASK_SIZE / 8 * 3)
arch/powerpc/include/asm/task_size_32.h
43
#define DEFAULT_MAP_WINDOW TASK_SIZE
arch/powerpc/include/asm/task_size_32.h
44
#define STACK_TOP TASK_SIZE
arch/powerpc/include/asm/task_size_64.h
67
#define DEFAULT_MAP_WINDOW TASK_SIZE
arch/powerpc/include/asm/task_size_64.h
76
(((addr) > DEFAULT_MAP_WINDOW) ? (base) + TASK_SIZE - DEFAULT_MAP_WINDOW : (base))
arch/powerpc/include/asm/task_size_64.h
80
(((flags) & MAP_FIXED) && ((addr) + (len) > DEFAULT_MAP_WINDOW)) ? TASK_SIZE : \
arch/powerpc/include/asm/uaccess.h
482
asm("cmplw %1, %2; iselgt %0, %2, %1" : "=r"(addr) : "r"(ptr), "r"(TASK_SIZE) : "cr0");
arch/powerpc/include/asm/uaccess.h
491
unsigned long mask = (unsigned long)((long)((TASK_SIZE >> 17) - 1 - (addr >> 17)) >> 31);
arch/powerpc/include/asm/uaccess.h
493
addr = (addr & ~mask) | (TASK_SIZE & mask);
arch/powerpc/include/asm/uaccess.h
502
return (void __user *)(likely(addr < TASK_SIZE) ? addr : TASK_SIZE);
arch/powerpc/include/asm/uaccess.h
517
if (TASK_SIZE <= UL(SZ_2G) && border >= UL(SZ_2G))
arch/powerpc/kernel/asm-offsets.c
333
DEFINE(TASK_SIZE, TASK_SIZE);
arch/powerpc/kernel/asm-offsets.c
334
DEFINE(NUM_USER_SEGMENTS, ALIGN(TASK_SIZE, SZ_256M) >> 28);
arch/powerpc/kernel/ptrace/ptrace-adv.c
101
if ((data & ~0x7UL) >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
154
if (bp_info->addr >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
159
if (bp_info->addr2 >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
296
if (bp_info->addr >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
401
if (bp_info->addr >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
408
if (~((unsigned long)bp_info->addr2) >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-adv.c
414
if (bp_info->addr2 >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-noadv.c
214
if ((unsigned long)bp_info->addr >= TASK_SIZE)
arch/powerpc/kernel/ptrace/ptrace-noadv.c
97
if ((data & ~0x7UL) >= TASK_SIZE)
arch/powerpc/mm/book3s32/mmu.c
226
for (i = ALIGN(TASK_SIZE, SZ_256M) >> 28; i < 16; i++) {
arch/powerpc/mm/book3s32/mmu.c
329
if (!pte_young(*ptep) || address >= TASK_SIZE)
arch/powerpc/mm/book3s32/tlb.c
102
mm = (vmaddr < TASK_SIZE)? vma->vm_mm: &init_mm;
arch/powerpc/mm/book3s64/hash_utils.c
2132
if (!pte_young(*ptep) || address >= TASK_SIZE)
arch/powerpc/mm/book3s64/slice.c
442
high_limit = TASK_SIZE;
arch/powerpc/mm/fault.c
207
address >= TASK_SIZE ? "exec-protected" : "user",
arch/powerpc/mm/fault.c
216
if (address >= TASK_SIZE)
arch/powerpc/mm/fault.c
320
if (is_user && address >= TASK_SIZE) {
arch/powerpc/mm/ptdump/ptdump.c
265
address_markers[i++].start_address = TASK_SIZE;
arch/powerpc/mm/ptdump/segment_regs.c
34
for (i = 0; i < ALIGN(TASK_SIZE, SZ_256M) >> 28; i++)
arch/powerpc/net/bpf_jit_comp32.c
1005
PPC_LI32(_R0, TASK_SIZE - off);
arch/powerpc/perf/callchain.h
29
if (addr > TASK_SIZE - size || (addr & (size - 1)))
arch/riscv/include/asm/pgtable.h
1226
#define KERN_VIRT_START (-(BIT(VA_BITS)) + TASK_SIZE)
arch/riscv/include/asm/pgtable.h
1268
#define VMALLOC_END TASK_SIZE
arch/riscv/include/asm/pgtable.h
169
#define USER_PTRS_PER_PGD (TASK_SIZE / PGDIR_SIZE)
arch/riscv/include/asm/processor.h
35
#define DEFAULT_MAP_WINDOW TASK_SIZE
arch/riscv/include/asm/processor.h
36
#define STACK_TOP_MAX TASK_SIZE
arch/riscv/include/asm/processor.h
55
#define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
arch/riscv/mm/fault.c
334
if (!user_mode(regs) && addr < TASK_SIZE && unlikely(!(regs->status & SR_SUM))) {
arch/s390/include/asm/pgalloc.h
43
addr + len <= TASK_SIZE) {
arch/s390/kvm/kvm-s390.c
3272
.memory_size = ALIGN_DOWN(TASK_SIZE, _SEGMENT_SIZE),
arch/s390/mm/mmap.c
102
info.high_limit = TASK_SIZE;
arch/s390/mm/mmap.c
123
if (len > TASK_SIZE - mmap_min_addr)
arch/s390/mm/mmap.c
133
if (TASK_SIZE - len >= addr && addr >= mmap_min_addr &&
arch/s390/mm/mmap.c
157
info.high_limit = TASK_SIZE;
arch/s390/mm/mmap.c
86
if (len > TASK_SIZE - mmap_min_addr)
arch/s390/mm/mmap.c
95
if (TASK_SIZE - len >= addr && addr >= mmap_min_addr &&
arch/sh/include/asm/elf.h
114
#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3)
arch/sh/include/asm/kexec.h
24
#define KEXEC_CONTROL_MEMORY_LIMIT TASK_SIZE
arch/sh/include/asm/pgtable-2level.h
22
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/sh/include/asm/processor_32.h
30
#define STACK_TOP TASK_SIZE
arch/sh/include/asm/processor_32.h
36
#define TASK_UNMAPPED_BASE PAGE_ALIGN(TASK_SIZE / 3)
arch/sh/kernel/hw_breakpoint.c
132
return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE);
arch/sh/mm/fault.c
381
return address >= TASK_SIZE;
arch/sh/mm/mmap.c
121
if (unlikely(len > TASK_SIZE))
arch/sh/mm/mmap.c
136
if (TASK_SIZE - len >= addr &&
arch/sh/mm/mmap.c
159
info.high_limit = TASK_SIZE;
arch/sh/mm/mmap.c
73
if (unlikely(len > TASK_SIZE))
arch/sh/mm/mmap.c
87
if (TASK_SIZE - len >= addr &&
arch/sh/mm/mmap.c
94
info.high_limit = TASK_SIZE;
arch/sh/mm/tlbex_32.c
41
if (unlikely(address >= TASK_SIZE || !current->mm))
arch/sparc/kernel/perf_event.c
1793
return (__range_not_ok(fp, size, TASK_SIZE) == 0);
arch/sparc/kernel/sys_sparc_32.c
101
if (len > TASK_SIZE - PAGE_SIZE || addr + len > TASK_SIZE - PAGE_SIZE)
arch/sparc/kernel/sys_sparc_32.c
62
if (len > TASK_SIZE - PAGE_SIZE)
arch/sparc/kernel/sys_sparc_32.c
69
info.high_limit = TASK_SIZE;
arch/sparc/kernel/sys_sparc_64.c
105
unsigned long task_size = TASK_SIZE;
arch/sparc/mm/fault_32.c
136
if (address >= TASK_SIZE)
arch/um/include/asm/pgtable-2level.h
24
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
arch/um/include/asm/pgtable-4level.h
43
#define USER_PTRS_PER_PGD ((TASK_SIZE + (PGDIR_SIZE - 1)) / PGDIR_SIZE)
arch/um/include/asm/pgtable.h
54
#define VMALLOC_END (TASK_SIZE-2*PAGE_SIZE)
arch/um/include/asm/processor-generic.h
61
#define STACK_TOP (TASK_SIZE - 2 * PAGE_SIZE)
arch/um/include/asm/uaccess.h
15
(((unsigned long) (addr) < TASK_SIZE) && \
arch/um/include/asm/uaccess.h
16
(((unsigned long) (addr) + (size)) < TASK_SIZE))
arch/um/kernel/trap.c
349
else if (!is_user && address > PAGE_SIZE && address < TASK_SIZE) {
arch/um/kernel/um_arch.c
372
max_physmem = TASK_SIZE - uml_physmem - MIN_VMALLOC;
arch/x86/include/asm/kexec.h
52
# define KEXEC_CONTROL_MEMORY_LIMIT TASK_SIZE
arch/x86/include/asm/page_32_types.h
49
#define TASK_SIZE_LOW TASK_SIZE
arch/x86/include/asm/page_32_types.h
50
#define TASK_SIZE_MAX TASK_SIZE
arch/x86/include/asm/page_32_types.h
51
#define DEFAULT_MAP_WINDOW TASK_SIZE
arch/x86/include/asm/page_32_types.h
52
#define STACK_TOP TASK_SIZE
arch/x86/kernel/sys_x86_64.c
177
if (len > TASK_SIZE)
arch/x86/kernel/uprobes.c
1844
vaddr = get_unmapped_area(NULL, TASK_SIZE - PAGE_SIZE, PAGE_SIZE, 0, 0);
arch/x86/kernel/uprobes.c
663
info.high_limit = min(high_limit, TASK_SIZE);
arch/x86/kernel/uprobes.c
719
if (vaddr > TASK_SIZE || vaddr < PAGE_SIZE)
arch/x86/mm/init.c
840
(TASK_SIZE - TASK_UNMAPPED_BASE - 3 * PAGE_SIZE);
arch/x86/mm/mmap.c
199
if (TASK_SIZE - len < addr)
arch/x86/um/asm/elf.h
176
#define ELF_ET_DYN_BASE (TASK_SIZE / 3 * 2)
arch/xtensa/include/asm/asm-uaccess.h
50
movi \at, __XTENSA_UL_CONST(TASK_SIZE)
arch/xtensa/include/asm/elf.h
125
#define ELF_ET_DYN_BASE (2 * TASK_SIZE / 3)
arch/xtensa/include/asm/pgtable.h
60
#define USER_PTRS_PER_PGD (TASK_SIZE/PGDIR_SIZE)
arch/xtensa/include/asm/processor.h
168
#define TASK_UNMAPPED_BASE (TASK_SIZE / 2)
arch/xtensa/include/asm/processor.h
39
#define STACK_TOP TASK_SIZE
arch/xtensa/kernel/hw_breakpoint.c
45
return (va >= TASK_SIZE) && ((va + len - 1) >= TASK_SIZE);
arch/xtensa/kernel/signal.c
215
&& ((regs->lbeg > TASK_SIZE) || (regs->lend > TASK_SIZE)) )
arch/xtensa/kernel/stacktrace.c
111
if (pc == 0 || pc >= TASK_SIZE || ufn(&frame, data))
arch/xtensa/kernel/stacktrace.c
46
if (pc == 0 || pc >= TASK_SIZE || ufn(&frame, data))
arch/xtensa/kernel/stacktrace.c
84
if (pc == 0 || pc >= TASK_SIZE || ufn(&frame, data))
arch/xtensa/kernel/syscall.c
74
if (len > TASK_SIZE)
arch/xtensa/kernel/syscall.c
95
if (TASK_SIZE - len < addr)
arch/xtensa/mm/fault.c
105
if (address >= TASK_SIZE && !user_mode(regs)) {
arch/xtensa/mm/tlb.c
153
if (end > start && start >= TASK_SIZE && end <= PAGE_OFFSET &&
arch/xtensa/mm/tlb.c
246
if (tlb_asid > 0 && ((vpn < TASK_SIZE) == kernel)) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
919
info->maxUserVa = TASK_SIZE;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
169
info->maxUserVa = TASK_SIZE;
drivers/infiniband/hw/mlx5/odp.c
1924
u32 log_va_pages = ilog2(TASK_SIZE) - PAGE_SHIFT;
drivers/infiniband/hw/mlx5/odp.c
1939
mlx5_imr_ksm_entries = BIT_ULL(get_order(TASK_SIZE) -
drivers/misc/lkdtm/perms.c
218
if (user_addr >= TASK_SIZE) {
drivers/misc/lkdtm/perms.c
239
if (user_addr >= TASK_SIZE) {
drivers/misc/lkdtm/usercopy.c
152
if (user_addr >= TASK_SIZE) {
drivers/misc/lkdtm/usercopy.c
228
if (user_alloc >= TASK_SIZE) {
drivers/misc/lkdtm/usercopy.c
318
if (user_addr >= TASK_SIZE) {
drivers/misc/lkdtm/usercopy.c
357
if (uaddr >= TASK_SIZE) {
drivers/misc/lkdtm/usercopy.c
84
if (user_addr >= TASK_SIZE) {
fs/binfmt_elf.c
105
#define BAD_ADDR(x) (unlikely((unsigned long)(x) >= TASK_SIZE))
fs/binfmt_elf.c
1228
elf_ppnt->p_memsz > TASK_SIZE ||
fs/binfmt_elf.c
1229
TASK_SIZE - elf_ppnt->p_memsz < k) {
fs/binfmt_elf.c
711
eppnt->p_memsz > TASK_SIZE ||
fs/binfmt_elf.c
712
TASK_SIZE - eppnt->p_memsz < k) {
fs/exec.c
1328
me->mm->task_size = TASK_SIZE;
include/asm-generic/access_ok.h
16
#define TASK_SIZE_MAX TASK_SIZE
include/asm-generic/tlb.h
409
tlb->start = TASK_SIZE;
include/linux/rseq_entry.h
271
u64 start_ip, abort_ip, offset, cs_end, head, tasksize = TASK_SIZE;
include/linux/rseq_entry.h
396
unsigned long tasksize = TASK_SIZE;
include/linux/sched.h
2305
#define TASK_SIZE_OF(tsk) TASK_SIZE
include/linux/sched/mm.h
173
#define arch_get_mmap_end(addr, len, flags) (TASK_SIZE)
kernel/bpf/core.c
3182
return TASK_SIZE;
kernel/bpf/helpers.c
754
if ((unsigned long)unsafe_ptr < TASK_SIZE)
kernel/events/core.c
7782
if (!addr || addr >= TASK_SIZE)
kernel/events/core.c
7785
return TASK_SIZE - addr;
kernel/events/core.c
8380
if (virt >= TASK_SIZE) {
kernel/events/uprobes.c
1700
return get_unmapped_area(NULL, TASK_SIZE - PAGE_SIZE, PAGE_SIZE, 0, 0);
kernel/rseq.c
369
if (unlikely(csaddr >= TASK_SIZE))
kernel/sys.c
2000
unsigned long mmap_max_addr = TASK_SIZE;
kernel/sys.c
2229
if (addr >= TASK_SIZE || addr < mmap_min_addr)
kernel/trace/bpf_trace.c
289
if ((unsigned long)unsafe_ptr < TASK_SIZE) {
kernel/trace/bpf_trace.c
308
if ((unsigned long)unsafe_ptr < TASK_SIZE) {
kernel/trace/trace_events_synth.c
452
if ((unsigned long)str_val < TASK_SIZE)
kernel/trace/trace_probe_kernel.h
113
if ((unsigned long)src < TASK_SIZE)
kernel/trace/trace_probe_kernel.h
28
if (addr < TASK_SIZE)
kernel/trace/trace_probe_kernel.h
82
if ((unsigned long)addr < TASK_SIZE)
lib/kunit/user_alloc.c
41
mm->task_size = TASK_SIZE;
lib/tests/usercopy_kunit.c
312
KUNIT_ASSERT_LT_MSG(test, user_addr, (unsigned long)TASK_SIZE,
mm/debug_vm_pgtable.c
966
total_user_pages = (TASK_SIZE - FIRST_USER_ADDRESS) / PAGE_SIZE;
mm/kasan/kasan_test_c.c
2165
KUNIT_ASSERT_LT_MSG(test, useraddr, (unsigned long)TASK_SIZE,
mm/kasan/report.c
666
else if (orig_addr < TASK_SIZE)
mm/kasan/report_generic.c
134
else if ((unsigned long)info->access_addr < TASK_SIZE)
mm/kmsan/hooks.c
269
(u64)to < TASK_SIZE) {
mm/kmsan/instrumentation.c
25
(u64)addr < TASK_SIZE)
mm/memory-failure.c
836
ret = walk_page_range(p->mm, 0, TASK_SIZE, &hwpoison_walk_ops,
mm/mlock.c
769
mm_populate(0, TASK_SIZE);
mm/mmap.c
824
if (len > TASK_SIZE)
mm/mmap.c
857
if (addr > TASK_SIZE - len)
mm/mremap.c
1779
if (vrm->new_len > TASK_SIZE)
mm/mremap.c
1787
if (vrm->new_addr > TASK_SIZE - vrm->new_len)
mm/nommu.c
714
if (!rlen || rlen > TASK_SIZE)
mm/pagewalk.c
76
if (walk->mm == &init_mm || addr >= TASK_SIZE)
mm/pagewalk.c
82
if (walk->mm != &init_mm && addr < TASK_SIZE)
mm/shmem.c
2790
if (len > TASK_SIZE)
mm/shmem.c
2801
if (addr > TASK_SIZE - len)
mm/shmem.c
2866
if (inflated_len > TASK_SIZE)
mm/shmem.c
2882
if (inflated_addr > TASK_SIZE - len)
mm/vma.c
1618
if ((offset_in_page(start)) || start > TASK_SIZE || len > TASK_SIZE-start)
mm/vma.c
3105
if (address >= (TASK_SIZE & PAGE_MASK))
mm/vma.c
3113
if (gap_addr < address || gap_addr > TASK_SIZE)
mm/vma.c
3114
gap_addr = TASK_SIZE;