Symbol: VM_EXEC
arch/alpha/include/asm/cacheflush.h
41
if (vma->vm_flags & VM_EXEC) {
arch/alpha/include/asm/tlbflush.h
32
if (vma->vm_flags & VM_EXEC)
arch/alpha/kernel/smp.c
735
if ((vma->vm_flags & VM_EXEC) == 0)
arch/alpha/mm/fault.c
130
if (!(vma->vm_flags & VM_EXEC))
arch/alpha/mm/init.c
253
[VM_EXEC] = _PAGE_P(_PAGE_FOW | _PAGE_FOR),
arch/alpha/mm/init.c
254
[VM_EXEC | VM_READ] = _PAGE_P(_PAGE_FOW),
arch/alpha/mm/init.c
255
[VM_EXEC | VM_WRITE] = _PAGE_P(0),
arch/alpha/mm/init.c
256
[VM_EXEC | VM_WRITE | VM_READ] = _PAGE_P(0),
arch/alpha/mm/init.c
262
[VM_SHARED | VM_EXEC] = _PAGE_S(_PAGE_FOW | _PAGE_FOR),
arch/alpha/mm/init.c
263
[VM_SHARED | VM_EXEC | VM_READ] = _PAGE_S(_PAGE_FOW),
arch/alpha/mm/init.c
264
[VM_SHARED | VM_EXEC | VM_WRITE] = _PAGE_S(0),
arch/alpha/mm/init.c
265
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = _PAGE_S(0)
arch/alpha/mm/tlbflush.c
35
int tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2;
arch/alpha/mm/tlbflush.c
84
.tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2,
arch/arc/include/asm/cacheflush.h
62
if (vma->vm_flags & VM_EXEC) \
arch/arc/mm/fault.c
128
mask = VM_EXEC;
arch/arc/mm/mmap.c
67
[VM_EXEC] = PAGE_U_X_R,
arch/arc/mm/mmap.c
68
[VM_EXEC | VM_READ] = PAGE_U_X_R,
arch/arc/mm/mmap.c
69
[VM_EXEC | VM_WRITE] = PAGE_U_X_R,
arch/arc/mm/mmap.c
70
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_U_X_R,
arch/arc/mm/mmap.c
75
[VM_SHARED | VM_EXEC] = PAGE_U_X_R,
arch/arc/mm/mmap.c
76
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_U_X_R,
arch/arc/mm/mmap.c
77
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_U_X_W_R,
arch/arc/mm/mmap.c
78
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_U_X_W_R
arch/arc/mm/tlb.c
489
if (vma->vm_flags & VM_EXEC) {
arch/arc/mm/tlb.c
501
if (vma->vm_flags & VM_EXEC)
arch/arm/kernel/asm-offsets.c
114
DEFINE(VM_EXEC, VM_EXEC);
arch/arm/kernel/process.c
317
vm_flags_init(&gate_vma, VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYEXEC);
arch/arm/kernel/process.c
422
VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC,
arch/arm/kernel/vdso.c
226
VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC,
arch/arm/mach-rpc/ecard.c
216
struct vm_area_struct vma = TLB_FLUSH_VMA(mm, VM_EXEC);
arch/arm/mm/fault-armv.c
211
else if (vma->vm_flags & VM_EXEC)
arch/arm/mm/fault.c
332
vm_flags = VM_EXEC;
arch/arm/mm/flush.c
110
if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged())
arch/arm/mm/flush.c
165
if (vma->vm_flags & VM_EXEC)
arch/arm/mm/flush.c
94
if (vma->vm_flags & VM_EXEC)
arch/arm/mm/mmu.c
424
[VM_EXEC] = __PAGE_READONLY_EXEC,
arch/arm/mm/mmu.c
425
[VM_EXEC | VM_READ] = __PAGE_READONLY_EXEC,
arch/arm/mm/mmu.c
426
[VM_EXEC | VM_WRITE] = __PAGE_COPY_EXEC,
arch/arm/mm/mmu.c
427
[VM_EXEC | VM_WRITE | VM_READ] = __PAGE_COPY_EXEC,
arch/arm/mm/mmu.c
432
[VM_SHARED | VM_EXEC] = __PAGE_READONLY_EXEC,
arch/arm/mm/mmu.c
433
[VM_SHARED | VM_EXEC | VM_READ] = __PAGE_READONLY_EXEC,
arch/arm/mm/mmu.c
434
[VM_SHARED | VM_EXEC | VM_WRITE] = __PAGE_SHARED_EXEC,
arch/arm/mm/mmu.c
435
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __PAGE_SHARED_EXEC
arch/arm/mm/nommu.c
189
if (vma->vm_flags & VM_EXEC)
arch/arm64/include/asm/mman.h
84
if (vm_flags & VM_EXEC)
arch/arm64/kernel/vdso.c
130
VM_READ|VM_EXEC|gp_flags|
arch/arm64/kernel/vdso.c
257
VM_READ | VM_EXEC |
arch/arm64/kernel/vdso.c
281
VM_READ | VM_EXEC | VM_MAYREAD |
arch/arm64/mm/fault.c
590
vm_flags = VM_EXEC;
arch/arm64/mm/fault.c
611
vm_flags |= VM_EXEC;
arch/arm64/mm/flush.c
35
if (vma->vm_flags & VM_EXEC)
arch/arm64/mm/mmap.c
101
(VM_READ|VM_WRITE|VM_EXEC|VM_SHARED)]);
arch/arm64/mm/mmap.c
22
[VM_EXEC] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
23
[VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
24
[VM_EXEC | VM_WRITE] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
25
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
31
[VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
32
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/arm64/mm/mmap.c
33
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC,
arch/arm64/mm/mmap.c
34
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC
arch/arm64/mm/mmap.c
74
protection_map[VM_EXEC] = PAGE_EXECONLY;
arch/arm64/mm/mmap.c
75
protection_map[VM_EXEC | VM_SHARED] = PAGE_EXECONLY;
arch/csky/abiv1/cacheflush.c
63
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv1/cacheflush.c
73
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv2/cacheflush.c
33
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv2/inc/abi/cacheflush.h
50
if (vma->vm_flags & VM_EXEC) { \
arch/csky/kernel/vdso.c
70
(VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC),
arch/csky/mm/init.c
139
[VM_EXEC] = PAGE_READ,
arch/csky/mm/init.c
140
[VM_EXEC | VM_READ] = PAGE_READ,
arch/csky/mm/init.c
141
[VM_EXEC | VM_WRITE] = PAGE_READ,
arch/csky/mm/init.c
142
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_READ,
arch/csky/mm/init.c
147
[VM_SHARED | VM_EXEC] = PAGE_READ,
arch/csky/mm/init.c
148
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READ,
arch/csky/mm/init.c
149
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_WRITE,
arch/csky/mm/init.c
150
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_WRITE
arch/hexagon/kernel/vdso.c
75
VM_READ|VM_EXEC|
arch/hexagon/mm/cache.c
122
if (vma->vm_flags & VM_EXEC) {
arch/hexagon/mm/init.c
213
[VM_EXEC] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
215
[VM_EXEC | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
218
[VM_EXEC | VM_WRITE] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
220
[VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
232
[VM_SHARED | VM_EXEC] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
234
[VM_SHARED | VM_EXEC | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
237
[VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/init.c
240
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_PAGE_PRESENT | _PAGE_USER |
arch/hexagon/mm/vm_fault.c
70
if (!(vma->vm_flags & VM_EXEC))
arch/loongarch/kernel/vdso.c
111
VM_READ | VM_EXEC |
arch/loongarch/mm/cache.c
174
[VM_EXEC] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
176
[VM_EXEC | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
178
[VM_EXEC | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
180
[VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
194
[VM_SHARED | VM_EXEC] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
196
[VM_SHARED | VM_EXEC | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
198
[VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/cache.c
201
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(_CACHE_CC | _PAGE_VALID |
arch/loongarch/mm/fault.c
235
if (!(vma->vm_flags & VM_EXEC) && address == exception_era(regs)) {
arch/loongarch/mm/fault.c
298
if (!(vma->vm_flags & VM_EXEC) && address == exception_era(regs))
arch/m68k/mm/mcfmmu.c
230
[VM_EXEC] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
233
[VM_EXEC | VM_READ] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
237
[VM_EXEC | VM_WRITE] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
241
[VM_EXEC | VM_WRITE | VM_READ] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
255
[VM_SHARED | VM_EXEC] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
258
[VM_SHARED | VM_EXEC | VM_READ] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
262
[VM_SHARED | VM_EXEC | VM_WRITE] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/mcfmmu.c
266
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = __pgprot(CF_PAGE_VALID |
arch/m68k/mm/motorola.c
411
[VM_EXEC] = PAGE_READONLY_C,
arch/m68k/mm/motorola.c
412
[VM_EXEC | VM_READ] = PAGE_READONLY_C,
arch/m68k/mm/motorola.c
413
[VM_EXEC | VM_WRITE] = PAGE_COPY_C,
arch/m68k/mm/motorola.c
414
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_C,
arch/m68k/mm/motorola.c
419
[VM_SHARED | VM_EXEC] = PAGE_READONLY_C,
arch/m68k/mm/motorola.c
420
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_C,
arch/m68k/mm/motorola.c
421
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_C,
arch/m68k/mm/motorola.c
422
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_C
arch/m68k/mm/sun3mmu.c
100
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
arch/m68k/mm/sun3mmu.c
89
[VM_EXEC] = PAGE_READONLY,
arch/m68k/mm/sun3mmu.c
90
[VM_EXEC | VM_READ] = PAGE_READONLY,
arch/m68k/mm/sun3mmu.c
91
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/m68k/mm/sun3mmu.c
92
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY,
arch/m68k/mm/sun3mmu.c
97
[VM_SHARED | VM_EXEC] = PAGE_READONLY,
arch/m68k/mm/sun3mmu.c
98
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY,
arch/m68k/mm/sun3mmu.c
99
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED,
arch/microblaze/include/asm/cacheflush.h
94
if (vma->vm_flags & VM_EXEC) {
arch/microblaze/mm/fault.c
212
if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC))))
arch/microblaze/mm/init.c
246
[VM_EXEC] = PAGE_READONLY,
arch/microblaze/mm/init.c
247
[VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/microblaze/mm/init.c
248
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/microblaze/mm/init.c
249
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X,
arch/microblaze/mm/init.c
254
[VM_SHARED | VM_EXEC] = PAGE_READONLY,
arch/microblaze/mm/init.c
255
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/microblaze/mm/init.c
256
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED,
arch/microblaze/mm/init.c
257
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
arch/mips/kernel/smp.c
632
int exec = vma->vm_flags & VM_EXEC;
arch/mips/kernel/vdso.c
169
VM_READ | VM_EXEC |
arch/mips/mm/c-octeon.c
142
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/c-octeon.c
157
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/c-r3k.c
240
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
473
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
493
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
543
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/fault.c
118
if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) {
arch/mips/mm/init.c
204
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/tlb-r4k.c
54
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
138
if (vma == NULL || (vma->vm_flags & VM_EXEC))
arch/nios2/mm/cacheflush.c
159
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
236
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
268
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
279
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/fault.c
105
if (!(vma->vm_flags & VM_EXEC))
arch/nios2/mm/init.c
103
VM_READ | VM_EXEC | VM_MAYREAD |
arch/nios2/mm/init.c
121
[VM_EXEC] = MKP(1, 0, 0),
arch/nios2/mm/init.c
122
[VM_EXEC | VM_READ] = MKP(1, 0, 1),
arch/nios2/mm/init.c
123
[VM_EXEC | VM_WRITE] = MKP(1, 0, 0),
arch/nios2/mm/init.c
124
[VM_EXEC | VM_WRITE | VM_READ] = MKP(1, 0, 1),
arch/nios2/mm/init.c
129
[VM_SHARED | VM_EXEC] = MKP(1, 0, 0),
arch/nios2/mm/init.c
130
[VM_SHARED | VM_EXEC | VM_READ] = MKP(1, 0, 1),
arch/nios2/mm/init.c
131
[VM_SHARED | VM_EXEC | VM_WRITE] = MKP(1, 1, 0),
arch/nios2/mm/init.c
132
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = MKP(1, 1, 1)
arch/openrisc/include/asm/cacheflush.h
90
if (vma->vm_flags & VM_EXEC) \
arch/openrisc/mm/cache.c
93
if ((vma->vm_flags & VM_EXEC) && dirty) {
arch/openrisc/mm/fault.c
152
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/openrisc/mm/init.c
245
[VM_EXEC] = PAGE_READONLY,
arch/openrisc/mm/init.c
246
[VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/openrisc/mm/init.c
247
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/openrisc/mm/init.c
248
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X,
arch/openrisc/mm/init.c
253
[VM_SHARED | VM_EXEC] = PAGE_READONLY,
arch/openrisc/mm/init.c
254
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/openrisc/mm/init.c
255
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED,
arch/openrisc/mm/init.c
256
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
arch/parisc/kernel/cache.c
350
if (vma->vm_flags & VM_EXEC)
arch/parisc/kernel/cache.c
761
if (vma->vm_flags & VM_EXEC)
arch/parisc/kernel/traps.c
721
&& (vma->vm_flags & VM_EXEC)) {
arch/parisc/kernel/vdso.c
84
VM_READ|VM_EXEC|
arch/parisc/mm/fault.c
57
return VM_EXEC;
arch/parisc/mm/init.c
967
[VM_EXEC] = PAGE_EXECREAD,
arch/parisc/mm/init.c
968
[VM_EXEC | VM_READ] = PAGE_EXECREAD,
arch/parisc/mm/init.c
969
[VM_EXEC | VM_WRITE] = PAGE_EXECREAD,
arch/parisc/mm/init.c
970
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_EXECREAD,
arch/parisc/mm/init.c
975
[VM_SHARED | VM_EXEC] = PAGE_EXECREAD,
arch/parisc/mm/init.c
976
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_EXECREAD,
arch/parisc/mm/init.c
977
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_RWX,
arch/parisc/mm/init.c
978
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_RWX
arch/powerpc/kernel/vdso.c
143
VM_READ | VM_EXEC | VM_MAYREAD |
arch/powerpc/mm/book3s64/pgtable.c
658
if (!radix_enabled() && ((vm_flags & VM_ACCESS_FLAGS) == VM_EXEC))
arch/powerpc/mm/book3s64/pkeys.c
382
if ((vma->vm_flags & VM_ACCESS_FLAGS) != VM_EXEC)
arch/powerpc/mm/cacheflush.c
206
if (!vma->vm_file && ((vma->vm_flags & VM_EXEC) == 0))
arch/powerpc/mm/copro_fault.c
44
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/powerpc/mm/fault.c
265
return !(vma->vm_flags & VM_EXEC) &&
arch/powerpc/mm/fault.c
285
if ((vma->vm_flags & VM_ACCESS_FLAGS) == VM_EXEC)
arch/powerpc/mm/pgtable.c
169
if (WARN_ON(!(vma->vm_flags & VM_EXEC)))
arch/powerpc/mm/pgtable.c
551
[VM_EXEC] = PAGE_EXECONLY_X,
arch/powerpc/mm/pgtable.c
552
[VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/powerpc/mm/pgtable.c
553
[VM_EXEC | VM_WRITE] = PAGE_COPY_X,
arch/powerpc/mm/pgtable.c
554
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X,
arch/powerpc/mm/pgtable.c
559
[VM_SHARED | VM_EXEC] = PAGE_EXECONLY_X,
arch/powerpc/mm/pgtable.c
560
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X,
arch/powerpc/mm/pgtable.c
561
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_X,
arch/powerpc/mm/pgtable.c
562
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X
arch/riscv/include/asm/cacheflush.h
39
if (vma->vm_flags & VM_EXEC) \
arch/riscv/kernel/vdso.c
144
(VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC | VM_SEALED_SYSMAP),
arch/riscv/mm/fault.c
253
if (!(vma->vm_flags & VM_EXEC)) {
arch/riscv/mm/init.c
375
[VM_EXEC] = PAGE_EXEC,
arch/riscv/mm/init.c
376
[VM_EXEC | VM_READ] = PAGE_READ_EXEC,
arch/riscv/mm/init.c
377
[VM_EXEC | VM_WRITE] = PAGE_COPY_EXEC,
arch/riscv/mm/init.c
378
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_EXEC,
arch/riscv/mm/init.c
383
[VM_SHARED | VM_EXEC] = PAGE_EXEC,
arch/riscv/mm/init.c
384
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READ_EXEC,
arch/riscv/mm/init.c
385
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC,
arch/riscv/mm/init.c
386
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC
arch/s390/kernel/vdso.c
69
VM_READ|VM_EXEC|VM_SEALED_SYSMAP|
arch/s390/mm/mmap.c
201
pm[VM_EXEC] = PAGE_RX;
arch/s390/mm/mmap.c
202
pm[VM_EXEC | VM_READ] = PAGE_RX;
arch/s390/mm/mmap.c
203
pm[VM_EXEC | VM_WRITE] = PAGE_RX;
arch/s390/mm/mmap.c
204
pm[VM_EXEC | VM_WRITE | VM_READ] = PAGE_RX;
arch/s390/mm/mmap.c
209
pm[VM_SHARED | VM_EXEC] = PAGE_RX;
arch/s390/mm/mmap.c
210
pm[VM_SHARED | VM_EXEC | VM_READ] = PAGE_RX;
arch/s390/mm/mmap.c
211
pm[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_RWX;
arch/s390/mm/mmap.c
212
pm[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_RWX;
arch/sh/kernel/vsyscall/vsyscall.c
101
VM_READ | VM_EXEC |
arch/sh/mm/cache-sh4.c
263
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/cache-sh4.c
305
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/cache.c
119
(vma->vm_flags & VM_EXEC))
arch/sh/mm/cache.c
77
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/fault.c
369
!(vma->vm_flags & VM_EXEC)))
arch/sh/mm/mmap.c
27
[VM_EXEC] = PAGE_EXECREAD,
arch/sh/mm/mmap.c
28
[VM_EXEC | VM_READ] = PAGE_EXECREAD,
arch/sh/mm/mmap.c
29
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/sh/mm/mmap.c
30
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY,
arch/sh/mm/mmap.c
35
[VM_SHARED | VM_EXEC] = PAGE_EXECREAD,
arch/sh/mm/mmap.c
36
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_EXECREAD,
arch/sh/mm/mmap.c
37
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_RWX,
arch/sh/mm/mmap.c
38
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_RWX
arch/sparc/mm/fault_32.c
165
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/fault_32.c
324
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/fault_64.c
402
if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) {
arch/sparc/mm/fault_64.c
417
(vma->vm_flags & VM_EXEC) != 0 &&
arch/sparc/mm/fault_64.c
425
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/init_32.c
290
[VM_EXEC] = PAGE_READONLY,
arch/sparc/mm/init_32.c
291
[VM_EXEC | VM_READ] = PAGE_READONLY,
arch/sparc/mm/init_32.c
292
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/sparc/mm/init_32.c
293
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY,
arch/sparc/mm/init_32.c
298
[VM_SHARED | VM_EXEC] = PAGE_READONLY,
arch/sparc/mm/init_32.c
299
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY,
arch/sparc/mm/init_32.c
300
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED,
arch/sparc/mm/init_32.c
301
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
arch/sparc/mm/init_64.c
3194
(VM_READ|VM_WRITE|VM_EXEC|VM_SHARED)]);
arch/sparc/mm/leon_mm.c
190
if (vma->vm_flags & VM_EXEC)
arch/sparc/mm/srmmu.c
1247
if (vma->vm_flags & VM_EXEC)
arch/sparc/vdso/vma.c
402
VM_READ|VM_EXEC|
arch/um/kernel/mem.c
134
[VM_EXEC] = PAGE_READONLY,
arch/um/kernel/mem.c
135
[VM_EXEC | VM_READ] = PAGE_READONLY,
arch/um/kernel/mem.c
136
[VM_EXEC | VM_WRITE] = PAGE_COPY,
arch/um/kernel/mem.c
137
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY,
arch/um/kernel/mem.c
142
[VM_SHARED | VM_EXEC] = PAGE_READONLY,
arch/um/kernel/mem.c
143
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY,
arch/um/kernel/mem.c
144
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED,
arch/um/kernel/mem.c
145
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED
arch/um/kernel/trap.c
173
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/x86/entry/vdso/vma.c
158
VM_READ|VM_EXEC|
arch/x86/entry/vsyscall/vsyscall_64.c
299
.vm_flags = VM_READ | VM_EXEC,
arch/x86/entry/vsyscall/vsyscall_64.c
378
vm_flags_init(&gate_vma, VM_EXEC);
arch/x86/kernel/uprobes.c
705
VM_READ|VM_EXEC|VM_MAYEXEC|VM_MAYREAD|VM_DONTCOPY|VM_IO,
arch/x86/mm/pgprot.c
13
[VM_EXEC] = PAGE_READONLY_EXEC,
arch/x86/mm/pgprot.c
14
[VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/x86/mm/pgprot.c
15
[VM_EXEC | VM_WRITE] = PAGE_COPY_EXEC,
arch/x86/mm/pgprot.c
16
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_EXEC,
arch/x86/mm/pgprot.c
21
[VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC,
arch/x86/mm/pgprot.c
22
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/x86/mm/pgprot.c
23
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC,
arch/x86/mm/pgprot.c
24
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC
arch/x86/mm/pgprot.c
38
(VM_READ|VM_WRITE|VM_EXEC|VM_SHARED)]);
arch/x86/mm/pkeys.c
65
if ((vma->vm_flags & VM_ACCESS_FLAGS) != VM_EXEC)
arch/x86/um/vdso/vma.c
48
VM_READ|VM_EXEC|
arch/xtensa/mm/cache.c
254
&& (vma->vm_flags & VM_EXEC) != 0) {
arch/xtensa/mm/cache.c
303
if ((vma->vm_flags & VM_EXEC) != 0)
arch/xtensa/mm/cache.c
307
} else if ((vma->vm_flags & VM_EXEC) != 0) {
arch/xtensa/mm/fault.c
149
if (!(vma->vm_flags & VM_EXEC))
arch/xtensa/mm/init.c
188
[VM_EXEC] = PAGE_READONLY_EXEC,
arch/xtensa/mm/init.c
189
[VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/xtensa/mm/init.c
190
[VM_EXEC | VM_WRITE] = PAGE_COPY_EXEC,
arch/xtensa/mm/init.c
191
[VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_EXEC,
arch/xtensa/mm/init.c
196
[VM_SHARED | VM_EXEC] = PAGE_READONLY_EXEC,
arch/xtensa/mm/init.c
197
[VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_EXEC,
arch/xtensa/mm/init.c
198
[VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_EXEC,
arch/xtensa/mm/init.c
199
[VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_EXEC
arch/xtensa/mm/tlb.c
108
if (vma->vm_flags & VM_EXEC)
arch/xtensa/mm/tlb.c
142
if (vma->vm_flags & VM_EXEC)
drivers/gpu/drm/panthor/panthor_device.c
465
(vma->vm_flags & (VM_WRITE | VM_EXEC)))
drivers/gpu/drm/xe/xe_device.c
331
if (vma->vm_flags & (VM_READ | VM_EXEC))
drivers/gpu/drm/xe/xe_oa.c
1694
if (vma->vm_flags & (VM_WRITE | VM_EXEC | VM_SHARED | VM_MAYSHARE)) {
drivers/infiniband/hw/mlx5/main.c
2446
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
drivers/iommu/iommu-sva.c
253
access_flags |= VM_EXEC;
drivers/iommu/iommufd/main.c
575
if (vma->vm_flags & VM_EXEC)
drivers/misc/ocxl/context.c
180
if ((vma->vm_flags & VM_READ) || (vma->vm_flags & VM_EXEC) ||
fs/binfmt_elf.c
2085
if (meta->flags & VM_EXEC)
fs/binfmt_elf.c
456
prot & PROT_EXEC ? VM_EXEC : 0);
fs/binfmt_elf_fdpic.c
1605
if (meta->flags & VM_EXEC)
fs/binfmt_elf_fdpic.c
417
(executable_stack == EXSTACK_DEFAULT && VM_STACK_FLAGS & VM_EXEC))
fs/exec.c
661
vm_flags |= VM_EXEC;
fs/exec.c
663
vm_flags &= ~VM_EXEC;
fs/exec.c
678
if (unlikely(vm_flags & VM_EXEC)) {
fs/proc/nommu.c
54
flags & VM_EXEC ? 'x' : '-',
fs/proc/task_mmu.c
1155
[ilog2(VM_EXEC)] = "ex",
fs/proc/task_mmu.c
453
seq_putc(m, flags & VM_EXEC ? 'x' : '-');
fs/proc/task_mmu.c
630
perm |= VM_EXEC;
fs/proc/task_mmu.c
710
if (vma->vm_flags & VM_EXEC)
fs/proc/task_nommu.c
153
flags & VM_EXEC ? 'x' : '-',
fs/proc/vmcore.c
603
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
include/asm-generic/tlb.h
460
.vm_flags = (tlb->vma_exec ? VM_EXEC : 0) |
include/asm-generic/tlb.h
486
tlb->vma_exec = !!(vma->vm_flags & VM_EXEC);
include/linux/mm.h
516
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
include/linux/mm.h
523
#define VM_DATA_FLAGS_EXEC (VM_READ | VM_WRITE | VM_EXEC | \
include/linux/mm.h
545
#define VM_ACCESS_FLAGS (VM_READ | VM_WRITE | VM_EXEC)
include/linux/mman.h
145
_calc_vm_trans(prot, PROT_EXEC, VM_EXEC) |
include/linux/mman.h
208
if (!(new & VM_EXEC))
include/linux/mman.h
216
if (!(old & VM_EXEC))
include/linux/pgtable.h
2368
(VM_READ | VM_WRITE | VM_EXEC | VM_SHARED)]; \
include/trace/events/mmflags.h
206
{VM_EXEC, "exec" }, \
kernel/bpf/sysfs_btf.c
33
if (vma->vm_flags & (VM_WRITE | VM_EXEC | VM_MAYSHARE))
kernel/events/core.c
9611
int executable = vma->vm_flags & VM_EXEC;
kernel/events/core.c
9700
if (vma->vm_flags & VM_EXEC)
kernel/events/core.c
9777
if (!(vma->vm_flags & VM_EXEC))
kernel/events/core.c
9879
if (!(vma->vm_flags & VM_EXEC))
kernel/events/uprobes.c
1726
VM_EXEC|VM_MAYEXEC|VM_DONTCOPY|VM_IO|
kernel/trace/ring_buffer.c
7161
if (vma->vm_flags & VM_WRITE || vma->vm_flags & VM_EXEC ||
lib/tests/printf_kunit.c
661
flags = VM_READ | VM_EXEC | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
mm/debug_vm_pgtable.c
1294
#define VM_FLAGS_END (VM_SHARED | VM_EXEC | VM_WRITE | VM_READ)
mm/filemap.c
3326
if ((vm_flags & (VM_RAND_READ | VM_EXEC)) == VM_RAND_READ)
mm/filemap.c
3369
if (vm_flags & VM_EXEC) {
mm/migrate.c
2672
if ((vma->vm_flags & VM_EXEC) && folio_maybe_mapped_shared(folio))
mm/mmap.c
1128
prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0;
mm/mmap.c
1221
if ((vm_flags & (~VM_EXEC)) != 0)
mm/mmap.c
471
if (vm_flags & VM_EXEC)
mm/nommu.c
1208
if (vma->vm_flags & VM_EXEC && !region->vm_icache_flushed) {
mm/vma.h
504
return (flags & (VM_EXEC | VM_WRITE | VM_STACK)) == VM_EXEC;
mm/vmscan.c
2161
if ((vm_flags & VM_EXEC) && folio_is_file_lru(folio)) {
mm/vmscan.c
940
if ((vm_flags & VM_EXEC) && folio_is_file_lru(folio))
net/ipv4/tcp.c
1896
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
rust/bindings/bindings_helper.h
126
const vm_flags_t RUST_CONST_HELPER_VM_EXEC = VM_EXEC;
security/integrity/ima/ima_main.c
541
!(prot & PROT_EXEC) || (vma->vm_flags & VM_EXEC))
security/ipe/hooks.c
114
if (vma->vm_flags & VM_EXEC)
security/selinux/hooks.c
4024
(prot & PROT_EXEC) && !(vma->vm_flags & VM_EXEC)) {
security/selinux/hooks.c
7732
default_noexec = !(VM_DATA_DEFAULT_FLAGS & VM_EXEC);
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
45
perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-';
tools/testing/selftests/bpf/progs/find_vma.c
31
if (vma->vm_flags & VM_EXEC)
tools/testing/vma/include/dup.h
1013
((flags & (VM_EXEC | VM_WRITE | VM_STACK)) == VM_EXEC)
tools/testing/vma/include/dup.h
1266
if (!(new & VM_EXEC))
tools/testing/vma/include/dup.h
1274
if (!(old & VM_EXEC))
tools/testing/vma/include/dup.h
314
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
tools/testing/vma/include/dup.h
321
#define VM_DATA_FLAGS_EXEC (VM_READ | VM_WRITE | VM_EXEC | \
tools/testing/vma/include/dup.h
337
#define VM_ACCESS_FLAGS (VM_READ | VM_WRITE | VM_EXEC)
tools/testing/vma/include/dup.h
353
#define TASK_EXEC ((current->personality & READ_IMPLIES_EXEC) ? VM_EXEC : 0)
tools/testing/vma/tests/vma.c
156
vma_flags_clear_word(&flags, VM_EXEC);
virt/kvm/kvm_main.c
4090
((vma->vm_flags & VM_EXEC) || !(vma->vm_flags & VM_SHARED)))