arch/alpha/include/asm/cacheflush.h
41
if (vma->vm_flags & VM_EXEC) {
arch/alpha/include/asm/tlbflush.h
32
if (vma->vm_flags & VM_EXEC)
arch/alpha/kernel/osf_sys.c
1218
unsigned long flags, vm_flags_t vm_flags)
arch/alpha/kernel/smp.c
735
if ((vma->vm_flags & VM_EXEC) == 0)
arch/alpha/mm/fault.c
130
if (!(vma->vm_flags & VM_EXEC))
arch/alpha/mm/fault.c
134
if (!(vma->vm_flags & (VM_READ | VM_WRITE)))
arch/alpha/mm/fault.c
137
if (!(vma->vm_flags & VM_WRITE))
arch/alpha/mm/tlbflush.c
35
int tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2;
arch/alpha/mm/tlbflush.c
84
.tbi_type = (vma->vm_flags & VM_EXEC) ? 3 : 2,
arch/arc/include/asm/cacheflush.h
62
if (vma->vm_flags & VM_EXEC) \
arch/arc/mm/fault.c
130
if (!(vma->vm_flags & mask)) {
arch/arc/mm/mmap.c
27
unsigned long flags, vm_flags_t vm_flags)
arch/arc/mm/tlb.c
489
if (vma->vm_flags & VM_EXEC) {
arch/arc/mm/tlb.c
501
if (vma->vm_flags & VM_EXEC)
arch/arm/include/asm/cacheflush.h
231
vma->vm_flags);
arch/arm/include/asm/cacheflush.h
242
vma->vm_flags);
arch/arm/kernel/asm-offsets.c
112
DEFINE(VMA_VM_FLAGS, offsetof(struct vm_area_struct, vm_flags));
arch/arm/mm/fault-armv.c
159
if (!(mpnt->vm_flags & VM_MAYSHARE))
arch/arm/mm/fault-armv.c
211
else if (vma->vm_flags & VM_EXEC)
arch/arm/mm/fault.c
300
vm_flags_t vm_flags = VM_ACCESS_FLAGS;
arch/arm/mm/fault.c
328
vm_flags = VM_WRITE;
arch/arm/mm/fault.c
332
vm_flags = VM_EXEC;
arch/arm/mm/fault.c
356
if (!(vma->vm_flags & vm_flags)) {
arch/arm/mm/fault.c
395
if (!(vma->vm_flags & vm_flags)) {
arch/arm/mm/flush.c
110
if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged())
arch/arm/mm/flush.c
165
if (vma->vm_flags & VM_EXEC)
arch/arm/mm/flush.c
263
if (!(vma->vm_flags & VM_MAYSHARE))
arch/arm/mm/flush.c
94
if (vma->vm_flags & VM_EXEC)
arch/arm/mm/mmap.c
32
unsigned long flags, vm_flags_t vm_flags)
arch/arm/mm/mmap.c
83
const unsigned long flags, vm_flags_t vm_flags)
arch/arm/mm/nommu.c
189
if (vma->vm_flags & VM_EXEC)
arch/arm64/include/asm/mman.h
71
static inline bool arch_validate_flags(vm_flags_t vm_flags)
arch/arm64/include/asm/mman.h
78
if ((vm_flags & VM_MTE) && !(vm_flags & VM_MTE_ALLOWED))
arch/arm64/include/asm/mman.h
82
if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) {
arch/arm64/include/asm/mman.h
84
if (vm_flags & VM_EXEC)
arch/arm64/include/asm/mman.h
88
VM_WARN_ON(vm_flags & VM_SHARED);
arch/arm64/include/asm/mman.h
94
#define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags)
arch/arm64/include/asm/pkeys.h
25
return (vma->vm_flags & ARCH_VM_PKEY_FLAGS) >> VM_PKEY_SHIFT;
arch/arm64/kernel/mte.c
471
if (!(vma->vm_flags & VM_MTE)) {
arch/arm64/kvm/mmu.c
1061
if (!(vma->vm_flags & VM_PFNMAP)) {
arch/arm64/kvm/mmu.c
1437
if (is_vm_hugetlb_page(vma) && !(vma->vm_flags & VM_PFNMAP))
arch/arm64/kvm/mmu.c
1440
if (!(vma->vm_flags & VM_PFNMAP))
arch/arm64/kvm/mmu.c
1500
return vma->vm_flags & VM_MTE_ALLOWED;
arch/arm64/kvm/mmu.c
1666
vm_flags_t vm_flags;
arch/arm64/kvm/mmu.c
1768
vfio_allow_any_uc = vma->vm_flags & VM_ALLOW_ANY_UNCACHED;
arch/arm64/kvm/mmu.c
1770
vm_flags = vma->vm_flags;
arch/arm64/kvm/mmu.c
1801
if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP) && !pfn_is_map_memory(pfn)) {
arch/arm64/kvm/mmu.c
2459
if (vma->vm_flags & VM_PFNMAP) {
arch/arm64/mm/fault.c
546
if (!(vma->vm_flags & VM_SHADOW_STACK))
arch/arm64/mm/fault.c
548
} else if (unlikely(vma->vm_flags & VM_SHADOW_STACK)) {
arch/arm64/mm/fault.c
562
vm_flags_t vm_flags;
arch/arm64/mm/fault.c
590
vm_flags = VM_EXEC;
arch/arm64/mm/fault.c
598
vm_flags = VM_WRITE;
arch/arm64/mm/fault.c
602
vm_flags = VM_WRITE;
arch/arm64/mm/fault.c
606
vm_flags = VM_READ;
arch/arm64/mm/fault.c
608
vm_flags |= VM_WRITE;
arch/arm64/mm/fault.c
611
vm_flags |= VM_EXEC;
arch/arm64/mm/fault.c
640
if (!(vma->vm_flags & vm_flags)) {
arch/arm64/mm/fault.c
685
if (!(vma->vm_flags & vm_flags)) {
arch/arm64/mm/fault.c
968
if (vma->vm_flags & VM_MTE)
arch/arm64/mm/flush.c
35
if (vma->vm_flags & VM_EXEC)
arch/arm64/mm/mmap.c
100
prot = pgprot_val(protection_map[vm_flags &
arch/arm64/mm/mmap.c
104
if (vm_flags & VM_ARM64_BTI)
arch/arm64/mm/mmap.c
117
if (vm_flags & VM_MTE)
arch/arm64/mm/mmap.c
122
if (vm_flags & VM_PKEY_BIT0)
arch/arm64/mm/mmap.c
124
if (vm_flags & VM_PKEY_BIT1)
arch/arm64/mm/mmap.c
126
if (vm_flags & VM_PKEY_BIT2)
arch/arm64/mm/mmap.c
88
pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
arch/arm64/mm/mmap.c
93
if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) {
arch/arm64/mm/mmap.c
95
if (vm_flags & VM_ACCESS_FLAGS)
arch/arm64/mm/mmu.c
1204
unsigned long vm_flags)
arch/arm64/mm/mmu.c
1212
if (!(vm_flags & VM_NO_GUARD))
arch/arm64/mm/mmu.c
1218
vma->flags = VM_MAP | vm_flags;
arch/csky/abiv1/cacheflush.c
63
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv1/cacheflush.c
73
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv1/mmap.c
27
unsigned long flags, vm_flags_t vm_flags)
arch/csky/abiv2/cacheflush.c
33
if (vma->vm_flags & VM_EXEC)
arch/csky/abiv2/inc/abi/cacheflush.h
50
if (vma->vm_flags & VM_EXEC) { \
arch/csky/mm/fault.c
174
if (!(vma->vm_flags & VM_WRITE))
arch/hexagon/mm/cache.c
122
if (vma->vm_flags & VM_EXEC) {
arch/hexagon/mm/vm_fault.c
70
if (!(vma->vm_flags & VM_EXEC))
arch/hexagon/mm/vm_fault.c
74
if (!(vma->vm_flags & VM_READ))
arch/hexagon/mm/vm_fault.c
78
if (!(vma->vm_flags & VM_WRITE))
arch/loongarch/mm/fault.c
228
if (!(vma->vm_flags & VM_WRITE)) {
arch/loongarch/mm/fault.c
235
if (!(vma->vm_flags & VM_EXEC) && address == exception_era(regs)) {
arch/loongarch/mm/fault.c
241
if (!(vma->vm_flags & (VM_READ | VM_WRITE)) && address != exception_era(regs)) {
arch/loongarch/mm/fault.c
295
if (!(vma->vm_flags & VM_WRITE))
arch/loongarch/mm/fault.c
298
if (!(vma->vm_flags & VM_EXEC) && address == exception_era(regs))
arch/loongarch/mm/fault.c
300
if (!(vma->vm_flags & (VM_READ | VM_WRITE)) && address != exception_era(regs))
arch/loongarch/mm/mmap.c
109
unsigned long flags, vm_flags_t vm_flags)
arch/loongarch/mm/mmap.c
97
vm_flags_t vm_flags)
arch/m68k/mm/fault.c
100
if (!(vma->vm_flags & VM_GROWSDOWN))
arch/m68k/mm/fault.c
124
if (!(vma->vm_flags & VM_WRITE))
arch/microblaze/include/asm/cacheflush.h
94
if (vma->vm_flags & VM_EXEC) {
arch/microblaze/mm/fault.c
157
if (unlikely(!(vma->vm_flags & VM_GROWSDOWN)))
arch/microblaze/mm/fault.c
204
if (unlikely(!(vma->vm_flags & VM_WRITE)))
arch/microblaze/mm/fault.c
212
if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC))))
arch/mips/kernel/smp.c
632
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-octeon.c
142
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/c-octeon.c
157
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/c-r3k.c
240
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
473
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
493
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/c-r4k.c
543
int exec = vma->vm_flags & VM_EXEC;
arch/mips/mm/fault.c
113
if (!(vma->vm_flags & VM_WRITE))
arch/mips/mm/fault.c
118
if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) {
arch/mips/mm/fault.c
128
if (!(vma->vm_flags & VM_READ) &&
arch/mips/mm/init.c
204
if (vma->vm_flags & VM_EXEC)
arch/mips/mm/mmap.c
102
vm_flags_t vm_flags)
arch/mips/mm/mmap.c
114
unsigned long flags, vm_flags_t vm_flags)
arch/mips/mm/tlb-r4k.c
54
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
138
if (vma == NULL || (vma->vm_flags & VM_EXEC))
arch/nios2/mm/cacheflush.c
159
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
236
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
268
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
279
if (vma->vm_flags & VM_EXEC)
arch/nios2/mm/cacheflush.c
90
if (!(vma->vm_flags & VM_MAYSHARE))
arch/nios2/mm/fault.c
105
if (!(vma->vm_flags & VM_EXEC))
arch/nios2/mm/fault.c
109
if (!(vma->vm_flags & VM_READ))
arch/nios2/mm/fault.c
113
if (!(vma->vm_flags & VM_WRITE))
arch/openrisc/include/asm/cacheflush.h
90
if (vma->vm_flags & VM_EXEC) \
arch/openrisc/mm/cache.c
93
if ((vma->vm_flags & VM_EXEC) && dirty) {
arch/openrisc/mm/fault.c
119
if (!(vma->vm_flags & VM_GROWSDOWN))
arch/openrisc/mm/fault.c
147
if (!(vma->vm_flags & VM_WRITE))
arch/openrisc/mm/fault.c
152
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/parisc/kernel/cache.c
350
if (vma->vm_flags & VM_EXEC)
arch/parisc/kernel/cache.c
531
if (!(vma->vm_flags & VM_SHARED))
arch/parisc/kernel/cache.c
761
if (vma->vm_flags & VM_EXEC)
arch/parisc/kernel/sys_parisc.c
171
vm_flags_t vm_flags)
arch/parisc/kernel/sys_parisc.c
179
unsigned long flags, vm_flags_t vm_flags)
arch/parisc/kernel/traps.c
721
&& (vma->vm_flags & VM_EXEC)) {
arch/parisc/mm/fault.c
297
if (!prev_vma || !(prev_vma->vm_flags & VM_GROWSUP))
arch/parisc/mm/fault.c
309
if ((vma->vm_flags & acc_type) != acc_type)
arch/parisc/mm/fault.c
384
if ((vma->vm_flags & acc_type) != acc_type) {
arch/parisc/mm/fault.c
506
&& (vma->vm_flags & acc_type) == acc_type)
arch/powerpc/include/asm/book3s/64/hash-pkey.h
10
return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT0 : 0x0UL) |
arch/powerpc/include/asm/book3s/64/hash-pkey.h
11
((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT1 : 0x0UL) |
arch/powerpc/include/asm/book3s/64/hash-pkey.h
12
((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) |
arch/powerpc/include/asm/book3s/64/hash-pkey.h
13
((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT3 : 0x0UL) |
arch/powerpc/include/asm/book3s/64/hash-pkey.h
14
((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT4 : 0x0UL));
arch/powerpc/include/asm/book3s/64/hash-pkey.h
8
static inline u64 hash__vmflag_to_pte_pkey_bits(u64 vm_flags)
arch/powerpc/include/asm/book3s/64/pkeys.h
15
return hash__vmflag_to_pte_pkey_bits(vm_flags);
arch/powerpc/include/asm/book3s/64/pkeys.h
8
static inline u64 vmflag_to_pte_pkey_bits(vm_flags_t vm_flags)
arch/powerpc/include/asm/pkeys.h
42
return (vma->vm_flags & ARCH_VM_PKEY_FLAGS) >> VM_PKEY_SHIFT;
arch/powerpc/kernel/pci-common.c
681
if (vma->vm_flags & VM_SHARED)
arch/powerpc/kvm/book3s_hv.c
5514
if (!vma || (vma->vm_flags & VM_IO))
arch/powerpc/kvm/book3s_hv_uvmem.c
396
vm_flags_t vm_flags;
arch/powerpc/kvm/book3s_hv_uvmem.c
415
vm_flags = vma->vm_flags;
arch/powerpc/kvm/book3s_hv_uvmem.c
417
merge_flag, &vm_flags);
arch/powerpc/kvm/book3s_hv_uvmem.c
422
vm_flags_reset(vma, vm_flags);
arch/powerpc/mm/book3s64/pgtable.c
653
pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
arch/powerpc/mm/book3s64/pgtable.c
658
if (!radix_enabled() && ((vm_flags & VM_ACCESS_FLAGS) == VM_EXEC))
arch/powerpc/mm/book3s64/pgtable.c
659
vm_flags |= VM_READ;
arch/powerpc/mm/book3s64/pgtable.c
661
prot = pgprot_val(protection_map[vm_flags & (VM_ACCESS_FLAGS | VM_SHARED)]);
arch/powerpc/mm/book3s64/pgtable.c
663
if (vm_flags & VM_SAO)
arch/powerpc/mm/book3s64/pgtable.c
667
prot |= vmflag_to_pte_pkey_bits(vm_flags);
arch/powerpc/mm/book3s64/pkeys.c
382
if ((vma->vm_flags & VM_ACCESS_FLAGS) != VM_EXEC)
arch/powerpc/mm/book3s64/slice.c
657
vm_flags_t vm_flags)
arch/powerpc/mm/book3s64/slice.c
662
return generic_get_unmapped_area(filp, addr, len, pgoff, flags, vm_flags);
arch/powerpc/mm/book3s64/slice.c
677
vm_flags_t vm_flags)
arch/powerpc/mm/book3s64/slice.c
682
return generic_get_unmapped_area_topdown(filp, addr0, len, pgoff, flags, vm_flags);
arch/powerpc/mm/cacheflush.c
206
if (!vma->vm_file && ((vma->vm_flags & VM_EXEC) == 0))
arch/powerpc/mm/copro_fault.c
41
if (!(vma->vm_flags & VM_WRITE))
arch/powerpc/mm/copro_fault.c
44
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/powerpc/mm/fault.c
265
return !(vma->vm_flags & VM_EXEC) &&
arch/powerpc/mm/fault.c
267
!(vma->vm_flags & (VM_READ | VM_WRITE)));
arch/powerpc/mm/fault.c
271
if (unlikely(!(vma->vm_flags & VM_WRITE)))
arch/powerpc/mm/fault.c
285
if ((vma->vm_flags & VM_ACCESS_FLAGS) == VM_EXEC)
arch/powerpc/mm/pgtable.c
169
if (WARN_ON(!(vma->vm_flags & VM_EXEC)))
arch/powerpc/platforms/cell/spufs/file.c
1036
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
1170
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
1291
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
1352
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
1411
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
269
if (write && !(vma->vm_flags & VM_WRITE))
arch/powerpc/platforms/cell/spufs/file.c
291
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/cell/spufs/file.c
381
if (!(vma->vm_flags & VM_SHARED))
arch/powerpc/platforms/powernv/opal-prd.c
120
vma->vm_flags);
arch/riscv/include/asm/cacheflush.h
39
if (vma->vm_flags & VM_EXEC) \
arch/riscv/kvm/mmu.c
222
if (writable && !(vma->vm_flags & VM_WRITE)) {
arch/riscv/kvm/mmu.c
230
if (vma->vm_flags & VM_PFNMAP) {
arch/riscv/kvm/mmu.c
495
if (logging || (vma->vm_flags & VM_PFNMAP))
arch/riscv/mm/fault.c
253
if (!(vma->vm_flags & VM_EXEC)) {
arch/riscv/mm/fault.c
259
if (!(vma->vm_flags & (VM_READ | VM_WRITE))) {
arch/riscv/mm/fault.c
264
if (!(vma->vm_flags & VM_WRITE)) {
arch/riscv/mm/pgtable.c
171
if (vma->vm_flags & VM_SHADOW_STACK)
arch/riscv/mm/pgtable.c
179
if (vma->vm_flags & VM_SHADOW_STACK)
arch/s390/include/uapi/asm/ipl.h
108
__u8 vm_flags;
arch/s390/kernel/ipl.c
1130
reipl_block_nss->ccw.vm_flags |= IPL_PB0_CCW_VM_FLAG_NSS;
arch/s390/kernel/ipl.c
1135
reipl_block_nss->ccw.vm_flags &= ~IPL_PB0_CCW_VM_FLAG_NSS;
arch/s390/kernel/ipl.c
1298
(ipl_block.ccw.vm_flags & IPL_PB0_CCW_VM_FLAG_VP)) {
arch/s390/kernel/ipl.c
1300
ipb->ccw.vm_flags |= IPL_PB0_CCW_VM_FLAG_VP;
arch/s390/kernel/ipl.c
778
ipb->ccw.vm_flags |= IPL_PB0_CCW_VM_FLAG_VP;
arch/s390/kernel/ipl.c
782
ipb->ccw.vm_flags &= ~IPL_PB0_CCW_VM_FLAG_VP;
arch/s390/kernel/ipl_vmparm.c
16
if ((ipb->ccw.vm_flags & IPL_PB0_CCW_VM_FLAG_VP) &&
arch/s390/mm/fault.c
300
if (!(vma->vm_flags & access)) {
arch/s390/mm/fault.c
326
if (unlikely(!(vma->vm_flags & access)))
arch/s390/mm/gmap_helpers.c
190
if (!is_cow_mapping(walk->vma->vm_flags))
arch/s390/mm/gmap_helpers.c
234
if ((vma->vm_flags & VM_PFNMAP) || is_vm_hugetlb_page(vma))
arch/s390/mm/mmap.c
116
unsigned long flags, vm_flags_t vm_flags)
arch/s390/mm/mmap.c
80
unsigned long flags, vm_flags_t vm_flags)
arch/s390/pci/pci_mmio.c
175
if (!(vma->vm_flags & (VM_IO | VM_PFNMAP)))
arch/s390/pci/pci_mmio.c
178
if (!(vma->vm_flags & VM_WRITE))
arch/s390/pci/pci_mmio.c
327
if (!(vma->vm_flags & (VM_IO | VM_PFNMAP)))
arch/s390/pci/pci_mmio.c
330
if (!(vma->vm_flags & VM_READ))
arch/sh/mm/cache-sh4.c
263
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/cache-sh4.c
305
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/cache.c
119
(vma->vm_flags & VM_EXEC))
arch/sh/mm/cache.c
77
if (vma->vm_flags & VM_EXEC)
arch/sh/mm/fault.c
362
if (unlikely(!(vma->vm_flags & VM_WRITE)))
arch/sh/mm/fault.c
369
!(vma->vm_flags & VM_EXEC)))
arch/sh/mm/mmap.c
103
const unsigned long flags, vm_flags_t vm_flags)
arch/sh/mm/mmap.c
56
vm_flags_t vm_flags)
arch/sparc/include/asm/mman.h
57
#define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags)
arch/sparc/include/asm/mman.h
61
static inline bool arch_validate_flags(vm_flags_t vm_flags)
arch/sparc/include/asm/mman.h
67
if (vm_flags & VM_SPARC_ADI) {
arch/sparc/include/asm/mman.h
72
if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP))
arch/sparc/include/asm/mman.h
83
if (vm_flags & VM_MERGEABLE)
arch/sparc/kernel/sys_sparc_32.c
43
unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags)
arch/sparc/kernel/sys_sparc_64.c
101
unsigned long arch_get_unmapped_area(struct file *filp, unsigned long addr, unsigned long len, unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags)
arch/sparc/kernel/sys_sparc_64.c
165
const unsigned long flags, vm_flags_t vm_flags)
arch/sparc/mm/fault_32.c
161
if (!(vma->vm_flags & VM_WRITE))
arch/sparc/mm/fault_32.c
165
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/fault_32.c
320
if (!(vma->vm_flags & VM_WRITE))
arch/sparc/mm/fault_32.c
324
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/fault_64.c
352
(vma->vm_flags & VM_WRITE) != 0) {
arch/sparc/mm/fault_64.c
373
if (!(vma->vm_flags & VM_GROWSDOWN))
arch/sparc/mm/fault_64.c
402
if ((fault_code & FAULT_CODE_ITLB) && !(vma->vm_flags & VM_EXEC)) {
arch/sparc/mm/fault_64.c
410
if (!(vma->vm_flags & VM_WRITE))
arch/sparc/mm/fault_64.c
417
(vma->vm_flags & VM_EXEC) != 0 &&
arch/sparc/mm/fault_64.c
425
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/sparc/mm/init_64.c
3136
if (vma->vm_flags & VM_SPARC_ADI) {
arch/sparc/mm/init_64.c
3191
pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
arch/sparc/mm/init_64.c
3193
unsigned long prot = pgprot_val(protection_map[vm_flags &
arch/sparc/mm/init_64.c
3196
if (vm_flags & VM_SPARC_ADI)
arch/sparc/mm/leon_mm.c
190
if (vma->vm_flags & VM_EXEC)
arch/sparc/mm/srmmu.c
1247
if (vma->vm_flags & VM_EXEC)
arch/um/kernel/trap.c
120
if (!(vma->vm_flags & VM_GROWSDOWN))
arch/um/kernel/trap.c
168
if (!(vma->vm_flags & VM_WRITE))
arch/um/kernel/trap.c
173
if (!(vma->vm_flags & (VM_READ | VM_EXEC)))
arch/um/kernel/trap.c
97
if (!vma || !(vma->vm_flags & VM_GROWSDOWN)) {
arch/x86/entry/vsyscall/vsyscall_64.c
299
.vm_flags = VM_READ | VM_EXEC,
arch/x86/include/asm/pkeys.h
123
return (vma->vm_flags & vma_pkey_mask) >> VM_PKEY_SHIFT;
arch/x86/include/asm/tlbflush.h
311
((vma)->vm_flags & VM_HUGETLB) \
arch/x86/kernel/cpu/sgx/driver.c
106
ret = sgx_encl_may_map(encl, vma->vm_start, vma->vm_end, vma->vm_flags);
arch/x86/kernel/cpu/sgx/encl.c
282
vm_flags_t vm_flags)
arch/x86/kernel/cpu/sgx/encl.c
284
unsigned long vm_prot_bits = vm_flags & VM_ACCESS_FLAGS;
arch/x86/kernel/cpu/sgx/encl.c
462
entry = sgx_encl_load_page_in_vma(encl, addr, vma->vm_flags);
arch/x86/kernel/cpu/sgx/encl.c
523
unsigned long end, vm_flags_t vm_flags)
arch/x86/kernel/cpu/sgx/encl.c
525
vm_flags_t vm_prot_bits = vm_flags & VM_ACCESS_FLAGS;
arch/x86/kernel/cpu/sgx/encl.c
608
vm_flags_t vm_flags)
arch/x86/kernel/cpu/sgx/encl.c
615
entry = sgx_encl_load_page_in_vma(encl, addr, vm_flags);
arch/x86/kernel/cpu/sgx/encl.c
652
vma->vm_flags);
arch/x86/kernel/cpu/sgx/encl.h
104
unsigned long end, vm_flags_t vm_flags);
arch/x86/kernel/cpu/sgx/ioctl.c
222
if (!(vma->vm_flags & VM_MAYEXEC))
arch/x86/kernel/cpu/sgx/virt.c
104
if (!(vma->vm_flags & VM_SHARED))
arch/x86/kernel/shstk.c
363
if (!vma || !(vma->vm_flags & VM_SHADOW_STACK)) {
arch/x86/kernel/sys_x86_64.c
118
static inline unsigned long stack_guard_placement(vm_flags_t vm_flags)
arch/x86/kernel/sys_x86_64.c
120
if (vm_flags & VM_SHADOW_STACK)
arch/x86/kernel/sys_x86_64.c
128
unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags)
arch/x86/kernel/sys_x86_64.c
156
info.start_gap = stack_guard_placement(vm_flags);
arch/x86/kernel/sys_x86_64.c
169
unsigned long flags, vm_flags_t vm_flags)
arch/x86/kernel/sys_x86_64.c
209
info.start_gap = stack_guard_placement(vm_flags);
arch/x86/mm/fault.c
1088
if (unlikely(!(vma->vm_flags & VM_SHADOW_STACK)))
arch/x86/mm/fault.c
1090
if (unlikely(!(vma->vm_flags & VM_WRITE)))
arch/x86/mm/fault.c
1097
if (unlikely(vma->vm_flags & VM_SHADOW_STACK))
arch/x86/mm/fault.c
1099
if (unlikely(!(vma->vm_flags & VM_WRITE)))
arch/x86/mm/pgprot.c
35
pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
arch/x86/mm/pgprot.c
37
unsigned long val = pgprot_val(protection_map[vm_flags &
arch/x86/mm/pgprot.c
48
if (vm_flags & VM_PKEY_BIT0)
arch/x86/mm/pgprot.c
50
if (vm_flags & VM_PKEY_BIT1)
arch/x86/mm/pgprot.c
52
if (vm_flags & VM_PKEY_BIT2)
arch/x86/mm/pgprot.c
54
if (vm_flags & VM_PKEY_BIT3)
arch/x86/mm/pgtable.c
804
if (vma->vm_flags & VM_SHADOW_STACK)
arch/x86/mm/pgtable.c
814
if (vma->vm_flags & VM_SHADOW_STACK)
arch/x86/mm/pgtable.c
831
VM_WARN_ON_ONCE(!(vma->vm_flags & VM_SHADOW_STACK) &&
arch/x86/mm/pgtable.c
838
VM_WARN_ON_ONCE(!(vma->vm_flags & VM_SHADOW_STACK) &&
arch/x86/mm/pgtable.c
845
VM_WARN_ON_ONCE(!(vma->vm_flags & VM_SHADOW_STACK) && pud_shstk(pud));
arch/x86/mm/pkeys.c
65
if ((vma->vm_flags & VM_ACCESS_FLAGS) != VM_EXEC)
arch/x86/xen/mmu_pv.c
2490
BUG_ON(!((vma->vm_flags & (VM_PFNMAP | VM_IO)) == (VM_PFNMAP | VM_IO)));
arch/xtensa/kernel/syscall.c
59
vm_flags_t vm_flags)
arch/xtensa/mm/cache.c
254
&& (vma->vm_flags & VM_EXEC) != 0) {
arch/xtensa/mm/cache.c
303
if ((vma->vm_flags & VM_EXEC) != 0)
arch/xtensa/mm/cache.c
307
} else if ((vma->vm_flags & VM_EXEC) != 0) {
arch/xtensa/mm/fault.c
145
if (!(vma->vm_flags & VM_WRITE))
arch/xtensa/mm/fault.c
149
if (!(vma->vm_flags & VM_EXEC))
arch/xtensa/mm/fault.c
152
if (!(vma->vm_flags & (VM_READ | VM_WRITE)))
arch/xtensa/mm/tlb.c
108
if (vma->vm_flags & VM_EXEC)
arch/xtensa/mm/tlb.c
142
if (vma->vm_flags & VM_EXEC)
drivers/acpi/pfr_telemetry.c
306
if (vma->vm_flags & VM_WRITE)
drivers/android/binder.c
5998
(vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
drivers/android/binder.c
6009
(vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
drivers/android/binder.c
6036
(vma->vm_end - vma->vm_start) / SZ_1K, vma->vm_flags,
drivers/android/binder.c
6039
if (vma->vm_flags & FORBIDDEN_MMAP_FLAGS) {
drivers/block/ublk_drv.c
2562
if (vma->vm_flags & VM_WRITE)
drivers/char/uv_mmtimer.c
154
if (vma->vm_flags & VM_WRITE)
drivers/comedi/comedi_fops.c
2509
if (vma->vm_flags & VM_WRITE)
drivers/dma-buf/heaps/cma_heap.c
202
if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0)
drivers/dma-buf/udmabuf.c
100
if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0)
drivers/firewire/core-cdev.c
1792
if (!(vma->vm_flags & VM_SHARED))
drivers/fpga/dfl-afu-main.c
817
if (!(vma->vm_flags & VM_SHARED))
drivers/fpga/dfl-afu-main.c
830
if ((vma->vm_flags & VM_READ) && !(region.flags & DFL_PORT_REGION_READ))
drivers/fpga/dfl-afu-main.c
833
if ((vma->vm_flags & VM_WRITE) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
381
if (is_cow_mapping(vma->vm_flags) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
382
!(vma->vm_flags & VM_ACCESS_FLAGS))
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
172
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
369
#define amdgpu_gmc_get_vm_pte(adev, vm, bo, vm_flags, pte_flags) \
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
370
((adev)->gmc.gmc_funcs->get_vm_pte((adev), (vm), (bo), (vm_flags), \
drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c
480
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c
483
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c
488
switch (vm_flags & AMDGPU_VM_MTYPE_MASK) {
drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c
505
if (vm_flags & AMDGPU_VM_PAGE_NOALLOC)
drivers/gpu/drm/amd/amdgpu/gmc_v10_0.c
510
if (vm_flags & AMDGPU_VM_PAGE_PRT) {
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
471
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
474
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
479
switch (vm_flags & AMDGPU_VM_MTYPE_MASK) {
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
496
if (vm_flags & AMDGPU_VM_PAGE_NOALLOC)
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
501
if (vm_flags & AMDGPU_VM_PAGE_PRT) {
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
510
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
513
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
518
switch (vm_flags & AMDGPU_VM_MTYPE_MASK) {
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
531
if (vm_flags & AMDGPU_VM_PAGE_NOALLOC)
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
536
if (vm_flags & AMDGPU_VM_PAGE_PRT) {
drivers/gpu/drm/amd/amdgpu/gmc_v12_1.c
577
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v12_1.c
580
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v12_1.c
585
switch (vm_flags & AMDGPU_VM_MTYPE_MASK) {
drivers/gpu/drm/amd/amdgpu/gmc_v6_0.c
387
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c
509
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c
721
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c
724
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1083
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1199
uint32_t vm_flags,
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1202
if (vm_flags & AMDGPU_VM_PAGE_EXECUTABLE)
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1207
switch (vm_flags & AMDGPU_VM_MTYPE_MASK) {
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1227
if (vm_flags & AMDGPU_VM_PAGE_PRT) {
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1233
gmc_v9_0_get_coherence_flags(adev, vm, bo, vm_flags, flags);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3476
address, vma->vm_flags, PAGE_SIZE);
drivers/gpu/drm/amd/amdkfd/kfd_doorbell.c
137
(unsigned long long) vma->vm_start, address, vma->vm_flags,
drivers/gpu/drm/amd/amdkfd/kfd_events.c
1090
pr_debug(" vm_flags == 0x%08lX\n", vma->vm_flags);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1763
readonly = !(vma->vm_flags & VM_WRITE);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1771
if (!(vma->vm_flags & VM_READ)) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1775
if (vma->vm_flags & VM_WRITE)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3028
vma->vm_flags);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3029
return (vma->vm_flags & requested) == requested;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
3465
if (!vma || (vma->vm_flags & device_vma))
drivers/gpu/drm/drm_gem.c
1212
WARN_ON(!(vma->vm_flags & VM_DONTEXPAND));
drivers/gpu/drm/drm_gem.c
1220
vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/drm_gem_dma_helper.c
540
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/drm_gem_shmem_helper.c
690
if (is_cow_mapping(vma->vm_flags))
drivers/gpu/drm/drm_gem_shmem_helper.c
701
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/drm_gpusvm.c
1063
if (!ctx->read_only && !(vas->vm_flags & VM_WRITE)) {
drivers/gpu/drm/etnaviv/etnaviv_gem.c
135
vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/exynos/exynos_drm_gem.c
380
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/exynos/exynos_drm_gem.c
383
pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/exynos/exynos_drm_gem.c
386
pgprot_noncached(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
1027
vma->vm_page_prot = pgprot_decrypted(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
1038
pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
1046
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
1052
pgprot_noncached(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
1058
pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
115
pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/i915/gem/i915_gem_mman.c
263
area->vm_flags & VM_WRITE))
drivers/gpu/drm/i915/gem/i915_gem_mman.c
285
if (area->vm_flags & VM_WRITE) {
drivers/gpu/drm/i915/gem/i915_gem_mman.c
348
bool write = area->vm_flags & VM_WRITE;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
999
if (vma->vm_flags & VM_WRITE) {
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1050
area->vm_flags & VM_WRITE))
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
414
if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP))
drivers/gpu/drm/i915/gvt/kvmgt.c
1029
if ((vma->vm_flags & VM_SHARED) == 0)
drivers/gpu/drm/i915/i915_mm.c
101
GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS);
drivers/gpu/drm/i915/i915_mm.c
144
GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS);
drivers/gpu/drm/mediatek/mtk_gem.c
201
vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/msm/msm_gem.c
1125
vma->vm_page_prot = msm_gem_pgprot(msm_obj, vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/nouveau/nouveau_gem.c
58
prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/omapdrm/omap_fbdev.c
87
vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/omapdrm/omap_gem.c
540
vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/omapdrm/omap_gem.c
542
vma->vm_page_prot = pgprot_noncached(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/omapdrm/omap_gem.c
560
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/panthor/panthor_device.c
459
if ((vma->vm_flags & VM_SHARED) == 0)
drivers/gpu/drm/panthor/panthor_device.c
465
(vma->vm_flags & (VM_WRITE | VM_EXEC)))
drivers/gpu/drm/rockchip/rockchip_drm_gem.c
258
vma->vm_page_prot = pgprot_writecombine(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/tegra/gem.c
610
pgprot_t prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/ttm/ttm_bo_vm.c
491
if (is_cow_mapping(vma->vm_flags))
drivers/gpu/drm/vc4/vc4_bo.c
736
if (bo->validated_shader && (vma->vm_flags & VM_WRITE)) {
drivers/gpu/drm/virtio/virtgpu_vram.c
51
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
456
num_prefault = (vma->vm_flags & VM_RAND_READ) ? 1 :
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
482
prot = vm_get_page_prot(vma->vm_flags & ~VM_SHARED);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
484
prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/xe/xe_device.c
279
prot = vm_get_page_prot(vma->vm_flags);
drivers/gpu/drm/xe/xe_device.c
328
if (is_cow_mapping(vma->vm_flags))
drivers/gpu/drm/xe/xe_device.c
331
if (vma->vm_flags & (VM_READ | VM_EXEC))
drivers/gpu/drm/xe/xe_mmio_gem.c
147
if ((vma->vm_flags & VM_SHARED) == 0)
drivers/gpu/drm/xe/xe_mmio_gem.c
152
vma->vm_page_prot = pgprot_noncached(vm_get_page_prot(vma->vm_flags));
drivers/gpu/drm/xe/xe_oa.c
1694
if (vma->vm_flags & (VM_WRITE | VM_EXEC | VM_SHARED | VM_MAYSHARE)) {
drivers/gpu/drm/xen/xen_drm_front_gem.c
83
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/hv/mshv_root_main.c
1314
is_mmio = vma ? !!(vma->vm_flags & (VM_IO | VM_PFNMAP)) : 0;
drivers/infiniband/core/ib_core_uverbs.c
170
if (!(vma->vm_flags & VM_SHARED))
drivers/infiniband/core/ib_core_uverbs.c
78
if (!(vma->vm_flags & VM_SHARED))
drivers/infiniband/core/uverbs_main.c
798
if (!(vmf->vma->vm_flags & (VM_WRITE | VM_MAYWRITE))) {
drivers/infiniband/hw/bnxt_re/ib_verbs.c
4626
if (vma->vm_flags & VM_WRITE)
drivers/infiniband/hw/hfi1/file_ops.c
315
vma->vm_end - vma->vm_start, vma->vm_flags);
drivers/infiniband/hw/hfi1/file_ops.c
334
!(vma->vm_flags & VM_SHARED)) {
drivers/infiniband/hw/hfi1/file_ops.c
352
flags = vma->vm_flags;
drivers/infiniband/hw/hfi1/file_ops.c
421
if (vma->vm_flags & VM_WRITE) {
drivers/infiniband/hw/mlx4/mr.c
128
if (vma->vm_flags & VM_WRITE)
drivers/infiniband/hw/mlx5/main.c
2440
!(vma->vm_flags & VM_SHARED))
drivers/infiniband/hw/mlx5/main.c
2446
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
drivers/infiniband/hw/mlx5/main.c
2734
if (vma->vm_flags & VM_WRITE)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
558
if (vma->vm_flags & VM_READ)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
569
if (vma->vm_flags & VM_READ)
drivers/iommu/iommu-sva.c
260
if (access_flags & ~vma->vm_flags)
drivers/iommu/iommufd/main.c
573
if (!(vma->vm_flags & VM_SHARED))
drivers/iommu/iommufd/main.c
575
if (vma->vm_flags & VM_EXEC)
drivers/media/common/videobuf2/videobuf2-core.c
2498
if (!(vma->vm_flags & VM_SHARED)) {
drivers/media/common/videobuf2/videobuf2-core.c
2503
if (!(vma->vm_flags & VM_WRITE)) {
drivers/media/common/videobuf2/videobuf2-core.c
2508
if (!(vma->vm_flags & VM_READ)) {
drivers/misc/ocxl/context.c
180
if ((vma->vm_flags & VM_READ) || (vma->vm_flags & VM_EXEC) ||
drivers/misc/ocxl/context.c
181
!(vma->vm_flags & VM_WRITE))
drivers/misc/open-dice.c
93
if (vma->vm_flags & VM_MAYSHARE) {
drivers/misc/open-dice.c
95
if (vma->vm_flags & VM_WRITE)
drivers/misc/sgi-gru/grufile.c
97
if ((vma->vm_flags & (VM_SHARED | VM_WRITE)) != (VM_SHARED | VM_WRITE))
drivers/mtd/mtdchar.c
1399
return vma->vm_flags & VM_SHARED ? 0 : -EACCES;
drivers/pci/p2pdma.c
104
if ((vma->vm_flags & VM_MAYSHARE) != VM_MAYSHARE) {
drivers/platform/x86/intel/pmt/class.c
121
if (vma->vm_flags & (VM_WRITE | VM_MAYWRITE))
drivers/ptp/ptp_vmclock.c
377
if ((vma->vm_flags & (VM_READ|VM_WRITE)) != VM_READ)
drivers/sbus/char/flash.c
44
if ((vma->vm_flags & VM_READ) &&
drivers/sbus/char/flash.c
45
(vma->vm_flags & VM_WRITE)) {
drivers/sbus/char/flash.c
49
if (vma->vm_flags & VM_READ) {
drivers/sbus/char/flash.c
52
} else if (vma->vm_flags & VM_WRITE) {
drivers/sbus/char/oradax.c
376
dax_dbg("len=0x%lx, flags=0x%lx", len, vma->vm_flags);
drivers/sbus/char/oradax.c
389
if (vma->vm_flags & VM_WRITE)
drivers/usb/mon/mon_bin.c
1281
if (vma->vm_flags & VM_WRITE)
drivers/vfio/cdx/main.c
264
(vma->vm_flags & VM_READ))
drivers/vfio/cdx/main.c
268
(vma->vm_flags & VM_WRITE))
drivers/vfio/fsl-mc/vfio_fsl_mc.c
394
if (!(vma->vm_flags & VM_SHARED))
drivers/vfio/fsl-mc/vfio_fsl_mc.c
403
&& (vma->vm_flags & VM_READ))
drivers/vfio/fsl-mc/vfio_fsl_mc.c
407
&& (vma->vm_flags & VM_WRITE))
drivers/vfio/pci/vfio_pci_core.c
1743
if ((vma->vm_flags & VM_SHARED) == 0)
drivers/vfio/platform/vfio_platform_common.c
578
if (!(vma->vm_flags & VM_SHARED))
drivers/vfio/platform/vfio_platform_common.c
591
&& (vma->vm_flags & VM_READ))
drivers/vfio/platform/vfio_platform_common.c
595
&& (vma->vm_flags & VM_WRITE))
drivers/vfio/vfio_iommu_type1.c
614
if (vma && vma->vm_flags & VM_PFNMAP) {
drivers/vhost/vdpa.c
1063
if (!(vma->vm_file && (vma->vm_flags & VM_SHARED) &&
drivers/vhost/vdpa.c
1064
!(vma->vm_flags & (VM_IO | VM_PFNMAP))))
drivers/vhost/vdpa.c
1513
if ((vma->vm_flags & VM_SHARED) == 0)
drivers/vhost/vdpa.c
1515
if (vma->vm_flags & VM_READ)
drivers/video/fbdev/core/fb_io_fops.c
164
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
drivers/video/fbdev/sbuslib.c
53
if (!(vma->vm_flags & (VM_SHARED | VM_MAYSHARE)))
drivers/virt/acrn/mm.c
178
if (vma && ((vma->vm_flags & VM_PFNMAP) != 0)) {
drivers/xen/gntalloc.c
499
if (!(vma->vm_flags & VM_SHARED)) {
drivers/xen/gntdev.c
1062
if ((vma->vm_flags & VM_WRITE) && !(vma->vm_flags & VM_SHARED))
drivers/xen/gntdev.c
1086
if ((vma->vm_flags & VM_WRITE) &&
drivers/xen/gntdev.c
1091
if (!(vma->vm_flags & VM_WRITE))
drivers/xen/privcmd-buf.c
141
if (!(vma->vm_flags & VM_SHARED))
drivers/xen/xlate_mmu.c
156
BUG_ON(!((vma->vm_flags & (VM_PFNMAP | VM_IO)) == (VM_PFNMAP | VM_IO)));
fs/9p/vfs_file.c
491
if (!(vma->vm_flags & VM_SHARED))
fs/coredump.c
1603
if (vma->vm_flags & VM_DONTDUMP)
fs/coredump.c
1608
if ((vma->vm_flags & VM_SHARED) && FILTER(DAX_SHARED))
fs/coredump.c
1610
if (!(vma->vm_flags & VM_SHARED) && FILTER(DAX_PRIVATE))
fs/coredump.c
1617
if ((vma->vm_flags & VM_SHARED) && FILTER(HUGETLB_SHARED))
fs/coredump.c
1619
if (!(vma->vm_flags & VM_SHARED) && FILTER(HUGETLB_PRIVATE))
fs/coredump.c
1625
if (vma->vm_flags & VM_IO)
fs/coredump.c
1629
if (vma->vm_flags & VM_SHARED) {
fs/coredump.c
1650
vma->vm_pgoff == 0 && (vma->vm_flags & VM_READ)) {
fs/coredump.c
1750
m->flags = vma->vm_flags;
fs/cramfs/inode.c
375
if (vma->vm_flags & VM_WRITE)
fs/cramfs/inode.c
449
return is_nommu_shared_mapping(vma->vm_flags) ? 0 : -ENOSYS;
fs/dax.c
1031
return (iter->flags & IOMAP_WRITE) && (vma->vm_flags & VM_SYNC) &&
fs/dax.c
1956
if (write && !(vmf->vma->vm_flags & VM_SHARED))
fs/exec.c
607
vm_flags_t vm_flags;
fs/exec.c
653
vm_flags = VM_STACK_FLAGS;
fs/exec.c
661
vm_flags |= VM_EXEC;
fs/exec.c
663
vm_flags &= ~VM_EXEC;
fs/exec.c
664
vm_flags |= mm->def_flags;
fs/exec.c
665
vm_flags |= VM_STACK_INCOMPLETE_SETUP;
fs/exec.c
671
vm_flags);
fs/exec.c
678
if (unlikely(vm_flags & VM_EXEC)) {
fs/ext2/file.c
99
(vmf->vma->vm_flags & VM_SHARED);
fs/ext4/file.c
744
(vmf->vma->vm_flags & VM_SHARED);
fs/f2fs/file.c
189
trace_f2fs_vm_page_mkwrite(inode, folio->index, vmf->vma->vm_flags, ret);
fs/f2fs/file.c
62
vm_flags_t flags = vmf->vma->vm_flags;
fs/fuse/file.c
2398
if ((vma->vm_flags & VM_MAYSHARE) && !fc->direct_io_allow_mmap)
fs/fuse/file.c
2403
if (!(vma->vm_flags & VM_MAYSHARE)) {
fs/fuse/file.c
2420
if ((vma->vm_flags & VM_SHARED) && (vma->vm_flags & VM_MAYWRITE))
fs/proc/nommu.c
38
flags = region->vm_flags;
fs/proc/task_mmu.c
1023
bool locked = !!(vma->vm_flags & VM_LOCKED);
fs/proc/task_mmu.c
1070
bool locked = !!(vma->vm_flags & VM_LOCKED);
fs/proc/task_mmu.c
1225
if (vma->vm_flags & (1UL << i))
fs/proc/task_mmu.c
1311
if (!start && (!shmem_swapped || (vma->vm_flags & VM_SHARED) ||
fs/proc/task_mmu.c
1312
!(vma->vm_flags & VM_WRITE))) {
fs/proc/task_mmu.c
1387
!!thp_vma_allowable_orders(vma, vma->vm_flags, TVA_SMAPS,
fs/proc/task_mmu.c
1606
if (!is_cow_mapping(vma->vm_flags))
fs/proc/task_mmu.c
1746
if (vma->vm_flags & VM_PFNMAP)
fs/proc/task_mmu.c
1816
if (!(vma->vm_flags & VM_SOFTDIRTY))
fs/proc/task_mmu.c
1922
if (vma->vm_flags & VM_SOFTDIRTY)
fs/proc/task_mmu.c
1994
if (vma->vm_flags & VM_SOFTDIRTY)
fs/proc/task_mmu.c
2012
if (vma->vm_flags & VM_SOFTDIRTY)
fs/proc/task_mmu.c
2134
if (vma->vm_flags & VM_SOFTDIRTY)
fs/proc/task_mmu.c
2602
if (vma->vm_flags & VM_PFNMAP)
fs/proc/task_mmu.c
2608
if (vma->vm_flags & VM_SOFTDIRTY)
fs/proc/task_mmu.c
467
vm_flags_t flags = vma->vm_flags;
fs/proc/task_mmu.c
634
if ((vma->vm_flags & perm) != perm)
fs/proc/task_mmu.c
706
if (vma->vm_flags & VM_READ)
fs/proc/task_mmu.c
708
if (vma->vm_flags & VM_WRITE)
fs/proc/task_mmu.c
710
if (vma->vm_flags & VM_EXEC)
fs/proc/task_mmu.c
712
if (vma->vm_flags & VM_MAYSHARE)
fs/proc/task_nommu.c
136
flags = vma->vm_flags;
fs/proc/task_nommu.c
41
is_nommu_shared_mapping(vma->vm_flags)) {
fs/proc/vmcore.c
603
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
fs/userfaultfd.c
1269
vm_flags_t vm_flags;
fs/userfaultfd.c
1288
vm_flags = 0;
fs/userfaultfd.c
1290
vm_flags |= VM_UFFD_MISSING;
fs/userfaultfd.c
1295
vm_flags |= VM_UFFD_WP;
fs/userfaultfd.c
1301
vm_flags |= VM_UFFD_MINOR;
fs/userfaultfd.c
1344
!!(cur->vm_flags & __VM_UFFD_FLAGS));
fs/userfaultfd.c
1348
if (!vma_can_userfault(cur, vm_flags, wp_async))
fs/userfaultfd.c
1360
if (unlikely(!(cur->vm_flags & VM_MAYWRITE)))
fs/userfaultfd.c
1376
if ((vm_flags & VM_UFFD_WP) && !(cur->vm_flags & VM_MAYWRITE))
fs/userfaultfd.c
1400
ret = userfaultfd_register_range(ctx, vma, vm_flags, start, end,
fs/userfaultfd.c
1491
!!(cur->vm_flags & __VM_UFFD_FLAGS));
fs/userfaultfd.c
1508
if (!vma_can_userfault(cur, cur->vm_flags, wp_async))
fs/userfaultfd.c
1529
VM_WARN_ON_ONCE(!vma_can_userfault(vma, vma->vm_flags, wp_async));
fs/userfaultfd.c
1530
VM_WARN_ON_ONCE(!(vma->vm_flags & VM_MAYWRITE));
fs/xfs/xfs_file.c
1940
(vmf->vma->vm_flags & VM_SHARED);
include/asm-generic/tlb.h
460
.vm_flags = (tlb->vma_exec ? VM_EXEC : 0) |
include/asm-generic/tlb.h
486
tlb->vma_exec = !!(vma->vm_flags & VM_EXEC);
include/asm-generic/tlb.h
492
tlb->vma_pfn |= !!(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP));
include/linux/huge_mm.h
104
#define thp_vma_allowable_order(vma, vm_flags, type, order) \
include/linux/huge_mm.h
105
(!!thp_vma_allowable_orders(vma, vm_flags, type, BIT(order)))
include/linux/huge_mm.h
269
vm_flags_t vm_flags,
include/linux/huge_mm.h
290
vm_flags_t vm_flags,
include/linux/huge_mm.h
301
if (vm_flags & VM_HUGEPAGE)
include/linux/huge_mm.h
304
((vm_flags & VM_HUGEPAGE) && hugepage_global_enabled()))
include/linux/huge_mm.h
312
return __thp_vma_allowable_orders(vma, vm_flags, type, orders);
include/linux/huge_mm.h
332
vm_flags_t vm_flags, bool forced_collapse)
include/linux/huge_mm.h
335
if (vm_flags & VM_NOHUGEPAGE)
include/linux/huge_mm.h
344
if (vm_flags & VM_HUGEPAGE)
include/linux/huge_mm.h
365
vm_flags_t vm_flags);
include/linux/huge_mm.h
483
int hugepage_madvise(struct vm_area_struct *vma, vm_flags_t *vm_flags,
include/linux/huge_mm.h
591
vm_flags_t vm_flags,
include/linux/huge_mm.h
605
unsigned long flags, vm_flags_t vm_flags)
include/linux/huge_mm.h
676
vm_flags_t *vm_flags, int advice)
include/linux/hugetlb.h
1350
return (vma->vm_flags & VM_MAYSHARE) && vma->vm_private_data;
include/linux/hugetlb_inline.h
11
return !!(vm_flags & VM_HUGETLB);
include/linux/hugetlb_inline.h
21
static inline bool is_vm_hugetlb_flags(vm_flags_t vm_flags)
include/linux/hugetlb_inline.h
35
return is_vm_hugetlb_flags(vma->vm_flags);
include/linux/hugetlb_inline.h
9
static inline bool is_vm_hugetlb_flags(vm_flags_t vm_flags)
include/linux/khugepaged.h
17
vm_flags_t vm_flags);
include/linux/khugepaged.h
42
vm_flags_t vm_flags)
include/linux/ksm.h
107
const struct file *file, vm_flags_t vm_flags)
include/linux/ksm.h
109
return vm_flags;
include/linux/ksm.h
142
unsigned long end, int advice, vm_flags_t *vm_flags)
include/linux/ksm.h
19
unsigned long end, int advice, vm_flags_t *vm_flags);
include/linux/ksm.h
21
vm_flags_t vm_flags);
include/linux/mm.h
1028
return test_bit((__force int)bit, &vma->vm_flags);
include/linux/mm.h
1267
int maybe_stack = vma->vm_flags & (VM_GROWSDOWN | VM_GROWSUP);
include/linux/mm.h
1272
if ((vma->vm_flags & VM_STACK_INCOMPLETE_SETUP) ==
include/linux/mm.h
1292
return vma->vm_flags & VM_ACCESS_FLAGS;
include/linux/mm.h
1295
static inline bool is_shared_maywrite_vm_flags(vm_flags_t vm_flags)
include/linux/mm.h
1297
return (vm_flags & (VM_SHARED | VM_MAYWRITE)) ==
include/linux/mm.h
1400
#define TLB_FLUSH_VMA(mm,flags) { .vm_mm = (mm), .vm_flags = (flags) }
include/linux/mm.h
1692
if (likely(vma->vm_flags & VM_WRITE))
include/linux/mm.h
3857
vm_flags_t vm_flags,
include/linux/mm.h
3865
unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags);
include/linux/mm.h
3876
vm_flags_t vm_flags, unsigned long pgoff, unsigned long *populate,
include/linux/mm.h
3964
if (vma->vm_flags & VM_GROWSDOWN)
include/linux/mm.h
3968
if (vma->vm_flags & VM_SHADOW_STACK)
include/linux/mm.h
3989
if (vma->vm_flags & VM_GROWSUP) {
include/linux/mm.h
4105
pgprot_t vm_get_page_prot(vm_flags_t vm_flags);
include/linux/mm.h
4108
static inline pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
include/linux/mm.h
4114
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
include/linux/mm.h
4713
(vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP)));
include/linux/mm.h
980
vm_flags_init(vma, (vma->vm_flags | set) & ~clear);
include/linux/mm_inline.h
612
if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ))
include/linux/mm_types.h
706
vm_flags_t vm_flags; /* VMA vm_flags */
include/linux/mm_types.h
939
const vm_flags_t vm_flags;
include/linux/pgtable.h
2365
pgprot_t vm_get_page_prot(vm_flags_t vm_flags) \
include/linux/pgtable.h
2367
return protection_map[vm_flags & \
include/linux/rmap.h
846
struct mem_cgroup *memcg, vm_flags_t *vm_flags);
include/linux/rmap.h
979
vm_flags_t *vm_flags)
include/linux/rmap.h
981
*vm_flags = 0;
include/linux/sched/mm.h
186
unsigned long flags, vm_flags_t vm_flags);
include/linux/sched/mm.h
201
vm_flags_t vm_flags);
include/linux/sched/mm.h
206
unsigned long flags, vm_flags_t vm_flags);
include/linux/sched/mm.h
210
unsigned long flags, vm_flags_t vm_flags);
include/linux/userfaultfd_k.h
164
return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR);
include/linux/userfaultfd_k.h
176
return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR);
include/linux/userfaultfd_k.h
181
return vma->vm_flags & VM_UFFD_MISSING;
include/linux/userfaultfd_k.h
186
return vma->vm_flags & VM_UFFD_WP;
include/linux/userfaultfd_k.h
191
return vma->vm_flags & VM_UFFD_MINOR;
include/linux/userfaultfd_k.h
208
return vma->vm_flags & __VM_UFFD_FLAGS;
include/linux/userfaultfd_k.h
212
vm_flags_t vm_flags,
include/linux/userfaultfd_k.h
215
vm_flags &= __VM_UFFD_FLAGS;
include/linux/userfaultfd_k.h
217
if (vma->vm_flags & VM_DROPPABLE)
include/linux/userfaultfd_k.h
220
if ((vm_flags & VM_UFFD_MINOR) &&
include/linux/userfaultfd_k.h
228
if (wp_async && (vm_flags == VM_UFFD_WP))
include/linux/userfaultfd_k.h
236
if (!uffd_supports_wp_marker() && (vm_flags & VM_UFFD_WP) &&
include/linux/userfaultfd_k.h
284
vm_flags_t vm_flags,
include/linux/vmalloc.h
176
pgprot_t prot, unsigned long vm_flags, int node,
include/trace/events/fs_dax.h
110
__field(vm_flags_t, vm_flags)
include/trace/events/fs_dax.h
120
__entry->vm_flags = vmf->vma->vm_flags;
include/trace/events/fs_dax.h
130
__entry->vm_flags & VM_SHARED ? "shared" : "private",
include/trace/events/fs_dax.h
18
__field(vm_flags_t, vm_flags)
include/trace/events/fs_dax.h
31
__entry->vm_flags = vmf->vma->vm_flags;
include/trace/events/fs_dax.h
43
__entry->vm_flags & VM_SHARED ? "shared" : "private",
include/trace/events/fs_dax.h
70
__field(vm_flags_t, vm_flags)
include/trace/events/fs_dax.h
79
__entry->vm_flags = vmf->vma->vm_flags;
include/trace/events/fs_dax.h
89
__entry->vm_flags & VM_SHARED ? "shared" : "private",
io_uring/memmap.c
371
return is_nommu_shared_mapping(vma->vm_flags) ? 0 : -EINVAL;
kernel/bpf/ringbuf.c
281
if (vma->vm_flags & VM_WRITE) {
kernel/bpf/ringbuf.c
297
if (vma->vm_flags & VM_WRITE) {
kernel/bpf/syscall.c
1069
if (vma->vm_flags & VM_MAYWRITE)
kernel/bpf/syscall.c
1078
if (vma->vm_flags & VM_MAYWRITE)
kernel/bpf/syscall.c
1095
if (!(vma->vm_flags & VM_SHARED))
kernel/bpf/syscall.c
1100
if (vma->vm_flags & VM_WRITE) {
kernel/bpf/syscall.c
1132
if (!(vma->vm_flags & VM_WRITE))
kernel/bpf/syscall.c
1137
if (vma->vm_flags & VM_WRITE)
kernel/bpf/sysfs_btf.c
33
if (vma->vm_flags & (VM_WRITE | VM_EXEC | VM_MAYSHARE))
kernel/events/core.c
7208
vm_get_page_prot(vma->vm_flags & ~VM_SHARED));
kernel/events/core.c
7317
if (vma->vm_flags & VM_WRITE)
kernel/events/core.c
7399
if (vma->vm_flags & VM_WRITE)
kernel/events/core.c
7434
if (!(vma->vm_flags & VM_SHARED))
kernel/events/core.c
9611
int executable = vma->vm_flags & VM_EXEC;
kernel/events/core.c
9696
if (vma->vm_flags & VM_READ)
kernel/events/core.c
9698
if (vma->vm_flags & VM_WRITE)
kernel/events/core.c
9700
if (vma->vm_flags & VM_EXEC)
kernel/events/core.c
9703
if (vma->vm_flags & VM_MAYSHARE)
kernel/events/core.c
9708
if (vma->vm_flags & VM_LOCKED)
kernel/events/core.c
9777
if (!(vma->vm_flags & VM_EXEC))
kernel/events/core.c
9879
if (!(vma->vm_flags & VM_EXEC))
kernel/events/uprobes.c
142
return vma->vm_file && (vma->vm_flags & flags) == VM_MAYEXEC;
kernel/events/uprobes.c
1607
(vma->vm_flags & (VM_WRITE|VM_SHARED)) == VM_WRITE &&
kernel/events/uprobes.c
287
(vma->vm_flags & (VM_WRITE|VM_SHARED)) == VM_WRITE &&
kernel/events/uprobes.c
514
if (WARN_ON_ONCE(!is_cow_mapping(vma->vm_flags)))
kernel/liveupdate/kexec_handover.c
1080
unsigned int align, order, shift, vm_flags;
kernel/liveupdate/kexec_handover.c
1088
vm_flags = kho_flags_to_vmalloc(preservation->flags);
kernel/liveupdate/kexec_handover.c
1089
if (vm_flags & ~KHO_VMALLOC_SUPPORTED_FLAGS)
kernel/liveupdate/kexec_handover.c
1131
vm_flags, VMALLOC_START, VMALLOC_END,
kernel/liveupdate/kexec_handover.c
918
static unsigned short vmalloc_flags_to_kho(unsigned int vm_flags)
kernel/liveupdate/kexec_handover.c
922
if (vm_flags & VM_ALLOC)
kernel/liveupdate/kexec_handover.c
924
if (vm_flags & VM_ALLOW_HUGE_VMAP)
kernel/liveupdate/kexec_handover.c
932
unsigned int vm_flags = 0;
kernel/liveupdate/kexec_handover.c
935
vm_flags |= VM_ALLOC;
kernel/liveupdate/kexec_handover.c
937
vm_flags |= VM_ALLOW_HUGE_VMAP;
kernel/liveupdate/kexec_handover.c
939
return vm_flags;
kernel/sched/fair.c
3457
is_vm_hugetlb_page(vma) || (vma->vm_flags & VM_MIXEDMAP)) {
kernel/sched/fair.c
3469
(vma->vm_file && (vma->vm_flags & (VM_READ|VM_WRITE)) == (VM_READ))) {
kernel/trace/ring_buffer.c
7161
if (vma->vm_flags & VM_WRITE || vma->vm_flags & VM_EXEC ||
kernel/trace/ring_buffer.c
7162
!(vma->vm_flags & VM_MAYSHARE))
lib/test_hmm.c
1068
if (!vma || !(vma->vm_flags & VM_READ)) {
lib/test_hmm.c
1142
if (!vma || !(vma->vm_flags & VM_READ)) {
mm/debug.c
171
vma->vm_flags, &vma->vm_flags);
mm/debug.c
259
vmg->start, vmg->end, vmg->vm_flags,
mm/execmem.c
283
unsigned long vm_flags = VM_ALLOW_HUGE_VMAP;
mm/execmem.c
29
pgprot_t pgprot, unsigned long vm_flags)
mm/execmem.c
290
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
mm/execmem.c
293
p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags);
mm/execmem.c
39
vm_flags |= VM_DEFER_KMEMLEAK;
mm/execmem.c
42
pgprot, vm_flags, NUMA_NO_NODE,
mm/execmem.c
465
unsigned long vm_flags = VM_FLUSH_RESET_PERMS;
mm/execmem.c
474
p = execmem_vmalloc(range, size, pgprot, vm_flags);
mm/execmem.c
48
pgprot, vm_flags, NUMA_NO_NODE,
mm/execmem.c
82
pgprot_t pgprot, unsigned long vm_flags)
mm/filemap.c
3312
vm_flags_t vm_flags = vmf->vma->vm_flags;
mm/filemap.c
3318
(vm_flags & VM_HUGEPAGE) && HPAGE_PMD_ORDER <= MAX_PAGECACHE_ORDER)
mm/filemap.c
3326
if ((vm_flags & (VM_RAND_READ | VM_EXEC)) == VM_RAND_READ)
mm/filemap.c
3332
if (vm_flags & VM_SEQ_READ) {
mm/filemap.c
3339
if (!(vm_flags & VM_SEQ_READ)) {
mm/filemap.c
3361
if (!(vm_flags & VM_RAND_READ))
mm/filemap.c
3369
if (vm_flags & VM_EXEC) {
mm/filemap.c
3424
if (vmf->vma->vm_flags & VM_RAND_READ || !ra->ra_pages)
mm/filemap.c
3466
if (!(vma->vm_flags & VM_LOCKED))
mm/gup.c
1202
vm_flags_t vm_flags = vma->vm_flags;
mm/gup.c
1207
if (vm_flags & (VM_IO | VM_PFNMAP))
mm/gup.c
1227
if (!(vm_flags & VM_WRITE) || (vm_flags & VM_SHADOW_STACK)) {
mm/gup.c
1239
if (!is_cow_mapping(vm_flags))
mm/gup.c
1242
} else if (!(vm_flags & VM_READ)) {
mm/gup.c
1249
if (!(vm_flags & VM_MAYREAD))
mm/gup.c
1280
if (!(vma->vm_flags & VM_GROWSDOWN))
mm/gup.c
1517
vm_flags_t vm_flags = write ? VM_WRITE : VM_READ;
mm/gup.c
1519
if (!(vm_flags & vma->vm_flags))
mm/gup.c
1832
if (vma->vm_flags & VM_LOCKONFAULT)
mm/gup.c
1848
if ((vma->vm_flags & (VM_WRITE | VM_SHARED)) == VM_WRITE)
mm/gup.c
1954
if (vma->vm_flags & (VM_IO | VM_PFNMAP))
mm/gup.c
1985
vm_flags_t vm_flags;
mm/gup.c
2005
vm_flags = (foll_flags & FOLL_WRITE) ?
mm/gup.c
2007
vm_flags &= (foll_flags & FOLL_FORCE) ?
mm/gup.c
2016
if ((vma->vm_flags & (VM_IO | VM_PFNMAP)) ||
mm/gup.c
2017
!(vm_flags & vma->vm_flags))
mm/gup.c
593
if (vma->vm_flags & (VM_MAYSHARE | VM_SHARED))
mm/gup.c
597
if (!(vma->vm_flags & VM_MAYWRITE))
mm/gup.c
601
if (vma->vm_flags & VM_WRITE)
mm/hmm.c
604
if (!(vma->vm_flags & (VM_IO | VM_PFNMAP)) &&
mm/hmm.c
605
vma->vm_flags & VM_READ)
mm/hmm.c
84
if (!(vma->vm_flags & VM_WRITE))
mm/huge_memory.c
104
vm_flags_t vm_flags,
mm/huge_memory.c
1078
if (likely(vma->vm_flags & VM_WRITE))
mm/huge_memory.c
1192
vm_flags_t vm_flags)
mm/huge_memory.c
1209
off >> PAGE_SHIFT, flags, vm_flags);
mm/huge_memory.c
1236
vm_flags_t vm_flags)
mm/huge_memory.c
1241
ret = __thp_get_unmapped_area(filp, addr, len, off, flags, PMD_SIZE, vm_flags);
mm/huge_memory.c
1246
vm_flags);
mm/huge_memory.c
128
if (thp_disabled_by_hw() || vma_thp_disabled(vma, vm_flags, forced_collapse))
mm/huge_memory.c
140
if (!in_pf && !smaps && (vm_flags & VM_NO_KHUGEPAGED))
mm/huge_memory.c
1424
const bool vma_madvised = vma && (vma->vm_flags & VM_HUGEPAGE);
mm/huge_memory.c
1472
khugepaged_enter_vma(vma, vma->vm_flags);
mm/huge_memory.c
1622
BUG_ON(!(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)));
mm/huge_memory.c
1623
BUG_ON((vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) ==
mm/huge_memory.c
1625
BUG_ON((vma->vm_flags & VM_PFNMAP) && is_cow_mapping(vma->vm_flags));
mm/huge_memory.c
1653
if (likely(vma->vm_flags & VM_WRITE))
mm/huge_memory.c
1730
BUG_ON(!(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)));
mm/huge_memory.c
1731
BUG_ON((vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) ==
mm/huge_memory.c
1733
BUG_ON((vma->vm_flags & VM_PFNMAP) && is_cow_mapping(vma->vm_flags));
mm/huge_memory.c
181
(!hugepage_global_enabled() || (!(vm_flags & VM_HUGEPAGE) &&
mm/huge_memory.c
1875
VM_WARN_ON_ONCE(is_cow_mapping(src_vma->vm_flags) && pmd_write(pmd));
mm/huge_memory.c
1990
if (is_cow_mapping(vma->vm_flags) && pud_write(pud)) {
mm/huge_memory.c
2159
if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE)))
mm/huge_memory.c
2174
if (!(vma->vm_flags & VM_SHARED)) {
mm/huge_memory.c
3365
if (folio_test_dirty(folio) && !(vma->vm_flags & VM_DROPPABLE)) {
mm/huge_memory.c
3397
if (folio_test_dirty(folio) && !(vma->vm_flags & VM_DROPPABLE)) {
mm/huge_memory.c
3411
if (vma->vm_flags & VM_LOCKED)
mm/huge_memory.c
4579
return vma_is_special_huge(vma) || (vma->vm_flags & VM_IO) ||
mm/hugetlb.c
1174
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
1189
VM_WARN_ON_ONCE_VMA(vma->vm_flags & VM_MAYSHARE, vma);
mm/hugetlb.c
1226
return !(vma->vm_flags & VM_MAYSHARE) &&
mm/hugetlb.c
1246
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
2516
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
2526
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
2539
if (vma->vm_flags & VM_MAYSHARE || mode == VMA_DEL_RESV)
mm/hugetlb.c
2668
if (!(vma->vm_flags & VM_MAYSHARE))
mm/hugetlb.c
425
if (!vma || !(vma->vm_flags & VM_MAYSHARE))
mm/hugetlb.c
4717
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
4842
if (try_mkwrite && (vma->vm_flags & VM_WRITE)) {
mm/hugetlb.c
4848
entry = arch_make_huge_pte(entry, shift, vma->vm_flags);
mm/hugetlb.c
4866
if (vma->vm_flags & VM_WRITE)
mm/hugetlb.c
4892
bool cow = is_cow_mapping(src_vma->vm_flags);
mm/hugetlb.c
5426
if (iter_vma->vm_flags & VM_MAYSHARE)
mm/hugetlb.c
5475
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
5786
if (!(vma->vm_flags & VM_MAYSHARE)) {
mm/hugetlb.c
5816
if (vma->vm_flags & VM_MAYSHARE) {
mm/hugetlb.c
5869
if ((vmf->flags & FAULT_FLAG_WRITE) && !(vma->vm_flags & VM_SHARED)) {
mm/hugetlb.c
5888
new_pte = make_huge_pte(vma, folio, vma->vm_flags & VM_SHARED);
mm/hugetlb.c
5898
if ((vmf->flags & FAULT_FLAG_WRITE) && !(vma->vm_flags & VM_SHARED)) {
mm/hugetlb.c
6076
!(vma->vm_flags & VM_MAYSHARE) && !huge_pte_write(vmf.orig_pte)) {
mm/hugetlb.c
6208
int vm_shared = dst_vma->vm_flags & VM_SHARED;
mm/hugetlb.c
6530
pte = arch_make_huge_pte(pte, shift, vma->vm_flags);
mm/hugetlb.c
6803
vm_flags_t vm_flags = vma->vm_flags & ~VM_LOCKED_MASK;
mm/hugetlb.c
6804
vm_flags_t svm_flags = svma->vm_flags & ~VM_LOCKED_MASK;
mm/hugetlb.c
6813
vm_flags != svm_flags ||
mm/hugetlb.c
6833
if (!(vma->vm_flags & VM_MAYSHARE))
mm/hugetlb.c
6857
if (!(vma->vm_flags & VM_MAYSHARE) || !(v_end > v_start) ||
mm/hugetlb.c
7274
if (!(vma->vm_flags & VM_MAYSHARE))
mm/internal.h
1122
if (unlikely((vma->vm_flags & (VM_LOCKED|VM_SPECIAL)) == VM_LOCKED))
mm/internal.h
1139
if (unlikely(vma->vm_flags & VM_LOCKED))
mm/internal.h
1463
if (vma->vm_flags & VM_SHARED)
mm/internal.h
1490
unsigned long vm_flags, unsigned long start,
mm/internal.h
1602
return is_cow_mapping(vma->vm_flags);
mm/internal.h
1644
return !(vma->vm_flags & VM_SOFTDIRTY);
mm/khugepaged.c
1499
if (!thp_vma_allowable_order(vma, vma->vm_flags, TVA_FORCED_COLLAPSE, PMD_ORDER))
mm/khugepaged.c
1576
if (userfaultfd_armed(vma) && !(vma->vm_flags & VM_SHARED))
mm/khugepaged.c
2435
if (!thp_vma_allowable_order(vma, vma->vm_flags, TVA_KHUGEPAGED, PMD_ORDER)) {
mm/khugepaged.c
2769
if (!thp_vma_allowable_order(vma, vma->vm_flags, TVA_FORCED_COLLAPSE, PMD_ORDER))
mm/khugepaged.c
344
vm_flags_t *vm_flags, int advice)
mm/khugepaged.c
348
*vm_flags &= ~VM_NOHUGEPAGE;
mm/khugepaged.c
349
*vm_flags |= VM_HUGEPAGE;
mm/khugepaged.c
355
khugepaged_enter_vma(vma, *vm_flags);
mm/khugepaged.c
358
*vm_flags &= ~VM_HUGEPAGE;
mm/khugepaged.c
359
*vm_flags |= VM_NOHUGEPAGE;
mm/khugepaged.c
455
vm_flags_t vm_flags)
mm/khugepaged.c
459
if (thp_vma_allowable_order(vma, vm_flags, TVA_KHUGEPAGED, PMD_ORDER))
mm/khugepaged.c
907
if (!thp_vma_allowable_order(vma, vma->vm_flags, type, PMD_ORDER))
mm/ksm.c
1221
if (!(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma)
mm/ksm.c
2653
if (!(vma->vm_flags & VM_MERGEABLE))
mm/ksm.c
2828
static bool __ksm_should_add_vma(const struct file *file, vm_flags_t vm_flags)
mm/ksm.c
2830
if (vm_flags & VM_MERGEABLE)
mm/ksm.c
2833
return ksm_compatible(file, vm_flags);
mm/ksm.c
2838
if (__ksm_should_add_vma(vma->vm_file, vma->vm_flags))
mm/ksm.c
2846
if (!(vma->vm_flags & VM_MERGEABLE))
mm/ksm.c
2868
vm_flags_t vm_flags)
mm/ksm.c
2871
__ksm_should_add_vma(file, vm_flags)) {
mm/ksm.c
2872
vm_flags |= VM_MERGEABLE;
mm/ksm.c
2882
return vm_flags;
mm/ksm.c
2976
unsigned long end, int advice, vm_flags_t *vm_flags)
mm/ksm.c
2983
if (vma->vm_flags & VM_MERGEABLE)
mm/ksm.c
2994
*vm_flags |= VM_MERGEABLE;
mm/ksm.c
2998
if (!(*vm_flags & VM_MERGEABLE))
mm/ksm.c
3007
*vm_flags &= ~VM_MERGEABLE;
mm/ksm.c
3434
if (vma->vm_flags & VM_MERGEABLE)
mm/ksm.c
738
static bool ksm_compatible(const struct file *file, vm_flags_t vm_flags)
mm/ksm.c
740
if (vm_flags & (VM_SHARED | VM_MAYSHARE | VM_SPECIAL |
mm/ksm.c
748
if (vm_flags & VM_SAO)
mm/ksm.c
752
if (vm_flags & VM_SPARC_ADI)
mm/ksm.c
761
return ksm_compatible(vma->vm_file, vma->vm_flags);
mm/ksm.c
771
if (!vma || !(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma)
mm/madvise.c
1018
if (vma->vm_flags & VM_LOCKED)
mm/madvise.c
1064
return !(vma->vm_flags & disallowed);
mm/madvise.c
1325
if ((vma->vm_flags & VM_WRITE) &&
mm/madvise.c
1349
vm_flags_t new_flags = vma->vm_flags;
mm/madvise.c
159
if (new_flags == vma->vm_flags && (!set_new_anon_name ||
mm/madvise.c
591
return !(vma->vm_flags & (VM_LOCKED|VM_PFNMAP|VM_HUGETLB));
mm/madvise.c
640
(vma->vm_flags & VM_MAYSHARE)))
mm/madvise.c
883
return !(vma->vm_flags & forbidden);
mm/mapping_dirty_helpers.c
221
vm_flags_t vm_flags = READ_ONCE(walk->vma->vm_flags);
mm/mapping_dirty_helpers.c
224
if ((vm_flags & (VM_SHARED | VM_MAYWRITE | VM_HUGETLB)) !=
mm/memfd.c
374
vm_flags_t vm_flags = *vm_flags_ptr;
mm/memfd.c
375
vm_flags_t mask = vm_flags & (VM_SHARED | VM_WRITE);
mm/memory.c
1014
is_cow_mapping(vm_flags)) {
mm/memory.c
1029
VM_BUG_ON(!is_cow_mapping(src_vma->vm_flags));
mm/memory.c
1102
if (is_cow_mapping(src_vma->vm_flags) && pte_write(pte)) {
mm/memory.c
1108
if (src_vma->vm_flags & VM_SHARED)
mm/memory.c
1147
if (!(src_vma->vm_flags & VM_SHARED))
mm/memory.c
1485
if (dst_vma->vm_flags & VM_COPY_ON_FORK)
mm/memory.c
1528
is_cow = is_cow_mapping(src_vma->vm_flags);
mm/memory.c
2254
!(vma->vm_flags & VM_PFNMAP))
mm/memory.c
2295
VM_WARN_ON_ONCE(vma->vm_flags & VM_PFNMAP);
mm/memory.c
2305
if (is_cow_mapping(vma->vm_flags))
mm/memory.c
2308
if (!(vma->vm_flags & (VM_WRITE | VM_MAYWRITE)))
mm/memory.c
2320
(vma_is_fsdax(vma) || vma->vm_flags & VM_IO);
mm/memory.c
2497
if (!(vma->vm_flags & VM_MIXEDMAP)) {
mm/memory.c
2499
BUG_ON(vma->vm_flags & VM_PFNMAP);
mm/memory.c
2542
if (!(vma->vm_flags & VM_MIXEDMAP)) {
mm/memory.c
2544
BUG_ON(vma->vm_flags & VM_PFNMAP);
mm/memory.c
2719
BUG_ON(!(vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)));
mm/memory.c
2720
BUG_ON((vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) ==
mm/memory.c
2722
BUG_ON((vma->vm_flags & VM_PFNMAP) && is_cow_mapping(vma->vm_flags));
mm/memory.c
2723
BUG_ON((vma->vm_flags & VM_MIXEDMAP) && pfn_valid(pfn));
mm/memory.c
2771
if (vma->vm_flags & VM_MIXEDMAP)
mm/memory.c
3126
err = get_remap_pgoff(is_cow_mapping(vma->vm_flags), addr, end,
mm/memory.c
3928
WARN_ON_ONCE(!(vmf->vma->vm_flags & VM_SHARED));
mm/memory.c
4197
if (vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) {
mm/memory.c
4439
if (mem_cgroup_swap_full(folio) || (vma->vm_flags & VM_LOCKED) ||
mm/memory.c
4633
orders = thp_vma_allowable_orders(vma, vma->vm_flags, TVA_PAGEFAULT,
mm/memory.c
5018
if ((vma->vm_flags & VM_WRITE) && !userfaultfd_pte_wp(vma, pte) &&
mm/memory.c
5150
orders = thp_vma_allowable_orders(vma, vma->vm_flags, TVA_PAGEFAULT,
mm/memory.c
5227
if (vma->vm_flags & VM_SHARED)
mm/memory.c
5284
if (vma->vm_flags & VM_WRITE)
mm/memory.c
5422
if (thp_disabled_by_hw() || vma_thp_disabled(vma, vma->vm_flags,
mm/memory.c
5520
if (write && !(vma->vm_flags & VM_SHARED)) {
mm/memory.c
5563
!(vma->vm_flags & VM_SHARED);
mm/memory.c
5582
if (!(vma->vm_flags & VM_SHARED)) {
mm/memory.c
5934
else if (!(vma->vm_flags & VM_SHARED))
mm/memory.c
5968
if (folio_maybe_mapped_shared(folio) && (vma->vm_flags & VM_SHARED))
mm/memory.c
6166
if (vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) {
mm/memory.c
618
(void *)addr, vma->vm_flags, vma->anon_vma, mapping, index);
mm/memory.c
6205
if (vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) {
mm/memory.c
6367
vm_flags_t vm_flags = vma->vm_flags;
mm/memory.c
6382
thp_vma_allowable_order(vma, vm_flags, TVA_PAGEFAULT, PUD_ORDER)) {
mm/memory.c
6416
thp_vma_allowable_order(vma, vm_flags, TVA_PAGEFAULT, PMD_ORDER)) {
mm/memory.c
6557
if (!is_cow_mapping(vma->vm_flags))
mm/memory.c
6561
if (WARN_ON_ONCE(!(vma->vm_flags & VM_MAYWRITE)))
mm/memory.c
6564
if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE) &&
mm/memory.c
6565
!is_cow_mapping(vma->vm_flags)))
mm/memory.c
6610
is_droppable = !!(vma->vm_flags & VM_DROPPABLE);
mm/memory.c
6804
if (!(vma->vm_flags & (VM_IO | VM_PFNMAP)))
mm/memory.c
706
if (vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP))
mm/memory.c
719
if (unlikely(vma->vm_flags & (VM_PFNMAP | VM_MIXEDMAP))) {
mm/memory.c
720
if (vma->vm_flags & VM_MIXEDMAP) {
mm/memory.c
730
if (!is_cow_mapping(vma->vm_flags))
mm/memory.c
896
if ((vma->vm_flags & VM_WRITE) &&
mm/memory.c
942
vm_flags_t vm_flags = dst_vma->vm_flags;
mm/memory.c
973
is_cow_mapping(vm_flags)) {
mm/mempolicy.c
1992
if (vma->vm_flags & (VM_IO | VM_PFNMAP))
mm/mempolicy.c
2540
if (vma->vm_flags & VM_DROPPABLE)
mm/mempolicy.c
844
if (is_cow_mapping(vma->vm_flags) && folio_maybe_mapped_shared(folio))
mm/migrate.c
2672
if ((vma->vm_flags & VM_EXEC) && folio_maybe_mapped_shared(folio))
mm/migrate.c
312
if (folio_test_mlocked(folio) || (pvmw->vma->vm_flags & VM_LOCKED) ||
mm/migrate.c
420
pte = arch_make_huge_pte(pte, shift, vma->vm_flags);
mm/migrate.c
438
if (READ_ONCE(vma->vm_flags) & VM_LOCKED)
mm/migrate_device.c
1036
if (vma->vm_flags & VM_WRITE)
mm/migrate_device.c
1050
if (vma->vm_flags & VM_WRITE)
mm/migrate_device.c
742
(args->vma->vm_flags & VM_SPECIAL) || vma_is_dax(args->vma))
mm/migrate_device.c
839
if (vma->vm_flags & VM_WRITE)
mm/migrate_device.c
852
if (vma->vm_flags & VM_WRITE)
mm/mlock.c
332
if (!(vma->vm_flags & VM_LOCKED))
mm/mlock.c
371
if (vma->vm_flags & VM_LOCKED)
mm/mlock.c
396
if (vma->vm_flags & VM_LOCKED)
mm/mlock.c
473
vm_flags_t oldflags = vma->vm_flags;
mm/mlock.c
545
newflags = vma->vm_flags & ~VM_LOCKED_MASK;
mm/mlock.c
586
if (vma->vm_flags & VM_LOCKED) {
mm/mlock.c
666
vm_flags_t vm_flags = VM_LOCKED;
mm/mlock.c
672
vm_flags |= VM_LOCKONFAULT;
mm/mlock.c
674
return do_mlock(start, len, vm_flags);
mm/mlock.c
731
newflags = vma->vm_flags & ~VM_LOCKED_MASK;
mm/mmap.c
1094
vm_flags_t vm_flags;
mm/mmap.c
1121
if (!vma || !(vma->vm_flags & VM_SHARED)) {
mm/mmap.c
1126
prot |= vma->vm_flags & VM_READ ? PROT_READ : 0;
mm/mmap.c
1127
prot |= vma->vm_flags & VM_WRITE ? PROT_WRITE : 0;
mm/mmap.c
1128
prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0;
mm/mmap.c
1132
if (vma->vm_flags & VM_LOCKED)
mm/mmap.c
1136
vm_flags = vma->vm_flags;
mm/mmap.c
1162
if (vma->vm_flags != vm_flags)
mm/mmap.c
1179
if (next->vm_flags != vma->vm_flags)
mm/mmap.c
1204
int vm_brk_flags(unsigned long addr, unsigned long request, vm_flags_t vm_flags)
mm/mmap.c
1221
if ((vm_flags & (~VM_EXEC)) != 0)
mm/mmap.c
1236
ret = do_brk_flags(&vmi, vma, addr, len, vm_flags);
mm/mmap.c
1261
if (vma->vm_flags & VM_ACCOUNT)
mm/mmap.c
1454
vm_flags_t vm_flags, void *priv,
mm/mmap.c
1465
vm_flags |= mm->def_flags | VM_DONTEXPAND;
mm/mmap.c
1467
vm_flags |= VM_SOFTDIRTY;
mm/mmap.c
1468
vm_flags_init(vma, vm_flags & ~VM_LOCKED_MASK);
mm/mmap.c
1469
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
mm/mmap.c
1478
vm_stat_account(mm, vma->vm_flags, len >> PAGE_SHIFT);
mm/mmap.c
1508
vm_flags_t vm_flags, const struct vm_special_mapping *spec)
mm/mmap.c
1510
return __install_special_mapping(mm, addr, len, vm_flags, (void *)spec,
mm/mmap.c
1719
if (!(new_vma->vm_flags & VM_GROWSDOWN))
mm/mmap.c
1769
if (mpnt->vm_flags & VM_DONTCOPY) {
mm/mmap.c
1775
vm_stat_account(mm, mpnt->vm_flags, -vma_pages(mpnt));
mm/mmap.c
1779
if (mpnt->vm_flags & VM_ACCOUNT) {
mm/mmap.c
1797
if (tmp->vm_flags & VM_WIPEONFORK) {
mm/mmap.c
1840
if (!(tmp->vm_flags & VM_WIPEONFORK))
mm/mmap.c
337
unsigned long flags, vm_flags_t vm_flags,
mm/mmap.c
401
vm_flags |= calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(file, flags) |
mm/mmap.c
407
addr = __get_unmapped_area(file, addr, len, pgoff, flags, vm_flags);
mm/mmap.c
420
if (!mlock_future_ok(mm, vm_flags & VM_LOCKED, len))
mm/mmap.c
463
vm_flags |= VM_SHARED | VM_MAYSHARE;
mm/mmap.c
465
vm_flags &= ~(VM_MAYWRITE | VM_SHARED);
mm/mmap.c
471
if (vm_flags & VM_EXEC)
mm/mmap.c
473
vm_flags &= ~VM_MAYEXEC;
mm/mmap.c
478
if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP))
mm/mmap.c
490
err = memfd_check_seals_mmap(file, &vm_flags);
mm/mmap.c
496
if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP))
mm/mmap.c
502
vm_flags |= VM_SHARED | VM_MAYSHARE;
mm/mmap.c
517
if (vm_flags & (VM_GROWSDOWN | VM_GROWSUP))
mm/mmap.c
520
vm_flags |= VM_DROPPABLE;
mm/mmap.c
526
vm_flags |= VM_NORESERVE;
mm/mmap.c
532
vm_flags |= VM_WIPEONFORK | VM_DONTDUMP;
mm/mmap.c
552
vm_flags |= VM_NORESERVE;
mm/mmap.c
556
vm_flags |= VM_NORESERVE;
mm/mmap.c
559
addr = mmap_region(file, addr, len, vm_flags, pgoff, uf);
mm/mmap.c
561
((vm_flags & VM_LOCKED) ||
mm/mmap.c
647
static inline unsigned long stack_guard_placement(vm_flags_t vm_flags)
mm/mmap.c
649
if (vm_flags & VM_SHADOW_STACK)
mm/mmap.c
691
unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
716
info.start_gap = stack_guard_placement(vm_flags);
mm/mmap.c
726
unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
729
vm_flags);
mm/mmap.c
740
unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
768
info.start_gap = stack_guard_placement(vm_flags);
mm/mmap.c
794
unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
797
vm_flags);
mm/mmap.c
803
unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
807
flags, vm_flags);
mm/mmap.c
808
return arch_get_unmapped_area(filp, addr, len, pgoff, flags, vm_flags);
mm/mmap.c
813
unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags)
mm/mmap.c
83
vm_flags_t vm_flags = vma->vm_flags;
mm/mmap.c
849
pgoff, flags, vm_flags);
mm/mmap.c
852
pgoff, flags, vm_flags);
mm/mmap.c
86
vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags);
mm/mmap.c
88
vm_flags &= ~VM_SHARED;
mm/mmap.c
89
vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags);
mm/mmap.c
972
if (prev->vm_flags & VM_LOCKED)
mm/mmap.c
996
if (vma->vm_flags & VM_LOCKED)
mm/mmap_lock.c
512
if (!vma || !(vma->vm_flags & VM_GROWSDOWN)) {
mm/mmap_lock.c
535
if (!(vma->vm_flags & VM_GROWSDOWN))
mm/mprotect.c
100
if (!(vma->vm_flags & VM_SHARED))
mm/mprotect.c
197
if (vma->vm_flags & VM_SHARED) {
mm/mprotect.c
43
if (WARN_ON_ONCE(!(vma->vm_flags & VM_WRITE)))
mm/mprotect.c
700
vm_flags_t oldflags = READ_ONCE(vma->vm_flags);
mm/mprotect.c
854
if (!(vma->vm_flags & VM_GROWSDOWN))
mm/mprotect.c
862
if (!(vma->vm_flags & VM_GROWSUP))
mm/mprotect.c
885
if (rier && (vma->vm_flags & VM_MAYEXEC))
mm/mprotect.c
897
newflags |= (vma->vm_flags & ~mask_off_old_flags);
mm/mprotect.c
905
if (map_deny_write_exec(vma->vm_flags, newflags)) {
mm/mremap.c
1007
if (!(vrm->vma->vm_flags & VM_ACCOUNT))
mm/mremap.c
1026
vm_stat_account(mm, vma->vm_flags, pages);
mm/mremap.c
1027
if (vma->vm_flags & VM_LOCKED)
mm/mremap.c
1041
vm_flags_t dummy = vma->vm_flags;
mm/mremap.c
1101
bool accountable_move = (vma->vm_flags & VM_ACCOUNT) &&
mm/mremap.c
1405
vm_flags_t vm_flags = vrm->vma->vm_flags;
mm/mremap.c
1408
if (!may_expand_vm(mm, vm_flags, pages))
mm/mremap.c
1694
if (!old_len && !(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) {
mm/mremap.c
1701
(vma->vm_flags & (VM_DONTEXPAND | VM_PFNMAP)))
mm/mremap.c
1731
if (vma->vm_flags & VM_LOCKED)
mm/mremap.c
1740
if (vma->vm_flags & (VM_DONTEXPAND | VM_PFNMAP))
mm/mremap.c
1743
if (!mlock_future_ok(mm, vma->vm_flags & VM_LOCKED, vrm->delta))
mm/mremap.c
1746
if (!may_expand_vm(mm, vma->vm_flags, vrm->delta >> PAGE_SHIFT))
mm/mremap.c
958
if (vma->vm_flags & VM_MAYSHARE)
mm/mremap.c
980
if (!(vrm->vma->vm_flags & VM_ACCOUNT))
mm/mseal.c
71
if (!(vma->vm_flags & VM_SEALED)) {
mm/mseal.c
72
vm_flags_t vm_flags = vma->vm_flags | VM_SEALED;
mm/mseal.c
75
curr_end, &vm_flags);
mm/msync.c
83
(vma->vm_flags & VM_LOCKED)) {
mm/msync.c
93
(vma->vm_flags & VM_SHARED)) {
mm/nommu.c
1017
vm_flags_t vm_flags,
mm/nommu.c
1044
vm_flags |= determine_vm_flags(file, prot, flags, capabilities);
mm/nommu.c
1057
region->vm_flags = vm_flags;
mm/nommu.c
1060
vm_flags_init(vma, vm_flags);
mm/nommu.c
1078
if (is_nommu_shared_mapping(vm_flags)) {
mm/nommu.c
1088
if (!is_nommu_shared_mapping(pregion->vm_flags))
mm/nommu.c
1123
if (pregion->vm_flags & VM_MAPPED_COPY)
mm/nommu.c
1175
if (file && vma->vm_flags & VM_SHARED)
mm/nommu.c
1208
if (vma->vm_flags & VM_EXEC && !region->vm_icache_flushed) {
mm/nommu.c
130
pgprot_t prot, unsigned long vm_flags, int node,
mm/nommu.c
1567
if (is_nommu_shared_mapping(vma->vm_flags))
mm/nommu.c
1616
if (!(vma->vm_flags & VM_USERMAP))
mm/nommu.c
1658
if (write && vma->vm_flags & VM_MAYWRITE)
mm/nommu.c
1661
else if (!write && vma->vm_flags & VM_MAYREAD)
mm/nommu.c
1744
if (vma->vm_flags & VM_MAYREAD) {
mm/nommu.c
1822
if (vma->vm_flags & VM_SHARED) {
mm/nommu.c
1836
if (!(vma->vm_flags & VM_SHARED))
mm/nommu.c
545
if (region->vm_flags & VM_MAPPED_COPY)
mm/nommu.c
853
vm_flags_t vm_flags;
mm/nommu.c
855
vm_flags = calc_vm_prot_bits(prot, 0) | calc_vm_flag_bits(file, flags);
mm/nommu.c
862
vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
mm/nommu.c
866
vm_flags |= (capabilities & NOMMU_VMFLAGS);
mm/nommu.c
868
vm_flags |= VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC;
mm/nommu.c
879
vm_flags |= VM_MAYOVERLAY;
mm/nommu.c
882
vm_flags |= VM_SHARED | VM_MAYSHARE |
mm/nommu.c
886
return vm_flags;
mm/nommu.c
932
if (WARN_ON_ONCE(!is_nommu_shared_mapping(vma->vm_flags)))
mm/nommu.c
966
region->vm_flags = vma->vm_flags;
mm/oom_kill.c
537
if (vma->vm_flags & (VM_HUGETLB|VM_PFNMAP))
mm/oom_kill.c
550
if (vma_is_anonymous(vma) || !(vma->vm_flags & VM_SHARED)) {
mm/pagewalk.c
399
if (vma->vm_flags & VM_PFNMAP) {
mm/rmap.c
1014
pra->vm_flags |= vma->vm_flags & ~VM_LOCKED;
mm/rmap.c
1060
struct mem_cgroup *memcg, vm_flags_t *vm_flags)
mm/rmap.c
1075
*vm_flags = 0;
mm/rmap.c
1089
*vm_flags = pra.vm_flags;
mm/rmap.c
1187
if (vma->vm_flags & VM_SHARED)
mm/rmap.c
1649
if (!folio_test_swapbacked(folio) && !(vma->vm_flags & VM_DROPPABLE))
mm/rmap.c
2031
(vma->vm_flags & VM_LOCKED)) {
mm/rmap.c
2251
if (folio_test_dirty(folio) && !(vma->vm_flags & VM_DROPPABLE)) {
mm/rmap.c
2342
if (vma->vm_flags & VM_LOCKED)
mm/rmap.c
2713
if (vma->vm_flags & VM_LOCKED)
mm/rmap.c
910
vm_flags_t vm_flags;
mm/rmap.c
929
if (vma->vm_flags & VM_LOCKED) {
mm/rmap.c
950
pra->vm_flags |= VM_LOCKED;
mm/shmem.c
1839
vm_flags_t vm_flags = vma ? vma->vm_flags : 0;
mm/shmem.c
1842
if (thp_disabled_by_hw() || (vma && vma_thp_disabled(vma, vm_flags, shmem_huge_force)))
mm/shmem.c
1846
shmem_huge_force, vma, vm_flags);
mm/shmem.c
1868
if (vm_flags & VM_HUGEPAGE)
mm/shmem.c
629
vm_flags_t vm_flags)
mm/shmem.c
662
if (vm_flags & VM_HUGEPAGE)
mm/shmem.c
864
vm_flags_t vm_flags)
mm/swap.c
527
if (unlikely((vma->vm_flags & (VM_LOCKED | VM_SPECIAL)) == VM_LOCKED))
mm/userfaultfd.c
1532
return !(vma->vm_flags & (VM_PFNMAP | VM_IO | VM_HUGETLB |
mm/userfaultfd.c
1541
if ((src_vma->vm_flags & VM_ACCESS_FLAGS) != (dst_vma->vm_flags & VM_ACCESS_FLAGS) ||
mm/userfaultfd.c
1546
if ((src_vma->vm_flags & VM_LOCKED) != (dst_vma->vm_flags & VM_LOCKED))
mm/userfaultfd.c
1553
if (!(src_vma->vm_flags & VM_WRITE))
mm/userfaultfd.c
176
bool writable = dst_vma->vm_flags & VM_WRITE;
mm/userfaultfd.c
177
bool vm_shared = dst_vma->vm_flags & VM_SHARED;
mm/userfaultfd.c
1800
if (src_vma->vm_flags & VM_SHARED)
mm/userfaultfd.c
1805
if (dst_vma->vm_flags & VM_SHARED)
mm/userfaultfd.c
1942
vm_flags_t vm_flags)
mm/userfaultfd.c
1944
const bool uffd_wp_changed = (vma->vm_flags ^ vm_flags) & VM_UFFD_WP;
mm/userfaultfd.c
1946
vm_flags_reset(vma, vm_flags);
mm/userfaultfd.c
1952
if ((vma->vm_flags & VM_SHARED) && uffd_wp_changed)
mm/userfaultfd.c
1958
vm_flags_t vm_flags)
mm/userfaultfd.c
1963
(vma->vm_flags & ~__VM_UFFD_FLAGS) | vm_flags);
mm/userfaultfd.c
1992
vma->vm_flags & ~__VM_UFFD_FLAGS,
mm/userfaultfd.c
2009
vm_flags_t vm_flags,
mm/userfaultfd.c
2024
VM_WARN_ON_ONCE(!vma_can_userfault(vma, vm_flags, wp_async));
mm/userfaultfd.c
2027
VM_WARN_ON_ONCE(!(vma->vm_flags & VM_MAYWRITE));
mm/userfaultfd.c
2034
(vma->vm_flags & vm_flags) == vm_flags)
mm/userfaultfd.c
2041
new_flags = (vma->vm_flags & ~__VM_UFFD_FLAGS) | vm_flags;
mm/userfaultfd.c
2054
userfaultfd_set_ctx(vma, ctx, vm_flags);
mm/userfaultfd.c
2104
!!(vma->vm_flags & __VM_UFFD_FLAGS));
mm/userfaultfd.c
51
else if (!(vma->vm_flags & VM_SHARED) &&
mm/userfaultfd.c
687
if (!(dst_vma->vm_flags & VM_SHARED)) {
mm/userfaultfd.c
759
dst_vma->vm_flags & VM_SHARED))
mm/userfaultfd.c
766
if ((flags & MFILL_ATOMIC_WP) && !(dst_vma->vm_flags & VM_UFFD_WP))
mm/userfaultfd.c
80
if (!(vma->vm_flags & VM_SHARED) && unlikely(!vma->anon_vma))
mm/vma.c
1062
if ((vmg->vm_flags & VM_SPECIAL) || (!prev && !next))
mm/vma.c
1100
khugepaged_enter_vma(vmg->target, vmg->vm_flags);
mm/vma.c
1177
sticky_flags = vmg->vm_flags & VM_STICKY;
mm/vma.c
1178
sticky_flags |= target->vm_flags & VM_STICKY;
mm/vma.c
1180
sticky_flags |= next->vm_flags & VM_STICKY;
mm/vma.c
1443
if (next->vm_flags & VM_LOCKED)
mm/vma.c
1446
if (next->vm_flags & VM_ACCOUNT)
mm/vma.c
1449
if (is_exec_mapping(next->vm_flags))
mm/vma.c
1451
else if (is_stack_mapping(next->vm_flags))
mm/vma.c
1453
else if (is_data_mapping(next->vm_flags))
mm/vma.c
1695
const vm_flags_t vm_flags = *vm_flags_ptr;
mm/vma.c
1698
vmg.vm_flags = vm_flags;
mm/vma.c
1710
*vm_flags_ptr = ret->vm_flags;
mm/vma.c
1740
unsigned long start, unsigned long end, vm_flags_t vm_flags,
mm/vma.c
1745
vmg.vm_flags = vm_flags;
mm/vma.c
19
vm_flags_t vm_flags;
mm/vma.c
1956
!((a->vm_flags ^ b->vm_flags) & ~(VM_ACCESS_FLAGS | VM_IGNORE_MERGE)) &&
mm/vma.c
2044
return (vma->vm_flags & (VM_WRITE | VM_SHARED)) ==
mm/vma.c
2051
if (vma->vm_flags & VM_PFNMAP)
mm/vma.c
2098
pgprot_val(vm_pgprot_modify(vm_page_prot, vma->vm_flags)))
mm/vma.c
2317
static bool accountable_mapping(struct file *file, vm_flags_t vm_flags)
mm/vma.c
2326
return (vm_flags & (VM_NORESERVE | VM_SHARED | VM_WRITE)) == VM_WRITE;
mm/vma.c
2364
map->vm_flags = ksm_vma_flags(map->mm, map->file, map->vm_flags);
mm/vma.c
2425
if (!may_expand_vm(map->mm, map->vm_flags, map->pglen - vms->nr_pages))
mm/vma.c
2429
if (accountable_mapping(map->file, map->vm_flags)) {
mm/vma.c
2439
map->vm_flags |= VM_ACCOUNT;
mm/vma.c
2487
VM_WARN_ON_ONCE(map->vm_flags != vma->vm_flags &&
mm/vma.c
2488
!(map->vm_flags & VM_MAYWRITE) &&
mm/vma.c
2489
(vma->vm_flags & VM_MAYWRITE));
mm/vma.c
2492
map->vm_flags = vma->vm_flags;
mm/vma.c
2523
vm_flags_init(vma, map->vm_flags);
mm/vma.c
2533
else if (map->vm_flags & VM_SHARED)
mm/vma.c
2543
vm_flags_init(vma, map->vm_flags);
mm/vma.c
2548
WARN_ON_ONCE(!arch_validate_flags(map->vm_flags));
mm/vma.c
2562
khugepaged_enter_vma(vma, map->vm_flags);
mm/vma.c
2583
vm_flags_t vm_flags = vma->vm_flags;
mm/vma.c
2590
vm_stat_account(mm, vma->vm_flags, map->pglen);
mm/vma.c
2591
if (vm_flags & VM_LOCKED) {
mm/vma.c
2592
if ((vm_flags & VM_SPECIAL) || vma_is_dax(vma) ||
mm/vma.c
2721
unsigned long len, vm_flags_t vm_flags, unsigned long pgoff,
mm/vma.c
2728
MMAP_STATE(map, mm, &vmi, addr, len, pgoff, vm_flags, file);
mm/vma.c
2819
unsigned long len, vm_flags_t vm_flags, unsigned long pgoff,
mm/vma.c
2828
if (map_deny_write_exec(vm_flags, vm_flags))
mm/vma.c
2832
if (!arch_validate_flags(vm_flags))
mm/vma.c
2836
if (file && is_shared_maywrite_vm_flags(vm_flags)) {
mm/vma.c
2844
ret = __mmap_region(file, addr, len, vm_flags, pgoff, uf);
mm/vma.c
2867
unsigned long addr, unsigned long len, vm_flags_t vm_flags)
mm/vma.c
2875
vm_flags |= VM_DATA_DEFAULT_FLAGS | VM_ACCOUNT | mm->def_flags;
mm/vma.c
2876
vm_flags = ksm_vma_flags(mm, NULL, vm_flags);
mm/vma.c
2877
if (!may_expand_vm(mm, vm_flags, len >> PAGE_SHIFT))
mm/vma.c
2891
VMG_STATE(vmg, mm, vmi, addr, addr + len, vm_flags, PHYS_PFN(addr));
mm/vma.c
2912
vm_flags_init(vma, vm_flags);
mm/vma.c
2913
vma->vm_page_prot = vm_get_page_prot(vm_flags);
mm/vma.c
2924
if (vm_flags & VM_LOCKED)
mm/vma.c
2976
if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */
mm/vma.c
3028
if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */
mm/vma.c
3058
if (!may_expand_vm(mm, vma->vm_flags, grow))
mm/vma.c
3066
if (!mlock_future_ok(mm, vma->vm_flags & VM_LOCKED, grow << PAGE_SHIFT))
mm/vma.c
3070
new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start :
mm/vma.c
3098
if (!(vma->vm_flags & VM_GROWSUP))
mm/vma.c
3118
if (!(next->vm_flags & VM_GROWSUP))
mm/vma.c
3152
if (vma->vm_flags & VM_LOCKED)
mm/vma.c
3154
vm_stat_account(mm, vma->vm_flags, grow);
mm/vma.c
3183
if (!(vma->vm_flags & VM_GROWSDOWN))
mm/vma.c
3196
if (!(prev->vm_flags & VM_GROWSDOWN) &&
mm/vma.c
3231
if (vma->vm_flags & VM_LOCKED)
mm/vma.c
3233
vm_stat_account(mm, vma->vm_flags, grow);
mm/vma.c
3281
if ((vma->vm_flags & VM_ACCOUNT) &&
mm/vma.c
3303
if (vma->vm_flags & VM_ACCOUNT)
mm/vma.c
55
.vm_flags = vm_flags_, \
mm/vma.c
66
.vm_flags = (map_)->vm_flags, \
mm/vma.c
808
vm_flags_t sticky_flags = vmg->vm_flags & VM_STICKY;
mm/vma.c
847
if (vmg->vm_flags & VM_SPECIAL || (!left_side && !right_side))
mm/vma.c
903
sticky_flags |= (next->vm_flags & VM_STICKY);
mm/vma.c
909
sticky_flags |= (prev->vm_flags & VM_STICKY);
mm/vma.c
92
if ((vma->vm_flags ^ vmg->vm_flags) & ~VM_IGNORE_MERGE)
mm/vma.c
980
khugepaged_enter_vma(vmg->target, vmg->vm_flags);
mm/vma.h
101
vm_flags_t vm_flags;
mm/vma.h
242
.vm_flags = vm_flags_, \
mm/vma.h
256
.vm_flags = vma_->vm_flags, \
mm/vma.h
432
unsigned long start, unsigned long end, vm_flags_t vm_flags,
mm/vma.h
460
unsigned long len, vm_flags_t vm_flags, unsigned long pgoff,
mm/vma.h
477
if (vma->vm_flags & VM_SHARED)
mm/vma.h
479
return !!(vma->vm_flags & VM_WRITE);
mm/vma.h
483
static inline pgprot_t vm_pgprot_modify(pgprot_t oldprot, vm_flags_t vm_flags)
mm/vma.h
485
return pgprot_modify(oldprot, vm_get_page_prot(vm_flags));
mm/vma.h
664
return (vma->vm_flags & VM_SEALED);
mm/vma_exec.c
144
vma->vm_page_prot = vm_get_page_prot(vma->vm_flags);
mm/vma_init.c
128
ASSERT_EXCLUSIVE_WRITER(orig->vm_flags);
mm/vma_init.c
52
vm_flags_init(dest, src->vm_flags);
mm/vmalloc.c
3988
pgprot_t prot, unsigned long vm_flags, int node,
mm/vmalloc.c
4007
if (vmap_allow_huge && (vm_flags & VM_ALLOW_HUGE_VMAP)) {
mm/vmalloc.c
4025
VM_UNINITIALIZED | vm_flags, start, end, node,
mm/vmalloc.c
4090
if (!(vm_flags & VM_DEFER_KMEMLEAK))
mm/vmscan.c
2105
vm_flags_t vm_flags;
mm/vmscan.c
2151
&vm_flags) != 0) {
mm/vmscan.c
2161
if ((vm_flags & VM_EXEC) && folio_is_file_lru(folio)) {
mm/vmscan.c
3340
if (vma->vm_flags & (VM_LOCKED | VM_SPECIAL))
mm/vmscan.c
4231
if (vma->vm_flags & VM_SPECIAL)
mm/vmscan.c
887
vm_flags_t vm_flags;
mm/vmscan.c
890
&vm_flags);
mm/vmscan.c
896
if (vm_flags & VM_LOCKED)
mm/vmscan.c
940
if ((vm_flags & VM_EXEC) && folio_is_file_lru(folio))
net/ipv4/tcp.c
1896
if (vma->vm_flags & (VM_WRITE | VM_EXEC))
samples/vfio-mdev/mbochs.c
804
if ((vma->vm_flags & VM_SHARED) == 0)
samples/vfio-mdev/mbochs.c
836
if ((vma->vm_flags & VM_SHARED) == 0)
samples/vfio-mdev/mdpy.c
432
if ((vma->vm_flags & VM_SHARED) == 0)
security/apparmor/lsm.c
587
!(vma->vm_flags & VM_SHARED) ? MAP_PRIVATE : 0);
security/integrity/ima/ima_main.c
541
!(prot & PROT_EXEC) || (vma->vm_flags & VM_EXEC))
security/ipe/hooks.c
114
if (vma->vm_flags & VM_EXEC)
security/selinux/hooks.c
4024
(prot & PROT_EXEC) && !(vma->vm_flags & VM_EXEC)) {
security/selinux/hooks.c
4057
return file_map_prot_check(vma->vm_file, prot, vma->vm_flags&VM_SHARED);
security/selinux/selinuxfs.c
254
if (vma->vm_flags & VM_WRITE)
security/selinux/selinuxfs.c
470
if (vma->vm_flags & VM_SHARED) {
security/selinux/selinuxfs.c
474
if (vma->vm_flags & VM_WRITE)
sound/core/memalloc.c
854
area->vm_page_prot = vm_get_page_prot(area->vm_flags);
sound/core/oss/pcm_oss.c
2884
switch ((area->vm_flags & (VM_READ | VM_WRITE))) {
sound/core/pcm_native.c
3747
if (!(area->vm_flags & VM_READ))
sound/core/pcm_native.c
3785
if (!(area->vm_flags & VM_READ))
sound/core/pcm_native.c
3979
if (!(area->vm_flags & (VM_WRITE|VM_READ)))
sound/core/pcm_native.c
3982
if (!(area->vm_flags & VM_READ))
sound/usb/usx2y/us122l.c
183
if (read && area->vm_flags & VM_WRITE)
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
43
perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-';
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
44
perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-';
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
45
perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-';
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
46
perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p';
tools/testing/selftests/bpf/progs/find_vma.c
31
if (vma->vm_flags & VM_EXEC)
tools/testing/vma/include/dup.h
1085
.vm_flags = vma->vm_flags,
tools/testing/vma/include/dup.h
1187
vm_flags_t vm_flags = vma->vm_flags;
tools/testing/vma/include/dup.h
1191
vm_page_prot = pgprot_modify(vma->vm_page_prot, vm_get_page_prot(vm_flags));
tools/testing/vma/include/dup.h
1194
vm_flags &= ~VM_SHARED;
tools/testing/vma/include/dup.h
1196
vm_page_prot = pgprot_modify(vm_page_prot, vm_get_page_prot(vm_flags));
tools/testing/vma/include/dup.h
1204
if (vma->vm_flags & VM_GROWSDOWN)
tools/testing/vma/include/dup.h
1208
if (vma->vm_flags & VM_SHADOW_STACK)
tools/testing/vma/include/dup.h
1229
if (vma->vm_flags & VM_GROWSUP) {
tools/testing/vma/include/dup.h
1239
return vma->vm_flags & VM_ACCESS_FLAGS;
tools/testing/vma/include/dup.h
1243
vm_flags_t vm_flags, unsigned long bytes)
tools/testing/vma/include/dup.h
1247
if (!(vm_flags & VM_LOCKED) || capable(CAP_IPC_LOCK))
tools/testing/vma/include/dup.h
498
vm_flags_t vm_flags;
tools/testing/vma/include/dup.h
531
const vm_flags_t vm_flags;
tools/testing/vma/include/dup.h
733
static inline pgprot_t vm_get_page_prot(vm_flags_t vm_flags)
tools/testing/vma/include/dup.h
735
return __pgprot(vm_flags);
tools/testing/vma/include/dup.h
901
static inline bool is_shared_maywrite_vm_flags(vm_flags_t vm_flags)
tools/testing/vma/include/dup.h
903
return (vm_flags & (VM_SHARED | VM_MAYWRITE)) ==
tools/testing/vma/include/stubs.h
105
const struct file *file, vm_flags_t vm_flags)
tools/testing/vma/include/stubs.h
107
return vm_flags;
tools/testing/vma/include/stubs.h
186
vm_flags_t vm_flags)
tools/testing/vma/shared.c
17
pgoff_t pgoff, vm_flags_t vm_flags)
tools/testing/vma/shared.c
27
vm_flags_reset(vma, vm_flags);
tools/testing/vma/shared.c
41
pgoff_t pgoff, vm_flags_t vm_flags)
tools/testing/vma/shared.c
43
struct vm_area_struct *vma = alloc_vma(mm, start, end, pgoff, vm_flags);
tools/testing/vma/shared.h
79
pgoff_t pgoff, vm_flags_t vm_flags);
tools/testing/vma/shared.h
87
pgoff_t pgoff, vm_flags_t vm_flags);
tools/testing/vma/tests/merge.c
100
struct vm_area_struct *vma_left = alloc_vma(&mm, 0, 0x1000, 0, vm_flags);
tools/testing/vma/tests/merge.c
101
struct vm_area_struct *vma_right = alloc_vma(&mm, 0x2000, 0x3000, 2, vm_flags);
tools/testing/vma/tests/merge.c
1011
vma = alloc_and_link_vma(&mm, 0x3000, 0x8000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1014
vmg_set_range(&vmg, 0x4000, 0x5000, 4, vm_flags);
tools/testing/vma/tests/merge.c
1020
vmg_set_range(&vmg, 0x5000, 0x6000, 5, vm_flags);
tools/testing/vma/tests/merge.c
1026
vmg_set_range(&vmg, 0x6000, 0x7000, 6, vm_flags);
tools/testing/vma/tests/merge.c
1032
vmg_set_range(&vmg, 0x4000, 0x7000, 4, vm_flags);
tools/testing/vma/tests/merge.c
1038
vmg_set_range(&vmg, 0x4000, 0x6000, 4, vm_flags);
tools/testing/vma/tests/merge.c
1044
vmg_set_range(&vmg, 0x5000, 0x6000, 5, vm_flags);
tools/testing/vma/tests/merge.c
1070
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
108
.vm_flags = vm_flags,
tools/testing/vma/tests/merge.c
1094
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1095
vma = alloc_and_link_vma(&mm, 0x3000, 0x7000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1096
vma_next = alloc_and_link_vma(&mm, 0x7000, 0x9000, 7, vm_flags);
tools/testing/vma/tests/merge.c
1104
vmg_set_range_anon_vma(&vmg, 0x3000, 0x7000, 3, vm_flags, NULL);
tools/testing/vma/tests/merge.c
1132
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1133
vma_next = alloc_and_link_vma(&mm, 0x7000, 0x9000, 7, vm_flags);
tools/testing/vma/tests/merge.c
1135
vmg_set_range_anon_vma(&vmg, 0x3000, 0x7000, 3, vm_flags, NULL);
tools/testing/vma/tests/merge.c
1157
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
1178
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1179
vma_next = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1182
vmg_set_range(&vmg, 0, 0x5000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1204
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1205
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1206
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x8000, 5, vm_flags);
tools/testing/vma/tests/merge.c
121
ASSERT_EQ(vma->vm_flags, vm_flags);
tools/testing/vma/tests/merge.c
1213
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1237
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1238
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1239
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x8000, 5, vm_flags);
tools/testing/vma/tests/merge.c
1242
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1266
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1267
vma = alloc_and_link_vma(&mm, 0x3000, 0x8000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1270
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1294
vma = alloc_and_link_vma(&mm, 0, 0x5000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1295
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x8000, 5, vm_flags);
tools/testing/vma/tests/merge.c
1298
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1317
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
132
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
1333
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1334
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1337
vmg_set_range_anon_vma(&vmg, 0x3000, 0x5000, 3, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
134
struct vm_area_struct *init_vma = alloc_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1361
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1362
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1365
vmg_set_range(&vmg, 0, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1383
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
1388
vma = alloc_and_link_vma(&mm, 0, 0x1000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1389
alloc_and_link_vma(&mm, 0x3000, 0x4000, 3, vm_flags);
tools/testing/vma/tests/merge.c
1413
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
1417
VMG_STATE(vmg, &mm, &vmi, 0x5000, 0x9000, vm_flags, 5);
tools/testing/vma/tests/merge.c
1425
alloc_and_link_vma(&mm, 0, 0x2000, 0, vm_flags);
tools/testing/vma/tests/merge.c
1432
vma_prev = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
192
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
194
struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x1000, 0, vm_flags);
tools/testing/vma/tests/merge.c
220
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
222
struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
241
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
268
vm_flags |= VM_STICKY;
tools/testing/vma/tests/merge.c
274
vma_a = alloc_and_link_vma(&mm, 0, 0x2000, 0, vm_flags);
tools/testing/vma/tests/merge.c
282
vma_b = alloc_and_link_vma(&mm, 0x3000, 0x4000, 3, vm_flags);
tools/testing/vma/tests/merge.c
289
vma_c = alloc_and_link_vma(&mm, 0xb000, 0xc000, 0xb, vm_flags);
tools/testing/vma/tests/merge.c
302
vma_d = try_merge_new_vma(&mm, &vmg, 0x7000, 0x9000, 7, vm_flags, &merged);
tools/testing/vma/tests/merge.c
317
vma = try_merge_new_vma(&mm, &vmg, 0x2000, 0x3000, 2, vm_flags, &merged);
tools/testing/vma/tests/merge.c
328
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
336
vma = try_merge_new_vma(&mm, &vmg, 0x4000, 0x5000, 4, vm_flags, &merged);
tools/testing/vma/tests/merge.c
347
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
357
vma = try_merge_new_vma(&mm, &vmg, 0x6000, 0x7000, 6, vm_flags, &merged);
tools/testing/vma/tests/merge.c
36
unsigned long end, pgoff_t pgoff, vm_flags_t vm_flags)
tools/testing/vma/tests/merge.c
368
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
377
vma = try_merge_new_vma(&mm, &vmg, 0x5000, 0x6000, 5, vm_flags, &merged);
tools/testing/vma/tests/merge.c
388
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
397
vma = try_merge_new_vma(&mm, &vmg, 0xa000, 0xb000, 0xa, vm_flags, &merged);
tools/testing/vma/tests/merge.c
408
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
416
vma = try_merge_new_vma(&mm, &vmg, 0x9000, 0xa000, 0x9, vm_flags, &merged);
tools/testing/vma/tests/merge.c
427
ASSERT_TRUE(IS_SET(vma->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
472
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
48
vmg->vm_flags = vm_flags;
tools/testing/vma/tests/merge.c
494
vma_left = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
505
vmg_set_range(&vmg, 0x3000, 0x4000, 3, vm_flags);
tools/testing/vma/tests/merge.c
509
vm_flags_reset(vma_left, vm_flags | special_flag);
tools/testing/vma/tests/merge.c
510
vmg.vm_flags = vm_flags | special_flag;
tools/testing/vma/tests/merge.c
524
vma = alloc_and_link_vma(&mm, 0x3000, 0x4000, 3, vm_flags);
tools/testing/vma/tests/merge.c
531
vm_flags_reset(vma_left, vm_flags | special_flag);
tools/testing/vma/tests/merge.c
532
vmg.vm_flags = vm_flags | special_flag;
tools/testing/vma/tests/merge.c
544
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
59
unsigned long end, pgoff_t pgoff, vm_flags_t vm_flags,
tools/testing/vma/tests/merge.c
62
vmg_set_range(vmg, start, end, pgoff, vm_flags);
tools/testing/vma/tests/merge.c
624
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
625
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x9000, 5, vm_flags);
tools/testing/vma/tests/merge.c
628
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
649
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
650
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
653
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
677
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
678
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x9000, 5, vm_flags);
tools/testing/vma/tests/merge.c
681
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
705
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
706
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
707
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x9000, 5, vm_flags);
tools/testing/vma/tests/merge.c
710
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
731
vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, vm_flags);
tools/testing/vma/tests/merge.c
732
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
733
vma_next = alloc_and_link_vma(&mm, 0x5000, 0x9000, 5, vm_flags);
tools/testing/vma/tests/merge.c
736
vmg_set_range(&vmg, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/merge.c
74
unsigned long end, pgoff_t pgoff, vm_flags_t vm_flags,
tools/testing/vma/tests/merge.c
753
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
760
struct vm_area_struct *vma_prev = alloc_and_link_vma(&mm, 0, 0x2000, 0, vm_flags);
tools/testing/vma/tests/merge.c
761
struct vm_area_struct *vma_next = alloc_and_link_vma(&mm, 0x5000, 0x7000, 5, vm_flags);
tools/testing/vma/tests/merge.c
79
vmg_set_range(vmg, start, end, pgoff, vm_flags);
tools/testing/vma/tests/merge.c
791
vmg_set_range(&vmg, 0x2000, 0x5000, 2, vm_flags);
tools/testing/vma/tests/merge.c
808
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
809
vm_flags_t prev_flags = vm_flags;
tools/testing/vma/tests/merge.c
810
vm_flags_t next_flags = vm_flags;
tools/testing/vma/tests/merge.c
826
vm_flags |= VM_STICKY;
tools/testing/vma/tests/merge.c
840
vma = alloc_and_link_vma(&mm, 0x2000, 0x6000, 2, vm_flags);
tools/testing/vma/tests/merge.c
844
vmg_set_range_anon_vma(&vmg, 0x3000, 0x6000, 3, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
861
ASSERT_TRUE(IS_SET(vma_next->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
876
vma = alloc_and_link_vma(&mm, 0x2000, 0x6000, 2, vm_flags);
tools/testing/vma/tests/merge.c
879
vmg_set_range_anon_vma(&vmg, 0x2000, 0x6000, 2, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
891
ASSERT_TRUE(IS_SET(vma_next->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
908
vma = alloc_and_link_vma(&mm, 0x3000, 0x7000, 3, vm_flags);
tools/testing/vma/tests/merge.c
910
vmg_set_range_anon_vma(&vmg, 0x3000, 0x6000, 3, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
92
return alloc_and_link_vma(mm, start, end, pgoff, vm_flags);
tools/testing/vma/tests/merge.c
927
ASSERT_TRUE(IS_SET(vma_prev->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
944
vma = alloc_and_link_vma(&mm, 0x3000, 0x7000, 3, vm_flags);
tools/testing/vma/tests/merge.c
945
vmg_set_range_anon_vma(&vmg, 0x3000, 0x7000, 3, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
958
ASSERT_TRUE(IS_SET(vma_prev->vm_flags, VM_STICKY));
tools/testing/vma/tests/merge.c
975
vma = alloc_and_link_vma(&mm, 0x3000, 0x7000, 3, vm_flags);
tools/testing/vma/tests/merge.c
977
vmg_set_range_anon_vma(&vmg, 0x3000, 0x7000, 3, vm_flags, &dummy_anon_vma);
tools/testing/vma/tests/merge.c
98
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/merge.c
990
ASSERT_TRUE(IS_SET(vma_prev->vm_flags, VM_STICKY));
tools/testing/vma/tests/vma.c
25
vm_flags_t vm_flags = VM_READ | VM_WRITE | VM_MAYREAD | VM_MAYWRITE;
tools/testing/vma/tests/vma.c
33
vma = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, vm_flags);
tools/testing/vma/tests/vma.c
45
vma = alloc_and_link_vma(&mm, 0, 0x2000, 0, vm_flags);
tools/testing/vma/tests/vma.c
46
vma_next = alloc_and_link_vma(&mm, 0x6000, 0x8000, 6, vm_flags);
virt/kvm/guest_memfd.c
473
if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) !=
virt/kvm/kvm_main.c
2937
if (unlikely(!(vma->vm_flags & VM_READ)))
virt/kvm/kvm_main.c
2940
if (write_fault && (unlikely(!(vma->vm_flags & VM_WRITE))))
virt/kvm/kvm_main.c
3020
else if (vma->vm_flags & (VM_IO | VM_PFNMAP)) {
virt/kvm/kvm_main.c
4090
((vma->vm_flags & VM_EXEC) || !(vma->vm_flags & VM_SHARED)))