VM_PROT_ALL
VM_PROT_ALL, 0, NULL);
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
pti_obj = vm_pager_allocate(OBJT_PHYS, NULL, 0, VM_PROT_ALL, 0, NULL);
KASSERT((prot & ~VM_PROT_ALL) == 0, ("invalid prot %x", prot));
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
KASSERT((prot & ~VM_PROT_ALL) == 0, ("invalid prot %x", prot));
VM_PROT_ALL, 0, NULL);
KASSERT((prot & ~VM_PROT_ALL) == 0, ("invalid prot %x", prot));
KINST_TRAMPCHUNK_SIZE, 0, VMFS_ANY_SPACE, VM_PROT_ALL, VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
return (kmap_atomic_prot(page, VM_PROT_ALL));
return (vm_flags & VM_PROT_ALL);
CTASSERT((VM_PROT_ALL & -(1 << 8)) == 0);
CTASSERT((VM_PROT_ALL & -LINUXKPI_PROT_VALID) == 0);
(((prot) & VM_PROT_ALL) | cachemode2protval(VM_MEMATTR_UNCACHEABLE))
(((prot) & VM_PROT_ALL) | cachemode2protval(VM_MEMATTR_WRITE_COMBINING))
vmap->vm_flags = vmap->vm_page_prot = (nprot & VM_PROT_ALL);
VM_PROT_ALL, 0, td->td_ucred);
VM_PROT_ALL, 0, NULL);
if (prot == 0 || (prot & ~(VM_PROT_ALL)) != 0)
mem_obj = vm_pager_allocate(OBJT_PHYS, NULL, size, VM_PROT_ALL, 0,
.sv_stackprot = VM_PROT_ALL,
KASSERT((prot & ~VM_PROT_ALL) == 0, ("invalid prot %x", prot));
VM_PROT_ALL, VM_PROT_ALL, 0);
VM_PROT_READ | VM_PROT_EXECUTE, VM_PROT_ALL,
vmaddr + a_out->a_text + a_out->a_data, VM_PROT_ALL, 0,
bss_size, 0, VMFS_NO_SPACE, VM_PROT_ALL, VM_PROT_ALL, 0);
VM_PROT_ALL, VM_PROT_ALL, 0);
a_out->a_text + a_out->a_data, VM_PROT_ALL, VM_PROT_ALL,
&vmaddr, bss_size, 0, VMFS_NO_SPACE, VM_PROT_ALL,
VM_PROT_ALL, 0);
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
VM_PROT_READ | VM_PROT_EXECUTE, VM_PROT_ALL,
VM_PROT_ALL, VM_PROT_ALL,
VM_PROT_ALL, VM_PROT_ALL, 0);
.sv_stackprot = VM_PROT_ALL,
if ((entry->protection & VM_PROT_ALL) == 0)
trunc_page(start), VM_PROT_ALL, VM_PROT_ALL, MAP_CHECK_EXCL);
prot | VM_PROT_WRITE, VM_PROT_ALL, MAP_CHECK_EXCL);
prot, VM_PROT_ALL, cow | MAP_CHECK_EXCL |
.sv_stackprot = VM_PROT_ALL,
*maxprotp = VM_PROT_ALL;
sv->sv_usrstack, find_space, stack_prot, VM_PROT_ALL,
ef->object = vm_pager_allocate(OBJT_PHYS, NULL, mapsize, VM_PROT_ALL,
VM_PROT_ALL, VM_PROT_ALL, 0);
error = preload_protect(ef, VM_PROT_ALL);
round_page(mapsize), 0, VMFS_OPTIMAL_SPACE, VM_PROT_ALL,
VM_PROT_ALL, 0);
round_page((vm_offset_t)ef->address + ef->lf.size), VM_PROT_ALL);
VM_PROT_ALL, 0, thread0.td_ucred);
VM_PROT_WRITE, VM_PROT_ALL, MAP_PRIVATE | MAP_ANON, NULL, 0,
KASSERT((prot & ~VM_PROT_ALL) == 0, ("invalid prot %x", prot));
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
.sv_stackprot = VM_PROT_ALL,
switch (prot & VM_PROT_ALL) {
#define VM_PROT_DEFAULT VM_PROT_ALL
KASSERT((rw & ~VM_PROT_ALL) == 0,
KASSERT((rw & ~VM_PROT_ALL) == 0,
prot = (flags & M_EXEC) != 0 ? VM_PROT_ALL : VM_PROT_RW;
VMFS_SUPER_SPACE : VMFS_ANY_SPACE, VM_PROT_ALL, VM_PROT_ALL,
prot = (flags & M_EXEC) != 0 ? VM_PROT_ALL : VM_PROT_RW;
VMFS_SUPER_SPACE, VM_PROT_ALL, VM_PROT_ALL, MAP_NOFAULT);
start, VM_PROT_ALL, VM_PROT_ALL, MAP_NOFAULT);
VM_PROT_ALL)
fault_type = fault_typea & VM_PROT_ALL;
*maxprotp = VM_PROT_ALL;
rv = vm_map_insert(map, NULL, 0, old, new, prot, VM_PROT_ALL,
VM_PROT_ALL, 0, NULL);