PROT_NONE
PROT_NONE) != 0)
PROT_NONE) != 0)
if (mprotect(start, 1, PROT_NONE) == -1)
if (mprotect(start + nbytes, 1, PROT_NONE) == -1)
if (mmap((void *)laddr, size, PROT_NONE,
if (mmap((void *)laddr, syspagsz, PROT_NONE,
(head->ch_resend - head->ch_resbgn), PROT_NONE,
(void) mprotect(addr, redzonesize, PROT_NONE);
if (arg == PROT_NONE)
(void) mprotect(stk, guardsize, PROT_NONE);
prot = PROT_NONE;
#define FREE_PROT PROT_NONE
vmm->vmm_mem = mmap(NULL, last_gpa, PROT_NONE,
ptr = mmap(NULL, len, PROT_NONE, MAP_GUARD | MAP_ALIGNED_SUPER, -1, 0);
base = mmap(NULL, len2, PROT_NONE, MAP_GUARD | MAP_ALIGNED_SUPER, -1,
if (mprotect(buf + pagesize, pagesize, PROT_NONE) < 0) {
void *addr = mmap(NULL, pgsz, PROT_NONE, MAP_PRIVATE | MAP_ANON,
ioctl_fault_addr = mmap(NULL, sysconf(_SC_PAGESIZE) * 4, PROT_NONE,
if (mprotect(addr, pgsz, PROT_NONE) != 0) {
if (mprotect(addr, pgsz, PROT_NONE) != 0) {
if (mprotect(guarded_page + page_size, page_size, PROT_NONE) < 0) {
unmap = mmap(NULL, 1024 * 1024, PROT_NONE, MAP_PRIVATE | MAP_ANON, -1,
unmap = mmap(NULL, page, PROT_NONE, MAP_PRIVATE | MAP_ANON, -1, 0);
void *addr = mmap(NULL, sysconf(_SC_PAGESIZE), PROT_NONE,
void *addr = mmap(NULL, pgsz * 2, PROT_NONE, MAP_PRIVATE | MAP_ANON,
if (mprotect((void *)((uintptr_t)addr + pgsz), pgsz, PROT_NONE) != 0) {
if (mprotect(addr, len, PROT_NONE) != 0) {
mprotect(addr, size, PROT_NONE);
dev_a.prot = dev_a.maxprot = (uchar_t)PROT_NONE;
valid_usr_range(addr_start, sg_sz, PROT_NONE, as,
dev_a.prot = dev_a.maxprot = (uchar_t)PROT_NONE;
return (PROT_NONE); /* can't happen */
if ((prot & ~PROT_USER) == PROT_NONE) {
prot = PROT_NONE;
protchk = PROT_NONE;
prot = PROT_NONE;
protchk = PROT_NONE;
(prot & ~PROT_USER) == PROT_NONE) {
ASSERT((prot & (PROT_READ | PROT_WRITE)) != PROT_NONE);
ASSERT((req_prot & (PROT_READ | PROT_WRITE | PROT_EXEC)) != PROT_NONE);