vm_map
kp->ki_size = vmspace.vm_map.size;
vmentry = vmspace.vm_map.header;
entryp != NULL && entryp != &kp->ki_vmspace->vm_map.header;
struct vm_map *map;
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
map = &p->p_vmspace->vm_map;
map = &curproc->p_vmspace->vm_map;
struct vm_map *map;
map = &vm_vmspace(vm)->vm_map;
vm_map_remove(&vmspace->vm_map, gpa, gpa + len);
vm_map_lock(&vmspace->vm_map);
error = vm_map_insert(&vmspace->vm_map, obj, 0, gpa, gpa + len,
vm_map_unlock(&vmspace->vm_map);
if (!p || !p->p_vmspace || !p->p_vmspace->vm_map.pmap ||
pmap = p->p_vmspace->vm_map.pmap;
struct vm_map *map;
map = &vm->vm_map;
struct vm_map *map;
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
struct vm_map *map;
map = &vm_vmspace(vm)->vm_map;
map = &vm->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
map = &vmspace->vm_map;
map = &curthread->td_proc->p_vmspace->vm_map;
map = &task->task_thread->td_proc->p_vmspace->vm_map;
map = &curthread->td_proc->p_vmspace->vm_map;
(p->p_vmspace != NULL) ? &p->p_vmspace->vm_map : 0);
(p->p_vmspace != NULL) ? &p->p_vmspace->vm_map.pmap : 0);
struct vm_map;
struct vm_map *db_map_addr(vm_offset_t);
bool db_map_current(struct vm_map *);
bool db_map_equal(struct vm_map *, struct vm_map *);
map = &vm->vm_map;
return (ps->vm != vm || ps->vm_timestamp != vm->vm_map.timestamp);
map = &vm->vm_map;
retcode = vm_mmap(&vms->vm_map, &virtual, map->size,
retcode = vm_mmap(&vms->vm_map, &virtual, dma->byte_count,
map = &vm->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
map = &curthread->td_proc->p_vmspace->vm_map;
err = vm_fault_hold_pages(&curproc->p_vmspace->vm_map,
map = &td->td_proc->p_vmspace->vm_map;
struct vm_map *vmmap;
vmmap = &mem->mem_vmspace->vm_map;
error = vm_map_remove(&vm_vmspace(vm)->vm_map, mm->gpa,
&vm_vmspace(vm)->vm_map, trunc_page(gpa),
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
map = &vm->vm_map;
map = &vm->vm_map;
i = vm_fault_quick_hold_pages(&curproc->p_vmspace->vm_map, uva, len,
map = usermode ? &p->p_vmspace->vm_map : kernel_map;
error = vm_map_find(&vmspace->vm_map, NULL, 0, &vmaddr,
error = vm_map_protect(&vmspace->vm_map, vmaddr,
error = vm_mmap(&vmspace->vm_map, &vmaddr,
error = vm_map_protect(&vmspace->vm_map, vmaddr + a_out->a_text,
error = vm_map_find(&vmspace->vm_map, NULL, 0, &vmaddr,
error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0,
map = &td->td_proc->p_vmspace->vm_map;
error = vm_map_find(&td->td_proc->p_vmspace->vm_map, NULL, 0,
map = &vmspace->vm_map;
map = &vmspace->vm_map;
vm_map_t map = &p->p_vmspace->vm_map;
map = &imgp->proc->p_vmspace->vm_map;
vm_map_init(&vmspace0.vm_map, vmspace_pmap(&vmspace0),
map = &vmspace->vm_map;
map = &vmspace->vm_map;
map = &vmspace->vm_map;
map = &cp->td->td_proc->p_vmspace->vm_map;
&curproc->p_vmspace->vm_map,
kp->ki_size = vm->vm_map.size;
map = &vm->vm_map;
map = &vm->vm_map;
kvm.kvm_min_user_addr = vm_map_min(&vmspace->vm_map);
kvm.kvm_max_user_addr = vm_map_max(&vmspace->vm_map);
if ((vmspace->vm_map.flags & MAP_WIREFUTURE) != 0)
if ((vmspace->vm_map.flags & MAP_ASLR) != 0)
if ((vmspace->vm_map.flags & MAP_ASLR_IGNSTART) != 0)
if ((vmspace->vm_map.flags & MAP_WXORX) != 0)
if ((vmspace->vm_map.flags & MAP_ASLR_STACK) != 0)
if ((vm->vm_map.flags & MAP_ASLR) != 0)
map = &vm->vm_map;
if ((vm->vm_map.flags & MAP_WXORX) != 0)
(void)vm_map_protect(&p->p_vmspace->vm_map,
*res = vm->vm_map.size;
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
error = vm_mmap_object(&vms->vm_map, addr, size, VM_PROT_READ |
map = &td->td_proc->p_vmspace->vm_map;
i = vm_fault_quick_hold_pages(&curproc->p_vmspace->vm_map,
p->p_pid, p->p_vmspace->vm_map.timestamp);
td->td_retval[0] = p->p_vmspace->vm_map.timestamp;
map = &p->p_vmspace->vm_map;
map = &vm->vm_map;
result = vm_map_remove(&vm->vm_map, shmmap_s->va, shmmap_s->va + size);
rv = vm_map_find(&p->p_vmspace->vm_map, shmseg->object, 0, &attach_va,
curmap = map == &vms->vm_map;
npages = vm_fault_quick_hold_pages(&curproc->p_vmspace->vm_map,
pidx = vm_fault_quick_hold_pages(&curproc->p_vmspace->vm_map,
cnt = vm_fault_quick_hold_pages(&td->td_proc->p_vmspace->vm_map,
zbuf_sfbuf_get(struct vm_map *map, vm_offset_t uaddr)
struct vm_map *map;
map = &td->td_proc->p_vmspace->vm_map;
&curproc->p_vmspace->vm_map, (vm_offset_t)sbp,
map = &p->p_vmspace->vm_map;
map = &p->p_vmspace->vm_map;
struct vm_map *map;
map = &p->p_vmspace->vm_map;
map = &p->p_vmspace->vm_map;
struct vm_map *map;
map = &vm_vmspace(vm)->vm_map;
&td->td_proc->p_vmspace->vm_map);
struct vm_map *map)
struct ucred *cred, struct vm_map *map);
void kern_proc_vmmap_resident(struct vm_map *map, struct vm_map_entry *entry,
bool cpu_exec_vmspace_reuse(struct proc *p, struct vm_map *map);
struct vm_map;
int kern_mmap_racct_check(struct thread *td, struct vm_map *map,
struct vm_map;
memguard_fudge(unsigned long km_size, const struct vm_map *parent_map)
struct vm_map;
unsigned long memguard_fudge(unsigned long, const struct vm_map *);
map = &vmspace->vm_map;
struct vm_map;
typedef struct vm_map *vm_map_t;
map = &curproc->p_vmspace->vm_map;
error = vm_map_wire(&curproc->p_vmspace->vm_map, start, end,
(void)vm_map_unwire(&curproc->p_vmspace->vm_map,
struct vm_map exec_map_store;
struct vm_map pipe_map_store;
struct vm_map kernel_map_store;
extern struct vm_map kernel_map_store;
extern struct vm_map exec_map_store;
extern struct vm_map pipe_map_store;
map = &vm->vm_map;
KASSERT(vm->vm_map.nentries == 0,
("vmspace %p nentries == %d on free", vm, vm->vm_map.nentries));
KASSERT(vm->vm_map.size == 0,
("vmspace %p size == %ju on free", vm, (uintmax_t)vm->vm_map.size));
KASSERT(vm->vm_map.pmap == NULL, ("vm_map.pmap must be NULL"));
_vm_map_init(&vm->vm_map, vmspace_pmap(vm), min, max);
(void)vm_map_remove(&vm->vm_map, vm_map_min(&vm->vm_map),
vm_map_max(&vm->vm_map));
vm->vm_map.pmap = NULL;
vm2->vm_map.size += entrysize;
old_map = &vm1->vm_map;
new_map = &vm2->vm_map;
if (p != initproc && (map != &p->p_vmspace->vm_map ||
vm_map_max_KBI(const struct vm_map *map)
vm_map_min_KBI(const struct vm_map *map)
(void *)p, (void *)p->p_vmspace, (void *)&p->p_vmspace->vm_map,
vm_map_print((vm_map_t)&p->p_vmspace->vm_map);
vm_map_max(const struct vm_map *map)
vm_map_min(const struct vm_map *map)
struct vm_map vm_map; /* VM address map */
vm_offset_t vm_map_max_KBI(const struct vm_map *map);
vm_offset_t vm_map_min_KBI(const struct vm_map *map);
struct vm_map *sub_map; /* belongs to another map */
map = &proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
error = vm_map_unwire(&td->td_proc->p_vmspace->vm_map, start, end,
map = &td->td_proc->p_vmspace->vm_map;
curmap = map == &td->td_proc->p_vmspace->vm_map;
if (!vm_map_range_valid(&vms->vm_map, addr, addr + size)) {
error = vm_mmap_object(&vms->vm_map, &addr, size, VM_PROT_NONE,
error = vm_mmap_object(&vms->vm_map, &addr, size, prot,
error = fo_mmap(fp, &vms->vm_map, &addr, size, prot,
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
vm_error = vm_map_protect(&td->td_proc->p_vmspace->vm_map,
switch (vm_map_inherit(&td->td_proc->p_vmspace->vm_map, addr,
map = &td->td_proc->p_vmspace->vm_map;
map = &td->td_proc->p_vmspace->vm_map;
if (_vm_object_in_map(&p->p_vmspace->vm_map, object, 0)) {
map = &vmspace->vm_map;
if (!vm_map_trylock_read(&vm->vm_map)) {
vm_map_unlock_read(&vm->vm_map);
&vm->vm_map, limit);
&vm->vm_map,
vm_map_t map = &vm->vm_map;