vm_map_max
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (end - 1 >= vm_map_max(kernel_map)) {
end = vm_map_max(kernel_map);
if (end - 1 >= vm_map_max(kernel_map)) {
end = vm_map_max(kernel_map);
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (minv >= maxv || minv + align >= maxv || maxv > vm_map_max(map)) {
vm_map_max(map) == sv->sv_maxuser &&
vm_map_remove(map, vm_map_min(map), vm_map_max(map));
kvm.kvm_max_user_addr = vm_map_max(&vmspace->vm_map);
maxaddr = vm_map_max(map);
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (addr - 1 >= vm_map_max(kernel_map))
addr = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
if (kernel_vm_end - 1 >= vm_map_max(kernel_map)) {
kernel_vm_end = vm_map_max(kernel_map);
parent_size = vm_map_max(parent_map) - vm_map_min(parent_map) +
if ((vm_offset_t)addr + len > vm_map_max(kernel_map) ||
if ((vm_offset_t)addr + len > vm_map_max(map) ||
physmem_est = lmin(physmem, btoc(vm_map_max(kernel_map) -
if (addr + size <= vm_map_max(map))
if (vm_map_max(map) - vm_map_min(map) < size) {
if (end > vm_map_max(map)) \
end = vm_map_max(map); \
if (start >= vm_map_max(map) || length > vm_map_max(map) - start)
return (vm_map_max(map) - length + 1);
return (vm_map_max(map) - length + 1);
if (addr + bytes < addr || addr + bytes > vm_map_max(kernel_map))
if (*addr + length > vm_map_max(map) ||
if (*addr + length > vm_map_max(map) ||
gap = vm_map_max(map) > MAP_32BIT_MAX_ADDR &&
vm_map_max(map))
if (*addr + length > vm_map_max(map) ||
vm_map_max(&vm->vm_map));
vm2 = vmspace_alloc(vm_map_min(old_map), vm_map_max(old_map),
return (vm_map_max(map));
if (start < vm_map_min(map) || end > vm_map_max(map))
error = vm_map_wire(map, vm_map_min(map), vm_map_max(map),
error = vm_map_unwire(map, vm_map_min(map), vm_map_max(map),
if (end > vm_map_max(map) || end < addr)
vm_map_max(map));
if (new > vm_map_max(map)) {