kernel_object
error = kmem_back(kernel_object, trampaddr, KINST_TRAMPCHUNK_SIZE,
kmem_unback(kernel_object, (vm_offset_t)chunk->addr,
if (page->object != kernel_object) {
kernel_object->domain.dr_policy = domainset2;
kernel_object->domain.dr_policy = cpuset_kernel->cs_domain;
if (kmem_back_domain(domain, kernel_object, addr, bytes,
rv = kmem_back(kernel_object, addr, size_p, flags);
kmem_unback(kernel_object, addr, size);
object = kernel_object;
object = kernel_object;
rv = kmem_back_domain(domain, kernel_object, addr, asize, flags);
KASSERT(object == kernel_object,
KASSERT(object == kernel_object,
KASSERT(object == kernel_object,
arena = _kmem_unback(kernel_object, (uintptr_t)addr, size);
KASSERT(object != kernel_object ||
error = kmem_back_domain(domain, kernel_object, addr, bytes, M_NOWAIT |
kmem_unback(kernel_object, addr, size);
object == kernel_object) {
object == kernel_object)) {
rw_init(&kernel_object->lock, "kernel vm object");
vm_radix_init(&kernel_object->rtree);
VM_MIN_KERNEL_ADDRESS), OBJ_UNMANAGED, kernel_object, NULL);
kernel_object->flags |= OBJ_COLORED;
kernel_object->pg_color = (u_short)atop(VM_MIN_KERNEL_ADDRESS);
kernel_object->un_pager.phys.ops = &default_phys_pg_ops;
KASSERT(object != kernel_object || rv->domain == domain,
KASSERT(object != kernel_object || rv->domain == domain,