VMEM_ALLOC
if (vsp->vs_type == VMEM_ALLOC && (void *)vsp->vs_start == *target) {
if (vs->vs_type != VMEM_ALLOC && vs->vs_type != VMEM_FREE)
(vs->vs_type == VMEM_ALLOC && vs->vs_depth != 0))) {
(vs->vs_type == VMEM_ALLOC) ? "allocated" : "freed",
return (vmem_seg_walk_common_init(wsp, VMEM_ALLOC, "alloc"));
t = VMEM_ALLOC;
no_debug = (t != VMEM_ALLOC) ||
type = (t == VMEM_ALLOC ? "ALLC" :
if (vs->vs_type != VMEM_ALLOC && vs->vs_type != VMEM_FREE)
(vs->vs_type == VMEM_ALLOC && vs->vs_depth != 0))) {
(vs->vs_type == VMEM_ALLOC) ? "allocated" : "freed",
return (vmem_seg_walk_common_init(wsp, VMEM_ALLOC, "alloc"));
t = VMEM_ALLOC;
no_debug = (t != VMEM_ALLOC) ||
type = (t == VMEM_ALLOC ? "ALLC" :
if (typemask & VMEM_ALLOC)
leaked = vmem_size(vmp, VMEM_ALLOC);
vsp->vs_type = VMEM_ALLOC;
ASSERT(vsp->vs_type == VMEM_ALLOC &&
vmem_size(heap_lp_arena, VMEM_ALLOC) < size) {
vmem_walk(heap_lp_arena, VMEM_ALLOC, cpr_xwalk, arg);
vmem_walk(heap_arena, VMEM_ALLOC, cpr_walk, &cwinfo);
VMEM_ALLOC | VMEM_FREE) + 1), NM_INOQUANT, VM_SLEEP);
(vmem_size(zio_arena, VMEM_ALLOC) >>
vmem_size(heap_arena, VMEM_ALLOC | VMEM_FREE) / 2);
vmem_size(cur->nse_port_arena, VMEM_ALLOC) != 0) {
vmem_walk(tcl_minor_arena, VMEM_ALLOC, tclvm_backenable,
vmem_walk(tcl_minor_arena, VMEM_ALLOC,
vmem_walk(tcl_minor_arena, VMEM_ALLOC, tclvm_remove_tll, tll);
vmem_walk(ksyms_arena, VMEM_ALLOC, ksyms_walk_one, kwp);
if (typemask & VMEM_ALLOC)
leaked = vmem_size(vmp, VMEM_ALLOC);
vsp->vs_type = VMEM_ALLOC;
ASSERT(vsp->vs_type == VMEM_ALLOC &&
if (vmem_size(vmp, VMEM_ALLOC) >= segkmem_kmemlp_max) {
vmem_size(heap_lp_arena, VMEM_ALLOC) < size) {
vmem_walk(heap_lp_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(heap_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(heaptext_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(heap_core_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(heap32_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(heaptext_arena, VMEM_ALLOC | VMEM_REENTRANT,
vmem_walk(domain->dom_dvma_arena, VMEM_ALLOC, immu_fault_walk,