range_tree_space
if (range_tree_space(rt) == 0)
ASSERT0(range_tree_space(svr->svr_allocd_segs));
ASSERT0(range_tree_space(allocs));
zcb->zcb_removing_size += range_tree_space(svr->svr_allocd_segs);
int free_pct = range_tree_space(rt) * 100 / msp->ms_size;
int free_pct = range_tree_space(rt) * 100 / msp->ms_size;
range_tree_space(msp->ms_unflushed_frees));
range_tree_space(msp->ms_unflushed_allocs) -
range_tree_space(msp->ms_unflushed_frees));
range_tree_space(msp->ms_allocating[(txg + t) & TXG_MASK]);
range_tree_space(msp->ms_defer[0]) +
range_tree_space(msp->ms_defer[1]));
msp_free_space = range_tree_space(msp->ms_allocatable) + allocating +
msp->ms_deferspace + range_tree_space(msp->ms_freed);
range_tree_space(msp->ms_unflushed_allocs),
range_tree_space(msp->ms_unflushed_frees),
range_tree_space(msp->ms_freed),
range_tree_space(msp->ms_defer[0]),
range_tree_space(msp->ms_defer[1]),
ASSERT0(range_tree_space(msp->ms_allocatable));
range_tree_space(msp->ms_unflushed_allocs),
range_tree_space(msp->ms_unflushed_frees),
msp->ms_allocated_space += range_tree_space(alloctree);
range_tree_space(msp->ms_freeing));
msp->ms_allocated_space -= range_tree_space(msp->ms_freeing);
range_tree_space(msp->ms_checkpointing);
range_tree_space(msp->ms_checkpointing);
msp->ms_allocated_this_txg += range_tree_space(alloctree);
ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK]));
ASSERT0(range_tree_space(msp->ms_allocating[TXG_CLEAN(txg)
ASSERT0(range_tree_space(msp->ms_freeing));
ASSERT0(range_tree_space(msp->ms_checkpointing));
VERIFY0(range_tree_space(
ASSERT0(range_tree_space(msp->ms_freeing));
ASSERT0(range_tree_space(msp->ms_checkpointing));
range_tree_space(msp->ms_freed);
defer_delta = range_tree_space(msp->ms_freed) -
range_tree_space(*defer_tree);
defer_delta -= range_tree_space(*defer_tree);
VERIFY0(range_tree_space(
ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK]));
ASSERT0(range_tree_space(msp->ms_freeing));
ASSERT0(range_tree_space(msp->ms_freed));
ASSERT0(range_tree_space(msp->ms_checkpointing));
VERIFY3U(range_tree_space(rt) - size, <=, msp->ms_size);
VERIFY3U(range_tree_space(msp->ms_allocatable) + size, <=,
VERIFY3U(range_tree_space(msp->ms_allocatable) - size, <=,
ASSERT0(range_tree_space(*rtdst));
return (range_tree_space(rt) == 0);
ASSERT0(range_tree_space(vd->vdev_obsolete_segments));
range_tree_space(m->ms_unflushed_allocs) -
range_tree_space(m->ms_unflushed_frees);
range_tree_space(m->ms_unflushed_allocs), 0, 0);
-range_tree_space(m->ms_unflushed_frees), 0, 0);
VERIFY3U(range_tree_space(smla->smla_rt) + sme->sme_run, <=,
VERIFY0(range_tree_space(rt));
sm->sm_phys->smp_alloc += range_tree_space(rt);
sm->sm_phys->smp_alloc -= range_tree_space(rt);
uint64_t rt_space = range_tree_space(rt);
VERIFY3U(range_tree_space(rt), ==, rt_space);
uint64_t range_tree_space(range_tree_t *rt);
ASSERT3U(range_tree_space(vd->vdev_dtl[DTL_MISSING]), !=, 0);
ASSERT3U(range_tree_space(vd->vdev_dtl[DTL_MISSING]), !=, 0);
if (range_tree_space(vd->vdev_obsolete_segments) > 0) {
ASSERT(range_tree_space(vd->vdev_obsolete_segments) > 0);
svr->svr_bytes_done[txg & TXG_MASK] += range_tree_space(segs);
ASSERT0(range_tree_space(svr->svr_allocd_segs));
ASSERT0(range_tree_space(msp->ms_allocating[i]));
ASSERT0(range_tree_space(svr->svr_allocd_segs));
ASSERT0(range_tree_space(svr->svr_allocd_segs));
ASSERT0(range_tree_space(msp->ms_allocating[i]));
ASSERT0(range_tree_space(msp->ms_defer[i]));
ASSERT0(range_tree_space(msp->ms_freed));
range_tree_space(ms->ms_freeing);
ASSERT0(range_tree_space(ms->ms_freed));
ASSERT0(range_tree_space(ms->ms_allocating[t]));
entry->vime_obsolete_count = range_tree_space(obsolete_segs);