SPA_MINBLOCKSIZE
#define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
"[n*%llu, (n+1)*%llu)", SPA_MINBLOCKSIZE,
SPA_MINBLOCKSIZE);
"[2^(n-1)*%llu, 2^n*%llu)", SPA_MINBLOCKSIZE,
SPA_MINBLOCKSIZE);
#define PSIZE_HISTO_SIZE (SPA_OLD_MAXBLOCKSIZE / SPA_MINBLOCKSIZE + 2)
idx = MIN(idx, SPA_OLD_MAXBLOCKSIZE / SPA_MINBLOCKSIZE + 1);
lsize -= SPA_MINBLOCKSIZE) {
lsize -= SPA_MINBLOCKSIZE;
if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) {
chunksize < (SPA_MINBLOCKSIZE * 2)) {
if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) {
if (i != 5 || chunksize < (SPA_MINBLOCKSIZE * 2)) {
MAX(SPA_MINBLOCKSIZE << 1, value);
if (size < SPA_MINBLOCKSIZE)
if (size < SPA_MINBLOCKSIZE)
#define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
if (intval < SPA_MINBLOCKSIZE ||
(intval < SPA_MINBLOCKSIZE ||
BP_GET_LSIZE(blkptr) : SPA_MINBLOCKSIZE;
blksz = SPA_MINBLOCKSIZE;
blksz = P2ROUNDUP(blksz, SPA_MINBLOCKSIZE);
*nblk512 = ((DN_USED_BYTES(dn->dn_phys) + SPA_MINBLOCKSIZE/2) >>
P2PHASE(drro->drr_blksz, SPA_MINBLOCKSIZE) ||
drro->drr_blksz < SPA_MINBLOCKSIZE ||
if (drrs->drr_length < SPA_MINBLOCKSIZE ||
size = SPA_MINBLOCKSIZE;
size = P2ROUNDUP(size, SPA_MINBLOCKSIZE);
ASSERT0(P2PHASE(size, SPA_MINBLOCKSIZE));
ASSERT3U(size, >=, SPA_MINBLOCKSIZE);
blocksize = P2ROUNDUP(blocksize, SPA_MINBLOCKSIZE);
ASSERT3U(blocksize, >=, SPA_MINBLOCKSIZE);
ASSERT0(blocksize % SPA_MINBLOCKSIZE);
SPA_MINBLOCKSIZE) == 0);
if (ret != 0 || blksz < SPA_MINBLOCKSIZE)
ASSERT(weight >= SPA_MINBLOCKSIZE || weight == 0);
} else if (mc->mc_alloc_groups == 0 || psize == SPA_MINBLOCKSIZE) {
if (space > 0 && space < SPA_MINBLOCKSIZE)
space = SPA_MINBLOCKSIZE;
ASSERT(size >= SPA_MINBLOCKSIZE ||
blocksize = SPA_MINBLOCKSIZE;
blocksize = P2ROUNDUP_TYPED(size, SPA_MINBLOCKSIZE, uint32_t);
if (vd->vdev_deflate_ratio != SPA_MINBLOCKSIZE)
uint64_t bufsz = MAX(sm->sm_blksz, SPA_MINBLOCKSIZE);
CTASSERT(sizeof (l2arc_dev_hdr_phys_t) == SPA_MINBLOCKSIZE);
CTASSERT(sizeof (l2arc_log_blk_phys_t) >= SPA_MINBLOCKSIZE);
#define FUID_SIZE_ESTIMATE(z) ((z)->z_fuid_size + (SPA_MINBLOCKSIZE << 1))
#define SPA_GANGBLOCKSIZE SPA_MINBLOCKSIZE
ASSERT((space & (SPA_MINBLOCKSIZE-1)) == 0);
*ashift = highbit64(MAX(pbsize, SPA_MINBLOCKSIZE)) - 1;
IS_P2ALIGNED(inner_size, 2 * SPA_MINBLOCKSIZE)) {
ta.trim_extent_bytes_min = SPA_MINBLOCKSIZE;
uint64_t newsz = db->db_size + SPA_MINBLOCKSIZE;
ASSERT3U(newval, >=, SPA_MINBLOCKSIZE);
*valp = (ulong_t)SPA_MINBLOCKSIZE;
if (size <= 4 * SPA_MINBLOCKSIZE) {
align = SPA_MINBLOCKSIZE;
SPA_MINBLOCKSIZE);
ASSERT(lsize >= SPA_MINBLOCKSIZE && lsize <= resid);
if (error == ENOSPC && zio->io_size > SPA_MINBLOCKSIZE)
ASSERT0(P2PHASE(zio->io_offset, SPA_MINBLOCKSIZE));
ASSERT0(P2PHASE(zio->io_size, SPA_MINBLOCKSIZE));
ASSERT(P2PHASE(psize, SPA_MINBLOCKSIZE) == 0);
ASSERT(P2PHASE(offset, SPA_MINBLOCKSIZE) == 0);
BP_SET_PSIZE(bp, SPA_MINBLOCKSIZE);
BP_SET_PSIZE(bp, SPA_MINBLOCKSIZE);
if (volblocksize < SPA_MINBLOCKSIZE ||