SPA_MINBLOCKSHIFT
size_t bsize = fp->f_dnode.dn_datablkszsec << SPA_MINBLOCKSHIFT;
int bsize = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
int bsize = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
size_t size = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
int bsize = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
size_t size = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
int bsize = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
size_t size = dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
size = child_dir_zap.dn_datablkszsec << SPA_MINBLOCKSHIFT;
size = dir.dn_datablkszsec << SPA_MINBLOCKSHIFT;
bsize = vim->vim_dn->dn_datablkszsec << SPA_MINBLOCKSHIFT;
#define DN_MAX_LEVELS (2 + ((DN_MAX_OFFSET_SHIFT - SPA_MINBLOCKSHIFT) / \
(dnp)->dn_used : (dnp)->dn_used << SPA_MINBLOCKSHIFT)
#define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT)
BF64_GET_SB((vimep)->vimep_src, 0, 63, SPA_MINBLOCKSHIFT, 0)
BF64_SET_SB((vimep)->vimep_src, 0, 63, SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((dva)->dva_word[0], 0, SPA_ASIZEBITS, SPA_MINBLOCKSHIFT, 0)
SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0)
BF64_SET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((bp)->blk_prop, 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1))
0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1, x); \
BF64_GET_SB((bp)->blk_prop, 16, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((bp)->blk_prop, 16, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1, x)
unsigned int max_shifted = SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT;
MAX(SPA_MINBLOCKSHIFT, value));
sls->sls_sm_obj, 0, UINT64_MAX, SPA_MINBLOCKSHIFT));
dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT);
ashift = SPA_MINBLOCKSHIFT;
ashift = SPA_MINBLOCKSHIFT;
unsigned idx = BP_GET_PSIZE(bp) >> SPA_MINBLOCKSHIFT;
sls->sls_sm_obj, 0, UINT64_MAX, SPA_MINBLOCKSHIFT));
return (1 << (SPA_MINBLOCKSHIFT + block_shift));
.zo_ashift = SPA_MINBLOCKSHIFT,
return (SPA_MINBLOCKSHIFT + ztest_random(5));
sector += (VDEV_SKIP_SIZE >> SPA_MINBLOCKSHIFT);
uint64_t adjpl = (uint64_t)part_length << SPA_MINBLOCKSHIFT;
label) >> SPA_MINBLOCKSHIFT;
((VDEV_SKIP_SIZE + VDEV_PHYS_SIZE) >> SPA_MINBLOCKSHIFT),
blksz = DNODE->dn_datablkszsec << SPA_MINBLOCKSHIFT;
offset = (sector << SPA_MINBLOCKSHIFT) +
dn->dn_datablkszsec << SPA_MINBLOCKSHIFT);
size = zap_dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
uint32_t size = zap_dnode->dn_datablkszsec << SPA_MINBLOCKSHIFT;
blksz = mdn->dn_datablkszsec << SPA_MINBLOCKSHIFT;
((offset + VDEV_LABEL_START_SIZE) >> SPA_MINBLOCKSHIFT)
BF64_GET_SB((dva)->dva_word[0], 0, SPA_ASIZEBITS, SPA_MINBLOCKSHIFT, 0)
SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0)
BF64_SET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((bp)->blk_prop, 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((bp)->blk_prop, 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1, x)
BF64_GET_SB((bp)->blk_prop, 16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((bp)->blk_prop, 16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1, x)
#define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT)
int sectors = len >> SPA_MINBLOCKSHIFT;
SPA_MINBLOCKSHIFT;
for (i = 0; i < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; i++) {
uint64_t log_entries = write_sz >> SPA_MINBLOCKSHIFT;
SPA_MINBLOCKSHIFT) + dn->dn_num_slots;
(((uint64_t)dnp->dn_datablkszsec) << (SPA_MINBLOCKSHIFT + \
drro->drr_blksz = dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT;
(dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT), DMU_OBJECT_END) != 0)
record_size = dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT;
(((uint64_t)datablkszsec) << (SPA_MINBLOCKSHIFT + \
int blksz = dblkszsec << SPA_MINBLOCKSHIFT;
int blksz = dblkszsec << SPA_MINBLOCKSHIFT;
if (size >> SPA_MINBLOCKSHIFT == dn->dn_datablkszsec && ibs == 0)
ASSERT3U(dn->dn_datablkshift, >=, SPA_MINBLOCKSHIFT);
dn->dn_datablkszsec << SPA_MINBLOCKSHIFT);
ASSERT3U(size >> SPA_MINBLOCKSHIFT, <,
dn->dn_datablkszsec = size >> SPA_MINBLOCKSHIFT;
dnode_setdblksz(dn, dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT);
int zfs_default_bs = SPA_MINBLOCKSHIFT;
(dn->dn_phys->dn_datablkszsec << SPA_MINBLOCKSHIFT);
dnp->dn_datablkszsec << SPA_MINBLOCKSHIFT);
dn->dn_next_blksz[txgoff] >> SPA_MINBLOCKSHIFT ==
dn->dn_next_blksz[txgoff] >> SPA_MINBLOCKSHIFT;
int idx = MIN(shift - SPA_MINBLOCKSHIFT + i,
for (int i = RANGE_TREE_HISTOGRAM_SIZE - 1; i >= SPA_MINBLOCKSHIFT;
ASSERT3U(spa->spa_min_ashift, >=, SPA_MINBLOCKSHIFT);
sls->sls_sm_obj, 0, UINT64_MAX, SPA_MINBLOCKSHIFT);
0, UINT64_MAX, SPA_MINBLOCKSHIFT));
dsize = (asize >> SPA_MINBLOCKSHIFT) * vd->vdev_deflate_ratio;
ASSERT(IS_P2ALIGNED(x, 1U << SPA_MINBLOCKSHIFT)); \
(hdr)->b_lsize = ((x) >> SPA_MINBLOCKSHIFT); \
ASSERT(IS_P2ALIGNED((x), 1U << SPA_MINBLOCKSHIFT)); \
(hdr)->b_psize = ((x) >> SPA_MINBLOCKSHIFT); \
#define HDR_GET_LSIZE(hdr) ((hdr)->b_lsize << SPA_MINBLOCKSHIFT)
#define HDR_GET_PSIZE(hdr) ((hdr)->b_psize << SPA_MINBLOCKSHIFT)
1ULL << SPA_MINBLOCKSHIFT));
BF64_GET_SB((field), 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((field), 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1, x)
BF64_GET_SB((field), 16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((field), 16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1, x)
BF64_GET_SB((ddk)->ddk_prop, 0, 16, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((ddk)->ddk_prop, 0, 16, SPA_MINBLOCKSHIFT, 1, x)
BF64_GET_SB((ddk)->ddk_prop, 16, 16, SPA_MINBLOCKSHIFT, 1)
BF64_SET_SB((ddk)->ddk_prop, 16, 16, SPA_MINBLOCKSHIFT, 1, x)
#define DN_MAX_LEVELS (2 + ((DN_MAX_OFFSET_SHIFT - SPA_MINBLOCKSHIFT) / \
(dnp)->dn_used : (dnp)->dn_used << SPA_MINBLOCKSHIFT)
BF64_GET_SB((dva)->dva_word[0], 0, SPA_ASIZEBITS, SPA_MINBLOCKSHIFT, 0)
SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0)
BF64_SET_SB((dva)->dva_word[1], 0, 63, SPA_MINBLOCKSHIFT, 0, x)
BF64_GET_SB((bp)->blk_prop, 0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1))
0, SPA_LSIZEBITS, SPA_MINBLOCKSHIFT, 1, x); \
BF64_GET_SB((bp)->blk_prop, 16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1))
16, SPA_PSIZEBITS, SPA_MINBLOCKSHIFT, 1, x); \
#define SPA_MINBLOCKSIZE (1ULL << SPA_MINBLOCKSHIFT)
BF64_GET_SB((vimep)->vimep_src, 0, 63, SPA_MINBLOCKSHIFT, 0)
BF64_SET_SB((vimep)->vimep_src, 0, 63, SPA_MINBLOCKSHIFT, 0, x)
(vdev_psize_to_asize(vd, 1 << 17) >> SPA_MINBLOCKSHIFT);
return ((space >> SPA_MINBLOCKSHIFT) * vd->vdev_deflate_ratio);
*ashift = SPA_MINBLOCKSHIFT;
bench_rm = vdev_raidz_map_alloc(bench_zio, SPA_MINBLOCKSHIFT,
bench_rm = vdev_raidz_map_alloc(bench_zio, SPA_MINBLOCKSHIFT,
ASSERT3U(spa->spa_max_ashift, >=, SPA_MINBLOCKSHIFT);
ASSERT(leaf_blockshift >= SPA_MINBLOCKSHIFT &&
indirect_blockshift >= SPA_MINBLOCKSHIFT &&
statp->f_frsize = 1UL << SPA_MINBLOCKSHIFT;
statp->f_blocks = (refdbytes + availbytes) >> SPA_MINBLOCKSHIFT;
statp->f_bfree = availbytes >> SPA_MINBLOCKSHIFT;
for (c = 0; c < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; c++) {
size_t size = (c + 1) << SPA_MINBLOCKSHIFT;
ASSERT3U(spa->spa_min_ashift, >=, SPA_MINBLOCKSHIFT);
for (c = 0; c < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; c++) {
size_t c = (size - 1) >> SPA_MINBLOCKSHIFT;
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
size_t c = (size - 1) >> SPA_MINBLOCKSHIFT;
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
size_t c = (size - 1) >> SPA_MINBLOCKSHIFT;
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
size_t c = (size - 1) >> SPA_MINBLOCKSHIFT;
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
zb1obj = zb1L0 * (dbss1 << (SPA_MINBLOCKSHIFT - DNODE_SHIFT));
zb2obj = zb2L0 * (dbss2 << (SPA_MINBLOCKSHIFT - DNODE_SHIFT));
1ULL << (DNODE_BLOCK_SHIFT - SPA_MINBLOCKSHIFT), 0, &mod_zb,
kmem_cache_t *zio_buf_cache[SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT];
kmem_cache_t *zio_data_buf_cache[SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT];