SPA_MAXBLOCKSIZE
dnode_cache_buf = malloc(SPA_MAXBLOCKSIZE);
if (bsize > SPA_MAXBLOCKSIZE) {
"supported\n", SPA_MAXBLOCKSIZE);
unsigned int max_shifted = SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT;
static size_t max_data_size = SPA_MAXBLOCKSIZE;
1 << 17, (1 << 20) - (1 << 12), SPA_MAXBLOCKSIZE };
rand_data = (int *)umem_alloc(SPA_MAXBLOCKSIZE, UMEM_NOFAIL);
for (i = 0; i < SPA_MAXBLOCKSIZE / sizeof (int); i++)
mprotect((void *)rand_data, SPA_MAXBLOCKSIZE, PROT_READ);
umem_free(rand_data, SPA_MAXBLOCKSIZE);
char *buf = umem_alloc(SPA_MAXBLOCKSIZE, UMEM_NOFAIL);
len = SPA_MAXBLOCKSIZE;
umem_free(buf, SPA_MAXBLOCKSIZE);
umem_free(buf, SPA_MAXBLOCKSIZE);
pabd = abd_alloc_linear(SPA_MAXBLOCKSIZE, B_FALSE);
lbuf = umem_alloc(SPA_MAXBLOCKSIZE, UMEM_NOFAIL);
void *pbuf2 = umem_alloc(SPA_MAXBLOCKSIZE, UMEM_NOFAIL);
void *lbuf2 = umem_alloc(SPA_MAXBLOCKSIZE, UMEM_NOFAIL);
VERIFY0(abd_iterate_func(pabd, psize, SPA_MAXBLOCKSIZE - psize,
SPA_MAXBLOCKSIZE - psize));
for (lsize = SPA_MAXBLOCKSIZE; lsize > psize;
umem_free(pbuf2, SPA_MAXBLOCKSIZE);
umem_free(lbuf2, SPA_MAXBLOCKSIZE);
umem_free(lbuf, SPA_MAXBLOCKSIZE);
ASSERT3U(BPE_GET_LSIZE(&bp), <=, SPA_MAXBLOCKSIZE);
buf = malloc(SPA_MAXBLOCKSIZE);
0, MIN(lr->lr_length, (verbose < 6 ? 20 : SPA_MAXBLOCKSIZE)),
char *buf = safe_malloc(SPA_MAXBLOCKSIZE);
if (sz > SPA_MAXBLOCKSIZE) {
if (spa_maxblocksize(ztest_spa) == SPA_MAXBLOCKSIZE)
newsize = psize + MAX(psize / 8, SPA_MAXBLOCKSIZE);
stackbase += SPA_MAXBLOCKSIZE;
if (blksz > SPA_MAXBLOCKSIZE) {
int maxbs = SPA_MAXBLOCKSIZE;
char *buf = zfs_alloc(dda->dedup_hdl, SPA_MAXBLOCKSIZE);
if (sz > SPA_MAXBLOCKSIZE) {
SPA_MAXBLOCKSIZE, sz);
void *buf = zfs_alloc(hdl, SPA_MAXBLOCKSIZE);
assert(payload_size <= SPA_MAXBLOCKSIZE);
hdl->libzfs_max_nvlist = (SPA_MAXBLOCKSIZE * 4);
ASSERT3U(abd->abd_size, <=, SPA_MAXBLOCKSIZE);
VERIFY3U(size, <=, SPA_MAXBLOCKSIZE);
VERIFY3U(size, <=, SPA_MAXBLOCKSIZE);
VERIFY3U(size, <=, SPA_MAXBLOCKSIZE);
for (i = 0; i < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; i++) {
uint64_t overflow = MAX(SPA_MAXBLOCKSIZE,
uint64_t dmu_prefetch_max = 8 * SPA_MAXBLOCKSIZE;
ASSERT3U(len, <=, SPA_MAXBLOCKSIZE);
int zfs_recv_queue_length = SPA_MAXBLOCKSIZE;
int zfs_send_queue_length = SPA_MAXBLOCKSIZE;
ASSERT3U(size, <=, SPA_MAXBLOCKSIZE);
uint64_t metaslab_force_ganging = SPA_MAXBLOCKSIZE + 1; /* force gang blocks */
MIN(zfs_max_recordsize, SPA_MAXBLOCKSIZE), ZPROP_SRC_NONE);
return (SPA_MAXBLOCKSIZE);
CTASSERT(sizeof (l2arc_log_blk_phys_t) <= SPA_MAXBLOCKSIZE);
ASSERT3U(size, <=, SPA_MAXBLOCKSIZE);
uint64_t thismax = SPA_MAXBLOCKSIZE;
ASSERT3U(maxalloc, <=, SPA_MAXBLOCKSIZE);
ASSERT3U(size, <=, SPA_MAXBLOCKSIZE);
intval > SPA_MAXBLOCKSIZE)
zr.zr_lr = kmem_alloc(2 * SPA_MAXBLOCKSIZE, KM_SLEEP);
kmem_free(zr.zr_lr, 2 * SPA_MAXBLOCKSIZE);
for (c = 0; c < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; c++) {
for (c = 0; c < SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT; c++) {
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
VERIFY3U(c, <, SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT);
ASSERT((uintptr_t)zio->io_abd < SPA_MAXBLOCKSIZE);
IMPLY(type != ZIO_TYPE_TRIM, psize <= SPA_MAXBLOCKSIZE);
kmem_cache_t *zio_buf_cache[SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT];
kmem_cache_t *zio_data_buf_cache[SPA_MAXBLOCKSIZE >> SPA_MINBLOCKSHIFT];
if (BP_GET_LSIZE(bp) > SPA_MAXBLOCKSIZE) {
if (BP_GET_PSIZE(bp) > SPA_MAXBLOCKSIZE) {