P2PHASE
if (P2PHASE(size, align) != 0)
#define BF32_DECODE(x, low, len) P2PHASE((x) >> (low), 1U << (len))
#define BF64_DECODE(x, low, len) P2PHASE((x) >> (low), 1ULL << (len))
#define BF32_ENCODE(x, low, len) (P2PHASE((x), 1U << (len)) << (low))
#define BF64_ENCODE(x, low, len) (P2PHASE((x), 1ULL << (len)) << (low))
P2PHASE(cp->cache_slabsize, vm_quantum) != 0 ||
P2PHASE(cp->cache_slabsize, vm_quantum) != 0 ||
return (P2PHASE(DATAOFF(pgnum), coreblksize) == 0);
P2PHASE(offset, length) == 0)
#define BF32_DECODE(x, low, len) P2PHASE((x) >> (low), 1U << (len))
#define BF64_DECODE(x, low, len) P2PHASE((x) >> (low), 1ULL << (len))
#define BF32_ENCODE(x, low, len) (P2PHASE((x), 1U << (len)) << (low))
#define BF64_ENCODE(x, low, len) (P2PHASE((x), 1ULL << (len)) << (low))
ASSERT(P2PHASE((uintptr_t)ret, align) == 0);
var = P2PHASE((var), ERR_SIZE); \
umem_error_begin = P2PHASE(umem_error_end + 1, ERR_SIZE);
NULL || P2PHASE((uintptr_t)bcp, UMEM_ALIGN) ||
ASSERT(P2PHASE((uintptr_t)slab, vmp->vm_quantum) == 0);
ASSERT(P2PHASE(phase, UMEM_ALIGN) == 0);
P2PHASE(bufsize, umem_lite_maxalign) != 0) {
if (P2PHASE(size, UMEM_ALIGN) != 0) {
ASSERT(P2PHASE(vs_start, vmp->vm_quantum) == 0);
ASSERT(P2PHASE(addr, vmp->vm_quantum) == 0);
ASSERT(P2PHASE(addr, align) == phase);
unaligned = P2PHASE((uintptr_t)dp->db_base, sizeof (uint_t));
poff = P2PHASE((int)foff, blksz);
base += P2PHASE(off, aslr_max_brk_skew);
blkoff = P2PHASE(i, epb);
blkoff = P2PHASE(i, bpo->bpo_epb);
blkoff = P2PHASE(i, epb);
blkoff = P2PHASE(bpo->bpo_phys->bpo_num_blkptrs, bpo->bpo_epb);
P2PHASE(nextblkid, 1ULL << dpa->dpa_epbs);
bp = bpp[P2PHASE(curblkid, 1 << epbs)];
if (P2PHASE(object, L1_dnode_count) == 0) {
if ((P2PHASE(object, dnodes_per_chunk) == 0) ||
(P2PHASE(object + dn_slots - 1, dnodes_per_chunk) <
ASSERT0(P2PHASE(os->os_obj_next_chunk,
P2PHASE(drro->drr_blksz, SPA_MINBLOCKSIZE) ||
P2PHASE(drror->drr_firstobj, DNODES_PER_BLOCK) != 0 ||
if (P2PHASE(off, dn->dn_datablksz) || len < dn->dn_datablksz) {
P2PHASE(off + len, dn->dn_datablksz)) {
blkoff = P2PHASE(off, blksz);
tail = P2PHASE(len, blksz);
ASSERT0(P2PHASE(off, blksz));
ASSERT0(P2PHASE(space, 1<<DEV_BSHIFT));
ASSERT0(P2PHASE(size, SPA_MINBLOCKSIZE));
ASSERT(P2PHASE(dn->dn_next_blksz[txgoff],
VERIFY0(P2PHASE(start, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(size, 1ULL << vd->vdev_ashift));
ASSERT(P2PHASE(asize, 1ULL << vd->vdev_ashift) == 0);
VERIFY0(P2PHASE(offset, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(asize, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(offset, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(size, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(offset, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(size, 1ULL << vd->vdev_ashift));
VERIFY0(P2PHASE(block_length, sizeof (uint64_t)));
VERIFY0(P2PHASE(entry_offset, 1ULL << sm->sm_shift));
VERIFY0(P2PHASE(entry_run, 1ULL << sm->sm_shift));
VERIFY0(P2PHASE(entry_offset, 1ULL << sm->sm_shift));
VERIFY0(P2PHASE(entry_run, 1ULL << sm->sm_shift));
ASSERT0(P2PHASE(end, sizeof (uint64_t)));
#define BF32_DECODE(x, low, len) P2PHASE((x) >> (low), 1U << (len))
#define BF64_DECODE(x, low, len) P2PHASE((x) >> (low), 1ULL << (len))
#define BF32_ENCODE(x, low, len) (P2PHASE((x), 1U << (len)) << (low))
#define BF64_ENCODE(x, low, len) (P2PHASE((x), 1ULL << (len)) << (low))
uint64_t cache_phase = P2PHASE(zio->io_offset, VCBS);
uint64_t cache_phase = P2PHASE(zio->io_offset, VCBS);
len = MIN(blocksize - P2PHASE(off, blocksize), resid);
P2PHASE(woff, max_blksz) == 0 &&
nbytes = MIN(n, max_blksz - P2PHASE(woff, max_blksz));
blkoff = ISP2(size) ? P2PHASE(offset, size) : offset;
preamble = P2PHASE(offset, blksz);
postamble = P2PHASE(size, blksz);
P2PHASE(uio->uio_loffset, zfs_read_chunk_size));
ASSERT0(P2PHASE(lwb->lwb_nused, sizeof (uint64_t)));
ASSERT0(P2PHASE(offset, 1ULL << vd->vdev_ashift));
ASSERT0(P2PHASE(size, 1ULL << vd->vdev_ashift));
P2PHASE(zio->io_size, align) != 0) {
ASSERT0(P2PHASE(zio->io_offset, align));
ASSERT0(P2PHASE(zio->io_size, align));
ASSERT0(P2PHASE(zio->io_offset, SPA_MINBLOCKSIZE));
ASSERT0(P2PHASE(zio->io_size, SPA_MINBLOCKSIZE));
ASSERT(P2PHASE(psize, SPA_MINBLOCKSIZE) == 0);
ASSERT(P2PHASE(offset, SPA_MINBLOCKSIZE) == 0);
len = MIN(blocksize - P2PHASE(off, blocksize), resid);
if (P2PHASE(offset, DEV_BSIZE) || P2PHASE(size, DEV_BSIZE) ||
(P2PHASE(len, (1U << shift)) != 0)) {
if ((P2PHASE(uio->uio_loffset, (1U << shift)) != 0) ||
(P2PHASE(uio->uio_iov->iov_len, (1U << shift)) != 0)) {
if ((P2PHASE(bp->b_lblkno, (1U << (shift - DEV_BSHIFT))) != 0) ||
(P2PHASE(bp->b_bcount, (1U << shift)) != 0) ||
if (P2PHASE(length, (1U << bd->d_blkshift)) != 0) {
(P2PHASE(bd->d_maxxfer, media.m_blksize))) {
(P2PHASE(len, (1U << xi->i_blkshift)) != 0)) {
if (P2PHASE(length, (1U << lsp->ls_lbshift)) != 0) {
unaligned = P2PHASE((uintptr_t)dp->db_base, sizeof (uint_t));
ASSERT(P2PHASE((uintptr_t)kaddr, bp_align) == color);
color = P2PHASE(base, bp_align);
off = P2PHASE(dumpbuf.vp_off, dumpbuf.size);
baseoff = P2PHASE(base, CBUF_MAPNP);
off = P2PHASE(pfn, CBUF_MAPNP);
off = P2PHASE(pfn, CBUF_MAPNP);
return (SA(P2PHASE(off, aslr_max_stack_skew)));
ASSERT(P2PHASE((uintptr_t)usrstack, args->stk_align) == 0);
ASSERT(P2PHASE(size, args->stk_align) == 0);
ASSERT(P2PHASE(sp_slew, args->stk_align) == 0);
ASSERT(P2PHASE(oldsize, PAGESIZE) == 0);
NULL || P2PHASE((uintptr_t)bcp, KMEM_ALIGN) ||
ASSERT(P2PHASE((uintptr_t)slab, vmp->vm_quantum) == 0);
P2PHASE(bufsize, kmem_lite_maxalign) != 0) {
ASSERT(P2PHASE(cache_size, table_unit) == 0);
aphase = (P2PHASE(phase, aquantum) != 0) ?
ASSERT(P2PHASE(asize,
ASSERT(P2PHASE(addr, align) == phase);
ASSERT(P2PHASE(vs_start, vmp->vm_quantum) == 0);
ASSERT(P2PHASE(addr, vmp->vm_quantum) == 0);
P2PHASE(segkmem_heaplp_quantum, segkmem_lpsize)) {
ASSERT(P2PHASE(segkmem_kmemlp_min,
if (anon_locked && P2PHASE(an_idx, amp_pgs) == 0) {
if (P2PHASE(pfn, curnpgs) == curnpgs - 1) {
ASSERT(P2PHASE(pfn, pages) == pgidx);
aphase = btop(P2PHASE((uintptr_t)a, ppgsz));
if (aphase != P2PHASE(pfn, ppages)) {
if (szc > 1 && (tszcvec = P2PHASE(szcvec, 1 << szc)) > 1) {
if (szc > 1 && (tszcvec = P2PHASE(szcvec, 1 << szc)) > 1) {
if (a == addr || P2PHASE(anon_index, anpgcnt) == 0) {
if (P2PHASE(anon_index, anpgcnt) == anpgcnt - 1) {
rem = P2PHASE(len, ((uint_t)BLOCKZEROALIGN));
if (P2PHASE((uintptr_t)addr ^ off, pgsz)) {
ASSERT(P2PHASE(lastpp->p_pagenum, npgs) == npgs - 1);
if (P2PHASE((uintptr_t)addr ^ off, pgsz)) {
szcvec |= P2PHASE(~disable_lpgs, (1 << i));
ttes2map = (MMU_PAGE_SIZE - P2PHASE(ra, MMU_PAGE_SIZE)) >> 3;
ASSERT(P2PHASE(pfn, szcpgcnt) == 0);
if (P2PHASE(base_ra_to_pa_pfn, szcpgcnt)) {
if (P2PHASE(mpo_mblock[i].ra_to_pa, szcpagesize))