SMR_SEQ_INVALID
goal = SMR_SEQ_INVALID;
c_seq = SMR_SEQ_INVALID;
if (c_seq == SMR_SEQ_INVALID)
if (c_seq != SMR_SEQ_INVALID)
c->c_seq = SMR_SEQ_INVALID;
KASSERT(inp->inp_smr == SMR_SEQ_INVALID,
if (inp->inp_smr != SMR_SEQ_INVALID) {
inp->inp_smr = SMR_SEQ_INVALID;
KASSERT(inp->inp_smr == SMR_SEQ_INVALID,
inp->inp_smr = SMR_SEQ_INVALID;
KASSERT(inp->inp_smr == SMR_SEQ_INVALID,
(curthread->td_critnest != 0 && zpcpu_get((smr))->c_seq != SMR_SEQ_INVALID)
KASSERT(smr->c_seq != SMR_SEQ_INVALID,
atomic_store_rel_int(&smr->c_seq, SMR_SEQ_INVALID);
KASSERT(smr->c_seq != SMR_SEQ_INVALID,
atomic_store_int(&smr->c_seq, SMR_SEQ_INVALID);
(seq = atomic_load_32(&zdom->uzd_seq)) != SMR_SEQ_INVALID &&
bucket->ub_seq != SMR_SEQ_INVALID) {
bucket->ub_seq = SMR_SEQ_INVALID;
seq = SMR_SEQ_INVALID;
seq = SMR_SEQ_INVALID;
bucket->ub_seq = SMR_SEQ_INVALID;
bucket->ub_seq = SMR_SEQ_INVALID;
KASSERT(bucket->ub_seq == SMR_SEQ_INVALID,
bucket->ub_seq != SMR_SEQ_INVALID) {
bucket->ub_seq = SMR_SEQ_INVALID;
if (bucket->ub_seq == SMR_SEQ_INVALID)
MPASS(b->ub_seq == SMR_SEQ_INVALID);