intr_vec_t
static intr_vec_t *
intr_vec_t *iv_p;
(intr_vec_t *)&intr_vecx_pool[i] : &intr_vec_pool[i];
sizeof (intr_vec_t), 64, NULL, NULL, NULL, NULL, NULL, 0);
bzero(iv_p, sizeof (intr_vec_t));
iv_free(intr_vec_t *iv_p)
sizeof (intr_vecx_t) : sizeof (intr_vec_t));
intr_vec_t *iv_p, *new_iv_p;
for (iv_p = (intr_vec_t *)intr_vec_table[inum];
new_iv_p->iv_vec_next = (intr_vec_t *)intr_vec_table[inum];
intr_vec_t *iv_p, *prev_iv_p;
for (iv_p = prev_iv_p = (intr_vec_t *)intr_vec_table[inum];
intr_vec_t *iv_p;
intr_vec_t *iv_p = (intr_vec_t *)softint_id;
intr_vec_t *list = softint_list;
intr_vec_t *iv_p = (intr_vec_t *)softint_id;
intr_vec_t *iv_p = (intr_vec_t *)softint_id;
intr_vec_t *softint_list = NULL;
intr_vec_t *intr_vec_pool = NULL; /* For HW and single target SW intrs */
intr_vec_pool = (intr_vec_t *)((caddr_t)intr_vec_table +
(MAXIVNUM * sizeof (intr_vec_t *)));
(MAX_RSVD_IV * sizeof (intr_vec_t)));
bzero(intr_vec_table, MAXIVNUM * sizeof (intr_vec_t *));
bzero(intr_vec_pool, MAX_RSVD_IV * sizeof (intr_vec_t));
intr_vec_t *iv, *next, *prev;
iv = (intr_vec_t *)inum;
intr_vec_t *next_iv = IV_GET_PIL_NEXT(next, cpu_id);
intr_vec_t iv_vec; /* CPU0 uses iv_pil_next */
intr_vec_t *iv_pil_xnext[NCPU -1]; /* For CPU1 through N-1 */
#define IVSIZE roundup(((MAXIVNUM * sizeof (intr_vec_t *)) + \
(MAX_RSVD_IV * sizeof (intr_vec_t)) + \
intr_vec_t *iv_vec_next; /* Per vector list */
intr_vec_t *iv_pil_next; /* Per PIL list */