ql_alloc_phys
if ((rval = ql_alloc_phys(ha, mem, KM_SLEEP)) != QL_SUCCESS) {
if ((rval = ql_alloc_phys(ha, &ha->req_q[0]->req_ring, KM_SLEEP)) !=
if ((rval = ql_alloc_phys(ha, &ha->req_q[1]->req_ring,
if ((rval = ql_alloc_phys(ha, &ha->rcv_ring, KM_SLEEP)) !=
rval = ql_alloc_phys(ha, &rsp_q->rsp_ring, KM_SLEEP);
rval = ql_alloc_phys(ha, &ha->dmp_template, KM_SLEEP);
rval = ql_alloc_phys(ha, &ha->dmp_template, KM_SLEEP);
if ((rval = ql_alloc_phys(ha, &mem, KM_SLEEP)) != QL_SUCCESS) {
if (ql_alloc_phys(ha, &sp->sg_dma, KM_SLEEP) != QL_SUCCESS) {
if (ql_alloc_phys(qlge->dip, &qlge->ioctl_buf_dma_attr.dma_handle,
if (ql_alloc_phys(qlge->dip,
if (ql_alloc_phys(qlge->dip,
if (ql_alloc_phys(qlge->dip, &tx_ring->wq_dma.dma_handle,
if (ql_alloc_phys(qlge->dip,
if (ql_alloc_phys(qlge->dip, &tx_ring->wqicb_dma.dma_handle,
if (ql_alloc_phys(qlge->dip, &rx_ring->cq_dma.dma_handle,
if (ql_alloc_phys(qlge->dip, &rx_ring->sbq_dma.dma_handle,
if (ql_alloc_phys(qlge->dip, &rx_ring->lbq_dma.dma_handle,
if (ql_alloc_phys(qlge->dip, &rx_ring->cqicb_dma.dma_handle,
if (ql_alloc_phys(qlge->dip, &qlge->ricb_dma.dma_handle,
static int ql_alloc_phys(dev_info_t *, ddi_dma_handle_t *,
int ql_alloc_phys(ql_adapter_state_t *, dma_mem_t *, int);