bxe_dma_alloc
if (bxe_dma_alloc(sc, sizeof(struct host_sp_status_block),
if (bxe_dma_alloc(sc, BCM_PAGE_SIZE,
if (bxe_dma_alloc(sc, sizeof(struct bxe_slowpath),
if (bxe_dma_alloc(sc, BCM_PAGE_SIZE,
if (bxe_dma_alloc(sc, FW_BUF_SIZE, &sc->gz_buf_dma,
if (bxe_dma_alloc(sc, sizeof(union bxe_host_hc_status_block),
if (bxe_dma_alloc(sc, (BCM_PAGE_SIZE * TX_BD_NUM_PAGES),
if (bxe_dma_alloc(sc, (BCM_PAGE_SIZE * RX_BD_NUM_PAGES),
if (bxe_dma_alloc(sc, (BCM_PAGE_SIZE * RCQ_NUM_PAGES),
if (bxe_dma_alloc(sc, (BCM_PAGE_SIZE * RX_SGE_NUM_PAGES),
if (bxe_dma_alloc(sc, sc->context[i].size,
if (bxe_dma_alloc(sc, (sc->fw_stats_req_size + sc->fw_stats_data_size),
int bxe_dma_alloc(struct bxe_softc *sc, bus_size_t size,
if (bxe_dma_alloc((struct bxe_softc *)sc, \