BGE_SPLIT
dma_area_t buf[BGE_SPLIT]; /* (const) related */
dma_area_t buf[BGE_SEND_BUF_ARRAY][BGE_SPLIT];
dma_area_t tx_buff[BGE_SPLIT];
dma_area_t rx_buff[BGE_SPLIT];
ASSERT(brp->buf[0].nslots == nslots/BGE_SPLIT);
for (split = 0; split < BGE_SPLIT; ++split) {
for (slot = 0; slot < nslots/BGE_SPLIT; ++srbdp, ++slot)
for (split = 0; split < BGE_SPLIT; ++split) {
for (slot = 0; slot < BGE_SEND_BUF_NUM/BGE_SPLIT; ++slot) {
for (split = 0; split < BGE_SPLIT; ++split)
for (split = 0; split < BGE_SPLIT; ++split)
ASSERT((txbuffsize % BGE_SPLIT) == 0);
for (split = 0; split < BGE_SPLIT; ++split) {
err = bge_alloc_dma_mem(bgep, txbuffsize/BGE_SPLIT,
for (split = 0; split < BGE_SPLIT; ++split) {
for (slot = 0; slot < BGE_SEND_BUF_NUM/BGE_SPLIT; ++slot) {
ASSERT((rxbuffsize % BGE_SPLIT) == 0);
for (split = 0; split < BGE_SPLIT; ++split) {
err = bge_alloc_dma_mem(bgep, rxbuffsize/BGE_SPLIT,
rxbuffsize/BGE_SPLIT,
ASSERT((txbuffsize % BGE_SPLIT) == 0);
for (split = 0; split < BGE_SPLIT; ++split) {
err = bge_alloc_dma_mem(bgep, txbuffsize/BGE_SPLIT,
txbuffsize/BGE_SPLIT,
for (split = 0; split < BGE_SPLIT; ++split) {
&area, BGE_STD_SLOTS_USED/BGE_SPLIT,
BGE_STD_SLOTS_USED/BGE_SPLIT,
&area, bgep->chipid.jumbo_slots/BGE_SPLIT,
if ((bgep->chipid.jumbo_slots / BGE_SPLIT) > 0)
ASSERT(brp->buf[0].nslots == nslots/BGE_SPLIT);
bgep->chipid.jumbo_slots/BGE_SPLIT,
&area, BGE_MINI_SLOTS_USED/BGE_SPLIT,
if ((BGE_MINI_SLOTS_USED / BGE_SPLIT) > 0)
BGE_MINI_SLOTS_USED/BGE_SPLIT,
for (split = 0; split < BGE_SPLIT; ++split) {
&area, BGE_SEND_BUF_NUM/BGE_SPLIT,
BGE_SEND_BUF_NUM/BGE_SPLIT,
for (split = 0; split < BGE_SPLIT; ++split)
for (split = 0; split < BGE_SPLIT; ++split)