dma_pool_alloc
pp->sg_tbl[tag] = dma_pool_alloc(hpriv->sg_tbl_pool,
tpd = dma_pool_alloc(he_dev->tpd_pool, GFP_ATOMIC, &mapping);
heb = dma_pool_alloc(he_dev->rbpl_pool, GFP_ATOMIC, &mapping);
heb = dma_pool_alloc(he_dev->rbpl_pool, GFP_KERNEL, &mapping);
dma_pool_alloc(mlli_params->curr_pool, flags,
req_ctx->hmac_virt = dma_pool_alloc(buffer_pool, flags, &dma);
dir->npe_ctx = dma_pool_alloc(ctx_pool, GFP_KERNEL, &dir->npe_ctx_phys);
pad = dma_pool_alloc(ctx_pool, GFP_KERNEL, &pad_phys);
buf = dma_pool_alloc(buffer_pool, GFP_KERNEL, &buf_phys);
next_buf = dma_pool_alloc(buffer_pool, flags, &next_buf_phys);
req->cache = dma_pool_alloc(cesa_dev->dma->cache_pool, flags,
req->padding = dma_pool_alloc(cesa_dev->dma->padding_pool, flags,
op = dma_pool_alloc(cesa_dev->dma->op_pool, flags, &dma_handle);
ctx->sc = dma_pool_alloc(data->sc_pool, GFP_KERNEL, &ctx->sc_phys);
txd->llis_va = dma_pool_alloc(pl08x->pool, GFP_NOWAIT, &txd->llis_bus);
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
vaddr = dma_pool_alloc(atdma->memset_pool, GFP_NOWAIT, &paddr);
vaddr = dma_pool_alloc(atdma->memset_pool, GFP_NOWAIT, &paddr);
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool,
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool,
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_ATOMIC,
atdma_sg->lli = dma_pool_alloc(atdma->lli_pool, GFP_NOWAIT,
cb_entry->cb = dma_pool_alloc(c->cb_pool, gfp,
desc->desc = dma_pool_alloc(jzchan->desc_pool, GFP_NOWAIT,
dma_pool_alloc(dpaa2_chan->fd_pool, GFP_NOWAIT,
dma_pool_alloc(dpaa2_chan->fl_pool, GFP_NOWAIT,
dma_pool_alloc(dpaa2_chan->sdd_pool, GFP_NOWAIT,
fsl_desc->tcd[i].vtcd = dma_pool_alloc(fsl_chan->tcd_pool,
dma_pool_alloc(queue->comp_pool, GFP_KERNEL,
dma_pool_alloc(queue->desc_pool, GFP_KERNEL,
cf = dma_pool_alloc(re_chan->re_dev->cf_desc_pool, GFP_NOWAIT,
cf = dma_pool_alloc(re_chan->re_dev->cf_desc_pool, GFP_KERNEL,
chan->inb_ring_virt_addr = dma_pool_alloc(chan->re_dev->hw_desc_pool,
chan->oub_ring_virt_addr = dma_pool_alloc(chan->re_dev->hw_desc_pool,
hw->lli = dma_pool_alloc(idma64c->pool, GFP_NOWAIT, &hw->llp);
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT, &curr_phys);
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT,
curr = dma_pool_alloc(mdma->desc_pool, GFP_NOWAIT,
sed->hw = dma_pool_alloc(ioat_dma->sed_hw_pool[hw_pool],
ldma_sg->hw = dma_pool_alloc(lchan->pool, GFP_NOWAIT, &ldma_sg->llp);
ldma_sg->hw = dma_pool_alloc(lchan->pool, GFP_NOWAIT, &ldma_sg->llp);
lli = dma_pool_alloc(od->lli_pool, GFP_NOWAIT, &phys);
desc = dma_pool_alloc(sw_desc->desc_pool, GFP_NOWAIT, &dma);
fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
dma_pool_alloc(chan->desc_pool, GFP_NOWAIT,
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
v_lli = dma_pool_alloc(sdev->pool, GFP_DMA32 | GFP_NOWAIT, &p_lli);
osg->t2_desc = dma_pool_alloc(od->desc_pool, GFP_ATOMIC,
addr = dma_pool_alloc(chan->desc_pool, GFP_NOWAIT, &dma_addr);
dscr = dma_pool_alloc(dc->dscrpool, GFP_KERNEL, &dscr_dma);
dma_pool_alloc(hr_dev->cmd.pool, GFP_KERNEL, &mailbox->dma);
mailbox->buf = dma_pool_alloc(dev->cmd.pool, gfp_mask, &mailbox->dma);
ring->bd_base = dma_pool_alloc(ring->mbox->bd_pool,
req_addr = dma_pool_alloc(bd->hwrm_dma_pool, GFP_KERNEL | __GFP_ZERO,
req_addr = dma_pool_alloc(bp->hwrm_dma_pool, GFP_KERNEL | __GFP_ZERO,
descriptor = dma_pool_alloc(priv->adminq_pool, GFP_KERNEL,
descriptor = dma_pool_alloc(priv->adminq_pool, GFP_KERNEL, &descriptor_bus);
descriptor = dma_pool_alloc(priv->adminq_pool, GFP_KERNEL, &descriptor_bus);
priv->adminq = dma_pool_alloc(priv->adminq_pool, GFP_KERNEL,
cmdq_buf->buf = dma_pool_alloc(cmdqs->cmdq_buf_pool, GFP_KERNEL,
cmd_buf->buf = dma_pool_alloc(cmdqs->cmd_buf_pool, GFP_ATOMIC,
ddp->udl = dma_pool_alloc(ddp_pool->pool, GFP_ATOMIC, &ddp->udp);
desc = dma_pool_alloc(sdma->desc_pool, GFP_DMA | GFP_KERNEL, &dma);
virts[total_nr] = dma_pool_alloc(pool, GFP_KERNEL,
txq->bc_tbl.addr = dma_pool_alloc(trans_pcie->txqs.bc_pool, GFP_KERNEL,
prp_list = dma_pool_alloc(pool, GFP_ATOMIC, &prp_dma);
prp_list = dma_pool_alloc(pool, GFP_ATOMIC, &prp_dma);
prp_list = dma_pool_alloc(nvme_dma_pool(nvmeq, iod), GFP_ATOMIC,
prp_list = dma_pool_alloc(nvmeq->descriptor_pools.large,
sg_list = dma_pool_alloc(nvme_dma_pool(nvmeq, iod), GFP_ATOMIC,
sg_list = dma_pool_alloc(nvmeq->descriptor_pools.small, GFP_ATOMIC,
sgblkp = dma_pool_alloc(boardp->adv_sgblk_pool, GFP_ATOMIC,
seq->next_scb.vaddr = dma_pool_alloc(asd_ha->scb_pool, GFP_KERNEL,
io_task->cmd_bhs = dma_pool_alloc(beiscsi_sess->bhs_pool,
dma_buf->vaddr = dma_pool_alloc(hw->scsi_dma_pool, GFP_KERNEL,
evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC,
pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool,
ret = dma_pool_alloc(phba->lpfc_mbuf_pool, GFP_KERNEL, handle);
ret = dma_pool_alloc(phba->lpfc_sg_dma_buf_pool, GFP_KERNEL, handle);
hbqbp->dbuf.virt = dma_pool_alloc(phba->lpfc_hbq_pool, GFP_KERNEL,
dma_buf->hbuf.virt = dma_pool_alloc(phba->lpfc_hrb_pool, GFP_KERNEL,
dma_buf->dbuf.virt = dma_pool_alloc(phba->lpfc_drb_pool, GFP_KERNEL,
dma_buf->hbuf.virt = dma_pool_alloc(phba->lpfc_hrb_pool, GFP_KERNEL,
dma_buf->dbuf.virt = dma_pool_alloc(phba->lpfc_nvmet_drb_pool,
pcmd->virt = dma_pool_alloc(phba->lpfc_drb_pool, GFP_KERNEL,
tmp->dma_sgl = dma_pool_alloc(phba->lpfc_sg_dma_buf_pool,
mbox_pci_blk[i].vaddr = dma_pool_alloc(
epthru_pci_blk[i].vaddr = dma_pool_alloc(
sg_pci_blk[i].vaddr = dma_pool_alloc(
pool->vaddr = dma_pool_alloc(pool->handle, GFP_KERNEL,
kioc->buf_vaddr = dma_pool_alloc(pool->handle, GFP_ATOMIC,
kioc->pthru32 = dma_pool_alloc(adapter->pthru_dma_pool,
cmd->sense = dma_pool_alloc(instance->sense_dma_pool,
fusion->sense = dma_pool_alloc(fusion->sense_dma_pool,
fusion->sense = dma_pool_alloc(fusion->sense_dma_pool,
cmd->sg_frame = dma_pool_alloc(fusion->sg_dma_pool,
dma_pool_alloc(fusion->io_request_frames_pool,
dma_pool_alloc(fusion->io_request_frames_pool,
dma_pool_alloc(fusion->reply_frames_desc_pool,
dma_pool_alloc(fusion->reply_frames_desc_pool,
dma_pool_alloc(fusion->reply_frames_desc_pool,
dma_pool_alloc(fusion->reply_frames_desc_pool_align,
dma_pool_alloc(ioc->pcie_sgl_dma_pool, GFP_KERNEL,
ctr->chain_buffer = dma_pool_alloc(ioc->chain_dma_pool,
ioc->sense = dma_pool_alloc(ioc->sense_dma_pool,
ioc->reply = dma_pool_alloc(ioc->reply_dma_pool, GFP_KERNEL,
ioc->reply_free = dma_pool_alloc(ioc->reply_free_dma_pool,
dma_pool_alloc(ioc->reply_post_free_array_dma_pool,
dcdb = dma_pool_alloc(cb->dcdb_pool, GFP_ATOMIC, &dcdb_addr);
hw_sgl = dma_pool_alloc(cb->sg_pool, GFP_ATOMIC, &hw_sgl_addr);
cmd_blk->sense = dma_pool_alloc(cs->sense_pool, GFP_ATOMIC,
cmd_blk->dcdb = dma_pool_alloc(cs->dcdb_pool, GFP_ATOMIC,
hw_sgl = dma_pool_alloc(cs->sg_pool, GFP_ATOMIC,
void *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
uint8_t *sfp = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &sfp_dma);
sp->remap.req.buf = dma_pool_alloc(ha->purex_dma_pool,
sp->remap.rsp.buf = dma_pool_alloc(ha->purex_dma_pool,
dsd_ptr->dsd_addr = dma_pool_alloc
dma_pool_alloc(ha->dl_dma_pool, GFP_ATOMIC,
dma_pool_alloc(ha->dl_dma_pool, GFP_ATOMIC,
dsd_ptr->dsd_addr = dma_pool_alloc(ha->dl_dma_pool,
dsd_ptr->dsd_addr = dma_pool_alloc(ha->dl_dma_pool,
dma_pool_alloc(ha->dl_dma_pool, GFP_ATOMIC,
dma_pool_alloc(ha->dl_dma_pool, GFP_ATOMIC,
mn = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma);
str = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL, &str_dma);
dsd->dsd_addr = dma_pool_alloc(
ha->ms_iocb = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL,
ha->ex_init_cb = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL,
ha->async_pd = dma_pool_alloc(ha->s_dma_pool, GFP_KERNEL,
fw_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
flash_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
fw_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
fw_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
fw_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
fw_ddb_entry = dma_pool_alloc(ha->fw_ddb_dma_pool, GFP_KERNEL,
virt_cmd_desc = dma_pool_alloc(ctrl->dma_cmd_pool, GFP_ATOMIC | __GFP_ZERO, &dma_cmd_desc);
pkg->buffer = dma_pool_alloc(ctl->frame_pool, GFP_KERNEL,
priv_ep->trb_pool = dma_pool_alloc(priv_dev->eps_dma_pool,
return dma_pool_alloc(hcd->pool[i], mem_flags, dma);
qtd = dma_pool_alloc(fotg210->qtd_pool, flags, &dma);
itd = dma_pool_alloc(fotg210->itd_pool, mem_flags,
ring->trbs = dma_pool_alloc(pdev->eps_dma_pool,
dtd = dma_pool_alloc(udc_controller->td_pool, gfp_flags, dma);
dd = dma_pool_alloc(udc->dd_cache, GFP_ATOMIC | GFP_DMA, &dma);
td = dma_pool_alloc(dev->requests, GFP_KERNEL,
td = dma_pool_alloc(ep->dev->requests, gfp_flags,
td = dma_pool_alloc(ep->dev->data_requests, gfp_flags,
dma_desc = dma_pool_alloc(ep->dev->data_requests, gfp,
td_stp = dma_pool_alloc(dev->stp_requests, GFP_KERNEL,
td_data = dma_pool_alloc(dev->data_requests, GFP_KERNEL,
td_stp = dma_pool_alloc(dev->stp_requests, GFP_KERNEL,
td_data = dma_pool_alloc(dev->stp_requests, GFP_KERNEL,
dma_desc = dma_pool_alloc(ep->dev->data_requests, gfp,
td = dma_pool_alloc(ep->dev->data_requests,
ep->transfer_ring = dma_pool_alloc(xudc->transfer_ring_pool,
qtd = dma_pool_alloc (ehci->qtd_pool, flags, &dma);
itd = dma_pool_alloc(ehci->itd_pool, mem_flags,
sitd = dma_pool_alloc(ehci->sitd_pool, mem_flags,
td = dma_pool_alloc(uhci->td_pool, GFP_ATOMIC, &dma_handle);
void *dma_pool_alloc(struct dma_pool *pool, gfp_t mem_flags,
return dma_pool_alloc(pool, mem_flags | __GFP_ZERO, handle);
EXPORT_SYMBOL(dma_pool_alloc);
p[i].v = dma_pool_alloc(pool, GFP_KERNEL,