ring_to_dev
dma_unmap_single(ring_to_dev(ring), ring->desc_dma_addr,
ring->desc_dma_addr = dma_map_single(ring_to_dev(ring),
if (dma_mapping_error(ring_to_dev(ring), ring->desc_dma_addr)) {
cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0,
if (dma_mapping_error(ring_to_dev(ring), cb->dma))
dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length,
dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length,
struct device *dev = ring_to_dev(ring);
dev_warn(ring_to_dev(ring), "failed to allocate tx spare buffer, exceed to max order\n");
tx_spare = devm_kzalloc(ring_to_dev(ring), sizeof(*tx_spare),
dev_warn(ring_to_dev(ring), "failed to allocate hns3_tx_spare\n");
page = alloc_pages_node(dev_to_node(ring_to_dev(ring)),
dev_warn(ring_to_dev(ring), "failed to allocate tx spare pages\n");
dma = dma_map_page(ring_to_dev(ring), page, 0,
if (dma_mapping_error(ring_to_dev(ring), dma)) {
dev_warn(ring_to_dev(ring), "failed to map pages for tx spare\n");
devm_kfree(ring_to_dev(ring), tx_spare);
dma_sync_single_for_cpu(ring_to_dev(ring), dma, len,
dma_unmap_sg(ring_to_dev(ring), sgt->sgl, sgt->orig_nents,
struct device *dev = ring_to_dev(ring);
struct device *dev = ring_to_dev(ring);
dma_sync_single_for_device(ring_to_dev(ring), dma, size,
sgt->nents = dma_map_sg(ring_to_dev(ring), sgt->sgl, sgt->orig_nents,
cb->dma = dma_map_page(ring_to_dev(ring), cb->priv, 0,
if (unlikely(dma_mapping_error(ring_to_dev(ring), cb->dma)))
dma_unmap_single(ring_to_dev(ring), cb->dma, cb->length,
dma_unmap_page(ring_to_dev(ring), cb->dma, cb->length,
dma_free_coherent(ring_to_dev(ring), size,
ring->desc = dma_alloc_coherent(ring_to_dev(ring), size,
dma_sync_single_for_device(ring_to_dev(ring),
dma_sync_single_for_cpu(ring_to_dev(ring),
dma_sync_single_for_cpu(ring_to_dev(ring),
.nid = dev_to_node(ring_to_dev(ring)),
.dev = ring_to_dev(ring),
dev_warn(ring_to_dev(ring), "page pool creation failed: %ld\n",
ring->desc_cb = devm_kcalloc(ring_to_dev(ring), ring->desc_num,
devm_kfree(ring_to_dev(ring), ring->desc_cb);
devm_kfree(ring_to_dev(ring), ring->desc_cb);
dma_unmap_page(ring_to_dev(ring), tx_spare->dma, tx_spare->len,
devm_kfree(ring_to_dev(ring), tx_spare);