mlx5_core_dma_dev
dma_free_coherent(mlx5_core_dma_dev(dev), PAGE_SIZE, buf->frags[i].buf,
dma_free_coherent(mlx5_core_dma_dev(dev), frag_sz, buf->frags[i].buf,
dma_free_coherent(mlx5_core_dma_dev(dev), PAGE_SIZE,
struct device *device = mlx5_core_dma_dev(dev);
dma_free_coherent(mlx5_core_dma_dev(dev), frag_sz,
cmd->cmd_alloc_buf = dma_alloc_coherent(mlx5_core_dma_dev(dev), MLX5_ADAPTER_PAGE_SIZE,
dma_free_coherent(mlx5_core_dma_dev(dev), MLX5_ADAPTER_PAGE_SIZE, cmd->cmd_alloc_buf,
cmd->cmd_alloc_buf = dma_alloc_coherent(mlx5_core_dma_dev(dev),
dma_free_coherent(mlx5_core_dma_dev(dev), cmd->alloc_size, cmd->cmd_alloc_buf,
cmd->pool = dma_pool_create("mlx5_cmd", mlx5_core_dma_dev(dev), size, align, 0);
ddev = mlx5_core_dma_dev(dev);
ddev = mlx5_core_dma_dev(dev);
struct device *ddev = mlx5_core_dma_dev(dev);
param->wq.buf_numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
param->wq.buf_numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
param->wq.buf_numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
node = dev_to_node(mlx5_core_dma_dev(mdev));
err = mlx5e_ptp_alloc_traffic_db(ptpsq, dev_to_node(mlx5_core_dma_dev(c->mdev)));
ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev));
ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev));
c = kvzalloc_node(sizeof(*c), GFP_KERNEL, dev_to_node(mlx5_core_dma_dev(mdev)));
c->pdev = mlx5_core_dma_dev(priv->mdev);
t->pdev = mlx5_core_dma_dev(priv->mdev);
struct device *dev = mlx5_core_dma_dev(mdev);
pdev = mlx5_core_dma_dev(mdev);
pdev = mlx5_core_dma_dev(mdev);
pdev = mlx5_core_dma_dev(sq->channel->mdev);
dev = mlx5_core_dma_dev(sq->channel->mdev);
dma_device = mlx5_core_dma_dev(mdev);
c->pdev = mlx5_core_dma_dev(mdev);
param->wq.buf_numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
param->wq.db_numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
node = dev_to_node(mlx5_core_dma_dev(mdev));
dma_device = mlx5_core_dma_dev(conn->fdev->mdev);
dma_device = mlx5_core_dma_dev(conn->fdev->mdev);
int numa_node = dev_to_node(mlx5_core_dma_dev(mdev));
priv->numa_node = dev_to_node(mlx5_core_dma_dev(dev));
dma_unmap_page(mlx5_core_dma_dev(dev), fwp->addr & MLX5_U64_4K_PAGE_MASK,
struct device *device = mlx5_core_dma_dev(dev);
int numa_node = dev_to_node(mlx5_core_dma_dev(ctx->mdev));
dma_device = mlx5_core_dma_dev(mdev);
dma_unmap_single(mlx5_core_dma_dev(mdev), mr->dma_addr, mr->size,