mdev_to_gc
err = mana_gd_send_request(mdev_to_gc(mdev), sizeof(req), &req,
err = mana_gd_send_request(mdev_to_gc(mdev), sizeof(req), &req,
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
gc = mdev_to_gc(dev);
gc = mdev_to_gc(mdev);
gc = mdev_to_gc(mdev);
mana_gd_destroy_queue(mdev_to_gc(mdev), queue->kmem);
gc = mdev_to_gc(dev);
struct gdma_context *gc = mdev_to_gc(dev);
gc = mdev_to_gc(mdev);
struct pci_dev *pdev = to_pci_dev(mdev_to_gc(dev)->dev);
err = mana_gd_send_request(mdev_to_gc(dev), sizeof(req),
if (mdev_to_gc(dev)->pf_cap_flags1 & GDMA_DRV_CAP_FLAG_1_GDMA_PAGES_4MB_1GB_2GB)
err = mana_gd_send_request(mdev_to_gc(dev), sizeof(req), &req, sizeof(resp), &resp);
struct gdma_context *gc = mdev_to_gc(mdev);
gc = mdev_to_gc(dev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(mdev);
mana_gd_destroy_dma_region(mdev_to_gc(dev), dma_region_handle);
mana_gd_destroy_dma_region(mdev_to_gc(dev), dma_region_handle);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(dev);
struct gdma_context *gc = mdev_to_gc(mdev);
struct gdma_context *gc = mdev_to_gc(dev);
gc = mdev_to_gc(dev);
struct gdma_context *gc = mdev_to_gc(mdev);
mana_gd_wq_ring_doorbell(mdev_to_gc(mdev), queue);
mana_gd_wq_ring_doorbell(mdev_to_gc(mdev), queue);