#include <subdev/mmu/vmm.h>
#include <nvhw/drf.h>
#include "nvrm/vmm.h"
void
r535_mmu_vaspace_del(struct nvkm_vmm *vmm)
{
if (vmm->rm.external) {
NV0080_CTRL_DMA_UNSET_PAGE_DIRECTORY_PARAMS *ctrl;
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
NV0080_CTRL_CMD_DMA_UNSET_PAGE_DIRECTORY,
sizeof(*ctrl));
if (!IS_ERR(ctrl)) {
ctrl->hVASpace = vmm->rm.object.handle;
WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl));
}
vmm->rm.external = false;
}
nvkm_gsp_rm_free(&vmm->rm.object);
nvkm_gsp_device_dtor(&vmm->rm.device);
nvkm_gsp_client_dtor(&vmm->rm.client);
nvkm_vmm_put(vmm, &vmm->rm.rsvd);
}
int
r535_mmu_vaspace_new(struct nvkm_vmm *vmm, u32 handle, bool external)
{
NV_VASPACE_ALLOCATION_PARAMETERS *args;
int ret;
ret = nvkm_gsp_client_device_ctor(vmm->mmu->subdev.device->gsp,
&vmm->rm.client, &vmm->rm.device);
if (ret)
return ret;
args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, handle, FERMI_VASPACE_A,
sizeof(*args), &vmm->rm.object);
if (IS_ERR(args))
return PTR_ERR(args);
args->index = NV_VASPACE_ALLOCATION_INDEX_GPU_NEW;
if (external)
args->flags = NV_VASPACE_ALLOCATION_FLAGS_IS_EXTERNALLY_OWNED;
ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args);
if (ret)
return ret;
if (!external) {
NV90F1_CTRL_VASPACE_COPY_SERVER_RESERVED_PDES_PARAMS *ctrl;
u8 page_shift = 29;
const u64 page_size = BIT_ULL(page_shift);
const struct nvkm_vmm_page *page;
const struct nvkm_vmm_desc *desc;
struct nvkm_vmm_pt *pd = vmm->pd;
for (page = vmm->func->page; page->shift; page++) {
if (page->shift == page_shift)
break;
}
if (WARN_ON(!page->shift))
return -EINVAL;
mutex_lock(&vmm->mutex.vmm);
ret = nvkm_vmm_get_locked(vmm, true, false, false, page_shift, 32, page_size,
&vmm->rm.rsvd);
mutex_unlock(&vmm->mutex.vmm);
if (ret)
return ret;
if (WARN_ON(vmm->rm.rsvd->addr != SPLIT_VAS_SERVER_RM_MANAGED_VA_START ||
vmm->rm.rsvd->size != SPLIT_VAS_SERVER_RM_MANAGED_VA_SIZE))
return -EINVAL;
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object,
NV90F1_CTRL_CMD_VASPACE_COPY_SERVER_RESERVED_PDES,
sizeof(*ctrl));
if (IS_ERR(ctrl))
return PTR_ERR(ctrl);
ctrl->pageSize = page_size;
ctrl->virtAddrLo = vmm->rm.rsvd->addr;
ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1;
for (desc = page->desc; desc->bits; desc++) {
ctrl->numLevelsToCopy++;
page_shift += desc->bits;
}
desc--;
for (int i = 0; i < ctrl->numLevelsToCopy; i++, desc--) {
page_shift -= desc->bits;
ctrl->levels[i].physAddress = pd->pt[0]->addr;
ctrl->levels[i].size = BIT_ULL(desc->bits) * desc->size;
ctrl->levels[i].aperture = 1;
ctrl->levels[i].pageShift = page_shift;
pd = pd->pde[0];
}
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl);
} else {
NV0080_CTRL_DMA_SET_PAGE_DIRECTORY_PARAMS *ctrl;
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
NV0080_CTRL_CMD_DMA_SET_PAGE_DIRECTORY,
sizeof(*ctrl));
if (IS_ERR(ctrl))
return PTR_ERR(ctrl);
ctrl->physAddress = vmm->pd->pt[0]->addr;
ctrl->numEntries = 1 << vmm->func->page[0].desc->bits;
ctrl->flags = NVDEF(NV0080_CTRL_DMA_SET_PAGE_DIRECTORY, FLAGS, APERTURE, VIDMEM);
ctrl->hVASpace = vmm->rm.object.handle;
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl);
if (ret == 0)
vmm->rm.external = true;
}
return ret;
}
static int
r535_mmu_promote_vmm(struct nvkm_vmm *vmm)
{
return r535_mmu_vaspace_new(vmm, NVKM_RM_VASPACE, true);
}
static void
r535_mmu_dtor(struct nvkm_mmu *mmu)
{
kfree(mmu->func);
}
int
r535_mmu_new(const struct nvkm_mmu_func *hw,
struct nvkm_device *device, enum nvkm_subdev_type type, int inst,
struct nvkm_mmu **pmmu)
{
struct nvkm_mmu_func *rm;
int ret;
if (!(rm = kzalloc_obj(*rm)))
return -ENOMEM;
rm->dtor = r535_mmu_dtor;
rm->dma_bits = hw->dma_bits;
rm->mmu = hw->mmu;
rm->mem = hw->mem;
rm->vmm = hw->vmm;
rm->kind = hw->kind;
rm->kind_sys = hw->kind_sys;
rm->promote_vmm = r535_mmu_promote_vmm;
ret = nvkm_mmu_new_(rm, device, type, inst, pmmu);
if (ret)
kfree(rm);
return ret;
}