to_msm_vm
a2xx_gpummu_params(to_msm_vm(gpu->vm)->mmu, &pt_base, &tran_error);
msm_mmu_set_fault_handler(to_msm_vm(gpu->vm)->mmu, gpu,
struct msm_mmu *mmu = to_msm_vm(gmu->vm)->mmu;
if (msm_iommu_pagetable_params(to_msm_vm(vm)->mmu, &ttbr, &asid))
mmu = msm_iommu_pagetable_create(to_msm_vm(gpu->vm)->mmu, kernel_managed);
msm_mmu_set_fault_handler(to_msm_vm(gpu->vm)->mmu, gpu,
msm_iommu_pagetable_params(to_msm_vm(gpu->vm)->mmu, &ttbr, &asid);
struct msm_mmu *mmu = to_msm_vm(gpu->vm)->mmu;
struct msm_mmu *mmu = to_msm_vm(gpu->vm)->mmu;
*value = gpu->global_faults + to_msm_vm(vm)->faults;
mmu = to_msm_vm(dpu_kms->base.vm)->mmu;
struct msm_mmu *mmu = to_msm_vm(kms->vm)->mmu;
struct msm_mmu *mmu = to_msm_vm(kms->vm)->mmu;
struct msm_gem_vm *vm = to_msm_vm(vma->vm);
struct msm_gem_vm *vm = to_msm_vm(submit->vm);
if (to_msm_vm(ctx->vm)->unusable)
struct msm_mmu *mmu = to_msm_vm(job->vm)->mmu;
struct msm_gem_vm *vm = to_msm_vm(job->vm);
struct msm_gem_vm *vm = to_msm_vm(job->vm);
if (to_msm_vm(msm_context_vm(dev, ctx))->unusable)
struct msm_gem_vm *vm = to_msm_vm(gpuvm);
struct msm_gem_vm *vm = to_msm_vm(vma->vm);
struct msm_gem_vm *vm = to_msm_vm(vma->vm);
struct msm_gem_vm *vm = to_msm_vm(vma->vm);
struct msm_gem_vm *vm = to_msm_vm(gpuvm);
struct msm_gem_vm *vm = to_msm_vm(job->vm);
struct msm_gem_vm *vm = to_msm_vm(job->vm);
struct msm_gem_vm *vm = to_msm_vm(gpuvm);
struct msm_mmu *mmu = to_msm_vm(gpu->vm)->mmu;
struct msm_mmu *mmu = to_msm_vm(submit->vm)->mmu;
crashstate_get_vm_logs(state, to_msm_vm(submit->vm));
struct msm_gem_vm *vm = to_msm_vm(submit->vm);
if (to_msm_vm(submit->vm)->unusable)
to_msm_vm(vm)->pid = get_pid(task_pid(task));
msm_mmu_set_fault_handler(to_msm_vm(vm)->mmu, kms, msm_kms_fault_handler);
struct drm_gpu_scheduler *sched = &to_msm_vm(msm_context_vm(drm, ctx))->sched;