xe_vma_vm
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_vm *vm = xe_vma_vm(vma);
xe_vma_vm(vma)->flags, NULL);
vm = xe_vma_vm(vma);
vm = xe_vma_vm(vma);
xe_vma_vm(vma)->flags, deferred);
xe_pt_destroy(newpte, xe_vma_vm(vma)->flags, NULL);
vm_dbg(&xe_vma_vm(vma)->xe->drm,
vm_dbg(&xe_vma_vm(vma)->xe->drm,
vm_dbg(&xe_vma_vm(vma)->xe->drm,
pt_op->num_entries = xe_pt_stage_unbind(tile, xe_vma_vm(vma),
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_pt *pt = xe_vma_vm(vma)->pt_root[tile->id];
lockdep_assert_held(&xe_vma_vm(vma)->svm.gpusvm.notifier_lock);
__entry->vm = xe_vma_vm(vma);
__entry->asid = xe_vma_vm(vma)->usm.asid;
#define __dev_name_vma(vma) __dev_name_vm(xe_vma_vm(vma))
struct xe_vm *vm = xe_vma_vm(vma);
vm_dbg(&xe_vma_vm(vma)->xe->drm,
struct xe_vm *vm = xe_vma_vm(&uvma->vma);
struct xe_vm *vm = xe_vma_vm(&uvma->vma);
struct xe_vm *vm = xe_vma_vm(&uvma->vma);
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_device *xe = xe_vma_vm(vma)->xe;
xe_assert(vm->xe, xe_vma_vm(vma) == vm);
xe_assert(vm->xe, xe_vma_vm(vma) == vm);
struct xe_vm *vm = xe_vma_vm(vma);
struct xe_device *xe = xe_vma_vm(vma)->xe;
struct xe_vm *vm = xe_vma_vm(vma);
ret = xe_vm_range_tilemask_tlb_inval(xe_vma_vm(vma), xe_vma_start(vma),
xe_assert(xe_vma_vm(vma)->xe, xe_vma_is_userptr(vma));