vmx_get_exit_qual
exit_qualification = vmx_get_exit_qual(vcpu) & INTR_INFO_UNBLOCK_NMI;
exit_qualification |= vmx_get_exit_qual(vcpu) &
if (get_vmx_mem_address(vcpu, vmx_get_exit_qual(vcpu),
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qual = vmx_get_exit_qual(vcpu);
if (get_vmx_mem_address(vcpu, vmx_get_exit_qual(vcpu),
if (get_vmx_mem_address(vcpu, vmx_get_exit_qual(vcpu),
vmx_get_exit_qual(vcpu));
exit_qualification = vmx_get_exit_qual(vcpu);
msr_index = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
exit_qual = vmx_get_exit_qual(vcpu);
u64 eq = vmx_get_exit_qual(vcpu);
exit_qual = vmx_get_exit_qual(vcpu);
*info1 = vmx_get_exit_qual(vcpu);
unsigned long cr2 = vmx_get_exit_qual(vcpu);
dr6 = vmx_get_exit_qual(vcpu);
exit_qualification = vmx_get_exit_qual(vcpu);
exit_qualification = vmx_get_exit_qual(vcpu);
exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qualification = vmx_get_exit_qual(vcpu);
if (get_vmx_mem_address(vcpu, vmx_get_exit_qual(vcpu),
exit_qualification = vmx_get_exit_qual(vcpu);
unsigned long exit_qual = vmx_get_exit_qual(vcpu);
return kvm_emulate_rdmsr_imm(vcpu, vmx_get_exit_qual(vcpu),
return kvm_emulate_wrmsr_imm(vcpu, vmx_get_exit_qual(vcpu),
*info1 = vmx_get_exit_qual(vcpu);
return handle_fastpath_wrmsr_imm(vcpu, vmx_get_exit_qual(vcpu),