Symbol: emul_to_vcpu
arch/x86/kvm/x86.c
10500
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
7906
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
7951
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
7996
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8239
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8317
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8467
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8495
return emulator_pio_out(emul_to_vcpu(ctxt), size, port, val, count);
arch/x86/kvm/x86.c
8505
kvm_mmu_invlpg(emul_to_vcpu(ctxt), address);
arch/x86/kvm/x86.c
8536
kvm_emulate_wbinvd_noskip(emul_to_vcpu(ctxt));
arch/x86/kvm/x86.c
8541
return kvm_get_dr(emul_to_vcpu(ctxt), dr);
arch/x86/kvm/x86.c
8548
return kvm_set_dr(emul_to_vcpu(ctxt), dr, value);
arch/x86/kvm/x86.c
8558
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8587
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8616
return kvm_x86_call(get_cpl)(emul_to_vcpu(ctxt));
arch/x86/kvm/x86.c
8621
kvm_x86_call(get_gdt)(emul_to_vcpu(ctxt), dt);
arch/x86/kvm/x86.c
8626
kvm_x86_call(get_idt)(emul_to_vcpu(ctxt), dt);
arch/x86/kvm/x86.c
8631
kvm_x86_call(set_gdt)(emul_to_vcpu(ctxt), dt);
arch/x86/kvm/x86.c
8636
kvm_x86_call(set_idt)(emul_to_vcpu(ctxt), dt);
arch/x86/kvm/x86.c
8642
return get_segment_base(emul_to_vcpu(ctxt), seg);
arch/x86/kvm/x86.c
8651
kvm_get_segment(emul_to_vcpu(ctxt), &var, seg);
arch/x86/kvm/x86.c
8685
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8714
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8737
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8767
return kvm_msr_read(emul_to_vcpu(ctxt), msr_index, pdata);
arch/x86/kvm/x86.c
8769
return __kvm_emulate_msr_read(emul_to_vcpu(ctxt), msr_index, pdata);
arch/x86/kvm/x86.c
8774
return kvm_pmu_check_rdpmc_early(emul_to_vcpu(ctxt), pmc);
arch/x86/kvm/x86.c
8780
return kvm_pmu_rdpmc(emul_to_vcpu(ctxt), pmc, pdata);
arch/x86/kvm/x86.c
8785
emul_to_vcpu(ctxt)->arch.halt_request = 1;
arch/x86/kvm/x86.c
8792
return kvm_x86_call(check_intercept)(emul_to_vcpu(ctxt), info, stage,
arch/x86/kvm/x86.c
8800
return kvm_cpuid(emul_to_vcpu(ctxt), eax, ebx, ecx, edx, exact_only);
arch/x86/kvm/x86.c
8805
return guest_cpu_cap_has(emul_to_vcpu(ctxt), X86_FEATURE_MOVBE);
arch/x86/kvm/x86.c
8810
return guest_cpu_cap_has(emul_to_vcpu(ctxt), X86_FEATURE_FXSR);
arch/x86/kvm/x86.c
8815
return guest_cpu_cap_has(emul_to_vcpu(ctxt), X86_FEATURE_RDPID);
arch/x86/kvm/x86.c
8820
return guest_cpuid_is_intel_compatible(emul_to_vcpu(ctxt));
arch/x86/kvm/x86.c
8825
return kvm_register_read_raw(emul_to_vcpu(ctxt), reg);
arch/x86/kvm/x86.c
8830
kvm_register_write_raw(emul_to_vcpu(ctxt), reg, val);
arch/x86/kvm/x86.c
8835
kvm_x86_call(set_nmi_mask)(emul_to_vcpu(ctxt), masked);
arch/x86/kvm/x86.c
8840
return is_smm(emul_to_vcpu(ctxt));
arch/x86/kvm/x86.c
8853
kvm_make_request(KVM_REQ_TRIPLE_FAULT, emul_to_vcpu(ctxt));
arch/x86/kvm/x86.c
8860
*xcr = emul_to_vcpu(ctxt)->arch.xcr0;
arch/x86/kvm/x86.c
8866
return __kvm_set_xcr(emul_to_vcpu(ctxt), index, xcr);
arch/x86/kvm/x86.c
8871
struct kvm *kvm = emul_to_vcpu(ctxt)->kvm;
arch/x86/kvm/x86.c
8883
return kvm_x86_call(get_untagged_addr)(emul_to_vcpu(ctxt),
arch/x86/kvm/x86.c
8890
return !is_noncanonical_address(addr, emul_to_vcpu(ctxt), flags);