Symbol: kvm_x86_ops
arch/x86/include/asm/kvm_host.h
1993
struct kvm_x86_ops *runtime_ops;
arch/x86/include/asm/kvm_host.h
2010
extern struct kvm_x86_ops kvm_x86_ops;
arch/x86/include/asm/kvm_host.h
2016
DECLARE_STATIC_CALL(kvm_x86_##func, *(((struct kvm_x86_ops *)0)->func));
arch/x86/include/asm/kvm_host.h
2027
return kvzalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT);
arch/x86/include/asm/kvm_host.h
2037
if (kvm_x86_ops.flush_remote_tlbs &&
arch/x86/include/asm/kvm_host.h
2048
if (!kvm_x86_ops.flush_remote_tlbs_range)
arch/x86/kvm/hyperv.c
2403
kvm_x86_ops.nested_ops->hv_inject_synthetic_vmexit_post_tlb_flush(vcpu);
arch/x86/kvm/hyperv.c
2787
if (kvm_x86_ops.nested_ops->get_evmcs_version)
arch/x86/kvm/hyperv.c
2788
evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu);
arch/x86/kvm/kvm_onhyperv.c
116
if (kvm_x86_ops.flush_remote_tlbs == hv_flush_remote_tlbs) {
arch/x86/kvm/lapic.c
196
return kvm_x86_ops.set_hv_timer
arch/x86/kvm/lapic.c
2626
if (kvm_x86_ops.x2apic_icr_is_split) {
arch/x86/kvm/lapic.c
2648
if (kvm_x86_ops.x2apic_icr_is_split)
arch/x86/kvm/lapic.c
3067
if (kvm_x86_ops.alloc_apic_backing_page)
arch/x86/kvm/lapic.c
3228
if (!kvm_x86_ops.x2apic_icr_is_split) {
arch/x86/kvm/lapic.c
943
if (kvm_x86_ops.sync_pir_to_irr)
arch/x86/kvm/mmu/mmu.c
1670
if (kvm_x86_ops.set_apic_access_page_addr &&
arch/x86/kvm/mmu/mmu.c
275
return kvm_x86_ops.flush_remote_tlbs_range;
arch/x86/kvm/mmu/paging_tmpl.h
225
if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr))
arch/x86/kvm/mmu/spte.c
242
if (kvm_x86_ops.get_mt_mask)
arch/x86/kvm/svm/svm.c
5128
struct kvm_x86_ops svm_x86_ops __initdata = {
arch/x86/kvm/svm/svm.h
56
extern struct kvm_x86_ops svm_x86_ops __initdata;
arch/x86/kvm/vmx/main.c
867
struct kvm_x86_ops vt_x86_ops __initdata = {
arch/x86/kvm/vmx/x86_ops.h
11
extern struct kvm_x86_ops vt_x86_ops __initdata;
arch/x86/kvm/x86.c
10003
memcpy(&kvm_x86_ops, ops->runtime_ops, sizeof(kvm_x86_ops));
arch/x86/kvm/x86.c
10006
static_call_update(kvm_x86_##func, kvm_x86_ops.func);
arch/x86/kvm/x86.c
10008
WARN_ON(!kvm_x86_ops.func); __KVM_X86_OP(func)
arch/x86/kvm/x86.c
10011
static_call_update(kvm_x86_##func, (void *)kvm_x86_ops.func ? : \
arch/x86/kvm/x86.c
10051
if (kvm_x86_ops.enable_virtualization_cpu) {
arch/x86/kvm/x86.c
10052
pr_err("already loaded vendor module '%s'\n", kvm_x86_ops.name);
arch/x86/kvm/x86.c
10190
kvm_x86_ops.enable_virtualization_cpu = NULL;
arch/x86/kvm/x86.c
10230
kvm_x86_ops.enable_virtualization_cpu = NULL;
arch/x86/kvm/x86.c
10550
if (!kvm_x86_ops.update_cr8_intercept)
arch/x86/kvm/x86.c
10576
kvm_x86_ops.nested_ops->triple_fault(vcpu);
arch/x86/kvm/x86.c
10580
return kvm_x86_ops.nested_ops->check_events(vcpu);
arch/x86/kvm/x86.c
10821
kvm_x86_ops.nested_ops->has_events &&
arch/x86/kvm/x86.c
10822
kvm_x86_ops.nested_ops->has_events(vcpu, true))
arch/x86/kvm/x86.c
10958
kvm_x86_ops.allow_apicv_in_x2apic_without_x2apic_virtualization)
arch/x86/kvm/x86.c
10971
if (!(kvm_x86_ops.required_apicv_inhibits & BIT(reason)))
arch/x86/kvm/x86.c
11102
if (unlikely(!kvm_x86_ops.nested_ops->get_nested_state_pages(vcpu))) {
arch/x86/kvm/x86.c
11154
kvm_x86_ops.nested_ops->triple_fault(vcpu);
arch/x86/kvm/x86.c
11368
if ((debug_ctl ^ vcpu->arch.host_debugctl) & kvm_x86_ops.HOST_OWNED_DEBUGCTL &&
arch/x86/kvm/x86.c
11576
kvm_x86_ops.nested_ops->has_events &&
arch/x86/kvm/x86.c
11577
kvm_x86_ops.nested_ops->has_events(vcpu, false))
arch/x86/kvm/x86.c
11995
kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, ex->vector,
arch/x86/kvm/x86.c
13076
cpu_emergency_register_virt_callback(kvm_x86_ops.emergency_disable_virtualization_cpu);
arch/x86/kvm/x86.c
13081
cpu_emergency_unregister_virt_callback(kvm_x86_ops.emergency_disable_virtualization_cpu);
arch/x86/kvm/x86.c
143
struct kvm_x86_ops kvm_x86_ops __read_mostly;
arch/x86/kvm/x86.c
147
*(((struct kvm_x86_ops *)0)->func));
arch/x86/kvm/x86.c
4951
r = kvm_x86_ops.nested_ops->get_state ?
arch/x86/kvm/x86.c
4952
kvm_x86_ops.nested_ops->get_state(NULL, NULL, 0) : 0;
arch/x86/kvm/x86.c
4956
r = kvm_x86_ops.enable_l2_tlb_flush != NULL;
arch/x86/kvm/x86.c
4959
r = kvm_x86_ops.nested_ops->enable_evmcs != NULL;
arch/x86/kvm/x86.c
5005
if (kvm_x86_ops.dev_get_attr)
arch/x86/kvm/x86.c
6031
if (!kvm_x86_ops.nested_ops->enable_evmcs)
arch/x86/kvm/x86.c
6033
r = kvm_x86_ops.nested_ops->enable_evmcs(vcpu, &vmcs_version);
arch/x86/kvm/x86.c
6043
if (!kvm_x86_ops.enable_l2_tlb_flush)
arch/x86/kvm/x86.c
6531
if (!kvm_x86_ops.nested_ops->get_state)
arch/x86/kvm/x86.c
6539
r = kvm_x86_ops.nested_ops->get_state(vcpu, user_kvm_nested_state,
arch/x86/kvm/x86.c
6561
if (!kvm_x86_ops.nested_ops->set_state)
arch/x86/kvm/x86.c
6584
r = kvm_x86_ops.nested_ops->set_state(vcpu, user_kvm_nested_state, &kvm_state);
arch/x86/kvm/x86.c
6654
if (!kvm_x86_ops.vcpu_mem_enc_ioctl)
arch/x86/kvm/x86.c
6656
r = kvm_x86_ops.vcpu_mem_enc_ioctl(vcpu, argp);
arch/x86/kvm/x86.c
6861
if (!kvm_x86_ops.vm_copy_enc_context_from)
arch/x86/kvm/x86.c
6868
if (!kvm_x86_ops.vm_move_enc_context_from)
arch/x86/kvm/x86.c
7247
kvm_x86_ops.vcpu_mem_enc_unlocked_ioctl)
arch/x86/kvm/x86.c
7555
if (!kvm_x86_ops.mem_enc_ioctl)
arch/x86/kvm/x86.c
7568
if (!kvm_x86_ops.mem_enc_register_region)
arch/x86/kvm/x86.c
7582
if (!kvm_x86_ops.mem_enc_unregister_region)
arch/x86/kvm/x86.c
851
kvm_x86_ops.nested_ops->is_exception_vmexit(vcpu, nr, error_code)) {
arch/x86/kvm/x86.c
8880
if (!kvm_x86_ops.get_untagged_addr)
arch/x86/kvm/x86.h
154
kvm_x86_ops.nested_ops->leave_nested(vcpu);