arch/x86/kvm/kvm-asm-offsets.c
27
OFFSET(VMX_spec_ctrl, vcpu_vmx, spec_ctrl);
arch/x86/kvm/vmx/hyperv.c
202
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/hyperv.c
215
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/hyperv.h
25
static inline bool nested_vmx_is_evmptr12_valid(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/hyperv.h
35
static inline bool nested_vmx_is_evmptr12_set(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/hyperv.h
40
static inline struct hv_enlightened_vmcs *nested_vmx_evmcs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/hyperv.h
69
static inline bool nested_vmx_is_evmptr12_valid(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/hyperv.h
79
static inline bool nested_vmx_is_evmptr12_set(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/hyperv.h
84
static inline struct hv_enlightened_vmcs *nested_vmx_evmcs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/main.c
1056
vcpu_size = sizeof(struct vcpu_vmx);
arch/x86/kvm/vmx/main.c
1057
vcpu_align = __alignof__(struct vcpu_vmx);
arch/x86/kvm/vmx/main.c
14
static_assert(offsetof(struct vcpu_vmx, vt) == offsetof(struct vcpu_tdx, vt));
arch/x86/kvm/vmx/nested.c
1075
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
1228
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
1289
static int vmx_restore_vmx_basic(struct vcpu_vmx *vmx, u64 data)
arch/x86/kvm/vmx/nested.c
1361
vmx_restore_control_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data)
arch/x86/kvm/vmx/nested.c
1384
static int vmx_restore_vmx_misc(struct vcpu_vmx *vmx, u64 data)
arch/x86/kvm/vmx/nested.c
1432
static int vmx_restore_vmx_ept_vpid_cap(struct vcpu_vmx *vmx, u64 data)
arch/x86/kvm/vmx/nested.c
1458
static int vmx_restore_fixed0_msr(struct vcpu_vmx *vmx, u32 msr_index, u64 data)
arch/x86/kvm/vmx/nested.c
1480
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
1628
static void copy_shadow_to_vmcs12(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
1655
static void copy_vmcs12_to_shadow(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
1689
static void copy_enlightened_to_vmcs12(struct vcpu_vmx *vmx, u32 hv_clean_fields)
arch/x86/kvm/vmx/nested.c
1937
static void copy_vmcs12_to_enlightened(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
196
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2123
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2211
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2223
struct vcpu_vmx *vmx =
arch/x86/kvm/vmx/nested.c
2224
container_of(timer, struct vcpu_vmx, nested.preemption_timer);
arch/x86/kvm/vmx/nested.c
2235
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2252
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
226
static void vmx_disable_shadow_vmcs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
2274
static u64 nested_vmx_calc_efer(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/nested.c
2285
static void prepare_vmcs02_constant_state(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
2340
static void prepare_vmcs02_early_rare(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/nested.c
2366
static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01,
arch/x86/kvm/vmx/nested.c
237
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
254
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2558
static void prepare_vmcs02_rare(struct vcpu_vmx *vmx, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/nested.c
2709
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
278
static void vmx_sync_vmcs_host_state(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/nested.c
2868
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2917
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2980
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
299
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
2997
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
324
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3248
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
338
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3392
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3422
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3537
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3609
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
3773
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4008
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4159
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
417
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4285
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4557
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4602
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4630
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
470
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
4912
static inline u64 nested_vmx_get_vmcs01_guest_efer(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/nested.c
4938
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5052
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5373
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5396
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5449
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5537
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5579
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5637
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5743
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5844
static void set_current_vmptr(struct vcpu_vmx *vmx, gpa_t vmptr)
arch/x86/kvm/vmx/nested.c
5860
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
5958
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
598
void nested_vmx_set_msr_##rw##_intercept(struct vcpu_vmx *vmx, \
arch/x86/kvm/vmx/nested.c
6038
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
611
static inline void nested_vmx_set_intercept_for_msr(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/nested.c
6170
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
642
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6677
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
672
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
6731
struct vcpu_vmx *vmx;
arch/x86/kvm/vmx/nested.c
6885
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
792
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
811
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.c
894
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.h
81
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/nested.h
90
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/sgx.c
300
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/sgx.c
430
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1003
unsigned int __vmx_vcpu_run_flags(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1025
static __always_inline void clear_atomic_switch_msr_special(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/vmx.c
1057
static void clear_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr)
arch/x86/kvm/vmx/vmx.c
1084
static __always_inline void add_atomic_switch_msr_special(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/vmx.c
1113
static void add_atomic_switch_msr(struct vcpu_vmx *vmx, unsigned msr,
arch/x86/kvm/vmx/vmx.c
1155
static bool update_transition_efer(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1207
static void vmx_add_autostore_msr(struct vcpu_vmx *vmx, u32 msr)
arch/x86/kvm/vmx/vmx.c
1213
static void vmx_remove_autostore_msr(struct vcpu_vmx *vmx, u32 msr)
arch/x86/kvm/vmx/vmx.c
1247
static inline bool pt_can_write_msr(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1287
static void pt_guest_enter(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1304
static void pt_guest_exit(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1351
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1422
static void vmx_prepare_switch_to_host(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1462
static u64 vmx_read_guest_host_msr(struct vcpu_vmx *vmx, u32 msr, u64 *cache)
arch/x86/kvm/vmx/vmx.c
1471
static void vmx_write_guest_host_msr(struct vcpu_vmx *vmx, u32 msr, u64 data,
arch/x86/kvm/vmx/vmx.c
1481
static u64 vmx_read_guest_kernel_gs_base(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
1487
static void vmx_write_guest_kernel_gs_base(struct vcpu_vmx *vmx, u64 data)
arch/x86/kvm/vmx/vmx.c
1496
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1512
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1528
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1606
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1617
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1645
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1663
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1726
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1891
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1942
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
1983
static void vmx_setup_uret_msr(struct vcpu_vmx *vmx, unsigned int msr,
arch/x86/kvm/vmx/vmx.c
2001
static void vmx_setup_uret_msrs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
2083
static inline bool is_vmx_feature_control_msr_valid(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/vmx.c
2126
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
2332
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3194
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3264
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3311
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3360
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3504
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3646
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3709
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3756
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
3798
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4168
static void vmx_msr_bitmap_l01_changed(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4188
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4220
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4273
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4293
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4432
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4513
void vmx_set_constant_host_state(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4605
void set_cr4_guest_host_mask(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4621
static u32 vmx_pin_based_exec_ctrl(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4676
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
4693
static u32 vmx_exec_control(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
473
static __always_inline void vmx_disable_fb_clear(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4737
static u64 vmx_tertiary_exec_control(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4757
vmx_adjust_secondary_exec_control(struct vcpu_vmx *vmx, u32 *exec_control,
arch/x86/kvm/vmx/vmx.c
4815
static u32 vmx_secondary_exec_control(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
487
static __always_inline void vmx_enable_fb_clear(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
4935
static void init_vmcs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
496
static void vmx_update_fb_clear_dis(struct kvm_vcpu *vcpu, struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
5061
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5093
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5176
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5204
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5235
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5249
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5479
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
5980
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6108
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6135
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6274
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6467
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6494
static void vmx_destroy_pml_buffer(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
6504
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6543
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6578
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6783
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
6972
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7301
static void vmx_recover_nmi_blocking(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
7399
static void vmx_complete_interrupts(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
7416
static void atomic_switch_perf_msrs(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
7445
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7467
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7491
void noinstr vmx_update_host_rsp(struct vcpu_vmx *vmx, unsigned long host_rsp)
arch/x86/kvm/vmx/vmx.c
7499
void noinstr vmx_spec_ctrl_restore_host(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/vmx.c
7569
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7608
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
763
struct vmx_uret_msr *vmx_find_uret_msr(struct vcpu_vmx *vmx, u32 msr)
arch/x86/kvm/vmx/vmx.c
773
static int vmx_set_guest_uret_msr(struct vcpu_vmx *vmx,
arch/x86/kvm/vmx/vmx.c
7760
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
7773
struct vcpu_vmx *vmx;
arch/x86/kvm/vmx/vmx.c
7776
BUILD_BUG_ON(offsetof(struct vcpu_vmx, vcpu) != 0);
arch/x86/kvm/vmx/vmx.c
7935
static void vmcs_set_secondary_exec_control(struct vcpu_vmx *vmx, u32 new_ctl)
arch/x86/kvm/vmx/vmx.c
7960
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8005
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8074
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8417
struct vcpu_vmx *vmx;
arch/x86/kvm/vmx/vmx.c
8462
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8501
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
8522
struct vcpu_vmx *vmx = to_vmx(vcpu);
arch/x86/kvm/vmx/vmx.c
876
static bool vmx_segment_cache_test_set(struct vcpu_vmx *vmx, unsigned seg,
arch/x86/kvm/vmx/vmx.c
891
static u16 vmx_read_guest_seg_selector(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
900
static ulong vmx_read_guest_seg_base(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
909
static u32 vmx_read_guest_seg_limit(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
918
static u32 vmx_read_guest_seg_ar(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
995
static bool msr_write_intercepted(struct vcpu_vmx *vmx, u32 msr)
arch/x86/kvm/vmx/vmx.h
306
return &(container_of(vcpu, struct vcpu_vmx, vcpu)->vt);
arch/x86/kvm/vmx/vmx.h
311
return &(container_of(vt, struct vcpu_vmx, vt)->vcpu);
arch/x86/kvm/vmx/vmx.h
344
void vmx_set_constant_host_state(struct vcpu_vmx *vmx);
arch/x86/kvm/vmx/vmx.h
358
void set_cr4_guest_host_mask(struct vcpu_vmx *vmx);
arch/x86/kvm/vmx/vmx.h
371
struct vmx_uret_msr *vmx_find_uret_msr(struct vcpu_vmx *vmx, u32 msr);
arch/x86/kvm/vmx/vmx.h
373
void vmx_update_host_rsp(struct vcpu_vmx *vmx, unsigned long host_rsp);
arch/x86/kvm/vmx/vmx.h
374
void vmx_spec_ctrl_restore_host(struct vcpu_vmx *vmx, unsigned int flags);
arch/x86/kvm/vmx/vmx.h
375
unsigned int __vmx_vcpu_run_flags(struct vcpu_vmx *vmx);
arch/x86/kvm/vmx/vmx.h
376
bool __vmx_vcpu_run(struct vcpu_vmx *vmx, unsigned long *regs,
arch/x86/kvm/vmx/vmx.h
581
static inline void lname##_controls_set(struct vcpu_vmx *vmx, u##bits val) \
arch/x86/kvm/vmx/vmx.h
592
static inline u##bits lname##_controls_get(struct vcpu_vmx *vmx) \
arch/x86/kvm/vmx/vmx.h
596
static __always_inline void lname##_controls_setbit(struct vcpu_vmx *vmx, u##bits val) \
arch/x86/kvm/vmx/vmx.h
601
static __always_inline void lname##_controls_clearbit(struct vcpu_vmx *vmx, u##bits val) \
arch/x86/kvm/vmx/vmx.h
606
static __always_inline void lname##_controls_changebit(struct vcpu_vmx *vmx, u##bits val, \
arch/x86/kvm/vmx/vmx.h
660
static __always_inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
662
return container_of(vcpu, struct vcpu_vmx, vcpu);
arch/x86/kvm/vmx/vmx.h
680
static inline bool vmx_has_waitpkg(struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.h
725
static inline void vmx_segment_cache_clear(struct vcpu_vmx *vmx)