Symbol: kvm_vcpu
arch/arm64/include/asm/kvm_asm.h
220
struct kvm_vcpu;
arch/arm64/include/asm/kvm_asm.h
249
extern int __kvm_at_s1e01(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_asm.h
250
extern int __kvm_at_s1e2(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_asm.h
251
extern int __kvm_at_s12(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_asm.h
253
extern int __kvm_vcpu_run(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_asm.h
255
extern void __kvm_adjust_pc(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
104
static inline void vcpu_reset_hcr(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
119
static inline unsigned long *vcpu_hcr(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
124
static inline unsigned long vcpu_get_vsesr(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
129
static inline void vcpu_set_vsesr(struct kvm_vcpu *vcpu, u64 vsesr)
arch/arm64/include/asm/kvm_emulate.h
134
static __always_inline unsigned long *vcpu_pc(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
139
static __always_inline unsigned long *vcpu_cpsr(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
144
static __always_inline bool vcpu_mode_is_32bit(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
149
static __always_inline bool kvm_condition_valid(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
157
static inline void vcpu_set_thumb(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
167
static __always_inline unsigned long vcpu_get_reg(const struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_emulate.h
173
static __always_inline void vcpu_set_reg(struct kvm_vcpu *vcpu, u8 reg_num,
arch/arm64/include/asm/kvm_emulate.h
191
static inline bool vcpu_is_el2(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
196
static inline bool vcpu_el2_e2h_is_set(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
202
static inline bool vcpu_el2_tge_is_set(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
207
static inline bool vcpu_el2_amo_is_set(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
226
static inline bool is_hyp_ctxt(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
250
static inline bool vcpu_is_host_el0(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
255
static inline bool is_nested_ctxt(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
260
static inline bool vserror_state_is_nested(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
301
static inline bool vcpu_mode_priv(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
315
static __always_inline u64 kvm_vcpu_get_esr(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
320
static inline bool guest_hyp_wfx_traps_enabled(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
333
static __always_inline int kvm_vcpu_get_condition(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
343
static __always_inline unsigned long kvm_vcpu_get_hfar(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
348
static __always_inline phys_addr_t kvm_vcpu_get_fault_ipa(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
358
static inline u64 kvm_vcpu_get_disr(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
363
static inline u32 kvm_vcpu_hvc_get_imm(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
368
static __always_inline bool kvm_vcpu_dabt_isvalid(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
373
static inline unsigned long kvm_vcpu_dabt_iss_nisv_sanitized(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
378
static inline bool kvm_vcpu_dabt_issext(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
383
static inline bool kvm_vcpu_dabt_issf(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
388
static __always_inline int kvm_vcpu_dabt_get_rd(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
393
static __always_inline bool kvm_vcpu_abt_iss1tw(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
399
static __always_inline bool kvm_vcpu_dabt_iswrite(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
404
static inline bool kvm_vcpu_dabt_is_cm(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
409
static __always_inline unsigned int kvm_vcpu_dabt_get_as(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
415
static __always_inline bool kvm_vcpu_trap_il_is32bit(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
420
static __always_inline u8 kvm_vcpu_trap_get_class(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
425
static inline bool kvm_vcpu_trap_is_iabt(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
430
static inline bool kvm_vcpu_trap_is_exec_fault(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
435
static __always_inline u8 kvm_vcpu_trap_get_fault(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
44
bool kvm_condition_valid32(const struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
441
bool kvm_vcpu_trap_is_permission_fault(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
447
bool kvm_vcpu_trap_is_translation_fault(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
45
void kvm_skip_instr32(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
453
u64 kvm_vcpu_trap_get_perm_fault_granule(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
461
static __always_inline bool kvm_vcpu_abt_issea(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
47
void kvm_inject_undefined(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
474
static __always_inline int kvm_vcpu_sys_get_rt(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
48
void kvm_inject_sync(struct kvm_vcpu *vcpu, u64 esr);
arch/arm64/include/asm/kvm_emulate.h
480
static inline bool kvm_is_write_fault(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
49
int kvm_inject_serror_esr(struct kvm_vcpu *vcpu, u64 esr);
arch/arm64/include/asm/kvm_emulate.h
50
int kvm_inject_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr);
arch/arm64/include/asm/kvm_emulate.h
504
static inline unsigned long kvm_vcpu_get_mpidr_aff(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
509
static inline void kvm_vcpu_set_be(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
51
int kvm_inject_dabt_excl_atomic(struct kvm_vcpu *vcpu, u64 addr);
arch/arm64/include/asm/kvm_emulate.h
52
void kvm_inject_size_fault(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
525
static inline bool kvm_vcpu_is_be(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
539
static inline unsigned long vcpu_data_guest_to_host(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_emulate.h
54
static inline int kvm_inject_sea_dabt(struct kvm_vcpu *vcpu, u64 addr)
arch/arm64/include/asm/kvm_emulate.h
570
static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_emulate.h
59
static inline int kvm_inject_sea_iabt(struct kvm_vcpu *vcpu, u64 addr)
arch/arm64/include/asm/kvm_emulate.h
601
static __always_inline void kvm_incr_pc(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
618
static inline u64 vcpu_sanitised_cptr_el2(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
628
static inline bool ____cptr_xen_trap_enabled(const struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_emulate.h
64
static inline int kvm_inject_serror(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
649
static inline bool guest_hyp_fpsimd_traps_enabled(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
654
static inline bool guest_hyp_sve_traps_enabled(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
659
static inline void vcpu_set_hcrx(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
76
void kvm_vcpu_wfi(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
78
void kvm_emulate_nested_eret(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
79
int kvm_inject_nested_sync(struct kvm_vcpu *vcpu, u64 esr_el2);
arch/arm64/include/asm/kvm_emulate.h
80
int kvm_inject_nested_irq(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_emulate.h
81
int kvm_inject_nested_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr);
arch/arm64/include/asm/kvm_emulate.h
82
int kvm_inject_nested_serror(struct kvm_vcpu *vcpu, u64 esr);
arch/arm64/include/asm/kvm_emulate.h
84
static inline void kvm_inject_nested_sve_trap(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
93
static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_emulate.h
98
static __always_inline bool vcpu_el1_is_32bit(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_host.h
1167
u64 kvm_vcpu_apply_reg_masks(const struct kvm_vcpu *, enum vcpu_sysreg, u64);
arch/arm64/include/asm/kvm_host.h
1199
u64 vcpu_read_sys_reg(const struct kvm_vcpu *, enum vcpu_sysreg);
arch/arm64/include/asm/kvm_host.h
1200
void vcpu_write_sys_reg(struct kvm_vcpu *, u64, enum vcpu_sysreg);
arch/arm64/include/asm/kvm_host.h
1217
unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1218
int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices);
arch/arm64/include/asm/kvm_host.h
1219
int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
arch/arm64/include/asm/kvm_host.h
1220
int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
arch/arm64/include/asm/kvm_host.h
1222
unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1223
int kvm_arm_copy_sys_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices);
arch/arm64/include/asm/kvm_host.h
1225
int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1228
int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1280
int handle_exit(struct kvm_vcpu *vcpu, int exception_index);
arch/arm64/include/asm/kvm_host.h
1281
void handle_exit_early(struct kvm_vcpu *vcpu, int exception_index);
arch/arm64/include/asm/kvm_host.h
1283
int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1284
int kvm_handle_cp14_32(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1285
int kvm_handle_cp14_64(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1286
int kvm_handle_cp15_32(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1287
int kvm_handle_cp15_64(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1288
int kvm_handle_sys_reg(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1289
int kvm_handle_cp10_id(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1292
void kvm_reset_sys_regs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1300
void kvm_calculate_traps(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1306
int kvm_handle_mmio_return(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1307
int io_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa);
arch/arm64/include/asm/kvm_host.h
1314
static inline bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_host.h
1319
long kvm_hypercall_pv_features(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1320
gpa_t kvm_init_stolen_time(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1321
void kvm_update_stolen_time(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1324
int kvm_arm_pvtime_set_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1326
int kvm_arm_pvtime_get_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1328
int kvm_arm_pvtime_has_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1347
struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr);
arch/arm64/include/asm/kvm_host.h
1409
void kvm_vcpu_load_debug(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1410
void kvm_vcpu_put_debug(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1411
void kvm_debug_set_guest_ownership(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1412
void kvm_debug_handle_oslar(struct kvm_vcpu *vcpu, u64 val);
arch/arm64/include/asm/kvm_host.h
1424
int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1426
int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1428
int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_host.h
1439
void kvm_arch_vcpu_load_fp(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1440
void kvm_arch_vcpu_ctxflush_fp(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1441
void kvm_arch_vcpu_ctxsync_fp(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1442
void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1468
void kvm_vcpu_load_vhe(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1469
void kvm_vcpu_put_vhe(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1485
int kvm_arm_vcpu_finalize(struct kvm_vcpu *vcpu, int feature);
arch/arm64/include/asm/kvm_host.h
1486
bool kvm_arm_vcpu_is_finalized(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1511
int kvm_trng_call(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1520
void kvm_arm_vcpu_power_off(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1521
bool kvm_arm_vcpu_stopped(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
1639
void kvm_vcpu_load_fgt(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
703
struct kvm_vcpu *__hyp_running_vcpu;
arch/arm64/include/asm/kvm_host.h
85
void kvm_reset_vcpu(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_host.h
86
void kvm_arm_vcpu_destroy(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
100
void __vcpu_put_switch_sysregs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
107
void __debug_switch_to_guest(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
108
void __debug_switch_to_host(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
111
void __debug_save_host_buffers_nvhe(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
112
void __debug_restore_host_buffers_nvhe(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
120
u64 __guest_enter(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
77
int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
88
int __vgic_v3_perform_cpuif_access(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
91
void __timer_enable_traps(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
92
void __timer_disable_traps(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_hyp.h
99
void __vcpu_load_switch_sysregs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_mmu.h
184
int kvm_handle_guest_sea(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_mmu.h
185
int kvm_handle_guest_abort(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_mmu.h
204
static inline bool vcpu_has_cache_enabled(struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_mmu.h
261
void kvm_set_way_flush(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_mmu.h
262
void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled);
arch/arm64/include/asm/kvm_nested.h
10
static inline bool vcpu_has_nv(const struct kvm_vcpu *vcpu)
arch/arm64/include/asm/kvm_nested.h
161
extern int kvm_walk_nested_s2(struct kvm_vcpu *vcpu, phys_addr_t gipa,
arch/arm64/include/asm/kvm_nested.h
163
extern int kvm_s2_handle_perm_fault(struct kvm_vcpu *vcpu,
arch/arm64/include/asm/kvm_nested.h
165
extern int kvm_inject_s2_fault(struct kvm_vcpu *vcpu, u64 esr_el2);
arch/arm64/include/asm/kvm_nested.h
172
static inline bool kvm_supported_tlbi_s1e1_op(struct kvm_vcpu *vpcu, u32 instr)
arch/arm64/include/asm/kvm_nested.h
198
static inline bool kvm_supported_tlbi_s1e2_op(struct kvm_vcpu *vpcu, u32 instr)
arch/arm64/include/asm/kvm_nested.h
227
int kvm_init_nv_sysregs(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
231
bool kvm_auth_eretax(struct kvm_vcpu *vcpu, u64 *elr);
arch/arm64/include/asm/kvm_nested.h
233
static inline bool kvm_auth_eretax(struct kvm_vcpu *vcpu, u64 *elr)
arch/arm64/include/asm/kvm_nested.h
390
int __kvm_translate_va(struct kvm_vcpu *vcpu, struct s1_walk_info *wi,
arch/arm64/include/asm/kvm_nested.h
392
int __kvm_find_s1_desc_level(struct kvm_vcpu *vcpu, u64 va, u64 ipa,
arch/arm64/include/asm/kvm_nested.h
396
int kvm_vcpu_allocate_vncr_tlb(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
397
int kvm_handle_vncr_abort(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
398
void kvm_handle_s1e2_tlbi(struct kvm_vcpu *vcpu, u32 inst, u64 val);
arch/arm64/include/asm/kvm_nested.h
400
u16 get_asid_by_regime(struct kvm_vcpu *vcpu, enum trans_regime regime);
arch/arm64/include/asm/kvm_nested.h
66
extern bool forward_smc_trap(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
67
extern bool forward_debug_exception(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
69
extern int kvm_vcpu_init_nested(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
71
extern struct kvm_s2_mmu *lookup_s2_mmu(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
79
extern void kvm_vcpu_load_hw_mmu(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
80
extern void kvm_vcpu_put_hw_mmu(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
82
extern void check_nested_vcpu_requests(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
83
extern void kvm_nested_flush_hwstate(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
84
extern void kvm_nested_sync_hwstate(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_nested.h
86
extern void kvm_nested_setup_mdcr_el2(struct kvm_vcpu *vcpu);
arch/arm64/include/asm/kvm_pkvm.h
24
int pkvm_create_hyp_vcpu(struct kvm_vcpu *vcpu);
arch/arm64/kernel/asm-offsets.c
105
DEFINE(VCPU_CONTEXT, offsetof(struct kvm_vcpu, arch.ctxt));
arch/arm64/kernel/asm-offsets.c
106
DEFINE(VCPU_FAULT_DISR, offsetof(struct kvm_vcpu, arch.fault.disr_el1));
arch/arm64/kernel/asm-offsets.c
107
DEFINE(VCPU_HCR_EL2, offsetof(struct kvm_vcpu, arch.hcr_el2));
arch/arm64/kvm/arch_timer.c
1006
void kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
1052
static void timer_context_init(struct kvm_vcpu *vcpu, int timerid)
arch/arm64/kvm/arch_timer.c
107
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctxt);
arch/arm64/kvm/arch_timer.c
1082
void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
1136
static u64 kvm_arm_timer_read(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
1171
u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
1196
static void kvm_arm_timer_write(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
1223
void kvm_arm_timer_write_sysreg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
129
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctxt);
arch/arm64/kvm/arch_timer.c
1481
void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
1488
static bool timer_irqs_are_valid(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
1523
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
arch/arm64/kvm/arch_timer.c
154
void get_timer_map(struct kvm_vcpu *vcpu, struct timer_map *map)
arch/arm64/kvm/arch_timer.c
1542
int kvm_timer_enable(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
1595
int kvm_arm_timer_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/arch_timer.c
1647
int kvm_arm_timer_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/arch_timer.c
1674
int kvm_arm_timer_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/arch_timer.c
201
struct kvm_vcpu *vcpu = *(struct kvm_vcpu **)dev_id;
arch/arm64/kvm/arch_timer.c
262
static bool vcpu_has_wfit_active(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
268
static u64 wfit_delay_ns(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
282
static u64 kvm_timer_earliest_exp(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
308
struct kvm_vcpu *vcpu;
arch/arm64/kvm/arch_timer.c
312
vcpu = container_of(timer, struct kvm_vcpu, arch.timer_cpu);
arch/arm64/kvm/arch_timer.c
332
struct kvm_vcpu *vcpu;
arch/arm64/kvm/arch_timer.c
397
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
405
void kvm_timer_update_run(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
429
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctx);
arch/arm64/kvm/arch_timer.c
43
static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level,
arch/arm64/kvm/arch_timer.c
438
static void kvm_timer_update_irq(struct kvm_vcpu *vcpu, bool new_level,
arch/arm64/kvm/arch_timer.c
46
static void kvm_arm_timer_write(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
50
static u64 kvm_arm_timer_read(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
569
static void kvm_timer_blocking(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
59
static int nr_timers(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
594
static void kvm_timer_unblocking(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
662
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctx);
arch/arm64/kvm/arch_timer.c
681
static void kvm_timer_vcpu_load_nogic(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
69
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctxt);
arch/arm64/kvm/arch_timer.c
718
static void kvm_timer_vcpu_load_nested_switch(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arch_timer.c
754
static void timer_set_traps(struct kvm_vcpu *vcpu, struct timer_map *map)
arch/arm64/kvm/arch_timer.c
856
void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
88
struct kvm_vcpu *vcpu = timer_context_to_vcpu(ctxt);
arch/arm64/kvm/arch_timer.c
890
bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
907
void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
939
void kvm_timer_sync_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
982
static void unmask_vtimer_irq_user(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arch_timer.c
995
void kvm_timer_sync_user(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1009
void kvm_vcpu_wfi(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1035
static int kvm_vcpu_suspend(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1077
static int check_vcpu_requests(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1128
static bool vcpu_mode_is_bad_32bit(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1153
static bool kvm_vcpu_exit_request(struct kvm_vcpu *vcpu, int *ret)
arch/arm64/kvm/arm.c
1192
static int noinstr kvm_arm_vcpu_enter_exit(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1213
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
123
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1405
static int vcpu_interrupt_line(struct kvm_vcpu *vcpu, int number, bool level)
arch/arm64/kvm/arm.c
1444
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/arm.c
1516
static int kvm_vcpu_init_check_features(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1555
static bool kvm_vcpu_init_changed(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1564
static int kvm_setup_vcpu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
1583
static int __kvm_vcpu_set_target(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1613
static int kvm_vcpu_set_target(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1636
static int kvm_arch_vcpu_ioctl_vcpu_init(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1689
static int kvm_arm_vcpu_set_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1703
static int kvm_arm_vcpu_get_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1717
static int kvm_arm_vcpu_has_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1731
static int kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1739
static int kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
1760
struct kvm_vcpu *vcpu = filp->private_data;
arch/arm64/kvm/arm.c
262
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/arm64/kvm/arm.c
2797
struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr)
arch/arm64/kvm/arm.c
2799
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/arm.c
2969
err = kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/arm64/kvm/arm.c
507
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
553
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
557
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
569
void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
574
void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
579
static void vcpu_set_pauth_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
611
static bool kvm_vcpu_should_clear_twi(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
622
static bool kvm_vcpu_should_clear_twe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
630
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/arm64/kvm/arm.c
710
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
732
static void __kvm_arm_vcpu_power_off(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
739
void kvm_arm_vcpu_power_off(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
746
bool kvm_arm_vcpu_stopped(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
751
static void kvm_arm_vcpu_suspend(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
758
static bool kvm_arm_vcpu_suspended(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
763
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
771
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/arm64/kvm/arm.c
804
int kvm_arch_vcpu_runnable(struct kvm_vcpu *v)
arch/arm64/kvm/arm.c
812
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
818
unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
829
struct kvm_vcpu *vcpu;
arch/arm64/kvm/arm.c
881
int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu)
arch/arm64/kvm/arm.c
962
struct kvm_vcpu *vcpu;
arch/arm64/kvm/arm.c
972
struct kvm_vcpu *vcpu;
arch/arm64/kvm/arm.c
980
static void kvm_vcpu_sleep(struct kvm_vcpu *vcpu)
arch/arm64/kvm/at.c
100
static u64 effective_tcr2(struct kvm_vcpu *vcpu, enum trans_regime regime)
arch/arm64/kvm/at.c
1004
static void compute_s1_hierarchical_permissions(struct kvm_vcpu *vcpu,
arch/arm64/kvm/at.c
1121
static void compute_s1_indirect_permissions(struct kvm_vcpu *vcpu,
arch/arm64/kvm/at.c
113
static bool s1pie_enabled(struct kvm_vcpu *vcpu, enum trans_regime regime)
arch/arm64/kvm/at.c
1161
static void compute_s1_overlay_permissions(struct kvm_vcpu *vcpu,
arch/arm64/kvm/at.c
122
static void compute_s1poe(struct kvm_vcpu *vcpu, struct s1_walk_info *wi)
arch/arm64/kvm/at.c
1220
static void compute_s1_permissions(struct kvm_vcpu *vcpu,
arch/arm64/kvm/at.c
1250
static int handle_at_slow(struct kvm_vcpu *vcpu, u32 op, u64 vaddr, u64 *par)
arch/arm64/kvm/at.c
1326
static u64 __kvm_at_s1e01_fast(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
138
static int setup_s1_walk(struct kvm_vcpu *vcpu, struct s1_walk_info *wi,
arch/arm64/kvm/at.c
1442
int __kvm_at_s1e01(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1467
int __kvm_at_s1e2(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1525
int __kvm_at_s12(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1590
int __kvm_translate_va(struct kvm_vcpu *vcpu, struct s1_walk_info *wi,
arch/arm64/kvm/at.c
1635
int __kvm_find_s1_desc_level(struct kvm_vcpu *vcpu, u64 va, u64 ipa, int *level)
arch/arm64/kvm/at.c
36
static bool has_52bit_pa(struct kvm_vcpu *vcpu, struct s1_walk_info *wi, u64 tcr)
arch/arm64/kvm/at.c
370
static int kvm_read_s1_desc(struct kvm_vcpu *vcpu, u64 pa, u64 *desc,
arch/arm64/kvm/at.c
388
static int kvm_swap_s1_desc(struct kvm_vcpu *vcpu, u64 pa, u64 old, u64 new,
arch/arm64/kvm/at.c
402
static int walk_s1(struct kvm_vcpu *vcpu, struct s1_walk_info *wi,
arch/arm64/kvm/at.c
621
static bool at_s1e1p_fast(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
785
static u64 compute_par_s12(struct kvm_vcpu *vcpu, u64 s1_par,
arch/arm64/kvm/at.c
83
static enum trans_regime compute_translation_regime(struct kvm_vcpu *vcpu, u32 op)
arch/arm64/kvm/at.c
872
static u64 compute_par_s1(struct kvm_vcpu *vcpu, struct s1_walk_info *wi,
arch/arm64/kvm/at.c
928
static bool pan3_enabled(struct kvm_vcpu *vcpu, enum trans_regime regime)
arch/arm64/kvm/at.c
946
static void compute_s1_direct_permissions(struct kvm_vcpu *vcpu,
arch/arm64/kvm/config.c
1573
static __always_inline void __compute_fgt(struct kvm_vcpu *vcpu, enum vcpu_sysreg reg)
arch/arm64/kvm/config.c
1593
static void __compute_hfgwtr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/config.c
1601
static void __compute_hdfgwtr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/config.c
1609
void kvm_vcpu_load_fgt(struct kvm_vcpu *vcpu)
arch/arm64/kvm/debug.c
126
static void setup_external_mdscr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/debug.c
149
void kvm_vcpu_load_debug(struct kvm_vcpu *vcpu)
arch/arm64/kvm/debug.c
203
void kvm_vcpu_put_debug(struct kvm_vcpu *vcpu)
arch/arm64/kvm/debug.c
232
void kvm_debug_set_guest_ownership(struct kvm_vcpu *vcpu)
arch/arm64/kvm/debug.c
241
void kvm_debug_handle_oslar(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/debug.c
37
static void kvm_arm_setup_mdcr_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
2390
static enum trap_behaviour get_behaviour(struct kvm_vcpu *vcpu,
arch/arm64/kvm/emulate-nested.c
2403
static enum trap_behaviour __compute_trap_behaviour(struct kvm_vcpu *vcpu,
arch/arm64/kvm/emulate-nested.c
2429
static enum trap_behaviour compute_trap_behaviour(struct kvm_vcpu *vcpu,
arch/arm64/kvm/emulate-nested.c
2442
static bool check_fgt_bit(struct kvm_vcpu *vcpu, enum vcpu_sysreg sr,
arch/arm64/kvm/emulate-nested.c
2474
bool triage_sysreg_trap(struct kvm_vcpu *vcpu, int *sr_index)
arch/arm64/kvm/emulate-nested.c
2627
static bool __forward_traps(struct kvm_vcpu *vcpu, unsigned int reg, u64 control_bit)
arch/arm64/kvm/emulate-nested.c
2637
static bool forward_hcr_traps(struct kvm_vcpu *vcpu, u64 control_bit)
arch/arm64/kvm/emulate-nested.c
2642
bool forward_smc_trap(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
2647
static bool forward_mdcr_traps(struct kvm_vcpu *vcpu, u64 control_bit)
arch/arm64/kvm/emulate-nested.c
2652
bool forward_debug_exception(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
2657
static u64 kvm_check_illegal_exception_return(struct kvm_vcpu *vcpu, u64 spsr)
arch/arm64/kvm/emulate-nested.c
2690
void kvm_emulate_nested_eret(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
2738
static void kvm_inject_el2_exception(struct kvm_vcpu *vcpu, u64 esr_el2,
arch/arm64/kvm/emulate-nested.c
2763
static int kvm_inject_nested(struct kvm_vcpu *vcpu, u64 esr_el2,
arch/arm64/kvm/emulate-nested.c
2827
int kvm_inject_nested_sync(struct kvm_vcpu *vcpu, u64 esr_el2)
arch/arm64/kvm/emulate-nested.c
2832
int kvm_inject_nested_irq(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
2852
int kvm_inject_nested_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr)
arch/arm64/kvm/emulate-nested.c
2866
int kvm_inject_nested_serror(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/emulate-nested.c
498
static u64 get_sanitized_cnthctl(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
508
static enum trap_behaviour check_cnthctl_el1pcten(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
516
static enum trap_behaviour check_cnthctl_el1pten(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
524
static bool is_nested_nv2_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
532
static enum trap_behaviour check_cnthctl_el1nvpct(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
541
static enum trap_behaviour check_cnthctl_el1nvvct(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
550
static enum trap_behaviour check_cptr_tta(struct kvm_vcpu *vcpu)
arch/arm64/kvm/emulate-nested.c
563
static enum trap_behaviour check_mdcr_hpmn(struct kvm_vcpu *vcpu)
arch/arm64/kvm/fpsimd.c
101
void kvm_arch_vcpu_put_fp(struct kvm_vcpu *vcpu)
arch/arm64/kvm/fpsimd.c
24
void kvm_arch_vcpu_load_fp(struct kvm_vcpu *vcpu)
arch/arm64/kvm/fpsimd.c
52
void kvm_arch_vcpu_ctxflush_fp(struct kvm_vcpu *vcpu)
arch/arm64/kvm/fpsimd.c
65
void kvm_arch_vcpu_ctxsync_fp(struct kvm_vcpu *vcpu)
arch/arm64/kvm/guest.c
121
static void *core_reg_addr(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
186
static int get_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
215
static int set_core_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
313
static int get_sve_vls(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
337
static int set_sve_vls(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
419
struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
485
static int get_sve_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
511
static int set_sve_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
536
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/arm64/kvm/guest.c
541
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/arm64/kvm/guest.c
546
static int copy_core_reg_indices(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
589
static unsigned long num_core_regs(const struct kvm_vcpu *vcpu)
arch/arm64/kvm/guest.c
594
static unsigned long num_sve_regs(const struct kvm_vcpu *vcpu)
arch/arm64/kvm/guest.c
608
static int copy_sve_reg_indices(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
661
unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/guest.c
680
int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
arch/arm64/kvm/guest.c
702
int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
719
int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/guest.c
736
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
742
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
748
int __kvm_arm_vcpu_get_events(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
76
static int core_reg_size_from_offset(const struct kvm_vcpu *vcpu, u64 off)
arch/arm64/kvm/guest.c
767
static void commit_pending_events(struct kvm_vcpu *vcpu)
arch/arm64/kvm/guest.c
780
int __kvm_arm_vcpu_set_events(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
855
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/arm64/kvm/guest.c
860
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/arm64/kvm/guest.c
865
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
881
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
904
int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
929
int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/guest.c
952
int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/handle_exit.c
105
static int kvm_handle_fpasimd(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
130
static int kvm_handle_wfx(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
191
static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
216
static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
231
static int handle_sve(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
253
static int kvm_handle_ptrauth(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
271
static int kvm_handle_eret(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
296
static int handle_svc(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
30
typedef int (*exit_handle_fn)(struct kvm_vcpu *);
arch/arm64/kvm/handle_exit.c
307
static int kvm_handle_gcs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
317
static int handle_other(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
32
static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/handle_exit.c
38
static int handle_hvc(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
407
static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
421
static int handle_trap_exceptions(struct kvm_vcpu *vcpu)
arch/arm64/kvm/handle_exit.c
446
int handle_exit(struct kvm_vcpu *vcpu, int exception_index)
arch/arm64/kvm/handle_exit.c
490
void handle_exit_early(struct kvm_vcpu *vcpu, int exception_index)
arch/arm64/kvm/handle_exit.c
57
static int handle_smc(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/aarch32.c
110
static void kvm_adjust_itstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/aarch32.c
140
void kvm_skip_instr32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/aarch32.c
47
bool kvm_condition_valid32(const struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/exception.c
186
static unsigned long get_except32_cpsr(struct kvm_vcpu *vcpu, u32 mode)
arch/arm64/kvm/hyp/exception.c
23
static inline u64 __vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, int reg)
arch/arm64/kvm/hyp/exception.c
280
static void enter_exception32(struct kvm_vcpu *vcpu, u32 mode, u32 vect_offset)
arch/arm64/kvm/hyp/exception.c
31
static inline void __vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, int reg)
arch/arm64/kvm/hyp/exception.c
313
static void kvm_inject_exception(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/exception.c
367
void __kvm_adjust_pc(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/exception.c
39
static void __vcpu_write_spsr(struct kvm_vcpu *vcpu, unsigned long target_mode,
arch/arm64/kvm/hyp/exception.c
52
static void __vcpu_write_spsr_abt(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/hyp/exception.c
60
static void __vcpu_write_spsr_und(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/hyp/exception.c
85
static void enter_exception64(struct kvm_vcpu *vcpu, unsigned long target_mode,
arch/arm64/kvm/hyp/include/hyp/adjust_pc.h
16
static inline void kvm_skip_instr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/adjust_pc.h
33
static inline void __kvm_skip_instr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
134
static inline void __debug_switch_to_guest_common(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
153
static inline void __debug_switch_to_host_common(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
91
static struct kvm_guest_debug_arch *__vcpu_debug_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
153
static inline void __activate_cptr_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
164
static inline void __deactivate_cptr_traps_nvhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
176
static inline void __deactivate_cptr_traps_vhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
188
static inline void __deactivate_cptr_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
210
static inline void __activate_traps_hfgxtr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
242
static inline void __deactivate_traps_hfgxtr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
268
static inline void __activate_traps_mpam(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
297
static inline void __activate_traps_common(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
334
static inline void __deactivate_traps_common(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
351
static inline void ___activate_traps(struct kvm_vcpu *vcpu, u64 hcr)
arch/arm64/kvm/hyp/include/hyp/switch.h
387
static inline void ___deactivate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
411
static inline bool __populate_fault_info(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
416
static inline bool kvm_hyp_handle_mops(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/include/hyp/switch.h
43
static inline void __fpsimd_save_fpexc32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
432
static inline void __hyp_sve_restore_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
465
static inline void fpsimd_lazy_switch_to_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
486
static inline void fpsimd_lazy_switch_to_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
51
static inline void __activate_traps_fpsimd32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
524
static void kvm_hyp_save_fpsimd_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
548
static inline bool kvm_hyp_handle_fpsimd(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/include/hyp/switch.h
615
static inline bool handle_tx2_tvm(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
66
static inline void __activate_cptr_traps_nvhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
688
static bool kvm_handle_cntxct(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
745
static bool handle_ampere1_tcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
769
static inline bool kvm_hyp_handle_sysreg(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/include/hyp/switch.h
789
static inline bool kvm_hyp_handle_cp15_32(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/include/hyp/switch.h
798
static inline bool kvm_hyp_handle_memory_fault(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/include/hyp/switch.h
809
static inline bool kvm_hyp_handle_dabt_low(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/include/hyp/switch.h
837
typedef bool (*exit_handler_fn)(struct kvm_vcpu *, u64 *);
arch/arm64/kvm/hyp/include/hyp/switch.h
845
static inline bool kvm_hyp_handle_exit(struct kvm_vcpu *vcpu, u64 *exit_code,
arch/arm64/kvm/hyp/include/hyp/switch.h
85
static inline void __activate_cptr_traps_vhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
855
static inline void synchronize_vcpu_pstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/switch.h
878
static inline bool __fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_code,
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
103
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
114
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
125
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
21
static inline struct kvm_vcpu *ctxt_to_vcpu(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
23
struct kvm_vcpu *vcpu = ctxt->__hyp_running_vcpu;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
26
vcpu = container_of(ctxt, struct kvm_vcpu, arch.ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
343
static inline void __sysreg32_save_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
360
static inline void __sysreg32_restore_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
38
struct kvm_vcpu *vcpu = ctxt_to_vcpu(ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
74
struct kvm_vcpu *vcpu = ctxt_to_vcpu(ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
81
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
92
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
19
struct kvm_vcpu vcpu;
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
22
struct kvm_vcpu *host_vcpu;
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
74
int __pkvm_init_vcpu(pkvm_handle_t handle, struct kvm_vcpu *host_vcpu,
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
87
bool kvm_handle_pvm_sysreg(struct kvm_vcpu *vcpu, u64 *exit_code);
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
88
bool kvm_handle_pvm_restricted(struct kvm_vcpu *vcpu, u64 *exit_code);
arch/arm64/kvm/hyp/include/nvhe/pkvm.h
89
void kvm_init_pvm_id_regs(struct kvm_vcpu *vcpu);
arch/arm64/kvm/hyp/nvhe/debug-sr.c
121
void __debug_save_host_buffers_nvhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/debug-sr.c
135
void __debug_switch_to_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/debug-sr.c
140
void __debug_restore_host_buffers_nvhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/debug-sr.c
150
void __debug_switch_to_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
100
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
112
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
122
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
143
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
203
DECLARE_REG(struct kvm_vcpu *, host_vcpu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
231
struct kvm_vcpu *vcpu = kern_hyp_va(host_vcpu);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
244
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/hyp-main.c
27
static void __hyp_sve_save_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
379
DECLARE_REG(struct kvm_vcpu *, vcpu, host_ctxt, 1);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
578
DECLARE_REG(struct kvm_vcpu *, host_vcpu, host_ctxt, 2);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
65
static void fpsimd_sve_sync(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
134
static int pkvm_check_pvm_cpu_features(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
163
struct kvm_vcpu *vcpu = &hyp_vcpu->vcpu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
171
struct kvm_vcpu *host_vcpu = hyp_vcpu->host_vcpu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
31
static void pkvm_vcpu_reset_hcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
382
static void unpin_host_vcpu(struct kvm_vcpu *host_vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
439
static int pkvm_vcpu_init_sve(struct pkvm_hyp_vcpu *hyp_vcpu, struct kvm_vcpu *host_vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
441
struct kvm_vcpu *vcpu = &hyp_vcpu->vcpu;
arch/arm64/kvm/hyp/nvhe/pkvm.c
506
struct kvm_vcpu *host_vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
62
static void pvm_init_traps_hcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/pkvm.c
806
int __pkvm_init_vcpu(pkvm_handle_t handle, struct kvm_vcpu *host_vcpu,
arch/arm64/kvm/hyp/nvhe/pkvm.c
96
static void pvm_init_traps_mdcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
111
static void __hyp_vgic_save_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
120
static void __hyp_vgic_restore_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
132
static bool __pmu_switch_to_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
148
static void __pmu_switch_to_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
169
static bool kvm_handle_pvm_sys64(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/nvhe/switch.c
202
static const exit_handler_fn *kvm_get_exit_handler_array(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
210
static inline bool fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/nvhe/switch.c
242
int __kvm_vcpu_run(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
362
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/nvhe/switch.c
50
static void __activate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/switch.c
77
static void __deactivate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
188
static u64 get_restricted_features(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
217
static u64 pvm_calc_id_reg(const struct kvm_vcpu *vcpu, u32 id)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
246
static void inject_sync64(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
271
static void inject_undef64(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
276
static u64 read_id_reg(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
292
static bool pvm_access_raz_wi(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
307
static bool pvm_access_id_aarch32(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
326
static bool pvm_access_id_aarch64(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
339
static bool pvm_gic_read_sre(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
350
static bool pvm_idst_access(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/nvhe/sys_regs.c
513
void kvm_init_pvm_id_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
557
bool kvm_handle_pvm_sysreg(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/nvhe/sys_regs.c
595
bool kvm_handle_pvm_restricted(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/nvhe/timer-sr.c
23
void __timer_disable_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/timer-sr.c
41
void __timer_enable_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/nvhe/tlb.c
122
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/nvhe/tlb.c
25
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
18
static bool __is_be(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
37
int __vgic_v2_perform_cpuif_access(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1018
static void __vgic_v3_read_pmr(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1023
static void __vgic_v3_write_pmr(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1032
static void __vgic_v3_read_rpr(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1038
static void __vgic_v3_read_ctlr(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1059
static void __vgic_v3_write_ctlr(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1072
static bool __vgic_v3_check_trap_forwarding(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/vgic-v3-sr.c
1141
int __vgic_v3_perform_cpuif_access(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vgic-v3-sr.c
1146
void (*fn)(struct kvm_vcpu *, u32, int);
arch/arm64/kvm/hyp/vgic-v3-sr.c
546
static int __vgic_v3_get_group(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vgic-v3-sr.c
556
static int __vgic_v3_highest_priority_lr(struct kvm_vcpu *vcpu, u32 vmcr,
arch/arm64/kvm/hyp/vgic-v3-sr.c
595
static int __vgic_v3_find_active_lr(struct kvm_vcpu *vcpu, int intid,
arch/arm64/kvm/hyp/vgic-v3-sr.c
746
static void __vgic_v3_read_iar(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
802
static int ___vgic_v3_write_dir(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
825
static void __vgic_v3_write_dir(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
831
static void __vgic_v3_write_eoir(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
866
static void __vgic_v3_read_igrpen0(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
871
static void __vgic_v3_read_igrpen1(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
876
static void __vgic_v3_write_igrpen0(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
885
static void __vgic_v3_write_igrpen1(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
894
static void __vgic_v3_read_bpr0(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
899
static void __vgic_v3_read_bpr1(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
904
static void __vgic_v3_write_bpr0(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
918
static void __vgic_v3_write_bpr1(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
935
static void __vgic_v3_read_apxrn(struct kvm_vcpu *vcpu, int rt, int n)
arch/arm64/kvm/hyp/vgic-v3-sr.c
947
static void __vgic_v3_write_apxrn(struct kvm_vcpu *vcpu, int rt, int n)
arch/arm64/kvm/hyp/vgic-v3-sr.c
957
static void __vgic_v3_read_apxr0(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/vgic-v3-sr.c
963
static void __vgic_v3_read_apxr1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/vgic-v3-sr.c
969
static void __vgic_v3_read_apxr2(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
974
static void __vgic_v3_read_apxr3(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
979
static void __vgic_v3_write_apxr0(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
984
static void __vgic_v3_write_apxr1(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
989
static void __vgic_v3_write_apxr2(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
994
static void __vgic_v3_write_apxr3(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vgic-v3-sr.c
999
static void __vgic_v3_read_hppir(struct kvm_vcpu *vcpu, u32 vmcr, int rt)
arch/arm64/kvm/hyp/vhe/debug-sr.c
13
void __debug_switch_to_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/debug-sr.c
18
void __debug_switch_to_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
113
static void __activate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
145
static void __deactivate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
198
static void __vcpu_load_activate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
207
static void __vcpu_put_deactivate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
216
void kvm_vcpu_load_vhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
225
void kvm_vcpu_put_vhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
233
static u64 compute_emulated_cntx_ctl_el0(struct kvm_vcpu *vcpu,
arch/arm64/kvm/hyp/vhe/switch.c
261
static bool kvm_hyp_handle_timer(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
337
static bool kvm_hyp_handle_eret(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
390
static bool kvm_hyp_handle_tlbi_el2(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
439
static bool kvm_hyp_handle_cpacr_el1(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
461
static bool kvm_hyp_handle_zcr_el2(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
485
static bool kvm_hyp_handle_sysreg_vhe(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
502
static bool kvm_hyp_handle_impdef(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
52
static u64 __compute_hcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
537
static inline bool fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_code)
arch/arm64/kvm/hyp/vhe/switch.c
572
static int __kvm_vcpu_run_vhe(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
630
int __kvm_vcpu_run(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/switch.c
661
struct kvm_vcpu *vcpu;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
18
static void __sysreg_save_vel2_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
197
void __vcpu_load_switch_sysregs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
258
void __vcpu_put_switch_sysregs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
85
static void __sysreg_restore_vel2_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hyp/vhe/tlb.c
23
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
arch/arm64/kvm/hypercalls.c
21
static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val)
arch/arm64/kvm/hypercalls.c
229
static u8 kvm_smccc_get_action(struct kvm_vcpu *vcpu, u32 func_id)
arch/arm64/kvm/hypercalls.c
246
static void kvm_prepare_hypercall_exit(struct kvm_vcpu *vcpu, u32 func_id)
arch/arm64/kvm/hypercalls.c
265
int kvm_smccc_call_handler(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hypercalls.c
415
int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/hypercalls.c
420
int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
arch/arm64/kvm/hypercalls.c
438
static int get_kernel_wa_level(struct kvm_vcpu *vcpu, u64 regid)
arch/arm64/kvm/hypercalls.c
483
int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/hypercalls.c
520
static int kvm_arm_set_fw_reg_bmap(struct kvm_vcpu *vcpu, u64 reg_id, u64 val)
arch/arm64/kvm/hypercalls.c
565
int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/hypercalls.c
98
static bool kvm_smccc_test_fw_bmap(struct kvm_vcpu *vcpu, u32 func_id)
arch/arm64/kvm/inject_fault.c
100
static bool effective_sctlr2_nmea(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
105
static void inject_abt64(struct kvm_vcpu *vcpu, bool is_iabt, unsigned long addr)
arch/arm64/kvm/inject_fault.c
165
void kvm_inject_sync(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/inject_fault.c
171
static void inject_undef64(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
18
static unsigned int exception_target_el(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
190
static void inject_undef32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
199
static void inject_abt32(struct kvm_vcpu *vcpu, bool is_pabt, u32 addr)
arch/arm64/kvm/inject_fault.c
229
static void __kvm_inject_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr)
arch/arm64/kvm/inject_fault.c
237
static bool kvm_sea_target_is_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
249
int kvm_inject_sea(struct kvm_vcpu *vcpu, bool iabt, u64 addr)
arch/arm64/kvm/inject_fault.c
260
static int kvm_inject_nested_excl_atomic(struct kvm_vcpu *vcpu, u64 addr)
arch/arm64/kvm/inject_fault.c
279
int kvm_inject_dabt_excl_atomic(struct kvm_vcpu *vcpu, u64 addr)
arch/arm64/kvm/inject_fault.c
294
void kvm_inject_size_fault(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
326
void kvm_inject_undefined(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
334
static bool serror_is_masked(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
339
static bool kvm_serror_target_is_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
363
static bool kvm_serror_undeliverable_at_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
368
int kvm_inject_serror_esr(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/inject_fault.c
44
static enum vcpu_sysreg exception_esr_elx(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
52
static enum vcpu_sysreg exception_far_elx(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
60
static void pend_sync_exception(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
68
static void pend_serror_exception(struct kvm_vcpu *vcpu)
arch/arm64/kvm/inject_fault.c
76
static bool __effective_sctlr2_bit(struct kvm_vcpu *vcpu, unsigned int idx)
arch/arm64/kvm/inject_fault.c
95
static bool effective_sctlr2_ease(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmio.c
108
int kvm_handle_mmio_return(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmio.c
153
int io_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa)
arch/arm64/kvm/mmio.c
75
static bool kvm_pending_external_abort(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmu.c
1515
static int prepare_mmu_memcache(struct kvm_vcpu *vcpu, bool topup_memcache,
arch/arm64/kvm/mmu.c
1567
static int gmem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
arch/arm64/kvm/mmu.c
1642
static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa,
arch/arm64/kvm/mmu.c
1946
static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa)
arch/arm64/kvm/mmu.c
1963
static bool host_owns_sea(struct kvm_vcpu *vcpu, u64 esr)
arch/arm64/kvm/mmu.c
1987
int kvm_handle_guest_sea(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmu.c
2049
int kvm_handle_guest_abort(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmu.c
2530
void kvm_set_way_flush(struct kvm_vcpu *vcpu)
arch/arm64/kvm/mmu.c
2551
void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled)
arch/arm64/kvm/nested.c
1120
void kvm_handle_s1e2_tlbi(struct kvm_vcpu *vcpu, u32 inst, u64 val)
arch/arm64/kvm/nested.c
1242
int kvm_vcpu_allocate_vncr_tlb(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1255
static u64 read_vncr_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1260
static int kvm_translate_vncr(struct kvm_vcpu *vcpu, bool *is_gmem)
arch/arm64/kvm/nested.c
1343
static void inject_vncr_perm(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1356
static bool kvm_vncr_tlb_lookup(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1374
int kvm_handle_vncr_abort(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
141
static int esr_s2_fault(struct kvm_vcpu *vcpu, int level, u32 fsc)
arch/arm64/kvm/nested.c
1445
static void kvm_map_l1_vncr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
155
static int check_base_s2_limits(struct kvm_vcpu *vcpu, struct s2_walk_info *wi,
arch/arm64/kvm/nested.c
1680
u64 kvm_vcpu_apply_reg_masks(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/nested.c
1701
int kvm_init_nv_sysregs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1835
void check_nested_vcpu_requests(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1864
void kvm_nested_flush_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1873
void kvm_nested_sync_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
1911
void kvm_nested_setup_mdcr_el2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
201
static int read_guest_s2_desc(struct kvm_vcpu *vcpu, phys_addr_t pa, u64 *desc,
arch/arm64/kvm/nested.c
223
static int swap_guest_s2_desc(struct kvm_vcpu *vcpu, phys_addr_t pa, u64 old, u64 new,
arch/arm64/kvm/nested.c
244
static int walk_nested_s2_pgd(struct kvm_vcpu *vcpu, phys_addr_t ipa,
arch/arm64/kvm/nested.c
403
int kvm_walk_nested_s2(struct kvm_vcpu *vcpu, phys_addr_t gipa,
arch/arm64/kvm/nested.c
658
struct kvm_s2_mmu *lookup_s2_mmu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
68
int kvm_vcpu_init_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
707
static struct kvm_s2_mmu *get_s2_mmu_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
776
void kvm_vcpu_load_hw_mmu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
796
void kvm_vcpu_put_hw_mmu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/nested.c
827
int kvm_s2_handle_perm_fault(struct kvm_vcpu *vcpu, struct kvm_s2_trans *trans)
arch/arm64/kvm/nested.c
854
int kvm_inject_s2_fault(struct kvm_vcpu *vcpu, u64 esr_el2)
arch/arm64/kvm/nested.c
862
u16 get_asid_by_regime(struct kvm_vcpu *vcpu, enum trans_regime regime)
arch/arm64/kvm/nested.c
898
struct kvm_vcpu *vcpu;
arch/arm64/kvm/nested.c
952
struct kvm_vcpu *vcpu;
arch/arm64/kvm/nested.c
999
static void compute_s1_tlbi_range(struct kvm_vcpu *vcpu, u32 inst, u64 val,
arch/arm64/kvm/pauth.c
101
static u64 compute_pac_mask(struct kvm_vcpu *vcpu, bool bit55)
arch/arm64/kvm/pauth.c
115
static u64 to_canonical_addr(struct kvm_vcpu *vcpu, u64 ptr, u64 mask)
arch/arm64/kvm/pauth.c
125
static u64 corrupt_addr(struct kvm_vcpu *vcpu, u64 ptr)
arch/arm64/kvm/pauth.c
156
bool kvm_auth_eretax(struct kvm_vcpu *vcpu, u64 *elr)
arch/arm64/kvm/pauth.c
31
static u64 compute_pac(struct kvm_vcpu *vcpu, u64 ptr,
arch/arm64/kvm/pauth.c
61
static bool effective_tbi(struct kvm_vcpu *vcpu, bool bit55)
arch/arm64/kvm/pauth.c
86
static int compute_bottom_pac(struct kvm_vcpu *vcpu, bool bit55)
arch/arm64/kvm/pkvm.c
107
static int __pkvm_create_hyp_vcpu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pkvm.c
203
int pkvm_create_hyp_vcpu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
1034
struct kvm_vcpu *vcpu;
arch/arm64/kvm/pmu-emul.c
1077
static int kvm_arm_pmu_v3_set_pmu(struct kvm_vcpu *vcpu, int pmu_id)
arch/arm64/kvm/pmu-emul.c
1107
static int kvm_arm_pmu_v3_set_nr_counters(struct kvm_vcpu *vcpu, unsigned int n)
arch/arm64/kvm/pmu-emul.c
1121
int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/pmu-emul.c
1233
int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/pmu-emul.c
1257
int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/pmu-emul.c
130
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
1300
u64 kvm_vcpu_read_pmcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
1311
void kvm_pmu_nested_transition(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
155
u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx)
arch/arm64/kvm/pmu-emul.c
162
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
193
void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val)
arch/arm64/kvm/pmu-emul.c
204
void kvm_pmu_set_counter_value_user(struct kvm_vcpu *vcpu, u64 select_idx, u64 val)
arch/arm64/kvm/pmu-emul.c
232
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
252
void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
266
void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
275
static u64 kvm_pmu_hyp_counter_mask(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
303
bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx)
arch/arm64/kvm/pmu-emul.c
308
u64 kvm_pmu_accessible_counter_mask(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
318
u64 kvm_pmu_implemented_counter_mask(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
33
static struct kvm_vcpu *kvm_pmc_to_vcpu(const struct kvm_pmc *pmc)
arch/arm64/kvm/pmu-emul.c
346
void kvm_pmu_reprogram_counter_mask(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/pmu-emul.c
35
return container_of(pmc, struct kvm_vcpu, arch.pmu.pmc[pmc->idx]);
arch/arm64/kvm/pmu-emul.c
373
static bool kvm_pmu_overflow_status(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
38
static struct kvm_pmc *kvm_vcpu_idx_to_pmc(struct kvm_vcpu *vcpu, int cnt_idx)
arch/arm64/kvm/pmu-emul.c
396
static void kvm_pmu_update_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
414
bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
429
void kvm_pmu_update_run(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
446
void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
458
void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
470
struct kvm_vcpu *vcpu;
arch/arm64/kvm/pmu-emul.c
472
vcpu = container_of(work, struct kvm_vcpu, arch.pmu.overflow_work);
arch/arm64/kvm/pmu-emul.c
481
static void kvm_pmu_counter_increment(struct kvm_vcpu *vcpu,
arch/arm64/kvm/pmu-emul.c
543
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
582
void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/pmu-emul.c
592
void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/pmu-emul.c
624
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
656
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
685
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
775
void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data,
arch/arm64/kvm/pmu-emul.c
879
u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1)
arch/arm64/kvm/pmu-emul.c
89
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
913
void kvm_vcpu_reload_pmu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
924
int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
951
static int kvm_arm_pmu_v3_init(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu-emul.c
97
struct kvm_vcpu *vcpu = kvm_pmc_to_vcpu(pmc);
arch/arm64/kvm/pmu-emul.c
988
struct kvm_vcpu *vcpu;
arch/arm64/kvm/pmu.c
131
void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu.c
152
void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pmu.c
181
struct kvm_vcpu *vcpu;
arch/arm64/kvm/pmu.c
201
struct kvm_vcpu *vcpu;
arch/arm64/kvm/psci.c
121
static unsigned long kvm_psci_vcpu_affinity_info(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
129
struct kvm_vcpu *tmp;
arch/arm64/kvm/psci.c
164
static void kvm_prepare_system_event(struct kvm_vcpu *vcpu, u32 type, u64 flags)
arch/arm64/kvm/psci.c
167
struct kvm_vcpu *tmp;
arch/arm64/kvm/psci.c
192
static void kvm_psci_system_off(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
197
static void kvm_psci_system_off2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
203
static void kvm_psci_system_reset(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
208
static void kvm_psci_system_reset2(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
214
static void kvm_psci_system_suspend(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
223
static void kvm_psci_narrow_to_32bit(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
235
static unsigned long kvm_psci_check_allowed_function(struct kvm_vcpu *vcpu, u32 fn)
arch/arm64/kvm/psci.c
246
static int kvm_psci_0_2_call(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
321
static int kvm_psci_1_x_call(struct kvm_vcpu *vcpu, u32 minor)
arch/arm64/kvm/psci.c
34
static unsigned long kvm_psci_vcpu_suspend(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
441
static int kvm_psci_0_1_call(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
477
int kvm_psci_call(struct kvm_vcpu *vcpu)
arch/arm64/kvm/psci.c
54
static inline bool kvm_psci_valid_affinity(struct kvm_vcpu *vcpu,
arch/arm64/kvm/psci.c
60
static unsigned long kvm_psci_vcpu_on(struct kvm_vcpu *source_vcpu)
arch/arm64/kvm/psci.c
64
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/pvtime.c
107
int kvm_arm_pvtime_get_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/pvtime.c
124
int kvm_arm_pvtime_has_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/pvtime.c
13
void kvm_update_stolen_time(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pvtime.c
35
long kvm_hypercall_pv_features(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pvtime.c
51
gpa_t kvm_init_stolen_time(struct kvm_vcpu *vcpu)
arch/arm64/kvm/pvtime.c
75
int kvm_arm_pvtime_set_attr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/reset.c
129
int kvm_arm_vcpu_finalize(struct kvm_vcpu *vcpu, int feature)
arch/arm64/kvm/reset.c
145
bool kvm_arm_vcpu_is_finalized(struct kvm_vcpu *vcpu)
arch/arm64/kvm/reset.c
153
void kvm_arm_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/arm64/kvm/reset.c
166
static void kvm_vcpu_reset_sve(struct kvm_vcpu *vcpu)
arch/arm64/kvm/reset.c
190
void kvm_reset_vcpu(struct kvm_vcpu *vcpu)
arch/arm64/kvm/reset.c
79
static void kvm_vcpu_enable_sve(struct kvm_vcpu *vcpu)
arch/arm64/kvm/reset.c
95
static int kvm_vcpu_finalize_sve(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
1004
static bool pmu_write_swinc_el0_disabled(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
1009
static bool pmu_access_cycle_counter_el0_disabled(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
1014
static bool pmu_access_event_counter_el0_disabled(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
1019
static bool access_pmcr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1048
static bool access_pmselr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1064
static bool access_pmceid(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1085
static bool pmu_counter_idx_valid(struct kvm_vcpu *vcpu, u64 idx)
arch/arm64/kvm/sys_regs.c
1099
static int get_pmu_evcntr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/sys_regs.c
1115
static int set_pmu_evcntr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/sys_regs.c
1131
static bool access_pmu_evcntr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1184
static bool access_pmu_evtyper(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1220
static int set_pmreg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r, u64 val)
arch/arm64/kvm/sys_regs.c
1230
static int get_pmreg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r, u64 *val)
arch/arm64/kvm/sys_regs.c
1238
static bool access_pmcnten(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1264
static bool access_pminten(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1288
static bool access_pmovs(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1310
static bool access_pmswinc(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1326
static bool access_pmuserenr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
1343
static int get_pmcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/sys_regs.c
1350
static int set_pmcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/sys_regs.c
1429
static unsigned int ptrauth_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1449
static bool access_arch_timer(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
147
static void locate_mapped_el2_register(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1612
static int arch_timer_set_user(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1637
static int arch_timer_get_user(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1693
static int arm64_check_features(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
175
static void locate_register(const struct kvm_vcpu *vcpu, enum vcpu_sysreg reg,
arch/arm64/kvm/sys_regs.c
1759
static u64 sanitise_id_aa64pfr0_el1(const struct kvm_vcpu *vcpu, u64 val);
arch/arm64/kvm/sys_regs.c
1760
static u64 sanitise_id_aa64pfr1_el1(const struct kvm_vcpu *vcpu, u64 val);
arch/arm64/kvm/sys_regs.c
1761
static u64 sanitise_id_aa64dfr0_el1(const struct kvm_vcpu *vcpu, u64 val);
arch/arm64/kvm/sys_regs.c
1764
static u64 __kvm_read_sanitised_id_reg(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1834
static u64 kvm_read_sanitised_id_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1840
static u64 read_id_reg(const struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
1889
static unsigned int id_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1904
static unsigned int aa32_id_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1918
static unsigned int raz_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1926
static bool access_id_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1939
static unsigned int sve_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1948
static unsigned int sme_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1957
static unsigned int fp8_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
1966
static u64 sanitise_id_aa64pfr0_el1(const struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/sys_regs.c
2005
static u64 sanitise_id_aa64pfr1_el1(const struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/sys_regs.c
2030
static u64 sanitise_id_aa64dfr0_el1(const struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/sys_regs.c
2056
static bool ignore_feat_doublelock(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/sys_regs.c
2071
static int set_id_aa64dfr0_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2110
static u64 read_sanitised_id_dfr0_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2127
static int set_id_dfr0_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2154
static int set_id_aa64pfr0_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2191
static int set_id_aa64pfr1_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2239
static int set_id_aa64mmfr0_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2255
static int set_id_aa64mmfr2_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2271
static int set_ctr_el0(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2306
static int get_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2325
static int set_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2377
static int get_raz_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2384
static int set_wi_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2390
static bool access_ctr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
2400
static bool access_clidr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
2414
static u64 reset_clidr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
2466
static int set_clidr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2480
static bool access_csselr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
2492
static bool access_ccsidr(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
2508
static unsigned int mte_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2525
static unsigned int el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2534
static bool bad_vncr_trap(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2547
static bool bad_redir_trap(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2664
static bool access_sp_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2676
static bool access_elr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2688
static bool access_spsr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2700
static bool access_cntkctl_el12(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2712
static u64 reset_hcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
2724
static unsigned int __el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2726
unsigned int (*fn)(const struct kvm_vcpu *,
arch/arm64/kvm/sys_regs.c
2732
static unsigned int sve_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2738
static unsigned int vncr_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2748
static unsigned int sctlr2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2757
static unsigned int sctlr2_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2763
static bool access_zcr_el2(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2785
static bool access_gic_vtr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2797
static bool access_gic_misr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2809
static bool access_gic_eisr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2821
static bool access_gic_elrsr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2833
static unsigned int s1poe_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2842
static unsigned int s1poe_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2848
static unsigned int tcr2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2857
static unsigned int tcr2_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2863
static unsigned int fgt2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2873
static unsigned int fgt_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2883
static unsigned int s1pie_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2892
static unsigned int s1pie_el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2898
static unsigned int cnthv_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2908
static bool access_mdcr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
294
u64 vcpu_read_sys_reg(const struct kvm_vcpu *vcpu, enum vcpu_sysreg reg)
arch/arm64/kvm/sys_regs.c
2944
static bool access_ras(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
2988
static bool access_imp_id_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
3031
static u64 reset_imp_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
3046
static int set_imp_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/sys_regs.c
3088
static u64 reset_mdcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
341
void vcpu_write_sys_reg(struct kvm_vcpu *vcpu, u64 val, enum vcpu_sysreg reg)
arch/arm64/kvm/sys_regs.c
3784
static bool handle_at_s1e01(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
3795
static bool handle_at_s1e2(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
3813
static bool handle_at_s12(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
3824
static bool kvm_supported_tlbi_s12_op(struct kvm_vcpu *vpcu, u32 instr)
arch/arm64/kvm/sys_regs.c
3840
static bool handle_alle1is(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
3861
static bool kvm_supported_tlbi_ipas2_op(struct kvm_vcpu *vpcu, u32 instr)
arch/arm64/kvm/sys_regs.c
3932
static bool handle_vmalls12e1is(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
3956
static bool handle_ripas2e1is(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
4011
static bool handle_ipas2e1is(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
4037
static bool handle_tlbi_el2(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
4049
static bool handle_tlbi_el1(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
423
static u32 get_ccsidr(struct kvm_vcpu *vcpu, u32 csselr)
arch/arm64/kvm/sys_regs.c
4282
static bool trap_dbgdidr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
458
static int set_ccsidr(struct kvm_vcpu *vcpu, u32 csselr, u32 val)
arch/arm64/kvm/sys_regs.c
4649
int kvm_handle_cp14_load_store(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
4655
static void perform_access(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4692
static bool emulate_cp(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4713
static void unhandled_cp_access(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4744
static int kvm_handle_cp_64(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4789
static bool emulate_sys_reg(struct kvm_vcpu *vcpu, struct sys_reg_params *params);
arch/arm64/kvm/sys_regs.c
4844
int kvm_handle_cp10_id(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
487
static bool access_rw(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4879
static int kvm_emulate_cp15_id_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
49
static int set_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
4913
static int kvm_handle_cp_32(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
4932
int kvm_handle_cp15_64(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
4937
int kvm_handle_cp15_32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
4959
int kvm_handle_cp14_64(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
4964
int kvm_handle_cp14_32(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
4980
static bool emulate_sys_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
502
static bool access_dcsw(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
5136
static void reset_vm_ftr_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *reg)
arch/arm64/kvm/sys_regs.c
5147
static void reset_vcpu_ftr_id_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
5163
void kvm_reset_sys_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
5196
int kvm_handle_sys_reg(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
52
static bool undef_access(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
522
static bool access_dcgsw(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
5275
id_to_sys_reg_desc(struct kvm_vcpu *vcpu, u64 id,
arch/arm64/kvm/sys_regs.c
5294
static int demux_c15_get(struct kvm_vcpu *vcpu, u64 id, void __user *uaddr)
arch/arm64/kvm/sys_regs.c
5319
static int demux_c15_set(struct kvm_vcpu *vcpu, u64 id, void __user *uaddr)
arch/arm64/kvm/sys_regs.c
5359
int kvm_sys_reg_get_user(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg,
arch/arm64/kvm/sys_regs.c
5385
int kvm_arm_sys_reg_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/sys_regs.c
5396
int kvm_sys_reg_set_user(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg,
arch/arm64/kvm/sys_regs.c
5425
int kvm_arm_sys_reg_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/arm64/kvm/sys_regs.c
5491
static int walk_one_sys_reg(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
5514
static int walk_sys_regs(struct kvm_vcpu *vcpu, u64 __user *uind)
arch/arm64/kvm/sys_regs.c
5531
unsigned long kvm_arm_num_sys_reg_descs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
5537
int kvm_arm_copy_sys_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices)
arch/arm64/kvm/sys_regs.c
556
static bool access_vm_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
5591
static void vcpu_set_hcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
5631
void kvm_calculate_traps(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
5664
int kvm_finalize_sys_regs(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.c
581
static bool access_actlr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
59
static bool bad_trap(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
602
static bool access_gic_sgi(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
650
static bool access_gic_sre(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
669
static bool access_gic_dir(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
684
static bool trap_raz_wi(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
69
static bool read_from_write_only(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
700
static bool trap_loregion(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
715
static bool trap_oslar_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
726
static bool trap_oslsr_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
737
static int set_oslsr_el1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
751
static bool trap_dbgauthstatus_el1(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
763
static bool trap_debug_regs(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
77
static bool write_to_read_only(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
779
static void reg_to_dbg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
794
static void dbg_to_reg(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
805
static u64 *demux_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
824
static bool trap_dbg_wb_reg(struct kvm_vcpu *vcpu, struct sys_reg_params *p,
arch/arm64/kvm/sys_regs.c
841
static int set_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
853
static int get_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
865
static u64 reset_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
881
static u64 reset_amair_el1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
888
static u64 reset_actlr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
895
static u64 reset_mpidr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
915
static unsigned int hidden_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
921
static unsigned int pmu_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
930
static u64 reset_pmu_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
944
static u64 reset_pmevcntr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
952
static u64 reset_pmevtyper(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
964
static u64 reset_pmselr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
972
static u64 reset_pmcr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.c
988
static bool check_pmu_access_disabled(struct kvm_vcpu *vcpu, u64 flags)
arch/arm64/kvm/sys_regs.c
99
static enum sr_loc_attr locate_direct_register(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.c
999
static bool pmu_access_el0_disabled(struct kvm_vcpu *vcpu)
arch/arm64/kvm/sys_regs.h
103
unsigned int (*visibility)(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
131
static inline bool ignore_write(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
137
static inline bool read_zero(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
145
static inline u64 reset_unknown(struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
154
static inline u64 reset_val(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r)
arch/arm64/kvm/sys_regs.h
162
static inline unsigned int sysreg_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
171
static inline bool sysreg_hidden(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
177
static inline bool sysreg_visible_as_raz(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
183
static inline bool sysreg_user_write_ignore(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/sys_regs.h
229
int kvm_arm_sys_reg_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *);
arch/arm64/kvm/sys_regs.h
230
int kvm_arm_sys_reg_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *);
arch/arm64/kvm/sys_regs.h
231
int kvm_sys_reg_get_user(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg,
arch/arm64/kvm/sys_regs.h
233
int kvm_sys_reg_set_user(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg,
arch/arm64/kvm/sys_regs.h
236
bool triage_sysreg_trap(struct kvm_vcpu *vcpu, int *sr_index);
arch/arm64/kvm/sys_regs.h
238
int kvm_finalize_sys_regs(struct kvm_vcpu *vcpu);
arch/arm64/kvm/sys_regs.h
80
bool (*access)(struct kvm_vcpu *,
arch/arm64/kvm/sys_regs.h
88
u64 (*reset)(struct kvm_vcpu *, const struct sys_reg_desc *);
arch/arm64/kvm/sys_regs.h
97
int (*get_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.h
99
int (*set_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/trace_arm.h
335
TP_PROTO(struct kvm_vcpu *vcpu, unsigned long elr_el2,
arch/arm64/kvm/trace_arm.h
340
__field(struct kvm_vcpu *, vcpu)
arch/arm64/kvm/trace_arm.h
362
TP_PROTO(struct kvm_vcpu *vcpu, u64 esr_el2, int type),
arch/arm64/kvm/trace_arm.h
366
__field(struct kvm_vcpu *, vcpu)
arch/arm64/kvm/trace_arm.h
393
TP_PROTO(struct kvm_vcpu *vcpu, u32 sysreg, bool is_read),
arch/arm64/kvm/trace_handle_exit.h
120
TP_PROTO(struct kvm_vcpu *vcpu, __u32 guest_debug),
arch/arm64/kvm/trace_handle_exit.h
124
__field(struct kvm_vcpu *, vcpu)
arch/arm64/kvm/trng.c
24
static int kvm_trng_do_rnd(struct kvm_vcpu *vcpu, int size)
arch/arm64/kvm/trng.c
51
int kvm_trng_call(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic-sys-reg-v3.c
110
static int set_gic_bpr0(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
122
static int get_gic_bpr0(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
13
static int set_gic_ctlr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
133
static int set_gic_bpr1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
147
static int get_gic_bpr1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
162
static int set_gic_grpen0(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
174
static int get_gic_grpen0(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
185
static int set_gic_grpen1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
197
static int get_gic_grpen1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
208
static void set_apr_reg(struct kvm_vcpu *vcpu, u64 val, u8 apr, u8 idx)
arch/arm64/kvm/vgic-sys-reg-v3.c
218
static u64 get_apr_reg(struct kvm_vcpu *vcpu, u8 apr, u8 idx)
arch/arm64/kvm/vgic-sys-reg-v3.c
228
static int set_gic_ap0r(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
241
static int get_gic_ap0r(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
254
static int set_gic_ap1r(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
267
static int get_gic_ap1r(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
280
static int set_gic_sre(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
290
static int get_gic_sre(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
300
static int set_gic_ich_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
307
static int get_gic_ich_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
314
static int set_gic_ich_apr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
325
static int get_gic_ich_apr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
336
static int set_gic_icc_sre(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
344
static int get_gic_icc_sre(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
351
static int set_gic_ich_vtr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
359
static int get_gic_ich_vtr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
366
static unsigned int el2_visibility(const struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic-sys-reg-v3.c
461
int vgic_v3_has_cpu_sysregs_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr)
arch/arm64/kvm/vgic-sys-reg-v3.c
474
int vgic_v3_cpu_sysregs_uaccess(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic-sys-reg-v3.c
59
static int get_gic_ctlr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
87
static int set_gic_pmr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic-sys-reg-v3.c
99
static int get_gic_pmr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *r,
arch/arm64/kvm/vgic/vgic-debug.c
166
struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-debug.c
182
struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-debug.c
239
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/vgic/vgic-init.c
198
struct kvm_vcpu *vcpu0 = kvm_get_vcpu(kvm, 0);
arch/arm64/kvm/vgic/vgic-init.c
244
int kvm_vgic_vcpu_nv_init(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-init.c
262
static int vgic_allocate_private_irqs_locked(struct kvm_vcpu *vcpu, u32 type)
arch/arm64/kvm/vgic/vgic-init.c
316
static int vgic_allocate_private_irqs(struct kvm_vcpu *vcpu, u32 type)
arch/arm64/kvm/vgic/vgic-init.c
336
int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-init.c
367
static void kvm_vgic_vcpu_reset(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-init.c
387
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-init.c
458
static void __kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-init.c
496
void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-init.c
507
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-init.c
61
static int vgic_allocate_private_irqs_locked(struct kvm_vcpu *vcpu, u32 type);
arch/arm64/kvm/vgic/vgic-init.c
635
struct kvm_vcpu *vcpu = *(struct kvm_vcpu **)data;
arch/arm64/kvm/vgic/vgic-init.c
73
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-its.c
1029
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/vgic/vgic-its.c
1208
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-its.c
1294
int vgic_its_invall(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-its.c
1329
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-its.c
1353
struct kvm_vcpu *vcpu1, *vcpu2;
arch/arm64/kvm/vgic/vgic-its.c
1796
void vgic_enable_lpis(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-its.c
2134
struct kvm_vcpu *vcpu = NULL;
arch/arm64/kvm/vgic/vgic-its.c
281
struct kvm_vcpu *filter_vcpu, bool needs_inv)
arch/arm64/kvm/vgic/vgic-its.c
313
static int update_affinity(struct vgic_irq *irq, struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-its.c
32
struct kvm_vcpu *filter_vcpu, bool needs_inv);
arch/arm64/kvm/vgic/vgic-its.c
336
static struct kvm_vcpu *collection_to_vcpu(struct kvm *kvm,
arch/arm64/kvm/vgic/vgic-its.c
350
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-its.c
389
static int its_sync_lpi_pending_table(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-its.c
626
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-its.c
77
struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-its.c
825
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-kvm-device.c
371
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-kvm-device.c
538
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
110
static int vgic_mmio_uaccess_write_v2_group(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
120
static void vgic_mmio_write_sgir(struct kvm_vcpu *source_vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
128
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
162
static unsigned long vgic_mmio_read_target(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
180
static void vgic_mmio_write_target(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
208
static unsigned long vgic_mmio_read_sgipend(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
225
static void vgic_mmio_write_sgipendc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
247
static void vgic_mmio_write_sgipends(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
25
static unsigned long vgic_mmio_read_v2_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
275
static unsigned long vgic_mmio_read_vcpuif(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
322
static void vgic_mmio_write_vcpuif(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
362
static void vgic_mmio_write_dir(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
372
static unsigned long vgic_mmio_read_apr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
397
static void vgic_mmio_write_apr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
52
static void vgic_mmio_write_v2_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
526
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
563
int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
575
int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic-mmio-v2.c
72
static int vgic_mmio_uaccess_write_v2_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1004
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1053
static void vgic_v3_queue_sgi(struct kvm_vcpu *vcpu, u32 sgi, bool allow_group1)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1101
void vgic_v3_dispatch_sgi(struct kvm_vcpu *vcpu, u64 reg, bool allow_group1)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1104
struct kvm_vcpu *c_vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1138
int vgic_v3_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1149
int vgic_v3_redist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1160
int vgic_v3_line_level_info_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
121
static void vgic_mmio_write_v3_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
170
static int vgic_mmio_uaccess_write_v3_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
220
static unsigned long vgic_mmio_read_irouter(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
238
static void vgic_mmio_write_irouter(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
265
bool vgic_lpis_enabled(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
272
static unsigned long vgic_mmio_read_v3r_ctlr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
285
static void vgic_mmio_write_v3r_ctlr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
319
static bool vgic_mmio_vcpu_rdist_is_last(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
346
static unsigned long vgic_mmio_read_v3r_typer(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
365
static unsigned long vgic_mmio_read_v3r_iidr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
371
static unsigned long vgic_mmio_read_v3_idregs(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
383
static int vgic_v3_uaccess_write_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
479
static unsigned long vgic_mmio_read_propbase(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
487
static void vgic_mmio_write_propbase(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
507
static unsigned long vgic_mmio_read_pendbase(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
518
static void vgic_mmio_write_pendbase(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
538
static unsigned long vgic_mmio_read_sync(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
544
static void vgic_set_rdist_busy(struct kvm_vcpu *vcpu, bool busy)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
555
static void vgic_mmio_write_invlpi(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
587
static void vgic_mmio_write_invall(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
778
int vgic_register_redist_iodev(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
81
static unsigned long vgic_mmio_read_v3_misc(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio-v3.c
837
void vgic_unregister_redist_iodev(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
846
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
956
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-mmio.c
1005
int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev,
arch/arm64/kvm/vgic/vgic-mmio.c
1014
static int dispatch_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/arm64/kvm/vgic/vgic-mmio.c
1046
static int dispatch_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/arm64/kvm/vgic/vgic-mmio.c
116
void vgic_mmio_write_senable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
165
void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
187
int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
19
unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
208
int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
229
static unsigned long __read_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
25
unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
286
unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
292
unsigned long vgic_uaccess_read_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
298
static bool is_vgic_v2_sgi(struct kvm_vcpu *vcpu, struct vgic_irq *irq)
arch/arm64/kvm/vgic/vgic-mmio.c
304
static void __set_pending(struct kvm_vcpu *vcpu, gpa_t addr, unsigned int len,
arch/arm64/kvm/vgic/vgic-mmio.c
31
void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.c
353
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
360
int vgic_uaccess_write_spending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
369
static void vgic_hw_irq_cpending(struct kvm_vcpu *vcpu, struct vgic_irq *irq)
arch/arm64/kvm/vgic/vgic-mmio.c
37
int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.c
389
static void __clear_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
44
unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
440
void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
447
int vgic_uaccess_write_cpending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
471
static void vgic_access_active_prepare(struct kvm_vcpu *vcpu, u32 intid)
arch/arm64/kvm/vgic/vgic-mmio.c
480
static void vgic_access_active_finish(struct kvm_vcpu *vcpu, u32 intid)
arch/arm64/kvm/vgic/vgic-mmio.c
488
static unsigned long __vgic_mmio_read_active(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
512
unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
529
unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
536
static void vgic_hw_irq_change_active(struct kvm_vcpu *vcpu, struct vgic_irq *irq,
arch/arm64/kvm/vgic/vgic-mmio.c
546
static void vgic_mmio_change_active(struct kvm_vcpu *vcpu, struct vgic_irq *irq,
arch/arm64/kvm/vgic/vgic-mmio.c
550
struct kvm_vcpu *requester_vcpu = kvm_get_running_vcpu();
arch/arm64/kvm/vgic/vgic-mmio.c
593
static void __vgic_mmio_write_cactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
607
void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
622
int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
630
static void __vgic_mmio_write_sactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
644
void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
659
int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
667
unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
69
void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.c
692
void vgic_mmio_write_priority(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
714
unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
733
void vgic_mmio_write_config(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
766
u32 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid)
arch/arm64/kvm/vgic/vgic-mmio.c
788
void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid,
arch/arm64/kvm/vgic/vgic-mmio.c
843
void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
arch/arm64/kvm/vgic/vgic-mmio.c
851
void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr)
arch/arm64/kvm/vgic/vgic-mmio.c
950
vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
arch/arm64/kvm/vgic/vgic-mmio.c
96
unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.c
963
static int vgic_uaccess_read(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
arch/arm64/kvm/vgic/vgic-mmio.c
967
struct kvm_vcpu *r_vcpu;
arch/arm64/kvm/vgic/vgic-mmio.c
984
static int vgic_uaccess_write(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
arch/arm64/kvm/vgic/vgic-mmio.c
988
struct kvm_vcpu *r_vcpu;
arch/arm64/kvm/vgic/vgic-mmio.h
112
unsigned long vgic_mmio_read_raz(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
115
unsigned long vgic_mmio_read_rao(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
118
void vgic_mmio_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
121
int vgic_mmio_uaccess_write_wi(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
124
unsigned long vgic_mmio_read_group(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
127
void vgic_mmio_write_group(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
130
unsigned long vgic_mmio_read_enable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
133
void vgic_mmio_write_senable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
137
void vgic_mmio_write_cenable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
14
unsigned long (*read)(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
141
int vgic_uaccess_write_senable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
145
int vgic_uaccess_write_cenable(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
149
unsigned long vgic_mmio_read_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
152
unsigned long vgic_uaccess_read_pending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
155
void vgic_mmio_write_spending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
159
void vgic_mmio_write_cpending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
163
int vgic_uaccess_write_spending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
167
int vgic_uaccess_write_cpending(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
171
unsigned long vgic_mmio_read_active(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
174
unsigned long vgic_uaccess_read_active(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
177
void vgic_mmio_write_cactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
181
void vgic_mmio_write_sactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
185
int vgic_mmio_uaccess_write_cactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
189
int vgic_mmio_uaccess_write_sactive(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
193
unsigned long vgic_mmio_read_priority(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
196
void vgic_mmio_write_priority(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
20
void (*write)(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
200
unsigned long vgic_mmio_read_config(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
203
void vgic_mmio_write_config(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-mmio.h
207
int vgic_uaccess(struct kvm_vcpu *vcpu, struct vgic_io_device *dev,
arch/arm64/kvm/vgic/vgic-mmio.h
210
u32 vgic_read_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid);
arch/arm64/kvm/vgic/vgic-mmio.h
212
void vgic_write_irq_line_level_info(struct kvm_vcpu *vcpu, u32 intid,
arch/arm64/kvm/vgic/vgic-mmio.h
26
unsigned long (*uaccess_read)(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-mmio.h
29
int (*uaccess_write)(struct kvm_vcpu *vcpu, gpa_t addr,
arch/arm64/kvm/vgic/vgic-v2.c
104
static u32 vgic_v2_compute_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq);
arch/arm64/kvm/vgic/vgic-v2.c
113
void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v2.c
151
void vgic_v2_deactivate(struct kvm_vcpu *vcpu, u32 val)
arch/arm64/kvm/vgic/vgic-v2.c
155
struct kvm_vcpu *target_vcpu = NULL;
arch/arm64/kvm/vgic/vgic-v2.c
224
static u32 vgic_v2_compute_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq)
arch/arm64/kvm/vgic/vgic-v2.c
30
void vgic_v2_configure_hcr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-v2.c
300
void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr)
arch/arm64/kvm/vgic/vgic-v2.c
334
void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr)
arch/arm64/kvm/vgic/vgic-v2.c
339
void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
arch/arm64/kvm/vgic/vgic-v2.c
366
void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
arch/arm64/kvm/vgic/vgic-v2.c
394
void vgic_v2_reset(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v2.c
546
static void save_lrs(struct kvm_vcpu *vcpu, void __iomem *base)
arch/arm64/kvm/vgic/vgic-v2.c
56
static void vgic_v2_fold_lr(struct kvm_vcpu *vcpu, u32 val)
arch/arm64/kvm/vgic/vgic-v2.c
567
void vgic_v2_save_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v2.c
591
void vgic_v2_restore_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v2.c
607
void vgic_v2_load(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v2.c
617
void vgic_v2_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
118
bool vgic_state_is_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
143
static void vgic_compute_mi_state(struct kvm_vcpu *vcpu, struct mi_state *mi_state)
arch/arm64/kvm/vgic/vgic-v3-nested.c
163
u16 vgic_v3_get_eisr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
171
u16 vgic_v3_get_elrsr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
179
u64 vgic_v3_get_misr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
220
static u64 translate_lr_pintid(struct kvm_vcpu *vcpu, u64 lr)
arch/arm64/kvm/vgic/vgic-v3-nested.c
248
static void vgic_v3_create_shadow_lr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-v3-nested.c
271
void vgic_v3_flush_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
278
void vgic_v3_sync_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
323
static void vgic_v3_create_shadow_state(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-v3-nested.c
341
void vgic_v3_load_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
363
void vgic_v3_put_nested(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
389
void vgic_v3_handle_nested_maint_irq(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3-nested.c
400
void vgic_v3_nested_update_mi(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3.c
136
static u64 vgic_v3_compute_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq);
arch/arm64/kvm/vgic/vgic-v3.c
146
void vgic_v3_fold_lr_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3.c
195
void vgic_v3_deactivate(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/vgic/vgic-v3.c
200
struct kvm_vcpu *target_vcpu = NULL;
arch/arm64/kvm/vgic/vgic-v3.c
24
void vgic_v3_configure_hcr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic-v3.c
298
static u64 vgic_v3_compute_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq)
arch/arm64/kvm/vgic/vgic-v3.c
366
void vgic_v3_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr)
arch/arm64/kvm/vgic/vgic-v3.c
399
void vgic_v3_clear_lr(struct kvm_vcpu *vcpu, int lr)
arch/arm64/kvm/vgic/vgic-v3.c
404
void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
arch/arm64/kvm/vgic/vgic-v3.c
432
void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcrp)
arch/arm64/kvm/vgic/vgic-v3.c
466
void vgic_v3_reset(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3.c
498
void vcpu_set_ich_hcr(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3.c
514
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-v3.c
609
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-v3.c
74
static void vgic_v3_fold_lr(struct kvm_vcpu *vcpu, u64 val)
arch/arm64/kvm/vgic/vgic-v3.c
749
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-v3.c
976
void vgic_v3_load(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v3.c
995
void vgic_v3_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
115
static void vgic_v4_enable_vsgis(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
158
static void vgic_v4_disable_vsgis(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
193
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-v4.c
229
int vgic_v4_request_vpe_irq(struct kvm_vcpu *vcpu, int irq)
arch/arm64/kvm/vgic/vgic-v4.c
245
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic-v4.c
329
struct kvm_vcpu *vcpu = kvm_get_vcpu(kvm, i);
arch/arm64/kvm/vgic/vgic-v4.c
342
static inline bool vgic_v4_want_doorbell(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
358
int vgic_v4_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
368
int vgic_v4_load(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
404
void vgic_v4_commit(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic-v4.c
86
struct kvm_vcpu *vcpu = info;
arch/arm64/kvm/vgic/vgic.c
1003
static inline void vgic_save_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1012
void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
103
struct vgic_irq *vgic_get_vcpu_irq(struct kvm_vcpu *vcpu, u32 intid)
arch/arm64/kvm/vgic/vgic.c
1031
void kvm_vgic_process_async_update(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1041
static inline void vgic_restore_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1050
void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1095
void kvm_vgic_load(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1109
void kvm_vgic_put(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1123
int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
1159
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic.c
1174
bool kvm_vgic_map_is_active(struct kvm_vcpu *vcpu, unsigned int vintid)
arch/arm64/kvm/vgic/vgic.c
179
void vgic_flush_pending_lpis(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
247
struct kvm_vcpu *vgic_target_oracle(struct vgic_irq *irq)
arch/arm64/kvm/vgic/vgic.c
276
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic.c
299
struct kvm_vcpu *vcpu = info->vcpu;
arch/arm64/kvm/vgic/vgic.c
346
static void vgic_sort_ap_list(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
402
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic.c
513
int kvm_vgic_inject_irq(struct kvm *kvm, struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic.c
557
static int kvm_vgic_map_irq(struct kvm_vcpu *vcpu, struct vgic_irq *irq,
arch/arm64/kvm/vgic/vgic.c
591
int kvm_vgic_map_phys_irq(struct kvm_vcpu *vcpu, unsigned int host_irq,
arch/arm64/kvm/vgic/vgic.c
617
void kvm_vgic_reset_mapped_irq(struct kvm_vcpu *vcpu, u32 vintid)
arch/arm64/kvm/vgic/vgic.c
634
int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu, unsigned int vintid)
arch/arm64/kvm/vgic/vgic.c
653
int kvm_vgic_get_map(struct kvm_vcpu *vcpu, unsigned int vintid)
arch/arm64/kvm/vgic/vgic.c
678
int kvm_vgic_set_owner(struct kvm_vcpu *vcpu, unsigned int intid, void *owner)
arch/arm64/kvm/vgic/vgic.c
710
static void vgic_prune_ap_list(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
722
struct kvm_vcpu *target_vcpu, *vcpuA, *vcpuB;
arch/arm64/kvm/vgic/vgic.c
815
static inline void vgic_fold_lr_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.c
827
static inline void vgic_populate_lr(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic.c
838
static inline void vgic_clear_lr(struct kvm_vcpu *vcpu, int lr)
arch/arm64/kvm/vgic/vgic.c
846
static void summarize_ap_list(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic.c
952
static void vgic_flush_lr_state(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.h
121
static inline u32 vgic_get_implementation_rev(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.h
205
struct kvm_vcpu *vcpu;
arch/arm64/kvm/vgic/vgic.h
259
vgic_get_mmio_region(struct kvm_vcpu *vcpu, struct vgic_io_device *iodev,
arch/arm64/kvm/vgic/vgic.h
262
struct vgic_irq *vgic_get_vcpu_irq(struct kvm_vcpu *vcpu, u32 intid);
arch/arm64/kvm/vgic/vgic.h
264
struct kvm_vcpu *vgic_target_oracle(struct vgic_irq *irq);
arch/arm64/kvm/vgic/vgic.h
278
void vgic_v2_fold_lr_state(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
279
void vgic_v2_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr);
arch/arm64/kvm/vgic/vgic.h
280
void vgic_v2_deactivate(struct kvm_vcpu *vcpu, u32 val);
arch/arm64/kvm/vgic/vgic.h
281
void vgic_v2_clear_lr(struct kvm_vcpu *vcpu, int lr);
arch/arm64/kvm/vgic/vgic.h
282
void vgic_v2_configure_hcr(struct kvm_vcpu *vcpu, struct ap_list_summary *als);
arch/arm64/kvm/vgic/vgic.h
284
int vgic_v2_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic.h
286
int vgic_v2_cpuif_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic.h
288
void vgic_v2_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
289
void vgic_v2_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
290
void vgic_v2_reset(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
297
void vgic_v2_load(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
298
void vgic_v2_put(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
300
void vgic_v2_save_state(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
301
void vgic_v2_restore_state(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
319
void vgic_v3_fold_lr_state(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
320
void vgic_v3_populate_lr(struct kvm_vcpu *vcpu, struct vgic_irq *irq, int lr);
arch/arm64/kvm/vgic/vgic.h
321
void vgic_v3_clear_lr(struct kvm_vcpu *vcpu, int lr);
arch/arm64/kvm/vgic/vgic.h
322
void vgic_v3_deactivate(struct kvm_vcpu *vcpu, u64 val);
arch/arm64/kvm/vgic/vgic.h
323
void vgic_v3_configure_hcr(struct kvm_vcpu *vcpu, struct ap_list_summary *als);
arch/arm64/kvm/vgic/vgic.h
324
void vgic_v3_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
325
void vgic_v3_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
326
void vgic_v3_reset(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
333
int vgic_register_redist_iodev(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
334
void vgic_unregister_redist_iodev(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
337
void vgic_v3_load(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
338
void vgic_v3_put(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
342
void vgic_enable_lpis(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
343
void vgic_flush_pending_lpis(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
346
int vgic_v3_dist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic.h
348
int vgic_v3_redist_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic.h
350
int vgic_v3_cpu_sysregs_uaccess(struct kvm_vcpu *vcpu,
arch/arm64/kvm/vgic/vgic.h
352
int vgic_v3_has_cpu_sysregs_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr);
arch/arm64/kvm/vgic/vgic.h
354
int vgic_v3_line_level_info_uaccess(struct kvm_vcpu *vcpu, bool is_write,
arch/arm64/kvm/vgic/vgic.h
357
void vgic_set_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
358
void vgic_get_vmcr(struct kvm_vcpu *vcpu, struct vgic_vmcr *vmcr);
arch/arm64/kvm/vgic/vgic.h
367
static inline int vgic_v3_max_apr_idx(struct kvm_vcpu *vcpu)
arch/arm64/kvm/vgic/vgic.h
417
bool vgic_lpis_enabled(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
426
int vgic_its_invall(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
441
int vgic_v4_request_vpe_irq(struct kvm_vcpu *vcpu, int irq);
arch/arm64/kvm/vgic/vgic.h
443
void vcpu_set_ich_hcr(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
450
void vgic_v3_flush_nested(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
451
void vgic_v3_sync_nested(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
452
void vgic_v3_load_nested(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
453
void vgic_v3_put_nested(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
454
void vgic_v3_handle_nested_maint_irq(struct kvm_vcpu *vcpu);
arch/arm64/kvm/vgic/vgic.h
455
void vgic_v3_nested_update_mi(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_csr.h
187
int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
110
struct kvm_vcpu *vcpu;
arch/loongarch/include/asm/kvm_host.h
182
int (*handle_exit)(struct kvm_run *run, struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
313
bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
316
int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
320
void kvm_flush_tlb_gpa(struct kvm_vcpu *vcpu, unsigned long gpa);
arch/loongarch/include/asm/kvm_host.h
321
int kvm_handle_mm_fault(struct kvm_vcpu *vcpu, unsigned long badv, bool write, int ecode);
arch/loongarch/include/asm/kvm_host.h
347
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
arch/loongarch/include/asm/kvm_host.h
348
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
arch/loongarch/include/asm/kvm_host.h
349
static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {}
arch/loongarch/include/asm/kvm_host.h
351
void kvm_check_vpid(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
356
int kvm_enter_guest(struct kvm_run *run, struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_host.h
80
struct kvm_vcpu *last_vcpu;
arch/loongarch/include/asm/kvm_host.h
88
int (*enter_guest)(struct kvm_run *run, struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
102
static inline void kvm_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int irq)
arch/loongarch/include/asm/kvm_vcpu.h
108
static inline int kvm_queue_exception(struct kvm_vcpu *vcpu,
arch/loongarch/include/asm/kvm_vcpu.h
120
static inline unsigned long kvm_read_reg(struct kvm_vcpu *vcpu, int num)
arch/loongarch/include/asm/kvm_vcpu.h
125
static inline void kvm_write_reg(struct kvm_vcpu *vcpu, int num, unsigned long val)
arch/loongarch/include/asm/kvm_vcpu.h
135
static inline bool kvm_guest_has_pv_feature(struct kvm_vcpu *vcpu, unsigned int feature)
arch/loongarch/include/asm/kvm_vcpu.h
41
typedef int (*exit_handle_fn)(struct kvm_vcpu *, int);
arch/loongarch/include/asm/kvm_vcpu.h
43
int kvm_emu_mmio_read(struct kvm_vcpu *vcpu, larch_inst inst);
arch/loongarch/include/asm/kvm_vcpu.h
44
int kvm_emu_mmio_write(struct kvm_vcpu *vcpu, larch_inst inst);
arch/loongarch/include/asm/kvm_vcpu.h
45
int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/loongarch/include/asm/kvm_vcpu.h
46
int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/loongarch/include/asm/kvm_vcpu.h
47
int kvm_complete_user_service(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/loongarch/include/asm/kvm_vcpu.h
48
int kvm_emu_idle(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
49
int kvm_pending_timer(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
50
int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault);
arch/loongarch/include/asm/kvm_vcpu.h
51
void kvm_deliver_intr(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
52
void kvm_deliver_exception(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
54
void kvm_own_fpu(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
55
void kvm_lose_fpu(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
61
int kvm_own_lsx(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
65
static inline int kvm_own_lsx(struct kvm_vcpu *vcpu) { return -EINVAL; }
arch/loongarch/include/asm/kvm_vcpu.h
71
int kvm_own_lasx(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
75
static inline int kvm_own_lasx(struct kvm_vcpu *vcpu) { return -EINVAL; }
arch/loongarch/include/asm/kvm_vcpu.h
81
int kvm_own_lbt(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
83
static inline int kvm_own_lbt(struct kvm_vcpu *vcpu) { return -EINVAL; }
arch/loongarch/include/asm/kvm_vcpu.h
86
void kvm_init_timer(struct kvm_vcpu *vcpu, unsigned long hz);
arch/loongarch/include/asm/kvm_vcpu.h
87
void kvm_save_timer(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
88
void kvm_restore_timer(struct kvm_vcpu *vcpu);
arch/loongarch/include/asm/kvm_vcpu.h
90
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq);
arch/loongarch/include/asm/kvm_vcpu.h
91
struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid);
arch/loongarch/include/asm/kvm_vcpu.h
96
static inline void kvm_queue_irq(struct kvm_vcpu *vcpu, unsigned int irq)
arch/loongarch/kernel/asm-offsets.c
295
OFFSET(KVM_VCPU_ARCH, kvm_vcpu, arch);
arch/loongarch/kernel/asm-offsets.c
296
OFFSET(KVM_VCPU_KVM, kvm_vcpu, kvm);
arch/loongarch/kernel/asm-offsets.c
297
OFFSET(KVM_VCPU_RUN, kvm_vcpu, run);
arch/loongarch/kvm/exit.c
113
static int kvm_handle_csr(struct kvm_vcpu *vcpu, larch_inst inst)
arch/loongarch/kvm/exit.c
157
int kvm_emu_iocsr(larch_inst inst, struct kvm_run *run, struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
237
int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/loongarch/kvm/exit.c
24
static int kvm_emu_cpucfg(struct kvm_vcpu *vcpu, larch_inst inst)
arch/loongarch/kvm/exit.c
265
int kvm_emu_idle(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
276
static int kvm_trap_handle_gspr(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
346
static int kvm_handle_gspr(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
369
int kvm_emu_mmio_read(struct kvm_vcpu *vcpu, larch_inst inst)
arch/loongarch/kvm/exit.c
502
int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/loongarch/kvm/exit.c
544
int kvm_emu_mmio_write(struct kvm_vcpu *vcpu, larch_inst inst)
arch/loongarch/kvm/exit.c
65
static unsigned long kvm_emu_read_csr(struct kvm_vcpu *vcpu, int csrid)
arch/loongarch/kvm/exit.c
667
static int kvm_handle_rdwr_fault(struct kvm_vcpu *vcpu, bool write, int ecode)
arch/loongarch/kvm/exit.c
711
static int kvm_handle_read_fault(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
716
static int kvm_handle_write_fault(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
721
int kvm_complete_user_service(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/loongarch/kvm/exit.c
737
static int kvm_handle_fpu_disabled(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
763
static long kvm_save_notify(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
794
static int kvm_handle_lsx_disabled(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
814
static int kvm_handle_lasx_disabled(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
82
static unsigned long kvm_emu_write_csr(struct kvm_vcpu *vcpu, int csrid, unsigned long val)
arch/loongarch/kvm/exit.c
826
static int kvm_handle_lbt_disabled(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
838
static void kvm_send_pv_ipi(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
841
struct kvm_vcpu *dest;
arch/loongarch/kvm/exit.c
862
static void kvm_handle_service(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/exit.c
885
static int kvm_handle_hypercall(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
945
static int kvm_fault_ni(struct kvm_vcpu *vcpu, int ecode)
arch/loongarch/kvm/exit.c
96
static unsigned long kvm_emu_xchg_csr(struct kvm_vcpu *vcpu, int csrid,
arch/loongarch/kvm/exit.c
975
int kvm_handle_fault(struct kvm_vcpu *vcpu, int fault)
arch/loongarch/kvm/intc/eiointc.c
119
static int loongarch_eiointc_read(struct kvm_vcpu *vcpu, struct loongarch_eiointc *s,
arch/loongarch/kvm/intc/eiointc.c
13
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/eiointc.c
160
static int kvm_eiointc_read(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/eiointc.c
203
static int loongarch_eiointc_write(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/eiointc.c
292
static int kvm_eiointc_write(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/eiointc.c
339
static int kvm_eiointc_virt_read(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/eiointc.c
369
static int kvm_eiointc_virt_write(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/eiointc.c
38
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/eiointc.c
78
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/ipi.c
10
static void ipi_set(struct kvm_vcpu *vcpu, uint32_t data)
arch/loongarch/kvm/intc/ipi.c
108
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/ipi.c
132
static int send_ipi_data(struct kvm_vcpu *vcpu, gpa_t addr, uint64_t data)
arch/loongarch/kvm/intc/ipi.c
171
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/ipi.c
184
static int loongarch_ipi_readl(struct kvm_vcpu *vcpu, gpa_t addr, int len, void *val)
arch/loongarch/kvm/intc/ipi.c
223
static int loongarch_ipi_writel(struct kvm_vcpu *vcpu, gpa_t addr, int len, const void *val)
arch/loongarch/kvm/intc/ipi.c
273
static int kvm_ipi_read(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/ipi.c
28
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/ipi.c
281
static int kvm_ipi_write(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/ipi.c
302
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/intc/ipi.c
40
static void ipi_clear(struct kvm_vcpu *vcpu, uint64_t data)
arch/loongarch/kvm/intc/ipi.c
55
static uint64_t read_mailbox(struct kvm_vcpu *vcpu, int offset, int len)
arch/loongarch/kvm/intc/ipi.c
78
static void write_mailbox(struct kvm_vcpu *vcpu, int offset, uint64_t data, int len)
arch/loongarch/kvm/intc/pch_pic.c
135
static int kvm_pch_pic_read(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/intc/pch_pic.c
234
static int kvm_pch_pic_write(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/interrupt.c
104
int kvm_pending_timer(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/interrupt.c
113
static void _kvm_deliver_exception(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/interrupt.c
166
void kvm_deliver_exception(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/interrupt.c
27
static int kvm_irq_deliver(struct kvm_vcpu *vcpu, unsigned int priority)
arch/loongarch/kvm/interrupt.c
59
static int kvm_irq_clear(struct kvm_vcpu *vcpu, unsigned int priority)
arch/loongarch/kvm/interrupt.c
91
void kvm_deliver_intr(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/main.c
205
static void kvm_update_vpid(struct kvm_vcpu *vcpu, int cpu)
arch/loongarch/kvm/main.c
227
void kvm_check_vpid(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/main.c
453
return kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/loongarch/kvm/mmu.c
552
static int kvm_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, bool write)
arch/loongarch/kvm/mmu.c
728
static kvm_pte_t *kvm_split_huge(struct kvm_vcpu *vcpu, kvm_pte_t *ptep, gfn_t gfn)
arch/loongarch/kvm/mmu.c
772
static int kvm_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, bool write)
arch/loongarch/kvm/mmu.c
915
int kvm_handle_mm_fault(struct kvm_vcpu *vcpu, unsigned long gpa, bool write, int ecode)
arch/loongarch/kvm/timer.c
14
static inline u64 ktime_to_tick(struct kvm_vcpu *vcpu, ktime_t now)
arch/loongarch/kvm/timer.c
140
static void _kvm_save_timer(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/timer.c
178
void kvm_save_timer(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/timer.c
22
static inline u64 tick_to_ns(struct kvm_vcpu *vcpu, u64 tick)
arch/loongarch/kvm/timer.c
30
struct kvm_vcpu *vcpu;
arch/loongarch/kvm/timer.c
32
vcpu = container_of(timer, struct kvm_vcpu, arch.swtimer);
arch/loongarch/kvm/timer.c
42
void kvm_init_timer(struct kvm_vcpu *vcpu, unsigned long timer_hz)
arch/loongarch/kvm/timer.c
53
void kvm_restore_timer(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/tlb.c
24
void kvm_flush_tlb_gpa(struct kvm_vcpu *vcpu, unsigned long gpa)
arch/loongarch/kvm/trace.h
102
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int inst_word),
arch/loongarch/kvm/trace.h
141
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op,
arch/loongarch/kvm/trace.h
19
TP_PROTO(struct kvm_vcpu *vcpu),
arch/loongarch/kvm/trace.h
200
TP_PROTO(struct kvm_vcpu *vcpu, unsigned long vpid),
arch/loongarch/kvm/trace.h
35
TP_PROTO(struct kvm_vcpu *vcpu),
arch/loongarch/kvm/trace.h
39
TP_PROTO(struct kvm_vcpu *vcpu),
arch/loongarch/kvm/trace.h
43
TP_PROTO(struct kvm_vcpu *vcpu),
arch/loongarch/kvm/trace.h
60
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/trace.h
82
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/trace.h
86
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/trace.h
90
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/trace.h
94
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/trace.h
98
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/loongarch/kvm/vcpu.c
100
static int kvm_own_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1005
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/loongarch/kvm/vcpu.c
1010
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/loongarch/kvm/vcpu.c
1015
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/loongarch/kvm/vcpu.c
1027
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/loongarch/kvm/vcpu.c
1040
static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1047
static int kvm_loongarch_cpucfg_has_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1063
static int kvm_loongarch_pvtime_has_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1073
static int kvm_loongarch_vcpu_has_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1092
static int kvm_loongarch_cpucfg_get_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1117
static int kvm_loongarch_pvtime_get_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1134
static int kvm_loongarch_vcpu_get_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1153
static int kvm_loongarch_cpucfg_set_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1180
static int kvm_loongarch_pvtime_set_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
119
static void kvm_lose_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1217
static int kvm_loongarch_vcpu_set_attr(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
1242
struct kvm_vcpu *vcpu = filp->private_data;
arch/loongarch/kvm/vcpu.c
1309
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/loongarch/kvm/vcpu.c
1321
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/loongarch/kvm/vcpu.c
1334
int kvm_own_lbt(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1345
static void kvm_lose_lbt(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1356
static void kvm_check_fcsr(struct kvm_vcpu *vcpu, unsigned long fcsr)
arch/loongarch/kvm/vcpu.c
1366
static void kvm_check_fcsr_alive(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1375
static inline void kvm_lose_lbt(struct kvm_vcpu *vcpu) { }
arch/loongarch/kvm/vcpu.c
1376
static inline void kvm_check_fcsr(struct kvm_vcpu *vcpu, unsigned long fcsr) { }
arch/loongarch/kvm/vcpu.c
1377
static inline void kvm_check_fcsr_alive(struct kvm_vcpu *vcpu) { }
arch/loongarch/kvm/vcpu.c
1381
void kvm_own_fpu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1397
int kvm_own_lsx(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1427
int kvm_own_lasx(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1456
void kvm_lose_fpu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1488
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq)
arch/loongarch/kvm/vcpu.c
1510
struct kvm_vcpu *vcpu = filp->private_data;
arch/loongarch/kvm/vcpu.c
152
static void kvm_check_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1531
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1589
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1593
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
160
static void kvm_update_stolen_time(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1614
static int _kvm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/loongarch/kvm/vcpu.c
1715
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/loongarch/kvm/vcpu.c
1725
static int _kvm_vcpu_put(struct kvm_vcpu *vcpu, int cpu)
arch/loongarch/kvm/vcpu.c
1806
static void kvm_vcpu_set_pv_preempted(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1842
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
1866
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
215
static int kvm_check_requests(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
232
static void kvm_late_check_requests(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
272
static int kvm_enter_guest_check(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
296
static int kvm_pre_enter_guest(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
345
static int kvm_handle_exit(struct kvm_run *run, struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
386
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
392
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
397
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
409
unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
41
static inline void kvm_save_host_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
419
bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
425
bool kvm_arch_vcpu_preempted_in_kernel(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
430
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/loongarch/kvm/vcpu.c
435
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
441
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
454
int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
477
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
485
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
501
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
515
static inline int kvm_set_cpuid(struct kvm_vcpu *vcpu, u64 val)
arch/loongarch/kvm/vcpu.c
56
static inline void kvm_restore_host_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
566
static inline void kvm_drop_cpuid(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
587
struct kvm_vcpu *kvm_get_vcpu_by_cpuid(struct kvm *kvm, int cpuid)
arch/loongarch/kvm/vcpu.c
604
static int _kvm_getcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 *val)
arch/loongarch/kvm/vcpu.c
639
static int _kvm_setcsr(struct kvm_vcpu *vcpu, unsigned int id, u64 val)
arch/loongarch/kvm/vcpu.c
72
static inline void kvm_save_guest_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
816
static int kvm_get_one_reg(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
86
static inline void kvm_restore_guest_pmu(struct kvm_vcpu *vcpu)
arch/loongarch/kvm/vcpu.c
883
static int kvm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/loongarch/kvm/vcpu.c
903
static int kvm_set_one_reg(struct kvm_vcpu *vcpu,
arch/loongarch/kvm/vcpu.c
987
static int kvm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/mips/include/asm/kvm_host.h
292
int (*vcpu_run)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
717
int (*handle_cop_unusable)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
718
int (*handle_tlb_mod)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
719
int (*handle_tlb_ld_miss)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
720
int (*handle_tlb_st_miss)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
721
int (*handle_addr_err_st)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
722
int (*handle_addr_err_ld)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
723
int (*handle_syscall)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
724
int (*handle_res_inst)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
725
int (*handle_break)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
726
int (*handle_trap)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
727
int (*handle_msa_fpe)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
728
int (*handle_fpe)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
729
int (*handle_msa_disabled)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
730
int (*handle_guest_exit)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
734
int (*vcpu_init)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
735
void (*vcpu_uninit)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
736
int (*vcpu_setup)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
739
void (*queue_timer_int)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
740
void (*dequeue_timer_int)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
741
void (*queue_io_int)(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
743
void (*dequeue_io_int)(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
745
int (*irq_deliver)(struct kvm_vcpu *vcpu, unsigned int priority,
arch/mips/include/asm/kvm_host.h
747
int (*irq_clear)(struct kvm_vcpu *vcpu, unsigned int priority,
arch/mips/include/asm/kvm_host.h
749
unsigned long (*num_regs)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
750
int (*copy_reg_indices)(struct kvm_vcpu *vcpu, u64 __user *indices);
arch/mips/include/asm/kvm_host.h
751
int (*get_one_reg)(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
753
int (*set_one_reg)(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
755
int (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu);
arch/mips/include/asm/kvm_host.h
756
int (*vcpu_put)(struct kvm_vcpu *vcpu, int cpu);
arch/mips/include/asm/kvm_host.h
757
int (*vcpu_run)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
758
void (*vcpu_reenter)(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
764
int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
766
extern int kvm_mips_handle_exit(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
783
void kvm_own_fpu(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
784
void kvm_own_msa(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
785
void kvm_drop_fpu(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
786
void kvm_lose_fpu(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
790
struct kvm_vcpu *vcpu, bool write_fault);
arch/mips/include/asm/kvm_host.h
792
int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long entryhi);
arch/mips/include/asm/kvm_host.h
793
int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
arch/mips/include/asm/kvm_host.h
811
void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
814
enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause);
arch/mips/include/asm/kvm_host.h
815
int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
arch/mips/include/asm/kvm_host.h
816
int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out);
arch/mips/include/asm/kvm_host.h
845
extern enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
847
u32 kvm_mips_read_count(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
848
void kvm_mips_write_count(struct kvm_vcpu *vcpu, u32 count);
arch/mips/include/asm/kvm_host.h
849
void kvm_mips_write_compare(struct kvm_vcpu *vcpu, u32 compare, bool ack);
arch/mips/include/asm/kvm_host.h
850
void kvm_mips_init_count(struct kvm_vcpu *vcpu, unsigned long count_hz);
arch/mips/include/asm/kvm_host.h
851
int kvm_mips_set_count_ctl(struct kvm_vcpu *vcpu, s64 count_ctl);
arch/mips/include/asm/kvm_host.h
852
int kvm_mips_set_count_resume(struct kvm_vcpu *vcpu, s64 count_resume);
arch/mips/include/asm/kvm_host.h
853
int kvm_mips_set_count_hz(struct kvm_vcpu *vcpu, s64 count_hz);
arch/mips/include/asm/kvm_host.h
854
void kvm_mips_count_enable_cause(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
855
void kvm_mips_count_disable_cause(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
856
enum hrtimer_restart kvm_mips_count_timeout(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
859
int kvm_mips_count_disabled(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
860
ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count);
arch/mips/include/asm/kvm_host.h
861
int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before,
arch/mips/include/asm/kvm_host.h
864
void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
865
void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
869
struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
872
struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
875
enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
879
enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
881
int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
884
extern void kvm_mips_dump_stats(struct kvm_vcpu *vcpu);
arch/mips/include/asm/kvm_host.h
886
extern int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu,
arch/mips/include/asm/kvm_host.h
892
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
arch/mips/include/asm/kvm_host.h
893
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
arch/mips/kvm/emulate.c
1269
u32 cause, struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
1509
enum emulation_result kvm_mips_complete_mmio_load(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
243
enum emulation_result update_pc(struct kvm_vcpu *vcpu, u32 cause)
arch/mips/kvm/emulate.c
272
int kvm_get_badinstr(u32 *opc, struct kvm_vcpu *vcpu, u32 *out)
arch/mips/kvm/emulate.c
294
int kvm_get_badinstrp(u32 *opc, struct kvm_vcpu *vcpu, u32 *out)
arch/mips/kvm/emulate.c
313
int kvm_mips_count_disabled(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
328
static u32 kvm_mips_ktime_to_count(struct kvm_vcpu *vcpu, ktime_t now)
arch/mips/kvm/emulate.c
367
static inline ktime_t kvm_mips_count_time(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
385
static u32 kvm_mips_read_count_running(struct kvm_vcpu *vcpu, ktime_t now)
arch/mips/kvm/emulate.c
40
static int kvm_compute_return_epc(struct kvm_vcpu *vcpu, unsigned long instpc,
arch/mips/kvm/emulate.c
445
u32 kvm_mips_read_count(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
472
ktime_t kvm_mips_freeze_hrtimer(struct kvm_vcpu *vcpu, u32 *count)
arch/mips/kvm/emulate.c
502
static void kvm_mips_resume_hrtimer(struct kvm_vcpu *vcpu,
arch/mips/kvm/emulate.c
543
int kvm_mips_restore_hrtimer(struct kvm_vcpu *vcpu, ktime_t before,
arch/mips/kvm/emulate.c
604
void kvm_mips_write_count(struct kvm_vcpu *vcpu, u32 count)
arch/mips/kvm/emulate.c
629
void kvm_mips_init_count(struct kvm_vcpu *vcpu, unsigned long count_hz)
arch/mips/kvm/emulate.c
650
int kvm_mips_set_count_hz(struct kvm_vcpu *vcpu, s64 count_hz)
arch/mips/kvm/emulate.c
697
void kvm_mips_write_compare(struct kvm_vcpu *vcpu, u32 compare, bool ack)
arch/mips/kvm/emulate.c
780
static ktime_t kvm_mips_count_disable(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
807
void kvm_mips_count_disable_cause(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
827
void kvm_mips_count_enable_cause(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
853
int kvm_mips_set_count_ctl(struct kvm_vcpu *vcpu, s64 count_ctl)
arch/mips/kvm/emulate.c
915
int kvm_mips_set_count_resume(struct kvm_vcpu *vcpu, s64 count_resume)
arch/mips/kvm/emulate.c
937
enum hrtimer_restart kvm_mips_count_timeout(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
945
enum emulation_result kvm_mips_emul_wait(struct kvm_vcpu *vcpu)
arch/mips/kvm/emulate.c
970
struct kvm_vcpu *vcpu)
arch/mips/kvm/entry.c
184
UASM_i_ADDIU(&p, GPR_K1, GPR_A0, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/entry.c
264
UASM_i_LW(&p, GPR_S0, (int)offsetof(struct kvm_vcpu, kvm) -
arch/mips/kvm/entry.c
265
(int)offsetof(struct kvm_vcpu, arch), GPR_K1);
arch/mips/kvm/entry.c
400
UASM_i_SW(&p, GPR_K0, offsetof(struct kvm_vcpu, arch.gprs[GPR_K0]), GPR_K1);
arch/mips/kvm/entry.c
448
UASM_i_LW(&p, GPR_K0, offsetof(struct kvm_vcpu, arch.gprs[GPR_K0]), GPR_K1);
arch/mips/kvm/entry.c
484
UASM_i_ADDIU(&p, GPR_K1, GPR_K1, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/entry.c
759
UASM_i_ADDIU(&p, GPR_K1, GPR_K1, offsetof(struct kvm_vcpu, arch));
arch/mips/kvm/hypcall.c
17
enum emulation_result kvm_mips_emul_hypcall(struct kvm_vcpu *vcpu,
arch/mips/kvm/hypcall.c
32
static int kvm_mips_hypercall(struct kvm_vcpu *vcpu, unsigned long num,
arch/mips/kvm/hypcall.c
40
int kvm_mips_handle_hypcall(struct kvm_vcpu *vcpu)
arch/mips/kvm/interrupt.c
24
void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause)
arch/mips/kvm/interrupt.c
37
int kvm_mips_pending_timer(struct kvm_vcpu *vcpu)
arch/mips/kvm/interrupt.h
37
int kvm_mips_pending_timer(struct kvm_vcpu *vcpu);
arch/mips/kvm/interrupt.h
39
void kvm_mips_deliver_interrupts(struct kvm_vcpu *vcpu, u32 cause);
arch/mips/kvm/loongson_ipi.c
153
static int kvm_ipi_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/mips/kvm/loongson_ipi.c
170
static int kvm_ipi_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/mips/kvm/mips.c
1005
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
1011
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1015
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/mips/kvm/mips.c
1020
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/mips/kvm/mips.c
1025
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/mips/kvm/mips.c
1076
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1082
int kvm_arch_vcpu_dump_regs(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1113
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/mips/kvm/mips.c
113
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1130
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/mips/kvm/mips.c
1147
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
1167
static int __kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
118
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
123
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1356
int noinstr kvm_mips_handle_exit(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1368
void kvm_own_fpu(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1412
void kvm_own_msa(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1474
void kvm_drop_fpu(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1491
void kvm_lose_fpu(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
1624
ret = kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/mips/kvm/mips.c
261
struct kvm_vcpu *vcpu;
arch/mips/kvm/mips.c
263
vcpu = container_of(timer, struct kvm_vcpu, arch.comparecount_timer);
arch/mips/kvm/mips.c
278
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
386
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
398
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
411
static int noinstr kvm_mips_vcpu_enter_exit(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
422
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
479
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
483
struct kvm_vcpu *dvcpu = NULL;
arch/mips/kvm/mips.c
515
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
521
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
578
static unsigned long kvm_mips_num_regs(struct kvm_vcpu *vcpu)
arch/mips/kvm/mips.c
596
static int kvm_mips_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices)
arch/mips/kvm/mips.c
646
static int kvm_mips_get_reg(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
758
static int kvm_mips_set_reg(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
871
static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
arch/mips/kvm/mips.c
901
struct kvm_vcpu *vcpu = filp->private_data;
arch/mips/kvm/mips.c
921
struct kvm_vcpu *vcpu = filp->private_data;
arch/mips/kvm/mips.c
999
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/mips/kvm/mmu.c
28
void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu)
arch/mips/kvm/mmu.c
480
static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa,
arch/mips/kvm/mmu.c
547
static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa,
arch/mips/kvm/mmu.c
644
struct kvm_vcpu *vcpu,
arch/mips/kvm/mmu.c
668
static void kvm_mips_migrate_count(struct kvm_vcpu *vcpu)
arch/mips/kvm/mmu.c
675
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/mips/kvm/mmu.c
702
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/mips/kvm/stats.c
49
void kvm_mips_dump_stats(struct kvm_vcpu *vcpu)
arch/mips/kvm/tlb.c
113
int kvm_vz_host_tlb_inv(struct kvm_vcpu *vcpu, unsigned long va)
arch/mips/kvm/tlb.c
166
int kvm_vz_guest_tlb_lookup(struct kvm_vcpu *vcpu, unsigned long gva,
arch/mips/kvm/tlb.c
40
static u32 kvm_mips_get_root_asid(struct kvm_vcpu *vcpu)
arch/mips/kvm/trace.h
116
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int reason),
arch/mips/kvm/trace.h
203
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op, unsigned int reg,
arch/mips/kvm/trace.h
253
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op,
arch/mips/kvm/trace.h
277
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int old_asid,
arch/mips/kvm/trace.h
299
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int guestid),
arch/mips/kvm/trace.h
31
TP_PROTO(struct kvm_vcpu *vcpu),
arch/mips/kvm/trace.h
314
TP_PROTO(struct kvm_vcpu *vcpu),
arch/mips/kvm/trace.h
46
TP_PROTO(struct kvm_vcpu *vcpu),
arch/mips/kvm/trace.h
50
TP_PROTO(struct kvm_vcpu *vcpu),
arch/mips/kvm/trace.h
54
TP_PROTO(struct kvm_vcpu *vcpu),
arch/mips/kvm/vz.c
104
static inline unsigned int kvm_vz_config4_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1076
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
110
static inline unsigned int kvm_vz_config5_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1145
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1213
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1295
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
132
static inline unsigned int kvm_vz_config6_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1421
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1433
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1465
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
148
static inline unsigned int kvm_vz_config_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1482
static int kvm_trap_vz_handle_guest_exit(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
153
static inline unsigned int kvm_vz_config1_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1553
static int kvm_trap_vz_handle_cop_unusable(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1602
static int kvm_trap_vz_handle_msa_disabled(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1623
static int kvm_trap_vz_handle_tlb_ld_miss(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
164
static inline unsigned int kvm_vz_config2_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1670
static int kvm_trap_vz_handle_tlb_st_miss(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
169
static inline unsigned int kvm_vz_config3_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1782
static unsigned long kvm_vz_num_regs(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1806
static int kvm_vz_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices)
arch/mips/kvm/vz.c
181
static inline unsigned int kvm_vz_config4_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
186
static inline unsigned int kvm_vz_config5_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
191
static inline unsigned int kvm_vz_config6_user_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
1910
static int kvm_vz_get_one_reg(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
203
static void kvm_vz_queue_irq(struct kvm_vcpu *vcpu, unsigned int priority)
arch/mips/kvm/vz.c
209
static void kvm_vz_dequeue_irq(struct kvm_vcpu *vcpu, unsigned int priority)
arch/mips/kvm/vz.c
2134
static int kvm_vz_set_one_reg(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
215
static void kvm_vz_queue_timer_int_cb(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
224
static void kvm_vz_dequeue_timer_int_cb(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
233
static void kvm_vz_queue_io_int_cb(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
2412
static void kvm_vz_get_new_guestid(unsigned long cpu, struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
2434
static int kvm_vz_check_requests(struct kvm_vcpu *vcpu, int cpu)
arch/mips/kvm/vz.c
245
static void kvm_vz_dequeue_io_int_cb(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
2461
static void kvm_vz_vcpu_save_wired(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
2494
static void kvm_vz_vcpu_load_wired(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
2502
static void kvm_vz_vcpu_load_tlb(struct kvm_vcpu *vcpu, int cpu)
arch/mips/kvm/vz.c
2563
static int kvm_vz_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/mips/kvm/vz.c
257
static int kvm_vz_irq_deliver_cb(struct kvm_vcpu *vcpu, unsigned int priority,
arch/mips/kvm/vz.c
2705
static int kvm_vz_vcpu_put(struct kvm_vcpu *vcpu, int cpu)
arch/mips/kvm/vz.c
286
static int kvm_vz_irq_clear_cb(struct kvm_vcpu *vcpu, unsigned int priority,
arch/mips/kvm/vz.c
3050
static int kvm_vz_vcpu_init(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
3060
static void kvm_vz_vcpu_uninit(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
3077
static int kvm_vz_vcpu_setup(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
3234
static void kvm_vz_vcpu_reenter(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
3250
static int kvm_vz_vcpu_run(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
340
static bool kvm_vz_should_use_htimer(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
365
static void _kvm_vz_restore_stimer(struct kvm_vcpu *vcpu, u32 compare,
arch/mips/kvm/vz.c
387
static void _kvm_vz_restore_htimer(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
39
static struct kvm_vcpu *last_vcpu[NR_CPUS];
arch/mips/kvm/vz.c
41
static struct kvm_vcpu *last_exec_vcpu[NR_CPUS];
arch/mips/kvm/vz.c
423
static void kvm_vz_restore_timer(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
444
void kvm_vz_acquire_htimer(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
468
static void _kvm_vz_save_htimer(struct kvm_vcpu *vcpu,
arch/mips/kvm/vz.c
518
static void kvm_vz_save_timer(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
547
void kvm_vz_lose_htimer(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
617
static bool is_eva_am_mapped(struct kvm_vcpu *vcpu, unsigned int am, bool eu)
arch/mips/kvm/vz.c
689
static int kvm_vz_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva,
arch/mips/kvm/vz.c
796
static int kvm_vz_badvaddr_to_gpa(struct kvm_vcpu *vcpu, unsigned long badvaddr,
arch/mips/kvm/vz.c
817
static int kvm_trap_vz_no_handler(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
84
static inline unsigned int kvm_vz_config_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
864
static void kvm_write_maari(struct kvm_vcpu *vcpu, unsigned long val)
arch/mips/kvm/vz.c
877
struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
89
static inline unsigned int kvm_vz_config1_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
94
static inline unsigned int kvm_vz_config2_guest_wrmask(struct kvm_vcpu *vcpu)
arch/mips/kvm/vz.c
99
static inline unsigned int kvm_vz_config3_guest_wrmask(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/asm-prototypes.h
55
struct kvm_vcpu;
arch/powerpc/include/asm/asm-prototypes.h
56
void _kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr);
arch/powerpc/include/asm/asm-prototypes.h
57
void _kvmppc_save_tm_pr(struct kvm_vcpu *vcpu, u64 guest_msr);
arch/powerpc/include/asm/asm-prototypes.h
60
void kvmppc_save_tm_hv(struct kvm_vcpu *vcpu, u64 msr, bool preserve_nv);
arch/powerpc/include/asm/asm-prototypes.h
61
void kvmppc_restore_tm_hv(struct kvm_vcpu *vcpu, u64 msr, bool preserve_nv);
arch/powerpc/include/asm/asm-prototypes.h
63
static inline void kvmppc_save_tm_hv(struct kvm_vcpu *vcpu, u64 msr,
arch/powerpc/include/asm/asm-prototypes.h
65
static inline void kvmppc_restore_tm_hv(struct kvm_vcpu *vcpu, u64 msr,
arch/powerpc/include/asm/asm-prototypes.h
69
void kvmppc_p9_enter_guest(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/asm-prototypes.h
71
long kvmppc_h_set_dabr(struct kvm_vcpu *vcpu, unsigned long dabr);
arch/powerpc/include/asm/asm-prototypes.h
72
long kvmppc_h_set_xdabr(struct kvm_vcpu *vcpu, unsigned long dabr,
arch/powerpc/include/asm/kvm_book3s.h
146
extern void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong ea, ulong ea_mask);
arch/powerpc/include/asm/kvm_book3s.h
147
extern void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 vp, u64 vp_mask);
arch/powerpc/include/asm/kvm_book3s.h
148
extern void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end);
arch/powerpc/include/asm/kvm_book3s.h
149
extern void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 new_msr);
arch/powerpc/include/asm/kvm_book3s.h
150
extern void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
151
extern void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
152
extern void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
153
extern int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte,
arch/powerpc/include/asm/kvm_book3s.h
155
extern void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte);
arch/powerpc/include/asm/kvm_book3s.h
156
extern int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr);
arch/powerpc/include/asm/kvm_book3s.h
157
extern void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong eaddr, ulong seg_size);
arch/powerpc/include/asm/kvm_book3s.h
158
extern void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
159
extern int kvmppc_book3s_hv_page_fault(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
163
extern int kvmppc_hv_emulate_mmio(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
166
extern void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
arch/powerpc/include/asm/kvm_book3s.h
167
extern struct hpte_cache *kvmppc_mmu_hpte_cache_next(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
169
extern void kvmppc_mmu_hpte_destroy(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
170
extern int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
171
extern void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte);
arch/powerpc/include/asm/kvm_book3s.h
177
extern int kvmppc_book3s_radix_page_fault(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
182
extern long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
184
extern long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
186
extern int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
189
extern int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
192
extern int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_book3s.h
203
extern int kvmppc_book3s_instantiate_page(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
227
extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr, bool data);
arch/powerpc/include/asm/kvm_book3s.h
228
extern void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec);
arch/powerpc/include/asm/kvm_book3s.h
229
extern void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
231
extern void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags);
arch/powerpc/include/asm/kvm_book3s.h
232
extern void kvmppc_trigger_fac_interrupt(struct kvm_vcpu *vcpu, ulong fac);
arch/powerpc/include/asm/kvm_book3s.h
233
extern void kvmppc_set_bat(struct kvm_vcpu *vcpu, struct kvmppc_bat *bat,
arch/powerpc/include/asm/kvm_book3s.h
235
extern void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr);
arch/powerpc/include/asm/kvm_book3s.h
236
extern int kvmppc_emulate_paired_single(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
237
extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/powerpc/include/asm/kvm_book3s.h
266
extern void kvmppc_set_fscr(struct kvm_vcpu *vcpu, u64 fscr);
arch/powerpc/include/asm/kvm_book3s.h
268
extern int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
269
extern int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
270
extern void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
274
extern u32 kvmppc_alignment_dsisr(struct kvm_vcpu *vcpu, unsigned int inst);
arch/powerpc/include/asm/kvm_book3s.h
275
extern ulong kvmppc_alignment_dar(struct kvm_vcpu *vcpu, unsigned int inst);
arch/powerpc/include/asm/kvm_book3s.h
276
extern int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd);
arch/powerpc/include/asm/kvm_book3s.h
280
extern void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
281
extern void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
284
void kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 msr);
arch/powerpc/include/asm/kvm_book3s.h
285
void kvmppc_inject_interrupt_hv(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags);
arch/powerpc/include/asm/kvm_book3s.h
288
void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
289
void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
290
void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
291
void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
293
static inline void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_book3s.h
294
static inline void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_book3s.h
295
static inline void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_book3s.h
296
static inline void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_book3s.h
303
long kvmhv_set_partition_table(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
304
long kvmhv_copy_tofrom_guest_nested(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
308
long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
309
long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
310
long do_h_rpt_invalidate_pat(struct kvm_vcpu *vcpu, unsigned long lpid,
arch/powerpc/include/asm/kvm_book3s.h
313
int kvmhv_run_single_vcpu(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
315
void kvmhv_save_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr);
arch/powerpc/include/asm/kvm_book3s.h
316
void kvmhv_restore_hv_return_state(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
318
long int kvmhv_nested_page_fault(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
320
void kvmppc_giveup_fac(struct kvm_vcpu *vcpu, ulong fac);
arch/powerpc/include/asm/kvm_book3s.h
351
int __kvmhv_nestedv2_reload_ptregs(struct kvm_vcpu *vcpu, struct pt_regs *regs);
arch/powerpc/include/asm/kvm_book3s.h
352
int __kvmhv_nestedv2_mark_dirty_ptregs(struct kvm_vcpu *vcpu, struct pt_regs *regs);
arch/powerpc/include/asm/kvm_book3s.h
353
int __kvmhv_nestedv2_mark_dirty(struct kvm_vcpu *vcpu, u16 iden);
arch/powerpc/include/asm/kvm_book3s.h
354
int __kvmhv_nestedv2_cached_reload(struct kvm_vcpu *vcpu, u16 iden);
arch/powerpc/include/asm/kvm_book3s.h
356
static inline int kvmhv_nestedv2_reload_ptregs(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
363
static inline int kvmhv_nestedv2_mark_dirty_ptregs(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_book3s.h
371
static inline int kvmhv_nestedv2_mark_dirty(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/include/asm/kvm_book3s.h
378
static inline int kvmhv_nestedv2_cached_reload(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/include/asm/kvm_book3s.h
387
static inline struct kvmppc_vcpu_book3s *to_book3s(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
401
static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val)
arch/powerpc/include/asm/kvm_book3s.h
407
static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num)
arch/powerpc/include/asm/kvm_book3s.h
413
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
arch/powerpc/include/asm/kvm_book3s.h
419
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
425
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_book3s.h
431
static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
437
static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_book3s.h
443
static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
449
static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_book3s.h
455
static inline ulong kvmppc_get_lr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
461
static inline void kvmppc_set_pc(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_book3s.h
467
static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
473
static inline u64 kvmppc_get_msr(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
474
static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
479
static inline ulong kvmppc_get_fault_dar(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
484
static inline u64 kvmppc_get_fpr(struct kvm_vcpu *vcpu, int i)
arch/powerpc/include/asm/kvm_book3s.h
490
static inline void kvmppc_set_fpr(struct kvm_vcpu *vcpu, int i, u64 val)
arch/powerpc/include/asm/kvm_book3s.h
496
static inline u64 kvmppc_get_fpscr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
502
static inline void kvmppc_set_fpscr(struct kvm_vcpu *vcpu, u64 val)
arch/powerpc/include/asm/kvm_book3s.h
509
static inline u64 kvmppc_get_vsx_fpr(struct kvm_vcpu *vcpu, int i, int j)
arch/powerpc/include/asm/kvm_book3s.h
515
static inline void kvmppc_set_vsx_fpr(struct kvm_vcpu *vcpu, int i, int j,
arch/powerpc/include/asm/kvm_book3s.h
523
static inline void kvmppc_get_vsx_vr(struct kvm_vcpu *vcpu, int i, vector128 *v)
arch/powerpc/include/asm/kvm_book3s.h
529
static inline void kvmppc_set_vsx_vr(struct kvm_vcpu *vcpu, int i,
arch/powerpc/include/asm/kvm_book3s.h
536
static inline u32 kvmppc_get_vscr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
542
static inline void kvmppc_set_vscr(struct kvm_vcpu *vcpu, u32 val)
arch/powerpc/include/asm/kvm_book3s.h
550
static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, u##size val) \
arch/powerpc/include/asm/kvm_book3s.h
558
static inline u##size kvmppc_get_##reg(struct kvm_vcpu *vcpu) \
arch/powerpc/include/asm/kvm_book3s.h
578
static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, u##size val) \
arch/powerpc/include/asm/kvm_book3s.h
585
static inline u##size kvmppc_get_##reg(struct kvm_vcpu *vcpu) \
arch/powerpc/include/asm/kvm_book3s.h
602
static inline u64 kvmppc_get_tb_offset(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
607
static inline u64 kvmppc_get_dec_expires(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
613
static inline void kvmppc_set_dec_expires(struct kvm_vcpu *vcpu, u64 val)
arch/powerpc/include/asm/kvm_book3s.h
620
static inline u64 kvmppc_dec_expires_host_tb(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
631
static inline bool kvmppc_supports_magic_page(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s.h
637
extern int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
638
extern int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s.h
79
struct kvm_vcpu *runnable_threads[MAX_SMT_THREADS];
arch/powerpc/include/asm/kvm_book3s.h
86
struct kvm_vcpu *runner;
arch/powerpc/include/asm/kvm_book3s_32.h
12
static inline struct kvmppc_book3s_shadow_vcpu *svcpu_get(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s_64.h
112
static inline struct kvmppc_book3s_shadow_vcpu *svcpu_get(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s_64.h
131
static inline bool kvmhv_vcpu_is_radix(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s_64.h
143
unsigned long kvmppc_msr_hard_disable_set_facilities(struct kvm_vcpu *vcpu, unsigned long msr);
arch/powerpc/include/asm/kvm_book3s_64.h
145
int kvmhv_vcpu_entry_p9(struct kvm_vcpu *vcpu, u64 time_limit, unsigned long lpcr, u64 *tb);
arch/powerpc/include/asm/kvm_book3s_64.h
590
static inline void copy_from_checkpoint(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s_64.h
607
static inline void copy_to_checkpoint(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_book3s_64.h
680
int kvmhv_nestedv2_vcpu_create(struct kvm_vcpu *vcpu, struct kvmhv_nestedv2_io *io);
arch/powerpc/include/asm/kvm_book3s_64.h
681
void kvmhv_nestedv2_vcpu_free(struct kvm_vcpu *vcpu, struct kvmhv_nestedv2_io *io);
arch/powerpc/include/asm/kvm_book3s_64.h
682
int kvmhv_nestedv2_flush_vcpu(struct kvm_vcpu *vcpu, u64 time_limit);
arch/powerpc/include/asm/kvm_book3s_64.h
684
int kvmhv_nestedv2_parse_output(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_book3s_64.h
685
int kvmhv_nestedv2_set_vpa(struct kvm_vcpu *vcpu, unsigned long vpa);
arch/powerpc/include/asm/kvm_book3s_asm.h
101
struct kvm_vcpu *kvm_vcpu;
arch/powerpc/include/asm/kvm_booke.h
103
static inline ulong kvmppc_get_fault_dar(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
109
static inline bool kvmppc_supports_magic_page(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
26
static inline void kvmppc_set_gpr(struct kvm_vcpu *vcpu, int num, ulong val)
arch/powerpc/include/asm/kvm_booke.h
31
static inline ulong kvmppc_get_gpr(struct kvm_vcpu *vcpu, int num)
arch/powerpc/include/asm/kvm_booke.h
36
static inline void kvmppc_set_cr(struct kvm_vcpu *vcpu, u32 val)
arch/powerpc/include/asm/kvm_booke.h
41
static inline u32 kvmppc_get_cr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
46
static inline void kvmppc_set_xer(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_booke.h
51
static inline ulong kvmppc_get_xer(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
56
static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
62
static inline void kvmppc_set_ctr(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_booke.h
67
static inline ulong kvmppc_get_ctr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
72
static inline void kvmppc_set_lr(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_booke.h
77
static inline ulong kvmppc_get_lr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
82
static inline void kvmppc_set_pc(struct kvm_vcpu *vcpu, ulong val)
arch/powerpc/include/asm/kvm_booke.h
87
static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_booke.h
92
static inline void kvmppc_set_fpr(struct kvm_vcpu *vcpu, int i, u64 val)
arch/powerpc/include/asm/kvm_booke.h
97
static inline u64 kvmppc_get_fpr(struct kvm_vcpu *vcpu, int i)
arch/powerpc/include/asm/kvm_host.h
391
void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs);
arch/powerpc/include/asm/kvm_host.h
392
u64 (*slbmfee)(struct kvm_vcpu *vcpu, u64 slb_nr);
arch/powerpc/include/asm/kvm_host.h
393
u64 (*slbmfev)(struct kvm_vcpu *vcpu, u64 slb_nr);
arch/powerpc/include/asm/kvm_host.h
394
int (*slbfee)(struct kvm_vcpu *vcpu, gva_t eaddr, ulong *ret_slb);
arch/powerpc/include/asm/kvm_host.h
395
void (*slbie)(struct kvm_vcpu *vcpu, u64 slb_nr);
arch/powerpc/include/asm/kvm_host.h
396
void (*slbia)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_host.h
398
void (*mtsrin)(struct kvm_vcpu *vcpu, u32 srnum, ulong value);
arch/powerpc/include/asm/kvm_host.h
399
u32 (*mfsrin)(struct kvm_vcpu *vcpu, u32 srnum);
arch/powerpc/include/asm/kvm_host.h
400
int (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/include/asm/kvm_host.h
402
void (*tlbie)(struct kvm_vcpu *vcpu, ulong addr, bool large);
arch/powerpc/include/asm/kvm_host.h
403
int (*esid_to_vsid)(struct kvm_vcpu *vcpu, ulong esid, u64 *vsid);
arch/powerpc/include/asm/kvm_host.h
404
u64 (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data);
arch/powerpc/include/asm/kvm_host.h
405
bool (*is_dcbz32)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_host.h
907
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_host.h
908
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
arch/powerpc/include/asm/kvm_ppc.h
100
extern void kvmppc_decrementer_func(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
101
extern int kvmppc_sanity_check(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
102
extern int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
103
extern void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
1030
static inline void kvmppc_set_msr_fast(struct kvm_vcpu *vcpu, u64 val)
arch/powerpc/include/asm/kvm_ppc.h
1045
static inline u32 kvmppc_get_sr(struct kvm_vcpu *vcpu, int nr)
arch/powerpc/include/asm/kvm_ppc.h
1053
static inline void kvmppc_set_sr(struct kvm_vcpu *vcpu, int nr, u32 val)
arch/powerpc/include/asm/kvm_ppc.h
107
extern void kvmppc_mmu_map(struct kvm_vcpu *vcpu, u64 gvaddr, gpa_t gpaddr,
arch/powerpc/include/asm/kvm_ppc.h
109
extern void kvmppc_mmu_switch_pid(struct kvm_vcpu *vcpu, u32 pid);
arch/powerpc/include/asm/kvm_ppc.h
1095
static inline ulong kvmppc_get_ea_indexed(struct kvm_vcpu *vcpu, int ra, int rb)
arch/powerpc/include/asm/kvm_ppc.h
110
extern int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
arch/powerpc/include/asm/kvm_ppc.h
111
extern int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr);
arch/powerpc/include/asm/kvm_ppc.h
112
extern gpa_t kvmppc_mmu_xlate(struct kvm_vcpu *vcpu, unsigned int gtlb_index,
arch/powerpc/include/asm/kvm_ppc.h
114
extern void kvmppc_mmu_dtlb_miss(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
115
extern void kvmppc_mmu_itlb_miss(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
116
extern int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr,
arch/powerpc/include/asm/kvm_ppc.h
120
extern int kvmppc_core_vcpu_create(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
121
extern void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
122
extern int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
123
extern int kvmppc_core_vcpu_translate(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
126
extern void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
arch/powerpc/include/asm/kvm_ppc.h
127
extern void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
129
extern int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
130
extern int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
132
extern void kvmppc_core_queue_machine_check(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
134
extern void kvmppc_core_queue_syscall(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
135
extern void kvmppc_core_queue_program(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
137
extern void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
139
extern void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
141
extern void kvmppc_core_queue_vsx_unavail(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
143
extern void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
144
extern void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
145
extern void kvmppc_core_queue_external(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
147
extern void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
148
extern void kvmppc_core_queue_dtlb_miss(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
151
extern void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
155
extern void kvmppc_core_queue_itlb_miss(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
156
extern void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
159
extern void kvmppc_core_flush_tlb(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
160
extern int kvmppc_core_check_requests(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
165
extern int kvmppc_kvm_pv(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
166
extern void kvmppc_map_magic(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
173
extern void kvmppc_map_vrma(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
175
extern int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
190
extern long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn,
arch/powerpc/include/asm/kvm_ppc.h
192
extern long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
195
extern long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
198
extern long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn,
arch/powerpc/include/asm/kvm_ppc.h
222
extern int kvmppc_prepare_to_enter(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
230
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq);
arch/powerpc/include/asm/kvm_ppc.h
233
extern int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
243
void kvmppc_core_dequeue_debug(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
244
void kvmppc_core_queue_debug(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
263
int (*get_sregs)(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
264
int (*set_sregs)(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
265
int (*get_one_reg)(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/include/asm/kvm_ppc.h
267
int (*set_one_reg)(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/include/asm/kvm_ppc.h
269
void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu);
arch/powerpc/include/asm/kvm_ppc.h
270
void (*vcpu_put)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
271
void (*inject_interrupt)(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags);
arch/powerpc/include/asm/kvm_ppc.h
272
void (*set_msr)(struct kvm_vcpu *vcpu, u64 msr);
arch/powerpc/include/asm/kvm_ppc.h
273
int (*vcpu_run)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
274
int (*vcpu_create)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
275
void (*vcpu_free)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
276
int (*check_requests)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
294
int (*emulate_op)(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
296
int (*emulate_mtspr)(struct kvm_vcpu *vcpu, int sprn, ulong spr_val);
arch/powerpc/include/asm/kvm_ppc.h
297
int (*emulate_mfspr)(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val);
arch/powerpc/include/asm/kvm_ppc.h
298
void (*fast_vcpu_kick)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
310
void (*giveup_ext)(struct kvm_vcpu *vcpu, ulong msr);
arch/powerpc/include/asm/kvm_ppc.h
312
int (*load_from_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
arch/powerpc/include/asm/kvm_ppc.h
314
int (*store_to_eaddr)(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
arch/powerpc/include/asm/kvm_ppc.h
321
int (*create_vcpu_debugfs)(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry);
arch/powerpc/include/asm/kvm_ppc.h
327
static inline int kvmppc_get_last_inst(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
429
int kvmppc_core_get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
430
int kvmppc_core_set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
432
int kvmppc_get_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
433
int kvmppc_set_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/include/asm/kvm_ppc.h
435
int kvm_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg);
arch/powerpc/include/asm/kvm_ppc.h
436
int kvm_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg);
arch/powerpc/include/asm/kvm_ppc.h
437
int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id, union kvmppc_one_reg *);
arch/powerpc/include/asm/kvm_ppc.h
438
int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id, union kvmppc_one_reg *);
arch/powerpc/include/asm/kvm_ppc.h
440
void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid);
arch/powerpc/include/asm/kvm_ppc.h
564
static inline void kvmppc_fast_vcpu_kick(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
598
static inline void kvmppc_fast_vcpu_kick(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
62
extern int kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
63
extern int __kvmppc_vcpu_run(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
632
static inline int kvmhv_nestedv2_reload_ptregs(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
637
static inline int kvmhv_nestedv2_mark_dirty_ptregs(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
643
static inline int kvmhv_nestedv2_mark_dirty(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/include/asm/kvm_ppc.h
648
static inline int kvmhv_nestedv2_cached_reload(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/include/asm/kvm_ppc.h
656
static inline int kvmppc_xics_enabled(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
66
extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
67
extern int kvmppc_handle_load(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
672
extern int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall);
arch/powerpc/include/asm/kvm_ppc.h
673
extern void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
674
extern int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd);
arch/powerpc/include/asm/kvm_ppc.h
675
extern int kvmppc_xive_xics_hcall(struct kvm_vcpu *vcpu, u32 req);
arch/powerpc/include/asm/kvm_ppc.h
676
extern u64 kvmppc_xics_get_icp(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
677
extern int kvmppc_xics_set_icp(struct kvm_vcpu *vcpu, u64 icpval);
arch/powerpc/include/asm/kvm_ppc.h
679
struct kvm_vcpu *vcpu, u32 cpu);
arch/powerpc/include/asm/kvm_ppc.h
685
extern long kvmppc_deliver_irq_passthru(struct kvm_vcpu *vcpu, __be32 xirr,
arch/powerpc/include/asm/kvm_ppc.h
70
extern int kvmppc_handle_loads(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
701
static inline int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall)
arch/powerpc/include/asm/kvm_ppc.h
703
static inline int kvmppc_xics_enabled(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
705
static inline void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu) { }
arch/powerpc/include/asm/kvm_ppc.h
706
static inline int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd)
arch/powerpc/include/asm/kvm_ppc.h
708
static inline int kvmppc_xive_xics_hcall(struct kvm_vcpu *vcpu, u32 req)
arch/powerpc/include/asm/kvm_ppc.h
729
struct kvm_vcpu *vcpu, u32 cpu);
arch/powerpc/include/asm/kvm_ppc.h
73
extern int kvmppc_handle_vsx_load(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
730
extern void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
735
extern u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
736
extern int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval);
arch/powerpc/include/asm/kvm_ppc.h
740
extern void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
741
extern void kvmppc_xive_pull_vcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
742
extern bool kvmppc_xive_rearm_escalation(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
744
static inline int kvmppc_xive_enabled(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
750
struct kvm_vcpu *vcpu, u32 cpu);
arch/powerpc/include/asm/kvm_ppc.h
751
extern void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
752
extern int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
754
extern int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
76
extern int kvmppc_handle_vmx_load(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
767
struct kvm_vcpu *vcpu, u32 cpu) { return -EBUSY; }
arch/powerpc/include/asm/kvm_ppc.h
768
static inline void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu) { }
arch/powerpc/include/asm/kvm_ppc.h
773
static inline u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu) { return 0; }
arch/powerpc/include/asm/kvm_ppc.h
774
static inline int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval) { return -ENOENT; }
arch/powerpc/include/asm/kvm_ppc.h
778
static inline void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu) { }
arch/powerpc/include/asm/kvm_ppc.h
779
static inline void kvmppc_xive_pull_vcpu(struct kvm_vcpu *vcpu) { }
arch/powerpc/include/asm/kvm_ppc.h
78
extern int kvmppc_handle_vmx_store(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
780
static inline bool kvmppc_xive_rearm_escalation(struct kvm_vcpu *vcpu) { return true; }
arch/powerpc/include/asm/kvm_ppc.h
782
static inline int kvmppc_xive_enabled(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
785
struct kvm_vcpu *vcpu, u32 cpu) { return -EBUSY; }
arch/powerpc/include/asm/kvm_ppc.h
786
static inline void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu) { }
arch/powerpc/include/asm/kvm_ppc.h
787
static inline int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
790
static inline int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
80
extern int kvmppc_handle_store(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
812
long kvmppc_rm_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn,
arch/powerpc/include/asm/kvm_ppc.h
814
long kvmppc_rm_h_put_tce_indirect(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
817
long kvmppc_rm_h_stuff_tce(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
820
long int kvmppc_rm_h_confer(struct kvm_vcpu *vcpu, int target,
arch/powerpc/include/asm/kvm_ppc.h
822
long kvmppc_rm_h_random(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
824
void kvmppc_realmode_machine_check(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
828
long kvmppc_p9_realmode_hmi_handler(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
829
long kvmppc_h_enter(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
83
extern int kvmppc_handle_vsx_store(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
831
long kvmppc_h_remove(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
833
long kvmppc_h_bulk_remove(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
834
long kvmppc_h_protect(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
836
long kvmppc_h_read(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
838
long kvmppc_h_clear_ref(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
840
long kvmppc_h_clear_mod(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
842
long kvmppc_rm_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/include/asm/kvm_ppc.h
844
long kvmppc_hpte_hv_fault(struct kvm_vcpu *vcpu, unsigned long addr,
arch/powerpc/include/asm/kvm_ppc.h
846
void kvmppc_guest_entry_inject_int(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
87
extern int kvmppc_load_last_inst(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
870
void (*vcpu_kick)(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
875
static inline unsigned long kvmppc_get_epr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
886
static inline void kvmppc_set_epr(struct kvm_vcpu *vcpu, u32 epr)
arch/powerpc/include/asm/kvm_ppc.h
897
void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
898
int kvmppc_mpic_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
900
void kvmppc_mpic_disconnect_vcpu(struct openpic *opp, struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
904
static inline void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
909
struct kvm_vcpu *vcpu, u32 cpu)
arch/powerpc/include/asm/kvm_ppc.h
91
extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
arch/powerpc/include/asm/kvm_ppc.h
915
struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
921
int kvm_vcpu_ioctl_config_tlb(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
923
int kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu,
arch/powerpc/include/asm/kvm_ppc.h
93
extern int kvmppc_st(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
arch/powerpc/include/asm/kvm_ppc.h
95
extern int kvmppc_emulate_instruction(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
952
static inline bool kvmppc_shared_big_endian(struct kvm_vcpu *vcpu)
arch/powerpc/include/asm/kvm_ppc.h
96
extern int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
966
static inline ulong kvmppc_get_##reg(struct kvm_vcpu *vcpu) \
arch/powerpc/include/asm/kvm_ppc.h
97
extern int kvmppc_emulate_mmio(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
972
static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, ulong val) \
arch/powerpc/include/asm/kvm_ppc.h
978
static inline u##size kvmppc_get_##reg(struct kvm_vcpu *vcpu) \
arch/powerpc/include/asm/kvm_ppc.h
98
extern void kvmppc_emulate_dec(struct kvm_vcpu *vcpu);
arch/powerpc/include/asm/kvm_ppc.h
989
static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, u##size val) \
arch/powerpc/include/asm/kvm_ppc.h
99
extern u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb);
arch/powerpc/include/asm/processor.h
230
struct kvm_vcpu *kvm_vcpu;
arch/powerpc/kernel/asm-offsets.c
155
OFFSET(THREAD_KVM_VCPU, thread_struct, kvm_vcpu);
arch/powerpc/kernel/asm-offsets.c
355
OFFSET(VCPU_HOST_STACK, kvm_vcpu, arch.host_stack);
arch/powerpc/kernel/asm-offsets.c
356
OFFSET(VCPU_HOST_PID, kvm_vcpu, arch.host_pid);
arch/powerpc/kernel/asm-offsets.c
357
OFFSET(VCPU_GUEST_PID, kvm_vcpu, arch.pid);
arch/powerpc/kernel/asm-offsets.c
358
OFFSET(VCPU_GPRS, kvm_vcpu, arch.regs.gpr);
arch/powerpc/kernel/asm-offsets.c
359
OFFSET(VCPU_VRSAVE, kvm_vcpu, arch.vrsave);
arch/powerpc/kernel/asm-offsets.c
360
OFFSET(VCPU_FPRS, kvm_vcpu, arch.fp.fpr);
arch/powerpc/kernel/asm-offsets.c
362
OFFSET(VCPU_VRS, kvm_vcpu, arch.vr.vr);
arch/powerpc/kernel/asm-offsets.c
364
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
arch/powerpc/kernel/asm-offsets.c
365
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
arch/powerpc/kernel/asm-offsets.c
366
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
arch/powerpc/kernel/asm-offsets.c
368
OFFSET(VCPU_TAR, kvm_vcpu, arch.tar);
arch/powerpc/kernel/asm-offsets.c
370
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
arch/powerpc/kernel/asm-offsets.c
371
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
arch/powerpc/kernel/asm-offsets.c
373
OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr);
arch/powerpc/kernel/asm-offsets.c
374
OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0);
arch/powerpc/kernel/asm-offsets.c
375
OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1);
arch/powerpc/kernel/asm-offsets.c
376
OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0);
arch/powerpc/kernel/asm-offsets.c
377
OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1);
arch/powerpc/kernel/asm-offsets.c
378
OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2);
arch/powerpc/kernel/asm-offsets.c
379
OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3);
arch/powerpc/kernel/asm-offsets.c
382
OFFSET(VCPU_TB_RMENTRY, kvm_vcpu, arch.rm_entry);
arch/powerpc/kernel/asm-offsets.c
383
OFFSET(VCPU_TB_RMINTR, kvm_vcpu, arch.rm_intr);
arch/powerpc/kernel/asm-offsets.c
384
OFFSET(VCPU_TB_RMEXIT, kvm_vcpu, arch.rm_exit);
arch/powerpc/kernel/asm-offsets.c
385
OFFSET(VCPU_TB_GUEST, kvm_vcpu, arch.guest_time);
arch/powerpc/kernel/asm-offsets.c
386
OFFSET(VCPU_TB_CEDE, kvm_vcpu, arch.cede_time);
arch/powerpc/kernel/asm-offsets.c
387
OFFSET(VCPU_CUR_ACTIVITY, kvm_vcpu, arch.cur_activity);
arch/powerpc/kernel/asm-offsets.c
388
OFFSET(VCPU_ACTIVITY_START, kvm_vcpu, arch.cur_tb_start);
arch/powerpc/kernel/asm-offsets.c
399
OFFSET(VCPU_SHADOW_PID, kvm_vcpu, arch.shadow_pid);
arch/powerpc/kernel/asm-offsets.c
400
OFFSET(VCPU_SHADOW_PID1, kvm_vcpu, arch.shadow_pid1);
arch/powerpc/kernel/asm-offsets.c
401
OFFSET(VCPU_SHARED, kvm_vcpu, arch.shared);
arch/powerpc/kernel/asm-offsets.c
403
OFFSET(VCPU_SHADOW_MSR, kvm_vcpu, arch.shadow_msr);
arch/powerpc/kernel/asm-offsets.c
405
OFFSET(VCPU_SHAREDBE, kvm_vcpu, arch.shared_big_endian);
arch/powerpc/kernel/asm-offsets.c
415
OFFSET(VCPU_KVM, kvm_vcpu, kvm);
arch/powerpc/kernel/asm-offsets.c
427
OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr);
arch/powerpc/kernel/asm-offsets.c
428
OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar);
arch/powerpc/kernel/asm-offsets.c
429
OFFSET(VCPU_VPA, kvm_vcpu, arch.vpa.pinned_addr);
arch/powerpc/kernel/asm-offsets.c
430
OFFSET(VCPU_VPA_DIRTY, kvm_vcpu, arch.vpa.dirty);
arch/powerpc/kernel/asm-offsets.c
431
OFFSET(VCPU_HEIR, kvm_vcpu, arch.emul_inst);
arch/powerpc/kernel/asm-offsets.c
432
OFFSET(VCPU_CPU, kvm_vcpu, cpu);
arch/powerpc/kernel/asm-offsets.c
433
OFFSET(VCPU_THREAD_CPU, kvm_vcpu, arch.thread_cpu);
arch/powerpc/kernel/asm-offsets.c
436
OFFSET(VCPU_PURR, kvm_vcpu, arch.purr);
arch/powerpc/kernel/asm-offsets.c
437
OFFSET(VCPU_SPURR, kvm_vcpu, arch.spurr);
arch/powerpc/kernel/asm-offsets.c
438
OFFSET(VCPU_IC, kvm_vcpu, arch.ic);
arch/powerpc/kernel/asm-offsets.c
439
OFFSET(VCPU_DSCR, kvm_vcpu, arch.dscr);
arch/powerpc/kernel/asm-offsets.c
440
OFFSET(VCPU_AMR, kvm_vcpu, arch.amr);
arch/powerpc/kernel/asm-offsets.c
441
OFFSET(VCPU_UAMOR, kvm_vcpu, arch.uamor);
arch/powerpc/kernel/asm-offsets.c
442
OFFSET(VCPU_IAMR, kvm_vcpu, arch.iamr);
arch/powerpc/kernel/asm-offsets.c
443
OFFSET(VCPU_CTRL, kvm_vcpu, arch.ctrl);
arch/powerpc/kernel/asm-offsets.c
444
OFFSET(VCPU_DABR, kvm_vcpu, arch.dabr);
arch/powerpc/kernel/asm-offsets.c
445
OFFSET(VCPU_DABRX, kvm_vcpu, arch.dabrx);
arch/powerpc/kernel/asm-offsets.c
446
OFFSET(VCPU_DAWR0, kvm_vcpu, arch.dawr0);
arch/powerpc/kernel/asm-offsets.c
447
OFFSET(VCPU_DAWRX0, kvm_vcpu, arch.dawrx0);
arch/powerpc/kernel/asm-offsets.c
448
OFFSET(VCPU_CIABR, kvm_vcpu, arch.ciabr);
arch/powerpc/kernel/asm-offsets.c
449
OFFSET(VCPU_HFLAGS, kvm_vcpu, arch.hflags);
arch/powerpc/kernel/asm-offsets.c
450
OFFSET(VCPU_DEC_EXPIRES, kvm_vcpu, arch.dec_expires);
arch/powerpc/kernel/asm-offsets.c
451
OFFSET(VCPU_PENDING_EXC, kvm_vcpu, arch.pending_exceptions);
arch/powerpc/kernel/asm-offsets.c
452
OFFSET(VCPU_CEDED, kvm_vcpu, arch.ceded);
arch/powerpc/kernel/asm-offsets.c
453
OFFSET(VCPU_PRODDED, kvm_vcpu, arch.prodded);
arch/powerpc/kernel/asm-offsets.c
454
OFFSET(VCPU_MMCR, kvm_vcpu, arch.mmcr);
arch/powerpc/kernel/asm-offsets.c
455
OFFSET(VCPU_MMCRA, kvm_vcpu, arch.mmcra);
arch/powerpc/kernel/asm-offsets.c
456
OFFSET(VCPU_MMCRS, kvm_vcpu, arch.mmcrs);
arch/powerpc/kernel/asm-offsets.c
457
OFFSET(VCPU_PMC, kvm_vcpu, arch.pmc);
arch/powerpc/kernel/asm-offsets.c
458
OFFSET(VCPU_SIAR, kvm_vcpu, arch.siar);
arch/powerpc/kernel/asm-offsets.c
459
OFFSET(VCPU_SDAR, kvm_vcpu, arch.sdar);
arch/powerpc/kernel/asm-offsets.c
460
OFFSET(VCPU_SIER, kvm_vcpu, arch.sier);
arch/powerpc/kernel/asm-offsets.c
461
OFFSET(VCPU_SLB, kvm_vcpu, arch.slb);
arch/powerpc/kernel/asm-offsets.c
462
OFFSET(VCPU_SLB_MAX, kvm_vcpu, arch.slb_max);
arch/powerpc/kernel/asm-offsets.c
463
OFFSET(VCPU_SLB_NR, kvm_vcpu, arch.slb_nr);
arch/powerpc/kernel/asm-offsets.c
464
OFFSET(VCPU_FAULT_DSISR, kvm_vcpu, arch.fault_dsisr);
arch/powerpc/kernel/asm-offsets.c
465
OFFSET(VCPU_FAULT_DAR, kvm_vcpu, arch.fault_dar);
arch/powerpc/kernel/asm-offsets.c
466
OFFSET(VCPU_INTR_MSR, kvm_vcpu, arch.intr_msr);
arch/powerpc/kernel/asm-offsets.c
467
OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst);
arch/powerpc/kernel/asm-offsets.c
468
OFFSET(VCPU_TRAP, kvm_vcpu, arch.trap);
arch/powerpc/kernel/asm-offsets.c
469
OFFSET(VCPU_CFAR, kvm_vcpu, arch.cfar);
arch/powerpc/kernel/asm-offsets.c
470
OFFSET(VCPU_PPR, kvm_vcpu, arch.ppr);
arch/powerpc/kernel/asm-offsets.c
471
OFFSET(VCPU_FSCR, kvm_vcpu, arch.fscr);
arch/powerpc/kernel/asm-offsets.c
472
OFFSET(VCPU_PSPB, kvm_vcpu, arch.pspb);
arch/powerpc/kernel/asm-offsets.c
473
OFFSET(VCPU_EBBHR, kvm_vcpu, arch.ebbhr);
arch/powerpc/kernel/asm-offsets.c
474
OFFSET(VCPU_EBBRR, kvm_vcpu, arch.ebbrr);
arch/powerpc/kernel/asm-offsets.c
475
OFFSET(VCPU_BESCR, kvm_vcpu, arch.bescr);
arch/powerpc/kernel/asm-offsets.c
476
OFFSET(VCPU_CSIGR, kvm_vcpu, arch.csigr);
arch/powerpc/kernel/asm-offsets.c
477
OFFSET(VCPU_TACR, kvm_vcpu, arch.tacr);
arch/powerpc/kernel/asm-offsets.c
478
OFFSET(VCPU_TCSCR, kvm_vcpu, arch.tcscr);
arch/powerpc/kernel/asm-offsets.c
479
OFFSET(VCPU_ACOP, kvm_vcpu, arch.acop);
arch/powerpc/kernel/asm-offsets.c
480
OFFSET(VCPU_WORT, kvm_vcpu, arch.wort);
arch/powerpc/kernel/asm-offsets.c
481
OFFSET(VCPU_HFSCR, kvm_vcpu, arch.hfscr);
arch/powerpc/kernel/asm-offsets.c
496
OFFSET(VCPU_TFHAR, kvm_vcpu, arch.tfhar);
arch/powerpc/kernel/asm-offsets.c
497
OFFSET(VCPU_TFIAR, kvm_vcpu, arch.tfiar);
arch/powerpc/kernel/asm-offsets.c
498
OFFSET(VCPU_TEXASR, kvm_vcpu, arch.texasr);
arch/powerpc/kernel/asm-offsets.c
499
OFFSET(VCPU_ORIG_TEXASR, kvm_vcpu, arch.orig_texasr);
arch/powerpc/kernel/asm-offsets.c
500
OFFSET(VCPU_GPR_TM, kvm_vcpu, arch.gpr_tm);
arch/powerpc/kernel/asm-offsets.c
501
OFFSET(VCPU_FPRS_TM, kvm_vcpu, arch.fp_tm.fpr);
arch/powerpc/kernel/asm-offsets.c
502
OFFSET(VCPU_VRS_TM, kvm_vcpu, arch.vr_tm.vr);
arch/powerpc/kernel/asm-offsets.c
503
OFFSET(VCPU_VRSAVE_TM, kvm_vcpu, arch.vrsave_tm);
arch/powerpc/kernel/asm-offsets.c
504
OFFSET(VCPU_CR_TM, kvm_vcpu, arch.cr_tm);
arch/powerpc/kernel/asm-offsets.c
505
OFFSET(VCPU_XER_TM, kvm_vcpu, arch.xer_tm);
arch/powerpc/kernel/asm-offsets.c
506
OFFSET(VCPU_LR_TM, kvm_vcpu, arch.lr_tm);
arch/powerpc/kernel/asm-offsets.c
507
OFFSET(VCPU_CTR_TM, kvm_vcpu, arch.ctr_tm);
arch/powerpc/kernel/asm-offsets.c
508
OFFSET(VCPU_AMR_TM, kvm_vcpu, arch.amr_tm);
arch/powerpc/kernel/asm-offsets.c
509
OFFSET(VCPU_PPR_TM, kvm_vcpu, arch.ppr_tm);
arch/powerpc/kernel/asm-offsets.c
510
OFFSET(VCPU_DSCR_TM, kvm_vcpu, arch.dscr_tm);
arch/powerpc/kernel/asm-offsets.c
511
OFFSET(VCPU_TAR_TM, kvm_vcpu, arch.tar_tm);
arch/powerpc/kernel/asm-offsets.c
573
HSTATE_FIELD(HSTATE_KVM_VCPU, kvm_vcpu);
arch/powerpc/kernel/asm-offsets.c
611
OFFSET(VCPU_CR, kvm_vcpu, arch.regs.ccr);
arch/powerpc/kernel/asm-offsets.c
612
OFFSET(VCPU_XER, kvm_vcpu, arch.regs.xer);
arch/powerpc/kernel/asm-offsets.c
613
OFFSET(VCPU_LR, kvm_vcpu, arch.regs.link);
arch/powerpc/kernel/asm-offsets.c
614
OFFSET(VCPU_CTR, kvm_vcpu, arch.regs.ctr);
arch/powerpc/kernel/asm-offsets.c
615
OFFSET(VCPU_PC, kvm_vcpu, arch.regs.nip);
arch/powerpc/kernel/asm-offsets.c
616
OFFSET(VCPU_SPRG9, kvm_vcpu, arch.sprg9);
arch/powerpc/kernel/asm-offsets.c
617
OFFSET(VCPU_LAST_INST, kvm_vcpu, arch.last_inst);
arch/powerpc/kernel/asm-offsets.c
618
OFFSET(VCPU_FAULT_DEAR, kvm_vcpu, arch.fault_dear);
arch/powerpc/kernel/asm-offsets.c
619
OFFSET(VCPU_FAULT_ESR, kvm_vcpu, arch.fault_esr);
arch/powerpc/kernel/asm-offsets.c
620
OFFSET(VCPU_CRIT_SAVE, kvm_vcpu, arch.crit_save);
arch/powerpc/kernel/asm-offsets.c
648
OFFSET(VCPU_EVR, kvm_vcpu, arch.evr[0]);
arch/powerpc/kernel/asm-offsets.c
649
OFFSET(VCPU_ACC, kvm_vcpu, arch.acc);
arch/powerpc/kernel/asm-offsets.c
650
OFFSET(VCPU_SPEFSCR, kvm_vcpu, arch.spefscr);
arch/powerpc/kernel/asm-offsets.c
651
OFFSET(VCPU_HOST_SPEFSCR, kvm_vcpu, arch.host_spefscr);
arch/powerpc/kernel/asm-offsets.c
655
OFFSET(VCPU_HOST_MAS4, kvm_vcpu, arch.host_mas4);
arch/powerpc/kernel/asm-offsets.c
656
OFFSET(VCPU_HOST_MAS6, kvm_vcpu, arch.host_mas6);
arch/powerpc/kernel/asm-offsets.c
660
OFFSET(VCPU_TIMING_EXIT_TBU, kvm_vcpu, arch.timing_exit.tv32.tbu);
arch/powerpc/kernel/asm-offsets.c
661
OFFSET(VCPU_TIMING_EXIT_TBL, kvm_vcpu, arch.timing_exit.tv32.tbl);
arch/powerpc/kernel/asm-offsets.c
662
OFFSET(VCPU_TIMING_LAST_ENTER_TBU, kvm_vcpu, arch.timing_last_enter.tv32.tbu);
arch/powerpc/kernel/asm-offsets.c
663
OFFSET(VCPU_TIMING_LAST_ENTER_TBL, kvm_vcpu, arch.timing_last_enter.tv32.tbl);
arch/powerpc/kvm/book3s.c
1069
r = kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/powerpc/kvm/book3s.c
109
static inline bool kvmppc_critical_section(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
135
void kvmppc_inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 flags)
arch/powerpc/kvm/book3s.c
167
void kvmppc_book3s_dequeue_irqprio(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
179
void kvmppc_book3s_queue_irqprio(struct kvm_vcpu *vcpu, unsigned int vec)
arch/powerpc/kvm/book3s.c
191
void kvmppc_core_queue_machine_check(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
198
void kvmppc_core_queue_syscall(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
204
void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
211
void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
217
void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
223
void kvmppc_core_queue_vsx_unavail(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
229
void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
235
int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
241
void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
247
void kvmppc_core_queue_external(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
276
void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
281
void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, ulong srr1_flags,
arch/powerpc/kvm/book3s.c
290
void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/book3s.c
296
static int kvmppc_book3s_irqprio_deliver(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
372
static bool clear_irqprio(struct kvm_vcpu *vcpu, unsigned int priority)
arch/powerpc/kvm/book3s.c
394
int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
424
kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, bool writing,
arch/powerpc/kvm/book3s.c
451
int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid,
arch/powerpc/kvm/book3s.c
489
int kvmppc_load_last_inst(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
520
int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
525
void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
529
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
541
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
553
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/powerpc/kvm/book3s.c
581
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/powerpc/kvm/book3s.c
608
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/powerpc/kvm/book3s.c
613
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/powerpc/kvm/book3s.c
618
int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s.c
706
int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s.c
791
void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/book3s.c
796
void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
801
void kvmppc_set_msr(struct kvm_vcpu *vcpu, u64 msr)
arch/powerpc/kvm/book3s.c
807
int kvmppc_vcpu_run(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
812
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
818
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.c
827
void kvmppc_decrementer_func(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
833
int kvmppc_core_vcpu_create(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
838
void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
843
int kvmppc_core_check_requests(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
934
int kvmppc_h_logical_ci_load(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
976
int kvmppc_h_logical_ci_store(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s.c
98
static inline void kvmppc_update_int_pending(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.h
16
extern int kvmppc_mmu_init_pr(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s.h
17
extern void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s.h
18
extern int kvmppc_core_emulate_op_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.h
20
extern int kvmppc_core_emulate_mtspr_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.h
22
extern int kvmppc_core_emulate_mfspr_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s.h
26
extern int kvmppc_handle_exit_pr(struct kvm_vcpu *vcpu, unsigned int exit_nr);
arch/powerpc/kvm/book3s.h
29
extern void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val);
arch/powerpc/kvm/book3s.h
31
static inline void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val) {}
arch/powerpc/kvm/book3s.h
34
extern void kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 msr);
arch/powerpc/kvm/book3s.h
35
extern void kvmppc_inject_interrupt_hv(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags);
arch/powerpc/kvm/book3s_32_mmu.c
127
static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
182
static int kvmppc_mmu_book3s_32_xlate_pte(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
290
static int kvmppc_mmu_book3s_32_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
326
static u32 kvmppc_mmu_book3s_32_mfsrin(struct kvm_vcpu *vcpu, u32 srnum)
arch/powerpc/kvm/book3s_32_mmu.c
331
static void kvmppc_mmu_book3s_32_mtsrin(struct kvm_vcpu *vcpu, u32 srnum,
arch/powerpc/kvm/book3s_32_mmu.c
338
static void kvmppc_mmu_book3s_32_tlbie(struct kvm_vcpu *vcpu, ulong ea, bool large)
arch/powerpc/kvm/book3s_32_mmu.c
341
struct kvm_vcpu *v;
arch/powerpc/kvm/book3s_32_mmu.c
348
static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid,
arch/powerpc/kvm/book3s_32_mmu.c
391
static bool kvmppc_mmu_book3s_32_is_dcbz32(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu.c
397
void kvmppc_mmu_book3s_32_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu.c
40
static inline bool check_debug_ip(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu.c
69
static int kvmppc_mmu_book3s_32_xlate_bat(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
72
static int kvmppc_mmu_book3s_32_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid,
arch/powerpc/kvm/book3s_32_mmu.c
75
static u32 find_sr(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/book3s_32_mmu.c
80
static u64 kvmppc_mmu_book3s_32_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_32_mmu.c
93
static hva_t kvmppc_mmu_book3s_32_get_pteg(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_32_mmu_host.c
106
static u32 *kvmppc_mmu_get_pteg(struct kvm_vcpu *vcpu, u32 vsid, u32 eaddr,
arch/powerpc/kvm/book3s_32_mmu_host.c
128
int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte,
arch/powerpc/kvm/book3s_32_mmu_host.c
257
void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_32_mmu_host.c
262
static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_32_mmu_host.c
301
int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr)
arch/powerpc/kvm/book3s_32_mmu_host.c
332
void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu_host.c
344
void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu_host.c
355
int kvmppc_mmu_init_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_32_mmu_host.c
49
void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_32_mmu_host.c
66
static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_32_mmu_host.c
79
static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_64_mmu.c
116
static hva_t kvmppc_mmu_book3s_64_get_pteg(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu.c
191
static int kvmppc_mmu_book3s_64_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
28
struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu.c
363
static void kvmppc_mmu_book3s_64_slbmte(struct kvm_vcpu *vcpu, u64 rs, u64 rb)
arch/powerpc/kvm/book3s_64_mmu.c
412
static int kvmppc_mmu_book3s_64_slbfee(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu.c
425
static u64 kvmppc_mmu_book3s_64_slbmfee(struct kvm_vcpu *vcpu, u64 slb_nr)
arch/powerpc/kvm/book3s_64_mmu.c
437
static u64 kvmppc_mmu_book3s_64_slbmfev(struct kvm_vcpu *vcpu, u64 slb_nr)
arch/powerpc/kvm/book3s_64_mmu.c
449
static void kvmppc_mmu_book3s_64_slbie(struct kvm_vcpu *vcpu, u64 ea)
arch/powerpc/kvm/book3s_64_mmu.c
471
static void kvmppc_mmu_book3s_64_slbia(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu.c
489
static void kvmppc_mmu_book3s_64_mtsrin(struct kvm_vcpu *vcpu, u32 srnum,
arch/powerpc/kvm/book3s_64_mmu.c
529
static void kvmppc_mmu_book3s_64_tlbie(struct kvm_vcpu *vcpu, ulong va,
arch/powerpc/kvm/book3s_64_mmu.c
534
struct kvm_vcpu *v;
arch/powerpc/kvm/book3s_64_mmu.c
563
static int segment_contains_magic_page(struct kvm_vcpu *vcpu, ulong esid)
arch/powerpc/kvm/book3s_64_mmu.c
572
static int kvmppc_mmu_book3s_64_esid_to_vsid(struct kvm_vcpu *vcpu, ulong esid,
arch/powerpc/kvm/book3s_64_mmu.c
646
static bool kvmppc_mmu_book3s_64_is_dcbz32(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu.c
651
void kvmppc_mmu_book3s_64_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu.c
81
static u64 kvmppc_mmu_book3s_64_ea_to_vp(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_host.c
214
void kvmppc_mmu_unmap_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_64_mmu_host.c
225
static struct kvmppc_sid_map *create_sid_map(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_64_mmu_host.c
24
void kvmppc_mmu_invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_64_mmu_host.c
271
static int kvmppc_mmu_next_segment(struct kvm_vcpu *vcpu, ulong esid)
arch/powerpc/kvm/book3s_64_mmu_host.c
312
int kvmppc_mmu_map_segment(struct kvm_vcpu *vcpu, ulong eaddr)
arch/powerpc/kvm/book3s_64_mmu_host.c
33
static u16 kvmppc_sid_hash(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_64_mmu_host.c
358
void kvmppc_mmu_flush_segment(struct kvm_vcpu *vcpu, ulong ea, ulong seg_size)
arch/powerpc/kvm/book3s_64_mmu_host.c
375
void kvmppc_mmu_flush_segments(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu_host.c
383
void kvmppc_mmu_destroy_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu_host.c
389
int kvmppc_mmu_init_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu_host.c
46
static struct kvmppc_sid_map *find_sid_vsid(struct kvm_vcpu *vcpu, u64 gvsid)
arch/powerpc/kvm/book3s_64_mmu_host.c
71
int kvmppc_mmu_map_page(struct kvm_vcpu *vcpu, struct kvmppc_pte *orig_pte,
arch/powerpc/kvm/book3s_64_mmu_hv.c
206
void kvmppc_map_vrma(struct kvm_vcpu *vcpu, struct kvm_memory_slot *memslot,
arch/powerpc/kvm/book3s_64_mmu_hv.c
2112
void kvmppc_mmu_book3s_hv_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_64_mmu_hv.c
311
static struct kvmppc_slb *kvmppc_mmu_book3s_hv_find_slbe(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu_hv.c
341
static int kvmppc_mmu_book3s_64_hv_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_hv.c
433
int kvmppc_hv_emulate_mmio(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu_hv.c
506
int kvmppc_book3s_hv_page_fault(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu_radix.c
124
long kvmhv_copy_from_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *to,
arch/powerpc/kvm/book3s_64_mmu_radix.c
136
long kvmhv_copy_to_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr, void *from,
arch/powerpc/kvm/book3s_64_mmu_radix.c
142
int kvmppc_mmu_walk_radix_tree(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
238
int kvmppc_mmu_radix_translate_table(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
269
int kvmppc_mmu_radix_xlate(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_mmu_radix.c
821
int kvmppc_book3s_instantiate_page(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu_radix.c
926
int kvmppc_book3s_radix_page_fault(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_mmu_radix.c
99
static long kvmhv_copy_tofrom_guest_radix(struct kvm_vcpu *vcpu, gva_t eaddr,
arch/powerpc/kvm/book3s_64_vio.c
570
long kvmppc_h_put_tce(struct kvm_vcpu *vcpu, unsigned long liobn,
arch/powerpc/kvm/book3s_64_vio.c
629
long kvmppc_h_put_tce_indirect(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_vio.c
722
long kvmppc_h_stuff_tce(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_64_vio.c
767
long kvmppc_h_get_tce(struct kvm_vcpu *vcpu, unsigned long liobn,
arch/powerpc/kvm/book3s_emulate.c
103
static inline void kvmppc_copyfrom_vcpu_tm(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_emulate.c
1034
u32 kvmppc_alignment_dsisr(struct kvm_vcpu *vcpu, unsigned int inst)
arch/powerpc/kvm/book3s_emulate.c
1039
ulong kvmppc_alignment_dar(struct kvm_vcpu *vcpu, unsigned int inst)
arch/powerpc/kvm/book3s_emulate.c
122
static void kvmppc_emulate_treclaim(struct kvm_vcpu *vcpu, int ra_val)
arch/powerpc/kvm/book3s_emulate.c
168
static void kvmppc_emulate_trchkpt(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_emulate.c
193
void kvmppc_emulate_tabort(struct kvm_vcpu *vcpu, int ra_val)
arch/powerpc/kvm/book3s_emulate.c
234
int kvmppc_core_emulate_op_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_emulate.c
633
void kvmppc_set_bat(struct kvm_vcpu *vcpu, struct kvmppc_bat *bat, bool upper,
arch/powerpc/kvm/book3s_emulate.c
653
static struct kvmppc_bat *kvmppc_find_bat(struct kvm_vcpu *vcpu, int sprn)
arch/powerpc/kvm/book3s_emulate.c
678
int kvmppc_core_emulate_mtspr_pr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
arch/powerpc/kvm/book3s_emulate.c
70
static bool spr_allowed(struct kvm_vcpu *vcpu, enum priv_level level)
arch/powerpc/kvm/book3s_emulate.c
84
static inline void kvmppc_copyto_vcpu_tm(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_emulate.c
864
int kvmppc_core_emulate_mfspr_pr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
arch/powerpc/kvm/book3s_hv.c
1011
static long kvmppc_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv.c
1043
static int kvm_arch_vcpu_yield_to(struct kvm_vcpu *target)
arch/powerpc/kvm/book3s_hv.c
1070
static int kvmppc_get_yield_count(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1088
static int kvmppc_nested_h_rpt_invalidate(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1111
static long kvmppc_h_rpt_invalidate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
1145
int kvmppc_pseries_do_hcall(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1151
struct kvm_vcpu *tvcpu;
arch/powerpc/kvm/book3s_hv.c
135
static int kvmppc_hv_setup_htab_rma(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_hv.c
1429
static void kvmppc_cede(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1474
static int kvmppc_emulate_debug_inst(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1502
static unsigned long kvmppc_read_dpdes(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1505
struct kvm_vcpu *v;
arch/powerpc/kvm/book3s_hv.c
1535
static int kvmppc_emulate_doorbell_instr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1540
struct kvm_vcpu *tvcpu;
arch/powerpc/kvm/book3s_hv.c
1600
static int kvmppc_pmu_unavailable(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1610
static int kvmppc_ebb_unavailable(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
1620
static int kvmppc_tm_unavailable(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
163
static inline struct kvm_vcpu *next_runnable_thread(struct kvmppc_vcore *vc,
arch/powerpc/kvm/book3s_hv.c
1630
static int kvmppc_handle_exit_hv(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
167
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
1969
static int kvmppc_handle_nested_exit(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
2101
static int kvm_arch_vcpu_ioctl_get_sregs_hv(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
2116
static int kvm_arch_vcpu_ioctl_set_sregs_hv(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
2181
static void kvmppc_set_lpcr(struct kvm_vcpu *vcpu, u64 new_lpcr,
arch/powerpc/kvm/book3s_hv.c
2211
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
2230
static int kvmppc_get_one_reg_hv(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s_hv.c
227
static void kvmppc_fast_vcpu_kick_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
2488
static int kvmppc_set_one_reg_hv(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s_hv.c
2816
{"vcpu_entry", offsetof(struct kvm_vcpu, arch.vcpu_entry)},
arch/powerpc/kvm/book3s_hv.c
2817
{"guest_entry", offsetof(struct kvm_vcpu, arch.guest_entry)},
arch/powerpc/kvm/book3s_hv.c
2818
{"in_guest", offsetof(struct kvm_vcpu, arch.in_guest)},
arch/powerpc/kvm/book3s_hv.c
2819
{"guest_exit", offsetof(struct kvm_vcpu, arch.guest_exit)},
arch/powerpc/kvm/book3s_hv.c
2820
{"vcpu_exit", offsetof(struct kvm_vcpu, arch.vcpu_exit)},
arch/powerpc/kvm/book3s_hv.c
2821
{"hypercall", offsetof(struct kvm_vcpu, arch.hcall)},
arch/powerpc/kvm/book3s_hv.c
2822
{"page_fault", offsetof(struct kvm_vcpu, arch.pg_fault)},
arch/powerpc/kvm/book3s_hv.c
2824
{"rm_entry", offsetof(struct kvm_vcpu, arch.rm_entry)},
arch/powerpc/kvm/book3s_hv.c
2825
{"rm_intr", offsetof(struct kvm_vcpu, arch.rm_intr)},
arch/powerpc/kvm/book3s_hv.c
2826
{"rm_exit", offsetof(struct kvm_vcpu, arch.rm_exit)},
arch/powerpc/kvm/book3s_hv.c
2827
{"guest", offsetof(struct kvm_vcpu, arch.guest_time)},
arch/powerpc/kvm/book3s_hv.c
2828
{"cede", offsetof(struct kvm_vcpu, arch.cede_time)},
arch/powerpc/kvm/book3s_hv.c
2835
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
2842
struct kvm_vcpu *vcpu = inode->i_private;
arch/powerpc/kvm/book3s_hv.c
2869
struct kvm_vcpu *vcpu = p->vcpu;
arch/powerpc/kvm/book3s_hv.c
2946
static int kvmppc_arch_create_vcpu_debugfs_hv(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry)
arch/powerpc/kvm/book3s_hv.c
2955
static int kvmppc_arch_create_vcpu_debugfs_hv(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry)
arch/powerpc/kvm/book3s_hv.c
2961
static int kvmppc_core_vcpu_create_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3140
static void kvmppc_core_vcpu_free_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3151
static int kvmppc_core_check_requests_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3157
static void kvmppc_set_timer(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3176
struct kvm_vcpu *vcpu, u64 tb)
arch/powerpc/kvm/book3s_hv.c
318
static void kvmppc_core_vcpu_load_hv(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/book3s_hv.c
3201
tpaca->kvm_hstate.kvm_vcpu = NULL;
arch/powerpc/kvm/book3s_hv.c
3233
tpaca->kvm_hstate.kvm_vcpu = NULL;
arch/powerpc/kvm/book3s_hv.c
3240
static void radix_flush_cpu(struct kvm *kvm, int cpu, struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3276
struct kvm_vcpu *vcpu = arg;
arch/powerpc/kvm/book3s_hv.c
3293
static void kvmppc_prepare_radix_vcpu(struct kvm_vcpu *vcpu, int pcpu)
arch/powerpc/kvm/book3s_hv.c
3334
static void kvmppc_start_thread(struct kvm_vcpu *vcpu, struct kvmppc_vcore *vc)
arch/powerpc/kvm/book3s_hv.c
3350
tpaca->kvm_hstate.kvm_vcpu = vcpu;
arch/powerpc/kvm/book3s_hv.c
353
static void kvmppc_core_vcpu_put_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
3584
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
3634
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
3653
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
3779
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
386
static void kvmppc_set_pvr_hv(struct kvm_vcpu *vcpu, u32 pvr)
arch/powerpc/kvm/book3s_hv.c
4101
static void vcpu_vpa_increment_dispatch(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
415
static int kvmppc_set_arch_compat(struct kvm_vcpu *vcpu, u32 arch_compat)
arch/powerpc/kvm/book3s_hv.c
4153
static void do_trace_nested_cs_time(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4195
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
4198
vcpu = local_paca->kvm_hstate.kvm_vcpu;
arch/powerpc/kvm/book3s_hv.c
4210
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
4213
vcpu = local_paca->kvm_hstate.kvm_vcpu;
arch/powerpc/kvm/book3s_hv.c
4225
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
4228
vcpu = local_paca->kvm_hstate.kvm_vcpu;
arch/powerpc/kvm/book3s_hv.c
4244
static void do_trace_nested_cs_time(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4249
static int kvmhv_vcpu_entry_nestedv2(struct kvm_vcpu *vcpu, u64 time_limit,
arch/powerpc/kvm/book3s_hv.c
4310
static int kvmhv_vcpu_entry_p9_nested(struct kvm_vcpu *vcpu, u64 time_limit, unsigned long lpcr, u64 *tb)
arch/powerpc/kvm/book3s_hv.c
4414
static int kvmhv_p9_guest_entry(struct kvm_vcpu *vcpu, u64 time_limit,
arch/powerpc/kvm/book3s_hv.c
4522
struct kvm_vcpu *vcpu, int wait_state)
arch/powerpc/kvm/book3s_hv.c
4554
static inline bool xive_interrupt_pending(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4562
static inline bool xive_interrupt_pending(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4568
static bool kvmppc_vcpu_woken(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4577
static bool kvmppc_vcpu_check_block(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4590
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
4719
static int kvmhv_setup_mmu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4738
static int kvmppc_run_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4743
struct kvm_vcpu *v;
arch/powerpc/kvm/book3s_hv.c
486
static void kvmppc_dump_regs(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
4868
int kvmhv_run_single_vcpu(struct kvm_vcpu *vcpu, u64 time_limit,
arch/powerpc/kvm/book3s_hv.c
4926
local_paca->kvm_hstate.kvm_vcpu = vcpu;
arch/powerpc/kvm/book3s_hv.c
5085
static int kvmppc_vcpu_run_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
519
static struct kvm_vcpu *kvmppc_find_vcpu(struct kvm *kvm, int id)
arch/powerpc/kvm/book3s_hv.c
524
static void init_vpa(struct kvm_vcpu *vcpu, struct lppaca *vpa)
arch/powerpc/kvm/book3s_hv.c
5270
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
530
static int set_vpa(struct kvm_vcpu *vcpu, struct kvmppc_vpa *v,
arch/powerpc/kvm/book3s_hv.c
5437
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
5470
static int kvmppc_hv_setup_htab_rma(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
562
static unsigned long do_h_register_vpa(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
569
struct kvm_vcpu *tvcpu;
arch/powerpc/kvm/book3s_hv.c
5923
static int kvmppc_core_emulate_op_hv(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
5929
static int kvmppc_core_emulate_mtspr_hv(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/book3s_hv.c
5935
static int kvmppc_core_emulate_mfspr_hv(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/book3s_hv.c
6338
static int kvmhv_load_from_eaddr(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
arch/powerpc/kvm/book3s_hv.c
6357
static int kvmhv_store_to_eaddr(struct kvm_vcpu *vcpu, ulong *eaddr, void *ptr,
arch/powerpc/kvm/book3s_hv.c
6410
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv.c
679
static void kvmppc_update_vpa(struct kvm_vcpu *vcpu, struct kvmppc_vpa *vpap,
arch/powerpc/kvm/book3s_hv.c
729
static void kvmppc_update_vpas(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
792
static void __kvmppc_create_dtl_entry(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
826
static void kvmppc_update_vpa_dispatch(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
856
static void kvmppc_update_vpa_dispatch_p9(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.c
880
static bool kvmppc_doorbell_pending(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
900
static bool kvmppc_power8_compatible(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.c
910
static int kvmppc_h_set_mode(struct kvm_vcpu *vcpu, unsigned long mflags,
arch/powerpc/kvm/book3s_hv.h
35
bool load_vcpu_state(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.h
37
void store_vcpu_state(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_hv.h
39
void restore_p9_host_os_sprs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.h
41
void switch_pmu_to_guest(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.h
43
void switch_pmu_to_host(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv.h
47
void accumulate_time(struct kvm_vcpu *vcpu, struct kvmhv_tb_accumulator *next);
arch/powerpc/kvm/book3s_hv.h
56
static inline void __kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 val)
arch/powerpc/kvm/book3s_hv.h
62
static inline u64 __kvmppc_get_msr_hv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv.h
69
static inline void kvmppc_set_##reg ##_hv(struct kvm_vcpu *vcpu, u##size val) \
arch/powerpc/kvm/book3s_hv.h
76
static inline u##size kvmppc_get_##reg ##_hv(struct kvm_vcpu *vcpu) \
arch/powerpc/kvm/book3s_hv.h
87
static inline void kvmppc_set_##reg ##_hv(struct kvm_vcpu *vcpu, int i, u##size val) \
arch/powerpc/kvm/book3s_hv.h
94
static inline u##size kvmppc_get_##reg ##_hv(struct kvm_vcpu *vcpu, int i) \
arch/powerpc/kvm/book3s_hv_builtin.c
110
long int kvmppc_rm_h_confer(struct kvm_vcpu *vcpu, int target,
arch/powerpc/kvm/book3s_hv_builtin.c
184
long kvmppc_rm_h_random(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_builtin.c
347
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv_builtin.c
349
vcpu = local_paca->kvm_hstate.kvm_vcpu;
arch/powerpc/kvm/book3s_hv_builtin.c
498
static void kvmppc_end_cede(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_builtin.c
507
void kvmppc_set_msr_hv(struct kvm_vcpu *vcpu, u64 msr)
arch/powerpc/kvm/book3s_hv_builtin.c
523
static void inject_interrupt(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags)
arch/powerpc/kvm/book3s_hv_builtin.c
559
void kvmppc_inject_interrupt_hv(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags)
arch/powerpc/kvm/book3s_hv_builtin.c
570
void kvmppc_guest_entry_inject_int(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nested.c
103
static void save_hv_return_state(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1073
static bool kvmhv_invalidate_shadow_pte(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1132
static int kvmhv_emulate_tlbie_tlb_addr(struct kvm_vcpu *vcpu, int lpid,
arch/powerpc/kvm/book3s_hv_nested.c
1168
static void kvmhv_emulate_tlbie_lpid(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1199
static void kvmhv_emulate_tlbie_all_lpid(struct kvm_vcpu *vcpu, int ric)
arch/powerpc/kvm/book3s_hv_nested.c
1214
static int kvmhv_emulate_priv_tlbie(struct kvm_vcpu *vcpu, unsigned int instr,
arch/powerpc/kvm/book3s_hv_nested.c
1277
long kvmhv_do_nested_tlbie(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nested.c
1288
static long do_tlb_invalidate_nested_all(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1308
static long do_tlb_invalidate_nested_tlb(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1348
long do_h_rpt_invalidate_pat(struct kvm_vcpu *vcpu, unsigned long lpid,
arch/powerpc/kvm/book3s_hv_nested.c
1396
static int kvmhv_translate_addr_nested(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
141
static void restore_hv_regs(struct kvm_vcpu *vcpu, const struct hv_guest_state *hr)
arch/powerpc/kvm/book3s_hv_nested.c
1458
static long kvmhv_handle_nested_set_rc(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1523
static long int __kvmhv_nested_page_fault(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
168
void kvmhv_restore_hv_return_state(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
1693
long int kvmhv_nested_page_fault(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nested.c
200
static void kvmhv_nested_mmio_needed(struct kvm_vcpu *vcpu, u64 regs_ptr)
arch/powerpc/kvm/book3s_hv_nested.c
220
static int kvmhv_read_guest_state_and_regs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
243
static int kvmhv_write_guest_state_and_regs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
259
static void load_l2_hv_regs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nested.c
295
long kvmhv_enter_nested_guest(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nested.c
30
void kvmhv_save_hv_regs(struct kvm_vcpu *vcpu, struct hv_guest_state *hr)
arch/powerpc/kvm/book3s_hv_nested.c
568
long kvmhv_set_partition_table(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nested.c
596
long kvmhv_copy_tofrom_guest_nested(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nestedv2.c
1013
int __kvmhv_nestedv2_mark_dirty_ptregs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nestedv2.c
1036
int kvmhv_nestedv2_vcpu_create(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nestedv2.c
1067
void kvmhv_nestedv2_vcpu_free(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nestedv2.c
143
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv_nestedv2.c
405
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_hv_nestedv2.c
641
static int kvmhv_nestedv2_host_create(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nestedv2.c
756
int __kvmhv_nestedv2_mark_dirty(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/kvm/book3s_hv_nestedv2.c
784
int __kvmhv_nestedv2_cached_reload(struct kvm_vcpu *vcpu, u16 iden)
arch/powerpc/kvm/book3s_hv_nestedv2.c
822
int kvmhv_nestedv2_flush_vcpu(struct kvm_vcpu *vcpu, u64 time_limit)
arch/powerpc/kvm/book3s_hv_nestedv2.c
912
int kvmhv_nestedv2_set_vpa(struct kvm_vcpu *vcpu, unsigned long vpa)
arch/powerpc/kvm/book3s_hv_nestedv2.c
942
int kvmhv_nestedv2_parse_output(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_nestedv2.c
961
static void kvmhv_nestedv2_host_free(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_nestedv2.c
970
int __kvmhv_nestedv2_reload_ptregs(struct kvm_vcpu *vcpu, struct pt_regs *regs)
arch/powerpc/kvm/book3s_hv_p9_entry.c
10
static void load_spr_state(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_p9_entry.c
117
void store_vcpu_state(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_p9_entry.c
156
void restore_p9_host_os_sprs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_p9_entry.c
220
void accumulate_time(struct kvm_vcpu *vcpu, struct kvmhv_tb_accumulator *next)
arch/powerpc/kvm/book3s_hv_p9_entry.c
301
static void switch_mmu_to_guest_radix(struct kvm *kvm, struct kvm_vcpu *vcpu, u64 lpcr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
326
static void switch_mmu_to_guest_hpt(struct kvm *kvm, struct kvm_vcpu *vcpu, u64 lpcr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
391
static void save_clear_guest_mmu(struct kvm *kvm, struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_p9_entry.c
494
unsigned long kvmppc_msr_hard_disable_set_facilities(struct kvm_vcpu *vcpu, unsigned long msr)
arch/powerpc/kvm/book3s_hv_p9_entry.c
532
int kvmhv_vcpu_entry_p9(struct kvm_vcpu *vcpu, u64 time_limit, unsigned long lpcr, u64 *tb)
arch/powerpc/kvm/book3s_hv_p9_entry.c
58
static void store_spr_state(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_p9_entry.c
86
bool load_vcpu_state(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_p9_perf.c
118
void switch_pmu_to_host(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_p9_perf.c
35
void switch_pmu_to_guest(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_ras.c
112
void kvmppc_realmode_machine_check(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_ras.c
141
long kvmppc_p9_realmode_hmi_handler(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_ras.c
37
static void reload_slb(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_ras.c
69
static long kvmppc_realmode_mc_power7(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1062
static struct mmio_hpte_cache_entry *mmio_cache_search(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1082
next_mmio_cache_entry(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
1190
long kvmppc_hpte_hv_fault(struct kvm_vcpu *vcpu, unsigned long addr,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
401
long kvmppc_h_enter(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
44
if (kvm->arch.online_vcores == 1 && local_paca->kvm_hstate.kvm_vcpu)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
547
long kvmppc_h_remove(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
555
long kvmppc_h_bulk_remove(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_mmu.c
674
long kvmppc_h_protect(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
745
long kvmppc_h_read(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
786
long kvmppc_h_clear_ref(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
835
long kvmppc_h_clear_mod(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
883
static int kvmppc_get_hpa(struct kvm_vcpu *vcpu, unsigned long mmu_seq,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
925
static long kvmppc_do_h_page_init_zero(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
954
static long kvmppc_do_h_page_init_copy(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_rm_mmu.c
985
long kvmppc_rm_h_page_init(struct kvm_vcpu *vcpu, unsigned long flags,
arch/powerpc/kvm/book3s_hv_rm_xics.c
127
static void icp_rm_set_vcpu_irq(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_rm_xics.c
128
struct kvm_vcpu *this_vcpu)
arch/powerpc/kvm/book3s_hv_rm_xics.c
167
static void icp_rm_clr_vcpu_irq(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_xics.c
178
struct kvm_vcpu *this_vcpu = local_paca->kvm_hstate.kvm_vcpu;
arch/powerpc/kvm/book3s_hv_rm_xics.c
482
unsigned long xics_rm_h_xirr_x(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_xics.c
488
unsigned long xics_rm_h_xirr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_xics.c
526
int xics_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server,
arch/powerpc/kvm/book3s_hv_rm_xics.c
54
static inline void icp_send_hcore_msg(int hcore, struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_rm_xics.c
613
int xics_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr)
arch/powerpc/kvm/book3s_hv_rm_xics.c
66
static inline void icp_send_hcore_msg(int hcore, struct kvm_vcpu *vcpu) { }
arch/powerpc/kvm/book3s_hv_rm_xics.c
675
static int ics_rm_eoi(struct kvm_vcpu *vcpu, u32 irq)
arch/powerpc/kvm/book3s_hv_rm_xics.c
734
int xics_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr)
arch/powerpc/kvm/book3s_hv_rm_xics.c
844
long kvmppc_deliver_irq_passthru(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_hv_tm.c
16
static void emulate_tx_failure(struct kvm_vcpu *vcpu, u64 failure_cause)
arch/powerpc/kvm/book3s_hv_tm.c
42
int kvmhv_p9_tm_emulation(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_tm_builtin.c
113
void kvmhv_emulate_tm_rollback(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_hv_tm_builtin.c
20
int kvmhv_p9_tm_emulation_early(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_mmu_hpte.c
126
static void kvmppc_mmu_pte_flush_all(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_mmu_hpte.c
144
static void kvmppc_mmu_pte_flush_page(struct kvm_vcpu *vcpu, ulong guest_ea)
arch/powerpc/kvm/book3s_mmu_hpte.c
163
static void kvmppc_mmu_pte_flush_long(struct kvm_vcpu *vcpu, ulong guest_ea)
arch/powerpc/kvm/book3s_mmu_hpte.c
183
void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong guest_ea, ulong ea_mask)
arch/powerpc/kvm/book3s_mmu_hpte.c
206
static void kvmppc_mmu_pte_vflush_short(struct kvm_vcpu *vcpu, u64 guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
227
static void kvmppc_mmu_pte_vflush_64k(struct kvm_vcpu *vcpu, u64 guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
249
static void kvmppc_mmu_pte_vflush_long(struct kvm_vcpu *vcpu, u64 guest_vp)
arch/powerpc/kvm/book3s_mmu_hpte.c
269
void kvmppc_mmu_pte_vflush(struct kvm_vcpu *vcpu, u64 guest_vp, u64 vp_mask)
arch/powerpc/kvm/book3s_mmu_hpte.c
292
void kvmppc_mmu_pte_pflush(struct kvm_vcpu *vcpu, ulong pa_start, ulong pa_end)
arch/powerpc/kvm/book3s_mmu_hpte.c
314
struct hpte_cache *kvmppc_mmu_hpte_cache_next(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_mmu_hpte.c
332
void kvmppc_mmu_hpte_destroy(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_mmu_hpte.c
345
int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_mmu_hpte.c
56
void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_mmu_hpte.c
95
static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte)
arch/powerpc/kvm/book3s_paired_singles.c
150
static inline void kvmppc_sync_qpr(struct kvm_vcpu *vcpu, int rt)
arch/powerpc/kvm/book3s_paired_singles.c
155
static void kvmppc_inject_pf(struct kvm_vcpu *vcpu, ulong eaddr, bool is_store)
arch/powerpc/kvm/book3s_paired_singles.c
172
static int kvmppc_emulate_fpr_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_paired_singles.c
216
static int kvmppc_emulate_fpr_store(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_paired_singles.c
262
static int kvmppc_emulate_psq_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_paired_singles.c
305
static int kvmppc_emulate_psq_store(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_paired_singles.c
344
static bool kvmppc_inst_is_paired_single(struct kvm_vcpu *vcpu, u32 inst)
arch/powerpc/kvm/book3s_paired_singles.c
490
static int kvmppc_ps_three_in(struct kvm_vcpu *vcpu, bool rc,
arch/powerpc/kvm/book3s_paired_singles.c
538
static int kvmppc_ps_two_in(struct kvm_vcpu *vcpu, bool rc,
arch/powerpc/kvm/book3s_paired_singles.c
590
static int kvmppc_ps_one_in(struct kvm_vcpu *vcpu, bool rc,
arch/powerpc/kvm/book3s_paired_singles.c
621
int kvmppc_emulate_paired_single(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1036
void kvmppc_set_fscr(struct kvm_vcpu *vcpu, u64 fscr)
arch/powerpc/kvm/book3s_pr.c
1055
static void kvmppc_setup_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1064
static void kvmppc_clear_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
107
static void kvmppc_inject_interrupt_pr(struct kvm_vcpu *vcpu, int vec, u64 srr1_flags)
arch/powerpc/kvm/book3s_pr.c
1073
static int kvmppc_exit_pr_progint(struct kvm_vcpu *vcpu, unsigned int exit_nr)
arch/powerpc/kvm/book3s_pr.c
1135
int kvmppc_handle_exit_pr(struct kvm_vcpu *vcpu, unsigned int exit_nr)
arch/powerpc/kvm/book3s_pr.c
132
static void kvmppc_core_vcpu_load_pr(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/book3s_pr.c
1455
static int kvm_arch_vcpu_ioctl_get_sregs_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_pr.c
1482
static int kvm_arch_vcpu_ioctl_set_sregs_pr(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_pr.c
1528
static int kvmppc_get_one_reg_pr(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s_pr.c
161
static void kvmppc_core_vcpu_put_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1628
static void kvmppc_set_lpcr_pr(struct kvm_vcpu *vcpu, u64 new_lpcr)
arch/powerpc/kvm/book3s_pr.c
1636
static int kvmppc_set_one_reg_pr(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/book3s_pr.c
1727
static int kvmppc_core_vcpu_create_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1795
static void kvmppc_core_vcpu_free_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1807
static int kvmppc_vcpu_run_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1866
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_pr.c
192
void kvmppc_copy_to_svcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
1928
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_pr.c
231
static void kvmppc_recalc_shadow_msr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
264
void kvmppc_copy_from_svcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
343
void kvmppc_save_tm_sprs(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
352
void kvmppc_restore_tm_sprs(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
364
static void kvmppc_handle_lost_math_exts(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
383
void kvmppc_save_tm_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
398
void kvmppc_restore_tm_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
422
static int kvmppc_core_check_requests_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
438
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_pr.c
466
static void kvmppc_set_msr_pr(struct kvm_vcpu *vcpu, u64 msr)
arch/powerpc/kvm/book3s_pr.c
53
static int kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr,
arch/powerpc/kvm/book3s_pr.c
550
static void kvmppc_set_pvr_pr(struct kvm_vcpu *vcpu, u32 pvr)
arch/powerpc/kvm/book3s_pr.c
56
static int kvmppc_handle_fac(struct kvm_vcpu *vcpu, ulong fac);
arch/powerpc/kvm/book3s_pr.c
640
static void kvmppc_patch_dcbz(struct kvm_vcpu *vcpu, struct kvmppc_pte *pte)
arch/powerpc/kvm/book3s_pr.c
665
static bool kvmppc_visible_gpa(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/powerpc/kvm/book3s_pr.c
67
static bool kvmppc_is_split_real(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
680
static int kvmppc_handle_pagefault(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_pr.c
73
static void kvmppc_fixup_split_real(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
797
void kvmppc_giveup_ext(struct kvm_vcpu *vcpu, ulong msr)
arch/powerpc/kvm/book3s_pr.c
840
void kvmppc_giveup_fac(struct kvm_vcpu *vcpu, ulong fac)
arch/powerpc/kvm/book3s_pr.c
859
static int kvmppc_handle_ext(struct kvm_vcpu *vcpu, unsigned int exit_nr,
arch/powerpc/kvm/book3s_pr.c
930
static void kvmppc_handle_lost_ext(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
94
static void kvmppc_unfixup_split_real(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr.c
959
void kvmppc_trigger_fac_interrupt(struct kvm_vcpu *vcpu, ulong fac)
arch/powerpc/kvm/book3s_pr.c
967
static void kvmppc_emulate_fac(struct kvm_vcpu *vcpu, ulong fac)
arch/powerpc/kvm/book3s_pr.c
981
static int kvmppc_handle_fac(struct kvm_vcpu *vcpu, ulong fac)
arch/powerpc/kvm/book3s_pr_papr.c
145
static int kvmppc_h_pr_bulk_remove(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
214
static int kvmppc_h_pr_protect(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
23
static unsigned long get_pteg_addr(struct kvm_vcpu *vcpu, long pte_index)
arch/powerpc/kvm/book3s_pr_papr.c
262
static int kvmppc_h_pr_logical_ci_load(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
273
static int kvmppc_h_pr_logical_ci_store(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
284
static int kvmppc_h_pr_set_mode(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
301
static int kvmppc_h_pr_put_tce(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
315
static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
331
static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
347
static int kvmppc_h_pr_put_tce(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
352
static int kvmppc_h_pr_put_tce_indirect(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
357
static int kvmppc_h_pr_stuff_tce(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
36
static int kvmppc_h_pr_enter(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_pr_papr.c
363
static int kvmppc_h_pr_xics_hcall(struct kvm_vcpu *vcpu, u32 cmd)
arch/powerpc/kvm/book3s_pr_papr.c
370
int kvmppc_h_pr(struct kvm_vcpu *vcpu, unsigned long cmd)
arch/powerpc/kvm/book3s_pr_papr.c
86
static int kvmppc_h_pr_remove(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_rtas.c
117
void (*handler)(struct kvm_vcpu *vcpu, struct rtas_args *args);
arch/powerpc/kvm/book3s_rtas.c
19
static void kvm_rtas_set_xive(struct kvm_vcpu *vcpu, struct rtas_args *args)
arch/powerpc/kvm/book3s_rtas.c
218
int kvmppc_rtas_hcall(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_rtas.c
43
static void kvm_rtas_get_xive(struct kvm_vcpu *vcpu, struct rtas_args *args)
arch/powerpc/kvm/book3s_rtas.c
71
static void kvm_rtas_int_off(struct kvm_vcpu *vcpu, struct rtas_args *args)
arch/powerpc/kvm/book3s_rtas.c
93
static void kvm_rtas_int_on(struct kvm_vcpu *vcpu, struct rtas_args *args)
arch/powerpc/kvm/book3s_xics.c
1062
static int kvmppc_xics_create_icp(struct kvm_vcpu *vcpu, unsigned long server_num)
arch/powerpc/kvm/book3s_xics.c
1087
u64 kvmppc_xics_get_icp(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xics.c
1101
int kvmppc_xics_set_icp(struct kvm_vcpu *vcpu, u64 icpval)
arch/powerpc/kvm/book3s_xics.c
1336
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xics.c
1449
int kvmppc_xics_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu,
arch/powerpc/kvm/book3s_xics.c
1469
void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xics.c
586
static noinline unsigned long kvmppc_h_xirr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xics.c
619
static noinline int kvmppc_h_ipi(struct kvm_vcpu *vcpu, unsigned long server,
arch/powerpc/kvm/book3s_xics.c
704
static int kvmppc_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server)
arch/powerpc/kvm/book3s_xics.c
721
static noinline void kvmppc_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr)
arch/powerpc/kvm/book3s_xics.c
777
static int ics_eoi(struct kvm_vcpu *vcpu, u32 irq)
arch/powerpc/kvm/book3s_xics.c
817
static noinline int kvmppc_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr)
arch/powerpc/kvm/book3s_xics.c
848
int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall)
arch/powerpc/kvm/book3s_xics.c
875
int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 req)
arch/powerpc/kvm/book3s_xics.c
944
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xics.h
118
struct kvm_vcpu *vcpu = NULL;
arch/powerpc/kvm/book3s_xics.h
145
extern unsigned long xics_rm_h_xirr(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_xics.h
146
extern unsigned long xics_rm_h_xirr_x(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_xics.h
147
extern int xics_rm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server,
arch/powerpc/kvm/book3s_xics.h
149
extern int xics_rm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr);
arch/powerpc/kvm/book3s_xics.h
150
extern int xics_rm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr);
arch/powerpc/kvm/book3s_xics.h
67
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xics.h
79
struct kvm_vcpu *rm_kick_target;
arch/powerpc/kvm/book3s_xics.h
94
struct kvm_vcpu *rm_dbgtgt;
arch/powerpc/kvm/book3s_xive.c
1006
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.c
1024
static int xive_try_pick_queue(struct kvm_vcpu *vcpu, u8 prio)
arch/powerpc/kvm/book3s_xive.c
1046
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.c
1478
u64 kvmppc_xive_get_icp(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
1491
int kvmppc_xive_set_icp(struct kvm_vcpu *vcpu, u64 icpval)
arch/powerpc/kvm/book3s_xive.c
1728
void kvmppc_xive_disable_vcpu_interrupts(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
1787
void xive_cleanup_single_escalation(struct kvm_vcpu *vcpu, int irq)
arch/powerpc/kvm/book3s_xive.c
1802
void kvmppc_xive_cleanup_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
1902
struct kvm_vcpu *vcpu, u32 cpu)
arch/powerpc/kvm/book3s_xive.c
2124
struct kvm_vcpu *vcpu = NULL;
arch/powerpc/kvm/book3s_xive.c
2306
struct kvm_vcpu *vcpu = NULL;
arch/powerpc/kvm/book3s_xive.c
2643
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.c
2778
int kvmppc_xive_xics_hcall(struct kvm_vcpu *vcpu, u32 req)
arch/powerpc/kvm/book3s_xive.c
2806
int kvmppc_xive_debug_show_queues(struct seq_file *m, struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
287
static unsigned long xive_vm_h_xirr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
2883
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.c
336
static unsigned long xive_vm_h_ipoll(struct kvm_vcpu *vcpu, unsigned long server)
arch/powerpc/kvm/book3s_xive.c
456
static int xive_vm_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr)
arch/powerpc/kvm/book3s_xive.c
515
static int xive_vm_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr)
arch/powerpc/kvm/book3s_xive.c
620
static int xive_vm_h_ipi(struct kvm_vcpu *vcpu, unsigned long server,
arch/powerpc/kvm/book3s_xive.c
663
static bool kvmppc_xive_vcpu_has_save_restore(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
671
bool kvmppc_xive_check_save_restore(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
686
void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
756
void kvmppc_xive_pull_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
784
bool kvmppc_xive_rearm_escalation(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive.c
842
struct kvm_vcpu *vcpu = data;
arch/powerpc/kvm/book3s_xive.c
866
int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio,
arch/powerpc/kvm/book3s_xive.c
935
static int xive_provision_queue(struct kvm_vcpu *vcpu, u8 prio)
arch/powerpc/kvm/book3s_xive.c
974
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.h
154
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive.h
199
static inline struct kvm_vcpu *kvmppc_xive_find_server(struct kvm *kvm, u32 nr)
arch/powerpc/kvm/book3s_xive.h
201
struct kvm_vcpu *vcpu = NULL;
arch/powerpc/kvm/book3s_xive.h
242
struct kvm_vcpu *vcpu = NULL;
arch/powerpc/kvm/book3s_xive.h
291
void kvmppc_xive_disable_vcpu_interrupts(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_xive.h
292
int kvmppc_xive_debug_show_queues(struct seq_file *m, struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_xive.h
299
int kvmppc_xive_attach_escalation(struct kvm_vcpu *vcpu, u8 prio,
arch/powerpc/kvm/book3s_xive.h
302
void xive_cleanup_single_escalation(struct kvm_vcpu *vcpu, int irq);
arch/powerpc/kvm/book3s_xive.h
305
bool kvmppc_xive_check_save_restore(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/book3s_xive_native.c
1018
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive_native.c
1136
int kvmppc_xive_native_get_vp(struct kvm_vcpu *vcpu, union kvmppc_one_reg *val)
arch/powerpc/kvm/book3s_xive_native.c
1174
int kvmppc_xive_native_set_vp(struct kvm_vcpu *vcpu, union kvmppc_one_reg *val)
arch/powerpc/kvm/book3s_xive_native.c
1215
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive_native.c
124
struct kvm_vcpu *vcpu, u32 server_num)
arch/powerpc/kvm/book3s_xive_native.c
44
static void kvmppc_xive_native_cleanup_queue(struct kvm_vcpu *vcpu, int prio)
arch/powerpc/kvm/book3s_xive_native.c
563
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive_native.c
711
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive_native.c
74
void kvmppc_xive_native_cleanup_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive_native.c
808
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/book3s_xive_native.c
891
static int kvmppc_xive_native_vcpu_eq_sync(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/book3s_xive_native.c
917
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/booke.c
1011
int kvmppc_handle_exit(struct kvm_vcpu *vcpu, unsigned int exit_nr)
arch/powerpc/kvm/booke.c
113
void kvmppc_vcpu_disable_spe(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
123
static void kvmppc_vcpu_enable_spe(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
133
static void kvmppc_vcpu_sync_spe(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1416
static void kvmppc_set_tsr(struct kvm_vcpu *vcpu, u32 new_tsr)
arch/powerpc/kvm/booke.c
1428
int kvmppc_subarch_vcpu_init(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
143
static void kvmppc_vcpu_sync_spe(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1442
void kvmppc_subarch_vcpu_uninit(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1447
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/powerpc/kvm/booke.c
1478
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/powerpc/kvm/booke.c
1509
static void get_sregs_base(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
1528
static int set_sregs_base(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
1553
static void get_sregs_arch206(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
1565
static int set_sregs_arch206(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
157
static inline void kvmppc_load_guest_fp(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1582
int kvmppc_get_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/powerpc/kvm/booke.c
1605
int kvmppc_set_sregs_ivor(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/powerpc/kvm/booke.c
1630
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
1647
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
1671
int kvmppc_get_one_reg(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/booke.c
1727
int kvmppc_set_one_reg(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/booke.c
174
static inline void kvmppc_save_guest_fp(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1796
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/powerpc/kvm/booke.c
1801
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/powerpc/kvm/booke.c
1806
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
183
static void kvmppc_vcpu_sync_fpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1850
void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr)
arch/powerpc/kvm/booke.c
1862
void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr)
arch/powerpc/kvm/booke.c
1869
void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits)
arch/powerpc/kvm/booke.c
1877
void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits)
arch/powerpc/kvm/booke.c
1891
void kvmppc_decrementer_func(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1956
static void kvm_guest_protect_msr(struct kvm_vcpu *vcpu, ulong prot_bitmap,
arch/powerpc/kvm/booke.c
198
static inline void kvmppc_load_guest_altivec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
1980
int kvmppc_xlate(struct kvm_vcpu *vcpu, ulong eaddr, enum xlate_instdata xlid,
arch/powerpc/kvm/booke.c
2031
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
2109
void kvmppc_booke_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/booke.c
2112
current->thread.kvm_vcpu = vcpu;
arch/powerpc/kvm/booke.c
2115
void kvmppc_booke_vcpu_put(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
2117
current->thread.kvm_vcpu = NULL;
arch/powerpc/kvm/booke.c
2129
int kvmppc_core_vcpu_create(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
2165
void kvmppc_core_vcpu_free(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
217
static inline void kvmppc_save_guest_altivec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
2175
void kvmppc_core_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/booke.c
2180
void kvmppc_core_vcpu_put(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
228
static void kvmppc_vcpu_sync_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
255
void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr)
arch/powerpc/kvm/booke.c
271
static void kvmppc_booke_queue_irqprio(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
278
void kvmppc_core_queue_dtlb_miss(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
286
void kvmppc_core_queue_data_storage(struct kvm_vcpu *vcpu, ulong srr1_flags,
arch/powerpc/kvm/booke.c
295
void kvmppc_core_queue_itlb_miss(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
300
void kvmppc_core_queue_inst_storage(struct kvm_vcpu *vcpu, ulong esr_flags)
arch/powerpc/kvm/booke.c
306
static void kvmppc_core_queue_alignment(struct kvm_vcpu *vcpu, ulong dear_flags,
arch/powerpc/kvm/booke.c
314
void kvmppc_core_queue_program(struct kvm_vcpu *vcpu, ulong esr_flags)
arch/powerpc/kvm/booke.c
320
void kvmppc_core_queue_fpunavail(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/booke.c
327
void kvmppc_core_queue_vec_unavail(struct kvm_vcpu *vcpu, ulong srr1_flags)
arch/powerpc/kvm/booke.c
334
void kvmppc_core_queue_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
339
int kvmppc_core_pending_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
344
void kvmppc_core_dequeue_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
349
void kvmppc_core_queue_external(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
360
void kvmppc_core_dequeue_external(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
366
static void kvmppc_core_queue_watchdog(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
371
static void kvmppc_core_dequeue_watchdog(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
376
void kvmppc_core_queue_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
381
void kvmppc_core_dequeue_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
386
static void set_guest_srr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1)
arch/powerpc/kvm/booke.c
392
static void set_guest_csrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1)
arch/powerpc/kvm/booke.c
398
static void set_guest_dsrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1)
arch/powerpc/kvm/booke.c
408
static void set_guest_mcsrr(struct kvm_vcpu *vcpu, unsigned long srr0, u32 srr1)
arch/powerpc/kvm/booke.c
415
static int kvmppc_booke_irqprio_deliver(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
578
static unsigned long watchdog_next_timeout(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
604
static void arm_next_watchdog(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
631
struct kvm_vcpu *vcpu = timer_container_of(vcpu, t, arch.wdt_timer);
arch/powerpc/kvm/booke.c
677
static void update_timer_ints(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
690
static void kvmppc_core_check_exceptions(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
710
int kvmppc_core_prepare_to_enter(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
734
int kvmppc_core_check_requests(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
760
int kvmppc_vcpu_run(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
830
static int emulation_exit(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
864
static int kvmppc_handle_debug(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
90
void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke.c
937
static void kvmppc_restart_interrupt(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.c
984
static int kvmppc_resume_inst_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.h
101
extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/booke.h
103
extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/booke.h
105
extern int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/booke.h
113
int kvmppc_handle_exit(struct kvm_vcpu *vcpu, unsigned int exit_nr);
arch/powerpc/kvm/booke.h
65
void kvmppc_set_msr(struct kvm_vcpu *vcpu, u32 new_msr);
arch/powerpc/kvm/booke.h
66
void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr);
arch/powerpc/kvm/booke.h
68
void kvmppc_set_epcr(struct kvm_vcpu *vcpu, u32 new_epcr);
arch/powerpc/kvm/booke.h
69
void kvmppc_set_tcr(struct kvm_vcpu *vcpu, u32 new_tcr);
arch/powerpc/kvm/booke.h
70
void kvmppc_set_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
arch/powerpc/kvm/booke.h
71
void kvmppc_clr_tsr_bits(struct kvm_vcpu *vcpu, u32 tsr_bits);
arch/powerpc/kvm/booke.h
73
int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.h
75
int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val);
arch/powerpc/kvm/booke.h
76
int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val);
arch/powerpc/kvm/booke.h
79
void kvmppc_load_guest_spe(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/booke.h
80
void kvmppc_save_guest_spe(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/booke.h
83
void kvmppc_vcpu_disable_spe(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/booke.h
85
void kvmppc_booke_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
arch/powerpc/kvm/booke.h
86
void kvmppc_booke_vcpu_put(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/booke.h
95
void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type);
arch/powerpc/kvm/booke.h
97
extern int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/booke.h
99
extern int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn,
arch/powerpc/kvm/booke_emulate.c
120
int kvmppc_booke_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
arch/powerpc/kvm/booke_emulate.c
24
static void kvmppc_emul_rfi(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke_emulate.c
30
static void kvmppc_emul_rfdi(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke_emulate.c
36
static void kvmppc_emul_rfci(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/booke_emulate.c
379
int kvmppc_booke_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
arch/powerpc/kvm/booke_emulate.c
42
int kvmppc_booke_emulate_op(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.c
214
unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.c
221
void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid)
arch/powerpc/kvm/e500.c
293
void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr)
arch/powerpc/kvm/e500.c
299
static void kvmppc_core_vcpu_load_e500(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/e500.c
307
static void kvmppc_core_vcpu_put_e500(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.c
346
int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.c
361
static int kvmppc_core_get_sregs_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.c
386
static int kvmppc_core_set_sregs_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.c
422
static int kvmppc_get_one_reg_e500(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500.c
429
static int kvmppc_set_one_reg_e500(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500.c
436
static int kvmppc_core_vcpu_create_e500(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.c
466
static void kvmppc_core_vcpu_free_e500(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
125
int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/e500.h
126
int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/e500.h
127
int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea);
arch/powerpc/kvm/e500.h
128
int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea);
arch/powerpc/kvm/e500.h
129
int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea);
arch/powerpc/kvm/e500.h
133
void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/kvm/e500.h
134
int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
arch/powerpc/kvm/e500.h
136
int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500.h
138
int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500.h
206
static inline unsigned int get_cur_pid(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
211
static inline unsigned int get_cur_as(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
216
static inline unsigned int get_cur_pr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
221
static inline unsigned int get_cur_spid(const struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
226
static inline unsigned int get_cur_sas(const struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
231
static inline unsigned int get_tlb_tlbsel(const struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
240
static inline unsigned int get_tlb_nv_bit(const struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
245
static inline unsigned int get_tlb_esel_bit(const struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
250
static inline int tlbe_is_host_safe(const struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.h
303
static inline int get_lpid(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
308
unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.h
311
static inline unsigned int get_tlbmiss_tid(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500.h
323
static inline bool has_feature(const struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500.h
56
struct kvm_vcpu vcpu;
arch/powerpc/kvm/e500.h
98
static inline struct kvmppc_vcpu_e500 *to_e500(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_emulate.c
106
static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_emulate.c
115
static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
arch/powerpc/kvm/e500_emulate.c
128
int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_emulate.c
204
int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
arch/powerpc/kvm/e500_emulate.c
321
int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
arch/powerpc/kvm/e500_emulate.c
51
static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
arch/powerpc/kvm/e500_emulate.c
63
static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
arch/powerpc/kvm/e500_emulate.c
69
struct kvm_vcpu *cvcpu;
arch/powerpc/kvm/e500_emulate.c
86
static int kvmppc_e500_emul_ehpriv(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
126
static inline void kvmppc_e500_deliver_tlb_miss(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
193
static void kvmppc_set_tlb1map_range(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
245
int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
307
int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
322
int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu.c
341
int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea)
arch/powerpc/kvm/e500_mmu.c
390
int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu.c
453
static int kvmppc_e500_tlb_search(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
469
int kvmppc_core_vcpu_translate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
495
int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/e500_mmu.c
502
int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
arch/powerpc/kvm/e500_mmu.c
509
void kvmppc_mmu_itlb_miss(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu.c
516
void kvmppc_mmu_dtlb_miss(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu.c
523
gpa_t kvmppc_mmu_xlate(struct kvm_vcpu *vcpu, unsigned int index,
arch/powerpc/kvm/e500_mmu.c
567
void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/powerpc/kvm/e500_mmu.c
583
int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/powerpc/kvm/e500_mmu.c
597
int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500_mmu.c
64
static unsigned int get_tlb_esel(struct kvm_vcpu *vcpu, int tlbsel)
arch/powerpc/kvm/e500_mmu.c
650
int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500_mmu.c
717
static int vcpu_mmu_geometry_update(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
731
int kvm_vcpu_ioctl_config_tlb(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
856
int kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
866
static int vcpu_mmu_init(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu.c
900
struct kvm_vcpu *vcpu = &vcpu_e500->vcpu;
arch/powerpc/kvm/e500_mmu_host.c
161
void kvmppc_map_magic(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu_host.c
292
void kvmppc_core_flush_tlb(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500_mmu_host.c
302
struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu_host.c
545
void kvmppc_mmu_map(struct kvm_vcpu *vcpu, u64 eaddr, gpa_t gpaddr,
arch/powerpc/kvm/e500_mmu_host.c
584
int kvmppc_load_last_inst(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500_mmu_host.c
674
int kvmppc_load_last_inst(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500mc.c
105
void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid)
arch/powerpc/kvm/e500mc.c
110
void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr)
arch/powerpc/kvm/e500mc.c
115
static DEFINE_PER_CPU(struct kvm_vcpu *[KVMPPC_NR_LPIDS], last_vcpu_of_lpid);
arch/powerpc/kvm/e500mc.c
117
static void kvmppc_core_vcpu_load_e500mc(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/e500mc.c
154
static void kvmppc_core_vcpu_put_e500mc(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500mc.c
200
int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500mc.c
219
static int kvmppc_core_get_sregs_e500mc(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500mc.c
243
static int kvmppc_core_set_sregs_e500mc(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/e500mc.c
277
static int kvmppc_get_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500mc.c
28
void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type)
arch/powerpc/kvm/e500mc.c
293
static int kvmppc_set_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id,
arch/powerpc/kvm/e500mc.c
309
static int kvmppc_core_vcpu_create_e500mc(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/e500mc.c
337
static void kvmppc_core_vcpu_free_e500mc(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate.c
128
static int kvmppc_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, int rt)
arch/powerpc/kvm/emulate.c
194
int kvmppc_emulate_instruction(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate.c
26
void kvmppc_emulate_dec(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate.c
63
u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb)
arch/powerpc/kvm/emulate.c
75
static int kvmppc_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, int rs)
arch/powerpc/kvm/emulate_loadstore.c
28
static bool kvmppc_check_fp_disabled(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate_loadstore.c
40
static bool kvmppc_check_vsx_disabled(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate_loadstore.c
52
static bool kvmppc_check_altivec_disabled(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/emulate_loadstore.c
72
int kvmppc_emulate_loadstore(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/mpic.c
116
struct kvm_vcpu *vcpu = current->thread.kvm_vcpu;
arch/powerpc/kvm/mpic.c
1176
void kvmppc_mpic_set_epr(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/mpic.c
1371
static int kvm_mpic_read(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/mpic.c
1413
static int kvm_mpic_write(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/mpic.c
1732
int kvmppc_mpic_connect_vcpu(struct kvm_device *dev, struct kvm_vcpu *vcpu,
arch/powerpc/kvm/mpic.c
1777
void kvmppc_mpic_disconnect_vcpu(struct openpic *opp, struct kvm_vcpu *vcpu)
arch/powerpc/kvm/mpic.c
178
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/powerpc.c
1007
static inline int kvmppc_get_vmx_dword_offset(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1013
static inline int kvmppc_get_vmx_word_offset(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1019
static inline int kvmppc_get_vmx_hword_offset(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1025
static inline int kvmppc_get_vmx_byte_offset(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1032
static inline void kvmppc_set_vmx_dword(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1048
static inline void kvmppc_set_vmx_word(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1064
static inline void kvmppc_set_vmx_hword(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1080
static inline void kvmppc_set_vmx_byte(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1127
static void kvmppc_complete_mmio_load(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1237
static int __kvmppc_handle_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1281
int kvmppc_handle_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1290
int kvmppc_handle_loads(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1298
int kvmppc_handle_vsx_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1324
int kvmppc_handle_store(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1385
static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val)
arch/powerpc/kvm/powerpc.c
142
static void kvmppc_swab_shared(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1439
int kvmppc_handle_vsx_store(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1470
static int kvmppc_emulate_mmio_vsx_loadstore(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1506
int kvmppc_handle_vmx_load(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
1529
static int kvmppc_get_vmx_dword(struct kvm_vcpu *vcpu, int index, u64 *val)
arch/powerpc/kvm/powerpc.c
1547
static int kvmppc_get_vmx_word(struct kvm_vcpu *vcpu, int index, u64 *val)
arch/powerpc/kvm/powerpc.c
1565
static int kvmppc_get_vmx_hword(struct kvm_vcpu *vcpu, int index, u64 *val)
arch/powerpc/kvm/powerpc.c
1583
static int kvmppc_get_vmx_byte(struct kvm_vcpu *vcpu, int index, u64 *val)
arch/powerpc/kvm/powerpc.c
1601
int kvmppc_handle_vmx_store(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
162
int kvmppc_kvm_pv(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1649
static int kvmppc_emulate_mmio_vmx_loadstore(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1684
int kvm_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg)
arch/powerpc/kvm/powerpc.c
1732
int kvm_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu, struct kvm_one_reg *reg)
arch/powerpc/kvm/powerpc.c
1781
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
1865
int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, struct kvm_interrupt *irq)
arch/powerpc/kvm/powerpc.c
1879
static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
2013
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
2019
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
2028
struct kvm_vcpu *vcpu = filp->private_data;
arch/powerpc/kvm/powerpc.c
2043
struct kvm_vcpu *vcpu = filp->private_data;
arch/powerpc/kvm/powerpc.c
2094
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/powerpc/kvm/powerpc.c
2515
void kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry)
arch/powerpc/kvm/powerpc.c
253
int kvmppc_sanity_check(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
282
int kvmppc_emulate_mmio(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
348
int kvmppc_st(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
arch/powerpc/kvm/powerpc.c
391
int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size, void *ptr,
arch/powerpc/kvm/powerpc.c
50
int kvm_arch_vcpu_runnable(struct kvm_vcpu *v)
arch/powerpc/kvm/powerpc.c
55
bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
60
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
65
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
748
struct kvm_vcpu *vcpu;
arch/powerpc/kvm/powerpc.c
750
vcpu = container_of(timer, struct kvm_vcpu, arch.dec_timer);
arch/powerpc/kvm/powerpc.c
756
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
783
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
787
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
79
int kvmppc_prepare_to_enter(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
812
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
817
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/powerpc/kvm/powerpc.c
832
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/powerpc.c
908
static inline void kvmppc_set_vsr_dword(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
927
static inline void kvmppc_set_vsr_dword_dump(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
944
static inline void kvmppc_set_vsr_word_dump(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
964
static inline void kvmppc_set_vsr_word(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/powerpc.c
990
static inline int kvmppc_get_vmx_offset_generic(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/timing.c
131
struct kvm_vcpu *vcpu = m->private;
arch/powerpc/kvm/timing.c
179
struct kvm_vcpu *vcpu = seqf->private;
arch/powerpc/kvm/timing.c
207
int kvmppc_create_vcpu_debugfs_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/timing.c
22
void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/timing.c
44
static void add_exit_timing(struct kvm_vcpu *vcpu, u64 duration, int type)
arch/powerpc/kvm/timing.c
82
void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu)
arch/powerpc/kvm/timing.h
15
void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/timing.h
16
void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu);
arch/powerpc/kvm/timing.h
17
int kvmppc_create_vcpu_debugfs_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/timing.h
20
static inline void kvmppc_set_exit_type(struct kvm_vcpu *vcpu, int type)
arch/powerpc/kvm/timing.h
27
static inline void kvmppc_init_timing_stats(struct kvm_vcpu *vcpu) {}
arch/powerpc/kvm/timing.h
28
static inline void kvmppc_update_timing_stats(struct kvm_vcpu *vcpu) {}
arch/powerpc/kvm/timing.h
29
static inline int kvmppc_create_vcpu_debugfs_e500(struct kvm_vcpu *vcpu,
arch/powerpc/kvm/timing.h
34
static inline void kvmppc_set_exit_type(struct kvm_vcpu *vcpu, int type) {}
arch/powerpc/kvm/timing.h
38
static inline void kvmppc_account_exit_stat(struct kvm_vcpu *vcpu, int type)
arch/powerpc/kvm/timing.h
89
static inline void kvmppc_account_exit(struct kvm_vcpu *vcpu, int type)
arch/powerpc/kvm/trace.h
101
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_booke.h
180
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int priority),
arch/powerpc/kvm/trace_booke.h
39
TP_PROTO(unsigned int exit_nr, struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
242
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
266
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
293
TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep,
arch/powerpc/kvm/trace_hv.h
329
TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, long ret),
arch/powerpc/kvm/trace_hv.h
353
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
382
TP_PROTO(struct kvm_vcpu *vcpu, int ret),
arch/powerpc/kvm/trace_hv.h
431
TP_PROTO(struct kvm_vcpu *vcpu, int where),
arch/powerpc/kvm/trace_hv.h
477
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
495
TP_PROTO(struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_hv.h
518
TP_PROTO(struct kvm_vcpu *vcpu, u64 l1_to_l2_cs, u64 l2_to_l1_cs, u64 l2_runtime),
arch/powerpc/kvm/trace_pr.h
122
TP_PROTO(const char *type, struct kvm_vcpu *vcpu, unsigned long long p1,
arch/powerpc/kvm/trace_pr.h
13
TP_PROTO(int r, struct kvm_vcpu *vcpu),
arch/powerpc/kvm/trace_pr.h
218
TP_PROTO(unsigned int exit_nr, struct kvm_vcpu *vcpu),
arch/riscv/include/asm/kvm_aia.h
103
void kvm_riscv_vcpu_aia_imsic_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
104
int kvm_riscv_vcpu_aia_imsic_inject(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_aia.h
106
int kvm_riscv_vcpu_aia_imsic_init(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
107
void kvm_riscv_vcpu_aia_imsic_cleanup(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
117
void kvm_riscv_vcpu_aia_flush_interrupts(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
118
void kvm_riscv_vcpu_aia_sync_interrupts(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
120
static inline void kvm_riscv_vcpu_aia_flush_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_aia.h
123
static inline void kvm_riscv_vcpu_aia_sync_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_aia.h
127
bool kvm_riscv_vcpu_aia_has_interrupts(struct kvm_vcpu *vcpu, u64 mask);
arch/riscv/include/asm/kvm_aia.h
129
void kvm_riscv_vcpu_aia_update_hvip(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
130
void kvm_riscv_vcpu_aia_load(struct kvm_vcpu *vcpu, int cpu);
arch/riscv/include/asm/kvm_aia.h
131
void kvm_riscv_vcpu_aia_put(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
132
int kvm_riscv_vcpu_aia_get_csr(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_aia.h
135
int kvm_riscv_vcpu_aia_set_csr(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_aia.h
139
int kvm_riscv_vcpu_aia_rmw_topei(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_aia.h
144
int kvm_riscv_vcpu_aia_rmw_ireg(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/include/asm/kvm_aia.h
151
int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
152
void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
153
void kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
154
void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
164
int kvm_riscv_aia_alloc_hgei(int cpu, struct kvm_vcpu *owner,
arch/riscv/include/asm/kvm_aia.h
90
bool kvm_riscv_vcpu_aia_imsic_has_interrupt(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
91
void kvm_riscv_vcpu_aia_imsic_load(struct kvm_vcpu *vcpu, int cpu);
arch/riscv/include/asm/kvm_aia.h
92
void kvm_riscv_vcpu_aia_imsic_put(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
93
void kvm_riscv_vcpu_aia_imsic_release(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
94
int kvm_riscv_vcpu_aia_imsic_update(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_aia.h
97
int kvm_riscv_vcpu_aia_imsic_rmw(struct kvm_vcpu *vcpu, unsigned long isel,
arch/riscv/include/asm/kvm_host.h
288
static inline bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_host.h
293
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
arch/riscv/include/asm/kvm_host.h
294
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
arch/riscv/include/asm/kvm_host.h
300
unsigned long kvm_riscv_vcpu_unpriv_read(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_host.h
304
void kvm_riscv_vcpu_trap_redirect(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_host.h
306
int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/include/asm/kvm_host.h
311
void kvm_riscv_vcpu_setup_isa(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
312
unsigned long kvm_riscv_vcpu_num_regs(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
313
int kvm_riscv_vcpu_copy_reg_indices(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_host.h
315
int kvm_riscv_vcpu_get_reg(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_host.h
317
int kvm_riscv_vcpu_set_reg(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_host.h
320
int kvm_riscv_vcpu_set_interrupt(struct kvm_vcpu *vcpu, unsigned int irq);
arch/riscv/include/asm/kvm_host.h
321
int kvm_riscv_vcpu_unset_interrupt(struct kvm_vcpu *vcpu, unsigned int irq);
arch/riscv/include/asm/kvm_host.h
322
void kvm_riscv_vcpu_flush_interrupts(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
323
void kvm_riscv_vcpu_sync_interrupts(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
324
bool kvm_riscv_vcpu_has_interrupts(struct kvm_vcpu *vcpu, u64 mask);
arch/riscv/include/asm/kvm_host.h
325
void __kvm_riscv_vcpu_power_off(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
326
void kvm_riscv_vcpu_power_off(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
327
void __kvm_riscv_vcpu_power_on(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
328
void kvm_riscv_vcpu_power_on(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
329
bool kvm_riscv_vcpu_stopped(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_host.h
331
void kvm_riscv_vcpu_record_steal_time(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_mmu.h
14
int kvm_riscv_mmu_map(struct kvm_vcpu *vcpu, struct kvm_memory_slot *memslot,
arch/riscv/include/asm/kvm_mmu.h
19
void kvm_riscv_mmu_update_hgatp(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_tlb.h
52
void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_tlb.h
54
void kvm_riscv_tlb_flush_process(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_tlb.h
56
void kvm_riscv_fence_i_process(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_tlb.h
57
void kvm_riscv_hfence_vvma_all_process(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_tlb.h
58
void kvm_riscv_hfence_process(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_fp.h
23
void kvm_riscv_vcpu_fp_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_fp.h
31
static inline void kvm_riscv_vcpu_fp_reset(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_vcpu_fp.h
52
int kvm_riscv_vcpu_get_reg_fp(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_fp.h
55
int kvm_riscv_vcpu_set_reg_fp(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_insn.h
35
void kvm_riscv_vcpu_wfi(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_insn.h
36
int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/riscv/include/asm/kvm_vcpu_insn.h
37
int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/include/asm/kvm_vcpu_insn.h
40
int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/include/asm/kvm_vcpu_insn.h
43
int kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/include/asm/kvm_vcpu_insn.h
46
int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/riscv/include/asm/kvm_vcpu_insn.h
9
struct kvm_vcpu;
arch/riscv/include/asm/kvm_vcpu_pmu.h
101
int kvm_riscv_vcpu_pmu_event_info(struct kvm_vcpu *vcpu, unsigned long saddr_low,
arch/riscv/include/asm/kvm_vcpu_pmu.h
104
void kvm_riscv_vcpu_pmu_deinit(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_pmu.h
105
void kvm_riscv_vcpu_pmu_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_pmu.h
111
static inline int kvm_riscv_vcpu_pmu_read_legacy(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/include/asm/kvm_vcpu_pmu.h
126
static inline void kvm_riscv_vcpu_pmu_init(struct kvm_vcpu *vcpu) {}
arch/riscv/include/asm/kvm_vcpu_pmu.h
127
static inline int kvm_riscv_vcpu_pmu_incr_fw(struct kvm_vcpu *vcpu, unsigned long fid)
arch/riscv/include/asm/kvm_vcpu_pmu.h
132
static inline void kvm_riscv_vcpu_pmu_deinit(struct kvm_vcpu *vcpu) {}
arch/riscv/include/asm/kvm_vcpu_pmu.h
133
static inline void kvm_riscv_vcpu_pmu_reset(struct kvm_vcpu *vcpu) {}
arch/riscv/include/asm/kvm_vcpu_pmu.h
40
struct kvm_vcpu *vcpu;
arch/riscv/include/asm/kvm_vcpu_pmu.h
64
#define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu_context))
arch/riscv/include/asm/kvm_vcpu_pmu.h
75
int kvm_riscv_vcpu_pmu_incr_fw(struct kvm_vcpu *vcpu, unsigned long fid);
arch/riscv/include/asm/kvm_vcpu_pmu.h
76
int kvm_riscv_vcpu_pmu_read_hpm(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/include/asm/kvm_vcpu_pmu.h
80
int kvm_riscv_vcpu_pmu_num_ctrs(struct kvm_vcpu *vcpu, struct kvm_vcpu_sbi_return *retdata);
arch/riscv/include/asm/kvm_vcpu_pmu.h
81
int kvm_riscv_vcpu_pmu_ctr_info(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/include/asm/kvm_vcpu_pmu.h
83
int kvm_riscv_vcpu_pmu_ctr_start(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/include/asm/kvm_vcpu_pmu.h
86
int kvm_riscv_vcpu_pmu_ctr_stop(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/include/asm/kvm_vcpu_pmu.h
89
int kvm_riscv_vcpu_pmu_ctr_cfg_match(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/include/asm/kvm_vcpu_pmu.h
93
int kvm_riscv_vcpu_pmu_fw_ctr_read(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/include/asm/kvm_vcpu_pmu.h
95
int kvm_riscv_vcpu_pmu_fw_ctr_read_hi(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/include/asm/kvm_vcpu_pmu.h
97
void kvm_riscv_vcpu_pmu_init(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_pmu.h
98
int kvm_riscv_vcpu_pmu_snapshot_set_shmem(struct kvm_vcpu *vcpu, unsigned long saddr_low,
arch/riscv/include/asm/kvm_vcpu_sbi.h
47
int (*handler)(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/include/asm/kvm_vcpu_sbi.h
51
unsigned long (*probe)(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
58
int (*init)(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
59
void (*deinit)(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
61
void (*reset)(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
64
unsigned long (*get_state_reg_count)(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
65
int (*get_state_reg_id)(struct kvm_vcpu *vcpu, int index, u64 *reg_id);
arch/riscv/include/asm/kvm_vcpu_sbi.h
66
int (*get_state_reg)(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/include/asm/kvm_vcpu_sbi.h
68
int (*set_state_reg)(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/include/asm/kvm_vcpu_sbi.h
72
int kvm_riscv_vcpu_sbi_forward_handler(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_sbi.h
75
void kvm_riscv_vcpu_sbi_system_reset(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_sbi.h
78
void kvm_riscv_vcpu_sbi_request_reset(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_sbi.h
80
void kvm_riscv_vcpu_sbi_load_reset_state(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
81
int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/riscv/include/asm/kvm_vcpu_sbi.h
82
int kvm_riscv_vcpu_reg_indices_sbi_ext(struct kvm_vcpu *vcpu, u64 __user *uindices);
arch/riscv/include/asm/kvm_vcpu_sbi.h
83
int kvm_riscv_vcpu_set_reg_sbi_ext(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_sbi.h
85
int kvm_riscv_vcpu_get_reg_sbi_ext(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_sbi.h
87
int kvm_riscv_vcpu_reg_indices_sbi(struct kvm_vcpu *vcpu, u64 __user *uindices);
arch/riscv/include/asm/kvm_vcpu_sbi.h
88
int kvm_riscv_vcpu_set_reg_sbi(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
arch/riscv/include/asm/kvm_vcpu_sbi.h
89
int kvm_riscv_vcpu_get_reg_sbi(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
arch/riscv/include/asm/kvm_vcpu_sbi.h
91
struct kvm_vcpu *vcpu, unsigned long extid);
arch/riscv/include/asm/kvm_vcpu_sbi.h
92
int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run);
arch/riscv/include/asm/kvm_vcpu_sbi.h
93
void kvm_riscv_vcpu_sbi_init(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
94
void kvm_riscv_vcpu_sbi_deinit(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_sbi.h
95
void kvm_riscv_vcpu_sbi_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
35
int (*timer_next_event)(struct kvm_vcpu *vcpu, u64 ncycles);
arch/riscv/include/asm/kvm_vcpu_timer.h
38
int kvm_riscv_vcpu_timer_next_event(struct kvm_vcpu *vcpu, u64 ncycles);
arch/riscv/include/asm/kvm_vcpu_timer.h
39
int kvm_riscv_vcpu_get_reg_timer(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_timer.h
41
int kvm_riscv_vcpu_set_reg_timer(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_timer.h
43
int kvm_riscv_vcpu_timer_init(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
44
int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
45
int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
46
void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
48
void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
49
void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_timer.h
50
bool kvm_riscv_vcpu_timer_pending(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_vector.h
29
void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_vector.h
36
int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_vector.h
37
void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu);
arch/riscv/include/asm/kvm_vcpu_vector.h
42
static inline void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_vcpu_vector.h
64
static inline int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_vcpu_vector.h
69
static inline void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu)
arch/riscv/include/asm/kvm_vcpu_vector.h
74
int kvm_riscv_vcpu_get_reg_vector(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vcpu_vector.h
76
int kvm_riscv_vcpu_set_reg_vector(struct kvm_vcpu *vcpu,
arch/riscv/include/asm/kvm_vmid.h
24
void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu);
arch/riscv/kvm/aia.c
111
void kvm_riscv_vcpu_aia_load(struct kvm_vcpu *vcpu, int cpu)
arch/riscv/kvm/aia.c
146
void kvm_riscv_vcpu_aia_put(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia.c
181
int kvm_riscv_vcpu_aia_get_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/aia.c
202
int kvm_riscv_vcpu_aia_set_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/aia.c
228
int kvm_riscv_vcpu_aia_rmw_topei(struct kvm_vcpu *vcpu,
arch/riscv/kvm/aia.c
25
struct kvm_vcpu *owners[BITS_PER_LONG];
arch/riscv/kvm/aia.c
262
static u8 aia_get_iprio8(struct kvm_vcpu *vcpu, unsigned int irq)
arch/riscv/kvm/aia.c
295
static void aia_set_iprio8(struct kvm_vcpu *vcpu, unsigned int irq, u8 prio)
arch/riscv/kvm/aia.c
351
static int aia_rmw_iprio(struct kvm_vcpu *vcpu, unsigned int isel,
arch/riscv/kvm/aia.c
388
int kvm_riscv_vcpu_aia_rmw_ireg(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/kvm/aia.c
411
int kvm_riscv_aia_alloc_hgei(int cpu, struct kvm_vcpu *owner,
arch/riscv/kvm/aia.c
49
void kvm_riscv_vcpu_aia_flush_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia.c
574
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia.c
66
void kvm_riscv_vcpu_aia_sync_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia.c
75
bool kvm_riscv_vcpu_aia_has_interrupts(struct kvm_vcpu *vcpu, u64 mask)
arch/riscv/kvm/aia.c
98
void kvm_riscv_vcpu_aia_update_hvip(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_aplic.c
443
static int aplic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/riscv/kvm/aia_aplic.c
515
static int aplic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/riscv/kvm/aia_device.c
167
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_device.c
21
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_device.c
230
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_device.c
496
int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_device.c
506
void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_device.c
522
void kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_device.c
541
void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_device.c
555
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_device.c
576
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_imsic.c
1014
void kvm_riscv_vcpu_aia_imsic_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
1026
int kvm_riscv_vcpu_aia_imsic_inject(struct kvm_vcpu *vcpu,
arch/riscv/kvm/aia_imsic.c
1058
static int imsic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/riscv/kvm/aia_imsic.c
1069
static int imsic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/riscv/kvm/aia_imsic.c
1090
int kvm_riscv_vcpu_aia_imsic_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
1147
void kvm_riscv_vcpu_aia_imsic_cleanup(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
614
static void imsic_swfile_extirq_update(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
637
static void imsic_swfile_read(struct kvm_vcpu *vcpu, bool clear,
arch/riscv/kvm/aia_imsic.c
656
static void imsic_swfile_update(struct kvm_vcpu *vcpu,
arch/riscv/kvm/aia_imsic.c
680
bool kvm_riscv_vcpu_aia_imsic_has_interrupt(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
711
void kvm_riscv_vcpu_aia_imsic_load(struct kvm_vcpu *vcpu, int cpu)
arch/riscv/kvm/aia_imsic.c
720
void kvm_riscv_vcpu_aia_imsic_put(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
734
void kvm_riscv_vcpu_aia_imsic_release(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
784
int kvm_riscv_vcpu_aia_imsic_update(struct kvm_vcpu *vcpu)
arch/riscv/kvm/aia_imsic.c
902
int kvm_riscv_vcpu_aia_imsic_rmw(struct kvm_vcpu *vcpu, unsigned long isel,
arch/riscv/kvm/aia_imsic.c
952
struct kvm_vcpu *vcpu;
arch/riscv/kvm/aia_imsic.c
996
struct kvm_vcpu *vcpu;
arch/riscv/kvm/main.c
179
rc = kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/riscv/kvm/mmu.c
448
int kvm_riscv_mmu_map(struct kvm_vcpu *vcpu, struct kvm_memory_slot *memslot,
arch/riscv/kvm/mmu.c
604
void kvm_riscv_mmu_update_hgatp(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
161
void kvm_riscv_local_tlb_sanitize(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
191
void kvm_riscv_fence_i_process(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
197
void kvm_riscv_tlb_flush_process(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
208
void kvm_riscv_hfence_vvma_all_process(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
219
static bool vcpu_hfence_dequeue(struct kvm_vcpu *vcpu,
arch/riscv/kvm/tlb.c
244
static bool vcpu_hfence_enqueue(struct kvm_vcpu *vcpu,
arch/riscv/kvm/tlb.c
268
void kvm_riscv_hfence_process(struct kvm_vcpu *vcpu)
arch/riscv/kvm/tlb.c
334
struct kvm_vcpu *vcpu;
arch/riscv/kvm/trace.h
17
TP_PROTO(struct kvm_vcpu *vcpu),
arch/riscv/kvm/vcpu.c
127
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
178
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
189
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
208
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
213
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
219
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
224
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
230
unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
236
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/riscv/kvm/vcpu.c
244
struct kvm_vcpu *vcpu = filp->private_data;
arch/riscv/kvm/vcpu.c
265
struct kvm_vcpu *vcpu = filp->private_data;
arch/riscv/kvm/vcpu.c
309
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
315
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
321
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/riscv/kvm/vcpu.c
326
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/riscv/kvm/vcpu.c
331
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
337
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/riscv/kvm/vcpu.c
342
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/riscv/kvm/vcpu.c
347
void kvm_riscv_vcpu_flush_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
364
void kvm_riscv_vcpu_sync_interrupts(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
401
int kvm_riscv_vcpu_set_interrupt(struct kvm_vcpu *vcpu, unsigned int irq)
arch/riscv/kvm/vcpu.c
424
int kvm_riscv_vcpu_unset_interrupt(struct kvm_vcpu *vcpu, unsigned int irq)
arch/riscv/kvm/vcpu.c
445
bool kvm_riscv_vcpu_has_interrupts(struct kvm_vcpu *vcpu, u64 mask)
arch/riscv/kvm/vcpu.c
460
void __kvm_riscv_vcpu_power_off(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
467
void kvm_riscv_vcpu_power_off(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
474
void __kvm_riscv_vcpu_power_on(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
480
void kvm_riscv_vcpu_power_on(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
487
bool kvm_riscv_vcpu_stopped(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
492
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
500
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
529
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
54
static void kvm_riscv_vcpu_context_reset(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
543
static void kvm_riscv_vcpu_setup_config(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
578
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/riscv/kvm/vcpu.c
643
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
692
static int kvm_riscv_check_vcpu_requests(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
741
static void kvm_riscv_update_hvip(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
749
static __always_inline void kvm_riscv_vcpu_swap_in_guest_state(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
763
static __always_inline void kvm_riscv_vcpu_swap_in_host_state(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu.c
784
static void noinstr kvm_riscv_vcpu_enter_exit(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu.c
79
static void kvm_riscv_reset_vcpu(struct kvm_vcpu *vcpu, bool kvm_sbi_reset)
arch/riscv/kvm/vcpu.c
864
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_exit.c
139
void kvm_riscv_vcpu_trap_redirect(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_exit.c
15
static int gstage_page_fault(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_exit.c
172
static inline int vcpu_redirect(struct kvm_vcpu *vcpu, struct kvm_cpu_trap *trap)
arch/riscv/kvm/vcpu_exit.c
187
int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_exit.c
63
unsigned long kvm_riscv_vcpu_unpriv_read(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_fp.c
127
int kvm_riscv_vcpu_set_reg_fp(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_fp.c
18
void kvm_riscv_vcpu_fp_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_fp.c
78
int kvm_riscv_vcpu_get_reg_fp(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_insn.c
102
static int seed_csr_rmw(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/kvm/vcpu_insn.c
127
int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/riscv/kvm/vcpu_insn.c
147
static int csr_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn)
arch/riscv/kvm/vcpu_insn.c
27
int (*func)(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn);
arch/riscv/kvm/vcpu_insn.c
272
static int system_opcode_insn(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
30
static int truly_illegal_insn(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
328
int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
373
int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
46
static int truly_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
497
int kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_insn.c
627
int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/riscv/kvm/vcpu_insn.c
67
void kvm_riscv_vcpu_wfi(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_insn.c
76
static int wfi_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn)
arch/riscv/kvm/vcpu_insn.c
83
static int wrs_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, ulong insn)
arch/riscv/kvm/vcpu_insn.c
97
int (*func)(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/kvm/vcpu_onereg.c
1007
static inline unsigned long num_fp_f_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1017
static int copy_fp_f_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1036
static inline unsigned long num_fp_d_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1046
static int copy_fp_d_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1076
static int copy_isa_ext_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1102
static inline unsigned long num_isa_ext_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1107
static unsigned long num_sbi_ext_regs(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1112
static inline unsigned long num_sbi_regs(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1117
static inline unsigned long num_vector_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1126
static int copy_vector_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1171
unsigned long kvm_riscv_vcpu_num_regs(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
1192
int kvm_riscv_vcpu_copy_reg_indices(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1250
int kvm_riscv_vcpu_set_reg(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
1283
int kvm_riscv_vcpu_get_reg(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
263
void kvm_riscv_vcpu_setup_isa(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
275
static int kvm_riscv_vcpu_get_reg_config(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
329
static int kvm_riscv_vcpu_set_reg_config(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
439
static int kvm_riscv_vcpu_get_reg_core(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
475
static int kvm_riscv_vcpu_set_reg_core(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
513
static int kvm_riscv_vcpu_general_get_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
535
static int kvm_riscv_vcpu_general_set_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
560
static inline int kvm_riscv_vcpu_smstateen_set_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
579
static int kvm_riscv_vcpu_smstateen_get_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
598
static int kvm_riscv_vcpu_get_reg_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
637
static int kvm_riscv_vcpu_set_reg_csr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
676
static int riscv_vcpu_get_isa_ext_single(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
694
static int riscv_vcpu_set_isa_ext_single(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
729
static int riscv_vcpu_get_isa_ext_multi(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
752
static int riscv_vcpu_set_isa_ext_multi(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
772
static int kvm_riscv_vcpu_get_reg_isa_ext(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
812
static int kvm_riscv_vcpu_set_reg_isa_ext(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
845
static int copy_config_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_onereg.c
884
static unsigned long num_config_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
913
static inline unsigned long num_csr_regs(const struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_onereg.c
925
static int copy_csr_reg_indices(const struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_pmu.c
206
static int pmu_fw_ctr_read_hi(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/kvm/vcpu_pmu.c
237
static int pmu_ctr_read(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/kvm/vcpu_pmu.c
281
struct kvm_vcpu *vcpu = pmc->vcpu;
arch/riscv/kvm/vcpu_pmu.c
349
int kvm_riscv_vcpu_pmu_incr_fw(struct kvm_vcpu *vcpu, unsigned long fid)
arch/riscv/kvm/vcpu_pmu.c
364
int kvm_riscv_vcpu_pmu_read_hpm(struct kvm_vcpu *vcpu, unsigned int csr_num,
arch/riscv/kvm/vcpu_pmu.c
399
static void kvm_pmu_clear_snapshot_area(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_pmu.c
408
int kvm_riscv_vcpu_pmu_snapshot_set_shmem(struct kvm_vcpu *vcpu, unsigned long saddr_low,
arch/riscv/kvm/vcpu_pmu.c
456
int kvm_riscv_vcpu_pmu_event_info(struct kvm_vcpu *vcpu, unsigned long saddr_low,
arch/riscv/kvm/vcpu_pmu.c
512
int kvm_riscv_vcpu_pmu_num_ctrs(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_pmu.c
522
int kvm_riscv_vcpu_pmu_ctr_info(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/kvm/vcpu_pmu.c
538
int kvm_riscv_vcpu_pmu_ctr_start(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/kvm/vcpu_pmu.c
615
int kvm_riscv_vcpu_pmu_ctr_stop(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/kvm/vcpu_pmu.c
713
int kvm_riscv_vcpu_pmu_ctr_cfg_match(struct kvm_vcpu *vcpu, unsigned long ctr_base,
arch/riscv/kvm/vcpu_pmu.c
795
int kvm_riscv_vcpu_pmu_fw_ctr_read_hi(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/kvm/vcpu_pmu.c
807
int kvm_riscv_vcpu_pmu_fw_ctr_read(struct kvm_vcpu *vcpu, unsigned long cidx,
arch/riscv/kvm/vcpu_pmu.c
819
void kvm_riscv_vcpu_pmu_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_pmu.c
887
void kvm_riscv_vcpu_pmu_deinit(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_pmu.c
908
void kvm_riscv_vcpu_pmu_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi.c
100
riscv_vcpu_get_sbi_ext(struct kvm_vcpu *vcpu, unsigned long idx)
arch/riscv/kvm/vcpu_sbi.c
117
static bool riscv_vcpu_supports_sbi_ext(struct kvm_vcpu *vcpu, int idx)
arch/riscv/kvm/vcpu_sbi.c
127
int kvm_riscv_vcpu_sbi_forward_handler(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
150
void kvm_riscv_vcpu_sbi_system_reset(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
155
struct kvm_vcpu *tmp;
arch/riscv/kvm/vcpu_sbi.c
171
void kvm_riscv_vcpu_sbi_request_reset(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
182
void kvm_riscv_vcpu_sbi_load_reset_state(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi.c
199
int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/riscv/kvm/vcpu_sbi.c
218
static int riscv_vcpu_set_sbi_ext_single(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
239
static int riscv_vcpu_get_sbi_ext_single(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
256
static int riscv_vcpu_set_sbi_ext_multi(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
276
static int riscv_vcpu_get_sbi_ext_multi(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
299
int kvm_riscv_vcpu_reg_indices_sbi_ext(struct kvm_vcpu *vcpu, u64 __user *uindices)
arch/riscv/kvm/vcpu_sbi.c
324
int kvm_riscv_vcpu_set_reg_sbi_ext(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
360
int kvm_riscv_vcpu_get_reg_sbi_ext(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
400
int kvm_riscv_vcpu_reg_indices_sbi(struct kvm_vcpu *vcpu, u64 __user *uindices)
arch/riscv/kvm/vcpu_sbi.c
446
static const struct kvm_vcpu_sbi_extension *kvm_vcpu_sbi_find_ext_withstate(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi.c
467
int kvm_riscv_vcpu_set_reg_sbi(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/riscv/kvm/vcpu_sbi.c
512
int kvm_riscv_vcpu_get_reg_sbi(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg)
arch/riscv/kvm/vcpu_sbi.c
563
struct kvm_vcpu *vcpu, unsigned long extid)
arch/riscv/kvm/vcpu_sbi.c
587
int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run)
arch/riscv/kvm/vcpu_sbi.c
652
void kvm_riscv_vcpu_sbi_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi.c
681
void kvm_riscv_vcpu_sbi_deinit(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi.c
704
void kvm_riscv_vcpu_sbi_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_base.c
16
static int kvm_sbi_ext_base_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_fwft.c
120
static long kvm_sbi_fwft_get_misaligned_delegation(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi_fwft.c
138
static bool kvm_sbi_fwft_pointer_masking_pmlen_supported(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
151
static void kvm_sbi_fwft_reset_pointer_masking_pmlen(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
156
static long kvm_sbi_fwft_set_pointer_masking_pmlen(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi_fwft.c
195
static long kvm_sbi_fwft_get_pointer_masking_pmlen(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi_fwft.c
256
kvm_sbi_fwft_get_config(struct kvm_vcpu *vcpu, enum sbi_fwft_feature_t feature)
arch/riscv/kvm/vcpu_sbi_fwft.c
269
static int kvm_fwft_get_feature(struct kvm_vcpu *vcpu, u32 feature,
arch/riscv/kvm/vcpu_sbi_fwft.c
290
static int kvm_sbi_fwft_set(struct kvm_vcpu *vcpu, u32 feature,
arch/riscv/kvm/vcpu_sbi_fwft.c
311
static int kvm_sbi_fwft_get(struct kvm_vcpu *vcpu, unsigned long feature,
arch/riscv/kvm/vcpu_sbi_fwft.c
324
static int kvm_sbi_ext_fwft_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_fwft.c
348
static int kvm_sbi_ext_fwft_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
36
bool (*supported)(struct kvm_vcpu *vcpu);
arch/riscv/kvm/vcpu_sbi_fwft.c
375
static void kvm_sbi_ext_fwft_deinit(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
382
static void kvm_sbi_ext_fwft_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
396
static unsigned long kvm_sbi_ext_fwft_get_reg_count(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
418
static int kvm_sbi_ext_fwft_get_reg_id(struct kvm_vcpu *vcpu, int index, u64 *reg_id)
arch/riscv/kvm/vcpu_sbi_fwft.c
43
void (*reset)(struct kvm_vcpu *vcpu);
arch/riscv/kvm/vcpu_sbi_fwft.c
449
static int kvm_sbi_ext_fwft_get_reg(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/kvm/vcpu_sbi_fwft.c
486
static int kvm_sbi_ext_fwft_set_reg(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/kvm/vcpu_sbi_fwft.c
52
long (*set)(struct kvm_vcpu *vcpu, struct kvm_sbi_fwft_config *conf,
arch/riscv/kvm/vcpu_sbi_fwft.c
62
long (*get)(struct kvm_vcpu *vcpu, struct kvm_sbi_fwft_config *conf,
arch/riscv/kvm/vcpu_sbi_fwft.c
87
static bool kvm_sbi_fwft_misaligned_delegation_supported(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
92
static void kvm_sbi_fwft_reset_misaligned_delegation(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_fwft.c
99
static long kvm_sbi_fwft_set_misaligned_delegation(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi_hsm.c
16
static int kvm_sbi_hsm_vcpu_start(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_hsm.c
19
struct kvm_vcpu *target_vcpu;
arch/riscv/kvm/vcpu_sbi_hsm.c
44
static int kvm_sbi_hsm_vcpu_stop(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_hsm.c
63
static int kvm_sbi_hsm_vcpu_get_status(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_hsm.c
67
struct kvm_vcpu *target_vcpu;
arch/riscv/kvm/vcpu_sbi_hsm.c
80
static int kvm_sbi_ext_hsm_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_pmu.c
16
static int kvm_sbi_ext_pmu_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_pmu.c
86
static unsigned long kvm_sbi_ext_pmu_probe(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_replace.c
147
static int kvm_sbi_ext_srst_handler(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_sbi_replace.c
17
static int kvm_sbi_ext_time_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_replace.c
45
static int kvm_sbi_ext_ipi_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_replace.c
50
struct kvm_vcpu *tmp;
arch/riscv/kvm/vcpu_sbi_replace.c
92
static int kvm_sbi_ext_rfence_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_sta.c
123
static int kvm_sbi_ext_sta_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_sta.c
144
static unsigned long kvm_sbi_ext_sta_probe(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_sta.c
149
static unsigned long kvm_sbi_ext_sta_get_state_reg_count(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_sta.c
154
static int kvm_sbi_ext_sta_get_reg(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/kvm/vcpu_sbi_sta.c
180
static int kvm_sbi_ext_sta_set_reg(struct kvm_vcpu *vcpu, unsigned long reg_num,
arch/riscv/kvm/vcpu_sbi_sta.c
19
static void kvm_riscv_vcpu_sbi_sta_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_sta.c
25
void kvm_riscv_vcpu_record_steal_time(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_sta.c
81
static int kvm_sbi_sta_steal_time_set_shmem(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_sbi_system.c
12
static int kvm_sbi_ext_susp_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_system.c
18
struct kvm_vcpu *tmp;
arch/riscv/kvm/vcpu_sbi_v01.c
16
static int kvm_sbi_ext_v01_handler(struct kvm_vcpu *vcpu, struct kvm_run *run,
arch/riscv/kvm/vcpu_sbi_v01.c
22
struct kvm_vcpu *rvcpu;
arch/riscv/kvm/vcpu_timer.c
103
int kvm_riscv_vcpu_timer_next_event(struct kvm_vcpu *vcpu, u64 ncycles)
arch/riscv/kvm/vcpu_timer.c
114
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
129
bool kvm_riscv_vcpu_timer_pending(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
141
static void kvm_riscv_vcpu_timer_blocking(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
155
static void kvm_riscv_vcpu_timer_unblocking(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
160
int kvm_riscv_vcpu_get_reg_timer(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_timer.c
200
int kvm_riscv_vcpu_set_reg_timer(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_timer.c
245
int kvm_riscv_vcpu_timer_init(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
271
int kvm_riscv_vcpu_timer_deinit(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
281
int kvm_riscv_vcpu_timer_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
289
static void kvm_riscv_vcpu_update_timedelta(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
301
void kvm_riscv_vcpu_timer_restore(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
325
void kvm_riscv_vcpu_timer_sync(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
340
void kvm_riscv_vcpu_timer_save(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_timer.c
46
struct kvm_vcpu *vcpu = container_of(t, struct kvm_vcpu, arch.timer);
arch/riscv/kvm/vcpu_timer.c
72
static int kvm_riscv_vcpu_update_vstimecmp(struct kvm_vcpu *vcpu, u64 ncycles)
arch/riscv/kvm/vcpu_timer.c
84
static int kvm_riscv_vcpu_update_hrtimer(struct kvm_vcpu *vcpu, u64 ncycles)
arch/riscv/kvm/vcpu_vector.c
139
int kvm_riscv_vcpu_get_reg_vector(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_vector.c
165
int kvm_riscv_vcpu_set_reg_vector(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vcpu_vector.c
19
void kvm_riscv_vcpu_vector_reset(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_vector.c
76
int kvm_riscv_vcpu_alloc_vector_context(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_vector.c
89
void kvm_riscv_vcpu_free_vector_context(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vcpu_vector.c
96
static int kvm_riscv_vcpu_vreg_addr(struct kvm_vcpu *vcpu,
arch/riscv/kvm/vmid.c
71
void kvm_riscv_gstage_vmid_update(struct kvm_vcpu *vcpu)
arch/riscv/kvm/vmid.c
74
struct kvm_vcpu *v;
arch/s390/include/asm/kvm_host.h
507
typedef int (*crypto_hook)(struct kvm_vcpu *vcpu);
arch/s390/include/asm/kvm_host.h
696
bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu);
arch/s390/include/asm/kvm_host.h
698
void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
arch/s390/include/asm/kvm_host.h
701
bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
arch/s390/include/asm/kvm_host.h
704
void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
arch/s390/include/asm/kvm_host.h
707
static inline void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu) {}
arch/s390/include/asm/kvm_host.h
727
bool kvm_s390_pv_cpu_is_protected(struct kvm_vcpu *vcpu);
arch/s390/include/asm/kvm_host.h
743
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu) {}
arch/s390/include/asm/kvm_host.h
744
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu) {}
arch/s390/kvm/diag.c
165
static int __diag_time_slice_end(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
186
static int __diag_time_slice_end_directed(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
188
struct kvm_vcpu *tcpu;
arch/s390/kvm/diag.c
21
static void do_discard_gfn_range(struct kvm_vcpu *vcpu, gfn_t gfn_start, gfn_t gfn_end)
arch/s390/kvm/diag.c
232
static int __diag_ipl_functions(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
266
static int __diag_virtio_hypercall(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
302
int kvm_s390_handle_diag(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
38
static int diag_release_pages(struct kvm_vcpu *vcpu)
arch/s390/kvm/diag.c
78
static int __diag_page_ref_service(struct kvm_vcpu *vcpu)
arch/s390/kvm/faultin.c
15
bool kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu);
arch/s390/kvm/faultin.c
33
int kvm_s390_faultin_gfn(struct kvm_vcpu *vcpu, struct kvm *kvm, struct guest_fault *f)
arch/s390/kvm/faultin.h
16
int kvm_s390_faultin_gfn(struct kvm_vcpu *vcpu, struct kvm *kvm, struct guest_fault *f);
arch/s390/kvm/faultin.h
19
static inline int kvm_s390_faultin_gfn_simple(struct kvm_vcpu *vcpu, struct kvm *kvm,
arch/s390/kvm/gaccess.c
1009
int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra,
arch/s390/kvm/gaccess.c
1184
int guest_translate_address_with_key(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.c
1208
int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.c
1258
int kvm_s390_check_low_addr_prot_real(struct kvm_vcpu *vcpu, unsigned long gra)
arch/s390/kvm/gaccess.c
1583
static inline int _gaccess_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *sg, gpa_t saddr,
arch/s390/kvm/gaccess.c
1630
static int __gaccess_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *sg, gpa_t saddr,
arch/s390/kvm/gaccess.c
1655
int gaccess_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *sg, gpa_t saddr,
arch/s390/kvm/gaccess.c
234
static int ar_translation(struct kvm_vcpu *vcpu, union asce *asce, u8 ar,
arch/s390/kvm/gaccess.c
334
static int trans_exc_ending(struct kvm_vcpu *vcpu, int code, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.c
404
static int trans_exc(struct kvm_vcpu *vcpu, int code, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.c
410
static int get_vcpu_asce(struct kvm_vcpu *vcpu, union asce *asce,
arch/s390/kvm/gaccess.c
470
static unsigned long guest_translate_gva(struct kvm_vcpu *vcpu, unsigned long gva,
arch/s390/kvm/gaccess.c
640
static int low_address_protection_enabled(struct kvm_vcpu *vcpu,
arch/s390/kvm/gaccess.c
670
static bool fetch_prot_override_applicable(struct kvm_vcpu *vcpu, enum gacc_mode mode,
arch/s390/kvm/gaccess.c
692
static bool storage_prot_override_applicable(struct kvm_vcpu *vcpu)
arch/s390/kvm/gaccess.c
704
static int vcpu_check_access_key_gpa(struct kvm_vcpu *vcpu, u8 access_key,
arch/s390/kvm/gaccess.c
774
static int guest_range_to_gpas(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
arch/s390/kvm/gaccess.c
929
int access_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
arch/s390/kvm/gaccess.h
122
struct kvm_vcpu *__vcpu = (vcpu); \
arch/s390/kvm/gaccess.h
149
int write_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data,
arch/s390/kvm/gaccess.h
175
int read_guest_lc(struct kvm_vcpu *vcpu, unsigned long gra, void *data,
arch/s390/kvm/gaccess.h
189
int guest_translate_address_with_key(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.h
193
int check_gva_range(struct kvm_vcpu *vcpu, unsigned long gva, u8 ar,
arch/s390/kvm/gaccess.h
202
int access_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
arch/s390/kvm/gaccess.h
206
int access_guest_real(struct kvm_vcpu *vcpu, unsigned long gra,
arch/s390/kvm/gaccess.h
259
int write_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
arch/s390/kvm/gaccess.h
278
int write_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data,
arch/s390/kvm/gaccess.h
301
int read_guest_with_key(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar,
arch/s390/kvm/gaccess.h
322
int read_guest(struct kvm_vcpu *vcpu, unsigned long ga, u8 ar, void *data,
arch/s390/kvm/gaccess.h
345
int read_guest_instr(struct kvm_vcpu *vcpu, unsigned long ga, void *data,
arch/s390/kvm/gaccess.h
371
int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data,
arch/s390/kvm/gaccess.h
394
int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data,
arch/s390/kvm/gaccess.h
418
int write_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, void *data,
arch/s390/kvm/gaccess.h
44
static inline unsigned long kvm_s390_real_to_abs(struct kvm_vcpu *vcpu,
arch/s390/kvm/gaccess.h
442
int read_guest_real(struct kvm_vcpu *vcpu, unsigned long gra, void *data,
arch/s390/kvm/gaccess.h
451
int kvm_s390_check_low_addr_prot_real(struct kvm_vcpu *vcpu, unsigned long gra);
arch/s390/kvm/gaccess.h
463
int gaccess_shadow_fault(struct kvm_vcpu *vcpu, struct gmap *sg, gpa_t saddr,
arch/s390/kvm/gaccess.h
86
static inline unsigned long kvm_s390_logical_to_effective(struct kvm_vcpu *vcpu,
arch/s390/kvm/gmap.c
257
struct kvm_vcpu *vcpu;
arch/s390/kvm/gmap.c
27
static inline bool kvm_s390_is_in_sie(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
130
void kvm_s390_backup_guest_per_regs(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
138
void kvm_s390_restore_guest_per_regs(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
146
void kvm_s390_patch_guest_per_regs(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
174
static int __import_wp_info(struct kvm_vcpu *vcpu,
arch/s390/kvm/guestdbg.c
203
int kvm_s390_import_bp_data(struct kvm_vcpu *vcpu,
arch/s390/kvm/guestdbg.c
278
void kvm_s390_clear_bp_data(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
309
static struct kvm_hw_bp_info_arch *find_hw_bp(struct kvm_vcpu *vcpu,
arch/s390/kvm/guestdbg.c
334
static struct kvm_hw_wp_info_arch *any_wp_changed(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
367
void kvm_s390_prepare_debug_exit(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
384
static int debug_exit_required(struct kvm_vcpu *vcpu, u8 perc,
arch/s390/kvm/guestdbg.c
432
static int per_fetched_addr(struct kvm_vcpu *vcpu, unsigned long *addr)
arch/s390/kvm/guestdbg.c
493
int kvm_s390_handle_per_ifetch_icpt(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
533
static int filter_guest_per_event(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
582
int kvm_s390_handle_per_event(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
59
static void enable_all_hw_bp(struct kvm_vcpu *vcpu)
arch/s390/kvm/guestdbg.c
99
static void enable_all_hw_wp(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
107
static int handle_instruction(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
144
static int inject_prog_on_prog_intercept(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
212
static int handle_itdb(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
232
static bool should_handle_per_event(const struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
247
static int handle_prog(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
26
u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
303
static int handle_external_interrupt(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
359
static int handle_mvpg_pei(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
395
static int handle_partial_execution(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
412
int handle_sthyi(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
469
static int handle_operexc(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
508
static int handle_pv_spx(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
517
static int handle_pv_sclp(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
54
static int handle_stop(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
541
static int handle_pv_uvc(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
572
static int handle_pv_notification(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
599
static bool should_handle_per_ifetch(const struct kvm_vcpu *vcpu, int rc)
arch/s390/kvm/intercept.c
612
int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu)
arch/s390/kvm/intercept.c
92
static int handle_validity(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
101
static int psw_ioint_disabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1034
static int __must_check __deliver_virtio(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
106
static int psw_mchk_disabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1080
static int __do_deliver_io(struct kvm_vcpu *vcpu, struct kvm_s390_io_info *io)
arch/s390/kvm/interrupt.c
1106
static int __must_check __deliver_io(struct kvm_vcpu *vcpu,
arch/s390/kvm/interrupt.c
111
static int psw_interrupts_disabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1176
int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
118
static int ckc_interrupts_enabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1186
int kvm_s390_vcpu_has_irq(struct kvm_vcpu *vcpu, int exclude_stop)
arch/s390/kvm/interrupt.c
1205
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1210
static u64 __calculate_sltime(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1242
int kvm_s390_handle_wait(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1289
void kvm_s390_vcpu_wakeup(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
129
static int ckc_irq_pending(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1303
struct kvm_vcpu *vcpu;
arch/s390/kvm/interrupt.c
1306
vcpu = container_of(timer, struct kvm_vcpu, arch.ckc_timer);
arch/s390/kvm/interrupt.c
1319
void kvm_s390_clear_local_irqs(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1332
int __must_check kvm_s390_deliver_pending_interrupts(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
143
static int cpu_timer_interrupts_enabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1430
static int __inject_prog(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
1472
static int __inject_pfault_init(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
1489
static int __inject_extcall(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
149
static int cpu_timer_irq_pending(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1515
static int __inject_set_prefix(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
1535
static int __inject_sigp_stop(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
1561
static int __inject_sigp_restart(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1573
static int __inject_sigp_emergency(struct kvm_vcpu *vcpu,
arch/s390/kvm/interrupt.c
1594
static int __inject_mchk(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
1626
static int __inject_ckc(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1640
static int __inject_cpu_timer(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
1899
struct kvm_vcpu *dst_vcpu;
arch/s390/kvm/interrupt.c
2064
int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
2071
int kvm_s390_is_restart_irq_pending(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
2078
void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
2088
static int do_inject_vcpu(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
2133
int kvm_s390_inject_vcpu(struct kvm_vcpu *vcpu, struct kvm_s390_irq *irq)
arch/s390/kvm/interrupt.c
2623
struct kvm_vcpu *vcpu;
arch/s390/kvm/interrupt.c
267
static inline unsigned long pending_irqs_no_gisa(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
276
static inline unsigned long pending_irqs(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
2819
void kvm_s390_reinject_machine_check(struct kvm_vcpu *vcpu,
arch/s390/kvm/interrupt.c
2909
int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu, void __user *irqstate, int len)
arch/s390/kvm/interrupt.c
297
static unsigned long disable_iscs(struct kvm_vcpu *vcpu,
arch/s390/kvm/interrupt.c
2991
int kvm_s390_get_irq_state(struct kvm_vcpu *vcpu, __u8 __user *buf, int len)
arch/s390/kvm/interrupt.c
3051
struct kvm_vcpu *vcpu;
arch/s390/kvm/interrupt.c
309
static unsigned long deliverable_irqs(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
3165
struct kvm_vcpu *vcpu;
arch/s390/kvm/interrupt.c
3206
struct kvm_vcpu *vcpu;
arch/s390/kvm/interrupt.c
361
static void __set_cpu_idle(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
367
static void __unset_cpu_idle(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
373
static void __reset_intercept_indicators(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
387
static void set_intercept_indicators_io(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
397
static void set_intercept_indicators_ext(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
407
static void set_intercept_indicators_mchk(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
417
static void set_intercept_indicators_stop(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
424
static void set_intercept_indicators(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
432
static int __must_check __deliver_cpu_timer(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
45
static int sca_ext_call_pending(struct kvm_vcpu *vcpu, int *src_id)
arch/s390/kvm/interrupt.c
456
static int __must_check __deliver_ckc(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
480
static int __must_check __deliver_pfault_init(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
508
static int __write_machine_check(struct kvm_vcpu *vcpu,
arch/s390/kvm/interrupt.c
61
static int sca_inject_ext_call(struct kvm_vcpu *vcpu, int src_id)
arch/s390/kvm/interrupt.c
620
static int __must_check __deliver_machine_check(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
671
static int __must_check __deliver_restart(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
693
static int __must_check __deliver_set_prefix(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
713
static int __must_check __deliver_emergency_signal(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
747
static int __must_check __deliver_external_call(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
781
static int __deliver_prog_pv(struct kvm_vcpu *vcpu, u16 code)
arch/s390/kvm/interrupt.c
796
static int __must_check __deliver_prog(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
84
static void sca_clear_ext_call(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
917
static int write_sclp(struct kvm_vcpu *vcpu, u32 parm)
arch/s390/kvm/interrupt.c
940
static int __must_check __deliver_service(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
96
int psw_extint_disabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
968
static int __must_check __deliver_service_ev(struct kvm_vcpu *vcpu)
arch/s390/kvm/interrupt.c
992
static int __must_check __deliver_pfault_done(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
1026
static void kvm_s390_vcpu_crypto_setup(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.c
1030
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
1114
static void kvm_s390_vcpu_pci_setup(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
1126
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
1153
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
2322
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
2367
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
305
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
3306
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3365
static void sca_del_vcpu(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3376
static void sca_add_vcpu(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3402
static void __start_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3411
static void __stop_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3421
static void __enable_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3429
static void __disable_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3436
static void enable_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3443
static void disable_cpu_timer_accounting(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3451
void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm)
arch/s390/kvm/kvm-s390.c
3463
__u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3488
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/s390/kvm/kvm-s390.c
3497
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3506
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3548
static void kvm_s390_vcpu_crypto_setup(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3579
void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3585
int kvm_s390_vcpu_setup_cmma(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3596
static void kvm_s390_vcpu_setup_model(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3605
static int kvm_s390_vcpu_setup(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3699
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3782
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3788
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3793
void kvm_s390_vcpu_block(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3799
void kvm_s390_vcpu_unblock(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3804
static void kvm_s390_vcpu_request(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3810
bool kvm_s390_vcpu_sie_inhibited(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3816
static void kvm_s390_vcpu_request_handled(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3825
void exit_sie(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3834
void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3840
bool kvm_arch_no_poll(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3851
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3858
static int kvm_arch_vcpu_ioctl_get_one_reg(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
3907
static int kvm_arch_vcpu_ioctl_set_one_reg(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
3960
static void kvm_arch_vcpu_ioctl_normal_reset(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
3972
static void kvm_arch_vcpu_ioctl_initial_reset(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4016
static void kvm_arch_vcpu_ioctl_clear_reset(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4032
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/s390/kvm/kvm-s390.c
4040
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/s390/kvm/kvm-s390.c
4048
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4060
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4072
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/s390/kvm/kvm-s390.c
4087
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/s390/kvm/kvm-s390.c
4102
static int kvm_arch_vcpu_ioctl_set_initial_psw(struct kvm_vcpu *vcpu, psw_t psw)
arch/s390/kvm/kvm-s390.c
4115
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4125
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4167
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4182
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4216
static bool ibs_enabled(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4221
static int vcpu_ucontrol_translate(struct kvm_vcpu *vcpu, gpa_t *gaddr)
arch/s390/kvm/kvm-s390.c
4237
static int kvm_s390_fixup_prefix(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4259
static int kvm_s390_handle_requests(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4337
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
4372
static void __kvm_inject_pfault_token(struct kvm_vcpu *vcpu, bool start_token,
arch/s390/kvm/kvm-s390.c
4389
bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4398
void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4405
void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
4411
bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4420
bool kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4446
static int vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4486
static int vcpu_post_run_addressing_exception(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4522
static void kvm_s390_assert_primary_as(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4529
static int vcpu_dat_fault_handler(struct kvm_vcpu *vcpu, gpa_t gaddr, bool wr)
arch/s390/kvm/kvm-s390.c
4550
static int vcpu_post_run_handle_fault(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4620
static int vcpu_post_run(struct kvm_vcpu *vcpu, int sie_return)
arch/s390/kvm/kvm-s390.c
4680
static int __vcpu_run(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4757
static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4830
static void sync_regs(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4870
static void store_regs_fmt2(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4894
static void store_regs(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
4915
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
5001
int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long gpa)
arch/s390/kvm/kvm-s390.c
5053
int kvm_s390_vcpu_store_status(struct kvm_vcpu *vcpu, unsigned long addr)
arch/s390/kvm/kvm-s390.c
5066
static void __disable_ibs_on_vcpu(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
5075
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.c
5082
static void __enable_ibs_on_vcpu(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
5090
int kvm_s390_vcpu_start(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
5145
int kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.c
5148
struct kvm_vcpu *started_vcpu = NULL;
arch/s390/kvm/kvm-s390.c
5179
struct kvm_vcpu *tmp = kvm_get_vcpu(vcpu->kvm, i);
arch/s390/kvm/kvm-s390.c
5199
static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
5223
static long kvm_s390_vcpu_sida_op(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
5255
static long kvm_s390_vcpu_mem_op(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
5300
static long kvm_s390_vcpu_memsida_op(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
5328
struct kvm_vcpu *vcpu = filp->private_data;
arch/s390/kvm/kvm-s390.c
5370
static int kvm_s390_handle_pv_vcpu_dump(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.c
5415
struct kvm_vcpu *vcpu = filp->private_data;
arch/s390/kvm/kvm-s390.c
5627
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/s390/kvm/kvm-s390.c
5810
r = kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
arch/s390/kvm/kvm-s390.c
729
static void sca_del_vcpu(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.c
769
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.h
101
static inline bool kvm_s390_test_cpuflags(struct kvm_vcpu *vcpu, u32 flags)
arch/s390/kvm/kvm-s390.h
106
static inline int is_vcpu_stopped(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.h
111
static inline int is_vcpu_idle(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.h
126
static inline u32 kvm_s390_get_prefix(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.h
131
static inline void kvm_s390_set_prefix(struct kvm_vcpu *vcpu, u32 prefix)
arch/s390/kvm/kvm-s390.h
140
static inline u64 kvm_s390_get_base_disp_s(struct kvm_vcpu *vcpu, u8 *ar)
arch/s390/kvm/kvm-s390.h
151
static inline u64 kvm_s390_get_base_disp_siy(struct kvm_vcpu *vcpu, u8 *ar)
arch/s390/kvm/kvm-s390.h
166
static inline void kvm_s390_get_base_disp_sse(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
184
static inline void kvm_s390_get_regs_rre(struct kvm_vcpu *vcpu, int *r1, int *r2)
arch/s390/kvm/kvm-s390.h
192
static inline u64 kvm_s390_get_base_disp_rsy(struct kvm_vcpu *vcpu, u8 *ar)
arch/s390/kvm/kvm-s390.h
207
static inline u64 kvm_s390_get_base_disp_rs(struct kvm_vcpu *vcpu, u8 *ar)
arch/s390/kvm/kvm-s390.h
219
static inline void kvm_s390_set_psw_cc(struct kvm_vcpu *vcpu, unsigned long cc)
arch/s390/kvm/kvm-s390.h
302
int kvm_s390_pv_destroy_cpu(struct kvm_vcpu *vcpu, u16 *rc, u16 *rrc);
arch/s390/kvm/kvm-s390.h
303
int kvm_s390_pv_create_cpu(struct kvm_vcpu *vcpu, u16 *rc, u16 *rrc);
arch/s390/kvm/kvm-s390.h
313
int kvm_s390_pv_set_cpu_state(struct kvm_vcpu *vcpu, u8 state);
arch/s390/kvm/kvm-s390.h
314
int kvm_s390_pv_dump_cpu(struct kvm_vcpu *vcpu, void *buff, u16 *rc, u16 *rrc);
arch/s390/kvm/kvm-s390.h
328
static inline u64 kvm_s390_pv_cpu_get_handle(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.h
369
int kvm_s390_handle_wait(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
370
void kvm_s390_vcpu_wakeup(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
372
int __must_check kvm_s390_deliver_pending_interrupts(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
373
void kvm_s390_clear_local_irqs(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
377
int __must_check kvm_s390_inject_vcpu(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
379
static inline int kvm_s390_inject_prog_irq(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
389
static inline int kvm_s390_inject_program_int(struct kvm_vcpu *vcpu, u16 code)
arch/s390/kvm/kvm-s390.h
405
u8 kvm_s390_get_ilen(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
406
int kvm_handle_sie_intercept(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
407
static inline void kvm_s390_rewind_psw(struct kvm_vcpu *vcpu, int ilen)
arch/s390/kvm/kvm-s390.h
413
static inline void kvm_s390_forward_psw(struct kvm_vcpu *vcpu, int ilen)
arch/s390/kvm/kvm-s390.h
417
static inline void kvm_s390_retry_instr(struct kvm_vcpu *vcpu)
arch/s390/kvm/kvm-s390.h
424
int handle_sthyi(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
428
int kvm_s390_handle_aa(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
429
int kvm_s390_handle_b2(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
430
int kvm_s390_handle_e3(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
431
int kvm_s390_handle_e5(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
432
int kvm_s390_handle_01(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
433
int kvm_s390_handle_b9(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
434
int kvm_s390_handle_lpsw(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
435
int kvm_s390_handle_stctl(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
436
int kvm_s390_handle_lctl(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
437
int kvm_s390_handle_eb(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
438
int kvm_s390_skey_check_enable(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
441
int kvm_s390_handle_vsie(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
442
void kvm_s390_vsie_kick(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
448
int kvm_s390_handle_sigp(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
449
int kvm_s390_handle_sigp_pei(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
453
int kvm_s390_store_status_unloaded(struct kvm_vcpu *vcpu, unsigned long addr);
arch/s390/kvm/kvm-s390.h
454
int kvm_s390_vcpu_store_status(struct kvm_vcpu *vcpu, unsigned long addr);
arch/s390/kvm/kvm-s390.h
455
int kvm_s390_vcpu_start(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
456
int kvm_s390_vcpu_stop(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
457
void kvm_s390_vcpu_block(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
458
void kvm_s390_vcpu_unblock(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
459
bool kvm_s390_vcpu_sie_inhibited(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
460
void exit_sie(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
461
void kvm_s390_sync_request(int req, struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
462
int kvm_s390_vcpu_setup_cmma(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
463
void kvm_s390_vcpu_unsetup_cmma(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
464
void kvm_s390_set_cpu_timer(struct kvm_vcpu *vcpu, __u64 cputm);
arch/s390/kvm/kvm-s390.h
465
__u64 kvm_s390_get_cpu_timer(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
470
bool kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
473
int kvm_s390_handle_diag(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
478
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.h
488
struct kvm_vcpu *vcpu;
arch/s390/kvm/kvm-s390.h
530
static inline int kvm_s390_inject_prog_cond(struct kvm_vcpu *vcpu, int rc)
arch/s390/kvm/kvm-s390.h
541
int kvm_s390_vcpu_has_irq(struct kvm_vcpu *vcpu, int exclude_stop);
arch/s390/kvm/kvm-s390.h
542
int psw_extint_disabled(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
544
int kvm_s390_ext_call_pending(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
546
int kvm_s390_is_stop_irq_pending(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
547
int kvm_s390_is_restart_irq_pending(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
548
void kvm_s390_clear_stop_irq(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
549
int kvm_s390_set_irq_state(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
551
int kvm_s390_get_irq_state(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
562
void kvm_s390_backup_guest_per_regs(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
563
void kvm_s390_restore_guest_per_regs(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
564
void kvm_s390_patch_guest_per_regs(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
565
int kvm_s390_import_bp_data(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
567
void kvm_s390_clear_bp_data(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
568
void kvm_s390_prepare_debug_exit(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
569
int kvm_s390_handle_per_ifetch_icpt(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
570
int kvm_s390_handle_per_event(struct kvm_vcpu *vcpu);
arch/s390/kvm/kvm-s390.h
581
void kvm_s390_reinject_machine_check(struct kvm_vcpu *vcpu,
arch/s390/kvm/kvm-s390.h
91
static inline void kvm_s390_set_cpuflags(struct kvm_vcpu *vcpu, u32 flags)
arch/s390/kvm/kvm-s390.h
96
static inline void kvm_s390_clear_cpuflags(struct kvm_vcpu *vcpu, u32 flags)
arch/s390/kvm/priv.c
1010
static int handle_epsw(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1039
static int handle_pfmf(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1144
static inline int __do_essa(struct kvm_vcpu *vcpu, const int orc)
arch/s390/kvm/priv.c
1182
static void _essa_clear_cbrl(struct kvm_vcpu *vcpu, unsigned long *cbrl, int len)
arch/s390/kvm/priv.c
1204
static int handle_essa(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
123
static int handle_set_prefix(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1270
int kvm_s390_handle_b9(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1288
int kvm_s390_handle_lctl(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1327
int kvm_s390_handle_stctl(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1361
static int handle_lctlg(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1399
static int handle_stctg(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1433
int kvm_s390_handle_eb(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1451
static int handle_tprot(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1510
int kvm_s390_handle_e5(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1520
static int handle_sckpf(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1539
static int handle_ptff(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
1548
int kvm_s390_handle_01(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
161
static int handle_store_prefix(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
191
static int handle_store_cpu_address(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
217
int kvm_s390_skey_check_enable(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
241
static int try_handle_skey(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
257
static int handle_iske(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
289
static int handle_rrbe(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
323
static int handle_sske(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
33
static int handle_ri(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
398
static int handle_ipte_interlock(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
409
static int handle_test_block(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
439
static int handle_tpi(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
46
int kvm_s390_handle_aa(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
505
static int handle_tsch(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
538
static int handle_io_inst(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
54
static int handle_gs(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
590
static int handle_pqap(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
667
static int handle_stfl(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
713
int kvm_s390_handle_lpsw(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
74
int kvm_s390_handle_e3(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
743
static int handle_lpswe(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
767
static int handle_lpswey(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
797
static int handle_stidp(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
822
static void handle_stsi_3_2_2(struct kvm_vcpu *vcpu, struct sysinfo_3_2_2 *mem)
arch/s390/kvm/priv.c
84
static int handle_set_clock(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
849
static void insert_stsi_usr_data(struct kvm_vcpu *vcpu, u64 addr, u8 ar,
arch/s390/kvm/priv.c
860
static int handle_stsi(struct kvm_vcpu *vcpu)
arch/s390/kvm/priv.c
952
int kvm_s390_handle_b2(struct kvm_vcpu *vcpu)
arch/s390/kvm/pv.c
228
int kvm_s390_pv_destroy_cpu(struct kvm_vcpu *vcpu, u16 *rc, u16 *rrc)
arch/s390/kvm/pv.c
262
int kvm_s390_pv_create_cpu(struct kvm_vcpu *vcpu, u16 *rc, u16 *rrc)
arch/s390/kvm/pv.c
33
bool kvm_s390_pv_cpu_is_protected(struct kvm_vcpu *vcpu)
arch/s390/kvm/pv.c
828
int kvm_s390_pv_set_cpu_state(struct kvm_vcpu *vcpu, u8 state)
arch/s390/kvm/pv.c
846
int kvm_s390_pv_dump_cpu(struct kvm_vcpu *vcpu, void *buff, u16 *rc, u16 *rrc)
arch/s390/kvm/sigp.c
115
static int __sigp_stop(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu)
arch/s390/kvm/sigp.c
132
static int __sigp_stop_and_store_status(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
133
struct kvm_vcpu *dst_vcpu, u64 *reg)
arch/s390/kvm/sigp.c
151
static int __sigp_set_arch(struct kvm_vcpu *vcpu, u32 parameter,
arch/s390/kvm/sigp.c
161
static int __sigp_set_prefix(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu,
arch/s390/kvm/sigp.c
191
static int __sigp_store_status_at_addr(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
192
struct kvm_vcpu *dst_vcpu,
arch/s390/kvm/sigp.c
20
static int __sigp_sense(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu,
arch/s390/kvm/sigp.c
213
static int __sigp_sense_running(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
214
struct kvm_vcpu *dst_vcpu, u64 *reg)
arch/s390/kvm/sigp.c
240
static int __prepare_sigp_re_start(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
241
struct kvm_vcpu *dst_vcpu, u8 order_code)
arch/s390/kvm/sigp.c
256
static int __prepare_sigp_cpu_reset(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
257
struct kvm_vcpu *dst_vcpu, u8 order_code)
arch/s390/kvm/sigp.c
263
static int __prepare_sigp_unknown(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
264
struct kvm_vcpu *dst_vcpu)
arch/s390/kvm/sigp.c
270
static int handle_sigp_dst(struct kvm_vcpu *vcpu, u8 order_code,
arch/s390/kvm/sigp.c
274
struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr);
arch/s390/kvm/sigp.c
375
static int handle_sigp_order_in_user_space(struct kvm_vcpu *vcpu, u8 order_code,
arch/s390/kvm/sigp.c
425
int kvm_s390_handle_sigp(struct kvm_vcpu *vcpu)
arch/s390/kvm/sigp.c
44
static int __inject_sigp_emergency(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
45
struct kvm_vcpu *dst_vcpu)
arch/s390/kvm/sigp.c
476
int kvm_s390_handle_sigp_pei(struct kvm_vcpu *vcpu)
arch/s390/kvm/sigp.c
480
struct kvm_vcpu *dest_vcpu;
arch/s390/kvm/sigp.c
61
static int __sigp_emergency(struct kvm_vcpu *vcpu, struct kvm_vcpu *dst_vcpu)
arch/s390/kvm/sigp.c
66
static int __sigp_conditional_emergency(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
67
struct kvm_vcpu *dst_vcpu,
arch/s390/kvm/sigp.c
93
static int __sigp_external_call(struct kvm_vcpu *vcpu,
arch/s390/kvm/sigp.c
94
struct kvm_vcpu *dst_vcpu, u64 *reg)
arch/s390/kvm/trace-s390.h
43
TP_PROTO(unsigned int id, struct kvm_vcpu *vcpu,
arch/s390/kvm/trace-s390.h
49
__field(struct kvm_vcpu *, vcpu)
arch/s390/kvm/trace.h
20
#define VCPU_PROTO_COMMON struct kvm_vcpu *vcpu
arch/s390/kvm/vsie.c
1025
static u64 vsie_get_register(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, u8 reg)
arch/s390/kvm/vsie.c
1039
static int vsie_handle_mvpg(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, struct gmap *sg)
arch/s390/kvm/vsie.c
1119
static int do_vsie_run(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, struct gmap *sg)
arch/s390/kvm/vsie.c
122
static int prepare_cpuflags(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
1232
static struct gmap *acquire_gmap_shadow(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
1284
static void register_shadow_scb(struct kvm_vcpu *vcpu,
arch/s390/kvm/vsie.c
1314
static void unregister_shadow_scb(struct kvm_vcpu *vcpu)
arch/s390/kvm/vsie.c
1328
static int vsie_run(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
1488
int kvm_s390_handle_vsie(struct kvm_vcpu *vcpu)
arch/s390/kvm/vsie.c
1574
void kvm_s390_vsie_kick(struct kvm_vcpu *vcpu)
arch/s390/kvm/vsie.c
159
static int setup_apcb10(struct kvm_vcpu *vcpu, struct kvm_s390_apcb1 *apcb_s,
arch/s390/kvm/vsie.c
188
static int setup_apcb00(struct kvm_vcpu *vcpu, unsigned long *apcb_s,
arch/s390/kvm/vsie.c
214
static int setup_apcb11(struct kvm_vcpu *vcpu, unsigned long *apcb_s,
arch/s390/kvm/vsie.c
246
static int setup_apcb(struct kvm_vcpu *vcpu, struct kvm_s390_crypto_cb *crycb_s,
arch/s390/kvm/vsie.c
313
static int shadow_crycb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
389
static void prepare_ibc(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
412
static void unshadow_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
463
static int shadow_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
630
static int map_prefix(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, struct gmap *sg)
arch/s390/kvm/vsie.c
686
static void unpin_blocks(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
742
static int pin_blocks(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/s390/kvm/vsie.c
860
static void unpin_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page,
arch/s390/kvm/vsie.c
876
static int pin_scb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page,
arch/s390/kvm/vsie.c
898
static int inject_fault(struct kvm_vcpu *vcpu, __u16 code, __u64 vaddr,
arch/s390/kvm/vsie.c
928
static int handle_fault(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, struct gmap *sg)
arch/s390/kvm/vsie.c
954
static void handle_last_fault(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page, struct gmap *sg)
arch/s390/kvm/vsie.c
989
static int handle_stfle(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page)
arch/x86/include/asm/kvm_host.h
1742
void (*vcpu_after_set_cpuid)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1751
int (*vcpu_create)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1752
void (*vcpu_free)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1753
void (*vcpu_reset)(struct kvm_vcpu *vcpu, bool init_event);
arch/x86/include/asm/kvm_host.h
1755
void (*prepare_switch_to_guest)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1756
void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu);
arch/x86/include/asm/kvm_host.h
1757
void (*vcpu_put)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1765
void (*update_exception_bitmap)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1766
int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/include/asm/kvm_host.h
1767
int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/include/asm/kvm_host.h
1768
u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg);
arch/x86/include/asm/kvm_host.h
1769
void (*get_segment)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1771
int (*get_cpl)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1772
int (*get_cpl_no_cache)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1773
void (*set_segment)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1775
void (*get_cs_db_l_bits)(struct kvm_vcpu *vcpu, int *db, int *l);
arch/x86/include/asm/kvm_host.h
1776
bool (*is_valid_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/include/asm/kvm_host.h
1777
void (*set_cr0)(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/include/asm/kvm_host.h
1778
void (*post_set_cr3)(struct kvm_vcpu *vcpu, unsigned long cr3);
arch/x86/include/asm/kvm_host.h
1779
bool (*is_valid_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
1780
void (*set_cr4)(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
1781
int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer);
arch/x86/include/asm/kvm_host.h
1782
void (*get_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/include/asm/kvm_host.h
1783
void (*set_idt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/include/asm/kvm_host.h
1784
void (*get_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/include/asm/kvm_host.h
1785
void (*set_gdt)(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/include/asm/kvm_host.h
1786
void (*sync_dirty_debug_regs)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1787
void (*set_dr7)(struct kvm_vcpu *vcpu, unsigned long value);
arch/x86/include/asm/kvm_host.h
1788
void (*cache_reg)(struct kvm_vcpu *vcpu, enum kvm_reg reg);
arch/x86/include/asm/kvm_host.h
1789
unsigned long (*get_rflags)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1790
void (*set_rflags)(struct kvm_vcpu *vcpu, unsigned long rflags);
arch/x86/include/asm/kvm_host.h
1791
bool (*get_if_flag)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1793
void (*flush_tlb_all)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1794
void (*flush_tlb_current)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1807
void (*flush_tlb_gva)(struct kvm_vcpu *vcpu, gva_t addr);
arch/x86/include/asm/kvm_host.h
1813
void (*flush_tlb_guest)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1815
int (*vcpu_pre_run)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1816
enum exit_fastpath_completion (*vcpu_run)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1818
int (*handle_exit)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1820
int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1821
void (*update_emulated_instruction)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1822
void (*set_interrupt_shadow)(struct kvm_vcpu *vcpu, int mask);
arch/x86/include/asm/kvm_host.h
1823
u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1824
void (*patch_hypercall)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1826
void (*inject_irq)(struct kvm_vcpu *vcpu, bool reinjected);
arch/x86/include/asm/kvm_host.h
1827
void (*inject_nmi)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1828
void (*inject_exception)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1829
void (*cancel_injection)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1830
int (*interrupt_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/include/asm/kvm_host.h
1831
int (*nmi_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/include/asm/kvm_host.h
1832
bool (*get_nmi_mask)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1833
void (*set_nmi_mask)(struct kvm_vcpu *vcpu, bool masked);
arch/x86/include/asm/kvm_host.h
1835
bool (*is_vnmi_pending)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1840
bool (*set_vnmi_pending)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1841
void (*enable_nmi_window)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1842
void (*enable_irq_window)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1843
void (*update_cr8_intercept)(struct kvm_vcpu *vcpu, int tpr, int irr);
arch/x86/include/asm/kvm_host.h
1848
void (*refresh_apicv_exec_ctrl)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1849
void (*hwapic_isr_update)(struct kvm_vcpu *vcpu, int isr);
arch/x86/include/asm/kvm_host.h
1850
void (*load_eoi_exitmap)(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap);
arch/x86/include/asm/kvm_host.h
1851
void (*set_virtual_apic_mode)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1852
void (*set_apic_access_page_addr)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1855
int (*sync_pir_to_irr)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1858
u8 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
arch/x86/include/asm/kvm_host.h
1860
void (*load_mmu_pgd)(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/include/asm/kvm_host.h
1880
u64 (*get_l2_tsc_offset)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1881
u64 (*get_l2_tsc_multiplier)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1882
void (*write_tsc_offset)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1883
void (*write_tsc_multiplier)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1889
void (*get_exit_info)(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/include/asm/kvm_host.h
1893
void (*get_entry_info)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1896
int (*check_intercept)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1900
void (*handle_exit_irqoff)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1902
void (*update_cpu_dirty_logging)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1906
void (*vcpu_blocking)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1907
void (*vcpu_unblocking)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1911
struct kvm_vcpu *vcpu, u32 vector);
arch/x86/include/asm/kvm_host.h
1913
void (*apicv_pre_state_restore)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1914
void (*apicv_post_state_restore)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1915
bool (*dy_apicv_has_pending_interrupt)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1916
bool (*protected_apic_has_interrupt)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1918
int (*set_hv_timer)(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc,
arch/x86/include/asm/kvm_host.h
1920
void (*cancel_hv_timer)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1922
void (*setup_mce)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1925
int (*smi_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/include/asm/kvm_host.h
1926
int (*enter_smm)(struct kvm_vcpu *vcpu, union kvm_smram *smram);
arch/x86/include/asm/kvm_host.h
1927
int (*leave_smm)(struct kvm_vcpu *vcpu, const union kvm_smram *smram);
arch/x86/include/asm/kvm_host.h
1928
void (*enable_smi_window)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1933
int (*vcpu_mem_enc_ioctl)(struct kvm_vcpu *vcpu, void __user *argp);
arch/x86/include/asm/kvm_host.h
1934
int (*vcpu_mem_enc_unlocked_ioctl)(struct kvm_vcpu *vcpu, void __user *argp);
arch/x86/include/asm/kvm_host.h
1943
int (*check_emulate_instruction)(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/include/asm/kvm_host.h
1946
bool (*apic_init_signal_blocked)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1947
int (*enable_l2_tlb_flush)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1949
void (*migrate_timers)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1950
void (*recalc_intercepts)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1951
int (*complete_emulated_msr)(struct kvm_vcpu *vcpu, int err);
arch/x86/include/asm/kvm_host.h
1953
void (*vcpu_deliver_sipi_vector)(struct kvm_vcpu *vcpu, u8 vector);
arch/x86/include/asm/kvm_host.h
1958
unsigned long (*vcpu_get_apicv_inhibit_reasons)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1960
gva_t (*get_untagged_addr)(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags);
arch/x86/include/asm/kvm_host.h
1961
void *(*alloc_apic_backing_page)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1968
void (*leave_nested)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1969
bool (*is_exception_vmexit)(struct kvm_vcpu *vcpu, u8 vector,
arch/x86/include/asm/kvm_host.h
1971
int (*check_events)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1972
bool (*has_events)(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/include/asm/kvm_host.h
1973
void (*triple_fault)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1974
int (*get_state)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1977
int (*set_state)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1980
bool (*get_nested_state_pages)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1981
int (*write_log_dirty)(struct kvm_vcpu *vcpu, gpa_t l2_gpa);
arch/x86/include/asm/kvm_host.h
1983
int (*enable_evmcs)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
1985
uint16_t (*get_evmcs_version)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
1986
void (*hv_inject_synthetic_vmexit_post_tlb_flush)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2070
void kvm_mmu_destroy(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2071
int kvm_mmu_create(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2078
void kvm_mmu_after_set_cpuid(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2079
void kvm_mmu_reset_context(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2098
int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3);
arch/x86/include/asm/kvm_host.h
2100
int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/include/asm/kvm_host.h
2105
u64 vcpu_tsc_khz(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2189
int kvm_emulate_instruction(struct kvm_vcpu *vcpu, int emulation_type);
arch/x86/include/asm/kvm_host.h
2190
int kvm_emulate_instruction_from_buffer(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2192
void __kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2194
void kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2196
void kvm_prepare_event_vectoring_exit(struct kvm_vcpu *vcpu, gpa_t gpa);
arch/x86/include/asm/kvm_host.h
2197
void kvm_prepare_unexpected_reason_exit(struct kvm_vcpu *vcpu, u64 exit_reason);
arch/x86/include/asm/kvm_host.h
2200
bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
arch/x86/include/asm/kvm_host.h
2201
int kvm_emulate_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data);
arch/x86/include/asm/kvm_host.h
2202
int kvm_emulate_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data);
arch/x86/include/asm/kvm_host.h
2203
int __kvm_emulate_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data);
arch/x86/include/asm/kvm_host.h
2204
int __kvm_emulate_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data);
arch/x86/include/asm/kvm_host.h
2205
int kvm_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data);
arch/x86/include/asm/kvm_host.h
2206
int kvm_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data);
arch/x86/include/asm/kvm_host.h
2207
int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2208
int kvm_emulate_rdmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg);
arch/x86/include/asm/kvm_host.h
2209
int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2210
int kvm_emulate_wrmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg);
arch/x86/include/asm/kvm_host.h
2211
int kvm_emulate_as_nop(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2212
int kvm_emulate_invd(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2213
int kvm_emulate_mwait(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2214
int kvm_handle_invalid_op(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2215
int kvm_emulate_monitor(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2217
int kvm_fast_pio(struct kvm_vcpu *vcpu, int size, unsigned short port, int in);
arch/x86/include/asm/kvm_host.h
2218
int kvm_emulate_cpuid(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2219
int kvm_emulate_halt(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2220
int kvm_emulate_halt_noskip(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2221
int kvm_emulate_ap_reset_hold(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2222
int kvm_emulate_wbinvd(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2224
void kvm_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
2225
void kvm_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
2226
int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg);
arch/x86/include/asm/kvm_host.h
2227
void kvm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector);
arch/x86/include/asm/kvm_host.h
2229
int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int idt_index,
arch/x86/include/asm/kvm_host.h
2232
void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0);
arch/x86/include/asm/kvm_host.h
2233
void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
2234
int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/include/asm/kvm_host.h
2235
int kvm_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3);
arch/x86/include/asm/kvm_host.h
2236
int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/include/asm/kvm_host.h
2237
int kvm_set_cr8(struct kvm_vcpu *vcpu, unsigned long cr8);
arch/x86/include/asm/kvm_host.h
2238
int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val);
arch/x86/include/asm/kvm_host.h
2239
unsigned long kvm_get_dr(struct kvm_vcpu *vcpu, int dr);
arch/x86/include/asm/kvm_host.h
2240
unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2241
void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw);
arch/x86/include/asm/kvm_host.h
2242
int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr);
arch/x86/include/asm/kvm_host.h
2243
int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2245
int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/include/asm/kvm_host.h
2246
int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/include/asm/kvm_host.h
2248
unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2249
void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
arch/x86/include/asm/kvm_host.h
2250
int kvm_emulate_rdpmc(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2252
void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr);
arch/x86/include/asm/kvm_host.h
2253
void kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code);
arch/x86/include/asm/kvm_host.h
2254
void kvm_queue_exception_p(struct kvm_vcpu *vcpu, unsigned nr, unsigned long payload);
arch/x86/include/asm/kvm_host.h
2255
void kvm_requeue_exception(struct kvm_vcpu *vcpu, unsigned int nr,
arch/x86/include/asm/kvm_host.h
2257
void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault);
arch/x86/include/asm/kvm_host.h
2258
void kvm_inject_emulated_page_fault(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2260
bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl);
arch/x86/include/asm/kvm_host.h
2261
bool kvm_require_dr(struct kvm_vcpu *vcpu, int dr);
arch/x86/include/asm/kvm_host.h
2275
void kvm_inject_nmi(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2276
int kvm_get_nr_pending_nmis(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2278
void kvm_update_dr7(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2280
bool __kvm_mmu_unprotect_gfn_and_retry(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/include/asm/kvm_host.h
2283
static inline bool kvm_mmu_unprotect_gfn_and_retry(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2292
gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2294
gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2296
gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/include/asm/kvm_host.h
2300
bool kvm_vcpu_apicv_activated(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2301
void __kvm_vcpu_update_apicv(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2319
int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
arch/x86/include/asm/kvm_host.h
2321
void kvm_mmu_print_sptes(struct kvm_vcpu *vcpu, gpa_t gpa, const char *msg);
arch/x86/include/asm/kvm_host.h
2322
void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva);
arch/x86/include/asm/kvm_host.h
2323
void kvm_mmu_invalidate_addr(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/include/asm/kvm_host.h
2325
void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid);
arch/x86/include/asm/kvm_host.h
2326
void kvm_mmu_new_pgd(struct kvm_vcpu *vcpu, gpa_t new_pgd);
arch/x86/include/asm/kvm_host.h
2360
static inline void kvm_inject_gp(struct kvm_vcpu *vcpu, u32 error_code)
arch/x86/include/asm/kvm_host.h
2394
int kvm_cpu_has_injectable_intr(struct kvm_vcpu *v);
arch/x86/include/asm/kvm_host.h
2395
int kvm_cpu_has_interrupt(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2396
int kvm_cpu_has_extint(struct kvm_vcpu *v);
arch/x86/include/asm/kvm_host.h
2397
int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2398
int kvm_cpu_get_extint(struct kvm_vcpu *v);
arch/x86/include/asm/kvm_host.h
2399
int kvm_cpu_get_interrupt(struct kvm_vcpu *v);
arch/x86/include/asm/kvm_host.h
2400
void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
arch/x86/include/asm/kvm_host.h
2417
u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc);
arch/x86/include/asm/kvm_host.h
2421
unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2422
bool kvm_is_linear_rip(struct kvm_vcpu *vcpu, unsigned long linear_rip);
arch/x86/include/asm/kvm_host.h
2428
bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2430
void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2432
void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
2434
void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2435
bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2436
extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
arch/x86/include/asm/kvm_host.h
2438
int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2439
int kvm_complete_insn_gp(struct kvm_vcpu *vcpu, int err);
arch/x86/include/asm/kvm_host.h
2443
bool kvm_vcpu_is_reset_bsp(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2444
bool kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
2453
static inline void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu)
arch/x86/include/asm/kvm_host.h
2458
static inline void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu)
arch/x86/include/asm/kvm_host.h
464
unsigned long (*get_guest_pgd)(struct kvm_vcpu *vcpu);
arch/x86/include/asm/kvm_host.h
465
u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
arch/x86/include/asm/kvm_host.h
466
int (*page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/include/asm/kvm_host.h
467
void (*inject_page_fault)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
469
gpa_t (*gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/include/asm/kvm_host.h
472
int (*sync_spte)(struct kvm_vcpu *vcpu,
arch/x86/include/asm/kvm_host.h
542
struct kvm_vcpu *vcpu;
arch/x86/include/asm/kvm_host.h
696
struct kvm_vcpu *vcpu;
arch/x86/include/asm/kvm_host.h
945
int (*complete_userspace_io)(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.c
139
static int kvm_check_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
172
static u32 kvm_apply_cpuid_pv_features_quirk(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.c
173
static void kvm_update_cpuid_runtime(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.c
176
static int kvm_cpuid_check_equal(struct kvm_vcpu *vcpu, struct kvm_cpuid_entry2 *e2,
arch/x86/kvm/cpuid.c
2048
get_out_of_range_cpuid_entry(struct kvm_vcpu *vcpu, u32 *fn_ptr, u32 index)
arch/x86/kvm/cpuid.c
207
static struct kvm_hypervisor_cpuid kvm_get_hypervisor_cpuid(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.c
2087
bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx,
arch/x86/kvm/cpuid.c
2159
int kvm_emulate_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
235
static u32 kvm_apply_cpuid_pv_features_quirk(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
258
static u64 cpuid_get_supported_xcr0(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
269
static u64 cpuid_get_supported_xss(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
280
static __always_inline void kvm_update_feature_runtime(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.c
289
static void kvm_update_cpuid_runtime(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
326
static bool kvm_cpuid_has_hyperv(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
338
static bool guest_cpuid_is_amd_or_hygon(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
375
void kvm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
471
int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
485
int cpuid_query_maxguestphyaddr(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
504
u64 kvm_vcpu_reserved_gpa_bits_raw(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.c
509
static int kvm_set_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid_entry2 *e2,
arch/x86/kvm/cpuid.c
589
int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.c
634
int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.c
657
int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
102
static __always_inline bool guest_cpuid_has(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
136
static inline bool guest_cpuid_is_amd_compatible(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
141
static inline bool guest_cpuid_is_intel_compatible(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
146
static inline int guest_cpuid_family(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
157
static inline int guest_cpuid_model(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
168
static inline bool cpuid_model_is_consistent(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
173
static inline int guest_cpuid_stepping(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
184
static inline bool supports_cpuid_fault(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
189
static inline bool cpuid_fault_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
21
void kvm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.h
229
static __always_inline bool guest_pv_has(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
238
static __always_inline void guest_cpu_cap_set(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
246
static __always_inline void guest_cpu_cap_clear(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
254
static __always_inline void guest_cpu_cap_change(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
264
static __always_inline bool guest_cpu_cap_has(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
280
static inline bool kvm_vcpu_is_legal_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
arch/x86/kvm/cpuid.h
288
static inline bool guest_has_spec_ctrl_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
296
static inline bool guest_has_pred_cmd_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
36
static inline struct kvm_cpuid_entry2 *kvm_find_cpuid_entry_index(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
43
static inline struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
53
int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
56
int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
59
int kvm_vcpu_ioctl_get_cpuid2(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
62
bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx,
arch/x86/kvm/cpuid.h
68
int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.h
69
int cpuid_query_maxguestphyaddr(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.h
70
u64 kvm_vcpu_reserved_gpa_bits_raw(struct kvm_vcpu *vcpu);
arch/x86/kvm/cpuid.h
72
static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu)
arch/x86/kvm/cpuid.h
77
static inline bool kvm_vcpu_is_legal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/x86/kvm/cpuid.h
82
static inline bool kvm_vcpu_is_legal_aligned_gpa(struct kvm_vcpu *vcpu,
arch/x86/kvm/cpuid.h
88
static inline bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/x86/kvm/debugfs.c
17
struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data;
arch/x86/kvm/debugfs.c
26
struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data;
arch/x86/kvm/debugfs.c
35
struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data;
arch/x86/kvm/debugfs.c
44
struct kvm_vcpu *vcpu = (struct kvm_vcpu *) data;
arch/x86/kvm/debugfs.c
59
void kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry)
arch/x86/kvm/hyperv.c
110
struct kvm_vcpu *vcpu = hv_synic_to_vcpu(synic);
arch/x86/kvm/hyperv.c
1336
static void __kvm_hv_xsaves_xsavec_maybe_warn(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
1363
void kvm_hv_xsaves_xsavec_maybe_warn(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
1376
static int kvm_hv_set_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 data,
arch/x86/kvm/hyperv.c
1519
static int kvm_hv_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host)
arch/x86/kvm/hyperv.c
1631
static int kvm_hv_get_msr_pw(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata,
arch/x86/kvm/hyperv.c
1687
static int kvm_hv_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata,
arch/x86/kvm/hyperv.c
1752
int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host)
arch/x86/kvm/hyperv.c
1773
int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host)
arch/x86/kvm/hyperv.c
1800
struct kvm_vcpu *vcpu;
arch/x86/kvm/hyperv.c
190
static struct kvm_vcpu *get_vcpu_by_vpidx(struct kvm *kvm, u32 vpidx)
arch/x86/kvm/hyperv.c
192
struct kvm_vcpu *vcpu = NULL;
arch/x86/kvm/hyperv.c
1933
static void hv_tlb_flush_enqueue(struct kvm_vcpu *vcpu,
arch/x86/kvm/hyperv.c
1965
int kvm_hv_vcpu_flush_tlb(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
2007
static u64 kvm_hv_flush_tlb(struct kvm_vcpu *vcpu, struct kvm_hv_hcall *hc)
arch/x86/kvm/hyperv.c
2025
struct kvm_vcpu *v;
arch/x86/kvm/hyperv.c
209
struct kvm_vcpu *vcpu;
arch/x86/kvm/hyperv.c
219
static void kvm_hv_notify_acked_sint(struct kvm_vcpu *vcpu, u32 sint)
arch/x86/kvm/hyperv.c
2209
struct kvm_vcpu *vcpu;
arch/x86/kvm/hyperv.c
2223
static u64 kvm_hv_send_ipi(struct kvm_vcpu *vcpu, struct kvm_hv_hcall *hc)
arch/x86/kvm/hyperv.c
2306
void kvm_hv_set_cpuid(struct kvm_vcpu *vcpu, bool hyperv_enabled)
arch/x86/kvm/hyperv.c
2351
int kvm_hv_set_enforce_cpuid(struct kvm_vcpu *vcpu, bool enforce)
arch/x86/kvm/hyperv.c
2372
static void kvm_hv_hypercall_set_result(struct kvm_vcpu *vcpu, u64 result)
arch/x86/kvm/hyperv.c
2385
static int kvm_hv_hypercall_complete(struct kvm_vcpu *vcpu, u64 result)
arch/x86/kvm/hyperv.c
2408
static int kvm_hv_hypercall_complete_userspace(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
2413
static u16 kvm_hvcall_signal_event(struct kvm_vcpu *vcpu, struct kvm_hv_hcall *hc)
arch/x86/kvm/hyperv.c
247
struct kvm_vcpu *vcpu = hv_synic_to_vcpu(synic);
arch/x86/kvm/hyperv.c
2528
int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
262
struct kvm_vcpu *vcpu = hv_synic_to_vcpu(synic);
arch/x86/kvm/hyperv.c
2769
int kvm_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid,
arch/x86/kvm/hyperv.c
328
static bool kvm_hv_is_syndbg_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
336
static int kvm_hv_syndbg_complete_userspace(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
346
static void syndbg_exit(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/hyperv.c
363
static int syndbg_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host)
arch/x86/kvm/hyperv.c
402
static int syndbg_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host)
arch/x86/kvm/hyperv.c
474
struct kvm_vcpu *vcpu = hv_synic_to_vcpu(synic);
arch/x86/kvm/hyperv.c
515
void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/hyperv.c
575
struct kvm_vcpu *vcpu;
arch/x86/kvm/hyperv.c
594
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
605
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
693
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
723
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
762
struct kvm_vcpu *vcpu = hv_synic_to_vcpu(synic);
arch/x86/kvm/hyperv.c
818
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
838
struct kvm_vcpu *vcpu = hv_stimer_to_vcpu(stimer);
arch/x86/kvm/hyperv.c
867
void kvm_hv_process_stimers(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
900
void kvm_hv_vcpu_uninit(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
915
bool kvm_hv_assist_page_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
928
int kvm_hv_get_assist_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
963
int kvm_hv_vcpu_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.c
994
int kvm_hv_activate_synic(struct kvm_vcpu *vcpu, bool dont_zero_synic_pages)
arch/x86/kvm/hyperv.h
103
int kvm_hv_hypercall(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
108
void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector);
arch/x86/kvm/hyperv.h
109
int kvm_hv_activate_synic(struct kvm_vcpu *vcpu, bool dont_zero_synic_pages);
arch/x86/kvm/hyperv.h
111
static inline bool kvm_hv_synic_has_vector(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/hyperv.h
116
static inline bool kvm_hv_synic_auto_eoi_set(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/hyperv.h
122
void kvm_hv_vcpu_uninit(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
124
bool kvm_hv_assist_page_enabled(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
125
int kvm_hv_get_assist_page(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
127
static inline struct kvm_vcpu_hv_stimer *to_hv_stimer(struct kvm_vcpu *vcpu,
arch/x86/kvm/hyperv.h
133
static inline struct kvm_vcpu *hv_stimer_to_vcpu(struct kvm_vcpu_hv_stimer *stimer)
arch/x86/kvm/hyperv.h
142
static inline bool kvm_hv_has_stimer_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
157
static inline bool kvm_hv_invtsc_suppressed(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
180
void kvm_hv_process_stimers(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
186
void kvm_hv_xsaves_xsavec_maybe_warn(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
190
int kvm_hv_vcpu_init(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
191
void kvm_hv_set_cpuid(struct kvm_vcpu *vcpu, bool hyperv_enabled);
arch/x86/kvm/hyperv.h
192
int kvm_hv_set_enforce_cpuid(struct kvm_vcpu *vcpu, bool enforce);
arch/x86/kvm/hyperv.h
194
int kvm_get_hv_cpuid(struct kvm_vcpu *vcpu, struct kvm_cpuid2 *cpuid,
arch/x86/kvm/hyperv.h
197
static inline struct kvm_vcpu_hv_tlb_flush_fifo *kvm_hv_get_tlb_flush_fifo(struct kvm_vcpu *vcpu,
arch/x86/kvm/hyperv.h
207
static inline void kvm_hv_vcpu_purge_flush_tlb(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
219
static inline bool guest_hv_cpuid_has_l2_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
227
static inline bool kvm_hv_is_tlb_flush_hcall(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
244
static inline int kvm_hv_verify_vp_assist(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
255
static inline void kvm_hv_nested_transtion_tlb_flush(struct kvm_vcpu *vcpu,
arch/x86/kvm/hyperv.h
268
int kvm_hv_vcpu_flush_tlb(struct kvm_vcpu *vcpu);
arch/x86/kvm/hyperv.h
273
static inline void kvm_hv_xsaves_xsavec_maybe_warn(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/hyperv.h
276
static inline int kvm_hv_vcpu_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
280
static inline void kvm_hv_vcpu_uninit(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/hyperv.h
281
static inline bool kvm_hv_hypercall_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
285
static inline int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
289
static inline void kvm_hv_vcpu_purge_flush_tlb(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/hyperv.h
290
static inline bool kvm_hv_synic_has_vector(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/hyperv.h
294
static inline bool kvm_hv_synic_auto_eoi_set(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/hyperv.h
298
static inline void kvm_hv_synic_send_eoi(struct kvm_vcpu *vcpu, int vector) {}
arch/x86/kvm/hyperv.h
299
static inline bool kvm_hv_invtsc_suppressed(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
303
static inline void kvm_hv_set_cpuid(struct kvm_vcpu *vcpu, bool hyperv_enabled) {}
arch/x86/kvm/hyperv.h
304
static inline bool kvm_hv_has_stimer_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
308
static inline bool kvm_hv_is_tlb_flush_hcall(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
312
static inline bool guest_hv_cpuid_has_l2_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
316
static inline int kvm_hv_verify_vp_assist(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
320
static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
324
static inline void kvm_hv_nested_transtion_tlb_flush(struct kvm_vcpu *vcpu, bool tdp_enabled) {}
arch/x86/kvm/hyperv.h
64
static inline struct kvm_vcpu_hv *to_hv_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
69
static inline struct kvm_vcpu_hv_synic *to_hv_synic(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
76
static inline struct kvm_vcpu *hv_synic_to_vcpu(struct kvm_vcpu_hv_synic *synic)
arch/x86/kvm/hyperv.h
83
static inline struct kvm_hv_syndbg *to_hv_syndbg(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
88
static inline u32 kvm_hv_get_vpindex(struct kvm_vcpu *vcpu)
arch/x86/kvm/hyperv.h
95
int kvm_hv_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data, bool host);
arch/x86/kvm/hyperv.h
96
int kvm_hv_get_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host);
arch/x86/kvm/hyperv.h
98
static inline bool kvm_hv_hypercall_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/i8254.c
218
void __kvm_migrate_pit_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/i8254.c
244
struct kvm_vcpu *vcpu;
arch/x86/kvm/i8254.c
438
static int pit_ioport_write(struct kvm_vcpu *vcpu,
arch/x86/kvm/i8254.c
513
static int pit_ioport_read(struct kvm_vcpu *vcpu,
arch/x86/kvm/i8254.c
583
static int speaker_ioport_write(struct kvm_vcpu *vcpu,
arch/x86/kvm/i8254.c
603
static int speaker_ioport_read(struct kvm_vcpu *vcpu,
arch/x86/kvm/i8259.c
271
struct kvm_vcpu *vcpu;
arch/x86/kvm/i8259.c
516
static int picdev_master_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
523
static int picdev_master_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
530
static int picdev_slave_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
537
static int picdev_slave_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
544
static int picdev_elcr_write(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
551
static int picdev_elcr_read(struct kvm_vcpu *vcpu, struct kvm_io_device *dev,
arch/x86/kvm/i8259.c
56
struct kvm_vcpu *vcpu;
arch/x86/kvm/ioapic.c
121
void kvm_rtc_eoi_tracking_restore_one(struct kvm_vcpu *vcpu)
arch/x86/kvm/ioapic.c
132
struct kvm_vcpu *vcpu;
arch/x86/kvm/ioapic.c
143
static void rtc_irq_eoi(struct kvm_ioapic *ioapic, struct kvm_vcpu *vcpu,
arch/x86/kvm/ioapic.c
168
struct kvm_vcpu *vcpu;
arch/x86/kvm/ioapic.c
259
void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu, ulong *ioapic_handled_vectors)
arch/x86/kvm/ioapic.c
538
static void kvm_ioapic_update_eoi_one(struct kvm_vcpu *vcpu,
arch/x86/kvm/ioapic.c
585
void kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, int vector, int trigger_mode)
arch/x86/kvm/ioapic.c
613
static int ioapic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
arch/x86/kvm/ioapic.c
653
static int ioapic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
arch/x86/kvm/ioapic.c
91
static void __rtc_irq_eoi_tracking_restore_one(struct kvm_vcpu *vcpu)
arch/x86/kvm/ioapic.h
104
void kvm_rtc_eoi_tracking_restore_one(struct kvm_vcpu *vcpu);
arch/x86/kvm/ioapic.h
105
void kvm_ioapic_update_eoi(struct kvm_vcpu *vcpu, int vector,
arch/x86/kvm/ioapic.h
114
void kvm_ioapic_scan_entry(struct kvm_vcpu *vcpu,
arch/x86/kvm/ioapic.h
12
struct kvm_vcpu;
arch/x86/kvm/ioapic.h
123
void kvm_scan_ioapic_routes(struct kvm_vcpu *vcpu,
arch/x86/kvm/ioapic.h
125
void kvm_scan_ioapic_irq(struct kvm_vcpu *vcpu, u32 dest_id, u16 dest_mode,
arch/x86/kvm/irq.c
112
int kvm_cpu_has_interrupt(struct kvm_vcpu *v)
arch/x86/kvm/irq.c
128
int kvm_cpu_get_extint(struct kvm_vcpu *v)
arch/x86/kvm/irq.c
156
int kvm_cpu_get_interrupt(struct kvm_vcpu *v)
arch/x86/kvm/irq.c
169
void kvm_inject_pending_timer_irqs(struct kvm_vcpu *vcpu)
arch/x86/kvm/irq.c
177
void __kvm_migrate_timers(struct kvm_vcpu *vcpu)
arch/x86/kvm/irq.c
27
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/irq.c
357
void kvm_scan_ioapic_irq(struct kvm_vcpu *vcpu, u32 dest_id, u16 dest_mode,
arch/x86/kvm/irq.c
384
void kvm_scan_ioapic_routes(struct kvm_vcpu *vcpu,
arch/x86/kvm/irq.c
42
static int pending_userspace_extint(struct kvm_vcpu *v)
arch/x86/kvm/irq.c
431
struct kvm_vcpu *vcpu = NULL;
arch/x86/kvm/irq.c
47
static int get_userspace_extint(struct kvm_vcpu *vcpu)
arch/x86/kvm/irq.c
59
int kvm_cpu_has_extint(struct kvm_vcpu *v)
arch/x86/kvm/irq.c
96
int kvm_cpu_has_injectable_intr(struct kvm_vcpu *v)
arch/x86/kvm/irq.h
115
void kvm_inject_pending_timer_irqs(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
116
void kvm_inject_apic_timer_irqs(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
117
void kvm_apic_nmi_wd_deliver(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
118
void __kvm_migrate_apic_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
119
void __kvm_migrate_pit_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
120
void __kvm_migrate_timers(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
122
int apic_has_pending_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/irq.h
28
struct kvm_vcpu;
arch/x86/kvm/kvm_cache_regs.h
101
static __always_inline bool kvm_register_test_and_mark_available(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
113
static inline unsigned long kvm_register_read_raw(struct kvm_vcpu *vcpu, int reg)
arch/x86/kvm/kvm_cache_regs.h
124
static inline void kvm_register_write_raw(struct kvm_vcpu *vcpu, int reg,
arch/x86/kvm/kvm_cache_regs.h
134
static inline unsigned long kvm_rip_read(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
139
static inline void kvm_rip_write(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/kvm_cache_regs.h
144
static inline unsigned long kvm_rsp_read(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
149
static inline void kvm_rsp_write(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/kvm_cache_regs.h
154
static inline u64 kvm_pdptr_read(struct kvm_vcpu *vcpu, int index)
arch/x86/kvm/kvm_cache_regs.h
164
static inline void kvm_pdptr_write(struct kvm_vcpu *vcpu, int index, u64 value)
arch/x86/kvm/kvm_cache_regs.h
169
static inline ulong kvm_read_cr0_bits(struct kvm_vcpu *vcpu, ulong mask)
arch/x86/kvm/kvm_cache_regs.h
178
static __always_inline bool kvm_is_cr0_bit_set(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
186
static inline ulong kvm_read_cr0(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
191
static inline ulong kvm_read_cr4_bits(struct kvm_vcpu *vcpu, ulong mask)
arch/x86/kvm/kvm_cache_regs.h
20
static __always_inline unsigned long kvm_##lname##_read(struct kvm_vcpu *vcpu)\
arch/x86/kvm/kvm_cache_regs.h
200
static __always_inline bool kvm_is_cr4_bit_set(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
208
static inline ulong kvm_read_cr3(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
215
static inline ulong kvm_read_cr4(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
220
static inline u64 kvm_read_edx_eax(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
226
static inline void enter_guest_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
232
static inline void leave_guest_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
24
static __always_inline void kvm_##lname##_write(struct kvm_vcpu *vcpu, \
arch/x86/kvm/kvm_cache_regs.h
244
static inline bool is_guest_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_cache_regs.h
66
static inline bool kvm_register_is_available(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
73
static inline bool kvm_register_is_dirty(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
80
static inline void kvm_register_mark_available(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_cache_regs.h
87
static inline void kvm_register_mark_dirty(struct kvm_vcpu *vcpu,
arch/x86/kvm/kvm_onhyperv.c
112
void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp)
arch/x86/kvm/kvm_onhyperv.c
41
struct kvm_vcpu *vcpu;
arch/x86/kvm/kvm_onhyperv.h
12
void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp);
arch/x86/kvm/kvm_onhyperv.h
13
static inline hpa_t hv_get_partition_assist_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/kvm_onhyperv.h
39
static inline void hv_track_root_tdp(struct kvm_vcpu *vcpu, hpa_t root_tdp)
arch/x86/kvm/lapic.c
100
bool kvm_apic_pending_eoi(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/lapic.c
1063
static u32 kvm_apic_mda(struct kvm_vcpu *vcpu, unsigned int dest_id,
arch/x86/kvm/lapic.c
1075
bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source,
arch/x86/kvm/lapic.c
1140
static int kvm_apic_compare_prio(struct kvm_vcpu *vcpu1, struct kvm_vcpu *vcpu2)
arch/x86/kvm/lapic.c
1279
struct kvm_vcpu **dest_vcpu)
arch/x86/kvm/lapic.c
1307
struct kvm_vcpu **dest_vcpu)
arch/x86/kvm/lapic.c
1311
struct kvm_vcpu *vcpu;
arch/x86/kvm/lapic.c
1339
struct kvm_vcpu *vcpu, *lowest = NULL;
arch/x86/kvm/lapic.c
1401
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/lapic.c
1505
struct kvm_vcpu *vcpu;
arch/x86/kvm/lapic.c
1612
void kvm_apic_set_eoi_accelerated(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/lapic.c
1676
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/lapic.c
1812
static int apic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
arch/x86/kvm/lapic.c
188
static bool kvm_can_post_timer_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
1901
static bool lapic_timer_int_injected(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
1928
static inline void __wait_lapic_expire(struct kvm_vcpu *vcpu, u64 guest_cycles)
arch/x86/kvm/lapic.c
194
static bool kvm_can_use_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
1948
static inline void adjust_lapic_timer_advance(struct kvm_vcpu *vcpu,
arch/x86/kvm/lapic.c
1977
static void __kvm_wait_lapic_expire(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2000
void kvm_wait_lapic_expire(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
201
static bool kvm_use_posted_timer_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2025
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/lapic.c
2067
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/lapic.c
2233
bool kvm_lapic_hv_timer_in_use(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2252
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/lapic.c
2321
void kvm_lapic_expired_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2342
void kvm_lapic_switch_to_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2347
void kvm_lapic_switch_to_sw_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2358
void kvm_lapic_restart_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
252
struct kvm_vcpu *vcpu,
arch/x86/kvm/lapic.c
2556
static int apic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
arch/x86/kvm/lapic.c
2589
void kvm_lapic_set_eoi(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2656
void kvm_apic_write_nodecode(struct kvm_vcpu *vcpu, u32 offset)
arch/x86/kvm/lapic.c
2678
void kvm_free_lapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2706
u64 kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2716
void kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/lapic.c
2728
void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8)
arch/x86/kvm/lapic.c
2733
u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2742
static void __kvm_apic_set_base(struct kvm_vcpu *vcpu, u64 value)
arch/x86/kvm/lapic.c
2790
int kvm_apic_set_base(struct kvm_vcpu *vcpu, u64 value, bool host_initiated)
arch/x86/kvm/lapic.c
2816
void kvm_apic_update_apicv(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2872
void kvm_inhibit_apic_access_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
2907
void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/lapic.c
2994
int apic_has_pending_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3024
void kvm_apic_nmi_wd_deliver(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3052
int kvm_create_lapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3116
int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3132
int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3144
void kvm_inject_apic_timer_irqs(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3154
void kvm_apic_ack_interrupt(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/lapic.c
3191
static int kvm_apic_state_fixup(struct kvm_vcpu *vcpu,
arch/x86/kvm/lapic.c
3243
int kvm_apic_get_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s)
arch/x86/kvm/lapic.c
3256
int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s)
arch/x86/kvm/lapic.c
328
struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3300
void __kvm_migrate_apic_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3320
static void apic_sync_pv_eoi_from_guest(struct kvm_vcpu *vcpu,
arch/x86/kvm/lapic.c
3343
void kvm_lapic_sync_from_vapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3366
static void apic_sync_pv_eoi_to_guest(struct kvm_vcpu *vcpu,
arch/x86/kvm/lapic.c
3386
void kvm_lapic_sync_to_vapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
3410
int kvm_lapic_set_vapic_addr(struct kvm_vcpu *vcpu, gpa_t vapic_addr)
arch/x86/kvm/lapic.c
3460
int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/lapic.c
3471
int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data)
arch/x86/kvm/lapic.c
3482
int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 reg, u64 data)
arch/x86/kvm/lapic.c
3490
int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 reg, u64 *data)
arch/x86/kvm/lapic.c
3498
int kvm_lapic_set_pv_eoi(struct kvm_vcpu *vcpu, u64 data, unsigned long len)
arch/x86/kvm/lapic.c
3524
int kvm_apic_accept_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
408
struct kvm_vcpu *vcpu;
arch/x86/kvm/lapic.c
599
static inline int kvm_apic_calc_nr_lvt_entries(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
604
void kvm_apic_set_version(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
621
void kvm_apic_after_set_mcg_cap(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
702
bool kvm_apic_update_irr(struct kvm_vcpu *vcpu, unsigned long *pir, int *max_irr)
arch/x86/kvm/lapic.c
742
void kvm_apic_clear_irr(struct kvm_vcpu *vcpu, int vec)
arch/x86/kvm/lapic.c
814
int kvm_lapic_find_highest_irr(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
829
int kvm_apic_set_irq(struct kvm_vcpu *vcpu, struct kvm_lapic_irq *irq,
arch/x86/kvm/lapic.c
842
struct kvm_vcpu *vcpu;
arch/x86/kvm/lapic.c
891
static int pv_eoi_put_user(struct kvm_vcpu *vcpu, u8 val)
arch/x86/kvm/lapic.c
898
static int pv_eoi_get_user(struct kvm_vcpu *vcpu, u8 *val)
arch/x86/kvm/lapic.c
905
static inline bool pv_eoi_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
910
static void pv_eoi_set_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
918
static bool pv_eoi_test_and_clr_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.c
983
void kvm_apic_update_ppr(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
100
void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event);
arch/x86/kvm/lapic.h
101
u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
102
void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8);
arch/x86/kvm/lapic.h
103
void kvm_lapic_set_eoi(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
104
void kvm_apic_set_version(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
105
void kvm_apic_after_set_mcg_cap(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
106
bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source,
arch/x86/kvm/lapic.h
108
void kvm_apic_clear_irr(struct kvm_vcpu *vcpu, int vec);
arch/x86/kvm/lapic.h
110
bool kvm_apic_update_irr(struct kvm_vcpu *vcpu, unsigned long *pir, int *max_irr);
arch/x86/kvm/lapic.h
111
void kvm_apic_update_ppr(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
112
int kvm_apic_set_irq(struct kvm_vcpu *vcpu, struct kvm_lapic_irq *irq,
arch/x86/kvm/lapic.h
115
void kvm_apic_update_apicv(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
117
void kvm_inhibit_apic_access_page(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
134
int kvm_apic_set_base(struct kvm_vcpu *vcpu, u64 value, bool host_initiated);
arch/x86/kvm/lapic.h
135
int kvm_apic_get_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s);
arch/x86/kvm/lapic.h
136
int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s);
arch/x86/kvm/lapic.h
137
int kvm_lapic_find_highest_irr(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
139
u64 kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
140
void kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu, u64 data);
arch/x86/kvm/lapic.h
142
void kvm_apic_write_nodecode(struct kvm_vcpu *vcpu, u32 offset);
arch/x86/kvm/lapic.h
143
void kvm_apic_set_eoi_accelerated(struct kvm_vcpu *vcpu, int vector);
arch/x86/kvm/lapic.h
145
int kvm_lapic_set_vapic_addr(struct kvm_vcpu *vcpu, gpa_t vapic_addr);
arch/x86/kvm/lapic.h
146
void kvm_lapic_sync_from_vapic(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
147
void kvm_lapic_sync_to_vapic(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
150
int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data);
arch/x86/kvm/lapic.h
151
int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
arch/x86/kvm/lapic.h
153
int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data);
arch/x86/kvm/lapic.h
154
int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data);
arch/x86/kvm/lapic.h
156
int kvm_lapic_set_pv_eoi(struct kvm_vcpu *vcpu, u64 data, unsigned long len);
arch/x86/kvm/lapic.h
178
static inline bool lapic_in_kernel(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
203
static inline bool kvm_apic_present(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
208
static inline int kvm_lapic_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
218
static inline bool kvm_vcpu_apicv_active(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
223
static inline bool kvm_apic_has_pending_init_or_sipi(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
228
static inline bool kvm_apic_init_sipi_allowed(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
234
static inline int kvm_lapic_latched_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
239
bool kvm_apic_pending_eoi(struct kvm_vcpu *vcpu, int vector);
arch/x86/kvm/lapic.h
243
void kvm_wait_lapic_expire(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
249
struct kvm_vcpu **dest_vcpu);
arch/x86/kvm/lapic.h
250
void kvm_lapic_switch_to_sw_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
251
void kvm_lapic_switch_to_hv_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
252
void kvm_lapic_expired_hv_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
253
bool kvm_lapic_hv_timer_in_use(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
254
void kvm_lapic_restart_hv_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
261
static inline enum lapic_mode kvm_get_apic_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/lapic.h
67
struct kvm_vcpu *vcpu;
arch/x86/kvm/lapic.h
93
int kvm_create_lapic(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
94
void kvm_free_lapic(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
96
int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
97
void kvm_apic_ack_interrupt(struct kvm_vcpu *vcpu, int vector);
arch/x86/kvm/lapic.h
98
int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu);
arch/x86/kvm/lapic.h
99
int kvm_apic_accept_events(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
100
void kvm_mmu_free_obsolete_roots(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
101
void kvm_mmu_sync_roots(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
102
void kvm_mmu_sync_prev_roots(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
103
void kvm_mmu_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
arch/x86/kvm/mmu.h
106
static inline int kvm_mmu_reload(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu.h
126
static inline unsigned long kvm_get_pcid(struct kvm_vcpu *vcpu, gpa_t cr3)
arch/x86/kvm/mmu.h
135
static inline unsigned long kvm_get_active_pcid(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu.h
140
static inline unsigned long kvm_get_active_cr3_lam_bits(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu.h
148
static inline void kvm_mmu_load_pgd(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu.h
159
static inline void kvm_mmu_refresh_passthrough_bits(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu.h
185
static inline u8 permission_fault(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu.h
258
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn);
arch/x86/kvm/mmu.h
291
gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u64 access,
arch/x86/kvm/mmu.h
294
static inline gpa_t kvm_translate_gpa(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu.h
86
void kvm_init_mmu(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
87
void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, unsigned long cr0,
arch/x86/kvm/mmu.h
89
void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly,
arch/x86/kvm/mmu.h
92
bool kvm_can_do_async_pf(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
93
int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code,
arch/x86/kvm/mmu.h
95
void __kvm_mmu_refresh_passthrough_bits(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu.h
98
int kvm_mmu_load(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu.h
99
void kvm_mmu_unload(struct kvm_vcpu *vcpu);
arch/x86/kvm/mmu/mmu.c
1497
static bool kvm_vcpu_write_protect_gfn(struct kvm_vcpu *vcpu, u64 gfn)
arch/x86/kvm/mmu/mmu.c
1703
static void rmap_add(struct kvm_vcpu *vcpu, const struct kvm_memory_slot *slot,
arch/x86/kvm/mmu/mmu.c
2021
static bool kvm_sync_page_check(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2054
static int kvm_sync_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i)
arch/x86/kvm/mmu/mmu.c
2063
static int __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2091
static int kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2197
static int mmu_sync_children(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2256
struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2377
struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2404
static struct kvm_mmu_page *kvm_mmu_get_shadow_page(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2463
static struct kvm_mmu_page *kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
247
static struct kvm_mmu_role_regs vcpu_to_role_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
2477
struct kvm_vcpu *vcpu, hpa_t root,
arch/x86/kvm/mmu/mmu.c
2506
struct kvm_vcpu *vcpu, u64 addr)
arch/x86/kvm/mmu/mmu.c
2574
static void link_shadow_page(struct kvm_vcpu *vcpu, u64 *sptep,
arch/x86/kvm/mmu/mmu.c
258
static unsigned long get_guest_cr3(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
2580
static void validate_direct_spte(struct kvm_vcpu *vcpu, u64 *sptep,
arch/x86/kvm/mmu/mmu.c
263
static inline unsigned long kvm_mmu_get_guest_pgd(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2825
static int make_mmu_pages_available(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
2868
bool __kvm_mmu_unprotect_gfn_and_retry(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/kvm/mmu/mmu.c
292
static void mark_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, u64 gfn,
arch/x86/kvm/mmu/mmu.c
3030
static int mmu_set_spte(struct kvm_vcpu *vcpu, struct kvm_memory_slot *slot,
arch/x86/kvm/mmu/mmu.c
3106
static bool kvm_mmu_prefetch_sptes(struct kvm_vcpu *vcpu, gfn_t gfn, u64 *sptep,
arch/x86/kvm/mmu/mmu.c
3143
static bool direct_pte_prefetch_many(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
3153
static void __direct_pte_prefetch(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
316
static bool check_mmio_spte(struct kvm_vcpu *vcpu, u64 spte)
arch/x86/kvm/mmu/mmu.c
3179
static void direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep)
arch/x86/kvm/mmu/mmu.c
3380
void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3436
static int direct_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3487
static int kvm_handle_error_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3510
static int kvm_handle_noslot_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
3609
static bool fast_pf_fix_direct_spte(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
3643
static u64 *fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gpa_t gpa, u64 *spte)
arch/x86/kvm/mmu/mmu.c
3660
static int fast_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
3900
static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, int quadrant,
arch/x86/kvm/mmu/mmu.c
3918
static int mmu_alloc_direct_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4057
static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4175
static int mmu_alloc_special_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4282
void kvm_mmu_sync_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4323
void kvm_mmu_sync_prev_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4336
static gpa_t nonpaging_gva_to_gpa(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu/mmu.c
4345
static bool mmio_info_in_cache(struct kvm_vcpu *vcpu, u64 addr, bool direct)
arch/x86/kvm/mmu/mmu.c
4366
static int get_walk(struct kvm_vcpu *vcpu, u64 addr, u64 *sptes, int *root_level)
arch/x86/kvm/mmu/mmu.c
4385
static int get_sptes_lockless(struct kvm_vcpu *vcpu, u64 addr, u64 *sptes,
arch/x86/kvm/mmu/mmu.c
4402
static bool get_mmio_spte(struct kvm_vcpu *vcpu, u64 addr, u64 *sptep)
arch/x86/kvm/mmu/mmu.c
4443
static int handle_mmio_page_fault(struct kvm_vcpu *vcpu, u64 addr, bool direct)
arch/x86/kvm/mmu/mmu.c
4477
static bool page_fault_handle_page_track(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4496
static void shadow_page_table_clear_flood(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/mmu/mmu.c
4507
static u32 alloc_apf_token(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
4518
static bool kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4536
void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work)
arch/x86/kvm/mmu/mmu.c
4567
static void kvm_mmu_finish_page_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4574
static int kvm_mmu_faultin_pfn_gmem(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4597
static int __kvm_mmu_faultin_pfn(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4642
static int kvm_mmu_faultin_pfn(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4765
static bool is_page_fault_stale(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4794
static int direct_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4835
static int nonpaging_page_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4843
int kvm_handle_page_fault(struct kvm_vcpu *vcpu, u64 error_code,
arch/x86/kvm/mmu/mmu.c
4888
static int kvm_tdp_mmu_page_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
4923
int kvm_tdp_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/mmu.c
4933
static int kvm_tdp_page_prefault(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/mmu/mmu.c
4977
long kvm_arch_vcpu_pre_fault_memory(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5036
int kvm_tdp_mmu_map_private_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn)
arch/x86/kvm/mmu/mmu.c
5207
void kvm_mmu_new_pgd(struct kvm_vcpu *vcpu, gpa_t new_pgd)
arch/x86/kvm/mmu/mmu.c
5253
static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn,
arch/x86/kvm/mmu/mmu.c
5371
static void reset_guest_rsvds_bits_mask(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5420
static void reset_rsvds_bits_mask_ept(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5438
static void reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
555
static inline bool is_tdp_mmu_active(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
560
static void walk_shadow_page_lockless_begin(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
5673
static void reset_guest_paging_metadata(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5698
static union kvm_cpu_role kvm_calc_cpu_role(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5738
void __kvm_mmu_refresh_passthrough_bits(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5753
static inline int kvm_mmu_get_tdp_level(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
5779
kvm_calc_tdp_mmu_root_page_role(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
579
static void walk_shadow_page_lockless_end(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
5797
static void init_kvm_tdp_mmu(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5826
static void shadow_mmu_init_context(struct kvm_vcpu *vcpu, struct kvm_mmu *context,
arch/x86/kvm/mmu/mmu.c
5848
static void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5873
void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, unsigned long cr0,
arch/x86/kvm/mmu/mmu.c
5900
kvm_calc_shadow_ept_root_page_role(struct kvm_vcpu *vcpu, bool accessed_dirty,
arch/x86/kvm/mmu/mmu.c
5924
void kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, bool execonly,
arch/x86/kvm/mmu/mmu.c
594
static int mmu_topup_memory_caches(struct kvm_vcpu *vcpu, bool maybe_indirect)
arch/x86/kvm/mmu/mmu.c
5953
static void init_kvm_softmmu(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
5965
static void init_kvm_nested_mmu(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
6004
void kvm_init_mmu(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6018
void kvm_mmu_after_set_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6043
void kvm_mmu_reset_context(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6050
int kvm_mmu_load(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6083
void kvm_mmu_unload(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6138
void kvm_mmu_free_obsolete_roots(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6145
static u64 mmu_pte_write_fetch_gpte(struct kvm_vcpu *vcpu, gpa_t *gpa,
arch/x86/kvm/mmu/mmu.c
623
static void mmu_free_memory_caches(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
6244
void kvm_mmu_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
arch/x86/kvm/mmu/mmu.c
6305
static int kvm_mmu_write_protect_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/kvm/mmu/mmu.c
6398
int noinline kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
arch/x86/kvm/mmu/mmu.c
6471
void kvm_mmu_print_sptes(struct kvm_vcpu *vcpu, gpa_t gpa, const char *msg)
arch/x86/kvm/mmu/mmu.c
6487
static void __kvm_mmu_invalidate_addr(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu/mmu.c
6524
void kvm_mmu_invalidate_addr(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu/mmu.c
6553
void kvm_mmu_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/mmu/mmu.c
6571
void kvm_mmu_invpcid_gva(struct kvm_vcpu *vcpu, gva_t gva, unsigned long pcid)
arch/x86/kvm/mmu/mmu.c
6632
static int __kvm_mmu_create(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu)
arch/x86/kvm/mmu/mmu.c
6686
int kvm_mmu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
7618
void kvm_mmu_destroy(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/mmu.c
842
static struct kvm_memory_slot *gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu_internal.h
160
static inline void kvm_mmu_alloc_external_spt(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
293
int kvm_tdp_page_fault(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/mmu/mmu_internal.h
332
static inline void kvm_mmu_prepare_memory_fault_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu_internal.h
340
static inline int kvm_mmu_do_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/kvm/mmu/mmu_internal.h
406
void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/mmu/mmutrace.h
260
TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/page_track.h
50
static inline void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/mmu/paging_tmpl.h
151
static bool FNAME(prefetch_invalid_gpte)(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/paging_tmpl.h
197
static int FNAME(update_accessed_dirty_bits)(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/paging_tmpl.h
259
static inline unsigned FNAME(gpte_pkeys)(struct kvm_vcpu *vcpu, u64 gpte)
arch/x86/kvm/mmu/paging_tmpl.h
303
struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu/paging_tmpl.h
525
struct kvm_vcpu *vcpu, gpa_t addr, u64 access)
arch/x86/kvm/mmu/paging_tmpl.h
532
FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/paging_tmpl.h
548
static bool FNAME(gpte_changed)(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/paging_tmpl.h
571
static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, struct guest_walker *gw,
arch/x86/kvm/mmu/paging_tmpl.h
614
static int FNAME(fetch)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault,
arch/x86/kvm/mmu/paging_tmpl.h
765
static int FNAME(page_fault)(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/paging_tmpl.h
860
static gpa_t FNAME(gva_to_gpa)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu,
arch/x86/kvm/mmu/paging_tmpl.h
895
static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i)
arch/x86/kvm/mmu/spte.c
141
static void kvm_track_host_mmio_mapping(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/spte.c
186
bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/spte.c
92
u64 make_mmio_spte(struct kvm_vcpu *vcpu, u64 gfn, unsigned int access)
arch/x86/kvm/mmu/spte.h
293
static inline bool kvm_vcpu_can_access_host_mmio(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/spte.h
543
bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/spte.h
552
u64 make_mmio_spte(struct kvm_vcpu *vcpu, u64 gfn, unsigned int access);
arch/x86/kvm/mmu/tdp_mmu.c
1168
static int tdp_mmu_map_handle_target_level(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/tdp_mmu.c
1263
int kvm_tdp_mmu_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault)
arch/x86/kvm/mmu/tdp_mmu.c
1939
int kvm_tdp_mmu_get_walk(struct kvm_vcpu *vcpu, u64 addr, u64 *sptes,
arch/x86/kvm/mmu/tdp_mmu.c
1968
u64 *kvm_tdp_mmu_fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.c
214
static struct kvm_mmu_page *tdp_mmu_alloc_sp(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/tdp_mmu.c
253
void kvm_tdp_mmu_alloc_root(struct kvm_vcpu *vcpu, bool mirror)
arch/x86/kvm/mmu/tdp_mmu.h
111
int kvm_tdp_mmu_get_walk(struct kvm_vcpu *vcpu, u64 addr, u64 *sptes,
arch/x86/kvm/mmu/tdp_mmu.h
113
u64 *kvm_tdp_mmu_fast_pf_get_last_sptep(struct kvm_vcpu *vcpu, gfn_t gfn,
arch/x86/kvm/mmu/tdp_mmu.h
13
void kvm_tdp_mmu_alloc_root(struct kvm_vcpu *vcpu, bool private);
arch/x86/kvm/mmu/tdp_mmu.h
48
static inline struct kvm_mmu_page *tdp_mmu_get_root_for_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/tdp_mmu.h
57
static inline struct kvm_mmu_page *tdp_mmu_get_root(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/tdp_mmu.h
74
int kvm_tdp_mmu_map(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);
arch/x86/kvm/mtrr.c
111
int kvm_mtrr_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata)
arch/x86/kvm/mtrr.c
24
static u64 *find_mtrr(struct kvm_vcpu *vcpu, unsigned int msr)
arch/x86/kvm/mtrr.c
61
static bool kvm_mtrr_valid(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/mtrr.c
96
int kvm_mtrr_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/pmu.c
1007
void kvm_pmu_cleanup(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1029
void kvm_pmu_destroy(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1048
struct kvm_vcpu *vcpu = pmc->vcpu;
arch/x86/kvm/pmu.c
1102
static void kvm_pmu_trigger_event(struct kvm_vcpu *vcpu,
arch/x86/kvm/pmu.c
1131
void kvm_pmu_instruction_retired(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1137
void kvm_pmu_branch_retired(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1229
struct kvm_vcpu *vcpu;
arch/x86/kvm/pmu.c
1301
static void kvm_pmu_load_guest_pmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1327
void kvm_mediated_pmu_load(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1359
static void kvm_pmu_put_guest_pmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
1388
void kvm_mediated_pmu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
165
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
arch/x86/kvm/pmu.c
634
void kvm_pmu_handle_event(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
674
int kvm_pmu_check_rdpmc_early(struct kvm_vcpu *vcpu, unsigned int idx)
arch/x86/kvm/pmu.c
701
static int kvm_pmu_rdpmc_vmware(struct kvm_vcpu *vcpu, unsigned idx, u64 *data)
arch/x86/kvm/pmu.c
724
int kvm_pmu_rdpmc(struct kvm_vcpu *vcpu, unsigned idx, u64 *data)
arch/x86/kvm/pmu.c
749
static bool kvm_need_any_pmc_intercept(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
765
bool kvm_need_perf_global_ctrl_intercept(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
772
bool kvm_need_rdpmc_intercept(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
789
void kvm_pmu_deliver_pmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
797
bool kvm_pmu_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/pmu.c
811
static void kvm_pmu_mark_pmc_in_use(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/pmu.c
820
int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/pmu.c
846
int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/pmu.c
915
static void kvm_pmu_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
947
void kvm_pmu_refresh(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.c
998
void kvm_pmu_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.h
10
#define pmu_to_vcpu(pmu) (container_of((pmu), struct kvm_vcpu, arch.pmu))
arch/x86/kvm/pmu.h
237
static inline bool kvm_pmu_is_fastpath_emulation_allowed(struct kvm_vcpu *vcpu)
arch/x86/kvm/pmu.h
247
void kvm_pmu_deliver_pmi(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
248
void kvm_pmu_handle_event(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
249
int kvm_pmu_rdpmc(struct kvm_vcpu *vcpu, unsigned pmc, u64 *data);
arch/x86/kvm/pmu.h
250
int kvm_pmu_check_rdpmc_early(struct kvm_vcpu *vcpu, unsigned int idx);
arch/x86/kvm/pmu.h
251
bool kvm_pmu_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr);
arch/x86/kvm/pmu.h
252
int kvm_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
253
int kvm_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
254
void kvm_pmu_refresh(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
255
void kvm_pmu_init(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
256
void kvm_pmu_cleanup(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
257
void kvm_pmu_destroy(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
259
void kvm_pmu_instruction_retired(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
260
void kvm_pmu_branch_retired(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
261
void kvm_mediated_pmu_load(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
262
void kvm_mediated_pmu_put(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
265
bool kvm_need_perf_global_ctrl_intercept(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
266
bool kvm_need_rdpmc_intercept(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
27
struct kvm_pmc *(*rdpmc_ecx_to_pmc)(struct kvm_vcpu *vcpu,
arch/x86/kvm/pmu.h
29
struct kvm_pmc *(*msr_idx_to_pmc)(struct kvm_vcpu *vcpu, u32 msr);
arch/x86/kvm/pmu.h
30
int (*check_rdpmc_early)(struct kvm_vcpu *vcpu, unsigned int idx);
arch/x86/kvm/pmu.h
31
bool (*is_valid_msr)(struct kvm_vcpu *vcpu, u32 msr);
arch/x86/kvm/pmu.h
32
int (*get_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
33
int (*set_msr)(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/pmu.h
34
void (*refresh)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
35
void (*init)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
36
void (*reset)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
37
void (*deliver_pmi)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
38
void (*cleanup)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
41
void (*mediated_load)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
42
void (*mediated_put)(struct kvm_vcpu *vcpu);
arch/x86/kvm/pmu.h
74
static inline bool kvm_vcpu_has_mediated_pmu(struct kvm_vcpu *vcpu)
arch/x86/kvm/smm.c
112
void kvm_smm_changed(struct kvm_vcpu *vcpu, bool entering_smm)
arch/x86/kvm/smm.c
136
void process_smi(struct kvm_vcpu *vcpu)
arch/x86/kvm/smm.c
156
static void enter_smm_save_seg_32(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
170
static void enter_smm_save_seg_64(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
184
static void enter_smm_save_state_32(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
228
static void enter_smm_save_state_64(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
279
void enter_smm(struct kvm_vcpu *vcpu)
arch/x86/kvm/smm.c
388
static int rsm_load_seg_32(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
404
static int rsm_load_seg_64(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
419
static int rsm_enter_protected_mode(struct kvm_vcpu *vcpu,
arch/x86/kvm/smm.c
467
struct kvm_vcpu *vcpu = ctxt->vcpu;
arch/x86/kvm/smm.c
519
struct kvm_vcpu *vcpu = ctxt->vcpu;
arch/x86/kvm/smm.c
575
struct kvm_vcpu *vcpu = ctxt->vcpu;
arch/x86/kvm/smm.h
143
static inline int kvm_inject_smi(struct kvm_vcpu *vcpu)
arch/x86/kvm/smm.h
152
static inline bool is_smm(struct kvm_vcpu *vcpu)
arch/x86/kvm/smm.h
157
void kvm_smm_changed(struct kvm_vcpu *vcpu, bool in_smm);
arch/x86/kvm/smm.h
158
void enter_smm(struct kvm_vcpu *vcpu);
arch/x86/kvm/smm.h
160
void process_smi(struct kvm_vcpu *vcpu);
arch/x86/kvm/smm.h
162
static inline int kvm_inject_smi(struct kvm_vcpu *vcpu) { return -ENOTTY; }
arch/x86/kvm/smm.h
163
static inline bool is_smm(struct kvm_vcpu *vcpu) { return false; }
arch/x86/kvm/svm/avic.c
1001
static void __avic_vcpu_load(struct kvm_vcpu *vcpu, int cpu,
arch/x86/kvm/svm/avic.c
1054
void avic_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/svm/avic.c
1069
static void __avic_vcpu_put(struct kvm_vcpu *vcpu, enum avic_vcpu_action action)
arch/x86/kvm/svm/avic.c
1117
void avic_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
1150
void avic_refresh_virtual_apic_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
1174
void avic_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
1190
void avic_vcpu_blocking(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
1216
void avic_vcpu_unblocking(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
161
static u32 __avic_get_max_physical_id(struct kvm *kvm, struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
181
static u32 avic_get_max_physical_id(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
189
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/avic.c
252
struct kvm_vcpu *vcpu = NULL;
arch/x86/kvm/svm/avic.c
380
static int avic_init_backing_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
440
void avic_ring_doorbell(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
459
static void avic_kick_vcpu(struct kvm_vcpu *vcpu, u32 icrl)
arch/x86/kvm/svm/avic.c
475
struct kvm_vcpu *target_vcpu = kvm_get_vcpu_by_id(kvm, physical_id);
arch/x86/kvm/svm/avic.c
589
struct kvm_vcpu *vcpu;
arch/x86/kvm/svm/avic.c
609
int avic_incomplete_ipi_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
661
unsigned long avic_vcpu_get_apicv_inhibit_reasons(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
668
static u32 *avic_get_logical_id_entry(struct kvm_vcpu *vcpu, u32 ldr, bool flat)
arch/x86/kvm/svm/avic.c
694
static void avic_ldr_write(struct kvm_vcpu *vcpu, u8 g_physical_id, u32 ldr)
arch/x86/kvm/svm/avic.c
711
static void avic_invalidate_logical_id_entry(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
726
static void avic_handle_ldr_update(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
745
static void avic_handle_dfr_update(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
757
static int avic_unaccel_trap_write(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
809
int avic_unaccelerated_access_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
838
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/avic.c
855
void avic_apicv_post_state_restore(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/avic.c
863
struct kvm_vcpu *vcpu = irqfd->irq_bypass_vcpu;
arch/x86/kvm/svm/avic.c
876
struct kvm_vcpu *vcpu, u32 vector)
arch/x86/kvm/svm/avic.c
973
static void avic_update_iommu_vcpu_affinity(struct kvm_vcpu *vcpu, int cpu,
arch/x86/kvm/svm/hyperv.c
9
void svm_hv_inject_synthetic_vmexit_post_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/hyperv.h
15
static inline void nested_svm_hv_update_vm_vp_ids(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/hyperv.h
29
static inline bool nested_svm_l2_tlb_flush_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/hyperv.h
44
void svm_hv_inject_synthetic_vmexit_post_tlb_flush(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/hyperv.h
46
static inline void nested_svm_hv_update_vm_vp_ids(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/svm/hyperv.h
47
static inline bool nested_svm_l2_tlb_flush_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/hyperv.h
51
static inline void svm_hv_inject_synthetic_vmexit_post_tlb_flush(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/svm/nested.c
105
static void nested_svm_uninit_mmu_context(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1130
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/nested.c
1343
static void nested_svm_triple_fault(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1405
void svm_leave_nested(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1523
int nested_svm_check_permissions(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1538
static bool nested_svm_is_exception_vmexit(struct kvm_vcpu *vcpu, u8 vector,
arch/x86/kvm/svm/nested.c
1546
static void nested_svm_inject_exception_vmexit(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1586
static int svm_check_nested_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1664
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/nested.c
1697
void nested_svm_update_tsc_ratio_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
1743
static int svm_get_nested_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
1810
static int svm_set_nested_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
1942
static bool svm_get_nested_state_pages(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
279
static bool nested_svm_merge_msrpm(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
334
static bool nested_svm_check_bitmap_pa(struct kvm_vcpu *vcpu, u64 pa, u32 size)
arch/x86/kvm/svm/nested.c
342
static bool __nested_vmcb_check_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
36
static void nested_svm_inject_npf_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
370
static bool __nested_vmcb_check_save(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
405
static bool nested_vmcb_check_save(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
413
static bool nested_vmcb_check_controls(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
421
int nested_svm_check_cached_vmcb12(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
444
void __nested_copy_vmcb_control_to_cache(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/nested.c
565
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/nested.c
58
static u64 nested_svm_get_tdp_pdptr(struct kvm_vcpu *vcpu, int index)
arch/x86/kvm/svm/nested.c
595
static void nested_svm_transition_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
621
static int nested_svm_load_cr3(struct kvm_vcpu *vcpu, unsigned long cr3,
arch/x86/kvm/svm/nested.c
656
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/nested.c
76
static unsigned long nested_svm_get_tdp_cr3(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
760
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/nested.c
83
static void nested_svm_init_mmu_context(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/nested.c
943
int enter_svm_guest_mode(struct kvm_vcpu *vcpu, u64 vmcb12_gpa,
arch/x86/kvm/svm/nested.c
995
int nested_svm_vmrun(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/pmu.c
104
static bool amd_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/svm/pmu.c
128
static int amd_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/pmu.c
150
static int amd_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/pmu.c
179
static void amd_pmu_refresh(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/pmu.c
217
static void amd_pmu_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/pmu.c
237
static void amd_mediated_pmu_load(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/pmu.c
251
static void amd_mediated_pmu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/pmu.c
41
struct kvm_vcpu *vcpu = pmu_to_vcpu(pmu);
arch/x86/kvm/svm/pmu.c
76
static int amd_check_rdpmc_early(struct kvm_vcpu *vcpu, unsigned int idx)
arch/x86/kvm/svm/pmu.c
87
static struct kvm_pmc *amd_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/pmu.c
93
static struct kvm_pmc *amd_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/svm/sev.c
1023
struct kvm_vcpu *vcpu;
arch/x86/kvm/svm/sev.c
161
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
1968
struct kvm_vcpu *dst_vcpu, *src_vcpu;
arch/x86/kvm/svm/sev.c
2047
struct kvm_vcpu *src_vcpu;
arch/x86/kvm/svm/sev.c
2446
struct kvm_vcpu *vcpu;
arch/x86/kvm/svm/sev.c
3205
static void sev_flush_encrypted_page(struct kvm_vcpu *vcpu, void *va)
arch/x86/kvm/svm/sev.c
3254
void sev_free_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
3318
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
3338
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
3385
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
3719
static int snp_complete_psc_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
3735
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
3812
static int snp_complete_one_psc(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
3831
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
3962
static void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
4035
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
4036
struct kvm_vcpu *target_vcpu;
arch/x86/kvm/svm/sev.c
4161
static int snp_complete_req_certs(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
4208
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
4248
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
4389
int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
4554
void sev_es_recalc_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
4588
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
4671
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
4689
int sev_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
4774
void sev_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector)
arch/x86/kvm/svm/sev.c
4843
void sev_handle_rmp_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code)
arch/x86/kvm/svm/sev.c
5112
struct vmcb_save_area *sev_decrypt_vmsa(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/sev.c
5197
void sev_free_decrypted_vmsa(struct kvm_vcpu *vcpu, struct vmcb_save_area *vmsa)
arch/x86/kvm/svm/sev.c
877
struct kvm_vcpu *vcpu = &svm->vcpu;
arch/x86/kvm/svm/sev.c
970
static int __sev_launch_update_vmsa(struct kvm *kvm, struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
1004
void svm_write_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1013
static void svm_recalc_instruction_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1061
static void svm_recalc_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1067
static void init_vmcb(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/svm/svm.c
1221
static void __svm_vcpu_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1235
static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/svm/svm.c
1259
static int svm_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1306
static void svm_vcpu_free(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1378
static void svm_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1418
static void svm_prepare_host_switch(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1423
static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/svm/svm.c
1432
static void svm_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1442
static unsigned long svm_get_rflags(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1457
static void svm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
arch/x86/kvm/svm/svm.c
1470
static bool svm_get_if_flag(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1479
static void svm_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg)
arch/x86/kvm/svm/svm.c
1551
static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/svm/svm.c
1570
static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/svm/svm.c
1577
static void svm_get_segment(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
1646
static int svm_get_cpl(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1653
static void svm_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l)
arch/x86/kvm/svm/svm.c
1662
static void svm_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/svm/svm.c
1670
static void svm_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/svm/svm.c
1679
static void svm_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/svm/svm.c
1687
static void svm_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/svm/svm.c
1696
static void sev_post_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
arch/x86/kvm/svm/svm.c
1714
static bool svm_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/svm/svm.c
1719
void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/svm/svm.c
1776
static bool svm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/svm/svm.c
1781
void svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/svm/svm.c
1801
static void svm_set_segment(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
1832
static void svm_update_exception_bitmap(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1857
static void svm_set_dr6(struct kvm_vcpu *vcpu, unsigned long value)
arch/x86/kvm/svm/svm.c
1870
static void svm_sync_dirty_debug_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1891
static void svm_set_dr7(struct kvm_vcpu *vcpu, unsigned long value)
arch/x86/kvm/svm/svm.c
1902
static int pf_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1915
static int svm_check_emulate_instruction(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/svm/svm.c
1918
static int npf_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
1968
static int db_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2001
static int bp_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2012
static int ud_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2017
static int ac_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
203
int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer)
arch/x86/kvm/svm/svm.c
2055
static void svm_handle_mce(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2076
static int mc_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2081
static int shutdown_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2111
static int io_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2136
static int nmi_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2141
static int smi_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2146
static int intr_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2152
static int vmload_vmsave_interception(struct kvm_vcpu *vcpu, bool vmload)
arch/x86/kvm/svm/svm.c
2187
static int vmload_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2192
static int vmsave_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2197
static int vmrun_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2213
static int svm_instr_opcode(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2234
static int emulate_svm_instr(struct kvm_vcpu *vcpu, int opcode)
arch/x86/kvm/svm/svm.c
2241
int (*const svm_instr_handlers[])(struct kvm_vcpu *vcpu) = {
arch/x86/kvm/svm/svm.c
2267
static int gp_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2340
static int stgi_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2352
static int clgi_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2364
static int invlpga_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2381
static int skinit_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2389
static int task_switch_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2467
static int iret_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2483
static int invlpg_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2492
static int emulate_on_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2497
static int rsm_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2502
static bool check_selective_cr0_intercepted(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
2526
static int cr_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
254
static u32 svm_get_interrupt_shadow(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2600
static int cr_trap(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2635
static int dr_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
264
static void svm_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
arch/x86/kvm/svm/svm.c
2674
static int cr8_write_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2691
static int efer_trap(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
2726
static bool sev_es_prevent_msr_access(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
2734
static int svm_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/svm/svm.c
275
static int __svm_skip_emulated_instruction(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
2862
static int svm_complete_emulated_msr(struct kvm_vcpu *vcpu, int err)
arch/x86/kvm/svm/svm.c
2872
static int svm_set_vm_cr(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/svm/svm.c
2897
static int svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr)
arch/x86/kvm/svm/svm.c
3121
static int msr_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3129
static int interrupt_window_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
314
static int svm_skip_emulated_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3152
static int pause_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3168
static int invpcid_interception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
319
static int svm_update_soft_interrupt_rip(struct kvm_vcpu *vcpu, u8 vector)
arch/x86/kvm/svm/svm.c
3201
static inline int complete_userspace_buslock(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3216
static int bus_lock_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3232
static int (*const svm_exit_handlers[])(struct kvm_vcpu *vcpu) = {
arch/x86/kvm/svm/svm.c
3310
static void dump_vmcb(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3520
int svm_invoke_exit_handler(struct kvm_vcpu *vcpu, u64 __exit_code)
arch/x86/kvm/svm/svm.c
3567
static void svm_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/svm/svm.c
3584
static void svm_get_entry_info(struct kvm_vcpu *vcpu, u32 *intr_info,
arch/x86/kvm/svm/svm.c
3599
static int svm_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath)
arch/x86/kvm/svm/svm.c
3641
static int pre_svm_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
366
static void svm_inject_exception(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3667
static void svm_inject_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3688
static bool svm_is_vnmi_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3698
static bool svm_set_vnmi_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3721
static void svm_inject_irq(struct kvm_vcpu *vcpu, bool reinjected)
arch/x86/kvm/svm/svm.c
3742
void svm_complete_interrupt_delivery(struct kvm_vcpu *vcpu, int delivery_mode,
arch/x86/kvm/svm/svm.c
3792
static void svm_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
arch/x86/kvm/svm/svm.c
3815
static bool svm_get_nmi_mask(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3825
static void svm_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked)
arch/x86/kvm/svm/svm.c
3844
bool svm_nmi_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3861
static int svm_nmi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/svm/svm.c
3876
bool svm_interrupt_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3902
static int svm_interrupt_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/svm/svm.c
3922
static void svm_enable_irq_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
3952
static void svm_enable_nmi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4007
static void svm_flush_tlb_asid(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
402
static void svm_init_osvw(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4031
static void svm_flush_tlb_current(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4047
static void svm_flush_tlb_all(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4062
static void svm_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/svm/svm.c
4069
static void svm_flush_tlb_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4076
static inline void sync_cr8_to_lapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4089
static inline void sync_lapic_to_cr8(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4102
static void svm_complete_soft_interrupt(struct kvm_vcpu *vcpu, u8 vector,
arch/x86/kvm/svm/svm.c
4134
static void svm_complete_interrupts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4206
static void svm_cancel_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4217
static int svm_vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4225
static fastpath_t svm_exit_handlers_fastpath(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4256
static noinstr void svm_vcpu_enter_exit(struct kvm_vcpu *vcpu, bool spec_ctrl_intercepted)
arch/x86/kvm/svm/svm.c
4288
static __no_kcsan fastpath_t svm_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags)
arch/x86/kvm/svm/svm.c
4442
static void svm_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/kvm/svm/svm.c
4468
svm_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall)
arch/x86/kvm/svm/svm.c
4502
static void svm_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4601
static int svm_check_intercept(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.c
4735
static void svm_handle_exit_irqoff(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4749
static void svm_setup_mce(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4756
bool svm_smi_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4767
static int svm_smi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/svm/svm.c
4783
static int svm_enter_smm(struct kvm_vcpu *vcpu, union kvm_smram *smram)
arch/x86/kvm/svm/svm.c
4835
static int svm_leave_smm(struct kvm_vcpu *vcpu, const union kvm_smram *smram)
arch/x86/kvm/svm/svm.c
4901
static void svm_enable_smi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
4915
static int svm_check_emulate_instruction(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/svm/svm.c
5068
static bool svm_apic_init_signal_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
5075
static void svm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector)
arch/x86/kvm/svm/svm.c
5118
static void *svm_alloc_apic_backing_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
650
static bool msr_write_intercepted(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/svm/svm.c
667
void svm_set_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type, bool set)
arch/x86/kvm/svm/svm.c
710
static void svm_recalc_lbr_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
734
static void svm_recalc_pmu_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
768
static void svm_recalc_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
855
static void __svm_enable_lbrv(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
860
void svm_enable_lbrv(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
866
static void __svm_disable_lbrv(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
872
void svm_update_lbrv(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
907
static void grow_ple_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
928
static void shrink_ple_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
981
static u64 svm_get_l2_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
988
static u64 svm_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.c
995
static void svm_write_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.h
133
struct kvm_vcpu;
arch/x86/kvm/svm/svm.h
253
struct kvm_vcpu vcpu;
arch/x86/kvm/svm/svm.h
423
static __always_inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.h
718
void svm_enable_lbrv(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
719
void svm_update_lbrv(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
721
int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
arch/x86/kvm/svm/svm.h
722
void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/kvm/svm/svm.h
723
void svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/svm/svm.h
725
bool svm_smi_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
726
bool svm_nmi_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
727
bool svm_interrupt_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
729
int svm_invoke_exit_handler(struct kvm_vcpu *vcpu, u64 exit_code);
arch/x86/kvm/svm/svm.h
730
void set_msr_interception(struct kvm_vcpu *vcpu, u32 *msrpm, u32 msr,
arch/x86/kvm/svm/svm.h
732
void svm_complete_interrupt_delivery(struct kvm_vcpu *vcpu, int delivery_mode,
arch/x86/kvm/svm/svm.h
735
void svm_set_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type, bool set);
arch/x86/kvm/svm/svm.h
737
static inline void svm_disable_intercept_for_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.h
743
static inline void svm_enable_intercept_for_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.h
755
static inline bool nested_svm_virtualize_tpr(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.h
779
int enter_svm_guest_mode(struct kvm_vcpu *vcpu,
arch/x86/kvm/svm/svm.h
781
void svm_leave_nested(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
784
int nested_svm_vmrun(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
799
int nested_svm_check_permissions(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
800
int nested_svm_check_cached_vmcb12(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
804
void nested_svm_update_tsc_ratio_msr(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
805
void svm_write_tsc_multiplier(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
840
int avic_incomplete_ipi_interception(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
841
int avic_unaccelerated_access_interception(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
843
void avic_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
arch/x86/kvm/svm/svm.h
844
void avic_vcpu_put(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
845
void avic_apicv_post_state_restore(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
846
void avic_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
849
struct kvm_vcpu *vcpu, u32 vector);
arch/x86/kvm/svm/svm.h
850
void avic_vcpu_blocking(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
851
void avic_vcpu_unblocking(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
852
void avic_ring_doorbell(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
853
unsigned long avic_vcpu_get_apicv_inhibit_reasons(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
854
void avic_refresh_virtual_apic_mode(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
863
void sev_es_recalc_msr_intercepts(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
864
void sev_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector);
arch/x86/kvm/svm/svm.h
877
int sev_handle_vmgexit(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
887
int sev_vcpu_create(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
888
void sev_free_vcpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
896
void sev_handle_rmp_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code);
arch/x86/kvm/svm/svm.h
900
struct vmcb_save_area *sev_decrypt_vmsa(struct kvm_vcpu *vcpu);
arch/x86/kvm/svm/svm.h
901
void sev_free_decrypted_vmsa(struct kvm_vcpu *vcpu, struct vmcb_save_area *vmsa);
arch/x86/kvm/svm/svm.h
913
static inline int sev_vcpu_create(struct kvm_vcpu *vcpu) { return 0; }
arch/x86/kvm/svm/svm.h
914
static inline void sev_free_vcpu(struct kvm_vcpu *vcpu) {}
arch/x86/kvm/svm/svm.h
922
static inline void sev_handle_rmp_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code) {}
arch/x86/kvm/svm/svm.h
933
static inline struct vmcb_save_area *sev_decrypt_vmsa(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm.h
937
static inline void sev_free_decrypted_vmsa(struct kvm_vcpu *vcpu, struct vmcb_save_area *vmsa) {}
arch/x86/kvm/svm/svm_onhyperv.c
18
static int svm_hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
18
static inline bool svm_hv_is_enlightened_tlb_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
42
struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
51
static inline void svm_hv_update_vp_id(struct vmcb *vmcb, struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
63
static inline bool svm_hv_is_enlightened_tlb_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
77
struct kvm_vcpu *vcpu)
arch/x86/kvm/svm/svm_onhyperv.h
82
struct kvm_vcpu *vcpu)
arch/x86/kvm/trace.h
1180
TP_PROTO(unsigned int host_irq, struct kvm_vcpu *vcpu,
arch/x86/kvm/trace.h
1941
TP_PROTO(struct kvm_vcpu *vcpu, u64 gpa, u64 pfn, u64 error_code,
arch/x86/kvm/trace.h
25
TP_PROTO(struct kvm_vcpu *vcpu, bool force_immediate_exit),
arch/x86/kvm/trace.h
393
TP_PROTO(struct kvm_vcpu *vcpu, u32 isa), \
arch/x86/kvm/trace.h
502
TP_PROTO(struct kvm_vcpu *vcpu, u64 fault_address, u64 error_code),
arch/x86/kvm/trace.h
917
TP_PROTO(struct kvm_vcpu *vcpu, __u8 failed),
arch/x86/kvm/vmx/capabilities.h
128
static inline bool cpu_need_tpr_shadow(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/common.h
111
static inline void kvm_vcpu_trigger_posted_interrupt(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/common.h
159
static inline void __vmx_deliver_posted_interrupt(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/common.h
178
noinstr void vmx_handle_nmi(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/common.h
65
static __always_inline bool is_td_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/common.h
73
static __always_inline bool is_td_vcpu(struct kvm_vcpu *vcpu) { return false; }
arch/x86/kvm/vmx/common.h
83
static inline int __vmx_handle_ept_violation(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/vmx/hyperv.c
112
void nested_evmcs_filter_control_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata)
arch/x86/kvm/vmx/hyperv.c
17
u64 nested_get_evmptr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.c
199
int nested_enable_evmcs(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/hyperv.c
212
bool nested_evmcs_l2_tlb_flush_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.c
227
void vmx_hv_inject_synthetic_vmexit_post_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.c
30
uint16_t nested_get_evmcs_version(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.c
95
static bool evmcs_has_perf_global_ctrl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.h
45
static inline bool guest_cpu_cap_has_evmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/hyperv.h
55
u64 nested_get_evmptr(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/hyperv.h
56
uint16_t nested_get_evmcs_version(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/hyperv.h
57
int nested_enable_evmcs(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/hyperv.h
59
void nested_evmcs_filter_control_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata);
arch/x86/kvm/vmx/hyperv.h
61
bool nested_evmcs_l2_tlb_flush_enabled(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/hyperv.h
62
void vmx_hv_inject_synthetic_vmexit_post_tlb_flush(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/main.c
106
static void vt_update_cpu_dirty_logging(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
118
static void vt_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
128
static void vt_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
138
static int vt_vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
146
static fastpath_t vt_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags)
arch/x86/kvm/vmx/main.c
154
static int vt_handle_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/main.c
163
static int vt_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/main.c
183
static int vt_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/main.c
191
static void vt_recalc_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
205
static int vt_complete_emulated_msr(struct kvm_vcpu *vcpu, int err)
arch/x86/kvm/vmx/main.c
214
static int vt_smi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/main.c
222
static int vt_enter_smm(struct kvm_vcpu *vcpu, union kvm_smram *smram)
arch/x86/kvm/vmx/main.c
230
static int vt_leave_smm(struct kvm_vcpu *vcpu, const union kvm_smram *smram)
arch/x86/kvm/vmx/main.c
238
static void vt_enable_smi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
248
static int vt_check_emulate_instruction(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/vmx/main.c
264
static bool vt_apic_init_signal_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
276
static void vt_set_virtual_apic_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
285
static void vt_hwapic_isr_update(struct kvm_vcpu *vcpu, int max_isr)
arch/x86/kvm/vmx/main.c
293
static int vt_sync_pir_to_irr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
313
static void vt_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
321
static void vt_update_exception_bitmap(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
329
static u64 vt_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/main.c
337
static void vt_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var,
arch/x86/kvm/vmx/main.c
348
static void vt_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var,
arch/x86/kvm/vmx/main.c
357
static int vt_get_cpl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
365
static int vt_get_cpl_no_cache(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
373
static void vt_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l)
arch/x86/kvm/vmx/main.c
384
static bool vt_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/vmx/main.c
392
static void vt_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/vmx/main.c
400
static bool vt_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/main.c
408
static void vt_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/main.c
416
static int vt_set_efer(struct kvm_vcpu *vcpu, u64 efer)
arch/x86/kvm/vmx/main.c
424
static void vt_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/main.c
434
static void vt_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/main.c
442
static void vt_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/main.c
452
static void vt_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/main.c
460
static void vt_set_dr7(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/main.c
468
static void vt_sync_dirty_debug_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
481
static void vt_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg)
arch/x86/kvm/vmx/main.c
489
static unsigned long vt_get_rflags(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
497
static void vt_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
arch/x86/kvm/vmx/main.c
505
static bool vt_get_if_flag(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
513
static void vt_flush_tlb_all(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
523
static void vt_flush_tlb_current(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
533
static void vt_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/vmx/main.c
541
static void vt_flush_tlb_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
549
static void vt_inject_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
559
static int vt_nmi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/main.c
571
static bool vt_get_nmi_mask(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
583
static void vt_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked)
arch/x86/kvm/vmx/main.c
591
static void vt_enable_nmi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
600
static void vt_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa,
arch/x86/kvm/vmx/main.c
611
static void vt_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
arch/x86/kvm/vmx/main.c
619
static u32 vt_get_interrupt_shadow(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
627
static void vt_patch_hypercall(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/main.c
640
static void vt_inject_irq(struct kvm_vcpu *vcpu, bool reinjected)
arch/x86/kvm/vmx/main.c
648
static void vt_inject_exception(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
656
static void vt_cancel_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
664
static int vt_interrupt_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/main.c
672
static void vt_enable_irq_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
68
static int vt_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
680
static void vt_get_entry_info(struct kvm_vcpu *vcpu, u32 *intr_info, u32 *error_code)
arch/x86/kvm/vmx/main.c
691
static void vt_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/vmx/main.c
703
static void vt_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
arch/x86/kvm/vmx/main.c
711
static void vt_set_apic_access_page_addr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
719
static void vt_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
729
static void vt_load_eoi_exitmap(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap)
arch/x86/kvm/vmx/main.c
753
static u64 vt_get_l2_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
76
static void vt_vcpu_free(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
762
static u64 vt_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
771
static void vt_write_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
780
static void vt_write_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
790
static int vt_set_hv_timer(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc,
arch/x86/kvm/vmx/main.c
800
static void vt_cancel_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
810
static void vt_setup_mce(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/main.c
826
static int vt_vcpu_mem_enc_ioctl(struct kvm_vcpu *vcpu, void __user *argp)
arch/x86/kvm/vmx/main.c
834
static int vt_vcpu_mem_enc_unlocked_ioctl(struct kvm_vcpu *vcpu, void __user *argp)
arch/x86/kvm/vmx/main.c
86
static void vt_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/vmx/main.c
96
static void vt_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/vmx/nested.c
1006
static int nested_vmx_load_msr_check(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
1017
static int nested_vmx_store_msr_check(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
1035
static u32 nested_vmx_load_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count)
arch/x86/kvm/vmx/nested.c
1071
static bool nested_vmx_get_vmexit_msr_value(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
1098
static bool read_and_check_msr_entry(struct kvm_vcpu *vcpu, u64 gpa, int i,
arch/x86/kvm/vmx/nested.c
1118
static int nested_vmx_store_msr(struct kvm_vcpu *vcpu, u64 gpa, u32 count)
arch/x86/kvm/vmx/nested.c
1148
static bool nested_msr_store_list_has_msr(struct kvm_vcpu *vcpu, u32 msr_index)
arch/x86/kvm/vmx/nested.c
1172
static int nested_vmx_load_cr3(struct kvm_vcpu *vcpu, unsigned long cr3,
arch/x86/kvm/vmx/nested.c
1216
static bool nested_has_guest_tlb_tag(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
1224
static void nested_vmx_transition_tlb_flush(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
1478
int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data)
arch/x86/kvm/vmx/nested.c
158
static int nested_vmx_succeed(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
166
static int nested_vmx_failInvalid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
175
static int nested_vmx_failValid(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
194
static int nested_vmx_fail(struct kvm_vcpu *vcpu, u32 vm_instruction_error)
arch/x86/kvm/vmx/nested.c
209
static void nested_vmx_abort(struct kvm_vcpu *vcpu, u32 indicator)
arch/x86/kvm/vmx/nested.c
2120
struct kvm_vcpu *vcpu, bool from_launch)
arch/x86/kvm/vmx/nested.c
2209
void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
2233
static u64 vmx_calc_preemption_timer_value(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
2249
static void vmx_start_preemption_timer(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
233
static inline void nested_release_evmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
251
static bool nested_evmcs_handle_vmclear(struct kvm_vcpu *vcpu, gpa_t vmptr)
arch/x86/kvm/vmx/nested.c
2532
static void vmcs_read_cet_state(struct kvm_vcpu *vcpu, u64 *s_cet,
arch/x86/kvm/vmx/nested.c
2545
static void vmcs_write_cet_state(struct kvm_vcpu *vcpu, u64 s_cet,
arch/x86/kvm/vmx/nested.c
2705
static int prepare_vmcs02(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12,
arch/x86/kvm/vmx/nested.c
2866
static bool nested_vmx_check_eptp(struct kvm_vcpu *vcpu, u64 new_eptp)
arch/x86/kvm/vmx/nested.c
2914
static int nested_check_vm_execution_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
297
static void vmx_switch_vmcs(struct kvm_vcpu *vcpu, struct loaded_vmcs *vmcs)
arch/x86/kvm/vmx/nested.c
2977
static int nested_check_vm_exit_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
2994
static int nested_check_vm_entry_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3072
static int nested_vmx_check_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3088
static int nested_vmx_check_controls_late(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3120
static int nested_vmx_check_address_space_size(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3142
static int nested_vmx_check_cet_state_common(struct kvm_vcpu *vcpu, u64 s_cet,
arch/x86/kvm/vmx/nested.c
3152
static int nested_vmx_check_host_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
322
static void nested_put_vmcs12_pages(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3245
static int nested_vmx_check_vmcs_link_ptr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3288
static int nested_vmx_check_guest_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
336
static void free_nested(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3390
static bool nested_get_evmcs_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3419
static bool nested_get_vmcs12_pages(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3507
static bool vmx_get_nested_state_pages(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3534
static int nested_vmx_write_pml_buffer(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/x86/kvm/vmx/nested.c
3578
static int nested_vmx_check_permission(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3593
static void load_vmcs12_host_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3606
enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
3769
static int nested_vmx_run(struct kvm_vcpu *vcpu, bool launch)
arch/x86/kvm/vmx/nested.c
378
void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
3921
vmcs12_guest_cr0(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/nested.c
393
static void nested_ept_invalidate_addr(struct kvm_vcpu *vcpu, gpa_t eptp,
arch/x86/kvm/vmx/nested.c
3931
vmcs12_guest_cr4(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/nested.c
3940
static void vmcs12_save_pending_event(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
4006
static int vmx_complete_nested_posted_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4049
static void nested_vmx_inject_exception_vmexit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
413
static void nested_ept_inject_page_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
4142
static void nested_vmx_update_pending_dbg(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4151
static bool nested_vmx_preemption_timer_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4157
static bool vmx_has_nested_events(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/nested.c
4282
static int vmx_check_nested_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4493
static u32 vmx_get_preemption_timer_value(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4554
static void sync_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
4599
static void copy_vmcs02_to_vmcs12_rare(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
4628
static void sync_vmcs02_to_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/nested.c
468
static void nested_ept_new_eptp(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4719
static void prepare_vmcs12(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12,
arch/x86/kvm/vmx/nested.c
4775
static void load_vmcs12_host_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
479
static void nested_ept_init_mmu_context(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
492
static void nested_ept_uninit_mmu_context(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
4935
static void nested_vmx_restore_host_state(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5048
void __nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason,
arch/x86/kvm/vmx/nested.c
510
static bool nested_vmx_is_exception_vmexit(struct kvm_vcpu *vcpu, u8 vector,
arch/x86/kvm/vmx/nested.c
5206
static void nested_vmx_triple_fault(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5218
int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification,
arch/x86/kvm/vmx/nested.c
528
static int nested_vmx_check_io_bitmap_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
5343
static int nested_vmx_get_vmptr(struct kvm_vcpu *vcpu, gpa_t *vmpointer,
arch/x86/kvm/vmx/nested.c
5371
static struct vmcs *alloc_shadow_vmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5394
static int enter_vmx_operation(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
541
static int nested_vmx_check_msr_bitmap_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
5444
static int handle_vmxon(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
553
static int nested_vmx_check_tpr_shadow_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
5535
static inline void nested_release_vmcs12(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5563
static int handle_vmxoff(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5577
static int handle_vmclear(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5619
static int handle_vmlaunch(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5625
static int handle_vmresume(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5631
static int handle_vmread(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5737
static int handle_vmwrite(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5858
static int handle_vmptrld(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5928
static int handle_vmptrst(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
5956
static int handle_invept(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
6036
static int handle_invvpid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
6135
static int nested_vmx_eptp_switching(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6168
static int handle_vmfunc(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
6223
bool nested_vmx_check_io_bitmaps(struct kvm_vcpu *vcpu, unsigned int port,
arch/x86/kvm/vmx/nested.c
6256
static bool nested_vmx_exit_handled_io(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6280
static bool nested_vmx_exit_handled_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6325
static bool nested_vmx_exit_handled_cr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
637
static void nested_vmx_merge_pmu_msr_bitmaps(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6395
static bool nested_vmx_exit_handled_encls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6410
static bool nested_vmx_exit_handled_vmcs_access(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6455
static bool nested_vmx_l0_wants_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6535
static bool nested_vmx_l1_wants_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6675
bool nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
669
static inline bool nested_vmx_prepare_msr_bitmap(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6727
static int vmx_get_nested_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
6850
void vmx_leave_nested(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
6859
int nested_vmx_check_restored_vmcs12(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
6881
static int vmx_set_nested_state(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
7392
__init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *))
arch/x86/kvm/vmx/nested.c
789
static void nested_cache_shadow_vmcs12(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
808
static void nested_flush_cached_shadow_vmcs12(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
831
static bool nested_exit_intr_ack_set(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
837
static int nested_vmx_check_apic_access_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
847
static int nested_vmx_check_apicv_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
892
static u32 nested_vmx_max_atomic_switch_msrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.c
901
static int nested_vmx_check_msr_switch(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
923
static int nested_vmx_check_exit_msr_switch_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
937
static int nested_vmx_check_entry_msr_switch_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
948
static int nested_vmx_check_pml_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
961
static int nested_vmx_check_unrestricted_guest_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
970
static int nested_vmx_check_mode_based_ept_exec_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
979
static int nested_vmx_check_shadow_vmcs_controls(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.c
992
static int nested_vmx_msr_check_common(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.h
101
static inline bool nested_ept_ad_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
123
static inline unsigned nested_cpu_vmx_misc_cr3_count(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
133
static inline bool nested_cpu_has_vmwrite_any_field(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
139
static inline bool nested_cpu_has_zero_length_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
144
static inline bool nested_cpu_supports_monitor_trap_flag(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
150
static inline bool nested_cpu_has_vmx_shadow_vmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
20
void vmx_leave_nested(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/nested.h
23
__init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *));
arch/x86/kvm/vmx/nested.h
25
int nested_vmx_check_restored_vmcs12(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/nested.h
252
static inline bool nested_exit_on_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
26
void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/nested.h
261
static inline bool nested_exit_on_intr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
27
enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/nested.h
281
static inline bool nested_guest_cr0_valid(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/nested.h
29
bool nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/nested.h
295
static inline bool nested_host_cr0_valid(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/nested.h
30
void __nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason,
arch/x86/kvm/vmx/nested.h
303
static inline bool nested_cr4_valid(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/nested.h
312
static inline bool nested_cpu_has_no_hw_errcode_cc(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
34
static inline void nested_vmx_vmexit(struct kvm_vcpu *vcpu, u32 vm_exit_reason,
arch/x86/kvm/vmx/nested.h
50
void nested_sync_vmcs12_to_shadow(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/nested.h
51
int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data);
arch/x86/kvm/vmx/nested.h
53
int get_vmx_mem_address(struct kvm_vcpu *vcpu, unsigned long exit_qualification,
arch/x86/kvm/vmx/nested.h
55
bool nested_vmx_check_io_bitmaps(struct kvm_vcpu *vcpu, unsigned int port,
arch/x86/kvm/vmx/nested.h
58
static inline struct vmcs12 *get_vmcs12(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
66
static inline struct vmcs12 *get_shadow_vmcs12(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
79
static inline int vmx_has_valid_vmcs12(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
88
static inline u16 nested_get_vpid02(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/nested.h
95
static inline unsigned long nested_ept_get_eptp(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
140
static bool intel_pmu_lbr_is_compatible(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
148
bool intel_pmu_lbr_is_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
156
static bool intel_pmu_is_valid_lbr_msr(struct kvm_vcpu *vcpu, u32 index)
arch/x86/kvm/vmx/pmu_intel.c
174
static bool intel_is_valid_msr(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/vmx/pmu_intel.c
205
static struct kvm_pmc *intel_msr_idx_to_pmc(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/vmx/pmu_intel.c
217
static inline void intel_pmu_release_guest_lbr_event(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
231
int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
291
static bool intel_pmu_handle_lbr_msrs_access(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/pmu_intel.c
328
static int intel_pmu_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/pmu_intel.c
371
static int intel_pmu_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/pmu_intel.c
39
static struct lbr_desc *vcpu_to_lbr_desc(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
47
static struct x86_pmu_lbr *vcpu_to_lbr_records(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
490
static void intel_pmu_refresh(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
592
static void intel_pmu_init(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
621
static void intel_pmu_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
634
static void intel_pmu_legacy_freezing_lbrs_on_pmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
644
static void intel_pmu_deliver_pmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
655
static void vmx_update_intercept_for_lbr_msrs(struct kvm_vcpu *vcpu, bool set)
arch/x86/kvm/vmx/pmu_intel.c
671
static inline void vmx_disable_lbr_msrs_passthrough(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
682
static inline void vmx_enable_lbr_msrs_passthrough(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
703
void vmx_passthrough_lbr_msrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
733
static void intel_pmu_cleanup(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
789
static void intel_mediated_pmu_load(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.c
79
static struct kvm_pmc *intel_rdpmc_ecx_to_pmc(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/pmu_intel.c
804
static void intel_mediated_pmu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.h
17
static inline bool fw_writes_is_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/pmu_intel.h
22
bool intel_pmu_lbr_is_enabled(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/pmu_intel.h
23
int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/pmu_intel.h
9
static inline u64 vcpu_get_perf_capabilities(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
162
static void pi_enable_wakeup_handler(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
209
static bool vmx_needs_pi_wakeup(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
223
void vmx_vcpu_pi_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
275
void pi_apicv_pre_state_restore(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
283
bool pi_has_pending_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
307
struct kvm_vcpu *vcpu, u32 vector)
arch/x86/kvm/vmx/posted_intr.c
38
static struct pi_desc *vcpu_to_pi_desc(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/posted_intr.c
57
void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/vmx/posted_intr.h
11
void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu);
arch/x86/kvm/vmx/posted_intr.h
12
void vmx_vcpu_pi_put(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/posted_intr.h
15
void pi_apicv_pre_state_restore(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/posted_intr.h
16
bool pi_has_pending_interrupt(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/posted_intr.h
19
struct kvm_vcpu *vcpu, u32 vector);
arch/x86/kvm/vmx/sgx.c
105
static int sgx_inject_fault(struct kvm_vcpu *vcpu, gva_t gva, int trapnr)
arch/x86/kvm/vmx/sgx.c
141
static int __handle_encls_ecreate(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/sgx.c
217
static int handle_encls_ecreate(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
24
static int sgx_get_encls_gva(struct kvm_vcpu *vcpu, unsigned long offset,
arch/x86/kvm/vmx/sgx.c
297
static int handle_encls_einit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
359
static inline bool encls_leaf_enabled_in_guest(struct kvm_vcpu *vcpu, u32 leaf)
arch/x86/kvm/vmx/sgx.c
374
static inline bool sgx_enabled_in_guest_bios(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
381
int handle_encls(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
428
void vcpu_setup_sgx_lepubkeyhash(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
440
static bool sgx_intercept_encls_ecreate(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/sgx.c
468
void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12)
arch/x86/kvm/vmx/sgx.c
55
static void sgx_handle_emulation_failure(struct kvm_vcpu *vcpu, u64 addr,
arch/x86/kvm/vmx/sgx.c
63
static int sgx_read_hva(struct kvm_vcpu *vcpu, unsigned long hva, void *data,
arch/x86/kvm/vmx/sgx.c
74
static int sgx_gva_to_gpa(struct kvm_vcpu *vcpu, gva_t gva, bool write,
arch/x86/kvm/vmx/sgx.c
92
static int sgx_gpa_to_hva(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned long *hva)
arch/x86/kvm/vmx/sgx.h
13
int handle_encls(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/sgx.h
16
void vcpu_setup_sgx_lepubkeyhash(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/sgx.h
18
void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu, struct vmcs12 *vmcs12);
arch/x86/kvm/vmx/sgx.h
23
static inline void vcpu_setup_sgx_lepubkeyhash(struct kvm_vcpu *vcpu) { }
arch/x86/kvm/vmx/sgx.h
25
static inline void vmx_write_encls_bitmap(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/tdx.c
1038
static void tdx_load_host_xsave_state(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1065
fastpath_t tdx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags)
arch/x86/kvm/vmx/tdx.c
1121
void tdx_inject_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1143
static int tdx_handle_exception_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1162
static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1168
static int tdx_emulate_vmcall(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1187
static int tdx_complete_vmcall_map_gpa(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1243
static int tdx_map_gpa(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1288
static int tdx_report_fatal_error(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1314
static int tdx_emulate_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1333
static int tdx_complete_pio_out(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1339
static int tdx_complete_pio_in(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1355
static int tdx_emulate_io(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1391
static int tdx_complete_mmio_read(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1406
static inline int tdx_mmio_write(struct kvm_vcpu *vcpu, gpa_t gpa, int size,
arch/x86/kvm/vmx/tdx.c
1421
static inline int tdx_mmio_read(struct kvm_vcpu *vcpu, gpa_t gpa, int size)
arch/x86/kvm/vmx/tdx.c
1433
static int tdx_emulate_mmio(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1492
static int tdx_complete_get_td_vm_call_info(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1511
static int tdx_get_td_vm_call_info(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1541
static int tdx_complete_simple(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1547
static int tdx_get_quote(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1571
static int tdx_setup_event_notify_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1592
static int handle_tdvmcall(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1613
void tdx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int pgd_level)
arch/x86/kvm/vmx/tdx.c
1841
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/vmx/tdx.c
1850
static inline bool tdx_is_sept_violation_unexpected_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1861
static int tdx_handle_ept_violation(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
1964
int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err)
arch/x86/kvm/vmx/tdx.c
1978
int tdx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t fastpath)
arch/x86/kvm/vmx/tdx.c
2085
void tdx_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/vmx/tdx.c
2157
int tdx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr)
arch/x86/kvm/vmx/tdx.c
2182
int tdx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr)
arch/x86/kvm/vmx/tdx.c
244
static __always_inline unsigned long tdvmcall_exit_type(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
249
static __always_inline unsigned long tdvmcall_leaf(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
254
static __always_inline void tdvmcall_set_return_code(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/tdx.c
2582
static int tdx_read_cpuid(struct kvm_vcpu *vcpu, u32 leaf, u32 sub_leaf,
arch/x86/kvm/vmx/tdx.c
260
static __always_inline void tdvmcall_set_return_val(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/tdx.c
2778
void tdx_flush_tlb_current(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
2797
void tdx_flush_tlb_all(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
281
static inline void tdx_disassociate_vp(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
2882
static int tdx_td_vcpu_init(struct kvm_vcpu *vcpu, u64 vcpu_rcx)
arch/x86/kvm/vmx/tdx.c
2977
static int tdx_vcpu_get_cpuid_leaf(struct kvm_vcpu *vcpu, u32 leaf, int *entry_index,
arch/x86/kvm/vmx/tdx.c
3006
static int tdx_vcpu_get_cpuid(struct kvm_vcpu *vcpu, struct kvm_tdx_cmd *cmd)
arch/x86/kvm/vmx/tdx.c
3067
static int tdx_vcpu_init(struct kvm_vcpu *vcpu, struct kvm_tdx_cmd *cmd)
arch/x86/kvm/vmx/tdx.c
3101
void tdx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/vmx/tdx.c
3115
struct kvm_vcpu *vcpu;
arch/x86/kvm/vmx/tdx.c
3157
static int tdx_vcpu_init_mem_region(struct kvm_vcpu *vcpu, struct kvm_tdx_cmd *cmd)
arch/x86/kvm/vmx/tdx.c
3223
int tdx_vcpu_unlocked_ioctl(struct kvm_vcpu *vcpu, void __user *argp)
arch/x86/kvm/vmx/tdx.c
3260
int tdx_vcpu_ioctl(struct kvm_vcpu *vcpu, void __user *argp)
arch/x86/kvm/vmx/tdx.c
373
struct kvm_vcpu *vcpu;
arch/x86/kvm/vmx/tdx.c
380
struct kvm_vcpu *vcpu = arg->vcpu;
arch/x86/kvm/vmx/tdx.c
417
static void tdx_flush_vp_on_cpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
496
struct kvm_vcpu *vcpu;
arch/x86/kvm/vmx/tdx.c
676
int tdx_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
723
void tdx_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/vmx/tdx.c
745
bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
756
static bool tdx_protected_apic_has_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
793
void tdx_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
822
static void tdx_prepare_switch_to_host(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
835
void tdx_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
85
static __always_inline struct vcpu_tdx *to_tdx(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
871
void tdx_vcpu_free(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
908
int tdx_vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
917
static __always_inline u32 tdcall_to_vmx_exit_reason(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
935
static __always_inline u32 tdx_to_vmx_exit_reason(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
972
static noinstr void tdx_vcpu_enter_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
993
static bool tdx_failed_vmentry(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.c
999
static fastpath_t tdx_exit_handlers_fastpath(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/tdx.h
179
bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/tdx.h
180
int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err);
arch/x86/kvm/vmx/tdx.h
200
struct kvm_vcpu vcpu;
arch/x86/kvm/vmx/tdx.h
203
static inline bool tdx_interrupt_allowed(struct kvm_vcpu *vcpu) { return false; }
arch/x86/kvm/vmx/tdx.h
204
static inline int tdx_complete_emulated_msr(struct kvm_vcpu *vcpu, int err) { return 0; }
arch/x86/kvm/vmx/tdx.h
61
struct kvm_vcpu vcpu;
arch/x86/kvm/vmx/vmx.c
1253
static inline bool pt_output_base_valid(struct kvm_vcpu *vcpu, u64 base)
arch/x86/kvm/vmx/vmx.c
1349
void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1494
static void grow_ple_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1510
static void shrink_ple_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1526
void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/vmx/vmx.c
1586
void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/vmx/vmx.c
1596
void vmx_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1603
static void vmx_switch_loaded_vmcs(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
1615
static void vmx_load_vmcs01(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1628
static void vmx_put_vmcs01(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1635
DEFINE_GUARD(vmx_vmcs01, struct kvm_vcpu *,
arch/x86/kvm/vmx/vmx.c
1638
bool vmx_emulation_required(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1643
unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1661
void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
arch/x86/kvm/vmx/vmx.c
1690
bool vmx_get_if_flag(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1695
u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1708
void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
arch/x86/kvm/vmx/vmx.c
1724
static int vmx_rtit_ctl_check(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/vmx/vmx.c
1797
int vmx_check_emulate_instruction(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/vmx/vmx.c
1820
static int skip_emulated_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1888
void vmx_update_emulated_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1919
int vmx_skip_emulated_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1925
static void vmx_clear_hlt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
1938
void vmx_inject_exception(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
2038
u64 vmx_get_l2_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
2048
u64 vmx_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
2059
void vmx_write_tsc_offset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
2064
void vmx_write_tsc_multiplier(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
2124
int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/vmx.c
2284
static u64 nested_vmx_truncate_sysenter_addr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
2294
u64 vmx_get_supported_debugctl(struct kvm_vcpu *vcpu, bool host_initiated)
arch/x86/kvm/vmx/vmx.c
2313
bool vmx_is_valid_debugctl(struct kvm_vcpu *vcpu, u64 data, bool host_initiated)
arch/x86/kvm/vmx/vmx.c
2330
int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/vmx/vmx.c
2633
void vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg)
arch/x86/kvm/vmx/vmx.c
3172
static void fix_pmode_seg(struct kvm_vcpu *vcpu, int seg,
arch/x86/kvm/vmx/vmx.c
3191
static void enter_pmode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3261
static void enter_rmode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3309
int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer)
arch/x86/kvm/vmx/vmx.c
3334
static void enter_lmode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3351
static void exit_lmode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3358
void vmx_flush_tlb_all(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3381
static inline int vmx_get_current_vpid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3422
void vmx_flush_tlb_current(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3437
void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr)
arch/x86/kvm/vmx/vmx.c
3446
void vmx_flush_tlb_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3458
void vmx_ept_load_pdptrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3473
void ept_save_pdptrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3491
bool vmx_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/vmx/vmx.c
3502
void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/vmx/vmx.c
3598
void vmx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level)
arch/x86/kvm/vmx/vmx.c
3627
bool vmx_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/vmx.c
3643
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/vmx/vmx.c
3707
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3743
u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
3754
static int __vmx_get_cpl(struct kvm_vcpu *vcpu, bool no_cache)
arch/x86/kvm/vmx/vmx.c
3769
int vmx_get_cpl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3774
int vmx_get_cpl_no_cache(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3796
void __vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3833
void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3840
void vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l)
arch/x86/kvm/vmx/vmx.c
3848
void vmx_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/vmx.c
3854
void vmx_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/vmx.c
3860
void vmx_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/vmx.c
3866
void vmx_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt)
arch/x86/kvm/vmx/vmx.c
3872
static bool rmode_segment_valid(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
388
static noinstr void vmx_l1d_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3893
static bool code_segment_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3921
static bool stack_segment_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3943
static bool data_segment_valid(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
3968
static bool tr_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
3986
static bool ldtr_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4004
static bool cs_ss_rpl_check(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4020
bool __vmx_guest_state_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4186
void vmx_set_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type, bool set)
arch/x86/kvm/vmx/vmx.c
4211
static void vmx_update_msr_bitmap_x2apic(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4271
void pt_update_intercept_for_msr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4287
static void vmx_recalc_pmu_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4343
static void vmx_recalc_msr_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4417
static void vmx_recalc_instruction_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4423
void vmx_recalc_intercepts(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4429
static int vmx_deliver_nested_posted_interrupt(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
4475
static int vmx_deliver_posted_interrupt(struct kvm_vcpu *vcpu, int vector)
arch/x86/kvm/vmx/vmx.c
4495
struct kvm_vcpu *vcpu = apic->vcpu;
arch/x86/kvm/vmx/vmx.c
452
static __always_inline void vmx_l1d_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4607
struct kvm_vcpu *vcpu = &vmx->vcpu;
arch/x86/kvm/vmx/vmx.c
4674
void vmx_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
4798
struct kvm_vcpu *__vcpu = &(vmx)->vcpu; \
arch/x86/kvm/vmx/vmx.c
4817
struct kvm_vcpu *vcpu = &vmx->vcpu;
arch/x86/kvm/vmx/vmx.c
496
static void vmx_update_fb_clear_dis(struct kvm_vcpu *vcpu, struct vcpu_vmx *vmx)
arch/x86/kvm/vmx/vmx.c
5059
static void __vmx_vcpu_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5091
void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/vmx/vmx.c
5158
void vmx_enable_irq_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5163
void vmx_enable_nmi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5174
void vmx_inject_irq(struct kvm_vcpu *vcpu, bool reinjected)
arch/x86/kvm/vmx/vmx.c
5202
void vmx_inject_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5233
bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5247
void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked)
arch/x86/kvm/vmx/vmx.c
5267
bool vmx_nmi_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5280
int vmx_nmi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/vmx.c
5292
bool __vmx_interrupt_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5299
bool vmx_interrupt_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5307
int vmx_interrupt_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/vmx.c
5348
static bool rmode_exception(struct kvm_vcpu *vcpu, int vec)
arch/x86/kvm/vmx/vmx.c
5377
static int handle_rmode_exception(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
5404
static int handle_machine_check(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5421
bool vmx_guest_inject_ac(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5430
static bool is_xfd_nm_fault(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5436
static int vmx_handle_page_fault(struct kvm_vcpu *vcpu, u32 error_code)
arch/x86/kvm/vmx/vmx.c
5477
static int handle_exception_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5641
static __always_inline int handle_external_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5647
static int handle_triple_fault(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5654
static int handle_io(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5675
void vmx_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall)
arch/x86/kvm/vmx/vmx.c
5686
static int handle_set_cr0(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/vmx.c
5712
static int handle_set_cr4(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/vmx.c
5729
static int handle_desc(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5741
static int handle_cr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5821
static int handle_dr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5881
void vmx_sync_dirty_debug_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5900
void vmx_set_dr7(struct kvm_vcpu *vcpu, unsigned long val)
arch/x86/kvm/vmx/vmx.c
5905
static int handle_tpr_below_threshold(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5911
static int handle_interrupt_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5921
static int handle_invlpg(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5929
static int handle_apic_access(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5951
static int handle_apic_eoi_induced(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5961
static int handle_apic_write(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
5978
static int handle_task_switch(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6035
static int handle_ept_violation(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6068
static int handle_ept_misconfig(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6089
static int handle_nmi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6106
static bool vmx_unhandleable_emulation_required(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6133
static int handle_invalid_guest_state(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6181
int vmx_vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6195
static int handle_pause(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6210
static int handle_monitor_trap(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6215
static int handle_invpcid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6246
static int handle_pml_full(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
627
static int hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6271
static fastpath_t handle_fastpath_preemption_timer(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
6302
static int handle_preemption_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6318
static int handle_vmx_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6324
static int handle_tdx_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6331
static int handle_encls(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6343
static int handle_bus_lock_vmexit(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6354
static int handle_notify(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6380
static int vmx_get_msr_imm_reg(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6385
static int handle_rdmsr_imm(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6391
static int handle_wrmsr_imm(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6402
static int (*kvm_vmx_exit_handlers[])(struct kvm_vcpu *vcpu) = {
arch/x86/kvm/vmx/vmx.c
6464
void vmx_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/vmx/vmx.c
6485
void vmx_get_entry_info(struct kvm_vcpu *vcpu, u32 *intr_info, u32 *error_code)
arch/x86/kvm/vmx/vmx.c
6502
static void vmx_flush_pml_buffer(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6541
static void nested_vmx_mark_all_vmcs12_pages_dirty(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6576
void dump_vmcs(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
6781
static int __vmx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath)
arch/x86/kvm/vmx/vmx.c
6937
int vmx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath)
arch/x86/kvm/vmx/vmx.c
6955
void vmx_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
arch/x86/kvm/vmx/vmx.c
6970
void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7029
void vmx_set_apic_access_page_addr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7097
void vmx_hwapic_isr_update(struct kvm_vcpu *vcpu, int max_isr)
arch/x86/kvm/vmx/vmx.c
7138
int vmx_sync_pir_to_irr(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7184
void vmx_load_eoi_exitmap(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap)
arch/x86/kvm/vmx/vmx.c
7198
static void handle_nm_fault_irqoff(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7217
static void handle_exception_irqoff(struct kvm_vcpu *vcpu, u32 intr_info)
arch/x86/kvm/vmx/vmx.c
7230
static void handle_external_interrupt_irqoff(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
7255
void vmx_handle_exit_irqoff(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7341
static void __vmx_complete_interrupts(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
7406
void vmx_cancel_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7442
static void vmx_refresh_guest_perf_global_control(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7465
static void vmx_update_hv_timer(struct kvm_vcpu *vcpu, bool force_immediate_exit)
arch/x86/kvm/vmx/vmx.c
7524
static fastpath_t vmx_exit_handlers_fastpath(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
7552
noinstr void vmx_handle_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7566
static noinstr void vmx_vcpu_enter_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
758
static inline bool cpu_need_virtualize_apic_accesses(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7605
fastpath_t vmx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags)
arch/x86/kvm/vmx/vmx.c
7758
void vmx_vcpu_free(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7770
int vmx_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
7919
u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio)
arch/x86/kvm/vmx/vmx.c
7958
static void nested_vmx_cr_fixed1_bits_update(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8003
static void update_intel_pt_cfg(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8072
void vmx_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8253
static bool vmx_is_io_intercepted(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
8299
int vmx_check_intercept(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.c
8414
int vmx_set_hv_timer(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc,
arch/x86/kvm/vmx/vmx.c
8454
void vmx_cancel_hv_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8460
void vmx_update_cpu_dirty_logging(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8480
void vmx_setup_mce(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8491
int vmx_smi_allowed(struct kvm_vcpu *vcpu, bool for_injection)
arch/x86/kvm/vmx/vmx.c
8499
int vmx_enter_smm(struct kvm_vcpu *vcpu, union kvm_smram *smram)
arch/x86/kvm/vmx/vmx.c
8520
int vmx_leave_smm(struct kvm_vcpu *vcpu, const union kvm_smram *smram)
arch/x86/kvm/vmx/vmx.c
8545
void vmx_enable_smi_window(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8551
bool vmx_apic_init_signal_blocked(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8556
void vmx_migrate_timers(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.c
8590
gva_t vmx_get_untagged_addr(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags)
arch/x86/kvm/vmx/vmx.c
8629
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
arch/x86/kvm/vmx/vmx.c
927
void vmx_update_exception_bitmap(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
204
struct kvm_vcpu vcpu;
arch/x86/kvm/vmx/vmx.h
304
static __always_inline struct vcpu_vt *to_vt(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
309
static __always_inline struct kvm_vcpu *vt_to_vcpu(struct vcpu_vt *vt)
arch/x86/kvm/vmx/vmx.h
314
static __always_inline union vmx_exit_reason vmx_get_exit_reason(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
319
static __always_inline unsigned long vmx_get_exit_qual(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
330
static __always_inline u32 vmx_get_intr_info(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
341
void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu);
arch/x86/kvm/vmx/vmx.h
345
void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
348
int vmx_get_cpl(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
349
int vmx_get_cpl_no_cache(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
350
bool vmx_emulation_required(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
351
unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
352
void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
arch/x86/kvm/vmx/vmx.h
353
u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
354
void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask);
arch/x86/kvm/vmx/vmx.h
355
int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
arch/x86/kvm/vmx/vmx.h
356
void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/kvm/vmx/vmx.h
357
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/vmx.h
359
void ept_save_pdptrs(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
360
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/vmx.h
361
void __vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/vmx.h
363
bool vmx_guest_inject_ac(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
364
void vmx_update_exception_bitmap(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
365
bool vmx_nmi_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
366
bool __vmx_interrupt_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
367
bool vmx_interrupt_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
368
bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
369
void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked);
arch/x86/kvm/vmx/vmx.h
370
void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
372
void pt_update_intercept_for_msr(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
378
void vmx_ept_load_pdptrs(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
380
void vmx_set_intercept_for_msr(struct kvm_vcpu *vcpu, u32 msr, int type, bool set);
arch/x86/kvm/vmx/vmx.h
382
static inline void vmx_disable_intercept_for_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.h
388
static inline void vmx_enable_intercept_for_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/vmx.h
394
u64 vmx_get_l2_tsc_offset(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
395
u64 vmx_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
397
gva_t vmx_get_untagged_addr(struct kvm_vcpu *vcpu, gva_t gva, unsigned int flags);
arch/x86/kvm/vmx/vmx.h
399
void vmx_update_cpu_dirty_logging(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
401
u64 vmx_get_supported_debugctl(struct kvm_vcpu *vcpu, bool host_initiated);
arch/x86/kvm/vmx/vmx.h
402
bool vmx_is_valid_debugctl(struct kvm_vcpu *vcpu, u64 data, bool host_initiated);
arch/x86/kvm/vmx/vmx.h
406
static inline void vmx_guest_debugctl_write(struct kvm_vcpu *vcpu, u64 val)
arch/x86/kvm/vmx/vmx.h
419
static inline void vmx_reload_guest_debugctl(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
660
static __always_inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
666
int intel_pmu_create_guest_lbr_event(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
667
void vmx_passthrough_lbr_msrs(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
686
static inline bool vmx_need_pf_intercept(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
695
static inline bool is_unrestricted_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
702
bool __vmx_guest_state_valid(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
703
static inline bool vmx_guest_state_valid(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/vmx.h
708
void dump_vmcs(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/vmx.h
720
static inline bool vmx_can_use_ipiv(struct kvm_vcpu *vcpu)
arch/x86/kvm/vmx/x86_ops.h
100
void vmx_set_apic_access_page_addr(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
101
void vmx_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
102
void vmx_load_eoi_exitmap(struct kvm_vcpu *vcpu, u64 *eoi_exit_bitmap);
arch/x86/kvm/vmx/x86_ops.h
105
u8 vmx_get_mt_mask(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio);
arch/x86/kvm/vmx/x86_ops.h
107
void vmx_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/vmx/x86_ops.h
109
void vmx_get_entry_info(struct kvm_vcpu *vcpu, u32 *intr_info, u32 *error_code);
arch/x86/kvm/vmx/x86_ops.h
111
u64 vmx_get_l2_tsc_offset(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
112
u64 vmx_get_l2_tsc_multiplier(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
113
void vmx_write_tsc_offset(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
114
void vmx_write_tsc_multiplier(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
115
void vmx_update_cpu_dirty_logging(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
117
int vmx_set_hv_timer(struct kvm_vcpu *vcpu, u64 guest_deadline_tsc,
arch/x86/kvm/vmx/x86_ops.h
119
void vmx_cancel_hv_timer(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
121
void vmx_setup_mce(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
130
int tdx_vcpu_create(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
131
void tdx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
arch/x86/kvm/vmx/x86_ops.h
132
void tdx_vcpu_free(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
133
void tdx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
arch/x86/kvm/vmx/x86_ops.h
134
int tdx_vcpu_pre_run(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
135
fastpath_t tdx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags);
arch/x86/kvm/vmx/x86_ops.h
136
void tdx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
137
void tdx_vcpu_put(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
138
int tdx_handle_exit(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/x86_ops.h
143
void tdx_inject_nmi(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
144
void tdx_get_exit_info(struct kvm_vcpu *vcpu, u32 *reason,
arch/x86/kvm/vmx/x86_ops.h
147
int tdx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/kvm/vmx/x86_ops.h
148
int tdx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr);
arch/x86/kvm/vmx/x86_ops.h
150
int tdx_vcpu_ioctl(struct kvm_vcpu *vcpu, void __user *argp);
arch/x86/kvm/vmx/x86_ops.h
151
int tdx_vcpu_unlocked_ioctl(struct kvm_vcpu *vcpu, void __user *argp);
arch/x86/kvm/vmx/x86_ops.h
153
void tdx_flush_tlb_current(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
154
void tdx_flush_tlb_all(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
155
void tdx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
arch/x86/kvm/vmx/x86_ops.h
22
int vmx_vcpu_create(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
23
int vmx_vcpu_pre_run(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
24
fastpath_t vmx_vcpu_run(struct kvm_vcpu *vcpu, u64 run_flags);
arch/x86/kvm/vmx/x86_ops.h
25
void vmx_vcpu_free(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
26
void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event);
arch/x86/kvm/vmx/x86_ops.h
27
void vmx_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
arch/x86/kvm/vmx/x86_ops.h
28
void vmx_vcpu_put(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
29
int vmx_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath);
arch/x86/kvm/vmx/x86_ops.h
30
void vmx_handle_exit_irqoff(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
31
int vmx_skip_emulated_instruction(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
32
void vmx_update_emulated_instruction(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
33
int vmx_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/vmx/x86_ops.h
35
int vmx_smi_allowed(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/kvm/vmx/x86_ops.h
36
int vmx_enter_smm(struct kvm_vcpu *vcpu, union kvm_smram *smram);
arch/x86/kvm/vmx/x86_ops.h
37
int vmx_leave_smm(struct kvm_vcpu *vcpu, const union kvm_smram *smram);
arch/x86/kvm/vmx/x86_ops.h
38
void vmx_enable_smi_window(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
40
int vmx_check_emulate_instruction(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/vmx/x86_ops.h
42
int vmx_check_intercept(struct kvm_vcpu *vcpu,
arch/x86/kvm/vmx/x86_ops.h
46
bool vmx_apic_init_signal_blocked(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
47
void vmx_migrate_timers(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
48
void vmx_set_virtual_apic_mode(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
49
void vmx_hwapic_isr_update(struct kvm_vcpu *vcpu, int max_isr);
arch/x86/kvm/vmx/x86_ops.h
50
int vmx_sync_pir_to_irr(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
53
void vmx_vcpu_after_set_cpuid(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
55
void vmx_recalc_intercepts(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
56
void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
57
void vmx_update_exception_bitmap(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
59
int vmx_get_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info);
arch/x86/kvm/vmx/x86_ops.h
61
u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg);
arch/x86/kvm/vmx/x86_ops.h
62
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/x86_ops.h
63
void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/x86_ops.h
64
int vmx_get_cpl(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
65
void vmx_get_cs_db_l_bits(struct kvm_vcpu *vcpu, int *db, int *l);
arch/x86/kvm/vmx/x86_ops.h
66
bool vmx_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/kvm/vmx/x86_ops.h
67
void vmx_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
arch/x86/kvm/vmx/x86_ops.h
68
void vmx_load_mmu_pgd(struct kvm_vcpu *vcpu, hpa_t root_hpa, int root_level);
arch/x86/kvm/vmx/x86_ops.h
69
void vmx_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/x86_ops.h
70
bool vmx_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
arch/x86/kvm/vmx/x86_ops.h
71
int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
arch/x86/kvm/vmx/x86_ops.h
72
void vmx_get_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/kvm/vmx/x86_ops.h
73
void vmx_set_idt(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/kvm/vmx/x86_ops.h
74
void vmx_get_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/kvm/vmx/x86_ops.h
75
void vmx_set_gdt(struct kvm_vcpu *vcpu, struct desc_ptr *dt);
arch/x86/kvm/vmx/x86_ops.h
76
void vmx_set_dr7(struct kvm_vcpu *vcpu, unsigned long val);
arch/x86/kvm/vmx/x86_ops.h
77
void vmx_sync_dirty_debug_regs(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
78
void vmx_cache_reg(struct kvm_vcpu *vcpu, enum kvm_reg reg);
arch/x86/kvm/vmx/x86_ops.h
79
unsigned long vmx_get_rflags(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
80
void vmx_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
arch/x86/kvm/vmx/x86_ops.h
81
bool vmx_get_if_flag(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
82
void vmx_flush_tlb_all(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
83
void vmx_flush_tlb_current(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
84
void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr);
arch/x86/kvm/vmx/x86_ops.h
85
void vmx_flush_tlb_guest(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
86
void vmx_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask);
arch/x86/kvm/vmx/x86_ops.h
87
u32 vmx_get_interrupt_shadow(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
88
void vmx_patch_hypercall(struct kvm_vcpu *vcpu, unsigned char *hypercall);
arch/x86/kvm/vmx/x86_ops.h
89
void vmx_inject_irq(struct kvm_vcpu *vcpu, bool reinjected);
arch/x86/kvm/vmx/x86_ops.h
90
void vmx_inject_nmi(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
91
void vmx_inject_exception(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
92
void vmx_cancel_injection(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
93
int vmx_interrupt_allowed(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/kvm/vmx/x86_ops.h
94
int vmx_nmi_allowed(struct kvm_vcpu *vcpu, bool for_injection);
arch/x86/kvm/vmx/x86_ops.h
95
bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
96
void vmx_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked);
arch/x86/kvm/vmx/x86_ops.h
97
void vmx_enable_nmi_window(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
98
void vmx_enable_irq_window(struct kvm_vcpu *vcpu);
arch/x86/kvm/vmx/x86_ops.h
99
void vmx_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr);
arch/x86/kvm/x86.c
1010
void kvm_inject_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1016
void kvm_queue_exception_e(struct kvm_vcpu *vcpu, unsigned nr, u32 error_code)
arch/x86/kvm/x86.c
10236
static int kvm_pv_clock_pairing(struct kvm_vcpu *vcpu, gpa_t paddr,
arch/x86/kvm/x86.c
1026
bool kvm_require_cpl(struct kvm_vcpu *vcpu, int required_cpl)
arch/x86/kvm/x86.c
10299
bool kvm_vcpu_apicv_activated(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10334
static void kvm_sched_yield(struct kvm_vcpu *vcpu, unsigned long dest_id)
arch/x86/kvm/x86.c
10336
struct kvm_vcpu *target = NULL;
arch/x86/kvm/x86.c
1034
bool kvm_require_dr(struct kvm_vcpu *vcpu, int dr)
arch/x86/kvm/x86.c
10371
static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10381
int ____kvm_emulate_hypercall(struct kvm_vcpu *vcpu, int cpl,
arch/x86/kvm/x86.c
10382
int (*complete_hypercall)(struct kvm_vcpu *))
arch/x86/kvm/x86.c
1044
static bool kvm_pv_async_pf_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10485
int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10500
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
1051
static inline u64 pdptr_rsvd_bits(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10521
static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10528
static void post_kvm_run_save(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10546
static void update_cr8_intercept(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10573
int kvm_check_nested_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10583
static void kvm_inject_exception(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1059
int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3)
arch/x86/kvm/x86.c
10641
static int kvm_check_and_inject_events(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
107
((struct kvm_vcpu *)(ctxt)->vcpu)
arch/x86/kvm/x86.c
10849
static void process_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10887
int kvm_get_nr_pending_nmis(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10904
void __kvm_vcpu_update_apicv(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
10941
static void kvm_vcpu_update_apicv(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11017
static void vcpu_scan_ioapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11040
static void vcpu_load_eoi_exitmap(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1106
static bool kvm_is_valid_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/x86.c
11065
static void kvm_vcpu_reload_apic_access_page(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11079
static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1122
void kvm_post_set_cr0(struct kvm_vcpu *vcpu, unsigned long old_cr0, unsigned long cr0)
arch/x86/kvm/x86.c
11533
static bool kvm_vcpu_running(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11539
bool kvm_vcpu_has_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11587
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11594
static inline int vcpu_block(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1164
int kvm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0)
arch/x86/kvm/x86.c
11662
static int vcpu_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11712
static int __kvm_emulate_halt(struct kvm_vcpu *vcpu, int state, int reason)
arch/x86/kvm/x86.c
11733
int kvm_emulate_halt_noskip(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11739
int kvm_emulate_halt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11750
fastpath_t handle_fastpath_hlt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11765
int kvm_emulate_ap_reset_hold(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11774
bool kvm_arch_dy_has_pending_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11780
bool kvm_arch_vcpu_preempted_in_kernel(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11785
bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11800
static inline int complete_emulated_io(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11805
static int complete_emulated_pio(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11830
static int complete_emulated_mmio(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11876
static void kvm_load_guest_fpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11887
static void kvm_put_guest_fpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11897
static int kvm_x86_vcpu_pre_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
11919
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12006
int (*cui)(struct kvm_vcpu *) = vcpu->arch.complete_userspace_io;
arch/x86/kvm/x86.c
12039
static void __get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/x86/kvm/x86.c
12075
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/x86/kvm/x86.c
1208
void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw)
arch/x86/kvm/x86.c
12087
static void __set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/x86/kvm/x86.c
12120
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
arch/x86/kvm/x86.c
12132
static void __get_sregs_common(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/x86/kvm/x86.c
1214
static void kvm_load_xfeatures(struct kvm_vcpu *vcpu, bool load_guest)
arch/x86/kvm/x86.c
12167
static void __get_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/x86/kvm/x86.c
12179
static void __get_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2)
arch/x86/kvm/x86.c
12197
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
12210
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
12236
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
12279
int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int idt_index,
arch/x86/kvm/x86.c
1231
static void kvm_load_guest_pkru(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12329
static bool kvm_is_valid_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/x86/kvm/x86.c
12354
static int __set_sregs_common(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs,
arch/x86/kvm/x86.c
12423
static int __set_sregs(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
arch/x86/kvm/x86.c
1243
static void kvm_load_host_pkru(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12449
static int __set_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2)
arch/x86/kvm/x86.c
12483
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
12501
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
12519
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
1258
static inline u64 kvm_guest_supported_xfd(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12583
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
12604
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/x86/kvm/x86.c
12627
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
arch/x86/kvm/x86.c
1264
int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr)
arch/x86/kvm/x86.c
12651
static void store_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12666
static int sync_regs(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12736
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12833
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12847
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
12878
static void kvm_xstate_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/x86.c
129
static void update_cr8_intercept(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
12925
void kvm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
arch/x86/kvm/x86.c
130
static void process_nmi(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
13062
void kvm_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector)
arch/x86/kvm/x86.c
13087
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
131
static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags);
arch/x86/kvm/x86.c
1310
int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
132
static void store_regs(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
13200
bool kvm_vcpu_is_reset_bsp(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13206
bool kvm_vcpu_is_bsp(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1323
static bool kvm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.c
1329
void kvm_post_set_cr4(struct kvm_vcpu *vcpu, unsigned long old_cr4, unsigned long cr4)
arch/x86/kvm/x86.c
133
static int sync_regs(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
134
static int kvm_vcpu_do_singlestep(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
13533
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
136
static int __set_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2);
arch/x86/kvm/x86.c
137
static void __get_sregs2(struct kvm_vcpu *vcpu, struct kvm_sregs2 *sregs2);
arch/x86/kvm/x86.c
1371
int kvm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.c
13731
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13741
unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13751
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13756
int kvm_arch_interrupt_allowed(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13761
unsigned long kvm_get_linear_rip(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13774
bool kvm_is_linear_rip(struct kvm_vcpu *vcpu, unsigned long linear_rip)
arch/x86/kvm/x86.c
13780
unsigned long kvm_get_rflags(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13791
static void __kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
arch/x86/kvm/x86.c
13799
void kvm_set_rflags(struct kvm_vcpu *vcpu, unsigned long rflags)
arch/x86/kvm/x86.c
13818
static void kvm_add_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13828
static u32 kvm_async_pf_gfn_slot(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13841
bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13846
static void kvm_del_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
arch/x86/kvm/x86.c
13873
static inline int apf_put_user_notpresent(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13881
static inline int apf_put_user_ready(struct kvm_vcpu *vcpu, u32 token)
arch/x86/kvm/x86.c
13889
static inline bool apf_pageready_slot_free(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13901
static bool kvm_can_deliver_async_pf(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13927
bool kvm_can_do_async_pf(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
13944
bool kvm_arch_async_page_not_present(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
13976
void kvm_arch_async_page_present(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
140
static void kvm_load_guest_fpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
14001
void kvm_arch_async_page_present_queued(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
14012
bool kvm_arch_can_dequeue_async_page_present(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1405
static void kvm_invalidate_pcid(struct kvm_vcpu *vcpu, unsigned long pcid)
arch/x86/kvm/x86.c
14053
bool kvm_arch_no_poll(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
141
static void kvm_put_guest_fpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
14110
void kvm_fixup_and_inject_pf_error(struct kvm_vcpu *vcpu, gva_t gva, u16 error_code)
arch/x86/kvm/x86.c
14140
int kvm_handle_memory_failure(struct kvm_vcpu *vcpu, int r,
arch/x86/kvm/x86.c
14164
int kvm_handle_invpcid(struct kvm_vcpu *vcpu, unsigned long type, gva_t gva)
arch/x86/kvm/x86.c
14240
static int complete_sev_es_emulated_mmio(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
14286
int kvm_sev_es_mmio_write(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned int bytes,
arch/x86/kvm/x86.c
14325
int kvm_sev_es_mmio_read(struct kvm_vcpu *vcpu, gpa_t gpa, unsigned int bytes,
arch/x86/kvm/x86.c
14363
static void advance_sev_es_emulated_pio(struct kvm_vcpu *vcpu, unsigned count, int size)
arch/x86/kvm/x86.c
14369
static int kvm_sev_es_outs(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.c
14372
static int complete_sev_es_emulated_outs(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
14383
static int kvm_sev_es_outs(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.c
14405
static int kvm_sev_es_ins(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.c
14408
static int complete_sev_es_emulated_ins(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
14421
static int kvm_sev_es_ins(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.c
14440
int kvm_sev_es_string_io(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.c
1448
int kvm_set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
arch/x86/kvm/x86.c
1497
int kvm_set_cr8(struct kvm_vcpu *vcpu, unsigned long cr8)
arch/x86/kvm/x86.c
1509
unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1518
static void kvm_update_dr0123(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1528
void kvm_update_dr7(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1543
static u64 kvm_dr6_fixed(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1555
int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val)
arch/x86/kvm/x86.c
1584
unsigned long kvm_get_dr(struct kvm_vcpu *vcpu, int dr)
arch/x86/kvm/x86.c
1601
int kvm_emulate_rdpmc(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
1696
static int kvm_get_feature_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data,
arch/x86/kvm/x86.c
1720
static int do_get_feature_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data)
arch/x86/kvm/x86.c
1726
static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer)
arch/x86/kvm/x86.c
1747
bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer)
arch/x86/kvm/x86.c
1756
static int set_efer(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
1799
bool kvm_msr_allowed(struct kvm_vcpu *vcpu, u32 index, u32 type)
arch/x86/kvm/x86.c
1848
static int __kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 data,
arch/x86/kvm/x86.c
1948
static int _kvm_set_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data,
arch/x86/kvm/x86.c
1954
static int kvm_set_msr_ignored_check(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
1967
static int __kvm_get_msr(struct kvm_vcpu *vcpu, u32 index, u64 *data,
arch/x86/kvm/x86.c
2008
int kvm_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data)
arch/x86/kvm/x86.c
2013
int kvm_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data)
arch/x86/kvm/x86.c
2018
static int kvm_get_msr_ignored_check(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
2025
int __kvm_emulate_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data)
arch/x86/kvm/x86.c
2031
int __kvm_emulate_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data)
arch/x86/kvm/x86.c
2037
int kvm_emulate_msr_read(struct kvm_vcpu *vcpu, u32 index, u64 *data)
arch/x86/kvm/x86.c
2046
int kvm_emulate_msr_write(struct kvm_vcpu *vcpu, u32 index, u64 data)
arch/x86/kvm/x86.c
2056
static void complete_userspace_rdmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2064
static int complete_emulated_msr_access(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2069
static int complete_emulated_rdmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2075
static int complete_fast_msr_access(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2080
static int complete_fast_rdmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2086
static int complete_fast_rdmsr_imm(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2107
static int kvm_msr_user_space(struct kvm_vcpu *vcpu, u32 index,
arch/x86/kvm/x86.c
2109
int (*completion)(struct kvm_vcpu *vcpu),
arch/x86/kvm/x86.c
2129
static int __kvm_emulate_rdmsr(struct kvm_vcpu *vcpu, u32 msr, int reg,
arch/x86/kvm/x86.c
2130
int (*complete_rdmsr)(struct kvm_vcpu *))
arch/x86/kvm/x86.c
2157
int kvm_emulate_rdmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2164
int kvm_emulate_rdmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg)
arch/x86/kvm/x86.c
2172
static int __kvm_emulate_wrmsr(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/x86.c
2193
int kvm_emulate_wrmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2200
int kvm_emulate_wrmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg)
arch/x86/kvm/x86.c
2206
int kvm_emulate_as_nop(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2211
int kvm_emulate_invd(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2218
fastpath_t handle_fastpath_invd(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2230
int kvm_handle_invalid_op(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2238
static int kvm_emulate_monitor_mwait(struct kvm_vcpu *vcpu, const char *insn)
arch/x86/kvm/x86.c
2257
int kvm_emulate_mwait(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2263
int kvm_emulate_monitor(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2269
static inline bool kvm_vcpu_exit_request(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2277
static fastpath_t __handle_fastpath_wrmsr(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/x86.c
2303
fastpath_t handle_fastpath_wrmsr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
2310
fastpath_t handle_fastpath_wrmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg)
arch/x86/kvm/x86.c
2319
static int do_get_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data)
arch/x86/kvm/x86.c
2324
static int do_set_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data)
arch/x86/kvm/x86.c
2450
static void kvm_write_system_time(struct kvm_vcpu *vcpu, gpa_t system_time,
arch/x86/kvm/x86.c
2523
static void kvm_vcpu_write_tsc_multiplier(struct kvm_vcpu *vcpu, u64 l1_multiplier);
arch/x86/kvm/x86.c
2525
static int set_tsc_khz(struct kvm_vcpu *vcpu, u32 user_tsc_khz, bool scale)
arch/x86/kvm/x86.c
2561
static int kvm_set_tsc_khz(struct kvm_vcpu *vcpu, u32 user_tsc_khz)
arch/x86/kvm/x86.c
2595
static u64 compute_guest_tsc(struct kvm_vcpu *vcpu, s64 kernel_ns)
arch/x86/kvm/x86.c
2611
static void kvm_track_tsc_matching(struct kvm_vcpu *vcpu, bool new_generation)
arch/x86/kvm/x86.c
2667
static u64 kvm_compute_l1_tsc_offset(struct kvm_vcpu *vcpu, u64 target_tsc)
arch/x86/kvm/x86.c
2676
u64 kvm_read_l1_tsc(struct kvm_vcpu *vcpu, u64 host_tsc)
arch/x86/kvm/x86.c
2708
static void kvm_vcpu_write_tsc_offset(struct kvm_vcpu *vcpu, u64 l1_offset)
arch/x86/kvm/x86.c
2735
static void kvm_vcpu_write_tsc_multiplier(struct kvm_vcpu *vcpu, u64 l1_multiplier)
arch/x86/kvm/x86.c
2769
static void __kvm_synchronize_tsc(struct kvm_vcpu *vcpu, u64 offset, u64 tsc,
arch/x86/kvm/x86.c
2822
static void kvm_synchronize_tsc(struct kvm_vcpu *vcpu, u64 *user_value)
arch/x86/kvm/x86.c
2890
static inline void adjust_tsc_offset_guest(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
2897
static inline void adjust_tsc_offset_host(struct kvm_vcpu *vcpu, s64 adjustment)
arch/x86/kvm/x86.c
3172
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
3265
struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
3312
int kvm_guest_time_update(struct kvm_vcpu *v)
arch/x86/kvm/x86.c
3523
static void kvm_gen_kvmclock_update(struct kvm_vcpu *v)
arch/x86/kvm/x86.c
3526
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
3548
static bool can_set_mci_status(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3557
static int set_msr_mce(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3629
static int kvm_pv_enable_async_pf(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/x86.c
3668
static int kvm_pv_enable_async_pf_int(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/x86.c
3684
static void kvmclock_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3690
static void kvm_vcpu_flush_tlb_all(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3699
static void kvm_vcpu_flush_tlb_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3724
static inline void kvm_vcpu_flush_tlb_current(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3736
void kvm_service_local_tlb_flush_requests(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3746
static void record_steal_time(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
3850
static bool is_xstate_managed_msr(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/x86.c
3877
static __always_inline void kvm_access_xstate_msr(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
3894
static void kvm_set_xstate_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3899
static void kvm_get_xstate_msr(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
3904
int kvm_set_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
4305
static int get_msr_mce(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata, bool host)
arch/x86/kvm/x86.c
4355
int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr_info)
arch/x86/kvm/x86.c
4663
static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs,
arch/x86/kvm/x86.c
4665
int (*do_msr)(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
4696
static int msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs __user *user_msrs,
arch/x86/kvm/x86.c
4697
int (*do_msr)(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
4756
static int kvm_ioctl_get_supported_hv_cpuid(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5156
static bool need_emulate_wbinvd(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
516
typedef int (*msr_access_t)(struct kvm_vcpu *vcpu, u32 index, u64 *data,
arch/x86/kvm/x86.c
5161
static DEFINE_PER_CPU(struct kvm_vcpu *, last_vcpu);
arch/x86/kvm/x86.c
5163
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
arch/x86/kvm/x86.c
519
static __always_inline int kvm_do_msr_access(struct kvm_vcpu *vcpu, u32 msr,
arch/x86/kvm/x86.c
5238
static void kvm_steal_time_set_preempted(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5285
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5314
static int kvm_vcpu_ioctl_get_lapic(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5325
static int kvm_vcpu_ioctl_set_lapic(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5341
static int kvm_cpu_accept_dm_intr(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5357
static int kvm_vcpu_ready_for_interrupt_injection(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5372
static int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5399
static int kvm_vcpu_ioctl_nmi(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5406
static int vcpu_ioctl_tpr_access_reporting(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5415
static int kvm_vcpu_ioctl_x86_setup_mce(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5462
static int kvm_vcpu_x86_set_ucna(struct kvm_vcpu *vcpu, struct kvm_x86_mce *mce, u64* banks)
arch/x86/kvm/x86.c
5481
static int kvm_vcpu_ioctl_x86_set_mce(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5534
static void kvm_vcpu_ioctl_x86_get_vcpu_events(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5627
static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
574
static inline void kvm_async_pf_hash_reset(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5740
static int kvm_vcpu_ioctl_x86_get_debugregs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5760
static int kvm_vcpu_ioctl_x86_set_debugregs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5789
static int kvm_vcpu_ioctl_x86_get_xsave2(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5815
static int kvm_vcpu_ioctl_x86_get_xsave(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5822
static int kvm_vcpu_ioctl_x86_set_xsave(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5843
static int kvm_vcpu_ioctl_x86_get_xcrs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5862
static int kvm_vcpu_ioctl_x86_set_xcrs(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5895
static int kvm_set_guest_paused(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
5904
static int kvm_arch_tsc_has_attr(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5920
static int kvm_arch_tsc_get_attr(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5940
static int kvm_arch_tsc_set_attr(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
5979
static int kvm_vcpu_ioctl_device_attr(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
6007
static int kvm_vcpu_ioctl_enable_cap(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
6069
static int kvm_translate_kvm_reg(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
6092
static int kvm_get_one_msr(struct kvm_vcpu *vcpu, u32 msr, u64 __user *user_val)
arch/x86/kvm/x86.c
6105
static int kvm_set_one_msr(struct kvm_vcpu *vcpu, u32 msr, u64 __user *user_val)
arch/x86/kvm/x86.c
6118
static int kvm_get_set_one_reg(struct kvm_vcpu *vcpu, unsigned int ioctl,
arch/x86/kvm/x86.c
6166
static int kvm_get_reg_list(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
6191
struct kvm_vcpu *vcpu = filp->private_data;
arch/x86/kvm/x86.c
6668
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf)
arch/x86/kvm/x86.c
6713
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
7158
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
7243
struct kvm_vcpu *vcpu = filp->private_data;
arch/x86/kvm/x86.c
770
void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
7770
static int vcpu_mmio_write(struct kvm_vcpu *vcpu, gpa_t addr, int len,
arch/x86/kvm/x86.c
7791
static int vcpu_mmio_read(struct kvm_vcpu *vcpu, gpa_t addr, int len, void *v)
arch/x86/kvm/x86.c
7813
void kvm_set_segment(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
7819
void kvm_get_segment(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
7825
gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u64 access,
arch/x86/kvm/x86.c
7840
gpa_t kvm_mmu_gva_to_gpa_read(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7850
gpa_t kvm_mmu_gva_to_gpa_write(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7862
gpa_t kvm_mmu_gva_to_gpa_system(struct kvm_vcpu *vcpu, gva_t gva,
arch/x86/kvm/x86.c
7871
struct kvm_vcpu *vcpu, u64 access,
arch/x86/kvm/x86.c
7906
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
7929
int kvm_read_guest_virt(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
7951
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
7963
struct kvm_vcpu *vcpu, u64 access,
arch/x86/kvm/x86.c
7996
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8008
int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu, gva_t addr, void *val,
arch/x86/kvm/x86.c
8019
static int kvm_check_emulate_insn(struct kvm_vcpu *vcpu, int emul_type,
arch/x86/kvm/x86.c
8026
int handle_ud(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
8053
static int vcpu_is_mmio_gpa(struct kvm_vcpu *vcpu, unsigned long gva,
arch/x86/kvm/x86.c
8068
static int vcpu_mmio_gva_to_gpa(struct kvm_vcpu *vcpu, unsigned long gva,
arch/x86/kvm/x86.c
8098
int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8111
int (*read_write_prepare)(struct kvm_vcpu *vcpu, void *val,
arch/x86/kvm/x86.c
8113
int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8115
int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8117
int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8122
static int read_prepare(struct kvm_vcpu *vcpu, void *val, int bytes)
arch/x86/kvm/x86.c
8134
static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8140
static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8146
static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val)
arch/x86/kvm/x86.c
8152
static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8159
static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa,
arch/x86/kvm/x86.c
8185
struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
822
static void kvm_queue_exception_vmexit(struct kvm_vcpu *vcpu, unsigned int vector,
arch/x86/kvm/x86.c
8239
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8317
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
837
static void kvm_multiple_exception(struct kvm_vcpu *vcpu, unsigned int nr,
arch/x86/kvm/x86.c
8393
static int emulator_pio_in_out(struct kvm_vcpu *vcpu, int size,
arch/x86/kvm/x86.c
8444
static int emulator_pio_in(struct kvm_vcpu *vcpu, int size,
arch/x86/kvm/x86.c
8454
static void complete_emulator_pio_in(struct kvm_vcpu *vcpu, void *val)
arch/x86/kvm/x86.c
8467
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8483
static int emulator_pio_out(struct kvm_vcpu *vcpu, int size,
arch/x86/kvm/x86.c
8498
static unsigned long get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/x86.c
8508
static int kvm_emulate_wbinvd_noskip(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
8525
int kvm_emulate_wbinvd(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
8558
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8587
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8685
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8714
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8737
struct kvm_vcpu *vcpu = emul_to_vcpu(ctxt);
arch/x86/kvm/x86.c
8942
static void toggle_interruptibility(struct kvm_vcpu *vcpu, u32 mask)
arch/x86/kvm/x86.c
8961
static void inject_emulated_exception(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
8974
static struct x86_emulate_ctxt *alloc_emulate_ctxt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
8991
static void init_emulate_ctxt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
900
void kvm_queue_exception(struct kvm_vcpu *vcpu, unsigned nr)
arch/x86/kvm/x86.c
9017
void kvm_inject_realmode_interrupt(struct kvm_vcpu *vcpu, int irq, int inc_eip)
arch/x86/kvm/x86.c
9039
static void prepare_emulation_failure_exit(struct kvm_vcpu *vcpu, u64 *data,
arch/x86/kvm/x86.c
907
void kvm_queue_exception_p(struct kvm_vcpu *vcpu, unsigned nr,
arch/x86/kvm/x86.c
9089
static void prepare_emulation_ctxt_failure_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9097
void __kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu, u64 *data,
arch/x86/kvm/x86.c
9104
void kvm_prepare_emulation_failure_exit(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9110
void kvm_prepare_event_vectoring_exit(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/x86/kvm/x86.c
9132
void kvm_prepare_unexpected_reason_exit(struct kvm_vcpu *vcpu, u64 exit_reason)
arch/x86/kvm/x86.c
914
static void kvm_queue_exception_e_p(struct kvm_vcpu *vcpu, unsigned nr,
arch/x86/kvm/x86.c
9144
static int handle_emulation_failure(struct kvm_vcpu *vcpu, int emulation_type)
arch/x86/kvm/x86.c
9172
static bool kvm_unprotect_and_retry_on_failure(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
920
void kvm_requeue_exception(struct kvm_vcpu *vcpu, unsigned int nr,
arch/x86/kvm/x86.c
9210
static int complete_emulated_mmio(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
9211
static int complete_emulated_pio(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.c
9228
static int kvm_vcpu_do_singlestep(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9243
int kvm_skip_emulated_instruction(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9268
static bool kvm_is_code_breakpoint_inhibited(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9283
static bool kvm_vcpu_check_code_breakpoint(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
9399
int x86_decode_emulated_instruction(struct kvm_vcpu *vcpu, int emulation_type,
arch/x86/kvm/x86.c
9416
int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/kvm/x86.c
949
int kvm_complete_insn_gp(struct kvm_vcpu *vcpu, int err)
arch/x86/kvm/x86.c
960
static int complete_emulated_insn_gp(struct kvm_vcpu *vcpu, int err)
arch/x86/kvm/x86.c
9638
int kvm_emulate_instruction(struct kvm_vcpu *vcpu, int emulation_type)
arch/x86/kvm/x86.c
9644
int kvm_emulate_instruction_from_buffer(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
9651
static int complete_fast_pio_out_port_0x7e(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9657
static int complete_fast_pio_out(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
9667
static int kvm_fast_pio_out(struct kvm_vcpu *vcpu, int size,
arch/x86/kvm/x86.c
9692
static int complete_fast_pio_in(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.c
971
void kvm_inject_page_fault(struct kvm_vcpu *vcpu, struct x86_exception *fault)
arch/x86/kvm/x86.c
9713
static int kvm_fast_pio_in(struct kvm_vcpu *vcpu, int size,
arch/x86/kvm/x86.c
9734
int kvm_fast_pio(struct kvm_vcpu *vcpu, int size, unsigned short port, int in)
arch/x86/kvm/x86.c
9801
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.c
988
void kvm_inject_emulated_page_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.c
9932
struct kvm_vcpu *vcpu;
arch/x86/kvm/x86.h
148
void kvm_service_local_tlb_flush_requests(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
149
int kvm_check_nested_events(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
152
static inline void kvm_leave_nested(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
165
static inline void kvm_nested_vmexit_handle_ibrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
186
static inline bool kvm_can_set_cpuid_and_feature_msrs(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
191
static inline void kvm_set_mp_state(struct kvm_vcpu *vcpu, int mp_state)
arch/x86/kvm/x86.h
198
static inline bool kvm_is_exception_pending(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
205
static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
212
static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector,
arch/x86/kvm/x86.h
220
static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
225
static inline bool kvm_event_needs_reinjection(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
236
static inline bool is_protmode(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
241
static inline bool is_long_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
250
static inline bool is_64_bit_mode(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
262
static inline bool is_64_bit_hypercall(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
281
static inline bool mmu_is_nested(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
286
static inline bool is_pae(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
291
static inline bool is_pse(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
296
static inline bool is_paging(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
301
static inline bool is_pae_paging(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
306
static inline u8 vcpu_virt_addr_bits(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
335
static inline bool is_noncanonical_address(u64 la, struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
344
static inline bool is_noncanonical_msr_address(u64 la, struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
349
static inline bool is_noncanonical_base_address(u64 la, struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
354
static inline bool is_noncanonical_invlpg_address(u64 la, struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
359
static inline void vcpu_cache_mmio_info(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
377
static inline bool vcpu_match_mmio_gen(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
388
static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva)
arch/x86/kvm/x86.h
396
static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva)
arch/x86/kvm/x86.h
405
static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa)
arch/x86/kvm/x86.h
414
static inline unsigned long kvm_register_read(struct kvm_vcpu *vcpu, int reg)
arch/x86/kvm/x86.h
421
static inline void kvm_register_write(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
448
void kvm_inject_realmode_interrupt(struct kvm_vcpu *vcpu, int irq, int inc_eip);
arch/x86/kvm/x86.h
453
int kvm_guest_time_update(struct kvm_vcpu *v);
arch/x86/kvm/x86.h
455
int kvm_read_guest_virt(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
459
int kvm_write_guest_virt_system(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
463
int handle_ud(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
465
void kvm_deliver_exception_payload(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
468
int kvm_mtrr_set_msr(struct kvm_vcpu *vcpu, u32 msr, u64 data);
arch/x86/kvm/x86.h
469
int kvm_mtrr_get_msr(struct kvm_vcpu *vcpu, u32 msr, u64 *pdata);
arch/x86/kvm/x86.h
470
void kvm_fixup_and_inject_pf_error(struct kvm_vcpu *vcpu, gva_t gva, u16 error_code);
arch/x86/kvm/x86.h
471
int x86_decode_emulated_instruction(struct kvm_vcpu *vcpu, int emulation_type,
arch/x86/kvm/x86.h
473
int x86_emulate_instruction(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
arch/x86/kvm/x86.h
475
fastpath_t handle_fastpath_wrmsr(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
476
fastpath_t handle_fastpath_wrmsr_imm(struct kvm_vcpu *vcpu, u32 msr, int reg);
arch/x86/kvm/x86.h
477
fastpath_t handle_fastpath_hlt(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
478
fastpath_t handle_fastpath_invd(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
532
static inline void kvm_pr_unimpl_wrmsr(struct kvm_vcpu *vcpu, u32 msr, u64 data)
arch/x86/kvm/x86.h
538
static inline void kvm_pr_unimpl_rdmsr(struct kvm_vcpu *vcpu, u32 msr)
arch/x86/kvm/x86.h
544
static inline u64 nsec_to_cycles(struct kvm_vcpu *vcpu, u64 nsec)
arch/x86/kvm/x86.h
599
static __always_inline void kvm_before_interrupt(struct kvm_vcpu *vcpu,
arch/x86/kvm/x86.h
605
static __always_inline void kvm_after_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
610
static inline bool kvm_handling_nmi_from_guest(struct kvm_vcpu *vcpu)
arch/x86/kvm/x86.h
654
int kvm_handle_memory_failure(struct kvm_vcpu *vcpu, int r,
arch/x86/kvm/x86.h
656
int kvm_handle_invpcid(struct kvm_vcpu *vcpu, unsigned long type, gva_t gva);
arch/x86/kvm/x86.h
657
bool kvm_msr_allowed(struct kvm_vcpu *vcpu, u32 index, u32 type);
arch/x86/kvm/x86.h
680
static inline bool __kvm_is_valid_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
arch/x86/kvm/x86.h
715
int kvm_sev_es_mmio_write(struct kvm_vcpu *vcpu, gpa_t src, unsigned int bytes,
arch/x86/kvm/x86.h
717
int kvm_sev_es_mmio_read(struct kvm_vcpu *vcpu, gpa_t src, unsigned int bytes,
arch/x86/kvm/x86.h
719
int kvm_sev_es_string_io(struct kvm_vcpu *vcpu, unsigned int size,
arch/x86/kvm/x86.h
728
int ____kvm_emulate_hypercall(struct kvm_vcpu *vcpu, int cpl,
arch/x86/kvm/x86.h
729
int (*complete_hypercall)(struct kvm_vcpu *));
arch/x86/kvm/x86.h
741
int kvm_emulate_hypercall(struct kvm_vcpu *vcpu);
arch/x86/kvm/x86.h
748
static inline bool kvm_is_valid_u_s_cet(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/xen.c
108
void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
1150
int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data)
arch/x86/kvm/xen.c
127
struct kvm_vcpu *vcpu = container_of(timer, struct kvm_vcpu,
arch/x86/kvm/xen.c
1276
int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/xen.c
1409
static int kvm_xen_hypercall_set_result(struct kvm_vcpu *vcpu, u64 result)
arch/x86/kvm/xen.c
1415
static int kvm_xen_hypercall_complete_userspace(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
1433
static bool wait_pending_event(struct kvm_vcpu *vcpu, int nr_ports,
arch/x86/kvm/xen.c
1471
static bool kvm_xen_schedop_poll(struct kvm_vcpu *vcpu, bool longmode,
arch/x86/kvm/xen.c
153
static int xen_get_guest_pvclock(struct kvm_vcpu *vcpu,
arch/x86/kvm/xen.c
1573
struct kvm_vcpu *vcpu = timer_container_of(vcpu, t,
arch/x86/kvm/xen.c
1580
static bool kvm_xen_hcall_sched_op(struct kvm_vcpu *vcpu, bool longmode,
arch/x86/kvm/xen.c
1604
static bool kvm_xen_hcall_vcpu_op(struct kvm_vcpu *vcpu, bool longmode, int cmd,
arch/x86/kvm/xen.c
1659
static bool kvm_xen_hcall_set_timer_op(struct kvm_vcpu *vcpu, uint64_t timeout,
arch/x86/kvm/xen.c
1674
int kvm_xen_hypercall(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
1771
static void kvm_xen_check_poller(struct kvm_vcpu *vcpu, int port)
arch/x86/kvm/xen.c
1795
struct kvm_vcpu *vcpu;
arch/x86/kvm/xen.c
186
static void kvm_xen_start_timer(struct kvm_vcpu *vcpu, u64 guest_abs,
arch/x86/kvm/xen.c
1971
struct kvm_vcpu *vcpu;
arch/x86/kvm/xen.c
2262
static bool kvm_xen_hcall_evtchn_send(struct kvm_vcpu *vcpu, u64 param, u64 *r)
arch/x86/kvm/xen.c
2297
void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
2312
void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
314
static void kvm_xen_stop_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.c
321
static void kvm_xen_update_runstate_guest(struct kvm_vcpu *v, bool atomic)
arch/x86/kvm/xen.c
34
static bool kvm_xen_hcall_evtchn_send(struct kvm_vcpu *vcpu, u64 param, u64 *r);
arch/x86/kvm/xen.c
587
void kvm_xen_update_runstate(struct kvm_vcpu *v, int state)
arch/x86/kvm/xen.c
618
void kvm_xen_inject_vcpu_vector(struct kvm_vcpu *v)
arch/x86/kvm/xen.c
639
void kvm_xen_inject_pending_events(struct kvm_vcpu *v)
arch/x86/kvm/xen.c
697
int __kvm_xen_has_interrupt(struct kvm_vcpu *v)
arch/x86/kvm/xen.c
906
int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data)
arch/x86/kvm/xen.h
103
static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
111
void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
113
static inline int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data)
arch/x86/kvm/xen.h
126
static inline void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
130
static inline void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
134
static inline void kvm_xen_sw_enable_lapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
153
static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
158
static inline void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
162
static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
167
static inline int kvm_xen_has_pending_timer(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
172
static inline void kvm_xen_inject_timer_irqs(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
176
static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
181
static inline bool kvm_xen_is_tsc_leaf(struct kvm_vcpu *vcpu, u32 function)
arch/x86/kvm/xen.h
187
int kvm_xen_hypercall(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
193
void kvm_xen_update_runstate(struct kvm_vcpu *vcpu, int state);
arch/x86/kvm/xen.h
195
static inline void kvm_xen_runstate_set_running(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
20
int __kvm_xen_has_interrupt(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
200
static inline void kvm_xen_runstate_set_preempted(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
21
void kvm_xen_inject_pending_events(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
22
void kvm_xen_inject_vcpu_vector(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
23
int kvm_xen_vcpu_set_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
arch/x86/kvm/xen.h
24
int kvm_xen_vcpu_get_attr(struct kvm_vcpu *vcpu, struct kvm_xen_vcpu_attr *data);
arch/x86/kvm/xen.h
28
int kvm_xen_write_hypercall_page(struct kvm_vcpu *vcpu, u64 data);
arch/x86/kvm/xen.h
32
void kvm_xen_init_vcpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
33
void kvm_xen_destroy_vcpu(struct kvm_vcpu *vcpu);
arch/x86/kvm/xen.h
40
static inline void kvm_xen_sw_enable_lapic(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
53
static inline bool kvm_xen_is_tsc_leaf(struct kvm_vcpu *vcpu, u32 function)
arch/x86/kvm/xen.h
82
static inline int kvm_xen_has_interrupt(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
92
static inline bool kvm_xen_has_pending_events(struct kvm_vcpu *vcpu)
arch/x86/kvm/xen.h
98
static inline bool kvm_xen_timer_enabled(struct kvm_vcpu *vcpu)
drivers/s390/crypto/vfio_ap_ops.c
351
static int vfio_ap_validate_nib(struct kvm_vcpu *vcpu, dma_addr_t *nib)
drivers/s390/crypto/vfio_ap_ops.c
418
struct kvm_vcpu *vcpu)
drivers/s390/crypto/vfio_ap_ops.c
581
static int handle_pqap(struct kvm_vcpu *vcpu)
include/kvm/arm_arch_timer.h
100
void kvm_timer_vcpu_reset(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
101
void kvm_timer_vcpu_init(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
102
void kvm_timer_sync_nested(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
103
void kvm_timer_sync_user(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
104
bool kvm_timer_should_notify_user(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
105
void kvm_timer_update_run(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
106
void kvm_timer_vcpu_terminate(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
110
int kvm_arm_timer_set_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr);
include/kvm/arm_arch_timer.h
111
int kvm_arm_timer_get_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr);
include/kvm/arm_arch_timer.h
112
int kvm_arm_timer_has_attr(struct kvm_vcpu *vcpu, struct kvm_device_attr *attr);
include/kvm/arm_arch_timer.h
116
void kvm_timer_vcpu_load(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
117
void kvm_timer_vcpu_put(struct kvm_vcpu *vcpu);
include/kvm/arm_arch_timer.h
129
#define timer_context_to_vcpu(ctx) container_of((ctx), struct kvm_vcpu, arch.timer_cpu.timers[(ctx)->timer_id])
include/kvm/arm_arch_timer.h
133
u64 kvm_arm_timer_read_sysreg(struct kvm_vcpu *vcpu,
include/kvm/arm_arch_timer.h
136
void kvm_arm_timer_write_sysreg(struct kvm_vcpu *vcpu,
include/kvm/arm_arch_timer.h
86
void get_timer_map(struct kvm_vcpu *vcpu, struct timer_map *map);
include/kvm/arm_arch_timer.h
99
int kvm_timer_enable(struct kvm_vcpu *vcpu);
include/kvm/arm_hypercalls.h
11
static inline u32 smccc_get_function(struct kvm_vcpu *vcpu)
include/kvm/arm_hypercalls.h
16
static inline unsigned long smccc_get_arg1(struct kvm_vcpu *vcpu)
include/kvm/arm_hypercalls.h
21
static inline unsigned long smccc_get_arg2(struct kvm_vcpu *vcpu)
include/kvm/arm_hypercalls.h
26
static inline unsigned long smccc_get_arg3(struct kvm_vcpu *vcpu)
include/kvm/arm_hypercalls.h
31
static inline void smccc_set_retval(struct kvm_vcpu *vcpu,
include/kvm/arm_hypercalls.h
47
int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu);
include/kvm/arm_hypercalls.h
48
int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices);
include/kvm/arm_hypercalls.h
49
int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
include/kvm/arm_hypercalls.h
50
int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg);
include/kvm/arm_hypercalls.h
9
int kvm_smccc_call_handler(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
105
static inline u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
110
static inline void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
112
static inline void kvm_pmu_set_counter_value_user(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
114
static inline u64 kvm_pmu_implemented_counter_mask(struct kvm_vcpu *vcpu)
include/kvm/arm_pmu.h
118
static inline u64 kvm_pmu_accessible_counter_mask(struct kvm_vcpu *vcpu)
include/kvm/arm_pmu.h
122
static inline void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
123
static inline void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
124
static inline void kvm_pmu_reprogram_counter_mask(struct kvm_vcpu *vcpu, u64 val) {}
include/kvm/arm_pmu.h
125
static inline void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
126
static inline void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
127
static inline bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu)
include/kvm/arm_pmu.h
131
static inline void kvm_pmu_update_run(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
132
static inline void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val) {}
include/kvm/arm_pmu.h
133
static inline void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val) {}
include/kvm/arm_pmu.h
134
static inline void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
136
static inline int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
141
static inline int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
146
static inline int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
151
static inline int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu)
include/kvm/arm_pmu.h
155
static inline u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1)
include/kvm/arm_pmu.h
161
static inline void kvm_pmu_update_vcpu_events(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
162
static inline void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
163
static inline void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
164
static inline void kvm_vcpu_reload_pmu(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
185
static inline u64 kvm_vcpu_read_pmcr(struct kvm_vcpu *vcpu)
include/kvm/arm_pmu.h
190
static inline bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx)
include/kvm/arm_pmu.h
195
static inline void kvm_pmu_nested_transition(struct kvm_vcpu *vcpu) {}
include/kvm/arm_pmu.h
42
u64 kvm_pmu_get_counter_value(struct kvm_vcpu *vcpu, u64 select_idx);
include/kvm/arm_pmu.h
43
void kvm_pmu_set_counter_value(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
include/kvm/arm_pmu.h
44
void kvm_pmu_set_counter_value_user(struct kvm_vcpu *vcpu, u64 select_idx, u64 val);
include/kvm/arm_pmu.h
45
u64 kvm_pmu_implemented_counter_mask(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
46
u64 kvm_pmu_accessible_counter_mask(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
47
u64 kvm_pmu_get_pmceid(struct kvm_vcpu *vcpu, bool pmceid1);
include/kvm/arm_pmu.h
48
void kvm_pmu_vcpu_init(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
49
void kvm_pmu_vcpu_destroy(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
50
void kvm_pmu_reprogram_counter_mask(struct kvm_vcpu *vcpu, u64 val);
include/kvm/arm_pmu.h
51
void kvm_pmu_flush_hwstate(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
52
void kvm_pmu_sync_hwstate(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
53
bool kvm_pmu_should_notify_user(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
54
void kvm_pmu_update_run(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
55
void kvm_pmu_software_increment(struct kvm_vcpu *vcpu, u64 val);
include/kvm/arm_pmu.h
56
void kvm_pmu_handle_pmcr(struct kvm_vcpu *vcpu, u64 val);
include/kvm/arm_pmu.h
57
void kvm_pmu_set_counter_event_type(struct kvm_vcpu *vcpu, u64 data,
include/kvm/arm_pmu.h
59
void kvm_vcpu_reload_pmu(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
60
int kvm_arm_pmu_v3_set_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
62
int kvm_arm_pmu_v3_get_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
64
int kvm_arm_pmu_v3_has_attr(struct kvm_vcpu *vcpu,
include/kvm/arm_pmu.h
66
int kvm_arm_pmu_v3_enable(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
69
void kvm_vcpu_pmu_restore_guest(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
70
void kvm_vcpu_pmu_restore_host(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
92
u64 kvm_vcpu_read_pmcr(struct kvm_vcpu *vcpu);
include/kvm/arm_pmu.h
93
bool kvm_pmu_counter_is_hyp(struct kvm_vcpu *vcpu, unsigned int idx);
include/kvm/arm_pmu.h
94
void kvm_pmu_nested_transition(struct kvm_vcpu *vcpu);
include/kvm/arm_psci.h
22
static inline int kvm_psci_version(struct kvm_vcpu *vcpu)
include/kvm/arm_psci.h
42
int kvm_psci_call(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
130
struct kvm_vcpu *vcpu; /* SGIs and PPIs: The VCPU
include/kvm/arm_vgic.h
135
struct kvm_vcpu *target_vcpu; /* The VCPU that this interrupt should
include/kvm/arm_vgic.h
189
struct kvm_vcpu *redist_vcpu;
include/kvm/arm_vgic.h
389
int kvm_vgic_vcpu_init(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
390
int kvm_vgic_vcpu_nv_init(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
393
void kvm_vgic_vcpu_destroy(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
398
int kvm_vgic_inject_irq(struct kvm *kvm, struct kvm_vcpu *vcpu,
include/kvm/arm_vgic.h
400
int kvm_vgic_map_phys_irq(struct kvm_vcpu *vcpu, unsigned int host_irq,
include/kvm/arm_vgic.h
402
int kvm_vgic_unmap_phys_irq(struct kvm_vcpu *vcpu, unsigned int vintid);
include/kvm/arm_vgic.h
403
int kvm_vgic_get_map(struct kvm_vcpu *vcpu, unsigned int vintid);
include/kvm/arm_vgic.h
404
bool kvm_vgic_map_is_active(struct kvm_vcpu *vcpu, unsigned int vintid);
include/kvm/arm_vgic.h
406
int kvm_vgic_vcpu_pending_irq(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
408
void kvm_vgic_load(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
409
void kvm_vgic_put(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
411
u16 vgic_v3_get_eisr(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
412
u16 vgic_v3_get_elrsr(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
413
u64 vgic_v3_get_misr(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
420
bool kvm_vcpu_has_pending_irqs(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
421
void kvm_vgic_sync_hwstate(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
422
void kvm_vgic_flush_hwstate(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
423
void kvm_vgic_reset_mapped_irq(struct kvm_vcpu *vcpu, u32 vintid);
include/kvm/arm_vgic.h
424
void kvm_vgic_process_async_update(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
426
void vgic_v3_dispatch_sgi(struct kvm_vcpu *vcpu, u64 reg, bool allow_group1);
include/kvm/arm_vgic.h
445
int kvm_vgic_set_owner(struct kvm_vcpu *vcpu, unsigned int intid, void *owner);
include/kvm/arm_vgic.h
454
int vgic_v4_load(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
455
void vgic_v4_commit(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
456
int vgic_v4_put(struct kvm_vcpu *vcpu);
include/kvm/arm_vgic.h
458
bool vgic_state_is_nested(struct kvm_vcpu *vcpu);
include/kvm/iodev.h
10
struct kvm_vcpu;
include/kvm/iodev.h
18
int (*read)(struct kvm_vcpu *vcpu,
include/kvm/iodev.h
23
int (*write)(struct kvm_vcpu *vcpu,
include/kvm/iodev.h
42
static inline int kvm_iodevice_read(struct kvm_vcpu *vcpu,
include/kvm/iodev.h
50
static inline int kvm_iodevice_write(struct kvm_vcpu *vcpu,
include/linux/kvm_dirty_ring.h
58
static inline void kvm_dirty_ring_push(struct kvm_vcpu *vcpu,
include/linux/kvm_dirty_ring.h
83
void kvm_dirty_ring_push(struct kvm_vcpu *vcpu, u32 slot, u64 offset);
include/linux/kvm_dirty_ring.h
85
bool kvm_dirty_ring_check_request(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1015
static inline struct kvm_vcpu *kvm_get_vcpu_by_id(struct kvm *kvm, int id)
include/linux/kvm_host.h
1017
struct kvm_vcpu *vcpu = NULL;
include/linux/kvm_host.h
1038
void vcpu_load(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1039
void vcpu_put(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1084
static inline struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
1212
struct kvm_memslots *kvm_vcpu_memslots(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1213
struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1307
static inline kvm_pfn_t kvm_faultin_pfn(struct kvm_vcpu *vcpu, gfn_t gfn,
include/linux/kvm_host.h
1379
bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1380
unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1383
void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1385
int __kvm_vcpu_map(struct kvm_vcpu *vcpu, gpa_t gpa, struct kvm_host_map *map,
include/linux/kvm_host.h
1387
void kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map);
include/linux/kvm_host.h
1389
static inline int kvm_vcpu_map(struct kvm_vcpu *vcpu, gpa_t gpa,
include/linux/kvm_host.h
1395
static inline int kvm_vcpu_map_readonly(struct kvm_vcpu *vcpu, gpa_t gpa,
include/linux/kvm_host.h
1401
static inline void kvm_vcpu_map_mark_dirty(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1408
unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn);
include/linux/kvm_host.h
1409
unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable);
include/linux/kvm_host.h
1410
int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data, int offset,
include/linux/kvm_host.h
1412
int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa, void *data,
include/linux/kvm_host.h
1414
int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data,
include/linux/kvm_host.h
1416
int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, const void *data,
include/linux/kvm_host.h
1418
int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data,
include/linux/kvm_host.h
1523
void kvm_sigset_activate(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1524
void kvm_sigset_deactivate(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1526
void kvm_vcpu_halt(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1527
bool kvm_vcpu_block(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1528
void kvm_arch_vcpu_blocking(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1529
void kvm_arch_vcpu_unblocking(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1530
bool kvm_vcpu_wake_up(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1533
void __kvm_vcpu_kick(struct kvm_vcpu *vcpu, bool wait);
include/linux/kvm_host.h
1535
static inline void kvm_vcpu_kick(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
1541
int kvm_vcpu_yield_to(struct kvm_vcpu *target);
include/linux/kvm_host.h
1542
void kvm_vcpu_on_spin(struct kvm_vcpu *vcpu, bool yield_to_kernel_mode);
include/linux/kvm_host.h
1568
vm_fault_t kvm_arch_vcpu_fault(struct kvm_vcpu *vcpu, struct vm_fault *vmf);
include/linux/kvm_host.h
1592
int kvm_arch_vcpu_ioctl_get_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu);
include/linux/kvm_host.h
1593
int kvm_arch_vcpu_ioctl_set_fpu(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu);
include/linux/kvm_host.h
1595
int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1598
int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs);
include/linux/kvm_host.h
1599
int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs);
include/linux/kvm_host.h
1600
int kvm_arch_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1602
int kvm_arch_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1604
int kvm_arch_vcpu_ioctl_get_mpstate(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1606
int kvm_arch_vcpu_ioctl_set_mpstate(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1608
int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
1610
int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1612
void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
include/linux/kvm_host.h
1613
void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1615
int kvm_arch_vcpu_create(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1616
void kvm_arch_vcpu_postcreate(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1617
void kvm_arch_vcpu_destroy(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1624
void kvm_arch_create_vcpu_debugfs(struct kvm_vcpu *vcpu, struct dentry *debugfs_dentry);
include/linux/kvm_host.h
1626
static inline void kvm_create_vcpu_debugfs(struct kvm_vcpu *vcpu) {}
include/linux/kvm_host.h
1647
bool kvm_vcpu_has_events(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1648
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1649
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1650
int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1651
bool kvm_arch_dy_runnable(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1652
bool kvm_arch_dy_has_pending_interrupt(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1653
bool kvm_arch_vcpu_preempted_in_kernel(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1718
static inline struct rcuwait *kvm_arch_vcpu_get_wait(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
1731
static inline bool __kvm_vcpu_wake_up(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
1736
static inline bool kvm_vcpu_is_blocking(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
1756
unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
1775
int kvm_cpu_has_pending_timer(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
183
BUILD_BUG_ON((unsigned)(nr) >= (sizeof_field(struct kvm_vcpu, requests) * 8) - KVM_REQUEST_ARCH_BASE); \
include/linux/kvm_host.h
223
int kvm_io_bus_write(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr,
include/linux/kvm_host.h
2239
static inline void __kvm_make_request(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2249
static __always_inline void kvm_make_request(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
225
int kvm_io_bus_write_cookie(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx,
include/linux/kvm_host.h
2263
static inline void kvm_make_request_and_kick(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
227
int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr,
include/linux/kvm_host.h
2270
static inline bool kvm_request_pending(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2275
static inline bool kvm_test_request(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2280
static inline void kvm_clear_request(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2285
static inline bool kvm_check_request(int req, struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2372
static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val)
include/linux/kvm_host.h
2376
static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val)
include/linux/kvm_host.h
2383
static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val)
include/linux/kvm_host.h
2387
static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val)
include/linux/kvm_host.h
2398
struct kvm_vcpu *kvm_get_running_vcpu(void);
include/linux/kvm_host.h
2399
struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void);
include/linux/kvm_host.h
241
struct kvm_vcpu *vcpu;
include/linux/kvm_host.h
2418
static inline bool vcpu_valid_wakeup(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2424
static inline bool vcpu_valid_wakeup(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2432
bool kvm_arch_no_poll(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
2434
static inline bool kvm_arch_no_poll(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2443
int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
2445
static inline int kvm_arch_vcpu_run_pid_change(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2452
static inline void kvm_handle_signal_exit(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
2458
static inline int kvm_xfer_to_guest_mode_handle_work(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
249
void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
2493
static inline void kvm_prepare_memory_fault_exit(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
250
void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
251
bool kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
include/linux/kvm_host.h
253
int kvm_async_pf_wakeup_all(struct kvm_vcpu *vcpu);
include/linux/kvm_host.h
2593
long kvm_arch_vcpu_pre_fault_memory(struct kvm_vcpu *vcpu,
include/linux/kvm_host.h
559
static inline int kvm_vcpu_exiting_guest_mode(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
718
static inline int kvm_arch_vcpu_memslots_id(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
956
static inline void kvm_vcpu_srcu_read_lock(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
965
static inline void kvm_vcpu_srcu_read_unlock(struct kvm_vcpu *vcpu)
include/linux/kvm_host.h
991
static inline struct kvm_vcpu *kvm_get_vcpu(struct kvm *kvm, int i)
include/linux/kvm_irqfd.h
62
struct kvm_vcpu *irq_bypass_vcpu;
include/linux/kvm_types.h
47
struct kvm_vcpu;
include/trace/events/kvm.h
347
TP_PROTO(struct kvm_vcpu *vcpu),
tools/testing/selftests/kvm/access_tracking_perf_test.c
300
static void assert_ucall(struct kvm_vcpu *vcpu, uint64_t expected_ucall)
tools/testing/selftests/kvm/access_tracking_perf_test.c
327
struct kvm_vcpu *vcpu = vcpu_args->vcpu;
tools/testing/selftests/kvm/arch_timer.c
40
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
tools/testing/selftests/kvm/arch_timer.c
51
struct kvm_vcpu *vcpu = vcpus[vcpu_idx];
tools/testing/selftests/kvm/arm64/aarch32_id_regs.c
121
static void test_user_raz_invariant(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/aarch32_id_regs.c
143
static bool vcpu_aarch64_only(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/aarch32_id_regs.c
155
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/aarch32_id_regs.c
52
static void test_guest_raz(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/aarch32_id_regs.c
92
static void test_user_raz_wi(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
1032
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
865
static void kvm_set_cntxct(struct kvm_vcpu *vcpu, uint64_t cnt,
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
874
static void handle_sync(struct kvm_vcpu *vcpu, struct ucall *uc)
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
898
static void test_run(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
925
static void test_init_timer_irq(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/arch_timer_edge_cases.c
936
static void test_vm_create(struct kvm_vm **vm, struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/arm64/at.c
124
static void run_test(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/at.c
148
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/at.c
99
static void handle_sync(struct kvm_vcpu *vcpu, struct ucall *uc)
tools/testing/selftests/kvm/arm64/debug-exceptions.c
426
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/debug-exceptions.c
467
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/debug-exceptions.c
579
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
100
static void vcpu_run_expect_sync(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/external_aborts.c
124
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
159
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
175
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
207
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
242
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
279
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
307
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
31
static struct kvm_vm *vm_create_with_dabt_handler(struct kvm_vcpu **vcpu, void *guest_code,
tools/testing/selftests/kvm/arm64/external_aborts.c
333
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
389
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/external_aborts.c
45
static void vcpu_inject_sea(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/external_aborts.c
53
static bool vcpu_has_ras(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/external_aborts.c
65
static void vcpu_inject_serror(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/external_aborts.c
78
static void __vcpu_run_expect(struct kvm_vcpu *vcpu, unsigned int cmd)
tools/testing/selftests/kvm/arm64/external_aborts.c
95
static void vcpu_run_expect_done(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/get-reg-list.c
106
bool check_supported_reg(struct kvm_vcpu *vcpu, __u64 reg)
tools/testing/selftests/kvm/arm64/get-reg-list.c
116
void finalize_vcpu(struct kvm_vcpu *vcpu, struct vcpu_reg_list *c)
tools/testing/selftests/kvm/arm64/get-reg-list.c
87
static bool check_supported_feat_reg(struct kvm_vcpu *vcpu, __u64 reg)
tools/testing/selftests/kvm/arm64/hello_el2.c
45
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/host_sve.c
79
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/hypercalls.c
163
static void steal_time_init(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/hypercalls.c
175
static void test_fw_regs_before_vm_start(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/hypercalls.c
230
static void test_fw_regs_after_vm_start(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/hypercalls.c
260
static struct kvm_vm *test_vm_create(struct kvm_vcpu **vcpu)
tools/testing/selftests/kvm/arm64/hypercalls.c
271
static void test_guest_stage(struct kvm_vm **vm, struct kvm_vcpu **vcpu)
tools/testing/selftests/kvm/arm64/hypercalls.c
300
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/idreg-idst.c
102
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/idreg-idst.c
57
static void test_run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/idreg-idst.c
81
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/kvm-uuid.c
40
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/no-vgic-v3.c
119
static void test_run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/no-vgic-v3.c
143
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/no-vgic-v3.c
162
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/page_fault_test.c
540
static void setup_abort_handlers(struct kvm_vm *vm, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/arm64/page_fault_test.c
656
static void vcpu_run_loop(struct kvm_vm *vm, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/arm64/page_fault_test.c
702
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/psci_test.c
102
static void enter_guest(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/psci_test.c
111
static void assert_vcpu_reset(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/psci_test.c
144
struct kvm_vcpu *source, *target;
tools/testing/selftests/kvm/arm64/psci_test.c
181
struct kvm_vcpu *source, *target;
tools/testing/selftests/kvm/arm64/psci_test.c
239
struct kvm_vcpu *source, *target;
tools/testing/selftests/kvm/arm64/psci_test.c
75
static void vcpu_power_off(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/psci_test.c
84
static struct kvm_vm *setup_vm(void *guest_code, struct kvm_vcpu **source,
tools/testing/selftests/kvm/arm64/psci_test.c
85
struct kvm_vcpu **target)
tools/testing/selftests/kvm/arm64/sea_to_user.c
180
static void vcpu_inject_sea(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/sea_to_user.c
188
static void run_vm(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/sea_to_user.c
251
static struct kvm_vm *vm_create_with_sea_handler(struct kvm_vcpu **vcpu)
tools/testing/selftests/kvm/arm64/sea_to_user.c
318
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/set_id_regs.c
375
static uint64_t test_reg_set_success(struct kvm_vcpu *vcpu, uint64_t reg,
tools/testing/selftests/kvm/arm64/set_id_regs.c
398
static void test_reg_set_fail(struct kvm_vcpu *vcpu, uint64_t reg,
tools/testing/selftests/kvm/arm64/set_id_regs.c
432
static void test_vm_ftr_id_regs(struct kvm_vcpu *vcpu, bool aarch64_only)
tools/testing/selftests/kvm/arm64/set_id_regs.c
480
static void test_user_set_mpam_reg(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
575
static void test_user_set_mte_reg(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
637
static void test_guest_reg_read(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
670
static void test_clidr(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
696
static void test_ctr(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
709
static void test_id_reg(struct kvm_vcpu *vcpu, u32 id)
tools/testing/selftests/kvm/arm64/set_id_regs.c
719
static void test_vcpu_ftr_id_regs(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
728
static void test_vcpu_non_ftr_id_regs(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
737
static void test_assert_id_reg_unchanged(struct kvm_vcpu *vcpu, uint32_t encoding)
tools/testing/selftests/kvm/arm64/set_id_regs.c
746
static void test_reset_preserves_id_regs(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/set_id_regs.c
769
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
113
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
135
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
148
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
161
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
176
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
189
static void expect_call_denied(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/smccc_filter.c
204
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
220
static void expect_call_fwd_to_user(struct kvm_vcpu *vcpu, uint32_t func_id,
tools/testing/selftests/kvm/arm64/smccc_filter.c
242
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/smccc_filter.c
73
static struct kvm_vm *setup_vm(struct kvm_vcpu **vcpu)
tools/testing/selftests/kvm/arm64/smccc_filter.c
94
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vcpu_width_config.c
24
struct kvm_vcpu *vcpu0, *vcpu1;
tools/testing/selftests/kvm/arm64/vcpu_width_config.c
50
struct kvm_vcpu *vcpu0, *vcpu1;
tools/testing/selftests/kvm/arm64/vgic_init.c
337
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
358
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
408
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
606
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
639
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
681
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
69
static int run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/arm64/vgic_init.c
722
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/arm64/vgic_init.c
76
struct kvm_vcpu *vcpus[])
tools/testing/selftests/kvm/arm64/vgic_init.c
939
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vgic_irq.c
606
struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/arm64/vgic_irq.c
671
static void run_guest_cmd(struct kvm_vcpu *vcpu, int gic_fd,
tools/testing/selftests/kvm/arm64/vgic_irq.c
752
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vgic_irq.c
963
struct kvm_vcpu *vcpu = arg;
tools/testing/selftests/kvm/arm64/vgic_irq.c
986
struct kvm_vcpu *vcpus[2];
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
264
struct kvm_vcpu *vcpu = data;
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
29
static struct kvm_vcpu **vcpus;
tools/testing/selftests/kvm/arm64/vgic_lpi_stress.c
339
vcpus = malloc(test_data.nr_cpus * sizeof(struct kvm_vcpu *));
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
30
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
446
static void run_vcpu(struct kvm_vcpu *vcpu, uint64_t pmcr_n)
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
466
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
495
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
537
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/coalesced_io_test.c
112
static void vcpu_run_and_verify_coalesced_io(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/coalesced_io_test.c
150
static void test_coalesced_io(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/coalesced_io_test.c
188
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/coalesced_io_test.c
71
static void vcpu_run_and_verify_io_exit(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/demand_paging_test.c
34
struct kvm_vcpu *vcpu = vcpu_args->vcpu;
tools/testing/selftests/kvm/dirty_log_perf_test.c
38
struct kvm_vcpu *vcpu = vcpu_args->vcpu;
tools/testing/selftests/kvm/dirty_log_test.c
257
static void dirty_log_collect_dirty_pages(struct kvm_vcpu *vcpu, int slot,
tools/testing/selftests/kvm/dirty_log_test.c
264
static void clear_log_collect_dirty_pages(struct kvm_vcpu *vcpu, int slot,
tools/testing/selftests/kvm/dirty_log_test.c
281
static void default_after_vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/dirty_log_test.c
361
static void dirty_ring_collect_dirty_pages(struct kvm_vcpu *vcpu, int slot,
tools/testing/selftests/kvm/dirty_log_test.c
382
static void dirty_ring_after_vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/dirty_log_test.c
406
void (*collect_dirty_pages) (struct kvm_vcpu *vcpu, int slot,
tools/testing/selftests/kvm/dirty_log_test.c
410
void (*after_vcpu_run)(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/dirty_log_test.c
461
static void log_mode_collect_dirty_pages(struct kvm_vcpu *vcpu, int slot,
tools/testing/selftests/kvm/dirty_log_test.c
472
static void log_mode_after_vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/dirty_log_test.c
482
struct kvm_vcpu *vcpu = data;
tools/testing/selftests/kvm/dirty_log_test.c
577
static struct kvm_vm *create_vm(enum vm_guest_mode mode, struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/dirty_log_test.c
601
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/get-reg-list.c
114
void __weak finalize_vcpu(struct kvm_vcpu *vcpu, struct vcpu_reg_list *c)
tools/testing/selftests/kvm/get-reg-list.c
131
static struct kvm_vcpu *vcpu_config_get_vcpu(struct vcpu_reg_list *c, struct kvm_vm *vm)
tools/testing/selftests/kvm/get-reg-list.c
134
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/get-reg-list.c
143
static struct kvm_vcpu *vcpu_config_get_vcpu(struct vcpu_reg_list *c, struct kvm_vm *vm)
tools/testing/selftests/kvm/get-reg-list.c
171
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/get-reg-list.c
84
bool __weak check_supported_reg(struct kvm_vcpu *vcpu, __u64 reg)
tools/testing/selftests/kvm/guest_memfd_test.c
431
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/guest_print_test.c
114
static void do_vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/guest_print_test.c
125
static void run_test(struct kvm_vcpu *vcpu, const char *expected_printf,
tools/testing/selftests/kvm/guest_print_test.c
170
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/guest_print_test.c
192
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/guest_print_test.c
46
static void run_test(struct kvm_vcpu *vcpu, const char *expected_printf,
tools/testing/selftests/kvm/guest_print_test.c
52
static void fn(struct kvm_vcpu *vcpu, T a, T b) \
tools/testing/selftests/kvm/hardware_disable_test.c
35
struct kvm_vcpu *vcpu = arg;
tools/testing/selftests/kvm/hardware_disable_test.c
85
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/include/arm64/arch_timer.h
158
static inline u32 vcpu_get_vtimer_irq(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/arm64/arch_timer.h
170
static inline u32 vcpu_get_ptimer_irq(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/arm64/processor.h
130
void aarch64_vcpu_setup(struct kvm_vcpu *vcpu, struct kvm_vcpu_init *init);
tools/testing/selftests/kvm/include/arm64/processor.h
131
struct kvm_vcpu *aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id,
tools/testing/selftests/kvm/include/arm64/processor.h
174
void vcpu_init_descriptor_tables(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/arm64/processor.h
322
static inline bool vcpu_has_el2(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/arm64/processor.h
334
static __always_inline u64 ctxt_reg_alias(struct kvm_vcpu *vcpu, u32 encoding)
tools/testing/selftests/kvm/include/arm64/vgic.h
33
void kvm_irq_write_ispendr(int gic_fd, uint32_t intid, struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/arm64/vgic.h
34
void kvm_irq_write_isactiver(int gic_fd, uint32_t intid, struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_test_harness.h
15
struct kvm_vcpu *vcpu; \
tools/testing/selftests/kvm/include/kvm_test_harness.h
27
static void __suite##_##test(struct kvm_vcpu *vcpu); \
tools/testing/selftests/kvm/include/kvm_test_harness.h
34
static void __suite##_##test(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/kvm_util.h
1044
void *guest_code, struct kvm_vcpu *vcpus[]);
tools/testing/selftests/kvm/include/kvm_util.h
1048
struct kvm_vcpu *vcpus[])
tools/testing/selftests/kvm/include/kvm_util.h
1056
struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1064
static inline struct kvm_vm *__vm_create_with_one_vcpu(struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1072
static inline struct kvm_vm *vm_create_with_one_vcpu(struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1079
struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1085
struct kvm_vcpu *vm_recreate_with_one_vcpu(struct kvm_vm *vm);
tools/testing/selftests/kvm/include/kvm_util.h
1160
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
1162
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1165
static inline void vcpu_dump(FILE *stream, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
1178
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id);
tools/testing/selftests/kvm/include/kvm_util.h
1179
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code);
tools/testing/selftests/kvm/include/kvm_util.h
1181
static inline struct kvm_vcpu *vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id,
tools/testing/selftests/kvm/include/kvm_util.h
1184
struct kvm_vcpu *vcpu = vm_arch_vcpu_add(vm, vcpu_id);
tools/testing/selftests/kvm/include/kvm_util.h
1192
struct kvm_vcpu *vm_arch_vcpu_recreate(struct kvm_vm *vm, uint32_t vcpu_id);
tools/testing/selftests/kvm/include/kvm_util.h
1194
static inline struct kvm_vcpu *vm_vcpu_recreate(struct kvm_vm *vm,
tools/testing/selftests/kvm/include/kvm_util.h
1200
void vcpu_arch_free(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
380
static __always_inline void static_assert_is_vcpu(struct kvm_vcpu *vcpu) { }
tools/testing/selftests/kvm/include/kvm_util.h
717
struct kvm_vcpu *__vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id);
tools/testing/selftests/kvm/include/kvm_util.h
747
void vcpu_run(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
748
int _vcpu_run(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
750
static inline int __vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/kvm_util.h
755
void vcpu_run_complete_io(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
756
struct kvm_reg_list *vcpu_get_reg_list(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
758
static inline void vcpu_enable_cap(struct kvm_vcpu *vcpu, uint32_t cap,
tools/testing/selftests/kvm/include/kvm_util.h
766
static inline void vcpu_guest_debug_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
772
static inline void vcpu_mp_state_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
777
static inline void vcpu_mp_state_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
783
static inline void vcpu_regs_get(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
tools/testing/selftests/kvm/include/kvm_util.h
788
static inline void vcpu_regs_set(struct kvm_vcpu *vcpu, struct kvm_regs *regs)
tools/testing/selftests/kvm/include/kvm_util.h
792
static inline void vcpu_sregs_get(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
tools/testing/selftests/kvm/include/kvm_util.h
797
static inline void vcpu_sregs_set(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
tools/testing/selftests/kvm/include/kvm_util.h
801
static inline int _vcpu_sregs_set(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
tools/testing/selftests/kvm/include/kvm_util.h
805
static inline void vcpu_fpu_get(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
tools/testing/selftests/kvm/include/kvm_util.h
809
static inline void vcpu_fpu_set(struct kvm_vcpu *vcpu, struct kvm_fpu *fpu)
tools/testing/selftests/kvm/include/kvm_util.h
814
static inline int __vcpu_get_reg(struct kvm_vcpu *vcpu, uint64_t id, void *addr)
tools/testing/selftests/kvm/include/kvm_util.h
820
static inline int __vcpu_set_reg(struct kvm_vcpu *vcpu, uint64_t id, uint64_t val)
tools/testing/selftests/kvm/include/kvm_util.h
826
static inline uint64_t vcpu_get_reg(struct kvm_vcpu *vcpu, uint64_t id)
tools/testing/selftests/kvm/include/kvm_util.h
836
static inline void vcpu_set_reg(struct kvm_vcpu *vcpu, uint64_t id, uint64_t val)
tools/testing/selftests/kvm/include/kvm_util.h
846
static inline void vcpu_events_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
851
static inline void vcpu_events_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
858
static inline void vcpu_nested_state_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
863
static inline int __vcpu_nested_state_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
869
static inline void vcpu_nested_state_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/kvm_util.h
875
static inline int vcpu_get_stats_fd(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/kvm_util.h
912
static inline int __vcpu_has_device_attr(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
918
static inline void vcpu_has_device_attr(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
924
static inline int __vcpu_device_attr_get(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
930
static inline void vcpu_device_attr_get(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
936
static inline int __vcpu_device_attr_set(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
942
static inline void vcpu_device_attr_set(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
959
void *vcpu_map_dirty_ring(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/kvm_util.h
978
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...);
tools/testing/selftests/kvm/include/memstress.h
28
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/include/memstress.h
72
void memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu *vcpus[]);
tools/testing/selftests/kvm/include/riscv/processor.h
125
void vcpu_init_vector_tables(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/riscv/processor.h
65
bool __vcpu_has_ext(struct kvm_vcpu *vcpu, uint64_t ext);
tools/testing/selftests/kvm/include/riscv/processor.h
67
static inline bool __vcpu_has_isa_ext(struct kvm_vcpu *vcpu, uint64_t isa_ext)
tools/testing/selftests/kvm/include/riscv/processor.h
72
static inline bool __vcpu_has_sbi_ext(struct kvm_vcpu *vcpu, uint64_t sbi_ext)
tools/testing/selftests/kvm/include/timer_test.h
39
extern struct kvm_vcpu *vcpus[];
tools/testing/selftests/kvm/include/ucall_common.h
34
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/ucall_common.h
41
uint64_t get_ucall(struct kvm_vcpu *vcpu, struct ucall *uc);
tools/testing/selftests/kvm/include/x86/evmcs.h
241
int vcpu_enable_evmcs(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/x86/hyperv.h
356
const struct kvm_cpuid2 *vcpu_get_supported_hv_cpuid(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/x86/hyperv.h
357
void vcpu_set_hv_cpuid(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/x86/processor.h
1050
void vcpu_init_cpuid(struct kvm_vcpu *vcpu, const struct kvm_cpuid2 *cpuid);
tools/testing/selftests/kvm/include/x86/processor.h
1052
static inline void vcpu_get_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/x86/processor.h
1057
static inline struct kvm_cpuid_entry2 *__vcpu_get_cpuid_entry(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1069
static inline struct kvm_cpuid_entry2 *vcpu_get_cpuid_entry(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1075
static inline int __vcpu_set_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/x86/processor.h
1089
static inline void vcpu_set_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/include/x86/processor.h
1098
void vcpu_set_cpuid_property(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1101
void vcpu_set_cpuid_maxphyaddr(struct kvm_vcpu *vcpu, uint8_t maxphyaddr);
tools/testing/selftests/kvm/include/x86/processor.h
1103
void vcpu_clear_cpuid_entry(struct kvm_vcpu *vcpu, uint32_t function);
tools/testing/selftests/kvm/include/x86/processor.h
1105
static inline bool vcpu_cpuid_has(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1114
void vcpu_set_or_clear_cpuid_feature(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1118
static inline void vcpu_set_cpuid_feature(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1125
static inline void vcpu_clear_cpuid_feature(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
1131
uint64_t vcpu_get_msr(struct kvm_vcpu *vcpu, uint64_t msr_index);
tools/testing/selftests/kvm/include/x86/processor.h
1132
int _vcpu_set_msr(struct kvm_vcpu *vcpu, uint64_t msr_index, uint64_t msr_value);
tools/testing/selftests/kvm/include/x86/processor.h
888
struct kvm_x86_state *vcpu_save_state(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/include/x86/processor.h
889
void vcpu_load_state(struct kvm_vcpu *vcpu, struct kvm_x86_state *state);
tools/testing/selftests/kvm/include/x86/processor.h
897
static inline void vcpu_msrs_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
906
static inline void vcpu_msrs_set(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs)
tools/testing/selftests/kvm/include/x86/processor.h
914
static inline void vcpu_debugregs_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
919
static inline void vcpu_debugregs_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
924
static inline void vcpu_xsave_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
929
static inline void vcpu_xsave2_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
934
static inline void vcpu_xsave_set(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
939
static inline void vcpu_xcrs_get(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/processor.h
944
static inline void vcpu_xcrs_set(struct kvm_vcpu *vcpu, struct kvm_xcrs *xcrs)
tools/testing/selftests/kvm/include/x86/sev.h
57
struct kvm_vcpu **cpu);
tools/testing/selftests/kvm/include/x86/smm.h
11
void setup_smram(struct kvm_vm *vm, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/include/x86/smm.h
15
void inject_smi(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/irqfd_test.c
92
struct kvm_vcpu *unused;
tools/testing/selftests/kvm/kvm_binary_stats_test.c
188
struct kvm_vcpu **vcpus;
tools/testing/selftests/kvm/kvm_binary_stats_test.c
216
vcpus = malloc(sizeof(struct kvm_vcpu *) * max_vm * max_vcpu);
tools/testing/selftests/kvm/kvm_page_table_test.c
185
struct kvm_vcpu *vcpu = data;
tools/testing/selftests/kvm/kvm_page_table_test.c
56
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
tools/testing/selftests/kvm/lib/arm64/processor.c
297
void aarch64_vcpu_setup(struct kvm_vcpu *vcpu, struct kvm_vcpu_init *init)
tools/testing/selftests/kvm/lib/arm64/processor.c
400
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
tools/testing/selftests/kvm/lib/arm64/processor.c
411
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
tools/testing/selftests/kvm/lib/arm64/processor.c
416
static struct kvm_vcpu *__aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id,
tools/testing/selftests/kvm/lib/arm64/processor.c
421
struct kvm_vcpu *vcpu = __vm_vcpu_add(vm, vcpu_id);
tools/testing/selftests/kvm/lib/arm64/processor.c
435
struct kvm_vcpu *aarch64_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id,
tools/testing/selftests/kvm/lib/arm64/processor.c
438
struct kvm_vcpu *vcpu = __aarch64_vcpu_add(vm, vcpu_id, init);
tools/testing/selftests/kvm/lib/arm64/processor.c
445
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/arm64/processor.c
450
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...)
tools/testing/selftests/kvm/lib/arm64/processor.c
475
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/arm64/processor.c
497
void vcpu_init_descriptor_tables(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/arm64/ucall.c
22
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/arm64/vgic.c
154
static void vgic_poke_irq(int gic_fd, uint32_t intid, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/lib/arm64/vgic.c
186
void kvm_irq_write_ispendr(int gic_fd, uint32_t intid, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/arm64/vgic.c
191
void kvm_irq_write_isactiver(int gic_fd, uint32_t intid, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1329
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/kvm_util.c
1343
struct kvm_vcpu *__vm_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/kvm_util.c
1345
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/kvm_util.c
1732
int _vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1750
void vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1757
void vcpu_run_complete_io(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1775
struct kvm_reg_list *vcpu_get_reg_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1789
void *vcpu_map_dirty_ring(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
1970
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/kvm_util.c
551
void *guest_code, struct kvm_vcpu *vcpus[])
tools/testing/selftests/kvm/lib/kvm_util.c
568
struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/lib/kvm_util.c
572
struct kvm_vcpu *vcpus[1];
tools/testing/selftests/kvm/lib/kvm_util.c
616
__weak struct kvm_vcpu *vm_arch_vcpu_recreate(struct kvm_vm *vm,
tools/testing/selftests/kvm/lib/kvm_util.c
622
struct kvm_vcpu *vm_recreate_with_one_vcpu(struct kvm_vm *vm)
tools/testing/selftests/kvm/lib/kvm_util.c
754
__weak void vcpu_arch_free(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
771
static void vm_vcpu_rm(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/kvm_util.c
791
struct kvm_vcpu *vcpu, *tmp;
tools/testing/selftests/kvm/lib/loongarch/processor.c
171
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
tools/testing/selftests/kvm/lib/loongarch/processor.c
175
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/loongarch/processor.c
230
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...)
tools/testing/selftests/kvm/lib/loongarch/processor.c
249
static void loongarch_set_reg(struct kvm_vcpu *vcpu, uint64_t id, uint64_t val)
tools/testing/selftests/kvm/lib/loongarch/processor.c
254
static void loongarch_get_csr(struct kvm_vcpu *vcpu, uint64_t id, void *addr)
tools/testing/selftests/kvm/lib/loongarch/processor.c
262
static void loongarch_set_csr(struct kvm_vcpu *vcpu, uint64_t id, uint64_t val)
tools/testing/selftests/kvm/lib/loongarch/processor.c
270
static void loongarch_vcpu_setup(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/loongarch/processor.c
359
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/loongarch/processor.c
364
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/loongarch/processor.c
381
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
tools/testing/selftests/kvm/lib/loongarch/ucall.c
25
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/memstress.c
256
void __weak memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu **vcpus)
tools/testing/selftests/kvm/lib/memstress.c
41
static struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
tools/testing/selftests/kvm/lib/memstress.c
89
struct kvm_vcpu *vcpus[],
tools/testing/selftests/kvm/lib/riscv/processor.c
193
void riscv_vcpu_mmu_setup(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/riscv/processor.c
20
bool __vcpu_has_ext(struct kvm_vcpu *vcpu, uint64_t ext)
tools/testing/selftests/kvm/lib/riscv/processor.c
236
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
tools/testing/selftests/kvm/lib/riscv/processor.c
309
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
tools/testing/selftests/kvm/lib/riscv/processor.c
314
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/riscv/processor.c
321
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/riscv/processor.c
358
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...)
tools/testing/selftests/kvm/lib/riscv/processor.c
407
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/riscv/processor.c
443
void vcpu_init_vector_tables(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/riscv/ucall.c
14
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/s390/diag318_test_handler.c
28
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/s390/processor.c
158
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
tools/testing/selftests/kvm/lib/s390/processor.c
163
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/s390/processor.c
169
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/s390/processor.c
195
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...)
tools/testing/selftests/kvm/lib/s390/processor.c
215
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
tools/testing/selftests/kvm/lib/s390/processor.c
221
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/s390/ucall.c
9
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/ucall_common.c
143
uint64_t get_ucall(struct kvm_vcpu *vcpu, struct ucall *uc)
tools/testing/selftests/kvm/lib/x86/hyperv.c
28
void vcpu_set_hv_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/hyperv.c
61
const struct kvm_cpuid2 *vcpu_get_supported_hv_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/memstress.c
104
void memstress_setup_nested(struct kvm_vm *vm, int nr_vcpus, struct kvm_vcpu *vcpus[])
tools/testing/selftests/kvm/lib/x86/processor.c
1016
void vcpu_set_cpuid_property(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/lib/x86/processor.c
1033
void vcpu_clear_cpuid_entry(struct kvm_vcpu *vcpu, uint32_t function)
tools/testing/selftests/kvm/lib/x86/processor.c
1044
void vcpu_set_or_clear_cpuid_feature(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/lib/x86/processor.c
1062
uint64_t vcpu_get_msr(struct kvm_vcpu *vcpu, uint64_t msr_index)
tools/testing/selftests/kvm/lib/x86/processor.c
1077
int _vcpu_set_msr(struct kvm_vcpu *vcpu, uint64_t msr_index, uint64_t msr_value)
tools/testing/selftests/kvm/lib/x86/processor.c
1092
void vcpu_args_set(struct kvm_vcpu *vcpu, unsigned int num, ...)
tools/testing/selftests/kvm/lib/x86/processor.c
1126
void vcpu_arch_dump(FILE *stream, struct kvm_vcpu *vcpu, uint8_t indent)
tools/testing/selftests/kvm/lib/x86/processor.c
1208
static void vcpu_save_xsave_state(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/lib/x86/processor.c
1222
struct kvm_x86_state *vcpu_save_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
1279
void vcpu_load_state(struct kvm_vcpu *vcpu, struct kvm_x86_state *state)
tools/testing/selftests/kvm/lib/x86/processor.c
1449
void setup_smram(struct kvm_vm *vm, struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/lib/x86/processor.c
1464
void inject_smi(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
644
static void vcpu_init_sregs(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
678
static void vcpu_init_xcrs(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
776
void assert_on_unhandled_exception(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
811
void vcpu_arch_set_entry_point(struct kvm_vcpu *vcpu, void *guest_code)
tools/testing/selftests/kvm/lib/x86/processor.c
820
struct kvm_vcpu *vm_arch_vcpu_add(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/x86/processor.c
825
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/lib/x86/processor.c
871
struct kvm_vcpu *vm_arch_vcpu_recreate(struct kvm_vm *vm, uint32_t vcpu_id)
tools/testing/selftests/kvm/lib/x86/processor.c
873
struct kvm_vcpu *vcpu = __vm_vcpu_add(vm, vcpu_id);
tools/testing/selftests/kvm/lib/x86/processor.c
880
void vcpu_arch_free(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/processor.c
999
void vcpu_init_cpuid(struct kvm_vcpu *vcpu, const struct kvm_cpuid2 *cpuid)
tools/testing/selftests/kvm/lib/x86/sev.c
162
struct kvm_vcpu **cpu)
tools/testing/selftests/kvm/lib/x86/sev.c
169
struct kvm_vcpu *cpus[1];
tools/testing/selftests/kvm/lib/x86/ucall.c
45
void *ucall_arch_get_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/lib/x86/vmx.c
28
int vcpu_enable_evmcs(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/memslot_modification_stress_test.c
37
struct kvm_vcpu *vcpu = vcpu_args->vcpu;
tools/testing/selftests/kvm/memslot_perf_test.c
143
struct kvm_vcpu *vcpu = data->vcpu;
tools/testing/selftests/kvm/memslot_perf_test.c
86
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/mmu_stress_test.c
102
static void assert_sync_stage(struct kvm_vcpu *vcpu, int stage)
tools/testing/selftests/kvm/mmu_stress_test.c
110
static void run_vcpu(struct kvm_vcpu *vcpu, int stage)
tools/testing/selftests/kvm/mmu_stress_test.c
120
struct kvm_vcpu *vcpu = info->vcpu;
tools/testing/selftests/kvm/mmu_stress_test.c
205
static pthread_t *spawn_workers(struct kvm_vm *vm, struct kvm_vcpu **vcpus,
tools/testing/selftests/kvm/mmu_stress_test.c
288
struct kvm_vcpu **vcpus;
tools/testing/selftests/kvm/mmu_stress_test.c
78
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/pre_fault_memory_test.c
169
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/pre_fault_memory_test.c
64
static void pre_fault_memory(struct kvm_vcpu *vcpu, u64 base_gpa, u64 offset,
tools/testing/selftests/kvm/riscv/ebreak_test.c
39
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/riscv/get-reg-list.c
164
static int override_vector_reg_size(struct kvm_vcpu *vcpu, struct vcpu_reg_sublist *s,
tools/testing/selftests/kvm/riscv/get-reg-list.c
196
void finalize_vcpu(struct kvm_vcpu *vcpu, struct vcpu_reg_list *c)
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
531
static void run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
559
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
577
struct kvm_vcpu *vcpu = NULL;
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
587
static void test_vm_setup_snapshot_mem(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
603
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/riscv/sbi_pmu_test.c
619
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/rseq_test.c
211
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/cmma_test.c
230
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/cmma_test.c
264
static void assert_exit_was_hypercall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/cmma_test.c
275
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/cmma_test.c
450
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/cmma_test.c
508
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/debug_test.c
19
static struct kvm_vm *test_step_int_1(struct kvm_vcpu **vcpu, void *guest_code,
tools/testing/selftests/kvm/s390/debug_test.c
44
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/debug_test.c
81
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/irq_routing.c
28
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/keyop.c
147
static void test_init(struct kvm_vcpu *vcpu, int fault_in)
tools/testing/selftests/kvm/s390/keyop.c
167
static void test_rrbe(struct kvm_vcpu *vcpu, int fault_in)
tools/testing/selftests/kvm/s390/keyop.c
197
static void test_iske(struct kvm_vcpu *vcpu, int fault_in)
tools/testing/selftests/kvm/s390/keyop.c
221
static void test_sske(struct kvm_vcpu *vcpu, int fault_in)
tools/testing/selftests/kvm/s390/keyop.c
249
void (*test)(struct kvm_vcpu *vcpu, int fault_in_location);
tools/testing/selftests/kvm/s390/keyop.c
258
static void run_test(void (*the_test)(struct kvm_vcpu *, int), int fault_in_location)
tools/testing/selftests/kvm/s390/keyop.c
260
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/keyop.c
40
static int _get_skeys(struct kvm_vcpu *vcpu, unsigned char skeys[])
tools/testing/selftests/kvm/s390/keyop.c
51
static void get_skeys(struct kvm_vcpu *vcpu, unsigned char skeys[])
tools/testing/selftests/kvm/s390/keyop.c
58
static void set_skeys(struct kvm_vcpu *vcpu, unsigned char skeys[])
tools/testing/selftests/kvm/s390/keyop.c
71
static int do_keyop(struct kvm_vcpu *vcpu, int op, unsigned long page_idx, unsigned char skey)
tools/testing/selftests/kvm/s390/keyop.c
89
static void fault_in_buffer(struct kvm_vcpu *vcpu, int where, int cur_loc)
tools/testing/selftests/kvm/s390/memop.c
117
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/memop.c
121
static void print_memop(struct kvm_vcpu *vcpu, const struct kvm_s390_mem_op *ksmo)
tools/testing/selftests/kvm/s390/memop.c
168
struct kvm_vcpu *vcpu = info.vcpu;
tools/testing/selftests/kvm/s390/memop.c
246
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/memop.c
273
struct kvm_vcpu *__vcpu = __info.vcpu; \
tools/testing/selftests/kvm/s390/resets.c
112
static void assert_initial_noclear(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
129
static void assert_initial(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
172
static void assert_normal_noclear(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
182
static void assert_normal(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
190
static void inject_irq(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
205
static struct kvm_vm *create_vm(struct kvm_vcpu **vcpu)
tools/testing/selftests/kvm/s390/resets.c
218
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/resets.c
241
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/resets.c
264
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/resets.c
60
static void test_one_reg(struct kvm_vcpu *vcpu, uint64_t id, uint64_t value)
tools/testing/selftests/kvm/s390/resets.c
68
static void assert_noirq(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/resets.c
85
static void assert_clear(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
52
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/sync_regs_test.c
116
void test_req_and_verify_all_valid_regs(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/sync_regs_test.c
142
void test_set_and_verify_various_reg_values(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/sync_regs_test.c
181
void test_clear_kvm_dirty_regs_bits(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/sync_regs_test.c
206
void (*test)(struct kvm_vcpu *vcpu);
tools/testing/selftests/kvm/s390/sync_regs_test.c
217
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/sync_regs_test.c
74
void test_read_invalid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/sync_regs_test.c
95
void test_set_invalid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/s390/tprot.c
187
struct kvm_vcpu *__vcpu = (vcpup); \
tools/testing/selftests/kvm/s390/tprot.c
207
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/user_operexec.c
23
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/user_operexec.c
46
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/s390/user_operexec.c
80
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/set_memory_region_test.c
110
static struct kvm_vm *spawn_vm(struct kvm_vcpu **vcpu, pthread_t *vcpu_thread,
tools/testing/selftests/kvm/set_memory_region_test.c
181
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/set_memory_region_test.c
275
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/set_memory_region_test.c
330
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/set_memory_region_test.c
54
struct kvm_vcpu *vcpu = data;
tools/testing/selftests/kvm/set_memory_region_test.c
577
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/steal_time.c
159
static bool is_steal_time_supported(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/steal_time.c
169
static void steal_time_init(struct kvm_vcpu *vcpu, uint32_t i)
tools/testing/selftests/kvm/steal_time.c
269
static bool is_steal_time_supported(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/steal_time.c
279
static void steal_time_init(struct kvm_vcpu *vcpu, uint32_t i)
tools/testing/selftests/kvm/steal_time.c
346
static bool is_steal_time_supported(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/steal_time.c
367
static void steal_time_init(struct kvm_vcpu *vcpu, uint32_t i)
tools/testing/selftests/kvm/steal_time.c
418
static void run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/steal_time.c
438
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/steal_time.c
65
static bool is_steal_time_supported(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/steal_time.c
70
static void steal_time_init(struct kvm_vcpu *vcpu, uint32_t i)
tools/testing/selftests/kvm/system_counter_offset_test.c
119
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/system_counter_offset_test.c
29
static void check_preconditions(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/system_counter_offset_test.c
36
static void setup_system_counter(struct kvm_vcpu *vcpu, struct test_case *test)
tools/testing/selftests/kvm/system_counter_offset_test.c
89
static void enter_guest(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/amx_test.c
234
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/aperfmperf_test.c
127
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/apic_bus_clock_test.c
100
static void test_apic_bus_clock(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/apic_bus_clock_test.c
127
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/cpuid_test.c
120
static void run_vcpu(struct kvm_vcpu *vcpu, int stage)
tools/testing/selftests/kvm/x86/cpuid_test.c
155
static void set_cpuid_after_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/cpuid_test.c
198
static void test_get_cpuid2(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/cpuid_test.c
219
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/cr4_cpuid_sync_test.c
64
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/debug_regs.c
72
static void vcpu_skip_insn(struct kvm_vcpu *vcpu, int insn_len)
tools/testing/selftests/kvm/x86/debug_regs.c
85
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/dirty_log_page_splitting_test.c
68
struct kvm_vcpu *vcpu = vcpu_args->vcpu;
tools/testing/selftests/kvm/x86/evmcs_smm_controls_test.c
79
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/exit_on_emulation_failure_test.c
23
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/fastops_test.c
198
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/feature_msrs_test.c
46
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/fix_hypercall_test.c
109
static void test_fix_hypercall(struct kvm_vcpu *vcpu, bool disable_quirk)
tools/testing/selftests/kvm/x86/fix_hypercall_test.c
89
static void enter_guest(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/flds_emulation.h
20
static inline void handle_flds_emulation_failure_exit(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hwcr_msr_test.c
11
void test_hwcr_bit(struct kvm_vcpu *vcpu, unsigned int bit)
tools/testing/selftests/kvm/x86/hwcr_msr_test.c
36
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_clock.c
176
static void host_check_tsc_msr_rdtsc(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_clock.c
208
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
115
static void test_hv_cpuid_e2big(struct kvm_vm *vm, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
133
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_cpuid.c
25
static void test_hv_cpuid(struct kvm_vcpu *vcpu, bool evmcs_expected)
tools/testing/selftests/kvm/x86/hyperv_evmcs.c
193
void inject_nmi(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_evmcs.c
205
static struct kvm_vcpu *save_restore_vm(struct kvm_vm *vm,
tools/testing/selftests/kvm/x86/hyperv_evmcs.c
206
struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_evmcs.c
237
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_extended_hypercalls.c
40
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_features.c
117
static void vcpu_reset_hv_cpuid(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_features.c
133
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_features.c
522
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_ipi.c
211
struct kvm_vcpu *vcpu = (struct kvm_vcpu *)arg;
tools/testing/selftests/kvm/x86/hyperv_ipi.c
225
static void cancel_join_vcpu_thread(pthread_t thread, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_ipi.c
245
struct kvm_vcpu *vcpu[3];
tools/testing/selftests/kvm/x86/hyperv_svm_test.c
154
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
539
struct kvm_vcpu *vcpu = (struct kvm_vcpu *)arg;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
562
static void cancel_join_vcpu_thread(pthread_t thread, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
582
struct kvm_vcpu *vcpu[3];
tools/testing/selftests/kvm/x86/kvm_buslock_test.c
77
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/kvm_clock_test.c
105
static void enter_guest(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/kvm_clock_test.c
137
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/kvm_pv_test.c
112
static void enter_guest(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/kvm_pv_test.c
138
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/kvm_pv_test.c
203
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/monitor_mwait_test.c
71
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/msrs_test.c
213
static bool vcpu_has_reg(struct kvm_vcpu *vcpu, u64 reg)
tools/testing/selftests/kvm/x86/msrs_test.c
251
static void host_test_kvm_reg(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/msrs_test.c
288
static void host_test_msr(struct kvm_vcpu *vcpu, u64 guest_val)
tools/testing/selftests/kvm/x86/msrs_test.c
318
static void do_vcpu_run(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/msrs_test.c
342
static void vcpus_run(struct kvm_vcpu **vcpus, const int NR_VCPUS)
tools/testing/selftests/kvm/x86/msrs_test.c
414
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/x86/nested_close_kvm_test.c
71
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_dirty_log_test.c
203
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_emulation_test.c
126
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
167
static void assert_ucall_vector(struct kvm_vcpu *vcpu, int vector)
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
190
static void queue_ss_exception(struct kvm_vcpu *vcpu, bool inject)
tools/testing/selftests/kvm/x86/nested_exceptions_test.c
221
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_invalid_cr3_test.c
79
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_set_state_test.c
251
static void vcpu_efer_enable_svm(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/nested_set_state_test.c
258
static void vcpu_efer_disable_svm(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/nested_set_state_test.c
273
void test_svm_nested_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/nested_set_state_test.c
28
void test_nested_state(struct kvm_vcpu *vcpu, struct kvm_nested_state *state)
tools/testing/selftests/kvm/x86/nested_set_state_test.c
33
void test_nested_state_expect_errno(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/nested_set_state_test.c
353
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_set_state_test.c
46
void test_nested_state_expect_einval(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/nested_set_state_test.c
52
void test_nested_state_expect_efault(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/nested_set_state_test.c
87
void test_vmx_nested_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/nested_tsc_adjust_test.c
129
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_tsc_scaling_test.c
153
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nested_vmsave_vmload_test.c
133
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/nx_huge_pages_test.c
108
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/platform_info_test.c
43
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_counters_test.c
112
static void run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_counters_test.c
334
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_counters_test.c
501
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_counters_test.c
568
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_counters_test.c
90
static struct kvm_vm *pmu_vm_create_with_one_vcpu(struct kvm_vcpu **vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
150
static uint64_t run_vcpu_to_sync(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
162
static void run_vcpu_and_sync_pmc_results(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
183
static bool sanity_check_pmu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
240
static void test_without_filter(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
247
static void test_with_filter(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
256
static void test_amd_deny_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
271
static void test_member_deny_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
281
static void test_member_allow_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
291
static void test_not_member_deny_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
305
static void test_not_member_allow_list(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
326
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
477
static void run_masked_events_test(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
605
static void run_masked_events_tests(struct kvm_vcpu *vcpu, uint64_t *events,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
649
static void test_masked_events(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
662
static int set_pmu_event_filter(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
670
static int set_pmu_single_event_filter(struct kvm_vcpu *vcpu, uint64_t event,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
685
static void test_filter_ioctl(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
747
static uint64_t test_with_fixed_counter_filter(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
759
static uint64_t test_set_gp_and_fixed_event_filter(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
772
static void __test_fixed_counter_bitmap(struct kvm_vcpu *vcpu, uint8_t idx,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
819
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
838
struct kvm_vcpu *vcpu, *vcpu2 = NULL;
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
289
static void handle_exit_hypercall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
316
struct kvm_vcpu *vcpu = __vcpu;
tools/testing/selftests/kvm/x86/private_mem_conversions_test.c
380
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
tools/testing/selftests/kvm/x86/private_mem_kvm_exits_test.c
30
static uint32_t run_vcpu_get_exit_reason(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/private_mem_kvm_exits_test.c
50
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/private_mem_kvm_exits_test.c
88
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/recalc_apic_map_test.c
24
struct kvm_vcpu *vcpu = arg;
tools/testing/selftests/kvm/x86/recalc_apic_map_test.c
37
struct kvm_vcpu *vcpus[KVM_MAX_VCPUS];
tools/testing/selftests/kvm/x86/recalc_apic_map_test.c
38
struct kvm_vcpu *vcpuN;
tools/testing/selftests/kvm/x86/set_boot_cpu_id.c
109
struct kvm_vcpu *vcpus[2];
tools/testing/selftests/kvm/x86/set_boot_cpu_id.c
122
struct kvm_vcpu *vcpus[2];
tools/testing/selftests/kvm/x86/set_boot_cpu_id.c
50
static void test_set_bsp_busy(struct kvm_vcpu *vcpu, const char *msg)
tools/testing/selftests/kvm/x86/set_boot_cpu_id.c
58
static void run_vcpu(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/set_boot_cpu_id.c
90
struct kvm_vcpu *vcpus[])
tools/testing/selftests/kvm/x86/set_sregs_test.c
124
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/set_sregs_test.c
77
static void test_cr_bits(struct kvm_vcpu *vcpu, uint64_t cr4)
tools/testing/selftests/kvm/x86/sev_smoke_test.c
109
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/sev_smoke_test.c
155
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/sev_smoke_test.c
206
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/smaller_maxphyaddr_emulation_test.c
47
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/smm_test.c
118
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/state_test.c
231
struct kvm_vcpu *vcpu, *vcpuN;
tools/testing/selftests/kvm/x86/svm_int_ctl_test.c
84
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/svm_nested_shutdown_test.c
44
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/svm_nested_soft_inject_test.c
145
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/svm_vmcall_test.c
38
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/sync_regs_test.c
156
static void race_sync_regs(struct kvm_vcpu *vcpu, void *racer)
tools/testing/selftests/kvm/x86/triple_fault_event_test.c
61
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/tsc_msrs_test.c
67
static void run_vcpu(struct kvm_vcpu *vcpu, int stage)
tools/testing/selftests/kvm/x86/tsc_msrs_test.c
96
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/tsc_scaling_sync.c
48
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/ucna_injection_test.c
135
static void run_vcpu_expect_gp(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/ucna_injection_test.c
148
static void inject_ucna(struct kvm_vcpu *vcpu, uint64_t addr) {
tools/testing/selftests/kvm/x86/ucna_injection_test.c
215
static void test_ucna_injection(struct kvm_vcpu *vcpu, struct thread_params *params)
tools/testing/selftests/kvm/x86/ucna_injection_test.c
243
static void setup_mce_cap(struct kvm_vcpu *vcpu, bool enable_cmci_p)
tools/testing/selftests/kvm/x86/ucna_injection_test.c
253
static struct kvm_vcpu *create_vcpu_with_mce_cap(struct kvm_vm *vm, uint32_t vcpuid,
tools/testing/selftests/kvm/x86/ucna_injection_test.c
256
struct kvm_vcpu *vcpu = vm_vcpu_add(vm, vcpuid, guest_code);
tools/testing/selftests/kvm/x86/ucna_injection_test.c
265
struct kvm_vcpu *ucna_vcpu;
tools/testing/selftests/kvm/x86/ucna_injection_test.c
266
struct kvm_vcpu *cmcidis_vcpu;
tools/testing/selftests/kvm/x86/ucna_injection_test.c
267
struct kvm_vcpu *cmci_vcpu;
tools/testing/selftests/kvm/x86/ucna_injection_test.c
62
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/userspace_io_test.c
55
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
384
static void check_for_guest_assert(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
394
static void process_rdmsr(struct kvm_vcpu *vcpu, uint32_t msr_index)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
426
static void process_wrmsr(struct kvm_vcpu *vcpu, uint32_t msr_index)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
454
static void process_ucall_done(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
467
static uint64_t process_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
491
static void run_guest_then_process_rdmsr(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
498
static void run_guest_then_process_wrmsr(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
505
static uint64_t run_guest_then_process_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
511
static void run_guest_then_process_ucall_done(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/userspace_msr_exit_test.c
575
static int handle_ucall(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_apic_access_test.c
80
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_apicv_updates_test.c
115
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
109
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
23
static void __run_vcpu_with_invalid_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
35
static void run_vcpu_with_invalid_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
56
static void set_or_clear_invalid_guest_state(struct kvm_vcpu *vcpu, bool set)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
66
static void set_invalid_guest_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
71
static void clear_invalid_guest_state(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
76
static struct kvm_vcpu *get_set_sigalrm_vcpu(struct kvm_vcpu *__vcpu)
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
78
static struct kvm_vcpu *vcpu = NULL;
tools/testing/selftests/kvm/x86/vmx_exception_with_invalid_guest_state.c
87
struct kvm_vcpu *vcpu = get_set_sigalrm_vcpu(NULL);
tools/testing/selftests/kvm/x86/vmx_invalid_nested_guest_state.c
57
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
118
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
15
static void vmx_fixed1_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index,
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
29
static void vmx_fixed0_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index,
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
43
static void vmx_fixed0and1_msr_test(struct kvm_vcpu *vcpu, uint32_t msr_index)
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
49
static void vmx_save_restore_msrs_test(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
70
static void __ia32_feature_control_msr_test(struct kvm_vcpu *vcpu,
tools/testing/selftests/kvm/x86/vmx_msrs_test.c
91
static void ia32_feature_control_msr_test(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/vmx_nested_la57_state_test.c
78
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/vmx_preemption_timer_test.c
159
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xapic_ipi_test.c
196
struct kvm_vcpu *vcpu = params->vcpu;
tools/testing/selftests/kvm/x86/xapic_ipi_test.c
231
static void cancel_join_vcpu_thread(pthread_t thread, struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/xapic_ipi_test.c
77
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xapic_state_test.c
111
struct kvm_vcpu *vcpu = x->vcpu;
tools/testing/selftests/kvm/x86/xapic_state_test.c
14
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xapic_state_test.c
145
static void __test_apic_id(struct kvm_vcpu *vcpu, uint64_t apic_base)
tools/testing/selftests/kvm/x86/xapic_state_test.c
174
struct kvm_vcpu *vcpus[NR_VCPUS];
tools/testing/selftests/kvm/x86/xapic_state_test.c
201
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xapic_state_test.c
61
struct kvm_vcpu *vcpu = x->vcpu;
tools/testing/selftests/kvm/x86/xapic_tpr_test.c
177
static void test_tpr_check_tpr_zero(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/xapic_tpr_test.c
186
static void test_tpr_check_tpr_cr8_equal(struct kvm_vcpu *vcpu)
tools/testing/selftests/kvm/x86/xapic_tpr_test.c
197
static void test_tpr_set_tpr_for_irq(struct kvm_vcpu *vcpu, bool mask)
tools/testing/selftests/kvm/x86/xapic_tpr_test.c
214
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xcr0_cpuid_test.c
107
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
383
static struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xen_vmcall_test.c
85
struct kvm_vcpu *vcpu;
tools/testing/selftests/kvm/x86/xss_msr_test.c
19
struct kvm_vcpu *vcpu;
virt/kvm/async_pf.c
120
void kvm_clear_async_pf_completion_queue(struct kvm_vcpu *vcpu)
virt/kvm/async_pf.c
153
void kvm_check_async_pf_completion(struct kvm_vcpu *vcpu)
virt/kvm/async_pf.c
179
bool kvm_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
virt/kvm/async_pf.c
216
int kvm_async_pf_wakeup_all(struct kvm_vcpu *vcpu)
virt/kvm/async_pf.c
38
void kvm_async_pf_vcpu_init(struct kvm_vcpu *vcpu)
virt/kvm/async_pf.c
49
struct kvm_vcpu *vcpu = apf->vcpu;
virt/kvm/async_pf.h
17
void kvm_async_pf_vcpu_init(struct kvm_vcpu *vcpu);
virt/kvm/coalesced_mmio.c
43
static int coalesced_mmio_write(struct kvm_vcpu *vcpu,
virt/kvm/dirty_ring.c
218
void kvm_dirty_ring_push(struct kvm_vcpu *vcpu, u32 slot, u64 offset)
virt/kvm/dirty_ring.c
243
bool kvm_dirty_ring_check_request(struct kvm_vcpu *vcpu)
virt/kvm/eventfd.c
807
ioeventfd_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this, gpa_t addr,
virt/kvm/kvm_main.c
117
static DEFINE_PER_CPU(struct kvm_vcpu *, kvm_running_vcpu);
virt/kvm/kvm_main.c
1367
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
1389
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
1414
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
165
void vcpu_load(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
176
void vcpu_put(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
187
static bool kvm_request_needs_ipi(struct kvm_vcpu *vcpu, unsigned req)
virt/kvm/kvm_main.c
217
static void kvm_make_vcpu_request(struct kvm_vcpu *vcpu, unsigned int req,
virt/kvm/kvm_main.c
248
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
2641
struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2683
bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2691
unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
273
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
2754
unsigned long kvm_vcpu_gfn_to_hva(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
2786
unsigned long kvm_vcpu_gfn_to_hva_prot(struct kvm_vcpu *vcpu, gfn_t gfn, bool *writable)
virt/kvm/kvm_main.c
3117
int __kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map,
virt/kvm/kvm_main.c
3151
void kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map)
virt/kvm/kvm_main.c
3215
int kvm_vcpu_read_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn, void *data,
virt/kvm/kvm_main.c
3244
int kvm_vcpu_read_guest(struct kvm_vcpu *vcpu, gpa_t gpa, void *data, unsigned long len)
virt/kvm/kvm_main.c
3284
int kvm_vcpu_read_guest_atomic(struct kvm_vcpu *vcpu, gpa_t gpa,
virt/kvm/kvm_main.c
3325
int kvm_vcpu_write_guest_page(struct kvm_vcpu *vcpu, gfn_t gfn,
virt/kvm/kvm_main.c
3355
int kvm_vcpu_write_guest(struct kvm_vcpu *vcpu, gpa_t gpa, const void *data,
virt/kvm/kvm_main.c
3524
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
virt/kvm/kvm_main.c
3554
void kvm_vcpu_mark_page_dirty(struct kvm_vcpu *vcpu, gfn_t gfn)
virt/kvm/kvm_main.c
3563
void kvm_sigset_activate(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3577
void kvm_sigset_deactivate(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3586
static void grow_halt_poll_ns(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3605
static void shrink_halt_poll_ns(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3624
static int kvm_vcpu_check_block(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3649
bool kvm_vcpu_block(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3681
static inline void update_halt_poll_stats(struct kvm_vcpu *vcpu, ktime_t start,
virt/kvm/kvm_main.c
3703
static unsigned int kvm_vcpu_max_halt_poll_ns(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3727
void kvm_vcpu_halt(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3800
bool kvm_vcpu_wake_up(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3816
void __kvm_vcpu_kick(struct kvm_vcpu *vcpu, bool wait)
virt/kvm/kvm_main.c
3865
int kvm_vcpu_yield_to(struct kvm_vcpu *target)
virt/kvm/kvm_main.c
3909
static bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3931
bool __weak kvm_arch_dy_runnable(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3936
static bool vcpu_dy_runnable(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3956
bool __weak kvm_arch_vcpu_preempted_in_kernel(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3961
bool __weak kvm_arch_dy_has_pending_interrupt(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
3966
void kvm_vcpu_on_spin(struct kvm_vcpu *me, bool yield_to_kernel_mode)
virt/kvm/kvm_main.c
3970
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
4055
struct kvm_vcpu *vcpu = vmf->vma->vm_file->private_data;
virt/kvm/kvm_main.c
4085
struct kvm_vcpu *vcpu = file->private_data;
virt/kvm/kvm_main.c
4099
struct kvm_vcpu *vcpu = filp->private_data;
virt/kvm/kvm_main.c
4116
static int create_vcpu_fd(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
4127
struct kvm_vcpu *vcpu = data;
virt/kvm/kvm_main.c
4137
static void kvm_create_vcpu_debugfs(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
4161
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
4278
static int kvm_vcpu_ioctl_set_sigmask(struct kvm_vcpu *vcpu, sigset_t *sigset)
virt/kvm/kvm_main.c
4292
struct kvm_vcpu *vcpu = file->private_data;
virt/kvm/kvm_main.c
4301
struct kvm_vcpu *vcpu = file->private_data;
virt/kvm/kvm_main.c
4314
static int kvm_vcpu_ioctl_get_stats_fd(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
4340
static int kvm_vcpu_pre_fault_memory(struct kvm_vcpu *vcpu,
virt/kvm/kvm_main.c
4385
static int kvm_wait_for_vcpu_online(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
4415
struct kvm_vcpu *vcpu = filp->private_data;
virt/kvm/kvm_main.c
442
static void kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id)
virt/kvm/kvm_main.c
4666
struct kvm_vcpu *vcpu = filp->private_data;
virt/kvm/kvm_main.c
467
static void kvm_vcpu_destroy(struct kvm_vcpu *vcpu)
virt/kvm/kvm_main.c
486
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
4988
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
5849
static int __kvm_io_bus_write(struct kvm_vcpu *vcpu, struct kvm_io_bus *bus,
virt/kvm/kvm_main.c
5881
int kvm_io_bus_write(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr,
virt/kvm/kvm_main.c
5901
int kvm_io_bus_write_cookie(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx,
virt/kvm/kvm_main.c
5930
static int __kvm_io_bus_read(struct kvm_vcpu *vcpu, struct kvm_io_bus *bus,
virt/kvm/kvm_main.c
5950
int kvm_io_bus_read(struct kvm_vcpu *vcpu, enum kvm_bus bus_idx, gpa_t addr,
virt/kvm/kvm_main.c
6145
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
6158
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
6377
struct kvm_vcpu *preempt_notifier_to_vcpu(struct preempt_notifier *pn)
virt/kvm/kvm_main.c
6379
return container_of(pn, struct kvm_vcpu, preempt_notifier);
virt/kvm/kvm_main.c
6384
struct kvm_vcpu *vcpu = preempt_notifier_to_vcpu(pn);
virt/kvm/kvm_main.c
6398
struct kvm_vcpu *vcpu = preempt_notifier_to_vcpu(pn);
virt/kvm/kvm_main.c
6419
struct kvm_vcpu *kvm_get_running_vcpu(void)
virt/kvm/kvm_main.c
6421
struct kvm_vcpu *vcpu;
virt/kvm/kvm_main.c
6434
struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void)
virt/kvm/kvm_main.c
6442
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
virt/kvm/kvm_main.c
6457
struct kvm_vcpu *vcpu = kvm_get_running_vcpu();
virt/kvm/kvm_main.c
6494
vcpu_align = __alignof__(struct kvm_vcpu);
virt/kvm/kvm_main.c
6498
offsetof(struct kvm_vcpu, arch),
virt/kvm/kvm_main.c
6499
offsetofend(struct kvm_vcpu, stats_id)
virt/kvm/kvm_main.c
6500
- offsetof(struct kvm_vcpu, arch),