Symbol: kvm_cpu_context
arch/arm64/include/asm/kvm_asm.h
292
struct kvm_cpu_context;
arch/arm64/include/asm/kvm_asm.h
293
void handle_trap(struct kvm_cpu_context *host_ctxt);
arch/arm64/include/asm/kvm_emulate.h
180
static inline bool vcpu_is_el2_ctxt(const struct kvm_cpu_context *ctxt)
arch/arm64/include/asm/kvm_host.h
1148
static inline u64 *___ctxt_sys_reg(const struct kvm_cpu_context *ctxt, int r)
arch/arm64/include/asm/kvm_host.h
1171
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1181
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1192
const struct kvm_cpu_context *ctxt = &(v)->arch.ctxt; \
arch/arm64/include/asm/kvm_host.h
1396
static inline void kvm_init_host_cpu_context(struct kvm_cpu_context *cpu_ctxt)
arch/arm64/include/asm/kvm_host.h
744
struct kvm_cpu_context host_ctxt;
arch/arm64/include/asm/kvm_host.h
825
struct kvm_cpu_context ctxt;
arch/arm64/include/asm/kvm_hyp.h
101
void sysreg_save_host_state_vhe(struct kvm_cpu_context *ctxt);
arch/arm64/include/asm/kvm_hyp.h
102
void sysreg_restore_host_state_vhe(struct kvm_cpu_context *ctxt);
arch/arm64/include/asm/kvm_hyp.h
103
void sysreg_save_guest_state_vhe(struct kvm_cpu_context *ctxt);
arch/arm64/include/asm/kvm_hyp.h
104
void sysreg_restore_guest_state_vhe(struct kvm_cpu_context *ctxt);
arch/arm64/include/asm/kvm_hyp.h
122
bool kvm_host_psci_handler(struct kvm_cpu_context *host_ctxt, u32 func_id);
arch/arm64/include/asm/kvm_hyp.h
125
void __noreturn __hyp_do_panic(struct kvm_cpu_context *host_ctxt, u64 spsr,
arch/arm64/include/asm/kvm_hyp.h
134
void __noreturn __host_enter(struct kvm_cpu_context *host_ctxt);
arch/arm64/include/asm/kvm_hyp.h
15
DECLARE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/include/asm/kvm_hyp.h
96
void __sysreg_save_state_nvhe(struct kvm_cpu_context *ctxt);
arch/arm64/include/asm/kvm_hyp.h
97
void __sysreg_restore_state_nvhe(struct kvm_cpu_context *ctxt);
arch/arm64/kernel/asm-offsets.c
108
DEFINE(CPU_USER_PT_REGS, offsetof(struct kvm_cpu_context, regs));
arch/arm64/kernel/asm-offsets.c
109
DEFINE(CPU_ELR_EL2, offsetof(struct kvm_cpu_context, sys_regs[ELR_EL2]));
arch/arm64/kernel/asm-offsets.c
110
DEFINE(CPU_RGSR_EL1, offsetof(struct kvm_cpu_context, sys_regs[RGSR_EL1]));
arch/arm64/kernel/asm-offsets.c
111
DEFINE(CPU_GCR_EL1, offsetof(struct kvm_cpu_context, sys_regs[GCR_EL1]));
arch/arm64/kernel/asm-offsets.c
112
DEFINE(CPU_APIAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APIAKEYLO_EL1]));
arch/arm64/kernel/asm-offsets.c
113
DEFINE(CPU_APIBKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APIBKEYLO_EL1]));
arch/arm64/kernel/asm-offsets.c
114
DEFINE(CPU_APDAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APDAKEYLO_EL1]));
arch/arm64/kernel/asm-offsets.c
115
DEFINE(CPU_APDBKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APDBKEYLO_EL1]));
arch/arm64/kernel/asm-offsets.c
116
DEFINE(CPU_APGAKEYLO_EL1, offsetof(struct kvm_cpu_context, sys_regs[APGAKEYLO_EL1]));
arch/arm64/kernel/asm-offsets.c
117
DEFINE(HOST_CONTEXT_VCPU, offsetof(struct kvm_cpu_context, __hyp_running_vcpu));
arch/arm64/kvm/arm.c
112
DECLARE_KVM_NVHE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/kvm/arm.c
2605
struct kvm_cpu_context *hyp_ctxt;
arch/arm64/kvm/arm.c
603
struct kvm_cpu_context *ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
107
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
121
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
136
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
137
struct kvm_cpu_context *guest_ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
155
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/include/hyp/debug-sr.h
156
struct kvm_cpu_context *guest_ctxt;
arch/arm64/kvm/hyp/include/hyp/switch.h
212
struct kvm_cpu_context *hctxt = host_data_ptr(host_ctxt);
arch/arm64/kvm/hyp/include/hyp/switch.h
244
struct kvm_cpu_context *hctxt = host_data_ptr(host_ctxt);
arch/arm64/kvm/hyp/include/hyp/switch.h
299
struct kvm_cpu_context *hctxt = host_data_ptr(host_ctxt);
arch/arm64/kvm/hyp/include/hyp/switch.h
336
struct kvm_cpu_context *hctxt = host_data_ptr(host_ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
101
static inline bool ctxt_has_s1poe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
112
static inline bool ctxt_has_ras(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
123
static inline bool ctxt_has_sctlr2(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
134
static inline void __sysreg_save_el1_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
177
static inline void __sysreg_save_el2_return_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
19
static inline bool ctxt_has_s1poe(struct kvm_cpu_context *ctxt);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
196
static inline void __sysreg_restore_common_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
205
static inline void __sysreg_restore_user_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
21
static inline struct kvm_vcpu *ctxt_to_vcpu(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
211
static inline void __sysreg_restore_el1_state(struct kvm_cpu_context *ctxt,
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
291
static inline u64 to_hw_pstate(const struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
307
static inline void __sysreg_restore_el2_return_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
31
static inline bool ctxt_is_guest(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
36
static inline u64 *ctxt_mdscr_el1(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
46
static inline u64 ctxt_midr_el1(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
57
static inline void __sysreg_save_common_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
66
static inline void __sysreg_save_user_state(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
72
static inline bool ctxt_has_mte(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
79
static inline bool ctxt_has_s1pie(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
90
static inline bool ctxt_has_tcrx(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/include/nvhe/ffa.h
15
bool kvm_host_ffa_handler(struct kvm_cpu_context *host_ctxt, u32 func_id);
arch/arm64/kvm/hyp/include/nvhe/mem_protect.h
55
void handle_host_mem_abort(struct kvm_cpu_context *host_ctxt);
arch/arm64/kvm/hyp/nvhe/ffa.c
234
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
313
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
414
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
473
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
572
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
698
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
768
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
812
struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/ffa.c
865
bool kvm_host_ffa_handler(struct kvm_cpu_context *host_ctxt, u32 func_id)
arch/arm64/kvm/hyp/nvhe/ffa.c
97
static void ffa_set_retval(struct kvm_cpu_context *ctxt,
arch/arm64/kvm/hyp/nvhe/hyp-main.c
165
static void handle___pkvm_vcpu_load(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
189
static void handle___pkvm_vcpu_put(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
201
static void handle___kvm_vcpu_run(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
25
void __kvm_hyp_host_forward_smc(struct kvm_cpu_context *host_ctxt);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
251
static void handle___pkvm_host_share_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
276
static void handle___pkvm_host_unshare_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
297
static void handle___pkvm_host_relax_perms_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
316
static void handle___pkvm_host_wrprotect_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
337
static void handle___pkvm_host_test_clear_young_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
359
static void handle___pkvm_host_mkyoung_guest(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
377
static void handle___kvm_adjust_pc(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
384
static void handle___kvm_flush_vm_context(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
389
static void handle___kvm_tlb_flush_vmid_ipa(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
398
static void handle___kvm_tlb_flush_vmid_ipa_nsh(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
408
handle___kvm_tlb_flush_vmid_range(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
417
static void handle___kvm_tlb_flush_vmid(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
424
static void handle___pkvm_tlb_flush_vmid(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
440
static void handle___kvm_flush_cpu_context(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
447
static void handle___kvm_timer_set_cntvoff(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
452
static void handle___kvm_enable_ssbs(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
461
static void handle___vgic_v3_get_gic_config(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
466
static void handle___vgic_v3_init_lrs(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
471
static void handle___vgic_v3_save_aprs(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
478
static void handle___vgic_v3_restore_vmcr_aprs(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
485
static void handle___pkvm_init(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
502
static void handle___pkvm_cpu_set_vector(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
509
static void handle___pkvm_host_share_hyp(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
516
static void handle___pkvm_host_unshare_hyp(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
523
static void handle___pkvm_create_private_mapping(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
548
static void handle___pkvm_prot_finalize(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
553
static void handle___pkvm_reserve_vm(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
558
static void handle___pkvm_unreserve_vm(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
565
static void handle___pkvm_init_vm(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
575
static void handle___pkvm_init_vcpu(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
585
static void handle___pkvm_teardown_vm(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
592
typedef void (*hcall_t)(struct kvm_cpu_context *);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
635
static void handle_host_hcall(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
671
static void default_host_smc_handler(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
676
static void handle_host_smc(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/hyp-main.c
756
void handle_trap(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/mem_protect.c
608
void handle_host_mem_abort(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
107
static int psci_cpu_on(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
151
static int psci_cpu_suspend(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
179
static int psci_system_suspend(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
20
void __noreturn __host_enter(struct kvm_cpu_context *host_ctxt);
arch/arm64/kvm/hyp/nvhe/psci-relay.c
206
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/nvhe/psci-relay.c
227
static unsigned long psci_0_1_handler(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
239
static unsigned long psci_0_2_handler(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
265
static unsigned long psci_1_0_handler(u64 func_id, struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
281
bool kvm_host_psci_handler(struct kvm_cpu_context *host_ctxt, u32 func_id)
arch/arm64/kvm/hyp/nvhe/psci-relay.c
72
static unsigned long psci_forward(struct kvm_cpu_context *host_ctxt)
arch/arm64/kvm/hyp/nvhe/setup.c
289
struct kvm_cpu_context *host_ctxt = host_data_ptr(host_ctxt);
arch/arm64/kvm/hyp/nvhe/switch.c
244
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/nvhe/switch.c
245
struct kvm_cpu_context *guest_ctxt;
arch/arm64/kvm/hyp/nvhe/switch.c
33
DEFINE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/kvm/hyp/nvhe/switch.c
361
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/nvhe/switch.c
63
struct kvm_cpu_context *ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/nvhe/sysreg-sr.c
21
void __sysreg_save_state_nvhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/sysreg-sr.c
29
void __sysreg_restore_state_nvhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/nvhe/tlb.c
121
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/nvhe/tlb.c
24
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/vhe/switch.c
33
DEFINE_PER_CPU(struct kvm_cpu_context, kvm_hyp_ctxt);
arch/arm64/kvm/hyp/vhe/switch.c
574
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/vhe/switch.c
575
struct kvm_cpu_context *guest_ctxt;
arch/arm64/kvm/hyp/vhe/switch.c
660
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
160
void sysreg_save_host_state_vhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
166
void sysreg_save_guest_state_vhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
173
void sysreg_restore_host_state_vhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
179
void sysreg_restore_guest_state_vhe(struct kvm_cpu_context *ctxt)
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
199
struct kvm_cpu_context *guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
200
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
260
struct kvm_cpu_context *guest_ctxt = &vcpu->arch.ctxt;
arch/arm64/kvm/hyp/vhe/sysreg-sr.c
261
struct kvm_cpu_context *host_ctxt;
arch/arm64/kvm/pmu.c
180
struct kvm_cpu_context *hctxt;
arch/riscv/include/asm/kvm_host.h
209
struct kvm_cpu_context host_context;
arch/riscv/include/asm/kvm_host.h
212
struct kvm_cpu_context guest_context;
arch/riscv/include/asm/kvm_vcpu_fp.h
15
struct kvm_cpu_context;
arch/riscv/include/asm/kvm_vcpu_fp.h
18
void __kvm_riscv_fp_f_save(struct kvm_cpu_context *context);
arch/riscv/include/asm/kvm_vcpu_fp.h
19
void __kvm_riscv_fp_f_restore(struct kvm_cpu_context *context);
arch/riscv/include/asm/kvm_vcpu_fp.h
20
void __kvm_riscv_fp_d_save(struct kvm_cpu_context *context);
arch/riscv/include/asm/kvm_vcpu_fp.h
21
void __kvm_riscv_fp_d_restore(struct kvm_cpu_context *context);
arch/riscv/include/asm/kvm_vcpu_fp.h
24
void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_fp.h
26
void kvm_riscv_vcpu_guest_fp_restore(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_fp.h
28
void kvm_riscv_vcpu_host_fp_save(struct kvm_cpu_context *cntx);
arch/riscv/include/asm/kvm_vcpu_fp.h
29
void kvm_riscv_vcpu_host_fp_restore(struct kvm_cpu_context *cntx);
arch/riscv/include/asm/kvm_vcpu_fp.h
34
static inline void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_fp.h
39
struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_fp.h
43
static inline void kvm_riscv_vcpu_host_fp_save(struct kvm_cpu_context *cntx)
arch/riscv/include/asm/kvm_vcpu_fp.h
47
struct kvm_cpu_context *cntx)
arch/riscv/include/asm/kvm_vcpu_vector.h
19
static __always_inline void __kvm_riscv_vector_save(struct kvm_cpu_context *context)
arch/riscv/include/asm/kvm_vcpu_vector.h
24
static __always_inline void __kvm_riscv_vector_restore(struct kvm_cpu_context *context)
arch/riscv/include/asm/kvm_vcpu_vector.h
30
void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_vector.h
32
void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_vector.h
34
void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx);
arch/riscv/include/asm/kvm_vcpu_vector.h
35
void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx);
arch/riscv/include/asm/kvm_vcpu_vector.h
40
struct kvm_cpu_context;
arch/riscv/include/asm/kvm_vcpu_vector.h
46
static inline void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_vector.h
51
static inline void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
arch/riscv/include/asm/kvm_vcpu_vector.h
56
static inline void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx)
arch/riscv/include/asm/kvm_vcpu_vector.h
60
static inline void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx)
arch/riscv/kernel/asm-offsets.c
224
OFFSET(KVM_ARCH_FP_F_F0, kvm_cpu_context, fp.f.f[0]);
arch/riscv/kernel/asm-offsets.c
225
OFFSET(KVM_ARCH_FP_F_F1, kvm_cpu_context, fp.f.f[1]);
arch/riscv/kernel/asm-offsets.c
226
OFFSET(KVM_ARCH_FP_F_F2, kvm_cpu_context, fp.f.f[2]);
arch/riscv/kernel/asm-offsets.c
227
OFFSET(KVM_ARCH_FP_F_F3, kvm_cpu_context, fp.f.f[3]);
arch/riscv/kernel/asm-offsets.c
228
OFFSET(KVM_ARCH_FP_F_F4, kvm_cpu_context, fp.f.f[4]);
arch/riscv/kernel/asm-offsets.c
229
OFFSET(KVM_ARCH_FP_F_F5, kvm_cpu_context, fp.f.f[5]);
arch/riscv/kernel/asm-offsets.c
230
OFFSET(KVM_ARCH_FP_F_F6, kvm_cpu_context, fp.f.f[6]);
arch/riscv/kernel/asm-offsets.c
231
OFFSET(KVM_ARCH_FP_F_F7, kvm_cpu_context, fp.f.f[7]);
arch/riscv/kernel/asm-offsets.c
232
OFFSET(KVM_ARCH_FP_F_F8, kvm_cpu_context, fp.f.f[8]);
arch/riscv/kernel/asm-offsets.c
233
OFFSET(KVM_ARCH_FP_F_F9, kvm_cpu_context, fp.f.f[9]);
arch/riscv/kernel/asm-offsets.c
234
OFFSET(KVM_ARCH_FP_F_F10, kvm_cpu_context, fp.f.f[10]);
arch/riscv/kernel/asm-offsets.c
235
OFFSET(KVM_ARCH_FP_F_F11, kvm_cpu_context, fp.f.f[11]);
arch/riscv/kernel/asm-offsets.c
236
OFFSET(KVM_ARCH_FP_F_F12, kvm_cpu_context, fp.f.f[12]);
arch/riscv/kernel/asm-offsets.c
237
OFFSET(KVM_ARCH_FP_F_F13, kvm_cpu_context, fp.f.f[13]);
arch/riscv/kernel/asm-offsets.c
238
OFFSET(KVM_ARCH_FP_F_F14, kvm_cpu_context, fp.f.f[14]);
arch/riscv/kernel/asm-offsets.c
239
OFFSET(KVM_ARCH_FP_F_F15, kvm_cpu_context, fp.f.f[15]);
arch/riscv/kernel/asm-offsets.c
240
OFFSET(KVM_ARCH_FP_F_F16, kvm_cpu_context, fp.f.f[16]);
arch/riscv/kernel/asm-offsets.c
241
OFFSET(KVM_ARCH_FP_F_F17, kvm_cpu_context, fp.f.f[17]);
arch/riscv/kernel/asm-offsets.c
242
OFFSET(KVM_ARCH_FP_F_F18, kvm_cpu_context, fp.f.f[18]);
arch/riscv/kernel/asm-offsets.c
243
OFFSET(KVM_ARCH_FP_F_F19, kvm_cpu_context, fp.f.f[19]);
arch/riscv/kernel/asm-offsets.c
244
OFFSET(KVM_ARCH_FP_F_F20, kvm_cpu_context, fp.f.f[20]);
arch/riscv/kernel/asm-offsets.c
245
OFFSET(KVM_ARCH_FP_F_F21, kvm_cpu_context, fp.f.f[21]);
arch/riscv/kernel/asm-offsets.c
246
OFFSET(KVM_ARCH_FP_F_F22, kvm_cpu_context, fp.f.f[22]);
arch/riscv/kernel/asm-offsets.c
247
OFFSET(KVM_ARCH_FP_F_F23, kvm_cpu_context, fp.f.f[23]);
arch/riscv/kernel/asm-offsets.c
248
OFFSET(KVM_ARCH_FP_F_F24, kvm_cpu_context, fp.f.f[24]);
arch/riscv/kernel/asm-offsets.c
249
OFFSET(KVM_ARCH_FP_F_F25, kvm_cpu_context, fp.f.f[25]);
arch/riscv/kernel/asm-offsets.c
250
OFFSET(KVM_ARCH_FP_F_F26, kvm_cpu_context, fp.f.f[26]);
arch/riscv/kernel/asm-offsets.c
251
OFFSET(KVM_ARCH_FP_F_F27, kvm_cpu_context, fp.f.f[27]);
arch/riscv/kernel/asm-offsets.c
252
OFFSET(KVM_ARCH_FP_F_F28, kvm_cpu_context, fp.f.f[28]);
arch/riscv/kernel/asm-offsets.c
253
OFFSET(KVM_ARCH_FP_F_F29, kvm_cpu_context, fp.f.f[29]);
arch/riscv/kernel/asm-offsets.c
254
OFFSET(KVM_ARCH_FP_F_F30, kvm_cpu_context, fp.f.f[30]);
arch/riscv/kernel/asm-offsets.c
255
OFFSET(KVM_ARCH_FP_F_F31, kvm_cpu_context, fp.f.f[31]);
arch/riscv/kernel/asm-offsets.c
256
OFFSET(KVM_ARCH_FP_F_FCSR, kvm_cpu_context, fp.f.fcsr);
arch/riscv/kernel/asm-offsets.c
260
OFFSET(KVM_ARCH_FP_D_F0, kvm_cpu_context, fp.d.f[0]);
arch/riscv/kernel/asm-offsets.c
261
OFFSET(KVM_ARCH_FP_D_F1, kvm_cpu_context, fp.d.f[1]);
arch/riscv/kernel/asm-offsets.c
262
OFFSET(KVM_ARCH_FP_D_F2, kvm_cpu_context, fp.d.f[2]);
arch/riscv/kernel/asm-offsets.c
263
OFFSET(KVM_ARCH_FP_D_F3, kvm_cpu_context, fp.d.f[3]);
arch/riscv/kernel/asm-offsets.c
264
OFFSET(KVM_ARCH_FP_D_F4, kvm_cpu_context, fp.d.f[4]);
arch/riscv/kernel/asm-offsets.c
265
OFFSET(KVM_ARCH_FP_D_F5, kvm_cpu_context, fp.d.f[5]);
arch/riscv/kernel/asm-offsets.c
266
OFFSET(KVM_ARCH_FP_D_F6, kvm_cpu_context, fp.d.f[6]);
arch/riscv/kernel/asm-offsets.c
267
OFFSET(KVM_ARCH_FP_D_F7, kvm_cpu_context, fp.d.f[7]);
arch/riscv/kernel/asm-offsets.c
268
OFFSET(KVM_ARCH_FP_D_F8, kvm_cpu_context, fp.d.f[8]);
arch/riscv/kernel/asm-offsets.c
269
OFFSET(KVM_ARCH_FP_D_F9, kvm_cpu_context, fp.d.f[9]);
arch/riscv/kernel/asm-offsets.c
270
OFFSET(KVM_ARCH_FP_D_F10, kvm_cpu_context, fp.d.f[10]);
arch/riscv/kernel/asm-offsets.c
271
OFFSET(KVM_ARCH_FP_D_F11, kvm_cpu_context, fp.d.f[11]);
arch/riscv/kernel/asm-offsets.c
272
OFFSET(KVM_ARCH_FP_D_F12, kvm_cpu_context, fp.d.f[12]);
arch/riscv/kernel/asm-offsets.c
273
OFFSET(KVM_ARCH_FP_D_F13, kvm_cpu_context, fp.d.f[13]);
arch/riscv/kernel/asm-offsets.c
274
OFFSET(KVM_ARCH_FP_D_F14, kvm_cpu_context, fp.d.f[14]);
arch/riscv/kernel/asm-offsets.c
275
OFFSET(KVM_ARCH_FP_D_F15, kvm_cpu_context, fp.d.f[15]);
arch/riscv/kernel/asm-offsets.c
276
OFFSET(KVM_ARCH_FP_D_F16, kvm_cpu_context, fp.d.f[16]);
arch/riscv/kernel/asm-offsets.c
277
OFFSET(KVM_ARCH_FP_D_F17, kvm_cpu_context, fp.d.f[17]);
arch/riscv/kernel/asm-offsets.c
278
OFFSET(KVM_ARCH_FP_D_F18, kvm_cpu_context, fp.d.f[18]);
arch/riscv/kernel/asm-offsets.c
279
OFFSET(KVM_ARCH_FP_D_F19, kvm_cpu_context, fp.d.f[19]);
arch/riscv/kernel/asm-offsets.c
280
OFFSET(KVM_ARCH_FP_D_F20, kvm_cpu_context, fp.d.f[20]);
arch/riscv/kernel/asm-offsets.c
281
OFFSET(KVM_ARCH_FP_D_F21, kvm_cpu_context, fp.d.f[21]);
arch/riscv/kernel/asm-offsets.c
282
OFFSET(KVM_ARCH_FP_D_F22, kvm_cpu_context, fp.d.f[22]);
arch/riscv/kernel/asm-offsets.c
283
OFFSET(KVM_ARCH_FP_D_F23, kvm_cpu_context, fp.d.f[23]);
arch/riscv/kernel/asm-offsets.c
284
OFFSET(KVM_ARCH_FP_D_F24, kvm_cpu_context, fp.d.f[24]);
arch/riscv/kernel/asm-offsets.c
285
OFFSET(KVM_ARCH_FP_D_F25, kvm_cpu_context, fp.d.f[25]);
arch/riscv/kernel/asm-offsets.c
286
OFFSET(KVM_ARCH_FP_D_F26, kvm_cpu_context, fp.d.f[26]);
arch/riscv/kernel/asm-offsets.c
287
OFFSET(KVM_ARCH_FP_D_F27, kvm_cpu_context, fp.d.f[27]);
arch/riscv/kernel/asm-offsets.c
288
OFFSET(KVM_ARCH_FP_D_F28, kvm_cpu_context, fp.d.f[28]);
arch/riscv/kernel/asm-offsets.c
289
OFFSET(KVM_ARCH_FP_D_F29, kvm_cpu_context, fp.d.f[29]);
arch/riscv/kernel/asm-offsets.c
290
OFFSET(KVM_ARCH_FP_D_F30, kvm_cpu_context, fp.d.f[30]);
arch/riscv/kernel/asm-offsets.c
291
OFFSET(KVM_ARCH_FP_D_F31, kvm_cpu_context, fp.d.f[31]);
arch/riscv/kernel/asm-offsets.c
292
OFFSET(KVM_ARCH_FP_D_FCSR, kvm_cpu_context, fp.d.fcsr);
arch/riscv/kvm/vcpu.c
58
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu.c
788
struct kvm_cpu_context *gcntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu.c
789
struct kvm_cpu_context *hcntx = &vcpu->arch.host_context;
arch/riscv/kvm/vcpu_fp.c
131
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_fp.c
20
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_fp.c
30
static void kvm_riscv_vcpu_fp_clean(struct kvm_cpu_context *cntx)
arch/riscv/kvm/vcpu_fp.c
36
void kvm_riscv_vcpu_guest_fp_save(struct kvm_cpu_context *cntx,
arch/riscv/kvm/vcpu_fp.c
48
void kvm_riscv_vcpu_guest_fp_restore(struct kvm_cpu_context *cntx,
arch/riscv/kvm/vcpu_fp.c
60
void kvm_riscv_vcpu_host_fp_save(struct kvm_cpu_context *cntx)
arch/riscv/kvm/vcpu_fp.c
69
void kvm_riscv_vcpu_host_fp_restore(struct kvm_cpu_context *cntx)
arch/riscv/kvm/vcpu_fp.c
82
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_insn.c
333
struct kvm_cpu_context *ct;
arch/riscv/kvm/vcpu_insn.c
381
struct kvm_cpu_context *ct = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_insn.c
509
struct kvm_cpu_context *ct = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
1009
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
1038
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
1129
const struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
442
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_onereg.c
478
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
131
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
185
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
201
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi.c
591
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_base.c
19
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_fwft.c
328
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_hsm.c
18
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_hsm.c
65
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_hsm.c
84
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_pmu.c
20
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
151
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
20
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
51
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_replace.c
95
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_sta.c
126
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_sta.c
83
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_system.c
15
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_sbi_v01.c
24
struct kvm_cpu_context *cp = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
101
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
182
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
22
struct kvm_cpu_context *cntx = &vcpu->arch.guest_context;
arch/riscv/kvm/vcpu_vector.c
37
static void kvm_riscv_vcpu_vector_clean(struct kvm_cpu_context *cntx)
arch/riscv/kvm/vcpu_vector.c
43
void kvm_riscv_vcpu_guest_vector_save(struct kvm_cpu_context *cntx,
arch/riscv/kvm/vcpu_vector.c
53
void kvm_riscv_vcpu_guest_vector_restore(struct kvm_cpu_context *cntx,
arch/riscv/kvm/vcpu_vector.c
63
void kvm_riscv_vcpu_host_vector_save(struct kvm_cpu_context *cntx)
arch/riscv/kvm/vcpu_vector.c
70
void kvm_riscv_vcpu_host_vector_restore(struct kvm_cpu_context *cntx)