Symbol: kern_hyp_va
arch/arm64/include/asm/kvm_host.h
1079
#define vcpu_sve_pffr(vcpu) (kern_hyp_va((vcpu)->arch.sve_state) + \
arch/arm64/include/asm/kvm_host.h
1112
#define vcpu_has_sve(vcpu) kvm_has_sve(kern_hyp_va((vcpu)->kvm))
arch/arm64/kvm/arm.c
2041
base = kern_hyp_va(kvm_ksym_ref(__kvm_hyp_vector));
arch/arm64/kvm/arm.c
2044
base = kern_hyp_va(kvm_ksym_ref(__bp_harden_hyp_vecs));
arch/arm64/kvm/arm.c
2468
num_possible_cpus(), kern_hyp_va(per_cpu_base),
arch/arm64/kvm/arm.c
2598
kern_hyp_va(sve_state);
arch/arm64/kvm/hyp/exception.c
128
if (kvm_has_mte(kern_hyp_va(vcpu->kvm)))
arch/arm64/kvm/hyp/include/hyp/switch.h
378
else if (kvm_has_ras(kern_hyp_va(vcpu->kvm)))
arch/arm64/kvm/hyp/include/hyp/switch.h
537
if (kvm_has_fpmr(kern_hyp_va(vcpu->kvm)))
arch/arm64/kvm/hyp/include/hyp/switch.h
596
if (kvm_has_fpmr(kern_hyp_va(vcpu->kvm)))
arch/arm64/kvm/hyp/include/hyp/switch.h
676
offset += *kern_hyp_va(ctxt->offset.vm_offset);
arch/arm64/kvm/hyp/include/hyp/switch.h
678
offset += *kern_hyp_va(ctxt->offset.vcpu_offset);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
109
return kvm_has_s1poe(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
120
return kvm_has_ras(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
131
return kvm_has_sctlr2(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
48
struct kvm *kvm = kern_hyp_va(ctxt_to_vcpu(ctxt)->kvm);
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
76
return kvm_has_mte(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
87
return kvm_has_s1pie(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/include/hyp/sysreg-sr.h
98
return kvm_has_tcr2(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
231
struct kvm_vcpu *vcpu = kern_hyp_va(host_vcpu);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
381
__kvm_adjust_pc(kern_hyp_va(vcpu));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
395
__kvm_tlb_flush_vmid_ipa(kern_hyp_va(mmu), ipa, level);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
404
__kvm_tlb_flush_vmid_ipa_nsh(kern_hyp_va(mmu), ipa, level);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
414
__kvm_tlb_flush_vmid_range(kern_hyp_va(mmu), start, pages);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
421
__kvm_tlb_flush_vmid(kern_hyp_va(mmu));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
444
__kvm_flush_cpu_context(kern_hyp_va(mmu));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
475
__vgic_v3_save_aprs(kern_hyp_va(cpu_if));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
482
__vgic_v3_restore_vmcr_aprs(kern_hyp_va(cpu_if));
arch/arm64/kvm/hyp/nvhe/hyp-main.c
571
host_kvm = kern_hyp_va(host_kvm);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
581
host_vcpu = kern_hyp_va(host_vcpu);
arch/arm64/kvm/hyp/nvhe/hyp-main.c
83
has_fpmr = kvm_has_fpmr(kern_hyp_va(vcpu->kvm));
arch/arm64/kvm/hyp/nvhe/hyp-smp.c
37
this_cpu_base = kern_hyp_va(cpu_base_array[cpu]);
arch/arm64/kvm/hyp/nvhe/pkvm.c
455
sve_state = kern_hyp_va(READ_ONCE(host_vcpu->arch.sve_state));
arch/arm64/kvm/hyp/nvhe/pkvm.c
640
void *va = (void *)kern_hyp_va(host_va);
arch/arm64/kvm/hyp/nvhe/setup.c
147
start = (void *)kern_hyp_va(per_cpu_base[i]);
arch/arm64/kvm/hyp/nvhe/setup.c
87
start = kern_hyp_va(sve_state);
arch/arm64/kvm/hyp/nvhe/switch.c
301
mmu = kern_hyp_va(vcpu->arch.hw_mmu);
arch/arm64/kvm/hyp/nvhe/switch.c
302
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/nvhe/timer-sr.c
52
!kern_hyp_va(vcpu->kvm)->arch.timer_data.poffset)
arch/arm64/kvm/hyp/nvhe/tlb.c
113
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/nvhe/tlb.c
131
__load_stage2(mmu, kern_hyp_va(mmu->arch));
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
39
struct kvm *kvm = kern_hyp_va(vcpu->kvm);
arch/arm64/kvm/hyp/vgic-v3-sr.c
1150
if (kern_hyp_va(vcpu->kvm)->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V3)
arch/arm64/kvm/mmu.c
2319
kern_hyp_va(PAGE_OFFSET),
arch/arm64/kvm/mmu.c
2320
kern_hyp_va((unsigned long)high_memory - 1));
arch/arm64/kvm/mmu.c
2322
if (hyp_idmap_start >= kern_hyp_va(PAGE_OFFSET) &&
arch/arm64/kvm/mmu.c
2323
hyp_idmap_start < kern_hyp_va((unsigned long)high_memory - 1) &&
arch/arm64/kvm/mmu.c
595
unsigned long start = kern_hyp_va((unsigned long)from);
arch/arm64/kvm/mmu.c
596
unsigned long end = kern_hyp_va((unsigned long)to);