kern_hyp_va
#define vcpu_sve_pffr(vcpu) (kern_hyp_va((vcpu)->arch.sve_state) + \
#define vcpu_has_sve(vcpu) kvm_has_sve(kern_hyp_va((vcpu)->kvm))
base = kern_hyp_va(kvm_ksym_ref(__kvm_hyp_vector));
base = kern_hyp_va(kvm_ksym_ref(__bp_harden_hyp_vecs));
num_possible_cpus(), kern_hyp_va(per_cpu_base),
kern_hyp_va(sve_state);
if (kvm_has_mte(kern_hyp_va(vcpu->kvm)))
else if (kvm_has_ras(kern_hyp_va(vcpu->kvm)))
if (kvm_has_fpmr(kern_hyp_va(vcpu->kvm)))
if (kvm_has_fpmr(kern_hyp_va(vcpu->kvm)))
offset += *kern_hyp_va(ctxt->offset.vm_offset);
offset += *kern_hyp_va(ctxt->offset.vcpu_offset);
return kvm_has_s1poe(kern_hyp_va(vcpu->kvm));
return kvm_has_ras(kern_hyp_va(vcpu->kvm));
return kvm_has_sctlr2(kern_hyp_va(vcpu->kvm));
struct kvm *kvm = kern_hyp_va(ctxt_to_vcpu(ctxt)->kvm);
return kvm_has_mte(kern_hyp_va(vcpu->kvm));
return kvm_has_s1pie(kern_hyp_va(vcpu->kvm));
return kvm_has_tcr2(kern_hyp_va(vcpu->kvm));
struct kvm_vcpu *vcpu = kern_hyp_va(host_vcpu);
__kvm_adjust_pc(kern_hyp_va(vcpu));
__kvm_tlb_flush_vmid_ipa(kern_hyp_va(mmu), ipa, level);
__kvm_tlb_flush_vmid_ipa_nsh(kern_hyp_va(mmu), ipa, level);
__kvm_tlb_flush_vmid_range(kern_hyp_va(mmu), start, pages);
__kvm_tlb_flush_vmid(kern_hyp_va(mmu));
__kvm_flush_cpu_context(kern_hyp_va(mmu));
__vgic_v3_save_aprs(kern_hyp_va(cpu_if));
__vgic_v3_restore_vmcr_aprs(kern_hyp_va(cpu_if));
host_kvm = kern_hyp_va(host_kvm);
host_vcpu = kern_hyp_va(host_vcpu);
has_fpmr = kvm_has_fpmr(kern_hyp_va(vcpu->kvm));
this_cpu_base = kern_hyp_va(cpu_base_array[cpu]);
sve_state = kern_hyp_va(READ_ONCE(host_vcpu->arch.sve_state));
void *va = (void *)kern_hyp_va(host_va);
start = (void *)kern_hyp_va(per_cpu_base[i]);
start = kern_hyp_va(sve_state);
mmu = kern_hyp_va(vcpu->arch.hw_mmu);
__load_stage2(mmu, kern_hyp_va(mmu->arch));
!kern_hyp_va(vcpu->kvm)->arch.timer_data.poffset)
__load_stage2(mmu, kern_hyp_va(mmu->arch));
__load_stage2(mmu, kern_hyp_va(mmu->arch));
struct kvm *kvm = kern_hyp_va(vcpu->kvm);
if (kern_hyp_va(vcpu->kvm)->arch.vgic.vgic_model != KVM_DEV_TYPE_ARM_VGIC_V3)
kern_hyp_va(PAGE_OFFSET),
kern_hyp_va((unsigned long)high_memory - 1));
if (hyp_idmap_start >= kern_hyp_va(PAGE_OFFSET) &&
hyp_idmap_start < kern_hyp_va((unsigned long)high_memory - 1) &&
unsigned long start = kern_hyp_va((unsigned long)from);
unsigned long end = kern_hyp_va((unsigned long)to);