kvm_read_guest
int ret = kvm_read_guest(kvm, gpa, data, len);
r = kvm_read_guest(vcpu->kvm, pa, &val, sizeof(val));
r = kvm_read_guest(vcpu->kvm, pa, &val, sizeof(val));
ret = kvm_read_guest(kvm, addr, &rpte, sizeof(rpte));
ret = kvm_read_guest(kvm, ptbl, &entry, sizeof(entry));
ret = kvm_read_guest(kvm, ptbl_addr,
rc = kvm_read_guest(vcpu->kvm, args_phys, &args, sizeof(args));
rc = kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size);
return kvm_read_guest(kvm, gpa, val, sizeof(*val));
return kvm_read_guest(vcpu->kvm, gpa, data, len);
return kvm_read_guest(vcpu->kvm, gpa, data, len);
if (unlikely(kvm_read_guest(kvm, gfn_to_gpa(gfn),
if (kvm_read_guest(kvm, gfn_to_gpa(gfn), &hv->tsc_ref, sizeof(hv->tsc_ref)))
return kvm_read_guest(kvm, hc->ingpa + hc->data_offset, data,
if (unlikely(kvm_read_guest(kvm, hc->ingpa,
if (unlikely(kvm_read_guest(kvm, hc->ingpa, &flush_ex,
if (unlikely(kvm_read_guest(kvm, hc->ingpa, &send_ipi,
if (unlikely(kvm_read_guest(kvm, hc->ingpa, &send_ipi_ex,
kvm_read_guest(vcpu->kvm, to_hv_vcpu(vcpu)->nested.pa_page_gpa,
if (kvm_read_guest(svm->vcpu.kvm, scratch_gpa_beg, scratch_va, len)) {
if (kvm_read_guest(kvm, req_gpa, sev->guest_req_buf, PAGE_SIZE)) {
if (kvm_read_guest(kvm, req_gpa + offsetof(struct snp_guest_msg_hdr, msg_type),
if (kvm_read_guest(vcpu->kvm, vmptr, &revision, sizeof(revision)) ||
r = kvm_read_guest(kvm, wall_clock, &version, sizeof(version));
int kvm_read_guest(struct kvm *kvm, gpa_t gpa, void *data, unsigned long len);
EXPORT_SYMBOL_FOR_KVM_INTERNAL(kvm_read_guest);
return kvm_read_guest(kvm, gpa, data, len);