root/tools/testing/selftests/kvm/include/x86/svm_util.h
/* SPDX-License-Identifier: GPL-2.0-only */
/*
 * Copyright (C) 2020, Red Hat, Inc.
 */

#ifndef SELFTEST_KVM_SVM_UTILS_H
#define SELFTEST_KVM_SVM_UTILS_H

#include <asm/svm.h>

#include <stdint.h>
#include "svm.h"
#include "processor.h"

struct svm_test_data {
        /* VMCB */
        struct vmcb *vmcb; /* gva */
        void *vmcb_hva;
        uint64_t vmcb_gpa;

        /* host state-save area */
        struct vmcb_save_area *save_area; /* gva */
        void *save_area_hva;
        uint64_t save_area_gpa;

        /* MSR-Bitmap */
        void *msr; /* gva */
        void *msr_hva;
        uint64_t msr_gpa;

        /* NPT */
        uint64_t ncr3_gpa;
};

static inline void vmmcall(void)
{
        /*
         * Stuff RAX and RCX with "safe" values to make sure L0 doesn't handle
         * it as a valid hypercall (e.g. Hyper-V L2 TLB flush) as the intended
         * use of this function is to exit to L1 from L2.  Clobber all other
         * GPRs as L1 doesn't correctly preserve them during vmexits.
         */
        __asm__ __volatile__("push %%rbp; vmmcall; pop %%rbp"
                             : : "a"(0xdeadbeef), "c"(0xbeefdead)
                             : "rbx", "rdx", "rsi", "rdi", "r8", "r9",
                               "r10", "r11", "r12", "r13", "r14", "r15");
}

#define stgi()                  \
        __asm__ __volatile__(   \
                "stgi\n"        \
                )

#define clgi()                  \
        __asm__ __volatile__(   \
                "clgi\n"        \
                )

struct svm_test_data *vcpu_alloc_svm(struct kvm_vm *vm, vm_vaddr_t *p_svm_gva);
void generic_svm_setup(struct svm_test_data *svm, void *guest_rip, void *guest_rsp);
void run_guest(struct vmcb *vmcb, uint64_t vmcb_gpa);

static inline bool kvm_cpu_has_npt(void)
{
        return kvm_cpu_has(X86_FEATURE_NPT);
}
void vm_enable_npt(struct kvm_vm *vm);

int open_sev_dev_path_or_exit(void);

#endif /* SELFTEST_KVM_SVM_UTILS_H */