root/tools/testing/selftests/kvm/lib/x86/hyperv.c
// SPDX-License-Identifier: GPL-2.0-only
/*
 * Hyper-V specific functions.
 *
 * Copyright (C) 2021, Red Hat Inc.
 */
#include <stdint.h>
#include "processor.h"
#include "hyperv.h"

const struct kvm_cpuid2 *kvm_get_supported_hv_cpuid(void)
{
        static struct kvm_cpuid2 *cpuid;
        int kvm_fd;

        if (cpuid)
                return cpuid;

        cpuid = allocate_kvm_cpuid2(MAX_NR_CPUID_ENTRIES);
        kvm_fd = open_kvm_dev_path_or_exit();

        kvm_ioctl(kvm_fd, KVM_GET_SUPPORTED_HV_CPUID, cpuid);

        close(kvm_fd);
        return cpuid;
}

void vcpu_set_hv_cpuid(struct kvm_vcpu *vcpu)
{
        static struct kvm_cpuid2 *cpuid_full;
        const struct kvm_cpuid2 *cpuid_sys, *cpuid_hv;
        int i, nent = 0;

        if (!cpuid_full) {
                cpuid_sys = kvm_get_supported_cpuid();
                cpuid_hv = kvm_get_supported_hv_cpuid();

                cpuid_full = allocate_kvm_cpuid2(cpuid_sys->nent + cpuid_hv->nent);
                if (!cpuid_full) {
                        perror("malloc");
                        abort();
                }

                /* Need to skip KVM CPUID leaves 0x400000xx */
                for (i = 0; i < cpuid_sys->nent; i++) {
                        if (cpuid_sys->entries[i].function >= 0x40000000 &&
                            cpuid_sys->entries[i].function < 0x40000100)
                                continue;
                        cpuid_full->entries[nent] = cpuid_sys->entries[i];
                        nent++;
                }

                memcpy(&cpuid_full->entries[nent], cpuid_hv->entries,
                       cpuid_hv->nent * sizeof(struct kvm_cpuid_entry2));
                cpuid_full->nent = nent + cpuid_hv->nent;
        }

        vcpu_init_cpuid(vcpu, cpuid_full);
}

const struct kvm_cpuid2 *vcpu_get_supported_hv_cpuid(struct kvm_vcpu *vcpu)
{
        struct kvm_cpuid2 *cpuid = allocate_kvm_cpuid2(MAX_NR_CPUID_ENTRIES);

        vcpu_ioctl(vcpu, KVM_GET_SUPPORTED_HV_CPUID, cpuid);

        return cpuid;
}

bool kvm_hv_cpu_has(struct kvm_x86_cpu_feature feature)
{
        if (!kvm_has_cap(KVM_CAP_SYS_HYPERV_CPUID))
                return false;

        return kvm_cpuid_has(kvm_get_supported_hv_cpuid(), feature);
}

struct hyperv_test_pages *vcpu_alloc_hyperv_test_pages(struct kvm_vm *vm,
                                                       vm_vaddr_t *p_hv_pages_gva)
{
        vm_vaddr_t hv_pages_gva = vm_vaddr_alloc_page(vm);
        struct hyperv_test_pages *hv = addr_gva2hva(vm, hv_pages_gva);

        /* Setup of a region of guest memory for the VP Assist page. */
        hv->vp_assist = (void *)vm_vaddr_alloc_page(vm);
        hv->vp_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->vp_assist);
        hv->vp_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->vp_assist);

        /* Setup of a region of guest memory for the partition assist page. */
        hv->partition_assist = (void *)vm_vaddr_alloc_page(vm);
        hv->partition_assist_hva = addr_gva2hva(vm, (uintptr_t)hv->partition_assist);
        hv->partition_assist_gpa = addr_gva2gpa(vm, (uintptr_t)hv->partition_assist);

        /* Setup of a region of guest memory for the enlightened VMCS. */
        hv->enlightened_vmcs = (void *)vm_vaddr_alloc_page(vm);
        hv->enlightened_vmcs_hva = addr_gva2hva(vm, (uintptr_t)hv->enlightened_vmcs);
        hv->enlightened_vmcs_gpa = addr_gva2gpa(vm, (uintptr_t)hv->enlightened_vmcs);

        *p_hv_pages_gva = hv_pages_gva;
        return hv;
}

int enable_vp_assist(uint64_t vp_assist_pa, void *vp_assist)
{
        uint64_t val = (vp_assist_pa & HV_X64_MSR_VP_ASSIST_PAGE_ADDRESS_MASK) |
                HV_X64_MSR_VP_ASSIST_PAGE_ENABLE;

        wrmsr(HV_X64_MSR_VP_ASSIST_PAGE, val);

        current_vp_assist = vp_assist;

        return 0;
}