root/arch/loongarch/kvm/main.c
// SPDX-License-Identifier: GPL-2.0
/*
 * Copyright (C) 2020-2023 Loongson Technology Corporation Limited
 */

#include <linux/err.h>
#include <linux/module.h>
#include <linux/kvm_host.h>
#include <asm/cacheflush.h>
#include <asm/cpufeature.h>
#include <asm/kvm_csr.h>
#include <asm/kvm_eiointc.h>
#include <asm/kvm_pch_pic.h>
#include "trace.h"

unsigned long vpid_mask;
struct kvm_world_switch *kvm_loongarch_ops;
static int gcsr_flag[CSR_MAX_NUMS];
static struct kvm_context __percpu *vmcs;

int get_gcsr_flag(int csr)
{
        if (csr < CSR_MAX_NUMS)
                return gcsr_flag[csr];

        return INVALID_GCSR;
}

static inline void set_gcsr_sw_flag(int csr)
{
        if (csr < CSR_MAX_NUMS)
                gcsr_flag[csr] |= SW_GCSR;
}

static inline void set_gcsr_hw_flag(int csr)
{
        if (csr < CSR_MAX_NUMS)
                gcsr_flag[csr] |= HW_GCSR;
}

/*
 * The default value of gcsr_flag[CSR] is 0, and we use this
 * function to set the flag to 1 (SW_GCSR) or 2 (HW_GCSR) if the
 * gcsr is software or hardware. It will be used by get/set_gcsr,
 * if gcsr_flag is HW we should use gcsrrd/gcsrwr to access it,
 * else use software csr to emulate it.
 */
static void kvm_init_gcsr_flag(void)
{
        set_gcsr_hw_flag(LOONGARCH_CSR_CRMD);
        set_gcsr_hw_flag(LOONGARCH_CSR_PRMD);
        set_gcsr_hw_flag(LOONGARCH_CSR_EUEN);
        set_gcsr_hw_flag(LOONGARCH_CSR_MISC);
        set_gcsr_hw_flag(LOONGARCH_CSR_ECFG);
        set_gcsr_hw_flag(LOONGARCH_CSR_ESTAT);
        set_gcsr_hw_flag(LOONGARCH_CSR_ERA);
        set_gcsr_hw_flag(LOONGARCH_CSR_BADV);
        set_gcsr_hw_flag(LOONGARCH_CSR_BADI);
        set_gcsr_hw_flag(LOONGARCH_CSR_EENTRY);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBIDX);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBEHI);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBELO0);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBELO1);
        set_gcsr_hw_flag(LOONGARCH_CSR_ASID);
        set_gcsr_hw_flag(LOONGARCH_CSR_PGDL);
        set_gcsr_hw_flag(LOONGARCH_CSR_PGDH);
        set_gcsr_hw_flag(LOONGARCH_CSR_PGD);
        set_gcsr_hw_flag(LOONGARCH_CSR_PWCTL0);
        set_gcsr_hw_flag(LOONGARCH_CSR_PWCTL1);
        set_gcsr_hw_flag(LOONGARCH_CSR_STLBPGSIZE);
        set_gcsr_hw_flag(LOONGARCH_CSR_RVACFG);
        set_gcsr_hw_flag(LOONGARCH_CSR_CPUID);
        set_gcsr_hw_flag(LOONGARCH_CSR_PRCFG1);
        set_gcsr_hw_flag(LOONGARCH_CSR_PRCFG2);
        set_gcsr_hw_flag(LOONGARCH_CSR_PRCFG3);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS0);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS1);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS2);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS3);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS4);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS5);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS6);
        set_gcsr_hw_flag(LOONGARCH_CSR_KS7);
        set_gcsr_hw_flag(LOONGARCH_CSR_TMID);
        set_gcsr_hw_flag(LOONGARCH_CSR_TCFG);
        set_gcsr_hw_flag(LOONGARCH_CSR_TVAL);
        set_gcsr_hw_flag(LOONGARCH_CSR_TINTCLR);
        set_gcsr_hw_flag(LOONGARCH_CSR_CNTC);
        set_gcsr_hw_flag(LOONGARCH_CSR_LLBCTL);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRENTRY);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRBADV);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRERA);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRSAVE);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRELO0);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRELO1);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBREHI);
        set_gcsr_hw_flag(LOONGARCH_CSR_TLBRPRMD);
        set_gcsr_hw_flag(LOONGARCH_CSR_DMWIN0);
        set_gcsr_hw_flag(LOONGARCH_CSR_DMWIN1);
        set_gcsr_hw_flag(LOONGARCH_CSR_DMWIN2);
        set_gcsr_hw_flag(LOONGARCH_CSR_DMWIN3);

        set_gcsr_sw_flag(LOONGARCH_CSR_IMPCTL1);
        set_gcsr_sw_flag(LOONGARCH_CSR_IMPCTL2);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRCTL);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRINFO1);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRINFO2);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRENTRY);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRERA);
        set_gcsr_sw_flag(LOONGARCH_CSR_MERRSAVE);
        set_gcsr_sw_flag(LOONGARCH_CSR_CTAG);
        set_gcsr_sw_flag(LOONGARCH_CSR_DEBUG);
        set_gcsr_sw_flag(LOONGARCH_CSR_DERA);
        set_gcsr_sw_flag(LOONGARCH_CSR_DESAVE);

        set_gcsr_sw_flag(LOONGARCH_CSR_FWPC);
        set_gcsr_sw_flag(LOONGARCH_CSR_FWPS);
        set_gcsr_sw_flag(LOONGARCH_CSR_MWPC);
        set_gcsr_sw_flag(LOONGARCH_CSR_MWPS);

        set_gcsr_sw_flag(LOONGARCH_CSR_DB0ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB0MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB0CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB0ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB1ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB1MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB1CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB1ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB2ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB2MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB2CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB2ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB3ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB3MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB3CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB3ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB4ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB4MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB4CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB4ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB5ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB5MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB5CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB5ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB6ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB6MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB6CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB6ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB7ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB7MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB7CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_DB7ASID);

        set_gcsr_sw_flag(LOONGARCH_CSR_IB0ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB0MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB0CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB0ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB1ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB1MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB1CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB1ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB2ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB2MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB2CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB2ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB3ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB3MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB3CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB3ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB4ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB4MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB4CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB4ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB5ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB5MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB5CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB5ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB6ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB6MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB6CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB6ASID);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB7ADDR);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB7MASK);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB7CTRL);
        set_gcsr_sw_flag(LOONGARCH_CSR_IB7ASID);

        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCTRL0);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCNTR0);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCTRL1);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCNTR1);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCTRL2);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCNTR2);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCTRL3);
        set_gcsr_sw_flag(LOONGARCH_CSR_PERFCNTR3);

        if (cpu_has_msgint) {
                set_gcsr_hw_flag(LOONGARCH_CSR_IPR);
                set_gcsr_hw_flag(LOONGARCH_CSR_ISR0);
                set_gcsr_hw_flag(LOONGARCH_CSR_ISR1);
                set_gcsr_hw_flag(LOONGARCH_CSR_ISR2);
                set_gcsr_hw_flag(LOONGARCH_CSR_ISR3);
        }
}

static void kvm_update_vpid(struct kvm_vcpu *vcpu, int cpu)
{
        unsigned long vpid;
        struct kvm_context *context;

        context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
        vpid = context->vpid_cache + 1;
        if (!(vpid & vpid_mask)) {
                /* finish round of vpid loop */
                if (unlikely(!vpid))
                        vpid = vpid_mask + 1;

                ++vpid; /* vpid 0 reserved for root */

                /* start new vpid cycle */
                kvm_flush_tlb_all();
        }

        context->vpid_cache = vpid;
        vcpu->arch.vpid = vpid;
}

void kvm_check_vpid(struct kvm_vcpu *vcpu)
{
        int cpu;
        bool migrated;
        unsigned long ver, old, vpid;
        struct kvm_context *context;

        cpu = smp_processor_id();
        /*
         * Are we entering guest context on a different CPU to last time?
         * If so, the vCPU's guest TLB state on this CPU may be stale.
         */
        context = per_cpu_ptr(vcpu->kvm->arch.vmcs, cpu);
        migrated = (vcpu->cpu != cpu);

        /*
         * Check if our vpid is of an older version
         *
         * We also discard the stored vpid if we've executed on
         * another CPU, as the guest mappings may have changed without
         * hypervisor knowledge.
         */
        ver = vcpu->arch.vpid & ~vpid_mask;
        old = context->vpid_cache  & ~vpid_mask;
        if (migrated || (ver != old)) {
                kvm_update_vpid(vcpu, cpu);
                trace_kvm_vpid_change(vcpu, vcpu->arch.vpid);
                vcpu->cpu = cpu;
                kvm_clear_request(KVM_REQ_TLB_FLUSH_GPA, vcpu);

                /*
                 * LLBCTL is a separated guest CSR register from host, a general
                 * exception ERET instruction clears the host LLBCTL register in
                 * host mode, and clears the guest LLBCTL register in guest mode.
                 * ERET in tlb refill exception does not clear LLBCTL register.
                 *
                 * When secondary mmu mapping is changed, guest OS does not know
                 * even if the content is changed after mapping is changed.
                 *
                 * Here clear WCLLB of the guest LLBCTL register when mapping is
                 * changed. Otherwise, if mmu mapping is changed while guest is
                 * executing LL/SC pair, LL loads with the old address and set
                 * the LLBCTL flag, SC checks the LLBCTL flag and will store the
                 * new address successfully since LLBCTL_WCLLB is on, even if
                 * memory with new address is changed on other VCPUs.
                 */
                set_gcsr_llbctl(CSR_LLBCTL_WCLLB);
        }

        /* Restore GSTAT(0x50).vpid */
        vpid = (vcpu->arch.vpid & vpid_mask) << CSR_GSTAT_GID_SHIFT;
        change_csr_gstat(vpid_mask << CSR_GSTAT_GID_SHIFT, vpid);
}

void kvm_init_vmcs(struct kvm *kvm)
{
        kvm->arch.vmcs = vmcs;
}

long kvm_arch_dev_ioctl(struct file *filp,
                        unsigned int ioctl, unsigned long arg)
{
        return -ENOIOCTLCMD;
}

int kvm_arch_enable_virtualization_cpu(void)
{
        unsigned long env, gcfg = 0;

        env = read_csr_gcfg();

        /* First init gcfg, gstat, gintc, gtlbc. All guest use the same config */
        write_csr_gcfg(0);
        write_csr_gstat(0);
        write_csr_gintc(0);
        clear_csr_gtlbc(CSR_GTLBC_USETGID | CSR_GTLBC_TOTI);

        /*
         * Enable virtualization features granting guest direct control of
         * certain features:
         * GCI=2:       Trap on init or unimplemented cache instruction.
         * TORU=0:      Trap on Root Unimplement.
         * CACTRL=1:    Root control cache.
         * TOP=0:       Trap on Privilege.
         * TOE=0:       Trap on Exception.
         * TIT=0:       Trap on Timer.
         */
        if (env & CSR_GCFG_GCIP_SECURE)
                gcfg |= CSR_GCFG_GCI_SECURE;
        if (env & CSR_GCFG_MATP_ROOT)
                gcfg |= CSR_GCFG_MATC_ROOT;

        write_csr_gcfg(gcfg);

        kvm_flush_tlb_all();

        /* Enable using TGID  */
        set_csr_gtlbc(CSR_GTLBC_USETGID);
        kvm_debug("GCFG:%lx GSTAT:%lx GINTC:%lx GTLBC:%lx",
                  read_csr_gcfg(), read_csr_gstat(), read_csr_gintc(), read_csr_gtlbc());

        /*
         * HW Guest CSR registers are lost after CPU suspend and resume.
         * Clear last_vcpu so that Guest CSR registers forced to reload
         * from vCPU SW state.
         */
        this_cpu_ptr(vmcs)->last_vcpu = NULL;

        return 0;
}

void kvm_arch_disable_virtualization_cpu(void)
{
        write_csr_gcfg(0);
        write_csr_gstat(0);
        write_csr_gintc(0);
        clear_csr_gtlbc(CSR_GTLBC_USETGID | CSR_GTLBC_TOTI);

        /* Flush any remaining guest TLB entries */
        kvm_flush_tlb_all();
}

static int kvm_loongarch_env_init(void)
{
        int cpu, order, ret;
        void *addr;
        struct kvm_context *context;

        vmcs = alloc_percpu(struct kvm_context);
        if (!vmcs) {
                pr_err("kvm: failed to allocate percpu kvm_context\n");
                return -ENOMEM;
        }

        kvm_loongarch_ops = kzalloc_obj(*kvm_loongarch_ops);
        if (!kvm_loongarch_ops) {
                free_percpu(vmcs);
                vmcs = NULL;
                return -ENOMEM;
        }

        /*
         * PGD register is shared between root kernel and kvm hypervisor.
         * So world switch entry should be in DMW area rather than TLB area
         * to avoid page fault reenter.
         *
         * In future if hardware pagetable walking is supported, we won't
         * need to copy world switch code to DMW area.
         */
        order = get_order(kvm_exception_size + kvm_enter_guest_size);
        addr = (void *)__get_free_pages(GFP_KERNEL, order);
        if (!addr) {
                free_percpu(vmcs);
                vmcs = NULL;
                kfree(kvm_loongarch_ops);
                kvm_loongarch_ops = NULL;
                return -ENOMEM;
        }

        memcpy(addr, kvm_exc_entry, kvm_exception_size);
        memcpy(addr + kvm_exception_size, kvm_enter_guest, kvm_enter_guest_size);
        flush_icache_range((unsigned long)addr, (unsigned long)addr + kvm_exception_size + kvm_enter_guest_size);
        kvm_loongarch_ops->exc_entry = addr;
        kvm_loongarch_ops->enter_guest = addr + kvm_exception_size;
        kvm_loongarch_ops->page_order = order;

        vpid_mask = read_csr_gstat();
        vpid_mask = (vpid_mask & CSR_GSTAT_GIDBIT) >> CSR_GSTAT_GIDBIT_SHIFT;
        if (vpid_mask)
                vpid_mask = GENMASK(vpid_mask - 1, 0);

        for_each_possible_cpu(cpu) {
                context = per_cpu_ptr(vmcs, cpu);
                context->vpid_cache = vpid_mask + 1;
                context->last_vcpu = NULL;
        }

        kvm_init_gcsr_flag();
        kvm_register_perf_callbacks();

        /* Register LoongArch IPI interrupt controller interface. */
        ret = kvm_loongarch_register_ipi_device();
        if (ret)
                return ret;

        /* Register LoongArch EIOINTC interrupt controller interface. */
        ret = kvm_loongarch_register_eiointc_device();
        if (ret)
                return ret;

        /* Register LoongArch PCH-PIC interrupt controller interface. */
        ret = kvm_loongarch_register_pch_pic_device();

        return ret;
}

static void kvm_loongarch_env_exit(void)
{
        unsigned long addr;

        if (vmcs)
                free_percpu(vmcs);

        if (kvm_loongarch_ops) {
                if (kvm_loongarch_ops->exc_entry) {
                        addr = (unsigned long)kvm_loongarch_ops->exc_entry;
                        free_pages(addr, kvm_loongarch_ops->page_order);
                }
                kfree(kvm_loongarch_ops);
        }

        kvm_unregister_perf_callbacks();
}

static int kvm_loongarch_init(void)
{
        int r;

        if (!cpu_has_lvz) {
                kvm_info("Hardware virtualization not available\n");
                return -ENODEV;
        }
        r = kvm_loongarch_env_init();
        if (r)
                return r;

        return kvm_init(sizeof(struct kvm_vcpu), 0, THIS_MODULE);
}

static void kvm_loongarch_exit(void)
{
        kvm_exit();
        kvm_loongarch_env_exit();
}

module_init(kvm_loongarch_init);
module_exit(kvm_loongarch_exit);

#ifdef MODULE
static const struct cpu_feature kvm_feature[] = {
        { .feature = cpu_feature(LOONGARCH_LVZ) },
        {},
};
MODULE_DEVICE_TABLE(cpu, kvm_feature);
#endif