root/arch/x86/kvm/fpu.h
/* SPDX-License-Identifier: GPL-2.0 */

#ifndef __KVM_FPU_H_
#define __KVM_FPU_H_

#include <asm/fpu/api.h>

typedef u32             __attribute__((vector_size(16))) sse128_t;
#define __sse128_u      union { sse128_t vec; u64 as_u64[2]; u32 as_u32[4]; }
#define sse128_lo(x)    ({ __sse128_u t; t.vec = x; t.as_u64[0]; })
#define sse128_hi(x)    ({ __sse128_u t; t.vec = x; t.as_u64[1]; })
#define sse128_l0(x)    ({ __sse128_u t; t.vec = x; t.as_u32[0]; })
#define sse128_l1(x)    ({ __sse128_u t; t.vec = x; t.as_u32[1]; })
#define sse128_l2(x)    ({ __sse128_u t; t.vec = x; t.as_u32[2]; })
#define sse128_l3(x)    ({ __sse128_u t; t.vec = x; t.as_u32[3]; })
#define sse128(lo, hi)  ({ __sse128_u t; t.as_u64[0] = lo; t.as_u64[1] = hi; t.vec; })

typedef u32             __attribute__((vector_size(32))) avx256_t;

static inline void _kvm_read_avx_reg(int reg, avx256_t *data)
{
        switch (reg) {
        case 0:  asm("vmovdqa %%ymm0,  %0" : "=m"(*data)); break;
        case 1:  asm("vmovdqa %%ymm1,  %0" : "=m"(*data)); break;
        case 2:  asm("vmovdqa %%ymm2,  %0" : "=m"(*data)); break;
        case 3:  asm("vmovdqa %%ymm3,  %0" : "=m"(*data)); break;
        case 4:  asm("vmovdqa %%ymm4,  %0" : "=m"(*data)); break;
        case 5:  asm("vmovdqa %%ymm5,  %0" : "=m"(*data)); break;
        case 6:  asm("vmovdqa %%ymm6,  %0" : "=m"(*data)); break;
        case 7:  asm("vmovdqa %%ymm7,  %0" : "=m"(*data)); break;
#ifdef CONFIG_X86_64
        case 8:  asm("vmovdqa %%ymm8,  %0" : "=m"(*data)); break;
        case 9:  asm("vmovdqa %%ymm9,  %0" : "=m"(*data)); break;
        case 10: asm("vmovdqa %%ymm10, %0" : "=m"(*data)); break;
        case 11: asm("vmovdqa %%ymm11, %0" : "=m"(*data)); break;
        case 12: asm("vmovdqa %%ymm12, %0" : "=m"(*data)); break;
        case 13: asm("vmovdqa %%ymm13, %0" : "=m"(*data)); break;
        case 14: asm("vmovdqa %%ymm14, %0" : "=m"(*data)); break;
        case 15: asm("vmovdqa %%ymm15, %0" : "=m"(*data)); break;
#endif
        default: BUG();
        }
}

static inline void _kvm_write_avx_reg(int reg, const avx256_t *data)
{
        switch (reg) {
        case 0:  asm("vmovdqa %0, %%ymm0"  : : "m"(*data)); break;
        case 1:  asm("vmovdqa %0, %%ymm1"  : : "m"(*data)); break;
        case 2:  asm("vmovdqa %0, %%ymm2"  : : "m"(*data)); break;
        case 3:  asm("vmovdqa %0, %%ymm3"  : : "m"(*data)); break;
        case 4:  asm("vmovdqa %0, %%ymm4"  : : "m"(*data)); break;
        case 5:  asm("vmovdqa %0, %%ymm5"  : : "m"(*data)); break;
        case 6:  asm("vmovdqa %0, %%ymm6"  : : "m"(*data)); break;
        case 7:  asm("vmovdqa %0, %%ymm7"  : : "m"(*data)); break;
#ifdef CONFIG_X86_64
        case 8:  asm("vmovdqa %0, %%ymm8"  : : "m"(*data)); break;
        case 9:  asm("vmovdqa %0, %%ymm9"  : : "m"(*data)); break;
        case 10: asm("vmovdqa %0, %%ymm10" : : "m"(*data)); break;
        case 11: asm("vmovdqa %0, %%ymm11" : : "m"(*data)); break;
        case 12: asm("vmovdqa %0, %%ymm12" : : "m"(*data)); break;
        case 13: asm("vmovdqa %0, %%ymm13" : : "m"(*data)); break;
        case 14: asm("vmovdqa %0, %%ymm14" : : "m"(*data)); break;
        case 15: asm("vmovdqa %0, %%ymm15" : : "m"(*data)); break;
#endif
        default: BUG();
        }
}

static inline void _kvm_read_sse_reg(int reg, sse128_t *data)
{
        switch (reg) {
        case 0: asm("movdqa %%xmm0, %0" : "=m"(*data)); break;
        case 1: asm("movdqa %%xmm1, %0" : "=m"(*data)); break;
        case 2: asm("movdqa %%xmm2, %0" : "=m"(*data)); break;
        case 3: asm("movdqa %%xmm3, %0" : "=m"(*data)); break;
        case 4: asm("movdqa %%xmm4, %0" : "=m"(*data)); break;
        case 5: asm("movdqa %%xmm5, %0" : "=m"(*data)); break;
        case 6: asm("movdqa %%xmm6, %0" : "=m"(*data)); break;
        case 7: asm("movdqa %%xmm7, %0" : "=m"(*data)); break;
#ifdef CONFIG_X86_64
        case 8: asm("movdqa %%xmm8, %0" : "=m"(*data)); break;
        case 9: asm("movdqa %%xmm9, %0" : "=m"(*data)); break;
        case 10: asm("movdqa %%xmm10, %0" : "=m"(*data)); break;
        case 11: asm("movdqa %%xmm11, %0" : "=m"(*data)); break;
        case 12: asm("movdqa %%xmm12, %0" : "=m"(*data)); break;
        case 13: asm("movdqa %%xmm13, %0" : "=m"(*data)); break;
        case 14: asm("movdqa %%xmm14, %0" : "=m"(*data)); break;
        case 15: asm("movdqa %%xmm15, %0" : "=m"(*data)); break;
#endif
        default: BUG();
        }
}

static inline void _kvm_write_sse_reg(int reg, const sse128_t *data)
{
        switch (reg) {
        case 0: asm("movdqa %0, %%xmm0" : : "m"(*data)); break;
        case 1: asm("movdqa %0, %%xmm1" : : "m"(*data)); break;
        case 2: asm("movdqa %0, %%xmm2" : : "m"(*data)); break;
        case 3: asm("movdqa %0, %%xmm3" : : "m"(*data)); break;
        case 4: asm("movdqa %0, %%xmm4" : : "m"(*data)); break;
        case 5: asm("movdqa %0, %%xmm5" : : "m"(*data)); break;
        case 6: asm("movdqa %0, %%xmm6" : : "m"(*data)); break;
        case 7: asm("movdqa %0, %%xmm7" : : "m"(*data)); break;
#ifdef CONFIG_X86_64
        case 8: asm("movdqa %0, %%xmm8" : : "m"(*data)); break;
        case 9: asm("movdqa %0, %%xmm9" : : "m"(*data)); break;
        case 10: asm("movdqa %0, %%xmm10" : : "m"(*data)); break;
        case 11: asm("movdqa %0, %%xmm11" : : "m"(*data)); break;
        case 12: asm("movdqa %0, %%xmm12" : : "m"(*data)); break;
        case 13: asm("movdqa %0, %%xmm13" : : "m"(*data)); break;
        case 14: asm("movdqa %0, %%xmm14" : : "m"(*data)); break;
        case 15: asm("movdqa %0, %%xmm15" : : "m"(*data)); break;
#endif
        default: BUG();
        }
}

static inline void _kvm_read_mmx_reg(int reg, u64 *data)
{
        switch (reg) {
        case 0: asm("movq %%mm0, %0" : "=m"(*data)); break;
        case 1: asm("movq %%mm1, %0" : "=m"(*data)); break;
        case 2: asm("movq %%mm2, %0" : "=m"(*data)); break;
        case 3: asm("movq %%mm3, %0" : "=m"(*data)); break;
        case 4: asm("movq %%mm4, %0" : "=m"(*data)); break;
        case 5: asm("movq %%mm5, %0" : "=m"(*data)); break;
        case 6: asm("movq %%mm6, %0" : "=m"(*data)); break;
        case 7: asm("movq %%mm7, %0" : "=m"(*data)); break;
        default: BUG();
        }
}

static inline void _kvm_write_mmx_reg(int reg, const u64 *data)
{
        switch (reg) {
        case 0: asm("movq %0, %%mm0" : : "m"(*data)); break;
        case 1: asm("movq %0, %%mm1" : : "m"(*data)); break;
        case 2: asm("movq %0, %%mm2" : : "m"(*data)); break;
        case 3: asm("movq %0, %%mm3" : : "m"(*data)); break;
        case 4: asm("movq %0, %%mm4" : : "m"(*data)); break;
        case 5: asm("movq %0, %%mm5" : : "m"(*data)); break;
        case 6: asm("movq %0, %%mm6" : : "m"(*data)); break;
        case 7: asm("movq %0, %%mm7" : : "m"(*data)); break;
        default: BUG();
        }
}

static inline void kvm_fpu_get(void)
{
        fpregs_lock();

        fpregs_assert_state_consistent();
        if (test_thread_flag(TIF_NEED_FPU_LOAD))
                switch_fpu_return();
}

static inline void kvm_fpu_put(void)
{
        fpregs_unlock();
}

static inline void kvm_read_avx_reg(int reg, avx256_t *data)
{
        kvm_fpu_get();
        _kvm_read_avx_reg(reg, data);
        kvm_fpu_put();
}

static inline void kvm_write_avx_reg(int reg, const avx256_t  *data)
{
        kvm_fpu_get();
        _kvm_write_avx_reg(reg, data);
        kvm_fpu_put();
}

static inline void kvm_read_sse_reg(int reg, sse128_t *data)
{
        kvm_fpu_get();
        _kvm_read_sse_reg(reg, data);
        kvm_fpu_put();
}

static inline void kvm_write_sse_reg(int reg, const sse128_t *data)
{
        kvm_fpu_get();
        _kvm_write_sse_reg(reg, data);
        kvm_fpu_put();
}

static inline void kvm_read_mmx_reg(int reg, u64 *data)
{
        kvm_fpu_get();
        _kvm_read_mmx_reg(reg, data);
        kvm_fpu_put();
}

static inline void kvm_write_mmx_reg(int reg, const u64 *data)
{
        kvm_fpu_get();
        _kvm_write_mmx_reg(reg, data);
        kvm_fpu_put();
}

#endif