root/arch/mips/include/asm/msa.h
/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * Copyright (C) 2013 Imagination Technologies
 * Author: Paul Burton <paul.burton@mips.com>
 */
#ifndef _ASM_MSA_H
#define _ASM_MSA_H

#include <asm/mipsregs.h>

#ifndef __ASSEMBLER__

#include <asm/inst.h>

extern void _save_msa(struct task_struct *);
extern void _restore_msa(struct task_struct *);
extern void _init_msa_upper(void);

extern void read_msa_wr_b(unsigned idx, union fpureg *to);
extern void read_msa_wr_h(unsigned idx, union fpureg *to);
extern void read_msa_wr_w(unsigned idx, union fpureg *to);
extern void read_msa_wr_d(unsigned idx, union fpureg *to);

/**
 * read_msa_wr() - Read a single MSA vector register
 * @idx:        The index of the vector register to read
 * @to:         The FPU register union to store the registers value in
 * @fmt:        The format of the data in the vector register
 *
 * Read the value of MSA vector register idx into the FPU register
 * union to, using the format fmt.
 */
static inline void read_msa_wr(unsigned idx, union fpureg *to,
                               enum msa_2b_fmt fmt)
{
        switch (fmt) {
        case msa_fmt_b:
                read_msa_wr_b(idx, to);
                break;

        case msa_fmt_h:
                read_msa_wr_h(idx, to);
                break;

        case msa_fmt_w:
                read_msa_wr_w(idx, to);
                break;

        case msa_fmt_d:
                read_msa_wr_d(idx, to);
                break;

        default:
                BUG();
        }
}

extern void write_msa_wr_b(unsigned idx, union fpureg *from);
extern void write_msa_wr_h(unsigned idx, union fpureg *from);
extern void write_msa_wr_w(unsigned idx, union fpureg *from);
extern void write_msa_wr_d(unsigned idx, union fpureg *from);

/**
 * write_msa_wr() - Write a single MSA vector register
 * @idx:        The index of the vector register to write
 * @from:       The FPU register union to take the registers value from
 * @fmt:        The format of the data in the vector register
 *
 * Write the value from the FPU register union from into MSA vector
 * register idx, using the format fmt.
 */
static inline void write_msa_wr(unsigned idx, union fpureg *from,
                                enum msa_2b_fmt fmt)
{
        switch (fmt) {
        case msa_fmt_b:
                write_msa_wr_b(idx, from);
                break;

        case msa_fmt_h:
                write_msa_wr_h(idx, from);
                break;

        case msa_fmt_w:
                write_msa_wr_w(idx, from);
                break;

        case msa_fmt_d:
                write_msa_wr_d(idx, from);
                break;

        default:
                BUG();
        }
}

static inline void enable_msa(void)
{
        if (cpu_has_msa) {
                set_c0_config5(MIPS_CONF5_MSAEN);
                enable_fpu_hazard();
        }
}

static inline void disable_msa(void)
{
        if (cpu_has_msa) {
                clear_c0_config5(MIPS_CONF5_MSAEN);
                disable_fpu_hazard();
        }
}

static inline int is_msa_enabled(void)
{
        if (!cpu_has_msa)
                return 0;

        return read_c0_config5() & MIPS_CONF5_MSAEN;
}

static inline int thread_msa_context_live(void)
{
        /*
         * Check cpu_has_msa only if it's a constant. This will allow the
         * compiler to optimise out code for CPUs without MSA without adding
         * an extra redundant check for CPUs with MSA.
         */
        if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa)
                return 0;

        return test_thread_flag(TIF_MSA_CTX_LIVE);
}

static inline void save_msa(struct task_struct *t)
{
        if (cpu_has_msa)
                _save_msa(t);
}

static inline void restore_msa(struct task_struct *t)
{
        if (cpu_has_msa)
                _restore_msa(t);
}

static inline void init_msa_upper(void)
{
        /*
         * Check cpu_has_msa only if it's a constant. This will allow the
         * compiler to optimise out code for CPUs without MSA without adding
         * an extra redundant check for CPUs with MSA.
         */
        if (__builtin_constant_p(cpu_has_msa) && !cpu_has_msa)
                return;

        _init_msa_upper();
}

#ifndef TOOLCHAIN_SUPPORTS_MSA
/*
 * Define assembler macros using .word for the c[ft]cmsa instructions in order
 * to allow compilation with toolchains that do not support MSA. Once all
 * toolchains in use support MSA these can be removed.
 */

#define _ASM_SET_CFCMSA                                                 \
        _ASM_MACRO_2R(cfcmsa, rd, cs,                                   \
                      _ASM_INSN_IF_MIPS(0x787e0019 | __cs << 11 | __rd << 6)    \
                      _ASM_INSN32_IF_MM(0x587e0016 | __cs << 11 | __rd << 6))
#define _ASM_UNSET_CFCMSA ".purgem cfcmsa\n\t"
#define _ASM_SET_CTCMSA                                                 \
        _ASM_MACRO_2R(ctcmsa, cd, rs,                                   \
                      _ASM_INSN_IF_MIPS(0x783e0019 | __rs << 11 | __cd << 6)    \
                      _ASM_INSN32_IF_MM(0x583e0016 | __rs << 11 | __cd << 6))
#define _ASM_UNSET_CTCMSA ".purgem ctcmsa\n\t"
#else /* TOOLCHAIN_SUPPORTS_MSA */
#define _ASM_SET_CFCMSA                                         \
                ".set\tfp=64\n\t"                               \
                ".set\tmsa\n\t"
#define _ASM_UNSET_CFCMSA
#define _ASM_SET_CTCMSA                                         \
                ".set\tfp=64\n\t"                               \
                ".set\tmsa\n\t"
#define _ASM_UNSET_CTCMSA
#endif

#define __BUILD_MSA_CTL_REG(name, cs)                           \
static inline unsigned int read_msa_##name(void)                \
{                                                               \
        unsigned int reg;                                       \
        __asm__ __volatile__(                                   \
        "       .set    push\n"                                 \
        _ASM_SET_CFCMSA                                         \
        "       cfcmsa  %0, $" #cs "\n"                         \
        _ASM_UNSET_CFCMSA                                       \
        "       .set    pop\n"                                  \
        : "=r"(reg));                                           \
        return reg;                                             \
}                                                               \
                                                                \
static inline void write_msa_##name(unsigned int val)           \
{                                                               \
        __asm__ __volatile__(                                   \
        "       .set    push\n"                                 \
        _ASM_SET_CTCMSA                                         \
        "       ctcmsa  $" #cs ", %0\n"                         \
        _ASM_UNSET_CTCMSA                                       \
        "       .set    pop\n"                                  \
        : : "r"(val));                                          \
}

__BUILD_MSA_CTL_REG(ir, 0)
__BUILD_MSA_CTL_REG(csr, 1)
__BUILD_MSA_CTL_REG(access, 2)
__BUILD_MSA_CTL_REG(save, 3)
__BUILD_MSA_CTL_REG(modify, 4)
__BUILD_MSA_CTL_REG(request, 5)
__BUILD_MSA_CTL_REG(map, 6)
__BUILD_MSA_CTL_REG(unmap, 7)

#endif /* !__ASSEMBLER__ */

#define MSA_IR          0
#define MSA_CSR         1
#define MSA_ACCESS      2
#define MSA_SAVE        3
#define MSA_MODIFY      4
#define MSA_REQUEST     5
#define MSA_MAP         6
#define MSA_UNMAP       7

/* MSA Implementation Register (MSAIR) */
#define MSA_IR_REVB             0
#define MSA_IR_REVF             (_ULCAST_(0xff) << MSA_IR_REVB)
#define MSA_IR_PROCB            8
#define MSA_IR_PROCF            (_ULCAST_(0xff) << MSA_IR_PROCB)
#define MSA_IR_WRPB             16
#define MSA_IR_WRPF             (_ULCAST_(0x1) << MSA_IR_WRPB)

/* MSA Control & Status Register (MSACSR) */
#define MSA_CSR_RMB             0
#define MSA_CSR_RMF             (_ULCAST_(0x3) << MSA_CSR_RMB)
#define MSA_CSR_RM_NEAREST      0
#define MSA_CSR_RM_TO_ZERO      1
#define MSA_CSR_RM_TO_POS       2
#define MSA_CSR_RM_TO_NEG       3
#define MSA_CSR_FLAGSB          2
#define MSA_CSR_FLAGSF          (_ULCAST_(0x1f) << MSA_CSR_FLAGSB)
#define MSA_CSR_FLAGS_IB        2
#define MSA_CSR_FLAGS_IF        (_ULCAST_(0x1) << MSA_CSR_FLAGS_IB)
#define MSA_CSR_FLAGS_UB        3
#define MSA_CSR_FLAGS_UF        (_ULCAST_(0x1) << MSA_CSR_FLAGS_UB)
#define MSA_CSR_FLAGS_OB        4
#define MSA_CSR_FLAGS_OF        (_ULCAST_(0x1) << MSA_CSR_FLAGS_OB)
#define MSA_CSR_FLAGS_ZB        5
#define MSA_CSR_FLAGS_ZF        (_ULCAST_(0x1) << MSA_CSR_FLAGS_ZB)
#define MSA_CSR_FLAGS_VB        6
#define MSA_CSR_FLAGS_VF        (_ULCAST_(0x1) << MSA_CSR_FLAGS_VB)
#define MSA_CSR_ENABLESB        7
#define MSA_CSR_ENABLESF        (_ULCAST_(0x1f) << MSA_CSR_ENABLESB)
#define MSA_CSR_ENABLES_IB      7
#define MSA_CSR_ENABLES_IF      (_ULCAST_(0x1) << MSA_CSR_ENABLES_IB)
#define MSA_CSR_ENABLES_UB      8
#define MSA_CSR_ENABLES_UF      (_ULCAST_(0x1) << MSA_CSR_ENABLES_UB)
#define MSA_CSR_ENABLES_OB      9
#define MSA_CSR_ENABLES_OF      (_ULCAST_(0x1) << MSA_CSR_ENABLES_OB)
#define MSA_CSR_ENABLES_ZB      10
#define MSA_CSR_ENABLES_ZF      (_ULCAST_(0x1) << MSA_CSR_ENABLES_ZB)
#define MSA_CSR_ENABLES_VB      11
#define MSA_CSR_ENABLES_VF      (_ULCAST_(0x1) << MSA_CSR_ENABLES_VB)
#define MSA_CSR_CAUSEB          12
#define MSA_CSR_CAUSEF          (_ULCAST_(0x3f) << MSA_CSR_CAUSEB)
#define MSA_CSR_CAUSE_IB        12
#define MSA_CSR_CAUSE_IF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_IB)
#define MSA_CSR_CAUSE_UB        13
#define MSA_CSR_CAUSE_UF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_UB)
#define MSA_CSR_CAUSE_OB        14
#define MSA_CSR_CAUSE_OF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_OB)
#define MSA_CSR_CAUSE_ZB        15
#define MSA_CSR_CAUSE_ZF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_ZB)
#define MSA_CSR_CAUSE_VB        16
#define MSA_CSR_CAUSE_VF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_VB)
#define MSA_CSR_CAUSE_EB        17
#define MSA_CSR_CAUSE_EF        (_ULCAST_(0x1) << MSA_CSR_CAUSE_EB)
#define MSA_CSR_NXB             18
#define MSA_CSR_NXF             (_ULCAST_(0x1) << MSA_CSR_NXB)
#define MSA_CSR_FSB             24
#define MSA_CSR_FSF             (_ULCAST_(0x1) << MSA_CSR_FSB)

#endif /* _ASM_MSA_H */