#include <linux/bug.h>
#include "bpf_jit.h"
enum {
ARC_R_0, ARC_R_1, ARC_R_2, ARC_R_3, ARC_R_4, ARC_R_5,
ARC_R_6, ARC_R_7, ARC_R_8, ARC_R_9, ARC_R_10, ARC_R_11,
ARC_R_12, ARC_R_13, ARC_R_14, ARC_R_15, ARC_R_16, ARC_R_17,
ARC_R_18, ARC_R_19, ARC_R_20, ARC_R_21, ARC_R_22, ARC_R_23,
ARC_R_24, ARC_R_25, ARC_R_26, ARC_R_FP, ARC_R_SP, ARC_R_ILINK,
ARC_R_30, ARC_R_BLINK,
ARC_R_IMM = 62
};
static const u8 bpf2arc[][2] = {
[BPF_REG_0] = {ARC_R_8, ARC_R_9},
[BPF_REG_1] = {ARC_R_0, ARC_R_1},
[BPF_REG_2] = {ARC_R_2, ARC_R_3},
[BPF_REG_3] = {ARC_R_4, ARC_R_5},
[BPF_REG_4] = {ARC_R_6, ARC_R_7},
[BPF_REG_5] = {ARC_R_22, ARC_R_23},
[BPF_REG_6] = {ARC_R_14, ARC_R_15},
[BPF_REG_7] = {ARC_R_16, ARC_R_17},
[BPF_REG_8] = {ARC_R_18, ARC_R_19},
[BPF_REG_9] = {ARC_R_20, ARC_R_21},
[BPF_REG_FP] = {ARC_R_FP, },
[BPF_REG_AX] = {ARC_R_24, ARC_R_25},
[JIT_REG_TMP] = {ARC_R_10, ARC_R_11}
};
#define ARC_CALLEE_SAVED_REG_FIRST ARC_R_13
#define ARC_CALLEE_SAVED_REG_LAST ARC_R_25
#define REG_LO(r) (bpf2arc[(r)][0])
#define REG_HI(r) (bpf2arc[(r)][1])
#define ARG5_SIZE 8
enum {
INSN_len_normal = 4,
INSN_len_imm = 4
};
enum {
ZZ_1_byte = 1,
ZZ_2_byte = 2,
ZZ_4_byte = 0,
ZZ_8_byte = 3
};
enum {
AA_none = 0,
AA_pre = 1,
AA_post = 2,
AA_scale = 3
};
enum {
X_zero = 0,
X_sign = 1
};
enum {
CC_always = 0,
CC_equal = 1,
CC_unequal = 2,
CC_positive = 3,
CC_negative = 4,
CC_less_u = 5,
CC_less_eq_u = 14,
CC_great_eq_u = 6,
CC_great_u = 13,
CC_less_s = 11,
CC_less_eq_s = 12,
CC_great_eq_s = 10,
CC_great_s = 9
};
#define IN_U6_RANGE(x) ((x) <= (0x40 - 1) && (x) >= 0)
#define IN_S9_RANGE(x) ((x) <= (0x100 - 1) && (x) >= -0x100)
#define IN_S12_RANGE(x) ((x) <= (0x800 - 1) && (x) >= -0x800)
#define IN_S21_RANGE(x) ((x) <= (0x100000 - 1) && (x) >= -0x100000)
#define IN_S25_RANGE(x) ((x) <= (0x1000000 - 1) && (x) >= -0x1000000)
#define OP_A(x) ((x) & 0x03f)
#define OP_B(x) ((((x) & 0x07) << 24) | (((x) & 0x38) << 9))
#define OP_C(x) (((x) & 0x03f) << 6)
#define OP_IMM (OP_C(ARC_R_IMM))
#define COND(x) (OP_A((x) & 31))
#define FLAG(x) (((x) & 1) << 15)
#define OPC_MOV 0x200a0000
#define OPC_MOVI 0x208a0000
#define MOVI_S12(x) ((((x) & 0xfc0) >> 6) | (((x) & 0x3f) << 6))
#define OPC_MOV_CC 0x20ca0000
#define MOV_CC_I BIT(5)
#define OPC_MOVU_CC (OPC_MOV_CC | MOV_CC_I)
#define OPC_SEXB 0x202f0005
#define OPC_SEXH 0x202f0006
#define OPC_LOAD 0x10000000
#define LOAD_X(x) ((x) << 6)
#define LOAD_ZZ(x) ((x) << 7)
#define LOAD_AA(x) ((x) << 9)
#define LOAD_S9(x) ((((x) & 0x0ff) << 16) | (((x) & 0x100) << 7))
#define LOAD_C(x) ((x) & 0x03f)
#define OPC_LDU (OPC_LOAD | LOAD_X(X_zero))
#define OPC_LDS (OPC_LOAD | LOAD_X(X_sign))
#define OPC_LD32 (OPC_LDU | LOAD_ZZ(ZZ_4_byte))
#define OPC_POP \
(OPC_LD32 | LOAD_AA(AA_post) | LOAD_S9(4) | OP_B(ARC_R_SP))
#define OPC_STORE 0x18000000
#define STORE_ZZ(x) ((x) << 1)
#define STORE_AA(x) ((x) << 3)
#define STORE_S9(x) ((((x) & 0x0ff) << 16) | (((x) & 0x100) << 7))
#define OPC_ST32 (OPC_STORE | STORE_ZZ(ZZ_4_byte))
#define OPC_PUSH \
(OPC_ST32 | STORE_AA(AA_pre) | STORE_S9(-4) | OP_B(ARC_R_SP))
#define OPC_ADD 0x20000000
#define OPC_ADDF (OPC_ADD | FLAG(1))
#define ADDI BIT(22)
#define ADDI_U6(x) OP_C(x)
#define OPC_ADDI (OPC_ADD | ADDI)
#define OPC_ADDIF (OPC_ADDI | FLAG(1))
#define OPC_ADD_I (OPC_ADD | OP_IMM)
#define OPC_ADC 0x20010000
#define ADCI BIT(22)
#define ADCI_U6(x) OP_C(x)
#define OPC_ADCI (OPC_ADC | ADCI)
#define OPC_SUB 0x20020000
#define OPC_SUBF (OPC_SUB | FLAG(1))
#define SUBI BIT(22)
#define SUBI_U6(x) OP_C(x)
#define OPC_SUBI (OPC_SUB | SUBI)
#define OPC_SUB_I (OPC_SUB | OP_IMM)
#define OPC_SBC 0x20030000
#define OPC_CMP 0x20cc8000
#define OPC_NEG 0x204e0000
#define OPC_MPY 0x201a0000
#define OPC_MPYI (OPC_MPY | OP_IMM)
#define OPC_MPYDU 0x28190000
#define OPC_MPYDUI (OPC_MPYDU | OP_IMM)
#define OPC_DIVU 0x28050000
#define OPC_DIVUI (OPC_DIVU | OP_IMM)
#define OPC_DIVS 0x28040000
#define OPC_DIVSI (OPC_DIVS | OP_IMM)
#define OPC_REMU 0x28090000
#define OPC_REMUI (OPC_REMU | OP_IMM)
#define OPC_REMS 0x28080000
#define OPC_REMSI (OPC_REMS | OP_IMM)
#define OPC_AND 0x20040000
#define OPC_ANDI (OPC_AND | OP_IMM)
#define OPC_TST 0x20cb8000
#define OPC_OR 0x20050000
#define OPC_ORI (OPC_OR | OP_IMM)
#define OPC_XOR 0x20070000
#define OPC_XORI (OPC_XOR | OP_IMM)
#define OPC_NOT 0x202f000a
#define OPC_BTSTU6 0x20518000
#define BTST_U6(x) (OP_C((x) & 63))
#define OPC_ASL 0x28000000
#define ASL_I BIT(22)
#define ASLI_U6(x) OP_C((x) & 31)
#define OPC_ASLI (OPC_ASL | ASL_I)
#define OPC_ASR 0x28020000
#define ASR_I ASL_I
#define ASRI_U6(x) ASLI_U6(x)
#define OPC_ASRI (OPC_ASR | ASR_I)
#define OPC_LSR 0x28010000
#define LSR_I ASL_I
#define LSRI_U6(x) ASLI_U6(x)
#define OPC_LSRI (OPC_LSR | LSR_I)
#define OPC_SWAPE 0x282f0009
#define OPC_JMP 0x20e00000
#define OPC_J_BLINK (OPC_JMP | OP_C(ARC_R_BLINK))
#define OPC_JL 0x20220000
#define OPC_BCC 0x00000000
#define BCC_S21(d) ((((d) & 0x7fe) << 16) | (((d) & 0x1ff800) >> 5))
#define OPC_B 0x00010000
#define B_S25(d) ((((d) & 0x1e00000) >> 21) | BCC_S21(d))
static inline void emit_2_bytes(u8 *buf, u16 bytes)
{
*((u16 *)buf) = bytes;
}
static inline void emit_4_bytes(u8 *buf, u32 bytes)
{
emit_2_bytes(buf, bytes >> 16);
emit_2_bytes(buf + 2, bytes & 0xffff);
}
static inline u8 bpf_to_arc_size(u8 size)
{
switch (size) {
case BPF_B:
return ZZ_1_byte;
case BPF_H:
return ZZ_2_byte;
case BPF_W:
return ZZ_4_byte;
case BPF_DW:
return ZZ_8_byte;
default:
return ZZ_4_byte;
}
}
static u8 arc_movi_r(u8 *buf, u8 reg, s16 imm)
{
const u32 insn = OPC_MOVI | OP_B(reg) | MOVI_S12(imm);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_mov_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_MOV | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_mov_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_MOV | OP_B(rd) | OP_IMM;
if (IN_S12_RANGE(imm))
return arc_movi_r(buf, rd, imm);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_mov_i_fixed(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_MOV | OP_B(rd) | OP_IMM;
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_mov_cc_r(u8 *buf, u8 cc, u8 rd, u8 rs)
{
const u32 insn = OPC_MOV_CC | OP_B(rd) | OP_C(rs) | COND(cc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_movu_cc_r(u8 *buf, u8 cc, u8 rd, u8 imm)
{
const u32 insn = OPC_MOVU_CC | OP_B(rd) | OP_C(imm) | COND(cc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_sexb_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_SEXB | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_sexh_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_SEXH | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_st_r(u8 *buf, u8 reg, u8 reg_mem, s16 off, u8 zz)
{
const u32 insn = OPC_STORE | STORE_ZZ(zz) | OP_C(reg) |
OP_B(reg_mem) | STORE_S9(off);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_push_r(u8 *buf, u8 reg)
{
const u32 insn = OPC_PUSH | OP_C(reg);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_ld_r(u8 *buf, u8 reg, u8 reg_mem, s16 off, u8 zz)
{
const u32 insn = OPC_LDU | LOAD_ZZ(zz) | LOAD_C(reg) |
OP_B(reg_mem) | LOAD_S9(off);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_ldx_r(u8 *buf, u8 reg, u8 reg_mem, s16 off, u8 zz)
{
const u32 insn = OPC_LDS | LOAD_ZZ(zz) | LOAD_C(reg) |
OP_B(reg_mem) | LOAD_S9(off);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_pop_r(u8 *buf, u8 reg)
{
const u32 insn = OPC_POP | LOAD_C(reg);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_add_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_ADD | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_addf_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_ADDF | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_addif_r(u8 *buf, u8 ra, u8 u6)
{
const u32 insn = OPC_ADDIF | OP_A(ra) | OP_B(ra) | ADDI_U6(u6);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_addi_r(u8 *buf, u8 ra, u8 u6)
{
const u32 insn = OPC_ADDI | OP_A(ra) | OP_B(ra) | ADDI_U6(u6);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_add_i(u8 *buf, u8 ra, u8 rb, s32 imm)
{
const u32 insn = OPC_ADD_I | OP_A(ra) | OP_B(rb);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_adc_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_ADC | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_adci_r(u8 *buf, u8 ra, u8 u6)
{
const u32 insn = OPC_ADCI | OP_A(ra) | OP_B(ra) | ADCI_U6(u6);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_sub_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_SUB | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_subf_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_SUBF | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_subi_r(u8 *buf, u8 ra, u8 u6)
{
const u32 insn = OPC_SUBI | OP_A(ra) | OP_B(ra) | SUBI_U6(u6);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_sub_i(u8 *buf, u8 ra, s32 imm)
{
const u32 insn = OPC_SUB_I | OP_A(ra) | OP_B(ra);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_sbc_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_SBC | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_cmp_r(u8 *buf, u8 rb, u8 rc)
{
const u32 insn = OPC_CMP | OP_B(rb) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_cmpz_r(u8 *buf, u8 rb, u8 rc)
{
const u32 insn = OPC_CMP | OP_B(rb) | OP_C(rc) | CC_equal;
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_neg_r(u8 *buf, u8 ra, u8 rb)
{
const u32 insn = OPC_NEG | OP_A(ra) | OP_B(rb);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_mpy_r(u8 *buf, u8 ra, u8 rb, u8 rc)
{
const u32 insn = OPC_MPY | OP_A(ra) | OP_B(rb) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_mpy_i(u8 *buf, u8 ra, u8 rb, s32 imm)
{
const u32 insn = OPC_MPYI | OP_A(ra) | OP_B(rb);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_mpydu_r(u8 *buf, u8 ra, u8 rc)
{
const u32 insn = OPC_MPYDU | OP_A(ra) | OP_B(ra) | OP_C(rc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_mpydu_i(u8 *buf, u8 ra, s32 imm)
{
const u32 insn = OPC_MPYDUI | OP_A(ra) | OP_B(ra);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_divu_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_DIVU | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_divu_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_DIVUI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_divs_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_DIVS | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_divs_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_DIVSI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_remu_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_REMU | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_remu_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_REMUI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_rems_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_REMS | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_rems_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_REMSI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_and_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_AND | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_and_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_ANDI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_tst_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_TST | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_tstz_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_TST | OP_B(rd) | OP_C(rs) | CC_equal;
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_or_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
{
const u32 insn = OPC_OR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_or_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_ORI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_xor_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_XOR | OP_A(rd) | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_xor_i(u8 *buf, u8 rd, s32 imm)
{
const u32 insn = OPC_XORI | OP_A(rd) | OP_B(rd);
if (buf) {
emit_4_bytes(buf, insn);
emit_4_bytes(buf + INSN_len_normal, imm);
}
return INSN_len_normal + INSN_len_imm;
}
static u8 arc_not_r(u8 *buf, u8 rd, u8 rs)
{
const u32 insn = OPC_NOT | OP_B(rd) | OP_C(rs);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_btst_i(u8 *buf, u8 rs, u8 imm)
{
const u32 insn = OPC_BTSTU6 | OP_B(rs) | BTST_U6(imm);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_asl_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
{
const u32 insn = OPC_ASL | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_asli_r(u8 *buf, u8 rd, u8 rs, u8 imm)
{
const u32 insn = OPC_ASLI | OP_A(rd) | OP_B(rs) | ASLI_U6(imm);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_asr_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
{
const u32 insn = OPC_ASR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_asri_r(u8 *buf, u8 rd, u8 rs, u8 imm)
{
const u32 insn = OPC_ASRI | OP_A(rd) | OP_B(rs) | ASRI_U6(imm);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_lsr_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
{
const u32 insn = OPC_LSR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_lsri_r(u8 *buf, u8 rd, u8 rs, u8 imm)
{
const u32 insn = OPC_LSRI | OP_A(rd) | OP_B(rs) | LSRI_U6(imm);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_swape_r(u8 *buf, u8 r)
{
const u32 insn = OPC_SWAPE | OP_B(r) | OP_C(r);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_jmp_return(u8 *buf)
{
if (buf)
emit_4_bytes(buf, OPC_J_BLINK);
return INSN_len_normal;
}
static u8 arc_jl(u8 *buf, u8 reg)
{
const u32 insn = OPC_JL | OP_C(reg);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_bcc(u8 *buf, u8 cc, int offset)
{
const u32 insn = OPC_BCC | BCC_S21(offset) | COND(cc);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
static u8 arc_b(u8 *buf, s32 offset)
{
const u32 insn = OPC_B | B_S25(offset);
if (buf)
emit_4_bytes(buf, insn);
return INSN_len_normal;
}
u8 zext(u8 *buf, u8 rd)
{
if (rd != BPF_REG_FP)
return arc_movi_r(buf, REG_HI(rd), 0);
else
return 0;
}
u8 mov_r32(u8 *buf, u8 rd, u8 rs, u8 sign_ext)
{
u8 len = 0;
if (sign_ext) {
if (sign_ext == 8)
len = arc_sexb_r(buf, REG_LO(rd), REG_LO(rs));
else if (sign_ext == 16)
len = arc_sexh_r(buf, REG_LO(rd), REG_LO(rs));
else if (sign_ext == 32 && rd != rs)
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
return len;
}
if (rd != rs)
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
return len;
}
u8 mov_r32_i32(u8 *buf, u8 reg, s32 imm)
{
return arc_mov_i(buf, REG_LO(reg), imm);
}
u8 mov_r64(u8 *buf, u8 rd, u8 rs, u8 sign_ext)
{
u8 len = 0;
if (sign_ext) {
len = mov_r32(buf, rd, rs, sign_ext);
if (sign_ext == 8 || sign_ext == 16 || sign_ext == 32) {
len += arc_asri_r(BUF(buf, len),
REG_HI(rd), REG_LO(rd), 31);
}
return len;
}
if (rd == rs)
return 0;
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
if (rs != BPF_REG_FP)
len += arc_mov_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
else
len += arc_movi_r(BUF(buf, len), REG_HI(rd), 0);
return len;
}
u8 mov_r64_i32(u8 *buf, u8 reg, s32 imm)
{
u8 len = 0;
len = arc_mov_i(buf, REG_LO(reg), imm);
if (reg != BPF_REG_FP) {
if (imm >= 0)
len += arc_movi_r(BUF(buf, len), REG_HI(reg), 0);
else
len += arc_movi_r(BUF(buf, len), REG_HI(reg), -1);
}
return len;
}
u8 mov_r64_i64(u8 *buf, u8 reg, u32 lo, u32 hi)
{
u8 len;
len = arc_mov_i_fixed(buf, REG_LO(reg), lo);
len += arc_mov_i_fixed(BUF(buf, len), REG_HI(reg), hi);
return len;
}
static u8 adjust_mem_access(u8 *buf, s16 *off, u8 size,
u8 rm, u8 *arc_reg_mem)
{
u8 len = 0;
*arc_reg_mem = REG_LO(rm);
if (!IN_S9_RANGE(*off) ||
(size == BPF_DW && !IN_S9_RANGE(*off + 4))) {
len += arc_add_i(BUF(buf, len),
REG_LO(JIT_REG_TMP), REG_LO(rm), (u32)(*off));
*arc_reg_mem = REG_LO(JIT_REG_TMP);
*off = 0;
}
return len;
}
u8 store_r(u8 *buf, u8 rs, u8 rd, s16 off, u8 size)
{
u8 len, arc_reg_mem;
len = adjust_mem_access(buf, &off, size, rd, &arc_reg_mem);
if (size == BPF_DW) {
len += arc_st_r(BUF(buf, len), REG_LO(rs), arc_reg_mem,
off, ZZ_4_byte);
len += arc_st_r(BUF(buf, len), REG_HI(rs), arc_reg_mem,
off + 4, ZZ_4_byte);
} else {
u8 zz = bpf_to_arc_size(size);
len += arc_st_r(BUF(buf, len), REG_LO(rs), arc_reg_mem,
off, zz);
}
return len;
}
u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size)
{
u8 len, arc_reg_mem;
const u8 arc_rs = REG_HI(JIT_REG_TMP);
len = adjust_mem_access(buf, &off, size, rd, &arc_reg_mem);
if (size == BPF_DW) {
len += arc_mov_i(BUF(buf, len), arc_rs, imm);
len += arc_st_r(BUF(buf, len), arc_rs, arc_reg_mem,
off, ZZ_4_byte);
imm = (imm >= 0 ? 0 : -1);
len += arc_mov_i(BUF(buf, len), arc_rs, imm);
len += arc_st_r(BUF(buf, len), arc_rs, arc_reg_mem,
off + 4, ZZ_4_byte);
} else {
u8 zz = bpf_to_arc_size(size);
len += arc_mov_i(BUF(buf, len), arc_rs, imm);
len += arc_st_r(BUF(buf, len), arc_rs, arc_reg_mem, off, zz);
}
return len;
}
static u8 push_r64(u8 *buf, u8 reg)
{
u8 len = 0;
#ifdef __LITTLE_ENDIAN
if (reg != BPF_REG_FP)
len += arc_push_r(BUF(buf, len), REG_HI(reg));
len += arc_push_r(BUF(buf, len), REG_LO(reg));
#else
len += arc_push_r(BUF(buf, len), REG_LO(reg));
if (reg != BPF_REG_FP)
len += arc_push_r(BUF(buf, len), REG_HI(reg));
#endif
return len;
}
u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext)
{
u8 len, arc_reg_mem;
len = adjust_mem_access(buf, &off, size, rs, &arc_reg_mem);
if (size == BPF_B || size == BPF_H || size == BPF_W) {
const u8 zz = bpf_to_arc_size(size);
if (sign_ext && (zz == ZZ_1_byte || zz == ZZ_2_byte)) {
len += arc_ldx_r(BUF(buf, len), REG_LO(rd),
arc_reg_mem, off, zz);
} else {
len += arc_ld_r(BUF(buf, len), REG_LO(rd),
arc_reg_mem, off, zz);
}
if (sign_ext) {
len += arc_asri_r(BUF(buf, len),
REG_HI(rd), REG_LO(rd), 31);
} else {
len += arc_movi_r(BUF(buf, len), REG_HI(rd), 0);
}
} else if (size == BPF_DW) {
if (REG_LO(rd) != arc_reg_mem) {
len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
off, ZZ_4_byte);
len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
off + 4, ZZ_4_byte);
} else {
len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
off + 4, ZZ_4_byte);
len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
off, ZZ_4_byte);
}
}
return len;
}
u8 add_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_add_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 add_r32_i32(u8 *buf, u8 rd, s32 imm)
{
if (IN_U6_RANGE(imm))
return arc_addi_r(buf, REG_LO(rd), imm);
else
return arc_add_i(buf, REG_LO(rd), REG_LO(rd), imm);
}
u8 add_r64(u8 *buf, u8 rd, u8 rs)
{
u8 len;
len = arc_addf_r(buf, REG_LO(rd), REG_LO(rs));
len += arc_adc_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
return len;
}
u8 add_r64_i32(u8 *buf, u8 rd, s32 imm)
{
u8 len;
if (IN_U6_RANGE(imm)) {
len = arc_addif_r(buf, REG_LO(rd), imm);
len += arc_adci_r(BUF(buf, len), REG_HI(rd), 0);
} else {
len = mov_r64_i32(buf, JIT_REG_TMP, imm);
len += add_r64(BUF(buf, len), rd, JIT_REG_TMP);
}
return len;
}
u8 sub_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_sub_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 sub_r32_i32(u8 *buf, u8 rd, s32 imm)
{
if (IN_U6_RANGE(imm))
return arc_subi_r(buf, REG_LO(rd), imm);
else
return arc_sub_i(buf, REG_LO(rd), imm);
}
u8 sub_r64(u8 *buf, u8 rd, u8 rs)
{
u8 len;
len = arc_subf_r(buf, REG_LO(rd), REG_LO(rs));
len += arc_sbc_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
return len;
}
u8 sub_r64_i32(u8 *buf, u8 rd, s32 imm)
{
u8 len;
len = mov_r64_i32(buf, JIT_REG_TMP, imm);
len += sub_r64(BUF(buf, len), rd, JIT_REG_TMP);
return len;
}
static u8 cmp_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_cmp_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 neg_r32(u8 *buf, u8 r)
{
return arc_neg_r(buf, REG_LO(r), REG_LO(r));
}
u8 neg_r64(u8 *buf, u8 r)
{
u8 len;
len = arc_not_r(buf, REG_LO(r), REG_LO(r));
len += arc_not_r(BUF(buf, len), REG_HI(r), REG_HI(r));
len += add_r64_i32(BUF(buf, len), r, 1);
return len;
}
u8 mul_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_mpy_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
}
u8 mul_r32_i32(u8 *buf, u8 rd, s32 imm)
{
return arc_mpy_i(buf, REG_LO(rd), REG_LO(rd), imm);
}
u8 mul_r64(u8 *buf, u8 rd, u8 rs)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 t1 = REG_HI(JIT_REG_TMP);
const u8 C_lo = REG_LO(rs);
const u8 C_hi = REG_HI(rs);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
u8 len;
len = arc_mpy_r(buf, t0, B_hi, C_lo);
len += arc_mpy_r(BUF(buf, len), t1, B_lo, C_hi);
len += arc_mpydu_r(BUF(buf, len), B_lo, C_lo);
len += arc_add_r(BUF(buf, len), B_hi, t0);
len += arc_add_r(BUF(buf, len), B_hi, t1);
return len;
}
u8 mul_r64_i32(u8 *buf, u8 rd, s32 imm)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 t1 = REG_HI(JIT_REG_TMP);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
u8 len = 0;
if (imm == 1)
return 0;
if (imm < 0)
len += arc_neg_r(BUF(buf, len), t1, B_lo);
len += arc_mpy_i(BUF(buf, len), t0, B_hi, imm);
len += arc_mpydu_i(BUF(buf, len), B_lo, imm);
len += arc_add_r(BUF(buf, len), B_hi, t0);
if (imm < 0)
len += arc_add_r(BUF(buf, len), B_hi, t1);
return len;
}
u8 div_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext)
{
if (sign_ext)
return arc_divs_r(buf, REG_LO(rd), REG_LO(rs));
else
return arc_divu_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 div_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext)
{
if (imm == 0)
return 0;
if (sign_ext)
return arc_divs_i(buf, REG_LO(rd), imm);
else
return arc_divu_i(buf, REG_LO(rd), imm);
}
u8 mod_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext)
{
if (sign_ext)
return arc_rems_r(buf, REG_LO(rd), REG_LO(rs));
else
return arc_remu_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 mod_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext)
{
if (imm == 0)
return 0;
if (sign_ext)
return arc_rems_i(buf, REG_LO(rd), imm);
else
return arc_remu_i(buf, REG_LO(rd), imm);
}
u8 and_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_and_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 and_r32_i32(u8 *buf, u8 rd, s32 imm)
{
return arc_and_i(buf, REG_LO(rd), imm);
}
u8 and_r64(u8 *buf, u8 rd, u8 rs)
{
u8 len;
len = arc_and_r(buf, REG_LO(rd), REG_LO(rs));
len += arc_and_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
return len;
}
u8 and_r64_i32(u8 *buf, u8 rd, s32 imm)
{
u8 len;
len = mov_r64_i32(buf, JIT_REG_TMP, imm);
len += and_r64(BUF(buf, len), rd, JIT_REG_TMP);
return len;
}
static u8 tst_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_tst_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 or_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_or_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
}
u8 or_r32_i32(u8 *buf, u8 rd, s32 imm)
{
return arc_or_i(buf, REG_LO(rd), imm);
}
u8 or_r64(u8 *buf, u8 rd, u8 rs)
{
u8 len;
len = arc_or_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
len += arc_or_r(BUF(buf, len), REG_HI(rd), REG_HI(rd), REG_HI(rs));
return len;
}
u8 or_r64_i32(u8 *buf, u8 rd, s32 imm)
{
u8 len;
len = mov_r64_i32(buf, JIT_REG_TMP, imm);
len += or_r64(BUF(buf, len), rd, JIT_REG_TMP);
return len;
}
u8 xor_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_xor_r(buf, REG_LO(rd), REG_LO(rs));
}
u8 xor_r32_i32(u8 *buf, u8 rd, s32 imm)
{
return arc_xor_i(buf, REG_LO(rd), imm);
}
u8 xor_r64(u8 *buf, u8 rd, u8 rs)
{
u8 len;
len = arc_xor_r(buf, REG_LO(rd), REG_LO(rs));
len += arc_xor_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
return len;
}
u8 xor_r64_i32(u8 *buf, u8 rd, s32 imm)
{
u8 len;
len = mov_r64_i32(buf, JIT_REG_TMP, imm);
len += xor_r64(BUF(buf, len), rd, JIT_REG_TMP);
return len;
}
u8 lsh_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_asl_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
}
u8 lsh_r32_i32(u8 *buf, u8 rd, u8 imm)
{
return arc_asli_r(buf, REG_LO(rd), REG_LO(rd), imm);
}
u8 lsh_r64(u8 *buf, u8 rd, u8 rs)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 t1 = REG_HI(JIT_REG_TMP);
const u8 C_lo = REG_LO(rs);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
u8 len;
len = arc_not_r(buf, t0, C_lo);
len += arc_lsri_r(BUF(buf, len), t1, B_lo, 1);
len += arc_lsr_r(BUF(buf, len), t1, t1, t0);
len += arc_mov_r(BUF(buf, len), t0, C_lo);
len += arc_asl_r(BUF(buf, len), B_lo, B_lo, t0);
len += arc_asl_r(BUF(buf, len), B_hi, B_hi, t0);
len += arc_or_r(BUF(buf, len), B_hi, B_hi, t1);
len += arc_btst_i(BUF(buf, len), t0, 5);
len += arc_mov_cc_r(BUF(buf, len), CC_unequal, B_hi, B_lo);
len += arc_movu_cc_r(BUF(buf, len), CC_unequal, B_lo, 0);
return len;
}
u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
const u8 n = (u8)imm;
u8 len = 0;
if (n == 0) {
return 0;
} else if (n <= 31) {
len = arc_lsri_r(buf, t0, B_lo, 32 - n);
len += arc_asli_r(BUF(buf, len), B_lo, B_lo, n);
len += arc_asli_r(BUF(buf, len), B_hi, B_hi, n);
len += arc_or_r(BUF(buf, len), B_hi, B_hi, t0);
} else if (n <= 63) {
len = arc_asli_r(buf, B_hi, B_lo, n - 32);
len += arc_movi_r(BUF(buf, len), B_lo, 0);
}
return len;
}
u8 rsh_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_lsr_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
}
u8 rsh_r32_i32(u8 *buf, u8 rd, u8 imm)
{
return arc_lsri_r(buf, REG_LO(rd), REG_LO(rd), imm);
}
u8 rsh_r64(u8 *buf, u8 rd, u8 rs)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 t1 = REG_HI(JIT_REG_TMP);
const u8 C_lo = REG_LO(rs);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
u8 len;
len = arc_not_r(buf, t0, C_lo);
len += arc_asli_r(BUF(buf, len), t1, B_hi, 1);
len += arc_asl_r(BUF(buf, len), t1, t1, t0);
len += arc_mov_r(BUF(buf, len), t0, C_lo);
len += arc_lsr_r(BUF(buf, len), B_hi, B_hi, t0);
len += arc_lsr_r(BUF(buf, len), B_lo, B_lo, t0);
len += arc_or_r(BUF(buf, len), B_lo, B_lo, t1);
len += arc_btst_i(BUF(buf, len), t0, 5);
len += arc_mov_cc_r(BUF(buf, len), CC_unequal, B_lo, B_hi);
len += arc_movu_cc_r(BUF(buf, len), CC_unequal, B_hi, 0);
return len;
}
u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
const u8 n = (u8)imm;
u8 len = 0;
if (n == 0) {
return 0;
} else if (n <= 31) {
len = arc_asli_r(buf, t0, B_hi, 32 - n);
len += arc_lsri_r(BUF(buf, len), B_lo, B_lo, n);
len += arc_lsri_r(BUF(buf, len), B_hi, B_hi, n);
len += arc_or_r(BUF(buf, len), B_lo, B_lo, t0);
} else if (n <= 63) {
len = arc_lsri_r(buf, B_lo, B_hi, n - 32);
len += arc_movi_r(BUF(buf, len), B_hi, 0);
}
return len;
}
u8 arsh_r32(u8 *buf, u8 rd, u8 rs)
{
return arc_asr_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
}
u8 arsh_r32_i32(u8 *buf, u8 rd, u8 imm)
{
return arc_asri_r(buf, REG_LO(rd), REG_LO(rd), imm);
}
u8 arsh_r64(u8 *buf, u8 rd, u8 rs)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 t1 = REG_HI(JIT_REG_TMP);
const u8 C_lo = REG_LO(rs);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
u8 len;
len = arc_not_r(buf, t0, C_lo);
len += arc_asli_r(BUF(buf, len), t1, B_hi, 1);
len += arc_asl_r(BUF(buf, len), t1, t1, t0);
len += arc_mov_r(BUF(buf, len), t0, C_lo);
len += arc_asr_r(BUF(buf, len), B_hi, B_hi, t0);
len += arc_lsr_r(BUF(buf, len), B_lo, B_lo, t0);
len += arc_or_r(BUF(buf, len), B_lo, B_lo, t1);
len += arc_btst_i(BUF(buf, len), t0, 5);
len += arc_asri_r(BUF(buf, len), t0, B_hi, 31);
len += arc_mov_cc_r(BUF(buf, len), CC_unequal, B_lo, B_hi);
len += arc_mov_cc_r(BUF(buf, len), CC_unequal, B_hi, t0);
return len;
}
u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm)
{
const u8 t0 = REG_LO(JIT_REG_TMP);
const u8 B_lo = REG_LO(rd);
const u8 B_hi = REG_HI(rd);
const u8 n = (u8)imm;
u8 len = 0;
if (n == 0) {
return 0;
} else if (n <= 31) {
len = arc_asli_r(buf, t0, B_hi, 32 - n);
len += arc_lsri_r(BUF(buf, len), B_lo, B_lo, n);
len += arc_asri_r(BUF(buf, len), B_hi, B_hi, n);
len += arc_or_r(BUF(buf, len), B_lo, B_lo, t0);
} else if (n <= 63) {
len = arc_asri_r(buf, B_lo, B_hi, n - 32);
len += arc_movi_r(BUF(buf, len), B_hi, -1);
len += arc_btst_i(BUF(buf, len), B_lo, 31);
len += arc_movu_cc_r(BUF(buf, len), CC_equal, B_hi, 0);
}
return len;
}
u8 gen_swap(u8 *buf, u8 rd, u8 size, u8 endian, bool force, bool do_zext)
{
u8 len = 0;
#ifdef __BIG_ENDIAN
const u8 host_endian = BPF_FROM_BE;
#else
const u8 host_endian = BPF_FROM_LE;
#endif
if (host_endian != endian || force) {
switch (size) {
case 16:
len = arc_asli_r(buf, REG_LO(rd), REG_LO(rd), 16);
fallthrough;
case 32:
len += arc_swape_r(BUF(buf, len), REG_LO(rd));
if (do_zext)
len += zext(BUF(buf, len), rd);
break;
case 64:
len = arc_xor_r(buf, REG_HI(rd), REG_LO(rd));
len += arc_xor_r(BUF(buf, len), REG_LO(rd), REG_HI(rd));
len += arc_xor_r(BUF(buf, len), REG_HI(rd), REG_LO(rd));
len += arc_swape_r(BUF(buf, len), REG_LO(rd));
len += arc_swape_r(BUF(buf, len), REG_HI(rd));
break;
default:
break;
}
} else {
switch (size) {
case 16:
len = arc_and_i(buf, REG_LO(rd), 0xffff);
fallthrough;
case 32:
if (do_zext)
len += zext(BUF(buf, len), rd);
break;
case 64:
break;
default:
break;
}
}
return len;
}
static inline u8 frame_create(u8 *buf, u16 size)
{
u8 len;
len = arc_mov_r(buf, ARC_R_FP, ARC_R_SP);
if (IN_U6_RANGE(size))
len += arc_subi_r(BUF(buf, len), ARC_R_SP, size);
else
len += arc_sub_i(BUF(buf, len), ARC_R_SP, size);
return len;
}
static inline u8 frame_restore(u8 *buf)
{
return arc_mov_r(buf, ARC_R_SP, ARC_R_FP);
}
static u8 bpf_to_arc_return(u8 *buf)
{
u8 len;
len = arc_mov_r(buf, ARC_R_0, REG_LO(BPF_REG_0));
len += arc_mov_r(BUF(buf, len), ARC_R_1, REG_HI(BPF_REG_0));
return len;
}
u8 arc_to_bpf_return(u8 *buf)
{
u8 len;
len = arc_mov_r(buf, REG_LO(BPF_REG_0), ARC_R_0);
len += arc_mov_r(BUF(buf, len), REG_HI(BPF_REG_0), ARC_R_1);
return len;
}
static u8 jump_and_link(u8 *buf, u32 addr)
{
u8 len;
len = arc_mov_i_fixed(buf, REG_LO(JIT_REG_TMP), addr);
len += arc_jl(BUF(buf, len), REG_LO(JIT_REG_TMP));
return len;
}
u32 mask_for_used_regs(u8 bpf_reg, bool is_call)
{
u32 usage = 0;
if (bpf_reg >= BPF_REG_6 && bpf_reg <= BPF_REG_9) {
usage |= BIT(REG_LO(bpf_reg));
usage |= BIT(REG_HI(bpf_reg));
} else if (bpf_reg == BPF_REG_FP) {
usage |= BIT(REG_LO(BPF_REG_FP));
} else {
if (REG_LO(bpf_reg) >= ARC_CALLEE_SAVED_REG_FIRST &&
REG_LO(bpf_reg) <= ARC_CALLEE_SAVED_REG_LAST)
usage |= BIT(REG_LO(bpf_reg));
if (REG_HI(bpf_reg) >= ARC_CALLEE_SAVED_REG_FIRST &&
REG_HI(bpf_reg) <= ARC_CALLEE_SAVED_REG_LAST)
usage |= BIT(REG_HI(bpf_reg));
}
usage |= is_call ? BIT(ARC_R_BLINK) : 0;
return usage;
}
u8 arc_prologue(u8 *buf, u32 usage, u16 frame_size)
{
u8 len = 0;
u32 gp_regs = 0;
if (usage & BIT(ARC_R_BLINK))
len += arc_push_r(BUF(buf, len), ARC_R_BLINK);
gp_regs = usage & ~(BIT(ARC_R_BLINK) | BIT(ARC_R_FP));
while (gp_regs) {
u8 reg = __builtin_ffs(gp_regs) - 1;
len += arc_push_r(BUF(buf, len), reg);
gp_regs &= ~BIT(reg);
}
if ((usage & BIT(ARC_R_FP)) || frame_size > 0)
len += arc_push_r(BUF(buf, len), ARC_R_FP);
if (frame_size > 0)
len += frame_create(BUF(buf, len), frame_size);
#ifdef ARC_BPF_JIT_DEBUG
if ((usage & BIT(ARC_R_FP)) && frame_size == 0) {
pr_err("FP is being saved while there is no frame.");
BUG();
}
#endif
return len;
}
u8 arc_epilogue(u8 *buf, u32 usage, u16 frame_size)
{
u32 len = 0;
u32 gp_regs = 0;
#ifdef ARC_BPF_JIT_DEBUG
if ((usage & BIT(ARC_R_FP)) && frame_size == 0) {
pr_err("FP is being saved while there is no frame.");
BUG();
}
#endif
if (frame_size > 0)
len += frame_restore(BUF(buf, len));
if ((usage & BIT(ARC_R_FP)) || frame_size > 0)
len += arc_pop_r(BUF(buf, len), ARC_R_FP);
gp_regs = usage & ~(BIT(ARC_R_BLINK) | BIT(ARC_R_FP));
while (gp_regs) {
u8 reg = 31 - __builtin_clz(gp_regs);
len += arc_pop_r(BUF(buf, len), reg);
gp_regs &= ~BIT(reg);
}
if (usage & BIT(ARC_R_BLINK))
len += arc_pop_r(BUF(buf, len), ARC_R_BLINK);
len += bpf_to_arc_return(BUF(buf, len));
len += arc_jmp_return(BUF(buf, len));
return len;
}
#define JCC64_NR_OF_JMPS 3
#define JCC64_INSNS_TO_END 3
#define JCC64_SKIP_JMP 1
static const struct {
u8 jit_off[JCC64_NR_OF_JMPS];
struct {
u8 cond[JCC64_NR_OF_JMPS];
} jmp[ARC_CC_SLE + 1];
} arcv2_64_jccs = {
.jit_off = {
INSN_len_normal * 1,
INSN_len_normal * 2,
INSN_len_normal * 4
},
.jmp[ARC_CC_UGT] = {
.cond = {CC_great_u, CC_less_u, CC_great_u}
},
.jmp[ARC_CC_UGE] = {
.cond = {CC_great_u, CC_less_u, CC_great_eq_u}
},
.jmp[ARC_CC_ULT] = {
.cond = {CC_less_u, CC_great_u, CC_less_u}
},
.jmp[ARC_CC_ULE] = {
.cond = {CC_less_u, CC_great_u, CC_less_eq_u}
},
.jmp[ARC_CC_SGT] = {
.cond = {CC_great_s, CC_less_s, CC_great_u}
},
.jmp[ARC_CC_SGE] = {
.cond = {CC_great_s, CC_less_s, CC_great_eq_u}
},
.jmp[ARC_CC_SLT] = {
.cond = {CC_less_s, CC_great_s, CC_less_u}
},
.jmp[ARC_CC_SLE] = {
.cond = {CC_less_s, CC_great_s, CC_less_eq_u}
}
};
static inline s32 get_displacement(u32 curr_off, u32 targ_off)
{
return (s32)(targ_off - (curr_off & ~3L));
}
static inline bool is_valid_far_disp(s32 disp)
{
return (!(disp & 1) && IN_S25_RANGE(disp));
}
static inline bool is_valid_near_disp(s32 disp)
{
return (!(disp & 1) && IN_S21_RANGE(disp));
}
static int gen_j_eq_64(u8 *buf, u8 rd, u8 rs, bool eq,
u32 curr_off, u32 targ_off)
{
s32 disp;
u8 len = 0;
len += arc_cmp_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
len += arc_cmpz_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
disp = get_displacement(curr_off + len, targ_off);
len += arc_bcc(BUF(buf, len), eq ? CC_equal : CC_unequal, disp);
return len;
}
static u8 gen_jset_64(u8 *buf, u8 rd, u8 rs, u32 curr_off, u32 targ_off)
{
u8 len = 0;
s32 disp;
len += arc_tst_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
len += arc_tstz_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
disp = get_displacement(curr_off + len, targ_off);
len += arc_bcc(BUF(buf, len), CC_unequal, disp);
return len;
}
static bool check_jcc_64(u32 curr_off, u32 targ_off, u8 cond)
{
size_t i;
if (cond >= ARC_CC_LAST)
return false;
for (i = 0; i < JCC64_NR_OF_JMPS; i++) {
u32 from, to;
from = curr_off + arcv2_64_jccs.jit_off[i];
if (i != JCC64_SKIP_JMP)
to = targ_off;
else
to = from + (JCC64_INSNS_TO_END * INSN_len_normal);
if (!is_valid_near_disp(get_displacement(from, to)))
return false;
}
return true;
}
bool check_jmp_64(u32 curr_off, u32 targ_off, u8 cond)
{
s32 disp;
switch (cond) {
case ARC_CC_UGT:
case ARC_CC_UGE:
case ARC_CC_ULT:
case ARC_CC_ULE:
case ARC_CC_SGT:
case ARC_CC_SGE:
case ARC_CC_SLT:
case ARC_CC_SLE:
return check_jcc_64(curr_off, targ_off, cond);
case ARC_CC_EQ:
case ARC_CC_NE:
case ARC_CC_SET:
curr_off += 2 * INSN_len_normal;
disp = get_displacement(curr_off, targ_off);
return is_valid_near_disp(disp);
case ARC_CC_AL:
disp = get_displacement(curr_off, targ_off);
return is_valid_far_disp(disp);
default:
return false;
}
}
static u8 gen_jcc_64(u8 *buf, u8 rd, u8 rs, u8 cond,
u32 curr_off, u32 targ_off)
{
s32 disp;
u32 end_off;
const u8 *cc = arcv2_64_jccs.jmp[cond].cond;
u8 len = 0;
len += arc_cmp_r(buf, REG_HI(rd), REG_HI(rs));
disp = get_displacement(curr_off + len, targ_off);
len += arc_bcc(BUF(buf, len), cc[0], disp);
end_off = curr_off + len + (JCC64_INSNS_TO_END * INSN_len_normal);
disp = get_displacement(curr_off + len, end_off);
len += arc_bcc(BUF(buf, len), cc[1], disp);
len += arc_cmp_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
disp = get_displacement(curr_off + len, targ_off);
len += arc_bcc(BUF(buf, len), cc[2], disp);
return len;
}
u8 gen_jmp_64(u8 *buf, u8 rd, u8 rs, u8 cond, u32 curr_off, u32 targ_off)
{
u8 len = 0;
bool eq = false;
s32 disp;
switch (cond) {
case ARC_CC_AL:
disp = get_displacement(curr_off, targ_off);
len = arc_b(buf, disp);
break;
case ARC_CC_UGT:
case ARC_CC_UGE:
case ARC_CC_ULT:
case ARC_CC_ULE:
case ARC_CC_SGT:
case ARC_CC_SGE:
case ARC_CC_SLT:
case ARC_CC_SLE:
len = gen_jcc_64(buf, rd, rs, cond, curr_off, targ_off);
break;
case ARC_CC_EQ:
eq = true;
fallthrough;
case ARC_CC_NE:
len = gen_j_eq_64(buf, rd, rs, eq, curr_off, targ_off);
break;
case ARC_CC_SET:
len = gen_jset_64(buf, rd, rs, curr_off, targ_off);
break;
default:
#ifdef ARC_BPF_JIT_DEBUG
pr_err("64-bit jump condition is not known.");
BUG();
#endif
}
return len;
}
static const u8 arcv2_32_jmps[ARC_CC_LAST] = {
[ARC_CC_UGT] = CC_great_u,
[ARC_CC_UGE] = CC_great_eq_u,
[ARC_CC_ULT] = CC_less_u,
[ARC_CC_ULE] = CC_less_eq_u,
[ARC_CC_SGT] = CC_great_s,
[ARC_CC_SGE] = CC_great_eq_s,
[ARC_CC_SLT] = CC_less_s,
[ARC_CC_SLE] = CC_less_eq_s,
[ARC_CC_AL] = CC_always,
[ARC_CC_EQ] = CC_equal,
[ARC_CC_NE] = CC_unequal,
[ARC_CC_SET] = CC_unequal
};
bool check_jmp_32(u32 curr_off, u32 targ_off, u8 cond)
{
u8 addendum;
s32 disp;
if (cond >= ARC_CC_LAST)
return false;
addendum = (cond == ARC_CC_AL) ? 0 : INSN_len_normal;
disp = get_displacement(curr_off + addendum, targ_off);
if (cond == ARC_CC_AL)
return is_valid_far_disp(disp);
else
return is_valid_near_disp(disp);
}
u8 gen_jmp_32(u8 *buf, u8 rd, u8 rs, u8 cond, u32 curr_off, u32 targ_off)
{
s32 disp;
u8 len = 0;
if (cond >= ARC_CC_LAST) {
#ifdef ARC_BPF_JIT_DEBUG
pr_err("32-bit jump condition is not known.");
BUG();
#endif
return 0;
}
if (cond != ARC_CC_AL) {
if (cond == ARC_CC_SET)
len = tst_r32(buf, rd, rs);
else
len = cmp_r32(buf, rd, rs);
disp = get_displacement(curr_off + len, targ_off);
len += arc_bcc(BUF(buf, len), arcv2_32_jmps[cond], disp);
} else {
disp = get_displacement(curr_off, targ_off);
len = arc_b(buf, disp);
}
return len;
}
u8 gen_func_call(u8 *buf, ARC_ADDR func_addr, bool external_func)
{
u8 len = 0;
if (external_func)
len += push_r64(BUF(buf, len), BPF_REG_5);
len += jump_and_link(BUF(buf, len), func_addr);
if (external_func)
len += arc_add_i(BUF(buf, len), ARC_R_SP, ARC_R_SP, ARG5_SIZE);
return len;
}