BPF_JLT
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
return BPF_JLT;
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JGE: return BPF_JLT;
case BPF_JLT: return BPF_JGE;
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JLT:
return BPF_JLT;
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JLT:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
return BPF_JLT;
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_K: /* (dst_reg < imm) */
case BPF_JMP32 | BPF_JLT | BPF_K: /* ((u32) dst_reg < (u32) imm) */
case BPF_JMP | BPF_JLT | BPF_X: /* (dst < src) */
case BPF_JMP32 | BPF_JLT | BPF_X: /* ((u32) dst < (u32) src) */
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
case BPF_JLT:
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_X:
case BPF_JMP32 | BPF_JLT | BPF_X:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
[BPF_JLT >> 4] = { BR_BLO, false },
[BPF_JMP | BPF_JLT | BPF_K] = cmp_imm,
[BPF_JMP | BPF_JLT | BPF_X] = cmp_reg,
[BPF_JMP32 | BPF_JLT | BPF_K] = cmp_imm,
[BPF_JMP32 | BPF_JLT | BPF_X] = cmp_reg,
case BPF_JLT:
case BPF_JMP | BPF_JLT | BPF_K:
case BPF_JMP32 | BPF_JLT | BPF_K:
[BPF_JLT >> 4] = "<",
_J(BPF_JLT, {.can_jump = true, .can_fallthrough = true}),
case BPF_JLT:
case BPF_JLT:
[BPF_JGT >> 4] = BPF_JLT,
[BPF_JLT >> 4] = BPF_JGT,
case BPF_JLT:
case BPF_JGE: return BPF_JLT;
case BPF_JLT: return BPF_JGE;
case BPF_JLT:
case BPF_JLT:
BPF_JMP_IMM(BPF_JLT, R1, 3, 1),
BPF_JMP_IMM(BPF_JLT, R1, -1, 1),
BPF_JMP_IMM(BPF_JLT, R1, 3, -6), /* goto out */
BPF_JMP_REG(BPF_JLT, R2, R1, 1),
BPF_JMP_REG(BPF_JLT, R2, R1, 1),
BPF_JMP_IMM(BPF_JLT, R1, 0, 1),
BPF_JMP_REG(BPF_JLT, R1, R1, 1),
case BPF_JLT:
return __bpf_fill_jmp_imm(self, BPF_JLT);
return __bpf_fill_jmp32_imm(self, BPF_JLT);
return __bpf_fill_jmp_reg(self, BPF_JLT);
return __bpf_fill_jmp32_reg(self, BPF_JLT);
struct bpf_insn jmp = BPF_JMP_IMM(BPF_JLT, R1, 0x80000000, 0);
struct bpf_insn jmp = BPF_JMP_REG(BPF_JLT, R1, R2, 0);
struct bpf_insn jmp = BPF_JMP32_IMM(BPF_JLT, R1, 0x80000000, 0);
struct bpf_insn jmp = BPF_JMP32_REG(BPF_JLT, R1, R2, 0);
return __bpf_fill_max_jmp(self, BPF_JLT, 0, true);
return __bpf_fill_max_jmp(self, BPF_JLT, 0, false);
BPF_JMP32_IMM(BPF_JLT, R0, 123, 1),
BPF_JMP32_IMM(BPF_JLT, R0, 124, 1),
BPF_JMP32_IMM(BPF_JLT, R0, 0xfffffffd, 1),
BPF_JMP32_IMM(BPF_JLT, R0, 0xffffffff, 1),
BPF_JMP32_REG(BPF_JLT, R0, R1, 2),
BPF_JMP32_REG(BPF_JLT, R0, R1, 1),
insn->code = BPF_JMP | BPF_JLT | bpf_src;
BPF_JMP_IMM(BPF_JLT, BPF_REG_0, 0, 1),
BPF_JMP32_IMM(BPF_JLT, BPF_REG_0, 0, 1),
[BPF_JLT >> 4] = "<",
emit(BPF_JMP32_REG(spec.setup_signed ? BPF_JSLT : BPF_JLT,
emit(BPF_JMP_REG(spec.setup_signed ? BPF_JSLT : BPF_JLT,
emit(BPF_JMP32_REG(spec.setup_signed ? BPF_JSLT : BPF_JLT,
emit(BPF_JMP_REG(spec.setup_signed ? BPF_JSLT : BPF_JLT,
case OP_LT: op_code = spec.compare_signed ? BPF_JSLT : BPF_JLT; break;
BPF_JMP_IMM(BPF_JLT, BPF_REG_1, 10, 1),
BPF_JMP_REG(BPF_JLT, BPF_REG_2, BPF_REG_3, 1), // 8. if r2 < r3 goto 10
BPF_JMP32_IMM(BPF_JLT, BPF_REG_7, INT_MAX, 1),
BPF_JMP32_REG(BPF_JLT, BPF_REG_7, BPF_REG_8, 1),
BPF_JMP32_REG(BPF_JLT, BPF_REG_7, BPF_REG_8, 1),
BPF_JMP_IMM(BPF_JLT, BPF_REG_1, 0x10, 1),
BPF_JMP_IMM(BPF_JLT, BPF_REG_2, 8, 1),
BPF_JMP_IMM(BPF_JLT, BPF_REG_2, 8, 1),