root/tools/testing/selftests/bpf/verifier/atomic_cmpxchg.c
{
        "atomic compare-and-exchange smoketest - 64bit",
        .insns = {
                /* val = 3; */
                BPF_ST_MEM(BPF_DW, BPF_REG_10, -8, 3),
                /* old = atomic_cmpxchg(&val, 2, 4); */
                BPF_MOV64_IMM(BPF_REG_1, 4),
                BPF_MOV64_IMM(BPF_REG_0, 2),
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -8),
                /* if (old != 3) exit(2); */
                BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV64_IMM(BPF_REG_0, 2),
                BPF_EXIT_INSN(),
                /* if (val != 3) exit(3); */
                BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_10, -8),
                BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV64_IMM(BPF_REG_0, 3),
                BPF_EXIT_INSN(),
                /* old = atomic_cmpxchg(&val, 3, 4); */
                BPF_MOV64_IMM(BPF_REG_1, 4),
                BPF_MOV64_IMM(BPF_REG_0, 3),
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -8),
                /* if (old != 3) exit(4); */
                BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV64_IMM(BPF_REG_0, 4),
                BPF_EXIT_INSN(),
                /* if (val != 4) exit(5); */
                BPF_LDX_MEM(BPF_DW, BPF_REG_0, BPF_REG_10, -8),
                BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 4, 2),
                BPF_MOV64_IMM(BPF_REG_0, 5),
                BPF_EXIT_INSN(),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
},
{
        "atomic compare-and-exchange smoketest - 32bit",
        .insns = {
                /* val = 3; */
                BPF_ST_MEM(BPF_W, BPF_REG_10, -4, 3),
                /* old = atomic_cmpxchg(&val, 2, 4); */
                BPF_MOV32_IMM(BPF_REG_1, 4),
                BPF_MOV32_IMM(BPF_REG_0, 2),
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -4),
                /* if (old != 3) exit(2); */
                BPF_JMP32_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV32_IMM(BPF_REG_0, 2),
                BPF_EXIT_INSN(),
                /* if (val != 3) exit(3); */
                BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_10, -4),
                BPF_JMP32_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV32_IMM(BPF_REG_0, 3),
                BPF_EXIT_INSN(),
                /* old = atomic_cmpxchg(&val, 3, 4); */
                BPF_MOV32_IMM(BPF_REG_1, 4),
                BPF_MOV32_IMM(BPF_REG_0, 3),
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -4),
                /* if (old != 3) exit(4); */
                BPF_JMP32_IMM(BPF_JEQ, BPF_REG_0, 3, 2),
                BPF_MOV32_IMM(BPF_REG_0, 4),
                BPF_EXIT_INSN(),
                /* if (val != 4) exit(5); */
                BPF_LDX_MEM(BPF_W, BPF_REG_0, BPF_REG_10, -4),
                BPF_JMP32_IMM(BPF_JEQ, BPF_REG_0, 4, 2),
                BPF_MOV32_IMM(BPF_REG_0, 5),
                BPF_EXIT_INSN(),
                /* exit(0); */
                BPF_MOV32_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
},
{
        "Can't use cmpxchg on uninit src reg",
        .insns = {
                BPF_ST_MEM(BPF_DW, BPF_REG_10, -8, 3),
                BPF_MOV64_IMM(BPF_REG_0, 3),
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_2, -8),
                BPF_EXIT_INSN(),
        },
        .result = REJECT,
        .errstr = "!read_ok",
},
{
        "BPF_W cmpxchg should zero top 32 bits",
        .insns = {
                /* r0 = U64_MAX; */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 1),
                /* u64 val = r0; */
                BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_0, -8),
                /* r0 = (u32)atomic_cmpxchg((u32 *)&val, r0, 1); */
                BPF_MOV32_IMM(BPF_REG_1, 1),
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -8),
                /* r1 = 0x00000000FFFFFFFFull; */
                BPF_MOV64_IMM(BPF_REG_1, 1),
                BPF_ALU64_IMM(BPF_LSH, BPF_REG_1, 32),
                BPF_ALU64_IMM(BPF_SUB, BPF_REG_1, 1),
                /* if (r0 != r1) exit(1); */
                BPF_JMP_REG(BPF_JEQ, BPF_REG_0, BPF_REG_1, 2),
                BPF_MOV32_IMM(BPF_REG_0, 1),
                BPF_EXIT_INSN(),
                /* exit(0); */
                BPF_MOV32_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
},
{
        "Dest pointer in r0 - fail",
        .insns = {
                /* val = 0; */
                BPF_ST_MEM(BPF_DW, BPF_REG_10, -8, 0),
                /* r0 = &val */
                BPF_MOV64_REG(BPF_REG_0, BPF_REG_10),
                /* r0 = atomic_cmpxchg(&val, r0, 1); */
                BPF_MOV64_IMM(BPF_REG_1, 1),
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -8),
                /* if (r0 != 0) exit(1); */
                BPF_JMP_IMM(BPF_JEQ, BPF_REG_0, 0, 2),
                BPF_MOV64_IMM(BPF_REG_0, 1),
                BPF_EXIT_INSN(),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
        .result_unpriv = REJECT,
        .errstr_unpriv = "R0 leaks addr into mem",
},
{
        "Dest pointer in r0 - succeed",
        .insns = {
                /* r0 = &val */
                BPF_MOV64_REG(BPF_REG_0, BPF_REG_10),
                /* val = r0; */
                BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_0, -8),
                /* r0 = atomic_cmpxchg(&val, r0, 0); */
                BPF_MOV64_IMM(BPF_REG_1, 0),
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_1, -8),
                /* r1 = *r0 */
                BPF_LDX_MEM(BPF_DW, BPF_REG_1, BPF_REG_0, -8),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
        .result_unpriv = REJECT,
        .errstr_unpriv = "R0 leaks addr into mem",
},
{
        "Dest pointer in r0 - succeed, check 2",
        .insns = {
                /* r0 = &val */
                BPF_MOV64_REG(BPF_REG_0, BPF_REG_10),
                /* val = r0; */
                BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_0, -8),
                /* r5 = &val */
                BPF_MOV64_REG(BPF_REG_5, BPF_REG_10),
                /* r0 = atomic_cmpxchg(&val, r0, r5); */
                BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_5, -8),
                /* r1 = *r0 */
                BPF_LDX_MEM(BPF_DW, BPF_REG_1, BPF_REG_0, -8),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
        .result_unpriv = REJECT,
        .errstr_unpriv = "R0 leaks addr into mem",
},
{
        "Dest pointer in r0 - succeed, check 3",
        .insns = {
                /* r0 = &val */
                BPF_MOV64_REG(BPF_REG_0, BPF_REG_10),
                /* val = r0; */
                BPF_STX_MEM(BPF_DW, BPF_REG_10, BPF_REG_0, -8),
                /* r5 = &val */
                BPF_MOV64_REG(BPF_REG_5, BPF_REG_10),
                /* r0 = atomic_cmpxchg(&val, r0, r5); */
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_5, -8),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = REJECT,
        .errstr = "invalid size of register fill",
        .errstr_unpriv = "R0 leaks addr into mem",
},
{
        "Dest pointer in r0 - succeed, check 4",
        .insns = {
                /* r0 = &val */
                BPF_MOV32_REG(BPF_REG_0, BPF_REG_10),
                /* val = r0; */
                BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_0, -8),
                /* r5 = &val */
                BPF_MOV32_REG(BPF_REG_5, BPF_REG_10),
                /* r0 = atomic_cmpxchg(&val, r0, r5); */
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_5, -8),
                /* r1 = *r10 */
                BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_10, -8),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = ACCEPT,
        .result_unpriv = REJECT,
        .errstr_unpriv = "R10 partial copy of pointer",
},
{
        "Dest pointer in r0 - succeed, check 5",
        .insns = {
                /* r0 = &val */
                BPF_MOV32_REG(BPF_REG_0, BPF_REG_10),
                /* val = r0; */
                BPF_STX_MEM(BPF_W, BPF_REG_10, BPF_REG_0, -8),
                /* r5 = &val */
                BPF_MOV32_REG(BPF_REG_5, BPF_REG_10),
                /* r0 = atomic_cmpxchg(&val, r0, r5); */
                BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_10, BPF_REG_5, -8),
                /* r1 = *r0 */
                BPF_LDX_MEM(BPF_W, BPF_REG_1, BPF_REG_0, -8),
                /* exit(0); */
                BPF_MOV64_IMM(BPF_REG_0, 0),
                BPF_EXIT_INSN(),
        },
        .result = REJECT,
        .errstr = "R0 invalid mem access",
        .errstr_unpriv = "R10 partial copy of pointer",
        .flags = F_NEEDS_EFFICIENT_UNALIGNED_ACCESS,
},