Symbol: rd
arch/arc/net/bpf_jit.h
100
u8 rsh_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
101
u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
103
u8 arsh_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
104
u8 arsh_r32_i32(u8 *buf, u8 rd, u8 imm);
arch/arc/net/bpf_jit.h
105
u8 arsh_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
106
u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
150
u8 gen_jmp_32(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
arch/arc/net/bpf_jit.h
151
u8 gen_jmp_64(u8 *buf, u8 rd, u8 rs, u8 cond, u32 c_off, u32 t_off);
arch/arc/net/bpf_jit.h
162
u8 gen_swap(u8 *buf, u8 rd, u8 size, u8 endian, bool force, bool do_zext);
arch/arc/net/bpf_jit.h
42
u8 zext(u8 *buf, u8 rd);
arch/arc/net/bpf_jit.h
44
u8 mov_r32(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
arch/arc/net/bpf_jit.h
46
u8 mov_r64(u8 *buf, u8 rd, u8 rs, u8 sign_ext);
arch/arc/net/bpf_jit.h
50
u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext);
arch/arc/net/bpf_jit.h
51
u8 store_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size);
arch/arc/net/bpf_jit.h
52
u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size);
arch/arc/net/bpf_jit.h
54
u8 add_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
55
u8 add_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
56
u8 add_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
57
u8 add_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
59
u8 sub_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
60
u8 sub_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
61
u8 sub_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
62
u8 sub_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
64
u8 mul_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
65
u8 mul_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
66
u8 mul_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
67
u8 mul_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
69
u8 div_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
arch/arc/net/bpf_jit.h
70
u8 div_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
arch/arc/net/bpf_jit.h
72
u8 mod_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext);
arch/arc/net/bpf_jit.h
73
u8 mod_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext);
arch/arc/net/bpf_jit.h
75
u8 and_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
76
u8 and_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
77
u8 and_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
78
u8 and_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
80
u8 or_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
81
u8 or_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
82
u8 or_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
83
u8 or_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
85
u8 xor_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
86
u8 xor_r32_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
87
u8 xor_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
88
u8 xor_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
93
u8 lsh_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
94
u8 lsh_r32_i32(u8 *buf, u8 rd, u8 imm);
arch/arc/net/bpf_jit.h
95
u8 lsh_r64(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
96
u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm);
arch/arc/net/bpf_jit.h
98
u8 rsh_r32(u8 *buf, u8 rd, u8 rs);
arch/arc/net/bpf_jit.h
99
u8 rsh_r32_i32(u8 *buf, u8 rd, u8 imm);
arch/arc/net/bpf_jit_arcv2.c
1004
static u8 arc_divu_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1006
const u32 insn = OPC_DIVU | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1014
static u8 arc_divu_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1016
const u32 insn = OPC_DIVUI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1026
static u8 arc_divs_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1028
const u32 insn = OPC_DIVS | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1036
static u8 arc_divs_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1038
const u32 insn = OPC_DIVSI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1048
static u8 arc_remu_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1050
const u32 insn = OPC_REMU | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1058
static u8 arc_remu_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1060
const u32 insn = OPC_REMUI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1070
static u8 arc_rems_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1072
const u32 insn = OPC_REMS | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1080
static u8 arc_rems_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1082
const u32 insn = OPC_REMSI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1092
static u8 arc_and_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1094
const u32 insn = OPC_AND | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1102
static u8 arc_and_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1104
const u32 insn = OPC_ANDI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1114
static u8 arc_tst_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1116
const u32 insn = OPC_TST | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1129
static u8 arc_tstz_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1131
const u32 insn = OPC_TST | OP_B(rd) | OP_C(rs) | CC_equal;
arch/arc/net/bpf_jit_arcv2.c
1138
static u8 arc_or_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
arch/arc/net/bpf_jit_arcv2.c
1140
const u32 insn = OPC_OR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
arch/arc/net/bpf_jit_arcv2.c
1147
static u8 arc_or_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1149
const u32 insn = OPC_ORI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1158
static u8 arc_xor_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1160
const u32 insn = OPC_XOR | OP_A(rd) | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1167
static u8 arc_xor_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1169
const u32 insn = OPC_XORI | OP_A(rd) | OP_B(rd);
arch/arc/net/bpf_jit_arcv2.c
1178
static u8 arc_not_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1180
const u32 insn = OPC_NOT | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
1196
static u8 arc_asl_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
arch/arc/net/bpf_jit_arcv2.c
1198
const u32 insn = OPC_ASL | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
arch/arc/net/bpf_jit_arcv2.c
1205
static u8 arc_asli_r(u8 *buf, u8 rd, u8 rs, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
1207
const u32 insn = OPC_ASLI | OP_A(rd) | OP_B(rs) | ASLI_U6(imm);
arch/arc/net/bpf_jit_arcv2.c
1214
static u8 arc_asr_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
arch/arc/net/bpf_jit_arcv2.c
1216
const u32 insn = OPC_ASR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
arch/arc/net/bpf_jit_arcv2.c
1223
static u8 arc_asri_r(u8 *buf, u8 rd, u8 rs, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
1225
const u32 insn = OPC_ASRI | OP_A(rd) | OP_B(rs) | ASRI_U6(imm);
arch/arc/net/bpf_jit_arcv2.c
1232
static u8 arc_lsr_r(u8 *buf, u8 rd, u8 rs1, u8 rs2)
arch/arc/net/bpf_jit_arcv2.c
1234
const u32 insn = OPC_LSR | OP_A(rd) | OP_B(rs1) | OP_C(rs2);
arch/arc/net/bpf_jit_arcv2.c
1241
static u8 arc_lsri_r(u8 *buf, u8 rd, u8 rs, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
1243
const u32 insn = OPC_LSRI | OP_A(rd) | OP_B(rs) | LSRI_U6(imm);
arch/arc/net/bpf_jit_arcv2.c
1305
u8 zext(u8 *buf, u8 rd)
arch/arc/net/bpf_jit_arcv2.c
1307
if (rd != BPF_REG_FP)
arch/arc/net/bpf_jit_arcv2.c
1308
return arc_movi_r(buf, REG_HI(rd), 0);
arch/arc/net/bpf_jit_arcv2.c
1313
u8 mov_r32(u8 *buf, u8 rd, u8 rs, u8 sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1319
len = arc_sexb_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1321
len = arc_sexh_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1322
else if (sign_ext == 32 && rd != rs)
arch/arc/net/bpf_jit_arcv2.c
1323
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1330
if (rd != rs)
arch/arc/net/bpf_jit_arcv2.c
1331
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1341
u8 mov_r64(u8 *buf, u8 rd, u8 rs, u8 sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1347
len = mov_r32(buf, rd, rs, sign_ext);
arch/arc/net/bpf_jit_arcv2.c
1352
REG_HI(rd), REG_LO(rd), 31);
arch/arc/net/bpf_jit_arcv2.c
1360
if (rd == rs)
arch/arc/net/bpf_jit_arcv2.c
1363
len = arc_mov_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1366
len += arc_mov_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1369
len += arc_movi_r(BUF(buf, len), REG_HI(rd), 0);
arch/arc/net/bpf_jit_arcv2.c
1460
u8 store_r(u8 *buf, u8 rs, u8 rd, s16 off, u8 size)
arch/arc/net/bpf_jit_arcv2.c
1464
len = adjust_mem_access(buf, &off, size, rd, &arc_reg_mem);
arch/arc/net/bpf_jit_arcv2.c
1491
u8 store_i(u8 *buf, s32 imm, u8 rd, s16 off, u8 size)
arch/arc/net/bpf_jit_arcv2.c
1497
len = adjust_mem_access(buf, &off, size, rd, &arc_reg_mem);
arch/arc/net/bpf_jit_arcv2.c
1540
u8 load_r(u8 *buf, u8 rd, u8 rs, s16 off, u8 size, bool sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1551
len += arc_ldx_r(BUF(buf, len), REG_LO(rd),
arch/arc/net/bpf_jit_arcv2.c
1554
len += arc_ld_r(BUF(buf, len), REG_LO(rd),
arch/arc/net/bpf_jit_arcv2.c
1561
REG_HI(rd), REG_LO(rd), 31);
arch/arc/net/bpf_jit_arcv2.c
1563
len += arc_movi_r(BUF(buf, len), REG_HI(rd), 0);
arch/arc/net/bpf_jit_arcv2.c
1579
if (REG_LO(rd) != arc_reg_mem) {
arch/arc/net/bpf_jit_arcv2.c
1580
len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
arch/arc/net/bpf_jit_arcv2.c
1582
len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
arch/arc/net/bpf_jit_arcv2.c
1585
len += arc_ld_r(BUF(buf, len), REG_HI(rd), arc_reg_mem,
arch/arc/net/bpf_jit_arcv2.c
1587
len += arc_ld_r(BUF(buf, len), REG_LO(rd), arc_reg_mem,
arch/arc/net/bpf_jit_arcv2.c
1595
u8 add_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1597
return arc_add_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1600
u8 add_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1603
return arc_addi_r(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1605
return arc_add_i(buf, REG_LO(rd), REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1608
u8 add_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1612
len = arc_addf_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1613
len += arc_adc_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1617
u8 add_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1622
len = arc_addif_r(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1623
len += arc_adci_r(BUF(buf, len), REG_HI(rd), 0);
arch/arc/net/bpf_jit_arcv2.c
1626
len += add_r64(BUF(buf, len), rd, JIT_REG_TMP);
arch/arc/net/bpf_jit_arcv2.c
1631
u8 sub_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1633
return arc_sub_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1636
u8 sub_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1639
return arc_subi_r(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1641
return arc_sub_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1644
u8 sub_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1648
len = arc_subf_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1649
len += arc_sbc_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1653
u8 sub_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1658
len += sub_r64(BUF(buf, len), rd, JIT_REG_TMP);
arch/arc/net/bpf_jit_arcv2.c
1662
static u8 cmp_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1664
return arc_cmp_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1683
u8 mul_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1685
return arc_mpy_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1688
u8 mul_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1690
return arc_mpy_i(buf, REG_LO(rd), REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1702
u8 mul_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1708
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
1709
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
1748
u8 mul_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1752
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
1753
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
1774
u8 div_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1777
return arc_divs_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1779
return arc_divu_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1782
u8 div_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1788
return arc_divs_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1790
return arc_divu_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1793
u8 mod_r32(u8 *buf, u8 rd, u8 rs, bool sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1796
return arc_rems_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1798
return arc_remu_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1801
u8 mod_r32_i32(u8 *buf, u8 rd, s32 imm, bool sign_ext)
arch/arc/net/bpf_jit_arcv2.c
1807
return arc_rems_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1809
return arc_remu_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1812
u8 and_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1814
return arc_and_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1817
u8 and_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1819
return arc_and_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1822
u8 and_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1826
len = arc_and_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1827
len += arc_and_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1831
u8 and_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1836
len += and_r64(BUF(buf, len), rd, JIT_REG_TMP);
arch/arc/net/bpf_jit_arcv2.c
1840
static u8 tst_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1842
return arc_tst_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1845
u8 or_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1847
return arc_or_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1850
u8 or_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1852
return arc_or_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1855
u8 or_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1859
len = arc_or_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1860
len += arc_or_r(BUF(buf, len), REG_HI(rd), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1864
u8 or_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1869
len += or_r64(BUF(buf, len), rd, JIT_REG_TMP);
arch/arc/net/bpf_jit_arcv2.c
1873
u8 xor_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1875
return arc_xor_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1878
u8 xor_r32_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1880
return arc_xor_i(buf, REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1883
u8 xor_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1887
len = arc_xor_r(buf, REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1888
len += arc_xor_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
1892
u8 xor_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1897
len += xor_r64(BUF(buf, len), rd, JIT_REG_TMP);
arch/arc/net/bpf_jit_arcv2.c
1902
u8 lsh_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1904
return arc_asl_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
1907
u8 lsh_r32_i32(u8 *buf, u8 rd, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
1909
return arc_asli_r(buf, REG_LO(rd), REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
1948
u8 lsh_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
1953
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
1954
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
1981
u8 lsh_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
1984
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
1985
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
2006
u8 rsh_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
2008
return arc_lsr_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
2011
u8 rsh_r32_i32(u8 *buf, u8 rd, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
2013
return arc_lsri_r(buf, REG_LO(rd), REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
2043
u8 rsh_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
2048
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
2049
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
2076
u8 rsh_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
2079
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
2080
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
2101
u8 arsh_r32(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
2103
return arc_asr_r(buf, REG_LO(rd), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
2106
u8 arsh_r32_i32(u8 *buf, u8 rd, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
2108
return arc_asri_r(buf, REG_LO(rd), REG_LO(rd), imm);
arch/arc/net/bpf_jit_arcv2.c
2140
u8 arsh_r64(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
2145
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
2146
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
2174
u8 arsh_r64_i32(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
2177
const u8 B_lo = REG_LO(rd);
arch/arc/net/bpf_jit_arcv2.c
2178
const u8 B_hi = REG_HI(rd);
arch/arc/net/bpf_jit_arcv2.c
2200
u8 gen_swap(u8 *buf, u8 rd, u8 size, u8 endian, bool force, bool do_zext)
arch/arc/net/bpf_jit_arcv2.c
2215
len = arc_asli_r(buf, REG_LO(rd), REG_LO(rd), 16);
arch/arc/net/bpf_jit_arcv2.c
2218
len += arc_swape_r(BUF(buf, len), REG_LO(rd));
arch/arc/net/bpf_jit_arcv2.c
2220
len += zext(BUF(buf, len), rd);
arch/arc/net/bpf_jit_arcv2.c
2230
len = arc_xor_r(buf, REG_HI(rd), REG_LO(rd));
arch/arc/net/bpf_jit_arcv2.c
2231
len += arc_xor_r(BUF(buf, len), REG_LO(rd), REG_HI(rd));
arch/arc/net/bpf_jit_arcv2.c
2232
len += arc_xor_r(BUF(buf, len), REG_HI(rd), REG_LO(rd));
arch/arc/net/bpf_jit_arcv2.c
2233
len += arc_swape_r(BUF(buf, len), REG_LO(rd));
arch/arc/net/bpf_jit_arcv2.c
2234
len += arc_swape_r(BUF(buf, len), REG_HI(rd));
arch/arc/net/bpf_jit_arcv2.c
2247
len = arc_and_i(buf, REG_LO(rd), 0xffff);
arch/arc/net/bpf_jit_arcv2.c
2251
len += zext(BUF(buf, len), rd);
arch/arc/net/bpf_jit_arcv2.c
2676
static int gen_j_eq_64(u8 *buf, u8 rd, u8 rs, bool eq,
arch/arc/net/bpf_jit_arcv2.c
2682
len += arc_cmp_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
2683
len += arc_cmpz_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
2695
static u8 gen_jset_64(u8 *buf, u8 rd, u8 rs, u32 curr_off, u32 targ_off)
arch/arc/net/bpf_jit_arcv2.c
2700
len += arc_tst_r(BUF(buf, len), REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
2701
len += arc_tstz_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
2804
static u8 gen_jcc_64(u8 *buf, u8 rd, u8 rs, u8 cond,
arch/arc/net/bpf_jit_arcv2.c
2813
len += arc_cmp_r(buf, REG_HI(rd), REG_HI(rs));
arch/arc/net/bpf_jit_arcv2.c
2825
len += arc_cmp_r(BUF(buf, len), REG_LO(rd), REG_LO(rs));
arch/arc/net/bpf_jit_arcv2.c
2839
u8 gen_jmp_64(u8 *buf, u8 rd, u8 rs, u8 cond, u32 curr_off, u32 targ_off)
arch/arc/net/bpf_jit_arcv2.c
2858
len = gen_jcc_64(buf, rd, rs, cond, curr_off, targ_off);
arch/arc/net/bpf_jit_arcv2.c
2864
len = gen_j_eq_64(buf, rd, rs, eq, curr_off, targ_off);
arch/arc/net/bpf_jit_arcv2.c
2867
len = gen_jset_64(buf, rd, rs, curr_off, targ_off);
arch/arc/net/bpf_jit_arcv2.c
2939
u8 gen_jmp_32(u8 *buf, u8 rd, u8 rs, u8 cond, u32 curr_off, u32 targ_off)
arch/arc/net/bpf_jit_arcv2.c
2960
len = tst_r32(buf, rd, rs);
arch/arc/net/bpf_jit_arcv2.c
2962
len = cmp_r32(buf, rd, rs);
arch/arc/net/bpf_jit_arcv2.c
669
static u8 arc_mov_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
671
const u32 insn = OPC_MOV | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
679
static u8 arc_mov_i(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
681
const u32 insn = OPC_MOV | OP_B(rd) | OP_IMM;
arch/arc/net/bpf_jit_arcv2.c
684
return arc_movi_r(buf, rd, imm);
arch/arc/net/bpf_jit_arcv2.c
694
static u8 arc_mov_i_fixed(u8 *buf, u8 rd, s32 imm)
arch/arc/net/bpf_jit_arcv2.c
696
const u32 insn = OPC_MOV | OP_B(rd) | OP_IMM;
arch/arc/net/bpf_jit_arcv2.c
706
static u8 arc_mov_cc_r(u8 *buf, u8 cc, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
708
const u32 insn = OPC_MOV_CC | OP_B(rd) | OP_C(rs) | COND(cc);
arch/arc/net/bpf_jit_arcv2.c
716
static u8 arc_movu_cc_r(u8 *buf, u8 cc, u8 rd, u8 imm)
arch/arc/net/bpf_jit_arcv2.c
718
const u32 insn = OPC_MOVU_CC | OP_B(rd) | OP_C(imm) | COND(cc);
arch/arc/net/bpf_jit_arcv2.c
726
static u8 arc_sexb_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
728
const u32 insn = OPC_SEXB | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_arcv2.c
736
static u8 arc_sexh_r(u8 *buf, u8 rd, u8 rs)
arch/arc/net/bpf_jit_arcv2.c
738
const u32 insn = OPC_SEXH | OP_B(rd) | OP_C(rs);
arch/arc/net/bpf_jit_core.c
403
static int handle_swap(u8 *buf, u8 rd, u8 size, u8 endian,
arch/arc/net/bpf_jit_core.c
417
*len = gen_swap(buf, rd, size, endian, force, do_zext);
arch/arc/net/bpf_jit_core.c
589
const u8 rd = insn->dst_reg;
arch/arc/net/bpf_jit_core.c
631
*len += gen_jmp_32(BUF(buf, *len), rd, rs, cond,
arch/arc/net/bpf_jit_core.c
634
*len += gen_jmp_64(BUF(buf, *len), rd, rs, cond,
arch/arm/include/asm/assembler.h
208
.macro badr\c, rd, sym
arch/arm/include/asm/assembler.h
220
.macro get_thread_info, rd
arch/arm/include/asm/assembler.h
293
.macro this_cpu_offset, rd:req
arch/arm/include/asm/assembler.h
331
.macro get_current, rd:req
arch/arm/include/asm/assembler.h
705
.macro ldr_va, rd:req, sym:req, cond, tmp, offset=0
arch/arm/include/asm/assembler.h
725
.macro ldr_this_cpu_armv6, rd:req, sym:req
arch/arm/include/asm/assembler.h
743
.macro ldr_this_cpu, rd:req, sym:req, t1:req, t2:req
arch/arm/include/asm/vfpmacros.h
11
.macro VFPFMRX, rd, sysreg, cond
arch/arm/include/asm/vfpmacros.h
15
.macro VFPFMXR, sysreg, rd, cond
arch/arm/mach-tegra/sleep.h
51
.macro cpu_to_halt_reg rd, rcpu
arch/arm/mach-tegra/sleep.h
60
.macro cpu_to_csr_reg rd, rcpu
arch/arm/mach-tegra/sleep.h
69
.macro cpu_id, rd
arch/arm/mm/alignment.c
343
unsigned int rd = RD_BITS(instr);
arch/arm/mm/alignment.c
358
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
360
put16_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
376
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
379
put16t_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
392
unsigned int rd = RD_BITS(instr);
arch/arm/mm/alignment.c
400
} else if (((rd & 1) == 1) || (rd == 14))
arch/arm/mm/alignment.c
404
rd2 = rd + 1;
arch/arm/mm/alignment.c
415
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
419
put32_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
435
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
439
put32t_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
454
unsigned int rd = RD_BITS(instr);
arch/arm/mm/alignment.c
464
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
466
put32_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
475
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
478
put32t_unaligned_check(regs->uregs[rd], addr);
arch/arm/mm/alignment.c
503
unsigned int rd, rn, correction, nr_regs, regbits;
arch/arm/mm/alignment.c
551
for (regbits = REGMASK_BITS(instr), rd = 0; regbits;
arch/arm/mm/alignment.c
552
regbits >>= 1, rd += 1)
arch/arm/mm/alignment.c
557
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
559
put32t_unaligned_check(regs->uregs[rd], eaddr);
arch/arm/mm/alignment.c
564
for (regbits = REGMASK_BITS(instr), rd = 0; regbits;
arch/arm/mm/alignment.c
565
regbits >>= 1, rd += 1)
arch/arm/mm/alignment.c
570
regs->uregs[rd] = val;
arch/arm/mm/alignment.c
572
put32_unaligned_check(regs->uregs[rd], eaddr);
arch/arm/net/bpf_jit_32.c
1001
emit(ARM_RSBS_I(rd[1], rd[1], 0), ctx);
arch/arm/net/bpf_jit_32.c
1002
emit(ARM_RSC_I(rd[0], rd[0], 0), ctx);
arch/arm/net/bpf_jit_32.c
1004
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1012
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1017
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1022
emit(ARM_MOV_SR(ARM_LR, rd[0], SRTYPE_ASL, rt), ctx);
arch/arm/net/bpf_jit_32.c
1023
emit(ARM_ORR_SR(ARM_LR, ARM_LR, rd[1], SRTYPE_ASL, ARM_IP), ctx);
arch/arm/net/bpf_jit_32.c
1024
emit(ARM_ORR_SR(ARM_IP, ARM_LR, rd[1], SRTYPE_LSR, tmp2[0]), ctx);
arch/arm/net/bpf_jit_32.c
1025
emit(ARM_MOV_SR(ARM_LR, rd[1], SRTYPE_ASL, rt), ctx);
arch/arm/net/bpf_jit_32.c
1036
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1041
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1046
emit(ARM_MOV_SR(ARM_LR, rd[1], SRTYPE_LSR, rt), ctx);
arch/arm/net/bpf_jit_32.c
1047
emit(ARM_ORR_SR(ARM_LR, ARM_LR, rd[0], SRTYPE_ASL, ARM_IP), ctx);
arch/arm/net/bpf_jit_32.c
1049
ARM_ORR_SR(ARM_LR, ARM_LR, rd[0], SRTYPE_ASR, tmp2[0]), ctx);
arch/arm/net/bpf_jit_32.c
1050
emit(ARM_MOV_SR(ARM_IP, rd[0], SRTYPE_ASR, rt), ctx);
arch/arm/net/bpf_jit_32.c
1061
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1066
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1071
emit(ARM_MOV_SR(ARM_LR, rd[1], SRTYPE_LSR, rt), ctx);
arch/arm/net/bpf_jit_32.c
1072
emit(ARM_ORR_SR(ARM_LR, ARM_LR, rd[0], SRTYPE_ASL, ARM_IP), ctx);
arch/arm/net/bpf_jit_32.c
1073
emit(ARM_ORR_SR(ARM_LR, ARM_LR, rd[0], SRTYPE_LSR, tmp2[0]), ctx);
arch/arm/net/bpf_jit_32.c
1074
emit(ARM_MOV_SR(ARM_IP, rd[0], SRTYPE_LSR, rt), ctx);
arch/arm/net/bpf_jit_32.c
1085
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1088
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1092
emit(ARM_MOV_SI(tmp2[0], rd[0], SRTYPE_ASL, val), ctx);
arch/arm/net/bpf_jit_32.c
1093
emit(ARM_ORR_SI(rd[0], tmp2[0], rd[1], SRTYPE_LSR, 32 - val), ctx);
arch/arm/net/bpf_jit_32.c
1094
emit(ARM_MOV_SI(rd[1], rd[1], SRTYPE_ASL, val), ctx);
arch/arm/net/bpf_jit_32.c
1097
emit(ARM_MOV_R(rd[0], rd[1]), ctx);
arch/arm/net/bpf_jit_32.c
1099
emit(ARM_MOV_SI(rd[0], rd[1], SRTYPE_ASL, val - 32), ctx);
arch/arm/net/bpf_jit_32.c
1100
emit(ARM_EOR_R(rd[1], rd[1], rd[1]), ctx);
arch/arm/net/bpf_jit_32.c
1103
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1111
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1114
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1122
emit(ARM_MOV_SI(tmp2[1], rd[1], SRTYPE_LSR, val), ctx);
arch/arm/net/bpf_jit_32.c
1123
emit(ARM_ORR_SI(rd[1], tmp2[1], rd[0], SRTYPE_ASL, 32 - val), ctx);
arch/arm/net/bpf_jit_32.c
1124
emit(ARM_MOV_SI(rd[0], rd[0], SRTYPE_LSR, val), ctx);
arch/arm/net/bpf_jit_32.c
1126
emit(ARM_MOV_R(rd[1], rd[0]), ctx);
arch/arm/net/bpf_jit_32.c
1127
emit(ARM_MOV_I(rd[0], 0), ctx);
arch/arm/net/bpf_jit_32.c
1129
emit(ARM_MOV_SI(rd[1], rd[0], SRTYPE_LSR, val - 32), ctx);
arch/arm/net/bpf_jit_32.c
1130
emit(ARM_MOV_I(rd[0], 0), ctx);
arch/arm/net/bpf_jit_32.c
1133
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1141
const s8 *rd;
arch/arm/net/bpf_jit_32.c
1144
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1152
emit(ARM_MOV_SI(tmp2[1], rd[1], SRTYPE_LSR, val), ctx);
arch/arm/net/bpf_jit_32.c
1153
emit(ARM_ORR_SI(rd[1], tmp2[1], rd[0], SRTYPE_ASL, 32 - val), ctx);
arch/arm/net/bpf_jit_32.c
1154
emit(ARM_MOV_SI(rd[0], rd[0], SRTYPE_ASR, val), ctx);
arch/arm/net/bpf_jit_32.c
1156
emit(ARM_MOV_R(rd[1], rd[0]), ctx);
arch/arm/net/bpf_jit_32.c
1157
emit(ARM_MOV_SI(rd[0], rd[0], SRTYPE_ASR, 31), ctx);
arch/arm/net/bpf_jit_32.c
1159
emit(ARM_MOV_SI(rd[1], rd[0], SRTYPE_ASR, val - 32), ctx);
arch/arm/net/bpf_jit_32.c
1160
emit(ARM_MOV_SI(rd[0], rd[0], SRTYPE_ASR, 31), ctx);
arch/arm/net/bpf_jit_32.c
1163
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1170
const s8 *rd, *rt;
arch/arm/net/bpf_jit_32.c
1173
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1177
emit(ARM_MUL(ARM_IP, rd[1], rt[0]), ctx);
arch/arm/net/bpf_jit_32.c
1178
emit(ARM_MUL(ARM_LR, rd[0], rt[1]), ctx);
arch/arm/net/bpf_jit_32.c
1181
emit(ARM_UMULL(ARM_IP, rd[0], rd[1], rt[1]), ctx);
arch/arm/net/bpf_jit_32.c
1182
emit(ARM_ADD_R(rd[0], ARM_LR, rd[0]), ctx);
arch/arm/net/bpf_jit_32.c
1185
arm_bpf_put_reg32(dst_hi, rd[0], ctx);
arch/arm/net/bpf_jit_32.c
1230
s8 rd;
arch/arm/net/bpf_jit_32.c
1232
rd = arm_bpf_get_reg32(dst, tmp[1], ctx);
arch/arm/net/bpf_jit_32.c
1236
emit(ARM_ADD_R(tmp[0], tmp[0], rd), ctx);
arch/arm/net/bpf_jit_32.c
1237
rd = tmp[0];
arch/arm/net/bpf_jit_32.c
1243
emit(ARM_STRB_I(src_lo, rd, off), ctx);
arch/arm/net/bpf_jit_32.c
1247
emit(ARM_STRH_I(src_lo, rd, off), ctx);
arch/arm/net/bpf_jit_32.c
1251
emit(ARM_STR_I(src_lo, rd, off), ctx);
arch/arm/net/bpf_jit_32.c
1255
emit(ARM_STR_I(src_lo, rd, off), ctx);
arch/arm/net/bpf_jit_32.c
1256
emit(ARM_STR_I(src_hi, rd, off + 4), ctx);
arch/arm/net/bpf_jit_32.c
1265
const s8 *rd = is_stacked(dst_lo) ? tmp : dst;
arch/arm/net/bpf_jit_32.c
1273
} else if (rd[1] == rm) {
arch/arm/net/bpf_jit_32.c
1280
emit(ARM_LDRB_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1282
emit_a32_mov_i(rd[0], 0, ctx);
arch/arm/net/bpf_jit_32.c
1286
emit(ARM_LDRH_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1288
emit_a32_mov_i(rd[0], 0, ctx);
arch/arm/net/bpf_jit_32.c
1292
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1294
emit_a32_mov_i(rd[0], 0, ctx);
arch/arm/net/bpf_jit_32.c
1298
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1299
emit(ARM_LDR_I(rd[0], rm, off + 4), ctx);
arch/arm/net/bpf_jit_32.c
1302
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1309
const s8 *rd = is_stacked(dst_lo) ? tmp : dst;
arch/arm/net/bpf_jit_32.c
1333
emit(ARM_LDRSB_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1337
emit(ARM_LDRSH_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1341
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1345
emit(ARM_ASR_I(rd[0], rd[1], 31), ctx);
arch/arm/net/bpf_jit_32.c
1346
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1350
static inline void emit_ar_r(const u8 rd, const u8 rt, const u8 rm,
arch/arm/net/bpf_jit_32.c
1357
emit(ARM_AND_R(ARM_LR, rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1370
emit(ARM_CMP_R(rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1381
emit(ARM_SBCS_R(ARM_IP, rm, rd), ctx);
arch/arm/net/bpf_jit_32.c
1387
emit(ARM_SBCS_R(ARM_IP, rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1475
static inline void emit_rev16(const u8 rd, const u8 rn, struct jit_ctx *ctx)
arch/arm/net/bpf_jit_32.c
1483
emit(ARM_ORR_SI(rd, tmp2[0], tmp2[1], SRTYPE_LSL, 8), ctx);
arch/arm/net/bpf_jit_32.c
1485
emit(ARM_REV16(rd, rn), ctx);
arch/arm/net/bpf_jit_32.c
1490
static inline void emit_rev32(const u8 rd, const u8 rn, struct jit_ctx *ctx)
arch/arm/net/bpf_jit_32.c
1505
emit(ARM_ORR_R(rd, ARM_IP, tmp2[0]), ctx);
arch/arm/net/bpf_jit_32.c
1508
emit(ARM_REV(rd, rn), ctx);
arch/arm/net/bpf_jit_32.c
1600
const s8 *rd, *rs;
arch/arm/net/bpf_jit_32.c
1717
rd = arm_bpf_get_reg64(dst, tmp2, ctx);
arch/arm/net/bpf_jit_32.c
1727
emit_udivmod64(rd, rd, rs, ctx, BPF_OP(code), off);
arch/arm/net/bpf_jit_32.c
1728
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1808
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1813
emit_rev16(rd[1], rd[1], ctx);
arch/arm/net/bpf_jit_32.c
1816
emit_rev32(rd[1], rd[1], ctx);
arch/arm/net/bpf_jit_32.c
1819
emit_rev32(ARM_LR, rd[1], ctx);
arch/arm/net/bpf_jit_32.c
1820
emit_rev32(rd[1], rd[0], ctx);
arch/arm/net/bpf_jit_32.c
1821
emit(ARM_MOV_R(rd[0], ARM_LR), ctx);
arch/arm/net/bpf_jit_32.c
1831
emit(ARM_AND_R(rd[1], rd[1], tmp2[1]), ctx);
arch/arm/net/bpf_jit_32.c
1833
emit(ARM_UXTH(rd[1], rd[1]), ctx);
arch/arm/net/bpf_jit_32.c
1836
emit(ARM_EOR_R(rd[0], rd[0], rd[0]), ctx);
arch/arm/net/bpf_jit_32.c
1841
emit(ARM_EOR_R(rd[0], rd[0], rd[0]), ctx);
arch/arm/net/bpf_jit_32.c
1848
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
1985
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
1988
emit_ar_r(rd[0], rd[1], rm, rn, ctx, BPF_OP(code),
arch/arm/net/bpf_jit_32.c
469
static inline void emit_mov_i_no8m(const u8 rd, u32 val, struct jit_ctx *ctx)
arch/arm/net/bpf_jit_32.c
472
emit(ARM_LDR_I(rd, ARM_PC, imm_offset(val, ctx)), ctx);
arch/arm/net/bpf_jit_32.c
474
emit(ARM_MOVW(rd, val & 0xffff), ctx);
arch/arm/net/bpf_jit_32.c
476
emit(ARM_MOVT(rd, val >> 16), ctx);
arch/arm/net/bpf_jit_32.c
480
static inline void emit_mov_i(const u8 rd, u32 val, struct jit_ctx *ctx)
arch/arm/net/bpf_jit_32.c
485
emit(ARM_MOV_I(rd, imm12), ctx);
arch/arm/net/bpf_jit_32.c
487
emit_mov_i_no8m(rd, val, ctx);
arch/arm/net/bpf_jit_32.c
520
static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op, u8 sign)
arch/arm/net/bpf_jit_32.c
529
emit(sign ? ARM_SDIV(rd, rm, rn) : ARM_UDIV(rd, rm, rn), ctx);
arch/arm/net/bpf_jit_32.c
532
emit(ARM_MLS(rd, rn, ARM_IP, rm), ctx);
arch/arm/net/bpf_jit_32.c
578
if (rd != ARM_R0)
arch/arm/net/bpf_jit_32.c
579
emit(ARM_MOV_R(rd, ARM_R0), ctx);
arch/arm/net/bpf_jit_32.c
588
static inline void emit_udivmod64(const s8 *rd, const s8 *rm, const s8 *rn, struct jit_ctx *ctx,
arch/arm/net/bpf_jit_32.c
643
if (rd[1] != ARM_R0) {
arch/arm/net/bpf_jit_32.c
644
emit(ARM_MOV_R(rd[0], ARM_R1), ctx);
arch/arm/net/bpf_jit_32.c
645
emit(ARM_MOV_R(rd[1], ARM_R0), ctx);
arch/arm/net/bpf_jit_32.c
649
if (rd[1] != ARM_R0 && rd[1] != ARM_R2) {
arch/arm/net/bpf_jit_32.c
651
} else if (rd[1] != ARM_R0) {
arch/arm/net/bpf_jit_32.c
748
const s8 *rd = is_stacked(dst_lo) ? tmp : dst;
arch/arm/net/bpf_jit_32.c
750
emit_mov_i(rd[1], (u32)val, ctx);
arch/arm/net/bpf_jit_32.c
751
emit_mov_i(rd[0], val >> 32, ctx);
arch/arm/net/bpf_jit_32.c
753
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
848
const s8 *rd;
arch/arm/net/bpf_jit_32.c
850
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.c
857
emit_alu_r(rd[1], rs[1], true, false, op, ctx);
arch/arm/net/bpf_jit_32.c
858
emit_alu_r(rd[0], rs[0], true, true, op, ctx);
arch/arm/net/bpf_jit_32.c
865
emit_alu_r(rd[1], rs, true, false, op, ctx);
arch/arm/net/bpf_jit_32.c
867
emit_a32_mov_i(rd[0], 0, ctx);
arch/arm/net/bpf_jit_32.c
870
arm_bpf_put_reg64(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
916
s8 rd;
arch/arm/net/bpf_jit_32.c
919
rd = tmp[1];
arch/arm/net/bpf_jit_32.c
921
rd = dst_lo;
arch/arm/net/bpf_jit_32.c
922
rs = arm_bpf_get_reg32(src_lo, rd, ctx);
arch/arm/net/bpf_jit_32.c
930
emit(ARM_LSL_I(rd, rs, 32 - off), ctx);
arch/arm/net/bpf_jit_32.c
931
emit(ARM_ASR_I(rd, rd, 32 - off), ctx);
arch/arm/net/bpf_jit_32.c
933
rd = rs;
arch/arm/net/bpf_jit_32.c
948
arm_bpf_put_reg32(dst_lo, rd, ctx);
arch/arm/net/bpf_jit_32.c
956
emit(ARM_ASR_I(tmp[0], rd, 31), ctx);
arch/arm/net/bpf_jit_32.c
959
emit(ARM_ASR_I(dst_hi, rd, 31), ctx);
arch/arm/net/bpf_jit_32.c
968
s8 rd;
arch/arm/net/bpf_jit_32.c
970
rd = arm_bpf_get_reg32(dst, tmp[0], ctx);
arch/arm/net/bpf_jit_32.c
975
emit(ARM_LSL_I(rd, rd, val), ctx);
arch/arm/net/bpf_jit_32.c
978
emit(ARM_LSR_I(rd, rd, val), ctx);
arch/arm/net/bpf_jit_32.c
981
emit(ARM_ASR_I(rd, rd, val), ctx);
arch/arm/net/bpf_jit_32.c
984
emit(ARM_RSB_I(rd, rd, val), ctx);
arch/arm/net/bpf_jit_32.c
988
arm_bpf_put_reg32(dst, rd, ctx);
arch/arm/net/bpf_jit_32.c
995
const s8 *rd;
arch/arm/net/bpf_jit_32.c
998
rd = arm_bpf_get_reg64(dst, tmp, ctx);
arch/arm/net/bpf_jit_32.h
165
#define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm))
arch/arm/net/bpf_jit_32.h
167
#define _AL3_I(op, rd, rn, imm) ((op ## _I) | (rd) << 12 | (rn) << 16 | (imm))
arch/arm/net/bpf_jit_32.h
171
#define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
172
#define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
173
#define ARM_ADD_I(rd, rn, imm) _AL3_I(ARM_INST_ADD, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
174
#define ARM_ADDS_I(rd, rn, imm) _AL3_I(ARM_INST_ADDS, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
175
#define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
176
#define ARM_ADC_I(rd, rn, imm) _AL3_I(ARM_INST_ADC, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
178
#define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
179
#define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
180
#define ARM_AND_I(rd, rn, imm) _AL3_I(ARM_INST_AND, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
182
#define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
183
#define ARM_BIC_I(rd, rn, imm) _AL3_I(ARM_INST_BIC, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
192
#define ARM_EOR_R(rd, rn, rm) _AL3_R(ARM_INST_EOR, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
193
#define ARM_EOR_I(rd, rn, imm) _AL3_I(ARM_INST_EOR, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
212
#define ARM_LSL_R(rd, rn, rm) (_AL3_R(ARM_INST_LSL, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
213
#define ARM_LSL_I(rd, rn, imm) (_AL3_I(ARM_INST_LSL, rd, 0, rn) | (imm) << 7)
arch/arm/net/bpf_jit_32.h
215
#define ARM_LSR_R(rd, rn, rm) (_AL3_R(ARM_INST_LSR, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
216
#define ARM_LSR_I(rd, rn, imm) (_AL3_I(ARM_INST_LSR, rd, 0, rn) | (imm) << 7)
arch/arm/net/bpf_jit_32.h
217
#define ARM_ASR_R(rd, rn, rm) (_AL3_R(ARM_INST_ASR, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
218
#define ARM_ASR_I(rd, rn, imm) (_AL3_I(ARM_INST_ASR, rd, 0, rn) | (imm) << 7)
arch/arm/net/bpf_jit_32.h
220
#define ARM_MOV_R(rd, rm) _AL3_R(ARM_INST_MOV, rd, 0, rm)
arch/arm/net/bpf_jit_32.h
221
#define ARM_MOVS_R(rd, rm) _AL3_R(ARM_INST_MOVS, rd, 0, rm)
arch/arm/net/bpf_jit_32.h
222
#define ARM_MOV_I(rd, imm) _AL3_I(ARM_INST_MOV, rd, 0, imm)
arch/arm/net/bpf_jit_32.h
223
#define ARM_MOV_SR(rd, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
224
(_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
225
#define ARM_MOV_SI(rd, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
226
(ARM_MOV_R(rd, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
228
#define ARM_MOVW(rd, imm) \
arch/arm/net/bpf_jit_32.h
229
(ARM_INST_MOVW | ((imm) >> 12) << 16 | (rd) << 12 | ((imm) & 0x0fff))
arch/arm/net/bpf_jit_32.h
231
#define ARM_MOVT(rd, imm) \
arch/arm/net/bpf_jit_32.h
232
(ARM_INST_MOVT | ((imm) >> 12) << 16 | (rd) << 12 | ((imm) & 0x0fff))
arch/arm/net/bpf_jit_32.h
234
#define ARM_MUL(rd, rm, rn) (ARM_INST_MUL | (rd) << 16 | (rm) << 8 | (rn))
arch/arm/net/bpf_jit_32.h
239
#define ARM_ORR_R(rd, rn, rm) _AL3_R(ARM_INST_ORR, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
240
#define ARM_ORR_I(rd, rn, imm) _AL3_I(ARM_INST_ORR, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
241
#define ARM_ORR_SR(rd, rn, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
242
(_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
243
#define ARM_ORRS_R(rd, rn, rm) _AL3_R(ARM_INST_ORRS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
244
#define ARM_ORRS_SR(rd, rn, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
245
(_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
246
#define ARM_ORR_SI(rd, rn, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
247
(ARM_ORR_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
248
#define ARM_ORRS_SI(rd, rn, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
249
(ARM_ORRS_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
251
#define ARM_REV(rd, rm) (ARM_INST_REV | (rd) << 12 | (rm))
arch/arm/net/bpf_jit_32.h
252
#define ARM_REV16(rd, rm) (ARM_INST_REV16 | (rd) << 12 | (rm))
arch/arm/net/bpf_jit_32.h
254
#define ARM_RSB_I(rd, rn, imm) _AL3_I(ARM_INST_RSB, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
255
#define ARM_RSBS_I(rd, rn, imm) _AL3_I(ARM_INST_RSBS, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
256
#define ARM_RSC_I(rd, rn, imm) _AL3_I(ARM_INST_RSC, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
258
#define ARM_SUB_R(rd, rn, rm) _AL3_R(ARM_INST_SUB, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
259
#define ARM_SUBS_R(rd, rn, rm) _AL3_R(ARM_INST_SUBS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
260
#define ARM_RSB_R(rd, rn, rm) _AL3_R(ARM_INST_RSB, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
261
#define ARM_SBC_R(rd, rn, rm) _AL3_R(ARM_INST_SBC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
262
#define ARM_SBCS_R(rd, rn, rm) _AL3_R(ARM_INST_SBCS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
263
#define ARM_SUB_I(rd, rn, imm) _AL3_I(ARM_INST_SUB, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
264
#define ARM_SUBS_I(rd, rn, imm) _AL3_I(ARM_INST_SUBS, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
265
#define ARM_SBC_I(rd, rn, imm) _AL3_I(ARM_INST_SBC, rd, rn, imm)
arch/arm/net/bpf_jit_32.h
270
#define ARM_UDIV(rd, rn, rm) (ARM_INST_UDIV | (rd) << 16 | (rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
271
#define ARM_SDIV(rd, rn, rm) (ARM_INST_SDIV | (rd) << 16 | (rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
276
#define ARM_MLS(rd, rn, rm, ra) (ARM_INST_MLS | (rd) << 16 | (rn) | (rm) << 8 \
arch/arm/net/bpf_jit_32.h
278
#define ARM_UXTH(rd, rm) (ARM_INST_UXTH | (rd) << 12 | (rm))
arch/arm/probes/decode-arm.c
94
int rd = (insn >> 12) & 0xf;
arch/arm/probes/decode-arm.c
96
regs->uregs[rd] = regs->ARM_cpsr & mask;
arch/arm/probes/kprobes/actions-arm.c
159
int rd = (insn >> 12) & 0xf;
arch/arm/probes/kprobes/actions-arm.c
164
register unsigned long rdv asm("r0") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-arm.c
182
if (rd == 15)
arch/arm/probes/kprobes/actions-arm.c
185
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-arm.c
193
int rd = (insn >> 12) & 0xf;
arch/arm/probes/kprobes/actions-arm.c
197
register unsigned long rdv asm("r0") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-arm.c
212
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-arm.c
221
int rd = (insn >> 16) & 0xf;
arch/arm/probes/kprobes/actions-arm.c
226
register unsigned long rdv asm("r2") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-arm.c
242
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-arm.c
250
int rd = (insn >> 12) & 0xf;
arch/arm/probes/kprobes/actions-arm.c
253
register unsigned long rdv asm("r0") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-arm.c
263
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-thumb.c
214
int rd = (insn >> 8) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
218
register unsigned long rdv asm("r1") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-thumb.c
233
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-thumb.c
242
int rd = (insn >> 8) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
244
register unsigned long rdv asm("r1") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-thumb.c
254
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-thumb.c
261
int rd = (insn >> 8) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
264
register unsigned long rdv asm("r1") = regs->uregs[rd];
arch/arm/probes/kprobes/actions-thumb.c
274
regs->uregs[rd] = rdv;
arch/arm/probes/kprobes/actions-thumb.c
47
int rd = (insn >> 8) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
49
regs->uregs[rd] = regs->ARM_cpsr & mask;
arch/arm64/include/asm/assembler.h
160
.macro regs_to_64, rd, lbits, hbits
arch/arm64/include/asm/assembler.h
162
.macro regs_to_64, rd, hbits, lbits
arch/arm64/include/asm/assembler.h
580
.macro get_current_task, rd
arch/arm64/kernel/compat_alignment.c
117
unsigned int rd, rn, nr_regs, regbits;
arch/arm64/kernel/compat_alignment.c
136
for (regbits = REGMASK_BITS(instr), rd = 0; regbits;
arch/arm64/kernel/compat_alignment.c
137
regbits >>= 1, rd += 1)
arch/arm64/kernel/compat_alignment.c
142
if (rd < 15)
arch/arm64/kernel/compat_alignment.c
143
regs->regs[rd] = val;
arch/arm64/kernel/compat_alignment.c
154
val = (rd < 15) ? regs->regs[rd] : regs->pc + 8;
arch/arm64/kernel/compat_alignment.c
70
unsigned int rd = RD_BITS(instr);
arch/arm64/kernel/compat_alignment.c
78
} else if (((rd & 1) == 1) || (rd == 14)) {
arch/arm64/kernel/compat_alignment.c
82
rd2 = rd + 1;
arch/arm64/kernel/compat_alignment.c
91
regs->regs[rd] = val;
arch/arm64/kernel/compat_alignment.c
94
if (put_user(regs->regs[rd], (u32 __user *)addr) ||
arch/arm64/kernel/module-plts.c
109
int rd;
arch/arm64/kernel/module-plts.c
118
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD,
arch/arm64/kernel/module-plts.c
124
plt[i] = __get_adrp_add_pair(val, (u64)&plt[i], rd);
arch/arm64/kernel/proton-pack.c
1124
u8 rd;
arch/arm64/kernel/proton-pack.c
1133
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
arch/arm64/kernel/proton-pack.c
1134
insn = aarch64_insn_gen_movewide(rd, max_bhb_k, 0,
arch/arm64/kernel/proton-pack.c
1144
u8 rd;
arch/arm64/kernel/proton-pack.c
1154
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, insn);
arch/arm64/kernel/proton-pack.c
1158
AARCH64_INSN_REG_ZR, rd,
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
43
int rd;
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
70
rd = kvm_vcpu_dabt_get_rd(vcpu);
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
75
u32 data = vcpu_get_reg(vcpu, rd);
arch/arm64/kvm/hyp/vgic-v2-cpuif-proxy.c
87
vcpu_set_reg(vcpu, rd, data);
arch/arm64/kvm/sys_regs.c
1430
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
1613
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
1616
switch (reg_to_encoding(rd)) {
arch/arm64/kvm/sys_regs.c
1633
__vcpu_assign_sys_reg(vcpu, rd->reg, val);
arch/arm64/kvm/sys_regs.c
1638
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
1641
switch (reg_to_encoding(rd)) {
arch/arm64/kvm/sys_regs.c
1649
*val = __vcpu_sys_reg(vcpu, rd->reg);
arch/arm64/kvm/sys_regs.c
1694
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
1699
u32 id = reg_to_encoding(rd);
arch/arm64/kvm/sys_regs.c
1700
u64 writable_mask = rd->val;
arch/arm64/kvm/sys_regs.c
1701
u64 limit = rd->reset(vcpu, rd);
arch/arm64/kvm/sys_regs.c
1709
if (sysreg_visible_as_raz(vcpu, rd))
arch/arm64/kvm/sys_regs.c
1940
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
1949
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
1958
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2072
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2107
return set_id_reg(vcpu, rd, val);
arch/arm64/kvm/sys_regs.c
2111
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2128
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2151
return set_id_reg(vcpu, rd, val);
arch/arm64/kvm/sys_regs.c
2155
const struct sys_reg_desc *rd, u64 user_val)
arch/arm64/kvm/sys_regs.c
2188
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2192
const struct sys_reg_desc *rd, u64 user_val)
arch/arm64/kvm/sys_regs.c
2224
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2240
const struct sys_reg_desc *rd, u64 user_val)
arch/arm64/kvm/sys_regs.c
2242
u64 sanitized_val = kvm_read_sanitised_id_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
2245
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2252
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2256
const struct sys_reg_desc *rd, u64 user_val)
arch/arm64/kvm/sys_regs.c
2268
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2272
const struct sys_reg_desc *rd, u64 user_val)
arch/arm64/kvm/sys_regs.c
2293
return set_id_reg(vcpu, rd, user_val);
arch/arm64/kvm/sys_regs.c
2306
static int get_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2314
*val = read_id_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
2319
*val = read_id_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
2325
static int set_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2328
u32 id = reg_to_encoding(rd);
arch/arm64/kvm/sys_regs.c
2338
if (val != read_id_reg(vcpu, rd))
arch/arm64/kvm/sys_regs.c
2347
ret = arm64_check_features(vcpu, rd, val);
arch/arm64/kvm/sys_regs.c
2377
static int get_raz_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2384
static int set_wi_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2466
static int set_clidr(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2475
__vcpu_assign_sys_reg(vcpu, rd->reg, val);
arch/arm64/kvm/sys_regs.c
2509
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2526
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2725
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
2729
return el2_visibility(vcpu, rd) ?: fn(vcpu, rd);
arch/arm64/kvm/sys_regs.c
2733
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2735
return __el2_visibility(vcpu, rd, sve_visibility);
arch/arm64/kvm/sys_regs.c
2739
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2741
if (el2_visibility(vcpu, rd) == 0 &&
arch/arm64/kvm/sys_regs.c
2749
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2758
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2760
return __el2_visibility(vcpu, rd, sctlr2_visibility);
arch/arm64/kvm/sys_regs.c
2834
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2843
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2845
return __el2_visibility(vcpu, rd, s1poe_visibility);
arch/arm64/kvm/sys_regs.c
2849
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2858
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2860
return __el2_visibility(vcpu, rd, tcr2_visibility);
arch/arm64/kvm/sys_regs.c
2864
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2866
if (el2_visibility(vcpu, rd) == 0 &&
arch/arm64/kvm/sys_regs.c
2874
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2876
if (el2_visibility(vcpu, rd) == 0 &&
arch/arm64/kvm/sys_regs.c
2884
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2893
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
2895
return __el2_visibility(vcpu, rd, s1pie_visibility);
arch/arm64/kvm/sys_regs.c
2899
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
49
static int set_id_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
5492
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
5500
if (!(rd->reg || rd->get_user))
arch/arm64/kvm/sys_regs.c
5503
if (sysreg_hidden(vcpu, rd))
arch/arm64/kvm/sys_regs.c
5506
if (!copy_reg_to_user(rd, uind))
arch/arm64/kvm/sys_regs.c
737
static int set_oslsr_el1(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
744
if ((val ^ rd->val) & ~OSLSR_EL1_OSLK)
arch/arm64/kvm/sys_regs.c
747
__vcpu_assign_sys_reg(vcpu, rd->reg, val);
arch/arm64/kvm/sys_regs.c
781
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
786
get_access_mask(rd, &mask, &shift);
arch/arm64/kvm/sys_regs.c
796
const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
801
get_access_mask(rd, &mask, &shift);
arch/arm64/kvm/sys_regs.c
805
static u64 *demux_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
809
switch (rd->Op2) {
arch/arm64/kvm/sys_regs.c
811
return &dbg->dbg_bvr[rd->CRm];
arch/arm64/kvm/sys_regs.c
813
return &dbg->dbg_bcr[rd->CRm];
arch/arm64/kvm/sys_regs.c
815
return &dbg->dbg_wvr[rd->CRm];
arch/arm64/kvm/sys_regs.c
817
return &dbg->dbg_wcr[rd->CRm];
arch/arm64/kvm/sys_regs.c
825
const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
827
u64 *reg = demux_wb_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
833
reg_to_dbg(vcpu, p, rd, reg);
arch/arm64/kvm/sys_regs.c
835
dbg_to_reg(vcpu, p, rd, reg);
arch/arm64/kvm/sys_regs.c
841
static int set_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
844
u64 *reg = demux_wb_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
853
static int get_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.c
856
u64 *reg = demux_wb_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
865
static u64 reset_dbg_wb_reg(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd)
arch/arm64/kvm/sys_regs.c
867
u64 *reg = demux_wb_reg(vcpu, rd);
arch/arm64/kvm/sys_regs.c
877
*reg = rd->val;
arch/arm64/kvm/sys_regs.c
878
return rd->val;
arch/arm64/kvm/sys_regs.h
104
const struct sys_reg_desc *rd);
arch/arm64/kvm/sys_regs.h
97
int (*get_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/sys_regs.h
99
int (*set_user)(struct kvm_vcpu *vcpu, const struct sys_reg_desc *rd,
arch/arm64/kvm/va_layout.c
135
static u32 compute_instruction(int n, u32 rd, u32 rn)
arch/arm64/kvm/va_layout.c
143
rn, rd, va_mask);
arch/arm64/kvm/va_layout.c
149
rn, rn, rd,
arch/arm64/kvm/va_layout.c
154
insn = aarch64_insn_gen_add_sub_imm(rd, rn,
arch/arm64/kvm/va_layout.c
161
insn = aarch64_insn_gen_add_sub_imm(rd, rn,
arch/arm64/kvm/va_layout.c
170
rn, rn, rd, 64 - tag_lsb);
arch/arm64/kvm/va_layout.c
185
u32 rd, rn, insn, oinsn;
arch/arm64/kvm/va_layout.c
201
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, oinsn);
arch/arm64/kvm/va_layout.c
204
insn = compute_instruction(i, rd, rn);
arch/arm64/kvm/va_layout.c
269
u32 insn, oinsn, rd;
arch/arm64/kvm/va_layout.c
275
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, oinsn);
arch/arm64/kvm/va_layout.c
278
insn = aarch64_insn_gen_movewide(rd,
arch/arm64/kvm/va_layout.c
286
insn = aarch64_insn_gen_movewide(rd,
arch/arm64/kvm/va_layout.c
294
insn = aarch64_insn_gen_movewide(rd,
arch/arm64/kvm/va_layout.c
302
insn = aarch64_insn_gen_movewide(rd,
arch/arm64/kvm/vgic-sys-reg-v3.c
367
const struct sys_reg_desc *rd)
arch/arm64/kvm/vgic/vgic-its.c
1741
#define REGISTER_ITS_DESC(off, rd, wr, length, acc) \
arch/arm64/kvm/vgic/vgic-its.c
1746
.its_read = rd, \
arch/arm64/kvm/vgic/vgic-its.c
1750
#define REGISTER_ITS_DESC_UACCESS(off, rd, wr, uwr, length, acc)\
arch/arm64/kvm/vgic/vgic-its.c
1755
.its_read = rd, \
arch/arm64/kvm/vgic/vgic-mmio-v3.c
607
#define REGISTER_DESC_WITH_BITS_PER_IRQ_SHARED(off, rd, wr, ur, uw, bpi, acc) \
arch/arm64/kvm/vgic/vgic-mmio-v3.c
620
.read = rd, \
arch/arm64/kvm/vgic/vgic-mmio.h
67
#define REGISTER_DESC_WITH_BITS_PER_IRQ(off, rd, wr, ur, uw, bpi, acc) \
arch/arm64/kvm/vgic/vgic-mmio.h
73
.read = rd, \
arch/arm64/kvm/vgic/vgic-mmio.h
79
#define REGISTER_DESC_WITH_LENGTH(off, rd, wr, length, acc) \
arch/arm64/kvm/vgic/vgic-mmio.h
85
.read = rd, \
arch/arm64/kvm/vgic/vgic-mmio.h
89
#define REGISTER_DESC_WITH_LENGTH_UACCESS(off, rd, wr, urd, uwr, length, acc) \
arch/arm64/kvm/vgic/vgic-mmio.h
95
.read = rd, \
arch/arm64/kvm/vgic/vgic-v3.c
839
u32 insn, oinsn, rd;
arch/arm64/kvm/vgic/vgic-v3.c
868
rd = aarch64_insn_decode_register(AARCH64_INSN_REGTYPE_RD, oinsn);
arch/arm64/kvm/vgic/vgic-v3.c
871
insn = aarch64_insn_gen_movewide(rd,
arch/loongarch/include/asm/inst.h
277
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
290
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
296
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
303
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
310
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
317
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
324
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
331
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
339
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
346
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
353
unsigned int rd : 5;
arch/loongarch/include/asm/inst.h
446
ip->reg2i12_format.rd == LOONGARCH_GPR_RA &&
arch/loongarch/include/asm/inst.h
457
ip->reg2i12_format.rd == LOONGARCH_GPR_SP &&
arch/loongarch/include/asm/inst.h
517
u32 larch_insn_gen_or(enum loongarch_gpr rd, enum loongarch_gpr rj, enum loongarch_gpr rk);
arch/loongarch/include/asm/inst.h
518
u32 larch_insn_gen_move(enum loongarch_gpr rd, enum loongarch_gpr rj);
arch/loongarch/include/asm/inst.h
520
u32 larch_insn_gen_lu12iw(enum loongarch_gpr rd, int imm);
arch/loongarch/include/asm/inst.h
521
u32 larch_insn_gen_lu32id(enum loongarch_gpr rd, int imm);
arch/loongarch/include/asm/inst.h
522
u32 larch_insn_gen_lu52id(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm);
arch/loongarch/include/asm/inst.h
523
u32 larch_insn_gen_beq(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm);
arch/loongarch/include/asm/inst.h
524
u32 larch_insn_gen_bne(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm);
arch/loongarch/include/asm/inst.h
525
u32 larch_insn_gen_jirl(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm);
arch/loongarch/include/asm/inst.h
570
enum loongarch_gpr rd, int imm) \
arch/loongarch/include/asm/inst.h
574
insn->reg1i20_format.rd = rd; \
arch/loongarch/include/asm/inst.h
583
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
587
insn->reg2_format.rd = rd; \
arch/loongarch/include/asm/inst.h
599
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
605
insn->reg2i5_format.rd = rd; \
arch/loongarch/include/asm/inst.h
615
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
621
insn->reg2i6_format.rd = rd; \
arch/loongarch/include/asm/inst.h
631
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
637
insn->reg2i12_format.rd = rd; \
arch/loongarch/include/asm/inst.h
661
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
667
insn->reg2i14_format.rd = rd; \
arch/loongarch/include/asm/inst.h
683
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
689
insn->reg2i16_format.rd = rd; \
arch/loongarch/include/asm/inst.h
700
enum loongarch_gpr rd,
arch/loongarch/include/asm/inst.h
706
insn->reg2i16_format.rd = rd;
arch/loongarch/include/asm/inst.h
712
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
721
insn->reg2bstrd_format.rd = rd; \
arch/loongarch/include/asm/inst.h
728
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
733
insn->reg3_format.rd = rd; \
arch/loongarch/include/asm/inst.h
779
enum loongarch_gpr rd, \
arch/loongarch/include/asm/inst.h
786
insn->reg3sa2_format.rd = rd; \
arch/loongarch/kernel/inst.c
105
if ((long)regs->regs[rj] < (long)regs->regs[rd])
arch/loongarch/kernel/inst.c
111
if ((long)regs->regs[rj] >= (long)regs->regs[rd])
arch/loongarch/kernel/inst.c
117
if (regs->regs[rj] < regs->regs[rd])
arch/loongarch/kernel/inst.c
123
if (regs->regs[rj] >= regs->regs[rd])
arch/loongarch/kernel/inst.c
130
regs->regs[rd] = pc + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/inst.c
18
unsigned int rd = insn.reg1i20_format.rd;
arch/loongarch/kernel/inst.c
28
regs->regs[rd] = pc + sign_extend64(imm << 2, 21);
arch/loongarch/kernel/inst.c
31
regs->regs[rd] = pc + sign_extend64(imm << 12, 31);
arch/loongarch/kernel/inst.c
34
regs->regs[rd] = pc + sign_extend64(imm << 18, 37);
arch/loongarch/kernel/inst.c
347
u32 larch_insn_gen_or(enum loongarch_gpr rd, enum loongarch_gpr rj, enum loongarch_gpr rk)
arch/loongarch/kernel/inst.c
351
emit_or(&insn, rd, rj, rk);
arch/loongarch/kernel/inst.c
356
u32 larch_insn_gen_move(enum loongarch_gpr rd, enum loongarch_gpr rj)
arch/loongarch/kernel/inst.c
358
return larch_insn_gen_or(rd, rj, 0);
arch/loongarch/kernel/inst.c
361
u32 larch_insn_gen_lu12iw(enum loongarch_gpr rd, int imm)
arch/loongarch/kernel/inst.c
37
regs->regs[rd] = pc + sign_extend64(imm << 12, 31);
arch/loongarch/kernel/inst.c
370
emit_lu12iw(&insn, rd, imm);
arch/loongarch/kernel/inst.c
375
u32 larch_insn_gen_lu32id(enum loongarch_gpr rd, int imm)
arch/loongarch/kernel/inst.c
38
regs->regs[rd] &= ~((1 << 12) - 1);
arch/loongarch/kernel/inst.c
384
emit_lu32id(&insn, rd, imm);
arch/loongarch/kernel/inst.c
389
u32 larch_insn_gen_lu52id(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm)
arch/loongarch/kernel/inst.c
398
emit_lu52id(&insn, rd, rj, imm);
arch/loongarch/kernel/inst.c
403
u32 larch_insn_gen_beq(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm)
arch/loongarch/kernel/inst.c
412
emit_beq(&insn, rj, rd, imm >> 2);
arch/loongarch/kernel/inst.c
417
u32 larch_insn_gen_bne(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm)
arch/loongarch/kernel/inst.c
426
emit_bne(&insn, rj, rd, imm >> 2);
arch/loongarch/kernel/inst.c
431
u32 larch_insn_gen_jirl(enum loongarch_gpr rd, enum loongarch_gpr rj, int imm)
arch/loongarch/kernel/inst.c
440
emit_jirl(&insn, rd, rj, imm >> 2);
arch/loongarch/kernel/inst.c
50
unsigned int imm, imm_l, imm_h, rd, rj;
arch/loongarch/kernel/inst.c
90
rd = insn.reg2i16_format.rd;
arch/loongarch/kernel/inst.c
93
if (regs->regs[rj] == regs->regs[rd])
arch/loongarch/kernel/inst.c
99
if (regs->regs[rj] != regs->regs[rd])
arch/loongarch/kernel/kgdb.c
298
unsigned int si, si_l, si_h, rd, rj, cj;
arch/loongarch/kernel/kgdb.c
344
rd = ip->reg2i16_format.rd;
arch/loongarch/kernel/kgdb.c
348
if (regs->regs[rj] == regs->regs[rd])
arch/loongarch/kernel/kgdb.c
352
if (regs->regs[rj] != regs->regs[rd])
arch/loongarch/kernel/kgdb.c
356
if ((long)regs->regs[rj] < (long)regs->regs[rd])
arch/loongarch/kernel/kgdb.c
360
if ((long)regs->regs[rj] >= (long)regs->regs[rd])
arch/loongarch/kernel/kgdb.c
364
if (regs->regs[rj] < regs->regs[rd])
arch/loongarch/kernel/kgdb.c
368
if (regs->regs[rj] >= regs->regs[rd])
arch/loongarch/kernel/kgdb.c
372
regs->regs[rd] = pc + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/traps.c
654
if (insn.reg3_format.rd != 0)
arch/loongarch/kernel/traps.c
661
if (insn.reg3_format.rd != 0)
arch/loongarch/kernel/unaligned.c
449
regs->regs[insn.reg3_format.rd] = value;
arch/loongarch/kernel/unaligned.c
452
write_fpr(insn.reg3_format.rd, value);
arch/loongarch/kernel/unaligned.c
454
set_fpr64(&current->thread.fpu.fpr[insn.reg3_format.rd], 0, value);
arch/loongarch/kernel/unaligned.c
459
value = regs->regs[insn.reg3_format.rd];
arch/loongarch/kernel/unaligned.c
462
value = read_fpr(insn.reg3_format.rd);
arch/loongarch/kernel/unaligned.c
464
value = get_fpr64(&current->thread.fpu.fpr[insn.reg3_format.rd], 0);
arch/loongarch/kvm/exit.c
115
unsigned int rd, rj, csrid;
arch/loongarch/kvm/exit.c
124
rd = inst.reg2csr_format.rd;
arch/loongarch/kvm/exit.c
140
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
143
val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
145
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
148
val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
151
vcpu->arch.gprs[rd] = val;
arch/loongarch/kvm/exit.c
161
u32 addr, rd, rj, opcode;
arch/loongarch/kvm/exit.c
166
rd = inst.reg2_format.rd;
arch/loongarch/kvm/exit.c
172
val = &vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
221
vcpu->arch.io_gpr = rd; /* Set register id for iocsr read completion */
arch/loongarch/kvm/exit.c
26
int rd, rj;
arch/loongarch/kvm/exit.c
32
rd = inst.reg2_format.rd;
arch/loongarch/kvm/exit.c
372
unsigned int op8, opcode, rd;
arch/loongarch/kvm/exit.c
382
rd = inst.reg2i14_format.rd;
arch/loongarch/kvm/exit.c
397
rd = inst.reg2i12_format.rd;
arch/loongarch/kvm/exit.c
431
rd = inst.reg3_format.rd;
arch/loongarch/kvm/exit.c
46
vcpu->arch.gprs[rd] = vcpu->arch.cpucfg[index];
arch/loongarch/kvm/exit.c
471
vcpu->arch.io_gpr = rd; /* Set for kvm_complete_mmio_read() use */
arch/loongarch/kvm/exit.c
50
vcpu->arch.gprs[rd] = *(unsigned int *)KVM_SIGNATURE;
arch/loongarch/kvm/exit.c
54
vcpu->arch.gprs[rd] = ret;
arch/loongarch/kvm/exit.c
547
unsigned int rd, op8, opcode;
arch/loongarch/kvm/exit.c
564
rd = inst.reg2i14_format.rd;
arch/loongarch/kvm/exit.c
57
vcpu->arch.gprs[rd] = 0;
arch/loongarch/kvm/exit.c
570
*(unsigned int *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
574
*(unsigned long *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
582
rd = inst.reg2i12_format.rd;
arch/loongarch/kvm/exit.c
584
rd_val = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
609
rd = inst.reg3_format.rd;
arch/loongarch/kvm/exit.c
615
*(unsigned char *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
619
*(unsigned short *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
623
*(unsigned int *)data = vcpu->arch.gprs[rd];
arch/loongarch/kvm/exit.c
627
*(unsigned long *)data = vcpu->arch.gprs[rd];
arch/loongarch/net/bpf_jit.c
1343
static int emit_jump_and_link(struct jit_ctx *ctx, u8 rd, u64 target)
arch/loongarch/net/bpf_jit.c
1351
emit_insn(ctx, jirl, rd, LOONGARCH_GPR_T1, 0);
arch/loongarch/net/bpf_jit.c
1585
static void sign_extend(struct jit_ctx *ctx, int rd, int rj, u8 size, bool sign)
arch/loongarch/net/bpf_jit.c
1589
if (rd != rj)
arch/loongarch/net/bpf_jit.c
1590
move_reg(ctx, rd, rj);
arch/loongarch/net/bpf_jit.c
1596
emit_insn(ctx, extwb, rd, rj);
arch/loongarch/net/bpf_jit.c
1599
emit_insn(ctx, extwh, rd, rj);
arch/loongarch/net/bpf_jit.c
1602
emit_insn(ctx, addiw, rd, rj, 0);
arch/loongarch/net/bpf_jit.c
1605
if (rd != rj)
arch/loongarch/net/bpf_jit.c
1606
move_reg(ctx, rd, rj);
arch/loongarch/net/bpf_jit.h
119
static inline void move_addr(struct jit_ctx *ctx, enum loongarch_gpr rd, u64 addr)
arch/loongarch/net/bpf_jit.h
125
emit_insn(ctx, lu12iw, rd, imm_31_12);
arch/loongarch/net/bpf_jit.h
129
emit_insn(ctx, ori, rd, rd, imm_11_0);
arch/loongarch/net/bpf_jit.h
133
emit_insn(ctx, lu32id, rd, imm_51_32);
arch/loongarch/net/bpf_jit.h
137
emit_insn(ctx, lu52id, rd, rd, imm_63_52);
arch/loongarch/net/bpf_jit.h
140
static inline void move_imm(struct jit_ctx *ctx, enum loongarch_gpr rd, long imm, bool is32)
arch/loongarch/net/bpf_jit.h
146
emit_insn(ctx, or, rd, LOONGARCH_GPR_ZERO, LOONGARCH_GPR_ZERO);
arch/loongarch/net/bpf_jit.h
152
emit_insn(ctx, addiw, rd, LOONGARCH_GPR_ZERO, imm);
arch/loongarch/net/bpf_jit.h
158
emit_insn(ctx, ori, rd, LOONGARCH_GPR_ZERO, imm);
arch/loongarch/net/bpf_jit.h
166
emit_insn(ctx, lu52id, rd, LOONGARCH_GPR_ZERO, imm_63_52);
arch/loongarch/net/bpf_jit.h
172
emit_insn(ctx, lu12iw, rd, imm_31_12);
arch/loongarch/net/bpf_jit.h
177
emit_insn(ctx, ori, rd, rd, imm_11_0);
arch/loongarch/net/bpf_jit.h
190
emit_insn(ctx, lu32id, rd, imm_51_32);
arch/loongarch/net/bpf_jit.h
196
emit_insn(ctx, lu52id, rd, rd, imm_63_52);
arch/loongarch/net/bpf_jit.h
200
emit_zext_32(ctx, rd, is32);
arch/loongarch/net/bpf_jit.h
203
static inline void move_reg(struct jit_ctx *ctx, enum loongarch_gpr rd,
arch/loongarch/net/bpf_jit.h
206
emit_insn(ctx, or, rd, rj, LOONGARCH_GPR_ZERO);
arch/loongarch/net/bpf_jit.h
238
enum loongarch_gpr rd, int jmp_offset)
arch/loongarch/net/bpf_jit.h
243
emit_insn(ctx, beq, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
248
emit_insn(ctx, bne, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
252
emit_insn(ctx, bltu, rd, rj, jmp_offset);
arch/loongarch/net/bpf_jit.h
256
emit_insn(ctx, bltu, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
260
emit_insn(ctx, bgeu, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
264
emit_insn(ctx, bgeu, rd, rj, jmp_offset);
arch/loongarch/net/bpf_jit.h
268
emit_insn(ctx, blt, rd, rj, jmp_offset);
arch/loongarch/net/bpf_jit.h
272
emit_insn(ctx, blt, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
276
emit_insn(ctx, bge, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
280
emit_insn(ctx, bge, rd, rj, jmp_offset);
arch/loongarch/net/bpf_jit.h
286
enum loongarch_gpr rd, int jmp_offset)
arch/loongarch/net/bpf_jit.h
289
cond_jmp_offset(ctx, cond, rj, rd, 2);
arch/loongarch/net/bpf_jit.h
299
enum loongarch_gpr rd, int jmp_offset)
arch/loongarch/net/bpf_jit.h
313
cond_jmp_offs26(ctx, cond, rj, rd, jmp_offset);
arch/loongarch/net/bpf_jit.h
331
enum loongarch_gpr rd, int jmp_offset)
arch/loongarch/net/bpf_jit.h
334
cond_jmp_offset(ctx, cond, rj, rd, jmp_offset);
arch/microblaze/include/asm/page.h
110
#define tophys(rd, rs) \
arch/microblaze/include/asm/page.h
111
addik rd, rs, (CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START)
arch/microblaze/include/asm/page.h
112
#define tovirt(rd, rs) \
arch/microblaze/include/asm/page.h
113
addik rd, rs, (CONFIG_KERNEL_START - CONFIG_KERNEL_BASE_ADDR)
arch/mips/include/asm/asmmacro.h
205
.macro _EXT rd, rs, p, s
arch/mips/include/asm/asmmacro.h
209
.macro _EXT rd, rs, p, s
arch/mips/include/asm/asmmacro.h
249
.macro _cfcmsa rd, cs
arch/mips/include/asm/asmmacro.h
379
.macro _cfcmsa rd, cs
arch/mips/include/asm/mipsmtregs.h
410
#define mttr(rd, u, sel, v) \
arch/mips/include/asm/mipsmtregs.h
413
"mttr %0," #rd ", " #u ", " #sel \
arch/mips/include/asm/mipsregs.h
1771
_ASM_MACRO_2R_1S(mthc0, rt, rd, sel, \
arch/mips/include/asm/mipsregs.h
2214
_ASM_MACRO_2R_1S(mtgc0, rt, rd, sel, \
arch/mips/include/asm/mipsregs.h
2219
_ASM_MACRO_2R_1S(dmtgc0, rt, rd, sel, \
arch/mips/include/asm/msa.h
167
_ASM_MACRO_2R(cfcmsa, rd, cs, \
arch/mips/include/asm/octeon/cvmx-asm.h
130
asm ("pop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input))
arch/mips/include/asm/octeon/cvmx-asm.h
132
asm ("dpop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input))
arch/mips/include/asm/pci/bridge.h
190
u32 rd; /* read-only */
arch/mips/include/asm/pci/bridge.h
199
u32 rd; /* read-only */
arch/mips/include/asm/uasm.h
211
# define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_daddu(buf, rs, rt, rd)
arch/mips/include/asm/uasm.h
214
# define UASM_i_LWX(buf, rs, rt, rd) uasm_i_ldx(buf, rs, rt, rd)
arch/mips/include/asm/uasm.h
215
# define UASM_i_MFC0(buf, rt, rd...) uasm_i_dmfc0(buf, rt, rd)
arch/mips/include/asm/uasm.h
216
# define UASM_i_MTC0(buf, rt, rd...) uasm_i_dmtc0(buf, rt, rd)
arch/mips/include/asm/uasm.h
223
# define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_dsubu(buf, rs, rt, rd)
arch/mips/include/asm/uasm.h
227
# define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_addu(buf, rs, rt, rd)
arch/mips/include/asm/uasm.h
230
# define UASM_i_LWX(buf, rs, rt, rd) uasm_i_lwx(buf, rs, rt, rd)
arch/mips/include/asm/uasm.h
231
# define UASM_i_MFC0(buf, rt, rd...) uasm_i_mfc0(buf, rt, rd)
arch/mips/include/asm/uasm.h
232
# define UASM_i_MTC0(buf, rt, rd...) uasm_i_mtc0(buf, rt, rd)
arch/mips/include/asm/uasm.h
239
# define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_subu(buf, rs, rt, rd)
arch/mips/include/uapi/asm/inst.h
1051
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
682
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
692
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
702
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
722
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
733
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
792
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
802
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
948
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/include/uapi/asm/inst.h
959
__BITFIELD_FIELD(unsigned int rd : 5,
arch/mips/kernel/branch.c
436
regs->regs[insn.r_format.rd] = epc + 8;
arch/mips/kernel/process.c
275
if (ip->mm_m_format.rd < 0x10)
arch/mips/kernel/process.c
281
*poff += (ip->mm_m_format.rd & 0xf) * sizeof(u32);
arch/mips/kernel/traps.c
637
static int simulate_rdhwr(struct pt_regs *regs, int rd, int rt)
arch/mips/kernel/traps.c
643
switch (rd) {
arch/mips/kernel/traps.c
675
int rd = (opcode & RD) >> 11;
arch/mips/kernel/traps.c
678
simulate_rdhwr(regs, rd, rt);
arch/mips/kernel/traps.c
689
int rd = (opcode & MM_RS) >> 16;
arch/mips/kernel/traps.c
691
simulate_rdhwr(regs, rd, rt);
arch/mips/kernel/traps.c
731
int rd = (opcode & RD) >> 11;
arch/mips/kernel/traps.c
743
regs->regs[rd] = loongson3_cpucfg_read_synthesized(
arch/mips/kernel/unaligned.c
177
regs->regs[insn.mxu_lx_format.rd] = value;
arch/mips/kernel/unaligned.c
186
regs->regs[insn.dsp_format.rd] = value;
arch/mips/kernel/unaligned.c
195
regs->regs[insn.dsp_format.rd] = value;
arch/mips/kernel/unaligned.c
215
regs->regs[insn.dsp_format.rd] = value;
arch/mips/kernel/unaligned.c
224
regs->regs[insn.dsp_format.rd] = value;
arch/mips/kernel/unaligned.c
680
reg = insn.mm_x_format.rd;
arch/mips/kernel/unaligned.c
689
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
708
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
728
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
751
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
773
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
809
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
846
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
887
reg = insn.mm_m_format.rd;
arch/mips/kernel/unaligned.c
934
reg = insn.mm_m_format.rd;
arch/mips/kvm/emulate.c
65
arch->gprs[insn.r_format.rd] = epc + 8;
arch/mips/kvm/vz.c
1002
} else if (rd == MIPS_CP0_LLADDR &&
arch/mips/kvm/vz.c
1014
} else if (rd == MIPS_CP0_LLADDR &&
arch/mips/kvm/vz.c
1019
} else if (rd == MIPS_CP0_CONFIG &&
arch/mips/kvm/vz.c
1021
cop0->reg[rd][sel] = (int)val;
arch/mips/kvm/vz.c
1022
} else if (rd == MIPS_CP0_ERRCTL &&
arch/mips/kvm/vz.c
1026
} else if (rd == MIPS_CP0_DIAG &&
arch/mips/kvm/vz.c
1147
unsigned int rs, rd;
arch/mips/kvm/vz.c
1162
rd = inst.loongson3_lscsr_format.rd;
arch/mips/kvm/vz.c
1170
vcpu->arch.gprs[rd] = 0x14c000;
arch/mips/kvm/vz.c
1176
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1181
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1184
vcpu->arch.gprs[rd] = hostcfg;
arch/mips/kvm/vz.c
1188
vcpu->arch.gprs[rd] = 0;
arch/mips/kvm/vz.c
1218
int rd, rt, sel;
arch/mips/kvm/vz.c
1257
rd = inst.r_format.rd;
arch/mips/kvm/vz.c
1261
switch (rd) {
arch/mips/kvm/vz.c
1268
KVM_TRACE_HWR(rd, sel), 0);
arch/mips/kvm/vz.c
1273
KVM_TRACE_HWR(rd, sel), arch->gprs[rt]);
arch/mips/kvm/vz.c
1316
int rd = inst.c0r_format.rd;
arch/mips/kvm/vz.c
1321
trace_kvm_hwr(vcpu, KVM_TRACE_MTC0, KVM_TRACE_COP0(rd, sel),
arch/mips/kvm/vz.c
1324
if ((rd == MIPS_CP0_STATUS) && (sel == 0)) {
arch/mips/kvm/vz.c
1360
} else if ((rd == MIPS_CP0_CAUSE) && (sel == 0)) {
arch/mips/kvm/vz.c
1382
} else if ((rd == MIPS_CP0_STATUS) && (sel == 1)) { /* IntCtl */
arch/mips/kvm/vz.c
1384
} else if ((rd == MIPS_CP0_CONFIG) && (sel == 5)) {
arch/mips/kvm/vz.c
881
u32 rt, rd, sel;
arch/mips/kvm/vz.c
904
rd = inst.c0r_format.rd;
arch/mips/kvm/vz.c
911
cop0->stat[rd][sel]++;
arch/mips/kvm/vz.c
913
if (rd == MIPS_CP0_COUNT &&
arch/mips/kvm/vz.c
916
} else if (rd == MIPS_CP0_COMPARE &&
arch/mips/kvm/vz.c
919
} else if (rd == MIPS_CP0_LLADDR &&
arch/mips/kvm/vz.c
926
} else if (rd == MIPS_CP0_LLADDR &&
arch/mips/kvm/vz.c
935
} else if ((rd == MIPS_CP0_PRID &&
arch/mips/kvm/vz.c
939
(rd == MIPS_CP0_STATUS &&
arch/mips/kvm/vz.c
942
(rd == MIPS_CP0_CONFIG &&
arch/mips/kvm/vz.c
945
(rd == MIPS_CP0_LLADDR &&
arch/mips/kvm/vz.c
949
(rd == MIPS_CP0_ERRCTL &&
arch/mips/kvm/vz.c
951
val = cop0->reg[rd][sel];
arch/mips/kvm/vz.c
953
} else if (rd == MIPS_CP0_DIAG &&
arch/mips/kvm/vz.c
955
val = cop0->reg[rd][sel];
arch/mips/kvm/vz.c
971
KVM_TRACE_COP0(rd, sel), val);
arch/mips/kvm/vz.c
977
cop0->stat[rd][sel]++;
arch/mips/kvm/vz.c
982
KVM_TRACE_COP0(rd, sel), val);
arch/mips/kvm/vz.c
984
if (rd == MIPS_CP0_COUNT &&
arch/mips/kvm/vz.c
988
} else if (rd == MIPS_CP0_COMPARE &&
arch/mips/kvm/vz.c
993
} else if (rd == MIPS_CP0_LLADDR &&
arch/mips/math-emu/cp1emu.c
172
mips32_insn.r_format.rd = 0;
arch/mips/math-emu/cp1emu.c
257
mips32_insn.r_format.rd = insn.mm_fp4_format.rt;
arch/mips/math-emu/cp1emu.c
437
if (insn.r_format.rd != 0) {
arch/mips/math-emu/cp1emu.c
438
regs->regs[insn.r_format.rd] =
arch/mips/net/bpf_jit_comp32.c
1002
emit_mov_i(ctx, rd, (s32)imm); /* rd = imm */
arch/mips/net/bpf_jit_comp32.c
1003
emit(ctx, sltu, rd, lo(rs), rd); /* rd = rsl < rd */
arch/mips/net/bpf_jit_comp32.c
1005
emit(ctx, sltiu, rd, lo(rs), imm); /* rd = rsl < imm */
arch/mips/net/bpf_jit_comp32.c
1007
emit_movn_r(ctx, rd, MIPS_R_ZERO, hi(rs)); /* rd = 0 if rsh */
arch/mips/net/bpf_jit_comp32.c
1012
static void emit_sltu_r64(struct jit_context *ctx, u8 rd,
arch/mips/net/bpf_jit_comp32.c
1017
emit(ctx, sltu, rd, lo(rs), lo(rt)); /* rd = rsl < rtl */
arch/mips/net/bpf_jit_comp32.c
1019
emit_movn_r(ctx, rd, MIPS_R_ZERO, tmp); /* rd = 0 if tmp != 0 */
arch/mips/net/bpf_jit_comp32.c
1021
emit(ctx, or, rd, rd, tmp); /* rd = rd | tmp */
arch/mips/net/bpf_jit_comp32.c
1025
static void emit_slti_r64(struct jit_context *ctx, u8 rd,
arch/mips/net/bpf_jit_comp32.c
1036
emit_mov_i(ctx, rd, (s32)imm);
arch/mips/net/bpf_jit_comp32.c
1037
emit(ctx, sltu, t1, lo(rs), rd); /* t1 = rsl <u imm */
arch/mips/net/bpf_jit_comp32.c
1038
emit(ctx, sltu, t2, rd, lo(rs)); /* t2 = imm <u rsl */
arch/mips/net/bpf_jit_comp32.c
1039
emit(ctx, srl, rd, hi(rs), 31); /* rd = rsh >> 31 */
arch/mips/net/bpf_jit_comp32.c
1041
emit_movz_r(ctx, t1, t2, rd); /* t1 = rd ? t1 : t2 */
arch/mips/net/bpf_jit_comp32.c
1043
emit_movn_r(ctx, t1, t2, rd); /* t1 = rd ? t2 : t1 */
arch/mips/net/bpf_jit_comp32.c
1050
emit(ctx, addiu, rd, hi(rs), 1); /* rd = rsh + 1 */
arch/mips/net/bpf_jit_comp32.c
1051
cmp = rd;
arch/mips/net/bpf_jit_comp32.c
1062
emit(ctx, slti, rd, hi(rs), imm < 0 ? -1 : 0); /* rd = rsh < hi(imm) */
arch/mips/net/bpf_jit_comp32.c
1063
emit(ctx, or, rd, rd, t1); /* rd = rd | t1 */
arch/mips/net/bpf_jit_comp32.c
1067
static void emit_slt_r64(struct jit_context *ctx, u8 rd,
arch/mips/net/bpf_jit_comp32.c
1082
emit(ctx, srl, rd, t3, 31); /* rd = t3 >> 31 */
arch/mips/net/bpf_jit_comp32.c
1083
emit_movn_r(ctx, t1, t2, rd); /* t1 = rd ? t2 : t1 */
arch/mips/net/bpf_jit_comp32.c
1087
emit(ctx, slt, rd, hi(rs), hi(rt)); /* rd = rsh <s rth */
arch/mips/net/bpf_jit_comp32.c
1088
emit(ctx, or, rd, rd, t1); /* rd = rd | t1 */
arch/mips/net/bpf_jit_comp32.c
947
static void emit_movz_r(struct jit_context *ctx, u8 rd, u8 rs, u8 rt)
arch/mips/net/bpf_jit_comp32.c
950
emit(ctx, movz, rd, rs, rt); /* rd = rt ? rd : rs */
arch/mips/net/bpf_jit_comp32.c
954
emit(ctx, selnez, rd, rd, rt); /* rd = 0 if rt != 0 */
arch/mips/net/bpf_jit_comp32.c
956
emit(ctx, or, rd, rd, rs); /* rd = rd | rs */
arch/mips/net/bpf_jit_comp32.c
960
emit(ctx, or, rd, rs, MIPS_R_ZERO); /* +4: rd = rs */
arch/mips/net/bpf_jit_comp32.c
962
clobber_reg(ctx, rd);
arch/mips/net/bpf_jit_comp32.c
970
static void emit_movn_r(struct jit_context *ctx, u8 rd, u8 rs, u8 rt)
arch/mips/net/bpf_jit_comp32.c
973
emit(ctx, movn, rd, rs, rt); /* rd = rt ? rs : rd */
arch/mips/net/bpf_jit_comp32.c
977
emit(ctx, seleqz, rd, rd, rt); /* rd = 0 if rt != 0 */
arch/mips/net/bpf_jit_comp32.c
979
emit(ctx, or, rd, rd, rs); /* rd = rd | rs */
arch/mips/net/bpf_jit_comp32.c
983
emit(ctx, or, rd, rs, MIPS_R_ZERO); /* +4: rd = rs */
arch/mips/net/bpf_jit_comp32.c
985
clobber_reg(ctx, rd);
arch/mips/net/bpf_jit_comp32.c
990
static void emit_sltiu_r64(struct jit_context *ctx, u8 rd,
arch/mips/net/bpf_jit_comp32.c
996
emit_mov_i(ctx, rd, imm); /* rd = imm */
arch/mips/net/bpf_jit_comp32.c
997
emit(ctx, sltu, rd, lo(rs), rd); /* rd = rsl < rd */
arch/mips/net/bpf_jit_comp32.c
999
emit(ctx, or, rd, rd, tmp); /* rd = rd | tmp */
arch/openrisc/kernel/traps.c
321
unsigned int ra, rd;
arch/openrisc/kernel/traps.c
332
rd = (insn >> 21) & 0x1f;
arch/openrisc/kernel/traps.c
357
regs->gpr[rd] = value;
arch/parisc/net/bpf_jit.h
156
hppa_zdep(r, len, len, lo(rd))
arch/parisc/net/bpf_jit_comp32.c
1011
const s8 *rd = bpf_get_reg64_ref(dst, tmp1, ctx->prog->aux->verifier_zext, ctx);
arch/parisc/net/bpf_jit_comp32.c
1028
emit(hppa_ldb(off + 0, srcreg, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1030
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1033
emit(hppa_ldh(off + 0, srcreg, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1035
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1038
emit(hppa_ldw(off + 0, srcreg, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1040
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1043
emit(hppa_ldw(off + 0, srcreg, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1044
emit(hppa_ldw(off + 4, srcreg, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1048
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
1058
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
1064
dstreg = lo(rd);
arch/parisc/net/bpf_jit_comp32.c
1068
emit(hppa_addil(off, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1092
static void emit_rev16(const s8 rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
1094
emit(hppa_extru(rd, 23, 8, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp32.c
1095
emit(hppa_depwz(rd, 23, 8, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp32.c
1096
emit(hppa_extru(HPPA_REG_T1, 31, 16, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1099
static void emit_rev32(const s8 rs, const s8 rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
1103
emit(hppa_shrpw(rs, HPPA_REG_T1, 8, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1108
const s8 *rd;
arch/parisc/net/bpf_jit_comp32.c
1111
rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
1112
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1113
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
120
static void emit_hppa_copy(const s8 rs, const s8 rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
122
REG_SET_SEEN(ctx, rd);
arch/parisc/net/bpf_jit_comp32.c
123
if (OPTIMIZE_HPPA && (rs == rd))
arch/parisc/net/bpf_jit_comp32.c
1245
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
1250
emit(hppa_extru(lo(rd), 31, 16, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
1255
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
126
emit(hppa_copy(rs, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1265
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
1271
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
1275
emit_rev16(lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1277
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1280
emit_rev32(lo(rd), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1282
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1286
emit_hppa_copy(hi(rd), HPPA_REG_T0, ctx);
arch/parisc/net/bpf_jit_comp32.c
1287
emit_rev32(lo(rd), hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1288
emit_rev32(HPPA_REG_T0, lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
1295
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
1406
const s8 *rd = bpf_get_reg64_ref(dst, tmp1, false, ctx);
arch/parisc/net/bpf_jit_comp32.c
141
static void emit_imm(const s8 rd, s32 imm, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
1414
emit_imm64(rd, upper, lower, ctx);
arch/parisc/net/bpf_jit_comp32.c
1415
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
145
REG_SET_SEEN(ctx, rd);
arch/parisc/net/bpf_jit_comp32.c
147
emit(hppa_ldi(imm, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
150
emit(hppa_ldil(imm, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
153
emit(hppa_ldo(lower, rd, rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
156
static void emit_imm32(const s8 *rd, s32 imm, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
159
REG_SET_SEEN(ctx, lo(rd));
arch/parisc/net/bpf_jit_comp32.c
160
emit_imm(lo(rd), imm, ctx);
arch/parisc/net/bpf_jit_comp32.c
163
REG_SET_SEEN(ctx, hi(rd));
arch/parisc/net/bpf_jit_comp32.c
165
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
167
emit(hppa_ldi(-1, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
170
static void emit_imm64(const s8 *rd, s32 imm_hi, s32 imm_lo,
arch/parisc/net/bpf_jit_comp32.c
173
emit_imm(hi(rd), imm_hi, ctx);
arch/parisc/net/bpf_jit_comp32.c
174
emit_imm(lo(rd), imm_lo, ctx);
arch/parisc/net/bpf_jit_comp32.c
425
const s8 *rd;
arch/parisc/net/bpf_jit_comp32.c
428
rd = bpf_get_reg64_ref(dst, tmp1, false, ctx);
arch/parisc/net/bpf_jit_comp32.c
430
rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
435
emit_imm32(rd, imm, ctx);
arch/parisc/net/bpf_jit_comp32.c
439
emit(hppa_and(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
441
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
445
emit(hppa_or(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
447
emit_imm(hi(rd), -1, ctx);
arch/parisc/net/bpf_jit_comp32.c
451
emit_hppa_xor(lo(rd), HPPA_REG_T0, lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
454
emit_hppa_xor(hi(rd), HPPA_REG_T0, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
462
emit(hppa_zdep(lo(rd), imm, imm, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
463
emit_hppa_copy(HPPA_REG_ZERO, lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
465
emit_hppa_copy(lo(rd), hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
466
emit_hppa_copy(HPPA_REG_ZERO, lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
468
emit(hppa_shd(hi(rd), lo(rd), 32 - imm, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
469
emit(hppa_zdep(lo(rd), imm, imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
477
emit(hppa_shr(hi(rd), imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
478
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
480
emit_hppa_copy(hi(rd), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
481
emit_hppa_copy(HPPA_REG_ZERO, hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
483
emit(hppa_shrpw(hi(rd), lo(rd), imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
484
emit(hppa_shr(hi(rd), imm, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
492
emit(hppa_extrws(hi(rd), 31 - imm, imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
493
emit(hppa_extrws(hi(rd), 0, 31, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
495
emit_hppa_copy(hi(rd), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
496
emit(hppa_extrws(hi(rd), 0, 31, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
498
emit(hppa_shrpw(hi(rd), lo(rd), imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
499
emit(hppa_extrws(hi(rd), 31 - imm, imm, hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
506
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
513
const s8 *rd = bpf_get_reg32(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
516
rd = bpf_get_reg32_ref(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
518
rd = bpf_get_reg32(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
523
emit_imm(lo(rd), imm, ctx);
arch/parisc/net/bpf_jit_comp32.c
527
emit(hppa_add(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
531
emit(hppa_sub(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
535
emit(hppa_and(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
539
emit(hppa_or(lo(rd), HPPA_REG_T0, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
543
emit_hppa_xor(lo(rd), HPPA_REG_T0, lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
547
emit(hppa_zdep(lo(rd), imm, imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
551
emit(hppa_shr(lo(rd), imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
555
emit(hppa_extrws(lo(rd), 31 - imm, imm, lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
561
bpf_put_reg32(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
569
const s8 *rd;
arch/parisc/net/bpf_jit_comp32.c
573
rd = bpf_get_reg64_ref(dst, tmp1, false, ctx);
arch/parisc/net/bpf_jit_comp32.c
575
rd = bpf_get_reg64(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
580
emit_hppa_copy(lo(rs), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
581
emit_hppa_copy(hi(rs), hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
584
emit(hppa_add(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
585
emit(hppa_addc(hi(rd), hi(rs), hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
588
emit(hppa_sub(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
589
emit(hppa_subb(hi(rd), hi(rs), hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
592
emit(hppa_and(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
593
emit(hppa_and(hi(rd), hi(rs), hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
596
emit(hppa_or(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
597
emit(hppa_or(hi(rd), hi(rs), hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
600
emit_hppa_xor(lo(rd), lo(rs), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
601
emit_hppa_xor(hi(rd), hi(rs), hi(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
604
emit_call_libgcc_ll(__muldi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
607
emit_call_libgcc_ll(&hppa_div64, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
610
emit_call_libgcc_ll(&hppa_div64_rem, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
613
emit_call_libgcc_ll(__ashldi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
616
emit_call_libgcc_ll(__lshrdi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
619
emit_call_libgcc_ll(__ashrdi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
622
emit(hppa_sub(HPPA_REG_ZERO, lo(rd), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
623
emit(hppa_subb(HPPA_REG_ZERO, hi(rd), hi(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
629
bpf_put_reg64(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
637
const s8 *rd;
arch/parisc/net/bpf_jit_comp32.c
641
rd = bpf_get_reg32_ref(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
643
rd = bpf_get_reg32(dst, tmp1, ctx);
arch/parisc/net/bpf_jit_comp32.c
648
emit_hppa_copy(lo(rs), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
651
emit(hppa_add(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
654
emit(hppa_sub(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
657
emit(hppa_and(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
660
emit(hppa_or(lo(rd), lo(rs), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
663
emit_hppa_xor(lo(rd), lo(rs), lo(rd), ctx);
arch/parisc/net/bpf_jit_comp32.c
666
emit_call_millicode($$mulI, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
669
emit_call_millicode($$divU, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
672
emit_call_millicode($$remU, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
677
emit(hppa_depwz_sar(lo(rd), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
681
emit(hppa_shrpw_sar(lo(rd), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
687
emit(hppa_extrws_sar(lo(rd), lo(rd)), ctx);
arch/parisc/net/bpf_jit_comp32.c
690
emit(hppa_sub(HPPA_REG_ZERO, lo(rd), lo(rd)), ctx); // sub r0,rd,rd
arch/parisc/net/bpf_jit_comp32.c
696
bpf_put_reg32(dst, rd, ctx);
arch/parisc/net/bpf_jit_comp32.c
789
static int emit_bcc(u8 op, u8 rd, u8 rs, int paoff, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
800
emit(hppa_and(rd, rs, HPPA_REG_T0), ctx);
arch/parisc/net/bpf_jit_comp32.c
802
rd = HPPA_REG_T0;
arch/parisc/net/bpf_jit_comp32.c
824
emit(hppa_beq(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
827
emit(hppa_bgtu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
830
emit(hppa_bltu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
833
emit(hppa_bgeu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
836
emit(hppa_bleu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
839
emit(hppa_bne(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
842
emit(hppa_bgt(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
845
emit(hppa_blt(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
848
emit(hppa_bge(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp32.c
851
emit(hppa_ble(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
1005
emit_imm(rd, imm64, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
1029
BUG_ON(rd == HPPA_REG_R1);
arch/parisc/net/bpf_jit_comp64.c
1036
emit(hppa_ldb(off, srcreg, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1041
emit(hppa_ldh(off, srcreg, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1046
emit(hppa_ldw(off, srcreg, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1053
emit(hppa64_ldd_reg(HPPA_REG_ZERO, HPPA_REG_R1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1055
emit(hppa64_ldd_im5(off, srcreg, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1057
emit(hppa64_ldd_im16(off, srcreg, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
1082
emit_store(rd, rs, off, ctx, BPF_SIZE(code), BPF_MODE(code));
arch/parisc/net/bpf_jit_comp64.c
135
static void emit_imm32(u8 rd, s32 imm, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
139
REG_SET_SEEN(ctx, rd);
arch/parisc/net/bpf_jit_comp64.c
141
emit(hppa_ldi(imm, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
145
emit(hppa_ldo(lower, HPPA_REG_ZERO, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
148
emit(hppa_ldil(imm, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
151
emit(hppa_ldo(lower, rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
160
static void emit_imm(u8 rd, s64 imm, u8 tmpreg, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
165
emit_imm32(rd, imm, ctx);
arch/parisc/net/bpf_jit_comp64.c
175
emit_hppa64_depd(tmpreg, 31, 32, rd, 1, ctx);
arch/parisc/net/bpf_jit_comp64.c
177
emit_hppa64_depd(HPPA_REG_ZERO, 31, 32, rd, 1, ctx);
arch/parisc/net/bpf_jit_comp64.c
248
static int emit_branch(u8 op, u8 rd, u8 rs, signed long paoff,
arch/parisc/net/bpf_jit_comp64.c
260
emit(hppa_and(rd, rs, HPPA_REG_T0), ctx);
arch/parisc/net/bpf_jit_comp64.c
262
rd = HPPA_REG_T0;
arch/parisc/net/bpf_jit_comp64.c
285
emit(hppa_beq(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
288
emit(hppa_bgtu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
291
emit(hppa_bltu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
294
emit(hppa_bgeu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
297
emit(hppa_bleu(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
300
emit(hppa_bne(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
303
emit(hppa_bgt(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
306
emit(hppa_blt(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
309
emit(hppa_bge(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
312
emit(hppa_ble(rd, rs, off), ctx);
arch/parisc/net/bpf_jit_comp64.c
407
static void init_regs(u8 *rd, u8 *rs, const struct bpf_insn *insn,
arch/parisc/net/bpf_jit_comp64.c
419
*rd = bpf_to_hppa_reg(insn->dst_reg, ctx);
arch/parisc/net/bpf_jit_comp64.c
428
static void emit_zext_32_rd_rs(u8 *rd, u8 *rs, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
430
emit_hppa64_zext32(*rd, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
431
*rd = HPPA_REG_T2;
arch/parisc/net/bpf_jit_comp64.c
436
static void emit_sext_32_rd_rs(u8 *rd, u8 *rs, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
438
emit_hppa64_sext32(*rd, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
439
*rd = HPPA_REG_T2;
arch/parisc/net/bpf_jit_comp64.c
444
static void emit_zext_32_rd_t1(u8 *rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
446
emit_hppa64_zext32(*rd, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
447
*rd = HPPA_REG_T2;
arch/parisc/net/bpf_jit_comp64.c
451
static void emit_sext_32_rd(u8 *rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
453
emit_hppa64_sext32(*rd, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
454
*rd = HPPA_REG_T2;
arch/parisc/net/bpf_jit_comp64.c
554
static void emit_store(const s8 rd, const s8 rs, s16 off,
arch/parisc/net/bpf_jit_comp64.c
562
dstreg = rd;
arch/parisc/net/bpf_jit_comp64.c
566
emit(hppa_addil(off, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
600
u8 rd = -1, rs = -1, code = insn->code;
arch/parisc/net/bpf_jit_comp64.c
604
init_regs(&rd, &rs, insn, ctx);
arch/parisc/net/bpf_jit_comp64.c
612
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
616
emit_hppa64_zext32(rs, rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
618
emit_hppa_copy(rs, rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
624
emit(hppa_add(rd, rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
626
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
630
emit(hppa_sub(rd, rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
632
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
636
emit(hppa_and(rd, rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
638
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
642
emit(hppa_or(rd, rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
644
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
648
emit(hppa_xor(rd, rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
649
if (!is64 && !aux->verifier_zext && rs != rd)
arch/parisc/net/bpf_jit_comp64.c
650
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
659
emit_call_libgcc_ll(__muldi3, rd, rs, code, ctx);
arch/parisc/net/bpf_jit_comp64.c
661
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
670
emit_call_libgcc_ll(&hppa_div64, rd, rs, code, ctx);
arch/parisc/net/bpf_jit_comp64.c
672
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
681
emit_call_libgcc_ll(&hppa_div64_rem, rd, rs, code, ctx);
arch/parisc/net/bpf_jit_comp64.c
683
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
691
emit(hppa64_depdz_sar(rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
693
emit(hppa_depwz_sar(rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
695
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
70
static void emit_hppa_copy(const s8 rs, const s8 rd, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp64.c
701
emit(hppa64_shrpd_sar(rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
703
emit(hppa_shrpw_sar(rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
705
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
712
emit(hppa_extrd_sar(rd, rd, 1), ctx);
arch/parisc/net/bpf_jit_comp64.c
714
emit(hppa_extrws_sar(rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
716
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
72
REG_SET_SEEN(ctx, rd);
arch/parisc/net/bpf_jit_comp64.c
722
emit(hppa_sub(HPPA_REG_ZERO, rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
724
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
73
if (OPTIMIZE_HPPA && (rs == rd))
arch/parisc/net/bpf_jit_comp64.c
732
emit_hppa64_depd(HPPA_REG_ZERO, 63-16, 64-16, rd, 1, ctx);
arch/parisc/net/bpf_jit_comp64.c
736
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
747
emit(hppa_extru(rd, 31 - 8, 8, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
748
emit(hppa_depwz(rd, 23, 8, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
749
emit(hppa_extru(HPPA_REG_T1, 31, 16, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
750
emit_hppa64_extrd(HPPA_REG_T1, 63, 16, rd, 0, ctx);
arch/parisc/net/bpf_jit_comp64.c
753
emit(hppa_shrpw(rd, rd, 16, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
755
emit(hppa_shrpw(rd, HPPA_REG_T1, 8, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
756
emit_hppa64_extrd(HPPA_REG_T1, 63, 32, rd, 0, ctx);
arch/parisc/net/bpf_jit_comp64.c
759
emit(hppa64_permh_3210(rd, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
76
emit(hppa_copy(rs, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
762
emit(hppa_or(HPPA_REG_T2, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
773
emit_imm(rd, imm, HPPA_REG_T2, ctx);
arch/parisc/net/bpf_jit_comp64.c
775
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
782
emit(hppa_ldo(imm, rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
785
emit(hppa_add(rd, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
788
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
793
emit(hppa_ldo(-imm, rd, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
796
emit(hppa_sub(rd, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
799
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
804
emit(hppa_and(rd, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
806
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
811
emit(hppa_or(rd, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
813
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
818
emit(hppa_xor(rd, HPPA_REG_T1, rd), ctx);
arch/parisc/net/bpf_jit_comp64.c
820
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
825
emit_hppa64_shld(rd, imm, rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
829
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
835
emit_hppa64_shrd(rd, imm, rd, false, ctx);
arch/parisc/net/bpf_jit_comp64.c
837
emit_hppa64_shrw(rd, imm, rd, false, ctx);
arch/parisc/net/bpf_jit_comp64.c
841
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
847
emit_hppa64_shrd(rd, imm, rd, true, ctx);
arch/parisc/net/bpf_jit_comp64.c
849
emit_hppa64_shrw(rd, imm, rd, true, ctx);
arch/parisc/net/bpf_jit_comp64.c
853
emit_zext_32(rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
891
emit_sext_32_rd_rs(&rd, &rs, ctx);
arch/parisc/net/bpf_jit_comp64.c
893
emit_zext_32_rd_rs(&rd, &rs, ctx);
arch/parisc/net/bpf_jit_comp64.c
902
emit(hppa_and(rs, rd, HPPA_REG_T1), ctx);
arch/parisc/net/bpf_jit_comp64.c
906
emit_branch(BPF_OP(code), rd, rs, paoff, ctx);
arch/parisc/net/bpf_jit_comp64.c
941
emit_sext_32_rd(&rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
943
emit_zext_32_rd_t1(&rd, ctx);
arch/parisc/net/bpf_jit_comp64.c
949
emit_branch(BPF_OP(code), rd, rs, paoff, ctx);
arch/parisc/net/bpf_jit_comp64.c
956
emit(hppa_and(HPPA_REG_T1, rd, HPPA_REG_T1), ctx);
arch/powerpc/include/asm/ppc_asm.h
540
#define toreal(rd)
arch/powerpc/include/asm/ppc_asm.h
541
#define fromreal(rd)
arch/powerpc/include/asm/ppc_asm.h
549
#define tophys(rd,rs) \
arch/powerpc/include/asm/ppc_asm.h
550
addis rd,rs,0
arch/powerpc/include/asm/ppc_asm.h
552
#define tovirt(rd,rs) \
arch/powerpc/include/asm/ppc_asm.h
553
addis rd,rs,0
arch/powerpc/include/asm/ppc_asm.h
556
#define toreal(rd) /* we can access c000... in real mode */
arch/powerpc/include/asm/ppc_asm.h
557
#define fromreal(rd)
arch/powerpc/include/asm/ppc_asm.h
559
#define tophys(rd,rs) \
arch/powerpc/include/asm/ppc_asm.h
560
clrldi rd,rs,2
arch/powerpc/include/asm/ppc_asm.h
562
#define tovirt(rd,rs) \
arch/powerpc/include/asm/ppc_asm.h
563
rotldi rd,rs,16; \
arch/powerpc/include/asm/ppc_asm.h
564
ori rd,rd,((KERNELBASE>>48)&0xFFFF);\
arch/powerpc/include/asm/ppc_asm.h
565
rotldi rd,rd,48
arch/powerpc/include/asm/ppc_asm.h
567
#define toreal(rd) tophys(rd,rd)
arch/powerpc/include/asm/ppc_asm.h
568
#define fromreal(rd) tovirt(rd,rd)
arch/powerpc/include/asm/ppc_asm.h
570
#define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h
arch/powerpc/include/asm/ppc_asm.h
571
#define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h
arch/powerpc/kernel/traps.c
1350
u32 rd;
arch/powerpc/kernel/traps.c
1361
rd = (instword >> 21) & 0x1f;
arch/powerpc/kernel/traps.c
1362
regs->gpr[rd] = mfspr(SPRN_PVR);
arch/powerpc/kernel/traps.c
1419
rd = (instword >> 21) & 0x1f;
arch/powerpc/kernel/traps.c
1420
regs->gpr[rd] = mfspr(SPRN_DSCR);
arch/powerpc/kernel/traps.c
1430
rd = (instword >> 21) & 0x1f;
arch/powerpc/kernel/traps.c
1431
current->thread.dscr = regs->gpr[rd];
arch/powerpc/kernel/traps.c
1757
u32 instword, rd;
arch/powerpc/kernel/traps.c
1807
rd = (instword >> 21) & 0x1f;
arch/powerpc/kernel/traps.c
1808
current->thread.dscr = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1164
struct instruction_op *op, int rd,
arch/powerpc/lib/sstep.c
1173
op->reg = rd;
arch/powerpc/lib/sstep.c
1357
unsigned int opcode, ra, rb, rc, rd, spr, u;
arch/powerpc/lib/sstep.c
1408
rd = 7 - ((word >> 23) & 0x7);
arch/powerpc/lib/sstep.c
1410
rd *= 4;
arch/powerpc/lib/sstep.c
1413
op->ccval = (regs->ccr & ~(0xfUL << rd)) | (val << rd);
arch/powerpc/lib/sstep.c
1448
rd = (word >> 21) & 0x1f;
arch/powerpc/lib/sstep.c
1452
op->ccval = (regs->ccr & ~(1UL << (31 - rd))) |
arch/powerpc/lib/sstep.c
1453
(val << (31 - rd));
arch/powerpc/lib/sstep.c
1480
rd = (word >> 21) & 0x1f;
arch/powerpc/lib/sstep.c
1493
rd = (suffix >> 21) & 0x1f;
arch/powerpc/lib/sstep.c
1494
op->reg = rd;
arch/powerpc/lib/sstep.c
1495
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1511
if (rd & trap_compare(regs->gpr[ra], (short) word))
arch/powerpc/lib/sstep.c
1516
if (rd & trap_compare((int)regs->gpr[ra], (short) word))
arch/powerpc/lib/sstep.c
1563
add_with_carry(regs, op, rd, ~regs->gpr[ra], imm, 1);
arch/powerpc/lib/sstep.c
1570
if ((rd & 1) == 0)
arch/powerpc/lib/sstep.c
1573
do_cmp_unsigned(regs, op, val, imm, rd >> 2);
arch/powerpc/lib/sstep.c
1580
if ((rd & 1) == 0)
arch/powerpc/lib/sstep.c
1583
do_cmp_signed(regs, op, val, imm, rd >> 2);
arch/powerpc/lib/sstep.c
1588
add_with_carry(regs, op, rd, regs->gpr[ra], imm, 0);
arch/powerpc/lib/sstep.c
1593
add_with_carry(regs, op, rd, regs->gpr[ra], imm, 0);
arch/powerpc/lib/sstep.c
1627
val = DATA32(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
1635
val = DATA32(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
1643
val = DATA32(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
1648
op->val = regs->gpr[rd] | (unsigned short) word;
arch/powerpc/lib/sstep.c
1653
op->val = regs->gpr[rd] | (imm << 16);
arch/powerpc/lib/sstep.c
1657
op->val = regs->gpr[rd] ^ (unsigned short) word;
arch/powerpc/lib/sstep.c
1662
op->val = regs->gpr[rd] ^ (imm << 16);
arch/powerpc/lib/sstep.c
1666
op->val = regs->gpr[rd] & (unsigned short) word;
arch/powerpc/lib/sstep.c
1672
op->val = regs->gpr[rd] & (imm << 16);
arch/powerpc/lib/sstep.c
1679
val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1729
if (rd == 0x1f ||
arch/powerpc/lib/sstep.c
1730
(rd & trap_compare((int)regs->gpr[ra],
arch/powerpc/lib/sstep.c
1736
if (rd & trap_compare(regs->gpr[ra], regs->gpr[rb]))
arch/powerpc/lib/sstep.c
1744
op->reg = rd;
arch/powerpc/lib/sstep.c
1750
op->reg = rd;
arch/powerpc/lib/sstep.c
1758
op->reg = rd;
arch/powerpc/lib/sstep.c
1804
val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1817
op->reg = rd;
arch/powerpc/lib/sstep.c
1827
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1841
if ((rd & 1) == 0) {
arch/powerpc/lib/sstep.c
1847
do_cmp_signed(regs, op, val, val2, rd >> 2);
arch/powerpc/lib/sstep.c
1854
if ((rd & 1) == 0) {
arch/powerpc/lib/sstep.c
1860
do_cmp_unsigned(regs, op, val, val2, rd >> 2);
arch/powerpc/lib/sstep.c
1864
do_cmpb(regs, op, regs->gpr[rd], regs->gpr[rb]);
arch/powerpc/lib/sstep.c
1871
add_with_carry(regs, op, rd, ~regs->gpr[ra],
arch/powerpc/lib/sstep.c
1881
add_with_carry(regs, op, rd, regs->gpr[ra],
arch/powerpc/lib/sstep.c
1909
add_with_carry(regs, op, rd, ~regs->gpr[ra],
arch/powerpc/lib/sstep.c
1914
add_with_carry(regs, op, rd, regs->gpr[ra],
arch/powerpc/lib/sstep.c
1919
add_with_carry(regs, op, rd, ~regs->gpr[ra], 0L,
arch/powerpc/lib/sstep.c
1924
add_with_carry(regs, op, rd, regs->gpr[ra], 0L,
arch/powerpc/lib/sstep.c
1929
add_with_carry(regs, op, rd, ~regs->gpr[ra], -1L,
arch/powerpc/lib/sstep.c
1938
add_with_carry(regs, op, rd, regs->gpr[ra], -1L,
arch/powerpc/lib/sstep.c
2036
val = (unsigned int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2041
val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
2046
op->val = regs->gpr[rd] & regs->gpr[rb];
arch/powerpc/lib/sstep.c
2050
op->val = regs->gpr[rd] & ~regs->gpr[rb];
arch/powerpc/lib/sstep.c
2054
do_popcnt(regs, op, regs->gpr[rd], 8);
arch/powerpc/lib/sstep.c
2058
op->val = ~(regs->gpr[rd] | regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2062
do_prty(regs, op, regs->gpr[rd], 32);
arch/powerpc/lib/sstep.c
2066
do_prty(regs, op, regs->gpr[rd], 64);
arch/powerpc/lib/sstep.c
2070
do_bpermd(regs, op, regs->gpr[rd], regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2074
op->val = ~(regs->gpr[rd] ^ regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2078
op->val = regs->gpr[rd] ^ regs->gpr[rb];
arch/powerpc/lib/sstep.c
2082
do_popcnt(regs, op, regs->gpr[rd], 32);
arch/powerpc/lib/sstep.c
2086
op->val = regs->gpr[rd] | ~regs->gpr[rb];
arch/powerpc/lib/sstep.c
2090
op->val = regs->gpr[rd] | regs->gpr[rb];
arch/powerpc/lib/sstep.c
2094
op->val = ~(regs->gpr[rd] & regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2098
do_popcnt(regs, op, regs->gpr[rd], 64);
arch/powerpc/lib/sstep.c
2104
val = (unsigned int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2111
val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
2116
op->val = (signed short) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2120
op->val = (signed char) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2124
op->val = (signed int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2134
op->val = (regs->gpr[rd] << sh) & 0xffffffffUL;
arch/powerpc/lib/sstep.c
2142
op->val = (regs->gpr[rd] & 0xffffffffUL) >> sh;
arch/powerpc/lib/sstep.c
2150
ival = (signed int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2163
ival = (signed int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2177
op->val = regs->gpr[rd] << sh;
arch/powerpc/lib/sstep.c
2185
op->val = regs->gpr[rd] >> sh;
arch/powerpc/lib/sstep.c
2193
ival = (signed long int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2207
ival = (signed long int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2223
val = (signed int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2248
op->reg = rd;
arch/powerpc/lib/sstep.c
2254
op->reg = rd;
arch/powerpc/lib/sstep.c
2275
op->reg = rd;
arch/powerpc/lib/sstep.c
2276
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
2320
if (!((rd & 1) || rd == ra || rd == rb))
arch/powerpc/lib/sstep.c
2325
if (!(rd & 1))
arch/powerpc/lib/sstep.c
2499
op->val = byterev_8(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2509
op->val = byterev_4(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2525
op->val = byterev_2(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2530
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2536
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2542
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2550
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2561
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2573
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2582
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2590
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2599
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2610
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2624
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2629
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2636
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2642
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2649
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2655
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2663
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2672
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2681
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2688
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2696
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2703
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2711
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2720
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2729
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2736
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2744
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2797
if (ra >= rd)
arch/powerpc/lib/sstep.c
2799
op->type = MKOP(LOAD_MULTI, 0, 4 * (32 - rd));
arch/powerpc/lib/sstep.c
2804
op->type = MKOP(STORE_MULTI, 0, 4 * (32 - rd));
arch/powerpc/lib/sstep.c
2836
if (!((rd & 1) || (rd == ra)))
arch/powerpc/lib/sstep.c
2847
if (rd & 1)
arch/powerpc/lib/sstep.c
2854
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2862
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2893
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2918
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2929
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2940
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2951
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2971
if (!(rd & 1))
arch/powerpc/lib/sstep.c
2983
rd = (suffix >> 21) & 0x1f;
arch/powerpc/lib/sstep.c
2984
op->reg = rd;
arch/powerpc/lib/sstep.c
2985
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
3000
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3006
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3012
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3018
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3032
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3048
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
3061
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
3120
if (ra == rd)
arch/powerpc/lib/sstep.c
3157
op->reg = rd;
arch/powerpc/lib/sstep.c
33
#define VSX_REGISTER_XTP(rd) ((((rd) & 1) << 5) | ((rd) & 0xfe))
arch/powerpc/lib/sstep.c
3323
int i, rd, nb;
arch/powerpc/lib/sstep.c
3467
rd = op->reg;
arch/powerpc/lib/sstep.c
3479
regs->gpr[rd] = v32;
arch/powerpc/lib/sstep.c
3482
rd = (rd + 1) & 0x1f;
arch/powerpc/lib/sstep.c
3537
rd = op->reg;
arch/powerpc/lib/sstep.c
3539
unsigned int v32 = regs->gpr[rd];
arch/powerpc/lib/sstep.c
3551
rd = (rd + 1) & 0x1f;
arch/powerpc/platforms/pseries/hotplug-cpu.c
841
struct of_reconfig_data *rd = data;
arch/powerpc/platforms/pseries/hotplug-cpu.c
846
err = pseries_add_processor(rd->dn);
arch/powerpc/platforms/pseries/hotplug-cpu.c
849
pseries_remove_processor(rd->dn);
arch/powerpc/platforms/pseries/hotplug-memory.c
900
struct of_reconfig_data *rd = data;
arch/powerpc/platforms/pseries/hotplug-memory.c
905
err = pseries_add_mem_node(rd->dn);
arch/powerpc/platforms/pseries/hotplug-memory.c
908
err = pseries_remove_mem_node(rd->dn);
arch/powerpc/platforms/pseries/hotplug-memory.c
911
if (!strcmp(rd->dn->name,
arch/powerpc/platforms/pseries/hotplug-memory.c
913
drmem_update_lmbs(rd->prop);
arch/powerpc/platforms/pseries/iommu.c
2471
struct of_reconfig_data *rd = data;
arch/powerpc/platforms/pseries/iommu.c
2472
struct device_node *np = rd->dn;
arch/powerpc/platforms/pseries/setup.c
280
struct of_reconfig_data *rd = data;
arch/powerpc/platforms/pseries/setup.c
281
struct device_node *parent, *np = rd->dn;
arch/powerpc/platforms/pseries/vas.c
938
struct of_reconfig_data *rd = data;
arch/powerpc/platforms/pseries/vas.c
939
struct device_node *dn = rd->dn;
arch/powerpc/sysdev/fsl_pci.c
1003
regs->gpr[rd] = 0xffffffff;
arch/powerpc/sysdev/fsl_pci.c
1007
regs->gpr[rd] = 0xffffffff;
arch/powerpc/sysdev/fsl_pci.c
1012
regs->gpr[rd] = 0xff;
arch/powerpc/sysdev/fsl_pci.c
1016
regs->gpr[rd] = 0xff;
arch/powerpc/sysdev/fsl_pci.c
1021
regs->gpr[rd] = 0xffff;
arch/powerpc/sysdev/fsl_pci.c
1025
regs->gpr[rd] = 0xffff;
arch/powerpc/sysdev/fsl_pci.c
1030
regs->gpr[rd] = ~0UL;
arch/powerpc/sysdev/fsl_pci.c
1034
regs->gpr[rd] = ~0UL;
arch/powerpc/sysdev/fsl_pci.c
949
unsigned int rd, ra, rb, d;
arch/powerpc/sysdev/fsl_pci.c
951
rd = get_rt(inst);
arch/powerpc/sysdev/fsl_pci.c
961
regs->gpr[rd] = 0xffffffff;
arch/powerpc/sysdev/fsl_pci.c
965
regs->gpr[rd] = 0xffffffff;
arch/powerpc/sysdev/fsl_pci.c
970
regs->gpr[rd] = 0xff;
arch/powerpc/sysdev/fsl_pci.c
974
regs->gpr[rd] = 0xff;
arch/powerpc/sysdev/fsl_pci.c
980
regs->gpr[rd] = 0xffff;
arch/powerpc/sysdev/fsl_pci.c
984
regs->gpr[rd] = 0xffff;
arch/powerpc/sysdev/fsl_pci.c
989
regs->gpr[rd] = ~0UL;
arch/powerpc/sysdev/fsl_pci.c
993
regs->gpr[rd] = ~0UL;
arch/riscv/include/asm/insn-def.h
139
#define __INSN_R(opcode, func3, func7, rd, rs1, rs2) \
arch/riscv/include/asm/insn-def.h
141
"insn_r " opcode ", " func3 ", " func7 ", " rd ", " rs1 ", " rs2 "\n" \
arch/riscv/include/asm/insn-def.h
144
#define __INSN_I(opcode, func3, rd, rs1, simm12) \
arch/riscv/include/asm/insn-def.h
146
"insn_i " opcode ", " func3 ", " rd ", " rs1 ", " simm12 "\n" \
arch/riscv/include/asm/insn-def.h
158
#define INSN_R(opcode, func3, func7, rd, rs1, rs2) \
arch/riscv/include/asm/insn-def.h
160
RV_##rd, RV_##rs1, RV_##rs2)
arch/riscv/include/asm/insn-def.h
162
#define INSN_I(opcode, func3, rd, rs1, simm12) \
arch/riscv/include/asm/insn-def.h
163
__INSN_I(RV_##opcode, RV_##func3, RV_##rd, \
arch/riscv/include/asm/insn-def.h
32
.macro insn_r, opcode, func3, func7, rd, rs1, rs2
arch/riscv/include/asm/insn-def.h
36
.macro insn_i, opcode, func3, rd, rs1, simm12
arch/riscv/include/asm/insn-def.h
48
.macro insn_r, opcode, func3, func7, rd, rs1, rs2
arch/riscv/include/asm/insn-def.h
57
.macro insn_i, opcode, func3, rd, rs1, simm12
arch/riscv/include/asm/insn-def.h
84
#define __INSN_R(opcode, func3, func7, rd, rs1, rs2) \
arch/riscv/include/asm/insn-def.h
85
".insn r " opcode ", " func3 ", " func7 ", " rd ", " rs1 ", " rs2 "\n"
arch/riscv/include/asm/insn-def.h
87
#define __INSN_I(opcode, func3, rd, rs1, simm12) \
arch/riscv/include/asm/insn-def.h
88
".insn i " opcode ", " func3 ", " rd ", " rs1 ", " simm12 "\n"
arch/riscv/net/bpf_jit.h
1002
static inline void emit_mv(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1004
if (rvc_enabled() && rd && rs)
arch/riscv/net/bpf_jit.h
1005
emitc(rvc_mv(rd, rs), ctx);
arch/riscv/net/bpf_jit.h
1007
emit(rv_addi(rd, rs, 0), ctx);
arch/riscv/net/bpf_jit.h
1010
static inline void emit_add(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1012
if (rvc_enabled() && rd && rd == rs1 && rs2)
arch/riscv/net/bpf_jit.h
1013
emitc(rvc_add(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1015
emit(rv_add(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1018
static inline void emit_addi(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1020
if (rvc_enabled() && rd == RV_REG_SP && rd == rs && is_10b_int(imm) && imm && !(imm & 0xf))
arch/riscv/net/bpf_jit.h
1022
else if (rvc_enabled() && is_creg(rd) && rs == RV_REG_SP && is_10b_uint(imm) &&
arch/riscv/net/bpf_jit.h
1024
emitc(rvc_addi4spn(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1025
else if (rvc_enabled() && rd && rd == rs && imm && is_6b_int(imm))
arch/riscv/net/bpf_jit.h
1026
emitc(rvc_addi(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1028
emit(rv_addi(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1031
static inline void emit_li(u8 rd, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1033
if (rvc_enabled() && rd && is_6b_int(imm))
arch/riscv/net/bpf_jit.h
1034
emitc(rvc_li(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1036
emit(rv_addi(rd, RV_REG_ZERO, imm), ctx);
arch/riscv/net/bpf_jit.h
1039
static inline void emit_lui(u8 rd, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1041
if (rvc_enabled() && rd && rd != RV_REG_SP && is_6b_int(imm) && imm)
arch/riscv/net/bpf_jit.h
1042
emitc(rvc_lui(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1044
emit(rv_lui(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1047
static inline void emit_slli(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1049
if (rvc_enabled() && rd && rd == rs && imm && (u32)imm < __riscv_xlen)
arch/riscv/net/bpf_jit.h
1050
emitc(rvc_slli(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1052
emit(rv_slli(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1055
static inline void emit_andi(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1057
if (rvc_enabled() && is_creg(rd) && rd == rs && is_6b_int(imm))
arch/riscv/net/bpf_jit.h
1058
emitc(rvc_andi(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1060
emit(rv_andi(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1063
static inline void emit_srli(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1065
if (rvc_enabled() && is_creg(rd) && rd == rs && imm && (u32)imm < __riscv_xlen)
arch/riscv/net/bpf_jit.h
1066
emitc(rvc_srli(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1068
emit(rv_srli(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1071
static inline void emit_srai(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1073
if (rvc_enabled() && is_creg(rd) && rd == rs && imm && (u32)imm < __riscv_xlen)
arch/riscv/net/bpf_jit.h
1074
emitc(rvc_srai(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1076
emit(rv_srai(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1079
static inline void emit_sub(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1081
if (rvc_enabled() && is_creg(rd) && rd == rs1 && is_creg(rs2))
arch/riscv/net/bpf_jit.h
1082
emitc(rvc_sub(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1084
emit(rv_sub(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1087
static inline void emit_or(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1089
if (rvc_enabled() && is_creg(rd) && rd == rs1 && is_creg(rs2))
arch/riscv/net/bpf_jit.h
1090
emitc(rvc_or(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1092
emit(rv_or(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1095
static inline void emit_and(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1097
if (rvc_enabled() && is_creg(rd) && rd == rs1 && is_creg(rs2))
arch/riscv/net/bpf_jit.h
1098
emitc(rvc_and(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1100
emit(rv_and(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1103
static inline void emit_xor(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1105
if (rvc_enabled() && is_creg(rd) && rd == rs1 && is_creg(rs2))
arch/riscv/net/bpf_jit.h
1106
emitc(rvc_xor(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1108
emit(rv_xor(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1111
static inline void emit_lw(u8 rd, s32 off, u8 rs1, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1113
if (rvc_enabled() && rs1 == RV_REG_SP && rd && is_8b_uint(off) && !(off & 0x3))
arch/riscv/net/bpf_jit.h
1114
emitc(rvc_lwsp(rd, off), ctx);
arch/riscv/net/bpf_jit.h
1115
else if (rvc_enabled() && is_creg(rd) && is_creg(rs1) && is_7b_uint(off) && !(off & 0x3))
arch/riscv/net/bpf_jit.h
1116
emitc(rvc_lw(rd, off, rs1), ctx);
arch/riscv/net/bpf_jit.h
1118
emit(rv_lw(rd, off, rs1), ctx);
arch/riscv/net/bpf_jit.h
1131
static inline void emit_sh2add(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1134
emit(rvzba_sh2add(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1138
emit_slli(rd, rs1, 2, ctx);
arch/riscv/net/bpf_jit.h
1139
emit_add(rd, rd, rs2, ctx);
arch/riscv/net/bpf_jit.h
1142
static inline void emit_sh3add(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1145
emit(rvzba_sh3add(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1149
emit_slli(rd, rs1, 3, ctx);
arch/riscv/net/bpf_jit.h
1150
emit_add(rd, rd, rs2, ctx);
arch/riscv/net/bpf_jit.h
1156
static inline void emit_addiw(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1158
if (rvc_enabled() && rd && rd == rs && is_6b_int(imm))
arch/riscv/net/bpf_jit.h
1159
emitc(rvc_addiw(rd, imm), ctx);
arch/riscv/net/bpf_jit.h
1161
emit(rv_addiw(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit.h
1164
static inline void emit_ld(u8 rd, s32 off, u8 rs1, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1166
if (rvc_enabled() && rs1 == RV_REG_SP && rd && is_9b_uint(off) && !(off & 0x7))
arch/riscv/net/bpf_jit.h
1167
emitc(rvc_ldsp(rd, off), ctx);
arch/riscv/net/bpf_jit.h
1168
else if (rvc_enabled() && is_creg(rd) && is_creg(rs1) && is_8b_uint(off) && !(off & 0x7))
arch/riscv/net/bpf_jit.h
1169
emitc(rvc_ld(rd, off, rs1), ctx);
arch/riscv/net/bpf_jit.h
1171
emit(rv_ld(rd, off, rs1), ctx);
arch/riscv/net/bpf_jit.h
1184
static inline void emit_subw(u8 rd, u8 rs1, u8 rs2, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1186
if (rvc_enabled() && is_creg(rd) && rd == rs1 && is_creg(rs2))
arch/riscv/net/bpf_jit.h
1187
emitc(rvc_subw(rd, rs2), ctx);
arch/riscv/net/bpf_jit.h
1189
emit(rv_subw(rd, rs1, rs2), ctx);
arch/riscv/net/bpf_jit.h
1192
static inline void emit_sextb(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1195
emit(rvzbb_sextb(rd, rs), ctx);
arch/riscv/net/bpf_jit.h
1199
emit_slli(rd, rs, 56, ctx);
arch/riscv/net/bpf_jit.h
1200
emit_srai(rd, rd, 56, ctx);
arch/riscv/net/bpf_jit.h
1203
static inline void emit_sexth(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1206
emit(rvzbb_sexth(rd, rs), ctx);
arch/riscv/net/bpf_jit.h
1210
emit_slli(rd, rs, 48, ctx);
arch/riscv/net/bpf_jit.h
1211
emit_srai(rd, rd, 48, ctx);
arch/riscv/net/bpf_jit.h
1214
static inline void emit_sextw(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1216
emit_addiw(rd, rs, 0, ctx);
arch/riscv/net/bpf_jit.h
1219
static inline void emit_zexth(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1222
emit(rvzbb_zexth(rd, rs), ctx);
arch/riscv/net/bpf_jit.h
1226
emit_slli(rd, rs, 48, ctx);
arch/riscv/net/bpf_jit.h
1227
emit_srli(rd, rd, 48, ctx);
arch/riscv/net/bpf_jit.h
1230
static inline void emit_zextw(u8 rd, u8 rs, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1233
emit(rvzba_zextw(rd, rs), ctx);
arch/riscv/net/bpf_jit.h
1237
emit_slli(rd, rs, 32, ctx);
arch/riscv/net/bpf_jit.h
1238
emit_srli(rd, rd, 32, ctx);
arch/riscv/net/bpf_jit.h
1241
static inline void emit_bswap(u8 rd, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1246
emit(rvzbb_rev8(rd, rd), ctx);
arch/riscv/net/bpf_jit.h
1248
emit_srli(rd, rd, bits, ctx);
arch/riscv/net/bpf_jit.h
1254
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1257
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1261
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1264
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1266
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1269
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1273
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1276
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1278
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1281
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1283
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1286
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1288
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1291
emit_srli(rd, rd, 8, ctx);
arch/riscv/net/bpf_jit.h
1293
emit_andi(RV_REG_T1, rd, 0xff, ctx);
arch/riscv/net/bpf_jit.h
1296
emit_mv(rd, RV_REG_T2, ctx);
arch/riscv/net/bpf_jit.h
1299
static inline void emit_cmpxchg(u8 rd, u8 rs, u8 r0, bool is64, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
1305
emit(is64 ? rvzacas_amocas_d(r0, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit.h
1306
rvzacas_amocas_w(r0, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit.h
1317
emit(is64 ? rv_lr_d(r0, 0, rd, 0, 0) :
arch/riscv/net/bpf_jit.h
1318
rv_lr_w(r0, 0, rd, 0, 0), ctx);
arch/riscv/net/bpf_jit.h
1321
emit(is64 ? rv_sc_d(RV_REG_T3, rs, rd, 0, 1) :
arch/riscv/net/bpf_jit.h
1322
rv_sc_w(RV_REG_T3, rs, rd, 0, 1), ctx);
arch/riscv/net/bpf_jit.h
237
static inline u32 rv_r_insn(u8 funct7, u8 rs2, u8 rs1, u8 funct3, u8 rd,
arch/riscv/net/bpf_jit.h
241
(rd << 7) | opcode;
arch/riscv/net/bpf_jit.h
244
static inline u32 rv_i_insn(u16 imm11_0, u8 rs1, u8 funct3, u8 rd, u8 opcode)
arch/riscv/net/bpf_jit.h
246
return (imm11_0 << 20) | (rs1 << 15) | (funct3 << 12) | (rd << 7) |
arch/riscv/net/bpf_jit.h
267
static inline u32 rv_u_insn(u32 imm31_12, u8 rd, u8 opcode)
arch/riscv/net/bpf_jit.h
269
return (imm31_12 << 12) | (rd << 7) | opcode;
arch/riscv/net/bpf_jit.h
272
static inline u32 rv_j_insn(u32 imm20_1, u8 rd, u8 opcode)
arch/riscv/net/bpf_jit.h
279
return (imm << 12) | (rd << 7) | opcode;
arch/riscv/net/bpf_jit.h
283
u8 funct3, u8 rd, u8 opcode)
arch/riscv/net/bpf_jit.h
287
return rv_r_insn(funct7, rs2, rs1, funct3, rd, opcode);
arch/riscv/net/bpf_jit.h
292
static inline u16 rv_cr_insn(u8 funct4, u8 rd, u8 rs2, u8 op)
arch/riscv/net/bpf_jit.h
294
return (funct4 << 12) | (rd << 7) | (rs2 << 2) | op;
arch/riscv/net/bpf_jit.h
297
static inline u16 rv_ci_insn(u8 funct3, u32 imm6, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
302
return (funct3 << 13) | (rd << 7) | op | imm;
arch/riscv/net/bpf_jit.h
310
static inline u16 rv_ciw_insn(u8 funct3, u32 uimm, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
312
return (funct3 << 13) | (uimm << 5) | ((rd & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
315
static inline u16 rv_cl_insn(u8 funct3, u32 imm_hi, u8 rs1, u32 imm_lo, u8 rd,
arch/riscv/net/bpf_jit.h
319
(imm_lo << 5) | ((rd & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
329
static inline u16 rv_ca_insn(u8 funct6, u8 rd, u8 funct2, u8 rs2, u8 op)
arch/riscv/net/bpf_jit.h
331
return (funct6 << 10) | ((rd & 0x7) << 7) | (funct2 << 5) |
arch/riscv/net/bpf_jit.h
335
static inline u16 rv_cb_insn(u8 funct3, u32 imm6, u8 funct2, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
340
return (funct3 << 13) | (funct2 << 10) | ((rd & 0x7) << 7) | op | imm;
arch/riscv/net/bpf_jit.h
345
static inline u32 rv_addi(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
347
return rv_i_insn(imm11_0, rs1, 0, rd, 0x13);
arch/riscv/net/bpf_jit.h
350
static inline u32 rv_andi(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
352
return rv_i_insn(imm11_0, rs1, 7, rd, 0x13);
arch/riscv/net/bpf_jit.h
355
static inline u32 rv_ori(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
357
return rv_i_insn(imm11_0, rs1, 6, rd, 0x13);
arch/riscv/net/bpf_jit.h
360
static inline u32 rv_xori(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
362
return rv_i_insn(imm11_0, rs1, 4, rd, 0x13);
arch/riscv/net/bpf_jit.h
365
static inline u32 rv_slli(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
367
return rv_i_insn(imm11_0, rs1, 1, rd, 0x13);
arch/riscv/net/bpf_jit.h
370
static inline u32 rv_srli(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
372
return rv_i_insn(imm11_0, rs1, 5, rd, 0x13);
arch/riscv/net/bpf_jit.h
375
static inline u32 rv_srai(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
377
return rv_i_insn(0x400 | imm11_0, rs1, 5, rd, 0x13);
arch/riscv/net/bpf_jit.h
380
static inline u32 rv_lui(u8 rd, u32 imm31_12)
arch/riscv/net/bpf_jit.h
382
return rv_u_insn(imm31_12, rd, 0x37);
arch/riscv/net/bpf_jit.h
385
static inline u32 rv_auipc(u8 rd, u32 imm31_12)
arch/riscv/net/bpf_jit.h
387
return rv_u_insn(imm31_12, rd, 0x17);
arch/riscv/net/bpf_jit.h
390
static inline u32 rv_add(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
392
return rv_r_insn(0, rs2, rs1, 0, rd, 0x33);
arch/riscv/net/bpf_jit.h
395
static inline u32 rv_sub(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
397
return rv_r_insn(0x20, rs2, rs1, 0, rd, 0x33);
arch/riscv/net/bpf_jit.h
400
static inline u32 rv_sltu(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
402
return rv_r_insn(0, rs2, rs1, 3, rd, 0x33);
arch/riscv/net/bpf_jit.h
405
static inline u32 rv_and(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
407
return rv_r_insn(0, rs2, rs1, 7, rd, 0x33);
arch/riscv/net/bpf_jit.h
410
static inline u32 rv_or(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
412
return rv_r_insn(0, rs2, rs1, 6, rd, 0x33);
arch/riscv/net/bpf_jit.h
415
static inline u32 rv_xor(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
417
return rv_r_insn(0, rs2, rs1, 4, rd, 0x33);
arch/riscv/net/bpf_jit.h
420
static inline u32 rv_sll(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
422
return rv_r_insn(0, rs2, rs1, 1, rd, 0x33);
arch/riscv/net/bpf_jit.h
425
static inline u32 rv_srl(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
427
return rv_r_insn(0, rs2, rs1, 5, rd, 0x33);
arch/riscv/net/bpf_jit.h
430
static inline u32 rv_sra(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
432
return rv_r_insn(0x20, rs2, rs1, 5, rd, 0x33);
arch/riscv/net/bpf_jit.h
435
static inline u32 rv_mul(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
437
return rv_r_insn(1, rs2, rs1, 0, rd, 0x33);
arch/riscv/net/bpf_jit.h
440
static inline u32 rv_mulhu(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
442
return rv_r_insn(1, rs2, rs1, 3, rd, 0x33);
arch/riscv/net/bpf_jit.h
445
static inline u32 rv_div(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
447
return rv_r_insn(1, rs2, rs1, 4, rd, 0x33);
arch/riscv/net/bpf_jit.h
450
static inline u32 rv_divu(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
452
return rv_r_insn(1, rs2, rs1, 5, rd, 0x33);
arch/riscv/net/bpf_jit.h
455
static inline u32 rv_rem(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
457
return rv_r_insn(1, rs2, rs1, 6, rd, 0x33);
arch/riscv/net/bpf_jit.h
460
static inline u32 rv_remu(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
462
return rv_r_insn(1, rs2, rs1, 7, rd, 0x33);
arch/riscv/net/bpf_jit.h
465
static inline u32 rv_jal(u8 rd, u32 imm20_1)
arch/riscv/net/bpf_jit.h
467
return rv_j_insn(imm20_1, rd, 0x6f);
arch/riscv/net/bpf_jit.h
470
static inline u32 rv_jalr(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
472
return rv_i_insn(imm11_0, rs1, 0, rd, 0x67);
arch/riscv/net/bpf_jit.h
525
static inline u32 rv_lb(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
527
return rv_i_insn(imm11_0, rs1, 0, rd, 0x03);
arch/riscv/net/bpf_jit.h
530
static inline u32 rv_lh(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
532
return rv_i_insn(imm11_0, rs1, 1, rd, 0x03);
arch/riscv/net/bpf_jit.h
535
static inline u32 rv_lw(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
537
return rv_i_insn(imm11_0, rs1, 2, rd, 0x03);
arch/riscv/net/bpf_jit.h
540
static inline u32 rv_lbu(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
542
return rv_i_insn(imm11_0, rs1, 4, rd, 0x03);
arch/riscv/net/bpf_jit.h
545
static inline u32 rv_lhu(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
547
return rv_i_insn(imm11_0, rs1, 5, rd, 0x03);
arch/riscv/net/bpf_jit.h
565
static inline u32 rv_amoadd_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
567
return rv_amo_insn(0, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
570
static inline u32 rv_amoand_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
572
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
575
static inline u32 rv_amoor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
577
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
580
static inline u32 rv_amoxor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
582
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
585
static inline u32 rv_amoswap_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
587
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
590
static inline u32 rv_lr_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
592
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
595
static inline u32 rv_sc_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
597
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
629
static inline u16 rvc_addi4spn(u8 rd, u32 imm10)
arch/riscv/net/bpf_jit.h
635
return rv_ciw_insn(0x0, imm, rd, 0x0);
arch/riscv/net/bpf_jit.h
638
static inline u16 rvc_lw(u8 rd, u32 imm7, u8 rs1)
arch/riscv/net/bpf_jit.h
644
return rv_cl_insn(0x2, imm_hi, rs1, imm_lo, rd, 0x0);
arch/riscv/net/bpf_jit.h
656
static inline u16 rvc_addi(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
658
return rv_ci_insn(0, imm6, rd, 0x1);
arch/riscv/net/bpf_jit.h
661
static inline u16 rvc_li(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
663
return rv_ci_insn(0x2, imm6, rd, 0x1);
arch/riscv/net/bpf_jit.h
675
static inline u16 rvc_lui(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
677
return rv_ci_insn(0x3, imm6, rd, 0x1);
arch/riscv/net/bpf_jit.h
680
static inline u16 rvc_srli(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
682
return rv_cb_insn(0x4, imm6, 0, rd, 0x1);
arch/riscv/net/bpf_jit.h
685
static inline u16 rvc_srai(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
687
return rv_cb_insn(0x4, imm6, 0x1, rd, 0x1);
arch/riscv/net/bpf_jit.h
690
static inline u16 rvc_andi(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
692
return rv_cb_insn(0x4, imm6, 0x2, rd, 0x1);
arch/riscv/net/bpf_jit.h
695
static inline u16 rvc_sub(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
697
return rv_ca_insn(0x23, rd, 0, rs, 0x1);
arch/riscv/net/bpf_jit.h
700
static inline u16 rvc_xor(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
702
return rv_ca_insn(0x23, rd, 0x1, rs, 0x1);
arch/riscv/net/bpf_jit.h
705
static inline u16 rvc_or(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
707
return rv_ca_insn(0x23, rd, 0x2, rs, 0x1);
arch/riscv/net/bpf_jit.h
710
static inline u16 rvc_and(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
712
return rv_ca_insn(0x23, rd, 0x3, rs, 0x1);
arch/riscv/net/bpf_jit.h
715
static inline u16 rvc_slli(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
717
return rv_ci_insn(0, imm6, rd, 0x2);
arch/riscv/net/bpf_jit.h
720
static inline u16 rvc_lwsp(u8 rd, u32 imm8)
arch/riscv/net/bpf_jit.h
725
return rv_ci_insn(0x2, imm, rd, 0x2);
arch/riscv/net/bpf_jit.h
733
static inline u16 rvc_mv(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
735
return rv_cr_insn(0x8, rd, rs, 0x2);
arch/riscv/net/bpf_jit.h
743
static inline u16 rvc_add(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
745
return rv_cr_insn(0x9, rd, rs, 0x2);
arch/riscv/net/bpf_jit.h
757
static inline u32 rvzacas_amocas_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
759
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
762
static inline u32 rvzacas_amocas_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
764
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
768
static inline u32 rvzba_sh2add(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
770
return rv_r_insn(0x10, rs2, rs1, 0x4, rd, 0x33);
arch/riscv/net/bpf_jit.h
773
static inline u32 rvzba_sh3add(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
775
return rv_r_insn(0x10, rs2, rs1, 0x6, rd, 0x33);
arch/riscv/net/bpf_jit.h
779
static inline u32 rvzbb_sextb(u8 rd, u8 rs1)
arch/riscv/net/bpf_jit.h
781
return rv_i_insn(0x604, rs1, 1, rd, 0x13);
arch/riscv/net/bpf_jit.h
784
static inline u32 rvzbb_sexth(u8 rd, u8 rs1)
arch/riscv/net/bpf_jit.h
786
return rv_i_insn(0x605, rs1, 1, rd, 0x13);
arch/riscv/net/bpf_jit.h
789
static inline u32 rvzbb_zexth(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
792
return rv_i_insn(0x80, rs, 4, rd, 0x3b);
arch/riscv/net/bpf_jit.h
794
return rv_i_insn(0x80, rs, 4, rd, 0x33);
arch/riscv/net/bpf_jit.h
797
static inline u32 rvzbb_rev8(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
800
return rv_i_insn(0x6b8, rs, 5, rd, 0x13);
arch/riscv/net/bpf_jit.h
802
return rv_i_insn(0x698, rs, 5, rd, 0x13);
arch/riscv/net/bpf_jit.h
814
static inline u32 rv_addiw(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
816
return rv_i_insn(imm11_0, rs1, 0, rd, 0x1b);
arch/riscv/net/bpf_jit.h
819
static inline u32 rv_slliw(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
821
return rv_i_insn(imm11_0, rs1, 1, rd, 0x1b);
arch/riscv/net/bpf_jit.h
824
static inline u32 rv_srliw(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
826
return rv_i_insn(imm11_0, rs1, 5, rd, 0x1b);
arch/riscv/net/bpf_jit.h
829
static inline u32 rv_sraiw(u8 rd, u8 rs1, u16 imm11_0)
arch/riscv/net/bpf_jit.h
831
return rv_i_insn(0x400 | imm11_0, rs1, 5, rd, 0x1b);
arch/riscv/net/bpf_jit.h
834
static inline u32 rv_addw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
836
return rv_r_insn(0, rs2, rs1, 0, rd, 0x3b);
arch/riscv/net/bpf_jit.h
839
static inline u32 rv_subw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
841
return rv_r_insn(0x20, rs2, rs1, 0, rd, 0x3b);
arch/riscv/net/bpf_jit.h
844
static inline u32 rv_sllw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
846
return rv_r_insn(0, rs2, rs1, 1, rd, 0x3b);
arch/riscv/net/bpf_jit.h
849
static inline u32 rv_srlw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
851
return rv_r_insn(0, rs2, rs1, 5, rd, 0x3b);
arch/riscv/net/bpf_jit.h
854
static inline u32 rv_sraw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
856
return rv_r_insn(0x20, rs2, rs1, 5, rd, 0x3b);
arch/riscv/net/bpf_jit.h
859
static inline u32 rv_mulw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
861
return rv_r_insn(1, rs2, rs1, 0, rd, 0x3b);
arch/riscv/net/bpf_jit.h
864
static inline u32 rv_divw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
866
return rv_r_insn(1, rs2, rs1, 4, rd, 0x3b);
arch/riscv/net/bpf_jit.h
869
static inline u32 rv_divuw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
871
return rv_r_insn(1, rs2, rs1, 5, rd, 0x3b);
arch/riscv/net/bpf_jit.h
874
static inline u32 rv_remw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
876
return rv_r_insn(1, rs2, rs1, 6, rd, 0x3b);
arch/riscv/net/bpf_jit.h
879
static inline u32 rv_remuw(u8 rd, u8 rs1, u8 rs2)
arch/riscv/net/bpf_jit.h
881
return rv_r_insn(1, rs2, rs1, 7, rd, 0x3b);
arch/riscv/net/bpf_jit.h
884
static inline u32 rv_ld(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
886
return rv_i_insn(imm11_0, rs1, 3, rd, 0x03);
arch/riscv/net/bpf_jit.h
889
static inline u32 rv_lwu(u8 rd, u16 imm11_0, u8 rs1)
arch/riscv/net/bpf_jit.h
891
return rv_i_insn(imm11_0, rs1, 6, rd, 0x03);
arch/riscv/net/bpf_jit.h
899
static inline u32 rv_amoadd_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
901
return rv_amo_insn(0, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
904
static inline u32 rv_amoand_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
906
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
909
static inline u32 rv_amoor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
911
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
914
static inline u32 rv_amoxor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
916
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
919
static inline u32 rv_amoswap_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
921
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
924
static inline u32 rv_lr_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
926
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
929
static inline u32 rv_sc_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
931
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
936
static inline u16 rvc_ld(u8 rd, u32 imm8, u8 rs1)
arch/riscv/net/bpf_jit.h
942
return rv_cl_insn(0x3, imm_hi, rs1, imm_lo, rd, 0x0);
arch/riscv/net/bpf_jit.h
954
static inline u16 rvc_subw(u8 rd, u8 rs)
arch/riscv/net/bpf_jit.h
956
return rv_ca_insn(0x27, rd, 0, rs, 0x1);
arch/riscv/net/bpf_jit.h
959
static inline u16 rvc_addiw(u8 rd, u32 imm6)
arch/riscv/net/bpf_jit.h
961
return rv_ci_insn(0x1, imm6, rd, 0x1);
arch/riscv/net/bpf_jit.h
964
static inline u16 rvc_ldsp(u8 rd, u32 imm9)
arch/riscv/net/bpf_jit.h
969
return rv_ci_insn(0x3, imm, rd, 0x2);
arch/riscv/net/bpf_jit.h
982
static inline u32 rvzba_zextw(u8 rd, u8 rs1)
arch/riscv/net/bpf_jit.h
985
return rv_r_insn(0x04, RV_REG_ZERO, rs1, 0, rd, 0x3b);
arch/riscv/net/bpf_jit.h
992
static inline void emit_jalr(u8 rd, u8 rs, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit.h
994
if (rvc_enabled() && rd == RV_REG_RA && rs && !imm)
arch/riscv/net/bpf_jit.h
996
else if (rvc_enabled() && !rd && rs && !imm)
arch/riscv/net/bpf_jit.h
999
emit(rv_jalr(rd, rs, imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
1073
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
1077
emit(rv_slli(lo(rd), lo(rd), 16), ctx);
arch/riscv/net/bpf_jit_comp32.c
1078
emit(rv_srli(lo(rd), lo(rd), 16), ctx);
arch/riscv/net/bpf_jit_comp32.c
1082
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
1092
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
1098
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
1102
emit_rev16(lo(rd), ctx);
arch/riscv/net/bpf_jit_comp32.c
1104
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
1107
emit_rev32(lo(rd), ctx);
arch/riscv/net/bpf_jit_comp32.c
1109
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
111
static void emit_imm(const s8 rd, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
1113
emit(rv_addi(RV_REG_T0, lo(rd), 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
1114
emit(rv_addi(lo(rd), hi(rd), 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
1115
emit(rv_addi(hi(rd), RV_REG_T0, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
1118
emit_rev32(lo(rd), ctx);
arch/riscv/net/bpf_jit_comp32.c
1119
emit_rev32(hi(rd), ctx);
arch/riscv/net/bpf_jit_comp32.c
1126
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
117
emit(rv_lui(rd, upper), ctx);
arch/riscv/net/bpf_jit_comp32.c
118
emit(rv_addi(rd, rd, lower), ctx);
arch/riscv/net/bpf_jit_comp32.c
120
emit(rv_addi(rd, RV_REG_ZERO, lower), ctx);
arch/riscv/net/bpf_jit_comp32.c
1236
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
1238
emit_imm64(rd, imm_hi, imm_lo, ctx);
arch/riscv/net/bpf_jit_comp32.c
1239
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
124
static void emit_imm32(const s8 *rd, s32 imm, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
127
emit_imm(lo(rd), imm, ctx);
arch/riscv/net/bpf_jit_comp32.c
131
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
133
emit(rv_addi(hi(rd), RV_REG_ZERO, -1), ctx);
arch/riscv/net/bpf_jit_comp32.c
136
static void emit_imm64(const s8 *rd, s32 imm_hi, s32 imm_lo,
arch/riscv/net/bpf_jit_comp32.c
139
emit_imm(lo(rd), imm_lo, ctx);
arch/riscv/net/bpf_jit_comp32.c
140
emit_imm(hi(rd), imm_hi, ctx);
arch/riscv/net/bpf_jit_comp32.c
227
static void emit_jump_and_link(u8 rd, s32 rvoff, bool force_jalr,
arch/riscv/net/bpf_jit_comp32.c
233
emit(rv_jal(rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
240
emit(rv_jalr(rd, RV_REG_T1, lower), ctx);
arch/riscv/net/bpf_jit_comp32.c
247
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
251
emit_imm32(rd, imm, ctx);
arch/riscv/net/bpf_jit_comp32.c
255
emit(rv_andi(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
258
emit(rv_and(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
261
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
265
emit(rv_ori(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
268
emit(rv_or(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
271
emit(rv_ori(hi(rd), RV_REG_ZERO, -1), ctx);
arch/riscv/net/bpf_jit_comp32.c
275
emit(rv_xori(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
278
emit(rv_xor(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
281
emit(rv_xori(hi(rd), hi(rd), -1), ctx);
arch/riscv/net/bpf_jit_comp32.c
285
emit(rv_slli(hi(rd), lo(rd), imm - 32), ctx);
arch/riscv/net/bpf_jit_comp32.c
286
emit(rv_addi(lo(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
290
emit(rv_srli(RV_REG_T0, lo(rd), 32 - imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
291
emit(rv_slli(hi(rd), hi(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
292
emit(rv_or(hi(rd), RV_REG_T0, hi(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
293
emit(rv_slli(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
298
emit(rv_srli(lo(rd), hi(rd), imm - 32), ctx);
arch/riscv/net/bpf_jit_comp32.c
299
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
303
emit(rv_slli(RV_REG_T0, hi(rd), 32 - imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
304
emit(rv_srli(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
305
emit(rv_or(lo(rd), RV_REG_T0, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
306
emit(rv_srli(hi(rd), hi(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
311
emit(rv_srai(lo(rd), hi(rd), imm - 32), ctx);
arch/riscv/net/bpf_jit_comp32.c
312
emit(rv_srai(hi(rd), hi(rd), 31), ctx);
arch/riscv/net/bpf_jit_comp32.c
316
emit(rv_slli(RV_REG_T0, hi(rd), 32 - imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
317
emit(rv_srli(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
318
emit(rv_or(lo(rd), RV_REG_T0, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
319
emit(rv_srai(hi(rd), hi(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
324
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
331
const s8 *rd = bpf_get_reg32(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
335
emit_imm(lo(rd), imm, ctx);
arch/riscv/net/bpf_jit_comp32.c
339
emit(rv_addi(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
342
emit(rv_add(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
347
emit(rv_addi(lo(rd), lo(rd), -imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
350
emit(rv_sub(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
355
emit(rv_andi(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
358
emit(rv_and(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
363
emit(rv_ori(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
366
emit(rv_or(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
371
emit(rv_xori(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
374
emit(rv_xor(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
379
emit(rv_slli(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
382
emit(rv_sll(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
387
emit(rv_srli(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
390
emit(rv_srl(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
395
emit(rv_srai(lo(rd), lo(rd), imm), ctx);
arch/riscv/net/bpf_jit_comp32.c
398
emit(rv_sra(lo(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
403
bpf_put_reg32(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
411
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
416
emit(rv_addi(lo(rd), lo(rs), 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
417
emit(rv_addi(hi(rd), hi(rs), 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
420
if (rd == rs) {
arch/riscv/net/bpf_jit_comp32.c
421
emit(rv_srli(RV_REG_T0, lo(rd), 31), ctx);
arch/riscv/net/bpf_jit_comp32.c
422
emit(rv_slli(hi(rd), hi(rd), 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
423
emit(rv_or(hi(rd), RV_REG_T0, hi(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
424
emit(rv_slli(lo(rd), lo(rd), 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
426
emit(rv_add(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
427
emit(rv_sltu(RV_REG_T0, lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
428
emit(rv_add(hi(rd), hi(rd), hi(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
429
emit(rv_add(hi(rd), hi(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
433
emit(rv_sub(RV_REG_T1, hi(rd), hi(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
434
emit(rv_sltu(RV_REG_T0, lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
435
emit(rv_sub(hi(rd), RV_REG_T1, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
436
emit(rv_sub(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
439
emit(rv_and(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
440
emit(rv_and(hi(rd), hi(rd), hi(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
443
emit(rv_or(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
444
emit(rv_or(hi(rd), hi(rd), hi(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
447
emit(rv_xor(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
448
emit(rv_xor(hi(rd), hi(rd), hi(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
451
emit(rv_mul(RV_REG_T0, hi(rs), lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
452
emit(rv_mul(hi(rd), hi(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
453
emit(rv_mulhu(RV_REG_T1, lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
454
emit(rv_add(hi(rd), hi(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
455
emit(rv_mul(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
456
emit(rv_add(hi(rd), hi(rd), RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp32.c
461
emit(rv_sll(hi(rd), lo(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
462
emit(rv_addi(lo(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
465
emit(rv_srli(RV_REG_T0, lo(rd), 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
468
emit(rv_sll(hi(rd), hi(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
469
emit(rv_or(hi(rd), RV_REG_T0, hi(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
470
emit(rv_sll(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
475
emit(rv_srl(lo(rd), hi(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
476
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
479
emit(rv_slli(RV_REG_T0, hi(rd), 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
482
emit(rv_srl(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
483
emit(rv_or(lo(rd), RV_REG_T0, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
484
emit(rv_srl(hi(rd), hi(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
489
emit(rv_sra(lo(rd), hi(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
490
emit(rv_srai(hi(rd), hi(rd), 31), ctx);
arch/riscv/net/bpf_jit_comp32.c
493
emit(rv_slli(RV_REG_T0, hi(rd), 1), ctx);
arch/riscv/net/bpf_jit_comp32.c
496
emit(rv_srl(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
497
emit(rv_or(lo(rd), RV_REG_T0, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
498
emit(rv_sra(hi(rd), hi(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
501
emit(rv_sub(lo(rd), RV_REG_ZERO, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
502
emit(rv_sltu(RV_REG_T0, RV_REG_ZERO, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
503
emit(rv_sub(hi(rd), RV_REG_ZERO, hi(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
504
emit(rv_sub(hi(rd), hi(rd), RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
508
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
516
const s8 *rd = bpf_get_reg32(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
521
emit(rv_addi(lo(rd), lo(rs), 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
524
emit(rv_add(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
527
emit(rv_sub(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
530
emit(rv_and(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
533
emit(rv_or(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
536
emit(rv_xor(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
539
emit(rv_mul(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
542
emit(rv_divu(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
545
emit(rv_remu(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
548
emit(rv_sll(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
551
emit(rv_srl(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
554
emit(rv_sra(lo(rd), lo(rd), lo(rs)), ctx);
arch/riscv/net/bpf_jit_comp32.c
557
emit(rv_sub(lo(rd), RV_REG_ZERO, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
561
bpf_put_reg32(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
652
static int emit_bcc(u8 op, u8 rd, u8 rs, int rvoff, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
678
emit(rv_beq(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
681
emit(rv_bgtu(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
684
emit(rv_bltu(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
687
emit(rv_bgeu(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
690
emit(rv_bleu(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
693
emit(rv_bne(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
696
emit(rv_bgt(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
699
emit(rv_blt(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
702
emit(rv_bge(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
705
emit(rv_ble(rd, rs, off), ctx);
arch/riscv/net/bpf_jit_comp32.c
708
emit(rv_and(RV_REG_T0, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp32.c
840
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
848
emit(rv_lbu(lo(rd), 0, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
850
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
853
emit(rv_lhu(lo(rd), 0, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
855
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
858
emit(rv_lw(lo(rd), 0, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
860
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
863
emit(rv_lw(lo(rd), 0, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
864
emit(rv_lw(hi(rd), 4, RV_REG_T0), ctx);
arch/riscv/net/bpf_jit_comp32.c
868
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp32.c
878
const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
885
emit(rv_add(RV_REG_T0, RV_REG_T0, lo(rd)), ctx);
arch/riscv/net/bpf_jit_comp32.c
914
static void emit_rev16(const s8 rd, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
916
emit(rv_slli(rd, rd, 16), ctx);
arch/riscv/net/bpf_jit_comp32.c
917
emit(rv_slli(RV_REG_T1, rd, 8), ctx);
arch/riscv/net/bpf_jit_comp32.c
918
emit(rv_srli(rd, rd, 8), ctx);
arch/riscv/net/bpf_jit_comp32.c
919
emit(rv_add(RV_REG_T1, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp32.c
920
emit(rv_srli(rd, RV_REG_T1, 16), ctx);
arch/riscv/net/bpf_jit_comp32.c
923
static void emit_rev32(const s8 rd, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
926
emit(rv_andi(RV_REG_T0, rd, 255), ctx);
arch/riscv/net/bpf_jit_comp32.c
929
emit(rv_srli(rd, rd, 8), ctx);
arch/riscv/net/bpf_jit_comp32.c
930
emit(rv_andi(RV_REG_T0, rd, 255), ctx);
arch/riscv/net/bpf_jit_comp32.c
933
emit(rv_srli(rd, rd, 8), ctx);
arch/riscv/net/bpf_jit_comp32.c
934
emit(rv_andi(RV_REG_T0, rd, 255), ctx);
arch/riscv/net/bpf_jit_comp32.c
937
emit(rv_srli(rd, rd, 8), ctx);
arch/riscv/net/bpf_jit_comp32.c
938
emit(rv_andi(RV_REG_T0, rd, 255), ctx);
arch/riscv/net/bpf_jit_comp32.c
940
emit(rv_addi(rd, RV_REG_T1, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
945
const s8 *rd;
arch/riscv/net/bpf_jit_comp32.c
948
rd = bpf_get_reg64(dst, tmp1, ctx);
arch/riscv/net/bpf_jit_comp32.c
949
emit(rv_addi(hi(rd), RV_REG_ZERO, 0), ctx);
arch/riscv/net/bpf_jit_comp32.c
950
bpf_put_reg64(dst, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1376
u8 rd = -1, rs = -1, code = insn->code;
arch/riscv/net/bpf_jit_comp64.c
1380
init_regs(&rd, &rs, insn, ctx);
arch/riscv/net/bpf_jit_comp64.c
1389
emit_imm(rd, (ctx->user_vm_start >> 32) << 32, ctx);
arch/riscv/net/bpf_jit_comp64.c
1391
emit_or(RV_REG_T1, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1392
emit_mv(rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1395
if (rd != rs)
arch/riscv/net/bpf_jit_comp64.c
1396
emit_mv(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1408
emit_add(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1413
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1418
emit_mv(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1421
emit_sextb(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1424
emit_sexth(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1427
emit_sextw(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1431
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1437
emit_add(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1439
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1444
emit_sub(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1446
emit_subw(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1449
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1453
emit_and(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1455
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1459
emit_or(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1461
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1465
emit_xor(rd, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1467
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1471
emit(is64 ? rv_mul(rd, rd, rs) : rv_mulw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1473
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1478
emit(is64 ? rv_div(rd, rd, rs) : rv_divw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1480
emit(is64 ? rv_divu(rd, rd, rs) : rv_divuw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1482
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1487
emit(is64 ? rv_rem(rd, rd, rs) : rv_remw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1489
emit(is64 ? rv_remu(rd, rd, rs) : rv_remuw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1491
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1495
emit(is64 ? rv_sll(rd, rd, rs) : rv_sllw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1497
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1501
emit(is64 ? rv_srl(rd, rd, rs) : rv_srlw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1503
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1507
emit(is64 ? rv_sra(rd, rd, rs) : rv_sraw(rd, rd, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
1509
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1515
emit_sub(rd, RV_REG_ZERO, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1517
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1524
emit_zexth(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1528
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
153
static void emit_sextw_alt(u8 *rd, u8 ra, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
1537
emit_bswap(rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1543
emit_imm(rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1545
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
155
emit_sextw(ra, *rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1552
emit_addi(rd, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1555
emit_add(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1558
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
156
*rd = ra;
arch/riscv/net/bpf_jit_comp64.c
1563
emit_addi(rd, rd, -imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1566
emit_sub(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1569
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1574
emit_andi(rd, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1577
emit_and(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1580
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1585
emit(rv_ori(rd, rd, imm), ctx);
arch/riscv/net/bpf_jit_comp64.c
1588
emit_or(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
159
static void emit_zextw_alt(u8 *rd, u8 ra, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
1591
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1596
emit(rv_xori(rd, rd, imm), ctx);
arch/riscv/net/bpf_jit_comp64.c
1599
emit_xor(rd, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
1602
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1607
emit(is64 ? rv_mul(rd, rd, RV_REG_T1) :
arch/riscv/net/bpf_jit_comp64.c
1608
rv_mulw(rd, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp64.c
161
emit_zextw(ra, *rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1610
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1616
emit(is64 ? rv_div(rd, rd, RV_REG_T1) :
arch/riscv/net/bpf_jit_comp64.c
1617
rv_divw(rd, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp64.c
1619
emit(is64 ? rv_divu(rd, rd, RV_REG_T1) :
arch/riscv/net/bpf_jit_comp64.c
162
*rd = ra;
arch/riscv/net/bpf_jit_comp64.c
1620
rv_divuw(rd, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp64.c
1622
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1628
emit(is64 ? rv_rem(rd, rd, RV_REG_T1) :
arch/riscv/net/bpf_jit_comp64.c
1629
rv_remw(rd, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp64.c
1631
emit(is64 ? rv_remu(rd, rd, RV_REG_T1) :
arch/riscv/net/bpf_jit_comp64.c
1632
rv_remuw(rd, rd, RV_REG_T1), ctx);
arch/riscv/net/bpf_jit_comp64.c
1634
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1638
emit_slli(rd, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1641
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1646
emit_srli(rd, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1648
emit(rv_srliw(rd, rd, imm), ctx);
arch/riscv/net/bpf_jit_comp64.c
1651
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1656
emit_srai(rd, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1658
emit(rv_sraiw(rd, rd, imm), ctx);
arch/riscv/net/bpf_jit_comp64.c
166
static int emit_addr(u8 rd, u64 addr, bool extra_pass, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
1661
emit_zextw(rd, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1704
emit_sextw_alt(&rd, RV_REG_T2, ctx);
arch/riscv/net/bpf_jit_comp64.c
1707
emit_zextw_alt(&rd, RV_REG_T2, ctx);
arch/riscv/net/bpf_jit_comp64.c
1718
emit_and(RV_REG_T1, rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
1721
emit_branch(BPF_OP(code), rd, rs, rvoff, ctx);
arch/riscv/net/bpf_jit_comp64.c
1753
emit_sextw_alt(&rd, RV_REG_T2, ctx);
arch/riscv/net/bpf_jit_comp64.c
1756
emit_zextw_alt(&rd, RV_REG_T2, ctx);
arch/riscv/net/bpf_jit_comp64.c
1765
emit_branch(BPF_OP(code), rd, rs, rvoff, ctx);
arch/riscv/net/bpf_jit_comp64.c
1773
emit_andi(RV_REG_T1, rd, imm, ctx);
arch/riscv/net/bpf_jit_comp64.c
1776
emit_and(RV_REG_T1, rd, RV_REG_T1, ctx);
arch/riscv/net/bpf_jit_comp64.c
182
emit(rv_auipc(rd, upper), ctx);
arch/riscv/net/bpf_jit_comp64.c
183
emit(rv_addi(rd, rd, lower), ctx);
arch/riscv/net/bpf_jit_comp64.c
1867
ret = emit_addr(rd, imm64, extra_pass, ctx);
arch/riscv/net/bpf_jit_comp64.c
1871
emit_imm(rd, imm64, ctx);
arch/riscv/net/bpf_jit_comp64.c
188
static void emit_imm(u8 rd, s64 val, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
1909
emit_ldx(rd, off, rs, BPF_SIZE(code), sign_ext, ctx);
arch/riscv/net/bpf_jit_comp64.c
1911
ret = add_exception_handler(insn, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
1935
emit_add(RV_REG_T3, rd, RV_REG_ARENA, ctx);
arch/riscv/net/bpf_jit_comp64.c
1936
rd = RV_REG_T3;
arch/riscv/net/bpf_jit_comp64.c
1939
emit_st(rd, off, imm, BPF_SIZE(code), ctx);
arch/riscv/net/bpf_jit_comp64.c
1957
emit_add(RV_REG_T2, rd, RV_REG_ARENA, ctx);
arch/riscv/net/bpf_jit_comp64.c
1958
rd = RV_REG_T2;
arch/riscv/net/bpf_jit_comp64.c
1961
emit_stx(rd, off, rs, BPF_SIZE(code), ctx);
arch/riscv/net/bpf_jit_comp64.c
1978
ret = emit_atomic_ld_st(rd, rs, insn, ctx);
arch/riscv/net/bpf_jit_comp64.c
1980
ret = emit_atomic_rmw(rd, rs, insn, ctx);
arch/riscv/net/bpf_jit_comp64.c
209
emit_lui(rd, upper, ctx);
arch/riscv/net/bpf_jit_comp64.c
212
emit_li(rd, lower, ctx);
arch/riscv/net/bpf_jit_comp64.c
216
emit_addiw(rd, rd, lower, ctx);
arch/riscv/net/bpf_jit_comp64.c
224
emit_imm(rd, upper, ctx);
arch/riscv/net/bpf_jit_comp64.c
226
emit_slli(rd, rd, shift, ctx);
arch/riscv/net/bpf_jit_comp64.c
228
emit_addi(rd, rd, lower, ctx);
arch/riscv/net/bpf_jit_comp64.c
280
static void emit_bcc(u8 cond, u8 rd, u8 rs, int rvoff,
arch/riscv/net/bpf_jit_comp64.c
285
emit(rv_beq(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
288
emit(rv_bltu(rs, rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
291
emit(rv_bltu(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
294
emit(rv_bgeu(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
297
emit(rv_bgeu(rs, rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
300
emit(rv_bne(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
303
emit(rv_blt(rs, rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
306
emit(rv_blt(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
309
emit(rv_bge(rd, rs, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
312
emit(rv_bge(rs, rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
316
static void emit_branch(u8 cond, u8 rd, u8 rs, int rvoff,
arch/riscv/net/bpf_jit_comp64.c
322
emit_bcc(cond, rd, rs, rvoff, ctx);
arch/riscv/net/bpf_jit_comp64.c
339
emit_bcc(cond, rd, rs, 8, ctx);
arch/riscv/net/bpf_jit_comp64.c
350
emit_bcc(cond, rd, rs, 12, ctx);
arch/riscv/net/bpf_jit_comp64.c
406
static void init_regs(u8 *rd, u8 *rs, const struct bpf_insn *insn,
arch/riscv/net/bpf_jit_comp64.c
418
*rd = bpf_to_rv_reg(insn->dst_reg, ctx);
arch/riscv/net/bpf_jit_comp64.c
427
static int emit_jump_and_link(u8 rd, s64 rvoff, bool fixed_addr,
arch/riscv/net/bpf_jit_comp64.c
433
emit(rv_jal(rd, rvoff >> 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
439
emit(rv_jalr(rd, RV_REG_T1, lower), ctx);
arch/riscv/net/bpf_jit_comp64.c
476
static void emit_ldx_insn(u8 rd, s16 off, u8 rs, u8 size, bool sign_ext,
arch/riscv/net/bpf_jit_comp64.c
481
emit(sign_ext ? rv_lb(rd, off, rs) : rv_lbu(rd, off, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
484
emit(sign_ext ? rv_lh(rd, off, rs) : rv_lhu(rd, off, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
487
emit(sign_ext ? rv_lw(rd, off, rs) : rv_lwu(rd, off, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
490
emit_ld(rd, off, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
496
static void emit_stx_insn(u8 rd, s16 off, u8 rs, u8 size, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
500
emit(rv_sb(rd, off, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
503
emit(rv_sh(rd, off, rs), ctx);
arch/riscv/net/bpf_jit_comp64.c
506
emit_sw(rd, off, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
509
emit_sd(rd, off, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
514
static void emit_ldx(u8 rd, s16 off, u8 rs, u8 size, bool sign_ext,
arch/riscv/net/bpf_jit_comp64.c
519
emit_ldx_insn(rd, off, rs, size, sign_ext, ctx);
arch/riscv/net/bpf_jit_comp64.c
527
emit_ldx_insn(rd, 0, RV_REG_T1, size, sign_ext, ctx);
arch/riscv/net/bpf_jit_comp64.c
531
static void emit_st(u8 rd, s16 off, s32 imm, u8 size, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
536
emit_stx_insn(rd, off, RV_REG_T1, size, ctx);
arch/riscv/net/bpf_jit_comp64.c
542
emit_add(RV_REG_T2, RV_REG_T2, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
548
static void emit_stx(u8 rd, s16 off, u8 rs, u8 size, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
552
emit_stx_insn(rd, off, rs, size, ctx);
arch/riscv/net/bpf_jit_comp64.c
558
emit_add(RV_REG_T1, RV_REG_T1, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
564
static int emit_atomic_ld_st(u8 rd, u8 rs, const struct bpf_insn *insn,
arch/riscv/net/bpf_jit_comp64.c
579
emit_ldx(rd, off, rs, BPF_SIZE(code), false, ctx);
arch/riscv/net/bpf_jit_comp64.c
591
emit_add(RV_REG_T2, rd, RV_REG_ARENA, ctx);
arch/riscv/net/bpf_jit_comp64.c
592
rd = RV_REG_T2;
arch/riscv/net/bpf_jit_comp64.c
596
emit_stx(rd, off, rs, BPF_SIZE(code), ctx);
arch/riscv/net/bpf_jit_comp64.c
606
static int emit_atomic_rmw(u8 rd, u8 rs, const struct bpf_insn *insn,
arch/riscv/net/bpf_jit_comp64.c
621
emit_addi(RV_REG_T1, rd, off, ctx);
arch/riscv/net/bpf_jit_comp64.c
624
emit_add(RV_REG_T1, RV_REG_T1, rd, ctx);
arch/riscv/net/bpf_jit_comp64.c
626
rd = RV_REG_T1;
arch/riscv/net/bpf_jit_comp64.c
630
emit_add(RV_REG_T1, rd, RV_REG_ARENA, ctx);
arch/riscv/net/bpf_jit_comp64.c
631
rd = RV_REG_T1;
arch/riscv/net/bpf_jit_comp64.c
638
emit(is64 ? rv_amoadd_d(RV_REG_ZERO, rs, rd, 0, 0) :
arch/riscv/net/bpf_jit_comp64.c
639
rv_amoadd_w(RV_REG_ZERO, rs, rd, 0, 0), ctx);
arch/riscv/net/bpf_jit_comp64.c
644
emit(is64 ? rv_amoand_d(RV_REG_ZERO, rs, rd, 0, 0) :
arch/riscv/net/bpf_jit_comp64.c
645
rv_amoand_w(RV_REG_ZERO, rs, rd, 0, 0), ctx);
arch/riscv/net/bpf_jit_comp64.c
650
emit(is64 ? rv_amoor_d(RV_REG_ZERO, rs, rd, 0, 0) :
arch/riscv/net/bpf_jit_comp64.c
651
rv_amoor_w(RV_REG_ZERO, rs, rd, 0, 0), ctx);
arch/riscv/net/bpf_jit_comp64.c
656
emit(is64 ? rv_amoxor_d(RV_REG_ZERO, rs, rd, 0, 0) :
arch/riscv/net/bpf_jit_comp64.c
657
rv_amoxor_w(RV_REG_ZERO, rs, rd, 0, 0), ctx);
arch/riscv/net/bpf_jit_comp64.c
663
emit(is64 ? rv_amoadd_d(rs, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit_comp64.c
664
rv_amoadd_w(rs, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
671
emit(is64 ? rv_amoand_d(rs, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit_comp64.c
672
rv_amoand_w(rs, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
679
emit(is64 ? rv_amoor_d(rs, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit_comp64.c
680
rv_amoor_w(rs, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
687
emit(is64 ? rv_amoxor_d(rs, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit_comp64.c
688
rv_amoxor_w(rs, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
696
emit(is64 ? rv_amoswap_d(rs, rs, rd, 1, 1) :
arch/riscv/net/bpf_jit_comp64.c
697
rv_amoswap_w(rs, rs, rd, 1, 1), ctx);
arch/riscv/net/bpf_jit_comp64.c
704
emit_cmpxchg(rd, rs, regmap[BPF_REG_0], is64, ctx);
arch/riscv/net/bpf_jit_comp64.c
717
static int sign_extend(u8 rd, u8 rs, u8 sz, bool sign, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp64.c
720
if (rd != rs)
arch/riscv/net/bpf_jit_comp64.c
721
emit_mv(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
727
emit_sextb(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
730
emit_sexth(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
733
emit_sextw(rd, rs, ctx);
arch/riscv/net/bpf_jit_comp64.c
736
if (rd != rs)
arch/riscv/net/bpf_jit_comp64.c
737
emit_mv(rd, rs, ctx);
arch/sparc/include/asm/backoff.h
57
88: rd %ccr, %g0; \
arch/sparc/include/asm/backoff.h
58
rd %ccr, %g0; \
arch/sparc/include/asm/backoff.h
59
rd %ccr, %g0; \
arch/sparc/include/asm/head_32.h
13
rd %psr, %l0; b label; rd %wim, %l3; nop;
arch/sparc/include/asm/head_32.h
16
#define SRMMU_TFAULT rd %psr, %l0; rd %wim, %l3; b srmmu_fault; mov 1, %l7;
arch/sparc/include/asm/head_32.h
17
#define SRMMU_DFAULT rd %psr, %l0; rd %wim, %l3; b srmmu_fault; mov 0, %l7;
arch/sparc/include/asm/head_32.h
21
rd %psr, %l0; mov num, %l7; b bad_trap_handler; rd %wim, %l3;
arch/sparc/include/asm/head_32.h
38
rd %psr, %l0;
arch/sparc/include/asm/head_32.h
42
rd %psr,%l0; \
arch/sparc/include/asm/head_32.h
50
rd %psr,%l0; \
arch/sparc/include/asm/head_32.h
59
b getcc_trap_handler; rd %psr, %l0; nop; nop;
arch/sparc/include/asm/head_32.h
63
b setcc_trap_handler; rd %psr, %l0; nop; nop;
arch/sparc/include/asm/head_32.h
67
rd %psr, %i0; jmp %l2; rett %l2 + 4; nop;
arch/sparc/include/asm/head_32.h
73
mov int_level, %l7; rd %psr, %l0; b real_irq_entry; rd %wim, %l3;
arch/sparc/include/asm/head_32.h
79
rd %psr, %l0; rd %wim, %l3; b spill_window_entry; andcc %l0, PSR_PS, %g0;
arch/sparc/include/asm/head_32.h
82
rd %psr, %l0; rd %wim, %l3; b fill_window_entry; andcc %l0, PSR_PS, %g0;
arch/sparc/include/asm/ttable.h
150
rd %pc, %g7; \
arch/sparc/include/asm/ttable.h
162
rd %pc, %g7; \
arch/sparc/include/asm/ttable.h
176
rd %pc, %g7; \
arch/sparc/include/asm/ttable.h
227
rd %pc, %g7; \
arch/sparc/include/asm/uaccess_64.h
250
unsigned int rd);
arch/sparc/include/asm/visasm.h
16
rd %fprs, %o5; \
arch/sparc/include/asm/visasm.h
38
rd %fprs, %o5; \
arch/sparc/include/asm/winmacro.h
108
661: rd %tbr, %idreg; \
arch/sparc/include/asm/winmacro.h
119
rd %asr17, %idreg; \
arch/sparc/include/asm/winmacro.h
78
rd %y, %scratch; \
arch/sparc/kernel/kprobes.c
248
unsigned long rd = ((insn >> 25) & 0x1f);
arch/sparc/kernel/kprobes.c
250
if (rd <= 15) {
arch/sparc/kernel/kprobes.c
251
slot = &regs->u_regs[rd];
arch/sparc/kernel/kprobes.c
256
rd -= 16;
arch/sparc/kernel/kprobes.c
259
slot += rd;
arch/sparc/kernel/unaligned_32.c
141
unsigned int rd = (insn >> 25) & 0x1f;
arch/sparc/kernel/unaligned_32.c
144
maybe_flush_windows(rs1, 0, rd);
arch/sparc/kernel/unaligned_32.c
147
maybe_flush_windows(rs1, rs2, rd);
arch/sparc/kernel/unaligned_32.c
157
unsigned int rd = (insn >> 25) & 0x1f;
arch/sparc/kernel/unaligned_32.c
160
maybe_flush_windows(rs1, 0, rd);
arch/sparc/kernel/unaligned_32.c
163
maybe_flush_windows(rs1, rs2, rd);
arch/sparc/kernel/unaligned_32.c
73
unsigned int rd)
arch/sparc/kernel/unaligned_32.c
75
if(rs2 >= 16 || rs1 >= 16 || rd >= 16) {
arch/sparc/kernel/unaligned_64.c
105
unsigned int rd, int from_kernel)
arch/sparc/kernel/unaligned_64.c
107
if (rs2 >= 16 || rs1 >= 16 || rd >= 16) {
arch/sparc/kernel/unaligned_64.c
170
unsigned int insn, unsigned int rd)
arch/sparc/kernel/unaligned_64.c
178
maybe_flush_windows(rs1, 0, rd, from_kernel);
arch/sparc/kernel/unaligned_64.c
181
maybe_flush_windows(rs1, rs2, rd, from_kernel);
arch/sparc/kernel/unaligned_64.c
399
int ret, rd = ((insn >> 25) & 0x1f);
arch/sparc/kernel/unaligned_64.c
404
maybe_flush_windows(0, 0, rd, from_kernel);
arch/sparc/kernel/unaligned_64.c
407
maybe_flush_windows(0, insn & 0x1f, rd, from_kernel);
arch/sparc/kernel/unaligned_64.c
411
if (rd < 16) {
arch/sparc/kernel/unaligned_64.c
412
if (rd)
arch/sparc/kernel/unaligned_64.c
413
regs->u_regs[rd] = ret;
arch/sparc/kernel/unaligned_64.c
420
put_user(ret, &win32->locals[rd - 16]);
arch/sparc/kernel/unaligned_64.c
424
put_user(ret, &win->locals[rd - 16]);
arch/sparc/kernel/unaligned_64.c
572
int rd = ((insn >> 25) & 0x1f);
arch/sparc/kernel/unaligned_64.c
578
maybe_flush_windows(0, 0, rd, from_kernel);
arch/sparc/kernel/unaligned_64.c
579
reg = fetch_reg_addr(rd, regs);
arch/sparc/kernel/unaligned_64.c
580
if (from_kernel || rd < 16) {
arch/sparc/kernel/uprobes.c
137
unsigned long rd = ((insn >> 25) & 0x1f);
arch/sparc/kernel/uprobes.c
139
if (rd <= 15) {
arch/sparc/kernel/uprobes.c
140
slot = &regs->u_regs[rd];
arch/sparc/kernel/uprobes.c
146
rd -= 16;
arch/sparc/kernel/uprobes.c
149
(unsigned long __user *) (fp + STACK_BIAS) + rd;
arch/sparc/kernel/uprobes.c
153
__user *) fp + rd;
arch/sparc/kernel/visemul.c
141
unsigned int rd, int from_kernel)
arch/sparc/kernel/visemul.c
143
if (rs2 >= 16 || rs1 >= 16 || rd >= 16) {
arch/sparc/kernel/visemul.c
204
static void store_reg(struct pt_regs *regs, unsigned long val, unsigned long rd)
arch/sparc/kernel/visemul.c
206
if (rd < 16) {
arch/sparc/kernel/visemul.c
207
unsigned long *rd_kern = __fetch_reg_addr_kern(rd, regs);
arch/sparc/kernel/visemul.c
211
unsigned long __user *rd_user = __fetch_reg_addr_user(rd, regs);
arch/sparc/kernel/visemul.c
451
unsigned long rs1, rs2, *rd, rd_val;
arch/sparc/kernel/visemul.c
456
rd = fpd_regaddr(f, RD(insn));
arch/sparc/kernel/visemul.c
458
rd_val = *rd;
arch/sparc/kernel/visemul.c
474
*rd = rd_val;
arch/sparc/math-emu/math_32.c
284
argp rs1 = NULL, rs2 = NULL, rd = NULL;
arch/sparc/math-emu/math_32.c
419
rd = (void *)&fregs[freg];
arch/sparc/math-emu/math_32.c
452
case FMOVS: rd->s = rs2->s; break;
arch/sparc/math-emu/math_32.c
453
case FABSS: rd->s = rs2->s & 0x7fffffff; break;
arch/sparc/math-emu/math_32.c
454
case FNEGS: rd->s = rs2->s ^ 0x80000000; break;
arch/sparc/math-emu/math_32.c
506
case 1: rd->s = IR; break;
arch/sparc/math-emu/math_32.c
507
case 5: FP_PACK_SP (rd, SR); break;
arch/sparc/math-emu/math_32.c
508
case 6: FP_PACK_DP (rd, DR); break;
arch/sparc/math-emu/math_32.c
509
case 7: FP_PACK_QP (rd, QR); break;
arch/sparc/math-emu/math_64.c
359
argp rs1 = NULL, rs2 = NULL, rd = NULL;
arch/sparc/math-emu/math_64.c
416
case 1: rd = (argp)&f->regs[freg];
arch/sparc/math-emu/math_64.c
457
case FMOVQ: rd->q[0] = rs2->q[0]; rd->q[1] = rs2->q[1]; break;
arch/sparc/math-emu/math_64.c
458
case FABSQ: rd->q[0] = rs2->q[0] & 0x7fffffffffffffffUL; rd->q[1] = rs2->q[1]; break;
arch/sparc/math-emu/math_64.c
459
case FNEGQ: rd->q[0] = rs2->q[0] ^ 0x8000000000000000UL; rd->q[1] = rs2->q[1]; break;
arch/sparc/math-emu/math_64.c
507
case 1: rd->s = IR; break;
arch/sparc/math-emu/math_64.c
508
case 2: rd->d = XR; break;
arch/sparc/math-emu/math_64.c
509
case 5: FP_PACK_SP (rd, SR); break;
arch/sparc/math-emu/math_64.c
510
case 6: FP_PACK_DP (rd, DR); break;
arch/sparc/math-emu/math_64.c
511
case 7: FP_PACK_QP (rd, QR); break;
arch/x86/hyperv/irqdomain.c
107
struct rid_data *rd = data;
arch/x86/hyperv/irqdomain.c
108
u8 bus = PCI_BUS_NUM(rd->rid);
arch/x86/hyperv/irqdomain.c
111
rd->bridge = pdev;
arch/x86/hyperv/irqdomain.c
112
rd->rid = alias;
arch/xtensa/platforms/iss/console.c
67
int rd = 1;
arch/xtensa/platforms/iss/console.c
71
rd = simc_read(0, &c, 1);
arch/xtensa/platforms/iss/console.c
72
if (rd <= 0)
arch/xtensa/platforms/iss/console.c
80
if (rd)
drivers/bus/imx-weim.c
304
struct of_reconfig_data *rd = arg;
drivers/bus/imx-weim.c
309
switch (of_reconfig_get_state_change(action, rd)) {
drivers/bus/imx-weim.c
311
of_id = of_match_node(weim_id_table, rd->dn->parent);
drivers/bus/imx-weim.c
317
pdev = of_find_device_by_node(rd->dn->parent);
drivers/bus/imx-weim.c
320
__func__, rd->dn->parent);
drivers/bus/imx-weim.c
325
if (weim_timing_setup(&pdev->dev, rd->dn, devtype))
drivers/bus/imx-weim.c
327
"Failed to setup timing for '%pOF'\n", rd->dn);
drivers/bus/imx-weim.c
329
if (!of_node_check_flag(rd->dn, OF_POPULATED)) {
drivers/bus/imx-weim.c
335
rd->dn->fwnode.flags &= ~FWNODE_FLAG_NOT_DEVICE;
drivers/bus/imx-weim.c
336
if (!of_platform_device_create(rd->dn, NULL, &pdev->dev)) {
drivers/bus/imx-weim.c
339
rd->dn);
drivers/bus/imx-weim.c
348
if (!of_node_check_flag(rd->dn, OF_POPULATED))
drivers/bus/imx-weim.c
351
of_id = of_match_node(weim_id_table, rd->dn->parent);
drivers/bus/imx-weim.c
355
pdev = of_find_device_by_node(rd->dn);
drivers/bus/imx-weim.c
358
rd->dn);
drivers/clk/samsung/clk-exynos5-subcmu.c
22
struct exynos5_subcmu_reg_dump *rd,
drivers/clk/samsung/clk-exynos5-subcmu.c
25
for (; num_regs > 0; --num_regs, ++rd) {
drivers/clk/samsung/clk-exynos5-subcmu.c
26
rd->save = readl(base + rd->offset);
drivers/clk/samsung/clk-exynos5-subcmu.c
27
writel((rd->save & ~rd->mask) | rd->value, base + rd->offset);
drivers/clk/samsung/clk-exynos5-subcmu.c
28
rd->save &= rd->mask;
drivers/clk/samsung/clk-exynos5-subcmu.c
33
struct exynos5_subcmu_reg_dump *rd,
drivers/clk/samsung/clk-exynos5-subcmu.c
36
for (; num_regs > 0; --num_regs, ++rd)
drivers/clk/samsung/clk-exynos5-subcmu.c
37
writel((readl(base + rd->offset) & ~rd->mask) | rd->save,
drivers/clk/samsung/clk-exynos5-subcmu.c
38
base + rd->offset);
drivers/clk/samsung/clk.c
27
struct samsung_clk_reg_dump *rd,
drivers/clk/samsung/clk.c
30
for (; num_regs > 0; --num_regs, ++rd) {
drivers/clk/samsung/clk.c
32
rd->value = readl(base + rd->offset);
drivers/clk/samsung/clk.c
34
regmap_read(regmap, rd->offset, &rd->value);
drivers/clk/samsung/clk.c
40
const struct samsung_clk_reg_dump *rd,
drivers/clk/samsung/clk.c
43
for (; num_regs > 0; --num_regs, ++rd) {
drivers/clk/samsung/clk.c
45
writel(rd->value, base + rd->offset);
drivers/clk/samsung/clk.c
47
regmap_write(regmap, rd->offset, rd->value);
drivers/clk/samsung/clk.c
55
struct samsung_clk_reg_dump *rd;
drivers/clk/samsung/clk.c
58
rd = kzalloc_objs(*rd, nr_rdump);
drivers/clk/samsung/clk.c
59
if (!rd)
drivers/clk/samsung/clk.c
63
rd[i].offset = rdump[i];
drivers/clk/samsung/clk.c
65
return rd;
drivers/clk/samsung/clk.h
467
struct samsung_clk_reg_dump *rd,
drivers/clk/samsung/clk.h
471
const struct samsung_clk_reg_dump *rd,
drivers/clk/versatile/icst.c
51
unsigned int i = 0, rd, best = (unsigned int)-1;
drivers/clk/versatile/icst.c
74
for (rd = p->rd_min; rd <= p->rd_max; rd++) {
drivers/clk/versatile/icst.c
79
fref_div = (2 * p->ref) / rd;
drivers/clk/versatile/icst.c
92
vco.r = rd - 2;
drivers/clocksource/arm_arch_timer.c
910
u64 (*rd)(void);
drivers/clocksource/arm_arch_timer.c
917
rd = arch_counter_get_cntvct_stable;
drivers/clocksource/arm_arch_timer.c
920
rd = arch_counter_get_cntvct;
drivers/clocksource/arm_arch_timer.c
925
rd = arch_counter_get_cntpct_stable;
drivers/clocksource/arm_arch_timer.c
928
rd = arch_counter_get_cntpct;
drivers/clocksource/arm_arch_timer.c
933
arch_timer_read_counter = rd;
drivers/dma/dw-edma/dw-edma-v0-core.c
103
return &__dw_regs(dw)->type.unroll.ch[ch].rd;
drivers/dma/dw-edma/dw-edma-v0-debugfs.c
30
__ch_regs = REGS_ADDR(dw, type.unroll.ch[_ch].rd); \
drivers/dma/dw-edma/dw-edma-v0-regs.h
57
struct dw_edma_v0_ch_regs rd; /* 0x0300 */
drivers/dma/dw-edma/dw-hdma-v0-core.c
37
return &(__dw_regs(dw)->ch[ch].rd);
drivers/dma/dw-edma/dw-hdma-v0-debugfs.c
28
__ch_regs = REGS_ADDR(dw, ch[_ch].rd); \
drivers/dma/dw-edma/dw-hdma-v0-regs.h
91
struct dw_hdma_v0_ch_regs rd; /* 0x0100 */
drivers/firmware/arm_scmi/raw_mode.c
788
struct scmi_dbg_raw_data *rd = filp->private_data;
drivers/firmware/arm_scmi/raw_mode.c
790
if (!rd->rx_size) {
drivers/firmware/arm_scmi/raw_mode.c
793
ret = scmi_raw_message_receive(rd->raw, rd->rx.buf, rd->rx.len,
drivers/firmware/arm_scmi/raw_mode.c
794
&rd->rx_size, idx, rd->chan_id,
drivers/firmware/arm_scmi/raw_mode.c
797
rd->rx_size = 0;
drivers/firmware/arm_scmi/raw_mode.c
803
} else if (*ppos == rd->rx_size) {
drivers/firmware/arm_scmi/raw_mode.c
805
rd->rx_size = 0;
drivers/firmware/arm_scmi/raw_mode.c
810
rd->rx.buf, rd->rx_size);
drivers/firmware/arm_scmi/raw_mode.c
821
struct scmi_dbg_raw_data *rd = filp->private_data;
drivers/firmware/arm_scmi/raw_mode.c
823
if (count > rd->tx.len - rd->tx_size)
drivers/firmware/arm_scmi/raw_mode.c
827
if (!rd->tx_size)
drivers/firmware/arm_scmi/raw_mode.c
828
rd->tx_req_size = count;
drivers/firmware/arm_scmi/raw_mode.c
834
if (rd->tx_size < rd->tx_req_size) {
drivers/firmware/arm_scmi/raw_mode.c
837
cnt = simple_write_to_buffer(rd->tx.buf, rd->tx.len, ppos,
drivers/firmware/arm_scmi/raw_mode.c
842
rd->tx_size += cnt;
drivers/firmware/arm_scmi/raw_mode.c
847
ret = scmi_raw_message_send(rd->raw, rd->tx.buf, rd->tx_size,
drivers/firmware/arm_scmi/raw_mode.c
848
rd->chan_id, async, poll);
drivers/firmware/arm_scmi/raw_mode.c
851
rd->tx_size = 0;
drivers/firmware/arm_scmi/raw_mode.c
862
struct scmi_dbg_raw_data *rd = filp->private_data;
drivers/firmware/arm_scmi/raw_mode.c
866
q = scmi_raw_queue_select(rd->raw, idx, rd->chan_id);
drivers/firmware/arm_scmi/raw_mode.c
905
struct scmi_dbg_raw_data *rd;
drivers/firmware/arm_scmi/raw_mode.c
911
rd = kzalloc_obj(*rd);
drivers/firmware/arm_scmi/raw_mode.c
912
if (!rd)
drivers/firmware/arm_scmi/raw_mode.c
915
rd->rx.len = raw->desc->max_msg_size + sizeof(u32);
drivers/firmware/arm_scmi/raw_mode.c
916
rd->rx.buf = kzalloc(rd->rx.len, GFP_KERNEL);
drivers/firmware/arm_scmi/raw_mode.c
917
if (!rd->rx.buf) {
drivers/firmware/arm_scmi/raw_mode.c
918
kfree(rd);
drivers/firmware/arm_scmi/raw_mode.c
922
rd->tx.len = raw->desc->max_msg_size + sizeof(u32);
drivers/firmware/arm_scmi/raw_mode.c
923
rd->tx.buf = kzalloc(rd->tx.len, GFP_KERNEL);
drivers/firmware/arm_scmi/raw_mode.c
924
if (!rd->tx.buf) {
drivers/firmware/arm_scmi/raw_mode.c
925
kfree(rd->rx.buf);
drivers/firmware/arm_scmi/raw_mode.c
926
kfree(rd);
drivers/firmware/arm_scmi/raw_mode.c
932
rd->chan_id = debugfs_get_aux_num(filp);
drivers/firmware/arm_scmi/raw_mode.c
934
rd->raw = raw;
drivers/firmware/arm_scmi/raw_mode.c
935
filp->private_data = rd;
drivers/firmware/arm_scmi/raw_mode.c
942
struct scmi_dbg_raw_data *rd = filp->private_data;
drivers/firmware/arm_scmi/raw_mode.c
944
kfree(rd->rx.buf);
drivers/firmware/arm_scmi/raw_mode.c
945
kfree(rd->tx.buf);
drivers/firmware/arm_scmi/raw_mode.c
946
kfree(rd);
drivers/firmware/arm_scmi/raw_mode.c
955
struct scmi_dbg_raw_data *rd = filp->private_data;
drivers/firmware/arm_scmi/raw_mode.c
957
scmi_xfer_raw_reset(rd->raw);
drivers/gpio/gpiolib-of.c
904
struct of_reconfig_data *rd = arg;
drivers/gpio/gpiolib-of.c
915
if (!of_property_read_bool(rd->dn, "gpio-hog"))
drivers/gpio/gpiolib-of.c
918
if (of_node_test_and_set_flag(rd->dn, OF_POPULATED))
drivers/gpio/gpiolib-of.c
921
gdev = of_find_gpio_device_by_node(rd->dn->parent);
drivers/gpio/gpiolib-of.c
925
ret = of_gpiochip_add_hog(gpio_device_get_chip(gdev), rd->dn);
drivers/gpio/gpiolib-of.c
928
rd->dn);
drivers/gpio/gpiolib-of.c
929
of_node_clear_flag(rd->dn, OF_POPULATED);
drivers/gpio/gpiolib-of.c
935
if (!of_node_check_flag(rd->dn, OF_POPULATED))
drivers/gpio/gpiolib-of.c
938
gdev = of_find_gpio_device_by_node(rd->dn->parent);
drivers/gpio/gpiolib-of.c
942
of_gpiochip_remove_hog(gpio_device_get_chip(gdev), rd->dn);
drivers/gpio/gpiolib-of.c
943
of_node_clear_flag(rd->dn, OF_POPULATED);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
207
struct amdgpu_debugfs_regs2_data *rd;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
209
rd = kzalloc_obj(*rd);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
210
if (!rd)
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
212
rd->adev = file_inode(file)->i_private;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
213
file->private_data = rd;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
214
mutex_init(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
221
struct amdgpu_debugfs_regs2_data *rd = file->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
223
mutex_destroy(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
230
struct amdgpu_debugfs_regs2_data *rd = f->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
231
struct amdgpu_device *adev = rd->adev;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
251
mutex_lock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
253
if (rd->id.use_grbm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
254
if ((rd->id.grbm.sh != 0xFFFFFFFF && rd->id.grbm.sh >= adev->gfx.config.max_sh_per_se) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
255
(rd->id.grbm.se != 0xFFFFFFFF && rd->id.grbm.se >= adev->gfx.config.max_shader_engines)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
258
mutex_unlock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
262
amdgpu_gfx_select_se_sh(adev, rd->id.grbm.se,
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
263
rd->id.grbm.sh,
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
264
rd->id.grbm.instance, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
267
if (rd->id.use_srbm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
269
amdgpu_gfx_select_me_pipe_q(adev, rd->id.srbm.me, rd->id.srbm.pipe,
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
270
rd->id.srbm.queue, rd->id.srbm.vmid, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
273
if (rd->id.pg_lock)
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
283
amdgpu_mm_wreg_mmio_rlc(adev, offset >> 2, value, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
295
if (rd->id.use_grbm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
296
amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
300
if (rd->id.use_srbm) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
301
amdgpu_gfx_select_me_pipe_q(adev, 0, 0, 0, 0, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
305
if (rd->id.pg_lock)
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
308
mutex_unlock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
318
struct amdgpu_debugfs_regs2_data *rd = f->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
322
mutex_lock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
326
r = copy_from_user(&rd->id, (struct amdgpu_debugfs_regs2_iocdata_v2 *)data,
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
327
sizeof(rd->id));
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
345
rd->id.use_srbm = v1_data.use_srbm;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
346
rd->id.use_grbm = v1_data.use_grbm;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
347
rd->id.pg_lock = v1_data.pg_lock;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
348
rd->id.grbm.se = v1_data.grbm.se;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
349
rd->id.grbm.sh = v1_data.grbm.sh;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
350
rd->id.grbm.instance = v1_data.grbm.instance;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
351
rd->id.srbm.me = v1_data.srbm.me;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
352
rd->id.srbm.pipe = v1_data.srbm.pipe;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
353
rd->id.srbm.queue = v1_data.srbm.queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
354
rd->id.xcc_id = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
356
mutex_unlock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
372
struct amdgpu_debugfs_gprwave_data *rd;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
374
rd = kzalloc_obj(*rd);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
375
if (!rd)
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
377
rd->adev = file_inode(file)->i_private;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
378
file->private_data = rd;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
379
mutex_init(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
386
struct amdgpu_debugfs_gprwave_data *rd = file->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
388
mutex_destroy(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
395
struct amdgpu_debugfs_gprwave_data *rd = f->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
396
struct amdgpu_device *adev = rd->adev;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
425
amdgpu_gfx_select_se_sh(adev, rd->id.se, rd->id.sh, rd->id.cu, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
427
if (!rd->id.gpr_or_wave) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
430
adev->gfx.funcs->read_wave_data(adev, rd->id.xcc_id, rd->id.simd, rd->id.wave, data, &x);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
433
if (rd->id.gpr.vpgr_or_sgpr) {
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
435
adev->gfx.funcs->read_wave_vgprs(adev, rd->id.xcc_id, rd->id.simd, rd->id.wave, rd->id.gpr.thread, *pos, size>>2, data);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
438
adev->gfx.funcs->read_wave_sgprs(adev, rd->id.xcc_id, rd->id.simd, rd->id.wave, *pos, size>>2, data);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
442
amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
476
struct amdgpu_debugfs_gprwave_data *rd = f->private_data;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
479
mutex_lock(&rd->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
483
if (copy_from_user(&rd->id,
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
485
sizeof(rd->id)))
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
494
mutex_unlock(&rd->lock);
drivers/gpu/drm/i915/gt/uc/guc_capture_fwif.h
25
u32 rd;
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
1340
buf.rd = read_offset;
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
745
if (buf->wr >= buf->rd)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
746
return (buf->wr - buf->rd);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
747
return (buf->size - buf->rd) + buf->wr;
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
752
if (buf->rd > buf->wr)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
753
return (buf->size - buf->rd);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
754
return (buf->wr - buf->rd);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
791
src_data = (u32 *)(buf->data + buf->rd);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
793
buf->rd += 4;
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
798
buf->rd = 0;
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
809
memcpy(dest, (b->data + b->rd), size);
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
810
b->rd += size;
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
74
ot_params.rd = false;
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c
109
u32 xin_id, bool rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c
117
if (rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c
88
u32 xin_id, bool rd, u32 limit)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.c
95
if (rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
27
u32 xin_id, bool rd, u32 limit);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_vbif.h
37
u32 xin_id, bool rd);
drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c
378
ot_params.rd = true;
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
100
tbl = params->rd ? &vbif->cap->dynamic_ot_rd_tbl :
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
133
if (vbif->cap->default_ot_wr_limit && !params->rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
135
else if (vbif->cap->default_ot_rd_limit && params->rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
150
params->xin_id, params->rd);
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
185
if (vbif->ops.set_write_gather_en && !params->rd)
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.c
196
vbif->ops.set_limit_conf(vbif, params->xin_id, params->rd, ot_lim);
drivers/gpu/drm/msm/disp/dpu1/dpu_vbif.h
16
bool rd;
drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c
1002
struct hdmi_hdcp_reg_data *rd;
drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c
1009
rd = &reg_data[i];
drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c
1011
rd->off, (u8 *)&data[i], (u16)sizeof(data[i]));
drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c
1013
pr_err("%s: Read %s failed\n", __func__, rd->name);
drivers/gpu/drm/msm/hdmi/hdmi_hdcp.c
1017
DBG("%s =%x", rd->name, data[i]);
drivers/gpu/drm/msm/msm_drv.h
443
void msm_rd_dump_submit(struct msm_rd_state *rd, struct msm_gem_submit *submit,
drivers/gpu/drm/msm/msm_drv.h
450
static inline void msm_rd_dump_submit(struct msm_rd_state *rd,
drivers/gpu/drm/msm/msm_drv.h
90
struct msm_rd_state *rd; /* debugfs to dump all submits */
drivers/gpu/drm/msm/msm_gem_submit.c
803
msm_rd_dump_submit(priv->rd, submit, NULL);
drivers/gpu/drm/msm/msm_rd.c
106
wait_event(rd->fifo_event, circ_space(&rd->fifo) > 0 || !rd->open);
drivers/gpu/drm/msm/msm_rd.c
107
if (!rd->open)
drivers/gpu/drm/msm/msm_rd.c
114
n = min(sz, circ_space_to_end(&rd->fifo));
drivers/gpu/drm/msm/msm_rd.c
121
wake_up_all(&rd->fifo_event);
drivers/gpu/drm/msm/msm_rd.c
125
static void rd_write_section(struct msm_rd_state *rd,
drivers/gpu/drm/msm/msm_rd.c
128
rd_write(rd, &type, 4);
drivers/gpu/drm/msm/msm_rd.c
129
rd_write(rd, &sz, 4);
drivers/gpu/drm/msm/msm_rd.c
130
rd_write(rd, buf, sz);
drivers/gpu/drm/msm/msm_rd.c
136
struct msm_rd_state *rd = file->private_data;
drivers/gpu/drm/msm/msm_rd.c
137
struct circ_buf *fifo = &rd->fifo;
drivers/gpu/drm/msm/msm_rd.c
141
mutex_lock(&rd->read_lock);
drivers/gpu/drm/msm/msm_rd.c
143
ret = wait_event_interruptible(rd->fifo_event,
drivers/gpu/drm/msm/msm_rd.c
144
circ_count(&rd->fifo) > 0);
drivers/gpu/drm/msm/msm_rd.c
152
n = min_t(int, sz, circ_count_to_end(&rd->fifo));
drivers/gpu/drm/msm/msm_rd.c
161
wake_up_all(&rd->fifo_event);
drivers/gpu/drm/msm/msm_rd.c
164
mutex_unlock(&rd->read_lock);
drivers/gpu/drm/msm/msm_rd.c
172
struct msm_rd_state *rd = inode->i_private;
drivers/gpu/drm/msm/msm_rd.c
173
struct drm_device *dev = rd->dev;
drivers/gpu/drm/msm/msm_rd.c
186
if (rd->open) {
drivers/gpu/drm/msm/msm_rd.c
191
file->private_data = rd;
drivers/gpu/drm/msm/msm_rd.c
192
rd->open = true;
drivers/gpu/drm/msm/msm_rd.c
195
rd->fifo.head = rd->fifo.tail = 0;
drivers/gpu/drm/msm/msm_rd.c
205
rd_write_section(rd, RD_GPU_ID, &gpu_id, sizeof(gpu_id));
drivers/gpu/drm/msm/msm_rd.c
208
rd_write_section(rd, RD_CHIP_ID, &val, sizeof(val));
drivers/gpu/drm/msm/msm_rd.c
217
struct msm_rd_state *rd = inode->i_private;
drivers/gpu/drm/msm/msm_rd.c
219
rd->open = false;
drivers/gpu/drm/msm/msm_rd.c
220
wake_up_all(&rd->fifo_event);
drivers/gpu/drm/msm/msm_rd.c
234
static void rd_cleanup(struct msm_rd_state *rd)
drivers/gpu/drm/msm/msm_rd.c
236
if (!rd)
drivers/gpu/drm/msm/msm_rd.c
239
mutex_destroy(&rd->read_lock);
drivers/gpu/drm/msm/msm_rd.c
240
mutex_destroy(&rd->write_lock);
drivers/gpu/drm/msm/msm_rd.c
241
kfree(rd);
drivers/gpu/drm/msm/msm_rd.c
246
struct msm_rd_state *rd;
drivers/gpu/drm/msm/msm_rd.c
248
rd = kzalloc_obj(*rd);
drivers/gpu/drm/msm/msm_rd.c
249
if (!rd)
drivers/gpu/drm/msm/msm_rd.c
252
rd->dev = minor->dev;
drivers/gpu/drm/msm/msm_rd.c
253
rd->fifo.buf = rd->buf;
drivers/gpu/drm/msm/msm_rd.c
255
mutex_init(&rd->read_lock);
drivers/gpu/drm/msm/msm_rd.c
256
mutex_init(&rd->write_lock);
drivers/gpu/drm/msm/msm_rd.c
258
init_waitqueue_head(&rd->fifo_event);
drivers/gpu/drm/msm/msm_rd.c
260
debugfs_create_file(name, S_IFREG | S_IRUGO, minor->debugfs_root, rd,
drivers/gpu/drm/msm/msm_rd.c
263
return rd;
drivers/gpu/drm/msm/msm_rd.c
269
struct msm_rd_state *rd;
drivers/gpu/drm/msm/msm_rd.c
276
if (priv->rd)
drivers/gpu/drm/msm/msm_rd.c
279
rd = rd_init(minor, "rd");
drivers/gpu/drm/msm/msm_rd.c
280
if (IS_ERR(rd)) {
drivers/gpu/drm/msm/msm_rd.c
281
ret = PTR_ERR(rd);
drivers/gpu/drm/msm/msm_rd.c
285
priv->rd = rd;
drivers/gpu/drm/msm/msm_rd.c
287
rd = rd_init(minor, "hangrd");
drivers/gpu/drm/msm/msm_rd.c
288
if (IS_ERR(rd)) {
drivers/gpu/drm/msm/msm_rd.c
289
ret = PTR_ERR(rd);
drivers/gpu/drm/msm/msm_rd.c
293
priv->hangrd = rd;
drivers/gpu/drm/msm/msm_rd.c
304
rd_cleanup(priv->rd);
drivers/gpu/drm/msm/msm_rd.c
305
priv->rd = NULL;
drivers/gpu/drm/msm/msm_rd.c
311
static void snapshot_buf(struct msm_rd_state *rd, struct drm_gem_object *obj,
drivers/gpu/drm/msm/msm_rd.c
320
rd_write_section(rd, RD_GPUADDR,
drivers/gpu/drm/msm/msm_rd.c
332
rd_write_section(rd, RD_BUFFER_CONTENTS, buf, size);
drivers/gpu/drm/msm/msm_rd.c
338
void msm_rd_dump_submit(struct msm_rd_state *rd, struct msm_gem_submit *submit,
drivers/gpu/drm/msm/msm_rd.c
346
if (!rd->open)
drivers/gpu/drm/msm/msm_rd.c
349
mutex_lock(&rd->write_lock);
drivers/gpu/drm/msm/msm_rd.c
358
rd_write_section(rd, RD_CMD, msg, ALIGN(n, 4));
drivers/gpu/drm/msm/msm_rd.c
373
rd_write_section(rd, RD_CMD, msg, ALIGN(n, 4));
drivers/gpu/drm/msm/msm_rd.c
387
snapshot_buf(rd, vma->gem.obj, vma->va.addr, dump,
drivers/gpu/drm/msm/msm_rd.c
396
snapshot_buf(rd, obj, submit->bos[i].iova, dump, 0, obj->size);
drivers/gpu/drm/msm/msm_rd.c
409
snapshot_buf(rd, obj, submit->cmd[i].iova, true,
drivers/gpu/drm/msm/msm_rd.c
428
rd_write_section(rd, RD_CMDSTREAM_ADDR,
drivers/gpu/drm/msm/msm_rd.c
434
mutex_unlock(&rd->write_lock);
drivers/gpu/drm/msm/msm_rd.c
97
static void rd_write(struct msm_rd_state *rd, const void *buf, int sz)
drivers/gpu/drm/msm/msm_rd.c
99
struct circ_buf *fifo = &rd->fifo;
drivers/gpu/drm/nouveau/include/nvkm/core/falcon.h
31
void (*rd)(struct nvkm_falcon *, u8 port, const u8 *img, int len);
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
147
if (WARN_ON(!pio || !pio->rd))
drivers/gpu/drm/nouveau/nvkm/falcon/base.c
157
pio->rd(falcon, port, img, xfer_len);
drivers/gpu/drm/nouveau/nvkm/falcon/gm200.c
94
.rd = gm200_flcn_pio_dmem_rd,
drivers/gpu/drm/nouveau/nvkm/falcon/gp102.c
63
.rd = gp102_flcn_pio_emem_rd,
drivers/gpu/drm/xe/xe_guc_capture.c
1040
if (!buf->rd)
drivers/gpu/drm/xe/xe_guc_capture.c
1042
buf->rd = 0;
drivers/gpu/drm/xe/xe_guc_capture.c
1049
buf->data_offset + buf->rd, copy_size);
drivers/gpu/drm/xe/xe_guc_capture.c
1050
buf->rd += copy_size;
drivers/gpu/drm/xe/xe_guc_capture.c
1422
buf.rd = read_offset;
drivers/gpu/drm/xe/xe_guc_capture.c
49
u32 rd;
drivers/gpu/drm/xe/xe_guc_capture.c
986
if (buf->wr >= buf->rd)
drivers/gpu/drm/xe/xe_guc_capture.c
987
return (buf->wr - buf->rd);
drivers/gpu/drm/xe/xe_guc_capture.c
988
return (buf->size - buf->rd) + buf->wr;
drivers/gpu/drm/xe/xe_guc_capture.c
993
if (buf->rd > buf->wr)
drivers/gpu/drm/xe/xe_guc_capture.c
994
return (buf->size - buf->rd);
drivers/gpu/drm/xe/xe_guc_capture.c
995
return (buf->wr - buf->rd);
drivers/hid/hid-lg.c
740
u8 *rd, int size)
drivers/hid/hid-lg.c
745
return lg4ff_raw_event(hdev, report, rd, size, drv_data);
drivers/hid/hid-lg4ff.c
327
u8 *rd, int size, struct lg_drv_data *drv_data)
drivers/hid/hid-lg4ff.c
339
rd[5] = rd[3];
drivers/hid/hid-lg4ff.c
340
rd[6] = 0x7F;
drivers/hid/hid-lg4ff.c
346
rd[4] = rd[3];
drivers/hid/hid-lg4ff.c
347
rd[5] = 0x7F;
drivers/hid/hid-lg4ff.c
350
rd[5] = rd[4];
drivers/hid/hid-lg4ff.c
351
rd[6] = 0x7F;
drivers/hid/hid-lg4ff.c
369
rd[offset] = (0xFF + rd[offset] - rd[offset+1]) >> 1;
drivers/hid/hid-lg4ff.c
370
rd[offset+1] = 0x7F;
drivers/hid/hid-lg4ff.h
11
u8 *rd, int size, struct lg_drv_data *drv_data);
drivers/hid/hid-lg4ff.h
18
u8 *rd, int size, struct lg_drv_data *drv_data) { return 0; }
drivers/hid/hid-sony.c
1003
input_report_abs(sc->input_dev, ABS_Z, rd[41]);
drivers/hid/hid-sony.c
1004
input_report_abs(sc->input_dev, ABS_RZ, rd[42]);
drivers/hid/hid-sony.c
1010
u8 *rd, int size)
drivers/hid/hid-sony.c
1018
if ((sc->quirks & SIXAXIS_CONTROLLER) && rd[0] == 0x01 && size == 49) {
drivers/hid/hid-sony.c
1027
if (rd[1] == 0xff)
drivers/hid/hid-sony.c
1030
swap(rd[41], rd[42]);
drivers/hid/hid-sony.c
1031
swap(rd[43], rd[44]);
drivers/hid/hid-sony.c
1032
swap(rd[45], rd[46]);
drivers/hid/hid-sony.c
1033
swap(rd[47], rd[48]);
drivers/hid/hid-sony.c
1035
sixaxis_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1036
} else if ((sc->quirks & MOTION_CONTROLLER_BT) && rd[0] == 0x01 && size == 49) {
drivers/hid/hid-sony.c
1037
sixaxis_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1038
} else if ((sc->quirks & NAVIGATION_CONTROLLER) && rd[0] == 0x01 &&
drivers/hid/hid-sony.c
1040
sixaxis_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1041
} else if ((sc->quirks & NSG_MRXU_REMOTE) && rd[0] == 0x02) {
drivers/hid/hid-sony.c
1042
nsg_mrxu_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1044
} else if ((sc->quirks & RB4_GUITAR_PS4_USB) && rd[0] == 0x01 && size == 64) {
drivers/hid/hid-sony.c
1045
rb4_ps4_guitar_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1047
} else if ((sc->quirks & RB4_GUITAR_PS4_BT) && rd[0] == 0x01 && size == 78) {
drivers/hid/hid-sony.c
1048
rb4_ps4_guitar_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
1050
} else if ((sc->quirks & RB4_GUITAR_PS5) && rd[0] == 0x01 && size == 64) {
drivers/hid/hid-sony.c
1051
rb4_ps5_guitar_parse_report(sc, rd, size);
drivers/hid/hid-sony.c
853
static void sixaxis_parse_report(struct sony_sc *sc, u8 *rd, int size)
drivers/hid/hid-sony.c
869
if (rd[offset] >= 0xee) {
drivers/hid/hid-sony.c
871
battery_status = (rd[offset] & 0x01) ? POWER_SUPPLY_STATUS_FULL : POWER_SUPPLY_STATUS_CHARGING;
drivers/hid/hid-sony.c
873
u8 index = rd[offset] <= 5 ? rd[offset] : 5;
drivers/hid/hid-sony.c
887
val = ((rd[offset+1] << 8) | rd[offset]) - 511;
drivers/hid/hid-sony.c
891
val = 511 - ((rd[offset+5] << 8) | rd[offset+4]);
drivers/hid/hid-sony.c
894
val = 511 - ((rd[offset+3] << 8) | rd[offset+2]);
drivers/hid/hid-sony.c
901
static void nsg_mrxu_parse_report(struct sony_sc *sc, u8 *rd, int size)
drivers/hid/hid-sony.c
926
input_report_key(sc->touchpad, BTN_LEFT, rd[offset] & 0x0F);
drivers/hid/hid-sony.c
927
active = (rd[offset] >> 4);
drivers/hid/hid-sony.c
928
relx = (s8) rd[offset+5];
drivers/hid/hid-sony.c
929
rely = ((s8) rd[offset+10]) * -1;
drivers/hid/hid-sony.c
937
x = rd[offset] | ((rd[offset+1] & 0x0F) << 8);
drivers/hid/hid-sony.c
938
y = ((rd[offset+1] & 0xF0) >> 4) | (rd[offset+2] << 4);
drivers/hid/hid-sony.c
944
contactx = rd[offset+3] & 0x0F;
drivers/hid/hid-sony.c
945
contacty = rd[offset+3] >> 4;
drivers/hid/hid-sony.c
975
static void rb4_ps4_guitar_parse_report(struct sony_sc *sc, u8 *rd, int size)
drivers/hid/hid-sony.c
986
input_report_abs(sc->input_dev, ABS_Z, rd[44]);
drivers/hid/hid-sony.c
987
input_report_abs(sc->input_dev, ABS_RZ, rd[45]);
drivers/hid/hid-sony.c
992
static void rb4_ps5_guitar_parse_report(struct sony_sc *sc, u8 *rd, int size)
drivers/i2c/busses/i2c-au1550.c
115
if (rd)
drivers/i2c/busses/i2c-au1550.c
98
do_address(struct i2c_au1550_data *adap, unsigned int addr, int rd, int q)
drivers/i2c/busses/i2c-ocores.c
486
u32 rd;
drivers/i2c/busses/i2c-ocores.c
491
rd = ioread32be(i2c->base + (rreg << i2c->reg_shift));
drivers/i2c/busses/i2c-ocores.c
493
return (u8)(rd >> 8);
drivers/i2c/busses/i2c-ocores.c
495
return (u8)rd;
drivers/i2c/busses/i2c-octeon-core.c
679
__be64 rd = cpu_to_be64(__raw_readq(i2c->twsi_base + OCTEON_REG_BLOCK_FIFO(i2c)));
drivers/i2c/busses/i2c-octeon-core.c
681
memcpy(&msgs[1].buf[i], &rd, min(8, msgs[1].len - i));
drivers/i2c/busses/i2c-pasemi-core.c
198
u32 rd;
drivers/i2c/busses/i2c-pasemi-core.c
213
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
214
if (rd & MRXFIFO_EMPTY) {
drivers/i2c/busses/i2c-pasemi-core.c
218
msg->buf[i] = rd & MRXFIFO_DATA_M;
drivers/i2c/busses/i2c-pasemi-core.c
262
unsigned int rd;
drivers/i2c/busses/i2c-pasemi-core.c
313
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
314
len = min_t(u8, (rd & MRXFIFO_DATA_M),
drivers/i2c/busses/i2c-pasemi-core.c
345
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
346
len = min_t(u8, (rd & MRXFIFO_DATA_M),
drivers/i2c/busses/i2c-pasemi-core.c
366
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
367
if (rd & MRXFIFO_EMPTY) {
drivers/i2c/busses/i2c-pasemi-core.c
371
data->byte = rd & MRXFIFO_DATA_M;
drivers/i2c/busses/i2c-pasemi-core.c
375
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
376
if (rd & MRXFIFO_EMPTY) {
drivers/i2c/busses/i2c-pasemi-core.c
380
data->word = rd & MRXFIFO_DATA_M;
drivers/i2c/busses/i2c-pasemi-core.c
381
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
382
if (rd & MRXFIFO_EMPTY) {
drivers/i2c/busses/i2c-pasemi-core.c
386
data->word |= (rd & MRXFIFO_DATA_M) << 8;
drivers/i2c/busses/i2c-pasemi-core.c
392
rd = RXFIFO_RD(smbus);
drivers/i2c/busses/i2c-pasemi-core.c
393
if (rd & MRXFIFO_EMPTY) {
drivers/i2c/busses/i2c-pasemi-core.c
397
data->block[i] = rd & MRXFIFO_DATA_M;
drivers/i2c/i2c-core-of.c
164
struct of_reconfig_data *rd = arg;
drivers/i2c/i2c-core-of.c
168
switch (of_reconfig_get_state_change(action, rd)) {
drivers/i2c/i2c-core-of.c
170
adap = of_find_i2c_adapter_by_node(rd->dn->parent);
drivers/i2c/i2c-core-of.c
174
if (of_node_test_and_set_flag(rd->dn, OF_POPULATED)) {
drivers/i2c/i2c-core-of.c
183
rd->dn->fwnode.flags &= ~FWNODE_FLAG_NOT_DEVICE;
drivers/i2c/i2c-core-of.c
184
client = of_i2c_register_device(adap, rd->dn);
drivers/i2c/i2c-core-of.c
187
rd->dn);
drivers/i2c/i2c-core-of.c
189
of_node_clear_flag(rd->dn, OF_POPULATED);
drivers/i2c/i2c-core-of.c
196
if (!of_node_check_flag(rd->dn, OF_POPULATED))
drivers/i2c/i2c-core-of.c
200
client = of_find_i2c_device_by_node(rd->dn);
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
104
int ms_sensors_convert_and_read(void *cli, u8 conv, u8 rd,
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
118
if (rd != MS_SENSORS_NO_READ_CMD)
drivers/iio/common/ms_sensors/ms_sensors_i2c.c
119
ret = i2c_smbus_read_i2c_block_data(client, rd, 3, (u8 *)&buf);
drivers/iio/common/ms_sensors/ms_sensors_i2c.h
56
int ms_sensors_convert_and_read(void *cli, u8 conv, u8 rd,
drivers/iio/temperature/tsys01.c
35
int (*convert_and_read)(void *cli, u8 conv, u8 rd,
drivers/md/dm-raid.c
1139
struct raid_dev *rd;
drivers/md/dm-raid.c
1328
rd = rs->dev + value;
drivers/md/dm-raid.c
1329
clear_bit(In_sync, &rd->rdev.flags);
drivers/md/dm-raid.c
1330
clear_bit(Faulty, &rd->rdev.flags);
drivers/md/dm-raid.c
1331
rd->rdev.recovery_offset = 0;
drivers/media/common/videobuf2/videobuf2-dma-contig.c
323
struct scatterlist *rd, *wr;
drivers/media/common/videobuf2/videobuf2-dma-contig.c
342
rd = buf->sgt_base->sgl;
drivers/media/common/videobuf2/videobuf2-dma-contig.c
345
sg_set_page(wr, sg_page(rd), rd->length, rd->offset);
drivers/media/common/videobuf2/videobuf2-dma-contig.c
346
rd = sg_next(rd);
drivers/media/common/videobuf2/videobuf2-dma-sg.c
373
struct scatterlist *rd, *wr;
drivers/media/common/videobuf2/videobuf2-dma-sg.c
392
rd = buf->dma_sgt->sgl;
drivers/media/common/videobuf2/videobuf2-dma-sg.c
395
sg_set_page(wr, sg_page(rd), rd->length, rd->offset);
drivers/media/common/videobuf2/videobuf2-dma-sg.c
396
rd = sg_next(rd);
drivers/media/dvb-frontends/dib3000mb.c
342
rd(DIB3000MB_REG_AS_IRQ_PENDING),
drivers/media/dvb-frontends/dib3000mb.c
343
rd(DIB3000MB_REG_LOCK2_VALUE))) < 0 && as_count++ < 100)
drivers/media/dvb-frontends/dib3000mb.c
450
if (!rd(DIB3000MB_REG_TPS_LOCK))
drivers/media/dvb-frontends/dib3000mb.c
453
dds_val = ((rd(DIB3000MB_REG_DDS_VALUE_MSB) & 0xff) << 16) + rd(DIB3000MB_REG_DDS_VALUE_LSB);
drivers/media/dvb-frontends/dib3000mb.c
454
deb_getf("DDS_VAL: %x %x %x\n", dds_val, rd(DIB3000MB_REG_DDS_VALUE_MSB), rd(DIB3000MB_REG_DDS_VALUE_LSB));
drivers/media/dvb-frontends/dib3000mb.c
462
dds_val = ((rd(DIB3000MB_REG_DDS_FREQ_MSB) & 0xff) << 16) + rd(DIB3000MB_REG_DDS_FREQ_LSB);
drivers/media/dvb-frontends/dib3000mb.c
463
deb_getf("DDS_FREQ: %x %x %x\n", dds_val, rd(DIB3000MB_REG_DDS_FREQ_MSB), rd(DIB3000MB_REG_DDS_FREQ_LSB));
drivers/media/dvb-frontends/dib3000mb.c
478
switch ((tps_val = rd(DIB3000MB_REG_TPS_QAM))) {
drivers/media/dvb-frontends/dib3000mb.c
497
if (rd(DIB3000MB_REG_TPS_HRCH)) {
drivers/media/dvb-frontends/dib3000mb.c
501
switch ((tps_val = rd(DIB3000MB_REG_TPS_VIT_ALPHA))) {
drivers/media/dvb-frontends/dib3000mb.c
524
tps_val = rd(DIB3000MB_REG_TPS_CODE_RATE_LP);
drivers/media/dvb-frontends/dib3000mb.c
531
tps_val = rd(DIB3000MB_REG_TPS_CODE_RATE_HP);
drivers/media/dvb-frontends/dib3000mb.c
561
switch ((tps_val = rd(DIB3000MB_REG_TPS_GUARD_TIME))) {
drivers/media/dvb-frontends/dib3000mb.c
584
switch ((tps_val = rd(DIB3000MB_REG_TPS_FFT))) {
drivers/media/dvb-frontends/dib3000mb.c
609
if (rd(DIB3000MB_REG_AGC_LOCK))
drivers/media/dvb-frontends/dib3000mb.c
611
if (rd(DIB3000MB_REG_CARRIER_LOCK))
drivers/media/dvb-frontends/dib3000mb.c
613
if (rd(DIB3000MB_REG_VIT_LCK))
drivers/media/dvb-frontends/dib3000mb.c
615
if (rd(DIB3000MB_REG_TS_SYNC_LOCK))
drivers/media/dvb-frontends/dib3000mb.c
621
rd(DIB3000MB_REG_TPS_LOCK),
drivers/media/dvb-frontends/dib3000mb.c
622
rd(DIB3000MB_REG_TPS_QAM),
drivers/media/dvb-frontends/dib3000mb.c
623
rd(DIB3000MB_REG_TPS_HRCH),
drivers/media/dvb-frontends/dib3000mb.c
624
rd(DIB3000MB_REG_TPS_VIT_ALPHA),
drivers/media/dvb-frontends/dib3000mb.c
625
rd(DIB3000MB_REG_TPS_CODE_RATE_HP),
drivers/media/dvb-frontends/dib3000mb.c
626
rd(DIB3000MB_REG_TPS_CODE_RATE_LP),
drivers/media/dvb-frontends/dib3000mb.c
627
rd(DIB3000MB_REG_TPS_GUARD_TIME),
drivers/media/dvb-frontends/dib3000mb.c
628
rd(DIB3000MB_REG_TPS_FFT),
drivers/media/dvb-frontends/dib3000mb.c
629
rd(DIB3000MB_REG_TPS_CELL_ID));
drivers/media/dvb-frontends/dib3000mb.c
639
*ber = ((rd(DIB3000MB_REG_BER_MSB) << 16) | rd(DIB3000MB_REG_BER_LSB));
drivers/media/dvb-frontends/dib3000mb.c
648
*strength = rd(DIB3000MB_REG_SIGNAL_POWER) * 0xffff / 0x170;
drivers/media/dvb-frontends/dib3000mb.c
655
short sigpow = rd(DIB3000MB_REG_SIGNAL_POWER);
drivers/media/dvb-frontends/dib3000mb.c
656
int icipow = ((rd(DIB3000MB_REG_NOISE_POWER_MSB) & 0xff) << 16) |
drivers/media/dvb-frontends/dib3000mb.c
657
rd(DIB3000MB_REG_NOISE_POWER_LSB);
drivers/media/dvb-frontends/dib3000mb.c
666
*unc = rd(DIB3000MB_REG_PACKET_ERROR_RATE);
drivers/media/dvb-frontends/dib3000mb.c
758
if (rd(DIB3000_REG_MANUFACTOR_ID) != DIB3000_I2C_ID_DIBCOM)
drivers/media/dvb-frontends/dib3000mb.c
761
if (rd(DIB3000_REG_DEVICE_ID) != DIB3000MB_DEVICE_ID)
drivers/media/dvb-frontends/dib3000mb_priv.h
26
#define set_or(reg,val) wr(reg,rd(reg) | val)
drivers/media/dvb-frontends/dib3000mb_priv.h
28
#define set_and(reg,val) wr(reg,rd(reg) & val)
drivers/media/dvb-frontends/mb86a20s.c
233
u8 i2c_addr, struct regdata *rd, int size)
drivers/media/dvb-frontends/mb86a20s.c
238
rc = mb86a20s_i2c_writereg(state, i2c_addr, rd[i].reg,
drivers/media/dvb-frontends/mb86a20s.c
239
rd[i].data);
drivers/media/dvb-frontends/s921.c
218
struct regdata *rd, int size)
drivers/media/dvb-frontends/s921.c
223
rc = s921_i2c_writereg(state, i2c_addr, rd[i].reg, rd[i].data);
drivers/media/pci/bt8xx/bt878.c
344
retval = bttv_read_gpio(bt->bttv_nr, &mp->rd.value);
drivers/media/pci/bt8xx/dst.c
99
*result = (u8) rd_packet.rd.value;
drivers/media/pci/bt8xx/dst_priv.h
25
struct dst_gpio_read rd;
drivers/media/pci/cobalt/cobalt-cpld.c
27
u32 rd;
drivers/media/pci/cobalt/cobalt-cpld.c
66
rd = cpld_read(cobalt, 0x3c);
drivers/media/pci/cobalt/cobalt-cpld.c
67
tmp = (rd * 33 * 1000) / (483 * 10);
drivers/media/pci/cobalt/cobalt-cpld.c
69
rd = cpld_read(cobalt, 0x40);
drivers/media/pci/cobalt/cobalt-cpld.c
70
tmp = (rd * 74 * 2197) / (27 * 1000);
drivers/media/pci/cobalt/cobalt-cpld.c
72
rd = cpld_read(cobalt, 0x44);
drivers/media/pci/cobalt/cobalt-cpld.c
73
tmp = (rd * 74 * 2197) / (47 * 1000);
drivers/media/pci/cobalt/cobalt-cpld.c
75
rd = cpld_read(cobalt, 0x48);
drivers/media/pci/cobalt/cobalt-cpld.c
76
tmp = (rd * 57 * 2197) / (47 * 1000);
drivers/media/pci/cobalt/cobalt-cpld.c
78
rd = cpld_read(cobalt, 0x4c);
drivers/media/pci/cobalt/cobalt-cpld.c
79
tmp = (rd * 2197) / 1000;
drivers/media/pci/cobalt/cobalt-cpld.c
81
rd = cpld_read(cobalt, 0x50);
drivers/media/pci/cobalt/cobalt-cpld.c
82
tmp = (rd * 2197) / 1000;
drivers/media/pci/cobalt/cobalt-cpld.c
84
rd = cpld_read(cobalt, 0x54);
drivers/media/pci/cobalt/cobalt-cpld.c
85
tmp = (rd * 2197) / 1000;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
341
unsigned int wr, rd;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
346
rd = readl(q_dmem + FW_COM_RD_REG);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
348
if (WARN_ON_ONCE(wr >= q->size || rd >= q->size))
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
351
if (wr < rd)
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
352
packets = rd - wr - 1;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
354
packets = q->size - (wr - rd + 1);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
382
unsigned int wr, rd;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
386
rd = readl(q_dmem + FW_COM_RD_REG);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
388
if (WARN_ON_ONCE(wr >= q->size || rd >= q->size))
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
391
if (wr < rd)
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
394
packets = wr - rd;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
398
return (void *)((uintptr_t)q->host_address + rd * q->token_size);
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
406
unsigned int rd = readl(q_dmem + FW_COM_RD_REG) + 1;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
408
if (rd >= q->size)
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
409
rd = 0;
drivers/media/pci/intel/ipu6/ipu6-fw-com.c
411
writel(rd, q_dmem + FW_COM_RD_REG);
drivers/media/pci/pt3/pt3_i2c.c
107
bool rd;
drivers/media/pci/pt3/pt3_i2c.c
111
rd = !!(msgs[i].flags & I2C_M_RD);
drivers/media/pci/pt3/pt3_i2c.c
113
put_byte_write(cbuf, msgs[i].addr << 1 | rd);
drivers/media/pci/pt3/pt3_i2c.c
114
if (rd)
drivers/media/platform/amlogic/c3/mipi-adapter/c3-mipi-adap.c
208
void __iomem *rd;
drivers/media/platform/amlogic/c3/mipi-adapter/c3-mipi-adap.c
262
addr = adap->rd + ADAP_REG_ADDR(reg);
drivers/media/platform/amlogic/c3/mipi-adapter/c3-mipi-adap.c
745
adap->rd = devm_platform_ioremap_resource_byname(pdev, "rd");
drivers/media/platform/amlogic/c3/mipi-adapter/c3-mipi-adap.c
746
if (IS_ERR(adap->rd))
drivers/media/platform/amlogic/c3/mipi-adapter/c3-mipi-adap.c
747
return PTR_ERR(adap->rd);
drivers/media/tuners/qt1010.c
123
rd[2].val = reg05;
drivers/media/tuners/qt1010.c
126
rd[4].val = (freq + QT1010_OFFSET) / FREQ1;
drivers/media/tuners/qt1010.c
129
if (mod1 < 8000000) rd[6].val = 0x1d;
drivers/media/tuners/qt1010.c
130
else rd[6].val = 0x1c;
drivers/media/tuners/qt1010.c
133
if (mod1 < 1*FREQ2) rd[7].val = 0x09; /* +0 MHz */
drivers/media/tuners/qt1010.c
134
else if (mod1 < 2*FREQ2) rd[7].val = 0x08; /* +4 MHz */
drivers/media/tuners/qt1010.c
135
else if (mod1 < 3*FREQ2) rd[7].val = 0x0f; /* +8 MHz */
drivers/media/tuners/qt1010.c
136
else if (mod1 < 4*FREQ2) rd[7].val = 0x0e; /* +12 MHz */
drivers/media/tuners/qt1010.c
137
else if (mod1 < 5*FREQ2) rd[7].val = 0x0d; /* +16 MHz */
drivers/media/tuners/qt1010.c
138
else if (mod1 < 6*FREQ2) rd[7].val = 0x0c; /* +20 MHz */
drivers/media/tuners/qt1010.c
139
else if (mod1 < 7*FREQ2) rd[7].val = 0x0b; /* +24 MHz */
drivers/media/tuners/qt1010.c
140
else rd[7].val = 0x0a; /* +28 MHz */
drivers/media/tuners/qt1010.c
143
if (mod2 < 2000000) rd[8].val = 0x45;
drivers/media/tuners/qt1010.c
144
else rd[8].val = 0x44;
drivers/media/tuners/qt1010.c
148
rd[10].val = tmpval-((mod2/QT1010_STEP)*0x08);
drivers/media/tuners/qt1010.c
151
rd[13].val = 0xfd; /* TODO: correct value calculation */
drivers/media/tuners/qt1010.c
154
rd[14].val = 0x91; /* TODO: correct value calculation */
drivers/media/tuners/qt1010.c
157
if (freq < 450000000) rd[15].val = 0xd0; /* 450 MHz */
drivers/media/tuners/qt1010.c
158
else if (freq < 482000000) rd[15].val = 0xd1; /* 482 MHz */
drivers/media/tuners/qt1010.c
159
else if (freq < 514000000) rd[15].val = 0xd4; /* 514 MHz */
drivers/media/tuners/qt1010.c
160
else if (freq < 546000000) rd[15].val = 0xd7; /* 546 MHz */
drivers/media/tuners/qt1010.c
161
else if (freq < 610000000) rd[15].val = 0xda; /* 610 MHz */
drivers/media/tuners/qt1010.c
162
else rd[15].val = 0xd0;
drivers/media/tuners/qt1010.c
165
rd[35].val = (reg05 & 0xf0);
drivers/media/tuners/qt1010.c
174
rd[40].val = (priv->reg1f_init_val + 0x0e + tmpval);
drivers/media/tuners/qt1010.c
183
rd[41].val = (priv->reg20_init_val + 0x0d + tmpval);
drivers/media/tuners/qt1010.c
186
rd[43].val = priv->reg25_init_val;
drivers/media/tuners/qt1010.c
189
rd[45].val = 0x92; /* TODO: correct value calculation */
drivers/media/tuners/qt1010.c
195
freq, rd[2].val, rd[4].val, rd[6].val, rd[7].val, \
drivers/media/tuners/qt1010.c
196
rd[8].val, rd[10].val, rd[13].val, rd[14].val, \
drivers/media/tuners/qt1010.c
197
rd[15].val, rd[35].val, rd[40].val, rd[41].val, \
drivers/media/tuners/qt1010.c
198
rd[43].val, rd[45].val);
drivers/media/tuners/qt1010.c
200
for (i = 0; i < ARRAY_SIZE(rd); i++) {
drivers/media/tuners/qt1010.c
201
if (rd[i].oper == QT1010_WR) {
drivers/media/tuners/qt1010.c
202
err = qt1010_writereg(priv, rd[i].reg, rd[i].val);
drivers/media/tuners/qt1010.c
204
err = qt1010_readreg(priv, rd[i].reg, &tmpval);
drivers/media/tuners/qt1010.c
51
qt1010_i2c_oper_t rd[48] = {
drivers/memory/samsung/exynos-srom.c
158
struct exynos_srom_reg_dump *rd,
drivers/memory/samsung/exynos-srom.c
161
for (; num_regs > 0; --num_regs, ++rd)
drivers/memory/samsung/exynos-srom.c
162
rd->value = readl(base + rd->offset);
drivers/memory/samsung/exynos-srom.c
166
const struct exynos_srom_reg_dump *rd,
drivers/memory/samsung/exynos-srom.c
169
for (; num_regs > 0; --num_regs, ++rd)
drivers/memory/samsung/exynos-srom.c
170
writel(rd->value, base + rd->offset);
drivers/memory/samsung/exynos-srom.c
54
struct exynos_srom_reg_dump *rd;
drivers/memory/samsung/exynos-srom.c
57
rd = kzalloc_objs(*rd, nr_rdump);
drivers/memory/samsung/exynos-srom.c
58
if (!rd)
drivers/memory/samsung/exynos-srom.c
62
rd[i].offset = rdump[i];
drivers/memory/samsung/exynos-srom.c
64
return rd;
drivers/misc/bcm-vk/bcm_vk.h
274
u32 rd; /* read offset shadow */
drivers/misc/bcm-vk/bcm_vk_tty.c
106
while (vk->tty[i].rd != wr) {
drivers/misc/bcm-vk/bcm_vk_tty.c
108
VK_BAR_CHAN_DATA(vktty, from, vktty->rd));
drivers/misc/bcm-vk/bcm_vk_tty.c
109
vktty->rd++;
drivers/misc/bcm-vk/bcm_vk_tty.c
110
if (vktty->rd >= vktty->from_size)
drivers/misc/bcm-vk/bcm_vk_tty.c
111
vktty->rd = 0;
drivers/misc/bcm-vk/bcm_vk_tty.c
120
vkwrite32(vk, vktty->rd, BAR_1,
drivers/misc/bcm-vk/bcm_vk_tty.c
160
vktty->rd = vkread32(vk, BAR_1, VK_BAR_CHAN_RD(vktty, from));
drivers/misc/bcm-vk/bcm_vk_tty.c
28
u32 rd;
drivers/misc/bcm-vk/bcm_vk_tty.c
36
#define VK_BAR_CHAN_RD(v, DIR) VK_BAR_CHAN(v, DIR, rd)
drivers/mtd/nand/raw/nand_bbt.c
1017
if (rd) {
drivers/mtd/nand/raw/nand_bbt.c
1018
res = read_abs_bbt(this, buf, rd, chipsel);
drivers/mtd/nand/raw/nand_bbt.c
1021
rd->pages[i] = -1;
drivers/mtd/nand/raw/nand_bbt.c
1022
rd->version[i] = 0;
drivers/mtd/nand/raw/nand_bbt.c
955
struct nand_bbt_descr *rd, *rd2;
drivers/mtd/nand/raw/nand_bbt.c
966
rd = NULL;
drivers/mtd/nand/raw/nand_bbt.c
977
rd = md;
drivers/mtd/nand/raw/nand_bbt.c
980
rd = td;
drivers/mtd/nand/raw/nand_bbt.c
983
rd = td;
drivers/mtd/nand/raw/nand_bbt.c
987
rd = td;
drivers/mtd/nand/raw/nand_bbt.c
990
rd = md;
drivers/mtd/nand/raw/nand_bbt.c
998
rd = td;
drivers/mtd/spi-nor/sfdp.c
501
const struct sfdp_bfpt_read *rd = &sfdp_bfpt_reads[i];
drivers/mtd/spi-nor/sfdp.c
504
if (!(bfpt.dwords[rd->supported_dword] & rd->supported_bit)) {
drivers/mtd/spi-nor/sfdp.c
505
params->hwcaps.mask &= ~rd->hwcaps;
drivers/mtd/spi-nor/sfdp.c
509
params->hwcaps.mask |= rd->hwcaps;
drivers/mtd/spi-nor/sfdp.c
510
cmd = spi_nor_hwcaps_read2cmd(rd->hwcaps);
drivers/mtd/spi-nor/sfdp.c
512
half = bfpt.dwords[rd->settings_dword] >> rd->settings_shift;
drivers/mtd/spi-nor/sfdp.c
513
spi_nor_set_read_settings_from_bfpt(read, half, rd->proto);
drivers/net/can/grcan.c
1163
u32 wr, rd, startrd;
drivers/net/can/grcan.c
1168
rd = grcan_read_reg(&regs->rxrd);
drivers/net/can/grcan.c
1169
startrd = rd;
drivers/net/can/grcan.c
1173
if (rd == wr)
drivers/net/can/grcan.c
1185
slot = dma->rx.buf + rd;
drivers/net/can/grcan.c
1213
rd = grcan_ring_add(rd, GRCAN_MSG_SIZE, dma->rx.size);
drivers/net/can/grcan.c
1222
if (likely(rd != startrd))
drivers/net/can/grcan.c
1223
grcan_write_reg(&regs->rxrd, rd);
drivers/net/ethernet/alteon/acenic.c
1635
struct rx_desc *rd;
drivers/net/ethernet/alteon/acenic.c
1650
rd = &ap->rx_std_ring[idx];
drivers/net/ethernet/alteon/acenic.c
1651
set_aceaddr(&rd->addr, mapping);
drivers/net/ethernet/alteon/acenic.c
1652
rd->size = ACE_STD_BUFSIZE;
drivers/net/ethernet/alteon/acenic.c
1653
rd->idx = idx;
drivers/net/ethernet/alteon/acenic.c
1696
struct rx_desc *rd;
drivers/net/ethernet/alteon/acenic.c
1711
rd = &ap->rx_mini_ring[idx];
drivers/net/ethernet/alteon/acenic.c
1712
set_aceaddr(&rd->addr, mapping);
drivers/net/ethernet/alteon/acenic.c
1713
rd->size = ACE_MINI_BUFSIZE;
drivers/net/ethernet/alteon/acenic.c
1714
rd->idx = idx;
drivers/net/ethernet/alteon/acenic.c
1752
struct rx_desc *rd;
drivers/net/ethernet/alteon/acenic.c
1767
rd = &ap->rx_jumbo_ring[idx];
drivers/net/ethernet/alteon/acenic.c
1768
set_aceaddr(&rd->addr, mapping);
drivers/net/ethernet/alteon/acenic.c
1769
rd->size = ACE_JUMBO_BUFSIZE;
drivers/net/ethernet/alteon/acenic.c
1770
rd->idx = idx;
drivers/net/ethernet/amd/7990.c
276
volatile struct lance_rx_desc *rd;
drivers/net/ethernet/amd/7990.c
298
for (rd = &ib->brx_ring[lp->rx_new]; /* For each Rx ring we own... */
drivers/net/ethernet/amd/7990.c
299
!((bits = rd->rmd1_bits) & LE_R1_OWN);
drivers/net/ethernet/amd/7990.c
300
rd = &ib->brx_ring[lp->rx_new]) {
drivers/net/ethernet/amd/7990.c
322
int len = (rd->mblength & 0xfff) - 4;
drivers/net/ethernet/amd/7990.c
327
rd->mblength = 0;
drivers/net/ethernet/amd/7990.c
328
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/7990.c
345
rd->mblength = 0;
drivers/net/ethernet/amd/7990.c
346
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/a2065.c
246
volatile struct lance_rx_desc *rd;
drivers/net/ethernet/amd/a2065.c
266
for (rd = &ib->brx_ring[lp->rx_new];
drivers/net/ethernet/amd/a2065.c
267
!((bits = rd->rmd1_bits) & LE_R1_OWN);
drivers/net/ethernet/amd/a2065.c
268
rd = &ib->brx_ring[lp->rx_new]) {
drivers/net/ethernet/amd/a2065.c
290
int len = (rd->mblength & 0xfff) - 4;
drivers/net/ethernet/amd/a2065.c
295
rd->mblength = 0;
drivers/net/ethernet/amd/a2065.c
296
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/a2065.c
313
rd->mblength = 0;
drivers/net/ethernet/amd/a2065.c
314
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/declance.c
241
#define rds_ptr(rd, rt, type) \
drivers/net/ethernet/amd/declance.c
242
((volatile u16 *)((u8 *)(rd) + rds_off(rt, type)))
drivers/net/ethernet/amd/declance.c
560
volatile u16 *rd;
drivers/net/ethernet/amd/declance.c
584
for (rd = lib_ptr(ib, brx_ring[lp->rx_new], lp->type);
drivers/net/ethernet/amd/declance.c
585
!((bits = *rds_ptr(rd, rmd1, lp->type)) & LE_R1_OWN);
drivers/net/ethernet/amd/declance.c
586
rd = lib_ptr(ib, brx_ring[lp->rx_new], lp->type)) {
drivers/net/ethernet/amd/declance.c
608
len = (*rds_ptr(rd, mblength, lp->type) & 0xfff) - 4;
drivers/net/ethernet/amd/declance.c
613
*rds_ptr(rd, mblength, lp->type) = 0;
drivers/net/ethernet/amd/declance.c
614
*rds_ptr(rd, rmd1, lp->type) =
drivers/net/ethernet/amd/declance.c
634
*rds_ptr(rd, mblength, lp->type) = 0;
drivers/net/ethernet/amd/declance.c
635
*rds_ptr(rd, length, lp->type) = -RX_BUFF_SIZE | 0xf000;
drivers/net/ethernet/amd/declance.c
636
*rds_ptr(rd, rmd1, lp->type) =
drivers/net/ethernet/amd/sunlance.c
507
struct lance_rx_desc *rd;
drivers/net/ethernet/amd/sunlance.c
512
for (rd = &ib->brx_ring [entry];
drivers/net/ethernet/amd/sunlance.c
513
!((bits = rd->rmd1_bits) & LE_R1_OWN);
drivers/net/ethernet/amd/sunlance.c
514
rd = &ib->brx_ring [entry]) {
drivers/net/ethernet/amd/sunlance.c
530
len = (rd->mblength & 0xfff) - 4;
drivers/net/ethernet/amd/sunlance.c
535
rd->mblength = 0;
drivers/net/ethernet/amd/sunlance.c
536
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/sunlance.c
554
rd->mblength = 0;
drivers/net/ethernet/amd/sunlance.c
555
rd->rmd1_bits = LE_R1_OWN;
drivers/net/ethernet/amd/sunlance.c
676
struct lance_rx_desc __iomem *rd;
drivers/net/ethernet/amd/sunlance.c
682
for (rd = &ib->brx_ring [entry];
drivers/net/ethernet/amd/sunlance.c
683
!((bits = sbus_readb(&rd->rmd1_bits)) & LE_R1_OWN);
drivers/net/ethernet/amd/sunlance.c
684
rd = &ib->brx_ring [entry]) {
drivers/net/ethernet/amd/sunlance.c
700
len = (sbus_readw(&rd->mblength) & 0xfff) - 4;
drivers/net/ethernet/amd/sunlance.c
705
sbus_writew(0, &rd->mblength);
drivers/net/ethernet/amd/sunlance.c
706
sbus_writeb(LE_R1_OWN, &rd->rmd1_bits);
drivers/net/ethernet/amd/sunlance.c
722
sbus_writew(0, &rd->mblength);
drivers/net/ethernet/amd/sunlance.c
723
sbus_writeb(LE_R1_OWN, &rd->rmd1_bits);
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
316
void __iomem *addr, *rd, *cmd, *cmd_done;
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
329
rd = pdata->mcx_mac_addr + MAC_READ_REG_OFFSET;
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
343
rd_data = ioread32(rd);
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
352
void __iomem *addr, *rd, *cmd, *cmd_done;
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
357
rd = pdata->mcx_stats_addr + STAT_READ_REG_OFFSET;
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
372
rd_data = ioread32(rd);
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
107
static bool xgene_enet_rd_indirect(void __iomem *addr, void __iomem *rd,
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
124
*rd_data = ioread32(rd);
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
133
void __iomem *addr, *rd, *cmd, *cmd_done;
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
137
rd = pdata->pcs_addr + PCS_READ_REG_OFFSET;
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
141
success = xgene_enet_rd_indirect(addr, rd, cmd, cmd_done, rd_addr, rd_data);
drivers/net/ethernet/apple/bmac.c
1332
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
1349
rd->control = cpu_to_le32(DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE)); /* clear run bit */
drivers/net/ethernet/apple/bmac.c
1416
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
1444
cp = bus_to_virt(le32_to_cpu(rd->cmdptr));
drivers/net/ethernet/apple/bmac.c
1445
out_le32(&rd->control, DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE|ACTIVE|DEAD));
drivers/net/ethernet/apple/bmac.c
1447
out_le32(&rd->cmdptr, virt_to_bus(cp));
drivers/net/ethernet/apple/bmac.c
1448
out_le32(&rd->control, DBDMA_SET(RUN|WAKE));
drivers/net/ethernet/apple/bmac.c
224
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
227
if (rd)
drivers/net/ethernet/apple/bmac.c
228
dbdma_reset(rd);
drivers/net/ethernet/apple/bmac.c
401
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
405
dbdma_continue(rd);
drivers/net/ethernet/apple/bmac.c
472
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
481
rd->control = cpu_to_le32(DBDMA_CLEAR(RUN|PAUSE|FLUSH|WAKE)); /* clear run bit */
drivers/net/ethernet/apple/bmac.c
603
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
627
dbdma_reset(rd);
drivers/net/ethernet/apple/bmac.c
628
out_le32(&rd->cmdptr, virt_to_bus(bp->rx_cmds));
drivers/net/ethernet/apple/bmac.c
673
volatile struct dbdma_regs __iomem *rd = bp->rx_dma;
drivers/net/ethernet/apple/bmac.c
732
dbdma_continue(rd);
drivers/net/ethernet/apple/mace.c
435
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
475
out_le32(&rd->control, (RUN|PAUSE|FLUSH|WAKE) << 16); /* clear run bit */
drivers/net/ethernet/apple/mace.c
476
out_le32(&rd->cmdptr, virt_to_bus(mp->rx_cmds));
drivers/net/ethernet/apple/mace.c
477
out_le32(&rd->control, (RUN << 16) | RUN);
drivers/net/ethernet/apple/mace.c
505
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
513
rd->control = cpu_to_le32((RUN|PAUSE|FLUSH|WAKE) << 16); /* clear run bit */
drivers/net/ethernet/apple/mace.c
812
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
834
cp = bus_to_virt(le32_to_cpu(rd->cmdptr));
drivers/net/ethernet/apple/mace.c
835
dbdma_reset(rd);
drivers/net/ethernet/apple/mace.c
837
out_le32(&rd->cmdptr, virt_to_bus(cp));
drivers/net/ethernet/apple/mace.c
838
out_le32(&rd->control, (RUN << 16) | RUN);
drivers/net/ethernet/apple/mace.c
881
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
972
if ((le32_to_cpu(rd->status) & ACTIVE) != 0) {
drivers/net/ethernet/apple/mace.c
973
out_le32(&rd->control, (PAUSE << 16) | PAUSE);
drivers/net/ethernet/apple/mace.c
974
while ((in_le32(&rd->status) & ACTIVE) != 0)
drivers/net/ethernet/apple/mace.c
981
out_le32(&rd->control, ((RUN|WAKE) << 16) | (RUN|WAKE));
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1302
struct hns_nic_ring_data *rd;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1308
rd = &priv->ring_data[i];
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1310
if (rd->ring->irq_init_flag == RCB_IRQ_INITED)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1313
snprintf(rd->ring->ring_name, RCB_RING_NAME_LEN,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1315
(is_tx_ring(rd->ring) ? "tx" : "rx"), rd->queue_index);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1317
rd->ring->ring_name[RCB_RING_NAME_LEN - 1] = '\0';
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1319
irq_set_status_flags(rd->ring->irq, IRQ_NOAUTOEN);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1320
ret = request_irq(rd->ring->irq,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1321
hns_irq_handle, 0, rd->ring->ring_name, rd);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1324
rd->ring->irq);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1329
rd->ring, &rd->mask);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1332
irq_set_affinity_hint(rd->ring->irq,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1333
&rd->mask);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1335
rd->ring->irq_init_flag = RCB_IRQ_INITED;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1628
struct hns_nic_ring_data *rd;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1693
rd = &tx_ring_data(priv, skb->queue_mapping);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1694
hns_nic_net_xmit_hw(ndev, skb, rd);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1700
rd = &rx_ring_data(priv, i);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1701
if (rd->poll_one(rd, fetch_num,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1710
rd = &tx_ring_data(priv,
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1712
if (rd->poll_one(rd, fetch_num, NULL))
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2105
struct hns_nic_ring_data *rd;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2121
rd = &priv->ring_data[i];
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2122
rd->queue_index = i;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2123
rd->ring = &h->qs[i]->tx_ring;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2124
rd->poll_one = hns_nic_tx_poll_one;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2125
rd->fini_process = is_ver1 ? hns_nic_tx_fini_pro :
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2128
netif_napi_add(priv->netdev, &rd->napi, hns_nic_common_poll);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2129
rd->ring->irq_init_flag = RCB_IRQ_NOT_INITED;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2132
rd = &priv->ring_data[i];
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2133
rd->queue_index = i - h->q_num;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2134
rd->ring = &h->qs[i - h->q_num]->rx_ring;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2135
rd->poll_one = hns_nic_rx_poll_one;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2136
rd->ex_process = hns_nic_rx_up_pro;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2137
rd->fini_process = is_ver1 ? hns_nic_rx_fini_pro :
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2140
netif_napi_add(priv->netdev, &rd->napi, hns_nic_common_poll);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2141
rd->ring->irq_init_flag = RCB_IRQ_NOT_INITED;
drivers/net/ethernet/korina.c
599
struct dma_desc *rd = &lp->rd_ring[lp->rx_next_done];
drivers/net/ethernet/korina.c
609
devcs = rd->devcs;
drivers/net/ethernet/korina.c
611
if ((KORINA_RBSIZE - (u32)DMA_COUNT(rd->control)) == 0)
drivers/net/ethernet/korina.c
671
rd->devcs = 0;
drivers/net/ethernet/korina.c
674
rd->ca = lp->rx_skb_dma[lp->rx_next_done];
drivers/net/ethernet/korina.c
676
rd->control = DMA_COUNT(KORINA_RBSIZE) |
drivers/net/ethernet/korina.c
683
rd = &lp->rd_ring[lp->rx_next_done];
drivers/net/ethernet/korina.c
694
rd->devcs = 0;
drivers/net/ethernet/korina.c
695
rd->ca = lp->rx_skb_dma[lp->rx_next_done];
drivers/net/ethernet/korina.c
696
writel(korina_rx_dma(lp, rd - lp->rd_ring),
drivers/net/ethernet/marvell/pxa168_eth.c
438
u32 rd, u32 skip, int del)
drivers/net/ethernet/marvell/pxa168_eth.c
453
| (skip << SKIP) | (rd << HASH_ENTRY_RECEIVE_DISCARD_BIT)
drivers/net/ethernet/marvell/skge.c
3209
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
3214
control = rd->control;
drivers/net/ethernet/marvell/skge.c
3218
skb = skge_rx_get(dev, e, control, rd->status, rd->csum2);
drivers/net/ethernet/marvell/skge.c
944
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
953
rd->dma_lo = lower_32_bits(map);
drivers/net/ethernet/marvell/skge.c
954
rd->dma_hi = upper_32_bits(map);
drivers/net/ethernet/marvell/skge.c
956
rd->csum1_start = ETH_HLEN;
drivers/net/ethernet/marvell/skge.c
957
rd->csum2_start = ETH_HLEN;
drivers/net/ethernet/marvell/skge.c
958
rd->csum1 = 0;
drivers/net/ethernet/marvell/skge.c
959
rd->csum2 = 0;
drivers/net/ethernet/marvell/skge.c
963
rd->control = BMU_OWN | BMU_STF | BMU_IRQ_EOF | BMU_TCP_CHECK | bufsize;
drivers/net/ethernet/marvell/skge.c
975
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
977
rd->csum2 = 0;
drivers/net/ethernet/marvell/skge.c
978
rd->csum2_start = ETH_HLEN;
drivers/net/ethernet/marvell/skge.c
982
rd->control = BMU_OWN | BMU_STF | BMU_IRQ_EOF | BMU_TCP_CHECK | size;
drivers/net/ethernet/marvell/skge.c
995
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
996
rd->control = 0;
drivers/net/ethernet/micrel/ks8851_common.c
1000
if ((rd & both_done) != both_done) {
drivers/net/ethernet/micrel/ks8851_common.c
1005
if (rd & MBIR_TXMBFA)
drivers/net/ethernet/micrel/ks8851_common.c
1008
if (rd & MBIR_RXMBFA)
drivers/net/ethernet/micrel/ks8851_common.c
996
unsigned rd;
drivers/net/ethernet/micrel/ks8851_common.c
998
rd = ks8851_rdreg16(ks, KS_MBIR);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1305
ms->rd[0] = QLCNIC_MS_RDDATA_LO;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1307
ms->rd[1] = QLCNIC_MS_RDDATA_HI;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1310
ms->rd[2] = QLCNIC_MS_RDDATA_ULO;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1311
ms->rd[3] = QLCNIC_MS_RDDATA_UHI;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1314
ms->rd[0] = QLCNIC_MS_RDDATA_ULO;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1316
ms->rd[1] = QLCNIC_MS_RDDATA_UHI;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1319
ms->rd[2] = QLCNIC_MS_RDDATA_LO;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1320
ms->rd[3] = QLCNIC_MS_RDDATA_HI;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1372
qlcnic_ind_wr(adapter, ms.wd[0], qlcnic_ind_rd(adapter, ms.rd[0]));
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1373
qlcnic_ind_wr(adapter, ms.wd[1], qlcnic_ind_rd(adapter, ms.rd[1]));
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1447
temp = qlcnic_ind_rd(adapter, ms.rd[3]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1449
val |= qlcnic_ind_rd(adapter, ms.rd[2]);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
30
u32 rd[4];
drivers/net/ethernet/seeq/sgiseeq.c
345
struct sgiseeq_rx_desc *rd;
drivers/net/ethernet/seeq/sgiseeq.c
353
rd = &sp->rx_desc[sp->rx_new];
drivers/net/ethernet/seeq/sgiseeq.c
354
dma_sync_desc_cpu(dev, rd);
drivers/net/ethernet/seeq/sgiseeq.c
355
while (!(rd->rdma.cntinfo & HPCDMA_OWN)) {
drivers/net/ethernet/seeq/sgiseeq.c
356
len = PKT_BUF_SZ - (rd->rdma.cntinfo & HPCDMA_BCNT) - 3;
drivers/net/ethernet/seeq/sgiseeq.c
357
dma_unmap_single(dev->dev.parent, rd->rdma.pbuf,
drivers/net/ethernet/seeq/sgiseeq.c
359
pkt_status = rd->skb->data[len];
drivers/net/ethernet/seeq/sgiseeq.c
363
if (!ether_addr_equal(rd->skb->data + 6, dev->dev_addr)) {
drivers/net/ethernet/seeq/sgiseeq.c
365
skb = rd->skb;
drivers/net/ethernet/seeq/sgiseeq.c
376
skb_copy_to_linear_data(skb, rd->skb->data, len);
drivers/net/ethernet/seeq/sgiseeq.c
378
newskb = rd->skb;
drivers/net/ethernet/seeq/sgiseeq.c
392
newskb = rd->skb;
drivers/net/ethernet/seeq/sgiseeq.c
396
newskb = rd->skb;
drivers/net/ethernet/seeq/sgiseeq.c
398
rd->skb = newskb;
drivers/net/ethernet/seeq/sgiseeq.c
399
rd->rdma.pbuf = dma_map_single(dev->dev.parent,
drivers/net/ethernet/seeq/sgiseeq.c
404
rd->rdma.cntinfo = RCNTINFO_INIT;
drivers/net/ethernet/seeq/sgiseeq.c
406
dma_sync_desc_dev(dev, rd);
drivers/net/ethernet/seeq/sgiseeq.c
407
rd = &sp->rx_desc[sp->rx_new];
drivers/net/ethernet/seeq/sgiseeq.c
408
dma_sync_desc_cpu(dev, rd);
drivers/net/ethernet/seeq/sgiseeq.c
410
dma_sync_desc_dev(dev, rd);
drivers/net/ethernet/sun/niu.c
5906
u64 rd, wr, val;
drivers/net/ethernet/sun/niu.c
5909
rd = nr64_ipp(IPP_DFIFO_RD_PTR);
drivers/net/ethernet/sun/niu.c
5912
while (--limit >= 0 && (rd != wr)) {
drivers/net/ethernet/sun/niu.c
5913
rd = nr64_ipp(IPP_DFIFO_RD_PTR);
drivers/net/ethernet/sun/niu.c
5917
(rd != 0 && wr != 1)) {
drivers/net/ethernet/via/via-velocity.c
1534
struct rx_desc *rd = &(vptr->rx.ring[idx]);
drivers/net/ethernet/via/via-velocity.c
1554
*((u32 *) & (rd->rdesc0)) = 0;
drivers/net/ethernet/via/via-velocity.c
1555
rd->size = cpu_to_le16(vptr->rx.buf_sz) | RX_INTEN;
drivers/net/ethernet/via/via-velocity.c
1556
rd->pa_low = cpu_to_le32(rd_info->skb_dma);
drivers/net/ethernet/via/via-velocity.c
1557
rd->pa_high = 0;
drivers/net/ethernet/via/via-velocity.c
1567
struct rx_desc *rd = vptr->rx.ring + dirty;
drivers/net/ethernet/via/via-velocity.c
1570
if (rd->rdesc0.len & OWNED_BY_NIC)
drivers/net/ethernet/via/via-velocity.c
1605
struct rx_desc *rd = vptr->rx.ring + i;
drivers/net/ethernet/via/via-velocity.c
1607
memset(rd, 0, sizeof(*rd));
drivers/net/ethernet/via/via-velocity.c
1974
static inline void velocity_rx_csum(struct rx_desc *rd, struct sk_buff *skb)
drivers/net/ethernet/via/via-velocity.c
1978
if (rd->rdesc1.CSM & CSM_IPKT) {
drivers/net/ethernet/via/via-velocity.c
1979
if (rd->rdesc1.CSM & CSM_IPOK) {
drivers/net/ethernet/via/via-velocity.c
1980
if ((rd->rdesc1.CSM & CSM_TCPKT) ||
drivers/net/ethernet/via/via-velocity.c
1981
(rd->rdesc1.CSM & CSM_UDPKT)) {
drivers/net/ethernet/via/via-velocity.c
1982
if (!(rd->rdesc1.CSM & CSM_TUPOK))
drivers/net/ethernet/via/via-velocity.c
2050
struct rx_desc *rd = &(vptr->rx.ring[idx]);
drivers/net/ethernet/via/via-velocity.c
2051
int pkt_len = le16_to_cpu(rd->rdesc0.len) & 0x3fff;
drivers/net/ethernet/via/via-velocity.c
2054
if (unlikely(rd->rdesc0.RSR & (RSR_STP | RSR_EDP | RSR_RL))) {
drivers/net/ethernet/via/via-velocity.c
2055
if (rd->rdesc0.RSR & (RSR_STP | RSR_EDP))
drivers/net/ethernet/via/via-velocity.c
2061
if (rd->rdesc0.RSR & RSR_MAR)
drivers/net/ethernet/via/via-velocity.c
2069
velocity_rx_csum(rd, skb);
drivers/net/ethernet/via/via-velocity.c
2084
if (rd->rdesc0.RSR & RSR_DETAG) {
drivers/net/ethernet/via/via-velocity.c
2085
u16 vid = swab16(le16_to_cpu(rd->rdesc1.PQTAG));
drivers/net/ethernet/via/via-velocity.c
2113
struct rx_desc *rd = vptr->rx.ring + rd_curr;
drivers/net/ethernet/via/via-velocity.c
2118
if (rd->rdesc0.len & OWNED_BY_NIC)
drivers/net/ethernet/via/via-velocity.c
2126
if (rd->rdesc0.RSR & (RSR_RXOK | RSR_CE | RSR_RL)) {
drivers/net/ethernet/via/via-velocity.c
2130
if (rd->rdesc0.RSR & RSR_CRC)
drivers/net/ethernet/via/via-velocity.c
2132
if (rd->rdesc0.RSR & RSR_FAE)
drivers/net/ethernet/via/via-velocity.c
2138
rd->size |= RX_INTEN;
drivers/net/hamradio/hdlcdrv.c
445
s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0;
drivers/net/hamradio/hdlcdrv.c
449
s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0;
drivers/net/hamradio/hdlcdrv.c
579
if (s->bitbuf_channel.rd == s->bitbuf_channel.wr)
drivers/net/hamradio/hdlcdrv.c
582
s->bitbuf_channel.buffer[s->bitbuf_channel.rd];
drivers/net/hamradio/hdlcdrv.c
583
s->bitbuf_channel.rd = (s->bitbuf_channel.rd+1) %
drivers/net/hamradio/hdlcdrv.c
592
if (s->bitbuf_hdlc.rd == s->bitbuf_hdlc.wr)
drivers/net/hamradio/hdlcdrv.c
595
s->bitbuf_hdlc.buffer[s->bitbuf_hdlc.rd];
drivers/net/hamradio/hdlcdrv.c
596
s->bitbuf_hdlc.rd = (s->bitbuf_hdlc.rd+1) %
drivers/net/hamradio/hdlcdrv.c
644
s->hdlcrx.hbuf.rd = s->hdlcrx.hbuf.wr = 0;
drivers/net/hamradio/hdlcdrv.c
649
s->hdlctx.hbuf.rd = s->hdlctx.hbuf.wr = 0;
drivers/net/hamradio/hdlcdrv.c
659
s->bitbuf_channel.rd = s->bitbuf_channel.wr = 0;
drivers/net/hamradio/hdlcdrv.c
662
s->bitbuf_hdlc.rd = s->bitbuf_hdlc.wr = 0;
drivers/net/mdio/mdio-xgene.c
27
void __iomem *addr, *rd, *cmd, *cmd_done;
drivers/net/mdio/mdio-xgene.c
32
rd = pdata->mac_csr_addr + MAC_READ_REG_OFFSET;
drivers/net/mdio/mdio-xgene.c
44
rd_data = ioread32(rd);
drivers/net/vxlan/vxlan_core.c
1032
rc = vxlan_fdb_append(f, ip, port, vni, ifindex, &rd);
drivers/net/vxlan/vxlan_core.c
1043
if (rd == NULL)
drivers/net/vxlan/vxlan_core.c
1044
rd = first_remote_rtnl(f);
drivers/net/vxlan/vxlan_core.c
1047
err = vxlan_fdb_notify(vxlan, f, rd, RTM_NEWNEIGH,
drivers/net/vxlan/vxlan_core.c
1059
*rd = oldrd;
drivers/net/vxlan/vxlan_core.c
1061
list_del_rcu(&rd->list);
drivers/net/vxlan/vxlan_core.c
1062
call_rcu(&rd->rcu, vxlan_dst_free);
drivers/net/vxlan/vxlan_core.c
1136
struct vxlan_rdst *rd, bool swdev_notify)
drivers/net/vxlan/vxlan_core.c
1138
list_del_rcu(&rd->list);
drivers/net/vxlan/vxlan_core.c
1139
vxlan_fdb_notify(vxlan, f, rd, RTM_DELNEIGH, swdev_notify, NULL);
drivers/net/vxlan/vxlan_core.c
1140
call_rcu(&rd->rcu, vxlan_dst_free);
drivers/net/vxlan/vxlan_core.c
1278
struct vxlan_rdst *rd = NULL;
drivers/net/vxlan/vxlan_core.c
1287
rd = vxlan_fdb_find_rdst(f, &ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
1288
if (!rd)
drivers/net/vxlan/vxlan_core.c
1295
if (rd && !list_is_singular(&f->remotes)) {
drivers/net/vxlan/vxlan_core.c
1296
vxlan_fdb_dst_destroy(vxlan, f, rd, swdev_notify);
drivers/net/vxlan/vxlan_core.c
1347
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_core.c
1365
list_for_each_entry_rcu(rd, &f->remotes, list) {
drivers/net/vxlan/vxlan_core.c
1372
RTM_NEWNEIGH, NLM_F_MULTI, rd);
drivers/net/vxlan/vxlan_core.c
265
struct vxlan_rdst *rd, int type)
drivers/net/vxlan/vxlan_core.c
275
err = vxlan_fdb_info(skb, vxlan, fdb, 0, 0, type, 0, rd);
drivers/net/vxlan/vxlan_core.c
291
const struct vxlan_rdst *rd,
drivers/net/vxlan/vxlan_core.c
297
fdb_info->remote_ip = rd->remote_ip;
drivers/net/vxlan/vxlan_core.c
298
fdb_info->remote_port = rd->remote_port;
drivers/net/vxlan/vxlan_core.c
299
fdb_info->remote_vni = rd->remote_vni;
drivers/net/vxlan/vxlan_core.c
300
fdb_info->remote_ifindex = rd->remote_ifindex;
drivers/net/vxlan/vxlan_core.c
303
fdb_info->offloaded = rd->offloaded;
drivers/net/vxlan/vxlan_core.c
3034
const struct vxlan_rdst *rd)
drivers/net/vxlan/vxlan_core.c
3036
if (desc->vni && rd->remote_vni != desc->vni)
drivers/net/vxlan/vxlan_core.c
3039
if (desc->port && rd->remote_port != desc->port)
drivers/net/vxlan/vxlan_core.c
3043
!vxlan_addr_equal(&rd->remote_ip, &desc->dst_ip))
drivers/net/vxlan/vxlan_core.c
3055
struct vxlan_rdst *rd, *tmp;
drivers/net/vxlan/vxlan_core.c
3057
list_for_each_entry_safe(rd, tmp, &f->remotes, list) {
drivers/net/vxlan/vxlan_core.c
3058
if (!vxlan_fdb_flush_remote_matches(desc, rd))
drivers/net/vxlan/vxlan_core.c
3061
vxlan_fdb_dst_destroy(vxlan, f, rd, true);
drivers/net/vxlan/vxlan_core.c
309
struct vxlan_rdst *rd,
drivers/net/vxlan/vxlan_core.c
317
if (WARN_ON(!rd))
drivers/net/vxlan/vxlan_core.c
322
vxlan_fdb_switchdev_notifier_info(vxlan, fdb, rd, NULL, &info);
drivers/net/vxlan/vxlan_core.c
329
struct vxlan_rdst *rd, int type, bool swdev_notify,
drivers/net/vxlan/vxlan_core.c
334
if (swdev_notify && rd) {
drivers/net/vxlan/vxlan_core.c
337
err = vxlan_fdb_switchdev_call_notifiers(vxlan, fdb, rd,
drivers/net/vxlan/vxlan_core.c
343
vxlan_fdb_switchdev_call_notifiers(vxlan, fdb, rd,
drivers/net/vxlan/vxlan_core.c
349
__vxlan_fdb_notify(vxlan, fdb, rd, type);
drivers/net/vxlan/vxlan_core.c
431
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_core.c
433
list_for_each_entry(rd, &f->remotes, list) {
drivers/net/vxlan/vxlan_core.c
434
if (vxlan_addr_equal(&rd->remote_ip, ip) &&
drivers/net/vxlan/vxlan_core.c
435
rd->remote_port == port &&
drivers/net/vxlan/vxlan_core.c
436
rd->remote_vni == vni &&
drivers/net/vxlan/vxlan_core.c
437
rd->remote_ifindex == ifindex)
drivers/net/vxlan/vxlan_core.c
438
return rd;
drivers/net/vxlan/vxlan_core.c
552
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_core.c
554
rd = vxlan_fdb_find_rdst(f, ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
555
if (rd)
drivers/net/vxlan/vxlan_core.c
558
rd = list_first_entry_or_null(&f->remotes, struct vxlan_rdst, list);
drivers/net/vxlan/vxlan_core.c
559
if (!rd)
drivers/net/vxlan/vxlan_core.c
562
*oldrd = *rd;
drivers/net/vxlan/vxlan_core.c
563
dst_cache_reset(&rd->dst_cache);
drivers/net/vxlan/vxlan_core.c
564
rd->remote_ip = *ip;
drivers/net/vxlan/vxlan_core.c
565
rd->remote_port = port;
drivers/net/vxlan/vxlan_core.c
566
rd->remote_vni = vni;
drivers/net/vxlan/vxlan_core.c
567
rd->remote_ifindex = ifindex;
drivers/net/vxlan/vxlan_core.c
568
rd->offloaded = false;
drivers/net/vxlan/vxlan_core.c
577
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_core.c
579
rd = vxlan_fdb_find_rdst(f, ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
580
if (rd)
drivers/net/vxlan/vxlan_core.c
583
rd = kmalloc_obj(*rd, GFP_ATOMIC);
drivers/net/vxlan/vxlan_core.c
584
if (rd == NULL)
drivers/net/vxlan/vxlan_core.c
590
dst_cache_init(&rd->dst_cache, GFP_ATOMIC | __GFP_NOWARN);
drivers/net/vxlan/vxlan_core.c
592
rd->remote_ip = *ip;
drivers/net/vxlan/vxlan_core.c
593
rd->remote_port = port;
drivers/net/vxlan/vxlan_core.c
594
rd->offloaded = false;
drivers/net/vxlan/vxlan_core.c
595
rd->remote_vni = vni;
drivers/net/vxlan/vxlan_core.c
596
rd->remote_ifindex = ifindex;
drivers/net/vxlan/vxlan_core.c
598
list_add_tail_rcu(&rd->list, &f->remotes);
drivers/net/vxlan/vxlan_core.c
600
*rdp = rd;
drivers/net/vxlan/vxlan_core.c
863
struct vxlan_rdst *rd = NULL;
drivers/net/vxlan/vxlan_core.c
879
rc = vxlan_fdb_append(f, ip, port, vni, ifindex, &rd);
drivers/net/vxlan/vxlan_core.c
900
list_del(&rd->list);
drivers/net/vxlan/vxlan_core.c
901
dst_cache_destroy(&rd->dst_cache);
drivers/net/vxlan/vxlan_core.c
902
kfree(rd);
drivers/net/vxlan/vxlan_core.c
911
struct vxlan_rdst *rd, *nd;
drivers/net/vxlan/vxlan_core.c
921
list_for_each_entry_safe(rd, nd, &f->remotes, list) {
drivers/net/vxlan/vxlan_core.c
922
dst_cache_destroy(&rd->dst_cache);
drivers/net/vxlan/vxlan_core.c
923
kfree(rd);
drivers/net/vxlan/vxlan_core.c
938
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_core.c
948
list_for_each_entry(rd, &f->remotes, list)
drivers/net/vxlan/vxlan_core.c
949
vxlan_fdb_notify(vxlan, f, rd, RTM_DELNEIGH,
drivers/net/vxlan/vxlan_core.c
962
struct vxlan_rdst *rd = container_of(head, struct vxlan_rdst, rcu);
drivers/net/vxlan/vxlan_core.c
964
dst_cache_destroy(&rd->dst_cache);
drivers/net/vxlan/vxlan_core.c
965
kfree(rd);
drivers/net/vxlan/vxlan_core.c
978
struct vxlan_rdst *rd = NULL;
drivers/net/vxlan/vxlan_mdb.c
1093
struct vxlan_rdst *new_rd, *old_rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
1100
new_rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
1116
rcu_assign_pointer(remote->rd, old_rd);
drivers/net/vxlan/vxlan_mdb.c
1394
struct vxlan_rdst *rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
1398
!vxlan_addr_equal(&desc->remote_ip, &rd->remote_ip))
drivers/net/vxlan/vxlan_mdb.c
1404
remote_vni = rd->remote_vni ? : mdb_entry->key.vni;
drivers/net/vxlan/vxlan_mdb.c
1408
if (desc->remote_port && desc->remote_port != rd->remote_port)
drivers/net/vxlan/vxlan_mdb.c
167
struct vxlan_rdst *rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
1708
rcu_dereference(remote->rd), false);
drivers/net/vxlan/vxlan_mdb.c
1713
rcu_dereference(fremote->rd), false);
drivers/net/vxlan/vxlan_mdb.c
188
vxlan_nla_put_addr(skb, MDBA_MDB_EATTR_DST, &rd->remote_ip))
drivers/net/vxlan/vxlan_mdb.c
191
if (rd->remote_port && rd->remote_port != vxlan->cfg.dst_port &&
drivers/net/vxlan/vxlan_mdb.c
193
be16_to_cpu(rd->remote_port)))
drivers/net/vxlan/vxlan_mdb.c
196
if (rd->remote_vni != vxlan->default_dst.remote_vni &&
drivers/net/vxlan/vxlan_mdb.c
197
nla_put_u32(skb, MDBA_MDB_EATTR_VNI, be32_to_cpu(rd->remote_vni)))
drivers/net/vxlan/vxlan_mdb.c
200
if (rd->remote_ifindex &&
drivers/net/vxlan/vxlan_mdb.c
201
nla_put_u32(skb, MDBA_MDB_EATTR_IFINDEX, rd->remote_ifindex))
drivers/net/vxlan/vxlan_mdb.c
36
struct vxlan_rdst __rcu *rd;
drivers/net/vxlan/vxlan_mdb.c
677
struct vxlan_rdst *rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
679
if (vxlan_addr_equal(addr, &rd->remote_ip))
drivers/net/vxlan/vxlan_mdb.c
688
struct vxlan_rdst *rd = container_of(head, struct vxlan_rdst, rcu);
drivers/net/vxlan/vxlan_mdb.c
690
dst_cache_destroy(&rd->dst_cache);
drivers/net/vxlan/vxlan_mdb.c
691
kfree(rd);
drivers/net/vxlan/vxlan_mdb.c
697
struct vxlan_rdst *rd;
drivers/net/vxlan/vxlan_mdb.c
700
rd = kzalloc_obj(*rd);
drivers/net/vxlan/vxlan_mdb.c
701
if (!rd)
drivers/net/vxlan/vxlan_mdb.c
704
err = dst_cache_init(&rd->dst_cache, GFP_KERNEL);
drivers/net/vxlan/vxlan_mdb.c
708
rd->remote_ip = cfg->remote_ip;
drivers/net/vxlan/vxlan_mdb.c
709
rd->remote_port = cfg->remote_port;
drivers/net/vxlan/vxlan_mdb.c
710
rd->remote_vni = cfg->remote_vni;
drivers/net/vxlan/vxlan_mdb.c
711
rd->remote_ifindex = cfg->remote_ifindex;
drivers/net/vxlan/vxlan_mdb.c
712
rcu_assign_pointer(remote->rd, rd);
drivers/net/vxlan/vxlan_mdb.c
717
kfree(rd);
drivers/net/vxlan/vxlan_mdb.c
721
static void vxlan_mdb_remote_rdst_fini(struct vxlan_rdst *rd)
drivers/net/vxlan/vxlan_mdb.c
723
call_rcu(&rd->rcu, vxlan_mdb_rdst_free);
drivers/net/vxlan/vxlan_mdb.c
747
vxlan_mdb_remote_rdst_fini(rtnl_dereference(remote->rd));
drivers/net/vxlan/vxlan_mdb.c
819
struct vxlan_rdst *rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
828
sg_cfg.remote_ip = rd->remote_ip;
drivers/net/vxlan/vxlan_mdb.c
942
struct vxlan_rdst *rd = rtnl_dereference(remote->rd);
drivers/net/vxlan/vxlan_mdb.c
960
nlmsg_size += nla_total_size(vxlan_addr_size(&rd->remote_ip));
drivers/net/vxlan/vxlan_mdb.c
962
if (rd->remote_port && rd->remote_port != vxlan->cfg.dst_port)
drivers/net/vxlan/vxlan_mdb.c
965
if (rd->remote_vni != vxlan->default_dst.remote_vni)
drivers/net/vxlan/vxlan_mdb.c
968
if (rd->remote_ifindex)
drivers/net/wireless/ath/ath10k/mac.c
9908
static int ath10k_mac_get_wrdd_regulatory(struct ath10k *ar, u16 *rd)
drivers/net/wireless/ath/ath10k/mac.c
9947
*rd = ath_regd_find_country_by_name(alpha2);
drivers/net/wireless/ath/ath10k/mac.c
9948
if (*rd == 0xffff)
drivers/net/wireless/ath/ath10k/mac.c
9951
*rd |= COUNTRY_ERD_FLAG;
drivers/net/wireless/ath/ath10k/mac.c
9958
u16 rd;
drivers/net/wireless/ath/ath10k/mac.c
9960
ret = ath10k_mac_get_wrdd_regulatory(ar, &rd);
drivers/net/wireless/ath/ath10k/mac.c
9964
rd = ar->hw_eeprom_rd;
drivers/net/wireless/ath/ath10k/mac.c
9967
ar->ath_common.regulatory.current_rd = rd;
drivers/net/wireless/ath/ath10k/wmi-ops.h
517
ath10k_wmi_pdev_set_regdomain(struct ath10k *ar, u16 rd, u16 rd2g, u16 rd5g,
drivers/net/wireless/ath/ath10k/wmi-ops.h
526
skb = ar->wmi.ops->gen_pdev_set_rd(ar, rd, rd2g, rd5g, ctl2g, ctl5g,
drivers/net/wireless/ath/ath10k/wmi-ops.h
63
struct sk_buff *(*gen_pdev_set_rd)(struct ath10k *ar, u16 rd, u16 rd2g,
drivers/net/wireless/ath/ath10k/wmi-tlv.c
1758
u16 rd, u16 rd2g, u16 rd5g,
drivers/net/wireless/ath/ath10k/wmi-tlv.c
1774
cmd->regd = __cpu_to_le32(rd);
drivers/net/wireless/ath/ath10k/wmi.c
6463
ath10k_wmi_op_gen_pdev_set_rd(struct ath10k *ar, u16 rd, u16 rd2g, u16 rd5g,
drivers/net/wireless/ath/ath10k/wmi.c
6475
cmd->reg_domain = __cpu_to_le32(rd);
drivers/net/wireless/ath/ath10k/wmi.c
6483
rd, rd2g, rd5g, ctl2g, ctl5g);
drivers/net/wireless/ath/ath10k/wmi.c
6488
ath10k_wmi_10x_op_gen_pdev_set_rd(struct ath10k *ar, u16 rd, u16 rd2g, u16
drivers/net/wireless/ath/ath10k/wmi.c
6500
cmd->reg_domain = __cpu_to_le32(rd);
drivers/net/wireless/ath/ath10k/wmi.c
6509
rd, rd2g, rd5g, ctl2g, ctl5g, dfs_reg);
drivers/net/wireless/ath/ath5k/debug.c
1050
struct ath5k_hw_all_rx_desc *rd = &ds->ud.ds_rx;
drivers/net/wireless/ath/ath5k/debug.c
1055
rd->rx_ctl.rx_control_0, rd->rx_ctl.rx_control_1,
drivers/net/wireless/ath/ath5k/debug.c
1056
rd->rx_stat.rx_status_0, rd->rx_stat.rx_status_1,
drivers/net/wireless/ath/regd.c
553
u16 rd = ath_regd_get_eepromRD(reg);
drivers/net/wireless/ath/regd.c
556
if (rd & COUNTRY_ERD_FLAG) {
drivers/net/wireless/ath/regd.c
558
u16 cc = rd & ~COUNTRY_ERD_FLAG;
drivers/net/wireless/ath/regd.c
567
if (rd != CTRY_DEFAULT)
drivers/net/wireless/ath/regd.c
571
if (regDomainPairs[i].reg_domain == rd)
drivers/net/wireless/ath/regd.c
575
"ath: invalid regulatory domain/country code 0x%x\n", rd);
drivers/net/wireless/ath/regd.c
606
static u16 ath_regd_get_default_country(u16 rd)
drivers/net/wireless/ath/regd.c
608
if (rd & COUNTRY_ERD_FLAG) {
drivers/net/wireless/ath/regd.c
610
u16 cc = rd & ~COUNTRY_ERD_FLAG;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1837
struct brcmf_skb_reorder_data *rd;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1897
rd = (struct brcmf_skb_reorder_data *)skb->cb;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1898
rd->reorder = data;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h
105
struct brcmf_skb_reorder_data *rd;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h
107
rd = (struct brcmf_skb_reorder_data *)skb->cb;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/proto.h
108
return !!rd->reorder;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1371
struct brcmf_sdio_hdrinfo *rd,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1399
(roundup(len, bus->blocksize) != rd->len)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1403
if (type == BRCMF_SDIO_FT_SUB && len > rd->len) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1407
rd->len = len;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1414
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1418
rd->channel = (swheader & SDPCM_CHANNEL_MASK) >> SDPCM_CHANNEL_SHIFT;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1419
if (len > MAX_RX_DATASZ && rd->channel != SDPCM_CONTROL_CHANNEL &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1424
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1427
if (type == BRCMF_SDIO_FT_SUPER && rd->channel != SDPCM_GLOM_CHANNEL) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1429
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1432
if (type == BRCMF_SDIO_FT_SUB && rd->channel != SDPCM_DATA_CHANNEL &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1433
rd->channel != SDPCM_EVENT_CHANNEL) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1435
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1438
rd->dat_offset = brcmf_sdio_getdatoffset(header);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1439
if (rd->dat_offset < SDPCM_HDRLEN || rd->dat_offset > rd->len) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1443
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1446
if (rd->seq_num != rx_seq) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1447
brcmf_dbg(SDIO, "seq %d, expected %d\n", rx_seq, rd->seq_num);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1449
rd->seq_num = rx_seq;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1454
rd->len_nxtfrm = (swheader & SDPCM_NEXTLEN_MASK) >> SDPCM_NEXTLEN_SHIFT;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1455
if (rd->len_nxtfrm << 4 > MAX_RX_DATASZ) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1457
if (rd->channel != SDPCM_GLOM_CHANNEL)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1459
rd->len_nxtfrm = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1863
struct brcmf_sdio_hdrinfo *rd = &bus->cur_read, rd_new;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1871
for (rd->seq_num = bus->rx_seq, rxleft = maxframes;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1873
rd->seq_num++, rxleft--) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1880
cnt = brcmf_sdio_rxglom(bus, rd->seq_num);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1882
rd->seq_num += cnt - 1;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1887
rd->len_left = rd->len;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1890
if (!rd->len) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1907
if (brcmf_sdio_hdparse(bus, bus->rxhdr, rd,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1916
if (rd->channel == SDPCM_CONTROL_CHANNEL) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1918
rd->len,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1919
rd->dat_offset);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1921
rd->len = rd->len_nxtfrm << 4;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1922
rd->len_nxtfrm = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1924
rd->channel = SDPCM_EVENT_CHANNEL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1928
rd->len_left = rd->len > BRCMF_FIRSTREAD ?
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1929
rd->len - BRCMF_FIRSTREAD : 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1933
brcmf_sdio_pad(bus, &pad, &rd->len_left);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1935
pkt = brcmu_pkt_buf_get_skb(rd->len_left + head_read +
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1941
RETRYCHAN(rd->channel));
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1946
pkt_align(pkt, rd->len_left, bus->head_align);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1954
rd->len, rd->channel, ret);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1958
RETRYCHAN(rd->channel));
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1969
rd_new.seq_num = rd->seq_num;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1973
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1980
if (rd->len != roundup(rd_new.len, 16)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1982
rd->len,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1984
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1991
rd->len_nxtfrm = rd_new.len_nxtfrm;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1992
rd->channel = rd_new.channel;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1993
rd->dat_offset = rd_new.dat_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2005
rd->len = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2015
pkt->data, rd->len, "Rx Data:\n");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2018
if (rd->channel == SDPCM_GLOM_CHANNEL) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2021
rd->len);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2023
pkt->data, rd->len,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2025
__skb_trim(pkt, rd->len);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2036
rd->len = rd->len_nxtfrm << 4;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2037
rd->len_nxtfrm = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2039
rd->channel = SDPCM_EVENT_CHANNEL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2044
__skb_trim(pkt, rd->len);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2045
skb_pull(pkt, rd->dat_offset);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2049
else if (rd->channel == SDPCM_EVENT_CHANNEL)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2056
rd->len = rd->len_nxtfrm << 4;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2057
rd->len_nxtfrm = 0;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2059
rd->channel = SDPCM_EVENT_CHANNEL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2070
rd->seq_num--;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2071
bus->rx_seq = rd->seq_num;
drivers/net/wireless/intel/iwlwifi/mei/main.c
1040
u32 rd = *_rd;
drivers/net/wireless/intel/iwlwifi/mei/main.c
1042
if (rd + len <= q_sz) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
1043
memcpy(buf, q_head + rd, len);
drivers/net/wireless/intel/iwlwifi/mei/main.c
1044
rd += len;
drivers/net/wireless/intel/iwlwifi/mei/main.c
1046
memcpy(buf, q_head + rd, q_sz - rd);
drivers/net/wireless/intel/iwlwifi/mei/main.c
1047
memcpy(buf + q_sz - rd, q_head, len - (q_sz - rd));
drivers/net/wireless/intel/iwlwifi/mei/main.c
1048
rd = len - (q_sz - rd);
drivers/net/wireless/intel/iwlwifi/mei/main.c
1051
*_rd = rd;
drivers/net/wireless/intel/iwlwifi/mei/main.c
1060
u32 rd, u32 wr, ssize_t valid_rx_sz,
drivers/net/wireless/intel/iwlwifi/mei/main.c
1077
iwl_mei_read_from_q(q_head, q_sz, &rd, wr, &hdr, sizeof(hdr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
1110
iwl_mei_read_from_q(q_head, q_sz, &rd, wr,
drivers/net/wireless/intel/iwlwifi/mei/main.c
1119
iwl_mei_read_from_q(q_head, q_sz, &rd, wr, data, len);
drivers/net/wireless/intel/iwlwifi/mei/main.c
1134
u32 rd, u32 wr, ssize_t valid_rx_sz)
drivers/net/wireless/intel/iwlwifi/mei/main.c
1147
iwl_mei_read_from_q(q_head, q_sz, &rd, wr, hdr, sizeof(*hdr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
1154
iwl_mei_read_from_q(q_head, q_sz, &rd, wr, hdr + 1, len);
drivers/net/wireless/intel/iwlwifi/mei/main.c
1175
u32 rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
1179
if (rd > q_sz || wr > q_sz) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
1185
if (rd == wr)
drivers/net/wireless/intel/iwlwifi/mei/main.c
1188
valid_rx_sz = wr > rd ? wr - rd : q_sz - rd + wr;
drivers/net/wireless/intel/iwlwifi/mei/main.c
1191
iwl_mei_handle_sap_data(cldev, q_head, q_sz, rd, wr,
drivers/net/wireless/intel/iwlwifi/mei/main.c
1194
iwl_mei_handle_sap_rx_cmd(cldev, q_head, q_sz, rd, wr,
drivers/net/wireless/intel/iwlwifi/mei/main.c
337
u32 rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
342
if (rd > q_sz || wr > q_sz) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
348
room_in_buf = wr >= rd ? q_sz - wr + rd : rd - wr;
drivers/net/wireless/intel/iwlwifi/mei/main.c
479
u32 rd;
drivers/net/wireless/intel/iwlwifi/mei/main.c
512
rd = le32_to_cpu(READ_ONCE(notif_q->rd_ptr));
drivers/net/wireless/intel/iwlwifi/mei/main.c
518
if (rd > q_sz || wr > q_sz) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
524
room_in_buf = wr >= rd ? q_sz - wr + rd : rd - wr;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1013
rd->val[0] = mt7615_mac_tx_rate_val(dev, mphy, &rates[0], stbc, &bw);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1017
rd->probe_val = mt7615_mac_tx_rate_val(dev, mphy, probe_rate,
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1020
rd->bw_idx = 1;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1024
rd->probe_val = rd->val[0];
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1027
rd->val[1] = mt7615_mac_tx_rate_val(dev, mphy, &rates[1], stbc, &bw);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1029
rd->bw_idx = 3;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1033
rd->val[2] = mt7615_mac_tx_rate_val(dev, mphy, &rates[2], stbc, &bw);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1035
rd->bw_idx = 5;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1039
rd->val[3] = mt7615_mac_tx_rate_val(dev, mphy, &rates[3], stbc, &bw);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1041
rd->bw_idx = 7;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1043
rd->rateset = rateset;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1044
rd->bw = bw;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1100
struct mt7615_rate_desc rd;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1112
memset(&rd, 0, sizeof(struct mt7615_rate_desc));
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1113
mt7615_mac_update_rate_desc(phy, sta, probe_rate, rates, &rd);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1118
w27 |= FIELD_PREP(MT_WTBL_W27_CC_BW_SEL, rd.bw);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1125
w5 |= FIELD_PREP(MT_WTBL_W5_BW_CAP, rd.bw) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1127
rd.bw_idx ? rd.bw_idx - 1 : 7);
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1132
FIELD_PREP(MT_WTBL_RIUCR1_RATE0, rd.probe_val) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1133
FIELD_PREP(MT_WTBL_RIUCR1_RATE1, rd.val[0]) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1134
FIELD_PREP(MT_WTBL_RIUCR1_RATE2_LO, rd.val[1]));
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1137
FIELD_PREP(MT_WTBL_RIUCR2_RATE2_HI, rd.val[1] >> 8) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1138
FIELD_PREP(MT_WTBL_RIUCR2_RATE3, rd.val[1]) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1139
FIELD_PREP(MT_WTBL_RIUCR2_RATE4, rd.val[2]) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1140
FIELD_PREP(MT_WTBL_RIUCR2_RATE5_LO, rd.val[2]));
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1143
FIELD_PREP(MT_WTBL_RIUCR3_RATE5_HI, rd.val[2] >> 4) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1144
FIELD_PREP(MT_WTBL_RIUCR3_RATE6, rd.val[3]) |
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1145
FIELD_PREP(MT_WTBL_RIUCR3_RATE7, rd.val[3]));
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
1159
sta->rate_set_tsf |= rd.rateset;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
961
struct mt7615_rate_desc *rd)
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1202
regdomain_is_known = isalpha(mac->rd->alpha2[0]) &&
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1203
isalpha(mac->rd->alpha2[1]);
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1208
if (mac->rd->alpha2[0] == '9' && mac->rd->alpha2[1] == '9') {
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1211
wiphy_apply_custom_regulatory(wiphy, mac->rd);
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1237
ret = regulatory_set_wiphy_regd(wiphy, mac->rd);
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
1239
ret = regulatory_hint(wiphy, mac->rd->alpha2);
drivers/net/wireless/quantenna/qtnfmac/commands.c
1034
mac->rd = kzalloc_flex(*mac->rd, reg_rules, resp->n_reg_rules);
drivers/net/wireless/quantenna/qtnfmac/commands.c
1035
if (!mac->rd)
drivers/net/wireless/quantenna/qtnfmac/commands.c
1038
mac->rd->n_reg_rules = resp->n_reg_rules;
drivers/net/wireless/quantenna/qtnfmac/commands.c
1039
mac->rd->alpha2[0] = resp->alpha2[0];
drivers/net/wireless/quantenna/qtnfmac/commands.c
1040
mac->rd->alpha2[1] = resp->alpha2[1];
drivers/net/wireless/quantenna/qtnfmac/commands.c
1044
mac->rd->dfs_region = NL80211_DFS_FCC;
drivers/net/wireless/quantenna/qtnfmac/commands.c
1047
mac->rd->dfs_region = NL80211_DFS_ETSI;
drivers/net/wireless/quantenna/qtnfmac/commands.c
1050
mac->rd->dfs_region = NL80211_DFS_JP;
drivers/net/wireless/quantenna/qtnfmac/commands.c
1054
mac->rd->dfs_region = NL80211_DFS_UNSET;
drivers/net/wireless/quantenna/qtnfmac/commands.c
1157
rule = &mac->rd->reg_rules[rule_idx++];
drivers/net/wireless/quantenna/qtnfmac/core.c
538
kfree(mac->rd);
drivers/net/wireless/quantenna/qtnfmac/core.c
539
mac->rd = NULL;
drivers/net/wireless/quantenna/qtnfmac/core.h
107
struct ieee80211_regdomain *rd;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
180
struct rt2x00_async_read_data *rd = urb->context;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
181
if (rd->callback(rd->rt2x00dev, urb->status, le32_to_cpu(rd->reg))) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
182
usb_anchor_urb(urb, rd->rt2x00dev->anchor);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
185
kfree(rd);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
188
kfree(rd);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
197
struct rt2x00_async_read_data *rd;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
199
rd = kmalloc_obj(*rd, GFP_ATOMIC);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
200
if (!rd)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
205
kfree(rd);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
209
rd->rt2x00dev = rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
210
rd->callback = callback;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
211
rd->cr.bRequestType = USB_VENDOR_REQUEST_IN;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
212
rd->cr.bRequest = USB_MULTI_READ;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
213
rd->cr.wValue = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
214
rd->cr.wIndex = cpu_to_le16(offset);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
215
rd->cr.wLength = cpu_to_le16(sizeof(u32));
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
218
(u8 *)(&rd->cr), &rd->reg, sizeof(rd->reg),
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
219
rt2x00usb_register_read_async_cb, rd);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
223
kfree(rd);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1648
struct reg_data *rd = &regs->regs[i];
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1649
if (rd->addr != req->addr[i]) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1652
le16_to_cpu(rd->addr),
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1686
struct reg_data *rd = &regs->regs[i];
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1687
values[i] = le16_to_cpu(rd->value);
drivers/of/dynamic.c
241
struct of_reconfig_data rd;
drivers/of/dynamic.c
243
memset(&rd, 0, sizeof(rd));
drivers/of/dynamic.c
244
rd.dn = np;
drivers/of/dynamic.c
250
of_reconfig_notify(OF_RECONFIG_ATTACH_NODE, &rd);
drivers/of/dynamic.c
295
struct of_reconfig_data rd;
drivers/of/dynamic.c
297
memset(&rd, 0, sizeof(rd));
drivers/of/dynamic.c
298
rd.dn = np;
drivers/of/dynamic.c
304
of_reconfig_notify(OF_RECONFIG_DETACH_NODE, &rd);
drivers/of/dynamic.c
572
struct of_reconfig_data rd;
drivers/of/dynamic.c
584
memset(&rd, 0, sizeof(rd));
drivers/of/dynamic.c
585
rd.dn = ce->np;
drivers/of/dynamic.c
586
ret = of_reconfig_notify(ce->action, &rd);
drivers/of/of_reserved_mem.c
626
struct rmem_assigned_device *rd;
drivers/of/of_reserved_mem.c
649
rd = kmalloc_obj(struct rmem_assigned_device);
drivers/of/of_reserved_mem.c
650
if (!rd)
drivers/of/of_reserved_mem.c
655
rd->dev = dev;
drivers/of/of_reserved_mem.c
656
rd->rmem = rmem;
drivers/of/of_reserved_mem.c
659
list_add(&rd->list, &of_rmem_assigned_device_list);
drivers/of/of_reserved_mem.c
664
kfree(rd);
drivers/of/of_reserved_mem.c
699
struct rmem_assigned_device *rd, *tmp;
drivers/of/of_reserved_mem.c
703
list_for_each_entry_safe(rd, tmp, &of_rmem_assigned_device_list, list) {
drivers/of/of_reserved_mem.c
704
if (rd->dev == dev)
drivers/of/of_reserved_mem.c
705
list_move_tail(&rd->list, &release_list);
drivers/of/of_reserved_mem.c
709
list_for_each_entry_safe(rd, tmp, &release_list, list) {
drivers/of/of_reserved_mem.c
710
if (rd->rmem && rd->rmem->ops && rd->rmem->ops->device_release)
drivers/of/of_reserved_mem.c
711
rd->rmem->ops->device_release(rd->rmem, dev);
drivers/of/of_reserved_mem.c
713
kfree(rd);
drivers/of/platform.c
724
struct of_reconfig_data *rd = arg;
drivers/of/platform.c
729
switch (of_reconfig_get_state_change(action, rd)) {
drivers/of/platform.c
731
parent = rd->dn->parent;
drivers/of/platform.c
738
if (of_node_check_flag(rd->dn, OF_POPULATED))
drivers/of/platform.c
745
rd->dn->fwnode.flags &= ~FWNODE_FLAG_NOT_DEVICE;
drivers/of/platform.c
748
pdev = of_platform_device_create(rd->dn, NULL,
drivers/of/platform.c
754
__func__, rd->dn);
drivers/of/platform.c
763
if (!of_node_check_flag(rd->dn, OF_POPULATED))
drivers/of/platform.c
767
pdev = of_find_device_by_node(rd->dn);
drivers/perf/arm_pmuv3.c
1608
struct clock_read_data *rd;
drivers/perf/arm_pmuv3.c
1625
rd = sched_clock_read_begin(&seq);
drivers/perf/arm_pmuv3.c
1627
if (rd->read_sched_clock != arch_timer_read_counter)
drivers/perf/arm_pmuv3.c
1630
userpg->time_mult = rd->mult;
drivers/perf/arm_pmuv3.c
1631
userpg->time_shift = rd->shift;
drivers/perf/arm_pmuv3.c
1632
userpg->time_zero = rd->epoch_ns;
drivers/perf/arm_pmuv3.c
1633
userpg->time_cycles = rd->epoch_cyc;
drivers/perf/arm_pmuv3.c
1634
userpg->time_mask = rd->sched_clock_mask;
drivers/perf/arm_pmuv3.c
1641
ns = mul_u64_u32_shr(rd->epoch_cyc, rd->mult, rd->shift);
drivers/perf/riscv_pmu.c
33
struct clock_read_data *rd;
drivers/perf/riscv_pmu.c
51
rd = sched_clock_read_begin(&seq);
drivers/perf/riscv_pmu.c
53
userpg->time_mult = rd->mult;
drivers/perf/riscv_pmu.c
54
userpg->time_shift = rd->shift;
drivers/perf/riscv_pmu.c
55
userpg->time_zero = rd->epoch_ns;
drivers/perf/riscv_pmu.c
56
userpg->time_cycles = rd->epoch_cyc;
drivers/perf/riscv_pmu.c
57
userpg->time_mask = rd->sched_clock_mask;
drivers/perf/riscv_pmu.c
64
ns = mul_u64_u32_shr(rd->epoch_cyc, rd->mult, rd->shift);
drivers/perf/xgene_pmu.c
329
XGENE_PMU_EVENT_ATTR(rd-cmd-sent, 0x04),
drivers/perf/xgene_pmu.c
337
XGENE_PMU_EVENT_ATTR(rd-rda-cmd-sent, 0x0c),
drivers/perf/xgene_pmu.c
339
XGENE_PMU_EVENT_ATTR(in-rd-collision, 0x0e),
drivers/perf/xgene_pmu.c
344
XGENE_PMU_EVENT_ATTR(mcu-rd-request, 0x13),
drivers/perf/xgene_pmu.c
345
XGENE_PMU_EVENT_ATTR(mcu-hp-rd-request, 0x14),
drivers/perf/xgene_pmu.c
347
XGENE_PMU_EVENT_ATTR(mcu-rd-proceed-all, 0x16),
drivers/perf/xgene_pmu.c
348
XGENE_PMU_EVENT_ATTR(mcu-rd-proceed-cancel, 0x17),
drivers/perf/xgene_pmu.c
349
XGENE_PMU_EVENT_ATTR(mcu-rd-response, 0x18),
drivers/perf/xgene_pmu.c
350
XGENE_PMU_EVENT_ATTR(mcu-rd-proceed-speculative-all, 0x19),
drivers/perf/xgene_pmu.c
351
XGENE_PMU_EVENT_ATTR(mcu-rd-proceed-speculative-cancel, 0x1a),
drivers/perf/xgene_pmu.c
423
XGENE_PMU_EVENT_ATTR(pa-req-buf-alloc-rd, 0x02),
drivers/perf/xgene_pmu.c
428
XGENE_PMU_EVENT_ATTR(pa-cp-rd-req, 0x07),
drivers/perf/xgene_pmu.c
431
XGENE_PMU_EVENT_ATTR(ba-rd-req, 0x0a),
drivers/perf/xgene_pmu.c
433
XGENE_PMU_EVENT_ATTR(pa-rd-shared-req-issued, 0x10),
drivers/perf/xgene_pmu.c
434
XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-req-issued, 0x11),
drivers/perf/xgene_pmu.c
440
XGENE_PMU_EVENT_ATTR(pa-ptl-rd-req, 0x17),
drivers/perf/xgene_pmu.c
446
XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop, 0x21),
drivers/perf/xgene_pmu.c
447
XGENE_PMU_EVENT_ATTR(pa-rd-shared-snoop-hit, 0x22),
drivers/perf/xgene_pmu.c
448
XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop, 0x23),
drivers/perf/xgene_pmu.c
449
XGENE_PMU_EVENT_ATTR(pa-rd-exclusive-snoop-hit, 0x24),
drivers/perf/xgene_pmu.c
450
XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop, 0x25),
drivers/perf/xgene_pmu.c
451
XGENE_PMU_EVENT_ATTR(pa-rd-wr-invalid-snoop-hit, 0x26),
drivers/perf/xgene_pmu.c
465
XGENE_PMU_EVENT_ATTR(pa-axi0-rd-req, 0x01),
drivers/perf/xgene_pmu.c
467
XGENE_PMU_EVENT_ATTR(pa-axi1-rd-req, 0x03),
drivers/perf/xgene_pmu.c
470
XGENE_PMU_EVENT_ATTR(ba-axi-rd-req, 0x08),
drivers/perf/xgene_pmu.c
479
XGENE_PMU_EVENT_ATTR(rd-req-recv, 0x02),
drivers/perf/xgene_pmu.c
480
XGENE_PMU_EVENT_ATTR(rd-req-recv-2, 0x03),
drivers/perf/xgene_pmu.c
483
XGENE_PMU_EVENT_ATTR(rd-req-sent-to-mcu, 0x06),
drivers/perf/xgene_pmu.c
484
XGENE_PMU_EVENT_ATTR(rd-req-sent-to-mcu-2, 0x07),
drivers/perf/xgene_pmu.c
485
XGENE_PMU_EVENT_ATTR(rd-req-sent-to-spec-mcu, 0x08),
drivers/perf/xgene_pmu.c
486
XGENE_PMU_EVENT_ATTR(rd-req-sent-to-spec-mcu-2, 0x09),
drivers/perf/xgene_pmu.c
487
XGENE_PMU_EVENT_ATTR(glbl-ack-recv-for-rd-sent-to-spec-mcu, 0x0a),
drivers/perf/xgene_pmu.c
488
XGENE_PMU_EVENT_ATTR(glbl-ack-go-recv-for-rd-sent-to-spec-mcu, 0x0b),
drivers/perf/xgene_pmu.c
489
XGENE_PMU_EVENT_ATTR(glbl-ack-nogo-recv-for-rd-sent-to-spec-mcu, 0x0c),
drivers/perf/xgene_pmu.c
490
XGENE_PMU_EVENT_ATTR(glbl-ack-go-recv-any-rd-req, 0x0d),
drivers/perf/xgene_pmu.c
491
XGENE_PMU_EVENT_ATTR(glbl-ack-go-recv-any-rd-req-2, 0x0e),
drivers/perf/xgene_pmu.c
494
XGENE_PMU_EVENT_ATTR(rd-gack-recv, 0x11),
drivers/perf/xgene_pmu.c
496
XGENE_PMU_EVENT_ATTR(cancel-rd-gack, 0x13),
drivers/perf/xgene_pmu.c
500
XGENE_PMU_EVENT_ATTR(mcb-mcu-rd-intf-stall, 0x17),
drivers/perf/xgene_pmu.c
501
XGENE_PMU_EVENT_ATTR(csw-rd-intf-blocked, 0x18),
drivers/perf/xgene_pmu.c
520
XGENE_PMU_EVENT_ATTR(rd-sent, 0x03),
drivers/perf/xgene_pmu.c
528
XGENE_PMU_EVENT_ATTR(rd-rda-sent, 0x0b),
drivers/perf/xgene_pmu.c
541
XGENE_PMU_EVENT_ATTR(rd-retry, 0x18),
drivers/perf/xgene_pmu.c
558
XGENE_PMU_EVENT_ATTR(rd-enq, 0x29),
drivers/power/reset/syscon-reboot.c
32
const struct reboot_data *rd; /* from of match data, if any */
drivers/power/reset/syscon-reboot.c
46
if (ctx->rd) {
drivers/power/reset/syscon-reboot.c
47
if (mode < ARRAY_SIZE(ctx->rd->mode_bits) &&
drivers/power/reset/syscon-reboot.c
48
ctx->rd->mode_bits[mode].valid)
drivers/power/reset/syscon-reboot.c
49
mode_bits = &ctx->rd->mode_bits[mode];
drivers/power/reset/syscon-reboot.c
51
mode_bits = &ctx->rd->catchall;
drivers/power/reset/syscon-reboot.c
87
ctx->rd = of_device_get_match_data(dev);
drivers/power/reset/syscon-reboot.c
88
if (!ctx->rd) {
drivers/powercap/intel_rapl_common.c
1059
static void set_floor_freq_default(struct rapl_domain *rd, bool mode)
drivers/powercap/intel_rapl_common.c
1066
rapl_write_pl_data(rd, POWER_LIMIT1, PL_CLAMP, mode);
drivers/powercap/intel_rapl_common.c
1069
rapl_write_pl_data(rd, i, PL_ENABLE, mode);
drivers/powercap/intel_rapl_common.c
1070
rapl_write_pl_data(rd, i, PL_CLAMP, mode);
drivers/powercap/intel_rapl_common.c
1074
static void set_floor_freq_atom(struct rapl_domain *rd, bool enable)
drivers/powercap/intel_rapl_common.c
1077
struct rapl_defaults *defaults = get_defaults(rd->rp);
drivers/powercap/intel_rapl_common.c
1098
static u64 rapl_compute_time_window_core(struct rapl_domain *rd, u64 value,
drivers/powercap/intel_rapl_common.c
1110
value = (1 << y) * (4 + f) * rd->time_unit / 4;
drivers/powercap/intel_rapl_common.c
1112
if (value < rd->time_unit)
drivers/powercap/intel_rapl_common.c
1115
do_div(value, rd->time_unit);
drivers/powercap/intel_rapl_common.c
1131
static u64 rapl_compute_time_window_atom(struct rapl_domain *rd, u64 value,
drivers/powercap/intel_rapl_common.c
1139
return (value) ? value * rd->time_unit : rd->time_unit;
drivers/powercap/intel_rapl_common.c
1141
value = div64_u64(value, rd->time_unit);
drivers/powercap/intel_rapl_common.c
1154
static int rapl_check_unit_tpmi(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
1159
ra.reg = rd->regs[RAPL_DOMAIN_REG_UNIT];
drivers/powercap/intel_rapl_common.c
1161
if (rd->rp->priv->read_raw(get_rid(rd->rp), &ra, false)) {
drivers/powercap/intel_rapl_common.c
1163
ra.reg.val, rd->rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
1168
rd->energy_unit = ENERGY_UNIT_SCALE * 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
1171
rd->power_unit = 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
1174
rd->time_unit = 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
1177
rd->rp->name, rd->name, rd->energy_unit, rd->time_unit, rd->power_unit);
drivers/powercap/intel_rapl_common.c
126
static bool is_pl_valid(struct rapl_domain *rd, int pl)
drivers/powercap/intel_rapl_common.c
130
return rd->rpl[pl].name ? true : false;
drivers/powercap/intel_rapl_common.c
133
static int get_pl_lock_prim(struct rapl_domain *rd, int pl)
drivers/powercap/intel_rapl_common.c
1342
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
135
if (rd->rp->priv->type == RAPL_IF_TPMI) {
drivers/powercap/intel_rapl_common.c
1350
for (rd = rp->domains; rd < rp->domains + rp->nr_domains; rd++) {
drivers/powercap/intel_rapl_common.c
1351
if (rd->id == RAPL_DOMAIN_PACKAGE) {
drivers/powercap/intel_rapl_common.c
1352
nr_pl = find_nr_power_limit(rd);
drivers/powercap/intel_rapl_common.c
1354
power_zone = powercap_register_zone(&rd->power_zone,
drivers/powercap/intel_rapl_common.c
1356
NULL, &zone_ops[rd->id], nr_pl,
drivers/powercap/intel_rapl_common.c
1374
for (rd = rp->domains; rd < rp->domains + rp->nr_domains; rd++) {
drivers/powercap/intel_rapl_common.c
1377
if (rd->id == RAPL_DOMAIN_PACKAGE)
drivers/powercap/intel_rapl_common.c
1379
if (rd->id == RAPL_DOMAIN_PLATFORM)
drivers/powercap/intel_rapl_common.c
1382
nr_pl = find_nr_power_limit(rd);
drivers/powercap/intel_rapl_common.c
1383
power_zone = powercap_register_zone(&rd->power_zone,
drivers/powercap/intel_rapl_common.c
1385
rd->name, parent,
drivers/powercap/intel_rapl_common.c
1386
&zone_ops[rd->id], nr_pl,
drivers/powercap/intel_rapl_common.c
1391
rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
1403
while (--rd >= rp->domains) {
drivers/powercap/intel_rapl_common.c
1404
pr_debug("unregister %s domain %s\n", rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
1406
&rd->power_zone);
drivers/powercap/intel_rapl_common.c
1444
static int rapl_get_domain_unit(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
1446
struct rapl_defaults *defaults = get_defaults(rd->rp);
drivers/powercap/intel_rapl_common.c
1449
if (!rd->regs[RAPL_DOMAIN_REG_UNIT].val) {
drivers/powercap/intel_rapl_common.c
1450
if (!rd->rp->priv->reg_unit.val) {
drivers/powercap/intel_rapl_common.c
1454
rd->regs[RAPL_DOMAIN_REG_UNIT] = rd->rp->priv->reg_unit;
drivers/powercap/intel_rapl_common.c
1462
ret = defaults->check_unit(rd);
drivers/powercap/intel_rapl_common.c
1466
if (rd->id == RAPL_DOMAIN_DRAM && defaults->dram_domain_energy_unit)
drivers/powercap/intel_rapl_common.c
1467
rd->energy_unit = defaults->dram_domain_energy_unit;
drivers/powercap/intel_rapl_common.c
1468
if (rd->id == RAPL_DOMAIN_PLATFORM && defaults->psys_domain_energy_unit)
drivers/powercap/intel_rapl_common.c
1469
rd->energy_unit = defaults->psys_domain_energy_unit;
drivers/powercap/intel_rapl_common.c
1482
static void rapl_detect_powerlimit(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
1488
if (!rapl_read_pl_data(rd, i, PL_LOCK, false, &val64)) {
drivers/powercap/intel_rapl_common.c
1490
rd->rpl[i].locked = true;
drivers/powercap/intel_rapl_common.c
1492
rd->rp->name, rd->name, pl_names[i]);
drivers/powercap/intel_rapl_common.c
1496
if (rapl_read_pl_data(rd, i, PL_LIMIT, false, &val64))
drivers/powercap/intel_rapl_common.c
1497
rd->rpl[i].name = NULL;
drivers/powercap/intel_rapl_common.c
1506
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
152
if (rd->rp->priv->limits[rd->id] & BIT(POWER_LIMIT2))
drivers/powercap/intel_rapl_common.c
1529
for (rd = rp->domains; rd < rp->domains + rp->nr_domains; rd++) {
drivers/powercap/intel_rapl_common.c
1530
rapl_get_domain_unit(rd);
drivers/powercap/intel_rapl_common.c
1531
rapl_detect_powerlimit(rd);
drivers/powercap/intel_rapl_common.c
157
static int get_pl_prim(struct rapl_domain *rd, int pl, enum pl_prims prim)
drivers/powercap/intel_rapl_common.c
163
if (prim == PL_CLAMP && rd->rp->priv->type != RAPL_IF_TPMI)
drivers/powercap/intel_rapl_common.c
172
return get_pl_lock_prim(rd, pl);
drivers/powercap/intel_rapl_common.c
177
if (prim == PL_CLAMP && rd->rp->priv->type != RAPL_IF_TPMI)
drivers/powercap/intel_rapl_common.c
186
return get_pl_lock_prim(rd, pl);
drivers/powercap/intel_rapl_common.c
197
return get_pl_lock_prim(rd, pl);
drivers/powercap/intel_rapl_common.c
2038
struct rapl_domain *rd = &rp->domains[idx];
drivers/powercap/intel_rapl_common.c
2039
int domain = rd->id;
drivers/powercap/intel_rapl_common.c
2049
val = rd->energy_unit * (1ULL << 32);
drivers/powercap/intel_rapl_common.c
2068
val *= rd->energy_unit;
drivers/powercap/intel_rapl_common.c
2075
pr_debug("Domain %s: hw unit %lld * 2^-32 Joules\n", rd->name, data->scale[domain]);
drivers/powercap/intel_rapl_common.c
209
int (*check_unit)(struct rapl_domain *rd);
drivers/powercap/intel_rapl_common.c
210
void (*set_floor_freq)(struct rapl_domain *rd, bool mode);
drivers/powercap/intel_rapl_common.c
211
u64 (*compute_time_window)(struct rapl_domain *rd, u64 val,
drivers/powercap/intel_rapl_common.c
2126
struct rapl_domain *rd, *rd_package = NULL;
drivers/powercap/intel_rapl_common.c
2130
for (rd = rp->domains; rd < rp->domains + rp->nr_domains; rd++) {
drivers/powercap/intel_rapl_common.c
2134
rapl_write_pl_data(rd, i, PL_ENABLE, 0);
drivers/powercap/intel_rapl_common.c
2135
rapl_write_pl_data(rd, i, PL_CLAMP, 0);
drivers/powercap/intel_rapl_common.c
2138
if (rd->id == RAPL_DOMAIN_PACKAGE) {
drivers/powercap/intel_rapl_common.c
2139
rd_package = rd;
drivers/powercap/intel_rapl_common.c
2143
rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
2145
&rd->power_zone);
drivers/powercap/intel_rapl_common.c
2276
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
2283
rd = power_zone_to_rapl_domain(rp->power_zone);
drivers/powercap/intel_rapl_common.c
2285
ret = rapl_read_pl_data(rd, i, PL_LIMIT, true,
drivers/powercap/intel_rapl_common.c
2286
&rd->rpl[i].last_power_limit);
drivers/powercap/intel_rapl_common.c
2288
rd->rpl[i].last_power_limit = 0;
drivers/powercap/intel_rapl_common.c
2297
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
2304
rd = power_zone_to_rapl_domain(rp->power_zone);
drivers/powercap/intel_rapl_common.c
2306
if (rd->rpl[i].last_power_limit)
drivers/powercap/intel_rapl_common.c
2307
rapl_write_pl_data(rd, i, PL_LIMIT,
drivers/powercap/intel_rapl_common.c
2308
rd->rpl[i].last_power_limit);
drivers/powercap/intel_rapl_common.c
254
static int rapl_read_data_raw(struct rapl_domain *rd,
drivers/powercap/intel_rapl_common.c
258
static int rapl_write_data_raw(struct rapl_domain *rd,
drivers/powercap/intel_rapl_common.c
261
static int rapl_read_pl_data(struct rapl_domain *rd, int pl,
drivers/powercap/intel_rapl_common.c
264
static int rapl_write_pl_data(struct rapl_domain *rd, int pl,
drivers/powercap/intel_rapl_common.c
267
static u64 rapl_unit_xlate(struct rapl_domain *rd,
drivers/powercap/intel_rapl_common.c
284
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
291
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
293
if (!rapl_read_data_raw(rd, ENERGY_COUNTER, true, &energy_now, false)) {
drivers/powercap/intel_rapl_common.c
306
struct rapl_domain *rd = power_zone_to_rapl_domain(pcd_dev);
drivers/powercap/intel_rapl_common.c
308
*energy = rapl_unit_xlate(rd, ENERGY_UNIT, ENERGY_STATUS_MASK, 0);
drivers/powercap/intel_rapl_common.c
314
struct rapl_domain *rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
315
struct rapl_package *rp = rd->rp;
drivers/powercap/intel_rapl_common.c
320
if (rd->id == RAPL_DOMAIN_PACKAGE) {
drivers/powercap/intel_rapl_common.c
321
kfree(rd);
drivers/powercap/intel_rapl_common.c
329
static int find_nr_power_limit(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
334
if (is_pl_valid(rd, i))
drivers/powercap/intel_rapl_common.c
343
struct rapl_domain *rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
344
struct rapl_defaults *defaults = get_defaults(rd->rp);
drivers/powercap/intel_rapl_common.c
349
ret = rapl_write_pl_data(rd, POWER_LIMIT1, PL_ENABLE, mode);
drivers/powercap/intel_rapl_common.c
353
ret = rapl_read_pl_data(rd, POWER_LIMIT1, PL_ENABLE, false, &val);
drivers/powercap/intel_rapl_common.c
364
defaults->set_floor_freq(rd, mode);
drivers/powercap/intel_rapl_common.c
374
struct rapl_domain *rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
378
if (rd->rpl[POWER_LIMIT1].locked) {
drivers/powercap/intel_rapl_common.c
383
ret = rapl_read_pl_data(rd, POWER_LIMIT1, PL_ENABLE, true, &val);
drivers/powercap/intel_rapl_common.c
440
static int contraint_to_pl(struct rapl_domain *rd, int cid)
drivers/powercap/intel_rapl_common.c
445
if (is_pl_valid(rd, i) && j++ == cid) {
drivers/powercap/intel_rapl_common.c
458
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
464
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
465
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
466
rp = rd->rp;
drivers/powercap/intel_rapl_common.c
468
ret = rapl_write_pl_data(rd, id, PL_LIMIT, power_limit);
drivers/powercap/intel_rapl_common.c
478
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
484
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
485
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
487
ret = rapl_read_pl_data(rd, id, PL_LIMIT, true, &val);
drivers/powercap/intel_rapl_common.c
499
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
504
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
505
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
507
ret = rapl_write_pl_data(rd, id, PL_TIME_WINDOW, window);
drivers/powercap/intel_rapl_common.c
516
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
522
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
523
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
525
ret = rapl_read_pl_data(rd, id, PL_TIME_WINDOW, true, &val);
drivers/powercap/intel_rapl_common.c
537
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
540
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
541
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
543
return rd->rpl[id].name;
drivers/powercap/intel_rapl_common.c
550
struct rapl_domain *rd;
drivers/powercap/intel_rapl_common.c
556
rd = power_zone_to_rapl_domain(power_zone);
drivers/powercap/intel_rapl_common.c
557
id = contraint_to_pl(rd, cid);
drivers/powercap/intel_rapl_common.c
559
ret = rapl_read_pl_data(rd, id, PL_MAX_POWER, true, &val);
drivers/powercap/intel_rapl_common.c
592
struct rapl_domain *rd = rp->domains;
drivers/powercap/intel_rapl_common.c
601
rd->rp = rp;
drivers/powercap/intel_rapl_common.c
604
snprintf(rd->name, RAPL_DOMAIN_NAME_LENGTH, "psys-%d",
drivers/powercap/intel_rapl_common.c
608
snprintf(rd->name, RAPL_DOMAIN_NAME_LENGTH, "%s",
drivers/powercap/intel_rapl_common.c
612
rd->id = i;
drivers/powercap/intel_rapl_common.c
619
rd->rpl[t].name = pl_names[t];
drivers/powercap/intel_rapl_common.c
623
rd->regs[j] = rp->priv->regs[i][j];
drivers/powercap/intel_rapl_common.c
625
rd++;
drivers/powercap/intel_rapl_common.c
629
static u64 rapl_unit_xlate(struct rapl_domain *rd, enum unit_type type,
drivers/powercap/intel_rapl_common.c
633
struct rapl_defaults *defaults = get_defaults(rd->rp);
drivers/powercap/intel_rapl_common.c
638
units = rd->power_unit;
drivers/powercap/intel_rapl_common.c
642
units = rd->energy_unit;
drivers/powercap/intel_rapl_common.c
645
return defaults->compute_time_window(rd, value, to_raw);
drivers/powercap/intel_rapl_common.c
792
prim_fixups(struct rapl_domain *rd, enum rapl_primitives prim)
drivers/powercap/intel_rapl_common.c
794
struct rapl_defaults *defaults = get_defaults(rd->rp);
drivers/powercap/intel_rapl_common.c
799
if (rd->id != RAPL_DOMAIN_PLATFORM)
drivers/powercap/intel_rapl_common.c
833
static int rapl_read_data_raw(struct rapl_domain *rd,
drivers/powercap/intel_rapl_common.c
838
enum rapl_primitives prim_fixed = prim_fixups(rd, prim);
drivers/powercap/intel_rapl_common.c
839
struct rapl_primitive_info *rpi = get_rpi(rd->rp, prim_fixed);
drivers/powercap/intel_rapl_common.c
845
ra.reg = rd->regs[rpi->id];
drivers/powercap/intel_rapl_common.c
851
*data = rd->rdd.primitives[prim];
drivers/powercap/intel_rapl_common.c
857
if (rd->rp->priv->read_raw(get_rid(rd->rp), &ra, pmu_ctx)) {
drivers/powercap/intel_rapl_common.c
858
pr_debug("failed to read reg 0x%llx for %s:%s\n", ra.reg.val, rd->rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
865
*data = rapl_unit_xlate(rd, rpi->unit, value, 0);
drivers/powercap/intel_rapl_common.c
873
static int rapl_write_data_raw(struct rapl_domain *rd,
drivers/powercap/intel_rapl_common.c
877
enum rapl_primitives prim_fixed = prim_fixups(rd, prim);
drivers/powercap/intel_rapl_common.c
878
struct rapl_primitive_info *rpi = get_rpi(rd->rp, prim_fixed);
drivers/powercap/intel_rapl_common.c
886
bits = rapl_unit_xlate(rd, rpi->unit, value, 1);
drivers/powercap/intel_rapl_common.c
892
ra.reg = rd->regs[rpi->id];
drivers/powercap/intel_rapl_common.c
896
ret = rd->rp->priv->write_raw(get_rid(rd->rp), &ra);
drivers/powercap/intel_rapl_common.c
901
static int rapl_read_pl_data(struct rapl_domain *rd, int pl,
drivers/powercap/intel_rapl_common.c
904
enum rapl_primitives prim = get_pl_prim(rd, pl, pl_prim);
drivers/powercap/intel_rapl_common.c
906
if (!is_pl_valid(rd, pl))
drivers/powercap/intel_rapl_common.c
909
return rapl_read_data_raw(rd, prim, xlate, data, false);
drivers/powercap/intel_rapl_common.c
912
static int rapl_write_pl_data(struct rapl_domain *rd, int pl,
drivers/powercap/intel_rapl_common.c
916
enum rapl_primitives prim = get_pl_prim(rd, pl, pl_prim);
drivers/powercap/intel_rapl_common.c
918
if (!is_pl_valid(rd, pl))
drivers/powercap/intel_rapl_common.c
921
if (rd->rpl[pl].locked) {
drivers/powercap/intel_rapl_common.c
922
pr_debug("%s:%s:%s locked by BIOS\n", rd->rp->name, rd->name, pl_names[pl]);
drivers/powercap/intel_rapl_common.c
926
return rapl_write_data_raw(rd, prim, value);
drivers/powercap/intel_rapl_common.c
939
static int rapl_check_unit_core(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
944
ra.reg = rd->regs[RAPL_DOMAIN_REG_UNIT];
drivers/powercap/intel_rapl_common.c
946
if (rd->rp->priv->read_raw(get_rid(rd->rp), &ra, false)) {
drivers/powercap/intel_rapl_common.c
948
ra.reg.val, rd->rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
953
rd->energy_unit = ENERGY_UNIT_SCALE * 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
956
rd->power_unit = 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
959
rd->time_unit = 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
962
rd->rp->name, rd->name, rd->energy_unit, rd->time_unit, rd->power_unit);
drivers/powercap/intel_rapl_common.c
967
static int rapl_check_unit_atom(struct rapl_domain *rd)
drivers/powercap/intel_rapl_common.c
972
ra.reg = rd->regs[RAPL_DOMAIN_REG_UNIT];
drivers/powercap/intel_rapl_common.c
974
if (rd->rp->priv->read_raw(get_rid(rd->rp), &ra, false)) {
drivers/powercap/intel_rapl_common.c
976
ra.reg.val, rd->rp->name, rd->name);
drivers/powercap/intel_rapl_common.c
981
rd->energy_unit = ENERGY_UNIT_SCALE * 1 << value;
drivers/powercap/intel_rapl_common.c
984
rd->power_unit = (1 << value) * 1000;
drivers/powercap/intel_rapl_common.c
987
rd->time_unit = 1000000 / (1 << value);
drivers/powercap/intel_rapl_common.c
990
rd->rp->name, rd->name, rd->energy_unit, rd->time_unit, rd->power_unit);
drivers/regulator/bd71828-regulator.c
1667
struct bd71828_regulator_data *rd;
drivers/regulator/bd71828-regulator.c
1669
rd = &rdata[i];
drivers/regulator/bd71828-regulator.c
1671
config.driver_data = rd;
drivers/regulator/bd71828-regulator.c
1673
&rd->desc, &config);
drivers/regulator/bd71828-regulator.c
1677
rd->desc.name);
drivers/regulator/bd71828-regulator.c
1679
for (j = 0; j < rd->reg_init_amnt; j++) {
drivers/regulator/bd71828-regulator.c
1681
rd->reg_inits[j].reg,
drivers/regulator/bd71828-regulator.c
1682
rd->reg_inits[j].mask,
drivers/regulator/bd71828-regulator.c
1683
rd->reg_inits[j].val);
drivers/regulator/bd71828-regulator.c
1687
rd->desc.name);
drivers/reset/reset-pistachio.c
104
struct pistachio_reset_data *rd;
drivers/reset/reset-pistachio.c
108
rd = devm_kzalloc(dev, sizeof(*rd), GFP_KERNEL);
drivers/reset/reset-pistachio.c
109
if (!rd)
drivers/reset/reset-pistachio.c
112
rd->periph_regs = syscon_node_to_regmap(np->parent);
drivers/reset/reset-pistachio.c
113
if (IS_ERR(rd->periph_regs))
drivers/reset/reset-pistachio.c
114
return PTR_ERR(rd->periph_regs);
drivers/reset/reset-pistachio.c
116
rd->rcdev.owner = THIS_MODULE;
drivers/reset/reset-pistachio.c
117
rd->rcdev.nr_resets = PISTACHIO_RESET_MAX + 1;
drivers/reset/reset-pistachio.c
118
rd->rcdev.ops = &pistachio_reset_ops;
drivers/reset/reset-pistachio.c
119
rd->rcdev.of_node = np;
drivers/reset/reset-pistachio.c
121
return devm_reset_controller_register(dev, &rd->rcdev);
drivers/reset/reset-pistachio.c
66
struct pistachio_reset_data *rd;
drivers/reset/reset-pistachio.c
70
rd = container_of(rcdev, struct pistachio_reset_data, rcdev);
drivers/reset/reset-pistachio.c
76
return regmap_update_bits(rd->periph_regs, PISTACHIO_SOFT_RESET,
drivers/reset/reset-pistachio.c
83
struct pistachio_reset_data *rd;
drivers/reset/reset-pistachio.c
87
rd = container_of(rcdev, struct pistachio_reset_data, rcdev);
drivers/reset/reset-pistachio.c
93
return regmap_update_bits(rd->periph_regs, PISTACHIO_SOFT_RESET,
drivers/s390/net/ctcm_fsms.c
1078
int rd = CHANNEL_DIRECTION(ch->flags);
drivers/s390/net/ctcm_fsms.c
1083
CTCM_FUNTAIL, ch->id, rd == CTCM_READ ? "RX" : "TX");
drivers/s390/net/ctcm_fsms.c
1089
if (rd == CTCM_READ) {
drivers/sbus/char/envctrl.c
312
unsigned char rd;
drivers/sbus/char/envctrl.c
317
rd = envctrl_i2c_read_data();
drivers/sbus/char/envctrl.c
319
return rd;
drivers/scsi/imm.c
45
unsigned rd:1; /* Read data in data phase */
drivers/scsi/imm.c
683
if (dev->rd == 0) {
drivers/scsi/imm.c
863
dev->rd = (x & 0x10) ? 1 : 0;
drivers/scsi/imm.c
866
if ((dev->dp) && (dev->rd))
drivers/scsi/imm.c
889
if ((dev->dp) && (dev->rd)) {
drivers/scsi/mpt3sas/mpt3sas_base.c
1700
union reply_descriptor rd;
drivers/scsi/mpt3sas/mpt3sas_base.c
1725
rd.word = le64_to_cpu(rpf->Words);
drivers/scsi/mpt3sas/mpt3sas_base.c
1726
if (rd.u.low == UINT_MAX || rd.u.high == UINT_MAX)
drivers/scsi/qla2xxx/qla_mbx.c
3895
qla2x00_disable_fce_trace(scsi_qla_host_t *vha, uint64_t *wr, uint64_t *rd)
drivers/scsi/qla2xxx/qla_mbx.c
3932
if (rd)
drivers/scsi/qla2xxx/qla_mbx.c
3933
*rd = (uint64_t) mcp->mb[9] << 48 |
drivers/scsi/raid_class.c
101
list_for_each_entry_safe(rc, next, &rd->component_list, node) {
drivers/scsi/raid_class.c
107
kfree(rd);
drivers/scsi/raid_class.c
179
struct raid_data *rd = dev_get_drvdata(dev); \
drivers/scsi/raid_class.c
188
name = raid_##states##_name(rd->attr); \
drivers/scsi/raid_class.c
194
raid_attr_show_internal(attr, %d, rd->attr, code) \
drivers/scsi/raid_class.c
80
struct raid_data *rd;
drivers/scsi/raid_class.c
84
rd = kzalloc_obj(*rd);
drivers/scsi/raid_class.c
85
if (!rd)
drivers/scsi/raid_class.c
88
INIT_LIST_HEAD(&rd->component_list);
drivers/scsi/raid_class.c
89
dev_set_drvdata(cdev, rd);
drivers/scsi/raid_class.c
97
struct raid_data *rd = dev_get_drvdata(cdev);
drivers/soc/fsl/dpio/dpio-service.c
592
struct qbman_release_desc rd;
drivers/soc/fsl/dpio/dpio-service.c
598
qbman_release_desc_clear(&rd);
drivers/soc/fsl/dpio/dpio-service.c
599
qbman_release_desc_set_bpid(&rd, bpid);
drivers/soc/fsl/dpio/dpio-service.c
601
return qbman_swp_release(d->swp, &rd, buffers, num_buffers);
drivers/spi/spi-atmel.c
1062
u16 rd; /* RD field is the lowest 16 bits of RDR */
drivers/spi/spi-atmel.c
1080
rd = spi_readl(as, RDR);
drivers/spi/spi-atmel.c
1082
*words++ = rd;
drivers/spi/spi-atmel.c
1084
*bytes++ = rd;
drivers/spi/spi-bcm-qspi.c
585
u32 rd = 0;
drivers/spi/spi-bcm-qspi.c
589
rd = bcm_qspi_read(qspi, CHIP_SELECT, 0);
drivers/spi/spi-bcm-qspi.c
590
wr = (rd & ~0xff) | (1 << cs);
drivers/spi/spi-bcm-qspi.c
591
if (rd == wr)
drivers/spi/spi.c
4927
struct of_reconfig_data *rd = arg;
drivers/spi/spi.c
4933
ctlr = of_find_spi_controller_by_node(rd->dn->parent);
drivers/spi/spi.c
4937
if (of_node_test_and_set_flag(rd->dn, OF_POPULATED)) {
drivers/spi/spi.c
4946
rd->dn->fwnode.flags &= ~FWNODE_FLAG_NOT_DEVICE;
drivers/spi/spi.c
4947
spi = of_register_spi_device(ctlr, rd->dn);
drivers/spi/spi.c
4952
__func__, rd->dn);
drivers/spi/spi.c
4953
of_node_clear_flag(rd->dn, OF_POPULATED);
drivers/spi/spi.c
4960
if (!of_node_check_flag(rd->dn, OF_POPULATED))
drivers/spi/spi.c
4964
spi = of_find_spi_device_by_node(rd->dn);
drivers/staging/fbtft/fbtft-core.c
101
ret = fbtft_request_one_gpio(par, "rd", 0, &par->gpio.rd);
drivers/staging/fbtft/fbtft.h
212
struct gpio_desc *rd;
drivers/video/fbdev/omap/lcd_mipid.c
343
} *rd, rd_ctrl[7] = {
drivers/video/fbdev/omap/lcd_mipid.c
353
rd = rd_ctrl;
drivers/video/fbdev/omap/lcd_mipid.c
354
for (i = 0; i < 3; i++, rd++)
drivers/video/fbdev/omap/lcd_mipid.c
355
mipid_write(md, rd->cmd, (u8 *)rd->wbuf, rd->wlen);
drivers/video/fbdev/omap/lcd_mipid.c
358
mipid_read(md, rd->cmd, rbuf, 2);
drivers/video/fbdev/omap/lcd_mipid.c
359
rd++;
drivers/video/fbdev/omap/lcd_mipid.c
361
for (i = 0; i < 3; i++, rd++) {
drivers/video/fbdev/omap/lcd_mipid.c
363
mipid_write(md, rd->cmd, (u8 *)rd->wbuf, rd->wlen);
fs/cachefiles/namei.c
385
struct renamedata rd = {
fs/cachefiles/namei.c
395
ret = vfs_rename(&rd);
fs/debugfs/inode.c
827
struct renamedata rd = {};
fs/debugfs/inode.c
840
rd.old_parent = dget_parent(dentry);
fs/debugfs/inode.c
841
rd.new_parent = rd.old_parent;
fs/debugfs/inode.c
842
rd.flags = RENAME_NOREPLACE;
fs/debugfs/inode.c
843
target = lookup_noperm_unlocked(&QSTR(new_name), rd.new_parent);
fs/debugfs/inode.c
849
error = start_renaming_two_dentries(&rd, dentry, target);
fs/debugfs/inode.c
857
dir = d_inode(rd.old_parent);
fs/debugfs/inode.c
859
simple_rename_timestamp(dir, dentry, dir, rd.new_dentry);
fs/debugfs/inode.c
860
d_move(dentry, rd.new_dentry);
fs/debugfs/inode.c
863
end_renaming(&rd);
fs/debugfs/inode.c
865
dput(rd.old_parent);
fs/dlm/midcomms.c
1471
struct dlm_rawmsg_data *rd = data;
fs/dlm/midcomms.c
1472
struct dlm_header *h = rd->buf;
fs/dlm/midcomms.c
1481
h->u.h_seq = cpu_to_le32(atomic_fetch_inc(&rd->node->seq_send));
fs/dlm/midcomms.c
1493
struct dlm_rawmsg_data rd;
fs/dlm/midcomms.c
1497
rd.node = node;
fs/dlm/midcomms.c
1498
rd.buf = buf;
fs/dlm/midcomms.c
1501
midcomms_new_rawmsg_cb, &rd);
fs/ecryptfs/inode.c
605
struct renamedata rd = {};
fs/ecryptfs/inode.c
618
rd.mnt_idmap = &nop_mnt_idmap;
fs/ecryptfs/inode.c
619
rd.old_parent = lower_old_dir_dentry;
fs/ecryptfs/inode.c
620
rd.new_parent = lower_new_dir_dentry;
fs/ecryptfs/inode.c
621
rc = start_renaming_two_dentries(&rd, lower_old_dentry, lower_new_dentry);
fs/ecryptfs/inode.c
625
rc = vfs_rename(&rd);
fs/ecryptfs/inode.c
635
end_renaming(&rd);
fs/hfs/catalog.c
343
struct hfs_readdir_data *rd;
fs/hfs/catalog.c
371
list_for_each_entry(rd, &HFS_I(dir)->open_dir_list, list) {
fs/hfs/catalog.c
372
if (fd.tree->keycmp(fd.search_key, (void *)&rd->key) < 0)
fs/hfs/catalog.c
373
rd->file->f_pos--;
fs/hfs/dir.c
149
rd = file->private_data;
fs/hfs/dir.c
150
if (!rd) {
fs/hfs/dir.c
151
rd = kmalloc_obj(struct hfs_readdir_data);
fs/hfs/dir.c
152
if (!rd) {
fs/hfs/dir.c
156
file->private_data = rd;
fs/hfs/dir.c
157
rd->file = file;
fs/hfs/dir.c
159
list_add(&rd->list, &HFS_I(inode)->open_dir_list);
fs/hfs/dir.c
166
memcpy(&rd->key, &fd.key->cat, sizeof(struct hfs_cat_key));
fs/hfs/dir.c
174
struct hfs_readdir_data *rd = file->private_data;
fs/hfs/dir.c
175
if (rd) {
fs/hfs/dir.c
177
list_del(&rd->list);
fs/hfs/dir.c
179
kfree(rd);
fs/hfs/dir.c
56
struct hfs_readdir_data *rd;
fs/hfsplus/catalog.c
397
struct hfsplus_readdir_data *rd =
fs/hfsplus/catalog.c
399
if (fd.tree->keycmp(fd.search_key, (void *)&rd->key) < 0)
fs/hfsplus/catalog.c
400
rd->file->f_pos--;
fs/hfsplus/dir.c
138
struct hfsplus_readdir_data *rd;
fs/hfsplus/dir.c
264
rd = file->private_data;
fs/hfsplus/dir.c
265
if (!rd) {
fs/hfsplus/dir.c
266
rd = kmalloc_obj(struct hfsplus_readdir_data);
fs/hfsplus/dir.c
267
if (!rd) {
fs/hfsplus/dir.c
271
file->private_data = rd;
fs/hfsplus/dir.c
272
rd->file = file;
fs/hfsplus/dir.c
274
list_add(&rd->list, &HFSPLUS_I(inode)->open_dir_list);
fs/hfsplus/dir.c
281
memcpy(&rd->key, fd.key, sizeof(struct hfsplus_cat_key));
fs/hfsplus/dir.c
290
struct hfsplus_readdir_data *rd = file->private_data;
fs/hfsplus/dir.c
291
if (rd) {
fs/hfsplus/dir.c
293
list_del(&rd->list);
fs/hfsplus/dir.c
295
kfree(rd);
fs/hpfs/dnode.c
242
struct dnode *d, *ad, *rd, *nd = NULL;
fs/hpfs/dnode.c
341
if (!(rd = hpfs_alloc_dnode(i->i_sb, le32_to_cpu(d->up), &rdno, &qbh2))) {
fs/hpfs/dnode.c
351
rd->root_dnode = 1;
fs/hpfs/dnode.c
352
rd->up = d->up;
fs/hpfs/dnode.c
373
set_last_pointer(i->i_sb, rd, dno);
fs/hpfs/dnode.c
375
d = rd;
fs/jffs2/dir.c
294
struct jffs2_raw_dirent *rd;
fs/jffs2/dir.c
388
ret = jffs2_reserve_space(c, sizeof(*rd)+namelen, &alloclen,
fs/jffs2/dir.c
393
rd = jffs2_alloc_raw_dirent();
fs/jffs2/dir.c
394
if (!rd) {
fs/jffs2/dir.c
404
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/dir.c
405
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/dir.c
406
rd->totlen = cpu_to_je32(sizeof(*rd) + namelen);
fs/jffs2/dir.c
407
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/dir.c
409
rd->pino = cpu_to_je32(dir_i->i_ino);
fs/jffs2/dir.c
410
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/dir.c
411
rd->ino = cpu_to_je32(inode->i_ino);
fs/jffs2/dir.c
412
rd->mctime = cpu_to_je32(JFFS2_NOW());
fs/jffs2/dir.c
413
rd->nsize = namelen;
fs/jffs2/dir.c
414
rd->type = DT_LNK;
fs/jffs2/dir.c
415
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/dir.c
416
rd->name_crc = cpu_to_je32(crc32(0, dentry->d_name.name, namelen));
fs/jffs2/dir.c
418
fd = jffs2_write_dirent(c, dir_f, rd, dentry->d_name.name, namelen, ALLOC_NORMAL);
fs/jffs2/dir.c
424
jffs2_free_raw_dirent(rd);
fs/jffs2/dir.c
431
inode_set_ctime_to_ts(dir_i, ITIME(je32_to_cpu(rd->mctime))));
fs/jffs2/dir.c
433
jffs2_free_raw_dirent(rd);
fs/jffs2/dir.c
458
struct jffs2_raw_dirent *rd;
fs/jffs2/dir.c
533
ret = jffs2_reserve_space(c, sizeof(*rd)+namelen, &alloclen,
fs/jffs2/dir.c
538
rd = jffs2_alloc_raw_dirent();
fs/jffs2/dir.c
539
if (!rd) {
fs/jffs2/dir.c
549
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/dir.c
550
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/dir.c
551
rd->totlen = cpu_to_je32(sizeof(*rd) + namelen);
fs/jffs2/dir.c
552
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/dir.c
554
rd->pino = cpu_to_je32(dir_i->i_ino);
fs/jffs2/dir.c
555
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/dir.c
556
rd->ino = cpu_to_je32(inode->i_ino);
fs/jffs2/dir.c
557
rd->mctime = cpu_to_je32(JFFS2_NOW());
fs/jffs2/dir.c
558
rd->nsize = namelen;
fs/jffs2/dir.c
559
rd->type = DT_DIR;
fs/jffs2/dir.c
560
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/dir.c
561
rd->name_crc = cpu_to_je32(crc32(0, dentry->d_name.name, namelen));
fs/jffs2/dir.c
563
fd = jffs2_write_dirent(c, dir_f, rd, dentry->d_name.name, namelen, ALLOC_NORMAL);
fs/jffs2/dir.c
569
jffs2_free_raw_dirent(rd);
fs/jffs2/dir.c
576
inode_set_ctime_to_ts(dir_i, ITIME(je32_to_cpu(rd->mctime))));
fs/jffs2/dir.c
579
jffs2_free_raw_dirent(rd);
fs/jffs2/dir.c
632
struct jffs2_raw_dirent *rd;
fs/jffs2/dir.c
709
ret = jffs2_reserve_space(c, sizeof(*rd)+namelen, &alloclen,
fs/jffs2/dir.c
714
rd = jffs2_alloc_raw_dirent();
fs/jffs2/dir.c
715
if (!rd) {
fs/jffs2/dir.c
725
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/dir.c
726
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/dir.c
727
rd->totlen = cpu_to_je32(sizeof(*rd) + namelen);
fs/jffs2/dir.c
728
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/dir.c
730
rd->pino = cpu_to_je32(dir_i->i_ino);
fs/jffs2/dir.c
731
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/dir.c
732
rd->ino = cpu_to_je32(inode->i_ino);
fs/jffs2/dir.c
733
rd->mctime = cpu_to_je32(JFFS2_NOW());
fs/jffs2/dir.c
734
rd->nsize = namelen;
fs/jffs2/dir.c
737
rd->type = (mode & S_IFMT) >> 12;
fs/jffs2/dir.c
739
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/dir.c
740
rd->name_crc = cpu_to_je32(crc32(0, dentry->d_name.name, namelen));
fs/jffs2/dir.c
742
fd = jffs2_write_dirent(c, dir_f, rd, dentry->d_name.name, namelen, ALLOC_NORMAL);
fs/jffs2/dir.c
748
jffs2_free_raw_dirent(rd);
fs/jffs2/dir.c
755
inode_set_ctime_to_ts(dir_i, ITIME(je32_to_cpu(rd->mctime))));
fs/jffs2/dir.c
757
jffs2_free_raw_dirent(rd);
fs/jffs2/gc.c
854
struct jffs2_raw_dirent rd;
fs/jffs2/gc.c
858
rd.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/gc.c
859
rd.nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/gc.c
860
rd.nsize = strlen(fd->name);
fs/jffs2/gc.c
861
rd.totlen = cpu_to_je32(sizeof(rd) + rd.nsize);
fs/jffs2/gc.c
862
rd.hdr_crc = cpu_to_je32(crc32(0, &rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/gc.c
864
rd.pino = cpu_to_je32(f->inocache->ino);
fs/jffs2/gc.c
865
rd.version = cpu_to_je32(++f->highest_version);
fs/jffs2/gc.c
866
rd.ino = cpu_to_je32(fd->ino);
fs/jffs2/gc.c
870
rd.mctime = cpu_to_je32(JFFS2_F_I_MTIME(f));
fs/jffs2/gc.c
872
rd.mctime = cpu_to_je32(0);
fs/jffs2/gc.c
873
rd.type = fd->type;
fs/jffs2/gc.c
874
rd.node_crc = cpu_to_je32(crc32(0, &rd, sizeof(rd)-8));
fs/jffs2/gc.c
875
rd.name_crc = cpu_to_je32(crc32(0, fd->name, rd.nsize));
fs/jffs2/gc.c
877
ret = jffs2_reserve_space_gc(c, sizeof(rd)+rd.nsize, &alloclen,
fs/jffs2/gc.c
878
JFFS2_SUMMARY_DIRENT_SIZE(rd.nsize));
fs/jffs2/gc.c
881
sizeof(rd)+rd.nsize, ret);
fs/jffs2/gc.c
884
new_fd = jffs2_write_dirent(c, f, &rd, fd->name, rd.nsize, ALLOC_GC);
fs/jffs2/gc.c
907
struct jffs2_raw_dirent *rd;
fs/jffs2/gc.c
915
rd = kmalloc(rawlen, GFP_KERNEL);
fs/jffs2/gc.c
916
if (!rd)
fs/jffs2/gc.c
946
ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)rd);
fs/jffs2/gc.c
960
if (je16_to_cpu(rd->nodetype) != JFFS2_NODETYPE_DIRENT)
fs/jffs2/gc.c
964
if (je32_to_cpu(rd->name_crc) != name_crc)
fs/jffs2/gc.c
968
if (rd->nsize != name_len || !je32_to_cpu(rd->ino))
fs/jffs2/gc.c
972
if (memcmp(rd->name, fd->name, name_len))
fs/jffs2/gc.c
982
ref_offset(raw), je32_to_cpu(rd->ino));
fs/jffs2/gc.c
983
kfree(rd);
fs/jffs2/gc.c
989
kfree(rd);
fs/jffs2/nodelist.h
144
#define dirent_node_state(rd) ( (je32_to_cpu((rd)->ino)?REF_PRISTINE:REF_NORMAL) )
fs/jffs2/nodelist.h
402
struct jffs2_raw_dirent *rd, const unsigned char *name,
fs/jffs2/readinode.c
587
struct jffs2_raw_dirent *rd, size_t read,
fs/jffs2/readinode.c
596
crc = crc32(0, rd, sizeof(*rd) - 8);
fs/jffs2/readinode.c
597
if (unlikely(crc != je32_to_cpu(rd->node_crc))) {
fs/jffs2/readinode.c
599
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
610
if (unlikely(PAD((rd->nsize + sizeof(*rd))) != PAD(je32_to_cpu(rd->totlen)))) {
fs/jffs2/readinode.c
612
ref_offset(ref), rd->nsize, je32_to_cpu(rd->totlen));
fs/jffs2/readinode.c
625
ref->flash_offset = ref_offset(ref) | dirent_node_state(rd);
fs/jffs2/readinode.c
629
fd = jffs2_alloc_full_dirent(rd->nsize + 1);
fs/jffs2/readinode.c
634
fd->version = je32_to_cpu(rd->version);
fs/jffs2/readinode.c
635
fd->ino = je32_to_cpu(rd->ino);
fs/jffs2/readinode.c
636
fd->type = rd->type;
fs/jffs2/readinode.c
642
if(fd->version > rii->mctime_ver && je32_to_cpu(rd->mctime)) {
fs/jffs2/readinode.c
644
rii->latest_mctime = je32_to_cpu(rd->mctime);
fs/jffs2/readinode.c
651
if (read > sizeof(*rd))
fs/jffs2/readinode.c
652
memcpy(&fd->name[0], &rd->name[0],
fs/jffs2/readinode.c
653
min_t(uint32_t, rd->nsize, (read - sizeof(*rd)) ));
fs/jffs2/readinode.c
656
if (rd->nsize + sizeof(*rd) > read) {
fs/jffs2/readinode.c
659
int already = read - sizeof(*rd);
fs/jffs2/readinode.c
662
rd->nsize - already, &read, &fd->name[already]);
fs/jffs2/readinode.c
663
if (unlikely(read != rd->nsize - already) && likely(!err)) {
fs/jffs2/readinode.c
666
rd->nsize - already, read);
fs/jffs2/readinode.c
681
crc = crc32(0, fd->name, rd->nsize);
fs/jffs2/readinode.c
682
if (unlikely(crc != je32_to_cpu(rd->name_crc))) {
fs/jffs2/readinode.c
685
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
693
fd->nhash = full_name_hash(NULL, fd->name, rd->nsize);
fs/jffs2/readinode.c
695
fd->name[rd->nsize] = '\0';
fs/jffs2/readinode.c
714
struct jffs2_raw_inode *rd, int rdlen,
fs/jffs2/readinode.c
725
crc = crc32(0, rd, sizeof(*rd) - 8);
fs/jffs2/readinode.c
726
if (unlikely(crc != je32_to_cpu(rd->node_crc))) {
fs/jffs2/readinode.c
728
ref_offset(ref), je32_to_cpu(rd->node_crc), crc);
fs/jffs2/readinode.c
740
csize = je32_to_cpu(rd->csize);
fs/jffs2/readinode.c
746
if (unlikely(je32_to_cpu(rd->offset) > je32_to_cpu(rd->isize)) ||
fs/jffs2/readinode.c
747
unlikely(PAD(je32_to_cpu(rd->csize) + sizeof(*rd)) != PAD(je32_to_cpu(rd->totlen)))) {
fs/jffs2/readinode.c
793
buf = (unsigned char *)rd + sizeof(*rd);
fs/jffs2/readinode.c
795
len = min_t(uint32_t, rdlen - sizeof(*rd), csize);
fs/jffs2/readinode.c
802
if (len >= csize && unlikely(tn->partial_crc != je32_to_cpu(rd->data_crc))) {
fs/jffs2/readinode.c
804
ref_offset(ref), tn->partial_crc, je32_to_cpu(rd->data_crc));
fs/jffs2/readinode.c
839
tn->version = je32_to_cpu(rd->version);
fs/jffs2/readinode.c
840
tn->fn->ofs = je32_to_cpu(rd->offset);
fs/jffs2/readinode.c
841
tn->data_crc = je32_to_cpu(rd->data_crc);
fs/jffs2/readinode.c
851
if (rd->compr == JFFS2_COMPR_ZERO && !je32_to_cpu(rd->dsize) && csize)
fs/jffs2/readinode.c
854
tn->fn->size = je32_to_cpu(rd->dsize);
fs/jffs2/readinode.c
857
ref_offset(ref), je32_to_cpu(rd->version),
fs/jffs2/readinode.c
858
je32_to_cpu(rd->offset), je32_to_cpu(rd->dsize), csize);
fs/jffs2/readinode.c
869
dbg_readinode2("After adding ver %d:\n", je32_to_cpu(rd->version));
fs/jffs2/scan.c
1048
struct jffs2_raw_dirent *rd, uint32_t ofs, struct jffs2_summary *s)
fs/jffs2/scan.c
1060
crc = crc32(0, rd, sizeof(*rd)-8);
fs/jffs2/scan.c
1062
if (crc != je32_to_cpu(rd->node_crc)) {
fs/jffs2/scan.c
1064
__func__, ofs, je32_to_cpu(rd->node_crc), crc);
fs/jffs2/scan.c
1066
if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(rd->totlen)))))
fs/jffs2/scan.c
1071
pseudo_random += je32_to_cpu(rd->version);
fs/jffs2/scan.c
1074
checkedlen = strnlen(rd->name, rd->nsize);
fs/jffs2/scan.c
1075
if (checkedlen < rd->nsize) {
fs/jffs2/scan.c
1083
memcpy(&fd->name, rd->name, checkedlen);
fs/jffs2/scan.c
1087
if (crc != je32_to_cpu(rd->name_crc)) {
fs/jffs2/scan.c
1089
__func__, ofs, je32_to_cpu(rd->name_crc), crc);
fs/jffs2/scan.c
1091
fd->name, je32_to_cpu(rd->ino));
fs/jffs2/scan.c
1095
if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(rd->totlen)))))
fs/jffs2/scan.c
1099
ic = jffs2_scan_make_ino_cache(c, je32_to_cpu(rd->pino));
fs/jffs2/scan.c
1105
fd->raw = jffs2_link_node_ref(c, jeb, ofs | dirent_node_state(rd),
fs/jffs2/scan.c
1106
PAD(je32_to_cpu(rd->totlen)), ic);
fs/jffs2/scan.c
1109
fd->version = je32_to_cpu(rd->version);
fs/jffs2/scan.c
1110
fd->ino = je32_to_cpu(rd->ino);
fs/jffs2/scan.c
1112
fd->type = rd->type;
fs/jffs2/scan.c
1116
jffs2_sum_add_dirent_mem(s, rd, ofs - jeb->offset);
fs/jffs2/scan.c
49
struct jffs2_raw_dirent *rd, uint32_t ofs, struct jffs2_summary *s);
fs/jffs2/summary.c
133
int jffs2_sum_add_dirent_mem(struct jffs2_summary *s, struct jffs2_raw_dirent *rd,
fs/jffs2/summary.c
137
kmalloc(sizeof(struct jffs2_sum_dirent_mem) + rd->nsize, GFP_KERNEL);
fs/jffs2/summary.c
142
temp->nodetype = rd->nodetype;
fs/jffs2/summary.c
143
temp->totlen = rd->totlen;
fs/jffs2/summary.c
145
temp->pino = rd->pino;
fs/jffs2/summary.c
146
temp->version = rd->version;
fs/jffs2/summary.c
147
temp->ino = rd->ino;
fs/jffs2/summary.c
148
temp->nsize = rd->nsize;
fs/jffs2/summary.c
149
temp->type = rd->type;
fs/jffs2/summary.c
152
memcpy(temp->name, rd->name, rd->nsize);
fs/jffs2/summary.h
186
int jffs2_sum_add_dirent_mem(struct jffs2_summary *s, struct jffs2_raw_dirent *rd, uint32_t ofs);
fs/jffs2/write.c
206
struct jffs2_raw_dirent *rd, const unsigned char *name,
fs/jffs2/write.c
218
je32_to_cpu(rd->pino), name, name, je32_to_cpu(rd->ino),
fs/jffs2/write.c
219
je32_to_cpu(rd->name_crc));
fs/jffs2/write.c
221
D1(if(je32_to_cpu(rd->hdr_crc) != crc32(0, rd, sizeof(struct jffs2_unknown_node)-4)) {
fs/jffs2/write.c
231
je32_to_cpu(rd->pino), name, name, je32_to_cpu(rd->ino),
fs/jffs2/write.c
232
je32_to_cpu(rd->name_crc));
fs/jffs2/write.c
237
vecs[0].iov_base = rd;
fs/jffs2/write.c
238
vecs[0].iov_len = sizeof(*rd);
fs/jffs2/write.c
246
fd->version = je32_to_cpu(rd->version);
fs/jffs2/write.c
247
fd->ino = je32_to_cpu(rd->ino);
fs/jffs2/write.c
249
fd->type = rd->type;
fs/jffs2/write.c
258
if ((alloc_mode!=ALLOC_GC) && (je32_to_cpu(rd->version) < f->highest_version)) {
fs/jffs2/write.c
262
je32_to_cpu(rd->version), f->highest_version);
fs/jffs2/write.c
263
rd->version = cpu_to_je32(++f->highest_version);
fs/jffs2/write.c
264
fd->version = je32_to_cpu(rd->version);
fs/jffs2/write.c
265
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/write.c
269
(alloc_mode==ALLOC_GC)?0:je32_to_cpu(rd->pino));
fs/jffs2/write.c
270
if (ret || (retlen != sizeof(*rd) + namelen)) {
fs/jffs2/write.c
272
sizeof(*rd) + namelen, flash_ofs, ret, retlen);
fs/jffs2/write.c
275
jffs2_add_physical_node_ref(c, flash_ofs | REF_OBSOLETE, PAD(sizeof(*rd)+namelen), NULL);
fs/jffs2/write.c
293
ret = jffs2_reserve_space_gc(c, sizeof(*rd) + namelen, &dummy,
fs/jffs2/write.c
300
ret = jffs2_reserve_space(c, sizeof(*rd) + namelen, &dummy,
fs/jffs2/write.c
321
fd->raw = jffs2_add_physical_node_ref(c, flash_ofs | dirent_node_state(rd),
fs/jffs2/write.c
322
PAD(sizeof(*rd)+namelen), f->inocache);
fs/jffs2/write.c
445
struct jffs2_raw_dirent *rd;
fs/jffs2/write.c
492
ret = jffs2_reserve_space(c, sizeof(*rd)+qstr->len, &alloclen,
fs/jffs2/write.c
501
rd = jffs2_alloc_raw_dirent();
fs/jffs2/write.c
502
if (!rd) {
fs/jffs2/write.c
510
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/write.c
511
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/write.c
512
rd->totlen = cpu_to_je32(sizeof(*rd) + qstr->len);
fs/jffs2/write.c
513
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/write.c
515
rd->pino = cpu_to_je32(dir_f->inocache->ino);
fs/jffs2/write.c
516
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/write.c
517
rd->ino = ri->ino;
fs/jffs2/write.c
518
rd->mctime = ri->ctime;
fs/jffs2/write.c
519
rd->nsize = qstr->len;
fs/jffs2/write.c
520
rd->type = DT_REG;
fs/jffs2/write.c
521
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/write.c
522
rd->name_crc = cpu_to_je32(crc32(0, qstr->name, qstr->len));
fs/jffs2/write.c
524
fd = jffs2_write_dirent(c, dir_f, rd, qstr->name, qstr->len, ALLOC_NORMAL);
fs/jffs2/write.c
526
jffs2_free_raw_dirent(rd);
fs/jffs2/write.c
551
struct jffs2_raw_dirent *rd;
fs/jffs2/write.c
559
rd = jffs2_alloc_raw_dirent();
fs/jffs2/write.c
560
if (!rd)
fs/jffs2/write.c
563
ret = jffs2_reserve_space(c, sizeof(*rd)+namelen, &alloclen,
fs/jffs2/write.c
566
jffs2_free_raw_dirent(rd);
fs/jffs2/write.c
573
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/write.c
574
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/write.c
575
rd->totlen = cpu_to_je32(sizeof(*rd) + namelen);
fs/jffs2/write.c
576
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/write.c
578
rd->pino = cpu_to_je32(dir_f->inocache->ino);
fs/jffs2/write.c
579
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/write.c
580
rd->ino = cpu_to_je32(0);
fs/jffs2/write.c
581
rd->mctime = cpu_to_je32(time);
fs/jffs2/write.c
582
rd->nsize = namelen;
fs/jffs2/write.c
583
rd->type = DT_UNKNOWN;
fs/jffs2/write.c
584
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/write.c
585
rd->name_crc = cpu_to_je32(crc32(0, name, namelen));
fs/jffs2/write.c
587
fd = jffs2_write_dirent(c, dir_f, rd, name, namelen, ALLOC_DELETION);
fs/jffs2/write.c
589
jffs2_free_raw_dirent(rd);
fs/jffs2/write.c
671
struct jffs2_raw_dirent *rd;
fs/jffs2/write.c
676
rd = jffs2_alloc_raw_dirent();
fs/jffs2/write.c
677
if (!rd)
fs/jffs2/write.c
680
ret = jffs2_reserve_space(c, sizeof(*rd)+namelen, &alloclen,
fs/jffs2/write.c
683
jffs2_free_raw_dirent(rd);
fs/jffs2/write.c
690
rd->magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/write.c
691
rd->nodetype = cpu_to_je16(JFFS2_NODETYPE_DIRENT);
fs/jffs2/write.c
692
rd->totlen = cpu_to_je32(sizeof(*rd) + namelen);
fs/jffs2/write.c
693
rd->hdr_crc = cpu_to_je32(crc32(0, rd, sizeof(struct jffs2_unknown_node)-4));
fs/jffs2/write.c
695
rd->pino = cpu_to_je32(dir_f->inocache->ino);
fs/jffs2/write.c
696
rd->version = cpu_to_je32(++dir_f->highest_version);
fs/jffs2/write.c
697
rd->ino = cpu_to_je32(ino);
fs/jffs2/write.c
698
rd->mctime = cpu_to_je32(time);
fs/jffs2/write.c
699
rd->nsize = namelen;
fs/jffs2/write.c
701
rd->type = type;
fs/jffs2/write.c
703
rd->node_crc = cpu_to_je32(crc32(0, rd, sizeof(*rd)-8));
fs/jffs2/write.c
704
rd->name_crc = cpu_to_je32(crc32(0, name, namelen));
fs/jffs2/write.c
706
fd = jffs2_write_dirent(c, dir_f, rd, name, namelen, ALLOC_NORMAL);
fs/jffs2/write.c
708
jffs2_free_raw_dirent(rd);
fs/namei.c
3845
__start_renaming(struct renamedata *rd, int lookup_flags,
fs/namei.c
3853
if (rd->flags & RENAME_EXCHANGE)
fs/namei.c
3855
if (rd->flags & RENAME_NOREPLACE)
fs/namei.c
3858
trap = lock_rename(rd->old_parent, rd->new_parent);
fs/namei.c
3862
d1 = lookup_one_qstr_excl(old_last, rd->old_parent,
fs/namei.c
3868
d2 = lookup_one_qstr_excl(new_last, rd->new_parent,
fs/namei.c
3882
if (rd->flags & RENAME_EXCHANGE)
fs/namei.c
3889
rd->old_dentry = d1;
fs/namei.c
3890
rd->new_dentry = d2;
fs/namei.c
3891
dget(rd->old_parent);
fs/namei.c
3899
unlock_rename(rd->old_parent, rd->new_parent);
fs/namei.c
3924
int start_renaming(struct renamedata *rd, int lookup_flags,
fs/namei.c
3929
err = lookup_one_common(rd->mnt_idmap, old_last, rd->old_parent);
fs/namei.c
3932
err = lookup_one_common(rd->mnt_idmap, new_last, rd->new_parent);
fs/namei.c
3935
return __start_renaming(rd, lookup_flags, old_last, new_last);
fs/namei.c
3940
__start_renaming_dentry(struct renamedata *rd, int lookup_flags,
fs/namei.c
3948
if (rd->flags & RENAME_EXCHANGE)
fs/namei.c
3950
if (rd->flags & RENAME_NOREPLACE)
fs/namei.c
3954
trap = lock_rename_child(old_dentry, rd->new_parent);
fs/namei.c
3958
(rd->old_parent && rd->old_parent != old_dentry->d_parent)) {
fs/namei.c
3964
d2 = lookup_one_qstr_excl(new_last, rd->new_parent,
fs/namei.c
3978
if (rd->flags & RENAME_EXCHANGE)
fs/namei.c
3985
rd->old_dentry = dget(old_dentry);
fs/namei.c
3986
rd->new_dentry = d2;
fs/namei.c
3987
rd->old_parent = dget(old_dentry->d_parent);
fs/namei.c
3993
unlock_rename(old_dentry->d_parent, rd->new_parent);
fs/namei.c
4021
int start_renaming_dentry(struct renamedata *rd, int lookup_flags,
fs/namei.c
4026
err = lookup_one_common(rd->mnt_idmap, new_last, rd->new_parent);
fs/namei.c
4029
return __start_renaming_dentry(rd, lookup_flags, old_dentry, new_last);
fs/namei.c
4050
start_renaming_two_dentries(struct renamedata *rd,
fs/namei.c
4057
trap = lock_rename_child(old_dentry, rd->new_parent);
fs/namei.c
4062
(rd->old_parent && rd->old_parent != old_dentry->d_parent))
fs/namei.c
4066
rd->new_parent != new_dentry->d_parent)
fs/namei.c
4076
if (rd->flags & RENAME_EXCHANGE)
fs/namei.c
4084
if (d_is_positive(new_dentry) && (rd->flags & RENAME_NOREPLACE))
fs/namei.c
4087
rd->old_dentry = dget(old_dentry);
fs/namei.c
4088
rd->new_dentry = dget(new_dentry);
fs/namei.c
4089
rd->old_parent = dget(old_dentry->d_parent);
fs/namei.c
4093
unlock_rename(old_dentry->d_parent, rd->new_parent);
fs/namei.c
4098
void end_renaming(struct renamedata *rd)
fs/namei.c
4100
unlock_rename(rd->old_parent, rd->new_parent);
fs/namei.c
4101
dput(rd->old_dentry);
fs/namei.c
4102
dput(rd->new_dentry);
fs/namei.c
4103
dput(rd->old_parent);
fs/namei.c
5902
int vfs_rename(struct renamedata *rd)
fs/namei.c
5905
struct inode *old_dir = d_inode(rd->old_parent);
fs/namei.c
5906
struct inode *new_dir = d_inode(rd->new_parent);
fs/namei.c
5907
struct dentry *old_dentry = rd->old_dentry;
fs/namei.c
5908
struct dentry *new_dentry = rd->new_dentry;
fs/namei.c
5909
struct delegated_inode *delegated_inode = rd->delegated_inode;
fs/namei.c
5910
unsigned int flags = rd->flags;
fs/namei.c
5922
error = may_delete_dentry(rd->mnt_idmap, old_dir, old_dentry, is_dir);
fs/namei.c
5927
error = may_create_dentry(rd->mnt_idmap, new_dir, new_dentry);
fs/namei.c
5932
error = may_delete_dentry(rd->mnt_idmap, new_dir,
fs/namei.c
5935
error = may_delete_dentry(rd->mnt_idmap, new_dir,
fs/namei.c
5950
error = inode_permission(rd->mnt_idmap, source,
fs/namei.c
5956
error = inode_permission(rd->mnt_idmap, target,
fs/namei.c
6032
error = old_dir->i_op->rename(rd->mnt_idmap, old_dir, old_dentry,
fs/namei.c
6074
struct renamedata rd;
fs/namei.c
6119
rd.old_parent = old_path.dentry;
fs/namei.c
6120
rd.mnt_idmap = mnt_idmap(old_path.mnt);
fs/namei.c
6121
rd.new_parent = new_path.dentry;
fs/namei.c
6122
rd.delegated_inode = &delegated_inode;
fs/namei.c
6123
rd.flags = flags;
fs/namei.c
6125
error = __start_renaming(&rd, lookup_flags, &old_last, &new_last);
fs/namei.c
6130
if (!d_is_dir(rd.new_dentry)) {
fs/namei.c
6137
if (!d_is_dir(rd.old_dentry)) {
fs/namei.c
6145
error = security_path_rename(&old_path, rd.old_dentry,
fs/namei.c
6146
&new_path, rd.new_dentry, flags);
fs/namei.c
6150
error = vfs_rename(&rd);
fs/namei.c
6152
end_renaming(&rd);
fs/nfsd/vfs.c
2141
struct renamedata rd = {};
fs/nfsd/vfs.c
2177
rd.mnt_idmap = &nop_mnt_idmap;
fs/nfsd/vfs.c
2178
rd.old_parent = fdentry;
fs/nfsd/vfs.c
2179
rd.new_parent = tdentry;
fs/nfsd/vfs.c
2181
host_err = start_renaming(&rd, 0, &QSTR_LEN(fname, flen),
fs/nfsd/vfs.c
2195
type = d_inode(rd.old_dentry)->i_mode & S_IFMT;
fs/nfsd/vfs.c
2197
if (d_inode(rd.new_dentry))
fs/nfsd/vfs.c
2198
type = d_inode(rd.new_dentry)->i_mode & S_IFMT;
fs/nfsd/vfs.c
2200
if ((rd.new_dentry->d_sb->s_export_op->flags & EXPORT_OP_CLOSE_BEFORE_UNLINK) &&
fs/nfsd/vfs.c
2201
nfsd_has_cached_files(rd.new_dentry)) {
fs/nfsd/vfs.c
2202
close_cached = dget(rd.new_dentry);
fs/nfsd/vfs.c
2208
host_err = vfs_rename(&rd);
fs/nfsd/vfs.c
2211
if (!nfsd_wait_for_delegreturn(rqstp, d_inode(rd.old_dentry)))
fs/nfsd/vfs.c
2238
end_renaming(&rd);
fs/overlayfs/copy_up.c
527
struct renamedata rd = {};
fs/overlayfs/copy_up.c
559
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/copy_up.c
560
rd.old_parent = indexdir;
fs/overlayfs/copy_up.c
561
rd.new_parent = indexdir;
fs/overlayfs/copy_up.c
562
err = start_renaming_dentry(&rd, 0, temp, &name);
fs/overlayfs/copy_up.c
566
err = ovl_do_rename_rd(&rd);
fs/overlayfs/copy_up.c
567
end_renaming(&rd);
fs/overlayfs/copy_up.c
763
struct renamedata rd = {};
fs/overlayfs/copy_up.c
807
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/copy_up.c
808
rd.old_parent = c->workdir;
fs/overlayfs/copy_up.c
809
rd.new_parent = c->destdir;
fs/overlayfs/copy_up.c
810
rd.flags = 0;
fs/overlayfs/copy_up.c
811
err = start_renaming_dentry(&rd, 0, temp,
fs/overlayfs/copy_up.c
822
err = ovl_do_rename_rd(&rd);
fs/overlayfs/copy_up.c
823
end_renaming(&rd);
fs/overlayfs/dir.c
1201
struct renamedata rd = {};
fs/overlayfs/dir.c
1233
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/dir.c
1234
rd.old_parent = old_upperdir;
fs/overlayfs/dir.c
1235
rd.new_parent = new_upperdir;
fs/overlayfs/dir.c
1236
rd.flags = ovlrd->flags;
fs/overlayfs/dir.c
1238
err = start_renaming(&rd, 0,
fs/overlayfs/dir.c
1245
if (!ovl_matches_upper(old, rd.old_dentry))
fs/overlayfs/dir.c
1254
if (rd.new_dentry != ovlrd->opaquedir)
fs/overlayfs/dir.c
1257
if (!ovl_matches_upper(new, rd.new_dentry))
fs/overlayfs/dir.c
126
struct renamedata rd = {};
fs/overlayfs/dir.c
1261
if (!d_is_negative(rd.new_dentry)) {
fs/overlayfs/dir.c
1262
if (!new_opaque || !ovl_upper_is_whiteout(ofs, rd.new_dentry))
fs/overlayfs/dir.c
1270
if (rd.old_dentry->d_inode == rd.new_dentry->d_inode)
fs/overlayfs/dir.c
1277
err = ovl_set_opaque_xerr(old, rd.old_dentry, -EXDEV);
fs/overlayfs/dir.c
1285
err = ovl_set_opaque_xerr(new, rd.new_dentry, -EXDEV);
fs/overlayfs/dir.c
1289
err = ovl_do_rename_rd(&rd);
fs/overlayfs/dir.c
1292
whiteout = dget(rd.new_dentry);
fs/overlayfs/dir.c
1295
end_renaming(&rd);
fs/overlayfs/dir.c
138
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/dir.c
139
rd.old_parent = ofs->workdir;
fs/overlayfs/dir.c
140
rd.new_parent = dir;
fs/overlayfs/dir.c
141
rd.flags = flags;
fs/overlayfs/dir.c
142
err = start_renaming_two_dentries(&rd, whiteout, dentry);
fs/overlayfs/dir.c
144
err = ovl_do_rename_rd(&rd);
fs/overlayfs/dir.c
145
end_renaming(&rd);
fs/overlayfs/dir.c
390
struct renamedata rd = {};
fs/overlayfs/dir.c
416
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/dir.c
417
rd.old_parent = workdir;
fs/overlayfs/dir.c
418
rd.new_parent = upperdir;
fs/overlayfs/dir.c
419
rd.flags = RENAME_EXCHANGE;
fs/overlayfs/dir.c
420
err = start_renaming_two_dentries(&rd, opaquedir, upper);
fs/overlayfs/dir.c
438
err = ovl_do_rename_rd(&rd);
fs/overlayfs/dir.c
439
end_renaming(&rd);
fs/overlayfs/dir.c
452
end_renaming(&rd);
fs/overlayfs/dir.c
475
struct renamedata rd = {};
fs/overlayfs/dir.c
507
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/dir.c
508
rd.old_parent = workdir;
fs/overlayfs/dir.c
509
rd.new_parent = upperdir;
fs/overlayfs/dir.c
510
rd.flags = 0;
fs/overlayfs/dir.c
511
err = start_renaming_two_dentries(&rd, newdentry, upper);
fs/overlayfs/dir.c
548
rd.flags = RENAME_EXCHANGE;
fs/overlayfs/dir.c
549
err = ovl_do_rename_rd(&rd);
fs/overlayfs/dir.c
550
end_renaming(&rd);
fs/overlayfs/dir.c
556
err = ovl_do_rename_rd(&rd);
fs/overlayfs/dir.c
557
end_renaming(&rd);
fs/overlayfs/dir.c
577
end_renaming(&rd);
fs/overlayfs/overlayfs.h
364
static inline int ovl_do_rename_rd(struct renamedata *rd)
fs/overlayfs/overlayfs.h
368
pr_debug("rename(%pd2, %pd2, 0x%x)\n", rd->old_dentry, rd->new_dentry,
fs/overlayfs/overlayfs.h
369
rd->flags);
fs/overlayfs/overlayfs.h
370
err = vfs_rename(rd);
fs/overlayfs/overlayfs.h
373
rd->old_dentry, rd->new_dentry, err);
fs/overlayfs/overlayfs.h
382
struct renamedata rd = {
fs/overlayfs/overlayfs.h
391
return ovl_do_rename_rd(&rd);
fs/overlayfs/super.c
576
struct renamedata rd = {};
fs/overlayfs/super.c
585
rd.mnt_idmap = ovl_upper_mnt_idmap(ofs);
fs/overlayfs/super.c
586
rd.old_parent = workdir;
fs/overlayfs/super.c
587
rd.new_parent = workdir;
fs/overlayfs/super.c
588
rd.flags = RENAME_WHITEOUT;
fs/overlayfs/super.c
590
err = start_renaming_dentry(&rd, 0, temp, &QSTR(name2));
fs/overlayfs/super.c
598
err = ovl_do_rename_rd(&rd);
fs/overlayfs/super.c
599
end_renaming(&rd);
fs/qnx4/inode.c
161
int rd, rl;
fs/qnx4/inode.c
167
rd = le32_to_cpu(s->RootDir.di_first_xtnt.xtnt_blk) - 1;
fs/qnx4/inode.c
170
bh = sb_bread(sb, rd + j); /* root dir, first block */
fs/smb/server/vfs.c
668
struct renamedata rd;
fs/smb/server/vfs.c
695
rd.mnt_idmap = mnt_idmap(old_path->mnt);
fs/smb/server/vfs.c
696
rd.old_parent = NULL;
fs/smb/server/vfs.c
697
rd.new_parent = new_path.dentry;
fs/smb/server/vfs.c
698
rd.flags = flags;
fs/smb/server/vfs.c
699
rd.delegated_inode = NULL;
fs/smb/server/vfs.c
700
err = start_renaming_dentry(&rd, lookup_flags, old_child, &new_last);
fs/smb/server/vfs.c
715
if (d_is_symlink(rd.new_dentry)) {
fs/smb/server/vfs.c
720
err = vfs_rename(&rd);
fs/smb/server/vfs.c
725
end_renaming(&rd);
fs/ubifs/lpt.c
1889
int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr)
fs/ubifs/lpt.c
1893
if (rd) {
fs/ubifs/lpt.c
1910
if (rd)
fs/ubifs/ubifs.h
1958
int ubifs_lpt_init(struct ubifs_info *c, int rd, int wr);
fs/xfs/scrub/dir_repair.c
1013
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1015
struct xfs_inode *dp = rd->sc->ip;
fs/xfs/scrub/dir_repair.c
1023
error = xrep_dir_recover_sf(rd);
fs/xfs/scrub/dir_repair.c
1025
error = xfs_iread_extents(rd->sc->tp, dp, XFS_DATA_FORK);
fs/xfs/scrub/dir_repair.c
1029
error = xrep_dir_recover(rd);
fs/xfs/scrub/dir_repair.c
1034
return xrep_dir_flush_stashed(rd);
fs/xfs/scrub/dir_repair.c
1040
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1042
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1052
error = xrep_dir_find_parent(rd);
fs/xfs/scrub/dir_repair.c
1062
error = xrep_dir_find_entries(rd);
fs/xfs/scrub/dir_repair.c
1113
struct xrep_dir *rd = priv;
fs/xfs/scrub/dir_repair.c
1132
mutex_lock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1133
error = xrep_dir_stash_createname(rd, &xname, ip->i_ino);
fs/xfs/scrub/dir_repair.c
1134
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1151
struct xrep_dir *rd = priv;
fs/xfs/scrub/dir_repair.c
1154
if (ino != rd->sc->ip->i_ino)
fs/xfs/scrub/dir_repair.c
1158
if (!xfs_verify_dir_ino(rd->sc->mp, ino))
fs/xfs/scrub/dir_repair.c
1173
xrep_findparent_scan_found(&rd->pscan, dp->i_ino);
fs/xfs/scrub/dir_repair.c
1183
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
1186
return ip != rd->sc->ip && !xrep_is_tempfile(ip);
fs/xfs/scrub/dir_repair.c
1197
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
1203
if (!xrep_dir_want_scan(rd, ip))
fs/xfs/scrub/dir_repair.c
1225
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
1231
lock_mode = xrep_dir_scan_ilock(rd, ip);
fs/xfs/scrub/dir_repair.c
1233
if (!xrep_dir_want_scan(rd, ip))
fs/xfs/scrub/dir_repair.c
1245
error = xchk_xattr_walk(rd->sc, ip, xrep_dir_scan_pptr, NULL, rd);
fs/xfs/scrub/dir_repair.c
1259
error = xchk_dir_walk(rd->sc, ip, xrep_dir_scan_dirent, rd);
fs/xfs/scrub/dir_repair.c
1265
xchk_iscan_mark_visited(&rd->pscan.iscan, ip);
fs/xfs/scrub/dir_repair.c
1276
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1278
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1284
xrep_findparent_scan_found(&rd->pscan, sc->ip->i_ino);
fs/xfs/scrub/dir_repair.c
1298
while ((error = xchk_iscan_iter(&rd->pscan.iscan, &ip)) == 1) {
fs/xfs/scrub/dir_repair.c
1301
error = xrep_dir_scan_file(rd, ip);
fs/xfs/scrub/dir_repair.c
1307
mutex_lock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1308
flush = xrep_dir_want_flush_stashed(rd);
fs/xfs/scrub/dir_repair.c
1309
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1317
error = xrep_dir_replay_updates(rd);
fs/xfs/scrub/dir_repair.c
1328
xchk_iscan_iter_finish(&rd->pscan.iscan);
fs/xfs/scrub/dir_repair.c
1345
xchk_trans_cancel(rd->sc);
fs/xfs/scrub/dir_repair.c
1360
struct xrep_dir *rd;
fs/xfs/scrub/dir_repair.c
1364
rd = container_of(nb, struct xrep_dir, pscan.dhook.dirent_hook.nb);
fs/xfs/scrub/dir_repair.c
1365
sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1373
xchk_iscan_want_live_update(&rd->pscan.iscan, p->ip->i_ino)) {
fs/xfs/scrub/dir_repair.c
1374
mutex_lock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1376
error = xrep_dir_stash_createname(rd, p->name,
fs/xfs/scrub/dir_repair.c
1379
error = xrep_dir_stash_removename(rd, p->name,
fs/xfs/scrub/dir_repair.c
1381
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
1392
xchk_iscan_want_live_update(&rd->pscan.iscan, p->dp->i_ino)) {
fs/xfs/scrub/dir_repair.c
1398
xrep_findparent_scan_found(&rd->pscan, p->dp->i_ino);
fs/xfs/scrub/dir_repair.c
1402
rd->pscan.parent_ino);
fs/xfs/scrub/dir_repair.c
1404
xrep_findparent_scan_found(&rd->pscan, NULLFSINO);
fs/xfs/scrub/dir_repair.c
1410
xchk_iscan_abort(&rd->pscan.iscan);
fs/xfs/scrub/dir_repair.c
1421
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
1424
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1443
xrep_dir_init_args(rd, sc->tempip, NULL);
fs/xfs/scrub/dir_repair.c
1444
return xfs_dir2_sf_create(&rd->args, parent_ino);
fs/xfs/scrub/dir_repair.c
1515
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
1521
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1530
xrep_dir_init_args(rd, dp, name);
fs/xfs/scrub/dir_repair.c
1531
rd->args.inumber = inum;
fs/xfs/scrub/dir_repair.c
1532
rd->args.total = total;
fs/xfs/scrub/dir_repair.c
1533
return xfs_dir_replace_args(&rd->args);
fs/xfs/scrub/dir_repair.c
1542
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1544
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1548
rd->subdirs + 2,
fs/xfs/scrub/dir_repair.c
1566
if (rd->dirents == 0) {
fs/xfs/scrub/dir_repair.c
1567
rd->needs_adoption = false;
fs/xfs/scrub/dir_repair.c
1603
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1605
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1609
return xrep_tempexch_trans_alloc(sc, XFS_DATA_FORK, &rd->tx);
fs/xfs/scrub/dir_repair.c
1618
error = xrep_dir_replay_updates(rd);
fs/xfs/scrub/dir_repair.c
1622
error = xrep_tempexch_trans_alloc(sc, XFS_DATA_FORK, &rd->tx);
fs/xfs/scrub/dir_repair.c
1626
if (xfarray_length(rd->dir_entries) == 0)
fs/xfs/scrub/dir_repair.c
1638
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1640
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1652
if (rd->pscan.parent_ino == NULLFSINO) {
fs/xfs/scrub/dir_repair.c
1653
rd->needs_adoption = true;
fs/xfs/scrub/dir_repair.c
1654
rd->pscan.parent_ino = rd->sc->mp->m_sb.sb_rootino;
fs/xfs/scrub/dir_repair.c
1665
error = xchk_dir_lookup(sc, rd->sc->tempip, &xfs_name_dotdot, &ino);
fs/xfs/scrub/dir_repair.c
1669
if (rd->pscan.parent_ino != ino) {
fs/xfs/scrub/dir_repair.c
1670
error = xrep_dir_replace(rd, rd->sc->tempip, &xfs_name_dotdot,
fs/xfs/scrub/dir_repair.c
1671
rd->pscan.parent_ino, rd->tx.req.resblks);
fs/xfs/scrub/dir_repair.c
1692
return xrep_dir_set_nlink(rd);
fs/xfs/scrub/dir_repair.c
1699
error = xrep_tempfile_roll_trans(rd->sc);
fs/xfs/scrub/dir_repair.c
1712
error = xrep_dir_set_nlink(rd);
fs/xfs/scrub/dir_repair.c
1716
return xrep_tempexch_contents(sc, &rd->tx);
fs/xfs/scrub/dir_repair.c
172
struct xrep_dir *rd = sc->buf;
fs/xfs/scrub/dir_repair.c
1725
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1727
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1730
trace_xrep_dir_rebuild_tree(sc->ip, rd->pscan.parent_ino);
fs/xfs/scrub/dir_repair.c
1737
error = xrep_tempfile_iolock_polled(rd->sc);
fs/xfs/scrub/dir_repair.c
174
xrep_findparent_scan_teardown(&rd->pscan);
fs/xfs/scrub/dir_repair.c
1746
error = xrep_dir_finalize_tempdir(rd);
fs/xfs/scrub/dir_repair.c
175
if (rd->dir_names)
fs/xfs/scrub/dir_repair.c
1750
if (xchk_iscan_aborted(&rd->pscan.iscan))
fs/xfs/scrub/dir_repair.c
1758
error = xrep_dir_swap(rd);
fs/xfs/scrub/dir_repair.c
176
xfblob_destroy(rd->dir_names);
fs/xfs/scrub/dir_repair.c
1767
error = xrep_dir_reset_fork(rd, sc->mp->m_rootip->i_ino);
fs/xfs/scrub/dir_repair.c
177
rd->dir_names = NULL;
fs/xfs/scrub/dir_repair.c
178
if (rd->dir_entries)
fs/xfs/scrub/dir_repair.c
1788
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
179
xfarray_destroy(rd->dir_entries);
fs/xfs/scrub/dir_repair.c
1790
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1795
sizeof(struct xrep_dirent), &rd->dir_entries);
fs/xfs/scrub/dir_repair.c
1799
error = xfblob_create("directory entry names", &rd->dir_names);
fs/xfs/scrub/dir_repair.c
180
rd->dir_entries = NULL;
fs/xfs/scrub/dir_repair.c
1804
error = __xrep_findparent_scan_start(sc, &rd->pscan,
fs/xfs/scrub/dir_repair.c
1807
error = xrep_findparent_scan_start(sc, &rd->pscan);
fs/xfs/scrub/dir_repair.c
1814
xfblob_destroy(rd->dir_names);
fs/xfs/scrub/dir_repair.c
1815
rd->dir_names = NULL;
fs/xfs/scrub/dir_repair.c
1817
xfarray_destroy(rd->dir_entries);
fs/xfs/scrub/dir_repair.c
1818
rd->dir_entries = NULL;
fs/xfs/scrub/dir_repair.c
1832
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
1834
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
1866
error = xrep_adoption_trans_alloc(sc, &rd->adoption);
fs/xfs/scrub/dir_repair.c
1870
error = xrep_adoption_compute_name(&rd->adoption, &rd->xname);
fs/xfs/scrub/dir_repair.c
188
struct xrep_dir *rd;
fs/xfs/scrub/dir_repair.c
1889
error = xrep_adoption_move(&rd->adoption);
fs/xfs/scrub/dir_repair.c
1898
error = xrep_adoption_trans_roll(&rd->adoption);
fs/xfs/scrub/dir_repair.c
1922
struct xrep_dir *rd = sc->buf;
fs/xfs/scrub/dir_repair.c
1932
error = xrep_dir_setup_scan(rd);
fs/xfs/scrub/dir_repair.c
1937
error = xrep_dir_scan_dirtree(rd);
fs/xfs/scrub/dir_repair.c
1939
error = xrep_dir_salvage_entries(rd);
fs/xfs/scrub/dir_repair.c
1947
error = xrep_dir_rebuild_tree(rd);
fs/xfs/scrub/dir_repair.c
1951
if (rd->needs_adoption) {
fs/xfs/scrub/dir_repair.c
1952
if (!xrep_orphanage_can_adopt(rd->sc))
fs/xfs/scrub/dir_repair.c
1955
error = xrep_dir_move_to_orphanage(rd);
fs/xfs/scrub/dir_repair.c
201
rd = kvzalloc_obj(struct xrep_dir, XCHK_GFP_FLAGS);
fs/xfs/scrub/dir_repair.c
202
if (!rd)
fs/xfs/scrub/dir_repair.c
204
rd->sc = sc;
fs/xfs/scrub/dir_repair.c
205
rd->xname.name = rd->namebuf;
fs/xfs/scrub/dir_repair.c
206
sc->buf = rd;
fs/xfs/scrub/dir_repair.c
217
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
219
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
242
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
244
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
262
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
266
ino = xrep_findparent_self_reference(rd->sc);
fs/xfs/scrub/dir_repair.c
268
xrep_findparent_scan_finish_early(&rd->pscan, ino);
fs/xfs/scrub/dir_repair.c
272
ino = xrep_dir_dcache_parent(rd);
fs/xfs/scrub/dir_repair.c
274
xrep_findparent_scan_finish_early(&rd->pscan, ino);
fs/xfs/scrub/dir_repair.c
278
ino = xrep_dir_lookup_parent(rd);
fs/xfs/scrub/dir_repair.c
280
xrep_findparent_scan_finish_early(&rd->pscan, ino);
fs/xfs/scrub/dir_repair.c
290
return xrep_findparent_scan(&rd->pscan);
fs/xfs/scrub/dir_repair.c
299
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
304
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
307
if (ino == rd->sc->ip->i_ino)
fs/xfs/scrub/dir_repair.c
329
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
341
trace_xrep_dir_stash_createname(rd->sc->tempip, name, ino);
fs/xfs/scrub/dir_repair.c
343
error = xfblob_storename(rd->dir_names, &dirent.name_cookie, name);
fs/xfs/scrub/dir_repair.c
347
return xfarray_append(rd->dir_entries, &dirent);
fs/xfs/scrub/dir_repair.c
356
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
368
trace_xrep_dir_stash_removename(rd->sc->tempip, name, ino);
fs/xfs/scrub/dir_repair.c
370
error = xfblob_storename(rd->dir_names, &dirent.name_cookie, name);
fs/xfs/scrub/dir_repair.c
374
return xfarray_append(rd->dir_entries, &dirent);
fs/xfs/scrub/dir_repair.c
380
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
388
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
423
if (xfs_is_metadir_inode(ip) != xfs_is_metadir_inode(rd->sc->ip)) {
fs/xfs/scrub/dir_repair.c
431
return xrep_dir_stash_createname(rd, &xname, ino);
fs/xfs/scrub/dir_repair.c
437
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
443
ino = xfs_dir2_sf_get_ino(rd->sc->mp, sfp, sfep);
fs/xfs/scrub/dir_repair.c
444
if (!xrep_dir_want_salvage(rd, sfep->name, sfep->namelen, ino))
fs/xfs/scrub/dir_repair.c
447
return xrep_dir_salvage_entry(rd, sfep->name, sfep->namelen, ino);
fs/xfs/scrub/dir_repair.c
453
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
459
if (!xrep_dir_want_salvage(rd, dep->name, dep->namelen, ino))
fs/xfs/scrub/dir_repair.c
462
return xrep_dir_salvage_entry(rd, dep->name, dep->namelen, ino);
fs/xfs/scrub/dir_repair.c
468
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
471
struct xfs_da_geometry *geo = rd->sc->mp->m_dir_geo;
fs/xfs/scrub/dir_repair.c
488
if (xchk_should_terminate(rd->sc, &error))
fs/xfs/scrub/dir_repair.c
498
offset += xfs_dir2_data_entsize(rd->sc->mp, dep->namelen);
fs/xfs/scrub/dir_repair.c
503
error = xrep_dir_salvage_data_entry(rd, dep);
fs/xfs/scrub/dir_repair.c
515
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
525
ifp = xfs_ifork_ptr(rd->sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/dir_repair.c
530
trace_xrep_dir_salvaged_parent(rd->sc->ip, ino);
fs/xfs/scrub/dir_repair.c
534
if (xchk_should_terminate(rd->sc, &error))
fs/xfs/scrub/dir_repair.c
537
next = xfs_dir2_sf_nextentry(rd->sc->mp, hdr, sfep);
fs/xfs/scrub/dir_repair.c
542
error = xrep_dir_salvage_sf_entry(rd, hdr, sfep);
fs/xfs/scrub/dir_repair.c
562
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
565
struct xfs_inode *dp = rd->sc->ip;
fs/xfs/scrub/dir_repair.c
566
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
592
error = xfs_bmap_last_before(rd->sc->tp, dp, &last, XFS_DATA_FORK);
fs/xfs/scrub/dir_repair.c
604
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
617
error = xfs_da_read_buf(rd->sc->tp, rd->sc->ip, dabno,
fs/xfs/scrub/dir_repair.c
625
trace_xrep_dir_recover_dirblock(rd->sc->ip, dabno,
fs/xfs/scrub/dir_repair.c
647
if (xfs_dir3_block_header_check(bp, rd->sc->ip->i_ino) != NULL)
fs/xfs/scrub/dir_repair.c
654
if (xfs_dir3_data_header_check(bp, rd->sc->ip->i_ino) != NULL)
fs/xfs/scrub/dir_repair.c
662
error = xrep_dir_recover_data(rd, bp);
fs/xfs/scrub/dir_repair.c
666
xfs_trans_brelse(rd->sc->tp, bp);
fs/xfs/scrub/dir_repair.c
672
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
676
memset(&rd->args, 0, sizeof(struct xfs_da_args));
fs/xfs/scrub/dir_repair.c
677
rd->args.geo = rd->sc->mp->m_dir_geo;
fs/xfs/scrub/dir_repair.c
678
rd->args.whichfork = XFS_DATA_FORK;
fs/xfs/scrub/dir_repair.c
679
rd->args.owner = rd->sc->ip->i_ino;
fs/xfs/scrub/dir_repair.c
680
rd->args.trans = rd->sc->tp;
fs/xfs/scrub/dir_repair.c
681
rd->args.dp = dp;
fs/xfs/scrub/dir_repair.c
684
rd->args.name = name->name;
fs/xfs/scrub/dir_repair.c
685
rd->args.namelen = name->len;
fs/xfs/scrub/dir_repair.c
686
rd->args.filetype = name->type;
fs/xfs/scrub/dir_repair.c
687
rd->args.hashval = xfs_dir2_hashname(rd->sc->mp, name);
fs/xfs/scrub/dir_repair.c
693
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
698
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
699
struct xfs_inode *dp = rd->sc->tempip;
fs/xfs/scrub/dir_repair.c
710
xrep_dir_init_args(rd, dp, name);
fs/xfs/scrub/dir_repair.c
711
rd->args.inumber = inum;
fs/xfs/scrub/dir_repair.c
712
rd->args.total = total;
fs/xfs/scrub/dir_repair.c
713
rd->args.op_flags = XFS_DA_OP_ADDNAME | XFS_DA_OP_OKNOENT;
fs/xfs/scrub/dir_repair.c
714
return xfs_dir_createname_args(&rd->args);
fs/xfs/scrub/dir_repair.c
720
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
724
struct xfs_inode *dp = rd->args.dp;
fs/xfs/scrub/dir_repair.c
728
xrep_dir_init_args(rd, dp, name);
fs/xfs/scrub/dir_repair.c
729
rd->args.op_flags = 0;
fs/xfs/scrub/dir_repair.c
730
rd->args.total = total;
fs/xfs/scrub/dir_repair.c
733
return xfs_dir_removename_args(&rd->args);
fs/xfs/scrub/dir_repair.c
743
struct xrep_dir *rd,
fs/xfs/scrub/dir_repair.c
747
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
755
error = xchk_trans_alloc(rd->sc, resblks);
fs/xfs/scrub/dir_repair.c
760
xrep_tempfile_ilock(rd->sc);
fs/xfs/scrub/dir_repair.c
761
xfs_trans_ijoin(rd->sc->tp, rd->sc->tempip, 0);
fs/xfs/scrub/dir_repair.c
772
error = xchk_dir_lookup(rd->sc, rd->sc->tempip, xname, &ino);
fs/xfs/scrub/dir_repair.c
779
error = xrep_dir_replay_createname(rd, xname, dirent->ino,
fs/xfs/scrub/dir_repair.c
785
rd->subdirs++;
fs/xfs/scrub/dir_repair.c
786
rd->dirents++;
fs/xfs/scrub/dir_repair.c
796
error = xchk_dir_lookup(rd->sc, rd->sc->tempip, xname, &ino);
fs/xfs/scrub/dir_repair.c
808
error = xrep_dir_replay_removename(rd, xname, resblks);
fs/xfs/scrub/dir_repair.c
813
rd->subdirs--;
fs/xfs/scrub/dir_repair.c
814
rd->dirents--;
fs/xfs/scrub/dir_repair.c
823
error = xrep_trans_commit(rd->sc);
fs/xfs/scrub/dir_repair.c
827
xrep_tempfile_iunlock(rd->sc);
fs/xfs/scrub/dir_repair.c
830
xchk_trans_cancel(rd->sc);
fs/xfs/scrub/dir_repair.c
831
xrep_tempfile_iunlock(rd->sc);
fs/xfs/scrub/dir_repair.c
845
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
851
mutex_lock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
852
foreach_xfarray_idx(rd->dir_entries, array_cur) {
fs/xfs/scrub/dir_repair.c
855
error = xfarray_load(rd->dir_entries, array_cur, &dirent);
fs/xfs/scrub/dir_repair.c
859
error = xfblob_loadname(rd->dir_names, dirent.name_cookie,
fs/xfs/scrub/dir_repair.c
860
&rd->xname, dirent.namelen);
fs/xfs/scrub/dir_repair.c
863
rd->xname.type = dirent.ftype;
fs/xfs/scrub/dir_repair.c
864
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
866
error = xrep_dir_replay_update(rd, &rd->xname, &dirent);
fs/xfs/scrub/dir_repair.c
869
mutex_lock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
873
xfarray_truncate(rd->dir_entries);
fs/xfs/scrub/dir_repair.c
874
xfblob_truncate(rd->dir_names);
fs/xfs/scrub/dir_repair.c
875
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
878
mutex_unlock(&rd->pscan.lock);
fs/xfs/scrub/dir_repair.c
889
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
912
error = xrep_trans_commit(rd->sc);
fs/xfs/scrub/dir_repair.c
915
xchk_iunlock(rd->sc, XFS_ILOCK_EXCL);
fs/xfs/scrub/dir_repair.c
923
error = xrep_tempfile_iolock_polled(rd->sc);
fs/xfs/scrub/dir_repair.c
928
error = xrep_dir_replay_updates(rd);
fs/xfs/scrub/dir_repair.c
929
xrep_tempfile_iounlock(rd->sc);
fs/xfs/scrub/dir_repair.c
936
error = xchk_trans_alloc(rd->sc, 0);
fs/xfs/scrub/dir_repair.c
939
xchk_ilock(rd->sc, XFS_ILOCK_EXCL);
fs/xfs/scrub/dir_repair.c
946
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
950
bytes = xfarray_bytes(rd->dir_entries) + xfblob_bytes(rd->dir_names);
fs/xfs/scrub/dir_repair.c
957
struct xrep_dir *rd)
fs/xfs/scrub/dir_repair.c
960
struct xfs_scrub *sc = rd->sc;
fs/xfs/scrub/dir_repair.c
968
xrep_dir_guess_format(rd, &magic_guess);
fs/xfs/scrub/dir_repair.c
987
if (xchk_should_terminate(rd->sc, &error))
fs/xfs/scrub/dir_repair.c
990
error = xrep_dir_recover_dirblock(rd,
fs/xfs/scrub/dir_repair.c
996
if (xrep_dir_want_flush_stashed(rd)) {
fs/xfs/scrub/dir_repair.c
997
error = xrep_dir_flush_stashed(rd);
include/linux/hdlcdrv.h
164
ret = !((HDLCDRV_HDLCBUFFER - 1 + hb->rd - hb->wr) % HDLCDRV_HDLCBUFFER);
include/linux/hdlcdrv.h
177
ret = (hb->rd == hb->wr);
include/linux/hdlcdrv.h
191
if (hb->rd == hb->wr)
include/linux/hdlcdrv.h
194
newr = (hb->rd+1) % HDLCDRV_HDLCBUFFER;
include/linux/hdlcdrv.h
195
val = hb->buf[hb->rd];
include/linux/hdlcdrv.h
196
hb->rd = newr;
include/linux/hdlcdrv.h
212
if (newp != hb->rd) {
include/linux/hdlcdrv.h
28
unsigned rd, wr;
include/linux/hdlcdrv.h
34
unsigned int rd;
include/linux/namei.h
174
int start_renaming(struct renamedata *rd, int lookup_flags,
include/linux/namei.h
176
int start_renaming_dentry(struct renamedata *rd, int lookup_flags,
include/linux/namei.h
178
int start_renaming_two_dentries(struct renamedata *rd,
include/linux/namei.h
180
void end_renaming(struct renamedata *rd);
include/linux/netdevice.h
3267
netdev_offload_xstats_report_delta(struct netdev_notifier_offload_xstats_rd *rd,
include/linux/of.h
1637
extern int of_reconfig_notify(unsigned long, struct of_reconfig_data *rd);
include/linux/raid_class.h
59
struct raid_data *rd; \
include/linux/raid_class.h
61
rd = dev_get_drvdata(device); \
include/linux/raid_class.h
62
rd->attr = value; \
include/linux/raid_class.h
68
struct raid_data *rd; \
include/linux/raid_class.h
70
rd = dev_get_drvdata(device); \
include/linux/raid_class.h
71
return rd->attr; \
include/linux/sched/deadline.h
34
extern void dl_clear_root_domain(struct root_domain *rd);
include/net/cfg80211.h
7635
struct ieee80211_regdomain *rd);
include/net/cfg80211.h
7649
struct ieee80211_regdomain *rd);
include/trace/events/sched.h
862
TP_PROTO(struct root_domain *rd, bool overutilized),
include/trace/events/sched.h
863
TP_ARGS(rd, overutilized));
io_uring/io_uring.c
2692
struct io_uring_region_desc rd;
io_uring/io_uring.c
2700
memset(&rd, 0, sizeof(rd));
io_uring/io_uring.c
2701
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/io_uring.c
2703
rd.user_addr = p->cq_off.user_addr;
io_uring/io_uring.c
2704
rd.flags |= IORING_MEM_REGION_TYPE_USER;
io_uring/io_uring.c
2706
ret = io_create_region(ctx, &ctx->ring_region, &rd, IORING_OFF_CQ_RING);
io_uring/io_uring.c
2714
memset(&rd, 0, sizeof(rd));
io_uring/io_uring.c
2715
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/io_uring.c
2717
rd.user_addr = p->sq_off.user_addr;
io_uring/io_uring.c
2718
rd.flags |= IORING_MEM_REGION_TYPE_USER;
io_uring/io_uring.c
2720
ret = io_create_region(ctx, &ctx->sq_region, &rd, IORING_OFF_SQES);
io_uring/kbuf.c
620
struct io_uring_region_desc rd;
io_uring/kbuf.c
655
memset(&rd, 0, sizeof(rd));
io_uring/kbuf.c
656
rd.size = PAGE_ALIGN(ring_size);
io_uring/kbuf.c
658
rd.user_addr = reg.ring_addr;
io_uring/kbuf.c
659
rd.flags |= IORING_MEM_REGION_TYPE_USER;
io_uring/kbuf.c
661
ret = io_create_region(ctx, &bl->region, &rd, mmap_offset);
io_uring/register.c
501
struct io_uring_region_desc rd;
io_uring/register.c
525
memset(&rd, 0, sizeof(rd));
io_uring/register.c
526
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/register.c
528
rd.user_addr = p->cq_off.user_addr;
io_uring/register.c
529
rd.flags |= IORING_MEM_REGION_TYPE_USER;
io_uring/register.c
531
ret = io_create_region(ctx, &n.ring_region, &rd, IORING_OFF_CQ_RING);
io_uring/register.c
555
memset(&rd, 0, sizeof(rd));
io_uring/register.c
556
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/register.c
558
rd.user_addr = p->sq_off.user_addr;
io_uring/register.c
559
rd.flags |= IORING_MEM_REGION_TYPE_USER;
io_uring/register.c
561
ret = io_create_region(ctx, &n.sq_region, &rd, IORING_OFF_SQES);
io_uring/register.c
677
struct io_uring_region_desc rd;
io_uring/register.c
686
if (copy_from_user(&rd, rd_uptr, sizeof(rd)))
io_uring/register.c
702
ret = io_create_region(ctx, &region, &rd, IORING_MAP_OFF_PARAM_REGION);
io_uring/register.c
705
if (copy_to_user(rd_uptr, &rd, sizeof(rd))) {
io_uring/register.c
712
ctx->cq_wait_size = rd.size;
io_uring/zcrx.c
378
struct io_uring_region_desc *rd,
io_uring/zcrx.c
389
if (size > rd->size)
io_uring/zcrx.c
395
ret = io_create_region(ctx, &ifq->region, rd, mmap_offset);
io_uring/zcrx.c
760
struct io_uring_region_desc rd;
io_uring/zcrx.c
783
if (copy_from_user(&rd, u64_to_user_ptr(reg.region_ptr), sizeof(rd)))
io_uring/zcrx.c
818
ret = io_allocate_rbuf_ring(ctx, ifq, &reg, &rd, id);
io_uring/zcrx.c
862
copy_to_user(u64_to_user_ptr(reg.region_ptr), &rd, sizeof(rd)) ||
kernel/sched/core.c
8235
cpumask_set_cpu(rq->cpu, rq->rd->online);
kernel/sched/core.c
8256
cpumask_clear_cpu(rq->cpu, rq->rd->online);
kernel/sched/core.c
8266
if (rq->rd) {
kernel/sched/core.c
8267
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
kernel/sched/core.c
8278
if (rq->rd) {
kernel/sched/core.c
8279
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
kernel/sched/core.c
8712
rq->rd = NULL;
kernel/sched/deadline.c
122
return &cpu_rq(i)->rd->dl_bw;
kernel/sched/deadline.c
127
struct root_domain *rd = cpu_rq(i)->rd;
kernel/sched/deadline.c
132
return cpumask_weight_and(rd->span, cpu_active_mask);
kernel/sched/deadline.c
159
return __dl_bw_capacity(cpu_rq(i)->rd->span);
kernel/sched/deadline.c
165
struct root_domain *rd = cpu_rq(cpu)->rd;
kernel/sched/deadline.c
167
if (rd->visit_cookie == cookie)
kernel/sched/deadline.c
170
rd->visit_cookie = cookie;
kernel/sched/deadline.c
177
struct root_domain *rd = container_of(dl_b, struct root_domain, dl_bw);
kernel/sched/deadline.c
182
for_each_cpu_and(i, rd->span, cpu_active_mask) {
kernel/sched/deadline.c
2035
cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_HIGHER);
kernel/sched/deadline.c
2037
cpudl_set(&rq->rd->cpudl, rq->cpu, deadline);
kernel/sched/deadline.c
2052
cpudl_clear(&rq->rd->cpudl, rq->cpu, rq->online);
kernel/sched/deadline.c
2053
cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio.curr);
kernel/sched/deadline.c
2059
cpudl_set(&rq->rd->cpudl, rq->cpu, entry->deadline);
kernel/sched/deadline.c
2494
!cpudl_find(&rq->rd->cpudl, rq->donor, NULL))
kernel/sched/deadline.c
2502
cpudl_find(&rq->rd->cpudl, p, NULL))
kernel/sched/deadline.c
2733
if (!cpudl_find(&task_rq(task)->rd->cpudl, task, later_mask))
kernel/sched/deadline.c
3003
for_each_cpu(cpu, this_rq->rd->dlo_mask) {
kernel/sched/deadline.c
3101
src_rd = rq->rd;
kernel/sched/deadline.c
3132
cpudl_set(&rq->rd->cpudl, rq->cpu, rq->dl.earliest_dl.curr);
kernel/sched/deadline.c
3134
cpudl_clear(&rq->rd->cpudl, rq->cpu, true);
kernel/sched/deadline.c
3143
cpudl_clear(&rq->rd->cpudl, rq->cpu, false);
kernel/sched/deadline.c
3204
dl_b = &rq->rd->dl_bw;
kernel/sched/deadline.c
3207
__dl_add(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span));
kernel/sched/deadline.c
3212
static void dl_server_add_bw(struct root_domain *rd, int cpu)
kernel/sched/deadline.c
3218
__dl_add(&rd->dl_bw, dl_se->dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3223
__dl_add(&rd->dl_bw, dl_se->dl_bw, dl_bw_cpus(cpu));
kernel/sched/deadline.c
3242
void dl_clear_root_domain(struct root_domain *rd)
kernel/sched/deadline.c
3246
guard(raw_spinlock_irqsave)(&rd->dl_bw.lock);
kernel/sched/deadline.c
3252
rd->dl_bw.total_bw = 0;
kernel/sched/deadline.c
3253
for_each_cpu(i, rd->span)
kernel/sched/deadline.c
3260
for_each_cpu(i, rd->span)
kernel/sched/deadline.c
3261
dl_server_add_bw(rd, i);
kernel/sched/deadline.c
3266
dl_clear_root_domain(cpu_rq(cpu)->rd);
kernel/sched/deadline.c
533
return atomic_read(&rq->rd->dlo_count);
kernel/sched/deadline.c
541
cpumask_set_cpu(rq->cpu, rq->rd->dlo_mask);
kernel/sched/deadline.c
549
atomic_inc(&rq->rd->dlo_count);
kernel/sched/deadline.c
557
atomic_dec(&rq->rd->dlo_count);
kernel/sched/deadline.c
558
cpumask_clear_cpu(rq->cpu, rq->rd->dlo_mask);
kernel/sched/deadline.c
699
dl_b = &rq->rd->dl_bw;
kernel/sched/deadline.c
701
__dl_sub(dl_b, p->dl.dl_bw, cpumask_weight(rq->rd->span));
kernel/sched/deadline.c
704
dl_b = &later_rq->rd->dl_bw;
kernel/sched/deadline.c
706
__dl_add(dl_b, p->dl.dl_bw, cpumask_weight(later_rq->rd->span));
kernel/sched/debug.c
1022
dl_bw = &cpu_rq(cpu)->rd->dl_bw;
kernel/sched/fair.c
11190
set_rd_overloaded(env->dst_rq->rd, sg_overloaded);
kernel/sched/fair.c
11193
set_rd_overutilized(env->dst_rq->rd, sg_overutilized);
kernel/sched/fair.c
11195
set_rd_overutilized(env->dst_rq->rd, sg_overutilized);
kernel/sched/fair.c
11432
if (!is_rd_overutilized(env->dst_rq->rd) &&
kernel/sched/fair.c
11433
rcu_dereference_all(env->dst_rq->rd->pd))
kernel/sched/fair.c
12966
if (!get_rd_overloaded(this_rq->rd) ||
kernel/sched/fair.c
6867
static inline bool is_rd_overutilized(struct root_domain *rd)
kernel/sched/fair.c
6869
return !sched_energy_enabled() || READ_ONCE(rd->overutilized);
kernel/sched/fair.c
6872
static inline void set_rd_overutilized(struct root_domain *rd, bool flag)
kernel/sched/fair.c
6877
WRITE_ONCE(rd->overutilized, flag);
kernel/sched/fair.c
6878
trace_sched_overutilized_tp(rd, flag);
kernel/sched/fair.c
6888
if (!is_rd_overutilized(rq->rd) && cpu_overutilized(rq->cpu))
kernel/sched/fair.c
6889
set_rd_overutilized(rq->rd, 1);
kernel/sched/fair.c
8390
struct root_domain *rd = this_rq()->rd;
kernel/sched/fair.c
8400
pd = rcu_dereference_all(rd->pd);
kernel/sched/fair.c
8601
if (!is_rd_overutilized(this_rq()->rd)) {
kernel/sched/rt.c
1061
cpupri_set(&rq->rd->cpupri, rq->cpu, prio);
kernel/sched/rt.c
1076
cpupri_set(&rq->rd->cpupri, rq->cpu, rt_rq->highest_prio.curr);
kernel/sched/rt.c
1574
!cpupri_find(&rq->rd->cpupri, rq->donor, NULL))
kernel/sched/rt.c
1582
cpupri_find(&rq->rd->cpupri, p, NULL))
kernel/sched/rt.c
1784
ret = cpupri_find_fitness(&task_rq(task)->rd->cpupri,
kernel/sched/rt.c
1789
ret = cpupri_find(&task_rq(task)->rd->cpupri,
kernel/sched/rt.c
2107
static int rto_next_cpu(struct root_domain *rd)
kernel/sched/rt.c
2129
cpu = cpumask_next(rd->rto_cpu, rd->rto_mask);
kernel/sched/rt.c
2131
rd->rto_cpu = cpu;
kernel/sched/rt.c
2140
rd->rto_cpu = -1;
kernel/sched/rt.c
2148
next = atomic_read_acquire(&rd->rto_loop_next);
kernel/sched/rt.c
2150
if (rd->rto_loop == next)
kernel/sched/rt.c
2153
rd->rto_loop = next;
kernel/sched/rt.c
2174
atomic_inc(&rq->rd->rto_loop_next);
kernel/sched/rt.c
2177
if (!rto_start_trylock(&rq->rd->rto_loop_start))
kernel/sched/rt.c
2180
raw_spin_lock(&rq->rd->rto_lock);
kernel/sched/rt.c
2188
if (rq->rd->rto_cpu < 0)
kernel/sched/rt.c
2189
cpu = rto_next_cpu(rq->rd);
kernel/sched/rt.c
2191
raw_spin_unlock(&rq->rd->rto_lock);
kernel/sched/rt.c
2193
rto_start_unlock(&rq->rd->rto_loop_start);
kernel/sched/rt.c
2197
sched_get_rd(rq->rd);
kernel/sched/rt.c
2198
irq_work_queue_on(&rq->rd->rto_push_work, cpu);
kernel/sched/rt.c
2205
struct root_domain *rd =
kernel/sched/rt.c
2223
raw_spin_lock(&rd->rto_lock);
kernel/sched/rt.c
2226
cpu = rto_next_cpu(rd);
kernel/sched/rt.c
2228
raw_spin_unlock(&rd->rto_lock);
kernel/sched/rt.c
2231
sched_put_rd(rd);
kernel/sched/rt.c
2236
irq_work_queue_on(&rd->rto_push_work, cpu);
kernel/sched/rt.c
2259
cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask))
kernel/sched/rt.c
2269
for_each_cpu(cpu, this_rq->rd->rto_mask) {
kernel/sched/rt.c
2374
cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio.curr);
kernel/sched/rt.c
2385
cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_INVALID);
kernel/sched/rt.c
341
return atomic_read(&rq->rd->rto_count);
kernel/sched/rt.c
349
cpumask_set_cpu(rq->cpu, rq->rd->rto_mask);
kernel/sched/rt.c
360
atomic_inc(&rq->rd->rto_count);
kernel/sched/rt.c
369
atomic_dec(&rq->rd->rto_count);
kernel/sched/rt.c
370
cpumask_clear_cpu(rq->cpu, rq->rd->rto_mask);
kernel/sched/rt.c
583
return this_rq()->rd->span;
kernel/sched/rt.c
611
struct root_domain *rd = rq_of_rt_rq(rt_rq)->rd;
kernel/sched/rt.c
615
weight = cpumask_weight(rd->span);
kernel/sched/rt.c
619
for_each_cpu(i, rd->span) {
kernel/sched/rt.c
662
struct root_domain *rd = rq->rd;
kernel/sched/rt.c
696
for_each_cpu(i, rd->span) {
kernel/sched/sched.h
1054
extern void rq_attach_root(struct rq *rq, struct root_domain *rd);
kernel/sched/sched.h
1055
extern void sched_get_rd(struct root_domain *rd);
kernel/sched/sched.h
1056
extern void sched_put_rd(struct root_domain *rd);
kernel/sched/sched.h
1058
static inline int get_rd_overloaded(struct root_domain *rd)
kernel/sched/sched.h
1060
return READ_ONCE(rd->overloaded);
kernel/sched/sched.h
1063
static inline void set_rd_overloaded(struct root_domain *rd, int status)
kernel/sched/sched.h
1065
if (get_rd_overloaded(rd) != status)
kernel/sched/sched.h
1066
WRITE_ONCE(rd->overloaded, status);
kernel/sched/sched.h
1233
struct root_domain *rd;
kernel/sched/sched.h
2943
set_rd_overloaded(rq->rd, 1);
kernel/sched/syscalls.c
1124
if (!cpumask_subset(task_rq(p)->rd->span, mask))
kernel/sched/syscalls.c
624
cpumask_t *span = rq->rd->span;
kernel/sched/syscalls.c
632
rq->rd->dl_bw.bw == 0) {
kernel/sched/topology.c
1542
if (!atomic_read(&d->rd->refcount))
kernel/sched/topology.c
1543
free_rootdomain(&d->rd->rcu);
kernel/sched/topology.c
1566
d->rd = alloc_rootdomain();
kernel/sched/topology.c
1567
if (!d->rd)
kernel/sched/topology.c
2683
cpu_attach_domain(sd, d.rd, i);
kernel/sched/topology.c
2912
cpu_rq(cpumask_first(doms_cur[j]))->rd->pd) {
kernel/sched/topology.c
414
struct root_domain *rd = cpu_rq(cpu)->rd;
kernel/sched/topology.c
438
tmp = rd->pd;
kernel/sched/topology.c
439
rcu_assign_pointer(rd->pd, pd);
kernel/sched/topology.c
447
tmp = rd->pd;
kernel/sched/topology.c
448
rcu_assign_pointer(rd->pd, NULL);
kernel/sched/topology.c
460
struct root_domain *rd = container_of(rcu, struct root_domain, rcu);
kernel/sched/topology.c
462
cpupri_cleanup(&rd->cpupri);
kernel/sched/topology.c
463
cpudl_cleanup(&rd->cpudl);
kernel/sched/topology.c
464
free_cpumask_var(rd->dlo_mask);
kernel/sched/topology.c
465
free_cpumask_var(rd->rto_mask);
kernel/sched/topology.c
466
free_cpumask_var(rd->online);
kernel/sched/topology.c
467
free_cpumask_var(rd->span);
kernel/sched/topology.c
468
free_pd(rd->pd);
kernel/sched/topology.c
469
kfree(rd);
kernel/sched/topology.c
472
void rq_attach_root(struct rq *rq, struct root_domain *rd)
kernel/sched/topology.c
479
if (rq->rd) {
kernel/sched/topology.c
480
old_rd = rq->rd;
kernel/sched/topology.c
496
atomic_inc(&rd->refcount);
kernel/sched/topology.c
497
rq->rd = rd;
kernel/sched/topology.c
499
cpumask_set_cpu(rq->cpu, rd->span);
kernel/sched/topology.c
522
void sched_get_rd(struct root_domain *rd)
kernel/sched/topology.c
524
atomic_inc(&rd->refcount);
kernel/sched/topology.c
527
void sched_put_rd(struct root_domain *rd)
kernel/sched/topology.c
529
if (!atomic_dec_and_test(&rd->refcount))
kernel/sched/topology.c
532
call_rcu(&rd->rcu, free_rootdomain);
kernel/sched/topology.c
535
static int init_rootdomain(struct root_domain *rd)
kernel/sched/topology.c
537
if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL))
kernel/sched/topology.c
539
if (!zalloc_cpumask_var(&rd->online, GFP_KERNEL))
kernel/sched/topology.c
541
if (!zalloc_cpumask_var(&rd->dlo_mask, GFP_KERNEL))
kernel/sched/topology.c
543
if (!zalloc_cpumask_var(&rd->rto_mask, GFP_KERNEL))
kernel/sched/topology.c
547
rd->rto_cpu = -1;
kernel/sched/topology.c
548
raw_spin_lock_init(&rd->rto_lock);
kernel/sched/topology.c
549
rd->rto_push_work = IRQ_WORK_INIT_HARD(rto_push_irq_work_func);
kernel/sched/topology.c
552
rd->visit_cookie = 0;
kernel/sched/topology.c
553
init_dl_bw(&rd->dl_bw);
kernel/sched/topology.c
554
if (cpudl_init(&rd->cpudl) != 0)
kernel/sched/topology.c
557
if (cpupri_init(&rd->cpupri) != 0)
kernel/sched/topology.c
562
cpudl_cleanup(&rd->cpudl);
kernel/sched/topology.c
564
free_cpumask_var(rd->rto_mask);
kernel/sched/topology.c
566
free_cpumask_var(rd->dlo_mask);
kernel/sched/topology.c
568
free_cpumask_var(rd->online);
kernel/sched/topology.c
570
free_cpumask_var(rd->span);
kernel/sched/topology.c
590
struct root_domain *rd;
kernel/sched/topology.c
592
rd = kzalloc_obj(*rd);
kernel/sched/topology.c
593
if (!rd)
kernel/sched/topology.c
596
if (init_rootdomain(rd) != 0) {
kernel/sched/topology.c
597
kfree(rd);
kernel/sched/topology.c
601
return rd;
kernel/sched/topology.c
721
cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu)
kernel/sched/topology.c
774
rq_attach_root(rq, rd);
kernel/sched/topology.c
785
struct root_domain *rd;
kernel/time/sched_clock.c
132
static void update_clock_read_data(struct clock_read_data *rd)
kernel/time/sched_clock.c
138
cd.read_data[0] = *rd;
kernel/time/sched_clock.c
144
cd.read_data[1] = *rd;
kernel/time/sched_clock.c
156
struct clock_read_data rd;
kernel/time/sched_clock.c
158
rd = cd.read_data[0];
kernel/time/sched_clock.c
161
ns = rd.epoch_ns + cyc_to_ns((cyc - rd.epoch_cyc) & rd.sched_clock_mask, rd.mult, rd.shift);
kernel/time/sched_clock.c
163
rd.epoch_ns = ns;
kernel/time/sched_clock.c
164
rd.epoch_cyc = cyc;
kernel/time/sched_clock.c
166
update_clock_read_data(&rd);
kernel/time/sched_clock.c
183
struct clock_read_data rd;
kernel/time/sched_clock.c
201
rd = cd.read_data[0];
kernel/time/sched_clock.c
206
ns = rd.epoch_ns + cyc_to_ns((cyc - rd.epoch_cyc) & rd.sched_clock_mask, rd.mult, rd.shift);
kernel/time/sched_clock.c
209
rd.read_sched_clock = read;
kernel/time/sched_clock.c
210
rd.sched_clock_mask = new_mask;
kernel/time/sched_clock.c
211
rd.mult = new_mult;
kernel/time/sched_clock.c
212
rd.shift = new_shift;
kernel/time/sched_clock.c
213
rd.epoch_cyc = new_epoch;
kernel/time/sched_clock.c
214
rd.epoch_ns = ns;
kernel/time/sched_clock.c
216
update_clock_read_data(&rd);
kernel/time/sched_clock.c
290
struct clock_read_data *rd = &cd.read_data[0];
kernel/time/sched_clock.c
294
rd->read_sched_clock = suspended_sched_clock_read;
kernel/time/sched_clock.c
306
struct clock_read_data *rd = &cd.read_data[0];
kernel/time/sched_clock.c
308
rd->epoch_cyc = cd.actual_read_sched_clock();
kernel/time/sched_clock.c
310
rd->read_sched_clock = cd.actual_read_sched_clock;
kernel/time/sched_clock.c
85
struct clock_read_data *rd;
kernel/time/sched_clock.c
91
rd = cd.read_data + (seq & 1);
kernel/time/sched_clock.c
93
cyc = (rd->read_sched_clock() - rd->epoch_cyc) &
kernel/time/sched_clock.c
94
rd->sched_clock_mask;
kernel/time/sched_clock.c
95
res = rd->epoch_ns + cyc_to_ns(cyc, rd->mult, rd->shift);
kernel/trace/ring_buffer.c
4956
unsigned int rd;
kernel/trace/ring_buffer.c
4959
rd = atomic_read(&buffer->record_disabled);
kernel/trace/ring_buffer.c
4961
new_rd = rd | RB_BUFFER_OFF;
kernel/trace/ring_buffer.c
4962
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
kernel/trace/ring_buffer.c
4979
unsigned int rd;
kernel/trace/ring_buffer.c
4982
rd = atomic_read(&buffer->record_disabled);
kernel/trace/ring_buffer.c
4984
new_rd = rd & ~RB_BUFFER_OFF;
kernel/trace/ring_buffer.c
4985
} while (!atomic_try_cmpxchg(&buffer->record_disabled, &rd, new_rd));
lib/test_bpf.c
1245
int rd;
lib/test_bpf.c
1267
for (rd = R0; rd <= R9; rd++) {
lib/test_bpf.c
1268
i += __bpf_ld_imm64(&insns[i], rd, dst);
lib/test_bpf.c
1271
insns[i++] = BPF_ALU32_IMM(op, rd, imm);
lib/test_bpf.c
1273
insns[i++] = BPF_ALU64_IMM(op, rd, imm);
lib/test_bpf.c
1275
insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res, 2);
lib/test_bpf.c
1279
insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32);
lib/test_bpf.c
1280
insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, res >> 32, 2);
lib/test_bpf.c
1426
int rd, rs;
lib/test_bpf.c
1454
for (rd = R0; rd <= R9; rd++) {
lib/test_bpf.c
1456
u64 val = rd == rs ? same : res;
lib/test_bpf.c
1458
i += __bpf_ld_imm64(&insns[i], rd, dst);
lib/test_bpf.c
1462
insns[i++] = BPF_ALU32_REG(op, rd, rs);
lib/test_bpf.c
1464
insns[i++] = BPF_ALU64_REG(op, rd, rs);
lib/test_bpf.c
1466
insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val, 2);
lib/test_bpf.c
1470
insns[i++] = BPF_ALU64_IMM(BPF_RSH, rd, 32);
lib/test_bpf.c
1471
insns[i++] = BPF_JMP32_IMM(BPF_JEQ, rd, val >> 32, 2);
lib/test_bpf.c
1918
int rd, rs, i = 0;
lib/test_bpf.c
1946
for (rd = R0; rd <= R9; rd++) {
lib/test_bpf.c
1957
insn[i++] = BPF_MOV64_REG(rd, R10);
lib/test_bpf.c
1960
insn[i++] = BPF_ATOMIC_OP(width, op, rd, rs, -8);
lib/test_bpf.c
1967
else if (R0 == rd || R0 == rs)
lib/test_bpf.c
1986
else if (rs == rd && (op == BPF_CMPXCHG ||
lib/test_bpf.c
2008
if (!(rd == R0 && op == BPF_CMPXCHG) &&
lib/test_bpf.c
2009
!(rd == rs && (op & BPF_FETCH))) {
lib/test_bpf.c
2010
insn[i++] = BPF_JMP_REG(BPF_JEQ, rd, R10, 2);
lib/test_bpf.c
2016
if (rs != rd) { /* No aliasing */
net/9p/trans_fd.c
108
struct file *rd;
net/9p/trans_fd.c
202
ret = vfs_poll(ts->rd, pt);
net/9p/trans_fd.c
203
if (ts->rd != ts->wr)
net/9p/trans_fd.c
228
if (!(ts->rd->f_flags & O_NONBLOCK))
net/9p/trans_fd.c
231
pos = ts->rd->f_pos;
net/9p/trans_fd.c
232
ret = kernel_read(ts->rd, v, len, &pos);
net/9p/trans_fd.c
726
ts->rd = fget(rfd);
net/9p/trans_fd.c
727
if (!ts->rd)
net/9p/trans_fd.c
729
if (!(ts->rd->f_mode & FMODE_READ))
net/9p/trans_fd.c
739
data_race(ts->rd->f_flags |= O_NONBLOCK);
net/9p/trans_fd.c
755
fput(ts->rd);
net/9p/trans_fd.c
783
p->wr = p->rd = file;
net/9p/trans_fd.c
787
p->rd->f_flags |= O_NONBLOCK;
net/9p/trans_fd.c
842
if (ts->rd)
net/9p/trans_fd.c
843
fput(ts->rd);
net/nfc/digital_technology.c
1132
size -= sizeof(sensf_res->rd);
net/nfc/digital_technology.c
1151
sensf_res->rd[0] = sensf_req->sc1;
net/nfc/digital_technology.c
1152
sensf_res->rd[1] = sensf_req->sc2;
net/nfc/digital_technology.c
1155
sensf_res->rd[0] = DIGITAL_SENSF_RES_RD_AP_B1;
net/nfc/digital_technology.c
1156
sensf_res->rd[1] = DIGITAL_SENSF_RES_RD_AP_B2;
net/nfc/digital_technology.c
148
u8 rd[2];
net/sunrpc/xprtrdma/ib_client.c
109
struct rpcrdma_device *rd;
net/sunrpc/xprtrdma/ib_client.c
111
rd = kzalloc_obj(*rd);
net/sunrpc/xprtrdma/ib_client.c
112
if (!rd)
net/sunrpc/xprtrdma/ib_client.c
115
kref_init(&rd->rd_kref);
net/sunrpc/xprtrdma/ib_client.c
116
xa_init_flags(&rd->rd_xa, XA_FLAGS_ALLOC);
net/sunrpc/xprtrdma/ib_client.c
117
rd->rd_device = device;
net/sunrpc/xprtrdma/ib_client.c
118
init_completion(&rd->rd_done);
net/sunrpc/xprtrdma/ib_client.c
119
ib_set_client_data(device, &rpcrdma_ib_client, rd);
net/sunrpc/xprtrdma/ib_client.c
136
struct rpcrdma_device *rd = client_data;
net/sunrpc/xprtrdma/ib_client.c
142
set_bit(RPCRDMA_RD_F_REMOVING, &rd->rd_flags);
net/sunrpc/xprtrdma/ib_client.c
143
xa_for_each(&rd->rd_xa, index, rn)
net/sunrpc/xprtrdma/ib_client.c
150
if (!refcount_dec_and_test(&rd->rd_kref.refcount)) {
net/sunrpc/xprtrdma/ib_client.c
152
wait_for_completion(&rd->rd_done);
net/sunrpc/xprtrdma/ib_client.c
156
xa_destroy(&rd->rd_xa);
net/sunrpc/xprtrdma/ib_client.c
157
kfree(rd);
net/sunrpc/xprtrdma/ib_client.c
60
struct rpcrdma_device *rd = rpcrdma_get_client_data(device);
net/sunrpc/xprtrdma/ib_client.c
62
if (!rd || test_bit(RPCRDMA_RD_F_REMOVING, &rd->rd_flags))
net/sunrpc/xprtrdma/ib_client.c
65
if (xa_alloc(&rd->rd_xa, &rn->rn_index, rn, xa_limit_32b, GFP_KERNEL) < 0)
net/sunrpc/xprtrdma/ib_client.c
67
kref_get(&rd->rd_kref);
net/sunrpc/xprtrdma/ib_client.c
75
struct rpcrdma_device *rd = container_of(kref, struct rpcrdma_device,
net/sunrpc/xprtrdma/ib_client.c
78
trace_rpcrdma_client_completion(rd->rd_device);
net/sunrpc/xprtrdma/ib_client.c
79
complete(&rd->rd_done);
net/sunrpc/xprtrdma/ib_client.c
90
struct rpcrdma_device *rd = rpcrdma_get_client_data(device);
net/sunrpc/xprtrdma/ib_client.c
92
if (!rd)
net/sunrpc/xprtrdma/ib_client.c
96
xa_erase(&rd->rd_xa, rn->rn_index);
net/sunrpc/xprtrdma/ib_client.c
97
kref_put(&rd->rd_kref, rpcrdma_rn_release);
net/wireless/nl80211.c
10133
struct ieee80211_regdomain *rd;
net/wireless/nl80211.c
10159
rd = kzalloc_flex(*rd, reg_rules, num_rules);
net/wireless/nl80211.c
10160
if (!rd) {
net/wireless/nl80211.c
10165
rd->n_reg_rules = num_rules;
net/wireless/nl80211.c
10166
rd->alpha2[0] = alpha2[0];
net/wireless/nl80211.c
10167
rd->alpha2[1] = alpha2[1];
net/wireless/nl80211.c
10174
rd->dfs_region = dfs_region;
net/wireless/nl80211.c
10183
r = parse_reg_rule(tb, &rd->reg_rules[rule_idx]);
net/wireless/nl80211.c
10195
r = set_regdom(rd, REGD_SOURCE_CRDA);
net/wireless/nl80211.c
10197
rd = NULL;
net/wireless/nl80211.c
10199
kfree(rd);
net/wireless/reg.c
1160
reg_get_max_bandwidth_from_range(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
1168
for (idx = 0; idx < rd->n_reg_rules; idx++)
net/wireless/reg.c
1169
if (rule == &rd->reg_rules[idx])
net/wireless/reg.c
1172
if (idx == rd->n_reg_rules)
net/wireless/reg.c
1179
tmp = &rd->reg_rules[--no];
net/wireless/reg.c
1194
while (no < rd->n_reg_rules - 1) {
net/wireless/reg.c
1195
tmp = &rd->reg_rules[++no];
net/wireless/reg.c
1209
unsigned int reg_get_max_bandwidth(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
1212
unsigned int bw = reg_get_max_bandwidth_from_range(rd, rule);
net/wireless/reg.c
1253
static bool is_valid_rd(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
1258
if (!rd->n_reg_rules)
net/wireless/reg.c
1261
if (WARN_ON(rd->n_reg_rules > NL80211_MAX_SUPP_REG_RULES))
net/wireless/reg.c
1264
for (i = 0; i < rd->n_reg_rules; i++) {
net/wireless/reg.c
1265
reg_rule = &rd->reg_rules[i];
net/wireless/reg.c
137
static void print_regdomain(const struct ieee80211_regdomain *rd);
net/wireless/reg.c
1507
struct ieee80211_regdomain *rd;
net/wireless/reg.c
1533
rd = kzalloc_flex(*rd, reg_rules, num_rules);
net/wireless/reg.c
1534
if (!rd)
net/wireless/reg.c
1550
add_rule(&intersected_rule, rd->reg_rules,
net/wireless/reg.c
1551
&rd->n_reg_rules);
net/wireless/reg.c
1555
rd->alpha2[0] = '9';
net/wireless/reg.c
1556
rd->alpha2[1] = '8';
net/wireless/reg.c
1557
rd->dfs_region = reg_intersect_dfs_region(rd1->dfs_region,
net/wireless/reg.c
1560
return rd;
net/wireless/reg.c
356
static void update_world_regdomain(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
364
reset_regdomains(false, rd);
net/wireless/reg.c
366
cfg80211_world_regdom = rd;
net/wireless/reg.c
3691
static void print_rd_rules(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
3701
for (i = 0; i < rd->n_reg_rules; i++) {
net/wireless/reg.c
3702
reg_rule = &rd->reg_rules[i];
net/wireless/reg.c
3709
reg_get_max_bandwidth(rd, reg_rule));
net/wireless/reg.c
3757
static void print_regdomain(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
3761
if (is_intersected_alpha2(rd->alpha2)) {
net/wireless/reg.c
3773
} else if (is_world_regdom(rd->alpha2)) {
net/wireless/reg.c
3776
if (is_unknown_alpha2(rd->alpha2))
net/wireless/reg.c
3781
rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3784
rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3788
pr_debug(" DFS Master region: %s", reg_dfs_region_str(rd->dfs_region));
net/wireless/reg.c
3789
print_rd_rules(rd);
net/wireless/reg.c
3792
static void print_regdomain_info(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
3794
pr_debug("Regulatory domain: %c%c\n", rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3795
print_rd_rules(rd);
net/wireless/reg.c
3798
static int reg_set_rd_core(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
3800
if (!is_world_regdom(rd->alpha2))
net/wireless/reg.c
3802
update_world_regdomain(rd);
net/wireless/reg.c
3806
static int reg_set_rd_user(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
3811
if (!regdom_changes(rd->alpha2))
net/wireless/reg.c
3814
if (!is_valid_rd(rd)) {
net/wireless/reg.c
3816
rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3817
print_regdomain_info(rd);
net/wireless/reg.c
3822
reset_regdomains(false, rd);
net/wireless/reg.c
3826
intersected_rd = regdom_intersect(rd, get_cfg80211_regdom());
net/wireless/reg.c
3830
kfree(rd);
net/wireless/reg.c
3831
rd = NULL;
net/wireless/reg.c
3837
static int reg_set_rd_driver(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
3845
if (is_world_regdom(rd->alpha2))
net/wireless/reg.c
3848
if (!regdom_changes(rd->alpha2))
net/wireless/reg.c
3851
if (!is_valid_rd(rd)) {
net/wireless/reg.c
3853
rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3854
print_regdomain_info(rd);
net/wireless/reg.c
3868
regd = reg_copy_regd(rd);
net/wireless/reg.c
3876
reset_regdomains(false, rd);
net/wireless/reg.c
3880
intersected_rd = regdom_intersect(rd, get_cfg80211_regdom());
net/wireless/reg.c
3890
rcu_assign_pointer(request_wiphy->regd, rd);
net/wireless/reg.c
3893
rd = NULL;
net/wireless/reg.c
3900
static int reg_set_rd_country_ie(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
3905
if (!is_alpha2_set(rd->alpha2) && !is_an_alpha2(rd->alpha2) &&
net/wireless/reg.c
3906
!is_unknown_alpha2(rd->alpha2))
net/wireless/reg.c
3915
if (!is_valid_rd(rd)) {
net/wireless/reg.c
3917
rd->alpha2[0], rd->alpha2[1]);
net/wireless/reg.c
3918
print_regdomain_info(rd);
net/wireless/reg.c
3929
reset_regdomains(false, rd);
net/wireless/reg.c
3938
int set_regdom(const struct ieee80211_regdomain *rd,
net/wireless/reg.c
3945
if (IS_ERR_OR_NULL(rd))
net/wireless/reg.c
3948
if (!reg_is_valid_request(rd->alpha2)) {
net/wireless/reg.c
3949
kfree(rd);
net/wireless/reg.c
3961
r = reg_set_rd_core(rd);
net/wireless/reg.c
3964
cfg80211_save_user_regdom(rd);
net/wireless/reg.c
3965
r = reg_set_rd_user(rd, lr);
net/wireless/reg.c
3969
r = reg_set_rd_driver(rd, lr);
net/wireless/reg.c
3972
r = reg_set_rd_country_ie(rd, lr);
net/wireless/reg.c
3976
kfree(rd);
net/wireless/reg.c
3990
kfree(rd);
net/wireless/reg.c
3995
if (WARN_ON(!lr->intersect && rd != get_cfg80211_regdom()))
net/wireless/reg.c
4011
struct ieee80211_regdomain *rd)
net/wireless/reg.c
4017
if (WARN_ON(!wiphy || !rd))
net/wireless/reg.c
4024
if (WARN(!is_valid_rd(rd),
net/wireless/reg.c
4026
rd->alpha2[0], rd->alpha2[1])) {
net/wireless/reg.c
4027
print_regdomain_info(rd);
net/wireless/reg.c
4031
regd = reg_copy_regd(rd);
net/wireless/reg.c
4047
struct ieee80211_regdomain *rd)
net/wireless/reg.c
4049
int ret = __regulatory_set_wiphy_regd(wiphy, rd);
net/wireless/reg.c
4060
struct ieee80211_regdomain *rd)
net/wireless/reg.c
4066
ret = __regulatory_set_wiphy_regd(wiphy, rd);
net/wireless/reg.c
468
static void cfg80211_save_user_regdom(const struct ieee80211_regdomain *rd)
net/wireless/reg.c
474
cfg80211_user_regdom = reg_copy_regd(rd);
net/wireless/reg.h
59
int set_regdom(const struct ieee80211_regdomain *rd,
net/wireless/reg.h
62
unsigned int reg_get_max_bandwidth(const struct ieee80211_regdomain *rd,
security/selinux/selinuxfs.c
509
struct renamedata rd = {};
security/selinux/selinuxfs.c
543
rd.old_parent = tmp_parent;
security/selinux/selinuxfs.c
544
rd.new_parent = fsi->sb->s_root;
security/selinux/selinuxfs.c
547
ret = start_renaming_two_dentries(&rd, tmp_bool_dir, fsi->bool_dir);
security/selinux/selinuxfs.c
558
end_renaming(&rd);
security/selinux/selinuxfs.c
561
ret = start_renaming_two_dentries(&rd, tmp_class_dir, fsi->class_dir);
security/selinux/selinuxfs.c
568
end_renaming(&rd);
sound/soc/codecs/tscs42xx.c
950
#define PLL_CTL(f, rt, rd, r1b_l, r9, ra, rb, \
sound/soc/codecs/tscs42xx.c
956
{R_PLLCTLD, rd, 0xFF}, \
sound/soc/codecs/wm_adsp.c
505
unsigned int out, rd, wr, vol;
sound/soc/codecs/wm_adsp.c
508
rd = SNDRV_CTL_ELEM_ACCESS_TLV_READ;
sound/soc/codecs/wm_adsp.c
514
rd = SNDRV_CTL_ELEM_ACCESS_READ;
sound/soc/codecs/wm_adsp.c
522
out |= rd;
sound/soc/codecs/wm_adsp.c
528
out |= rd | wr | vol;
sound/soc/ti/davinci-mcasp.c
1556
struct davinci_mcasp_ruledata *rd = rule->private;
sound/soc/ti/davinci-mcasp.c
1563
slot_width = mcasp_get_slot_width(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1579
struct davinci_mcasp_ruledata *rd = rule->private;
sound/soc/ti/davinci-mcasp.c
1586
format_width = rd->mcasp->max_format_width;
sound/soc/ti/davinci-mcasp.c
1609
struct davinci_mcasp_ruledata *rd = rule->private;
sound/soc/ti/davinci-mcasp.c
1617
slots = mcasp_get_tdm_slots(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1619
slot_width = mcasp_get_slot_width(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1634
ratio = mcasp_get_auxclk_fs_ratio(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1638
sysclk_freq = mcasp_get_sysclk_freq(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1640
ppm = davinci_mcasp_calc_clk_div(rd->mcasp, sysclk_freq,
sound/soc/ti/davinci-mcasp.c
1641
bclk_freq, rd->stream, false);
sound/soc/ti/davinci-mcasp.c
1652
dev_dbg(rd->mcasp->dev,
sound/soc/ti/davinci-mcasp.c
1663
struct davinci_mcasp_ruledata *rd = rule->private;
sound/soc/ti/davinci-mcasp.c
1671
slots = mcasp_get_tdm_slots(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1682
ratio = mcasp_get_auxclk_fs_ratio(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1686
sysclk_freq = mcasp_get_sysclk_freq(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1688
slot_width = mcasp_get_slot_width(rd->mcasp, rd->stream);
sound/soc/ti/davinci-mcasp.c
1692
ppm = davinci_mcasp_calc_clk_div(rd->mcasp, sysclk_freq,
sound/soc/ti/davinci-mcasp.c
1694
rd->stream, false);
sound/soc/ti/davinci-mcasp.c
1701
dev_dbg(rd->mcasp->dev,
sound/soc/uniphier/aio-core.c
19
static u64 rb_cnt(u64 wr, u64 rd, u64 len)
sound/soc/uniphier/aio-core.c
21
if (rd <= wr)
sound/soc/uniphier/aio-core.c
22
return wr - rd;
sound/soc/uniphier/aio-core.c
24
return len - (rd - wr);
sound/soc/uniphier/aio-core.c
27
static u64 rb_cnt_to_end(u64 wr, u64 rd, u64 len)
sound/soc/uniphier/aio-core.c
29
if (rd <= wr)
sound/soc/uniphier/aio-core.c
30
return wr - rd;
sound/soc/uniphier/aio-core.c
32
return len - rd;
sound/soc/uniphier/aio-core.c
35
static u64 rb_space(u64 wr, u64 rd, u64 len)
sound/soc/uniphier/aio-core.c
37
if (rd <= wr)
sound/soc/uniphier/aio-core.c
38
return len - (wr - rd) - 8;
sound/soc/uniphier/aio-core.c
40
return rd - wr - 8;
sound/soc/uniphier/aio-core.c
43
static u64 rb_space_to_end(u64 wr, u64 rd, u64 len)
sound/soc/uniphier/aio-core.c
45
if (rd > wr)
sound/soc/uniphier/aio-core.c
46
return rd - wr - 8;
sound/soc/uniphier/aio-core.c
47
else if (rd > 0)
sound/sparc/dbri.c
1806
int rd = dbri->pipes[pipe].desc;
sound/sparc/dbri.c
1809
if (rd < 0 || rd >= DBRI_NO_DESCS) {
sound/sparc/dbri.c
1814
dbri->pipes[pipe].desc = dbri->next_desc[rd];
sound/sparc/dbri.c
1815
status = dbri->dma->desc[rd].word1;
sound/sparc/dbri.c
1816
dbri->dma->desc[rd].word1 = 0; /* Reset it for next time. */
sound/sparc/dbri.c
1824
rd, DBRI_RD_STATUS(status), DBRI_RD_CNT(status));
tools/arch/loongarch/include/asm/inst.h
104
unsigned int rd : 5;
tools/arch/loongarch/include/asm/inst.h
162
enum loongarch_gpr rd, \
tools/arch/loongarch/include/asm/inst.h
168
insn->reg2i16_format.rd = rd; \
tools/arch/loongarch/include/asm/inst.h
77
unsigned int rd : 5;
tools/arch/loongarch/include/asm/inst.h
83
unsigned int rd : 5;
tools/arch/loongarch/include/asm/inst.h
90
unsigned int rd : 5;
tools/arch/loongarch/include/asm/inst.h
97
unsigned int rd : 5;
tools/gpio/gpio-watch.c
30
ssize_t rd;
tools/gpio/gpio-watch.c
65
rd = read(pfd.fd, &chg, sizeof(chg));
tools/gpio/gpio-watch.c
66
if (rd < 0 || rd != sizeof(chg)) {
tools/gpio/gpio-watch.c
67
if (rd != sizeof(chg))
tools/objtool/arch/loongarch/decode.c
133
if ((inst.reg2i12_format.rd == CFI_SP) || (inst.reg2i12_format.rj == CFI_SP)) {
tools/objtool/arch/loongarch/decode.c
141
op->dest.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
144
if ((inst.reg2i12_format.rd == CFI_SP) && (inst.reg2i12_format.rj == CFI_FP)) {
tools/objtool/arch/loongarch/decode.c
163
op->dest.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
173
op->src.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
181
if (inst.reg2i12_format.rd == 0 &&
tools/objtool/arch/loongarch/decode.c
209
op->dest.reg = inst.reg2i14_format.rd;
tools/objtool/arch/loongarch/decode.c
216
if (inst.reg2i14_format.rd == LOONGARCH_GPR_RA &&
tools/objtool/arch/loongarch/decode.c
224
op->src.reg = inst.reg2i14_format.rd;
tools/objtool/arch/loongarch/decode.c
243
if (inst.reg2i16_format.rd == 0 &&
tools/objtool/arch/loongarch/decode.c
248
} else if (inst.reg2i16_format.rd == CFI_RA) {
tools/objtool/arch/loongarch/decode.c
251
} else if (inst.reg2i16_format.rd == CFI_A0 &&
tools/objtool/arch/loongarch/decode.c
259
} else if (inst.reg2i16_format.rd == 0 &&
tools/objtool/arch/loongarch/decode.c
263
} else if (inst.reg2i16_format.rd == 0 &&
tools/objtool/arch/loongarch/decode.c
298
if (inst.reg3_format.rd == LOONGARCH_GPR_ZERO &&
tools/perf/util/session.c
2282
reader__init(struct reader *rd, bool *one_mmap)
tools/perf/util/session.c
2284
u64 data_size = rd->data_size;
tools/perf/util/session.c
2285
char **mmaps = rd->mmaps;
tools/perf/util/session.c
2287
rd->head = rd->data_offset;
tools/perf/util/session.c
2288
data_size += rd->data_offset;
tools/perf/util/session.c
2290
rd->mmap_size = MMAP_SIZE;
tools/perf/util/session.c
2291
if (rd->mmap_size > data_size) {
tools/perf/util/session.c
2292
rd->mmap_size = data_size;
tools/perf/util/session.c
2297
memset(mmaps, 0, sizeof(rd->mmaps));
tools/perf/util/session.c
2299
if (zstd_init(&rd->zstd_data, 0))
tools/perf/util/session.c
2301
rd->decomp_data.zstd_decomp = &rd->zstd_data;
tools/perf/util/session.c
2307
reader__release_decomp(struct reader *rd)
tools/perf/util/session.c
2309
perf_decomp__release_events(rd->decomp_data.decomp);
tools/perf/util/session.c
2310
zstd_fini(&rd->zstd_data);
tools/perf/util/session.c
2314
reader__mmap(struct reader *rd, struct perf_session *session)
tools/perf/util/session.c
2317
char *buf, **mmaps = rd->mmaps;
tools/perf/util/session.c
2323
if (rd->in_place_update) {
tools/perf/util/session.c
2330
if (mmaps[rd->mmap_idx]) {
tools/perf/util/session.c
2331
munmap(mmaps[rd->mmap_idx], rd->mmap_size);
tools/perf/util/session.c
2332
mmaps[rd->mmap_idx] = NULL;
tools/perf/util/session.c
2335
page_offset = page_size * (rd->head / page_size);
tools/perf/util/session.c
2336
rd->file_offset += page_offset;
tools/perf/util/session.c
2337
rd->head -= page_offset;
tools/perf/util/session.c
2339
buf = mmap(NULL, rd->mmap_size, mmap_prot, mmap_flags, rd->fd,
tools/perf/util/session.c
2340
rd->file_offset);
tools/perf/util/session.c
2345
mmaps[rd->mmap_idx] = rd->mmap_cur = buf;
tools/perf/util/session.c
2346
rd->mmap_idx = (rd->mmap_idx + 1) & (ARRAY_SIZE(rd->mmaps) - 1);
tools/perf/util/session.c
2347
rd->file_pos = rd->file_offset + rd->head;
tools/perf/util/session.c
2350
session->one_mmap_offset = rd->file_offset;
tools/perf/util/session.c
2362
reader__read_event(struct reader *rd, struct perf_session *session,
tools/perf/util/session.c
2370
event = fetch_mmaped_event(rd->head, rd->mmap_size, rd->mmap_cur,
tools/perf/util/session.c
2383
(skip = rd->process(session, event, rd->file_pos, rd->path)) < 0) {
tools/perf/util/session.c
2386
rd->file_offset + rd->head, event->header.size,
tools/perf/util/session.c
2395
rd->size += size;
tools/perf/util/session.c
2396
rd->head += size;
tools/perf/util/session.c
2397
rd->file_pos += size;
tools/perf/util/session.c
2410
reader__eof(struct reader *rd)
tools/perf/util/session.c
2412
return (rd->file_pos >= rd->data_size + rd->data_offset);
tools/perf/util/session.c
2416
reader__process_events(struct reader *rd, struct perf_session *session,
tools/perf/util/session.c
2421
err = reader__init(rd, &session->one_mmap);
tools/perf/util/session.c
2425
session->active_decomp = &rd->decomp_data;
tools/perf/util/session.c
2428
err = reader__mmap(rd, session);
tools/perf/util/session.c
2433
err = reader__read_event(rd, session, prog);
tools/perf/util/session.c
2442
if (!reader__eof(rd))
tools/perf/util/session.c
2460
struct reader rd = {
tools/perf/util/session.c
2473
if (rd.data_size == 0)
tools/perf/util/session.c
2476
ui_progress__init_size(&prog, rd.data_size, "Processing events...");
tools/perf/util/session.c
2478
err = reader__process_events(&rd, session, &prog);
tools/perf/util/session.c
2502
reader__release_decomp(&rd);
tools/perf/util/session.c
2526
struct reader *rd;
tools/perf/util/session.c
2536
rd = zalloc(nr_readers * sizeof(struct reader));
tools/perf/util/session.c
2537
if (!rd)
tools/perf/util/session.c
2540
rd[0] = (struct reader) {
tools/perf/util/session.c
2548
ret = reader__init(&rd[0], NULL);
tools/perf/util/session.c
2551
ret = reader__mmap(&rd[0], session);
tools/perf/util/session.c
2559
rd[readers] = (struct reader) {
tools/perf/util/session.c
2567
ret = reader__init(&rd[readers], NULL);
tools/perf/util/session.c
2570
ret = reader__mmap(&rd[readers], session);
tools/perf/util/session.c
2581
if (rd[i].done) {
tools/perf/util/session.c
2585
if (reader__eof(&rd[i])) {
tools/perf/util/session.c
2586
rd[i].done = true;
tools/perf/util/session.c
2591
session->active_decomp = &rd[i].decomp_data;
tools/perf/util/session.c
2592
ret = reader__read_event(&rd[i], session, &prog);
tools/perf/util/session.c
2596
ret = reader__mmap(&rd[i], session);
tools/perf/util/session.c
2601
if (rd[i].size >= READER_MAX_SIZE) {
tools/perf/util/session.c
2602
rd[i].size = 0;
tools/perf/util/session.c
2632
reader__release_decomp(&rd[i]);
tools/perf/util/session.c
2633
zfree(&rd);
tools/perf/util/stat.c
697
struct perf_record_stat_round *rd = (struct perf_record_stat_round *)event;
tools/perf/util/stat.c
700
ret = fprintf(fp, "\n... time %" PRI_lu64 ", type %s\n", rd->time,
tools/perf/util/stat.c
701
rd->type == PERF_STAT_ROUND_TYPE__FINAL ? "FINAL" : "INTERVAL");
tools/testing/selftests/bpf/prog_tests/module_attach.c
15
ssize_t rd;
tools/testing/selftests/bpf/prog_tests/module_attach.c
22
rd = read(fd, buf, sizeof(buf) - 1);
tools/testing/selftests/bpf/prog_tests/module_attach.c
24
if (!ASSERT_GT(rd, 0, "testmod_file_rd_val")) {
tools/testing/selftests/bpf/prog_tests/module_attach.c
29
buf[rd] = '\0';
tools/testing/selftests/net/psock_tpacket.c
239
while (__v1_v2_rx_kernel_ready(ring->rd[frame_num].iov_base,
tools/testing/selftests/net/psock_tpacket.c
241
ppd.raw = ring->rd[frame_num].iov_base;
tools/testing/selftests/net/psock_tpacket.c
356
uint8_t *f0 = ring->rd[0].iov_base;
tools/testing/selftests/net/psock_tpacket.c
361
return ring->rd[n].iov_base;
tools/testing/selftests/net/psock_tpacket.c
600
pbd = (struct block_desc *) ring->rd[block_num].iov_base;
tools/testing/selftests/net/psock_tpacket.c
67
struct iovec *rd;
tools/testing/selftests/net/psock_tpacket.c
697
ring->rd_len = ring->rd_num * sizeof(*ring->rd);
tools/testing/selftests/net/psock_tpacket.c
698
ring->rd = malloc(ring->rd_len);
tools/testing/selftests/net/psock_tpacket.c
699
if (ring->rd == NULL) {
tools/testing/selftests/net/psock_tpacket.c
719
memset(ring->rd, 0, ring->rd_len);
tools/testing/selftests/net/psock_tpacket.c
721
ring->rd[i].iov_base = ring->mm_space + (i * ring->flen);
tools/testing/selftests/net/psock_tpacket.c
722
ring->rd[i].iov_len = ring->flen;
tools/testing/selftests/net/psock_tpacket.c
754
free(ring->rd);
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
540
#define toreal(rd)
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
541
#define fromreal(rd)
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
549
#define tophys(rd,rs) \
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
550
addis rd,rs,0
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
552
#define tovirt(rd,rs) \
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
553
addis rd,rs,0
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
556
#define toreal(rd) /* we can access c000... in real mode */
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
557
#define fromreal(rd)
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
559
#define tophys(rd,rs) \
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
560
clrldi rd,rs,2
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
562
#define tovirt(rd,rs) \
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
563
rotldi rd,rs,16; \
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
564
ori rd,rd,((KERNELBASE>>48)&0xFFFF);\
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
565
rotldi rd,rd,48
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
567
#define toreal(rd) tophys(rd,rd)
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
568
#define fromreal(rd) tovirt(rd,rd)
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
570
#define tophys(rd, rs) addis rd, rs, -PAGE_OFFSET@h
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
571
#define tovirt(rd, rs) addis rd, rs, PAGE_OFFSET@h