Symbol: rm
arch/arc/net/bpf_jit_arcv2.c
1443
u8 rm, u8 *arc_reg_mem)
arch/arc/net/bpf_jit_arcv2.c
1446
*arc_reg_mem = REG_LO(rm);
arch/arc/net/bpf_jit_arcv2.c
1451
REG_LO(JIT_REG_TMP), REG_LO(rm), (u32)(*off));
arch/arm/net/bpf_jit_32.c
1266
s8 rm = src;
arch/arm/net/bpf_jit_32.c
1271
rm = tmp[0];
arch/arm/net/bpf_jit_32.c
1273
} else if (rd[1] == rm) {
arch/arm/net/bpf_jit_32.c
1274
emit(ARM_MOV_R(tmp[0], rm), ctx);
arch/arm/net/bpf_jit_32.c
1275
rm = tmp[0];
arch/arm/net/bpf_jit_32.c
1280
emit(ARM_LDRB_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1286
emit(ARM_LDRH_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1292
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1298
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1299
emit(ARM_LDR_I(rd[0], rm, off + 4), ctx);
arch/arm/net/bpf_jit_32.c
1310
s8 rm = src;
arch/arm/net/bpf_jit_32.c
1321
rm = tmp[0];
arch/arm/net/bpf_jit_32.c
1325
rm = tmp[0];
arch/arm/net/bpf_jit_32.c
1333
emit(ARM_LDRSB_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1337
emit(ARM_LDRSH_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1341
emit(ARM_LDR_I(rd[1], rm, off), ctx);
arch/arm/net/bpf_jit_32.c
1350
static inline void emit_ar_r(const u8 rd, const u8 rt, const u8 rm,
arch/arm/net/bpf_jit_32.c
1357
emit(ARM_AND_R(ARM_LR, rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1370
emit(ARM_CMP_R(rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1381
emit(ARM_SBCS_R(ARM_IP, rm, rd), ctx);
arch/arm/net/bpf_jit_32.c
1387
emit(ARM_SBCS_R(ARM_IP, rd, rm), ctx);
arch/arm/net/bpf_jit_32.c
1601
s8 rd_lo, rt, rm, rn;
arch/arm/net/bpf_jit_32.c
1941
rm = arm_bpf_get_reg32(src_hi, tmp2[0], ctx);
arch/arm/net/bpf_jit_32.c
1979
rm = tmp2[0];
arch/arm/net/bpf_jit_32.c
1988
emit_ar_r(rd[0], rd[1], rm, rn, ctx, BPF_OP(code),
arch/arm/net/bpf_jit_32.c
520
static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op, u8 sign)
arch/arm/net/bpf_jit_32.c
529
emit(sign ? ARM_SDIV(rd, rm, rn) : ARM_UDIV(rd, rm, rn), ctx);
arch/arm/net/bpf_jit_32.c
531
emit(sign ? ARM_SDIV(ARM_IP, rm, rn) : ARM_UDIV(ARM_IP, rm, rn), ctx);
arch/arm/net/bpf_jit_32.c
532
emit(ARM_MLS(rd, rn, ARM_IP, rm), ctx);
arch/arm/net/bpf_jit_32.c
550
if (rm != ARM_R0) {
arch/arm/net/bpf_jit_32.c
552
emit(ARM_MOV_R(ARM_R0, rm), ctx);
arch/arm/net/bpf_jit_32.c
584
if (rm != ARM_R0)
arch/arm/net/bpf_jit_32.c
588
static inline void emit_udivmod64(const s8 *rd, const s8 *rm, const s8 *rn, struct jit_ctx *ctx,
arch/arm/net/bpf_jit_32.c
601
if (rm[1] != ARM_R0 || rn[1] != ARM_R2) {
arch/arm/net/bpf_jit_32.c
605
if (rm[1] != ARM_R0) {
arch/arm/net/bpf_jit_32.c
608
emit(ARM_MOV_R(ARM_R1, rm[0]), ctx);
arch/arm/net/bpf_jit_32.c
609
emit(ARM_MOV_R(ARM_R0, rm[1]), ctx);
arch/arm/net/bpf_jit_32.h
165
#define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm))
arch/arm/net/bpf_jit_32.h
171
#define ARM_ADD_R(rd, rn, rm) _AL3_R(ARM_INST_ADD, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
172
#define ARM_ADDS_R(rd, rn, rm) _AL3_R(ARM_INST_ADDS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
175
#define ARM_ADC_R(rd, rn, rm) _AL3_R(ARM_INST_ADC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
178
#define ARM_AND_R(rd, rn, rm) _AL3_R(ARM_INST_AND, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
179
#define ARM_ANDS_R(rd, rn, rm) _AL3_R(ARM_INST_ANDS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
182
#define ARM_BIC_R(rd, rn, rm) _AL3_R(ARM_INST_BIC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
186
#define ARM_BX(rm) (ARM_INST_BX | (rm))
arch/arm/net/bpf_jit_32.h
187
#define ARM_BLX_R(rm) (ARM_INST_BLX_R | (rm))
arch/arm/net/bpf_jit_32.h
189
#define ARM_CMP_R(rn, rm) _AL3_R(ARM_INST_CMP, 0, rn, rm)
arch/arm/net/bpf_jit_32.h
192
#define ARM_EOR_R(rd, rn, rm) _AL3_R(ARM_INST_EOR, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
195
#define ARM_LDR_R(rt, rn, rm) (ARM_INST_LDR_R | ARM_INST_LDST__U \
arch/arm/net/bpf_jit_32.h
197
| (rm))
arch/arm/net/bpf_jit_32.h
198
#define ARM_LDR_R_SI(rt, rn, rm, type, imm) \
arch/arm/net/bpf_jit_32.h
201
| (imm) << 7 | (type) << 5 | (rm))
arch/arm/net/bpf_jit_32.h
202
#define ARM_LDRB_R(rt, rn, rm) (ARM_INST_LDRB_R | ARM_INST_LDST__U \
arch/arm/net/bpf_jit_32.h
204
| (rm))
arch/arm/net/bpf_jit_32.h
205
#define ARM_LDRH_R(rt, rn, rm) (ARM_INST_LDRH_R | ARM_INST_LDST__U \
arch/arm/net/bpf_jit_32.h
207
| (rm))
arch/arm/net/bpf_jit_32.h
212
#define ARM_LSL_R(rd, rn, rm) (_AL3_R(ARM_INST_LSL, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
215
#define ARM_LSR_R(rd, rn, rm) (_AL3_R(ARM_INST_LSR, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
217
#define ARM_ASR_R(rd, rn, rm) (_AL3_R(ARM_INST_ASR, rd, 0, rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
220
#define ARM_MOV_R(rd, rm) _AL3_R(ARM_INST_MOV, rd, 0, rm)
arch/arm/net/bpf_jit_32.h
221
#define ARM_MOVS_R(rd, rm) _AL3_R(ARM_INST_MOVS, rd, 0, rm)
arch/arm/net/bpf_jit_32.h
223
#define ARM_MOV_SR(rd, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
224
(_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
225
#define ARM_MOV_SI(rd, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
226
(ARM_MOV_R(rd, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
234
#define ARM_MUL(rd, rm, rn) (ARM_INST_MUL | (rd) << 16 | (rm) << 8 | (rn))
arch/arm/net/bpf_jit_32.h
239
#define ARM_ORR_R(rd, rn, rm) _AL3_R(ARM_INST_ORR, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
241
#define ARM_ORR_SR(rd, rn, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
242
(_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
243
#define ARM_ORRS_R(rd, rn, rm) _AL3_R(ARM_INST_ORRS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
244
#define ARM_ORRS_SR(rd, rn, rm, type, rs) \
arch/arm/net/bpf_jit_32.h
245
(_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
arch/arm/net/bpf_jit_32.h
246
#define ARM_ORR_SI(rd, rn, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
247
(ARM_ORR_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
248
#define ARM_ORRS_SI(rd, rn, rm, type, imm6) \
arch/arm/net/bpf_jit_32.h
249
(ARM_ORRS_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
arch/arm/net/bpf_jit_32.h
251
#define ARM_REV(rd, rm) (ARM_INST_REV | (rd) << 12 | (rm))
arch/arm/net/bpf_jit_32.h
252
#define ARM_REV16(rd, rm) (ARM_INST_REV16 | (rd) << 12 | (rm))
arch/arm/net/bpf_jit_32.h
258
#define ARM_SUB_R(rd, rn, rm) _AL3_R(ARM_INST_SUB, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
259
#define ARM_SUBS_R(rd, rn, rm) _AL3_R(ARM_INST_SUBS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
260
#define ARM_RSB_R(rd, rn, rm) _AL3_R(ARM_INST_RSB, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
261
#define ARM_SBC_R(rd, rn, rm) _AL3_R(ARM_INST_SBC, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
262
#define ARM_SBCS_R(rd, rn, rm) _AL3_R(ARM_INST_SBCS, rd, rn, rm)
arch/arm/net/bpf_jit_32.h
267
#define ARM_TST_R(rn, rm) _AL3_R(ARM_INST_TST, 0, rn, rm)
arch/arm/net/bpf_jit_32.h
270
#define ARM_UDIV(rd, rn, rm) (ARM_INST_UDIV | (rd) << 16 | (rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
271
#define ARM_SDIV(rd, rn, rm) (ARM_INST_SDIV | (rd) << 16 | (rn) | (rm) << 8)
arch/arm/net/bpf_jit_32.h
273
#define ARM_UMULL(rd_lo, rd_hi, rn, rm) (ARM_INST_UMULL | (rd_hi) << 16 \
arch/arm/net/bpf_jit_32.h
274
| (rd_lo) << 12 | (rm) << 8 | rn)
arch/arm/net/bpf_jit_32.h
276
#define ARM_MLS(rd, rn, rm, ra) (ARM_INST_MLS | (rd) << 16 | (rn) | (rm) << 8 \
arch/arm/net/bpf_jit_32.h
278
#define ARM_UXTH(rd, rm) (ARM_INST_UXTH | (rd) << 12 | (rm))
arch/arm/probes/decode-arm.c
79
int rm = insn & 0xf;
arch/arm/probes/decode-arm.c
80
long rmv = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
104
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
109
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
135
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
141
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
161
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
167
register unsigned long rmv asm("r3") = (rm == 15) ? pc
arch/arm/probes/kprobes/actions-arm.c
168
: regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
195
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
199
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
223
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
228
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
251
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
254
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
274
int rm = (insn >> 8) & 0xf;
arch/arm/probes/kprobes/actions-arm.c
279
register unsigned long rmv asm("r1") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-arm.c
75
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-arm.c
81
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
190
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
194
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
216
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
220
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
285
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
29
int rm = insn & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
290
register unsigned long rmv asm("r3") = regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
310
int rm = (insn >> 3) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
311
unsigned long rmv = (rm == 15) ? pc : regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
32
unsigned long rmv = regs->uregs[rm];
arch/arm/probes/kprobes/actions-thumb.c
488
int rm = (insn >> 3) & 0xf;
arch/arm/probes/kprobes/actions-thumb.c
495
rmv = (rm == 15) ? pc : regs->uregs[rm];
arch/mips/math-emu/cp1emu.c
1998
oldrm = ieee754_csr.rm;
arch/mips/math-emu/cp1emu.c
2000
ieee754_csr.rm = MIPSInst_FUNC(ir);
arch/mips/math-emu/cp1emu.c
2002
ieee754_csr.rm = oldrm;
arch/mips/math-emu/cp1emu.c
2044
oldrm = ieee754_csr.rm;
arch/mips/math-emu/cp1emu.c
2046
ieee754_csr.rm = MIPSInst_FUNC(ir);
arch/mips/math-emu/cp1emu.c
2048
ieee754_csr.rm = oldrm;
arch/mips/math-emu/cp1emu.c
2349
oldrm = ieee754_csr.rm;
arch/mips/math-emu/cp1emu.c
2351
ieee754_csr.rm = MIPSInst_FUNC(ir);
arch/mips/math-emu/cp1emu.c
2353
ieee754_csr.rm = oldrm;
arch/mips/math-emu/cp1emu.c
2395
oldrm = ieee754_csr.rm;
arch/mips/math-emu/cp1emu.c
2397
ieee754_csr.rm = MIPSInst_FUNC(ir);
arch/mips/math-emu/cp1emu.c
2399
ieee754_csr.rm = oldrm;
arch/mips/math-emu/dp_add.c
153
return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/dp_add.c
83
return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/dp_div.c
115
rm = 0;
arch/mips/math-emu/dp_div.c
121
rm |= bm;
arch/mips/math-emu/dp_div.c
128
rm <<= 1;
arch/mips/math-emu/dp_div.c
130
rm |= 1; /* have remainder, set sticky */
arch/mips/math-emu/dp_div.c
132
assert(rm);
arch/mips/math-emu/dp_div.c
137
while ((rm >> (DP_FBITS + 3)) == 0) {
arch/mips/math-emu/dp_div.c
138
rm <<= 1;
arch/mips/math-emu/dp_div.c
14
u64 rm;
arch/mips/math-emu/dp_div.c
142
return ieee754dp_format(xs == ys ? 0 : 1, re, rm);
arch/mips/math-emu/dp_maddf.c
146
return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/dp_maddf.c
283
return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/dp_mul.c
143
rm = hrm | (lrm != 0);
arch/mips/math-emu/dp_mul.c
148
if ((s64) rm < 0) {
arch/mips/math-emu/dp_mul.c
149
rm = (rm >> (64 - (DP_FBITS + 1 + 3))) |
arch/mips/math-emu/dp_mul.c
150
((rm << (DP_FBITS + 1 + 3)) != 0);
arch/mips/math-emu/dp_mul.c
153
rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) |
arch/mips/math-emu/dp_mul.c
154
((rm << (DP_FBITS + 1 + 3 + 1)) != 0);
arch/mips/math-emu/dp_mul.c
156
assert(rm & (DP_HIDDEN_BIT << 3));
arch/mips/math-emu/dp_mul.c
158
return ieee754dp_format(rs, re, rm);
arch/mips/math-emu/dp_mul.c
16
u64 rm;
arch/mips/math-emu/dp_rint.c
54
switch (ieee754_csr.rm) {
arch/mips/math-emu/dp_simple.c
22
oldrm = ieee754_csr.rm;
arch/mips/math-emu/dp_simple.c
23
ieee754_csr.rm = FPU_CSR_RD;
arch/mips/math-emu/dp_simple.c
25
ieee754_csr.rm = oldrm;
arch/mips/math-emu/dp_simple.c
40
oldrm = ieee754_csr.rm;
arch/mips/math-emu/dp_simple.c
41
ieee754_csr.rm = FPU_CSR_RD;
arch/mips/math-emu/dp_simple.c
46
ieee754_csr.rm = oldrm;
arch/mips/math-emu/dp_sqrt.c
113
ieee754_csr.rm = FPU_CSR_RZ;
arch/mips/math-emu/dp_sqrt.c
129
switch (oldcsr.rm) {
arch/mips/math-emu/dp_sqrt.c
69
ieee754_csr.rm = FPU_CSR_RN;
arch/mips/math-emu/dp_sub.c
157
if (ieee754_csr.rm == FPU_CSR_RD)
arch/mips/math-emu/dp_sub.c
83
return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/dp_tint.c
67
switch (ieee754_csr.rm) {
arch/mips/math-emu/dp_tlong.c
72
switch (ieee754_csr.rm) {
arch/mips/math-emu/ieee754.h
167
__BITFIELD_FIELD(unsigned rm:2, /* current rounding mode */
arch/mips/math-emu/ieee754.h
174
return (ieee754_csr.rm);
arch/mips/math-emu/ieee754.h
177
static inline unsigned int ieee754_setrm(unsigned int rm)
arch/mips/math-emu/ieee754.h
179
return (ieee754_csr.rm = rm);
arch/mips/math-emu/ieee754dp.c
102
switch(ieee754_csr.rm) {
arch/mips/math-emu/ieee754dp.c
166
switch (ieee754_csr.rm) {
arch/mips/math-emu/ieee754dp.c
61
switch (ieee754_csr.rm) {
arch/mips/math-emu/ieee754sp.c
102
switch(ieee754_csr.rm) {
arch/mips/math-emu/ieee754sp.c
165
switch (ieee754_csr.rm) {
arch/mips/math-emu/ieee754sp.c
61
switch (ieee754_csr.rm) {
arch/mips/math-emu/sp_add.c
152
return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/sp_add.c
83
return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/sp_div.c
115
rm = 0;
arch/mips/math-emu/sp_div.c
121
rm |= bm;
arch/mips/math-emu/sp_div.c
128
rm <<= 1;
arch/mips/math-emu/sp_div.c
130
rm |= 1; /* have remainder, set sticky */
arch/mips/math-emu/sp_div.c
132
assert(rm);
arch/mips/math-emu/sp_div.c
136
while ((rm >> (SP_FBITS + 3)) == 0) {
arch/mips/math-emu/sp_div.c
137
rm <<= 1;
arch/mips/math-emu/sp_div.c
14
unsigned int rm;
arch/mips/math-emu/sp_div.c
141
return ieee754sp_format(xs == ys ? 0 : 1, re, rm);
arch/mips/math-emu/sp_fdp.c
22
u32 rm;
arch/mips/math-emu/sp_fdp.c
57
if ((ieee754_csr.rm == FPU_CSR_RU && !xs) ||
arch/mips/math-emu/sp_fdp.c
58
(ieee754_csr.rm == FPU_CSR_RD && xs))
arch/mips/math-emu/sp_fdp.c
69
rm = (xm >> (DP_FBITS - (SP_FBITS + 3))) |
arch/mips/math-emu/sp_fdp.c
72
return ieee754sp_format(xs, xe, rm);
arch/mips/math-emu/sp_maddf.c
115
return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/sp_maddf.c
177
rm = XSPSRS64(rm64, (62 - 26));
arch/mips/math-emu/sp_maddf.c
178
return ieee754sp_format(rs, re, rm);
arch/mips/math-emu/sp_maddf.c
20
unsigned int rm;
arch/mips/math-emu/sp_maddf.c
223
return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/sp_mul.c
138
rm = hrm | (lrm != 0);
arch/mips/math-emu/sp_mul.c
143
if ((int) rm < 0) {
arch/mips/math-emu/sp_mul.c
144
rm = (rm >> (32 - (SP_FBITS + 1 + 3))) |
arch/mips/math-emu/sp_mul.c
145
((rm << (SP_FBITS + 1 + 3)) != 0);
arch/mips/math-emu/sp_mul.c
148
rm = (rm >> (32 - (SP_FBITS + 1 + 3 + 1))) |
arch/mips/math-emu/sp_mul.c
149
((rm << (SP_FBITS + 1 + 3 + 1)) != 0);
arch/mips/math-emu/sp_mul.c
151
assert(rm & (SP_HIDDEN_BIT << 3));
arch/mips/math-emu/sp_mul.c
153
return ieee754sp_format(rs, re, rm);
arch/mips/math-emu/sp_mul.c
16
unsigned int rm;
arch/mips/math-emu/sp_rint.c
55
switch (ieee754_csr.rm) {
arch/mips/math-emu/sp_simple.c
22
oldrm = ieee754_csr.rm;
arch/mips/math-emu/sp_simple.c
23
ieee754_csr.rm = FPU_CSR_RD;
arch/mips/math-emu/sp_simple.c
25
ieee754_csr.rm = oldrm;
arch/mips/math-emu/sp_simple.c
40
oldrm = ieee754_csr.rm;
arch/mips/math-emu/sp_simple.c
41
ieee754_csr.rm = FPU_CSR_RD;
arch/mips/math-emu/sp_simple.c
46
ieee754_csr.rm = oldrm;
arch/mips/math-emu/sp_sqrt.c
90
switch (ieee754_csr.rm) {
arch/mips/math-emu/sp_sub.c
154
if (ieee754_csr.rm == FPU_CSR_RD)
arch/mips/math-emu/sp_sub.c
83
return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
arch/mips/math-emu/sp_tint.c
72
switch (ieee754_csr.rm) {
arch/mips/math-emu/sp_tlong.c
68
switch (ieee754_csr.rm) {
arch/powerpc/platforms/ps3/mm.c
1000
if (r->offset + r->len > map.rm.size) {
arch/powerpc/platforms/ps3/mm.c
1002
virt_addr = map.rm.size;
arch/powerpc/platforms/ps3/mm.c
1004
if (r->offset >= map.rm.size)
arch/powerpc/platforms/ps3/mm.c
1005
virt_addr += r->offset - map.rm.size;
arch/powerpc/platforms/ps3/mm.c
1007
len -= map.rm.size - r->offset;
arch/powerpc/platforms/ps3/mm.c
101
struct mem_region rm;
arch/powerpc/platforms/ps3/mm.c
1030
if (r->offset < map.rm.size) {
arch/powerpc/platforms/ps3/mm.c
1032
lpar_addr = map.rm.base + r->offset;
arch/powerpc/platforms/ps3/mm.c
1033
len = map.rm.size - r->offset;
arch/powerpc/platforms/ps3/mm.c
1041
if (r->offset + r->len > map.rm.size) {
arch/powerpc/platforms/ps3/mm.c
1045
if (r->offset >= map.rm.size)
arch/powerpc/platforms/ps3/mm.c
1046
lpar_addr += r->offset - map.rm.size;
arch/powerpc/platforms/ps3/mm.c
1048
len -= map.rm.size - r->offset;
arch/powerpc/platforms/ps3/mm.c
110
DBG("%s:%d: map.rm.size = %llxh\n", func, line, m->rm.size);
arch/powerpc/platforms/ps3/mm.c
1131
if (r->offset >= map.rm.size)
arch/powerpc/platforms/ps3/mm.c
1207
result = ps3_repository_read_mm_info(&map.rm.base, &map.rm.size,
arch/powerpc/platforms/ps3/mm.c
1213
map.rm.offset = map.rm.base;
arch/powerpc/platforms/ps3/mm.c
1218
BUG_ON(map.rm.base);
arch/powerpc/platforms/ps3/mm.c
1219
BUG_ON(!map.rm.size);
arch/powerpc/platforms/ps3/mm.c
1224
result = ps3_mm_region_create(&map.r1, map.total - map.rm.size);
arch/powerpc/platforms/ps3/mm.c
1231
map.total = map.rm.size + map.r1.size;
arch/powerpc/platforms/ps3/mm.c
1237
__func__, __LINE__, map.rm.size,
arch/powerpc/platforms/ps3/mm.c
1238
map.total - map.rm.size);
arch/powerpc/platforms/ps3/mm.c
1239
memblock_add(map.rm.size, map.total - map.rm.size);
arch/powerpc/platforms/ps3/mm.c
128
return (phys_addr < map.rm.size || phys_addr >= map.total)
arch/powerpc/platforms/ps3/mm.c
233
r->offset = r->base - map.rm.size;
arch/powerpc/platforms/ps3/mm.c
285
if (result || r->base < map.rm.size) {
arch/powerpc/platforms/ps3/mm.c
292
r->offset = r->base - map.rm.size;
arch/powerpc/platforms/ps3/mm.c
321
map.total = map.rm.size;
arch/powerpc/platforms/ps3/mm.c
340
if (lpar_addr >= map.rm.size)
arch/powerpc/platforms/ps3/mm.c
988
if (r->offset < map.rm.size) {
arch/powerpc/platforms/ps3/mm.c
990
virt_addr = map.rm.base + r->offset;
arch/powerpc/platforms/ps3/mm.c
991
len = map.rm.size - r->offset;
arch/sh/kernel/disassemble.c
302
int rm = 0;
arch/sh/kernel/disassemble.c
364
rm = nibs[n];
arch/sh/kernel/disassemble.c
368
rm = (nibs[n] & 0x3);
arch/sh/kernel/disassemble.c
408
pr_cont("r%d", rm);
arch/sh/kernel/disassemble.c
411
pr_cont("@r%d+", rm);
arch/sh/kernel/disassemble.c
414
pr_cont("@-r%d", rm);
arch/sh/kernel/disassemble.c
417
pr_cont("@r%d", rm);
arch/sh/kernel/disassemble.c
420
pr_cont("@(%d,r%d)", imm, rm);
arch/sh/kernel/disassemble.c
435
pr_cont("@(r0,r%d)", rm);
arch/sh/kernel/disassemble.c
482
pr_cont("fr%d", rm);
arch/sh/kernel/disassemble.c
494
if (rm & 1) {
arch/sh/kernel/disassemble.c
495
pr_cont("xd%d", rm & ~1);
arch/sh/kernel/disassemble.c
500
pr_cont("dr%d", rm);
arch/sh/kernel/disassemble.c
517
pr_cont("fv%d", rm*4);
arch/sh/kernel/traps_32.c
107
unsigned long *rm, *rn;
arch/sh/kernel/traps_32.c
115
rm = &regs->regs[index];
arch/sh/kernel/traps_32.c
131
srcu = (unsigned char __user *)*rm;
arch/sh/kernel/traps_32.c
145
src = (unsigned char *)rm;
arch/sh/kernel/traps_32.c
159
src = (unsigned char*) rm;
arch/sh/kernel/traps_32.c
171
src = (unsigned char*) rm;
arch/sh/kernel/traps_32.c
182
srcu = (unsigned char __user *)*rm;
arch/sh/kernel/traps_32.c
193
srcu = (unsigned char __user *)*rm;
arch/sh/kernel/traps_32.c
195
*rm += count;
arch/sh/kernel/traps_32.c
215
dstu = (unsigned char __user *)*rm; /* called Rn in the spec */
arch/sh/kernel/traps_32.c
224
srcu = (unsigned char __user *)*rm;
arch/sh/kernel/traps_32.c
324
u_int rm;
arch/sh/kernel/traps_32.c
334
rm = regs->regs[index];
arch/sh/kernel/traps_32.c
362
regs->pc += rm + 4;
arch/sh/kernel/traps_32.c
369
regs->pc += rm + 4;
arch/sh/kernel/traps_32.c
389
regs->pc = rm;
arch/sh/kernel/traps_32.c
396
regs->pc = rm;
arch/x86/include/asm/fpu/types.h
95
u8 rm;
arch/x86/kernel/traps.c
114
u8 v, reg, rm, rex = 0;
arch/x86/kernel/traps.c
173
rm = X86_MODRM_RM(v) + 8*!!X86_REX_B(rex);
arch/x86/kernel/traps.c
180
if (rm == 0) /* (%eax) */
arch/x86/kernel/traps.c
183
if (rm == 2) { /* (%edx) */
arch/x86/kernel/traps.c
191
if (rm == 0) /* (%eax) */
arch/x86/kernel/traps.c
197
if (rm == 0) /* (%eax) */
arch/x86/lib/insn.c
501
insn_byte_t mod, rm, base;
arch/x86/lib/insn.c
530
rm = X86_MODRM_RM(insn->modrm.value);
arch/x86/lib/insn.c
538
if ((mod == 0 && rm == 6) || mod == 2) {
arch/x86/lib/insn.c
543
if ((mod == 0 && rm == 5) || mod == 2 ||
arch/x86/math-emu/fpu_proto.h
111
extern int FPU_sub(int flags, int rm, int control_w);
arch/x86/math-emu/fpu_system.h
92
#define FPU_rm (*(u_char *)&(I387->soft.rm))
arch/x86/math-emu/get_address.c
214
unsigned rm = FPU_modrm & 7;
arch/x86/math-emu/get_address.c
229
if (rm == 4 && mod != 3) {
arch/x86/math-emu/get_address.c
232
cpu_reg_ptr = &REG_(rm);
arch/x86/math-emu/get_address.c
235
if (rm == 5) {
arch/x86/math-emu/get_address.c
299
unsigned rm = FPU_modrm & 7;
arch/x86/math-emu/get_address.c
315
if (rm == 6) {
arch/x86/math-emu/get_address.c
347
switch (rm) {
arch/x86/math-emu/reg_add_sub.c
129
int FPU_sub(int flags, int rm, int control_w)
arch/x86/math-emu/reg_add_sub.c
141
b = (FPU_REG *) rm;
arch/x86/math-emu/reg_add_sub.c
144
b = &st(rm);
arch/x86/math-emu/reg_add_sub.c
145
tagb = FPU_gettagi(rm);
arch/x86/math-emu/reg_add_sub.c
148
deststnr = rm;
arch/x86/math-emu/reg_divide.c
133
return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0,
arch/x86/math-emu/reg_divide.c
141
return real_2op_NaN(st0_ptr, tag, rm,
arch/x86/math-emu/reg_divide.c
142
(flags & REV) ? st0_ptr : &st(rm));
arch/x86/math-emu/reg_divide.c
145
tag = FPU_gettagi(rm);
arch/x86/math-emu/reg_divide.c
147
tag = FPU_Special(&st(rm));
arch/x86/math-emu/reg_divide.c
148
return real_2op_NaN(&st(rm), tag, 0,
arch/x86/math-emu/reg_divide.c
149
(flags & REV) ? st0_ptr : &st(rm));
arch/x86/math-emu/reg_divide.c
28
int FPU_div(int flags, int rm, int control_w)
arch/x86/math-emu/reg_divide.c
37
deststnr = rm;
arch/x86/math-emu/reg_divide.c
46
a = (FPU_REG *) rm;
arch/x86/math-emu/reg_divide.c
49
a = &st(rm);
arch/x86/math-emu/reg_divide.c
51
taga = FPU_gettagi(rm);
arch/x86/math-emu/reg_divide.c
58
b = (FPU_REG *) rm;
arch/x86/math-emu/reg_divide.c
61
b = &st(rm);
arch/x86/math-emu/reg_divide.c
63
tagb = FPU_gettagi(rm);
drivers/gpu/drm/armada/armada_crtc.c
336
uint32_t lm, rm, tm, bm, val, sclk;
drivers/gpu/drm/armada/armada_crtc.c
342
rm = adj->crtc_hsync_start - adj->crtc_hdisplay;
drivers/gpu/drm/armada/armada_crtc.c
349
DRM_DEBUG_KMS("lm %d rm %d tm %d bm %d\n", lm, rm, tm, bm);
drivers/gpu/drm/armada/armada_crtc.c
380
armada_reg_queue_set(regs, i, (lm << 16) | rm, LCD_SPU_H_PORCH);
drivers/gpu/drm/i915/gvt/handlers.c
2165
#define MMIO_F(reg, s, f, am, rm, d, r, w) do { \
drivers/gpu/drm/i915/gvt/handlers.c
2167
s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2184
#define MMIO_RO(reg, d, f, rm, r, w) \
drivers/gpu/drm/i915/gvt/handlers.c
2185
MMIO_F(reg, 4, F_RO | f, 0, rm, d, r, w)
drivers/gpu/drm/i915/gvt/handlers.c
2187
#define MMIO_RING_F(prefix, s, f, am, rm, d, r, w) do { \
drivers/gpu/drm/i915/gvt/handlers.c
2188
MMIO_F(prefix(RENDER_RING_BASE), s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2189
MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2190
MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2191
MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2193
MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \
drivers/gpu/drm/i915/gvt/handlers.c
2205
#define MMIO_RING_RO(prefix, d, f, rm, r, w) \
drivers/gpu/drm/i915/gvt/handlers.c
2206
MMIO_RING_F(prefix, 4, F_RO | f, 0, rm, d, r, w)
drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
1450
ret = dpu_rm_reserve(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
1457
num_ctl = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
1461
num_lm = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_crtc.c
1465
num_dspp = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1194
num_pp = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1198
num_cwb = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1203
num_pp = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1216
num_ctl = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1223
num_dsc = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
1237
dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2193
num_lm = dpu_rm_get_assigned_resources(&phys_enc->dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2374
num_pp = dpu_rm_get_assigned_resources(&dpu_kms->rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2677
phys_params.hw_intf = dpu_encoder_get_intf(dpu_kms->catalog, &dpu_kms->rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2682
phys_params.hw_wb = dpu_rm_get_wb(&dpu_kms->rm, controller_id);
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
1205
rc = dpu_rm_init(dev, &dpu_kms->rm, dpu_kms->catalog, dpu_kms->mdss, dpu_kms->mmio);
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
282
struct dpu_hw_sspp *hw = dpu_rm_get_sspp(&dpu_kms->rm, i);
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
396
state->rm = &dpu_kms->rm;
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
732
maxlinewidth = dpu_rm_get_wb(&dpu_kms->rm, info.h_tile_instance[0])->caps->maxlinewidth;
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
125
struct dpu_rm *rm;
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h
82
struct dpu_rm rm;
drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c
1123
pipe->sspp = dpu_rm_get_sspp(&dpu_kms->rm, pdpu->pipe);
drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c
1215
pipe->sspp = dpu_rm_reserve_sspp(&dpu_kms->rm, global_state, crtc, reqs);
drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c
1230
r_pipe->sspp = dpu_rm_reserve_sspp(&dpu_kms->rm, global_state, crtc, reqs);
drivers/gpu/drm/msm/disp/dpu1/dpu_plane.c
1831
pipe_hw = dpu_rm_get_sspp(&kms->rm, pipe);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1002
const struct dpu_rm *rm = global_state->rm;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1008
dpu_rm_print_state_helper(p, rm->pingpong_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1014
dpu_rm_print_state_helper(p, rm->mixer_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1020
dpu_rm_print_state_helper(p, rm->ctl_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1026
dpu_rm_print_state_helper(p, rm->dspp_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1032
dpu_rm_print_state_helper(p, rm->dsc_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1037
dpu_rm_print_state_helper(p, rm->cdm_blk,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1044
dpu_rm_print_state_helper(p, rm->hw_sspp[i] ? &rm->hw_sspp[i]->base : NULL,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
1050
dpu_rm_print_state_helper(p, rm->cwb_blks[i],
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
112
rm->hw_intf[intf->id - INTF_0] = hw;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
125
rm->hw_wb[wb->id - WB_0] = hw;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
138
rm->cwb_blks[cwb->id - CWB_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
151
rm->ctl_blks[ctl->id - CTL_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
164
rm->dspp_blks[dspp->id - DSPP_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
181
rm->dsc_blks[dsc->id - DSC_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
194
rm->hw_sspp[sspp->id - SSPP_NONE] = hw;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
206
rm->cdm_blk = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
227
static int _dpu_rm_get_lm_peer(struct dpu_rm *rm, int primary_idx)
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
231
prim_lm_cfg = to_dpu_hw_mixer(rm->mixer_blks[primary_idx])->cap;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
238
static int _dpu_rm_reserve_cwb_mux_and_pingpongs(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
255
for (int i = 0; i < ARRAY_SIZE(rm->mixer_blks) &&
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
257
for (int j = 0; j < ARRAY_SIZE(rm->cwb_blks); j++) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
308
static bool _dpu_rm_check_lm_and_get_connected_blks(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
322
lm_cfg = to_dpu_hw_mixer(rm->mixer_blks[lm_idx])->cap;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
324
if (idx < 0 || idx >= ARRAY_SIZE(rm->pingpong_blks) || !rm->pingpong_blks[idx]) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
339
if (idx < 0 || idx >= ARRAY_SIZE(rm->dspp_blks) || !rm->dspp_blks[idx]) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
353
static int _dpu_rm_reserve_lms(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
370
for (i = 0; i < ARRAY_SIZE(rm->mixer_blks) &&
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
372
if (!rm->mixer_blks[i])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
382
if (!_dpu_rm_check_lm_and_get_connected_blks(rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
392
int j = _dpu_rm_get_lm_peer(rm, i);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
398
if (!rm->mixer_blks[j])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
401
if (!_dpu_rm_check_lm_and_get_connected_blks(rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
41
struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
432
struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
441
if (rm->has_legacy_ctls) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
455
for (j = 0; j < ARRAY_SIZE(rm->ctl_blks); j++) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
460
if (!rm->ctl_blks[j])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
465
ctl = to_dpu_hw_ctl(rm->ctl_blks[j]);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
48
if (!rm || !cat || !mmio) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
519
static int _dpu_rm_dsc_alloc(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
529
for (dsc_idx = 0; dsc_idx < ARRAY_SIZE(rm->dsc_blks) &&
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
531
if (!rm->dsc_blks[dsc_idx])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
54
memset(rm, 0, sizeof(*rm));
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
559
static int _dpu_rm_dsc_alloc_pair(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
56
rm->has_legacy_ctls = (cat->mdss_ver->core_major_ver < 5);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
569
for (dsc_idx = 0; dsc_idx < ARRAY_SIZE(rm->dsc_blks) &&
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
571
if (!rm->dsc_blks[dsc_idx] ||
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
572
!rm->dsc_blks[dsc_idx + 1])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
615
static int _dpu_rm_reserve_dsc(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
635
return _dpu_rm_dsc_alloc_pair(rm, global_state, crtc_id, top);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
637
return _dpu_rm_dsc_alloc(rm, global_state, crtc_id, top);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
642
static int _dpu_rm_reserve_cdm(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
648
if (!rm->cdm_blk) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
669
struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
676
ret = _dpu_rm_reserve_lms(rm, global_state, crtc_id, topology);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
683
ret = _dpu_rm_reserve_cwb_mux_and_pingpongs(rm, global_state,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
689
ret = _dpu_rm_reserve_ctls(rm, global_state, crtc_id,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
69
rm->mixer_blks[lm->id - LM_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
696
ret = _dpu_rm_reserve_dsc(rm, global_state, crtc_id, topology);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
701
ret = _dpu_rm_reserve_cdm(rm, global_state, crtc_id, topology->num_cdm);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
763
struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
781
ret = _dpu_rm_make_reservation(rm, global_state, crtc->base.id, topology);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
788
static struct dpu_hw_sspp *dpu_rm_try_sspp(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
798
for (i = 0; i < ARRAY_SIZE(rm->hw_sspp); i++) {
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
799
if (!rm->hw_sspp[i])
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
805
hw_sspp = rm->hw_sspp[i];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
825
return rm->hw_sspp[i];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
83
rm->merge_3d_blks[merge_3d->id - MERGE_3D_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
838
struct dpu_hw_sspp *dpu_rm_reserve_sspp(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
846
hw_sspp = dpu_rm_try_sspp(rm, global_state, crtc, reqs, SSPP_TYPE_DMA);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
848
hw_sspp = dpu_rm_try_sspp(rm, global_state, crtc, reqs, SSPP_TYPE_RGB);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
850
hw_sspp = dpu_rm_try_sspp(rm, global_state, crtc, reqs, SSPP_TYPE_VIG);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
895
int dpu_rm_get_assigned_resources(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
907
hw_blks = rm->pingpong_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
909
max_blks = ARRAY_SIZE(rm->pingpong_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
912
hw_blks = rm->mixer_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
914
max_blks = ARRAY_SIZE(rm->mixer_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
917
hw_blks = rm->ctl_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
919
max_blks = ARRAY_SIZE(rm->ctl_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
922
hw_blks = rm->dspp_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
924
max_blks = ARRAY_SIZE(rm->dspp_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
927
hw_blks = rm->dsc_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
929
max_blks = ARRAY_SIZE(rm->dsc_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
932
hw_blks = &rm->cdm_blk;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
937
hw_blks = rm->cwb_blks;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
939
max_blks = ARRAY_SIZE(rm->cwb_blks);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
98
hw->merge_3d = to_dpu_hw_merge_3d(rm->merge_3d_blks[pp->merge_3d - MERGE_3D_0]);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
99
rm->pingpong_blks[pp->id - PINGPONG_0] = &hw->base;
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
103
static inline struct dpu_hw_intf *dpu_rm_get_intf(struct dpu_rm *rm, enum dpu_intf intf_idx)
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
105
return rm->hw_intf[intf_idx - INTF_0];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
113
static inline struct dpu_hw_wb *dpu_rm_get_wb(struct dpu_rm *rm, enum dpu_wb wb_idx)
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
115
return rm->hw_wb[wb_idx - WB_0];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
123
static inline struct dpu_hw_sspp *dpu_rm_get_sspp(struct dpu_rm *rm, enum dpu_sspp sspp_idx)
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
125
return rm->hw_sspp[sspp_idx - SSPP_NONE];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
70
struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
75
int dpu_rm_reserve(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
83
struct dpu_hw_sspp *dpu_rm_reserve_sspp(struct dpu_rm *rm,
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h
91
int dpu_rm_get_assigned_resources(struct dpu_rm *rm,
drivers/gpu/drm/nouveau/include/nvkm/engine/disp.h
29
} rm;
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
46
} rm;
drivers/gpu/drm/nouveau/include/nvkm/engine/fifo.h
85
} rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
236
struct nvkm_rm *rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
259
struct dentry *rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
273
return gsp && (gsp->fws.rm || gsp->fw.img);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
281
return gsp->rm->api->rpc->get(gsp, fn, argc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
288
return gsp->rm->api->rpc->push(gsp, argv, policy, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
317
gsp->rm->api->rpc->done(gsp, repv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
323
return object->client->gsp->rm->api->ctrl->get(object, cmd, argc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
329
return object->client->gsp->rm->api->ctrl->push(object, argv, repc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
360
object->client->gsp->rm->api->ctrl->done(object, repv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
375
argv = gsp->rm->api->alloc->get(object, oclass, argc);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
387
void *repv = object->client->gsp->rm->api->alloc->push(object, argv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
409
object->client->gsp->rm->api->alloc->done(object, repv);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
428
int ret = object->client->gsp->rm->api->alloc->free(object);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
442
return client->gsp->rm->api->device->ctor(client, device);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
449
device->object.client->gsp->rm->api->device->dtor(device);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
481
struct nvkm_rm *rm = device->object.client->gsp->rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
483
return rm->api->device->event.ctor(device, handle, id, func, event);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
492
device->object.client->gsp->rm->api->device->event.dtor(event);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
79
const struct firmware *rm;
drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
77
} rm;
drivers/gpu/drm/nouveau/nvkm/engine/disp/chan.h
28
} rm;
drivers/gpu/drm/nouveau/nvkm/engine/disp/outp.c
389
if (!disp->rm.client.gsp)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
105
if (!disp->rm.client.gsp && conn->info.hpd == DCB_GPIO_UNUSED)
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
121
if (disp->rm.client.gsp) {
drivers/gpu/drm/nouveau/nvkm/engine/disp/uconn.c
126
return nvkm_uevent_add(uevent, &disp->rm.event, outp->index, bits,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
65
u32 rm = ((1ULL << c->bits) - 1) << c->regs;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv04.c
67
u32 rv = (nvkm_rd32(device, c->regp) & rm) >> c->regs;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
35
} rm;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
36
struct nvkm_devinit_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
39
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
42
rm->dtor = r535_devinit_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
43
rm->post = hw->post;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
44
rm->disable = hw->disable;
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
46
ret = nv50_devinit_new_(rm, device, type, inst, pdevinit);
drivers/gpu/drm/nouveau/nvkm/subdev/devinit/r535.c
48
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
71
struct nvkm_fb_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
74
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
77
rm->dtor = r535_fb_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
78
rm->sysmem.flush_page_init = hw->sysmem.flush_page_init;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
79
rm->vidmem.size = hw->vidmem.size;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
80
rm->ram_new = r535_fb_ram_new;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
82
ret = nvkm_fb_new_(rm, device, type, inst, pfb);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
84
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ad102.c
42
.rm.gpu = &ad10x_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
107
kfree(gsp->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
148
if (fwif->rm) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
151
gsp->rm = kzalloc_obj(*gsp->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
152
if (!gsp->rm)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
155
gsp->rm->device = device;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
156
gsp->rm->gpu = fwif->func->rm.gpu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
157
gsp->rm->wpr = fwif->rm->wpr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
158
gsp->rm->api = fwif->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
94
nvkm_firmware_put(gsp->fws.rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/base.c
95
gsp->fws.rm = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga100.c
59
.rm.gpu = &ga100_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
170
.rm.gpu = &ga1xx_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gb100.c
18
.rm.gpu = &gb10x_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gb202.c
18
.rm.gpu = &gb20x_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
102
if (gsp->rm->wpr->rsvd_size_pmu)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
103
rsvd_size = ALIGN(rsvd_size + gsp->rm->wpr->rsvd_size_pmu, 0x200000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
142
gsp->fb.heap.size = gsp->rm->wpr->heap_size_non_wpr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
165
meta->pmuReservedSize = gsp->rm->wpr->rsvd_size_pmu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
324
.rm.gpu = &gh100_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
87
gsp->rm->api->gsp->set_rmargs(gsp, true);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/priv.h
17
const struct nvkm_rm_impl *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/priv.h
67
} rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
31
if (WARN_ON(!gsp->rm))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/client.c
44
ret = gsp->rm->api->client->ctor(client, NVKM_RM_CLIENT(id));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
103
nvkm_rm_engine_ctor(void *(*dtor)(struct nvkm_engine *), struct nvkm_rm *rm,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
122
nvkm_engine_ctor(func, rm->device, type, inst, true, engine);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
127
nvkm_rm_engine_new_(struct nvkm_rm *rm, enum nvkm_subdev_type type, int inst, u32 class,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
137
ret = nvkm_rm_engine_ctor(nvkm_rm_engine_dtor, rm, type, inst, &class, 1, engine);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
148
nvkm_rm_engine_new(struct nvkm_rm *rm, enum nvkm_subdev_type type, int inst)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
150
const struct nvkm_rm_gpu *gpu = rm->gpu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
151
struct nvkm_device *device = rm->device;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
158
return nvkm_rm_engine_new_(rm, type, inst, gpu->ce.class, &device->ce[inst]);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
163
return nvkm_rm_gr_new(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
168
return nvkm_rm_nvdec_new(rm, inst);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
173
return nvkm_rm_nvenc_new(rm, inst);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
178
return nvkm_rm_engine_new_(rm, type, inst, gpu->nvjpg.class, &device->nvjpg[inst]);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
183
return nvkm_rm_engine_new_(rm, type, inst, gpu->ofa.class, &device->ofa[inst]);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
19
struct nvkm_gsp_object rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
27
nvkm_gsp_rm_free(&obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
40
struct nvkm_rm *rm = chan->client->gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
53
ret = rm->api->ce->alloc(chan, handle, class, inst, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
56
ret = nvkm_gsp_rm_alloc(chan, handle, class, 0, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
59
ret = rm->api->nvdec->alloc(chan, handle, class, inst, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
62
ret = rm->api->nvenc->alloc(chan, handle, class, inst, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
65
ret = rm->api->nvjpg->alloc(chan, handle, class, inst, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
68
ret = rm->api->ofa->alloc(chan, handle, class, inst, &obj->rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/engine.c
92
return nvkm_rm_engine_obj_new(&chan->rm.object, chan->id, oclass, pobject);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
16
return nvkm_rm_engine_obj_new(&chan->chan->rm.object, chan->chan->id, oclass, pobject);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
22
struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
25
if (rm->api->gr->scrubber.fini)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
26
rm->api->gr->scrubber.fini(gr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
34
struct nvkm_rm *rm = base->engine.subdev.device->gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
38
if (rm->api->gr->scrubber.init) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
39
ret = rm->api->gr->scrubber.init(gr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
48
nvkm_rm_gr_new(struct nvkm_rm *rm)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
51
rm->gpu->gr.class.i2m,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
52
rm->gpu->gr.class.twod,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
53
rm->gpu->gr.class.threed,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
54
rm->gpu->gr.class.compute,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
83
nvkm_gr_ctor(func, rm->device, NVKM_ENGINE_GR, 0, true, &gr->base);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/gr.c
85
rm->device->gr = &gr->base;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvdec.c
15
nvkm_rm_nvdec_new(struct nvkm_rm *rm, int inst)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvdec.c
24
ret = nvkm_rm_engine_ctor(nvkm_rm_nvdec_dtor, rm, NVKM_ENGINE_NVDEC, inst,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvdec.c
25
&rm->gpu->nvdec.class, 1, &nvdec->engine);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvdec.c
31
rm->device->nvdec[inst] = nvdec;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvenc.c
15
nvkm_rm_nvenc_new(struct nvkm_rm *rm, int inst)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvenc.c
24
ret = nvkm_rm_engine_ctor(nvkm_rm_nvenc_dtor, rm, NVKM_ENGINE_NVENC, inst,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvenc.c
25
&rm->gpu->nvenc.class, 1, &nvenc->engine);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/nvenc.c
31
rm->device->nvenc[inst] = nvenc;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
168
struct nvkm_bar_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
172
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
175
rm->dtor = r535_bar_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
176
rm->oneinit = hw->oneinit;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
177
rm->bar1.init = r535_bar_bar1_init;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
178
rm->bar1.fini = r535_bar_bar1_fini;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
179
rm->bar1.wait = r535_bar_bar1_wait;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
180
rm->bar1.vmm = hw->bar1.vmm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
181
rm->bar2.init = r535_bar_bar2_init;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
182
rm->bar2.fini = r535_bar_bar2_fini;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
183
rm->bar2.wait = r535_bar_bar2_wait;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
184
rm->bar2.vmm = hw->bar2.vmm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
185
rm->flush = r535_bar_flush;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
187
ret = gf100_bar_new_(rm, device, type, inst, &bar);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
189
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
98
vmm->rm.bar2_pdb = gsp->bar.rm_bar2_pdb;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1003
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1013
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1018
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1049
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1058
return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1064
struct nvkm_rm *rm = outp->disp->rm.objcom.client->gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1070
return rm->api->disp->dp.set_indexed_link_rates(outp);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1083
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, NV0073_CTRL_CMD_DP_AUXCH_CTRL, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1099
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1109
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1114
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
116
const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1176
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1183
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1185
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1209
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1220
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1228
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1230
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
124
args = nvkm_gsp_rm_alloc_get(&chan->disp->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1241
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1257
const struct nvkm_rm_api *rmapi = disp->rm.objcom.client->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
126
chan->object.oclass, sizeof(*args), &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1266
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1274
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1276
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1317
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
132
return nvkm_gsp_rm_alloc_wr(&chan->rm.object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1370
struct nvkm_disp *disp = container_of(event, typeof(*disp), rm.irq);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1379
nvkm_event_ntfy(&disp->rm.event, fls(irq->displayId) - 1, NVKM_DPYID_IRQ);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1385
struct nvkm_disp *disp = container_of(event, typeof(*disp), rm.hpd);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1403
nvkm_event_ntfy(&disp->rm.event, i, mask);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1446
nvkm_gsp_rm_free(&disp->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1449
nvkm_gsp_event_dtor(&disp->rm.irq);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1450
nvkm_gsp_event_dtor(&disp->rm.hpd);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1451
nvkm_event_fini(&disp->rm.event);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1453
nvkm_gsp_rm_free(&disp->rm.objcom);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1454
nvkm_gsp_device_dtor(&disp->rm.device);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1455
nvkm_gsp_client_dtor(&disp->rm.client);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1464
ret = nvkm_gsp_rm_alloc(&disp->rm.device.object, disp->func->root.oclass << 16,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1465
disp->func->root.oclass, 0, &disp->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1477
ctrl = nvkm_gsp_rm_ctrl_rd(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1483
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1491
struct nvkm_gsp *gsp = disp->rm.objcom.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1511
const struct nvkm_rm_api *rmapi = gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
154
(chan->disp->rm.client.object.handle & 0x3fff));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1540
ret = nvkm_gsp_client_device_ctor(gsp, &disp->rm.client, &disp->rm.device);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1544
ret = nvkm_gsp_rm_alloc(&disp->rm.device.object, NVKM_RM_DISP, NV04_DISPLAY_COMMON, 0,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1545
&disp->rm.objcom);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1557
struct nvkm_gsp_object *subdevice = &disp->rm.client.gsp->internal.device.subdevice;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1628
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1634
ret = nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1643
ctrl = nvkm_gsp_rm_ctrl_rd(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1649
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1656
ctrl = nvkm_gsp_rm_ctrl_rd(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1663
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1690
ret = nvkm_event_init(&r535_disp_event, &gsp->subdev, 3, 32, &disp->rm.event);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1694
ret = nvkm_gsp_device_event_ctor(&disp->rm.device, 0x007e0000, NV2080_NOTIFIERS_HOTPLUG,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1695
r535_disp_hpd, &disp->rm.hpd);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1699
ret = nvkm_gsp_device_event_ctor(&disp->rm.device, 0x007e0001, NV2080_NOTIFIERS_DP_IRQ,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1700
r535_disp_irq, &disp->rm.irq);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
173
args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1733
const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1734
struct nvkm_disp_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1737
if (!(rm = kzalloc(sizeof(*rm) + 6 * sizeof(rm->user[0]), GFP_KERNEL)))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1740
rm->dtor = r535_disp_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1741
rm->oneinit = r535_disp_oneinit;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1742
rm->init = r535_disp_init;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1743
rm->fini = r535_disp_fini;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1744
rm->uevent = hw->uevent;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1745
rm->sor.cnt = r535_sor_cnt;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1746
rm->sor.new = r535_sor_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1747
rm->ramht_size = hw->ramht_size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1749
rm->root.oclass = gpu->disp.class.root;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1751
rm->user[0].base.oclass = gpu->disp.class.caps;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1752
rm->user[0].ctor = gv100_disp_caps_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1754
rm->user[1].base.oclass = gpu->disp.class.core;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1755
rm->user[1].ctor = nvkm_disp_core_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1756
rm->user[1].chan = &r535_core;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1758
rm->user[2].base.oclass = gpu->disp.class.wndw;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1759
rm->user[2].ctor = nvkm_disp_wndw_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1760
rm->user[2].chan = &r535_wndw;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1762
rm->user[3].base.oclass = gpu->disp.class.wimm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1763
rm->user[3].ctor = nvkm_disp_wndw_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1764
rm->user[3].chan = &r535_wimm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1766
rm->user[4].base.oclass = gpu->disp.class.curs;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1767
rm->user[4].ctor = nvkm_disp_chan_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1768
rm->user[4].chan = &r535_curs;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1770
ret = nvkm_disp_new_(rm, device, type, inst, pdisp);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
1772
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
187
const struct nvkm_rm_api *rmapi = chan->disp->rm.objcom.client->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
195
chan->suspend_put, &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
272
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, cmd, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
279
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
285
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
293
const struct nvkm_rm_api *rmapi = disp->engine.subdev.device->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
302
const struct nvkm_rm_api *rmapi = disp->engine.subdev.device->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
323
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
336
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
348
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
356
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
371
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
378
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
390
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
397
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
409
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
432
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
442
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
461
return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
479
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
493
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
502
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
509
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
518
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
542
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
569
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
577
WARN_ON(nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
648
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
656
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
658
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
664
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
671
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
684
outp->disp->rm.assigned_sors &= ~BIT(outp->ior->id);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
69
nvkm_gsp_rm_free(&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
697
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
704
ctrl->sorExcludeMask = disp->rm.assigned_sors;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
708
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
710
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
716
disp->rm.assigned_sors |= BIT(or);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
721
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
740
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
748
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
75
struct nvkm_gsp *gsp = disp->rm.objcom.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
750
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
755
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
768
const struct nvkm_rm_api *rmapi = disp->rm.objcom.client->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
779
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
788
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
790
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
796
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
830
disp->rm.assigned_sors |= BIT(ior->id);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
845
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, NV0073_CTRL_CMD_DFP_GET_INFO, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
851
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
853
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
860
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
870
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
878
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
882
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
889
const struct nvkm_rm_api *rmapi = outp->disp->rm.objcom.client->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
908
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
915
return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
925
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
933
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
935
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
940
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
950
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
961
return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/disp.c
992
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, NV0073_CTRL_CMD_DP_CTRL,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
309
struct nvkm_instmem_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
312
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
315
rm->dtor = r535_instmem_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
316
rm->fini = hw->fini;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
317
rm->memory_new = hw->memory_new;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
318
rm->memory_wrap = hw->memory_wrap;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
319
rm->zero = false;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
320
rm->set_bar0_window_addr = hw->set_bar0_window_addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
322
ret = nv50_instmem_new_(rm, device, type, inst, pinstmem);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fbsr.c
324
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
120
args->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
139
args->mthdbufMem.size = fifo->rm.mthdbuf_size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
159
const struct nvkm_rm_api *rmapi = device->gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
177
chan->rm.mthdbuf.ptr = dma_alloc_coherent(fifo->engine.subdev.device->dev,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
178
fifo->rm.mthdbuf_size,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
179
&chan->rm.mthdbuf.addr, GFP_KERNEL);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
180
if (!chan->rm.mthdbuf.ptr)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
183
ret = rmapi->fifo->chan.alloc(&chan->vmm->rm.device, NVKM_RM_CHAN(chan->id),
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
186
chan->rm.mthdbuf.addr, chan->vmm, offset, length,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
187
&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
197
ctrl = nvkm_gsp_rm_ctrl_get(&chan->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
204
ret = nvkm_gsp_rm_ctrl_wr(&chan->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
209
ctrl = nvkm_gsp_rm_ctrl_get(&chan->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
215
ret = nvkm_gsp_rm_ctrl_wr(&chan->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
266
chan->rm.grctx = vctx;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
279
struct nvkm_gsp_client *client = &chan->vmm->rm.client;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
282
ctrl = nvkm_gsp_rm_ctrl_get(&chan->vmm->rm.device.subdevice,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
288
ctrl->hObject = chan->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
295
return nvkm_gsp_rm_ctrl_wr(&chan->vmm->rm.device.subdevice, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
303
if (WARN_ON(!engn->rm.size))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
306
ret = nvkm_gpuobj_new(engn->engine->subdev.device, engn->rm.size, 0, true, NULL,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
378
r535_fifo_xlat_rm_engine_type(u32 rm, enum nvkm_subdev_type *ptype, int *p2080)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
386
switch (rm) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
44
struct nvkm_gsp *gsp = chan->rm.object.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
443
if (engn->rm.desc == ctrl->constructedFalconsTable[i].engDesc) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
444
engn->rm.size =
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
46
return gsp->rm->gpu->fifo.chan.doorbell_handle(chan);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
462
struct nvkm_rm *rm = gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
466
u32 first = rm->api->fifo->rsvd_chids;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
503
inst = rm->api->fifo->xlat_rm_engine_type(rmid, &type, &nv2080);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
518
rm->gpu->ce.grce_mask &&
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
519
(rm->gpu->ce.grce_mask(device) & BIT(inst)) &&
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
526
ret = nvkm_rm_engine_new(gsp->rm, type, inst);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
559
engn->rm.desc = ctrl->entries[i].engineData[ENGINE_INFO_TYPE_ENG_DESC];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
573
fifo->rm.mthdbuf_size = ctrl->size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
578
return rm->api->fifo->ectx_size(fifo);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
591
const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
592
struct nvkm_fifo_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
594
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
597
rm->dtor = r535_fifo_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
598
rm->runl_ctor = r535_fifo_runl_ctor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
599
rm->runl = &r535_runl;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
600
rm->chan.user.oclass = gpu->fifo.chan.class;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
601
rm->chan.func = &r535_chan;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
602
rm->nonstall = &ga100_fifo_nonstall;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
603
rm->nonstall_ctor = ga100_fifo_nonstall_ctor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
604
rm->nonstall_dtor = ga100_fifo_nonstall_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
606
return nvkm_fifo_new_(rm, device, type, inst, pfifo);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
64
nvkm_gsp_rm_free(&chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
66
dma_free_coherent(fifo->engine.subdev.device->dev, fifo->rm.mthdbuf_size,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
67
chan->rm.mthdbuf.ptr, chan->rm.mthdbuf.addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
69
nvkm_cgrp_vctx_put(chan->cgrp, &chan->rm.grctx);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
139
return nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.subdevice, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
159
ret = r535_gr_promote_ctx(gr, false, grc->vmm, grc->mem, grc->vma, &chan->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
278
struct nvkm_rm *rm = gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
301
ret = rm->api->fifo->chan.alloc(&golden.vmm->rm.device, NVKM_RM_CHAN(0),
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
302
1, 0, true, rm->api->fifo->rsvd_chids,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
315
ret = rm->api->gr->get_ctxbufs_info(gr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
325
ret = nvkm_gsp_rm_alloc(&golden.chan, NVKM_RM_THREED, rm->gpu->gr.class.threed, 0, &threed);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
68
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.subdevice,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gr.c
74
ctrl->hChanClient = vmm->rm.client.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1222
gsp->rm->api->gsp->set_rmargs(gsp, resume);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
135
const struct nvkm_rm_api *rmapi = gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1392
gsp->debugfs.rm = create_debugfs(gsp, "logrm", &gsp->blob_rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1393
if (!gsp->debugfs.rm)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1413
i_size_write(d_inode(gsp->debugfs.rm), gsp->blob_rm.size);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1726
struct nvkm_rm *rm = gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1730
u32 len = rm->api->gsp->sr_data_size(gsp);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1751
ret = rm->api->fbsr->suspend(gsp, suspend == NVKM_RUNTIME_SUSPEND);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1798
gsp->rm->api->fbsr->resume(gsp);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1851
const u8 *img = gsp->fws.rm->data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2018
debugfs_remove(gsp->debugfs.rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
2134
const struct nvkm_rm_api *rmapi = gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
296
const struct nvkm_rm_api *rmapi = gsp->rm->api;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
100
if (WARN_ON(vmm->rm.rsvd->addr != SPLIT_VAS_SERVER_RM_MANAGED_VA_START ||
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
101
vmm->rm.rsvd->size != SPLIT_VAS_SERVER_RM_MANAGED_VA_SIZE))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
104
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
111
ctrl->virtAddrLo = vmm->rm.rsvd->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
112
ctrl->virtAddrHi = vmm->rm.rsvd->addr + vmm->rm.rsvd->size - 1;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
131
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
135
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
144
ctrl->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
146
ret = nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
148
vmm->rm.external = true;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
171
struct nvkm_mmu_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
174
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
177
rm->dtor = r535_mmu_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
178
rm->dma_bits = hw->dma_bits;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
179
rm->mmu = hw->mmu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
180
rm->mem = hw->mem;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
181
rm->vmm = hw->vmm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
182
rm->kind = hw->kind;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
183
rm->kind_sys = hw->kind_sys;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
184
rm->promote_vmm = r535_mmu_promote_vmm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
186
ret = nvkm_mmu_new_(rm, device, type, inst, pmmu);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
188
kfree(rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
30
if (vmm->rm.external) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
33
ctrl = nvkm_gsp_rm_ctrl_get(&vmm->rm.device.object,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
37
ctrl->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
39
WARN_ON(nvkm_gsp_rm_ctrl_wr(&vmm->rm.device.object, ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
42
vmm->rm.external = false;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
45
nvkm_gsp_rm_free(&vmm->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
46
nvkm_gsp_device_dtor(&vmm->rm.device);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
47
nvkm_gsp_client_dtor(&vmm->rm.client);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
49
nvkm_vmm_put(vmm, &vmm->rm.rsvd);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
59
&vmm->rm.client, &vmm->rm.device);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
63
args = nvkm_gsp_rm_alloc_get(&vmm->rm.device.object, handle, FERMI_VASPACE_A,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
64
sizeof(*args), &vmm->rm.object);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
72
ret = nvkm_gsp_rm_alloc_wr(&vmm->rm.object, args);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/vmm.c
94
&vmm->rm.rsvd);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
104
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
111
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
113
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
137
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
149
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom, cmd, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
157
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
163
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
173
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
181
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
183
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
188
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
197
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
20
args = nvkm_gsp_rm_alloc_get(&disp->rm.object, (oclass << 16) | inst, oclass,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
205
ret = nvkm_gsp_rm_ctrl_push(&disp->rm.objcom, &ctrl, sizeof(*ctrl));
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
209
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
218
ctrl = nvkm_gsp_rm_ctrl_rd(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
225
nvkm_gsp_rm_ctrl_done(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
35
struct nvkm_gsp *gsp = disp->rm.objcom.client->gsp;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
86
ctrl = nvkm_gsp_rm_ctrl_get(&disp->rm.objcom,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/disp.c
95
return nvkm_gsp_rm_ctrl_wr(&disp->rm.objcom, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
134
if (engn->rm.desc == ctrl->constructedFalconsTable[i].engDesc) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
135
engn->rm.size =
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
148
r570_fifo_xlat_rm_engine_type(u32 rm, enum nvkm_subdev_type *ptype, int *p2080)
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
156
switch (rm) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
65
args->hVASpace = vmm->rm.object.handle;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
84
args->mthdbufMem.size = fifo->rm.mthdbuf_size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
119
0x2000 + device->fifo->rm.mthdbuf_size, 0, true,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
133
ret = rm->api->fifo->chan.alloc(&gr->scrubber.vmm->rm.device, KGRAPHICS_SCRUBBER_HANDLE_CHANNEL,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
148
rm->gpu->gr.class.threed, 0, &gr->scrubber.threed);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
59
ctrl = nvkm_gsp_rm_ctrl_get(&gr->scrubber.vmm->rm.device.subdevice,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
67
return nvkm_gsp_rm_ctrl_wr(&gr->scrubber.vmm->rm.device.subdevice, ctrl);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gr.c
98
struct nvkm_rm *rm = gsp->rm;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
107
if (gsp->rm->wpr->offset_set_by_acr) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
198
gsp->rm->api->gsp->set_rmargs(gsp, true);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
267
heap_size = gsp->rm->wpr->os_carveout_size +
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
268
gsp->rm->wpr->base_size +
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
272
return max(heap_size, gsp->rm->wpr->heap_size_min);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
394
.rm.gpu = &tu1xx_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
406
ret = nvkm_gsp_load_fw(gsp, "gsp", fwif->ver, &gsp->fws.rm);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu116.c
42
.rm.gpu = &tu1xx_gpu,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1057
if (vmm->rm.client.gsp)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c
38
if (!vmm->rm.bar2_pdb)
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c
41
nvkm_wr32(device, 0xb830a0, vmm->rm.bar2_pdb >> 8);
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
37
const struct nvkm_rm_gpu *gpu = device->gsp->rm->gpu;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
38
struct nvkm_vfn_func *rm;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
41
if (!(rm = kzalloc_obj(*rm)))
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
44
rm->dtor = r535_vfn_dtor;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
45
rm->intr = &tu102_vfn_intr;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
46
rm->user.addr = 0x030000;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
47
rm->user.size = 0x010000;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
48
rm->user.base.minver = -1;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
49
rm->user.base.maxver = -1;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
50
rm->user.base.oclass = gpu->usermode.class;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
52
ret = nvkm_vfn_new_(rm, device, type, inst, addr, pvfn);
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/r535.c
54
kfree(rm);
drivers/hwmon/asc7621.c
799
#define PREAD(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
801
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
804
#define PWRITE(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
806
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
813
#define PWRITEM(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
815
.priority = pri, .msb = rm, .lsb = rl, .mask = m, .shift = s,}
drivers/isdn/hardware/mISDN/hfcmulti.c
1000
hfcmulti_resync(hc, NULL, rm);
drivers/isdn/hardware/mISDN/hfcmulti.c
882
hfcmulti_resync(struct hfc_multi *locked, struct hfc_multi *newmaster, int rm)
drivers/isdn/hardware/mISDN/hfcmulti.c
972
if (!rm)
drivers/isdn/hardware/mISDN/hfcmulti.c
984
plxsd_checksync(struct hfc_multi *hc, int rm)
drivers/isdn/hardware/mISDN/hfcmulti.c
992
hfcmulti_resync(hc, hc, rm);
drivers/macintosh/rack-meter.c
106
pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1);
drivers/macintosh/rack-meter.c
121
out_le32(rm->i2s_regs + 0x10, 0x01fa0000);
drivers/macintosh/rack-meter.c
122
(void)in_le32(rm->i2s_regs + 0x10);
drivers/macintosh/rack-meter.c
132
static void rackmeter_set_default_pattern(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
138
rm->ubuf[i] = (i & 1) * 255;
drivers/macintosh/rack-meter.c
140
rm->ubuf[i] = ((~i) & 1) * 255;
drivers/macintosh/rack-meter.c
144
static void rackmeter_do_pause(struct rackmeter *rm, int pause)
drivers/macintosh/rack-meter.c
146
struct rackmeter_dma *rdma = rm->dma_buf_v;
drivers/macintosh/rack-meter.c
150
rm->paused = pause;
drivers/macintosh/rack-meter.c
152
DBDMA_DO_STOP(rm->dma_regs);
drivers/macintosh/rack-meter.c
158
rm->dma_buf_v->mark = 0;
drivers/macintosh/rack-meter.c
161
out_le32(&rm->dma_regs->cmdptr_hi, 0);
drivers/macintosh/rack-meter.c
162
out_le32(&rm->dma_regs->cmdptr, rm->dma_buf_p);
drivers/macintosh/rack-meter.c
163
out_le32(&rm->dma_regs->control, (RUN << 16) | RUN);
drivers/macintosh/rack-meter.c
166
static void rackmeter_setup_dbdma(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
168
struct rackmeter_dma *db = rm->dma_buf_v;
drivers/macintosh/rack-meter.c
172
DBDMA_DO_RESET(rm->dma_regs);
drivers/macintosh/rack-meter.c
185
cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
drivers/macintosh/rack-meter.c
192
cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
drivers/macintosh/rack-meter.c
198
cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
drivers/macintosh/rack-meter.c
205
cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
drivers/macintosh/rack-meter.c
207
cmd->cmd_dep = cpu_to_le32(rm->dma_buf_p);
drivers/macintosh/rack-meter.c
209
rackmeter_do_pause(rm, 0);
drivers/macintosh/rack-meter.c
216
struct rackmeter *rm = rcpu->rm;
drivers/macintosh/rack-meter.c
240
rm->ubuf[i + offset] = ub;
drivers/macintosh/rack-meter.c
246
pause = (rm->cpu[0].zero && rm->cpu[1].zero);
drivers/macintosh/rack-meter.c
247
if (pause != rm->paused) {
drivers/macintosh/rack-meter.c
248
mutex_lock(&rm->sem);
drivers/macintosh/rack-meter.c
249
pause = (rm->cpu[0].zero && rm->cpu[1].zero);
drivers/macintosh/rack-meter.c
250
rackmeter_do_pause(rm, pause);
drivers/macintosh/rack-meter.c
251
mutex_unlock(&rm->sem);
drivers/macintosh/rack-meter.c
257
static void rackmeter_init_cpu_sniffer(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
267
rm->cpu[0].rm = rm;
drivers/macintosh/rack-meter.c
268
INIT_DELAYED_WORK(&rm->cpu[0].sniffer, rackmeter_do_timer);
drivers/macintosh/rack-meter.c
269
rm->cpu[1].rm = rm;
drivers/macintosh/rack-meter.c
270
INIT_DELAYED_WORK(&rm->cpu[1].sniffer, rackmeter_do_timer);
drivers/macintosh/rack-meter.c
277
rcpu = &rm->cpu[cpu];
drivers/macintosh/rack-meter.c
280
schedule_delayed_work_on(cpu, &rm->cpu[cpu].sniffer,
drivers/macintosh/rack-meter.c
285
static void rackmeter_stop_cpu_sniffer(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
287
cancel_delayed_work_sync(&rm->cpu[0].sniffer);
drivers/macintosh/rack-meter.c
288
cancel_delayed_work_sync(&rm->cpu[1].sniffer);
drivers/macintosh/rack-meter.c
291
static int rackmeter_setup(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
294
rackmeter_setup_i2s(rm);
drivers/macintosh/rack-meter.c
297
rackmeter_set_default_pattern(rm);
drivers/macintosh/rack-meter.c
300
rackmeter_setup_dbdma(rm);
drivers/macintosh/rack-meter.c
303
rackmeter_init_cpu_sniffer(rm);
drivers/macintosh/rack-meter.c
311
static u32 rackmeter_calc_sample(struct rackmeter *rm, unsigned int index)
drivers/macintosh/rack-meter.c
318
sample |= ((rm->ubuf[led] >= 0x80) << 15);
drivers/macintosh/rack-meter.c
325
struct rackmeter *rm = arg;
drivers/macintosh/rack-meter.c
326
struct rackmeter_dma *db = rm->dma_buf_v;
drivers/macintosh/rack-meter.c
334
(void)in_le32(&rm->dma_regs->status);
drivers/macintosh/rack-meter.c
345
if (++rm->stale_irq > 3) {
drivers/macintosh/rack-meter.c
348
DBDMA_DO_RESET(rm->dma_regs);
drivers/macintosh/rack-meter.c
360
buf[i] = rackmeter_calc_sample(rm, i);
drivers/macintosh/rack-meter.c
370
struct rackmeter *rm = NULL;
drivers/macintosh/rack-meter.c
399
rm = kzalloc_obj(*rm);
drivers/macintosh/rack-meter.c
400
if (rm == NULL) {
drivers/macintosh/rack-meter.c
405
rm->mdev = mdev;
drivers/macintosh/rack-meter.c
406
rm->i2s = i2s;
drivers/macintosh/rack-meter.c
407
mutex_init(&rm->sem);
drivers/macintosh/rack-meter.c
408
dev_set_drvdata(&mdev->ofdev.dev, rm);
drivers/macintosh/rack-meter.c
426
rm->irq = macio_irq(mdev, 1);
drivers/macintosh/rack-meter.c
428
rm->irq = irq_of_parse_and_map(i2s, 1);
drivers/macintosh/rack-meter.c
429
if (!rm->irq ||
drivers/macintosh/rack-meter.c
442
pr_debug(" irq %d\n", rm->irq);
drivers/macintosh/rack-meter.c
444
rm->ubuf = (u8 *)__get_free_page(GFP_KERNEL);
drivers/macintosh/rack-meter.c
445
if (rm->ubuf == NULL) {
drivers/macintosh/rack-meter.c
452
rm->dma_buf_v = dma_alloc_coherent(&macio_get_pci_dev(mdev)->dev,
drivers/macintosh/rack-meter.c
454
&rm->dma_buf_p, GFP_KERNEL);
drivers/macintosh/rack-meter.c
455
if (rm->dma_buf_v == NULL) {
drivers/macintosh/rack-meter.c
462
rm->i2s_regs = ioremap(macio_resource_start(mdev, 0), 0x1000);
drivers/macintosh/rack-meter.c
464
rm->i2s_regs = ioremap(ri2s.start, 0x1000);
drivers/macintosh/rack-meter.c
466
if (rm->i2s_regs == NULL) {
drivers/macintosh/rack-meter.c
473
rm->dma_regs = ioremap(macio_resource_start(mdev, 1), 0x100);
drivers/macintosh/rack-meter.c
475
rm->dma_regs = ioremap(rdma.start, 0x100);
drivers/macintosh/rack-meter.c
477
if (rm->dma_regs == NULL) {
drivers/macintosh/rack-meter.c
484
rc = rackmeter_setup(rm);
drivers/macintosh/rack-meter.c
492
rc = request_irq(rm->irq, rackmeter_irq, 0, "rackmeter", rm);
drivers/macintosh/rack-meter.c
502
DBDMA_DO_RESET(rm->dma_regs);
drivers/macintosh/rack-meter.c
504
iounmap(rm->dma_regs);
drivers/macintosh/rack-meter.c
506
iounmap(rm->i2s_regs);
drivers/macintosh/rack-meter.c
51
struct rackmeter *rm;
drivers/macintosh/rack-meter.c
510
rm->dma_buf_v, rm->dma_buf_p);
drivers/macintosh/rack-meter.c
512
free_page((unsigned long)rm->ubuf);
drivers/macintosh/rack-meter.c
518
kfree(rm);
drivers/macintosh/rack-meter.c
528
struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev);
drivers/macintosh/rack-meter.c
531
rackmeter_stop_cpu_sniffer(rm);
drivers/macintosh/rack-meter.c
537
DBDMA_DO_RESET(rm->dma_regs);
drivers/macintosh/rack-meter.c
540
free_irq(rm->irq, rm);
drivers/macintosh/rack-meter.c
543
iounmap(rm->dma_regs);
drivers/macintosh/rack-meter.c
544
iounmap(rm->i2s_regs);
drivers/macintosh/rack-meter.c
549
rm->dma_buf_v, rm->dma_buf_p);
drivers/macintosh/rack-meter.c
552
free_page((unsigned long)rm->ubuf);
drivers/macintosh/rack-meter.c
560
kfree(rm);
drivers/macintosh/rack-meter.c
565
struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev);
drivers/macintosh/rack-meter.c
567
if (rm == NULL)
drivers/macintosh/rack-meter.c
571
rackmeter_stop_cpu_sniffer(rm);
drivers/macintosh/rack-meter.c
574
DBDMA_DO_RESET(rm->dma_regs);
drivers/macintosh/rack-meter.c
95
static void rackmeter_setup_i2s(struct rackmeter *rm)
drivers/macintosh/rack-meter.c
97
struct macio_chip *macio = rm->mdev->bus->chip;
drivers/media/i2c/ov4689.c
418
struct regmap *rm = ov4689->regmap;
drivers/media/i2c/ov4689.c
421
cci_write(rm, OV4689_REG_H_CROP_START, 8, &ret);
drivers/media/i2c/ov4689.c
422
cci_write(rm, OV4689_REG_V_CROP_START, 8, &ret);
drivers/media/i2c/ov4689.c
423
cci_write(rm, OV4689_REG_H_CROP_END, 2711, &ret);
drivers/media/i2c/ov4689.c
424
cci_write(rm, OV4689_REG_V_CROP_END, 1531, &ret);
drivers/media/i2c/ov4689.c
426
cci_write(rm, OV4689_REG_H_OUTPUT_SIZE, mode->width, &ret);
drivers/media/i2c/ov4689.c
427
cci_write(rm, OV4689_REG_V_OUTPUT_SIZE, mode->height, &ret);
drivers/media/i2c/ov4689.c
429
cci_write(rm, OV4689_REG_H_WIN_OFF, 8, &ret);
drivers/media/i2c/ov4689.c
430
cci_write(rm, OV4689_REG_V_WIN_OFF, 4, &ret);
drivers/media/i2c/ov4689.c
432
cci_write(rm, OV4689_REG_VFIFO_CTRL_01, 167, &ret);
drivers/media/i2c/ov4689.c
439
struct regmap *rm = ov4689->regmap;
drivers/media/i2c/ov4689.c
442
cci_write(rm, OV4689_REG_ANCHOR_LEFT_START, 16, &ret);
drivers/media/i2c/ov4689.c
443
cci_write(rm, OV4689_REG_ANCHOR_LEFT_END, 1999, &ret);
drivers/media/i2c/ov4689.c
444
cci_write(rm, OV4689_REG_ANCHOR_RIGHT_START, 2400, &ret);
drivers/media/i2c/ov4689.c
445
cci_write(rm, OV4689_REG_ANCHOR_RIGHT_END, 2415, &ret);
drivers/media/pci/ddbridge/ddbridge-core.c
2243
const struct ddb_regmap *rm = io_regmap(io, 0);
drivers/media/pci/ddbridge/ddbridge-core.c
2253
dma->regs = rm->odma->base + rm->odma->size * nr;
drivers/media/pci/ddbridge/ddbridge-core.c
2254
dma->bufregs = rm->odma_buf->base + rm->odma_buf->size * nr;
drivers/media/pci/ddbridge/ddbridge-core.c
2260
dma->regs = rm->idma->base + rm->idma->size * nr;
drivers/media/pci/ddbridge/ddbridge-core.c
2261
dma->bufregs = rm->idma_buf->base + rm->idma_buf->size * nr;
drivers/media/pci/ddbridge/ddbridge-core.c
2275
const struct ddb_regmap *rm;
drivers/media/pci/ddbridge/ddbridge-core.c
2280
rm = io_regmap(input, 1);
drivers/media/pci/ddbridge/ddbridge-core.c
2282
(rm->input->base + rm->input->size * nr);
drivers/media/pci/ddbridge/ddbridge-core.c
2306
const struct ddb_regmap *rm;
drivers/media/pci/ddbridge/ddbridge-core.c
2311
rm = io_regmap(output, 1);
drivers/media/pci/ddbridge/ddbridge-core.c
2313
(rm->output->base + rm->output->size * nr);
drivers/media/pci/ddbridge/ddbridge-core.c
2361
const struct ddb_regmap *rm;
drivers/media/pci/ddbridge/ddbridge-core.c
2367
rm = info->regmap;
drivers/media/pci/ddbridge/ddbridge-core.c
2368
if (!rm)
drivers/mtd/nand/raw/mpc5121_nfc.c
518
struct mpc512x_reset_module *rm;
drivers/mtd/nand/raw/mpc5121_nfc.c
534
rm = of_iomap(rmnode, 0);
drivers/mtd/nand/raw/mpc5121_nfc.c
535
if (!rm) {
drivers/mtd/nand/raw/mpc5121_nfc.c
541
rcwh = in_be32(&rm->rcwhr);
drivers/mtd/nand/raw/mpc5121_nfc.c
586
iounmap(rm);
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
532
struct pucan_rx_msg *rm = (struct pucan_rx_msg *)rx_msg;
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
537
const u16 rx_msg_flags = le16_to_cpu(rm->flags);
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
539
if (pucan_msg_get_channel(rm) >= ARRAY_SIZE(usb_if->dev))
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
542
dev = usb_if->dev[pucan_msg_get_channel(rm)];
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
557
cfd->len = can_fd_dlc2len(pucan_msg_get_dlc(rm));
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
565
pucan_msg_get_dlc(rm),
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
569
cfd->can_id = le32_to_cpu(rm->can_id);
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
577
memcpy(cfd->data, rm->d, cfd->len);
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
582
peak_usb_netif_rx_64(skb, le32_to_cpu(rm->ts_low),
drivers/net/can/usb/peak_usb/pcan_usb_fd.c
583
le32_to_cpu(rm->ts_high));
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1895
pdata->rm = RM3;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1903
pdata->rm = RM1;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1912
pdata->rm = RM0;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
1970
pdata->rm = RM0;
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
213
enum xgene_enet_rm rm;
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
259
return ((u16)pdata->rm << 10) | ring->num;
drivers/net/ethernet/chelsio/cxgb/common.h
65
#define t1_rx_mode_promisc(rm) (rm->dev->flags & IFF_PROMISC)
drivers/net/ethernet/chelsio/cxgb/common.h
66
#define t1_rx_mode_allmulti(rm) (rm->dev->flags & IFF_ALLMULTI)
drivers/net/ethernet/chelsio/cxgb/common.h
67
#define t1_rx_mode_mc_cnt(rm) (netdev_mc_count(rm->dev))
drivers/net/ethernet/chelsio/cxgb/common.h
68
#define t1_get_netdev(rm) (rm->dev)
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
119
struct t1_rx_mode rm;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
121
rm.dev = dev;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
122
mac->ops->set_rx_mode(mac, &rm);
drivers/net/ethernet/chelsio/cxgb/gmac.h
102
int (*set_rx_mode)(struct cmac *, struct t1_rx_mode *rm);
drivers/net/ethernet/chelsio/cxgb/pm3393.c
336
static int pm3393_set_rx_mode(struct cmac *cmac, struct t1_rx_mode *rm)
drivers/net/ethernet/chelsio/cxgb/pm3393.c
351
if (t1_rx_mode_promisc(rm)) {
drivers/net/ethernet/chelsio/cxgb/pm3393.c
355
if (t1_rx_mode_allmulti(rm)) {
drivers/net/ethernet/chelsio/cxgb/pm3393.c
362
} else if (t1_rx_mode_mc_cnt(rm)) {
drivers/net/ethernet/chelsio/cxgb/pm3393.c
368
netdev_for_each_mc_addr(ha, t1_get_netdev(rm)) {
drivers/net/ethernet/chelsio/cxgb/vsc7326.c
433
static int mac_set_rx_mode(struct cmac *mac, struct t1_rx_mode *rm)
drivers/net/ethernet/chelsio/cxgb/vsc7326.c
441
if (t1_rx_mode_promisc(rm))
drivers/net/ethernet/intel/ice/ice_switch.c
4880
ice_fill_fv_word_index(struct ice_hw *hw, struct ice_sw_recipe *rm)
drivers/net/ethernet/intel/ice/ice_switch.c
4886
if (list_empty(&rm->fv_list))
drivers/net/ethernet/intel/ice/ice_switch.c
4889
fv = list_first_entry(&rm->fv_list, struct ice_sw_fv_list_entry,
drivers/net/ethernet/intel/ice/ice_switch.c
4894
rm->fv_idx[0] = ICE_AQ_SW_ID_LKUP_IDX;
drivers/net/ethernet/intel/ice/ice_switch.c
4895
rm->fv_mask[0] = ICE_AQ_SW_ID_LKUP_MASK;
drivers/net/ethernet/intel/ice/ice_switch.c
4896
rm->n_ext_words++;
drivers/net/ethernet/intel/ice/ice_switch.c
4898
for (i = 1; i < rm->n_ext_words; i++) {
drivers/net/ethernet/intel/ice/ice_switch.c
4899
struct ice_fv_word *fv_word = &rm->ext_words[i - 1];
drivers/net/ethernet/intel/ice/ice_switch.c
4900
u16 fv_mask = rm->word_masks[i - 1];
drivers/net/ethernet/intel/ice/ice_switch.c
4910
rm->fv_idx[i] = j;
drivers/net/ethernet/intel/ice/ice_switch.c
4911
rm->fv_mask[i] = fv_mask;
drivers/net/ethernet/intel/ice/ice_switch.c
5021
const struct ice_sw_recipe *rm)
drivers/net/ethernet/intel/ice/ice_switch.c
5034
recp->content.act_ctrl_fwd_priority = rm->priority;
drivers/net/ethernet/intel/ice/ice_switch.c
5036
if (rm->need_pass_l2)
drivers/net/ethernet/intel/ice/ice_switch.c
5039
if (rm->allow_pass_l2)
drivers/net/ethernet/intel/ice/ice_switch.c
5045
const struct ice_sw_recipe *rm)
drivers/net/ethernet/intel/ice/ice_switch.c
5050
recipe->tun_type = rm->tun_type;
drivers/net/ethernet/intel/ice/ice_switch.c
5051
recipe->need_pass_l2 = rm->need_pass_l2;
drivers/net/ethernet/intel/ice/ice_switch.c
5052
recipe->allow_pass_l2 = rm->allow_pass_l2;
drivers/net/ethernet/intel/ice/ice_switch.c
5067
ice_add_sw_recipe(struct ice_hw *hw, struct ice_sw_recipe *rm,
drivers/net/ethernet/intel/ice/ice_switch.c
5081
recp_cnt = ice_calc_recp_cnt(rm->n_ext_words);
drivers/net/ethernet/intel/ice/ice_switch.c
5084
bitmap_zero(rm->r_bitmap, ICE_MAX_NUM_RECIPES);
drivers/net/ethernet/intel/ice/ice_switch.c
5117
fill_recipe_template(&buf[i], rid, rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5135
set_bit(rid, rm->r_bitmap);
drivers/net/ethernet/intel/ice/ice_switch.c
5138
while (lookup < rm->n_ext_words &&
drivers/net/ethernet/intel/ice/ice_switch.c
5140
content->lkup_indx[word] = rm->fv_idx[lookup];
drivers/net/ethernet/intel/ice/ice_switch.c
5141
content->mask[word] = cpu_to_le16(rm->fv_mask[lookup]);
drivers/net/ethernet/intel/ice/ice_switch.c
5149
bookkeep_recipe(recipe, &buf[i], rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5159
fill_recipe_template(root, rid, rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5162
set_bit(rid, rm->r_bitmap);
drivers/net/ethernet/intel/ice/ice_switch.c
5163
memcpy(root->recipe_bitmap, rm->r_bitmap, sizeof(root->recipe_bitmap));
drivers/net/ethernet/intel/ice/ice_switch.c
5172
while (lookup < rm->n_ext_words &&
drivers/net/ethernet/intel/ice/ice_switch.c
5174
root->content.lkup_indx[word] = rm->fv_idx[lookup];
drivers/net/ethernet/intel/ice/ice_switch.c
5175
root->content.mask[word] = cpu_to_le16(rm->fv_mask[lookup]);
drivers/net/ethernet/intel/ice/ice_switch.c
5197
rm->root_rid = rid;
drivers/net/ethernet/intel/ice/ice_switch.c
5198
bookkeep_recipe(&hw->switch_info->recp_list[rid], root, rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5319
struct ice_sw_recipe *rm;
drivers/net/ethernet/intel/ice/ice_switch.c
5349
rm = kzalloc_obj(*rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5350
if (!rm) {
drivers/net/ethernet/intel/ice/ice_switch.c
5358
INIT_LIST_HEAD(&rm->fv_list);
drivers/net/ethernet/intel/ice/ice_switch.c
5366
status = ice_get_sw_fv_list(hw, lkup_exts, fv_bitmap, &rm->fv_list);
drivers/net/ethernet/intel/ice/ice_switch.c
5371
rm->n_ext_words = lkup_exts->n_val_words;
drivers/net/ethernet/intel/ice/ice_switch.c
5372
memcpy(rm->ext_words, lkup_exts->fv_words, sizeof(rm->ext_words));
drivers/net/ethernet/intel/ice/ice_switch.c
5373
memcpy(rm->word_masks, lkup_exts->field_mask, sizeof(rm->word_masks));
drivers/net/ethernet/intel/ice/ice_switch.c
5376
rm->priority = (u8)rinfo->priority;
drivers/net/ethernet/intel/ice/ice_switch.c
5378
rm->need_pass_l2 = rinfo->need_pass_l2;
drivers/net/ethernet/intel/ice/ice_switch.c
5379
rm->allow_pass_l2 = rinfo->allow_pass_l2;
drivers/net/ethernet/intel/ice/ice_switch.c
5384
status = ice_fill_fv_word_index(hw, rm);
drivers/net/ethernet/intel/ice/ice_switch.c
5390
list_for_each_entry(fvit, &rm->fv_list, list_entry) {
drivers/net/ethernet/intel/ice/ice_switch.c
5405
rm->tun_type = rinfo->tun_type;
drivers/net/ethernet/intel/ice/ice_switch.c
5407
status = ice_add_sw_recipe(hw, rm, profiles);
drivers/net/ethernet/intel/ice/ice_switch.c
5414
list_for_each_entry(fvit, &rm->fv_list, list_entry) {
drivers/net/ethernet/intel/ice/ice_switch.c
5425
bitmap_or(r_bitmap, r_bitmap, rm->r_bitmap,
drivers/net/ethernet/intel/ice/ice_switch.c
5444
for_each_set_bit(j, rm->r_bitmap, ICE_MAX_NUM_RECIPES)
drivers/net/ethernet/intel/ice/ice_switch.c
5448
*rid = rm->root_rid;
drivers/net/ethernet/intel/ice/ice_switch.c
5455
for_each_set_bit(rid_tmp, rm->r_bitmap, ICE_MAX_NUM_RECIPES) {
drivers/net/ethernet/intel/ice/ice_switch.c
5457
clear_bit(rid_tmp, rm->r_bitmap);
drivers/net/ethernet/intel/ice/ice_switch.c
5462
list_for_each_entry_safe(fvit, tmp, &rm->fv_list, list_entry) {
drivers/net/ethernet/intel/ice/ice_switch.c
5467
kfree(rm);
drivers/net/ethernet/intel/ice/virt/rss.c
932
u16 rm = actions[i].remove_list[j];
drivers/net/ethernet/intel/ice/virt/rss.c
934
ret = ice_hash_remove(vf, &ctx->ctx[rm]);
drivers/parport/parport_gsc.h
136
const unsigned char rm = (PARPORT_CONTROL_STROBE |
drivers/parport/parport_gsc.h
141
return priv->ctr & rm; /* Use soft copy */
drivers/parport/parport_ip32.c
937
const unsigned int rm =
drivers/parport/parport_ip32.c
939
return __parport_ip32_read_control(p) & rm;
drivers/pinctrl/pinctrl-st.c
1050
struct regmap *rm = info->regmap;
drivers/pinctrl/pinctrl-st.c
1065
rt_p->clk1notclk0 = devm_regmap_field_alloc(dev, rm, clk1notclk0);
drivers/pinctrl/pinctrl-st.c
1066
rt_p->delay_0 = devm_regmap_field_alloc(dev, rm, delay_0);
drivers/pinctrl/pinctrl-st.c
1067
rt_p->delay_1 = devm_regmap_field_alloc(dev, rm, delay_1);
drivers/pinctrl/pinctrl-st.c
1068
rt_p->invertclk = devm_regmap_field_alloc(dev, rm, invertclk);
drivers/pinctrl/pinctrl-st.c
1069
rt_p->retime = devm_regmap_field_alloc(dev, rm, retime);
drivers/pinctrl/pinctrl-st.c
1070
rt_p->clknotdata = devm_regmap_field_alloc(dev, rm, clknotdata);
drivers/pinctrl/pinctrl-st.c
1071
rt_p->double_edge = devm_regmap_field_alloc(dev, rm, double_edge);
drivers/pinctrl/pinctrl-st.c
1086
struct regmap *rm = info->regmap;
drivers/pinctrl/pinctrl-st.c
1097
rt_d->rt[j] = devm_regmap_field_alloc(dev, rm, reg);
drivers/pinctrl/stm32/pinctrl-stm32.c
1765
struct regmap *rm;
drivers/pinctrl/stm32/pinctrl-stm32.c
1773
rm = pctl->regmap;
drivers/pinctrl/stm32/pinctrl-stm32.c
1795
pctl->irqmux[i] = devm_regmap_field_alloc(dev, rm, mux);
drivers/pinctrl/tegra/pinctrl-tegra20.c
2127
MUX_PG(rm, I2C1, RSVD2, RSVD3, RSVD4, 0x14, 25, 0x80, 14, 0xa4, 0),
drivers/s390/net/ctcm_mpc.c
131
__u32 ct, sw, rm, dup;
drivers/s390/net/ctcm_mpc.c
141
rm = 16;
drivers/s390/net/ctcm_mpc.c
169
rm--;
drivers/s390/net/ctcm_mpc.c
187
rm = 16;
drivers/s390/net/ctcm_mpc.c
191
for ( ; rm > 0; rm--, sw++) {
drivers/video/fbdev/pm3fb.c
289
unsigned long rm = 1 | (3 << 7);
drivers/video/fbdev/pm3fb.c
295
rm |= 3 << 15;
drivers/video/fbdev/pm3fb.c
302
rm |= 2 << 15;
drivers/video/fbdev/pm3fb.c
314
PM3_WRITE_REG(par, PM3RasterizerMode, rm);
drivers/video/sticore.c
937
u32 *rm;
drivers/video/sticore.c
954
rm = (u32*) &sti->rm_entry;
drivers/video/sticore.c
955
*rm++ = gsc_readl(address+rm_offset+0x00);
drivers/video/sticore.c
956
*rm++ = gsc_readl(address+rm_offset+0x04);
drivers/video/sticore.c
957
*rm++ = gsc_readl(address+rm_offset+0x08);
drivers/video/sticore.c
958
*rm++ = gsc_readl(address+rm_offset+0x0c);
fs/ocfs2/journal.c
1221
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/journal.c
1224
empty = (rm->rm_used == 0);
fs/ocfs2/journal.c
1463
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/journal.c
1514
while (rm->rm_used) {
fs/ocfs2/journal.c
1517
node_num = rm->rm_entries[0];
fs/ocfs2/journal.c
172
struct ocfs2_recovery_map *rm;
fs/ocfs2/journal.c
179
rm = kzalloc_flex(*rm, rm_entries, osb->max_slots);
fs/ocfs2/journal.c
180
if (!rm) {
fs/ocfs2/journal.c
185
osb->recovery_map = rm;
fs/ocfs2/journal.c
232
struct ocfs2_recovery_map *rm;
fs/ocfs2/journal.c
242
rm = osb->recovery_map;
fs/ocfs2/journal.c
245
kfree(rm);
fs/ocfs2/journal.c
252
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/journal.c
256
for (i = 0; i < rm->rm_used; i++) {
fs/ocfs2/journal.c
257
if (rm->rm_entries[i] == node_num)
fs/ocfs2/journal.c
268
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/journal.c
277
BUG_ON(rm->rm_used >= osb->max_slots);
fs/ocfs2/journal.c
279
rm->rm_entries[rm->rm_used] = node_num;
fs/ocfs2/journal.c
280
rm->rm_used++;
fs/ocfs2/journal.c
290
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/journal.c
294
for (i = 0; i < rm->rm_used; i++) {
fs/ocfs2/journal.c
295
if (rm->rm_entries[i] == node_num)
fs/ocfs2/journal.c
299
if (i < rm->rm_used) {
fs/ocfs2/journal.c
301
memmove(&(rm->rm_entries[i]), &(rm->rm_entries[i + 1]),
fs/ocfs2/journal.c
302
(rm->rm_used - i - 1) * sizeof(unsigned int));
fs/ocfs2/journal.c
303
rm->rm_used--;
fs/ocfs2/super.c
219
struct ocfs2_recovery_map *rm = osb->recovery_map;
fs/ocfs2/super.c
272
if (rm->rm_used == 0)
fs/ocfs2/super.c
275
for (i = 0; i < rm->rm_used; i++)
fs/ocfs2/super.c
277
rm->rm_entries[i]);
fs/xfs/scrub/refcount.c
148
memcpy(&frag->rm, rec, sizeof(frag->rm));
fs/xfs/scrub/refcount.c
193
if (frag->rm.rm_startblock < bno)
fs/xfs/scrub/refcount.c
195
bno = frag->rm.rm_startblock;
fs/xfs/scrub/refcount.c
204
if (frag->rm.rm_startblock > refchk->bno || nr > target_nr)
fs/xfs/scrub/refcount.c
206
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
225
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
238
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
239
if (frag->rm.rm_startblock != rbno)
fs/xfs/scrub/refcount.c
85
struct xfs_rmap_irec rm;
fs/xfs/scrub/rtrefcount.c
162
memcpy(&frag->rm, rec, sizeof(frag->rm));
fs/xfs/scrub/rtrefcount.c
207
if (frag->rm.rm_startblock < bno)
fs/xfs/scrub/rtrefcount.c
209
bno = frag->rm.rm_startblock;
fs/xfs/scrub/rtrefcount.c
218
if (frag->rm.rm_startblock > refchk->bno || nr > target_nr)
fs/xfs/scrub/rtrefcount.c
220
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
239
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
252
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
253
if (frag->rm.rm_startblock != rbno)
fs/xfs/scrub/rtrefcount.c
99
struct xfs_rmap_irec rm;
include/linux/math64.h
215
} rl, rm, rn, rh, a0, b0;
include/linux/math64.h
222
rm.ll = mul_u32_u32(a0.l.low, b0.l.high);
include/linux/math64.h
231
rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low;
include/linux/math64.h
232
rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low;
include/linux/parport_pc.h
179
const unsigned char rm = (PARPORT_CONTROL_STROBE |
include/linux/parport_pc.h
184
return priv->ctr & rm; /* Use soft copy */
include/linux/transport_class.h
27
#define DECLARE_TRANSPORT_CLASS(cls, nm, su, rm, cfg) \
include/linux/transport_class.h
33
.remove = rm, \
net/ipv4/ipconfig.c
412
struct rtentry rm;
net/ipv4/ipconfig.c
415
memset(&rm, 0, sizeof(rm));
net/ipv4/ipconfig.c
420
set_sockaddr((struct sockaddr_in *) &rm.rt_dst, 0, 0);
net/ipv4/ipconfig.c
421
set_sockaddr((struct sockaddr_in *) &rm.rt_genmask, 0, 0);
net/ipv4/ipconfig.c
422
set_sockaddr((struct sockaddr_in *) &rm.rt_gateway, ic_gateway, 0);
net/ipv4/ipconfig.c
423
rm.rt_flags = RTF_UP | RTF_GATEWAY;
net/ipv4/ipconfig.c
424
if ((err = ip_rt_ioctl(&init_net, SIOCADDRT, &rm)) < 0) {
net/rds/cong.c
421
struct rds_message *rm;
net/rds/cong.c
423
rm = rds_message_map_pages(map->m_page_addrs, RDS_CONG_MAP_BYTES);
net/rds/cong.c
424
if (!IS_ERR(rm))
net/rds/cong.c
425
rm->m_inc.i_hdr.h_flags = RDS_FLAG_CONG_BITMAP;
net/rds/cong.c
427
return rm;
net/rds/connection.c
470
struct rds_message *rm, *rtmp;
net/rds/connection.c
483
list_for_each_entry_safe(rm, rtmp,
net/rds/connection.c
486
list_del_init(&rm->m_conn_item);
net/rds/connection.c
487
BUG_ON(!list_empty(&rm->m_sock_item));
net/rds/connection.c
488
rds_message_put(rm);
net/rds/connection.c
574
struct rds_message *rm;
net/rds/connection.c
609
list_for_each_entry(rm, list, m_conn_item) {
net/rds/connection.c
612
__rds_inc_msg_cp(&rm->m_inc,
net/rds/ib.h
424
int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm,
net/rds/ib_cm.c
1163
struct rds_message *rm;
net/rds/ib_cm.c
1165
rm = container_of(ic->i_data_op, struct rds_message, data);
net/rds/ib_cm.c
1166
rds_message_put(rm);
net/rds/ib_send.c
154
struct rds_message *rm = NULL;
net/rds/ib_send.c
160
rm = container_of(send->s_op, struct rds_message, data);
net/rds/ib_send.c
167
rm = container_of(send->s_op, struct rds_message, rdma);
net/rds/ib_send.c
174
rm = container_of(send->s_op, struct rds_message, atomic);
net/rds/ib_send.c
187
return rm;
net/rds/ib_send.c
245
struct rds_message *rm = NULL;
net/rds/ib_send.c
276
rm = rds_ib_send_unmap_op(ic, send, wc->status);
net/rds/ib_send.c
282
if (send->s_op == rm->m_final_op) {
net/rds/ib_send.c
286
rds_message_unmapped(rm);
net/rds/ib_send.c
288
rds_message_put(rm);
net/rds/ib_send.c
48
static void rds_ib_send_complete(struct rds_message *rm,
net/rds/ib_send.c
484
int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm,
net/rds/ib_send.c
50
void (*complete)(struct rds_message *rm, int status))
net/rds/ib_send.c
511
&& rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) {
net/rds/ib_send.c
513
scat = &rm->data.op_sg[sg];
net/rds/ib_send.c
519
if (be32_to_cpu(rm->m_inc.i_hdr.h_len) == 0)
net/rds/ib_send.c
522
i = DIV_ROUND_UP(be32_to_cpu(rm->m_inc.i_hdr.h_len), RDS_FRAG_SIZE);
net/rds/ib_send.c
550
if (rm->data.op_nents) {
net/rds/ib_send.c
551
rm->data.op_count = ib_dma_map_sg(dev,
net/rds/ib_send.c
552
rm->data.op_sg,
net/rds/ib_send.c
553
rm->data.op_nents,
net/rds/ib_send.c
555
rdsdebug("ic %p mapping rm %p: %d\n", ic, rm, rm->data.op_count);
net/rds/ib_send.c
556
if (rm->data.op_count == 0) {
net/rds/ib_send.c
563
rm->data.op_count = 0;
net/rds/ib_send.c
566
rds_message_addref(rm);
net/rds/ib_send.c
567
rm->data.op_dmasg = 0;
net/rds/ib_send.c
568
rm->data.op_dmaoff = 0;
net/rds/ib_send.c
569
ic->i_data_op = &rm->data;
net/rds/ib_send.c
572
if (test_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags))
net/rds/ib_send.c
573
rm->m_inc.i_hdr.h_flags |= RDS_FLAG_ACK_REQUIRED;
net/rds/ib_send.c
574
if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))
net/rds/ib_send.c
575
rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED;
net/rds/ib_send.c
579
if (rm->rdma.op_active) {
net/rds/ib_send.c
584
ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey);
net/rds/ib_send.c
585
if (rds_message_add_extension(&rm->m_inc.i_hdr,
net/rds/ib_send.c
590
rm->rdma.op_write ?
net/rds/ib_send.c
594
cpu_to_be32(rm->rdma.op_bytes);
net/rds/ib_send.c
599
if (rds_message_add_extension(&rm->m_inc.i_hdr,
net/rds/ib_send.c
605
rm->m_inc.i_hdr.h_flags |=
net/rds/ib_send.c
611
if (rm->m_rdma_cookie &&
net/rds/ib_send.c
612
!rds_message_add_rdma_dest_extension(&rm->m_inc.i_hdr,
net/rds/ib_send.c
613
rds_rdma_cookie_key(rm->m_rdma_cookie),
net/rds/ib_send.c
614
rds_rdma_cookie_offset(rm->m_rdma_cookie))) {
net/rds/ib_send.c
622
rm->m_inc.i_hdr.h_ack = cpu_to_be64(rds_ib_piggyb_ack(ic));
net/rds/ib_send.c
623
rds_message_make_checksum(&rm->m_inc.i_hdr);
net/rds/ib_send.c
641
if (rm->rdma.op_active && rm->rdma.op_fence)
net/rds/ib_send.c
648
scat = &ic->i_data_op->op_sg[rm->data.op_dmasg];
net/rds/ib_send.c
670
memcpy(ic->i_send_hdrs[pos], &rm->m_inc.i_hdr,
net/rds/ib_send.c
676
&& scat != &rm->data.op_sg[rm->data.op_count]) {
net/rds/ib_send.c
678
sg_dma_len(scat) - rm->data.op_dmaoff);
net/rds/ib_send.c
682
send->s_sge[1].addr += rm->data.op_dmaoff;
net/rds/ib_send.c
687
rm->data.op_dmaoff += len;
net/rds/ib_send.c
688
if (rm->data.op_dmaoff == sg_dma_len(scat)) {
net/rds/ib_send.c
690
rm->data.op_dmasg++;
net/rds/ib_send.c
691
rm->data.op_dmaoff = 0;
net/rds/ib_send.c
70
complete(rm, notify_status);
net/rds/ib_send.c
734
&& scat != &rm->data.op_sg[rm->data.op_count]);
net/rds/ib_send.c
742
if (scat == &rm->data.op_sg[rm->data.op_count]) {
net/rds/loop.c
102
rds_inc_put(&rm->m_inc);
net/rds/loop.c
113
struct rds_message *rm = container_of(inc, struct rds_message, m_inc);
net/rds/loop.c
115
rds_message_put(rm);
net/rds/loop.c
75
static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm,
net/rds/loop.c
79
struct scatterlist *sgp = &rm->data.op_sg[sg];
net/rds/loop.c
81
be32_to_cpu(rm->m_inc.i_hdr.h_len);
net/rds/loop.c
84
if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) {
net/rds/loop.c
92
rds_inc_init(&rm->m_inc, conn, &conn->c_laddr);
net/rds/loop.c
94
rds_message_addref(rm);
net/rds/loop.c
96
rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc,
net/rds/loop.c
99
rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence),
net/rds/message.c
132
static void rds_message_purge(struct rds_message *rm)
net/rds/message.c
137
if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags)))
net/rds/message.c
140
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/message.c
141
if (rm->m_rs) {
net/rds/message.c
142
struct rds_sock *rs = rm->m_rs;
net/rds/message.c
144
if (rm->data.op_mmp_znotifier) {
net/rds/message.c
146
rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier);
net/rds/message.c
148
rm->data.op_mmp_znotifier = NULL;
net/rds/message.c
151
rm->m_rs = NULL;
net/rds/message.c
153
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/message.c
155
for (i = 0; i < rm->data.op_nents; i++) {
net/rds/message.c
158
__free_page(sg_page(&rm->data.op_sg[i]));
net/rds/message.c
160
put_page(sg_page(&rm->data.op_sg[i]));
net/rds/message.c
162
rm->data.op_nents = 0;
net/rds/message.c
164
if (rm->rdma.op_active)
net/rds/message.c
165
rds_rdma_free_op(&rm->rdma);
net/rds/message.c
166
if (rm->rdma.op_rdma_mr)
net/rds/message.c
167
kref_put(&rm->rdma.op_rdma_mr->r_kref, __rds_put_mr_final);
net/rds/message.c
169
if (rm->atomic.op_active)
net/rds/message.c
170
rds_atomic_free_op(&rm->atomic);
net/rds/message.c
171
if (rm->atomic.op_rdma_mr)
net/rds/message.c
172
kref_put(&rm->atomic.op_rdma_mr->r_kref, __rds_put_mr_final);
net/rds/message.c
175
void rds_message_put(struct rds_message *rm)
net/rds/message.c
177
rdsdebug("put rm %p ref %d\n", rm, refcount_read(&rm->m_refcount));
net/rds/message.c
178
WARN(!refcount_read(&rm->m_refcount), "danger refcount zero on %p\n", rm);
net/rds/message.c
179
if (refcount_dec_and_test(&rm->m_refcount)) {
net/rds/message.c
180
BUG_ON(!list_empty(&rm->m_sock_item));
net/rds/message.c
181
BUG_ON(!list_empty(&rm->m_conn_item));
net/rds/message.c
182
rds_message_purge(rm);
net/rds/message.c
184
kfree(rm);
net/rds/message.c
326
struct rds_message *rm;
net/rds/message.c
331
rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp);
net/rds/message.c
332
if (!rm)
net/rds/message.c
335
rm->m_used_sgs = 0;
net/rds/message.c
336
rm->m_total_sgs = extra_len / sizeof(struct scatterlist);
net/rds/message.c
338
refcount_set(&rm->m_refcount, 1);
net/rds/message.c
339
INIT_LIST_HEAD(&rm->m_sock_item);
net/rds/message.c
340
INIT_LIST_HEAD(&rm->m_conn_item);
net/rds/message.c
341
spin_lock_init(&rm->m_rs_lock);
net/rds/message.c
342
init_waitqueue_head(&rm->m_flush_wait);
net/rds/message.c
345
return rm;
net/rds/message.c
351
struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents)
net/rds/message.c
353
struct scatterlist *sg_first = (struct scatterlist *) &rm[1];
net/rds/message.c
361
if (rm->m_used_sgs + nents > rm->m_total_sgs) {
net/rds/message.c
363
rm->m_total_sgs, rm->m_used_sgs, nents);
net/rds/message.c
367
sg_ret = &sg_first[rm->m_used_sgs];
net/rds/message.c
369
rm->m_used_sgs += nents;
net/rds/message.c
376
struct rds_message *rm;
net/rds/message.c
381
rm = rds_message_alloc(extra_bytes, GFP_NOWAIT);
net/rds/message.c
382
if (!rm)
net/rds/message.c
385
set_bit(RDS_MSG_PAGEVEC, &rm->m_flags);
net/rds/message.c
386
rm->m_inc.i_hdr.h_len = cpu_to_be32(total_len);
net/rds/message.c
387
rm->data.op_nents = DIV_ROUND_UP(total_len, PAGE_SIZE);
net/rds/message.c
388
rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs);
net/rds/message.c
389
if (IS_ERR(rm->data.op_sg)) {
net/rds/message.c
390
void *err = ERR_CAST(rm->data.op_sg);
net/rds/message.c
391
rds_message_put(rm);
net/rds/message.c
395
for (i = 0; i < rm->data.op_nents; ++i) {
net/rds/message.c
396
sg_set_page(&rm->data.op_sg[i],
net/rds/message.c
401
return rm;
net/rds/message.c
404
static int rds_message_zcopy_from_user(struct rds_message *rm, struct iov_iter *from)
net/rds/message.c
411
rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from));
net/rds/message.c
416
sg = rm->data.op_sg;
net/rds/message.c
422
rm->data.op_mmp_znotifier = &info->znotif;
net/rds/message.c
423
if (mm_account_pinned_pages(&rm->data.op_mmp_znotifier->z_mmp,
net/rds/message.c
439
for (i = 0; i < rm->data.op_nents; i++)
net/rds/message.c
440
put_page(sg_page(&rm->data.op_sg[i]));
net/rds/message.c
441
mmp = &rm->data.op_mmp_znotifier->z_mmp;
net/rds/message.c
448
rm->data.op_nents++;
net/rds/message.c
455
rm->data.op_mmp_znotifier = NULL;
net/rds/message.c
459
int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from,
net/rds/message.c
467
rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from));
net/rds/message.c
470
sg = rm->data.op_sg;
net/rds/message.c
474
return rds_message_zcopy_from_user(rm, from);
net/rds/message.c
482
rm->data.op_nents++;
net/rds/message.c
506
struct rds_message *rm;
net/rds/message.c
514
rm = container_of(inc, struct rds_message, m_inc);
net/rds/message.c
515
len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
net/rds/message.c
517
sg = rm->data.op_sg;
net/rds/message.c
53
void rds_message_addref(struct rds_message *rm)
net/rds/message.c
548
void rds_message_wait(struct rds_message *rm)
net/rds/message.c
55
rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount));
net/rds/message.c
550
wait_event_interruptible(rm->m_flush_wait,
net/rds/message.c
551
!test_bit(RDS_MSG_MAPPED, &rm->m_flags));
net/rds/message.c
554
void rds_message_unmapped(struct rds_message *rm)
net/rds/message.c
556
clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
net/rds/message.c
557
wake_up_interruptible(&rm->m_flush_wait);
net/rds/message.c
56
refcount_inc(&rm->m_refcount);
net/rds/rdma.c
609
int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
net/rds/rdma.c
614
struct rm_rdma_op *op = &rm->rdma;
net/rds/rdma.c
624
|| rm->rdma.op_active)
net/rds/rdma.c
670
op->op_sg = rds_message_alloc_sgs(rm, nr_pages);
net/rds/rdma.c
808
int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
net/rds/rdma.c
817
rm->m_rdma_cookie != 0)
net/rds/rdma.c
820
memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie));
net/rds/rdma.c
827
r_key = rds_rdma_cookie_key(rm->m_rdma_cookie);
net/rds/rdma.c
840
rm->rdma.op_rdma_mr = mr;
net/rds/rdma.c
851
int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
net/rds/rdma.c
855
rm->m_rdma_cookie != 0)
net/rds/rdma.c
858
return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie,
net/rds/rdma.c
859
&rm->rdma.op_rdma_mr, rm->m_conn_path);
net/rds/rdma.c
865
int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
net/rds/rdma.c
873
|| rm->atomic.op_active)
net/rds/rdma.c
881
rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD;
net/rds/rdma.c
882
rm->atomic.op_m_fadd.add = args->fadd.add;
net/rds/rdma.c
883
rm->atomic.op_m_fadd.nocarry_mask = 0;
net/rds/rdma.c
886
rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD;
net/rds/rdma.c
887
rm->atomic.op_m_fadd.add = args->m_fadd.add;
net/rds/rdma.c
888
rm->atomic.op_m_fadd.nocarry_mask = args->m_fadd.nocarry_mask;
net/rds/rdma.c
891
rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP;
net/rds/rdma.c
892
rm->atomic.op_m_cswp.compare = args->cswp.compare;
net/rds/rdma.c
893
rm->atomic.op_m_cswp.swap = args->cswp.swap;
net/rds/rdma.c
894
rm->atomic.op_m_cswp.compare_mask = ~0;
net/rds/rdma.c
895
rm->atomic.op_m_cswp.swap_mask = ~0;
net/rds/rdma.c
898
rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP;
net/rds/rdma.c
899
rm->atomic.op_m_cswp.compare = args->m_cswp.compare;
net/rds/rdma.c
900
rm->atomic.op_m_cswp.swap = args->m_cswp.swap;
net/rds/rdma.c
901
rm->atomic.op_m_cswp.compare_mask = args->m_cswp.compare_mask;
net/rds/rdma.c
902
rm->atomic.op_m_cswp.swap_mask = args->m_cswp.swap_mask;
net/rds/rdma.c
908
rm->atomic.op_notify = !!(args->flags & RDS_RDMA_NOTIFY_ME);
net/rds/rdma.c
909
rm->atomic.op_silent = !!(args->flags & RDS_RDMA_SILENT);
net/rds/rdma.c
910
rm->atomic.op_active = 1;
net/rds/rdma.c
911
rm->atomic.op_recverr = rs->rs_recverr;
net/rds/rdma.c
912
rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1);
net/rds/rdma.c
913
if (IS_ERR(rm->atomic.op_sg)) {
net/rds/rdma.c
914
ret = PTR_ERR(rm->atomic.op_sg);
net/rds/rdma.c
929
sg_set_page(rm->atomic.op_sg, page, 8, offset_in_page(args->local_addr));
net/rds/rdma.c
931
if (rm->atomic.op_notify || rm->atomic.op_recverr) {
net/rds/rdma.c
937
rm->atomic.op_notifier = kmalloc_obj(*rm->atomic.op_notifier);
net/rds/rdma.c
938
if (!rm->atomic.op_notifier) {
net/rds/rdma.c
943
rm->atomic.op_notifier->n_user_token = args->user_token;
net/rds/rdma.c
944
rm->atomic.op_notifier->n_status = RDS_RDMA_SUCCESS;
net/rds/rdma.c
947
rm->atomic.op_rkey = rds_rdma_cookie_key(args->cookie);
net/rds/rdma.c
948
rm->atomic.op_remote_addr = args->remote_addr + rds_rdma_cookie_offset(args->cookie);
net/rds/rdma.c
954
rm->atomic.op_active = 0;
net/rds/rdma.c
955
kfree(rm->atomic.op_notifier);
net/rds/rds.h
584
int (*xmit)(struct rds_connection *conn, struct rds_message *rm,
net/rds/rds.h
888
struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents);
net/rds/rds.h
889
int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from,
net/rds/rds.h
900
void rds_message_addref(struct rds_message *rm);
net/rds/rds.h
901
void rds_message_put(struct rds_message *rm);
net/rds/rds.h
902
void rds_message_wait(struct rds_message *rm);
net/rds/rds.h
903
void rds_message_unmapped(struct rds_message *rm);
net/rds/rds.h
950
typedef int (*is_acked_func)(struct rds_message *rm, uint64_t ack);
net/rds/rds.h
966
int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
net/rds/rds.h
968
int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
net/rds/rds.h
971
int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
net/rds/rds.h
975
void rds_rdma_send_complete(struct rds_message *rm, int wc_status);
net/rds/rds.h
976
void rds_atomic_send_complete(struct rds_message *rm, int wc_status);
net/rds/rds.h
977
int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
net/rds/recv.c
137
struct rds_message *rm, *tmp;
net/rds/recv.c
151
list_for_each_entry_safe(rm, tmp,
net/rds/recv.c
154
set_bit(RDS_MSG_FLUSH, &rm->m_flags);
net/rds/send.c
1030
static int rds_cmsg_zcopy(struct rds_sock *rs, struct rds_message *rm,
net/rds/send.c
1036
!rm->data.op_mmp_znotifier)
net/rds/send.c
1039
rm->data.op_mmp_znotifier->z_cookie = *cookie;
net/rds/send.c
1043
static int rds_cmsg_send(struct rds_sock *rs, struct rds_message *rm,
net/rds/send.c
1064
ret = rds_cmsg_rdma_args(rs, rm, cmsg, &vct->vec[ind]);
net/rds/send.c
1069
ret = rds_cmsg_rdma_dest(rs, rm, cmsg);
net/rds/send.c
1073
ret = rds_cmsg_rdma_map(rs, rm, cmsg);
net/rds/send.c
1086
ret = rds_cmsg_atomic(rs, rm, cmsg);
net/rds/send.c
1090
ret = rds_cmsg_zcopy(rs, rm, cmsg);
net/rds/send.c
1134
struct rds_message *rm = NULL;
net/rds/send.c
1294
rm = rds_message_alloc(ret, GFP_KERNEL);
net/rds/send.c
1295
if (!rm) {
net/rds/send.c
1302
rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs);
net/rds/send.c
1303
if (IS_ERR(rm->data.op_sg)) {
net/rds/send.c
1304
ret = PTR_ERR(rm->data.op_sg);
net/rds/send.c
1307
ret = rds_message_copy_from_user(rm, &msg->msg_iter, zcopy);
net/rds/send.c
1311
rm->data.op_active = 1;
net/rds/send.c
1313
rm->m_daddr = daddr;
net/rds/send.c
1360
rm->m_conn_path = cpath;
net/rds/send.c
1363
ret = rds_cmsg_send(rs, rm, msg, &allocated_mr, &vct);
net/rds/send.c
1367
if (rm->rdma.op_active && !conn->c_trans->xmit_rdma) {
net/rds/send.c
1369
&rm->rdma, conn->c_trans->xmit_rdma);
net/rds/send.c
1374
if (rm->atomic.op_active && !conn->c_trans->xmit_atomic) {
net/rds/send.c
1376
&rm->atomic, conn->c_trans->xmit_atomic);
net/rds/send.c
1394
while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port,
net/rds/send.c
1404
rds_send_queue_rm(rs, conn, cpath, rm,
net/rds/send.c
1439
rds_message_put(rm);
net/rds/send.c
1456
rds_rdma_unuse(rs, rds_rdma_cookie_key(rm->m_rdma_cookie), 1);
net/rds/send.c
1458
if (rm)
net/rds/send.c
1459
rds_message_put(rm);
net/rds/send.c
1475
struct rds_message *rm;
net/rds/send.c
1479
rm = rds_message_alloc(0, GFP_ATOMIC);
net/rds/send.c
1480
if (!rm) {
net/rds/send.c
1485
rm->m_daddr = cp->cp_conn->c_faddr;
net/rds/send.c
1486
rm->data.op_active = 1;
net/rds/send.c
1495
list_add_tail(&rm->m_conn_item, &cp->cp_send_queue);
net/rds/send.c
1496
set_bit(RDS_MSG_ON_CONN, &rm->m_flags);
net/rds/send.c
1497
rds_message_addref(rm);
net/rds/send.c
1498
rm->m_inc.i_conn = cp->cp_conn;
net/rds/send.c
1499
rm->m_inc.i_conn_path = cp;
net/rds/send.c
1501
rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport,
net/rds/send.c
1503
rm->m_inc.i_hdr.h_flags |= h_flags;
net/rds/send.c
1512
rds_message_add_extension(&rm->m_inc.i_hdr,
net/rds/send.c
1514
rds_message_add_extension(&rm->m_inc.i_hdr,
net/rds/send.c
1517
rds_message_add_extension(&rm->m_inc.i_hdr,
net/rds/send.c
1532
rds_message_put(rm);
net/rds/send.c
1536
if (rm)
net/rds/send.c
1537
rds_message_put(rm);
net/rds/send.c
190
struct rds_message *rm;
net/rds/send.c
252
rm = cp->cp_xmit_rm;
net/rds/send.c
254
if (!rm) {
net/rds/send.c
269
if (!rm && test_and_clear_bit(0, &conn->c_map_queued)) {
net/rds/send.c
270
rm = rds_cong_update_alloc(conn);
net/rds/send.c
271
if (IS_ERR(rm)) {
net/rds/send.c
272
ret = PTR_ERR(rm);
net/rds/send.c
275
rm->data.op_active = 1;
net/rds/send.c
276
rm->m_inc.i_conn_path = cp;
net/rds/send.c
277
rm->m_inc.i_conn = cp->cp_conn;
net/rds/send.c
279
cp->cp_xmit_rm = rm;
net/rds/send.c
289
if (!rm) {
net/rds/send.c
313
rm = list_entry(cp->cp_send_queue.next,
net/rds/send.c
316
rds_message_addref(rm);
net/rds/send.c
322
list_move_tail(&rm->m_conn_item,
net/rds/send.c
328
if (!rm)
net/rds/send.c
338
if (test_bit(RDS_MSG_FLUSH, &rm->m_flags) ||
net/rds/send.c
339
(rm->rdma.op_active &&
net/rds/send.c
340
test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))) {
net/rds/send.c
342
if (test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags))
net/rds/send.c
343
list_move(&rm->m_conn_item, &to_be_dropped);
net/rds/send.c
349
len = ntohl(rm->m_inc.i_hdr.h_len);
net/rds/send.c
352
set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
net/rds/send.c
364
cp->cp_xmit_rm = rm;
net/rds/send.c
368
if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) {
net/rds/send.c
369
rm->m_final_op = &rm->rdma;
net/rds/send.c
373
set_bit(RDS_MSG_MAPPED, &rm->m_flags);
net/rds/send.c
374
ret = conn->c_trans->xmit_rdma(conn, &rm->rdma);
net/rds/send.c
376
clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
net/rds/send.c
377
wake_up_interruptible(&rm->m_flush_wait);
net/rds/send.c
384
if (rm->atomic.op_active && !cp->cp_xmit_atomic_sent) {
net/rds/send.c
385
rm->m_final_op = &rm->atomic;
net/rds/send.c
389
set_bit(RDS_MSG_MAPPED, &rm->m_flags);
net/rds/send.c
390
ret = conn->c_trans->xmit_atomic(conn, &rm->atomic);
net/rds/send.c
392
clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
net/rds/send.c
393
wake_up_interruptible(&rm->m_flush_wait);
net/rds/send.c
407
if (rm->data.op_nents == 0) {
net/rds/send.c
411
ops_present = (rm->atomic.op_active || rm->rdma.op_active);
net/rds/send.c
412
if (rm->atomic.op_active && !rm->atomic.op_silent)
net/rds/send.c
414
if (rm->rdma.op_active && !rm->rdma.op_silent)
net/rds/send.c
418
&& !rm->m_rdma_cookie)
net/rds/send.c
419
rm->data.op_active = 0;
net/rds/send.c
422
if (rm->data.op_active && !cp->cp_xmit_data_sent) {
net/rds/send.c
423
rm->m_final_op = &rm->data;
net/rds/send.c
425
ret = conn->c_trans->xmit(conn, rm,
net/rds/send.c
440
sg = &rm->data.op_sg[cp->cp_xmit_sg];
net/rds/send.c
451
rm->data.op_nents);
net/rds/send.c
456
(cp->cp_xmit_sg == rm->data.op_nents))
net/rds/send.c
465
if (!rm->data.op_active || cp->cp_xmit_data_sent) {
net/rds/send.c
474
rds_message_put(rm);
net/rds/send.c
486
list_for_each_entry(rm, &to_be_dropped, m_conn_item)
net/rds/send.c
487
rds_message_put(rm);
net/rds/send.c
532
static void rds_send_sndbuf_remove(struct rds_sock *rs, struct rds_message *rm)
net/rds/send.c
534
u32 len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
net/rds/send.c
545
static inline int rds_send_is_acked(struct rds_message *rm, u64 ack,
net/rds/send.c
549
return is_acked(rm, ack);
net/rds/send.c
550
return be64_to_cpu(rm->m_inc.i_hdr.h_sequence) <= ack;
net/rds/send.c
559
void rds_rdma_send_complete(struct rds_message *rm, int status)
net/rds/send.c
566
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/send.c
568
ro = &rm->rdma;
net/rds/send.c
569
if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags) &&
net/rds/send.c
572
rs = rm->m_rs;
net/rds/send.c
583
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/send.c
595
void rds_atomic_send_complete(struct rds_message *rm, int status)
net/rds/send.c
602
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/send.c
604
ao = &rm->atomic;
net/rds/send.c
605
if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags)
net/rds/send.c
608
rs = rm->m_rs;
net/rds/send.c
619
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/send.c
634
__rds_send_complete(struct rds_sock *rs, struct rds_message *rm, int status)
net/rds/send.c
639
ro = &rm->rdma;
net/rds/send.c
646
ao = &rm->atomic;
net/rds/send.c
668
struct rds_message *rm;
net/rds/send.c
67
struct rds_message *rm, *tmp;
net/rds/send.c
673
rm = list_entry(messages->next, struct rds_message,
net/rds/send.c
675
list_del_init(&rm->m_conn_item);
net/rds/send.c
687
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/send.c
688
if (!test_bit(RDS_MSG_ON_SOCK, &rm->m_flags))
net/rds/send.c
691
if (rs != rm->m_rs) {
net/rds/send.c
696
rs = rm->m_rs;
net/rds/send.c
704
if (test_and_clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags)) {
net/rds/send.c
705
struct rm_rdma_op *ro = &rm->rdma;
net/rds/send.c
708
list_del_init(&rm->m_sock_item);
net/rds/send.c
709
rds_send_sndbuf_remove(rs, rm);
net/rds/send.c
71
rm = cp->cp_xmit_rm;
net/rds/send.c
718
rm->rdma.op_notifier = NULL;
net/rds/send.c
725
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/send.c
726
rds_message_put(rm);
net/rds/send.c
728
rds_message_put(rm);
net/rds/send.c
748
struct rds_message *rm, *tmp;
net/rds/send.c
754
list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) {
net/rds/send.c
755
if (!rds_send_is_acked(rm, ack, is_acked))
net/rds/send.c
758
list_move(&rm->m_conn_item, &list);
net/rds/send.c
759
clear_bit(RDS_MSG_ON_CONN, &rm->m_flags);
net/rds/send.c
77
rds_message_unmapped(rm);
net/rds/send.c
78
rds_message_put(rm);
net/rds/send.c
783
struct rds_message *rm, *tmp;
net/rds/send.c
792
list_for_each_entry_safe(rm, tmp, &rs->rs_send_queue, m_sock_item) {
net/rds/send.c
794
(!ipv6_addr_equal(&dest->sin6_addr, &rm->m_daddr) ||
net/rds/send.c
795
dest->sin6_port != rm->m_inc.i_hdr.h_dport))
net/rds/send.c
798
list_move(&rm->m_sock_item, &list);
net/rds/send.c
799
rds_send_sndbuf_remove(rs, rm);
net/rds/send.c
800
clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags);
net/rds/send.c
812
list_for_each_entry(rm, &list, m_sock_item) {
net/rds/send.c
814
conn = rm->m_inc.i_conn;
net/rds/send.c
816
cp = rm->m_inc.i_conn_path;
net/rds/send.c
826
if (!test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags)) {
net/rds/send.c
830
list_del_init(&rm->m_conn_item);
net/rds/send.c
837
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/send.c
840
__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
net/rds/send.c
843
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/send.c
845
rds_message_put(rm);
net/rds/send.c
851
rm = list_entry(list.next, struct rds_message, m_sock_item);
net/rds/send.c
852
list_del_init(&rm->m_sock_item);
net/rds/send.c
853
rds_message_wait(rm);
net/rds/send.c
860
spin_lock_irqsave(&rm->m_rs_lock, flags);
net/rds/send.c
863
__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
net/rds/send.c
866
spin_unlock_irqrestore(&rm->m_rs_lock, flags);
net/rds/send.c
868
rds_message_put(rm);
net/rds/send.c
879
struct rds_message *rm, __be16 sport,
net/rds/send.c
888
len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
net/rds/send.c
911
set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
net/rds/send.c
913
list_add_tail(&rm->m_sock_item, &rs->rs_send_queue);
net/rds/send.c
914
set_bit(RDS_MSG_ON_SOCK, &rm->m_flags);
net/rds/send.c
915
rds_message_addref(rm);
net/rds/send.c
917
rm->m_rs = rs;
net/rds/send.c
921
rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport, 0);
net/rds/send.c
922
rm->m_inc.i_conn = conn;
net/rds/send.c
923
rm->m_inc.i_conn_path = cp;
net/rds/send.c
924
rds_message_addref(rm);
net/rds/send.c
927
rm->m_inc.i_hdr.h_sequence = cpu_to_be64(cp->cp_next_tx_seq++);
net/rds/send.c
928
list_add_tail(&rm->m_conn_item, &cp->cp_send_queue);
net/rds/send.c
929
set_bit(RDS_MSG_ON_CONN, &rm->m_flags);
net/rds/send.c
933
rm, len, rs, rs->rs_snd_bytes,
net/rds/send.c
934
(unsigned long long)be64_to_cpu(rm->m_inc.i_hdr.h_sequence));
net/rds/send.c
95
list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) {
net/rds/send.c
96
set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
net/rds/send.c
97
set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags);
net/rds/tcp.h
109
int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
net/rds/tcp.h
111
int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack);
net/rds/tcp_send.c
106
(void *)&rm->m_inc.i_hdr + hdr_off,
net/rds/tcp_send.c
107
sizeof(rm->m_inc.i_hdr) - hdr_off);
net/rds/tcp_send.c
115
while (sg < rm->data.op_nents) {
net/rds/tcp_send.c
117
if (sg + 1 < rm->data.op_nents)
net/rds/tcp_send.c
120
bvec_set_page(&bvec, sg_page(&rm->data.op_sg[sg]),
net/rds/tcp_send.c
121
rm->data.op_sg[sg].length - off,
net/rds/tcp_send.c
122
rm->data.op_sg[sg].offset + off);
net/rds/tcp_send.c
124
rm->data.op_sg[sg].length - off);
net/rds/tcp_send.c
126
rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]),
net/rds/tcp_send.c
127
rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off,
net/rds/tcp_send.c
134
if (off == rm->data.op_sg[sg].length) {
net/rds/tcp_send.c
172
int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack)
net/rds/tcp_send.c
174
if (!test_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags))
net/rds/tcp_send.c
176
return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0;
net/rds/tcp_send.c
70
int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
net/rds/tcp_send.c
73
struct rds_conn_path *cp = rm->m_inc.i_conn_path;
net/rds/tcp_send.c
86
rm->m_ack_seq = tc->t_last_sent_nxt +
net/rds/tcp_send.c
88
be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1;
net/rds/tcp_send.c
90
set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags);
net/rds/tcp_send.c
91
tc->t_last_expected_una = rm->m_ack_seq + 1;
net/rds/tcp_send.c
93
if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))
net/rds/tcp_send.c
94
rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED;
net/rds/tcp_send.c
97
rm, rds_tcp_write_seq(tc),
net/rds/tcp_send.c
98
(unsigned long long)rm->m_ack_seq);
net/sunrpc/xprtsock.c
1111
rpc_fraghdr rm = xs_stream_record_marker(xdr);
net/sunrpc/xprtsock.c
1112
unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen;
net/sunrpc/xprtsock.c
1148
transport->xmit.offset, rm, &sent);
net/sunrpc/xprtsock.c
955
rpc_fraghdr rm = xs_stream_record_marker(xdr);
net/sunrpc/xprtsock.c
956
unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen;
net/sunrpc/xprtsock.c
977
transport->xmit.offset, rm, &sent);
net/tipc/group.c
595
struct tipc_member *m, *rm, *pm;
net/tipc/group.c
621
rm = list_first_entry(active, struct tipc_member, list);
net/tipc/group.c
622
rm->state = MBR_RECLAIMING;
net/tipc/group.c
623
list_del_init(&rm->list);
net/tipc/group.c
624
tipc_group_proto_xmit(grp, rm, GRP_RECLAIM_MSG, xmitq);
net/wireless/core.h
275
struct cfg80211_roam_info rm;
net/wireless/sme.c
1213
ev->rm.req_ie = next;
net/wireless/sme.c
1214
ev->rm.req_ie_len = info->req_ie_len;
net/wireless/sme.c
1215
memcpy((void *)ev->rm.req_ie, info->req_ie, info->req_ie_len);
net/wireless/sme.c
1219
ev->rm.resp_ie = next;
net/wireless/sme.c
1220
ev->rm.resp_ie_len = info->resp_ie_len;
net/wireless/sme.c
1221
memcpy((void *)ev->rm.resp_ie, info->resp_ie,
net/wireless/sme.c
1226
ev->rm.fils.kek = next;
net/wireless/sme.c
1227
ev->rm.fils.kek_len = info->fils.kek_len;
net/wireless/sme.c
1228
memcpy((void *)ev->rm.fils.kek, info->fils.kek,
net/wireless/sme.c
1233
ev->rm.fils.pmk = next;
net/wireless/sme.c
1234
ev->rm.fils.pmk_len = info->fils.pmk_len;
net/wireless/sme.c
1235
memcpy((void *)ev->rm.fils.pmk, info->fils.pmk,
net/wireless/sme.c
1240
ev->rm.fils.pmkid = next;
net/wireless/sme.c
1241
memcpy((void *)ev->rm.fils.pmkid, info->fils.pmkid,
net/wireless/sme.c
1245
ev->rm.fils.update_erp_next_seq_num = info->fils.update_erp_next_seq_num;
net/wireless/sme.c
1247
ev->rm.fils.erp_next_seq_num = info->fils.erp_next_seq_num;
net/wireless/sme.c
1249
ev->rm.ap_mld_addr = next;
net/wireless/sme.c
1250
memcpy((void *)ev->rm.ap_mld_addr, info->ap_mld_addr,
net/wireless/sme.c
1254
ev->rm.valid_links = info->valid_links;
net/wireless/sme.c
1256
ev->rm.links[link].bss = info->links[link].bss;
net/wireless/sme.c
1259
ev->rm.links[link].addr = next;
net/wireless/sme.c
1260
memcpy((void *)ev->rm.links[link].addr,
net/wireless/sme.c
1267
ev->rm.links[link].bssid = next;
net/wireless/sme.c
1268
memcpy((void *)ev->rm.links[link].bssid,
net/wireless/util.c
1134
__cfg80211_roamed(wdev, &ev->rm);
net/xfrm/xfrm_policy.c
1118
static void xfrm_policy_inexact_gc_tree(struct rb_root *r, bool rm)
net/xfrm/xfrm_policy.c
1126
xfrm_policy_inexact_gc_tree(&node->root, rm);
net/xfrm/xfrm_policy.c
1130
WARN_ON_ONCE(rm);
sound/mips/ad1843.c
298
int lg, rg, lm, rm;
sound/mips/ad1843.c
308
ad1843_read_multi(ad1843, 2, gp->lmute, &lm, gp->rmute, &rm);
sound/mips/ad1843.c
311
if (rm)
sound/mips/ad1843.c
331
int rm = (rg == 0) ? 1 : 0;
sound/mips/ad1843.c
338
ad1843_write_multi(ad1843, 2, gp->lmute, lm, gp->rmute, rm);
sound/soc/codecs/tas5805m.c
174
static void set_dsp_scale(struct regmap *rm, int offset, int vol)
sound/soc/codecs/tas5805m.c
185
regmap_bulk_write(rm, offset, v, ARRAY_SIZE(v));
sound/soc/codecs/tas5805m.c
190
struct regmap *rm = tas5805m->regmap;
sound/soc/codecs/tas5805m.c
195
regmap_write(rm, REG_PAGE, 0x00);
sound/soc/codecs/tas5805m.c
196
regmap_write(rm, REG_BOOK, 0x8c);
sound/soc/codecs/tas5805m.c
197
regmap_write(rm, REG_PAGE, 0x2a);
sound/soc/codecs/tas5805m.c
203
set_dsp_scale(rm, 0x24, tas5805m->vol[0]);
sound/soc/codecs/tas5805m.c
204
set_dsp_scale(rm, 0x28, tas5805m->vol[1]);
sound/soc/codecs/tas5805m.c
206
regmap_write(rm, REG_PAGE, 0x00);
sound/soc/codecs/tas5805m.c
207
regmap_write(rm, REG_BOOK, 0x00);
sound/soc/codecs/tas5805m.c
210
regmap_write(rm, REG_DEVICE_CTRL_2,
sound/soc/codecs/tas5805m.c
287
static void send_cfg(struct regmap *rm,
sound/soc/codecs/tas5805m.c
293
regmap_write(rm, s[i], s[i + 1]);
sound/soc/codecs/tas5805m.c
331
struct regmap *rm = tas5805m->regmap;
sound/soc/codecs/tas5805m.c
342
send_cfg(rm, dsp_cfg_preboot, ARRAY_SIZE(dsp_cfg_preboot));
sound/soc/codecs/tas5805m.c
344
send_cfg(rm, tas5805m->dsp_cfg_data, tas5805m->dsp_cfg_len);
sound/soc/codecs/tas5805m.c
357
struct regmap *rm = tas5805m->regmap;
sound/soc/codecs/tas5805m.c
369
regmap_write(rm, REG_PAGE, 0x00);
sound/soc/codecs/tas5805m.c
370
regmap_write(rm, REG_BOOK, 0x00);
sound/soc/codecs/tas5805m.c
372
regmap_read(rm, REG_CHAN_FAULT, &chan);
sound/soc/codecs/tas5805m.c
373
regmap_read(rm, REG_GLOBAL_FAULT1, &global1);
sound/soc/codecs/tas5805m.c
374
regmap_read(rm, REG_GLOBAL_FAULT2, &global2);
sound/soc/codecs/tas5805m.c
380
regmap_write(rm, REG_DEVICE_CTRL_2, DCTRL2_MODE_HIZ);
sound/soc/codecs/wcd9335.c
5023
struct regmap *rm = wcd->regmap;
sound/soc/codecs/wcd9335.c
5026
regmap_read(rm, WCD9335_CHIP_TIER_CTRL_EFUSE_VAL_OUT0, &val);
sound/soc/codecs/wcd9335.c
5027
regmap_read(rm, WCD9335_CHIP_TIER_CTRL_CHIP_ID_BYTE0, &byte0);
sound/soc/codecs/wcd9335.c
5036
regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x01);
sound/soc/codecs/wcd9335.c
5037
regmap_write(rm, WCD9335_SIDO_SIDO_TEST_2, 0x00);
sound/soc/codecs/wcd9335.c
5038
regmap_write(rm, WCD9335_SIDO_SIDO_CCL_8, 0x6F);
sound/soc/codecs/wcd9335.c
5039
regmap_write(rm, WCD9335_BIAS_VBG_FINE_ADJ, 0x65);
sound/soc/codecs/wcd9335.c
5040
regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x5);
sound/soc/codecs/wcd9335.c
5041
regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x7);
sound/soc/codecs/wcd9335.c
5042
regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x3);
sound/soc/codecs/wcd9335.c
5043
regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x3);
sound/soc/codecs/wcd934x.c
2204
struct regmap *rm = wcd->regmap;
sound/soc/codecs/wcd934x.c
2207
regmap_update_bits(rm, WCD934X_CDC_RX7_RX_PATH_CFG1, 0x08, 0x08);
sound/soc/codecs/wcd934x.c
2208
regmap_update_bits(rm, WCD934X_CDC_RX8_RX_PATH_CFG1, 0x08, 0x08);
sound/soc/codecs/wcd934x.c
2211
regmap_update_bits(rm, WCD934X_CPE_SS_DMIC_CFG, 0x80, 0x00);
sound/soc/codecs/wcd938x.c
344
struct regmap *rm = wcd938x->regmap;
sound/soc/codecs/wcd938x.c
346
regmap_update_bits(rm, WCD938X_SLEEP_CTL, 0x0E, 0x0E);
sound/soc/codecs/wcd938x.c
347
regmap_update_bits(rm, WCD938X_SLEEP_CTL, 0x80, 0x80);
sound/soc/codecs/wcd938x.c
350
regmap_update_bits(rm, WCD938X_SLEEP_CTL, 0x40, 0x40);
sound/soc/codecs/wcd938x.c
353
regmap_update_bits(rm, WCD938X_LDORXTX_CONFIG, 0x10, 0x00);
sound/soc/codecs/wcd938x.c
354
regmap_update_bits(rm, WCD938X_BIAS_VBG_FINE_ADJ,
sound/soc/codecs/wcd938x.c
356
regmap_update_bits(rm, WCD938X_ANA_BIAS, 0x80, 0x80);
sound/soc/codecs/wcd938x.c
357
regmap_update_bits(rm, WCD938X_ANA_BIAS, 0x40, 0x40);
sound/soc/codecs/wcd938x.c
361
regmap_update_bits(rm, WCD938X_ANA_BIAS, 0x40, 0x00);
sound/soc/codecs/wcd938x.c
362
regmap_update_bits(rm, WCD938X_HPH_NEW_INT_RDAC_GAIN_CTL,
sound/soc/codecs/wcd938x.c
364
regmap_update_bits(rm, WCD938X_HPH_NEW_INT_RDAC_HD2_CTL_L_NEW,
sound/soc/codecs/wcd938x.c
366
regmap_update_bits(rm, WCD938X_HPH_NEW_INT_RDAC_HD2_CTL_R_NEW,
sound/soc/codecs/wcd938x.c
368
regmap_update_bits(rm, WCD938X_HPH_REFBUFF_UHQA_CTL,
sound/soc/codecs/wcd938x.c
370
regmap_update_bits(rm, WCD938X_DIGITAL_CDC_DMIC_CTL,
sound/soc/codecs/wcd938x.c
373
regmap_update_bits(rm, WCD938X_TX_COM_NEW_INT_TXFE_ICTRL_STG2CASC_ULP,
sound/soc/codecs/wcd938x.c
375
regmap_update_bits(rm, WCD938X_TX_COM_NEW_INT_TXFE_ICTRL_STG2MAIN_ULP,
sound/soc/codecs/wcd938x.c
378
regmap_update_bits(rm, WCD938X_DIGITAL_TX_REQ_FB_CTL_0, 0xFF, 0x55);
sound/soc/codecs/wcd938x.c
379
regmap_update_bits(rm, WCD938X_DIGITAL_TX_REQ_FB_CTL_1, 0xFF, 0x44);
sound/soc/codecs/wcd938x.c
380
regmap_update_bits(rm, WCD938X_DIGITAL_TX_REQ_FB_CTL_2, 0xFF, 0x11);
sound/soc/codecs/wcd938x.c
381
regmap_update_bits(rm, WCD938X_DIGITAL_TX_REQ_FB_CTL_3, 0xFF, 0x00);
sound/soc/codecs/wcd938x.c
382
regmap_update_bits(rm, WCD938X_DIGITAL_TX_REQ_FB_CTL_4, 0xFF, 0x00);
sound/soc/codecs/wcd938x.c
385
regmap_update_bits(rm, WCD938X_MICB1_TEST_CTL_1, 0xE0, 0xE0);
sound/soc/codecs/wcd938x.c
386
regmap_update_bits(rm, WCD938X_MICB2_TEST_CTL_1, 0xE0, 0xE0);
sound/soc/codecs/wcd938x.c
387
regmap_update_bits(rm, WCD938X_MICB3_TEST_CTL_1, 0xE0, 0xE0);
sound/soc/codecs/wcd938x.c
388
regmap_update_bits(rm, WCD938X_MICB4_TEST_CTL_1, 0xE0, 0xE0);
sound/soc/codecs/wcd938x.c
390
regmap_update_bits(rm, WCD938X_TX_3_4_TEST_BLK_EN2, 0x01, 0x00);
sound/soc/codecs/wcd938x.c
391
regmap_update_bits(rm, WCD938X_HPH_SURGE_HPHLR_SURGE_EN, 0xC0, 0xC0);
sound/soc/codecs/wsa881x.c
688
struct regmap *rm = wsa881x->regmap;
sound/soc/codecs/wsa881x.c
698
regmap_update_bits(rm, WSA881X_SWR_RESET_EN, 0x07, 0x07);
sound/soc/codecs/wsa881x.c
701
regmap_update_bits(rm, WSA881X_CDC_RST_CTL, 0x02, 0x02);
sound/soc/codecs/wsa881x.c
704
regmap_update_bits(rm, WSA881X_CDC_RST_CTL, 0x01, 0x01);
sound/soc/codecs/wsa881x.c
705
regmap_update_bits(rm, WSA881X_CLOCK_CONFIG, 0x10, 0x10);
sound/soc/codecs/wsa881x.c
706
regmap_update_bits(rm, WSA881X_SPKR_OCP_CTL, 0x02, 0x02);
sound/soc/codecs/wsa881x.c
707
regmap_update_bits(rm, WSA881X_SPKR_MISC_CTL1, 0xC0, 0x80);
sound/soc/codecs/wsa881x.c
708
regmap_update_bits(rm, WSA881X_SPKR_MISC_CTL1, 0x06, 0x06);
sound/soc/codecs/wsa881x.c
709
regmap_update_bits(rm, WSA881X_SPKR_BIAS_INT, 0xFF, 0x00);
sound/soc/codecs/wsa881x.c
710
regmap_update_bits(rm, WSA881X_SPKR_PA_INT, 0xF0, 0x40);
sound/soc/codecs/wsa881x.c
711
regmap_update_bits(rm, WSA881X_SPKR_PA_INT, 0x0E, 0x0E);
sound/soc/codecs/wsa881x.c
712
regmap_update_bits(rm, WSA881X_BOOST_LOOP_STABILITY, 0x03, 0x03);
sound/soc/codecs/wsa881x.c
713
regmap_update_bits(rm, WSA881X_BOOST_MISC2_CTL, 0xFF, 0x14);
sound/soc/codecs/wsa881x.c
714
regmap_update_bits(rm, WSA881X_BOOST_START_CTL, 0x80, 0x80);
sound/soc/codecs/wsa881x.c
715
regmap_update_bits(rm, WSA881X_BOOST_START_CTL, 0x03, 0x00);
sound/soc/codecs/wsa881x.c
716
regmap_update_bits(rm, WSA881X_BOOST_SLOPE_COMP_ISENSE_FB, 0x0C, 0x04);
sound/soc/codecs/wsa881x.c
717
regmap_update_bits(rm, WSA881X_BOOST_SLOPE_COMP_ISENSE_FB, 0x03, 0x00);
sound/soc/codecs/wsa881x.c
719
regmap_read(rm, WSA881X_OTP_REG_0, &val);
sound/soc/codecs/wsa881x.c
721
regmap_update_bits(rm, WSA881X_BOOST_PRESET_OUT1, 0xF0, 0x70);
sound/soc/codecs/wsa881x.c
723
regmap_update_bits(rm, WSA881X_BOOST_PRESET_OUT2, 0xF0, 0x30);
sound/soc/codecs/wsa881x.c
724
regmap_update_bits(rm, WSA881X_SPKR_DRV_EN, 0x08, 0x08);
sound/soc/codecs/wsa881x.c
725
regmap_update_bits(rm, WSA881X_BOOST_CURRENT_LIMIT, 0x0F, 0x08);
sound/soc/codecs/wsa881x.c
726
regmap_update_bits(rm, WSA881X_SPKR_OCP_CTL, 0x30, 0x30);
sound/soc/codecs/wsa881x.c
727
regmap_update_bits(rm, WSA881X_SPKR_OCP_CTL, 0x0C, 0x00);
sound/soc/codecs/wsa881x.c
728
regmap_update_bits(rm, WSA881X_OTP_REG_28, 0x3F, 0x3A);
sound/soc/codecs/wsa881x.c
729
regmap_update_bits(rm, WSA881X_BONGO_RESRV_REG1, 0xFF, 0xB2);
sound/soc/codecs/wsa881x.c
730
regmap_update_bits(rm, WSA881X_BONGO_RESRV_REG2, 0xFF, 0x05);
tools/arch/x86/lib/insn.c
501
insn_byte_t mod, rm, base;
tools/arch/x86/lib/insn.c
530
rm = X86_MODRM_RM(insn->modrm.value);
tools/arch/x86/lib/insn.c
538
if ((mod == 0 && rm == 6) || mod == 2) {
tools/arch/x86/lib/insn.c
543
if ((mod == 0 && rm == 5) || mod == 2 ||
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
62
regmatch_t rm[5];
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
68
if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) {
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
71
new_len += (int)(rm[1].rm_eo - rm[1].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
78
(int)(rm[1].rm_eo - rm[1].rm_so), old_op + rm[1].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
79
} else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) {
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
84
if (rm[2].rm_so == -1)
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
87
new_len += (int)(rm[2].rm_eo - rm[2].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
93
if (rm[2].rm_so == -1)
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
97
(int)(rm[2].rm_eo - rm[2].rm_so),
tools/perf/util/perf-regs-arch/perf_regs_aarch64.c
98
old_op + rm[2].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
101
(int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
102
} else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) {
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
107
prefix = (rm[1].rm_so == -1) ? '+' : '-';
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
110
new_len += (int)(rm[2].rm_eo - rm[2].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
111
new_len += (int)(rm[4].rm_eo - rm[4].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
118
(int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so,
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
119
(int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
77
regmatch_t rm[5];
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
90
if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) {
tools/perf/util/perf-regs-arch/perf_regs_powerpc.c
94
new_len += (int)(rm[2].rm_eo - rm[2].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_x86.c
166
regmatch_t rm[6];
tools/perf/util/perf-regs-arch/perf_regs_x86.c
185
regexec(&sdt_op_regex, old_op, 6, rm, 0) ||
tools/perf/util/perf-regs-arch/perf_regs_x86.c
186
rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) {
tools/perf/util/perf-regs-arch/perf_regs_x86.c
203
if (rm[3].rm_so != rm[3].rm_eo) {
tools/perf/util/perf-regs-arch/perf_regs_x86.c
204
if (rm[1].rm_so != rm[1].rm_eo)
tools/perf/util/perf-regs-arch/perf_regs_x86.c
205
prefix[0] = *(old_op + rm[1].rm_so);
tools/perf/util/perf-regs-arch/perf_regs_x86.c
206
else if (rm[2].rm_so != rm[2].rm_eo)
tools/perf/util/perf-regs-arch/perf_regs_x86.c
213
sdt_rename_register(old_op + rm[4].rm_so, rm[4].rm_eo - rm[4].rm_so,
tools/perf/util/perf-regs-arch/perf_regs_x86.c
218
(rm[2].rm_eo - rm[2].rm_so) +
tools/perf/util/perf-regs-arch/perf_regs_x86.c
219
(rm[3].rm_eo - rm[3].rm_so) +
tools/perf/util/perf-regs-arch/perf_regs_x86.c
221
(rm[5].rm_eo - rm[5].rm_so) +
tools/perf/util/perf-regs-arch/perf_regs_x86.c
230
(int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so,
tools/perf/util/perf-regs-arch/perf_regs_x86.c
231
(int)(rm[3].rm_eo - rm[3].rm_so), old_op + rm[3].rm_so,
tools/perf/util/perf-regs-arch/perf_regs_x86.c
233
(int)(rm[5].rm_eo - rm[5].rm_so), old_op + rm[5].rm_so);
tools/testing/selftests/exec/non-regular.c
152
rm(_metadata, self->pathname, variant->is_dir);
tools/testing/selftests/exec/non-regular.c
158
rm(_metadata, self->pathname, self->is_dir);
tools/testing/selftests/kvm/x86/fastops_test.c
110
type_t _a = __val1, _d = __val1, rm = __val2; \
tools/testing/selftests/kvm/x86/fastops_test.c
115
ex_v = guest_execute_fastop_div(KVM_ASM_SAFE, insn, ex_a, ex_d, rm, ex_flags); \
tools/testing/selftests/kvm/x86/fastops_test.c
116
v = guest_execute_fastop_div(KVM_ASM_SAFE_FEP, insn, a, d, rm, flags); \
tools/testing/selftests/kvm/x86/fastops_test.c
121
ex_v, insn, (uint64_t)_a, (uint64_t)_d, (uint64_t)rm, v); \
tools/testing/selftests/kvm/x86/fastops_test.c
125
(uint64_t)_d, (uint64_t)rm, (uint64_t)a, (uint64_t)d); \
tools/testing/selftests/kvm/x86/fastops_test.c
128
ex_flags, insn, (uint64_t)_a, (uint64_t)_d, (uint64_t)rm, flags);\
tools/testing/selftests/kvm/x86/hyperv_clock.c
27
} rm, rn, rh, a0, b0;
tools/testing/selftests/kvm/x86/hyperv_clock.c
33
rm.ll = (u64)a0.l.low * b0.l.high;
tools/testing/selftests/kvm/x86/hyperv_clock.c
37
rh.l.low = c = rm.l.high + rn.l.high + rh.l.low;