Symbol: Rn
arch/arm/mm/alignment.c
678
u32 Rn = (tinstr & (7<<8)) >> 8;
arch/arm/mm/alignment.c
679
u32 W = ((L<<Rn) & (tinstr&255)) ? 0 : 1<<21;
arch/arm/mm/alignment.c
681
return 0xe8800000 | W | (L<<20) | (Rn<<16) |
arch/arm64/include/asm/insn.h
666
enum aarch64_insn_register Rn,
arch/arm64/include/asm/insn.h
671
enum aarch64_insn_register Rn,
arch/arm64/kernel/compat_alignment.c
192
u32 Rn = (tinstr & (7<<8)) >> 8;
arch/arm64/kernel/compat_alignment.c
193
u32 W = ((L<<Rn) & (tinstr&255)) ? 0 : 1<<21;
arch/arm64/kernel/compat_alignment.c
195
return 0xe8800000 | W | (L<<20) | (Rn<<16) |
arch/arm64/kvm/pauth.c
29
: [Rn] "r" ((n)), [Rm] "r" ((m)))
arch/arm64/lib/insn.c
1441
enum aarch64_insn_register Rn,
arch/arm64/lib/insn.c
1466
insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, Rn);
arch/arm64/lib/insn.c
1472
enum aarch64_insn_register Rn,
arch/arm64/lib/insn.c
1498
insn = aarch64_insn_encode_register(AARCH64_INSN_REGTYPE_RN, insn, Rn);
arch/arm64/net/bpf_jit.h
101
#define A64_PUSH(Rt, Rt2, Rn) A64_LS_PAIR(Rt, Rt2, Rn, -16, STORE, PRE_INDEX)
arch/arm64/net/bpf_jit.h
103
#define A64_POP(Rt, Rt2, Rn) A64_LS_PAIR(Rt, Rt2, Rn, 16, LOAD, POST_INDEX)
arch/arm64/net/bpf_jit.h
108
#define A64_LSX(sf, Rt, Rn, Rs, type) \
arch/arm64/net/bpf_jit.h
109
aarch64_insn_gen_load_store_ex(Rt, Rn, Rs, A64_SIZE(sf), \
arch/arm64/net/bpf_jit.h
112
#define A64_LDXR(sf, Rt, Rn) \
arch/arm64/net/bpf_jit.h
113
A64_LSX(sf, Rt, Rn, A64_ZR, LOAD_EX)
arch/arm64/net/bpf_jit.h
115
#define A64_STXR(sf, Rt, Rn, Rs) \
arch/arm64/net/bpf_jit.h
116
A64_LSX(sf, Rt, Rn, Rs, STORE_EX)
arch/arm64/net/bpf_jit.h
118
#define A64_STLXR(sf, Rt, Rn, Rs) \
arch/arm64/net/bpf_jit.h
119
aarch64_insn_gen_load_store_ex(Rt, Rn, Rs, A64_SIZE(sf), \
arch/arm64/net/bpf_jit.h
123
#define A64_LDAR(Rt, Rn, size) \
arch/arm64/net/bpf_jit.h
124
aarch64_insn_gen_load_acq_store_rel(Rt, Rn, AARCH64_INSN_SIZE_##size, \
arch/arm64/net/bpf_jit.h
126
#define A64_STLR(Rt, Rn, size) \
arch/arm64/net/bpf_jit.h
127
aarch64_insn_gen_load_acq_store_rel(Rt, Rn, AARCH64_INSN_SIZE_##size, \
arch/arm64/net/bpf_jit.h
148
#define A64_ST_OP(sf, Rn, Rs, op) \
arch/arm64/net/bpf_jit.h
149
aarch64_insn_gen_atomic_ld_op(A64_ZR, Rn, Rs, \
arch/arm64/net/bpf_jit.h
153
#define A64_STADD(sf, Rn, Rs) A64_ST_OP(sf, Rn, Rs, ADD)
arch/arm64/net/bpf_jit.h
154
#define A64_STCLR(sf, Rn, Rs) A64_ST_OP(sf, Rn, Rs, CLR)
arch/arm64/net/bpf_jit.h
155
#define A64_STEOR(sf, Rn, Rs) A64_ST_OP(sf, Rn, Rs, EOR)
arch/arm64/net/bpf_jit.h
156
#define A64_STSET(sf, Rn, Rs) A64_ST_OP(sf, Rn, Rs, SET)
arch/arm64/net/bpf_jit.h
158
#define A64_LD_OP_AL(sf, Rt, Rn, Rs, op) \
arch/arm64/net/bpf_jit.h
159
aarch64_insn_gen_atomic_ld_op(Rt, Rn, Rs, \
arch/arm64/net/bpf_jit.h
163
#define A64_LDADDAL(sf, Rt, Rn, Rs) A64_LD_OP_AL(sf, Rt, Rn, Rs, ADD)
arch/arm64/net/bpf_jit.h
164
#define A64_LDCLRAL(sf, Rt, Rn, Rs) A64_LD_OP_AL(sf, Rt, Rn, Rs, CLR)
arch/arm64/net/bpf_jit.h
165
#define A64_LDEORAL(sf, Rt, Rn, Rs) A64_LD_OP_AL(sf, Rt, Rn, Rs, EOR)
arch/arm64/net/bpf_jit.h
166
#define A64_LDSETAL(sf, Rt, Rn, Rs) A64_LD_OP_AL(sf, Rt, Rn, Rs, SET)
arch/arm64/net/bpf_jit.h
168
#define A64_SWPAL(sf, Rt, Rn, Rs) A64_LD_OP_AL(sf, Rt, Rn, Rs, SWP)
arch/arm64/net/bpf_jit.h
170
#define A64_CASAL(sf, Rt, Rn, Rs) \
arch/arm64/net/bpf_jit.h
171
aarch64_insn_gen_cas(Rt, Rn, Rs, A64_SIZE(sf), \
arch/arm64/net/bpf_jit.h
175
#define A64_ADDSUB_IMM(sf, Rd, Rn, imm12, type) \
arch/arm64/net/bpf_jit.h
176
aarch64_insn_gen_add_sub_imm(Rd, Rn, imm12, \
arch/arm64/net/bpf_jit.h
179
#define A64_ADD_I(sf, Rd, Rn, imm12) A64_ADDSUB_IMM(sf, Rd, Rn, imm12, ADD)
arch/arm64/net/bpf_jit.h
180
#define A64_SUB_I(sf, Rd, Rn, imm12) A64_ADDSUB_IMM(sf, Rd, Rn, imm12, SUB)
arch/arm64/net/bpf_jit.h
181
#define A64_ADDS_I(sf, Rd, Rn, imm12) \
arch/arm64/net/bpf_jit.h
182
A64_ADDSUB_IMM(sf, Rd, Rn, imm12, ADD_SETFLAGS)
arch/arm64/net/bpf_jit.h
183
#define A64_SUBS_I(sf, Rd, Rn, imm12) \
arch/arm64/net/bpf_jit.h
184
A64_ADDSUB_IMM(sf, Rd, Rn, imm12, SUB_SETFLAGS)
arch/arm64/net/bpf_jit.h
186
#define A64_CMN_I(sf, Rn, imm12) A64_ADDS_I(sf, A64_ZR, Rn, imm12)
arch/arm64/net/bpf_jit.h
188
#define A64_CMP_I(sf, Rn, imm12) A64_SUBS_I(sf, A64_ZR, Rn, imm12)
arch/arm64/net/bpf_jit.h
190
#define A64_MOV(sf, Rd, Rn) A64_ADD_I(sf, Rd, Rn, 0)
arch/arm64/net/bpf_jit.h
193
#define A64_BITFIELD(sf, Rd, Rn, immr, imms, type) \
arch/arm64/net/bpf_jit.h
194
aarch64_insn_gen_bitfield(Rd, Rn, immr, imms, \
arch/arm64/net/bpf_jit.h
197
#define A64_SBFM(sf, Rd, Rn, ir, is) A64_BITFIELD(sf, Rd, Rn, ir, is, SIGNED)
arch/arm64/net/bpf_jit.h
199
#define A64_UBFM(sf, Rd, Rn, ir, is) A64_BITFIELD(sf, Rd, Rn, ir, is, UNSIGNED)
arch/arm64/net/bpf_jit.h
202
#define A64_LSL(sf, Rd, Rn, shift) ({ \
arch/arm64/net/bpf_jit.h
204
A64_UBFM(sf, Rd, Rn, (unsigned)-(shift) % sz, sz - 1 - (shift)); \
arch/arm64/net/bpf_jit.h
207
#define A64_LSR(sf, Rd, Rn, shift) A64_UBFM(sf, Rd, Rn, shift, (sf) ? 63 : 31)
arch/arm64/net/bpf_jit.h
209
#define A64_ASR(sf, Rd, Rn, shift) A64_SBFM(sf, Rd, Rn, shift, (sf) ? 63 : 31)
arch/arm64/net/bpf_jit.h
212
#define A64_UXTH(sf, Rd, Rn) A64_UBFM(sf, Rd, Rn, 0, 15)
arch/arm64/net/bpf_jit.h
213
#define A64_UXTW(sf, Rd, Rn) A64_UBFM(sf, Rd, Rn, 0, 31)
arch/arm64/net/bpf_jit.h
216
#define A64_SXTB(sf, Rd, Rn) A64_SBFM(sf, Rd, Rn, 0, 7)
arch/arm64/net/bpf_jit.h
217
#define A64_SXTH(sf, Rd, Rn) A64_SBFM(sf, Rd, Rn, 0, 15)
arch/arm64/net/bpf_jit.h
218
#define A64_SXTW(sf, Rd, Rn) A64_SBFM(sf, Rd, Rn, 0, 31)
arch/arm64/net/bpf_jit.h
232
#define A64_ADDSUB_SREG(sf, Rd, Rn, Rm, type) \
arch/arm64/net/bpf_jit.h
233
aarch64_insn_gen_add_sub_shifted_reg(Rd, Rn, Rm, 0, \
arch/arm64/net/bpf_jit.h
236
#define A64_ADD(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, ADD)
arch/arm64/net/bpf_jit.h
237
#define A64_SUB(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB)
arch/arm64/net/bpf_jit.h
238
#define A64_SUBS(sf, Rd, Rn, Rm) A64_ADDSUB_SREG(sf, Rd, Rn, Rm, SUB_SETFLAGS)
arch/arm64/net/bpf_jit.h
242
#define A64_CMP(sf, Rn, Rm) A64_SUBS(sf, A64_ZR, Rn, Rm)
arch/arm64/net/bpf_jit.h
245
#define A64_DATA1(sf, Rd, Rn, type) aarch64_insn_gen_data1(Rd, Rn, \
arch/arm64/net/bpf_jit.h
248
#define A64_REV16(sf, Rd, Rn) A64_DATA1(sf, Rd, Rn, REVERSE_16)
arch/arm64/net/bpf_jit.h
249
#define A64_REV32(sf, Rd, Rn) A64_DATA1(sf, Rd, Rn, REVERSE_32)
arch/arm64/net/bpf_jit.h
250
#define A64_REV64(Rd, Rn) A64_DATA1(1, Rd, Rn, REVERSE_64)
arch/arm64/net/bpf_jit.h
254
#define A64_DATA2(sf, Rd, Rn, Rm, type) aarch64_insn_gen_data2(Rd, Rn, Rm, \
arch/arm64/net/bpf_jit.h
256
#define A64_UDIV(sf, Rd, Rn, Rm) A64_DATA2(sf, Rd, Rn, Rm, UDIV)
arch/arm64/net/bpf_jit.h
257
#define A64_SDIV(sf, Rd, Rn, Rm) A64_DATA2(sf, Rd, Rn, Rm, SDIV)
arch/arm64/net/bpf_jit.h
258
#define A64_LSLV(sf, Rd, Rn, Rm) A64_DATA2(sf, Rd, Rn, Rm, LSLV)
arch/arm64/net/bpf_jit.h
259
#define A64_LSRV(sf, Rd, Rn, Rm) A64_DATA2(sf, Rd, Rn, Rm, LSRV)
arch/arm64/net/bpf_jit.h
260
#define A64_ASRV(sf, Rd, Rn, Rm) A64_DATA2(sf, Rd, Rn, Rm, ASRV)
arch/arm64/net/bpf_jit.h
264
#define A64_MADD(sf, Rd, Ra, Rn, Rm) aarch64_insn_gen_data3(Rd, Ra, Rn, Rm, \
arch/arm64/net/bpf_jit.h
267
#define A64_MSUB(sf, Rd, Ra, Rn, Rm) aarch64_insn_gen_data3(Rd, Ra, Rn, Rm, \
arch/arm64/net/bpf_jit.h
270
#define A64_MUL(sf, Rd, Rn, Rm) A64_MADD(sf, Rd, A64_ZR, Rn, Rm)
arch/arm64/net/bpf_jit.h
273
#define A64_LOGIC_SREG(sf, Rd, Rn, Rm, type) \
arch/arm64/net/bpf_jit.h
274
aarch64_insn_gen_logical_shifted_reg(Rd, Rn, Rm, 0, \
arch/arm64/net/bpf_jit.h
277
#define A64_AND(sf, Rd, Rn, Rm) A64_LOGIC_SREG(sf, Rd, Rn, Rm, AND)
arch/arm64/net/bpf_jit.h
278
#define A64_ORR(sf, Rd, Rn, Rm) A64_LOGIC_SREG(sf, Rd, Rn, Rm, ORR)
arch/arm64/net/bpf_jit.h
279
#define A64_EOR(sf, Rd, Rn, Rm) A64_LOGIC_SREG(sf, Rd, Rn, Rm, EOR)
arch/arm64/net/bpf_jit.h
280
#define A64_ANDS(sf, Rd, Rn, Rm) A64_LOGIC_SREG(sf, Rd, Rn, Rm, AND_SETFLAGS)
arch/arm64/net/bpf_jit.h
282
#define A64_TST(sf, Rn, Rm) A64_ANDS(sf, A64_ZR, Rn, Rm)
arch/arm64/net/bpf_jit.h
288
#define A64_LOGIC_IMM(sf, Rd, Rn, imm, type) ({ \
arch/arm64/net/bpf_jit.h
291
A64_VARIANT(sf), Rn, Rd, imm64); \
arch/arm64/net/bpf_jit.h
294
#define A64_AND_I(sf, Rd, Rn, imm) A64_LOGIC_IMM(sf, Rd, Rn, imm, AND)
arch/arm64/net/bpf_jit.h
295
#define A64_ORR_I(sf, Rd, Rn, imm) A64_LOGIC_IMM(sf, Rd, Rn, imm, ORR)
arch/arm64/net/bpf_jit.h
296
#define A64_EOR_I(sf, Rd, Rn, imm) A64_LOGIC_IMM(sf, Rd, Rn, imm, EOR)
arch/arm64/net/bpf_jit.h
297
#define A64_ANDS_I(sf, Rd, Rn, imm) A64_LOGIC_IMM(sf, Rd, Rn, imm, AND_SETFLAGS)
arch/arm64/net/bpf_jit.h
299
#define A64_TST_I(sf, Rn, imm) A64_ANDS_I(sf, A64_ZR, Rn, imm)
arch/arm64/net/bpf_jit.h
51
#define A64_BR(Rn) aarch64_insn_gen_branch_reg(Rn, AARCH64_INSN_BRANCH_NOLINK)
arch/arm64/net/bpf_jit.h
52
#define A64_BLR(Rn) aarch64_insn_gen_branch_reg(Rn, AARCH64_INSN_BRANCH_LINK)
arch/arm64/net/bpf_jit.h
53
#define A64_RET(Rn) aarch64_insn_gen_branch_reg(Rn, AARCH64_INSN_BRANCH_RETURN)
arch/arm64/net/bpf_jit.h
56
#define A64_LS_REG(Rt, Rn, Rm, size, type) \
arch/arm64/net/bpf_jit.h
57
aarch64_insn_gen_load_store_reg(Rt, Rn, Rm, \
arch/arm64/net/bpf_jit.h
73
#define A64_LS_IMM(Rt, Rn, imm, size, type) \
arch/arm64/net/bpf_jit.h
74
aarch64_insn_gen_load_store_imm(Rt, Rn, imm, \
arch/arm64/net/bpf_jit.h
96
#define A64_LS_PAIR(Rt, Rt2, Rn, offset, ls, type) \
arch/arm64/net/bpf_jit.h
97
aarch64_insn_gen_load_store_pair(Rt, Rt2, Rn, offset, \
arch/sh/math-emu/math.c
212
MWRITE(FRm, Rn + R0 + 4);
arch/sh/math-emu/math.c
214
MWRITE(FRm, Rn + R0);
arch/sh/math-emu/math.c
216
MWRITE(FRm, Rn + R0);
arch/sh/math-emu/math.c
228
MWRITE(FRm, Rn + 4);
arch/sh/math-emu/math.c
230
MWRITE(FRm, Rn);
arch/sh/math-emu/math.c
232
MWRITE(FRm, Rn);
arch/sh/math-emu/math.c
244
Rn -= 8;
arch/sh/math-emu/math.c
245
MWRITE(FRm, Rn + 4);
arch/sh/math-emu/math.c
247
MWRITE(FRm, Rn);
arch/sh/math-emu/math.c
249
Rn -= 4;
arch/sh/math-emu/math.c
250
MWRITE(FRm, Rn);
arch/sh/math-emu/math.c
441
Rn = *reg;
arch/sh/math-emu/math.c
445
*reg = Rn;
arch/sh/math-emu/math.c
449
Rn -= 4;
arch/sh/math-emu/math.c
450
MWRITE(*reg, Rn);
arch/sh/math-emu/math.c
454
MREAD(*reg, Rn);
arch/sh/math-emu/math.c
455
Rn += 4;