Symbol: op
arch/alpha/include/asm/atomic.h
110
#define ATOMIC64_OP_RETURN(op, asm_op) \
arch/alpha/include/asm/atomic.h
112
arch_atomic64_##op##_return_relaxed(s64 i, atomic64_t * v) \
arch/alpha/include/asm/atomic.h
130
#define ATOMIC64_FETCH_OP(op, asm_op) \
arch/alpha/include/asm/atomic.h
132
arch_atomic64_fetch_##op##_relaxed(s64 i, atomic64_t * v) \
arch/alpha/include/asm/atomic.h
149
#define ATOMIC_OPS(op) \
arch/alpha/include/asm/atomic.h
150
ATOMIC_OP(op, op##l) \
arch/alpha/include/asm/atomic.h
151
ATOMIC_OP_RETURN(op, op##l) \
arch/alpha/include/asm/atomic.h
152
ATOMIC_FETCH_OP(op, op##l) \
arch/alpha/include/asm/atomic.h
153
ATOMIC64_OP(op, op##q) \
arch/alpha/include/asm/atomic.h
154
ATOMIC64_OP_RETURN(op, op##q) \
arch/alpha/include/asm/atomic.h
155
ATOMIC64_FETCH_OP(op, op##q)
arch/alpha/include/asm/atomic.h
174
#define ATOMIC_OPS(op, asm) \
arch/alpha/include/asm/atomic.h
175
ATOMIC_OP(op, asm) \
arch/alpha/include/asm/atomic.h
176
ATOMIC_FETCH_OP(op, asm) \
arch/alpha/include/asm/atomic.h
177
ATOMIC64_OP(op, asm) \
arch/alpha/include/asm/atomic.h
178
ATOMIC64_FETCH_OP(op, asm)
arch/alpha/include/asm/atomic.h
41
#define ATOMIC_OP(op, asm_op) \
arch/alpha/include/asm/atomic.h
42
static __inline__ void arch_atomic_##op(int i, atomic_t * v) \
arch/alpha/include/asm/atomic.h
57
#define ATOMIC_OP_RETURN(op, asm_op) \
arch/alpha/include/asm/atomic.h
58
static inline int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \
arch/alpha/include/asm/atomic.h
76
#define ATOMIC_FETCH_OP(op, asm_op) \
arch/alpha/include/asm/atomic.h
77
static inline int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \
arch/alpha/include/asm/atomic.h
94
#define ATOMIC64_OP(op, asm_op) \
arch/alpha/include/asm/atomic.h
95
static __inline__ void arch_atomic64_##op(s64 i, atomic64_t * v) \
arch/alpha/include/asm/futex.h
29
static inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/alpha/include/asm/futex.h
37
switch (op) {
arch/alpha/kernel/osf_sys.c
749
SYSCALL_DEFINE5(osf_getsysinfo, unsigned long, op, void __user *, buffer,
arch/alpha/kernel/osf_sys.c
755
switch (op) {
arch/alpha/kernel/osf_sys.c
806
SYSCALL_DEFINE5(osf_setsysinfo, unsigned long, op, void __user *, buffer,
arch/alpha/kernel/osf_sys.c
809
switch (op) {
arch/arc/include/asm/atomic-llsc.h
24
#define ATOMIC_OP_RETURN(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
25
static inline int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \
arch/arc/include/asm/atomic-llsc.h
45
#define ATOMIC_FETCH_OP(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
46
static inline int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \
arch/arc/include/asm/atomic-llsc.h
72
#define ATOMIC_OPS(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
73
ATOMIC_OP(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
74
ATOMIC_OP_RETURN(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
75
ATOMIC_FETCH_OP(op, asm_op)
arch/arc/include/asm/atomic-llsc.h
8
#define ATOMIC_OP(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
81
#define ATOMIC_OPS(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
82
ATOMIC_OP(op, asm_op) \
arch/arc/include/asm/atomic-llsc.h
83
ATOMIC_FETCH_OP(op, asm_op)
arch/arc/include/asm/atomic-llsc.h
9
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/arc/include/asm/atomic-spinlock.h
31
#define ATOMIC_OP(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
32
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/arc/include/asm/atomic-spinlock.h
41
#define ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
42
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/arc/include/asm/atomic-spinlock.h
59
#define ATOMIC_FETCH_OP(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
60
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/arc/include/asm/atomic-spinlock.h
76
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
77
ATOMIC_OP(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
78
ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
79
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/arc/include/asm/atomic-spinlock.h
90
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
91
ATOMIC_OP(op, c_op, asm_op) \
arch/arc/include/asm/atomic-spinlock.h
92
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/arc/include/asm/atomic64-arcv2.h
115
#define ATOMIC64_OPS(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
116
ATOMIC64_OP(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
117
ATOMIC64_OP_RETURN(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
118
ATOMIC64_FETCH_OP(op, op1, op2)
arch/arc/include/asm/atomic64-arcv2.h
124
#define ATOMIC64_OPS(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
125
ATOMIC64_OP(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
126
ATOMIC64_FETCH_OP(op, op1, op2)
arch/arc/include/asm/atomic64-arcv2.h
49
#define ATOMIC64_OP(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
50
static inline void arch_atomic64_##op(s64 a, atomic64_t *v) \
arch/arc/include/asm/atomic64-arcv2.h
66
#define ATOMIC64_OP_RETURN(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
67
static inline s64 arch_atomic64_##op##_return_relaxed(s64 a, atomic64_t *v) \
arch/arc/include/asm/atomic64-arcv2.h
88
#define ATOMIC64_FETCH_OP(op, op1, op2) \
arch/arc/include/asm/atomic64-arcv2.h
89
static inline s64 arch_atomic64_fetch_##op##_relaxed(s64 a, atomic64_t *v) \
arch/arc/include/asm/futex.h
73
static inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/arc/include/asm/futex.h
85
switch (op) {
arch/arc/include/asm/pgtable-bits-arcv2.h
85
#define PTE_BIT_FUNC(fn, op) \
arch/arc/include/asm/pgtable-bits-arcv2.h
86
static inline pte_t pte_##fn(pte_t pte) { pte_val(pte) op; return pte; }
arch/arc/include/asm/uaccess.h
101
#define __arc_put_user_one(src, dst, op, ret) \
arch/arc/include/asm/uaccess.h
103
"1: "op" %1,[%2]\n" \
arch/arc/include/asm/uaccess.h
46
#define __arc_get_user_one(dst, src, op, ret) \
arch/arc/include/asm/uaccess.h
48
"1: "op" %1,[%2]\n" \
arch/arc/mm/cache.c
189
unsigned long sz, const int op, const int full_page)
arch/arc/mm/cache.c
194
if (op == OP_INV_IC) {
arch/arc/mm/cache.c
198
aux_cmd = op & OP_INV ? ARC_REG_DC_IVDL : ARC_REG_DC_FLDL;
arch/arc/mm/cache.c
229
if (is_pae40_enabled() && op == OP_INV_IC)
arch/arc/mm/cache.c
249
unsigned long sz, const int op, const int full_page)
arch/arc/mm/cache.c
254
if (op == OP_INV_IC) {
arch/arc/mm/cache.c
258
aux_cmd = op & OP_INV ? ARC_REG_DC_IVDL : ARC_REG_DC_FLDL;
arch/arc/mm/cache.c
280
if (op == OP_INV_IC)
arch/arc/mm/cache.c
303
unsigned long sz, const int op, const int full_page)
arch/arc/mm/cache.c
308
if (op == OP_INV_IC) {
arch/arc/mm/cache.c
330
if (op == OP_INV_IC)
arch/arc/mm/cache.c
36
unsigned long sz, const int op, const int full_page);
arch/arc/mm/cache.c
362
static inline void __before_dc_op(const int op)
arch/arc/mm/cache.c
364
if (op == OP_FLUSH_N_INV) {
arch/arc/mm/cache.c
377
static inline void __before_dc_op(const int op)
arch/arc/mm/cache.c
382
if (op == OP_FLUSH_N_INV) {
arch/arc/mm/cache.c
386
if (op != OP_INV_IC) {
arch/arc/mm/cache.c
392
if (op & OP_INV)
arch/arc/mm/cache.c
401
static inline void __after_dc_op(const int op)
arch/arc/mm/cache.c
403
if (op & OP_FLUSH) {
arch/arc/mm/cache.c
412
if (op == OP_FLUSH_N_INV)
arch/arc/mm/cache.c
423
static inline void __dc_entire_op(const int op)
arch/arc/mm/cache.c
427
__before_dc_op(op);
arch/arc/mm/cache.c
429
if (op & OP_INV) /* Inv or flush-n-inv use same cmd reg */
arch/arc/mm/cache.c
436
__after_dc_op(op);
arch/arc/mm/cache.c
455
#define __dc_line_op_k(p, sz, op) __dc_line_op(p, p, sz, op)
arch/arc/mm/cache.c
461
unsigned long sz, const int op)
arch/arc/mm/cache.c
468
__before_dc_op(op);
arch/arc/mm/cache.c
470
__cache_line_loop(paddr, vaddr, sz, op, full_page);
arch/arc/mm/cache.c
472
__after_dc_op(op);
arch/arc/mm/cache.c
479
#define __dc_entire_op(op)
arch/arc/mm/cache.c
482
#define __dc_line_op(paddr, vaddr, sz, op)
arch/arc/mm/cache.c
483
#define __dc_line_op_k(paddr, sz, op)
arch/arc/mm/cache.c
546
static noinline void slc_op_rgn(phys_addr_t paddr, unsigned long sz, const int op)
arch/arc/mm/cache.c
572
if (!(op & OP_FLUSH)) /* i.e. OP_INV */
arch/arc/mm/cache.c
577
if (op & OP_INV)
arch/arc/mm/cache.c
609
static __maybe_unused noinline void slc_op_line(phys_addr_t paddr, unsigned long sz, const int op)
arch/arc/mm/cache.c
631
if (!(op & OP_FLUSH)) /* i.e. OP_INV */
arch/arc/mm/cache.c
638
cmd = op & OP_INV ? ARC_AUX_SLC_IVDL : ARC_AUX_SLC_FLDL;
arch/arc/mm/cache.c
659
#define slc_op(paddr, sz, op) slc_op_rgn(paddr, sz, op)
arch/arc/mm/cache.c
661
noinline static void slc_entire_op(const int op)
arch/arc/mm/cache.c
667
if (!(op & OP_FLUSH)) /* i.e. OP_INV */
arch/arc/mm/cache.c
674
if (op & OP_INV) /* Inv or flush-n-inv use same cmd reg */
arch/arc/net/bpf_jit_core.c
434
static int bpf_cond_to_arc(const u8 op, u8 *arc_cc)
arch/arc/net/bpf_jit_core.c
436
switch (op) {
arch/arc/net/bpf_jit_core.c
474
pr_err("bpf-jit: can't handle condition 0x%02X\n", op);
arch/arc/net/bpf_jit_core.c
491
const u8 op = BPF_OP(insn->code);
arch/arc/net/bpf_jit_core.c
495
(op == BPF_CALL || op == BPF_EXIT)) {
arch/arm/include/asm/assembler.h
595
.macro __adldst_l, op, reg, sym, tmp, c
arch/arm/include/asm/assembler.h
679
.macro __ldst_va, op, reg, tmp, sym, cond, offset
arch/arm/include/asm/atomic.h
162
#define ATOMIC_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
163
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
172
#define ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
173
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
186
#define ATOMIC_FETCH_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
187
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
227
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
228
ATOMIC_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
229
ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
230
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/arm/include/asm/atomic.h
238
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
239
ATOMIC_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
240
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/arm/include/asm/atomic.h
311
#define ATOMIC64_OP(op, op1, op2) \
arch/arm/include/asm/atomic.h
312
static inline void arch_atomic64_##op(s64 i, atomic64_t *v) \
arch/arm/include/asm/atomic.h
318
__asm__ __volatile__("@ atomic64_" #op "\n" \
arch/arm/include/asm/atomic.h
330
#define ATOMIC64_OP_RETURN(op, op1, op2) \
arch/arm/include/asm/atomic.h
332
arch_atomic64_##op##_return_relaxed(s64 i, atomic64_t *v) \
arch/arm/include/asm/atomic.h
339
__asm__ __volatile__("@ atomic64_" #op "_return\n" \
arch/arm/include/asm/atomic.h
353
#define ATOMIC64_FETCH_OP(op, op1, op2) \
arch/arm/include/asm/atomic.h
355
arch_atomic64_fetch_##op##_relaxed(s64 i, atomic64_t *v) \
arch/arm/include/asm/atomic.h
36
#define ATOMIC_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
362
__asm__ __volatile__("@ atomic64_fetch_" #op "\n" \
arch/arm/include/asm/atomic.h
37
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
376
#define ATOMIC64_OPS(op, op1, op2) \
arch/arm/include/asm/atomic.h
377
ATOMIC64_OP(op, op1, op2) \
arch/arm/include/asm/atomic.h
378
ATOMIC64_OP_RETURN(op, op1, op2) \
arch/arm/include/asm/atomic.h
379
ATOMIC64_FETCH_OP(op, op1, op2)
arch/arm/include/asm/atomic.h
390
#define ATOMIC64_OPS(op, op1, op2) \
arch/arm/include/asm/atomic.h
391
ATOMIC64_OP(op, op1, op2) \
arch/arm/include/asm/atomic.h
392
ATOMIC64_FETCH_OP(op, op1, op2)
arch/arm/include/asm/atomic.h
43
__asm__ __volatile__("@ atomic_" #op "\n" \
arch/arm/include/asm/atomic.h
54
#define ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
55
static inline int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
62
__asm__ __volatile__("@ atomic_" #op "_return\n" \
arch/arm/include/asm/atomic.h
75
#define ATOMIC_FETCH_OP(op, c_op, asm_op) \
arch/arm/include/asm/atomic.h
76
static inline int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \
arch/arm/include/asm/atomic.h
83
__asm__ __volatile__("@ atomic_fetch_" #op "\n" \
arch/arm/include/asm/firmware.h
64
#define call_firmware_op(op, ...) \
arch/arm/include/asm/firmware.h
65
((firmware_ops->op) ? firmware_ops->op(__VA_ARGS__) : (-ENOSYS))
arch/arm/include/asm/futex.h
133
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/arm/include/asm/futex.h
144
switch (op) {
arch/arm/include/asm/pgtable-3level.h
198
#define PMD_BIT_FUNC(fn,op) \
arch/arm/include/asm/pgtable-3level.h
199
static inline pmd_t pmd_##fn(pmd_t pmd) { pmd_val(pmd) op; return pmd; }
arch/arm/include/asm/syscalls.h
28
asmlinkage long sys_oabi_epoll_ctl(int epfd, int op, int fd,
arch/arm/kernel/sys_oabi-compat.c
283
asmlinkage long sys_oabi_epoll_ctl(int epfd, int op, int fd,
arch/arm/kernel/sys_oabi-compat.c
289
if (ep_op_has_event(op) &&
arch/arm/kernel/sys_oabi-compat.c
296
return do_epoll_ctl(epfd, op, fd, &kernel, false);
arch/arm/kernel/sys_oabi-compat.c
299
asmlinkage long sys_oabi_epoll_ctl(int epfd, int op, int fd,
arch/arm/mach-imx/mxc.h
79
extern struct cpu_op *(*get_cpu_op)(int *op);
arch/arm/mm/cache-uniphier.c
57
#define UNIPHIER_SSCOQM_S_IS_RANGE(op) \
arch/arm/mm/cache-uniphier.c
58
((op & UNIPHIER_SSCOQM_S_MASK) == UNIPHIER_SSCOQM_S_RANGE)
arch/arm/mm/dma-mapping.c
629
void (*op)(const void *, size_t, int))
arch/arm/mm/dma-mapping.c
652
op(vaddr + offset, len, dir);
arch/arm/mm/dma-mapping.c
659
op(vaddr + offset, len, dir);
arch/arm/mm/dma-mapping.c
666
op(vaddr, len, dir);
arch/arm/net/bpf_jit_32.c
1351
const u8 rn, struct jit_ctx *ctx, u8 op,
arch/arm/net/bpf_jit_32.c
1353
switch (op) {
arch/arm/net/bpf_jit_32.c
351
static u32 arm_bpf_ldst_imm12(u32 op, u8 rt, u8 rn, s16 imm12)
arch/arm/net/bpf_jit_32.c
353
op |= rt << 12 | rn << 16;
arch/arm/net/bpf_jit_32.c
355
op |= ARM_INST_LDST__U;
arch/arm/net/bpf_jit_32.c
358
return op | (imm12 & ARM_INST_LDST__IMM12);
arch/arm/net/bpf_jit_32.c
361
static u32 arm_bpf_ldst_imm8(u32 op, u8 rt, u8 rn, s16 imm8)
arch/arm/net/bpf_jit_32.c
363
op |= rt << 12 | rn << 16;
arch/arm/net/bpf_jit_32.c
365
op |= ARM_INST_LDST__U;
arch/arm/net/bpf_jit_32.c
368
return op | (imm8 & 0xf0) << 4 | (imm8 & 0x0f);
arch/arm/net/bpf_jit_32.c
520
static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op, u8 sign)
arch/arm/net/bpf_jit_32.c
528
if (op == BPF_DIV) {
arch/arm/net/bpf_jit_32.c
560
if (op == BPF_DIV)
arch/arm/net/bpf_jit_32.c
565
if (op == BPF_DIV)
arch/arm/net/bpf_jit_32.c
589
u8 op, u8 sign)
arch/arm/net/bpf_jit_32.c
628
if (op == BPF_DIV)
arch/arm/net/bpf_jit_32.c
633
if (op == BPF_DIV)
arch/arm/net/bpf_jit_32.c
801
const bool hi, const u8 op, struct jit_ctx *ctx){
arch/arm/net/bpf_jit_32.c
802
switch (BPF_OP(op)) {
arch/arm/net/bpf_jit_32.c
845
const u8 op) {
arch/arm/net/bpf_jit_32.c
857
emit_alu_r(rd[1], rs[1], true, false, op, ctx);
arch/arm/net/bpf_jit_32.c
858
emit_alu_r(rd[0], rs[0], true, true, op, ctx);
arch/arm/net/bpf_jit_32.c
865
emit_alu_r(rd[1], rs, true, false, op, ctx);
arch/arm/net/bpf_jit_32.c
966
struct jit_ctx *ctx, const u8 op) {
arch/arm/net/bpf_jit_32.c
973
switch (op) {
arch/arm/net/bpf_jit_32.h
165
#define _AL3_R(op, rd, rn, rm) ((op ## _R) | (rd) << 12 | (rn) << 16 | (rm))
arch/arm/net/bpf_jit_32.h
167
#define _AL3_I(op, rd, rn, imm) ((op ## _I) | (rd) << 12 | (rn) << 16 | (imm))
arch/arm/probes/kprobes/opt-arm.c
119
int arch_check_optimized_kprobe(struct optimized_kprobe *op)
arch/arm/probes/kprobes/opt-arm.c
140
__arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty)
arch/arm/probes/kprobes/opt-arm.c
142
if (op->optinsn.insn) {
arch/arm/probes/kprobes/opt-arm.c
143
free_optinsn_slot(op->optinsn.insn, dirty);
arch/arm/probes/kprobes/opt-arm.c
144
op->optinsn.insn = NULL;
arch/arm/probes/kprobes/opt-arm.c
149
optimized_callback(struct optimized_kprobe *op, struct pt_regs *regs)
arch/arm/probes/kprobes/opt-arm.c
152
struct kprobe *p = &op->kp;
arch/arm/probes/kprobes/opt-arm.c
156
regs->ARM_pc = (unsigned long)op->kp.addr;
arch/arm/probes/kprobes/opt-arm.c
163
kprobes_inc_nmissed_count(&op->kp);
arch/arm/probes/kprobes/opt-arm.c
165
__this_cpu_write(current_kprobe, &op->kp);
arch/arm/probes/kprobes/opt-arm.c
167
opt_pre_handler(&op->kp, regs);
arch/arm/probes/kprobes/opt-arm.c
176
op->kp.ainsn.insn_singlestep(p->opcode, &p->ainsn, regs);
arch/arm/probes/kprobes/opt-arm.c
182
int arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *orig)
arch/arm/probes/kprobes/opt-arm.c
252
val = (unsigned long)op;
arch/arm/probes/kprobes/opt-arm.c
264
(unsigned long)(op->kp.addr) + 4);
arch/arm/probes/kprobes/opt-arm.c
285
op->optinsn.insn = code;
arch/arm/probes/kprobes/opt-arm.c
291
struct optimized_kprobe *op, *tmp;
arch/arm/probes/kprobes/opt-arm.c
293
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/arm/probes/kprobes/opt-arm.c
295
WARN_ON(kprobe_disabled(&op->kp));
arch/arm/probes/kprobes/opt-arm.c
301
memcpy(op->optinsn.copied_insn, op->kp.addr,
arch/arm/probes/kprobes/opt-arm.c
304
insn = arm_gen_branch((unsigned long)op->kp.addr,
arch/arm/probes/kprobes/opt-arm.c
305
(unsigned long)op->optinsn.insn);
arch/arm/probes/kprobes/opt-arm.c
313
op->optinsn.copied_insn[0]) & 0xf0000000) |
arch/arm/probes/kprobes/opt-arm.c
321
kprobes_remove_breakpoint(op->kp.addr, insn);
arch/arm/probes/kprobes/opt-arm.c
323
list_del_init(&op->list);
arch/arm/probes/kprobes/opt-arm.c
327
void arch_unoptimize_kprobe(struct optimized_kprobe *op)
arch/arm/probes/kprobes/opt-arm.c
329
arch_arm_kprobe(&op->kp);
arch/arm/probes/kprobes/opt-arm.c
339
struct optimized_kprobe *op, *tmp;
arch/arm/probes/kprobes/opt-arm.c
341
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/arm/probes/kprobes/opt-arm.c
342
arch_unoptimize_kprobe(op);
arch/arm/probes/kprobes/opt-arm.c
343
list_move(&op->list, done_list);
arch/arm/probes/kprobes/opt-arm.c
347
int arch_within_optimized_kprobe(struct optimized_kprobe *op,
arch/arm/probes/kprobes/opt-arm.c
350
return (op->kp.addr <= addr &&
arch/arm/probes/kprobes/opt-arm.c
351
op->kp.addr + (RELATIVEJUMP_SIZE / sizeof(kprobe_opcode_t)) > addr);
arch/arm/probes/kprobes/opt-arm.c
355
void arch_remove_optimized_kprobe(struct optimized_kprobe *op)
arch/arm/probes/kprobes/opt-arm.c
357
__arch_remove_optimized_kprobe(op, 1);
arch/arm/probes/kprobes/test-arm.c
100
TEST_RR( op "le r13" ", r",14,val, ", ror r",14,7,"") \
arch/arm/probes/kprobes/test-arm.c
101
TEST_RR( op "gt r",0, val, ", r13" ", lsl r",14,8,"") \
arch/arm/probes/kprobes/test-arm.c
102
TEST_R( op "eq r",11,VAL1,", #0xf5") \
arch/arm/probes/kprobes/test-arm.c
103
TEST_R( op "ne r",0, VAL1,", #0xf5000000") \
arch/arm/probes/kprobes/test-arm.c
104
TEST_R( op " r",8, VAL2,", #0x000af000")
arch/arm/probes/kprobes/test-arm.c
106
#define _DATA_PROCESSING_DM(op,s,val) \
arch/arm/probes/kprobes/test-arm.c
107
TEST_R( op s "eq r0, r",1, val, "") \
arch/arm/probes/kprobes/test-arm.c
108
TEST_R( op s "ne r1, r",1, val, ", lsl #3") \
arch/arm/probes/kprobes/test-arm.c
109
TEST_R( op s "cs r2, r",3, val, ", lsr #4") \
arch/arm/probes/kprobes/test-arm.c
110
TEST_R( op s "cc r3, r",3, val, ", asr #5") \
arch/arm/probes/kprobes/test-arm.c
111
TEST_R( op s "mi r4, r",5, N(val),", asr #6") \
arch/arm/probes/kprobes/test-arm.c
112
TEST_R( op s "pl r5, r",5, val, ", ror #7") \
arch/arm/probes/kprobes/test-arm.c
113
TEST_R( op s "vs r6, r",10,val, ", rrx") \
arch/arm/probes/kprobes/test-arm.c
114
TEST( op s "vs r7, pc, lsl #3") \
arch/arm/probes/kprobes/test-arm.c
115
TEST( op s "vs r7, sp, lsr #4") \
arch/arm/probes/kprobes/test-arm.c
116
TEST_RR( op s "vc r8, r",7, val, ", lsl r",0, 3,"") \
arch/arm/probes/kprobes/test-arm.c
117
TEST_RR( op s "hi r9, r",9, val, ", lsr r",7, 4,"") \
arch/arm/probes/kprobes/test-arm.c
118
TEST_RR( op s "ls r10, r",9, val, ", asr r",7, 5,"") \
arch/arm/probes/kprobes/test-arm.c
119
TEST_RR( op s "ge r11, r",11,N(val),", asr r",7, 6,"") \
arch/arm/probes/kprobes/test-arm.c
120
TEST_RR( op s "lt r12, r",11,val, ", ror r",14,7,"") \
arch/arm/probes/kprobes/test-arm.c
121
TEST_R( op s "gt r14, r13" ", lsl r",14,8,"") \
arch/arm/probes/kprobes/test-arm.c
122
TEST( op s "eq r0, #0xf5") \
arch/arm/probes/kprobes/test-arm.c
123
TEST( op s "ne r11, #0xf5000000") \
arch/arm/probes/kprobes/test-arm.c
124
TEST( op s " r7, #0x000af000") \
arch/arm/probes/kprobes/test-arm.c
125
TEST( op s " r4, #0x00005a00")
arch/arm/probes/kprobes/test-arm.c
127
#define DATA_PROCESSING_DM(op,val) \
arch/arm/probes/kprobes/test-arm.c
128
_DATA_PROCESSING_DM(op,"",val) \
arch/arm/probes/kprobes/test-arm.c
129
_DATA_PROCESSING_DM(op,"s",val)
arch/arm/probes/kprobes/test-arm.c
57
#define _DATA_PROCESSING_DNM(op,s,val) \
arch/arm/probes/kprobes/test-arm.c
58
TEST_RR( op s "eq r0, r",1, VAL1,", r",2, val, "") \
arch/arm/probes/kprobes/test-arm.c
59
TEST_RR( op s "ne r1, r",1, VAL1,", r",2, val, ", lsl #3") \
arch/arm/probes/kprobes/test-arm.c
60
TEST_RR( op s "cs r2, r",3, VAL1,", r",2, val, ", lsr #4") \
arch/arm/probes/kprobes/test-arm.c
61
TEST_RR( op s "cc r3, r",3, VAL1,", r",2, val, ", asr #5") \
arch/arm/probes/kprobes/test-arm.c
62
TEST_RR( op s "mi r4, r",5, VAL1,", r",2, N(val),", asr #6") \
arch/arm/probes/kprobes/test-arm.c
63
TEST_RR( op s "pl r5, r",5, VAL1,", r",2, val, ", ror #7") \
arch/arm/probes/kprobes/test-arm.c
64
TEST_RR( op s "vs r6, r",7, VAL1,", r",2, val, ", rrx") \
arch/arm/probes/kprobes/test-arm.c
65
TEST_R( op s "vc r6, r",7, VAL1,", pc, lsl #3") \
arch/arm/probes/kprobes/test-arm.c
66
TEST_R( op s "vc r6, r",7, VAL1,", sp, lsr #4") \
arch/arm/probes/kprobes/test-arm.c
67
TEST_R( op s "vc r6, pc, r",7, VAL1,", asr #5") \
arch/arm/probes/kprobes/test-arm.c
68
TEST_R( op s "vc r6, sp, r",7, VAL1,", ror #6") \
arch/arm/probes/kprobes/test-arm.c
69
TEST_RRR( op s "hi r8, r",9, VAL1,", r",14,val, ", lsl r",0, 3,"")\
arch/arm/probes/kprobes/test-arm.c
70
TEST_RRR( op s "ls r9, r",9, VAL1,", r",14,val, ", lsr r",7, 4,"")\
arch/arm/probes/kprobes/test-arm.c
71
TEST_RRR( op s "ge r10, r",11,VAL1,", r",14,val, ", asr r",7, 5,"")\
arch/arm/probes/kprobes/test-arm.c
72
TEST_RRR( op s "lt r11, r",11,VAL1,", r",14,N(val),", asr r",7, 6,"")\
arch/arm/probes/kprobes/test-arm.c
73
TEST_RR( op s "gt r12, r13" ", r",14,val, ", ror r",14,7,"")\
arch/arm/probes/kprobes/test-arm.c
74
TEST_RR( op s "le r14, r",0, val, ", r13" ", lsl r",14,8,"")\
arch/arm/probes/kprobes/test-arm.c
75
TEST_R( op s "eq r0, r",11,VAL1,", #0xf5") \
arch/arm/probes/kprobes/test-arm.c
76
TEST_R( op s "ne r11, r",0, VAL1,", #0xf5000000") \
arch/arm/probes/kprobes/test-arm.c
77
TEST_R( op s " r7, r",8, VAL2,", #0x000af000") \
arch/arm/probes/kprobes/test-arm.c
78
TEST( op s " r4, pc" ", #0x00005a00")
arch/arm/probes/kprobes/test-arm.c
80
#define DATA_PROCESSING_DNM(op,val) \
arch/arm/probes/kprobes/test-arm.c
81
_DATA_PROCESSING_DNM(op,"",val) \
arch/arm/probes/kprobes/test-arm.c
82
_DATA_PROCESSING_DNM(op,"s",val)
arch/arm/probes/kprobes/test-arm.c
84
#define DATA_PROCESSING_NM(op,val) \
arch/arm/probes/kprobes/test-arm.c
85
TEST_RR( op "ne r",1, VAL1,", r",2, val, "") \
arch/arm/probes/kprobes/test-arm.c
86
TEST_RR( op "eq r",1, VAL1,", r",2, val, ", lsl #3") \
arch/arm/probes/kprobes/test-arm.c
87
TEST_RR( op "cc r",3, VAL1,", r",2, val, ", lsr #4") \
arch/arm/probes/kprobes/test-arm.c
88
TEST_RR( op "cs r",3, VAL1,", r",2, val, ", asr #5") \
arch/arm/probes/kprobes/test-arm.c
89
TEST_RR( op "pl r",5, VAL1,", r",2, N(val),", asr #6") \
arch/arm/probes/kprobes/test-arm.c
90
TEST_RR( op "mi r",5, VAL1,", r",2, val, ", ror #7") \
arch/arm/probes/kprobes/test-arm.c
91
TEST_RR( op "vc r",7, VAL1,", r",2, val, ", rrx") \
arch/arm/probes/kprobes/test-arm.c
92
TEST_R ( op "vs r",7, VAL1,", pc, lsl #3") \
arch/arm/probes/kprobes/test-arm.c
93
TEST_R ( op "vs r",7, VAL1,", sp, lsr #4") \
arch/arm/probes/kprobes/test-arm.c
94
TEST_R( op "vs pc, r",7, VAL1,", asr #5") \
arch/arm/probes/kprobes/test-arm.c
95
TEST_R( op "vs sp, r",7, VAL1,", ror #6") \
arch/arm/probes/kprobes/test-arm.c
96
TEST_RRR( op "ls r",9, VAL1,", r",14,val, ", lsl r",0, 3,"") \
arch/arm/probes/kprobes/test-arm.c
97
TEST_RRR( op "hi r",9, VAL1,", r",14,val, ", lsr r",7, 4,"") \
arch/arm/probes/kprobes/test-arm.c
98
TEST_RRR( op "lt r",11,VAL1,", r",14,val, ", asr r",7, 5,"") \
arch/arm/probes/kprobes/test-arm.c
99
TEST_RRR( op "ge r",11,VAL1,", r",14,N(val),", asr r",7, 6,"") \
arch/arm/probes/kprobes/test-thumb.c
478
#define _DATA_PROCESSING32_DNM(op,s,val) \
arch/arm/probes/kprobes/test-thumb.c
479
TEST_RR(op s".w r0, r",1, VAL1,", r",2, val, "") \
arch/arm/probes/kprobes/test-thumb.c
480
TEST_RR(op s" r1, r",1, VAL1,", r",2, val, ", lsl #3") \
arch/arm/probes/kprobes/test-thumb.c
481
TEST_RR(op s" r2, r",3, VAL1,", r",2, val, ", lsr #4") \
arch/arm/probes/kprobes/test-thumb.c
482
TEST_RR(op s" r3, r",3, VAL1,", r",2, val, ", asr #5") \
arch/arm/probes/kprobes/test-thumb.c
483
TEST_RR(op s" r4, r",5, VAL1,", r",2, N(val),", asr #6") \
arch/arm/probes/kprobes/test-thumb.c
484
TEST_RR(op s" r5, r",5, VAL1,", r",2, val, ", ror #7") \
arch/arm/probes/kprobes/test-thumb.c
485
TEST_RR(op s" r8, r",9, VAL1,", r",10,val, ", rrx") \
arch/arm/probes/kprobes/test-thumb.c
486
TEST_R( op s" r0, r",11,VAL1,", #0x00010001") \
arch/arm/probes/kprobes/test-thumb.c
487
TEST_R( op s" r11, r",0, VAL1,", #0xf5000000") \
arch/arm/probes/kprobes/test-thumb.c
488
TEST_R( op s" r7, r",8, VAL2,", #0x000af000")
arch/arm/probes/kprobes/test-thumb.c
490
#define DATA_PROCESSING32_DNM(op,val) \
arch/arm/probes/kprobes/test-thumb.c
491
_DATA_PROCESSING32_DNM(op,"",val) \
arch/arm/probes/kprobes/test-thumb.c
492
_DATA_PROCESSING32_DNM(op,"s",val)
arch/arm/probes/kprobes/test-thumb.c
494
#define DATA_PROCESSING32_NM(op,val) \
arch/arm/probes/kprobes/test-thumb.c
495
TEST_RR(op".w r",1, VAL1,", r",2, val, "") \
arch/arm/probes/kprobes/test-thumb.c
496
TEST_RR(op" r",1, VAL1,", r",2, val, ", lsl #3") \
arch/arm/probes/kprobes/test-thumb.c
497
TEST_RR(op" r",3, VAL1,", r",2, val, ", lsr #4") \
arch/arm/probes/kprobes/test-thumb.c
498
TEST_RR(op" r",3, VAL1,", r",2, val, ", asr #5") \
arch/arm/probes/kprobes/test-thumb.c
499
TEST_RR(op" r",5, VAL1,", r",2, N(val),", asr #6") \
arch/arm/probes/kprobes/test-thumb.c
500
TEST_RR(op" r",5, VAL1,", r",2, val, ", ror #7") \
arch/arm/probes/kprobes/test-thumb.c
501
TEST_RR(op" r",9, VAL1,", r",10,val, ", rrx") \
arch/arm/probes/kprobes/test-thumb.c
502
TEST_R( op" r",11,VAL1,", #0x00010001") \
arch/arm/probes/kprobes/test-thumb.c
503
TEST_R( op" r",0, VAL1,", #0xf5000000") \
arch/arm/probes/kprobes/test-thumb.c
504
TEST_R( op" r",8, VAL2,", #0x000af000")
arch/arm/probes/kprobes/test-thumb.c
506
#define _DATA_PROCESSING32_DM(op,s,val) \
arch/arm/probes/kprobes/test-thumb.c
507
TEST_R( op s".w r0, r",14, val, "") \
arch/arm/probes/kprobes/test-thumb.c
508
TEST_R( op s" r1, r",12, val, ", lsl #3") \
arch/arm/probes/kprobes/test-thumb.c
509
TEST_R( op s" r2, r",11, val, ", lsr #4") \
arch/arm/probes/kprobes/test-thumb.c
510
TEST_R( op s" r3, r",10, val, ", asr #5") \
arch/arm/probes/kprobes/test-thumb.c
511
TEST_R( op s" r4, r",9, N(val),", asr #6") \
arch/arm/probes/kprobes/test-thumb.c
512
TEST_R( op s" r5, r",8, val, ", ror #7") \
arch/arm/probes/kprobes/test-thumb.c
513
TEST_R( op s" r8, r",7,val, ", rrx") \
arch/arm/probes/kprobes/test-thumb.c
514
TEST( op s" r0, #0x00010001") \
arch/arm/probes/kprobes/test-thumb.c
515
TEST( op s" r11, #0xf5000000") \
arch/arm/probes/kprobes/test-thumb.c
516
TEST( op s" r7, #0x000af000") \
arch/arm/probes/kprobes/test-thumb.c
517
TEST( op s" r4, #0x00005a00")
arch/arm/probes/kprobes/test-thumb.c
519
#define DATA_PROCESSING32_DM(op,val) \
arch/arm/probes/kprobes/test-thumb.c
520
_DATA_PROCESSING32_DM(op,"",val) \
arch/arm/probes/kprobes/test-thumb.c
521
_DATA_PROCESSING32_DM(op,"s",val)
arch/arm/probes/kprobes/test-thumb.c
91
#define DATA_PROCESSING16(op,val) \
arch/arm/probes/kprobes/test-thumb.c
92
TEST_RR( op" r",0,VAL1,", r",7,val,"") \
arch/arm/probes/kprobes/test-thumb.c
93
TEST_RR( op" r",7,VAL2,", r",0,val,"")
arch/arm/probes/kprobes/test-thumb.c
934
#define SHIFTS32(op) \
arch/arm/probes/kprobes/test-thumb.c
935
TEST_RR(op" r0, r",1, VAL1,", r",2, 3, "") \
arch/arm/probes/kprobes/test-thumb.c
936
TEST_RR(op" r14, r",12,VAL2,", r",11,10,"")
arch/arm/probes/kprobes/test-thumb.c
986
#define PARALLEL_ADD_SUB(op) \
arch/arm/probes/kprobes/test-thumb.c
987
TEST_RR( op"add16 r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
988
TEST_RR( op"add16 r14, r",12,HH2,", r",10,HH1,"") \
arch/arm/probes/kprobes/test-thumb.c
989
TEST_RR( op"asx r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
990
TEST_RR( op"asx r14, r",12,HH2,", r",10,HH1,"") \
arch/arm/probes/kprobes/test-thumb.c
991
TEST_RR( op"sax r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
992
TEST_RR( op"sax r14, r",12,HH2,", r",10,HH1,"") \
arch/arm/probes/kprobes/test-thumb.c
993
TEST_RR( op"sub16 r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
994
TEST_RR( op"sub16 r14, r",12,HH2,", r",10,HH1,"") \
arch/arm/probes/kprobes/test-thumb.c
995
TEST_RR( op"add8 r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
996
TEST_RR( op"add8 r14, r",12,HH2,", r",10,HH1,"") \
arch/arm/probes/kprobes/test-thumb.c
997
TEST_RR( op"sub8 r0, r",0, HH1,", r",1, HH2,"") \
arch/arm/probes/kprobes/test-thumb.c
998
TEST_RR( op"sub8 r14, r",12,HH2,", r",10,HH1,"")
arch/arm/vfp/vfpdouble.c
1112
static struct op fops[16] = {
arch/arm/vfp/vfpdouble.c
1129
u32 op = inst & FOP_MASK;
arch/arm/vfp/vfpdouble.c
1135
struct op *fop;
arch/arm/vfp/vfpdouble.c
1139
fop = (op == FOP_EXT) ? &fops_ext[FEXT_TO_IDX(inst)] : &fops[FOP_TO_IDX(op)];
arch/arm/vfp/vfpdouble.c
1178
if (op == FOP_EXT)
arch/arm/vfp/vfpdouble.c
1185
type, dest, dn, FOP_TO_IDX(op), dm);
arch/arm/vfp/vfpdouble.c
661
static struct op fops_ext[32] = {
arch/arm/vfp/vfpsingle.c
1157
static struct op fops[16] = {
arch/arm/vfp/vfpsingle.c
1174
u32 op = inst & FOP_MASK;
arch/arm/vfp/vfpsingle.c
1180
struct op *fop;
arch/arm/vfp/vfpsingle.c
1184
fop = (op == FOP_EXT) ? &fops_ext[FEXT_TO_IDX(inst)] : &fops[FOP_TO_IDX(op)];
arch/arm/vfp/vfpsingle.c
1218
if (op == FOP_EXT)
arch/arm/vfp/vfpsingle.c
1225
FOP_TO_IDX(op), sm, m);
arch/arm/vfp/vfpsingle.c
706
static struct op fops_ext[32] = {
arch/arm/xen/enlighten.c
105
struct xen_platform_op op;
arch/arm/xen/enlighten.c
120
op.cmd = XENPF_settime64;
arch/arm/xen/enlighten.c
121
op.u.settime64.mbz = 0;
arch/arm/xen/enlighten.c
122
op.u.settime64.secs = now.tv_sec;
arch/arm/xen/enlighten.c
123
op.u.settime64.nsecs = now.tv_nsec;
arch/arm/xen/enlighten.c
124
op.u.settime64.system_time = timespec64_to_ns(&system_time);
arch/arm/xen/enlighten.c
125
(void)HYPERVISOR_platform_op(&op);
arch/arm/xen/mm.c
133
cflush.op = 0;
arch/arm/xen/mm.c
46
size_t size, u32 op)
arch/arm/xen/mm.c
51
cflush.op = op;
arch/arm64/include/asm/assembler.h
361
.macro __dcache_op_workaround_clean_cache, op, addr
arch/arm64/include/asm/assembler.h
381
.macro dcache_by_myline_op op, domain, start, end, linesz, tmp, fixup
arch/arm64/include/asm/assembler.h
431
.macro dcache_by_line_op op, domain, start, end, tmp1, tmp2, fixup
arch/arm64/include/asm/assembler.h
683
.macro __frame_regs, reg1, reg2, op, num
arch/arm64/include/asm/assembler.h
691
.macro __frame, op, regcount, extra=0
arch/arm64/include/asm/atomic.h
19
#define ATOMIC_OP(op) \
arch/arm64/include/asm/atomic.h
20
static __always_inline void arch_##op(int i, atomic_t *v) \
arch/arm64/include/asm/atomic.h
22
__lse_ll_sc_body(op, i, v); \
arch/arm64/include/asm/atomic.h
34
#define ATOMIC_FETCH_OP(name, op) \
arch/arm64/include/asm/atomic.h
35
static __always_inline int arch_##op##name(int i, atomic_t *v) \
arch/arm64/include/asm/atomic.h
37
return __lse_ll_sc_body(op##name, i, v); \
arch/arm64/include/asm/atomic.h
40
#define ATOMIC_FETCH_OPS(op) \
arch/arm64/include/asm/atomic.h
41
ATOMIC_FETCH_OP(_relaxed, op) \
arch/arm64/include/asm/atomic.h
42
ATOMIC_FETCH_OP(_acquire, op) \
arch/arm64/include/asm/atomic.h
43
ATOMIC_FETCH_OP(_release, op) \
arch/arm64/include/asm/atomic.h
44
ATOMIC_FETCH_OP( , op)
arch/arm64/include/asm/atomic.h
58
#define ATOMIC64_OP(op) \
arch/arm64/include/asm/atomic.h
59
static __always_inline void arch_##op(long i, atomic64_t *v) \
arch/arm64/include/asm/atomic.h
61
__lse_ll_sc_body(op, i, v); \
arch/arm64/include/asm/atomic.h
73
#define ATOMIC64_FETCH_OP(name, op) \
arch/arm64/include/asm/atomic.h
74
static __always_inline long arch_##op##name(long i, atomic64_t *v) \
arch/arm64/include/asm/atomic.h
76
return __lse_ll_sc_body(op##name, i, v); \
arch/arm64/include/asm/atomic.h
79
#define ATOMIC64_FETCH_OPS(op) \
arch/arm64/include/asm/atomic.h
80
ATOMIC64_FETCH_OP(_relaxed, op) \
arch/arm64/include/asm/atomic.h
81
ATOMIC64_FETCH_OP(_acquire, op) \
arch/arm64/include/asm/atomic.h
82
ATOMIC64_FETCH_OP(_release, op) \
arch/arm64/include/asm/atomic.h
83
ATOMIC64_FETCH_OP( , op)
arch/arm64/include/asm/atomic_ll_sc.h
121
#define ATOMIC64_OP(op, asm_op, constraint) \
arch/arm64/include/asm/atomic_ll_sc.h
123
__ll_sc_atomic64_##op(s64 i, atomic64_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
128
asm volatile("// atomic64_" #op "\n" \
arch/arm64/include/asm/atomic_ll_sc.h
138
#define ATOMIC64_OP_RETURN(name, mb, acq, rel, cl, op, asm_op, constraint)\
arch/arm64/include/asm/atomic_ll_sc.h
140
__ll_sc_atomic64_##op##_return##name(s64 i, atomic64_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
145
asm volatile("// atomic64_" #op "_return" #name "\n" \
arch/arm64/include/asm/atomic_ll_sc.h
159
#define ATOMIC64_FETCH_OP(name, mb, acq, rel, cl, op, asm_op, constraint)\
arch/arm64/include/asm/atomic_ll_sc.h
161
__ll_sc_atomic64_fetch_##op##name(s64 i, atomic64_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
166
asm volatile("// atomic64_fetch_" #op #name "\n" \
arch/arm64/include/asm/atomic_ll_sc.h
25
#define ATOMIC_OP(op, asm_op, constraint) \
arch/arm64/include/asm/atomic_ll_sc.h
27
__ll_sc_atomic_##op(int i, atomic_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
32
asm volatile("// atomic_" #op "\n" \
arch/arm64/include/asm/atomic_ll_sc.h
42
#define ATOMIC_OP_RETURN(name, mb, acq, rel, cl, op, asm_op, constraint)\
arch/arm64/include/asm/atomic_ll_sc.h
44
__ll_sc_atomic_##op##_return##name(int i, atomic_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
49
asm volatile("// atomic_" #op "_return" #name "\n" \
arch/arm64/include/asm/atomic_ll_sc.h
63
#define ATOMIC_FETCH_OP(name, mb, acq, rel, cl, op, asm_op, constraint) \
arch/arm64/include/asm/atomic_ll_sc.h
65
__ll_sc_atomic_fetch_##op##name(int i, atomic_t *v) \
arch/arm64/include/asm/atomic_ll_sc.h
70
asm volatile("// atomic_fetch_" #op #name "\n" \
arch/arm64/include/asm/atomic_lse.h
120
#define ATOMIC64_OP(op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
122
__lse_atomic64_##op(s64 i, atomic64_t *v) \
arch/arm64/include/asm/atomic_lse.h
13
#define ATOMIC_OP(op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
143
#define ATOMIC64_FETCH_OP(name, mb, op, asm_op, cl...) \
arch/arm64/include/asm/atomic_lse.h
145
__lse_atomic64_fetch_##op##name(s64 i, atomic64_t *v) \
arch/arm64/include/asm/atomic_lse.h
15
__lse_atomic_##op(int i, atomic_t *v) \
arch/arm64/include/asm/atomic_lse.h
160
#define ATOMIC64_FETCH_OPS(op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
161
ATOMIC64_FETCH_OP(_relaxed, , op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
162
ATOMIC64_FETCH_OP(_acquire, a, op, asm_op, "memory") \
arch/arm64/include/asm/atomic_lse.h
163
ATOMIC64_FETCH_OP(_release, l, op, asm_op, "memory") \
arch/arm64/include/asm/atomic_lse.h
164
ATOMIC64_FETCH_OP( , al, op, asm_op, "memory")
arch/arm64/include/asm/atomic_lse.h
36
#define ATOMIC_FETCH_OP(name, mb, op, asm_op, cl...) \
arch/arm64/include/asm/atomic_lse.h
38
__lse_atomic_fetch_##op##name(int i, atomic_t *v) \
arch/arm64/include/asm/atomic_lse.h
53
#define ATOMIC_FETCH_OPS(op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
54
ATOMIC_FETCH_OP(_relaxed, , op, asm_op) \
arch/arm64/include/asm/atomic_lse.h
55
ATOMIC_FETCH_OP(_acquire, a, op, asm_op, "memory") \
arch/arm64/include/asm/atomic_lse.h
56
ATOMIC_FETCH_OP(_release, l, op, asm_op, "memory") \
arch/arm64/include/asm/atomic_lse.h
57
ATOMIC_FETCH_OP( , al, op, asm_op, "memory")
arch/arm64/include/asm/ftrace.h
173
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/arm64/include/asm/futex.h
41
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *_uaddr)
arch/arm64/include/asm/futex.h
49
switch (op) {
arch/arm64/include/asm/insn.h
573
aarch64_insn_gen_hint(enum aarch64_insn_hint_cr_op op)
arch/arm64/include/asm/insn.h
575
return aarch64_insn_get_hint_value() | op;
arch/arm64/include/asm/insn.h
678
enum aarch64_insn_mem_atomic_op op,
arch/arm64/include/asm/kvm_asm.h
249
extern int __kvm_at_s1e01(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_asm.h
250
extern int __kvm_at_s1e2(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_asm.h
251
extern int __kvm_at_s12(struct kvm_vcpu *vcpu, u32 op, u64 vaddr);
arch/arm64/include/asm/kvm_host.h
1179
#define __vcpu_rmw_sys_reg(v, r, op, val) \
arch/arm64/include/asm/kvm_host.h
1183
__v op (val); \
arch/arm64/include/asm/kvm_host.h
1571
#define kvm_cmp_feat_signed(kvm, id, fld, op, limit) \
arch/arm64/include/asm/kvm_host.h
1572
(get_idreg_field_signed((kvm), id, fld) op __expand_field_sign_signed(id, fld, limit))
arch/arm64/include/asm/kvm_host.h
1574
#define kvm_cmp_feat_unsigned(kvm, id, fld, op, limit) \
arch/arm64/include/asm/kvm_host.h
1575
(get_idreg_field_unsigned((kvm), id, fld) op __expand_field_sign_unsigned(id, fld, limit))
arch/arm64/include/asm/kvm_host.h
1577
#define kvm_cmp_feat(kvm, id, fld, op, limit) \
arch/arm64/include/asm/kvm_host.h
1579
kvm_cmp_feat_signed(kvm, id, fld, op, limit) : \
arch/arm64/include/asm/kvm_host.h
1580
kvm_cmp_feat_unsigned(kvm, id, fld, op, limit))
arch/arm64/include/asm/lse.h
17
#define __lse_ll_sc_body(op, ...) \
arch/arm64/include/asm/lse.h
20
__lse_##op(__VA_ARGS__) : \
arch/arm64/include/asm/lse.h
21
__ll_sc_##op(__VA_ARGS__); \
arch/arm64/include/asm/percpu.h
156
#define _pcp_protect(op, pcp, ...) \
arch/arm64/include/asm/percpu.h
159
op(raw_cpu_ptr(&(pcp)), __VA_ARGS__); \
arch/arm64/include/asm/percpu.h
163
#define _pcp_protect_return(op, pcp, args...) \
arch/arm64/include/asm/percpu.h
167
__retval = (typeof(pcp))op(raw_cpu_ptr(&(pcp)), ##args); \
arch/arm64/include/asm/tlbflush.h
100
#define __tlbi_level(op, addr, level) do { \
arch/arm64/include/asm/tlbflush.h
111
__tlbi(op, arg); \
arch/arm64/include/asm/tlbflush.h
114
#define __tlbi_user_level(op, arg, level) do { \
arch/arm64/include/asm/tlbflush.h
116
__tlbi_level(op, (arg | USER_ASID_FLAG), level); \
arch/arm64/include/asm/tlbflush.h
176
#define __repeat_tlbi_sync(op, arg...) \
arch/arm64/include/asm/tlbflush.h
180
__tlbi(op, ##arg); \
arch/arm64/include/asm/tlbflush.h
32
#define __TLBI_0(op, arg) asm (ARM64_ASM_PREAMBLE \
arch/arm64/include/asm/tlbflush.h
33
"tlbi " #op "\n" \
arch/arm64/include/asm/tlbflush.h
36
#define __TLBI_1(op, arg) asm (ARM64_ASM_PREAMBLE \
arch/arm64/include/asm/tlbflush.h
37
"tlbi " #op ", %x0\n" \
arch/arm64/include/asm/tlbflush.h
40
#define __TLBI_N(op, arg, n, ...) __TLBI_##n(op, arg)
arch/arm64/include/asm/tlbflush.h
42
#define __tlbi(op, ...) __TLBI_N(op, ##__VA_ARGS__, 1, 0)
arch/arm64/include/asm/tlbflush.h
430
#define __flush_tlb_range_op(op, start, pages, stride, \
arch/arm64/include/asm/tlbflush.h
44
#define __tlbi_user(op, arg) do { \
arch/arm64/include/asm/tlbflush.h
445
__tlbi_level(op, addr, tlb_level); \
arch/arm64/include/asm/tlbflush.h
447
__tlbi_user_level(op, addr, tlb_level); \
arch/arm64/include/asm/tlbflush.h
457
__tlbi(r##op, addr); \
arch/arm64/include/asm/tlbflush.h
459
__tlbi_user(r##op, addr); \
arch/arm64/include/asm/tlbflush.h
46
__tlbi(op, (arg) | USER_ASID_FLAG); \
arch/arm64/include/asm/tlbflush.h
467
#define __flush_s2_tlb_range_op(op, start, pages, stride, tlb_level) \
arch/arm64/include/asm/tlbflush.h
468
__flush_tlb_range_op(op, start, pages, stride, 0, tlb_level, false, kvm_lpa2_is_enabled());
arch/arm64/kernel/ftrace.c
551
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/arm64/kernel/module.c
107
pr_err("Invalid 32-bit data relocation (%d)\n", op);
arch/arm64/kernel/module.c
126
static int reloc_insn_movw(enum aarch64_reloc_op op, __le32 *place, u64 val,
arch/arm64/kernel/module.c
134
sval = do_reloc(op, place, val);
arch/arm64/kernel/module.c
168
static int reloc_insn_imm(enum aarch64_reloc_op op, __le32 *place, u64 val,
arch/arm64/kernel/module.c
177
sval = do_reloc(op, place, val);
arch/arm64/kernel/module.c
61
static int reloc_data(enum aarch64_reloc_op op, void *place, u64 val, int len,
arch/arm64/kernel/module.c
64
s64 sval = do_reloc(op, place, val);
arch/arm64/kernel/module.c
81
switch (op) {
arch/arm64/kernel/module.c
91
pr_err("Invalid 16-bit data relocation (%d)\n", op);
arch/arm64/kernel/module.c
97
switch (op) {
arch/arm64/kvm/at.c
1250
static int handle_at_slow(struct kvm_vcpu *vcpu, u32 op, u64 vaddr, u64 *par)
arch/arm64/kvm/at.c
1257
wi.regime = compute_translation_regime(vcpu, op);
arch/arm64/kvm/at.c
1258
wi.as_el0 = (op == OP_AT_S1E0R || op == OP_AT_S1E0W);
arch/arm64/kvm/at.c
1259
wi.pan = (op == OP_AT_S1E1RP || op == OP_AT_S1E1WP) &&
arch/arm64/kvm/at.c
1285
switch (op) {
arch/arm64/kvm/at.c
1326
static u64 __kvm_at_s1e01_fast(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1389
switch (op) {
arch/arm64/kvm/at.c
1392
fail = at_s1e1p_fast(vcpu, op, vaddr);
arch/arm64/kvm/at.c
1442
int __kvm_at_s1e01(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1444
u64 par = __kvm_at_s1e01_fast(vcpu, op, vaddr);
arch/arm64/kvm/at.c
1458
ret = handle_at_slow(vcpu, op, vaddr, &par);
arch/arm64/kvm/at.c
1467
int __kvm_at_s1e2(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1492
switch (op) {
arch/arm64/kvm/at.c
1516
ret = handle_at_slow(vcpu, op, vaddr, &par);
arch/arm64/kvm/at.c
1525
int __kvm_at_s12(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
1533
switch (op) {
arch/arm64/kvm/at.c
1535
op = OP_AT_S1E1R;
arch/arm64/kvm/at.c
1539
op = OP_AT_S1E1W;
arch/arm64/kvm/at.c
1543
op = OP_AT_S1E0R;
arch/arm64/kvm/at.c
1547
op = OP_AT_S1E0W;
arch/arm64/kvm/at.c
1555
__kvm_at_s1e01(vcpu, op, vaddr);
arch/arm64/kvm/at.c
1564
if (compute_translation_regime(vcpu, op) == TR_EL20 ||
arch/arm64/kvm/at.c
621
static bool at_s1e1p_fast(struct kvm_vcpu *vcpu, u32 op, u64 vaddr)
arch/arm64/kvm/at.c
629
switch (op) {
arch/arm64/kvm/at.c
83
static enum trans_regime compute_translation_regime(struct kvm_vcpu *vcpu, u32 op)
arch/arm64/kvm/at.c
89
switch (op) {
arch/arm64/kvm/hyp/nvhe/gen-hyprel.c
180
#define assert_op(lhs, rhs, fmt, op) \
arch/arm64/kvm/hyp/nvhe/gen-hyprel.c
185
if (!(_lhs op _rhs)) { \
arch/arm64/kvm/hyp/nvhe/gen-hyprel.c
186
fatal_error("assertion " #lhs " " #op " " #rhs \
arch/arm64/kvm/sys_regs.c
3787
u32 op = sys_insn(p->Op0, p->Op1, p->CRn, p->CRm, p->Op2);
arch/arm64/kvm/sys_regs.c
3789
if (__kvm_at_s1e01(vcpu, op, p->regval))
arch/arm64/kvm/sys_regs.c
3798
u32 op = sys_insn(p->Op0, p->Op1, p->CRn, p->CRm, p->Op2);
arch/arm64/kvm/sys_regs.c
3801
if (op == OP_AT_S1E2A &&
arch/arm64/kvm/sys_regs.c
3807
if (__kvm_at_s1e2(vcpu, op, p->regval))
arch/arm64/kvm/sys_regs.c
3816
u32 op = sys_insn(p->Op0, p->Op1, p->CRn, p->CRm, p->Op2);
arch/arm64/kvm/sys_regs.c
3818
if (__kvm_at_s12(vcpu, op, p->regval))
arch/arm64/lib/insn.c
647
enum aarch64_insn_mem_atomic_op op,
arch/arm64/lib/insn.c
652
switch (op) {
arch/arm64/lib/insn.c
669
pr_err("%s: unimplemented mem atomic op %d\n", __func__, op);
arch/arm64/net/bpf_jit.h
148
#define A64_ST_OP(sf, Rn, Rs, op) \
arch/arm64/net/bpf_jit.h
150
A64_SIZE(sf), AARCH64_INSN_MEM_ATOMIC_##op, \
arch/arm64/net/bpf_jit.h
158
#define A64_LD_OP_AL(sf, Rt, Rn, Rs, op) \
arch/arm64/net/bpf_jit.h
160
A64_SIZE(sf), AARCH64_INSN_MEM_ATOMIC_##op, \
arch/csky/include/asm/atomic.h
25
#define ATOMIC_OP(op) \
arch/csky/include/asm/atomic.h
27
void arch_atomic_##op(int i, atomic_t *v) \
arch/csky/include/asm/atomic.h
32
" " #op " %0, %1 \n" \
arch/csky/include/asm/atomic.h
48
#define ATOMIC_FETCH_OP(op) \
arch/csky/include/asm/atomic.h
50
int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \
arch/csky/include/asm/atomic.h
56
" " #op " %0, %2 \n" \
arch/csky/include/asm/atomic.h
65
#define ATOMIC_OP_RETURN(op, c_op) \
arch/csky/include/asm/atomic.h
67
int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \
arch/csky/include/asm/atomic.h
69
return arch_atomic_fetch_##op##_relaxed(i, v) c_op i; \
arch/csky/include/asm/atomic.h
72
#define ATOMIC_OPS(op, c_op) \
arch/csky/include/asm/atomic.h
73
ATOMIC_FETCH_OP(op) \
arch/csky/include/asm/atomic.h
74
ATOMIC_OP_RETURN(op, c_op)
arch/csky/include/asm/atomic.h
88
#define ATOMIC_OPS(op) \
arch/csky/include/asm/atomic.h
89
ATOMIC_FETCH_OP(op)
arch/csky/include/asm/futex.h
35
: [op] "Jr" (oparg), [e] "jr" (-EFAULT) \
arch/csky/include/asm/futex.h
42
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/csky/include/asm/futex.h
49
switch (op) {
arch/hexagon/include/asm/atomic.h
33
#define ATOMIC_OP(op) \
arch/hexagon/include/asm/atomic.h
34
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/hexagon/include/asm/atomic.h
40
" %0 = "#op "(%0,%2);\n" \
arch/hexagon/include/asm/atomic.h
49
#define ATOMIC_OP_RETURN(op) \
arch/hexagon/include/asm/atomic.h
50
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/hexagon/include/asm/atomic.h
56
" %0 = "#op "(%0,%2);\n" \
arch/hexagon/include/asm/atomic.h
66
#define ATOMIC_FETCH_OP(op) \
arch/hexagon/include/asm/atomic.h
67
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/hexagon/include/asm/atomic.h
73
" %1 = "#op "(%0,%3);\n" \
arch/hexagon/include/asm/atomic.h
83
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) ATOMIC_FETCH_OP(op)
arch/hexagon/include/asm/atomic.h
94
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op)
arch/hexagon/include/asm/futex.h
35
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/hexagon/include/asm/futex.h
42
switch (op) {
arch/hexagon/include/asm/hexagon_vm.h
77
long __vmcache(enum VM_CACHE_OPS op, unsigned long addr, unsigned long len);
arch/loongarch/include/asm/atomic-amo.h
111
#define ATOMIC64_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
112
static inline void arch_atomic64_##op(long i, atomic64_t *v) \
arch/loongarch/include/asm/atomic-amo.h
121
#define ATOMIC64_OP_RETURN(op, I, asm_op, c_op, mb, suffix) \
arch/loongarch/include/asm/atomic-amo.h
122
static inline long arch_atomic64_##op##_return##suffix(long i, atomic64_t *v) \
arch/loongarch/include/asm/atomic-amo.h
134
#define ATOMIC64_FETCH_OP(op, I, asm_op, mb, suffix) \
arch/loongarch/include/asm/atomic-amo.h
135
static inline long arch_atomic64_fetch_##op##suffix(long i, atomic64_t *v) \
arch/loongarch/include/asm/atomic-amo.h
148
#define ATOMIC64_OPS(op, I, asm_op, c_op) \
arch/loongarch/include/asm/atomic-amo.h
149
ATOMIC64_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
15
#define ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
150
ATOMIC64_OP_RETURN(op, I, asm_op, c_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
151
ATOMIC64_OP_RETURN(op, I, asm_op, c_op, , _relaxed) \
arch/loongarch/include/asm/atomic-amo.h
152
ATOMIC64_FETCH_OP(op, I, asm_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
153
ATOMIC64_FETCH_OP(op, I, asm_op, , _relaxed)
arch/loongarch/include/asm/atomic-amo.h
16
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-amo.h
177
#define ATOMIC64_OPS(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
178
ATOMIC64_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
179
ATOMIC64_FETCH_OP(op, I, asm_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
180
ATOMIC64_FETCH_OP(op, I, asm_op, , _relaxed)
arch/loongarch/include/asm/atomic-amo.h
25
#define ATOMIC_OP_RETURN(op, I, asm_op, c_op, mb, suffix) \
arch/loongarch/include/asm/atomic-amo.h
26
static inline int arch_atomic_##op##_return##suffix(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-amo.h
39
#define ATOMIC_FETCH_OP(op, I, asm_op, mb, suffix) \
arch/loongarch/include/asm/atomic-amo.h
40
static inline int arch_atomic_fetch_##op##suffix(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-amo.h
53
#define ATOMIC_OPS(op, I, asm_op, c_op) \
arch/loongarch/include/asm/atomic-amo.h
54
ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
55
ATOMIC_OP_RETURN(op, I, asm_op, c_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
56
ATOMIC_OP_RETURN(op, I, asm_op, c_op, , _relaxed) \
arch/loongarch/include/asm/atomic-amo.h
57
ATOMIC_FETCH_OP(op, I, asm_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
58
ATOMIC_FETCH_OP(op, I, asm_op, , _relaxed)
arch/loongarch/include/asm/atomic-amo.h
82
#define ATOMIC_OPS(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
83
ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-amo.h
84
ATOMIC_FETCH_OP(op, I, asm_op, _db, ) \
arch/loongarch/include/asm/atomic-amo.h
85
ATOMIC_FETCH_OP(op, I, asm_op, , _relaxed)
arch/loongarch/include/asm/atomic-llsc.h
15
#define ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
16
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-llsc.h
21
"1: ll.w %0, %1 #atomic_" #op " \n" \
arch/loongarch/include/asm/atomic-llsc.h
30
#define ATOMIC_OP_RETURN(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
31
static inline int arch_atomic_##op##_return_relaxed(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-llsc.h
36
"1: ll.w %1, %2 # atomic_" #op "_return \n" \
arch/loongarch/include/asm/atomic-llsc.h
47
#define ATOMIC_FETCH_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
48
static inline int arch_atomic_fetch_##op##_relaxed(int i, atomic_t *v) \
arch/loongarch/include/asm/atomic-llsc.h
53
"1: ll.w %1, %2 # atomic_fetch_" #op " \n" \
arch/loongarch/include/asm/atomic-llsc.h
64
#define ATOMIC_OPS(op,I ,asm_op, c_op) \
arch/loongarch/include/asm/atomic-llsc.h
65
ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
66
ATOMIC_OP_RETURN(op, I , asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
67
ATOMIC_FETCH_OP(op, I, asm_op)
arch/loongarch/include/asm/atomic-llsc.h
79
#define ATOMIC_OPS(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
80
ATOMIC_OP(op, I, asm_op) \
arch/loongarch/include/asm/atomic-llsc.h
81
ATOMIC_FETCH_OP(op, I, asm_op)
arch/loongarch/include/asm/cacheflush.h
51
#define cache_op(op, addr) \
arch/loongarch/include/asm/cacheflush.h
55
: "i" (op), "ZC" (*(unsigned char *)(addr)))
arch/loongarch/include/asm/ftrace.h
72
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/loongarch/include/asm/futex.h
31
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/loongarch/include/asm/futex.h
37
switch (op) {
arch/loongarch/include/asm/percpu.h
41
#define PERCPU_OP(op, asm_op, c_op) \
arch/loongarch/include/asm/percpu.h
42
static __always_inline unsigned long __percpu_##op(void *ptr, \
arch/loongarch/include/asm/percpu.h
77
#define __pcpu_op_1(op) op ".b "
arch/loongarch/include/asm/percpu.h
78
#define __pcpu_op_2(op) op ".h "
arch/loongarch/include/asm/percpu.h
79
#define __pcpu_op_4(op) op ".w "
arch/loongarch/include/asm/percpu.h
80
#define __pcpu_op_8(op) op ".d "
arch/loongarch/include/asm/tlb.h
101
static __always_inline void invtlb_addr(u32 op, u32 info, u64 addr)
arch/loongarch/include/asm/tlb.h
107
: "i"(op), "r"(addr)
arch/loongarch/include/asm/tlb.h
112
static __always_inline void invtlb_info(u32 op, u32 info, u64 addr)
arch/loongarch/include/asm/tlb.h
118
: "i"(op), "r"(info)
arch/loongarch/include/asm/tlb.h
123
static __always_inline void invtlb_all(u32 op, u32 info, u64 addr)
arch/loongarch/include/asm/tlb.h
130
: "i"(op)
arch/loongarch/include/asm/tlb.h
91
static __always_inline void invtlb(u32 op, u32 info, u64 addr)
arch/loongarch/include/asm/tlb.h
96
: "i"(op), "r"(info), "r"(addr)
arch/loongarch/kernel/ftrace_dyn.c
241
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/loongarch/kvm/trace.h
141
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op,
arch/loongarch/kvm/trace.h
143
TP_ARGS(vcpu, op, state),
arch/loongarch/kvm/trace.h
146
__field(u8, op)
arch/loongarch/kvm/trace.h
152
__entry->op = op;
arch/loongarch/kvm/trace.h
157
__print_symbolic(__entry->op,
arch/m68k/apollo/config.c
181
int dn_dummy_hwclk(int op, struct rtc_time *t) {
arch/m68k/apollo/config.c
184
if(!op) { /* read */
arch/m68k/atari/atari.h
14
int atari_mste_hwclk(int op, struct rtc_time *t);
arch/m68k/atari/atari.h
15
int atari_tt_hwclk(int op, struct rtc_time *t);
arch/m68k/atari/time.c
142
int atari_mste_hwclk( int op, struct rtc_time *t )
arch/m68k/atari/time.c
152
if (op) {
arch/m68k/atari/time.c
201
int atari_tt_hwclk( int op, struct rtc_time *t )
arch/m68k/atari/time.c
211
if (op) {
arch/m68k/atari/time.c
265
if (!op) {
arch/m68k/atari/time.c
286
if (!op) {
arch/m68k/bvme6000/config.c
283
int bvme6000_hwclk(int op, struct rtc_time *t)
arch/m68k/bvme6000/config.c
290
if (op)
arch/m68k/hp300/config.c
202
static int hp300_hwclk(int op, struct rtc_time *t)
arch/m68k/hp300/config.c
204
if (!op) { /* read */
arch/m68k/ifpsp060/src/fpsp.S
18235
# if the bit string is a zero, then the operation is a no-op
arch/m68k/ifpsp060/src/pfpsp.S
4284
# if the bit string is a zero, then the operation is a no-op
arch/m68k/include/asm/atomic.h
101
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
102
ATOMIC_OP(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
103
ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
104
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/m68k/include/asm/atomic.h
115
#define ATOMIC_OPS(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
116
ATOMIC_OP(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
117
ATOMIC_FETCH_OP(op, c_op, asm_op)
arch/m68k/include/asm/atomic.h
32
#define ATOMIC_OP(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
33
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/m68k/include/asm/atomic.h
40
#define ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
41
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/m68k/include/asm/atomic.h
55
#define ATOMIC_FETCH_OP(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
56
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/m68k/include/asm/atomic.h
72
#define ATOMIC_OP_RETURN(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
73
static inline int arch_atomic_##op##_return(int i, atomic_t * v) \
arch/m68k/include/asm/atomic.h
85
#define ATOMIC_FETCH_OP(op, c_op, asm_op) \
arch/m68k/include/asm/atomic.h
86
static inline int arch_atomic_fetch_##op(int i, atomic_t * v) \
arch/m68k/mac/mac.h
12
int mac_hwclk(int op, struct rtc_time *t);
arch/m68k/mac/misc.c
610
int mac_hwclk(int op, struct rtc_time *t)
arch/m68k/mac/misc.c
614
if (!op) { /* read */
arch/m68k/q40/config.c
203
static int q40_hwclk(int op, struct rtc_time *t)
arch/m68k/q40/config.c
205
if (op) {
arch/microblaze/include/asm/futex.h
33
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/microblaze/include/asm/futex.h
40
switch (op) {
arch/microblaze/kernel/cpu/cache.c
104
#define CACHE_ALL_LOOP(cache_size, line_length, op) \
arch/microblaze/kernel/cpu/cache.c
110
__asm__ __volatile__ (" 1: " #op " %0, r0;" \
arch/microblaze/kernel/cpu/cache.c
125
#define CACHE_RANGE_LOOP_2(start, end, line_length, op) \
arch/microblaze/kernel/cpu/cache.c
134
__asm__ __volatile__ (" 1: " #op " %0, %1;" \
arch/microblaze/kernel/cpu/cache.c
142
#define CACHE_RANGE_LOOP_1(start, end, line_length, op) \
arch/microblaze/kernel/cpu/cache.c
149
__asm__ __volatile__ (" 1: " #op " %1, r0;" \
arch/mips/include/asm/asm-eva.h
104
#define kernel_cache(op, base) cache op, base
arch/mips/include/asm/asm-eva.h
141
#define user_cache(op, base) __BUILD_EVA_INSN(cachee, op, base)
arch/mips/include/asm/asm-eva.h
162
#define user_cache(op, base) kernel_cache(op, base)
arch/mips/include/asm/asm-eva.h
17
#define kernel_cache(op, base) "cache " op ", " base "\n"
arch/mips/include/asm/asm-eva.h
54
#define user_cache(op, base) __BUILD_EVA_INSN("cachee", op, base)
arch/mips/include/asm/asm-eva.h
76
#define user_cache(op, base) kernel_cache(op, base)
arch/mips/include/asm/atomic.h
105
#define ATOMIC_FETCH_OP(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
107
arch_##pfx##_fetch_##op##_relaxed(type i, pfx##_t * v) \
arch/mips/include/asm/atomic.h
125
"1: " #ll " %1, %2 # " #pfx "_fetch_" #op "\n" \
arch/mips/include/asm/atomic.h
139
#define ATOMIC_OPS(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
140
ATOMIC_OP(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
141
ATOMIC_OP_RETURN(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
142
ATOMIC_FETCH_OP(pfx, op, type, c_op, asm_op, ll, sc)
arch/mips/include/asm/atomic.h
162
#define ATOMIC_OPS(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
163
ATOMIC_OP(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
164
ATOMIC_FETCH_OP(pfx, op, type, c_op, asm_op, ll, sc)
arch/mips/include/asm/atomic.h
196
#define ATOMIC_SIP_OP(pfx, type, op, ll, sc) \
arch/mips/include/asm/atomic.h
222
" " #op " %0, %1, %3 \n" \
arch/mips/include/asm/atomic.h
44
#define ATOMIC_OP(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
45
static __inline__ void arch_##pfx##_##op(type i, pfx##_t * v) \
arch/mips/include/asm/atomic.h
62
"1: " #ll " %0, %1 # " #pfx "_" #op " \n" \
arch/mips/include/asm/atomic.h
71
#define ATOMIC_OP_RETURN(pfx, op, type, c_op, asm_op, ll, sc) \
arch/mips/include/asm/atomic.h
73
arch_##pfx##_##op##_return_relaxed(type i, pfx##_t * v) \
arch/mips/include/asm/atomic.h
92
"1: " #ll " %1, %2 # " #pfx "_" #op "_return\n" \
arch/mips/include/asm/futex.h
100
switch (op) {
arch/mips/include/asm/futex.h
102
__futex_atomic_op(op, "move $1, %z5", ret, oldval, uaddr, oparg);
arch/mips/include/asm/futex.h
106
__futex_atomic_op(op, "addu $1, %1, %z5",
arch/mips/include/asm/futex.h
110
__futex_atomic_op(op, "or $1, %1, %z5",
arch/mips/include/asm/futex.h
114
__futex_atomic_op(op, "and $1, %1, %z5",
arch/mips/include/asm/futex.h
118
__futex_atomic_op(op, "xor $1, %1, %z5",
arch/mips/include/asm/futex.h
25
#define __futex_atomic_op(op, insn, ret, oldval, uaddr, oparg) \
arch/mips/include/asm/futex.h
88
ret = futex_atomic_op_inuser_local(op, oparg, oval, uaddr); \
arch/mips/include/asm/futex.h
93
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/mips/include/asm/octeon/cvmx-asm.h
117
#define CVMX_CACHE(op, address, offset) \
arch/mips/include/asm/octeon/cvmx-asm.h
118
asm volatile ("cache " CVMX_TMP_STR(op) ", " CVMX_TMP_STR(offset) "(%[rbase])" \
arch/mips/include/asm/octeon/cvmx-mio-defs.h
3510
uint64_t op:4;
arch/mips/include/asm/octeon/cvmx-mio-defs.h
3528
uint64_t op:4;
arch/mips/include/asm/octeon/cvmx-pow.h
1566
tag_req.s.op = CVMX_POW_TAG_OP_SWTAG;
arch/mips/include/asm/octeon/cvmx-pow.h
1677
tag_req.s.op = CVMX_POW_TAG_OP_SWTAG_FULL;
arch/mips/include/asm/octeon/cvmx-pow.h
1756
tag_req.s.op = CVMX_POW_TAG_OP_SWTAG;
arch/mips/include/asm/octeon/cvmx-pow.h
1820
tag_req.s.op = CVMX_POW_TAG_OP_ADDWQ;
arch/mips/include/asm/octeon/cvmx-pow.h
192
cvmx_pow_tag_op_t op:4;
arch/mips/include/asm/octeon/cvmx-pow.h
1977
tag_req.s.op = CVMX_POW_TAG_OP_SWTAG_DESCH;
arch/mips/include/asm/octeon/cvmx-pow.h
2080
tag_req.s.op = CVMX_POW_TAG_OP_DESCH;
arch/mips/include/asm/octeon/cvmx-pow.h
227
cvmx_pow_tag_op_t op:4;
arch/mips/include/asm/octeon/cvmx-sriox-defs.h
984
uint64_t op:1;
arch/mips/include/asm/octeon/cvmx-sriox-defs.h
988
uint64_t op:1;
arch/mips/include/asm/octeon/cvmx.h
456
#define CVMX_WAIT_FOR_FIELD64(address, type, field, op, value, timeout_usec)\
arch/mips/include/asm/octeon/cvmx.h
466
if ((c.s.field) op(value)) { \
arch/mips/include/asm/r4kcache.h
110
#define protected_cache_op(op, addr) \
arch/mips/include/asm/r4kcache.h
129
: "i" (op), "r" (addr), "i" (-EFAULT)); \
arch/mips/include/asm/r4kcache.h
134
#define protected_cache_op(op, addr) \
arch/mips/include/asm/r4kcache.h
152
: "i" (op), "r" (addr), "i" (-EFAULT)); \
arch/mips/include/asm/r4kcache.h
195
#define cache_unroll(times, insn, op, addr, lsize) do { \
arch/mips/include/asm/r4kcache.h
197
unroll(times, _cache_op, insn, op, (addr) + (i++ * (lsize))); \
arch/mips/include/asm/r4kcache.h
47
#define _cache_op(insn, op, addr) \
arch/mips/include/asm/r4kcache.h
55
: "i" (op), "R" (*(unsigned char *)(addr)))
arch/mips/include/asm/r4kcache.h
57
#define cache_op(op, addr) \
arch/mips/include/asm/r4kcache.h
58
_cache_op(kernel_cache, op, addr)
arch/mips/include/asm/uasm.h
24
#define Ip_u1u2u3(op) \
arch/mips/include/asm/uasm.h
25
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c)
arch/mips/include/asm/uasm.h
27
#define Ip_u2u1u3(op) \
arch/mips/include/asm/uasm.h
28
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c)
arch/mips/include/asm/uasm.h
30
#define Ip_u3u2u1(op) \
arch/mips/include/asm/uasm.h
31
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c)
arch/mips/include/asm/uasm.h
33
#define Ip_u3u1u2(op) \
arch/mips/include/asm/uasm.h
34
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c)
arch/mips/include/asm/uasm.h
36
#define Ip_u1u2s3(op) \
arch/mips/include/asm/uasm.h
37
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, signed int c)
arch/mips/include/asm/uasm.h
39
#define Ip_u2s3u1(op) \
arch/mips/include/asm/uasm.h
40
void uasm_i##op(u32 **buf, unsigned int a, signed int b, unsigned int c)
arch/mips/include/asm/uasm.h
42
#define Ip_s3s1s2(op) \
arch/mips/include/asm/uasm.h
43
void uasm_i##op(u32 **buf, int a, int b, int c)
arch/mips/include/asm/uasm.h
45
#define Ip_u2u1s3(op) \
arch/mips/include/asm/uasm.h
46
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, signed int c)
arch/mips/include/asm/uasm.h
48
#define Ip_u2u1msbu3(op) \
arch/mips/include/asm/uasm.h
49
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b, unsigned int c, \
arch/mips/include/asm/uasm.h
52
#define Ip_u1u2(op) \
arch/mips/include/asm/uasm.h
53
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b)
arch/mips/include/asm/uasm.h
55
#define Ip_u2u1(op) \
arch/mips/include/asm/uasm.h
56
void uasm_i##op(u32 **buf, unsigned int a, unsigned int b)
arch/mips/include/asm/uasm.h
58
#define Ip_u1s2(op) \
arch/mips/include/asm/uasm.h
59
void uasm_i##op(u32 **buf, unsigned int a, signed int b)
arch/mips/include/asm/uasm.h
61
#define Ip_u1(op) void uasm_i##op(u32 **buf, unsigned int a)
arch/mips/include/asm/uasm.h
63
#define Ip_0(op) void uasm_i##op(u32 **buf)
arch/mips/include/uapi/asm/inst.h
793
__BITFIELD_FIELD(unsigned int op : 5,
arch/mips/include/uapi/asm/inst.h
804
__BITFIELD_FIELD(unsigned int op : 3,
arch/mips/include/uapi/asm/inst.h
850
__BITFIELD_FIELD(unsigned int op : 2,
arch/mips/include/uapi/asm/inst.h
857
__BITFIELD_FIELD(unsigned int op : 5,
arch/mips/include/uapi/asm/inst.h
870
__BITFIELD_FIELD(unsigned int op : 8,
arch/mips/include/uapi/asm/inst.h
882
__BITFIELD_FIELD(unsigned int op : 3,
arch/mips/include/uapi/asm/inst.h
892
__BITFIELD_FIELD(unsigned int op : 7,
arch/mips/include/uapi/asm/inst.h
913
__BITFIELD_FIELD(unsigned int op : 5,
arch/mips/kernel/pm-cps.c
191
unsigned op, int lbl)
arch/mips/kernel/pm-cps.c
216
uasm_i_cache(pp, op, 0, GPR_T0);
arch/mips/kernel/pm-cps.c
219
uasm_i_cache(pp, op, i * cache->linesz, GPR_T0);
arch/mips/kernel/syscall.c
240
SYSCALL_DEFINE3(cachectl, char *, addr, int, nbytes, int, op)
arch/mips/kernel/traps.c
726
int op = opcode & OPCODE;
arch/mips/kernel/traps.c
730
if (op == LWC2 && op2 == CSR_OPCODE2 && csr_func == CSR_FUNC_CPUCFG) {
arch/mips/kernel/unaligned.c
169
switch (insn.mxu_lx_format.op) {
arch/mips/kernel/unaligned.c
207
switch (insn.dsp_format.op) {
arch/mips/kvm/emulate.c
1275
u32 op, rt;
arch/mips/kvm/emulate.c
1279
op = inst.i_format.opcode;
arch/mips/kvm/emulate.c
1301
switch (op) {
arch/mips/kvm/mmu.c
277
#define BUILD_PTE_RANGE_OP(name, op) \
arch/mips/kvm/mmu.c
292
new = op(old); \
arch/mips/kvm/trace.h
203
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op, unsigned int reg,
arch/mips/kvm/trace.h
205
TP_ARGS(vcpu, op, reg, val),
arch/mips/kvm/trace.h
209
__field(u8, op)
arch/mips/kvm/trace.h
215
__entry->op = op;
arch/mips/kvm/trace.h
219
__print_symbolic(__entry->op,
arch/mips/kvm/trace.h
253
TP_PROTO(struct kvm_vcpu *vcpu, unsigned int op,
arch/mips/kvm/trace.h
255
TP_ARGS(vcpu, op, state),
arch/mips/kvm/trace.h
258
__field(u8, op)
arch/mips/kvm/trace.h
264
__entry->op = op;
arch/mips/kvm/trace.h
269
__print_symbolic(__entry->op,
arch/mips/kvm/vz.c
1079
u32 cache, op_inst, op, base;
arch/mips/kvm/vz.c
1100
op = op_inst & CacheOp_Op;
arch/mips/kvm/vz.c
1105
cache, op, base, arch->gprs[base], offset);
arch/mips/kvm/vz.c
1134
curr_pc, vcpu->arch.gprs[31], cache, op, base, arch->gprs[base],
arch/mips/kvm/vz.c
840
static unsigned long mips_process_maar(unsigned int op, unsigned long val)
arch/mips/kvm/vz.c
851
if (op == mtc_op) {
arch/mips/kvm/vz.c
854
} else if (op == dmtc_op) {
arch/mips/math-emu/cp1emu.c
130
op = insn.mm_fp0_format.func;
arch/mips/math-emu/cp1emu.c
131
if (op == mm_32f_01_op)
arch/mips/math-emu/cp1emu.c
133
else if (op == mm_32f_11_op)
arch/mips/math-emu/cp1emu.c
135
else if (op == mm_32f_02_op)
arch/mips/math-emu/cp1emu.c
137
else if (op == mm_32f_12_op)
arch/mips/math-emu/cp1emu.c
139
else if (op == mm_32f_41_op)
arch/mips/math-emu/cp1emu.c
141
else if (op == mm_32f_51_op)
arch/mips/math-emu/cp1emu.c
143
else if (op == mm_32f_42_op)
arch/mips/math-emu/cp1emu.c
156
op = insn.mm_fp5_format.op & 0x7;
arch/mips/math-emu/cp1emu.c
157
if (op == mm_ldxc1_op)
arch/mips/math-emu/cp1emu.c
159
else if (op == mm_sdxc1_op)
arch/mips/math-emu/cp1emu.c
161
else if (op == mm_lwxc1_op)
arch/mips/math-emu/cp1emu.c
163
else if (op == mm_swxc1_op)
arch/mips/math-emu/cp1emu.c
179
op = -1; /* Invalid */
arch/mips/math-emu/cp1emu.c
180
if (insn.mm_fp2_format.op == mm_fmovt_op)
arch/mips/math-emu/cp1emu.c
181
op = 1;
arch/mips/math-emu/cp1emu.c
182
else if (insn.mm_fp2_format.op == mm_fmovf_op)
arch/mips/math-emu/cp1emu.c
183
op = 0;
arch/mips/math-emu/cp1emu.c
184
if (op != -1) {
arch/mips/math-emu/cp1emu.c
189
(insn.mm_fp2_format.cc<<2) + op;
arch/mips/math-emu/cp1emu.c
200
if (insn.mm_fp0_format.op == mm_fadd_op)
arch/mips/math-emu/cp1emu.c
202
else if (insn.mm_fp0_format.op == mm_fsub_op)
arch/mips/math-emu/cp1emu.c
204
else if (insn.mm_fp0_format.op == mm_fmul_op)
arch/mips/math-emu/cp1emu.c
206
else if (insn.mm_fp0_format.op == mm_fdiv_op)
arch/mips/math-emu/cp1emu.c
224
if (insn.mm_fp0_format.op == mm_fmovn_op)
arch/mips/math-emu/cp1emu.c
226
else if (insn.mm_fp0_format.op == mm_fmovz_op)
arch/mips/math-emu/cp1emu.c
243
switch (insn.mm_fp1_format.op) {
arch/mips/math-emu/cp1emu.c
248
if ((insn.mm_fp1_format.op & 0x7f) ==
arch/mips/math-emu/cp1emu.c
250
op = 0;
arch/mips/math-emu/cp1emu.c
252
op = 1;
arch/mips/math-emu/cp1emu.c
256
(insn.mm_fp4_format.cc << 2) + op;
arch/mips/math-emu/cp1emu.c
265
if ((insn.mm_fp1_format.op & 0x7f) ==
arch/mips/math-emu/cp1emu.c
288
if ((insn.mm_fp1_format.op & 0x7f) ==
arch/mips/math-emu/cp1emu.c
291
else if ((insn.mm_fp1_format.op & 0x7f) ==
arch/mips/math-emu/cp1emu.c
316
if (insn.mm_fp1_format.op == mm_ffloorl_op)
arch/mips/math-emu/cp1emu.c
318
else if (insn.mm_fp1_format.op == mm_ffloorw_op)
arch/mips/math-emu/cp1emu.c
320
else if (insn.mm_fp1_format.op == mm_fceill_op)
arch/mips/math-emu/cp1emu.c
322
else if (insn.mm_fp1_format.op == mm_fceilw_op)
arch/mips/math-emu/cp1emu.c
324
else if (insn.mm_fp1_format.op == mm_ftruncl_op)
arch/mips/math-emu/cp1emu.c
326
else if (insn.mm_fp1_format.op == mm_ftruncw_op)
arch/mips/math-emu/cp1emu.c
328
else if (insn.mm_fp1_format.op == mm_froundl_op)
arch/mips/math-emu/cp1emu.c
330
else if (insn.mm_fp1_format.op == mm_froundw_op)
arch/mips/math-emu/cp1emu.c
332
else if (insn.mm_fp1_format.op == mm_fcvtl_op)
arch/mips/math-emu/cp1emu.c
349
if (insn.mm_fp1_format.op == mm_frsqrt_op)
arch/mips/math-emu/cp1emu.c
351
else if (insn.mm_fp1_format.op == mm_fsqrt_op)
arch/mips/math-emu/cp1emu.c
371
if (insn.mm_fp1_format.op == mm_mfc1_op)
arch/mips/math-emu/cp1emu.c
372
op = mfc_op;
arch/mips/math-emu/cp1emu.c
373
else if (insn.mm_fp1_format.op == mm_mtc1_op)
arch/mips/math-emu/cp1emu.c
374
op = mtc_op;
arch/mips/math-emu/cp1emu.c
375
else if (insn.mm_fp1_format.op == mm_cfc1_op)
arch/mips/math-emu/cp1emu.c
376
op = cfc_op;
arch/mips/math-emu/cp1emu.c
377
else if (insn.mm_fp1_format.op == mm_ctc1_op)
arch/mips/math-emu/cp1emu.c
378
op = ctc_op;
arch/mips/math-emu/cp1emu.c
379
else if (insn.mm_fp1_format.op == mm_mfhc1_op)
arch/mips/math-emu/cp1emu.c
380
op = mfhc_op;
arch/mips/math-emu/cp1emu.c
382
op = mthc_op;
arch/mips/math-emu/cp1emu.c
384
mips32_insn.fp1_format.op = op;
arch/mips/math-emu/cp1emu.c
86
int func, fmt, op;
arch/mips/mm/uasm-micromips.c
160
u32 op;
arch/mips/mm/uasm-micromips.c
169
op = ip->match;
arch/mips/mm/uasm-micromips.c
174
op |= build_rt(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
176
op |= build_rs(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
181
op |= build_rs(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
183
op |= build_rt(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
186
op |= build_rd(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
188
op |= build_re(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
190
op |= build_simm(va_arg(ap, s32));
arch/mips/mm/uasm-micromips.c
192
op |= build_uimm(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
194
op |= build_bimm(va_arg(ap, s32));
arch/mips/mm/uasm-micromips.c
196
op |= build_jimm(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
198
op |= build_func(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
200
op |= build_set(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
202
op |= build_scimm(va_arg(ap, u32));
arch/mips/mm/uasm-micromips.c
206
**buf = ((op & 0xffff) << 16) | (op >> 16);
arch/mips/mm/uasm-micromips.c
208
**buf = op;
arch/mips/mm/uasm-mips.c
236
u32 op;
arch/mips/mm/uasm-mips.c
245
op = ip->match;
arch/mips/mm/uasm-mips.c
248
op |= build_rs(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
250
op |= build_rt(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
252
op |= build_rd(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
254
op |= build_re(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
256
op |= build_simm(va_arg(ap, s32));
arch/mips/mm/uasm-mips.c
258
op |= build_uimm(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
260
op |= build_bimm(va_arg(ap, s32));
arch/mips/mm/uasm-mips.c
262
op |= build_jimm(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
264
op |= build_func(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
266
op |= build_set(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
268
op |= build_scimm(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
270
op |= build_scimm9(va_arg(ap, u32));
arch/mips/mm/uasm-mips.c
273
**buf = op;
arch/mips/mm/uasm.c
152
#define I_u1u2u3(op) \
arch/mips/mm/uasm.c
153
Ip_u1u2u3(op) \
arch/mips/mm/uasm.c
155
build_insn(buf, insn##op, a, b, c); \
arch/mips/mm/uasm.c
157
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
159
#define I_s3s1s2(op) \
arch/mips/mm/uasm.c
160
Ip_s3s1s2(op) \
arch/mips/mm/uasm.c
162
build_insn(buf, insn##op, b, c, a); \
arch/mips/mm/uasm.c
164
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
166
#define I_u2u1u3(op) \
arch/mips/mm/uasm.c
167
Ip_u2u1u3(op) \
arch/mips/mm/uasm.c
169
build_insn(buf, insn##op, b, a, c); \
arch/mips/mm/uasm.c
171
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
173
#define I_u3u2u1(op) \
arch/mips/mm/uasm.c
174
Ip_u3u2u1(op) \
arch/mips/mm/uasm.c
176
build_insn(buf, insn##op, c, b, a); \
arch/mips/mm/uasm.c
178
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
180
#define I_u3u1u2(op) \
arch/mips/mm/uasm.c
181
Ip_u3u1u2(op) \
arch/mips/mm/uasm.c
183
build_insn(buf, insn##op, b, c, a); \
arch/mips/mm/uasm.c
185
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
187
#define I_u1u2s3(op) \
arch/mips/mm/uasm.c
188
Ip_u1u2s3(op) \
arch/mips/mm/uasm.c
190
build_insn(buf, insn##op, a, b, c); \
arch/mips/mm/uasm.c
192
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
194
#define I_u2s3u1(op) \
arch/mips/mm/uasm.c
195
Ip_u2s3u1(op) \
arch/mips/mm/uasm.c
197
build_insn(buf, insn##op, c, a, b); \
arch/mips/mm/uasm.c
199
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
201
#define I_u2u1s3(op) \
arch/mips/mm/uasm.c
202
Ip_u2u1s3(op) \
arch/mips/mm/uasm.c
204
build_insn(buf, insn##op, b, a, c); \
arch/mips/mm/uasm.c
206
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
208
#define I_u2u1msbu3(op) \
arch/mips/mm/uasm.c
209
Ip_u2u1msbu3(op) \
arch/mips/mm/uasm.c
211
build_insn(buf, insn##op, b, a, c+d-1, c); \
arch/mips/mm/uasm.c
213
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
215
#define I_u2u1msb32u3(op) \
arch/mips/mm/uasm.c
216
Ip_u2u1msbu3(op) \
arch/mips/mm/uasm.c
218
build_insn(buf, insn##op, b, a, c+d-33, c); \
arch/mips/mm/uasm.c
220
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
222
#define I_u2u1msb32msb3(op) \
arch/mips/mm/uasm.c
223
Ip_u2u1msbu3(op) \
arch/mips/mm/uasm.c
225
build_insn(buf, insn##op, b, a, c+d-33, c-32); \
arch/mips/mm/uasm.c
227
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
229
#define I_u2u1msbdu3(op) \
arch/mips/mm/uasm.c
230
Ip_u2u1msbu3(op) \
arch/mips/mm/uasm.c
232
build_insn(buf, insn##op, b, a, d-1, c); \
arch/mips/mm/uasm.c
234
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
236
#define I_u1u2(op) \
arch/mips/mm/uasm.c
237
Ip_u1u2(op) \
arch/mips/mm/uasm.c
239
build_insn(buf, insn##op, a, b); \
arch/mips/mm/uasm.c
241
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
243
#define I_u2u1(op) \
arch/mips/mm/uasm.c
244
Ip_u1u2(op) \
arch/mips/mm/uasm.c
246
build_insn(buf, insn##op, b, a); \
arch/mips/mm/uasm.c
248
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
250
#define I_u1s2(op) \
arch/mips/mm/uasm.c
251
Ip_u1s2(op) \
arch/mips/mm/uasm.c
253
build_insn(buf, insn##op, a, b); \
arch/mips/mm/uasm.c
255
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
257
#define I_u1(op) \
arch/mips/mm/uasm.c
258
Ip_u1(op) \
arch/mips/mm/uasm.c
260
build_insn(buf, insn##op, a); \
arch/mips/mm/uasm.c
262
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/mm/uasm.c
264
#define I_0(op) \
arch/mips/mm/uasm.c
265
Ip_0(op) \
arch/mips/mm/uasm.c
267
build_insn(buf, insn##op); \
arch/mips/mm/uasm.c
269
UASM_EXPORT_SYMBOL(uasm_i##op);
arch/mips/net/bpf_jit_comp.c
211
bool valid_alu_i(u8 op, s32 imm)
arch/mips/net/bpf_jit_comp.c
213
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp.c
247
bool rewrite_alu_i(u8 op, s32 imm, u8 *alu, s32 *val)
arch/mips/net/bpf_jit_comp.c
251
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp.c
268
op = BPF_AND;
arch/mips/net/bpf_jit_comp.c
271
op = BPF_LSH;
arch/mips/net/bpf_jit_comp.c
281
op = BPF_RSH;
arch/mips/net/bpf_jit_comp.c
287
op = BPF_AND;
arch/mips/net/bpf_jit_comp.c
292
*alu = op;
arch/mips/net/bpf_jit_comp.c
298
void emit_alu_i(struct jit_context *ctx, u8 dst, s32 imm, u8 op)
arch/mips/net/bpf_jit_comp.c
300
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp.c
342
void emit_alu_r(struct jit_context *ctx, u8 dst, u8 src, u8 op)
arch/mips/net/bpf_jit_comp.c
344
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp.c
502
bool valid_jmp_i(u8 op, s32 imm)
arch/mips/net/bpf_jit_comp.c
504
switch (op) {
arch/mips/net/bpf_jit_comp.c
533
static u8 invert_jmp(u8 op)
arch/mips/net/bpf_jit_comp.c
535
switch (op) {
arch/mips/net/bpf_jit_comp.c
557
int op = bpf_op;
arch/mips/net/bpf_jit_comp.c
594
op = invert_jmp(bpf_op);
arch/mips/net/bpf_jit_comp.c
601
*jit_op = op;
arch/mips/net/bpf_jit_comp.c
698
void emit_jmp_i(struct jit_context *ctx, u8 dst, s32 imm, s32 off, u8 op)
arch/mips/net/bpf_jit_comp.c
700
switch (op) {
arch/mips/net/bpf_jit_comp.c
758
void emit_jmp_r(struct jit_context *ctx, u8 dst, u8 src, s32 off, u8 op)
arch/mips/net/bpf_jit_comp.c
760
switch (op) {
arch/mips/net/bpf_jit_comp.h
166
bool valid_alu_i(u8 op, s32 imm);
arch/mips/net/bpf_jit_comp.h
169
bool rewrite_alu_i(u8 op, s32 imm, u8 *alu, s32 *val);
arch/mips/net/bpf_jit_comp.h
172
void emit_alu_i(struct jit_context *ctx, u8 dst, s32 imm, u8 op);
arch/mips/net/bpf_jit_comp.h
175
void emit_alu_r(struct jit_context *ctx, u8 dst, u8 src, u8 op);
arch/mips/net/bpf_jit_comp.h
187
bool valid_jmp_i(u8 op, s32 imm);
arch/mips/net/bpf_jit_comp.h
201
void emit_jmp_i(struct jit_context *ctx, u8 dst, s32 imm, s32 off, u8 op);
arch/mips/net/bpf_jit_comp.h
204
void emit_jmp_r(struct jit_context *ctx, u8 dst, u8 src, s32 off, u8 op);
arch/mips/net/bpf_jit_comp32.c
1093
const u8 dst[], s32 imm, s32 off, u8 op)
arch/mips/net/bpf_jit_comp32.c
1097
switch (op) {
arch/mips/net/bpf_jit_comp32.c
1119
if (op == BPF_JEQ)
arch/mips/net/bpf_jit_comp32.c
1136
if (op == BPF_JSET)
arch/mips/net/bpf_jit_comp32.c
1186
const u8 dst[], const u8 src[], s32 off, u8 op)
arch/mips/net/bpf_jit_comp32.c
1191
switch (op) {
arch/mips/net/bpf_jit_comp32.c
1202
if (op == BPF_JEQ)
arch/mips/net/bpf_jit_comp32.c
1214
if (op == BPF_JSET)
arch/mips/net/bpf_jit_comp32.c
202
const u8 dst[], s32 imm, u8 op)
arch/mips/net/bpf_jit_comp32.c
211
switch (op) {
arch/mips/net/bpf_jit_comp32.c
213
op = BPF_SUB;
arch/mips/net/bpf_jit_comp32.c
217
op = BPF_ADD;
arch/mips/net/bpf_jit_comp32.c
225
switch (op) {
arch/mips/net/bpf_jit_comp32.c
268
const u8 dst[], const u8 src[], u8 op)
arch/mips/net/bpf_jit_comp32.c
270
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp32.c
322
const u8 dst[], u32 imm, u8 op)
arch/mips/net/bpf_jit_comp32.c
324
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp32.c
367
const u8 dst[], u8 src, u8 op)
arch/mips/net/bpf_jit_comp32.c
376
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp32.c
526
const u8 dst[], const u8 src[], u8 op)
arch/mips/net/bpf_jit_comp32.c
546
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp64.c
158
static void emit_alu_i64(struct jit_context *ctx, u8 dst, s32 imm, u8 op)
arch/mips/net/bpf_jit_comp64.c
160
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp64.c
195
emit_alu_i(ctx, dst, imm, op);
arch/mips/net/bpf_jit_comp64.c
201
static void emit_alu_r64(struct jit_context *ctx, u8 dst, u8 src, u8 op)
arch/mips/net/bpf_jit_comp64.c
203
switch (BPF_OP(op)) {
arch/mips/net/bpf_jit_comp64.c
256
emit_alu_r(ctx, dst, src, op);
arch/nios2/include/asm/syscalls.h
9
unsigned int op);
arch/nios2/kernel/sys_nios2.c
22
unsigned int op)
arch/nios2/kernel/sys_nios2.c
31
if (op)
arch/openrisc/include/asm/atomic.h
15
#define ATOMIC_OP(op) \
arch/openrisc/include/asm/atomic.h
16
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/openrisc/include/asm/atomic.h
22
" l." #op " %0,%0,%2 \n" \
arch/openrisc/include/asm/atomic.h
32
#define ATOMIC_OP_RETURN(op) \
arch/openrisc/include/asm/atomic.h
33
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/openrisc/include/asm/atomic.h
39
" l." #op " %0,%0,%2 \n" \
arch/openrisc/include/asm/atomic.h
51
#define ATOMIC_FETCH_OP(op) \
arch/openrisc/include/asm/atomic.h
52
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/openrisc/include/asm/atomic.h
58
" l." #op " %1,%0,%3 \n" \
arch/openrisc/include/asm/futex.h
34
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/openrisc/include/asm/futex.h
41
switch (op) {
arch/openrisc/include/asm/uaccess.h
161
#define __get_user_asm(x, addr, err, op) \
arch/openrisc/include/asm/uaccess.h
165
"1: "op" %1,0(%2)\n" \
arch/openrisc/include/asm/uaccess.h
95
#define __put_user_asm(x, addr, err, op) \
arch/openrisc/include/asm/uaccess.h
97
"1: "op" 0(%2),%1\n" \
arch/openrisc/kernel/traps.c
250
unsigned int op, insn;
arch/openrisc/kernel/traps.c
253
op = insn >> 26;
arch/openrisc/kernel/traps.c
254
switch (op) {
arch/openrisc/kernel/traps.c
273
unsigned int rb, op, jmp;
arch/openrisc/kernel/traps.c
281
op = jmp >> 26;
arch/openrisc/kernel/traps.c
283
switch (op) {
arch/openrisc/kernel/traps.c
413
unsigned int op;
arch/openrisc/kernel/traps.c
416
op = insn >> 26;
arch/openrisc/kernel/traps.c
418
switch (op) {
arch/parisc/include/asm/atomic.h
100
static __inline__ int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/parisc/include/asm/atomic.h
113
#define ATOMIC_OPS(op, c_op) \
arch/parisc/include/asm/atomic.h
114
ATOMIC_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
115
ATOMIC_OP_RETURN(op, c_op) \
arch/parisc/include/asm/atomic.h
116
ATOMIC_FETCH_OP(op, c_op)
arch/parisc/include/asm/atomic.h
127
#define ATOMIC_OPS(op, c_op) \
arch/parisc/include/asm/atomic.h
128
ATOMIC_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
129
ATOMIC_FETCH_OP(op, c_op)
arch/parisc/include/asm/atomic.h
148
#define ATOMIC64_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
149
static __inline__ void arch_atomic64_##op(s64 i, atomic64_t *v) \
arch/parisc/include/asm/atomic.h
158
#define ATOMIC64_OP_RETURN(op, c_op) \
arch/parisc/include/asm/atomic.h
159
static __inline__ s64 arch_atomic64_##op##_return(s64 i, atomic64_t *v) \
arch/parisc/include/asm/atomic.h
171
#define ATOMIC64_FETCH_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
172
static __inline__ s64 arch_atomic64_fetch_##op(s64 i, atomic64_t *v) \
arch/parisc/include/asm/atomic.h
185
#define ATOMIC64_OPS(op, c_op) \
arch/parisc/include/asm/atomic.h
186
ATOMIC64_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
187
ATOMIC64_OP_RETURN(op, c_op) \
arch/parisc/include/asm/atomic.h
188
ATOMIC64_FETCH_OP(op, c_op)
arch/parisc/include/asm/atomic.h
199
#define ATOMIC64_OPS(op, c_op) \
arch/parisc/include/asm/atomic.h
200
ATOMIC64_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
201
ATOMIC64_FETCH_OP(op, c_op)
arch/parisc/include/asm/atomic.h
76
#define ATOMIC_OP(op, c_op) \
arch/parisc/include/asm/atomic.h
77
static __inline__ void arch_atomic_##op(int i, atomic_t *v) \
arch/parisc/include/asm/atomic.h
86
#define ATOMIC_OP_RETURN(op, c_op) \
arch/parisc/include/asm/atomic.h
87
static __inline__ int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/parisc/include/asm/atomic.h
99
#define ATOMIC_FETCH_OP(op, c_op) \
arch/parisc/include/asm/futex.h
35
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/parisc/include/asm/futex.h
56
switch (op) {
arch/parisc/kernel/smp.c
197
ipi_send(int cpu, enum ipi_message_type op)
arch/parisc/kernel/smp.c
204
p->pending_ipi |= 1 << op;
arch/parisc/kernel/smp.c
210
send_IPI_mask(const struct cpumask *mask, enum ipi_message_type op)
arch/parisc/kernel/smp.c
215
ipi_send(cpu, op);
arch/parisc/kernel/smp.c
219
send_IPI_single(int dest_cpu, enum ipi_message_type op)
arch/parisc/kernel/smp.c
223
ipi_send(dest_cpu, op);
arch/parisc/kernel/smp.c
227
send_IPI_allbutself(enum ipi_message_type op)
arch/parisc/kernel/smp.c
234
send_IPI_single(i, op);
arch/parisc/math-emu/dbl_float.h
305
#define Dbl_setwrapped_exponent(dbl_value,exponent,op) \
arch/parisc/math-emu/dbl_float.h
306
Deposit_dexponent(dbl_value,(exponent op DBL_WRAP))
arch/parisc/math-emu/fpudispatch.c
103
#define get_major(op) extru(op,fpmajorpos,6)
arch/parisc/math-emu/fpudispatch.c
108
#define get_class(op) extru(op,fpclasspos,2)
arch/parisc/math-emu/fpudispatch.c
113
#define get_subop(op) extru(op,fpsubpos,3)
arch/parisc/math-emu/fpudispatch.c
118
#define get_subop1_PA1_1(op) extru(op,fpclass1subpos,2) /* PA89 (1.1) fmt */
arch/parisc/math-emu/fpudispatch.c
119
#define get_subop1_PA2_0(op) extru(op,fpclass1subpos,3) /* PA 2.0 fmt */
arch/parisc/math-emu/sgl_float.h
182
#define Sgl_setwrapped_exponent(sgl_value,exponent,op) \
arch/parisc/math-emu/sgl_float.h
183
Deposit_sexponent(sgl_value,(exponent op SGL_WRAP))
arch/parisc/net/bpf_jit_comp32.c
422
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
427
if (0 && op == BPF_MOV)
arch/parisc/net/bpf_jit_comp32.c
433
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
510
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
515
if (op == BPF_MOV)
arch/parisc/net/bpf_jit_comp32.c
521
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
565
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
572
if (op == BPF_MOV)
arch/parisc/net/bpf_jit_comp32.c
578
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
604
emit_call_libgcc_ll(__muldi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
607
emit_call_libgcc_ll(&hppa_div64, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
610
emit_call_libgcc_ll(&hppa_div64_rem, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
613
emit_call_libgcc_ll(__ashldi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
616
emit_call_libgcc_ll(__lshrdi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
619
emit_call_libgcc_ll(__ashrdi3, rd, rs, op, ctx);
arch/parisc/net/bpf_jit_comp32.c
633
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
640
if (op == BPF_MOV)
arch/parisc/net/bpf_jit_comp32.c
646
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
666
emit_call_millicode($$mulI, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
669
emit_call_millicode($$divU, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
672
emit_call_millicode($$remU, lo(rd), lo(rs), op, ctx);
arch/parisc/net/bpf_jit_comp32.c
700
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
720
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
789
static int emit_bcc(u8 op, u8 rd, u8 rs, int paoff, struct hppa_jit_context *ctx)
arch/parisc/net/bpf_jit_comp32.c
795
if (op == BPF_JSET) {
arch/parisc/net/bpf_jit_comp32.c
804
op = BPF_JNE;
arch/parisc/net/bpf_jit_comp32.c
810
op = invert_bpf_cond(op);
arch/parisc/net/bpf_jit_comp32.c
821
switch (op) {
arch/parisc/net/bpf_jit_comp32.c
867
struct hppa_jit_context *ctx, const u8 op)
arch/parisc/net/bpf_jit_comp32.c
880
if (emit_bcc(op, lo(rs1), lo(rs2), paoff, ctx))
arch/parisc/net/bpf_jit_comp64.c
248
static int emit_branch(u8 op, u8 rd, u8 rs, signed long paoff,
arch/parisc/net/bpf_jit_comp64.c
255
if (op == BPF_JSET) {
arch/parisc/net/bpf_jit_comp64.c
264
op = BPF_JNE;
arch/parisc/net/bpf_jit_comp64.c
271
op = invert_bpf_cond(op);
arch/parisc/net/bpf_jit_comp64.c
282
switch (op) {
arch/powerpc/boot/cpm-serial.c
104
out_be32(cpcr, op | cpm_cmd | 0x10000);
arch/powerpc/boot/cpm-serial.c
80
static void (*do_cmd)(int op);
arch/powerpc/boot/cpm-serial.c
88
static void cpm1_cmd(int op)
arch/powerpc/boot/cpm-serial.c
93
out_be16(cpcr, (op << 8) | cpm_cmd | 1);
arch/powerpc/boot/cpm-serial.c
99
static void cpm2_cmd(int op)
arch/powerpc/include/asm/atomic.h
100
ATOMIC_OP_RETURN_RELAXED(op, asm_op, suffix, sign, ##__VA_ARGS__)\
arch/powerpc/include/asm/atomic.h
101
ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign, ##__VA_ARGS__)
arch/powerpc/include/asm/atomic.h
113
#define ATOMIC_OPS(op, asm_op, suffix, sign) \
arch/powerpc/include/asm/atomic.h
114
ATOMIC_OP(op, asm_op, suffix, sign) \
arch/powerpc/include/asm/atomic.h
115
ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign)
arch/powerpc/include/asm/atomic.h
215
#define ATOMIC64_OP(op, asm_op) \
arch/powerpc/include/asm/atomic.h
216
static __inline__ void arch_atomic64_##op(s64 a, atomic64_t *v) \
arch/powerpc/include/asm/atomic.h
221
"1: ldarx %0,0,%3 # atomic64_" #op "\n" \
arch/powerpc/include/asm/atomic.h
230
#define ATOMIC64_OP_RETURN_RELAXED(op, asm_op) \
arch/powerpc/include/asm/atomic.h
232
arch_atomic64_##op##_return_relaxed(s64 a, atomic64_t *v) \
arch/powerpc/include/asm/atomic.h
237
"1: ldarx %0,0,%3 # atomic64_" #op "_return_relaxed\n" \
arch/powerpc/include/asm/atomic.h
248
#define ATOMIC64_FETCH_OP_RELAXED(op, asm_op) \
arch/powerpc/include/asm/atomic.h
250
arch_atomic64_fetch_##op##_relaxed(s64 a, atomic64_t *v) \
arch/powerpc/include/asm/atomic.h
255
"1: ldarx %0,0,%4 # atomic64_fetch_" #op "_relaxed\n" \
arch/powerpc/include/asm/atomic.h
266
#define ATOMIC64_OPS(op, asm_op) \
arch/powerpc/include/asm/atomic.h
267
ATOMIC64_OP(op, asm_op) \
arch/powerpc/include/asm/atomic.h
268
ATOMIC64_OP_RETURN_RELAXED(op, asm_op) \
arch/powerpc/include/asm/atomic.h
269
ATOMIC64_FETCH_OP_RELAXED(op, asm_op)
arch/powerpc/include/asm/atomic.h
281
#define ATOMIC64_OPS(op, asm_op) \
arch/powerpc/include/asm/atomic.h
282
ATOMIC64_OP(op, asm_op) \
arch/powerpc/include/asm/atomic.h
283
ATOMIC64_FETCH_OP_RELAXED(op, asm_op)
arch/powerpc/include/asm/atomic.h
49
#define ATOMIC_OP(op, asm_op, suffix, sign, ...) \
arch/powerpc/include/asm/atomic.h
50
static __inline__ void arch_atomic_##op(int a, atomic_t *v) \
arch/powerpc/include/asm/atomic.h
55
"1: lwarx %0,0,%3 # atomic_" #op "\n" \
arch/powerpc/include/asm/atomic.h
64
#define ATOMIC_OP_RETURN_RELAXED(op, asm_op, suffix, sign, ...) \
arch/powerpc/include/asm/atomic.h
65
static inline int arch_atomic_##op##_return_relaxed(int a, atomic_t *v) \
arch/powerpc/include/asm/atomic.h
70
"1: lwarx %0,0,%3 # atomic_" #op "_return_relaxed\n" \
arch/powerpc/include/asm/atomic.h
81
#define ATOMIC_FETCH_OP_RELAXED(op, asm_op, suffix, sign, ...) \
arch/powerpc/include/asm/atomic.h
82
static inline int arch_atomic_fetch_##op##_relaxed(int a, atomic_t *v) \
arch/powerpc/include/asm/atomic.h
87
"1: lwarx %0,0,%4 # atomic_fetch_" #op "_relaxed\n" \
arch/powerpc/include/asm/atomic.h
98
#define ATOMIC_OPS(op, asm_op, suffix, sign, ...) \
arch/powerpc/include/asm/atomic.h
99
ATOMIC_OP(op, asm_op, suffix, sign, ##__VA_ARGS__) \
arch/powerpc/include/asm/bitops.h
151
#define DEFINE_TESTOP(fn, op, prefix, postfix, eh) \
arch/powerpc/include/asm/bitops.h
161
#op "%I2 %1,%0,%2\n" \
arch/powerpc/include/asm/bitops.h
65
#define DEFINE_BITOP(fn, op, prefix) \
arch/powerpc/include/asm/bitops.h
74
#op "%I2 %0,%0,%2\n" \
arch/powerpc/include/asm/ftrace.h
85
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/powerpc/include/asm/futex.h
31
static inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/powerpc/include/asm/futex.h
39
switch (op) {
arch/powerpc/include/asm/io.h
373
#define __do_in_asm(name, op) \
arch/powerpc/include/asm/io.h
379
"0:" op " %0,0,%1\n" \
arch/powerpc/include/asm/io.h
398
#define __do_out_asm(name, op) \
arch/powerpc/include/asm/io.h
403
"0:" op " %0,0,%1\n" \
arch/powerpc/include/asm/local.h
30
#define LOCAL_OP(op, c_op) \
arch/powerpc/include/asm/local.h
31
static __inline__ void local_##op(long i, local_t *l) \
arch/powerpc/include/asm/local.h
40
#define LOCAL_OP_RETURN(op, c_op) \
arch/powerpc/include/asm/local.h
41
static __inline__ long local_##op##_return(long a, local_t *l) \
arch/powerpc/include/asm/local.h
53
#define LOCAL_OPS(op, c_op) \
arch/powerpc/include/asm/local.h
54
LOCAL_OP(op, c_op) \
arch/powerpc/include/asm/local.h
55
LOCAL_OP_RETURN(op, c_op)
arch/powerpc/include/asm/opal.h
176
int64_t opal_manage_flash(uint8_t op);
arch/powerpc/include/asm/opal.h
302
s64 opal_mpipl_update(enum opal_mpipl_ops op, u64 src, u64 dest, u64 size);
arch/powerpc/include/asm/ppc_asm.h
28
.macro OP_REGS op, width, start, end, base, offset
arch/powerpc/include/asm/sstep.h
147
extern int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
arch/powerpc/include/asm/sstep.h
154
void emulate_update_regs(struct pt_regs *reg, struct instruction_op *op);
arch/powerpc/include/asm/sstep.h
174
extern int emulate_loadstore(struct pt_regs *regs, struct instruction_op *op);
arch/powerpc/include/asm/uaccess.h
164
#define __get_user_asm_goto(x, addr, label, op) \
arch/powerpc/include/asm/uaccess.h
166
"1: "op" %0,0(%1) # get_user\n" \
arch/powerpc/include/asm/uaccess.h
173
#define __get_user_asm_goto(x, addr, label, op) \
arch/powerpc/include/asm/uaccess.h
175
"1: "op"%U1%X1 %0, %1 # get_user\n" \
arch/powerpc/include/asm/uaccess.h
236
#define __get_user_asm(x, addr, err, op) \
arch/powerpc/include/asm/uaccess.h
238
"1: "op"%U2%X2 %1, %2 # get_user\n" \
arch/powerpc/include/asm/uaccess.h
82
#define __put_user_asm_goto(x, addr, label, op) \
arch/powerpc/include/asm/uaccess.h
84
"1: " op " %0,0(%1) # put_user\n" \
arch/powerpc/include/asm/uaccess.h
91
#define __put_user_asm_goto(x, addr, label, op) \
arch/powerpc/include/asm/uaccess.h
93
"1: " op "%U1%X1 %0,%1 # put_user\n" \
arch/powerpc/include/asm/vio.h
140
extern int vio_h_cop_sync(struct vio_dev *vdev, struct vio_pfo_op *op);
arch/powerpc/kernel/align.c
304
struct instruction_op op;
arch/powerpc/kernel/align.c
343
r = analyse_instr(&op, regs, instr);
arch/powerpc/kernel/align.c
347
type = GETTYPE(op.type);
arch/powerpc/kernel/align.c
349
if (op.type != CACHEOP + DCBZ)
arch/powerpc/kernel/align.c
353
r = emulate_dcbz(op.ea, regs);
arch/powerpc/kernel/align.c
358
r = emulate_loadstore(regs, &op);
arch/powerpc/kernel/hw_breakpoint_constraints.c
133
struct instruction_op op;
arch/powerpc/kernel/hw_breakpoint_constraints.c
143
analyse_instr(&op, regs, *instr);
arch/powerpc/kernel/hw_breakpoint_constraints.c
144
*type = GETTYPE(op.type);
arch/powerpc/kernel/hw_breakpoint_constraints.c
145
*ea = op.ea;
arch/powerpc/kernel/hw_breakpoint_constraints.c
151
*size = GETSIZE(op.type);
arch/powerpc/kernel/mce_power.c
460
struct instruction_op op;
arch/powerpc/kernel/mce_power.c
467
if (!analyse_instr(&op, &tmp, instr)) {
arch/powerpc/kernel/mce_power.c
468
pfn = addr_to_pfn(regs, op.ea);
arch/powerpc/kernel/mce_power.c
469
*addr = op.ea;
arch/powerpc/kernel/optprobes.c
100
if (kprobe_disabled(&op->kp))
arch/powerpc/kernel/optprobes.c
106
kprobes_inc_nmissed_count(&op->kp);
arch/powerpc/kernel/optprobes.c
108
__this_cpu_write(current_kprobe, &op->kp);
arch/powerpc/kernel/optprobes.c
109
regs_set_return_ip(regs, (unsigned long)op->kp.addr);
arch/powerpc/kernel/optprobes.c
111
opt_pre_handler(&op->kp, regs);
arch/powerpc/kernel/optprobes.c
119
void arch_remove_optimized_kprobe(struct optimized_kprobe *op)
arch/powerpc/kernel/optprobes.c
121
if (op->optinsn.insn) {
arch/powerpc/kernel/optprobes.c
122
free_optinsn_slot(op->optinsn.insn, 1);
arch/powerpc/kernel/optprobes.c
123
op->optinsn.insn = NULL;
arch/powerpc/kernel/optprobes.c
154
int arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *p)
arch/powerpc/kernel/optprobes.c
204
patch_imm_load_insns((unsigned long)op, 3, buff + TMPL_OP_IDX);
arch/powerpc/kernel/optprobes.c
241
op->optinsn.insn = buff;
arch/powerpc/kernel/optprobes.c
261
int arch_check_optimized_kprobe(struct optimized_kprobe *op)
arch/powerpc/kernel/optprobes.c
269
struct optimized_kprobe *op;
arch/powerpc/kernel/optprobes.c
272
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/powerpc/kernel/optprobes.c
277
memcpy(op->optinsn.copied_insn, op->kp.addr, RELATIVEJUMP_SIZE);
arch/powerpc/kernel/optprobes.c
278
create_branch(&instr, op->kp.addr, (unsigned long)op->optinsn.insn, 0);
arch/powerpc/kernel/optprobes.c
279
patch_instruction(op->kp.addr, instr);
arch/powerpc/kernel/optprobes.c
280
list_del_init(&op->list);
arch/powerpc/kernel/optprobes.c
284
void arch_unoptimize_kprobe(struct optimized_kprobe *op)
arch/powerpc/kernel/optprobes.c
286
arch_arm_kprobe(&op->kp);
arch/powerpc/kernel/optprobes.c
291
struct optimized_kprobe *op;
arch/powerpc/kernel/optprobes.c
294
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/powerpc/kernel/optprobes.c
295
arch_unoptimize_kprobe(op);
arch/powerpc/kernel/optprobes.c
296
list_move(&op->list, done_list);
arch/powerpc/kernel/optprobes.c
300
int arch_within_optimized_kprobe(struct optimized_kprobe *op, kprobe_opcode_t *addr)
arch/powerpc/kernel/optprobes.c
302
return (op->kp.addr <= addr &&
arch/powerpc/kernel/optprobes.c
303
op->kp.addr + (RELATIVEJUMP_SIZE / sizeof(kprobe_opcode_t)) > addr);
arch/powerpc/kernel/optprobes.c
50
struct instruction_op op;
arch/powerpc/kernel/optprobes.c
88
analyse_instr(&op, &regs, ppc_inst_read(p->ainsn.insn)) == 1) {
arch/powerpc/kernel/optprobes.c
89
emulate_update_regs(&regs, &op);
arch/powerpc/kernel/optprobes.c
96
static void optimized_callback(struct optimized_kprobe *op,
arch/powerpc/kernel/rtas_flash.c
366
static void manage_flash(struct rtas_manage_flash_t *args_buf, unsigned int op)
arch/powerpc/kernel/rtas_flash.c
372
NULL, op);
arch/powerpc/kernel/rtas_flash.c
400
int op;
arch/powerpc/kernel/rtas_flash.c
407
op = -1;
arch/powerpc/kernel/rtas_flash.c
413
op = RTAS_REJECT_TMP_IMG;
arch/powerpc/kernel/rtas_flash.c
415
op = RTAS_COMMIT_TMP_IMG;
arch/powerpc/kernel/rtas_flash.c
418
if (op == -1) { /* buf is empty, or contains invalid string */
arch/powerpc/kernel/rtas_flash.c
422
manage_flash(args_buf, op);
arch/powerpc/kernel/signal_32.c
1197
struct sig_dbg_op op;
arch/powerpc/kernel/signal_32.c
1205
if (copy_from_user(&op, dbg + i, sizeof(op)))
arch/powerpc/kernel/signal_32.c
1207
switch (op.dbg_type) {
arch/powerpc/kernel/signal_32.c
1210
if (op.dbg_value) {
arch/powerpc/kernel/signal_32.c
1222
if (op.dbg_value)
arch/powerpc/kernel/signal_32.c
1232
if (op.dbg_value)
arch/powerpc/kernel/trace/ftrace.c
113
static int is_bl_op(ppc_inst_t op)
arch/powerpc/kernel/trace/ftrace.c
115
return (ppc_inst_val(op) & ~PPC_LI_MASK) == PPC_RAW_BL(0);
arch/powerpc/kernel/trace/ftrace.c
680
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/powerpc/kernel/trace/ftrace.c
70
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace.c
73
create_branch(&op, (u32 *)ip, addr, link ? BRANCH_SET_LINK : 0);
arch/powerpc/kernel/trace/ftrace.c
75
return op;
arch/powerpc/kernel/trace/ftrace.c
78
static inline int ftrace_read_inst(unsigned long ip, ppc_inst_t *op)
arch/powerpc/kernel/trace/ftrace.c
80
if (copy_inst_from_kernel_nofault(op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace.c
90
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace.c
93
ret = ftrace_read_inst(ip, &op);
arch/powerpc/kernel/trace/ftrace.c
94
if (!ret && !ppc_inst_equal(op, inst)) {
arch/powerpc/kernel/trace/ftrace.c
96
ip, ppc_inst_as_ulong(inst), ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
101
static int is_b_op(ppc_inst_t op)
arch/powerpc/kernel/trace/ftrace_64_pg.c
103
return (ppc_inst_val(op) & ~PPC_LI_MASK) == PPC_RAW_BRANCH(0);
arch/powerpc/kernel/trace/ftrace_64_pg.c
106
static unsigned long find_bl_target(unsigned long ip, ppc_inst_t op)
arch/powerpc/kernel/trace/ftrace_64_pg.c
110
offset = PPC_LI(ppc_inst_val(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
137
ppc_inst_t op, pop;
arch/powerpc/kernel/trace/ftrace_64_pg.c
146
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
152
if (!is_bl_op(op)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
153
pr_err("Not expected bl: opcode is %08lx\n", ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
158
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
177
if (copy_inst_from_kernel_nofault(&op, (void *)(ip - 4))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
183
if (!ppc_inst_equal(op, ppc_inst(PPC_RAW_MFLR(_R0))) &&
arch/powerpc/kernel/trace/ftrace_64_pg.c
184
!ppc_inst_equal(op, ppc_inst(PPC_INST_STD_LR))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
186
ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
194
if (copy_inst_from_kernel_nofault(&op, (void *)(ip + 4))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
199
if (!ppc_inst_equal(op, ppc_inst(PPC_INST_LD_TOC))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
201
ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
279
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
288
if (copy_inst_from_kernel_nofault(&op, (void *)tramp)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
294
if (!is_b_op(op)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
300
ptr = find_bl_target(tramp, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
329
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
332
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
338
if (!is_bl_op(op)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
339
pr_err("Not expected bl: opcode is %08lx\n", ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
344
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
408
ppc_inst_t op[2];
arch/powerpc/kernel/trace/ftrace_64_pg.c
417
if (copy_inst_from_kernel_nofault(op, ip))
arch/powerpc/kernel/trace/ftrace_64_pg.c
421
copy_inst_from_kernel_nofault(op + 1, ip + 4))
arch/powerpc/kernel/trace/ftrace_64_pg.c
424
if (!expected_nop_sequence(ip, op[0], op[1])) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
426
ppc_inst_as_ulong(op[0]), ppc_inst_as_ulong(op[1]));
arch/powerpc/kernel/trace/ftrace_64_pg.c
472
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
48
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
489
if (copy_inst_from_kernel_nofault(&op, ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
494
if (!ppc_inst_equal(op, ppc_inst(PPC_RAW_NOP()))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
496
ip, ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
53
create_branch(&op, (u32 *)ip, addr, link ? BRANCH_SET_LINK : 0);
arch/powerpc/kernel/trace/ftrace_64_pg.c
545
ppc_inst_t op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
55
return op;
arch/powerpc/kernel/trace/ftrace_64_pg.c
560
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
566
if (!is_bl_op(op)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
567
pr_err("Not expected bl: opcode is %08lx\n", ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
572
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
810
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/powerpc/kernel/trace/ftrace_64_pg.c
96
static int is_bl_op(ppc_inst_t op)
arch/powerpc/kernel/trace/ftrace_64_pg.c
98
return (ppc_inst_val(op) & ~PPC_LI_MASK) == PPC_RAW_BL(0);
arch/powerpc/kvm/emulate_loadstore.c
104
int instr_byte_swap = op.type & BYTEREV;
arch/powerpc/kvm/emulate_loadstore.c
106
if (op.type & SIGNEXT)
arch/powerpc/kvm/emulate_loadstore.c
108
op.reg, size, !instr_byte_swap);
arch/powerpc/kvm/emulate_loadstore.c
111
op.reg, size, !instr_byte_swap);
arch/powerpc/kvm/emulate_loadstore.c
113
if ((op.type & UPDATE) && (emulated != EMULATE_FAIL))
arch/powerpc/kvm/emulate_loadstore.c
114
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
123
if (op.type & FPCONV)
arch/powerpc/kvm/emulate_loadstore.c
126
if (op.type & SIGNEXT)
arch/powerpc/kvm/emulate_loadstore.c
128
KVM_MMIO_REG_FPR|op.reg, size, 1);
arch/powerpc/kvm/emulate_loadstore.c
131
KVM_MMIO_REG_FPR|op.reg, size, 1);
arch/powerpc/kvm/emulate_loadstore.c
133
if ((op.type & UPDATE) && (emulated != EMULATE_FAIL))
arch/powerpc/kvm/emulate_loadstore.c
134
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
168
KVM_MMIO_REG_VMX|op.reg,
arch/powerpc/kvm/emulate_loadstore.c
173
KVM_MMIO_REG_VMX|op.reg,
arch/powerpc/kvm/emulate_loadstore.c
182
if (op.vsx_flags & VSX_CHECK_VEC) {
arch/powerpc/kvm/emulate_loadstore.c
190
if (op.vsx_flags & VSX_FPCONV)
arch/powerpc/kvm/emulate_loadstore.c
193
if (op.element_size == 8) {
arch/powerpc/kvm/emulate_loadstore.c
194
if (op.vsx_flags & VSX_SPLAT)
arch/powerpc/kvm/emulate_loadstore.c
200
} else if (op.element_size == 4) {
arch/powerpc/kvm/emulate_loadstore.c
201
if (op.vsx_flags & VSX_SPLAT)
arch/powerpc/kvm/emulate_loadstore.c
210
if (size < op.element_size) {
arch/powerpc/kvm/emulate_loadstore.c
216
size/op.element_size;
arch/powerpc/kvm/emulate_loadstore.c
217
io_size_each = op.element_size;
arch/powerpc/kvm/emulate_loadstore.c
221
KVM_MMIO_REG_VSX|op.reg, io_size_each,
arch/powerpc/kvm/emulate_loadstore.c
222
1, op.type & SIGNEXT);
arch/powerpc/kvm/emulate_loadstore.c
227
int instr_byte_swap = op.type & BYTEREV;
arch/powerpc/kvm/emulate_loadstore.c
229
emulated = kvmppc_handle_store(vcpu, kvmppc_get_gpr(vcpu, op.reg),
arch/powerpc/kvm/emulate_loadstore.c
232
if ((op.type & UPDATE) && (emulated != EMULATE_FAIL))
arch/powerpc/kvm/emulate_loadstore.c
233
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
250
if (op.type & FPCONV)
arch/powerpc/kvm/emulate_loadstore.c
254
kvmppc_get_fpr(vcpu, op.reg), size, 1);
arch/powerpc/kvm/emulate_loadstore.c
256
if ((op.type & UPDATE) && (emulated != EMULATE_FAIL))
arch/powerpc/kvm/emulate_loadstore.c
257
kvmppc_set_gpr(vcpu, op.update_reg, vcpu->arch.vaddr_accessed);
arch/powerpc/kvm/emulate_loadstore.c
294
op.reg, 8, 1);
arch/powerpc/kvm/emulate_loadstore.c
298
op.reg, size, 1);
arch/powerpc/kvm/emulate_loadstore.c
307
if (op.vsx_flags & VSX_CHECK_VEC) {
arch/powerpc/kvm/emulate_loadstore.c
319
if (op.vsx_flags & VSX_FPCONV)
arch/powerpc/kvm/emulate_loadstore.c
322
if (op.element_size == 8)
arch/powerpc/kvm/emulate_loadstore.c
325
else if (op.element_size == 4)
arch/powerpc/kvm/emulate_loadstore.c
331
if (size < op.element_size) {
arch/powerpc/kvm/emulate_loadstore.c
337
size/op.element_size;
arch/powerpc/kvm/emulate_loadstore.c
338
io_size_each = op.element_size;
arch/powerpc/kvm/emulate_loadstore.c
342
op.reg, io_size_each, 1);
arch/powerpc/kvm/emulate_loadstore.c
76
struct instruction_op op;
arch/powerpc/kvm/emulate_loadstore.c
96
if (analyse_instr(&op, &vcpu->arch.regs, inst) == 0) {
arch/powerpc/kvm/emulate_loadstore.c
97
int type = op.type & INSTR_TYPE_MASK;
arch/powerpc/kvm/emulate_loadstore.c
98
int size = GETSIZE(op.type);
arch/powerpc/lib/sstep.c
1004
static nokprobe_inline int do_vsx_store(struct instruction_op *op,
arch/powerpc/lib/sstep.c
1008
int reg = op->reg;
arch/powerpc/lib/sstep.c
1012
int size = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
1047
emulate_vsx_store(op, buf, mem, cross_endian);
arch/powerpc/lib/sstep.c
1093
#define __put_user_asmx(x, addr, err, op, cr) \
arch/powerpc/lib/sstep.c
1097
"1: " op " %2,0,%3\n" \
arch/powerpc/lib/sstep.c
1109
#define __get_user_asmx(x, addr, err, op) \
arch/powerpc/lib/sstep.c
1113
"1: "op" %1,0,%2\n" \
arch/powerpc/lib/sstep.c
1124
#define __cacheop_user_asmx(addr, err, op) \
arch/powerpc/lib/sstep.c
1126
"1: "op" 0,%1\n" \
arch/powerpc/lib/sstep.c
1137
struct instruction_op *op)
arch/powerpc/lib/sstep.c
1139
long val = op->val;
arch/powerpc/lib/sstep.c
1141
op->type |= SETCC;
arch/powerpc/lib/sstep.c
1142
op->ccval = (regs->ccr & 0x0fffffff) | ((regs->xer >> 3) & 0x10000000);
arch/powerpc/lib/sstep.c
1146
op->ccval |= 0x80000000;
arch/powerpc/lib/sstep.c
1148
op->ccval |= 0x40000000;
arch/powerpc/lib/sstep.c
1150
op->ccval |= 0x20000000;
arch/powerpc/lib/sstep.c
1153
static nokprobe_inline void set_ca32(struct instruction_op *op, bool val)
arch/powerpc/lib/sstep.c
1157
op->xerval |= XER_CA32;
arch/powerpc/lib/sstep.c
1159
op->xerval &= ~XER_CA32;
arch/powerpc/lib/sstep.c
1164
struct instruction_op *op, int rd,
arch/powerpc/lib/sstep.c
1172
op->type = COMPUTE | SETREG | SETXER;
arch/powerpc/lib/sstep.c
1173
op->reg = rd;
arch/powerpc/lib/sstep.c
1174
op->val = val;
arch/powerpc/lib/sstep.c
1177
op->xerval = regs->xer;
arch/powerpc/lib/sstep.c
1179
op->xerval |= XER_CA;
arch/powerpc/lib/sstep.c
1181
op->xerval &= ~XER_CA;
arch/powerpc/lib/sstep.c
1183
set_ca32(op, (unsigned int)val < (unsigned int)val1 ||
arch/powerpc/lib/sstep.c
1188
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1193
op->type = COMPUTE | SETCC;
arch/powerpc/lib/sstep.c
1202
op->ccval = (regs->ccr & ~(0xf << shift)) | (crval << shift);
arch/powerpc/lib/sstep.c
1206
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1212
op->type = COMPUTE | SETCC;
arch/powerpc/lib/sstep.c
1221
op->ccval = (regs->ccr & ~(0xf << shift)) | (crval << shift);
arch/powerpc/lib/sstep.c
1225
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1237
op->val = out_val;
arch/powerpc/lib/sstep.c
1245
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1256
op->val = out;
arch/powerpc/lib/sstep.c
1262
op->val = out & 0x0000003f0000003fULL;
arch/powerpc/lib/sstep.c
1267
op->val = out; /* popcntd */
arch/powerpc/lib/sstep.c
1272
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1285
op->val = perm;
arch/powerpc/lib/sstep.c
1293
struct instruction_op *op,
arch/powerpc/lib/sstep.c
1300
op->val = res & 0x0000000100000001ULL;
arch/powerpc/lib/sstep.c
1305
op->val = res & 1; /*prtyd */
arch/powerpc/lib/sstep.c
1351
int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
arch/powerpc/lib/sstep.c
1367
op->type = COMPUTE;
arch/powerpc/lib/sstep.c
1372
op->type = BRANCH;
arch/powerpc/lib/sstep.c
1376
op->val = truncate_if_32bit(regs->msr, imm);
arch/powerpc/lib/sstep.c
1378
op->type |= SETLK;
arch/powerpc/lib/sstep.c
1379
if (branch_taken(word, regs, op))
arch/powerpc/lib/sstep.c
1380
op->type |= BRTAKEN;
arch/powerpc/lib/sstep.c
1384
op->type = SYSCALL;
arch/powerpc/lib/sstep.c
1387
op->type = SYSCALL_VECTORED_0;
arch/powerpc/lib/sstep.c
1391
op->type = UNKNOWN;
arch/powerpc/lib/sstep.c
1394
op->type = BRANCH | BRTAKEN;
arch/powerpc/lib/sstep.c
1400
op->val = truncate_if_32bit(regs->msr, imm);
arch/powerpc/lib/sstep.c
1402
op->type |= SETLK;
arch/powerpc/lib/sstep.c
1407
op->type = COMPUTE + SETCC;
arch/powerpc/lib/sstep.c
1413
op->ccval = (regs->ccr & ~(0xfUL << rd)) | (val << rd);
arch/powerpc/lib/sstep.c
1418
op->type = BRANCH;
arch/powerpc/lib/sstep.c
1420
op->val = truncate_if_32bit(regs->msr, imm);
arch/powerpc/lib/sstep.c
1422
op->type |= SETLK;
arch/powerpc/lib/sstep.c
1423
if (branch_taken(word, regs, op))
arch/powerpc/lib/sstep.c
1424
op->type |= BRTAKEN;
arch/powerpc/lib/sstep.c
1430
op->type = RFI;
arch/powerpc/lib/sstep.c
1434
op->type = BARRIER | BARRIER_ISYNC;
arch/powerpc/lib/sstep.c
1445
op->type = COMPUTE + SETCC;
arch/powerpc/lib/sstep.c
1452
op->ccval = (regs->ccr & ~(1UL << (31 - rd))) |
arch/powerpc/lib/sstep.c
1460
op->type = BARRIER + BARRIER_SYNC;
arch/powerpc/lib/sstep.c
1464
op->type = BARRIER + BARRIER_LWSYNC;
arch/powerpc/lib/sstep.c
1467
op->type = BARRIER + BARRIER_PTESYNC;
arch/powerpc/lib/sstep.c
1474
op->type = BARRIER + BARRIER_EIEIO;
arch/powerpc/lib/sstep.c
1494
op->reg = rd;
arch/powerpc/lib/sstep.c
1495
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1504
op->type = COMPUTE | PREFIXED;
arch/powerpc/lib/sstep.c
1505
op->val = mlsd_8lsd_ea(word, suffix, regs);
arch/powerpc/lib/sstep.c
1533
"=r" (op->val) : "r" (regs->gpr[ra]),
arch/powerpc/lib/sstep.c
1539
"=r" (op->val) : "r" (regs->gpr[ra]),
arch/powerpc/lib/sstep.c
1545
"=r" (op->val) : "r" (regs->gpr[ra]),
arch/powerpc/lib/sstep.c
1558
op->val = regs->gpr[ra] * (short) word;
arch/powerpc/lib/sstep.c
1563
add_with_carry(regs, op, rd, ~regs->gpr[ra], imm, 1);
arch/powerpc/lib/sstep.c
1573
do_cmp_unsigned(regs, op, val, imm, rd >> 2);
arch/powerpc/lib/sstep.c
1583
do_cmp_signed(regs, op, val, imm, rd >> 2);
arch/powerpc/lib/sstep.c
1588
add_with_carry(regs, op, rd, regs->gpr[ra], imm, 0);
arch/powerpc/lib/sstep.c
1593
add_with_carry(regs, op, rd, regs->gpr[ra], imm, 0);
arch/powerpc/lib/sstep.c
1594
set_cr0(regs, op);
arch/powerpc/lib/sstep.c
1601
op->val = imm;
arch/powerpc/lib/sstep.c
1608
op->val = imm;
arch/powerpc/lib/sstep.c
1618
op->val = regs->nip + (imm << 16) + 4;
arch/powerpc/lib/sstep.c
1621
op->type = UNKNOWN;
arch/powerpc/lib/sstep.c
1629
op->val = (regs->gpr[ra] & ~imm) | (ROTATE(val, rb) & imm);
arch/powerpc/lib/sstep.c
1636
op->val = ROTATE(val, rb) & MASK32(mb, me);
arch/powerpc/lib/sstep.c
1644
op->val = ROTATE(val, rb) & MASK32(mb, me);
arch/powerpc/lib/sstep.c
1648
op->val = regs->gpr[rd] | (unsigned short) word;
arch/powerpc/lib/sstep.c
1653
op->val = regs->gpr[rd] | (imm << 16);
arch/powerpc/lib/sstep.c
1657
op->val = regs->gpr[rd] ^ (unsigned short) word;
arch/powerpc/lib/sstep.c
1662
op->val = regs->gpr[rd] ^ (imm << 16);
arch/powerpc/lib/sstep.c
1666
op->val = regs->gpr[rd] & (unsigned short) word;
arch/powerpc/lib/sstep.c
1667
set_cr0(regs, op);
arch/powerpc/lib/sstep.c
1672
op->val = regs->gpr[rd] & (imm << 16);
arch/powerpc/lib/sstep.c
1673
set_cr0(regs, op);
arch/powerpc/lib/sstep.c
1698
op->val = val;
arch/powerpc/lib/sstep.c
1705
op->val = val & MASK64_L(mb);
arch/powerpc/lib/sstep.c
1708
op->val = val & MASK64_R(mb);
arch/powerpc/lib/sstep.c
1713
op->type = UNKNOWN; /* illegal instruction */
arch/powerpc/lib/sstep.c
1723
op->val = (val) ? val2 : regs->gpr[rb];
arch/powerpc/lib/sstep.c
1743
op->type = MFMSR;
arch/powerpc/lib/sstep.c
1744
op->reg = rd;
arch/powerpc/lib/sstep.c
1749
op->type = MTMSR;
arch/powerpc/lib/sstep.c
1750
op->reg = rd;
arch/powerpc/lib/sstep.c
1751
op->val = 0xffffffff & ~(MSR_ME | MSR_LE);
arch/powerpc/lib/sstep.c
1757
op->type = MTMSR;
arch/powerpc/lib/sstep.c
1758
op->reg = rd;
arch/powerpc/lib/sstep.c
1762
op->val = imm;
arch/powerpc/lib/sstep.c
1776
op->val = regs->ccr & imm;
arch/powerpc/lib/sstep.c
1793
op->val = -1;
arch/powerpc/lib/sstep.c
1796
op->val = 1;
arch/powerpc/lib/sstep.c
1798
op->val = 0;
arch/powerpc/lib/sstep.c
1802
op->type = COMPUTE + SETCC;
arch/powerpc/lib/sstep.c
1805
op->ccval = regs->ccr;
arch/powerpc/lib/sstep.c
1808
op->ccval = (op->ccval & ~imm) |
arch/powerpc/lib/sstep.c
1816
op->type = MFSPR;
arch/powerpc/lib/sstep.c
1817
op->reg = rd;
arch/powerpc/lib/sstep.c
1818
op->spr = spr;
arch/powerpc/lib/sstep.c
1826
op->type = MTSPR;
arch/powerpc/lib/sstep.c
1827
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
1828
op->spr = spr;
arch/powerpc/lib/sstep.c
1847
do_cmp_signed(regs, op, val, val2, rd >> 2);
arch/powerpc/lib/sstep.c
1860
do_cmp_unsigned(regs, op, val, val2, rd >> 2);
arch/powerpc/lib/sstep.c
1864
do_cmpb(regs, op, regs->gpr[rd], regs->gpr[rb]);
arch/powerpc/lib/sstep.c
1871
add_with_carry(regs, op, rd, ~regs->gpr[ra],
arch/powerpc/lib/sstep.c
1876
asm("mulhdu %0,%1,%2" : "=r" (op->val) :
arch/powerpc/lib/sstep.c
1881
add_with_carry(regs, op, rd, regs->gpr[ra],
arch/powerpc/lib/sstep.c
1886
asm("mulhwu %0,%1,%2" : "=r" (op->val) :
arch/powerpc/lib/sstep.c
1891
op->val = regs->gpr[rb] - regs->gpr[ra];
arch/powerpc/lib/sstep.c
1895
asm("mulhd %0,%1,%2" : "=r" (op->val) :
arch/powerpc/lib/sstep.c
1900
asm("mulhw %0,%1,%2" : "=r" (op->val) :
arch/powerpc/lib/sstep.c
1905
op->val = -regs->gpr[ra];
arch/powerpc/lib/sstep.c
1909
add_with_carry(regs, op, rd, ~regs->gpr[ra],
arch/powerpc/lib/sstep.c
1914
add_with_carry(regs, op, rd, regs->gpr[ra],
arch/powerpc/lib/sstep.c
1919
add_with_carry(regs, op, rd, ~regs->gpr[ra], 0L,
arch/powerpc/lib/sstep.c
1924
add_with_carry(regs, op, rd, regs->gpr[ra], 0L,
arch/powerpc/lib/sstep.c
1929
add_with_carry(regs, op, rd, ~regs->gpr[ra], -1L,
arch/powerpc/lib/sstep.c
1934
op->val = regs->gpr[ra] * regs->gpr[rb];
arch/powerpc/lib/sstep.c
1938
add_with_carry(regs, op, rd, regs->gpr[ra], -1L,
arch/powerpc/lib/sstep.c
1943
op->val = (long)(int) regs->gpr[ra] *
arch/powerpc/lib/sstep.c
1951
op->val = regs->gpr[ra] % regs->gpr[rb];
arch/powerpc/lib/sstep.c
1955
op->val = regs->gpr[ra] + regs->gpr[rb];
arch/powerpc/lib/sstep.c
1961
op->val = (unsigned int) regs->gpr[ra] %
arch/powerpc/lib/sstep.c
1966
op->val = regs->gpr[ra] / regs->gpr[rb];
arch/powerpc/lib/sstep.c
1970
op->val = (unsigned int) regs->gpr[ra] /
arch/powerpc/lib/sstep.c
1975
op->val = (long int) regs->gpr[ra] /
arch/powerpc/lib/sstep.c
1980
op->val = (int) regs->gpr[ra] /
arch/powerpc/lib/sstep.c
1986
"=r" (op->val) : "r" (regs->gpr[ra]),
arch/powerpc/lib/sstep.c
1991
"=r" (op->val) : "r" (regs->gpr[ra]),
arch/powerpc/lib/sstep.c
2001
asm volatile(PPC_DARN(%0, 0) : "=r" (op->val));
arch/powerpc/lib/sstep.c
2006
asm volatile(PPC_DARN(%0, 1) : "=r" (op->val));
arch/powerpc/lib/sstep.c
2011
asm volatile(PPC_DARN(%0, 2) : "=r" (op->val));
arch/powerpc/lib/sstep.c
2020
op->val = (long int) regs->gpr[ra] %
arch/powerpc/lib/sstep.c
2027
op->val = (int) regs->gpr[ra] %
arch/powerpc/lib/sstep.c
2037
op->val = ( val ? __builtin_clz(val) : 32 );
arch/powerpc/lib/sstep.c
2042
op->val = ( val ? __builtin_clzl(val) : 64 );
arch/powerpc/lib/sstep.c
2046
op->val = regs->gpr[rd] & regs->gpr[rb];
arch/powerpc/lib/sstep.c
2050
op->val = regs->gpr[rd] & ~regs->gpr[rb];
arch/powerpc/lib/sstep.c
2054
do_popcnt(regs, op, regs->gpr[rd], 8);
arch/powerpc/lib/sstep.c
2058
op->val = ~(regs->gpr[rd] | regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2062
do_prty(regs, op, regs->gpr[rd], 32);
arch/powerpc/lib/sstep.c
2066
do_prty(regs, op, regs->gpr[rd], 64);
arch/powerpc/lib/sstep.c
2070
do_bpermd(regs, op, regs->gpr[rd], regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2074
op->val = ~(regs->gpr[rd] ^ regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2078
op->val = regs->gpr[rd] ^ regs->gpr[rb];
arch/powerpc/lib/sstep.c
2082
do_popcnt(regs, op, regs->gpr[rd], 32);
arch/powerpc/lib/sstep.c
2086
op->val = regs->gpr[rd] | ~regs->gpr[rb];
arch/powerpc/lib/sstep.c
2090
op->val = regs->gpr[rd] | regs->gpr[rb];
arch/powerpc/lib/sstep.c
2094
op->val = ~(regs->gpr[rd] & regs->gpr[rb]);
arch/powerpc/lib/sstep.c
2098
do_popcnt(regs, op, regs->gpr[rd], 64);
arch/powerpc/lib/sstep.c
2105
op->val = (val ? __builtin_ctz(val) : 32);
arch/powerpc/lib/sstep.c
2112
op->val = (val ? __builtin_ctzl(val) : 64);
arch/powerpc/lib/sstep.c
2116
op->val = (signed short) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2120
op->val = (signed char) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2124
op->val = (signed int) regs->gpr[rd];
arch/powerpc/lib/sstep.c
2134
op->val = (regs->gpr[rd] << sh) & 0xffffffffUL;
arch/powerpc/lib/sstep.c
2136
op->val = 0;
arch/powerpc/lib/sstep.c
2142
op->val = (regs->gpr[rd] & 0xffffffffUL) >> sh;
arch/powerpc/lib/sstep.c
2144
op->val = 0;
arch/powerpc/lib/sstep.c
2148
op->type = COMPUTE + SETREG + SETXER;
arch/powerpc/lib/sstep.c
2151
op->val = ival >> (sh < 32 ? sh : 31);
arch/powerpc/lib/sstep.c
2152
op->xerval = regs->xer;
arch/powerpc/lib/sstep.c
2154
op->xerval |= XER_CA;
arch/powerpc/lib/sstep.c
2156
op->xerval &= ~XER_CA;
arch/powerpc/lib/sstep.c
2157
set_ca32(op, op->xerval & XER_CA);
arch/powerpc/lib/sstep.c
2161
op->type = COMPUTE + SETREG + SETXER;
arch/powerpc/lib/sstep.c
2164
op->val = ival >> sh;
arch/powerpc/lib/sstep.c
2165
op->xerval = regs->xer;
arch/powerpc/lib/sstep.c
2167
op->xerval |= XER_CA;
arch/powerpc/lib/sstep.c
2169
op->xerval &= ~XER_CA;
arch/powerpc/lib/sstep.c
2170
set_ca32(op, op->xerval & XER_CA);
arch/powerpc/lib/sstep.c
2177
op->val = regs->gpr[rd] << sh;
arch/powerpc/lib/sstep.c
2179
op->val = 0;
arch/powerpc/lib/sstep.c
2185
op->val = regs->gpr[rd] >> sh;
arch/powerpc/lib/sstep.c
2187
op->val = 0;
arch/powerpc/lib/sstep.c
2191
op->type = COMPUTE + SETREG + SETXER;
arch/powerpc/lib/sstep.c
2194
op->val = ival >> (sh < 64 ? sh : 63);
arch/powerpc/lib/sstep.c
2195
op->xerval = regs->xer;
arch/powerpc/lib/sstep.c
2197
op->xerval |= XER_CA;
arch/powerpc/lib/sstep.c
2199
op->xerval &= ~XER_CA;
arch/powerpc/lib/sstep.c
2200
set_ca32(op, op->xerval & XER_CA);
arch/powerpc/lib/sstep.c
2205
op->type = COMPUTE + SETREG + SETXER;
arch/powerpc/lib/sstep.c
2208
op->val = ival >> sh;
arch/powerpc/lib/sstep.c
2209
op->xerval = regs->xer;
arch/powerpc/lib/sstep.c
2211
op->xerval |= XER_CA;
arch/powerpc/lib/sstep.c
2213
op->xerval &= ~XER_CA;
arch/powerpc/lib/sstep.c
2214
set_ca32(op, op->xerval & XER_CA);
arch/powerpc/lib/sstep.c
2221
op->type = COMPUTE + SETREG;
arch/powerpc/lib/sstep.c
2225
op->val = ROTATE(val, sh) & MASK64(0, 63 - sh);
arch/powerpc/lib/sstep.c
2227
op->val = val;
arch/powerpc/lib/sstep.c
2236
op->type = MKOP(CACHEOP, DCBST, 0);
arch/powerpc/lib/sstep.c
2237
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2241
op->type = MKOP(CACHEOP, DCBF, 0);
arch/powerpc/lib/sstep.c
2242
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2246
op->type = MKOP(CACHEOP, DCBTST, 0);
arch/powerpc/lib/sstep.c
2247
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2248
op->reg = rd;
arch/powerpc/lib/sstep.c
2252
op->type = MKOP(CACHEOP, DCBTST, 0);
arch/powerpc/lib/sstep.c
2253
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2254
op->reg = rd;
arch/powerpc/lib/sstep.c
2258
op->type = MKOP(CACHEOP, ICBI, 0);
arch/powerpc/lib/sstep.c
2259
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2263
op->type = MKOP(CACHEOP, DCBZ, 0);
arch/powerpc/lib/sstep.c
2264
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2273
op->type = UNKNOWN;
arch/powerpc/lib/sstep.c
2274
op->update_reg = ra;
arch/powerpc/lib/sstep.c
2275
op->reg = rd;
arch/powerpc/lib/sstep.c
2276
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
2278
op->vsx_flags = 0;
arch/powerpc/lib/sstep.c
2283
op->ea = xform_ea(word, regs);
arch/powerpc/lib/sstep.c
2286
op->type = MKOP(LARX, 0, 4);
arch/powerpc/lib/sstep.c
2290
op->type = MKOP(STCX, 0, 4);
arch/powerpc/lib/sstep.c
2295
op->type = MKOP(LARX, 0, 1);
arch/powerpc/lib/sstep.c
2299
op->type = MKOP(STCX, 0, 1);
arch/powerpc/lib/sstep.c
2303
op->type = MKOP(LARX, 0, 2);
arch/powerpc/lib/sstep.c
2307
op->type = MKOP(STCX, 0, 2);
arch/powerpc/lib/sstep.c
2312
op->type = MKOP(LARX, 0, 8);
arch/powerpc/lib/sstep.c
2316
op->type = MKOP(STCX, 0, 8);
arch/powerpc/lib/sstep.c
2321
op->type = MKOP(LARX, 0, 16);
arch/powerpc/lib/sstep.c
2326
op->type = MKOP(STCX, 0, 16);
arch/powerpc/lib/sstep.c
2332
op->type = MKOP(LOAD, u, 4);
arch/powerpc/lib/sstep.c
2337
op->type = MKOP(LOAD, u, 1);
arch/powerpc/lib/sstep.c
2346
op->type = MKOP(LOAD_VMX, 0, 1);
arch/powerpc/lib/sstep.c
2347
op->element_size = 1;
arch/powerpc/lib/sstep.c
2351
op->type = MKOP(LOAD_VMX, 0, 2);
arch/powerpc/lib/sstep.c
2352
op->element_size = 2;
arch/powerpc/lib/sstep.c
2356
op->type = MKOP(LOAD_VMX, 0, 4);
arch/powerpc/lib/sstep.c
2357
op->element_size = 4;
arch/powerpc/lib/sstep.c
2362
op->type = MKOP(LOAD_VMX, 0, 16);
arch/powerpc/lib/sstep.c
2363
op->element_size = 16;
arch/powerpc/lib/sstep.c
2367
op->type = MKOP(STORE_VMX, 0, 1);
arch/powerpc/lib/sstep.c
2368
op->element_size = 1;
arch/powerpc/lib/sstep.c
2372
op->type = MKOP(STORE_VMX, 0, 2);
arch/powerpc/lib/sstep.c
2373
op->element_size = 2;
arch/powerpc/lib/sstep.c
2377
op->type = MKOP(STORE_VMX, 0, 4);
arch/powerpc/lib/sstep.c
2378
op->element_size = 4;
arch/powerpc/lib/sstep.c
2383
op->type = MKOP(STORE_VMX, 0, 16);
arch/powerpc/lib/sstep.c
2390
op->type = MKOP(LOAD, u, 8);
arch/powerpc/lib/sstep.c
2395
op->type = MKOP(STORE, u, 8);
arch/powerpc/lib/sstep.c
2401
op->type = MKOP(STORE, u, 4);
arch/powerpc/lib/sstep.c
2406
op->type = MKOP(STORE, u, 1);
arch/powerpc/lib/sstep.c
2411
op->type = MKOP(LOAD, u, 2);
arch/powerpc/lib/sstep.c
2417
op->type = MKOP(LOAD, SIGNEXT | u, 4);
arch/powerpc/lib/sstep.c
2423
op->type = MKOP(LOAD, SIGNEXT | u, 2);
arch/powerpc/lib/sstep.c
2428
op->type = MKOP(STORE, u, 2);
arch/powerpc/lib/sstep.c
2433
op->type = MKOP(LOAD, BYTEREV, 8);
arch/powerpc/lib/sstep.c
2438
op->type = MKOP(LOAD_MULTI, 0, regs->xer & 0x7f);
arch/powerpc/lib/sstep.c
2442
op->type = MKOP(LOAD, BYTEREV, 4);
arch/powerpc/lib/sstep.c
2448
op->type = MKOP(LOAD_MULTI, 0, rb);
arch/powerpc/lib/sstep.c
2449
op->ea = ra ? regs->gpr[ra] : 0;
arch/powerpc/lib/sstep.c
2455
op->type = MKOP(LOAD_FP, u | FPCONV, 4);
arch/powerpc/lib/sstep.c
2460
op->type = MKOP(LOAD_FP, u, 8);
arch/powerpc/lib/sstep.c
2465
op->type = MKOP(STORE_FP, u | FPCONV, 4);
arch/powerpc/lib/sstep.c
2470
op->type = MKOP(STORE_FP, u, 8);
arch/powerpc/lib/sstep.c
2475
op->type = MKOP(LOAD_FP, 0, 16);
arch/powerpc/lib/sstep.c
2479
op->type = MKOP(LOAD_FP, SIGNEXT, 4);
arch/powerpc/lib/sstep.c
2483
op->type = MKOP(LOAD_FP, 0, 4);
arch/powerpc/lib/sstep.c
2487
op->type = MKOP(STORE_FP, 0, 16);
arch/powerpc/lib/sstep.c
2491
op->type = MKOP(STORE_FP, 0, 4);
arch/powerpc/lib/sstep.c
2498
op->type = MKOP(STORE, BYTEREV, 8);
arch/powerpc/lib/sstep.c
2499
op->val = byterev_8(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2504
op->type = MKOP(STORE_MULTI, 0, regs->xer & 0x7f);
arch/powerpc/lib/sstep.c
2508
op->type = MKOP(STORE, BYTEREV, 4);
arch/powerpc/lib/sstep.c
2509
op->val = byterev_4(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2515
op->type = MKOP(STORE_MULTI, 0, rb);
arch/powerpc/lib/sstep.c
2516
op->ea = ra ? regs->gpr[ra] : 0;
arch/powerpc/lib/sstep.c
2520
op->type = MKOP(LOAD, BYTEREV, 2);
arch/powerpc/lib/sstep.c
2524
op->type = MKOP(STORE, BYTEREV, 2);
arch/powerpc/lib/sstep.c
2525
op->val = byterev_2(regs->gpr[rd]);
arch/powerpc/lib/sstep.c
2530
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2531
op->type = MKOP(LOAD_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2532
op->element_size = 8;
arch/powerpc/lib/sstep.c
2536
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2537
op->type = MKOP(LOAD_VSX, SIGNEXT, 4);
arch/powerpc/lib/sstep.c
2538
op->element_size = 8;
arch/powerpc/lib/sstep.c
2542
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2543
op->type = MKOP(STORE_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2544
op->element_size = 8;
arch/powerpc/lib/sstep.c
2550
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2551
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2552
op->element_size = 16;
arch/powerpc/lib/sstep.c
2553
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2561
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2562
op->ea = ra ? regs->gpr[ra] : 0;
arch/powerpc/lib/sstep.c
2566
op->type = MKOP(LOAD_VSX, 0, nb);
arch/powerpc/lib/sstep.c
2567
op->element_size = 16;
arch/powerpc/lib/sstep.c
2568
op->vsx_flags = ((word & 0x20) ? VSX_LDLEFT : 0) |
arch/powerpc/lib/sstep.c
2573
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2574
op->type = MKOP(LOAD_VSX, 0, 8);
arch/powerpc/lib/sstep.c
2575
op->element_size = 8;
arch/powerpc/lib/sstep.c
2576
op->vsx_flags = VSX_SPLAT;
arch/powerpc/lib/sstep.c
2582
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2583
op->type = MKOP(LOAD_VSX, 0, 32);
arch/powerpc/lib/sstep.c
2584
op->element_size = 32;
arch/powerpc/lib/sstep.c
2590
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2591
op->type = MKOP(LOAD_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2592
op->element_size = 4;
arch/powerpc/lib/sstep.c
2593
op->vsx_flags = VSX_SPLAT | VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2599
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2600
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2601
op->element_size = 16;
arch/powerpc/lib/sstep.c
2602
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2610
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2611
op->ea = ra ? regs->gpr[ra] : 0;
arch/powerpc/lib/sstep.c
2615
op->type = MKOP(STORE_VSX, 0, nb);
arch/powerpc/lib/sstep.c
2616
op->element_size = 16;
arch/powerpc/lib/sstep.c
2617
op->vsx_flags = ((word & 0x20) ? VSX_LDLEFT : 0) |
arch/powerpc/lib/sstep.c
2624
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2625
op->type = MKOP(STORE_VSX, 0, 32);
arch/powerpc/lib/sstep.c
2626
op->element_size = 32;
arch/powerpc/lib/sstep.c
2629
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2630
op->type = MKOP(LOAD_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2631
op->element_size = 8;
arch/powerpc/lib/sstep.c
2632
op->vsx_flags = VSX_FPCONV;
arch/powerpc/lib/sstep.c
2636
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2637
op->type = MKOP(LOAD_VSX, 0, 8);
arch/powerpc/lib/sstep.c
2638
op->element_size = 8;
arch/powerpc/lib/sstep.c
2642
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2643
op->type = MKOP(STORE_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2644
op->element_size = 8;
arch/powerpc/lib/sstep.c
2645
op->vsx_flags = VSX_FPCONV;
arch/powerpc/lib/sstep.c
2649
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2650
op->type = MKOP(STORE_VSX, 0, 8);
arch/powerpc/lib/sstep.c
2651
op->element_size = 8;
arch/powerpc/lib/sstep.c
2655
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2656
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2657
op->element_size = 4;
arch/powerpc/lib/sstep.c
2663
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2664
op->type = MKOP(LOAD_VSX, 0, 1);
arch/powerpc/lib/sstep.c
2665
op->element_size = 8;
arch/powerpc/lib/sstep.c
2666
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2672
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2673
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2674
op->element_size = 2;
arch/powerpc/lib/sstep.c
2675
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2681
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2682
op->type = MKOP(LOAD_VSX, 0, 2);
arch/powerpc/lib/sstep.c
2683
op->element_size = 8;
arch/powerpc/lib/sstep.c
2684
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2688
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2689
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2690
op->element_size = 8;
arch/powerpc/lib/sstep.c
2696
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2697
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2698
op->element_size = 1;
arch/powerpc/lib/sstep.c
2699
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2703
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2704
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2705
op->element_size = 4;
arch/powerpc/lib/sstep.c
2711
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2712
op->type = MKOP(STORE_VSX, 0, 1);
arch/powerpc/lib/sstep.c
2713
op->element_size = 8;
arch/powerpc/lib/sstep.c
2714
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2720
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2721
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2722
op->element_size = 2;
arch/powerpc/lib/sstep.c
2723
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2729
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2730
op->type = MKOP(STORE_VSX, 0, 2);
arch/powerpc/lib/sstep.c
2731
op->element_size = 8;
arch/powerpc/lib/sstep.c
2732
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2736
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2737
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2738
op->element_size = 8;
arch/powerpc/lib/sstep.c
2744
op->reg = rd | ((word & 1) << 5);
arch/powerpc/lib/sstep.c
2745
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2746
op->element_size = 1;
arch/powerpc/lib/sstep.c
2747
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2756
op->type = MKOP(LOAD, u, 4);
arch/powerpc/lib/sstep.c
2757
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2762
op->type = MKOP(LOAD, u, 1);
arch/powerpc/lib/sstep.c
2763
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2768
op->type = MKOP(STORE, u, 4);
arch/powerpc/lib/sstep.c
2769
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2774
op->type = MKOP(STORE, u, 1);
arch/powerpc/lib/sstep.c
2775
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2780
op->type = MKOP(LOAD, u, 2);
arch/powerpc/lib/sstep.c
2781
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2786
op->type = MKOP(LOAD, SIGNEXT | u, 2);
arch/powerpc/lib/sstep.c
2787
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2792
op->type = MKOP(STORE, u, 2);
arch/powerpc/lib/sstep.c
2793
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2799
op->type = MKOP(LOAD_MULTI, 0, 4 * (32 - rd));
arch/powerpc/lib/sstep.c
2800
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2804
op->type = MKOP(STORE_MULTI, 0, 4 * (32 - rd));
arch/powerpc/lib/sstep.c
2805
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2811
op->type = MKOP(LOAD_FP, u | FPCONV, 4);
arch/powerpc/lib/sstep.c
2812
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2817
op->type = MKOP(LOAD_FP, u, 8);
arch/powerpc/lib/sstep.c
2818
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2823
op->type = MKOP(STORE_FP, u | FPCONV, 4);
arch/powerpc/lib/sstep.c
2824
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2829
op->type = MKOP(STORE_FP, u, 8);
arch/powerpc/lib/sstep.c
2830
op->ea = dform_ea(word, regs);
arch/powerpc/lib/sstep.c
2837
op->type = MKOP(LOAD, 0, 16);
arch/powerpc/lib/sstep.c
2838
op->ea = dqform_ea(word, regs);
arch/powerpc/lib/sstep.c
2844
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2849
op->type = MKOP(LOAD_FP, 0, 16);
arch/powerpc/lib/sstep.c
2854
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2855
op->type = MKOP(LOAD_VSX, 0, 8);
arch/powerpc/lib/sstep.c
2856
op->element_size = 8;
arch/powerpc/lib/sstep.c
2857
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2862
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2863
op->type = MKOP(LOAD_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2864
op->element_size = 8;
arch/powerpc/lib/sstep.c
2865
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2873
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2876
op->type = MKOP(LOAD, 0, 8);
arch/powerpc/lib/sstep.c
2879
op->type = MKOP(LOAD, UPDATE, 8);
arch/powerpc/lib/sstep.c
2882
op->type = MKOP(LOAD, SIGNEXT, 4);
arch/powerpc/lib/sstep.c
2892
op->ea = dqform_ea(word, regs);
arch/powerpc/lib/sstep.c
2893
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
2894
op->element_size = 32;
arch/powerpc/lib/sstep.c
2897
op->type = MKOP(LOAD_VSX, 0, 32);
arch/powerpc/lib/sstep.c
2900
op->type = MKOP(STORE_VSX, 0, 32);
arch/powerpc/lib/sstep.c
2909
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2910
op->type = MKOP(STORE_FP, 0, 16);
arch/powerpc/lib/sstep.c
2916
op->ea = dqform_ea(word, regs);
arch/powerpc/lib/sstep.c
2918
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2919
op->type = MKOP(LOAD_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2920
op->element_size = 16;
arch/powerpc/lib/sstep.c
2921
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2928
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2929
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2930
op->type = MKOP(STORE_VSX, 0, 8);
arch/powerpc/lib/sstep.c
2931
op->element_size = 8;
arch/powerpc/lib/sstep.c
2932
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2939
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2940
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2941
op->type = MKOP(STORE_VSX, 0, 4);
arch/powerpc/lib/sstep.c
2942
op->element_size = 8;
arch/powerpc/lib/sstep.c
2943
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2949
op->ea = dqform_ea(word, regs);
arch/powerpc/lib/sstep.c
2951
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
2952
op->type = MKOP(STORE_VSX, 0, 16);
arch/powerpc/lib/sstep.c
2953
op->element_size = 16;
arch/powerpc/lib/sstep.c
2954
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
2962
op->ea = dsform_ea(word, regs);
arch/powerpc/lib/sstep.c
2965
op->type = MKOP(STORE, 0, 8);
arch/powerpc/lib/sstep.c
2968
op->type = MKOP(STORE, UPDATE, 8);
arch/powerpc/lib/sstep.c
2972
op->type = MKOP(STORE, 0, 16);
arch/powerpc/lib/sstep.c
2982
op->update_reg = ra;
arch/powerpc/lib/sstep.c
2984
op->reg = rd;
arch/powerpc/lib/sstep.c
2985
op->val = regs->gpr[rd];
arch/powerpc/lib/sstep.c
2993
op->ea = mlsd_8lsd_ea(word, suffix, regs);
arch/powerpc/lib/sstep.c
2996
op->type = MKOP(LOAD, PREFIXED | SIGNEXT, 4);
arch/powerpc/lib/sstep.c
3000
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3001
op->type = MKOP(LOAD_VSX, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3002
op->element_size = 8;
arch/powerpc/lib/sstep.c
3003
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3006
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3007
op->type = MKOP(LOAD_VSX, PREFIXED, 4);
arch/powerpc/lib/sstep.c
3008
op->element_size = 8;
arch/powerpc/lib/sstep.c
3009
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3012
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3013
op->type = MKOP(STORE_VSX, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3014
op->element_size = 8;
arch/powerpc/lib/sstep.c
3015
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3018
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3019
op->type = MKOP(STORE_VSX, PREFIXED, 4);
arch/powerpc/lib/sstep.c
3020
op->element_size = 8;
arch/powerpc/lib/sstep.c
3021
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3024
op->reg += 32;
arch/powerpc/lib/sstep.c
3027
op->type = MKOP(LOAD_VSX, PREFIXED, 16);
arch/powerpc/lib/sstep.c
3028
op->element_size = 16;
arch/powerpc/lib/sstep.c
3029
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3032
op->reg = rd + 32;
arch/powerpc/lib/sstep.c
3035
op->type = MKOP(STORE_VSX, PREFIXED, 16);
arch/powerpc/lib/sstep.c
3036
op->element_size = 16;
arch/powerpc/lib/sstep.c
3037
op->vsx_flags = VSX_CHECK_VEC;
arch/powerpc/lib/sstep.c
3041
op->type = MKOP(LOAD, PREFIXED, 16);
arch/powerpc/lib/sstep.c
3044
op->type = MKOP(LOAD, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3048
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
3049
op->type = MKOP(LOAD_VSX, PREFIXED, 32);
arch/powerpc/lib/sstep.c
3050
op->element_size = 32;
arch/powerpc/lib/sstep.c
3054
op->type = MKOP(STORE, PREFIXED, 16);
arch/powerpc/lib/sstep.c
3057
op->type = MKOP(STORE, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3061
op->reg = VSX_REGISTER_XTP(rd);
arch/powerpc/lib/sstep.c
3062
op->type = MKOP(STORE_VSX, PREFIXED, 32);
arch/powerpc/lib/sstep.c
3063
op->element_size = 32;
arch/powerpc/lib/sstep.c
3073
op->ea = mlsd_8lsd_ea(word, suffix, regs);
arch/powerpc/lib/sstep.c
3076
op->type = MKOP(LOAD, PREFIXED, 4);
arch/powerpc/lib/sstep.c
3079
op->type = MKOP(LOAD, PREFIXED, 1);
arch/powerpc/lib/sstep.c
3082
op->type = MKOP(STORE, PREFIXED, 4);
arch/powerpc/lib/sstep.c
3085
op->type = MKOP(STORE, PREFIXED, 1);
arch/powerpc/lib/sstep.c
3088
op->type = MKOP(LOAD, PREFIXED, 2);
arch/powerpc/lib/sstep.c
3091
op->type = MKOP(LOAD, PREFIXED | SIGNEXT, 2);
arch/powerpc/lib/sstep.c
3094
op->type = MKOP(STORE, PREFIXED, 2);
arch/powerpc/lib/sstep.c
3097
op->type = MKOP(LOAD_FP, PREFIXED | FPCONV, 4);
arch/powerpc/lib/sstep.c
3100
op->type = MKOP(LOAD_FP, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3103
op->type = MKOP(STORE_FP, PREFIXED | FPCONV, 4);
arch/powerpc/lib/sstep.c
3106
op->type = MKOP(STORE_FP, PREFIXED, 8);
arch/powerpc/lib/sstep.c
3117
if (OP_IS_LOAD_STORE(op->type) && (op->type & UPDATE)) {
arch/powerpc/lib/sstep.c
3118
switch (GETTYPE(op->type)) {
arch/powerpc/lib/sstep.c
3132
if ((GETTYPE(op->type) == LOAD_VSX ||
arch/powerpc/lib/sstep.c
3133
GETTYPE(op->type) == STORE_VSX) &&
arch/powerpc/lib/sstep.c
3142
op->type = UNKNOWN;
arch/powerpc/lib/sstep.c
3147
set_cr0(regs, op);
arch/powerpc/lib/sstep.c
3149
op->reg = ra;
arch/powerpc/lib/sstep.c
3150
op->type |= SETREG;
arch/powerpc/lib/sstep.c
3155
set_cr0(regs, op);
arch/powerpc/lib/sstep.c
3157
op->reg = rd;
arch/powerpc/lib/sstep.c
3158
op->type |= SETREG;
arch/powerpc/lib/sstep.c
3162
op->type = INTERRUPT | 0x700;
arch/powerpc/lib/sstep.c
3163
op->val = SRR1_PROGPRIV;
arch/powerpc/lib/sstep.c
3167
op->type = INTERRUPT | 0x700;
arch/powerpc/lib/sstep.c
3168
op->val = SRR1_PROGTRAP;
arch/powerpc/lib/sstep.c
3226
void emulate_update_regs(struct pt_regs *regs, struct instruction_op *op)
arch/powerpc/lib/sstep.c
3230
next_pc = truncate_if_32bit(regs->msr, regs->nip + GETLENGTH(op->type));
arch/powerpc/lib/sstep.c
3231
switch (GETTYPE(op->type)) {
arch/powerpc/lib/sstep.c
3233
if (op->type & SETREG)
arch/powerpc/lib/sstep.c
3234
regs->gpr[op->reg] = op->val;
arch/powerpc/lib/sstep.c
3235
if (op->type & SETCC)
arch/powerpc/lib/sstep.c
3236
regs->ccr = op->ccval;
arch/powerpc/lib/sstep.c
3237
if (op->type & SETXER)
arch/powerpc/lib/sstep.c
3238
regs->xer = op->xerval;
arch/powerpc/lib/sstep.c
3242
if (op->type & SETLK)
arch/powerpc/lib/sstep.c
3244
if (op->type & BRTAKEN)
arch/powerpc/lib/sstep.c
3245
next_pc = op->val;
arch/powerpc/lib/sstep.c
3246
if (op->type & DECCTR)
arch/powerpc/lib/sstep.c
3251
switch (op->type & BARRIER_MASK) {
arch/powerpc/lib/sstep.c
3273
switch (op->spr) {
arch/powerpc/lib/sstep.c
3275
regs->gpr[op->reg] = regs->xer & 0xffffffffUL;
arch/powerpc/lib/sstep.c
3278
regs->gpr[op->reg] = regs->link;
arch/powerpc/lib/sstep.c
3281
regs->gpr[op->reg] = regs->ctr;
arch/powerpc/lib/sstep.c
3289
switch (op->spr) {
arch/powerpc/lib/sstep.c
3291
regs->xer = op->val & 0xffffffffUL;
arch/powerpc/lib/sstep.c
3294
regs->link = op->val;
arch/powerpc/lib/sstep.c
3297
regs->ctr = op->val;
arch/powerpc/lib/sstep.c
3320
int emulate_loadstore(struct pt_regs *regs, struct instruction_op *op)
arch/powerpc/lib/sstep.c
3330
size = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
3331
type = GETTYPE(op->type);
arch/powerpc/lib/sstep.c
3333
ea = truncate_if_32bit(regs->msr, op->ea);
arch/powerpc/lib/sstep.c
3360
err = do_lqarx(ea, &regs->gpr[op->reg]);
arch/powerpc/lib/sstep.c
3371
regs->gpr[op->reg] = val;
arch/powerpc/lib/sstep.c
3383
__put_user_asmx(op->val, ea, err, "stbcx.", cr);
arch/powerpc/lib/sstep.c
3386
__put_user_asmx(op->val, ea, err, "sthcx.", cr);
arch/powerpc/lib/sstep.c
3390
__put_user_asmx(op->val, ea, err, "stwcx.", cr);
arch/powerpc/lib/sstep.c
3394
__put_user_asmx(op->val, ea, err, "stdcx.", cr);
arch/powerpc/lib/sstep.c
3397
err = do_stqcx(ea, regs->gpr[op->reg],
arch/powerpc/lib/sstep.c
3398
regs->gpr[op->reg + 1], &cr);
arch/powerpc/lib/sstep.c
3415
err = emulate_lq(regs, ea, op->reg, cross_endian);
arch/powerpc/lib/sstep.c
3419
err = read_mem(&regs->gpr[op->reg], ea, size, regs);
arch/powerpc/lib/sstep.c
3421
if (op->type & SIGNEXT)
arch/powerpc/lib/sstep.c
3422
do_signext(&regs->gpr[op->reg], size);
arch/powerpc/lib/sstep.c
3423
if ((op->type & BYTEREV) == (cross_endian ? 0 : BYTEREV))
arch/powerpc/lib/sstep.c
3424
do_byterev(&regs->gpr[op->reg], size);
arch/powerpc/lib/sstep.c
3438
err = do_fp_load(op, ea, regs, cross_endian);
arch/powerpc/lib/sstep.c
3445
err = do_vec_load(op->reg, ea, size, regs, cross_endian);
arch/powerpc/lib/sstep.c
3456
if (op->reg >= 32 && (op->vsx_flags & VSX_CHECK_VEC))
arch/powerpc/lib/sstep.c
3460
err = do_vsx_load(op, ea, regs, cross_endian);
arch/powerpc/lib/sstep.c
3467
rd = op->reg;
arch/powerpc/lib/sstep.c
3489
err = emulate_stq(regs, ea, op->reg, cross_endian);
arch/powerpc/lib/sstep.c
3493
if ((op->type & UPDATE) && size == sizeof(long) &&
arch/powerpc/lib/sstep.c
3494
op->reg == 1 && op->update_reg == 1 && !user_mode(regs) &&
arch/powerpc/lib/sstep.c
3500
do_byterev(&op->val, size);
arch/powerpc/lib/sstep.c
3501
err = write_mem(op->val, ea, size, regs);
arch/powerpc/lib/sstep.c
3508
err = do_fp_store(op, ea, regs, cross_endian);
arch/powerpc/lib/sstep.c
3515
err = do_vec_store(op->reg, ea, size, regs, cross_endian);
arch/powerpc/lib/sstep.c
3526
if (op->reg >= 32 && (op->vsx_flags & VSX_CHECK_VEC))
arch/powerpc/lib/sstep.c
3530
err = do_vsx_store(op, ea, regs, cross_endian);
arch/powerpc/lib/sstep.c
3537
rd = op->reg;
arch/powerpc/lib/sstep.c
3562
if (op->type & UPDATE)
arch/powerpc/lib/sstep.c
3563
regs->gpr[op->update_reg] = op->ea;
arch/powerpc/lib/sstep.c
3578
struct instruction_op op;
arch/powerpc/lib/sstep.c
3583
r = analyse_instr(&op, regs, instr);
arch/powerpc/lib/sstep.c
3587
emulate_update_regs(regs, &op);
arch/powerpc/lib/sstep.c
3592
type = GETTYPE(op.type);
arch/powerpc/lib/sstep.c
3595
err = emulate_loadstore(regs, &op);
arch/powerpc/lib/sstep.c
3603
ea = truncate_if_32bit(regs->msr, op.ea);
arch/powerpc/lib/sstep.c
3606
switch (op.type & CACHEOP_MASK) {
arch/powerpc/lib/sstep.c
3614
if (op.reg == 0)
arch/powerpc/lib/sstep.c
3618
if (op.reg == 0)
arch/powerpc/lib/sstep.c
3635
regs->gpr[op.reg] = regs->msr & MSR_MASK;
arch/powerpc/lib/sstep.c
3639
val = regs->gpr[op.reg];
arch/powerpc/lib/sstep.c
3644
regs_set_return_msr(regs, (regs->msr & ~op.val) | (val & op.val));
arch/powerpc/lib/sstep.c
3668
truncate_if_32bit(regs->msr, regs->nip + GETLENGTH(op.type)));
arch/powerpc/lib/sstep.c
575
static int do_fp_load(struct instruction_op *op, unsigned long ea,
arch/powerpc/lib/sstep.c
588
nb = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
593
rn = op->reg;
arch/powerpc/lib/sstep.c
604
if (op->type & FPCONV)
arch/powerpc/lib/sstep.c
606
else if (op->type & SIGNEXT)
arch/powerpc/lib/sstep.c
628
static int do_fp_store(struct instruction_op *op, unsigned long ea,
arch/powerpc/lib/sstep.c
640
nb = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
645
rn = op->reg;
arch/powerpc/lib/sstep.c
652
if (op->type & FPCONV)
arch/powerpc/lib/sstep.c
783
static nokprobe_inline void emulate_vsx_load(struct instruction_op *op, union vsx_reg *reg,
arch/powerpc/lib/sstep.c
792
size = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
795
switch (op->element_size) {
arch/powerpc/lib/sstep.c
803
if (IS_LE && (op->vsx_flags & VSX_LDLEFT))
arch/powerpc/lib/sstep.c
816
if (op->type & SIGNEXT) {
arch/powerpc/lib/sstep.c
819
} else if (op->vsx_flags & VSX_FPCONV) {
arch/powerpc/lib/sstep.c
829
} else if (op->vsx_flags & VSX_SPLAT)
arch/powerpc/lib/sstep.c
840
if (op->vsx_flags & VSX_SPLAT) {
arch/powerpc/lib/sstep.c
85
struct instruction_op *op)
arch/powerpc/lib/sstep.c
867
static nokprobe_inline void emulate_vsx_store(struct instruction_op *op, const union vsx_reg *reg,
arch/powerpc/lib/sstep.c
877
size = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
879
switch (op->element_size) {
arch/powerpc/lib/sstep.c
900
if (IS_LE && (op->vsx_flags & VSX_LDLEFT))
arch/powerpc/lib/sstep.c
914
if (size < 8 && op->vsx_flags & VSX_FPCONV) {
arch/powerpc/lib/sstep.c
92
op->type |= DECCTR;
arch/powerpc/lib/sstep.c
957
static nokprobe_inline int do_vsx_load(struct instruction_op *op,
arch/powerpc/lib/sstep.c
961
int reg = op->reg;
arch/powerpc/lib/sstep.c
965
int size = GETSIZE(op->type);
arch/powerpc/lib/sstep.c
971
emulate_vsx_load(op, buf, mem, cross_endian);
arch/powerpc/lib/test_emulate_step.c
1604
struct instruction_op op;
arch/powerpc/lib/test_emulate_step.c
1612
analysed = analyse_instr(&op, regs, instr);
arch/powerpc/lib/test_emulate_step.c
1613
if (analysed != 1 || GETTYPE(op.type) != COMPUTE) {
arch/powerpc/lib/test_emulate_step.c
1622
emulate_update_regs(regs, &op);
arch/powerpc/perf/core-book3s.c
1963
unsigned long type, op, result;
arch/powerpc/perf/core-book3s.c
1971
op = (config >> 8) & 0xff;
arch/powerpc/perf/core-book3s.c
1975
op >= PERF_COUNT_HW_CACHE_OP_MAX ||
arch/powerpc/perf/core-book3s.c
1979
ev = (*ppmu->cache_events)[type][op][result];
arch/powerpc/perf/core-fsl-emb.c
444
unsigned long type, op, result;
arch/powerpc/perf/core-fsl-emb.c
452
op = (config >> 8) & 0xff;
arch/powerpc/perf/core-fsl-emb.c
456
op >= PERF_COUNT_HW_CACHE_OP_MAX ||
arch/powerpc/perf/core-fsl-emb.c
460
ev = (*ppmu->cache_events)[type][op][result];
arch/powerpc/platforms/powernv/opal-flash.c
225
static inline void opal_flash_manage(uint8_t op)
arch/powerpc/platforms/powernv/opal-flash.c
229
args_buf->status = opal_manage_flash(op);
arch/powerpc/platforms/powernv/opal-flash.c
256
uint8_t op;
arch/powerpc/platforms/powernv/opal-flash.c
259
op = FLASH_REJECT_TMP_SIDE;
arch/powerpc/platforms/powernv/opal-flash.c
262
op = FLASH_COMMIT_TMP_SIDE;
arch/powerpc/platforms/powernv/opal-flash.c
269
opal_flash_manage(op);
arch/powerpc/platforms/powernv/opal-flash.c
276
static int opal_flash_update(int op)
arch/powerpc/platforms/powernv/opal-flash.c
282
if (op == FLASH_UPDATE_CANCEL) {
arch/powerpc/platforms/powernv/pci-ioda.c
609
uint8_t op = is_add ? OPAL_ADD_PE_TO_DOMAIN :
arch/powerpc/platforms/powernv/pci-ioda.c
616
child->pe_number, op);
arch/powerpc/platforms/powernv/pci-ioda.c
629
slave->pe_number, op);
arch/powerpc/platforms/ps3/device-init.c
728
const char *op = str_write_read(write);
arch/powerpc/platforms/ps3/device-init.c
740
pr_err("%s:%u: %s failed %d\n", __func__, __LINE__, op, res);
arch/powerpc/platforms/ps3/device-init.c
743
pr_debug("%s:%u: notification %s issued\n", __func__, __LINE__, op);
arch/powerpc/platforms/ps3/device-init.c
752
__LINE__, op, dev->lv1_status);
arch/powerpc/platforms/ps3/device-init.c
755
pr_debug("%s:%u: notification %s completed\n", __func__, __LINE__, op);
arch/powerpc/platforms/pseries/vio.c
1098
int vio_h_cop_sync(struct vio_dev *vdev, struct vio_pfo_op *op)
arch/powerpc/platforms/pseries/vio.c
1105
if (op->timeout)
arch/powerpc/platforms/pseries/vio.c
1106
deadline = jiffies + msecs_to_jiffies(op->timeout);
arch/powerpc/platforms/pseries/vio.c
1109
hret = plpar_hcall_norets(H_COP, op->flags,
arch/powerpc/platforms/pseries/vio.c
1111
op->in, op->inlen, op->out,
arch/powerpc/platforms/pseries/vio.c
1112
op->outlen, op->csbcpb);
arch/powerpc/platforms/pseries/vio.c
1117
(op->timeout && time_after(deadline, jiffies)))
arch/powerpc/platforms/pseries/vio.c
1151
op->hcall_err = hret;
arch/powerpc/sysdev/fsl_rio.c
73
#define ___fsl_read_rio_config(x, addr, err, op, barrier) \
arch/powerpc/sysdev/fsl_rio.c
75
"1: "op" %1,0(%2)\n" \
arch/powerpc/sysdev/fsl_rio.c
88
#define __fsl_read_rio_config(x, addr, err, op) \
arch/powerpc/sysdev/fsl_rio.c
89
___fsl_read_rio_config(x, addr, err, op, "mbar")
arch/powerpc/sysdev/fsl_rio.c
91
#define __fsl_read_rio_config(x, addr, err, op) \
arch/powerpc/sysdev/fsl_rio.c
92
___fsl_read_rio_config(x, addr, err, op, "eieio")
arch/powerpc/sysdev/tsi108_pci.c
120
#define __tsi108_read_pci_config(x, addr, op) \
arch/powerpc/sysdev/tsi108_pci.c
122
" "op" %0,0,%1\n" \
arch/powerpc/xmon/ppc-opc.c
2325
#define A(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 1) | (((unsigned long)(rc)) & 1))
arch/powerpc/xmon/ppc-opc.c
2341
#define B(op, aa, lk) (OP (op) | ((((unsigned long)(aa)) & 1) << 1) | ((lk) & 1))
arch/powerpc/xmon/ppc-opc.c
2345
#define BD8(op, aa, lk) (((((unsigned long)(op)) & 0x3f) << 10) | (((aa) & 1) << 9) | (((lk) & 1) << 8))
arch/powerpc/xmon/ppc-opc.c
2349
#define BD8IO(op) ((((unsigned long)(op)) & 0x1f) << 11)
arch/powerpc/xmon/ppc-opc.c
2353
#define EBD8IO(op, bo, bi) (BD8IO ((op)) | ((bo) << 10) | ((bi) << 8))
arch/powerpc/xmon/ppc-opc.c
2362
#define BD15(op, aa, lk) (OP (op) | ((((unsigned long)(aa)) & 0xf) << 22) | ((lk) & 1))
arch/powerpc/xmon/ppc-opc.c
2366
#define EBD15(op, aa, bo, lk) (((op) & 0x3f) << 26) | (((aa) & 0xf) << 22) | (((bo) & 0x3) << 20) | ((lk) & 1)
arch/powerpc/xmon/ppc-opc.c
2370
#define EBD15BI(op, aa, bo, bi, lk) (((op) & 0x3f) << 26) \
arch/powerpc/xmon/ppc-opc.c
2378
#define BD24(op, aa, lk) (OP (op) | ((((unsigned long)(aa)) & 1) << 25) | ((lk) & 1))
arch/powerpc/xmon/ppc-opc.c
2382
#define BBO(op, bo, aa, lk) (B ((op), (aa), (lk)) | ((((unsigned long)(bo)) & 0x1f) << 21))
arch/powerpc/xmon/ppc-opc.c
2396
#define BBOCB(op, bo, cb, aa, lk) \
arch/powerpc/xmon/ppc-opc.c
2397
(BBO ((op), (bo), (aa), (lk)) | ((((unsigned long)(cb)) & 0x3) << 16))
arch/powerpc/xmon/ppc-opc.c
2416
#define CTX(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x7))
arch/powerpc/xmon/ppc-opc.c
2420
#define UCTX(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1f))
arch/powerpc/xmon/ppc-opc.c
2427
#define DQX(op, xop) (OP (op) | ((xop) & 0x7))
arch/powerpc/xmon/ppc-opc.c
2431
#define DSO(op, xop) (OP (op) | ((xop) & 0x3))
arch/powerpc/xmon/ppc-opc.c
2435
#define DX(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 1))
arch/powerpc/xmon/ppc-opc.c
2439
#define EVSEL(op, xop) (OP (op) | (((unsigned long)(xop)) & 0xff) << 3)
arch/powerpc/xmon/ppc-opc.c
2443
#define IA16(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1f) << 11)
arch/powerpc/xmon/ppc-opc.c
2447
#define I16A(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1f) << 11)
arch/powerpc/xmon/ppc-opc.c
2451
#define I16L(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1f) << 11)
arch/powerpc/xmon/ppc-opc.c
2455
#define IM7(op) ((((unsigned long)(op)) & 0x1f) << 11)
arch/powerpc/xmon/ppc-opc.c
2459
#define M(op, rc) (OP (op) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2463
#define LI20(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1) << 15)
arch/powerpc/xmon/ppc-opc.c
2467
#define MME(op, me, rc) (M ((op), (rc)) | ((((unsigned long)(me)) & 0x1f) << 1))
arch/powerpc/xmon/ppc-opc.c
2476
#define MD(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0x7) << 2) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2486
#define MDS(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0xf) << 1) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2493
#define SC(op, sa, lk) (OP (op) | ((((unsigned long)(sa)) & 1) << 1) | ((lk) & 1))
arch/powerpc/xmon/ppc-opc.c
2497
#define SCI8(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 11))
arch/powerpc/xmon/ppc-opc.c
2501
#define SCI8BF(op, fop, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 11) | (((fop) & 7) << 23))
arch/powerpc/xmon/ppc-opc.c
2505
#define SD4(op) ((((unsigned long)(op)) & 0xf) << 12)
arch/powerpc/xmon/ppc-opc.c
2509
#define SE_IM5(op, xop) (((((unsigned long)(op)) & 0x3f) << 10) | (((xop) & 0x1) << 9))
arch/powerpc/xmon/ppc-opc.c
2513
#define SE_R(op, xop) (((((unsigned long)(op)) & 0x3f) << 10) | (((xop) & 0x3f) << 4))
arch/powerpc/xmon/ppc-opc.c
2517
#define SE_RR(op, xop) (((((unsigned long)(op)) & 0x3f) << 10) | (((xop) & 0x3) << 8))
arch/powerpc/xmon/ppc-opc.c
2521
#define VX(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x7ff))
arch/powerpc/xmon/ppc-opc.c
2554
#define VXA(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x03f))
arch/powerpc/xmon/ppc-opc.c
2563
#define VXR(op, xop, rc) (OP (op) | (((rc) & 1) << 10) | (((unsigned long)(xop)) & 0x3ff))
arch/powerpc/xmon/ppc-opc.c
2569
#define VXVA(op, xop, vaop) (VX(op,xop) | (((vaop) & 0x1f) << 16))
arch/powerpc/xmon/ppc-opc.c
2571
#define VXASH(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 1))
arch/powerpc/xmon/ppc-opc.c
2575
#define X(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x3ff) << 1))
arch/powerpc/xmon/ppc-opc.c
2578
#define XVA(op, xop, vaop) (X(op,xop) | (((vaop) & 0x1f) << 16))
arch/powerpc/xmon/ppc-opc.c
2581
#define EX(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x7ff))
arch/powerpc/xmon/ppc-opc.c
2587
#define XX2(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1ff) << 2))
arch/powerpc/xmon/ppc-opc.c
2590
#define XX2VA(op, xop, vaop) (XX2(op,xop) | (((vaop) & 0x1f) << 16))
arch/powerpc/xmon/ppc-opc.c
2593
#define XX3(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0xff) << 3))
arch/powerpc/xmon/ppc-opc.c
2596
#define XX3RC(op, xop, rc) (OP (op) | (((rc) & 1) << 10) | ((((unsigned long)(xop)) & 0x7f) << 3))
arch/powerpc/xmon/ppc-opc.c
2599
#define XX4(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x3) << 4))
arch/powerpc/xmon/ppc-opc.c
2602
#define Z(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1ff) << 1))
arch/powerpc/xmon/ppc-opc.c
2605
#define XRC(op, xop, rc) (X ((op), (xop)) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2608
#define XVARC(op, xop, vaop, rc) (XVA ((op), (xop), (vaop)) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2611
#define XMMF(op, xop, mop0, mop1) (X ((op), (xop)) | ((mop0) & 3) << 19 | ((mop1) & 7) << 16)
arch/powerpc/xmon/ppc-opc.c
2614
#define ZRC(op, xop, rc) (Z ((op), (xop)) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2723
#define XOPL(op, xop, l) (X ((op), (xop)) | ((((unsigned long)(l)) & 1) << 21))
arch/powerpc/xmon/ppc-opc.c
2726
#define XOPL2(op, xop, l) (X ((op), (xop)) | ((((unsigned long)(l)) & 3) << 21))
arch/powerpc/xmon/ppc-opc.c
2729
#define XRCL(op, xop, l, rc) (XRC ((op), (xop), (rc)) | ((((unsigned long)(l)) & 1) << 21))
arch/powerpc/xmon/ppc-opc.c
2732
#define XRT(op, xop, rt) (X ((op), (xop)) \
arch/powerpc/xmon/ppc-opc.c
2736
#define XRTRA(op, xop, rt, ra) (X ((op), (xop)) \
arch/powerpc/xmon/ppc-opc.c
2748
#define XTO(op, xop, to) (X ((op), (xop)) | ((((unsigned long)(to)) & 0x1f) << 21))
arch/powerpc/xmon/ppc-opc.c
2752
#define XTLB(op, xop, sh) (X ((op), (xop)) | ((((unsigned long)(sh)) & 0x1f) << 11))
arch/powerpc/xmon/ppc-opc.c
2756
#define XSYNC(op, xop, l) (X ((op), (xop)) | ((((unsigned long)(l)) & 3) << 21))
arch/powerpc/xmon/ppc-opc.c
2768
#define XDSS(op, xop, a) (X ((op), (xop)) | ((((unsigned long)(a)) & 1) << 25))
arch/powerpc/xmon/ppc-opc.c
2772
#define XFL(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0x3ff) << 1) | (((unsigned long)(rc)) & 1))
arch/powerpc/xmon/ppc-opc.c
2776
#define XISEL(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x1f) << 1))
arch/powerpc/xmon/ppc-opc.c
2780
#define XL(op, xop) (OP (op) | ((((unsigned long)(xop)) & 0x3ff) << 1))
arch/powerpc/xmon/ppc-opc.c
2783
#define XLLK(op, xop, lk) (XL ((op), (xop)) | ((lk) & 1))
arch/powerpc/xmon/ppc-opc.c
2792
#define XLO(op, bo, xop, lk) \
arch/powerpc/xmon/ppc-opc.c
2793
(XLLK ((op), (xop), (lk)) | ((((unsigned long)(bo)) & 0x1f) << 21))
arch/powerpc/xmon/ppc-opc.c
2798
#define XLYLK(op, xop, y, lk) (XLLK ((op), (xop), (lk)) | ((((unsigned long)(y)) & 1) << 21))
arch/powerpc/xmon/ppc-opc.c
2803
#define XLOCB(op, bo, cb, xop, lk) \
arch/powerpc/xmon/ppc-opc.c
2804
(XLO ((op), (bo), (xop), (lk)) | ((((unsigned long)(cb)) & 3) << 16))
arch/powerpc/xmon/ppc-opc.c
2822
#define XMBAR(op, xop, mo) (X ((op), (xop)) | ((((unsigned long)(mo)) & 1) << 21))
arch/powerpc/xmon/ppc-opc.c
2825
#define XO(op, xop, oe, rc) \
arch/powerpc/xmon/ppc-opc.c
2826
(OP (op) | ((((unsigned long)(xop)) & 0x1ff) << 1) | ((((unsigned long)(oe)) & 1) << 10) | (((unsigned long)(rc)) & 1))
arch/powerpc/xmon/ppc-opc.c
2833
#define XOPS(op, xop, rc) \
arch/powerpc/xmon/ppc-opc.c
2834
(OP (op) | ((((unsigned long)(xop)) & 0x3ff) << 1) | (((unsigned long)(rc)) & 1))
arch/powerpc/xmon/ppc-opc.c
2839
#define XS(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0x1ff) << 2) | (((unsigned long)(rc)) & 1))
arch/powerpc/xmon/ppc-opc.c
2846
#define XFXM(op, xop, fxm, p4) \
arch/powerpc/xmon/ppc-opc.c
2847
(X ((op), (xop)) | ((((unsigned long)(fxm)) & 0xff) << 12) \
arch/powerpc/xmon/ppc-opc.c
2851
#define XSPR(op, xop, spr) \
arch/powerpc/xmon/ppc-opc.c
2852
(X ((op), (xop)) | ((((unsigned long)(spr)) & 0x1f) << 16) | ((((unsigned long)(spr)) & 0x3e0) << 6))
arch/powerpc/xmon/ppc-opc.c
2867
#define XUC(op, xop) (OP (op) | (((unsigned long)(xop)) & 0x1f))
arch/powerpc/xmon/ppc-opc.c
2871
#define XW(op, xop, rc) (OP (op) | ((((unsigned long)(xop)) & 0x3f) << 1) | ((rc) & 1))
arch/powerpc/xmon/ppc-opc.c
2876
#define APU(op, xop, rc) (OP (op) | (((unsigned long)(xop)) & 0x3ff) << 1 | ((rc) & 1))
arch/powerpc/xmon/ppc.h
263
(unsigned long instruction, long op, ppc_cpu_t dialect, const char **errmsg);
arch/riscv/include/asm/atomic.h
101
c_type arch_atomic##prefix##_fetch_##op(c_type i, atomic##prefix##_t *v) \
arch/riscv/include/asm/atomic.h
112
#define ATOMIC_OP_RETURN(op, asm_op, c_op, I, asm_type, c_type, prefix) \
arch/riscv/include/asm/atomic.h
114
c_type arch_atomic##prefix##_##op##_return_relaxed(c_type i, \
arch/riscv/include/asm/atomic.h
117
return arch_atomic##prefix##_fetch_##op##_relaxed(i, v) c_op I; \
arch/riscv/include/asm/atomic.h
120
c_type arch_atomic##prefix##_##op##_return(c_type i, atomic##prefix##_t *v) \
arch/riscv/include/asm/atomic.h
122
return arch_atomic##prefix##_fetch_##op(i, v) c_op I; \
arch/riscv/include/asm/atomic.h
126
#define ATOMIC_OPS(op, asm_op, c_op, I) \
arch/riscv/include/asm/atomic.h
127
ATOMIC_FETCH_OP( op, asm_op, I, w, int, ) \
arch/riscv/include/asm/atomic.h
128
ATOMIC_OP_RETURN(op, asm_op, c_op, I, w, int, )
arch/riscv/include/asm/atomic.h
130
#define ATOMIC_OPS(op, asm_op, c_op, I) \
arch/riscv/include/asm/atomic.h
131
ATOMIC_FETCH_OP( op, asm_op, I, w, int, ) \
arch/riscv/include/asm/atomic.h
132
ATOMIC_OP_RETURN(op, asm_op, c_op, I, w, int, ) \
arch/riscv/include/asm/atomic.h
133
ATOMIC_FETCH_OP( op, asm_op, I, d, s64, 64) \
arch/riscv/include/asm/atomic.h
134
ATOMIC_OP_RETURN(op, asm_op, c_op, I, d, s64, 64)
arch/riscv/include/asm/atomic.h
165
#define ATOMIC_OPS(op, asm_op, I) \
arch/riscv/include/asm/atomic.h
166
ATOMIC_FETCH_OP(op, asm_op, I, w, int, )
arch/riscv/include/asm/atomic.h
168
#define ATOMIC_OPS(op, asm_op, I) \
arch/riscv/include/asm/atomic.h
169
ATOMIC_FETCH_OP(op, asm_op, I, w, int, ) \
arch/riscv/include/asm/atomic.h
170
ATOMIC_FETCH_OP(op, asm_op, I, d, s64, 64)
arch/riscv/include/asm/atomic.h
53
#define ATOMIC_OP(op, asm_op, I, asm_type, c_type, prefix) \
arch/riscv/include/asm/atomic.h
55
void arch_atomic##prefix##_##op(c_type i, atomic##prefix##_t *v) \
arch/riscv/include/asm/atomic.h
65
#define ATOMIC_OPS(op, asm_op, I) \
arch/riscv/include/asm/atomic.h
66
ATOMIC_OP (op, asm_op, I, w, int, )
arch/riscv/include/asm/atomic.h
68
#define ATOMIC_OPS(op, asm_op, I) \
arch/riscv/include/asm/atomic.h
69
ATOMIC_OP (op, asm_op, I, w, int, ) \
arch/riscv/include/asm/atomic.h
70
ATOMIC_OP (op, asm_op, I, d, s64, 64)
arch/riscv/include/asm/atomic.h
87
#define ATOMIC_FETCH_OP(op, asm_op, I, asm_type, c_type, prefix) \
arch/riscv/include/asm/atomic.h
89
c_type arch_atomic##prefix##_fetch_##op##_relaxed(c_type i, \
arch/riscv/include/asm/bitops.h
172
#define __AMO(op) "amo" #op ".d"
arch/riscv/include/asm/bitops.h
174
#define __AMO(op) "amo" #op ".w"
arch/riscv/include/asm/bitops.h
179
#define __test_and_op_bit_ord(op, mod, nr, addr, ord) \
arch/riscv/include/asm/bitops.h
184
__AMO(op) #ord " %0, %2, %1" \
arch/riscv/include/asm/bitops.h
191
#define __op_bit_ord(op, mod, nr, addr, ord) \
arch/riscv/include/asm/bitops.h
193
__AMO(op) #ord " zero, %1, %0" \
arch/riscv/include/asm/bitops.h
198
#define __test_and_op_bit(op, mod, nr, addr) \
arch/riscv/include/asm/bitops.h
199
__test_and_op_bit_ord(op, mod, nr, addr, .aqrl)
arch/riscv/include/asm/bitops.h
200
#define __op_bit(op, mod, nr, addr) \
arch/riscv/include/asm/bitops.h
201
__op_bit_ord(op, mod, nr, addr, )
arch/riscv/include/asm/ftrace.h
226
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/riscv/include/asm/futex.h
31
: [op] "Jr" (oparg) \
arch/riscv/include/asm/futex.h
37
arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/riscv/include/asm/futex.h
44
switch (op) {
arch/riscv/include/asm/kvm_gstage.h
63
pte_t *ptep, u32 ptep_level, enum kvm_riscv_gstage_op op);
arch/riscv/include/asm/spinlock.h
23
#define SPINLOCK_BASE_DECLARE(op, type, type_lock) \
arch/riscv/include/asm/spinlock.h
24
static __always_inline type arch_spin_##op(type_lock lock) \
arch/riscv/include/asm/spinlock.h
27
return queued_spin_##op(lock); \
arch/riscv/include/asm/spinlock.h
28
return ticket_spin_##op(lock); \
arch/riscv/kernel/ftrace.c
251
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/riscv/kernel/smp.c
115
static void send_ipi_mask(const struct cpumask *mask, enum ipi_message_type op)
arch/riscv/kernel/smp.c
117
__ipi_send_mask(ipi_desc[op], mask);
arch/riscv/kernel/smp.c
120
static void send_ipi_single(int cpu, enum ipi_message_type op)
arch/riscv/kernel/smp.c
122
__ipi_send_mask(ipi_desc[op], cpumask_of(cpu));
arch/riscv/kvm/gstage.c
213
pte_t *ptep, u32 ptep_level, enum kvm_riscv_gstage_op op)
arch/riscv/kvm/gstage.c
236
if (op == GSTAGE_OP_CLEAR)
arch/riscv/kvm/gstage.c
240
&next_ptep[i], next_ptep_level, op);
arch/riscv/kvm/gstage.c
241
if (op == GSTAGE_OP_CLEAR)
arch/riscv/kvm/gstage.c
245
if (op == GSTAGE_OP_CLEAR)
arch/riscv/kvm/gstage.c
247
else if (op == GSTAGE_OP_WP)
arch/riscv/net/bpf_jit.h
292
static inline u16 rv_cr_insn(u8 funct4, u8 rd, u8 rs2, u8 op)
arch/riscv/net/bpf_jit.h
294
return (funct4 << 12) | (rd << 7) | (rs2 << 2) | op;
arch/riscv/net/bpf_jit.h
297
static inline u16 rv_ci_insn(u8 funct3, u32 imm6, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
302
return (funct3 << 13) | (rd << 7) | op | imm;
arch/riscv/net/bpf_jit.h
305
static inline u16 rv_css_insn(u8 funct3, u32 uimm, u8 rs2, u8 op)
arch/riscv/net/bpf_jit.h
307
return (funct3 << 13) | (uimm << 7) | (rs2 << 2) | op;
arch/riscv/net/bpf_jit.h
310
static inline u16 rv_ciw_insn(u8 funct3, u32 uimm, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
312
return (funct3 << 13) | (uimm << 5) | ((rd & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
316
u8 op)
arch/riscv/net/bpf_jit.h
319
(imm_lo << 5) | ((rd & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
323
u8 op)
arch/riscv/net/bpf_jit.h
326
(imm_lo << 5) | ((rs2 & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
329
static inline u16 rv_ca_insn(u8 funct6, u8 rd, u8 funct2, u8 rs2, u8 op)
arch/riscv/net/bpf_jit.h
332
((rs2 & 0x7) << 2) | op;
arch/riscv/net/bpf_jit.h
335
static inline u16 rv_cb_insn(u8 funct3, u32 imm6, u8 funct2, u8 rd, u8 op)
arch/riscv/net/bpf_jit.h
340
return (funct3 << 13) | (funct2 << 10) | ((rd & 0x7) << 7) | op | imm;
arch/riscv/net/bpf_jit_comp32.c
244
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
249
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
328
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
333
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
407
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
414
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
512
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
519
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
565
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
585
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
652
static int emit_bcc(u8 op, u8 rd, u8 rs, int rvoff, struct rv_jit_context *ctx)
arch/riscv/net/bpf_jit_comp32.c
658
if (op == BPF_JSET) {
arch/riscv/net/bpf_jit_comp32.c
665
op = invert_bpf_cond(op);
arch/riscv/net/bpf_jit_comp32.c
676
switch (op) {
arch/riscv/net/bpf_jit_comp32.c
723
struct rv_jit_context *ctx, const u8 op)
arch/riscv/net/bpf_jit_comp32.c
736
if (emit_bcc(op, lo(rs1), lo(rs2), rvoff, ctx))
arch/s390/crypto/paes_s390.c
1167
.op = {
arch/s390/crypto/paes_s390.c
1609
.op = {
arch/s390/crypto/paes_s390.c
591
.op = {
arch/s390/crypto/paes_s390.c
858
.op = {
arch/s390/crypto/phmac_s390.c
967
.op = { \
arch/s390/include/asm/atomic.h
203
#define ATOMIC64_OPS(op) \
arch/s390/include/asm/atomic.h
204
static __always_inline void arch_atomic64_##op(s64 i, atomic64_t *v) \
arch/s390/include/asm/atomic.h
206
__atomic64_##op(i, (long *)&v->counter); \
arch/s390/include/asm/atomic.h
208
static __always_inline long arch_atomic64_fetch_##op(s64 i, atomic64_t *v) \
arch/s390/include/asm/atomic.h
210
return __atomic64_##op##_barrier(i, (long *)&v->counter); \
arch/s390/include/asm/atomic.h
82
#define ATOMIC_OPS(op) \
arch/s390/include/asm/atomic.h
83
static __always_inline void arch_atomic_##op(int i, atomic_t *v) \
arch/s390/include/asm/atomic.h
85
__atomic_##op(i, &v->counter); \
arch/s390/include/asm/atomic.h
87
static __always_inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/s390/include/asm/atomic.h
89
return __atomic_##op##_barrier(i, &v->counter); \
arch/s390/include/asm/cio.h
377
int chsc_sstpc(void *page, unsigned int op, u16 ctrl, long *clock_delta);
arch/s390/include/asm/ftrace.h
117
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/s390/include/asm/futex.h
50
int arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr)
arch/s390/include/asm/futex.h
54
switch (op) {
arch/s390/include/asm/itcw.h
22
struct itcw *itcw_init(void *buffer, size_t size, int op, int intrg,
arch/s390/include/asm/percpu.h
104
#define arch_this_cpu_to_op(pcp, val, op) \
arch/s390/include/asm/percpu.h
112
op " %[old__],%[val__],%[ptr__]" \
arch/s390/include/asm/percpu.h
19
#define arch_this_cpu_to_op_simple(pcp, val, op) \
arch/s390/include/asm/percpu.h
29
new__ = old__ op (val); \
arch/s390/include/asm/percpu.h
85
#define arch_this_cpu_add_return(pcp, val, op) \
arch/s390/include/asm/percpu.h
93
op " %[old__],%[val__],%[ptr__]" \
arch/s390/include/uapi/asm/kvm.h
306
__u8 op; /* operation to perform */
arch/s390/include/uapi/asm/kvm.h
65
__u32 op; /* type of operation */
arch/s390/kernel/ftrace.c
202
u8 op;
arch/s390/kernel/ftrace.c
209
op = enable ? 0xf4 : 0x04;
arch/s390/kernel/ftrace.c
210
s390_kernel_write((char *)addr + 1, &op, sizeof(op));
arch/s390/kernel/ftrace.c
267
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/s390/kvm/kvm-s390.c
2762
acc_mode = mop->op == KVM_S390_MEMOP_ABSOLUTE_READ ? GACC_FETCH : GACC_STORE;
arch/s390/kvm/kvm-s390.c
2828
switch (mop->op) {
arch/s390/kvm/kvm-s390.c
5241
switch (mop->op) {
arch/s390/kvm/kvm-s390.c
5278
acc_mode = mop->op == KVM_S390_MEMOP_LOGICAL_READ ? GACC_FETCH : GACC_STORE;
arch/s390/kvm/kvm-s390.c
5307
switch (mop->op) {
arch/s390/kvm/kvm-s390.c
557
static int kvm_s390_keyop(struct kvm_s390_mmu_cache *mc, struct kvm *kvm, int op,
arch/s390/kvm/kvm-s390.c
566
switch (op) {
arch/s390/kvm/pci.c
638
switch (args->op) {
arch/s390/net/bpf_jit_comp.c
148
#define _EMIT2(op) \
arch/s390/net/bpf_jit_comp.c
151
*(u16 *) (jit->prg_buf + jit->prg) = (op); \
arch/s390/net/bpf_jit_comp.c
155
#define EMIT2(op, b1, b2) \
arch/s390/net/bpf_jit_comp.c
157
_EMIT2((op) | reg(b1, b2)); \
arch/s390/net/bpf_jit_comp.c
162
#define _EMIT4(op) \
arch/s390/net/bpf_jit_comp.c
165
*(u32 *) (jit->prg_buf + jit->prg) = (op); \
arch/s390/net/bpf_jit_comp.c
169
#define EMIT4(op, b1, b2) \
arch/s390/net/bpf_jit_comp.c
171
_EMIT4((op) | reg(b1, b2)); \
arch/s390/net/bpf_jit_comp.c
176
#define EMIT4_RRF(op, b1, b2, b3) \
arch/s390/net/bpf_jit_comp.c
178
_EMIT4((op) | reg_high(b3) << 8 | reg(b1, b2)); \
arch/s390/net/bpf_jit_comp.c
184
#define _EMIT4_DISP(op, disp) \
arch/s390/net/bpf_jit_comp.c
187
_EMIT4((op) | __disp); \
arch/s390/net/bpf_jit_comp.c
190
#define EMIT4_DISP(op, b1, b2, disp) \
arch/s390/net/bpf_jit_comp.c
192
_EMIT4_DISP((op) | reg_high(b1) << 16 | \
arch/s390/net/bpf_jit_comp.c
198
#define EMIT4_IMM(op, b1, imm) \
arch/s390/net/bpf_jit_comp.c
201
_EMIT4((op) | reg_high(b1) << 16 | __imm); \
arch/s390/net/bpf_jit_comp.c
205
#define EMIT4_PCREL(op, pcrel) \
arch/s390/net/bpf_jit_comp.c
208
_EMIT4((op) | __pcrel); \
arch/s390/net/bpf_jit_comp.c
211
#define EMIT4_PCREL_RIC(op, mask, target) \
arch/s390/net/bpf_jit_comp.c
214
_EMIT4((op) | (mask) << 20 | (__rel & 0xffff)); \
arch/s390/net/bpf_jit_comp.c
275
static void emit6_pcrel_ril(struct bpf_jit *jit, u32 op, s64 pcrel)
arch/s390/net/bpf_jit_comp.c
279
_EMIT6(op | pc32dbl >> 16, pc32dbl & 0xffff);
arch/s390/net/bpf_jit_comp.c
282
static void emit6_pcrel_rilb(struct bpf_jit *jit, u32 op, u8 b, s64 pcrel)
arch/s390/net/bpf_jit_comp.c
284
emit6_pcrel_ril(jit, op | reg_high(b) << 16, pcrel);
arch/s390/net/bpf_jit_comp.c
288
#define EMIT6_PCREL_RILB(op, b, target) \
arch/s390/net/bpf_jit_comp.c
289
emit6_pcrel_rilb(jit, op, b, off_to_pcrel(jit, target))
arch/s390/net/bpf_jit_comp.c
291
#define EMIT6_PCREL_RILB_PTR(op, b, target_ptr) \
arch/s390/net/bpf_jit_comp.c
292
emit6_pcrel_rilb(jit, op, b, ptr_to_pcrel(jit, target_ptr))
arch/s390/net/bpf_jit_comp.c
294
static void emit6_pcrel_rilc(struct bpf_jit *jit, u32 op, u8 mask, s64 pcrel)
arch/s390/net/bpf_jit_comp.c
296
emit6_pcrel_ril(jit, op | mask << 20, pcrel);
arch/s390/net/bpf_jit_comp.c
299
#define EMIT6_PCREL_RILC(op, mask, target) \
arch/s390/net/bpf_jit_comp.c
300
emit6_pcrel_rilc(jit, op, mask, off_to_pcrel(jit, target))
arch/s390/net/bpf_jit_comp.c
302
#define EMIT6_PCREL_RILC_PTR(op, mask, target_ptr) \
arch/s390/net/bpf_jit_comp.c
303
emit6_pcrel_rilc(jit, op, mask, ptr_to_pcrel(jit, target_ptr))
arch/s390/net/bpf_jit_comp.c
305
#define _EMIT6_IMM(op, imm) \
arch/s390/net/bpf_jit_comp.c
308
_EMIT6((op) | (__imm >> 16), __imm & 0xffff); \
arch/s390/net/bpf_jit_comp.c
311
#define EMIT6_IMM(op, b1, imm) \
arch/s390/net/bpf_jit_comp.c
313
_EMIT6_IMM((op) | reg_high(b1) << 16, imm); \
arch/s390/net/bpf_jit_comp.c
505
static void save_restore_regs(struct bpf_jit *jit, int op, u16 extra_regs)
arch/s390/net/bpf_jit_comp.c
525
if (op == REGS_SAVE)
arch/sh/include/asm/atomic-grb.h
16
" " #op " %2, %0 \n\t" /* $op */ \
arch/sh/include/asm/atomic-grb.h
25
#define ATOMIC_OP_RETURN(op) \
arch/sh/include/asm/atomic-grb.h
26
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/sh/include/asm/atomic-grb.h
36
" " #op " %2, %0 \n\t" /* $op */ \
arch/sh/include/asm/atomic-grb.h
47
#define ATOMIC_FETCH_OP(op) \
arch/sh/include/asm/atomic-grb.h
48
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-grb.h
5
#define ATOMIC_OP(op) \
arch/sh/include/asm/atomic-grb.h
59
" " #op " %3, %0 \n\t" /* $op */ \
arch/sh/include/asm/atomic-grb.h
6
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-grb.h
69
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) ATOMIC_FETCH_OP(op)
arch/sh/include/asm/atomic-grb.h
80
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op)
arch/sh/include/asm/atomic-irq.h
13
#define ATOMIC_OP(op, c_op) \
arch/sh/include/asm/atomic-irq.h
14
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-irq.h
23
#define ATOMIC_OP_RETURN(op, c_op) \
arch/sh/include/asm/atomic-irq.h
24
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/sh/include/asm/atomic-irq.h
37
#define ATOMIC_FETCH_OP(op, c_op) \
arch/sh/include/asm/atomic-irq.h
38
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-irq.h
50
#define ATOMIC_OPS(op, c_op) \
arch/sh/include/asm/atomic-irq.h
51
ATOMIC_OP(op, c_op) \
arch/sh/include/asm/atomic-irq.h
52
ATOMIC_OP_RETURN(op, c_op) \
arch/sh/include/asm/atomic-irq.h
53
ATOMIC_FETCH_OP(op, c_op)
arch/sh/include/asm/atomic-irq.h
64
#define ATOMIC_OPS(op, c_op) \
arch/sh/include/asm/atomic-irq.h
65
ATOMIC_OP(op, c_op) \
arch/sh/include/asm/atomic-irq.h
66
ATOMIC_FETCH_OP(op, c_op)
arch/sh/include/asm/atomic-llsc.h
19
#define ATOMIC_OP(op) \
arch/sh/include/asm/atomic-llsc.h
20
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-llsc.h
25
"1: movli.l @%2, %0 ! atomic_" #op "\n" \
arch/sh/include/asm/atomic-llsc.h
26
" " #op " %1, %0 \n" \
arch/sh/include/asm/atomic-llsc.h
34
#define ATOMIC_OP_RETURN(op) \
arch/sh/include/asm/atomic-llsc.h
35
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/sh/include/asm/atomic-llsc.h
40
"1: movli.l @%2, %0 ! atomic_" #op "_return \n" \
arch/sh/include/asm/atomic-llsc.h
41
" " #op " %1, %0 \n" \
arch/sh/include/asm/atomic-llsc.h
52
#define ATOMIC_FETCH_OP(op) \
arch/sh/include/asm/atomic-llsc.h
53
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/sh/include/asm/atomic-llsc.h
58
"1: movli.l @%3, %0 ! atomic_fetch_" #op " \n" \
arch/sh/include/asm/atomic-llsc.h
60
" " #op " %2, %0 \n" \
arch/sh/include/asm/atomic-llsc.h
71
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) ATOMIC_FETCH_OP(op)
arch/sh/include/asm/atomic-llsc.h
82
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op)
arch/sh/include/asm/futex.h
29
static inline int arch_futex_atomic_op_inuser(int op, u32 oparg, int *oval,
arch/sh/include/asm/futex.h
40
switch (op) {
arch/sh/include/asm/pgtable_32.h
351
#define PTE_BIT_FUNC(h,fn,op) \
arch/sh/include/asm/pgtable_32.h
352
static inline pte_t pte_##fn(pte_t pte) { pte.pte_##h op; return pte; }
arch/sh/include/asm/syscalls.h
11
asmlinkage int sys_cacheflush(unsigned long addr, unsigned long len, int op);
arch/sh/kernel/disassemble.c
296
struct sh_opcode_info *op = sh_table;
arch/sh/kernel/disassemble.c
298
for (; op->name; op++) {
arch/sh/kernel/disassemble.c
308
int i = op->nibbles[n];
arch/sh/kernel/disassemble.c
379
pr_cont("%-8s ", op->name);
arch/sh/kernel/disassemble.c
380
lastsp = (op->arg[0] == A_END);
arch/sh/kernel/disassemble.c
382
for (n = 0; n < 6 && op->arg[n] != A_END; n++) {
arch/sh/kernel/disassemble.c
383
if (n && op->arg[1] != A_END)
arch/sh/kernel/disassemble.c
385
switch (op->arg[n]) {
arch/sh/kernel/disassemble.c
527
if (disp_pc && strcmp(op->name, "mova") != 0) {
arch/sh/kernel/kgdb.c
100
+ linux_regs->regs[OPCODE_BSRF_REG(op)];
arch/sh/kernel/kgdb.c
103
else if (OPCODE_JMP(op))
arch/sh/kernel/kgdb.c
104
addr = linux_regs->regs[OPCODE_JMP_REG(op)];
arch/sh/kernel/kgdb.c
107
else if (OPCODE_JSR(op))
arch/sh/kernel/kgdb.c
108
addr = linux_regs->regs[OPCODE_JSR_REG(op)];
arch/sh/kernel/kgdb.c
111
else if (OPCODE_RTS(op))
arch/sh/kernel/kgdb.c
115
else if (OPCODE_RTE(op))
arch/sh/kernel/kgdb.c
120
addr = linux_regs->pc + instruction_size(op);
arch/sh/kernel/kgdb.c
122
flush_icache_range(addr, addr + instruction_size(op));
arch/sh/kernel/kgdb.c
20
#define OPCODE_BT(op) (((op) & 0xff00) == 0x8900)
arch/sh/kernel/kgdb.c
21
#define OPCODE_BF(op) (((op) & 0xff00) == 0x8b00)
arch/sh/kernel/kgdb.c
22
#define OPCODE_BTF_DISP(op) (((op) & 0x80) ? (((op) | 0xffffff80) << 1) : \
arch/sh/kernel/kgdb.c
23
(((op) & 0x7f ) << 1))
arch/sh/kernel/kgdb.c
24
#define OPCODE_BFS(op) (((op) & 0xff00) == 0x8f00)
arch/sh/kernel/kgdb.c
25
#define OPCODE_BTS(op) (((op) & 0xff00) == 0x8d00)
arch/sh/kernel/kgdb.c
26
#define OPCODE_BRA(op) (((op) & 0xf000) == 0xa000)
arch/sh/kernel/kgdb.c
27
#define OPCODE_BRA_DISP(op) (((op) & 0x800) ? (((op) | 0xfffff800) << 1) : \
arch/sh/kernel/kgdb.c
28
(((op) & 0x7ff) << 1))
arch/sh/kernel/kgdb.c
29
#define OPCODE_BRAF(op) (((op) & 0xf0ff) == 0x0023)
arch/sh/kernel/kgdb.c
30
#define OPCODE_BRAF_REG(op) (((op) & 0x0f00) >> 8)
arch/sh/kernel/kgdb.c
31
#define OPCODE_BSR(op) (((op) & 0xf000) == 0xb000)
arch/sh/kernel/kgdb.c
32
#define OPCODE_BSR_DISP(op) (((op) & 0x800) ? (((op) | 0xfffff800) << 1) : \
arch/sh/kernel/kgdb.c
33
(((op) & 0x7ff) << 1))
arch/sh/kernel/kgdb.c
34
#define OPCODE_BSRF(op) (((op) & 0xf0ff) == 0x0003)
arch/sh/kernel/kgdb.c
35
#define OPCODE_BSRF_REG(op) (((op) >> 8) & 0xf)
arch/sh/kernel/kgdb.c
36
#define OPCODE_JMP(op) (((op) & 0xf0ff) == 0x402b)
arch/sh/kernel/kgdb.c
37
#define OPCODE_JMP_REG(op) (((op) >> 8) & 0xf)
arch/sh/kernel/kgdb.c
38
#define OPCODE_JSR(op) (((op) & 0xf0ff) == 0x400b)
arch/sh/kernel/kgdb.c
39
#define OPCODE_JSR_REG(op) (((op) >> 8) & 0xf)
arch/sh/kernel/kgdb.c
40
#define OPCODE_RTS(op) ((op) == 0xb)
arch/sh/kernel/kgdb.c
41
#define OPCODE_RTE(op) ((op) == 0x2b)
arch/sh/kernel/kgdb.c
49
insn_size_t op = __raw_readw(linux_regs->pc);
arch/sh/kernel/kgdb.c
53
if (OPCODE_BT(op)) {
arch/sh/kernel/kgdb.c
55
addr = linux_regs->pc + 4 + OPCODE_BTF_DISP(op);
arch/sh/kernel/kgdb.c
61
else if (OPCODE_BTS(op)) {
arch/sh/kernel/kgdb.c
63
addr = linux_regs->pc + 4 + OPCODE_BTF_DISP(op);
arch/sh/kernel/kgdb.c
69
else if (OPCODE_BF(op)) {
arch/sh/kernel/kgdb.c
71
addr = linux_regs->pc + 4 + OPCODE_BTF_DISP(op);
arch/sh/kernel/kgdb.c
77
else if (OPCODE_BFS(op)) {
arch/sh/kernel/kgdb.c
79
addr = linux_regs->pc + 4 + OPCODE_BTF_DISP(op);
arch/sh/kernel/kgdb.c
85
else if (OPCODE_BRA(op))
arch/sh/kernel/kgdb.c
86
addr = linux_regs->pc + 4 + OPCODE_BRA_DISP(op);
arch/sh/kernel/kgdb.c
89
else if (OPCODE_BRAF(op))
arch/sh/kernel/kgdb.c
91
+ linux_regs->regs[OPCODE_BRAF_REG(op)];
arch/sh/kernel/kgdb.c
94
else if (OPCODE_BSR(op))
arch/sh/kernel/kgdb.c
95
addr = linux_regs->pc + 4 + OPCODE_BSR_DISP(op);
arch/sh/kernel/kgdb.c
98
else if (OPCODE_BSRF(op))
arch/sh/kernel/perf_event.c
75
unsigned long type, op, result;
arch/sh/kernel/perf_event.c
83
op = (config >> 8) & 0xff;
arch/sh/kernel/perf_event.c
87
op >= PERF_COUNT_HW_CACHE_OP_MAX ||
arch/sh/kernel/perf_event.c
91
ev = (*sh_pmu->cache_events)[type][op][result];
arch/sh/kernel/sys_sh.c
58
asmlinkage int sys_cacheflush(unsigned long addr, unsigned long len, int op)
arch/sh/kernel/sys_sh.c
62
if ((op <= 0) || (op > (CACHEFLUSH_D_PURGE|CACHEFLUSH_I)))
arch/sh/kernel/sys_sh.c
79
switch (op & CACHEFLUSH_D_PURGE) {
arch/sh/kernel/sys_sh.c
91
if (op & CACHEFLUSH_I)
arch/sh/math-emu/math.c
67
#define BOTH_PRmn(op,x) \
arch/sh/math-emu/math.c
68
FP_DECL_EX; if(FPSCR_PR) op(D,x,DRm,DRn); else op(S,x,FRm,FRn);
arch/sparc/crypto/camellia_glue.c
111
op(walk.src.virt.addr, walk.dst.virt.addr,
arch/sparc/crypto/camellia_glue.c
142
cbc_crypt_op *op;
arch/sparc/crypto/camellia_glue.c
147
op = camellia_sparc64_cbc_encrypt_3_grand_rounds;
arch/sparc/crypto/camellia_glue.c
149
op = camellia_sparc64_cbc_encrypt_4_grand_rounds;
arch/sparc/crypto/camellia_glue.c
158
op(walk.src.virt.addr, walk.dst.virt.addr,
arch/sparc/crypto/camellia_glue.c
171
cbc_crypt_op *op;
arch/sparc/crypto/camellia_glue.c
176
op = camellia_sparc64_cbc_decrypt_3_grand_rounds;
arch/sparc/crypto/camellia_glue.c
178
op = camellia_sparc64_cbc_decrypt_4_grand_rounds;
arch/sparc/crypto/camellia_glue.c
187
op(walk.src.virt.addr, walk.dst.virt.addr,
arch/sparc/crypto/camellia_glue.c
92
ecb_crypt_op *op;
arch/sparc/crypto/camellia_glue.c
97
op = camellia_sparc64_ecb_crypt_3_grand_rounds;
arch/sparc/crypto/camellia_glue.c
99
op = camellia_sparc64_ecb_crypt_4_grand_rounds;
arch/sparc/include/asm/atomic_64.h
23
#define ATOMIC_OP(op) \
arch/sparc/include/asm/atomic_64.h
24
void arch_atomic_##op(int, atomic_t *); \
arch/sparc/include/asm/atomic_64.h
25
void arch_atomic64_##op(s64, atomic64_t *);
arch/sparc/include/asm/atomic_64.h
27
#define ATOMIC_OP_RETURN(op) \
arch/sparc/include/asm/atomic_64.h
28
int arch_atomic_##op##_return(int, atomic_t *); \
arch/sparc/include/asm/atomic_64.h
29
s64 arch_atomic64_##op##_return(s64, atomic64_t *);
arch/sparc/include/asm/atomic_64.h
31
#define ATOMIC_FETCH_OP(op) \
arch/sparc/include/asm/atomic_64.h
32
int arch_atomic_fetch_##op(int, atomic_t *); \
arch/sparc/include/asm/atomic_64.h
33
s64 arch_atomic64_fetch_##op(s64, atomic64_t *);
arch/sparc/include/asm/atomic_64.h
35
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_OP_RETURN(op) ATOMIC_FETCH_OP(op)
arch/sparc/include/asm/atomic_64.h
51
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op)
arch/sparc/include/asm/device.h
17
struct platform_device *op;
arch/sparc/include/asm/floppy_32.h
274
struct platform_device *op;
arch/sparc/include/asm/floppy_32.h
321
op = NULL;
arch/sparc/include/asm/floppy_32.h
323
op = of_find_device_by_node(dp);
arch/sparc/include/asm/floppy_32.h
324
if (op)
arch/sparc/include/asm/floppy_32.h
327
if (!op) {
arch/sparc/include/asm/floppy_32.h
329
op = of_find_device_by_node(dp);
arch/sparc/include/asm/floppy_32.h
330
if (op)
arch/sparc/include/asm/floppy_32.h
334
if (!op)
arch/sparc/include/asm/floppy_32.h
337
FLOPPY_IRQ = op->archdata.irqs[0];
arch/sparc/include/asm/floppy_64.h
551
struct platform_device *op;
arch/sparc/include/asm/floppy_64.h
559
op = NULL;
arch/sparc/include/asm/floppy_64.h
564
op = of_find_device_by_node(dp);
arch/sparc/include/asm/floppy_64.h
565
if (op)
arch/sparc/include/asm/floppy_64.h
568
if (op) {
arch/sparc/include/asm/floppy_64.h
569
floppy_op = op;
arch/sparc/include/asm/floppy_64.h
570
FLOPPY_IRQ = op->archdata.irqs[0];
arch/sparc/include/asm/floppy_64.h
588
op = of_find_device_by_node(dp);
arch/sparc/include/asm/floppy_64.h
589
if (!op)
arch/sparc/include/asm/floppy_64.h
592
state_prop = of_get_property(op->dev.of_node, "status", NULL);
arch/sparc/include/asm/floppy_64.h
596
FLOPPY_IRQ = op->archdata.irqs[0];
arch/sparc/include/asm/floppy_64.h
601
auxio_reg = (void __iomem *) op->resource[2].start;
arch/sparc/include/asm/floppy_64.h
604
sun_floppy_dev = &op->dev;
arch/sparc/include/asm/floppy_64.h
610
op->resource[1].start;
arch/sparc/include/asm/floppy_64.h
624
sun_fdc = (struct sun_flpy_controller *) op->resource[0].start;
arch/sparc/include/asm/floppy_64.h
717
prop = of_get_property(op->dev.of_node, "status", NULL);
arch/sparc/include/asm/floppy_64.h
728
(op->resource[0].start +
arch/sparc/include/asm/floppy_64.h
729
((op->resource[0].flags & 0x1ffUL) << 32UL));
arch/sparc/include/asm/futex_64.h
33
static inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/sparc/include/asm/futex_64.h
41
switch (op) {
arch/sparc/include/asm/parport_64.h
111
static int ecpp_probe(struct platform_device *op)
arch/sparc/include/asm/parport_64.h
113
unsigned long base = op->resource[0].start;
arch/sparc/include/asm/parport_64.h
114
unsigned long config = op->resource[1].start;
arch/sparc/include/asm/parport_64.h
115
unsigned long d_base = op->resource[2].start;
arch/sparc/include/asm/parport_64.h
121
parent = op->dev.of_node->parent;
arch/sparc/include/asm/parport_64.h
124
op->archdata.irqs[0], PARPORT_DMA_NOFIFO,
arch/sparc/include/asm/parport_64.h
125
op->dev.parent->parent, 0);
arch/sparc/include/asm/parport_64.h
128
dev_set_drvdata(&op->dev, p);
arch/sparc/include/asm/parport_64.h
142
d_len = (op->resource[2].end - d_base) + 1UL;
arch/sparc/include/asm/parport_64.h
144
of_ioremap(&op->resource[2], 0, d_len, "ECPP DMA");
arch/sparc/include/asm/parport_64.h
174
op->archdata.irqs[0],
arch/sparc/include/asm/parport_64.h
176
op->dev.parent,
arch/sparc/include/asm/parport_64.h
182
dev_set_drvdata(&op->dev, p);
arch/sparc/include/asm/parport_64.h
191
of_iounmap(&op->resource[2], sparc_ebus_dmas[slot].info.regs, d_len);
arch/sparc/include/asm/parport_64.h
200
static void ecpp_remove(struct platform_device *op)
arch/sparc/include/asm/parport_64.h
202
struct parport *p = dev_get_drvdata(&op->dev);
arch/sparc/include/asm/parport_64.h
208
unsigned long d_base = op->resource[2].start;
arch/sparc/include/asm/parport_64.h
211
d_len = (op->resource[2].end - d_base) + 1UL;
arch/sparc/include/asm/parport_64.h
215
of_iounmap(&op->resource[2],
arch/sparc/kernel/apc.c
142
static int apc_probe(struct platform_device *op)
arch/sparc/kernel/apc.c
146
regs = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/apc.c
147
resource_size(&op->resource[0]), APC_OBPNAME);
arch/sparc/kernel/apc.c
156
apc_free(op);
arch/sparc/kernel/apc.c
73
static inline void apc_free(struct platform_device *op)
arch/sparc/kernel/apc.c
75
of_iounmap(&op->resource[0], regs, resource_size(&op->resource[0]));
arch/sparc/kernel/central.c
111
p->leds_pdev.dev.parent = &op->dev;
arch/sparc/kernel/central.c
129
of_iounmap(&op->resource[2], p->clock_ver_reg,
arch/sparc/kernel/central.c
130
resource_size(&op->resource[2]));
arch/sparc/kernel/central.c
133
of_iounmap(&op->resource[1], p->clock_regs,
arch/sparc/kernel/central.c
134
resource_size(&op->resource[1]));
arch/sparc/kernel/central.c
137
of_iounmap(&op->resource[0], p->clock_freq_regs,
arch/sparc/kernel/central.c
138
resource_size(&op->resource[0]));
arch/sparc/kernel/central.c
160
static int fhc_probe(struct platform_device *op)
arch/sparc/kernel/central.c
171
if (of_node_name_eq(op->dev.of_node->parent, "central"))
arch/sparc/kernel/central.c
174
p->pregs = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/central.c
175
resource_size(&op->resource[0]),
arch/sparc/kernel/central.c
186
p->board_num = of_getintprop_default(op->dev.of_node, "board#", -1);
arch/sparc/kernel/central.c
205
p->leds_pdev.dev.parent = &op->dev;
arch/sparc/kernel/central.c
242
of_iounmap(&op->resource[0], p->pregs, resource_size(&op->resource[0]));
arch/sparc/kernel/central.c
64
static int clock_board_probe(struct platform_device *op)
arch/sparc/kernel/central.c
74
p->clock_freq_regs = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/central.c
75
resource_size(&op->resource[0]),
arch/sparc/kernel/central.c
82
p->clock_regs = of_ioremap(&op->resource[1], 0,
arch/sparc/kernel/central.c
83
resource_size(&op->resource[1]),
arch/sparc/kernel/central.c
90
if (op->resource[2].flags) {
arch/sparc/kernel/central.c
91
p->clock_ver_reg = of_ioremap(&op->resource[2], 0,
arch/sparc/kernel/central.c
92
resource_size(&op->resource[2]),
arch/sparc/kernel/chmc.c
397
static int jbusmc_probe(struct platform_device *op)
arch/sparc/kernel/chmc.c
429
prop = of_get_property(op->dev.of_node, "portid", &len);
arch/sparc/kernel/chmc.c
437
prop = of_get_property(op->dev.of_node, "memory-control-register-1", &len);
arch/sparc/kernel/chmc.c
446
p->regs = of_ioremap(&op->resource[0], 0, JBUSMC_REGS_SIZE, "jbusmc");
arch/sparc/kernel/chmc.c
453
ml = of_get_property(op->dev.of_node, "memory-layout", &p->layout_len);
arch/sparc/kernel/chmc.c
470
op->dev.of_node);
arch/sparc/kernel/chmc.c
472
dev_set_drvdata(&op->dev, p);
arch/sparc/kernel/chmc.c
480
of_iounmap(&op->resource[0], p->regs, JBUSMC_REGS_SIZE);
arch/sparc/kernel/chmc.c
694
static int chmc_probe(struct platform_device *op)
arch/sparc/kernel/chmc.c
696
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/chmc.c
734
p->regs = of_ioremap(&op->resource[0], 0, 0x48, "chmc");
arch/sparc/kernel/chmc.c
756
dev_set_drvdata(&op->dev, p);
arch/sparc/kernel/chmc.c
768
static int us3mc_probe(struct platform_device *op)
arch/sparc/kernel/chmc.c
771
return chmc_probe(op);
arch/sparc/kernel/chmc.c
773
return jbusmc_probe(op);
arch/sparc/kernel/chmc.c
777
static void chmc_destroy(struct platform_device *op, struct chmc *p)
arch/sparc/kernel/chmc.c
780
of_iounmap(&op->resource[0], p->regs, 0x48);
arch/sparc/kernel/chmc.c
784
static void jbusmc_destroy(struct platform_device *op, struct jbusmc *p)
arch/sparc/kernel/chmc.c
787
of_iounmap(&op->resource[0], p->regs, JBUSMC_REGS_SIZE);
arch/sparc/kernel/chmc.c
791
static void us3mc_remove(struct platform_device *op)
arch/sparc/kernel/chmc.c
793
void *p = dev_get_drvdata(&op->dev);
arch/sparc/kernel/chmc.c
797
chmc_destroy(op, p);
arch/sparc/kernel/chmc.c
799
jbusmc_destroy(op, p);
arch/sparc/kernel/entry.h
181
unsigned long op);
arch/sparc/kernel/entry.h
183
unsigned long op);
arch/sparc/kernel/irq.h
59
unsigned int (*build_device_irq)(struct platform_device *op,
arch/sparc/kernel/leon_kernel.c
239
static unsigned int _leon_build_device_irq(struct platform_device *op,
arch/sparc/kernel/of_device_32.c
246
static void __init build_device_resources(struct platform_device *op,
arch/sparc/kernel/of_device_32.c
260
bus->count_cells(op->dev.of_node, &na, &ns);
arch/sparc/kernel/of_device_32.c
262
preg = of_get_property(op->dev.of_node, bus->addr_prop_name, &num_reg);
arch/sparc/kernel/of_device_32.c
272
op->resource = op->archdata.resource;
arch/sparc/kernel/of_device_32.c
273
op->num_resources = num_reg;
arch/sparc/kernel/of_device_32.c
275
struct resource *r = &op->resource[index];
arch/sparc/kernel/of_device_32.c
278
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/of_device_32.c
328
op->dev.of_node, index,
arch/sparc/kernel/of_device_32.c
336
r->name = op->dev.of_node->full_name;
arch/sparc/kernel/of_device_32.c
343
struct platform_device *op = kzalloc_obj(*op);
arch/sparc/kernel/of_device_32.c
348
if (!op)
arch/sparc/kernel/of_device_32.c
351
sd = &op->dev.archdata;
arch/sparc/kernel/of_device_32.c
352
sd->op = op;
arch/sparc/kernel/of_device_32.c
354
op->dev.of_node = dp;
arch/sparc/kernel/of_device_32.c
358
op->archdata.num_irqs = len / sizeof(struct linux_prom_irqs);
arch/sparc/kernel/of_device_32.c
359
for (i = 0; i < op->archdata.num_irqs; i++)
arch/sparc/kernel/of_device_32.c
360
op->archdata.irqs[i] =
arch/sparc/kernel/of_device_32.c
361
sparc_config.build_device_irq(op, intr[i].pri);
arch/sparc/kernel/of_device_32.c
367
op->archdata.num_irqs = len / sizeof(unsigned int);
arch/sparc/kernel/of_device_32.c
368
for (i = 0; i < op->archdata.num_irqs; i++)
arch/sparc/kernel/of_device_32.c
369
op->archdata.irqs[i] =
arch/sparc/kernel/of_device_32.c
370
sparc_config.build_device_irq(op, irq[i]);
arch/sparc/kernel/of_device_32.c
372
op->archdata.num_irqs = 0;
arch/sparc/kernel/of_device_32.c
376
build_device_resources(op, parent);
arch/sparc/kernel/of_device_32.c
378
op->dev.parent = parent;
arch/sparc/kernel/of_device_32.c
379
op->dev.bus = &platform_bus_type;
arch/sparc/kernel/of_device_32.c
381
dev_set_name(&op->dev, "root");
arch/sparc/kernel/of_device_32.c
383
dev_set_name(&op->dev, "%08x", dp->phandle);
arch/sparc/kernel/of_device_32.c
385
op->dev.coherent_dma_mask = DMA_BIT_MASK(32);
arch/sparc/kernel/of_device_32.c
386
op->dev.dma_mask = &op->dev.coherent_dma_mask;
arch/sparc/kernel/of_device_32.c
388
if (of_device_register(op)) {
arch/sparc/kernel/of_device_32.c
390
put_device(&op->dev);
arch/sparc/kernel/of_device_32.c
391
kfree(op);
arch/sparc/kernel/of_device_32.c
392
op = NULL;
arch/sparc/kernel/of_device_32.c
395
return op;
arch/sparc/kernel/of_device_32.c
401
struct platform_device *op = scan_one_device(dp, parent);
arch/sparc/kernel/of_device_32.c
403
if (op)
arch/sparc/kernel/of_device_32.c
404
scan_tree(dp->child, &op->dev);
arch/sparc/kernel/of_device_64.c
316
static void __init build_device_resources(struct platform_device *op,
arch/sparc/kernel/of_device_64.c
330
bus->count_cells(op->dev.of_node, &na, &ns);
arch/sparc/kernel/of_device_64.c
332
preg = of_get_property(op->dev.of_node, bus->addr_prop_name, &num_reg);
arch/sparc/kernel/of_device_64.c
346
op->dev.of_node, num_reg, PROMREG_MAX);
arch/sparc/kernel/of_device_64.c
350
op->resource = op->archdata.resource;
arch/sparc/kernel/of_device_64.c
351
op->num_resources = num_reg;
arch/sparc/kernel/of_device_64.c
353
struct resource *r = &op->resource[index];
arch/sparc/kernel/of_device_64.c
356
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/of_device_64.c
405
op->dev.of_node, index,
arch/sparc/kernel/of_device_64.c
416
r->name = op->dev.of_node->full_name;
arch/sparc/kernel/of_device_64.c
534
static unsigned int __init build_one_device_irq(struct platform_device *op,
arch/sparc/kernel/of_device_64.c
538
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/of_device_64.c
583
op->dev.of_node,
arch/sparc/kernel/of_device_64.c
601
op->dev.of_node,
arch/sparc/kernel/of_device_64.c
618
irq = ip->irq_trans->irq_build(op->dev.of_node, irq,
arch/sparc/kernel/of_device_64.c
622
op->dev.of_node, ip, orig_irq, irq);
arch/sparc/kernel/of_device_64.c
636
struct platform_device *op = kzalloc_obj(*op);
arch/sparc/kernel/of_device_64.c
641
if (!op)
arch/sparc/kernel/of_device_64.c
644
sd = &op->dev.archdata;
arch/sparc/kernel/of_device_64.c
645
sd->op = op;
arch/sparc/kernel/of_device_64.c
647
op->dev.of_node = dp;
arch/sparc/kernel/of_device_64.c
651
op->archdata.num_irqs = len / 4;
arch/sparc/kernel/of_device_64.c
654
if (op->archdata.num_irqs > PROMINTR_MAX) {
arch/sparc/kernel/of_device_64.c
657
dp, op->archdata.num_irqs, PROMINTR_MAX);
arch/sparc/kernel/of_device_64.c
658
op->archdata.num_irqs = PROMINTR_MAX;
arch/sparc/kernel/of_device_64.c
660
memcpy(op->archdata.irqs, irq, op->archdata.num_irqs * 4);
arch/sparc/kernel/of_device_64.c
662
op->archdata.num_irqs = 0;
arch/sparc/kernel/of_device_64.c
665
build_device_resources(op, parent);
arch/sparc/kernel/of_device_64.c
666
for (i = 0; i < op->archdata.num_irqs; i++)
arch/sparc/kernel/of_device_64.c
667
op->archdata.irqs[i] = build_one_device_irq(op, parent, op->archdata.irqs[i]);
arch/sparc/kernel/of_device_64.c
669
op->dev.parent = parent;
arch/sparc/kernel/of_device_64.c
670
op->dev.bus = &platform_bus_type;
arch/sparc/kernel/of_device_64.c
672
dev_set_name(&op->dev, "root");
arch/sparc/kernel/of_device_64.c
674
dev_set_name(&op->dev, "%08x", dp->phandle);
arch/sparc/kernel/of_device_64.c
675
op->dev.coherent_dma_mask = DMA_BIT_MASK(32);
arch/sparc/kernel/of_device_64.c
676
op->dev.dma_mask = &op->dev.coherent_dma_mask;
arch/sparc/kernel/of_device_64.c
678
if (of_device_register(op)) {
arch/sparc/kernel/of_device_64.c
680
put_device(&op->dev);
arch/sparc/kernel/of_device_64.c
681
kfree(op);
arch/sparc/kernel/of_device_64.c
682
op = NULL;
arch/sparc/kernel/of_device_64.c
685
return op;
arch/sparc/kernel/of_device_64.c
691
struct platform_device *op = scan_one_device(dp, parent);
arch/sparc/kernel/of_device_64.c
693
if (op)
arch/sparc/kernel/of_device_64.c
694
scan_tree(dp->child, &op->dev);
arch/sparc/kernel/of_device_common.c
18
struct platform_device *op = of_find_device_by_node(node);
arch/sparc/kernel/of_device_common.c
20
if (!op || index >= op->archdata.num_irqs)
arch/sparc/kernel/of_device_common.c
23
return op->archdata.irqs[index];
arch/sparc/kernel/of_device_common.c
30
struct platform_device *op = of_find_device_by_node(node);
arch/sparc/kernel/of_device_common.c
32
if (!op || index >= op->num_resources)
arch/sparc/kernel/of_device_common.c
35
memcpy(r, &op->archdata.resource[index], sizeof(*r));
arch/sparc/kernel/of_device_common.c
42
struct platform_device *op = of_find_device_by_node(node);
arch/sparc/kernel/of_device_common.c
45
if (!op || index >= op->num_resources)
arch/sparc/kernel/of_device_common.c
48
r = &op->archdata.resource[index];
arch/sparc/kernel/of_device_common.c
64
struct platform_device *op = of_find_device_by_node(dp);
arch/sparc/kernel/of_device_common.c
66
op->dev.archdata.iommu = bus_sd->iommu;
arch/sparc/kernel/of_device_common.c
67
op->dev.archdata.stc = bus_sd->stc;
arch/sparc/kernel/of_device_common.c
68
op->dev.archdata.host_controller = bus_sd->host_controller;
arch/sparc/kernel/of_device_common.c
69
op->dev.archdata.numa_node = bus_sd->numa_node;
arch/sparc/kernel/of_device_common.c
70
op->dev.dma_ops = bus->dev.dma_ops;
arch/sparc/kernel/of_device_common.c
73
of_propagate_archdata(op);
arch/sparc/kernel/pci.c
1003
node = pbm->op->dev.of_node;
arch/sparc/kernel/pci.c
228
static void pci_parse_of_addrs(struct platform_device *op,
arch/sparc/kernel/pci.c
242
op_res = &op->resource[0];
arch/sparc/kernel/pci.c
277
struct platform_device *op,
arch/sparc/kernel/pci.c
283
sd->op = op;
arch/sparc/kernel/pci.c
292
struct platform_device *op;
arch/sparc/kernel/pci.c
300
op = of_find_device_by_node(node);
arch/sparc/kernel/pci.c
302
pci_init_dev_archdata(sd, pbm->iommu, &pbm->stc, pbm, op,
arch/sparc/kernel/pci.c
304
sd = &op->dev.archdata;
arch/sparc/kernel/pci.c
310
of_propagate_archdata(op);
arch/sparc/kernel/pci.c
375
dev->irq = sd->op->archdata.irqs[0];
arch/sparc/kernel/pci.c
383
pci_parse_of_addrs(sd->op, node, dev);
arch/sparc/kernel/pci.c
722
struct device_node *node = pbm->op->dev.of_node;
arch/sparc/kernel/pci_common.c
321
const u32 *val = of_get_property(pbm->op->dev.of_node, "bus-range", NULL);
arch/sparc/kernel/pci_common.c
326
val = of_get_property(pbm->op->dev.of_node, "ino-bitmap", NULL);
arch/sparc/kernel/pci_common.c
335
const u32 *vdma = of_get_property(pbm->op->dev.of_node, "virtual-dma",
arch/sparc/kernel/pci_common.c
367
pbm_ranges = of_get_property(pbm->op->dev.of_node, "ranges", &i);
arch/sparc/kernel/pci_fire.c
415
struct platform_device *op, u32 portid)
arch/sparc/kernel/pci_fire.c
418
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_fire.c
429
pbm->op = op;
arch/sparc/kernel/pci_fire.c
450
pbm->pci_bus = pci_scan_one_pbm(pbm, &op->dev);
arch/sparc/kernel/pci_fire.c
460
static int fire_probe(struct platform_device *op)
arch/sparc/kernel/pci_fire.c
462
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_fire.c
485
err = pci_fire_pbm_init(pbm, op, portid);
arch/sparc/kernel/pci_fire.c
489
dev_set_drvdata(&op->dev, pbm);
arch/sparc/kernel/pci_impl.h
94
struct platform_device *op;
arch/sparc/kernel/pci_msi.c
325
val = of_get_property(pbm->op->dev.of_node, "#msi-eqs", &len);
arch/sparc/kernel/pci_msi.c
348
val = of_get_property(pbm->op->dev.of_node, "msi-eq-size", &len);
arch/sparc/kernel/pci_msi.c
354
mqp = of_get_property(pbm->op->dev.of_node,
arch/sparc/kernel/pci_msi.c
357
mqp = of_get_property(pbm->op->dev.of_node,
arch/sparc/kernel/pci_msi.c
365
val = of_get_property(pbm->op->dev.of_node, "#msi", &len);
arch/sparc/kernel/pci_msi.c
370
mrng = of_get_property(pbm->op->dev.of_node, "msi-ranges", &len);
arch/sparc/kernel/pci_msi.c
375
val = of_get_property(pbm->op->dev.of_node, "msi-data-mask", &len);
arch/sparc/kernel/pci_msi.c
380
val = of_get_property(pbm->op->dev.of_node, "msix-data-width", &len);
arch/sparc/kernel/pci_msi.c
385
arng = of_get_property(pbm->op->dev.of_node, "msi-address-ranges",
arch/sparc/kernel/pci_psycho.c
292
struct platform_device *op = of_find_device_by_node(pbm->op->dev.of_node);
arch/sparc/kernel/pci_psycho.c
297
if (!op)
arch/sparc/kernel/pci_psycho.c
309
if (op->archdata.num_irqs < 6)
arch/sparc/kernel/pci_psycho.c
316
err = request_irq(op->archdata.irqs[1], psycho_ue_intr, IRQF_SHARED,
arch/sparc/kernel/pci_psycho.c
318
err = request_irq(op->archdata.irqs[2], psycho_ce_intr, IRQF_SHARED,
arch/sparc/kernel/pci_psycho.c
325
err = request_irq(op->archdata.irqs[0], psycho_pcierr_intr, IRQF_SHARED,
arch/sparc/kernel/pci_psycho.c
490
struct platform_device *op, int is_pbm_a)
arch/sparc/kernel/pci_psycho.c
492
psycho_pbm_init_common(pbm, op, "PSYCHO", PBM_CHIP_TYPE_PSYCHO);
arch/sparc/kernel/pci_psycho.c
494
psycho_scan_bus(pbm, &op->dev);
arch/sparc/kernel/pci_psycho.c
510
static int psycho_probe(struct platform_device *op)
arch/sparc/kernel/pci_psycho.c
513
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_psycho.c
576
psycho_pbm_init(pbm, op, is_pbm_a);
arch/sparc/kernel/pci_psycho.c
584
dev_set_drvdata(&op->dev, pbm);
arch/sparc/kernel/pci_sabre.c
318
struct device_node *dp = pbm->op->dev.of_node;
arch/sparc/kernel/pci_sabre.c
319
struct platform_device *op;
arch/sparc/kernel/pci_sabre.c
327
op = of_find_device_by_node(dp);
arch/sparc/kernel/pci_sabre.c
328
if (!op)
arch/sparc/kernel/pci_sabre.c
337
if (op->archdata.num_irqs < 4)
arch/sparc/kernel/pci_sabre.c
349
err = request_irq(op->archdata.irqs[1], sabre_ue_intr, 0, "SABRE_UE", pbm);
arch/sparc/kernel/pci_sabre.c
359
err = request_irq(op->archdata.irqs[2], sabre_ce_intr, 0, "SABRE_CE", pbm);
arch/sparc/kernel/pci_sabre.c
363
err = request_irq(op->archdata.irqs[0], psycho_pcierr_intr, 0,
arch/sparc/kernel/pci_sabre.c
450
struct platform_device *op)
arch/sparc/kernel/pci_sabre.c
452
psycho_pbm_init_common(pbm, op, "SABRE", PBM_CHIP_TYPE_SABRE);
arch/sparc/kernel/pci_sabre.c
456
sabre_scan_bus(pbm, &op->dev);
arch/sparc/kernel/pci_sabre.c
460
static int sabre_probe(struct platform_device *op)
arch/sparc/kernel/pci_sabre.c
463
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_sabre.c
471
hummingbird_p = (uintptr_t)device_get_match_data(&op->dev);
arch/sparc/kernel/pci_sabre.c
570
sabre_pbm_init(pbm, op);
arch/sparc/kernel/pci_sabre.c
575
dev_set_drvdata(&op->dev, pbm);
arch/sparc/kernel/pci_schizo.c
1076
(of_find_property(pbm->op->dev.of_node, "66mhz-capable", NULL)
arch/sparc/kernel/pci_schizo.c
1146
vdma = of_get_property(pbm->op->dev.of_node, "virtual-dma", NULL);
arch/sparc/kernel/pci_schizo.c
1276
if (!of_property_read_bool(pbm->op->dev.of_node, "no-bus-parking"))
arch/sparc/kernel/pci_schizo.c
1315
struct platform_device *op, u32 portid,
arch/sparc/kernel/pci_schizo.c
1319
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_schizo.c
1362
pbm->op = op;
arch/sparc/kernel/pci_schizo.c
1392
schizo_scan_bus(pbm, &op->dev);
arch/sparc/kernel/pci_schizo.c
1418
static int __schizo_init(struct platform_device *op, unsigned long chip_type)
arch/sparc/kernel/pci_schizo.c
1420
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_schizo.c
1445
if (schizo_pbm_init(pbm, op, portid, chip_type))
arch/sparc/kernel/pci_schizo.c
1451
dev_set_drvdata(&op->dev, pbm);
arch/sparc/kernel/pci_schizo.c
1465
static int schizo_probe(struct platform_device *op)
arch/sparc/kernel/pci_schizo.c
1467
unsigned long chip_type = (unsigned long)device_get_match_data(&op->dev);
arch/sparc/kernel/pci_schizo.c
1471
return __schizo_init(op, chip_type);
arch/sparc/kernel/pci_schizo.c
853
struct platform_device *op = of_find_device_by_node(pbm->op->dev.of_node);
arch/sparc/kernel/pci_schizo.c
866
err = request_irq(op->archdata.irqs[1], schizo_ue_intr, 0,
arch/sparc/kernel/pci_schizo.c
873
err = request_irq(op->archdata.irqs[2], schizo_ce_intr, 0,
arch/sparc/kernel/pci_schizo.c
881
err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0,
arch/sparc/kernel/pci_schizo.c
884
err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0,
arch/sparc/kernel/pci_schizo.c
892
err = request_irq(op->archdata.irqs[3], schizo_safarierr_intr, 0,
arch/sparc/kernel/pci_schizo.c
948
struct platform_device *op = of_find_device_by_node(pbm->op->dev.of_node);
arch/sparc/kernel/pci_schizo.c
961
err = request_irq(op->archdata.irqs[1], schizo_ue_intr, 0,
arch/sparc/kernel/pci_schizo.c
968
err = request_irq(op->archdata.irqs[2], schizo_ce_intr, 0,
arch/sparc/kernel/pci_schizo.c
976
err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0,
arch/sparc/kernel/pci_schizo.c
979
err = request_irq(op->archdata.irqs[0], schizo_pcierr_intr, 0,
arch/sparc/kernel/pci_schizo.c
987
err = request_irq(op->archdata.irqs[3], schizo_safarierr_intr, 0,
arch/sparc/kernel/pci_sun4v.c
1179
struct platform_device *op, u32 devhandle)
arch/sparc/kernel/pci_sun4v.c
1181
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/pci_sun4v.c
1191
pbm->op = op;
arch/sparc/kernel/pci_sun4v.c
1210
pci_sun4v_scan_bus(pbm, &op->dev);
arch/sparc/kernel/pci_sun4v.c
1230
static int pci_sun4v_probe(struct platform_device *op)
arch/sparc/kernel/pci_sun4v.c
1242
dp = op->dev.of_node;
arch/sparc/kernel/pci_sun4v.c
1319
err = pci_sun4v_pbm_init(pbm, op, devhandle);
arch/sparc/kernel/pci_sun4v.c
1323
dev_set_drvdata(&op->dev, pbm);
arch/sparc/kernel/pci_sun4v.c
710
dp = pbm->op->dev.of_node;
arch/sparc/kernel/pci_sun4v.c
820
ranges = of_get_property(pbm->op->dev.of_node, "iommu-address-ranges",
arch/sparc/kernel/pci_sun4v.c
827
page_size = of_get_property(pbm->op->dev.of_node, "iommu-pagesizes",
arch/sparc/kernel/pci_sun4v.c
883
vdma = of_get_property(pbm->op->dev.of_node, "virtual-dma", NULL);
arch/sparc/kernel/pcic.c
172
unsigned int pcic_build_device_irq(struct platform_device *op,
arch/sparc/kernel/pcic.c
778
unsigned int pcic_build_device_irq(struct platform_device *op,
arch/sparc/kernel/pmc.c
57
static int pmc_probe(struct platform_device *op)
arch/sparc/kernel/pmc.c
59
regs = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/pmc.c
60
resource_size(&op->resource[0]), PMC_OBPNAME);
arch/sparc/kernel/power.c
38
static int power_probe(struct platform_device *op)
arch/sparc/kernel/power.c
40
struct resource *res = &op->resource[0];
arch/sparc/kernel/power.c
41
unsigned int irq = op->archdata.irqs[0];
arch/sparc/kernel/power.c
46
op->dev.of_node, res->start);
arch/sparc/kernel/power.c
48
if (has_button_interrupt(irq, op->dev.of_node)) {
arch/sparc/kernel/psycho_common.c
453
void psycho_pbm_init_common(struct pci_pbm_info *pbm, struct platform_device *op,
arch/sparc/kernel/psycho_common.c
456
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/psycho_common.c
463
pbm->op = op;
arch/sparc/kernel/psycho_common.h
46
struct platform_device *op,
arch/sparc/kernel/sbus.c
211
static unsigned int sbus_build_irq(struct platform_device *op, unsigned int ino)
arch/sparc/kernel/sbus.c
213
struct iommu *iommu = op->dev.archdata.iommu;
arch/sparc/kernel/sbus.c
274
struct platform_device *op = dev_id;
arch/sparc/kernel/sbus.c
275
struct iommu *iommu = op->dev.archdata.iommu;
arch/sparc/kernel/sbus.c
294
portid = of_getintprop_default(op->dev.of_node, "portid", -1);
arch/sparc/kernel/sbus.c
348
struct platform_device *op = dev_id;
arch/sparc/kernel/sbus.c
349
struct iommu *iommu = op->dev.archdata.iommu;
arch/sparc/kernel/sbus.c
368
portid = of_getintprop_default(op->dev.of_node, "portid", -1);
arch/sparc/kernel/sbus.c
427
struct platform_device *op = dev_id;
arch/sparc/kernel/sbus.c
428
struct iommu *iommu = op->dev.archdata.iommu;
arch/sparc/kernel/sbus.c
446
portid = of_getintprop_default(op->dev.of_node, "portid", -1);
arch/sparc/kernel/sbus.c
495
static void __init sysio_register_error_handlers(struct platform_device *op)
arch/sparc/kernel/sbus.c
497
struct iommu *iommu = op->dev.archdata.iommu;
arch/sparc/kernel/sbus.c
503
portid = of_getintprop_default(op->dev.of_node, "portid", -1);
arch/sparc/kernel/sbus.c
505
irq = sbus_build_irq(op, SYSIO_UE_INO);
arch/sparc/kernel/sbus.c
507
"SYSIO_UE", op) < 0) {
arch/sparc/kernel/sbus.c
513
irq = sbus_build_irq(op, SYSIO_CE_INO);
arch/sparc/kernel/sbus.c
515
"SYSIO_CE", op) < 0) {
arch/sparc/kernel/sbus.c
521
irq = sbus_build_irq(op, SYSIO_SBUSERR_INO);
arch/sparc/kernel/sbus.c
523
"SYSIO_SBERR", op) < 0) {
arch/sparc/kernel/sbus.c
541
static void __init sbus_iommu_init(struct platform_device *op)
arch/sparc/kernel/sbus.c
544
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/sbus.c
564
op->dev.archdata.iommu = iommu;
arch/sparc/kernel/sbus.c
565
op->dev.archdata.stc = strbuf;
arch/sparc/kernel/sbus.c
566
op->dev.archdata.numa_node = NUMA_NO_NODE;
arch/sparc/kernel/sbus.c
594
portid = of_getintprop_default(op->dev.of_node, "portid", -1);
arch/sparc/kernel/sbus.c
64
struct platform_device *op = to_platform_device(dev);
arch/sparc/kernel/sbus.c
656
sysio_register_error_handlers(op);
arch/sparc/kernel/sbus.c
670
struct platform_device *op = of_find_device_by_node(dp);
arch/sparc/kernel/sbus.c
672
sbus_iommu_init(op);
arch/sparc/kernel/sbus.c
673
of_propagate_archdata(op);
arch/sparc/kernel/sbus.c
70
regs = of_get_property(op->dev.of_node, "reg", NULL);
arch/sparc/kernel/sbus.c
73
op->dev.of_node);
arch/sparc/kernel/sun4d_irq.c
324
static unsigned int sun4d_build_device_irq(struct platform_device *op,
arch/sparc/kernel/sun4d_irq.c
327
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/sun4m_irq.c
250
static unsigned int sun4m_build_device_irq(struct platform_device *op,
arch/sparc/kernel/time_32.c
271
static int clock_probe(struct platform_device *op)
arch/sparc/kernel/time_32.c
273
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/time_32.c
283
m48t59_rtc.resource = &op->resource[0];
arch/sparc/kernel/time_32.c
286
m48t59_data.ioaddr = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/time_32.c
290
m48t59_data.ioaddr = of_ioremap(&op->resource[0], 0,
arch/sparc/kernel/time_64.c
445
static int rtc_probe(struct platform_device *op)
arch/sparc/kernel/time_64.c
450
op->dev.of_node, op->resource[0].start);
arch/sparc/kernel/time_64.c
460
r->name = op->resource[0].name;
arch/sparc/kernel/time_64.c
461
r->start = op->resource[0].start;
arch/sparc/kernel/time_64.c
462
r->end = op->resource[0].end;
arch/sparc/kernel/time_64.c
464
cmos_regs = op->resource[0].start;
arch/sparc/kernel/time_64.c
502
static int bq4802_probe(struct platform_device *op)
arch/sparc/kernel/time_64.c
506
op->dev.of_node, op->resource[0].start);
arch/sparc/kernel/time_64.c
508
rtc_bq4802_device.resource = &op->resource[0];
arch/sparc/kernel/time_64.c
559
static int mostek_probe(struct platform_device *op)
arch/sparc/kernel/time_64.c
561
struct device_node *dp = op->dev.of_node;
arch/sparc/kernel/time_64.c
571
dp, op->resource[0].start);
arch/sparc/kernel/time_64.c
573
m48t59_rtc.resource = &op->resource[0];
arch/sparc/kernel/traps_64.c
2304
void hypervisor_tlbop_error(unsigned long err, unsigned long op)
arch/sparc/kernel/traps_64.c
2307
err, op);
arch/sparc/kernel/traps_64.c
2310
void hypervisor_tlbop_error_xcall(unsigned long err, unsigned long op)
arch/sparc/kernel/traps_64.c
2313
err, op);
arch/sparc/kernel/uprobes.c
61
u32 op = (insn >> 30) & 0x3;
arch/sparc/kernel/uprobes.c
64
if (op == 0 &&
arch/sparc/lib/atomic32.c
31
#define ATOMIC_FETCH_OP(op, c_op) \
arch/sparc/lib/atomic32.c
32
int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/sparc/lib/atomic32.c
44
EXPORT_SYMBOL(arch_atomic_fetch_##op);
arch/sparc/lib/atomic32.c
46
#define ATOMIC_OP_RETURN(op, c_op) \
arch/sparc/lib/atomic32.c
47
int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/sparc/lib/atomic32.c
58
EXPORT_SYMBOL(arch_atomic_##op##_return);
arch/sparc/mm/io-unit.c
41
static void __init iounit_iommu_init(struct platform_device *op)
arch/sparc/mm/io-unit.c
60
xpt = of_ioremap(&op->resource[2], 0, PAGE_SIZE * 16, "XPT");
arch/sparc/mm/io-unit.c
66
op->dev.archdata.iommu = iounit;
arch/sparc/mm/io-unit.c
74
op->dev.dma_ops = &iounit_dma_ops;
arch/sparc/mm/io-unit.c
83
struct platform_device *op = of_find_device_by_node(dp);
arch/sparc/mm/io-unit.c
85
iounit_iommu_init(op);
arch/sparc/mm/io-unit.c
86
of_propagate_archdata(op);
arch/sparc/mm/iommu.c
132
op->dev.archdata.iommu = iommu;
arch/sparc/mm/iommu.c
135
op->dev.dma_ops = &sbus_iommu_dma_gflush_ops;
arch/sparc/mm/iommu.c
137
op->dev.dma_ops = &sbus_iommu_dma_pflush_ops;
arch/sparc/mm/iommu.c
145
struct platform_device *op = of_find_device_by_node(dp);
arch/sparc/mm/iommu.c
147
sbus_iommu_init(op);
arch/sparc/mm/iommu.c
148
of_propagate_archdata(op);
arch/sparc/mm/iommu.c
58
static void __init sbus_iommu_init(struct platform_device *op)
arch/sparc/mm/iommu.c
73
iommu->regs = of_ioremap(&op->resource[0], 0, PAGE_SIZE * 3,
arch/um/drivers/mconsole_kern.c
622
char *op = arg;
arch/um/drivers/mconsole_kern.c
623
handle_sysrq(*op);
arch/um/drivers/ubd_kern.c
1198
enum req_op op = req_op(req);
arch/um/drivers/ubd_kern.c
1200
if (op == REQ_OP_WRITE_ZEROES || op == REQ_OP_DISCARD) {
arch/um/drivers/ubd_kern.c
1261
enum req_op op = req_op(req);
arch/um/drivers/ubd_kern.c
1263
if (op == REQ_OP_FLUSH)
arch/um/drivers/ubd_kern.c
1265
else if (op == REQ_OP_WRITE_ZEROES || op == REQ_OP_DISCARD)
arch/um/drivers/virtio_pcidev.c
111
switch (cmd->op) {
arch/um/drivers/virtio_pcidev.c
211
.op = VIRTIO_PCIDEV_OP_CFG_READ,
arch/um/drivers/virtio_pcidev.c
251
.op = VIRTIO_PCIDEV_OP_CFG_WRITE,
arch/um/drivers/virtio_pcidev.c
284
.op = VIRTIO_PCIDEV_OP_MMIO_READ,
arch/um/drivers/virtio_pcidev.c
326
.op = VIRTIO_PCIDEV_OP_MMIO_WRITE,
arch/um/drivers/virtio_pcidev.c
372
.op = VIRTIO_PCIDEV_OP_CFG_WRITE,
arch/um/drivers/virtio_pcidev.c
415
switch (msg->op) {
arch/um/drivers/virtio_pcidev.c
430
dev_err(&vdev->dev, "unexpected virt-pci message %d\n", msg->op);
arch/um/include/asm/futex.h
10
int arch_futex_atomic_op_inuser(int op, u32 oparg, int *oval, u32 __user *uaddr);
arch/um/kernel/skas/uaccess.c
103
n = do_op_one_page(addr, size, is_write, op, arg);
arch/um/kernel/skas/uaccess.c
115
n = do_op_one_page(addr, PAGE_SIZE, is_write, op, arg);
arch/um/kernel/skas/uaccess.c
127
n = do_op_one_page(addr, remain, is_write, op, arg);
arch/um/kernel/skas/uaccess.c
251
int arch_futex_atomic_op_inuser(int op, u32 oparg, int *oval, u32 __user *uaddr)
arch/um/kernel/skas/uaccess.c
280
switch (op) {
arch/um/kernel/skas/uaccess.c
65
int (*op)(unsigned long addr, int len, void *arg), void *arg)
arch/um/kernel/skas/uaccess.c
84
n = (*op)(addr, len, arg);
arch/um/kernel/skas/uaccess.c
96
int (*op)(unsigned long, int, void *), void *arg)
arch/um/kernel/time.c
126
.op = UM_TIMETRAVEL_ACK,
arch/um/kernel/time.c
162
switch (msg->op) {
arch/um/kernel/time.c
165
(unsigned long long)msg->op);
arch/um/kernel/time.c
196
static u64 time_travel_ext_req(u32 op, u64 time)
arch/um/kernel/time.c
201
.op = op,
arch/um/kernel/time.c
226
if (msg.op == UM_TIMETRAVEL_WAIT && time_travel_shm)
arch/um/kernel/time.c
229
while (msg.op != UM_TIMETRAVEL_ACK)
arch/um/kernel/time.c
231
op == UM_TIMETRAVEL_START ?
arch/um/kernel/time.c
237
msg.op, msg.seq, mseq, msg.time);
arch/um/kernel/time.c
239
if (op == UM_TIMETRAVEL_GET)
arch/um/kernel/time.c
349
.op = UM_TIMETRAVEL_ACK,
arch/um/kernel/time.c
366
while (msg.op != UM_TIMETRAVEL_RUN)
arch/um/kernel/time.c
753
extern u64 time_travel_ext_req(u32 op, u64 time);
arch/um/os-Linux/skas/process.c
548
int err, status, op;
arch/um/os-Linux/skas/process.c
682
op = PTRACE_SYSEMU_SINGLESTEP;
arch/um/os-Linux/skas/process.c
684
op = PTRACE_SYSEMU;
arch/um/os-Linux/skas/process.c
686
if (ptrace(op, pid, 0, 0)) {
arch/um/os-Linux/skas/process.c
688
__func__, op, errno);
arch/x86/boot/startup/sev-shared.c
668
if (desc->op == SNP_PAGE_STATE_SHARED)
arch/x86/boot/startup/sev-shared.c
675
sev_es_wr_ghcb_msr(GHCB_MSR_PSC_REQ_GFN(paddr >> PAGE_SHIFT, desc->op));
arch/x86/boot/startup/sev-shared.c
690
if (desc->op == SNP_PAGE_STATE_PRIVATE)
arch/x86/coco/sev/core.c
320
unsigned long vaddr_end, int op)
arch/x86/coco/sev/core.c
349
e->operation = op;
arch/x86/coco/sev/core.c
365
if (op == SNP_PAGE_STATE_SHARED)
arch/x86/coco/sev/core.c
385
if (op == SNP_PAGE_STATE_PRIVATE)
arch/x86/coco/sev/core.c
391
static void set_pages_state(unsigned long vaddr, unsigned long npages, int op)
arch/x86/coco/sev/core.c
398
struct psc_desc d = { op, svsm_get_caa(), svsm_get_caa_pa() };
arch/x86/coco/sev/core.c
407
vaddr = __set_pages_state(&desc, vaddr, vaddr_end, op);
arch/x86/include/asm/cmpxchg.h
42
#define __xchg_op(ptr, arg, op, lock) \
arch/x86/include/asm/cmpxchg.h
47
asm_inline volatile (lock #op "b %b0, %1" \
arch/x86/include/asm/cmpxchg.h
52
asm_inline volatile (lock #op "w %w0, %1" \
arch/x86/include/asm/cmpxchg.h
57
asm_inline volatile (lock #op "l %0, %1" \
arch/x86/include/asm/cmpxchg.h
62
asm_inline volatile (lock #op "q %q0, %1" \
arch/x86/include/asm/cmpxchg.h
67
__ ## op ## _wrong_size(); \
arch/x86/include/asm/cpuid/api.h
100
static inline u32 cpuid_ebx(u32 op)
arch/x86/include/asm/cpuid/api.h
104
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
109
static inline u32 cpuid_ecx(u32 op)
arch/x86/include/asm/cpuid/api.h
113
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
118
static inline u32 cpuid_edx(u32 op)
arch/x86/include/asm/cpuid/api.h
122
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/cpuid/api.h
39
static inline u32 native_cpuid_##reg(u32 op) \
arch/x86/include/asm/cpuid/api.h
41
u32 eax = op, ebx, ecx = 0, edx; \
arch/x86/include/asm/cpuid/api.h
68
static inline void cpuid(u32 op,
arch/x86/include/asm/cpuid/api.h
72
*eax = op;
arch/x86/include/asm/cpuid/api.h
78
static inline void cpuid_count(u32 op, int count,
arch/x86/include/asm/cpuid/api.h
82
*eax = op;
arch/x86/include/asm/cpuid/api.h
91
static inline u32 cpuid_eax(u32 op)
arch/x86/include/asm/cpuid/api.h
95
cpuid(op, &eax, &ebx, &ecx, &edx);
arch/x86/include/asm/ftrace.h
84
struct ftrace_ops *op, struct ftrace_regs *fregs);
arch/x86/include/asm/futex.h
48
static __always_inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/x86/include/asm/futex.h
52
switch (op) {
arch/x86/include/asm/paravirt_types.h
198
#define paravirt_ptr(array, op) [paravirt_opptr] "m" (array.op)
arch/x86/include/asm/paravirt_types.h
348
#define ____PVOP_CALL(ret, array, op, call_clbr, extra_clbr, ...) \
arch/x86/include/asm/paravirt_types.h
354
: paravirt_ptr(array, op), \
arch/x86/include/asm/paravirt_types.h
360
#define ____PVOP_ALT_CALL(ret, array, op, alt, cond, call_clbr, \
arch/x86/include/asm/paravirt_types.h
368
: paravirt_ptr(array, op), \
arch/x86/include/asm/paravirt_types.h
374
#define __PVOP_CALL(rettype, array, op, ...) \
arch/x86/include/asm/paravirt_types.h
375
____PVOP_CALL(PVOP_RETVAL(rettype), array, op, \
arch/x86/include/asm/paravirt_types.h
378
#define __PVOP_ALT_CALL(rettype, array, op, alt, cond, ...) \
arch/x86/include/asm/paravirt_types.h
379
____PVOP_ALT_CALL(PVOP_RETVAL(rettype), array, op, alt, cond, \
arch/x86/include/asm/paravirt_types.h
383
#define __PVOP_CALLEESAVE(rettype, array, op, ...) \
arch/x86/include/asm/paravirt_types.h
384
____PVOP_CALL(PVOP_RETVAL(rettype), array, op.func, \
arch/x86/include/asm/paravirt_types.h
387
#define __PVOP_ALT_CALLEESAVE(rettype, array, op, alt, cond, ...) \
arch/x86/include/asm/paravirt_types.h
388
____PVOP_ALT_CALL(PVOP_RETVAL(rettype), array, op.func, alt, cond, \
arch/x86/include/asm/paravirt_types.h
392
#define __PVOP_VCALL(array, op, ...) \
arch/x86/include/asm/paravirt_types.h
393
(void)____PVOP_CALL(, array, op, PVOP_VCALL_CLOBBERS, \
arch/x86/include/asm/paravirt_types.h
396
#define __PVOP_ALT_VCALL(array, op, alt, cond, ...) \
arch/x86/include/asm/paravirt_types.h
397
(void)____PVOP_ALT_CALL(, array, op, alt, cond, \
arch/x86/include/asm/paravirt_types.h
401
#define __PVOP_VCALLEESAVE(array, op, ...) \
arch/x86/include/asm/paravirt_types.h
402
(void)____PVOP_CALL(, array, op.func, \
arch/x86/include/asm/paravirt_types.h
405
#define __PVOP_ALT_VCALLEESAVE(array, op, alt, cond, ...) \
arch/x86/include/asm/paravirt_types.h
406
(void)____PVOP_ALT_CALL(, array, op.func, alt, cond, \
arch/x86/include/asm/paravirt_types.h
410
#define PVOP_CALL0(rettype, array, op) \
arch/x86/include/asm/paravirt_types.h
411
__PVOP_CALL(rettype, array, op)
arch/x86/include/asm/paravirt_types.h
412
#define PVOP_VCALL0(array, op) \
arch/x86/include/asm/paravirt_types.h
413
__PVOP_VCALL(array, op)
arch/x86/include/asm/paravirt_types.h
414
#define PVOP_ALT_CALL0(rettype, array, op, alt, cond) \
arch/x86/include/asm/paravirt_types.h
415
__PVOP_ALT_CALL(rettype, array, op, alt, cond)
arch/x86/include/asm/paravirt_types.h
416
#define PVOP_ALT_VCALL0(array, op, alt, cond) \
arch/x86/include/asm/paravirt_types.h
417
__PVOP_ALT_VCALL(array, op, alt, cond)
arch/x86/include/asm/paravirt_types.h
419
#define PVOP_CALLEE0(rettype, array, op) \
arch/x86/include/asm/paravirt_types.h
420
__PVOP_CALLEESAVE(rettype, array, op)
arch/x86/include/asm/paravirt_types.h
421
#define PVOP_VCALLEE0(array, op) \
arch/x86/include/asm/paravirt_types.h
422
__PVOP_VCALLEESAVE(array, op)
arch/x86/include/asm/paravirt_types.h
423
#define PVOP_ALT_CALLEE0(rettype, array, op, alt, cond) \
arch/x86/include/asm/paravirt_types.h
424
__PVOP_ALT_CALLEESAVE(rettype, array, op, alt, cond)
arch/x86/include/asm/paravirt_types.h
425
#define PVOP_ALT_VCALLEE0(array, op, alt, cond) \
arch/x86/include/asm/paravirt_types.h
426
__PVOP_ALT_VCALLEESAVE(array, op, alt, cond)
arch/x86/include/asm/paravirt_types.h
429
#define PVOP_CALL1(rettype, array, op, arg1) \
arch/x86/include/asm/paravirt_types.h
430
__PVOP_CALL(rettype, array, op, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
431
#define PVOP_VCALL1(array, op, arg1) \
arch/x86/include/asm/paravirt_types.h
432
__PVOP_VCALL(array, op, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
433
#define PVOP_ALT_VCALL1(array, op, arg1, alt, cond) \
arch/x86/include/asm/paravirt_types.h
434
__PVOP_ALT_VCALL(array, op, alt, cond, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
436
#define PVOP_CALLEE1(rettype, array, op, arg1) \
arch/x86/include/asm/paravirt_types.h
437
__PVOP_CALLEESAVE(rettype, array, op, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
438
#define PVOP_VCALLEE1(array, op, arg1) \
arch/x86/include/asm/paravirt_types.h
439
__PVOP_VCALLEESAVE(array, op, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
440
#define PVOP_ALT_CALLEE1(rettype, array, op, arg1, alt, cond) \
arch/x86/include/asm/paravirt_types.h
441
__PVOP_ALT_CALLEESAVE(rettype, array, op, alt, cond, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
442
#define PVOP_ALT_VCALLEE1(array, op, arg1, alt, cond) \
arch/x86/include/asm/paravirt_types.h
443
__PVOP_ALT_VCALLEESAVE(array, op, alt, cond, PVOP_CALL_ARG1(arg1))
arch/x86/include/asm/paravirt_types.h
446
#define PVOP_CALL2(rettype, array, op, arg1, arg2) \
arch/x86/include/asm/paravirt_types.h
447
__PVOP_CALL(rettype, array, op, PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2))
arch/x86/include/asm/paravirt_types.h
448
#define PVOP_VCALL2(array, op, arg1, arg2) \
arch/x86/include/asm/paravirt_types.h
449
__PVOP_VCALL(array, op, PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2))
arch/x86/include/asm/paravirt_types.h
451
#define PVOP_CALL3(rettype, array, op, arg1, arg2, arg3) \
arch/x86/include/asm/paravirt_types.h
452
__PVOP_CALL(rettype, array, op, PVOP_CALL_ARG1(arg1), \
arch/x86/include/asm/paravirt_types.h
454
#define PVOP_VCALL3(array, op, arg1, arg2, arg3) \
arch/x86/include/asm/paravirt_types.h
455
__PVOP_VCALL(array, op, PVOP_CALL_ARG1(arg1), \
arch/x86/include/asm/paravirt_types.h
458
#define PVOP_CALL4(rettype, array, op, arg1, arg2, arg3, arg4) \
arch/x86/include/asm/paravirt_types.h
459
__PVOP_CALL(rettype, array, op, \
arch/x86/include/asm/paravirt_types.h
462
#define PVOP_VCALL4(array, op, arg1, arg2, arg3, arg4) \
arch/x86/include/asm/paravirt_types.h
463
__PVOP_VCALL(array, op, PVOP_CALL_ARG1(arg1), PVOP_CALL_ARG2(arg2), \
arch/x86/include/asm/percpu.h
121
#define __pcpu_op_1(op) op "b "
arch/x86/include/asm/percpu.h
122
#define __pcpu_op_2(op) op "w "
arch/x86/include/asm/percpu.h
123
#define __pcpu_op_4(op) op "l "
arch/x86/include/asm/percpu.h
124
#define __pcpu_op_8(op) op "q "
arch/x86/include/asm/percpu.h
199
#define percpu_unary_op(size, qual, op, _var) \
arch/x86/include/asm/percpu.h
201
asm qual (__pcpu_op_##size(op) __percpu_arg([var]) \
arch/x86/include/asm/percpu.h
205
#define percpu_binary_op(size, qual, op, _var, _val) \
arch/x86/include/asm/percpu.h
214
asm qual (__pcpu_op_##size(op) "%[val], " __percpu_arg([var]) \
arch/x86/include/asm/rmwcc.h
18
#define GEN_UNARY_RMWcc_4(op, var, cc, arg0) \
arch/x86/include/asm/rmwcc.h
19
__GEN_RMWcc(op " " arg0, var, cc, __CLOBBERS_MEM())
arch/x86/include/asm/rmwcc.h
21
#define GEN_UNARY_RMWcc_3(op, var, cc) \
arch/x86/include/asm/rmwcc.h
22
GEN_UNARY_RMWcc_4(op, var, cc, "%[var]")
arch/x86/include/asm/rmwcc.h
26
#define GEN_BINARY_RMWcc_6(op, var, cc, vcon, _val, arg0) \
arch/x86/include/asm/rmwcc.h
27
__GEN_RMWcc(op " %[val], " arg0, var, cc, \
arch/x86/include/asm/rmwcc.h
30
#define GEN_BINARY_RMWcc_5(op, var, cc, vcon, val) \
arch/x86/include/asm/rmwcc.h
31
GEN_BINARY_RMWcc_6(op, var, cc, vcon, val, "%[var]")
arch/x86/include/asm/rmwcc.h
35
#define GEN_UNARY_SUFFIXED_RMWcc(op, suffix, var, cc, clobbers...) \
arch/x86/include/asm/rmwcc.h
36
__GEN_RMWcc(op " %[var]\n\t" suffix, var, cc, \
arch/x86/include/asm/rmwcc.h
39
#define GEN_BINARY_SUFFIXED_RMWcc(op, suffix, var, cc, vcon, _val, clobbers...)\
arch/x86/include/asm/rmwcc.h
40
__GEN_RMWcc(op " %[val], %[var]\n\t" suffix, var, cc, \
arch/x86/include/asm/sev-common.h
96
#define GHCB_MSR_PSC_REQ_GFN(gfn, op) \
arch/x86/include/asm/sev-common.h
98
(((u64)((op) & 0xf) << 52) | \
arch/x86/include/asm/sev.h
577
enum psc_op op;
arch/x86/include/asm/static_call.h
66
extern bool __static_call_fixup(void *tramp, u8 op, void *dest);
arch/x86/include/asm/xen/hypercall.h
273
HYPERVISOR_mmuext_op(struct mmuext_op *op, int count,
arch/x86/include/asm/xen/hypercall.h
276
return _hypercall4(int, mmuext_op, op, count, success_count, domid);
arch/x86/include/asm/xen/hypercall.h
325
mcl->op = __HYPERVISOR_fpu_taskswitch;
arch/x86/include/asm/xen/hypercall.h
335
mcl->op = __HYPERVISOR_update_va_mapping;
arch/x86/include/asm/xen/hypercall.h
347
mcl->op = __HYPERVISOR_update_descriptor;
arch/x86/include/asm/xen/hypercall.h
358
mcl->op = __HYPERVISOR_mmu_update;
arch/x86/include/asm/xen/hypercall.h
368
MULTI_mmuext_op(struct multicall_entry *mcl, struct mmuext_op *op, int count,
arch/x86/include/asm/xen/hypercall.h
371
mcl->op = __HYPERVISOR_mmuext_op;
arch/x86/include/asm/xen/hypercall.h
372
mcl->args[0] = (unsigned long)op;
arch/x86/include/asm/xen/hypercall.h
384
mcl->op = __HYPERVISOR_stack_switch;
arch/x86/include/asm/xen/hypercall.h
414
HYPERVISOR_platform_op(struct xen_platform_op *op)
arch/x86/include/asm/xen/hypercall.h
416
op->interface_version = XENPF_INTERFACE_VERSION;
arch/x86/include/asm/xen/hypercall.h
417
return _hypercall1(int, platform_op, op);
arch/x86/include/asm/xen/hypercall.h
489
HYPERVISOR_hvm_op(int op, void *arg)
arch/x86/include/asm/xen/hypercall.h
491
return _hypercall2(unsigned long, hvm_op, op, arg);
arch/x86/include/asm/xen/hypercall.h
495
HYPERVISOR_xenpmu_op(unsigned int op, void *arg)
arch/x86/include/asm/xen/hypercall.h
497
return _hypercall2(int, xenpmu_op, op, arg);
arch/x86/include/asm/xor.h
52
#define BLK64(pf, op, i) \
arch/x86/include/asm/xor.h
54
op(i, 0) \
arch/x86/include/asm/xor.h
55
op(i + 1, 1) \
arch/x86/include/asm/xor.h
56
op(i + 2, 2) \
arch/x86/include/asm/xor.h
57
op(i + 3, 3)
arch/x86/include/uapi/asm/kvm_para.h
115
__u32 op;
arch/x86/kernel/alternative.c
1084
u8 op;
arch/x86/kernel/alternative.c
1090
op = insn.opcode.bytes[0];
arch/x86/kernel/alternative.c
1091
if (op == JMP32_INSN_OPCODE)
arch/x86/kernel/alternative.c
1094
if (__static_call_fixup(addr, op, dest) ||
arch/x86/kernel/alternative.c
1728
u8 op;
arch/x86/kernel/alternative.c
1748
op = insn.opcode.bytes[0];
arch/x86/kernel/alternative.c
1749
if (op != CALL_INSN_OPCODE && op != JMP32_INSN_OPCODE) {
arch/x86/kernel/alternative.c
1762
ret = emit_indirect(op, 11, bytes + fineibt_paranoid_ind, len);
arch/x86/kernel/alternative.c
732
static int emit_indirect(int op, int reg, u8 *bytes, int len)
arch/x86/kernel/alternative.c
744
switch (op) {
arch/x86/kernel/alternative.c
787
u8 op = insn->opcode.bytes[0];
arch/x86/kernel/alternative.c
795
bytes[i++] = op;
arch/x86/kernel/alternative.c
796
op = insn->opcode.bytes[1];
arch/x86/kernel/alternative.c
803
switch (op) {
arch/x86/kernel/alternative.c
805
__text_gen_insn(bytes+i, op, addr+i,
arch/x86/kernel/alternative.c
813
__text_gen_insn(bytes+i, op, addr+i,
arch/x86/kernel/alternative.c
891
u8 op, cc;
arch/x86/kernel/alternative.c
910
op = insn->opcode.bytes[0];
arch/x86/kernel/alternative.c
933
op = JMP32_INSN_OPCODE;
arch/x86/kernel/alternative.c
954
ret = emit_indirect(op, reg, bytes + i, insn->length - i);
arch/x86/kernel/cpu/cacheinfo.c
278
unsigned int eax, ebx, ecx, edx, op;
arch/x86/kernel/cpu/cacheinfo.c
283
op = (c->x86_vendor == X86_VENDOR_AMD || c->x86_vendor == X86_VENDOR_HYGON) ? 0x8000001d : 4;
arch/x86/kernel/cpu/cacheinfo.c
286
cpuid_count(op, i, &eax, &ebx, &ecx, &edx);
arch/x86/kernel/fpu/xstate.h
128
#define XSTATE_OP(op, st, lmask, hmask, err) \
arch/x86/kernel/fpu/xstate.h
129
asm volatile("1:" op "\n\t" \
arch/x86/kernel/ftrace.c
300
char op[3];
arch/x86/kernel/ftrace.c
393
if (WARN_ON(memcmp(op_ptr.op, op_ref, 3) != 0))
arch/x86/kernel/ftrace.c
653
struct ftrace_ops *op, struct ftrace_regs *fregs)
arch/x86/kernel/kprobes/core.c
110
__synthesize_relative_insn(void *dest, void *from, void *to, u8 op)
arch/x86/kernel/kprobes/core.c
113
u8 op;
arch/x86/kernel/kprobes/core.c
119
insn->op = op;
arch/x86/kernel/kprobes/opt.c
173
optimized_callback(struct optimized_kprobe *op, struct pt_regs *regs)
arch/x86/kernel/kprobes/opt.c
176
if (kprobe_disabled(&op->kp))
arch/x86/kernel/kprobes/opt.c
181
kprobes_inc_nmissed_count(&op->kp);
arch/x86/kernel/kprobes/opt.c
191
regs->ip = (unsigned long)op->kp.addr + INT3_INSN_SIZE;
arch/x86/kernel/kprobes/opt.c
194
__this_cpu_write(current_kprobe, &op->kp);
arch/x86/kernel/kprobes/opt.c
196
opt_pre_handler(&op->kp, regs);
arch/x86/kernel/kprobes/opt.c
337
int arch_check_optimized_kprobe(struct optimized_kprobe *op)
arch/x86/kernel/kprobes/opt.c
342
for (i = 1; i < op->optinsn.size; i++) {
arch/x86/kernel/kprobes/opt.c
343
p = get_kprobe(op->kp.addr + i);
arch/x86/kernel/kprobes/opt.c
352
int arch_within_optimized_kprobe(struct optimized_kprobe *op,
arch/x86/kernel/kprobes/opt.c
355
return (op->kp.addr <= addr &&
arch/x86/kernel/kprobes/opt.c
356
op->kp.addr + op->optinsn.size > addr);
arch/x86/kernel/kprobes/opt.c
361
void __arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty)
arch/x86/kernel/kprobes/opt.c
363
u8 *slot = op->optinsn.insn;
arch/x86/kernel/kprobes/opt.c
365
int len = TMPL_END_IDX + op->optinsn.size + JMP32_INSN_SIZE;
arch/x86/kernel/kprobes/opt.c
372
op->optinsn.insn = NULL;
arch/x86/kernel/kprobes/opt.c
373
op->optinsn.size = 0;
arch/x86/kernel/kprobes/opt.c
377
void arch_remove_optimized_kprobe(struct optimized_kprobe *op)
arch/x86/kernel/kprobes/opt.c
379
__arch_remove_optimized_kprobe(op, 1);
arch/x86/kernel/kprobes/opt.c
387
int arch_prepare_optimized_kprobe(struct optimized_kprobe *op,
arch/x86/kernel/kprobes/opt.c
39
struct optimized_kprobe *op;
arch/x86/kernel/kprobes/opt.c
394
if (!can_optimize((unsigned long)op->kp.addr))
arch/x86/kernel/kprobes/opt.c
401
op->optinsn.insn = slot = get_optinsn_slot();
arch/x86/kernel/kprobes/opt.c
411
rel = (long)slot - (long)op->kp.addr + JMP32_INSN_SIZE;
arch/x86/kernel/kprobes/opt.c
421
ret = copy_optimized_instructions(buf + TMPL_END_IDX, op->kp.addr,
arch/x86/kernel/kprobes/opt.c
425
op->optinsn.size = ret;
arch/x86/kernel/kprobes/opt.c
426
len = TMPL_END_IDX + op->optinsn.size;
arch/x86/kernel/kprobes/opt.c
431
synthesize_set_arg1(buf + TMPL_MOVE_IDX, (unsigned long)op);
arch/x86/kernel/kprobes/opt.c
439
(u8 *)op->kp.addr + op->optinsn.size);
arch/x86/kernel/kprobes/opt.c
457
__arch_remove_optimized_kprobe(op, 0);
arch/x86/kernel/kprobes/opt.c
471
struct optimized_kprobe *op, *tmp;
arch/x86/kernel/kprobes/opt.c
474
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/x86/kernel/kprobes/opt.c
475
s32 rel = (s32)((long)op->optinsn.insn -
arch/x86/kernel/kprobes/opt.c
476
((long)op->kp.addr + JMP32_INSN_SIZE));
arch/x86/kernel/kprobes/opt.c
478
WARN_ON(kprobe_disabled(&op->kp));
arch/x86/kernel/kprobes/opt.c
48
op = container_of(kp, struct optimized_kprobe, kp);
arch/x86/kernel/kprobes/opt.c
481
memcpy(op->optinsn.copied_insn, op->kp.addr + INT3_INSN_SIZE,
arch/x86/kernel/kprobes/opt.c
487
smp_text_poke_single(op->kp.addr, insn_buff, JMP32_INSN_SIZE, NULL);
arch/x86/kernel/kprobes/opt.c
489
list_del_init(&op->list);
arch/x86/kernel/kprobes/opt.c
50
if (list_empty(&op->list) || optprobe_queued_unopt(op))
arch/x86/kernel/kprobes/opt.c
500
void arch_unoptimize_kprobe(struct optimized_kprobe *op)
arch/x86/kernel/kprobes/opt.c
504
u8 *addr = op->kp.addr;
arch/x86/kernel/kprobes/opt.c
506
memcpy(old, op->kp.addr, JMP32_INSN_SIZE);
arch/x86/kernel/kprobes/opt.c
508
op->optinsn.copied_insn,
arch/x86/kernel/kprobes/opt.c
518
perf_event_text_poke(op->kp.addr, old, JMP32_INSN_SIZE, new, JMP32_INSN_SIZE);
arch/x86/kernel/kprobes/opt.c
528
struct optimized_kprobe *op, *tmp;
arch/x86/kernel/kprobes/opt.c
530
list_for_each_entry_safe(op, tmp, oplist, list) {
arch/x86/kernel/kprobes/opt.c
531
arch_unoptimize_kprobe(op);
arch/x86/kernel/kprobes/opt.c
532
list_move(&op->list, done_list);
arch/x86/kernel/kprobes/opt.c
538
struct optimized_kprobe *op;
arch/x86/kernel/kprobes/opt.c
542
op = container_of(p, struct optimized_kprobe, kp);
arch/x86/kernel/kprobes/opt.c
544
regs->ip = (unsigned long)op->optinsn.insn + TMPL_END_IDX;
arch/x86/kernel/kprobes/opt.c
68
memcpy(buf + 1, op->optinsn.copied_insn, DISP32_SIZE);
arch/x86/kernel/kprobes/opt.c
71
memcpy(buf, op->optinsn.copied_insn + offs, DISP32_SIZE - offs);
arch/x86/kernel/static_call.c
106
__text_gen_insn(buf+1, op, insn+1, func, 5);
arch/x86/kernel/static_call.c
204
bool __static_call_fixup(void *tramp, u8 op, void *dest)
arch/x86/kernel/static_call.c
225
if (op == RET_INSN_OPCODE || dest == &__x86_return_thunk)
arch/x86/kernel/static_call.c
64
u8 op, buf[6];
arch/x86/kernel/static_call.c
66
if ((type == JMP || type == RET) && (op = __is_Jcc(insn)))
arch/x86/kernel/uprobes.c
1050
u8 op;
arch/x86/kvm/emulate.c
1019
static void fetch_register_operand(struct operand *op)
arch/x86/kvm/emulate.c
1021
switch (op->bytes) {
arch/x86/kvm/emulate.c
1023
op->val = *(u8 *)op->addr.reg;
arch/x86/kvm/emulate.c
1026
op->val = *(u16 *)op->addr.reg;
arch/x86/kvm/emulate.c
1029
op->val = *(u32 *)op->addr.reg;
arch/x86/kvm/emulate.c
1032
op->val = *(u64 *)op->addr.reg;
arch/x86/kvm/emulate.c
1035
op->orig_val = op->val;
arch/x86/kvm/emulate.c
1082
struct operand *op, int reg)
arch/x86/kvm/emulate.c
1085
op->type = OP_YMM;
arch/x86/kvm/emulate.c
1086
op->bytes = 32;
arch/x86/kvm/emulate.c
1087
op->addr.xmm = reg;
arch/x86/kvm/emulate.c
1088
kvm_read_avx_reg(reg, &op->vec_val2);
arch/x86/kvm/emulate.c
1092
op->type = OP_XMM;
arch/x86/kvm/emulate.c
1093
op->bytes = 16;
arch/x86/kvm/emulate.c
1094
op->addr.xmm = reg;
arch/x86/kvm/emulate.c
1095
kvm_read_sse_reg(reg, &op->vec_val);
arch/x86/kvm/emulate.c
1100
op->type = OP_MM;
arch/x86/kvm/emulate.c
1101
op->bytes = 8;
arch/x86/kvm/emulate.c
1102
op->addr.mm = reg;
arch/x86/kvm/emulate.c
1106
op->type = OP_REG;
arch/x86/kvm/emulate.c
1107
op->bytes = (ctxt->d & ByteOp) ? 1 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
1108
op->addr.reg = decode_register(ctxt, reg, ctxt->d & ByteOp);
arch/x86/kvm/emulate.c
1109
fetch_register_operand(op);
arch/x86/kvm/emulate.c
1113
struct operand *op)
arch/x86/kvm/emulate.c
1122
__decode_register_operand(ctxt, op, reg);
arch/x86/kvm/emulate.c
1132
struct operand *op)
arch/x86/kvm/emulate.c
1149
__decode_register_operand(ctxt, op, ctxt->modrm_rm);
arch/x86/kvm/emulate.c
1153
op->type = OP_MEM;
arch/x86/kvm/emulate.c
1243
op->addr.mem.ea = modrm_ea;
arch/x86/kvm/emulate.c
1252
struct operand *op)
arch/x86/kvm/emulate.c
1256
op->type = OP_MEM;
arch/x86/kvm/emulate.c
1259
op->addr.mem.ea = insn_fetch(u16, ctxt);
arch/x86/kvm/emulate.c
1262
op->addr.mem.ea = insn_fetch(u32, ctxt);
arch/x86/kvm/emulate.c
1265
op->addr.mem.ea = insn_fetch(u64, ctxt);
arch/x86/kvm/emulate.c
1754
static void write_register_operand(struct operand *op)
arch/x86/kvm/emulate.c
1756
return assign_register(op->addr.reg, op->val, op->bytes);
arch/x86/kvm/emulate.c
1759
static int writeback(struct x86_emulate_ctxt *ctxt, struct operand *op)
arch/x86/kvm/emulate.c
1761
switch (op->type) {
arch/x86/kvm/emulate.c
1763
write_register_operand(op);
arch/x86/kvm/emulate.c
1768
op->addr.mem,
arch/x86/kvm/emulate.c
1769
&op->orig_val,
arch/x86/kvm/emulate.c
1770
&op->val,
arch/x86/kvm/emulate.c
1771
op->bytes);
arch/x86/kvm/emulate.c
1774
op->addr.mem,
arch/x86/kvm/emulate.c
1775
&op->val,
arch/x86/kvm/emulate.c
1776
op->bytes);
arch/x86/kvm/emulate.c
1779
op->addr.mem,
arch/x86/kvm/emulate.c
1780
op->data,
arch/x86/kvm/emulate.c
1781
op->bytes * op->count);
arch/x86/kvm/emulate.c
1784
kvm_write_sse_reg(op->addr.xmm, &op->vec_val);
arch/x86/kvm/emulate.c
1788
memset(op->valptr + 16, 0, 16);
arch/x86/kvm/emulate.c
1791
kvm_write_avx_reg(op->addr.xmm, &op->vec_val2);
arch/x86/kvm/emulate.c
1794
kvm_write_mmx_reg(op->addr.mm, &op->mm_val);
arch/x86/kvm/emulate.c
219
struct opcode op[8];
arch/x86/kvm/emulate.c
277
#define EM_ASM_START(op) \
arch/x86/kvm/emulate.c
278
static int em_##op(struct x86_emulate_ctxt *ctxt) \
arch/x86/kvm/emulate.c
297
#define __EM_ASM_1(op, dst) \
arch/x86/kvm/emulate.c
298
__EM_ASM(#op " %%" #dst " \n\t")
arch/x86/kvm/emulate.c
2997
struct operand *op)
arch/x86/kvm/emulate.c
2999
int df = (ctxt->eflags & X86_EFLAGS_DF) ? -op->count : op->count;
arch/x86/kvm/emulate.c
300
#define __EM_ASM_1_EX(op, dst) \
arch/x86/kvm/emulate.c
3001
register_address_increment(ctxt, reg, df * op->bytes);
arch/x86/kvm/emulate.c
3002
op->addr.mem.ea = register_address(ctxt, reg);
arch/x86/kvm/emulate.c
301
__EM_ASM(#op " %%" #dst " \n\t" \
arch/x86/kvm/emulate.c
304
#define __EM_ASM_2(op, dst, src) \
arch/x86/kvm/emulate.c
305
__EM_ASM(#op " %%" #src ", %%" #dst " \n\t")
arch/x86/kvm/emulate.c
307
#define __EM_ASM_3(op, dst, src, src2) \
arch/x86/kvm/emulate.c
308
__EM_ASM(#op " %%" #src2 ", %%" #src ", %%" #dst " \n\t")
arch/x86/kvm/emulate.c
317
#define EM_ASM_1(op) \
arch/x86/kvm/emulate.c
318
EM_ASM_START(op) \
arch/x86/kvm/emulate.c
319
case 1: __EM_ASM_1(op##b, al); break; \
arch/x86/kvm/emulate.c
320
case 2: __EM_ASM_1(op##w, ax); break; \
arch/x86/kvm/emulate.c
321
case 4: __EM_ASM_1(op##l, eax); break; \
arch/x86/kvm/emulate.c
322
ON64(case 8: __EM_ASM_1(op##q, rax); break;) \
arch/x86/kvm/emulate.c
326
#define EM_ASM_1SRC2(op, name) \
arch/x86/kvm/emulate.c
328
case 1: __EM_ASM_1(op##b, cl); break; \
arch/x86/kvm/emulate.c
329
case 2: __EM_ASM_1(op##w, cx); break; \
arch/x86/kvm/emulate.c
330
case 4: __EM_ASM_1(op##l, ecx); break; \
arch/x86/kvm/emulate.c
331
ON64(case 8: __EM_ASM_1(op##q, rcx); break;) \
arch/x86/kvm/emulate.c
335
#define EM_ASM_1SRC2EX(op, name) \
arch/x86/kvm/emulate.c
337
case 1: __EM_ASM_1_EX(op##b, cl); break; \
arch/x86/kvm/emulate.c
338
case 2: __EM_ASM_1_EX(op##w, cx); break; \
arch/x86/kvm/emulate.c
339
case 4: __EM_ASM_1_EX(op##l, ecx); break; \
arch/x86/kvm/emulate.c
340
ON64(case 8: __EM_ASM_1_EX(op##q, rcx); break;) \
arch/x86/kvm/emulate.c
344
#define EM_ASM_2(op) \
arch/x86/kvm/emulate.c
345
EM_ASM_START(op) \
arch/x86/kvm/emulate.c
346
case 1: __EM_ASM_2(op##b, al, dl); break; \
arch/x86/kvm/emulate.c
347
case 2: __EM_ASM_2(op##w, ax, dx); break; \
arch/x86/kvm/emulate.c
348
case 4: __EM_ASM_2(op##l, eax, edx); break; \
arch/x86/kvm/emulate.c
349
ON64(case 8: __EM_ASM_2(op##q, rax, rdx); break;) \
arch/x86/kvm/emulate.c
353
#define EM_ASM_2R(op, name) \
arch/x86/kvm/emulate.c
355
case 1: __EM_ASM_2(op##b, dl, al); break; \
arch/x86/kvm/emulate.c
356
case 2: __EM_ASM_2(op##w, dx, ax); break; \
arch/x86/kvm/emulate.c
357
case 4: __EM_ASM_2(op##l, edx, eax); break; \
arch/x86/kvm/emulate.c
358
ON64(case 8: __EM_ASM_2(op##q, rdx, rax); break;) \
arch/x86/kvm/emulate.c
362
#define EM_ASM_2W(op) \
arch/x86/kvm/emulate.c
363
EM_ASM_START(op) \
arch/x86/kvm/emulate.c
365
case 2: __EM_ASM_2(op##w, ax, dx); break; \
arch/x86/kvm/emulate.c
366
case 4: __EM_ASM_2(op##l, eax, edx); break; \
arch/x86/kvm/emulate.c
367
ON64(case 8: __EM_ASM_2(op##q, rax, rdx); break;) \
arch/x86/kvm/emulate.c
371
#define EM_ASM_2CL(op) \
arch/x86/kvm/emulate.c
372
EM_ASM_START(op) \
arch/x86/kvm/emulate.c
373
case 1: __EM_ASM_2(op##b, al, cl); break; \
arch/x86/kvm/emulate.c
374
case 2: __EM_ASM_2(op##w, ax, cl); break; \
arch/x86/kvm/emulate.c
375
case 4: __EM_ASM_2(op##l, eax, cl); break; \
arch/x86/kvm/emulate.c
376
ON64(case 8: __EM_ASM_2(op##q, rax, cl); break;) \
arch/x86/kvm/emulate.c
380
#define EM_ASM_3WCL(op) \
arch/x86/kvm/emulate.c
381
EM_ASM_START(op) \
arch/x86/kvm/emulate.c
383
case 2: __EM_ASM_3(op##w, ax, dx, cl); break; \
arch/x86/kvm/emulate.c
384
case 4: __EM_ASM_3(op##l, eax, edx, cl); break; \
arch/x86/kvm/emulate.c
385
ON64(case 8: __EM_ASM_3(op##q, rax, rdx, cl); break;) \
arch/x86/kvm/emulate.c
4564
static int decode_imm(struct x86_emulate_ctxt *ctxt, struct operand *op,
arch/x86/kvm/emulate.c
4569
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4570
op->bytes = size;
arch/x86/kvm/emulate.c
4571
op->addr.mem.ea = ctxt->_eip;
arch/x86/kvm/emulate.c
4573
switch (op->bytes) {
arch/x86/kvm/emulate.c
4575
op->val = insn_fetch(s8, ctxt);
arch/x86/kvm/emulate.c
4578
op->val = insn_fetch(s16, ctxt);
arch/x86/kvm/emulate.c
4581
op->val = insn_fetch(s32, ctxt);
arch/x86/kvm/emulate.c
4584
op->val = insn_fetch(s64, ctxt);
arch/x86/kvm/emulate.c
4588
switch (op->bytes) {
arch/x86/kvm/emulate.c
4590
op->val &= 0xff;
arch/x86/kvm/emulate.c
4593
op->val &= 0xffff;
arch/x86/kvm/emulate.c
4596
op->val &= 0xffffffff;
arch/x86/kvm/emulate.c
4604
static int decode_operand(struct x86_emulate_ctxt *ctxt, struct operand *op,
arch/x86/kvm/emulate.c
4611
decode_register_operand(ctxt, op);
arch/x86/kvm/emulate.c
4614
rc = decode_imm(ctxt, op, 1, false);
arch/x86/kvm/emulate.c
4619
*op = ctxt->memop;
arch/x86/kvm/emulate.c
4620
ctxt->memopp = op;
arch/x86/kvm/emulate.c
4623
op->orig_val = op->val;
arch/x86/kvm/emulate.c
4629
op->type = OP_REG;
arch/x86/kvm/emulate.c
4630
op->bytes = (ctxt->d & ByteOp) ? 1 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
4631
op->addr.reg = reg_rmw(ctxt, VCPU_REGS_RAX);
arch/x86/kvm/emulate.c
4632
fetch_register_operand(op);
arch/x86/kvm/emulate.c
4635
op->type = OP_REG;
arch/x86/kvm/emulate.c
4636
op->bytes = (ctxt->d & ByteOp) ? 2 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
4637
op->addr.reg = reg_rmw(ctxt, VCPU_REGS_RAX);
arch/x86/kvm/emulate.c
4638
fetch_register_operand(op);
arch/x86/kvm/emulate.c
4642
op->type = OP_NONE;
arch/x86/kvm/emulate.c
4645
op->type = OP_REG;
arch/x86/kvm/emulate.c
4646
op->bytes = ctxt->op_bytes;
arch/x86/kvm/emulate.c
4647
op->addr.reg = reg_rmw(ctxt, VCPU_REGS_RDX);
arch/x86/kvm/emulate.c
4648
fetch_register_operand(op);
arch/x86/kvm/emulate.c
4651
op->type = OP_MEM;
arch/x86/kvm/emulate.c
4652
op->bytes = (ctxt->d & ByteOp) ? 1 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
4653
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4655
op->addr.mem.seg = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4656
op->val = 0;
arch/x86/kvm/emulate.c
4657
op->count = 1;
arch/x86/kvm/emulate.c
4660
op->type = OP_REG;
arch/x86/kvm/emulate.c
4661
op->bytes = 2;
arch/x86/kvm/emulate.c
4662
op->addr.reg = reg_rmw(ctxt, VCPU_REGS_RDX);
arch/x86/kvm/emulate.c
4663
fetch_register_operand(op);
arch/x86/kvm/emulate.c
4666
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4667
op->bytes = 1;
arch/x86/kvm/emulate.c
4668
op->val = reg_read(ctxt, VCPU_REGS_RCX) & 0xff;
arch/x86/kvm/emulate.c
4671
rc = decode_imm(ctxt, op, 1, true);
arch/x86/kvm/emulate.c
4674
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4675
op->bytes = 1;
arch/x86/kvm/emulate.c
4676
op->val = 1;
arch/x86/kvm/emulate.c
4679
rc = decode_imm(ctxt, op, imm_size(ctxt), true);
arch/x86/kvm/emulate.c
4682
rc = decode_imm(ctxt, op, ctxt->op_bytes, true);
arch/x86/kvm/emulate.c
4699
rc = decode_imm(ctxt, op, 2, false);
arch/x86/kvm/emulate.c
4702
rc = decode_imm(ctxt, op, imm_size(ctxt), false);
arch/x86/kvm/emulate.c
4705
op->type = OP_MEM;
arch/x86/kvm/emulate.c
4706
op->bytes = (ctxt->d & ByteOp) ? 1 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
4707
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4709
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
4710
op->val = 0;
arch/x86/kvm/emulate.c
4711
op->count = 1;
arch/x86/kvm/emulate.c
4714
op->type = OP_MEM;
arch/x86/kvm/emulate.c
4715
op->bytes = (ctxt->d & ByteOp) ? 1 : ctxt->op_bytes;
arch/x86/kvm/emulate.c
4716
op->addr.mem.ea =
arch/x86/kvm/emulate.c
4720
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
4721
op->val = 0;
arch/x86/kvm/emulate.c
4724
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4725
op->addr.mem.ea = ctxt->_eip;
arch/x86/kvm/emulate.c
4726
op->bytes = ctxt->op_bytes + 2;
arch/x86/kvm/emulate.c
4727
insn_fetch_arr(op->valptr, op->bytes, ctxt);
arch/x86/kvm/emulate.c
4733
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4734
op->val = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4737
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4738
op->val = VCPU_SREG_CS;
arch/x86/kvm/emulate.c
4741
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4742
op->val = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
4745
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4746
op->val = VCPU_SREG_DS;
arch/x86/kvm/emulate.c
4749
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4750
op->val = VCPU_SREG_FS;
arch/x86/kvm/emulate.c
4753
op->type = OP_IMM;
arch/x86/kvm/emulate.c
4754
op->val = VCPU_SREG_GS;
arch/x86/kvm/emulate.c
4759
op->type = OP_NONE; /* Disable writeback. */
arch/x86/kvm/emulate.c
5036
opcode = opcode.u.esc->op[(ctxt->modrm >> 3) & 7];
arch/x86/kvm/emulate.c
5241
static void fetch_possible_mmx_operand(struct operand *op)
arch/x86/kvm/emulate.c
5243
if (op->type == OP_MM)
arch/x86/kvm/emulate.c
5244
kvm_read_mmx_reg(op->addr.mm, &op->mm_val);
arch/x86/kvm/svm/sev.c
3734
u8 op = GHCB_MSR_PSC_REQ_TO_OP(ghcb_msr);
arch/x86/kvm/svm/sev.c
3737
if (op != SNP_PAGE_STATE_PRIVATE && op != SNP_PAGE_STATE_SHARED) {
arch/x86/kvm/svm/sev.c
3758
vcpu->run->hypercall.args[2] = (op == SNP_PAGE_STATE_PRIVATE)
arch/x86/kvm/vmx/tdx.c
71
void tdh_vp_wr_failed(struct vcpu_tdx *tdx, char *uclass, char *op, u32 field,
arch/x86/kvm/vmx/tdx.c
75
pr_err("TDH_VP_WR[%s.0x%x]%s0x%llx failed: 0x%llx\n", uclass, field, op, val, err);
arch/x86/kvm/vmx/tdx.h
80
void tdh_vp_wr_failed(struct vcpu_tdx *tdx, char *uclass, char *op, u32 field,
arch/x86/kvm/x86.c
524
const char *op = rw == MSR_TYPE_W ? "wrmsr" : "rdmsr";
arch/x86/kvm/x86.c
551
op, msr, *data);
arch/x86/kvm/x86.c
556
kvm_pr_unimpl("ignored %s: 0x%x data 0x%llx\n", op, msr, *data);
arch/x86/lib/insn.c
280
insn_byte_t op;
arch/x86/lib/insn.c
290
op = get_next(insn_byte_t, insn);
arch/x86/lib/insn.c
291
insn_set_byte(opcode, 0, op);
arch/x86/lib/insn.c
301
insn->attr = inat_get_xop_attribute(op, m);
arch/x86/lib/insn.c
312
insn->attr = inat_get_avx_attribute(op, m, p);
arch/x86/lib/insn.c
335
insn->attr = inat_get_escape_attribute(op, pfx_id, esc_attr);
arch/x86/lib/insn.c
337
insn->attr = inat_get_opcode_attribute(op);
arch/x86/lib/insn.c
342
insn->attr = inat_get_opcode_attribute(op);
arch/x86/lib/insn.c
350
op = get_next(insn_byte_t, insn);
arch/x86/lib/insn.c
351
opcode->bytes[opcode->nbytes++] = op;
arch/x86/lib/insn.c
353
insn->attr = inat_get_escape_attribute(op, pfx_id, insn->attr);
arch/x86/net/bpf_jit_comp.c
1569
static void emit_shiftx(u8 **pprog, u32 dst_reg, u8 src_reg, bool is64, u8 op)
arch/x86/net/bpf_jit_comp.c
1575
emit_3vex(&prog, r, false, r, m, is64, src_reg, false, op);
arch/x86/net/bpf_jit_comp.c
2000
u8 op;
arch/x86/net/bpf_jit_comp.c
2004
op = 1; /* prefix 0x66 */
arch/x86/net/bpf_jit_comp.c
2007
op = 3; /* prefix 0xf2 */
arch/x86/net/bpf_jit_comp.c
2010
op = 2; /* prefix 0xf3 */
arch/x86/net/bpf_jit_comp.c
2014
emit_shiftx(&prog, dst_reg, src_reg, w, op);
arch/x86/net/bpf_jit_comp32.c
1422
static u8 get_cond_jmp_opcode(const u8 op, bool is_cmp_lo)
arch/x86/net/bpf_jit_comp32.c
1427
switch (op) {
arch/x86/net/bpf_jit_comp32.c
435
static inline void emit_ia32_div_mod_r(const u8 op, const u8 dst, const u8 src,
arch/x86/net/bpf_jit_comp32.c
462
if (op == BPF_MOD) {
arch/x86/net/bpf_jit_comp32.c
482
static inline void emit_ia32_shift_r(const u8 op, const u8 dst, const u8 src,
arch/x86/net/bpf_jit_comp32.c
501
switch (op) {
arch/x86/net/bpf_jit_comp32.c
523
static inline void emit_ia32_alu_r(const bool is64, const bool hi, const u8 op,
arch/x86/net/bpf_jit_comp32.c
540
switch (BPF_OP(op)) {
arch/x86/net/bpf_jit_comp32.c
577
static inline void emit_ia32_alu_r64(const bool is64, const u8 op,
arch/x86/net/bpf_jit_comp32.c
584
emit_ia32_alu_r(is64, false, op, dst_lo, src_lo, dstk, sstk, &prog);
arch/x86/net/bpf_jit_comp32.c
586
emit_ia32_alu_r(is64, true, op, dst_hi, src_hi, dstk, sstk,
arch/x86/net/bpf_jit_comp32.c
597
static inline void emit_ia32_alu_i(const bool is64, const bool hi, const u8 op,
arch/x86/net/bpf_jit_comp32.c
614
switch (op) {
arch/x86/net/bpf_jit_comp32.c
677
static inline void emit_ia32_alu_i64(const bool is64, const u8 op,
arch/x86/net/bpf_jit_comp32.c
688
emit_ia32_alu_i(is64, false, op, dst_lo, val, dstk, &prog);
arch/x86/net/bpf_jit_comp32.c
690
emit_ia32_alu_i(is64, true, op, dst_hi, hi, dstk, &prog);
arch/x86/platform/intel/iosf_mbi.c
34
static inline u32 iosf_mbi_form_mcr(u8 op, u8 port, u8 offset)
arch/x86/platform/intel/iosf_mbi.c
36
return (op << 24) | (port << 16) | (offset << 8) | MBI_ENABLE;
arch/x86/xen/apic.c
41
struct xen_platform_op op = {
arch/x86/xen/apic.c
56
op.u.pcpu_info.xen_cpuid = cpu;
arch/x86/xen/apic.c
58
ret = HYPERVISOR_platform_op(&op);
arch/x86/xen/apic.c
60
op.u.pcpu_info.apic_id = BAD_APICID;
arch/x86/xen/apic.c
62
return op.u.pcpu_info.apic_id << 24;
arch/x86/xen/efi.c
49
struct xen_platform_op op = {
arch/x86/xen/efi.c
56
union xenpf_efi_info *info = &op.u.firmware_info.u.efi_info;
arch/x86/xen/efi.c
58
if (!xen_initial_domain() || HYPERVISOR_platform_op(&op) < 0)
arch/x86/xen/efi.c
67
op.cmd = XENPF_firmware_info;
arch/x86/xen/efi.c
68
op.u.firmware_info.type = XEN_FW_EFI_INFO;
arch/x86/xen/efi.c
69
op.u.firmware_info.index = XEN_FW_EFI_VENDOR;
arch/x86/xen/efi.c
73
if (HYPERVISOR_platform_op(&op) == 0) {
arch/x86/xen/efi.c
79
op.cmd = XENPF_firmware_info;
arch/x86/xen/efi.c
80
op.u.firmware_info.type = XEN_FW_EFI_INFO;
arch/x86/xen/efi.c
81
op.u.firmware_info.index = XEN_FW_EFI_VERSION;
arch/x86/xen/efi.c
83
if (HYPERVISOR_platform_op(&op) == 0)
arch/x86/xen/efi.c
86
op.cmd = XENPF_firmware_info;
arch/x86/xen/efi.c
87
op.u.firmware_info.type = XEN_FW_EFI_INFO;
arch/x86/xen/efi.c
88
op.u.firmware_info.index = XEN_FW_EFI_RT_VERSION;
arch/x86/xen/efi.c
90
if (HYPERVISOR_platform_op(&op) == 0)
arch/x86/xen/enlighten_pv.c
1262
struct xen_platform_op op;
arch/x86/xen/enlighten_pv.c
1271
op.cmd = XENPF_firmware_info;
arch/x86/xen/enlighten_pv.c
1273
op.u.firmware_info.type = XEN_FW_DISK_INFO;
arch/x86/xen/enlighten_pv.c
1277
op.u.firmware_info.index = nr;
arch/x86/xen/enlighten_pv.c
1279
set_xen_guest_handle(op.u.firmware_info.u.disk_info.edd_params,
arch/x86/xen/enlighten_pv.c
1281
ret = HYPERVISOR_platform_op(&op);
arch/x86/xen/enlighten_pv.c
1285
#define C(x) info->x = op.u.firmware_info.u.disk_info.x
arch/x86/xen/enlighten_pv.c
1296
op.u.firmware_info.type = XEN_FW_DISK_MBR_SIGNATURE;
arch/x86/xen/enlighten_pv.c
1298
op.u.firmware_info.index = nr;
arch/x86/xen/enlighten_pv.c
1299
ret = HYPERVISOR_platform_op(&op);
arch/x86/xen/enlighten_pv.c
1302
mbr_signature[nr] = op.u.firmware_info.u.disk_mbr_signature.mbr_signature;
arch/x86/xen/enlighten_pv.c
1522
struct xen_platform_op op = {
arch/x86/xen/enlighten_pv.c
1535
if (HYPERVISOR_platform_op(&op) == 0)
arch/x86/xen/enlighten_pv.c
1536
boot_params.kbd_status = op.u.firmware_info.u.kbd_shift_flags;
arch/x86/xen/enlighten_pv.c
174
struct xen_platform_op op = {
arch/x86/xen/enlighten_pv.c
192
op.u.read_memtype.reg = reg;
arch/x86/xen/enlighten_pv.c
193
if (HYPERVISOR_platform_op(&op))
arch/x86/xen/enlighten_pv.c
201
var[reg].base_lo = op.u.read_memtype.type;
arch/x86/xen/enlighten_pv.c
202
var[reg].base_lo |= op.u.read_memtype.mfn << PAGE_SHIFT;
arch/x86/xen/enlighten_pv.c
203
var[reg].base_hi = op.u.read_memtype.mfn >> (32 - PAGE_SHIFT);
arch/x86/xen/enlighten_pv.c
204
mask = ~((op.u.read_memtype.nr_mfns << PAGE_SHIFT) - 1);
arch/x86/xen/enlighten_pv.c
301
struct xen_platform_op op = {
arch/x86/xen/enlighten_pv.c
357
set_xen_guest_handle(op.u.set_pminfo.pdc, buf);
arch/x86/xen/enlighten_pv.c
359
if ((HYPERVISOR_platform_op(&op) == 0) &&
arch/x86/xen/enlighten_pv.c
535
struct mmuext_op *op;
arch/x86/xen/enlighten_pv.c
536
struct multicall_space mcs = xen_mc_entry(sizeof(*op));
arch/x86/xen/enlighten_pv.c
540
op = mcs.args;
arch/x86/xen/enlighten_pv.c
541
op->cmd = MMUEXT_SET_LDT;
arch/x86/xen/enlighten_pv.c
542
op->arg1.linear_addr = (unsigned long)addr;
arch/x86/xen/enlighten_pv.c
543
op->arg2.nr_ents = entries;
arch/x86/xen/enlighten_pv.c
545
MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
arch/x86/xen/enlighten_pvh.c
159
struct xen_platform_op op = {
arch/x86/xen/enlighten_pvh.c
162
int ret = HYPERVISOR_platform_op(&op);
arch/x86/xen/enlighten_pvh.c
165
xen_init_vga(&op.u.dom0_console,
arch/x86/xen/enlighten_pvh.c
167
sizeof(op.u.dom0_console)),
arch/x86/xen/mmu_pv.c
1059
struct mmuext_op op;
arch/x86/xen/mmu_pv.c
1061
op.cmd = cmd;
arch/x86/xen/mmu_pv.c
1062
op.arg1.mfn = pfn_to_mfn(pfn);
arch/x86/xen/mmu_pv.c
1063
if (HYPERVISOR_mmuext_op(&op, 1, NULL, DOMID_SELF))
arch/x86/xen/mmu_pv.c
1293
struct mmuext_op *op;
arch/x86/xen/mmu_pv.c
1298
mcs = xen_mc_entry(sizeof(*op));
arch/x86/xen/mmu_pv.c
1300
op = mcs.args;
arch/x86/xen/mmu_pv.c
1301
op->cmd = MMUEXT_TLB_FLUSH_LOCAL;
arch/x86/xen/mmu_pv.c
1302
MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
arch/x86/xen/mmu_pv.c
1311
struct mmuext_op *op;
arch/x86/xen/mmu_pv.c
1318
mcs = xen_mc_entry(sizeof(*op));
arch/x86/xen/mmu_pv.c
1319
op = mcs.args;
arch/x86/xen/mmu_pv.c
1320
op->cmd = MMUEXT_INVLPG_LOCAL;
arch/x86/xen/mmu_pv.c
1321
op->arg1.linear_addr = addr & PAGE_MASK;
arch/x86/xen/mmu_pv.c
1322
MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
arch/x86/xen/mmu_pv.c
1333
struct mmuext_op op;
arch/x86/xen/mmu_pv.c
1337
const size_t mc_entry_size = sizeof(args->op) +
arch/x86/xen/mmu_pv.c
1347
args->op.arg2.vcpumask = to_cpumask(args->mask);
arch/x86/xen/mmu_pv.c
1352
args->op.cmd = MMUEXT_TLB_FLUSH_MULTI;
arch/x86/xen/mmu_pv.c
1355
args->op.cmd = MMUEXT_INVLPG_MULTI;
arch/x86/xen/mmu_pv.c
1356
args->op.arg1.linear_addr = info->start;
arch/x86/xen/mmu_pv.c
1359
MULTI_mmuext_op(mcs.mc, &args->op, 1, NULL, DOMID_SELF);
arch/x86/xen/mmu_pv.c
1376
struct mmuext_op op;
arch/x86/xen/mmu_pv.c
1388
op.cmd = kernel ? MMUEXT_NEW_BASEPTR : MMUEXT_NEW_USER_BASEPTR;
arch/x86/xen/mmu_pv.c
1389
op.arg1.mfn = mfn;
arch/x86/xen/mmu_pv.c
1391
xen_extend_mmuext_op(&op);
arch/x86/xen/mmu_pv.c
1571
struct mmuext_op *op;
arch/x86/xen/mmu_pv.c
1573
mcs = __xen_mc_entry(sizeof(*op));
arch/x86/xen/mmu_pv.c
1574
op = mcs.args;
arch/x86/xen/mmu_pv.c
1575
op->cmd = cmd;
arch/x86/xen/mmu_pv.c
1576
op->arg1.mfn = pfn_to_mfn(pfn);
arch/x86/xen/mmu_pv.c
2430
struct mmuext_op *op;
arch/x86/xen/mmu_pv.c
2435
mcs = xen_mc_entry(sizeof(*op));
arch/x86/xen/mmu_pv.c
2437
op = mcs.args;
arch/x86/xen/mmu_pv.c
2438
op->cmd = MMUEXT_TLB_FLUSH_ALL;
arch/x86/xen/mmu_pv.c
2439
MULTI_mmuext_op(mcs.mc, op, 1, NULL, DOMID_SELF);
arch/x86/xen/mmu_pv.c
254
static void xen_extend_mmuext_op(const struct mmuext_op *op)
arch/x86/xen/mmu_pv.c
269
*u = *op;
arch/x86/xen/mmu_pv.c
733
struct mmuext_op op;
arch/x86/xen/mmu_pv.c
735
op.cmd = level;
arch/x86/xen/mmu_pv.c
736
op.arg1.mfn = pfn_to_mfn(pfn);
arch/x86/xen/mmu_pv.c
738
xen_extend_mmuext_op(&op);
arch/x86/xen/multicalls.c
121
unsigned int opidx = mcdb->entries[idx].op & 0xff;
arch/x86/xen/multicalls.c
125
mcdb->entries[idx].op, b->entries[idx].result,
arch/x86/xen/multicalls.c
176
mc->result = xen_single_call(mc->op, mc->args[0], mc->args[1],
arch/x86/xen/multicalls.c
199
b->entries[i].op,
arch/x86/xen/multicalls.c
254
struct multicall_space xen_mc_extend_args(unsigned long op, size_t size)
arch/x86/xen/multicalls.c
263
b->entries[b->mcidx - 1].op != op)) {
arch/x86/xen/multicalls.c
264
trace_xen_mc_extend_args(op, size, XEN_MC_XE_BAD_OP);
arch/x86/xen/multicalls.c
269
trace_xen_mc_extend_args(op, size, XEN_MC_XE_NO_SPACE);
arch/x86/xen/multicalls.c
279
trace_xen_mc_extend_args(op, size, XEN_MC_XE_OK);
arch/x86/xen/setup.c
800
int op;
arch/x86/xen/setup.c
813
op = xen_initial_domain() ?
arch/x86/xen/setup.c
816
rc = HYPERVISOR_memory_op(op, &memmap);
arch/x86/xen/time.c
105
struct xen_platform_op op;
arch/x86/xen/time.c
123
op.cmd = XENPF_settime64;
arch/x86/xen/time.c
124
op.u.settime64.mbz = 0;
arch/x86/xen/time.c
125
op.u.settime64.secs = now.tv_sec;
arch/x86/xen/time.c
126
op.u.settime64.nsecs = now.tv_nsec;
arch/x86/xen/time.c
127
op.u.settime64.system_time = xen_clocksource_read();
arch/x86/xen/time.c
129
op.cmd = XENPF_settime32;
arch/x86/xen/time.c
130
op.u.settime32.secs = now.tv_sec;
arch/x86/xen/time.c
131
op.u.settime32.nsecs = now.tv_nsec;
arch/x86/xen/time.c
132
op.u.settime32.system_time = xen_clocksource_read();
arch/x86/xen/time.c
135
ret = HYPERVISOR_platform_op(&op);
arch/x86/xen/trace.c
12
static const char *xen_hypercall_name(unsigned op)
arch/x86/xen/trace.c
14
if (op < ARRAY_SIZE(xen_hypercall_names) && xen_hypercall_names[op] != NULL)
arch/x86/xen/trace.c
15
return xen_hypercall_names[op];
arch/x86/xen/xen-ops.h
261
struct multicall_space xen_mc_extend_args(unsigned long op, size_t arg_size);
arch/xtensa/include/asm/atomic.h
105
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
118
#define ATOMIC_OP(op) \
arch/xtensa/include/asm/atomic.h
119
static inline void arch_atomic_##op(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
127
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
137
#define ATOMIC_OP_RETURN(op) \
arch/xtensa/include/asm/atomic.h
138
static inline int arch_atomic_##op##_return(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
146
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
149
" " #op " %[result], %[result], %[i]\n" \
arch/xtensa/include/asm/atomic.h
159
#define ATOMIC_FETCH_OP(op) \
arch/xtensa/include/asm/atomic.h
160
static inline int arch_atomic_fetch_##op(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
168
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
182
#define ATOMIC_OP(op) \
arch/xtensa/include/asm/atomic.h
183
static inline void arch_atomic_##op(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
190
" " #op " %[result], %[result], %[i]\n" \
arch/xtensa/include/asm/atomic.h
200
#define ATOMIC_OP_RETURN(op) \
arch/xtensa/include/asm/atomic.h
201
static inline int arch_atomic_##op##_return(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
208
" " #op " %[result], %[result], %[i]\n" \
arch/xtensa/include/asm/atomic.h
220
#define ATOMIC_FETCH_OP(op) \
arch/xtensa/include/asm/atomic.h
221
static inline int arch_atomic_fetch_##op(int i, atomic_t * v) \
arch/xtensa/include/asm/atomic.h
228
" " #op " %[tmp], %[result], %[i]\n" \
arch/xtensa/include/asm/atomic.h
243
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op) ATOMIC_OP_RETURN(op)
arch/xtensa/include/asm/atomic.h
254
#define ATOMIC_OPS(op) ATOMIC_OP(op) ATOMIC_FETCH_OP(op)
arch/xtensa/include/asm/atomic.h
58
#define ATOMIC_OP(op) \
arch/xtensa/include/asm/atomic.h
59
static inline void arch_atomic_##op(int i, atomic_t *v) \
arch/xtensa/include/asm/atomic.h
66
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
76
#define ATOMIC_OP_RETURN(op) \
arch/xtensa/include/asm/atomic.h
77
static inline int arch_atomic_##op##_return(int i, atomic_t *v) \
arch/xtensa/include/asm/atomic.h
84
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
88
" " #op " %[result], %[tmp], %[i]\n" \
arch/xtensa/include/asm/atomic.h
97
#define ATOMIC_FETCH_OP(op) \
arch/xtensa/include/asm/atomic.h
98
static inline int arch_atomic_fetch_##op(int i, atomic_t *v) \
arch/xtensa/include/asm/bitops.h
101
#define BIT_OP(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
102
static inline void arch_##op##_bit(unsigned int bit, volatile unsigned long *p)\
arch/xtensa/include/asm/bitops.h
120
#define TEST_AND_BIT_OP(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
122
arch_test_and_##op##_bit(unsigned int bit, volatile unsigned long *p) \
arch/xtensa/include/asm/bitops.h
144
#define BIT_OP(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
145
static inline void arch_##op##_bit(unsigned int bit, volatile unsigned long *p)\
arch/xtensa/include/asm/bitops.h
164
#define TEST_AND_BIT_OP(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
166
arch_test_and_##op##_bit(unsigned int bit, volatile unsigned long *p) \
arch/xtensa/include/asm/bitops.h
189
#define BIT_OP(op, insn, inv)
arch/xtensa/include/asm/bitops.h
190
#define TEST_AND_BIT_OP(op, insn, inv)
arch/xtensa/include/asm/bitops.h
196
#define BIT_OPS(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
197
BIT_OP(op, insn, inv) \
arch/xtensa/include/asm/bitops.h
198
TEST_AND_BIT_OP(op, insn, inv)
arch/xtensa/include/asm/futex.h
112
return futex_atomic_op_inuser_local(op, oparg, oval, uaddr);
arch/xtensa/include/asm/futex.h
73
static inline int arch_futex_atomic_op_inuser(int op, int oparg, int *oval,
arch/xtensa/include/asm/futex.h
82
switch (op) {
block/bio.c
1527
size_t len, enum req_op op)
block/bio.c
1536
bio_init(&bio, bdev, &bv, 1, op);
block/blk-core.c
1040
unsigned long bdev_start_io_acct(struct block_device *bdev, enum req_op op,
block/blk-core.c
1045
part_stat_local_inc(bdev, in_flight[op_is_write(op)]);
block/blk-core.c
1064
void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
block/blk-core.c
1067
const int sgrp = op_stat_group(op);
block/blk-core.c
1076
part_stat_local_dec(bdev, in_flight[op_is_write(op)]);
block/blk-core.c
124
inline const char *blk_op_str(enum req_op op)
block/blk-core.c
128
if (op < ARRAY_SIZE(blk_op_name) && blk_op_name[op])
block/blk-core.c
129
op_str = blk_op_name[op];
block/blk-map.c
368
enum req_op op = req_op(rq);
block/blk-map.c
398
if (op_is_write(op))
block/blk-map.c
407
if (op_is_write(op)) {
block/blk-mq-debugfs.c
118
char opbuf[16] = { }, *op;
block/blk-mq-debugfs.c
134
op = strstrip(opbuf);
block/blk-mq-debugfs.c
135
if (strcmp(op, "run") == 0) {
block/blk-mq-debugfs.c
137
} else if (strcmp(op, "start") == 0) {
block/blk-mq-debugfs.c
139
} else if (strcmp(op, "kick") == 0) {
block/blk-mq-debugfs.c
142
pr_err("%s: unsupported operation '%s'\n", __func__, op);
block/blk-mq-debugfs.c
266
const enum req_op op = req_op(rq);
block/blk-mq-debugfs.c
267
const char *op_str = blk_op_str(op);
block/blk-mq-debugfs.c
274
seq_printf(m, "%u", op);
block/blk-wbt.c
703
const enum req_op op = req_op(rq);
block/blk-wbt.c
705
if (op == REQ_OP_READ)
block/blk-wbt.c
707
else if (op_is_write(op))
block/blk-zoned.c
292
int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
block/blk-zoned.c
307
if (!op_is_zone_mgmt(op))
block/blk-zoned.c
325
if (op == REQ_OP_ZONE_RESET && sector == 0 && nr_sectors == capacity)
block/blk-zoned.c
329
bio = blk_next_bio(bio, bdev, 0, op | REQ_SYNC, GFP_KERNEL);
block/blk-zoned.c
440
enum req_op op;
block/blk-zoned.c
457
op = REQ_OP_ZONE_RESET;
block/blk-zoned.c
467
op = REQ_OP_ZONE_OPEN;
block/blk-zoned.c
470
op = REQ_OP_ZONE_CLOSE;
block/blk-zoned.c
473
op = REQ_OP_ZONE_FINISH;
block/blk-zoned.c
479
ret = blkdev_zone_mgmt(bdev, op, zrange.sector, zrange.nr_sectors);
block/blk.h
209
enum req_op op = req_op(rq);
block/blk.h
211
if (unlikely(op == REQ_OP_DISCARD))
block/blk.h
215
if (unlikely(op == REQ_OP_SECURE_ERASE))
block/blk.h
219
if (unlikely(op == REQ_OP_WRITE_ZEROES))
block/ioctl.c
22
struct blkpg_partition __user *upart, int op)
block/ioctl.c
38
if (op == BLKPG_DEL_PARTITION)
block/ioctl.c
57
switch (op) {
block/ioctl.c
71
int op;
block/ioctl.c
73
if (get_user(op, &arg->op) || get_user(udata, &arg->data))
block/ioctl.c
76
return blkpg_do_ioctl(bdev, udata, op);
block/ioctl.c
81
compat_int_t op;
block/ioctl.c
91
int op;
block/ioctl.c
93
if (get_user(op, &arg->op) || get_user(udata, &arg->data))
block/ioctl.c
96
return blkpg_do_ioctl(bdev, compat_ptr(udata), op);
block/kyber-iosched.c
877
#define KYBER_LAT_ATTR(op) __ATTR(op##_lat_nsec, 0644, kyber_##op##_lat_show, kyber_##op##_lat_store)
crypto/af_alg.c
582
con->op = *(u32 *)CMSG_DATA(cmsg);
crypto/af_alg.c
930
switch (con.op) {
crypto/ahash.c
323
int (*const *op)(struct ahash_request *req))
crypto/ahash.c
329
return (*op)(req);
crypto/ahash.c
343
if (op == &crypto_ahash_alg(tfm)->digest) {
crypto/ahash.c
353
if (op == &crypto_ahash_alg(tfm)->finup) {
crypto/asymmetric_keys/public_key.c
292
&issig, params->op);
crypto/asymmetric_keys/public_key.c
338
switch (params->op) {
crypto/asymmetric_keys/public_key.c
70
enum kernel_pkey_operation op)
crypto/asymmetric_keys/public_key.c
84
*sig = op == kernel_pkey_sign ||
crypto/asymmetric_keys/public_key.c
85
op == kernel_pkey_verify;
crypto/crypto_engine.c
123
op = &alg->op;
crypto/crypto_engine.c
124
ret = op->do_one_request(engine, async_req);
crypto/crypto_engine.c
28
struct crypto_engine_op op;
crypto/crypto_engine.c
509
if (!alg->op.do_one_request)
crypto/crypto_engine.c
548
if (!alg->op.do_one_request)
crypto/crypto_engine.c
588
if (!alg->op.do_one_request)
crypto/crypto_engine.c
602
if (!alg->op.do_one_request)
crypto/crypto_engine.c
616
if (!alg->op.do_one_request)
crypto/crypto_engine.c
75
struct crypto_engine_op *op;
crypto/shash.c
105
err = op(desc, data, len, out);
crypto/shash.c
99
int (*op)(struct shash_desc *desc, const u8 *data,
crypto/testmgr.c
1277
static inline int check_shash_op(const char *op, int err,
crypto/testmgr.c
1283
driver, op, err, vec_name, cfg->name);
crypto/testmgr.c
1435
static int do_ahash_op(int (*op)(struct ahash_request *req),
crypto/testmgr.c
1444
err = op(req);
crypto/testmgr.c
1452
static int check_nonfinal_ahash_op(const char *op, int err,
crypto/testmgr.c
1459
driver, op, err, vec_name, cfg->name);
crypto/testmgr.c
1464
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2044
const char *op = enc ? "encryption" : "decryption";
crypto/testmgr.c
2110
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2138
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2161
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2167
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2186
driver, op, vec_name, expected_error, err,
crypto/testmgr.c
2191
driver, op, vec_name, expected_error, cfg->name);
crypto/testmgr.c
2204
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2209
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2811
const char *op = enc ? "encryption" : "decryption";
crypto/testmgr.c
2861
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2887
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2908
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2914
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2923
driver, op, vec_name, vec->crypt_error, err, cfg->name);
crypto/testmgr.c
2928
driver, op, vec_name, vec->crypt_error, cfg->name);
crypto/testmgr.c
2937
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2942
driver, op, vec_name, cfg->name);
crypto/testmgr.c
2949
driver, op, vec_name, cfg->name);
drivers/accel/amdxdna/aie2_message.c
1002
op = ERT_INVALID_CMD;
drivers/accel/amdxdna/aie2_message.c
1015
abo, &size, &op);
drivers/accel/amdxdna/aie2_message.c
1022
msg.opcode = EXEC_MSG_OPS(xdna)->get_chain_msg_op(op);
drivers/accel/amdxdna/aie2_message.c
1054
u32 op = ERT_INVALID_CMD;
drivers/accel/amdxdna/aie2_message.c
1059
ret = aie2_cmdlist_fill_slot(cmdbuf_abo->mem.kva, cmd_abo, &size, &op);
drivers/accel/amdxdna/aie2_message.c
1063
msg.opcode = EXEC_MSG_OPS(xdna)->get_chain_msg_op(op);
drivers/accel/amdxdna/aie2_message.c
27
#define DECLARE_AIE2_MSG(name, op) \
drivers/accel/amdxdna/aie2_message.c
28
DECLARE_XDNA_MSG_COMMON(name, op, MAX_AIE2_STATUS_CODE)
drivers/accel/amdxdna/aie2_message.c
846
u32 op;
drivers/accel/amdxdna/aie2_message.c
849
op = amdxdna_cmd_get_op(cmd_abo);
drivers/accel/amdxdna/aie2_message.c
850
switch (op) {
drivers/accel/amdxdna/aie2_message.c
867
XDNA_ERR(xdna, "Unsupported op %d", op);
drivers/accel/amdxdna/aie2_message.c
881
u32 op;
drivers/accel/amdxdna/aie2_message.c
883
op = amdxdna_cmd_get_op(cmd_abo);
drivers/accel/amdxdna/aie2_message.c
885
*cmd_op = op;
drivers/accel/amdxdna/aie2_message.c
886
else if (op != *cmd_op)
drivers/accel/amdxdna/aie2_message.c
889
switch (op) {
drivers/accel/amdxdna/aie2_message.c
907
XDNA_INFO(xdna, "Unsupported op %d", op);
drivers/accel/amdxdna/aie2_message.c
993
u32 op;
drivers/accel/amdxdna/aie2_message.c
996
op = amdxdna_cmd_get_op(cmd_abo);
drivers/accel/amdxdna/aie2_message.c
998
if (op != ERT_CMD_CHAIN || !payload ||
drivers/accel/amdxdna/amdxdna_mailbox_helper.h
22
#define DECLARE_XDNA_MSG_COMMON(name, op, s) \
drivers/accel/amdxdna/amdxdna_mailbox_helper.h
36
.opcode = op, \
drivers/accel/ethosu/ethosu_gem.c
236
u16 op, struct cmd_state *st,
drivers/accel/ethosu/ethosu_gem.c
259
op, st->ifm.region, st->ifm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
268
op, st->ifm2.region, st->ifm2.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
275
op, st->weight[0].region, st->weight[0].base,
drivers/accel/ethosu/ethosu_gem.c
287
op, st->scale[0].region, st->scale[0].base,
drivers/accel/ethosu/ethosu_gem.c
300
op, st->ofm.region, st->ofm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
310
u16 op, struct cmd_state *st,
drivers/accel/ethosu/ethosu_gem.c
324
op, st->ifm.region, st->ifm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
337
op, st->ifm2.region, st->ifm2.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
345
op, st->ofm.region, st->ofm.base[0], len);
drivers/accel/habanalabs/common/command_buffer.c
381
switch (args->in.op) {
drivers/accel/habanalabs/common/habanalabs.h
2303
u32 op;
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1022
switch (args->op) {
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1099
switch (args->op) {
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1150
dev_err(dev, "Invalid request %d\n", args->op);
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1169
switch (args->op) {
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1197
switch (args->op) {
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1210
args->input_size = min(args->input_size, hl_debug_struct_size[args->op]);
drivers/accel/habanalabs/common/habanalabs_ioctl.c
1219
dev_err(hdev->dev, "Invalid request %d\n", args->op);
drivers/accel/habanalabs/common/habanalabs_ioctl.c
210
params->op = args->op;
drivers/accel/habanalabs/common/habanalabs_ioctl.c
213
input = kzalloc(hl_debug_struct_size[args->op], GFP_KERNEL);
drivers/accel/habanalabs/common/memory.c
2212
switch (args->in.op) {
drivers/accel/habanalabs/gaudi/gaudi_coresight.c
869
switch (params->op) {
drivers/accel/habanalabs/gaudi/gaudi_coresight.c
893
dev_err(hdev->dev, "Unknown coresight id %d\n", params->op);
drivers/accel/habanalabs/gaudi2/gaudi2_coresight.c
2550
switch (params->op) {
drivers/accel/habanalabs/gaudi2/gaudi2_coresight.c
2573
dev_err(hdev->dev, "Unknown coresight id %d\n", params->op);
drivers/accel/habanalabs/goya/goya_coresight.c
664
switch (params->op) {
drivers/accel/habanalabs/goya/goya_coresight.c
688
dev_err(hdev->dev, "Unknown coresight id %d\n", params->op);
drivers/accel/ivpu/ivpu_mmu.c
849
u32 op = FIELD_GET(IVPU_MMU_EVT_OP_MASK, event[0]);
drivers/accel/ivpu/ivpu_mmu.c
855
op, ivpu_mmu_event_to_str(op), ssid, sid,
drivers/acpi/acpica/acconvert.h
103
cv_print_one_comment_type(union acpi_parse_object *op,
drivers/acpi/acpica/acconvert.h
43
u32 cv_calculate_comment_lengths(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
53
u32 cv_parse_op_block_type(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
57
void cg_write_aml_def_block_comment(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
60
cg_write_one_aml_comment(union acpi_parse_object *op,
drivers/acpi/acpica/acconvert.h
63
void cg_write_aml_comment(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
70
void cv_clear_op_comments(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
75
void cv_label_file_node(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
86
void cv_transfer_comments(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
91
void cv_switch_files(u32 level, union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
93
u8 cv_file_has_switched(union acpi_parse_object *op);
drivers/acpi/acpica/acconvert.h
95
void cv_close_paren_write_comment(union acpi_parse_object *op, u32 level);
drivers/acpi/acpica/acconvert.h
97
void cv_close_brace_write_comment(union acpi_parse_object *op, u32 level);
drivers/acpi/acpica/acdebug.h
117
union acpi_parse_object *op);
drivers/acpi/acpica/acdebug.h
119
void acpi_db_set_method_call_breakpoint(union acpi_parse_object *op);
drivers/acpi/acpica/acdebug.h
125
void acpi_db_disassemble_aml(char *statements, union acpi_parse_object *op);
drivers/acpi/acpica/acdebug.h
160
void acpi_db_display_method_info(union acpi_parse_object *op);
drivers/acpi/acpica/acdebug.h
244
union acpi_parse_object *op);
drivers/acpi/acpica/acdebug.h
55
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
100
acpi_ds_create_buffer_field(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
104
acpi_ds_init_field_objects(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
177
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
207
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
212
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
218
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
224
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
230
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
251
acpi_ds_is_result_used(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
255
acpi_ds_delete_result_if_not_used(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
304
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
342
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
36
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
40
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
47
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
51
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
55
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
59
union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
64
union acpi_parse_object *op);
drivers/acpi/acpica/acdispat.h
85
acpi_ds_create_field(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
90
acpi_ds_create_bank_field(union acpi_parse_object *op,
drivers/acpi/acpica/acdispat.h
95
acpi_ds_create_index_field(union acpi_parse_object *op,
drivers/acpi/acpica/acinterp.h
112
acpi_ex_start_trace_opcode(union acpi_parse_object *op,
drivers/acpi/acpica/acinterp.h
116
acpi_ex_stop_trace_opcode(union acpi_parse_object *op,
drivers/acpi/acpica/aclocal.h
1185
union acpi_parse_object *op;
drivers/acpi/acpica/aclocal.h
148
union acpi_parse_object *op;
drivers/acpi/acpica/aclocal.h
624
union acpi_parse_object *op; /* Current op being parsed */
drivers/acpi/acpica/acobject.h
473
union acpi_parse_object op;
drivers/acpi/acpica/acparser.h
103
union acpi_parse_object *op, acpi_status status);
drivers/acpi/acpica/acparser.h
125
union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
129
union acpi_parse_object *op,
drivers/acpi/acpica/acparser.h
151
union acpi_parse_object **op, u32 *arg_list, u32 *arg_count);
drivers/acpi/acpica/acparser.h
155
union acpi_parse_object *op,
drivers/acpi/acpica/acparser.h
164
acpi_ps_append_arg(union acpi_parse_object *op, union acpi_parse_object *arg);
drivers/acpi/acpica/acparser.h
169
union acpi_parse_object *acpi_ps_get_arg(union acpi_parse_object *op, u32 argn);
drivers/acpi/acpica/acparser.h
172
union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
190
union acpi_parse_object *op,
drivers/acpi/acpica/acparser.h
202
void acpi_ps_init_op(union acpi_parse_object *op, u16 opcode);
drivers/acpi/acpica/acparser.h
206
void acpi_ps_free_op(union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
210
u32 acpi_ps_get_name(union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
212
void acpi_ps_set_name(union acpi_parse_object *op, u32 name);
drivers/acpi/acpica/acparser.h
219
u32 buffer_size, union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
223
u32 buffer_size, union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
225
void acpi_ps_show(union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
82
union acpi_parse_object *acpi_ps_get_parent(union acpi_parse_object *op);
drivers/acpi/acpica/acparser.h
91
union acpi_parse_object **op);
drivers/acpi/acpica/acparser.h
99
union acpi_parse_object **op, acpi_status status);
drivers/acpi/acpica/acstruct.h
81
union acpi_parse_object *op; /* Current parser op */
drivers/acpi/acpica/amlresrc.h
568
mp_save_gpio_info(union acpi_parse_object *op,
drivers/acpi/acpica/amlresrc.h
573
mp_save_serial_info(union acpi_parse_object *op,
drivers/acpi/acpica/amlresrc.h
580
char *mp_get_connection_info(union acpi_parse_object *op,
drivers/acpi/acpica/amlresrc.h
585
char *mp_get_parent_device_hid(union acpi_parse_object *op,
drivers/acpi/acpica/dbdisply.c
110
static void acpi_db_dump_parser_descriptor(union acpi_parse_object *op)
drivers/acpi/acpica/dbdisply.c
114
info = acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/dbdisply.c
117
acpi_os_printf("%20.20s : %4.4X\n", "Opcode", op->common.aml_opcode);
drivers/acpi/acpica/dbdisply.c
122
acpi_os_printf("%20.20s : %p\n", "Value/ArgList", op->common.value.arg);
drivers/acpi/acpica/dbdisply.c
123
acpi_os_printf("%20.20s : %p\n", "Parent", op->common.parent);
drivers/acpi/acpica/dbdisply.c
124
acpi_os_printf("%20.20s : %p\n", "NextOp", op->common.next);
drivers/acpi/acpica/dbdisply.c
22
static void acpi_db_dump_parser_descriptor(union acpi_parse_object *op);
drivers/acpi/acpica/dbdisply.c
356
union acpi_parse_object *op;
drivers/acpi/acpica/dbdisply.c
386
op = root_op;
drivers/acpi/acpica/dbdisply.c
388
while (op) {
drivers/acpi/acpica/dbdisply.c
389
if (op == start_op) {
drivers/acpi/acpica/dbdisply.c
400
op_info = acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/dbdisply.c
427
op = acpi_ps_get_depth_next(start_op, op);
drivers/acpi/acpica/dbinput.c
1171
if (op) {
drivers/acpi/acpica/dbinput.c
697
union acpi_parse_object *op)
drivers/acpi/acpica/dbinput.c
744
if (op) {
drivers/acpi/acpica/dbinput.c
774
op);
drivers/acpi/acpica/dbinput.c
784
acpi_db_set_method_call_breakpoint(op);
drivers/acpi/acpica/dbinput.c
868
status = acpi_db_command_dispatch(command_line, walk_state, op);
drivers/acpi/acpica/dbinput.c
878
status = acpi_db_command_dispatch(command_line, walk_state, op);
drivers/acpi/acpica/dbinput.c
883
acpi_db_display_method_info(op);
drivers/acpi/acpica/dbinput.c
893
if (op) {
drivers/acpi/acpica/dbinput.c
928
acpi_db_disassemble_aml(acpi_gbl_db_args[1], op);
drivers/acpi/acpica/dbmethod.c
236
void acpi_db_disassemble_aml(char *statements, union acpi_parse_object *op)
drivers/acpi/acpica/dbmethod.c
240
if (!op) {
drivers/acpi/acpica/dbmethod.c
249
acpi_dm_disassemble(NULL, op, num_statements);
drivers/acpi/acpica/dbmethod.c
268
union acpi_parse_object *op;
drivers/acpi/acpica/dbmethod.c
286
op = acpi_ps_create_scope_op(obj_desc->method.aml_start);
drivers/acpi/acpica/dbmethod.c
287
if (!op) {
drivers/acpi/acpica/dbmethod.c
293
walk_state = acpi_ds_create_walk_state(0, op, NULL, NULL);
drivers/acpi/acpica/dbmethod.c
298
status = acpi_ds_init_aml_walk(walk_state, op, NULL,
drivers/acpi/acpica/dbmethod.c
330
(void)acpi_dm_parse_deferred_ops(op);
drivers/acpi/acpica/dbmethod.c
335
acpi_dm_disassemble(NULL, op, 0);
drivers/acpi/acpica/dbmethod.c
338
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/dbmethod.c
44
union acpi_parse_object *op)
drivers/acpi/acpica/dbmethod.c
49
if (!op) {
drivers/acpi/acpica/dbmethod.c
57
aml_offset = (u32)ACPI_PTR_DIFF(op->common.aml,
drivers/acpi/acpica/dbmethod.c
83
void acpi_db_set_method_call_breakpoint(union acpi_parse_object *op)
drivers/acpi/acpica/dbmethod.c
86
if (!op) {
drivers/acpi/acpica/dbstats.c
242
if (acpi_ps_is_named_op(op->opcode)) {
drivers/acpi/acpica/dbstats.c
251
op = acpi_ps_get_depth_next(root, op);
drivers/acpi/acpica/dbutils.c
326
union acpi_parse_object *op = root;
drivers/acpi/acpica/dbutils.c
338
while (op) {
drivers/acpi/acpica/dbutils.c
339
if (op->common.aml_opcode == AML_METHOD_OP) {
drivers/acpi/acpica/dbutils.c
340
method = op;
drivers/acpi/acpica/dbutils.c
358
walk_state->parser_state.start_scope = op;
drivers/acpi/acpica/dbutils.c
380
if (op->common.aml_opcode == AML_REGION_OP) {
drivers/acpi/acpica/dbutils.c
394
op = acpi_ps_get_depth_next(root, op);
drivers/acpi/acpica/dbxface.c
140
*op)
drivers/acpi/acpica/dbxface.c
145
display_op = op;
drivers/acpi/acpica/dbxface.c
146
parent_op = op->common.parent;
drivers/acpi/acpica/dbxface.c
203
union acpi_parse_object *op, u32 opcode_class)
drivers/acpi/acpica/dbxface.c
21
union acpi_parse_object *op);
drivers/acpi/acpica/dbxface.c
225
aml_offset = (u32)ACPI_PTR_DIFF(op->common.aml,
drivers/acpi/acpica/dbxface.c
257
if (op->common.aml_opcode == AML_INT_NAMEDFIELD_OP) {
drivers/acpi/acpica/dbxface.c
291
next = op->common.next;
drivers/acpi/acpica/dbxface.c
292
op->common.next = NULL;
drivers/acpi/acpica/dbxface.c
298
acpi_db_get_display_op(walk_state, op),
drivers/acpi/acpica/dbxface.c
305
acpi_os_printf("AML Opcode: %4.4X %s\n", op->common.aml_opcode,
drivers/acpi/acpica/dbxface.c
306
acpi_ps_get_opcode_name(op->common.aml_opcode));
drivers/acpi/acpica/dbxface.c
309
if ((op->common.aml_opcode == AML_IF_OP) ||
drivers/acpi/acpica/dbxface.c
31
*op);
drivers/acpi/acpica/dbxface.c
310
(op->common.aml_opcode == AML_WHILE_OP)) {
drivers/acpi/acpica/dbxface.c
318
} else if (op->common.aml_opcode == AML_ELSE_OP) {
drivers/acpi/acpica/dbxface.c
325
op->common.next = next;
drivers/acpi/acpica/dbxface.c
345
if (op->common.aml_opcode != AML_INT_METHODCALL_OP) {
drivers/acpi/acpica/dbxface.c
361
if (op->common.aml_opcode == AML_INT_METHODCALL_OP) {
drivers/acpi/acpica/dbxface.c
375
status = acpi_db_start_command(walk_state, op);
drivers/acpi/acpica/dbxface.c
49
union acpi_parse_object *op)
drivers/acpi/acpica/dbxface.c
80
op);
drivers/acpi/acpica/dsargs.c
103
op->common.node = scope_node;
drivers/acpi/acpica/dsargs.c
115
status = acpi_ds_init_aml_walk(walk_state, op, NULL, aml_start,
drivers/acpi/acpica/dsargs.c
128
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/dsargs.c
48
union acpi_parse_object *op;
drivers/acpi/acpica/dsargs.c
55
op = acpi_ps_alloc_op(AML_INT_EVAL_SUBTREE_OP, aml_start);
drivers/acpi/acpica/dsargs.c
56
if (!op) {
drivers/acpi/acpica/dsargs.c
62
op->common.node = scope_node;
drivers/acpi/acpica/dsargs.c
72
status = acpi_ds_init_aml_walk(walk_state, op, NULL, aml_start,
drivers/acpi/acpica/dsargs.c
93
op->common.node = node;
drivers/acpi/acpica/dsargs.c
94
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/dsargs.c
98
op = acpi_ps_alloc_op(AML_INT_EVAL_SUBTREE_OP, aml_start);
drivers/acpi/acpica/dsargs.c
99
if (!op) {
drivers/acpi/acpica/dscontrol.c
135
union acpi_parse_object *op)
drivers/acpi/acpica/dscontrol.c
142
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dscontrol.c
145
ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH, "[IF_OP] Op=%p\n", op));
drivers/acpi/acpica/dscontrol.c
169
ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH, "[WHILE_OP] Op=%p\n", op));
drivers/acpi/acpica/dscontrol.c
202
"[WHILE_OP] termination! Op=%p\n", op));
drivers/acpi/acpica/dscontrol.c
214
"[RETURN_OP] Op=%p Arg=%p\n", op,
drivers/acpi/acpica/dscontrol.c
215
op->common.value.arg));
drivers/acpi/acpica/dscontrol.c
222
if (op->common.value.arg) {
drivers/acpi/acpica/dscontrol.c
232
op->common.value.arg);
drivers/acpi/acpica/dscontrol.c
355
if (op->common.aml_opcode == AML_BREAK_OP) {
drivers/acpi/acpica/dscontrol.c
36
union acpi_parse_object *op)
drivers/acpi/acpica/dscontrol.c
365
op->common.aml_opcode, op));
drivers/acpi/acpica/dscontrol.c
44
op, op->common.aml_opcode, walk_state));
drivers/acpi/acpica/dscontrol.c
46
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dscontrol.c
86
control_state->control.opcode = op->common.aml_opcode;
drivers/acpi/acpica/dsdebug.c
157
if (op) {
drivers/acpi/acpica/dsdebug.c
161
next = op->common.next;
drivers/acpi/acpica/dsdebug.c
162
op->common.next = NULL;
drivers/acpi/acpica/dsdebug.c
171
acpi_dm_disassemble(next_walk_state, op,
drivers/acpi/acpica/dsdebug.c
175
op->common.next = next;
drivers/acpi/acpica/dsdebug.c
200
union acpi_parse_object *op)
drivers/acpi/acpica/dsdebug.c
91
union acpi_parse_object *op)
drivers/acpi/acpica/dsfield.c
120
acpi_ds_create_buffer_field(union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
135
if (op->common.aml_opcode == AML_CREATE_FIELD_OP) {
drivers/acpi/acpica/dsfield.c
139
arg = acpi_ps_get_arg(op, 3);
drivers/acpi/acpica/dsfield.c
143
arg = acpi_ps_get_arg(op, 2);
drivers/acpi/acpica/dsfield.c
195
op->common.node = node;
drivers/acpi/acpica/dsfield.c
225
second_desc->extra.aml_start = op->named.data;
drivers/acpi/acpica/dsfield.c
226
second_desc->extra.aml_length = op->named.length;
drivers/acpi/acpica/dsfield.c
30
union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
473
acpi_ds_create_field(union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
481
ACPI_FUNCTION_TRACE_PTR(ds_create_field, op);
drivers/acpi/acpica/dsfield.c
485
arg = op->common.value.arg;
drivers/acpi/acpica/dsfield.c
553
acpi_ds_init_field_objects(union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
562
ACPI_FUNCTION_TRACE_PTR(ds_init_field_objects, op);
drivers/acpi/acpica/dsfield.c
585
arg = acpi_ps_get_arg(op, 2);
drivers/acpi/acpica/dsfield.c
591
arg = acpi_ps_get_arg(op, 4);
drivers/acpi/acpica/dsfield.c
597
arg = acpi_ps_get_arg(op, 3);
drivers/acpi/acpica/dsfield.c
61
union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
673
acpi_ds_create_bank_field(union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
681
ACPI_FUNCTION_TRACE_PTR(ds_create_bank_field, op);
drivers/acpi/acpica/dsfield.c
685
arg = op->common.value.arg;
drivers/acpi/acpica/dsfield.c
745
info.data_register_node = (struct acpi_namespace_node *)op;
drivers/acpi/acpica/dsfield.c
766
acpi_ds_create_index_field(union acpi_parse_object *op,
drivers/acpi/acpica/dsfield.c
774
ACPI_FUNCTION_TRACE_PTR(ds_create_index_field, op);
drivers/acpi/acpica/dsfield.c
778
arg = op->common.value.arg;
drivers/acpi/acpica/dsfield.c
78
acpi_dm_add_op_to_external_list(op, path, ACPI_TYPE_REGION, 0, 0);
drivers/acpi/acpica/dsmethod.c
102
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/dsmethod.c
222
acpi_ds_dump_method_stack(status, walk_state, walk_state->op);
drivers/acpi/acpica/dsmethod.c
458
union acpi_parse_object *op)
drivers/acpi/acpica/dsmethod.c
59
union acpi_parse_object *op = NULL;
drivers/acpi/acpica/dsmethod.c
70
op = acpi_ps_alloc_op(AML_METHOD_OP, obj_desc->method.aml_start);
drivers/acpi/acpica/dsmethod.c
71
if (!op) {
drivers/acpi/acpica/dsmethod.c
75
acpi_ps_set_name(op, node->name.integer);
drivers/acpi/acpica/dsmethod.c
76
op->common.node = node;
drivers/acpi/acpica/dsmethod.c
83
acpi_ps_free_op(op);
drivers/acpi/acpica/dsmethod.c
87
status = acpi_ds_init_aml_walk(walk_state, op, node,
drivers/acpi/acpica/dsmethod.c
92
acpi_ps_free_op(op);
drivers/acpi/acpica/dsobject.c
102
acpi_ds_init_object_from_op(walk_state, op, op->common.aml_opcode,
drivers/acpi/acpica/dsobject.c
113
if ((op->common.parent->common.aml_opcode == AML_PACKAGE_OP) ||
drivers/acpi/acpica/dsobject.c
114
(op->common.parent->common.aml_opcode == AML_VARIABLE_PACKAGE_OP)) {
drivers/acpi/acpica/dsobject.c
117
if ((op->common.aml_opcode == AML_INT_NAMEPATH_OP) &&
drivers/acpi/acpica/dsobject.c
125
obj_desc->reference.aml = op->common.aml;
drivers/acpi/acpica/dsobject.c
152
union acpi_parse_object *op,
drivers/acpi/acpica/dsobject.c
185
arg = op->common.value.arg; /* skip first arg */
drivers/acpi/acpica/dsobject.c
234
op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc);
drivers/acpi/acpica/dsobject.c
255
union acpi_parse_object *op)
drivers/acpi/acpica/dsobject.c
260
ACPI_FUNCTION_TRACE_PTR(ds_create_node, op);
drivers/acpi/acpica/dsobject.c
271
if (!op->common.value.arg) {
drivers/acpi/acpica/dsobject.c
281
acpi_ds_build_internal_object(walk_state, op->common.value.arg,
drivers/acpi/acpica/dsobject.c
320
union acpi_parse_object *op,
drivers/acpi/acpica/dsobject.c
349
obj_desc->buffer.aml_start = op->named.data;
drivers/acpi/acpica/dsobject.c
350
obj_desc->buffer.aml_length = op->named.length;
drivers/acpi/acpica/dsobject.c
365
if (!op->named.data) {
drivers/acpi/acpica/dsobject.c
369
obj_desc->package.aml_start = op->named.data;
drivers/acpi/acpica/dsobject.c
37
union acpi_parse_object *op,
drivers/acpi/acpica/dsobject.c
370
obj_desc->package.aml_length = op->named.length;
drivers/acpi/acpica/dsobject.c
424
obj_desc->integer.value = op->common.value.integer;
drivers/acpi/acpica/dsobject.c
432
ACPI_FORMAT_UINT64(op->common.
drivers/acpi/acpica/dsobject.c
449
obj_desc->string.pointer = op->common.value.string;
drivers/acpi/acpica/dsobject.c
450
obj_desc->string.length = (u32)strlen(op->common.value.string);
drivers/acpi/acpica/dsobject.c
46
if (op->common.aml_opcode == AML_INT_NAMEPATH_OP) {
drivers/acpi/acpica/dsobject.c
506
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dsobject.c
511
obj_desc->reference.node = op->common.node;
drivers/acpi/acpica/dsobject.c
513
if (op->common.node) {
drivers/acpi/acpica/dsobject.c
515
op->common.node->object;
drivers/acpi/acpica/dsobject.c
52
if (!op->common.node) {
drivers/acpi/acpica/dsobject.c
56
if ((op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/dsobject.c
58
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/dsobject.c
69
op->common.value.string,
drivers/acpi/acpica/dsobject.c
78
&(op->common.node)));
drivers/acpi/acpica/dsobject.c
82
op->common.value.
drivers/acpi/acpica/dsobject.c
95
(op->common.aml_opcode))->
drivers/acpi/acpica/dsopcode.c
269
union acpi_parse_object *op)
drivers/acpi/acpica/dsopcode.c
276
ACPI_FUNCTION_TRACE_PTR(ds_eval_buffer_field_operands, op);
drivers/acpi/acpica/dsopcode.c
282
node = op->common.node;
drivers/acpi/acpica/dsopcode.c
286
next_op = op->common.value.arg;
drivers/acpi/acpica/dsopcode.c
303
acpi_ex_resolve_operands(op->common.aml_opcode, ACPI_WALK_OPERANDS,
drivers/acpi/acpica/dsopcode.c
307
acpi_ps_get_opcode_name(op->common.aml_opcode),
drivers/acpi/acpica/dsopcode.c
315
if (op->common.aml_opcode == AML_CREATE_FIELD_OP) {
drivers/acpi/acpica/dsopcode.c
320
acpi_ds_init_buffer_field(op->common.aml_opcode, obj_desc,
drivers/acpi/acpica/dsopcode.c
329
acpi_ds_init_buffer_field(op->common.aml_opcode, obj_desc,
drivers/acpi/acpica/dsopcode.c
354
union acpi_parse_object *op)
drivers/acpi/acpica/dsopcode.c
363
ACPI_FUNCTION_TRACE_PTR(ds_eval_region_operands, op);
drivers/acpi/acpica/dsopcode.c
369
node = op->common.node;
drivers/acpi/acpica/dsopcode.c
373
next_op = op->common.value.arg;
drivers/acpi/acpica/dsopcode.c
390
acpi_ex_resolve_operands(op->common.aml_opcode, ACPI_WALK_OPERANDS,
drivers/acpi/acpica/dsopcode.c
461
union acpi_parse_object *op)
drivers/acpi/acpica/dsopcode.c
471
ACPI_FUNCTION_TRACE_PTR(ds_eval_table_region_operands, op);
drivers/acpi/acpica/dsopcode.c
477
node = op->common.node;
drivers/acpi/acpica/dsopcode.c
481
next_op = op->common.value.arg;
drivers/acpi/acpica/dsopcode.c
499
acpi_ex_resolve_operands(op->common.aml_opcode, ACPI_WALK_OPERANDS,
drivers/acpi/acpica/dsopcode.c
570
union acpi_parse_object *op,
drivers/acpi/acpica/dsopcode.c
589
if (!op->common.value.arg) {
drivers/acpi/acpica/dsopcode.c
592
op->common.aml_opcode, op));
drivers/acpi/acpica/dsopcode.c
596
status = acpi_ds_create_operand(walk_state, op->common.value.arg, 1);
drivers/acpi/acpica/dsopcode.c
626
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dsopcode.c
630
acpi_ds_build_internal_buffer_obj(walk_state, op, length,
drivers/acpi/acpica/dsopcode.c
638
acpi_ds_build_internal_package_obj(walk_state, op, length,
drivers/acpi/acpica/dsopcode.c
653
if ((!op->common.parent) ||
drivers/acpi/acpica/dsopcode.c
654
((op->common.parent->common.aml_opcode != AML_PACKAGE_OP) &&
drivers/acpi/acpica/dsopcode.c
655
(op->common.parent->common.aml_opcode !=
drivers/acpi/acpica/dsopcode.c
657
&& (op->common.parent->common.aml_opcode !=
drivers/acpi/acpica/dsopcode.c
682
union acpi_parse_object *op)
drivers/acpi/acpica/dsopcode.c
691
ACPI_FUNCTION_TRACE_PTR(ds_eval_bank_field_operands, op);
drivers/acpi/acpica/dsopcode.c
700
next_op = op->common.value.arg;
drivers/acpi/acpica/dsopcode.c
729
acpi_ps_get_opcode_name(op->common.aml_opcode), 1);
drivers/acpi/acpica/dsopcode.c
738
arg = acpi_ps_get_arg(op, 4);
drivers/acpi/acpica/dspkginit.c
129
arg = op->common.value.arg;
drivers/acpi/acpica/dspkginit.c
307
op->common.node = ACPI_CAST_PTR(struct acpi_namespace_node, obj_desc);
drivers/acpi/acpica/dspkginit.c
56
union acpi_parse_object *op,
drivers/acpi/acpica/dspkginit.c
79
parent = op->common.parent;
drivers/acpi/acpica/dsutils.c
133
acpi_ds_is_result_used(union acpi_parse_object * op,
drivers/acpi/acpica/dsutils.c
138
ACPI_FUNCTION_TRACE_PTR(ds_is_result_used, op);
drivers/acpi/acpica/dsutils.c
142
if (!op) {
drivers/acpi/acpica/dsutils.c
167
if ((!op->common.parent) ||
drivers/acpi/acpica/dsutils.c
168
(op->common.parent->common.aml_opcode == AML_SCOPE_OP)) {
drivers/acpi/acpica/dsutils.c
174
acpi_ps_get_opcode_name(op->common.
drivers/acpi/acpica/dsutils.c
182
acpi_ps_get_opcode_info(op->common.parent->common.aml_opcode);
drivers/acpi/acpica/dsutils.c
184
ACPI_ERROR((AE_INFO, "Unknown parent opcode Op=%p", op));
drivers/acpi/acpica/dsutils.c
197
switch (op->common.parent->common.aml_opcode) {
drivers/acpi/acpica/dsutils.c
213
op)) {
drivers/acpi/acpica/dsutils.c
238
if ((op->common.parent->common.aml_opcode == AML_REGION_OP) ||
drivers/acpi/acpica/dsutils.c
239
(op->common.parent->common.aml_opcode == AML_DATA_REGION_OP)
drivers/acpi/acpica/dsutils.c
240
|| (op->common.parent->common.aml_opcode == AML_PACKAGE_OP)
drivers/acpi/acpica/dsutils.c
241
|| (op->common.parent->common.aml_opcode == AML_BUFFER_OP)
drivers/acpi/acpica/dsutils.c
242
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/dsutils.c
244
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/dsutils.c
246
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/dsutils.c
268
acpi_ps_get_opcode_name(op->common.aml_opcode),
drivers/acpi/acpica/dsutils.c
269
acpi_ps_get_opcode_name(op->common.parent->common.
drivers/acpi/acpica/dsutils.c
270
aml_opcode), op));
drivers/acpi/acpica/dsutils.c
277
acpi_ps_get_opcode_name(op->common.aml_opcode),
drivers/acpi/acpica/dsutils.c
278
acpi_ps_get_opcode_name(op->common.parent->common.
drivers/acpi/acpica/dsutils.c
279
aml_opcode), op));
drivers/acpi/acpica/dsutils.c
302
acpi_ds_delete_result_if_not_used(union acpi_parse_object *op,
drivers/acpi/acpica/dsutils.c
311
if (!op) {
drivers/acpi/acpica/dsutils.c
320
if (!acpi_ds_is_result_used(op, walk_state)) {
drivers/acpi/acpica/dsutils.c
755
union acpi_parse_object *op = walk_state->op;
drivers/acpi/acpica/dsutils.c
762
if (!op->common.parent) {
drivers/acpi/acpica/dsutils.c
769
if ((op->common.parent->common.aml_opcode == AML_PACKAGE_OP) ||
drivers/acpi/acpica/dsutils.c
770
(op->common.parent->common.aml_opcode == AML_VARIABLE_PACKAGE_OP) ||
drivers/acpi/acpica/dsutils.c
771
(op->common.parent->common.aml_opcode == AML_REF_OF_OP)) {
drivers/acpi/acpica/dsutils.c
778
status = acpi_ds_create_operand(walk_state, op, 0);
drivers/acpi/acpica/dsutils.c
783
if (op->common.flags & ACPI_PARSEOP_TARGET) {
drivers/acpi/acpica/dsutils.c
832
op->common.flags |= ACPI_PARSEOP_IN_STACK;
drivers/acpi/acpica/dswexec.c
148
walk_state->op));
drivers/acpi/acpica/dswexec.c
186
union acpi_parse_object *op;
drivers/acpi/acpica/dswexec.c
192
op = walk_state->op;
drivers/acpi/acpica/dswexec.c
193
if (!op) {
drivers/acpi/acpica/dswexec.c
199
op = *out_op;
drivers/acpi/acpica/dswexec.c
200
walk_state->op = op;
drivers/acpi/acpica/dswexec.c
201
walk_state->opcode = op->common.aml_opcode;
drivers/acpi/acpica/dswexec.c
203
acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/dswexec.c
211
op));
drivers/acpi/acpica/dswexec.c
220
if (op == walk_state->origin) {
drivers/acpi/acpica/dswexec.c
222
*out_op = op;
drivers/acpi/acpica/dswexec.c
238
op, walk_state));
drivers/acpi/acpica/dswexec.c
245
walk_state->control_state->control.predicate_op = op;
drivers/acpi/acpica/dswexec.c
252
if (op->common.aml_opcode == AML_INT_NAMEPATH_OP) {
drivers/acpi/acpica/dswexec.c
262
status = acpi_ds_exec_begin_control_op(walk_state, op);
drivers/acpi/acpica/dswexec.c
279
if (op->common.aml_opcode != AML_SCOPE_OP) {
drivers/acpi/acpica/dswexec.c
284
acpi_ds_scope_stack_push(op->named.node,
drivers/acpi/acpica/dswexec.c
285
op->named.node->
drivers/acpi/acpica/dswexec.c
329
union acpi_parse_object *op;
drivers/acpi/acpica/dswexec.c
342
op = walk_state->op;
drivers/acpi/acpica/dswexec.c
348
op->common.aml_opcode));
drivers/acpi/acpica/dswexec.c
352
first_arg = op->common.value.arg;
drivers/acpi/acpica/dswexec.c
363
status = acpi_db_single_step(walk_state, op, op_class);
drivers/acpi/acpica/dswexec.c
462
status = acpi_ds_exec_end_control_op(walk_state, op);
drivers/acpi/acpica/dswexec.c
472
if ((op->asl.parent) &&
drivers/acpi/acpica/dswexec.c
473
((op->asl.parent->asl.aml_opcode == AML_PACKAGE_OP)
drivers/acpi/acpica/dswexec.c
474
|| (op->asl.parent->asl.aml_opcode ==
drivers/acpi/acpica/dswexec.c
478
op));
drivers/acpi/acpica/dswexec.c
480
op->common.node = (struct acpi_namespace_node *)
drivers/acpi/acpica/dswexec.c
481
op->asl.value.arg->asl.node;
drivers/acpi/acpica/dswexec.c
482
acpi_ut_add_reference(op->asl.value.arg->asl.
drivers/acpi/acpica/dswexec.c
488
"Method invocation, Op=%p\n", op));
drivers/acpi/acpica/dswexec.c
540
op));
drivers/acpi/acpica/dswexec.c
548
acpi_ds_eval_buffer_field_operands(walk_state, op);
drivers/acpi/acpica/dswexec.c
558
acpi_ns_get_external_pathname(op->common.node);
drivers/acpi/acpica/dswexec.c
563
op->common.
drivers/acpi/acpica/dswexec.c
580
op, op->named.value.arg,
drivers/acpi/acpica/dswexec.c
581
op->common.parent->common.
drivers/acpi/acpica/dswexec.c
584
switch (op->common.parent->common.aml_opcode) {
drivers/acpi/acpica/dswexec.c
591
op->common.parent->common.node;
drivers/acpi/acpica/dswexec.c
595
op->common.parent->
drivers/acpi/acpica/dswexec.c
597
op->common.parent);
drivers/acpi/acpica/dswexec.c
608
(walk_state, op,
drivers/acpi/acpica/dswexec.c
609
acpi_ns_get_attached_object(op->common.
drivers/acpi/acpica/dswexec.c
618
(walk_state, op, NULL);
drivers/acpi/acpica/dswexec.c
643
if (op->common.aml_opcode == AML_REGION_OP) {
drivers/acpi/acpica/dswexec.c
646
op));
drivers/acpi/acpica/dswexec.c
650
op);
drivers/acpi/acpica/dswexec.c
654
} else if (op->common.aml_opcode == AML_DATA_REGION_OP) {
drivers/acpi/acpica/dswexec.c
657
op));
drivers/acpi/acpica/dswexec.c
661
(walk_state, op);
drivers/acpi/acpica/dswexec.c
665
} else if (op->common.aml_opcode == AML_BANK_FIELD_OP) {
drivers/acpi/acpica/dswexec.c
668
op));
drivers/acpi/acpica/dswexec.c
672
op);
drivers/acpi/acpica/dswexec.c
682
"Undefined opcode type Op=%p", op));
drivers/acpi/acpica/dswexec.c
689
walk_state->opcode, op));
drivers/acpi/acpica/dswexec.c
697
op_class, op_type, op->common.aml_opcode,
drivers/acpi/acpica/dswexec.c
698
op));
drivers/acpi/acpica/dswexec.c
719
(walk_state->control_state->control.predicate_op == op)) {
drivers/acpi/acpica/dswexec.c
740
acpi_ds_delete_result_if_not_used(op, walk_state->result_obj,
drivers/acpi/acpica/dswexec.c
78
status = acpi_ds_create_operand(walk_state, walk_state->op, 0);
drivers/acpi/acpica/dswload.c
106
union acpi_parse_object *op;
drivers/acpi/acpica/dswload.c
113
ACPI_FUNCTION_TRACE_PTR(ds_load1_begin_op, walk_state->op);
drivers/acpi/acpica/dswload.c
115
op = walk_state->op;
drivers/acpi/acpica/dswload.c
116
ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH, "Op=%p State=%p\n", op,
drivers/acpi/acpica/dswload.c
121
if (op) {
drivers/acpi/acpica/dswload.c
123
*out_op = op;
drivers/acpi/acpica/dswload.c
129
if (op->common.node) {
drivers/acpi/acpica/dswload.c
130
*out_op = op;
drivers/acpi/acpica/dswload.c
142
"State=%p Op=%p [%s]\n", walk_state, op,
drivers/acpi/acpica/dswload.c
163
acpi_dm_add_op_to_external_list(op, path,
drivers/acpi/acpica/dswload.c
352
if (!op) {
drivers/acpi/acpica/dswload.c
356
op = acpi_ps_alloc_op(walk_state->opcode, walk_state->aml);
drivers/acpi/acpica/dswload.c
357
if (!op) {
drivers/acpi/acpica/dswload.c
365
op->named.path = path;
drivers/acpi/acpica/dswload.c
373
op->common.node = node;
drivers/acpi/acpica/dswload.c
374
op->named.name = node->name.integer;
drivers/acpi/acpica/dswload.c
378
op);
drivers/acpi/acpica/dswload.c
379
*out_op = op;
drivers/acpi/acpica/dswload.c
398
union acpi_parse_object *op;
drivers/acpi/acpica/dswload.c
407
op = walk_state->op;
drivers/acpi/acpica/dswload.c
408
ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH, "Op=%p State=%p\n", op,
drivers/acpi/acpica/dswload.c
428
status = acpi_ds_create_buffer_field(op, walk_state);
drivers/acpi/acpica/dswload.c
452
acpi_ds_init_field_objects(op, walk_state);
drivers/acpi/acpica/dswload.c
463
if (op->common.aml_opcode == AML_REGION_OP) {
drivers/acpi/acpica/dswload.c
465
acpi_ex_create_region(op->named.data,
drivers/acpi/acpica/dswload.c
466
op->named.length,
drivers/acpi/acpica/dswload.c
468
((op->common.value.arg)->
drivers/acpi/acpica/dswload.c
474
} else if (op->common.aml_opcode == AML_DATA_REGION_OP) {
drivers/acpi/acpica/dswload.c
476
acpi_ex_create_region(op->named.data,
drivers/acpi/acpica/dswload.c
477
op->named.length,
drivers/acpi/acpica/dswload.c
486
if (op->common.aml_opcode == AML_NAME_OP) {
drivers/acpi/acpica/dswload.c
490
if (op->common.value.arg) {
drivers/acpi/acpica/dswload.c
491
object_type = (acpi_ps_get_opcode_info((op->common.
drivers/acpi/acpica/dswload.c
499
if (op->common.node) {
drivers/acpi/acpica/dswload.c
500
op->common.node->type = (u8) object_type;
drivers/acpi/acpica/dswload.c
510
op->common.node && op->common.aml_opcode == AML_EXTERNAL_OP) {
drivers/acpi/acpica/dswload.c
517
(u8)op->common.value.arg->common.next->common.value.integer;
drivers/acpi/acpica/dswload.c
518
object_type = (u8)op->common.value.arg->common.value.integer;
drivers/acpi/acpica/dswload.c
519
op->common.node->flags |= ANOBJ_IS_EXTERNAL;
drivers/acpi/acpica/dswload.c
520
op->common.node->type = (u8)object_type;
drivers/acpi/acpica/dswload.c
523
&op->common.node,
drivers/acpi/acpica/dswload.c
530
acpi_dm_add_op_to_external_list(op, op->named.path,
drivers/acpi/acpica/dswload.c
542
if (op->common.aml_opcode == AML_METHOD_OP) {
drivers/acpi/acpica/dswload.c
553
walk_state, op, op->named.node));
drivers/acpi/acpica/dswload.c
555
if (!acpi_ns_get_attached_object(op->named.node)) {
drivers/acpi/acpica/dswload.c
557
ACPI_CAST_PTR(void, op->named.node);
drivers/acpi/acpica/dswload.c
562
op->common.value.
drivers/acpi/acpica/dswload.c
566
acpi_ex_create_method(op->named.
drivers/acpi/acpica/dswload.c
568
op->named.
drivers/acpi/acpica/dswload.c
586
op->common.aml_opcode != AML_EXTERNAL_OP &&
drivers/acpi/acpica/dswload.c
590
acpi_ut_get_type_name(object_type), op));
drivers/acpi/acpica/dswload2.c
103
"State=%p Op=%p Type=%X\n", walk_state, op,
drivers/acpi/acpica/dswload2.c
131
if (op && (op->named.node == acpi_gbl_root_node)) {
drivers/acpi/acpica/dswload2.c
132
node = op->named.node;
drivers/acpi/acpica/dswload2.c
237
if (op && op->common.node) {
drivers/acpi/acpica/dswload2.c
241
node = op->common.node;
drivers/acpi/acpica/dswload2.c
331
if (!op) {
drivers/acpi/acpica/dswload2.c
335
op = acpi_ps_alloc_op(walk_state->opcode, walk_state->aml);
drivers/acpi/acpica/dswload2.c
336
if (!op) {
drivers/acpi/acpica/dswload2.c
343
op->named.name = node->name.integer;
drivers/acpi/acpica/dswload2.c
345
*out_op = op;
drivers/acpi/acpica/dswload2.c
352
op->common.node = node;
drivers/acpi/acpica/dswload2.c
371
union acpi_parse_object *op;
drivers/acpi/acpica/dswload2.c
386
op = walk_state->op;
drivers/acpi/acpica/dswload2.c
388
walk_state->op_info->name, op, walk_state));
drivers/acpi/acpica/dswload2.c
396
if (op->common.aml_opcode == AML_SCOPE_OP) {
drivers/acpi/acpica/dswload2.c
398
"Ending scope Op=%p State=%p\n", op,
drivers/acpi/acpica/dswload2.c
408
node = op->common.node;
drivers/acpi/acpica/dswload2.c
41
union acpi_parse_object *op;
drivers/acpi/acpica/dswload2.c
420
(op->common.aml_opcode != AML_INT_METHODCALL_OP)) {
drivers/acpi/acpica/dswload2.c
423
acpi_ut_get_type_name(object_type), op));
drivers/acpi/acpica/dswload2.c
461
acpi_ps_get_opcode_name(op->common.aml_opcode),
drivers/acpi/acpica/dswload2.c
462
walk_state, op, node));
drivers/acpi/acpica/dswload2.c
466
arg = op->common.value.arg;
drivers/acpi/acpica/dswload2.c
475
status = acpi_ds_create_buffer_field(op, walk_state);
drivers/acpi/acpica/dswload2.c
488
status = acpi_ds_init_field_objects(op, walk_state);
drivers/acpi/acpica/dswload2.c
491
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dswload2.c
495
acpi_ds_create_index_field(op,
drivers/acpi/acpica/dswload2.c
50
op = walk_state->op;
drivers/acpi/acpica/dswload2.c
503
acpi_ds_create_bank_field(op, arg->common.node,
drivers/acpi/acpica/dswload2.c
51
ACPI_DEBUG_PRINT((ACPI_DB_DISPATCH, "Op=%p State=%p\n", op,
drivers/acpi/acpica/dswload2.c
510
acpi_ds_create_field(op, arg->common.node,
drivers/acpi/acpica/dswload2.c
528
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dswload2.c
54
if (op) {
drivers/acpi/acpica/dswload2.c
573
switch (op->common.aml_opcode) {
drivers/acpi/acpica/dswload2.c
577
if (op->common.aml_opcode == AML_REGION_OP) {
drivers/acpi/acpica/dswload2.c
579
((op->common.value.arg)->common.value.
drivers/acpi/acpica/dswload2.c
602
status = acpi_ex_create_region(op->named.data,
drivers/acpi/acpica/dswload2.c
603
op->named.length,
drivers/acpi/acpica/dswload2.c
618
status = acpi_ds_create_node(walk_state, node, op);
drivers/acpi/acpica/dswload2.c
655
walk_state, op, op->named.node));
drivers/acpi/acpica/dswload2.c
657
if (!acpi_ns_get_attached_object(op->named.node)) {
drivers/acpi/acpica/dswload2.c
659
ACPI_CAST_PTR(void, op->named.node);
drivers/acpi/acpica/dswload2.c
664
op->common.value.
drivers/acpi/acpica/dswload2.c
668
acpi_ex_create_method(op->named.
drivers/acpi/acpica/dswload2.c
670
op->named.
drivers/acpi/acpica/dswload2.c
700
walk_state, op, node));
drivers/acpi/acpica/dswload2.c
726
op->common.node = new_node;
drivers/acpi/acpica/dswload2.c
79
buffer_ptr = op->common.value.string;
drivers/acpi/acpica/dswload2.c
89
buffer_ptr = ACPI_CAST_PTR(char, &op->named.name);
drivers/acpi/acpica/dswstate.c
566
union acpi_parse_object *op,
drivers/acpi/acpica/dswstate.c
598
status = acpi_ps_init_scope(&walk_state->parser_state, op);
drivers/acpi/acpica/excreate.c
226
node = walk_state->op->common.node;
drivers/acpi/acpica/extrace.c
330
acpi_ex_start_trace_opcode(union acpi_parse_object *op,
drivers/acpi/acpica/extrace.c
339
op->common.aml, op->common.aml_op_name);
drivers/acpi/acpica/extrace.c
358
acpi_ex_stop_trace_opcode(union acpi_parse_object *op,
drivers/acpi/acpica/extrace.c
367
op->common.aml, op->common.aml_op_name);
drivers/acpi/acpica/psargs.c
310
else if (walk_state->op->common.aml_opcode ==
drivers/acpi/acpica/psloop.c
123
acpi_ps_append_arg(op, arg);
drivers/acpi/acpica/psloop.c
136
switch (op->common.aml_opcode) {
drivers/acpi/acpica/psloop.c
144
op->named.data = walk_state->parser_state.aml;
drivers/acpi/acpica/psloop.c
145
op->named.length = (u32)
drivers/acpi/acpica/psloop.c
160
if ((op->common.parent) &&
drivers/acpi/acpica/psloop.c
161
(op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psloop.c
174
op->named.data = aml_op_start;
drivers/acpi/acpica/psloop.c
175
op->named.length = (u32)
drivers/acpi/acpica/psloop.c
224
union acpi_parse_object *op = NULL; /* current op */
drivers/acpi/acpica/psloop.c
249
if ((parser_state->scope->parse_scope.op) &&
drivers/acpi/acpica/psloop.c
250
((parser_state->scope->parse_scope.op->common.
drivers/acpi/acpica/psloop.c
252
|| (parser_state->scope->parse_scope.op->common.
drivers/acpi/acpica/psloop.c
261
walk_state->op = NULL;
drivers/acpi/acpica/psloop.c
279
acpi_ps_next_parse_state(walk_state, op,
drivers/acpi/acpica/psloop.c
283
acpi_ps_pop_scope(parser_state, &op,
drivers/acpi/acpica/psloop.c
287
"Popped scope, Op=%p\n", op));
drivers/acpi/acpica/psloop.c
292
op = walk_state->prev_op;
drivers/acpi/acpica/psloop.c
300
while ((parser_state->aml < parser_state->aml_end) || (op)) {
drivers/acpi/acpica/psloop.c
304
if (!op) {
drivers/acpi/acpica/psloop.c
306
acpi_ps_create_op(walk_state, aml_op_start, &op);
drivers/acpi/acpica/psloop.c
33
u8 * aml_op_start, union acpi_parse_object *op);
drivers/acpi/acpica/psloop.c
334
acpi_ps_complete_op(walk_state, &op,
drivers/acpi/acpica/psloop.c
371
acpi_ex_start_trace_opcode(op, walk_state);
drivers/acpi/acpica/psloop.c
380
switch (op->common.aml_opcode) {
drivers/acpi/acpica/psloop.c
401
acpi_ps_get_arguments(walk_state, aml_op_start, op);
drivers/acpi/acpica/psloop.c
404
acpi_ps_complete_op(walk_state, &op,
drivers/acpi/acpica/psloop.c
442
op = NULL;
drivers/acpi/acpica/psloop.c
458
status = acpi_ps_push_scope(parser_state, op,
drivers/acpi/acpica/psloop.c
463
acpi_ps_complete_op(walk_state, &op,
drivers/acpi/acpica/psloop.c
472
op = NULL;
drivers/acpi/acpica/psloop.c
481
acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/psloop.c
483
if (op->common.aml_opcode == AML_REGION_OP ||
drivers/acpi/acpica/psloop.c
484
op->common.aml_opcode == AML_DATA_REGION_OP) {
drivers/acpi/acpica/psloop.c
493
op->named.length =
drivers/acpi/acpica/psloop.c
494
(u32) (parser_state->aml - op->named.data);
drivers/acpi/acpica/psloop.c
505
op->named.length =
drivers/acpi/acpica/psloop.c
506
(u32) (parser_state->aml - op->named.data);
drivers/acpi/acpica/psloop.c
509
if (op->common.aml_opcode == AML_BANK_FIELD_OP) {
drivers/acpi/acpica/psloop.c
51
u8 * aml_op_start, union acpi_parse_object *op)
drivers/acpi/acpica/psloop.c
515
op->named.length =
drivers/acpi/acpica/psloop.c
516
(u32) (parser_state->aml - op->named.data);
drivers/acpi/acpica/psloop.c
522
walk_state->op = op;
drivers/acpi/acpica/psloop.c
523
walk_state->opcode = op->common.aml_opcode;
drivers/acpi/acpica/psloop.c
527
acpi_ps_next_parse_state(walk_state, op, status);
drivers/acpi/acpica/psloop.c
552
status = acpi_ps_complete_op(walk_state, &op, status);
drivers/acpi/acpica/psloop.c
559
status = acpi_ps_complete_final_op(walk_state, op, status);
drivers/acpi/acpica/psloop.c
60
op->common.aml_op_name));
drivers/acpi/acpica/psloop.c
62
switch (op->common.aml_opcode) {
drivers/acpi/acpica/psloop.c
74
op);
drivers/acpi/acpica/psloop.c
81
op,
drivers/acpi/acpica/psloop.c
98
switch (op->common.aml_opcode) {
drivers/acpi/acpica/psobject.c
157
union acpi_parse_object **op)
drivers/acpi/acpica/psobject.c
215
walk_state->op = NULL;
drivers/acpi/acpica/psobject.c
217
status = walk_state->descending_callback(walk_state, op);
drivers/acpi/acpica/psobject.c
226
if (!*op) {
drivers/acpi/acpica/psobject.c
230
status = acpi_ps_next_parse_state(walk_state, *op, status);
drivers/acpi/acpica/psobject.c
238
acpi_ps_append_arg(*op, unnamed_op->common.value.arg);
drivers/acpi/acpica/psobject.c
244
(*op)->common.inline_comment = unnamed_op->common.inline_comment;
drivers/acpi/acpica/psobject.c
245
(*op)->common.end_node_comment = unnamed_op->common.end_node_comment;
drivers/acpi/acpica/psobject.c
246
(*op)->common.close_brace_comment =
drivers/acpi/acpica/psobject.c
248
(*op)->common.name_comment = unnamed_op->common.name_comment;
drivers/acpi/acpica/psobject.c
249
(*op)->common.comment_list = unnamed_op->common.comment_list;
drivers/acpi/acpica/psobject.c
250
(*op)->common.end_blk_comment = unnamed_op->common.end_blk_comment;
drivers/acpi/acpica/psobject.c
251
(*op)->common.cv_filename = unnamed_op->common.cv_filename;
drivers/acpi/acpica/psobject.c
252
(*op)->common.cv_parent_filename =
drivers/acpi/acpica/psobject.c
254
(*op)->named.aml = unnamed_op->common.aml;
drivers/acpi/acpica/psobject.c
264
if ((*op)->common.aml_opcode == AML_REGION_OP ||
drivers/acpi/acpica/psobject.c
265
(*op)->common.aml_opcode == AML_DATA_REGION_OP) {
drivers/acpi/acpica/psobject.c
276
(*op)->named.data = aml_op_start;
drivers/acpi/acpica/psobject.c
277
(*op)->named.length = 0;
drivers/acpi/acpica/psobject.c
302
union acpi_parse_object *op;
drivers/acpi/acpica/psobject.c
321
op = acpi_ps_alloc_op(walk_state->opcode, aml_op_start);
drivers/acpi/acpica/psobject.c
322
if (!op) {
drivers/acpi/acpica/psobject.c
328
acpi_ps_build_named_op(walk_state, aml_op_start, op,
drivers/acpi/acpica/psobject.c
330
acpi_ps_free_op(op);
drivers/acpi/acpica/psobject.c
366
op->named.data = aml_op_start;
drivers/acpi/acpica/psobject.c
367
op->named.length = 0;
drivers/acpi/acpica/psobject.c
375
op->named.data = aml_op_start;
drivers/acpi/acpica/psobject.c
376
op->named.length = 0;
drivers/acpi/acpica/psobject.c
380
acpi_ps_append_arg(parent_scope, op);
drivers/acpi/acpica/psobject.c
390
op->common.flags |= ACPI_PARSEOP_TARGET;
drivers/acpi/acpica/psobject.c
401
op->common.flags |= ACPI_PARSEOP_TARGET;
drivers/acpi/acpica/psobject.c
410
walk_state->op = *new_op = op;
drivers/acpi/acpica/psobject.c
412
status = walk_state->descending_callback(walk_state, &op);
drivers/acpi/acpica/psobject.c
413
status = acpi_ps_next_parse_state(walk_state, op, status);
drivers/acpi/acpica/psobject.c
438
union acpi_parse_object **op, acpi_status status)
drivers/acpi/acpica/psobject.c
451
status2 = acpi_ps_complete_this_op(walk_state, *op);
drivers/acpi/acpica/psobject.c
456
*op = NULL;
drivers/acpi/acpica/psobject.c
473
acpi_ps_pop_scope(&(walk_state->parser_state), op,
drivers/acpi/acpica/psobject.c
477
if (*op) {
drivers/acpi/acpica/psobject.c
478
walk_state->op = *op;
drivers/acpi/acpica/psobject.c
480
acpi_ps_get_opcode_info((*op)->common.aml_opcode);
drivers/acpi/acpica/psobject.c
481
walk_state->opcode = (*op)->common.aml_opcode;
drivers/acpi/acpica/psobject.c
484
(void)acpi_ps_next_parse_state(walk_state, *op, status);
drivers/acpi/acpica/psobject.c
486
status2 = acpi_ps_complete_this_op(walk_state, *op);
drivers/acpi/acpica/psobject.c
499
while (!(*op) || ((*op)->common.aml_opcode != AML_WHILE_OP)) {
drivers/acpi/acpica/psobject.c
500
acpi_ps_pop_scope(&(walk_state->parser_state), op,
drivers/acpi/acpica/psobject.c
507
walk_state->op = *op;
drivers/acpi/acpica/psobject.c
509
acpi_ps_get_opcode_info((*op)->common.aml_opcode);
drivers/acpi/acpica/psobject.c
510
walk_state->opcode = (*op)->common.aml_opcode;
drivers/acpi/acpica/psobject.c
513
(void)acpi_ps_next_parse_state(walk_state, *op, status);
drivers/acpi/acpica/psobject.c
515
status2 = acpi_ps_complete_this_op(walk_state, *op);
drivers/acpi/acpica/psobject.c
526
if (*op) {
drivers/acpi/acpica/psobject.c
528
acpi_ps_complete_this_op(walk_state, *op);
drivers/acpi/acpica/psobject.c
538
acpi_ps_pop_scope(&(walk_state->parser_state), op,
drivers/acpi/acpica/psobject.c
542
} while (*op);
drivers/acpi/acpica/psobject.c
549
if (*op) {
drivers/acpi/acpica/psobject.c
555
if (((*op)->common.aml_opcode == AML_REGION_OP)
drivers/acpi/acpica/psobject.c
556
|| ((*op)->common.aml_opcode ==
drivers/acpi/acpica/psobject.c
558
acpi_ns_delete_children((*op)->common.
drivers/acpi/acpica/psobject.c
560
acpi_ns_remove_node((*op)->common.node);
drivers/acpi/acpica/psobject.c
561
(*op)->common.node = NULL;
drivers/acpi/acpica/psobject.c
562
acpi_ps_delete_parse_tree(*op);
drivers/acpi/acpica/psobject.c
566
acpi_ps_complete_this_op(walk_state, *op);
drivers/acpi/acpica/psobject.c
572
acpi_ps_pop_scope(&(walk_state->parser_state), op,
drivers/acpi/acpica/psobject.c
576
} while (*op);
drivers/acpi/acpica/psobject.c
582
if (*op == NULL) {
drivers/acpi/acpica/psobject.c
583
acpi_ps_pop_scope(parser_state, op,
drivers/acpi/acpica/psobject.c
609
acpi_ps_pop_scope(&(walk_state->parser_state), op,
drivers/acpi/acpica/psobject.c
612
ACPI_DEBUG_PRINT((ACPI_DB_PARSE, "Popped scope, Op=%p\n", *op));
drivers/acpi/acpica/psobject.c
614
*op = NULL;
drivers/acpi/acpica/psobject.c
637
union acpi_parse_object *op, acpi_status status)
drivers/acpi/acpica/psobject.c
651
op));
drivers/acpi/acpica/psobject.c
653
if (op) {
drivers/acpi/acpica/psobject.c
655
walk_state->op = op;
drivers/acpi/acpica/psobject.c
657
acpi_ps_get_opcode_info(op->common.
drivers/acpi/acpica/psobject.c
659
walk_state->opcode = op->common.aml_opcode;
drivers/acpi/acpica/psobject.c
664
acpi_ps_next_parse_state(walk_state, op,
drivers/acpi/acpica/psobject.c
668
acpi_ps_complete_op(walk_state, &op,
drivers/acpi/acpica/psobject.c
689
status = acpi_ps_complete_this_op(walk_state, op);
drivers/acpi/acpica/psobject.c
699
acpi_ps_pop_scope(&(walk_state->parser_state), &op,
drivers/acpi/acpica/psobject.c
703
} while (op);
drivers/acpi/acpica/psopinfo.c
131
const struct acpi_opcode_info *op;
drivers/acpi/acpica/psopinfo.c
133
op = acpi_ps_get_opcode_info(opcode);
drivers/acpi/acpica/psopinfo.c
137
return (op->name);
drivers/acpi/acpica/psparse.c
101
union acpi_parse_object *op)
drivers/acpi/acpica/psparse.c
109
ACPI_FUNCTION_TRACE_PTR(ps_complete_this_op, op);
drivers/acpi/acpica/psparse.c
113
if (!op) {
drivers/acpi/acpica/psparse.c
117
acpi_ex_stop_trace_opcode(op, walk_state);
drivers/acpi/acpica/psparse.c
129
if (op->common.parent) {
drivers/acpi/acpica/psparse.c
130
prev = op->common.parent->common.value.arg;
drivers/acpi/acpica/psparse.c
143
acpi_ps_get_opcode_info(op->common.parent->common.
drivers/acpi/acpica/psparse.c
158
op->common.aml);
drivers/acpi/acpica/psparse.c
169
if ((op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
171
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
173
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
175
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
177
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
179
|| (op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
183
op->common.aml);
drivers/acpi/acpica/psparse.c
188
if ((op->common.parent->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
192
if ((op->common.aml_opcode == AML_BUFFER_OP)
drivers/acpi/acpica/psparse.c
193
|| (op->common.aml_opcode == AML_PACKAGE_OP)
drivers/acpi/acpica/psparse.c
194
|| (op->common.aml_opcode ==
drivers/acpi/acpica/psparse.c
197
acpi_ps_alloc_op(op->common.
drivers/acpi/acpica/psparse.c
199
op->common.aml);
drivers/acpi/acpica/psparse.c
204
op->named.data;
drivers/acpi/acpica/psparse.c
206
op->named.length;
drivers/acpi/acpica/psparse.c
216
op->common.aml);
drivers/acpi/acpica/psparse.c
224
if (prev == op) {
drivers/acpi/acpica/psparse.c
230
op->common.parent;
drivers/acpi/acpica/psparse.c
232
replacement_op->common.node = op->common.node;
drivers/acpi/acpica/psparse.c
233
op->common.parent->common.value.arg =
drivers/acpi/acpica/psparse.c
235
replacement_op->common.next = op->common.next;
drivers/acpi/acpica/psparse.c
237
op->common.parent->common.value.arg =
drivers/acpi/acpica/psparse.c
238
op->common.next;
drivers/acpi/acpica/psparse.c
250
if (next == op) {
drivers/acpi/acpica/psparse.c
253
op->common.parent;
drivers/acpi/acpica/psparse.c
257
op->common.node;
drivers/acpi/acpica/psparse.c
261
op->common.next;
drivers/acpi/acpica/psparse.c
265
op->common.next;
drivers/acpi/acpica/psparse.c
277
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/psparse.c
298
union acpi_parse_object *op,
drivers/acpi/acpica/psparse.c
304
ACPI_FUNCTION_TRACE_PTR(ps_next_parse_state, op);
drivers/acpi/acpica/psparse.c
372
walk_state->prev_op = op;
drivers/acpi/acpica/psparse.c
373
walk_state->method_call_op = op;
drivers/acpi/acpica/psparse.c
375
(op->common.value.arg)->common.node;
drivers/acpi/acpica/psparse.c
380
acpi_ds_is_result_used(op, walk_state);
drivers/acpi/acpica/psscope.c
112
union acpi_parse_object *op,
drivers/acpi/acpica/psscope.c
117
ACPI_FUNCTION_TRACE_PTR(ps_push_scope, op);
drivers/acpi/acpica/psscope.c
125
scope->parse_scope.op = op;
drivers/acpi/acpica/psscope.c
166
union acpi_parse_object **op, u32 * arg_list, u32 * arg_count)
drivers/acpi/acpica/psscope.c
179
*op = scope->parse_scope.op;
drivers/acpi/acpica/psscope.c
190
*op = NULL;
drivers/acpi/acpica/psscope.c
196
"Popped Op %p Args %X\n", *op, *arg_count));
drivers/acpi/acpica/psscope.c
32
return (parser_state->scope->parse_scope.op);
drivers/acpi/acpica/psscope.c
84
scope->parse_scope.op = root_op;
drivers/acpi/acpica/pstree.c
101
if (!op) {
drivers/acpi/acpica/pstree.c
107
op_info = acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/pstree.c
113
op->common.aml_opcode));
drivers/acpi/acpica/pstree.c
128
if (op->common.value.arg) {
drivers/acpi/acpica/pstree.c
132
prev_arg = op->common.value.arg;
drivers/acpi/acpica/pstree.c
140
op->common.value.arg = arg;
drivers/acpi/acpica/pstree.c
146
arg->common.parent = op;
drivers/acpi/acpica/pstree.c
149
op->common.arg_list_length++;
drivers/acpi/acpica/pstree.c
170
union acpi_parse_object *op)
drivers/acpi/acpica/pstree.c
178
if (!op) {
drivers/acpi/acpica/pstree.c
184
next = acpi_ps_get_arg(op, 0);
drivers/acpi/acpica/pstree.c
192
next = op->common.next;
drivers/acpi/acpica/pstree.c
200
parent = op->common.parent;
drivers/acpi/acpica/pstree.c
204
while (arg && (arg != origin) && (arg != op)) {
drivers/acpi/acpica/pstree.c
21
union acpi_parse_object *acpi_ps_get_child(union acpi_parse_object *op);
drivers/acpi/acpica/pstree.c
225
op = parent;
drivers/acpi/acpica/pstree.c
246
union acpi_parse_object *acpi_ps_get_child(union acpi_parse_object *op)
drivers/acpi/acpica/pstree.c
252
switch (op->common.aml_opcode) {
drivers/acpi/acpica/pstree.c
259
child = acpi_ps_get_arg(op, 0);
drivers/acpi/acpica/pstree.c
270
child = acpi_ps_get_arg(op, 1);
drivers/acpi/acpica/pstree.c
276
child = acpi_ps_get_arg(op, 2);
drivers/acpi/acpica/pstree.c
282
child = acpi_ps_get_arg(op, 3);
drivers/acpi/acpica/pstree.c
37
union acpi_parse_object *acpi_ps_get_arg(union acpi_parse_object *op, u32 argn)
drivers/acpi/acpica/pstree.c
52
op_info = acpi_ps_get_opcode_info(op->common.aml_opcode);
drivers/acpi/acpica/pstree.c
71
arg = op->common.value.arg;
drivers/acpi/acpica/pstree.c
94
acpi_ps_append_arg(union acpi_parse_object *op, union acpi_parse_object *arg)
drivers/acpi/acpica/psutils.c
111
op = acpi_os_acquire_object(acpi_gbl_ps_node_cache);
drivers/acpi/acpica/psutils.c
115
op = acpi_os_acquire_object(acpi_gbl_ps_node_ext_cache);
drivers/acpi/acpica/psutils.c
120
if (op) {
drivers/acpi/acpica/psutils.c
121
acpi_ps_init_op(op, opcode);
drivers/acpi/acpica/psutils.c
122
op->common.aml = aml;
drivers/acpi/acpica/psutils.c
123
op->common.flags = flags;
drivers/acpi/acpica/psutils.c
124
ASL_CV_CLEAR_OP_COMMENTS(op);
drivers/acpi/acpica/psutils.c
127
acpi_gbl_current_scope = op;
drivers/acpi/acpica/psutils.c
131
ASL_CV_TRANSFER_COMMENTS(op);
drivers/acpi/acpica/psutils.c
135
return (op);
drivers/acpi/acpica/psutils.c
151
void acpi_ps_free_op(union acpi_parse_object *op)
drivers/acpi/acpica/psutils.c
155
ASL_CV_CLEAR_OP_COMMENTS(op);
drivers/acpi/acpica/psutils.c
156
if (op->common.aml_opcode == AML_INT_RETURN_VALUE_OP) {
drivers/acpi/acpica/psutils.c
158
"Free retval op: %p\n", op));
drivers/acpi/acpica/psutils.c
161
if (op->common.flags & ACPI_PARSEOP_GENERIC) {
drivers/acpi/acpica/psutils.c
162
(void)acpi_os_release_object(acpi_gbl_ps_node_cache, op);
drivers/acpi/acpica/psutils.c
164
(void)acpi_os_release_object(acpi_gbl_ps_node_ext_cache, op);
drivers/acpi/acpica/psutils.c
187
u32 acpi_ps_get_name(union acpi_parse_object * op)
drivers/acpi/acpica/psutils.c
192
if (op->common.flags & ACPI_PARSEOP_GENERIC) {
drivers/acpi/acpica/psutils.c
198
return (op->named.name);
drivers/acpi/acpica/psutils.c
204
void acpi_ps_set_name(union acpi_parse_object *op, u32 name)
drivers/acpi/acpica/psutils.c
209
if (op->common.flags & ACPI_PARSEOP_GENERIC) {
drivers/acpi/acpica/psutils.c
213
op->named.name = name;
drivers/acpi/acpica/psutils.c
56
void acpi_ps_init_op(union acpi_parse_object *op, u16 opcode)
drivers/acpi/acpica/psutils.c
60
op->common.descriptor_type = ACPI_DESC_TYPE_PARSER;
drivers/acpi/acpica/psutils.c
61
op->common.aml_opcode = opcode;
drivers/acpi/acpica/psutils.c
63
ACPI_DISASM_ONLY_MEMBERS(acpi_ut_safe_strncpy(op->common.aml_op_name,
drivers/acpi/acpica/psutils.c
66
sizeof(op->common.
drivers/acpi/acpica/psutils.c
87
union acpi_parse_object *op;
drivers/acpi/acpica/pswalk.c
100
op = parent;
drivers/acpi/acpica/pswalk.c
31
union acpi_parse_object *op = subtree_root;
drivers/acpi/acpica/pswalk.c
42
while (op) {
drivers/acpi/acpica/pswalk.c
43
if (op != parent) {
drivers/acpi/acpica/pswalk.c
54
acpi_ps_get_opcode_name(op->
drivers/acpi/acpica/pswalk.c
57
op);
drivers/acpi/acpica/pswalk.c
59
if (op->named.aml_opcode == AML_INT_NAMEPATH_OP) {
drivers/acpi/acpica/pswalk.c
61
op->common.value.string);
drivers/acpi/acpica/pswalk.c
63
if (op->named.aml_opcode == AML_STRING_OP) {
drivers/acpi/acpica/pswalk.c
65
op->common.value.string);
drivers/acpi/acpica/pswalk.c
72
next = acpi_ps_get_arg(op, 0);
drivers/acpi/acpica/pswalk.c
77
op = next;
drivers/acpi/acpica/pswalk.c
85
next = op->common.next;
drivers/acpi/acpica/pswalk.c
86
parent = op->common.parent;
drivers/acpi/acpica/pswalk.c
88
acpi_ps_free_op(op);
drivers/acpi/acpica/pswalk.c
92
if (op == subtree_root) {
drivers/acpi/acpica/pswalk.c
97
op = next;
drivers/acpi/acpica/psxface.c
124
op = acpi_ps_create_scope_op(info->obj_desc->method.aml_start);
drivers/acpi/acpica/psxface.c
125
if (!op) {
drivers/acpi/acpica/psxface.c
141
status = acpi_ds_init_aml_walk(walk_state, op, info->node,
drivers/acpi/acpica/psxface.c
195
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/psxface.c
242
union acpi_parse_object *op = NULL;
drivers/acpi/acpica/psxface.c
249
op = acpi_ps_create_scope_op(info->obj_desc->method.aml_start);
drivers/acpi/acpica/psxface.c
250
if (!op) {
drivers/acpi/acpica/psxface.c
265
status = acpi_ds_init_aml_walk(walk_state, op, info->node,
drivers/acpi/acpica/psxface.c
303
if (op) {
drivers/acpi/acpica/psxface.c
304
acpi_ps_delete_parse_tree(op);
drivers/acpi/acpica/psxface.c
87
union acpi_parse_object *op;
drivers/acpi/acpica/uttrack.c
664
descriptor->op.asl.
drivers/ata/pata_mpc52xx.c
672
static int mpc52xx_ata_probe(struct platform_device *op)
drivers/ata/pata_mpc52xx.c
686
ipb_freq = mpc5xxx_get_bus_frequency(&op->dev);
drivers/ata/pata_mpc52xx.c
688
dev_err(&op->dev, "could not determine IPB bus frequency\n");
drivers/ata/pata_mpc52xx.c
694
rv = of_address_to_resource(op->dev.of_node, 0, &res_mem);
drivers/ata/pata_mpc52xx.c
696
dev_err(&op->dev, "could not determine device base address\n");
drivers/ata/pata_mpc52xx.c
700
if (!devm_request_mem_region(&op->dev, res_mem.start,
drivers/ata/pata_mpc52xx.c
702
dev_err(&op->dev, "error requesting register region\n");
drivers/ata/pata_mpc52xx.c
706
ata_regs = devm_ioremap(&op->dev, res_mem.start, sizeof(*ata_regs));
drivers/ata/pata_mpc52xx.c
708
dev_err(&op->dev, "error mapping device registers\n");
drivers/ata/pata_mpc52xx.c
726
prop = of_get_property(op->dev.of_node, "mwdma-mode", &proplen);
drivers/ata/pata_mpc52xx.c
729
prop = of_get_property(op->dev.of_node, "udma-mode", &proplen);
drivers/ata/pata_mpc52xx.c
733
ata_irq = irq_of_parse_and_map(op->dev.of_node, 0);
drivers/ata/pata_mpc52xx.c
735
dev_err(&op->dev, "error mapping irq\n");
drivers/ata/pata_mpc52xx.c
740
priv = devm_kzalloc(&op->dev, sizeof(*priv), GFP_KERNEL);
drivers/ata/pata_mpc52xx.c
764
dev_err(&op->dev, "bestcomm initialization failed\n");
drivers/ata/pata_mpc52xx.c
770
rv = devm_request_irq(&op->dev, task_irq, &mpc52xx_ata_task_irq, 0,
drivers/ata/pata_mpc52xx.c
773
dev_err(&op->dev, "error requesting DMA IRQ\n");
drivers/ata/pata_mpc52xx.c
781
dev_err(&op->dev, "error initializing hardware\n");
drivers/ata/pata_mpc52xx.c
786
rv = mpc52xx_ata_init_one(&op->dev, priv, res_mem.start,
drivers/ata/pata_mpc52xx.c
789
dev_err(&op->dev, "error registering with ATA layer\n");
drivers/ata/pata_mpc52xx.c
803
static void mpc52xx_ata_remove(struct platform_device *op)
drivers/ata/pata_mpc52xx.c
805
struct ata_host *host = platform_get_drvdata(op);
drivers/ata/pata_mpc52xx.c
810
ata_platform_remove_one(op);
drivers/ata/pata_mpc52xx.c
821
mpc52xx_ata_suspend(struct platform_device *op, pm_message_t state)
drivers/ata/pata_mpc52xx.c
823
struct ata_host *host = platform_get_drvdata(op);
drivers/ata/pata_mpc52xx.c
830
mpc52xx_ata_resume(struct platform_device *op)
drivers/ata/pata_mpc52xx.c
832
struct ata_host *host = platform_get_drvdata(op);
drivers/ata/pata_parport/on20.c
111
op(1); vl(1); op(0);
drivers/ata/pata_parport/on20.c
43
op(1); vl(r); op(0);
drivers/ata/pata_parport/on20.c
64
op(1); vl(r);
drivers/ata/pata_parport/on20.c
65
op(0); vl(val);
drivers/ata/pata_parport/on20.c
66
op(0); vl(val);
drivers/ata/pata_parport/on20.c
76
op(2); vl(8); op(2); vl(9);
drivers/ata/pata_parport/on20.c
78
op(2); vl(0); op(2); vl(8);
drivers/ata/pata_parport/on20.c
93
op(1); vl(1); op(0);
drivers/ata/sata_fsl.c
1541
static int sata_fsl_suspend(struct platform_device *op, pm_message_t state)
drivers/ata/sata_fsl.c
1543
struct ata_host *host = platform_get_drvdata(op);
drivers/ata/sata_fsl.c
1549
static int sata_fsl_resume(struct platform_device *op)
drivers/ata/sata_fsl.c
1551
struct ata_host *host = platform_get_drvdata(op);
drivers/ata/sata_fsl.c
1560
dev_err(&op->dev, "Error initializing hardware\n");
drivers/atm/fore200e.c
2525
static int fore200e_sba_probe(struct platform_device *op)
drivers/atm/fore200e.c
2536
fore200e->dev = &op->dev;
drivers/atm/fore200e.c
2537
fore200e->irq = op->archdata.irqs[0];
drivers/atm/fore200e.c
2538
fore200e->phys_base = op->resource[0].start;
drivers/atm/fore200e.c
2542
err = fore200e_init(fore200e, &op->dev);
drivers/atm/fore200e.c
2550
dev_set_drvdata(&op->dev, fore200e);
drivers/atm/fore200e.c
2555
static void fore200e_sba_remove(struct platform_device *op)
drivers/atm/fore200e.c
2557
struct fore200e *fore200e = dev_get_drvdata(&op->dev);
drivers/atm/fore200e.c
669
struct platform_device *op = to_platform_device(fore200e->dev);
drivers/atm/fore200e.c
673
fore200e->regs.sba.hcr = of_ioremap(&op->resource[0], 0, SBA200E_HCR_LENGTH, "SBA HCR");
drivers/atm/fore200e.c
674
fore200e->regs.sba.bsr = of_ioremap(&op->resource[1], 0, SBA200E_BSR_LENGTH, "SBA BSR");
drivers/atm/fore200e.c
675
fore200e->regs.sba.isr = of_ioremap(&op->resource[2], 0, SBA200E_ISR_LENGTH, "SBA ISR");
drivers/atm/fore200e.c
676
fore200e->virt_base = of_ioremap(&op->resource[3], 0, SBA200E_RAM_LENGTH, "SBA RAM");
drivers/atm/fore200e.c
688
bursts = of_getintprop_default(op->dev.of_node->parent, "burst-sizes", 0x00);
drivers/atm/fore200e.c
691
sbus_set_sbus64(&op->dev, bursts);
drivers/atm/fore200e.c
699
struct platform_device *op = to_platform_device(fore200e->dev);
drivers/atm/fore200e.c
701
of_iounmap(&op->resource[0], fore200e->regs.sba.hcr, SBA200E_HCR_LENGTH);
drivers/atm/fore200e.c
702
of_iounmap(&op->resource[1], fore200e->regs.sba.bsr, SBA200E_BSR_LENGTH);
drivers/atm/fore200e.c
703
of_iounmap(&op->resource[2], fore200e->regs.sba.isr, SBA200E_ISR_LENGTH);
drivers/atm/fore200e.c
704
of_iounmap(&op->resource[3], fore200e->virt_base, SBA200E_RAM_LENGTH);
drivers/atm/fore200e.c
715
struct platform_device *op = to_platform_device(fore200e->dev);
drivers/atm/fore200e.c
719
prop = of_get_property(op->dev.of_node, "madaddrlo2", &len);
drivers/atm/fore200e.c
724
prop = of_get_property(op->dev.of_node, "madaddrhi4", &len);
drivers/atm/fore200e.c
729
prom->serial_number = of_getintprop_default(op->dev.of_node,
drivers/atm/fore200e.c
731
prom->hw_revision = of_getintprop_default(op->dev.of_node,
drivers/atm/fore200e.c
739
struct platform_device *op = to_platform_device(fore200e->dev);
drivers/atm/fore200e.c
742
regs = of_get_property(op->dev.of_node, "reg", NULL);
drivers/atm/fore200e.c
745
(regs ? regs->which_io : 0), op->dev.of_node);
drivers/base/devres.c
57
const char *op)
drivers/base/devres.c
61
op, node, node->name, node->size);
drivers/base/devres.c
64
#define devres_dbg(dev, node, op) do {} while (0)
drivers/base/devres.c
68
const char *op)
drivers/base/devres.c
70
trace_devres_log(dev, op, node, node->name, node->size);
drivers/base/devres.c
71
devres_dbg(dev, node, op);
drivers/base/power/generic_ops.c
11
#define CALL_PM_OP(dev, op) \
drivers/base/power/generic_ops.c
15
pm && pm->op ? pm->op(_dev) : 0; \
drivers/base/trace.h
20
TP_PROTO(struct device *dev, const char *op, void *node, const char *name, size_t size),
drivers/base/trace.h
21
TP_ARGS(dev, op, node, name, size),
drivers/base/trace.h
25
__field(const char *, op)
drivers/base/trace.h
32
__entry->op = op;
drivers/base/trace.h
38
__entry->op, __entry->node, __get_str(name), __entry->size)
drivers/base/trace.h
42
TP_PROTO(struct device *dev, const char *op, void *node, const char *name, size_t size),
drivers/base/trace.h
43
TP_ARGS(dev, op, node, name, size)
drivers/block/drbd/drbd_actlog.c
127
sector_t sector, enum req_op op)
drivers/block/drbd/drbd_actlog.c
138
if ((op == REQ_OP_WRITE) && !test_bit(MD_NO_FUA, &device->flags))
drivers/block/drbd/drbd_actlog.c
142
bio = bio_alloc_bioset(bdev->md_bdev, 1, op | op_flags, GFP_NOIO,
drivers/block/drbd/drbd_actlog.c
151
if (op != REQ_OP_WRITE && device->state.disk == D_DISKLESS && device->ldev == NULL)
drivers/block/drbd/drbd_actlog.c
164
if (drbd_insert_fault(device, (op == REQ_OP_WRITE) ? DRBD_FAULT_MD_WR : DRBD_FAULT_MD_RD))
drivers/block/drbd/drbd_actlog.c
178
sector_t sector, enum req_op op)
drivers/block/drbd/drbd_actlog.c
187
(unsigned long long)sector, (op == REQ_OP_WRITE) ? "WRITE" : "READ",
drivers/block/drbd/drbd_actlog.c
195
(op == REQ_OP_WRITE) ? "WRITE" : "READ");
drivers/block/drbd/drbd_actlog.c
197
err = _drbd_md_sync_page_io(device, bdev, sector, op);
drivers/block/drbd/drbd_actlog.c
201
(op == REQ_OP_WRITE) ? "WRITE" : "READ", err);
drivers/block/drbd/drbd_bitmap.c
1043
bio = bio_alloc_bioset(device->ldev->md_bdev, 1, op, GFP_NOIO,
drivers/block/drbd/drbd_bitmap.c
1050
if (drbd_insert_fault(device, (op == REQ_OP_WRITE) ? DRBD_FAULT_MD_WR : DRBD_FAULT_MD_RD)) {
drivers/block/drbd/drbd_bitmap.c
993
enum req_op op = ctx->flags & BM_AIO_READ ? REQ_OP_READ : REQ_OP_WRITE;
drivers/block/drbd/drbd_int.h
1420
struct drbd_backing_dev *bdev, sector_t sector, enum req_op op);
drivers/block/null_blk/main.c
1372
blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/main.c
1377
if (op == REQ_OP_DISCARD)
drivers/block/null_blk/main.c
1423
blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/main.c
1434
ret = null_handle_memory_backed(cmd, op, sector, nr_sectors);
drivers/block/null_blk/main.c
1443
sector_t nr_sectors, enum req_op op)
drivers/block/null_blk/main.c
1449
if (op == REQ_OP_FLUSH) {
drivers/block/null_blk/main.c
1455
sts = null_process_zoned_cmd(cmd, op, sector, nr_sectors);
drivers/block/null_blk/main.c
1457
sts = null_process_cmd(cmd, op, sector, nr_sectors);
drivers/block/null_blk/null_blk.h
134
blk_status_t null_process_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
138
blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
148
blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/null_blk.h
167
enum req_op op, sector_t sector, sector_t nr_sectors)
drivers/block/null_blk/trace.h
44
__field_struct(enum req_op, op)
drivers/block/null_blk/trace.h
49
__entry->op = req_op(blk_mq_rq_from_pdu(cmd));
drivers/block/null_blk/trace.h
57
blk_op_str(__entry->op),
drivers/block/null_blk/zoned.c
646
static blk_status_t null_zone_mgmt(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/zoned.c
655
if (op == REQ_OP_ZONE_RESET_ALL) {
drivers/block/null_blk/zoned.c
681
switch (op) {
drivers/block/null_blk/zoned.c
708
blk_status_t null_process_zoned_cmd(struct nullb_cmd *cmd, enum req_op op,
drivers/block/null_blk/zoned.c
715
switch (op) {
drivers/block/null_blk/zoned.c
725
return null_zone_mgmt(cmd, op, sector);
drivers/block/null_blk/zoned.c
733
sts = null_process_cmd(cmd, op, sector, nr_sectors);
drivers/block/ps3disk.c
104
const char *op = write ? "write" : "read";
drivers/block/ps3disk.c
117
__func__, __LINE__, op, n, blk_rq_sectors(req));
drivers/block/ps3disk.c
123
__func__, __LINE__, op, sectors, start_sector);
drivers/block/ps3disk.c
138
__LINE__, op, res);
drivers/block/ps3disk.c
209
const char *op;
drivers/block/ps3disk.c
237
op = "flush";
drivers/block/ps3disk.c
240
op = read ? "read" : "write";
drivers/block/ps3disk.c
244
__LINE__, op, status);
drivers/block/ps3disk.c
248
__LINE__, op);
drivers/block/ps3vram.c
533
const char *op = write ? "write" : "read";
drivers/block/ps3vram.c
545
dev_dbg(&dev->core, " %s %zu bytes at offset %llu\n", op,
drivers/block/ps3vram.c
553
dev_err(&dev->core, "%s failed\n", op);
drivers/block/ps3vram.c
558
dev_err(&dev->core, "Short %s\n", op);
drivers/block/ps3vram.c
566
dev_dbg(&dev->core, "%s completed\n", op);
drivers/block/sunvdc.c
471
u8 op;
drivers/block/sunvdc.c
480
op = VD_OP_BREAD;
drivers/block/sunvdc.c
483
op = VD_OP_BWRITE;
drivers/block/sunvdc.c
508
desc->operation = op;
drivers/block/sunvdc.c
572
static int generic_request(struct vdc_port *port, u8 op, void *buf, int len)
drivers/block/sunvdc.c
582
if (!(((u64)1 << (u64)op) & port->operations))
drivers/block/sunvdc.c
585
switch (op) {
drivers/block/sunvdc.c
678
desc->operation = op;
drivers/block/xen-blkback/blkback.c
148
unsigned short op, int st);
drivers/block/xen-blkback/blkback.c
1489
unsigned short op, int st)
drivers/block/xen-blkback/blkback.c
1517
resp->operation = op;
drivers/block/xen-blkfront.c
1540
unsigned int op;
drivers/block/xen-blkfront.c
1566
op = rinfo->shadow[id].req.operation;
drivers/block/xen-blkfront.c
1567
if (op == BLKIF_OP_INDIRECT)
drivers/block/xen-blkfront.c
1568
op = rinfo->shadow[id].req.u.indirect.indirect_op;
drivers/block/xen-blkfront.c
1569
if (bret.operation != op) {
drivers/block/xen-blkfront.c
1571
info->gd->disk_name, bret.operation, op);
drivers/block/xen-blkfront.c
426
static const char *op_name(int op)
drivers/block/xen-blkfront.c
435
if (op < 0 || op >= ARRAY_SIZE(names))
drivers/block/xen-blkfront.c
438
if (!names[op])
drivers/block/xen-blkfront.c
441
return names[op];
drivers/block/zloop.c
1399
struct zloop_ctl_op *op;
drivers/block/zloop.c
1411
op = &zloop_ctl_ops[i];
drivers/block/zloop.c
1412
if (!op->name) {
drivers/block/zloop.c
1417
if (!strncmp(buf, op->name, strlen(op->name)))
drivers/block/zloop.c
1421
if (count <= strlen(op->name))
drivers/block/zloop.c
1424
opts_buf = buf + strlen(op->name);
drivers/block/zloop.c
1432
switch (op->code) {
drivers/bluetooth/btmtk.c
1347
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtk.c
1376
wmt_params.op = BTMTK_WMT_SEMAPHORE;
drivers/bluetooth/btmtk.c
1418
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtk.c
1470
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtk.c
180
wmt_params.op = BTMTK_WMT_PATCH_DWNLD;
drivers/bluetooth/btmtk.c
209
wmt_params.op = BTMTK_WMT_PATCH_DWNLD;
drivers/bluetooth/btmtk.c
269
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtk.c
294
wmt_params.op = BTMTK_WMT_PATCH_DWNLD;
drivers/bluetooth/btmtk.c
321
wmt_params.op = BTMTK_WMT_RST;
drivers/bluetooth/btmtk.c
602
hdr->op = wmt_params->op;
drivers/bluetooth/btmtk.c
659
if (wmt_evt->whdr.op != hdr->op) {
drivers/bluetooth/btmtk.c
661
wmt_evt->whdr.op, hdr->op);
drivers/bluetooth/btmtk.c
666
switch (wmt_evt->whdr.op) {
drivers/bluetooth/btmtk.c
710
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtk.h
137
u8 op;
drivers/bluetooth/btmtk.h
69
u8 op;
drivers/bluetooth/btmtksdio.c
1209
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtksdio.c
169
hdr->op = wmt_params->op;
drivers/bluetooth/btmtksdio.c
208
if (wmt_evt->whdr.op != hdr->op) {
drivers/bluetooth/btmtksdio.c
210
wmt_evt->whdr.op, hdr->op);
drivers/bluetooth/btmtksdio.c
215
switch (wmt_evt->whdr.op) {
drivers/bluetooth/btmtksdio.c
766
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtksdio.c
791
wmt_params.op = BTMTK_WMT_SEMAPHORE;
drivers/bluetooth/btmtksdio.c
832
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtksdio.c
888
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtksdio.c
918
wmt_params.op = BTMTK_WMT_REGISTER;
drivers/bluetooth/btmtksdio.c
947
wmt_params.op = BTMTK_WMT_REGISTER;
drivers/bluetooth/btmtkuart.c
116
hdr->op = wmt_params->op;
drivers/bluetooth/btmtkuart.c
155
if (wmt_evt->whdr.op != hdr->op) {
drivers/bluetooth/btmtkuart.c
157
wmt_evt->whdr.op, hdr->op);
drivers/bluetooth/btmtkuart.c
162
switch (wmt_evt->whdr.op) {
drivers/bluetooth/btmtkuart.c
510
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtkuart.c
537
wmt_params.op = BTMTK_WMT_HIF;
drivers/bluetooth/btmtkuart.c
571
wmt_params.op = BTMTK_WMT_TEST;
drivers/bluetooth/btmtkuart.c
606
wmt_params.op = BTMTK_WMT_WAKEUP;
drivers/bluetooth/btmtkuart.c
625
wmt_params.op = BTMTK_WMT_SEMAPHORE;
drivers/bluetooth/btmtkuart.c
666
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/bluetooth/btmtkuart.c
711
wmt_params.op = BTMTK_WMT_FUNC_CTRL;
drivers/char/hw_random/n2-drv.c
392
dev_err(&np->op->dev,
drivers/char/hw_random/n2-drv.c
440
dev_err(&np->op->dev, "RNG error, retesting\n");
drivers/char/hw_random/n2-drv.c
518
dev_err(&np->op->dev, "Test buffer slot %d [0x%016llx]\n",
drivers/char/hw_random/n2-drv.c
550
dev_err(&np->op->dev, "Selftest failed on unit %lu\n", unit);
drivers/char/hw_random/n2-drv.c
553
dev_info(&np->op->dev, "Selftest passed on unit %lu\n", unit);
drivers/char/hw_random/n2-drv.c
679
dev_info(&np->op->dev, "RNG ready\n");
drivers/char/hw_random/n2-drv.c
683
dev_err(&np->op->dev, "Self-test retries failed, RNG not ready\n");
drivers/char/hw_random/n2-drv.c
697
static int n2rng_probe(struct platform_device *op)
drivers/char/hw_random/n2-drv.c
703
np = devm_kzalloc(&op->dev, sizeof(*np), GFP_KERNEL);
drivers/char/hw_random/n2-drv.c
706
np->op = op;
drivers/char/hw_random/n2-drv.c
707
np->data = (struct n2rng_template *)device_get_match_data(&op->dev);
drivers/char/hw_random/n2-drv.c
723
dev_err(&op->dev, "Cannot register suitable "
drivers/char/hw_random/n2-drv.c
731
dev_err(&op->dev, "multi-unit-capable RNG requires "
drivers/char/hw_random/n2-drv.c
736
np->num_units = of_getintprop_default(op->dev.of_node,
drivers/char/hw_random/n2-drv.c
739
dev_err(&op->dev, "VF RNG lacks rng-#units property\n");
drivers/char/hw_random/n2-drv.c
746
dev_info(&op->dev, "Registered RNG HVAPI major %lu minor %lu\n",
drivers/char/hw_random/n2-drv.c
748
np->units = devm_kcalloc(&op->dev, np->num_units, sizeof(*np->units),
drivers/char/hw_random/n2-drv.c
758
dev_info(&op->dev, "Found %s RNG, units: %d\n",
drivers/char/hw_random/n2-drv.c
767
err = devm_hwrng_register(&op->dev, &np->hwrng);
drivers/char/hw_random/n2-drv.c
771
platform_set_drvdata(op, np);
drivers/char/hw_random/n2-drv.c
784
static void n2rng_remove(struct platform_device *op)
drivers/char/hw_random/n2-drv.c
786
struct n2rng *np = platform_get_drvdata(op);
drivers/char/hw_random/n2rng.h
99
struct platform_device *op;
drivers/char/ipmi/ipmi_si_hotmod.c
103
static int parse_hotmod_str(const char *curr, enum hotmod_op *op,
drivers/char/ipmi/ipmi_si_hotmod.c
114
*op = ival;
drivers/char/ipmi/ipmi_si_hotmod.c
197
enum hotmod_op op;
drivers/char/ipmi/ipmi_si_hotmod.c
206
rv = parse_hotmod_str(curr, &op, &h);
drivers/char/ipmi/ipmi_si_hotmod.c
210
if (op == HM_ADD) {
drivers/char/xillybus/xillybus_of.c
34
static int xilly_drv_probe(struct platform_device *op)
drivers/char/xillybus/xillybus_of.c
36
struct device *dev = &op->dev;
drivers/char/xillybus/xillybus_of.c
50
endpoint->registers = devm_platform_ioremap_resource(op, 0);
drivers/char/xillybus/xillybus_of.c
54
irq = platform_get_irq(op, 0);
drivers/char/xillybus/xillybus_of.c
67
static void xilly_drv_remove(struct platform_device *op)
drivers/char/xillybus/xillybus_of.c
69
struct device *dev = &op->dev;
drivers/comedi/drivers/usbduxfast.c
184
u8 len, u8 op, u8 out, u8 log)
drivers/comedi/drivers/usbduxfast.c
190
devpriv->duxbuf[1 + 0x08 + index] = op;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
151
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
162
skcipher_request_set_tfm(&ctx->fallback_req, op->fallback_tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
179
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
18
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
180
struct sun4i_ss_ctx *ss = op->ss;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
19
struct sun4i_ss_ctx *ss = op->ss;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
254
for (i = 0; i < op->keylen / 4; i++)
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
255
writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
400
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
404
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
411
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
415
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
423
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
427
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
434
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
438
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
446
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
450
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
457
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
461
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
469
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
473
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
480
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
484
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
492
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
496
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
503
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
507
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
515
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
519
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
526
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
530
op->keymode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
536
struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
541
memset(op, 0, sizeof(struct sun4i_tfm_ctx));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
545
op->ss = algt->ss;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
547
op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
548
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
549
dev_err(op->ss->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
550
name, PTR_ERR(op->fallback_tfm));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
551
return PTR_ERR(op->fallback_tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
556
crypto_skcipher_reqsize(op->fallback_tfm));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
558
err = pm_runtime_resume_and_get(op->ss->dev);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
564
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
570
struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
572
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
573
pm_runtime_put(op->ss->dev);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
580
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
581
struct sun4i_ss_ctx *ss = op->ss;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
585
op->keymode = SS_AES_128BITS;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
588
op->keymode = SS_AES_192BITS;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
591
op->keymode = SS_AES_256BITS;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
597
op->keylen = keylen;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
598
memcpy(op->key, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
600
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
601
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
603
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
610
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
617
op->keylen = keylen;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
618
memcpy(op->key, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
62
for (i = 0; i < op->keylen / 4; i++)
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
620
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
621
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
623
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
63
writesl(ss->base + SS_KEY0 + i * 4, &op->key[i], 1);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
630
struct sun4i_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
637
op->keylen = keylen;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
638
memcpy(op->key, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
640
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
641
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-cipher.c
643
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
106
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
110
octx->count = op->byte_count + op->len;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
112
memcpy(octx->buffer, op->buf, op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
114
if (op->byte_count) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
116
octx->state[i] = op->hash[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
130
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
136
op->byte_count = ictx->count & ~0x3F;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
137
op->len = ictx->count & 0x3F;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
139
memcpy(op->buf, ictx->buffer, op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
142
op->hash[i] = ictx->state[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
192
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
20
struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
206
op->byte_count, areq->nbytes, op->mode,
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
207
op->len, op->hash[0]);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
209
if (unlikely(!areq->nbytes) && !(op->flags & SS_HASH_FINAL))
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
213
if (unlikely(areq->nbytes > UINT_MAX - op->len)) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
218
if (op->len + areq->nbytes < 64 && !(op->flags & SS_HASH_FINAL)) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
221
op->buf + op->len, areq->nbytes, 0);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
222
op->len += copied;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
232
if (op->byte_count) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
235
writel(op->hash[i], ss->base + SS_IV0 + i * 4);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
238
writel(op->mode | SS_ENABLED | ivmode, ss->base + SS_CTL);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
240
if (!(op->flags & SS_HASH_UPDATE))
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
244
if (!(op->flags & SS_HASH_FINAL)) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
245
end = ((areq->nbytes + op->len) / 64) * 64 - op->len;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
25
memset(op, 0, sizeof(struct sun4i_tfm_ctx));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
258
end = ((areq->nbytes + op->len) / 4) * 4 - op->len;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
268
if (i == 1 && !op->len && areq->nbytes)
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
28
op->ss = algt->ss;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
283
if (op->len || (mi.length - in_i) < 4) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
289
while (op->len < 64 && i < end) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
291
in_r = min(end - i, 64 - op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
293
memcpy(op->buf + op->len, mi.addr + in_i, in_r);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
294
op->len += in_r;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
30
err = pm_runtime_resume_and_get(op->ss->dev);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
302
if (op->len > 3 && !(op->len % 4)) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
304
writesl(ss->base + SS_RXFIFO, op->buf,
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
305
op->len / 4);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
306
op->byte_count += op->len;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
307
op->len = 0;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
317
op->byte_count += todo * 4;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
337
while (i < areq->nbytes && in_i < mi.length && op->len < 64) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
339
in_r = min(areq->nbytes - i, 64 - op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
341
memcpy(op->buf + op->len, mi.addr + in_i, in_r);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
342
op->len += in_r;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
359
if (op->flags & SS_HASH_FINAL)
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
362
writel(op->mode | SS_ENABLED | SS_DATA_END, ss->base + SS_CTL);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
386
op->hash[i] = readl(ss->base + SS_MD0 + i * 4);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
409
if (op->len) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
41
struct sun4i_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
410
nwait = op->len / 4;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
412
writesl(ss->base + SS_RXFIFO, op->buf, nwait);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
413
op->byte_count += 4 * nwait;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
416
nbw = op->len - 4 * nwait;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
418
wb = le32_to_cpup((__le32 *)(op->buf + nwait * 4));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
421
op->byte_count += nbw;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
43
pm_runtime_put(op->ss->dev);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
435
fill = 64 - (op->byte_count % 64);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
445
if (op->mode == SS_OP_SHA1) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
447
*bits = cpu_to_be64(op->byte_count << 3);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
451
*bits = cpu_to_le64(op->byte_count << 3);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
457
writel(op->mode | SS_ENABLED | SS_DATA_END, ss->base + SS_CTL);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
487
if (op->mode == SS_OP_SHA1) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
49
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
510
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
512
op->flags = SS_HASH_FINAL;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
518
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
520
op->flags = SS_HASH_UPDATE;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
527
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
529
op->flags = SS_HASH_UPDATE | SS_HASH_FINAL;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
537
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
54
memset(op, 0, sizeof(struct sun4i_req_ctx));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
543
op->flags = SS_HASH_UPDATE | SS_HASH_FINAL;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
57
op->mode = algt->mode;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
64
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
68
octx->byte_count = op->byte_count + op->len;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
70
memcpy(octx->block, op->buf, op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
72
if (op->byte_count) {
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
74
octx->hash[i] = op->hash[i];
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
87
struct sun4i_req_ctx *op = ahash_request_ctx(areq);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
93
op->byte_count = ictx->byte_count & ~0x3F;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
94
op->len = ictx->byte_count & 0x3F;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
96
memcpy(op->buf, ictx->block, op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
99
op->hash[i] = ictx->hash[i];
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
108
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
122
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
138
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
139
struct sun8i_ce_dev *ce = op->ce;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
159
op->keylen);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
177
len = op->keylen;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
193
rctx->addr_key = dma_map_single(ce->dev, op->key, op->keylen, DMA_TO_DEVICE);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
304
dma_unmap_single(ce->dev, rctx->addr_key, op->keylen, DMA_TO_DEVICE);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
314
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
315
struct sun8i_ce_dev *ce = op->ce;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
346
dma_unmap_single(ce->dev, rctx->addr_key, op->keylen, DMA_TO_DEVICE);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
380
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
389
e = sun8i_ce_get_engine_number(op->ce);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
391
engine = op->ce->chanlist[e].engine;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
399
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
408
e = sun8i_ce_get_engine_number(op->ce);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
410
engine = op->ce->chanlist[e].engine;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
417
struct sun8i_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
424
memset(op, 0, sizeof(struct sun8i_cipher_tfm_ctx));
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
427
op->ce = algt->ce;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
429
op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
430
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
431
dev_err(op->ce->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
432
name, PTR_ERR(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
433
return PTR_ERR(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
437
crypto_skcipher_reqsize(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
441
crypto_skcipher_driver_name(op->fallback_tfm),
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
444
err = pm_runtime_resume_and_get(op->ce->dev);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
450
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
456
struct sun8i_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
458
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
459
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
460
pm_runtime_put_sync_suspend(op->ce->dev);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
466
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
467
struct sun8i_ce_dev *ce = op->ce;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
480
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
481
op->keylen = keylen;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
482
op->key = kmemdup(key, keylen, GFP_KERNEL | GFP_DMA);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
483
if (!op->key)
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
486
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
487
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
489
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
495
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
502
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
503
op->keylen = keylen;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
504
op->key = kmemdup(key, keylen, GFP_KERNEL | GFP_DMA);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
505
if (!op->key)
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
508
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
509
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
511
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
330
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
359
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
389
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
418
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
451
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
484
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
516
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
548
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
580
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-core.c
612
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
40
struct sun8i_ce_hash_tfm_ctx *op = crypto_ahash_ctx(tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
46
op->ce = algt->ce;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
49
op->fallback_tfm = crypto_alloc_ahash(crypto_ahash_alg_name(tfm), 0,
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
51
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
53
return PTR_ERR(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
57
crypto_ahash_statesize(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
61
crypto_ahash_reqsize(op->fallback_tfm) +
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
66
crypto_ahash_driver_name(op->fallback_tfm),
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
69
err = pm_runtime_resume_and_get(op->ce->dev);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
74
crypto_free_ahash(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
108
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
123
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
124
struct sun8i_ss_dev *ss = op->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
186
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
187
struct sun8i_ss_dev *ss = op->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
207
op->keylen);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
215
rctx->keylen = op->keylen;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
217
rctx->p_key = dma_map_single(ss->dev, op->key, op->keylen, DMA_TO_DEVICE);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
325
dma_unmap_single(ss->dev, rctx->p_key, op->keylen, DMA_TO_DEVICE);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
348
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
359
e = sun8i_ss_get_engine_number(op->ss);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
360
engine = op->ss->flows[e].engine;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
369
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
380
e = sun8i_ss_get_engine_number(op->ss);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
381
engine = op->ss->flows[e].engine;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
389
struct sun8i_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
396
memset(op, 0, sizeof(struct sun8i_cipher_tfm_ctx));
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
399
op->ss = algt->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
401
op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
402
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
403
dev_err(op->ss->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
404
name, PTR_ERR(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
405
return PTR_ERR(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
409
crypto_skcipher_reqsize(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
412
crypto_tfm_alg_driver_name(crypto_skcipher_tfm(op->fallback_tfm)),
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
415
err = pm_runtime_resume_and_get(op->ss->dev);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
417
dev_err(op->ss->dev, "pm error %d\n", err);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
423
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
429
struct sun8i_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
431
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
432
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
433
pm_runtime_put_sync(op->ss->dev);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
439
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
440
struct sun8i_ss_dev *ss = op->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
453
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
454
op->keylen = keylen;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
455
op->key = kmemdup(key, keylen, GFP_KERNEL);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
456
if (!op->key)
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
459
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
460
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
462
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
468
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
469
struct sun8i_ss_dev *ss = op->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
476
kfree_sensitive(op->key);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
477
op->keylen = keylen;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
478
op->key = kmemdup(key, keylen, GFP_KERNEL);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
479
if (!op->key)
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
482
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
483
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
485
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
92
struct sun8i_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
194
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
223
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
253
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
282
.alg.skcipher.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
334
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
366
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
398
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
430
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-core.c
463
.alg.hash.op = {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
103
op->ss = algt->ss;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
106
op->fallback_tfm = crypto_alloc_ahash(crypto_ahash_alg_name(tfm), 0,
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
108
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
110
return PTR_ERR(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
114
crypto_ahash_statesize(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
118
crypto_ahash_reqsize(op->fallback_tfm));
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
120
memcpy(algt->fbname, crypto_ahash_driver_name(op->fallback_tfm),
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
123
err = pm_runtime_get_sync(op->ss->dev);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
128
pm_runtime_put_noidle(op->ss->dev);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
129
crypto_free_ahash(op->fallback_tfm);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
97
struct sun8i_ss_hash_tfm_ctx *op = crypto_ahash_ctx(tfm);
drivers/crypto/amcc/crypto4xx_alg.c
30
u32 op_grp, u32 op, u32 dir)
drivers/crypto/amcc/crypto4xx_alg.c
43
sa->sa_command_0.bf.opcode = op;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
110
op->keylen, flow);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
126
memcpy(bkeyiv, op->key, op->keylen);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
127
keyivlen = op->keylen;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
213
v = (op->keymode << 20) | DESC_OWN | todo | (algt->blockmode << 26);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
277
struct meson_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
285
e = get_engine_number(op->mc);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
286
engine = op->mc->chanlist[e].engine;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
295
struct meson_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
303
e = get_engine_number(op->mc);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
304
engine = op->mc->chanlist[e].engine;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
312
struct meson_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
318
memset(op, 0, sizeof(struct meson_cipher_tfm_ctx));
drivers/crypto/amlogic/amlogic-gxl-cipher.c
321
op->mc = algt->mc;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
323
op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
324
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/amlogic/amlogic-gxl-cipher.c
325
dev_err(op->mc->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
drivers/crypto/amlogic/amlogic-gxl-cipher.c
326
name, PTR_ERR(op->fallback_tfm));
drivers/crypto/amlogic/amlogic-gxl-cipher.c
327
return PTR_ERR(op->fallback_tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
331
crypto_skcipher_reqsize(op->fallback_tfm));
drivers/crypto/amlogic/amlogic-gxl-cipher.c
338
struct meson_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
340
kfree_sensitive(op->key);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
341
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
347
struct meson_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
348
struct meson_dev *mc = op->mc;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
352
op->keymode = MODE_AES_128;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
355
op->keymode = MODE_AES_192;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
358
op->keymode = MODE_AES_256;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
364
kfree_sensitive(op->key);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
365
op->keylen = keylen;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
366
op->key = kmemdup(key, keylen, GFP_KERNEL | GFP_DMA);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
367
if (!op->key)
drivers/crypto/amlogic/amlogic-gxl-cipher.c
370
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
61
struct meson_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
71
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
87
struct meson_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
89
struct meson_dev *mc = op->mc;
drivers/crypto/amlogic/amlogic-gxl-core.c
102
.alg.skcipher.op = {
drivers/crypto/amlogic/amlogic-gxl-core.c
74
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-acry.c
621
.akcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
694
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
721
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
747
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
774
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
800
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
827
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
856
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
882
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-crypto.c
908
.alg.skcipher.op = {
drivers/crypto/aspeed/aspeed-hace-hash.c
649
.alg.ahash.op = {
drivers/crypto/aspeed/aspeed-hace-hash.c
681
.alg.ahash.op = {
drivers/crypto/aspeed/aspeed-hace-hash.c
713
.alg.ahash.op = {
drivers/crypto/aspeed/aspeed-hace-hash.c
748
.alg.ahash.op = {
drivers/crypto/aspeed/aspeed-hace-hash.c
780
.alg.ahash.op = {
drivers/crypto/atmel-sha.c
1106
ctx->op, req->nbytes);
drivers/crypto/atmel-sha.c
1132
if (ctx->op == SHA_OP_UPDATE) {
drivers/crypto/atmel-sha.c
1137
} else if (ctx->op == SHA_OP_FINAL) {
drivers/crypto/atmel-sha.c
1150
static int atmel_sha_enqueue(struct ahash_request *req, unsigned int op)
drivers/crypto/atmel-sha.c
1156
ctx->op = op;
drivers/crypto/atmel-sha.c
1824
switch (ctx->op) {
drivers/crypto/atmel-sha.c
92
unsigned long op;
drivers/crypto/caam/blob_gen.c
101
op |= OP_PCL_BLOB_EKT;
drivers/crypto/caam/blob_gen.c
180
append_operation(desc, op);
drivers/crypto/caam/blob_gen.c
78
int op = OP_PCLID_BLOB;
drivers/crypto/caam/blob_gen.c
90
op |= OP_TYPE_ENCAP_PROTOCOL;
drivers/crypto/caam/blob_gen.c
93
op |= OP_TYPE_DECAP_PROTOCOL;
drivers/crypto/caam/blob_gen.c
99
op |= OP_PCL_BLOB_BLACK;
drivers/crypto/caam/caamalg.c
1989
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2008
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2027
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2046
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2066
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2089
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2113
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2131
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2149
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2167
.skcipher.op = {
drivers/crypto/caam/caamalg.c
2189
.aead.op = {
drivers/crypto/caam/caamalg.c
2211
.aead.op = {
drivers/crypto/caam/caamalg.c
2234
.aead.op = {
drivers/crypto/caam/caamalg.c
2259
.aead.op = {
drivers/crypto/caam/caamalg.c
2283
.aead.op = {
drivers/crypto/caam/caamalg.c
2307
.aead.op = {
drivers/crypto/caam/caamalg.c
2331
.aead.op = {
drivers/crypto/caam/caamalg.c
2355
.aead.op = {
drivers/crypto/caam/caamalg.c
2379
.aead.op = {
drivers/crypto/caam/caamalg.c
2402
.aead.op = {
drivers/crypto/caam/caamalg.c
2427
.aead.op = {
drivers/crypto/caam/caamalg.c
2452
.aead.op = {
drivers/crypto/caam/caamalg.c
2477
.aead.op = {
drivers/crypto/caam/caamalg.c
2502
.aead.op = {
drivers/crypto/caam/caamalg.c
2527
.aead.op = {
drivers/crypto/caam/caamalg.c
2552
.aead.op = {
drivers/crypto/caam/caamalg.c
2577
.aead.op = {
drivers/crypto/caam/caamalg.c
2602
.aead.op = {
drivers/crypto/caam/caamalg.c
2627
.aead.op = {
drivers/crypto/caam/caamalg.c
2652
.aead.op = {
drivers/crypto/caam/caamalg.c
2677
.aead.op = {
drivers/crypto/caam/caamalg.c
2702
.aead.op = {
drivers/crypto/caam/caamalg.c
2727
.aead.op = {
drivers/crypto/caam/caamalg.c
2753
.aead.op = {
drivers/crypto/caam/caamalg.c
2779
.aead.op = {
drivers/crypto/caam/caamalg.c
2805
.aead.op = {
drivers/crypto/caam/caamalg.c
2831
.aead.op = {
drivers/crypto/caam/caamalg.c
2857
.aead.op = {
drivers/crypto/caam/caamalg.c
2883
.aead.op = {
drivers/crypto/caam/caamalg.c
2909
.aead.op = {
drivers/crypto/caam/caamalg.c
2935
.aead.op = {
drivers/crypto/caam/caamalg.c
2961
.aead.op = {
drivers/crypto/caam/caamalg.c
2987
.aead.op = {
drivers/crypto/caam/caamalg.c
3012
.aead.op = {
drivers/crypto/caam/caamalg.c
3037
.aead.op = {
drivers/crypto/caam/caamalg.c
3062
.aead.op = {
drivers/crypto/caam/caamalg.c
3087
.aead.op = {
drivers/crypto/caam/caamalg.c
3112
.aead.op = {
drivers/crypto/caam/caamalg.c
3137
.aead.op = {
drivers/crypto/caam/caamalg.c
3162
.aead.op = {
drivers/crypto/caam/caamalg.c
3187
.aead.op = {
drivers/crypto/caam/caamalg.c
3212
.aead.op = {
drivers/crypto/caam/caamalg.c
3237
.aead.op = {
drivers/crypto/caam/caamalg.c
3262
.aead.op = {
drivers/crypto/caam/caamalg.c
3287
.aead.op = {
drivers/crypto/caam/caamalg.c
3313
.aead.op = {
drivers/crypto/caam/caamalg.c
3340
.aead.op = {
drivers/crypto/caam/caamalg.c
3368
.aead.op = {
drivers/crypto/caam/caamalg.c
3395
.aead.op = {
drivers/crypto/caam/caamalg.c
3423
.aead.op = {
drivers/crypto/caam/caamalg.c
3450
.aead.op = {
drivers/crypto/caam/caamalg.c
3478
.aead.op = {
drivers/crypto/caam/caamalg.c
3505
.aead.op = {
drivers/crypto/caam/caamalg.c
3533
.aead.op = {
drivers/crypto/caam/caamalg.c
3560
.aead.op = {
drivers/crypto/caam/caamalg.c
3588
.aead.op = {
drivers/crypto/caam/caamalg.c
3615
.aead.op = {
drivers/crypto/caam/caamalg.c
3642
.aead.op = {
drivers/crypto/caam/caamalg.c
3668
.aead.op = {
drivers/crypto/caam/caamhash.c
1940
t_alg->ahash_alg.op.do_one_request = ahash_do_one_req;
drivers/crypto/caam/caamhash_desc.c
31
u32 op = adata->algtype;
drivers/crypto/caam/caamhash_desc.c
53
op |= OP_ALG_AAI_HMAC_PRECOMP;
drivers/crypto/caam/caamhash_desc.c
62
append_operation(desc, op | state | OP_ALG_ENCRYPT);
drivers/crypto/caam/caampkc.c
1166
.akcipher.op = {
drivers/crypto/caam/desc_constr.h
233
#define APPEND_CMD_RET(cmd, op) \
drivers/crypto/caam/desc_constr.h
238
append_cmd(desc, CMD_##op | options); \
drivers/crypto/caam/desc_constr.h
260
#define APPEND_CMD(cmd, op) \
drivers/crypto/caam/desc_constr.h
264
append_cmd(desc, CMD_##op | options); \
drivers/crypto/caam/desc_constr.h
268
#define APPEND_CMD_LEN(cmd, op) \
drivers/crypto/caam/desc_constr.h
273
append_cmd(desc, CMD_##op | len | options); \
drivers/crypto/caam/desc_constr.h
281
#define APPEND_CMD_PTR(cmd, op) \
drivers/crypto/caam/desc_constr.h
286
append_cmd_ptr(desc, ptr, len, CMD_##op | options); \
drivers/crypto/caam/desc_constr.h
310
#define APPEND_SEQ_PTR_INTLEN(cmd, op) \
drivers/crypto/caam/desc_constr.h
318
append_cmd(desc, CMD_SEQ_##op##_PTR | len | options); \
drivers/crypto/caam/desc_constr.h
320
append_cmd_ptr(desc, ptr, len, CMD_SEQ_##op##_PTR | options); \
drivers/crypto/caam/desc_constr.h
325
#define APPEND_CMD_PTR_TO_IMM(cmd, op) \
drivers/crypto/caam/desc_constr.h
330
append_cmd_data(desc, data, len, CMD_##op | options); \
drivers/crypto/caam/desc_constr.h
335
#define APPEND_CMD_PTR_EXTLEN(cmd, op) \
drivers/crypto/caam/desc_constr.h
340
append_cmd_ptr_extlen(desc, ptr, len, CMD_##op | SQIN_EXT | options); \
drivers/crypto/caam/desc_constr.h
349
#define APPEND_CMD_PTR_LEN(cmd, op, type) \
drivers/crypto/caam/desc_constr.h
366
#define APPEND_CMD_PTR_TO_IMM2(cmd, op) \
drivers/crypto/caam/desc_constr.h
372
append_cmd(desc, CMD_##op | IMMEDIATE | len | options); \
drivers/crypto/caam/desc_constr.h
377
#define APPEND_CMD_RAW_IMM(cmd, op, type) \
drivers/crypto/caam/desc_constr.h
383
append_cmd(desc, CMD_##op | IMMEDIATE | options); \
drivers/crypto/caam/desc_constr.h
385
append_cmd(desc, CMD_##op | IMMEDIATE | options | \
drivers/crypto/caam/desc_constr.h
395
#define APPEND_CMD_RAW_IMM2(cmd, op, ee, size) \
drivers/crypto/caam/desc_constr.h
402
append_cmd(desc, CMD_##op | IMMEDIATE | options | sizeof(data)); \
drivers/crypto/caam/desc_constr.h
412
#define APPEND_MATH(op, desc, dest, src_0, src_1, len) \
drivers/crypto/caam/desc_constr.h
413
append_cmd(desc, CMD_MATH | MATH_FUN_##op | MATH_DEST_##dest | \
drivers/crypto/caam/desc_constr.h
438
#define APPEND_MATH_IMM_u32(op, desc, dest, src_0, src_1, data) \
drivers/crypto/caam/desc_constr.h
440
APPEND_MATH(op, desc, dest, src_0, src_1, CAAM_CMD_SZ); \
drivers/crypto/caam/desc_constr.h
464
#define APPEND_MATH_IMM_u64(op, desc, dest, src_0, src_1, data) \
drivers/crypto/caam/desc_constr.h
467
APPEND_MATH(op, desc, dest, src_0, src_1, CAAM_CMD_SZ * 2 | \
drivers/crypto/ccp/ccp-dev-v3.c
122
| op->jobid;
drivers/crypto/ccp/ccp-dev-v3.c
131
} else if (op->soc) {
drivers/crypto/ccp/ccp-dev-v3.c
135
| op->jobid;
drivers/crypto/ccp/ccp-dev-v3.c
148
static int ccp_perform_aes(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
154
| (op->u.aes.type << REQ1_AES_TYPE_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
155
| (op->u.aes.mode << REQ1_AES_MODE_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
156
| (op->u.aes.action << REQ1_AES_ACTION_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
157
| (op->sb_key << REQ1_KEY_KSB_SHIFT);
drivers/crypto/ccp/ccp-dev-v3.c
158
cr[1] = op->src.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
159
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
160
cr[3] = (op->sb_ctx << REQ4_KSB_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
162
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
163
cr[4] = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
165
| ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
167
if (op->u.aes.mode == CCP_AES_MODE_CFB)
drivers/crypto/ccp/ccp-dev-v3.c
170
if (op->eom)
drivers/crypto/ccp/ccp-dev-v3.c
173
if (op->init)
drivers/crypto/ccp/ccp-dev-v3.c
176
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
179
static int ccp_perform_xts_aes(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
185
| (op->u.xts.action << REQ1_AES_ACTION_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
186
| (op->u.xts.unit_size << REQ1_XTS_AES_SIZE_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
187
| (op->sb_key << REQ1_KEY_KSB_SHIFT);
drivers/crypto/ccp/ccp-dev-v3.c
188
cr[1] = op->src.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
189
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
190
cr[3] = (op->sb_ctx << REQ4_KSB_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
192
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
193
cr[4] = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
195
| ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
197
if (op->eom)
drivers/crypto/ccp/ccp-dev-v3.c
200
if (op->init)
drivers/crypto/ccp/ccp-dev-v3.c
203
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
206
static int ccp_perform_sha(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
212
| (op->u.sha.type << REQ1_SHA_TYPE_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
214
cr[1] = op->src.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
215
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
216
cr[3] = (op->sb_ctx << REQ4_KSB_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
218
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
220
if (op->eom) {
drivers/crypto/ccp/ccp-dev-v3.c
222
cr[4] = lower_32_bits(op->u.sha.msg_bits);
drivers/crypto/ccp/ccp-dev-v3.c
223
cr[5] = upper_32_bits(op->u.sha.msg_bits);
drivers/crypto/ccp/ccp-dev-v3.c
229
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
232
static int ccp_perform_rsa(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
238
| (op->u.rsa.mod_size << REQ1_RSA_MOD_SIZE_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
239
| (op->sb_key << REQ1_KEY_KSB_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
241
cr[1] = op->u.rsa.input_len - 1;
drivers/crypto/ccp/ccp-dev-v3.c
242
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
243
cr[3] = (op->sb_ctx << REQ4_KSB_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
245
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
246
cr[4] = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
248
| ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
250
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
253
static int ccp_perform_passthru(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
259
| (op->u.passthru.bit_mod << REQ1_PT_BW_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
260
| (op->u.passthru.byte_swap << REQ1_PT_BS_SHIFT);
drivers/crypto/ccp/ccp-dev-v3.c
262
if (op->src.type == CCP_MEMTYPE_SYSTEM)
drivers/crypto/ccp/ccp-dev-v3.c
263
cr[1] = op->src.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
265
cr[1] = op->dst.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
267
if (op->src.type == CCP_MEMTYPE_SYSTEM) {
drivers/crypto/ccp/ccp-dev-v3.c
268
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
270
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
272
if (op->u.passthru.bit_mod != CCP_PASSTHRU_BITWISE_NOOP)
drivers/crypto/ccp/ccp-dev-v3.c
273
cr[3] |= (op->sb_key << REQ4_KSB_SHIFT);
drivers/crypto/ccp/ccp-dev-v3.c
275
cr[2] = op->src.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v3.c
279
if (op->dst.type == CCP_MEMTYPE_SYSTEM) {
drivers/crypto/ccp/ccp-dev-v3.c
280
cr[4] = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
282
| ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
284
cr[4] = op->dst.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v3.c
288
if (op->eom)
drivers/crypto/ccp/ccp-dev-v3.c
291
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
294
static int ccp_perform_ecc(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v3.c
301
| (op->u.ecc.function << REQ1_ECC_FUNCTION_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
303
cr[1] = op->src.u.dma.length - 1;
drivers/crypto/ccp/ccp-dev-v3.c
304
cr[2] = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
306
| ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
307
cr[4] = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
309
| ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v3.c
311
return ccp_do_cmd(op, cr, ARRAY_SIZE(cr));
drivers/crypto/ccp/ccp-dev-v3.c
74
static int ccp_do_cmd(struct ccp_op *op, u32 *cr, unsigned int cr_count)
drivers/crypto/ccp/ccp-dev-v3.c
76
struct ccp_cmd_queue *cmd_q = op->cmd_q;
drivers/crypto/ccp/ccp-dev-v3.c
90
| (op->jobid << REQ0_JOBID_SHIFT)
drivers/crypto/ccp/ccp-dev-v3.c
93
if (op->soc)
drivers/crypto/ccp/ccp-dev-v3.c
97
if (op->ioc || !cmd_q->free_slots)
drivers/crypto/ccp/ccp-dev-v5.c
277
static int ccp5_perform_aes(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
281
u32 key_addr = op->sb_key * LSB_ITEM_SIZE;
drivers/crypto/ccp/ccp-dev-v5.c
283
op->cmd_q->total_aes_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
290
CCP5_CMD_SOC(&desc) = op->soc;
drivers/crypto/ccp/ccp-dev-v5.c
292
CCP5_CMD_INIT(&desc) = op->init;
drivers/crypto/ccp/ccp-dev-v5.c
293
CCP5_CMD_EOM(&desc) = op->eom;
drivers/crypto/ccp/ccp-dev-v5.c
297
CCP_AES_ENCRYPT(&function) = op->u.aes.action;
drivers/crypto/ccp/ccp-dev-v5.c
298
CCP_AES_MODE(&function) = op->u.aes.mode;
drivers/crypto/ccp/ccp-dev-v5.c
299
CCP_AES_TYPE(&function) = op->u.aes.type;
drivers/crypto/ccp/ccp-dev-v5.c
300
CCP_AES_SIZE(&function) = op->u.aes.size;
drivers/crypto/ccp/ccp-dev-v5.c
304
CCP5_CMD_LEN(&desc) = op->src.u.dma.length;
drivers/crypto/ccp/ccp-dev-v5.c
306
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
307
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
310
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
311
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
317
CCP5_CMD_LSB_ID(&desc) = op->sb_ctx;
drivers/crypto/ccp/ccp-dev-v5.c
319
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
322
static int ccp5_perform_xts_aes(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
326
u32 key_addr = op->sb_key * LSB_ITEM_SIZE;
drivers/crypto/ccp/ccp-dev-v5.c
328
op->cmd_q->total_xts_aes_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
335
CCP5_CMD_SOC(&desc) = op->soc;
drivers/crypto/ccp/ccp-dev-v5.c
337
CCP5_CMD_INIT(&desc) = op->init;
drivers/crypto/ccp/ccp-dev-v5.c
338
CCP5_CMD_EOM(&desc) = op->eom;
drivers/crypto/ccp/ccp-dev-v5.c
342
CCP_XTS_TYPE(&function) = op->u.xts.type;
drivers/crypto/ccp/ccp-dev-v5.c
343
CCP_XTS_ENCRYPT(&function) = op->u.xts.action;
drivers/crypto/ccp/ccp-dev-v5.c
344
CCP_XTS_SIZE(&function) = op->u.xts.unit_size;
drivers/crypto/ccp/ccp-dev-v5.c
347
CCP5_CMD_LEN(&desc) = op->src.u.dma.length;
drivers/crypto/ccp/ccp-dev-v5.c
349
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
350
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
353
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
354
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
360
CCP5_CMD_LSB_ID(&desc) = op->sb_ctx;
drivers/crypto/ccp/ccp-dev-v5.c
362
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
365
static int ccp5_perform_sha(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
370
op->cmd_q->total_sha_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
377
CCP5_CMD_SOC(&desc) = op->soc;
drivers/crypto/ccp/ccp-dev-v5.c
380
CCP5_CMD_EOM(&desc) = op->eom;
drivers/crypto/ccp/ccp-dev-v5.c
384
CCP_SHA_TYPE(&function) = op->u.sha.type;
drivers/crypto/ccp/ccp-dev-v5.c
387
CCP5_CMD_LEN(&desc) = op->src.u.dma.length;
drivers/crypto/ccp/ccp-dev-v5.c
389
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
390
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
393
CCP5_CMD_LSB_ID(&desc) = op->sb_ctx;
drivers/crypto/ccp/ccp-dev-v5.c
395
if (op->eom) {
drivers/crypto/ccp/ccp-dev-v5.c
396
CCP5_CMD_SHA_LO(&desc) = lower_32_bits(op->u.sha.msg_bits);
drivers/crypto/ccp/ccp-dev-v5.c
397
CCP5_CMD_SHA_HI(&desc) = upper_32_bits(op->u.sha.msg_bits);
drivers/crypto/ccp/ccp-dev-v5.c
403
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
406
static int ccp5_perform_des3(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
410
u32 key_addr = op->sb_key * LSB_ITEM_SIZE;
drivers/crypto/ccp/ccp-dev-v5.c
412
op->cmd_q->total_3des_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
419
CCP5_CMD_SOC(&desc) = op->soc;
drivers/crypto/ccp/ccp-dev-v5.c
421
CCP5_CMD_INIT(&desc) = op->init;
drivers/crypto/ccp/ccp-dev-v5.c
422
CCP5_CMD_EOM(&desc) = op->eom;
drivers/crypto/ccp/ccp-dev-v5.c
426
CCP_DES3_ENCRYPT(&function) = op->u.des3.action;
drivers/crypto/ccp/ccp-dev-v5.c
427
CCP_DES3_MODE(&function) = op->u.des3.mode;
drivers/crypto/ccp/ccp-dev-v5.c
428
CCP_DES3_TYPE(&function) = op->u.des3.type;
drivers/crypto/ccp/ccp-dev-v5.c
431
CCP5_CMD_LEN(&desc) = op->src.u.dma.length;
drivers/crypto/ccp/ccp-dev-v5.c
433
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
434
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
437
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
438
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
444
CCP5_CMD_LSB_ID(&desc) = op->sb_ctx;
drivers/crypto/ccp/ccp-dev-v5.c
446
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
449
static int ccp5_perform_rsa(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
454
op->cmd_q->total_rsa_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
461
CCP5_CMD_SOC(&desc) = op->soc;
drivers/crypto/ccp/ccp-dev-v5.c
468
CCP_RSA_SIZE(&function) = (op->u.rsa.mod_size + 7) >> 3;
drivers/crypto/ccp/ccp-dev-v5.c
471
CCP5_CMD_LEN(&desc) = op->u.rsa.input_len;
drivers/crypto/ccp/ccp-dev-v5.c
474
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
475
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
479
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
480
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
484
CCP5_CMD_KEY_LO(&desc) = ccp_addr_lo(&op->exp.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
485
CCP5_CMD_KEY_HI(&desc) = ccp_addr_hi(&op->exp.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
488
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
491
static int ccp5_perform_passthru(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
495
struct ccp_dma_info *saddr = &op->src.u.dma;
drivers/crypto/ccp/ccp-dev-v5.c
496
struct ccp_dma_info *daddr = &op->dst.u.dma;
drivers/crypto/ccp/ccp-dev-v5.c
499
op->cmd_q->total_pt_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
508
CCP5_CMD_EOM(&desc) = op->eom;
drivers/crypto/ccp/ccp-dev-v5.c
512
CCP_PT_BYTESWAP(&function) = op->u.passthru.byte_swap;
drivers/crypto/ccp/ccp-dev-v5.c
513
CCP_PT_BITWISE(&function) = op->u.passthru.bit_mod;
drivers/crypto/ccp/ccp-dev-v5.c
517
if (op->src.type == CCP_MEMTYPE_SYSTEM)
drivers/crypto/ccp/ccp-dev-v5.c
522
if (op->src.type == CCP_MEMTYPE_SYSTEM) {
drivers/crypto/ccp/ccp-dev-v5.c
523
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
524
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
527
if (op->u.passthru.bit_mod != CCP_PASSTHRU_BITWISE_NOOP)
drivers/crypto/ccp/ccp-dev-v5.c
528
CCP5_CMD_LSB_ID(&desc) = op->sb_key;
drivers/crypto/ccp/ccp-dev-v5.c
530
u32 key_addr = op->src.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v5.c
537
if (op->dst.type == CCP_MEMTYPE_SYSTEM) {
drivers/crypto/ccp/ccp-dev-v5.c
538
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
539
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
542
u32 key_addr = op->dst.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v5.c
549
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-dev-v5.c
552
static int ccp5_perform_ecc(struct ccp_op *op)
drivers/crypto/ccp/ccp-dev-v5.c
557
op->cmd_q->total_ecc_ops++;
drivers/crypto/ccp/ccp-dev-v5.c
571
function.ecc.mode = op->u.ecc.function;
drivers/crypto/ccp/ccp-dev-v5.c
574
CCP5_CMD_LEN(&desc) = op->src.u.dma.length;
drivers/crypto/ccp/ccp-dev-v5.c
576
CCP5_CMD_SRC_LO(&desc) = ccp_addr_lo(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
577
CCP5_CMD_SRC_HI(&desc) = ccp_addr_hi(&op->src.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
580
CCP5_CMD_DST_LO(&desc) = ccp_addr_lo(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
581
CCP5_CMD_DST_HI(&desc) = ccp_addr_hi(&op->dst.u.dma);
drivers/crypto/ccp/ccp-dev-v5.c
584
return ccp5_do_cmd(&desc, op->cmd_q);
drivers/crypto/ccp/ccp-ops.c
1009
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1019
op.u.aes.size = AES_BLOCK_SIZE * BITS_PER_BYTE - 1;
drivers/crypto/ccp/ccp-ops.c
1022
op.u.aes.size = 0;
drivers/crypto/ccp/ccp-ops.c
1049
ccp_prepare_data(&src, &dst, &op, AES_BLOCK_SIZE, true);
drivers/crypto/ccp/ccp-ops.c
1051
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
1058
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
1061
ret = cmd_q->ccp->vdata->perform->aes(&op);
drivers/crypto/ccp/ccp-ops.c
1067
ccp_process_data(&src, &dst, &op);
drivers/crypto/ccp/ccp-ops.c
1074
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1108
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
1156
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
1157
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
1158
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
1159
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
1160
op.sb_ctx = cmd_q->sb_ctx;
drivers/crypto/ccp/ccp-ops.c
1161
op.init = 1;
drivers/crypto/ccp/ccp-ops.c
1162
op.u.xts.type = aestype;
drivers/crypto/ccp/ccp-ops.c
1163
op.u.xts.action = xts->action;
drivers/crypto/ccp/ccp-ops.c
1164
op.u.xts.unit_size = xts->unit_size;
drivers/crypto/ccp/ccp-ops.c
1208
ret = ccp_copy_to_sb(cmd_q, &key, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
1228
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1259
ccp_prepare_data(&src, &dst, &op, unit_size, true);
drivers/crypto/ccp/ccp-ops.c
1261
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
1263
ret = cmd_q->ccp->vdata->perform->xts_aes(&op);
drivers/crypto/ccp/ccp-ops.c
1269
ccp_process_data(&src, &dst, &op);
drivers/crypto/ccp/ccp-ops.c
1275
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1309
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
1342
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
1345
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
1346
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
1347
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
1349
op.init = (des3->mode == CCP_DES3_MODE_ECB) ? 0 : 1;
drivers/crypto/ccp/ccp-ops.c
1350
op.u.des3.type = des3->type;
drivers/crypto/ccp/ccp-ops.c
1351
op.u.des3.mode = des3->mode;
drivers/crypto/ccp/ccp-ops.c
1352
op.u.des3.action = des3->action;
drivers/crypto/ccp/ccp-ops.c
1387
ret = ccp_copy_to_sb(cmd_q, &key, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
1400
op.sb_ctx = cmd_q->sb_ctx;
drivers/crypto/ccp/ccp-ops.c
1415
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1448
ccp_prepare_data(&src, &dst, &op, DES3_EDE_BLOCK_SIZE, true);
drivers/crypto/ccp/ccp-ops.c
1450
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
1456
op.soc = 0;
drivers/crypto/ccp/ccp-ops.c
1459
ret = cmd_q->ccp->vdata->perform->des3(&op);
drivers/crypto/ccp/ccp-ops.c
1465
ccp_process_data(&src, &dst, &op);
drivers/crypto/ccp/ccp-ops.c
1470
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1504
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
1654
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
1655
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
1656
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
1657
op.sb_ctx = cmd_q->sb_ctx; /* Pre-allocated */
drivers/crypto/ccp/ccp-ops.c
1658
op.u.sha.type = sha->type;
drivers/crypto/ccp/ccp-ops.c
1659
op.u.sha.msg_bits = sha->msg_bits;
drivers/crypto/ccp/ccp-ops.c
1696
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1711
ccp_prepare_data(&src, NULL, &op, block_size, false);
drivers/crypto/ccp/ccp-ops.c
1713
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
1715
ret = cmd_q->ccp->vdata->perform->sha(&op);
drivers/crypto/ccp/ccp-ops.c
1721
ccp_process_data(&src, NULL, &op);
drivers/crypto/ccp/ccp-ops.c
1724
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
1725
ret = cmd_q->ccp->vdata->perform->sha(&op);
drivers/crypto/ccp/ccp-ops.c
1735
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
1849
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
1860
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
1861
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
1862
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
1881
op.sb_key = cmd_q->ccp->vdata->perform->sballoc(cmd_q,
drivers/crypto/ccp/ccp-ops.c
1883
if (!op.sb_key)
drivers/crypto/ccp/ccp-ops.c
1890
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
1909
ret = ccp_copy_to_sb(cmd_q, &exp, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
1917
op.exp.u.dma.address = exp.dma.address;
drivers/crypto/ccp/ccp-ops.c
1918
op.exp.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
1941
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
1942
op.src.u.dma.address = src.dma.address;
drivers/crypto/ccp/ccp-ops.c
1943
op.src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
1944
op.src.u.dma.length = i_len;
drivers/crypto/ccp/ccp-ops.c
1945
op.dst.u.dma.address = dst.dma.address;
drivers/crypto/ccp/ccp-ops.c
1946
op.dst.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
1947
op.dst.u.dma.length = o_len;
drivers/crypto/ccp/ccp-ops.c
1949
op.u.rsa.mod_size = rsa->key_size;
drivers/crypto/ccp/ccp-ops.c
1950
op.u.rsa.input_len = i_len;
drivers/crypto/ccp/ccp-ops.c
1952
ret = cmd_q->ccp->vdata->perform->rsa(&op);
drivers/crypto/ccp/ccp-ops.c
1971
cmd_q->ccp->vdata->perform->sbfree(cmd_q, op.sb_key, sb_count);
drivers/crypto/ccp/ccp-ops.c
1982
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
2002
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
2003
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
2004
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
2008
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
2020
ret = ccp_copy_to_sb(cmd_q, &mask, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
2066
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
2067
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
2070
op.src.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
2071
op.src.u.dma.address = sg_dma_address(src.sg_wa.sg);
drivers/crypto/ccp/ccp-ops.c
2072
op.src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2073
op.src.u.dma.length = sg_dma_len(src.sg_wa.sg);
drivers/crypto/ccp/ccp-ops.c
2075
op.dst.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
2076
op.dst.u.dma.address = sg_dma_address(dst.sg_wa.sg);
drivers/crypto/ccp/ccp-ops.c
2077
op.dst.u.dma.offset = dst.sg_wa.sg_used;
drivers/crypto/ccp/ccp-ops.c
2078
op.dst.u.dma.length = op.src.u.dma.length;
drivers/crypto/ccp/ccp-ops.c
2080
ret = cmd_q->ccp->vdata->perform->passthru(&op);
drivers/crypto/ccp/ccp-ops.c
2114
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
2132
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
2133
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
2134
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
2138
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
2144
ret = ccp_copy_to_sb(cmd_q, &mask, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
2153
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
2154
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
2156
op.src.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
2157
op.src.u.dma.address = pt->src_dma;
drivers/crypto/ccp/ccp-ops.c
2158
op.src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2159
op.src.u.dma.length = pt->src_len;
drivers/crypto/ccp/ccp-ops.c
2161
op.dst.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
2162
op.dst.u.dma.address = pt->dst_dma;
drivers/crypto/ccp/ccp-ops.c
2163
op.dst.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2164
op.dst.u.dma.length = pt->src_len;
drivers/crypto/ccp/ccp-ops.c
2166
ret = cmd_q->ccp->vdata->perform->passthru(&op);
drivers/crypto/ccp/ccp-ops.c
2177
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
2194
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
2195
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
2196
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
2244
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
2245
op.src.u.dma.address = src.dma.address;
drivers/crypto/ccp/ccp-ops.c
2246
op.src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2247
op.src.u.dma.length = src.length;
drivers/crypto/ccp/ccp-ops.c
2248
op.dst.u.dma.address = dst.dma.address;
drivers/crypto/ccp/ccp-ops.c
2249
op.dst.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2250
op.dst.u.dma.length = dst.length;
drivers/crypto/ccp/ccp-ops.c
2252
op.u.ecc.function = cmd->u.ecc.function;
drivers/crypto/ccp/ccp-ops.c
2254
ret = cmd_q->ccp->vdata->perform->ecc(&op);
drivers/crypto/ccp/ccp-ops.c
2284
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
2317
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
2318
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
2319
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
2402
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
2403
op.src.u.dma.address = src.dma.address;
drivers/crypto/ccp/ccp-ops.c
2404
op.src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2405
op.src.u.dma.length = src.length;
drivers/crypto/ccp/ccp-ops.c
2406
op.dst.u.dma.address = dst.dma.address;
drivers/crypto/ccp/ccp-ops.c
2407
op.dst.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
2408
op.dst.u.dma.length = dst.length;
drivers/crypto/ccp/ccp-ops.c
2410
op.u.ecc.function = cmd->u.ecc.function;
drivers/crypto/ccp/ccp-ops.c
2412
ret = cmd_q->ccp->vdata->perform->ecc(&op);
drivers/crypto/ccp/ccp-ops.c
344
struct ccp_op *op, unsigned int block_size,
drivers/crypto/ccp/ccp-ops.c
372
op->soc = 0;
drivers/crypto/ccp/ccp-ops.c
380
op->soc = 1;
drivers/crypto/ccp/ccp-ops.c
381
op->src.u.dma.address = src->dm_wa.dma.address;
drivers/crypto/ccp/ccp-ops.c
382
op->src.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
383
op->src.u.dma.length = (blocksize_op) ? block_size : cp_len;
drivers/crypto/ccp/ccp-ops.c
388
op->src.u.dma.address = sg_dma_address(src->sg_wa.dma_sg);
drivers/crypto/ccp/ccp-ops.c
389
op->src.u.dma.offset = src->sg_wa.sg_used;
drivers/crypto/ccp/ccp-ops.c
390
op->src.u.dma.length = op_len & ~(block_size - 1);
drivers/crypto/ccp/ccp-ops.c
392
ccp_update_sg_workarea(&src->sg_wa, op->src.u.dma.length);
drivers/crypto/ccp/ccp-ops.c
401
op->soc = 1;
drivers/crypto/ccp/ccp-ops.c
402
op->dst.u.dma.address = dst->dm_wa.dma.address;
drivers/crypto/ccp/ccp-ops.c
403
op->dst.u.dma.offset = 0;
drivers/crypto/ccp/ccp-ops.c
404
op->dst.u.dma.length = op->src.u.dma.length;
drivers/crypto/ccp/ccp-ops.c
409
op->dst.u.dma.address = sg_dma_address(dst->sg_wa.dma_sg);
drivers/crypto/ccp/ccp-ops.c
410
op->dst.u.dma.offset = dst->sg_wa.sg_used;
drivers/crypto/ccp/ccp-ops.c
411
op->dst.u.dma.length = op->src.u.dma.length;
drivers/crypto/ccp/ccp-ops.c
417
struct ccp_op *op)
drivers/crypto/ccp/ccp-ops.c
419
op->init = 0;
drivers/crypto/ccp/ccp-ops.c
422
if (op->dst.u.dma.address == dst->dm_wa.dma.address)
drivers/crypto/ccp/ccp-ops.c
426
op->dst.u.dma.length);
drivers/crypto/ccp/ccp-ops.c
434
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
436
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
438
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
439
op.jobid = jobid;
drivers/crypto/ccp/ccp-ops.c
440
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
443
op.soc = 1;
drivers/crypto/ccp/ccp-ops.c
444
op.src.type = CCP_MEMTYPE_SB;
drivers/crypto/ccp/ccp-ops.c
445
op.src.u.sb = sb;
drivers/crypto/ccp/ccp-ops.c
446
op.dst.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
447
op.dst.u.dma.address = wa->dma.address;
drivers/crypto/ccp/ccp-ops.c
448
op.dst.u.dma.length = wa->length;
drivers/crypto/ccp/ccp-ops.c
450
op.src.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
451
op.src.u.dma.address = wa->dma.address;
drivers/crypto/ccp/ccp-ops.c
452
op.src.u.dma.length = wa->length;
drivers/crypto/ccp/ccp-ops.c
453
op.dst.type = CCP_MEMTYPE_SB;
drivers/crypto/ccp/ccp-ops.c
454
op.dst.u.sb = sb;
drivers/crypto/ccp/ccp-ops.c
457
op.u.passthru.byte_swap = byte_swap;
drivers/crypto/ccp/ccp-ops.c
459
return cmd_q->ccp->vdata->perform->passthru(&op);
drivers/crypto/ccp/ccp-ops.c
482
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
512
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
513
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
514
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
515
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
516
op.sb_ctx = cmd_q->sb_ctx;
drivers/crypto/ccp/ccp-ops.c
517
op.init = 1;
drivers/crypto/ccp/ccp-ops.c
518
op.u.aes.type = aes->type;
drivers/crypto/ccp/ccp-ops.c
519
op.u.aes.mode = aes->mode;
drivers/crypto/ccp/ccp-ops.c
520
op.u.aes.action = aes->action;
drivers/crypto/ccp/ccp-ops.c
537
ret = ccp_copy_to_sb(cmd_q, &key, op.jobid, op.sb_key,
drivers/crypto/ccp/ccp-ops.c
558
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
572
ccp_prepare_data(&src, NULL, &op, AES_BLOCK_SIZE, true);
drivers/crypto/ccp/ccp-ops.c
574
op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
577
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid,
drivers/crypto/ccp/ccp-ops.c
578
op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
589
ret = ccp_copy_to_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
597
ret = cmd_q->ccp->vdata->perform->aes(&op);
drivers/crypto/ccp/ccp-ops.c
603
ccp_process_data(&src, NULL, &op);
drivers/crypto/ccp/ccp-ops.c
609
ret = ccp_copy_from_sb(cmd_q, &ctx, op.jobid, op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
644
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
708
memset(&wa->op, 0, sizeof(wa->op));
drivers/crypto/ccp/ccp-ops.c
709
wa->op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
710
wa->op.jobid = jobid;
drivers/crypto/ccp/ccp-ops.c
711
wa->op.sb_key = cmd_q->sb_key; /* Pre-allocated */
drivers/crypto/ccp/ccp-ops.c
712
wa->op.sb_ctx = cmd_q->sb_ctx; /* Pre-allocated */
drivers/crypto/ccp/ccp-ops.c
713
wa->op.init = 1;
drivers/crypto/ccp/ccp-ops.c
714
wa->op.u.aes.type = aes->type;
drivers/crypto/ccp/ccp-ops.c
727
ret = ccp_copy_to_sb(cmd_q, &wa->key, wa->op.jobid, wa->op.sb_key,
drivers/crypto/ccp/ccp-ops.c
749
ret = ccp_copy_to_sb(cmd_q, &wa->ctx, wa->op.jobid, wa->op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
756
wa->op.init = 1;
drivers/crypto/ccp/ccp-ops.c
765
wa->op.u.aes.mode = CCP_AES_MODE_GHASH;
drivers/crypto/ccp/ccp-ops.c
766
wa->op.u.aes.action = CCP_AES_GHASHAAD;
drivers/crypto/ccp/ccp-ops.c
769
ccp_prepare_data(&wa->aad, NULL, &wa->op, AES_BLOCK_SIZE, true);
drivers/crypto/ccp/ccp-ops.c
771
ret = cmd_q->ccp->vdata->perform->aes(&wa->op);
drivers/crypto/ccp/ccp-ops.c
777
ccp_process_data(&wa->aad, NULL, &wa->op);
drivers/crypto/ccp/ccp-ops.c
778
wa->op.init = 0;
drivers/crypto/ccp/ccp-ops.c
782
wa->op.u.aes.mode = CCP_AES_MODE_GCTR;
drivers/crypto/ccp/ccp-ops.c
783
wa->op.u.aes.action = aes->action;
drivers/crypto/ccp/ccp-ops.c
805
wa->op.soc = 0;
drivers/crypto/ccp/ccp-ops.c
806
wa->op.eom = 0;
drivers/crypto/ccp/ccp-ops.c
807
wa->op.init = 1;
drivers/crypto/ccp/ccp-ops.c
809
ccp_prepare_data(&wa->src, &wa->dst, &wa->op, AES_BLOCK_SIZE, true);
drivers/crypto/ccp/ccp-ops.c
814
wa->op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
815
wa->op.u.aes.size = (nbytes * 8) - 1;
drivers/crypto/ccp/ccp-ops.c
819
ret = cmd_q->ccp->vdata->perform->aes(&wa->op);
drivers/crypto/ccp/ccp-ops.c
825
ccp_process_data(&wa->src, &wa->dst, &wa->op);
drivers/crypto/ccp/ccp-ops.c
826
wa->op.init = 0;
drivers/crypto/ccp/ccp-ops.c
831
ret = ccp_copy_from_sb(cmd_q, &wa->ctx, wa->op.jobid, wa->op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
842
ret = ccp_copy_to_sb(cmd_q, &wa->ctx, wa->op.jobid, wa->op.sb_ctx,
drivers/crypto/ccp/ccp-ops.c
860
memset(&wa->op, 0, sizeof(wa->op));
drivers/crypto/ccp/ccp-ops.c
861
wa->op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
862
wa->op.jobid = jobid;
drivers/crypto/ccp/ccp-ops.c
863
wa->op.sb_key = cmd_q->sb_key; /* Pre-allocated */
drivers/crypto/ccp/ccp-ops.c
864
wa->op.sb_ctx = cmd_q->sb_ctx; /* Pre-allocated */
drivers/crypto/ccp/ccp-ops.c
865
wa->op.init = 1;
drivers/crypto/ccp/ccp-ops.c
866
wa->op.u.aes.type = aes->type;
drivers/crypto/ccp/ccp-ops.c
867
wa->op.u.aes.mode = CCP_AES_MODE_GHASH;
drivers/crypto/ccp/ccp-ops.c
868
wa->op.u.aes.action = CCP_AES_GHASHFINAL;
drivers/crypto/ccp/ccp-ops.c
869
wa->op.src.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
870
wa->op.src.u.dma.address = wa->final.dma.address;
drivers/crypto/ccp/ccp-ops.c
871
wa->op.src.u.dma.length = AES_BLOCK_SIZE;
drivers/crypto/ccp/ccp-ops.c
872
wa->op.dst.type = CCP_MEMTYPE_SYSTEM;
drivers/crypto/ccp/ccp-ops.c
873
wa->op.dst.u.dma.address = wa->final.dma.address;
drivers/crypto/ccp/ccp-ops.c
874
wa->op.dst.u.dma.length = AES_BLOCK_SIZE;
drivers/crypto/ccp/ccp-ops.c
875
wa->op.eom = 1;
drivers/crypto/ccp/ccp-ops.c
876
wa->op.u.aes.size = 0;
drivers/crypto/ccp/ccp-ops.c
877
ret = cmd_q->ccp->vdata->perform->aes(&wa->op);
drivers/crypto/ccp/ccp-ops.c
931
struct ccp_op op;
drivers/crypto/ccp/ccp-ops.c
961
memset(&op, 0, sizeof(op));
drivers/crypto/ccp/ccp-ops.c
962
op.cmd_q = cmd_q;
drivers/crypto/ccp/ccp-ops.c
963
op.jobid = CCP_NEW_JOBID(cmd_q->ccp);
drivers/crypto/ccp/ccp-ops.c
964
op.sb_key = cmd_q->sb_key;
drivers/crypto/ccp/ccp-ops.c
965
op.sb_ctx = cmd_q->sb_ctx;
drivers/crypto/ccp/ccp-ops.c
966
op.init = (aes->mode == CCP_AES_MODE_ECB) ? 0 : 1;
drivers/crypto/ccp/ccp-ops.c
967
op.u.aes.type = aes->type;
drivers/crypto/ccp/ccp-ops.c
968
op.u.aes.mode = aes->mode;
drivers/crypto/ccp/ccp-ops.c
969
op.u.aes.action = aes->action;
drivers/crypto/ccp/ccp-ops.c
986
ret = ccp_copy_to_sb(cmd_q, &key, op.jobid, op.sb_key,
drivers/crypto/chelsio/chcr_algo.c
1084
if (reqctx->op)
drivers/crypto/chelsio/chcr_algo.c
1119
if (!reqctx->op)
drivers/crypto/chelsio/chcr_algo.c
1173
reqctx->op);
drivers/crypto/chelsio/chcr_algo.c
1309
reqctx->op = op_type;
drivers/crypto/chelsio/chcr_algo.c
2284
chcr_aead_dma_unmap(&u_ctx->lldi.pdev->dev, req, reqctx->op);
drivers/crypto/chelsio/chcr_algo.c
2298
if (reqctx->op && req->cryptlen < authsize)
drivers/crypto/chelsio/chcr_algo.c
2306
reqctx->op);
drivers/crypto/chelsio/chcr_algo.c
2387
(reqctx->op ? -authsize : authsize), CHCR_DST_SG_SIZE, 0);
drivers/crypto/chelsio/chcr_algo.c
2403
transhdr_len, reqctx->op)) {
drivers/crypto/chelsio/chcr_algo.c
2406
return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
drivers/crypto/chelsio/chcr_algo.c
2416
temp = (reqctx->op == CHCR_ENCRYPT_OP) ? 0 : authsize;
drivers/crypto/chelsio/chcr_algo.c
2440
chcr_req->sec_cpl.seqno_numivs = FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op,
drivers/crypto/chelsio/chcr_algo.c
2441
(reqctx->op == CHCR_ENCRYPT_OP) ? 1 : 0,
drivers/crypto/chelsio/chcr_algo.c
2449
if (reqctx->op == CHCR_ENCRYPT_OP ||
drivers/crypto/chelsio/chcr_algo.c
2635
(reqctx->op ? -authsize : authsize);
drivers/crypto/chelsio/chcr_algo.c
2977
error = aead_ccm_validate_input(reqctx->op, req, aeadctx, sub_type);
drivers/crypto/chelsio/chcr_algo.c
2981
+ (reqctx->op ? -authsize : authsize),
drivers/crypto/chelsio/chcr_algo.c
2999
reqctx->b0_len, transhdr_len, reqctx->op)) {
drivers/crypto/chelsio/chcr_algo.c
3002
return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
drivers/crypto/chelsio/chcr_algo.c
3013
fill_sec_cpl_for_aead(&chcr_req->sec_cpl, dst_size, req, reqctx->op);
drivers/crypto/chelsio/chcr_algo.c
3023
error = ccm_format_packet(req, ivptr, sub_type, reqctx->op, assoclen);
drivers/crypto/chelsio/chcr_algo.c
3077
(reqctx->op ? -authsize : authsize),
drivers/crypto/chelsio/chcr_algo.c
3092
transhdr_len, reqctx->op)) {
drivers/crypto/chelsio/chcr_algo.c
3096
return ERR_PTR(chcr_aead_fallback(req, reqctx->op));
drivers/crypto/chelsio/chcr_algo.c
3107
temp = (reqctx->op == CHCR_ENCRYPT_OP) ? 0 : authsize;
drivers/crypto/chelsio/chcr_algo.c
3120
FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op, (reqctx->op ==
drivers/crypto/chelsio/chcr_algo.c
3625
return chcr_aead_fallback(req, reqctx->op);
drivers/crypto/chelsio/chcr_algo.c
3669
reqctx->op = CHCR_ENCRYPT_OP;
drivers/crypto/chelsio/chcr_algo.c
3706
reqctx->op = CHCR_DECRYPT_OP;
drivers/crypto/chelsio/chcr_algo.c
817
chcr_req->sec_cpl.seqno_numivs = FILL_SEC_CPL_SCMD0_SEQNO(reqctx->op, 0,
drivers/crypto/chelsio/chcr_algo.c
824
if ((reqctx->op == CHCR_DECRYPT_OP) &&
drivers/crypto/chelsio/chcr_algo.c
858
if (reqctx->op && (ablkctx->ciph_mode ==
drivers/crypto/chelsio/chcr_crypto.h
187
unsigned int op;
drivers/crypto/chelsio/chcr_crypto.h
305
unsigned int op;
drivers/crypto/gemini/sl3516-ce-cipher.c
106
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
115
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
130
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
131
struct sl3516_ce_dev *ce = op->ce;
drivers/crypto/gemini/sl3516-ce-cipher.c
149
op->keylen);
drivers/crypto/gemini/sl3516-ce-cipher.c
235
cpu_to_be32_array((__be32 *)ecb->key, (u32 *)op->key, op->keylen / 4);
drivers/crypto/gemini/sl3516-ce-cipher.c
241
ecb->control.aesnk = op->keylen / 4;
drivers/crypto/gemini/sl3516-ce-cipher.c
28
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
281
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
29
struct sl3516_ce_dev *ce = op->ce;
drivers/crypto/gemini/sl3516-ce-cipher.c
291
engine = op->ce->engine;
drivers/crypto/gemini/sl3516-ce-cipher.c
299
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
309
engine = op->ce->engine;
drivers/crypto/gemini/sl3516-ce-cipher.c
316
struct sl3516_ce_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
323
memset(op, 0, sizeof(struct sl3516_ce_cipher_tfm_ctx));
drivers/crypto/gemini/sl3516-ce-cipher.c
326
op->ce = algt->ce;
drivers/crypto/gemini/sl3516-ce-cipher.c
328
op->fallback_tfm = crypto_alloc_skcipher(name, 0, CRYPTO_ALG_NEED_FALLBACK);
drivers/crypto/gemini/sl3516-ce-cipher.c
329
if (IS_ERR(op->fallback_tfm)) {
drivers/crypto/gemini/sl3516-ce-cipher.c
330
dev_err(op->ce->dev, "ERROR: Cannot allocate fallback for %s %ld\n",
drivers/crypto/gemini/sl3516-ce-cipher.c
331
name, PTR_ERR(op->fallback_tfm));
drivers/crypto/gemini/sl3516-ce-cipher.c
332
return PTR_ERR(op->fallback_tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
336
crypto_skcipher_reqsize(op->fallback_tfm));
drivers/crypto/gemini/sl3516-ce-cipher.c
338
dev_info(op->ce->dev, "Fallback for %s is %s\n",
drivers/crypto/gemini/sl3516-ce-cipher.c
340
crypto_tfm_alg_driver_name(crypto_skcipher_tfm(op->fallback_tfm)));
drivers/crypto/gemini/sl3516-ce-cipher.c
342
err = pm_runtime_get_sync(op->ce->dev);
drivers/crypto/gemini/sl3516-ce-cipher.c
348
pm_runtime_put_noidle(op->ce->dev);
drivers/crypto/gemini/sl3516-ce-cipher.c
349
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
355
struct sl3516_ce_cipher_tfm_ctx *op = crypto_tfm_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
357
kfree_sensitive(op->key);
drivers/crypto/gemini/sl3516-ce-cipher.c
358
crypto_free_skcipher(op->fallback_tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
359
pm_runtime_put_sync_suspend(op->ce->dev);
drivers/crypto/gemini/sl3516-ce-cipher.c
365
struct sl3516_ce_cipher_tfm_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/gemini/sl3516-ce-cipher.c
366
struct sl3516_ce_dev *ce = op->ce;
drivers/crypto/gemini/sl3516-ce-cipher.c
379
kfree_sensitive(op->key);
drivers/crypto/gemini/sl3516-ce-cipher.c
380
op->keylen = keylen;
drivers/crypto/gemini/sl3516-ce-cipher.c
381
op->key = kmemdup(key, keylen, GFP_KERNEL | GFP_DMA);
drivers/crypto/gemini/sl3516-ce-cipher.c
382
if (!op->key)
drivers/crypto/gemini/sl3516-ce-cipher.c
385
crypto_skcipher_clear_flags(op->fallback_tfm, CRYPTO_TFM_REQ_MASK);
drivers/crypto/gemini/sl3516-ce-cipher.c
386
crypto_skcipher_set_flags(op->fallback_tfm, tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK);
drivers/crypto/gemini/sl3516-ce-cipher.c
388
return crypto_skcipher_setkey(op->fallback_tfm, key, keylen);
drivers/crypto/gemini/sl3516-ce-core.c
242
.alg.skcipher.op = {
drivers/crypto/hifn_795x.c
1107
switch (rctx->op) {
drivers/crypto/hifn_795x.c
1124
if (rctx->op == ACRYPTO_OP_ENCRYPT || rctx->op == ACRYPTO_OP_DECRYPT) {
drivers/crypto/hifn_795x.c
1570
rctx->mode, rctx->op, rctx->type, err);
drivers/crypto/hifn_795x.c
1982
static int hifn_setup_crypto_req(struct skcipher_request *req, u8 op,
drivers/crypto/hifn_795x.c
2007
rctx->op = op;
drivers/crypto/hifn_795x.c
2051
static int hifn_setup_crypto(struct skcipher_request *req, u8 op,
drivers/crypto/hifn_795x.c
2058
err = hifn_setup_crypto_req(req, op, type, mode);
drivers/crypto/hifn_795x.c
606
u8 op, type, mode, unused;
drivers/crypto/hisilicon/qm.c
577
u64 base, u16 queue, bool op)
drivers/crypto/hisilicon/qm.c
580
((op) ? 0x1 << QM_MB_OP_SHIFT : 0) |
drivers/crypto/hisilicon/qm.c
694
bool op)
drivers/crypto/hisilicon/qm.c
711
qm_mb_pre_init(&mailbox, cmd, dma_addr, queue, op);
drivers/crypto/hisilicon/qm.c
741
int qm_set_and_get_xqc(struct hisi_qm *qm, u8 cmd, void *xqc, u32 qp_id, bool op)
drivers/crypto/hisilicon/qm.c
782
if (!op)
drivers/crypto/hisilicon/qm.c
785
qm_mb_pre_init(&mailbox, cmd, xqc_dma, qp_id, op);
drivers/crypto/hisilicon/qm.c
787
if (!ret && op)
drivers/crypto/hisilicon/qm_common.h
75
int qm_set_and_get_xqc(struct hisi_qm *qm, u8 cmd, void *xqc, u32 qp_id, bool op);
drivers/crypto/img-hash.c
102
unsigned long op;
drivers/crypto/img-hash.c
538
ctx->op, req->nbytes);
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
960
struct ixp_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/intel/ixp4xx/ixp4xx_crypto.c
964
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1279
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1302
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1324
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1347
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1369
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1391
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1412
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1434
.op.do_one_request = kmb_ocs_aes_sk_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1461
.op.do_one_request = kmb_ocs_aes_aead_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1484
.op.do_one_request = kmb_ocs_aes_aead_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1506
.op.do_one_request = kmb_ocs_aes_aead_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-aes-core.c
1528
.op.do_one_request = kmb_ocs_aes_aead_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-ecc.c
838
.op.do_one_request = kmb_ocs_ecc_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-ecc.c
855
.op.do_one_request = kmb_ocs_ecc_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1020
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1047
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1072
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1099
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1124
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
1151
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
915
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
942
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
968
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/keembay/keembay-ocs-hcu-core.c
995
.op.do_one_request = kmb_ocs_hcu_do_one_request,
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
18
static inline int check_stat(u32 (*op)(void __iomem *, u32), u32 expect_val,
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
21
u32 actual_val = op(base, bank);
drivers/crypto/marvell/cesa/cesa.h
315
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/cesa.h
561
struct mv_cesa_op_ctx op;
drivers/crypto/marvell/cesa/cesa.h
661
static inline void mv_cesa_update_op_cfg(struct mv_cesa_op_ctx *op,
drivers/crypto/marvell/cesa/cesa.h
664
op->desc.config &= cpu_to_le32(~mask);
drivers/crypto/marvell/cesa/cesa.h
665
op->desc.config |= cpu_to_le32(cfg);
drivers/crypto/marvell/cesa/cesa.h
668
static inline u32 mv_cesa_get_op_cfg(const struct mv_cesa_op_ctx *op)
drivers/crypto/marvell/cesa/cesa.h
670
return le32_to_cpu(op->desc.config);
drivers/crypto/marvell/cesa/cesa.h
673
static inline void mv_cesa_set_op_cfg(struct mv_cesa_op_ctx *op, u32 cfg)
drivers/crypto/marvell/cesa/cesa.h
675
op->desc.config = cpu_to_le32(cfg);
drivers/crypto/marvell/cesa/cesa.h
679
struct mv_cesa_op_ctx *op)
drivers/crypto/marvell/cesa/cesa.h
683
op->desc.enc_p = CESA_SA_DESC_CRYPT_DATA(offset);
drivers/crypto/marvell/cesa/cesa.h
684
op->desc.enc_key_p = CESA_SA_DESC_CRYPT_KEY(offset);
drivers/crypto/marvell/cesa/cesa.h
685
op->desc.enc_iv = CESA_SA_DESC_CRYPT_IV(offset);
drivers/crypto/marvell/cesa/cesa.h
686
op->desc.mac_src_p &= ~CESA_SA_DESC_MAC_DATA_MSK;
drivers/crypto/marvell/cesa/cesa.h
687
op->desc.mac_src_p |= CESA_SA_DESC_MAC_DATA(offset);
drivers/crypto/marvell/cesa/cesa.h
688
op->desc.mac_digest &= ~CESA_SA_DESC_MAC_DIGEST_MSK;
drivers/crypto/marvell/cesa/cesa.h
689
op->desc.mac_digest |= CESA_SA_DESC_MAC_DIGEST(offset);
drivers/crypto/marvell/cesa/cesa.h
690
op->desc.mac_iv = CESA_SA_DESC_MAC_IV(offset);
drivers/crypto/marvell/cesa/cesa.h
693
static inline void mv_cesa_set_crypt_op_len(struct mv_cesa_op_ctx *op, int len)
drivers/crypto/marvell/cesa/cesa.h
695
op->desc.enc_len = cpu_to_le32(len);
drivers/crypto/marvell/cesa/cesa.h
698
static inline void mv_cesa_set_mac_op_total_len(struct mv_cesa_op_ctx *op,
drivers/crypto/marvell/cesa/cesa.h
701
op->desc.mac_src_p &= ~CESA_SA_DESC_MAC_TOTAL_LEN_MSK;
drivers/crypto/marvell/cesa/cesa.h
702
op->desc.mac_src_p |= CESA_SA_DESC_MAC_TOTAL_LEN(len);
drivers/crypto/marvell/cesa/cesa.h
705
static inline void mv_cesa_set_mac_op_frag_len(struct mv_cesa_op_ctx *op,
drivers/crypto/marvell/cesa/cesa.h
708
op->desc.mac_digest &= ~CESA_SA_DESC_MAC_FRAG_LEN_MSK;
drivers/crypto/marvell/cesa/cesa.h
709
op->desc.mac_digest |= CESA_SA_DESC_MAC_FRAG_LEN(len);
drivers/crypto/marvell/cesa/cesa.h
727
static inline bool mv_cesa_mac_op_is_first_frag(const struct mv_cesa_op_ctx *op)
drivers/crypto/marvell/cesa/cesa.h
729
return (mv_cesa_get_op_cfg(op) & CESA_SA_DESC_CFG_FRAG_MSK) ==
drivers/crypto/marvell/cesa/cipher.c
105
mv_cesa_set_crypt_op_len(&sreq->op, len);
drivers/crypto/marvell/cesa/cipher.c
110
memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op));
drivers/crypto/marvell/cesa/cipher.c
112
memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op));
drivers/crypto/marvell/cesa/cipher.c
115
memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op.desc));
drivers/crypto/marvell/cesa/cipher.c
117
memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op.desc));
drivers/crypto/marvell/cesa/cipher.c
224
memcpy(skreq->iv, basereq->chain.last->op->ctx.skcipher.iv,
drivers/crypto/marvell/cesa/cipher.c
351
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/cipher.c
353
op = mv_cesa_dma_add_op(&basereq->chain, op_templ, skip_ctx,
drivers/crypto/marvell/cesa/cipher.c
355
if (IS_ERR(op)) {
drivers/crypto/marvell/cesa/cipher.c
356
ret = PTR_ERR(op);
drivers/crypto/marvell/cesa/cipher.c
361
mv_cesa_set_crypt_op_len(op, iter.base.op_len);
drivers/crypto/marvell/cesa/cipher.c
416
sreq->op = *op_templ;
drivers/crypto/marvell/cesa/cipher.c
94
mv_cesa_adjust_op(engine, &sreq->op);
drivers/crypto/marvell/cesa/cipher.c
96
memcpy(engine->sram_pool, &sreq->op, sizeof(sreq->op));
drivers/crypto/marvell/cesa/cipher.c
98
memcpy_toio(engine->sram, &sreq->op, sizeof(sreq->op));
drivers/crypto/marvell/cesa/hash.c
166
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/hash.c
211
op = &creq->op_tmpl;
drivers/crypto/marvell/cesa/hash.c
213
frag_mode = mv_cesa_get_op_cfg(op) & CESA_SA_DESC_CFG_FRAG_MSK;
drivers/crypto/marvell/cesa/hash.c
227
mv_cesa_set_mac_op_total_len(op, creq->len);
drivers/crypto/marvell/cesa/hash.c
265
mv_cesa_set_mac_op_frag_len(op, len);
drivers/crypto/marvell/cesa/hash.c
266
mv_cesa_update_op_cfg(op, frag_mode, CESA_SA_DESC_CFG_FRAG_MSK);
drivers/crypto/marvell/cesa/hash.c
270
memcpy(engine->sram_pool, op, sizeof(*op));
drivers/crypto/marvell/cesa/hash.c
272
memcpy_toio(engine->sram, op, sizeof(*op));
drivers/crypto/marvell/cesa/hash.c
275
mv_cesa_update_op_cfg(op, CESA_SA_DESC_CFG_MID_FRAG,
drivers/crypto/marvell/cesa/hash.c
374
data = creq->base.chain.last->op->ctx.hash.hash;
drivers/crypto/marvell/cesa/hash.c
494
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/hash.c
497
op = mv_cesa_dma_add_op(chain, tmpl, false, flags);
drivers/crypto/marvell/cesa/hash.c
498
if (IS_ERR(op))
drivers/crypto/marvell/cesa/hash.c
499
return op;
drivers/crypto/marvell/cesa/hash.c
502
mv_cesa_set_mac_op_frag_len(op, frag_len);
drivers/crypto/marvell/cesa/hash.c
514
return op;
drivers/crypto/marvell/cesa/hash.c
550
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/hash.c
558
op = mv_cesa_dma_add_frag(chain, &creq->op_tmpl, frag_len,
drivers/crypto/marvell/cesa/hash.c
560
if (IS_ERR(op))
drivers/crypto/marvell/cesa/hash.c
561
return op;
drivers/crypto/marvell/cesa/hash.c
563
mv_cesa_set_mac_op_total_len(op, creq->len);
drivers/crypto/marvell/cesa/hash.c
564
mv_cesa_update_op_cfg(op, mv_cesa_mac_op_is_first_frag(op) ?
drivers/crypto/marvell/cesa/hash.c
575
return op;
drivers/crypto/marvell/cesa/hash.c
600
op = mv_cesa_dma_add_frag(chain, &creq->op_tmpl, frag_len + len,
drivers/crypto/marvell/cesa/hash.c
602
if (IS_ERR(op))
drivers/crypto/marvell/cesa/hash.c
603
return op;
drivers/crypto/marvell/cesa/hash.c
606
return op;
drivers/crypto/marvell/cesa/hash.c
632
struct mv_cesa_op_ctx *op = NULL;
drivers/crypto/marvell/cesa/hash.c
682
op = mv_cesa_dma_add_frag(&basereq->chain,
drivers/crypto/marvell/cesa/hash.c
685
if (IS_ERR(op)) {
drivers/crypto/marvell/cesa/hash.c
686
ret = PTR_ERR(op);
drivers/crypto/marvell/cesa/hash.c
701
op = mv_cesa_ahash_dma_last_req(&basereq->chain, &iter, creq,
drivers/crypto/marvell/cesa/hash.c
704
op = mv_cesa_dma_add_frag(&basereq->chain, &creq->op_tmpl,
drivers/crypto/marvell/cesa/hash.c
707
if (IS_ERR(op)) {
drivers/crypto/marvell/cesa/hash.c
708
ret = PTR_ERR(op);
drivers/crypto/marvell/cesa/hash.c
720
if (op && type != CESA_TDMA_RESULT) {
drivers/crypto/marvell/cesa/tdma.c
101
mv_cesa_adjust_op(engine, tdma->op);
drivers/crypto/marvell/cesa/tdma.c
253
tdma->op = op_desc->op;
drivers/crypto/marvell/cesa/tdma.c
266
struct mv_cesa_op_ctx *op;
drivers/crypto/marvell/cesa/tdma.c
274
op = dma_pool_alloc(cesa_dev->dma->op_pool, flags, &dma_handle);
drivers/crypto/marvell/cesa/tdma.c
275
if (!op)
drivers/crypto/marvell/cesa/tdma.c
278
*op = *op_templ;
drivers/crypto/marvell/cesa/tdma.c
280
size = skip_ctx ? sizeof(op->desc) : sizeof(*op);
drivers/crypto/marvell/cesa/tdma.c
283
tdma->op = op;
drivers/crypto/marvell/cesa/tdma.c
289
return op;
drivers/crypto/marvell/cesa/tdma.c
76
dma_pool_free(cesa_dev->dma->op_pool, tdma->op,
drivers/crypto/nx/nx-aes-cbc.c
81
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-aes-cbc.c
86
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-ccm.c
247
nx_ctx->op.inlen = (nx_ctx->in_sg - nx_insg) *
drivers/crypto/nx/nx-aes-ccm.c
249
nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) *
drivers/crypto/nx/nx-aes-ccm.c
257
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-ccm.c
370
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-ccm.c
439
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-ctr.c
94
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-aes-ctr.c
99
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-ecb.c
80
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-aes-ecb.c
85
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-gcm.c
209
nx_ctx->op.inlen = (nx_ctx->in_sg - nx_sg)
drivers/crypto/nx/nx-aes-gcm.c
215
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-gcm.c
279
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-gcm.c
280
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-gcm.c
282
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-gcm.c
368
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op,
drivers/crypto/nx/nx-aes-xcbc.c
120
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
121
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
123
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-aes-xcbc.c
202
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
212
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) *
drivers/crypto/nx/nx-aes-xcbc.c
220
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-aes-xcbc.c
225
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-aes-xcbc.c
292
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
293
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
295
if (!nx_ctx->op.outlen) {
drivers/crypto/nx/nx-aes-xcbc.c
300
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-aes-xcbc.c
92
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
93
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-aes-xcbc.c
95
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-common-pseries.c
316
struct vio_pfo_op op = {
drivers/crypto/nx/nx-common-pseries.c
345
op.flags = NX842_OP_COMPRESS_CRC;
drivers/crypto/nx/nx-common-pseries.c
348
op.csbcpb = nx842_get_pa(csbcpb);
drivers/crypto/nx/nx-common-pseries.c
353
op.in = nx842_get_pa((void *)inbuf);
drivers/crypto/nx/nx-common-pseries.c
354
op.inlen = inlen;
drivers/crypto/nx/nx-common-pseries.c
358
op.in = nx842_get_pa(slin.entries);
drivers/crypto/nx/nx-common-pseries.c
359
op.inlen = -nx842_get_scatterlist_size(&slin);
drivers/crypto/nx/nx-common-pseries.c
365
op.out = nx842_get_pa((void *)outbuf);
drivers/crypto/nx/nx-common-pseries.c
366
op.outlen = *outlen;
drivers/crypto/nx/nx-common-pseries.c
370
op.out = nx842_get_pa(slout.entries);
drivers/crypto/nx/nx-common-pseries.c
371
op.outlen = -nx842_get_scatterlist_size(&slout);
drivers/crypto/nx/nx-common-pseries.c
375
__func__, (unsigned long)op.in, (long)op.inlen,
drivers/crypto/nx/nx-common-pseries.c
376
(unsigned long)op.out, (long)op.outlen);
drivers/crypto/nx/nx-common-pseries.c
379
ret = vio_h_cop_sync(local_devdata->vdev, &op);
drivers/crypto/nx/nx-common-pseries.c
384
__func__, ret, op.hcall_err);
drivers/crypto/nx/nx-common-pseries.c
446
struct vio_pfo_op op = {
drivers/crypto/nx/nx-common-pseries.c
477
op.flags = NX842_OP_DECOMPRESS_CRC;
drivers/crypto/nx/nx-common-pseries.c
480
op.csbcpb = nx842_get_pa(csbcpb);
drivers/crypto/nx/nx-common-pseries.c
485
op.in = nx842_get_pa((void *)inbuf);
drivers/crypto/nx/nx-common-pseries.c
486
op.inlen = inlen;
drivers/crypto/nx/nx-common-pseries.c
490
op.in = nx842_get_pa(slin.entries);
drivers/crypto/nx/nx-common-pseries.c
491
op.inlen = -nx842_get_scatterlist_size(&slin);
drivers/crypto/nx/nx-common-pseries.c
497
op.out = nx842_get_pa((void *)outbuf);
drivers/crypto/nx/nx-common-pseries.c
498
op.outlen = *outlen;
drivers/crypto/nx/nx-common-pseries.c
502
op.out = nx842_get_pa(slout.entries);
drivers/crypto/nx/nx-common-pseries.c
503
op.outlen = -nx842_get_scatterlist_size(&slout);
drivers/crypto/nx/nx-common-pseries.c
507
__func__, (unsigned long)op.in, (long)op.inlen,
drivers/crypto/nx/nx-common-pseries.c
508
(unsigned long)op.out, (long)op.outlen);
drivers/crypto/nx/nx-common-pseries.c
511
ret = vio_h_cop_sync(local_devdata->vdev, &op);
drivers/crypto/nx/nx-common-pseries.c
516
__func__, ret, op.hcall_err);
drivers/crypto/nx/nx-sha256.c
105
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha256.c
118
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-sha256.c
123
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-sha256.c
186
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha256.c
187
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha256.c
188
if (!nx_ctx->op.outlen) {
drivers/crypto/nx/nx-sha256.c
193
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-sha256.c
89
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha512.c
106
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha512.c
119
if (!nx_ctx->op.inlen || !nx_ctx->op.outlen) {
drivers/crypto/nx/nx-sha512.c
124
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-sha512.c
190
nx_ctx->op.inlen = (nx_ctx->in_sg - in_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha512.c
191
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx-sha512.c
193
if (!nx_ctx->op.outlen) {
drivers/crypto/nx/nx-sha512.c
198
rc = nx_hcall_sync(nx_ctx, &nx_ctx->op, 0);
drivers/crypto/nx/nx-sha512.c
90
nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg);
drivers/crypto/nx/nx.c
270
nx_ctx->op.inlen = trim_sg_list(nx_ctx->in_sg, nx_insg, delta, nbytes);
drivers/crypto/nx/nx.c
271
nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes);
drivers/crypto/nx/nx.c
288
nx_ctx->op.flags = function;
drivers/crypto/nx/nx.c
289
nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb);
drivers/crypto/nx/nx.c
290
nx_ctx->op.in = __pa(nx_ctx->in_sg);
drivers/crypto/nx/nx.c
291
nx_ctx->op.out = __pa(nx_ctx->out_sg);
drivers/crypto/nx/nx.c
41
struct vio_pfo_op *op,
drivers/crypto/nx/nx.c
50
rc = vio_h_cop_sync(viodev, op);
drivers/crypto/nx/nx.c
55
"hcall rc: %ld\n", rc, op->hcall_err);
drivers/crypto/nx/nx.c
57
atomic_set(&(nx_ctx->stats->last_error), op->hcall_err);
drivers/crypto/nx/nx.h
126
struct vio_pfo_op op; /* operation struct with hcall parameters */
drivers/crypto/nx/nx.h
161
int nx_hcall_sync(struct nx_crypto_ctx *ctx, struct vio_pfo_op *op,
drivers/crypto/omap-aes.c
677
.op.do_one_request = omap_aes_crypt_req,
drivers/crypto/omap-aes.c
700
.op.do_one_request = omap_aes_crypt_req,
drivers/crypto/omap-aes.c
726
.op.do_one_request = omap_aes_crypt_req,
drivers/crypto/omap-aes.c
759
.op.do_one_request = omap_aes_gcm_crypt_req,
drivers/crypto/omap-aes.c
782
.op.do_one_request = omap_aes_gcm_crypt_req,
drivers/crypto/omap-des.c
730
.op.do_one_request = omap_des_crypt_req,
drivers/crypto/omap-des.c
751
.op.do_one_request = omap_des_crypt_req,
drivers/crypto/omap-des.c
771
.op.do_one_request = omap_des_crypt_req,
drivers/crypto/omap-des.c
792
.op.do_one_request = omap_des_crypt_req,
drivers/crypto/omap-sham.c
1072
ctx->op, ctx->total, ctx->digcnt, final);
drivers/crypto/omap-sham.c
1090
if (ctx->op == OP_UPDATE)
drivers/crypto/omap-sham.c
1092
else if (ctx->op == OP_FINAL)
drivers/crypto/omap-sham.c
1157
omap_sham_enqueue(req, ctx->op);
drivers/crypto/omap-sham.c
1185
static int omap_sham_enqueue(struct ahash_request *req, unsigned int op)
drivers/crypto/omap-sham.c
1190
ctx->op = op;
drivers/crypto/omap-sham.c
1443
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
145
u8 op;
drivers/crypto/omap-sham.c
1465
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1489
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1513
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1539
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1561
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1585
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1609
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1634
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1656
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1680
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
1704
.op.do_one_request = omap_sham_hash_one_req,
drivers/crypto/omap-sham.c
246
static int omap_sham_enqueue(struct ahash_request *req, unsigned int op);
drivers/crypto/omap-sham.c
852
bool update = rctx->op == OP_UPDATE;
drivers/crypto/rockchip/rk3288_crypto_ahash.c
402
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
434
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_ahash.c
466
.alg.hash.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
481
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
507
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
532
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
558
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
583
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
609
.alg.skcipher.op = {
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
71
struct rk_cipher_ctx *op = crypto_skcipher_ctx(tfm);
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
79
skcipher_request_set_tfm(&rctx->fallback_req, op->fallback_tfm);
drivers/crypto/s5p-sss.c
1464
static int s5p_hash_enqueue(struct ahash_request *req, bool op)
drivers/crypto/s5p-sss.c
1469
ctx->op_update = op;
drivers/crypto/sahara.c
1118
.op = {
drivers/crypto/sahara.c
1141
.op = {
drivers/crypto/sahara.c
1172
.op = {
drivers/crypto/sahara.c
1203
.op = {
drivers/crypto/starfive/jh7110-aes.c
1034
.op = {
drivers/crypto/starfive/jh7110-aes.c
1057
.op = {
drivers/crypto/starfive/jh7110-aes.c
1080
.op = {
drivers/crypto/starfive/jh7110-aes.c
1107
.op = {
drivers/crypto/starfive/jh7110-aes.c
1130
.op = {
drivers/crypto/starfive/jh7110-hash.c
583
.op = {
drivers/crypto/starfive/jh7110-hash.c
612
.op = {
drivers/crypto/starfive/jh7110-hash.c
640
.op = {
drivers/crypto/starfive/jh7110-hash.c
669
.op = {
drivers/crypto/starfive/jh7110-hash.c
697
.op = {
drivers/crypto/starfive/jh7110-hash.c
726
.op = {
drivers/crypto/starfive/jh7110-hash.c
754
.op = {
drivers/crypto/starfive/jh7110-hash.c
783
.op = {
drivers/crypto/starfive/jh7110-hash.c
811
.op = {
drivers/crypto/starfive/jh7110-hash.c
840
.op = {
drivers/crypto/stm32/stm32-cryp.c
2301
.op = {
drivers/crypto/stm32/stm32-cryp.c
2324
.op = {
drivers/crypto/stm32/stm32-cryp.c
2347
.op = {
drivers/crypto/stm32/stm32-cryp.c
2369
.op = {
drivers/crypto/stm32/stm32-cryp.c
2392
.op = {
drivers/crypto/stm32/stm32-cryp.c
2414
.op = {
drivers/crypto/stm32/stm32-cryp.c
2437
.op = {
drivers/crypto/stm32/stm32-cryp.c
2463
.op = {
drivers/crypto/stm32/stm32-cryp.c
2486
.op = {
drivers/crypto/stm32/stm32-hash.c
1040
rctx->op, req->nbytes);
drivers/crypto/stm32/stm32-hash.c
1080
if (rctx->op == HASH_OP_UPDATE)
drivers/crypto/stm32/stm32-hash.c
1082
else if (rctx->op == HASH_OP_FINAL)
drivers/crypto/stm32/stm32-hash.c
1261
bool update = rctx->op & HASH_OP_UPDATE;
drivers/crypto/stm32/stm32-hash.c
1377
static int stm32_hash_enqueue(struct ahash_request *req, unsigned int op)
drivers/crypto/stm32/stm32-hash.c
1383
rctx->op = op;
drivers/crypto/stm32/stm32-hash.c
157
unsigned long op;
drivers/crypto/stm32/stm32-hash.c
1613
.op = {
drivers/crypto/stm32/stm32-hash.c
1642
.op = {
drivers/crypto/stm32/stm32-hash.c
1673
.op = {
drivers/crypto/stm32/stm32-hash.c
1702
.op = {
drivers/crypto/stm32/stm32-hash.c
1733
.op = {
drivers/crypto/stm32/stm32-hash.c
1762
.op = {
drivers/crypto/stm32/stm32-hash.c
1793
.op = {
drivers/crypto/stm32/stm32-hash.c
1822
.op = {
drivers/crypto/stm32/stm32-hash.c
1853
.op = {
drivers/crypto/stm32/stm32-hash.c
1882
.op = {
drivers/crypto/stm32/stm32-hash.c
1910
.op = {
drivers/crypto/stm32/stm32-hash.c
1939
.op = {
drivers/crypto/stm32/stm32-hash.c
1970
.op = {
drivers/crypto/stm32/stm32-hash.c
1999
.op = {
drivers/crypto/stm32/stm32-hash.c
2027
.op = {
drivers/crypto/stm32/stm32-hash.c
2056
.op = {
drivers/crypto/stm32/stm32-hash.c
2084
.op = {
drivers/crypto/stm32/stm32-hash.c
2113
.op = {
drivers/crypto/stm32/stm32-hash.c
2141
.op = {
drivers/crypto/stm32/stm32-hash.c
2170
.op = {
drivers/crypto/tegra/tegra-se-aes.c
1475
u32 *cpuvaddr = se->cmdbuf->addr, op;
drivers/crypto/tegra/tegra-se-aes.c
1479
op = SE_AES_OP_WRSTALL | SE_AES_OP_START | SE_AES_OP_LASTBUF;
drivers/crypto/tegra/tegra-se-aes.c
1482
op |= SE_AES_OP_FINAL;
drivers/crypto/tegra/tegra-se-aes.c
1514
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
1515
cpuvaddr[i++] = op;
drivers/crypto/tegra/tegra-se-aes.c
1912
.alg.aead.op.do_one_request = tegra_gcm_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
1934
.alg.aead.op.do_one_request = tegra_ccm_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
1961
.alg.ahash.op.do_one_request = tegra_cmac_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
250
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
518
.alg.skcipher.op.do_one_request = tegra_aes_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
540
.alg.skcipher.op.do_one_request = tegra_aes_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
561
.alg.skcipher.op.do_one_request = tegra_aes_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
583
.alg.skcipher.op.do_one_request = tegra_aes_do_one_req,
drivers/crypto/tegra/tegra-se-aes.c
635
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
652
u32 *cpuvaddr = se->cmdbuf->addr, op;
drivers/crypto/tegra/tegra-se-aes.c
656
op = SE_AES_OP_WRSTALL | SE_AES_OP_FINAL |
drivers/crypto/tegra/tegra-se-aes.c
664
op |= SE_AES_OP_INIT;
drivers/crypto/tegra/tegra-se-aes.c
696
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
697
cpuvaddr[i++] = op;
drivers/crypto/tegra/tegra-se-aes.c
711
u32 op;
drivers/crypto/tegra/tegra-se-aes.c
713
op = SE_AES_OP_WRSTALL | SE_AES_OP_FINAL |
drivers/crypto/tegra/tegra-se-aes.c
720
op |= SE_AES_OP_INIT;
drivers/crypto/tegra/tegra-se-aes.c
746
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
747
cpuvaddr[i++] = op;
drivers/crypto/tegra/tegra-se-aes.c
878
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-aes.c
917
cpuvaddr[i++] = se_host1x_opcode_nonincr(se->hw->regs->op, 1);
drivers/crypto/tegra/tegra-se-hash.c
1003
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
1031
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
1059
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
749
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
774
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
799
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
824
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
849
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
874
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
899
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
924
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
949
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-hash.c
975
.alg.ahash.op.do_one_request = tegra_sha_do_one_req,
drivers/crypto/tegra/tegra-se-key.c
57
cpuvaddr[i++] = se_host1x_opcode_incr_w(se->hw->regs->op);
drivers/crypto/tegra/tegra-se-key.c
85
cpuvaddr[i++] = se_host1x_opcode_incr_w(se->hw->regs->op);
drivers/crypto/tegra/tegra-se-main.c
340
.op = SE_AES1_OPERATION,
drivers/crypto/tegra/tegra-se-main.c
354
.op = SE_SHA_OPERATION,
drivers/crypto/tegra/tegra-se.h
397
u32 op;
drivers/crypto/ti/dthev2-aes.c
478
.op.do_one_request = dthe_aes_run,
drivers/crypto/ti/dthev2-aes.c
501
.op.do_one_request = dthe_aes_run,
drivers/crypto/ti/dthev2-aes.c
526
.op.do_one_request = dthe_aes_run,
drivers/crypto/virtio/virtio_crypto_akcipher_algs.c
482
.algo.op = {
drivers/crypto/virtio/virtio_crypto_akcipher_algs.c
510
.algo.op = {
drivers/crypto/virtio/virtio_crypto_skcipher_algs.c
117
int op = encrypt ? VIRTIO_CRYPTO_OP_ENCRYPT : VIRTIO_CRYPTO_OP_DECRYPT;
drivers/crypto/virtio/virtio_crypto_skcipher_algs.c
154
sym_create_session->u.cipher.para.op = cpu_to_le32(op);
drivers/crypto/virtio/virtio_crypto_skcipher_algs.c
584
.algo.op = {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
127
u32 op;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
176
hwreq->op = rq_ctx->op;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
178
if (hwreq->op == XILINX_AES_ENCRYPT)
drivers/crypto/xilinx/zynqmp-aes-gcm.c
223
if (hwreq->op == XILINX_AES_ENCRYPT)
drivers/crypto/xilinx/zynqmp-aes-gcm.c
294
if (rq_ctx->op == XILINX_AES_ENCRYPT) {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
295
hwreq->op = VERSAL_AES_ENCRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
299
hwreq->op = VERSAL_AES_DECRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
331
if (rq_ctx->op == XILINX_AES_ENCRYPT) {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
382
if (tfm_ctx->authsize != XILINX_AES_AUTH_SIZE && rq_ctx->op == XILINX_AES_DECRYPT)
drivers/crypto/xilinx/zynqmp-aes-gcm.c
395
if (rq_ctx->op == XILINX_AES_DECRYPT &&
drivers/crypto/xilinx/zynqmp-aes-gcm.c
407
if (tfm_ctx->authsize != XILINX_AES_AUTH_SIZE && rq_ctx->op == XILINX_AES_DECRYPT)
drivers/crypto/xilinx/zynqmp-aes-gcm.c
418
if (rq_ctx->op == XILINX_AES_DECRYPT &&
drivers/crypto/xilinx/zynqmp-aes-gcm.c
612
rq_ctx->op = XILINX_AES_ENCRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
633
rq_ctx->op = XILINX_AES_ENCRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
657
rq_ctx->op = XILINX_AES_DECRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
698
rq_ctx->op = XILINX_AES_DECRYPT;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
70
u64 op;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
804
.aead.op = {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
833
.aead.op = {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
85
enum xilinx_aead_op op;
drivers/crypto/xilinx/zynqmp-aes-gcm.c
867
.aead.op = {
drivers/crypto/xilinx/zynqmp-aes-gcm.c
896
.aead.op = {
drivers/dma/bestcomm/bestcomm.c
365
static int mpc52xx_bcom_probe(struct platform_device *op)
drivers/dma/bestcomm/bestcomm.c
376
of_node_get(op->dev.of_node);
drivers/dma/bestcomm/bestcomm.c
403
bcom_eng->ofnode = op->dev.of_node;
drivers/dma/bestcomm/bestcomm.c
406
if (of_address_to_resource(op->dev.of_node, 0, &res_bcom)) {
drivers/dma/bestcomm/bestcomm.c
450
of_node_put(op->dev.of_node);
drivers/dma/bestcomm/bestcomm.c
458
static void mpc52xx_bcom_remove(struct platform_device *op)
drivers/dma/dmatest.c
1003
dma_chan_name(chan), op, i);
drivers/dma/dmatest.c
1006
dma_chan_name(chan), op, i);
drivers/dma/dmatest.c
975
char *op;
drivers/dma/dmatest.c
979
op = "copy";
drivers/dma/dmatest.c
981
op = "set";
drivers/dma/dmatest.c
983
op = "xor";
drivers/dma/dmatest.c
985
op = "pq";
drivers/dma/dmatest.c
993
dma_chan_name(chan), op, i);
drivers/dma/fsldma.c
1214
static int fsldma_of_probe(struct platform_device *op)
drivers/dma/fsldma.c
1227
fdev->dev = &op->dev;
drivers/dma/fsldma.c
1233
fdev->regs = of_iomap(op->dev.of_node, 0);
drivers/dma/fsldma.c
1235
dev_err(&op->dev, "unable to ioremap registers\n");
drivers/dma/fsldma.c
1241
fdev->irq = irq_of_parse_and_map(op->dev.of_node, 0);
drivers/dma/fsldma.c
1252
fdev->common.dev = &op->dev;
drivers/dma/fsldma.c
1259
dma_set_mask(&(op->dev), DMA_BIT_MASK(fdev->addr_bits));
drivers/dma/fsldma.c
1261
platform_set_drvdata(op, fdev);
drivers/dma/fsldma.c
1268
for_each_child_of_node(op->dev.of_node, child) {
drivers/dma/fsldma.c
1311
static void fsldma_of_remove(struct platform_device *op)
drivers/dma/fsldma.c
1316
fdev = platform_get_drvdata(op);
drivers/dma/ioat/dma.c
502
if (hw->ctl_f.op == IOAT_OP_XOR ||
drivers/dma/ioat/dma.c
503
hw->ctl_f.op == IOAT_OP_XOR_VAL) {
drivers/dma/ioat/dma.c
508
} else if (hw->ctl_f.op == IOAT_OP_PQ ||
drivers/dma/ioat/dma.c
509
hw->ctl_f.op == IOAT_OP_PQ_VAL) {
drivers/dma/ioat/dma.c
561
switch (hw->ctl_f.op) {
drivers/dma/ioat/dma.c
801
switch (hw->ctl_f.op) {
drivers/dma/ioat/dma.h
230
hw->ctl, hw->ctl_f.op, hw->ctl_f.int_en, hw->ctl_f.compl_write);
drivers/dma/ioat/hw.h
119
unsigned int op:8;
drivers/dma/ioat/hw.h
173
unsigned int op:8;
drivers/dma/ioat/hw.h
217
unsigned int op:8;
drivers/dma/ioat/hw.h
86
unsigned int op:8;
drivers/dma/ioat/init.c
1017
if (op == IOAT_OP_XOR) {
drivers/dma/ioat/init.c
1022
} else if (op == IOAT_OP_XOR_VAL) {
drivers/dma/ioat/init.c
791
u8 op = 0;
drivers/dma/ioat/init.c
838
op = IOAT_OP_XOR;
drivers/dma/ioat/init.c
906
op = IOAT_OP_XOR_VAL;
drivers/dma/ioat/init.c
965
op = IOAT_OP_XOR_VAL;
drivers/dma/ioat/prep.c
166
u8 op = result ? IOAT_OP_XOR_VAL : IOAT_OP_XOR;
drivers/dma/ioat/prep.c
214
xor->ctl_f.op = op;
drivers/dma/ioat/prep.c
291
desc->txd.flags, pq->size, pq->ctl, pq->ctl_f.op,
drivers/dma/ioat/prep.c
325
pq->ctl_f.op, pq->ctl_f.int_en,
drivers/dma/ioat/prep.c
354
u8 op = result ? IOAT_OP_PQ_VAL : IOAT_OP_PQ;
drivers/dma/ioat/prep.c
420
pq->ctl_f.op = op;
drivers/dma/ioat/prep.c
474
u8 op;
drivers/dma/ioat/prep.c
478
op = result ? IOAT_OP_PQ_VAL_16S : IOAT_OP_PQ_16S;
drivers/dma/ioat/prep.c
534
pq->ctl_f.op = op;
drivers/dma/k3dma.c
832
static int k3_dma_probe(struct platform_device *op)
drivers/dma/k3dma.c
838
d = devm_kzalloc(&op->dev, sizeof(*d), GFP_KERNEL);
drivers/dma/k3dma.c
842
soc_data = device_get_match_data(&op->dev);
drivers/dma/k3dma.c
846
d->base = devm_platform_ioremap_resource(op, 0);
drivers/dma/k3dma.c
850
of_property_read_u32((&op->dev)->of_node,
drivers/dma/k3dma.c
852
of_property_read_u32((&op->dev)->of_node,
drivers/dma/k3dma.c
854
ret = of_property_read_u32((&op->dev)->of_node,
drivers/dma/k3dma.c
857
dev_warn(&op->dev,
drivers/dma/k3dma.c
863
d->clk = devm_clk_get(&op->dev, NULL);
drivers/dma/k3dma.c
865
dev_err(&op->dev, "no dma clk\n");
drivers/dma/k3dma.c
870
irq = platform_get_irq(op, 0);
drivers/dma/k3dma.c
871
ret = devm_request_irq(&op->dev, irq,
drivers/dma/k3dma.c
879
d->pool = dmam_pool_create(DRIVER_NAME, &op->dev,
drivers/dma/k3dma.c
885
d->phy = devm_kcalloc(&op->dev,
drivers/dma/k3dma.c
905
d->slave.dev = &op->dev;
drivers/dma/k3dma.c
920
d->chans = devm_kcalloc(&op->dev,
drivers/dma/k3dma.c
937
dev_err(&op->dev, "clk_prepare_enable failed: %d\n", ret);
drivers/dma/k3dma.c
947
ret = of_dma_controller_register((&op->dev)->of_node,
drivers/dma/k3dma.c
955
platform_set_drvdata(op, d);
drivers/dma/k3dma.c
956
dev_info(&op->dev, "initialized\n");
drivers/dma/k3dma.c
967
static void k3_dma_remove(struct platform_device *op)
drivers/dma/k3dma.c
970
struct k3_dma_dev *d = platform_get_drvdata(op);
drivers/dma/k3dma.c
973
of_dma_controller_free((&op->dev)->of_node);
drivers/dma/k3dma.c
975
devm_free_irq(&op->dev, d->irq, d);
drivers/dma/mmp_pdma.c
1108
static void mmp_pdma_remove(struct platform_device *op)
drivers/dma/mmp_pdma.c
1110
struct mmp_pdma_device *pdev = platform_get_drvdata(op);
drivers/dma/mmp_pdma.c
1114
if (op->dev.of_node)
drivers/dma/mmp_pdma.c
1115
of_dma_controller_free(op->dev.of_node);
drivers/dma/mmp_pdma.c
1118
if (platform_get_irq(op, i) > 0)
drivers/dma/mmp_pdma.c
1123
irq = platform_get_irq(op, 0);
drivers/dma/mmp_pdma.c
1124
devm_free_irq(&op->dev, irq, pdev);
drivers/dma/mmp_pdma.c
1128
irq = platform_get_irq(op, i);
drivers/dma/mmp_pdma.c
1129
devm_free_irq(&op->dev, irq, phy);
drivers/dma/mmp_pdma.c
1225
static int mmp_pdma_probe(struct platform_device *op)
drivers/dma/mmp_pdma.c
1228
struct mmp_dma_platdata *pdata = dev_get_platdata(&op->dev);
drivers/dma/mmp_pdma.c
1237
pdev = devm_kzalloc(&op->dev, sizeof(*pdev), GFP_KERNEL);
drivers/dma/mmp_pdma.c
1241
pdev->dev = &op->dev;
drivers/dma/mmp_pdma.c
1245
pdev->base = devm_platform_ioremap_resource(op, 0);
drivers/dma/mmp_pdma.c
1258
pdev->ops = of_device_get_match_data(&op->dev);
drivers/dma/mmp_pdma.c
1276
if (platform_get_irq_optional(op, i) > 0)
drivers/dma/mmp_pdma.c
1289
irq = platform_get_irq(op, 0);
drivers/dma/mmp_pdma.c
1297
irq = (irq_num != dma_channels) ? 0 : platform_get_irq(op, i);
drivers/dma/mmp_pdma.c
1307
pdev->device.dev = &op->dev;
drivers/dma/mmp_pdma.c
1333
if (op->dev.of_node) {
drivers/dma/mmp_pdma.c
1335
ret = of_dma_controller_register(op->dev.of_node,
drivers/dma/mmp_pdma.c
1338
dev_err(&op->dev, "of_dma_controller_register failed\n");
drivers/dma/mmp_pdma.c
1344
platform_set_drvdata(op, pdev);
drivers/dma/mpc512x_dma.c
1087
static void mpc_dma_remove(struct platform_device *op)
drivers/dma/mpc512x_dma.c
1089
struct device *dev = &op->dev;
drivers/dma/mpc512x_dma.c
898
static int mpc_dma_probe(struct platform_device *op)
drivers/dma/mpc512x_dma.c
900
struct device_node *dn = op->dev.of_node;
drivers/dma/mpc512x_dma.c
901
struct device *dev = &op->dev;
drivers/dma/ppc4xx/adma.c
2088
unsigned long op = 0;
drivers/dma/ppc4xx/adma.c
2091
set_bit(PPC440SPE_DESC_WXOR, &op);
drivers/dma/ppc4xx/adma.c
2104
set_bits(op, &sw_desc->flags);
drivers/dma/ppc4xx/adma.c
2174
unsigned long op = 0;
drivers/dma/ppc4xx/adma.c
2177
set_bit(PPC440SPE_DESC_WXOR, &op);
drivers/dma/ppc4xx/adma.c
2190
set_bits(op, &sw_desc->flags);
drivers/dma/ppc4xx/adma.c
2279
unsigned long op = 0;
drivers/dma/ppc4xx/adma.c
2288
set_bit(PPC440SPE_DESC_WXOR, &op);
drivers/dma/ppc4xx/adma.c
2300
set_bit(PPC440SPE_DESC_RXOR, &op);
drivers/dma/ppc4xx/adma.c
2306
&op);
drivers/dma/ppc4xx/adma.c
2309
set_bit(PPC440SPE_DESC_RXOR124, &op);
drivers/dma/ppc4xx/adma.c
2313
&op);
drivers/dma/ppc4xx/adma.c
2317
&op);
drivers/dma/ppc4xx/adma.c
2321
set_bit(PPC440SPE_DESC_RXOR12, &op);
drivers/dma/ppc4xx/adma.c
2325
if (!test_bit(PPC440SPE_DESC_RXOR, &op)) {
drivers/dma/ppc4xx/adma.c
2336
if (!test_bit(PPC440SPE_DESC_RXOR, &op)) {
drivers/dma/ppc4xx/adma.c
2345
set_bit(PPC440SPE_ZERO_P, &op);
drivers/dma/ppc4xx/adma.c
2349
set_bit(PPC440SPE_ZERO_Q, &op);
drivers/dma/ppc4xx/adma.c
2359
set_bit(PPC440SPE_ZERO_P, &op);
drivers/dma/ppc4xx/adma.c
2361
set_bit(PPC440SPE_ZERO_Q, &op);
drivers/dma/ppc4xx/adma.c
2363
if (test_bit(PPC440SPE_DESC_RXOR12, &op))
drivers/dma/ppc4xx/adma.c
2373
clear_bit(PPC440SPE_DESC_WXOR, &op);
drivers/dma/ppc4xx/adma.c
2381
flags, op);
drivers/dma/ppc4xx/adma.c
2424
unsigned long op = 0;
drivers/dma/ppc4xx/adma.c
2443
op = slot_cnt;
drivers/dma/ppc4xx/adma.c
2447
--op ? 0 : flags);
drivers/dma/ppc4xx/adma.c
2456
op = 0;
drivers/dma/ppc4xx/adma.c
2458
op++;
drivers/dma/ppc4xx/adma.c
2459
if (op % descs_per_op == 0)
drivers/dma/ppc4xx/adma.c
319
unsigned long op)
drivers/dma/ppc4xx/adma.c
326
set_bits(op, &desc->flags);
drivers/dma/ppc4xx/adma.c
3361
int k = 0, op = 0, lop = 0;
drivers/dma/ppc4xx/adma.c
3364
while (op <= index) {
drivers/dma/ppc4xx/adma.c
3365
lop = op;
drivers/dma/ppc4xx/adma.c
3375
op += 2;
drivers/dma/ppc4xx/adma.c
3377
op += 3;
drivers/dma/ppc4xx/adma.c
3384
if (index == op - 1)
drivers/dma/ppc4xx/adma.c
3402
int k = 0, op = 0, lop = 0;
drivers/dma/ppc4xx/adma.c
3405
while (op <= index) {
drivers/dma/ppc4xx/adma.c
3406
lop = op;
drivers/dma/ppc4xx/adma.c
3417
op += 2;
drivers/dma/ppc4xx/adma.c
3419
op += 3;
drivers/dma/ppc4xx/adma.c
3425
ppc440spe_rxor_set_mult(desc, k - 1, op - index - 1, mult);
drivers/dma/ppc4xx/adma.c
3704
unsigned long op = 0;
drivers/dma/ppc4xx/adma.c
3707
set_bit(PPC440SPE_DESC_WXOR, &op);
drivers/dma/ppc4xx/adma.c
3717
ppc440spe_desc_init_dma01pq(sw_desc, 1, 1, 1, op);
drivers/dma/ppc4xx/dma.h
148
u32 op;
drivers/dma/pxa_dma.c
1222
static void pxad_remove(struct platform_device *op)
drivers/dma/pxa_dma.c
1224
struct pxad_device *pdev = platform_get_drvdata(op);
drivers/dma/pxa_dma.c
1230
static int pxad_init_phys(struct platform_device *op,
drivers/dma/pxa_dma.c
1237
irq0 = platform_get_irq(op, 0);
drivers/dma/pxa_dma.c
1241
pdev->phys = devm_kcalloc(&op->dev, nb_phy_chans,
drivers/dma/pxa_dma.c
1247
if (platform_get_irq_optional(op, i) > 0)
drivers/dma/pxa_dma.c
1254
irq = platform_get_irq_optional(op, i);
drivers/dma/pxa_dma.c
1256
ret = devm_request_irq(&op->dev, irq,
drivers/dma/pxa_dma.c
1260
ret = devm_request_irq(&op->dev, irq0,
drivers/dma/pxa_dma.c
1296
static int pxad_init_dmadev(struct platform_device *op,
drivers/dma/pxa_dma.c
1316
if (op->dev.coherent_dma_mask)
drivers/dma/pxa_dma.c
1317
dma_set_mask(&op->dev, op->dev.coherent_dma_mask);
drivers/dma/pxa_dma.c
1319
dma_set_mask(&op->dev, DMA_BIT_MASK(32));
drivers/dma/pxa_dma.c
1321
ret = pxad_init_phys(op, pdev, nr_phy_chans);
drivers/dma/pxa_dma.c
1326
c = devm_kzalloc(&op->dev, sizeof(*c), GFP_KERNEL);
drivers/dma/pxa_dma.c
1340
static int pxad_probe(struct platform_device *op)
drivers/dma/pxa_dma.c
1344
struct mmp_dma_platdata *pdata = dev_get_platdata(&op->dev);
drivers/dma/pxa_dma.c
1350
pdev = devm_kzalloc(&op->dev, sizeof(*pdev), GFP_KERNEL);
drivers/dma/pxa_dma.c
1356
pdev->base = devm_platform_ioremap_resource(op, 0);
drivers/dma/pxa_dma.c
1360
if (op->dev.of_node) {
drivers/dma/pxa_dma.c
1362
if (of_property_read_u32(op->dev.of_node, "dma-channels",
drivers/dma/pxa_dma.c
1364
of_property_read_u32(op->dev.of_node, "#dma-channels",
drivers/dma/pxa_dma.c
1367
ret = of_property_read_u32(op->dev.of_node, "dma-requests",
drivers/dma/pxa_dma.c
1370
ret = of_property_read_u32(op->dev.of_node, "#dma-requests",
drivers/dma/pxa_dma.c
1405
pdev->slave.dev = &op->dev;
drivers/dma/pxa_dma.c
1406
ret = pxad_init_dmadev(op, pdev, dma_channels, nb_requestors);
drivers/dma/pxa_dma.c
1412
if (op->dev.of_node) {
drivers/dma/pxa_dma.c
1414
ret = of_dma_controller_register(op->dev.of_node,
drivers/dma/pxa_dma.c
1423
platform_set_drvdata(op, pdev);
drivers/dma/qcom/gpi.c
1654
if (i2c->op == I2C_WRITE) {
drivers/dma/qcom/gpi.c
1661
tre->dword[0] = u32_encode_bits(i2c->op, TRE_I2C_GO_CMD);
drivers/dma/qcom/gpi.c
1677
if (i2c->op == I2C_READ || i2c->multi_msg == false) {
drivers/dpll/zl3073x/core.c
419
zl3073x_do_hwreg_op(struct zl3073x_dev *zldev, u8 op)
drivers/dpll/zl3073x/core.c
424
rc = zl3073x_write_u8(zldev, ZL_REG_HWREG_OP, op | ZL_HWREG_OP_PENDING);
drivers/edac/fsl_ddr_edac.c
494
int fsl_mc_err_probe(struct platform_device *op)
drivers/edac/fsl_ddr_edac.c
504
if (!devres_open_group(&op->dev, fsl_mc_err_probe, GFP_KERNEL))
drivers/edac/fsl_ddr_edac.c
516
devres_release_group(&op->dev, fsl_mc_err_probe);
drivers/edac/fsl_ddr_edac.c
522
mci->pdev = &op->dev;
drivers/edac/fsl_ddr_edac.c
528
pdata->flag = (unsigned long)device_get_match_data(&op->dev);
drivers/edac/fsl_ddr_edac.c
534
pdata->little_endian = of_property_read_bool(op->dev.of_node, "little-endian");
drivers/edac/fsl_ddr_edac.c
536
res = of_address_to_resource(op->dev.of_node, 0, &r);
drivers/edac/fsl_ddr_edac.c
543
if (!devm_request_mem_region(&op->dev, r.start, resource_size(&r),
drivers/edac/fsl_ddr_edac.c
551
pdata->mc_vbase = devm_ioremap(&op->dev, r.start, resource_size(&r));
drivers/edac/fsl_ddr_edac.c
559
pdata->inject_vbase = devm_platform_ioremap_resource_byname(op, "inject");
drivers/edac/fsl_ddr_edac.c
625
pdata->irq = platform_get_irq(op, 0);
drivers/edac/fsl_ddr_edac.c
626
res = devm_request_irq(&op->dev, pdata->irq,
drivers/edac/fsl_ddr_edac.c
641
devres_remove_group(&op->dev, fsl_mc_err_probe);
drivers/edac/fsl_ddr_edac.c
648
edac_mc_del_mc(&op->dev);
drivers/edac/fsl_ddr_edac.c
650
devres_release_group(&op->dev, fsl_mc_err_probe);
drivers/edac/fsl_ddr_edac.c
655
void fsl_mc_err_remove(struct platform_device *op)
drivers/edac/fsl_ddr_edac.c
657
struct mem_ctl_info *mci = dev_get_drvdata(&op->dev);
drivers/edac/fsl_ddr_edac.c
671
edac_mc_del_mc(&op->dev);
drivers/edac/fsl_ddr_edac.h
87
int fsl_mc_err_probe(struct platform_device *op);
drivers/edac/fsl_ddr_edac.h
88
void fsl_mc_err_remove(struct platform_device *op);
drivers/edac/mpc85xx_edac.c
145
static int mpc85xx_pci_err_probe(struct platform_device *op)
drivers/edac/mpc85xx_edac.c
154
if (!devres_open_group(&op->dev, mpc85xx_pci_err_probe, GFP_KERNEL))
drivers/edac/mpc85xx_edac.c
174
plat_data = op->dev.platform_data;
drivers/edac/mpc85xx_edac.c
176
dev_err(&op->dev, "no platform data");
drivers/edac/mpc85xx_edac.c
185
dev_set_drvdata(&op->dev, pci);
drivers/edac/mpc85xx_edac.c
186
pci->dev = &op->dev;
drivers/edac/mpc85xx_edac.c
189
pci->dev_name = dev_name(&op->dev);
drivers/edac/mpc85xx_edac.c
209
if (!devm_request_mem_region(&op->dev, r.start, resource_size(&r),
drivers/edac/mpc85xx_edac.c
216
pdata->pci_vbase = devm_ioremap(&op->dev, r.start, resource_size(&r));
drivers/edac/mpc85xx_edac.c
257
res = devm_request_irq(&op->dev, pdata->irq,
drivers/edac/mpc85xx_edac.c
289
devres_remove_group(&op->dev, mpc85xx_pci_err_probe);
drivers/edac/mpc85xx_edac.c
296
edac_pci_del_device(&op->dev);
drivers/edac/mpc85xx_edac.c
299
devres_release_group(&op->dev, mpc85xx_pci_err_probe);
drivers/edac/mpc85xx_edac.c
303
static void mpc85xx_pci_err_remove(struct platform_device *op)
drivers/edac/mpc85xx_edac.c
305
struct edac_pci_ctl_info *pci = dev_get_drvdata(&op->dev);
drivers/edac/mpc85xx_edac.c
313
edac_pci_del_device(&op->dev);
drivers/edac/mpc85xx_edac.c
488
static int mpc85xx_l2_err_probe(struct platform_device *op)
drivers/edac/mpc85xx_edac.c
495
if (!devres_open_group(&op->dev, mpc85xx_l2_err_probe, GFP_KERNEL))
drivers/edac/mpc85xx_edac.c
502
devres_release_group(&op->dev, mpc85xx_l2_err_probe);
drivers/edac/mpc85xx_edac.c
508
edac_dev->dev = &op->dev;
drivers/edac/mpc85xx_edac.c
513
res = of_address_to_resource(op->dev.of_node, 0, &r);
drivers/edac/mpc85xx_edac.c
522
if (!devm_request_mem_region(&op->dev, r.start, resource_size(&r),
drivers/edac/mpc85xx_edac.c
529
pdata->l2_vbase = devm_ioremap(&op->dev, r.start, resource_size(&r));
drivers/edac/mpc85xx_edac.c
558
pdata->irq = irq_of_parse_and_map(op->dev.of_node, 0);
drivers/edac/mpc85xx_edac.c
559
res = devm_request_irq(&op->dev, pdata->irq,
drivers/edac/mpc85xx_edac.c
577
devres_remove_group(&op->dev, mpc85xx_l2_err_probe);
drivers/edac/mpc85xx_edac.c
585
edac_device_del_device(&op->dev);
drivers/edac/mpc85xx_edac.c
587
devres_release_group(&op->dev, mpc85xx_l2_err_probe);
drivers/edac/mpc85xx_edac.c
592
static void mpc85xx_l2_err_remove(struct platform_device *op)
drivers/edac/mpc85xx_edac.c
594
struct edac_device_ctl_info *edac_dev = dev_get_drvdata(&op->dev);
drivers/edac/mpc85xx_edac.c
605
edac_device_del_device(&op->dev);
drivers/edac/pnd2_edac.c
157
static int _apl_rd_reg(int port, int off, int op, u32 *data)
drivers/edac/pnd2_edac.c
177
P2SB_WRITE(word, P2SB_STAT_OFF, (op << 8) | P2SB_BUSY);
drivers/edac/pnd2_edac.c
196
static int apl_rd_reg(int port, int off, int op, void *data, size_t sz, char *name)
drivers/edac/pnd2_edac.c
200
edac_dbg(2, "Read %s port=%x off=%x op=%x\n", name, port, off, op);
drivers/edac/pnd2_edac.c
203
ret = _apl_rd_reg(port, off + 4, op, (u32 *)(data + 4));
drivers/edac/pnd2_edac.c
206
ret |= _apl_rd_reg(port, off, op, (u32 *)data);
drivers/edac/pnd2_edac.c
240
static int dnv_rd_reg(int port, int off, int op, void *data, size_t sz, char *name)
drivers/edac/pnd2_edac.c
247
if (op == 4) {
drivers/edac/pnd2_edac.c
256
if (op == 0 && port == 0x4c) {
drivers/edac/pnd2_edac.c
91
int (*rd_reg)(int port, int off, int op, void *data, size_t sz, char *name);
drivers/firmware/cirrus/cs_dsp.c
3529
struct cs_dsp_wseq_op *op, *op_tmp;
drivers/firmware/cirrus/cs_dsp.c
3531
list_for_each_entry_safe(op, op_tmp, &wseq->ops, list) {
drivers/firmware/cirrus/cs_dsp.c
3532
list_del(&op->list);
drivers/firmware/cirrus/cs_dsp.c
3533
devm_kfree(dsp->dev, op);
drivers/firmware/cirrus/cs_dsp.c
3539
struct cs_dsp_wseq_op *op = NULL;
drivers/firmware/cirrus/cs_dsp.c
3564
op = devm_kzalloc(dsp->dev, sizeof(*op), GFP_KERNEL);
drivers/firmware/cirrus/cs_dsp.c
3565
if (!op) {
drivers/firmware/cirrus/cs_dsp.c
3570
op->offset = cs_dsp_chunk_bytes(&chunk);
drivers/firmware/cirrus/cs_dsp.c
3571
op->operation = cs_dsp_chunk_read(&chunk, 8);
drivers/firmware/cirrus/cs_dsp.c
3573
switch (op->operation) {
drivers/firmware/cirrus/cs_dsp.c
3575
op->data = WSEQ_END_OF_SCRIPT;
drivers/firmware/cirrus/cs_dsp.c
3578
op->data = cs_dsp_chunk_read(&chunk, 16);
drivers/firmware/cirrus/cs_dsp.c
3581
op->address = cs_dsp_chunk_read(&chunk, 8);
drivers/firmware/cirrus/cs_dsp.c
3582
op->data = cs_dsp_chunk_read(&chunk, 32);
drivers/firmware/cirrus/cs_dsp.c
3586
op->address = cs_dsp_chunk_read(&chunk, 24);
drivers/firmware/cirrus/cs_dsp.c
3587
op->data = cs_dsp_chunk_read(&chunk, 16);
drivers/firmware/cirrus/cs_dsp.c
3590
op->address = cs_dsp_chunk_read(&chunk, 32);
drivers/firmware/cirrus/cs_dsp.c
3591
op->data = cs_dsp_chunk_read(&chunk, 32);
drivers/firmware/cirrus/cs_dsp.c
3595
cs_dsp_err(dsp, "Unsupported op: %X\n", op->operation);
drivers/firmware/cirrus/cs_dsp.c
3596
devm_kfree(dsp->dev, op);
drivers/firmware/cirrus/cs_dsp.c
3600
list_add_tail(&op->list, &wseq->ops);
drivers/firmware/cirrus/cs_dsp.c
3602
if (op->operation == CS_DSP_WSEQ_END)
drivers/firmware/cirrus/cs_dsp.c
3606
if (op && op->operation != CS_DSP_WSEQ_END) {
drivers/firmware/cirrus/cs_dsp.c
3646
struct cs_dsp_wseq_op *op;
drivers/firmware/cirrus/cs_dsp.c
3648
list_for_each_entry(op, wseq_ops, list) {
drivers/firmware/cirrus/cs_dsp.c
3649
if (op->operation == op_code && op->address == addr)
drivers/firmware/cirrus/cs_dsp.c
3650
return op;
drivers/firmware/efi/cper-x86.c
204
u8 op = CHECK_OPERATION(check);
drivers/firmware/efi/cper-x86.c
212
printk("%sOperation: %u, %s\n", pfx, op,
drivers/firmware/efi/cper-x86.c
213
op < max_ops ? ia_check_op_strs[op] : "unknown");
drivers/firmware/imx/sm-lmm.c
37
int scmi_imx_lmm_operation(u32 lmid, enum scmi_imx_lmm_op op, u32 flags)
drivers/firmware/imx/sm-lmm.c
42
switch (op) {
drivers/fsi/fsi-master-ast-cf.c
307
static int do_copro_command(struct fsi_master_acf *master, uint32_t op)
drivers/fsi/fsi-master-ast-cf.c
312
trace_fsi_master_acf_copro_command(master, op);
drivers/fsi/fsi-master-ast-cf.c
315
iowrite32be(op, master->sram + CMD_STAT_REG);
drivers/fsi/fsi-master-ast-cf.c
361
uint32_t op;
drivers/fsi/fsi-master-ast-cf.c
369
op = CMD_COMMAND;
drivers/fsi/fsi-master-ast-cf.c
370
op |= cmd->bits << CMD_REG_CLEN_SHIFT;
drivers/fsi/fsi-master-ast-cf.c
372
op |= resp_bits << CMD_REG_RLEN_SHIFT;
drivers/fsi/fsi-master-ast-cf.c
374
return do_copro_command(master, op);
drivers/fwctl/main.c
172
const struct fwctl_ioctl_op *op;
drivers/fwctl/main.c
182
op = &fwctl_ioctl_ops[nr - FWCTL_CMD_BASE];
drivers/fwctl/main.c
183
if (op->ioctl_num != cmd)
drivers/fwctl/main.c
193
if (ucmd.user_size < op->min_size)
drivers/fwctl/main.c
196
ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer,
drivers/fwctl/main.c
204
return op->execute(&ucmd);
drivers/fwctl/pds/main.c
338
if (PDS_FWCTL_RPC_OPCODE_CMP(rpc->in.op, le32_to_cpu(op_entry[i].id))) {
drivers/fwctl/pds/main.c
345
dev_dbg(dev, "Invalid operation %d for endpoint %d\n", rpc->in.op, rpc->in.ep);
drivers/fwctl/pds/main.c
407
.op = cpu_to_le32(rpc->in.op),
drivers/fwctl/pds/main.c
418
__func__, rpc->in.ep, rpc->in.op,
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
688
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1035
if (args->op == AMDGPU_GEM_OP_GET_MAPPING_INFO) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1043
switch (args->op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
699
if (args->op == AMDGPU_GEM_METADATA_OP_GET_METADATA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
705
} else if (args->op == AMDGPU_GEM_METADATA_OP_SET_METADATA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
485
op_input.op = MES_MISC_OP_READ_REG;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
515
op_input.op = MES_MISC_OP_WRITE_REG;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
544
op_input.op = MES_MISC_OP_WRM_REG_WR_WAIT;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
605
op_input.op = MES_MISC_OP_SET_SHADER_DEBUGGER;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
646
op_input.op = MES_MISC_OP_SET_SHADER_DEBUGGER;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
768
op_input.op = MES_MISC_OP_CHANGE_CONFIG;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
337
enum mes_misc_opcode op;
drivers/gpu/drm/amd/amdgpu/amdgpu_rap.c
50
uint32_t op;
drivers/gpu/drm/amd/amdgpu/amdgpu_rap.c
57
ret = kstrtouint_from_user(buf, size, *pos, &op);
drivers/gpu/drm/amd/amdgpu/amdgpu_rap.c
72
switch (op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_rap.c
74
ret = psp_rap_invoke(&adev->psp, op, &status);
drivers/gpu/drm/amd/amdgpu/amdgpu_rap.c
98
dev_info(adev->dev, "Unsupported op id: %d, ", op);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
338
int op = -1;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
356
op = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
358
op = 1;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
360
op = 2;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
362
op = 3;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
364
op = 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
369
if (op != -1) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
370
if (op == 3) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
375
data->op = op;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
379
} else if (op == 4) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
384
data->op = op;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
404
data->op = op;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
406
if (op == 2) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
571
if (data.op == 3) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
577
} else if (data.op == 4) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
585
switch (data.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
756
int op;
drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
102
DRM_ERROR("Invalid sched op specified: %d\n", args->in.op);
drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
111
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sched.c
97
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_securedisplay.c
118
sscanf(str, "%u ", &op);
drivers/gpu/drm/amd/amdgpu/amdgpu_securedisplay.c
120
sscanf(str, "%u %u", &op, &phy_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_securedisplay.c
122
switch (op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_securedisplay.c
99
uint32_t op;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1022
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1041
drm_dbg_driver(dev, "Invalid user queue op specified: %d\n", args->in.op);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
931
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2925
switch (args->in.op) {
drivers/gpu/drm/amd/amdgpu/atom.c
1225
unsigned char op;
drivers/gpu/drm/amd/amdgpu/atom.c
1261
op = CU8(ptr++);
drivers/gpu/drm/amd/amdgpu/atom.c
1262
if (op < ATOM_OP_NAMES_CNT)
drivers/gpu/drm/amd/amdgpu/atom.c
1263
SDEBUG("%s @ 0x%04X\n", atom_op_names[op], ptr - 1);
drivers/gpu/drm/amd/amdgpu/atom.c
1265
SDEBUG("[%d] @ 0x%04X\n", op, ptr - 1);
drivers/gpu/drm/amd/amdgpu/atom.c
1273
if (op < ATOM_OP_CNT && op > 0)
drivers/gpu/drm/amd/amdgpu/atom.c
1274
opcode_table[op].func(&ectx, &ptr,
drivers/gpu/drm/amd/amdgpu/atom.c
1275
opcode_table[op].arg);
drivers/gpu/drm/amd/amdgpu/atom.c
1279
if (op == ATOM_OP_EOT)
drivers/gpu/drm/amd/amdgpu/cikd.h
229
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/cikd.h
230
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/cikd.h
233
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/cikd.h
495
#define SDMA_PACKET(op, sub_op, e) ((((e) & 0xFFFF) << 16) | \
drivers/gpu/drm/amd/amdgpu/cikd.h
497
(((op) & 0xFF) << 0))
drivers/gpu/drm/amd/amdgpu/gfx_v12_1_pkt.h
48
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/gfx_v12_1_pkt.h
49
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/gfx_v12_1_pkt.h
52
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
611
switch (input->op) {
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
624
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WAIT_REG_MEM;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
632
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WR_WAIT_WR_REG;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
665
drm_err(adev_to_drm(mes->adev), "unsupported misc op (%d)\n", input->op);
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
652
switch (input->op) {
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
665
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WAIT_REG_MEM;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
673
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WR_WAIT_WR_REG;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
702
DRM_ERROR("unsupported misc op (%d)\n", input->op);
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
534
switch (input->op) {
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
561
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WAIT_REG_MEM;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
576
misc_pkt.wait_reg_mem.op = WRM_OPERATION__WR_WAIT_WR_REG;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
619
DRM_ERROR("unsupported misc op (%d) \n", input->op);
drivers/gpu/drm/amd/amdgpu/nvd.h
48
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/nvd.h
49
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/nvd.h
52
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/sid.h
336
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/sid.h
337
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/sid.h
340
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/soc15d.h
50
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/soc15d.h
51
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/soc15d.h
54
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/vid.h
105
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/amd/amdgpu/vid.h
106
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/amd/amdgpu/vid.h
109
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/amd/amdgpu/vpe_6_1_fw_if.h
59
#define VPE_CMD_HEADER(op, subop) \
drivers/gpu/drm/amd/amdgpu/vpe_6_1_fw_if.h
61
((op << VPE_HEADER_OPCODE__SHIFT) & VPE_HEADER_OPCODE_MASK))
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1742
args->start_addr, args->size, args->op, args->nattr);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1749
r = svm_ioctl(p, args->op, args->start_addr, args->size, args->nattr,
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2754
dev_dbg(kfd_device, "CRIU operation: %d\n", args->op);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2755
switch (args->op) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2772
dev_dbg(kfd_device, "Unsupported CRIU operation:%d\n", args->op);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2778
dev_dbg(kfd_device, "CRIU operation:%d err:%d\n", args->op, ret);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
2976
if (args->op == KFD_IOC_DBG_TRAP_ENABLE) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3004
if (target != p && args->op != KFD_IOC_DBG_TRAP_DISABLE
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3016
if (args->op != KFD_IOC_DBG_TRAP_ENABLE && !target->debug_trap_enabled) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3017
pr_err("PID %i not debug enabled for op %i\n", args->pid, args->op);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3023
(args->op == KFD_IOC_DBG_TRAP_SET_WAVE_LAUNCH_OVERRIDE ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3024
args->op == KFD_IOC_DBG_TRAP_SET_WAVE_LAUNCH_MODE ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3025
args->op == KFD_IOC_DBG_TRAP_SUSPEND_QUEUES ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3026
args->op == KFD_IOC_DBG_TRAP_RESUME_QUEUES ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3027
args->op == KFD_IOC_DBG_TRAP_SET_NODE_ADDRESS_WATCH ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3028
args->op == KFD_IOC_DBG_TRAP_CLEAR_NODE_ADDRESS_WATCH ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3029
args->op == KFD_IOC_DBG_TRAP_SET_FLAGS)) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3034
if (args->op == KFD_IOC_DBG_TRAP_SET_NODE_ADDRESS_WATCH ||
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3035
args->op == KFD_IOC_DBG_TRAP_CLEAR_NODE_ADDRESS_WATCH) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3037
args->op == KFD_IOC_DBG_TRAP_SET_NODE_ADDRESS_WATCH ?
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3048
switch (args->op) {
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
3142
pr_err("Invalid option: %i\n", args->op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1193
svm_range_add_child(struct svm_range *prange, struct svm_range *pchild, enum svm_work_list_ops op)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1196
pchild, pchild->start, pchild->last, prange, op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1199
pchild->work_item.op = op;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2328
switch (prange->work_item.op) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2366
prange->work_item.op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2419
prange->start, prange->last, prange->work_item.op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2445
pchild->work_item.op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2468
struct mm_struct *mm, enum svm_work_list_ops op)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2473
pr_debug("update exist prange 0x%p work op %d\n", prange, op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2475
if (op != SVM_OP_NULL &&
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2476
prange->work_item.op != SVM_OP_UNMAP_RANGE)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2477
prange->work_item.op = op;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2484
prange->work_item.op = op;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2488
prange, prange->start, prange->last, op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2509
if (prange->work_item.op == SVM_OP_UNMAP_RANGE) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2531
prange->work_item.op = SVM_OP_UNMAP_RANGE;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2978
if (prange->work_item.op == SVM_OP_UNMAP_RANGE) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2983
if (prange->work_item.op == SVM_OP_ADD_RANGE_AND_MAP ||
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
2984
prange->work_item.op == SVM_OP_ADD_RANGE) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4312
svm_ioctl(struct kfd_process *p, enum kfd_ioctl_svm_op op, uint64_t start,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
4321
switch (op) {
drivers/gpu/drm/amd/amdkfd/kfd_svm.h
164
int svm_ioctl(struct kfd_process *p, enum kfd_ioctl_svm_op op, uint64_t start,
drivers/gpu/drm/amd/amdkfd/kfd_svm.h
181
enum svm_work_list_ops op);
drivers/gpu/drm/amd/amdkfd/kfd_svm.h
63
enum svm_work_list_ops op;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
12
const struct mod_hdcp_atomic_op_i2c *op,
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
21
if (!op || op->size > sizeof(req->buffer))
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
28
loc->address = op->address;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
29
loc->offset = op->offset;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
30
loc->length = op->size;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
31
memcpy(req->buffer, op->data, op->size);
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
38
const struct mod_hdcp_atomic_op_aux *op,
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
46
if (!op || op->size > sizeof(req->buffer))
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
52
loc->address = op->address;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
53
loc->length = op->size;
drivers/gpu/drm/amd/display/dc/dc_fused_io.c
54
memcpy(req->buffer, op->data, op->size);
drivers/gpu/drm/amd/include/mes_v11_api_def.h
606
enum WRM_OPERATION op;
drivers/gpu/drm/amd/include/mes_v12_api_def.h
783
enum WRM_OPERATION op;
drivers/gpu/drm/bridge/analogix/anx7625.c
233
static int anx7625_aux_trans(struct anx7625_data *ctx, u8 op, u32 address,
drivers/gpu/drm/bridge/analogix/anx7625.c
240
bool is_write = !(op & DP_AUX_I2C_READ);
drivers/gpu/drm/bridge/analogix/anx7625.c
255
op &= ~DP_AUX_I2C_MOT;
drivers/gpu/drm/bridge/analogix/anx7625.c
256
cmd = DPCD_CMD(len, op);
drivers/gpu/drm/drm_gpuvm.c
2298
const struct drm_gpuva_op_map *op)
drivers/gpu/drm/drm_gpuvm.c
2300
drm_gpuva_init_from_op(va, op);
drivers/gpu/drm/drm_gpuvm.c
2318
const struct drm_gpuva_op_remap *op)
drivers/gpu/drm/drm_gpuvm.c
2320
struct drm_gpuva *va = op->unmap->va;
drivers/gpu/drm/drm_gpuvm.c
2325
if (op->prev) {
drivers/gpu/drm/drm_gpuvm.c
2326
drm_gpuva_init_from_op(prev, op->prev);
drivers/gpu/drm/drm_gpuvm.c
2330
if (op->next) {
drivers/gpu/drm/drm_gpuvm.c
2331
drm_gpuva_init_from_op(next, op->next);
drivers/gpu/drm/drm_gpuvm.c
2345
drm_gpuva_unmap(const struct drm_gpuva_op_unmap *op)
drivers/gpu/drm/drm_gpuvm.c
2347
drm_gpuva_remove(op->va);
drivers/gpu/drm/drm_gpuvm.c
2355
struct drm_gpuva_op op = {};
drivers/gpu/drm/drm_gpuvm.c
2360
op.op = DRM_GPUVA_OP_MAP;
drivers/gpu/drm/drm_gpuvm.c
2361
op.map.va.addr = req->map.va.addr;
drivers/gpu/drm/drm_gpuvm.c
2362
op.map.va.range = req->map.va.range;
drivers/gpu/drm/drm_gpuvm.c
2363
op.map.gem.obj = req->map.gem.obj;
drivers/gpu/drm/drm_gpuvm.c
2364
op.map.gem.offset = req->map.gem.offset;
drivers/gpu/drm/drm_gpuvm.c
2366
return fn->sm_step_map(&op, priv);
drivers/gpu/drm/drm_gpuvm.c
2375
struct drm_gpuva_op op = {};
drivers/gpu/drm/drm_gpuvm.c
2378
op.op = DRM_GPUVA_OP_REMAP;
drivers/gpu/drm/drm_gpuvm.c
2379
r = &op.remap;
drivers/gpu/drm/drm_gpuvm.c
2384
return fn->sm_step_remap(&op, priv);
drivers/gpu/drm/drm_gpuvm.c
2391
struct drm_gpuva_op op = {};
drivers/gpu/drm/drm_gpuvm.c
2396
op.op = DRM_GPUVA_OP_UNMAP;
drivers/gpu/drm/drm_gpuvm.c
2397
op.unmap.va = va;
drivers/gpu/drm/drm_gpuvm.c
2398
op.unmap.keep = merge;
drivers/gpu/drm/drm_gpuvm.c
2400
return fn->sm_step_unmap(&op, priv);
drivers/gpu/drm/drm_gpuvm.c
2727
drm_gpuva_sm_step_lock(struct drm_gpuva_op *op, void *priv)
drivers/gpu/drm/drm_gpuvm.c
2731
switch (op->op) {
drivers/gpu/drm/drm_gpuvm.c
2733
if (op->remap.unmap->va->gem.obj)
drivers/gpu/drm/drm_gpuvm.c
2734
return drm_exec_lock_obj(exec, op->remap.unmap->va->gem.obj);
drivers/gpu/drm/drm_gpuvm.c
2737
if (op->unmap.va->gem.obj)
drivers/gpu/drm/drm_gpuvm.c
2738
return drm_exec_lock_obj(exec, op->unmap.va->gem.obj);
drivers/gpu/drm/drm_gpuvm.c
2850
struct drm_gpuva_op *op;
drivers/gpu/drm/drm_gpuvm.c
2853
op = fn->op_alloc();
drivers/gpu/drm/drm_gpuvm.c
2855
op = kzalloc_obj(*op);
drivers/gpu/drm/drm_gpuvm.c
2857
if (unlikely(!op))
drivers/gpu/drm/drm_gpuvm.c
2860
return op;
drivers/gpu/drm/drm_gpuvm.c
2865
struct drm_gpuva_op *op)
drivers/gpu/drm/drm_gpuvm.c
2870
fn->op_free(op);
drivers/gpu/drm/drm_gpuvm.c
2872
kfree(op);
drivers/gpu/drm/drm_gpuvm.c
2885
struct drm_gpuva_op *op;
drivers/gpu/drm/drm_gpuvm.c
2887
op = gpuva_op_alloc(gpuvm);
drivers/gpu/drm/drm_gpuvm.c
2888
if (unlikely(!op))
drivers/gpu/drm/drm_gpuvm.c
2891
memcpy(op, __op, sizeof(*op));
drivers/gpu/drm/drm_gpuvm.c
2893
if (op->op == DRM_GPUVA_OP_REMAP) {
drivers/gpu/drm/drm_gpuvm.c
2895
struct drm_gpuva_op_remap *r = &op->remap;
drivers/gpu/drm/drm_gpuvm.c
2917
list_add_tail(&op->entry, &ops->list);
drivers/gpu/drm/drm_gpuvm.c
2922
kfree(op->remap.unmap);
drivers/gpu/drm/drm_gpuvm.c
2924
kfree(op->remap.prev);
drivers/gpu/drm/drm_gpuvm.c
2926
gpuva_op_free(gpuvm, op);
drivers/gpu/drm/drm_gpuvm.c
3128
struct drm_gpuva_op *op;
drivers/gpu/drm/drm_gpuvm.c
3140
op = gpuva_op_alloc(gpuvm);
drivers/gpu/drm/drm_gpuvm.c
3141
if (!op) {
drivers/gpu/drm/drm_gpuvm.c
3146
op->op = DRM_GPUVA_OP_PREFETCH;
drivers/gpu/drm/drm_gpuvm.c
3147
op->prefetch.va = va;
drivers/gpu/drm/drm_gpuvm.c
3148
list_add_tail(&op->entry, &ops->list);
drivers/gpu/drm/drm_gpuvm.c
3181
struct drm_gpuva_op *op;
drivers/gpu/drm/drm_gpuvm.c
3194
op = gpuva_op_alloc(vm_bo->vm);
drivers/gpu/drm/drm_gpuvm.c
3195
if (!op) {
drivers/gpu/drm/drm_gpuvm.c
3200
op->op = DRM_GPUVA_OP_UNMAP;
drivers/gpu/drm/drm_gpuvm.c
3201
op->unmap.va = va;
drivers/gpu/drm/drm_gpuvm.c
3202
list_add_tail(&op->entry, &ops->list);
drivers/gpu/drm/drm_gpuvm.c
3225
struct drm_gpuva_op *op, *next;
drivers/gpu/drm/drm_gpuvm.c
3227
drm_gpuva_for_each_op_safe(op, next, ops) {
drivers/gpu/drm/drm_gpuvm.c
3228
list_del(&op->entry);
drivers/gpu/drm/drm_gpuvm.c
3230
if (op->op == DRM_GPUVA_OP_REMAP) {
drivers/gpu/drm/drm_gpuvm.c
3231
kfree(op->remap.prev);
drivers/gpu/drm/drm_gpuvm.c
3232
kfree(op->remap.next);
drivers/gpu/drm/drm_gpuvm.c
3233
kfree(op->remap.unmap);
drivers/gpu/drm/drm_gpuvm.c
3236
gpuva_op_free(gpuvm, op);
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
165
unsigned int op = cmd >> 27;
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
167
switch (op) {
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
188
len = cmd_length[op];
drivers/gpu/drm/etnaviv/etnaviv_cmd_parser.c
191
__func__, op, buf - state.start);
drivers/gpu/drm/etnaviv/etnaviv_drv.c
304
if (args->op & ~(ETNA_PREP_READ | ETNA_PREP_WRITE | ETNA_PREP_NOSYNC))
drivers/gpu/drm/etnaviv/etnaviv_drv.c
311
ret = etnaviv_gem_cpu_prep(obj, args->op, &args->timeout);
drivers/gpu/drm/etnaviv/etnaviv_drv.h
67
int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
369
static inline enum dma_data_direction etnaviv_op_to_dma_dir(u32 op)
drivers/gpu/drm/etnaviv/etnaviv_gem.c
371
op &= ETNA_PREP_READ | ETNA_PREP_WRITE;
drivers/gpu/drm/etnaviv/etnaviv_gem.c
373
if (op == ETNA_PREP_READ)
drivers/gpu/drm/etnaviv/etnaviv_gem.c
375
else if (op == ETNA_PREP_WRITE)
drivers/gpu/drm/etnaviv/etnaviv_gem.c
381
int etnaviv_gem_cpu_prep(struct drm_gem_object *obj, u32 op,
drivers/gpu/drm/etnaviv/etnaviv_gem.c
386
bool write = !!(op & ETNA_PREP_WRITE);
drivers/gpu/drm/etnaviv/etnaviv_gem.c
399
if (op & ETNA_PREP_NOSYNC) {
drivers/gpu/drm/etnaviv/etnaviv_gem.c
414
etnaviv_op_to_dma_dir(op));
drivers/gpu/drm/etnaviv/etnaviv_gem.c
415
etnaviv_obj->last_cpu_prep_op = op;
drivers/gpu/drm/i915/gt/intel_gpu_commands.h
370
#define MEDIA_INSTR(pipe, op, sub_op, flags) \
drivers/gpu/drm/i915/gt/intel_gpu_commands.h
372
(op) << INSTR_26_TO_24_SHIFT | (sub_op) << 16 | (flags))
drivers/gpu/drm/i915/gt/intel_gt_sysfs_pm.c
55
enum intel_gt_sysfs_op op)
drivers/gpu/drm/i915/gt/intel_gt_sysfs_pm.c
60
ret = (op == INTEL_GT_SYSFS_MAX) ? 0 : (u32) -1;
drivers/gpu/drm/i915/gt/intel_gt_sysfs_pm.c
70
switch (op) {
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
34
static u32 *emit_wait(u32 *cs, u32 offset, int op, u32 value)
drivers/gpu/drm/i915/gt/selftest_engine_pm.c
39
op;
drivers/gpu/drm/i915/gt/selftest_timeline.c
893
bool (*op)(u32 hwsp, u32 seqno))
drivers/gpu/drm/i915/gt/selftest_timeline.c
914
if (!op(w->map[offset + 1], w->map[offset])) {
drivers/gpu/drm/i915/gvt/cmd_parser.c
321
#define OP_MFX(pipeline, op, sub_opa, sub_opb) \
drivers/gpu/drm/i915/gvt/cmd_parser.c
324
(op) << 8 | \
drivers/gpu/drm/i915/gvt/cmd_parser.c
375
#define OP_VEB(pipeline, op, sub_opa, sub_opb) \
drivers/gpu/drm/i915/gvt/cmd_parser.c
378
(op) << 8 | \
drivers/gpu/drm/i915/gvt/edid.c
489
int msg, addr, ctrl, op;
drivers/gpu/drm/i915/gvt/edid.c
505
op = ctrl >> 4;
drivers/gpu/drm/i915/gvt/edid.c
512
ret_msg_size = (((op & 0x1) == DP_AUX_I2C_READ) ? 2 : 1);
drivers/gpu/drm/i915/gvt/edid.c
518
if (!(op & DP_AUX_I2C_MOT)) {
drivers/gpu/drm/i915/gvt/edid.c
538
} else if ((op & 0x1) == DP_AUX_I2C_WRITE) {
drivers/gpu/drm/i915/gvt/edid.c
546
if (drm_WARN_ON(&i915->drm, (op & 0x1) != DP_AUX_I2C_READ))
drivers/gpu/drm/i915/gvt/handlers.c
1191
int msg, addr, ctrl, op, len;
drivers/gpu/drm/i915/gvt/handlers.c
1231
op = ctrl >> 4;
drivers/gpu/drm/i915/gvt/handlers.c
1233
if (op == DP_AUX_NATIVE_WRITE) {
drivers/gpu/drm/i915/gvt/handlers.c
1292
if (op == DP_AUX_NATIVE_READ) {
drivers/gpu/drm/i915/i915_cmd_parser.c
199
#define CMD(op, opm, f, lm, fl, ...) \
drivers/gpu/drm/i915/i915_cmd_parser.c
202
.cmd = { (op & ~0u << (opm)), ~0u << (opm) }, \
drivers/gpu/drm/i915/i915_cmd_parser.c
761
u32 op = (cmd_header & INSTR_26_TO_24_MASK) >> INSTR_26_TO_24_SHIFT;
drivers/gpu/drm/i915/i915_cmd_parser.c
767
if (op == 6)
drivers/gpu/drm/i915/intel_uncore.c
2877
i915_reg_t reg, unsigned int op)
drivers/gpu/drm/i915/intel_uncore.c
2881
drm_WARN_ON(&uncore->i915->drm, !op);
drivers/gpu/drm/i915/intel_uncore.c
2886
if (op & FW_REG_READ)
drivers/gpu/drm/i915/intel_uncore.c
2889
if (op & FW_REG_WRITE)
drivers/gpu/drm/i915/intel_uncore.h
255
i915_reg_t reg, unsigned int op);
drivers/gpu/drm/imagination/pvr_vm.c
354
pvr_vm_gpuva_map(struct drm_gpuva_op *op, void *op_ctx)
drivers/gpu/drm/imagination/pvr_vm.c
356
struct pvr_gem_object *pvr_gem = gem_to_pvr_gem(op->map.gem.obj);
drivers/gpu/drm/imagination/pvr_vm.c
360
if ((op->map.gem.offset | op->map.va.range) & ~PVR_DEVICE_PAGE_MASK)
drivers/gpu/drm/imagination/pvr_vm.c
363
err = pvr_mmu_map(ctx->mmu_op_ctx, op->map.va.range, pvr_gem->flags,
drivers/gpu/drm/imagination/pvr_vm.c
364
op->map.va.addr);
drivers/gpu/drm/imagination/pvr_vm.c
368
drm_gpuva_map(&ctx->vm_ctx->gpuvm_mgr, &ctx->new_va->base, &op->map);
drivers/gpu/drm/imagination/pvr_vm.c
388
pvr_vm_gpuva_unmap(struct drm_gpuva_op *op, void *op_ctx)
drivers/gpu/drm/imagination/pvr_vm.c
392
int err = pvr_mmu_unmap(ctx->mmu_op_ctx, op->unmap.va->va.addr,
drivers/gpu/drm/imagination/pvr_vm.c
393
op->unmap.va->va.range);
drivers/gpu/drm/imagination/pvr_vm.c
398
drm_gpuva_unmap(&op->unmap);
drivers/gpu/drm/imagination/pvr_vm.c
399
drm_gpuva_unlink(op->unmap.va);
drivers/gpu/drm/imagination/pvr_vm.c
400
kfree(to_pvr_vm_gpuva(op->unmap.va));
drivers/gpu/drm/imagination/pvr_vm.c
419
pvr_vm_gpuva_remap(struct drm_gpuva_op *op, void *op_ctx)
drivers/gpu/drm/imagination/pvr_vm.c
425
drm_gpuva_op_remap_to_unmap_range(&op->remap, &va_start, &va_range);
drivers/gpu/drm/imagination/pvr_vm.c
433
drm_gpuva_remap(&ctx->prev_va->base, &ctx->next_va->base, &op->remap);
drivers/gpu/drm/imagination/pvr_vm.c
435
if (op->remap.prev) {
drivers/gpu/drm/imagination/pvr_vm.c
441
if (op->remap.next) {
drivers/gpu/drm/imagination/pvr_vm.c
447
drm_gpuva_unlink(op->remap.unmap->va);
drivers/gpu/drm/imagination/pvr_vm.c
448
kfree(to_pvr_vm_gpuva(op->remap.unmap->va));
drivers/gpu/drm/lima/lima_drv.c
183
if (args->op & ~(LIMA_GEM_WAIT_READ|LIMA_GEM_WAIT_WRITE))
drivers/gpu/drm/lima/lima_drv.c
186
return lima_gem_wait(file, args->handle, args->op, args->timeout_ns);
drivers/gpu/drm/lima/lima_gem.c
401
int lima_gem_wait(struct drm_file *file, u32 handle, u32 op, s64 timeout_ns)
drivers/gpu/drm/lima/lima_gem.c
403
bool write = op & LIMA_GEM_WAIT_WRITE;
drivers/gpu/drm/lima/lima_gem.c
406
if (!op)
drivers/gpu/drm/lima/lima_gem.h
43
int lima_gem_wait(struct drm_file *file, u32 handle, u32 op, s64 timeout_ns);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
973
e->op, e->queue_id, e->iova,
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
191
uint32_t op;
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
194
op = MDP4_OVLP_STAGE_OP_FG_ALPHA(FG_PIXEL) |
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
198
op = MDP4_OVLP_STAGE_OP_FG_ALPHA(FG_CONST) |
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
204
mdp4_write(mdp4_kms, REG_MDP4_OVLP_STAGE_OP(ovlp, i), op);
drivers/gpu/drm/msm/msm_drv.c
387
if (args->op & ~MSM_PREP_FLAGS) {
drivers/gpu/drm/msm/msm_drv.c
388
DRM_ERROR("invalid op: %08x\n", args->op);
drivers/gpu/drm/msm/msm_drv.c
396
ret = msm_gem_cpu_prep(obj, args->op, &timeout);
drivers/gpu/drm/msm/msm_gem.c
899
int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout)
drivers/gpu/drm/msm/msm_gem.c
901
bool write = !!(op & MSM_PREP_WRITE);
drivers/gpu/drm/msm/msm_gem.c
903
op & MSM_PREP_NOSYNC ? 0 : timeout_to_jiffies(timeout);
drivers/gpu/drm/msm/msm_gem.c
906
if (op & MSM_PREP_BOOST) {
drivers/gpu/drm/msm/msm_gem.h
294
int msm_gem_cpu_prep(struct drm_gem_object *obj, uint32_t op, ktime_t *timeout);
drivers/gpu/drm/msm/msm_gem.h
35
const char *op;
drivers/gpu/drm/msm/msm_gem_vma.c
1001
if (invalid_alignment(op->iova))
drivers/gpu/drm/msm/msm_gem_vma.c
1002
ret = UERR(EINVAL, dev, "invalid address: %016llx\n", op->iova);
drivers/gpu/drm/msm/msm_gem_vma.c
1004
if (invalid_alignment(op->obj_offset))
drivers/gpu/drm/msm/msm_gem_vma.c
1005
ret = UERR(EINVAL, dev, "invalid bo_offset: %016llx\n", op->obj_offset);
drivers/gpu/drm/msm/msm_gem_vma.c
1007
if (invalid_alignment(op->range))
drivers/gpu/drm/msm/msm_gem_vma.c
1008
ret = UERR(EINVAL, dev, "invalid range: %016llx\n", op->range);
drivers/gpu/drm/msm/msm_gem_vma.c
1010
if (!drm_gpuvm_range_valid(job->vm, op->iova, op->range))
drivers/gpu/drm/msm/msm_gem_vma.c
1011
ret = UERR(EINVAL, dev, "invalid range: %016llx, %016llx\n", op->iova, op->range);
drivers/gpu/drm/msm/msm_gem_vma.c
1017
if (op->op == MSM_VM_BIND_OP_MAP) {
drivers/gpu/drm/msm/msm_gem_vma.c
1018
if (!op->handle)
drivers/gpu/drm/msm/msm_gem_vma.c
1020
} else if (op->handle) {
drivers/gpu/drm/msm/msm_gem_vma.c
1024
switch (op->op) {
drivers/gpu/drm/msm/msm_gem_vma.c
1030
ret = UERR(EINVAL, dev, "invalid op: %u\n", op->op);
drivers/gpu/drm/msm/msm_gem_vma.c
1034
if ((op->op == MSM_VM_BIND_OP_MAP_NULL) &&
drivers/gpu/drm/msm/msm_gem_vma.c
1056
ret = lookup_op(job, &args->op);
drivers/gpu/drm/msm/msm_gem_vma.c
1059
struct drm_msm_vm_bind_op op;
drivers/gpu/drm/msm/msm_gem_vma.c
1061
u64_to_user_ptr(args->ops + (i * sizeof(op)));
drivers/gpu/drm/msm/msm_gem_vma.c
1068
if (copy_from_user(&op, userptr, sizeof(op))) {
drivers/gpu/drm/msm/msm_gem_vma.c
1073
ret = lookup_op(job, &op);
drivers/gpu/drm/msm/msm_gem_vma.c
1087
struct msm_vm_bind_op *op = &job->ops[i];
drivers/gpu/drm/msm/msm_gem_vma.c
1090
if (!op->handle) {
drivers/gpu/drm/msm/msm_gem_vma.c
1091
op->obj = NULL;
drivers/gpu/drm/msm/msm_gem_vma.c
1099
obj = idr_find(&file->object_idr, op->handle);
drivers/gpu/drm/msm/msm_gem_vma.c
1101
ret = UERR(EINVAL, dev, "invalid handle %u at index %u\n", op->handle, i);
drivers/gpu/drm/msm/msm_gem_vma.c
1107
op->obj = obj;
drivers/gpu/drm/msm/msm_gem_vma.c
1110
if ((op->range + op->obj_offset) > obj->size) {
drivers/gpu/drm/msm/msm_gem_vma.c
1112
op->range, op->obj_offset, obj->size);
drivers/gpu/drm/msm/msm_gem_vma.c
1124
struct msm_vm_bind_op *op = &job->ops[i];
drivers/gpu/drm/msm/msm_gem_vma.c
1126
if (!op->obj)
drivers/gpu/drm/msm/msm_gem_vma.c
1129
drm_gem_object_put(op->obj);
drivers/gpu/drm/msm/msm_gem_vma.c
1130
op->obj = NULL;
drivers/gpu/drm/msm/msm_gem_vma.c
1181
struct msm_vm_bind_op *op = &job->ops[i];
drivers/gpu/drm/msm/msm_gem_vma.c
1184
if (op->op == MSM_VM_BIND_OP_UNMAP)
drivers/gpu/drm/msm/msm_gem_vma.c
1192
if (last && ops_are_same_pte(last, op)) {
drivers/gpu/drm/msm/msm_gem_vma.c
1193
last = op;
drivers/gpu/drm/msm/msm_gem_vma.c
1202
first = last = op;
drivers/gpu/drm/msm/msm_gem_vma.c
1239
const struct msm_vm_bind_op *op = &job->ops[i];
drivers/gpu/drm/msm/msm_gem_vma.c
1241
switch (op->op) {
drivers/gpu/drm/msm/msm_gem_vma.c
1244
op->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
1245
op->obj_offset);
drivers/gpu/drm/msm/msm_gem_vma.c
1250
.map.va.addr = op->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
1251
.map.va.range = op->range,
drivers/gpu/drm/msm/msm_gem_vma.c
1252
.map.gem.obj = op->obj,
drivers/gpu/drm/msm/msm_gem_vma.c
1253
.map.gem.offset = op->obj_offset,
drivers/gpu/drm/msm/msm_gem_vma.c
1351
const struct msm_vm_bind_op *op = &job->ops[i];
drivers/gpu/drm/msm/msm_gem_vma.c
1354
.op = op,
drivers/gpu/drm/msm/msm_gem_vma.c
1357
switch (op->op) {
drivers/gpu/drm/msm/msm_gem_vma.c
1359
ret = drm_gpuvm_sm_unmap(job->vm, &arg, op->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
1360
op->range);
drivers/gpu/drm/msm/msm_gem_vma.c
1363
if (op->flags & MSM_VM_BIND_OP_DUMP)
drivers/gpu/drm/msm/msm_gem_vma.c
1368
.map.va.addr = op->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
1369
.map.va.range = op->range,
drivers/gpu/drm/msm/msm_gem_vma.c
137
uint32_t op;
drivers/gpu/drm/msm/msm_gem_vma.c
1370
.map.gem.obj = op->obj,
drivers/gpu/drm/msm/msm_gem_vma.c
1371
.map.gem.offset = op->obj_offset,
drivers/gpu/drm/msm/msm_gem_vma.c
188
if (!vm->log || !vm->log[0].op)
drivers/gpu/drm/msm/msm_gem_vma.c
200
if (!vm->log[first].op) {
drivers/gpu/drm/msm/msm_gem_vma.c
217
e->op, e->queue_id, e->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
225
vm_log(struct msm_gem_vm *vm, const char *op, uint64_t iova, uint64_t range, int queue_id)
drivers/gpu/drm/msm/msm_gem_vma.c
232
vm_dbg("%s:%p:%d: %016llx %016llx", op, vm, queue_id, iova, iova + range);
drivers/gpu/drm/msm/msm_gem_vma.c
238
vm->log[idx].op = op;
drivers/gpu/drm/msm/msm_gem_vma.c
246
vm_unmap_op(struct msm_gem_vm *vm, const struct msm_vm_unmap_op *op)
drivers/gpu/drm/msm/msm_gem_vma.c
248
const char *reason = op->reason;
drivers/gpu/drm/msm/msm_gem_vma.c
253
vm_log(vm, reason, op->iova, op->range, op->queue_id);
drivers/gpu/drm/msm/msm_gem_vma.c
255
vm->mmu->funcs->unmap(vm->mmu, op->iova, op->range);
drivers/gpu/drm/msm/msm_gem_vma.c
259
vm_map_op(struct msm_gem_vm *vm, const struct msm_vm_map_op *op)
drivers/gpu/drm/msm/msm_gem_vma.c
261
vm_log(vm, "map", op->iova, op->range, op->queue_id);
drivers/gpu/drm/msm/msm_gem_vma.c
263
return vm->mmu->funcs->map(vm->mmu, op->iova, op->sgt, op->offset,
drivers/gpu/drm/msm/msm_gem_vma.c
264
op->range, op->prot);
drivers/gpu/drm/msm/msm_gem_vma.c
461
const struct msm_vm_bind_op *op;
drivers/gpu/drm/msm/msm_gem_vma.c
468
struct msm_vm_op *op = kmalloc_obj(*op);
drivers/gpu/drm/msm/msm_gem_vma.c
469
if (!op)
drivers/gpu/drm/msm/msm_gem_vma.c
472
*op = _op;
drivers/gpu/drm/msm/msm_gem_vma.c
473
list_add_tail(&op->node, &arg->job->vm_ops);
drivers/gpu/drm/msm/msm_gem_vma.c
475
if (op->obj)
drivers/gpu/drm/msm/msm_gem_vma.c
476
drm_gem_object_get(op->obj);
drivers/gpu/drm/msm/msm_gem_vma.c
482
vma_from_op(struct op_arg *arg, struct drm_gpuva_op_map *op)
drivers/gpu/drm/msm/msm_gem_vma.c
484
return msm_gem_vma_new(arg->job->vm, op->gem.obj, op->gem.offset,
drivers/gpu/drm/msm/msm_gem_vma.c
485
op->va.addr, op->va.addr + op->va.range);
drivers/gpu/drm/msm/msm_gem_vma.c
489
msm_gem_vm_sm_step_map(struct drm_gpuva_op *op, void *_arg)
drivers/gpu/drm/msm/msm_gem_vma.c
493
struct drm_gem_object *obj = op->map.gem.obj;
drivers/gpu/drm/msm/msm_gem_vma.c
502
vma = vma_from_op(arg, &op->map);
drivers/gpu/drm/msm/msm_gem_vma.c
518
.op = MSM_VM_OP_MAP,
drivers/gpu/drm/msm/msm_gem_vma.c
540
msm_gem_vm_sm_step_remap(struct drm_gpuva_op *op, void *arg)
drivers/gpu/drm/msm/msm_gem_vma.c
544
struct drm_gpuva *orig_vma = op->remap.unmap->va;
drivers/gpu/drm/msm/msm_gem_vma.c
557
drm_gpuva_op_remap_to_unmap_range(&op->remap, &unmap_start, &unmap_range);
drivers/gpu/drm/msm/msm_gem_vma.c
560
.op = MSM_VM_OP_UNMAP,
drivers/gpu/drm/msm/msm_gem_vma.c
597
if (op->remap.prev) {
drivers/gpu/drm/msm/msm_gem_vma.c
598
prev_vma = vma_from_op(arg, op->remap.prev);
drivers/gpu/drm/msm/msm_gem_vma.c
607
if (op->remap.next) {
drivers/gpu/drm/msm/msm_gem_vma.c
608
next_vma = vma_from_op(arg, op->remap.next);
drivers/gpu/drm/msm/msm_gem_vma.c
627
msm_gem_vm_sm_step_unmap(struct drm_gpuva_op *op, void *_arg)
drivers/gpu/drm/msm/msm_gem_vma.c
631
struct drm_gpuva *vma = op->unmap.va;
drivers/gpu/drm/msm/msm_gem_vma.c
644
if (op->unmap.keep &&
drivers/gpu/drm/msm/msm_gem_vma.c
645
(arg->op->op == MSM_VM_BIND_OP_MAP) &&
drivers/gpu/drm/msm/msm_gem_vma.c
646
(vma->gem.obj == arg->op->obj) &&
drivers/gpu/drm/msm/msm_gem_vma.c
647
(vma->gem.offset == arg->op->obj_offset) &&
drivers/gpu/drm/msm/msm_gem_vma.c
648
(vma->va.addr == arg->op->iova) &&
drivers/gpu/drm/msm/msm_gem_vma.c
649
(vma->va.range == arg->op->range)) {
drivers/gpu/drm/msm/msm_gem_vma.c
667
.op = MSM_VM_OP_UNMAP,
drivers/gpu/drm/msm/msm_gem_vma.c
709
struct msm_vm_op *op =
drivers/gpu/drm/msm/msm_gem_vma.c
712
switch (op->op) {
drivers/gpu/drm/msm/msm_gem_vma.c
720
ret = vm_map_op(vm, &op->map);
drivers/gpu/drm/msm/msm_gem_vma.c
723
vm_unmap_op(vm, &op->unmap);
drivers/gpu/drm/msm/msm_gem_vma.c
726
drm_gem_object_put(op->obj);
drivers/gpu/drm/msm/msm_gem_vma.c
727
list_del(&op->node);
drivers/gpu/drm/msm/msm_gem_vma.c
728
kfree(op);
drivers/gpu/drm/msm/msm_gem_vma.c
75
} op;
drivers/gpu/drm/msm/msm_gem_vma.c
772
struct msm_vm_op *op =
drivers/gpu/drm/msm/msm_gem_vma.c
774
list_del(&op->node);
drivers/gpu/drm/msm/msm_gem_vma.c
775
kfree(op);
drivers/gpu/drm/msm/msm_gem_vma.c
984
lookup_op(struct msm_vm_bind_job *job, const struct drm_msm_vm_bind_op *op)
drivers/gpu/drm/msm/msm_gem_vma.c
991
job->ops[i].op = op->op;
drivers/gpu/drm/msm/msm_gem_vma.c
992
job->ops[i].handle = op->handle;
drivers/gpu/drm/msm/msm_gem_vma.c
993
job->ops[i].obj_offset = op->obj_offset;
drivers/gpu/drm/msm/msm_gem_vma.c
994
job->ops[i].iova = op->iova;
drivers/gpu/drm/msm/msm_gem_vma.c
995
job->ops[i].range = op->range;
drivers/gpu/drm/msm/msm_gem_vma.c
996
job->ops[i].flags = op->flags;
drivers/gpu/drm/msm/msm_gem_vma.c
998
if (op->flags & ~MSM_VM_BIND_OP_FLAGS)
drivers/gpu/drm/msm/msm_gem_vma.c
999
ret = UERR(EINVAL, dev, "invalid flags: %x\n", op->flags);
drivers/gpu/drm/msm/msm_gpu.c
325
if (!vm->log || !vm->log[0].op)
drivers/gpu/drm/msm/msm_gpu.c
337
if (!vm->log[first].op) {
drivers/gpu/drm/nouveau/include/nvif/if000c.h
80
__u8 op;
drivers/gpu/drm/nouveau/nouveau_gem.c
411
validate_fini_no_ticket(struct validate_op *op, struct nouveau_channel *chan,
drivers/gpu/drm/nouveau/nouveau_gem.c
418
while (!list_empty(&op->list)) {
drivers/gpu/drm/nouveau/nouveau_gem.c
419
nvbo = list_entry(op->list.next, struct nouveau_bo, entry);
drivers/gpu/drm/nouveau/nouveau_gem.c
447
validate_fini(struct validate_op *op, struct nouveau_channel *chan,
drivers/gpu/drm/nouveau/nouveau_gem.c
451
validate_fini_no_ticket(op, chan, fence, pbbo);
drivers/gpu/drm/nouveau/nouveau_gem.c
452
ww_acquire_fini(&op->ticket);
drivers/gpu/drm/nouveau/nouveau_gem.c
458
int nr_buffers, struct validate_op *op)
drivers/gpu/drm/nouveau/nouveau_gem.c
468
ww_acquire_init(&op->ticket, &reservation_ww_class);
drivers/gpu/drm/nouveau/nouveau_gem.c
501
ret = ttm_bo_reserve(&nvbo->bo, true, false, &op->ticket);
drivers/gpu/drm/nouveau/nouveau_gem.c
503
list_splice_tail_init(&vram_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
504
list_splice_tail_init(&gart_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
505
list_splice_tail_init(&both_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
506
validate_fini_no_ticket(op, chan, NULL, NULL);
drivers/gpu/drm/nouveau/nouveau_gem.c
509
&op->ticket);
drivers/gpu/drm/nouveau/nouveau_gem.c
556
ww_acquire_done(&op->ticket);
drivers/gpu/drm/nouveau/nouveau_gem.c
557
list_splice_tail(&vram_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
558
list_splice_tail(&gart_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
559
list_splice_tail(&both_list, &op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
561
validate_fini(op, chan, NULL, NULL);
drivers/gpu/drm/nouveau/nouveau_gem.c
626
struct validate_op *op, bool *apply_relocs)
drivers/gpu/drm/nouveau/nouveau_gem.c
631
INIT_LIST_HEAD(&op->list);
drivers/gpu/drm/nouveau/nouveau_gem.c
636
ret = validate_init(chan, file_priv, pbbo, nr_buffers, op);
drivers/gpu/drm/nouveau/nouveau_gem.c
643
ret = validate_list(chan, &op->list, pbbo);
drivers/gpu/drm/nouveau/nouveau_gem.c
647
validate_fini(op, chan, NULL, NULL);
drivers/gpu/drm/nouveau/nouveau_gem.c
756
struct validate_op op;
drivers/gpu/drm/nouveau/nouveau_gem.c
826
req->nr_buffers, &op, &do_reloc);
drivers/gpu/drm/nouveau/nouveau_gem.c
836
validate_fini(&op, chan, NULL, bo);
drivers/gpu/drm/nouveau/nouveau_gem.c
938
validate_fini(&op, chan, fence, bo);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1037
struct bind_job_op *op, *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1039
list_for_each_op_safe(op, next, &job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1040
list_del(&op->entry);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1041
kfree(op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1062
struct bind_job_op *op)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1065
struct drm_gem_object *obj = op->gem.obj;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1067
if (op->op == OP_MAP) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1068
if (op->gem.offset & ~PAGE_MASK)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1071
if (obj->size <= op->gem.offset)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1074
if (op->va.range > (obj->size - op->gem.offset))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1078
return nouveau_uvmm_validate_range(uvmm, op->va.addr, op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1086
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1094
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1095
if (op->op == OP_UNMAP) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1096
u64 op_addr = op->va.addr;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1097
u64 op_end = op_addr + op->va.range;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1160
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1163
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1164
u64 op_addr = op->va.addr;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1165
u64 op_range = op->va.range;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1168
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1193
struct drm_gpuva_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1195
drm_gpuva_for_each_op(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1196
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1201
struct drm_gpuva *va = op->remap.unmap->va;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1203
if (op->remap.prev)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1205
if (op->remap.next)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1211
drm_gpuva_unlink(op->unmap.va);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1224
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1227
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1230
if (!op->ops)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1233
drm_gpuva_for_each_op(va_op, op->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1246
if (va_op->op == DRM_GPUVA_OP_UNMAP)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1266
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1269
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1270
if (op->op == OP_MAP) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1271
struct drm_gem_object *obj = op->gem.obj =
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1273
op->gem.handle);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1278
op->vm_bo = drm_gpuvm_bo_obtain_locked(&uvmm->base, obj);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1280
if (IS_ERR(op->vm_bo))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1281
return PTR_ERR(op->vm_bo);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1283
drm_gpuvm_bo_extobj_add(op->vm_bo);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1286
ret = bind_validate_op(job, op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1307
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1308
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1311
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1312
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1318
op->reg = nouveau_uvma_region_find(uvmm, op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1319
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1320
if (!op->reg || op->reg->dirty) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1325
op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1326
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1327
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1328
if (IS_ERR(op->ops)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1329
ret = PTR_ERR(op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1333
ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1334
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1336
drm_gpuva_ops_free(&uvmm->base, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1337
op->ops = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1338
op->reg = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1342
nouveau_uvma_region_dirty(op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1348
.map.va.addr = op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1349
.map.va.range = op->va.range,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1350
.map.gem.obj = op->gem.obj,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1351
.map.gem.offset = op->gem.offset,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1355
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1356
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1360
u64 op_addr = op->va.addr;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1361
u64 op_end = op_addr + op->va.range;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1377
op->ops = drm_gpuvm_sm_map_ops_create(&uvmm->base,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1379
if (IS_ERR(op->ops)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1380
ret = PTR_ERR(op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1384
ret = nouveau_uvmm_sm_map_prepare(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1385
reg, op->ops,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1386
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1387
op->va.range,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1388
op->flags & 0xff);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1390
drm_gpuva_ops_free(&uvmm->base, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1391
op->ops = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1398
op->ops = drm_gpuvm_sm_unmap_ops_create(&uvmm->base,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1399
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1400
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1401
if (IS_ERR(op->ops)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1402
ret = PTR_ERR(op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1406
ret = nouveau_uvmm_sm_unmap_prepare(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1407
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1409
drm_gpuva_ops_free(&uvmm->base, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1410
op->ops = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1426
op = list_last_op(&bind_job->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1451
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1452
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1456
bind_link_gpuvas(op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1467
op = list_prev_op(op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1469
list_for_each_op_from_reverse(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1470
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1472
nouveau_uvma_region_destroy(uvmm, op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1473
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1476
__nouveau_uvma_region_insert(uvmm, op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1477
nouveau_uvmm_sm_unmap_prepare_unwind(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1478
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1481
nouveau_uvmm_sm_map_prepare_unwind(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1482
op->ops,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1483
op->va.addr,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1484
op->va.range);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1487
nouveau_uvmm_sm_unmap_prepare_unwind(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1488
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1492
drm_gpuva_ops_free(&uvmm->base, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1493
op->ops = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1494
op->reg = NULL;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1516
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1519
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1520
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1525
ret = nouveau_uvmm_sm_map(uvmm, &op->new, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1532
ret = nouveau_uvmm_sm_unmap(uvmm, &op->new, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1550
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1552
list_for_each_op(op, &bind_job->ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1553
struct drm_gem_object *obj = op->gem.obj;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1558
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1563
if (!IS_ERR_OR_NULL(op->ops))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1564
nouveau_uvmm_sm_unmap_cleanup(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1565
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1567
if (op->reg) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1568
nouveau_uvma_region_sparse_unref(op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1570
nouveau_uvma_region_remove(op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1572
nouveau_uvma_region_complete(op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1573
nouveau_uvma_region_put(op->reg);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1578
if (!IS_ERR_OR_NULL(op->ops))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1579
nouveau_uvmm_sm_map_cleanup(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1580
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1583
if (!IS_ERR_OR_NULL(op->ops))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1584
nouveau_uvmm_sm_unmap_cleanup(uvmm, &op->new,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1585
op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1589
if (!IS_ERR_OR_NULL(op->ops))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1590
drm_gpuva_ops_free(&uvmm->base, op->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1592
if (!IS_ERR_OR_NULL(op->vm_bo)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1594
drm_gpuvm_bo_put(op->vm_bo);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1619
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1621
op = *pop = kzalloc_obj(*op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1622
if (!op)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1625
switch (uop->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1627
op->op = uop->flags & DRM_NOUVEAU_VM_BIND_SPARSE ?
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1631
op->op = uop->flags & DRM_NOUVEAU_VM_BIND_SPARSE ?
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1635
op->op = uop->op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1639
op->flags = uop->flags;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1640
op->va.addr = uop->addr;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1641
op->va.range = uop->range;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1642
op->gem.handle = uop->handle;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1643
op->gem.offset = uop->bo_offset;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1651
struct bind_job_op *op, *next;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1653
list_for_each_op_safe(op, next, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1654
list_del(&op->entry);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1655
kfree(op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1665
struct bind_job_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1674
for (i = 0; i < __args->op.count; i++) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1675
ret = bind_job_op_from_uop(&op, &__args->op.s[i]);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1679
list_add_tail(&op->entry, &job->ops);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1751
args->op.count = opc;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1752
args->op.s = u_memcpya(ops, opc,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1753
sizeof(*args->op.s));
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1754
if (IS_ERR(args->op.s))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1755
return PTR_ERR(args->op.s);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1783
u_free(args->op.s);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1792
u_free(args->op.s);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
458
op_map_aligned_to_page_shift(const struct drm_gpuva_op_map *op, u8 page_shift)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
462
return (op->va.addr & non_page_bits) == 0 &&
drivers/gpu/drm/nouveau/nouveau_uvmm.c
463
(op->va.range & non_page_bits) == 0 &&
drivers/gpu/drm/nouveau/nouveau_uvmm.c
464
(op->gem.offset & non_page_bits) == 0;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
468
select_page_shift(struct nouveau_uvmm *uvmm, struct drm_gpuva_op_map *op)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
470
struct nouveau_bo *nvbo = nouveau_gem_object(op->gem.obj);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
480
if (op_map_aligned_to_page_shift(op, nvbo->page))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
501
if (op_map_aligned_to_page_shift(op, vmm->page[i].shift))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
508
drm_WARN_ONCE(op->gem.obj->dev, 1, "Could not find an appropriate page size.\n");
drivers/gpu/drm/nouveau/nouveau_uvmm.c
520
struct drm_gpuva_op *op = last;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
525
drm_gpuva_for_each_op_from_reverse(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
526
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
531
struct drm_gpuva_op_remap *r = &op->remap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
544
op_unmap_prepare_unwind(op->unmap.va);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
557
drm_gpuva_for_each_op(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
558
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
565
select_page_shift(uvmm, &op->map));
drivers/gpu/drm/nouveau/nouveau_uvmm.c
569
struct drm_gpuva_op_remap *r = &op->remap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
587
struct drm_gpuva_op_unmap *u = &op->unmap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
613
if (op == last)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
62
enum vm_bind_op op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
646
struct drm_gpuva_op_map *op,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
658
uvma->page_shift = select_page_shift(uvmm, op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
660
drm_gpuva_map(&uvmm->base, &uvma->va, op);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
684
struct drm_gpuva_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
689
drm_gpuva_for_each_op(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
690
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
694
ret = op_map_prepare(uvmm, &new->map, &op->map, args);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
711
struct drm_gpuva_op_remap *r = &op->remap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
752
struct drm_gpuva_op_unmap *u = &op->unmap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
792
if (op != drm_gpuva_first_op(ops))
drivers/gpu/drm/nouveau/nouveau_uvmm.c
794
drm_gpuva_prev_op(op),
drivers/gpu/drm/nouveau/nouveau_uvmm.c
825
op_gem_obj(struct drm_gpuva_op *op)
drivers/gpu/drm/nouveau/nouveau_uvmm.c
827
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
829
return op->map.gem.obj;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
835
return op->remap.unmap->va->gem.obj;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
837
return op->unmap.va->gem.obj;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
898
struct drm_gpuva_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
900
drm_gpuva_for_each_op(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
901
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
906
op_remap(&op->remap, new);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
909
op_unmap(&op->unmap);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
940
struct drm_gpuva_op *op;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
942
drm_gpuva_for_each_op(op, ops) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
943
switch (op->op) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
947
struct drm_gpuva_op_remap *r = &op->remap;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
972
struct drm_gpuva_op_unmap *u = &op->unmap;
drivers/gpu/drm/nouveau/nouveau_uvmm.h
73
} op;
drivers/gpu/drm/nouveau/nvif/vmm.c
113
.op = NVIF_VMM_RAW_V0_GET,
drivers/gpu/drm/nouveau/nvif/vmm.c
128
.op = NVIF_VMM_RAW_V0_PUT,
drivers/gpu/drm/nouveau/nvif/vmm.c
144
.op = NVIF_VMM_RAW_V0_MAP,
drivers/gpu/drm/nouveau/nvif/vmm.c
165
.op = NVIF_VMM_RAW_V0_UNMAP,
drivers/gpu/drm/nouveau/nvif/vmm.c
181
.op = NVIF_VMM_RAW_V0_SPARSE,
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1476
u32 op = nvkm_rd32(device, 0x40449c);
drivers/gpu/drm/nouveau/nvkm/engine/gr/gf100.c
1483
op);
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
462
int class, op, valid = 1;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
467
op = (ctx1 >> 15) & 7;
drivers/gpu/drm/nouveau/nvkm/engine/gr/nv04.c
481
switch (op) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/uvmm.c
475
switch (args->v0.op) {
drivers/gpu/drm/panel/panel-ilitek-ili9881c.c
2053
if (instr->op == ILI9881C_SWITCH_PAGE)
drivers/gpu/drm/panel/panel-ilitek-ili9881c.c
2055
else if (instr->op == ILI9881C_COMMAND)
drivers/gpu/drm/panel/panel-ilitek-ili9881c.c
29
enum ili9881c_op op;
drivers/gpu/drm/panel/panel-ilitek-ili9881c.c
63
.op = ILI9881C_SWITCH_PAGE, \
drivers/gpu/drm/panel/panel-ilitek-ili9881c.c
71
.op = ILI9881C_COMMAND, \
drivers/gpu/drm/panfrost/panfrost_mmu.c
140
u64 iova, u64 size, u32 op)
drivers/gpu/drm/panfrost/panfrost_mmu.c
145
if (op != AS_COMMAND_UNLOCK)
drivers/gpu/drm/panfrost/panfrost_mmu.c
149
write_cmd(pfdev, as_nr, op);
drivers/gpu/drm/panfrost/panfrost_mmu.c
157
u64 iova, u64 size, u32 op)
drivers/gpu/drm/panfrost/panfrost_mmu.c
162
ret = mmu_hw_do_operation_locked(pfdev, mmu->as, iova, size, op);
drivers/gpu/drm/panthor/panthor_drv.c
1234
struct drm_panthor_vm_bind_op *op = &jobs_args[i];
drivers/gpu/drm/panthor/panthor_drv.c
1237
job = panthor_vm_bind_job_create(file, vm, op);
drivers/gpu/drm/panthor/panthor_drv.c
1243
ret = panthor_submit_ctx_add_job(&ctx, i, job, &op->syncs);
drivers/gpu/drm/panthor/panthor_mmu.c
2076
static int panthor_gpuva_sm_step_map(struct drm_gpuva_op *op, void *priv)
drivers/gpu/drm/panthor/panthor_mmu.c
2088
ret = panthor_vm_map_pages(vm, op->map.va.addr, flags_to_prot(vma->flags),
drivers/gpu/drm/panthor/panthor_mmu.c
2089
op_ctx->map.sgt, op->map.gem.offset,
drivers/gpu/drm/panthor/panthor_mmu.c
2090
op->map.va.range);
drivers/gpu/drm/panthor/panthor_mmu.c
2096
drm_gpuva_map(&vm->base, &vma->base, &op->map);
drivers/gpu/drm/panthor/panthor_mmu.c
2106
iova_mapped_as_huge_page(struct drm_gpuva_op_map *op, u64 addr)
drivers/gpu/drm/panthor/panthor_mmu.c
2111
bo_offset = addr - op->va.addr + op->gem.offset;
drivers/gpu/drm/panthor/panthor_mmu.c
2112
pg = to_panthor_bo(op->gem.obj)->base.pages[bo_offset >> PAGE_SHIFT];
drivers/gpu/drm/panthor/panthor_mmu.c
2118
unmap_hugepage_align(const struct drm_gpuva_op_remap *op,
drivers/gpu/drm/panthor/panthor_mmu.c
2130
if (op->prev && aligned_unmap_start < *unmap_start &&
drivers/gpu/drm/panthor/panthor_mmu.c
2131
op->prev->va.addr <= aligned_unmap_start &&
drivers/gpu/drm/panthor/panthor_mmu.c
2132
iova_mapped_as_huge_page(op->prev, *unmap_start)) {
drivers/gpu/drm/panthor/panthor_mmu.c
2140
if (op->next && aligned_unmap_end > unmap_end &&
drivers/gpu/drm/panthor/panthor_mmu.c
2141
op->next->va.addr + op->next->va.range >= aligned_unmap_end &&
drivers/gpu/drm/panthor/panthor_mmu.c
2142
iova_mapped_as_huge_page(op->next, unmap_end - 1)) {
drivers/gpu/drm/panthor/panthor_mmu.c
2147
static int panthor_gpuva_sm_step_remap(struct drm_gpuva_op *op,
drivers/gpu/drm/panthor/panthor_mmu.c
2150
struct panthor_vma *unmap_vma = container_of(op->remap.unmap->va, struct panthor_vma, base);
drivers/gpu/drm/panthor/panthor_mmu.c
2157
drm_gpuva_op_remap_to_unmap_range(&op->remap, &unmap_start, &unmap_range);
drivers/gpu/drm/panthor/panthor_mmu.c
2166
unmap_hugepage_align(&op->remap, &unmap_start, &unmap_range);
drivers/gpu/drm/panthor/panthor_mmu.c
2175
if (op->remap.prev) {
drivers/gpu/drm/panthor/panthor_mmu.c
2176
struct panthor_gem_object *bo = to_panthor_bo(op->remap.prev->gem.obj);
drivers/gpu/drm/panthor/panthor_mmu.c
2177
u64 offset = op->remap.prev->gem.offset + unmap_start - op->remap.prev->va.addr;
drivers/gpu/drm/panthor/panthor_mmu.c
2178
u64 size = op->remap.prev->va.addr + op->remap.prev->va.range - unmap_start;
drivers/gpu/drm/panthor/panthor_mmu.c
2189
if (op->remap.next) {
drivers/gpu/drm/panthor/panthor_mmu.c
2190
struct panthor_gem_object *bo = to_panthor_bo(op->remap.next->gem.obj);
drivers/gpu/drm/panthor/panthor_mmu.c
2191
u64 addr = op->remap.next->va.addr;
drivers/gpu/drm/panthor/panthor_mmu.c
2192
u64 size = unmap_start + unmap_range - op->remap.next->va.addr;
drivers/gpu/drm/panthor/panthor_mmu.c
2195
bo->base.sgt, op->remap.next->gem.offset, size);
drivers/gpu/drm/panthor/panthor_mmu.c
2205
&op->remap);
drivers/gpu/drm/panthor/panthor_mmu.c
2213
panthor_vma_link(vm, prev_vma, op->remap.unmap->va->vm_bo);
drivers/gpu/drm/panthor/panthor_mmu.c
2217
panthor_vma_link(vm, next_vma, op->remap.unmap->va->vm_bo);
drivers/gpu/drm/panthor/panthor_mmu.c
2224
static int panthor_gpuva_sm_step_unmap(struct drm_gpuva_op *op,
drivers/gpu/drm/panthor/panthor_mmu.c
2227
struct panthor_vma *unmap_vma = container_of(op->unmap.va, struct panthor_vma, base);
drivers/gpu/drm/panthor/panthor_mmu.c
2232
drm_gpuva_unmap(&op->unmap);
drivers/gpu/drm/panthor/panthor_mmu.c
2265
panthor_vm_exec_op(struct panthor_vm *vm, struct panthor_vm_op_ctx *op,
drivers/gpu/drm/panthor/panthor_mmu.c
2268
u32 op_type = op->flags & DRM_PANTHOR_VM_BIND_OP_TYPE_MASK;
drivers/gpu/drm/panthor/panthor_mmu.c
2275
vm->op_ctx = op;
drivers/gpu/drm/panthor/panthor_mmu.c
2277
ret = panthor_vm_lock_region(vm, op->va.addr, op->va.range);
drivers/gpu/drm/panthor/panthor_mmu.c
2284
.map.va.addr = op->va.addr,
drivers/gpu/drm/panthor/panthor_mmu.c
2285
.map.va.range = op->va.range,
drivers/gpu/drm/panthor/panthor_mmu.c
2286
.map.gem.obj = op->map.vm_bo->obj,
drivers/gpu/drm/panthor/panthor_mmu.c
2287
.map.gem.offset = op->map.bo_offset,
drivers/gpu/drm/panthor/panthor_mmu.c
2300
ret = drm_gpuvm_sm_unmap(&vm->base, vm, op->va.addr, op->va.range);
drivers/gpu/drm/panthor/panthor_mmu.c
2527
const struct drm_panthor_vm_bind_op *op,
drivers/gpu/drm/panthor/panthor_mmu.c
2535
if (!IS_ALIGNED(op->va | op->size | op->bo_offset, vm_pgsz))
drivers/gpu/drm/panthor/panthor_mmu.c
2538
switch (op->flags & DRM_PANTHOR_VM_BIND_OP_TYPE_MASK) {
drivers/gpu/drm/panthor/panthor_mmu.c
2540
gem = drm_gem_object_lookup(file, op->bo_handle);
drivers/gpu/drm/panthor/panthor_mmu.c
2543
op->bo_offset,
drivers/gpu/drm/panthor/panthor_mmu.c
2544
op->size,
drivers/gpu/drm/panthor/panthor_mmu.c
2545
op->va,
drivers/gpu/drm/panthor/panthor_mmu.c
2546
op->flags);
drivers/gpu/drm/panthor/panthor_mmu.c
2551
if (op->flags & ~DRM_PANTHOR_VM_BIND_OP_TYPE_MASK)
drivers/gpu/drm/panthor/panthor_mmu.c
2554
if (op->bo_handle || op->bo_offset)
drivers/gpu/drm/panthor/panthor_mmu.c
2557
return panthor_vm_prepare_unmap_op_ctx(op_ctx, vm, op->va, op->size);
drivers/gpu/drm/panthor/panthor_mmu.c
2560
if (op->flags & ~DRM_PANTHOR_VM_BIND_OP_TYPE_MASK)
drivers/gpu/drm/panthor/panthor_mmu.c
2563
if (op->bo_handle || op->bo_offset)
drivers/gpu/drm/panthor/panthor_mmu.c
2566
if (op->va || op->size)
drivers/gpu/drm/panthor/panthor_mmu.c
2569
if (!op->syncs.count)
drivers/gpu/drm/panthor/panthor_mmu.c
2599
const struct drm_panthor_vm_bind_op *op)
drivers/gpu/drm/panthor/panthor_mmu.c
2614
ret = panthor_vm_bind_prepare_op_ctx(file, vm, op, &job->ctx);
drivers/gpu/drm/panthor/panthor_mmu.c
2702
struct drm_panthor_vm_bind_op *op)
drivers/gpu/drm/panthor/panthor_mmu.c
2708
if (op->syncs.count)
drivers/gpu/drm/panthor/panthor_mmu.c
2711
if (!op->size)
drivers/gpu/drm/panthor/panthor_mmu.c
2714
ret = panthor_vm_bind_prepare_op_ctx(file, vm, op, &op_ctx);
drivers/gpu/drm/panthor/panthor_mmu.h
83
struct drm_panthor_vm_bind_op *op);
drivers/gpu/drm/panthor/panthor_mmu.h
88
const struct drm_panthor_vm_bind_op *op);
drivers/gpu/drm/radeon/atom.c
1174
unsigned char op;
drivers/gpu/drm/radeon/atom.c
1205
op = CU8(ptr++);
drivers/gpu/drm/radeon/atom.c
1206
if (op < ATOM_OP_NAMES_CNT)
drivers/gpu/drm/radeon/atom.c
1207
SDEBUG("%s @ 0x%04X\n", atom_op_names[op], ptr - 1);
drivers/gpu/drm/radeon/atom.c
1209
SDEBUG("[%d] @ 0x%04X\n", op, ptr - 1);
drivers/gpu/drm/radeon/atom.c
1217
if (op < ATOM_OP_CNT && op > 0)
drivers/gpu/drm/radeon/atom.c
1218
opcode_table[op].func(&ectx, &ptr,
drivers/gpu/drm/radeon/atom.c
1219
opcode_table[op].arg);
drivers/gpu/drm/radeon/atom.c
1223
if (op == ATOM_OP_EOT)
drivers/gpu/drm/radeon/cikd.h
1691
#define PACKET3(op, n) ((PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/cikd.h
1692
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/cikd.h
1695
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/radeon/cikd.h
2003
#define SDMA_PACKET(op, sub_op, e) ((((e) & 0xFFFF) << 16) | \
drivers/gpu/drm/radeon/cikd.h
2005
(((op) & 0xFF) << 0))
drivers/gpu/drm/radeon/evergreend.h
1543
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/evergreend.h
1544
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/nid.h
1157
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/nid.h
1158
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/r100d.h
63
#define PACKET3(op, n) (CP_PACKET3 | \
drivers/gpu/drm/radeon/r100d.h
64
REG_SET(PACKET3_IT_OPCODE, (op)) | \
drivers/gpu/drm/radeon/r300d.h
64
#define PACKET3(op, n) (CP_PACKET3 | \
drivers/gpu/drm/radeon/r300d.h
65
REG_SET(PACKET3_IT_OPCODE, (op)) | \
drivers/gpu/drm/radeon/r600d.h
1587
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/r600d.h
1588
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/radeon_gem.c
791
switch (args->op) {
drivers/gpu/drm/radeon/rv515d.h
204
#define PACKET3(op, n) (CP_PACKET3 | \
drivers/gpu/drm/radeon/rv515d.h
205
REG_SET(PACKET3_IT_OPCODE, (op)) | \
drivers/gpu/drm/radeon/rv770d.h
988
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/rv770d.h
989
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/sid.h
1595
#define PACKET3(op, n) ((RADEON_PACKET_TYPE3 << 30) | \
drivers/gpu/drm/radeon/sid.h
1596
(((op) & 0xFF) << 8) | \
drivers/gpu/drm/radeon/sid.h
1599
#define PACKET3_COMPUTE(op, n) (PACKET3(op, n) | 1 << 1)
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
575
switch (arg->op) {
drivers/gpu/drm/xe/instructions/xe_gsc_commands.h
26
#define __GSC_INSTR(op, dl) \
drivers/gpu/drm/xe/instructions/xe_gsc_commands.h
28
REG_FIELD_PREP(GSC_OPCODE, op) | \
drivers/gpu/drm/xe/instructions/xe_mfx_commands.h
16
#define XE_MFX_INSTR(subtype, op, sub_op) \
drivers/gpu/drm/xe/instructions/xe_mfx_commands.h
19
REG_FIELD_PREP(MFX_CMD_OPCODE, op) | \
drivers/gpu/drm/xe/xe_pt.c
1244
static int op_add_deps(struct xe_vm *vm, struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_pt.c
1254
switch (op->base.op) {
drivers/gpu/drm/xe/xe_pt.c
1256
if (!op->map.immediate && xe_vm_in_fault_mode(vm))
drivers/gpu/drm/xe/xe_pt.c
1259
err = vma_add_deps(op->map.vma, job);
drivers/gpu/drm/xe/xe_pt.c
1262
if (op->remap.prev)
drivers/gpu/drm/xe/xe_pt.c
1263
err = vma_add_deps(op->remap.prev, job);
drivers/gpu/drm/xe/xe_pt.c
1264
if (!err && op->remap.next)
drivers/gpu/drm/xe/xe_pt.c
1265
err = vma_add_deps(op->remap.next, job);
drivers/gpu/drm/xe/xe_pt.c
1270
err = vma_add_deps(gpuva_to_vma(op->base.prefetch.va), job);
drivers/gpu/drm/xe/xe_pt.c
1291
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_pt.c
1335
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_pt.c
1336
err = op_add_deps(vm, op, job);
drivers/gpu/drm/xe/xe_pt.c
1430
static int op_check_svm_userptr(struct xe_vm *vm, struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_pt.c
1437
switch (op->base.op) {
drivers/gpu/drm/xe/xe_pt.c
1439
if (!op->map.immediate && xe_vm_in_fault_mode(vm))
drivers/gpu/drm/xe/xe_pt.c
1442
err = vma_check_userptr(vm, op->map.vma, pt_update);
drivers/gpu/drm/xe/xe_pt.c
1445
if (op->remap.prev && !op->remap.skip_prev)
drivers/gpu/drm/xe/xe_pt.c
1446
err = vma_check_userptr(vm, op->remap.prev, pt_update);
drivers/gpu/drm/xe/xe_pt.c
1447
if (!err && op->remap.next && !op->remap.skip_next)
drivers/gpu/drm/xe/xe_pt.c
1448
err = vma_check_userptr(vm, op->remap.next, pt_update);
drivers/gpu/drm/xe/xe_pt.c
1453
if (xe_vma_is_cpu_addr_mirror(gpuva_to_vma(op->base.prefetch.va))) {
drivers/gpu/drm/xe/xe_pt.c
1454
struct xe_svm_range *range = op->map_range.range;
drivers/gpu/drm/xe/xe_pt.c
1458
xe_vma_is_cpu_addr_mirror(gpuva_to_vma(op->base.prefetch.va)));
drivers/gpu/drm/xe/xe_pt.c
1459
xa_for_each(&op->prefetch_range.range, i, range) {
drivers/gpu/drm/xe/xe_pt.c
1468
err = vma_check_userptr(vm, gpuva_to_vma(op->base.prefetch.va), pt_update);
drivers/gpu/drm/xe/xe_pt.c
1473
if (op->subop == XE_VMA_SUBOP_MAP_RANGE) {
drivers/gpu/drm/xe/xe_pt.c
1474
struct xe_svm_range *range = op->map_range.range;
drivers/gpu/drm/xe/xe_pt.c
1476
xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(op->map_range.vma));
drivers/gpu/drm/xe/xe_pt.c
1500
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_pt.c
1509
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_pt.c
1510
err = op_check_svm_userptr(vm, op, pt_update_ops);
drivers/gpu/drm/xe/xe_pt.c
2175
struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_pt.c
2181
switch (op->base.op) {
drivers/gpu/drm/xe/xe_pt.c
2183
if ((!op->map.immediate && xe_vm_in_fault_mode(vm) &&
drivers/gpu/drm/xe/xe_pt.c
2184
!op->map.invalidate_on_bind) ||
drivers/gpu/drm/xe/xe_pt.c
2185
(op->map.vma_flags & XE_VMA_SYSTEM_ALLOCATOR))
drivers/gpu/drm/xe/xe_pt.c
2188
err = bind_op_prepare(vm, tile, pt_update_ops, op->map.vma,
drivers/gpu/drm/xe/xe_pt.c
2189
op->map.invalidate_on_bind);
drivers/gpu/drm/xe/xe_pt.c
2194
struct xe_vma *old = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_pt.c
2201
if (!err && op->remap.prev && !op->remap.skip_prev) {
drivers/gpu/drm/xe/xe_pt.c
2203
op->remap.prev, false);
drivers/gpu/drm/xe/xe_pt.c
2206
if (!err && op->remap.next && !op->remap.skip_next) {
drivers/gpu/drm/xe/xe_pt.c
2208
op->remap.next, false);
drivers/gpu/drm/xe/xe_pt.c
2215
struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_pt.c
2225
struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_pt.c
2231
xa_for_each(&op->prefetch_range.range, i, range) {
drivers/gpu/drm/xe/xe_pt.c
2244
if (op->subop == XE_VMA_SUBOP_MAP_RANGE) {
drivers/gpu/drm/xe/xe_pt.c
2245
xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(op->map_range.vma));
drivers/gpu/drm/xe/xe_pt.c
2248
op->map_range.vma,
drivers/gpu/drm/xe/xe_pt.c
2249
op->map_range.range);
drivers/gpu/drm/xe/xe_pt.c
2250
} else if (op->subop == XE_VMA_SUBOP_UNMAP_RANGE) {
drivers/gpu/drm/xe/xe_pt.c
2252
op->unmap_range.range);
drivers/gpu/drm/xe/xe_pt.c
2286
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_pt.c
2300
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_pt.c
2301
err = op_prepare(vops->vm, tile, pt_update_ops, op);
drivers/gpu/drm/xe/xe_pt.c
2408
struct xe_vma_op *op, struct dma_fence *fence,
drivers/gpu/drm/xe/xe_pt.c
2413
switch (op->base.op) {
drivers/gpu/drm/xe/xe_pt.c
2415
if ((!op->map.immediate && xe_vm_in_fault_mode(vm)) ||
drivers/gpu/drm/xe/xe_pt.c
2416
(op->map.vma_flags & XE_VMA_SYSTEM_ALLOCATOR))
drivers/gpu/drm/xe/xe_pt.c
2419
bind_op_commit(vm, tile, pt_update_ops, op->map.vma, fence,
drivers/gpu/drm/xe/xe_pt.c
2420
fence2, op->map.invalidate_on_bind);
drivers/gpu/drm/xe/xe_pt.c
2424
struct xe_vma *old = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_pt.c
2431
if (op->remap.prev && !op->remap.skip_prev)
drivers/gpu/drm/xe/xe_pt.c
2432
bind_op_commit(vm, tile, pt_update_ops, op->remap.prev,
drivers/gpu/drm/xe/xe_pt.c
2434
if (op->remap.next && !op->remap.skip_next)
drivers/gpu/drm/xe/xe_pt.c
2435
bind_op_commit(vm, tile, pt_update_ops, op->remap.next,
drivers/gpu/drm/xe/xe_pt.c
2441
struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_pt.c
2450
struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_pt.c
2456
xa_for_each(&op->prefetch_range.range, i, range)
drivers/gpu/drm/xe/xe_pt.c
2467
if (op->subop == XE_VMA_SUBOP_MAP_RANGE)
drivers/gpu/drm/xe/xe_pt.c
2468
range_present_and_invalidated_tile(vm, op->map_range.range, tile->id);
drivers/gpu/drm/xe/xe_pt.c
2469
else if (op->subop == XE_VMA_SUBOP_UNMAP_RANGE)
drivers/gpu/drm/xe/xe_pt.c
2470
WRITE_ONCE(op->unmap_range.range->tile_present,
drivers/gpu/drm/xe/xe_pt.c
2471
op->unmap_range.range->tile_present &
drivers/gpu/drm/xe/xe_pt.c
2527
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_pt.c
2633
list_for_each_entry(op, &vops->list, link)
drivers/gpu/drm/xe/xe_pt.c
2634
op_commit(vops->vm, tile, pt_update_ops, op, fence, NULL);
drivers/gpu/drm/xe/xe_pt.c
2641
list_for_each_entry(op, &vops->list, link)
drivers/gpu/drm/xe/xe_pt.c
2642
op_commit(vops->vm, tile, pt_update_ops, op, ifence, NULL);
drivers/gpu/drm/xe/xe_pt.c
2654
list_for_each_entry(op, &vops->list, link)
drivers/gpu/drm/xe/xe_pt.c
2655
op_commit(vops->vm, tile, pt_update_ops, op, ifence,
drivers/gpu/drm/xe/xe_tlb_inval.c
254
#define xe_tlb_inval_issue(__tlb_inval, __fence, op, args...) \
drivers/gpu/drm/xe/xe_tlb_inval.c
263
__ret = op((__tlb_inval), (__fence)->seqno, ##args); \
drivers/gpu/drm/xe/xe_vm.c
1236
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
1238
op = kzalloc_obj(*op);
drivers/gpu/drm/xe/xe_vm.c
1240
if (unlikely(!op))
drivers/gpu/drm/xe/xe_vm.c
1243
return &op->base;
drivers/gpu/drm/xe/xe_vm.c
2174
static void print_op(struct xe_device *xe, struct drm_gpuva_op *op)
drivers/gpu/drm/xe/xe_vm.c
2178
switch (op->op) {
drivers/gpu/drm/xe/xe_vm.c
2181
(ULL)op->map.va.addr, (ULL)op->map.va.range);
drivers/gpu/drm/xe/xe_vm.c
2184
vma = gpuva_to_vma(op->remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
2187
op->remap.unmap->keep ? 1 : 0);
drivers/gpu/drm/xe/xe_vm.c
2188
if (op->remap.prev)
drivers/gpu/drm/xe/xe_vm.c
2191
(ULL)op->remap.prev->va.addr,
drivers/gpu/drm/xe/xe_vm.c
2192
(ULL)op->remap.prev->va.range);
drivers/gpu/drm/xe/xe_vm.c
2193
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
2196
(ULL)op->remap.next->va.addr,
drivers/gpu/drm/xe/xe_vm.c
2197
(ULL)op->remap.next->va.range);
drivers/gpu/drm/xe/xe_vm.c
2200
vma = gpuva_to_vma(op->unmap.va);
drivers/gpu/drm/xe/xe_vm.c
2203
op->unmap.keep ? 1 : 0);
drivers/gpu/drm/xe/xe_vm.c
2206
vma = gpuva_to_vma(op->prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
2215
static void print_op(struct xe_device *xe, struct drm_gpuva_op *op)
drivers/gpu/drm/xe/xe_vm.c
2239
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
2241
xe_vma_svm_prefetch_op_fini(op);
drivers/gpu/drm/xe/xe_vm.c
2321
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
2323
if (__op->op == DRM_GPUVA_OP_MAP) {
drivers/gpu/drm/xe/xe_vm.c
2324
op->map.immediate =
drivers/gpu/drm/xe/xe_vm.c
2327
op->map.vma_flags |= XE_VMA_READ_ONLY;
drivers/gpu/drm/xe/xe_vm.c
2329
op->map.vma_flags |= DRM_GPUVA_SPARSE;
drivers/gpu/drm/xe/xe_vm.c
2331
op->map.vma_flags |= XE_VMA_SYSTEM_ALLOCATOR;
drivers/gpu/drm/xe/xe_vm.c
2333
op->map.vma_flags |= XE_VMA_DUMPABLE;
drivers/gpu/drm/xe/xe_vm.c
2335
op->map.vma_flags |= XE_VMA_MADV_AUTORESET;
drivers/gpu/drm/xe/xe_vm.c
2336
op->map.pat_index = pat_index;
drivers/gpu/drm/xe/xe_vm.c
2337
op->map.invalidate_on_bind =
drivers/gpu/drm/xe/xe_vm.c
2339
} else if (__op->op == DRM_GPUVA_OP_PREFETCH) {
drivers/gpu/drm/xe/xe_vm.c
2340
struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
2349
op->prefetch.region = prefetch_region;
drivers/gpu/drm/xe/xe_vm.c
2360
xa_init_flags(&op->prefetch_range.range, XA_FLAGS_ALLOC);
drivers/gpu/drm/xe/xe_vm.c
2361
op->prefetch_range.ranges_count = 0;
drivers/gpu/drm/xe/xe_vm.c
2372
op->prefetch_range.dpagemap = dpagemap;
drivers/gpu/drm/xe/xe_vm.c
2396
err = xa_alloc(&op->prefetch_range.range,
drivers/gpu/drm/xe/xe_vm.c
2403
op->prefetch_range.ranges_count++;
drivers/gpu/drm/xe/xe_vm.c
2427
static struct xe_vma *new_vma(struct xe_vm *vm, struct drm_gpuva_op_map *op,
drivers/gpu/drm/xe/xe_vm.c
2430
struct xe_bo *bo = op->gem.obj ? gem_to_xe_bo(op->gem.obj) : NULL;
drivers/gpu/drm/xe/xe_vm.c
2453
vma = xe_vma_create(vm, bo, op->gem.offset,
drivers/gpu/drm/xe/xe_vm.c
2454
op->va.addr, op->va.addr +
drivers/gpu/drm/xe/xe_vm.c
2455
op->va.range - 1, attr, flags);
drivers/gpu/drm/xe/xe_vm.c
2470
vma = xe_vma_create(vm, NULL, op->gem.offset,
drivers/gpu/drm/xe/xe_vm.c
2471
op->va.addr, op->va.addr +
drivers/gpu/drm/xe/xe_vm.c
2472
op->va.range - 1, attr, flags);
drivers/gpu/drm/xe/xe_vm.c
2529
static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_vm.c
2535
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
2537
err |= xe_vm_insert_vma(vm, op->map.vma);
drivers/gpu/drm/xe/xe_vm.c
2539
op->flags |= XE_VMA_OP_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2544
gpuva_to_vma(op->base.remap.unmap->va)->tile_present;
drivers/gpu/drm/xe/xe_vm.c
2546
prep_vma_destroy(vm, gpuva_to_vma(op->base.remap.unmap->va),
drivers/gpu/drm/xe/xe_vm.c
2548
op->flags |= XE_VMA_OP_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2550
if (op->remap.prev) {
drivers/gpu/drm/xe/xe_vm.c
2551
err |= xe_vm_insert_vma(vm, op->remap.prev);
drivers/gpu/drm/xe/xe_vm.c
2553
op->flags |= XE_VMA_OP_PREV_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2554
if (!err && op->remap.skip_prev) {
drivers/gpu/drm/xe/xe_vm.c
2555
op->remap.prev->tile_present =
drivers/gpu/drm/xe/xe_vm.c
2559
if (op->remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2560
err |= xe_vm_insert_vma(vm, op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
2562
op->flags |= XE_VMA_OP_NEXT_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2563
if (!err && op->remap.skip_next) {
drivers/gpu/drm/xe/xe_vm.c
2564
op->remap.next->tile_present =
drivers/gpu/drm/xe/xe_vm.c
2574
op->base.remap.unmap->va->va.addr = op->remap.start;
drivers/gpu/drm/xe/xe_vm.c
2575
op->base.remap.unmap->va->va.range = op->remap.range;
drivers/gpu/drm/xe/xe_vm.c
2580
prep_vma_destroy(vm, gpuva_to_vma(op->base.unmap.va), true);
drivers/gpu/drm/xe/xe_vm.c
2581
op->flags |= XE_VMA_OP_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2584
op->flags |= XE_VMA_OP_COMMITTED;
drivers/gpu/drm/xe/xe_vm.c
2631
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
2635
INIT_LIST_HEAD(&op->link);
drivers/gpu/drm/xe/xe_vm.c
2636
list_add_tail(&op->link, &vops->list);
drivers/gpu/drm/xe/xe_vm.c
2637
op->tile_mask = tile_mask;
drivers/gpu/drm/xe/xe_vm.c
2639
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
2648
.default_pat_index = op->map.pat_index,
drivers/gpu/drm/xe/xe_vm.c
2649
.pat_index = op->map.pat_index,
drivers/gpu/drm/xe/xe_vm.c
2652
flags |= op->map.vma_flags & XE_VMA_CREATE_MASK;
drivers/gpu/drm/xe/xe_vm.c
2654
vma = new_vma(vm, &op->base.map, &default_attr,
drivers/gpu/drm/xe/xe_vm.c
2659
op->map.vma = vma;
drivers/gpu/drm/xe/xe_vm.c
2660
if (((op->map.immediate || !xe_vm_in_fault_mode(vm)) &&
drivers/gpu/drm/xe/xe_vm.c
2661
!(op->map.vma_flags & XE_VMA_SYSTEM_ALLOCATOR)) ||
drivers/gpu/drm/xe/xe_vm.c
2662
op->map.invalidate_on_bind)
drivers/gpu/drm/xe/xe_vm.c
2664
op->tile_mask, 1);
drivers/gpu/drm/xe/xe_vm.c
2670
gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
2675
if (op->base.remap.prev)
drivers/gpu/drm/xe/xe_vm.c
2676
start = op->base.remap.prev->va.addr +
drivers/gpu/drm/xe/xe_vm.c
2677
op->base.remap.prev->va.range;
drivers/gpu/drm/xe/xe_vm.c
2678
if (op->base.remap.next)
drivers/gpu/drm/xe/xe_vm.c
2679
end = op->base.remap.next->va.addr;
drivers/gpu/drm/xe/xe_vm.c
2689
op->remap.start = xe_vma_start(old);
drivers/gpu/drm/xe/xe_vm.c
2690
op->remap.range = xe_vma_size(old);
drivers/gpu/drm/xe/xe_vm.c
2691
op->remap.old_start = op->remap.start;
drivers/gpu/drm/xe/xe_vm.c
2692
op->remap.old_range = op->remap.range;
drivers/gpu/drm/xe/xe_vm.c
2694
flags |= op->base.remap.unmap->va->flags & XE_VMA_CREATE_MASK;
drivers/gpu/drm/xe/xe_vm.c
2695
if (op->base.remap.prev) {
drivers/gpu/drm/xe/xe_vm.c
2696
vma = new_vma(vm, op->base.remap.prev,
drivers/gpu/drm/xe/xe_vm.c
2701
op->remap.prev = vma;
drivers/gpu/drm/xe/xe_vm.c
2707
op->remap.skip_prev = skip ||
drivers/gpu/drm/xe/xe_vm.c
2711
if (op->remap.skip_prev) {
drivers/gpu/drm/xe/xe_vm.c
2713
op->remap.range -=
drivers/gpu/drm/xe/xe_vm.c
2716
op->remap.start = xe_vma_end(vma);
drivers/gpu/drm/xe/xe_vm.c
2718
(ULL)op->remap.start,
drivers/gpu/drm/xe/xe_vm.c
2719
(ULL)op->remap.range);
drivers/gpu/drm/xe/xe_vm.c
2725
if (op->base.remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2726
vma = new_vma(vm, op->base.remap.next,
drivers/gpu/drm/xe/xe_vm.c
2731
op->remap.next = vma;
drivers/gpu/drm/xe/xe_vm.c
2737
op->remap.skip_next = skip ||
drivers/gpu/drm/xe/xe_vm.c
2741
if (op->remap.skip_next) {
drivers/gpu/drm/xe/xe_vm.c
2743
op->remap.range -=
drivers/gpu/drm/xe/xe_vm.c
2747
(ULL)op->remap.start,
drivers/gpu/drm/xe/xe_vm.c
2748
(ULL)op->remap.range);
drivers/gpu/drm/xe/xe_vm.c
2756
xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask, num_remap_ops);
drivers/gpu/drm/xe/xe_vm.c
2760
vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_vm.c
2769
xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask, 1);
drivers/gpu/drm/xe/xe_vm.c
2772
vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
2781
xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask,
drivers/gpu/drm/xe/xe_vm.c
2782
op->prefetch_range.ranges_count);
drivers/gpu/drm/xe/xe_vm.c
2784
xe_vma_ops_incr_pt_update_ops(vops, op->tile_mask, 1);
drivers/gpu/drm/xe/xe_vm.c
2791
err = xe_vma_op_commit(vm, op);
drivers/gpu/drm/xe/xe_vm.c
2799
static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_vm.c
2805
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
2807
if (op->map.vma) {
drivers/gpu/drm/xe/xe_vm.c
2808
prep_vma_destroy(vm, op->map.vma, post_commit);
drivers/gpu/drm/xe/xe_vm.c
2809
xe_vma_destroy_unlocked(op->map.vma);
drivers/gpu/drm/xe/xe_vm.c
2814
struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_vm.c
2827
struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
2829
if (op->remap.prev) {
drivers/gpu/drm/xe/xe_vm.c
2830
prep_vma_destroy(vm, op->remap.prev, prev_post_commit);
drivers/gpu/drm/xe/xe_vm.c
2831
xe_vma_destroy_unlocked(op->remap.prev);
drivers/gpu/drm/xe/xe_vm.c
2833
if (op->remap.next) {
drivers/gpu/drm/xe/xe_vm.c
2834
prep_vma_destroy(vm, op->remap.next, next_post_commit);
drivers/gpu/drm/xe/xe_vm.c
2835
xe_vma_destroy_unlocked(op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
2848
op->base.remap.unmap->va->va.addr =
drivers/gpu/drm/xe/xe_vm.c
2849
op->remap.old_start;
drivers/gpu/drm/xe/xe_vm.c
2850
op->base.remap.unmap->va->va.range =
drivers/gpu/drm/xe/xe_vm.c
2851
op->remap.old_range;
drivers/gpu/drm/xe/xe_vm.c
2879
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
2881
xe_vma_op_unwind(vm, op,
drivers/gpu/drm/xe/xe_vm.c
2882
op->flags & XE_VMA_OP_COMMITTED,
drivers/gpu/drm/xe/xe_vm.c
2883
op->flags & XE_VMA_OP_PREV_COMMITTED,
drivers/gpu/drm/xe/xe_vm.c
2884
op->flags & XE_VMA_OP_NEXT_COMMITTED);
drivers/gpu/drm/xe/xe_vm.c
2923
static int prefetch_ranges(struct xe_vm *vm, struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_vm.c
2926
struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
2927
struct drm_pagemap *dpagemap = op->prefetch_range.dpagemap;
drivers/gpu/drm/xe/xe_vm.c
2943
xa_for_each(&op->prefetch_range.range, i, svm_range) {
drivers/gpu/drm/xe/xe_vm.c
2979
struct xe_vma_ops *vops, struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_vm.c
2991
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
2993
if (!op->map.invalidate_on_bind)
drivers/gpu/drm/xe/xe_vm.c
2994
err = vma_lock_and_validate(exec, op->map.vma,
drivers/gpu/drm/xe/xe_vm.c
2997
op->map.immediate);
drivers/gpu/drm/xe/xe_vm.c
3000
err = check_ufence(gpuva_to_vma(op->base.remap.unmap->va));
drivers/gpu/drm/xe/xe_vm.c
3005
gpuva_to_vma(op->base.remap.unmap->va),
drivers/gpu/drm/xe/xe_vm.c
3007
if (!err && op->remap.prev)
drivers/gpu/drm/xe/xe_vm.c
3008
err = vma_lock_and_validate(exec, op->remap.prev,
drivers/gpu/drm/xe/xe_vm.c
3010
if (!err && op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3011
err = vma_lock_and_validate(exec, op->remap.next,
drivers/gpu/drm/xe/xe_vm.c
3015
err = check_ufence(gpuva_to_vma(op->base.unmap.va));
drivers/gpu/drm/xe/xe_vm.c
3020
gpuva_to_vma(op->base.unmap.va),
drivers/gpu/drm/xe/xe_vm.c
3025
struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
3029
region = op->prefetch.region;
drivers/gpu/drm/xe/xe_vm.c
3035
gpuva_to_vma(op->base.prefetch.va),
drivers/gpu/drm/xe/xe_vm.c
3053
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
3059
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_vm.c
3060
if (op->base.op == DRM_GPUVA_OP_PREFETCH) {
drivers/gpu/drm/xe/xe_vm.c
3061
err = prefetch_ranges(vm, op);
drivers/gpu/drm/xe/xe_vm.c
3074
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
3081
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_vm.c
3082
err = op_lock_and_prep(exec, vm, vops, op);
drivers/gpu/drm/xe/xe_vm.c
3096
static void op_trace(struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_vm.c
3098
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
3100
trace_xe_vma_bind(op->map.vma);
drivers/gpu/drm/xe/xe_vm.c
3103
trace_xe_vma_unbind(gpuva_to_vma(op->base.remap.unmap->va));
drivers/gpu/drm/xe/xe_vm.c
3104
if (op->remap.prev)
drivers/gpu/drm/xe/xe_vm.c
3105
trace_xe_vma_bind(op->remap.prev);
drivers/gpu/drm/xe/xe_vm.c
3106
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3107
trace_xe_vma_bind(op->remap.next);
drivers/gpu/drm/xe/xe_vm.c
3110
trace_xe_vma_unbind(gpuva_to_vma(op->base.unmap.va));
drivers/gpu/drm/xe/xe_vm.c
3113
trace_xe_vma_bind(gpuva_to_vma(op->base.prefetch.va));
drivers/gpu/drm/xe/xe_vm.c
3124
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
3126
list_for_each_entry(op, &vops->list, link)
drivers/gpu/drm/xe/xe_vm.c
3127
op_trace(op);
drivers/gpu/drm/xe/xe_vm.c
3264
static void op_add_ufence(struct xe_vm *vm, struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_vm.c
3267
switch (op->base.op) {
drivers/gpu/drm/xe/xe_vm.c
3269
if (!xe_vma_is_cpu_addr_mirror(op->map.vma))
drivers/gpu/drm/xe/xe_vm.c
3270
vma_add_ufence(op->map.vma, ufence);
drivers/gpu/drm/xe/xe_vm.c
3273
if (op->remap.prev)
drivers/gpu/drm/xe/xe_vm.c
3274
vma_add_ufence(op->remap.prev, ufence);
drivers/gpu/drm/xe/xe_vm.c
3275
if (op->remap.next)
drivers/gpu/drm/xe/xe_vm.c
3276
vma_add_ufence(op->remap.next, ufence);
drivers/gpu/drm/xe/xe_vm.c
3281
vma_add_ufence(gpuva_to_vma(op->base.prefetch.va), ufence);
drivers/gpu/drm/xe/xe_vm.c
3292
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
3296
list_for_each_entry(op, &vops->list, link) {
drivers/gpu/drm/xe/xe_vm.c
3298
op_add_ufence(vm, op, ufence);
drivers/gpu/drm/xe/xe_vm.c
3300
if (op->base.op == DRM_GPUVA_OP_UNMAP)
drivers/gpu/drm/xe/xe_vm.c
3301
xe_vma_destroy(gpuva_to_vma(op->base.unmap.va), fence);
drivers/gpu/drm/xe/xe_vm.c
3302
else if (op->base.op == DRM_GPUVA_OP_REMAP)
drivers/gpu/drm/xe/xe_vm.c
3303
xe_vma_destroy(gpuva_to_vma(op->base.remap.unmap->va),
drivers/gpu/drm/xe/xe_vm.c
3410
u32 op = (*bind_ops)[i].op;
drivers/gpu/drm/xe/xe_vm.c
3448
if (XE_IOCTL_DBG(xe, op > DRM_XE_VM_BIND_OP_PREFETCH) ||
drivers/gpu/drm/xe/xe_vm.c
3453
XE_IOCTL_DBG(xe, op != DRM_XE_VM_BIND_OP_MAP &&
drivers/gpu/drm/xe/xe_vm.c
3456
op == DRM_XE_VM_BIND_OP_MAP &&
drivers/gpu/drm/xe/xe_vm.c
3459
op == DRM_XE_VM_BIND_OP_UNMAP_ALL) ||
drivers/gpu/drm/xe/xe_vm.c
3461
op == DRM_XE_VM_BIND_OP_UNMAP_ALL) ||
drivers/gpu/drm/xe/xe_vm.c
3463
op == DRM_XE_VM_BIND_OP_UNMAP_ALL) ||
drivers/gpu/drm/xe/xe_vm.c
3465
op == DRM_XE_VM_BIND_OP_MAP_USERPTR) ||
drivers/gpu/drm/xe/xe_vm.c
3467
op == DRM_XE_VM_BIND_OP_MAP_USERPTR) ||
drivers/gpu/drm/xe/xe_vm.c
3469
op == DRM_XE_VM_BIND_OP_MAP_USERPTR) ||
drivers/gpu/drm/xe/xe_vm.c
3470
XE_IOCTL_DBG(xe, op == DRM_XE_VM_BIND_OP_MAP_USERPTR &&
drivers/gpu/drm/xe/xe_vm.c
3473
op == DRM_XE_VM_BIND_OP_PREFETCH) ||
drivers/gpu/drm/xe/xe_vm.c
3475
op != DRM_XE_VM_BIND_OP_PREFETCH) ||
drivers/gpu/drm/xe/xe_vm.c
3481
op == DRM_XE_VM_BIND_OP_UNMAP) ||
drivers/gpu/drm/xe/xe_vm.c
3483
(!is_cpu_addr_mirror || op != DRM_XE_VM_BIND_OP_MAP))) {
drivers/gpu/drm/xe/xe_vm.c
3492
op != DRM_XE_VM_BIND_OP_UNMAP_ALL)) {
drivers/gpu/drm/xe/xe_vm.c
3545
u16 pat_index, u32 op, u32 bind_flags)
drivers/gpu/drm/xe/xe_vm.c
3604
op != DRM_XE_VM_BIND_OP_UNMAP && op != DRM_XE_VM_BIND_OP_UNMAP_ALL)
drivers/gpu/drm/xe/xe_vm.c
3701
u32 op = bind_ops[i].op;
drivers/gpu/drm/xe/xe_vm.c
3715
obj_offset, pat_index, op,
drivers/gpu/drm/xe/xe_vm.c
3764
u32 op = bind_ops[i].op;
drivers/gpu/drm/xe/xe_vm.c
3771
addr, range, op, flags,
drivers/gpu/drm/xe/xe_vm.c
4398
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
4402
if (__op->op == DRM_GPUVA_OP_UNMAP) {
drivers/gpu/drm/xe/xe_vm.c
4403
vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_vm.c
4409
if (__op->op == DRM_GPUVA_OP_REMAP) {
drivers/gpu/drm/xe/xe_vm.c
4410
vma = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
4415
if (__op->op == DRM_GPUVA_OP_MAP) {
drivers/gpu/drm/xe/xe_vm.c
4416
op->map.vma_flags |= vma_flags & XE_VMA_CREATE_MASK;
drivers/gpu/drm/xe/xe_vm.c
4417
op->map.pat_index = default_pat;
drivers/gpu/drm/xe/xe_vm.c
4420
if (__op->op == DRM_GPUVA_OP_REMAP) {
drivers/gpu/drm/xe/xe_vm.c
4421
vma = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
4428
if (__op->op == DRM_GPUVA_OP_MAP) {
drivers/gpu/drm/xe/xe_vm.c
4437
op->map.vma_flags |= vma_flags & XE_VMA_CREATE_MASK;
drivers/gpu/drm/xe/xe_vm.c
4457
struct xe_vma_op *op = gpuva_op_to_vma_op(__op);
drivers/gpu/drm/xe/xe_vm.c
4460
if (__op->op == DRM_GPUVA_OP_UNMAP) {
drivers/gpu/drm/xe/xe_vm.c
4461
vma = gpuva_to_vma(op->base.unmap.va);
drivers/gpu/drm/xe/xe_vm.c
4467
} else if (__op->op == DRM_GPUVA_OP_REMAP) {
drivers/gpu/drm/xe/xe_vm.c
4468
vma = gpuva_to_vma(op->base.remap.unmap->va);
drivers/gpu/drm/xe/xe_vm.c
4475
xe_vma_destroy(gpuva_to_vma(op->base.remap.unmap->va), NULL);
drivers/gpu/drm/xe/xe_vm.c
4476
} else if (__op->op == DRM_GPUVA_OP_MAP) {
drivers/gpu/drm/xe/xe_vm.c
4477
vma = op->map.vma;
drivers/gpu/drm/xe/xe_vm.c
600
static void xe_vma_svm_prefetch_op_fini(struct xe_vma_op *op)
drivers/gpu/drm/xe/xe_vm.c
604
vma = gpuva_to_vma(op->base.prefetch.va);
drivers/gpu/drm/xe/xe_vm.c
606
if (op->base.op == DRM_GPUVA_OP_PREFETCH && xe_vma_is_cpu_addr_mirror(vma))
drivers/gpu/drm/xe/xe_vm.c
607
xa_destroy(&op->prefetch_range.range);
drivers/gpu/drm/xe/xe_vm.c
612
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
617
list_for_each_entry(op, &vops->list, link)
drivers/gpu/drm/xe/xe_vm.c
618
xe_vma_svm_prefetch_op_fini(op);
drivers/gpu/drm/xe/xe_vm.c
650
static void xe_vm_populate_rebind(struct xe_vma_op *op, struct xe_vma *vma,
drivers/gpu/drm/xe/xe_vm.c
653
INIT_LIST_HEAD(&op->link);
drivers/gpu/drm/xe/xe_vm.c
654
op->tile_mask = tile_mask;
drivers/gpu/drm/xe/xe_vm.c
655
op->base.op = DRM_GPUVA_OP_MAP;
drivers/gpu/drm/xe/xe_vm.c
656
op->base.map.va.addr = vma->gpuva.va.addr;
drivers/gpu/drm/xe/xe_vm.c
657
op->base.map.va.range = vma->gpuva.va.range;
drivers/gpu/drm/xe/xe_vm.c
658
op->base.map.gem.obj = vma->gpuva.gem.obj;
drivers/gpu/drm/xe/xe_vm.c
659
op->base.map.gem.offset = vma->gpuva.gem.offset;
drivers/gpu/drm/xe/xe_vm.c
660
op->map.vma = vma;
drivers/gpu/drm/xe/xe_vm.c
661
op->map.immediate = true;
drivers/gpu/drm/xe/xe_vm.c
662
op->map.vma_flags = vma->gpuva.flags & XE_VMA_CREATE_MASK;
drivers/gpu/drm/xe/xe_vm.c
668
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
670
op = kzalloc_obj(*op);
drivers/gpu/drm/xe/xe_vm.c
671
if (!op)
drivers/gpu/drm/xe/xe_vm.c
674
xe_vm_populate_rebind(op, vma, tile_mask);
drivers/gpu/drm/xe/xe_vm.c
675
list_add_tail(&op->link, &vops->list);
drivers/gpu/drm/xe/xe_vm.c
692
struct xe_vma_op *op, *next_op;
drivers/gpu/drm/xe/xe_vm.c
733
list_for_each_entry_safe(op, next_op, &vops.list, link) {
drivers/gpu/drm/xe/xe_vm.c
734
list_del(&op->link);
drivers/gpu/drm/xe/xe_vm.c
735
kfree(op);
drivers/gpu/drm/xe/xe_vm.c
746
struct xe_vma_op *op, *next_op;
drivers/gpu/drm/xe/xe_vm.c
776
list_for_each_entry_safe(op, next_op, &vops.list, link) {
drivers/gpu/drm/xe/xe_vm.c
777
list_del(&op->link);
drivers/gpu/drm/xe/xe_vm.c
778
kfree(op);
drivers/gpu/drm/xe/xe_vm.c
785
static void xe_vm_populate_range_rebind(struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_vm.c
790
INIT_LIST_HEAD(&op->link);
drivers/gpu/drm/xe/xe_vm.c
791
op->tile_mask = tile_mask;
drivers/gpu/drm/xe/xe_vm.c
792
op->base.op = DRM_GPUVA_OP_DRIVER;
drivers/gpu/drm/xe/xe_vm.c
793
op->subop = XE_VMA_SUBOP_MAP_RANGE;
drivers/gpu/drm/xe/xe_vm.c
794
op->map_range.vma = vma;
drivers/gpu/drm/xe/xe_vm.c
795
op->map_range.range = range;
drivers/gpu/drm/xe/xe_vm.c
804
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
806
op = kzalloc_obj(*op);
drivers/gpu/drm/xe/xe_vm.c
807
if (!op)
drivers/gpu/drm/xe/xe_vm.c
810
xe_vm_populate_range_rebind(op, vma, range, tile_mask);
drivers/gpu/drm/xe/xe_vm.c
811
list_add_tail(&op->link, &vops->list);
drivers/gpu/drm/xe/xe_vm.c
836
struct xe_vma_op *op, *next_op;
drivers/gpu/drm/xe/xe_vm.c
867
list_for_each_entry_safe(op, next_op, &vops.list, link) {
drivers/gpu/drm/xe/xe_vm.c
868
list_del(&op->link);
drivers/gpu/drm/xe/xe_vm.c
869
kfree(op);
drivers/gpu/drm/xe/xe_vm.c
876
static void xe_vm_populate_range_unbind(struct xe_vma_op *op,
drivers/gpu/drm/xe/xe_vm.c
879
INIT_LIST_HEAD(&op->link);
drivers/gpu/drm/xe/xe_vm.c
880
op->tile_mask = range->tile_present;
drivers/gpu/drm/xe/xe_vm.c
881
op->base.op = DRM_GPUVA_OP_DRIVER;
drivers/gpu/drm/xe/xe_vm.c
882
op->subop = XE_VMA_SUBOP_UNMAP_RANGE;
drivers/gpu/drm/xe/xe_vm.c
883
op->unmap_range.range = range;
drivers/gpu/drm/xe/xe_vm.c
890
struct xe_vma_op *op;
drivers/gpu/drm/xe/xe_vm.c
892
op = kzalloc_obj(*op);
drivers/gpu/drm/xe/xe_vm.c
893
if (!op)
drivers/gpu/drm/xe/xe_vm.c
896
xe_vm_populate_range_unbind(op, range);
drivers/gpu/drm/xe/xe_vm.c
897
list_add_tail(&op->link, &vops->list);
drivers/gpu/drm/xe/xe_vm.c
918
struct xe_vma_op *op, *next_op;
drivers/gpu/drm/xe/xe_vm.c
950
list_for_each_entry_safe(op, next_op, &vops.list, link) {
drivers/gpu/drm/xe/xe_vm.c
951
list_del(&op->link);
drivers/gpu/drm/xe/xe_vm.c
952
kfree(op);
drivers/gpu/drm/xe/xe_vm.h
106
static inline struct xe_vma_op *gpuva_op_to_vma_op(struct drm_gpuva_op *op)
drivers/gpu/drm/xe/xe_vm.h
108
return container_of(op, struct xe_vma_op, base);
drivers/gpu/drm/xe/xe_vm_madvise.c
102
if ((loc->devmem_fd == op->preferred_mem_loc.devmem_fd &&
drivers/gpu/drm/xe/xe_vm_madvise.c
103
loc->migration_policy == op->preferred_mem_loc.migration_policy) ||
drivers/gpu/drm/xe/xe_vm_madvise.c
108
loc->devmem_fd = op->preferred_mem_loc.devmem_fd;
drivers/gpu/drm/xe/xe_vm_madvise.c
112
loc->migration_policy = op->preferred_mem_loc.migration_policy;
drivers/gpu/drm/xe/xe_vm_madvise.c
123
struct drm_xe_madvise *op,
drivers/gpu/drm/xe/xe_vm_madvise.c
129
xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_ATOMIC);
drivers/gpu/drm/xe/xe_vm_madvise.c
130
xe_assert(vm->xe, op->atomic.val <= DRM_XE_ATOMIC_CPU);
drivers/gpu/drm/xe/xe_vm_madvise.c
134
!(op->atomic.val == DRM_XE_ATOMIC_DEVICE &&
drivers/gpu/drm/xe/xe_vm_madvise.c
140
if (vmas[i]->attr.atomic_access == op->atomic.val) {
drivers/gpu/drm/xe/xe_vm_madvise.c
144
vmas[i]->attr.atomic_access = op->atomic.val;
drivers/gpu/drm/xe/xe_vm_madvise.c
148
if (!bo || bo->attr.atomic_access == op->atomic.val)
drivers/gpu/drm/xe/xe_vm_madvise.c
153
bo->attr.atomic_access = op->atomic.val;
drivers/gpu/drm/xe/xe_vm_madvise.c
165
struct drm_xe_madvise *op,
drivers/gpu/drm/xe/xe_vm_madvise.c
170
xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PAT);
drivers/gpu/drm/xe/xe_vm_madvise.c
173
if (vmas[i]->attr.pat_index == op->pat_index.val) {
drivers/gpu/drm/xe/xe_vm_madvise.c
177
vmas[i]->attr.pat_index = op->pat_index.val;
drivers/gpu/drm/xe/xe_vm_madvise.c
184
struct drm_xe_madvise *op,
drivers/gpu/drm/xe/xe_vm_madvise.c
90
struct drm_xe_madvise *op,
drivers/gpu/drm/xe/xe_vm_madvise.c
95
xe_assert(vm->xe, op->type == DRM_XE_MEM_RANGE_ATTR_PREFERRED_LOC);
drivers/gpu/drm/xe/xe_wait_user_fence.c
121
if (XE_IOCTL_DBG(xe, args->op > MAX_OP))
drivers/gpu/drm/xe/xe_wait_user_fence.c
139
err = do_compare(addr, args->value, args->mask, args->op);
drivers/gpu/drm/xe/xe_wait_user_fence.c
159
args->op);
drivers/gpu/drm/xe/xe_wait_user_fence.c
17
static int do_compare(u64 addr, u64 value, u64 mask, u16 op)
drivers/gpu/drm/xe/xe_wait_user_fence.c
27
switch (op) {
drivers/gpu/drm/xen/xen_drm_front.c
141
req->op.set_config.x = x;
drivers/gpu/drm/xen/xen_drm_front.c
142
req->op.set_config.y = y;
drivers/gpu/drm/xen/xen_drm_front.c
143
req->op.set_config.width = width;
drivers/gpu/drm/xen/xen_drm_front.c
144
req->op.set_config.height = height;
drivers/gpu/drm/xen/xen_drm_front.c
145
req->op.set_config.bpp = bpp;
drivers/gpu/drm/xen/xen_drm_front.c
146
req->op.set_config.fb_cookie = fb_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
195
req->op.dbuf_create.gref_directory =
drivers/gpu/drm/xen/xen_drm_front.c
197
req->op.dbuf_create.buffer_sz = size;
drivers/gpu/drm/xen/xen_drm_front.c
198
req->op.dbuf_create.data_ofs = offset;
drivers/gpu/drm/xen/xen_drm_front.c
199
req->op.dbuf_create.dbuf_cookie = dbuf_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
200
req->op.dbuf_create.width = width;
drivers/gpu/drm/xen/xen_drm_front.c
201
req->op.dbuf_create.height = height;
drivers/gpu/drm/xen/xen_drm_front.c
202
req->op.dbuf_create.bpp = bpp;
drivers/gpu/drm/xen/xen_drm_front.c
204
req->op.dbuf_create.flags |= XENDISPL_DBUF_FLG_REQ_ALLOC;
drivers/gpu/drm/xen/xen_drm_front.c
256
req->op.dbuf_destroy.dbuf_cookie = dbuf_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
299
req->op.fb_attach.dbuf_cookie = dbuf_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
300
req->op.fb_attach.fb_cookie = fb_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
301
req->op.fb_attach.width = width;
drivers/gpu/drm/xen/xen_drm_front.c
302
req->op.fb_attach.height = height;
drivers/gpu/drm/xen/xen_drm_front.c
303
req->op.fb_attach.pixel_format = pixel_format;
drivers/gpu/drm/xen/xen_drm_front.c
331
req->op.fb_detach.fb_cookie = fb_cookie;
drivers/gpu/drm/xen/xen_drm_front.c
360
req->op.pg_flip.fb_cookie = fb_cookie;
drivers/gpu/drm/xen/xen_drm_front_evtchnl.c
110
event->op.pg_flip.fb_cookie);
drivers/greybus/svc.c
1144
static int gb_svc_intf_reset_recv(struct gb_operation *op)
drivers/greybus/svc.c
1146
struct gb_svc *svc = gb_connection_get_data(op->connection);
drivers/greybus/svc.c
1147
struct gb_message *request = op->request;
drivers/greybus/svc.c
1162
static int gb_svc_module_inserted_recv(struct gb_operation *op)
drivers/greybus/svc.c
1164
struct gb_svc *svc = gb_connection_get_data(op->connection);
drivers/greybus/svc.c
1167
if (op->request->payload_size < sizeof(*request)) {
drivers/greybus/svc.c
1169
op->request->payload_size, sizeof(*request));
drivers/greybus/svc.c
1173
request = op->request->payload;
drivers/greybus/svc.c
1178
return gb_svc_queue_deferred_request(op);
drivers/greybus/svc.c
1181
static int gb_svc_module_removed_recv(struct gb_operation *op)
drivers/greybus/svc.c
1183
struct gb_svc *svc = gb_connection_get_data(op->connection);
drivers/greybus/svc.c
1186
if (op->request->payload_size < sizeof(*request)) {
drivers/greybus/svc.c
1188
op->request->payload_size, sizeof(*request));
drivers/greybus/svc.c
1192
request = op->request->payload;
drivers/greybus/svc.c
1197
return gb_svc_queue_deferred_request(op);
drivers/greybus/svc.c
1200
static int gb_svc_intf_oops_recv(struct gb_operation *op)
drivers/greybus/svc.c
1202
struct gb_svc *svc = gb_connection_get_data(op->connection);
drivers/greybus/svc.c
1205
if (op->request->payload_size < sizeof(*request)) {
drivers/greybus/svc.c
1207
op->request->payload_size, sizeof(*request));
drivers/greybus/svc.c
1211
return gb_svc_queue_deferred_request(op);
drivers/greybus/svc.c
1214
static int gb_svc_intf_mailbox_event_recv(struct gb_operation *op)
drivers/greybus/svc.c
1216
struct gb_svc *svc = gb_connection_get_data(op->connection);
drivers/greybus/svc.c
1219
if (op->request->payload_size < sizeof(*request)) {
drivers/greybus/svc.c
1221
op->request->payload_size, sizeof(*request));
drivers/greybus/svc.c
1225
request = op->request->payload;
drivers/greybus/svc.c
1229
return gb_svc_queue_deferred_request(op);
drivers/greybus/svc.c
1232
static int gb_svc_request_handler(struct gb_operation *op)
drivers/greybus/svc.c
1234
struct gb_connection *connection = op->connection;
drivers/greybus/svc.c
1236
u8 type = op->type;
drivers/greybus/svc.c
1272
ret = gb_svc_version_request(op);
drivers/greybus/svc.c
1277
ret = gb_svc_hello(op);
drivers/greybus/svc.c
1282
return gb_svc_intf_reset_recv(op);
drivers/greybus/svc.c
1284
return gb_svc_module_inserted_recv(op);
drivers/greybus/svc.c
1286
return gb_svc_module_removed_recv(op);
drivers/greybus/svc.c
1288
return gb_svc_intf_mailbox_event_recv(op);
drivers/greybus/svc.c
1290
return gb_svc_intf_oops_recv(op);
drivers/greybus/svc.c
642
static int gb_svc_version_request(struct gb_operation *op)
drivers/greybus/svc.c
644
struct gb_connection *connection = op->connection;
drivers/greybus/svc.c
649
if (op->request->payload_size < sizeof(*request)) {
drivers/greybus/svc.c
651
op->request->payload_size,
drivers/greybus/svc.c
656
request = op->request->payload;
drivers/greybus/svc.c
667
if (!gb_operation_response_alloc(op, sizeof(*response), GFP_KERNEL))
drivers/greybus/svc.c
670
response = op->response->payload;
drivers/greybus/svc.c
838
static int gb_svc_hello(struct gb_operation *op)
drivers/greybus/svc.c
840
struct gb_connection *connection = op->connection;
drivers/greybus/svc.c
845
if (op->request->payload_size < sizeof(*hello_request)) {
drivers/greybus/svc.c
847
op->request->payload_size,
drivers/greybus/svc.c
852
hello_request = op->request->payload;
drivers/greybus/svc.c
872
ret = gb_svc_queue_deferred_request(op);
drivers/hid/amd-sfh-hid/sfh1_1/amd_sfh_interface.c
161
int amd_get_sfh_info(struct amd_sfh_info *sfh_info, enum sfh_message_type op)
drivers/hid/amd-sfh-hid/sfh1_1/amd_sfh_interface.c
164
switch (op) {
drivers/hv/hv_kvp.c
261
static int process_ob_ipinfo(void *in_msg, void *out_msg, int op)
drivers/hv/hv_kvp.c
267
switch (op) {
drivers/hv/hv_kvp.c
321
static void process_ib_ipinfo(void *in_msg, void *out_msg, int op)
drivers/hv/hv_kvp.c
326
switch (op) {
drivers/hv/hv_snapshot.c
177
int op = vss_transaction.msg->vss_hdr.operation;
drivers/hv/hv_snapshot.c
191
vss_msg->vss_hdr.operation = op;
drivers/hv/hv_snapshot.c
195
schedule_delayed_work(&vss_timeout_work, op == VSS_OP_FREEZE ?
drivers/hwmon/drivetemp.c
165
enum req_op op;
drivers/hwmon/drivetemp.c
177
op = REQ_OP_DRV_OUT;
drivers/hwmon/drivetemp.c
185
op = REQ_OP_DRV_IN;
drivers/hwmon/drivetemp.c
194
err = scsi_execute_cmd(st->sdev, scsi_cmd, op, st->smartdata,
drivers/hwmon/pmbus/pmbus_core.c
192
static void pmbus_update_ts(struct i2c_client *client, int op)
drivers/hwmon/pmbus/pmbus_core.c
198
if (op & PMBUS_OP_WRITE)
drivers/hwmon/pmbus/pmbus_core.c
200
if (op & PMBUS_OP_PAGE_CHANGE)
drivers/hwmon/ultra45_env.c
255
static int env_probe(struct platform_device *op)
drivers/hwmon/ultra45_env.c
257
struct env *p = devm_kzalloc(&op->dev, sizeof(*p), GFP_KERNEL);
drivers/hwmon/ultra45_env.c
265
p->regs = of_ioremap(&op->resource[0], 0, REG_SIZE, "pic16f747");
drivers/hwmon/ultra45_env.c
269
err = sysfs_create_group(&op->dev.kobj, &env_group);
drivers/hwmon/ultra45_env.c
273
p->hwmon_dev = hwmon_device_register(&op->dev);
drivers/hwmon/ultra45_env.c
279
platform_set_drvdata(op, p);
drivers/hwmon/ultra45_env.c
286
sysfs_remove_group(&op->dev.kobj, &env_group);
drivers/hwmon/ultra45_env.c
289
of_iounmap(&op->resource[0], p->regs, REG_SIZE);
drivers/hwmon/ultra45_env.c
294
static void env_remove(struct platform_device *op)
drivers/hwmon/ultra45_env.c
296
struct env *p = platform_get_drvdata(op);
drivers/hwmon/ultra45_env.c
299
sysfs_remove_group(&op->dev.kobj, &env_group);
drivers/hwmon/ultra45_env.c
301
of_iounmap(&op->resource[0], p->regs, REG_SIZE);
drivers/hwtracing/coresight/coresight-cti-core.c
336
int cti_channel_trig_op(struct device *dev, enum cti_chan_op op,
drivers/hwtracing/coresight/coresight-cti-core.c
377
if (op == CTI_CHAN_ATTACH)
drivers/hwtracing/coresight/coresight-cti-core.c
395
int cti_channel_gate_op(struct device *dev, enum cti_chan_gate_op op,
drivers/hwtracing/coresight/coresight-cti-core.c
411
switch (op) {
drivers/hwtracing/coresight/coresight-cti-core.c
433
int cti_channel_setop(struct device *dev, enum cti_chan_set_op op,
drivers/hwtracing/coresight/coresight-cti-core.c
450
switch (op) {
drivers/hwtracing/coresight/coresight-cti-sysfs.c
535
cti_trig_op_parse(struct device *dev, enum cti_chan_op op,
drivers/hwtracing/coresight/coresight-cti-sysfs.c
545
err = cti_channel_trig_op(dev, op, dir, chan_idx, trig_idx);
drivers/hwtracing/coresight/coresight-cti-sysfs.c
635
chan_op_parse(struct device *dev, enum cti_chan_set_op op, const char *buf)
drivers/hwtracing/coresight/coresight-cti-sysfs.c
642
err = cti_channel_setop(dev, op, channel);
drivers/hwtracing/coresight/coresight-cti.h
225
int cti_channel_trig_op(struct device *dev, enum cti_chan_op op,
drivers/hwtracing/coresight/coresight-cti.h
228
int cti_channel_gate_op(struct device *dev, enum cti_chan_gate_op op,
drivers/hwtracing/coresight/coresight-cti.h
230
int cti_channel_setop(struct device *dev, enum cti_chan_set_op op,
drivers/hwtracing/coresight/coresight-etm4x.h
306
#define ETE_ONLY_SYSREG_LIST(op, val) \
drivers/hwtracing/coresight/coresight-etm4x.h
307
CASE_##op((val), TRCRSR) \
drivers/hwtracing/coresight/coresight-etm4x.h
308
CASE_##op((val), TRCEXTINSELRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
309
CASE_##op((val), TRCEXTINSELRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
310
CASE_##op((val), TRCEXTINSELRn(3))
drivers/hwtracing/coresight/coresight-etm4x.h
313
#define ETM4x_ONLY_SYSREG_LIST(op, val) \
drivers/hwtracing/coresight/coresight-etm4x.h
314
CASE_##op((val), TRCPROCSELR) \
drivers/hwtracing/coresight/coresight-etm4x.h
315
CASE_##op((val), TRCOSLAR)
drivers/hwtracing/coresight/coresight-etm4x.h
317
#define ETM_COMMON_SYSREG_LIST(op, val) \
drivers/hwtracing/coresight/coresight-etm4x.h
318
CASE_##op((val), TRCPRGCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
319
CASE_##op((val), TRCSTATR) \
drivers/hwtracing/coresight/coresight-etm4x.h
320
CASE_##op((val), TRCCONFIGR) \
drivers/hwtracing/coresight/coresight-etm4x.h
321
CASE_##op((val), TRCAUXCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
322
CASE_##op((val), TRCEVENTCTL0R) \
drivers/hwtracing/coresight/coresight-etm4x.h
323
CASE_##op((val), TRCEVENTCTL1R) \
drivers/hwtracing/coresight/coresight-etm4x.h
324
CASE_##op((val), TRCSTALLCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
325
CASE_##op((val), TRCTSCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
326
CASE_##op((val), TRCSYNCPR) \
drivers/hwtracing/coresight/coresight-etm4x.h
327
CASE_##op((val), TRCCCCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
328
CASE_##op((val), TRCBBCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
329
CASE_##op((val), TRCTRACEIDR) \
drivers/hwtracing/coresight/coresight-etm4x.h
330
CASE_##op((val), TRCQCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
331
CASE_##op((val), TRCVICTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
332
CASE_##op((val), TRCVIIECTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
333
CASE_##op((val), TRCVISSCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
334
CASE_##op((val), TRCVIPCSSCTLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
335
CASE_##op((val), TRCSEQEVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
336
CASE_##op((val), TRCSEQEVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
337
CASE_##op((val), TRCSEQEVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
338
CASE_##op((val), TRCSEQRSTEVR) \
drivers/hwtracing/coresight/coresight-etm4x.h
339
CASE_##op((val), TRCSEQSTR) \
drivers/hwtracing/coresight/coresight-etm4x.h
340
CASE_##op((val), TRCEXTINSELR) \
drivers/hwtracing/coresight/coresight-etm4x.h
341
CASE_##op((val), TRCCNTRLDVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
342
CASE_##op((val), TRCCNTRLDVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
343
CASE_##op((val), TRCCNTRLDVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
344
CASE_##op((val), TRCCNTRLDVRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
345
CASE_##op((val), TRCCNTCTLRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
346
CASE_##op((val), TRCCNTCTLRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
347
CASE_##op((val), TRCCNTCTLRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
348
CASE_##op((val), TRCCNTCTLRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
349
CASE_##op((val), TRCCNTVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
350
CASE_##op((val), TRCCNTVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
351
CASE_##op((val), TRCCNTVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
352
CASE_##op((val), TRCCNTVRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
353
CASE_##op((val), TRCIDR8) \
drivers/hwtracing/coresight/coresight-etm4x.h
354
CASE_##op((val), TRCIDR9) \
drivers/hwtracing/coresight/coresight-etm4x.h
355
CASE_##op((val), TRCIDR10) \
drivers/hwtracing/coresight/coresight-etm4x.h
356
CASE_##op((val), TRCIDR11) \
drivers/hwtracing/coresight/coresight-etm4x.h
357
CASE_##op((val), TRCIDR12) \
drivers/hwtracing/coresight/coresight-etm4x.h
358
CASE_##op((val), TRCIDR13) \
drivers/hwtracing/coresight/coresight-etm4x.h
359
CASE_##op((val), TRCIMSPECn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
360
CASE_##op((val), TRCIMSPECn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
361
CASE_##op((val), TRCIMSPECn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
362
CASE_##op((val), TRCIMSPECn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
363
CASE_##op((val), TRCIMSPECn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
364
CASE_##op((val), TRCIMSPECn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
365
CASE_##op((val), TRCIMSPECn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
366
CASE_##op((val), TRCIMSPECn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
367
CASE_##op((val), TRCIDR0) \
drivers/hwtracing/coresight/coresight-etm4x.h
368
CASE_##op((val), TRCIDR1) \
drivers/hwtracing/coresight/coresight-etm4x.h
369
CASE_##op((val), TRCIDR2) \
drivers/hwtracing/coresight/coresight-etm4x.h
370
CASE_##op((val), TRCIDR3) \
drivers/hwtracing/coresight/coresight-etm4x.h
371
CASE_##op((val), TRCIDR4) \
drivers/hwtracing/coresight/coresight-etm4x.h
372
CASE_##op((val), TRCIDR5) \
drivers/hwtracing/coresight/coresight-etm4x.h
373
CASE_##op((val), TRCIDR6) \
drivers/hwtracing/coresight/coresight-etm4x.h
374
CASE_##op((val), TRCIDR7) \
drivers/hwtracing/coresight/coresight-etm4x.h
375
CASE_##op((val), TRCRSCTLRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
376
CASE_##op((val), TRCRSCTLRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
377
CASE_##op((val), TRCRSCTLRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
378
CASE_##op((val), TRCRSCTLRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
379
CASE_##op((val), TRCRSCTLRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
380
CASE_##op((val), TRCRSCTLRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
381
CASE_##op((val), TRCRSCTLRn(8)) \
drivers/hwtracing/coresight/coresight-etm4x.h
382
CASE_##op((val), TRCRSCTLRn(9)) \
drivers/hwtracing/coresight/coresight-etm4x.h
383
CASE_##op((val), TRCRSCTLRn(10)) \
drivers/hwtracing/coresight/coresight-etm4x.h
384
CASE_##op((val), TRCRSCTLRn(11)) \
drivers/hwtracing/coresight/coresight-etm4x.h
385
CASE_##op((val), TRCRSCTLRn(12)) \
drivers/hwtracing/coresight/coresight-etm4x.h
386
CASE_##op((val), TRCRSCTLRn(13)) \
drivers/hwtracing/coresight/coresight-etm4x.h
387
CASE_##op((val), TRCRSCTLRn(14)) \
drivers/hwtracing/coresight/coresight-etm4x.h
388
CASE_##op((val), TRCRSCTLRn(15)) \
drivers/hwtracing/coresight/coresight-etm4x.h
389
CASE_##op((val), TRCRSCTLRn(16)) \
drivers/hwtracing/coresight/coresight-etm4x.h
390
CASE_##op((val), TRCRSCTLRn(17)) \
drivers/hwtracing/coresight/coresight-etm4x.h
391
CASE_##op((val), TRCRSCTLRn(18)) \
drivers/hwtracing/coresight/coresight-etm4x.h
392
CASE_##op((val), TRCRSCTLRn(19)) \
drivers/hwtracing/coresight/coresight-etm4x.h
393
CASE_##op((val), TRCRSCTLRn(20)) \
drivers/hwtracing/coresight/coresight-etm4x.h
394
CASE_##op((val), TRCRSCTLRn(21)) \
drivers/hwtracing/coresight/coresight-etm4x.h
395
CASE_##op((val), TRCRSCTLRn(22)) \
drivers/hwtracing/coresight/coresight-etm4x.h
396
CASE_##op((val), TRCRSCTLRn(23)) \
drivers/hwtracing/coresight/coresight-etm4x.h
397
CASE_##op((val), TRCRSCTLRn(24)) \
drivers/hwtracing/coresight/coresight-etm4x.h
398
CASE_##op((val), TRCRSCTLRn(25)) \
drivers/hwtracing/coresight/coresight-etm4x.h
399
CASE_##op((val), TRCRSCTLRn(26)) \
drivers/hwtracing/coresight/coresight-etm4x.h
400
CASE_##op((val), TRCRSCTLRn(27)) \
drivers/hwtracing/coresight/coresight-etm4x.h
401
CASE_##op((val), TRCRSCTLRn(28)) \
drivers/hwtracing/coresight/coresight-etm4x.h
402
CASE_##op((val), TRCRSCTLRn(29)) \
drivers/hwtracing/coresight/coresight-etm4x.h
403
CASE_##op((val), TRCRSCTLRn(30)) \
drivers/hwtracing/coresight/coresight-etm4x.h
404
CASE_##op((val), TRCRSCTLRn(31)) \
drivers/hwtracing/coresight/coresight-etm4x.h
405
CASE_##op((val), TRCSSCCRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
406
CASE_##op((val), TRCSSCCRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
407
CASE_##op((val), TRCSSCCRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
408
CASE_##op((val), TRCSSCCRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
409
CASE_##op((val), TRCSSCCRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
410
CASE_##op((val), TRCSSCCRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
411
CASE_##op((val), TRCSSCCRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
412
CASE_##op((val), TRCSSCCRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
413
CASE_##op((val), TRCSSCSRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
414
CASE_##op((val), TRCSSCSRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
415
CASE_##op((val), TRCSSCSRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
416
CASE_##op((val), TRCSSCSRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
417
CASE_##op((val), TRCSSCSRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
418
CASE_##op((val), TRCSSCSRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
419
CASE_##op((val), TRCSSCSRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
420
CASE_##op((val), TRCSSCSRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
421
CASE_##op((val), TRCSSPCICRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
422
CASE_##op((val), TRCSSPCICRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
423
CASE_##op((val), TRCSSPCICRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
424
CASE_##op((val), TRCSSPCICRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
425
CASE_##op((val), TRCSSPCICRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
426
CASE_##op((val), TRCSSPCICRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
427
CASE_##op((val), TRCSSPCICRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
428
CASE_##op((val), TRCSSPCICRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
429
CASE_##op((val), TRCOSLSR) \
drivers/hwtracing/coresight/coresight-etm4x.h
430
CASE_##op((val), TRCACVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
431
CASE_##op((val), TRCACVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
432
CASE_##op((val), TRCACVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
433
CASE_##op((val), TRCACVRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
434
CASE_##op((val), TRCACVRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
435
CASE_##op((val), TRCACVRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
436
CASE_##op((val), TRCACVRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
437
CASE_##op((val), TRCACVRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
438
CASE_##op((val), TRCACVRn(8)) \
drivers/hwtracing/coresight/coresight-etm4x.h
439
CASE_##op((val), TRCACVRn(9)) \
drivers/hwtracing/coresight/coresight-etm4x.h
440
CASE_##op((val), TRCACVRn(10)) \
drivers/hwtracing/coresight/coresight-etm4x.h
441
CASE_##op((val), TRCACVRn(11)) \
drivers/hwtracing/coresight/coresight-etm4x.h
442
CASE_##op((val), TRCACVRn(12)) \
drivers/hwtracing/coresight/coresight-etm4x.h
443
CASE_##op((val), TRCACVRn(13)) \
drivers/hwtracing/coresight/coresight-etm4x.h
444
CASE_##op((val), TRCACVRn(14)) \
drivers/hwtracing/coresight/coresight-etm4x.h
445
CASE_##op((val), TRCACVRn(15)) \
drivers/hwtracing/coresight/coresight-etm4x.h
446
CASE_##op((val), TRCACATRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
447
CASE_##op((val), TRCACATRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
448
CASE_##op((val), TRCACATRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
449
CASE_##op((val), TRCACATRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
450
CASE_##op((val), TRCACATRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
451
CASE_##op((val), TRCACATRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
452
CASE_##op((val), TRCACATRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
453
CASE_##op((val), TRCACATRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
454
CASE_##op((val), TRCACATRn(8)) \
drivers/hwtracing/coresight/coresight-etm4x.h
455
CASE_##op((val), TRCACATRn(9)) \
drivers/hwtracing/coresight/coresight-etm4x.h
456
CASE_##op((val), TRCACATRn(10)) \
drivers/hwtracing/coresight/coresight-etm4x.h
457
CASE_##op((val), TRCACATRn(11)) \
drivers/hwtracing/coresight/coresight-etm4x.h
458
CASE_##op((val), TRCACATRn(12)) \
drivers/hwtracing/coresight/coresight-etm4x.h
459
CASE_##op((val), TRCACATRn(13)) \
drivers/hwtracing/coresight/coresight-etm4x.h
460
CASE_##op((val), TRCACATRn(14)) \
drivers/hwtracing/coresight/coresight-etm4x.h
461
CASE_##op((val), TRCACATRn(15)) \
drivers/hwtracing/coresight/coresight-etm4x.h
462
CASE_##op((val), TRCCIDCVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
463
CASE_##op((val), TRCCIDCVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
464
CASE_##op((val), TRCCIDCVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
465
CASE_##op((val), TRCCIDCVRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
466
CASE_##op((val), TRCCIDCVRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
467
CASE_##op((val), TRCCIDCVRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
468
CASE_##op((val), TRCCIDCVRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
469
CASE_##op((val), TRCCIDCVRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
470
CASE_##op((val), TRCVMIDCVRn(0)) \
drivers/hwtracing/coresight/coresight-etm4x.h
471
CASE_##op((val), TRCVMIDCVRn(1)) \
drivers/hwtracing/coresight/coresight-etm4x.h
472
CASE_##op((val), TRCVMIDCVRn(2)) \
drivers/hwtracing/coresight/coresight-etm4x.h
473
CASE_##op((val), TRCVMIDCVRn(3)) \
drivers/hwtracing/coresight/coresight-etm4x.h
474
CASE_##op((val), TRCVMIDCVRn(4)) \
drivers/hwtracing/coresight/coresight-etm4x.h
475
CASE_##op((val), TRCVMIDCVRn(5)) \
drivers/hwtracing/coresight/coresight-etm4x.h
476
CASE_##op((val), TRCVMIDCVRn(6)) \
drivers/hwtracing/coresight/coresight-etm4x.h
477
CASE_##op((val), TRCVMIDCVRn(7)) \
drivers/hwtracing/coresight/coresight-etm4x.h
478
CASE_##op((val), TRCCIDCCTLR0) \
drivers/hwtracing/coresight/coresight-etm4x.h
479
CASE_##op((val), TRCCIDCCTLR1) \
drivers/hwtracing/coresight/coresight-etm4x.h
480
CASE_##op((val), TRCVMIDCCTLR0) \
drivers/hwtracing/coresight/coresight-etm4x.h
481
CASE_##op((val), TRCVMIDCCTLR1) \
drivers/hwtracing/coresight/coresight-etm4x.h
482
CASE_##op((val), TRCCLAIMSET) \
drivers/hwtracing/coresight/coresight-etm4x.h
483
CASE_##op((val), TRCCLAIMCLR) \
drivers/hwtracing/coresight/coresight-etm4x.h
484
CASE_##op((val), TRCAUTHSTATUS) \
drivers/hwtracing/coresight/coresight-etm4x.h
485
CASE_##op((val), TRCDEVARCH) \
drivers/hwtracing/coresight/coresight-etm4x.h
486
CASE_##op((val), TRCDEVID)
drivers/hwtracing/coresight/coresight-etm4x.h
489
#define ETM_MMAP_LIST(op, val) \
drivers/hwtracing/coresight/coresight-etm4x.h
490
CASE_##op((val), TRCDEVTYPE) \
drivers/hwtracing/coresight/coresight-etm4x.h
491
CASE_##op((val), TRCPDCR) \
drivers/hwtracing/coresight/coresight-etm4x.h
492
CASE_##op((val), TRCPDSR) \
drivers/hwtracing/coresight/coresight-etm4x.h
493
CASE_##op((val), TRCDEVAFF0) \
drivers/hwtracing/coresight/coresight-etm4x.h
494
CASE_##op((val), TRCDEVAFF1) \
drivers/hwtracing/coresight/coresight-etm4x.h
495
CASE_##op((val), TRCLAR) \
drivers/hwtracing/coresight/coresight-etm4x.h
496
CASE_##op((val), TRCLSR) \
drivers/hwtracing/coresight/coresight-etm4x.h
497
CASE_##op((val), TRCITCTRL) \
drivers/hwtracing/coresight/coresight-etm4x.h
498
CASE_##op((val), TRCPIDR4) \
drivers/hwtracing/coresight/coresight-etm4x.h
499
CASE_##op((val), TRCPIDR0) \
drivers/hwtracing/coresight/coresight-etm4x.h
500
CASE_##op((val), TRCPIDR1) \
drivers/hwtracing/coresight/coresight-etm4x.h
501
CASE_##op((val), TRCPIDR2) \
drivers/hwtracing/coresight/coresight-etm4x.h
502
CASE_##op((val), TRCPIDR3)
drivers/hwtracing/stm/p_sys-t.c
266
static inline bool sys_t_need_ts(struct sys_t_output *op)
drivers/hwtracing/stm/p_sys-t.c
268
if (op->node.ts_interval &&
drivers/hwtracing/stm/p_sys-t.c
269
time_after(jiffies, op->ts_jiffies + op->node.ts_interval)) {
drivers/hwtracing/stm/p_sys-t.c
270
op->ts_jiffies = jiffies;
drivers/hwtracing/stm/p_sys-t.c
278
static bool sys_t_need_clock_sync(struct sys_t_output *op)
drivers/hwtracing/stm/p_sys-t.c
280
if (op->node.clocksync_interval &&
drivers/hwtracing/stm/p_sys-t.c
282
op->clocksync_jiffies + op->node.clocksync_interval)) {
drivers/hwtracing/stm/p_sys-t.c
283
op->clocksync_jiffies = jiffies;
drivers/hwtracing/stm/p_sys-t.c
372
struct sys_t_output *op = output->pdrv_private;
drivers/hwtracing/stm/p_sys-t.c
380
if (!op)
drivers/hwtracing/stm/p_sys-t.c
383
if (sys_t_need_clock_sync(op)) {
drivers/hwtracing/stm/p_sys-t.c
389
if (op->node.do_len)
drivers/hwtracing/stm/p_sys-t.c
391
if (sys_t_need_ts(op))
drivers/hwtracing/stm/p_sys-t.c
407
export_uuid(uuid, &op->node.uuid);
drivers/hwtracing/stm/p_sys-t.c
408
sz = stm_data_write(data, m, c, false, uuid, sizeof(op->node.uuid));
drivers/hwtracing/stm/p_sys-t.c
413
if (op->node.do_len) {
drivers/i2c/busses/i2c-mpc.c
774
static int fsl_i2c_probe(struct platform_device *op)
drivers/i2c/busses/i2c-mpc.c
782
i2c = devm_kzalloc(&op->dev, sizeof(*i2c), GFP_KERNEL);
drivers/i2c/busses/i2c-mpc.c
786
i2c->dev = &op->dev; /* for debug and error output */
drivers/i2c/busses/i2c-mpc.c
791
i2c->base = devm_platform_ioremap_resource(op, 0);
drivers/i2c/busses/i2c-mpc.c
795
i2c->irq = platform_get_irq(op, 0);
drivers/i2c/busses/i2c-mpc.c
799
result = devm_request_irq(&op->dev, i2c->irq, mpc_i2c_isr,
drivers/i2c/busses/i2c-mpc.c
810
clk = devm_clk_get_optional_enabled(&op->dev, NULL);
drivers/i2c/busses/i2c-mpc.c
812
dev_err(&op->dev, "failed to enable clock\n");
drivers/i2c/busses/i2c-mpc.c
816
if (of_property_read_bool(op->dev.of_node, "fsl,preserve-clocking")) {
drivers/i2c/busses/i2c-mpc.c
819
result = of_property_read_u32(op->dev.of_node,
drivers/i2c/busses/i2c-mpc.c
825
data = device_get_match_data(&op->dev);
drivers/i2c/busses/i2c-mpc.c
827
data->setup(op->dev.of_node, i2c, clock);
drivers/i2c/busses/i2c-mpc.c
830
if (of_property_read_bool(op->dev.of_node, "dfsrr"))
drivers/i2c/busses/i2c-mpc.c
831
mpc_i2c_setup_8xxx(op->dev.of_node, i2c, clock);
drivers/i2c/busses/i2c-mpc.c
835
result = of_property_read_u32(op->dev.of_node,
drivers/i2c/busses/i2c-mpc.c
839
result = of_property_read_u32(op->dev.of_node,
drivers/i2c/busses/i2c-mpc.c
843
result = of_property_read_u32(op->dev.of_node,
drivers/i2c/busses/i2c-mpc.c
856
if (of_property_read_bool(op->dev.of_node, "fsl,i2c-erratum-a004447"))
drivers/i2c/busses/i2c-mpc.c
861
"MPC adapter (%s)", of_node_full_name(op->dev.of_node));
drivers/i2c/busses/i2c-mpc.c
862
i2c->adap.dev.parent = &op->dev;
drivers/i2c/busses/i2c-mpc.c
863
i2c->adap.nr = op->id;
drivers/i2c/busses/i2c-mpc.c
864
i2c->adap.dev.of_node = of_node_get(op->dev.of_node);
drivers/i2c/busses/i2c-mpc.c
866
platform_set_drvdata(op, i2c);
drivers/i2c/busses/i2c-mpc.c
876
static void fsl_i2c_remove(struct platform_device *op)
drivers/i2c/busses/i2c-mpc.c
878
struct mpc_i2c *i2c = platform_get_drvdata(op);
drivers/i2c/busses/i2c-mt65xx.c
1019
i2c->op != I2C_MASTER_WRRD && num > 1) {
drivers/i2c/busses/i2c-mt65xx.c
1045
if (i2c->op == I2C_MASTER_WRRD)
drivers/i2c/busses/i2c-mt65xx.c
1064
if (i2c->op == I2C_MASTER_WRRD) {
drivers/i2c/busses/i2c-mt65xx.c
1081
if (i2c->op == I2C_MASTER_WRRD)
drivers/i2c/busses/i2c-mt65xx.c
1086
if (i2c->op == I2C_MASTER_RD) {
drivers/i2c/busses/i2c-mt65xx.c
1109
} else if (i2c->op == I2C_MASTER_WR) {
drivers/i2c/busses/i2c-mt65xx.c
1203
if (i2c->op == I2C_MASTER_WR) {
drivers/i2c/busses/i2c-mt65xx.c
1208
} else if (i2c->op == I2C_MASTER_RD) {
drivers/i2c/busses/i2c-mt65xx.c
1279
i2c->op = I2C_MASTER_RD;
drivers/i2c/busses/i2c-mt65xx.c
1281
i2c->op = I2C_MASTER_WR;
drivers/i2c/busses/i2c-mt65xx.c
1285
i2c->op = I2C_MASTER_WRRD;
drivers/i2c/busses/i2c-mt65xx.c
1294
if (i2c->op == I2C_MASTER_WRRD)
drivers/i2c/busses/i2c-mt65xx.c
303
enum mtk_trans_op op;
drivers/i2c/busses/i2c-parport.c
193
const struct lineop *op)
drivers/i2c/busses/i2c-parport.c
195
u8 oldval = port_read[op->port](data);
drivers/i2c/busses/i2c-parport.c
198
if ((op->inverted && !state) || (!op->inverted && state))
drivers/i2c/busses/i2c-parport.c
199
port_write[op->port](data, oldval | op->val);
drivers/i2c/busses/i2c-parport.c
201
port_write[op->port](data, oldval & ~op->val);
drivers/i2c/busses/i2c-parport.c
205
const struct lineop *op)
drivers/i2c/busses/i2c-parport.c
207
u8 oldval = port_read[op->port](data);
drivers/i2c/busses/i2c-parport.c
209
return ((op->inverted && (oldval & op->val) != op->val)
drivers/i2c/busses/i2c-parport.c
210
|| (!op->inverted && (oldval & op->val) == op->val));
drivers/i2c/busses/i2c-qcom-geni.c
624
void **buf, unsigned int op, struct dma_chan *dma_chan)
drivers/i2c/busses/i2c-qcom-geni.c
648
if (op == I2C_WRITE)
drivers/i2c/busses/i2c-qcom-geni.c
673
peripheral->op = op;
drivers/i2c/busses/i2c-qcom-geni.c
677
dev_err(gi2c->se.dev, "dma config error: %d for op:%d\n", ret, op);
drivers/i2c/busses/i2c-qcom-geni.c
684
if (op == I2C_WRITE)
drivers/i2c/busses/i2c-qcom-geni.c
707
if (!((msgs[msg_idx].flags & I2C_M_RD) && op == I2C_WRITE))
drivers/i2c/busses/i2c-sh_mobile.c
304
static unsigned char i2c_op(struct sh_mobile_i2c_data *pd, enum sh_mobile_i2c_op op)
drivers/i2c/busses/i2c-sh_mobile.c
309
dev_dbg(pd->dev, "op %d\n", op);
drivers/i2c/busses/i2c-sh_mobile.c
313
switch (op) {
drivers/i2c/busses/i2c-sh_mobile.c
351
dev_dbg(pd->dev, "op %d, data out 0x%02x\n", op, ret);
drivers/i2c/busses/i2c-xgene-slimpro.c
69
#define SLIMPRO_IIC_ENCODE_MSG(dev, chip, op, proto, addrlen, datalen) \
drivers/i2c/busses/i2c-xgene-slimpro.c
75
((op << SLIMPRO_IIC_RW_SHIFT) & SLIMPRO_IIC_RW_MASK) | \
drivers/iio/industrialio-backend.c
103
#define iio_backend_ptr_op_call(back, op, args...) ({ \
drivers/iio/industrialio-backend.c
108
__ret = iio_backend_check_op(__back, op); \
drivers/iio/industrialio-backend.c
112
ptr_err = __back->ops->op(__back, ##args); \
drivers/iio/industrialio-backend.c
117
#define iio_backend_void_op_call(back, op, args...) { \
drivers/iio/industrialio-backend.c
121
__ret = iio_backend_check_op(__back, op); \
drivers/iio/industrialio-backend.c
123
__back->ops->op(__back, ##args); \
drivers/iio/industrialio-backend.c
126
__stringify(op)); \
drivers/iio/industrialio-backend.c
82
#define iio_backend_check_op(back, op) ({ \
drivers/iio/industrialio-backend.c
86
if (!____back->ops->op) \
drivers/iio/industrialio-backend.c
92
#define iio_backend_op_call(back, op, args...) ({ \
drivers/iio/industrialio-backend.c
96
__ret = iio_backend_check_op(__back, op); \
drivers/iio/industrialio-backend.c
98
__ret = __back->ops->op(__back, ##args); \
drivers/iio/light/gp2ap020a00f.c
367
enum gp2ap020a00f_opmode op)
drivers/iio/light/gp2ap020a00f.c
372
if (op != GP2AP020A00F_OPMODE_SHUTDOWN) {
drivers/iio/light/gp2ap020a00f.c
381
if ((opmode_regs_settings[op].op_reg & GP2AP020A00F_OP_MASK) !=
drivers/iio/light/gp2ap020a00f.c
392
GP2AP020A00F_PRST_MASK, opmode_regs_settings[op]
drivers/iio/light/gp2ap020a00f.c
398
GP2AP020A00F_INTTYPE_MASK, opmode_regs_settings[op]
drivers/iio/light/gp2ap020a00f.c
404
GP2AP020A00F_PIN_MASK, opmode_regs_settings[op]
drivers/iio/light/gp2ap020a00f.c
415
opmode_regs_settings[op].op_reg);
drivers/iio/light/gp2ap020a00f.c
419
data->cur_opmode = op;
drivers/iio/light/si1145.c
315
static int si1145_param_update(struct si1145_data *data, u8 op, u8 param,
drivers/iio/light/si1145.c
325
return si1145_command(data, op | (param & 0x1F));
drivers/infiniband/core/netlink.c
106
if (!cb_table || (!cb_table[op].dump && !cb_table[op].doit))
drivers/infiniband/core/netlink.c
132
int len, int client, int op, int flags)
drivers/infiniband/core/netlink.c
134
*nlh = nlmsg_put(skb, 0, seq, RDMA_NL_GET_TYPE(client, op), len, flags);
drivers/infiniband/core/netlink.c
157
unsigned int op = RDMA_NL_GET_OP(type);
drivers/infiniband/core/netlink.c
161
if (!is_nl_msg_valid(index, op))
drivers/infiniband/core/netlink.c
165
cb_table = get_cb_table(skb, index, op);
drivers/infiniband/core/netlink.c
169
if ((cb_table[op].flags & RDMA_NL_ADMIN_PERM) &&
drivers/infiniband/core/netlink.c
180
if (cb_table[op].doit)
drivers/infiniband/core/netlink.c
181
err = cb_table[op].doit(skb, nlh, extack);
drivers/infiniband/core/netlink.c
187
.dump = cb_table[op].dump,
drivers/infiniband/core/netlink.c
194
if (cb_table[op].doit)
drivers/infiniband/core/netlink.c
195
err = cb_table[op].doit(skb, nlh, extack);
drivers/infiniband/core/netlink.c
61
static bool is_nl_msg_valid(unsigned int type, unsigned int op)
drivers/infiniband/core/netlink.c
78
return op < max_num_ops[type];
drivers/infiniband/core/netlink.c
82
get_cb_table(const struct sk_buff *skb, unsigned int type, unsigned int op)
drivers/infiniband/hw/cxgb4/cq.c
135
res->u.cq.op = FW_RI_RES_OP_WRITE;
drivers/infiniband/hw/cxgb4/cq.c
57
res->u.cq.op = FW_RI_RES_OP_RESET;
drivers/infiniband/hw/cxgb4/qp.c
1739
init->u.write.u.immd_src[0].op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/qp.c
2517
res->u.srq.op = FW_RI_RES_OP_RESET;
drivers/infiniband/hw/cxgb4/qp.c
2605
res->u.srq.op = FW_RI_RES_OP_WRITE;
drivers/infiniband/hw/cxgb4/qp.c
322
res->u.sqrq.op = FW_RI_RES_OP_WRITE;
drivers/infiniband/hw/cxgb4/qp.c
351
res->u.sqrq.op = FW_RI_RES_OP_WRITE;
drivers/infiniband/hw/cxgb4/qp.c
442
immdp->op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/qp.c
477
isglp->op = FW_RI_DATA_ISGL;
drivers/infiniband/hw/cxgb4/qp.c
541
wqe->send.u.immd_src[0].op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/qp.c
592
wqe->write.u.immd_src[0].op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/qp.c
609
immdp->op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/qp.c
848
sglp->op = FW_RI_DATA_DSGL;
drivers/infiniband/hw/cxgb4/qp.c
857
imdp->op = FW_RI_DATA_IMMD;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
120
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
135
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
143
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
274
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
284
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
298
__u8 op;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
601
__u8 op;
drivers/infiniband/hw/efa/efa_verbs.c
35
#define EFA_DEFINE_DEVICE_STATS(op) \
drivers/infiniband/hw/efa/efa_verbs.c
36
op(EFA_SUBMITTED_CMDS, "submitted_cmds") \
drivers/infiniband/hw/efa/efa_verbs.c
37
op(EFA_COMPLETED_CMDS, "completed_cmds") \
drivers/infiniband/hw/efa/efa_verbs.c
38
op(EFA_CMDS_ERR, "cmds_err") \
drivers/infiniband/hw/efa/efa_verbs.c
39
op(EFA_NO_COMPLETION_CMDS, "no_completion_cmds") \
drivers/infiniband/hw/efa/efa_verbs.c
40
op(EFA_KEEP_ALIVE_RCVD, "keep_alive_rcvd") \
drivers/infiniband/hw/efa/efa_verbs.c
41
op(EFA_ALLOC_PD_ERR, "alloc_pd_err") \
drivers/infiniband/hw/efa/efa_verbs.c
42
op(EFA_CREATE_QP_ERR, "create_qp_err") \
drivers/infiniband/hw/efa/efa_verbs.c
43
op(EFA_CREATE_CQ_ERR, "create_cq_err") \
drivers/infiniband/hw/efa/efa_verbs.c
44
op(EFA_REG_MR_ERR, "reg_mr_err") \
drivers/infiniband/hw/efa/efa_verbs.c
45
op(EFA_ALLOC_UCONTEXT_ERR, "alloc_ucontext_err") \
drivers/infiniband/hw/efa/efa_verbs.c
46
op(EFA_CREATE_AH_ERR, "create_ah_err") \
drivers/infiniband/hw/efa/efa_verbs.c
47
op(EFA_MMAP_ERR, "mmap_err")
drivers/infiniband/hw/efa/efa_verbs.c
49
#define EFA_DEFINE_PORT_STATS(op) \
drivers/infiniband/hw/efa/efa_verbs.c
50
op(EFA_TX_BYTES, "tx_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
51
op(EFA_TX_PKTS, "tx_pkts") \
drivers/infiniband/hw/efa/efa_verbs.c
52
op(EFA_RX_BYTES, "rx_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
53
op(EFA_RX_PKTS, "rx_pkts") \
drivers/infiniband/hw/efa/efa_verbs.c
54
op(EFA_RX_DROPS, "rx_drops") \
drivers/infiniband/hw/efa/efa_verbs.c
55
op(EFA_SEND_BYTES, "send_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
56
op(EFA_SEND_WRS, "send_wrs") \
drivers/infiniband/hw/efa/efa_verbs.c
57
op(EFA_RECV_BYTES, "recv_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
58
op(EFA_RECV_WRS, "recv_wrs") \
drivers/infiniband/hw/efa/efa_verbs.c
59
op(EFA_RDMA_READ_WRS, "rdma_read_wrs") \
drivers/infiniband/hw/efa/efa_verbs.c
60
op(EFA_RDMA_READ_BYTES, "rdma_read_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
61
op(EFA_RDMA_READ_WR_ERR, "rdma_read_wr_err") \
drivers/infiniband/hw/efa/efa_verbs.c
62
op(EFA_RDMA_READ_RESP_BYTES, "rdma_read_resp_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
63
op(EFA_RDMA_WRITE_WRS, "rdma_write_wrs") \
drivers/infiniband/hw/efa/efa_verbs.c
64
op(EFA_RDMA_WRITE_BYTES, "rdma_write_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
65
op(EFA_RDMA_WRITE_WR_ERR, "rdma_write_wr_err") \
drivers/infiniband/hw/efa/efa_verbs.c
66
op(EFA_RDMA_WRITE_RECV_BYTES, "rdma_write_recv_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
67
op(EFA_RETRANS_BYTES, "retrans_bytes") \
drivers/infiniband/hw/efa/efa_verbs.c
68
op(EFA_RETRANS_PKTS, "retrans_pkts") \
drivers/infiniband/hw/efa/efa_verbs.c
69
op(EFA_RETRANS_TIMEOUT_EVENS, "retrans_timeout_events") \
drivers/infiniband/hw/efa/efa_verbs.c
70
op(EFA_UNRESPONSIVE_REMOTE_EVENTS, "unresponsive_remote_events") \
drivers/infiniband/hw/efa/efa_verbs.c
71
op(EFA_IMPAIRED_REMOTE_CONN_EVENTS, "impaired_remote_conn_events") \
drivers/infiniband/hw/erdma/erdma.h
266
void erdma_cmdq_build_reqhdr(u64 *hdr, u32 mod, u32 op);
drivers/infiniband/hw/erdma/erdma_cmdq.c
392
void erdma_cmdq_build_reqhdr(u64 *hdr, u32 mod, u32 op)
drivers/infiniband/hw/erdma/erdma_cmdq.c
395
FIELD_PREP(ERDMA_CMD_HDR_OPCODE_MASK, op);
drivers/infiniband/hw/erdma/erdma_qp.c
406
u32 op = ERDMA_OP_SEND;
drivers/infiniband/hw/erdma/erdma_qp.c
409
op = ERDMA_OP_SEND_WITH_IMM;
drivers/infiniband/hw/erdma/erdma_qp.c
412
op = ERDMA_OP_SEND_WITH_INV;
drivers/infiniband/hw/erdma/erdma_qp.c
416
*hw_op = op;
drivers/infiniband/hw/erdma/erdma_qp.c
424
u32 op = ERDMA_OP_SEND;
drivers/infiniband/hw/erdma/erdma_qp.c
427
op = ERDMA_OP_SEND_WITH_IMM;
drivers/infiniband/hw/erdma/erdma_qp.c
431
*hw_op = op;
drivers/infiniband/hw/erdma/erdma_qp.c
447
enum ib_wr_opcode op = send_wr->opcode;
drivers/infiniband/hw/erdma/erdma_qp.c
485
switch (op) {
drivers/infiniband/hw/erdma/erdma_qp.c
489
if (op == IB_WR_RDMA_WRITE_WITH_IMM)
drivers/infiniband/hw/erdma/erdma_qp.c
512
if (op == IB_WR_RDMA_READ_WITH_INV) {
drivers/infiniband/hw/erdma/erdma_qp.c
594
if (op == IB_WR_ATOMIC_CMP_AND_SWP) {
drivers/infiniband/hw/erdma/erdma_verbs.c
2150
static int erdma_set_gid(struct erdma_dev *dev, u8 op, u32 idx,
drivers/infiniband/hw/erdma/erdma_verbs.c
2157
FIELD_PREP(ERDMA_CMD_SET_GID_OP_MASK, op);
drivers/infiniband/hw/erdma/erdma_verbs.c
2159
if (op == ERDMA_SET_GID_OP_ADD) {
drivers/infiniband/hw/hfi1/chip.c
11997
void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op,
drivers/infiniband/hw/hfi1/chip.c
12009
hfi1_cdbg(RCVCTRL, "ctxt %d op 0x%x", ctxt, op);
drivers/infiniband/hw/hfi1/chip.c
12013
if ((op & HFI1_RCVCTRL_CTXT_ENB) &&
drivers/infiniband/hw/hfi1/chip.c
12078
if (op & HFI1_RCVCTRL_CTXT_DIS) {
drivers/infiniband/hw/hfi1/chip.c
12094
if (op & HFI1_RCVCTRL_INTRAVAIL_ENB) {
drivers/infiniband/hw/hfi1/chip.c
12099
if (op & HFI1_RCVCTRL_INTRAVAIL_DIS) {
drivers/infiniband/hw/hfi1/chip.c
12104
if ((op & HFI1_RCVCTRL_TAILUPD_ENB) && hfi1_rcvhdrtail_kvaddr(rcd))
drivers/infiniband/hw/hfi1/chip.c
12106
if (op & HFI1_RCVCTRL_TAILUPD_DIS) {
drivers/infiniband/hw/hfi1/chip.c
12108
if (!(op & HFI1_RCVCTRL_CTXT_DIS))
drivers/infiniband/hw/hfi1/chip.c
12111
if (op & HFI1_RCVCTRL_TIDFLOW_ENB)
drivers/infiniband/hw/hfi1/chip.c
12113
if (op & HFI1_RCVCTRL_TIDFLOW_DIS)
drivers/infiniband/hw/hfi1/chip.c
12115
if (op & HFI1_RCVCTRL_ONE_PKT_EGR_ENB) {
drivers/infiniband/hw/hfi1/chip.c
12123
if (op & HFI1_RCVCTRL_ONE_PKT_EGR_DIS)
drivers/infiniband/hw/hfi1/chip.c
12125
if (op & HFI1_RCVCTRL_NO_RHQ_DROP_ENB)
drivers/infiniband/hw/hfi1/chip.c
12127
if (op & HFI1_RCVCTRL_NO_RHQ_DROP_DIS)
drivers/infiniband/hw/hfi1/chip.c
12129
if (op & HFI1_RCVCTRL_NO_EGR_DROP_ENB)
drivers/infiniband/hw/hfi1/chip.c
12131
if (op & HFI1_RCVCTRL_NO_EGR_DROP_DIS)
drivers/infiniband/hw/hfi1/chip.c
12133
if (op & HFI1_RCVCTRL_URGENT_ENB)
drivers/infiniband/hw/hfi1/chip.c
12136
if (op & HFI1_RCVCTRL_URGENT_DIS)
drivers/infiniband/hw/hfi1/chip.c
12175
if (op & (HFI1_RCVCTRL_TAILUPD_DIS | HFI1_RCVCTRL_CTXT_DIS))
drivers/infiniband/hw/hfi1/chip.h
1382
void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op,
drivers/infiniband/hw/hfi1/pio.c
33
void pio_send_control(struct hfi1_devdata *dd, int op)
drivers/infiniband/hw/hfi1/pio.c
44
switch (op) {
drivers/infiniband/hw/hfi1/pio.c
76
dd_dev_err(dd, "%s: invalid control %d\n", __func__, op);
drivers/infiniband/hw/hfi1/pio.h
279
void pio_send_control(struct hfi1_devdata *dd, int op);
drivers/infiniband/hw/hfi1/sdma.c
1882
static void sdma_sendctrl(struct sdma_engine *sde, unsigned op)
drivers/infiniband/hw/hfi1/sdma.c
1891
(op & SDMA_SENDCTRL_OP_ENABLE) ? 1 : 0,
drivers/infiniband/hw/hfi1/sdma.c
1892
(op & SDMA_SENDCTRL_OP_INTENABLE) ? 1 : 0,
drivers/infiniband/hw/hfi1/sdma.c
1893
(op & SDMA_SENDCTRL_OP_HALT) ? 1 : 0,
drivers/infiniband/hw/hfi1/sdma.c
1894
(op & SDMA_SENDCTRL_OP_CLEANUP) ? 1 : 0);
drivers/infiniband/hw/hfi1/sdma.c
1897
if (op & SDMA_SENDCTRL_OP_ENABLE)
drivers/infiniband/hw/hfi1/sdma.c
1902
if (op & SDMA_SENDCTRL_OP_INTENABLE)
drivers/infiniband/hw/hfi1/sdma.c
1907
if (op & SDMA_SENDCTRL_OP_HALT)
drivers/infiniband/hw/hfi1/sdma.c
1917
if (op & SDMA_SENDCTRL_OP_CLEANUP)
drivers/infiniband/hw/hfi1/sdma.c
636
unsigned op = 0;
drivers/infiniband/hw/hfi1/sdma.c
653
op |= SDMA_SENDCTRL_OP_ENABLE;
drivers/infiniband/hw/hfi1/sdma.c
656
op |= SDMA_SENDCTRL_OP_INTENABLE;
drivers/infiniband/hw/hfi1/sdma.c
659
op |= SDMA_SENDCTRL_OP_HALT;
drivers/infiniband/hw/hfi1/sdma.c
662
op |= SDMA_SENDCTRL_OP_CLEANUP;
drivers/infiniband/hw/hfi1/sdma.c
670
ss->current_op = op;
drivers/infiniband/hw/hfi1/trace.c
102
#define OP(transport, op) IB_OPCODE_## transport ## _ ## op
drivers/infiniband/hw/hfi1/trace_ibhdrs.h
105
#define __parse_ib_ehdrs(op, l4, dest_qpn, src_qpn, ehdrs) \
drivers/infiniband/hw/hfi1/trace_ibhdrs.h
106
parse_everbs_hdrs(p, op, l4, dest_qpn, src_qpn, ehdrs)
drivers/infiniband/hw/ionic/ionic_admin.c
439
wr->wqe.op,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1169
.op = IONIC_V1_ADMIN_CREATE_CQ,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1198
.op = IONIC_V1_ADMIN_DESTROY_CQ,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1339
.op = IONIC_V1_ADMIN_CREATE_QP,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1399
.op = IONIC_V1_ADMIN_MODIFY_QP,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1530
.op = IONIC_V1_ADMIN_QUERY_QP,
drivers/infiniband/hw/ionic/ionic_controlpath.c
1677
.op = IONIC_V1_ADMIN_DESTROY_QP,
drivers/infiniband/hw/ionic/ionic_controlpath.c
627
.op = IONIC_V1_ADMIN_CREATE_AH,
drivers/infiniband/hw/ionic/ionic_controlpath.c
709
.op = IONIC_V1_ADMIN_DESTROY_AH,
drivers/infiniband/hw/ionic/ionic_controlpath.c
803
.op = IONIC_V1_ADMIN_CREATE_MR,
drivers/infiniband/hw/ionic/ionic_controlpath.c
836
.op = IONIC_V1_ADMIN_DESTROY_MR,
drivers/infiniband/hw/ionic/ionic_datapath.c
1023
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, ATOMIC_CS);
drivers/infiniband/hw/ionic/ionic_datapath.c
1031
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, ATOMIC_FA);
drivers/infiniband/hw/ionic/ionic_datapath.c
1066
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, LOCAL_INV);
drivers/infiniband/hw/ionic/ionic_datapath.c
1101
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, REG_MR);
drivers/infiniband/hw/ionic/ionic_datapath.c
138
u8 op;
drivers/infiniband/hw/ionic/ionic_datapath.c
208
op = src_qpn >> IONIC_V1_CQE_RECV_OP_SHIFT;
drivers/infiniband/hw/ionic/ionic_datapath.c
211
op &= IONIC_V1_CQE_RECV_OP_MASK;
drivers/infiniband/hw/ionic/ionic_datapath.c
214
switch (op) {
drivers/infiniband/hw/ionic/ionic_datapath.c
902
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, SEND);
drivers/infiniband/hw/ionic/ionic_datapath.c
905
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, SEND_IMM);
drivers/infiniband/hw/ionic/ionic_datapath.c
909
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, SEND_INV);
drivers/infiniband/hw/ionic/ionic_datapath.c
946
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, SEND);
drivers/infiniband/hw/ionic/ionic_datapath.c
949
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, SEND_IMM);
drivers/infiniband/hw/ionic/ionic_datapath.c
978
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, RDMA_READ);
drivers/infiniband/hw/ionic/ionic_datapath.c
983
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, RDMA_WRITE);
drivers/infiniband/hw/ionic/ionic_datapath.c
986
wqe->base.op = IONIC_OP(dev->lif_cfg.rdma_version, RDMA_WRITE_IMM);
drivers/infiniband/hw/ionic/ionic_fw.h
427
__u8 op;
drivers/infiniband/hw/ionic/ionic_fw.h
846
__u8 op;
drivers/infiniband/hw/ionic/ionic_hw_stats.c
100
if (dev->lif_cfg.admin_opcodes <= op)
drivers/infiniband/hw/ionic/ionic_hw_stats.c
85
dma_addr_t dma, size_t len, int qid, int op)
drivers/infiniband/hw/ionic/ionic_hw_stats.c
90
.op = op,
drivers/infiniband/hw/ionic/ionic_ibdev.h
414
static inline bool ionic_ibop_is_local(enum ib_wr_opcode op)
drivers/infiniband/hw/ionic/ionic_ibdev.h
416
return op == IB_WR_LOCAL_INV || op == IB_WR_REG_MR;
drivers/infiniband/hw/irdma/cm.c
4345
enum irdma_quad_hash_manage_type op = ifup ?
drivers/infiniband/hw/irdma/cm.c
4372
err = irdma_manage_qhash(iwdev, nfo, IRDMA_QHASH_TYPE_TCP_SYN, op, NULL,
drivers/infiniband/hw/irdma/cm.c
4401
enum irdma_quad_hash_manage_type op = ifup ?
drivers/infiniband/hw/irdma/cm.c
4427
IRDMA_QHASH_TYPE_TCP_SYN, op,
drivers/infiniband/hw/irdma/ctrl.c
43
void irdma_sc_suspend_resume_qps(struct irdma_sc_vsi *vsi, u8 op)
drivers/infiniband/hw/irdma/ctrl.c
52
if (op == IRDMA_OP_RESUME) {
drivers/infiniband/hw/irdma/ctrl.c
56
irdma_cqp_qp_suspend_resume(qp, op);
drivers/infiniband/hw/irdma/ctrl.c
58
irdma_cqp_qp_suspend_resume(qp, op);
drivers/infiniband/hw/irdma/ctrl.c
61
} else if (op == IRDMA_OP_SUSPEND) {
drivers/infiniband/hw/irdma/ctrl.c
63
if (!irdma_cqp_qp_suspend_resume(qp, op))
drivers/infiniband/hw/irdma/protos.h
36
u8 op);
drivers/infiniband/hw/irdma/protos.h
38
u8 op);
drivers/infiniband/hw/irdma/uda.c
116
struct irdma_mcast_grp_info *info, u32 op,
drivers/infiniband/hw/irdma/uda.c
158
FIELD_PREP(IRDMA_UDA_CQPSQ_MG_OPCODE, op) |
drivers/infiniband/hw/irdma/uda.c
21
u32 op, u64 scratch)
drivers/infiniband/hw/irdma/uda.c
70
FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_OPCODE, op) |
drivers/infiniband/hw/irdma/uda.h
40
u32 op, u64 scratch);
drivers/infiniband/hw/irdma/uda.h
42
struct irdma_mcast_grp_info *info, u32 op,
drivers/infiniband/hw/irdma/uk.c
240
op_info = &info->op.rdma_write;
drivers/infiniband/hw/irdma/uk.c
331
op_info = &info->op.atomic_fetch_add;
drivers/infiniband/hw/irdma/uk.c
382
op_info = &info->op.atomic_compare_swap;
drivers/infiniband/hw/irdma/uk.c
489
op_info = &info->op.rdma_read;
drivers/infiniband/hw/irdma/uk.c
568
op_info = &info->op.send;
drivers/infiniband/hw/irdma/uk.c
814
op_info = &info->op.rdma_write;
drivers/infiniband/hw/irdma/uk.c
882
op_info = &info->op.send;
drivers/infiniband/hw/irdma/uk.c
954
op_info = &info->op.inv_local_stag;
drivers/infiniband/hw/irdma/user.h
366
} op;
drivers/infiniband/hw/irdma/utils.c
1812
u8 op)
drivers/infiniband/hw/irdma/utils.c
1825
cqp_info->cqp_cmd = op;
drivers/infiniband/hw/irdma/utils.c
1842
u8 op)
drivers/infiniband/hw/irdma/utils.c
1855
cqp_info->cqp_cmd = op;
drivers/infiniband/hw/irdma/utils.c
914
int irdma_cqp_qp_suspend_resume(struct irdma_sc_qp *qp, u8 op)
drivers/infiniband/hw/irdma/utils.c
928
cqp_info->cqp_cmd = op;
drivers/infiniband/hw/irdma/verbs.c
3985
info.op.atomic_compare_swap.tagged_offset = ib_wr->sg_list[0].addr;
drivers/infiniband/hw/irdma/verbs.c
3986
info.op.atomic_compare_swap.remote_tagged_offset =
drivers/infiniband/hw/irdma/verbs.c
3988
info.op.atomic_compare_swap.swap_data_bytes = atomic_wr(ib_wr)->swap;
drivers/infiniband/hw/irdma/verbs.c
3989
info.op.atomic_compare_swap.compare_data_bytes =
drivers/infiniband/hw/irdma/verbs.c
3991
info.op.atomic_compare_swap.stag = ib_wr->sg_list[0].lkey;
drivers/infiniband/hw/irdma/verbs.c
3992
info.op.atomic_compare_swap.remote_stag = atomic_wr(ib_wr)->rkey;
drivers/infiniband/hw/irdma/verbs.c
4002
info.op.atomic_fetch_add.tagged_offset = ib_wr->sg_list[0].addr;
drivers/infiniband/hw/irdma/verbs.c
4003
info.op.atomic_fetch_add.remote_tagged_offset =
drivers/infiniband/hw/irdma/verbs.c
4005
info.op.atomic_fetch_add.fetch_add_data_bytes =
drivers/infiniband/hw/irdma/verbs.c
4007
info.op.atomic_fetch_add.stag = ib_wr->sg_list[0].lkey;
drivers/infiniband/hw/irdma/verbs.c
4008
info.op.atomic_fetch_add.remote_stag =
drivers/infiniband/hw/irdma/verbs.c
4037
info.op.send.num_sges = ib_wr->num_sge;
drivers/infiniband/hw/irdma/verbs.c
4038
info.op.send.sg_list = ib_wr->sg_list;
drivers/infiniband/hw/irdma/verbs.c
4042
info.op.send.ah_id = ah->sc_ah.ah_info.ah_idx;
drivers/infiniband/hw/irdma/verbs.c
4043
info.op.send.qkey = ud_wr(ib_wr)->remote_qkey;
drivers/infiniband/hw/irdma/verbs.c
4044
info.op.send.dest_qp = ud_wr(ib_wr)->remote_qpn;
drivers/infiniband/hw/irdma/verbs.c
4067
info.op.rdma_write.num_lo_sges = ib_wr->num_sge;
drivers/infiniband/hw/irdma/verbs.c
4068
info.op.rdma_write.lo_sg_list = ib_wr->sg_list;
drivers/infiniband/hw/irdma/verbs.c
4069
info.op.rdma_write.rem_addr.addr =
drivers/infiniband/hw/irdma/verbs.c
4071
info.op.rdma_write.rem_addr.lkey = rdma_wr(ib_wr)->rkey;
drivers/infiniband/hw/irdma/verbs.c
4087
info.op.rdma_read.rem_addr.addr = rdma_wr(ib_wr)->remote_addr;
drivers/infiniband/hw/irdma/verbs.c
4088
info.op.rdma_read.rem_addr.lkey = rdma_wr(ib_wr)->rkey;
drivers/infiniband/hw/irdma/verbs.c
4089
info.op.rdma_read.lo_sg_list = (void *)ib_wr->sg_list;
drivers/infiniband/hw/irdma/verbs.c
4090
info.op.rdma_read.num_lo_sges = ib_wr->num_sge;
drivers/infiniband/hw/irdma/verbs.c
4096
info.op.inv_local_stag.target_stag = ib_wr->ex.invalidate_rkey;
drivers/infiniband/hw/irdma/verbs.c
4749
struct irdma_mcast_grp_info *mc_grp_ctx, u8 op)
drivers/infiniband/hw/irdma/verbs.c
4761
cqp_info->cqp_cmd = op;
drivers/infiniband/hw/mana/main.c
897
req.op = ADDR_OP_ADD;
drivers/infiniband/hw/mana/main.c
927
req.op = ADDR_OP_REMOVE;
drivers/infiniband/hw/mana/main.c
940
int mana_ib_gd_config_mac(struct mana_ib_dev *mdev, enum mana_ib_addr_op op, u8 *mac)
drivers/infiniband/hw/mana/main.c
950
req.op = op;
drivers/infiniband/hw/mana/mana_ib.h
294
enum mana_ib_addr_op op;
drivers/infiniband/hw/mana/mana_ib.h
306
enum mana_ib_addr_op op;
drivers/infiniband/hw/mana/mana_ib.h
712
int mana_ib_gd_config_mac(struct mana_ib_dev *mdev, enum mana_ib_addr_op op, u8 *mac);
drivers/infiniband/hw/mlx5/dm.c
192
static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op,
drivers/infiniband/hw/mlx5/dm.c
197
op_entry = xa_load(&dm->ops, op);
drivers/infiniband/hw/mlx5/dm.c
214
u8 op;
drivers/infiniband/hw/mlx5/dm.c
216
err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP);
drivers/infiniband/hw/mlx5/dm.c
220
if (op >= BITS_PER_TYPE(u32))
drivers/infiniband/hw/mlx5/dm.c
223
if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op)))
drivers/infiniband/hw/mlx5/dm.c
227
err = map_existing_op(dm, op, attrs);
drivers/infiniband/hw/mlx5/dm.c
235
err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op,
drivers/infiniband/hw/mlx5/dm.c
241
op_entry->op = op;
drivers/infiniband/hw/mlx5/dm.c
248
mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op);
drivers/infiniband/hw/mlx5/dm.c
259
err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL);
drivers/infiniband/hw/mlx5/dm.c
553
op_entry->op);
drivers/infiniband/hw/mlx5/dm.h
25
u8 op;
drivers/infiniband/hw/mlx5/qp.c
4177
u16 op;
drivers/infiniband/hw/mlx5/qp.c
4360
op = optab[mlx5_cur][mlx5_new];
drivers/infiniband/hw/mlx5/qp.c
4373
raw_qp_param.operation = op;
drivers/infiniband/hw/mlx5/qp.c
4419
err = mlx5_core_qp_modify(dev, op, optpar, qpc, &base->mqp,
drivers/infiniband/hw/mlx5/wr.c
494
sblock_ctrl->op = cpu_to_be32(MLX5_STRIDE_BLOCK_OP);
drivers/infiniband/hw/mthca/mthca_cmd.c
1730
static const u16 op[IB_QPS_ERR + 1][IB_QPS_ERR + 1] = {
drivers/infiniband/hw/mthca/mthca_cmd.c
1774
if (op[cur][next] == CMD_ERR2RST_QPEE) {
drivers/infiniband/hw/mthca/mthca_cmd.c
1789
op[cur][next], CMD_TIME_CLASS_C);
drivers/infiniband/hw/mthca/mthca_cmd.c
1823
op_mod, op[cur][next], CMD_TIME_CLASS_C);
drivers/infiniband/hw/mthca/mthca_cmd.c
203
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
224
op), ptr + offs[6]);
drivers/infiniband/hw/mthca/mthca_cmd.c
235
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
270
op), dev->hcr + 6 * 4);
drivers/infiniband/hw/mthca/mthca_cmd.c
280
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
290
op_modifier, op, token);
drivers/infiniband/hw/mthca/mthca_cmd.c
293
op_modifier, op, token, event);
drivers/infiniband/hw/mthca/mthca_cmd.c
337
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
349
op, CMD_POLL_TOKEN, 0);
drivers/infiniband/hw/mthca/mthca_cmd.c
378
op, status);
drivers/infiniband/hw/mthca/mthca_cmd.c
412
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
432
op, context->token, 1);
drivers/infiniband/hw/mthca/mthca_cmd.c
447
op, context->status);
drivers/infiniband/hw/mthca/mthca_cmd.c
474
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
479
in_modifier, op_modifier, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
483
in_modifier, op_modifier, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
492
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
496
op_modifier, op, timeout);
drivers/infiniband/hw/mthca/mthca_cmd.c
509
u16 op,
drivers/infiniband/hw/mthca/mthca_cmd.c
514
in_modifier, op_modifier, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
518
in_modifier, op_modifier, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
656
static int mthca_map_cmd(struct mthca_dev *dev, u16 op, struct mthca_icm *icm,
drivers/infiniband/hw/mthca/mthca_cmd.c
704
err = mthca_cmd(dev, mailbox->dma, nent, 0, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
714
err = mthca_cmd(dev, mailbox->dma, nent, 0, op,
drivers/infiniband/hw/mthca/mthca_cmd.c
717
switch (op) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma.h
418
static inline enum pvrdma_wr_opcode ib_wr_opcode_to_pvrdma(enum ib_wr_opcode op)
drivers/infiniband/hw/vmw_pvrdma/pvrdma.h
420
switch (op) {
drivers/infiniband/sw/siw/siw_mem.c
230
void siw_wqe_put_mem(struct siw_wqe *wqe, enum siw_opcode op)
drivers/infiniband/sw/siw/siw_mem.c
232
switch (op) {
drivers/infiniband/sw/siw/siw_mem.h
21
void siw_wqe_put_mem(struct siw_wqe *wqe, enum siw_opcode op);
drivers/infiniband/sw/siw/siw_qp.c
551
enum rdma_opcode op = __rdmap_get_opcode(&rx_hdr->ctrl);
drivers/infiniband/sw/siw/siw_qp.c
553
real_ddp_len -= iwarp_pktinfo[op].hdr_len - MPA_HDR_SIZE;
drivers/infiniband/sw/siw/siw_qp_rx.c
885
enum rdma_opcode op;
drivers/infiniband/sw/siw/siw_qp_rx.c
917
op = __rdmap_get_opcode(&term_info.ctrl);
drivers/infiniband/sw/siw/siw_qp_rx.c
918
if (op >= RDMAP_TERMINATE)
drivers/infiniband/sw/siw/siw_qp_rx.c
926
to_copy = iwarp_pktinfo[op].hdr_len - to_copy;
drivers/infiniband/sw/siw/siw_qp_rx.c
936
op, be16_to_cpu(term_info.ctrl.mpa_len),
drivers/infiniband/sw/siw/siw_qp_rx.c
940
op, be16_to_cpu(term_info.ctrl.mpa_len),
drivers/input/misc/powermate.c
138
int op, arg;
drivers/input/misc/powermate.c
157
op = 0; // divide
drivers/input/misc/powermate.c
160
op = 2; // multiply
drivers/input/misc/powermate.c
163
op = 1; // normal speed
drivers/input/misc/powermate.c
167
pm->configcr->wIndex = cpu_to_le16( (arg << 8) | op );
drivers/input/misc/sparcspkr.c
191
static int bbc_beep_probe(struct platform_device *op)
drivers/input/misc/sparcspkr.c
197
state = devm_kzalloc(&op->dev, sizeof(*state), GFP_KERNEL);
drivers/input/misc/sparcspkr.c
214
info->regs = of_ioremap(&op->resource[0], 0, 6, "bbc beep");
drivers/input/misc/sparcspkr.c
218
platform_set_drvdata(op, state);
drivers/input/misc/sparcspkr.c
220
err = sparcspkr_probe(&op->dev);
drivers/input/misc/sparcspkr.c
222
of_iounmap(&op->resource[0], info->regs, 6);
drivers/input/misc/sparcspkr.c
229
static void bbc_remove(struct platform_device *op)
drivers/input/misc/sparcspkr.c
231
struct sparcspkr_state *state = platform_get_drvdata(op);
drivers/input/misc/sparcspkr.c
240
of_iounmap(&op->resource[0], info->regs, 6);
drivers/input/misc/sparcspkr.c
262
static int grover_beep_probe(struct platform_device *op)
drivers/input/misc/sparcspkr.c
268
state = devm_kzalloc(&op->dev, sizeof(*state), GFP_KERNEL);
drivers/input/misc/sparcspkr.c
277
info->freq_regs = of_ioremap(&op->resource[2], 0, 2, "grover beep freq");
drivers/input/misc/sparcspkr.c
281
info->enable_reg = of_ioremap(&op->resource[3], 0, 1, "grover beep enable");
drivers/input/misc/sparcspkr.c
285
platform_set_drvdata(op, state);
drivers/input/misc/sparcspkr.c
287
err = sparcspkr_probe(&op->dev);
drivers/input/misc/sparcspkr.c
294
of_iounmap(&op->resource[3], info->enable_reg, 1);
drivers/input/misc/sparcspkr.c
297
of_iounmap(&op->resource[2], info->freq_regs, 2);
drivers/input/misc/sparcspkr.c
302
static void grover_remove(struct platform_device *op)
drivers/input/misc/sparcspkr.c
304
struct sparcspkr_state *state = platform_get_drvdata(op);
drivers/input/misc/sparcspkr.c
313
of_iounmap(&op->resource[3], info->enable_reg, 1);
drivers/input/misc/sparcspkr.c
314
of_iounmap(&op->resource[2], info->freq_regs, 2);
drivers/input/rmi4/rmi_spi.c
130
switch (cmd->op) {
drivers/input/rmi4/rmi_spi.c
161
switch (cmd->op) {
drivers/input/rmi4/rmi_spi.c
204
__func__, cmd->op == RMI_SPI_WRITE ? "WRITE" : "READ",
drivers/input/rmi4/rmi_spi.c
261
cmd.op = RMI_SPI_WRITE;
drivers/input/rmi4/rmi_spi.c
288
cmd.op = RMI_SPI_WRITE;
drivers/input/rmi4/rmi_spi.c
314
cmd.op = RMI_SPI_READ;
drivers/input/rmi4/rmi_spi.c
32
enum rmi_spi_op op;
drivers/input/serio/i8042-sparcio.h
57
static int sparc_i8042_probe(struct platform_device *op)
drivers/input/serio/i8042-sparcio.h
61
for_each_child_of_node(op->dev.of_node, dp) {
drivers/input/serio/i8042-sparcio.h
67
irq = op->archdata.irqs[0];
drivers/input/serio/i8042-sparcio.h
77
irq = op->archdata.irqs[0];
drivers/input/serio/i8042-sparcio.h
85
static void sparc_i8042_remove(struct platform_device *op)
drivers/input/touchscreen/cyttsp_spi.c
105
dev_dbg(dev, "%s: operation %d failed\n", __func__, op);
drivers/input/touchscreen/cyttsp_spi.c
36
u8 op, u16 reg, u8 *buf, int length)
drivers/input/touchscreen/cyttsp_spi.c
58
wr_buf[3] = op; /* r/~w */
drivers/input/touchscreen/cyttsp_spi.c
59
if (op == CY_SPI_WR_OP)
drivers/input/touchscreen/cyttsp_spi.c
71
switch (op) {
drivers/input/touchscreen/cyttsp_spi.c
87
dev_err(dev, "%s: bad operation code=%d\n", __func__, op);
drivers/input/touchscreen/cyttsp_spi.c
94
__func__, retval, xfer[1].len, op);
drivers/iommu/generic_pt/iommu_pt.h
67
#define DOMAIN_NS(op) CONCATENATE(CONCATENATE(pt_iommu_, PTPFX), op)
drivers/iommu/intel/iommu.h
363
#define IOMMU_WAIT_OP(iommu, offset, op, cond, sts) \
drivers/iommu/intel/iommu.h
367
sts = op(iommu->reg + offset); \
drivers/iommu/iommufd/ioas.c
593
if (cmd->op == IOMMU_OPTION_OP_GET) {
drivers/iommu/iommufd/ioas.c
597
if (cmd->op == IOMMU_OPTION_OP_SET) {
drivers/iommu/iommufd/ioas.c
624
if (cmd->op == IOMMU_OPTION_OP_GET) {
drivers/iommu/iommufd/ioas.c
628
if (cmd->op == IOMMU_OPTION_OP_SET) {
drivers/iommu/iommufd/iommufd_test.h
84
__u32 op;
drivers/iommu/iommufd/main.c
505
const struct iommufd_ioctl_op *op;
drivers/iommu/iommufd/main.c
522
op = &iommufd_ioctl_ops[nr - IOMMUFD_CMD_BASE];
drivers/iommu/iommufd/main.c
523
if (op->ioctl_num != cmd)
drivers/iommu/iommufd/main.c
525
if (ucmd.user_size < op->min_size)
drivers/iommu/iommufd/main.c
529
ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer,
drivers/iommu/iommufd/main.c
533
ret = op->execute(&ucmd);
drivers/iommu/iommufd/selftest.c
1094
if (cmd->op == IOMMU_TEST_OP_MOCK_DOMAIN_FLAGS)
drivers/iommu/iommufd/selftest.c
2096
switch (cmd->op) {
drivers/iommu/iommufd/vfio_compat.c
130
switch (cmd->op) {
drivers/isdn/hardware/mISDN/avmfritz.c
872
switch (cq->op) {
drivers/isdn/hardware/mISDN/avmfritz.c
874
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/avmfritz.c
888
pr_info("%s: %s unknown Op %x\n", fc->name, __func__, cq->op);
drivers/isdn/hardware/mISDN/hfcmulti.c
3539
switch (cq->op) {
drivers/isdn/hardware/mISDN/hfcmulti.c
3542
cq->op |= MISDN_CTRL_HFC_OP | MISDN_CTRL_HW_FEATURES_OP;
drivers/isdn/hardware/mISDN/hfcmulti.c
4098
switch (cq->op) {
drivers/isdn/hardware/mISDN/hfcmulti.c
4100
cq->op = MISDN_CTRL_HFC_OP | MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/hfcmulti.c
4135
__func__, cq->op);
drivers/isdn/hardware/mISDN/hfcpci.c
1787
switch (cq->op) {
drivers/isdn/hardware/mISDN/hfcpci.c
1789
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_CONNECT |
drivers/isdn/hardware/mISDN/hfcpci.c
1872
__func__, cq->op);
drivers/isdn/hardware/mISDN/hfcsusb.c
502
hw->name, __func__, (cq->op), (cq->channel));
drivers/isdn/hardware/mISDN/hfcsusb.c
504
switch (cq->op) {
drivers/isdn/hardware/mISDN/hfcsusb.c
506
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_CONNECT |
drivers/isdn/hardware/mISDN/hfcsusb.c
511
hw->name, __func__, cq->op);
drivers/isdn/hardware/mISDN/mISDNipac.c
1502
switch (cq->op) {
drivers/isdn/hardware/mISDN/mISDNipac.c
1504
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/mISDNipac.c
1518
pr_info("%s: unknown CTRL OP %x\n", ipac->name, cq->op);
drivers/isdn/hardware/mISDN/netjet.c
826
switch (cq->op) {
drivers/isdn/hardware/mISDN/netjet.c
828
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/netjet.c
842
pr_info("%s: %s unknown Op %x\n", card->name, __func__, cq->op);
drivers/isdn/hardware/mISDN/speedfax.c
213
switch (cq->op) {
drivers/isdn/hardware/mISDN/speedfax.c
215
cq->op = MISDN_CTRL_LOOP | MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/speedfax.c
229
pr_info("%s: unknown Op %x\n", sf->name, cq->op);
drivers/isdn/hardware/mISDN/w6692.c
1009
switch (cq->op) {
drivers/isdn/hardware/mISDN/w6692.c
1011
cq->op = MISDN_CTRL_L1_TIMER3;
drivers/isdn/hardware/mISDN/w6692.c
1017
pr_info("%s: unknown CTRL OP %x\n", card->name, cq->op);
drivers/isdn/mISDN/dsp_cmx.c
356
cq.op = message;
drivers/isdn/mISDN/dsp_core.c
223
cq.op = MISDN_CTRL_RX_OFF;
drivers/isdn/mISDN/dsp_core.c
268
cq.op = MISDN_CTRL_FILL_EMPTY;
drivers/isdn/mISDN/dsp_core.c
639
cq.op = MISDN_CTRL_GETOP;
drivers/isdn/mISDN/dsp_core.c
645
if (cq.op & MISDN_CTRL_RX_OFF)
drivers/isdn/mISDN/dsp_core.c
647
if (cq.op & MISDN_CTRL_FILL_EMPTY)
drivers/isdn/mISDN/dsp_core.c
651
if ((cq.op & MISDN_CTRL_HW_FEATURES_OP)) {
drivers/isdn/mISDN/dsp_core.c
652
cq.op = MISDN_CTRL_HW_FEATURES;
drivers/isdn/mISDN/dsp_hwec.c
104
cq.op = MISDN_CTRL_HFC_ECHOCAN_OFF;
drivers/isdn/mISDN/dsp_hwec.c
83
cq.op = MISDN_CTRL_HFC_ECHOCAN_ON;
drivers/isdn/mISDN/hwchannel.c
160
switch (cq->op) {
drivers/isdn/mISDN/hwchannel.c
162
cq->op = MISDN_CTRL_RX_BUFFER | MISDN_CTRL_FILL_EMPTY |
drivers/isdn/mISDN/hwchannel.c
192
pr_info("mISDN unhandled control %x operation\n", cq->op);
drivers/isdn/mISDN/l1oip_core.c
1180
switch (cq->op) {
drivers/isdn/mISDN/l1oip_core.c
1182
cq->op = MISDN_CTRL_HW_FEATURES_OP;
drivers/isdn/mISDN/l1oip_core.c
1194
__func__, cq->op);
drivers/isdn/mISDN/l1oip_core.c
939
switch (cq->op) {
drivers/isdn/mISDN/l1oip_core.c
941
cq->op = MISDN_CTRL_SETPEER | MISDN_CTRL_UNSETPEER
drivers/isdn/mISDN/l1oip_core.c
971
__func__, cq->op);
drivers/leds/led-class-flash.c
16
#define has_flash_op(fled_cdev, op) \
drivers/leds/led-class-flash.c
17
(fled_cdev && fled_cdev->ops->op)
drivers/leds/led-class-flash.c
19
#define call_flash_op(fled_cdev, op, args...) \
drivers/leds/led-class-flash.c
20
((has_flash_op(fled_cdev, op)) ? \
drivers/leds/led-class-flash.c
21
(fled_cdev->ops->op(fled_cdev, args)) : \
drivers/macintosh/adb-iop.c
105
u8 op = amsg->cmd & OP_MASK;
drivers/macintosh/adb-iop.c
114
if (op == TALK && ((1 << addr) & autopoll_devs))
drivers/mailbox/mtk-adsp-mailbox.c
40
u32 op = readl(priv->va_mboxreg + priv->cfg->set_out);
drivers/mailbox/mtk-adsp-mailbox.c
42
writel(op, priv->va_mboxreg + priv->cfg->clr_out);
drivers/md/bcache/btree.c
1005
struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.c
1021
b = mca_alloc(c, op, k, level);
drivers/md/bcache/btree.c
1119
struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.c
1136
b = mca_alloc(c, op, &k.key, level);
drivers/md/bcache/btree.c
1163
struct btree_op *op, int level,
drivers/md/bcache/btree.c
1166
return __bch_btree_node_alloc(c, op, level, op != NULL, parent);
drivers/md/bcache/btree.c
1170
struct btree_op *op)
drivers/md/bcache/btree.c
1172
struct btree *n = bch_btree_node_alloc(b->c, op, b->level, b->parent);
drivers/md/bcache/btree.c
1203
static int btree_check_reserve(struct btree *b, struct btree_op *op)
drivers/md/bcache/btree.c
1212
if (op)
drivers/md/bcache/btree.c
1213
prepare_to_wait(&c->btree_cache_wait, &op->wait,
drivers/md/bcache/btree.c
1221
return mca_cannibalize_lock(b->c, op);
drivers/md/bcache/btree.c
1354
static int bch_btree_insert_node(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
1359
static int btree_gc_coalesce(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
1489
bch_btree_insert_node(b, op, &keylist, NULL, NULL);
drivers/md/bcache/btree.c
1530
static int btree_gc_rewrite_node(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
1558
bch_btree_insert_node(b, op, &keys, NULL, NULL);
drivers/md/bcache/btree.c
1606
static int btree_gc_recurse(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
1625
r->b = bch_btree_node_get(b->c, op, k, b->level - 1,
drivers/md/bcache/btree.c
1634
ret = btree_gc_coalesce(b, op, gc, r);
drivers/md/bcache/btree.c
1645
ret = btree_gc_rewrite_node(b, op, last->b);
drivers/md/bcache/btree.c
1651
ret = btree_gc_recurse(last->b, op, writes, gc);
drivers/md/bcache/btree.c
1696
static int bch_btree_gc_root(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
1721
ret = btree_gc_recurse(b, op, writes, gc);
drivers/md/bcache/btree.c
1831
struct btree_op op;
drivers/md/bcache/btree.c
1838
bch_btree_op_init(&op, SHRT_MAX);
drivers/md/bcache/btree.c
1844
ret = bcache_btree_root(gc_root, c, &op, &writes, &stats);
drivers/md/bcache/btree.c
1913
static int bch_btree_check_recurse(struct btree *b, struct btree_op *op)
drivers/md/bcache/btree.c
1940
ret = bcache_btree(check_recurse, p, b, op);
drivers/md/bcache/btree.c
2005
struct btree_op op;
drivers/md/bcache/btree.c
2009
bch_btree_op_init(&op, 0);
drivers/md/bcache/btree.c
2010
ret = bcache_btree(check_recurse, p, c->root, &op);
drivers/md/bcache/btree.c
2019
finish_wait(&c->btree_cache_wait, &(&op)->wait);
drivers/md/bcache/btree.c
2195
static bool bch_btree_insert_keys(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2229
op->insert_collision = true;
drivers/md/bcache/btree.c
2237
static int btree_split(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2250
if (btree_check_reserve(b, op)) {
drivers/md/bcache/btree.c
2257
n1 = btree_node_alloc_replacement(b, op);
drivers/md/bcache/btree.c
2269
n2 = bch_btree_node_alloc(b->c, op, b->level, b->parent);
drivers/md/bcache/btree.c
2274
n3 = bch_btree_node_alloc(b->c, op, b->level + 1, NULL);
drivers/md/bcache/btree.c
2282
bch_btree_insert_keys(n1, op, insert_keys, replace_key);
drivers/md/bcache/btree.c
2314
bch_btree_insert_keys(n1, op, insert_keys, replace_key);
drivers/md/bcache/btree.c
2325
bch_btree_insert_keys(n3, op, &parent_keys, NULL);
drivers/md/bcache/btree.c
2342
bch_btree_insert_node(b->parent, op, &parent_keys, NULL, NULL);
drivers/md/bcache/btree.c
2371
static int bch_btree_insert_node(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2395
if (bch_btree_insert_keys(b, op, insert_keys, replace_key)) {
drivers/md/bcache/btree.c
2410
op->lock = b->c->root->level + 1;
drivers/md/bcache/btree.c
2412
} else if (op->lock <= b->c->root->level) {
drivers/md/bcache/btree.c
2413
op->lock = b->c->root->level + 1;
drivers/md/bcache/btree.c
2417
int ret = btree_split(b, op, insert_keys, replace_key);
drivers/md/bcache/btree.c
2427
int bch_btree_insert_check_key(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2434
bool upgrade = op->lock == -1;
drivers/md/bcache/btree.c
2444
op->lock = b->level;
drivers/md/bcache/btree.c
2456
ret = bch_btree_insert_node(b, op, &insert, NULL, NULL);
drivers/md/bcache/btree.c
2466
struct btree_op op;
drivers/md/bcache/btree.c
2474
struct btree_insert_op *op = container_of(b_op,
drivers/md/bcache/btree.c
2475
struct btree_insert_op, op);
drivers/md/bcache/btree.c
2477
int ret = bch_btree_insert_node(b, &op->op, op->keys,
drivers/md/bcache/btree.c
2478
op->journal_ref, op->replace_key);
drivers/md/bcache/btree.c
2479
if (ret && !bch_keylist_empty(op->keys))
drivers/md/bcache/btree.c
2488
struct btree_insert_op op;
drivers/md/bcache/btree.c
2494
bch_btree_op_init(&op.op, 0);
drivers/md/bcache/btree.c
2495
op.keys = keys;
drivers/md/bcache/btree.c
2496
op.journal_ref = journal_ref;
drivers/md/bcache/btree.c
2497
op.replace_key = replace_key;
drivers/md/bcache/btree.c
2500
op.op.lock = 0;
drivers/md/bcache/btree.c
2501
ret = bch_btree_map_leaf_nodes(&op.op, c,
drivers/md/bcache/btree.c
2513
} else if (op.op.insert_collision)
drivers/md/bcache/btree.c
2545
static int bch_btree_map_nodes_recurse(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2560
op, from, fn, flags);
drivers/md/bcache/btree.c
2569
ret = fn(op, b);
drivers/md/bcache/btree.c
2574
int __bch_btree_map_nodes(struct btree_op *op, struct cache_set *c,
drivers/md/bcache/btree.c
2577
return bcache_btree_root(map_nodes_recurse, c, op, from, fn, flags);
drivers/md/bcache/btree.c
2580
int bch_btree_map_keys_recurse(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.c
2593
? fn(op, b, k)
drivers/md/bcache/btree.c
2595
b, op, from, fn, flags);
drivers/md/bcache/btree.c
2603
ret = fn(op, b, &KEY(KEY_INODE(&b->key),
drivers/md/bcache/btree.c
2609
int bch_btree_map_keys(struct btree_op *op, struct cache_set *c,
drivers/md/bcache/btree.c
2612
return bcache_btree_root(map_keys_recurse, c, op, from, fn, flags);
drivers/md/bcache/btree.c
2634
struct btree_op op;
drivers/md/bcache/btree.c
2641
static int refill_keybuf_fn(struct btree_op *op, struct btree *b,
drivers/md/bcache/btree.c
2644
struct refill *refill = container_of(op, struct refill, op);
drivers/md/bcache/btree.c
2693
bch_btree_op_init(&refill.op, -1);
drivers/md/bcache/btree.c
2699
bch_btree_map_keys(&refill.op, c, &buf->last_scanned,
drivers/md/bcache/btree.c
867
static int mca_cannibalize_lock(struct cache_set *c, struct btree_op *op)
drivers/md/bcache/btree.c
873
if (op)
drivers/md/bcache/btree.c
874
prepare_to_wait(&c->btree_cache_wait, &op->wait,
drivers/md/bcache/btree.c
884
static struct btree *mca_cannibalize(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.c
891
if (mca_cannibalize_lock(c, op))
drivers/md/bcache/btree.c
922
static struct btree *mca_alloc(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.c
986
b = mca_cannibalize(c, op, k);
drivers/md/bcache/btree.h
241
static inline void bch_btree_op_init(struct btree_op *op, int write_lock_level)
drivers/md/bcache/btree.h
243
memset(op, 0, sizeof(struct btree_op));
drivers/md/bcache/btree.h
244
init_wait(&op->wait);
drivers/md/bcache/btree.h
245
op->lock = write_lock_level;
drivers/md/bcache/btree.h
268
struct btree *__bch_btree_node_alloc(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.h
271
struct btree *bch_btree_node_get(struct cache_set *c, struct btree_op *op,
drivers/md/bcache/btree.h
275
int bch_btree_insert_check_key(struct btree *b, struct btree_op *op,
drivers/md/bcache/btree.h
328
#define bcache_btree(fn, key, b, op, ...) \
drivers/md/bcache/btree.h
331
bool _w = l <= (op)->lock; \
drivers/md/bcache/btree.h
332
struct btree *_child = bch_btree_node_get((b)->c, op, key, l, \
drivers/md/bcache/btree.h
335
_r = bch_btree_ ## fn(_child, op, ##__VA_ARGS__); \
drivers/md/bcache/btree.h
348
#define bcache_btree_root(fn, c, op, ...) \
drivers/md/bcache/btree.h
353
bool _w = insert_lock(op, _b); \
drivers/md/bcache/btree.h
356
_w == insert_lock(op, _b)) { \
drivers/md/bcache/btree.h
357
_r = bch_btree_ ## fn(_b, op, ##__VA_ARGS__); \
drivers/md/bcache/btree.h
365
finish_wait(&(c)->btree_cache_wait, &(op)->wait); \
drivers/md/bcache/btree.h
378
int __bch_btree_map_nodes(struct btree_op *op, struct cache_set *c,
drivers/md/bcache/btree.h
381
static inline int bch_btree_map_nodes(struct btree_op *op, struct cache_set *c,
drivers/md/bcache/btree.h
384
return __bch_btree_map_nodes(op, c, from, fn, MAP_ALL_NODES);
drivers/md/bcache/btree.h
387
static inline int bch_btree_map_leaf_nodes(struct btree_op *op,
drivers/md/bcache/btree.h
392
return __bch_btree_map_nodes(op, c, from, fn, MAP_LEAF_NODES);
drivers/md/bcache/btree.h
395
typedef int (btree_map_keys_fn)(struct btree_op *op, struct btree *b,
drivers/md/bcache/btree.h
397
int bch_btree_map_keys(struct btree_op *op, struct cache_set *c,
drivers/md/bcache/btree.h
399
int bch_btree_map_keys_recurse(struct btree *b, struct btree_op *op,
drivers/md/bcache/movinggc.c
101
op->write_prio = 1;
drivers/md/bcache/movinggc.c
102
op->bio = &io->bio.bio;
drivers/md/bcache/movinggc.c
104
op->writeback = KEY_DIRTY(&io->w->key);
drivers/md/bcache/movinggc.c
105
op->csum = KEY_CSUM(&io->w->key);
drivers/md/bcache/movinggc.c
107
bkey_copy(&op->replace_key, &io->w->key);
drivers/md/bcache/movinggc.c
108
op->replace = true;
drivers/md/bcache/movinggc.c
110
closure_call(&op->cl, bch_data_insert, NULL, cl);
drivers/md/bcache/movinggc.c
113
continue_at(cl, write_moving_finish, op->wq);
drivers/md/bcache/movinggc.c
121
bch_submit_bbio(bio, io->op.c, &io->w->key, 0);
drivers/md/bcache/movinggc.c
123
continue_at(cl, write_moving, io->op.wq);
drivers/md/bcache/movinggc.c
156
io->op.inode = KEY_INODE(&w->key);
drivers/md/bcache/movinggc.c
157
io->op.c = c;
drivers/md/bcache/movinggc.c
158
io->op.wq = c->moving_gc_wq;
drivers/md/bcache/movinggc.c
18
struct data_insert_op op;
drivers/md/bcache/movinggc.c
52
if (io->op.replace_collision)
drivers/md/bcache/movinggc.c
55
bch_keybuf_del(&io->op.c->moving_gc_keys, io->w);
drivers/md/bcache/movinggc.c
57
up(&io->op.c->moving_in_flight);
drivers/md/bcache/movinggc.c
69
io->op.status = bio->bi_status;
drivers/md/bcache/movinggc.c
71
ptr_stale(io->op.c, &b->key, 0)) {
drivers/md/bcache/movinggc.c
72
io->op.status = BLK_STS_IOERR;
drivers/md/bcache/movinggc.c
75
bch_bbio_endio(io->op.c, bio, bio->bi_status, "reading data to move");
drivers/md/bcache/movinggc.c
95
struct data_insert_op *op = &io->op;
drivers/md/bcache/movinggc.c
97
if (!op->status) {
drivers/md/bcache/request.c
110
struct data_insert_op *op = container_of(cl, struct data_insert_op, cl);
drivers/md/bcache/request.c
111
struct bio *bio = op->bio;
drivers/md/bcache/request.c
120
if (bch_keylist_realloc(&op->insert_keys, 2, op->c))
drivers/md/bcache/request.c
126
bch_keylist_add(&op->insert_keys,
drivers/md/bcache/request.c
127
&KEY(op->inode,
drivers/md/bcache/request.c
132
op->insert_data_done = true;
drivers/md/bcache/request.c
136
continue_at(cl, bch_data_insert_keys, op->wq);
drivers/md/bcache/request.c
141
closure_type(op, struct data_insert_op, cl);
drivers/md/bcache/request.c
152
struct bkey *src = op->insert_keys.keys, *dst = op->insert_keys.keys;
drivers/md/bcache/request.c
154
while (src != op->insert_keys.top) {
drivers/md/bcache/request.c
164
op->insert_keys.top = dst;
drivers/md/bcache/request.c
172
struct data_insert_op *op = container_of(cl, struct data_insert_op, cl);
drivers/md/bcache/request.c
176
if (op->writeback)
drivers/md/bcache/request.c
177
op->status = bio->bi_status;
drivers/md/bcache/request.c
178
else if (!op->replace)
drivers/md/bcache/request.c
179
set_closure_fn(cl, bch_data_insert_error, op->wq);
drivers/md/bcache/request.c
184
bch_bbio_endio(op->c, bio, bio->bi_status, "writing data to cache");
drivers/md/bcache/request.c
189
closure_type(op, struct data_insert_op, cl);
drivers/md/bcache/request.c
190
struct bio *bio = op->bio, *n;
drivers/md/bcache/request.c
192
if (op->bypass)
drivers/md/bcache/request.c
195
if (atomic_sub_return(bio_sectors(bio), &op->c->sectors_to_gc) < 0)
drivers/md/bcache/request.c
196
wake_up_gc(op->c);
drivers/md/bcache/request.c
207
struct bio_set *split = &op->c->bio_split;
drivers/md/bcache/request.c
210
if (bch_keylist_realloc(&op->insert_keys,
drivers/md/bcache/request.c
211
3 + (op->csum ? 1 : 0),
drivers/md/bcache/request.c
212
op->c)) {
drivers/md/bcache/request.c
213
continue_at(cl, bch_data_insert_keys, op->wq);
drivers/md/bcache/request.c
217
k = op->insert_keys.top;
drivers/md/bcache/request.c
219
SET_KEY_INODE(k, op->inode);
drivers/md/bcache/request.c
222
if (!bch_alloc_sectors(op->c, k, bio_sectors(bio),
drivers/md/bcache/request.c
223
op->write_point, op->write_prio,
drivers/md/bcache/request.c
224
op->writeback))
drivers/md/bcache/request.c
232
if (op->writeback) {
drivers/md/bcache/request.c
236
SET_GC_MARK(PTR_BUCKET(op->c, k, i),
drivers/md/bcache/request.c
240
SET_KEY_CSUM(k, op->csum);
drivers/md/bcache/request.c
245
bch_keylist_push(&op->insert_keys);
drivers/md/bcache/request.c
248
bch_submit_bbio(n, op->c, k, 0);
drivers/md/bcache/request.c
251
op->insert_data_done = true;
drivers/md/bcache/request.c
252
continue_at(cl, bch_data_insert_keys, op->wq);
drivers/md/bcache/request.c
256
BUG_ON(op->writeback);
drivers/md/bcache/request.c
264
if (!op->replace) {
drivers/md/bcache/request.c
271
op->bypass = true;
drivers/md/bcache/request.c
278
op->insert_data_done = true;
drivers/md/bcache/request.c
281
if (!bch_keylist_empty(&op->insert_keys))
drivers/md/bcache/request.c
282
continue_at(cl, bch_data_insert_keys, op->wq);
drivers/md/bcache/request.c
310
closure_type(op, struct data_insert_op, cl);
drivers/md/bcache/request.c
312
trace_bcache_write(op->c, op->inode, op->bio,
drivers/md/bcache/request.c
313
op->writeback, op->bypass);
drivers/md/bcache/request.c
315
bch_keylist_init(&op->insert_keys);
drivers/md/bcache/request.c
316
bio_get(op->bio);
drivers/md/bcache/request.c
495
struct btree_op op;
drivers/md/bcache/request.c
527
static int cache_lookup_fn(struct btree_op *op, struct btree *b, struct bkey *k)
drivers/md/bcache/request.c
529
struct search *s = container_of(op, struct search, op);
drivers/md/bcache/request.c
599
bch_btree_op_init(&s->op, -1);
drivers/md/bcache/request.c
60
closure_type(op, struct data_insert_op, cl);
drivers/md/bcache/request.c
601
ret = bch_btree_map_keys(&s->op, s->iop.c,
drivers/md/bcache/request.c
62
struct bkey *replace_key = op->replace ? &op->replace_key : NULL;
drivers/md/bcache/request.c
65
if (!op->replace)
drivers/md/bcache/request.c
66
journal_ref = bch_journal(op->c, &op->insert_keys,
drivers/md/bcache/request.c
67
op->flush_journal ? cl : NULL);
drivers/md/bcache/request.c
69
ret = bch_btree_insert(op->c, &op->insert_keys,
drivers/md/bcache/request.c
72
op->replace_collision = true;
drivers/md/bcache/request.c
74
op->status = BLK_STS_RESOURCE;
drivers/md/bcache/request.c
75
op->insert_data_done = true;
drivers/md/bcache/request.c
81
if (!op->insert_data_done) {
drivers/md/bcache/request.c
82
continue_at(cl, bch_data_insert_start, op->wq);
drivers/md/bcache/request.c
86
bch_keylist_free(&op->insert_keys);
drivers/md/bcache/request.c
917
ret = bch_btree_insert_check_key(b, &s->op, &s->iop.replace_key);
drivers/md/bcache/sysfs.c
616
struct btree_op op;
drivers/md/bcache/sysfs.c
623
struct bset_stats_op *op = container_of(b_op, struct bset_stats_op, op);
drivers/md/bcache/sysfs.c
625
op->nodes++;
drivers/md/bcache/sysfs.c
626
bch_btree_keys_stats(&b->keys, &op->stats);
drivers/md/bcache/sysfs.c
633
struct bset_stats_op op;
drivers/md/bcache/sysfs.c
636
memset(&op, 0, sizeof(op));
drivers/md/bcache/sysfs.c
637
bch_btree_op_init(&op.op, -1);
drivers/md/bcache/sysfs.c
639
ret = bch_btree_map_nodes(&op.op, c, &ZERO_KEY, bch_btree_bset_stats);
drivers/md/bcache/sysfs.c
651
op.nodes,
drivers/md/bcache/sysfs.c
652
op.stats.sets_written, op.stats.sets_unwritten,
drivers/md/bcache/sysfs.c
653
op.stats.bytes_written, op.stats.bytes_unwritten,
drivers/md/bcache/sysfs.c
654
op.stats.floats, op.stats.failed);
drivers/md/bcache/writeback.c
1002
if (KEY_INODE(k) != op.inode)
drivers/md/bcache/writeback.c
1004
sectors_dirty_init_fn(&op.op, c->root, k);
drivers/md/bcache/writeback.c
847
struct btree_op op;
drivers/md/bcache/writeback.c
855
struct sectors_dirty_init *op = container_of(_op,
drivers/md/bcache/writeback.c
856
struct sectors_dirty_init, op);
drivers/md/bcache/writeback.c
857
if (KEY_INODE(k) > op->inode)
drivers/md/bcache/writeback.c
864
op->count++;
drivers/md/bcache/writeback.c
865
if (!(op->count % INIT_KEYS_EACH_TIME))
drivers/md/bcache/writeback.c
875
struct sectors_dirty_init op;
drivers/md/bcache/writeback.c
878
bch_btree_op_init(&op.op, -1);
drivers/md/bcache/writeback.c
879
op.inode = d->id;
drivers/md/bcache/writeback.c
880
op.count = 0;
drivers/md/bcache/writeback.c
885
&op.op,
drivers/md/bcache/writeback.c
886
&KEY(op.inode, 0, 0),
drivers/md/bcache/writeback.c
900
finish_wait(&c->btree_cache_wait, &(&op.op)->wait);
drivers/md/bcache/writeback.c
982
struct sectors_dirty_init op;
drivers/md/bcache/writeback.c
996
bch_btree_op_init(&op.op, -1);
drivers/md/bcache/writeback.c
997
op.inode = d->id;
drivers/md/bcache/writeback.c
998
op.count = 0;
drivers/md/dm-audit.c
21
const char *op)
drivers/md/dm-audit.c
32
audit_log_format(ab, "module=%s op=%s", dm_msg_prefix, op);
drivers/md/dm-audit.c
36
void dm_audit_log_ti(int audit_type, const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.c
46
ab = dm_audit_log_start(audit_type, dm_msg_prefix, op);
drivers/md/dm-audit.c
54
ab = dm_audit_log_start(audit_type, dm_msg_prefix, op);
drivers/md/dm-audit.c
69
void dm_audit_log_bio(const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.c
76
ab = dm_audit_log_start(AUDIT_DM_EVENT, dm_msg_prefix, op);
drivers/md/dm-audit.h
17
void dm_audit_log_bio(const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.h
24
void dm_audit_log_ti(int audit_type, const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.h
39
static inline void dm_audit_log_target(const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.h
42
dm_audit_log_ti(AUDIT_DM_EVENT, dm_msg_prefix, op, ti, result);
drivers/md/dm-audit.h
45
static inline void dm_audit_log_bio(const char *dm_msg_prefix, const char *op,
drivers/md/dm-audit.h
51
const char *op, struct dm_target *ti,
drivers/md/dm-bufio.c
1297
static void use_dmio(struct dm_buffer *b, enum req_op op, sector_t sector,
drivers/md/dm-bufio.c
1303
.bi_opf = op,
drivers/md/dm-bufio.c
1337
static void use_bio(struct dm_buffer *b, enum req_op op, sector_t sector,
drivers/md/dm-bufio.c
1347
use_dmio(b, op, sector, n_sectors, offset, ioprio);
drivers/md/dm-bufio.c
1350
bio_init_inline(bio, b->c->bdev, 1, op);
drivers/md/dm-bufio.c
1377
static void submit_io(struct dm_buffer *b, enum req_op op, unsigned short ioprio,
drivers/md/dm-bufio.c
1388
if (op != REQ_OP_WRITE) {
drivers/md/dm-bufio.c
1409
use_bio(b, op, sector, n_sectors, offset, ioprio);
drivers/md/dm-bufio.c
1411
use_dmio(b, op, sector, n_sectors, offset, ioprio);
drivers/md/dm-cache-background-tracker.c
131
switch (w->op) {
drivers/md/dm-cache-background-tracker.c
221
struct policy_work *op)
drivers/md/dm-cache-background-tracker.c
223
struct bt_work *w = container_of(op, struct bt_work, work);
drivers/md/dm-cache-background-tracker.h
78
void btracker_complete(struct background_tracker *b, struct policy_work *op);
drivers/md/dm-cache-policy-smq.c
1196
work.op = POLICY_WRITEBACK;
drivers/md/dm-cache-policy-smq.c
1227
work.op = POLICY_DEMOTE;
drivers/md/dm-cache-policy-smq.c
1267
work.op = POLICY_PROMOTE;
drivers/md/dm-cache-policy-smq.c
1476
switch (work->op) {
drivers/md/dm-cache-policy.h
31
enum policy_operation op;
drivers/md/dm-cache-target.c
1106
o_region.sector = from_oblock(mg->op->oblock) * cache->sectors_per_block;
drivers/md/dm-cache-target.c
1110
c_region.sector = from_cblock(mg->op->cblock) * cache->sectors_per_block;
drivers/md/dm-cache-target.c
1154
if (mg->op->op == POLICY_PROMOTE)
drivers/md/dm-cache-target.c
1155
remap_to_cache(mg->cache, bio, mg->op->cblock);
drivers/md/dm-cache-target.c
1178
struct policy_work *op = mg->op;
drivers/md/dm-cache-target.c
1179
dm_cblock_t cblock = op->cblock;
drivers/md/dm-cache-target.c
1182
update_stats(&cache->stats, op->op);
drivers/md/dm-cache-target.c
1184
switch (op->op) {
drivers/md/dm-cache-target.c
1186
clear_discard(cache, oblock_to_dblock(cache, op->oblock));
drivers/md/dm-cache-target.c
1187
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1210
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1217
policy_complete_background_work(cache->policy, op, success);
drivers/md/dm-cache-target.c
1247
struct policy_work *op = mg->op;
drivers/md/dm-cache-target.c
1249
switch (op->op) {
drivers/md/dm-cache-target.c
1251
r = dm_cache_insert_mapping(cache->cmd, op->cblock, op->oblock);
drivers/md/dm-cache-target.c
1264
r = dm_cache_remove_mapping(cache->cmd, op->cblock);
drivers/md/dm-cache-target.c
1349
struct policy_work *op = mg->op;
drivers/md/dm-cache-target.c
1350
bool is_policy_promote = (op->op == POLICY_PROMOTE);
drivers/md/dm-cache-target.c
1352
if ((!is_policy_promote && !is_dirty(cache, op->cblock)) ||
drivers/md/dm-cache-target.c
1353
is_discarded_oblock(cache, op->oblock)) {
drivers/md/dm-cache-target.c
1372
if (!optimisable_bio(mg->cache, mg->overwrite_bio, mg->op->oblock)) {
drivers/md/dm-cache-target.c
1376
bool rb = bio_detain_shared(mg->cache, mg->op->oblock, mg->overwrite_bio);
drivers/md/dm-cache-target.c
1412
build_key(mg->op->oblock, oblock_succ(mg->op->oblock), &key);
drivers/md/dm-cache-target.c
1433
static int mg_start(struct cache *cache, struct policy_work *op, struct bio *bio)
drivers/md/dm-cache-target.c
1438
policy_complete_background_work(cache->policy, op, false);
drivers/md/dm-cache-target.c
1444
mg->op = op;
drivers/md/dm-cache-target.c
1642
struct policy_work *op = NULL;
drivers/md/dm-cache-target.c
1644
r = policy_lookup_with_work(cache->policy, block, &cblock, data_dir, true, &op);
drivers/md/dm-cache-target.c
1652
if (r == -ENOENT && op) {
drivers/md/dm-cache-target.c
1654
BUG_ON(op->op != POLICY_PROMOTE);
drivers/md/dm-cache-target.c
1655
mg_start(cache, op, bio);
drivers/md/dm-cache-target.c
1880
struct policy_work *op;
drivers/md/dm-cache-target.c
1887
r = policy_get_background_work(cache->policy, b == IDLE, &op);
drivers/md/dm-cache-target.c
1897
r = mg_start(cache, op, NULL);
drivers/md/dm-cache-target.c
429
struct policy_work *op;
drivers/md/dm-cache-target.c
938
static void metadata_operation_failed(struct cache *cache, const char *op, int r)
drivers/md/dm-cache-target.c
941
cache_device_name(cache), op, r);
drivers/md/dm-cache-target.c
974
static void update_stats(struct cache_stats *stats, enum policy_operation op)
drivers/md/dm-cache-target.c
976
switch (op) {
drivers/md/dm-clone-target.c
225
static void __metadata_operation_failed(struct clone *clone, const char *op, int r)
drivers/md/dm-clone-target.c
228
clone_device_name(clone), op, r);
drivers/md/dm-ebs-target.c
100
if (op == REQ_OP_READ) {
drivers/md/dm-ebs-target.c
122
static int __ebs_rw_bio(struct ebs_c *ec, enum req_op op, struct bio *bio)
drivers/md/dm-ebs-target.c
129
rr = __ebs_rw_bvec(ec, op, &bv, &iter);
drivers/md/dm-ebs-target.c
65
static int __ebs_rw_bvec(struct ebs_c *ec, enum req_op op, struct bio_vec *bv,
drivers/md/dm-ebs-target.c
86
if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio))
drivers/md/dm-integrity.c
1413
unsigned int *metadata_offset, unsigned int total_size, int op)
drivers/md/dm-integrity.c
1435
if (op == TAG_READ) {
drivers/md/dm-integrity.c
1437
} else if (op == TAG_WRITE) {
drivers/md/dm-integrity.c
1616
if (dio->op == REQ_OP_WRITE || unlikely(dio->op == REQ_OP_DISCARD))
drivers/md/dm-integrity.c
1899
if (likely(dio->op != REQ_OP_DISCARD))
drivers/md/dm-integrity.c
1913
if (unlikely(dio->op == REQ_OP_DISCARD)) {
drivers/md/dm-integrity.c
1962
checksums_ptr - checksums, dio->op == REQ_OP_READ ? TAG_CMP : TAG_WRITE);
drivers/md/dm-integrity.c
2004
this_len, dio->op == REQ_OP_READ ? TAG_READ : TAG_WRITE);
drivers/md/dm-integrity.c
2060
dio->op = bio_op(bio);
drivers/md/dm-integrity.c
2071
if (unlikely(dio->op == REQ_OP_DISCARD)) {
drivers/md/dm-integrity.c
2092
dio->fua = dio->op == REQ_OP_WRITE && bio->bi_opf & REQ_FUA;
drivers/md/dm-integrity.c
2125
if (unlikely(ic->mode == 'R') && unlikely(dio->op != REQ_OP_READ))
drivers/md/dm-integrity.c
2155
if (likely(dio->op == REQ_OP_WRITE))
drivers/md/dm-integrity.c
2161
if (unlikely(dio->op == REQ_OP_READ)) {
drivers/md/dm-integrity.c
2199
if (likely(dio->op == REQ_OP_WRITE))
drivers/md/dm-integrity.c
2207
} else if (likely(dio->op == REQ_OP_WRITE))
drivers/md/dm-integrity.c
2211
if (likely(dio->op == REQ_OP_WRITE)) {
drivers/md/dm-integrity.c
2251
if (unlikely(dio->op == REQ_OP_READ))
drivers/md/dm-integrity.c
2256
if (likely(dio->op == REQ_OP_WRITE)) {
drivers/md/dm-integrity.c
2288
bool need_sync_io = ic->internal_hash && dio->op == REQ_OP_READ;
drivers/md/dm-integrity.c
2290
if (unlikely(dio->op == REQ_OP_DISCARD) && ic->mode != 'D')
drivers/md/dm-integrity.c
2309
if (ic->mode == 'J' && likely(dio->op != REQ_OP_DISCARD)) {
drivers/md/dm-integrity.c
2310
if (dio->op == REQ_OP_WRITE) {
drivers/md/dm-integrity.c
2409
if (ic->mode == 'J' && likely(dio->op == REQ_OP_DISCARD) && !discard_retried) {
drivers/md/dm-integrity.c
2435
if (ic->mode == 'B' && (dio->op == REQ_OP_WRITE || unlikely(dio->op == REQ_OP_DISCARD))) {
drivers/md/dm-integrity.c
2464
if (unlikely(dio->op == REQ_OP_DISCARD) && likely(ic->mode != 'D')) {
drivers/md/dm-integrity.c
2603
if (dio->op == REQ_OP_WRITE) {
drivers/md/dm-integrity.c
2755
if (dio->op == REQ_OP_READ && likely(*status == BLK_STS_OK) && likely(dio->bio_details.bi_iter.bi_size != 0)) {
drivers/md/dm-integrity.c
317
enum req_op op;
drivers/md/dm-integrity.c
544
const enum req_op op = opf & REQ_OP_MASK;
drivers/md/dm-integrity.c
556
if (op == REQ_OP_WRITE) {
drivers/md/dm-integrity.c
569
if (op == REQ_OP_READ) {
drivers/md/dm-io.c
319
const enum req_op op = opf & REQ_OP_MASK;
drivers/md/dm-io.c
324
if (op == REQ_OP_DISCARD)
drivers/md/dm-io.c
326
else if (op == REQ_OP_WRITE_ZEROES)
drivers/md/dm-io.c
328
if ((op == REQ_OP_DISCARD || op == REQ_OP_WRITE_ZEROES) &&
drivers/md/dm-io.c
343
switch (op) {
drivers/md/dm-io.c
360
if (op == REQ_OP_DISCARD || op == REQ_OP_WRITE_ZEROES) {
drivers/md/dm-kcopyd.c
361
enum req_op op;
drivers/md/dm-kcopyd.c
427
if (job->op == REQ_OP_READ ||
drivers/md/dm-kcopyd.c
528
if (op_is_write(job->op))
drivers/md/dm-kcopyd.c
540
if (op_is_write(job->op))
drivers/md/dm-kcopyd.c
544
job->op = REQ_OP_WRITE;
drivers/md/dm-kcopyd.c
559
.bi_opf = job->op,
drivers/md/dm-kcopyd.c
580
if (job->op == REQ_OP_READ)
drivers/md/dm-kcopyd.c
623
if (op_is_write(job->op))
drivers/md/dm-kcopyd.c
827
job->op = REQ_OP_READ;
drivers/md/dm-kcopyd.c
836
job->op = REQ_OP_WRITE_ZEROES;
drivers/md/dm-kcopyd.c
839
job->op = REQ_OP_WRITE;
drivers/md/dm-log.c
299
static int rw_header(struct log_c *lc, enum req_op op)
drivers/md/dm-log.c
301
lc->io_req.bi_opf = op;
drivers/md/dm-table.c
1191
enum dm_wrappedkey_op op;
drivers/md/dm-table.c
1224
switch (args->op) {
drivers/md/dm-table.c
1295
.op = DERIVE_SW_SECRET,
drivers/md/dm-table.c
1310
.op = IMPORT_KEY,
drivers/md/dm-table.c
1324
.op = GENERATE_KEY,
drivers/md/dm-table.c
1337
.op = PREPARE_KEY,
drivers/md/dm-thin.c
1089
struct discard_op op;
drivers/md/dm-thin.c
1091
begin_discard(&op, tc, discard_parent);
drivers/md/dm-thin.c
1116
issue_discard(&op, b, e);
drivers/md/dm-thin.c
1121
end_discard(&op, r);
drivers/md/dm-thin.c
1186
struct discard_op op;
drivers/md/dm-thin.c
1188
begin_discard(&op, tc, discard_parent);
drivers/md/dm-thin.c
1189
issue_discard(&op, m->data_block, data_end);
drivers/md/dm-thin.c
1190
end_discard(&op, 0);
drivers/md/dm-thin.c
2654
static void metadata_operation_failed(struct pool *pool, const char *op, int r)
drivers/md/dm-thin.c
2657
dm_device_name(pool->pool_md), op, r);
drivers/md/dm-thin.c
290
static void metadata_operation_failed(struct pool *pool, const char *op, int r);
drivers/md/dm-thin.c
388
static void begin_discard(struct discard_op *op, struct thin_c *tc, struct bio *parent)
drivers/md/dm-thin.c
392
op->tc = tc;
drivers/md/dm-thin.c
393
blk_start_plug(&op->plug);
drivers/md/dm-thin.c
394
op->parent_bio = parent;
drivers/md/dm-thin.c
395
op->bio = NULL;
drivers/md/dm-thin.c
398
static void issue_discard(struct discard_op *op, dm_block_t data_b, dm_block_t data_e)
drivers/md/dm-thin.c
400
struct thin_c *tc = op->tc;
drivers/md/dm-thin.c
404
__blkdev_issue_discard(tc->pool_dev->bdev, s, len, GFP_NOIO, &op->bio);
drivers/md/dm-thin.c
407
static void end_discard(struct discard_op *op, int r)
drivers/md/dm-thin.c
409
if (op->bio) {
drivers/md/dm-thin.c
414
bio_chain(op->bio, op->parent_bio);
drivers/md/dm-thin.c
415
op->bio->bi_opf = REQ_OP_DISCARD;
drivers/md/dm-thin.c
416
submit_bio(op->bio);
drivers/md/dm-thin.c
419
blk_finish_plug(&op->plug);
drivers/md/dm-thin.c
425
if (r && !op->parent_bio->bi_status)
drivers/md/dm-thin.c
426
op->parent_bio->bi_status = errno_to_blk_status(r);
drivers/md/dm-thin.c
427
bio_endio(op->parent_bio);
drivers/md/dm-zoned-metadata.c
2042
unsigned int chunk, enum req_op op)
drivers/md/dm-zoned-metadata.c
2061
if (op != REQ_OP_WRITE)
drivers/md/dm-zoned-metadata.c
735
static int dmz_rdwr_block(struct dmz_dev *dev, enum req_op op,
drivers/md/dm-zoned-metadata.c
747
bio = bio_alloc(dev->bdev, 1, op | REQ_SYNC | REQ_META | REQ_PRIO,
drivers/md/dm-zoned.h
249
unsigned int chunk, enum req_op op);
drivers/md/dm.c
1328
enum req_op op = bio_op(bio);
drivers/md/dm.c
1330
BUG_ON(op_is_zone_mgmt(op));
drivers/md/dm.c
1331
BUG_ON(op == REQ_OP_WRITE);
drivers/md/dm.c
1332
BUG_ON(op == REQ_OP_WRITE_ZEROES);
drivers/md/dm.c
1333
BUG_ON(op == REQ_OP_ZONE_APPEND);
drivers/md/persistent-data/dm-space-map-metadata.c
195
static int commit_bop(struct sm_metadata *smm, struct block_op *op)
drivers/md/persistent-data/dm-space-map-metadata.c
200
switch (op->type) {
drivers/md/persistent-data/dm-space-map-metadata.c
202
r = sm_ll_inc(&smm->ll, op->b, op->e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
206
r = sm_ll_dec(&smm->ll, op->b, op->e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
317
struct block_op *op = smm->uncommitted.bops + i;
drivers/md/persistent-data/dm-space-map-metadata.c
319
if (b < op->b || b >= op->e)
drivers/md/persistent-data/dm-space-map-metadata.c
322
switch (op->type) {
drivers/md/persistent-data/dm-space-map-metadata.c
358
struct block_op *op = smm->uncommitted.bops + i;
drivers/md/persistent-data/dm-space-map-metadata.c
360
if (b < op->b || b >= op->e)
drivers/md/persistent-data/dm-space-map-metadata.c
363
switch (op->type) {
drivers/md/raid10.c
2586
int sectors, struct page *page, enum req_op op)
drivers/md/raid10.c
2589
(op == REQ_OP_READ || test_bit(WriteErrorSeen, &rdev->flags)))
drivers/md/raid10.c
2591
if (sync_page_io(rdev, sector, sectors << 9, page, op, false))
drivers/md/raid10.c
2594
if (op == REQ_OP_WRITE) {
drivers/md/raid5.c
1159
enum req_op op;
drivers/md/raid5.c
1167
op = REQ_OP_WRITE;
drivers/md/raid5.c
1171
op = REQ_OP_DISCARD;
drivers/md/raid5.c
1173
op = REQ_OP_READ;
drivers/md/raid5.c
1176
op = REQ_OP_WRITE;
drivers/md/raid5.c
1190
if (op_is_write(op)) {
drivers/md/raid5.c
1215
while (op_is_write(op) && rdev &&
drivers/md/raid5.c
1249
bio_init(bi, rdev->bdev, &dev->vec, 1, op | op_flags);
drivers/md/raid5.c
1250
bi->bi_end_io = op_is_write(op)
drivers/md/raid5.c
1273
if (!op_is_write(op) &&
drivers/md/raid5.c
1291
if (op == REQ_OP_DISCARD)
drivers/md/raid5.c
1297
if (should_defer && op_is_write(op))
drivers/md/raid5.c
1305
bio_init(rbi, rrdev->bdev, &dev->rvec, 1, op | op_flags);
drivers/md/raid5.c
1306
BUG_ON(!op_is_write(op));
drivers/md/raid5.c
1334
if (op == REQ_OP_DISCARD)
drivers/md/raid5.c
1337
if (should_defer && op_is_write(op))
drivers/media/cec/core/cec-pin-error-inj.c
191
} else if (!kstrtou8(token, 0, &op)) {
drivers/media/cec/core/cec-pin-error-inj.c
193
error = pin->error_inj + op;
drivers/media/cec/core/cec-pin-error-inj.c
194
args = pin->error_inj_args[op];
drivers/media/cec/core/cec-pin-error-inj.c
85
u8 op;
drivers/media/cec/core/cec-pin-priv.h
15
#define call_pin_op(pin, op, arg...) \
drivers/media/cec/core/cec-pin-priv.h
16
((pin && pin->ops->op && !pin->adap->devnode.unregistered) ? \
drivers/media/cec/core/cec-pin-priv.h
17
pin->ops->op(pin->adap, ## arg) : 0)
drivers/media/cec/core/cec-pin-priv.h
19
#define call_void_pin_op(pin, op, arg...) \
drivers/media/cec/core/cec-pin-priv.h
21
if (pin && pin->ops->op && \
drivers/media/cec/core/cec-pin-priv.h
23
pin->ops->op(pin->adap, ## arg); \
drivers/media/cec/core/cec-priv.h
20
#define call_op(adap, op, arg...) \
drivers/media/cec/core/cec-priv.h
21
((adap->ops->op && !adap->devnode.unregistered) ? \
drivers/media/cec/core/cec-priv.h
22
adap->ops->op(adap, ## arg) : 0)
drivers/media/cec/core/cec-priv.h
24
#define call_void_op(adap, op, arg...) \
drivers/media/cec/core/cec-priv.h
26
if (adap->ops->op && !adap->devnode.unregistered) \
drivers/media/cec/core/cec-priv.h
27
adap->ops->op(adap, ## arg); \
drivers/media/common/b2c2/flexcop-eeprom.c
108
flexcop_access_op_t op, u16 addr, u8 *buf, u16 len, int retries)
drivers/media/common/b2c2/flexcop-eeprom.c
113
ret = fc->i2c_request(&fc->fc_i2c_adap[1], op, chipaddr,
drivers/media/common/b2c2/flexcop-i2c.c
108
flexcop_access_op_t op, u8 chipaddr,
drivers/media/common/b2c2/flexcop-i2c.c
121
op == FC_READ ? "rd" : "wr",
drivers/media/common/b2c2/flexcop-i2c.c
125
r100.tw_sm_c_100.twoWS_rw = op;
drivers/media/common/b2c2/flexcop-i2c.c
131
if (i2c->no_base_addr && len == 0 && op == FC_WRITE) {
drivers/media/common/b2c2/flexcop-i2c.c
144
if (op == FC_READ)
drivers/media/common/b2c2/flexcop-i2c.c
158
op == FC_READ ? "rd" : "wr",
drivers/media/common/saa7146/saa7146_i2c.c
103
m[i].buf[j] = (le32_to_cpu(op[op_count/3]) >> ((3-(op_count%3))*8));
drivers/media/common/saa7146/saa7146_i2c.c
28
static int saa7146_i2c_msg_prepare(const struct i2c_msg *m, int num, __le32 *op)
drivers/media/common/saa7146/saa7146_i2c.c
52
memset(op,0,sizeof(__le32)*mem);
drivers/media/common/saa7146/saa7146_i2c.c
59
op[h1] |= cpu_to_le32( (u8)addr << ((3-h2)*8));
drivers/media/common/saa7146/saa7146_i2c.c
60
op[h1] |= cpu_to_le32(SAA7146_I2C_START << ((3-h2)*2));
drivers/media/common/saa7146/saa7146_i2c.c
67
op[h1] |= cpu_to_le32( (u32)((u8)m[i].buf[j]) << ((3-h2)*8));
drivers/media/common/saa7146/saa7146_i2c.c
68
op[h1] |= cpu_to_le32( SAA7146_I2C_CONT << ((3-h2)*2));
drivers/media/common/saa7146/saa7146_i2c.c
77
if ( SAA7146_I2C_CONT == (0x3 & (le32_to_cpu(op[h1]) >> ((3-h2)*2))) ) {
drivers/media/common/saa7146/saa7146_i2c.c
78
op[h1] &= ~cpu_to_le32(0x2 << ((3-h2)*2));
drivers/media/common/saa7146/saa7146_i2c.c
79
op[h1] |= cpu_to_le32(SAA7146_I2C_STOP << ((3-h2)*2));
drivers/media/common/saa7146/saa7146_i2c.c
90
static int saa7146_i2c_msg_cleanup(const struct i2c_msg *m, int num, __le32 *op)
drivers/media/common/videobuf2/videobuf2-core.c
100
(vb)->cnt_mem_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
103
#define log_qop(q, op) \
drivers/media/common/videobuf2/videobuf2-core.c
104
dprintk(q, 2, "call_qop(%s)%s\n", #op, \
drivers/media/common/videobuf2/videobuf2-core.c
105
(q)->ops->op ? "" : " (nop)")
drivers/media/common/videobuf2/videobuf2-core.c
107
#define call_qop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
111
log_qop(q, op); \
drivers/media/common/videobuf2/videobuf2-core.c
112
err = (q)->ops->op ? (q)->ops->op(args) : 0; \
drivers/media/common/videobuf2/videobuf2-core.c
114
(q)->cnt_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
118
#define call_void_qop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
120
log_qop(q, op); \
drivers/media/common/videobuf2/videobuf2-core.c
121
if ((q)->ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
122
(q)->ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
123
(q)->cnt_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
126
#define log_vb_qop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
128
(vb)->index, #op, \
drivers/media/common/videobuf2/videobuf2-core.c
129
(vb)->vb2_queue->ops->op ? "" : " (nop)")
drivers/media/common/videobuf2/videobuf2-core.c
131
#define call_vb_qop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
135
log_vb_qop(vb, op); \
drivers/media/common/videobuf2/videobuf2-core.c
136
err = (vb)->vb2_queue->ops->op ? \
drivers/media/common/videobuf2/videobuf2-core.c
137
(vb)->vb2_queue->ops->op(args) : 0; \
drivers/media/common/videobuf2/videobuf2-core.c
139
(vb)->cnt_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
143
#define call_void_vb_qop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
145
log_vb_qop(vb, op); \
drivers/media/common/videobuf2/videobuf2-core.c
146
if ((vb)->vb2_queue->ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
147
(vb)->vb2_queue->ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
148
(vb)->cnt_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
153
#define call_memop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
154
((vb)->vb2_queue->mem_ops->op ? \
drivers/media/common/videobuf2/videobuf2-core.c
155
(vb)->vb2_queue->mem_ops->op(args) : 0)
drivers/media/common/videobuf2/videobuf2-core.c
157
#define call_ptr_memop(op, vb, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
158
((vb)->vb2_queue->mem_ops->op ? \
drivers/media/common/videobuf2/videobuf2-core.c
159
(vb)->vb2_queue->mem_ops->op(vb, args) : NULL)
drivers/media/common/videobuf2/videobuf2-core.c
161
#define call_void_memop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
163
if ((vb)->vb2_queue->mem_ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
164
(vb)->vb2_queue->mem_ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
167
#define call_qop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
168
((q)->ops->op ? (q)->ops->op(args) : 0)
drivers/media/common/videobuf2/videobuf2-core.c
170
#define call_void_qop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
172
if ((q)->ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
173
(q)->ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
176
#define call_vb_qop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
177
((vb)->vb2_queue->ops->op ? (vb)->vb2_queue->ops->op(args) : 0)
drivers/media/common/videobuf2/videobuf2-core.c
179
#define call_void_vb_qop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
181
if ((vb)->vb2_queue->ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
182
(vb)->vb2_queue->ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
187
#define call_bufop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
190
if (q && q->buf_ops && q->buf_ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
191
ret = q->buf_ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
195
#define call_void_bufop(q, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
197
if (q && q->buf_ops && q->buf_ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
198
q->buf_ops->op(args); \
drivers/media/common/videobuf2/videobuf2-core.c
64
#define log_memop(vb, op) \
drivers/media/common/videobuf2/videobuf2-core.c
66
(vb)->index, #op, \
drivers/media/common/videobuf2/videobuf2-core.c
67
(vb)->vb2_queue->mem_ops->op ? "" : " (nop)")
drivers/media/common/videobuf2/videobuf2-core.c
69
#define call_memop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
74
log_memop(vb, op); \
drivers/media/common/videobuf2/videobuf2-core.c
75
err = _q->mem_ops->op ? _q->mem_ops->op(args) : 0; \
drivers/media/common/videobuf2/videobuf2-core.c
77
(vb)->cnt_mem_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
81
#define call_ptr_memop(op, vb, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
86
log_memop(vb, op); \
drivers/media/common/videobuf2/videobuf2-core.c
87
ptr = _q->mem_ops->op ? _q->mem_ops->op(vb, args) : NULL; \
drivers/media/common/videobuf2/videobuf2-core.c
89
(vb)->cnt_mem_ ## op++; \
drivers/media/common/videobuf2/videobuf2-core.c
93
#define call_void_memop(vb, op, args...) \
drivers/media/common/videobuf2/videobuf2-core.c
97
log_memop(vb, op); \
drivers/media/common/videobuf2/videobuf2-core.c
98
if (_q->mem_ops->op) \
drivers/media/common/videobuf2/videobuf2-core.c
99
_q->mem_ops->op(args); \
drivers/media/dvb-frontends/dib7000p.c
949
static u32 dib7000p_ctrl_timf(struct dvb_frontend *fe, u8 op, u32 timf)
drivers/media/dvb-frontends/dib7000p.c
952
switch (op) {
drivers/media/dvb-frontends/dib7000p.h
55
u32 (*ctrl_timf)(struct dvb_frontend *fe, u8 op, u32 timf);
drivers/media/dvb-frontends/dib8000.c
1969
static u32 dib8000_ctrl_timf(struct dvb_frontend *fe, uint8_t op, uint32_t timf)
drivers/media/dvb-frontends/dib8000.c
1973
switch (op) {
drivers/media/dvb-frontends/dib8000.h
53
u32 (*ctrl_timf)(struct dvb_frontend *fe, uint8_t op, uint32_t timf);
drivers/media/dvb-frontends/drx39xyj/drxj.c
7847
struct drx_channel *channel, s32 tuner_freq_offset, u32 op)
drivers/media/dvb-frontends/drx39xyj/drxj.c
7991
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_CONSTELLATION)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8049
if (op & QAM_SET_OP_ALL) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8105
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_CONSTELLATION)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8151
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_SPECTRUM)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8159
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_CONSTELLATION)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8173
if (op & QAM_SET_OP_ALL) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8490
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_CONSTELLATION)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8586
if ((op & QAM_SET_OP_ALL)) {
drivers/media/dvb-frontends/drx39xyj/drxj.c
8625
if ((op & QAM_SET_OP_ALL) || (op & QAM_SET_OP_CONSTELLATION)) {
drivers/media/dvb-frontends/mt352.c
157
struct dtv_frontend_properties *op = &fe->dtv_property_cache;
drivers/media/dvb-frontends/mt352.c
164
switch (op->code_rate_HP) {
drivers/media/dvb-frontends/mt352.c
184
switch (op->code_rate_LP) {
drivers/media/dvb-frontends/mt352.c
201
if (op->hierarchy == HIERARCHY_AUTO ||
drivers/media/dvb-frontends/mt352.c
202
op->hierarchy == HIERARCHY_NONE)
drivers/media/dvb-frontends/mt352.c
209
switch (op->modulation) {
drivers/media/dvb-frontends/mt352.c
223
switch (op->transmission_mode) {
drivers/media/dvb-frontends/mt352.c
234
switch (op->guard_interval) {
drivers/media/dvb-frontends/mt352.c
251
switch (op->hierarchy) {
drivers/media/dvb-frontends/mt352.c
277
mt352_calc_nominal_rate(state, op->bandwidth_hz, buf+4);
drivers/media/dvb-frontends/mt352.c
302
struct dtv_frontend_properties *op)
drivers/media/dvb-frontends/mt352.c
330
op->code_rate_HP = tps_fec_to_api[(tps >> 7) & 7];
drivers/media/dvb-frontends/mt352.c
331
op->code_rate_LP = tps_fec_to_api[(tps >> 4) & 7];
drivers/media/dvb-frontends/mt352.c
336
op->modulation = QPSK;
drivers/media/dvb-frontends/mt352.c
339
op->modulation = QAM_16;
drivers/media/dvb-frontends/mt352.c
342
op->modulation = QAM_64;
drivers/media/dvb-frontends/mt352.c
345
op->modulation = QAM_AUTO;
drivers/media/dvb-frontends/mt352.c
349
op->transmission_mode = (tps & 0x01) ? TRANSMISSION_MODE_8K : TRANSMISSION_MODE_2K;
drivers/media/dvb-frontends/mt352.c
354
op->guard_interval = GUARD_INTERVAL_1_32;
drivers/media/dvb-frontends/mt352.c
357
op->guard_interval = GUARD_INTERVAL_1_16;
drivers/media/dvb-frontends/mt352.c
360
op->guard_interval = GUARD_INTERVAL_1_8;
drivers/media/dvb-frontends/mt352.c
363
op->guard_interval = GUARD_INTERVAL_1_4;
drivers/media/dvb-frontends/mt352.c
366
op->guard_interval = GUARD_INTERVAL_AUTO;
drivers/media/dvb-frontends/mt352.c
373
op->hierarchy = HIERARCHY_NONE;
drivers/media/dvb-frontends/mt352.c
376
op->hierarchy = HIERARCHY_1;
drivers/media/dvb-frontends/mt352.c
379
op->hierarchy = HIERARCHY_2;
drivers/media/dvb-frontends/mt352.c
382
op->hierarchy = HIERARCHY_4;
drivers/media/dvb-frontends/mt352.c
385
op->hierarchy = HIERARCHY_AUTO;
drivers/media/dvb-frontends/mt352.c
389
op->frequency = (500 * (div - IF_FREQUENCYx6)) / 3 * 1000;
drivers/media/dvb-frontends/mt352.c
392
op->bandwidth_hz = 8000000;
drivers/media/dvb-frontends/mt352.c
394
op->bandwidth_hz = 7000000;
drivers/media/dvb-frontends/mt352.c
396
op->bandwidth_hz = 6000000;
drivers/media/dvb-frontends/mt352.c
400
op->inversion = INVERSION_OFF;
drivers/media/dvb-frontends/mt352.c
402
op->inversion = INVERSION_ON;
drivers/media/firewire/firedtv-avc.c
205
const char *op, *prefix;
drivers/media/firewire/firedtv-avc.c
211
op = debug_fcp_opcode(opcode, data, length);
drivers/media/firewire/firedtv-avc.c
213
if (op) {
drivers/media/firewire/firedtv-avc.c
216
debug_fcp_ctype(data[0]), op);
drivers/media/platform/amphion/vpu.h
285
#define call_vop(inst, op, args...) \
drivers/media/platform/amphion/vpu.h
286
((inst)->ops->op ? (inst)->ops->op(inst, ##args) : 0) \
drivers/media/platform/amphion/vpu.h
288
#define call_void_vop(inst, op, args...) \
drivers/media/platform/amphion/vpu.h
290
if ((inst)->ops->op) \
drivers/media/platform/amphion/vpu.h
291
(inst)->ops->op(inst, ##args); \
drivers/media/platform/m2m-deinterlace.c
207
static void deinterlace_issue_dma(struct deinterlace_ctx *ctx, int op,
drivers/media/platform/m2m-deinterlace.c
238
switch (op) {
drivers/media/platform/mediatek/mdp3/mtk-mdp3-cmdq.c
28
#define has_op(ctx, op) \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-cmdq.c
29
((ctx)->comp->ops && (ctx)->comp->ops->op)
drivers/media/platform/mediatek/mdp3/mtk-mdp3-cmdq.c
30
#define call_op(ctx, op, ...) \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-cmdq.c
31
(has_op(ctx, op) ? (ctx)->comp->ops->op(ctx, ##__VA_ARGS__) : 0)
drivers/media/platform/rockchip/rga/rga.h
64
u32 op;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
543
#define call_cop(c, op, args...) \
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
544
(((c)->c_ops->op) ? \
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
545
((c)->c_ops->op(args)) : 0)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
767
#define s5p_mfc_hw_call(f, op, args...) \
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
768
((f && f->op) ? f->op(args) : (typeof(f->op(args)))(-ENODEV))
drivers/media/platform/st/sti/delta/delta-v4l2.c
35
#define call_dec_op(dec, op, args...)\
drivers/media/platform/st/sti/delta/delta-v4l2.c
36
((dec && (dec)->op) ? (dec)->op(args) : 0)
drivers/media/test-drivers/vidtv/vidtv_pes.c
110
op = &pts_dts;
drivers/media/test-drivers/vidtv/vidtv_pes.c
118
op = &pts;
drivers/media/test-drivers/vidtv/vidtv_pes.c
126
op,
drivers/media/test-drivers/vidtv/vidtv_pes.c
87
void *op = NULL;
drivers/media/usb/b2c2/flexcop-usb.c
316
flexcop_access_op_t op, u8 chipaddr, u8 addr, u8 *buf, u16 len)
drivers/media/usb/b2c2/flexcop-usb.c
318
if (op == FC_READ)
drivers/media/usb/dvb-usb/cinergyT2-fe.c
107
switch (op->hierarchy) {
drivers/media/usb/dvb-usb/cinergyT2-fe.c
29
static uint16_t compute_tps(struct dtv_frontend_properties *op)
drivers/media/usb/dvb-usb/cinergyT2-fe.c
33
switch (op->code_rate_HP) {
drivers/media/usb/dvb-usb/cinergyT2-fe.c
52
switch (op->code_rate_LP) {
drivers/media/usb/dvb-usb/cinergyT2-fe.c
71
switch (op->modulation) {
drivers/media/usb/dvb-usb/cinergyT2-fe.c
83
switch (op->transmission_mode) {
drivers/media/usb/dvb-usb/cinergyT2-fe.c
92
switch (op->guard_interval) {
drivers/media/usb/gspca/vc032x.c
2800
u8 op;
drivers/media/usb/gspca/vc032x.c
3031
reg_w(gspca_dev, 0xa0, ptsensor_info->op, 0xb301);
drivers/media/usb/pwc/pwc-dec23.c
492
unsigned int channel_v, offset1, op;
drivers/media/usb/pwc/pwc-dec23.c
531
op = hash_table_ops[htable_idx * 4];
drivers/media/usb/pwc/pwc-dec23.c
533
if (op == 2) {
drivers/media/usb/pwc/pwc-dec23.c
536
} else if (op == 1) {
drivers/media/usb/pwc/pwc-dec23.c
582
} while (op != 2);
drivers/media/usb/uvc/uvc_video.c
1299
struct uvc_copy_op *op = &uvc_urb->copy_operations[i];
drivers/media/usb/uvc/uvc_video.c
1301
memcpy(op->dst, op->src, op->len);
drivers/media/usb/uvc/uvc_video.c
1304
uvc_queue_buffer_release(op->buf);
drivers/media/usb/uvc/uvc_video.c
1317
struct uvc_copy_op *op = &uvc_urb->copy_operations[active_op];
drivers/media/usb/uvc/uvc_video.c
1328
op->buf = buf;
drivers/media/usb/uvc/uvc_video.c
1329
op->src = data;
drivers/media/usb/uvc/uvc_video.c
1330
op->dst = buf->mem + buf->bytesused;
drivers/media/usb/uvc/uvc_video.c
1331
op->len = min_t(unsigned int, len, maxlen);
drivers/media/usb/uvc/uvc_video.c
1333
buf->bytesused += op->len;
drivers/media/v4l2-core/v4l2-ctrls-priv.h
17
#define has_op(master, op) \
drivers/media/v4l2-core/v4l2-ctrls-priv.h
18
((master)->ops && (master)->ops->op)
drivers/media/v4l2-core/v4l2-ctrls-priv.h
19
#define call_op(master, op) \
drivers/media/v4l2-core/v4l2-ctrls-priv.h
20
(has_op(master, op) ? (master)->ops->op(master) : 0)
drivers/media/v4l2-core/v4l2-dev.c
531
#define SET_VALID_IOCTL(ops, cmd, op) \
drivers/media/v4l2-core/v4l2-dev.c
532
do { if ((ops)->op) __set_bit(_IOC_NR(cmd), valid_ioctls); } while (0)
drivers/media/v4l2-core/v4l2-flash-led-class.c
17
#define has_flash_op(v4l2_flash, op) \
drivers/media/v4l2-core/v4l2-flash-led-class.c
18
(v4l2_flash && v4l2_flash->ops && v4l2_flash->ops->op)
drivers/media/v4l2-core/v4l2-flash-led-class.c
20
#define call_flash_op(v4l2_flash, op, arg) \
drivers/media/v4l2-core/v4l2-flash-led-class.c
21
(has_flash_op(v4l2_flash, op) ? \
drivers/media/v4l2-core/v4l2-flash-led-class.c
22
v4l2_flash->ops->op(v4l2_flash, arg) : \
drivers/memory/renesas-rpc-if.c
356
static void rpcif_prepare_impl(struct rpcif_priv *rpc, const struct rpcif_op *op,
drivers/memory/renesas-rpc-if.c
368
if (op->cmd.buswidth) {
drivers/memory/renesas-rpc-if.c
370
RPCIF_SMENR_CDB(rpcif_bit_size(op->cmd.buswidth));
drivers/memory/renesas-rpc-if.c
371
rpc->command = RPCIF_SMCMR_CMD(op->cmd.opcode);
drivers/memory/renesas-rpc-if.c
372
if (op->cmd.ddr)
drivers/memory/renesas-rpc-if.c
375
if (op->ocmd.buswidth) {
drivers/memory/renesas-rpc-if.c
377
RPCIF_SMENR_OCDB(rpcif_bit_size(op->ocmd.buswidth));
drivers/memory/renesas-rpc-if.c
378
rpc->command |= RPCIF_SMCMR_OCMD(op->ocmd.opcode);
drivers/memory/renesas-rpc-if.c
381
if (op->addr.buswidth) {
drivers/memory/renesas-rpc-if.c
383
RPCIF_SMENR_ADB(rpcif_bit_size(op->addr.buswidth));
drivers/memory/renesas-rpc-if.c
384
if (op->addr.nbytes == 4)
drivers/memory/renesas-rpc-if.c
388
2, 3 - op->addr.nbytes));
drivers/memory/renesas-rpc-if.c
389
if (op->addr.ddr)
drivers/memory/renesas-rpc-if.c
395
rpc->smadr = op->addr.val;
drivers/memory/renesas-rpc-if.c
398
if (op->dummy.buswidth) {
drivers/memory/renesas-rpc-if.c
400
rpc->dummy = RPCIF_SMDMCR_DMCYC(op->dummy.ncycles);
drivers/memory/renesas-rpc-if.c
403
if (op->option.buswidth) {
drivers/memory/renesas-rpc-if.c
405
rpcif_bits_set(rpc, op->option.nbytes)) |
drivers/memory/renesas-rpc-if.c
406
RPCIF_SMENR_OPDB(rpcif_bit_size(op->option.buswidth));
drivers/memory/renesas-rpc-if.c
407
if (op->option.ddr)
drivers/memory/renesas-rpc-if.c
409
rpc->option = op->option.val;
drivers/memory/renesas-rpc-if.c
412
rpc->dir = op->data.dir;
drivers/memory/renesas-rpc-if.c
413
if (op->data.buswidth) {
drivers/memory/renesas-rpc-if.c
416
rpc->buffer = op->data.buf.in;
drivers/memory/renesas-rpc-if.c
417
switch (op->data.dir) {
drivers/memory/renesas-rpc-if.c
427
if (op->data.ddr)
drivers/memory/renesas-rpc-if.c
433
nbytes = op->data.nbytes;
drivers/memory/renesas-rpc-if.c
436
rpc->enable |= RPCIF_SMENR_SPIDB(rpcif_bit_size(op->data.buswidth));
drivers/memory/renesas-rpc-if.c
440
static void xspi_prepare_impl(struct rpcif_priv *xspi, const struct rpcif_op *op,
drivers/memory/renesas-rpc-if.c
451
if (op->cmd.buswidth)
drivers/memory/renesas-rpc-if.c
452
xspi->command = op->cmd.opcode;
drivers/memory/renesas-rpc-if.c
454
if (op->ocmd.buswidth)
drivers/memory/renesas-rpc-if.c
455
xspi->command = (xspi->command << 8) | op->ocmd.opcode;
drivers/memory/renesas-rpc-if.c
457
if (op->addr.buswidth) {
drivers/memory/renesas-rpc-if.c
458
xspi->addr_nbytes = op->addr.nbytes;
drivers/memory/renesas-rpc-if.c
462
xspi->smadr = op->addr.val;
drivers/memory/renesas-rpc-if.c
465
if (op->dummy.buswidth)
drivers/memory/renesas-rpc-if.c
466
xspi->dummy = op->dummy.ncycles;
drivers/memory/renesas-rpc-if.c
468
xspi->dir = op->data.dir;
drivers/memory/renesas-rpc-if.c
469
if (op->data.buswidth) {
drivers/memory/renesas-rpc-if.c
472
xspi->buffer = op->data.buf.in;
drivers/memory/renesas-rpc-if.c
477
nbytes = op->data.nbytes;
drivers/memory/renesas-rpc-if.c
48
void (*prepare)(struct rpcif_priv *rpc, const struct rpcif_op *op,
drivers/memory/renesas-rpc-if.c
481
if (op->cmd.buswidth == 1) {
drivers/memory/renesas-rpc-if.c
482
if (op->addr.buswidth == 2 || op->data.buswidth == 2)
drivers/memory/renesas-rpc-if.c
484
else if (op->addr.buswidth == 4 || op->data.buswidth == 4)
drivers/memory/renesas-rpc-if.c
486
} else if (op->cmd.buswidth == 2 &&
drivers/memory/renesas-rpc-if.c
487
(op->addr.buswidth == 2 || op->data.buswidth == 2)) {
drivers/memory/renesas-rpc-if.c
489
} else if (op->cmd.buswidth == 4 &&
drivers/memory/renesas-rpc-if.c
490
(op->addr.buswidth == 4 || op->data.buswidth == 4)) {
drivers/memory/renesas-rpc-if.c
495
void rpcif_prepare(struct device *dev, const struct rpcif_op *op, u64 *offs,
drivers/memory/renesas-rpc-if.c
500
rpc->info->impl->prepare(rpc, op, offs, len);
drivers/misc/eeprom/at25.c
112
struct spi_mem_op op;
drivers/misc/eeprom/at25.c
116
op = (struct spi_mem_op)SPI_MEM_OP(SPI_MEM_OP_CMD(at25_instr(at25, AT25_READ,
drivers/misc/eeprom/at25.c
122
status = spi_mem_adjust_op_size(at25->spimem, &op);
drivers/misc/eeprom/at25.c
125
segment = op.data.nbytes;
drivers/misc/eeprom/at25.c
128
status = spi_mem_exec_op(at25->spimem, &op);
drivers/misc/eeprom/at25.c
154
struct spi_mem_op op;
drivers/misc/eeprom/at25.c
160
op = (struct spi_mem_op)SPI_MEM_OP(SPI_MEM_OP_CMD(command, 1),
drivers/misc/eeprom/at25.c
165
status = spi_mem_exec_op(at25->spimem, &op);
drivers/misc/eeprom/at25.c
215
struct spi_mem_op op;
drivers/misc/eeprom/at25.c
221
op = (struct spi_mem_op)SPI_MEM_OP(SPI_MEM_OP_CMD(AT25_RDSR, 1),
drivers/misc/eeprom/at25.c
229
at25->spimem, &op);
drivers/misc/eeprom/at25.c
264
struct spi_mem_op op = SPI_MEM_OP(SPI_MEM_OP_CMD(AT25_WREN, 1),
drivers/misc/eeprom/at25.c
270
status = spi_mem_exec_op(at25->spimem, &op);
drivers/misc/eeprom/at25.c
283
op = (struct spi_mem_op)SPI_MEM_OP(SPI_MEM_OP_CMD(at25_instr(at25, AT25_WRITE, off),
drivers/misc/eeprom/at25.c
289
status = spi_mem_adjust_op_size(at25->spimem, &op);
drivers/misc/eeprom/at25.c
292
segment = op.data.nbytes;
drivers/misc/eeprom/at25.c
296
status = spi_mem_exec_op(at25->spimem, &op);
drivers/misc/sgi-gru/grufault.c
863
gru_dbg(grudev, "op %d, gseg 0x%lx, value1 0x%lx\n", req.op, req.gseg, req.val1);
drivers/misc/sgi-gru/grufault.c
872
switch (req.op) {
drivers/misc/sgi-gru/gruhandles.c
24
static void update_mcs_stats(enum mcs_op op, unsigned long clks)
drivers/misc/sgi-gru/gruhandles.c
29
atomic_long_inc(&mcs_op_statistics[op].count);
drivers/misc/sgi-gru/gruhandles.c
30
atomic_long_add(nsec, &mcs_op_statistics[op].total);
drivers/misc/sgi-gru/gruhandles.c
31
if (mcs_op_statistics[op].max < nsec)
drivers/misc/sgi-gru/gruhandles.c
32
mcs_op_statistics[op].max = nsec;
drivers/misc/sgi-gru/grulib.h
115
unsigned int op;
drivers/misc/sgi-gru/grulib.h
96
int op;
drivers/misc/sgi-gru/gruprocfs.c
116
int op;
drivers/misc/sgi-gru/gruprocfs.c
123
for (op = 0; op < mcsop_last; op++) {
drivers/misc/sgi-gru/gruprocfs.c
124
count = atomic_long_read(&mcs_op_statistics[op].count);
drivers/misc/sgi-gru/gruprocfs.c
125
total = atomic_long_read(&mcs_op_statistics[op].total);
drivers/misc/sgi-gru/gruprocfs.c
126
max = mcs_op_statistics[op].max;
drivers/misc/sgi-gru/gruprocfs.c
127
seq_printf(s, "%-20s%12ld%12ld%12ld\n", id[op], count,
drivers/misc/vmw_balloon.c
1103
.op = VMW_BALLOON_INFLATE
drivers/misc/vmw_balloon.c
1191
.op = VMW_BALLOON_DEFLATE
drivers/misc/vmw_balloon.c
246
enum vmballoon_op op;
drivers/misc/vmw_balloon.c
411
static inline void vmballoon_stats_op_inc(struct vmballoon *b, unsigned int op,
drivers/misc/vmw_balloon.c
415
atomic64_inc(&b->stats->ops[op][type]);
drivers/misc/vmw_balloon.c
779
enum vmballoon_op op)
drivers/misc/vmw_balloon.c
786
if (op == VMW_BALLOON_INFLATE)
drivers/misc/vmw_balloon.c
797
cmd = op == VMW_BALLOON_INFLATE ? VMW_BALLOON_CMD_LOCK :
drivers/misc/vmw_balloon.c
867
ctl->op);
drivers/mmc/host/sdhci-uhs2.c
515
static int sdhci_uhs2_control(struct mmc_host *mmc, enum sd_uhs2_operation op)
drivers/mmc/host/sdhci-uhs2.c
521
DBG("Begin uhs2 control, act %d.\n", op);
drivers/mmc/host/sdhci-uhs2.c
523
switch (op) {
drivers/mmc/host/sdhci-uhs2.c
550
mmc_hostname(host->mmc), op);
drivers/mtd/devices/docg3.h
320
TP_PROTO(int op, int width, u16 reg, int val),
drivers/mtd/devices/docg3.h
321
TP_ARGS(op, width, reg, val),
drivers/mtd/devices/docg3.h
323
__field(int, op)
drivers/mtd/devices/docg3.h
328
__entry->op = op;
drivers/mtd/devices/docg3.h
333
__entry->op ? "write" : "read", __entry->width,
drivers/mtd/devices/powernv_flash.c
117
op, rc);
drivers/mtd/devices/powernv_flash.c
46
static int powernv_flash_async_op(struct mtd_info *mtd, enum flash_op op,
drivers/mtd/devices/powernv_flash.c
56
__func__, op, offset, len);
drivers/mtd/devices/powernv_flash.c
67
switch (op) {
drivers/mtd/devices/st_spi_fsm.c
172
#define STFSM_INSTR(cmd, op) ((cmd) | ((op) << 4))
drivers/mtd/hyperbus/rpc-if.c
49
struct rpcif_op op = rpcif_op_tmpl;
drivers/mtd/hyperbus/rpc-if.c
51
op.cmd.opcode = HYPERBUS_RW_READ | HYPERBUS_AS_MEM;
drivers/mtd/hyperbus/rpc-if.c
52
op.addr.val = from >> 1;
drivers/mtd/hyperbus/rpc-if.c
53
op.dummy.buswidth = 1;
drivers/mtd/hyperbus/rpc-if.c
54
op.dummy.ncycles = 15;
drivers/mtd/hyperbus/rpc-if.c
55
op.data.dir = RPCIF_DATA_IN;
drivers/mtd/hyperbus/rpc-if.c
56
op.data.nbytes = len;
drivers/mtd/hyperbus/rpc-if.c
57
op.data.buf.in = to;
drivers/mtd/hyperbus/rpc-if.c
59
rpcif_prepare(rpc->dev, &op, NULL, NULL);
drivers/mtd/hyperbus/rpc-if.c
65
struct rpcif_op op = rpcif_op_tmpl;
drivers/mtd/hyperbus/rpc-if.c
67
op.cmd.opcode = HYPERBUS_RW_WRITE | HYPERBUS_AS_MEM;
drivers/mtd/hyperbus/rpc-if.c
68
op.addr.val = to >> 1;
drivers/mtd/hyperbus/rpc-if.c
69
op.data.dir = RPCIF_DATA_OUT;
drivers/mtd/hyperbus/rpc-if.c
70
op.data.nbytes = len;
drivers/mtd/hyperbus/rpc-if.c
71
op.data.buf.out = from;
drivers/mtd/hyperbus/rpc-if.c
73
rpcif_prepare(rpc->dev, &op, NULL, NULL);
drivers/mtd/maps/sun_uflash.c
103
dev_set_drvdata(&op->dev, up);
drivers/mtd/maps/sun_uflash.c
108
static int uflash_probe(struct platform_device *op)
drivers/mtd/maps/sun_uflash.c
110
struct device_node *dp = op->dev.of_node;
drivers/mtd/maps/sun_uflash.c
118
return uflash_devinit(op, dp);
drivers/mtd/maps/sun_uflash.c
121
static void uflash_remove(struct platform_device *op)
drivers/mtd/maps/sun_uflash.c
123
struct uflash_dev *up = dev_get_drvdata(&op->dev);
drivers/mtd/maps/sun_uflash.c
130
of_iounmap(&op->resource[0], up->map.virt, up->map.size);
drivers/mtd/maps/sun_uflash.c
50
static int uflash_devinit(struct platform_device *op, struct device_node *dp)
drivers/mtd/maps/sun_uflash.c
54
if (op->resource[1].flags) {
drivers/mtd/maps/sun_uflash.c
59
dp, (unsigned long long)op->resource[0].start);
drivers/mtd/maps/sun_uflash.c
71
up->map.size = resource_size(&op->resource[0]);
drivers/mtd/maps/sun_uflash.c
77
up->map.phys = op->resource[0].start;
drivers/mtd/maps/sun_uflash.c
79
up->map.virt = of_ioremap(&op->resource[0], 0, up->map.size,
drivers/mtd/maps/sun_uflash.c
93
of_iounmap(&op->resource[0], up->map.virt, up->map.size);
drivers/mtd/mtdchar.c
1317
a.op = compat_arg.op;
drivers/mtd/mtdchar.c
552
switch (arg->op) {
drivers/mtd/nand/ecc-mtk.c
127
enum mtk_ecc_operation op)
drivers/mtd/nand/ecc-mtk.c
133
ret = readl_poll_timeout_atomic(ecc->regs + ECC_IDLE_REG(op), val,
drivers/mtd/nand/ecc-mtk.c
138
op == ECC_ENCODE ? "encoder" : "decoder");
drivers/mtd/nand/ecc-mtk.c
191
if (config->op == ECC_ENCODE) {
drivers/mtd/nand/ecc-mtk.c
305
enum mtk_ecc_operation op = config->op;
drivers/mtd/nand/ecc-mtk.c
315
mtk_ecc_wait_idle(ecc, op);
drivers/mtd/nand/ecc-mtk.c
323
if (config->mode != ECC_NFI_MODE || op != ECC_ENCODE) {
drivers/mtd/nand/ecc-mtk.c
333
if (op == ECC_ENCODE)
drivers/mtd/nand/ecc-mtk.c
341
writew(ECC_OP_ENABLE, ecc->regs + ECC_CTL_REG(op));
drivers/mtd/nand/ecc-mtk.c
349
enum mtk_ecc_operation op = ECC_ENCODE;
drivers/mtd/nand/ecc-mtk.c
352
if (readw(ecc->regs + ECC_CTL_REG(op)) != ECC_OP_ENABLE)
drivers/mtd/nand/ecc-mtk.c
353
op = ECC_DECODE;
drivers/mtd/nand/ecc-mtk.c
356
mtk_ecc_wait_idle(ecc, op);
drivers/mtd/nand/ecc-mtk.c
357
if (op == ECC_DECODE) {
drivers/mtd/nand/ecc-mtk.c
368
writew(ECC_OP_DISABLE, ecc->regs + ECC_CTL_REG(op));
drivers/mtd/nand/ecc-mtk.c
374
int mtk_ecc_wait_done(struct mtk_ecc *ecc, enum mtk_ecc_operation op)
drivers/mtd/nand/ecc-mtk.c
381
(op == ECC_ENCODE) ? "encoder" : "decoder");
drivers/mtd/nand/ecc-mtk.c
40
#define ECC_IDLE_REG(op) ((op) == ECC_ENCODE ? ECC_ENCIDLE : ECC_DECIDLE)
drivers/mtd/nand/ecc-mtk.c
403
config->op = ECC_ENCODE;
drivers/mtd/nand/ecc-mtk.c
41
#define ECC_CTL_REG(op) ((op) == ECC_ENCODE ? ECC_ENCCON : ECC_DECCON)
drivers/mtd/nand/raw/ams-delta.c
140
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/ams-delta.c
151
for (instr = op->instrs; instr < op->instrs + op->ninstrs; instr++) {
drivers/mtd/nand/raw/arasan-nand-controller.c
904
const struct nand_operation *op)
drivers/mtd/nand/raw/arasan-nand-controller.c
916
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/arasan-nand-controller.c
917
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/arasan-nand-controller.c
949
if (op->ninstrs == 2 &&
drivers/mtd/nand/raw/arasan-nand-controller.c
950
op->instrs[0].type == NAND_OP_CMD_INSTR &&
drivers/mtd/nand/raw/arasan-nand-controller.c
951
op->instrs[0].ctx.cmd.opcode != NAND_CMD_STATUS &&
drivers/mtd/nand/raw/arasan-nand-controller.c
952
op->instrs[1].type == NAND_OP_DATA_IN_INSTR)
drivers/mtd/nand/raw/arasan-nand-controller.c
955
return nand_op_parser_exec_op(chip, &anfc_op_parser, op, true);
drivers/mtd/nand/raw/arasan-nand-controller.c
959
const struct nand_operation *op,
drivers/mtd/nand/raw/arasan-nand-controller.c
965
return anfc_check_op(chip, op);
drivers/mtd/nand/raw/arasan-nand-controller.c
967
ret = anfc_select_target(chip, op->cs);
drivers/mtd/nand/raw/arasan-nand-controller.c
971
return nand_op_parser_exec_op(chip, &anfc_op_parser, op, check_only);
drivers/mtd/nand/raw/atmel/nand-controller.c
1000
nc->op.data = ATMEL_NFC_WRITE_DATA;
drivers/mtd/nand/raw/atmel/nand-controller.c
1064
nc->op.cmds[nc->op.ncmds++] = NAND_CMD_READ0;
drivers/mtd/nand/raw/atmel/nand-controller.c
1067
nc->op.cmds[nc->op.ncmds++] = NAND_CMD_READSTART;
drivers/mtd/nand/raw/atmel/nand-controller.c
1070
nc->op.cs = nand->activecs->id;
drivers/mtd/nand/raw/atmel/nand-controller.c
1071
nc->op.data = ATMEL_NFC_READ_DATA;
drivers/mtd/nand/raw/atmel/nand-controller.c
1560
const struct nand_operation *op,
drivers/mtd/nand/raw/atmel/nand-controller.c
1568
return nc->caps->ops->exec_op(nand, op, check_only);
drivers/mtd/nand/raw/atmel/nand-controller.c
206
const struct nand_operation *op, bool check_only);
drivers/mtd/nand/raw/atmel/nand-controller.c
262
struct atmel_nfc_op op;
drivers/mtd/nand/raw/atmel/nand-controller.c
278
static bool atmel_nfc_op_done(struct atmel_nfc_op *op, u32 status)
drivers/mtd/nand/raw/atmel/nand-controller.c
280
op->errors |= status & ATMEL_HSMC_NFC_SR_ERRORS;
drivers/mtd/nand/raw/atmel/nand-controller.c
281
op->wait ^= status & op->wait;
drivers/mtd/nand/raw/atmel/nand-controller.c
283
return !op->wait || op->errors;
drivers/mtd/nand/raw/atmel/nand-controller.c
294
rcvd = sr & (nc->op.wait | ATMEL_HSMC_NFC_SR_ERRORS);
drivers/mtd/nand/raw/atmel/nand-controller.c
295
done = atmel_nfc_op_done(&nc->op, sr);
drivers/mtd/nand/raw/atmel/nand-controller.c
319
atmel_nfc_op_done(&nc->op,
drivers/mtd/nand/raw/atmel/nand-controller.c
325
nc->op.wait | ATMEL_HSMC_NFC_SR_ERRORS);
drivers/mtd/nand/raw/atmel/nand-controller.c
336
if (nc->op.errors & ATMEL_HSMC_NFC_SR_DTOE) {
drivers/mtd/nand/raw/atmel/nand-controller.c
341
if (nc->op.errors & ATMEL_HSMC_NFC_SR_UNDEF) {
drivers/mtd/nand/raw/atmel/nand-controller.c
346
if (nc->op.errors & ATMEL_HSMC_NFC_SR_AWB) {
drivers/mtd/nand/raw/atmel/nand-controller.c
351
if (nc->op.errors & ATMEL_HSMC_NFC_SR_NFCASE) {
drivers/mtd/nand/raw/atmel/nand-controller.c
423
u8 *addrs = nc->op.addrs;
drivers/mtd/nand/raw/atmel/nand-controller.c
424
unsigned int op = 0;
drivers/mtd/nand/raw/atmel/nand-controller.c
428
nc->op.wait = ATMEL_HSMC_NFC_SR_CMDDONE;
drivers/mtd/nand/raw/atmel/nand-controller.c
430
for (i = 0; i < nc->op.ncmds; i++)
drivers/mtd/nand/raw/atmel/nand-controller.c
431
op |= ATMEL_NFC_CMD(i, nc->op.cmds[i]);
drivers/mtd/nand/raw/atmel/nand-controller.c
433
if (nc->op.naddrs == ATMEL_NFC_MAX_ADDR_CYCLES)
drivers/mtd/nand/raw/atmel/nand-controller.c
436
op |= ATMEL_NFC_CSID(nc->op.cs) |
drivers/mtd/nand/raw/atmel/nand-controller.c
437
ATMEL_NFC_ACYCLE(nc->op.naddrs);
drivers/mtd/nand/raw/atmel/nand-controller.c
439
if (nc->op.ncmds > 1)
drivers/mtd/nand/raw/atmel/nand-controller.c
440
op |= ATMEL_NFC_VCMD2;
drivers/mtd/nand/raw/atmel/nand-controller.c
445
if (nc->op.data != ATMEL_NFC_NO_DATA) {
drivers/mtd/nand/raw/atmel/nand-controller.c
446
op |= ATMEL_NFC_DATAEN;
drivers/mtd/nand/raw/atmel/nand-controller.c
447
nc->op.wait |= ATMEL_HSMC_NFC_SR_XFRDONE;
drivers/mtd/nand/raw/atmel/nand-controller.c
449
if (nc->op.data == ATMEL_NFC_WRITE_DATA)
drivers/mtd/nand/raw/atmel/nand-controller.c
450
op |= ATMEL_NFC_NFCWR;
drivers/mtd/nand/raw/atmel/nand-controller.c
457
regmap_write(nc->io, op, addr);
drivers/mtd/nand/raw/atmel/nand-controller.c
466
memset(&nc->op, 0, sizeof(nc->op));
drivers/mtd/nand/raw/atmel/nand-controller.c
611
const struct nand_operation *op,
drivers/mtd/nand/raw/atmel/nand-controller.c
620
atmel_nand_select_target(nand, op->cs);
drivers/mtd/nand/raw/atmel/nand-controller.c
622
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/atmel/nand-controller.c
623
ret = atmel_smc_nand_exec_instr(nand, &op->instrs[i]);
drivers/mtd/nand/raw/atmel/nand-controller.c
641
nc->op.cs = nand->activecs->id;
drivers/mtd/nand/raw/atmel/nand-controller.c
646
nc->op.cmds[nc->op.ncmds++] = instr->ctx.cmd.opcode;
drivers/mtd/nand/raw/atmel/nand-controller.c
652
nc->op.addrs[nc->op.naddrs] = instr->ctx.addr.addrs[j];
drivers/mtd/nand/raw/atmel/nand-controller.c
653
nc->op.naddrs++;
drivers/mtd/nand/raw/atmel/nand-controller.c
701
const struct nand_operation *op,
drivers/mtd/nand/raw/atmel/nand-controller.c
708
&atmel_hsmc_op_parser, op, true);
drivers/mtd/nand/raw/atmel/nand-controller.c
710
atmel_hsmc_nand_select_target(nand, op->cs);
drivers/mtd/nand/raw/atmel/nand-controller.c
711
ret = nand_op_parser_exec_op(&nand->base, &atmel_hsmc_op_parser, op,
drivers/mtd/nand/raw/atmel/nand-controller.c
770
nc->op.addrs[nc->op.naddrs++] = column;
drivers/mtd/nand/raw/atmel/nand-controller.c
776
nc->op.addrs[nc->op.naddrs++] = column >> 8;
drivers/mtd/nand/raw/atmel/nand-controller.c
780
nc->op.addrs[nc->op.naddrs++] = page;
drivers/mtd/nand/raw/atmel/nand-controller.c
781
nc->op.addrs[nc->op.naddrs++] = page >> 8;
drivers/mtd/nand/raw/atmel/nand-controller.c
784
nc->op.addrs[nc->op.naddrs++] = page >> 16;
drivers/mtd/nand/raw/atmel/nand-controller.c
788
static int atmel_nand_pmecc_enable(struct nand_chip *chip, int op, bool raw)
drivers/mtd/nand/raw/atmel/nand-controller.c
799
ret = atmel_pmecc_enable(nand->pmecc, op);
drivers/mtd/nand/raw/atmel/nand-controller.c
996
nc->op.cmds[0] = NAND_CMD_SEQIN;
drivers/mtd/nand/raw/atmel/nand-controller.c
997
nc->op.ncmds = 1;
drivers/mtd/nand/raw/atmel/nand-controller.c
999
nc->op.cs = nand->activecs->id;
drivers/mtd/nand/raw/atmel/pmecc.c
765
int atmel_pmecc_enable(struct atmel_pmecc_user *user, int op)
drivers/mtd/nand/raw/atmel/pmecc.c
770
if (op != NAND_ECC_READ && op != NAND_ECC_WRITE) {
drivers/mtd/nand/raw/atmel/pmecc.c
778
if (op == NAND_ECC_WRITE)
drivers/mtd/nand/raw/atmel/pmecc.h
59
int atmel_pmecc_enable(struct atmel_pmecc_user *user, int op);
drivers/mtd/nand/raw/au1550nd.c
211
const struct nand_operation *op,
drivers/mtd/nand/raw/au1550nd.c
226
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/au1550nd.c
227
ret = au1550nd_exec_instr(this, &op->instrs[i]);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2455
const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2457
const struct nand_op_instr *instr = &op->instrs[i];
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2468
last_op = ((i == (op->ninstrs - 1)) && (instr->type != NAND_OP_WAITRDY_INSTR)) ||
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2469
((i == (op->ninstrs - 2)) && (op->instrs[i + 1].type == NAND_OP_WAITRDY_INSTR));
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2512
static int brcmnand_op_is_status(const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2514
if (op->ninstrs == 2 &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2515
op->instrs[0].type == NAND_OP_CMD_INSTR &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2516
op->instrs[0].ctx.cmd.opcode == NAND_CMD_STATUS &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2517
op->instrs[1].type == NAND_OP_DATA_IN_INSTR)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2523
static int brcmnand_op_is_reset(const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2525
if (op->ninstrs == 2 &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2526
op->instrs[0].type == NAND_OP_CMD_INSTR &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2527
op->instrs[0].ctx.cmd.opcode == NAND_CMD_RESET &&
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2528
op->instrs[1].type == NAND_OP_WAITRDY_INSTR)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2535
const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2541
const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2547
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2548
ret = brcmnand_exec_instr(host, i, op);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2557
const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2563
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2564
instr = &op->instrs[i];
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2585
const struct nand_operation *op)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2596
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2597
instr = &op->instrs[i];
drivers/mtd/nand/raw/brcmnand/brcmnand.c
268
const struct nand_operation *op);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
270
const struct nand_operation *op);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2707
const struct nand_operation *op,
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2717
return ctrl->check_instr(chip, op);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2719
if (brcmnand_op_is_status(op)) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2720
status = op->instrs[1].ctx.data.buf.in;
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2728
} else if (brcmnand_op_is_reset(op)) {
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2738
if (op->deassert_wp)
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2741
ret = ctrl->exec_instr(chip, op);
drivers/mtd/nand/raw/brcmnand/brcmnand.c
2743
if (op->deassert_wp)
drivers/mtd/nand/raw/cadence-nand-controller.c
2317
const struct nand_operation *op,
drivers/mtd/nand/raw/cadence-nand-controller.c
2327
return nand_op_parser_exec_op(chip, &cadence_nand_op_parser, op,
drivers/mtd/nand/raw/cs553x_nand.c
193
const struct nand_operation *op,
drivers/mtd/nand/raw/cs553x_nand.c
205
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/cs553x_nand.c
206
ret = cs553x_exec_instr(cs553x, &op->instrs[i]);
drivers/mtd/nand/raw/davinci_nand.c
798
const struct nand_operation *op,
drivers/mtd/nand/raw/davinci_nand.c
807
info->current_cs = info->vaddr + (op->cs * info->mask_chipsel);
drivers/mtd/nand/raw/davinci_nand.c
809
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/davinci_nand.c
812
ret = davinci_nand_exec_instr(info, &op->instrs[i]);
drivers/mtd/nand/raw/denali.c
1149
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/denali.c
1156
denali_select_target(chip, op->cs);
drivers/mtd/nand/raw/denali.c
1164
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/denali.c
1165
ret = denali_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/diskonchip.c
354
struct nand_operation op = NAND_OPERATION(cs, instrs);
drivers/mtd/nand/raw/diskonchip.c
357
op.ninstrs--;
drivers/mtd/nand/raw/diskonchip.c
359
this->controller->ops->exec_op(this, &op, false);
drivers/mtd/nand/raw/diskonchip.c
564
const struct nand_operation *op,
drivers/mtd/nand/raw/diskonchip.c
573
doc->curchip = op->cs % doc->chips_per_floor;
drivers/mtd/nand/raw/diskonchip.c
574
doc->curfloor = op->cs / doc->chips_per_floor;
drivers/mtd/nand/raw/diskonchip.c
582
for (i = 0; i < op->ninstrs; i++)
drivers/mtd/nand/raw/diskonchip.c
583
doc200x_exec_instr(this, &op->instrs[i]);
drivers/mtd/nand/raw/diskonchip.c
639
const struct nand_operation *op,
drivers/mtd/nand/raw/diskonchip.c
648
doc->curchip = op->cs % doc->chips_per_floor;
drivers/mtd/nand/raw/diskonchip.c
649
doc->curfloor = op->cs / doc->chips_per_floor;
drivers/mtd/nand/raw/diskonchip.c
654
for (i = 0; i < op->ninstrs; i++)
drivers/mtd/nand/raw/diskonchip.c
655
doc2001plus_exec_instr(this, &op->instrs[i]);
drivers/mtd/nand/raw/fsl_upm.c
132
static int fun_exec_op(struct nand_chip *chip, const struct nand_operation *op,
drivers/mtd/nand/raw/fsl_upm.c
139
if (op->cs >= NAND_MAX_CHIPS)
drivers/mtd/nand/raw/fsl_upm.c
145
fun->mchip_number = op->cs;
drivers/mtd/nand/raw/fsl_upm.c
147
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/fsl_upm.c
148
ret = func_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/fsl_upm.c
152
if (op->instrs[i].delay_ns)
drivers/mtd/nand/raw/fsl_upm.c
153
ndelay(op->instrs[i].delay_ns);
drivers/mtd/nand/raw/fsmc_nand.c
632
static int fsmc_exec_op(struct nand_chip *chip, const struct nand_operation *op,
drivers/mtd/nand/raw/fsmc_nand.c
644
pr_debug("Executing operation [%d instructions]:\n", op->ninstrs);
drivers/mtd/nand/raw/fsmc_nand.c
646
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/fsmc_nand.c
647
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/gpio.c
138
const struct nand_operation *op,
drivers/mtd/nand/raw/gpio.c
150
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/gpio.c
151
ret = gpio_nand_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/gpio.c
155
if (op->instrs[i].delay_ns)
drivers/mtd/nand/raw/gpio.c
156
ndelay(op->instrs[i].delay_ns);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2503
const struct nand_operation *op,
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2541
dev_dbg(this->dev, "%s: %d instructions\n", __func__, op->ninstrs);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2543
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2544
instr = &op->instrs[i];
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2559
if (i + 1 != op->ninstrs &&
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2560
op->instrs[i + 1].type == NAND_OP_ADDR_INSTR)
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
312
const struct nand_operation *op,
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
324
cs = &nfc->cs[op->cs];
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
326
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
327
ret = ingenic_nand_exec_instr(chip, cs, &op->instrs[i]);
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
331
if (op->instrs[i].delay_ns)
drivers/mtd/nand/raw/ingenic/ingenic_nand_drv.c
332
ndelay(op->instrs[i].delay_ns);
drivers/mtd/nand/raw/intel-nand-controller.c
515
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/intel-nand-controller.c
525
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/intel-nand-controller.c
526
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/internals.h
125
const struct nand_operation *op)
drivers/mtd/nand/raw/internals.h
130
return chip->controller->ops->exec_op(chip, op, true);
drivers/mtd/nand/raw/internals.h
134
const struct nand_operation *op)
drivers/mtd/nand/raw/internals.h
139
if (WARN_ON(op->cs >= nanddev_ntargets(&chip->base)))
drivers/mtd/nand/raw/internals.h
142
return chip->controller->ops->exec_op(chip, op, false);
drivers/mtd/nand/raw/loongson-nand-controller.c
115
void (*set_addr)(struct loongson_nand_host *host, struct loongson_nand_op *op);
drivers/mtd/nand/raw/loongson-nand-controller.c
139
static int loongson_nand_op_cmd_mapping(struct nand_chip *chip, struct loongson_nand_op *op,
drivers/mtd/nand/raw/loongson-nand-controller.c
144
op->row_start = chip->page_shift + 1;
drivers/mtd/nand/raw/loongson-nand-controller.c
149
op->cmd_reg = LOONGSON_NAND_CMD_STATUS;
drivers/mtd/nand/raw/loongson-nand-controller.c
152
op->cmd_reg = LOONGSON_NAND_CMD_RESET;
drivers/mtd/nand/raw/loongson-nand-controller.c
155
op->is_readid = true;
drivers/mtd/nand/raw/loongson-nand-controller.c
156
op->cmd_reg = LOONGSON_NAND_CMD_READID;
drivers/mtd/nand/raw/loongson-nand-controller.c
159
op->is_erase = true;
drivers/mtd/nand/raw/loongson-nand-controller.c
160
op->addrs_offset = LOONGSON_NAND_COL_ADDR_CYC;
drivers/mtd/nand/raw/loongson-nand-controller.c
163
if (!op->is_erase)
drivers/mtd/nand/raw/loongson-nand-controller.c
166
op->row_start = chip->page_shift;
drivers/mtd/nand/raw/loongson-nand-controller.c
167
op->cmd_reg = LOONGSON_NAND_CMD_ERASE;
drivers/mtd/nand/raw/loongson-nand-controller.c
170
op->is_write = true;
drivers/mtd/nand/raw/loongson-nand-controller.c
173
if (!op->is_write)
drivers/mtd/nand/raw/loongson-nand-controller.c
175
op->cmd_reg = LOONGSON_NAND_CMD_WRITE;
drivers/mtd/nand/raw/loongson-nand-controller.c
178
op->is_read = true;
drivers/mtd/nand/raw/loongson-nand-controller.c
181
if (!op->is_read)
drivers/mtd/nand/raw/loongson-nand-controller.c
183
op->cmd_reg = LOONGSON_NAND_CMD_READ;
drivers/mtd/nand/raw/loongson-nand-controller.c
186
op->is_change_column = true;
drivers/mtd/nand/raw/loongson-nand-controller.c
189
if (!op->is_change_column)
drivers/mtd/nand/raw/loongson-nand-controller.c
191
op->cmd_reg = LOONGSON_NAND_CMD_READ;
drivers/mtd/nand/raw/loongson-nand-controller.c
202
struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
214
ret = loongson_nand_op_cmd_mapping(chip, op, instr->ctx.cmd.opcode);
drivers/mtd/nand/raw/loongson-nand-controller.c
223
op->naddrs = naddrs;
drivers/mtd/nand/raw/loongson-nand-controller.c
226
memcpy(op->addrs + op->addrs_offset, addrs, naddrs);
drivers/mtd/nand/raw/loongson-nand-controller.c
231
op->orig_len = nand_subop_get_data_len(subop, op_id);
drivers/mtd/nand/raw/loongson-nand-controller.c
233
op->buf = instr->ctx.data.buf.in + offset;
drivers/mtd/nand/raw/loongson-nand-controller.c
235
op->buf = (void *)instr->ctx.data.buf.out + offset;
drivers/mtd/nand/raw/loongson-nand-controller.c
239
op->rdy_timeout_ms = instr->ctx.waitrdy.timeout_ms;
drivers/mtd/nand/raw/loongson-nand-controller.c
269
static void ls1b_nand_set_addr(struct loongson_nand_host *host, struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
281
val = (u32)op->addrs[i] << shift;
drivers/mtd/nand/raw/loongson-nand-controller.c
283
} else if (!op->is_change_column) {
drivers/mtd/nand/raw/loongson-nand-controller.c
284
shift = op->row_start + (i - LOONGSON_NAND_COL_ADDR_CYC) * BITS_PER_BYTE;
drivers/mtd/nand/raw/loongson-nand-controller.c
286
val = (u32)op->addrs[i] << shift;
drivers/mtd/nand/raw/loongson-nand-controller.c
291
val = (u32)op->addrs[i] >> (BITS_PER_WORD - shift);
drivers/mtd/nand/raw/loongson-nand-controller.c
298
static void ls1c_nand_set_addr(struct loongson_nand_host *host, struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
308
val = (u32)op->addrs[i] << shift;
drivers/mtd/nand/raw/loongson-nand-controller.c
310
} else if (!op->is_change_column) {
drivers/mtd/nand/raw/loongson-nand-controller.c
313
val = (u32)op->addrs[i] << shift;
drivers/mtd/nand/raw/loongson-nand-controller.c
321
static void loongson_nand_trigger_op(struct loongson_nand_host *host, struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
325
int col0 = op->addrs[0];
drivers/mtd/nand/raw/loongson-nand-controller.c
329
col0 = ALIGN_DOWN(op->addrs[0], chip->buf_align);
drivers/mtd/nand/raw/loongson-nand-controller.c
330
op->aligned_offset = op->addrs[0] - col0;
drivers/mtd/nand/raw/loongson-nand-controller.c
331
op->addrs[0] = col0;
drivers/mtd/nand/raw/loongson-nand-controller.c
335
host->data->set_addr(host, op);
drivers/mtd/nand/raw/loongson-nand-controller.c
338
if (op->is_write || op->is_read || op->is_change_column)
drivers/mtd/nand/raw/loongson-nand-controller.c
339
op->len = ALIGN(op->orig_len + op->aligned_offset, chip->buf_align);
drivers/mtd/nand/raw/loongson-nand-controller.c
340
else if (op->is_erase)
drivers/mtd/nand/raw/loongson-nand-controller.c
341
op->len = 1;
drivers/mtd/nand/raw/loongson-nand-controller.c
343
op->len = op->orig_len;
drivers/mtd/nand/raw/loongson-nand-controller.c
345
writel(op->len, host->reg_base + LOONGSON_NAND_OP_NUM);
drivers/mtd/nand/raw/loongson-nand-controller.c
348
col = op->addrs[1] << BITS_PER_BYTE | op->addrs[0];
drivers/mtd/nand/raw/loongson-nand-controller.c
349
if (op->orig_len && !op->is_readid) {
drivers/mtd/nand/raw/loongson-nand-controller.c
353
op->cmd_reg |= LOONGSON_NAND_CMD_OP_MAIN;
drivers/mtd/nand/raw/loongson-nand-controller.c
357
op->cmd_reg |= LOONGSON_NAND_CMD_OP_SPARE;
drivers/mtd/nand/raw/loongson-nand-controller.c
366
writel(op->cmd_reg, host->reg_base + LOONGSON_NAND_CMD);
drivers/mtd/nand/raw/loongson-nand-controller.c
374
struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
379
if (op->rdy_timeout_ms) {
drivers/mtd/nand/raw/loongson-nand-controller.c
382
0, op->rdy_timeout_ms * MSEC_PER_SEC);
drivers/mtd/nand/raw/loongson-nand-controller.c
406
static int loongson_nand_dma_transfer(struct loongson_nand_host *host, struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
412
enum dma_data_direction data_dir = op->is_write ? DMA_TO_DEVICE : DMA_FROM_DEVICE;
drivers/mtd/nand/raw/loongson-nand-controller.c
413
enum dma_transfer_direction xfer_dir = op->is_write ? DMA_MEM_TO_DEV : DMA_DEV_TO_MEM;
drivers/mtd/nand/raw/loongson-nand-controller.c
414
void *buf = op->buf;
drivers/mtd/nand/raw/loongson-nand-controller.c
420
IS_ALIGNED(op->orig_len, chip->buf_align)) {
drivers/mtd/nand/raw/loongson-nand-controller.c
421
dma_addr = dma_map_single(dev, buf, op->orig_len, data_dir);
drivers/mtd/nand/raw/loongson-nand-controller.c
426
} else if (!op->is_write) {
drivers/mtd/nand/raw/loongson-nand-controller.c
427
dma_buf = dma_alloc_coherent(dev, op->len, &dma_addr, GFP_KERNEL);
drivers/mtd/nand/raw/loongson-nand-controller.c
435
desc = dmaengine_prep_slave_single(chan, dma_addr, op->len, xfer_dir, DMA_PREP_INTERRUPT);
drivers/mtd/nand/raw/loongson-nand-controller.c
462
memcpy(buf, dma_buf + op->aligned_offset, op->orig_len);
drivers/mtd/nand/raw/loongson-nand-controller.c
465
dma_free_coherent(dev, op->len, dma_buf, dma_addr);
drivers/mtd/nand/raw/loongson-nand-controller.c
467
dma_unmap_single(dev, dma_addr, op->orig_len, data_dir);
drivers/mtd/nand/raw/loongson-nand-controller.c
475
struct loongson_nand_op op = {};
drivers/mtd/nand/raw/loongson-nand-controller.c
478
ret = loongson_nand_parse_instructions(chip, subop, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
482
loongson_nand_trigger_op(host, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
484
ret = loongson_nand_dma_transfer(host, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
488
return loongson_nand_wait_for_op_done(host, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
492
struct loongson_nand_op *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
497
ret = loongson_nand_parse_instructions(chip, subop, op);
drivers/mtd/nand/raw/loongson-nand-controller.c
501
loongson_nand_trigger_op(host, op);
drivers/mtd/nand/raw/loongson-nand-controller.c
503
return loongson_nand_wait_for_op_done(host, op);
drivers/mtd/nand/raw/loongson-nand-controller.c
508
struct loongson_nand_op op = {};
drivers/mtd/nand/raw/loongson-nand-controller.c
510
return loongson_nand_misc_type_exec(chip, subop, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
516
struct loongson_nand_op op = {};
drivers/mtd/nand/raw/loongson-nand-controller.c
526
ret = loongson_nand_misc_type_exec(chip, subop, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
538
for (i = 0; i < min(host->data->max_id_cycle, op.orig_len); i++)
drivers/mtd/nand/raw/loongson-nand-controller.c
539
op.buf[i] = nand_id.ids[host->data->max_id_cycle - 1 - i];
drivers/mtd/nand/raw/loongson-nand-controller.c
548
struct loongson_nand_op op = {};
drivers/mtd/nand/raw/loongson-nand-controller.c
551
ret = loongson_nand_misc_type_exec(chip, subop, &op);
drivers/mtd/nand/raw/loongson-nand-controller.c
557
op.buf[0] = val << ffs(host->data->status_field);
drivers/mtd/nand/raw/loongson-nand-controller.c
623
static int loongson_nand_check_op(struct nand_chip *chip, const struct nand_operation *op)
drivers/mtd/nand/raw/loongson-nand-controller.c
628
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/loongson-nand-controller.c
629
const struct nand_op_instr *instr = &op->instrs[op_id];
drivers/mtd/nand/raw/loongson-nand-controller.c
650
static int loongson_nand_exec_op(struct nand_chip *chip, const struct nand_operation *op,
drivers/mtd/nand/raw/loongson-nand-controller.c
654
return loongson_nand_check_op(chip, op);
drivers/mtd/nand/raw/loongson-nand-controller.c
656
return nand_op_parser_exec_op(chip, &loongson_nand_op_parser, op, check_only);
drivers/mtd/nand/raw/marvell_nand.c
2182
const struct nand_operation *op,
drivers/mtd/nand/raw/marvell_nand.c
2188
marvell_nfc_select_target(chip, op->cs);
drivers/mtd/nand/raw/marvell_nand.c
2192
op, check_only);
drivers/mtd/nand/raw/marvell_nand.c
2195
op, check_only);
drivers/mtd/nand/raw/meson_nand.c
1003
const struct nand_operation *op)
drivers/mtd/nand/raw/meson_nand.c
1007
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/meson_nand.c
1010
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/meson_nand.c
1028
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/meson_nand.c
1038
err = meson_nfc_check_op(nand, op);
drivers/mtd/nand/raw/meson_nand.c
1045
meson_nfc_select_chip(nand, op->cs);
drivers/mtd/nand/raw/meson_nand.c
1046
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/meson_nand.c
1047
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/mpc5121_nfc.c
615
static int mpc5121_nfc_probe(struct platform_device *op)
drivers/mtd/nand/raw/mpc5121_nfc.c
617
struct device_node *dn = op->dev.of_node;
drivers/mtd/nand/raw/mpc5121_nfc.c
619
struct device *dev = &op->dev;
drivers/mtd/nand/raw/mpc5121_nfc.c
818
static void mpc5121_nfc_remove(struct platform_device *op)
drivers/mtd/nand/raw/mpc5121_nfc.c
820
struct device *dev = &op->dev;
drivers/mtd/nand/raw/mtk_nand.c
510
const struct nand_operation *op,
drivers/mtd/nand/raw/mtk_nand.c
522
mtk_nfc_select_target(chip, op->cs);
drivers/mtd/nand/raw/mtk_nand.c
524
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/mtk_nand.c
525
ret = mtk_nfc_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/mtk_nand.c
648
nfc->ecc_cfg.op = ECC_ENCODE;
drivers/mtd/nand/raw/mtk_nand.c
843
nfc->ecc_cfg.op = ECC_ENCODE;
drivers/mtd/nand/raw/mtk_nand.c
981
nfc->ecc_cfg.op = ECC_DECODE;
drivers/mtd/nand/raw/mxc_nand.c
1558
const struct nand_subop *op)
drivers/mtd/nand/raw/mxc_nand.c
1569
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/mxc_nand.c
1570
instr = &op->instrs[i];
drivers/mtd/nand/raw/mxc_nand.c
1676
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/mxc_nand.c
1679
op, check_only);
drivers/mtd/nand/raw/mxic_nand.c
389
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/mxic_nand.c
401
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/mxic_nand.c
402
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/nand_base.c
1153
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1158
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1178
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1196
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1201
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1215
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1409
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1413
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1415
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1468
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1477
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1481
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1551
struct nand_operation op = NAND_DESTRUCTIVE_OPERATION(chip->cur_cs,
drivers/mtd/nand/raw/nand_base.c
1567
op.ninstrs -= 2;
drivers/mtd/nand/raw/nand_base.c
1570
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1589
op.instrs++;
drivers/mtd/nand/raw/nand_base.c
1590
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1593
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1657
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1659
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1776
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1787
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1789
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1831
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1846
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1848
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1890
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1900
op.ninstrs--;
drivers/mtd/nand/raw/nand_base.c
1902
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1934
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
1936
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
1975
struct nand_operation op = NAND_DESTRUCTIVE_OPERATION(chip->cur_cs,
drivers/mtd/nand/raw/nand_base.c
1981
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2037
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2039
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2086
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2095
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2119
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2121
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2154
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2156
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2192
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2215
ret = nand_check_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2220
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2273
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_base.c
2277
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_base.c
2517
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/nand_base.c
2520
.subop.cs = op->cs,
drivers/mtd/nand/raw/nand_base.c
2521
.subop.instrs = op->instrs,
drivers/mtd/nand/raw/nand_base.c
2522
.instrs = op->instrs,
drivers/mtd/nand/raw/nand_base.c
2523
.ninstrs = op->ninstrs,
drivers/mtd/nand/raw/nand_base.c
2527
while (ctx.subop.instrs < op->instrs + op->ninstrs) {
drivers/mtd/nand/raw/nand_hynix.c
77
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_hynix.c
79
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_hynix.c
96
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_hynix.c
98
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_macronix.c
268
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_macronix.c
270
ret = nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nand_toshiba.c
42
struct nand_operation op = NAND_OPERATION(chip->cur_cs, instrs);
drivers/mtd/nand/raw/nand_toshiba.c
44
return nand_exec_op(chip, &op);
drivers/mtd/nand/raw/nandsim.c
1162
ns->op = NULL;
drivers/mtd/nand/raw/nandsim.c
1251
ns->op = &ops[idx].states[0];
drivers/mtd/nand/raw/nandsim.c
1265
ns->state = ns->op[ns->stateidx];
drivers/mtd/nand/raw/nandsim.c
1266
ns->nxstate = ns->op[ns->stateidx + 1];
drivers/mtd/nand/raw/nandsim.c
1722
if (ns->op) {
drivers/mtd/nand/raw/nandsim.c
1730
ns->nxstate = ns->op[ns->stateidx + 1];
drivers/mtd/nand/raw/nandsim.c
1793
ns->nxstate = ns->op[++ns->stateidx + 1];
drivers/mtd/nand/raw/nandsim.c
2154
static int ns_exec_op(struct nand_chip *chip, const struct nand_operation *op,
drivers/mtd/nand/raw/nandsim.c
2169
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/nandsim.c
2170
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/nandsim.c
2182
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/nandsim.c
2183
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/nandsim.c
301
uint32_t *op; /* current operation, NULL operations isn't known yet */
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
827
const struct nand_operation *op,
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
836
ma35_nand_target_enable(chip, op->cs);
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
838
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/nuvoton-ma35d1-nand-controller.c
839
ret = ma35_nfc_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/omap2.c
2184
const struct nand_operation *op,
drivers/mtd/nand/raw/omap2.c
2192
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/omap2.c
2195
ret = omap_nand_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/pl35x-nand-controller.c
774
const struct nand_operation *op,
drivers/mtd/nand/raw/pl35x-nand-controller.c
778
pl35x_nand_select_target(chip, op->cs);
drivers/mtd/nand/raw/pl35x-nand-controller.c
781
op, check_only);
drivers/mtd/nand/raw/qcom_nandc.c
1989
const struct nand_operation *op)
drivers/mtd/nand/raw/qcom_nandc.c
1994
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/qcom_nandc.c
1995
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/qcom_nandc.c
2019
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/qcom_nandc.c
2022
return qcom_check_op(chip, op);
drivers/mtd/nand/raw/qcom_nandc.c
2024
return nand_op_parser_exec_op(chip, &qcom_op_parser, op, check_only);
drivers/mtd/nand/raw/renesas-nand-controller.c
689
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/renesas-nand-controller.c
705
rnandc_select_target(chip, op->cs);
drivers/mtd/nand/raw/renesas-nand-controller.c
707
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/renesas-nand-controller.c
708
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/rockchip-nand-controller.c
404
const struct nand_operation *op,
drivers/mtd/nand/raw/rockchip-nand-controller.c
408
rk_nfc_select_chip(chip, op->cs);
drivers/mtd/nand/raw/rockchip-nand-controller.c
410
return nand_op_parser_exec_op(chip, &rk_nfc_op_parser, op,
drivers/mtd/nand/raw/stm32_fmc2_nand.c
1317
const struct nand_operation *op,
drivers/mtd/nand/raw/stm32_fmc2_nand.c
1328
ret = stm32_fmc2_nfc_select_chip(chip, op->cs);
drivers/mtd/nand/raw/stm32_fmc2_nand.c
1332
for (op_id = 0; op_id < op->ninstrs; op_id++) {
drivers/mtd/nand/raw/stm32_fmc2_nand.c
1333
instr = &op->instrs[op_id];
drivers/mtd/nand/raw/sunxi_nand.c
2077
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/sunxi_nand.c
2083
sunxi_nfc_select_chip(nand, op->cs);
drivers/mtd/nand/raw/sunxi_nand.c
2085
if (sunxi_nand->sels[op->cs].rb >= 0)
drivers/mtd/nand/raw/sunxi_nand.c
2090
return nand_op_parser_exec_op(nand, parser, op, check_only);
drivers/mtd/nand/raw/technologic-nand-controller.c
113
const struct nand_operation *op, bool check_only)
drivers/mtd/nand/raw/technologic-nand-controller.c
121
for (i = 0; i < op->ninstrs; i++) {
drivers/mtd/nand/raw/technologic-nand-controller.c
122
ret = ts72xx_nand_exec_instr(chip, &op->instrs[i]);
drivers/mtd/nand/raw/tegra_nand.c
470
const struct nand_operation *op,
drivers/mtd/nand/raw/tegra_nand.c
474
tegra_nand_select_target(chip, op->cs);
drivers/mtd/nand/raw/tegra_nand.c
476
return nand_op_parser_exec_op(chip, &tegra_nand_op_parser, op,
drivers/mtd/nand/raw/vf610_nfc.c
498
const struct nand_operation *op,
drivers/mtd/nand/raw/vf610_nfc.c
502
vf610_nfc_select_target(chip, op->cs);
drivers/mtd/nand/raw/vf610_nfc.c
504
return nand_op_parser_exec_op(chip, &vf610_nfc_op_parser, op,
drivers/mtd/nand/raw/xway_nand.c
71
static u8 xway_readb(struct mtd_info *mtd, int op)
drivers/mtd/nand/raw/xway_nand.c
76
return readb(data->nandaddr + op);
drivers/mtd/nand/raw/xway_nand.c
79
static void xway_writeb(struct mtd_info *mtd, int op, u8 value)
drivers/mtd/nand/raw/xway_nand.c
84
writeb(value, data->nandaddr + op);
drivers/mtd/nand/spi/core.c
100
return op;
drivers/mtd/nand/spi/core.c
106
struct spi_mem_op op = spinand->op_templates->prog_exec;
drivers/mtd/nand/spi/core.c
108
op.addr.val = addr;
drivers/mtd/nand/spi/core.c
110
return op;
drivers/mtd/nand/spi/core.c
115
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/core.c
119
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
129
struct spi_mem_op op = SPINAND_OP(spinand, set_feature,
drivers/mtd/nand/spi/core.c
133
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
1410
static bool spinand_op_is_odtr(const struct spi_mem_op *op)
drivers/mtd/nand/spi/core.c
1412
return op->cmd.dtr && op->cmd.buswidth == 8;
drivers/mtd/nand/spi/core.c
1447
const struct spi_mem_op *op = &info->vendor_ops->ops[i];
drivers/mtd/nand/spi/core.c
1449
if ((iface == SSDR && spinand_op_is_odtr(op)) ||
drivers/mtd/nand/spi/core.c
1450
(iface == ODTR && !spinand_op_is_odtr(op)))
drivers/mtd/nand/spi/core.c
1453
if (!spi_mem_supports_op(spinand->spimem, op))
drivers/mtd/nand/spi/core.c
1513
struct spi_mem_op op = variants->ops[i];
drivers/mtd/nand/spi/core.c
1518
if ((iface == SSDR && spinand_op_is_odtr(&op)) ||
drivers/mtd/nand/spi/core.c
1519
(iface == ODTR && !spinand_op_is_odtr(&op)))
drivers/mtd/nand/spi/core.c
1526
op.data.nbytes = nbytes;
drivers/mtd/nand/spi/core.c
1527
ret = spi_mem_adjust_op_size(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
1531
spi_mem_adjust_op_freq(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
1533
if (!spi_mem_supports_op(spinand->spimem, &op))
drivers/mtd/nand/spi/core.c
1536
nbytes -= op.data.nbytes;
drivers/mtd/nand/spi/core.c
1538
op_duration_ns += spi_mem_calc_op_duration(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
1577
const struct spi_mem_op *op;
drivers/mtd/nand/spi/core.c
1600
op = spinand_select_op_variant(spinand, SSDR,
drivers/mtd/nand/spi/core.c
1602
if (!op)
drivers/mtd/nand/spi/core.c
1605
spinand->ssdr_op_templates.read_cache = op;
drivers/mtd/nand/spi/core.c
1607
op = spinand_select_op_variant(spinand, SSDR,
drivers/mtd/nand/spi/core.c
1609
if (!op)
drivers/mtd/nand/spi/core.c
1612
spinand->ssdr_op_templates.write_cache = op;
drivers/mtd/nand/spi/core.c
1614
op = spinand_select_op_variant(spinand, SSDR,
drivers/mtd/nand/spi/core.c
1616
if (!op)
drivers/mtd/nand/spi/core.c
1619
spinand->ssdr_op_templates.update_cache = op;
drivers/mtd/nand/spi/core.c
1635
op = spinand_select_op_variant(spinand, ODTR,
drivers/mtd/nand/spi/core.c
1637
spinand->odtr_op_templates.read_cache = op;
drivers/mtd/nand/spi/core.c
1639
op = spinand_select_op_variant(spinand, ODTR,
drivers/mtd/nand/spi/core.c
1641
spinand->odtr_op_templates.write_cache = op;
drivers/mtd/nand/spi/core.c
1643
op = spinand_select_op_variant(spinand, ODTR,
drivers/mtd/nand/spi/core.c
1645
spinand->odtr_op_templates.update_cache = op;
drivers/mtd/nand/spi/core.c
33
struct spi_mem_op op = spinand->op_templates->readid;
drivers/mtd/nand/spi/core.c
35
op.addr.nbytes = naddr;
drivers/mtd/nand/spi/core.c
36
op.dummy.nbytes = ndummy;
drivers/mtd/nand/spi/core.c
37
op.data.buf.in = buf;
drivers/mtd/nand/spi/core.c
38
op.data.nbytes = len;
drivers/mtd/nand/spi/core.c
40
return op;
drivers/mtd/nand/spi/core.c
446
struct spi_mem_op op = SPINAND_OP(spinand, wr_en);
drivers/mtd/nand/spi/core.c
448
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
456
struct spi_mem_op op = SPINAND_OP(spinand, page_read, row);
drivers/mtd/nand/spi/core.c
458
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
58
struct spi_mem_op op = spinand->op_templates->set_feature;
drivers/mtd/nand/spi/core.c
60
if (op.cmd.dtr && op.cmd.buswidth == 8)
drivers/mtd/nand/spi/core.c
611
struct spi_mem_op op = SPINAND_OP(spinand, prog_exec, row);
drivers/mtd/nand/spi/core.c
613
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
621
struct spi_mem_op op = SPINAND_OP(spinand, blk_erase, row);
drivers/mtd/nand/spi/core.c
623
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
63
op.addr.val = reg;
drivers/mtd/nand/spi/core.c
64
op.data.buf.out = valptr;
drivers/mtd/nand/spi/core.c
641
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/core.c
646
ret = spi_mem_poll_status(spinand->spimem, &op, STATUS_BUSY, 0,
drivers/mtd/nand/spi/core.c
66
return op;
drivers/mtd/nand/spi/core.c
675
struct spi_mem_op op = SPINAND_OP(spinand, readid,
drivers/mtd/nand/spi/core.c
679
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
688
struct spi_mem_op op = SPINAND_OP(spinand, reset);
drivers/mtd/nand/spi/core.c
691
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/core.c
72
struct spi_mem_op op = spinand->op_templates->get_feature;
drivers/mtd/nand/spi/core.c
74
if (op.cmd.dtr && op.cmd.buswidth == 8)
drivers/mtd/nand/spi/core.c
77
op.addr.val = reg;
drivers/mtd/nand/spi/core.c
78
op.data.buf.in = valptr;
drivers/mtd/nand/spi/core.c
80
return op;
drivers/mtd/nand/spi/core.c
86
struct spi_mem_op op = spinand->op_templates->blk_erase;
drivers/mtd/nand/spi/core.c
88
op.addr.val = addr;
drivers/mtd/nand/spi/core.c
90
return op;
drivers/mtd/nand/spi/core.c
96
struct spi_mem_op op = spinand->op_templates->page_read;
drivers/mtd/nand/spi/core.c
98
op.addr.val = addr;
drivers/mtd/nand/spi/gigadevice.c
269
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/gigadevice.c
282
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/gigadevice.c
312
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/gigadevice.c
325
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/gigadevice.c
82
struct spi_mem_op op = SPI_MEM_OP(SPI_MEM_OP_CMD(0x7c, 1),
drivers/mtd/nand/spi/gigadevice.c
88
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/macronix.c
166
struct spi_mem_op op = SPINAND_OP(spinand, set_feature,
drivers/mtd/nand/spi/macronix.c
170
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/macronix.c
87
struct spi_mem_op op = SPINAND_OP(spinand, macronix_read_eccsr, eccsr);
drivers/mtd/nand/spi/macronix.c
90
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/micron.c
140
struct spi_mem_op op = SPINAND_OP(spinand, set_feature,
drivers/mtd/nand/spi/micron.c
148
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/toshiba.c
76
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/toshiba.c
93
if (spi_mem_exec_op(spinand->spimem, &op))
drivers/mtd/nand/spi/winbond.c
167
struct spi_mem_op op = SPINAND_OP(spinand, winbond_select_target,
drivers/mtd/nand/spi/winbond.c
171
return spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/nand/spi/winbond.c
295
struct spi_mem_op op = SPINAND_OP(spinand, get_feature,
drivers/mtd/nand/spi/winbond.c
312
if (spi_mem_exec_op(spinand->spimem, &op))
drivers/mtd/nand/spi/winbond.c
332
const struct spi_mem_op *op;
drivers/mtd/nand/spi/winbond.c
340
op = spinand->op_templates->read_cache;
drivers/mtd/nand/spi/winbond.c
341
if (op->cmd.dtr || op->addr.dtr || op->dummy.dtr || op->data.dtr)
drivers/mtd/nand/spi/winbond.c
343
else if (op->cmd.buswidth == 1 && op->addr.buswidth == 1 &&
drivers/mtd/nand/spi/winbond.c
344
op->dummy.buswidth == 1 && op->data.buswidth == 1)
drivers/mtd/nand/spi/winbond.c
346
else if (!op->max_freq)
drivers/mtd/nand/spi/winbond.c
369
struct spi_mem_op op = SPINAND_OP(spinand, winbond_write_vcr,
drivers/mtd/nand/spi/winbond.c
379
ret = spi_mem_exec_op(spinand->spimem, &op);
drivers/mtd/spi-nor/controllers/hisi-sfc.c
68
#define OP_CTRL_RW_OP(op) ((op) << 1)
drivers/mtd/spi-nor/controllers/nxp-spifi.c
39
#define SPIFI_CMD_OPCODE(op) ((op) << 24)
drivers/mtd/spi-nor/core.c
1013
struct spi_mem_op op = SPI_NOR_WRSR2_OP(sr2);
drivers/mtd/spi-nor/core.c
1015
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
1017
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
1045
struct spi_mem_op op = SPI_NOR_RDSR2_OP(sr2);
drivers/mtd/spi-nor/core.c
1047
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
1049
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
106
op->cmd.dtr = true;
drivers/mtd/spi-nor/core.c
107
op->addr.dtr = true;
drivers/mtd/spi-nor/core.c
1077
struct spi_mem_op op =
drivers/mtd/spi-nor/core.c
108
op->dummy.dtr = true;
drivers/mtd/spi-nor/core.c
1081
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
1083
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
109
op->data.dtr = true;
drivers/mtd/spi-nor/core.c
112
op->dummy.nbytes *= 2;
drivers/mtd/spi-nor/core.c
114
ext = spi_nor_get_cmd_ext(nor, op);
drivers/mtd/spi-nor/core.c
115
op->cmd.opcode = (op->cmd.opcode << 8) | ext;
drivers/mtd/spi-nor/core.c
116
op->cmd.nbytes = 2;
drivers/mtd/spi-nor/core.c
120
op->data.swap16 = true;
drivers/mtd/spi-nor/core.c
133
static bool spi_nor_spimem_bounce(struct spi_nor *nor, struct spi_mem_op *op)
drivers/mtd/spi-nor/core.c
136
if (object_is_on_stack(op->data.buf.in) ||
drivers/mtd/spi-nor/core.c
137
!virt_addr_valid(op->data.buf.in)) {
drivers/mtd/spi-nor/core.c
138
if (op->data.nbytes > nor->bouncebuf_size)
drivers/mtd/spi-nor/core.c
139
op->data.nbytes = nor->bouncebuf_size;
drivers/mtd/spi-nor/core.c
140
op->data.buf.in = nor->bouncebuf;
drivers/mtd/spi-nor/core.c
1451
struct spi_mem_op op =
drivers/mtd/spi-nor/core.c
1455
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
1457
return spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
154
static int spi_nor_spimem_exec_op(struct spi_nor *nor, struct spi_mem_op *op)
drivers/mtd/spi-nor/core.c
158
error = spi_mem_adjust_op_size(nor->spimem, op);
drivers/mtd/spi-nor/core.c
162
return spi_mem_exec_op(nor->spimem, op);
drivers/mtd/spi-nor/core.c
204
struct spi_mem_op op =
drivers/mtd/spi-nor/core.c
213
spi_nor_spimem_setup_op(nor, &op, nor->read_proto);
drivers/mtd/spi-nor/core.c
216
op.dummy.nbytes = (nor->read_dummy * op.dummy.buswidth) / 8;
drivers/mtd/spi-nor/core.c
218
op.dummy.nbytes *= 2;
drivers/mtd/spi-nor/core.c
220
usebouncebuf = spi_nor_spimem_bounce(nor, &op);
drivers/mtd/spi-nor/core.c
223
nbytes = spi_mem_dirmap_read(nor->dirmap.rdesc, op.addr.val,
drivers/mtd/spi-nor/core.c
224
op.data.nbytes, op.data.buf.in);
drivers/mtd/spi-nor/core.c
226
error = spi_nor_spimem_exec_op(nor, &op);
drivers/mtd/spi-nor/core.c
229
nbytes = op.data.nbytes;
drivers/mtd/spi-nor/core.c
233
memcpy(buf, op.data.buf.in, nbytes);
drivers/mtd/spi-nor/core.c
2356
struct spi_mem_op *op)
drivers/mtd/spi-nor/core.c
2364
op->addr.nbytes = 4;
drivers/mtd/spi-nor/core.c
2365
if (!spi_mem_supports_op(nor->spimem, op)) {
drivers/mtd/spi-nor/core.c
2370
op->addr.nbytes = 3;
drivers/mtd/spi-nor/core.c
2371
if (!spi_mem_supports_op(nor->spimem, op))
drivers/mtd/spi-nor/core.c
2389
struct spi_mem_op op = SPI_NOR_READ_OP(read->opcode);
drivers/mtd/spi-nor/core.c
2391
spi_nor_spimem_setup_op(nor, &op, read->proto);
drivers/mtd/spi-nor/core.c
2394
op.dummy.nbytes = (read->num_mode_clocks + read->num_wait_states) *
drivers/mtd/spi-nor/core.c
2395
op.dummy.buswidth / 8;
drivers/mtd/spi-nor/core.c
2397
op.dummy.nbytes *= 2;
drivers/mtd/spi-nor/core.c
2399
return spi_nor_spimem_check_read_pp_op(nor, &op);
drivers/mtd/spi-nor/core.c
2413
struct spi_mem_op op = SPI_NOR_PP_OP(pp->opcode);
drivers/mtd/spi-nor/core.c
2415
spi_nor_spimem_setup_op(nor, &op, pp->proto);
drivers/mtd/spi-nor/core.c
2417
return spi_nor_spimem_check_read_pp_op(nor, &op);
drivers/mtd/spi-nor/core.c
2465
struct spi_mem_op op = SPI_NOR_RDCR_OP(nor->bouncebuf);
drivers/mtd/spi-nor/core.c
2467
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
2469
if (!spi_mem_supports_op(nor->spimem, &op))
drivers/mtd/spi-nor/core.c
268
struct spi_mem_op op =
drivers/mtd/spi-nor/core.c
277
op.addr.nbytes = 0;
drivers/mtd/spi-nor/core.c
279
spi_nor_spimem_setup_op(nor, &op, nor->write_proto);
drivers/mtd/spi-nor/core.c
281
if (spi_nor_spimem_bounce(nor, &op))
drivers/mtd/spi-nor/core.c
282
memcpy(nor->bouncebuf, buf, op.data.nbytes);
drivers/mtd/spi-nor/core.c
285
nbytes = spi_mem_dirmap_write(nor->dirmap.wdesc, op.addr.val,
drivers/mtd/spi-nor/core.c
286
op.data.nbytes, op.data.buf.out);
drivers/mtd/spi-nor/core.c
288
error = spi_nor_spimem_exec_op(nor, &op);
drivers/mtd/spi-nor/core.c
291
nbytes = op.data.nbytes;
drivers/mtd/spi-nor/core.c
324
int spi_nor_read_any_reg(struct spi_nor *nor, struct spi_mem_op *op,
drivers/mtd/spi-nor/core.c
3294
struct spi_mem_op op;
drivers/mtd/spi-nor/core.c
3297
op = (struct spi_mem_op)SPINOR_SRSTEN_OP;
drivers/mtd/spi-nor/core.c
3299
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
330
spi_nor_spimem_setup_op(nor, op, proto);
drivers/mtd/spi-nor/core.c
3301
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
3308
op = (struct spi_mem_op)SPINOR_SRST_OP;
drivers/mtd/spi-nor/core.c
331
return spi_nor_spimem_exec_op(nor, op);
drivers/mtd/spi-nor/core.c
3310
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
3312
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
346
int spi_nor_write_any_volatile_reg(struct spi_nor *nor, struct spi_mem_op *op,
drivers/mtd/spi-nor/core.c
357
spi_nor_spimem_setup_op(nor, op, proto);
drivers/mtd/spi-nor/core.c
358
return spi_nor_spimem_exec_op(nor, op);
drivers/mtd/spi-nor/core.c
3651
struct spi_mem_op *op = &info.op_tmpl;
drivers/mtd/spi-nor/core.c
3653
spi_nor_spimem_setup_op(nor, op, nor->read_proto);
drivers/mtd/spi-nor/core.c
3656
op->dummy.nbytes = (nor->read_dummy * op->dummy.buswidth) / 8;
drivers/mtd/spi-nor/core.c
3658
op->dummy.nbytes *= 2;
drivers/mtd/spi-nor/core.c
3665
op->data.buswidth = spi_nor_get_protocol_data_nbits(nor->read_proto);
drivers/mtd/spi-nor/core.c
3682
struct spi_mem_op *op = &info.op_tmpl;
drivers/mtd/spi-nor/core.c
3685
op->addr.nbytes = 0;
drivers/mtd/spi-nor/core.c
3687
spi_nor_spimem_setup_op(nor, op, nor->write_proto);
drivers/mtd/spi-nor/core.c
3694
op->data.buswidth = spi_nor_get_protocol_data_nbits(nor->write_proto);
drivers/mtd/spi-nor/core.c
372
struct spi_mem_op op = SPI_NOR_WREN_OP;
drivers/mtd/spi-nor/core.c
374
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
376
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
399
struct spi_mem_op op = SPI_NOR_WRDI_OP;
drivers/mtd/spi-nor/core.c
401
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
403
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
434
struct spi_mem_op op =
drivers/mtd/spi-nor/core.c
437
spi_nor_spimem_setup_op(nor, &op, proto);
drivers/mtd/spi-nor/core.c
438
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
459
struct spi_mem_op op = SPI_NOR_RDSR_OP(sr);
drivers/mtd/spi-nor/core.c
462
op.addr.nbytes = nor->params->rdsr_addr_nbytes;
drivers/mtd/spi-nor/core.c
463
op.dummy.nbytes = nor->params->rdsr_dummy;
drivers/mtd/spi-nor/core.c
468
op.data.nbytes = 2;
drivers/mtd/spi-nor/core.c
471
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
473
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
499
struct spi_mem_op op = SPI_NOR_RDCR_OP(cr);
drivers/mtd/spi-nor/core.c
501
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
503
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
530
struct spi_mem_op op = SPI_NOR_EN4B_EX4B_OP(enable);
drivers/mtd/spi-nor/core.c
532
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
534
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
594
struct spi_mem_op op = SPI_NOR_BRWR_OP(nor->bouncebuf);
drivers/mtd/spi-nor/core.c
596
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
598
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
60
const struct spi_mem_op *op)
drivers/mtd/spi-nor/core.c
64
return ~op->cmd.opcode;
drivers/mtd/spi-nor/core.c
67
return op->cmd.opcode;
drivers/mtd/spi-nor/core.c
768
struct spi_mem_op op = SPI_NOR_GBULK_OP;
drivers/mtd/spi-nor/core.c
770
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
772
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
803
struct spi_mem_op op = SPI_NOR_WRSR_OP(sr, len);
drivers/mtd/spi-nor/core.c
805
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/core.c
807
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/core.c
83
struct spi_mem_op *op,
drivers/mtd/spi-nor/core.c
88
op->cmd.buswidth = spi_nor_get_protocol_inst_nbits(proto);
drivers/mtd/spi-nor/core.c
90
if (op->addr.nbytes)
drivers/mtd/spi-nor/core.c
91
op->addr.buswidth = spi_nor_get_protocol_addr_nbits(proto);
drivers/mtd/spi-nor/core.c
93
if (op->dummy.nbytes)
drivers/mtd/spi-nor/core.c
94
op->dummy.buswidth = spi_nor_get_protocol_addr_nbits(proto);
drivers/mtd/spi-nor/core.c
96
if (op->data.nbytes)
drivers/mtd/spi-nor/core.c
97
op->data.buswidth = spi_nor_get_protocol_data_nbits(proto);
drivers/mtd/spi-nor/core.h
611
struct spi_mem_op *op,
drivers/mtd/spi-nor/core.h
640
int spi_nor_read_any_reg(struct spi_nor *nor, struct spi_mem_op *op,
drivers/mtd/spi-nor/core.h
642
int spi_nor_write_any_volatile_reg(struct spi_nor *nor, struct spi_mem_op *op,
drivers/mtd/spi-nor/macronix.c
244
struct spi_mem_op op;
drivers/mtd/spi-nor/macronix.c
250
op = (struct spi_mem_op)MXIC_NOR_WR_CR2(MXIC_NOR_ADDR_CR2_DC, 1, buf);
drivers/mtd/spi-nor/macronix.c
251
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/macronix.c
257
op = (struct spi_mem_op)MXIC_NOR_WR_CR2(MXIC_NOR_ADDR_CR2_MODE, 1, buf);
drivers/mtd/spi-nor/macronix.c
258
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/macronix.c
280
struct spi_mem_op op;
drivers/mtd/spi-nor/macronix.c
292
op = (struct spi_mem_op)MXIC_NOR_WR_CR2(MXIC_NOR_ADDR_CR2_MODE, 2, buf);
drivers/mtd/spi-nor/macronix.c
293
ret = spi_nor_write_any_volatile_reg(nor, &op, SNOR_PROTO_8_8_8_DTR);
drivers/mtd/spi-nor/micron-st.c
104
op = (struct spi_mem_op)
drivers/mtd/spi-nor/micron-st.c
107
ret = spi_nor_write_any_volatile_reg(nor, &op, SNOR_PROTO_8_8_8_DTR);
drivers/mtd/spi-nor/micron-st.c
527
struct spi_mem_op op = MICRON_ST_RDFSR_OP(fsr);
drivers/mtd/spi-nor/micron-st.c
53
struct spi_mem_op op;
drivers/mtd/spi-nor/micron-st.c
530
op.addr.nbytes = nor->params->rdsr_addr_nbytes;
drivers/mtd/spi-nor/micron-st.c
531
op.dummy.nbytes = nor->params->rdsr_dummy;
drivers/mtd/spi-nor/micron-st.c
536
op.data.nbytes = 2;
drivers/mtd/spi-nor/micron-st.c
539
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/micron-st.c
541
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/micron-st.c
562
struct spi_mem_op op = MICRON_ST_CLFSR_OP;
drivers/mtd/spi-nor/micron-st.c
564
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/micron-st.c
566
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/micron-st.c
60
op = (struct spi_mem_op)
drivers/mtd/spi-nor/micron-st.c
63
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/micron-st.c
68
op = (struct spi_mem_op)
drivers/mtd/spi-nor/micron-st.c
71
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/micron-st.c
90
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
111
struct spi_mem_op op =
drivers/mtd/spi-nor/spansion.c
117
op.addr.nbytes = nor->addr_nbytes;
drivers/mtd/spi-nor/spansion.c
118
op.dummy.nbytes = params->rdsr_dummy;
drivers/mtd/spi-nor/spansion.c
119
op.data.nbytes = 2;
drivers/mtd/spi-nor/spansion.c
122
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
172
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
177
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
180
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
188
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
191
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
202
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
207
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
211
return spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
249
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
259
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
261
return spi_nor_write_any_volatile_reg(nor, &op, SNOR_PROTO_8_8_8_DTR);
drivers/mtd/spi-nor/spansion.c
293
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
298
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
302
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
311
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
314
ret = spi_nor_write_any_volatile_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
321
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
324
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
370
struct spi_mem_op op = CYPRESS_NOR_EN4B_EX4B_OP(enable);
drivers/mtd/spi-nor/spansion.c
372
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
374
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/spansion.c
397
struct spi_mem_op op =
drivers/mtd/spi-nor/spansion.c
407
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
413
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
416
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
445
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
471
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
474
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
504
struct spi_mem_op op =
drivers/mtd/spi-nor/spansion.c
517
op.addr.val = params->vreg_offset[i] + SPINOR_REG_CYPRESS_CFR3;
drivers/mtd/spi-nor/spansion.c
519
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
548
struct spi_mem_op op;
drivers/mtd/spi-nor/spansion.c
559
op = (struct spi_mem_op)
drivers/mtd/spi-nor/spansion.c
563
ret = spi_nor_read_any_reg(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
94
struct spi_mem_op op = SPANSION_OP(priv_params->clsr);
drivers/mtd/spi-nor/spansion.c
96
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/spansion.c
98
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/sst.c
173
u8 op = (len == 1) ? SPINOR_OP_BP : SPINOR_OP_AAI_WP;
drivers/mtd/spi-nor/sst.c
176
nor->program_opcode = op;
drivers/mtd/spi-nor/winbond.c
100
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/winbond.c
387
struct spi_mem_op op = WINBOND_NOR_WREAR_OP(nor->bouncebuf);
drivers/mtd/spi-nor/winbond.c
389
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/spi-nor/winbond.c
391
ret = spi_mem_exec_op(nor->spimem, &op);
drivers/mtd/spi-nor/winbond.c
96
struct spi_mem_op op = WINBOND_NOR_SELDIE_OP(nor->bouncebuf);
drivers/mtd/spi-nor/winbond.c
98
spi_nor_spimem_setup_op(nor, &op, nor->reg_proto);
drivers/mtd/tests/stresstest.c
130
int i, op;
drivers/mtd/tests/stresstest.c
197
for (op = 0; op < count; op++) {
drivers/mtd/tests/stresstest.c
198
if ((op & 1023) == 0)
drivers/mtd/tests/stresstest.c
199
pr_info("%d operations done\n", op);
drivers/mtd/tests/stresstest.c
208
pr_info("finished, %d operations done\n", op);
drivers/net/can/m_can/m_can.c
1986
struct m_can_tx_op *op = container_of(ws, struct m_can_tx_op, work);
drivers/net/can/m_can/m_can.c
1987
struct m_can_classdev *cdev = op->cdev;
drivers/net/can/m_can/m_can.c
1988
struct sk_buff *skb = op->skb;
drivers/net/can/m_can/m_can.c
1990
op->skb = NULL;
drivers/net/can/m_can/m_can.c
1992
if (op->submit)
drivers/net/dsa/b53/b53_common.c
1825
static int b53_arl_rw_op(struct b53_device *dev, unsigned int op)
drivers/net/dsa/b53/b53_common.c
1829
if (op > ARLTBL_RW)
drivers/net/dsa/b53/b53_common.c
1834
if (op)
drivers/net/dsa/b53/b53_common.c
1955
static int b53_arl_op(struct b53_device *dev, int op, int port,
drivers/net/dsa/b53/b53_common.c
1983
if (op)
drivers/net/dsa/b53/b53_common.c
233
static int b53_do_vlan_op(struct b53_device *dev, u8 op)
drivers/net/dsa/b53/b53_common.c
237
b53_write8(dev, B53_ARLIO_PAGE, dev->vta_regs[0], VTA_START_CMD | op);
drivers/net/dsa/b53/b53_mdio.c
42
static int b53_mdio_op(struct b53_device *dev, u8 page, u8 reg, u16 op)
drivers/net/dsa/b53/b53_mdio.c
60
v = (reg << 8) | op;
drivers/net/dsa/b53/b53_srab.c
125
static int b53_srab_op(struct b53_device *dev, u8 page, u8 reg, u32 op)
drivers/net/dsa/b53/b53_srab.c
136
op;
drivers/net/dsa/bcm_sf2.c
381
static int bcm_sf2_sw_indir_rw(struct bcm_sf2_priv *priv, int op, int addr,
drivers/net/dsa/bcm_sf2.c
400
if (op)
drivers/net/dsa/bcm_sf2_cfp.c
169
static int bcm_sf2_cfp_op(struct bcm_sf2_priv *priv, unsigned int op)
drivers/net/dsa/bcm_sf2_cfp.c
176
reg |= OP_STR_DONE | op;
drivers/net/dsa/mv88e6xxx/devlink.c
425
u16 op;
drivers/net/dsa/mv88e6xxx/devlink.c
466
&entry->op);
drivers/net/dsa/mv88e6xxx/global1_atu.c
132
static int mv88e6xxx_g1_atu_op(struct mv88e6xxx_chip *chip, u16 fid, u16 op)
drivers/net/dsa/mv88e6xxx/global1_atu.c
157
op |= (fid & 0x30) << 4;
drivers/net/dsa/mv88e6xxx/global1_atu.c
161
op |= fid & 0xf;
drivers/net/dsa/mv88e6xxx/global1_atu.c
165
MV88E6XXX_G1_ATU_OP_BUSY | op);
drivers/net/dsa/mv88e6xxx/global1_atu.c
179
u16 val = 0, upper = 0, op = 0;
drivers/net/dsa/mv88e6xxx/global1_atu.c
188
err = mv88e6xxx_g1_read(chip, MV88E6XXX_G1_ATU_OP, &op);
drivers/net/dsa/mv88e6xxx/global1_atu.c
201
upper = (op >> 4) & 0x30;
drivers/net/dsa/mv88e6xxx/global1_atu.c
205
val = (op & 0xf) | upper;
drivers/net/dsa/mv88e6xxx/global1_atu.c
340
u16 op;
drivers/net/dsa/mv88e6xxx/global1_atu.c
353
op = MV88E6XXX_G1_ATU_OP_FLUSH_MOVE_ALL_DB;
drivers/net/dsa/mv88e6xxx/global1_atu.c
355
op = MV88E6XXX_G1_ATU_OP_FLUSH_MOVE_NON_STATIC_DB;
drivers/net/dsa/mv88e6xxx/global1_atu.c
357
op = MV88E6XXX_G1_ATU_OP_FLUSH_MOVE_ALL;
drivers/net/dsa/mv88e6xxx/global1_atu.c
359
op = MV88E6XXX_G1_ATU_OP_FLUSH_MOVE_NON_STATIC;
drivers/net/dsa/mv88e6xxx/global1_atu.c
361
return mv88e6xxx_g1_atu_op(chip, fid, op);
drivers/net/dsa/mv88e6xxx/global1_vtu.c
376
u16 op = MV88E6XXX_G1_VTU_OP_VTU_LOAD_PURGE;
drivers/net/dsa/mv88e6xxx/global1_vtu.c
399
op |= entry->fid & 0x000f;
drivers/net/dsa/mv88e6xxx/global1_vtu.c
400
op |= (entry->fid & 0x00f0) << 4;
drivers/net/dsa/mv88e6xxx/global1_vtu.c
403
return mv88e6xxx_g1_vtu_op(chip, op);
drivers/net/dsa/mv88e6xxx/global1_vtu.c
78
static int mv88e6xxx_g1_vtu_op(struct mv88e6xxx_chip *chip, u16 op)
drivers/net/dsa/mv88e6xxx/global1_vtu.c
83
MV88E6XXX_G1_VTU_OP_BUSY | op);
drivers/net/dsa/mv88e6xxx/global2.c
186
static int mv88e6xxx_g2_irl_op(struct mv88e6xxx_chip *chip, u16 op, int port,
drivers/net/dsa/mv88e6xxx/global2.c
192
MV88E6XXX_G2_IRL_CMD_BUSY | op | (port << 8) |
drivers/net/dsa/mv88e6xxx/global2.c
224
int src_port, u16 op)
drivers/net/dsa/mv88e6xxx/global2.c
231
op |= MV88E6XXX_G2_PVT_ADDR_BUSY;
drivers/net/dsa/mv88e6xxx/global2.c
232
op |= (src_dev & 0x1f) << 4;
drivers/net/dsa/mv88e6xxx/global2.c
233
op |= (src_port & 0xf);
drivers/net/dsa/mv88e6xxx/global2.c
235
err = mv88e6xxx_g2_write(chip, MV88E6XXX_G2_PVT_ADDR, op);
drivers/net/dsa/mv88e6xxx/global2.c
634
bool external, bool c45, u16 op, int dev,
drivers/net/dsa/mv88e6xxx/global2.c
637
u16 cmd = op;
drivers/net/dsa/mv88e6xxx/global2.c
657
bool external, u16 op, int dev,
drivers/net/dsa/mv88e6xxx/global2.c
660
return mv88e6xxx_g2_smi_phy_access(chip, external, false, op, dev, reg);
drivers/net/dsa/mv88e6xxx/global2.c
668
u16 op = MV88E6XXX_G2_SMI_PHY_CMD_OP_22_READ_DATA;
drivers/net/dsa/mv88e6xxx/global2.c
675
err = mv88e6xxx_g2_smi_phy_access_c22(chip, external, op, dev, reg);
drivers/net/dsa/mv88e6xxx/global2.c
687
u16 op = MV88E6XXX_G2_SMI_PHY_CMD_OP_22_WRITE_DATA;
drivers/net/dsa/mv88e6xxx/global2.c
698
return mv88e6xxx_g2_smi_phy_access_c22(chip, external, op, dev, reg);
drivers/net/dsa/mv88e6xxx/global2.c
702
bool external, u16 op, int port,
drivers/net/dsa/mv88e6xxx/global2.c
705
return mv88e6xxx_g2_smi_phy_access(chip, external, true, op, port, dev);
drivers/net/dsa/mv88e6xxx/global2.c
713
u16 op = MV88E6XXX_G2_SMI_PHY_CMD_OP_45_WRITE_ADDR;
drivers/net/dsa/mv88e6xxx/global2.c
724
return mv88e6xxx_g2_smi_phy_access_c45(chip, external, op, port, dev);
drivers/net/dsa/mv88e6xxx/global2.c
732
u16 op = MV88E6XXX_G2_SMI_PHY_CMD_OP_45_READ_DATA;
drivers/net/dsa/mv88e6xxx/global2.c
735
err = mv88e6xxx_g2_smi_phy_access_c45(chip, external, op, port, dev);
drivers/net/dsa/mv88e6xxx/global2.c
762
u16 op = MV88E6XXX_G2_SMI_PHY_CMD_OP_45_WRITE_DATA;
drivers/net/dsa/mv88e6xxx/global2.c
769
return mv88e6xxx_g2_smi_phy_access_c45(chip, external, op, port, dev);
drivers/net/dsa/sja1105/sja1105.h
150
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_clocking.c
103
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
107
sja1105_packing(buf, &idiv->clksrc, 28, 24, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
108
sja1105_packing(buf, &idiv->autoblock, 11, 11, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
109
sja1105_packing(buf, &idiv->idiv, 5, 2, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
110
sja1105_packing(buf, &idiv->pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
142
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
146
sja1105_packing(buf, &cmd->clksrc, 28, 24, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
147
sja1105_packing(buf, &cmd->autoblock, 11, 11, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
148
sja1105_packing(buf, &cmd->pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
324
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
328
sja1105_packing(buf, &cmd->pllclksrc, 28, 24, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
329
sja1105_packing(buf, &cmd->msel, 23, 16, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
330
sja1105_packing(buf, &cmd->autoblock, 11, 11, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
331
sja1105_packing(buf, &cmd->psel, 9, 8, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
332
sja1105_packing(buf, &cmd->direct, 7, 7, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
333
sja1105_packing(buf, &cmd->fbsel, 6, 6, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
334
sja1105_packing(buf, &cmd->bypass, 1, 1, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
335
sja1105_packing(buf, &cmd->pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
377
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
381
sja1105_packing(buf, &cmd->d32_os, 28, 27, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
382
sja1105_packing(buf, &cmd->d32_ih, 26, 26, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
383
sja1105_packing(buf, &cmd->d32_ipud, 25, 24, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
384
sja1105_packing(buf, &cmd->d10_os, 20, 19, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
385
sja1105_packing(buf, &cmd->d10_ih, 18, 18, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
386
sja1105_packing(buf, &cmd->d10_ipud, 17, 16, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
387
sja1105_packing(buf, &cmd->ctrl_os, 12, 11, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
388
sja1105_packing(buf, &cmd->ctrl_ih, 10, 10, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
389
sja1105_packing(buf, &cmd->ctrl_ipud, 9, 8, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
390
sja1105_packing(buf, &cmd->clk_os, 4, 3, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
391
sja1105_packing(buf, &cmd->clk_ih, 2, 2, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
392
sja1105_packing(buf, &cmd->clk_ipud, 1, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
463
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
467
sja1105_packing(buf, &cmd->rxc_stable_ovr, 15, 15, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
468
sja1105_packing(buf, &cmd->rxc_delay, 14, 10, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
469
sja1105_packing(buf, &cmd->rxc_bypass, 9, 9, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
470
sja1105_packing(buf, &cmd->rxc_pd, 8, 8, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
471
sja1105_packing(buf, &cmd->txc_stable_ovr, 7, 7, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
472
sja1105_packing(buf, &cmd->txc_delay, 6, 2, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
473
sja1105_packing(buf, &cmd->txc_bypass, 1, 1, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
474
sja1105_packing(buf, &cmd->txc_pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
479
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
494
sja1105_packing(buf, &cmd->rxc_stable_ovr, 26, 26, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
495
sja1105_packing(buf, &cmd->rxc_delay, 25, 21, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
496
sja1105_packing(buf, &range, 20, 18, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
497
sja1105_packing(buf, &cmd->rxc_bypass, 17, 17, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
498
sja1105_packing(buf, &cmd->rxc_pd, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
499
sja1105_packing(buf, &cmd->txc_stable_ovr, 10, 10, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
500
sja1105_packing(buf, &cmd->txc_delay, 9, 5, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
501
sja1105_packing(buf, &range, 4, 2, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
502
sja1105_packing(buf, &cmd->txc_bypass, 1, 1, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
503
sja1105_packing(buf, &cmd->txc_pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
825
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_clocking.c
829
sja1105_packing(buf, &outclk->clksrc, 27, 24, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
830
sja1105_packing(buf, &outclk->autoblock, 11, 11, size, op);
drivers/net/dsa/sja1105/sja1105_clocking.c
831
sja1105_packing(buf, &outclk->pd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
200
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
204
sja1105_packing(buf, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
205
sja1105_packing(buf, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
206
sja1105_packing(buf, &cmd->rdwrset, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
207
sja1105_packing(buf, &cmd->index, 9, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
213
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
218
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
219
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
220
sja1105_packing(p, &cmd->rdwrset, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
221
sja1105_packing(p, &cmd->index, 9, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
226
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
231
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
232
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
233
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
234
sja1105_packing(p, &cmd->index, 11, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
238
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
243
sja1105_packing(buf, &entry->egrmirr, 21, 17, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
244
sja1105_packing(buf, &entry->ingrmirr, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
250
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
255
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
256
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
257
sja1105_packing(p, &cmd->index, 11, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
262
enum packing_op op, int entry_size)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
268
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
269
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
270
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
271
sja1105_packing(p, &cmd->valident, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
306
sja1105_packing(p, &hostcmd, 25, 23, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
311
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
315
sja1105pqrs_common_l2_lookup_cmd_packing(buf, cmd, op, entry_size);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
325
sja1105_packing(buf, &cmd->index, 15, 6, entry_size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
330
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
334
sja1105pqrs_common_l2_lookup_cmd_packing(buf, cmd, op, entry_size);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
336
sja1105_packing(buf, &cmd->index, 10, 1, entry_size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
386
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
392
sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
394
return sja1105pqrs_l2_lookup_entry_packing(buf, entry_ptr, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
398
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
404
sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
406
return sja1110_l2_lookup_entry_packing(buf, entry_ptr, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
411
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
416
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
417
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
418
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
419
sja1105_packing(p, &cmd->valident, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
422
SJA1105ET_SIZE_L2_LOOKUP_ENTRY, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
426
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
432
sja1105_packing(cmd, &entry->lockeds, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
434
return sja1105et_l2_lookup_entry_packing(buf, entry_ptr, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
439
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
444
sja1105et_l2_lookup_cmd_packing(buf, cmd, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
445
if (op == PACK)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
450
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
461
sja1105_packing(buf, &entry->tsreg, 85, 85, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
462
sja1105_packing(buf, &entry->takets, 84, 84, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
463
sja1105_packing(buf, &entry->macaddr, 83, 36, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
464
sja1105_packing(buf, &entry->destports, 35, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
465
sja1105_packing(buf, &entry->enfport, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
471
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
476
sja1105pqrs_l2_lookup_cmd_packing(buf, cmd, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
477
if (op == PACK)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
482
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
491
sja1105_packing(buf, &entry->tsreg, 71, 71, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
492
sja1105_packing(buf, &entry->takets, 70, 70, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
493
sja1105_packing(buf, &entry->macaddr, 69, 22, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
494
sja1105_packing(buf, &entry->destports, 21, 17, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
495
sja1105_packing(buf, &entry->enfport, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
505
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
510
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
511
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
512
sja1105_packing(p, &cmd->valident, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
517
SJA1105_SIZE_VLAN_LOOKUP_ENTRY, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
523
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
529
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
530
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
531
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
536
SJA1110_SIZE_VLAN_LOOKUP_ENTRY, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
543
if (op == PACK && !cmd->valident) {
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
546
} else if (op == UNPACK) {
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
555
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
560
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
561
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
562
sja1105_packing(p, &cmd->rdwrset, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
563
sja1105_packing(p, &cmd->index, 4, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
568
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
573
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
574
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
575
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
576
sja1105_packing(p, &cmd->index, 4, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
581
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
587
sja1105_packing(reg1, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
588
sja1105_packing(reg1, &cmd->index, 26, 24, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
592
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
600
sja1105_packing(reg1, &entry->speed, 30, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
601
sja1105_packing(reg1, &entry->drpdtag, 23, 23, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
602
sja1105_packing(reg1, &entry->drpuntag, 22, 22, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
603
sja1105_packing(reg1, &entry->retag, 21, 21, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
604
sja1105_packing(reg1, &entry->dyn_learn, 20, 20, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
605
sja1105_packing(reg1, &entry->egress, 19, 19, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
606
sja1105_packing(reg1, &entry->ingress, 18, 18, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
607
sja1105_packing(reg1, &entry->ing_mirr, 17, 17, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
608
sja1105_packing(reg1, &entry->egr_mirr, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
609
sja1105_packing(reg1, &entry->vlanprio, 14, 12, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
610
sja1105_packing(reg1, &entry->vlanid, 11, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
611
sja1105_packing(reg2, &entry->tp_delin, 31, 16, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
612
sja1105_packing(reg2, &entry->tp_delout, 15, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
622
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
627
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
628
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
629
sja1105_packing(p, &cmd->rdwrset, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
630
sja1105_packing(p, &cmd->index, 2, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
635
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
640
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
641
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
642
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
643
sja1105_packing(p, &cmd->index, 3, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
648
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
651
SJA1105ET_SIZE_L2_LOOKUP_PARAMS_DYN_CMD, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
656
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
661
SJA1105ET_SIZE_L2_LOOKUP_PARAMS_DYN_CMD, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
669
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
674
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
675
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
680
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
685
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
686
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
687
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
692
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
696
sja1105_packing(buf, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
697
sja1105_packing(buf, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
702
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
707
sja1105_packing(buf, &entry->mirr_port, 2, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
714
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
719
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
720
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
721
sja1105_packing(p, &cmd->rdwrset, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
726
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
731
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
732
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
733
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
738
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
743
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
744
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
745
sja1105_packing(p, &cmd->rdwrset, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
750
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
755
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
756
sja1105_packing(p, &cmd->errors, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
757
sja1105_packing(p, &cmd->valident, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
758
sja1105_packing(p, &cmd->rdwrset, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
759
sja1105_packing(p, &cmd->index, 5, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
764
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
769
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
770
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
771
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
772
sja1105_packing(p, &cmd->valident, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
773
sja1105_packing(p, &cmd->index, 4, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
777
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
782
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
783
sja1105_packing(p, &cmd->index, 19, 16, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
787
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
794
sja1105_packing(cmd, &entry->port, 5, 3, SJA1105_SIZE_DYN_CMD, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
795
sja1105_packing(cmd, &entry->prio, 2, 0, SJA1105_SIZE_DYN_CMD, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
796
sja1105_packing(p + 3, &entry->credit_lo, 31, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
797
sja1105_packing(p + 2, &entry->credit_hi, 31, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
798
sja1105_packing(p + 1, &entry->send_slope, 31, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
799
sja1105_packing(p + 0, &entry->idle_slope, 31, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
804
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
809
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
810
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
811
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
812
sja1105_packing(p, &cmd->index, 3, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
816
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
821
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
822
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
823
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
824
sja1105_packing(p, &cmd->index, 7, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
828
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
833
sja1105_packing(buf, &entry->port, 159, 157, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
834
sja1105_packing(buf, &entry->prio, 156, 154, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
835
sja1105_packing(buf, &entry->credit_lo, 153, 122, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
836
sja1105_packing(buf, &entry->credit_hi, 121, 90, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
837
sja1105_packing(buf, &entry->send_slope, 89, 58, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
838
sja1105_packing(buf, &entry->idle_slope, 57, 26, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
843
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
849
sja1105_packing(buf, &entry_type, 159, 159, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
850
sja1105_packing(buf, &entry->credit_lo, 151, 120, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
851
sja1105_packing(buf, &entry->credit_hi, 119, 88, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
852
sja1105_packing(buf, &entry->send_slope, 87, 56, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
853
sja1105_packing(buf, &entry->idle_slope, 55, 24, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
858
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
864
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
869
sja1105_packing(p, &cmd->valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
870
sja1105_packing(p, &cmd->rdwrset, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
871
sja1105_packing(p, &cmd->errors, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.c
872
sja1105_packing(p, &cmd->index, 6, 0, size, op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.h
19
size_t (*entry_packing)(void *buf, void *entry_ptr, enum packing_op op);
drivers/net/dsa/sja1105/sja1105_dynamic_config.h
21
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_mdio.c
144
int phy, enum sja1105_mdio_opcode op,
drivers/net/dsa/sja1105/sja1105_mdio.c
149
return regs->mdio_100base_t1 | (phy << 7) | (op << 5) | (xad << 0);
drivers/net/dsa/sja1105/sja1105_ptp.c
139
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_ptp.c
145
sja1105_packing(buf, &valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
146
sja1105_packing(buf, &cmd->ptpstrtsch, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
147
sja1105_packing(buf, &cmd->ptpstopsch, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
148
sja1105_packing(buf, &cmd->startptpcp, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
149
sja1105_packing(buf, &cmd->stopptpcp, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
150
sja1105_packing(buf, &cmd->resptp, 2, 2, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
151
sja1105_packing(buf, &cmd->corrclk4ts, 1, 1, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
152
sja1105_packing(buf, &cmd->ptpclkadd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
156
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_ptp.c
162
sja1105_packing(buf, &valid, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
163
sja1105_packing(buf, &cmd->ptpstrtsch, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
164
sja1105_packing(buf, &cmd->ptpstopsch, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
165
sja1105_packing(buf, &cmd->startptpcp, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
166
sja1105_packing(buf, &cmd->stopptpcp, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
167
sja1105_packing(buf, &cmd->resptp, 3, 3, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
168
sja1105_packing(buf, &cmd->corrclk4ts, 2, 2, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.c
169
sja1105_packing(buf, &cmd->ptpclkadd, 0, 0, size, op);
drivers/net/dsa/sja1105/sja1105_ptp.h
101
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_ptp.h
98
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.c
102
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
107
sja1105_packing(buf, &entry->cas_master, 126, 126, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
108
sja1105_packing(buf, &entry->destmeta, 125, 78, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
109
sja1105_packing(buf, &entry->srcmeta, 77, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
114
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
119
sja1105_packing(buf, &entry->vllupformat, 319, 319, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
120
sja1105_packing(buf, &entry->mirr_ptacu, 318, 318, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
121
sja1105_packing(buf, &entry->switchid, 317, 315, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
122
sja1105_packing(buf, &entry->hostprio, 314, 312, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
123
sja1105_packing(buf, &entry->mac_fltres1, 311, 264, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
124
sja1105_packing(buf, &entry->mac_fltres0, 263, 216, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
125
sja1105_packing(buf, &entry->mac_flt1, 215, 168, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
126
sja1105_packing(buf, &entry->mac_flt0, 167, 120, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
127
sja1105_packing(buf, &entry->incl_srcpt1, 119, 119, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
128
sja1105_packing(buf, &entry->incl_srcpt0, 118, 118, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
129
sja1105_packing(buf, &entry->send_meta1, 117, 117, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
130
sja1105_packing(buf, &entry->send_meta0, 116, 116, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
131
sja1105_packing(buf, &entry->casc_port, 115, 113, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
132
sja1105_packing(buf, &entry->host_port, 112, 110, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
133
sja1105_packing(buf, &entry->mirr_port, 109, 107, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
134
sja1105_packing(buf, &entry->vlmarker, 106, 75, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
135
sja1105_packing(buf, &entry->vlmask, 74, 43, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
136
sja1105_packing(buf, &entry->tpid, 42, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
137
sja1105_packing(buf, &entry->ignore2stf, 26, 26, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
138
sja1105_packing(buf, &entry->tpid2, 25, 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
146
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
151
sja1105_packing(buf, &entry->vllupformat, 351, 351, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
152
sja1105_packing(buf, &entry->mirr_ptacu, 350, 350, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
153
sja1105_packing(buf, &entry->switchid, 349, 347, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
154
sja1105_packing(buf, &entry->hostprio, 346, 344, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
155
sja1105_packing(buf, &entry->mac_fltres1, 343, 296, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
156
sja1105_packing(buf, &entry->mac_fltres0, 295, 248, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
157
sja1105_packing(buf, &entry->mac_flt1, 247, 200, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
158
sja1105_packing(buf, &entry->mac_flt0, 199, 152, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
159
sja1105_packing(buf, &entry->incl_srcpt1, 151, 151, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
160
sja1105_packing(buf, &entry->incl_srcpt0, 150, 150, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
161
sja1105_packing(buf, &entry->send_meta1, 149, 149, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
162
sja1105_packing(buf, &entry->send_meta0, 148, 148, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
163
sja1105_packing(buf, &entry->casc_port, 147, 145, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
164
sja1105_packing(buf, &entry->host_port, 144, 142, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
165
sja1105_packing(buf, &entry->mirr_port, 141, 139, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
166
sja1105_packing(buf, &entry->vlmarker, 138, 107, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
167
sja1105_packing(buf, &entry->vlmask, 106, 75, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
168
sja1105_packing(buf, &entry->tpid2, 74, 59, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
169
sja1105_packing(buf, &entry->ignore2stf, 58, 58, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
170
sja1105_packing(buf, &entry->tpid, 57, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
171
sja1105_packing(buf, &entry->queue_ts, 41, 41, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
172
sja1105_packing(buf, &entry->egrmirrvid, 40, 29, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
173
sja1105_packing(buf, &entry->egrmirrpcp, 28, 26, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
174
sja1105_packing(buf, &entry->egrmirrdei, 25, 25, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
175
sja1105_packing(buf, &entry->replay_port, 24, 22, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
180
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
185
sja1105_packing(buf, &entry->vllupformat, 447, 447, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
186
sja1105_packing(buf, &entry->mirr_ptacu, 446, 446, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
187
sja1105_packing(buf, &entry->switchid, 445, 442, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
188
sja1105_packing(buf, &entry->hostprio, 441, 439, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
189
sja1105_packing(buf, &entry->mac_fltres1, 438, 391, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
190
sja1105_packing(buf, &entry->mac_fltres0, 390, 343, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
191
sja1105_packing(buf, &entry->mac_flt1, 342, 295, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
192
sja1105_packing(buf, &entry->mac_flt0, 294, 247, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
193
sja1105_packing(buf, &entry->incl_srcpt1, 246, 246, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
194
sja1105_packing(buf, &entry->incl_srcpt0, 245, 245, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
195
sja1105_packing(buf, &entry->send_meta1, 244, 244, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
196
sja1105_packing(buf, &entry->send_meta0, 243, 243, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
197
sja1105_packing(buf, &entry->casc_port, 242, 232, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
198
sja1105_packing(buf, &entry->host_port, 231, 228, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
199
sja1105_packing(buf, &entry->mirr_port, 227, 224, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
200
sja1105_packing(buf, &entry->vlmarker, 223, 192, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
201
sja1105_packing(buf, &entry->vlmask, 191, 160, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
202
sja1105_packing(buf, &entry->tpid2, 159, 144, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
203
sja1105_packing(buf, &entry->ignore2stf, 143, 143, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
204
sja1105_packing(buf, &entry->tpid, 142, 127, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
205
sja1105_packing(buf, &entry->queue_ts, 126, 126, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
206
sja1105_packing(buf, &entry->egrmirrvid, 125, 114, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
207
sja1105_packing(buf, &entry->egrmirrpcp, 113, 111, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
208
sja1105_packing(buf, &entry->egrmirrdei, 110, 110, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
209
sja1105_packing(buf, &entry->replay_port, 109, 106, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
210
sja1105_packing(buf, &entry->tdmaconfigidx, 70, 67, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
211
sja1105_packing(buf, &entry->header_type, 64, 49, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
212
sja1105_packing(buf, &entry->tte_en, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
218
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
224
sja1105_packing(buf, &entry->max_dynp, 95, 93, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
227
offset + 9, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
232
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
238
sja1105_packing(buf, &entry->max_dynp, 95, 93, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
241
offset + 10, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
246
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
252
sja1105_packing(buf, &entry->bc_domain, 63, 59, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
253
sja1105_packing(buf, &entry->reach_port, 58, 54, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
254
sja1105_packing(buf, &entry->fl_domain, 53, 49, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
257
offset + 2, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
262
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
272
offset + 2, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
275
sja1105_packing(buf, &entry->bc_domain, 63, 53, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
276
sja1105_packing(buf, &entry->reach_port, 52, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
277
sja1105_packing(buf, &entry->fl_domain, 41, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
284
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
289
sja1105_packing(buf, &entry->maxage, 31, 17, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
290
sja1105_packing(buf, &entry->dyn_tbsz, 16, 14, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
291
sja1105_packing(buf, &entry->poly, 13, 6, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
292
sja1105_packing(buf, &entry->shared_learn, 5, 5, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
293
sja1105_packing(buf, &entry->no_enf_hostprt, 4, 4, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
294
sja1105_packing(buf, &entry->no_mgmt_learn, 3, 3, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
299
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
307
offset + 10, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
308
sja1105_packing(buf, &entry->maxage, 57, 43, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
309
sja1105_packing(buf, &entry->start_dynspc, 42, 33, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
310
sja1105_packing(buf, &entry->drpnolearn, 32, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
311
sja1105_packing(buf, &entry->shared_learn, 27, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
312
sja1105_packing(buf, &entry->no_enf_hostprt, 26, 26, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
313
sja1105_packing(buf, &entry->no_mgmt_learn, 25, 25, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
314
sja1105_packing(buf, &entry->use_static, 24, 24, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
315
sja1105_packing(buf, &entry->owr_dyn, 23, 23, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
316
sja1105_packing(buf, &entry->learn_once, 22, 22, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
321
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
329
offset + 10, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
330
sja1105_packing(buf, &entry->maxage, 69, 55, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
331
sja1105_packing(buf, &entry->start_dynspc, 54, 45, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
332
sja1105_packing(buf, &entry->drpnolearn, 44, 34, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
333
sja1105_packing(buf, &entry->shared_learn, 33, 33, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
334
sja1105_packing(buf, &entry->no_enf_hostprt, 32, 32, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
335
sja1105_packing(buf, &entry->no_mgmt_learn, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
336
sja1105_packing(buf, &entry->use_static, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
337
sja1105_packing(buf, &entry->owr_dyn, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
338
sja1105_packing(buf, &entry->learn_once, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
343
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
348
sja1105_packing(buf, &entry->vlanid, 95, 84, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
349
sja1105_packing(buf, &entry->macaddr, 83, 36, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
350
sja1105_packing(buf, &entry->destports, 35, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
351
sja1105_packing(buf, &entry->enfport, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
352
sja1105_packing(buf, &entry->index, 29, 20, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
357
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
363
sja1105_packing(buf, &entry->tsreg, 159, 159, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
364
sja1105_packing(buf, &entry->mirrvlan, 158, 147, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
365
sja1105_packing(buf, &entry->takets, 146, 146, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
366
sja1105_packing(buf, &entry->mirr, 145, 145, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
367
sja1105_packing(buf, &entry->retag, 144, 144, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
369
sja1105_packing(buf, &entry->touched, 159, 159, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
370
sja1105_packing(buf, &entry->age, 158, 144, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
372
sja1105_packing(buf, &entry->mask_iotag, 143, 143, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
373
sja1105_packing(buf, &entry->mask_vlanid, 142, 131, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
374
sja1105_packing(buf, &entry->mask_macaddr, 130, 83, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
375
sja1105_packing(buf, &entry->iotag, 82, 82, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
376
sja1105_packing(buf, &entry->vlanid, 81, 70, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
377
sja1105_packing(buf, &entry->macaddr, 69, 22, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
378
sja1105_packing(buf, &entry->destports, 21, 17, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
379
sja1105_packing(buf, &entry->enfport, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
380
sja1105_packing(buf, &entry->index, 15, 6, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
385
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
391
sja1105_packing(buf, &entry->trap, 168, 168, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
392
sja1105_packing(buf, &entry->mirrvlan, 167, 156, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
393
sja1105_packing(buf, &entry->takets, 155, 155, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
394
sja1105_packing(buf, &entry->mirr, 154, 154, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
395
sja1105_packing(buf, &entry->retag, 153, 153, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
397
sja1105_packing(buf, &entry->touched, 168, 168, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
398
sja1105_packing(buf, &entry->age, 167, 153, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
400
sja1105_packing(buf, &entry->mask_iotag, 152, 152, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
401
sja1105_packing(buf, &entry->mask_vlanid, 151, 140, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
402
sja1105_packing(buf, &entry->mask_macaddr, 139, 92, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
403
sja1105_packing(buf, &entry->mask_srcport, 91, 88, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
404
sja1105_packing(buf, &entry->iotag, 87, 87, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
405
sja1105_packing(buf, &entry->vlanid, 86, 75, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
406
sja1105_packing(buf, &entry->macaddr, 74, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
407
sja1105_packing(buf, &entry->srcport, 26, 23, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
408
sja1105_packing(buf, &entry->destports, 22, 12, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
409
sja1105_packing(buf, &entry->enfport, 11, 11, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
410
sja1105_packing(buf, &entry->index, 10, 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
415
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
420
sja1105_packing(buf, &entry->sharindx, 63, 58, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
421
sja1105_packing(buf, &entry->smax, 57, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
422
sja1105_packing(buf, &entry->rate, 41, 26, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
423
sja1105_packing(buf, &entry->maxlen, 25, 15, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
424
sja1105_packing(buf, &entry->partition, 14, 12, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
429
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
434
sja1105_packing(buf, &entry->sharindx, 63, 57, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
435
sja1105_packing(buf, &entry->smax, 56, 39, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
436
sja1105_packing(buf, &entry->rate, 38, 21, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
437
sja1105_packing(buf, &entry->maxlen, 20, 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
438
sja1105_packing(buf, &entry->partition, 9, 7, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
443
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
451
offset + 0, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
453
offset + 9, offset + 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
455
offset + 18, offset + 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
457
sja1105_packing(buf, &entry->ifg, 71, 67, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
458
sja1105_packing(buf, &entry->speed, 66, 65, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
459
sja1105_packing(buf, &entry->tp_delin, 64, 49, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
460
sja1105_packing(buf, &entry->tp_delout, 48, 33, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
461
sja1105_packing(buf, &entry->maxage, 32, 25, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
462
sja1105_packing(buf, &entry->vlanprio, 24, 22, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
463
sja1105_packing(buf, &entry->vlanid, 21, 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
464
sja1105_packing(buf, &entry->ing_mirr, 9, 9, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
465
sja1105_packing(buf, &entry->egr_mirr, 8, 8, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
466
sja1105_packing(buf, &entry->drpnona664, 7, 7, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
467
sja1105_packing(buf, &entry->drpdtag, 6, 6, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
468
sja1105_packing(buf, &entry->drpuntag, 5, 5, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
469
sja1105_packing(buf, &entry->retag, 4, 4, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
470
sja1105_packing(buf, &entry->dyn_learn, 3, 3, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
471
sja1105_packing(buf, &entry->egress, 2, 2, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
472
sja1105_packing(buf, &entry->ingress, 1, 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
477
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
485
offset + 0, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
487
offset + 9, offset + 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
489
offset + 18, offset + 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
491
sja1105_packing(buf, &entry->ifg, 103, 99, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
492
sja1105_packing(buf, &entry->speed, 98, 97, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
493
sja1105_packing(buf, &entry->tp_delin, 96, 81, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
494
sja1105_packing(buf, &entry->tp_delout, 80, 65, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
495
sja1105_packing(buf, &entry->maxage, 64, 57, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
496
sja1105_packing(buf, &entry->vlanprio, 56, 54, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
497
sja1105_packing(buf, &entry->vlanid, 53, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
498
sja1105_packing(buf, &entry->ing_mirr, 41, 41, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
499
sja1105_packing(buf, &entry->egr_mirr, 40, 40, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
500
sja1105_packing(buf, &entry->drpnona664, 39, 39, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
501
sja1105_packing(buf, &entry->drpdtag, 38, 38, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
502
sja1105_packing(buf, &entry->drpuntag, 35, 35, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
503
sja1105_packing(buf, &entry->retag, 34, 34, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
504
sja1105_packing(buf, &entry->dyn_learn, 33, 33, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
505
sja1105_packing(buf, &entry->egress, 32, 32, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
506
sja1105_packing(buf, &entry->ingress, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
511
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
519
offset + 0, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
521
offset + 9, offset + 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
523
offset + 18, offset + 10, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
525
sja1105_packing(buf, &entry->speed, 98, 96, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
526
sja1105_packing(buf, &entry->tp_delin, 95, 80, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
527
sja1105_packing(buf, &entry->tp_delout, 79, 64, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
528
sja1105_packing(buf, &entry->maxage, 63, 56, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
529
sja1105_packing(buf, &entry->vlanprio, 55, 53, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
53
size_t len, enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
530
sja1105_packing(buf, &entry->vlanid, 52, 41, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
531
sja1105_packing(buf, &entry->ing_mirr, 40, 40, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
532
sja1105_packing(buf, &entry->egr_mirr, 39, 39, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
533
sja1105_packing(buf, &entry->drpnona664, 38, 38, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
534
sja1105_packing(buf, &entry->drpdtag, 37, 37, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
535
sja1105_packing(buf, &entry->drpuntag, 34, 34, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
536
sja1105_packing(buf, &entry->retag, 33, 33, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
537
sja1105_packing(buf, &entry->dyn_learn, 32, 32, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
538
sja1105_packing(buf, &entry->egress, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
539
sja1105_packing(buf, &entry->ingress, 30, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
540
sja1105_packing(buf, &entry->ifg, 10, 5, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
546
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
55
int rc = packing(buf, val, start, end, len, op, QUIRK_LSW32_IS_FIRST);
drivers/net/dsa/sja1105/sja1105_static_config.c
551
sja1105_packing(buf, &entry->clksrc, 31, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
552
sja1105_packing(buf, &entry->actsubsch, 29, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
558
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
563
sja1105_packing(buf, &entry->subschindx, 31, 29, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
564
sja1105_packing(buf, &entry->delta, 28, 11, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
565
sja1105_packing(buf, &entry->address, 10, 1, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
571
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
576
sja1105_packing(buf, &entry->subschindx, 63, 61, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
577
sja1105_packing(buf, &entry->delta, 60, 43, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
578
sja1105_packing(buf, &entry->address, 42, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
583
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
591
offset + 9, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
596
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
604
offset + 11, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
609
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
614
sja1105_packing(buf, &entry->winstindex, 63, 54, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
615
sja1105_packing(buf, &entry->winend, 53, 53, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
616
sja1105_packing(buf, &entry->winst, 52, 52, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
617
sja1105_packing(buf, &entry->destports, 51, 47, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
618
sja1105_packing(buf, &entry->setvalid, 46, 46, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
619
sja1105_packing(buf, &entry->txen, 45, 45, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
620
sja1105_packing(buf, &entry->resmedia_en, 44, 44, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
621
sja1105_packing(buf, &entry->resmedia, 43, 36, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
622
sja1105_packing(buf, &entry->vlindex, 35, 26, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
623
sja1105_packing(buf, &entry->delta, 25, 8, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
628
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
633
sja1105_packing(buf, &entry->winstindex, 95, 84, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
634
sja1105_packing(buf, &entry->winend, 83, 83, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
635
sja1105_packing(buf, &entry->winst, 82, 82, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
636
sja1105_packing(buf, &entry->destports, 81, 71, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
637
sja1105_packing(buf, &entry->setvalid, 70, 70, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
638
sja1105_packing(buf, &entry->txen, 69, 69, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
639
sja1105_packing(buf, &entry->resmedia_en, 68, 68, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
640
sja1105_packing(buf, &entry->resmedia, 67, 60, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
641
sja1105_packing(buf, &entry->vlindex, 59, 48, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
642
sja1105_packing(buf, &entry->delta, 47, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
648
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
656
offset + 9, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
657
sja1105_packing(buf, &entry->debugen, 15, 15, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
663
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
671
offset + 10, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
672
sja1105_packing(buf, &entry->debugen, 7, 7, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
677
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
682
sja1105_packing(buf, &entry->type, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
683
sja1105_packing(buf, &entry->priority, 30, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
684
sja1105_packing(buf, &entry->partition, 27, 25, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
685
sja1105_packing(buf, &entry->destports, 24, 20, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
690
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
695
sja1105_packing(buf, &entry->type, 31, 31, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
696
sja1105_packing(buf, &entry->priority, 30, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
697
sja1105_packing(buf, &entry->partition, 27, 25, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
698
sja1105_packing(buf, &entry->destports, 24, 14, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
703
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
711
95, 91, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
713
90, 90, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
715
89, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
717
41, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
719
29, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
721
26, 24, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
725
95, 91, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
727
90, 90, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
729
57, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
731
29, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
737
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
745
94, 84, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
747
83, 83, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
749
82, 35, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
751
34, 23, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
753
22, 19, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
755
18, 16, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
759
94, 84, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
761
83, 83, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
763
50, 35, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
765
22, 19, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
771
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
776
sja1105_packing(buf, &entry->type, 63, 63, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
777
sja1105_packing(buf, &entry->maxlen, 62, 52, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
778
sja1105_packing(buf, &entry->sharindx, 51, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
780
sja1105_packing(buf, &entry->bag, 41, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
781
sja1105_packing(buf, &entry->jitter, 27, 18, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
787
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
792
sja1105_packing(buf, &entry->type, 63, 63, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
793
sja1105_packing(buf, &entry->maxlen, 62, 52, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
794
sja1105_packing(buf, &entry->sharindx, 51, 40, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
796
sja1105_packing(buf, &entry->bag, 41, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
797
sja1105_packing(buf, &entry->jitter, 27, 18, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
803
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
808
sja1105_packing(buf, &entry->ving_mirr, 63, 59, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
809
sja1105_packing(buf, &entry->vegr_mirr, 58, 54, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
810
sja1105_packing(buf, &entry->vmemb_port, 53, 49, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
811
sja1105_packing(buf, &entry->vlan_bc, 48, 44, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
812
sja1105_packing(buf, &entry->tag_port, 43, 39, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
813
sja1105_packing(buf, &entry->vlanid, 38, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
818
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
823
sja1105_packing(buf, &entry->ving_mirr, 95, 85, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
824
sja1105_packing(buf, &entry->vegr_mirr, 84, 74, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
825
sja1105_packing(buf, &entry->vmemb_port, 73, 63, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
826
sja1105_packing(buf, &entry->vlan_bc, 62, 52, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
827
sja1105_packing(buf, &entry->tag_port, 51, 41, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
828
sja1105_packing(buf, &entry->type_entry, 40, 39, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
829
sja1105_packing(buf, &entry->vlanid, 38, 27, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
834
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
842
offset + 1, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
844
offset + 2, offset + 2, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
850
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
858
offset + 1, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
860
offset + 2, offset + 2, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
862
offset + 3, offset + 3, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
868
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
873
sja1105_packing(buf, &entry->egr_port, 63, 59, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
874
sja1105_packing(buf, &entry->ing_port, 58, 54, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
875
sja1105_packing(buf, &entry->vlan_ing, 53, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
876
sja1105_packing(buf, &entry->vlan_egr, 41, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
877
sja1105_packing(buf, &entry->do_not_learn, 29, 29, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
878
sja1105_packing(buf, &entry->use_dest_ports, 28, 28, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
879
sja1105_packing(buf, &entry->destports, 27, 23, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
884
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
889
sja1105_packing(buf, &entry->egr_port, 63, 53, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
890
sja1105_packing(buf, &entry->ing_port, 52, 42, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
891
sja1105_packing(buf, &entry->vlan_ing, 41, 30, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
892
sja1105_packing(buf, &entry->vlan_egr, 29, 18, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
893
sja1105_packing(buf, &entry->do_not_learn, 17, 17, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
894
sja1105_packing(buf, &entry->use_dest_ports, 16, 16, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
895
sja1105_packing(buf, &entry->destports, 15, 5, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
900
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
908
offset + 2, offset + 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
91
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
914
enum packing_op op)
drivers/net/dsa/sja1105/sja1105_static_config.c
919
sja1105_packing(buf, &entry->block_id, 31, 24, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
920
sja1105_packing(buf, &entry->len, 55, 32, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
921
sja1105_packing(buf, &entry->crc, 95, 64, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
96
sja1105_packing(buf, &entry->destmeta, 95, 48, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.c
97
sja1105_packing(buf, &entry->srcmeta, 47, 0, size, op);
drivers/net/dsa/sja1105/sja1105_static_config.h
434
size_t (*packing)(void *buf, void *entry_ptr, enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
459
size_t sja1105_table_header_packing(void *buf, void *hdr, enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
500
size_t len, enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
504
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
506
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
508
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
510
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
512
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
514
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
516
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
518
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
520
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
522
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
524
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
526
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
528
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
530
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
532
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
534
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
536
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
538
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
540
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
542
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
544
enum packing_op op);
drivers/net/dsa/sja1105/sja1105_static_config.h
546
enum packing_op op);
drivers/net/ethernet/amd/sunlance.c
1258
of_iounmap(&lp->op->resource[0], lp->lregs, LANCE_REG_SIZE);
drivers/net/ethernet/amd/sunlance.c
1269
dma_free_coherent(&lp->op->dev,
drivers/net/ethernet/amd/sunlance.c
1297
static int sparc_lance_probe_one(struct platform_device *op,
drivers/net/ethernet/amd/sunlance.c
1301
struct device_node *dp = op->dev.of_node;
drivers/net/ethernet/amd/sunlance.c
1320
lp->lregs = of_ioremap(&op->resource[0], 0,
drivers/net/ethernet/amd/sunlance.c
1339
lp->op = op;
drivers/net/ethernet/amd/sunlance.c
1361
dma_alloc_coherent(&op->dev,
drivers/net/ethernet/amd/sunlance.c
1442
SET_NETDEV_DEV(dev, &op->dev);
drivers/net/ethernet/amd/sunlance.c
1447
dev->irq = op->archdata.irqs[0];
drivers/net/ethernet/amd/sunlance.c
1461
platform_set_drvdata(op, lp);
drivers/net/ethernet/amd/sunlance.c
1474
static int sunlance_sbus_probe(struct platform_device *op)
drivers/net/ethernet/amd/sunlance.c
1476
struct platform_device *parent = to_platform_device(op->dev.parent);
drivers/net/ethernet/amd/sunlance.c
1481
err = sparc_lance_probe_one(op, parent, NULL);
drivers/net/ethernet/amd/sunlance.c
1483
err = sparc_lance_probe_one(op, NULL, parent);
drivers/net/ethernet/amd/sunlance.c
1485
err = sparc_lance_probe_one(op, NULL, NULL);
drivers/net/ethernet/amd/sunlance.c
1490
static void sunlance_sbus_remove(struct platform_device *op)
drivers/net/ethernet/amd/sunlance.c
1492
struct lance_private *lp = platform_get_drvdata(op);
drivers/net/ethernet/amd/sunlance.c
262
struct platform_device *op;
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
118
if (state->op->cmd == XGBE_I2C_CMD_READ)
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
143
if (state->op->cmd != XGBE_I2C_CMD_READ)
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
186
state->tx_abort_source, state->op->target);
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
268
static int xgbe_i2c_xfer(struct xgbe_prv_data *pdata, struct xgbe_i2c_op *op)
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
283
xgbe_i2c_set_target(pdata, op->target);
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
286
state->op = op;
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
287
state->tx_len = op->len;
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
288
state->tx_buf = op->buf;
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
289
state->rx_len = op->len;
drivers/net/ethernet/amd/xgbe/xgbe-i2c.c
290
state->rx_buf = op->buf;
drivers/net/ethernet/amd/xgbe/xgbe.h
595
struct xgbe_i2c_op *op;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
203
u16 index, u8 op, u8 update);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
637
u8 segment, u16 index, u8 op,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
646
(op << IGU_REGULAR_ENABLE_INT_SHIFT));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
657
u8 storm, u16 index, u8 op, u8 update)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
668
(op << IGU_ACK_REGISTER_INTERRUPT_MODE_SHIFT));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
677
u16 index, u8 op, u8 update)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
680
bnx2x_hc_ack_sb(bp, igu_sb_id, storm, index, op, update);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
692
bnx2x_igu_ack_sb(bp, igu_sb_id, segment, index, op, update);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
49
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
55
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
61
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
67
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
79
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
85
u32 op:8;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
237
const union init_op *op;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
249
op = (const union init_op *)&(INIT_OPS(bp)[op_idx]);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
251
op_type = op->raw.op;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
252
addr = op->raw.offset;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
257
len = op->arr_wr.data_len;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
258
data = data_base + op->arr_wr.data_off;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
265
REG_WR(bp, addr, op->write.val);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
274
bnx2x_init_fill(bp, addr, 0, op->zero.len, 0);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
277
bnx2x_init_fill(bp, addr, 0, op->zero.len, 1);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
281
op->arr_wr.data_off);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
291
op->if_mode.mode_bit_map) !=
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
292
op->if_mode.mode_bit_map)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
293
op_idx += op->if_mode.cmd_offset;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
300
op->if_mode.mode_bit_map) == 0)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
301
op_idx += op->if_mode.cmd_offset;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
13338
target[i].op = (tmp >> 24) & 0xff;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
5244
u16 index, u8 op, u8 update)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
5248
bnx2x_igu_ack_sb_gen(bp, igu_sb_id, segment, index, op, update,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
79
u8 igu_sb_id, u8 segment, u16 index, u8 op,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
94
(op << IGU_REGULAR_ENABLE_INT_SHIFT));
drivers/net/ethernet/broadcom/cnic.c
3056
u16 index, u8 op, u8 update)
drivers/net/ethernet/broadcom/cnic.c
3068
(op << IGU_ACK_REGISTER_INTERRUPT_MODE_SHIFT));
drivers/net/ethernet/broadcom/cnic.c
3074
u16 index, u8 op, u8 update)
drivers/net/ethernet/broadcom/cnic.c
3083
(op << IGU_REGULAR_ENABLE_INT_SHIFT);
drivers/net/ethernet/broadcom/cnic_if.h
91
#define KCQE_OPCODE(op) \
drivers/net/ethernet/broadcom/cnic_if.h
92
(((op) & KCQE_FLAGS_OPCODE_MASK) >> KCQE_FLAGS_OPCODE_SHIFT)
drivers/net/ethernet/brocade/bna/bfa_ioc.c
110
bfa_fsm_state_decl(bfa_ioc, op, struct bfa_ioc, enum ioc_event);
drivers/net/ethernet/brocade/bna/bfa_ioc.c
1498
u8 rd_cnt, u8 ad_cnt, u8 op)
drivers/net/ethernet/brocade/bna/bfa_ioc.c
1507
cmd.r.cmd = op;
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
762
octnet_mdio45_access(struct lio *lio, int op, int loc, int *value)
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
781
mdio_cmd->op = op;
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
783
if (op)
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
821
if (!op)
drivers/net/ethernet/cavium/liquidio/liquidio_common.h
63
#define OPCODE_SUBCODE(op, sub) ((((op) & 0x0f) << 8) | ((sub) & 0x7f))
drivers/net/ethernet/cavium/liquidio/liquidio_common.h
930
u64 op;
drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c
112
if (op)
drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c
140
if (op) { /* read from core */
drivers/net/ethernet/cavium/liquidio/octeon_mem_ops.c
95
u8 *hostbuf, u32 len, u32 op)
drivers/net/ethernet/chelsio/cxgb3/common.h
753
int t3_sge_cqcntxt_op(struct adapter *adapter, unsigned int id, unsigned int op,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_ctl_defs.h
150
unsigned int op;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
304
ret = t3_sge_cqcntxt_op(adapter, rdma->id, rdma->op,
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2428
int t3_sge_cqcntxt_op(struct adapter *adapter, unsigned int id, unsigned int op,
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2437
t3_write_reg(adapter, A_SG_CONTEXT_CMD, V_CONTEXT_CMD_OPCODE(op) |
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2443
if (op >= 2 && op < 7) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1851
int t4_fwcache(struct adapter *adap, enum fw_params_param_dev_fwcache op);
drivers/net/ethernet/chelsio/cxgb4/sched.c
43
enum sched_fw_ops op)
drivers/net/ethernet/chelsio/cxgb4/sched.c
51
switch (op) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
1492
u32 wr_mid, ctrl0, op, sgl_off = 0;
drivers/net/ethernet/chelsio/cxgb4/sge.c
1659
op = FW_PTP_TX_PKT_WR;
drivers/net/ethernet/chelsio/cxgb4/sge.c
1661
op = FW_ETH_TX_PKT_WR;
drivers/net/ethernet/chelsio/cxgb4/sge.c
1662
wr->op_immdlen = htonl(FW_WR_OP_V(op) |
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3847
int t4_fwcache(struct adapter *adap, enum fw_params_param_dev_fwcache op)
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
3861
c.param[0].val = cpu_to_be32(op);
drivers/net/ethernet/emulex/benet/be_cmds.c
4683
int be_cmd_manage_iface(struct be_adapter *adapter, u32 iface, u8 op)
drivers/net/ethernet/emulex/benet/be_cmds.c
4704
req->op = op;
drivers/net/ethernet/emulex/benet/be_cmds.h
2369
u8 op;
drivers/net/ethernet/emulex/benet/be_cmds.h
2504
int be_cmd_manage_iface(struct be_adapter *adapter, u32 iface, u8 op);
drivers/net/ethernet/freescale/fec_mpc52xx.c
1004
static int mpc52xx_fec_of_suspend(struct platform_device *op, pm_message_t state)
drivers/net/ethernet/freescale/fec_mpc52xx.c
1006
struct net_device *dev = platform_get_drvdata(op);
drivers/net/ethernet/freescale/fec_mpc52xx.c
1014
static int mpc52xx_fec_of_resume(struct platform_device *op)
drivers/net/ethernet/freescale/fec_mpc52xx.c
1016
struct net_device *dev = platform_get_drvdata(op);
drivers/net/ethernet/freescale/fec_mpc52xx.c
809
static int mpc52xx_fec_probe(struct platform_device *op)
drivers/net/ethernet/freescale/fec_mpc52xx.c
817
struct device_node *np = op->dev.of_node;
drivers/net/ethernet/freescale/fec_mpc52xx.c
855
SET_NETDEV_DEV(ndev, &op->dev);
drivers/net/ethernet/freescale/fec_mpc52xx.c
927
priv->mdio_speed = ((mpc5xxx_get_bus_frequency(&op->dev) >> 20) / 5) << 1;
drivers/net/ethernet/freescale/fec_mpc52xx.c
954
platform_set_drvdata(op, ndev);
drivers/net/ethernet/freescale/fec_mpc52xx.c
956
op->dev.of_node, ndev->dev_addr);
drivers/net/ethernet/freescale/fec_mpc52xx.c
978
mpc52xx_fec_remove(struct platform_device *op)
drivers/net/ethernet/freescale/fec_mpc52xx.c
983
ndev = platform_get_drvdata(op);
drivers/net/ethernet/freescale/fs_enet/mac-fcc.c
73
static inline int fcc_cr_cmd(struct fs_enet_private *fep, u32 op)
drivers/net/ethernet/freescale/fs_enet/mac-fcc.c
77
return cpm_command(fpi->cp_command, op);
drivers/net/ethernet/freescale/fs_enet/mac-scc.c
79
static inline int scc_cr_cmd(struct fs_enet_private *fep, u32 op)
drivers/net/ethernet/freescale/fs_enet/mac-scc.c
83
return cpm_command(fpi->cp_command, op);
drivers/net/ethernet/fungible/funcore/fun_dev.c
191
rsp_common->op, rsp_common->ret);
drivers/net/ethernet/fungible/funcore/fun_dev.c
197
cid, rsp_common->op);
drivers/net/ethernet/fungible/funcore/fun_dev.c
450
rsp_common->op, rsp_len, ctx->rsp_len);
drivers/net/ethernet/fungible/funcore/fun_hci.h
1016
__u8 op;
drivers/net/ethernet/fungible/funcore/fun_hci.h
1025
__u8 op;
drivers/net/ethernet/fungible/funcore/fun_hci.h
1114
__u8 op;
drivers/net/ethernet/fungible/funcore/fun_hci.h
1162
__u8 op;
drivers/net/ethernet/fungible/funcore/fun_hci.h
31
__u8 op;
drivers/net/ethernet/fungible/funcore/fun_hci.h
41
.op = (_op), .len8 = (_len8), .flags = cpu_to_be16(_flags), \
drivers/net/ethernet/fungible/funcore/fun_hci.h
47
.op = (_op), .len8 = (_len) / 8, \
drivers/net/ethernet/fungible/funcore/fun_hci.h
51
__u8 op;
drivers/net/ethernet/fungible/funeth/funeth.h
169
const u32 *qtable, u8 op);
drivers/net/ethernet/fungible/funeth/funeth_main.c
177
const u32 *qtable, u8 op)
drivers/net/ethernet/fungible/funeth/funeth_main.c
1903
u8 op = ((struct fun_admin_rsp_common *)entry)->op;
drivers/net/ethernet/fungible/funeth/funeth_main.c
1905
if (op == FUN_ADMIN_OP_PORT) {
drivers/net/ethernet/fungible/funeth/funeth_main.c
1920
op, rsp->subop);
drivers/net/ethernet/fungible/funeth/funeth_main.c
1923
dev_info(fdev->dev, "adminq event unexpected op %u", op);
drivers/net/ethernet/fungible/funeth/funeth_main.c
194
if (op != FUN_ADMIN_SUBOP_CREATE && fp->rss_hw_id == FUN_HCI_ID_INVALID)
drivers/net/ethernet/fungible/funeth/funeth_main.c
197
flags = op == FUN_ADMIN_SUBOP_CREATE ?
drivers/net/ethernet/fungible/funeth/funeth_main.c
202
FUN_ADMIN_RSS_CREATE_REQ_INIT(op, flags, fp->rss_hw_id,
drivers/net/ethernet/fungible/funeth/funeth_main.c
218
if (!rc && op == FUN_ADMIN_SUBOP_CREATE)
drivers/net/ethernet/fungible/funeth/funeth_tx.c
172
req->op = FUN_ETH_OP_TX;
drivers/net/ethernet/fungible/funeth/funeth_tx.c
547
req->op = FUN_ETH_OP_TX;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
4996
enum hclge_opcode_type op;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
5000
op = en ? HCLGE_OPC_ADD_RING_TO_VECTOR : HCLGE_OPC_DEL_RING_TO_VECTOR;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
5001
hclge_cmd_setup_basic_desc(&desc, op, false);
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
5037
op,
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
8161
enum hclge_mac_vlan_tbl_opcode op)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
8172
if (op == HCLGE_MAC_VLAN_ADD) {
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
8183
} else if (op == HCLGE_MAC_VLAN_REMOVE) {
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
8196
} else if (op == HCLGE_MAC_VLAN_LKUP) {
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
8212
"unknown opcode for get_mac_vlan_cmd_status, opcode=%d.\n", op);
drivers/net/ethernet/hisilicon/hns_mdio.c
193
u8 is_c45, u8 op, u8 phy_id, u16 cmd)
drivers/net/ethernet/hisilicon/hns_mdio.c
199
cmd_reg_value |= op << MDIO_CMD_OP_S;
drivers/net/ethernet/hisilicon/hns_mdio.c
224
u8 op;
drivers/net/ethernet/hisilicon/hns_mdio.c
239
op = MDIO_C22_WRITE;
drivers/net/ethernet/hisilicon/hns_mdio.c
244
hns_mdio_cmd_write(mdio_dev, false, op, phy_id, cmd_reg_cfg);
drivers/net/ethernet/hisilicon/hns_mdio.c
266
u8 op;
drivers/net/ethernet/hisilicon/hns_mdio.c
295
op = MDIO_C45_WRITE_DATA;
drivers/net/ethernet/hisilicon/hns_mdio.c
300
hns_mdio_cmd_write(mdio_dev, true, op, phy_id, cmd_reg_cfg);
drivers/net/ethernet/huawei/hinic/hinic_port.c
35
u16 vlan_id, enum mac_op op)
drivers/net/ethernet/huawei/hinic/hinic_port.c
45
if (op == MAC_SET)
drivers/net/ethernet/huawei/hinic/hinic_port.c
68
(op == MAC_SET) ? "set" : "del");
drivers/net/ethernet/huawei/hinic3/hinic3_netdev_ops.c
500
#define SET_FEATURES_OP_STR(op) ((op) ? "Enable" : "Disable")
drivers/net/ethernet/intel/i40e/i40e_ddp.c
158
pinfo->op = I40E_DDP_ADD_TRACKID;
drivers/net/ethernet/intel/i40e/i40e_ddp.c
198
pinfo->op = I40E_DDP_REMOVE_TRACKID;
drivers/net/ethernet/intel/i40e/i40e_ddp.c
327
pinfo.op = I40E_DDP_ADD_TRACKID;
drivers/net/ethernet/intel/i40e/i40e_ddp.c
329
pinfo.op = I40E_DDP_REMOVE_TRACKID;
drivers/net/ethernet/intel/i40e/i40e_type.h
1385
u8 op;
drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
20
enum virtchnl_ops op, u8 *msg, u16 len)
drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
28
status = iavf_aq_send_msg_to_pf(hw, op, 0, msg, len, NULL);
drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
31
op, iavf_stat_str(hw, status),
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
167
static bool idpf_ptp_is_mb_msg(u32 op)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
169
switch (op) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
190
static void idpf_prepare_ptp_mb_msg(struct idpf_adapter *adapter, u32 op,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
196
if (!idpf_ptp_is_mb_msg(op) || !adapter->ptp->secondary_mbx.valid)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
204
static void idpf_prepare_ptp_mb_msg(struct idpf_adapter *adapter, u32 op,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
223
u32 op, u16 msg_size, u8 *msg, u16 cookie)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
254
idpf_prepare_ptp_mb_msg(adapter, op, ctlq_msg);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
257
ctlq_msg->cookie.mbx.chnl_opcode = op;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.h
129
u32 op, u16 msg_size, u8 *msg, u16 cookie);
drivers/net/ethernet/marvell/mvpp2/mvpp2_prs.c
320
unsigned int op)
drivers/net/ethernet/marvell/mvpp2/mvpp2_prs.c
337
mvpp2_prs_sram_bits_set(pe, MVPP2_PRS_SRAM_OP_SEL_SHIFT_OFFS, op);
drivers/net/ethernet/marvell/mvpp2/mvpp2_prs.c
348
unsigned int op)
drivers/net/ethernet/marvell/mvpp2/mvpp2_prs.c
373
op & MVPP2_PRS_SRAM_OP_SEL_UDF_MASK);
drivers/net/ethernet/marvell/mvpp2/mvpp2_tai.c
116
static void mvpp2_tai_op(u32 op, void __iomem *base)
drivers/net/ethernet/marvell/mvpp2/mvpp2_tai.c
123
op | TCFCR0_TCF_TRIGGER);
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1021
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1057
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1379
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1755
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1844
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
810
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
843
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
985
u8 op;
drivers/net/ethernet/marvell/octeontx2/af/npc.h
403
u64 op :4;
drivers/net/ethernet/marvell/octeontx2/af/npc.h
405
u64 op :4;
drivers/net/ethernet/marvell/octeontx2/af/npc.h
461
u64 op :4;
drivers/net/ethernet/marvell/octeontx2/af/npc.h
463
u64 op :4;
drivers/net/ethernet/marvell/octeontx2/af/ptp.c
634
switch (req->op) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
1250
aq_req.op = NPA_AQ_INSTOP_READ;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
1711
aq_req.op = NIX_AQ_INSTOP_READ;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
1976
aq_req.op = NIX_AQ_INSTOP_READ;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
2368
aq_req.op = NIX_AQ_INSTOP_READ;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3442
switch (rule->tx_action.op) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3463
switch (rule->rx_action.op) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1124
((req->op == NIX_AQ_INSTOP_INIT && req->sq.ena) ||
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1125
(req->op == NIX_AQ_INSTOP_WRITE &&
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1136
inst.op = req->op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1154
switch (req->op) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1208
if (req->op == NIX_AQ_INSTOP_INIT) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1217
if (req->op == NIX_AQ_INSTOP_WRITE) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1249
if (req->op == NIX_AQ_INSTOP_READ) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1354
if (!err && req->op == NIX_AQ_INSTOP_WRITE) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1418
aq_req.op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1441
if (req->op != NIX_AQ_INSTOP_INIT)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1451
lock_ctx_req.op = NIX_AQ_INSTOP_LOCK;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3177
int mce, u8 op, u16 pcifunc, int next,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3185
aq_req.op = op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3189
aq_req.mce.op = mce_op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5705
aq_req.op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6013
aq_req->op = NIX_AQ_INSTOP_READ;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6030
aq_req->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6166
aq_req.op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6373
ureq.op = NIX_MCAST_OP_DEL_ENTRY;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6545
if (req->op == NIX_MCAST_OP_DEL_ENTRY) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6566
if (req->op == NIX_MCAST_OP_ADD_ENTRY) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
115
switch (req->op) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
166
if (req->op == NPA_AQ_INSTOP_INIT && req->aura.ena)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
168
if (req->op == NPA_AQ_INSTOP_WRITE) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
181
if (req->op == NPA_AQ_INSTOP_INIT && req->pool.ena)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
183
if (req->op == NPA_AQ_INSTOP_WRITE) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
197
if (req->op == NPA_AQ_INSTOP_READ) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
239
aq_req.op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
263
if (req->op != NPA_AQ_INSTOP_INIT)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
269
lock_ctx_req.op = NPA_AQ_INSTOP_LOCK;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
97
inst.op = req->op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1037
action.op = NIX_RX_ACTIONOP_RSS;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1212
rule->rx_action.op != NIX_RX_ACTIONOP_MCAST) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
636
action.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
647
req.op = action.op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
694
if (action.op != NIX_RX_ACTIONOP_RSS) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
696
action.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
705
action.op = NIX_RX_ACTIONOP_MCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
735
req.op = action.op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
806
req.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
808
req.op = NIX_RX_ACTIONOP_MCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
872
if (action.op != NIX_RX_ACTIONOP_RSS) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
874
action.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
881
action.op = NIX_RX_ACTIONOP_MCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
903
req.op = action.op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1107
u64 op, void *action)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1118
(op == NIX_RX_ACTIONOP_MCAST || op == NIX_TX_ACTIONOP_MCAST)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1123
if (op == NIX_RX_ACTIONOP_MCAST)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1149
action.op = req->op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1152
ret = npc_mcast_update_action_index(rvu, req, action.op, (void *)&action);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1159
if (req->op == NIX_RX_ACTION_DEFAULT) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1170
action.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1211
action.op = req->op;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1214
ret = npc_mcast_update_action_index(rvu, req, action.op, (void *)&action);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1385
(req->op == NIX_RX_ACTIONOP_UCAST || req->op == NIX_RX_ACTIONOP_RSS))
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1389
if (owner && req->op == NIX_RX_ACTIONOP_MCAST)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc_fs.c
1660
if (rule->rx_action.op == NIX_RX_ACTION_DEFAULT) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
249
req.op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
254
req.op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
306
req.op = NIX_TX_ACTIONOP_UCAST_CHAN;
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
160
u64 op : 4; /* W0 */
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
172
u64 op : 4; /* W0 */
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
322
u64 op : 4;
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
334
u64 op : 4;
drivers/net/ethernet/marvell/octeontx2/af/rvu_struct.h
698
uint64_t op : 2;
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
45
req.op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/af/rvu_switch.c
76
req.op = NIX_TX_ACTIONOP_UCAST_CHAN;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
118
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
353
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
492
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
326
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
377
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
429
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1149
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1448
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1506
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
361
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
918
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
955
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
163
cn10k_sq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
178
sq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
369
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
384
npa_aq->op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1003
req->op = NIX_RX_ACTIONOP_DROP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1009
req->op = NIX_RX_ACTIONOP_RSS;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1013
req->op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1025
pfvf->pfc_en && req->op != NIX_RX_ACTIONOP_RSS) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1428
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
393
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
930
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2541
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2656
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2703
req->op = NIX_TX_ACTIONOP_UCAST_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
134
req->op = PTP_OP_SET_CLOCK;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
155
req->op = PTP_OP_ADJFINE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
172
req->op = PTP_OP_SET_THRESH;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
189
req->op = PTP_OP_PPS_ON;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
216
req->op = PTP_OP_GET_TSTMP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
402
req->op = PTP_OP_GET_CLOCK;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
64
req->op = PTP_OP_ADJTIME;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ptp.c
85
req->op = PTP_OP_GET_CLOCK;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
346
req->op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
405
ureq->op = NIX_MCAST_OP_ADD_ENTRY;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
424
req->op = NIX_RX_ACTIONOP_MCAST;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
461
req->op = NIX_RX_ACTIONOP_DROP;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
464
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
483
if (!req->op)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
484
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
526
req->op = NIX_RX_ACTION_DEFAULT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_tc.c
532
req->op = NIX_RX_ACTIONOP_UCAST;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
51
cn10k_rq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
60
rq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
71
aura_aq->op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
82
pool_aq->op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
181
cn10k_sq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
190
sq_aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
203
aura_aq->op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/qos_sq.c
216
pool_aq->op = NPA_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/sky2.c
1177
static void sky2_rx_add(struct sky2_port *sky2, u8 op,
drivers/net/ethernet/marvell/sky2.c
1191
le->opcode = op | HW_OWNER;
drivers/net/ethernet/mellanox/mlx4/cmd.c
1022
vhcr->op, MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
drivers/net/ethernet/mellanox/mlx4/cmd.c
1052
vhcr->in_modifier, vhcr->op_modifier, vhcr->op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
1702
vhcr->op = be16_to_cpu(vhcr_cmd->opcode) & 0xfff;
drivers/net/ethernet/mellanox/mlx4/cmd.c
1708
if (vhcr->op == cmd_info[i].opcode) {
drivers/net/ethernet/mellanox/mlx4/cmd.c
1715
vhcr->op, slave);
drivers/net/ethernet/mellanox/mlx4/cmd.c
1746
vhcr->op, slave, vhcr->in_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
1774
vhcr->op_modifier, vhcr->op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
1786
if (vhcr->op == MLX4_CMD_ALLOC_RES &&
drivers/net/ethernet/mellanox/mlx4/cmd.c
1794
vhcr->op, slave, vhcr->errno, err);
drivers/net/ethernet/mellanox/mlx4/cmd.c
189
static int mlx4_internal_err_ret_value(struct mlx4_dev *dev, u16 op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
192
switch (op) {
drivers/net/ethernet/mellanox/mlx4/cmd.c
218
static int mlx4_closing_cmd_fatal_error(u16 op, u8 fw_status)
drivers/net/ethernet/mellanox/mlx4/cmd.c
221
if (op == MLX4_CMD_CLOSE_HCA ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
222
op == MLX4_CMD_HW2SW_EQ ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
223
op == MLX4_CMD_HW2SW_CQ ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
224
op == MLX4_CMD_2RST_QP ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
225
op == MLX4_CMD_HW2SW_SRQ ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
226
op == MLX4_CMD_SYNC_TPT ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
227
op == MLX4_CMD_UNMAP_ICM ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
228
op == MLX4_CMD_UNMAP_ICM_AUX ||
drivers/net/ethernet/mellanox/mlx4/cmd.c
229
op == MLX4_CMD_UNMAP_FA)
drivers/net/ethernet/mellanox/mlx4/cmd.c
236
if (op == MLX4_CMD_HW2SW_MPT && fw_status != CMD_STAT_REG_BOUND)
drivers/net/ethernet/mellanox/mlx4/cmd.c
241
static int mlx4_cmd_reset_flow(struct mlx4_dev *dev, u16 op, u8 op_modifier,
drivers/net/ethernet/mellanox/mlx4/cmd.c
249
err = mlx4_internal_err_ret_value(dev, op, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
339
u16 param, u16 op, unsigned long timeout)
drivers/net/ethernet/mellanox/mlx4/cmd.c
368
vhcr_cmd, op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
376
if (mlx4_closing_cmd_fatal_error(op, context->fw_status))
drivers/net/ethernet/mellanox/mlx4/cmd.c
408
u16 op, unsigned long timeout)
drivers/net/ethernet/mellanox/mlx4/cmd.c
414
return mlx4_comm_cmd_wait(dev, cmd, param, op, timeout);
drivers/net/ethernet/mellanox/mlx4/cmd.c
433
u32 in_modifier, u8 op_modifier, u16 op, u16 token,
drivers/net/ethernet/mellanox/mlx4/cmd.c
496
op), hcr + 6);
drivers/net/ethernet/mellanox/mlx4/cmd.c
505
op, ret, in_param, in_modifier, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
513
u16 op, unsigned long timeout)
drivers/net/ethernet/mellanox/mlx4/cmd.c
524
vhcr->opcode = cpu_to_be16((((u16) op_modifier) << 12) | (op & 0xfff));
drivers/net/ethernet/mellanox/mlx4/cmd.c
538
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
546
ret = mlx4_internal_err_ret_value(dev, op, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
548
ret = mlx4_comm_cmd(dev, MLX4_COMM_CMD_VHCR_POST, 0, op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
557
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
565
ret = mlx4_internal_err_ret_value(dev, op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
568
mlx4_err(dev, "failed execution of VHCR_POST command opcode 0x%x\n", op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
578
u16 op, unsigned long timeout)
drivers/net/ethernet/mellanox/mlx4/cmd.c
593
err = mlx4_internal_err_ret_value(dev, op, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
599
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
605
in_modifier, op_modifier, op, CMD_POLL_TOKEN, 0);
drivers/net/ethernet/mellanox/mlx4/cmd.c
621
err = mlx4_internal_err_ret_value(dev, op, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
630
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
646
op, stat);
drivers/net/ethernet/mellanox/mlx4/cmd.c
647
if (mlx4_closing_cmd_fatal_error(op, stat))
drivers/net/ethernet/mellanox/mlx4/cmd.c
654
err = mlx4_cmd_reset_flow(dev, op, op_modifier, err);
drivers/net/ethernet/mellanox/mlx4/cmd.c
679
u16 op, unsigned long timeout)
drivers/net/ethernet/mellanox/mlx4/cmd.c
697
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
705
in_modifier, op_modifier, op, context->token, 1);
drivers/net/ethernet/mellanox/mlx4/cmd.c
709
if (op == MLX4_CMD_SENSE_PORT) {
drivers/net/ethernet/mellanox/mlx4/cmd.c
724
op);
drivers/net/ethernet/mellanox/mlx4/cmd.c
725
if (op == MLX4_CMD_NOP) {
drivers/net/ethernet/mellanox/mlx4/cmd.c
741
if (op == MLX4_CMD_SET_PORT &&
drivers/net/ethernet/mellanox/mlx4/cmd.c
746
op, context->fw_status);
drivers/net/ethernet/mellanox/mlx4/cmd.c
749
op, context->fw_status);
drivers/net/ethernet/mellanox/mlx4/cmd.c
751
err = mlx4_internal_err_ret_value(dev, op, op_modifier);
drivers/net/ethernet/mellanox/mlx4/cmd.c
752
else if (mlx4_closing_cmd_fatal_error(op, context->fw_status))
drivers/net/ethernet/mellanox/mlx4/cmd.c
763
err = mlx4_cmd_reset_flow(dev, op, op_modifier, err);
drivers/net/ethernet/mellanox/mlx4/cmd.c
776
u16 op, unsigned long timeout, int native)
drivers/net/ethernet/mellanox/mlx4/cmd.c
779
return mlx4_cmd_reset_flow(dev, op, op_modifier, -EIO);
drivers/net/ethernet/mellanox/mlx4/cmd.c
785
return mlx4_internal_err_ret_value(dev, op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
791
op_modifier, op, timeout);
drivers/net/ethernet/mellanox/mlx4/cmd.c
795
op_modifier, op, timeout);
drivers/net/ethernet/mellanox/mlx4/cmd.c
801
in_modifier, op_modifier, op, timeout);
drivers/net/ethernet/mellanox/mlx4/cmd.c
951
vhcr->op, MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
drivers/net/ethernet/mellanox/mlx4/cmd.c
976
vhcr->op,
drivers/net/ethernet/mellanox/mlx4/cmd.c
995
vhcr->op, MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
drivers/net/ethernet/mellanox/mlx4/fw.c
1519
int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt)
drivers/net/ethernet/mellanox/mlx4/fw.c
1566
err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
drivers/net/ethernet/mellanox/mlx4/fw.c
1577
err = mlx4_cmd(dev, mailbox->dma, nent, 0, op,
drivers/net/ethernet/mellanox/mlx4/fw.c
1582
switch (op) {
drivers/net/ethernet/mellanox/mlx4/fw.h
249
int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt);
drivers/net/ethernet/mellanox/mlx4/mlx4.h
1229
u16 op, unsigned long timeout);
drivers/net/ethernet/mellanox/mlx4/mlx4.h
181
u16 op;
drivers/net/ethernet/mellanox/mlx4/qp.c
146
!op[cur_state][new_state])
drivers/net/ethernet/mellanox/mlx4/qp.c
149
if (op[cur_state][new_state] == MLX4_CMD_2RST_QP) {
drivers/net/ethernet/mellanox/mlx4/qp.c
190
op[cur_state][new_state], MLX4_CMD_TIME_CLASS_C, native);
drivers/net/ethernet/mellanox/mlx4/qp.c
98
static const u16 op[MLX4_QP_NUM_STATE][MLX4_QP_NUM_STATE] = {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1779
static int qp_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1789
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1846
static int mtt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1853
if (op != RES_OP_RESERVE_AND_MAP)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1879
static int mpt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1887
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1928
static int cq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1934
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1963
static int srq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
1969
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2088
static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2096
if (op != RES_OP_RESERVE_AND_MAP)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2197
static int vlan_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2209
if (!port || op != RES_OP_RESERVE_AND_MAP)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2235
static int counter_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2241
if (op != RES_OP_RESERVE)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2265
static int xrcdn_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2271
if (op != RES_OP_RESERVE)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2352
static int qp_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2360
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2392
static int mtt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2399
if (op != RES_OP_RESERVE_AND_MAP)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2412
static int mpt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2420
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2454
static int cq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2460
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2479
static int srq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2485
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2504
static int mac_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2510
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2530
static int vlan_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2542
switch (op) {
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2559
static int counter_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2565
if (op != RES_OP_RESERVE)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2582
static int xrcdn_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
2588
if (op != RES_OP_RESERVE)
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
965
static int counter_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
105
static bool mlx5_cmd_is_throttle_opcode(u16 op)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1063
if (mlx5_cmd_is_down(dev) || !opcode_allowed(&dev->cmd, ent->op)) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
107
switch (op) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1155
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1160
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1190
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1193
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1196
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1317
stats = xa_load(&cmd->stats, ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1326
mlx5_command_str(ent->op), ds);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
139
ent->op = in_to_opcode(in->first.data);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1775
!opcode_allowed(cmd, ent->op))
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1797
stats = xa_load(&cmd->stats, ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
346
static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
352
switch (op) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
546
mlx5_core_err(dev, "Unknown FW command (%d)\n", op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
879
u16 op = ent->op;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
891
ent->idx, mlx5_command_str(op), op,
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
895
ent->idx, mlx5_command_str(op), op,
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
947
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
953
ent->idx, mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
999
mlx5_command_str(ent->op), ent->op);
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
191
static void esw_qos_sched_elem_warn(struct mlx5_esw_sched_node *node, int err, const char *op)
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
197
op, sched_node_type_str[node->type], node->tc, err);
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
202
op,
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
209
op, sched_node_type_str[node->type], node->vport->vport, err);
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
216
op, sched_node_type_str[node->type], err);
drivers/net/ethernet/mellanox/mlx5/core/esw/qos.c
221
op, err);
drivers/net/ethernet/mellanox/mlx5/core/fpga/cmd.c
82
int mlx5_fpga_ctrl_op(struct mlx5_core_dev *dev, u8 op)
drivers/net/ethernet/mellanox/mlx5/core/fpga/cmd.c
87
MLX5_SET(fpga_ctrl, in, operation, op);
drivers/net/ethernet/mellanox/mlx5/core/fpga/cmd.h
77
int mlx5_fpga_ctrl_op(struct mlx5_core_dev *dev, u8 op);
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
154
if (mpesww->op == MLX5_MPESW_OP_ENABLE)
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
156
else if (mpesww->op == MLX5_MPESW_OP_DISABLE)
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
165
enum mpesw_op op)
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.c
180
work->op = op;
drivers/net/ethernet/mellanox/mlx5/core/lag/mpesw.h
23
enum mpesw_op op;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
120
struct mlx5_flow_handle *op;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1988
mlx5_del_flow_rules(rx_rule->op);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2172
rx_rule->op = new_rule;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/cmd.c
567
stc_attr->reformat_trailer.op);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/cmd.h
152
u8 op;
drivers/net/ethernet/mellanox/mlxsw/core_linecards.c
452
mlxsw_linecards_event_op_t *op, void *priv)
drivers/net/ethernet/mellanox/mlxsw/core_linecards.c
456
if (!op)
drivers/net/ethernet/mellanox/mlxsw/core_linecards.c
458
op(mlxsw_core, linecard->slot_index, priv);
drivers/net/ethernet/mellanox/mlxsw/reg.h
11379
MLXSW_ITEM32(reg, mbct, op, 0x08, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
11436
enum mlxsw_reg_mbct_op op, bool oee)
drivers/net/ethernet/mellanox/mlxsw/reg.h
11440
mlxsw_reg_mbct_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
1237
MLXSW_ITEM32(reg, sldr, op, 0x00, 29, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
1850
MLXSW_ITEM32(reg, sfmr, op, 0x00, 24, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
1996
enum mlxsw_reg_sfmr_op op, u16 fid,
drivers/net/ethernet/mellanox/mlxsw/reg.h
2000
mlxsw_reg_sfmr_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
204
MLXSW_ITEM32(reg, sfd, op, 0x04, 30, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
224
static inline void mlxsw_reg_sfd_pack(char *payload, enum mlxsw_reg_sfd_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
228
mlxsw_reg_sfd_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2616
MLXSW_ITEM32(reg, ppbt, op, 0x00, 28, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2640
enum mlxsw_reg_pxbt_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
2645
mlxsw_reg_ppbt_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2783
MLXSW_ITEM32(reg, ptar, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2840
static inline void mlxsw_reg_ptar_pack(char *payload, enum mlxsw_reg_ptar_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
2846
mlxsw_reg_ptar_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2992
MLXSW_ITEM32(reg, prcr, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
3026
static inline void mlxsw_reg_prcr_pack(char *payload, enum mlxsw_reg_prcr_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
3033
mlxsw_reg_prcr_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
3184
MLXSW_ITEM32(reg, ptce2, op, 0x00, 20, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
3233
enum mlxsw_reg_ptce2_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
3239
mlxsw_reg_ptce2_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
3430
MLXSW_ITEM32(reg, ptce3, op, 0x00, 20, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
3537
enum mlxsw_reg_ptce3_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
3547
mlxsw_reg_ptce3_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
6842
MLXSW_ITEM32(reg, ritr, op, 0x00, 20, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7143
bool op = enable ? MLXSW_REG_RITR_RIF_CREATE : MLXSW_REG_RITR_RIF_DEL;
drivers/net/ethernet/mellanox/mlxsw/reg.h
7152
mlxsw_reg_ritr_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7243
MLXSW_ITEM32(reg, rtar, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7266
enum mlxsw_reg_rtar_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
7271
mlxsw_reg_rtar_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7308
MLXSW_ITEM32(reg, ratr, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7468
enum mlxsw_reg_ratr_op op, bool valid,
drivers/net/ethernet/mellanox/mlxsw/reg.h
7473
mlxsw_reg_ratr_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7595
MLXSW_ITEM32(reg, ricnt, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7663
enum mlxsw_reg_ricnt_opcode op)
drivers/net/ethernet/mellanox/mlxsw/reg.h
7666
mlxsw_reg_ricnt_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7691
MLXSW_ITEM32(reg, rrcr, op, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7718
static inline void mlxsw_reg_rrcr_pack(char *payload, enum mlxsw_reg_rrcr_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
7724
mlxsw_reg_rrcr_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7746
MLXSW_ITEM32(reg, ralta, op, 0x00, 28, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
7937
MLXSW_ITEM32(reg, ralue, op, 0x00, 20, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8096
enum mlxsw_reg_ralue_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8101
mlxsw_reg_ralue_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8111
enum mlxsw_reg_ralue_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8115
mlxsw_reg_ralue_pack(payload, protocol, op, virtual_router, prefix_len);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8121
enum mlxsw_reg_ralue_op op,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8125
mlxsw_reg_ralue_pack(payload, protocol, op, virtual_router, prefix_len);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8219
MLXSW_ITEM32(reg, rauht, op, 0x00, 20, 3);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8289
enum mlxsw_reg_rauht_op op, u16 rif,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8293
mlxsw_reg_rauht_op_set(payload, op);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8299
enum mlxsw_reg_rauht_op op, u16 rif,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8302
mlxsw_reg_rauht_pack(payload, op, rif, mac);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8307
enum mlxsw_reg_rauht_op op, u16 rif,
drivers/net/ethernet/mellanox/mlxsw/reg.h
8310
mlxsw_reg_rauht_pack(payload, op, rif, mac);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8426
MLXSW_ITEM32(reg, rauhtd, op, 0x04, 24, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
8767
MLXSW_ITEM32(reg, ratrad, op, 0x00, 30, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
9058
MLXSW_ITEM32(reg, rmft2, op, 0x00, 20, 2);
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
114
enum mlxsw_reg_mgpc_opcode op = clear ? MLXSW_REG_MGPC_OPCODE_CLEAR :
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
119
mlxsw_reg_mgpc_pack(mgpc_pl, counter_index, op,
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
106
enum mlxsw_reg_sfmr_op op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
519
enum mlxsw_reg_sfmr_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
525
mlxsw_reg_sfmr_pack(sfmr_pl, op, fid->fid_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
531
enum mlxsw_reg_sfmr_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
533
mlxsw_sp_fid_pack(sfmr_pl, fid, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
555
enum mlxsw_reg_sfmr_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_fid.c
560
mlxsw_sp_fid_pack(sfmr_pl, fid, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
176
enum mlxsw_reg_ratr_op op;
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
178
op = force ? MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY :
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
180
mlxsw_reg_ratr_pack(ratr_pl, op, true, MLXSW_REG_RATR_TYPE_IPIP,
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
393
enum mlxsw_reg_ratr_op op;
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
395
op = force ? MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY :
drivers/net/ethernet/mellanox/mlxsw/spectrum_ipip.c
397
mlxsw_reg_ratr_pack(ratr_pl, op, true, MLXSW_REG_RATR_TYPE_IPIP,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2674
enum mlxsw_reg_rauht_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2680
mlxsw_reg_rauht_pack4(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2691
enum mlxsw_reg_rauht_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2697
mlxsw_reg_rauht_pack6(rauht_pl, op, neigh_entry->rif, neigh_entry->ha,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2724
enum mlxsw_reg_rauht_op op = mlxsw_sp_rauht_op(adding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2732
op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2739
op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3747
enum mlxsw_reg_ratr_op op;
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3752
op = force ? MLXSW_REG_RATR_OP_WRITE_WRITE_ENTRY :
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
3754
mlxsw_reg_ratr_pack(ratr_pl, op, true, MLXSW_REG_RATR_TYPE_ETHERNET,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6164
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6166
switch (op) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6181
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6192
mlxsw_reg_ralue_pack4(ralue_pl, proto, op, fib->vr->id,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6197
mlxsw_reg_ralue_pack6(ralue_pl, proto, op, fib->vr->id,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6206
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6234
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6242
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6258
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6266
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6270
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6277
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6283
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6291
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6300
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6308
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6324
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6332
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6336
mlxsw_sp_fib_entry_ralue_pack(ralue_pl, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6344
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6348
return mlxsw_sp_fib_entry_op_remote(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6350
return mlxsw_sp_fib_entry_op_local(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6352
return mlxsw_sp_fib_entry_op_trap(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6354
return mlxsw_sp_fib_entry_op_blackhole(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6357
op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6360
fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6362
return mlxsw_sp_fib_entry_op_nve_decap(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6369
enum mlxsw_reg_ralue_op op)
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6371
int err = __mlxsw_sp_fib_entry_op(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
6376
mlxsw_sp_fib_entry_hw_flags_refresh(mlxsw_sp, fib_entry, op);
drivers/net/ethernet/micrel/ks8851_common.c
92
static void ks8851_soft_reset(struct ks8851_net *ks, unsigned op)
drivers/net/ethernet/micrel/ks8851_common.c
94
ks8851_wrreg16(ks, KS_GRR, op);
drivers/net/ethernet/micrel/ks8851_spi.c
145
static void ks8851_rdreg(struct ks8851_net *ks, unsigned int op,
drivers/net/ethernet/micrel/ks8851_spi.c
155
txb[0] = cpu_to_le16(op | KS_SPIOP_RD);
drivers/net/ethernet/microchip/enc28j60.c
142
static u8 spi_read_op(struct enc28j60_net *priv, u8 op, u8 addr)
drivers/net/ethernet/microchip/enc28j60.c
155
tx_buf[0] = op | (addr & ADDR_MASK);
drivers/net/ethernet/microchip/enc28j60.c
169
static int spi_write_op(struct enc28j60_net *priv, u8 op, u8 addr, u8 val)
drivers/net/ethernet/microchip/enc28j60.c
174
priv->spi_transfer_buf[0] = op | (addr & ADDR_MASK);
drivers/net/ethernet/microchip/lan743x_main.c
844
static u32 lan743x_mac_mmd_access(int id, int dev_addr, int op)
drivers/net/ethernet/microchip/lan743x_main.c
852
if (op == MMD_ACCESS_WRITE)
drivers/net/ethernet/microchip/lan743x_main.c
854
else if (op == MMD_ACCESS_READ)
drivers/net/ethernet/microchip/lan743x_main.c
856
else if (op == MMD_ACCESS_READ_INC)
drivers/net/ethernet/microchip/vcap/vcap_tc.c
335
if (mt.mask->op) {
drivers/net/ethernet/microchip/vcap/vcap_tc.c
338
value = mt.key->op == VCAP_ARP_OP_REQUEST ?
drivers/net/ethernet/microchip/vcap/vcap_tc.c
342
value = mt.key->op == VCAP_ARP_OP_REQUEST ?
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
181
static bool nfp_bpf_ctrl_op_cache_invalidate(enum nfp_ccm_type op)
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
183
return op == NFP_CCM_TYPE_BPF_MAP_UPDATE ||
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
184
op == NFP_CCM_TYPE_BPF_MAP_DELETE;
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
187
static bool nfp_bpf_ctrl_op_cache_capable(enum nfp_ccm_type op)
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
189
return op == NFP_CCM_TYPE_BPF_MAP_LOOKUP ||
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
190
op == NFP_CCM_TYPE_BPF_MAP_GETNEXT;
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
193
static bool nfp_bpf_ctrl_op_cache_fill(enum nfp_ccm_type op)
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
195
return op == NFP_CCM_TYPE_BPF_MAP_GETFIRST ||
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
196
op == NFP_CCM_TYPE_BPF_MAP_GETNEXT;
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
200
nfp_bpf_ctrl_op_cache_get(struct nfp_bpf_map *nfp_map, enum nfp_ccm_type op,
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
209
n_entries = nfp_bpf_ctrl_op_cache_fill(op) ? bpf->cmsg_cache_cnt : 1;
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
216
if (nfp_bpf_ctrl_op_cache_invalidate(op))
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
218
if (!nfp_bpf_ctrl_op_cache_capable(op))
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
236
if (op == NFP_CCM_TYPE_BPF_MAP_LOOKUP)
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
239
if (op == NFP_CCM_TYPE_BPF_MAP_GETNEXT) {
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
264
nfp_bpf_ctrl_op_cache_put(struct nfp_bpf_map *nfp_map, enum nfp_ccm_type op,
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
269
blocker = nfp_bpf_ctrl_op_cache_invalidate(op);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
270
filler = nfp_bpf_ctrl_op_cache_fill(op);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
294
nfp_bpf_ctrl_entry_op(struct bpf_offloaded_map *offmap, enum nfp_ccm_type op,
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
312
n_entries = nfp_bpf_ctrl_op_cache_get(nfp_map, op, key, out_key,
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
335
skb = nfp_ccm_communicate(&bpf->ccm, skb, op, 0);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
343
op, skb->len);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
362
op, skb->len, reply_entries);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
375
nfp_bpf_ctrl_op_cache_put(nfp_map, op, skb, cache_gen);
drivers/net/ethernet/netronome/nfp/bpf/cmsg.c
381
nfp_bpf_ctrl_op_cache_put(nfp_map, op, NULL, cache_gen);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
100
emit_cmd_any(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
122
__emit_cmd(nfp_prog, op, mode, xfer, reg.areg, reg.breg, size, ctx,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
127
emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
130
emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, ctx, false);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
134
emit_cmd_indir(struct nfp_prog *nfp_prog, enum cmd_tgt_map op, u8 mode, u8 xfer,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
137
emit_cmd_any(nfp_prog, op, mode, xfer, lreg, rreg, size, ctx, true);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1388
unsigned int op;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1390
op = BPF_OP(meta->insn.code) >> 4;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1392
if (WARN_ONCE(op >= ARRAY_SIZE(jmp_code_map) ||
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1393
!jmp_code_map[op].br_mask,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1397
return &jmp_code_map[op];
drivers/net/ethernet/netronome/nfp/bpf/jit.c
321
u16 areg, enum shf_op op, u16 breg, bool i8, bool sw, bool wr_both,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
353
FIELD_PREP(OP_SHF_OP, op) |
drivers/net/ethernet/netronome/nfp/bpf/jit.c
364
swreg lreg, enum shf_op op, swreg rreg, enum shf_sc sc, u8 shift)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
376
reg.areg, op, reg.breg, reg.i8, reg.swap, reg.wr_both,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
382
swreg lreg, enum shf_op op, swreg rreg, enum shf_sc sc)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
390
emit_shf(nfp_prog, dst, lreg, op, rreg, sc, 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
395
u16 areg, enum alu_op op, u16 breg, bool swap, bool wr_both,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
405
FIELD_PREP(OP_ALU_OP, op) |
drivers/net/ethernet/netronome/nfp/bpf/jit.c
416
swreg lreg, enum alu_op op, swreg rreg)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
428
reg.areg, op, reg.breg, reg.swap, reg.wr_both,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
79
__emit_cmd(struct nfp_prog *nfp_prog, enum cmd_tgt_map op,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
88
FIELD_PREP(OP_CMD_TOKEN, cmd_tgt_act[op].token) |
drivers/net/ethernet/netronome/nfp/bpf/jit.c
92
FIELD_PREP(OP_CMD_TGT_CMD, cmd_tgt_act[op].tgt_cmd) |
drivers/net/ethernet/netronome/nfp/bpf/main.h
448
u8 op;
drivers/net/ethernet/netronome/nfp/bpf/main.h
456
op = mbpf_op(meta);
drivers/net/ethernet/netronome/nfp/bpf/main.h
457
return op != BPF_JA && op != BPF_EXIT && op != BPF_CALL;
drivers/net/ethernet/netronome/nfp/flower/main.h
728
enum nfp_meter_op op,
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
623
enum nfp_meter_op op,
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
632
switch (op) {
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
1768
nfp_net_fs_fill_v4(struct nfp_net *nn, struct nfp_fs_entry *entry, u32 op, u32 *addr)
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
1782
nn_writel(nn, *addr, op);
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
1802
nfp_net_fs_fill_v6(struct nfp_net *nn, struct nfp_fs_entry *entry, u32 op, u32 *addr)
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
1816
nn_writel(nn, *addr, op);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1179
cpp->op = ops;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1214
if (cpp->op->init) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1215
err = cpp->op->init(cpp);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1292
if (cpp->op->func) \
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1293
err = cpp->op->func(expl, ##args); \
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1301
if (cpp->op->func) \
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1302
cpp->op->func(expl, ##args); \
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
1320
expl = kzalloc(sizeof(*expl) + cpp->op->explicit_priv_size, GFP_KERNEL);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
136
if (area->cpp->op->area_cleanup)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
137
area->cpp->op->area_cleanup(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
192
if (area->cpp->op->area_release)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
193
area->cpp->op->area_release(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
198
if (cpp->op->free)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
199
cpp->op->free(cpp);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
303
area = kzalloc(sizeof(*area) + cpp->op->area_priv_size + name_len,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
310
cpp->op->area_priv_size;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
322
if (cpp->op->area_init) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
325
err = cpp->op->area_init(area, dest, address, size);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
412
*status = area->cpp->op->area_acquire(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
424
if (!area->cpp->op->area_acquire)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
477
if (area->cpp->op->area_acquire) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
480
err = area->cpp->op->area_acquire(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
505
if (area->cpp->op->area_release) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
506
area->cpp->op->area_release(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
546
return area->cpp->op->area_read(area, kernel_vaddr, offset, length);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
567
return area->cpp->op->area_write(area, kernel_vaddr, offset, length);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
626
if (area->cpp->op->area_resource)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
627
res = area->cpp->op->area_resource(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
644
if (area->cpp->op->area_phys)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
645
addr = area->cpp->op->area_phys(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
665
if (area->cpp->op->area_iomem)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
666
iomem = area->cpp->op->area_iomem(area);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
68
const struct nfp_cpp_operations *op;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
880
if (cpp->op->area_init) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_cppcore.c
881
err = cpp->op->area_init(cache->area,
drivers/net/ethernet/qlogic/netxen/netxen_nic.h
1132
u8 op;
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
1367
u64 *data, int op)
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
1398
if (op == 0) /* read */
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
602
nx_p3_sre_macaddr_change(struct netxen_adapter *adapter, u8 *addr, unsigned op)
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
615
mac_req->op = op;
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
853
struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common;
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
856
if (cookie && op->dcbx_aen)
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
857
op->dcbx_aen(cookie, &hwfn->p_dcbx_info->get, mib_type);
drivers/net/ethernet/qlogic/qed/qed_l2.c
2797
struct qed_common_cb_ops *op = p_hwfn->cdev->protocol_ops.common;
drivers/net/ethernet/qlogic/qed/qed_l2.c
2800
op->arfs_filter_op(dev, cookie, fw_return_code);
drivers/net/ethernet/qlogic/qed/qed_main.c
2207
struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common;
drivers/net/ethernet/qlogic/qed/qed_main.c
2214
op->link_update(cookie, &if_link);
drivers/net/ethernet/qlogic/qed/qed_main.c
2220
struct qed_common_cb_ops *op = hwfn->cdev->protocol_ops.common;
drivers/net/ethernet/qlogic/qed/qed_main.c
2222
if (IS_LEAD_HWFN(hwfn) && cookie && op && op->bw_update)
drivers/net/ethernet/qlogic/qed/qed_main.c
2223
op->bw_update(cookie);
drivers/net/ethernet/qlogic/qed/qed_main.c
3132
struct qed_common_cb_ops *op = cdev->protocol_ops.common;
drivers/net/ethernet/qlogic/qed/qed_main.c
3139
op->get_generic_tlv_data(cdev->ops_cookie, &gen_tlvs);
drivers/net/ethernet/qlogic/qede/qede_filter.c
123
op->ntuple_filter_config(edev->cdev, n, &params);
drivers/net/ethernet/qlogic/qede/qede_filter.c
92
const struct qed_eth_ops *op = edev->ops;
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
990
u8 op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2096
u16 vlan_id, u8 op)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2117
op = (op == QLCNIC_MAC_ADD || op == QLCNIC_MAC_VLAN_ADD) ?
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2120
cmd->req.arg[1] = op | (1 << 8);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2422
u8 op = 0;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2432
cmd.req.arg[1] = op | BIT_31 | temp;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1245
u32 window, u64 off, u64 *data, int op)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
1259
if (op == 0) /* read */
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
437
u16 vlan_id, u8 op)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
451
mac_req->op = op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
161
u8 op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
163
op = vlan_id ? QLCNIC_MAC_VLAN_ADD : QLCNIC_MAC_ADD;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
164
ret = qlcnic_sre_macaddr_change(adapter, addr, vlan_id, op);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
168
op = vlan_id ? QLCNIC_MAC_VLAN_DEL : QLCNIC_MAC_DEL;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
169
ret = qlcnic_sre_macaddr_change(adapter, addr, vlan_id, op);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
199
u8 hindex, op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
241
op = vlan_id ? QLCNIC_MAC_VLAN_DEL : QLCNIC_MAC_DEL;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
244
vlan_id, op);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
291
mac_req->op = vlan_id ? QLCNIC_MAC_VLAN_ADD : QLCNIC_MAC_ADD;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1259
u8 op, new_op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1268
op = cmd->req.arg[1] & 0x7;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1270
new_op = (op == QLCNIC_MAC_ADD || op == QLCNIC_MAC_VLAN_ADD) ?
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1438
u8 op;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1445
op = cmd->req.arg[1] & 0xf;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
1447
if (op)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
790
u16 vlan, u8 op)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
818
op = ((op == QLCNIC_MAC_ADD || op == QLCNIC_MAC_VLAN_ADD) ?
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_pf.c
821
cmd->req.arg[1] = op | (1 << 8) | (3 << 6);
drivers/net/ethernet/sfc/siena/siena_sriov.c
898
if (req->op < VFDI_OP_LIMIT && vfdi_ops[req->op] != NULL) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
899
rc = vfdi_ops[req->op](vf);
drivers/net/ethernet/sfc/siena/siena_sriov.c
903
req->op, vf->pci_name);
drivers/net/ethernet/sfc/siena/siena_sriov.c
908
"%llx\n", req->op, vf->pci_name,
drivers/net/ethernet/sfc/siena/siena_sriov.c
919
req->op = VFDI_OP_RESPONSE;
drivers/net/ethernet/sfc/siena/siena_sriov.c
926
copy[1].from_buf = &req->op;
drivers/net/ethernet/sfc/siena/siena_sriov.c
928
copy[1].to_addr = vf->req_addr + offsetof(struct vfdi_req, op);
drivers/net/ethernet/sfc/siena/siena_sriov.c
929
copy[1].length = sizeof(req->op);
drivers/net/ethernet/sfc/siena/vfdi.h
163
u32 op;
drivers/net/ethernet/smsc/smsc911x.c
2016
static int smsc911x_eeprom_send_cmd(struct smsc911x_data *pdata, u32 op)
drivers/net/ethernet/smsc/smsc911x.c
2021
SMSC_TRACE(pdata, drv, "op 0x%08x", op);
drivers/net/ethernet/smsc/smsc911x.c
2027
e2cmd = op | E2P_CMD_EPC_BUSY_;
drivers/net/ethernet/smsc/smsc911x.c
2051
u32 op = E2P_CMD_EPC_CMD_READ_ | address;
drivers/net/ethernet/smsc/smsc911x.c
2055
ret = smsc911x_eeprom_send_cmd(pdata, op);
drivers/net/ethernet/smsc/smsc911x.c
2066
u32 op = E2P_CMD_EPC_CMD_ERASE_ | address;
drivers/net/ethernet/smsc/smsc911x.c
2070
ret = smsc911x_eeprom_send_cmd(pdata, op);
drivers/net/ethernet/smsc/smsc911x.c
2073
op = E2P_CMD_EPC_CMD_WRITE_ | address;
drivers/net/ethernet/smsc/smsc911x.c
2079
ret = smsc911x_eeprom_send_cmd(pdata, op);
drivers/net/ethernet/smsc/smsc9420.c
273
static int smsc9420_eeprom_send_cmd(struct smsc9420_pdata *pd, u32 op)
drivers/net/ethernet/smsc/smsc9420.c
278
netif_dbg(pd, hw, pd->dev, "op 0x%08x\n", op);
drivers/net/ethernet/smsc/smsc9420.c
284
e2cmd = op | E2P_CMD_EPC_BUSY_;
drivers/net/ethernet/smsc/smsc9420.c
309
u32 op = E2P_CMD_EPC_CMD_READ_ | address;
drivers/net/ethernet/smsc/smsc9420.c
313
ret = smsc9420_eeprom_send_cmd(pd, op);
drivers/net/ethernet/smsc/smsc9420.c
324
u32 op = E2P_CMD_EPC_CMD_ERASE_ | address;
drivers/net/ethernet/smsc/smsc9420.c
328
ret = smsc9420_eeprom_send_cmd(pd, op);
drivers/net/ethernet/smsc/smsc9420.c
331
op = E2P_CMD_EPC_CMD_WRITE_ | address;
drivers/net/ethernet/smsc/smsc9420.c
333
ret = smsc9420_eeprom_send_cmd(pd, op);
drivers/net/ethernet/sun/cassini.c
1166
val |= CAS_BASE(HP_INSTR_RAM_MID_OP, inst->op);
drivers/net/ethernet/sun/cassini.h
2124
u8 op;
drivers/net/ethernet/sun/niu.c
10063
static int niu_of_probe(struct platform_device *op)
drivers/net/ethernet/sun/niu.c
10073
reg = of_get_property(op->dev.of_node, "reg", NULL);
drivers/net/ethernet/sun/niu.c
10075
dev_err(&op->dev, "%pOF: No 'reg' property, aborting\n",
drivers/net/ethernet/sun/niu.c
10076
op->dev.of_node);
drivers/net/ethernet/sun/niu.c
10080
dev = niu_alloc_and_init(&op->dev, NULL, op,
drivers/net/ethernet/sun/niu.c
10089
parent_id.of = of_get_parent(op->dev.of_node);
drivers/net/ethernet/sun/niu.c
10100
np->regs = of_ioremap(&op->resource[1], 0,
drivers/net/ethernet/sun/niu.c
10101
resource_size(&op->resource[1]),
drivers/net/ethernet/sun/niu.c
10104
dev_err(&op->dev, "Cannot map device registers, aborting\n");
drivers/net/ethernet/sun/niu.c
10109
np->vir_regs_1 = of_ioremap(&op->resource[2], 0,
drivers/net/ethernet/sun/niu.c
10110
resource_size(&op->resource[2]),
drivers/net/ethernet/sun/niu.c
10113
dev_err(&op->dev, "Cannot map device vir registers 1, aborting\n");
drivers/net/ethernet/sun/niu.c
10118
np->vir_regs_2 = of_ioremap(&op->resource[3], 0,
drivers/net/ethernet/sun/niu.c
10119
resource_size(&op->resource[3]),
drivers/net/ethernet/sun/niu.c
10122
dev_err(&op->dev, "Cannot map device vir registers 2, aborting\n");
drivers/net/ethernet/sun/niu.c
10132
dev_err(&op->dev, "Problem fetching invariants of chip, aborting\n");
drivers/net/ethernet/sun/niu.c
10138
dev_err(&op->dev, "Cannot register net device, aborting\n");
drivers/net/ethernet/sun/niu.c
10142
platform_set_drvdata(op, dev);
drivers/net/ethernet/sun/niu.c
10150
of_iounmap(&op->resource[2], np->vir_regs_1,
drivers/net/ethernet/sun/niu.c
10151
resource_size(&op->resource[2]));
drivers/net/ethernet/sun/niu.c
10156
of_iounmap(&op->resource[3], np->vir_regs_2,
drivers/net/ethernet/sun/niu.c
10157
resource_size(&op->resource[3]));
drivers/net/ethernet/sun/niu.c
10162
of_iounmap(&op->resource[1], np->regs,
drivers/net/ethernet/sun/niu.c
10163
resource_size(&op->resource[1]));
drivers/net/ethernet/sun/niu.c
10177
static void niu_of_remove(struct platform_device *op)
drivers/net/ethernet/sun/niu.c
10179
struct net_device *dev = platform_get_drvdata(op);
drivers/net/ethernet/sun/niu.c
10187
of_iounmap(&op->resource[2], np->vir_regs_1,
drivers/net/ethernet/sun/niu.c
10188
resource_size(&op->resource[2]));
drivers/net/ethernet/sun/niu.c
10193
of_iounmap(&op->resource[3], np->vir_regs_2,
drivers/net/ethernet/sun/niu.c
10194
resource_size(&op->resource[3]));
drivers/net/ethernet/sun/niu.c
10199
of_iounmap(&op->resource[1], np->regs,
drivers/net/ethernet/sun/niu.c
10200
resource_size(&op->resource[1]));
drivers/net/ethernet/sun/niu.c
9104
struct platform_device *op = np->op;
drivers/net/ethernet/sun/niu.c
9108
int_prop = of_get_property(op->dev.of_node, "interrupts", NULL);
drivers/net/ethernet/sun/niu.c
9112
for (i = 0; i < op->archdata.num_irqs; i++) {
drivers/net/ethernet/sun/niu.c
9114
np->ldg[i].irq = op->archdata.irqs[i];
drivers/net/ethernet/sun/niu.c
9117
np->num_ldg = op->archdata.num_irqs;
drivers/net/ethernet/sun/niu.c
9259
dp = np->op->dev.of_node;
drivers/net/ethernet/sun/niu.c
9692
struct platform_device *op,
drivers/net/ethernet/sun/niu.c
9707
np->op = op;
drivers/net/ethernet/sun/niu.h
3306
struct platform_device *op;
drivers/net/ethernet/sun/sunbmac.c
1072
static int bigmac_ether_init(struct platform_device *op,
drivers/net/ethernet/sun/sunbmac.c
1093
bp->bigmac_op = op;
drivers/net/ethernet/sun/sunbmac.c
1095
SET_NETDEV_DEV(dev, &op->dev);
drivers/net/ethernet/sun/sunbmac.c
1133
bp->creg = of_ioremap(&op->resource[0], 0,
drivers/net/ethernet/sun/sunbmac.c
1141
bp->bregs = of_ioremap(&op->resource[1], 0,
drivers/net/ethernet/sun/sunbmac.c
1151
bp->tregs = of_ioremap(&op->resource[2], 0,
drivers/net/ethernet/sun/sunbmac.c
1207
of_iounmap(&op->resource[0], bp->creg, CREG_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1209
of_iounmap(&op->resource[1], bp->bregs, BMAC_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1211
of_iounmap(&op->resource[2], bp->tregs, TCVR_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1227
static int bigmac_sbus_probe(struct platform_device *op)
drivers/net/ethernet/sun/sunbmac.c
1229
struct device *parent = op->dev.parent;
drivers/net/ethernet/sun/sunbmac.c
1234
return bigmac_ether_init(op, qec_op);
drivers/net/ethernet/sun/sunbmac.c
1237
static void bigmac_sbus_remove(struct platform_device *op)
drivers/net/ethernet/sun/sunbmac.c
1239
struct bigmac *bp = platform_get_drvdata(op);
drivers/net/ethernet/sun/sunbmac.c
1240
struct device *parent = op->dev.parent;
drivers/net/ethernet/sun/sunbmac.c
1249
of_iounmap(&op->resource[0], bp->creg, CREG_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1250
of_iounmap(&op->resource[1], bp->bregs, BMAC_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1251
of_iounmap(&op->resource[2], bp->tregs, TCVR_REG_SIZE);
drivers/net/ethernet/sun/sunbmac.c
1252
dma_free_coherent(&op->dev,
drivers/net/ethernet/sun/sunhme.c
1411
struct platform_device *op = hp->happy_dev;
drivers/net/ethernet/sun/sunhme.c
1413
sbus_set_sbus64(&op->dev,
drivers/net/ethernet/sun/sunhme.c
2206
struct platform_device *op = hp->happy_dev;
drivers/net/ethernet/sun/sunhme.c
2207
regs = of_get_property(op->dev.of_node, "regs", NULL);
drivers/net/ethernet/sun/sunhme.c
2243
struct platform_device *op;
drivers/net/ethernet/sun/sunhme.c
2246
op = to_platform_device(parent);
drivers/net/ethernet/sun/sunhme.c
2247
qp = platform_get_drvdata(op);
drivers/net/ethernet/sun/sunhme.c
2259
platform_set_drvdata(op, qp);
drivers/net/ethernet/sun/sunhme.c
2488
static int happy_meal_sbus_probe_one(struct platform_device *op, int is_qfe)
drivers/net/ethernet/sun/sunhme.c
2490
struct device_node *dp = op->dev.of_node, *sbus_dp;
drivers/net/ethernet/sun/sunhme.c
2497
sbus_dp = op->dev.parent->of_node;
drivers/net/ethernet/sun/sunhme.c
2504
qp = quattro_sbus_find(op);
drivers/net/ethernet/sun/sunhme.c
2514
dev = devm_alloc_etherdev(&op->dev, sizeof(struct happy_meal));
drivers/net/ethernet/sun/sunhme.c
2517
SET_NETDEV_DEV(dev, &op->dev);
drivers/net/ethernet/sun/sunhme.c
2521
hp->happy_dev = op;
drivers/net/ethernet/sun/sunhme.c
2522
hp->dma_dev = &op->dev;
drivers/net/ethernet/sun/sunhme.c
2533
hp->gregs = devm_platform_ioremap_resource(op, 0);
drivers/net/ethernet/sun/sunhme.c
2535
dev_err(&op->dev, "Cannot map global registers.\n");
drivers/net/ethernet/sun/sunhme.c
2540
hp->etxregs = devm_platform_ioremap_resource(op, 1);
drivers/net/ethernet/sun/sunhme.c
2542
dev_err(&op->dev, "Cannot map MAC TX registers.\n");
drivers/net/ethernet/sun/sunhme.c
2547
hp->erxregs = devm_platform_ioremap_resource(op, 2);
drivers/net/ethernet/sun/sunhme.c
2549
dev_err(&op->dev, "Cannot map MAC RX registers.\n");
drivers/net/ethernet/sun/sunhme.c
2555
if ((op->resource[3].end - op->resource[3].start) >= BMAC_REG_SIZE)
drivers/net/ethernet/sun/sunhme.c
2556
op->resource[3].end = op->resource[3].start + BMAC_REG_SIZE - 1;
drivers/net/ethernet/sun/sunhme.c
2557
hp->bigmacregs = devm_platform_ioremap_resource(op, 3);
drivers/net/ethernet/sun/sunhme.c
2559
dev_err(&op->dev, "Cannot map BIGMAC registers.\n");
drivers/net/ethernet/sun/sunhme.c
2564
hp->tcvregs = devm_platform_ioremap_resource(op, 4);
drivers/net/ethernet/sun/sunhme.c
2566
dev_err(&op->dev, "Cannot map TCVR registers.\n");
drivers/net/ethernet/sun/sunhme.c
2576
hp->irq = op->archdata.irqs[0];
drivers/net/ethernet/sun/sunhme.c
2595
platform_set_drvdata(op, hp);
drivers/net/ethernet/sun/sunhme.c
2806
static int hme_sbus_probe(struct platform_device *op)
drivers/net/ethernet/sun/sunhme.c
2809
struct device_node *dp = op->dev.of_node;
drivers/net/ethernet/sun/sunhme.c
2813
match = of_match_device(hme_sbus_match, &op->dev);
drivers/net/ethernet/sun/sunhme.c
2821
return happy_meal_sbus_probe_one(op, is_qfe);
drivers/net/ethernet/sun/sunqe.c
685
struct platform_device *op;
drivers/net/ethernet/sun/sunqe.c
690
op = qep->op;
drivers/net/ethernet/sun/sunqe.c
691
regs = of_get_property(op->dev.of_node, "reg", NULL);
drivers/net/ethernet/sun/sunqe.c
717
static void qec_init_once(struct sunqec *qecp, struct platform_device *op)
drivers/net/ethernet/sun/sunqe.c
735
sbus_writel((resource_size(&op->resource[1]) >> 2),
drivers/net/ethernet/sun/sunqe.c
741
sbus_writel((resource_size(&op->resource[1]) >> 2) >> 1,
drivers/net/ethernet/sun/sunqe.c
743
sbus_writel((resource_size(&op->resource[1]) >> 2) >> 1,
drivers/net/ethernet/sun/sunqe.c
769
struct platform_device *op = to_platform_device(child->dev.parent);
drivers/net/ethernet/sun/sunqe.c
772
qecp = platform_get_drvdata(op);
drivers/net/ethernet/sun/sunqe.c
778
qecp->op = op;
drivers/net/ethernet/sun/sunqe.c
779
qecp->gregs = of_ioremap(&op->resource[0], 0,
drivers/net/ethernet/sun/sunqe.c
796
qecp->qec_bursts = qec_get_burst(op->dev.of_node);
drivers/net/ethernet/sun/sunqe.c
798
qec_init_once(qecp, op);
drivers/net/ethernet/sun/sunqe.c
800
if (request_irq(op->archdata.irqs[0], qec_interrupt,
drivers/net/ethernet/sun/sunqe.c
806
platform_set_drvdata(op, qecp);
drivers/net/ethernet/sun/sunqe.c
817
of_iounmap(&op->resource[0], qecp->gregs, GLOB_REG_SIZE);
drivers/net/ethernet/sun/sunqe.c
832
static int qec_ether_init(struct platform_device *op)
drivers/net/ethernet/sun/sunqe.c
853
i = of_getintprop_default(op->dev.of_node, "channel#", -1);
drivers/net/ethernet/sun/sunqe.c
859
qecp = get_qec(op);
drivers/net/ethernet/sun/sunqe.c
866
qe->op = op;
drivers/net/ethernet/sun/sunqe.c
869
qe->qcregs = of_ioremap(&op->resource[0], 0,
drivers/net/ethernet/sun/sunqe.c
876
qe->mregs = of_ioremap(&op->resource[1], 0,
drivers/net/ethernet/sun/sunqe.c
883
qe->qe_block = dma_alloc_coherent(&op->dev, PAGE_SIZE,
drivers/net/ethernet/sun/sunqe.c
885
qe->buffers = dma_alloc_coherent(&op->dev, sizeof(struct sunqe_buffers),
drivers/net/ethernet/sun/sunqe.c
894
SET_NETDEV_DEV(dev, &op->dev);
drivers/net/ethernet/sun/sunqe.c
897
dev->irq = op->archdata.irqs[0];
drivers/net/ethernet/sun/sunqe.c
906
platform_set_drvdata(op, qe);
drivers/net/ethernet/sun/sunqe.c
914
of_iounmap(&op->resource[0], qe->qcregs, CREG_REG_SIZE);
drivers/net/ethernet/sun/sunqe.c
916
of_iounmap(&op->resource[1], qe->mregs, MREGS_REG_SIZE);
drivers/net/ethernet/sun/sunqe.c
918
dma_free_coherent(&op->dev, PAGE_SIZE,
drivers/net/ethernet/sun/sunqe.c
921
dma_free_coherent(&op->dev,
drivers/net/ethernet/sun/sunqe.c
931
static int qec_sbus_probe(struct platform_device *op)
drivers/net/ethernet/sun/sunqe.c
933
return qec_ether_init(op);
drivers/net/ethernet/sun/sunqe.c
936
static void qec_sbus_remove(struct platform_device *op)
drivers/net/ethernet/sun/sunqe.c
938
struct sunqe *qp = platform_get_drvdata(op);
drivers/net/ethernet/sun/sunqe.c
943
of_iounmap(&op->resource[0], qp->qcregs, CREG_REG_SIZE);
drivers/net/ethernet/sun/sunqe.c
944
of_iounmap(&op->resource[1], qp->mregs, MREGS_REG_SIZE);
drivers/net/ethernet/sun/sunqe.c
945
dma_free_coherent(&op->dev, PAGE_SIZE,
drivers/net/ethernet/sun/sunqe.c
947
dma_free_coherent(&op->dev, sizeof(struct sunqe_buffers),
drivers/net/ethernet/sun/sunqe.c
982
struct platform_device *op = root_qec_dev->op;
drivers/net/ethernet/sun/sunqe.c
984
free_irq(op->archdata.irqs[0], (void *) root_qec_dev);
drivers/net/ethernet/sun/sunqe.c
985
of_iounmap(&op->resource[0], root_qec_dev->gregs,
drivers/net/ethernet/sun/sunqe.h
318
struct platform_device *op; /* QEC's OF device */
drivers/net/ethernet/sun/sunqe.h
346
struct platform_device *op; /* QE's OF device struct */
drivers/net/ethernet/xilinx/ll_temac_main.c
261
static int temac_dcr_setup(struct temac_local *lp, struct platform_device *op,
drivers/net/ethernet/xilinx/ll_temac_main.c
273
dev_dbg(&op->dev, "DCR base: %x\n", dcrs);
drivers/net/ethernet/xilinx/ll_temac_main.c
286
static int temac_dcr_setup(struct temac_local *lp, struct platform_device *op,
drivers/net/hyperv/hyperv_net.h
652
u32 op;
drivers/net/hyperv/hyperv_net.h
699
u32 op;
drivers/net/hyperv/hyperv_net.h
773
u32 op;
drivers/net/hyperv/rndis_filter.c
1299
init_packet->msg.v5_msg.subchn_req.op = NVSP_SUBCHANNEL_ALLOCATE;
drivers/net/mdio/mdio-aspeed.c
48
static int aspeed_mdio_op(struct mii_bus *bus, u8 st, u8 op, u8 phyad, u8 regad,
drivers/net/mdio/mdio-aspeed.c
55
__func__, st, op, phyad, regad, data);
drivers/net/mdio/mdio-aspeed.c
59
| FIELD_PREP(ASPEED_MDIO_CTRL_OP, op)
drivers/net/mdio/mdio-bitbang.c
117
if (op & MDIO_C45)
drivers/net/mdio/mdio-bitbang.c
121
mdiobb_send_bit(ctrl, (op >> 1) & 1);
drivers/net/mdio/mdio-bitbang.c
122
mdiobb_send_bit(ctrl, (op >> 0) & 1);
drivers/net/mdio/mdio-bitbang.c
94
static void mdiobb_cmd(struct mdiobb_ctrl *ctrl, int op, u8 phy, u8 reg)
drivers/net/mdio/mdio-mscc-miim.c
66
#define mscc_readx_poll_timeout(op, addr, val, cond, delay_us, timeout_us)\
drivers/net/mdio/mdio-mscc-miim.c
69
readx_poll_timeout_atomic(op, addr, val, cond, delay_us, \
drivers/net/mdio/mdio-mscc-miim.c
71
readx_poll_timeout(op, addr, val, cond, delay_us, timeout_us); \
drivers/net/mdio/mdio-mux-bcm-iproc.c
104
u16 phyid, u32 reg, u16 val, u32 op)
drivers/net/mdio/mdio-mux-bcm-iproc.c
124
writel(op, base + MDIO_CTRL_OFFSET);
drivers/net/mdio/mdio-mux-bcm-iproc.c
130
if (op == MDIO_CTRL_READ_OP)
drivers/net/ovpn/crypto_aead.c
125
op = ovpn_opcode_compose(OVPN_DATA_V2, ks->key_id, peer->id);
drivers/net/ovpn/crypto_aead.c
127
BUILD_BUG_ON(sizeof(op) != OVPN_OPCODE_SIZE);
drivers/net/ovpn/crypto_aead.c
128
*((__force __be32 *)skb->data) = htonl(op);
drivers/net/ovpn/crypto_aead.c
47
u32 pktid, op;
drivers/net/phy/mdio_bus.c
382
static void mdiobus_stats_acct(struct mdio_bus_stats *stats, bool op, int ret)
drivers/net/phy/mdio_bus.c
393
if (op)
drivers/net/phy/mscc/mscc_main.c
1885
u32 op)
drivers/net/phy/mscc/mscc_main.c
1892
op | (1 << mcb));
drivers/net/phy/mscc/mscc_main.c
1904
} while (time_before(jiffies, deadline) && (val & op));
drivers/net/phy/mscc/mscc_main.c
1906
if (val & op)
drivers/net/usb/smsc95xx.c
166
static u32 mii_address_cmd(int phy_id, int idx, u16 op)
drivers/net/usb/smsc95xx.c
168
return (phy_id & 0x1f) << 11 | (idx & 0x1f) << 6 | op;
drivers/net/wireless/ath/ar5523/ar5523.c
440
static int ar5523_set_rxfilter(struct ar5523 *ar, u32 bits, u32 op)
drivers/net/wireless/ath/ar5523/ar5523.c
445
rxfilter.op = cpu_to_be32(op);
drivers/net/wireless/ath/ar5523/ar5523.c
447
ar5523_dbg(ar, "setting Rx filter=0x%x flags=0x%x\n", bits, op);
drivers/net/wireless/ath/ar5523/ar5523_hw.h
301
__be32 op;
drivers/net/wireless/ath/wil6210/pmc.c
171
pmc_cmd.op = WMI_PMC_ALLOCATE;
drivers/net/wireless/ath/wil6210/pmc.c
245
pmc_cmd.op = WMI_PMC_RELEASE;
drivers/net/wireless/ath/wil6210/wmi.h
1450
u8 op;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
102
result = op;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
123
s16 qm_shl16(s16 op, int shift)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
127
result = op;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
147
s16 qm_shr16(s16 op, int shift)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
149
return qm_shl16(op, -shift);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
156
s16 qm_norm32(s32 op)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
159
if (op == 0) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
163
while ((op >> 31) == (op >> 30)) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
165
op = op << 1;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.c
98
s32 qm_shl32(s32 op, int shift)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.h
21
s32 qm_shl32(s32 op, int shift);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.h
23
s16 qm_shl16(s16 op, int shift);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.h
25
s16 qm_shr16(s16 op, int shift);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_qmath.h
27
s16 qm_norm32(s32 op);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
2601
static void eeprom_op(struct ipw_priv *priv, u8 op, u8 addr)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
2607
eeprom_write_bit(priv, op & 2);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
2608
eeprom_write_bit(priv, op & 1);
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
833
} op = NONE;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
843
op = ADD;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
846
op = REMOVE;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
849
op = HT_RATE_INIT;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
853
op = ADD_RATE_INIT;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
856
op = REMOVE;
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
859
switch (op) {
drivers/net/wireless/intel/iwlwifi/fw/api/debug.h
326
__le32 op;
drivers/net/wireless/intel/iwlwifi/fw/file.h
645
u8 op;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1534
_iwl_op_mode_start(struct iwl_drv *drv, struct iwlwifi_opmode_table *op)
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1536
const struct iwl_op_mode_ops *ops = op->ops;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1547
drv->dbgfs_op_mode = debugfs_create_dir(op->name,
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1604
struct iwlwifi_opmode_table *op;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1849
op = &iwlwifi_opmode_table[DVM_OP_MODE];
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1855
op = &iwlwifi_opmode_table[MVM_OP_MODE];
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1861
op = &iwlwifi_opmode_table[MLD_OP_MODE];
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1865
drv->fw.fw_version, op->name);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1868
list_add_tail(&drv->list, &op->drv);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1870
if (op->ops) {
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1871
drv->op_mode = _iwl_op_mode_start(drv, op);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
1878
request_module_nowait("%s", op->name);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2020
struct iwlwifi_opmode_table *op;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2024
op = &iwlwifi_opmode_table[i];
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2025
if (strcmp(op->name, name))
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2027
op->ops = ops;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2029
list_for_each_entry(drv, &op->drv, list)
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
2030
drv->op_mode = _iwl_op_mode_start(drv, op);
drivers/net/wireless/intel/iwlwifi/mld/notif.c
531
iwl_mld_log_async_handler_op(struct iwl_mld *mld, const char *op,
drivers/net/wireless/intel/iwlwifi/mld/notif.c
538
op, iwl_get_cmd_string(mld->trans,
drivers/net/wireless/intel/iwlwifi/mld/thermal.c
279
enum iwl_ctdp_cmd_operation op)
drivers/net/wireless/intel/iwlwifi/mld/thermal.c
282
.operation = cpu_to_le32(op),
drivers/net/wireless/intel/iwlwifi/mld/thermal.c
307
if (op == CTDP_CMD_OPERATION_START)
drivers/net/wireless/intel/iwlwifi/mld/thermal.h
26
enum iwl_ctdp_cmd_operation op);
drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c
1974
cmd.op = cpu_to_le32(DEBUG_MEM_OP_READ);
drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c
2026
u32 op, len;
drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c
2035
op = DEBUG_MEM_OP_WRITE_BYTES;
drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c
2039
op = DEBUG_MEM_OP_WRITE;
drivers/net/wireless/intel/iwlwifi/mvm/debugfs.c
2049
cmd->op = cpu_to_le32(op);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
2304
int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 budget);
drivers/net/wireless/intel/iwlwifi/mvm/tt.c
485
int iwl_mvm_ctdp_command(struct iwl_mvm *mvm, u32 op, u32 state)
drivers/net/wireless/intel/iwlwifi/mvm/tt.c
488
.operation = cpu_to_le32(op),
drivers/net/wireless/intel/iwlwifi/mvm/tt.c
517
switch (op) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
886
switch (dest->reg_ops[i].op) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
915
dest->reg_ops[i].op);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
25
u32 op = get ? PATCH_SEM_GET : PATCH_SEM_RELEASE;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
27
__le32 op;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
29
.op = cpu_to_le32(op),
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
504
.op.op_mode = cpu_to_le32(oper),
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
527
.op.rf.func_idx = cpu_to_le32(func_idx),
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
529
struct tm_tx_cont *tx_cont = &req.op.rf.param.tx_cont;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
566
req.op.rf.param.func_data = cpu_to_le32(band);
drivers/net/wireless/mediatek/mt76/mt7915/testmode.h
88
} op;
drivers/net/wireless/realtek/rtw89/fw.c
6521
struct rtw89_chan *op = &rtwdev->scan_info.op_chan;
drivers/net/wireless/realtek/rtw89/fw.c
6564
h2c->w1 |= le32_encode_bits(op->band_width,
drivers/net/wireless/realtek/rtw89/fw.c
6566
le32_encode_bits(op->primary_channel,
drivers/net/wireless/realtek/rtw89/fw.c
6568
le32_encode_bits(op->channel,
drivers/net/wireless/realtek/rtw89/fw.c
6570
h2c->w0 |= le32_encode_bits(op->band_type,
drivers/net/wireless/realtek/rtw89/fw.c
6636
struct rtw89_chan *op = &scan_info->op_chan;
drivers/net/wireless/realtek/rtw89/fw.c
6772
scan_op[0].chan = *op;
drivers/net/wireless/realtek/rtw89/fw.c
8324
struct rtw89_chan *op = &rtwdev->scan_info.op_chan;
drivers/net/wireless/realtek/rtw89/fw.c
8375
ch_info->central_ch = op->channel;
drivers/net/wireless/realtek/rtw89/fw.c
8376
ch_info->pri_ch = op->primary_channel;
drivers/net/wireless/realtek/rtw89/fw.c
8377
ch_info->ch_band = op->band_type;
drivers/net/wireless/realtek/rtw89/fw.c
8378
ch_info->bw = op->band_width;
drivers/net/wireless/realtek/rtw89/mac.c
5156
const struct rtw89_chan *op = &rtwdev->scan_info.op_chan;
drivers/net/wireless/realtek/rtw89/mac.c
5158
if (band == op->band_type && channel == op->primary_channel)
drivers/net/wireless/realtek/rtw89/mac.c
5159
return op;
drivers/net/wireless/realtek/rtw89/mac.c
5162
op = &scan_info->extra_op.chan;
drivers/net/wireless/realtek/rtw89/mac.c
5163
if (band == op->band_type && channel == op->primary_channel)
drivers/net/wireless/realtek/rtw89/mac.c
5164
return op;
drivers/net/xen-netback/common.h
129
struct gnttab_copy op[COPY_BATCH_SIZE];
drivers/net/xen-netback/rx.c
164
gnttab_batch_copy(queue->rx_copy.op, queue->rx_copy.num);
drivers/net/xen-netback/rx.c
167
struct gnttab_copy *op;
drivers/net/xen-netback/rx.c
169
op = &queue->rx_copy.op[i];
drivers/net/xen-netback/rx.c
174
if (unlikely(op->status != GNTST_okay)) {
drivers/net/xen-netback/rx.c
179
rsp->status = op->status;
drivers/net/xen-netback/rx.c
197
struct gnttab_copy *op;
drivers/net/xen-netback/rx.c
204
op = &queue->rx_copy.op[queue->rx_copy.num];
drivers/net/xen-netback/rx.c
208
op->flags = GNTCOPY_dest_gref;
drivers/net/xen-netback/rx.c
212
op->source.domid = foreign->domid;
drivers/net/xen-netback/rx.c
213
op->source.u.ref = foreign->gref;
drivers/net/xen-netback/rx.c
214
op->flags |= GNTCOPY_source_gref;
drivers/net/xen-netback/rx.c
216
op->source.u.gmfn = virt_to_gfn(data);
drivers/net/xen-netback/rx.c
217
op->source.domid = DOMID_SELF;
drivers/net/xen-netback/rx.c
220
op->source.offset = xen_offset_in_page(data);
drivers/net/xen-netback/rx.c
221
op->dest.u.ref = req->gref;
drivers/net/xen-netback/rx.c
222
op->dest.domid = queue->vif->domid;
drivers/net/xen-netback/rx.c
223
op->dest.offset = offset;
drivers/net/xen-netback/rx.c
224
op->len = len;
drivers/ntb/hw/mscc/ntb_hw_switchtec.c
100
u32 op, int wait_status)
drivers/ntb/hw/mscc/ntb_hw_switchtec.c
112
switch (op) {
drivers/ntb/hw/mscc/ntb_hw_switchtec.c
126
iowrite32(op, &ctl->partition_op);
drivers/ntb/hw/mscc/ntb_hw_switchtec.c
146
op_text[op], op,
drivers/nvdimm/btt.c
1419
enum req_op op, sector_t sector)
drivers/nvdimm/btt.c
1423
if (!op_is_write(op)) {
drivers/nvme/host/core.c
1005
if (op == nvme_cmd_write && ns->head->nr_plids) {
drivers/nvme/host/core.c
1020
cmnd->rw.opcode = op;
drivers/nvme/host/core.c
1052
if (op == nvme_cmd_zone_append)
drivers/nvme/host/core.c
1684
static int nvme_features(struct nvme_ctrl *dev, u8 op, unsigned int fid,
drivers/nvme/host/core.c
1691
c.features.opcode = op;
drivers/nvme/host/core.c
2620
u32 timeout, const char *op)
drivers/nvme/host/core.c
2638
op, csts);
drivers/nvme/host/core.c
992
enum nvme_opcode op)
drivers/nvme/host/fc.c
113
struct nvme_fc_fcp_op op;
drivers/nvme/host/fc.c
1791
struct nvme_fc_fcp_op *op)
drivers/nvme/host/fc.c
1793
fc_dma_unmap_single(ctrl->lport->dev, op->fcp_req.rspdma,
drivers/nvme/host/fc.c
1794
sizeof(op->rsp_iu), DMA_FROM_DEVICE);
drivers/nvme/host/fc.c
1795
fc_dma_unmap_single(ctrl->lport->dev, op->fcp_req.cmddma,
drivers/nvme/host/fc.c
1796
sizeof(op->cmd_iu), DMA_TO_DEVICE);
drivers/nvme/host/fc.c
1798
atomic_set(&op->state, FCPOP_STATE_UNINIT);
drivers/nvme/host/fc.c
1805
struct nvme_fc_fcp_op *op = blk_mq_rq_to_pdu(rq);
drivers/nvme/host/fc.c
1807
return __nvme_fc_exit_request(to_fc_ctrl(set->driver_data), op);
drivers/nvme/host/fc.c
1811
__nvme_fc_abort_op(struct nvme_fc_ctrl *ctrl, struct nvme_fc_fcp_op *op)
drivers/nvme/host/fc.c
1817
opstate = atomic_xchg(&op->state, FCPOP_STATE_ABORTED);
drivers/nvme/host/fc.c
1819
atomic_set(&op->state, opstate);
drivers/nvme/host/fc.c
1821
op->flags |= FCOP_FLAGS_TERMIO;
drivers/nvme/host/fc.c
1831
op->queue->lldd_handle,
drivers/nvme/host/fc.c
1832
&op->fcp_req);
drivers/nvme/host/fc.c
1853
struct nvme_fc_fcp_op *op, int opstate)
drivers/nvme/host/fc.c
1860
op->flags & FCOP_FLAGS_TERMIO) {
drivers/nvme/host/fc.c
1886
struct nvme_fc_fcp_op *op = fcp_req_to_fcp_op(req);
drivers/nvme/host/fc.c
1887
struct request *rq = op->rq;
drivers/nvme/host/fc.c
1898
struct nvme_fc_fcp_op *op = fcp_req_to_fcp_op(req);
drivers/nvme/host/fc.c
1899
struct request *rq = op->rq;
drivers/nvme/host/fc.c
1900
struct nvmefc_fcp_req *freq = &op->fcp_req;
drivers/nvme/host/fc.c
1901
struct nvme_fc_ctrl *ctrl = op->ctrl;
drivers/nvme/host/fc.c
1902
struct nvme_fc_queue *queue = op->queue;
drivers/nvme/host/fc.c
1903
struct nvme_completion *cqe = &op->rsp_iu.cqe;
drivers/nvme/host/fc.c
1904
struct nvme_command *sqe = &op->cmd_iu.sqe;
drivers/nvme/host/fc.c
1947
opstate = atomic_xchg(&op->state, FCPOP_STATE_COMPLETE);
drivers/nvme/host/fc.c
1949
fc_dma_sync_single_for_cpu(ctrl->lport->dev, op->fcp_req.rspdma,
drivers/nvme/host/fc.c
1950
sizeof(op->rsp_iu), DMA_FROM_DEVICE);
drivers/nvme/host/fc.c
1986
be32_to_cpu(op->cmd_iu.data_len)) {
drivers/nvme/host/fc.c
1992
be32_to_cpu(op->cmd_iu.data_len));
drivers/nvme/host/fc.c
2003
if (unlikely(be16_to_cpu(op->rsp_iu.iu_len) !=
drivers/nvme/host/fc.c
2005
be32_to_cpu(op->rsp_iu.xfrd_len) !=
drivers/nvme/host/fc.c
2007
op->rsp_iu.ersp_result ||
drivers/nvme/host/fc.c
2014
ctrl->cnum, be16_to_cpu(op->rsp_iu.iu_len),
drivers/nvme/host/fc.c
2015
be32_to_cpu(op->rsp_iu.xfrd_len),
drivers/nvme/host/fc.c
2017
op->rsp_iu.ersp_result,
drivers/nvme/host/fc.c
2038
if (op->flags & FCOP_FLAGS_AEN) {
drivers/nvme/host/fc.c
2040
__nvme_fc_fcpop_chk_teardowns(ctrl, op, opstate);
drivers/nvme/host/fc.c
2041
atomic_set(&op->state, FCPOP_STATE_IDLE);
drivers/nvme/host/fc.c
2042
op->flags = FCOP_FLAGS_AEN; /* clear other flags */
drivers/nvme/host/fc.c
2047
__nvme_fc_fcpop_chk_teardowns(ctrl, op, opstate);
drivers/nvme/host/fc.c
2059
struct nvme_fc_queue *queue, struct nvme_fc_fcp_op *op,
drivers/nvme/host/fc.c
2063
container_of(op, typeof(*op_w_sgl), op);
drivers/nvme/host/fc.c
2064
struct nvme_fc_cmd_iu *cmdiu = &op->cmd_iu;
drivers/nvme/host/fc.c
2067
memset(op, 0, sizeof(*op));
drivers/nvme/host/fc.c
2068
op->fcp_req.cmdaddr = &op->cmd_iu;
drivers/nvme/host/fc.c
2069
op->fcp_req.cmdlen = sizeof(op->cmd_iu);
drivers/nvme/host/fc.c
2070
op->fcp_req.rspaddr = &op->rsp_iu;
drivers/nvme/host/fc.c
2071
op->fcp_req.rsplen = sizeof(op->rsp_iu);
drivers/nvme/host/fc.c
2072
op->fcp_req.done = nvme_fc_fcpio_done;
drivers/nvme/host/fc.c
2073
op->ctrl = ctrl;
drivers/nvme/host/fc.c
2074
op->queue = queue;
drivers/nvme/host/fc.c
2075
op->rq = rq;
drivers/nvme/host/fc.c
2076
op->rqno = rqno;
drivers/nvme/host/fc.c
2087
op->fcp_req.cmddma = fc_dma_map_single(ctrl->lport->dev,
drivers/nvme/host/fc.c
2088
&op->cmd_iu, sizeof(op->cmd_iu), DMA_TO_DEVICE);
drivers/nvme/host/fc.c
2089
if (fc_dma_mapping_error(ctrl->lport->dev, op->fcp_req.cmddma)) {
drivers/nvme/host/fc.c
2096
op->fcp_req.rspdma = fc_dma_map_single(ctrl->lport->dev,
drivers/nvme/host/fc.c
2097
&op->rsp_iu, sizeof(op->rsp_iu),
drivers/nvme/host/fc.c
2099
if (fc_dma_mapping_error(ctrl->lport->dev, op->fcp_req.rspdma)) {
drivers/nvme/host/fc.c
2105
atomic_set(&op->state, FCPOP_STATE_IDLE);
drivers/nvme/host/fc.c
2115
struct nvme_fcp_op_w_sgl *op = blk_mq_rq_to_pdu(rq);
drivers/nvme/host/fc.c
2120
res = __nvme_fc_init_request(ctrl, queue, &op->op, rq, queue->rqcnt++);
drivers/nvme/host/fc.c
2123
op->op.fcp_req.first_sgl = op->sgl;
drivers/nvme/host/fc.c
2124
op->op.fcp_req.private = &op->priv[0];
drivers/nvme/host/fc.c
2126
nvme_req(rq)->cmd = &op->op.cmd_iu.sqe;
drivers/nvme/host/fc.c
2421
struct nvme_fc_fcp_op *op = blk_mq_rq_to_pdu(req);
drivers/nvme/host/fc.c
2423
op->nreq.flags |= NVME_REQ_CANCELLED;
drivers/nvme/host/fc.c
2424
__nvme_fc_abort_op(ctrl, op);
drivers/nvme/host/fc.c
2533
struct nvme_fc_fcp_op *op = blk_mq_rq_to_pdu(rq);
drivers/nvme/host/fc.c
2534
struct nvme_fc_ctrl *ctrl = op->ctrl;
drivers/nvme/host/fc.c
2535
u16 qnum = op->queue->qnum;
drivers/nvme/host/fc.c
2536
struct nvme_fc_cmd_iu *cmdiu = &op->cmd_iu;
drivers/nvme/host/fc.c
2549
if (__nvme_fc_abort_op(ctrl, op))
drivers/nvme/host/fc.c
2562
struct nvme_fc_fcp_op *op)
drivers/nvme/host/fc.c
2564
struct nvmefc_fcp_req *freq = &op->fcp_req;
drivers/nvme/host/fc.c
2579
op->nents = blk_rq_map_sg(rq, freq->sg_table.sgl);
drivers/nvme/host/fc.c
2580
WARN_ON(op->nents > blk_rq_nr_phys_segments(rq));
drivers/nvme/host/fc.c
2582
op->nents, rq_dma_dir(rq));
drivers/nvme/host/fc.c
2597
struct nvme_fc_fcp_op *op)
drivers/nvme/host/fc.c
2599
struct nvmefc_fcp_req *freq = &op->fcp_req;
drivers/nvme/host/fc.c
2604
fc_dma_unmap_sg(ctrl->lport->dev, freq->sg_table.sgl, op->nents,
drivers/nvme/host/fc.c
2637
struct nvme_fc_fcp_op *op, u32 data_len,
drivers/nvme/host/fc.c
2640
struct nvme_fc_cmd_iu *cmdiu = &op->cmd_iu;
drivers/nvme/host/fc.c
2668
op->fcp_req.payload_length = data_len;
drivers/nvme/host/fc.c
2669
op->fcp_req.io_dir = io_dir;
drivers/nvme/host/fc.c
2670
op->fcp_req.transferred_length = 0;
drivers/nvme/host/fc.c
2671
op->fcp_req.rcv_rsplen = 0;
drivers/nvme/host/fc.c
2672
op->fcp_req.status = NVME_SC_SUCCESS;
drivers/nvme/host/fc.c
2673
op->fcp_req.sqid = cpu_to_le16(queue->qnum);
drivers/nvme/host/fc.c
2694
if (!(op->flags & FCOP_FLAGS_AEN)) {
drivers/nvme/host/fc.c
2695
ret = nvme_fc_map_data(ctrl, op->rq, op);
drivers/nvme/host/fc.c
2697
nvme_cleanup_cmd(op->rq);
drivers/nvme/host/fc.c
2705
fc_dma_sync_single_for_device(ctrl->lport->dev, op->fcp_req.cmddma,
drivers/nvme/host/fc.c
2706
sizeof(op->cmd_iu), DMA_TO_DEVICE);
drivers/nvme/host/fc.c
2708
atomic_set(&op->state, FCPOP_STATE_ACTIVE);
drivers/nvme/host/fc.c
2710
if (!(op->flags & FCOP_FLAGS_AEN))
drivers/nvme/host/fc.c
2711
nvme_start_request(op->rq);
drivers/nvme/host/fc.c
2716
queue->lldd_handle, &op->fcp_req);
drivers/nvme/host/fc.c
2731
opstate = atomic_xchg(&op->state, FCPOP_STATE_COMPLETE);
drivers/nvme/host/fc.c
2732
__nvme_fc_fcpop_chk_teardowns(ctrl, op, opstate);
drivers/nvme/host/fc.c
2734
if (!(op->flags & FCOP_FLAGS_AEN)) {
drivers/nvme/host/fc.c
2735
nvme_fc_unmap_data(ctrl, op->rq, op);
drivers/nvme/host/fc.c
2736
nvme_cleanup_cmd(op->rq);
drivers/nvme/host/fc.c
2759
struct nvme_fc_fcp_op *op = blk_mq_rq_to_pdu(rq);
drivers/nvme/host/fc.c
2791
return nvme_fc_start_fcp_op(ctrl, queue, op, data_len, io_dir);
drivers/nvme/host/fc.c
2816
struct nvme_fc_fcp_op *op = blk_mq_rq_to_pdu(rq);
drivers/nvme/host/fc.c
2817
struct nvme_fc_ctrl *ctrl = op->ctrl;
drivers/nvme/host/fc.c
2819
atomic_set(&op->state, FCPOP_STATE_IDLE);
drivers/nvme/host/fc.c
2820
op->flags &= ~FCOP_FLAGS_TERMIO;
drivers/nvme/host/fc.c
2822
nvme_fc_unmap_data(ctrl, rq, op);
drivers/nvme/host/pr.c
100
c.common.opcode = op;
drivers/nvme/host/pr.c
111
u8 op, void *data, unsigned int data_len)
drivers/nvme/host/pr.c
115
ret = __nvme_send_pr_command(bdev, cdw10, cdw11, op, data, data_len);
drivers/nvme/host/pr.c
96
u32 cdw11, u8 op, void *data, unsigned int data_len)
drivers/nvme/host/rdma.c
1168
const char *op)
drivers/nvme/host/rdma.c
1176
op, wc->wr_cqe,
drivers/nvme/target/fc.c
2158
if (fod->fcpreq->op == NVMET_FCOP_READDATA_RSP)
drivers/nvme/target/fc.c
2238
fod->fcpreq->op = NVMET_FCOP_RSP;
drivers/nvme/target/fc.c
2250
struct nvmet_fc_fcp_iod *fod, u8 op)
drivers/nvme/target/fc.c
2259
fcpreq->op = op;
drivers/nvme/target/fc.c
2300
if ((op == NVMET_FCOP_READDATA) &&
drivers/nvme/target/fc.c
2303
fcpreq->op = NVMET_FCOP_READDATA_RSP;
drivers/nvme/target/fc.c
2316
if (op == NVMET_FCOP_WRITEDATA) {
drivers/nvme/target/fc.c
2337
if (fcpreq->op == NVMET_FCOP_WRITEDATA) {
drivers/nvme/target/fc.c
2365
switch (fcpreq->op) {
drivers/nvme/target/fc.c
2408
if (fcpreq->op == NVMET_FCOP_READDATA_RSP) {
drivers/nvme/target/fcloop.c
799
fcloop_fcp_copy_data(u8 op, struct scatterlist *data_sg,
drivers/nvme/target/fcloop.c
827
if (op == NVMET_FCOP_WRITEDATA)
drivers/nvme/target/fcloop.c
860
u8 op = tgt_fcpreq->op;
drivers/nvme/target/fcloop.c
891
switch (op) {
drivers/nvme/target/fcloop.c
895
fcloop_fcp_copy_data(op, tgt_fcpreq->sg,
drivers/nvme/target/fcloop.c
906
fcloop_fcp_copy_data(op, tgt_fcpreq->sg,
drivers/nvme/target/fcloop.c
911
if (op == NVMET_FCOP_READDATA)
drivers/nvme/target/zns.c
476
enum req_op op = zsa_req_op(req->cmd->zms.zsa);
drivers/nvme/target/zns.c
482
if (op == REQ_OP_LAST) {
drivers/nvme/target/zns.c
506
ret = blkdev_zone_mgmt(bdev, op, sect, zone_sectors);
drivers/nvmem/stm32-romem.c
56
static int stm32_bsec_smc(u8 op, u32 otp, u32 data, u32 *result)
drivers/nvmem/stm32-romem.c
61
arm_smccc_smc(STM32_SMC_BSEC, op, otp, data, 0, 0, 0, 0, &res);
drivers/parport/parport_sunbpp.c
269
static int bpp_probe(struct platform_device *op)
drivers/parport/parport_sunbpp.c
278
irq = op->archdata.irqs[0];
drivers/parport/parport_sunbpp.c
279
base = of_ioremap(&op->resource[0], 0,
drivers/parport/parport_sunbpp.c
280
resource_size(&op->resource[0]),
drivers/parport/parport_sunbpp.c
285
size = resource_size(&op->resource[0]);
drivers/parport/parport_sunbpp.c
302
p->dev = &op->dev;
drivers/parport/parport_sunbpp.c
319
dev_set_drvdata(&op->dev, p);
drivers/parport/parport_sunbpp.c
332
of_iounmap(&op->resource[0], base, size);
drivers/parport/parport_sunbpp.c
337
static void bpp_remove(struct platform_device *op)
drivers/parport/parport_sunbpp.c
339
struct parport *p = dev_get_drvdata(&op->dev);
drivers/parport/parport_sunbpp.c
349
of_iounmap(&op->resource[0], (void __iomem *) p->base, p->size);
drivers/parport/parport_sunbpp.c
353
dev_set_drvdata(&op->dev, NULL);
drivers/pci/endpoint/functions/pci-epf-test.c
333
const char *op, u64 size,
drivers/pci/endpoint/functions/pci-epf-test.c
347
op, size, dma ? "YES" : "NO", &ts, rate);
drivers/pci/hotplug/cpqphp_nvram.c
157
int op = operation;
drivers/pci/hotplug/cpqphp_nvram.c
172
: "a" (op), "c" (*buf_size), "S" (ev_name),
drivers/pci/xen-pcifront.c
111
static int do_pci_op(struct pcifront_device *pdev, struct xen_pci_op *op)
drivers/pci/xen-pcifront.c
114
struct xen_pci_op *active_op = &pdev->sh_info->op;
drivers/pci/xen-pcifront.c
122
memcpy(active_op, op, sizeof(struct xen_pci_op));
drivers/pci/xen-pcifront.c
166
memcpy(op, active_op, sizeof(struct xen_pci_op));
drivers/pci/xen-pcifront.c
168
err = op->err;
drivers/pci/xen-pcifront.c
179
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
195
err = do_pci_op(pdev, &op);
drivers/pci/xen-pcifront.c
199
op.value);
drivers/pci/xen-pcifront.c
201
*val = op.value;
drivers/pci/xen-pcifront.c
215
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
232
return errno_to_pcibios_err(do_pci_op(pdev, &op));
drivers/pci/xen-pcifront.c
246
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
265
op.msix_entries[i].entry = entry->msi_index;
drivers/pci/xen-pcifront.c
267
op.msix_entries[i].vector = -1;
drivers/pci/xen-pcifront.c
271
err = do_pci_op(pdev, &op);
drivers/pci/xen-pcifront.c
274
if (likely(!op.value)) {
drivers/pci/xen-pcifront.c
277
if (op.msix_entries[i].vector <= 0) {
drivers/pci/xen-pcifront.c
279
i, op.msix_entries[i].vector);
drivers/pci/xen-pcifront.c
284
vector[i] = op.msix_entries[i].vector;
drivers/pci/xen-pcifront.c
287
pr_info("enable msix get value %x\n", op.value);
drivers/pci/xen-pcifront.c
288
err = op.value;
drivers/pci/xen-pcifront.c
299
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
308
err = do_pci_op(pdev, &op);
drivers/pci/xen-pcifront.c
318
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
327
err = do_pci_op(pdev, &op);
drivers/pci/xen-pcifront.c
329
vector[0] = op.value;
drivers/pci/xen-pcifront.c
330
if (op.value <= 0) {
drivers/pci/xen-pcifront.c
332
op.value);
drivers/pci/xen-pcifront.c
338
"%x:%x\n", op.bus, op.devfn);
drivers/pci/xen-pcifront.c
347
struct xen_pci_op op = {
drivers/pci/xen-pcifront.c
356
err = do_pci_op(pdev, &op);
drivers/phy/mscc/phy-ocelot-serdes.c
268
static int __serdes_write_mcb_s1g(struct regmap *regmap, u8 macro, u32 op)
drivers/phy/mscc/phy-ocelot-serdes.c
272
regmap_write(regmap, HSIO_MCB_S1G_ADDR_CFG, op |
drivers/phy/mscc/phy-ocelot-serdes.c
276
(regval & op) != op, 100,
drivers/phy/mscc/phy-ocelot-serdes.c
36
static int __serdes_write_mcb_s6g(struct regmap *regmap, u8 macro, u32 op)
drivers/phy/mscc/phy-ocelot-serdes.c
40
regmap_write(regmap, HSIO_MCB_S6G_ADDR_CFG, op |
drivers/phy/mscc/phy-ocelot-serdes.c
44
(regval & op) != op, 100,
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
441
{ .name = NAME, .type = LN_FTYPE_##TYPE, .op = OP }
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
473
u8 op;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
796
unsigned int op)
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
810
if (val == op)
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
820
LOCHNAGAR2_GPIO_CHANNEL_SRC_MASK, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
826
dev_dbg(priv->dev, "Set channel %d to 0x%x\n", free, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
836
unsigned int op)
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
844
ret = lochnagar2_get_gpio_chan(priv, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
851
op = ret;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
855
dev_dbg(priv->dev, "Set pin %s to 0x%x\n", pin->name, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
857
ret = regmap_write(priv->lochnagar->regmap, pin->reg, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
867
unsigned int op)
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
874
ret = regmap_update_bits(regmap, aif->src_reg, aif->src_mask, op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
923
return lochnagar_aif_set_mux(priv, group, func->op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
929
return lochnagar_pin_set_mux(priv, pin, func->op);
drivers/pinctrl/pinctrl-equilibrium.c
603
unsigned int *nr_funcs, funcs_util_ops op)
drivers/pinctrl/pinctrl-equilibrium.c
625
switch (op) {
drivers/platform/chrome/cros_ec_vbc.c
83
params->op = EC_VBNV_CONTEXT_OP_WRITE;
drivers/platform/chrome/wilco_ec/properties.c
20
u8 op; /* One of enum ec_property_op */
drivers/platform/chrome/wilco_ec/properties.c
28
u8 op; /* One of enum ec_property_op */
drivers/platform/chrome/wilco_ec/properties.c
51
if (rs->op != rq->op)
drivers/platform/chrome/wilco_ec/properties.c
67
rq.op = EC_OP_GET;
drivers/platform/chrome/wilco_ec/properties.c
89
rq.op = EC_OP_SET;
drivers/platform/chrome/wilco_ec/sysfs.c
103
static ssize_t get_info(struct device *dev, char *buf, enum get_ec_info_op op)
drivers/platform/chrome/wilco_ec/sysfs.c
106
struct get_ec_info_req req = { .cmd = CMD_EC_INFO, .op = op };
drivers/platform/chrome/wilco_ec/sysfs.c
189
rq.op = USB_CHARGE_GET;
drivers/platform/chrome/wilco_ec/sysfs.c
216
rq.op = USB_CHARGE_SET;
drivers/platform/chrome/wilco_ec/sysfs.c
40
u8 op; /* One of enum usb_charge_op */
drivers/platform/chrome/wilco_ec/sysfs.c
61
u8 op; /* One of enum get_ec_info_op */
drivers/platform/cznic/turris-signing-key.c
104
if (params->op != kernel_pkey_sign)
drivers/platform/x86/amd/pmf/pmf.h
877
void amd_pmf_update_slider(struct amd_pmf_dev *dev, bool op, int idx,
drivers/platform/x86/amd/pmf/sps.c
208
void amd_pmf_update_slider(struct amd_pmf_dev *dev, bool op, int idx,
drivers/platform/x86/amd/pmf/sps.c
213
if (op == SLIDER_OP_SET) {
drivers/platform/x86/amd/pmf/sps.c
227
} else if (op == SLIDER_OP_GET) {
drivers/platform/x86/fujitsu-laptop.c
152
int func, int op, int feature, int state)
drivers/platform/x86/fujitsu-laptop.c
156
{ .integer.type = ACPI_TYPE_INTEGER, .integer.value = op },
drivers/platform/x86/fujitsu-laptop.c
173
func, op, feature, state, (int)value);
drivers/platform/x86/intel_scu_ipc.c
281
u32 count, u32 op, u32 id)
drivers/platform/x86/intel_scu_ipc.c
305
ipc_command(scu, (count * 2) << 16 | id << 12 | 0 << 8 | op);
drivers/platform/x86/intel_scu_ipc.c
311
ipc_command(scu, (count * 3) << 16 | id << 12 | 0 << 8 | op);
drivers/platform/x86/intel_scu_ipc.c
316
ipc_command(scu, 4 << 16 | id << 12 | 0 << 8 | op);
drivers/pmdomain/samsung/exynos-pm-domains.c
41
char *op;
drivers/pmdomain/samsung/exynos-pm-domains.c
54
op = (power_on) ? "enable" : "disable";
drivers/pmdomain/samsung/exynos-pm-domains.c
55
pr_err("Power domain %s %s failed\n", domain->name, op);
drivers/ps3/ps3-sys-manager.c
288
enum ps3_sys_manager_next_op op,
drivers/ps3/ps3-sys-manager.c
303
dev_dbg(&dev->core, "%s:%d: (%xh)\n", __func__, __LINE__, op);
drivers/ps3/ps3-sys-manager.c
313
payload.type = op;
drivers/ps3/ps3stor_lib.c
269
const char *op = str_write_read(write);
drivers/ps3/ps3stor_lib.c
273
__func__, __LINE__, op, sectors, start_sector);
drivers/ps3/ps3stor_lib.c
284
__LINE__, op, res);
drivers/ps3/ps3stor_lib.c
291
__LINE__, op, dev->lv1_status);
drivers/ps3/ps3stor_lib.c
296
op);
drivers/rapidio/devices/tsi721.c
120
u32 op = do_wr ? MAINT_WR : MAINT_RD;
drivers/rapidio/devices/tsi721.c
132
bd_ptr[0].type_id = cpu_to_le32((DTYPE2 << 29) | (op << 19) | destid);
drivers/rtc/rtc-isl12026.c
115
u8 op[3] = {0, ISL12026_REG_SR, 0};
drivers/rtc/rtc-isl12026.c
119
.len = sizeof(op),
drivers/rtc/rtc-isl12026.c
120
.buf = op
drivers/rtc/rtc-isl12026.c
138
u8 op[3] = {0, reg, val};
drivers/rtc/rtc-isl12026.c
142
.len = sizeof(op),
drivers/rtc/rtc-isl12026.c
143
.buf = op
drivers/rtc/rtc-isl12026.c
168
u8 op[10];
drivers/rtc/rtc-isl12026.c
172
.len = sizeof(op),
drivers/rtc/rtc-isl12026.c
173
.buf = op
drivers/rtc/rtc-isl12026.c
181
op[0] = 0;
drivers/rtc/rtc-isl12026.c
182
op[1] = ISL12026_REG_SC;
drivers/rtc/rtc-isl12026.c
183
op[2] = bin2bcd(tm->tm_sec); /* SC */
drivers/rtc/rtc-isl12026.c
184
op[3] = bin2bcd(tm->tm_min); /* MN */
drivers/rtc/rtc-isl12026.c
185
op[4] = bin2bcd(tm->tm_hour) | ISL12026_REG_HR_MIL; /* HR */
drivers/rtc/rtc-isl12026.c
186
op[5] = bin2bcd(tm->tm_mday); /* DT */
drivers/rtc/rtc-isl12026.c
187
op[6] = bin2bcd(tm->tm_mon + 1); /* MO */
drivers/rtc/rtc-isl12026.c
188
op[7] = bin2bcd(tm->tm_year % 100); /* YR */
drivers/rtc/rtc-isl12026.c
189
op[8] = bin2bcd(tm->tm_wday & 7); /* DW */
drivers/rtc/rtc-isl12026.c
190
op[9] = bin2bcd(tm->tm_year >= 100 ? 20 : 19); /* Y2K */
drivers/rtc/rtc-isl12026.c
76
u8 op[3];
drivers/rtc/rtc-isl12026.c
81
.buf = op
drivers/rtc/rtc-isl12026.c
85
op[0] = 0;
drivers/rtc/rtc-isl12026.c
86
op[1] = ISL12026_REG_SR;
drivers/rtc/rtc-isl12026.c
87
op[2] = ISL12026_REG_SR_WEL;
drivers/rtc/rtc-isl12026.c
97
op[2] = ISL12026_REG_SR_WEL | ISL12026_REG_SR_RWEL;
drivers/rtc/rtc-max77686.c
344
enum MAX77686_RTC_OP op)
drivers/rtc/rtc-max77686.c
350
if (op == MAX77686_RTC_WRITE)
drivers/rtc/rtc-mpc5121.c
291
static int mpc5121_rtc_probe(struct platform_device *op)
drivers/rtc/rtc-mpc5121.c
296
rtc = devm_kzalloc(&op->dev, sizeof(*rtc), GFP_KERNEL);
drivers/rtc/rtc-mpc5121.c
300
rtc->regs = devm_platform_ioremap_resource(op, 0);
drivers/rtc/rtc-mpc5121.c
302
dev_err(&op->dev, "%s: couldn't map io space\n", __func__);
drivers/rtc/rtc-mpc5121.c
306
device_init_wakeup(&op->dev, true);
drivers/rtc/rtc-mpc5121.c
308
platform_set_drvdata(op, rtc);
drivers/rtc/rtc-mpc5121.c
310
rtc->irq = irq_of_parse_and_map(op->dev.of_node, 1);
drivers/rtc/rtc-mpc5121.c
311
err = devm_request_irq(&op->dev, rtc->irq, mpc5121_rtc_handler, 0,
drivers/rtc/rtc-mpc5121.c
312
"mpc5121-rtc", &op->dev);
drivers/rtc/rtc-mpc5121.c
314
dev_err(&op->dev, "%s: could not request irq: %i\n",
drivers/rtc/rtc-mpc5121.c
319
rtc->irq_periodic = irq_of_parse_and_map(op->dev.of_node, 0);
drivers/rtc/rtc-mpc5121.c
320
err = devm_request_irq(&op->dev, rtc->irq_periodic,
drivers/rtc/rtc-mpc5121.c
322
&op->dev);
drivers/rtc/rtc-mpc5121.c
324
dev_err(&op->dev, "%s: could not request irq: %i\n",
drivers/rtc/rtc-mpc5121.c
329
rtc->rtc = devm_rtc_allocate_device(&op->dev);
drivers/rtc/rtc-mpc5121.c
341
if (of_device_is_compatible(op->dev.of_node, "fsl,mpc5121-rtc")) {
drivers/rtc/rtc-mpc5121.c
345
dev_warn(&op->dev,
drivers/rtc/rtc-mpc5121.c
373
static void mpc5121_rtc_remove(struct platform_device *op)
drivers/rtc/rtc-mpc5121.c
375
struct mpc5121_rtc_data *rtc = platform_get_drvdata(op);
drivers/s390/char/sclp_tty.c
359
unsigned char *ip, *op;
drivers/s390/char/sclp_tty.c
364
ip = op = buf;
drivers/s390/char/sclp_tty.c
374
*op++ = *ip++;
drivers/s390/char/sclp_tty.c
390
*op++ = _ebc_toupper[(int) *ip++];
drivers/s390/char/sclp_tty.c
393
*op++ = _ebc_tolower[(int) *ip++];
drivers/s390/char/sclp_tty.c
396
*op++ = *ip++;
drivers/s390/char/sclp_tty.c
399
return op - buf;
drivers/s390/char/tape.h
135
enum tape_op op;
drivers/s390/char/tape_3490.c
106
enum tape_op op;
drivers/s390/char/tape_3490.c
127
switch(p->op) {
drivers/s390/char/tape_3490.c
139
tape_3490_schedule_work(struct tape_device *device, enum tape_op op)
drivers/s390/char/tape_3490.c
149
p->op = op;
drivers/s390/char/tape_3490.c
161
DBF_EVENT(6, "%s done\n", tape_op_verbose[request->op]);
drivers/s390/char/tape_3490.c
169
tape_op_verbose[request->op], rc);
drivers/s390/char/tape_3490.c
177
tape_op_verbose[request->op]);
drivers/s390/char/tape_3490.c
184
DBF_EVENT(3, "xerp retr %s\n", tape_op_verbose[request->op]);
drivers/s390/char/tape_3490.c
210
if (request->op != TO_ASSIGN) {
drivers/s390/char/tape_3490.c
276
request->op == TO_DSE ||
drivers/s390/char/tape_3490.c
277
request->op == TO_WRI ||
drivers/s390/char/tape_3490.c
278
request->op == TO_WTM
drivers/s390/char/tape_3490.c
307
switch (request->op) {
drivers/s390/char/tape_3490.c
435
if (request->op == TO_RUN) {
drivers/s390/char/tape_3490.c
450
if (request->op == TO_DSE)
drivers/s390/char/tape_3490.c
487
if (request->op==TO_WRI ||
drivers/s390/char/tape_3490.c
488
request->op==TO_DSE ||
drivers/s390/char/tape_3490.c
489
request->op==TO_WTM)
drivers/s390/char/tape_3490.c
508
switch(request->op) {
drivers/s390/char/tape_3490.c
522
if (request->op != TO_BLOCK && request->op != TO_LBL)
drivers/s390/char/tape_3490.c
574
if (request->op == TO_RUN) {
drivers/s390/char/tape_3490.c
623
(request->op == TO_WRI)) {
drivers/s390/char/tape_3490.c
636
if (request->op == TO_BSB || request->op == TO_FSB)
drivers/s390/char/tape_3490.c
707
request->op = TO_LBL;
drivers/s390/char/tape_3490.c
80
request->op = TO_MSEN;
drivers/s390/char/tape_3490.c
97
request->op = TO_MSEN;
drivers/s390/char/tape_char.c
339
struct mtop op;
drivers/s390/char/tape_char.c
341
if (copy_from_user(&op, data, sizeof(op)) != 0)
drivers/s390/char/tape_char.c
343
if (op.mt_count < 0)
drivers/s390/char/tape_char.c
350
switch (op.mt_op) {
drivers/s390/char/tape_char.c
365
rc = tape_mtop(device, op.mt_op, op.mt_count);
drivers/s390/char/tape_char.c
367
if (op.mt_op == MTWEOF && rc == 0) {
drivers/s390/char/tape_char.c
368
if (op.mt_count > device->required_tapemarks)
drivers/s390/char/tape_char.c
371
device->required_tapemarks -= op.mt_count;
drivers/s390/char/tape_core.c
147
rc = sysfs_emit(buf, "%s\n", tape_op_verbose[req->op]);
drivers/s390/char/tape_core.c
897
const char* op;
drivers/s390/char/tape_core.c
900
op = tape_op_verbose[request->op];
drivers/s390/char/tape_core.c
902
op = "---";
drivers/s390/char/tape_core.c
905
DBF_EVENT(3, "DEVICE: %08x OP\t: %s\n", device->cdev_id, op);
drivers/s390/char/tape_core.c
923
switch (request->op) {
drivers/s390/char/tape_proc.c
73
str = tape_op_verbose[request->op];
drivers/s390/char/tape_std.c
107
request->op = TO_UNASSIGN;
drivers/s390/char/tape_std.c
132
request->op = TO_RBI;
drivers/s390/char/tape_std.c
230
request->op = TO_FSF;
drivers/s390/char/tape_std.c
255
request->op = TO_FSB;
drivers/s390/char/tape_std.c
287
request->op = TO_BSB;
drivers/s390/char/tape_std.c
317
request->op = TO_WTM;
drivers/s390/char/tape_std.c
342
request->op = TO_BSF;
drivers/s390/char/tape_std.c
367
request->op = TO_BSF;
drivers/s390/char/tape_std.c
398
request->op = TO_FSF;
drivers/s390/char/tape_std.c
426
request->op = TO_REW;
drivers/s390/char/tape_std.c
449
request->op = TO_RUN;
drivers/s390/char/tape_std.c
470
request->op = TO_NOP;
drivers/s390/char/tape_std.c
521
request->op = TO_FSF;
drivers/s390/char/tape_std.c
544
request->op = TO_DSE;
drivers/s390/char/tape_std.c
582
request->op = TO_NOP;
drivers/s390/char/tape_std.c
612
request->op = TO_RFO;
drivers/s390/char/tape_std.c
62
request->op = TO_ASSIGN;
drivers/s390/char/tape_std.c
639
request->op = TO_WRI;
drivers/s390/cio/chsc.c
1309
int chsc_sstpc(void *page, unsigned int op, u16 ctrl, long *clock_delta)
drivers/s390/cio/chsc.c
1314
unsigned int op : 8;
drivers/s390/cio/chsc.c
1329
rr->op = op;
drivers/s390/cio/chsc.c
1492
u16 op;
drivers/s390/cio/chsc.c
1513
sgib_area->op = 0x1;
drivers/s390/cio/chsc.c
527
u8 op;
drivers/s390/cio/chsc.c
548
events[data->op], chpid.cssid, chpid.id);
drivers/s390/cio/chsc.c
549
switch (data->op) {
drivers/s390/cio/itcw.c
180
struct itcw *itcw_init(void *buffer, size_t size, int op, int intrg,
drivers/s390/cio/itcw.c
216
tcw_init(itcw->tcw, (op == ITCW_OP_READ) ? 1 : 0,
drivers/s390/cio/itcw.c
217
(op == ITCW_OP_WRITE) ? 1 : 0);
drivers/sbus/char/bbc_envctrl.c
446
static void attach_one_temp(struct bbc_i2c_bus *bp, struct platform_device *op,
drivers/sbus/char/bbc_envctrl.c
458
tp->client = bbc_i2c_attach(bp, op);
drivers/sbus/char/bbc_envctrl.c
494
static void attach_one_fan(struct bbc_i2c_bus *bp, struct platform_device *op,
drivers/sbus/char/bbc_envctrl.c
506
fp->client = bbc_i2c_attach(bp, op);
drivers/sbus/char/bbc_envctrl.c
568
struct platform_device *op;
drivers/sbus/char/bbc_envctrl.c
573
while ((op = bbc_i2c_getdev(bp, devidx++)) != NULL) {
drivers/sbus/char/bbc_envctrl.c
574
if (of_node_name_eq(op->dev.of_node, "temperature"))
drivers/sbus/char/bbc_envctrl.c
575
attach_one_temp(bp, op, temp_index++);
drivers/sbus/char/bbc_envctrl.c
576
if (of_node_name_eq(op->dev.of_node, "fan-control"))
drivers/sbus/char/bbc_envctrl.c
577
attach_one_fan(bp, op, fan_index++);
drivers/sbus/char/bbc_i2c.c
101
reg = of_get_property(op->dev.of_node, "reg", NULL);
drivers/sbus/char/bbc_i2c.c
110
claim_device(bp, op);
drivers/sbus/char/bbc_i2c.c
118
struct platform_device *op = client->op;
drivers/sbus/char/bbc_i2c.c
120
release_device(bp, op);
drivers/sbus/char/bbc_i2c.c
295
static struct bbc_i2c_bus * attach_one_i2c(struct platform_device *op, int index)
drivers/sbus/char/bbc_i2c.c
308
bp->i2c_control_regs = of_ioremap(&op->resource[0], 0, 0x2, "bbc_i2c_regs");
drivers/sbus/char/bbc_i2c.c
312
if (op->num_resources == 2) {
drivers/sbus/char/bbc_i2c.c
313
bp->i2c_bussel_reg = of_ioremap(&op->resource[1], 0, 0x1, "bbc_i2c_bussel");
drivers/sbus/char/bbc_i2c.c
320
if (request_irq(op->archdata.irqs[0], bbc_i2c_interrupt,
drivers/sbus/char/bbc_i2c.c
325
bp->op = op;
drivers/sbus/char/bbc_i2c.c
330
for (dp = op->dev.of_node->child;
drivers/sbus/char/bbc_i2c.c
354
of_iounmap(&op->resource[1], bp->i2c_bussel_reg, 1);
drivers/sbus/char/bbc_i2c.c
356
of_iounmap(&op->resource[0], bp->i2c_control_regs, 2);
drivers/sbus/char/bbc_i2c.c
361
static int bbc_i2c_probe(struct platform_device *op)
drivers/sbus/char/bbc_i2c.c
366
bp = attach_one_i2c(op, index);
drivers/sbus/char/bbc_i2c.c
372
free_irq(op->archdata.irqs[0], bp);
drivers/sbus/char/bbc_i2c.c
374
of_iounmap(&op->resource[0], bp->i2c_bussel_reg, 1);
drivers/sbus/char/bbc_i2c.c
376
of_iounmap(&op->resource[1], bp->i2c_control_regs, 2);
drivers/sbus/char/bbc_i2c.c
379
dev_set_drvdata(&op->dev, bp);
drivers/sbus/char/bbc_i2c.c
385
static void bbc_i2c_remove(struct platform_device *op)
drivers/sbus/char/bbc_i2c.c
387
struct bbc_i2c_bus *bp = dev_get_drvdata(&op->dev);
drivers/sbus/char/bbc_i2c.c
391
free_irq(op->archdata.irqs[0], bp);
drivers/sbus/char/bbc_i2c.c
394
of_iounmap(&op->resource[0], bp->i2c_bussel_reg, 1);
drivers/sbus/char/bbc_i2c.c
396
of_iounmap(&op->resource[1], bp->i2c_control_regs, 2);
drivers/sbus/char/bbc_i2c.c
55
static void set_device_claimage(struct bbc_i2c_bus *bp, struct platform_device *op, int val)
drivers/sbus/char/bbc_i2c.c
60
if (bp->devs[i].device == op) {
drivers/sbus/char/bbc_i2c.c
72
struct platform_device *op = NULL;
drivers/sbus/char/bbc_i2c.c
76
if (!(op = bp->devs[i].device))
drivers/sbus/char/bbc_i2c.c
80
op = NULL;
drivers/sbus/char/bbc_i2c.c
86
return op;
drivers/sbus/char/bbc_i2c.c
90
struct bbc_i2c_client *bbc_i2c_attach(struct bbc_i2c_bus *bp, struct platform_device *op)
drivers/sbus/char/bbc_i2c.c
99
client->op = op;
drivers/sbus/char/bbc_i2c.h
10
struct platform_device *op;
drivers/sbus/char/bbc_i2c.h
67
struct platform_device *op;
drivers/sbus/char/display7seg.c
170
static int d7s_probe(struct platform_device *op)
drivers/sbus/char/display7seg.c
180
p = devm_kzalloc(&op->dev, sizeof(*p), GFP_KERNEL);
drivers/sbus/char/display7seg.c
185
p->regs = of_ioremap(&op->resource[0], 0, sizeof(u8), "d7s");
drivers/sbus/char/display7seg.c
214
op->dev.of_node,
drivers/sbus/char/display7seg.c
216
op->resource[0].start,
drivers/sbus/char/display7seg.c
219
dev_set_drvdata(&op->dev, p);
drivers/sbus/char/display7seg.c
228
of_iounmap(&op->resource[0], p->regs, sizeof(u8));
drivers/sbus/char/display7seg.c
232
static void d7s_remove(struct platform_device *op)
drivers/sbus/char/display7seg.c
234
struct d7s *p = dev_get_drvdata(&op->dev);
drivers/sbus/char/display7seg.c
247
of_iounmap(&op->resource[0], p->regs, sizeof(u8));
drivers/sbus/char/envctrl.c
1024
static int envctrl_probe(struct platform_device *op)
drivers/sbus/char/envctrl.c
1032
i2c = of_ioremap(&op->resource[0], 0, 0x2, DRIVER_NAME);
drivers/sbus/char/envctrl.c
1037
dp = op->dev.of_node->child;
drivers/sbus/char/envctrl.c
1093
of_iounmap(&op->resource[0], i2c, 0x2);
drivers/sbus/char/envctrl.c
1100
static void envctrl_remove(struct platform_device *op)
drivers/sbus/char/envctrl.c
1106
of_iounmap(&op->resource[0], i2c, 0x2);
drivers/sbus/char/flash.c
159
static int flash_probe(struct platform_device *op)
drivers/sbus/char/flash.c
161
struct device_node *dp = op->dev.of_node;
drivers/sbus/char/flash.c
171
flash.read_base = op->resource[0].start;
drivers/sbus/char/flash.c
172
flash.read_size = resource_size(&op->resource[0]);
drivers/sbus/char/flash.c
173
if (op->resource[1].flags) {
drivers/sbus/char/flash.c
174
flash.write_base = op->resource[1].start;
drivers/sbus/char/flash.c
175
flash.write_size = resource_size(&op->resource[1]);
drivers/sbus/char/flash.c
177
flash.write_base = op->resource[0].start;
drivers/sbus/char/flash.c
178
flash.write_size = resource_size(&op->resource[0]);
drivers/sbus/char/flash.c
183
op->dev.of_node,
drivers/sbus/char/flash.c
190
static void flash_remove(struct platform_device *op)
drivers/sbus/char/openprom.c
130
static int opromgetprop(void __user *argp, struct device_node *dp, struct openpromio *op, int bufsize)
drivers/sbus/char/openprom.c
136
!(pval = of_get_property(dp, op->oprom_array, &len)) ||
drivers/sbus/char/openprom.c
138
return copyout(argp, op, sizeof(int));
drivers/sbus/char/openprom.c
140
memcpy(op->oprom_array, pval, len);
drivers/sbus/char/openprom.c
141
op->oprom_array[len] = '\0';
drivers/sbus/char/openprom.c
142
op->oprom_size = len;
drivers/sbus/char/openprom.c
144
return copyout(argp, op, sizeof(int) + bufsize);
drivers/sbus/char/openprom.c
147
static int opromnxtprop(void __user *argp, struct device_node *dp, struct openpromio *op, int bufsize)
drivers/sbus/char/openprom.c
153
return copyout(argp, op, sizeof(int));
drivers/sbus/char/openprom.c
154
if (op->oprom_array[0] == '\0') {
drivers/sbus/char/openprom.c
157
return copyout(argp, op, sizeof(int));
drivers/sbus/char/openprom.c
160
prop = of_find_property(dp, op->oprom_array, NULL);
drivers/sbus/char/openprom.c
165
return copyout(argp, op, sizeof(int));
drivers/sbus/char/openprom.c
170
memcpy(op->oprom_array, prop->name, len);
drivers/sbus/char/openprom.c
171
op->oprom_array[len] = '\0';
drivers/sbus/char/openprom.c
172
op->oprom_size = ++len;
drivers/sbus/char/openprom.c
174
return copyout(argp, op, sizeof(int) + bufsize);
drivers/sbus/char/openprom.c
177
static int opromsetopt(struct device_node *dp, struct openpromio *op, int bufsize)
drivers/sbus/char/openprom.c
179
char *buf = op->oprom_array + strlen(op->oprom_array) + 1;
drivers/sbus/char/openprom.c
180
int len = op->oprom_array + bufsize - buf;
drivers/sbus/char/openprom.c
182
return of_set_property(options_node, op->oprom_array, buf, len);
drivers/sbus/char/openprom.c
185
static int opromnext(void __user *argp, unsigned int cmd, struct device_node *dp, struct openpromio *op, int bufsize, DATA *data)
drivers/sbus/char/openprom.c
194
ph = *((int *) op->oprom_array);
drivers/sbus/char/openprom.c
226
*((int *) op->oprom_array) = ph;
drivers/sbus/char/openprom.c
227
op->oprom_size = sizeof(phandle);
drivers/sbus/char/openprom.c
229
return copyout(argp, op, bufsize + sizeof(int));
drivers/sbus/char/openprom.c
232
static int oprompci2node(void __user *argp, struct device_node *dp, struct openpromio *op, int bufsize, DATA *data)
drivers/sbus/char/openprom.c
242
((int *) op->oprom_array)[0],
drivers/sbus/char/openprom.c
243
((int *) op->oprom_array)[1]);
drivers/sbus/char/openprom.c
247
*((int *)op->oprom_array) = dp->phandle;
drivers/sbus/char/openprom.c
248
op->oprom_size = sizeof(int);
drivers/sbus/char/openprom.c
249
err = copyout(argp, op, bufsize + sizeof(int));
drivers/sbus/char/openprom.c
258
static int oprompath2node(void __user *argp, struct device_node *dp, struct openpromio *op, int bufsize, DATA *data)
drivers/sbus/char/openprom.c
262
dp = of_find_node_by_path(op->oprom_array);
drivers/sbus/char/openprom.c
266
*((int *)op->oprom_array) = ph;
drivers/sbus/char/openprom.c
267
op->oprom_size = sizeof(int);
drivers/sbus/char/openprom.c
269
return copyout(argp, op, bufsize + sizeof(int));
drivers/sbus/char/openprom.c
272
static int opromgetbootargs(void __user *argp, struct openpromio *op, int bufsize)
drivers/sbus/char/openprom.c
280
strcpy(op->oprom_array, buf);
drivers/sbus/char/openprom.c
281
op->oprom_size = len;
drivers/sbus/char/openprom.c
283
return copyout(argp, op, bufsize + sizeof(int));
drivers/sbus/char/openprom.c
387
struct opiocdesc op;
drivers/sbus/char/openprom.c
393
if (copy_from_user(&op, argp, sizeof(op)))
drivers/sbus/char/openprom.c
396
dp = get_node(op.op_nodeid, data);
drivers/sbus/char/openprom.c
398
str = copyin_string(op.op_name, op.op_namelen);
drivers/sbus/char/openprom.c
404
if (!pval || len > op.op_buflen) {
drivers/sbus/char/openprom.c
407
op.op_buflen = len;
drivers/sbus/char/openprom.c
408
if (copy_to_user(argp, &op, sizeof(op)) ||
drivers/sbus/char/openprom.c
409
copy_to_user(op.op_buf, pval, len))
drivers/sbus/char/openprom.c
419
struct opiocdesc op;
drivers/sbus/char/openprom.c
425
if (copy_from_user(&op, argp, sizeof(op)))
drivers/sbus/char/openprom.c
428
dp = get_node(op.op_nodeid, data);
drivers/sbus/char/openprom.c
432
str = copyin_string(op.op_name, op.op_namelen);
drivers/sbus/char/openprom.c
450
if (len > op.op_buflen)
drivers/sbus/char/openprom.c
451
len = op.op_buflen;
drivers/sbus/char/openprom.c
453
if (copy_to_user(argp, &op, sizeof(op)))
drivers/sbus/char/openprom.c
457
copy_to_user(op.op_buf, prop->value, len))
drivers/sbus/char/openprom.c
465
struct opiocdesc op;
drivers/sbus/char/openprom.c
470
if (copy_from_user(&op, argp, sizeof(op)))
drivers/sbus/char/openprom.c
473
dp = get_node(op.op_nodeid, data);
drivers/sbus/char/openprom.c
477
str = copyin_string(op.op_name, op.op_namelen);
drivers/sbus/char/openprom.c
481
tmp = copyin_string(op.op_buf, op.op_buflen);
drivers/sbus/char/openprom.c
487
err = of_set_property(dp, str, tmp, op.op_buflen);
drivers/sbus/char/uctrl.c
346
static int uctrl_probe(struct platform_device *op)
drivers/sbus/char/uctrl.c
357
p->regs = of_ioremap(&op->resource[0], 0,
drivers/sbus/char/uctrl.c
358
resource_size(&op->resource[0]),
drivers/sbus/char/uctrl.c
365
p->irq = op->archdata.irqs[0];
drivers/sbus/char/uctrl.c
380
op->dev.of_node, p->regs, p->irq);
drivers/sbus/char/uctrl.c
384
dev_set_drvdata(&op->dev, p);
drivers/sbus/char/uctrl.c
394
of_iounmap(&op->resource[0], p->regs, resource_size(&op->resource[0]));
drivers/sbus/char/uctrl.c
401
static void uctrl_remove(struct platform_device *op)
drivers/sbus/char/uctrl.c
403
struct uctrl_driver *p = dev_get_drvdata(&op->dev);
drivers/sbus/char/uctrl.c
408
of_iounmap(&op->resource[0], p->regs, resource_size(&op->resource[0]));
drivers/scsi/aha1542.c
504
ccb[mbo].op = 0; /* SCSI Initiator Command */
drivers/scsi/aha1542.c
904
ccb[mbo].op = 0x81; /* BUS DEVICE RESET */
drivers/scsi/aha1542.h
83
u8 op; /* Command Control Block Operation Code: */
drivers/scsi/aic7xxx/aic79xx_core.c
519
ahd_sync_scb(struct ahd_softc *ahd, struct scb *scb, int op)
drivers/scsi/aic7xxx/aic79xx_core.c
524
/*len*/sizeof(*scb->hscb), op);
drivers/scsi/aic7xxx/aic79xx_core.c
528
ahd_sync_sglist(struct ahd_softc *ahd, struct scb *scb, int op)
drivers/scsi/aic7xxx/aic79xx_core.c
536
/*len*/ahd_sg_size(ahd) * scb->sg_count, op);
drivers/scsi/aic7xxx/aic79xx_core.c
540
ahd_sync_sense(struct ahd_softc *ahd, struct scb *scb, int op)
drivers/scsi/aic7xxx/aic79xx_core.c
545
/*len*/AHD_SENSE_BUFSIZE, op);
drivers/scsi/aic7xxx/aic79xx_core.c
887
ahd_sync_qoutfifo(struct ahd_softc *ahd, int op)
drivers/scsi/aic7xxx/aic79xx_core.c
891
/*len*/AHD_SCB_MAX * sizeof(struct ahd_completion), op);
drivers/scsi/aic7xxx/aic79xx_core.c
895
ahd_sync_tqinfifo(struct ahd_softc *ahd, int op)
drivers/scsi/aic7xxx/aic79xx_core.c
903
op);
drivers/scsi/aic7xxx/aic79xx_inline.h
107
struct scb *scb, int op);
drivers/scsi/aic7xxx/aic79xx_osm.h
204
#define ahd_dmamap_sync(ahd, dma_tag, dmamap, offset, len, op)
drivers/scsi/aic7xxx/aic7xxx_core.c
395
ahc_sync_scb(struct ahc_softc *ahc, struct scb *scb, int op)
drivers/scsi/aic7xxx/aic7xxx_core.c
400
/*len*/sizeof(*scb->hscb), op);
drivers/scsi/aic7xxx/aic7xxx_core.c
404
ahc_sync_sglist(struct ahc_softc *ahc, struct scb *scb, int op)
drivers/scsi/aic7xxx/aic7xxx_core.c
412
/*len*/sizeof(struct ahc_dma_seg) * scb->sg_count, op);
drivers/scsi/aic7xxx/aic7xxx_core.c
667
ahc_sync_qoutfifo(struct ahc_softc *ahc, int op)
drivers/scsi/aic7xxx/aic7xxx_core.c
670
/*offset*/0, /*len*/256, op);
drivers/scsi/aic7xxx/aic7xxx_core.c
674
ahc_sync_tqinfifo(struct ahc_softc *ahc, int op)
drivers/scsi/aic7xxx/aic7xxx_core.c
682
op);
drivers/scsi/aic7xxx/aic7xxx_inline.h
55
struct scb *scb, int op);
drivers/scsi/aic7xxx/aic7xxx_osm.h
217
#define ahc_dmamap_sync(ahc, dma_tag, dmamap, offset, len, op)
drivers/scsi/aic94xx/aic94xx_hwi.c
1256
void asd_turn_led(struct asd_ha_struct *asd_ha, int phy_id, int op)
drivers/scsi/aic94xx/aic94xx_hwi.c
1260
if (op)
drivers/scsi/aic94xx/aic94xx_hwi.c
1277
void asd_control_led(struct asd_ha_struct *asd_ha, int phy_id, int op)
drivers/scsi/aic94xx/aic94xx_hwi.c
1283
if (op)
drivers/scsi/aic94xx/aic94xx_hwi.c
1290
if (op)
drivers/scsi/aic94xx/aic94xx_hwi.h
372
void asd_control_led(struct asd_ha_struct *asd_ha, int phy_id, int op);
drivers/scsi/aic94xx/aic94xx_hwi.h
373
void asd_turn_led(struct asd_ha_struct *asd_ha, int phy_id, int op);
drivers/scsi/bfa/bfa_ioc.c
121
bfa_fsm_state_decl(bfa_ioc, op, struct bfa_ioc_s, enum ioc_event);
drivers/scsi/bfa/bfa_ioc.c
6699
u8 rd_cnt, u8 ad_cnt, u8 op)
drivers/scsi/bfa/bfa_ioc.c
6708
cmd.r.cmd = op;
drivers/scsi/bnx2fc/bnx2fc.h
576
struct fc_frame *fp, unsigned int op,
drivers/scsi/bnx2fc/bnx2fc_els.c
23
static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
drivers/scsi/bnx2fc/bnx2fc_els.c
670
static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
drivers/scsi/bnx2fc/bnx2fc_els.c
691
printk(KERN_ERR PFX "els 0x%x: rport not ready\n", op);
drivers/scsi/bnx2fc/bnx2fc_els.c
696
printk(KERN_ERR PFX "els 0x%x: link is not ready\n", op);
drivers/scsi/bnx2fc/bnx2fc_els.c
701
printk(KERN_ERR PFX "els 0x%x: tgt not ready\n", op);
drivers/scsi/bnx2fc/bnx2fc_els.c
738
if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) {
drivers/scsi/bnx2fc/bnx2fc_els.c
741
printk(KERN_ERR PFX "Invalid ELS op 0x%x\n", op);
drivers/scsi/bnx2fc/bnx2fc_els.c
759
if (op == ELS_SRR)
drivers/scsi/bnx2fc/bnx2fc_els.c
926
struct fc_frame *fp, unsigned int op,
drivers/scsi/bnx2fc/bnx2fc_els.c
937
switch (op) {
drivers/scsi/bnx2fc/bnx2fc_els.c
940
return fc_elsct_send(lport, did, fp, op, bnx2fc_flogi_resp,
drivers/scsi/bnx2fc/bnx2fc_els.c
946
return fc_elsct_send(lport, did, fp, op, bnx2fc_logo_resp,
drivers/scsi/bnx2fc/bnx2fc_els.c
949
return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
1101
u16 op;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
1106
op = ntohs(fiph->fip_op);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
1109
if (op == FIP_OP_CTRL && sub == FIP_SC_SOL && bnx2fc_log_fka)
drivers/scsi/bnx2fc/bnx2fc_hwi.c
561
u8 op;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
595
op = fc_frame_payload_op(fp);
drivers/scsi/bnx2fc/bnx2fc_hwi.c
596
if ((op == ELS_TEST) || (op == ELS_ESTC) ||
drivers/scsi/bnx2fc/bnx2fc_hwi.c
597
(op == ELS_FAN) || (op == ELS_CSU)) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
602
printk(KERN_ERR PFX "dropping ELS 0x%x\n", op);
drivers/scsi/ch.c
188
void *buffer, unsigned int buflength, enum req_op op)
drivers/scsi/ch.c
213
result = scsi_execute_cmd(ch->device, cmd, op, buffer, buflength,
drivers/scsi/csiostor/csio_hw.c
3936
__u8 op;
drivers/scsi/csiostor/csio_hw.c
3941
op = ((struct rss_header *) wr)->opcode;
drivers/scsi/csiostor/csio_hw.c
3942
if (op == CPL_FW6_PLD) {
drivers/scsi/csiostor/csio_hw.c
3952
} else if (op == CPL_FW6_MSG || op == CPL_FW4_MSG) {
drivers/scsi/csiostor/csio_hw.c
3957
msg_len = (op == CPL_FW6_MSG) ? sizeof(struct cpl_fw6_msg) :
drivers/scsi/csiostor/csio_hw.c
3960
csio_warn(hw, "unexpected CPL %#x on FW event queue\n", op);
drivers/scsi/csiostor/csio_lnode.c
211
csio_fill_ct_iu(void *buf, uint8_t type, uint8_t sub_type, uint16_t op)
drivers/scsi/csiostor/csio_lnode.c
217
cmd->ct_cmd = htons(op);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
685
u32 pdu_type = (info->op == ISCSI_OP_SCSI_CMD) ? 0 : 1;
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
709
info->flags, info->op, info->ahs, info->num_pdu,
drivers/scsi/cxgbi/libcxgbi.c
1885
int cxgbi_conn_alloc_pdu(struct iscsi_task *task, u8 op)
drivers/scsi/cxgbi/libcxgbi.c
1912
op &= ISCSI_OPCODE_MASK;
drivers/scsi/cxgbi/libcxgbi.c
1920
if ((op == ISCSI_OP_SCSI_DATA_OUT) ||
drivers/scsi/cxgbi/libcxgbi.c
1921
((op == ISCSI_OP_SCSI_CMD) &&
drivers/scsi/cxgbi/libcxgbi.c
1953
if ((op == ISCSI_OP_SCSI_CMD) &&
drivers/scsi/cxgbi/libcxgbi.c
2044
if (op != ISCSI_OP_SCSI_DATA_OUT)
drivers/scsi/cxgbi/libcxgbi.c
2049
task, op, tdata->skb, cdev->skb_tx_rsvd, headroom,
drivers/scsi/cxgbi/libcxgbi.c
2116
iso_info->op = task->hdr->opcode;
drivers/scsi/cxgbi/libcxgbi.h
563
u8 op;
drivers/scsi/elx/efct/efct_lio.c
107
if (kstrtoul(page, 0, &op) < 0)
drivers/scsi/elx/efct/efct_lio.c
110
if (op == 1) {
drivers/scsi/elx/efct/efct_lio.c
122
} else if (op == 0) {
drivers/scsi/elx/efct/efct_lio.c
156
unsigned long op;
drivers/scsi/elx/efct/efct_lio.c
158
if (kstrtoul(page, 0, &op) < 0)
drivers/scsi/elx/efct/efct_lio.c
169
if (op == 1) {
drivers/scsi/elx/efct/efct_lio.c
193
} else if (op == 0) {
drivers/scsi/elx/efct/efct_lio.c
97
unsigned long op;
drivers/scsi/fcoe/fcoe.c
120
unsigned int op,
drivers/scsi/fcoe/fcoe.c
2588
struct fc_frame *fp, unsigned int op,
drivers/scsi/fcoe/fcoe.c
2599
switch (op) {
drivers/scsi/fcoe/fcoe.c
2604
return fc_elsct_send(lport, did, fp, op, fcoe_flogi_resp,
drivers/scsi/fcoe/fcoe.c
2610
return fc_elsct_send(lport, did, fp, op, fcoe_logo_resp,
drivers/scsi/fcoe/fcoe.c
2613
return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
drivers/scsi/fcoe/fcoe_ctlr.c
1534
u16 op;
drivers/scsi/fcoe/fcoe_ctlr.c
1551
op = ntohs(fiph->fip_op);
drivers/scsi/fcoe/fcoe_ctlr.c
1570
if (fip->mode == FIP_MODE_VN2VN && op == FIP_OP_VN2VN)
drivers/scsi/fcoe/fcoe_ctlr.c
1573
if (fip_vlan_resp && op == FIP_OP_VLAN) {
drivers/scsi/fcoe/fcoe_ctlr.c
1582
if (op == FIP_OP_LS) {
drivers/scsi/fcoe/fcoe_ctlr.c
1590
if (op == FIP_OP_DISC && sub == FIP_SC_ADV)
drivers/scsi/fcoe/fcoe_ctlr.c
1592
else if (op == FIP_OP_CTRL && sub == FIP_SC_CLR_VLINK)
drivers/scsi/fcoe/fcoe_ctlr.c
1900
u8 op;
drivers/scsi/fcoe/fcoe_ctlr.c
1908
op = fc_frame_payload_op(fp);
drivers/scsi/fcoe/fcoe_ctlr.c
1909
if (op == ELS_LS_ACC && fh->fh_r_ctl == FC_RCTL_ELS_REP &&
drivers/scsi/fcoe/fcoe_ctlr.c
1936
} else if (op == ELS_FLOGI && fh->fh_r_ctl == FC_RCTL_ELS_REQ && sa) {
drivers/scsi/fcoe/fcoe_ctlr.c
617
u8 op;
drivers/scsi/fcoe/fcoe_ctlr.c
620
op = *(u8 *)(fh + 1);
drivers/scsi/fcoe/fcoe_ctlr.c
645
if (op == ELS_LS_ACC || op == ELS_LS_RJT)
drivers/scsi/fcoe/fcoe_ctlr.c
654
if (op != ELS_LS_RJT) {
drivers/scsi/fcoe/fcoe_ctlr.c
703
u8 op;
drivers/scsi/fcoe/fcoe_ctlr.c
708
op = *(u8 *)(fh + 1);
drivers/scsi/fcoe/fcoe_ctlr.c
710
if (op == ELS_FLOGI && fip->mode != FIP_MODE_VN2VN) {
drivers/scsi/fcoe/fcoe_ctlr.c
730
switch (op) {
drivers/scsi/fcoe/fcoe_ctlr.c
732
op = FIP_DT_FLOGI;
drivers/scsi/fcoe/fcoe_ctlr.c
745
op = FIP_DT_FDISC;
drivers/scsi/fcoe/fcoe_ctlr.c
759
op = FIP_DT_LOGO;
drivers/scsi/fcoe/fcoe_ctlr.c
778
op = fr_encaps(fp);
drivers/scsi/fcoe/fcoe_ctlr.c
779
if (op)
drivers/scsi/fcoe/fcoe_ctlr.c
789
op, ntoh24(fh->fh_d_id));
drivers/scsi/fcoe/fcoe_ctlr.c
790
if (fcoe_ctlr_encaps(fip, lport, op, skb, ntoh24(fh->fh_d_id)))
drivers/scsi/fcoe/fcoe_ctlr.c
796
op, ntoh24(fh->fh_d_id));
drivers/scsi/fnic/fip.c
693
u16 op;
drivers/scsi/fnic/fip.c
699
op = be16_to_cpu(fiph->fip_op);
drivers/scsi/fnic/fip.c
704
if (op == FIP_OP_DISC && sub == FIP_SC_REP)
drivers/scsi/fnic/fip.c
706
else if (op == FIP_OP_VLAN && sub == FIP_SC_REP)
drivers/scsi/fnic/fip.c
708
else if (op == FIP_OP_CTRL && sub == FIP_SC_REP)
drivers/scsi/fnic/fip.c
710
else if (op == FIP_OP_LS && sub == FIP_SC_REP)
drivers/scsi/fnic/fip.h
142
u16 op = be16_to_cpu(fiph->fip_op);
drivers/scsi/fnic/fip.h
147
pfx, op, sub, len);
drivers/scsi/isci/request.c
294
static void scu_ssp_ireq_dif_insert(struct isci_request *ireq, u8 type, u8 op)
drivers/scsi/isci/request.c
334
tc->bgctl_f.op = 0x2;
drivers/scsi/isci/request.c
349
static void scu_ssp_ireq_dif_strip(struct isci_request *ireq, u8 type, u8 op)
drivers/scsi/isci/request.c
394
tc->bgctl_f.op = 0x1;
drivers/scsi/isci/scu_task_context.h
927
u16 op:2;
drivers/scsi/libfc/fc_disc.c
180
u8 op;
drivers/scsi/libfc/fc_disc.c
183
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_disc.c
184
switch (op) {
drivers/scsi/libfc/fc_disc.c
192
"the opcode is (%x)\n", op);
drivers/scsi/libfc/fc_elsct.c
32
struct fc_frame *fp, unsigned int op,
drivers/scsi/libfc/fc_elsct.c
43
if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS))
drivers/scsi/libfc/fc_elsct.c
44
rc = fc_els_fill(lport, did, fp, op, &r_ctl, &fh_type);
drivers/scsi/libfc/fc_elsct.c
47
rc = fc_ct_fill(lport, did, fp, op, &r_ctl, &fh_type, &did);
drivers/scsi/libfc/fc_encode.h
101
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_gid_ft),
drivers/scsi/libfc/fc_encode.h
107
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_fid),
drivers/scsi/libfc/fc_encode.h
114
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_rft),
drivers/scsi/libfc/fc_encode.h
121
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_rff_id),
drivers/scsi/libfc/fc_encode.h
132
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_rn_id),
drivers/scsi/libfc/fc_encode.h
141
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_rspn) + len,
drivers/scsi/libfc/fc_encode.h
152
ct = fc_ct_hdr_fill(fp, op, sizeof(struct fc_ns_rsnn) + len,
drivers/scsi/libfc/fc_encode.h
189
unsigned int op, enum fc_rctl *r_ctl,
drivers/scsi/libfc/fc_encode.h
199
switch (op) {
drivers/scsi/libfc/fc_encode.h
228
ct = fc_ct_hdr_fill(fp, op, len, FC_FST_MGMT,
drivers/scsi/libfc/fc_encode.h
486
ct = fc_ct_hdr_fill(fp, op, len, FC_FST_MGMT,
drivers/scsi/libfc/fc_encode.h
65
unsigned int op, size_t req_size,
drivers/scsi/libfc/fc_encode.h
693
ct = fc_ct_hdr_fill(fp, op, len, FC_FST_MGMT,
drivers/scsi/libfc/fc_encode.h
701
ct = fc_ct_hdr_fill(fp, op, len, FC_FST_MGMT,
drivers/scsi/libfc/fc_encode.h
725
unsigned int op, enum fc_rctl *r_ctl,
drivers/scsi/libfc/fc_encode.h
732
rc = fc_ct_ms_fill(lport, fc_id, fp, op, r_ctl, fh_type);
drivers/scsi/libfc/fc_encode.h
737
rc = fc_ct_ns_fill(lport, fc_id, fp, op, r_ctl, fh_type);
drivers/scsi/libfc/fc_encode.h
748
unsigned int op)
drivers/scsi/libfc/fc_encode.h
756
plogi->fl_cmd = (u8) op;
drivers/scsi/libfc/fc_encode.h
78
ct->hdr.ct_cmd = htons((u16) op);
drivers/scsi/libfc/fc_encode.h
905
struct fc_frame *fp, unsigned int op,
drivers/scsi/libfc/fc_encode.h
908
switch (op) {
drivers/scsi/libfc/fc_encode.h
93
unsigned int op, enum fc_rctl *r_ctl,
drivers/scsi/libfc/fc_encode.h
99
switch (op) {
drivers/scsi/libfc/fc_exch.c
2104
unsigned int op;
drivers/scsi/libfc/fc_exch.c
2116
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_exch.c
2119
switch (op) {
drivers/scsi/libfc/fc_exch.c
2127
op);
drivers/scsi/libfc/fc_exch.c
221
static inline const char *fc_exch_name_lookup(unsigned int op, char **table,
drivers/scsi/libfc/fc_exch.c
226
if (op < max_index)
drivers/scsi/libfc/fc_exch.c
227
name = table[op];
drivers/scsi/libfc/fc_exch.c
237
static const char *fc_exch_rctl_name(unsigned int op)
drivers/scsi/libfc/fc_exch.c
239
return fc_exch_name_lookup(op, fc_exch_rctl_names,
drivers/scsi/libfc/fc_lport.c
1279
u8 op;
drivers/scsi/libfc/fc_lport.c
1301
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_lport.c
1302
if (op == ELS_LS_ACC)
drivers/scsi/libfc/fc_lport.c
1660
u8 op;
drivers/scsi/libfc/fc_lport.c
1682
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_lport.c
1683
if (op == ELS_LS_ACC)
drivers/scsi/libfc/fc_lport.c
357
unsigned int op)
drivers/scsi/libfc/fc_lport.c
363
flogi->fl_cmd = (u8) op;
drivers/scsi/libfc/fc_lport.c
373
if (op != ELS_FLOGI) {
drivers/scsi/libfc/fc_rport.c
1012
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_rport.c
1013
if (op == ELS_LS_ACC &&
drivers/scsi/libfc/fc_rport.c
1128
u8 op;
drivers/scsi/libfc/fc_rport.c
1154
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_rport.c
1155
if (op == ELS_LS_ACC) {
drivers/scsi/libfc/fc_rport.c
1322
u8 op;
drivers/scsi/libfc/fc_rport.c
1344
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_rport.c
1345
if (op == ELS_LS_ACC) {
drivers/scsi/libfc/fc_rport.c
1506
u8 op;
drivers/scsi/libfc/fc_rport.c
1533
op = fc_frame_payload_op(fp);
drivers/scsi/libfc/fc_rport.c
1535
if (op != ELS_LS_ACC || !adisc ||
drivers/scsi/libfc/fc_rport.c
990
u8 op;
drivers/scsi/lpfc/lpfc_debugfs.c
3283
if (debug->op == LPFC_IDIAG_OP_WR) {
drivers/scsi/lpfc/lpfc_debugfs.c
3342
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
3462
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
3637
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
3787
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
4294
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
4372
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
4668
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
4724
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
4886
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
4939
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
5079
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.c
5123
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
5397
debug->op = LPFC_IDIAG_OP_WR;
drivers/scsi/lpfc/lpfc_debugfs.c
5445
debug->op = LPFC_IDIAG_OP_RD;
drivers/scsi/lpfc/lpfc_debugfs.h
218
char op;
drivers/scsi/lpfc/lpfc_nvmet.c
1032
if (rsp->op == NVMET_FCOP_RSP)
drivers/scsi/lpfc/lpfc_nvmet.c
1081
ctxp->oxid, rsp->op, rsp->rsplen);
drivers/scsi/lpfc/lpfc_nvmet.c
1089
if (rsp->op == NVMET_FCOP_RSP)
drivers/scsi/lpfc/lpfc_nvmet.c
2094
if (ctxp->hdlrctx.fcp_req.op == NVMET_FCOP_RSP)
drivers/scsi/lpfc/lpfc_nvmet.c
2788
switch (rsp->op) {
drivers/scsi/lpfc/lpfc_nvmet.c
2851
if (rsp->op == NVMET_FCOP_READDATA_RSP) {
drivers/scsi/lpfc/lpfc_nvmet.c
3000
rsp->op);
drivers/scsi/lpfc/lpfc_nvmet.c
725
uint32_t status, result, op, logerr;
drivers/scsi/lpfc/lpfc_nvmet.c
735
op = rsp->op;
drivers/scsi/lpfc/lpfc_nvmet.c
746
ctxp->oxid, op, status);
drivers/scsi/lpfc/lpfc_nvmet.c
778
if (op == NVMET_FCOP_RSP)
drivers/scsi/lpfc/lpfc_nvmet.c
786
if ((op == NVMET_FCOP_READDATA_RSP) ||
drivers/scsi/lpfc/lpfc_nvmet.c
787
(op == NVMET_FCOP_RSP)) {
drivers/scsi/lpfc/lpfc_nvmet.c
794
if (rsp->op == NVMET_FCOP_READDATA_RSP) {
drivers/scsi/lpfc/lpfc_scsi.c
1030
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
1120
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
1151
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
1240
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
1272
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
1324
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
2418
unsigned char op = scsi_get_prot_op(sc);
drivers/scsi/lpfc/lpfc_scsi.c
2420
switch (op) {
drivers/scsi/lpfc/lpfc_scsi.c
2435
op);
drivers/scsi/lpfc/lpfc_scsi.c
969
uint32_t op = scsi_get_prot_op(sc);
drivers/scsi/lpfc/lpfc_scsi.c
976
if (op == SCSI_PROT_NORMAL)
drivers/scsi/myrb.c
206
enum myrb_cmd_opcode op, dma_addr_t addr)
drivers/scsi/myrb.c
215
mbox->type3.opcode = op;
drivers/scsi/myrb.c
228
enum myrb_cmd_opcode op, struct scsi_device *sdev,
drivers/scsi/myrb.c
245
mbox->type3D.opcode = op;
drivers/scsi/myrb.c
3329
enum myrb_cmd_opcode op;
drivers/scsi/myrb.c
3354
op = mbox->common.opcode;
drivers/scsi/myrb.c
3355
switch (op) {
drivers/scsi/ncr53c8xx.h
1088
#define SCR_SFBR_REG(reg,op,data) \
drivers/scsi/ncr53c8xx.h
1089
(0x68000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/ncr53c8xx.h
1091
#define SCR_REG_SFBR(reg,op,data) \
drivers/scsi/ncr53c8xx.h
1092
(0x70000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/ncr53c8xx.h
1094
#define SCR_REG_REG(reg,op,data) \
drivers/scsi/ncr53c8xx.h
1095
(0x78000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/pmcraid.h
994
({ u8 op = opcode; u8 __type = 0;\
drivers/scsi/pmcraid.h
995
if (op == READ_6 || op == READ_10 || op == READ_12 || op == READ_16)\
drivers/scsi/pmcraid.h
997
else if (op == WRITE_6 || op == WRITE_10 || op == WRITE_12 || \
drivers/scsi/pmcraid.h
998
op == WRITE_16)\
drivers/scsi/qedf/qedf.h
291
unsigned char op; /* SCSI CDB */
drivers/scsi/qedf/qedf.h
73
u8 op; /* Used to keep track of ELS op */
drivers/scsi/qedf/qedf_debugfs.c
265
seq_printf(s, "0x%02x:", io_log->op);
drivers/scsi/qedf/qedf_els.c
40
QEDF_ERR(&(qedf->dbg_ctx), "els 0x%x: rport not ready\n", op);
drivers/scsi/qedf/qedf_els.c
46
op);
drivers/scsi/qedf/qedf_els.c
466
if (cb_arg->op == ELS_ADISC)
drivers/scsi/qedf/qedf_els.c
52
QEDF_ERR(&(qedf->dbg_ctx), "els 0x%x: fcport not ready\n", op);
drivers/scsi/qedf/qedf_els.c
60
"Failed to alloc ELS request 0x%x\n", op);
drivers/scsi/qedf/qedf_els.c
73
cb_arg->op = op;
drivers/scsi/qedf/qedf_els.c
9
static int qedf_initiate_els(struct qedf_rport *fcport, unsigned int op,
drivers/scsi/qedf/qedf_els.c
91
if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) {
drivers/scsi/qedf/qedf_els.c
94
QEDF_ERR(&(qedf->dbg_ctx), "Invalid ELS op 0x%x\n", op);
drivers/scsi/qedf/qedf_fip.c
126
u16 op, vlan_tci = 0;
drivers/scsi/qedf/qedf_fip.c
138
op = ntohs(fiph->fip_op);
drivers/scsi/qedf/qedf_fip.c
151
"dest=%pM op=%x sub=%x vlan=%04x.", eth_hdr->h_dest, op, sub,
drivers/scsi/qedf/qedf_fip.c
177
u16 op;
drivers/scsi/qedf/qedf_fip.c
188
op = ntohs(fiph->fip_op);
drivers/scsi/qedf/qedf_fip.c
193
skb, fiph, eth_hdr->h_source, eth_hdr->h_dest, op,
drivers/scsi/qedf/qedf_fip.c
204
op, eth_hdr->h_dest, qedf->mac,
drivers/scsi/qedf/qedf_fip.c
211
if (op == FIP_OP_VLAN && sub == FIP_SC_VL_NOTE) {
drivers/scsi/qedf/qedf_fip.c
214
} else if (op == FIP_OP_CTRL && sub == FIP_SC_CLR_VLINK) {
drivers/scsi/qedf/qedf_io.c
815
io_log->op = sc_cmd->cmnd[0];
drivers/scsi/qedf/qedf_main.c
307
struct fc_frame *fp, unsigned int op,
drivers/scsi/qedf/qedf_main.c
333
return fc_elsct_send(lport, did, fp, op, qedf_flogi_resp,
drivers/scsi/qedf/qedf_main.c
337
return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
drivers/scsi/qedi/qedi.h
228
u8 op; /* SCSI CDB */
drivers/scsi/qedi/qedi_debugfs.c
189
seq_printf(s, "op-0x%02x:", io_log->op);
drivers/scsi/qedi/qedi_fw.c
1929
io_log->op = sc_cmd->cmnd[0];
drivers/scsi/qla2xxx/qla_attr.c
1969
static void qla_set_ini_mode(scsi_qla_host_t *vha, int op)
drivers/scsi/qla2xxx/qla_attr.c
1983
switch (op) {
drivers/scsi/qla2xxx/qla_attr.c
2051
switch (op) {
drivers/scsi/qla2xxx/qla_attr.c
2112
switch (op) {
drivers/scsi/qla2xxx/qla_attr.c
2135
switch (op) {
drivers/scsi/qla2xxx/qla_attr.c
2206
mode_to_str[vha->qlini_mode], mode_to_str[op],
drivers/scsi/qla2xxx/qla_attr.c
2210
vha->qlini_mode = op;
drivers/scsi/qla2xxx/qla_attr.c
2222
mode_to_str[vha->qlini_mode], mode_to_str[op],
drivers/scsi/qla2xxx/qla_attr.c
2225
vha->qlini_mode = op;
drivers/scsi/qla2xxx/qla_attr.c
2239
vha->qlini_mode, op,
drivers/scsi/qla2xxx/qla_target.c
1962
struct qla_tgt_sess_op *op;
drivers/scsi/qla2xxx/qla_target.c
1969
list_for_each_entry(op, &vha->unknown_atio_list, cmd_list) {
drivers/scsi/qla2xxx/qla_target.c
1973
op_key = sid_to_key(op->atio.u.isp24.fcp_hdr.s_id);
drivers/scsi/qla2xxx/qla_target.c
1975
(struct scsi_lun *)&op->atio.u.isp24.fcp_cmnd.lun);
drivers/scsi/qla2xxx/qla_target.c
1977
op->aborted = true;
drivers/scsi/qla2xxx/qla_target.c
4175
u8 op = cmd->cdb ? cmd->cdb[0] : 0;
drivers/scsi/qla2xxx/qla_target.c
4194
vha->vp_idx, cmd->se_cmd.tag, op,
drivers/scsi/qla2xxx/qla_target.c
4227
vha->vp_idx, cmd->se_cmd.tag, op,
drivers/scsi/qla2xxx/qla_target.c
4241
vha->vp_idx, cmd->se_cmd.tag, op,
drivers/scsi/qla2xxx/qla_target.c
4265
vha->vp_idx, cmd->se_cmd.tag, op, status,
drivers/scsi/qla2xxx/qla_target.c
4278
vha->vp_idx, cmd->se_cmd.tag, op, status,
drivers/scsi/qla2xxx/qla_target.c
4294
vha->vp_idx, cmd->se_cmd.tag, op, status,
drivers/scsi/qla2xxx/qla_target.c
4302
vha->vp_idx, cmd->se_cmd.tag, op, status,
drivers/scsi/qla2xxx/qla_target.c
4999
struct qla_tgt_sess_op *op;
drivers/scsi/qla2xxx/qla_target.c
5010
list_for_each_entry(op, &vha->unknown_atio_list, cmd_list) {
drivers/scsi/qla2xxx/qla_target.c
5011
uint32_t op_key = sid_to_key(op->atio.u.isp24.fcp_hdr.s_id);
drivers/scsi/qla2xxx/qla_target.c
5014
op->aborted = true;
drivers/scsi/qla2xxx/qla_target.c
5840
uint8_t op = cmd->cdb ? cmd->cdb[0] : 0;
drivers/scsi/qla2xxx/qla_target.c
5894
vha->vp_idx, se_cmd->tag, op,
drivers/scsi/qla2xxx/qla_target.c
5902
vha->vp_idx, se_cmd->tag, op);
drivers/scsi/qla2xxx/qla_target.c
5908
vha->vp_idx, se_cmd->tag, op,
drivers/scsi/qla2xxx/qla_target.c
5918
vha->vp_idx, se_cmd->tag, op, cmd->bufflen,
drivers/scsi/qla2xxx/qla_target.c
5949
vha->vp_idx, se_cmd->tag, op);
drivers/scsi/qla2xxx/qla_target.c
5967
vha->vp_idx, se_cmd->tag, op, cmd->bufflen,
drivers/scsi/qla2xxx/qla_target.c
5990
vha->vp_idx, se_cmd->tag, op);
drivers/scsi/qla2xxx/qla_target.c
6015
vha->vp_idx, se_cmd->tag, op, srr_ui, cmd->state,
drivers/scsi/qla2xxx/qla_target.c
8254
struct qla_tgt_sess_op *op = container_of(work,
drivers/scsi/qla2xxx/qla_target.c
8256
scsi_qla_host_t *vha = op->vha;
drivers/scsi/qla2xxx/qla_target.c
8261
(op->chip_reset != ha->base_qpair->chip_reset))
drivers/scsi/qla2xxx/qla_target.c
8269
qlt_response_pkt_all_vps(vha, op->rsp, (response_t *)&op->atio);
drivers/scsi/qla2xxx/qla_target.c
8272
kfree(op);
drivers/scsi/qla2xxx/qla_target.c
8279
struct qla_tgt_sess_op *op;
drivers/scsi/qla2xxx/qla_target.c
8281
op = kzalloc_obj(*op, GFP_ATOMIC);
drivers/scsi/qla2xxx/qla_target.c
8283
if (!op) {
drivers/scsi/qla2xxx/qla_target.c
8291
memcpy(&op->atio, pkt, sizeof(*pkt));
drivers/scsi/qla2xxx/qla_target.c
8292
op->vha = vha;
drivers/scsi/qla2xxx/qla_target.c
8293
op->chip_reset = vha->hw->base_qpair->chip_reset;
drivers/scsi/qla2xxx/qla_target.c
8294
op->rsp = rsp;
drivers/scsi/qla2xxx/qla_target.c
8295
INIT_WORK(&op->work, qlt_handle_abts_recv_work);
drivers/scsi/qla2xxx/qla_target.c
8296
queue_work(qla_tgt_wq, &op->work);
drivers/scsi/qlogicpti.c
1190
dma_unmap_sg(&qpti->op->dev,
drivers/scsi/qlogicpti.c
1305
static int qpti_sbus_probe(struct platform_device *op)
drivers/scsi/qlogicpti.c
1307
struct device_node *dp = op->dev.of_node;
drivers/scsi/qlogicpti.c
1316
if (op->archdata.irqs[0] == 0)
drivers/scsi/qlogicpti.c
1327
qpti->op = op;
drivers/scsi/qlogicpti.c
1329
qpti->is_pti = !of_node_name_eq(op->dev.of_node, "QLGC,isp");
drivers/scsi/qlogicpti.c
1373
if (scsi_add_host(host, &op->dev)) {
drivers/scsi/qlogicpti.c
1378
dev_set_drvdata(&op->dev, qpti);
drivers/scsi/qlogicpti.c
1389
dma_free_coherent(&op->dev,
drivers/scsi/qlogicpti.c
1392
dma_free_coherent(&op->dev,
drivers/scsi/qlogicpti.c
1401
of_iounmap(&op->resource[0], qpti->qregs,
drivers/scsi/qlogicpti.c
1402
resource_size(&op->resource[0]));
drivers/scsi/qlogicpti.c
1404
of_iounmap(&op->resource[0], qpti->sreg,
drivers/scsi/qlogicpti.c
1413
static void qpti_sbus_remove(struct platform_device *op)
drivers/scsi/qlogicpti.c
1415
struct qlogicpti *qpti = dev_get_drvdata(&op->dev);
drivers/scsi/qlogicpti.c
1428
dma_free_coherent(&op->dev,
drivers/scsi/qlogicpti.c
1431
dma_free_coherent(&op->dev,
drivers/scsi/qlogicpti.c
1436
of_iounmap(&op->resource[0], qpti->qregs,
drivers/scsi/qlogicpti.c
1437
resource_size(&op->resource[0]));
drivers/scsi/qlogicpti.c
1439
of_iounmap(&op->resource[0], qpti->sreg, sizeof(unsigned char));
drivers/scsi/qlogicpti.c
488
err = request_firmware(&fw, fwname, &qpti->op->dev);
drivers/scsi/qlogicpti.c
717
struct platform_device *op = qpti->op;
drivers/scsi/qlogicpti.c
719
qpti->qregs = of_ioremap(&op->resource[0], 0,
drivers/scsi/qlogicpti.c
720
resource_size(&op->resource[0]),
drivers/scsi/qlogicpti.c
727
qpti->sreg = of_ioremap(&op->resource[0], (16 * 4096),
drivers/scsi/qlogicpti.c
740
struct platform_device *op = qpti->op;
drivers/scsi/qlogicpti.c
742
qpti->qhost->irq = qpti->irq = op->archdata.irqs[0];
drivers/scsi/qlogicpti.c
765
struct platform_device *op = qpti->op;
drivers/scsi/qlogicpti.c
768
dp = op->dev.of_node;
drivers/scsi/qlogicpti.c
786
struct platform_device *op = qpti->op;
drivers/scsi/qlogicpti.c
789
bursts = of_getintprop_default(op->dev.of_node, "burst-sizes", 0xff);
drivers/scsi/qlogicpti.c
790
bmask = of_getintprop_default(op->dev.of_node->parent, "burst-sizes", 0xff);
drivers/scsi/qlogicpti.c
819
struct platform_device *op = qpti->op;
drivers/scsi/qlogicpti.c
822
qpti->res_cpu = dma_alloc_coherent(&op->dev,
drivers/scsi/qlogicpti.c
831
qpti->req_cpu = dma_alloc_coherent(&op->dev,
drivers/scsi/qlogicpti.c
836
dma_free_coherent(&op->dev, QSIZE(RES_QUEUE_LEN),
drivers/scsi/qlogicpti.c
909
sg_count = dma_map_sg(&qpti->op->dev, sg,
drivers/scsi/qlogicpti.h
346
struct platform_device *op;
drivers/scsi/scsi_lib.c
2159
enum req_op op = data_direction == DMA_TO_DEVICE ? REQ_OP_DRV_OUT :
drivers/scsi/scsi_lib.c
2164
rq = scsi_alloc_request(sdev->request_queue, op, flags);
drivers/scsi/scsi_transport_spi.c
108
enum req_op op, void *buffer, unsigned int bufflen,
drivers/scsi/scsi_transport_spi.c
111
blk_opf_t opf = op | REQ_FAILFAST_DEV | REQ_FAILFAST_TRANSPORT |
drivers/scsi/sd.h
239
unsigned char op, bool all);
drivers/scsi/sd.h
259
unsigned char op,
drivers/scsi/sd_zbc.c
325
unsigned char op, bool all)
drivers/scsi/sd_zbc.c
340
cmd->cmnd[1] = op;
drivers/scsi/sun_esp.c
116
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
118
host->irq = op->archdata.irqs[0];
drivers/scsi/sun_esp.c
124
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
127
dp = op->dev.of_node;
drivers/scsi/sun_esp.c
146
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
149
dp = op->dev.of_node;
drivers/scsi/sun_esp.c
158
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
162
dp = op->dev.of_node;
drivers/scsi/sun_esp.c
175
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
179
dp = op->dev.of_node;
drivers/scsi/sun_esp.c
226
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
262
sbus_set_sbus64(&op->dev, esp->bursts);
drivers/scsi/sun_esp.c
452
static int esp_sbus_probe_one(struct platform_device *op,
drivers/scsi/sun_esp.c
470
esp->dev = &op->dev;
drivers/scsi/sun_esp.c
504
dev_set_drvdata(&op->dev, esp);
drivers/scsi/sun_esp.c
515
dma_free_coherent(&op->dev, 16,
drivers/scsi/sun_esp.c
519
of_iounmap(&op->resource[(hme ? 1 : 0)], esp->regs, SBUS_ESP_REG_SIZE);
drivers/scsi/sun_esp.c
526
static int esp_sbus_probe(struct platform_device *op)
drivers/scsi/sun_esp.c
529
struct device_node *dp = op->dev.of_node;
drivers/scsi/sun_esp.c
538
dma_node = op->dev.of_node;
drivers/scsi/sun_esp.c
546
ret = esp_sbus_probe_one(op, dma_of, hme);
drivers/scsi/sun_esp.c
553
static void esp_sbus_remove(struct platform_device *op)
drivers/scsi/sun_esp.c
555
struct esp *esp = dev_get_drvdata(&op->dev);
drivers/scsi/sun_esp.c
571
dma_free_coherent(&op->dev, 16,
drivers/scsi/sun_esp.c
574
of_iounmap(&op->resource[(is_hme ? 1 : 0)], esp->regs,
drivers/scsi/sun_esp.c
581
dev_set_drvdata(&op->dev, NULL);
drivers/scsi/sun_esp.c
85
struct platform_device *op = to_platform_device(esp->dev);
drivers/scsi/sun_esp.c
92
res = &op->resource[1];
drivers/scsi/sun_esp.c
94
res = &op->resource[0];
drivers/scsi/sym53c8xx_2/sym_defs.h
571
#define SCR_SFBR_REG(reg,op,data) \
drivers/scsi/sym53c8xx_2/sym_defs.h
572
(0x68000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/sym53c8xx_2/sym_defs.h
574
#define SCR_REG_SFBR(reg,op,data) \
drivers/scsi/sym53c8xx_2/sym_defs.h
575
(0x70000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/sym53c8xx_2/sym_defs.h
577
#define SCR_REG_REG(reg,op,data) \
drivers/scsi/sym53c8xx_2/sym_defs.h
578
(0x78000000 | (SCR_REG_OFS(REG(reg))) | (op) | (((data)&0xff)<<8ul))
drivers/scsi/vmw_pvscsi.c
279
static int scsi_is_rw(unsigned char op)
drivers/scsi/vmw_pvscsi.c
281
return op == READ_6 || op == WRITE_6 ||
drivers/scsi/vmw_pvscsi.c
282
op == READ_10 || op == WRITE_10 ||
drivers/scsi/vmw_pvscsi.c
283
op == READ_12 || op == WRITE_12 ||
drivers/scsi/vmw_pvscsi.c
284
op == READ_16 || op == WRITE_16;
drivers/scsi/vmw_pvscsi.c
288
unsigned char op)
drivers/scsi/vmw_pvscsi.c
290
if (scsi_is_rw(op)) {
drivers/scsi/vmw_pvscsi.c
780
unsigned char op;
drivers/scsi/vmw_pvscsi.c
792
op = cmd->cmnd[0];
drivers/scsi/vmw_pvscsi.c
795
"queued cmd %p, ctx %p, op=%x\n", cmd, ctx, op);
drivers/scsi/vmw_pvscsi.c
799
pvscsi_kick_io(adapter, op);
drivers/scsi/xen-scsifront.c
1040
static void scsifront_do_lun_hotplug(struct vscsifrnt_info *info, int op)
drivers/scsi/xen-scsifront.c
1085
switch (op) {
drivers/soc/mediatek/mtk-cmdq-helper.c
212
.op = CMDQ_CODE_MASK,
drivers/soc/mediatek/mtk-cmdq-helper.c
221
.op = CMDQ_CODE_WRITE,
drivers/soc/mediatek/mtk-cmdq-helper.c
292
.op = CMDQ_CODE_READ_S,
drivers/soc/mediatek/mtk-cmdq-helper.c
306
.op = CMDQ_CODE_WRITE_S,
drivers/soc/mediatek/mtk-cmdq-helper.c
320
.op = CMDQ_CODE_WRITE_S_MASK,
drivers/soc/mediatek/mtk-cmdq-helper.c
340
.op = CMDQ_CODE_WRITE_S,
drivers/soc/mediatek/mtk-cmdq-helper.c
353
.op = CMDQ_CODE_WRITE_S_MASK,
drivers/soc/mediatek/mtk-cmdq-helper.c
400
.op = CMDQ_CODE_WFE,
drivers/soc/mediatek/mtk-cmdq-helper.c
415
.op = CMDQ_CODE_WFE,
drivers/soc/mediatek/mtk-cmdq-helper.c
430
.op = CMDQ_CODE_WFE,
drivers/soc/mediatek/mtk-cmdq-helper.c
445
.op = CMDQ_CODE_WFE,
drivers/soc/mediatek/mtk-cmdq-helper.c
461
.op = CMDQ_CODE_POLL,
drivers/soc/mediatek/mtk-cmdq-helper.c
47
u8 op;
drivers/soc/mediatek/mtk-cmdq-helper.c
50
static inline u8 cmdq_operand_get_type(struct cmdq_operand *op)
drivers/soc/mediatek/mtk-cmdq-helper.c
507
inst.op = CMDQ_CODE_MASK;
drivers/soc/mediatek/mtk-cmdq-helper.c
516
inst.op = CMDQ_CODE_POLL;
drivers/soc/mediatek/mtk-cmdq-helper.c
52
return op->reg ? CMDQ_REG_TYPE : CMDQ_IMMEDIATE_VALUE;
drivers/soc/mediatek/mtk-cmdq-helper.c
540
inst.op = CMDQ_CODE_LOGIC;
drivers/soc/mediatek/mtk-cmdq-helper.c
55
static inline u16 cmdq_operand_get_idx_value(struct cmdq_operand *op)
drivers/soc/mediatek/mtk-cmdq-helper.c
556
.op = CMDQ_CODE_LOGIC,
drivers/soc/mediatek/mtk-cmdq-helper.c
568
.op = CMDQ_CODE_JUMP,
drivers/soc/mediatek/mtk-cmdq-helper.c
57
return op->reg ? op->idx : op->value;
drivers/soc/mediatek/mtk-cmdq-helper.c
579
.op = CMDQ_CODE_JUMP,
drivers/soc/mediatek/mtk-cmdq-helper.c
589
.op = CMDQ_CODE_EOC,
drivers/spi/atmel-quadspi.c
296
int (*set_cfg)(struct atmel_qspi *aq, const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
298
int (*transfer)(struct spi_mem *mem, const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
443
static inline bool atmel_qspi_is_compatible(const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
446
if (op->cmd.buswidth != mode->cmd_buswidth)
drivers/spi/atmel-quadspi.c
449
if (op->addr.nbytes && op->addr.buswidth != mode->addr_buswidth)
drivers/spi/atmel-quadspi.c
452
if (op->data.nbytes && op->data.buswidth != mode->data_buswidth)
drivers/spi/atmel-quadspi.c
458
static int atmel_qspi_find_mode(const struct spi_mem_op *op)
drivers/spi/atmel-quadspi.c
463
if (atmel_qspi_is_compatible(op, &atmel_qspi_modes[i]))
drivers/spi/atmel-quadspi.c
469
static int atmel_qspi_sama7g5_find_mode(const struct spi_mem_op *op)
drivers/spi/atmel-quadspi.c
474
if (atmel_qspi_is_compatible(op, &atmel_qspi_sama7g5_modes[i]))
drivers/spi/atmel-quadspi.c
481
const struct spi_mem_op *op)
drivers/spi/atmel-quadspi.c
484
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/atmel-quadspi.c
488
if (atmel_qspi_sama7g5_find_mode(op) < 0)
drivers/spi/atmel-quadspi.c
494
if (atmel_qspi_find_mode(op) < 0)
drivers/spi/atmel-quadspi.c
498
if (op->addr.nbytes == 2 && op->cmd.buswidth != op->addr.buswidth &&
drivers/spi/atmel-quadspi.c
499
op->dummy.nbytes == 0)
drivers/spi/atmel-quadspi.c
525
const struct spi_mem_op *op, u32 *offset)
drivers/spi/atmel-quadspi.c
532
icr = QSPI_ICR_INST(op->cmd.opcode);
drivers/spi/atmel-quadspi.c
535
mode = atmel_qspi_find_mode(op);
drivers/spi/atmel-quadspi.c
540
if (op->dummy.nbytes)
drivers/spi/atmel-quadspi.c
541
dummy_cycles = op->dummy.nbytes * 8 / op->dummy.buswidth;
drivers/spi/atmel-quadspi.c
555
if (op->addr.buswidth) {
drivers/spi/atmel-quadspi.c
556
switch (op->addr.nbytes) {
drivers/spi/atmel-quadspi.c
561
icr |= QSPI_ICR_OPT(op->addr.val & 0xff);
drivers/spi/atmel-quadspi.c
564
if (dummy_cycles < 8 / op->addr.buswidth) {
drivers/spi/atmel-quadspi.c
567
iar = (op->cmd.opcode << 16) |
drivers/spi/atmel-quadspi.c
568
(op->addr.val & 0xffff);
drivers/spi/atmel-quadspi.c
571
iar = (op->addr.val << 8) & 0xffffff;
drivers/spi/atmel-quadspi.c
572
dummy_cycles -= 8 / op->addr.buswidth;
drivers/spi/atmel-quadspi.c
577
iar = op->addr.val & 0xffffff;
drivers/spi/atmel-quadspi.c
581
iar = op->addr.val & 0x7ffffff;
drivers/spi/atmel-quadspi.c
596
if (op->data.nbytes) {
drivers/spi/atmel-quadspi.c
599
if (op->addr.nbytes)
drivers/spi/atmel-quadspi.c
611
if (op->addr.nbytes && !op->data.nbytes)
drivers/spi/atmel-quadspi.c
615
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/atmel-quadspi.c
620
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/atmel-quadspi.c
655
const struct spi_mem_op *op, u32 offset)
drivers/spi/atmel-quadspi.c
660
if (!op->data.nbytes)
drivers/spi/atmel-quadspi.c
668
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/atmel-quadspi.c
669
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
670
op->data.nbytes);
drivers/spi/atmel-quadspi.c
675
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
676
op->data.nbytes);
drivers/spi/atmel-quadspi.c
689
const struct spi_mem_op *op, u32 *offset)
drivers/spi/atmel-quadspi.c
695
icr = FIELD_PREP(QSPI_ICR_INST_MASK_SAMA7G5, op->cmd.opcode);
drivers/spi/atmel-quadspi.c
698
mode = atmel_qspi_sama7g5_find_mode(op);
drivers/spi/atmel-quadspi.c
703
if (op->dummy.buswidth && op->dummy.nbytes) {
drivers/spi/atmel-quadspi.c
704
if (op->addr.dtr && op->dummy.dtr && op->data.dtr)
drivers/spi/atmel-quadspi.c
705
ifr |= QSPI_IFR_NBDUM(op->dummy.nbytes * 8 /
drivers/spi/atmel-quadspi.c
706
(2 * op->dummy.buswidth));
drivers/spi/atmel-quadspi.c
708
ifr |= QSPI_IFR_NBDUM(op->dummy.nbytes * 8 /
drivers/spi/atmel-quadspi.c
709
op->dummy.buswidth);
drivers/spi/atmel-quadspi.c
712
if (op->addr.buswidth && op->addr.nbytes) {
drivers/spi/atmel-quadspi.c
713
ifr |= FIELD_PREP(QSPI_IFR_ADDRL_SAMA7G5, op->addr.nbytes - 1) |
drivers/spi/atmel-quadspi.c
715
iar = FIELD_PREP(QSPI_IAR_ADDR, op->addr.val);
drivers/spi/atmel-quadspi.c
718
if (op->addr.dtr && op->dummy.dtr && op->data.dtr) {
drivers/spi/atmel-quadspi.c
720
if (op->cmd.dtr)
drivers/spi/atmel-quadspi.c
726
if (op->cmd.buswidth == 8 || op->addr.buswidth == 8 ||
drivers/spi/atmel-quadspi.c
727
op->data.buswidth == 8)
drivers/spi/atmel-quadspi.c
734
if (op->data.nbytes) {
drivers/spi/atmel-quadspi.c
737
if (op->addr.nbytes)
drivers/spi/atmel-quadspi.c
749
if (op->addr.nbytes && !op->data.nbytes)
drivers/spi/atmel-quadspi.c
752
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/atmel-quadspi.c
756
if (op->data.nbytes)
drivers/spi/atmel-quadspi.c
758
op->data.nbytes),
drivers/spi/atmel-quadspi.c
812
const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
837
const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
862
const struct spi_mem_op *op, loff_t loff)
drivers/spi/atmel-quadspi.c
867
ret = spi_controller_dma_map_mem_op_data(mem->spi->controller, op,
drivers/spi/atmel-quadspi.c
872
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/atmel-quadspi.c
873
ret = atmel_qspi_dma_rx_xfer(mem, op, &sgt, loff);
drivers/spi/atmel-quadspi.c
875
ret = atmel_qspi_dma_tx_xfer(mem, op, &sgt, loff);
drivers/spi/atmel-quadspi.c
877
spi_controller_dma_unmap_mem_op_data(mem->spi->controller, op, &sgt);
drivers/spi/atmel-quadspi.c
883
const struct spi_mem_op *op, u32 offset)
drivers/spi/atmel-quadspi.c
890
if (!op->data.nbytes) {
drivers/spi/atmel-quadspi.c
901
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/atmel-quadspi.c
902
if (aq->rx_chan && op->addr.nbytes &&
drivers/spi/atmel-quadspi.c
903
op->data.nbytes > ATMEL_QSPI_DMA_MIN_BYTES) {
drivers/spi/atmel-quadspi.c
904
ret = atmel_qspi_dma_transfer(mem, op, offset);
drivers/spi/atmel-quadspi.c
908
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
909
op->data.nbytes);
drivers/spi/atmel-quadspi.c
912
if (op->addr.nbytes) {
drivers/spi/atmel-quadspi.c
920
if (aq->tx_chan && op->addr.nbytes &&
drivers/spi/atmel-quadspi.c
921
op->data.nbytes > ATMEL_QSPI_DMA_MIN_BYTES) {
drivers/spi/atmel-quadspi.c
922
ret = atmel_qspi_dma_transfer(mem, op, offset);
drivers/spi/atmel-quadspi.c
926
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
927
op->data.nbytes);
drivers/spi/atmel-quadspi.c
944
static int atmel_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/atmel-quadspi.c
955
if (op->addr.val + op->data.nbytes > aq->mmap_size)
drivers/spi/atmel-quadspi.c
958
if (op->addr.nbytes > 4)
drivers/spi/atmel-quadspi.c
965
err = aq->ops->set_cfg(aq, op, &offset);
drivers/spi/atmel-quadspi.c
969
err = aq->ops->transfer(mem, op, offset);
drivers/spi/spi-airoha-snfi.c
1000
op->data.buswidth);
drivers/spi/spi-airoha-snfi.c
486
static bool airoha_snand_is_page_ops(const struct spi_mem_op *op)
drivers/spi/spi-airoha-snfi.c
488
if (op->addr.nbytes != 2)
drivers/spi/spi-airoha-snfi.c
491
if (op->addr.buswidth != 1 && op->addr.buswidth != 2 &&
drivers/spi/spi-airoha-snfi.c
492
op->addr.buswidth != 4)
drivers/spi/spi-airoha-snfi.c
495
switch (op->data.dir) {
drivers/spi/spi-airoha-snfi.c
497
if (op->dummy.nbytes * BITS_PER_BYTE / op->dummy.buswidth > 0xf)
drivers/spi/spi-airoha-snfi.c
501
if (op->addr.buswidth == 4)
drivers/spi/spi-airoha-snfi.c
502
return op->data.buswidth == 4;
drivers/spi/spi-airoha-snfi.c
504
if (op->addr.buswidth == 2)
drivers/spi/spi-airoha-snfi.c
505
return op->data.buswidth == 2;
drivers/spi/spi-airoha-snfi.c
508
return op->data.buswidth == 4 || op->data.buswidth == 2 ||
drivers/spi/spi-airoha-snfi.c
509
op->data.buswidth == 1;
drivers/spi/spi-airoha-snfi.c
511
return !op->dummy.nbytes && op->addr.buswidth == 1 &&
drivers/spi/spi-airoha-snfi.c
512
(op->data.buswidth == 4 || op->data.buswidth == 1);
drivers/spi/spi-airoha-snfi.c
519
const struct spi_mem_op *op)
drivers/spi/spi-airoha-snfi.c
521
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-airoha-snfi.c
524
if (op->cmd.buswidth != 1)
drivers/spi/spi-airoha-snfi.c
527
if (airoha_snand_is_page_ops(op))
drivers/spi/spi-airoha-snfi.c
530
return (!op->addr.nbytes || op->addr.buswidth == 1) &&
drivers/spi/spi-airoha-snfi.c
531
(!op->dummy.nbytes || op->dummy.buswidth == 1) &&
drivers/spi/spi-airoha-snfi.c
532
(!op->data.nbytes || op->data.buswidth == 1);
drivers/spi/spi-airoha-snfi.c
933
const struct spi_mem_op *op)
drivers/spi/spi-airoha-snfi.c
942
op_len = op->cmd.nbytes;
drivers/spi/spi-airoha-snfi.c
943
addr_len = op->addr.nbytes;
drivers/spi/spi-airoha-snfi.c
944
dummy_len = op->dummy.nbytes;
drivers/spi/spi-airoha-snfi.c
951
*data++ = op->cmd.opcode >> (8 * (op_len - i - 1));
drivers/spi/spi-airoha-snfi.c
953
*data++ = op->addr.val >> (8 * (addr_len - i - 1));
drivers/spi/spi-airoha-snfi.c
969
op->cmd.buswidth);
drivers/spi/spi-airoha-snfi.c
977
op->addr.buswidth);
drivers/spi/spi-airoha-snfi.c
986
op->dummy.buswidth);
drivers/spi/spi-airoha-snfi.c
992
if (op->data.nbytes) {
drivers/spi/spi-airoha-snfi.c
993
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-airoha-snfi.c
994
err = airoha_snand_read_data(as_ctrl, op->data.buf.in,
drivers/spi/spi-airoha-snfi.c
995
op->data.nbytes,
drivers/spi/spi-airoha-snfi.c
996
op->data.buswidth);
drivers/spi/spi-airoha-snfi.c
998
err = airoha_snand_write_data(as_ctrl, op->data.buf.out,
drivers/spi/spi-airoha-snfi.c
999
op->data.nbytes,
drivers/spi/spi-amd.c
400
static inline bool amd_is_spi_read_cmd_4b(const u16 op)
drivers/spi/spi-amd.c
402
switch (op) {
drivers/spi/spi-amd.c
414
static inline bool amd_is_spi_read_cmd(const u16 op)
drivers/spi/spi-amd.c
416
switch (op) {
drivers/spi/spi-amd.c
425
return amd_is_spi_read_cmd_4b(op);
drivers/spi/spi-amd.c
429
static inline bool amd_is_spi_write_cmd(const u16 op)
drivers/spi/spi-amd.c
431
switch (op) {
drivers/spi/spi-amd.c
441
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
446
if (op->cmd.buswidth > 1 || op->addr.buswidth > 4)
drivers/spi/spi-amd.c
450
if (amd_is_spi_read_cmd(op->cmd.opcode) || amd_is_spi_write_cmd(op->cmd.opcode)) {
drivers/spi/spi-amd.c
451
if (op->data.buswidth > 4)
drivers/spi/spi-amd.c
459
if ((amd_is_spi_read_cmd_4b(op->cmd.opcode) ||
drivers/spi/spi-amd.c
460
amd_is_spi_write_cmd(op->cmd.opcode)) &&
drivers/spi/spi-amd.c
461
op->data.nbytes > AMD_SPI_HID2_DMA_SIZE)
drivers/spi/spi-amd.c
463
} else if (op->data.nbytes > AMD_SPI_MAX_DATA) {
drivers/spi/spi-amd.c
466
} else if (op->data.buswidth > 1 || op->data.nbytes > AMD_SPI_MAX_DATA) {
drivers/spi/spi-amd.c
470
if (op->max_freq < mem->spi->controller->min_speed_hz)
drivers/spi/spi-amd.c
473
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-amd.c
476
static int amd_spi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-amd.c
486
if (amd_spi->version == AMD_HID2_SPI && (amd_is_spi_read_cmd(op->cmd.opcode) ||
drivers/spi/spi-amd.c
487
amd_is_spi_write_cmd(op->cmd.opcode)))
drivers/spi/spi-amd.c
488
op->data.nbytes = clamp_val(op->data.nbytes, 0, AMD_SPI_HID2_DMA_SIZE);
drivers/spi/spi-amd.c
490
op->data.nbytes = clamp_val(op->data.nbytes, 0, AMD_SPI_MAX_DATA);
drivers/spi/spi-amd.c
496
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
498
u8 nbytes = op->addr.nbytes;
drivers/spi/spi-amd.c
499
u64 addr_val = op->addr.val;
drivers/spi/spi-amd.c
511
static void amd_spi_hiddma_write(struct amd_spi *amd_spi, const struct spi_mem_op *op)
drivers/spi/spi-amd.c
525
hid_regval = (hid_regval & ~GENMASK(15, 0)) | ((op->data.nbytes) + 3);
drivers/spi/spi-amd.c
542
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
544
int base_addr = AMD_SPI_FIFO_BASE + op->addr.nbytes;
drivers/spi/spi-amd.c
545
u64 *buf_64 = (u64 *)op->data.buf.out;
drivers/spi/spi-amd.c
546
u64 addr_val = op->addr.val;
drivers/spi/spi-amd.c
547
u32 nbytes = op->data.nbytes;
drivers/spi/spi-amd.c
556
if (amd_spi->version == AMD_HID2_SPI && amd_is_spi_write_cmd(op->cmd.opcode)) {
drivers/spi/spi-amd.c
557
u64 *dma_buf64 = (u64 *)(amd_spi->dma_virt_addr + op->addr.nbytes + op->cmd.nbytes);
drivers/spi/spi-amd.c
561
*dma_buf = op->cmd.opcode;
drivers/spi/spi-amd.c
564
for (i = 0; i < op->addr.nbytes; i++) {
drivers/spi/spi-amd.c
580
amd_spi_hiddma_write(amd_spi, op);
drivers/spi/spi-amd.c
582
amd_spi_set_opcode(amd_spi, op->cmd.opcode);
drivers/spi/spi-amd.c
583
amd_spi_set_addr(amd_spi, op);
drivers/spi/spi-amd.c
586
amd_spi_writereg64(amd_spi, base_addr + op->dummy.nbytes + (i * 8),
drivers/spi/spi-amd.c
592
base_addr + op->dummy.nbytes + nbytes + i - left_data,
drivers/spi/spi-amd.c
596
amd_spi_set_tx_count(amd_spi, op->addr.nbytes + op->data.nbytes);
drivers/spi/spi-amd.c
603
static void amd_spi_hiddma_read(struct amd_spi *amd_spi, const struct spi_mem_op *op)
drivers/spi/spi-amd.c
610
hid_regval = (hid_regval & ~GENMASK(7, 0)) | op->cmd.opcode;
drivers/spi/spi-amd.c
618
hid_regval = (hid_regval & ~GENMASK(31, 8)) | (op->addr.val << 16);
drivers/spi/spi-amd.c
645
hid_regval = (hid_regval & ~GENMASK(15, 0)) | ((op->data.nbytes / 4) - 1);
drivers/spi/spi-amd.c
662
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
664
int base_addr = AMD_SPI_FIFO_BASE + op->addr.nbytes;
drivers/spi/spi-amd.c
665
u64 *buf_64 = (u64 *)op->data.buf.in;
drivers/spi/spi-amd.c
666
u32 nbytes = op->data.nbytes;
drivers/spi/spi-amd.c
676
if (amd_spi->version == AMD_HID2_SPI && amd_is_spi_read_cmd(op->cmd.opcode)) {
drivers/spi/spi-amd.c
680
amd_spi_hiddma_read(amd_spi, op);
drivers/spi/spi-amd.c
698
amd_spi_set_opcode(amd_spi, op->cmd.opcode);
drivers/spi/spi-amd.c
699
amd_spi_set_addr(amd_spi, op);
drivers/spi/spi-amd.c
700
amd_spi_set_tx_count(amd_spi, op->addr.nbytes + op->dummy.nbytes);
drivers/spi/spi-amd.c
702
for (i = 0; i < op->dummy.nbytes; i++)
drivers/spi/spi-amd.c
705
amd_spi_set_rx_count(amd_spi, op->data.nbytes);
drivers/spi/spi-amd.c
711
*buf_64++ = amd_spi_readreg64(amd_spi, base_addr + op->dummy.nbytes +
drivers/spi/spi-amd.c
716
buf[i] = amd_spi_readreg8(amd_spi, base_addr + op->dummy.nbytes +
drivers/spi/spi-amd.c
723
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
727
if (amd_is_spi_read_cmd_4b(op->cmd.opcode))
drivers/spi/spi-amd.c
734
const struct spi_mem_op *op)
drivers/spi/spi-amd.c
740
amd_set_spi_freq(amd_spi, op->max_freq);
drivers/spi/spi-amd.c
743
amd_set_spi_addr_mode(amd_spi, op);
drivers/spi/spi-amd.c
745
switch (op->data.dir) {
drivers/spi/spi-amd.c
747
amd_spi_mem_data_in(amd_spi, op);
drivers/spi/spi-amd.c
752
amd_spi_mem_data_out(amd_spi, op);
drivers/spi/spi-amlogic-spifc-a1.c
255
const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a1.c
259
size_t data_size = op->data.nbytes;
drivers/spi/spi-amlogic-spifc-a1.c
262
ret = amlogic_spifc_a1_set_freq(spifc, op->max_freq);
drivers/spi/spi-amlogic-spifc-a1.c
267
amlogic_spifc_a1_set_cmd(spifc, SPIFC_A1_USER_CMD(op));
drivers/spi/spi-amlogic-spifc-a1.c
269
if (op->addr.nbytes)
drivers/spi/spi-amlogic-spifc-a1.c
270
amlogic_spifc_a1_set_addr(spifc, op->addr.val,
drivers/spi/spi-amlogic-spifc-a1.c
271
SPIFC_A1_USER_ADDR(op));
drivers/spi/spi-amlogic-spifc-a1.c
273
if (op->dummy.nbytes)
drivers/spi/spi-amlogic-spifc-a1.c
274
amlogic_spifc_a1_set_dummy(spifc, SPIFC_A1_USER_DUMMY(op));
drivers/spi/spi-amlogic-spifc-a1.c
277
u32 mode = ilog2(op->data.buswidth);
drivers/spi/spi-amlogic-spifc-a1.c
281
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-amlogic-spifc-a1.c
282
ret = amlogic_spifc_a1_read(spifc, op->data.buf.in,
drivers/spi/spi-amlogic-spifc-a1.c
285
ret = amlogic_spifc_a1_write(spifc, op->data.buf.out,
drivers/spi/spi-amlogic-spifc-a1.c
295
struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a1.c
297
op->data.nbytes = min(op->data.nbytes, SPIFC_A1_BUFFER_SIZE);
drivers/spi/spi-amlogic-spifc-a1.c
80
#define SPIFC_A1_USER_CMD(op) ( \
drivers/spi/spi-amlogic-spifc-a1.c
82
FIELD_PREP(SPIFC_A1_USER_CMD_CODE, (op)->cmd.opcode) | \
drivers/spi/spi-amlogic-spifc-a1.c
83
FIELD_PREP(SPIFC_A1_USER_CMD_MODE, ilog2((op)->cmd.buswidth)))
drivers/spi/spi-amlogic-spifc-a1.c
85
#define SPIFC_A1_USER_ADDR(op) ( \
drivers/spi/spi-amlogic-spifc-a1.c
87
FIELD_PREP(SPIFC_A1_USER_ADDR_MODE, ilog2((op)->addr.buswidth)) | \
drivers/spi/spi-amlogic-spifc-a1.c
88
FIELD_PREP(SPIFC_A1_USER_ADDR_BYTES, (op)->addr.nbytes - 1))
drivers/spi/spi-amlogic-spifc-a1.c
90
#define SPIFC_A1_USER_DUMMY(op) ( \
drivers/spi/spi-amlogic-spifc-a1.c
92
FIELD_PREP(SPIFC_A1_USER_DUMMY_MODE, ilog2((op)->dummy.buswidth)) | \
drivers/spi/spi-amlogic-spifc-a1.c
93
FIELD_PREP(SPIFC_A1_USER_DUMMY_CLK_SYCLES, (op)->dummy.nbytes << 3))
drivers/spi/spi-amlogic-spifc-a4.c
298
static int aml_sfc_send_cmd(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
303
ret = aml_sfc_set_bus_width(sfc, op->cmd.buswidth, CMD_LANE);
drivers/spi/spi-amlogic-spifc-a4.c
307
for (i = 0; i < op->cmd.nbytes; i++) {
drivers/spi/spi-amlogic-spifc-a4.c
308
val = (op->cmd.opcode >> ((op->cmd.nbytes - i - 1) * 8)) & 0xff;
drivers/spi/spi-amlogic-spifc-a4.c
317
static int aml_sfc_send_addr(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
322
ret = aml_sfc_set_bus_width(sfc, op->addr.buswidth, ADDR_LANE);
drivers/spi/spi-amlogic-spifc-a4.c
326
for (i = 0; i < op->addr.nbytes; i++) {
drivers/spi/spi-amlogic-spifc-a4.c
327
val = (op->addr.val >> ((op->addr.nbytes - i - 1) * 8)) & 0xff;
drivers/spi/spi-amlogic-spifc-a4.c
337
static bool aml_sfc_is_xio_op(const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
339
switch (op->cmd.opcode) {
drivers/spi/spi-amlogic-spifc-a4.c
351
static int aml_sfc_send_cmd_addr_dummy(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
356
ret = aml_sfc_send_cmd(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
360
ret = aml_sfc_send_addr(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
364
if (op->dummy.nbytes) {
drivers/spi/spi-amlogic-spifc-a4.c
366
if (aml_sfc_is_xio_op(op))
drivers/spi/spi-amlogic-spifc-a4.c
367
dummy_cycle = op->dummy.nbytes * 8 / op->data.buswidth;
drivers/spi/spi-amlogic-spifc-a4.c
369
dummy_cycle = op->dummy.nbytes * 8;
drivers/spi/spi-amlogic-spifc-a4.c
377
static bool aml_sfc_is_snand_hwecc_page_op(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
379
switch (op->cmd.opcode) {
drivers/spi/spi-amlogic-spifc-a4.c
479
static void *aml_get_dma_safe_input_buf(const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
481
if (aml_sfc_dma_buffer_is_safe(op->data.buf.in))
drivers/spi/spi-amlogic-spifc-a4.c
482
return op->data.buf.in;
drivers/spi/spi-amlogic-spifc-a4.c
484
return kzalloc(op->data.nbytes, GFP_KERNEL);
drivers/spi/spi-amlogic-spifc-a4.c
487
static void aml_sfc_put_dma_safe_input_buf(const struct spi_mem_op *op, void *buf)
drivers/spi/spi-amlogic-spifc-a4.c
489
if (WARN_ON(op->data.dir != SPI_MEM_DATA_IN) || WARN_ON(!buf))
drivers/spi/spi-amlogic-spifc-a4.c
492
if (buf == op->data.buf.in)
drivers/spi/spi-amlogic-spifc-a4.c
495
memcpy(op->data.buf.in, buf, op->data.nbytes);
drivers/spi/spi-amlogic-spifc-a4.c
499
static void *aml_sfc_get_dma_safe_output_buf(const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
501
if (aml_sfc_dma_buffer_is_safe(op->data.buf.out))
drivers/spi/spi-amlogic-spifc-a4.c
502
return (void *)op->data.buf.out;
drivers/spi/spi-amlogic-spifc-a4.c
504
return kmemdup(op->data.buf.out, op->data.nbytes, GFP_KERNEL);
drivers/spi/spi-amlogic-spifc-a4.c
507
static void aml_sfc_put_dma_safe_output_buf(const struct spi_mem_op *op, const void *buf)
drivers/spi/spi-amlogic-spifc-a4.c
509
if (WARN_ON(op->data.dir != SPI_MEM_DATA_OUT) || WARN_ON(!buf))
drivers/spi/spi-amlogic-spifc-a4.c
512
if (buf != op->data.buf.out)
drivers/spi/spi-amlogic-spifc-a4.c
516
static u64 aml_sfc_cal_timeout_cycle(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
521
ms = 8 * MSEC_PER_SEC * op->data.nbytes / op->data.buswidth;
drivers/spi/spi-amlogic-spifc-a4.c
557
static int aml_sfc_raw_io_op(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
565
if (!op->data.nbytes)
drivers/spi/spi-amlogic-spifc-a4.c
568
conf = (op->data.nbytes >> RAW_SIZE_BW) << __ffs(RAW_EXT_SIZE);
drivers/spi/spi-amlogic-spifc-a4.c
573
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-amlogic-spifc-a4.c
576
buf = aml_get_dma_safe_input_buf(op);
drivers/spi/spi-amlogic-spifc-a4.c
582
cmd |= CMD_NAND2MEM(0, (op->data.nbytes & RAW_SIZE));
drivers/spi/spi-amlogic-spifc-a4.c
583
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-amlogic-spifc-a4.c
586
buf = aml_sfc_get_dma_safe_output_buf(op);
drivers/spi/spi-amlogic-spifc-a4.c
592
cmd |= CMD_MEM2NAND(0, (op->data.nbytes & RAW_SIZE));
drivers/spi/spi-amlogic-spifc-a4.c
597
ret = aml_sfc_dma_buffer_setup(sfc, buf, op->data.nbytes,
drivers/spi/spi-amlogic-spifc-a4.c
608
timeout_ms = aml_sfc_cal_timeout_cycle(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
616
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-amlogic-spifc-a4.c
617
aml_sfc_put_dma_safe_input_buf(op, buf);
drivers/spi/spi-amlogic-spifc-a4.c
618
else if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-amlogic-spifc-a4.c
619
aml_sfc_put_dma_safe_output_buf(op, buf);
drivers/spi/spi-amlogic-spifc-a4.c
621
aml_sfc_dma_buffer_release(sfc, op->data.nbytes,
drivers/spi/spi-amlogic-spifc-a4.c
690
static int aml_sfc_read_page_hwecc(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
713
timeout_ms = aml_sfc_cal_timeout_cycle(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
729
memcpy(op->data.buf.in, sfc->data_buf, page_size);
drivers/spi/spi-amlogic-spifc-a4.c
731
aml_sfc_get_user_byte(sfc, sfc->info_buf, op->data.buf.in);
drivers/spi/spi-amlogic-spifc-a4.c
733
memcpy(op->data.buf.in, sfc->data_buf, page_size);
drivers/spi/spi-amlogic-spifc-a4.c
734
aml_sfc_get_user_byte(sfc, sfc->info_buf, op->data.buf.in + page_size);
drivers/spi/spi-amlogic-spifc-a4.c
743
static int aml_sfc_write_page_hwecc(struct aml_sfc *sfc, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
757
memcpy(sfc->data_buf, op->data.buf.out, page_size);
drivers/spi/spi-amlogic-spifc-a4.c
762
(u8 *)op->data.buf.out + page_size, 1);
drivers/spi/spi-amlogic-spifc-a4.c
765
(u8 *)op->data.buf.out + page_size, 0);
drivers/spi/spi-amlogic-spifc-a4.c
778
timeout_ms = aml_sfc_cal_timeout_cycle(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
792
static int aml_sfc_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
805
op->cmd.opcode, op->addr.val, op->addr.buswidth, op->addr.nbytes,
drivers/spi/spi-amlogic-spifc-a4.c
806
op->dummy.buswidth, op->dummy.nbytes, op->data.buswidth, op->data.nbytes);
drivers/spi/spi-amlogic-spifc-a4.c
812
ret = aml_sfc_send_cmd_addr_dummy(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
816
ret = aml_sfc_set_bus_width(sfc, op->data.buswidth, DATA_LANE);
drivers/spi/spi-amlogic-spifc-a4.c
820
if (aml_sfc_is_snand_hwecc_page_op(sfc, op) &&
drivers/spi/spi-amlogic-spifc-a4.c
822
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-amlogic-spifc-a4.c
823
return aml_sfc_read_page_hwecc(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
825
return aml_sfc_write_page_hwecc(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
828
return aml_sfc_raw_io_op(sfc, op);
drivers/spi/spi-amlogic-spifc-a4.c
831
static int aml_sfc_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-amlogic-spifc-a4.c
839
if (aml_sfc_is_snand_hwecc_page_op(sfc, op) && ecc_cfg) {
drivers/spi/spi-amlogic-spifc-a4.c
840
if (op->data.nbytes > ecc_cfg->stepsize * ECC_BCH_MAX_SECT_SIZE)
drivers/spi/spi-amlogic-spifc-a4.c
842
} else if (op->data.nbytes & ~RAW_MAX_RW_SIZE_MASK) {
drivers/spi/spi-aspeed-smc.c
116
static u32 aspeed_spi_get_io_mode(const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
118
switch (op->data.buswidth) {
drivers/spi/spi-aspeed-smc.c
218
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
221
aspeed_spi_write_to_ahb(chip->ahb_base, &op->cmd.opcode, 1);
drivers/spi/spi-aspeed-smc.c
222
aspeed_spi_read_from_ahb(op->data.buf.in,
drivers/spi/spi-aspeed-smc.c
223
chip->ahb_base, op->data.nbytes);
drivers/spi/spi-aspeed-smc.c
229
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
232
aspeed_spi_write_to_ahb(chip->ahb_base, &op->cmd.opcode, 1);
drivers/spi/spi-aspeed-smc.c
233
aspeed_spi_write_to_ahb(chip->ahb_base, op->data.buf.out,
drivers/spi/spi-aspeed-smc.c
234
op->data.nbytes);
drivers/spi/spi-aspeed-smc.c
240
const struct spi_mem_op *op,
drivers/spi/spi-aspeed-smc.c
243
int io_mode = aspeed_spi_get_io_mode(op);
drivers/spi/spi-aspeed-smc.c
250
ret = aspeed_spi_send_cmd_addr(chip, op->addr.nbytes, offset, op->cmd.opcode);
drivers/spi/spi-aspeed-smc.c
254
if (op->dummy.buswidth && op->dummy.nbytes) {
drivers/spi/spi-aspeed-smc.c
255
for (i = 0; i < op->dummy.nbytes / op->dummy.buswidth; i++)
drivers/spi/spi-aspeed-smc.c
268
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
271
int io_mode = aspeed_spi_get_io_mode(op);
drivers/spi/spi-aspeed-smc.c
274
ret = aspeed_spi_send_cmd_addr(chip, op->addr.nbytes, op->addr.val, op->cmd.opcode);
drivers/spi/spi-aspeed-smc.c
280
aspeed_spi_write_to_ahb(chip->ahb_base, op->data.buf.out, op->data.nbytes);
drivers/spi/spi-aspeed-smc.c
288
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
290
if (op->cmd.buswidth > 1)
drivers/spi/spi-aspeed-smc.c
293
if (op->addr.nbytes != 0) {
drivers/spi/spi-aspeed-smc.c
294
if (op->addr.buswidth > 1)
drivers/spi/spi-aspeed-smc.c
296
if (op->addr.nbytes < 3 || op->addr.nbytes > 4)
drivers/spi/spi-aspeed-smc.c
300
if (op->dummy.nbytes != 0) {
drivers/spi/spi-aspeed-smc.c
301
if (op->dummy.buswidth > 1 || op->dummy.nbytes > 7)
drivers/spi/spi-aspeed-smc.c
305
if (op->data.nbytes != 0 && op->data.buswidth > 4)
drivers/spi/spi-aspeed-smc.c
308
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-aspeed-smc.c
314
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
328
ctl_val |= op->cmd.opcode << CTRL_COMMAND_SHIFT;
drivers/spi/spi-aspeed-smc.c
331
if (op->addr.nbytes) {
drivers/spi/spi-aspeed-smc.c
332
if (op->addr.nbytes == 4)
drivers/spi/spi-aspeed-smc.c
337
if (op->addr.nbytes == 4 && chip->aspi->data == &ast2400_spi_data)
drivers/spi/spi-aspeed-smc.c
341
if (op->dummy.nbytes)
drivers/spi/spi-aspeed-smc.c
342
ctl_val |= CTRL_IO_DUMMY_SET(op->dummy.nbytes / op->dummy.buswidth);
drivers/spi/spi-aspeed-smc.c
344
if (op->data.nbytes)
drivers/spi/spi-aspeed-smc.c
345
ctl_val |= aspeed_spi_get_io_mode(op);
drivers/spi/spi-aspeed-smc.c
347
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-aspeed-smc.c
356
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-aspeed-smc.c
357
if (!op->addr.nbytes)
drivers/spi/spi-aspeed-smc.c
358
ret = aspeed_spi_read_reg(chip, op);
drivers/spi/spi-aspeed-smc.c
360
ret = aspeed_spi_read_user(chip, op, op->addr.val,
drivers/spi/spi-aspeed-smc.c
361
op->data.nbytes, op->data.buf.in);
drivers/spi/spi-aspeed-smc.c
363
if (!op->addr.nbytes)
drivers/spi/spi-aspeed-smc.c
364
ret = aspeed_spi_write_reg(chip, op);
drivers/spi/spi-aspeed-smc.c
366
ret = aspeed_spi_write_user(chip, op);
drivers/spi/spi-aspeed-smc.c
377
const struct spi_mem_op *op)
drivers/spi/spi-aspeed-smc.c
381
ret = do_aspeed_spi_exec_mem_op(mem, op);
drivers/spi/spi-aspeed-smc.c
700
struct spi_mem_op *op = &desc->info.op_tmpl;
drivers/spi/spi-aspeed-smc.c
706
chip->cs, op->data.dir == SPI_MEM_DATA_IN ? "read" : "write",
drivers/spi/spi-aspeed-smc.c
708
op->cmd.opcode, op->cmd.buswidth, op->addr.buswidth,
drivers/spi/spi-aspeed-smc.c
709
op->dummy.buswidth, op->data.buswidth,
drivers/spi/spi-aspeed-smc.c
710
op->addr.nbytes, op->dummy.nbytes);
drivers/spi/spi-aspeed-smc.c
715
if (op->data.dir != SPI_MEM_DATA_IN)
drivers/spi/spi-aspeed-smc.c
726
ctl_val |= aspeed_spi_get_io_mode(op) |
drivers/spi/spi-aspeed-smc.c
727
op->cmd.opcode << CTRL_COMMAND_SHIFT |
drivers/spi/spi-aspeed-smc.c
730
if (op->dummy.nbytes)
drivers/spi/spi-aspeed-smc.c
731
ctl_val |= CTRL_IO_DUMMY_SET(op->dummy.nbytes / op->dummy.buswidth);
drivers/spi/spi-aspeed-smc.c
734
if (op->addr.nbytes) {
drivers/spi/spi-aspeed-smc.c
737
if (op->addr.nbytes == 4)
drivers/spi/spi-aspeed-smc.c
746
if (op->addr.nbytes == 4 && chip->aspi->data == &ast2400_spi_data)
drivers/spi/spi-aspeed-smc.c
757
chip->cs, op->data.buswidth, chip->ctl_val[ASPEED_SPI_READ]);
drivers/spi/spi-atcspi200.c
146
const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
148
void *rx_buf = op->data.buf.in;
drivers/spi/spi-atcspi200.c
149
const void *tx_buf = op->data.buf.out;
drivers/spi/spi-atcspi200.c
151
int trans_bytes = op->data.nbytes;
drivers/spi/spi-atcspi200.c
157
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-atcspi200.c
187
const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
191
if (op->cmd.nbytes)
drivers/spi/spi-atcspi200.c
193
if (op->addr.nbytes)
drivers/spi/spi-atcspi200.c
195
if (op->addr.buswidth > 1)
drivers/spi/spi-atcspi200.c
197
if (op->data.nbytes) {
drivers/spi/spi-atcspi200.c
200
width_code = ffs(op->data.buswidth) - 1;
drivers/spi/spi-atcspi200.c
207
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-atcspi200.c
208
if (op->dummy.nbytes)
drivers/spi/spi-atcspi200.c
210
TRANS_DUMMY_CNT(op->dummy.nbytes);
drivers/spi/spi-atcspi200.c
213
tc |= TRANS_RD_TRANS_CNT(op->data.nbytes);
drivers/spi/spi-atcspi200.c
216
TRANS_WR_TRANS_CNT(op->data.nbytes);
drivers/spi/spi-atcspi200.c
225
const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
230
if (op->data.nbytes) {
drivers/spi/spi-atcspi200.c
232
!(op->data.nbytes % 4)) {
drivers/spi/spi-atcspi200.c
242
TRANS_FMT_ADDR_LEN(op->addr.nbytes);
drivers/spi/spi-atcspi200.c
247
const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
249
atcspi_set_trans_fmt(spi, op);
drivers/spi/spi-atcspi200.c
250
atcspi_set_trans_ctl(spi, op);
drivers/spi/spi-atcspi200.c
251
if (op->addr.nbytes)
drivers/spi/spi-atcspi200.c
252
regmap_write(spi->regmap, ATCSPI_ADDR, op->addr.val);
drivers/spi/spi-atcspi200.c
253
regmap_write(spi->regmap, ATCSPI_CMD, op->cmd.opcode);
drivers/spi/spi-atcspi200.c
256
static int atcspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
261
op->data.nbytes = min(op->data.nbytes, ATCSPI_MAX_TRANS_LEN);
drivers/spi/spi-atcspi200.c
264
if (spi->use_dma && op->data.nbytes >= ATCSPI_DMA_THRESHOLD)
drivers/spi/spi-atcspi200.c
265
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 4);
drivers/spi/spi-atcspi200.c
306
const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
320
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-atcspi200.c
332
ret = spi_controller_dma_map_mem_op_data(spi->host, op, &sgt);
drivers/spi/spi-atcspi200.c
352
timeout = msecs_to_jiffies(ATCSPI_XFER_TIMEOUT(op->data.nbytes));
drivers/spi/spi-atcspi200.c
359
spi_controller_dma_unmap_mem_op_data(spi->host, op, &sgt);
drivers/spi/spi-atcspi200.c
364
static int atcspi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-atcspi200.c
373
atcspi_prepare_trans(spi, op);
drivers/spi/spi-atcspi200.c
374
if (op->data.nbytes) {
drivers/spi/spi-atcspi200.c
375
if (spi->use_dma && op->data.nbytes >= ATCSPI_DMA_THRESHOLD)
drivers/spi/spi-atcspi200.c
376
ret = atcspi_dma_trans(spi, op);
drivers/spi/spi-atcspi200.c
378
ret = atcspi_xfer_data_poll(spi, op);
drivers/spi/spi-ath79.c
138
const struct spi_mem_op *op)
drivers/spi/spi-ath79.c
147
if (op->cmd.opcode != 0x0b || op->data.dir != SPI_MEM_DATA_IN ||
drivers/spi/spi-ath79.c
148
op->addr.nbytes != 3 || op->dummy.nbytes != 1)
drivers/spi/spi-ath79.c
154
memcpy_fromio(op->data.buf.in, sp->base + op->addr.val, op->data.nbytes);
drivers/spi/spi-axiado.c
554
static int ax_spi_mem_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-axiado.c
564
u8 *rx_buf_ptr = (u8 *)op->data.buf.in;
drivers/spi/spi-axiado.c
565
u8 *tx_buf_ptr = (u8 *)op->data.buf.out;
drivers/spi/spi-axiado.c
570
__func__, op->cmd.opcode, op->cmd.buswidth, op->addr.buswidth,
drivers/spi/spi-axiado.c
571
op->dummy.buswidth, op->data.buswidth, op->addr.val,
drivers/spi/spi-axiado.c
572
op->data.nbytes);
drivers/spi/spi-axiado.c
575
if (op->cmd.buswidth != 1 ||
drivers/spi/spi-axiado.c
576
(op->addr.nbytes && op->addr.buswidth != 0 &&
drivers/spi/spi-axiado.c
577
op->addr.buswidth != 1) ||
drivers/spi/spi-axiado.c
578
(op->dummy.nbytes && op->dummy.buswidth != 0 &&
drivers/spi/spi-axiado.c
579
op->dummy.buswidth != 1) ||
drivers/spi/spi-axiado.c
580
(op->data.nbytes && op->data.buswidth != 1)) {
drivers/spi/spi-axiado.c
592
if (op->cmd.nbytes == 2) {
drivers/spi/spi-axiado.c
593
cmd_buf[cmd_len++] = (op->cmd.opcode >> 8) & 0xFF;
drivers/spi/spi-axiado.c
594
cmd_buf[cmd_len++] = op->cmd.opcode & 0xFF;
drivers/spi/spi-axiado.c
596
cmd_buf[cmd_len++] = op->cmd.opcode;
drivers/spi/spi-axiado.c
600
if (op->addr.nbytes) {
drivers/spi/spi-axiado.c
601
for (i = op->addr.nbytes - 1; i >= 0; i--) {
drivers/spi/spi-axiado.c
602
cmd_buf[cmd_len] = (op->addr.val >> (i * 8)) & 0xFF;
drivers/spi/spi-axiado.c
617
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-axiado.c
618
for (i = 0; i < op->dummy.nbytes; i++)
drivers/spi/spi-axiado.c
620
for (i = 0; i < op->data.nbytes; i++)
drivers/spi/spi-axiado.c
623
for (i = 0; i < op->data.nbytes; i++)
drivers/spi/spi-axiado.c
644
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-axiado.c
654
if (rx_count_reg >= op->data.nbytes)
drivers/spi/spi-axiado.c
668
bytes_to_discard_from_rx = op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-axiado.c
673
for (i = 0; i < op->data.nbytes; i++) {
drivers/spi/spi-axiado.c
677
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-axiado.c
699
static int ax_spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-axiado.c
709
protocol_overhead_bytes = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-axiado.c
723
if (op->data.nbytes > max_transfer_payload_bytes) {
drivers/spi/spi-axiado.c
724
op->data.nbytes = max_transfer_payload_bytes;
drivers/spi/spi-axiado.c
726
__func__, __LINE__, op->data.nbytes);
drivers/spi/spi-axiado.c
730
if (op->data.nbytes > SZ_64K) {
drivers/spi/spi-axiado.c
731
op->data.nbytes = SZ_64K;
drivers/spi/spi-axiado.c
733
__func__, __LINE__, op->data.nbytes);
drivers/spi/spi-bcm-qspi.c
1036
const struct spi_mem_op *op)
drivers/spi/spi-bcm-qspi.c
1045
if (op->addr.nbytes == BSPI_ADDRLEN_4BYTES)
drivers/spi/spi-bcm-qspi.c
1048
from = op->addr.val;
drivers/spi/spi-bcm-qspi.c
1075
len = op->data.nbytes;
drivers/spi/spi-bcm-qspi.c
1087
qspi->bspi_rf_op = op;
drivers/spi/spi-bcm-qspi.c
1151
const struct spi_mem_op *op)
drivers/spi/spi-bcm-qspi.c
1164
cmd[0] = op->cmd.opcode;
drivers/spi/spi-bcm-qspi.c
1165
for (i = 0; i < op->addr.nbytes; i++)
drivers/spi/spi-bcm-qspi.c
1166
cmd[1 + i] = op->addr.val >> (8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-bcm-qspi.c
1169
t[0].len = op->addr.nbytes + op->dummy.nbytes + 1;
drivers/spi/spi-bcm-qspi.c
1171
t[0].tx_nbits = op->cmd.buswidth;
drivers/spi/spi-bcm-qspi.c
1180
t[1].rx_buf = op->data.buf.in;
drivers/spi/spi-bcm-qspi.c
1181
t[1].len = op->data.nbytes;
drivers/spi/spi-bcm-qspi.c
1182
t[1].rx_nbits = op->data.buswidth;
drivers/spi/spi-bcm-qspi.c
1191
const struct spi_mem_op *op)
drivers/spi/spi-bcm-qspi.c
1200
if (!op->data.nbytes || !op->addr.nbytes || op->addr.nbytes > 4 ||
drivers/spi/spi-bcm-qspi.c
1201
op->data.dir != SPI_MEM_DATA_IN)
drivers/spi/spi-bcm-qspi.c
1204
buf = op->data.buf.in;
drivers/spi/spi-bcm-qspi.c
1205
addr = op->addr.val;
drivers/spi/spi-bcm-qspi.c
1206
len = op->data.nbytes;
drivers/spi/spi-bcm-qspi.c
1224
len < 4 || op->cmd.opcode == SPINOR_OP_RDSFDP)
drivers/spi/spi-bcm-qspi.c
1228
return bcm_qspi_mspi_exec_mem_op(spi, op);
drivers/spi/spi-bcm-qspi.c
1230
ret = bcm_qspi_bspi_set_mode(qspi, op, 0);
drivers/spi/spi-bcm-qspi.c
1233
ret = bcm_qspi_bspi_exec_mem_op(spi, op);
drivers/spi/spi-bcm-qspi.c
417
const struct spi_mem_op *op, int hp)
drivers/spi/spi-bcm-qspi.c
420
u8 command = op->cmd.opcode;
drivers/spi/spi-bcm-qspi.c
421
int width = op->data.buswidth ? op->data.buswidth : SPI_NBITS_SINGLE;
drivers/spi/spi-bcm-qspi.c
422
int addrlen = op->addr.nbytes;
drivers/spi/spi-bcm-qspi.c
431
if (op->dummy.nbytes)
drivers/spi/spi-bcm-qspi.c
432
bpp |= (op->dummy.nbytes * 8) / op->dummy.buswidth;
drivers/spi/spi-bcm-qspi.c
464
const struct spi_mem_op *op, int hp)
drivers/spi/spi-bcm-qspi.c
466
int width = op->data.buswidth ? op->data.buswidth : SPI_NBITS_SINGLE;
drivers/spi/spi-bcm-qspi.c
467
int addrlen = op->addr.nbytes;
drivers/spi/spi-bcm-qspi.c
503
bcm_qspi_bspi_set_xfer_params(qspi, op->cmd.opcode, 0, 0, 0);
drivers/spi/spi-bcm-qspi.c
509
const struct spi_mem_op *op, int hp)
drivers/spi/spi-bcm-qspi.c
512
int width = op->data.buswidth ? op->data.buswidth : SPI_NBITS_SINGLE;
drivers/spi/spi-bcm-qspi.c
513
int addrlen = op->addr.nbytes;
drivers/spi/spi-bcm-qspi.c
526
error = bcm_qspi_bspi_set_override(qspi, op, hp);
drivers/spi/spi-bcm-qspi.c
531
error = bcm_qspi_bspi_set_flex_mode(qspi, op, hp);
drivers/spi/spi-bcm63xx-hsspi.c
715
const struct spi_mem_op *op)
drivers/spi/spi-bcm63xx-hsspi.c
717
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-cadence-quadspi.c
1018
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1026
ret = cqspi_enable_dtr(f_pdata, op, CQSPI_REG_OP_EXT_WRITE_LSB);
drivers/spi/spi-cadence-quadspi.c
1030
if (op->cmd.dtr)
drivers/spi/spi-cadence-quadspi.c
1031
opcode = op->cmd.opcode >> 8;
drivers/spi/spi-cadence-quadspi.c
1033
opcode = op->cmd.opcode;
drivers/spi/spi-cadence-quadspi.c
1037
reg |= CQSPI_OP_WIDTH(op->data) << CQSPI_REG_WR_INSTR_TYPE_DATA_LSB;
drivers/spi/spi-cadence-quadspi.c
1038
reg |= CQSPI_OP_WIDTH(op->addr) << CQSPI_REG_WR_INSTR_TYPE_ADDR_LSB;
drivers/spi/spi-cadence-quadspi.c
1040
reg = cqspi_calc_rdreg(op);
drivers/spi/spi-cadence-quadspi.c
1068
reg |= (op->addr.nbytes - 1);
drivers/spi/spi-cadence-quadspi.c
1326
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1329
loff_t to = op->addr.val;
drivers/spi/spi-cadence-quadspi.c
1330
size_t len = op->data.nbytes;
drivers/spi/spi-cadence-quadspi.c
1331
const u_char *buf = op->data.buf.out;
drivers/spi/spi-cadence-quadspi.c
1334
ret = cqspi_write_setup(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1346
if ((!op->cmd.dtr && cqspi->use_direct_mode &&
drivers/spi/spi-cadence-quadspi.c
1423
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1427
loff_t from = op->addr.val;
drivers/spi/spi-cadence-quadspi.c
1428
size_t len = op->data.nbytes;
drivers/spi/spi-cadence-quadspi.c
1429
u_char *buf = op->data.buf.in;
drivers/spi/spi-cadence-quadspi.c
1433
ret = cqspi_read_setup(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1448
static int cqspi_mem_process(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1454
cqspi_configure(f_pdata, op->max_freq);
drivers/spi/spi-cadence-quadspi.c
1456
if (op->data.dir == SPI_MEM_DATA_IN && op->data.buf.in) {
drivers/spi/spi-cadence-quadspi.c
1462
if (!op->addr.nbytes ||
drivers/spi/spi-cadence-quadspi.c
1463
(op->data.nbytes <= CQSPI_STIG_DATA_LEN_MAX &&
drivers/spi/spi-cadence-quadspi.c
1465
return cqspi_command_read(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1467
return cqspi_read(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1470
if (!op->addr.nbytes || !op->data.buf.out)
drivers/spi/spi-cadence-quadspi.c
1471
return cqspi_command_write(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1473
return cqspi_write(f_pdata, op);
drivers/spi/spi-cadence-quadspi.c
1476
static int cqspi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1505
ret = cqspi_mem_process(mem, op);
drivers/spi/spi-cadence-quadspi.c
1521
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
1530
all_true = op->cmd.dtr &&
drivers/spi/spi-cadence-quadspi.c
1531
(!op->addr.nbytes || op->addr.dtr) &&
drivers/spi/spi-cadence-quadspi.c
1532
(!op->dummy.nbytes || op->dummy.dtr) &&
drivers/spi/spi-cadence-quadspi.c
1533
(!op->data.nbytes || op->data.dtr);
drivers/spi/spi-cadence-quadspi.c
1535
all_false = !op->cmd.dtr && !op->addr.dtr && !op->dummy.dtr &&
drivers/spi/spi-cadence-quadspi.c
1536
!op->data.dtr;
drivers/spi/spi-cadence-quadspi.c
1540
if (op->cmd.nbytes && op->cmd.buswidth != 8)
drivers/spi/spi-cadence-quadspi.c
1542
if (op->addr.nbytes && op->addr.buswidth != 8)
drivers/spi/spi-cadence-quadspi.c
1544
if (op->data.nbytes && op->data.buswidth != 8)
drivers/spi/spi-cadence-quadspi.c
1548
if ((op->cmd.opcode >> 8) != (op->cmd.opcode & 0xFF))
drivers/spi/spi-cadence-quadspi.c
1558
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-cadence-quadspi.c
398
static unsigned int cqspi_calc_rdreg(const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
402
rdreg |= CQSPI_OP_WIDTH(op->cmd) << CQSPI_REG_RD_INSTR_TYPE_INSTR_LSB;
drivers/spi/spi-cadence-quadspi.c
403
rdreg |= CQSPI_OP_WIDTH(op->addr) << CQSPI_REG_RD_INSTR_TYPE_ADDR_LSB;
drivers/spi/spi-cadence-quadspi.c
404
rdreg |= CQSPI_OP_WIDTH(op->data) << CQSPI_REG_RD_INSTR_TYPE_DATA_LSB;
drivers/spi/spi-cadence-quadspi.c
409
static unsigned int cqspi_calc_dummy(const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
413
if (!op->dummy.nbytes)
drivers/spi/spi-cadence-quadspi.c
416
dummy_clk = op->dummy.nbytes * (8 / op->dummy.buswidth);
drivers/spi/spi-cadence-quadspi.c
417
if (op->cmd.dtr)
drivers/spi/spi-cadence-quadspi.c
481
const struct spi_mem_op *op,
drivers/spi/spi-cadence-quadspi.c
489
if (op->cmd.nbytes != 2)
drivers/spi/spi-cadence-quadspi.c
493
ext = op->cmd.opcode & 0xff;
drivers/spi/spi-cadence-quadspi.c
504
const struct spi_mem_op *op, unsigned int shift)
drivers/spi/spi-cadence-quadspi.c
517
if (op->cmd.dtr) {
drivers/spi/spi-cadence-quadspi.c
522
ret = cqspi_setup_opcode_ext(f_pdata, op, shift);
drivers/spi/spi-cadence-quadspi.c
539
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
543
u8 *rxbuf = op->data.buf.in;
drivers/spi/spi-cadence-quadspi.c
545
size_t n_rx = op->data.nbytes;
drivers/spi/spi-cadence-quadspi.c
552
status = cqspi_enable_dtr(f_pdata, op, CQSPI_REG_OP_EXT_STIG_LSB);
drivers/spi/spi-cadence-quadspi.c
563
if (op->cmd.dtr)
drivers/spi/spi-cadence-quadspi.c
564
opcode = op->cmd.opcode >> 8;
drivers/spi/spi-cadence-quadspi.c
566
opcode = op->cmd.opcode;
drivers/spi/spi-cadence-quadspi.c
570
rdreg = cqspi_calc_rdreg(op);
drivers/spi/spi-cadence-quadspi.c
573
dummy_clk = cqspi_calc_dummy(op);
drivers/spi/spi-cadence-quadspi.c
588
if (op->addr.nbytes) {
drivers/spi/spi-cadence-quadspi.c
590
reg |= ((op->addr.nbytes - 1) &
drivers/spi/spi-cadence-quadspi.c
594
writel(op->addr.val, reg_base + CQSPI_REG_CMDADDRESS);
drivers/spi/spi-cadence-quadspi.c
622
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
627
const u8 *txbuf = op->data.buf.out;
drivers/spi/spi-cadence-quadspi.c
628
size_t n_tx = op->data.nbytes;
drivers/spi/spi-cadence-quadspi.c
634
ret = cqspi_enable_dtr(f_pdata, op, CQSPI_REG_OP_EXT_STIG_LSB);
drivers/spi/spi-cadence-quadspi.c
645
reg = cqspi_calc_rdreg(op);
drivers/spi/spi-cadence-quadspi.c
648
if (op->cmd.dtr)
drivers/spi/spi-cadence-quadspi.c
649
opcode = op->cmd.opcode >> 8;
drivers/spi/spi-cadence-quadspi.c
651
opcode = op->cmd.opcode;
drivers/spi/spi-cadence-quadspi.c
655
if (op->addr.nbytes) {
drivers/spi/spi-cadence-quadspi.c
657
reg |= ((op->addr.nbytes - 1) &
drivers/spi/spi-cadence-quadspi.c
661
writel(op->addr.val, reg_base + CQSPI_REG_CMDADDRESS);
drivers/spi/spi-cadence-quadspi.c
691
const struct spi_mem_op *op)
drivers/spi/spi-cadence-quadspi.c
700
ret = cqspi_enable_dtr(f_pdata, op, CQSPI_REG_OP_EXT_READ_LSB);
drivers/spi/spi-cadence-quadspi.c
704
if (op->cmd.dtr)
drivers/spi/spi-cadence-quadspi.c
705
opcode = op->cmd.opcode >> 8;
drivers/spi/spi-cadence-quadspi.c
707
opcode = op->cmd.opcode;
drivers/spi/spi-cadence-quadspi.c
710
reg |= cqspi_calc_rdreg(op);
drivers/spi/spi-cadence-quadspi.c
713
dummy_clk = cqspi_calc_dummy(op);
drivers/spi/spi-cadence-quadspi.c
727
reg |= (op->addr.nbytes - 1);
drivers/spi/spi-cadence-xspi.c
154
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_1(op, data_phase) ( \
drivers/spi/spi-cadence-xspi.c
157
FIELD_PREP(CDNS_XSPI_CMD_P1_R1_ADDR0, (op)->addr.val & 0xff))
drivers/spi/spi-cadence-xspi.c
159
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_2(op) ( \
drivers/spi/spi-cadence-xspi.c
160
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR1, ((op)->addr.val >> 8) & 0xFF) | \
drivers/spi/spi-cadence-xspi.c
161
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR2, ((op)->addr.val >> 16) & 0xFF) | \
drivers/spi/spi-cadence-xspi.c
162
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR3, ((op)->addr.val >> 24) & 0xFF) | \
drivers/spi/spi-cadence-xspi.c
163
FIELD_PREP(CDNS_XSPI_CMD_P1_R2_ADDR4, ((op)->addr.val >> 32) & 0xFF))
drivers/spi/spi-cadence-xspi.c
165
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, modebytes) ( \
drivers/spi/spi-cadence-xspi.c
166
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_ADDR5, ((op)->addr.val >> 40) & 0xFF) | \
drivers/spi/spi-cadence-xspi.c
167
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_CMD, (op)->cmd.opcode) | \
drivers/spi/spi-cadence-xspi.c
169
FIELD_PREP(CDNS_XSPI_CMD_P1_R3_NUM_ADDR_BYTES, (op)->addr.nbytes))
drivers/spi/spi-cadence-xspi.c
171
#define CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_4(op, chipsel) ( \
drivers/spi/spi-cadence-xspi.c
172
FIELD_PREP(CDNS_XSPI_CMD_P1_R4_ADDR_IOS, ilog2((op)->addr.buswidth)) | \
drivers/spi/spi-cadence-xspi.c
173
FIELD_PREP(CDNS_XSPI_CMD_P1_R4_CMD_IOS, ilog2((op)->cmd.buswidth)) | \
drivers/spi/spi-cadence-xspi.c
176
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_1(op) \
drivers/spi/spi-cadence-xspi.c
179
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_2(op) \
drivers/spi/spi-cadence-xspi.c
180
FIELD_PREP(CDNS_XSPI_CMD_DSEQ_R2_DCNT_L, (op)->data.nbytes & 0xFFFF)
drivers/spi/spi-cadence-xspi.c
182
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op, dummybytes) ( \
drivers/spi/spi-cadence-xspi.c
184
((op)->data.nbytes >> 16) & 0xffff) | \
drivers/spi/spi-cadence-xspi.c
186
(op)->dummy.buswidth != 0 ? \
drivers/spi/spi-cadence-xspi.c
187
(((dummybytes) * 8) / (op)->dummy.buswidth) : \
drivers/spi/spi-cadence-xspi.c
190
#define CDNS_XSPI_CMD_FLD_DSEQ_CMD_4(op, chipsel) ( \
drivers/spi/spi-cadence-xspi.c
193
ilog2((op)->data.buswidth)) | \
drivers/spi/spi-cadence-xspi.c
195
((op)->data.dir == SPI_MEM_DATA_IN) ? \
drivers/spi/spi-cadence-xspi.c
676
const struct spi_mem_op *op,
drivers/spi/spi-cadence-xspi.c
682
int dummybytes = op->dummy.nbytes;
drivers/spi/spi-cadence-xspi.c
695
cmd_regs[1] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_1(op, data_phase);
drivers/spi/spi-cadence-xspi.c
696
cmd_regs[2] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_2(op);
drivers/spi/spi-cadence-xspi.c
698
cmd_regs[3] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, 1);
drivers/spi/spi-cadence-xspi.c
701
cmd_regs[3] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_3(op, 0);
drivers/spi/spi-cadence-xspi.c
703
cmd_regs[4] = CDNS_XSPI_CMD_FLD_P1_INSTR_CMD_4(op,
drivers/spi/spi-cadence-xspi.c
710
cmd_regs[1] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_1(op);
drivers/spi/spi-cadence-xspi.c
711
cmd_regs[2] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_2(op);
drivers/spi/spi-cadence-xspi.c
712
cmd_regs[3] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_3(op, dummybytes);
drivers/spi/spi-cadence-xspi.c
713
cmd_regs[4] = CDNS_XSPI_CMD_FLD_DSEQ_CMD_4(op,
drivers/spi/spi-cadence-xspi.c
716
cdns_xspi->in_buffer = op->data.buf.in;
drivers/spi/spi-cadence-xspi.c
717
cdns_xspi->out_buffer = op->data.buf.out;
drivers/spi/spi-cadence-xspi.c
741
const struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
743
enum spi_mem_data_dir dir = op->data.dir;
drivers/spi/spi-cadence-xspi.c
748
return cdns_xspi_send_stig_command(cdns_xspi, op,
drivers/spi/spi-cadence-xspi.c
753
const struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
759
ret = cdns_xspi_mem_op(cdns_xspi, mem, op);
drivers/spi/spi-cadence-xspi.c
765
const struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
773
ret = cdns_xspi_mem_op(cdns_xspi, mem, op);
drivers/spi/spi-cadence-xspi.c
779
const struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
823
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-cadence-xspi.c
829
static int cdns_xspi_adjust_mem_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-cadence-xspi.c
834
op->data.nbytes = clamp_val(op->data.nbytes, 0, cdns_xspi->sdmasize);
drivers/spi/spi-dw-core.c
491
static int dw_spi_adjust_mem_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
493
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-dw-core.c
494
op->data.nbytes = clamp_val(op->data.nbytes, 0, DW_SPI_NDF_MASK + 1);
drivers/spi/spi-dw-core.c
500
const struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
502
if (op->data.buswidth > 1 || op->addr.buswidth > 1 ||
drivers/spi/spi-dw-core.c
503
op->dummy.buswidth > 1 || op->cmd.buswidth > 1)
drivers/spi/spi-dw-core.c
506
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-dw-core.c
509
static int dw_spi_init_mem_buf(struct dw_spi *dws, const struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
518
len = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-dw-core.c
519
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-dw-core.c
520
len += op->data.nbytes;
drivers/spi/spi-dw-core.c
535
for (i = 0; i < op->cmd.nbytes; ++i)
drivers/spi/spi-dw-core.c
536
out[i] = DW_SPI_GET_BYTE(op->cmd.opcode, op->cmd.nbytes - i - 1);
drivers/spi/spi-dw-core.c
537
for (j = 0; j < op->addr.nbytes; ++i, ++j)
drivers/spi/spi-dw-core.c
538
out[i] = DW_SPI_GET_BYTE(op->addr.val, op->addr.nbytes - j - 1);
drivers/spi/spi-dw-core.c
539
for (j = 0; j < op->dummy.nbytes; ++i, ++j)
drivers/spi/spi-dw-core.c
542
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-dw-core.c
543
memcpy(&out[i], op->data.buf.out, op->data.nbytes);
drivers/spi/spi-dw-core.c
548
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-dw-core.c
549
dws->rx = op->data.buf.in;
drivers/spi/spi-dw-core.c
550
dws->rx_len = op->data.nbytes;
drivers/spi/spi-dw-core.c
675
static int dw_spi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-dw-core.c
686
ret = dw_spi_init_mem_buf(dws, op);
drivers/spi/spi-dw-core.c
695
cfg.freq = clamp(op->max_freq, 0U, dws->max_mem_freq);
drivers/spi/spi-dw-core.c
696
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-dw-core.c
698
cfg.ndf = op->data.nbytes;
drivers/spi/spi-fsl-qspi.c
391
const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
396
ret = fsl_qspi_check_buswidth(q, op->cmd.buswidth);
drivers/spi/spi-fsl-qspi.c
398
if (op->addr.nbytes)
drivers/spi/spi-fsl-qspi.c
399
ret |= fsl_qspi_check_buswidth(q, op->addr.buswidth);
drivers/spi/spi-fsl-qspi.c
401
if (op->dummy.nbytes)
drivers/spi/spi-fsl-qspi.c
402
ret |= fsl_qspi_check_buswidth(q, op->dummy.buswidth);
drivers/spi/spi-fsl-qspi.c
404
if (op->data.nbytes)
drivers/spi/spi-fsl-qspi.c
405
ret |= fsl_qspi_check_buswidth(q, op->data.buswidth);
drivers/spi/spi-fsl-qspi.c
414
if (op->addr.nbytes +
drivers/spi/spi-fsl-qspi.c
415
(op->dummy.nbytes ? 1:0) +
drivers/spi/spi-fsl-qspi.c
416
(op->data.nbytes ? 1:0) > 6)
drivers/spi/spi-fsl-qspi.c
420
if (op->dummy.nbytes &&
drivers/spi/spi-fsl-qspi.c
421
(op->dummy.nbytes * 8 / op->dummy.buswidth > 64))
drivers/spi/spi-fsl-qspi.c
425
if (op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-fsl-qspi.c
426
(op->data.nbytes > q->devtype_data->ahb_buf_size ||
drivers/spi/spi-fsl-qspi.c
427
(op->data.nbytes > q->devtype_data->rxfifo - 4 &&
drivers/spi/spi-fsl-qspi.c
428
!IS_ALIGNED(op->data.nbytes, 8))))
drivers/spi/spi-fsl-qspi.c
431
if (op->data.dir == SPI_MEM_DATA_OUT &&
drivers/spi/spi-fsl-qspi.c
432
op->data.nbytes > q->devtype_data->txfifo)
drivers/spi/spi-fsl-qspi.c
435
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-fsl-qspi.c
439
const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
445
lutval[0] |= LUT_DEF(0, LUT_CMD, LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-fsl-qspi.c
446
op->cmd.opcode);
drivers/spi/spi-fsl-qspi.c
453
for (i = 0; i < op->addr.nbytes; i++) {
drivers/spi/spi-fsl-qspi.c
454
u8 addrbyte = op->addr.val >> (8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-fsl-qspi.c
457
LUT_PAD(op->addr.buswidth),
drivers/spi/spi-fsl-qspi.c
462
if (op->dummy.nbytes) {
drivers/spi/spi-fsl-qspi.c
464
LUT_PAD(op->dummy.buswidth),
drivers/spi/spi-fsl-qspi.c
465
op->dummy.nbytes * 8 /
drivers/spi/spi-fsl-qspi.c
466
op->dummy.buswidth);
drivers/spi/spi-fsl-qspi.c
470
if (op->data.nbytes) {
drivers/spi/spi-fsl-qspi.c
472
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-fsl-qspi.c
474
LUT_PAD(op->data.buswidth),
drivers/spi/spi-fsl-qspi.c
549
const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
551
unsigned long rate = op->max_freq;
drivers/spi/spi-fsl-qspi.c
578
static void fsl_qspi_read_ahb(struct fsl_qspi *q, const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
580
memcpy_fromio(op->data.buf.in,
drivers/spi/spi-fsl-qspi.c
582
op->data.nbytes);
drivers/spi/spi-fsl-qspi.c
586
const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
592
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 4); i += 4) {
drivers/spi/spi-fsl-qspi.c
593
memcpy(&val, op->data.buf.out + i, 4);
drivers/spi/spi-fsl-qspi.c
598
if (i < op->data.nbytes) {
drivers/spi/spi-fsl-qspi.c
599
memcpy(&val, op->data.buf.out + i, op->data.nbytes - i);
drivers/spi/spi-fsl-qspi.c
605
for (i = op->data.nbytes; i < 16; i += 4)
drivers/spi/spi-fsl-qspi.c
611
const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
615
u8 *buf = op->data.buf.in;
drivers/spi/spi-fsl-qspi.c
618
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 4); i += 4) {
drivers/spi/spi-fsl-qspi.c
624
if (i < op->data.nbytes) {
drivers/spi/spi-fsl-qspi.c
627
memcpy(buf + i, &val, op->data.nbytes - i);
drivers/spi/spi-fsl-qspi.c
631
static int fsl_qspi_do_op(struct fsl_qspi *q, const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
643
qspi_writel(q, op->data.nbytes | QUADSPI_IPCR_SEQID(SEQID_LUT),
drivers/spi/spi-fsl-qspi.c
650
if (!err && op->data.nbytes && op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-fsl-qspi.c
651
fsl_qspi_read_rxfifo(q, op);
drivers/spi/spi-fsl-qspi.c
668
static int fsl_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
682
fsl_qspi_select_mem(q, mem->spi, op);
drivers/spi/spi-fsl-qspi.c
702
fsl_qspi_prepare_lut(q, op);
drivers/spi/spi-fsl-qspi.c
709
if (op->data.nbytes > (q->devtype_data->rxfifo - 4) &&
drivers/spi/spi-fsl-qspi.c
710
op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-fsl-qspi.c
711
fsl_qspi_read_ahb(q, op);
drivers/spi/spi-fsl-qspi.c
716
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-fsl-qspi.c
717
fsl_qspi_fill_txfifo(q, op);
drivers/spi/spi-fsl-qspi.c
719
err = fsl_qspi_do_op(q, op);
drivers/spi/spi-fsl-qspi.c
730
static int fsl_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-fsl-qspi.c
734
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-fsl-qspi.c
735
if (op->data.nbytes > q->devtype_data->txfifo)
drivers/spi/spi-fsl-qspi.c
736
op->data.nbytes = q->devtype_data->txfifo;
drivers/spi/spi-fsl-qspi.c
738
if (op->data.nbytes > q->devtype_data->ahb_buf_size)
drivers/spi/spi-fsl-qspi.c
739
op->data.nbytes = q->devtype_data->ahb_buf_size;
drivers/spi/spi-fsl-qspi.c
740
else if (op->data.nbytes > (q->devtype_data->rxfifo - 4))
drivers/spi/spi-fsl-qspi.c
741
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 8);
drivers/spi/spi-gxp.c
101
memcpy_fromio(op->data.buf.in, spifi->dat_base, op->data.nbytes);
drivers/spi/spi-gxp.c
105
static int gxp_spi_write_reg(struct gxp_spi_chip *chip, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
121
writeb(op->cmd.opcode, reg_base + OFFSET_SPICMD);
drivers/spi/spi-gxp.c
123
memcpy_toio(spifi->dat_base, op->data.buf.in, op->data.nbytes);
drivers/spi/spi-gxp.c
125
writew(op->data.nbytes, reg_base + OFFSET_SPIDCNT);
drivers/spi/spi-gxp.c
142
static ssize_t gxp_spi_read(struct gxp_spi_chip *chip, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
145
u32 offset = op->addr.val;
drivers/spi/spi-gxp.c
150
memcpy_fromio(op->data.buf.in, spifi->dir_base + offset, op->data.nbytes);
drivers/spi/spi-gxp.c
155
static ssize_t gxp_spi_write(struct gxp_spi_chip *chip, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
163
write_len = op->data.nbytes;
drivers/spi/spi-gxp.c
171
value |= (op->addr.nbytes << 16);
drivers/spi/spi-gxp.c
175
writel(op->addr.val, reg_base + OFFSET_SPIADDR);
drivers/spi/spi-gxp.c
177
writeb(op->cmd.opcode, reg_base + OFFSET_SPICMD);
drivers/spi/spi-gxp.c
181
memcpy_toio(spifi->dat_base, op->data.buf.in, write_len);
drivers/spi/spi-gxp.c
200
static int do_gxp_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
206
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-gxp.c
207
if (!op->addr.nbytes)
drivers/spi/spi-gxp.c
208
ret = gxp_spi_read_reg(chip, op);
drivers/spi/spi-gxp.c
210
ret = gxp_spi_read(chip, op);
drivers/spi/spi-gxp.c
212
if (!op->addr.nbytes)
drivers/spi/spi-gxp.c
213
ret = gxp_spi_write_reg(chip, op);
drivers/spi/spi-gxp.c
215
ret = gxp_spi_write(chip, op);
drivers/spi/spi-gxp.c
221
static int gxp_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
225
ret = do_gxp_exec_mem_op(mem, op);
drivers/spi/spi-gxp.c
67
static int gxp_spi_read_reg(struct gxp_spi_chip *chip, const struct spi_mem_op *op)
drivers/spi/spi-gxp.c
83
writeb(op->cmd.opcode, reg_base + OFFSET_SPICMD);
drivers/spi/spi-gxp.c
85
writew(op->data.nbytes, reg_base + OFFSET_SPIDCNT);
drivers/spi/spi-hisi-sfc-v3xx.c
148
struct spi_mem_op *op)
drivers/spi/spi-hisi-sfc-v3xx.c
152
uintptr_t addr = (uintptr_t)op->data.buf.in;
drivers/spi/spi-hisi-sfc-v3xx.c
159
if (!IS_ALIGNED(addr, 4) && op->data.nbytes >= 4)
drivers/spi/spi-hisi-sfc-v3xx.c
160
op->data.nbytes = 4 - (addr % 4);
drivers/spi/spi-hisi-sfc-v3xx.c
161
else if (op->data.nbytes > max_byte_count)
drivers/spi/spi-hisi-sfc-v3xx.c
162
op->data.nbytes = max_byte_count;
drivers/spi/spi-hisi-sfc-v3xx.c
172
const struct spi_mem_op *op)
drivers/spi/spi-hisi-sfc-v3xx.c
179
if (op->data.buswidth > 4 || op->dummy.buswidth > 4 ||
drivers/spi/spi-hisi-sfc-v3xx.c
180
op->addr.buswidth > 4 || op->cmd.buswidth > 4)
drivers/spi/spi-hisi-sfc-v3xx.c
183
if (op->addr.nbytes != host->address_mode && op->addr.nbytes)
drivers/spi/spi-hisi-sfc-v3xx.c
186
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-hisi-sfc-v3xx.c
274
const struct spi_mem_op *op,
drivers/spi/spi-hisi-sfc-v3xx.c
277
int len = op->data.nbytes, buswidth_mode;
drivers/spi/spi-hisi-sfc-v3xx.c
280
if (op->addr.nbytes)
drivers/spi/spi-hisi-sfc-v3xx.c
283
if (op->data.buswidth == 0 || op->data.buswidth == 1) {
drivers/spi/spi-hisi-sfc-v3xx.c
288
data_idx = (op->data.buswidth - 1) / 2;
drivers/spi/spi-hisi-sfc-v3xx.c
289
addr_idx = op->addr.buswidth / 2;
drivers/spi/spi-hisi-sfc-v3xx.c
290
cmd_idx = op->cmd.buswidth / 2;
drivers/spi/spi-hisi-sfc-v3xx.c
297
if (op->data.dir != SPI_MEM_NO_DATA) {
drivers/spi/spi-hisi-sfc-v3xx.c
302
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-hisi-sfc-v3xx.c
305
config |= op->dummy.nbytes << HISI_SFC_V3XX_CMD_CFG_DUMMY_CNT_OFF |
drivers/spi/spi-hisi-sfc-v3xx.c
309
writel(op->addr.val, host->regbase + HISI_SFC_V3XX_CMD_ADDR);
drivers/spi/spi-hisi-sfc-v3xx.c
310
writel(op->cmd.opcode, host->regbase + HISI_SFC_V3XX_CMD_INS);
drivers/spi/spi-hisi-sfc-v3xx.c
318
const struct spi_mem_op *op,
drivers/spi/spi-hisi-sfc-v3xx.c
329
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-hisi-sfc-v3xx.c
330
hisi_sfc_v3xx_write_databuf(host, op->data.buf.out, op->data.nbytes);
drivers/spi/spi-hisi-sfc-v3xx.c
332
ret = hisi_sfc_v3xx_start_bus(host, op, chip_select);
drivers/spi/spi-hisi-sfc-v3xx.c
353
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-hisi-sfc-v3xx.c
354
hisi_sfc_v3xx_read_databuf(host, op->data.buf.in, op->data.nbytes);
drivers/spi/spi-hisi-sfc-v3xx.c
360
const struct spi_mem_op *op)
drivers/spi/spi-hisi-sfc-v3xx.c
368
return hisi_sfc_v3xx_generic_exec_op(host, op, chip_select);
drivers/spi/spi-intel.c
1285
struct spi_mem_op op =
drivers/spi/spi-intel.c
1293
op.addr.val = 0x10;
drivers/spi/spi-intel.c
1294
op.data.buf.in = buf;
drivers/spi/spi-intel.c
1295
op.data.nbytes = sizeof(buf);
drivers/spi/spi-intel.c
1297
ret = intel_spi_read(ispi, NULL, NULL, &op);
drivers/spi/spi-intel.c
1314
op.addr.val = fcba;
drivers/spi/spi-intel.c
1315
op.data.buf.in = &flcomp;
drivers/spi/spi-intel.c
1316
op.data.nbytes = sizeof(flcomp);
drivers/spi/spi-intel.c
1318
ret = intel_spi_read(ispi, NULL, NULL, &op);
drivers/spi/spi-intel.c
187
const struct spi_mem_op *op);
drivers/spi/spi-intel.c
469
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
471
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
472
size_t nbytes = op->data.nbytes;
drivers/spi/spi-intel.c
473
u8 opcode = op->cmd.opcode;
drivers/spi/spi-intel.c
487
return intel_spi_read_block(ispi, op->data.buf.in, nbytes);
drivers/spi/spi-intel.c
492
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
494
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
495
size_t nbytes = op->data.nbytes;
drivers/spi/spi-intel.c
496
u8 opcode = op->cmd.opcode;
drivers/spi/spi-intel.c
541
ret = intel_spi_write_block(ispi, op->data.buf.out, nbytes);
drivers/spi/spi-intel.c
553
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
555
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
556
size_t block_size, nbytes = op->data.nbytes;
drivers/spi/spi-intel.c
557
void *read_buf = op->data.buf.in;
drivers/spi/spi-intel.c
614
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
616
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
617
size_t block_size, nbytes = op->data.nbytes;
drivers/spi/spi-intel.c
618
const void *write_buf = op->data.buf.out;
drivers/spi/spi-intel.c
677
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
679
u32 addr = intel_spi_chip_addr(ispi, mem) + op->addr.val;
drivers/spi/spi-intel.c
680
u8 opcode = op->cmd.opcode;
drivers/spi/spi-intel.c
713
static int intel_spi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-intel.c
715
op->data.nbytes = clamp_val(op->data.nbytes, 0, INTEL_SPI_FIFO_SZ);
drivers/spi/spi-intel.c
720
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
722
if (iop->mem_op.cmd.nbytes != op->cmd.nbytes ||
drivers/spi/spi-intel.c
723
iop->mem_op.cmd.buswidth != op->cmd.buswidth ||
drivers/spi/spi-intel.c
724
iop->mem_op.cmd.dtr != op->cmd.dtr)
drivers/spi/spi-intel.c
727
if (iop->mem_op.addr.nbytes != op->addr.nbytes ||
drivers/spi/spi-intel.c
728
iop->mem_op.addr.dtr != op->addr.dtr)
drivers/spi/spi-intel.c
731
if (iop->mem_op.data.dir != op->data.dir ||
drivers/spi/spi-intel.c
732
iop->mem_op.data.dtr != op->data.dtr)
drivers/spi/spi-intel.c
736
if (iop->mem_op.data.buswidth != op->data.buswidth)
drivers/spi/spi-intel.c
744
intel_spi_match_mem_op(struct intel_spi *ispi, const struct spi_mem_op *op)
drivers/spi/spi-intel.c
749
if (iop->mem_op.cmd.opcode == op->cmd.opcode &&
drivers/spi/spi-intel.c
750
intel_spi_cmp_mem_op(iop, op))
drivers/spi/spi-intel.c
758
const struct spi_mem_op *op)
drivers/spi/spi-intel.c
763
iop = intel_spi_match_mem_op(ispi, op);
drivers/spi/spi-intel.c
765
dev_dbg(ispi->dev, "%#x not supported\n", op->cmd.opcode);
drivers/spi/spi-intel.c
778
if (ispi->opcodes[i] == op->cmd.opcode)
drivers/spi/spi-intel.c
782
dev_dbg(ispi->dev, "%#x not supported\n", op->cmd.opcode);
drivers/spi/spi-intel.c
789
static int intel_spi_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-intel.c
794
iop = intel_spi_match_mem_op(ispi, op);
drivers/spi/spi-intel.c
798
return iop->exec_op(ispi, mem, iop, op);
drivers/spi/spi-intel.c
830
struct spi_mem_op op = desc->info.op_tmpl;
drivers/spi/spi-intel.c
834
op.addr.val = offs;
drivers/spi/spi-intel.c
835
op.data.nbytes = len;
drivers/spi/spi-intel.c
836
op.data.buf.in = buf;
drivers/spi/spi-intel.c
838
ret = iop->exec_op(ispi, desc->mem, iop, &op);
drivers/spi/spi-intel.c
847
struct spi_mem_op op = desc->info.op_tmpl;
drivers/spi/spi-intel.c
850
op.addr.val = offs;
drivers/spi/spi-intel.c
851
op.data.nbytes = len;
drivers/spi/spi-intel.c
852
op.data.buf.out = buf;
drivers/spi/spi-intel.c
854
ret = iop->exec_op(ispi, desc->mem, iop, &op);
drivers/spi/spi-mem.c
1011
op, &status);
drivers/spi/spi-mem.c
102
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-mem.c
146
const struct spi_mem_op *op)
drivers/spi/spi-mem.c
148
if (spi_check_buswidth_req(mem, op->cmd.buswidth, true))
drivers/spi/spi-mem.c
151
if (op->addr.nbytes &&
drivers/spi/spi-mem.c
152
spi_check_buswidth_req(mem, op->addr.buswidth, true))
drivers/spi/spi-mem.c
155
if (op->dummy.nbytes &&
drivers/spi/spi-mem.c
156
spi_check_buswidth_req(mem, op->dummy.buswidth, true))
drivers/spi/spi-mem.c
159
if (op->data.dir != SPI_MEM_NO_DATA &&
drivers/spi/spi-mem.c
160
spi_check_buswidth_req(mem, op->data.buswidth,
drivers/spi/spi-mem.c
161
op->data.dir == SPI_MEM_DATA_OUT))
drivers/spi/spi-mem.c
168
const struct spi_mem_op *op)
drivers/spi/spi-mem.c
172
op->cmd.dtr || op->addr.dtr || op->dummy.dtr || op->data.dtr;
drivers/spi/spi-mem.c
178
if (op->data.swap16 && !spi_mem_controller_is_capable(ctlr, swap16))
drivers/spi/spi-mem.c
182
if (op->cmd.dtr && op->cmd.buswidth == 8) {
drivers/spi/spi-mem.c
183
if (op->cmd.nbytes != 2)
drivers/spi/spi-mem.c
186
if ((op->addr.nbytes % 2) ||
drivers/spi/spi-mem.c
187
(op->dummy.nbytes % 2) ||
drivers/spi/spi-mem.c
188
(op->data.nbytes % 2)) {
drivers/spi/spi-mem.c
195
if (op->cmd.nbytes != 1)
drivers/spi/spi-mem.c
199
if (op->data.ecc) {
drivers/spi/spi-mem.c
204
if (op->max_freq && mem->spi->controller->min_speed_hz &&
drivers/spi/spi-mem.c
205
op->max_freq < mem->spi->controller->min_speed_hz)
drivers/spi/spi-mem.c
208
if (op->max_freq &&
drivers/spi/spi-mem.c
209
op->max_freq < mem->spi->max_speed_hz) {
drivers/spi/spi-mem.c
214
return spi_mem_check_buswidth(mem, op);
drivers/spi/spi-mem.c
226
static int spi_mem_check_op(const struct spi_mem_op *op)
drivers/spi/spi-mem.c
228
if (!op->cmd.buswidth || !op->cmd.nbytes)
drivers/spi/spi-mem.c
231
if ((op->addr.nbytes && !op->addr.buswidth) ||
drivers/spi/spi-mem.c
232
(op->dummy.nbytes && !op->dummy.buswidth) ||
drivers/spi/spi-mem.c
233
(op->data.nbytes && !op->data.buswidth))
drivers/spi/spi-mem.c
236
if (!spi_mem_buswidth_is_valid(op->cmd.buswidth) ||
drivers/spi/spi-mem.c
237
!spi_mem_buswidth_is_valid(op->addr.buswidth) ||
drivers/spi/spi-mem.c
238
!spi_mem_buswidth_is_valid(op->dummy.buswidth) ||
drivers/spi/spi-mem.c
239
!spi_mem_buswidth_is_valid(op->data.buswidth))
drivers/spi/spi-mem.c
243
if (WARN_ON_ONCE(op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-mem.c
244
object_is_on_stack(op->data.buf.in)))
drivers/spi/spi-mem.c
247
if (WARN_ON_ONCE(op->data.dir == SPI_MEM_DATA_OUT &&
drivers/spi/spi-mem.c
248
object_is_on_stack(op->data.buf.out)))
drivers/spi/spi-mem.c
255
const struct spi_mem_op *op)
drivers/spi/spi-mem.c
260
return ctlr->mem_ops->supports_op(mem, op);
drivers/spi/spi-mem.c
262
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-mem.c
280
bool spi_mem_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mem.c
283
spi_mem_adjust_op_freq(mem, (struct spi_mem_op *)op);
drivers/spi/spi-mem.c
285
if (spi_mem_check_op(op))
drivers/spi/spi-mem.c
288
return spi_mem_internal_supports_op(mem, op);
drivers/spi/spi-mem.c
331
const struct spi_mem_op *op, int exec_op_ret)
drivers/spi/spi-mem.c
348
len = op->cmd.nbytes + op->addr.nbytes;
drivers/spi/spi-mem.c
349
len += op->dummy.nbytes + op->data.nbytes;
drivers/spi/spi-mem.c
355
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-mem.c
356
u64_stats_add(&stats->bytes_tx, op->data.nbytes);
drivers/spi/spi-mem.c
357
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mem.c
358
u64_stats_add(&stats->bytes_rx, op->data.nbytes);
drivers/spi/spi-mem.c
385
int spi_mem_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mem.c
395
spi_mem_adjust_op_freq(mem, (struct spi_mem_op *)op);
drivers/spi/spi-mem.c
398
op->cmd.opcode,
drivers/spi/spi-mem.c
399
op->addr.nbytes, (op->addr.nbytes ? op->addr.val : 0),
drivers/spi/spi-mem.c
40
const struct spi_mem_op *op,
drivers/spi/spi-mem.c
400
op->dummy.nbytes,
drivers/spi/spi-mem.c
401
op->data.nbytes, (op->data.nbytes ? (op->data.dir == SPI_MEM_DATA_IN ? " read" : "write") : " "),
drivers/spi/spi-mem.c
402
op->cmd.buswidth, op->cmd.dtr ? 'D' : 'S',
drivers/spi/spi-mem.c
403
op->addr.buswidth, op->addr.dtr ? 'D' : 'S',
drivers/spi/spi-mem.c
404
op->dummy.buswidth, op->dummy.dtr ? 'D' : 'S',
drivers/spi/spi-mem.c
405
op->data.buswidth, op->data.dtr ? 'D' : 'S',
drivers/spi/spi-mem.c
406
op->max_freq ? op->max_freq : mem->spi->max_speed_hz);
drivers/spi/spi-mem.c
408
ret = spi_mem_check_op(op);
drivers/spi/spi-mem.c
412
if (!spi_mem_internal_supports_op(mem, op))
drivers/spi/spi-mem.c
420
trace_spi_mem_start_op(mem, op);
drivers/spi/spi-mem.c
421
ret = ctlr->mem_ops->exec_op(mem, op);
drivers/spi/spi-mem.c
422
trace_spi_mem_stop_op(mem, op);
drivers/spi/spi-mem.c
432
spi_mem_add_op_stats(ctlr->pcpu_statistics, op, ret);
drivers/spi/spi-mem.c
433
spi_mem_add_op_stats(mem->spi->pcpu_statistics, op, ret);
drivers/spi/spi-mem.c
439
tmpbufsize = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-mem.c
45
if (!op->data.nbytes)
drivers/spi/spi-mem.c
452
tmpbuf[0] = op->cmd.opcode;
drivers/spi/spi-mem.c
454
xfers[xferpos].len = op->cmd.nbytes;
drivers/spi/spi-mem.c
455
xfers[xferpos].tx_nbits = op->cmd.buswidth;
drivers/spi/spi-mem.c
456
xfers[xferpos].speed_hz = op->max_freq;
drivers/spi/spi-mem.c
461
if (op->addr.nbytes) {
drivers/spi/spi-mem.c
464
for (i = 0; i < op->addr.nbytes; i++)
drivers/spi/spi-mem.c
465
tmpbuf[i + 1] = op->addr.val >>
drivers/spi/spi-mem.c
466
(8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-mem.c
469
xfers[xferpos].len = op->addr.nbytes;
drivers/spi/spi-mem.c
470
xfers[xferpos].tx_nbits = op->addr.buswidth;
drivers/spi/spi-mem.c
471
xfers[xferpos].speed_hz = op->max_freq;
drivers/spi/spi-mem.c
474
totalxferlen += op->addr.nbytes;
drivers/spi/spi-mem.c
477
if (op->dummy.nbytes) {
drivers/spi/spi-mem.c
478
memset(tmpbuf + op->addr.nbytes + 1, 0xff, op->dummy.nbytes);
drivers/spi/spi-mem.c
479
xfers[xferpos].tx_buf = tmpbuf + op->addr.nbytes + 1;
drivers/spi/spi-mem.c
48
if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx)
drivers/spi/spi-mem.c
480
xfers[xferpos].len = op->dummy.nbytes;
drivers/spi/spi-mem.c
481
xfers[xferpos].tx_nbits = op->dummy.buswidth;
drivers/spi/spi-mem.c
483
xfers[xferpos].speed_hz = op->max_freq;
drivers/spi/spi-mem.c
486
totalxferlen += op->dummy.nbytes;
drivers/spi/spi-mem.c
489
if (op->data.nbytes) {
drivers/spi/spi-mem.c
490
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mem.c
491
xfers[xferpos].rx_buf = op->data.buf.in;
drivers/spi/spi-mem.c
492
xfers[xferpos].rx_nbits = op->data.buswidth;
drivers/spi/spi-mem.c
494
xfers[xferpos].tx_buf = op->data.buf.out;
drivers/spi/spi-mem.c
495
xfers[xferpos].tx_nbits = op->data.buswidth;
drivers/spi/spi-mem.c
498
xfers[xferpos].len = op->data.nbytes;
drivers/spi/spi-mem.c
499
xfers[xferpos].speed_hz = op->max_freq;
drivers/spi/spi-mem.c
50
else if (op->data.dir == SPI_MEM_DATA_IN && ctlr->dma_rx)
drivers/spi/spi-mem.c
502
totalxferlen += op->data.nbytes;
drivers/spi/spi-mem.c
552
int spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
558
return ctlr->mem_ops->adjust_op_size(mem, op);
drivers/spi/spi-mem.c
561
len = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-mem.c
566
op->data.nbytes = min3((size_t)op->data.nbytes,
drivers/spi/spi-mem.c
570
if (!op->data.nbytes)
drivers/spi/spi-mem.c
58
return spi_map_buf(ctlr, dmadev, sgt, op->data.buf.in, op->data.nbytes,
drivers/spi/spi-mem.c
588
void spi_mem_adjust_op_freq(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
59
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-mem.c
590
if (!op->max_freq || op->max_freq > mem->spi->max_speed_hz)
drivers/spi/spi-mem.c
591
op->max_freq = mem->spi->max_speed_hz;
drivers/spi/spi-mem.c
614
u64 spi_mem_calc_op_duration(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mem.c
619
spi_mem_adjust_op_freq(mem, op);
drivers/spi/spi-mem.c
621
if (op->max_freq) {
drivers/spi/spi-mem.c
623
do_div(ps_per_cycles, op->max_freq);
drivers/spi/spi-mem.c
629
ncycles += ((op->cmd.nbytes * 8) / op->cmd.buswidth) / (op->cmd.dtr ? 2 : 1);
drivers/spi/spi-mem.c
630
ncycles += ((op->addr.nbytes * 8) / op->addr.buswidth) / (op->addr.dtr ? 2 : 1);
drivers/spi/spi-mem.c
633
if (op->dummy.nbytes)
drivers/spi/spi-mem.c
634
ncycles += ((op->dummy.nbytes * 8) / op->dummy.buswidth) / (op->dummy.dtr ? 2 : 1);
drivers/spi/spi-mem.c
636
ncycles += ((op->data.nbytes * 8) / op->data.buswidth) / (op->data.dtr ? 2 : 1);
drivers/spi/spi-mem.c
650
struct spi_mem_op op = desc->info.op_tmpl;
drivers/spi/spi-mem.c
653
op.addr.val = desc->info.offset + offs;
drivers/spi/spi-mem.c
654
op.data.buf.in = buf;
drivers/spi/spi-mem.c
655
op.data.nbytes = len;
drivers/spi/spi-mem.c
656
ret = spi_mem_adjust_op_size(desc->mem, &op);
drivers/spi/spi-mem.c
660
ret = spi_mem_exec_op(desc->mem, &op);
drivers/spi/spi-mem.c
664
return op.data.nbytes;
drivers/spi/spi-mem.c
670
struct spi_mem_op op = desc->info.op_tmpl;
drivers/spi/spi-mem.c
673
op.addr.val = desc->info.offset + offs;
drivers/spi/spi-mem.c
674
op.data.buf.out = buf;
drivers/spi/spi-mem.c
675
op.data.nbytes = len;
drivers/spi/spi-mem.c
676
ret = spi_mem_adjust_op_size(desc->mem, &op);
drivers/spi/spi-mem.c
680
ret = spi_mem_exec_op(desc->mem, &op);
drivers/spi/spi-mem.c
684
return op.data.nbytes;
drivers/spi/spi-mem.c
86
const struct spi_mem_op *op,
drivers/spi/spi-mem.c
91
if (!op->data.nbytes)
drivers/spi/spi-mem.c
936
const struct spi_mem_op *op,
drivers/spi/spi-mem.c
939
const u8 *bytes = (u8 *)op->data.buf.in;
drivers/spi/spi-mem.c
94
if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx)
drivers/spi/spi-mem.c
942
ret = spi_mem_exec_op(mem, op);
drivers/spi/spi-mem.c
946
if (op->data.nbytes > 1)
drivers/spi/spi-mem.c
96
else if (op->data.dir == SPI_MEM_DATA_IN && ctlr->dma_rx)
drivers/spi/spi-mem.c
971
const struct spi_mem_op *op,
drivers/spi/spi-mem.c
982
if (op->data.nbytes < 1 || op->data.nbytes > 2 ||
drivers/spi/spi-mem.c
983
op->data.dir != SPI_MEM_DATA_IN)
drivers/spi/spi-mem.c
991
ret = ctlr->mem_ops->poll_status(mem, op, mask, match,
drivers/spi/spi-mem.c
999
if (!spi_mem_supports_op(mem, op))
drivers/spi/spi-microchip-core-qspi.c
126
static int mchp_coreqspi_set_mode(struct mchp_coreqspi *qspi, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
141
if (op->data.buswidth == 4 || op->data.buswidth == 2) {
drivers/spi/spi-microchip-core-qspi.c
143
if (op->cmd.buswidth == 1 && (op->addr.buswidth == 1 || op->addr.buswidth == 0))
drivers/spi/spi-microchip-core-qspi.c
145
else if (op->cmd.buswidth == 1)
drivers/spi/spi-microchip-core-qspi.c
396
static inline void mchp_coreqspi_config_op(struct mchp_coreqspi *qspi, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
401
cmd_bytes = op->cmd.nbytes + op->addr.nbytes;
drivers/spi/spi-microchip-core-qspi.c
402
total_bytes = cmd_bytes + op->data.nbytes;
drivers/spi/spi-microchip-core-qspi.c
430
if (!(op->data.dir == SPI_MEM_DATA_IN))
drivers/spi/spi-microchip-core-qspi.c
438
if (op->dummy.buswidth)
drivers/spi/spi-microchip-core-qspi.c
439
idle_cycles = op->dummy.nbytes * 8 / op->dummy.buswidth;
drivers/spi/spi-microchip-core-qspi.c
460
static int mchp_coreqspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
464
u32 address = op->addr.val;
drivers/spi/spi-microchip-core-qspi.c
465
u8 opcode = op->cmd.opcode;
drivers/spi/spi-microchip-core-qspi.c
476
err = mchp_coreqspi_setup_clock(qspi, mem->spi, op->max_freq);
drivers/spi/spi-microchip-core-qspi.c
480
err = mchp_coreqspi_set_mode(qspi, op);
drivers/spi/spi-microchip-core-qspi.c
485
mchp_coreqspi_config_op(qspi, op);
drivers/spi/spi-microchip-core-qspi.c
486
if (op->cmd.opcode) {
drivers/spi/spi-microchip-core-qspi.c
489
qspi->tx_len = op->cmd.nbytes;
drivers/spi/spi-microchip-core-qspi.c
495
if (op->addr.nbytes) {
drivers/spi/spi-microchip-core-qspi.c
496
for (i = 0; i < op->addr.nbytes; i++)
drivers/spi/spi-microchip-core-qspi.c
497
qspi->txbuf[i] = address >> (8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-microchip-core-qspi.c
500
qspi->tx_len = op->addr.nbytes;
drivers/spi/spi-microchip-core-qspi.c
505
if (op->data.nbytes) {
drivers/spi/spi-microchip-core-qspi.c
506
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-microchip-core-qspi.c
507
qspi->txbuf = (u8 *)op->data.buf.out;
drivers/spi/spi-microchip-core-qspi.c
510
qspi->tx_len = op->data.nbytes;
drivers/spi/spi-microchip-core-qspi.c
514
qspi->rxbuf = (u8 *)op->data.buf.in;
drivers/spi/spi-microchip-core-qspi.c
515
qspi->rx_len = op->data.nbytes;
drivers/spi/spi-microchip-core-qspi.c
532
static bool mchp_coreqspi_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
534
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-microchip-core-qspi.c
537
if ((op->data.buswidth == 4 || op->data.buswidth == 2) &&
drivers/spi/spi-microchip-core-qspi.c
538
(op->cmd.buswidth == 1 && (op->addr.buswidth == 1 || op->addr.buswidth == 0))) {
drivers/spi/spi-microchip-core-qspi.c
552
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-microchip-core-qspi.c
559
static int mchp_coreqspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-microchip-core-qspi.c
561
if (op->data.dir == SPI_MEM_DATA_OUT || op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-microchip-core-qspi.c
562
if (op->data.nbytes > MAX_DATA_CMD_LEN)
drivers/spi/spi-microchip-core-qspi.c
563
op->data.nbytes = MAX_DATA_CMD_LEN;
drivers/spi/spi-mpc52xx.c
387
static int mpc52xx_spi_probe(struct platform_device *op)
drivers/spi/spi-mpc52xx.c
397
dev_dbg(&op->dev, "probing mpc5200 SPI device\n");
drivers/spi/spi-mpc52xx.c
398
regs = of_iomap(op->dev.of_node, 0);
drivers/spi/spi-mpc52xx.c
418
dev_err(&op->dev, "mode fault; is port_config correct?\n");
drivers/spi/spi-mpc52xx.c
423
dev_dbg(&op->dev, "allocating spi_controller struct\n");
drivers/spi/spi-mpc52xx.c
424
host = spi_alloc_host(&op->dev, sizeof(*ms));
drivers/spi/spi-mpc52xx.c
434
platform_set_drvdata(op, host);
drivers/spi/spi-mpc52xx.c
439
ms->irq0 = irq_of_parse_and_map(op->dev.of_node, 0);
drivers/spi/spi-mpc52xx.c
440
ms->irq1 = irq_of_parse_and_map(op->dev.of_node, 1);
drivers/spi/spi-mpc52xx.c
442
ms->ipb_freq = mpc5xxx_get_bus_frequency(&op->dev);
drivers/spi/spi-mpc52xx.c
443
ms->gpio_cs_count = gpiod_count(&op->dev, NULL);
drivers/spi/spi-mpc52xx.c
453
gpio_cs = gpiod_get_index(&op->dev,
drivers/spi/spi-mpc52xx.c
457
dev_err(&op->dev,
drivers/spi/spi-mpc52xx.c
488
dev_info(&op->dev, "using polled mode\n");
drivers/spi/spi-mpc52xx.c
490
dev_dbg(&op->dev, "registering spi_controller struct\n");
drivers/spi/spi-mpc52xx.c
514
static void mpc52xx_spi_remove(struct platform_device *op)
drivers/spi/spi-mpc52xx.c
516
struct spi_controller *host = spi_controller_get(platform_get_drvdata(op));
drivers/spi/spi-mt65xx.c
1010
mtk_spi_prepare_transfer(mem->spi->controller, op->max_freq);
drivers/spi/spi-mt65xx.c
1019
if (op->addr.nbytes || op->dummy.nbytes)
drivers/spi/spi-mt65xx.c
1020
reg_val |= (op->addr.nbytes + op->dummy.nbytes) <<
drivers/spi/spi-mt65xx.c
1024
if (op->data.dir == SPI_MEM_NO_DATA) {
drivers/spi/spi-mt65xx.c
1029
mdata->xfer_len = op->data.nbytes;
drivers/spi/spi-mt65xx.c
1033
if (op->addr.nbytes || op->dummy.nbytes) {
drivers/spi/spi-mt65xx.c
1034
if (op->addr.buswidth == 1 || op->dummy.buswidth == 1)
drivers/spi/spi-mt65xx.c
1040
if (op->addr.buswidth == 2 ||
drivers/spi/spi-mt65xx.c
1041
op->dummy.buswidth == 2 ||
drivers/spi/spi-mt65xx.c
1042
op->data.buswidth == 2)
drivers/spi/spi-mt65xx.c
1044
else if (op->addr.buswidth == 4 ||
drivers/spi/spi-mt65xx.c
1045
op->dummy.buswidth == 4 ||
drivers/spi/spi-mt65xx.c
1046
op->data.buswidth == 4)
drivers/spi/spi-mt65xx.c
1055
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mt65xx.c
1061
tx_size = 1 + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-mt65xx.c
1062
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-mt65xx.c
1063
tx_size += op->data.nbytes;
drivers/spi/spi-mt65xx.c
1073
tx_tmp_buf[0] = op->cmd.opcode;
drivers/spi/spi-mt65xx.c
1075
if (op->addr.nbytes) {
drivers/spi/spi-mt65xx.c
1078
for (i = 0; i < op->addr.nbytes; i++)
drivers/spi/spi-mt65xx.c
1079
tx_tmp_buf[i + 1] = op->addr.val >>
drivers/spi/spi-mt65xx.c
1080
(8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-mt65xx.c
1083
if (op->dummy.nbytes)
drivers/spi/spi-mt65xx.c
1084
memset(tx_tmp_buf + op->addr.nbytes + 1,
drivers/spi/spi-mt65xx.c
1086
op->dummy.nbytes);
drivers/spi/spi-mt65xx.c
1088
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-mt65xx.c
1089
memcpy(tx_tmp_buf + op->dummy.nbytes + op->addr.nbytes + 1,
drivers/spi/spi-mt65xx.c
1090
op->data.buf.out,
drivers/spi/spi-mt65xx.c
1091
op->data.nbytes);
drivers/spi/spi-mt65xx.c
1100
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mt65xx.c
1101
if (!IS_ALIGNED((size_t)op->data.buf.in, 4)) {
drivers/spi/spi-mt65xx.c
1102
rx_tmp_buf = kzalloc(op->data.nbytes,
drivers/spi/spi-mt65xx.c
1109
rx_tmp_buf = op->data.buf.in;
drivers/spi/spi-mt65xx.c
1114
op->data.nbytes,
drivers/spi/spi-mt65xx.c
1124
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mt65xx.c
1128
mtk_spi_mem_setup_dma_xfer(mem->spi->controller, op);
drivers/spi/spi-mt65xx.c
1133
ret = mtk_spi_transfer_wait(mem, op);
drivers/spi/spi-mt65xx.c
1140
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mt65xx.c
1145
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mt65xx.c
1147
op->data.nbytes, DMA_FROM_DEVICE);
drivers/spi/spi-mt65xx.c
1148
if (!IS_ALIGNED((size_t)op->data.buf.in, 4))
drivers/spi/spi-mt65xx.c
1149
memcpy(op->data.buf.in, rx_tmp_buf, op->data.nbytes);
drivers/spi/spi-mt65xx.c
1152
if (op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-mt65xx.c
1153
!IS_ALIGNED((size_t)op->data.buf.in, 4))
drivers/spi/spi-mt65xx.c
904
struct spi_mem_op *op)
drivers/spi/spi-mt65xx.c
908
if (op->data.dir != SPI_MEM_NO_DATA) {
drivers/spi/spi-mt65xx.c
909
opcode_len = 1 + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-mt65xx.c
910
if (opcode_len + op->data.nbytes > MTK_SPI_IPM_PACKET_SIZE) {
drivers/spi/spi-mt65xx.c
911
op->data.nbytes = MTK_SPI_IPM_PACKET_SIZE - opcode_len;
drivers/spi/spi-mt65xx.c
913
op->data.nbytes -= op->data.nbytes % 4;
drivers/spi/spi-mt65xx.c
921
const struct spi_mem_op *op)
drivers/spi/spi-mt65xx.c
923
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mt65xx.c
926
if (op->addr.nbytes && op->dummy.nbytes &&
drivers/spi/spi-mt65xx.c
927
op->addr.buswidth != op->dummy.buswidth)
drivers/spi/spi-mt65xx.c
930
if (op->addr.nbytes + op->dummy.nbytes > 16)
drivers/spi/spi-mt65xx.c
933
if (op->data.nbytes > MTK_SPI_IPM_PACKET_SIZE) {
drivers/spi/spi-mt65xx.c
934
if (op->data.nbytes / MTK_SPI_IPM_PACKET_SIZE >
drivers/spi/spi-mt65xx.c
936
op->data.nbytes % MTK_SPI_IPM_PACKET_SIZE != 0)
drivers/spi/spi-mt65xx.c
944
const struct spi_mem_op *op)
drivers/spi/spi-mt65xx.c
956
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mt65xx.c
968
const struct spi_mem_op *op)
drivers/spi/spi-mt65xx.c
978
if (op->data.dir == SPI_MEM_NO_DATA)
drivers/spi/spi-mt65xx.c
981
ms *= op->data.nbytes;
drivers/spi/spi-mt65xx.c
998
const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
163
static void mtk_nor_set_addr(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
165
u32 addr = op->addr.val;
drivers/spi/spi-mtk-nor.c
172
if (op->addr.nbytes == 4) {
drivers/spi/spi-mtk-nor.c
180
static bool need_bounce(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
182
return ((uintptr_t)op->data.buf.in & MTK_NOR_DMA_ALIGN_MASK);
drivers/spi/spi-mtk-nor.c
185
static bool mtk_nor_match_read(const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
189
if (op->dummy.nbytes)
drivers/spi/spi-mtk-nor.c
190
dummy = op->dummy.nbytes * BITS_PER_BYTE / op->dummy.buswidth;
drivers/spi/spi-mtk-nor.c
192
if ((op->data.buswidth == 2) || (op->data.buswidth == 4)) {
drivers/spi/spi-mtk-nor.c
193
if (op->addr.buswidth == 1)
drivers/spi/spi-mtk-nor.c
195
else if (op->addr.buswidth == 2)
drivers/spi/spi-mtk-nor.c
197
else if (op->addr.buswidth == 4)
drivers/spi/spi-mtk-nor.c
199
} else if ((op->addr.buswidth == 1) && (op->data.buswidth == 1)) {
drivers/spi/spi-mtk-nor.c
200
if (op->cmd.opcode == 0x03)
drivers/spi/spi-mtk-nor.c
202
else if (op->cmd.opcode == 0x0b)
drivers/spi/spi-mtk-nor.c
208
static bool mtk_nor_match_prg(const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
213
if ((op->cmd.buswidth > 1) || (op->addr.buswidth > 1) ||
drivers/spi/spi-mtk-nor.c
214
(op->dummy.buswidth > 1) || (op->data.buswidth > 1))
drivers/spi/spi-mtk-nor.c
217
tx_len = op->cmd.nbytes + op->addr.nbytes;
drivers/spi/spi-mtk-nor.c
219
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-nor.c
221
tx_len += op->dummy.nbytes;
drivers/spi/spi-mtk-nor.c
229
if ((!op->addr.nbytes) &&
drivers/spi/spi-mtk-nor.c
230
(tx_len + op->data.nbytes > MTK_NOR_REG_PRGDATA_MAX + 1))
drivers/spi/spi-mtk-nor.c
232
} else if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mtk-nor.c
236
rx_len = op->data.nbytes;
drivers/spi/spi-mtk-nor.c
237
prg_left = MTK_NOR_PRG_CNT_MAX / 8 - tx_len - op->dummy.nbytes;
drivers/spi/spi-mtk-nor.c
241
if (!op->addr.nbytes)
drivers/spi/spi-mtk-nor.c
246
prg_len = tx_len + op->dummy.nbytes + rx_len;
drivers/spi/spi-mtk-nor.c
250
prg_len = tx_len + op->dummy.nbytes;
drivers/spi/spi-mtk-nor.c
257
static void mtk_nor_adj_prg_size(struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
261
tx_len = op->cmd.nbytes + op->addr.nbytes;
drivers/spi/spi-mtk-nor.c
262
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-nor.c
263
tx_len += op->dummy.nbytes;
drivers/spi/spi-mtk-nor.c
265
if (op->data.nbytes > tx_left)
drivers/spi/spi-mtk-nor.c
266
op->data.nbytes = tx_left;
drivers/spi/spi-mtk-nor.c
267
} else if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mtk-nor.c
268
prg_left = MTK_NOR_PRG_CNT_MAX / 8 - tx_len - op->dummy.nbytes;
drivers/spi/spi-mtk-nor.c
271
if (op->data.nbytes > prg_left)
drivers/spi/spi-mtk-nor.c
272
op->data.nbytes = prg_left;
drivers/spi/spi-mtk-nor.c
276
static int mtk_nor_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
280
if (!op->data.nbytes)
drivers/spi/spi-mtk-nor.c
283
if ((op->addr.nbytes == 3) || (op->addr.nbytes == 4)) {
drivers/spi/spi-mtk-nor.c
284
if ((op->data.dir == SPI_MEM_DATA_IN) &&
drivers/spi/spi-mtk-nor.c
285
mtk_nor_match_read(op)) {
drivers/spi/spi-mtk-nor.c
287
if (op->data.nbytes > 0x400000)
drivers/spi/spi-mtk-nor.c
288
op->data.nbytes = 0x400000;
drivers/spi/spi-mtk-nor.c
290
if ((op->addr.val & MTK_NOR_DMA_ALIGN_MASK) ||
drivers/spi/spi-mtk-nor.c
291
(op->data.nbytes < MTK_NOR_DMA_ALIGN))
drivers/spi/spi-mtk-nor.c
292
op->data.nbytes = 1;
drivers/spi/spi-mtk-nor.c
293
else if (!need_bounce(sp, op))
drivers/spi/spi-mtk-nor.c
294
op->data.nbytes &= ~MTK_NOR_DMA_ALIGN_MASK;
drivers/spi/spi-mtk-nor.c
295
else if (op->data.nbytes > MTK_NOR_BOUNCE_BUF_SIZE)
drivers/spi/spi-mtk-nor.c
296
op->data.nbytes = MTK_NOR_BOUNCE_BUF_SIZE;
drivers/spi/spi-mtk-nor.c
298
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-nor.c
299
if (op->data.nbytes >= MTK_NOR_PP_SIZE)
drivers/spi/spi-mtk-nor.c
300
op->data.nbytes = MTK_NOR_PP_SIZE;
drivers/spi/spi-mtk-nor.c
302
op->data.nbytes = 1;
drivers/spi/spi-mtk-nor.c
307
mtk_nor_adj_prg_size(op);
drivers/spi/spi-mtk-nor.c
312
const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
314
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mtk-nor.c
317
if (op->cmd.buswidth != 1)
drivers/spi/spi-mtk-nor.c
320
if ((op->addr.nbytes == 3) || (op->addr.nbytes == 4)) {
drivers/spi/spi-mtk-nor.c
321
switch (op->data.dir) {
drivers/spi/spi-mtk-nor.c
323
if (mtk_nor_match_read(op))
drivers/spi/spi-mtk-nor.c
327
if ((op->addr.buswidth == 1) &&
drivers/spi/spi-mtk-nor.c
328
(op->dummy.nbytes == 0) &&
drivers/spi/spi-mtk-nor.c
329
(op->data.buswidth == 1))
drivers/spi/spi-mtk-nor.c
337
return mtk_nor_match_prg(op);
drivers/spi/spi-mtk-nor.c
340
static void mtk_nor_setup_bus(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
344
if (op->addr.nbytes == 4)
drivers/spi/spi-mtk-nor.c
347
if (op->data.buswidth == 4) {
drivers/spi/spi-mtk-nor.c
349
writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA(4));
drivers/spi/spi-mtk-nor.c
350
if (op->addr.buswidth == 4)
drivers/spi/spi-mtk-nor.c
352
} else if (op->data.buswidth == 2) {
drivers/spi/spi-mtk-nor.c
354
writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA(3));
drivers/spi/spi-mtk-nor.c
355
if (op->addr.buswidth == 2)
drivers/spi/spi-mtk-nor.c
358
if (op->cmd.opcode == 0x0b)
drivers/spi/spi-mtk-nor.c
410
static int mtk_nor_read_bounce(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
415
if (op->data.nbytes & MTK_NOR_DMA_ALIGN_MASK)
drivers/spi/spi-mtk-nor.c
416
rdlen = (op->data.nbytes + MTK_NOR_DMA_ALIGN) & ~MTK_NOR_DMA_ALIGN_MASK;
drivers/spi/spi-mtk-nor.c
418
rdlen = op->data.nbytes;
drivers/spi/spi-mtk-nor.c
420
ret = mtk_nor_dma_exec(sp, op->addr.val, rdlen, sp->buffer_dma);
drivers/spi/spi-mtk-nor.c
423
memcpy(op->data.buf.in, sp->buffer, op->data.nbytes);
drivers/spi/spi-mtk-nor.c
428
static int mtk_nor_read_dma(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
433
if (need_bounce(sp, op))
drivers/spi/spi-mtk-nor.c
434
return mtk_nor_read_bounce(sp, op);
drivers/spi/spi-mtk-nor.c
436
dma_addr = dma_map_single(sp->dev, op->data.buf.in,
drivers/spi/spi-mtk-nor.c
437
op->data.nbytes, DMA_FROM_DEVICE);
drivers/spi/spi-mtk-nor.c
442
ret = mtk_nor_dma_exec(sp, op->addr.val, op->data.nbytes, dma_addr);
drivers/spi/spi-mtk-nor.c
444
dma_unmap_single(sp->dev, dma_addr, op->data.nbytes, DMA_FROM_DEVICE);
drivers/spi/spi-mtk-nor.c
449
static int mtk_nor_read_pio(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
451
u8 *buf = op->data.buf.in;
drivers/spi/spi-mtk-nor.c
485
static int mtk_nor_pp_buffered(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
487
const u8 *buf = op->data.buf.out;
drivers/spi/spi-mtk-nor.c
495
for (i = 0; i < op->data.nbytes; i += 4) {
drivers/spi/spi-mtk-nor.c
501
(op->data.nbytes + 5) * BITS_PER_BYTE);
drivers/spi/spi-mtk-nor.c
505
const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
507
const u8 *buf = op->data.buf.out;
drivers/spi/spi-mtk-nor.c
517
static int mtk_nor_spi_mem_prg(struct mtk_nor *sp, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
526
tx_len = op->cmd.nbytes + op->addr.nbytes;
drivers/spi/spi-mtk-nor.c
529
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-mtk-nor.c
530
tx_len += op->dummy.nbytes + op->data.nbytes;
drivers/spi/spi-mtk-nor.c
531
else if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mtk-nor.c
532
rx_len = op->data.nbytes;
drivers/spi/spi-mtk-nor.c
534
prg_len = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes +
drivers/spi/spi-mtk-nor.c
535
op->data.nbytes;
drivers/spi/spi-mtk-nor.c
546
for (i = op->cmd.nbytes; i > 0; i--, reg_offset--) {
drivers/spi/spi-mtk-nor.c
548
bufbyte = (op->cmd.opcode >> ((i - 1) * BITS_PER_BYTE)) & 0xff;
drivers/spi/spi-mtk-nor.c
552
for (i = op->addr.nbytes; i > 0; i--, reg_offset--) {
drivers/spi/spi-mtk-nor.c
554
bufbyte = (op->addr.val >> ((i - 1) * BITS_PER_BYTE)) & 0xff;
drivers/spi/spi-mtk-nor.c
558
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-nor.c
559
for (i = 0; i < op->dummy.nbytes; i++, reg_offset--) {
drivers/spi/spi-mtk-nor.c
564
for (i = 0; i < op->data.nbytes; i++, reg_offset--) {
drivers/spi/spi-mtk-nor.c
566
writeb(((const u8 *)(op->data.buf.out))[i], reg);
drivers/spi/spi-mtk-nor.c
589
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mtk-nor.c
590
for (i = op->data.nbytes - 1; i >= 0; i--, reg_offset++) {
drivers/spi/spi-mtk-nor.c
592
((u8 *)(op->data.buf.in))[i] = readb(reg);
drivers/spi/spi-mtk-nor.c
599
static int mtk_nor_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mtk-nor.c
604
if ((op->data.nbytes == 0) ||
drivers/spi/spi-mtk-nor.c
605
((op->addr.nbytes != 3) && (op->addr.nbytes != 4)))
drivers/spi/spi-mtk-nor.c
606
return mtk_nor_spi_mem_prg(sp, op);
drivers/spi/spi-mtk-nor.c
608
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-nor.c
609
mtk_nor_set_addr(sp, op);
drivers/spi/spi-mtk-nor.c
610
writeb(op->cmd.opcode, sp->base + MTK_NOR_REG_PRGDATA0);
drivers/spi/spi-mtk-nor.c
611
if (op->data.nbytes == MTK_NOR_PP_SIZE)
drivers/spi/spi-mtk-nor.c
612
return mtk_nor_pp_buffered(sp, op);
drivers/spi/spi-mtk-nor.c
613
return mtk_nor_pp_unbuffered(sp, op);
drivers/spi/spi-mtk-nor.c
616
if ((op->data.dir == SPI_MEM_DATA_IN) && mtk_nor_match_read(op)) {
drivers/spi/spi-mtk-nor.c
620
mtk_nor_setup_bus(sp, op);
drivers/spi/spi-mtk-nor.c
621
if (op->data.nbytes == 1) {
drivers/spi/spi-mtk-nor.c
622
mtk_nor_set_addr(sp, op);
drivers/spi/spi-mtk-nor.c
623
return mtk_nor_read_pio(sp, op);
drivers/spi/spi-mtk-nor.c
625
ret = mtk_nor_read_dma(sp, op);
drivers/spi/spi-mtk-nor.c
629
mtk_nor_setup_bus(sp, op);
drivers/spi/spi-mtk-nor.c
630
return mtk_nor_read_dma(sp, op);
drivers/spi/spi-mtk-nor.c
637
return mtk_nor_spi_mem_prg(sp, op);
drivers/spi/spi-mtk-snfi.c
1008
memset(op->data.buf.in, 0xff, op->data.nbytes);
drivers/spi/spi-mtk-snfi.c
1013
if (buf == op->data.buf.in) {
drivers/spi/spi-mtk-snfi.c
1015
u32 req_left = op->data.nbytes - snf->nfi_cfg.page_size;
drivers/spi/spi-mtk-snfi.c
1018
memcpy(op->data.buf.in + snf->nfi_cfg.page_size,
drivers/spi/spi-mtk-snfi.c
1024
if (op->data.nbytes < cap_len)
drivers/spi/spi-mtk-snfi.c
1025
cap_len = op->data.nbytes;
drivers/spi/spi-mtk-snfi.c
1026
memcpy(op->data.buf.in, snf->buf + rd_offset, cap_len);
drivers/spi/spi-mtk-snfi.c
1030
if (op->data.ecc)
drivers/spi/spi-mtk-snfi.c
1055
const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1058
u32 op_addr = op->addr.val;
drivers/spi/spi-mtk-snfi.c
1075
if (op->data.ecc)
drivers/spi/spi-mtk-snfi.c
1090
if (op->data.nbytes < cap_len)
drivers/spi/spi-mtk-snfi.c
1091
cap_len = op->data.nbytes;
drivers/spi/spi-mtk-snfi.c
1092
memcpy(snf->buf + wr_offset, op->data.buf.out, cap_len);
drivers/spi/spi-mtk-snfi.c
1102
nfi_write32(snf, SNF_PG_CTL1, (op->cmd.opcode << PG_LOAD_CMD_S));
drivers/spi/spi-mtk-snfi.c
1108
if (op->data.buswidth == 4)
drivers/spi/spi-mtk-snfi.c
1133
if (op->data.ecc) {
drivers/spi/spi-mtk-snfi.c
1134
snf->ecc_cfg->op = ECC_ENCODE;
drivers/spi/spi-mtk-snfi.c
1165
if (op->data.ecc)
drivers/spi/spi-mtk-snfi.c
1199
static bool mtk_snand_is_page_ops(const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1201
if (op->addr.nbytes != 2)
drivers/spi/spi-mtk-snfi.c
1204
if (op->addr.buswidth != 1 && op->addr.buswidth != 2 &&
drivers/spi/spi-mtk-snfi.c
1205
op->addr.buswidth != 4)
drivers/spi/spi-mtk-snfi.c
1209
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mtk-snfi.c
1211
if (op->dummy.nbytes * BITS_PER_BYTE / op->dummy.buswidth >
drivers/spi/spi-mtk-snfi.c
1215
if ((op->addr.buswidth == 4 || op->addr.buswidth == 1) &&
drivers/spi/spi-mtk-snfi.c
1216
op->data.buswidth == 4)
drivers/spi/spi-mtk-snfi.c
1220
if ((op->addr.buswidth == 2 || op->addr.buswidth == 1) &&
drivers/spi/spi-mtk-snfi.c
1221
op->data.buswidth == 2)
drivers/spi/spi-mtk-snfi.c
1225
if (op->addr.buswidth == 1 && op->data.buswidth == 1)
drivers/spi/spi-mtk-snfi.c
1227
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-snfi.c
1229
if (op->dummy.nbytes)
drivers/spi/spi-mtk-snfi.c
1232
if (op->addr.buswidth == 1 && op->data.buswidth == 4)
drivers/spi/spi-mtk-snfi.c
1235
if (op->addr.buswidth == 1 && op->data.buswidth == 1)
drivers/spi/spi-mtk-snfi.c
1242
const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1244
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-mtk-snfi.c
1246
if (op->cmd.nbytes != 1 || op->cmd.buswidth != 1)
drivers/spi/spi-mtk-snfi.c
1248
if (mtk_snand_is_page_ops(op))
drivers/spi/spi-mtk-snfi.c
1250
return ((op->addr.nbytes == 0 || op->addr.buswidth == 1) &&
drivers/spi/spi-mtk-snfi.c
1251
(op->dummy.nbytes == 0 || op->dummy.buswidth == 1) &&
drivers/spi/spi-mtk-snfi.c
1252
(op->data.nbytes == 0 || op->data.buswidth == 1));
drivers/spi/spi-mtk-snfi.c
1255
static int mtk_snand_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1262
if (mtk_snand_is_page_ops(op)) {
drivers/spi/spi-mtk-snfi.c
1269
if (op->data.nbytes > l)
drivers/spi/spi-mtk-snfi.c
1270
op->data.nbytes = l;
drivers/spi/spi-mtk-snfi.c
1272
size_t hl = op->cmd.nbytes + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-mtk-snfi.c
1276
if (op->data.nbytes > SNF_GPRAM_SIZE - hl)
drivers/spi/spi-mtk-snfi.c
1277
op->data.nbytes = SNF_GPRAM_SIZE - hl;
drivers/spi/spi-mtk-snfi.c
1282
static int mtk_snand_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
1286
if (mtk_snand_is_page_ops(op)) {
drivers/spi/spi-mtk-snfi.c
1287
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-mtk-snfi.c
1288
return mtk_snand_read_page_cache(ms, op);
drivers/spi/spi-mtk-snfi.c
1290
return mtk_snand_write_page_cache(ms, op);
drivers/spi/spi-mtk-snfi.c
1292
return mtk_snand_mac_io(ms, op);
drivers/spi/spi-mtk-snfi.c
460
static int mtk_snand_mac_io(struct mtk_snand *snf, const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
470
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mtk-snfi.c
471
rx_len = op->data.nbytes;
drivers/spi/spi-mtk-snfi.c
472
rx_buf = op->data.buf.in;
drivers/spi/spi-mtk-snfi.c
474
tx_buf = op->data.buf.out;
drivers/spi/spi-mtk-snfi.c
479
for (i = 0; i < op->cmd.nbytes; i++, reg_offs++) {
drivers/spi/spi-mtk-snfi.c
480
b = (op->cmd.opcode >> ((op->cmd.nbytes - i - 1) * 8)) & 0xff;
drivers/spi/spi-mtk-snfi.c
488
for (i = 0; i < op->addr.nbytes; i++, reg_offs++) {
drivers/spi/spi-mtk-snfi.c
489
b = (op->addr.val >> ((op->addr.nbytes - i - 1) * 8)) & 0xff;
drivers/spi/spi-mtk-snfi.c
497
for (i = 0; i < op->dummy.nbytes; i++, reg_offs++) {
drivers/spi/spi-mtk-snfi.c
504
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-mtk-snfi.c
505
for (i = 0; i < op->data.nbytes; i++, reg_offs++) {
drivers/spi/spi-mtk-snfi.c
862
const struct spi_mem_op *op)
drivers/spi/spi-mtk-snfi.c
867
u32 op_addr = op->addr.val;
drivers/spi/spi-mtk-snfi.c
870
u32 dummy_clk = (op->dummy.nbytes * BITS_PER_BYTE / op->dummy.buswidth);
drivers/spi/spi-mtk-snfi.c
883
if (op->data.ecc)
drivers/spi/spi-mtk-snfi.c
896
if (rd_offset == 0 && op->data.nbytes >= snf->nfi_cfg.page_size)
drivers/spi/spi-mtk-snfi.c
897
buf = op->data.buf.in;
drivers/spi/spi-mtk-snfi.c
905
(op->cmd.opcode << DATA_READ_CMD_S));
drivers/spi/spi-mtk-snfi.c
911
if (op->data.buswidth == 4)
drivers/spi/spi-mtk-snfi.c
912
rd_mode = op->addr.buswidth == 4 ? DATA_READ_MODE_QUAD :
drivers/spi/spi-mtk-snfi.c
914
else if (op->data.buswidth == 2)
drivers/spi/spi-mtk-snfi.c
915
rd_mode = op->addr.buswidth == 2 ? DATA_READ_MODE_DUAL :
drivers/spi/spi-mtk-snfi.c
943
if (op->data.ecc) {
drivers/spi/spi-mtk-snfi.c
944
snf->ecc_cfg->op = ECC_DECODE;
drivers/spi/spi-mtk-snfi.c
985
if (op->data.ecc) {
drivers/spi/spi-mxic.c
319
static u32 mxic_spi_mem_prep_op_cfg(const struct spi_mem_op *op,
drivers/spi/spi-mxic.c
322
u32 cfg = OP_CMD_BYTES(op->cmd.nbytes) |
drivers/spi/spi-mxic.c
323
OP_CMD_BUSW(fls(op->cmd.buswidth) - 1) |
drivers/spi/spi-mxic.c
324
(op->cmd.dtr ? OP_CMD_DDR : 0);
drivers/spi/spi-mxic.c
326
if (op->addr.nbytes)
drivers/spi/spi-mxic.c
327
cfg |= OP_ADDR_BYTES(op->addr.nbytes) |
drivers/spi/spi-mxic.c
328
OP_ADDR_BUSW(fls(op->addr.buswidth) - 1) |
drivers/spi/spi-mxic.c
329
(op->addr.dtr ? OP_ADDR_DDR : 0);
drivers/spi/spi-mxic.c
331
if (op->dummy.nbytes)
drivers/spi/spi-mxic.c
332
cfg |= OP_DUMMY_CYC(op->dummy.nbytes);
drivers/spi/spi-mxic.c
336
cfg |= OP_DATA_BUSW(fls(op->data.buswidth) - 1) |
drivers/spi/spi-mxic.c
337
(op->data.dtr ? OP_DATA_DDR : 0);
drivers/spi/spi-mxic.c
338
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-mxic.c
340
if (op->data.dtr)
drivers/spi/spi-mxic.c
486
const struct spi_mem_op *op)
drivers/spi/spi-mxic.c
488
if (op->data.buswidth > 8 || op->addr.buswidth > 8 ||
drivers/spi/spi-mxic.c
489
op->dummy.buswidth > 8 || op->cmd.buswidth > 8)
drivers/spi/spi-mxic.c
492
if (op->data.nbytes && op->dummy.nbytes &&
drivers/spi/spi-mxic.c
493
op->data.buswidth != op->dummy.buswidth)
drivers/spi/spi-mxic.c
496
if (op->addr.nbytes > 7)
drivers/spi/spi-mxic.c
499
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-mxic.c
519
const struct spi_mem_op *op)
drivers/spi/spi-mxic.c
525
ret = mxic_spi_set_freq(mxic, op->max_freq);
drivers/spi/spi-mxic.c
529
writel(mxic_spi_prep_hc_cfg(mem->spi, HC_CFG_MAN_CS_EN, op->data.swap16),
drivers/spi/spi-mxic.c
534
writel(mxic_spi_mem_prep_op_cfg(op, op->data.nbytes),
drivers/spi/spi-mxic.c
540
for (i = 0; i < op->cmd.nbytes; i++)
drivers/spi/spi-mxic.c
541
cmd[i] = op->cmd.opcode >> (8 * (op->cmd.nbytes - i - 1));
drivers/spi/spi-mxic.c
543
ret = mxic_spi_data_xfer(mxic, cmd, NULL, op->cmd.nbytes);
drivers/spi/spi-mxic.c
547
for (i = 0; i < op->addr.nbytes; i++)
drivers/spi/spi-mxic.c
548
addr[i] = op->addr.val >> (8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-mxic.c
550
ret = mxic_spi_data_xfer(mxic, addr, NULL, op->addr.nbytes);
drivers/spi/spi-mxic.c
554
ret = mxic_spi_data_xfer(mxic, NULL, NULL, op->dummy.nbytes);
drivers/spi/spi-mxic.c
559
op->data.dir == SPI_MEM_DATA_OUT ?
drivers/spi/spi-mxic.c
560
op->data.buf.out : NULL,
drivers/spi/spi-mxic.c
561
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-mxic.c
562
op->data.buf.in : NULL,
drivers/spi/spi-mxic.c
563
op->data.nbytes);
drivers/spi/spi-npcm-fiu.c
267
const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
271
ilog2(op->addr.buswidth) <<
drivers/spi/spi-npcm-fiu.c
273
fiu->drd_op.addr.buswidth = op->addr.buswidth;
drivers/spi/spi-npcm-fiu.c
276
op->dummy.nbytes << NPCM_FIU_DRD_DBW_SHIFT);
drivers/spi/spi-npcm-fiu.c
277
fiu->drd_op.dummy.nbytes = op->dummy.nbytes;
drivers/spi/spi-npcm-fiu.c
279
NPCM_FIU_DRD_CFG_RDCMD, op->cmd.opcode);
drivers/spi/spi-npcm-fiu.c
280
fiu->drd_op.cmd.opcode = op->cmd.opcode;
drivers/spi/spi-npcm-fiu.c
283
(op->addr.nbytes - 3) << NPCM_FIU_DRD_ADDSIZ_SHIFT);
drivers/spi/spi-npcm-fiu.c
284
fiu->drd_op.addr.nbytes = op->addr.nbytes;
drivers/spi/spi-npcm-fiu.c
335
const struct spi_mem_op *op, u32 addr,
drivers/spi/spi-npcm-fiu.c
351
NPCM_FIU_UMA_CMD_CMD, op->cmd.opcode);
drivers/spi/spi-npcm-fiu.c
354
uma_cfg |= ilog2(op->cmd.buswidth);
drivers/spi/spi-npcm-fiu.c
355
uma_cfg |= ilog2(op->addr.buswidth)
drivers/spi/spi-npcm-fiu.c
357
if (op->dummy.nbytes)
drivers/spi/spi-npcm-fiu.c
358
uma_cfg |= ilog2(op->dummy.buswidth)
drivers/spi/spi-npcm-fiu.c
360
uma_cfg |= ilog2(op->data.buswidth)
drivers/spi/spi-npcm-fiu.c
362
uma_cfg |= op->dummy.nbytes << NPCM_FIU_UMA_CFG_DBSIZ_SHIFT;
drivers/spi/spi-npcm-fiu.c
363
uma_cfg |= op->addr.nbytes << NPCM_FIU_UMA_CFG_ADDSIZ_SHIFT;
drivers/spi/spi-npcm-fiu.c
391
const struct spi_mem_op *op, u8 cmd,
drivers/spi/spi-npcm-fiu.c
417
uma_cfg |= ilog2(op->cmd.buswidth);
drivers/spi/spi-npcm-fiu.c
418
uma_cfg |= ilog2(op->addr.buswidth) <<
drivers/spi/spi-npcm-fiu.c
420
uma_cfg |= ilog2(op->data.buswidth) <<
drivers/spi/spi-npcm-fiu.c
422
uma_cfg |= op->addr.nbytes << NPCM_FIU_UMA_CFG_ADDSIZ_SHIFT;
drivers/spi/spi-npcm-fiu.c
423
regmap_write(fiu->regmap, NPCM_FIU_UMA_ADDR, op->addr.val);
drivers/spi/spi-npcm-fiu.c
441
const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
445
u8 *data = (u8 *)op->data.buf.out;
drivers/spi/spi-npcm-fiu.c
451
num_data_chunks = op->data.nbytes / CHUNK_SIZE;
drivers/spi/spi-npcm-fiu.c
452
remain_data = op->data.nbytes % CHUNK_SIZE;
drivers/spi/spi-npcm-fiu.c
461
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, true, NULL, 0);
drivers/spi/spi-npcm-fiu.c
467
ret = npcm_fiu_uma_write(mem, op, data[0], false,
drivers/spi/spi-npcm-fiu.c
477
ret = npcm_fiu_uma_write(mem, op, data[0], false,
drivers/spi/spi-npcm-fiu.c
489
static int npcm_fiu_read(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
491
u8 *data = op->data.buf.in;
drivers/spi/spi-npcm-fiu.c
498
currlen = op->data.nbytes;
drivers/spi/spi-npcm-fiu.c
501
addr = ((u32)op->addr.val + i);
drivers/spi/spi-npcm-fiu.c
505
ret = npcm_fiu_uma_read(mem, op, addr, true, buf_ptr,
drivers/spi/spi-npcm-fiu.c
543
static int npcm_fiu_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-npcm-fiu.c
551
if (fiu->spix_mode || op->addr.nbytes > 4)
drivers/spi/spi-npcm-fiu.c
563
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-npcm-fiu.c
564
if (!op->addr.nbytes) {
drivers/spi/spi-npcm-fiu.c
565
buf = op->data.buf.in;
drivers/spi/spi-npcm-fiu.c
566
ret = npcm_fiu_uma_read(mem, op, op->addr.val, false,
drivers/spi/spi-npcm-fiu.c
567
buf, op->data.nbytes);
drivers/spi/spi-npcm-fiu.c
569
ret = npcm_fiu_read(mem, op);
drivers/spi/spi-npcm-fiu.c
572
if (!op->addr.nbytes && !op->data.nbytes)
drivers/spi/spi-npcm-fiu.c
573
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
575
if (op->addr.nbytes && !op->data.nbytes) {
drivers/spi/spi-npcm-fiu.c
578
u32 addr = op->addr.val;
drivers/spi/spi-npcm-fiu.c
580
for (i = op->addr.nbytes - 1; i >= 0; i--) {
drivers/spi/spi-npcm-fiu.c
584
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
585
buf_addr, op->addr.nbytes);
drivers/spi/spi-npcm-fiu.c
587
if (!op->addr.nbytes && op->data.nbytes)
drivers/spi/spi-npcm-fiu.c
588
ret = npcm_fiu_uma_write(mem, op, op->cmd.opcode, false,
drivers/spi/spi-npcm-fiu.c
589
(u8 *)op->data.buf.out,
drivers/spi/spi-npcm-fiu.c
590
op->data.nbytes);
drivers/spi/spi-npcm-fiu.c
591
if (op->addr.nbytes && op->data.nbytes)
drivers/spi/spi-npcm-fiu.c
592
ret = npcm_fiu_manualwrite(mem, op);
drivers/spi/spi-nxp-fspi.c
1001
fspi_writel(f, op->addr.val, base + FSPI_IPCR0);
drivers/spi/spi-nxp-fspi.c
1008
fspi_writel(f, op->data.nbytes |
drivers/spi/spi-nxp-fspi.c
1021
if (!err && op->data.nbytes && op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-nxp-fspi.c
1022
nxp_fspi_read_rxfifo(f, op);
drivers/spi/spi-nxp-fspi.c
1027
static int nxp_fspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
1045
nxp_fspi_select_mem(f, mem->spi, op);
drivers/spi/spi-nxp-fspi.c
1047
nxp_fspi_prepare_lut(f, op);
drivers/spi/spi-nxp-fspi.c
1054
if (op->data.nbytes > (f->devtype_data->rxfifo - 4) &&
drivers/spi/spi-nxp-fspi.c
1055
op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-nxp-fspi.c
1057
err = nxp_fspi_read_ahb(f, op);
drivers/spi/spi-nxp-fspi.c
1059
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-nxp-fspi.c
1060
nxp_fspi_fill_txfifo(f, op);
drivers/spi/spi-nxp-fspi.c
1062
err = nxp_fspi_do_op(f, op);
drivers/spi/spi-nxp-fspi.c
1073
static int nxp_fspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
1077
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-nxp-fspi.c
1078
if (op->data.nbytes > f->devtype_data->txfifo)
drivers/spi/spi-nxp-fspi.c
1079
op->data.nbytes = f->devtype_data->txfifo;
drivers/spi/spi-nxp-fspi.c
1081
if (op->data.nbytes > f->devtype_data->ahb_buf_size)
drivers/spi/spi-nxp-fspi.c
1082
op->data.nbytes = f->devtype_data->ahb_buf_size;
drivers/spi/spi-nxp-fspi.c
1083
else if (op->data.nbytes > (f->devtype_data->rxfifo - 4))
drivers/spi/spi-nxp-fspi.c
1084
op->data.nbytes = ALIGN_DOWN(op->data.nbytes, 8);
drivers/spi/spi-nxp-fspi.c
1088
if (op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-nxp-fspi.c
1090
op->data.nbytes > f->devtype_data->rxfifo)
drivers/spi/spi-nxp-fspi.c
1091
op->data.nbytes = f->devtype_data->rxfifo;
drivers/spi/spi-nxp-fspi.c
470
const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
475
ret = nxp_fspi_check_buswidth(f, op->cmd.buswidth);
drivers/spi/spi-nxp-fspi.c
477
if (op->addr.nbytes)
drivers/spi/spi-nxp-fspi.c
478
ret |= nxp_fspi_check_buswidth(f, op->addr.buswidth);
drivers/spi/spi-nxp-fspi.c
480
if (op->dummy.nbytes)
drivers/spi/spi-nxp-fspi.c
481
ret |= nxp_fspi_check_buswidth(f, op->dummy.buswidth);
drivers/spi/spi-nxp-fspi.c
483
if (op->data.nbytes)
drivers/spi/spi-nxp-fspi.c
484
ret |= nxp_fspi_check_buswidth(f, op->data.buswidth);
drivers/spi/spi-nxp-fspi.c
492
if (op->addr.nbytes > 4)
drivers/spi/spi-nxp-fspi.c
500
if (op->addr.val >= f->memmap_phy_size)
drivers/spi/spi-nxp-fspi.c
504
if (op->dummy.buswidth &&
drivers/spi/spi-nxp-fspi.c
505
(op->dummy.nbytes * 8 / op->dummy.buswidth > 64))
drivers/spi/spi-nxp-fspi.c
509
if (op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-nxp-fspi.c
510
(op->data.nbytes > f->devtype_data->ahb_buf_size ||
drivers/spi/spi-nxp-fspi.c
511
(op->data.nbytes > f->devtype_data->rxfifo - 4 &&
drivers/spi/spi-nxp-fspi.c
512
!IS_ALIGNED(op->data.nbytes, 8))))
drivers/spi/spi-nxp-fspi.c
515
if (op->data.dir == SPI_MEM_DATA_OUT &&
drivers/spi/spi-nxp-fspi.c
516
op->data.nbytes > f->devtype_data->txfifo)
drivers/spi/spi-nxp-fspi.c
519
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-nxp-fspi.c
560
const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
569
if (op->cmd.dtr) {
drivers/spi/spi-nxp-fspi.c
570
lutval[0] |= LUT_DEF(0, LUT_CMD_DDR, LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-fspi.c
571
op->cmd.opcode >> 8);
drivers/spi/spi-nxp-fspi.c
573
LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-fspi.c
574
op->cmd.opcode & 0xFF);
drivers/spi/spi-nxp-fspi.c
577
lutval[0] |= LUT_DEF(0, LUT_CMD, LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-fspi.c
578
op->cmd.opcode);
drivers/spi/spi-nxp-fspi.c
582
if (op->addr.nbytes) {
drivers/spi/spi-nxp-fspi.c
583
lutval[lutidx / 2] |= LUT_DEF(lutidx, op->addr.dtr ? LUT_ADDR_DDR : LUT_ADDR,
drivers/spi/spi-nxp-fspi.c
584
LUT_PAD(op->addr.buswidth),
drivers/spi/spi-nxp-fspi.c
585
op->addr.nbytes * 8);
drivers/spi/spi-nxp-fspi.c
590
if (op->dummy.nbytes) {
drivers/spi/spi-nxp-fspi.c
591
lutval[lutidx / 2] |= LUT_DEF(lutidx, op->dummy.dtr ? LUT_DUMMY_DDR : LUT_DUMMY,
drivers/spi/spi-nxp-fspi.c
596
LUT_PAD(op->data.buswidth),
drivers/spi/spi-nxp-fspi.c
597
op->dummy.nbytes * 8 /
drivers/spi/spi-nxp-fspi.c
598
op->dummy.buswidth);
drivers/spi/spi-nxp-fspi.c
603
if (op->data.nbytes) {
drivers/spi/spi-nxp-fspi.c
605
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-nxp-fspi.c
606
(op->data.dtr ? LUT_READ_DDR : LUT_NXP_READ) :
drivers/spi/spi-nxp-fspi.c
607
(op->data.dtr ? LUT_WRITE_DDR : LUT_NXP_WRITE),
drivers/spi/spi-nxp-fspi.c
608
LUT_PAD(op->data.buswidth),
drivers/spi/spi-nxp-fspi.c
627
op->cmd.opcode, lutval[0], lutval[1], lutval[2], lutval[3], op->data.nbytes);
drivers/spi/spi-nxp-fspi.c
787
const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
790
bool op_is_dtr = op->cmd.dtr && op->addr.dtr && op->dummy.dtr && op->data.dtr;
drivers/spi/spi-nxp-fspi.c
791
unsigned long rate = op->max_freq;
drivers/spi/spi-nxp-fspi.c
806
(f->pre_op_rate == op->max_freq))
drivers/spi/spi-nxp-fspi.c
824
rate = min(f->max_rate, op->max_freq);
drivers/spi/spi-nxp-fspi.c
855
f->pre_op_rate = op->max_freq;
drivers/spi/spi-nxp-fspi.c
860
static int nxp_fspi_read_ahb(struct nxp_fspi *f, const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
862
u32 start = op->addr.val;
drivers/spi/spi-nxp-fspi.c
863
u32 len = op->data.nbytes;
drivers/spi/spi-nxp-fspi.c
884
memcpy_fromio(op->data.buf.in,
drivers/spi/spi-nxp-fspi.c
891
const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
895
u8 *buf = (u8 *) op->data.buf.out;
drivers/spi/spi-nxp-fspi.c
905
for (i = 0; i < ALIGN_DOWN(op->data.nbytes, 8); i += 8) {
drivers/spi/spi-nxp-fspi.c
917
if (i < op->data.nbytes) {
drivers/spi/spi-nxp-fspi.c
920
int remaining = op->data.nbytes - i;
drivers/spi/spi-nxp-fspi.c
936
const struct spi_mem_op *op)
drivers/spi/spi-nxp-fspi.c
940
int len = op->data.nbytes;
drivers/spi/spi-nxp-fspi.c
941
u8 *buf = (u8 *) op->data.buf.in;
drivers/spi/spi-nxp-fspi.c
964
buf = op->data.buf.in + i;
drivers/spi/spi-nxp-fspi.c
971
len = op->data.nbytes - i;
drivers/spi/spi-nxp-fspi.c
972
for (j = 0; j < op->data.nbytes - i; j += 4) {
drivers/spi/spi-nxp-fspi.c
986
static int nxp_fspi_do_op(struct nxp_fspi *f, const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
1010
nxp_xspi_select_mem(xspi, mem->spi, op);
drivers/spi/spi-nxp-xspi.c
1012
nxp_xspi_prepare_lut(xspi, op);
drivers/spi/spi-nxp-xspi.c
1021
if ((op->data.dir == SPI_MEM_DATA_IN) && !needs_ip_only(xspi)
drivers/spi/spi-nxp-xspi.c
1022
&& ((op->addr.val + op->data.nbytes) <= xspi->memmap_phy_size))
drivers/spi/spi-nxp-xspi.c
1023
err = nxp_xspi_ahb_read(xspi, op);
drivers/spi/spi-nxp-xspi.c
1025
err = nxp_xspi_do_op(xspi, op);
drivers/spi/spi-nxp-xspi.c
1032
static int nxp_xspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
1036
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-nxp-xspi.c
1037
if (op->data.nbytes > xspi->devtype_data->txfifo)
drivers/spi/spi-nxp-xspi.c
1038
op->data.nbytes = xspi->devtype_data->txfifo;
drivers/spi/spi-nxp-xspi.c
1041
if (needs_ip_only(xspi) && (op->data.nbytes > xspi->devtype_data->rxfifo))
drivers/spi/spi-nxp-xspi.c
1042
op->data.nbytes = xspi->devtype_data->rxfifo;
drivers/spi/spi-nxp-xspi.c
1045
if (!needs_ip_only(xspi) && (op->addr.val < xspi->memmap_phy_size)
drivers/spi/spi-nxp-xspi.c
1046
&& ((op->addr.val + op->data.nbytes) > xspi->memmap_phy_size))
drivers/spi/spi-nxp-xspi.c
1047
op->data.nbytes = xspi->memmap_phy_size - op->addr.val;
drivers/spi/spi-nxp-xspi.c
374
const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
379
ret = nxp_xspi_check_buswidth(xspi, op->cmd.buswidth);
drivers/spi/spi-nxp-xspi.c
381
if (op->addr.nbytes)
drivers/spi/spi-nxp-xspi.c
382
ret |= nxp_xspi_check_buswidth(xspi, op->addr.buswidth);
drivers/spi/spi-nxp-xspi.c
384
if (op->dummy.nbytes)
drivers/spi/spi-nxp-xspi.c
385
ret |= nxp_xspi_check_buswidth(xspi, op->dummy.buswidth);
drivers/spi/spi-nxp-xspi.c
387
if (op->data.nbytes)
drivers/spi/spi-nxp-xspi.c
388
ret |= nxp_xspi_check_buswidth(xspi, op->data.buswidth);
drivers/spi/spi-nxp-xspi.c
396
if (op->addr.nbytes > 4)
drivers/spi/spi-nxp-xspi.c
400
if (op->dummy.buswidth &&
drivers/spi/spi-nxp-xspi.c
401
(op->dummy.nbytes * 8 / op->dummy.buswidth > 64))
drivers/spi/spi-nxp-xspi.c
404
if (needs_ip_only(xspi) && op->data.dir == SPI_MEM_DATA_IN &&
drivers/spi/spi-nxp-xspi.c
405
op->data.nbytes > xspi->devtype_data->rxfifo)
drivers/spi/spi-nxp-xspi.c
408
if (op->data.dir == SPI_MEM_DATA_OUT &&
drivers/spi/spi-nxp-xspi.c
409
op->data.nbytes > xspi->devtype_data->txfifo)
drivers/spi/spi-nxp-xspi.c
412
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-nxp-xspi.c
416
const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
423
if (op->cmd.dtr) {
drivers/spi/spi-nxp-xspi.c
424
lutval[0] |= LUT_DEF(0, LUT_CMD_DDR, LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-xspi.c
425
op->cmd.opcode >> 8);
drivers/spi/spi-nxp-xspi.c
427
LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-xspi.c
428
op->cmd.opcode & 0x00ff);
drivers/spi/spi-nxp-xspi.c
431
lutval[0] |= LUT_DEF(0, LUT_CMD_SDR, LUT_PAD(op->cmd.buswidth),
drivers/spi/spi-nxp-xspi.c
432
op->cmd.opcode);
drivers/spi/spi-nxp-xspi.c
436
if (op->addr.nbytes) {
drivers/spi/spi-nxp-xspi.c
437
lutval[lutidx / 2] |= LUT_DEF(lutidx, op->addr.dtr ?
drivers/spi/spi-nxp-xspi.c
439
LUT_PAD(op->addr.buswidth),
drivers/spi/spi-nxp-xspi.c
440
op->addr.nbytes * 8);
drivers/spi/spi-nxp-xspi.c
445
if (op->dummy.nbytes) {
drivers/spi/spi-nxp-xspi.c
447
LUT_PAD(op->data.buswidth),
drivers/spi/spi-nxp-xspi.c
448
op->dummy.nbytes * 8 /
drivers/spi/spi-nxp-xspi.c
450
op->dummy.buswidth / (op->dummy.dtr ? 2 : 1));
drivers/spi/spi-nxp-xspi.c
455
if (op->data.nbytes) {
drivers/spi/spi-nxp-xspi.c
457
op->data.dir == SPI_MEM_DATA_IN ?
drivers/spi/spi-nxp-xspi.c
458
(op->data.dtr ? LUT_READ_DDR : LUT_READ_SDR) :
drivers/spi/spi-nxp-xspi.c
459
(op->data.dtr ? LUT_WRITE_DDR : LUT_WRITE_SDR),
drivers/spi/spi-nxp-xspi.c
460
LUT_PAD(op->data.buswidth),
drivers/spi/spi-nxp-xspi.c
477
op->cmd.opcode, lutval[0], lutval[1], lutval[2], lutval[3], lutval[4],
drivers/spi/spi-nxp-xspi.c
478
op->data.nbytes);
drivers/spi/spi-nxp-xspi.c
687
const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
690
bool op_is_dtr = op->cmd.dtr && op->addr.dtr && op->dummy.dtr && op->data.dtr;
drivers/spi/spi-nxp-xspi.c
708
(xspi->pre_op_rate == op->max_freq))
drivers/spi/spi-nxp-xspi.c
718
rate = min_t(unsigned long, xspi->support_max_rate, op->max_freq);
drivers/spi/spi-nxp-xspi.c
745
xspi->pre_op_rate = op->max_freq;
drivers/spi/spi-nxp-xspi.c
759
if (op->data.swap16)
drivers/spi/spi-nxp-xspi.c
771
static int nxp_xspi_ahb_read(struct nxp_xspi *xspi, const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
773
u32 start = op->addr.val;
drivers/spi/spi-nxp-xspi.c
774
u32 len = op->data.nbytes;
drivers/spi/spi-nxp-xspi.c
796
memcpy_fromio(op->data.buf.in,
drivers/spi/spi-nxp-xspi.c
803
const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
806
u8 *buf = (u8 *)op->data.buf.out;
drivers/spi/spi-nxp-xspi.c
810
for (i = 0; i < ALIGN(op->data.nbytes, 4); i += 4) {
drivers/spi/spi-nxp-xspi.c
821
if (i == ALIGN_DOWN(op->data.nbytes, 4)) {
drivers/spi/spi-nxp-xspi.c
825
memcpy((u8 *)&left, buf + i, op->data.nbytes - i);
drivers/spi/spi-nxp-xspi.c
836
const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
840
u8 *buf = (u8 *) op->data.buf.in;
drivers/spi/spi-nxp-xspi.c
850
len = op->data.nbytes;
drivers/spi/spi-nxp-xspi.c
904
static int nxp_xspi_do_op(struct nxp_xspi *xspi, const struct spi_mem_op *op)
drivers/spi/spi-nxp-xspi.c
910
len = op->data.nbytes;
drivers/spi/spi-nxp-xspi.c
911
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-nxp-xspi.c
925
watermark = (xspi->devtype_data->txfifo - ALIGN(op->data.nbytes, 4)) / 4 + 1;
drivers/spi/spi-nxp-xspi.c
942
len = ALIGN(op->data.nbytes, 4);
drivers/spi/spi-nxp-xspi.c
944
} else if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-nxp-xspi.c
964
writel(op->addr.val + xspi->memmap_phy, base + XSPI_SFP_TG_SFAR);
drivers/spi/spi-nxp-xspi.c
971
if (op->data.nbytes && op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-nxp-xspi.c
972
err = nxp_xspi_fill_txfifo(xspi, op);
drivers/spi/spi-nxp-xspi.c
982
if (!err && op->data.nbytes && op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-nxp-xspi.c
983
err = nxp_xspi_read_rxfifo(xspi, op);
drivers/spi/spi-nxp-xspi.c
988
static int nxp_xspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-ppc4xx.c
335
static int spi_ppc4xx_of_probe(struct platform_device *op)
drivers/spi/spi-ppc4xx.c
341
struct device_node *np = op->dev.of_node;
drivers/spi/spi-ppc4xx.c
342
struct device *dev = &op->dev;
drivers/spi/spi-ppc4xx.c
351
platform_set_drvdata(op, host);
drivers/spi/spi-ppc4xx.c
413
ret = platform_get_irq(op, 0);
drivers/spi/spi-ppc4xx.c
466
static void spi_ppc4xx_of_remove(struct platform_device *op)
drivers/spi/spi-ppc4xx.c
468
struct spi_controller *host = platform_get_drvdata(op);
drivers/spi/spi-qcom-qspi.c
669
static int qcom_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-qcom-qspi.c
675
if (op->data.nbytes <= QSPI_MAX_BYTES_FIFO)
drivers/spi/spi-qcom-qspi.c
683
if (op->data.dir == SPI_MEM_DATA_IN && (op->data.nbytes & 0x3))
drivers/spi/spi-qcom-qspi.c
684
op->data.nbytes &= ~0x3;
drivers/spi/spi-qpic-snand.c
1002
return qcom_spi_read_page_oob(snandc, op);
drivers/spi/spi-qpic-snand.c
1027
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1109
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1186
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1239
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1242
return qcom_spi_program_raw(snandc, op);
drivers/spi/spi-qpic-snand.c
1245
return qcom_spi_program_ecc(snandc, op);
drivers/spi/spi-qpic-snand.c
1248
return qcom_spi_program_oob(snandc, op);
drivers/spi/spi-qpic-snand.c
1302
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1307
ret = qcom_spi_cmd_mapping(snandc, op->cmd.opcode, &cmd);
drivers/spi/spi-qpic-snand.c
1311
if (op->cmd.opcode == SPINAND_PROGRAM_LOAD)
drivers/spi/spi-qpic-snand.c
1312
snandc->qspi->data_buf = (u8 *)op->data.buf.out;
drivers/spi/spi-qpic-snand.c
1318
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1323
ret = qcom_spi_cmd_mapping(snandc, op->cmd.opcode, &cmd);
drivers/spi/spi-qpic-snand.c
1327
opcode = op->cmd.opcode;
drivers/spi/spi-qpic-snand.c
1333
snandc->qspi->addr1 = cpu_to_le32(op->addr.val << 16);
drivers/spi/spi-qpic-snand.c
1334
snandc->qspi->addr2 = cpu_to_le32(op->addr.val >> 16 & 0xff);
drivers/spi/spi-qpic-snand.c
1336
return qcom_spi_program_execute(snandc, op);
drivers/spi/spi-qpic-snand.c
1338
snandc->qspi->addr1 = cpu_to_le32(op->addr.val << 16);
drivers/spi/spi-qpic-snand.c
1339
snandc->qspi->addr2 = cpu_to_le32(op->addr.val >> 16 & 0xff);
drivers/spi/spi-qpic-snand.c
1343
snandc->qspi->addr1 = cpu_to_le32(op->addr.val << 16);
drivers/spi/spi-qpic-snand.c
1344
snandc->qspi->addr2 = cpu_to_le32(op->addr.val >> 16 & 0xffff);
drivers/spi/spi-qpic-snand.c
1358
snandc->regs->addr0 = cpu_to_le32(op->addr.val);
drivers/spi/spi-qpic-snand.c
1371
static int qcom_spi_io_op(struct qcom_nand_controller *snandc, const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1376
ret = qcom_spi_send_cmdaddr(snandc, op);
drivers/spi/spi-qpic-snand.c
1384
opcode = op->cmd.opcode;
drivers/spi/spi-qpic-snand.c
1398
snandc->regs->flash_feature = cpu_to_le32(*(u32 *)op->data.buf.out);
drivers/spi/spi-qpic-snand.c
1420
memcpy(op->data.buf.in, snandc->reg_read_buf, snandc->buf_count);
drivers/spi/spi-qpic-snand.c
1427
memcpy(op->data.buf.in, &val, snandc->buf_count);
drivers/spi/spi-qpic-snand.c
1433
static bool qcom_spi_is_page_op(const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1435
if (op->addr.buswidth != 1 && op->addr.buswidth != 2 && op->addr.buswidth != 4)
drivers/spi/spi-qpic-snand.c
1438
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-qpic-snand.c
1439
if (op->addr.buswidth == 4 && op->data.buswidth == 4)
drivers/spi/spi-qpic-snand.c
1442
if (op->addr.nbytes == 2 && op->addr.buswidth == 1)
drivers/spi/spi-qpic-snand.c
1445
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-qpic-snand.c
1446
if (op->data.buswidth == 4)
drivers/spi/spi-qpic-snand.c
1448
if (op->addr.nbytes == 2 && op->addr.buswidth == 1)
drivers/spi/spi-qpic-snand.c
1455
static bool qcom_spi_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1457
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-qpic-snand.c
1460
if (op->cmd.nbytes != 1 || op->cmd.buswidth != 1)
drivers/spi/spi-qpic-snand.c
1463
if (qcom_spi_is_page_op(op))
drivers/spi/spi-qpic-snand.c
1466
return ((!op->addr.nbytes || op->addr.buswidth == 1) &&
drivers/spi/spi-qpic-snand.c
1467
(!op->dummy.nbytes || op->dummy.buswidth == 1) &&
drivers/spi/spi-qpic-snand.c
1468
(!op->data.nbytes || op->data.buswidth == 1));
drivers/spi/spi-qpic-snand.c
1471
static int qcom_spi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
1475
dev_dbg(snandc->dev, "OP %02x ADDR %08llX@%d:%u DATA %d:%u", op->cmd.opcode,
drivers/spi/spi-qpic-snand.c
1476
op->addr.val, op->addr.buswidth, op->addr.nbytes,
drivers/spi/spi-qpic-snand.c
1477
op->data.buswidth, op->data.nbytes);
drivers/spi/spi-qpic-snand.c
1479
if (qcom_spi_is_page_op(op)) {
drivers/spi/spi-qpic-snand.c
1480
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-qpic-snand.c
1481
return qcom_spi_read_page(snandc, op);
drivers/spi/spi-qpic-snand.c
1482
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-qpic-snand.c
1483
return qcom_spi_write_page(snandc, op);
drivers/spi/spi-qpic-snand.c
1485
return qcom_spi_io_op(snandc, op);
drivers/spi/spi-qpic-snand.c
576
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
635
memcpy(op->data.buf.in, snandc->data_buffer + bbpos, op->data.nbytes);
drivers/spi/spi-qpic-snand.c
798
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
806
data_buf = op->data.buf.in;
drivers/spi/spi-qpic-snand.c
826
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
833
data_buf = op->data.buf.in;
drivers/spi/spi-qpic-snand.c
915
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
922
oob_buf = op->data.buf.in;
drivers/spi/spi-qpic-snand.c
990
const struct spi_mem_op *op)
drivers/spi/spi-qpic-snand.c
993
return qcom_spi_read_page_raw(snandc, op);
drivers/spi/spi-qpic-snand.c
996
return qcom_spi_read_page_ecc(snandc, op);
drivers/spi/spi-qpic-snand.c
999
return qcom_spi_read_last_cw(snandc, op);
drivers/spi/spi-realtek-rtl-snand.c
113
if (op->addr.buswidth > 1) {
drivers/spi/spi-realtek-rtl-snand.c
114
val = op->addr.val << ((3 - op->addr.nbytes) * 8);
drivers/spi/spi-realtek-rtl-snand.c
115
len = op->addr.nbytes;
drivers/spi/spi-realtek-rtl-snand.c
118
CMR_WID(op->addr.buswidth) | CMR_LEN(len));
drivers/spi/spi-realtek-rtl-snand.c
131
if (op->dummy.nbytes) {
drivers/spi/spi-realtek-rtl-snand.c
135
CMR_WID(op->dummy.buswidth) | CMR_LEN(op->dummy.nbytes));
drivers/spi/spi-realtek-rtl-snand.c
156
static int rtl_snand_xfer(struct rtl_snand *snand, int cs, const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
162
ret = rtl_snand_xfer_head(snand, cs, op);
drivers/spi/spi-realtek-rtl-snand.c
166
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-realtek-rtl-snand.c
168
len = op->data.nbytes;
drivers/spi/spi-realtek-rtl-snand.c
180
CMR_WID(op->data.buswidth) | CMR_LEN(nbytes));
drivers/spi/spi-realtek-rtl-snand.c
192
memcpy(op->data.buf.in + pos, &val, nbytes);
drivers/spi/spi-realtek-rtl-snand.c
196
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-realtek-rtl-snand.c
198
len = op->data.nbytes;
drivers/spi/spi-realtek-rtl-snand.c
205
memcpy(&val, op->data.buf.out + pos, nbytes);
drivers/spi/spi-realtek-rtl-snand.c
232
static int rtl_snand_dma_xfer(struct rtl_snand *snand, int cs, const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
240
ret = rtl_snand_xfer_head(snand, cs, op);
drivers/spi/spi-realtek-rtl-snand.c
244
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-realtek-rtl-snand.c
248
} else if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-realtek-rtl-snand.c
257
buf_dma = dma_map_single(snand->dev, op->data.buf.in, op->data.nbytes, dir);
drivers/spi/spi-realtek-rtl-snand.c
271
len = op->data.nbytes;
drivers/spi/spi-realtek-rtl-snand.c
287
CMR_WID(op->data.buswidth) | nbytes);
drivers/spi/spi-realtek-rtl-snand.c
305
dma_unmap_single(snand->dev, buf_dma, op->data.nbytes, dir);
drivers/spi/spi-realtek-rtl-snand.c
315
static bool rtl_snand_dma_op(const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
317
switch (op->data.dir) {
drivers/spi/spi-realtek-rtl-snand.c
320
return op->data.nbytes > 32;
drivers/spi/spi-realtek-rtl-snand.c
326
static int rtl_snand_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
332
cs, op->cmd.opcode,
drivers/spi/spi-realtek-rtl-snand.c
333
op->cmd.buswidth, op->cmd.nbytes, op->dummy.buswidth,
drivers/spi/spi-realtek-rtl-snand.c
334
op->dummy.nbytes, op->addr.val, op->addr.buswidth,
drivers/spi/spi-realtek-rtl-snand.c
335
op->addr.nbytes, op->data.buswidth, op->data.nbytes);
drivers/spi/spi-realtek-rtl-snand.c
337
if (rtl_snand_dma_op(op))
drivers/spi/spi-realtek-rtl-snand.c
338
return rtl_snand_dma_xfer(snand, cs, op);
drivers/spi/spi-realtek-rtl-snand.c
340
return rtl_snand_xfer(snand, cs, op);
drivers/spi/spi-realtek-rtl-snand.c
54
const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
56
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-realtek-rtl-snand.c
58
if (op->cmd.nbytes != 1 || op->cmd.buswidth != 1)
drivers/spi/spi-realtek-rtl-snand.c
83
static int rtl_snand_xfer_head(struct rtl_snand *snand, int cs, const struct spi_mem_op *op)
drivers/spi/spi-realtek-rtl-snand.c
90
val = op->cmd.opcode << 24;
drivers/spi/spi-realtek-rtl-snand.c
92
if (op->addr.nbytes && op->addr.buswidth == 1) {
drivers/spi/spi-realtek-rtl-snand.c
93
val |= op->addr.val << ((3 - op->addr.nbytes) * 8);
drivers/spi/spi-realtek-rtl-snand.c
94
len += op->addr.nbytes;
drivers/spi/spi-rockchip-sfc.c
302
static void rockchip_sfc_adjust_op_work(struct spi_mem_op *op)
drivers/spi/spi-rockchip-sfc.c
304
if (unlikely(op->dummy.nbytes && !op->addr.nbytes)) {
drivers/spi/spi-rockchip-sfc.c
309
op->addr.nbytes = op->dummy.nbytes;
drivers/spi/spi-rockchip-sfc.c
310
op->addr.buswidth = op->dummy.buswidth;
drivers/spi/spi-rockchip-sfc.c
311
op->addr.val = 0xFFFFFFFFF;
drivers/spi/spi-rockchip-sfc.c
313
op->dummy.nbytes = 0;
drivers/spi/spi-rockchip-sfc.c
319
const struct spi_mem_op *op,
drivers/spi/spi-rockchip-sfc.c
326
cmd = op->cmd.opcode;
drivers/spi/spi-rockchip-sfc.c
327
ctrl |= ((op->cmd.buswidth >> 1) << SFC_CTRL_CMD_BITS_SHIFT);
drivers/spi/spi-rockchip-sfc.c
330
if (op->addr.nbytes) {
drivers/spi/spi-rockchip-sfc.c
331
if (op->addr.nbytes == 4) {
drivers/spi/spi-rockchip-sfc.c
333
} else if (op->addr.nbytes == 3) {
drivers/spi/spi-rockchip-sfc.c
337
writel(op->addr.nbytes * 8 - 1,
drivers/spi/spi-rockchip-sfc.c
341
ctrl |= ((op->addr.buswidth >> 1) << SFC_CTRL_ADDR_BITS_SHIFT);
drivers/spi/spi-rockchip-sfc.c
345
if (op->dummy.nbytes) {
drivers/spi/spi-rockchip-sfc.c
346
if (op->dummy.buswidth == 4)
drivers/spi/spi-rockchip-sfc.c
347
cmd |= op->dummy.nbytes * 2 << SFC_CMD_DUMMY_SHIFT;
drivers/spi/spi-rockchip-sfc.c
348
else if (op->dummy.buswidth == 2)
drivers/spi/spi-rockchip-sfc.c
349
cmd |= op->dummy.nbytes * 4 << SFC_CMD_DUMMY_SHIFT;
drivers/spi/spi-rockchip-sfc.c
351
cmd |= op->dummy.nbytes * 8 << SFC_CMD_DUMMY_SHIFT;
drivers/spi/spi-rockchip-sfc.c
360
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-rockchip-sfc.c
363
ctrl |= ((op->data.buswidth >> 1) << SFC_CTRL_DATA_BITS_SHIFT);
drivers/spi/spi-rockchip-sfc.c
365
if (!len && op->addr.nbytes)
drivers/spi/spi-rockchip-sfc.c
373
op->addr.nbytes, op->addr.buswidth,
drivers/spi/spi-rockchip-sfc.c
374
op->dummy.nbytes, op->dummy.buswidth);
drivers/spi/spi-rockchip-sfc.c
376
ctrl, cmd, op->addr.val, len);
drivers/spi/spi-rockchip-sfc.c
380
if (op->addr.nbytes)
drivers/spi/spi-rockchip-sfc.c
381
writel(op->addr.val, sfc->regbase + SFC_ADDR);
drivers/spi/spi-rockchip-sfc.c
459
const struct spi_mem_op *op, u32 len)
drivers/spi/spi-rockchip-sfc.c
463
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-rockchip-sfc.c
464
return rockchip_sfc_write_fifo(sfc, op->data.buf.out, len);
drivers/spi/spi-rockchip-sfc.c
466
return rockchip_sfc_read_fifo(sfc, op->data.buf.in, len);
drivers/spi/spi-rockchip-sfc.c
470
const struct spi_mem_op *op, u32 len)
drivers/spi/spi-rockchip-sfc.c
476
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-rockchip-sfc.c
477
memcpy(sfc->buffer, op->data.buf.out, len);
drivers/spi/spi-rockchip-sfc.c
488
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-rockchip-sfc.c
490
memcpy(op->data.buf.in, sfc->buffer, len);
drivers/spi/spi-rockchip-sfc.c
524
static int rockchip_sfc_exec_mem_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-rockchip-sfc.c
527
u32 len = op->data.nbytes;
drivers/spi/spi-rockchip-sfc.c
537
if (unlikely(op->max_freq != sfc->speed[cs]) &&
drivers/spi/spi-rockchip-sfc.c
539
ret = rockchip_sfc_clk_set_rate(sfc, op->max_freq);
drivers/spi/spi-rockchip-sfc.c
542
sfc->speed[cs] = op->max_freq;
drivers/spi/spi-rockchip-sfc.c
547
rockchip_sfc_adjust_op_work((struct spi_mem_op *)op);
drivers/spi/spi-rockchip-sfc.c
548
rockchip_sfc_xfer_setup(sfc, mem, op, len);
drivers/spi/spi-rockchip-sfc.c
553
ret = rockchip_sfc_xfer_data_dma(sfc, op, len);
drivers/spi/spi-rockchip-sfc.c
555
ret = rockchip_sfc_xfer_data_poll(sfc, op, len);
drivers/spi/spi-rockchip-sfc.c
559
dev_err(sfc->dev, "xfer data failed ret %d dir %d\n", ret, op->data.dir);
drivers/spi/spi-rockchip-sfc.c
573
static int rockchip_sfc_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-rockchip-sfc.c
577
op->data.nbytes = min(op->data.nbytes, sfc->max_iosize);
drivers/spi/spi-rpc-if.c
126
const struct spi_mem_op *op)
drivers/spi/spi-rpc-if.c
131
rpcif_spi_mem_prepare(mem->spi, op, NULL, NULL);
drivers/spi/spi-rpc-if.c
65
const struct spi_mem_op *op)
drivers/spi/spi-rpc-if.c
67
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-rpc-if.c
70
if (op->data.buswidth > 4 || op->addr.buswidth > 4 ||
drivers/spi/spi-rpc-if.c
71
op->dummy.buswidth > 4 || op->cmd.buswidth > 4 ||
drivers/spi/spi-rpc-if.c
72
op->addr.nbytes > 4)
drivers/spi/spi-sg2044-nor.c
138
const struct spi_mem_op *op, loff_t from,
drivers/spi/spi-sg2044-nor.c
147
reg |= (op->addr.nbytes + op->dummy.nbytes) << SPIFMC_TRAN_CSR_ADDR_BYTES_SHIFT;
drivers/spi/spi-sg2044-nor.c
153
writeb(op->cmd.opcode, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
155
for (i = op->addr.nbytes - 1; i >= 0; i--)
drivers/spi/spi-sg2044-nor.c
158
for (i = 0; i < op->dummy.nbytes; i++)
drivers/spi/spi-sg2044-nor.c
194
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
198
loff_t from = op->addr.val;
drivers/spi/spi-sg2044-nor.c
199
size_t len = op->data.nbytes;
drivers/spi/spi-sg2044-nor.c
201
u8 *din = op->data.buf.in;
drivers/spi/spi-sg2044-nor.c
207
ret = sg2044_spifmc_read_64k(spifmc, op, from, xfer_size, din);
drivers/spi/spi-sg2044-nor.c
220
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
223
const u8 *dout = op->data.buf.out;
drivers/spi/spi-sg2044-nor.c
229
reg |= (op->addr.nbytes + op->dummy.nbytes) << SPIFMC_TRAN_CSR_ADDR_BYTES_SHIFT;
drivers/spi/spi-sg2044-nor.c
235
writeb(op->cmd.opcode, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
237
for (i = op->addr.nbytes - 1; i >= 0; i--)
drivers/spi/spi-sg2044-nor.c
238
writeb((op->addr.val >> i * 8) & 0xff, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
240
for (i = 0; i < op->dummy.nbytes; i++)
drivers/spi/spi-sg2044-nor.c
244
writel(op->data.nbytes, spifmc->io_base + SPIFMC_TRAN_NUM);
drivers/spi/spi-sg2044-nor.c
255
while (offset < op->data.nbytes) {
drivers/spi/spi-sg2044-nor.c
256
xfer_size = min_t(size_t, SPIFMC_MAX_FIFO_DEPTH, op->data.nbytes - offset);
drivers/spi/spi-sg2044-nor.c
278
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
284
reg |= (op->addr.nbytes + op->dummy.nbytes) << SPIFMC_TRAN_CSR_ADDR_BYTES_SHIFT;
drivers/spi/spi-sg2044-nor.c
289
writeb(op->cmd.opcode, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
291
for (i = op->addr.nbytes - 1; i >= 0; i--)
drivers/spi/spi-sg2044-nor.c
292
writeb((op->addr.val >> i * 8) & 0xff, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
294
for (i = 0; i < op->dummy.nbytes; i++)
drivers/spi/spi-sg2044-nor.c
311
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
313
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-sg2044-nor.c
314
sg2044_spifmc_read(spifmc, op);
drivers/spi/spi-sg2044-nor.c
315
else if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-sg2044-nor.c
316
sg2044_spifmc_write(spifmc, op);
drivers/spi/spi-sg2044-nor.c
318
sg2044_spifmc_tran_cmd(spifmc, op);
drivers/spi/spi-sg2044-nor.c
322
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
326
size_t len = op->data.nbytes;
drivers/spi/spi-sg2044-nor.c
330
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-sg2044-nor.c
331
din = op->data.buf.in;
drivers/spi/spi-sg2044-nor.c
333
dout = op->data.buf.out;
drivers/spi/spi-sg2044-nor.c
352
if (op->cmd.opcode == 0x01) {
drivers/spi/spi-sg2044-nor.c
360
writeb(op->cmd.opcode, spifmc->io_base + SPIFMC_FIFO_PORT);
drivers/spi/spi-sg2044-nor.c
389
const struct spi_mem_op *op)
drivers/spi/spi-sg2044-nor.c
397
if (op->addr.nbytes == 0)
drivers/spi/spi-sg2044-nor.c
398
sg2044_spifmc_trans_reg(spifmc, op);
drivers/spi/spi-sg2044-nor.c
400
sg2044_spifmc_trans(spifmc, op);
drivers/spi/spi-sn-f-ospi.c
117
static u32 f_ospi_get_dummy_cycle(const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
119
if (!op->dummy.nbytes)
drivers/spi/spi-sn-f-ospi.c
122
return (op->dummy.nbytes * 8) / op->dummy.buswidth;
drivers/spi/spi-sn-f-ospi.c
265
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
275
mode = f_ospi_get_mode(ospi, op->cmd.buswidth, 1);
drivers/spi/spi-sn-f-ospi.c
278
mode = f_ospi_get_mode(ospi, op->addr.buswidth, op->addr.nbytes);
drivers/spi/spi-sn-f-ospi.c
281
mode = f_ospi_get_mode(ospi, op->data.buswidth, op->data.nbytes);
drivers/spi/spi-sn-f-ospi.c
297
switch (op->data.nbytes & 0x3) {
drivers/spi/spi-sn-f-ospi.c
304
val = OSPI_DAT_SIZE_EN | (op->data.nbytes - 1);
drivers/spi/spi-sn-f-ospi.c
308
val = OSPI_DAT_SIZE_EN | (op->data.nbytes - 1);
drivers/spi/spi-sn-f-ospi.c
313
switch (op->data.dir) {
drivers/spi/spi-sn-f-ospi.c
331
prot |= FIELD_PREP(OSPI_PROT_ADDR_SIZE_MASK, op->addr.nbytes);
drivers/spi/spi-sn-f-ospi.c
339
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
348
f_ospi_config_clk(ospi, op->max_freq);
drivers/spi/spi-sn-f-ospi.c
350
f_ospi_config_indir_protocol(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
352
writel(f_ospi_get_dummy_cycle(op), ospi->base + OSPI_DMY_INDIR);
drivers/spi/spi-sn-f-ospi.c
353
writel(op->addr.val, ospi->base + OSPI_ADDR);
drivers/spi/spi-sn-f-ospi.c
354
writel(op->cmd.opcode, ospi->base + OSPI_CMD_IDX_INDIR);
drivers/spi/spi-sn-f-ospi.c
358
switch (op->data.dir) {
drivers/spi/spi-sn-f-ospi.c
404
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
406
u8 *buf = op->data.buf.in;
drivers/spi/spi-sn-f-ospi.c
413
ret = f_ospi_indir_prepare_op(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
420
for (i = 0; i < op->data.nbytes; i++) {
drivers/spi/spi-sn-f-ospi.c
458
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
460
u8 *buf = (u8 *)op->data.buf.out;
drivers/spi/spi-sn-f-ospi.c
467
ret = f_ospi_indir_prepare_op(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
477
for (i = 0; i < op->data.nbytes; i++) {
drivers/spi/spi-sn-f-ospi.c
504
static int f_ospi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
509
switch (op->data.dir) {
drivers/spi/spi-sn-f-ospi.c
511
err = f_ospi_indir_read(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
517
err = f_ospi_indir_write(ospi, mem, op);
drivers/spi/spi-sn-f-ospi.c
529
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
532
u8 width_op[] = { op->cmd.buswidth, op->addr.buswidth,
drivers/spi/spi-sn-f-ospi.c
533
op->dummy.buswidth, op->data.buswidth };
drivers/spi/spi-sn-f-ospi.c
555
const struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
557
if (f_ospi_get_dummy_cycle(op) > OSPI_DUMMY_CYCLE_MAX)
drivers/spi/spi-sn-f-ospi.c
560
if (op->addr.nbytes > 4)
drivers/spi/spi-sn-f-ospi.c
563
if (!f_ospi_supports_op_width(mem, op))
drivers/spi/spi-sn-f-ospi.c
566
return spi_mem_default_supports_op(mem, op);
drivers/spi/spi-sn-f-ospi.c
569
static int f_ospi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-sn-f-ospi.c
571
op->data.nbytes = min_t(int, op->data.nbytes, OSPI_DAT_SIZE_MAX);
drivers/spi/spi-stm32-ospi.c
332
const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
334
memcpy_fromio(op->data.buf.in, ospi->mm_base + op->addr.val,
drivers/spi/spi-stm32-ospi.c
335
op->data.nbytes);
drivers/spi/spi-stm32-ospi.c
340
const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
351
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-stm32-ospi.c
363
err = spi_controller_dma_map_mem_op_data(ospi->ctrl, op, &sgt);
drivers/spi/spi-stm32-ospi.c
399
spi_controller_dma_unmap_mem_op_data(ospi->ctrl, op, &sgt);
drivers/spi/spi-stm32-ospi.c
404
static int stm32_ospi_xfer(struct stm32_ospi *ospi, const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
408
if (!op->data.nbytes)
drivers/spi/spi-stm32-ospi.c
412
return stm32_ospi_tx_mm(ospi, op);
drivers/spi/spi-stm32-ospi.c
413
else if (((op->data.dir == SPI_MEM_DATA_IN && ospi->dma_chrx) ||
drivers/spi/spi-stm32-ospi.c
414
(op->data.dir == SPI_MEM_DATA_OUT && ospi->dma_chtx)) &&
drivers/spi/spi-stm32-ospi.c
415
op->data.nbytes > 8)
drivers/spi/spi-stm32-ospi.c
416
if (!stm32_ospi_tx_dma(ospi, op))
drivers/spi/spi-stm32-ospi.c
419
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-stm32-ospi.c
420
buf = op->data.buf.in;
drivers/spi/spi-stm32-ospi.c
422
buf = (void *)op->data.buf.out;
drivers/spi/spi-stm32-ospi.c
424
return stm32_ospi_poll(ospi, buf, op->data.nbytes,
drivers/spi/spi-stm32-ospi.c
425
op->data.dir == SPI_MEM_DATA_IN);
drivers/spi/spi-stm32-ospi.c
429
const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
464
static int stm32_ospi_send(struct spi_device *spi, const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
473
op->cmd.opcode, op->cmd.buswidth, op->addr.buswidth,
drivers/spi/spi-stm32-ospi.c
474
op->dummy.buswidth, op->data.buswidth,
drivers/spi/spi-stm32-ospi.c
475
op->addr.val, op->data.nbytes);
drivers/spi/spi-stm32-ospi.c
484
if (op->data.nbytes)
drivers/spi/spi-stm32-ospi.c
485
writel_relaxed(op->data.nbytes - 1, regs_base + OSPI_DLR);
drivers/spi/spi-stm32-ospi.c
492
ccr = FIELD_PREP(CCR_IMODE_MASK, stm32_ospi_get_mode(op->cmd.buswidth));
drivers/spi/spi-stm32-ospi.c
494
if (op->addr.nbytes) {
drivers/spi/spi-stm32-ospi.c
496
stm32_ospi_get_mode(op->addr.buswidth));
drivers/spi/spi-stm32-ospi.c
497
ccr |= FIELD_PREP(CCR_ADSIZE_MASK, op->addr.nbytes - 1);
drivers/spi/spi-stm32-ospi.c
501
if (op->dummy.buswidth && op->dummy.nbytes) {
drivers/spi/spi-stm32-ospi.c
503
op->dummy.nbytes * 8 / op->dummy.buswidth);
drivers/spi/spi-stm32-ospi.c
507
if (op->data.nbytes) {
drivers/spi/spi-stm32-ospi.c
509
stm32_ospi_get_mode(op->data.buswidth));
drivers/spi/spi-stm32-ospi.c
515
writel_relaxed(op->cmd.opcode, regs_base + OSPI_IR);
drivers/spi/spi-stm32-ospi.c
517
if (op->addr.nbytes && ospi->fmode != CR_FMODE_MM)
drivers/spi/spi-stm32-ospi.c
518
writel_relaxed(op->addr.val, regs_base + OSPI_AR);
drivers/spi/spi-stm32-ospi.c
521
err_poll_status = stm32_ospi_wait_poll_status(ospi, op);
drivers/spi/spi-stm32-ospi.c
523
err = stm32_ospi_xfer(ospi, op);
drivers/spi/spi-stm32-ospi.c
554
const struct spi_mem_op *op,
drivers/spi/spi-stm32-ospi.c
575
ret = stm32_ospi_send(mem->spi, op);
drivers/spi/spi-stm32-ospi.c
583
static int stm32_ospi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-stm32-ospi.c
593
if (op->data.dir == SPI_MEM_DATA_IN && op->data.nbytes)
drivers/spi/spi-stm32-ospi.c
598
ret = stm32_ospi_send(mem->spi, op);
drivers/spi/spi-stm32-ospi.c
627
struct spi_mem_op op;
drivers/spi/spi-stm32-ospi.c
641
memcpy(&op, &desc->info.op_tmpl, sizeof(struct spi_mem_op));
drivers/spi/spi-stm32-ospi.c
644
op.data.nbytes = len;
drivers/spi/spi-stm32-ospi.c
645
op.addr.val = desc->info.offset + offs;
drivers/spi/spi-stm32-ospi.c
646
op.data.buf.in = buf;
drivers/spi/spi-stm32-ospi.c
648
addr_max = op.addr.val + op.data.nbytes + 1;
drivers/spi/spi-stm32-ospi.c
649
if (addr_max < ospi->mm_size && op.addr.buswidth)
drivers/spi/spi-stm32-ospi.c
654
ret = stm32_ospi_send(desc->mem->spi, &op);
drivers/spi/spi-stm32-ospi.c
668
struct spi_mem_op op;
drivers/spi/spi-stm32-ospi.c
686
memset(&op, 0, sizeof(op));
drivers/spi/spi-stm32-ospi.c
699
op.dummy.buswidth = transfer->tx_nbits;
drivers/spi/spi-stm32-ospi.c
700
op.dummy.nbytes = transfer->len;
drivers/spi/spi-stm32-ospi.c
712
op.data.nbytes = transfer->len;
drivers/spi/spi-stm32-ospi.c
716
op.data.buswidth = transfer->rx_nbits;
drivers/spi/spi-stm32-ospi.c
717
op.data.dir = SPI_MEM_DATA_IN;
drivers/spi/spi-stm32-ospi.c
718
op.data.buf.in = transfer->rx_buf;
drivers/spi/spi-stm32-ospi.c
721
op.data.buswidth = transfer->tx_nbits;
drivers/spi/spi-stm32-ospi.c
722
op.data.dir = SPI_MEM_DATA_OUT;
drivers/spi/spi-stm32-ospi.c
723
op.data.buf.out = transfer->tx_buf;
drivers/spi/spi-stm32-ospi.c
726
ret = stm32_ospi_send(spi, &op);
drivers/spi/spi-stm32-qspi.c
174
const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
177
u32 len = op->data.nbytes, sr;
drivers/spi/spi-stm32-qspi.c
182
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-stm32-qspi.c
184
buf = op->data.buf.in;
drivers/spi/spi-stm32-qspi.c
188
buf = (void *)op->data.buf.out;
drivers/spi/spi-stm32-qspi.c
217
const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
219
memcpy_fromio(op->data.buf.in, qspi->mm_base + op->addr.val,
drivers/spi/spi-stm32-qspi.c
220
op->data.nbytes);
drivers/spi/spi-stm32-qspi.c
232
const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
242
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-stm32-qspi.c
254
err = spi_controller_dma_map_mem_op_data(qspi->ctrl, op, &sgt);
drivers/spi/spi-stm32-qspi.c
290
spi_controller_dma_unmap_mem_op_data(qspi->ctrl, op, &sgt);
drivers/spi/spi-stm32-qspi.c
295
static int stm32_qspi_tx(struct stm32_qspi *qspi, const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
297
if (!op->data.nbytes)
drivers/spi/spi-stm32-qspi.c
301
return stm32_qspi_tx_mm(qspi, op);
drivers/spi/spi-stm32-qspi.c
302
else if (((op->data.dir == SPI_MEM_DATA_IN && qspi->dma_chrx) ||
drivers/spi/spi-stm32-qspi.c
303
(op->data.dir == SPI_MEM_DATA_OUT && qspi->dma_chtx)) &&
drivers/spi/spi-stm32-qspi.c
304
op->data.nbytes > 4)
drivers/spi/spi-stm32-qspi.c
305
if (!stm32_qspi_tx_dma(qspi, op))
drivers/spi/spi-stm32-qspi.c
308
return stm32_qspi_tx_poll(qspi, op);
drivers/spi/spi-stm32-qspi.c
369
static int stm32_qspi_send(struct spi_device *spi, const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
382
if (op->data.nbytes)
drivers/spi/spi-stm32-qspi.c
383
writel_relaxed(op->data.nbytes - 1,
drivers/spi/spi-stm32-qspi.c
387
ccr |= FIELD_PREP(CCR_INST_MASK, op->cmd.opcode);
drivers/spi/spi-stm32-qspi.c
389
stm32_qspi_get_mode(op->cmd.buswidth));
drivers/spi/spi-stm32-qspi.c
391
if (op->addr.nbytes) {
drivers/spi/spi-stm32-qspi.c
393
stm32_qspi_get_mode(op->addr.buswidth));
drivers/spi/spi-stm32-qspi.c
394
ccr |= FIELD_PREP(CCR_ADSIZE_MASK, op->addr.nbytes - 1);
drivers/spi/spi-stm32-qspi.c
397
if (op->dummy.nbytes)
drivers/spi/spi-stm32-qspi.c
399
op->dummy.nbytes * 8 / op->dummy.buswidth);
drivers/spi/spi-stm32-qspi.c
401
if (op->data.nbytes) {
drivers/spi/spi-stm32-qspi.c
403
stm32_qspi_get_mode(op->data.buswidth));
drivers/spi/spi-stm32-qspi.c
408
if (op->addr.nbytes && qspi->fmode != CCR_FMODE_MM)
drivers/spi/spi-stm32-qspi.c
409
writel_relaxed(op->addr.val, qspi->io_base + QSPI_AR);
drivers/spi/spi-stm32-qspi.c
414
err = stm32_qspi_tx(qspi, op);
drivers/spi/spi-stm32-qspi.c
451
static int stm32_qspi_poll_status(struct spi_mem *mem, const struct spi_mem_op *op,
drivers/spi/spi-stm32-qspi.c
460
if (!spi_mem_supports_op(mem, op))
drivers/spi/spi-stm32-qspi.c
474
ret = stm32_qspi_send(mem->spi, op);
drivers/spi/spi-stm32-qspi.c
482
static int stm32_qspi_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-stm32-qspi.c
492
if (op->data.dir == SPI_MEM_DATA_IN && op->data.nbytes)
drivers/spi/spi-stm32-qspi.c
497
ret = stm32_qspi_send(mem->spi, op);
drivers/spi/spi-stm32-qspi.c
526
struct spi_mem_op op;
drivers/spi/spi-stm32-qspi.c
539
memcpy(&op, &desc->info.op_tmpl, sizeof(struct spi_mem_op));
drivers/spi/spi-stm32-qspi.c
542
op.data.nbytes = len;
drivers/spi/spi-stm32-qspi.c
543
op.addr.val = desc->info.offset + offs;
drivers/spi/spi-stm32-qspi.c
544
op.data.buf.in = buf;
drivers/spi/spi-stm32-qspi.c
546
addr_max = op.addr.val + op.data.nbytes + 1;
drivers/spi/spi-stm32-qspi.c
547
if (addr_max < qspi->mm_size && op.addr.buswidth)
drivers/spi/spi-stm32-qspi.c
552
ret = stm32_qspi_send(desc->mem->spi, &op);
drivers/spi/spi-stm32-qspi.c
566
struct spi_mem_op op;
drivers/spi/spi-stm32-qspi.c
583
memset(&op, 0, sizeof(op));
drivers/spi/spi-stm32-qspi.c
596
op.dummy.buswidth = transfer->tx_nbits;
drivers/spi/spi-stm32-qspi.c
597
op.dummy.nbytes = transfer->len;
drivers/spi/spi-stm32-qspi.c
609
op.data.nbytes = transfer->len;
drivers/spi/spi-stm32-qspi.c
613
op.data.buswidth = transfer->rx_nbits;
drivers/spi/spi-stm32-qspi.c
614
op.data.dir = SPI_MEM_DATA_IN;
drivers/spi/spi-stm32-qspi.c
615
op.data.buf.in = transfer->rx_buf;
drivers/spi/spi-stm32-qspi.c
618
op.data.buswidth = transfer->tx_nbits;
drivers/spi/spi-stm32-qspi.c
619
op.data.dir = SPI_MEM_DATA_OUT;
drivers/spi/spi-stm32-qspi.c
620
op.data.buf.out = transfer->tx_buf;
drivers/spi/spi-stm32-qspi.c
623
ret = stm32_qspi_send(spi, &op);
drivers/spi/spi-ti-qspi.c
575
static int ti_qspi_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-ti-qspi.c
580
if (op->data.dir == SPI_MEM_DATA_IN) {
drivers/spi/spi-ti-qspi.c
581
if (op->addr.val < qspi->mmap_size) {
drivers/spi/spi-ti-qspi.c
583
if (op->addr.val + op->data.nbytes > qspi->mmap_size) {
drivers/spi/spi-ti-qspi.c
584
max_len = qspi->mmap_size - op->addr.val;
drivers/spi/spi-ti-qspi.c
585
op->data.nbytes = min((size_t) op->data.nbytes,
drivers/spi/spi-ti-qspi.c
595
max_len -= 1 + op->addr.nbytes + op->dummy.nbytes;
drivers/spi/spi-ti-qspi.c
596
op->data.nbytes = min((size_t) op->data.nbytes,
drivers/spi/spi-ti-qspi.c
605
const struct spi_mem_op *op)
drivers/spi/spi-ti-qspi.c
612
if (!op->data.nbytes || op->data.dir != SPI_MEM_DATA_IN ||
drivers/spi/spi-ti-qspi.c
613
!op->addr.nbytes || op->addr.nbytes > 4)
drivers/spi/spi-ti-qspi.c
617
from = op->addr.val;
drivers/spi/spi-ti-qspi.c
618
if (from + op->data.nbytes > qspi->mmap_size)
drivers/spi/spi-ti-qspi.c
624
ti_qspi_setup_clk(qspi, op->max_freq);
drivers/spi/spi-ti-qspi.c
627
ti_qspi_setup_mmap_read(mem->spi, op->cmd.opcode, op->data.buswidth,
drivers/spi/spi-ti-qspi.c
628
op->addr.nbytes, op->dummy.nbytes);
drivers/spi/spi-ti-qspi.c
633
if (virt_addr_valid(op->data.buf.in) &&
drivers/spi/spi-ti-qspi.c
634
!spi_controller_dma_map_mem_op_data(mem->spi->controller, op,
drivers/spi/spi-ti-qspi.c
638
op, &sgt);
drivers/spi/spi-ti-qspi.c
641
op->data.buf.in,
drivers/spi/spi-ti-qspi.c
642
op->data.nbytes);
drivers/spi/spi-ti-qspi.c
645
memcpy_fromio(op->data.buf.in, qspi->mmap_base + from,
drivers/spi/spi-ti-qspi.c
646
op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
136
bool (*match)(const struct spi_mem_op *op);
drivers/spi/spi-wpcm-fiu.c
137
int (*exec)(struct spi_mem *mem, const struct spi_mem_op *op);
drivers/spi/spi-wpcm-fiu.c
140
static bool wpcm_fiu_normal_match(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
143
if (op->cmd.opcode == 0x0b)
drivers/spi/spi-wpcm-fiu.c
146
return (op->addr.nbytes == 0 || op->addr.nbytes == 3) &&
drivers/spi/spi-wpcm-fiu.c
147
op->dummy.nbytes == 0 && op->data.nbytes <= 4;
drivers/spi/spi-wpcm-fiu.c
150
static int wpcm_fiu_normal_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
155
wpcm_fiu_set_opcode(fiu, op->cmd.opcode);
drivers/spi/spi-wpcm-fiu.c
156
wpcm_fiu_set_addr(fiu, op->addr.val);
drivers/spi/spi-wpcm-fiu.c
157
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-wpcm-fiu.c
158
wpcm_fiu_set_data(fiu, op->data.buf.out, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
160
ret = wpcm_fiu_do_uma(fiu, spi_get_chipselect(mem->spi, 0), op->addr.nbytes == 3,
drivers/spi/spi-wpcm-fiu.c
161
op->data.dir == SPI_MEM_DATA_OUT, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
163
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-wpcm-fiu.c
164
wpcm_fiu_get_data(fiu, op->data.buf.in, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
169
static bool wpcm_fiu_fast_read_match(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
171
return op->cmd.opcode == 0x0b && op->addr.nbytes == 3 &&
drivers/spi/spi-wpcm-fiu.c
172
op->dummy.nbytes == 1 &&
drivers/spi/spi-wpcm-fiu.c
173
op->data.nbytes >= 1 && op->data.nbytes <= 4 &&
drivers/spi/spi-wpcm-fiu.c
174
op->data.dir == SPI_MEM_DATA_IN;
drivers/spi/spi-wpcm-fiu.c
177
static int wpcm_fiu_fast_read_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
190
static bool wpcm_fiu_4ba_match(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
192
return op->addr.nbytes == 4 && op->dummy.nbytes == 0 && op->data.nbytes <= 4;
drivers/spi/spi-wpcm-fiu.c
195
static int wpcm_fiu_4ba_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
202
wpcm_fiu_set_opcode(fiu, op->cmd.opcode);
drivers/spi/spi-wpcm-fiu.c
203
wpcm_fiu_set_addr(fiu, op->addr.val >> 8);
drivers/spi/spi-wpcm-fiu.c
206
wpcm_fiu_set_opcode(fiu, op->addr.val & 0xff);
drivers/spi/spi-wpcm-fiu.c
208
if (op->data.dir == SPI_MEM_DATA_OUT)
drivers/spi/spi-wpcm-fiu.c
209
wpcm_fiu_set_data(fiu, op->data.buf.out, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
210
wpcm_fiu_do_uma(fiu, cs, false, op->data.dir == SPI_MEM_DATA_OUT, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
214
if (op->data.dir == SPI_MEM_DATA_IN)
drivers/spi/spi-wpcm-fiu.c
215
wpcm_fiu_get_data(fiu, op->data.buf.in, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
233
static bool wpcm_fiu_rdid_match(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
235
return op->cmd.opcode == 0x9f && op->addr.nbytes == 0 &&
drivers/spi/spi-wpcm-fiu.c
236
op->dummy.nbytes == 0 && op->data.nbytes == 6 &&
drivers/spi/spi-wpcm-fiu.c
237
op->data.dir == SPI_MEM_DATA_IN;
drivers/spi/spi-wpcm-fiu.c
240
static int wpcm_fiu_rdid_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
246
wpcm_fiu_set_opcode(fiu, op->cmd.opcode);
drivers/spi/spi-wpcm-fiu.c
249
wpcm_fiu_get_data(fiu, op->data.buf.in, 3);
drivers/spi/spi-wpcm-fiu.c
252
wpcm_fiu_set_opcode(fiu, op->cmd.opcode);
drivers/spi/spi-wpcm-fiu.c
255
wpcm_fiu_get_data(fiu, op->data.buf.in + 3, 3);
drivers/spi/spi-wpcm-fiu.c
266
static bool wpcm_fiu_dummy_match(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
269
if (op->cmd.opcode == 0x0b)
drivers/spi/spi-wpcm-fiu.c
272
return (op->addr.nbytes == 0 || op->addr.nbytes == 3) &&
drivers/spi/spi-wpcm-fiu.c
273
op->dummy.nbytes >= 1 && op->dummy.nbytes <= 5 &&
drivers/spi/spi-wpcm-fiu.c
274
op->data.nbytes <= 4;
drivers/spi/spi-wpcm-fiu.c
277
static int wpcm_fiu_dummy_exec(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
285
wpcm_fiu_set_opcode(fiu, op->cmd.opcode);
drivers/spi/spi-wpcm-fiu.c
286
wpcm_fiu_set_addr(fiu, op->addr.val);
drivers/spi/spi-wpcm-fiu.c
287
wpcm_fiu_do_uma(fiu, cs, op->addr.nbytes != 0, true, op->dummy.nbytes - 1);
drivers/spi/spi-wpcm-fiu.c
292
wpcm_fiu_do_uma(fiu, cs, false, false, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
293
wpcm_fiu_get_data(fiu, op->data.buf.in, op->data.nbytes);
drivers/spi/spi-wpcm-fiu.c
308
static const struct wpcm_fiu_op_shape *wpcm_fiu_find_op_shape(const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
315
if (shape->match(op))
drivers/spi/spi-wpcm-fiu.c
322
static bool wpcm_fiu_supports_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
324
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-wpcm-fiu.c
327
if (op->cmd.dtr || op->addr.dtr || op->dummy.dtr || op->data.dtr)
drivers/spi/spi-wpcm-fiu.c
330
if (op->cmd.buswidth > 1 || op->addr.buswidth > 1 ||
drivers/spi/spi-wpcm-fiu.c
331
op->dummy.buswidth > 1 || op->data.buswidth > 1)
drivers/spi/spi-wpcm-fiu.c
334
return wpcm_fiu_find_op_shape(op) != NULL;
drivers/spi/spi-wpcm-fiu.c
352
static int wpcm_fiu_exec_op(struct spi_mem *mem, const struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
355
const struct wpcm_fiu_op_shape *shape = wpcm_fiu_find_op_shape(op);
drivers/spi/spi-wpcm-fiu.c
360
return shape->exec(mem, op);
drivers/spi/spi-wpcm-fiu.c
367
static int wpcm_fiu_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op)
drivers/spi/spi-wpcm-fiu.c
369
if (op->data.nbytes > 4)
drivers/spi/spi-wpcm-fiu.c
370
op->data.nbytes = 4;
drivers/spi/spi-zynq-qspi.c
226
const struct spi_mem_op *op)
drivers/spi/spi-zynq-qspi.c
228
if (!spi_mem_default_supports_op(mem, op))
drivers/spi/spi-zynq-qspi.c
234
if (op->addr.nbytes > 3)
drivers/spi/spi-zynq-qspi.c
336
const struct spi_mem_op *op)
drivers/spi/spi-zynq-qspi.c
351
op->max_freq)
drivers/spi/spi-zynq-qspi.c
537
const struct spi_mem_op *op)
drivers/spi/spi-zynq-qspi.c
544
zynq_qspi_config_op(xqspi, mem->spi, op);
drivers/spi/spi-zynq-qspi.c
546
if (op->cmd.opcode) {
drivers/spi/spi-zynq-qspi.c
548
xqspi->txbuf = (u8 *)&op->cmd.opcode;
drivers/spi/spi-zynq-qspi.c
550
xqspi->tx_bytes = op->cmd.nbytes;
drivers/spi/spi-zynq-qspi.c
551
xqspi->rx_bytes = op->cmd.nbytes;
drivers/spi/spi-zynq-qspi.c
560
if (op->addr.nbytes) {
drivers/spi/spi-zynq-qspi.c
561
for (i = 0; i < op->addr.nbytes; i++) {
drivers/spi/spi-zynq-qspi.c
562
xqspi->txbuf[i] = op->addr.val >>
drivers/spi/spi-zynq-qspi.c
563
(8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-zynq-qspi.c
568
xqspi->tx_bytes = op->addr.nbytes;
drivers/spi/spi-zynq-qspi.c
569
xqspi->rx_bytes = op->addr.nbytes;
drivers/spi/spi-zynq-qspi.c
578
if (op->dummy.nbytes) {
drivers/spi/spi-zynq-qspi.c
579
tmpbuf = kmalloc(op->dummy.nbytes, GFP_KERNEL);
drivers/spi/spi-zynq-qspi.c
583
memset(tmpbuf, 0xff, op->dummy.nbytes);
drivers/spi/spi-zynq-qspi.c
587
xqspi->tx_bytes = op->dummy.nbytes;
drivers/spi/spi-zynq-qspi.c
588
xqspi->rx_bytes = op->dummy.nbytes;
drivers/spi/spi-zynq-qspi.c
599
if (op->data.nbytes) {
drivers/spi/spi-zynq-qspi.c
601
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-zynq-qspi.c
602
xqspi->txbuf = (u8 *)op->data.buf.out;
drivers/spi/spi-zynq-qspi.c
603
xqspi->tx_bytes = op->data.nbytes;
drivers/spi/spi-zynq-qspi.c
605
xqspi->rx_bytes = op->data.nbytes;
drivers/spi/spi-zynq-qspi.c
608
xqspi->rxbuf = (u8 *)op->data.buf.in;
drivers/spi/spi-zynq-qspi.c
609
xqspi->rx_bytes = op->data.nbytes;
drivers/spi/spi-zynq-qspi.c
610
xqspi->tx_bytes = op->data.nbytes;
drivers/spi/spi-zynqmp-gqspi.c
1045
const struct spi_mem_op *op)
drivers/spi/spi-zynqmp-gqspi.c
1052
u16 opcode = op->cmd.opcode;
drivers/spi/spi-zynqmp-gqspi.c
1056
zynqmp_qspi_config_op(xqspi, op);
drivers/spi/spi-zynqmp-gqspi.c
1061
if (op->cmd.opcode) {
drivers/spi/spi-zynqmp-gqspi.c
1065
xqspi->bytes_to_transfer = op->cmd.nbytes;
drivers/spi/spi-zynqmp-gqspi.c
1067
zynqmp_qspi_write_op(xqspi, op->cmd.buswidth, genfifoentry);
drivers/spi/spi-zynqmp-gqspi.c
1074
timeout = zynqmp_qspi_timeout(xqspi, op->cmd.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1075
op->cmd.nbytes);
drivers/spi/spi-zynqmp-gqspi.c
1083
if (op->addr.nbytes) {
drivers/spi/spi-zynqmp-gqspi.c
1085
for (i = 0; i < op->addr.nbytes; i++) {
drivers/spi/spi-zynqmp-gqspi.c
1086
*(((u8 *)xqspi->txbuf) + i) = op->addr.val >>
drivers/spi/spi-zynqmp-gqspi.c
1087
(8 * (op->addr.nbytes - i - 1));
drivers/spi/spi-zynqmp-gqspi.c
1092
xqspi->bytes_to_transfer = op->addr.nbytes;
drivers/spi/spi-zynqmp-gqspi.c
1094
zynqmp_qspi_write_op(xqspi, op->addr.buswidth, genfifoentry);
drivers/spi/spi-zynqmp-gqspi.c
1103
timeout = zynqmp_qspi_timeout(xqspi, op->addr.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1104
op->addr.nbytes);
drivers/spi/spi-zynqmp-gqspi.c
1112
if (op->dummy.nbytes) {
drivers/spi/spi-zynqmp-gqspi.c
1119
xqspi->bytes_to_transfer = op->dummy.nbytes * 8 / op->dummy.buswidth;
drivers/spi/spi-zynqmp-gqspi.c
1125
zynqmp_qspi_write_op(xqspi, op->data.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1132
if (op->data.nbytes) {
drivers/spi/spi-zynqmp-gqspi.c
1134
if (op->data.dir == SPI_MEM_DATA_OUT) {
drivers/spi/spi-zynqmp-gqspi.c
1135
xqspi->txbuf = (u8 *)op->data.buf.out;
drivers/spi/spi-zynqmp-gqspi.c
1137
xqspi->bytes_to_transfer = op->data.nbytes;
drivers/spi/spi-zynqmp-gqspi.c
1139
zynqmp_qspi_write_op(xqspi, op->data.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1151
xqspi->rxbuf = (u8 *)op->data.buf.in;
drivers/spi/spi-zynqmp-gqspi.c
1152
xqspi->bytes_to_receive = op->data.nbytes;
drivers/spi/spi-zynqmp-gqspi.c
1154
err = zynqmp_qspi_read_op(xqspi, op->data.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1174
timeout = zynqmp_qspi_timeout(xqspi, op->data.buswidth,
drivers/spi/spi-zynqmp-gqspi.c
1175
op->data.nbytes);
drivers/spi/spi-zynqmp-gqspi.c
555
const struct spi_mem_op *op)
drivers/spi/spi-zynqmp-gqspi.c
560
req_speed_hz = op->max_freq;
drivers/staging/greybus/audio_module.c
141
static int gbaudio_codec_request_handler(struct gb_operation *op)
drivers/staging/greybus/audio_module.c
143
struct gb_connection *connection = op->connection;
drivers/staging/greybus/audio_module.c
146
struct gb_operation_msg_hdr *header = op->request->header;
drivers/staging/greybus/audio_module.c
154
stream_req = op->request->payload;
drivers/staging/greybus/audio_module.c
159
jack_req = op->request->payload;
drivers/staging/greybus/audio_module.c
164
button_req = op->request->payload;
drivers/staging/greybus/authentication.c
112
struct gb_operation *op;
drivers/staging/greybus/authentication.c
115
op = gb_operation_create_flags(connection,
drivers/staging/greybus/authentication.c
120
if (!op)
drivers/staging/greybus/authentication.c
123
request = op->request->payload;
drivers/staging/greybus/authentication.c
127
ret = gb_operation_request_send_sync(op);
drivers/staging/greybus/authentication.c
133
response = op->response->payload;
drivers/staging/greybus/authentication.c
135
*size = op->response->payload_size - sizeof(*response);
drivers/staging/greybus/authentication.c
139
gb_operation_put(op);
drivers/staging/greybus/authentication.c
151
struct gb_operation *op;
drivers/staging/greybus/authentication.c
154
op = gb_operation_create_flags(connection, GB_CAP_TYPE_AUTHENTICATE,
drivers/staging/greybus/authentication.c
158
if (!op)
drivers/staging/greybus/authentication.c
161
request = op->request->payload;
drivers/staging/greybus/authentication.c
166
ret = gb_operation_request_send_sync(op);
drivers/staging/greybus/authentication.c
172
response = op->response->payload;
drivers/staging/greybus/authentication.c
174
*signature_size = op->response->payload_size - sizeof(*response);
drivers/staging/greybus/authentication.c
179
gb_operation_put(op);
drivers/staging/greybus/bootrom.c
188
static int gb_bootrom_firmware_size_request(struct gb_operation *op)
drivers/staging/greybus/bootrom.c
190
struct gb_bootrom *bootrom = gb_connection_get_data(op->connection);
drivers/staging/greybus/bootrom.c
192
op->request->payload;
drivers/staging/greybus/bootrom.c
194
struct device *dev = &op->connection->bundle->dev;
drivers/staging/greybus/bootrom.c
200
if (op->request->payload_size != sizeof(*size_request)) {
drivers/staging/greybus/bootrom.c
202
__func__, op->request->payload_size,
drivers/staging/greybus/bootrom.c
214
if (!gb_operation_response_alloc(op, sizeof(*size_response),
drivers/staging/greybus/bootrom.c
222
size_response = op->response->payload;
drivers/staging/greybus/bootrom.c
241
static int gb_bootrom_get_firmware(struct gb_operation *op)
drivers/staging/greybus/bootrom.c
243
struct gb_bootrom *bootrom = gb_connection_get_data(op->connection);
drivers/staging/greybus/bootrom.c
246
struct device *dev = &op->connection->bundle->dev;
drivers/staging/greybus/bootrom.c
255
if (op->request->payload_size != sizeof(*firmware_request)) {
drivers/staging/greybus/bootrom.c
257
__func__, op->request->payload_size,
drivers/staging/greybus/bootrom.c
272
firmware_request = op->request->payload;
drivers/staging/greybus/bootrom.c
284
if (!gb_operation_response_alloc(op, size, GFP_KERNEL)) {
drivers/staging/greybus/bootrom.c
290
firmware_response = op->response->payload;
drivers/staging/greybus/bootrom.c
311
static int gb_bootrom_ready_to_boot(struct gb_operation *op)
drivers/staging/greybus/bootrom.c
313
struct gb_connection *connection = op->connection;
drivers/staging/greybus/bootrom.c
323
if (op->request->payload_size != sizeof(*rtb_request)) {
drivers/staging/greybus/bootrom.c
325
__func__, op->request->payload_size,
drivers/staging/greybus/bootrom.c
331
rtb_request = op->request->payload;
drivers/staging/greybus/bootrom.c
357
static int gb_bootrom_request_handler(struct gb_operation *op)
drivers/staging/greybus/bootrom.c
359
u8 type = op->type;
drivers/staging/greybus/bootrom.c
363
return gb_bootrom_firmware_size_request(op);
drivers/staging/greybus/bootrom.c
365
return gb_bootrom_get_firmware(op);
drivers/staging/greybus/bootrom.c
367
return gb_bootrom_ready_to_boot(op);
drivers/staging/greybus/bootrom.c
369
dev_err(&op->connection->bundle->dev,
drivers/staging/greybus/camera.c
1088
const struct gb_camera_debugfs_entry *op = file->private_data;
drivers/staging/greybus/camera.c
1094
if (!(op->mask & 0222)) {
drivers/staging/greybus/camera.c
1095
ret = op->execute(gcam, NULL, 0);
drivers/staging/greybus/camera.c
1100
buffer = &gcam->debugfs.buffers[op->buffer];
drivers/staging/greybus/camera.c
1110
const struct gb_camera_debugfs_entry *op = file->private_data;
drivers/staging/greybus/camera.c
1122
ret = op->execute(gcam, kbuf, len);
drivers/staging/greybus/camera.c
715
static int gb_camera_request_handler(struct gb_operation *op)
drivers/staging/greybus/camera.c
717
struct gb_camera *gcam = gb_connection_get_data(op->connection);
drivers/staging/greybus/camera.c
721
if (op->type != GB_CAMERA_TYPE_METADATA) {
drivers/staging/greybus/camera.c
722
dev_err(&gcam->bundle->dev, "Unsupported unsolicited event: %u\n", op->type);
drivers/staging/greybus/camera.c
726
request = op->request;
drivers/staging/greybus/firmware.h
26
int gb_fw_mgmt_request_handler(struct gb_operation *op);
drivers/staging/greybus/firmware.h
31
int gb_fw_download_request_handler(struct gb_operation *op);
drivers/staging/greybus/fw-download.c
221
static int fw_download_find_firmware(struct gb_operation *op)
drivers/staging/greybus/fw-download.c
223
struct gb_connection *connection = op->connection;
drivers/staging/greybus/fw-download.c
230
if (op->request->payload_size != sizeof(*request)) {
drivers/staging/greybus/fw-download.c
233
op->request->payload_size, sizeof(*request));
drivers/staging/greybus/fw-download.c
237
request = op->request->payload;
drivers/staging/greybus/fw-download.c
252
if (!gb_operation_response_alloc(op, sizeof(*response), GFP_KERNEL)) {
drivers/staging/greybus/fw-download.c
258
response = op->response->payload;
drivers/staging/greybus/fw-download.c
268
static int fw_download_fetch_firmware(struct gb_operation *op)
drivers/staging/greybus/fw-download.c
270
struct gb_connection *connection = op->connection;
drivers/staging/greybus/fw-download.c
280
if (op->request->payload_size != sizeof(*request)) {
drivers/staging/greybus/fw-download.c
283
op->request->payload_size, sizeof(*request));
drivers/staging/greybus/fw-download.c
287
request = op->request->payload;
drivers/staging/greybus/fw-download.c
328
if (!gb_operation_response_alloc(op, size, GFP_KERNEL)) {
drivers/staging/greybus/fw-download.c
335
response = op->response->payload;
drivers/staging/greybus/fw-download.c
351
static int fw_download_release_firmware(struct gb_operation *op)
drivers/staging/greybus/fw-download.c
353
struct gb_connection *connection = op->connection;
drivers/staging/greybus/fw-download.c
359
if (op->request->payload_size != sizeof(*request)) {
drivers/staging/greybus/fw-download.c
362
op->request->payload_size, sizeof(*request));
drivers/staging/greybus/fw-download.c
366
request = op->request->payload;
drivers/staging/greybus/fw-download.c
386
int gb_fw_download_request_handler(struct gb_operation *op)
drivers/staging/greybus/fw-download.c
388
u8 type = op->type;
drivers/staging/greybus/fw-download.c
392
return fw_download_find_firmware(op);
drivers/staging/greybus/fw-download.c
394
return fw_download_fetch_firmware(op);
drivers/staging/greybus/fw-download.c
396
return fw_download_release_firmware(op);
drivers/staging/greybus/fw-download.c
398
dev_err(&op->connection->bundle->dev,
drivers/staging/greybus/fw-management.c
185
static int fw_mgmt_interface_fw_loaded_operation(struct gb_operation *op)
drivers/staging/greybus/fw-management.c
187
struct gb_connection *connection = op->connection;
drivers/staging/greybus/fw-management.c
198
if (op->request->payload_size != sizeof(*request)) {
drivers/staging/greybus/fw-management.c
200
op->request->payload_size, sizeof(*request));
drivers/staging/greybus/fw-management.c
204
request = op->request->payload;
drivers/staging/greybus/fw-management.c
327
static int fw_mgmt_backend_fw_updated_operation(struct gb_operation *op)
drivers/staging/greybus/fw-management.c
329
struct gb_connection *connection = op->connection;
drivers/staging/greybus/fw-management.c
339
if (op->request->payload_size != sizeof(*request)) {
drivers/staging/greybus/fw-management.c
341
op->request->payload_size, sizeof(*request));
drivers/staging/greybus/fw-management.c
345
request = op->request->payload;
drivers/staging/greybus/fw-management.c
556
int gb_fw_mgmt_request_handler(struct gb_operation *op)
drivers/staging/greybus/fw-management.c
558
u8 type = op->type;
drivers/staging/greybus/fw-management.c
562
return fw_mgmt_interface_fw_loaded_operation(op);
drivers/staging/greybus/fw-management.c
564
return fw_mgmt_backend_fw_updated_operation(op);
drivers/staging/greybus/fw-management.c
566
dev_err(&op->connection->bundle->dev,
drivers/staging/greybus/gb-camera.h
120
#define gb_camera_call(f, op, args...) \
drivers/staging/greybus/gb-camera.h
121
(!(f) ? -ENODEV : (((f)->ops->op) ? \
drivers/staging/greybus/gb-camera.h
122
(f)->ops->op((f)->priv, ##args) : -ENOIOCTLCMD))
drivers/staging/greybus/gpio.c
357
static int gb_gpio_request_handler(struct gb_operation *op)
drivers/staging/greybus/gpio.c
359
struct gb_connection *connection = op->connection;
drivers/staging/greybus/gpio.c
364
u8 type = op->type;
drivers/staging/greybus/gpio.c
372
request = op->request;
drivers/staging/greybus/hid.c
133
static int gb_hid_request_handler(struct gb_operation *op)
drivers/staging/greybus/hid.c
135
struct gb_connection *connection = op->connection;
drivers/staging/greybus/hid.c
137
struct gb_hid_input_report_request *request = op->request->payload;
drivers/staging/greybus/hid.c
139
if (op->type != GB_HID_TYPE_IRQ_EVENT) {
drivers/staging/greybus/hid.c
147
request->report, op->request->payload_size, 1);
drivers/staging/greybus/i2c.c
112
request_size += msg_count * sizeof(*op);
drivers/staging/greybus/i2c.c
124
op = &request->ops[0];
drivers/staging/greybus/i2c.c
127
gb_i2c_fill_transfer_op(op++, msg++);
drivers/staging/greybus/i2c.c
133
data = op;
drivers/staging/greybus/i2c.c
68
gb_i2c_fill_transfer_op(struct gb_i2c_transfer_op *op, struct i2c_msg *msg)
drivers/staging/greybus/i2c.c
72
op->addr = cpu_to_le16(msg->addr);
drivers/staging/greybus/i2c.c
73
op->flags = cpu_to_le16(flags);
drivers/staging/greybus/i2c.c
74
op->size = cpu_to_le16(msg->len);
drivers/staging/greybus/i2c.c
84
struct gb_i2c_transfer_op *op;
drivers/staging/greybus/light.c
1194
static int gb_lights_request_handler(struct gb_operation *op)
drivers/staging/greybus/light.c
1196
struct gb_connection *connection = op->connection;
drivers/staging/greybus/light.c
1206
if (op->type != GB_LIGHTS_TYPE_EVENT) {
drivers/staging/greybus/light.c
1207
dev_err(dev, "Unsupported unsolicited event: %u\n", op->type);
drivers/staging/greybus/light.c
1211
request = op->request;
drivers/staging/greybus/log.c
18
static int gb_log_request_handler(struct gb_operation *op)
drivers/staging/greybus/log.c
20
struct gb_connection *connection = op->connection;
drivers/staging/greybus/log.c
25
if (op->type != GB_LOG_TYPE_SEND_LOG) {
drivers/staging/greybus/log.c
26
dev_err(dev, "unknown request type 0x%02x\n", op->type);
drivers/staging/greybus/log.c
31
if (op->request->payload_size < sizeof(*receive)) {
drivers/staging/greybus/log.c
33
op->request->payload_size, sizeof(*receive));
drivers/staging/greybus/log.c
36
receive = op->request->payload;
drivers/staging/greybus/log.c
38
if (len != (op->request->payload_size - sizeof(*receive))) {
drivers/staging/greybus/log.c
40
(op->request->payload_size - sizeof(*receive)));
drivers/staging/greybus/power_supply.c
1003
request = op->request;
drivers/staging/greybus/power_supply.c
511
struct gb_operation *op;
drivers/staging/greybus/power_supply.c
520
op = gb_operation_create(connection,
drivers/staging/greybus/power_supply.c
525
if (!op)
drivers/staging/greybus/power_supply.c
528
req = op->request->payload;
drivers/staging/greybus/power_supply.c
531
ret = gb_operation_request_send_sync(op);
drivers/staging/greybus/power_supply.c
535
resp = op->response->payload;
drivers/staging/greybus/power_supply.c
583
gb_operation_put(op);
drivers/staging/greybus/power_supply.c
986
static int gb_supplies_request_handler(struct gb_operation *op)
drivers/staging/greybus/power_supply.c
988
struct gb_connection *connection = op->connection;
drivers/staging/greybus/power_supply.c
997
if (op->type != GB_POWER_SUPPLY_TYPE_EVENT) {
drivers/staging/greybus/power_supply.c
999
"Unsupported unsolicited event: %u\n", op->type);
drivers/staging/greybus/raw.c
100
if (op->type != GB_RAW_TYPE_SEND) {
drivers/staging/greybus/raw.c
101
dev_err(dev, "unknown request type 0x%02x\n", op->type);
drivers/staging/greybus/raw.c
106
if (op->request->payload_size < sizeof(*receive)) {
drivers/staging/greybus/raw.c
108
op->request->payload_size, sizeof(*receive));
drivers/staging/greybus/raw.c
111
receive = op->request->payload;
drivers/staging/greybus/raw.c
113
if (len != (int)(op->request->payload_size - sizeof(__le32))) {
drivers/staging/greybus/raw.c
115
(int)(op->request->payload_size - sizeof(__le32)));
drivers/staging/greybus/raw.c
92
static int gb_raw_request_handler(struct gb_operation *op)
drivers/staging/greybus/raw.c
94
struct gb_connection *connection = op->connection;
drivers/staging/greybus/sdio.c
203
static int gb_sdio_request_handler(struct gb_operation *op)
drivers/staging/greybus/sdio.c
205
struct gb_sdio_host *host = gb_connection_get_data(op->connection);
drivers/staging/greybus/sdio.c
208
u8 type = op->type;
drivers/staging/greybus/sdio.c
218
request = op->request;
drivers/staging/greybus/uart.c
129
static int gb_uart_serial_state_handler(struct gb_operation *op)
drivers/staging/greybus/uart.c
131
struct gb_connection *connection = op->connection;
drivers/staging/greybus/uart.c
133
struct gb_message *request = op->request;
drivers/staging/greybus/uart.c
149
static int gb_uart_receive_credits_handler(struct gb_operation *op)
drivers/staging/greybus/uart.c
151
struct gb_connection *connection = op->connection;
drivers/staging/greybus/uart.c
153
struct gb_message *request = op->request;
drivers/staging/greybus/uart.c
199
static int gb_uart_request_handler(struct gb_operation *op)
drivers/staging/greybus/uart.c
201
struct gb_connection *connection = op->connection;
drivers/staging/greybus/uart.c
203
int type = op->type;
drivers/staging/greybus/uart.c
208
ret = gb_uart_receive_data_handler(op);
drivers/staging/greybus/uart.c
211
ret = gb_uart_serial_state_handler(op);
drivers/staging/greybus/uart.c
214
ret = gb_uart_receive_credits_handler(op);
drivers/staging/greybus/uart.c
73
static int gb_uart_receive_data_handler(struct gb_operation *op)
drivers/staging/greybus/uart.c
75
struct gb_connection *connection = op->connection;
drivers/staging/greybus/uart.c
78
struct gb_message *request = op->request;
drivers/staging/greybus/uart.c
91
receive_data = op->request->payload;
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
288
int cla, op;
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
292
for (op = 0; op_class[op].op_class; op++) {
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
294
const struct p2p_oper_class_map *o = &op_class[op];
drivers/target/iscsi/cxgbit/cxgbit_main.c
197
cxgbit_lro_add_packet_rsp(struct sk_buff *skb, u8 op, const __be64 *rsp)
drivers/target/iscsi/cxgbit/cxgbit_main.c
242
cxgbit_lro_add_packet_gl(struct sk_buff *skb, u8 op, const struct pkt_gl *gl)
drivers/target/iscsi/cxgbit/cxgbit_main.c
249
if (op == CPL_ISCSI_HDR) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
264
} else if (op == CPL_ISCSI_DATA) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
316
cxgbit_lro_init_skb(struct cxgbit_sock *csk, u8 op, const struct pkt_gl *gl,
drivers/target/iscsi/cxgbit/cxgbit_main.c
379
cxgbit_lro_receive(struct cxgbit_sock *csk, u8 op, const __be64 *rsp,
drivers/target/iscsi/cxgbit/cxgbit_main.c
387
pr_err("%s: csk NULL, op 0x%x.\n", __func__, op);
drivers/target/iscsi/cxgbit/cxgbit_main.c
400
skb = cxgbit_lro_init_skb(csk, op, gl, rsp, napi);
drivers/target/iscsi/cxgbit/cxgbit_main.c
421
cxgbit_lro_add_packet_gl(skb, op, gl);
drivers/target/iscsi/cxgbit/cxgbit_main.c
423
cxgbit_lro_add_packet_rsp(skb, op, rsp);
drivers/target/iscsi/cxgbit/cxgbit_main.c
444
unsigned int op = *(u8 *)rsp;
drivers/target/iscsi/cxgbit/cxgbit_main.c
447
switch (op) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
477
if (op == CPL_RX_ISCSI_DDP) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
478
if (!cxgbit_lro_receive(csk, op, rsp, NULL, lro_mgr,
drivers/target/iscsi/cxgbit/cxgbit_main.c
490
if (unlikely(op != *(u8 *)gl->va)) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
498
if ((op == CPL_ISCSI_HDR) || (op == CPL_ISCSI_DATA) ||
drivers/target/iscsi/cxgbit/cxgbit_main.c
499
(op == CPL_RX_ISCSI_CMP)) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
500
if (!cxgbit_lro_receive(csk, op, rsp, gl, lro_mgr,
drivers/target/iscsi/cxgbit/cxgbit_main.c
512
op = rpl->ot.opcode;
drivers/target/iscsi/cxgbit/cxgbit_main.c
513
cxgbit_skcb_rx_opcode(skb) = op;
drivers/target/iscsi/cxgbit/cxgbit_main.c
516
cdev, op, rpl->ot.opcode_tid,
drivers/target/iscsi/cxgbit/cxgbit_main.c
519
if (op < NUM_CPL_CMDS && cxgbit_cplhandlers[op]) {
drivers/target/iscsi/cxgbit/cxgbit_main.c
520
cxgbit_cplhandlers[op](cdev, skb);
drivers/target/iscsi/cxgbit/cxgbit_main.c
522
pr_err("No handler for opcode 0x%x.\n", op);
drivers/target/iscsi/iscsi_target_configfs.c
1292
u32 op;
drivers/target/iscsi/iscsi_target_configfs.c
1295
err = kstrtou32(page, 0, &op);
drivers/target/iscsi/iscsi_target_configfs.c
1298
if ((op != 1) && (op != 0)) {
drivers/target/iscsi/iscsi_target_configfs.c
1300
" %u\n", op);
drivers/target/iscsi/iscsi_target_configfs.c
1314
if (op) {
drivers/target/iscsi/iscsi_target_configfs.c
63
u32 op;
drivers/target/iscsi/iscsi_target_configfs.c
66
rc = kstrtou32(page, 0, &op);
drivers/target/iscsi/iscsi_target_configfs.c
69
if ((op != 1) && (op != 0)) {
drivers/target/iscsi/iscsi_target_configfs.c
70
pr_err("Illegal value for tpg_enable: %u\n", op);
drivers/target/iscsi/iscsi_target_configfs.c
84
if (op) {
drivers/target/target_core_fabric_configfs.c
831
bool op;
drivers/target/target_core_fabric_configfs.c
833
ret = kstrtobool(page, &op);
drivers/target/target_core_fabric_configfs.c
837
if (se_tpg->enabled == op)
drivers/target/target_core_fabric_configfs.c
839
if (op)
drivers/tee/amdtee/amdtee_if.h
153
struct tee_operation op;
drivers/tee/amdtee/amdtee_if.h
181
struct tee_operation op;
drivers/tee/amdtee/call.c
256
ret = tee_params_to_amd_params(p, arg->num_params, &cmd.op);
drivers/tee/amdtee/call.c
272
ret = amd_params_to_tee_params(p, arg->num_params, &cmd.op);
drivers/tee/amdtee/call.c
367
ret = tee_params_to_amd_params(p, arg->num_params, &cmd.op);
drivers/tee/amdtee/call.c
382
ret = amd_params_to_tee_params(p, arg->num_params, &cmd.op);
drivers/tee/qcomtee/async.c
158
switch (async_msg->op) {
drivers/tee/qcomtee/async.c
164
pr_err("Unsupported async message %u\n", async_msg->op);
drivers/tee/qcomtee/async.c
171
async_msg->op);
drivers/tee/qcomtee/async.c
35
u32 op;
drivers/tee/qcomtee/call.c
360
static int qcomtee_root_object_check(u32 op, struct tee_param *params,
drivers/tee/qcomtee/call.c
364
if (op == QCOMTEE_ROOT_OP_NOTIFY_DOMAIN_CHANGE ||
drivers/tee/qcomtee/call.c
365
op == QCOMTEE_ROOT_OP_ADCI_ACCEPT ||
drivers/tee/qcomtee/call.c
366
op == QCOMTEE_ROOT_OP_ADCI_SHUTDOWN)
drivers/tee/qcomtee/call.c
375
if (op == QCOMTEE_ROOT_OP_REG_WITH_CREDENTIALS && num_params == 2) {
drivers/tee/qcomtee/call.c
406
if (arg->op == QCOMTEE_MSG_OBJECT_OP_RELEASE) {
drivers/tee/qcomtee/call.c
427
if (qcomtee_root_object_check(arg->op, params, arg->num_params))
drivers/tee/qcomtee/call.c
439
ret = qcomtee_object_do_invoke(oic, object, arg->op, u, &result);
drivers/tee/qcomtee/call.c
490
static int qcomtee_supp_recv(struct tee_context *ctx, u32 *op, u32 *num_params,
drivers/tee/qcomtee/call.c
526
*op = data.op;
drivers/tee/qcomtee/core.c
396
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
461
qcomtee_msg_init(msg, object_id, op, ib, ob, io, oo);
drivers/tee/qcomtee/core.c
622
u32 op;
drivers/tee/qcomtee/core.c
644
op = msg->op & QCOMTEE_MSG_OBJECT_OP_MASK;
drivers/tee/qcomtee/core.c
646
switch (op) {
drivers/tee/qcomtee/core.c
668
errno = object->ops->dispatch(oic, object, op, oic->u);
drivers/tee/qcomtee/core.c
742
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
758
ret = qcomtee_prepare_msg(oic, object, op, u);
drivers/tee/qcomtee/core.c
849
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/core.c
853
if (op & ~QCOMTEE_MSG_OBJECT_OP_MASK)
drivers/tee/qcomtee/core.c
862
if (op == QCOMTEE_MSG_OBJECT_OP_RELEASE ||
drivers/tee/qcomtee/core.c
863
op == QCOMTEE_MSG_OBJECT_OP_RETAIN)
drivers/tee/qcomtee/core.c
866
return qcomtee_object_do_invoke_internal(oic, object, op, u, result);
drivers/tee/qcomtee/mem_obj.c
57
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/primordial_obj.c
37
u32 op, struct qcomtee_arg *args)
drivers/tee/qcomtee/primordial_obj.c
44
switch (op) {
drivers/tee/qcomtee/qcomtee.h
132
u32 op;
drivers/tee/qcomtee/qcomtee.h
66
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/qcomtee_msg.h
135
u32 op;
drivers/tee/qcomtee/qcomtee_msg.h
159
u32 op;
drivers/tee/qcomtee/qcomtee_msg.h
236
u32 cxt, u32 op, int in_buffer,
drivers/tee/qcomtee/qcomtee_msg.h
248
msg->op = op;
drivers/tee/qcomtee/qcomtee_object.h
202
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/qcomtee_object.h
214
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/user_obj.c
211
if (ureq->op != QCOMTEE_MSG_OBJECT_OP_RELEASE) {
drivers/tee/qcomtee/user_obj.c
226
struct qcomtee_object *object, u32 op,
drivers/tee/qcomtee/user_obj.c
240
ureq->op = op;
drivers/tee/qcomtee/user_obj.c
335
ureq->op = QCOMTEE_MSG_OBJECT_OP_RELEASE;
drivers/tee/qcomtee/user_obj.c
615
data->op = ureq->op;
drivers/tee/qcomtee/user_obj.c
640
if (data->op == QCOMTEE_MSG_OBJECT_OP_RELEASE) {
drivers/tee/qcomtee/user_obj.c
86
u32 op;
drivers/thermal/qcom/tsens.c
1075
put_device(&op->dev);
drivers/thermal/qcom/tsens.c
897
struct platform_device *op = of_find_device_by_node(priv->dev->of_node);
drivers/thermal/qcom/tsens.c
899
if (!op)
drivers/thermal/qcom/tsens.c
902
if (op->num_resources > 1) {
drivers/thermal/qcom/tsens.c
905
res = platform_get_resource(op, IORESOURCE_MEM, 1);
drivers/thermal/qcom/tsens.c
924
res = platform_get_resource(op, IORESOURCE_MEM, 0);
drivers/thermal/thermal_netlink.c
779
int cmd = info->op.cmd;
drivers/tty/n_tty.c
564
u8 op;
drivers/tty/n_tty.c
577
op = echo_buf(ldata, *tail + 1);
drivers/tty/n_tty.c
579
switch (op) {
drivers/tty/n_tty.c
644
tty_put_char(tty, op ^ 0100);
drivers/tty/serial/apbuart.c
521
static int apbuart_probe(struct platform_device *op)
drivers/tty/serial/apbuart.c
527
if (op->dev.of_node == grlib_apbuart_nodes[i])
drivers/tty/serial/apbuart.c
532
port->dev = &op->dev;
drivers/tty/serial/apbuart.c
533
port->irq = op->archdata.irqs[0];
drivers/tty/serial/mpc52xx_uart.c
1698
static int mpc52xx_uart_of_probe(struct platform_device *op)
drivers/tty/serial/mpc52xx_uart.c
1708
if (mpc52xx_uart_nodes[idx] == op->dev.of_node)
drivers/tty/serial/mpc52xx_uart.c
1718
uartclk = mpc5xxx_get_bus_frequency(&op->dev);
drivers/tty/serial/mpc52xx_uart.c
1720
dev_dbg(&op->dev, "Could not find uart clock frequency!\n");
drivers/tty/serial/mpc52xx_uart.c
1736
port->dev = &op->dev;
drivers/tty/serial/mpc52xx_uart.c
1739
ret = of_address_to_resource(op->dev.of_node, 0, &res);
drivers/tty/serial/mpc52xx_uart.c
1745
dev_dbg(&op->dev, "Could not allocate resources for PSC\n");
drivers/tty/serial/mpc52xx_uart.c
1749
psc_ops->get_irq(port, op->dev.of_node);
drivers/tty/serial/mpc52xx_uart.c
1751
dev_dbg(&op->dev, "Could not get irq\n");
drivers/tty/serial/mpc52xx_uart.c
1755
dev_dbg(&op->dev, "mpc52xx-psc uart at %p, irq=%x, freq=%i\n",
drivers/tty/serial/mpc52xx_uart.c
1763
platform_set_drvdata(op, (void *)port);
drivers/tty/serial/mpc52xx_uart.c
1767
static void mpc52xx_uart_of_remove(struct platform_device *op)
drivers/tty/serial/mpc52xx_uart.c
1769
struct uart_port *port = platform_get_drvdata(op);
drivers/tty/serial/mpc52xx_uart.c
1777
mpc52xx_uart_of_suspend(struct platform_device *op, pm_message_t state)
drivers/tty/serial/mpc52xx_uart.c
1779
struct uart_port *port = platform_get_drvdata(op);
drivers/tty/serial/mpc52xx_uart.c
1788
mpc52xx_uart_of_resume(struct platform_device *op)
drivers/tty/serial/mpc52xx_uart.c
1790
struct uart_port *port = platform_get_drvdata(op);
drivers/tty/serial/sunhv.c
523
static int hv_probe(struct platform_device *op)
drivers/tty/serial/sunhv.c
529
if (op->archdata.irqs[0] == 0xffffffff)
drivers/tty/serial/sunhv.c
562
port->irq = op->archdata.irqs[0];
drivers/tty/serial/sunhv.c
564
port->dev = &op->dev;
drivers/tty/serial/sunhv.c
570
sunserial_console_match(&sunhv_console, op->dev.of_node,
drivers/tty/serial/sunhv.c
581
platform_set_drvdata(op, port);
drivers/tty/serial/sunsab.c
1015
static int sab_probe(struct platform_device *op)
drivers/tty/serial/sunsab.c
1023
err = sunsab_init_one(&up[0], op,
drivers/tty/serial/sunsab.c
1029
err = sunsab_init_one(&up[1], op,
drivers/tty/serial/sunsab.c
1035
sunserial_console_match(SUNSAB_CONSOLE(), op->dev.of_node,
drivers/tty/serial/sunsab.c
1039
sunserial_console_match(SUNSAB_CONSOLE(), op->dev.of_node,
drivers/tty/serial/sunsab.c
1051
platform_set_drvdata(op, &up[0]);
drivers/tty/serial/sunsab.c
1060
of_iounmap(&op->resource[0],
drivers/tty/serial/sunsab.c
1064
of_iounmap(&op->resource[0],
drivers/tty/serial/sunsab.c
1071
static void sab_remove(struct platform_device *op)
drivers/tty/serial/sunsab.c
1073
struct uart_sunsab_port *up = platform_get_drvdata(op);
drivers/tty/serial/sunsab.c
1077
of_iounmap(&op->resource[0],
drivers/tty/serial/sunsab.c
1080
of_iounmap(&op->resource[0],
drivers/tty/serial/sunsab.c
962
struct platform_device *op,
drivers/tty/serial/sunsab.c
967
up->port.dev = &op->dev;
drivers/tty/serial/sunsab.c
969
up->port.mapbase = op->resource[0].start + offset;
drivers/tty/serial/sunsab.c
970
up->port.membase = of_ioremap(&op->resource[0], offset,
drivers/tty/serial/sunsab.c
977
up->port.irq = op->archdata.irqs[0];
drivers/tty/serial/sunsu.c
1386
static int su_probe(struct platform_device *op)
drivers/tty/serial/sunsu.c
1388
struct device_node *dp = op->dev.of_node;
drivers/tty/serial/sunsu.c
1412
rp = &op->resource[0];
drivers/tty/serial/sunsu.c
1422
up->port.irq = op->archdata.irqs[0];
drivers/tty/serial/sunsu.c
1424
up->port.dev = &op->dev;
drivers/tty/serial/sunsu.c
1434
of_iounmap(&op->resource[0],
drivers/tty/serial/sunsu.c
1439
platform_set_drvdata(op, up);
drivers/tty/serial/sunsu.c
1468
platform_set_drvdata(op, up);
drivers/tty/serial/sunsu.c
1475
of_iounmap(&op->resource[0], up->port.membase, up->reg_size);
drivers/tty/serial/sunsu.c
1480
static void su_remove(struct platform_device *op)
drivers/tty/serial/sunsu.c
1482
struct uart_sunsu_port *up = platform_get_drvdata(op);
drivers/tty/serial/sunsu.c
1497
of_iounmap(&op->resource[0], up->port.membase, up->reg_size);
drivers/tty/serial/sunzilog.c
1396
static int zs_probe(struct platform_device *op)
drivers/tty/serial/sunzilog.c
1405
if (of_property_present(op->dev.of_node, "keyboard"))
drivers/tty/serial/sunzilog.c
1414
sunzilog_chip_regs[inst] = of_ioremap(&op->resource[0], 0,
drivers/tty/serial/sunzilog.c
1423
zilog_irq = op->archdata.irqs[0];
drivers/tty/serial/sunzilog.c
1428
up[0].port.mapbase = op->resource[0].start + 0x00;
drivers/tty/serial/sunzilog.c
1431
up[0].port.irq = op->archdata.irqs[0];
drivers/tty/serial/sunzilog.c
1438
up[0].port.dev = &op->dev;
drivers/tty/serial/sunzilog.c
1446
up[1].port.mapbase = op->resource[0].start + 0x04;
drivers/tty/serial/sunzilog.c
1449
up[1].port.irq = op->archdata.irqs[0];
drivers/tty/serial/sunzilog.c
1456
up[1].port.dev = &op->dev;
drivers/tty/serial/sunzilog.c
1464
if (sunserial_console_match(SUNZILOG_CONSOLE(), op->dev.of_node,
drivers/tty/serial/sunzilog.c
1470
of_iounmap(&op->resource[0],
drivers/tty/serial/sunzilog.c
1474
if (sunserial_console_match(SUNZILOG_CONSOLE(), op->dev.of_node,
drivers/tty/serial/sunzilog.c
1481
of_iounmap(&op->resource[0],
drivers/tty/serial/sunzilog.c
1489
dev_name(&op->dev),
drivers/tty/serial/sunzilog.c
1491
op->archdata.irqs[0], sunzilog_type(&up[0].port));
drivers/tty/serial/sunzilog.c
1494
dev_name(&op->dev),
drivers/tty/serial/sunzilog.c
1496
op->archdata.irqs[0], sunzilog_type(&up[1].port));
drivers/tty/serial/sunzilog.c
1500
platform_set_drvdata(op, &up[0]);
drivers/tty/serial/sunzilog.c
1515
static void zs_remove(struct platform_device *op)
drivers/tty/serial/sunzilog.c
1517
struct uart_sunzilog_port *up = platform_get_drvdata(op);
drivers/tty/serial/sunzilog.c
1524
of_iounmap(&op->resource[0], regs, sizeof(struct zilog_layout));
drivers/tty/vt/vt.c
4836
static int con_font_get(struct vc_data *vc, struct console_font_op *op)
drivers/tty/vt/vt.c
4840
unsigned int vpitch = op->op == KD_FONT_OP_GET_TALL ? op->height : 32;
drivers/tty/vt/vt.c
4846
if (op->data) {
drivers/tty/vt/vt.c
4866
if (op->data && font.charcount > op->charcount)
drivers/tty/vt/vt.c
4868
if (font.width > op->width || font.height > op->height)
drivers/tty/vt/vt.c
4871
op->height = font.height;
drivers/tty/vt/vt.c
4872
op->width = font.width;
drivers/tty/vt/vt.c
4873
op->charcount = font.charcount;
drivers/tty/vt/vt.c
4875
if (op->data && copy_to_user(op->data, font.data, c))
drivers/tty/vt/vt.c
4881
static int con_font_set(struct vc_data *vc, const struct console_font_op *op)
drivers/tty/vt/vt.c
4885
unsigned int vpitch = op->op == KD_FONT_OP_SET_TALL ? op->height : 32;
drivers/tty/vt/vt.c
4887
if (!op->data)
drivers/tty/vt/vt.c
4889
if (op->charcount > max_font_glyphs)
drivers/tty/vt/vt.c
4891
if (op->width <= 0 || op->width > max_font_width || !op->height ||
drivers/tty/vt/vt.c
4892
op->height > max_font_height)
drivers/tty/vt/vt.c
4894
if (vpitch < op->height)
drivers/tty/vt/vt.c
4896
size = DIV_ROUND_UP(op->width, 8) * vpitch * op->charcount;
drivers/tty/vt/vt.c
4900
void *font_data __free(kfree) = font.data = memdup_user(op->data, size);
drivers/tty/vt/vt.c
4904
font.charcount = op->charcount;
drivers/tty/vt/vt.c
4905
font.width = op->width;
drivers/tty/vt/vt.c
4906
font.height = op->height;
drivers/tty/vt/vt.c
4918
return vc->vc_sw->con_font_set(vc, &font, vpitch, op->flags);
drivers/tty/vt/vt.c
4921
static int con_font_default(struct vc_data *vc, struct console_font_op *op)
drivers/tty/vt/vt.c
4923
struct console_font font = {.width = op->width, .height = op->height};
drivers/tty/vt/vt.c
4927
if (!op->data)
drivers/tty/vt/vt.c
4929
else if (strncpy_from_user(name, op->data, MAX_FONT_NAME - 1) < 0)
drivers/tty/vt/vt.c
4947
op->width = font.width;
drivers/tty/vt/vt.c
4948
op->height = font.height;
drivers/tty/vt/vt.c
4953
int con_font_op(struct vc_data *vc, struct console_font_op *op)
drivers/tty/vt/vt.c
4955
switch (op->op) {
drivers/tty/vt/vt.c
4958
return con_font_set(vc, op);
drivers/tty/vt/vt.c
4961
return con_font_get(vc, op);
drivers/tty/vt/vt.c
4963
return con_font_default(vc, op);
drivers/tty/vt/vt_ioctl.c
1001
compat_uint_t op; /* operation code KD_FONT_OP_* */
drivers/tty/vt/vt_ioctl.c
1010
int perm, struct console_font_op *op, struct vc_data *vc)
drivers/tty/vt/vt_ioctl.c
1014
if (copy_from_user(op, fontop, sizeof(struct compat_console_font_op)))
drivers/tty/vt/vt_ioctl.c
1016
if (!perm && op->op != KD_FONT_OP_GET)
drivers/tty/vt/vt_ioctl.c
1018
op->data = compat_ptr(((struct compat_console_font_op *)op)->data);
drivers/tty/vt/vt_ioctl.c
1019
i = con_font_op(vc, op);
drivers/tty/vt/vt_ioctl.c
1022
((struct compat_console_font_op *)op)->data = (unsigned long)op->data;
drivers/tty/vt/vt_ioctl.c
1023
if (copy_to_user(fontop, op, sizeof(struct compat_console_font_op)))
drivers/tty/vt/vt_ioctl.c
1060
struct console_font_op op; /* used in multiple places here */
drivers/tty/vt/vt_ioctl.c
1078
return compat_kdfontop_ioctl(up, perm, &op, vc);
drivers/tty/vt/vt_ioctl.c
466
struct console_font_op op;
drivers/tty/vt/vt_ioctl.c
468
if (copy_from_user(&op, up, sizeof(op)))
drivers/tty/vt/vt_ioctl.c
470
if (!perm && op.op != KD_FONT_OP_GET)
drivers/tty/vt/vt_ioctl.c
472
ret = con_font_op(vc, &op);
drivers/tty/vt/vt_ioctl.c
475
if (copy_to_user(up, &op, sizeof(op)))
drivers/ufs/core/ufshcd-priv.h
210
static inline int ufshcd_vops_suspend(struct ufs_hba *hba, enum ufs_pm_op op,
drivers/ufs/core/ufshcd-priv.h
214
return hba->vops->suspend(hba, op, status);
drivers/ufs/core/ufshcd-priv.h
219
static inline int ufshcd_vops_resume(struct ufs_hba *hba, enum ufs_pm_op op)
drivers/ufs/core/ufshcd-priv.h
222
return hba->vops->resume(hba, op);
drivers/ufs/host/ufs-mediatek-sip.h
88
#define ufs_mtk_mphy_ctrl(op, res) \
drivers/ufs/host/ufs-mediatek-sip.h
89
ufs_mtk_smc(UFS_MTK_SIP_MPHY_CTRL, &(res), op)
drivers/ufs/host/ufs-mediatek-sip.h
91
#define ufs_mtk_mtcmos_ctrl(op, res) \
drivers/ufs/host/ufs-mediatek-sip.h
92
ufs_mtk_smc(UFS_MTK_SIP_MTCMOS_CTRL, &(res), op)
drivers/ufs/host/ufshcd-pci.c
382
static int ufs_intel_resume(struct ufs_hba *hba, enum ufs_pm_op op)
drivers/usb/chipidea/ci.h
160
void __iomem *op;
drivers/usb/chipidea/core.c
107
(i <= CAP_LAST ? ci->hw_bank.cap : ci->hw_bank.op) +
drivers/usb/chipidea/core.c
111
ci->hw_bank.regmap[i] = ci->hw_bank.op +
drivers/usb/chipidea/core.c
228
ci->hw_bank.op = ci->hw_bank.cap + (ioread32(ci->hw_bank.cap) & 0xff);
drivers/usb/chipidea/core.c
236
ci->hw_bank.size = ci->hw_bank.op - ci->hw_bank.abs;
drivers/usb/chipidea/core.c
259
ci->rev, ci->hw_bank.lpm, ci->hw_bank.cap, ci->hw_bank.op);
drivers/usb/host/ehci-grlib.c
103
dev_err(&op->dev, "%s: irq_of_parse_and_map failed\n",
drivers/usb/host/ehci-grlib.c
109
hcd->regs = devm_ioremap_resource(&op->dev, &res);
drivers/usb/host/ehci-grlib.c
143
static void ehci_hcd_grlib_remove(struct platform_device *op)
drivers/usb/host/ehci-grlib.c
145
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/ehci-grlib.c
147
dev_dbg(&op->dev, "stopping GRLIB GRUSBHC EHCI USB Controller\n");
drivers/usb/host/ehci-grlib.c
72
static int ehci_hcd_grlib_probe(struct platform_device *op)
drivers/usb/host/ehci-grlib.c
74
struct device_node *dn = op->dev.of_node;
drivers/usb/host/ehci-grlib.c
85
dev_dbg(&op->dev, "initializing GRUSBHC EHCI USB Controller\n");
drivers/usb/host/ehci-grlib.c
92
op->dev.dma_mask = &op->dev.coherent_dma_mask;
drivers/usb/host/ehci-grlib.c
93
hcd = usb_create_hcd(&ehci_grlib_hc_driver, &op->dev,
drivers/usb/host/ehci-ppc-of.c
108
dev_dbg(&op->dev, "initializing PPC-OF USB Controller\n");
drivers/usb/host/ehci-ppc-of.c
114
hcd = usb_create_hcd(&ehci_ppc_of_hc_driver, &op->dev, "PPC-OF USB");
drivers/usb/host/ehci-ppc-of.c
123
dev_err(&op->dev, "%s: irq_of_parse_and_map failed\n",
drivers/usb/host/ehci-ppc-of.c
129
hcd->regs = devm_ioremap_resource(&op->dev, &res);
drivers/usb/host/ehci-ppc-of.c
141
devm_ioremap(&op->dev,
drivers/usb/host/ehci-ppc-of.c
187
static void ehci_hcd_ppc_of_remove(struct platform_device *op)
drivers/usb/host/ehci-ppc-of.c
189
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/ehci-ppc-of.c
195
dev_dbg(&op->dev, "stopping PPC-OF USB Controller\n");
drivers/usb/host/ehci-ppc-of.c
94
static int ehci_hcd_ppc_of_probe(struct platform_device *op)
drivers/usb/host/ehci-ppc-of.c
96
struct device_node *dn = op->dev.of_node;
drivers/usb/host/ehci-xilinx-of.c
118
static int ehci_hcd_xilinx_of_probe(struct platform_device *op)
drivers/usb/host/ehci-xilinx-of.c
120
struct device_node *dn = op->dev.of_node;
drivers/usb/host/ehci-xilinx-of.c
131
dev_dbg(&op->dev, "initializing XILINX-OF USB Controller\n");
drivers/usb/host/ehci-xilinx-of.c
137
hcd = usb_create_hcd(&ehci_xilinx_of_hc_driver, &op->dev,
drivers/usb/host/ehci-xilinx-of.c
147
dev_err(&op->dev, "%s: irq_of_parse_and_map failed\n",
drivers/usb/host/ehci-xilinx-of.c
153
hcd->regs = devm_ioremap_resource(&op->dev, &res);
drivers/usb/host/ehci-xilinx-of.c
204
static void ehci_hcd_xilinx_of_remove(struct platform_device *op)
drivers/usb/host/ehci-xilinx-of.c
206
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/ehci-xilinx-of.c
208
dev_dbg(&op->dev, "stopping XILINX-OF USB Controller\n");
drivers/usb/host/ohci-ppc-of.c
104
dev_dbg(&op->dev, "initializing PPC-OF USB Controller\n");
drivers/usb/host/ohci-ppc-of.c
110
hcd = usb_create_hcd(&ohci_ppc_of_hc_driver, &op->dev, "PPC-OF USB");
drivers/usb/host/ohci-ppc-of.c
117
hcd->regs = devm_ioremap_resource(&op->dev, &res);
drivers/usb/host/ohci-ppc-of.c
125
dev_err(&op->dev, "%s: irq_of_parse_and_map failed\n",
drivers/usb/host/ohci-ppc-of.c
180
static void ohci_hcd_ppc_of_remove(struct platform_device *op)
drivers/usb/host/ohci-ppc-of.c
182
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/ohci-ppc-of.c
184
dev_dbg(&op->dev, "stopping PPC-OF USB Controller\n");
drivers/usb/host/ohci-ppc-of.c
85
static int ohci_hcd_ppc_of_probe(struct platform_device *op)
drivers/usb/host/ohci-ppc-of.c
87
struct device_node *dn = op->dev.of_node;
drivers/usb/host/uhci-grlib.c
103
dev_dbg(&op->dev, "initializing GRUSBHC UHCI USB Controller\n");
drivers/usb/host/uhci-grlib.c
110
op->dev.dma_mask = &op->dev.coherent_dma_mask;
drivers/usb/host/uhci-grlib.c
111
hcd = usb_create_hcd(&uhci_grlib_hc_driver, &op->dev,
drivers/usb/host/uhci-grlib.c
126
hcd->regs = devm_ioremap_resource(&op->dev, &res);
drivers/usb/host/uhci-grlib.c
151
static void uhci_hcd_grlib_remove(struct platform_device *op)
drivers/usb/host/uhci-grlib.c
153
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/uhci-grlib.c
155
dev_dbg(&op->dev, "stopping GRLIB GRUSBHC UHCI USB Controller\n");
drivers/usb/host/uhci-grlib.c
170
static void uhci_hcd_grlib_shutdown(struct platform_device *op)
drivers/usb/host/uhci-grlib.c
172
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/host/uhci-grlib.c
91
static int uhci_hcd_grlib_probe(struct platform_device *op)
drivers/usb/host/uhci-grlib.c
93
struct device_node *dn = op->dev.of_node;
drivers/usb/host/uhci-platform.c
190
static void uhci_hcd_platform_shutdown(struct platform_device *op)
drivers/usb/host/uhci-platform.c
192
struct usb_hcd *hcd = platform_get_drvdata(op);
drivers/usb/typec/tcpm/tcpm.c
3929
unsigned int max, op, pdo_max, index;
drivers/usb/typec/tcpm/tcpm.c
3942
op = rdo_op_current(rdo);
drivers/usb/typec/tcpm/tcpm.c
3945
if (op > pdo_max)
drivers/usb/typec/tcpm/tcpm.c
3953
pdo_fixed_voltage(pdo), pdo_max, op, max);
drivers/usb/typec/tcpm/tcpm.c
3958
pdo_max, op, max);
drivers/usb/typec/tcpm/tcpm.c
3962
op = rdo_op_power(rdo);
drivers/usb/typec/tcpm/tcpm.c
3965
if (op > pdo_max)
drivers/usb/typec/tcpm/tcpm.c
3972
pdo_max, op, max);
drivers/vdpa/solidrun/snet_ctrl.c
107
iowrite32(val, &ctrl_regs->op);
drivers/vdpa/solidrun/snet_ctrl.c
56
struct snet_ctrl_reg_op op;
drivers/vdpa/solidrun/snet_ctrl.c
79
return readx_poll_timeout(ioread32, &regs->op, val, !val, 10, SNET_CTRL_TIMEOUT);
drivers/vfio/cdx/main.c
71
offsetofend(struct vfio_device_feature_bus_master, op);
drivers/vfio/cdx/main.c
89
switch (ops.op) {
drivers/vfio/vfio_iommu_spapr_tce.c
781
struct vfio_eeh_pe_op op;
drivers/vfio/vfio_iommu_spapr_tce.c
788
minsz = offsetofend(struct vfio_eeh_pe_op, op);
drivers/vfio/vfio_iommu_spapr_tce.c
789
if (copy_from_user(&op, (void __user *)arg, minsz))
drivers/vfio/vfio_iommu_spapr_tce.c
791
if (op.argsz < minsz || op.flags)
drivers/vfio/vfio_iommu_spapr_tce.c
794
switch (op.op) {
drivers/vfio/vfio_iommu_spapr_tce.c
816
if (op.argsz < minsz)
drivers/vfio/vfio_iommu_spapr_tce.c
818
if (copy_from_user(&op, (void __user *)arg, minsz))
drivers/vfio/vfio_iommu_spapr_tce.c
821
return eeh_pe_inject_err(pe, op.err.type, op.err.func,
drivers/vfio/vfio_iommu_spapr_tce.c
822
op.err.addr, op.err.mask);
drivers/video/console/newport_con.c
499
static int newport_set_font(int unit, const struct console_font *op,
drivers/video/console/newport_con.c
50
static int newport_set_def_font(int unit, struct console_font *op);
drivers/video/console/newport_con.c
502
int w = op->width;
drivers/video/console/newport_con.c
503
int h = op->height;
drivers/video/console/newport_con.c
504
int size = h * op->charcount;
drivers/video/console/newport_con.c
506
unsigned char *new_data, *data = op->data, *p;
drivers/video/console/newport_con.c
511
|| (op->charcount != 256 && op->charcount != 512))
drivers/video/console/newport_con.c
519
FNTCHARCNT(new_data) = op->charcount;
drivers/video/console/newport_con.c
524
for (i = 0; i < op->charcount; i++) {
drivers/video/console/newport_con.c
555
static int newport_set_def_font(int unit, struct console_font *op)
drivers/video/console/newport_con.c
567
static int newport_font_default(struct vc_data *vc, struct console_font *op,
drivers/video/console/newport_con.c
570
return newport_set_def_font(vc->vc_num, op);
drivers/video/console/sticon.c
156
static int sticon_set_font(struct vc_data *vc, const struct console_font *op,
drivers/video/console/sticon.c
162
int w = op->width;
drivers/video/console/sticon.c
163
int h = op->height;
drivers/video/console/sticon.c
167
unsigned char *data = op->data, *p;
drivers/video/console/sticon.c
170
|| (op->charcount != 256 && op->charcount != 512))
drivers/video/console/sticon.c
174
size = bpc * op->charcount;
drivers/video/console/sticon.c
181
new_font->last_char = op->charcount - 1;
drivers/video/console/sticon.c
202
for (i = 0; i < op->charcount; i++) {
drivers/video/console/sticon.c
249
static int sticon_font_default(struct vc_data *vc, struct console_font *op,
drivers/video/fbdev/broadsheetfb.c
438
u8 op, u32 addr)
drivers/video/fbdev/broadsheetfb.c
446
err = broadsheet_spiflash_write_byte(par, op);
drivers/video/fbdev/bw2.c
270
static int bw2_probe(struct platform_device *op)
drivers/video/fbdev/bw2.c
272
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/bw2.c
277
info = framebuffer_alloc(sizeof(struct bw2_par), &op->dev);
drivers/video/fbdev/bw2.c
286
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/bw2.c
287
par->which_io = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/bw2.c
298
par->regs = of_ioremap(&op->resource[0], BWTWO_REGISTER_OFFSET,
drivers/video/fbdev/bw2.c
313
info->screen_base = of_ioremap(&op->resource[0], 0,
drivers/video/fbdev/bw2.c
328
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/bw2.c
336
of_iounmap(&op->resource[0], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/bw2.c
339
of_iounmap(&op->resource[0], par->regs, sizeof(struct bw2_regs));
drivers/video/fbdev/bw2.c
348
static void bw2_remove(struct platform_device *op)
drivers/video/fbdev/bw2.c
350
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/bw2.c
355
of_iounmap(&op->resource[0], par->regs, sizeof(struct bw2_regs));
drivers/video/fbdev/bw2.c
356
of_iounmap(&op->resource[0], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/cg14.c
442
static void cg14_unmap_regs(struct platform_device *op, struct fb_info *info,
drivers/video/fbdev/cg14.c
446
of_iounmap(&op->resource[0],
drivers/video/fbdev/cg14.c
449
of_iounmap(&op->resource[0],
drivers/video/fbdev/cg14.c
452
of_iounmap(&op->resource[0],
drivers/video/fbdev/cg14.c
455
of_iounmap(&op->resource[1],
drivers/video/fbdev/cg14.c
459
static int cg14_probe(struct platform_device *op)
drivers/video/fbdev/cg14.c
461
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/cg14.c
466
info = framebuffer_alloc(sizeof(struct cg14_par), &op->dev);
drivers/video/fbdev/cg14.c
486
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/cg14.c
487
par->iospace = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/cg14.c
489
info->fix.smem_start = op->resource[1].start;
drivers/video/fbdev/cg14.c
490
par->iospace = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/cg14.c
493
par->regs = of_ioremap(&op->resource[0], 0,
drivers/video/fbdev/cg14.c
495
par->clut = of_ioremap(&op->resource[0], CG14_CLUT1,
drivers/video/fbdev/cg14.c
497
par->cursor = of_ioremap(&op->resource[0], CG14_CURSORREGS,
drivers/video/fbdev/cg14.c
500
info->screen_base = of_ioremap(&op->resource[1], 0,
drivers/video/fbdev/cg14.c
506
is_8mb = (resource_size(&op->resource[1]) == (8 * 1024 * 1024));
drivers/video/fbdev/cg14.c
519
(op->resource[0].start -
drivers/video/fbdev/cg14.c
520
op->resource[1].start);
drivers/video/fbdev/cg14.c
546
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/cg14.c
559
cg14_unmap_regs(op, info, par);
drivers/video/fbdev/cg14.c
566
static void cg14_remove(struct platform_device *op)
drivers/video/fbdev/cg14.c
568
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/cg14.c
574
cg14_unmap_regs(op, info, par);
drivers/video/fbdev/cg3.c
344
static int cg3_probe(struct platform_device *op)
drivers/video/fbdev/cg3.c
346
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/cg3.c
351
info = framebuffer_alloc(sizeof(struct cg3_par), &op->dev);
drivers/video/fbdev/cg3.c
360
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/cg3.c
361
par->which_io = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/cg3.c
376
par->regs = of_ioremap(&op->resource[0], CG3_REGS_OFFSET,
drivers/video/fbdev/cg3.c
382
info->screen_base = of_ioremap(&op->resource[0], CG3_RAM_OFFSET,
drivers/video/fbdev/cg3.c
407
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/cg3.c
418
of_iounmap(&op->resource[0], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/cg3.c
421
of_iounmap(&op->resource[0], par->regs, sizeof(struct cg3_regs));
drivers/video/fbdev/cg3.c
430
static void cg3_remove(struct platform_device *op)
drivers/video/fbdev/cg3.c
432
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/cg3.c
438
of_iounmap(&op->resource[0], par->regs, sizeof(struct cg3_regs));
drivers/video/fbdev/cg3.c
439
of_iounmap(&op->resource[0], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/cg6.c
722
static void cg6_unmap_regs(struct platform_device *op, struct fb_info *info,
drivers/video/fbdev/cg6.c
726
of_iounmap(&op->resource[0], par->fbc, 4096);
drivers/video/fbdev/cg6.c
728
of_iounmap(&op->resource[0], par->tec, sizeof(struct cg6_tec));
drivers/video/fbdev/cg6.c
730
of_iounmap(&op->resource[0], par->thc, sizeof(struct cg6_thc));
drivers/video/fbdev/cg6.c
732
of_iounmap(&op->resource[0], par->bt, sizeof(struct bt_regs));
drivers/video/fbdev/cg6.c
734
of_iounmap(&op->resource[0], par->fhc, sizeof(u32));
drivers/video/fbdev/cg6.c
737
of_iounmap(&op->resource[0], info->screen_base,
drivers/video/fbdev/cg6.c
741
static int cg6_probe(struct platform_device *op)
drivers/video/fbdev/cg6.c
743
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/cg6.c
749
info = framebuffer_alloc(sizeof(struct cg6_par), &op->dev);
drivers/video/fbdev/cg6.c
758
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/cg6.c
759
par->which_io = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/cg6.c
774
par->fbc = of_ioremap(&op->resource[0], CG6_FBC_OFFSET,
drivers/video/fbdev/cg6.c
776
par->tec = of_ioremap(&op->resource[0], CG6_TEC_OFFSET,
drivers/video/fbdev/cg6.c
778
par->thc = of_ioremap(&op->resource[0], CG6_THC_OFFSET,
drivers/video/fbdev/cg6.c
780
par->bt = of_ioremap(&op->resource[0], CG6_BROOKTREE_OFFSET,
drivers/video/fbdev/cg6.c
782
par->fhc = of_ioremap(&op->resource[0], CG6_FHC_OFFSET,
drivers/video/fbdev/cg6.c
790
info->screen_base = of_ioremap(&op->resource[0], CG6_RAM_OFFSET,
drivers/video/fbdev/cg6.c
812
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/cg6.c
824
cg6_unmap_regs(op, info, par);
drivers/video/fbdev/cg6.c
831
static void cg6_remove(struct platform_device *op)
drivers/video/fbdev/cg6.c
833
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/cg6.c
839
cg6_unmap_regs(op, info, par);
drivers/video/fbdev/cirrusfb.c
1815
unsigned char op = (info->var.bits_per_pixel == 24) ? 0xc : 0x4;
drivers/video/fbdev/cirrusfb.c
1823
op == 0xc)
drivers/video/fbdev/cirrusfb.c
1852
info->fix.line_length, op);
drivers/video/fbdev/cirrusfb.c
2708
u_char op;
drivers/video/fbdev/cirrusfb.c
2717
op = 0x80;
drivers/video/fbdev/cirrusfb.c
2721
op = 0x90;
drivers/video/fbdev/cirrusfb.c
2726
op = 0xa0;
drivers/video/fbdev/cirrusfb.c
2731
op = 0xb0;
drivers/video/fbdev/cirrusfb.c
2734
0, ndest, op | blitmode, line_length);
drivers/video/fbdev/ffb.c
1010
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/ffb.c
1025
of_iounmap(&op->resource[1], par->dac, sizeof(struct ffb_dac));
drivers/video/fbdev/ffb.c
1028
of_iounmap(&op->resource[2], par->fbc, sizeof(struct ffb_fbc));
drivers/video/fbdev/ffb.c
1037
static void ffb_remove(struct platform_device *op)
drivers/video/fbdev/ffb.c
1039
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/ffb.c
1045
of_iounmap(&op->resource[2], par->fbc, sizeof(struct ffb_fbc));
drivers/video/fbdev/ffb.c
1046
of_iounmap(&op->resource[1], par->dac, sizeof(struct ffb_dac));
drivers/video/fbdev/ffb.c
909
static int ffb_probe(struct platform_device *op)
drivers/video/fbdev/ffb.c
911
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/ffb.c
919
info = framebuffer_alloc(sizeof(struct ffb_par), &op->dev);
drivers/video/fbdev/ffb.c
928
par->fbc = of_ioremap(&op->resource[2], 0,
drivers/video/fbdev/ffb.c
933
par->dac = of_ioremap(&op->resource[1], 0,
drivers/video/fbdev/ffb.c
939
par->physbase = op->resource[0].start;
drivers/video/fbdev/leo.c
527
static void leo_unmap_regs(struct platform_device *op, struct fb_info *info,
drivers/video/fbdev/leo.c
531
of_iounmap(&op->resource[0], par->lc_ss0_usr, 0x1000);
drivers/video/fbdev/leo.c
533
of_iounmap(&op->resource[0], par->ld_ss0, 0x1000);
drivers/video/fbdev/leo.c
535
of_iounmap(&op->resource[0], par->ld_ss1, 0x1000);
drivers/video/fbdev/leo.c
537
of_iounmap(&op->resource[0], par->lx_krn, 0x1000);
drivers/video/fbdev/leo.c
539
of_iounmap(&op->resource[0],
drivers/video/fbdev/leo.c
542
of_iounmap(&op->resource[0], info->screen_base, 0x800000);
drivers/video/fbdev/leo.c
545
static int leo_probe(struct platform_device *op)
drivers/video/fbdev/leo.c
547
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/leo.c
552
info = framebuffer_alloc(sizeof(struct leo_par), &op->dev);
drivers/video/fbdev/leo.c
561
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/leo.c
562
par->which_io = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/leo.c
572
of_ioremap(&op->resource[0], LEO_OFF_LC_SS0_USR,
drivers/video/fbdev/leo.c
575
of_ioremap(&op->resource[0], LEO_OFF_LD_SS0,
drivers/video/fbdev/leo.c
578
of_ioremap(&op->resource[0], LEO_OFF_LD_SS1,
drivers/video/fbdev/leo.c
581
of_ioremap(&op->resource[0], LEO_OFF_LX_KRN,
drivers/video/fbdev/leo.c
584
of_ioremap(&op->resource[0], LEO_OFF_LX_CURSOR,
drivers/video/fbdev/leo.c
587
of_ioremap(&op->resource[0], LEO_OFF_SS0,
drivers/video/fbdev/leo.c
614
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/leo.c
626
leo_unmap_regs(op, info, par);
drivers/video/fbdev/leo.c
633
static void leo_remove(struct platform_device *op)
drivers/video/fbdev/leo.c
635
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/leo.c
641
leo_unmap_regs(op, info, par);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1000
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1002
if (op->enabled == enable)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1005
op->enabled = enable;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1006
op->extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1012
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1014
if (op->fifo_low == fifo_low && op->fifo_high == fifo_high)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1017
op->fifo_low = fifo_low;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1018
op->fifo_high = fifo_high;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1019
op->extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1024
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1028
if (!op->enabled && !op->enabling)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1302
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1312
op->user_info = *info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1313
op->user_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1323
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1328
*info = op->user_info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1336
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1358
if (op->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
136
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
138
op = &dss_data.ovl_priv_data_array[i];
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1387
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
140
op->info.color_mode = OMAP_DSS_COLOR_RGB16;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1401
if (op->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
141
op->info.rotation_type = OMAP_DSS_ROT_DMA;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1425
if (ovl_manual_update(ovl) && op->extra_info_dirty) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
143
op->info.global_alpha = 255;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1447
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1453
e = op->enabled;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1462
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1468
if (op->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
147
op->info.zorder = 0;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1480
op->enabling = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1491
op->enabling = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
150
op->info.zorder =
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1503
op->enabling = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1512
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1518
if (!op->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
154
op->info.zorder =
drivers/video/fbdev/omap2/omapfb/dss/apply.c
158
op->info.zorder =
drivers/video/fbdev/omap2/omapfb/dss/apply.c
163
op->user_info = op->info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
203
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
218
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
220
if (!op->enabled && !op->enabling)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
222
else if (applying && op->user_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
223
oi = &op->user_info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
225
oi = &op->info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
298
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
300
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
309
if (op->extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
313
if (op->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
316
if (!op->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
320
if (op->info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
324
if (op->shadow_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
337
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
345
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
346
if (op->shadow_info_dirty || op->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
377
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
379
if (op->extra_info_dirty || op->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
541
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
573
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
579
dirty = op->info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
580
shadow_dirty = op->shadow_info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
617
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
625
if (!op->enabled || !op->info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
628
oi = &op->info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
643
op->enabled = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
648
op->info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
650
op->shadow_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
655
struct ovl_priv_data *op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
660
if (!op->extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
666
dispc_ovl_enable(ovl->id, op->enabled);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
667
dispc_ovl_set_fifo_threshold(ovl->id, op->fifo_low, op->fifo_high);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
671
op->extra_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
673
op->shadow_extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
783
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
790
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
791
op->shadow_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
792
op->shadow_extra_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
938
struct ovl_priv_data *op;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
940
op = get_ovl_priv(ovl);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
942
if (!op->user_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
945
op->user_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
946
op->info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
947
op->info = op->user_info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
998
struct ovl_priv_data *op;
drivers/video/fbdev/p9100.c
246
static int p9100_probe(struct platform_device *op)
drivers/video/fbdev/p9100.c
248
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/p9100.c
253
info = framebuffer_alloc(sizeof(struct p9100_par), &op->dev);
drivers/video/fbdev/p9100.c
263
info->fix.smem_start = op->resource[2].start;
drivers/video/fbdev/p9100.c
264
par->which_io = op->resource[2].flags & IORESOURCE_BITS;
drivers/video/fbdev/p9100.c
274
par->regs = of_ioremap(&op->resource[0], 0,
drivers/video/fbdev/p9100.c
280
info->screen_base = of_ioremap(&op->resource[2], 0,
drivers/video/fbdev/p9100.c
298
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/p9100.c
310
of_iounmap(&op->resource[2], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/p9100.c
313
of_iounmap(&op->resource[0], par->regs, sizeof(struct p9100_regs));
drivers/video/fbdev/p9100.c
322
static void p9100_remove(struct platform_device *op)
drivers/video/fbdev/p9100.c
324
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/p9100.c
330
of_iounmap(&op->resource[0], par->regs, sizeof(struct p9100_regs));
drivers/video/fbdev/p9100.c
331
of_iounmap(&op->resource[2], info->screen_base, info->fix.smem_len);
drivers/video/fbdev/sm501fb.c
1129
unsigned int op;
drivers/video/fbdev/sm501fb.c
1134
for (op = 0; op < (64*64*2)/8; op+=4)
drivers/video/fbdev/sm501fb.c
1135
smc501_writel(0x0, dst + op);
drivers/video/fbdev/sm501fb.c
1148
op = (dcol & 1) ? 1 : 3;
drivers/video/fbdev/sm501fb.c
1149
op <<= ((x % 4) * 2);
drivers/video/fbdev/sm501fb.c
1151
op |= readb(dst + (x / 4));
drivers/video/fbdev/sm501fb.c
1152
writeb(op, dst + (x / 4));
drivers/video/fbdev/sunxvr1000.c
113
static int gfb_probe(struct platform_device *op)
drivers/video/fbdev/sunxvr1000.c
115
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/sunxvr1000.c
120
info = framebuffer_alloc(sizeof(struct gfb_info), &op->dev);
drivers/video/fbdev/sunxvr1000.c
130
gp->fb_base_phys = op->resource[6].start;
drivers/video/fbdev/sunxvr1000.c
140
gp->fb_base = of_ioremap(&op->resource[6], 0,
drivers/video/fbdev/sunxvr1000.c
160
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/sunxvr1000.c
165
of_iounmap(&op->resource[6], gp->fb_base, gp->fb_size);
drivers/video/fbdev/tcx.c
339
static void tcx_unmap_regs(struct platform_device *op, struct fb_info *info,
drivers/video/fbdev/tcx.c
343
of_iounmap(&op->resource[7],
drivers/video/fbdev/tcx.c
346
of_iounmap(&op->resource[9],
drivers/video/fbdev/tcx.c
349
of_iounmap(&op->resource[8],
drivers/video/fbdev/tcx.c
352
of_iounmap(&op->resource[4],
drivers/video/fbdev/tcx.c
355
of_iounmap(&op->resource[0],
drivers/video/fbdev/tcx.c
359
static int tcx_probe(struct platform_device *op)
drivers/video/fbdev/tcx.c
361
struct device_node *dp = op->dev.of_node;
drivers/video/fbdev/tcx.c
366
info = framebuffer_alloc(sizeof(struct tcx_par), &op->dev);
drivers/video/fbdev/tcx.c
386
par->tec = of_ioremap(&op->resource[7], 0,
drivers/video/fbdev/tcx.c
388
par->thc = of_ioremap(&op->resource[9], 0,
drivers/video/fbdev/tcx.c
390
par->bt = of_ioremap(&op->resource[8], 0,
drivers/video/fbdev/tcx.c
392
info->screen_base = of_ioremap(&op->resource[0], 0,
drivers/video/fbdev/tcx.c
400
par->cplane = of_ioremap(&op->resource[4], 0,
drivers/video/fbdev/tcx.c
412
info->fix.smem_start = op->resource[0].start;
drivers/video/fbdev/tcx.c
413
par->which_io = op->resource[0].flags & IORESOURCE_BITS;
drivers/video/fbdev/tcx.c
431
par->mmap_map[i].poff = op->resource[j].start - info->fix.smem_start;
drivers/video/fbdev/tcx.c
460
dev_set_drvdata(&op->dev, info);
drivers/video/fbdev/tcx.c
474
tcx_unmap_regs(op, info, par);
drivers/video/fbdev/tcx.c
481
static void tcx_remove(struct platform_device *op)
drivers/video/fbdev/tcx.c
483
struct fb_info *info = dev_get_drvdata(&op->dev);
drivers/video/fbdev/tcx.c
489
tcx_unmap_regs(op, info, par);
drivers/video/fbdev/vga16fb.c
143
static inline int setop(int op)
drivers/video/fbdev/vga16fb.c
148
vga_io_w(VGA_GFX_D, op);
drivers/video/fbdev/via/accel.c
110
if (op != VIA_BITBLT_COLOR)
drivers/video/fbdev/via/accel.c
113
if (op == VIA_BITBLT_MONO)
drivers/video/fbdev/via/accel.c
116
if (op != VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
135
if (op == VIA_BITBLT_FILL)
drivers/video/fbdev/via/accel.c
147
if (op == VIA_BITBLT_FILL)
drivers/video/fbdev/via/accel.c
153
if (op == VIA_BITBLT_MONO)
drivers/video/fbdev/via/accel.c
160
if (op == VIA_BITBLT_FILL || !src_mem)
drivers/video/fbdev/via/accel.c
163
tmp = (width * height * (op == VIA_BITBLT_MONO ? 1 : (dst_bpp >> 3)) +
drivers/video/fbdev/via/accel.c
172
static int hw_bitblt_2(void __iomem *engine, u8 op, u32 width, u32 height,
drivers/video/fbdev/via/accel.c
180
if (!op || op > 3) {
drivers/video/fbdev/via/accel.c
181
printk(KERN_WARNING "hw_bitblt_2: Invalid operation: %d\n", op);
drivers/video/fbdev/via/accel.c
185
if (op != VIA_BITBLT_FILL && !src_mem && src_addr == dst_addr) {
drivers/video/fbdev/via/accel.c
198
if (op == VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
216
if (op == VIA_BITBLT_FILL)
drivers/video/fbdev/via/accel.c
252
if (op != VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
253
if (src_x & (op == VIA_BITBLT_MONO ? 0xFFFF8000 : 0xFFFFF000)
drivers/video/fbdev/via/accel.c
272
if (op == VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
274
} else if (op == VIA_BITBLT_MONO) {
drivers/video/fbdev/via/accel.c
279
if (op == VIA_BITBLT_FILL)
drivers/video/fbdev/via/accel.c
285
if (op == VIA_BITBLT_MONO)
drivers/video/fbdev/via/accel.c
292
if (op == VIA_BITBLT_FILL || !src_mem)
drivers/video/fbdev/via/accel.c
295
tmp = (width * height * (op == VIA_BITBLT_MONO ? 1 : (dst_bpp >> 3)) +
drivers/video/fbdev/via/accel.c
39
static int hw_bitblt_1(void __iomem *engine, u8 op, u32 width, u32 height,
drivers/video/fbdev/via/accel.c
47
if (!op || op > 3) {
drivers/video/fbdev/via/accel.c
48
printk(KERN_WARNING "hw_bitblt_1: Invalid operation: %d\n", op);
drivers/video/fbdev/via/accel.c
52
if (op != VIA_BITBLT_FILL && !src_mem && src_addr == dst_addr) {
drivers/video/fbdev/via/accel.c
65
if (op == VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
83
if (op != VIA_BITBLT_FILL) {
drivers/video/fbdev/via/accel.c
84
if (src_x & (op == VIA_BITBLT_MONO ? 0xFFFF8000 : 0xFFFFF000)
drivers/video/fbdev/via/viafbdev.c
728
u8 op;
drivers/video/fbdev/via/viafbdev.c
737
op = VIA_BITBLT_MONO;
drivers/video/fbdev/via/viafbdev.c
748
op = VIA_BITBLT_COLOR;
drivers/video/fbdev/via/viafbdev.c
751
if (shared->hw_bitblt(shared->vdev->engine_mmio, op,
drivers/video/fbdev/via/viafbdev.h
52
int (*hw_bitblt)(void __iomem *engine, u8 op, u32 width, u32 height,
drivers/video/fbdev/xilinxfb.c
473
static void xilinxfb_of_remove(struct platform_device *op)
drivers/video/fbdev/xilinxfb.c
475
xilinxfb_release(&op->dev);
drivers/virt/acrn/hypercall.h
161
static inline long hcall_set_irqline(u64 vmid, u64 op)
drivers/virt/acrn/hypercall.h
163
return acrn_hypercall2(HC_SET_IRQLINE, vmid, op);
drivers/virtio/virtio_ring.c
2735
#define VIRTQUEUE_CALL(vq, op, ...) \
drivers/virtio/virtio_ring.c
2738
typeof(split_ops.op(__VIRTQUEUE_CALL_vq, ##__VA_ARGS__)) ret; \
drivers/virtio/virtio_ring.c
2742
ret = split_ops.op(__VIRTQUEUE_CALL_vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2745
ret = packed_ops.op(__VIRTQUEUE_CALL_vq, ##__VA_ARGS__);\
drivers/virtio/virtio_ring.c
2748
ret = split_in_order_ops.op(vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2751
ret = packed_in_order_ops.op(vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2760
#define VOID_VIRTQUEUE_CALL(vq, op, ...) \
drivers/virtio/virtio_ring.c
2766
split_ops.op(__VIRTQUEUE_CALL_vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2769
packed_ops.op(__VIRTQUEUE_CALL_vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2772
split_in_order_ops.op(vq, ##__VA_ARGS__); \
drivers/virtio/virtio_ring.c
2775
packed_in_order_ops.op(vq, ##__VA_ARGS__); \
drivers/watchdog/cpwd.c
512
static int cpwd_probe(struct platform_device *op)
drivers/watchdog/cpwd.c
523
p = devm_kzalloc(&op->dev, sizeof(*p), GFP_KERNEL);
drivers/watchdog/cpwd.c
527
p->irq = op->archdata.irqs[0];
drivers/watchdog/cpwd.c
531
p->regs = of_ioremap(&op->resource[0], 0,
drivers/watchdog/cpwd.c
561
str_prop = of_get_property(op->dev.of_node, "model", NULL);
drivers/watchdog/cpwd.c
602
platform_set_drvdata(op, p);
drivers/watchdog/cpwd.c
611
of_iounmap(&op->resource[0], p->regs, 4 * WD_TIMER_REGSZ);
drivers/watchdog/cpwd.c
616
static void cpwd_remove(struct platform_device *op)
drivers/watchdog/cpwd.c
618
struct cpwd *p = platform_get_drvdata(op);
drivers/watchdog/cpwd.c
637
of_iounmap(&op->resource[0], p->regs, 4 * WD_TIMER_REGSZ);
drivers/watchdog/riowd.c
176
static int riowd_probe(struct platform_device *op)
drivers/watchdog/riowd.c
185
p = devm_kzalloc(&op->dev, sizeof(*p), GFP_KERNEL);
drivers/watchdog/riowd.c
191
p->regs = of_ioremap(&op->resource[0], 0, 2, DRIVER_NAME);
drivers/watchdog/riowd.c
208
platform_set_drvdata(op, p);
drivers/watchdog/riowd.c
213
of_iounmap(&op->resource[0], p->regs, 2);
drivers/watchdog/riowd.c
219
static void riowd_remove(struct platform_device *op)
drivers/watchdog/riowd.c
221
struct riowd *p = platform_get_drvdata(op);
drivers/watchdog/riowd.c
224
of_iounmap(&op->resource[0], p->regs, 2);
drivers/xen/acpi.c
45
struct xen_platform_op op = {
drivers/xen/acpi.c
62
HYPERVISOR_platform_op(&op);
drivers/xen/dbgp.c
10
static int xen_dbgp_op(struct usb_hcd *hcd, int op)
drivers/xen/dbgp.c
20
dbgp.op = op;
drivers/xen/efi.c
104
struct xen_platform_op op = INIT_EFI_OP(set_wakeup_time);
drivers/xen/efi.c
106
BUILD_BUG_ON(sizeof(*tm) != sizeof(efi_data(op).u.set_wakeup_time));
drivers/xen/efi.c
108
efi_data(op).misc = XEN_EFI_SET_WAKEUP_TIME_ENABLE;
drivers/xen/efi.c
110
memcpy(&efi_data(op).u.set_wakeup_time, tm, sizeof(*tm));
drivers/xen/efi.c
112
efi_data(op).misc |= XEN_EFI_SET_WAKEUP_TIME_ENABLE_ONLY;
drivers/xen/efi.c
114
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
117
return efi_data(op).status;
drivers/xen/efi.c
124
struct xen_platform_op op = INIT_EFI_OP(get_variable);
drivers/xen/efi.c
126
set_xen_guest_handle(efi_data(op).u.get_variable.name, name);
drivers/xen/efi.c
128
sizeof(efi_data(op).u.get_variable.vendor_guid));
drivers/xen/efi.c
129
memcpy(&efi_data(op).u.get_variable.vendor_guid, vendor, sizeof(*vendor));
drivers/xen/efi.c
130
efi_data(op).u.get_variable.size = *data_size;
drivers/xen/efi.c
131
set_xen_guest_handle(efi_data(op).u.get_variable.data, data);
drivers/xen/efi.c
133
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
136
*data_size = efi_data(op).u.get_variable.size;
drivers/xen/efi.c
138
*attr = efi_data(op).misc;
drivers/xen/efi.c
140
return efi_data(op).status;
drivers/xen/efi.c
147
struct xen_platform_op op = INIT_EFI_OP(get_next_variable_name);
drivers/xen/efi.c
149
efi_data(op).u.get_next_variable_name.size = *name_size;
drivers/xen/efi.c
150
set_xen_guest_handle(efi_data(op).u.get_next_variable_name.name, name);
drivers/xen/efi.c
152
sizeof(efi_data(op).u.get_next_variable_name.vendor_guid));
drivers/xen/efi.c
153
memcpy(&efi_data(op).u.get_next_variable_name.vendor_guid, vendor,
drivers/xen/efi.c
156
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
159
*name_size = efi_data(op).u.get_next_variable_name.size;
drivers/xen/efi.c
160
memcpy(vendor, &efi_data(op).u.get_next_variable_name.vendor_guid,
drivers/xen/efi.c
163
return efi_data(op).status;
drivers/xen/efi.c
170
struct xen_platform_op op = INIT_EFI_OP(set_variable);
drivers/xen/efi.c
172
set_xen_guest_handle(efi_data(op).u.set_variable.name, name);
drivers/xen/efi.c
173
efi_data(op).misc = attr;
drivers/xen/efi.c
175
sizeof(efi_data(op).u.set_variable.vendor_guid));
drivers/xen/efi.c
176
memcpy(&efi_data(op).u.set_variable.vendor_guid, vendor, sizeof(*vendor));
drivers/xen/efi.c
177
efi_data(op).u.set_variable.size = data_size;
drivers/xen/efi.c
178
set_xen_guest_handle(efi_data(op).u.set_variable.data, data);
drivers/xen/efi.c
180
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
183
return efi_data(op).status;
drivers/xen/efi.c
190
struct xen_platform_op op = INIT_EFI_OP(query_variable_info);
drivers/xen/efi.c
195
efi_data(op).u.query_variable_info.attr = attr;
drivers/xen/efi.c
197
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
200
*storage_space = efi_data(op).u.query_variable_info.max_store_size;
drivers/xen/efi.c
201
*remaining_space = efi_data(op).u.query_variable_info.remain_store_size;
drivers/xen/efi.c
202
*max_variable_size = efi_data(op).u.query_variable_info.max_size;
drivers/xen/efi.c
204
return efi_data(op).status;
drivers/xen/efi.c
209
struct xen_platform_op op = INIT_EFI_OP(get_next_high_monotonic_count);
drivers/xen/efi.c
211
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
214
*count = efi_data(op).misc;
drivers/xen/efi.c
216
return efi_data(op).status;
drivers/xen/efi.c
222
struct xen_platform_op op = INIT_EFI_OP(update_capsule);
drivers/xen/efi.c
227
set_xen_guest_handle(efi_data(op).u.update_capsule.capsule_header_array,
drivers/xen/efi.c
229
efi_data(op).u.update_capsule.capsule_count = count;
drivers/xen/efi.c
230
efi_data(op).u.update_capsule.sg_list = sg_list;
drivers/xen/efi.c
232
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
235
return efi_data(op).status;
drivers/xen/efi.c
241
struct xen_platform_op op = INIT_EFI_OP(query_capsule_capabilities);
drivers/xen/efi.c
246
set_xen_guest_handle(efi_data(op).u.query_capsule_capabilities.capsule_header_array,
drivers/xen/efi.c
248
efi_data(op).u.query_capsule_capabilities.capsule_count = count;
drivers/xen/efi.c
250
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
253
*max_size = efi_data(op).u.query_capsule_capabilities.max_capsule_size;
drivers/xen/efi.c
254
*reset_type = efi_data(op).u.query_capsule_capabilities.reset_type;
drivers/xen/efi.c
256
return efi_data(op).status;
drivers/xen/efi.c
301
struct xen_platform_op op;
drivers/xen/efi.c
302
union xenpf_efi_info *info = &op.u.firmware_info.u.efi_info;
drivers/xen/efi.c
308
op = (struct xen_platform_op) {
drivers/xen/efi.c
318
rc = HYPERVISOR_platform_op(&op);
drivers/xen/efi.c
42
#define efi_data(op) (op.u.efi_runtime_call)
drivers/xen/efi.c
46
struct xen_platform_op op = INIT_EFI_OP(get_time);
drivers/xen/efi.c
48
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
52
BUILD_BUG_ON(sizeof(*tm) != sizeof(efi_data(op).u.get_time.time));
drivers/xen/efi.c
53
memcpy(tm, &efi_data(op).u.get_time.time, sizeof(*tm));
drivers/xen/efi.c
57
tc->resolution = efi_data(op).u.get_time.resolution;
drivers/xen/efi.c
58
tc->accuracy = efi_data(op).u.get_time.accuracy;
drivers/xen/efi.c
59
tc->sets_to_zero = !!(efi_data(op).misc &
drivers/xen/efi.c
63
return efi_data(op).status;
drivers/xen/efi.c
68
struct xen_platform_op op = INIT_EFI_OP(set_time);
drivers/xen/efi.c
70
BUILD_BUG_ON(sizeof(*tm) != sizeof(efi_data(op).u.set_time));
drivers/xen/efi.c
71
memcpy(&efi_data(op).u.set_time, tm, sizeof(*tm));
drivers/xen/efi.c
73
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
76
return efi_data(op).status;
drivers/xen/efi.c
83
struct xen_platform_op op = INIT_EFI_OP(get_wakeup_time);
drivers/xen/efi.c
85
if (HYPERVISOR_platform_op(&op) < 0)
drivers/xen/efi.c
89
BUILD_BUG_ON(sizeof(*tm) != sizeof(efi_data(op).u.get_wakeup_time));
drivers/xen/efi.c
90
memcpy(tm, &efi_data(op).u.get_wakeup_time, sizeof(*tm));
drivers/xen/efi.c
94
*enabled = !!(efi_data(op).misc & XEN_EFI_GET_WAKEUP_TIME_ENABLED);
drivers/xen/efi.c
97
*pending = !!(efi_data(op).misc & XEN_EFI_GET_WAKEUP_TIME_PENDING);
drivers/xen/efi.c
99
return efi_data(op).status;
drivers/xen/events/events_base.c
2220
xen_hvm_evtchn_upcall_vector_t op = {
drivers/xen/events/events_base.c
2225
rc = HYPERVISOR_hvm_op(HVMOP_set_evtchn_upcall_vector, &op);
drivers/xen/gntalloc.c
121
static int add_grefs(struct ioctl_gntalloc_alloc_gref *op,
drivers/xen/gntalloc.c
129
readonly = !(op->flags & GNTALLOC_FLAG_WRITABLE);
drivers/xen/gntalloc.c
130
for (i = 0; i < op->count; i++) {
drivers/xen/gntalloc.c
139
gref->file_index = op->index + i * PAGE_SIZE;
drivers/xen/gntalloc.c
147
rc = gnttab_grant_foreign_access(op->domid,
drivers/xen/gntalloc.c
165
gref_size -= (op->count - i);
drivers/xen/gntalloc.c
273
struct ioctl_gntalloc_alloc_gref op;
drivers/xen/gntalloc.c
278
if (copy_from_user(&op, arg, sizeof(op))) {
drivers/xen/gntalloc.c
283
gref_ids = kcalloc(op.count, sizeof(gref_ids[0]), GFP_KERNEL);
drivers/xen/gntalloc.c
295
if (gref_size + op.count > limit) {
drivers/xen/gntalloc.c
300
gref_size += op.count;
drivers/xen/gntalloc.c
301
op.index = priv->index;
drivers/xen/gntalloc.c
302
priv->index += op.count * PAGE_SIZE;
drivers/xen/gntalloc.c
305
rc = add_grefs(&op, gref_ids, priv);
drivers/xen/gntalloc.c
316
if (copy_to_user(arg, &op, sizeof(op))) {
drivers/xen/gntalloc.c
321
sizeof(gref_ids[0]) * op.count)) {
drivers/xen/gntalloc.c
336
struct ioctl_gntalloc_dealloc_gref op;
drivers/xen/gntalloc.c
341
if (copy_from_user(&op, arg, sizeof(op))) {
drivers/xen/gntalloc.c
347
gref = find_grefs(priv, op.index, op.count);
drivers/xen/gntalloc.c
353
for (i = 0; i < op.count; i++) {
drivers/xen/gntalloc.c
374
struct ioctl_gntalloc_unmap_notify op;
drivers/xen/gntalloc.c
380
if (copy_from_user(&op, arg, sizeof(op)))
drivers/xen/gntalloc.c
383
index = op.index & ~(PAGE_SIZE - 1);
drivers/xen/gntalloc.c
384
pgoff = op.index & (PAGE_SIZE - 1);
drivers/xen/gntalloc.c
394
if (op.action & ~(UNMAP_NOTIFY_CLEAR_BYTE|UNMAP_NOTIFY_SEND_EVENT)) {
drivers/xen/gntalloc.c
406
if (op.action & UNMAP_NOTIFY_SEND_EVENT) {
drivers/xen/gntalloc.c
407
if (evtchn_get(op.event_channel_port)) {
drivers/xen/gntalloc.c
416
gref->notify.flags = op.action;
drivers/xen/gntalloc.c
418
gref->notify.event = op.event_channel_port;
drivers/xen/gntdev-dmabuf.c
724
struct ioctl_gntdev_dmabuf_exp_from_refs op;
drivers/xen/gntdev-dmabuf.c
733
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev-dmabuf.c
736
if (unlikely(gntdev_test_page_count(op.count)))
drivers/xen/gntdev-dmabuf.c
739
refs = kcalloc(op.count, sizeof(*refs), GFP_KERNEL);
drivers/xen/gntdev-dmabuf.c
743
if (copy_from_user(refs, u->refs, sizeof(*refs) * op.count) != 0) {
drivers/xen/gntdev-dmabuf.c
748
ret = dmabuf_exp_from_refs(priv, op.flags, op.count,
drivers/xen/gntdev-dmabuf.c
749
op.domid, refs, &op.fd);
drivers/xen/gntdev-dmabuf.c
753
if (copy_to_user(u, &op, sizeof(op)) != 0)
drivers/xen/gntdev-dmabuf.c
764
struct ioctl_gntdev_dmabuf_exp_wait_released op;
drivers/xen/gntdev-dmabuf.c
766
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev-dmabuf.c
769
return dmabuf_exp_wait_released(priv->dmabuf_priv, op.fd,
drivers/xen/gntdev-dmabuf.c
770
op.wait_to_ms);
drivers/xen/gntdev-dmabuf.c
776
struct ioctl_gntdev_dmabuf_imp_to_refs op;
drivers/xen/gntdev-dmabuf.c
780
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev-dmabuf.c
783
if (unlikely(gntdev_test_page_count(op.count)))
drivers/xen/gntdev-dmabuf.c
787
priv->dma_dev, op.fd,
drivers/xen/gntdev-dmabuf.c
788
op.count, op.domid);
drivers/xen/gntdev-dmabuf.c
793
sizeof(*u->refs) * op.count) != 0) {
drivers/xen/gntdev-dmabuf.c
800
dmabuf_imp_release(priv->dmabuf_priv, op.fd);
drivers/xen/gntdev-dmabuf.c
807
struct ioctl_gntdev_dmabuf_imp_release op;
drivers/xen/gntdev-dmabuf.c
809
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev-dmabuf.c
812
return dmabuf_imp_release(priv->dmabuf_priv, op.fd);
drivers/xen/gntdev.c
649
struct ioctl_gntdev_map_grant_ref op;
drivers/xen/gntdev.c
653
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev.c
655
pr_debug("priv %p, add %d\n", priv, op.count);
drivers/xen/gntdev.c
656
if (unlikely(gntdev_test_page_count(op.count)))
drivers/xen/gntdev.c
660
map = gntdev_alloc_map(priv, op.count, 0 /* This is not a dma-buf. */);
drivers/xen/gntdev.c
665
sizeof(map->grants[0]) * op.count) != 0) {
drivers/xen/gntdev.c
672
op.index = map->index << PAGE_SHIFT;
drivers/xen/gntdev.c
675
if (copy_to_user(u, &op, sizeof(op)) != 0)
drivers/xen/gntdev.c
684
struct ioctl_gntdev_unmap_grant_ref op;
drivers/xen/gntdev.c
688
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev.c
690
pr_debug("priv %p, del %d+%d\n", priv, (int)op.index, (int)op.count);
drivers/xen/gntdev.c
693
map = gntdev_find_map_index(priv, op.index >> PAGE_SHIFT, op.count);
drivers/xen/gntdev.c
707
struct ioctl_gntdev_get_offset_for_vaddr op;
drivers/xen/gntdev.c
712
if (copy_from_user(&op, u, sizeof(op)) != 0)
drivers/xen/gntdev.c
714
pr_debug("priv %p, offset for vaddr %lx\n", priv, (unsigned long)op.vaddr);
drivers/xen/gntdev.c
717
vma = find_vma(current->mm, op.vaddr);
drivers/xen/gntdev.c
725
op.offset = map->index << PAGE_SHIFT;
drivers/xen/gntdev.c
726
op.count = map->count;
drivers/xen/gntdev.c
732
if (rv == 0 && copy_to_user(u, &op, sizeof(op)) != 0)
drivers/xen/gntdev.c
739
struct ioctl_gntdev_unmap_notify op;
drivers/xen/gntdev.c
745
if (copy_from_user(&op, u, sizeof(op)))
drivers/xen/gntdev.c
748
if (op.action & ~(UNMAP_NOTIFY_CLEAR_BYTE|UNMAP_NOTIFY_SEND_EVENT))
drivers/xen/gntdev.c
758
if (op.action & UNMAP_NOTIFY_SEND_EVENT) {
drivers/xen/gntdev.c
759
if (evtchn_get(op.event_channel_port))
drivers/xen/gntdev.c
763
out_flags = op.action;
drivers/xen/gntdev.c
764
out_event = op.event_channel_port;
drivers/xen/gntdev.c
771
if (op.index >= begin && op.index < end)
drivers/xen/gntdev.c
778
if ((op.action & UNMAP_NOTIFY_CLEAR_BYTE) &&
drivers/xen/gntdev.c
787
map->notify.flags = op.action;
drivers/xen/gntdev.c
788
map->notify.addr = op.index - (map->index << PAGE_SHIFT);
drivers/xen/gntdev.c
789
map->notify.event = op.event_channel_port;
drivers/xen/gntdev.c
890
struct gnttab_copy *op;
drivers/xen/gntdev.c
904
op = &batch->ops[batch->nr_ops];
drivers/xen/gntdev.c
905
op->flags = 0;
drivers/xen/gntdev.c
908
op->source.u.ref = seg->source.foreign.ref;
drivers/xen/gntdev.c
909
op->source.domid = seg->source.foreign.domid;
drivers/xen/gntdev.c
910
op->source.offset = seg->source.foreign.offset + copied;
drivers/xen/gntdev.c
911
op->flags |= GNTCOPY_source_gref;
drivers/xen/gntdev.c
922
op->source.u.gmfn = gfn;
drivers/xen/gntdev.c
923
op->source.domid = DOMID_SELF;
drivers/xen/gntdev.c
924
op->source.offset = off;
drivers/xen/gntdev.c
928
op->dest.u.ref = seg->dest.foreign.ref;
drivers/xen/gntdev.c
929
op->dest.domid = seg->dest.foreign.domid;
drivers/xen/gntdev.c
930
op->dest.offset = seg->dest.foreign.offset + copied;
drivers/xen/gntdev.c
931
op->flags |= GNTCOPY_dest_gref;
drivers/xen/gntdev.c
942
op->dest.u.gmfn = gfn;
drivers/xen/gntdev.c
943
op->dest.domid = DOMID_SELF;
drivers/xen/gntdev.c
944
op->dest.offset = off;
drivers/xen/gntdev.c
947
op->len = len;
drivers/xen/grant-table.c
1173
struct gnttab_map_grant_ref *op;
drivers/xen/grant-table.c
1177
for (op = batch; op < batch + count; op++)
drivers/xen/grant-table.c
1178
if (op->status == GNTST_eagain)
drivers/xen/grant-table.c
1179
gnttab_retry_eagain_gop(GNTTABOP_map_grant_ref, op,
drivers/xen/grant-table.c
1180
&op->status, __func__);
drivers/xen/grant-table.c
1186
struct gnttab_copy *op;
drivers/xen/grant-table.c
1190
for (op = batch; op < batch + count; op++)
drivers/xen/grant-table.c
1191
if (op->status == GNTST_eagain)
drivers/xen/grant-table.c
1192
gnttab_retry_eagain_gop(GNTTABOP_copy, op,
drivers/xen/grant-table.c
1193
&op->status, __func__);
drivers/xen/pcpu.c
279
struct xen_platform_op op = {
drivers/xen/pcpu.c
285
ret = HYPERVISOR_platform_op(&op);
drivers/xen/pcpu.c
289
info = &op.u.pcpu_info;
drivers/xen/pcpu.c
409
struct xen_platform_op op = {
drivers/xen/pcpu.c
417
set_xen_guest_handle(op.u.set_pminfo.pdc, buf);
drivers/xen/pcpu.c
418
ret = HYPERVISOR_platform_op(&op);
drivers/xen/pcpu.c
79
struct xen_platform_op op = {
drivers/xen/pcpu.c
85
return HYPERVISOR_platform_op(&op);
drivers/xen/pcpu.c
90
struct xen_platform_op op = {
drivers/xen/pcpu.c
96
return HYPERVISOR_platform_op(&op);
drivers/xen/privcmd.c
109
ret = privcmd_call(hypercall.op,
drivers/xen/xen-acpi-pad.c
26
struct xen_platform_op op;
drivers/xen/xen-acpi-pad.c
28
op.cmd = XENPF_core_parking;
drivers/xen/xen-acpi-pad.c
29
op.u.core_parking.type = XEN_CORE_PARKING_SET;
drivers/xen/xen-acpi-pad.c
30
op.u.core_parking.idle_nums = idle_nums;
drivers/xen/xen-acpi-pad.c
32
return HYPERVISOR_platform_op(&op);
drivers/xen/xen-acpi-pad.c
37
struct xen_platform_op op;
drivers/xen/xen-acpi-pad.c
39
op.cmd = XENPF_core_parking;
drivers/xen/xen-acpi-pad.c
40
op.u.core_parking.type = XEN_CORE_PARKING_GET;
drivers/xen/xen-acpi-pad.c
42
return HYPERVISOR_platform_op(&op)
drivers/xen/xen-acpi-pad.c
43
?: op.u.core_parking.idle_nums;
drivers/xen/xen-acpi-processor.c
102
op.u.set_pminfo.power.count = ok;
drivers/xen/xen-acpi-processor.c
103
op.u.set_pminfo.power.flags.bm_control = _pr->flags.bm_control;
drivers/xen/xen-acpi-processor.c
104
op.u.set_pminfo.power.flags.bm_check = _pr->flags.bm_check;
drivers/xen/xen-acpi-processor.c
105
op.u.set_pminfo.power.flags.has_cst = _pr->flags.has_cst;
drivers/xen/xen-acpi-processor.c
106
op.u.set_pminfo.power.flags.power_setup_done =
drivers/xen/xen-acpi-processor.c
109
set_xen_guest_handle(op.u.set_pminfo.power.states, dst_cx_states);
drivers/xen/xen-acpi-processor.c
112
ret = HYPERVISOR_platform_op(&op);
drivers/xen/xen-acpi-processor.c
206
struct xen_platform_op op = {
drivers/xen/xen-acpi-processor.c
215
dst_perf = &op.u.set_pminfo.perf;
drivers/xen/xen-acpi-processor.c
240
ret = HYPERVISOR_platform_op(&op);
drivers/xen/xen-acpi-processor.c
288
struct xen_platform_op op = {
drivers/xen/xen-acpi-processor.c
295
info = &op.u.pcpu_info;
drivers/xen/xen-acpi-processor.c
298
ret = HYPERVISOR_platform_op(&op);
drivers/xen/xen-acpi-processor.c
303
last_cpu = op.u.pcpu_info.max_present;
drivers/xen/xen-acpi-processor.c
306
ret = HYPERVISOR_platform_op(&op);
drivers/xen/xen-acpi-processor.c
53
struct xen_platform_op op = {
drivers/xen/xen-pciback/conf_space.c
27
#define DEFINE_PCI_CONFIG(op, size, type) \
drivers/xen/xen-pciback/conf_space.c
28
int xen_pcibk_##op##_config_##size \
drivers/xen/xen-pciback/conf_space.c
31
return pci_##op##_config_##size(dev, offset, value); \
drivers/xen/xen-pciback/pciback.h
44
struct xen_pci_op op;
drivers/xen/xen-pciback/pciback_ops.c
143
struct pci_dev *dev, struct xen_pci_op *op)
drivers/xen/xen-pciback/pciback_ops.c
158
op->value = 0;
drivers/xen/xen-pciback/pciback_ops.c
165
op->value = dev->irq ? xen_pirq_from_irq(dev->irq) : 0;
drivers/xen/xen-pciback/pciback_ops.c
167
dev_dbg(&dev->dev, "MSI: %d\n", op->value);
drivers/xen/xen-pciback/pciback_ops.c
178
struct pci_dev *dev, struct xen_pci_op *op)
drivers/xen/xen-pciback/pciback_ops.c
189
op->value = dev->irq ? xen_pirq_from_irq(dev->irq) : 0;
drivers/xen/xen-pciback/pciback_ops.c
191
dev_dbg(&dev->dev, "MSI: %d\n", op->value);
drivers/xen/xen-pciback/pciback_ops.c
198
struct pci_dev *dev, struct xen_pci_op *op)
drivers/xen/xen-pciback/pciback_ops.c
207
if (op->value > SH_INFO_MAX_VEC)
drivers/xen/xen-pciback/pciback_ops.c
222
entries = kmalloc_objs(*entries, op->value);
drivers/xen/xen-pciback/pciback_ops.c
226
for (i = 0; i < op->value; i++) {
drivers/xen/xen-pciback/pciback_ops.c
227
entries[i].entry = op->msix_entries[i].entry;
drivers/xen/xen-pciback/pciback_ops.c
228
entries[i].vector = op->msix_entries[i].vector;
drivers/xen/xen-pciback/pciback_ops.c
231
result = pci_enable_msix_exact(dev, entries, op->value);
drivers/xen/xen-pciback/pciback_ops.c
233
for (i = 0; i < op->value; i++) {
drivers/xen/xen-pciback/pciback_ops.c
234
op->msix_entries[i].entry = entries[i].entry;
drivers/xen/xen-pciback/pciback_ops.c
236
op->msix_entries[i].vector =
drivers/xen/xen-pciback/pciback_ops.c
239
op->msix_entries[i].vector);
drivers/xen/xen-pciback/pciback_ops.c
247
op->value = result;
drivers/xen/xen-pciback/pciback_ops.c
257
struct pci_dev *dev, struct xen_pci_op *op)
drivers/xen/xen-pciback/pciback_ops.c
272
op->value = dev->irq ? xen_pirq_from_irq(dev->irq) : 0;
drivers/xen/xen-pciback/pciback_ops.c
274
dev_dbg(&dev->dev, "MSI-X: %d\n", op->value);
drivers/xen/xen-pciback/pciback_ops.c
325
struct xen_pci_op *op = &pdev->op;
drivers/xen/xen-pciback/pciback_ops.c
331
*op = pdev->sh_info->op;
drivers/xen/xen-pciback/pciback_ops.c
333
dev = xen_pcibk_get_pci_dev(pdev, op->domain, op->bus, op->devfn);
drivers/xen/xen-pciback/pciback_ops.c
336
op->err = XEN_PCI_ERR_dev_not_found;
drivers/xen/xen-pciback/pciback_ops.c
341
switch (op->cmd) {
drivers/xen/xen-pciback/pciback_ops.c
343
op->err = xen_pcibk_config_read(dev,
drivers/xen/xen-pciback/pciback_ops.c
344
op->offset, op->size, &op->value);
drivers/xen/xen-pciback/pciback_ops.c
347
op->err = xen_pcibk_config_write(dev,
drivers/xen/xen-pciback/pciback_ops.c
348
op->offset, op->size, op->value);
drivers/xen/xen-pciback/pciback_ops.c
352
op->err = xen_pcibk_enable_msi(pdev, dev, op);
drivers/xen/xen-pciback/pciback_ops.c
355
op->err = xen_pcibk_disable_msi(pdev, dev, op);
drivers/xen/xen-pciback/pciback_ops.c
358
nr = op->value;
drivers/xen/xen-pciback/pciback_ops.c
359
op->err = xen_pcibk_enable_msix(pdev, dev, op);
drivers/xen/xen-pciback/pciback_ops.c
362
op->err = xen_pcibk_disable_msix(pdev, dev, op);
drivers/xen/xen-pciback/pciback_ops.c
366
op->err = XEN_PCI_ERR_not_implemented;
drivers/xen/xen-pciback/pciback_ops.c
370
if (!op->err && dev && dev_data) {
drivers/xen/xen-pciback/pciback_ops.c
375
pdev->sh_info->op.err = op->err;
drivers/xen/xen-pciback/pciback_ops.c
376
pdev->sh_info->op.value = op->value;
drivers/xen/xen-pciback/pciback_ops.c
378
if (op->cmd == XEN_PCI_OP_enable_msix && op->err == 0) {
drivers/xen/xen-pciback/pciback_ops.c
382
pdev->sh_info->op.msix_entries[i].vector =
drivers/xen/xen-pciback/pciback_ops.c
383
op->msix_entries[i].vector;
drivers/xen/xen-scsiback.c
1078
static void scsiback_do_1lun_hotplug(struct vscsibk_info *info, int op,
drivers/xen/xen-scsiback.c
1123
switch (op) {
drivers/xen/xen-scsiback.c
1158
static void scsiback_do_lun_hotplug(struct vscsibk_info *info, int op)
drivers/xen/xen-scsiback.c
1170
scsiback_do_1lun_hotplug(info, op, dir[i]);
drivers/xen/xenfs/xensyms.c
131
set_xen_guest_handle(xs->op.u.symdata.name, xs->name);
drivers/xen/xenfs/xensyms.c
132
xs->op.cmd = XENPF_get_symbol;
drivers/xen/xenfs/xensyms.c
133
xs->op.u.symdata.namelen = xs->namelen;
drivers/xen/xenfs/xensyms.c
17
struct xen_platform_op op;
drivers/xen/xenfs/xensyms.c
26
struct xenpf_symdata *symdata = &xs->op.u.symdata;
drivers/xen/xenfs/xensyms.c
34
ret = HYPERVISOR_platform_op(&xs->op);
drivers/xen/xenfs/xensyms.c
53
ret = HYPERVISOR_platform_op(&xs->op);
drivers/xen/xenfs/xensyms.c
69
xs->op.u.symdata.symnum = *pos;
drivers/xen/xenfs/xensyms.c
81
*pos = xs->op.u.symdata.symnum;
drivers/xen/xenfs/xensyms.c
92
struct xenpf_symdata *symdata = &xs->op.u.symdata;
fs/afs/dir.c
1222
void afs_check_for_remote_deletion(struct afs_operation *op)
fs/afs/dir.c
1224
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/dir.c
1226
switch (afs_op_abort_code(op)) {
fs/afs/dir.c
1237
static void afs_vnode_new_inode(struct afs_operation *op)
fs/afs/dir.c
1239
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir.c
1240
struct afs_vnode_param *vp = &op->file[1];
fs/afs/dir.c
1246
ASSERTCMP(afs_op_error(op), ==, 0);
fs/afs/dir.c
1248
inode = afs_iget(op, vp);
fs/afs/dir.c
1253
afs_op_accumulate_error(op, PTR_ERR(inode), 0);
fs/afs/dir.c
1262
afs_init_new_symlink(vnode, op);
fs/afs/dir.c
1263
if (!afs_op_error(op))
fs/afs/dir.c
1264
afs_cache_permit(vnode, op->key, vnode->cb_break, &vp->scb);
fs/afs/dir.c
1265
d_instantiate(op->dentry, inode);
fs/afs/dir.c
1268
static void afs_create_success(struct afs_operation *op)
fs/afs/dir.c
1270
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1271
op->ctime = op->file[0].scb.status.mtime_client;
fs/afs/dir.c
1272
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir.c
1273
afs_update_dentry_version(op, &op->file[0], op->dentry);
fs/afs/dir.c
1274
afs_vnode_new_inode(op);
fs/afs/dir.c
1277
static void afs_create_edit_dir(struct afs_operation *op)
fs/afs/dir.c
1280
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir.c
1281
struct afs_vnode_param *vp = &op->file[1];
fs/afs/dir.c
1284
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1290
afs_edit_dir_add(dvnode, &op->dentry->d_name, &vp->fid,
fs/afs/dir.c
1291
op->create.reason);
fs/afs/dir.c
1296
static void afs_create_put(struct afs_operation *op)
fs/afs/dir.c
1298
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1300
if (afs_op_error(op))
fs/afs/dir.c
1301
d_drop(op->dentry);
fs/afs/dir.c
1319
struct afs_operation *op;
fs/afs/dir.c
1326
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1327
if (IS_ERR(op)) {
fs/afs/dir.c
1329
return ERR_CAST(op);
fs/afs/dir.c
1334
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1335
op->file[0].dv_delta = 1;
fs/afs/dir.c
1336
op->file[0].modification = true;
fs/afs/dir.c
1337
op->file[0].update_ctime = true;
fs/afs/dir.c
1338
op->dentry = dentry;
fs/afs/dir.c
1339
op->create.mode = S_IFDIR | mode;
fs/afs/dir.c
1340
op->create.reason = afs_edit_dir_for_mkdir;
fs/afs/dir.c
1341
op->mtime = current_time(dir);
fs/afs/dir.c
1342
op->ops = &afs_mkdir_operation;
fs/afs/dir.c
1343
ret = afs_do_sync_operation(op);
fs/afs/dir.c
1363
static void afs_rmdir_success(struct afs_operation *op)
fs/afs/dir.c
1365
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1366
op->ctime = op->file[0].scb.status.mtime_client;
fs/afs/dir.c
1367
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir.c
1368
afs_update_dentry_version(op, &op->file[0], op->dentry);
fs/afs/dir.c
1371
static void afs_rmdir_edit_dir(struct afs_operation *op)
fs/afs/dir.c
1374
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir.c
1377
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1378
afs_dir_remove_subdir(op->dentry);
fs/afs/dir.c
1384
afs_edit_dir_remove(dvnode, &op->dentry->d_name,
fs/afs/dir.c
1390
static void afs_rmdir_put(struct afs_operation *op)
fs/afs/dir.c
1392
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1393
if (op->file[1].vnode)
fs/afs/dir.c
1394
up_write(&op->file[1].vnode->rmdir_lock);
fs/afs/dir.c
1411
struct afs_operation *op;
fs/afs/dir.c
1418
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1419
if (IS_ERR(op))
fs/afs/dir.c
1420
return PTR_ERR(op);
fs/afs/dir.c
1424
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1425
op->file[0].dv_delta = 1;
fs/afs/dir.c
1426
op->file[0].modification = true;
fs/afs/dir.c
1427
op->file[0].update_ctime = true;
fs/afs/dir.c
1429
op->dentry = dentry;
fs/afs/dir.c
1430
op->ops = &afs_rmdir_operation;
fs/afs/dir.c
1435
ret = afs_validate(vnode, op->key);
fs/afs/dir.c
1444
op->file[1].vnode = vnode;
fs/afs/dir.c
1447
ret = afs_do_sync_operation(op);
fs/afs/dir.c
1457
ret = afs_put_operation(op);
fs/afs/dir.c
1471
static void afs_dir_remove_link(struct afs_operation *op)
fs/afs/dir.c
1473
struct afs_vnode *dvnode = op->file[0].vnode;
fs/afs/dir.c
1474
struct afs_vnode *vnode = op->file[1].vnode;
fs/afs/dir.c
1475
struct dentry *dentry = op->dentry;
fs/afs/dir.c
1478
if (afs_op_error(op) ||
fs/afs/dir.c
1479
(op->file[1].scb.have_status && op->file[1].scb.have_error))
fs/afs/dir.c
1500
ret = afs_validate(vnode, op->key);
fs/afs/dir.c
1502
afs_op_set_error(op, ret);
fs/afs/dir.c
1505
_debug("nlink %d [val %d]", vnode->netfs.inode.i_nlink, afs_op_error(op));
fs/afs/dir.c
1508
static void afs_unlink_success(struct afs_operation *op)
fs/afs/dir.c
1510
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1511
op->ctime = op->file[0].scb.status.mtime_client;
fs/afs/dir.c
1512
afs_check_dir_conflict(op, &op->file[0]);
fs/afs/dir.c
1513
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir.c
1514
afs_vnode_commit_status(op, &op->file[1]);
fs/afs/dir.c
1515
afs_update_dentry_version(op, &op->file[0], op->dentry);
fs/afs/dir.c
1516
afs_dir_remove_link(op);
fs/afs/dir.c
1519
static void afs_unlink_edit_dir(struct afs_operation *op)
fs/afs/dir.c
1522
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir.c
1525
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1530
afs_edit_dir_remove(dvnode, &op->dentry->d_name,
fs/afs/dir.c
1536
static void afs_unlink_put(struct afs_operation *op)
fs/afs/dir.c
1538
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1539
if (op->unlink.need_rehash && afs_op_error(op) < 0 && afs_op_error(op) != -ENOENT)
fs/afs/dir.c
1540
d_rehash(op->dentry);
fs/afs/dir.c
1557
struct afs_operation *op;
fs/afs/dir.c
1568
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1569
if (IS_ERR(op))
fs/afs/dir.c
1570
return PTR_ERR(op);
fs/afs/dir.c
1574
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1575
op->file[0].dv_delta = 1;
fs/afs/dir.c
1576
op->file[0].modification = true;
fs/afs/dir.c
1577
op->file[0].update_ctime = true;
fs/afs/dir.c
1580
ret = afs_validate(vnode, op->key);
fs/afs/dir.c
1582
afs_op_set_error(op, ret);
fs/afs/dir.c
1591
afs_op_set_error(op, afs_sillyrename(dvnode, vnode, dentry, op->key));
fs/afs/dir.c
1597
op->unlink.need_rehash = true;
fs/afs/dir.c
1601
op->file[1].vnode = vnode;
fs/afs/dir.c
1602
op->file[1].update_ctime = true;
fs/afs/dir.c
1603
op->file[1].op_unlinked = true;
fs/afs/dir.c
1604
op->dentry = dentry;
fs/afs/dir.c
1605
op->ops = &afs_unlink_operation;
fs/afs/dir.c
1606
afs_begin_vnode_operation(op);
fs/afs/dir.c
1607
afs_wait_for_operation(op);
fs/afs/dir.c
1612
if (afs_op_error(op) == 0 && (op->flags & AFS_OPERATION_DIR_CONFLICT)) {
fs/afs/dir.c
1613
op->file[1].update_ctime = false;
fs/afs/dir.c
1614
op->fetch_status.which = 1;
fs/afs/dir.c
1615
op->ops = &afs_fetch_status_operation;
fs/afs/dir.c
1616
afs_begin_vnode_operation(op);
fs/afs/dir.c
1617
afs_wait_for_operation(op);
fs/afs/dir.c
1621
ret = afs_put_operation(op);
fs/afs/dir.c
1641
struct afs_operation *op;
fs/afs/dir.c
1651
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1652
if (IS_ERR(op)) {
fs/afs/dir.c
1653
ret = PTR_ERR(op);
fs/afs/dir.c
1659
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1660
op->file[0].dv_delta = 1;
fs/afs/dir.c
1661
op->file[0].modification = true;
fs/afs/dir.c
1662
op->file[0].update_ctime = true;
fs/afs/dir.c
1664
op->dentry = dentry;
fs/afs/dir.c
1665
op->create.mode = S_IFREG | mode;
fs/afs/dir.c
1666
op->create.reason = afs_edit_dir_for_create;
fs/afs/dir.c
1667
op->mtime = current_time(dir);
fs/afs/dir.c
1668
op->ops = &afs_create_operation;
fs/afs/dir.c
1669
ret = afs_do_sync_operation(op);
fs/afs/dir.c
1679
static void afs_link_success(struct afs_operation *op)
fs/afs/dir.c
1681
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir.c
1682
struct afs_vnode_param *vp = &op->file[1];
fs/afs/dir.c
1684
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1685
op->ctime = dvp->scb.status.mtime_client;
fs/afs/dir.c
1686
afs_vnode_commit_status(op, dvp);
fs/afs/dir.c
1687
afs_vnode_commit_status(op, vp);
fs/afs/dir.c
1688
afs_update_dentry_version(op, dvp, op->dentry);
fs/afs/dir.c
1689
if (op->dentry_2->d_parent == op->dentry->d_parent)
fs/afs/dir.c
1690
afs_update_dentry_version(op, dvp, op->dentry_2);
fs/afs/dir.c
1692
d_instantiate(op->dentry, &vp->vnode->netfs.inode);
fs/afs/dir.c
1695
static void afs_link_put(struct afs_operation *op)
fs/afs/dir.c
1697
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1698
if (afs_op_error(op))
fs/afs/dir.c
1699
d_drop(op->dentry);
fs/afs/dir.c
1717
struct afs_operation *op;
fs/afs/dir.c
1730
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1731
if (IS_ERR(op)) {
fs/afs/dir.c
1732
ret = PTR_ERR(op);
fs/afs/dir.c
1738
ret = afs_validate(vnode, op->key);
fs/afs/dir.c
1742
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1743
afs_op_set_vnode(op, 1, vnode);
fs/afs/dir.c
1744
op->file[0].dv_delta = 1;
fs/afs/dir.c
1745
op->file[0].modification = true;
fs/afs/dir.c
1746
op->file[0].update_ctime = true;
fs/afs/dir.c
1747
op->file[1].update_ctime = true;
fs/afs/dir.c
1749
op->dentry = dentry;
fs/afs/dir.c
1750
op->dentry_2 = from;
fs/afs/dir.c
1751
op->ops = &afs_link_operation;
fs/afs/dir.c
1752
op->create.reason = afs_edit_dir_for_link;
fs/afs/dir.c
1753
ret = afs_do_sync_operation(op);
fs/afs/dir.c
1758
afs_put_operation(op);
fs/afs/dir.c
1781
struct afs_operation *op;
fs/afs/dir.c
1797
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
1798
if (IS_ERR(op)) {
fs/afs/dir.c
1799
ret = PTR_ERR(op);
fs/afs/dir.c
1805
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
1806
op->file[0].dv_delta = 1;
fs/afs/dir.c
1808
op->dentry = dentry;
fs/afs/dir.c
1809
op->ops = &afs_symlink_operation;
fs/afs/dir.c
1810
op->create.reason = afs_edit_dir_for_symlink;
fs/afs/dir.c
1811
op->create.symlink = content;
fs/afs/dir.c
1812
op->mtime = current_time(dir);
fs/afs/dir.c
1813
ret = afs_do_sync_operation(op);
fs/afs/dir.c
1823
static void afs_rename_success(struct afs_operation *op)
fs/afs/dir.c
1825
struct afs_vnode *vnode = op->more_files[0].vnode;
fs/afs/dir.c
1826
struct afs_vnode *new_vnode = op->more_files[1].vnode;
fs/afs/dir.c
1828
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1830
op->ctime = op->file[0].scb.status.mtime_client;
fs/afs/dir.c
1831
afs_check_dir_conflict(op, &op->file[1]);
fs/afs/dir.c
1832
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir.c
1833
if (op->file[1].vnode != op->file[0].vnode) {
fs/afs/dir.c
1834
op->ctime = op->file[1].scb.status.mtime_client;
fs/afs/dir.c
1835
afs_vnode_commit_status(op, &op->file[1]);
fs/afs/dir.c
1837
if (op->more_files[0].scb.have_status)
fs/afs/dir.c
1838
afs_vnode_commit_status(op, &op->more_files[0]);
fs/afs/dir.c
1839
if (op->more_files[1].scb.have_status)
fs/afs/dir.c
1840
afs_vnode_commit_status(op, &op->more_files[1]);
fs/afs/dir.c
1845
if (op->file[0].vnode != op->file[1].vnode) {
fs/afs/dir.c
1859
if ((op->rename.rename_flags & RENAME_EXCHANGE) &&
fs/afs/dir.c
1874
static void afs_rename_edit_dir(struct afs_operation *op)
fs/afs/dir.c
1877
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/dir.c
1878
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/dir.c
1881
struct afs_vnode *vnode = AFS_FS_I(d_inode(op->dentry));
fs/afs/dir.c
1882
struct dentry *old_dentry = op->dentry;
fs/afs/dir.c
1883
struct dentry *new_dentry = op->dentry_2;
fs/afs/dir.c
1886
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
1888
if (op->rename.rehash) {
fs/afs/dir.c
1889
d_rehash(op->rename.rehash);
fs/afs/dir.c
1890
op->rename.rehash = NULL;
fs/afs/dir.c
1910
if (!op->rename.new_negative)
fs/afs/dir.c
1937
afs_update_dentry_version(op, new_dvp, op->dentry);
fs/afs/dir.c
1938
afs_update_dentry_version(op, new_dvp, op->dentry_2);
fs/afs/dir.c
1948
static void afs_rename_exchange_edit_dir(struct afs_operation *op)
fs/afs/dir.c
1950
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/dir.c
1951
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/dir.c
1954
struct afs_vnode *old_vnode = op->more_files[0].vnode;
fs/afs/dir.c
1955
struct afs_vnode *new_vnode = op->more_files[1].vnode;
fs/afs/dir.c
1956
struct dentry *old_dentry = op->dentry;
fs/afs/dir.c
1957
struct dentry *new_dentry = op->dentry_2;
fs/afs/dir.c
1959
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
2001
afs_update_dentry_version(op, new_dvp, old_dentry);
fs/afs/dir.c
2002
afs_update_dentry_version(op, orig_dvp, new_dentry);
fs/afs/dir.c
2009
static void afs_rename_put(struct afs_operation *op)
fs/afs/dir.c
2011
_enter("op=%08x", op->debug_id);
fs/afs/dir.c
2012
if (op->rename.rehash)
fs/afs/dir.c
2013
d_rehash(op->rename.rehash);
fs/afs/dir.c
2014
dput(op->rename.tmp);
fs/afs/dir.c
2015
if (afs_op_error(op))
fs/afs/dir.c
2016
d_rehash(op->dentry);
fs/afs/dir.c
2060
struct afs_operation *op;
fs/afs/dir.c
2083
op = afs_alloc_operation(NULL, orig_dvnode->volume);
fs/afs/dir.c
2084
if (IS_ERR(op))
fs/afs/dir.c
2085
return PTR_ERR(op);
fs/afs/dir.c
2091
ret = afs_validate(vnode, op->key);
fs/afs/dir.c
2092
afs_op_set_error(op, ret);
fs/afs/dir.c
2097
op->more_files = kvzalloc_objs(struct afs_vnode_param, 2);
fs/afs/dir.c
2098
if (!op->more_files)
fs/afs/dir.c
2101
afs_op_set_vnode(op, 0, orig_dvnode);
fs/afs/dir.c
2102
afs_op_set_vnode(op, 1, new_dvnode); /* May be same as orig_dvnode */
fs/afs/dir.c
2103
op->file[0].dv_delta = 1;
fs/afs/dir.c
2104
op->file[1].dv_delta = 1;
fs/afs/dir.c
2105
op->file[0].modification = true;
fs/afs/dir.c
2106
op->file[1].modification = true;
fs/afs/dir.c
2107
op->file[0].update_ctime = true;
fs/afs/dir.c
2108
op->file[1].update_ctime = true;
fs/afs/dir.c
2109
op->more_files[0].vnode = vnode;
fs/afs/dir.c
2110
op->more_files[0].speculative = true;
fs/afs/dir.c
2111
op->more_files[1].vnode = new_vnode;
fs/afs/dir.c
2112
op->more_files[1].speculative = true;
fs/afs/dir.c
2113
op->nr_files = 4;
fs/afs/dir.c
2115
op->dentry = old_dentry;
fs/afs/dir.c
2116
op->dentry_2 = new_dentry;
fs/afs/dir.c
2117
op->rename.rename_flags = flags;
fs/afs/dir.c
2118
op->rename.new_negative = d_is_negative(new_dentry);
fs/afs/dir.c
2121
op->ops = &afs_rename_noreplace_operation;
fs/afs/dir.c
2123
op->ops = &afs_rename_exchange_operation;
fs/afs/dir.c
2129
op->ops = &afs_rename_operation;
fs/afs/dir.c
2142
op->rename.rehash = new_dentry;
fs/afs/dir.c
2147
op->rename.tmp = d_alloc(new_dentry->d_parent,
fs/afs/dir.c
2149
if (!op->rename.tmp) {
fs/afs/dir.c
2150
afs_op_nomem(op);
fs/afs/dir.c
2156
new_dentry, op->key);
fs/afs/dir.c
2158
afs_op_set_error(op, ret);
fs/afs/dir.c
2162
op->dentry_2 = op->rename.tmp;
fs/afs/dir.c
2163
op->rename.rehash = NULL;
fs/afs/dir.c
2164
op->rename.new_negative = true;
fs/afs/dir.c
2180
ret = afs_do_sync_operation(op);
fs/afs/dir.c
2190
ret = afs_put_operation(op);
fs/afs/dir.c
664
static void afs_do_lookup_success(struct afs_operation *op)
fs/afs/dir.c
674
for (i = 0; i < op->nr_files; i++) {
fs/afs/dir.c
677
vp = &op->file[0];
fs/afs/dir.c
680
op->call_abort_code = abort_code;
fs/afs/dir.c
681
afs_op_set_error(op, afs_abort_to_error(abort_code));
fs/afs/dir.c
682
op->cumul_error.abort_code = abort_code;
fs/afs/dir.c
687
vp = &op->file[1];
fs/afs/dir.c
691
vp = &op->more_files[i - 2];
fs/afs/dir.c
696
trace_afs_bulkstat_error(op, &vp->fid, i, vp->scb.status.abort_code);
fs/afs/dir.c
703
afs_vnode_commit_status(op, vp);
fs/afs/dir.c
705
inode = afs_iget(op, vp);
fs/afs/dir.c
708
afs_cache_permit(vnode, op->key,
fs/afs/dir.c
710
op->cb_v_break,
fs/afs/dir.c
778
struct afs_operation *op;
fs/afs/dir.c
823
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir.c
824
if (IS_ERR(op)) {
fs/afs/dir.c
825
ret = PTR_ERR(op);
fs/afs/dir.c
829
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir.c
830
afs_op_set_fid(op, 1, &cookie->fids[1]);
fs/afs/dir.c
832
op->nr_files = cookie->nr_fids;
fs/afs/dir.c
833
_debug("nr_files %u", op->nr_files);
fs/afs/dir.c
836
if (op->nr_files > 2) {
fs/afs/dir.c
837
op->more_files = kvzalloc_objs(struct afs_vnode_param,
fs/afs/dir.c
838
op->nr_files - 2);
fs/afs/dir.c
839
if (!op->more_files) {
fs/afs/dir.c
840
afs_op_nomem(op);
fs/afs/dir.c
844
for (i = 2; i < op->nr_files; i++) {
fs/afs/dir.c
845
vp = &op->more_files[i - 2];
fs/afs/dir.c
868
afs_op_set_error(op, -ENOTSUPP);
fs/afs/dir.c
870
op->ops = &afs_inline_bulk_status_operation;
fs/afs/dir.c
871
afs_begin_vnode_operation(op);
fs/afs/dir.c
872
afs_wait_for_operation(op);
fs/afs/dir.c
875
if (afs_op_error(op) == -ENOTSUPP) {
fs/afs/dir.c
880
op->fetch_status.which = 1;
fs/afs/dir.c
881
op->ops = &afs_lookup_fetch_status_operation;
fs/afs/dir.c
882
afs_begin_vnode_operation(op);
fs/afs/dir.c
883
afs_wait_for_operation(op);
fs/afs/dir.c
887
if (!afs_op_error(op)) {
fs/afs/dir.c
888
if (op->file[1].scb.status.abort_code) {
fs/afs/dir.c
889
afs_op_accumulate_error(op, -ECONNABORTED,
fs/afs/dir.c
890
op->file[1].scb.status.abort_code);
fs/afs/dir.c
892
inode = &op->file[1].vnode->netfs.inode;
fs/afs/dir.c
893
op->file[1].vnode = NULL;
fs/afs/dir.c
897
if (op->file[0].scb.have_status)
fs/afs/dir.c
898
dentry->d_fsdata = (void *)(unsigned long)op->file[0].scb.status.data_version;
fs/afs/dir.c
900
dentry->d_fsdata = (void *)(unsigned long)op->file[0].dv_before;
fs/afs/dir.c
901
ret = afs_put_operation(op);
fs/afs/dir_silly.c
15
static void afs_silly_rename_success(struct afs_operation *op)
fs/afs/dir_silly.c
167
static void afs_silly_unlink_success(struct afs_operation *op)
fs/afs/dir_silly.c
169
_enter("op=%08x", op->debug_id);
fs/afs/dir_silly.c
17
_enter("op=%08x", op->debug_id);
fs/afs/dir_silly.c
170
afs_check_dir_conflict(op, &op->file[0]);
fs/afs/dir_silly.c
171
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir_silly.c
172
afs_vnode_commit_status(op, &op->file[1]);
fs/afs/dir_silly.c
173
afs_update_dentry_version(op, &op->file[0], op->dentry);
fs/afs/dir_silly.c
176
static void afs_silly_unlink_edit_dir(struct afs_operation *op)
fs/afs/dir_silly.c
178
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir_silly.c
181
_enter("op=%08x", op->debug_id);
fs/afs/dir_silly.c
185
afs_edit_dir_remove(dvnode, &op->dentry->d_name,
fs/afs/dir_silly.c
19
afs_check_dir_conflict(op, &op->file[0]);
fs/afs/dir_silly.c
20
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/dir_silly.c
204
struct afs_operation *op;
fs/afs/dir_silly.c
208
op = afs_alloc_operation(NULL, dvnode->volume);
fs/afs/dir_silly.c
209
if (IS_ERR(op))
fs/afs/dir_silly.c
210
return PTR_ERR(op);
fs/afs/dir_silly.c
212
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir_silly.c
213
afs_op_set_vnode(op, 1, vnode);
fs/afs/dir_silly.c
214
op->file[0].dv_delta = 1;
fs/afs/dir_silly.c
215
op->file[0].modification = true;
fs/afs/dir_silly.c
216
op->file[0].update_ctime = true;
fs/afs/dir_silly.c
217
op->file[1].op_unlinked = true;
fs/afs/dir_silly.c
218
op->file[1].update_ctime = true;
fs/afs/dir_silly.c
220
op->dentry = dentry;
fs/afs/dir_silly.c
221
op->ops = &afs_silly_unlink_operation;
fs/afs/dir_silly.c
224
afs_begin_vnode_operation(op);
fs/afs/dir_silly.c
225
afs_wait_for_operation(op);
fs/afs/dir_silly.c
23
static void afs_silly_rename_edit_dir(struct afs_operation *op)
fs/afs/dir_silly.c
230
if (op->cumul_error.error == 0 && (op->flags & AFS_OPERATION_DIR_CONFLICT)) {
fs/afs/dir_silly.c
231
op->file[1].update_ctime = false;
fs/afs/dir_silly.c
232
op->fetch_status.which = 1;
fs/afs/dir_silly.c
233
op->ops = &afs_fetch_status_operation;
fs/afs/dir_silly.c
234
afs_begin_vnode_operation(op);
fs/afs/dir_silly.c
235
afs_wait_for_operation(op);
fs/afs/dir_silly.c
238
return afs_put_operation(op);
fs/afs/dir_silly.c
25
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/dir_silly.c
27
struct afs_vnode *vnode = AFS_FS_I(d_inode(op->dentry));
fs/afs/dir_silly.c
28
struct dentry *old = op->dentry;
fs/afs/dir_silly.c
29
struct dentry *new = op->dentry_2;
fs/afs/dir_silly.c
34
if (dvnode->silly_key != op->key) {
fs/afs/dir_silly.c
36
dvnode->silly_key = key_get(op->key);
fs/afs/dir_silly.c
64
struct afs_operation *op;
fs/afs/dir_silly.c
68
op = afs_alloc_operation(key, dvnode->volume);
fs/afs/dir_silly.c
69
if (IS_ERR(op))
fs/afs/dir_silly.c
70
return PTR_ERR(op);
fs/afs/dir_silly.c
72
op->more_files = kvzalloc_objs(struct afs_vnode_param, 2);
fs/afs/dir_silly.c
73
if (!op->more_files) {
fs/afs/dir_silly.c
74
afs_put_operation(op);
fs/afs/dir_silly.c
78
afs_op_set_vnode(op, 0, dvnode);
fs/afs/dir_silly.c
79
afs_op_set_vnode(op, 1, dvnode);
fs/afs/dir_silly.c
80
op->file[0].dv_delta = 1;
fs/afs/dir_silly.c
81
op->file[1].dv_delta = 1;
fs/afs/dir_silly.c
82
op->file[0].modification = true;
fs/afs/dir_silly.c
83
op->file[1].modification = true;
fs/afs/dir_silly.c
84
op->file[0].update_ctime = true;
fs/afs/dir_silly.c
85
op->file[1].update_ctime = true;
fs/afs/dir_silly.c
86
op->more_files[0].vnode = AFS_FS_I(d_inode(old));
fs/afs/dir_silly.c
87
op->more_files[0].speculative = true;
fs/afs/dir_silly.c
88
op->more_files[1].vnode = AFS_FS_I(d_inode(new));
fs/afs/dir_silly.c
89
op->more_files[1].speculative = true;
fs/afs/dir_silly.c
90
op->nr_files = 4;
fs/afs/dir_silly.c
92
op->dentry = old;
fs/afs/dir_silly.c
93
op->dentry_2 = new;
fs/afs/dir_silly.c
94
op->ops = &afs_silly_rename_operation;
fs/afs/dir_silly.c
97
return afs_do_sync_operation(op);
fs/afs/file.c
203
static void afs_fetch_data_notify(struct afs_operation *op)
fs/afs/file.c
205
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/file.c
207
subreq->error = afs_op_error(op);
fs/afs/file.c
211
static void afs_fetch_data_success(struct afs_operation *op)
fs/afs/file.c
213
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/file.c
215
_enter("op=%08x", op->debug_id);
fs/afs/file.c
216
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/file.c
218
atomic_long_add(op->fetch.subreq->transferred, &op->net->n_fetch_bytes);
fs/afs/file.c
219
afs_fetch_data_notify(op);
fs/afs/file.c
222
static void afs_fetch_data_aborted(struct afs_operation *op)
fs/afs/file.c
224
afs_check_for_remote_deletion(op);
fs/afs/file.c
225
afs_fetch_data_notify(op);
fs/afs/file.c
236
static void afs_issue_read_call(struct afs_operation *op)
fs/afs/file.c
238
op->call_responded = false;
fs/afs/file.c
239
op->call_error = 0;
fs/afs/file.c
240
op->call_abort_code = 0;
fs/afs/file.c
241
if (test_bit(AFS_SERVER_FL_IS_YFS, &op->server->flags))
fs/afs/file.c
242
yfs_fs_fetch_data(op);
fs/afs/file.c
244
afs_fs_fetch_data(op);
fs/afs/file.c
247
static void afs_end_read(struct afs_operation *op)
fs/afs/file.c
249
if (op->call_responded && op->server)
fs/afs/file.c
250
set_bit(AFS_SERVER_FL_RESPONDING, &op->server->flags);
fs/afs/file.c
252
if (!afs_op_error(op))
fs/afs/file.c
253
afs_fetch_data_success(op);
fs/afs/file.c
254
else if (op->cumul_error.aborted)
fs/afs/file.c
255
afs_fetch_data_aborted(op);
fs/afs/file.c
257
afs_fetch_data_notify(op);
fs/afs/file.c
259
afs_end_vnode_operation(op);
fs/afs/file.c
260
afs_put_operation(op);
fs/afs/file.c
269
struct afs_operation *op = call->op;
fs/afs/file.c
277
trace_afs_read_recv(op, call);
fs/afs/file.c
286
netfs_read_subreq_progress(op->fetch.subreq);
fs/afs/file.c
293
op->call_abort_code = call->abort_code;
fs/afs/file.c
294
op->call_error = call->error;
fs/afs/file.c
295
op->call_responded = call->responded;
fs/afs/file.c
296
op->call = NULL;
fs/afs/file.c
297
call->op = NULL;
fs/afs/file.c
303
if (afs_select_fileserver(op)) {
fs/afs/file.c
304
afs_issue_read_call(op);
fs/afs/file.c
308
afs_end_read(op);
fs/afs/file.c
334
struct afs_operation *op;
fs/afs/file.c
345
op = afs_alloc_operation(key, vnode->volume);
fs/afs/file.c
346
if (IS_ERR(op)) {
fs/afs/file.c
347
subreq->error = PTR_ERR(op);
fs/afs/file.c
352
afs_op_set_vnode(op, 0, vnode);
fs/afs/file.c
354
op->fetch.subreq = subreq;
fs/afs/file.c
355
op->ops = &afs_fetch_data_operation;
fs/afs/file.c
361
op->flags |= AFS_OPERATION_ASYNC;
fs/afs/file.c
363
if (!afs_begin_vnode_operation(op)) {
fs/afs/file.c
364
subreq->error = afs_put_operation(op);
fs/afs/file.c
369
if (!afs_select_fileserver(op)) {
fs/afs/file.c
370
afs_end_read(op);
fs/afs/file.c
374
afs_issue_read_call(op);
fs/afs/file.c
376
afs_do_sync_operation(op);
fs/afs/flock.c
175
static void afs_lock_success(struct afs_operation *op)
fs/afs/flock.c
177
_enter("op=%08x", op->debug_id);
fs/afs/flock.c
178
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/flock.c
194
struct afs_operation *op;
fs/afs/flock.c
203
op = afs_alloc_operation(key, vnode->volume);
fs/afs/flock.c
204
if (IS_ERR(op))
fs/afs/flock.c
205
return PTR_ERR(op);
fs/afs/flock.c
207
afs_op_set_vnode(op, 0, vnode);
fs/afs/flock.c
209
op->lock.type = type;
fs/afs/flock.c
210
op->ops = &afs_set_lock_operation;
fs/afs/flock.c
211
return afs_do_sync_operation(op);
fs/afs/flock.c
225
struct afs_operation *op;
fs/afs/flock.c
234
op = afs_alloc_operation(key, vnode->volume);
fs/afs/flock.c
235
if (IS_ERR(op))
fs/afs/flock.c
236
return PTR_ERR(op);
fs/afs/flock.c
238
afs_op_set_vnode(op, 0, vnode);
fs/afs/flock.c
240
op->flags |= AFS_OPERATION_UNINTR;
fs/afs/flock.c
241
op->ops = &afs_extend_lock_operation;
fs/afs/flock.c
242
return afs_do_sync_operation(op);
fs/afs/flock.c
256
struct afs_operation *op;
fs/afs/flock.c
265
op = afs_alloc_operation(key, vnode->volume);
fs/afs/flock.c
266
if (IS_ERR(op))
fs/afs/flock.c
267
return PTR_ERR(op);
fs/afs/flock.c
269
afs_op_set_vnode(op, 0, vnode);
fs/afs/flock.c
271
op->flags |= AFS_OPERATION_UNINTR;
fs/afs/flock.c
272
op->ops = &afs_release_lock_operation;
fs/afs/flock.c
273
return afs_do_sync_operation(op);
fs/afs/flock.c
73
struct afs_operation *op = call->op;
fs/afs/flock.c
74
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/flock.c
767
enum afs_flock_operation op;
fs/afs/flock.c
787
case 0: op = afs_flock_op_return_ok; break;
fs/afs/flock.c
788
case -EAGAIN: op = afs_flock_op_return_eagain; break;
fs/afs/flock.c
789
case -EDEADLK: op = afs_flock_op_return_edeadlk; break;
fs/afs/flock.c
790
default: op = afs_flock_op_return_error; break;
fs/afs/flock.c
792
trace_afs_flock_op(vnode, fl, op);
fs/afs/flock.c
802
enum afs_flock_operation op;
fs/afs/flock.c
829
case 0: op = afs_flock_op_return_ok; break;
fs/afs/flock.c
830
case -EAGAIN: op = afs_flock_op_return_eagain; break;
fs/afs/flock.c
831
case -EDEADLK: op = afs_flock_op_return_edeadlk; break;
fs/afs/flock.c
832
default: op = afs_flock_op_return_error; break;
fs/afs/flock.c
834
trace_afs_flock_op(vnode, fl, op);
fs/afs/fs_operation.c
154
static bool afs_get_io_locks(struct afs_operation *op)
fs/afs/fs_operation.c
156
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/fs_operation.c
157
struct afs_vnode *vnode2 = op->file[1].vnode;
fs/afs/fs_operation.c
161
if (op->flags & AFS_OPERATION_UNINTR) {
fs/afs/fs_operation.c
163
op->flags |= AFS_OPERATION_LOCK_0;
fs/afs/fs_operation.c
168
if (!vnode2 || !op->file[1].need_io_lock || vnode == vnode2)
fs/afs/fs_operation.c
175
afs_op_set_error(op, -ERESTARTSYS);
fs/afs/fs_operation.c
176
op->flags |= AFS_OPERATION_STOP;
fs/afs/fs_operation.c
180
op->flags |= AFS_OPERATION_LOCK_0;
fs/afs/fs_operation.c
184
afs_op_set_error(op, -ERESTARTSYS);
fs/afs/fs_operation.c
185
op->flags |= AFS_OPERATION_STOP;
fs/afs/fs_operation.c
187
op->flags &= ~AFS_OPERATION_LOCK_0;
fs/afs/fs_operation.c
191
op->flags |= AFS_OPERATION_LOCK_1;
fs/afs/fs_operation.c
198
static void afs_drop_io_locks(struct afs_operation *op)
fs/afs/fs_operation.c
20
struct afs_operation *op;
fs/afs/fs_operation.c
200
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/fs_operation.c
201
struct afs_vnode *vnode2 = op->file[1].vnode;
fs/afs/fs_operation.c
205
if (op->flags & AFS_OPERATION_LOCK_1)
fs/afs/fs_operation.c
207
if (op->flags & AFS_OPERATION_LOCK_0)
fs/afs/fs_operation.c
211
static void afs_prepare_vnode(struct afs_operation *op, struct afs_vnode_param *vp,
fs/afs/fs_operation.c
221
op->flags |= AFS_OPERATION_CUR_ONLY;
fs/afs/fs_operation.c
237
bool afs_begin_vnode_operation(struct afs_operation *op)
fs/afs/fs_operation.c
239
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/fs_operation.c
24
op = kzalloc_obj(*op);
fs/afs/fs_operation.c
245
if (op->file[0].need_io_lock)
fs/afs/fs_operation.c
246
if (!afs_get_io_locks(op))
fs/afs/fs_operation.c
249
afs_prepare_vnode(op, &op->file[0], 0);
fs/afs/fs_operation.c
25
if (!op)
fs/afs/fs_operation.c
250
afs_prepare_vnode(op, &op->file[1], 1);
fs/afs/fs_operation.c
251
op->cb_v_break = atomic_read(&op->volume->cb_v_break);
fs/afs/fs_operation.c
259
void afs_end_vnode_operation(struct afs_operation *op)
fs/afs/fs_operation.c
263
switch (afs_op_error(op)) {
fs/afs/fs_operation.c
268
afs_dump_edestaddrreq(op);
fs/afs/fs_operation.c
272
afs_drop_io_locks(op);
fs/afs/fs_operation.c
278
void afs_wait_for_operation(struct afs_operation *op)
fs/afs/fs_operation.c
282
while (afs_select_fileserver(op)) {
fs/afs/fs_operation.c
283
op->call_responded = false;
fs/afs/fs_operation.c
284
op->call_error = 0;
fs/afs/fs_operation.c
285
op->call_abort_code = 0;
fs/afs/fs_operation.c
286
if (test_bit(AFS_SERVER_FL_IS_YFS, &op->server->flags) &&
fs/afs/fs_operation.c
287
op->ops->issue_yfs_rpc)
fs/afs/fs_operation.c
288
op->ops->issue_yfs_rpc(op);
fs/afs/fs_operation.c
289
else if (op->ops->issue_afs_rpc)
fs/afs/fs_operation.c
290
op->ops->issue_afs_rpc(op);
fs/afs/fs_operation.c
292
op->call_error = -ENOTSUPP;
fs/afs/fs_operation.c
294
if (op->call) {
fs/afs/fs_operation.c
295
afs_wait_for_call_to_complete(op->call);
fs/afs/fs_operation.c
296
op->call_abort_code = op->call->abort_code;
fs/afs/fs_operation.c
297
op->call_error = op->call->error;
fs/afs/fs_operation.c
298
op->call_responded = op->call->responded;
fs/afs/fs_operation.c
299
afs_put_call(op->call);
fs/afs/fs_operation.c
303
if (op->call_responded && op->server)
fs/afs/fs_operation.c
304
set_bit(AFS_SERVER_FL_RESPONDING, &op->server->flags);
fs/afs/fs_operation.c
306
if (!afs_op_error(op)) {
fs/afs/fs_operation.c
308
op->ops->success(op);
fs/afs/fs_operation.c
309
} else if (op->cumul_error.aborted) {
fs/afs/fs_operation.c
31
kfree(op);
fs/afs/fs_operation.c
310
if (op->ops->aborted)
fs/afs/fs_operation.c
311
op->ops->aborted(op);
fs/afs/fs_operation.c
313
if (op->ops->failed)
fs/afs/fs_operation.c
314
op->ops->failed(op);
fs/afs/fs_operation.c
317
afs_end_vnode_operation(op);
fs/afs/fs_operation.c
319
if (!afs_op_error(op) && op->ops->edit_dir) {
fs/afs/fs_operation.c
321
op->ops->edit_dir(op);
fs/afs/fs_operation.c
329
int afs_put_operation(struct afs_operation *op)
fs/afs/fs_operation.c
332
int i, ret = afs_op_error(op);
fs/afs/fs_operation.c
334
_enter("op=%08x,%d", op->debug_id, ret);
fs/afs/fs_operation.c
336
if (op->ops && op->ops->put)
fs/afs/fs_operation.c
337
op->ops->put(op);
fs/afs/fs_operation.c
338
if (op->file[0].modification)
fs/afs/fs_operation.c
339
clear_bit(AFS_VNODE_MODIFYING, &op->file[0].vnode->flags);
fs/afs/fs_operation.c
340
if (op->file[1].modification && op->file[1].vnode != op->file[0].vnode)
fs/afs/fs_operation.c
341
clear_bit(AFS_VNODE_MODIFYING, &op->file[1].vnode->flags);
fs/afs/fs_operation.c
342
if (op->file[0].put_vnode)
fs/afs/fs_operation.c
343
iput(&op->file[0].vnode->netfs.inode);
fs/afs/fs_operation.c
344
if (op->file[1].put_vnode)
fs/afs/fs_operation.c
345
iput(&op->file[1].vnode->netfs.inode);
fs/afs/fs_operation.c
347
if (op->more_files) {
fs/afs/fs_operation.c
348
for (i = 0; i < op->nr_files - 2; i++)
fs/afs/fs_operation.c
349
if (op->more_files[i].put_vnode)
fs/afs/fs_operation.c
350
iput(&op->more_files[i].vnode->netfs.inode);
fs/afs/fs_operation.c
351
kfree(op->more_files);
fs/afs/fs_operation.c
354
if (op->estate) {
fs/afs/fs_operation.c
355
alist = op->estate->addresses;
fs/afs/fs_operation.c
357
if (op->call_responded &&
fs/afs/fs_operation.c
358
op->addr_index != alist->preferred &&
fs/afs/fs_operation.c
359
test_bit(alist->preferred, &op->addr_tried))
fs/afs/fs_operation.c
360
WRITE_ONCE(alist->preferred, op->addr_index);
fs/afs/fs_operation.c
364
afs_clear_server_states(op);
fs/afs/fs_operation.c
365
afs_put_serverlist(op->net, op->server_list);
fs/afs/fs_operation.c
366
afs_put_volume(op->volume, afs_volume_trace_put_put_op);
fs/afs/fs_operation.c
367
key_put(op->key);
fs/afs/fs_operation.c
368
kfree(op);
fs/afs/fs_operation.c
372
int afs_do_sync_operation(struct afs_operation *op)
fs/afs/fs_operation.c
374
afs_begin_vnode_operation(op);
fs/afs/fs_operation.c
375
afs_wait_for_operation(op);
fs/afs/fs_operation.c
376
return afs_put_operation(op);
fs/afs/fs_operation.c
38
op->key = key;
fs/afs/fs_operation.c
39
op->volume = afs_get_volume(volume, afs_volume_trace_get_new_op);
fs/afs/fs_operation.c
40
op->net = volume->cell->net;
fs/afs/fs_operation.c
41
op->cb_v_break = atomic_read(&volume->cb_v_break);
fs/afs/fs_operation.c
42
op->pre_volsync.creation = volume->creation_time;
fs/afs/fs_operation.c
43
op->pre_volsync.update = volume->update_time;
fs/afs/fs_operation.c
44
op->debug_id = atomic_inc_return(&afs_operation_debug_counter);
fs/afs/fs_operation.c
45
op->nr_iterations = -1;
fs/afs/fs_operation.c
46
afs_op_set_error(op, -EDESTADDRREQ);
fs/afs/fs_operation.c
48
_leave(" = [op=%08x]", op->debug_id);
fs/afs/fs_operation.c
49
return op;
fs/afs/fs_probe.c
314
int afs_wait_for_fs_probes(struct afs_operation *op, struct afs_server_state *states, bool intr)
fs/afs/fs_probe.c
317
struct afs_server_list *slist = op->server_list;
fs/afs/fsclient.c
1025
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1033
struct afs_operation *op = call->op;
fs/afs/fsclient.c
1034
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1047
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
1058
.op = afs_FS_StoreData,
fs/afs/fsclient.c
1065
.op = afs_FS_StoreData64,
fs/afs/fsclient.c
1073
static void afs_fs_store_data64(struct afs_operation *op)
fs/afs/fsclient.c
1075
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1080
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
1082
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreData64,
fs/afs/fsclient.c
1086
return afs_op_nomem(op);
fs/afs/fsclient.c
1088
call->write_iter = op->store.write_iter;
fs/afs/fsclient.c
1098
*bp++ = htonl(op->mtime.tv_sec); /* mtime */
fs/afs/fsclient.c
1104
*bp++ = htonl(upper_32_bits(op->store.pos));
fs/afs/fsclient.c
1105
*bp++ = htonl(lower_32_bits(op->store.pos));
fs/afs/fsclient.c
1106
*bp++ = htonl(upper_32_bits(op->store.size));
fs/afs/fsclient.c
1107
*bp++ = htonl(lower_32_bits(op->store.size));
fs/afs/fsclient.c
1108
*bp++ = htonl(upper_32_bits(op->store.i_size));
fs/afs/fsclient.c
1109
*bp++ = htonl(lower_32_bits(op->store.i_size));
fs/afs/fsclient.c
1113
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1119
void afs_fs_store_data(struct afs_operation *op)
fs/afs/fsclient.c
1121
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1126
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
1129
(unsigned long long)op->store.size,
fs/afs/fsclient.c
1130
(unsigned long long)op->store.pos,
fs/afs/fsclient.c
1131
(unsigned long long)op->store.i_size);
fs/afs/fsclient.c
1133
if (test_bit(AFS_SERVER_FL_HAS_FS64, &op->server->flags))
fs/afs/fsclient.c
1134
return afs_fs_store_data64(op);
fs/afs/fsclient.c
1136
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreData,
fs/afs/fsclient.c
1140
return afs_op_nomem(op);
fs/afs/fsclient.c
1142
call->write_iter = op->store.write_iter;
fs/afs/fsclient.c
1152
*bp++ = htonl(op->mtime.tv_sec); /* mtime */
fs/afs/fsclient.c
1158
*bp++ = htonl(lower_32_bits(op->store.pos));
fs/afs/fsclient.c
1159
*bp++ = htonl(lower_32_bits(op->store.size));
fs/afs/fsclient.c
1160
*bp++ = htonl(lower_32_bits(op->store.i_size));
fs/afs/fsclient.c
1164
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1172
.op = afs_FS_StoreStatus,
fs/afs/fsclient.c
1179
.op = afs_FS_StoreData,
fs/afs/fsclient.c
1186
.op = afs_FS_StoreData64,
fs/afs/fsclient.c
1195
static void afs_fs_setattr_size64(struct afs_operation *op)
fs/afs/fsclient.c
1197
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1199
struct iattr *attr = op->setattr.attr;
fs/afs/fsclient.c
1203
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
1207
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreData64_as_Status,
fs/afs/fsclient.c
1211
return afs_op_nomem(op);
fs/afs/fsclient.c
1231
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1238
static void afs_fs_setattr_size(struct afs_operation *op)
fs/afs/fsclient.c
1240
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1242
struct iattr *attr = op->setattr.attr;
fs/afs/fsclient.c
1246
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
1249
if (test_bit(AFS_SERVER_FL_HAS_FS64, &op->server->flags))
fs/afs/fsclient.c
1250
return afs_fs_setattr_size64(op);
fs/afs/fsclient.c
1252
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreData_as_Status,
fs/afs/fsclient.c
1256
return afs_op_nomem(op);
fs/afs/fsclient.c
1273
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1280
void afs_fs_setattr(struct afs_operation *op)
fs/afs/fsclient.c
1282
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1284
struct iattr *attr = op->setattr.attr;
fs/afs/fsclient.c
1288
return afs_fs_setattr_size(op);
fs/afs/fsclient.c
1291
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
1293
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreStatus,
fs/afs/fsclient.c
1297
return afs_op_nomem(op);
fs/afs/fsclient.c
1306
xdr_encode_AFS_StoreStatus(&bp, op->setattr.attr);
fs/afs/fsclient.c
1310
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1318
struct afs_operation *op = call->op;
fs/afs/fsclient.c
1340
xdr_decode_AFSFetchVolumeStatus(&bp, &op->volstatus.vs);
fs/afs/fsclient.c
1446
.op = afs_FS_GetVolumeStatus,
fs/afs/fsclient.c
1454
void afs_fs_get_volume_status(struct afs_operation *op)
fs/afs/fsclient.c
1456
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1462
call = afs_alloc_flat_call(op->net, &afs_RXFSGetVolumeStatus, 2 * 4,
fs/afs/fsclient.c
1465
return afs_op_nomem(op);
fs/afs/fsclient.c
1474
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1482
struct afs_operation *op = call->op;
fs/afs/fsclient.c
1494
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
1505
.op = afs_FS_SetLock,
fs/afs/fsclient.c
1516
.op = afs_FS_ExtendLock,
fs/afs/fsclient.c
1527
.op = afs_FS_ReleaseLock,
fs/afs/fsclient.c
1535
void afs_fs_set_lock(struct afs_operation *op)
fs/afs/fsclient.c
1537
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1543
call = afs_alloc_flat_call(op->net, &afs_RXFSSetLock, 5 * 4, 6 * 4);
fs/afs/fsclient.c
1545
return afs_op_nomem(op);
fs/afs/fsclient.c
1553
*bp++ = htonl(op->lock.type);
fs/afs/fsclient.c
1556
trace_afs_make_fs_calli(call, &vp->fid, op->lock.type);
fs/afs/fsclient.c
1557
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1563
void afs_fs_extend_lock(struct afs_operation *op)
fs/afs/fsclient.c
1565
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1571
call = afs_alloc_flat_call(op->net, &afs_RXFSExtendLock, 4 * 4, 6 * 4);
fs/afs/fsclient.c
1573
return afs_op_nomem(op);
fs/afs/fsclient.c
1584
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1590
void afs_fs_release_lock(struct afs_operation *op)
fs/afs/fsclient.c
1592
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
1598
call = afs_alloc_flat_call(op->net, &afs_RXFSReleaseLock, 4 * 4, 6 * 4);
fs/afs/fsclient.c
1600
return afs_op_nomem(op);
fs/afs/fsclient.c
1611
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1627
.op = afs_FS_GiveUpAllCallBacks,
fs/afs/fsclient.c
1736
.op = afs_FS_GetCapabilities,
fs/afs/fsclient.c
1787
struct afs_operation *op = call->op;
fs/afs/fsclient.c
1809
_debug("status count: %u/%u", tmp, op->nr_files);
fs/afs/fsclient.c
1810
if (tmp != op->nr_files)
fs/afs/fsclient.c
1827
scb = &op->file[0].scb;
fs/afs/fsclient.c
1830
scb = &op->file[1].scb;
fs/afs/fsclient.c
1833
scb = &op->more_files[call->count - 2].scb;
fs/afs/fsclient.c
1841
if (call->count < op->nr_files)
fs/afs/fsclient.c
1858
if (tmp != op->nr_files)
fs/afs/fsclient.c
1875
scb = &op->file[0].scb;
fs/afs/fsclient.c
1878
scb = &op->file[1].scb;
fs/afs/fsclient.c
1881
scb = &op->more_files[call->count - 2].scb;
fs/afs/fsclient.c
1888
if (call->count < op->nr_files)
fs/afs/fsclient.c
1922
if (call->op)
fs/afs/fsclient.c
1923
set_bit(AFS_VOLUME_MAYBE_NO_IBULK, &call->op->volume->flags);
fs/afs/fsclient.c
1932
.op = afs_FS_InlineBulkStatus,
fs/afs/fsclient.c
1941
void afs_fs_inline_bulk_status(struct afs_operation *op)
fs/afs/fsclient.c
1943
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
1944
struct afs_vnode_param *vp = &op->file[1];
fs/afs/fsclient.c
1949
if (test_bit(AFS_SERVER_FL_NO_IBULK, &op->server->flags)) {
fs/afs/fsclient.c
1950
afs_op_set_error(op, -ENOTSUPP);
fs/afs/fsclient.c
1955
key_serial(op->key), vp->fid.vid, vp->fid.vnode, op->nr_files);
fs/afs/fsclient.c
1957
call = afs_alloc_flat_call(op->net, &afs_RXFSInlineBulkStatus,
fs/afs/fsclient.c
1958
(2 + op->nr_files * 3) * 4,
fs/afs/fsclient.c
1961
return afs_op_nomem(op);
fs/afs/fsclient.c
1966
*bp++ = htonl(op->nr_files);
fs/afs/fsclient.c
1973
for (i = 0; i < op->nr_files - 2; i++) {
fs/afs/fsclient.c
1974
*bp++ = htonl(op->more_files[i].fid.vid);
fs/afs/fsclient.c
1975
*bp++ = htonl(op->more_files[i].fid.vnode);
fs/afs/fsclient.c
1976
*bp++ = htonl(op->more_files[i].fid.unique);
fs/afs/fsclient.c
1981
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
1989
struct afs_operation *op = call->op;
fs/afs/fsclient.c
1990
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
2016
op->acl = acl;
fs/afs/fsclient.c
2040
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
2058
.op = afs_FS_FetchACL,
fs/afs/fsclient.c
2065
void afs_fs_fetch_acl(struct afs_operation *op)
fs/afs/fsclient.c
2067
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
2072
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
2074
call = afs_alloc_flat_call(op->net, &afs_RXFSFetchACL, 16, (21 + 6) * 4);
fs/afs/fsclient.c
2076
return afs_op_nomem(op);
fs/afs/fsclient.c
2087
afs_make_op_call(op, call, GFP_KERNEL);
fs/afs/fsclient.c
2095
.op = afs_FS_StoreACL,
fs/afs/fsclient.c
2103
void afs_fs_store_acl(struct afs_operation *op)
fs/afs/fsclient.c
2105
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
2107
const struct afs_acl *acl = op->acl;
fs/afs/fsclient.c
2112
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
2115
call = afs_alloc_flat_call(op->net, &afs_RXFSStoreACL,
fs/afs/fsclient.c
2118
return afs_op_nomem(op);
fs/afs/fsclient.c
2133
afs_make_op_call(op, call, GFP_KERNEL);
fs/afs/fsclient.c
240
struct afs_operation *op = call->op;
fs/afs/fsclient.c
241
struct afs_vnode_param *vp = &op->file[op->fetch_status.which];
fs/afs/fsclient.c
253
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
264
.op = afs_FS_FetchStatus,
fs/afs/fsclient.c
272
void afs_fs_fetch_status(struct afs_operation *op)
fs/afs/fsclient.c
274
struct afs_vnode_param *vp = &op->file[op->fetch_status.which];
fs/afs/fsclient.c
279
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/fsclient.c
281
call = afs_alloc_flat_call(op->net, &afs_RXFSFetchStatus,
fs/afs/fsclient.c
284
return afs_op_nomem(op);
fs/afs/fsclient.c
295
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
303
struct afs_operation *op = call->op;
fs/afs/fsclient.c
304
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/fsclient.c
305
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
389
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
410
.op = afs_FS_FetchData,
fs/afs/fsclient.c
419
.op = afs_FS_FetchData64,
fs/afs/fsclient.c
429
static void afs_fs_fetch_data64(struct afs_operation *op)
fs/afs/fsclient.c
431
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/fsclient.c
432
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
438
call = afs_alloc_flat_call(op->net, &afs_RXFSFetchData64, 32, (21 + 3 + 6) * 4);
fs/afs/fsclient.c
440
return afs_op_nomem(op);
fs/afs/fsclient.c
442
if (op->flags & AFS_OPERATION_ASYNC)
fs/afs/fsclient.c
458
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
464
void afs_fs_fetch_data(struct afs_operation *op)
fs/afs/fsclient.c
466
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/fsclient.c
467
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
471
if (test_bit(AFS_SERVER_FL_HAS_FS64, &op->server->flags))
fs/afs/fsclient.c
472
return afs_fs_fetch_data64(op);
fs/afs/fsclient.c
476
call = afs_alloc_flat_call(op->net, &afs_RXFSFetchData, 24, (21 + 3 + 6) * 4);
fs/afs/fsclient.c
478
return afs_op_nomem(op);
fs/afs/fsclient.c
491
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
499
struct afs_operation *op = call->op;
fs/afs/fsclient.c
500
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
501
struct afs_vnode_param *vp = &op->file[1];
fs/afs/fsclient.c
511
xdr_decode_AFSFid(&bp, &op->file[1].fid);
fs/afs/fsclient.c
515
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
526
.op = afs_FS_CreateFile,
fs/afs/fsclient.c
534
void afs_fs_create_file(struct afs_operation *op)
fs/afs/fsclient.c
536
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
537
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
548
call = afs_alloc_flat_call(op->net, &afs_RXFSCreateFile,
fs/afs/fsclient.c
551
return afs_op_nomem(op);
fs/afs/fsclient.c
567
*bp++ = htonl(op->mtime.tv_sec); /* mtime */
fs/afs/fsclient.c
570
*bp++ = htonl(op->create.mode & S_IALLUGO); /* unix mode */
fs/afs/fsclient.c
575
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
580
.op = afs_FS_MakeDir,
fs/afs/fsclient.c
588
void afs_fs_make_dir(struct afs_operation *op)
fs/afs/fsclient.c
590
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
591
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
602
call = afs_alloc_flat_call(op->net, &afs_RXFSMakeDir,
fs/afs/fsclient.c
605
return afs_op_nomem(op);
fs/afs/fsclient.c
621
*bp++ = htonl(op->mtime.tv_sec); /* mtime */
fs/afs/fsclient.c
624
*bp++ = htonl(op->create.mode & S_IALLUGO); /* unix mode */
fs/afs/fsclient.c
629
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
637
struct afs_operation *op = call->op;
fs/afs/fsclient.c
638
struct afs_vnode_param *vp = &op->file[0];
fs/afs/fsclient.c
649
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
660
.op = afs_FS_RemoveFile,
fs/afs/fsclient.c
668
void afs_fs_remove_file(struct afs_operation *op)
fs/afs/fsclient.c
670
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
671
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
682
call = afs_alloc_flat_call(op->net, &afs_RXFSRemoveFile,
fs/afs/fsclient.c
685
return afs_op_nomem(op);
fs/afs/fsclient.c
703
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
708
.op = afs_FS_RemoveDir,
fs/afs/fsclient.c
716
void afs_fs_remove_dir(struct afs_operation *op)
fs/afs/fsclient.c
718
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
719
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
730
call = afs_alloc_flat_call(op->net, &afs_RXFSRemoveDir,
fs/afs/fsclient.c
733
return afs_op_nomem(op);
fs/afs/fsclient.c
751
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
759
struct afs_operation *op = call->op;
fs/afs/fsclient.c
760
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
761
struct afs_vnode_param *vp = &op->file[1];
fs/afs/fsclient.c
775
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
786
.op = afs_FS_Link,
fs/afs/fsclient.c
794
void afs_fs_link(struct afs_operation *op)
fs/afs/fsclient.c
796
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
797
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
798
struct afs_vnode_param *vp = &op->file[1];
fs/afs/fsclient.c
809
call = afs_alloc_flat_call(op->net, &afs_RXFSLink, reqsz, (21 + 21 + 6) * 4);
fs/afs/fsclient.c
811
return afs_op_nomem(op);
fs/afs/fsclient.c
832
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
840
struct afs_operation *op = call->op;
fs/afs/fsclient.c
841
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
842
struct afs_vnode_param *vp = &op->file[1];
fs/afs/fsclient.c
857
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
868
.op = afs_FS_Symlink,
fs/afs/fsclient.c
876
void afs_fs_symlink(struct afs_operation *op)
fs/afs/fsclient.c
878
const struct qstr *name = &op->dentry->d_name;
fs/afs/fsclient.c
879
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/fsclient.c
889
c_namesz = strlen(op->create.symlink);
fs/afs/fsclient.c
894
call = afs_alloc_flat_call(op->net, &afs_RXFSSymlink, reqsz,
fs/afs/fsclient.c
897
return afs_op_nomem(op);
fs/afs/fsclient.c
913
memcpy(bp, op->create.symlink, c_namesz);
fs/afs/fsclient.c
920
*bp++ = htonl(op->mtime.tv_sec); /* mtime */
fs/afs/fsclient.c
928
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/fsclient.c
936
struct afs_operation *op = call->op;
fs/afs/fsclient.c
937
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/fsclient.c
938
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/fsclient.c
952
xdr_decode_AFSVolSync(&bp, &op->volsync);
fs/afs/fsclient.c
963
.op = afs_FS_Rename,
fs/afs/fsclient.c
971
void afs_fs_rename(struct afs_operation *op)
fs/afs/fsclient.c
973
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/fsclient.c
974
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/fsclient.c
975
const struct qstr *orig_name = &op->dentry->d_name;
fs/afs/fsclient.c
976
const struct qstr *new_name = &op->dentry_2->d_name;
fs/afs/fsclient.c
994
call = afs_alloc_flat_call(op->net, &afs_RXFSRename, reqsz, (21 + 21 + 6) * 4);
fs/afs/fsclient.c
996
return afs_op_nomem(op);
fs/afs/inode.c
153
static int afs_inode_init_from_status(struct afs_operation *op,
fs/afs/inode.c
163
op->type ? op->type->name : "???");
fs/afs/inode.c
174
vnode->cb_v_check = op->cb_v_break;
fs/afs/inode.c
222
dump_vnode(vnode, op->file[0].vnode != vnode ? op->file[0].vnode : NULL);
fs/afs/inode.c
239
vnode->cb_server = op->server;
fs/afs/inode.c
251
static void afs_apply_status(struct afs_operation *op,
fs/afs/inode.c
265
op->type ? op->type->name : "???");
fs/afs/inode.c
28
void afs_init_new_symlink(struct afs_vnode *vnode, struct afs_operation *op)
fs/afs/inode.c
298
inode_set_ctime_to_ts(inode, op->ctime);
fs/afs/inode.c
30
size_t size = strlen(op->create.symlink) + 1;
fs/afs/inode.c
317
op->type ? op->type->name : "???",
fs/afs/inode.c
318
op->debug_id);
fs/afs/inode.c
354
if (op->ops == &afs_fetch_data_operation)
fs/afs/inode.c
355
op->fetch.subreq->rreq->i_size = status->size;
fs/afs/inode.c
362
static void afs_apply_callback(struct afs_operation *op,
fs/afs/inode.c
369
if (op->volume->type == AFSVL_RWVOL)
fs/afs/inode.c
370
vnode->cb_server = op->server;
fs/afs/inode.c
379
void afs_vnode_commit_status(struct afs_operation *op, struct afs_vnode_param *vp)
fs/afs/inode.c
395
op->flags &= ~AFS_OPERATION_DIR_CONFLICT;
fs/afs/inode.c
40
memcpy(p, op->create.symlink, size);
fs/afs/inode.c
406
afs_apply_status(op, vp);
fs/afs/inode.c
408
afs_apply_callback(op, vp);
fs/afs/inode.c
409
} else if (vp->op_unlinked && !(op->flags & AFS_OPERATION_DIR_CONFLICT)) {
fs/afs/inode.c
421
afs_cache_permit(vnode, op->key, vp->cb_break_before, &vp->scb);
fs/afs/inode.c
424
static void afs_fetch_status_success(struct afs_operation *op)
fs/afs/inode.c
426
struct afs_vnode_param *vp = &op->file[op->fetch_status.which];
fs/afs/inode.c
431
ret = afs_inode_init_from_status(op, vp, vnode);
fs/afs/inode.c
432
afs_op_set_error(op, ret);
fs/afs/inode.c
434
afs_cache_permit(vnode, op->key, vp->cb_break_before, &vp->scb);
fs/afs/inode.c
436
afs_vnode_commit_status(op, vp);
fs/afs/inode.c
453
struct afs_operation *op;
fs/afs/inode.c
460
op = afs_alloc_operation(key, vnode->volume);
fs/afs/inode.c
461
if (IS_ERR(op))
fs/afs/inode.c
462
return PTR_ERR(op);
fs/afs/inode.c
464
afs_op_set_vnode(op, 0, vnode);
fs/afs/inode.c
466
op->nr_files = 1;
fs/afs/inode.c
467
op->ops = &afs_fetch_status_operation;
fs/afs/inode.c
468
afs_begin_vnode_operation(op);
fs/afs/inode.c
469
afs_wait_for_operation(op);
fs/afs/inode.c
472
*_caller_access = op->file[0].scb.status.caller_access;
fs/afs/inode.c
473
return afs_put_operation(op);
fs/afs/inode.c
560
struct inode *afs_iget(struct afs_operation *op, struct afs_vnode_param *vp)
fs/afs/inode.c
562
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/inode.c
587
ret = afs_inode_init_from_status(op, vp, vnode);
fs/afs/inode.c
627
struct afs_operation *op;
fs/afs/inode.c
648
op = afs_alloc_operation(key, as->volume);
fs/afs/inode.c
649
if (IS_ERR(op)) {
fs/afs/inode.c
650
ret = PTR_ERR(op);
fs/afs/inode.c
654
afs_op_set_vnode(op, 0, vnode);
fs/afs/inode.c
656
op->nr_files = 1;
fs/afs/inode.c
657
op->ops = &afs_fetch_status_operation;
fs/afs/inode.c
658
ret = afs_do_sync_operation(op);
fs/afs/inode.c
789
static void afs_setattr_success(struct afs_operation *op)
fs/afs/inode.c
791
struct afs_vnode_param *vp = &op->file[0];
fs/afs/inode.c
795
op->setattr.old_i_size = old_i_size;
fs/afs/inode.c
796
afs_vnode_commit_status(op, vp);
fs/afs/inode.c
799
if (op->setattr.attr->ia_valid & ATTR_SIZE) {
fs/afs/inode.c
800
loff_t size = op->setattr.attr->ia_size;
fs/afs/inode.c
806
static void afs_setattr_edit_file(struct afs_operation *op)
fs/afs/inode.c
808
struct afs_vnode_param *vp = &op->file[0];
fs/afs/inode.c
812
if (op->setattr.attr->ia_valid & ATTR_SIZE) {
fs/afs/inode.c
813
loff_t size = op->setattr.attr->ia_size;
fs/afs/inode.c
814
loff_t old = op->setattr.old_i_size;
fs/afs/inode.c
844
struct afs_operation *op;
fs/afs/inode.c
902
op = afs_alloc_operation(((attr->ia_valid & ATTR_FILE) ?
fs/afs/inode.c
905
if (IS_ERR(op)) {
fs/afs/inode.c
906
ret = PTR_ERR(op);
fs/afs/inode.c
910
afs_op_set_vnode(op, 0, vnode);
fs/afs/inode.c
911
op->setattr.attr = attr;
fs/afs/inode.c
914
op->file[0].dv_delta = 1;
fs/afs/inode.c
915
op->file[0].set_size = true;
fs/afs/inode.c
917
op->ctime = attr->ia_ctime;
fs/afs/inode.c
918
op->file[0].update_ctime = 1;
fs/afs/inode.c
919
op->file[0].modification = true;
fs/afs/inode.c
921
op->ops = &afs_setattr_operation;
fs/afs/inode.c
922
ret = afs_do_sync_operation(op);
fs/afs/internal.h
1212
extern void afs_end_vnode_operation(struct afs_operation *op);
fs/afs/internal.h
1216
static inline void afs_op_set_vnode(struct afs_operation *op, unsigned int n,
fs/afs/internal.h
1219
op->file[n].vnode = vnode;
fs/afs/internal.h
1220
op->file[n].need_io_lock = true;
fs/afs/internal.h
1223
static inline void afs_op_set_fid(struct afs_operation *op, unsigned int n,
fs/afs/internal.h
1226
op->file[n].fid = *fid;
fs/afs/internal.h
1238
int afs_wait_for_fs_probes(struct afs_operation *op, struct afs_server_state *states, bool intr);
fs/afs/internal.h
1250
void afs_init_new_symlink(struct afs_vnode *vnode, struct afs_operation *op);
fs/afs/internal.h
1314
static inline void afs_op_nomem(struct afs_operation *op)
fs/afs/internal.h
1316
op->cumul_error.error = -ENOMEM;
fs/afs/internal.h
1319
static inline int afs_op_error(const struct afs_operation *op)
fs/afs/internal.h
1321
return op->cumul_error.error;
fs/afs/internal.h
1324
static inline s32 afs_op_abort_code(const struct afs_operation *op)
fs/afs/internal.h
1326
return op->cumul_error.abort_code;
fs/afs/internal.h
1329
static inline int afs_op_set_error(struct afs_operation *op, int error)
fs/afs/internal.h
1331
return op->cumul_error.error = error;
fs/afs/internal.h
1334
static inline void afs_op_accumulate_error(struct afs_operation *op, int error, s32 abort_code)
fs/afs/internal.h
1336
afs_prioritise_error(&op->cumul_error, error, abort_code);
fs/afs/internal.h
1369
void afs_clear_server_states(struct afs_operation *op);
fs/afs/internal.h
1417
static inline void afs_make_op_call(struct afs_operation *op, struct afs_call *call,
fs/afs/internal.h
1420
struct afs_addr_list *alist = op->estate->addresses;
fs/afs/internal.h
1422
op->call = call;
fs/afs/internal.h
1423
op->type = call->type;
fs/afs/internal.h
1424
call->op = op;
fs/afs/internal.h
1425
call->key = op->key;
fs/afs/internal.h
1426
call->intr = !(op->flags & AFS_OPERATION_UNINTR);
fs/afs/internal.h
1427
call->peer = rxrpc_kernel_get_peer(alist->addrs[op->addr_index].peer);
fs/afs/internal.h
1428
call->service_id = op->server->service_id;
fs/afs/internal.h
1551
bool afs_check_server_record(struct afs_operation *op, struct afs_server *server, struct key *key);
fs/afs/internal.h
160
struct afs_operation *op;
fs/afs/internal.h
1607
int afs_update_volume_state(struct afs_operation *op);
fs/afs/internal.h
1709
void yfs_fs_rename_replace(struct afs_operation *op);
fs/afs/internal.h
1710
void yfs_fs_rename_noreplace(struct afs_operation *op);
fs/afs/internal.h
1711
void yfs_fs_rename_exchange(struct afs_operation *op);
fs/afs/internal.h
1753
static inline void afs_update_dentry_version(struct afs_operation *op,
fs/afs/internal.h
1757
if (!op->cumul_error.error)
fs/afs/internal.h
1777
static inline void afs_check_dir_conflict(struct afs_operation *op,
fs/afs/internal.h
1781
op->flags |= AFS_OPERATION_DIR_CONFLICT;
fs/afs/internal.h
198
unsigned int op; /* Really enum afs_fs_operation */
fs/afs/internal.h
833
void (*issue_afs_rpc)(struct afs_operation *op);
fs/afs/internal.h
834
void (*issue_yfs_rpc)(struct afs_operation *op);
fs/afs/internal.h
835
void (*success)(struct afs_operation *op);
fs/afs/internal.h
836
void (*aborted)(struct afs_operation *op);
fs/afs/internal.h
837
void (*failed)(struct afs_operation *op);
fs/afs/internal.h
838
void (*edit_dir)(struct afs_operation *op);
fs/afs/internal.h
839
void (*put)(struct afs_operation *op);
fs/afs/rotate.c
114
static void afs_busy(struct afs_operation *op, u32 abort_code)
fs/afs/rotate.c
126
op->volume->vid, op->volume->name, &op->server->uuid, m);
fs/afs/rotate.c
132
static bool afs_sleep_and_retry(struct afs_operation *op)
fs/afs/rotate.c
134
trace_afs_rotate(op, afs_rotate_trace_busy_sleep, 0);
fs/afs/rotate.c
135
if (!(op->flags & AFS_OPERATION_UNINTR)) {
fs/afs/rotate.c
138
afs_op_set_error(op, -ERESTARTSYS);
fs/afs/rotate.c
152
bool afs_select_fileserver(struct afs_operation *op)
fs/afs/rotate.c
156
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/rotate.c
158
s32 abort_code = op->call_abort_code;
fs/afs/rotate.c
160
int error = op->call_error, addr_index, i, j;
fs/afs/rotate.c
162
op->nr_iterations++;
fs/afs/rotate.c
165
op->debug_id, op->nr_iterations, op->volume->vid,
fs/afs/rotate.c
166
op->server_index, op->untried_servers,
fs/afs/rotate.c
167
op->addr_index, op->addr_tried,
fs/afs/rotate.c
170
if (op->flags & AFS_OPERATION_STOP) {
fs/afs/rotate.c
171
trace_afs_rotate(op, afs_rotate_trace_stopped, 0);
fs/afs/rotate.c
176
if (op->nr_iterations == 0)
fs/afs/rotate.c
179
WRITE_ONCE(op->estate->addresses->addrs[op->addr_index].last_error, error);
fs/afs/rotate.c
18
void afs_clear_server_states(struct afs_operation *op)
fs/afs/rotate.c
180
trace_afs_rotate(op, afs_rotate_trace_iter, op->call_error);
fs/afs/rotate.c
183
switch (op->call_error) {
fs/afs/rotate.c
186
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
188
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
189
op->cumul_error.responded = true;
fs/afs/rotate.c
197
error = afs_update_volume_state(op);
fs/afs/rotate.c
200
afs_sleep_and_retry(op);
fs/afs/rotate.c
203
afs_op_set_error(op, error);
fs/afs/rotate.c
209
afs_op_set_error(op, error);
fs/afs/rotate.c
210
op->flags |= AFS_OPERATION_STOP;
fs/afs/rotate.c
211
trace_afs_rotate(op, afs_rotate_trace_stop, error);
fs/afs/rotate.c
22
if (op->server_states) {
fs/afs/rotate.c
224
trace_afs_rotate(op, afs_rotate_trace_aborted, abort_code);
fs/afs/rotate.c
225
op->cumul_error.responded = true;
fs/afs/rotate.c
23
for (i = 0; i < op->server_list->nr_servers; i++)
fs/afs/rotate.c
238
if (op->flags & AFS_OPERATION_VNOVOL) {
fs/afs/rotate.c
239
afs_op_accumulate_error(op, -EREMOTEIO, abort_code);
fs/afs/rotate.c
24
afs_put_endpoint_state(op->server_states[i].endpoint_state,
fs/afs/rotate.c
243
write_lock(&op->volume->servers_lock);
fs/afs/rotate.c
244
op->server_list->vnovol_mask |= 1 << op->server_index;
fs/afs/rotate.c
245
write_unlock(&op->volume->servers_lock);
fs/afs/rotate.c
247
set_bit(AFS_VOLUME_NEEDS_UPDATE, &op->volume->flags);
fs/afs/rotate.c
248
error = afs_check_volume_status(op->volume, op);
fs/afs/rotate.c
250
afs_op_set_error(op, error);
fs/afs/rotate.c
254
if (test_bit(AFS_VOLUME_DELETED, &op->volume->flags)) {
fs/afs/rotate.c
255
afs_op_set_error(op, -ENOMEDIUM);
fs/afs/rotate.c
26
kfree(op->server_states);
fs/afs/rotate.c
262
if (rcu_access_pointer(op->volume->servers) == op->server_list) {
fs/afs/rotate.c
263
afs_op_accumulate_error(op, -EREMOTEIO, abort_code);
fs/afs/rotate.c
268
op->flags |= AFS_OPERATION_VNOVOL;
fs/afs/rotate.c
277
afs_op_accumulate_error(op, -EREMOTEIO, abort_code);
fs/afs/rotate.c
308
afs_op_accumulate_error(op, -ETIMEDOUT, abort_code);
fs/afs/rotate.c
323
&op->server_list->servers[op->server_index].flags)) {
fs/afs/rotate.c
324
afs_busy(op, abort_code);
fs/afs/rotate.c
326
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
328
if (op->flags & AFS_OPERATION_NO_VSLEEP) {
fs/afs/rotate.c
329
afs_op_set_error(op, -EADV);
fs/afs/rotate.c
34
static bool afs_start_fs_iteration(struct afs_operation *op,
fs/afs/rotate.c
349
if (op->flags & AFS_OPERATION_NO_VSLEEP) {
fs/afs/rotate.c
350
afs_op_set_error(op, -EBUSY);
fs/afs/rotate.c
354
&op->server_list->servers[op->server_index].flags)) {
fs/afs/rotate.c
355
afs_busy(op, abort_code);
fs/afs/rotate.c
357
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
360
if (op->flags & AFS_OPERATION_CUR_ONLY) {
fs/afs/rotate.c
361
if (!afs_sleep_and_retry(op))
fs/afs/rotate.c
369
op->flags |= AFS_OPERATION_VBUSY;
fs/afs/rotate.c
380
if (op->flags & AFS_OPERATION_VMOVED) {
fs/afs/rotate.c
381
afs_op_set_error(op, -EREMOTEIO);
fs/afs/rotate.c
384
op->flags |= AFS_OPERATION_VMOVED;
fs/afs/rotate.c
386
set_bit(AFS_VOLUME_WAIT, &op->volume->flags);
fs/afs/rotate.c
387
set_bit(AFS_VOLUME_NEEDS_UPDATE, &op->volume->flags);
fs/afs/rotate.c
388
error = afs_check_volume_status(op->volume, op);
fs/afs/rotate.c
390
afs_op_set_error(op, error);
fs/afs/rotate.c
403
if (rcu_access_pointer(op->volume->servers) == op->server_list) {
fs/afs/rotate.c
404
afs_op_accumulate_error(op, -ENOMEDIUM, abort_code);
fs/afs/rotate.c
41
trace_afs_rotate(op, afs_rotate_trace_start, 0);
fs/afs/rotate.c
412
afs_op_accumulate_error(op, -EREMOTEIO, abort_code);
fs/afs/rotate.c
413
if (op->volume->type != AFSVL_RWVOL)
fs/afs/rotate.c
423
afs_op_set_error(op, -ENOSPC);
fs/afs/rotate.c
43
read_lock(&op->volume->servers_lock);
fs/afs/rotate.c
432
afs_op_set_error(op, -EDQUOT);
fs/afs/rotate.c
438
afs_op_set_error(op, -ENOTSUPP);
fs/afs/rotate.c
439
if (op->flags & AFS_OPERATION_DOWNGRADE) {
fs/afs/rotate.c
44
op->server_list = afs_get_serverlist(
fs/afs/rotate.c
440
op->flags &= ~AFS_OPERATION_DOWNGRADE;
fs/afs/rotate.c
446
afs_op_accumulate_error(op, error, abort_code);
fs/afs/rotate.c
449
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
45
rcu_dereference_protected(op->volume->servers,
fs/afs/rotate.c
451
&op->server_list->servers[op->server_index].flags);
fs/afs/rotate.c
457
if (afs_op_error(op) != -EDESTADDRREQ)
fs/afs/rotate.c
46
lockdep_is_held(&op->volume->servers_lock)));
fs/afs/rotate.c
467
afs_op_accumulate_error(op, error, 0);
fs/afs/rotate.c
47
read_unlock(&op->volume->servers_lock);
fs/afs/rotate.c
472
op->type ? op->type->name : "???", op->debug_id);
fs/afs/rotate.c
476
afs_op_set_error(op, error);
fs/afs/rotate.c
481
trace_afs_rotate(op, afs_rotate_trace_restart, 0);
fs/afs/rotate.c
483
op->estate = NULL;
fs/afs/rotate.c
484
op->server = NULL;
fs/afs/rotate.c
485
afs_clear_server_states(op);
fs/afs/rotate.c
486
op->server_states = NULL;
fs/afs/rotate.c
487
afs_put_serverlist(op->net, op->server_list);
fs/afs/rotate.c
488
op->server_list = NULL;
fs/afs/rotate.c
49
op->server_states = kzalloc_objs(op->server_states[0],
fs/afs/rotate.c
491
ASSERTCMP(op->estate, ==, NULL);
fs/afs/rotate.c
495
error = afs_check_volume_status(op->volume, op);
fs/afs/rotate.c
496
trace_afs_rotate(op, afs_rotate_trace_check_vol_status, error);
fs/afs/rotate.c
498
afs_op_set_error(op, error);
fs/afs/rotate.c
50
op->server_list->nr_servers);
fs/afs/rotate.c
502
if (!afs_start_fs_iteration(op, vnode))
fs/afs/rotate.c
505
_debug("__ VOL %llx __", op->volume->vid);
fs/afs/rotate.c
508
_debug("pick [%lx]", op->untried_servers);
fs/afs/rotate.c
509
ASSERTCMP(op->estate, ==, NULL);
fs/afs/rotate.c
51
if (!op->server_states) {
fs/afs/rotate.c
511
error = afs_wait_for_fs_probes(op, op->server_states,
fs/afs/rotate.c
512
!(op->flags & AFS_OPERATION_UNINTR));
fs/afs/rotate.c
515
trace_afs_rotate(op, afs_rotate_trace_probe_none, 0);
fs/afs/rotate.c
518
trace_afs_rotate(op, afs_rotate_trace_probe_response, 0);
fs/afs/rotate.c
52
afs_op_nomem(op);
fs/afs/rotate.c
521
trace_afs_rotate(op, afs_rotate_trace_probe_superseded, 0);
fs/afs/rotate.c
524
trace_afs_rotate(op, afs_rotate_trace_probe_error, error);
fs/afs/rotate.c
525
afs_op_set_error(op, error);
fs/afs/rotate.c
53
trace_afs_rotate(op, afs_rotate_trace_nomem, 0);
fs/afs/rotate.c
533
if (op->server) {
fs/afs/rotate.c
534
_debug("server %u", op->server_index);
fs/afs/rotate.c
535
if (test_bit(op->server_index, &op->untried_servers))
fs/afs/rotate.c
537
op->server = NULL;
fs/afs/rotate.c
542
op->server_index = -1;
fs/afs/rotate.c
544
for (i = 0; i < op->server_list->nr_servers; i++) {
fs/afs/rotate.c
546
struct afs_server_entry *se = &op->server_list->servers[i];
fs/afs/rotate.c
550
if (!test_bit(i, &op->untried_servers) ||
fs/afs/rotate.c
554
es = op->server_states[i].endpoint_state;
fs/afs/rotate.c
557
afs_get_address_preferences_rcu(op->net, sal);
fs/afs/rotate.c
564
op->server_index = i;
fs/afs/rotate.c
571
if (op->server_index == -1)
fs/afs/rotate.c
575
trace_afs_rotate(op, afs_rotate_trace_selected_server, best_prio);
fs/afs/rotate.c
576
_debug("use %d prio %u", op->server_index, best_prio);
fs/afs/rotate.c
577
__clear_bit(op->server_index, &op->untried_servers);
fs/afs/rotate.c
58
for (i = 0; i < op->server_list->nr_servers; i++) {
fs/afs/rotate.c
583
ASSERTCMP(op->estate, ==, NULL);
fs/afs/rotate.c
584
server = op->server_list->servers[op->server_index].server;
fs/afs/rotate.c
586
if (!afs_check_server_record(op, server, op->key))
fs/afs/rotate.c
591
op->flags |= AFS_OPERATION_RETRY_SERVER;
fs/afs/rotate.c
592
op->server = server;
fs/afs/rotate.c
60
struct afs_server_state *s = &op->server_states[i];
fs/afs/rotate.c
600
op->addr_tried = 0;
fs/afs/rotate.c
601
op->addr_index = -1;
fs/afs/rotate.c
607
op->estate = op->server_states[op->server_index].endpoint_state;
fs/afs/rotate.c
608
set = READ_ONCE(op->estate->responsive_set);
fs/afs/rotate.c
609
failed = READ_ONCE(op->estate->failed_set);
fs/afs/rotate.c
610
_debug("iterate ES=%x rs=%lx fs=%lx", op->estate->probe_seq, set, failed);
fs/afs/rotate.c
611
set &= ~(failed | op->addr_tried);
fs/afs/rotate.c
612
trace_afs_rotate(op, afs_rotate_trace_iterate_addr, set);
fs/afs/rotate.c
616
alist = op->estate->addresses;
fs/afs/rotate.c
62
server = op->server_list->servers[i].server;
fs/afs/rotate.c
630
op->addr_index = addr_index;
fs/afs/rotate.c
631
set_bit(addr_index, &op->addr_tried);
fs/afs/rotate.c
634
op->server_index, addr_index, alist->nr_addrs,
fs/afs/rotate.c
635
rxrpc_kernel_remote_addr(alist->addrs[op->addr_index].peer));
fs/afs/rotate.c
637
op->volsync.creation = TIME64_MIN;
fs/afs/rotate.c
638
op->volsync.update = TIME64_MIN;
fs/afs/rotate.c
639
op->call_responded = false;
fs/afs/rotate.c
644
error = afs_wait_for_one_fs_probe(op->server, op->estate, op->addr_tried,
fs/afs/rotate.c
645
!(op->flags & AFS_OPERATION_UNINTR));
fs/afs/rotate.c
654
trace_afs_rotate(op, afs_rotate_trace_probe_fileserver, 0);
fs/afs/rotate.c
655
afs_probe_fileserver(op->net, op->server);
fs/afs/rotate.c
656
if (op->flags & AFS_OPERATION_RETRY_SERVER) {
fs/afs/rotate.c
657
error = afs_wait_for_one_fs_probe(op->server, op->estate, op->addr_tried,
fs/afs/rotate.c
658
!(op->flags & AFS_OPERATION_UNINTR));
fs/afs/rotate.c
661
op->flags &= ~AFS_OPERATION_RETRY_SERVER;
fs/afs/rotate.c
662
trace_afs_rotate(op, afs_rotate_trace_retry_server, 1);
fs/afs/rotate.c
665
trace_afs_rotate(op, afs_rotate_trace_retry_server, 0);
fs/afs/rotate.c
668
afs_op_set_error(op, error);
fs/afs/rotate.c
677
trace_afs_rotate(op, afs_rotate_trace_next_server, 0);
fs/afs/rotate.c
679
op->estate = NULL;
fs/afs/rotate.c
686
trace_afs_rotate(op, afs_rotate_trace_no_more_servers, 0);
fs/afs/rotate.c
687
if (op->flags & AFS_OPERATION_VBUSY) {
fs/afs/rotate.c
688
afs_sleep_and_retry(op);
fs/afs/rotate.c
689
op->flags &= ~AFS_OPERATION_VBUSY;
fs/afs/rotate.c
69
afs_get_address_preferences(op->net, estate->addresses);
fs/afs/rotate.c
694
for (i = 0; i < op->server_list->nr_servers; i++) {
fs/afs/rotate.c
697
estate = op->server_states[i].endpoint_state;
fs/afs/rotate.c
700
afs_op_accumulate_error(op, error, estate->abort_code);
fs/afs/rotate.c
705
trace_afs_rotate(op, afs_rotate_trace_failed, 0);
fs/afs/rotate.c
706
op->flags |= AFS_OPERATION_STOP;
fs/afs/rotate.c
707
op->estate = NULL;
fs/afs/rotate.c
708
_leave(" = f [failed %d]", afs_op_error(op));
fs/afs/rotate.c
715
void afs_dump_edestaddrreq(const struct afs_operation *op)
fs/afs/rotate.c
728
op->file[0].cb_break_before,
fs/afs/rotate.c
729
op->file[1].cb_break_before, op->flags, op->cumul_error.error);
fs/afs/rotate.c
731
op->untried_servers, op->server_index, op->nr_iterations);
fs/afs/rotate.c
733
op->call_error, op->call_abort_code, op->call_responded);
fs/afs/rotate.c
735
if (op->server_list) {
fs/afs/rotate.c
736
const struct afs_server_list *sl = op->server_list;
fs/afs/rotate.c
74
op->untried_servers = (1UL << op->server_list->nr_servers) - 1;
fs/afs/rotate.c
75
op->server_index = -1;
fs/afs/rotate.c
761
pr_notice("AC: t=%lx ax=%d\n", op->addr_tried, op->addr_index);
fs/afs/rotate.c
80
for (i = 0; i < op->server_list->nr_servers; i++) {
fs/afs/rotate.c
81
server = op->server_list->servers[i].server;
fs/afs/rotate.c
83
op->server_index = i;
fs/afs/rotate.c
92
if (op->flags & AFS_OPERATION_CUR_ONLY) {
fs/afs/rotate.c
93
afs_op_set_error(op, -ESTALE);
fs/afs/rotate.c
94
trace_afs_rotate(op, afs_rotate_trace_stale_lock, 0);
fs/afs/rxrpc.c
298
call->operation_ID = type->op;
fs/afs/rxrpc.c
559
if (call->op)
fs/afs/rxrpc.c
561
&call->op->server->flags);
fs/afs/server.c
542
static noinline bool afs_update_server_record(struct afs_operation *op,
fs/afs/server.c
556
alist = afs_vl_lookup_addrs(server, op->key);
fs/afs/server.c
565
(op->flags & AFS_OPERATION_UNINTR) &&
fs/afs/server.c
570
afs_op_set_error(op, PTR_ERR(alist));
fs/afs/server.c
571
_leave(" = f [%d]", afs_op_error(op));
fs/afs/server.c
576
afs_fs_probe_fileserver(op->net, server, alist, key);
fs/afs/server.c
586
bool afs_check_server_record(struct afs_operation *op, struct afs_server *server,
fs/afs/server.c
607
success = afs_update_server_record(op, server, key);
fs/afs/server.c
616
(op->flags & AFS_OPERATION_UNINTR) ?
fs/afs/server.c
619
afs_op_set_error(op, ret);
fs/afs/super.c
720
static void afs_get_volume_status_success(struct afs_operation *op)
fs/afs/super.c
722
struct afs_volume_status *vs = &op->volstatus.vs;
fs/afs/super.c
723
struct kstatfs *buf = op->volstatus.buf;
fs/afs/super.c
747
struct afs_operation *op;
fs/afs/super.c
761
op = afs_alloc_operation(NULL, as->volume);
fs/afs/super.c
762
if (IS_ERR(op))
fs/afs/super.c
763
return PTR_ERR(op);
fs/afs/super.c
765
afs_op_set_vnode(op, 0, vnode);
fs/afs/super.c
766
op->nr_files = 1;
fs/afs/super.c
767
op->volstatus.buf = buf;
fs/afs/super.c
768
op->ops = &afs_get_volume_status_operation;
fs/afs/super.c
769
return afs_do_sync_operation(op);
fs/afs/validation.c
153
static bool __afs_is_server_excluded(struct afs_operation *op, struct afs_volume *volume)
fs/afs/validation.c
165
if (op->server == se->server) {
fs/afs/validation.c
179
static int afs_is_server_excluded(struct afs_operation *op, struct afs_volume *volume)
fs/afs/validation.c
183
if (__afs_is_server_excluded(op, volume))
fs/afs/validation.c
187
ret = afs_check_volume_status(op->volume, op);
fs/afs/validation.c
191
return __afs_is_server_excluded(op, volume);
fs/afs/validation.c
197
static int afs_update_volume_creation_time(struct afs_operation *op, struct afs_volume *volume)
fs/afs/validation.c
201
time64_t old = op->pre_volsync.creation;
fs/afs/validation.c
202
time64_t new = op->volsync.creation;
fs/afs/validation.c
239
ret = afs_is_server_excluded(op, volume);
fs/afs/validation.c
264
static void afs_update_volume_update_time(struct afs_operation *op, struct afs_volume *volume)
fs/afs/validation.c
268
time64_t old = op->pre_volsync.update;
fs/afs/validation.c
269
time64_t new = op->volsync.update;
fs/afs/validation.c
303
static int afs_update_volume_times(struct afs_operation *op, struct afs_volume *volume)
fs/afs/validation.c
307
if (likely(op->volsync.creation == volume->creation_time &&
fs/afs/validation.c
308
op->volsync.update == volume->update_time))
fs/afs/validation.c
312
if (op->volsync.creation != volume->creation_time) {
fs/afs/validation.c
313
ret = afs_update_volume_creation_time(op, volume);
fs/afs/validation.c
317
if (op->volsync.update != volume->update_time)
fs/afs/validation.c
318
afs_update_volume_update_time(op, volume);
fs/afs/validation.c
328
int afs_update_volume_state(struct afs_operation *op)
fs/afs/validation.c
330
struct afs_server_list *slist = op->server_list;
fs/afs/validation.c
331
struct afs_server_entry *se = &slist->servers[op->server_index];
fs/afs/validation.c
332
struct afs_callback *cb = &op->file[0].scb.callback;
fs/afs/validation.c
333
struct afs_volume *volume = op->volume;
fs/afs/validation.c
338
_enter("%llx", op->volume->vid);
fs/afs/validation.c
340
if (op->volsync.creation != TIME64_MIN || op->volsync.update != TIME64_MIN) {
fs/afs/validation.c
341
ret = afs_update_volume_times(op, volume);
fs/afs/validation.c
348
if (op->cb_v_break == cb_v_break &&
fs/afs/validation.c
349
(op->file[0].scb.have_cb || op->file[1].scb.have_cb)) {
fs/afs/validation.c
352
if (!op->file[0].scb.have_cb)
fs/afs/validation.c
353
expires_at = op->file[1].scb.callback.expires_at;
fs/afs/validation.c
358
if (cb_v_check < op->cb_v_break)
fs/afs/validation.c
359
atomic_cmpxchg(&volume->cb_v_check, cb_v_check, op->cb_v_break);
fs/afs/vlclient.c
101
.op = afs_VL_GetEntryByNameU,
fs/afs/vlclient.c
249
.op = afs_VL_GetAddrsU,
fs/afs/vlclient.c
371
.op = afs_VL_GetCapabilities,
fs/afs/vlclient.c
625
.op = afs_YFSVL_GetEndpoints,
fs/afs/vlclient.c
744
.op = afs_YFSVL_GetCellName,
fs/afs/volume.c
424
int afs_check_volume_status(struct afs_volume *volume, struct afs_operation *op)
fs/afs/volume.c
442
ret = afs_update_volume_status(volume, op->key);
fs/afs/volume.c
459
(op->flags & AFS_OPERATION_UNINTR) ?
fs/afs/write.c
107
struct afs_operation *op;
fs/afs/write.c
131
op = afs_alloc_operation(wreq->netfs_priv, vnode->volume);
fs/afs/write.c
132
if (IS_ERR(op))
fs/afs/write.c
135
afs_op_set_vnode(op, 0, vnode);
fs/afs/write.c
136
op->file[0].dv_delta = 1;
fs/afs/write.c
137
op->file[0].modification = true;
fs/afs/write.c
138
op->store.pos = pos;
fs/afs/write.c
139
op->store.size = len;
fs/afs/write.c
140
op->flags |= AFS_OPERATION_UNINTR;
fs/afs/write.c
141
op->ops = &afs_store_data_operation;
fs/afs/write.c
143
afs_begin_vnode_operation(op);
fs/afs/write.c
145
op->store.write_iter = &subreq->io_iter;
fs/afs/write.c
146
op->store.i_size = umax(pos + len, vnode->netfs.remote_i_size);
fs/afs/write.c
147
op->mtime = inode_get_mtime(&vnode->netfs.inode);
fs/afs/write.c
149
afs_wait_for_operation(op);
fs/afs/write.c
150
ret = afs_put_operation(op);
fs/afs/write.c
67
static void afs_store_data_success(struct afs_operation *op)
fs/afs/write.c
69
struct afs_vnode *vnode = op->file[0].vnode;
fs/afs/write.c
71
op->ctime = op->file[0].scb.status.mtime_client;
fs/afs/write.c
72
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/write.c
73
if (!afs_op_error(op)) {
fs/afs/write.c
74
afs_pages_written_back(vnode, op->store.pos, op->store.size);
fs/afs/write.c
76
atomic_long_add(op->store.size, &afs_v2net(vnode)->n_store_bytes);
fs/afs/xattr.c
105
struct afs_operation *op;
fs/afs/xattr.c
111
op = afs_alloc_operation(NULL, vnode->volume);
fs/afs/xattr.c
112
if (IS_ERR(op))
fs/afs/xattr.c
115
afs_op_set_vnode(op, 0, vnode);
fs/afs/xattr.c
116
if (!afs_make_acl(op, buffer, size))
fs/afs/xattr.c
117
return afs_put_operation(op);
fs/afs/xattr.c
119
op->ops = &afs_store_acl_operation;
fs/afs/xattr.c
120
return afs_do_sync_operation(op);
fs/afs/xattr.c
143
struct afs_operation *op;
fs/afs/xattr.c
169
op = afs_alloc_operation(NULL, vnode->volume);
fs/afs/xattr.c
17
static void afs_acl_success(struct afs_operation *op)
fs/afs/xattr.c
170
if (IS_ERR(op))
fs/afs/xattr.c
173
afs_op_set_vnode(op, 0, vnode);
fs/afs/xattr.c
174
op->yacl = yacl;
fs/afs/xattr.c
175
op->ops = &yfs_fetch_opaque_acl_operation;
fs/afs/xattr.c
177
afs_begin_vnode_operation(op);
fs/afs/xattr.c
178
afs_wait_for_operation(op);
fs/afs/xattr.c
179
ret = afs_put_operation(op);
fs/afs/xattr.c
19
afs_vnode_commit_status(op, &op->file[0]);
fs/afs/xattr.c
22
static void afs_acl_put(struct afs_operation *op)
fs/afs/xattr.c
236
struct afs_operation *op;
fs/afs/xattr.c
24
kfree(op->acl);
fs/afs/xattr.c
244
op = afs_alloc_operation(NULL, vnode->volume);
fs/afs/xattr.c
245
if (IS_ERR(op))
fs/afs/xattr.c
248
afs_op_set_vnode(op, 0, vnode);
fs/afs/xattr.c
249
if (!afs_make_acl(op, buffer, size))
fs/afs/xattr.c
250
return afs_put_operation(op);
fs/afs/xattr.c
252
op->ops = &yfs_store_opaque_acl2_operation;
fs/afs/xattr.c
253
ret = afs_do_sync_operation(op);
fs/afs/xattr.c
41
struct afs_operation *op;
fs/afs/xattr.c
46
op = afs_alloc_operation(NULL, vnode->volume);
fs/afs/xattr.c
47
if (IS_ERR(op))
fs/afs/xattr.c
50
afs_op_set_vnode(op, 0, vnode);
fs/afs/xattr.c
51
op->ops = &afs_fetch_acl_operation;
fs/afs/xattr.c
53
afs_begin_vnode_operation(op);
fs/afs/xattr.c
54
afs_wait_for_operation(op);
fs/afs/xattr.c
55
acl = op->acl;
fs/afs/xattr.c
56
op->acl = NULL;
fs/afs/xattr.c
57
ret = afs_put_operation(op);
fs/afs/xattr.c
73
static bool afs_make_acl(struct afs_operation *op,
fs/afs/xattr.c
80
afs_op_nomem(op);
fs/afs/xattr.c
86
op->acl = acl;
fs/afs/yfsclient.c
1000
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1016
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
1026
.op = yfs_FS_Rename,
fs/afs/yfsclient.c
1034
void yfs_fs_rename(struct afs_operation *op)
fs/afs/yfsclient.c
1036
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1037
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1038
const struct qstr *orig_name = &op->dentry->d_name;
fs/afs/yfsclient.c
1039
const struct qstr *new_name = &op->dentry_2->d_name;
fs/afs/yfsclient.c
1045
if (!test_bit(AFS_SERVER_FL_NO_RENAME2, &op->server->flags))
fs/afs/yfsclient.c
1046
return yfs_fs_rename_replace(op);
fs/afs/yfsclient.c
1048
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRename,
fs/afs/yfsclient.c
1059
return afs_op_nomem(op);
fs/afs/yfsclient.c
1073
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1082
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
1083
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1084
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1085
struct afs_vnode_param *old_vp = &op->more_files[0];
fs/afs/yfsclient.c
1103
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
1115
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
1116
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1117
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1118
struct afs_vnode_param *old_vp = &op->more_files[0];
fs/afs/yfsclient.c
1119
struct afs_vnode_param *new_vp = &op->more_files[1];
fs/afs/yfsclient.c
1139
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
1149
set_bit(AFS_SERVER_FL_NO_RENAME2, &call->op->server->flags);
fs/afs/yfsclient.c
1150
call->op->flags |= AFS_OPERATION_DOWNGRADE;
fs/afs/yfsclient.c
1159
.op = yfs_FS_Rename_Replace,
fs/afs/yfsclient.c
1170
.op = yfs_FS_Rename_NoReplace,
fs/afs/yfsclient.c
1180
.op = yfs_FS_Rename_Exchange,
fs/afs/yfsclient.c
1189
void yfs_fs_rename_replace(struct afs_operation *op)
fs/afs/yfsclient.c
1191
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1192
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1193
const struct qstr *orig_name = &op->dentry->d_name;
fs/afs/yfsclient.c
1194
const struct qstr *new_name = &op->dentry_2->d_name;
fs/afs/yfsclient.c
1200
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRename_Replace,
fs/afs/yfsclient.c
1215
return afs_op_nomem(op);
fs/afs/yfsclient.c
1229
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1235
void yfs_fs_rename_noreplace(struct afs_operation *op)
fs/afs/yfsclient.c
1237
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1238
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1239
const struct qstr *orig_name = &op->dentry->d_name;
fs/afs/yfsclient.c
1240
const struct qstr *new_name = &op->dentry_2->d_name;
fs/afs/yfsclient.c
1246
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRename_NoReplace,
fs/afs/yfsclient.c
1259
return afs_op_nomem(op);
fs/afs/yfsclient.c
1273
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1279
void yfs_fs_rename_exchange(struct afs_operation *op)
fs/afs/yfsclient.c
1281
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/afs/yfsclient.c
1282
struct afs_vnode_param *new_dvp = &op->file[1];
fs/afs/yfsclient.c
1283
const struct qstr *orig_name = &op->dentry->d_name;
fs/afs/yfsclient.c
1284
const struct qstr *new_name = &op->dentry_2->d_name;
fs/afs/yfsclient.c
1290
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRename_Exchange,
fs/afs/yfsclient.c
1305
return afs_op_nomem(op);
fs/afs/yfsclient.c
1319
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1327
.op = yfs_FS_StoreData64,
fs/afs/yfsclient.c
1335
void yfs_fs_store_data(struct afs_operation *op)
fs/afs/yfsclient.c
1337
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1342
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
1345
(unsigned long long)op->store.size,
fs/afs/yfsclient.c
1346
(unsigned long long)op->store.pos,
fs/afs/yfsclient.c
1347
(unsigned long long)op->store.i_size);
fs/afs/yfsclient.c
1349
call = afs_alloc_flat_call(op->net, &yfs_RXYFSStoreData64,
fs/afs/yfsclient.c
1358
return afs_op_nomem(op);
fs/afs/yfsclient.c
1360
call->write_iter = op->store.write_iter;
fs/afs/yfsclient.c
1367
bp = xdr_encode_YFSStoreStatus(bp, NULL, &op->mtime);
fs/afs/yfsclient.c
1368
bp = xdr_encode_u64(bp, op->store.pos);
fs/afs/yfsclient.c
1369
bp = xdr_encode_u64(bp, op->store.size);
fs/afs/yfsclient.c
1370
bp = xdr_encode_u64(bp, op->store.i_size);
fs/afs/yfsclient.c
1375
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1383
.op = yfs_FS_StoreStatus,
fs/afs/yfsclient.c
1390
.op = yfs_FS_StoreData64,
fs/afs/yfsclient.c
1399
static void yfs_fs_setattr_size(struct afs_operation *op)
fs/afs/yfsclient.c
1401
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1403
struct iattr *attr = op->setattr.attr;
fs/afs/yfsclient.c
1407
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
1409
call = afs_alloc_flat_call(op->net, &yfs_RXYFSStoreData64_as_Status,
fs/afs/yfsclient.c
1417
return afs_op_nomem(op);
fs/afs/yfsclient.c
1432
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1439
void yfs_fs_setattr(struct afs_operation *op)
fs/afs/yfsclient.c
1441
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1443
struct iattr *attr = op->setattr.attr;
fs/afs/yfsclient.c
1447
return yfs_fs_setattr_size(op);
fs/afs/yfsclient.c
1450
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
1452
call = afs_alloc_flat_call(op->net, &yfs_RXYFSStoreStatus,
fs/afs/yfsclient.c
1459
return afs_op_nomem(op);
fs/afs/yfsclient.c
1471
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1479
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
1501
xdr_decode_YFSFetchVolumeStatus(&bp, &op->volstatus.vs);
fs/afs/yfsclient.c
1607
.op = yfs_FS_GetVolumeStatus,
fs/afs/yfsclient.c
1615
void yfs_fs_get_volume_status(struct afs_operation *op)
fs/afs/yfsclient.c
1617
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1623
call = afs_alloc_flat_call(op->net, &yfs_RXYFSGetVolumeStatus,
fs/afs/yfsclient.c
1631
return afs_op_nomem(op);
fs/afs/yfsclient.c
1642
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1650
.op = yfs_FS_SetLock,
fs/afs/yfsclient.c
1661
.op = yfs_FS_ExtendLock,
fs/afs/yfsclient.c
1672
.op = yfs_FS_ReleaseLock,
fs/afs/yfsclient.c
1680
void yfs_fs_set_lock(struct afs_operation *op)
fs/afs/yfsclient.c
1682
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1688
call = afs_alloc_flat_call(op->net, &yfs_RXYFSSetLock,
fs/afs/yfsclient.c
1695
return afs_op_nomem(op);
fs/afs/yfsclient.c
1702
bp = xdr_encode_u32(bp, op->lock.type);
fs/afs/yfsclient.c
1706
trace_afs_make_fs_calli(call, &vp->fid, op->lock.type);
fs/afs/yfsclient.c
1707
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1713
void yfs_fs_extend_lock(struct afs_operation *op)
fs/afs/yfsclient.c
1715
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1721
call = afs_alloc_flat_call(op->net, &yfs_RXYFSExtendLock,
fs/afs/yfsclient.c
1727
return afs_op_nomem(op);
fs/afs/yfsclient.c
1738
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1744
void yfs_fs_release_lock(struct afs_operation *op)
fs/afs/yfsclient.c
1746
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
1752
call = afs_alloc_flat_call(op->net, &yfs_RXYFSReleaseLock,
fs/afs/yfsclient.c
1758
return afs_op_nomem(op);
fs/afs/yfsclient.c
1769
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1777
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
1778
struct afs_vnode_param *vp = &op->file[op->fetch_status.which];
fs/afs/yfsclient.c
1790
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
1801
.op = yfs_FS_FetchStatus,
fs/afs/yfsclient.c
1809
void yfs_fs_fetch_status(struct afs_operation *op)
fs/afs/yfsclient.c
1811
struct afs_vnode_param *vp = &op->file[op->fetch_status.which];
fs/afs/yfsclient.c
1816
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
1818
call = afs_alloc_flat_call(op->net, &yfs_RXYFSFetchStatus,
fs/afs/yfsclient.c
1825
return afs_op_nomem(op);
fs/afs/yfsclient.c
1836
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
1844
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
1866
_debug("status count: %u/%u", tmp, op->nr_files);
fs/afs/yfsclient.c
1867
if (tmp != op->nr_files)
fs/afs/yfsclient.c
1884
scb = &op->file[0].scb;
fs/afs/yfsclient.c
1887
scb = &op->file[1].scb;
fs/afs/yfsclient.c
1890
scb = &op->more_files[call->count - 2].scb;
fs/afs/yfsclient.c
1898
if (call->count < op->nr_files)
fs/afs/yfsclient.c
1915
if (tmp != op->nr_files)
fs/afs/yfsclient.c
1932
scb = &op->file[0].scb;
fs/afs/yfsclient.c
1935
scb = &op->file[1].scb;
fs/afs/yfsclient.c
1938
scb = &op->more_files[call->count - 2].scb;
fs/afs/yfsclient.c
1945
if (call->count < op->nr_files)
fs/afs/yfsclient.c
1958
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
1976
.op = yfs_FS_InlineBulkStatus,
fs/afs/yfsclient.c
1984
void yfs_fs_inline_bulk_status(struct afs_operation *op)
fs/afs/yfsclient.c
1986
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
1987
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
1993
key_serial(op->key), vp->fid.vid, vp->fid.vnode, op->nr_files);
fs/afs/yfsclient.c
1995
call = afs_alloc_flat_call(op->net, &yfs_RXYFSInlineBulkStatus,
fs/afs/yfsclient.c
1999
sizeof(struct yfs_xdr_YFSFid) * op->nr_files,
fs/afs/yfsclient.c
2002
return afs_op_nomem(op);
fs/afs/yfsclient.c
2008
bp = xdr_encode_u32(bp, op->nr_files);
fs/afs/yfsclient.c
2011
for (i = 0; i < op->nr_files - 2; i++)
fs/afs/yfsclient.c
2012
bp = xdr_encode_YFSFid(bp, &op->more_files[i].fid);
fs/afs/yfsclient.c
2017
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
2025
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
2026
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
2027
struct yfs_acl *yacl = op->yacl;
fs/afs/yfsclient.c
2118
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
2145
.op = yfs_FS_FetchOpaqueACL,
fs/afs/yfsclient.c
2153
void yfs_fs_fetch_opaque_acl(struct afs_operation *op)
fs/afs/yfsclient.c
2155
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
2160
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
2162
call = afs_alloc_flat_call(op->net, &yfs_RXYFSFetchOpaqueACL,
fs/afs/yfsclient.c
2169
return afs_op_nomem(op);
fs/afs/yfsclient.c
2180
afs_make_op_call(op, call, GFP_KERNEL);
fs/afs/yfsclient.c
2188
.op = yfs_FS_StoreOpaqueACL2,
fs/afs/yfsclient.c
2196
void yfs_fs_store_opaque_acl2(struct afs_operation *op)
fs/afs/yfsclient.c
2198
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
2200
struct afs_acl *acl = op->acl;
fs/afs/yfsclient.c
2205
key_serial(op->key), vp->fid.vid, vp->fid.vnode);
fs/afs/yfsclient.c
2208
call = afs_alloc_flat_call(op->net, &yfs_RXYFSStoreOpaqueACL2,
fs/afs/yfsclient.c
2215
return afs_op_nomem(op);
fs/afs/yfsclient.c
2231
afs_make_op_call(op, call, GFP_KERNEL);
fs/afs/yfsclient.c
333
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
342
xdr_decode_YFSFetchStatus(&bp, call, &op->file[0].scb);
fs/afs/yfsclient.c
343
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
354
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
355
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/yfsclient.c
356
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
437
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
458
.op = yfs_FS_FetchData64,
fs/afs/yfsclient.c
468
void yfs_fs_fetch_data(struct afs_operation *op)
fs/afs/yfsclient.c
470
struct netfs_io_subrequest *subreq = op->fetch.subreq;
fs/afs/yfsclient.c
471
struct afs_vnode_param *vp = &op->file[0];
fs/afs/yfsclient.c
476
key_serial(op->key), vp->fid.vid, vp->fid.vnode,
fs/afs/yfsclient.c
480
call = afs_alloc_flat_call(op->net, &yfs_RXYFSFetchData64,
fs/afs/yfsclient.c
488
return afs_op_nomem(op);
fs/afs/yfsclient.c
490
if (op->flags & AFS_OPERATION_ASYNC)
fs/afs/yfsclient.c
504
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
512
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
513
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
514
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
526
xdr_decode_YFSFid(&bp, &op->file[1].fid);
fs/afs/yfsclient.c
530
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
541
.op = yfs_FS_CreateFile,
fs/afs/yfsclient.c
549
void yfs_fs_create_file(struct afs_operation *op)
fs/afs/yfsclient.c
551
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
552
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
571
call = afs_alloc_flat_call(op->net, &afs_RXFSCreateFile, reqsz, rplsz);
fs/afs/yfsclient.c
573
return afs_op_nomem(op);
fs/afs/yfsclient.c
581
bp = xdr_encode_YFSStoreStatus(bp, &op->create.mode, &op->mtime);
fs/afs/yfsclient.c
587
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
592
.op = yfs_FS_MakeDir,
fs/afs/yfsclient.c
600
void yfs_fs_make_dir(struct afs_operation *op)
fs/afs/yfsclient.c
602
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
603
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
621
call = afs_alloc_flat_call(op->net, &yfs_RXFSMakeDir, reqsz, rplsz);
fs/afs/yfsclient.c
623
return afs_op_nomem(op);
fs/afs/yfsclient.c
631
bp = xdr_encode_YFSStoreStatus(bp, &op->create.mode, &op->mtime);
fs/afs/yfsclient.c
636
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
644
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
645
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
646
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
663
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
672
set_bit(AFS_SERVER_FL_NO_RM2, &call->op->server->flags);
fs/afs/yfsclient.c
673
call->op->flags |= AFS_OPERATION_DOWNGRADE;
fs/afs/yfsclient.c
682
.op = yfs_FS_RemoveFile2,
fs/afs/yfsclient.c
691
void yfs_fs_remove_file2(struct afs_operation *op)
fs/afs/yfsclient.c
693
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
694
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
700
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRemoveFile2,
fs/afs/yfsclient.c
710
return afs_op_nomem(op);
fs/afs/yfsclient.c
722
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
730
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
731
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
743
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
752
.op = yfs_FS_RemoveFile,
fs/afs/yfsclient.c
760
void yfs_fs_remove_file(struct afs_operation *op)
fs/afs/yfsclient.c
762
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
763
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
769
if (!test_bit(AFS_SERVER_FL_NO_RM2, &op->server->flags))
fs/afs/yfsclient.c
770
return yfs_fs_remove_file2(op);
fs/afs/yfsclient.c
772
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRemoveFile,
fs/afs/yfsclient.c
780
return afs_op_nomem(op);
fs/afs/yfsclient.c
792
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
797
.op = yfs_FS_RemoveDir,
fs/afs/yfsclient.c
805
void yfs_fs_remove_dir(struct afs_operation *op)
fs/afs/yfsclient.c
807
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
808
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
814
call = afs_alloc_flat_call(op->net, &yfs_RXYFSRemoveDir,
fs/afs/yfsclient.c
822
return afs_op_nomem(op);
fs/afs/yfsclient.c
834
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
842
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
843
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
844
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
857
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
867
.op = yfs_FS_Link,
fs/afs/yfsclient.c
875
void yfs_fs_link(struct afs_operation *op)
fs/afs/yfsclient.c
877
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
878
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
879
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
885
call = afs_alloc_flat_call(op->net, &yfs_RXYFSLink,
fs/afs/yfsclient.c
895
return afs_op_nomem(op);
fs/afs/yfsclient.c
908
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
916
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
917
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
918
struct afs_vnode_param *vp = &op->file[1];
fs/afs/yfsclient.c
933
xdr_decode_YFSVolSync(&bp, &op->volsync);
fs/afs/yfsclient.c
944
.op = yfs_FS_Symlink,
fs/afs/yfsclient.c
952
void yfs_fs_symlink(struct afs_operation *op)
fs/afs/yfsclient.c
954
const struct qstr *name = &op->dentry->d_name;
fs/afs/yfsclient.c
955
struct afs_vnode_param *dvp = &op->file[0];
fs/afs/yfsclient.c
963
contents_sz = strlen(op->create.symlink);
fs/afs/yfsclient.c
964
call = afs_alloc_flat_call(op->net, &yfs_RXYFSSymlink,
fs/afs/yfsclient.c
976
return afs_op_nomem(op);
fs/afs/yfsclient.c
984
bp = xdr_encode_string(bp, op->create.symlink, contents_sz);
fs/afs/yfsclient.c
985
bp = xdr_encode_YFSStoreStatus(bp, &mode, &op->mtime);
fs/afs/yfsclient.c
990
afs_make_op_call(op, call, GFP_NOFS);
fs/afs/yfsclient.c
998
struct afs_operation *op = call->op;
fs/afs/yfsclient.c
999
struct afs_vnode_param *orig_dvp = &op->file[0];
fs/btrfs/compression.c
62
u64 start, blk_opf_t op,
fs/btrfs/compression.c
67
bbio = btrfs_bio(bio_alloc_bioset(NULL, BTRFS_MAX_COMPRESSED_PAGES, op,
fs/btrfs/delayed-ref.h
353
btrfs_free_delayed_extent_op(struct btrfs_delayed_extent_op *op)
fs/btrfs/delayed-ref.h
355
if (op)
fs/btrfs/delayed-ref.h
356
kmem_cache_free(btrfs_delayed_extent_op_cachep, op);
fs/btrfs/free-space-cache.c
2322
if (!ctl->op->use_bitmap(ctl, info))
fs/btrfs/free-space-cache.c
2325
if (ctl->op == &free_space_op)
fs/btrfs/free-space-cache.c
2956
ctl->op = &free_space_op;
fs/btrfs/free-space-cache.h
88
const struct btrfs_free_space_op *op;
fs/btrfs/fs.c
233
enum btrfs_exclusive_operation op)
fs/btrfs/fs.c
235
switch (op) {
fs/btrfs/fs.c
253
"invalid exclop balance operation %d requested", op);
fs/btrfs/fs.h
1080
enum btrfs_exclusive_operation op);
fs/btrfs/raid56.c
1239
unsigned int sector_nr, enum req_op op)
fs/btrfs/raid56.c
1298
op, GFP_NOFS);
fs/btrfs/relocation.c
3992
blk_opf_t op)
fs/btrfs/relocation.c
4001
bbio = btrfs_bio_alloc(BIO_MAX_VECS, op, BTRFS_I(fs_info->btree_inode),
fs/btrfs/relocation.c
4014
bbio = btrfs_bio_alloc(BIO_MAX_VECS, op,
fs/btrfs/tests/free-space-tests.c
1000
cache->free_space_ctl->op = orig_free_space_ops;
fs/btrfs/tests/free-space-tests.c
426
orig_free_space_ops = cache->free_space_ctl->op;
fs/btrfs/tests/free-space-tests.c
427
cache->free_space_ctl->op = &test_free_space_ops;
fs/btrfs/tests/free-space-tests.c
821
cache->free_space_ctl->op = orig_free_space_ops;
fs/btrfs/tests/free-space-tests.c
895
orig_free_space_ops = cache->free_space_ctl->op;
fs/btrfs/tests/free-space-tests.c
896
cache->free_space_ctl->op = &test_free_space_ops;
fs/btrfs/tree-mod-log.c
1024
if (tm->op == BTRFS_MOD_LOG_ROOT_REPLACE) {
fs/btrfs/tree-mod-log.c
1035
if (old_root && tm && tm->op != BTRFS_MOD_LOG_KEY_REMOVE_WHILE_FREEING) {
fs/btrfs/tree-mod-log.c
1116
if (tm && tm->op == BTRFS_MOD_LOG_ROOT_REPLACE)
fs/btrfs/tree-mod-log.c
19
enum btrfs_mod_log_op op;
fs/btrfs/tree-mod-log.c
238
enum btrfs_mod_log_op op)
fs/btrfs/tree-mod-log.c
243
ASSERT(op != BTRFS_MOD_LOG_MOVE_KEYS);
fs/btrfs/tree-mod-log.c
244
ASSERT(op != BTRFS_MOD_LOG_ROOT_REPLACE);
fs/btrfs/tree-mod-log.c
253
tm->op = op;
fs/btrfs/tree-mod-log.c
262
enum btrfs_mod_log_op op)
fs/btrfs/tree-mod-log.c
270
tm = alloc_tree_mod_elem(eb, slot, op);
fs/btrfs/tree-mod-log.c
312
tm->op = BTRFS_MOD_LOG_MOVE_KEYS;
fs/btrfs/tree-mod-log.c
468
tm->op = BTRFS_MOD_LOG_ROOT_REPLACE;
fs/btrfs/tree-mod-log.c
805
if (tm->op != BTRFS_MOD_LOG_ROOT_REPLACE)
fs/btrfs/tree-mod-log.c
863
switch (tm->op) {
fs/btrfs/tree-mod-log.c
964
if (tm->op == BTRFS_MOD_LOG_KEY_REMOVE_WHILE_FREEING) {
fs/btrfs/tree-mod-log.h
41
enum btrfs_mod_log_op op);
fs/btrfs/volumes.c
49
enum btrfs_map_op op;
fs/btrfs/volumes.c
6608
if (io_geom->op == BTRFS_MAP_GET_READ_MIRRORS && nr_extra_stripes == 2) {
fs/btrfs/volumes.c
6670
if (io_geom->op == BTRFS_MAP_WRITE)
fs/btrfs/volumes.c
6690
if (io_geom->op == BTRFS_MAP_READ && io_geom->use_rst)
fs/btrfs/volumes.c
6713
if (io_geom->use_rst && io_geom->op != BTRFS_MAP_READ)
fs/btrfs/volumes.c
6727
if (io_geom->op == BTRFS_MAP_READ)
fs/btrfs/volumes.c
6736
if (io_geom->op != BTRFS_MAP_READ) {
fs/btrfs/volumes.c
6754
if (io_geom->op != BTRFS_MAP_READ) {
fs/btrfs/volumes.c
6778
if (io_geom->op != BTRFS_MAP_READ) {
fs/btrfs/volumes.c
6840
if (io_geom->op == BTRFS_MAP_READ && io_geom->mirror_num < 1)
fs/btrfs/volumes.c
6887
int btrfs_map_block(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
fs/btrfs/volumes.c
6908
io_geom.op = op;
fs/btrfs/volumes.c
6972
if (op != BTRFS_MAP_READ || io_geom.mirror_num > 1)
fs/btrfs/volumes.c
6996
op != BTRFS_MAP_READ)
fs/btrfs/volumes.c
7035
(op != BTRFS_MAP_READ || io_geom.mirror_num > 1)) {
fs/btrfs/volumes.c
7078
if (op != BTRFS_MAP_READ)
fs/btrfs/volumes.c
7082
op != BTRFS_MAP_READ) {
fs/btrfs/volumes.h
710
int btrfs_map_block(struct btrfs_fs_info *fs_info, enum btrfs_map_op op,
fs/buffer.c
2781
const enum req_op op = opf & REQ_OP_MASK;
fs/buffer.c
2793
if (test_set_buffer_req(bh) && (op == REQ_OP_WRITE))
fs/ceph/addr.c
218
struct ceph_osd_req_op *op = &req->r_ops[0];
fs/ceph/addr.c
220
bool sparse = (op->op == CEPH_OSD_OP_SPARSE_READ);
fs/ceph/addr.c
239
err = ceph_sparse_ext_map_end(op);
fs/ceph/addr.c
247
op->extent.sparse_ext,
fs/ceph/addr.c
248
op->extent.sparse_ext_cnt);
fs/ceph/addr.c
911
if (req->r_ops[i].op != CEPH_OSD_OP_WRITE) {
fs/ceph/addr.c
914
ceph_vinop(inode), req->r_ops[i].op, req, i,
fs/ceph/caps.c
1230
int op, caps, wanted, dirty;
fs/ceph/caps.c
1255
ceph_cap_op_name(arg->op), arg->cid, arg->ino,
fs/ceph/caps.c
1269
fc->op = cpu_to_le32(arg->op);
fs/ceph/caps.c
1392
int op, int flags, int used, int want, int retain,
fs/ceph/caps.c
1455
arg->op = op;
fs/ceph/caps.c
1576
arg.op = CEPH_CAP_OP_FLUSHSNAP;
fs/ceph/caps.c
3683
if ((was_stale || le32_to_cpu(grant->op) == CEPH_CAP_OP_IMPORT) &&
fs/ceph/caps.c
3744
if (!revoke_wait && le32_to_cpu(grant->op) == CEPH_CAP_OP_REVOKE) {
fs/ceph/caps.c
3761
if (le32_to_cpu(grant->op) == CEPH_CAP_OP_IMPORT) {
fs/ceph/caps.c
4339
int op;
fs/ceph/caps.c
4359
op = le32_to_cpu(h->op);
fs/ceph/caps.c
4379
if (op == CEPH_CAP_OP_IMPORT) {
fs/ceph/caps.c
4384
} else if (op == CEPH_CAP_OP_EXPORT) {
fs/ceph/caps.c
4453
session->s_mds, ceph_cap_op_name(op), vino.ino, vino.snap, inode,
fs/ceph/caps.c
4456
trace_ceph_handle_caps(mdsc, session, op, &vino, ceph_inode(inode),
fs/ceph/caps.c
4464
switch (op) {
fs/ceph/caps.c
4478
switch (op) {
fs/ceph/caps.c
4521
switch (op) {
fs/ceph/caps.c
4533
switch (op) {
fs/ceph/caps.c
4557
pr_err_client(cl, "unknown cap op %d %s\n", op,
fs/ceph/caps.c
4558
ceph_cap_op_name(op));
fs/ceph/dir.c
1105
int op;
fs/ceph/dir.c
1113
op = CEPH_MDS_OP_MKSNAP;
fs/ceph/dir.c
1119
op = CEPH_MDS_OP_MKDIR;
fs/ceph/dir.c
1125
if (op == CEPH_MDS_OP_MKDIR &&
fs/ceph/dir.c
1130
if ((op == CEPH_MDS_OP_MKSNAP) && IS_ENCRYPTED(dir) &&
fs/ceph/dir.c
1137
req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS);
fs/ceph/dir.c
1156
if (op == CEPH_MDS_OP_MKDIR)
fs/ceph/dir.c
1347
int op;
fs/ceph/dir.c
1354
op = CEPH_MDS_OP_RMSNAP;
fs/ceph/dir.c
1358
op = d_is_dir(dentry) ?
fs/ceph/dir.c
1388
req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS);
fs/ceph/dir.c
1401
if (try_async && op == CEPH_MDS_OP_UNLINK &&
fs/ceph/dir.c
1476
int op = CEPH_MDS_OP_RENAME;
fs/ceph/dir.c
1486
op = CEPH_MDS_OP_RENAMESNAP;
fs/ceph/dir.c
1507
req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS);
fs/ceph/dir.c
2002
int op, err;
fs/ceph/dir.c
2010
op = ceph_snap(dir) == CEPH_SNAPDIR ?
fs/ceph/dir.c
2012
req = ceph_mdsc_create_request(mdsc, op, USE_ANY_MDS);
fs/ceph/dir.c
388
int op = ceph_snap(inode) == CEPH_SNAPDIR ?
fs/ceph/dir.c
409
req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS);
fs/ceph/dir.c
420
if (op == CEPH_MDS_OP_READDIR) {
fs/ceph/dir.c
802
int op;
fs/ceph/dir.c
850
op = ceph_snap(dir) == CEPH_SNAPDIR ?
fs/ceph/dir.c
852
req = ceph_mdsc_create_request(mdsc, op, USE_ANY_MDS);
fs/ceph/file.c
1085
struct ceph_osd_req_op *op;
fs/ceph/file.c
1113
op = &req->r_ops[0];
fs/ceph/file.c
1116
ret = ceph_alloc_sparse_ext_map(op, extent_cnt);
fs/ceph/file.c
1153
ret = ceph_sparse_ext_map_end(op);
fs/ceph/file.c
1168
read_off, op->extent.sparse_ext,
fs/ceph/file.c
1169
op->extent.sparse_ext_cnt);
fs/ceph/file.c
1331
struct ceph_osd_req_op *op = &req->r_ops[0];
fs/ceph/file.c
1334
bool sparse = (op->op == CEPH_OSD_OP_SPARSE_READ);
fs/ceph/file.c
1358
rc = ceph_sparse_ext_map_end(op);
fs/ceph/file.c
1525
struct ceph_osd_req_op *op;
fs/ceph/file.c
1548
op = &req->r_ops[0];
fs/ceph/file.c
1551
ret = ceph_alloc_sparse_ext_map(op, extent_cnt);
fs/ceph/file.c
1627
ret = ceph_sparse_ext_map_end(op);
fs/ceph/file.c
1791
struct ceph_osd_req_op *op;
fs/ceph/file.c
1825
op = &req->r_ops[0];
fs/ceph/file.c
1833
ret = __ceph_alloc_sparse_ext_map(op, 1);
fs/ceph/file.c
1845
op = &req->r_ops[1];
fs/ceph/file.c
185
int op = (flags & O_CREAT) ? CEPH_MDS_OP_CREATE : CEPH_MDS_OP_OPEN;
fs/ceph/file.c
1853
ret = __ceph_alloc_sparse_ext_map(op, 1);
fs/ceph/file.c
190
req = ceph_mdsc_create_request(mdsc, op, want_auth);
fs/ceph/file.c
1904
op = &req->r_ops[0];
fs/ceph/file.c
1905
if (op->extent.sparse_ext_cnt == 0) {
fs/ceph/file.c
1913
} else if (op->extent.sparse_ext_cnt != 1 ||
fs/ceph/file.c
1914
ceph_sparse_ext_map_end(op) !=
fs/ceph/file.c
1923
op = &req->r_ops[1];
fs/ceph/file.c
1924
if (op->extent.sparse_ext_cnt == 0) {
fs/ceph/file.c
1928
} else if (op->extent.sparse_ext_cnt != 1 ||
fs/ceph/file.c
1929
ceph_sparse_ext_map_end(op) !=
fs/ceph/file.c
2574
int op;
fs/ceph/file.c
2580
op = offset ? CEPH_OSD_OP_DELETE : CEPH_OSD_OP_TRUNCATE;
fs/ceph/file.c
2583
op = CEPH_OSD_OP_ZERO;
fs/ceph/file.c
2602
0, 1, op,
fs/ceph/inode.c
2000
if (le32_to_cpu(rinfo->head->op) == CEPH_MDS_OP_LSSNAP) {
fs/ceph/locks.c
256
u16 op = CEPH_MDS_OP_SETFILELOCK;
fs/ceph/locks.c
270
op = CEPH_MDS_OP_GETFILELOCK;
fs/ceph/locks.c
280
if (op == CEPH_MDS_OP_SETFILELOCK && lock_is_unlock(fl))
fs/ceph/locks.c
292
if (op == CEPH_MDS_OP_SETFILELOCK && lock_is_unlock(fl)) {
fs/ceph/locks.c
298
err = ceph_lock_message(CEPH_LOCK_FCNTL, op, inode, lock_cmd, wait, fl);
fs/ceph/locks.c
300
if (op == CEPH_MDS_OP_SETFILELOCK && F_UNLCK != fl->c.flc_type) {
fs/ceph/locks.c
307
ceph_lock_message(CEPH_LOCK_FCNTL, op, inode,
fs/ceph/mds_client.c
1443
struct ceph_msg *ceph_create_session_msg(u32 op, u64 seq)
fs/ceph/mds_client.c
1452
ceph_session_op_name(op));
fs/ceph/mds_client.c
1456
h->op = cpu_to_le32(op);
fs/ceph/mds_client.c
1540
create_session_full_msg(struct ceph_mds_client *mdsc, int op, u64 seq)
fs/ceph/mds_client.c
1598
h->op = cpu_to_le32(op);
fs/ceph/mds_client.c
2577
ceph_mdsc_create_request(struct ceph_mds_client *mdsc, int op, int mode)
fs/ceph/mds_client.c
2603
req->r_op = op;
fs/ceph/mds_client.c
3203
lhead->op = cpu_to_le32(req->r_op);
fs/ceph/mds_client.c
3988
le32_to_cpu(head->op) == CEPH_MDS_OP_RMSNAP,
fs/ceph/mds_client.c
4177
u32 op, cap_auths_num = 0;
fs/ceph/mds_client.c
4189
op = le32_to_cpu(h->op);
fs/ceph/mds_client.c
4231
if (cap_auths_num && op != CEPH_SESSION_OPEN) {
fs/ceph/mds_client.c
4232
WARN_ON_ONCE(op != CEPH_SESSION_OPEN);
fs/ceph/mds_client.c
4309
if (op == CEPH_SESSION_OPEN) {
fs/ceph/mds_client.c
4321
if (op == CEPH_SESSION_CLOSE) {
fs/ceph/mds_client.c
4332
ceph_session_op_name(op), session,
fs/ceph/mds_client.c
4340
switch (op) {
fs/ceph/mds_client.c
4430
pr_err_client(cl, "bad op %d mds%d\n", op, mds);
fs/ceph/mds_client.c
4442
if (op == CEPH_SESSION_CLOSE)
fs/ceph/mds_client.c
731
u32 op = le32_to_cpu(info->head->op);
fs/ceph/mds_client.c
733
if (op == CEPH_MDS_OP_GETFILELOCK)
fs/ceph/mds_client.c
735
else if (op == CEPH_MDS_OP_READDIR || op == CEPH_MDS_OP_LSSNAP)
fs/ceph/mds_client.c
737
else if (op == CEPH_MDS_OP_CREATE)
fs/ceph/mds_client.c
739
else if (op == CEPH_MDS_OP_GETVXATTR)
fs/ceph/mds_client.h
553
extern const char *ceph_mds_op_name(int op);
fs/ceph/mds_client.h
578
ceph_mdsc_create_request(struct ceph_mds_client *mdsc, int op, int mode);
fs/ceph/mds_client.h
604
extern struct ceph_msg *ceph_create_session_msg(u32 op, u64 seq);
fs/ceph/quota.c
326
static bool check_quota_exceeded(struct inode *inode, enum quota_check_op op,
fs/ceph/quota.c
369
if (op == QUOTA_CHECK_MAX_FILES_OP) {
fs/ceph/quota.c
377
switch (op) {
fs/ceph/quota.c
398
pr_warn_client(cl, "Invalid quota check op (%d)\n", op);
fs/ceph/snap.c
1024
int op;
fs/ceph/snap.c
1043
op = le32_to_cpu(h->op);
fs/ceph/snap.c
1052
ceph_snap_op_name(op), split, trace_len);
fs/ceph/snap.c
1057
if (op == CEPH_SNAP_OP_SPLIT) {
fs/ceph/snap.c
1156
op == CEPH_SNAP_OP_DESTROY,
fs/ceph/snap.c
1162
if (op == CEPH_SNAP_OP_SPLIT)
fs/ceph/strings.c
34
const char *ceph_session_op_name(int op)
fs/ceph/strings.c
36
switch (op) {
fs/ceph/strings.c
54
const char *ceph_mds_op_name(int op)
fs/ceph/strings.c
56
switch (op) {
fs/ceph/strings.c
90
const char *ceph_cap_op_name(int op)
fs/ceph/strings.c
92
switch (op) {
fs/ceph/xattr.c
1116
int op = CEPH_MDS_OP_SETXATTR;
fs/ceph/xattr.c
1130
op = CEPH_MDS_OP_RMXATTR;
fs/ceph/xattr.c
1138
req = ceph_mdsc_create_request(mdsc, op, USE_AUTH_MDS);
fs/ceph/xattr.c
1151
if (op == CEPH_MDS_OP_SETXATTR) {
fs/coda/upcall.c
61
#define UPARG(op)\
fs/coda/upcall.c
63
inp = (union inputArgs *)alloc_upcall(op, insize); \
fs/dcache.c
1896
static unsigned int d_op_flags(const struct dentry_operations *op)
fs/dcache.c
1899
if (op) {
fs/dcache.c
1900
if (op->d_hash)
fs/dcache.c
1902
if (op->d_compare)
fs/dcache.c
1904
if (op->d_revalidate)
fs/dcache.c
1906
if (op->d_weak_revalidate)
fs/dcache.c
1908
if (op->d_delete)
fs/dcache.c
1910
if (op->d_prune)
fs/dcache.c
1912
if (op->d_real)
fs/dcache.c
1918
static void d_set_d_op(struct dentry *dentry, const struct dentry_operations *op)
fs/dcache.c
1920
unsigned int flags = d_op_flags(op);
fs/dcache.c
1923
dentry->d_op = op;
fs/dlm/debug_fs.c
385
if (seq->op == &format1_seq_ops)
fs/dlm/debug_fs.c
387
else if (seq->op == &format2_seq_ops)
fs/dlm/debug_fs.c
389
else if (seq->op == &format3_seq_ops)
fs/dlm/debug_fs.c
391
else if (seq->op == &format4_seq_ops)
fs/dlm/debug_fs.c
403
if (seq->op == &format2_seq_ops)
fs/dlm/debug_fs.c
405
else if (seq->op == &format3_seq_ops)
fs/dlm/debug_fs.c
407
else if (seq->op == &format4_seq_ops)
fs/dlm/debug_fs.c
411
if (seq->op == &format4_seq_ops)
fs/dlm/debug_fs.c
425
if (seq->op == &format4_seq_ops)
fs/dlm/plock.c
102
struct plock_op *op;
fs/dlm/plock.c
105
op = kzalloc_obj(*op, GFP_NOFS);
fs/dlm/plock.c
106
if (!op)
fs/dlm/plock.c
109
op->info = *orig_info;
fs/dlm/plock.c
110
op->info.optype = DLM_PLOCK_OP_CANCEL;
fs/dlm/plock.c
111
op->info.wait = 0;
fs/dlm/plock.c
113
send_op(op);
fs/dlm/plock.c
114
wait_event(recv_wq, (op->done != 0));
fs/dlm/plock.c
116
rv = op->info.rv;
fs/dlm/plock.c
118
dlm_release_plock_op(op);
fs/dlm/plock.c
127
struct plock_op *op;
fs/dlm/plock.c
134
op = kzalloc_obj(*op, GFP_NOFS);
fs/dlm/plock.c
135
if (!op) {
fs/dlm/plock.c
140
op->info.optype = DLM_PLOCK_OP_LOCK;
fs/dlm/plock.c
141
op->info.pid = fl->c.flc_pid;
fs/dlm/plock.c
142
op->info.ex = lock_is_write(fl);
fs/dlm/plock.c
143
op->info.wait = !!(fl->c.flc_flags & FL_SLEEP);
fs/dlm/plock.c
144
op->info.fsid = ls->ls_global_id;
fs/dlm/plock.c
145
op->info.number = number;
fs/dlm/plock.c
146
op->info.start = fl->fl_start;
fs/dlm/plock.c
147
op->info.end = fl->fl_end;
fs/dlm/plock.c
148
op->info.owner = (__u64)(long) fl->c.flc_owner;
fs/dlm/plock.c
153
dlm_release_plock_op(op);
fs/dlm/plock.c
164
op->data = op_data;
fs/dlm/plock.c
166
send_op(op);
fs/dlm/plock.c
171
send_op(op);
fs/dlm/plock.c
173
if (op->info.wait) {
fs/dlm/plock.c
174
rv = wait_event_interruptible(recv_wq, (op->done != 0));
fs/dlm/plock.c
180
if (op->done != 0) {
fs/dlm/plock.c
186
rv = do_lock_cancel(&op->info);
fs/dlm/plock.c
195
list_del(&op->list);
fs/dlm/plock.c
209
(unsigned long long)number, op->info.pid);
fs/dlm/plock.c
210
dlm_release_plock_op(op);
fs/dlm/plock.c
215
wait_event(recv_wq, (op->done != 0));
fs/dlm/plock.c
220
WARN_ON(!list_empty(&op->list));
fs/dlm/plock.c
222
rv = op->info.rv;
fs/dlm/plock.c
230
dlm_release_plock_op(op);
fs/dlm/plock.c
238
static int dlm_plock_callback(struct plock_op *op)
fs/dlm/plock.c
240
struct plock_async_data *op_data = op->data;
fs/dlm/plock.c
247
WARN_ON(!list_empty(&op->list));
fs/dlm/plock.c
255
if (op->info.rv) {
fs/dlm/plock.c
256
notify(fl, op->info.rv);
fs/dlm/plock.c
272
(unsigned long long)op->info.number, file, fl);
fs/dlm/plock.c
284
dlm_release_plock_op(op);
fs/dlm/plock.c
292
struct plock_op *op;
fs/dlm/plock.c
300
op = kzalloc_obj(*op, GFP_NOFS);
fs/dlm/plock.c
301
if (!op) {
fs/dlm/plock.c
319
op->info.optype = DLM_PLOCK_OP_UNLOCK;
fs/dlm/plock.c
320
op->info.pid = fl->c.flc_pid;
fs/dlm/plock.c
321
op->info.fsid = ls->ls_global_id;
fs/dlm/plock.c
322
op->info.number = number;
fs/dlm/plock.c
323
op->info.start = fl->fl_start;
fs/dlm/plock.c
324
op->info.end = fl->fl_end;
fs/dlm/plock.c
325
op->info.owner = (__u64)(long) fl->c.flc_owner;
fs/dlm/plock.c
328
op->info.flags |= DLM_PLOCK_FL_CLOSE;
fs/dlm/plock.c
329
send_op(op);
fs/dlm/plock.c
334
send_op(op);
fs/dlm/plock.c
335
wait_event(recv_wq, (op->done != 0));
fs/dlm/plock.c
337
WARN_ON(!list_empty(&op->list));
fs/dlm/plock.c
339
rv = op->info.rv;
fs/dlm/plock.c
345
dlm_release_plock_op(op);
fs/dlm/plock.c
363
struct plock_op *op;
fs/dlm/plock.c
394
op = plock_lookup_waiter(&info);
fs/dlm/plock.c
395
if (WARN_ON_ONCE(!op)) {
fs/dlm/plock.c
401
list_del(&op->list);
fs/dlm/plock.c
403
WARN_ON(op->info.optype != DLM_PLOCK_OP_LOCK);
fs/dlm/plock.c
404
op->data->callback(op->data->fl, -EINTR);
fs/dlm/plock.c
405
dlm_release_plock_op(op);
fs/dlm/plock.c
426
struct plock_op *op;
fs/dlm/plock.c
433
op = kzalloc_obj(*op, GFP_NOFS);
fs/dlm/plock.c
434
if (!op) {
fs/dlm/plock.c
439
op->info.optype = DLM_PLOCK_OP_GET;
fs/dlm/plock.c
440
op->info.pid = fl->c.flc_pid;
fs/dlm/plock.c
441
op->info.ex = lock_is_write(fl);
fs/dlm/plock.c
442
op->info.fsid = ls->ls_global_id;
fs/dlm/plock.c
443
op->info.number = number;
fs/dlm/plock.c
444
op->info.start = fl->fl_start;
fs/dlm/plock.c
445
op->info.end = fl->fl_end;
fs/dlm/plock.c
446
op->info.owner = (__u64)(long) fl->c.flc_owner;
fs/dlm/plock.c
448
send_op(op);
fs/dlm/plock.c
449
wait_event(recv_wq, (op->done != 0));
fs/dlm/plock.c
451
WARN_ON(!list_empty(&op->list));
fs/dlm/plock.c
456
rv = op->info.rv;
fs/dlm/plock.c
463
fl->c.flc_type = (op->info.ex) ? F_WRLCK : F_RDLCK;
fs/dlm/plock.c
465
fl->c.flc_pid = op->info.pid;
fs/dlm/plock.c
466
if (op->info.nodeid != dlm_our_nodeid())
fs/dlm/plock.c
468
fl->fl_start = op->info.start;
fs/dlm/plock.c
469
fl->fl_end = op->info.end;
fs/dlm/plock.c
473
dlm_release_plock_op(op);
fs/dlm/plock.c
485
struct plock_op *op = NULL;
fs/dlm/plock.c
49
struct plock_op *op = NULL, *iter;
fs/dlm/plock.c
492
op = list_first_entry(&send_list, struct plock_op, list);
fs/dlm/plock.c
493
if (op->info.flags & DLM_PLOCK_FL_CLOSE)
fs/dlm/plock.c
494
list_del(&op->list);
fs/dlm/plock.c
496
list_move_tail(&op->list, &recv_list);
fs/dlm/plock.c
497
memcpy(&info, &op->info, sizeof(info));
fs/dlm/plock.c
501
if (!op)
fs/dlm/plock.c
510
if (op->info.flags & DLM_PLOCK_FL_CLOSE)
fs/dlm/plock.c
511
dlm_release_plock_op(op);
fs/dlm/plock.c
523
struct plock_op *op = NULL, *iter;
fs/dlm/plock.c
546
op = plock_lookup_waiter(&info);
fs/dlm/plock.c
551
op = iter;
fs/dlm/plock.c
557
if (op) {
fs/dlm/plock.c
560
WARN_ON(op->info.optype != DLM_PLOCK_OP_LOCK);
fs/dlm/plock.c
562
WARN_ON(op->info.number != info.number ||
fs/dlm/plock.c
563
op->info.owner != info.owner ||
fs/dlm/plock.c
564
op->info.optype != info.optype);
fs/dlm/plock.c
566
list_del_init(&op->list);
fs/dlm/plock.c
567
memcpy(&op->info, &info, sizeof(info));
fs/dlm/plock.c
568
if (op->data)
fs/dlm/plock.c
571
op->done = 1;
fs/dlm/plock.c
575
if (op) {
fs/dlm/plock.c
577
dlm_plock_callback(op);
fs/dlm/plock.c
60
op = iter;
fs/dlm/plock.c
65
return op;
fs/dlm/plock.c
85
static void dlm_release_plock_op(struct plock_op *op)
fs/dlm/plock.c
87
kfree(op->data);
fs/dlm/plock.c
88
kfree(op);
fs/dlm/plock.c
91
static void send_op(struct plock_op *op)
fs/dlm/plock.c
93
set_version(&op->info);
fs/dlm/plock.c
95
list_add_tail(&op->list, &send_list);
fs/ecryptfs/crypto.c
224
unsigned char *iv, int op)
fs/ecryptfs/crypto.c
264
rc = op == ENCRYPT ? crypto_skcipher_encrypt(req) :
fs/ecryptfs/crypto.c
302
unsigned long extent_offset, int op)
fs/ecryptfs/crypto.c
322
extent_iv, op);
fs/eventpoll.c
2236
int do_epoll_ctl(int epfd, int op, int fd, struct epoll_event *epds,
fs/eventpoll.c
2259
if (ep_op_has_event(op))
fs/eventpoll.c
2276
if (ep_op_has_event(op) && (epds->events & EPOLLEXCLUSIVE)) {
fs/eventpoll.c
2277
if (op == EPOLL_CTL_MOD)
fs/eventpoll.c
2279
if (op == EPOLL_CTL_ADD && (is_file_epoll(fd_file(tf)) ||
fs/eventpoll.c
2308
if (op == EPOLL_CTL_ADD) {
fs/eventpoll.c
2337
switch (op) {
fs/eventpoll.c
2383
SYSCALL_DEFINE4(epoll_ctl, int, epfd, int, op, int, fd,
fs/eventpoll.c
2388
if (ep_op_has_event(op) &&
fs/eventpoll.c
2392
return do_epoll_ctl(epfd, op, fd, &epds, false);
fs/ext4/fast_commit.c
392
int op;
fs/ext4/fast_commit.c
425
node->fcd_op = dentry_update->op;
fs/ext4/fast_commit.c
448
if (dentry_update->op == EXT4_FC_TAG_CREAT) {
fs/ext4/fast_commit.c
465
args.op = EXT4_FC_TAG_UNLINK;
fs/ext4/fast_commit.c
492
args.op = EXT4_FC_TAG_LINK;
fs/ext4/fast_commit.c
519
args.op = EXT4_FC_TAG_CREAT;
fs/f2fs/checkpoint.c
271
.op = REQ_OP_READ,
fs/f2fs/checkpoint.c
465
.op = REQ_OP_READ,
fs/f2fs/compress.c
1279
.op = REQ_OP_WRITE,
fs/f2fs/data.c
1012
f2fs_bug_on(sbi, is_read_io(fio->op));
fs/f2fs/data.c
2845
if (IS_F2FS_IPU_ASYNC(sbi) && fio && fio->op == REQ_OP_WRITE &&
fs/f2fs/data.c
3075
.op = REQ_OP_WRITE,
fs/f2fs/data.c
461
if (fio->op != REQ_OP_WRITE)
fs/f2fs/data.c
500
fio->op | fio->op_flags | f2fs_io_flags(fio),
fs/f2fs/data.c
503
if (is_read_io(fio->op)) {
fs/f2fs/data.c
576
if (is_read_io(fio->op)) {
fs/f2fs/data.c
768
if (fio->io_wbc && !is_read_io(fio->op))
fs/f2fs/data.c
771
inc_page_count(fio->sbi, is_read_io(fio->op) ?
fs/f2fs/data.c
797
if (io->fio.op != fio->op)
fs/f2fs/f2fs.h
1342
enum req_op op; /* contains REQ_OP_ */
fs/f2fs/gc.c
1228
.op = REQ_OP_READ,
fs/f2fs/gc.c
1317
.op = REQ_OP_READ,
fs/f2fs/gc.c
1443
fio.op = REQ_OP_WRITE;
fs/f2fs/gc.c
1504
.op = REQ_OP_WRITE,
fs/f2fs/inline.c
151
.op = REQ_OP_WRITE,
fs/f2fs/node.c
1454
.op = REQ_OP_READ,
fs/f2fs/node.c
1743
.op = REQ_OP_WRITE,
fs/f2fs/segment.c
4012
.op = REQ_OP_WRITE,
fs/fat/dir.c
149
unsigned char *op;
fs/fat/dir.c
153
op = ascii;
fs/fat/dir.c
157
charlen = nls->uni2char(ec, op, NLS_MAX_CHARSET_SIZE);
fs/fat/dir.c
159
op += charlen;
fs/fat/dir.c
163
*op++ = ':';
fs/fat/dir.c
164
op = hex_byte_pack(op, ec >> 8);
fs/fat/dir.c
165
op = hex_byte_pack(op, ec);
fs/fat/dir.c
168
*op++ = '?';
fs/fat/dir.c
179
*op = 0;
fs/fat/dir.c
180
return op - ascii;
fs/fat/namei_vfat.c
514
unsigned char *op;
fs/fat/namei_vfat.c
526
op = &outname[*outlen * sizeof(wchar_t)];
fs/fat/namei_vfat.c
528
for (i = 0, ip = name, op = outname, *outlen = 0;
fs/fat/namei_vfat.c
540
*(wchar_t *)op = uc[0] << 8 | uc[1];
fs/fat/namei_vfat.c
542
op += 2;
fs/fat/namei_vfat.c
547
(wchar_t *)op);
fs/fat/namei_vfat.c
552
op += 2;
fs/fat/namei_vfat.c
561
*op++ = 0;
fs/fat/namei_vfat.c
562
*op++ = 0;
fs/fat/namei_vfat.c
567
*op++ = 0xff;
fs/fat/namei_vfat.c
568
*op++ = 0xff;
fs/freevxfs/vxfs_olt.c
57
struct vxfs_olt *op;
fs/freevxfs/vxfs_olt.c
64
op = (struct vxfs_olt *)bp->b_data;
fs/freevxfs/vxfs_olt.c
65
if (fs32_to_cpu(infp, op->olt_magic) != VXFS_OLT_MAGIC) {
fs/freevxfs/vxfs_olt.c
80
oaddr = bp->b_data + fs32_to_cpu(infp, op->olt_size);
fs/hfsplus/unicode.c
156
u8 *op;
fs/hfsplus/unicode.c
161
op = astr;
fs/hfsplus/unicode.c
228
res = nls->uni2char(c0, op, len);
fs/hfsplus/unicode.c
232
*op = '?';
fs/hfsplus/unicode.c
235
op += res;
fs/hfsplus/unicode.c
271
res = nls->uni2char(cc, op, len);
fs/hfsplus/unicode.c
275
*op = '?';
fs/hfsplus/unicode.c
278
op += res;
fs/hfsplus/unicode.c
283
*len_p = (char *)op - astr;
fs/inode.c
1925
const struct super_operations *op = inode->i_sb->s_op;
fs/inode.c
1931
if (op->drop_inode)
fs/inode.c
1932
drop = op->drop_inode(inode);
fs/inode.c
825
const struct super_operations *op = inode->i_sb->s_op;
fs/inode.c
845
if (op->evict_inode) {
fs/inode.c
846
op->evict_inode(inode);
fs/internal.h
194
const struct open_flags *op);
fs/internal.h
198
extern int build_open_flags(const struct open_how *how, struct open_flags *op);
fs/iomap/direct-io.c
333
blk_opf_t op)
fs/iomap/direct-io.c
340
nr_vecs = bio_iov_bounce_nr_vecs(dio->submit.iter, op);
fs/iomap/direct-io.c
344
bio = iomap_dio_alloc_bio(iter, dio, nr_vecs, op);
fs/iomap/direct-io.c
366
if ((op & REQ_ATOMIC) && WARN_ON_ONCE(ret != iomap_length(iter))) {
fs/isofs/joliet.c
21
unsigned char *op;
fs/isofs/joliet.c
24
op = ascii;
fs/isofs/joliet.c
28
llen = nls->uni2char(be16_to_cpu(ch), op, NLS_MAX_CHARSET_SIZE);
fs/isofs/joliet.c
30
op += llen;
fs/isofs/joliet.c
32
*op++ = '?';
fs/isofs/joliet.c
37
*op = 0;
fs/isofs/joliet.c
38
return (op - ascii);
fs/namei.c
4385
const struct open_flags *op,
fs/namei.c
4391
int open_flag = op->open_flag;
fs/namei.c
4394
umode_t mode = op->mode;
fs/namei.c
4542
struct file *file, const struct open_flags *op)
fs/namei.c
4546
int open_flag = op->open_flag;
fs/namei.c
4551
nd->flags |= op->intent;
fs/namei.c
4589
dentry = lookup_open(nd, file, op, got_write, &delegated_inode);
fs/namei.c
4634
struct file *file, const struct open_flags *op)
fs/namei.c
4637
int open_flag = op->open_flag;
fs/namei.c
4664
acc_mode = op->acc_mode;
fs/namei.c
4679
error = security_file_post_open(file, op->acc_mode);
fs/namei.c
4782
const struct open_flags *op,
fs/namei.c
4793
error = vfs_tmpfile(mnt_idmap(path.mnt), &path, file, op->mode);
fs/namei.c
4817
const struct open_flags *op, unsigned flags)
fs/namei.c
4822
file = alloc_empty_file(op->open_flag, current_cred());
fs/namei.c
4827
error = do_tmpfile(nd, flags, op, file);
fs/namei.c
4833
(s = open_last_lookups(nd, file, op)) != NULL)
fs/namei.c
4836
error = do_open(nd, file, op);
fs/namei.c
4856
const struct open_flags *op)
fs/namei.c
4859
int flags = op->lookup_flags;
fs/namei.c
4865
filp = path_openat(&nd, op, flags | LOOKUP_RCU);
fs/namei.c
4867
filp = path_openat(&nd, op, flags);
fs/namei.c
4869
filp = path_openat(&nd, op, flags | LOOKUP_REVAL);
fs/namei.c
4875
const char *name, const struct open_flags *op)
fs/namei.c
4879
int flags = op->lookup_flags;
fs/namei.c
4881
if (d_is_symlink(root->dentry) && op->intent & LOOKUP_OPEN)
fs/namei.c
4889
file = path_openat(&nd, op, flags | LOOKUP_RCU);
fs/namei.c
4891
file = path_openat(&nd, op, flags);
fs/namei.c
4893
file = path_openat(&nd, op, flags | LOOKUP_REVAL);
fs/nfs/blocklayout/blocklayout.c
124
do_add_page_to_bio(struct bio *bio, int npg, enum req_op op, sector_t isect,
fs/nfs/blocklayout/blocklayout.c
134
npg, (__force u32)op, (unsigned long long)isect, offset, *len);
fs/nfs/blocklayout/blocklayout.c
157
bio = bio_alloc(map->bdev, bio_max_segs(npg), op, GFP_NOIO);
fs/nfs/callback_xdr.c
167
static __be32 decode_op_hdr(struct xdr_stream *xdr, unsigned int *op)
fs/nfs/callback_xdr.c
173
*op = ntohl(*p);
fs/nfs/callback_xdr.c
692
static __be32 encode_op_hdr(struct xdr_stream *xdr, uint32_t op, __be32 res)
fs/nfs/callback_xdr.c
699
*p++ = htonl(op);
fs/nfs/callback_xdr.c
784
preprocess_nfs41_op(int nop, unsigned int op_nr, struct callback_op **op)
fs/nfs/callback_xdr.c
803
*op = &callback_ops[op_nr];
fs/nfs/callback_xdr.c
843
preprocess_nfs42_op(int nop, unsigned int op_nr, struct callback_op **op)
fs/nfs/callback_xdr.c
845
__be32 status = preprocess_nfs41_op(nop, op_nr, op);
fs/nfs/callback_xdr.c
850
*op = &callback_ops[op_nr];
fs/nfs/callback_xdr.c
858
preprocess_nfs42_op(int nop, unsigned int op_nr, struct callback_op **op)
fs/nfs/callback_xdr.c
865
preprocess_nfs4_op(unsigned int op_nr, struct callback_op **op)
fs/nfs/callback_xdr.c
870
*op = &callback_ops[op_nr];
fs/nfs/callback_xdr.c
883
struct callback_op *op = &callback_ops[0];
fs/nfs/callback_xdr.c
895
status = preprocess_nfs4_op(op_nr, &op);
fs/nfs/callback_xdr.c
898
status = preprocess_nfs41_op(nop, op_nr, &op);
fs/nfs/callback_xdr.c
901
status = preprocess_nfs42_op(nop, op_nr, &op);
fs/nfs/callback_xdr.c
919
status = op->decode_args(rqstp, &rqstp->rq_arg_stream,
fs/nfs/callback_xdr.c
922
status = op->process_op(rqstp->rq_argp, rqstp->rq_resp,
fs/nfs/callback_xdr.c
931
if (op->encode_res != NULL && status == 0)
fs/nfs/callback_xdr.c
932
status = op->encode_res(rqstp, xdr_out, rqstp->rq_resp);
fs/nfs/nfs4trace.h
438
u32 op,
fs/nfs/nfs4trace.h
442
TP_ARGS(xdr, op, expected),
fs/nfs/nfs4trace.h
448
__field(u32, op)
fs/nfs/nfs4trace.h
459
__entry->op = op;
fs/nfs/nfs4trace.h
466
__entry->op, __entry->expected
fs/nfs/nfs4trace.h
473
u32 op,
fs/nfs/nfs4trace.h
477
TP_ARGS(xdr, op, error),
fs/nfs/nfs4trace.h
483
__field(u32, op)
fs/nfs/nfs4trace.h
494
__entry->op = op;
fs/nfs/nfs4trace.h
502
__entry->op
fs/nfs/nfs4trace.h
509
u32 op, \
fs/nfs/nfs4trace.h
512
TP_ARGS(xdr, op, error))
fs/nfs/nfs4xdr.c
1050
static void encode_op_hdr(struct xdr_stream *xdr, enum nfs_opnum4 op,
fs/nfs/nfs4xdr.c
1054
encode_uint32(xdr, op);
fs/nfsd/nfs4callback.c
127
static void encode_nfs_cb_opnum4(struct xdr_stream *xdr, enum nfs_cb_opnum4 op)
fs/nfsd/nfs4callback.c
132
*p = cpu_to_be32(op);
fs/nfsd/nfs4callback.c
1756
const struct nfsd4_callback_ops *ops, enum nfsd4_cb_op op)
fs/nfsd/nfs4callback.c
1759
cb->cb_msg.rpc_proc = &nfs4_cb_procedures[op];
fs/nfsd/nfs4callback.c
246
u32 op;
fs/nfsd/nfs4callback.c
251
op = be32_to_cpup(p++);
fs/nfsd/nfs4callback.c
252
if (unlikely(op != expected))
fs/nfsd/nfs4callback.c
260
"we issued a request for %d\n", op, expected);
fs/nfsd/nfs4proc.c
2949
const struct nfsd4_operation *OPDESC(struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
2951
return &nfsd4_ops[op->opnum];
fs/nfsd/nfs4proc.c
2954
bool nfsd4_cache_this_op(struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
2956
if (op->opnum == OP_ILLEGAL)
fs/nfsd/nfs4proc.c
2958
return OPDESC(op)->op_flags & OP_CACHEME;
fs/nfsd/nfs4proc.c
2998
struct nfsd4_op *op, *current_op = NULL, *saved_op = NULL;
fs/nfsd/nfs4proc.c
3007
op = &args->ops[i];
fs/nfsd/nfs4proc.c
3008
if (op->opnum == OP_PUTFH)
fs/nfsd/nfs4proc.c
3009
current_op = op;
fs/nfsd/nfs4proc.c
3010
else if (op->opnum == OP_SAVEFH)
fs/nfsd/nfs4proc.c
3012
else if (op->opnum == OP_RESTOREFH)
fs/nfsd/nfs4proc.c
3014
else if (op->opnum == OP_COPY) {
fs/nfsd/nfs4proc.c
3015
copy = (struct nfsd4_copy *)&op->u;
fs/nfsd/nfs4proc.c
3017
op->status = nfserr_nofilehandle;
fs/nfsd/nfs4proc.c
3041
struct nfsd4_op *op;
fs/nfsd/nfs4proc.c
3077
op = &args->ops[0];
fs/nfsd/nfs4proc.c
3078
op->status = status;
fs/nfsd/nfs4proc.c
3088
op = &args->ops[resp->opcnt++];
fs/nfsd/nfs4proc.c
3095
op->status = nfserr_resource;
fs/nfsd/nfs4proc.c
3105
if (op->status) {
fs/nfsd/nfs4proc.c
3106
if (op->opnum == OP_OPEN)
fs/nfsd/nfs4proc.c
3107
op->status = nfsd4_open_omfg(rqstp, cstate, op);
fs/nfsd/nfs4proc.c
3112
if (!(op->opdesc->op_flags & ALLOWED_WITHOUT_FH)) {
fs/nfsd/nfs4proc.c
3113
op->status = nfserr_nofilehandle;
fs/nfsd/nfs4proc.c
3118
!(op->opdesc->op_flags & ALLOWED_ON_ABSENT_FS)) {
fs/nfsd/nfs4proc.c
3119
op->status = nfserr_moved;
fs/nfsd/nfs4proc.c
3126
if (op->opdesc->op_flags & OP_MODIFIES_SOMETHING) {
fs/nfsd/nfs4proc.c
3131
u32 plen = op->opdesc->op_rsize_bop(rqstp, op);
fs/nfsd/nfs4proc.c
3138
op->status = nfsd4_check_resp_size(resp, plen);
fs/nfsd/nfs4proc.c
3141
if (op->status)
fs/nfsd/nfs4proc.c
3144
if (op->opdesc->op_get_currentstateid)
fs/nfsd/nfs4proc.c
3145
op->opdesc->op_get_currentstateid(cstate, &op->u);
fs/nfsd/nfs4proc.c
3146
op->status = op->opdesc->op_func(rqstp, cstate, &op->u);
fs/nfsd/nfs4proc.c
3147
trace_nfsd_compound_op_err(rqstp, op->opnum, op->status);
fs/nfsd/nfs4proc.c
3152
status = op->status;
fs/nfsd/nfs4proc.c
3155
if (!op->status) {
fs/nfsd/nfs4proc.c
3156
if (op->opdesc->op_set_currentstateid)
fs/nfsd/nfs4proc.c
3157
op->opdesc->op_set_currentstateid(cstate, &op->u);
fs/nfsd/nfs4proc.c
3159
if (op->opdesc->op_flags & OP_CLEAR_STATEID)
fs/nfsd/nfs4proc.c
3164
op->status = check_nfsd_access(current_fh->fh_export, rqstp, false);
fs/nfsd/nfs4proc.c
3167
if (op->status == nfserr_replay_me) {
fs/nfsd/nfs4proc.c
3168
op->replay = &cstate->replay_owner->so_replay;
fs/nfsd/nfs4proc.c
3169
nfsd4_encode_replay(resp->xdr, op);
fs/nfsd/nfs4proc.c
3170
status = op->status = op->replay->rp_status;
fs/nfsd/nfs4proc.c
3172
nfsd4_encode_operation(resp, op);
fs/nfsd/nfs4proc.c
3173
status = op->status;
fs/nfsd/nfs4proc.c
3177
status, nfsd4_op_name(op->opnum));
fs/nfsd/nfs4proc.c
3180
nfsd4_increment_op_stats(nn, op->opnum);
fs/nfsd/nfs4proc.c
3226
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3232
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3238
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3245
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3251
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3263
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3265
const u32 *bmap = op->u.getattr.ga_bmval;
fs/nfsd/nfs4proc.c
3301
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3307
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3314
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3321
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3330
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3332
u32 rlen = min(op->u.read.rd_length, nfsd4_max_payload(rqstp));
fs/nfsd/nfs4proc.c
3338
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3340
u32 rlen = min(op->u.read.rd_length, nfsd4_max_payload(rqstp));
fs/nfsd/nfs4proc.c
3352
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3354
u32 rlen = min(op->u.readdir.rd_maxcount, nfsd4_max_payload(rqstp));
fs/nfsd/nfs4proc.c
3361
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3367
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3374
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3381
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3388
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3390
return (op_encode_hdr_size + 1 + op->u.test_stateid.ts_num_ids)
fs/nfsd/nfs4proc.c
3395
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3401
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3408
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3415
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3421
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3436
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3444
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3454
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3467
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3475
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3490
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3503
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3505
u32 rlen = min(op->u.getdeviceinfo.gd_maxcount, nfsd4_max_payload(rqstp));
fs/nfsd/nfs4proc.c
3519
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3529
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3537
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3547
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3553
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3561
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3567
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
3569
u32 rlen = min(op->u.listxattrs.lsxa_maxcount, nfsd4_max_payload(rqstp));
fs/nfsd/nfs4proc.c
3575
const struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
4043
int nfsd4_max_reply(struct svc_rqst *rqstp, struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
4045
if (op->opnum == OP_ILLEGAL || op->status == nfserr_notsupp)
fs/nfsd/nfs4proc.c
4048
BUG_ON(OPDESC(op)->op_rsize_bop == NULL);
fs/nfsd/nfs4proc.c
4049
return OPDESC(op)->op_rsize_bop(rqstp, op);
fs/nfsd/nfs4proc.c
4052
void warn_on_nonidempotent_op(struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
4054
if (OPDESC(op)->op_flags & OP_MODIFIES_SOMETHING) {
fs/nfsd/nfs4proc.c
4056
op->opnum, nfsd4_op_name(op->opnum));
fs/nfsd/nfs4proc.c
688
static __be32 nfsd4_open_omfg(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate, struct nfsd4_op *op)
fs/nfsd/nfs4proc.c
690
struct nfsd4_open *open = &op->u.open;
fs/nfsd/nfs4proc.c
692
if (!seqid_mutating_err(ntohl(op->status)))
fs/nfsd/nfs4proc.c
693
return op->status;
fs/nfsd/nfs4proc.c
695
return op->status;
fs/nfsd/nfs4proc.c
696
open->op_xdr_error = op->status;
fs/nfsd/nfs4proc.c
697
return nfsd4_open(rqstp, cstate, &op->u);
fs/nfsd/nfs4state.c
3556
struct nfsd4_op *op;
fs/nfsd/nfs4state.c
3558
op = &args->ops[resp->opcnt++];
fs/nfsd/nfs4state.c
3559
op->status = nfserr_retry_uncached_rep;
fs/nfsd/nfs4state.c
3560
nfsd4_encode_operation(resp, op);
fs/nfsd/nfs4state.c
3561
return op->status;
fs/nfsd/nfs4xdr.c
2585
nfsd4_opnum_in_range(struct nfsd4_compoundargs *argp, struct nfsd4_op *op)
fs/nfsd/nfs4xdr.c
2587
if (op->opnum < FIRST_NFS4_OP)
fs/nfsd/nfs4xdr.c
2589
else if (argp->minorversion == 0 && op->opnum > LAST_NFS40_OP)
fs/nfsd/nfs4xdr.c
2591
else if (argp->minorversion == 1 && op->opnum > LAST_NFS41_OP)
fs/nfsd/nfs4xdr.c
2593
else if (argp->minorversion == 2 && op->opnum > LAST_NFS42_OP)
fs/nfsd/nfs4xdr.c
2601
struct nfsd4_op *op;
fs/nfsd/nfs4xdr.c
2645
op = &argp->ops[i];
fs/nfsd/nfs4xdr.c
2646
op->replay = NULL;
fs/nfsd/nfs4xdr.c
2647
op->opdesc = NULL;
fs/nfsd/nfs4xdr.c
2649
if (xdr_stream_decode_u32(argp->xdr, &op->opnum) < 0)
fs/nfsd/nfs4xdr.c
2651
if (nfsd4_opnum_in_range(argp, op)) {
fs/nfsd/nfs4xdr.c
2652
op->opdesc = OPDESC(op);
fs/nfsd/nfs4xdr.c
2653
op->status = nfsd4_dec_ops[op->opnum](argp, &op->u);
fs/nfsd/nfs4xdr.c
2654
if (op->status != nfs_ok)
fs/nfsd/nfs4xdr.c
2657
op->opnum,
fs/nfsd/nfs4xdr.c
2658
op->status);
fs/nfsd/nfs4xdr.c
2660
op->opnum = OP_ILLEGAL;
fs/nfsd/nfs4xdr.c
2661
op->status = nfserr_op_illegal;
fs/nfsd/nfs4xdr.c
2668
cachethis |= nfsd4_cache_this_op(op);
fs/nfsd/nfs4xdr.c
2670
if (op->opnum == OP_READ || op->opnum == OP_READ_PLUS) {
fs/nfsd/nfs4xdr.c
2672
readbytes += nfsd4_max_reply(argp->rqstp, op);
fs/nfsd/nfs4xdr.c
2674
max_reply += nfsd4_max_reply(argp->rqstp, op);
fs/nfsd/nfs4xdr.c
2681
if (op->opnum == OP_LOCK || op->opnum == OP_LOCKT)
fs/nfsd/nfs4xdr.c
2684
if (op->status) {
fs/nfsd/nfs4xdr.c
6225
nfsd4_encode_operation(struct nfsd4_compoundres *resp, struct nfsd4_op *op)
fs/nfsd/nfs4xdr.c
6230
const struct nfsd4_operation *opdesc = op->opdesc;
fs/nfsd/nfs4xdr.c
6234
if (xdr_stream_encode_u32(xdr, op->opnum) != XDR_UNIT)
fs/nfsd/nfs4xdr.c
6240
if (op->opnum == OP_ILLEGAL)
fs/nfsd/nfs4xdr.c
6242
if (op->status && opdesc &&
fs/nfsd/nfs4xdr.c
6245
BUG_ON(op->opnum >= ARRAY_SIZE(nfsd4_enc_ops) ||
fs/nfsd/nfs4xdr.c
6246
!nfsd4_enc_ops[op->opnum]);
fs/nfsd/nfs4xdr.c
6247
encoder = nfsd4_enc_ops[op->opnum];
fs/nfsd/nfs4xdr.c
6248
op->status = encoder(resp, op->status, &op->u);
fs/nfsd/nfs4xdr.c
6249
if (op->status)
fs/nfsd/nfs4xdr.c
6250
trace_nfsd_compound_encode_err(rqstp, op->opnum, op->status);
fs/nfsd/nfs4xdr.c
6254
if (!op->status) {
fs/nfsd/nfs4xdr.c
6258
op->status = nfsd4_check_resp_size(resp, space_needed);
fs/nfsd/nfs4xdr.c
6260
if (op->status == nfserr_resource && nfsd4_has_session(&resp->cstate)) {
fs/nfsd/nfs4xdr.c
6264
op->status = nfserr_rep_too_big_to_cache;
fs/nfsd/nfs4xdr.c
6266
op->status = nfserr_rep_too_big;
fs/nfsd/nfs4xdr.c
6268
if (op->status == nfserr_resource ||
fs/nfsd/nfs4xdr.c
6269
op->status == nfserr_rep_too_big ||
fs/nfsd/nfs4xdr.c
6270
op->status == nfserr_rep_too_big_to_cache) {
fs/nfsd/nfs4xdr.c
6278
warn_on_nonidempotent_op(op);
fs/nfsd/nfs4xdr.c
6283
so->so_replay.rp_status = op->status;
fs/nfsd/nfs4xdr.c
6294
op->status = nfsd4_map_status(op->status,
fs/nfsd/nfs4xdr.c
6297
&op->status, XDR_UNIT);
fs/nfsd/nfs4xdr.c
6300
opdesc->op_release(&op->u);
fs/nfsd/nfs4xdr.c
6316
void nfsd4_encode_replay(struct xdr_stream *xdr, struct nfsd4_op *op)
fs/nfsd/nfs4xdr.c
6318
struct nfs4_replay *rp = op->replay;
fs/nfsd/nfs4xdr.c
6320
trace_nfsd_stateowner_replay(op->opnum, rp);
fs/nfsd/nfs4xdr.c
6322
if (xdr_stream_encode_u32(xdr, op->opnum) != XDR_UNIT)
fs/nfsd/state.h
821
const struct nfsd4_callback_ops *ops, enum nfsd4_cb_op op);
fs/nfsd/trace.h
250
DEFINE_NFSD_COMPOUND_ERR_EVENT(op);
fs/nfsd/xdr4.h
1055
const struct nfsd4_op *op);
fs/nfsd/xdr4.h
942
const struct nfsd4_operation *OPDESC(struct nfsd4_op *op);
fs/nfsd/xdr4.h
943
int nfsd4_max_reply(struct svc_rqst *rqstp, struct nfsd4_op *op);
fs/nfsd/xdr4.h
944
void warn_on_nonidempotent_op(struct nfsd4_op *op);
fs/nfsd/xdr4.h
953
void nfsd4_encode_replay(struct xdr_stream *xdr, struct nfsd4_op *op);
fs/nls/nls_base.c
138
u16 *op;
fs/nls/nls_base.c
142
op = pwcs;
fs/nls/nls_base.c
155
put_utf16(op++, SURROGATE_PAIR |
fs/nls/nls_base.c
158
put_utf16(op++, SURROGATE_PAIR |
fs/nls/nls_base.c
164
put_utf16(op++, u, endian);
fs/nls/nls_base.c
168
put_utf16(op++, *s++, endian);
fs/nls/nls_base.c
173
return op - pwcs;
fs/nls/nls_base.c
192
u8 *op;
fs/nls/nls_base.c
196
op = s;
fs/nls/nls_base.c
222
size = utf32_to_utf8(u, op, maxout);
fs/nls/nls_base.c
234
op += size;
fs/nls/nls_base.c
238
*op++ = (u8) u;
fs/nls/nls_base.c
242
return op - s;
fs/ntfs3/bitmap.c
1094
size_t op, max_check;
fs/ntfs3/bitmap.c
1104
for (op = e->start.key + e->count.key; op < max_check;
fs/ntfs3/bitmap.c
1105
op++) {
fs/ntfs3/bitmap.c
1106
if (!wnd_is_free(wnd, op, 1))
fs/ntfs3/bitmap.c
1111
to_alloc = op - e->start.key;
fs/ntfs3/bitmap.c
1361
u32 op;
fs/ntfs3/bitmap.c
1369
op = b0 + bits > wbits ? wbits - b0 : bits;
fs/ntfs3/bitmap.c
1394
bits -= op;
fs/ntfs3/bitmap.c
734
u32 op;
fs/ntfs3/bitmap.c
736
for (; iw < wnd->nwnd && bits; iw++, bit += op, bits -= op, wbit = 0) {
fs/ntfs3/bitmap.c
740
op = min_t(u32, wbits - wbit, bits);
fs/ntfs3/bitmap.c
750
ntfs_bitmap_clear_le(bh->b_data, wbit, op);
fs/ntfs3/bitmap.c
752
wnd->free_bits[iw] += op;
fs/ntfs3/bitmap.c
753
wnd->total_zeroes += op;
fs/ntfs3/bitmap.c
760
wnd_add_free_ext(wnd, bit, op, false);
fs/ntfs3/bitmap.c
776
u32 op;
fs/ntfs3/bitmap.c
778
for (; iw < wnd->nwnd && bits; iw++, bit += op, bits -= op, wbit = 0) {
fs/ntfs3/bitmap.c
782
op = min_t(u32, wbits - wbit, bits);
fs/ntfs3/bitmap.c
792
ntfs_bitmap_set_le(bh->b_data, wbit, op);
fs/ntfs3/bitmap.c
793
wnd->free_bits[iw] -= op;
fs/ntfs3/bitmap.c
794
wnd->total_zeroes -= op;
fs/ntfs3/bitmap.c
802
wnd_remove_free_ext(wnd, bit, op);
fs/ntfs3/bitmap.c
856
u32 op;
fs/ntfs3/bitmap.c
858
for (; iw < wnd->nwnd && bits; iw++, bits -= op, wbit = 0) {
fs/ntfs3/bitmap.c
862
op = min_t(u32, wbits - wbit, bits);
fs/ntfs3/bitmap.c
871
ret = are_bits_clear(bh->b_data, wbit, op);
fs/ntfs3/bitmap.c
926
u32 op;
fs/ntfs3/bitmap.c
944
for (; iw < wnd->nwnd && bits; iw++, bits -= op, wbit = 0) {
fs/ntfs3/bitmap.c
948
op = min_t(u32, wbits - wbit, bits);
fs/ntfs3/bitmap.c
957
ret = are_bits_set(bh->b_data, wbit, op);
fs/ntfs3/dir.c
118
u16 *op;
fs/ntfs3/dir.c
122
op = pwcs;
fs/ntfs3/dir.c
136
put_utf16(op++,
fs/ntfs3/dir.c
140
put_utf16(op++,
fs/ntfs3/dir.c
149
put_utf16(op++, u, endian);
fs/ntfs3/dir.c
156
put_utf16(op++, *s++, endian);
fs/ntfs3/dir.c
161
return op - pwcs;
fs/ntfs3/dir.c
23
u8 *op;
fs/ntfs3/dir.c
36
op = buf;
fs/ntfs3/dir.c
51
charlen = nls->uni2char(ec, op, buf_len);
fs/ntfs3/dir.c
54
op += charlen;
fs/ntfs3/dir.c
59
*op++ = '_';
fs/ntfs3/dir.c
73
*op = '\0';
fs/ntfs3/dir.c
74
return op - buf;
fs/ntfs3/fslog.c
3058
const struct LOG_REC_HDR *lrh, u32 op, void *data,
fs/ntfs3/fslog.c
3093
switch (op) {
fs/ntfs3/fslog.c
3119
if (err && op == InitializeFileRecordSegment) {
fs/ntfs3/fslog.c
3130
if (op == DeallocateFileRecordSegment)
fs/ntfs3/fslog.c
3141
if (is_rec_base(rec) || InitializeFileRecordSegment == op) {
fs/ntfs3/fslog.c
3191
bytes = UpdateNonresidentValue == op ? dlen : 0;
fs/ntfs3/fslog.c
3225
switch (op) {
fs/ntfs3/fslog.c
349
static inline bool is_target_required(u16 op)
fs/ntfs3/fslog.c
351
bool ret = op <= UpdateRecordDataAllocation &&
fs/ntfs3/fslog.c
352
(AttributeRequired[op >> 3] >> (op & 7) & 1);
fs/ntfs3/fslog.c
356
static inline bool can_skip_action(enum NTFS_LOG_OPERATION op)
fs/ntfs3/fslog.c
358
switch (op) {
fs/ntfs3/fsntfs.c
1064
u32 op = blocksize - off;
fs/ntfs3/fsntfs.c
1070
for (; bytes; block += 1, off = 0, op = blocksize) {
fs/ntfs3/fsntfs.c
1071
if (op > bytes)
fs/ntfs3/fsntfs.c
1072
op = bytes;
fs/ntfs3/fsntfs.c
1074
if (op < blocksize) {
fs/ntfs3/fsntfs.c
1090
memcpy(bh->b_data + off, buf, op);
fs/ntfs3/fsntfs.c
1091
buf = Add2Ptr(buf, op);
fs/ntfs3/fsntfs.c
1093
memset(bh->b_data + off, -1, op);
fs/ntfs3/fsntfs.c
1115
bytes -= op;
fs/ntfs3/fsntfs.c
1140
u32 op = min_t(u64, len, bytes);
fs/ntfs3/fsntfs.c
1141
int err = ntfs_sb_write(sb, lbo, op, buf, sync);
fs/ntfs3/fsntfs.c
1146
bytes -= op;
fs/ntfs3/fsntfs.c
1159
buf = Add2Ptr(buf, op);
fs/ntfs3/fsntfs.c
1247
u32 op = blocksize - off;
fs/ntfs3/fsntfs.c
1249
if (op > len32)
fs/ntfs3/fsntfs.c
1250
op = len32;
fs/ntfs3/fsntfs.c
1259
memcpy(buf, bh->b_data + off, op);
fs/ntfs3/fsntfs.c
1260
buf = Add2Ptr(buf, op);
fs/ntfs3/fsntfs.c
1273
bytes -= op;
fs/ntfs3/fsntfs.c
1276
len32 -= op;
fs/ntfs3/fsntfs.c
1365
u32 op;
fs/ntfs3/fsntfs.c
1373
op = blocksize - off;
fs/ntfs3/fsntfs.c
1374
if (op > len32)
fs/ntfs3/fsntfs.c
1375
op = len32;
fs/ntfs3/fsntfs.c
1377
if (op == blocksize) {
fs/ntfs3/fsntfs.c
1399
bytes -= op;
fs/ntfs3/fsntfs.c
1406
len32 -= op;
fs/ntfs3/fsntfs.c
1452
u32 op = block_size - off;
fs/ntfs3/fsntfs.c
1457
if (op > bytes)
fs/ntfs3/fsntfs.c
1458
op = bytes;
fs/ntfs3/fsntfs.c
1464
end_data = Add2Ptr(bh_data, op);
fs/ntfs3/fsntfs.c
1465
memcpy(bh_data, rhdr, op);
fs/ntfs3/fsntfs.c
1502
bytes -= op;
fs/ntfs3/fsntfs.c
1503
rhdr = Add2Ptr(rhdr, op);
fs/ntfs3/fsntfs.c
1522
u32 off, op;
fs/ntfs3/fsntfs.c
1547
op = PAGE_SIZE - off;
fs/ntfs3/fsntfs.c
1549
if (op > len)
fs/ntfs3/fsntfs.c
1550
op = len;
fs/ntfs3/fsntfs.c
1551
if (op > bytes)
fs/ntfs3/fsntfs.c
1552
op = bytes;
fs/ntfs3/fsntfs.c
1556
memcpy(kaddr + off, buf, op);
fs/ntfs3/fsntfs.c
1559
memcpy(buf, kaddr + off, op);
fs/ntfs3/fsntfs.c
1565
bytes -= op;
fs/ntfs3/fsntfs.c
1569
buf += op;
fs/ntfs3/fsntfs.c
1570
len -= op;
fs/ntfs3/fsntfs.c
1573
lbo += op;
fs/ntfs3/inode.c
1115
size_t op = bytes > PAGE_SIZE ? PAGE_SIZE : bytes;
fs/ntfs3/inode.c
1123
memcpy(data, kaddr, op);
fs/ntfs3/inode.c
1128
bytes -= op;
fs/open.c
1181
inline int build_open_flags(const struct open_how *how, struct open_flags *op)
fs/open.c
1214
op->mode = how->mode | S_IFREG;
fs/open.c
1218
op->mode = 0;
fs/open.c
1257
op->open_flag = flags;
fs/open.c
1268
op->acc_mode = acc_mode;
fs/open.c
1270
op->intent = flags & O_PATH ? 0 : LOOKUP_OPEN;
fs/open.c
1273
op->intent |= LOOKUP_CREATE;
fs/open.c
1275
op->intent |= LOOKUP_EXCL;
fs/open.c
1302
op->lookup_flags = lookup_flags;
fs/open.c
1319
struct open_flags op;
fs/open.c
1321
int err = build_open_flags(&how, &op);
fs/open.c
1324
return do_file_open(AT_FDCWD, name, &op);
fs/open.c
1348
struct open_flags op;
fs/open.c
1350
int err = build_open_flags(&how, &op);
fs/open.c
1353
return do_file_open_root(root, filename, &op);
fs/open.c
1360
struct open_flags op;
fs/open.c
1361
int err = build_open_flags(how, &op);
fs/open.c
1366
return FD_ADD(how->flags, do_file_open(dfd, name, &op));
fs/orangefs/devorangefs-req.c
162
struct orangefs_kernel_op_s *op, *temp;
fs/orangefs/devorangefs-req.c
192
list_for_each_entry_safe(op, temp, &orangefs_request_list, list) {
fs/orangefs/devorangefs-req.c
195
spin_lock(&op->lock);
fs/orangefs/devorangefs-req.c
196
if (unlikely(op_state_purged(op) || op_state_given_up(op))) {
fs/orangefs/devorangefs-req.c
197
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
201
fsid = fsid_of_op(op);
fs/orangefs/devorangefs-req.c
211
llu(op->tag),
fs/orangefs/devorangefs-req.c
212
get_opname_string(op));
fs/orangefs/devorangefs-req.c
213
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
224
!(op->upcall.type ==
fs/orangefs/devorangefs-req.c
226
op->upcall.type ==
fs/orangefs/devorangefs-req.c
228
op->upcall.type ==
fs/orangefs/devorangefs-req.c
232
llu(op->tag), get_opname_string(op));
fs/orangefs/devorangefs-req.c
236
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
245
cur_op = op;
fs/orangefs/devorangefs-req.c
365
struct orangefs_kernel_op_s *op = NULL;
fs/orangefs/devorangefs-req.c
415
op = orangefs_devreq_remove_op(head.tag);
fs/orangefs/devorangefs-req.c
416
if (!op) {
fs/orangefs/devorangefs-req.c
423
if (!copy_from_iter_full(&op->downcall, downcall_size, iter)) {
fs/orangefs/devorangefs-req.c
428
if (op->downcall.status)
fs/orangefs/devorangefs-req.c
436
if ((head_size + downcall_size + op->downcall.trailer_size) != total) {
fs/orangefs/devorangefs-req.c
443
op->downcall.trailer_size,
fs/orangefs/devorangefs-req.c
449
if ((op->downcall.type != ORANGEFS_VFS_OP_READDIR) &&
fs/orangefs/devorangefs-req.c
450
(op->downcall.trailer_size != 0)) {
fs/orangefs/devorangefs-req.c
453
op->downcall.type);
fs/orangefs/devorangefs-req.c
458
if ((op->downcall.type == ORANGEFS_VFS_OP_READDIR) &&
fs/orangefs/devorangefs-req.c
459
(op->downcall.trailer_size == 0)) {
fs/orangefs/devorangefs-req.c
462
op->downcall.type);
fs/orangefs/devorangefs-req.c
466
if (op->downcall.type != ORANGEFS_VFS_OP_READDIR)
fs/orangefs/devorangefs-req.c
469
op->downcall.trailer_buf = vzalloc(op->downcall.trailer_size);
fs/orangefs/devorangefs-req.c
470
if (!op->downcall.trailer_buf)
fs/orangefs/devorangefs-req.c
473
if (!copy_from_iter_full(op->downcall.trailer_buf,
fs/orangefs/devorangefs-req.c
474
op->downcall.trailer_size, iter)) {
fs/orangefs/devorangefs-req.c
476
vfree(op->downcall.trailer_buf);
fs/orangefs/devorangefs-req.c
48
static void orangefs_devreq_add_op(struct orangefs_kernel_op_s *op)
fs/orangefs/devorangefs-req.c
485
spin_lock(&op->lock);
fs/orangefs/devorangefs-req.c
486
if (unlikely(op_is_cancel(op))) {
fs/orangefs/devorangefs-req.c
487
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
488
put_cancel(op);
fs/orangefs/devorangefs-req.c
489
} else if (unlikely(op_state_given_up(op))) {
fs/orangefs/devorangefs-req.c
490
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
491
complete(&op->waitq);
fs/orangefs/devorangefs-req.c
493
set_op_state_serviced(op);
fs/orangefs/devorangefs-req.c
497
get_opname_string(op),
fs/orangefs/devorangefs-req.c
498
op->op_state,
fs/orangefs/devorangefs-req.c
50
int index = hash_func(op->tag, hash_table_size);
fs/orangefs/devorangefs-req.c
500
spin_unlock(&op->lock);
fs/orangefs/devorangefs-req.c
505
op->downcall.status = -(ORANGEFS_ERROR_BIT | 9);
fs/orangefs/devorangefs-req.c
510
op->downcall.status = -(ORANGEFS_ERROR_BIT | 8);
fs/orangefs/devorangefs-req.c
52
list_add_tail(&op->list, &orangefs_htable_ops_in_progress[index]);
fs/orangefs/devorangefs-req.c
61
struct orangefs_kernel_op_s *op, *next;
fs/orangefs/devorangefs-req.c
67
list_for_each_entry_safe(op,
fs/orangefs/devorangefs-req.c
71
if (op->tag == tag && !op_state_purged(op) &&
fs/orangefs/devorangefs-req.c
72
!op_state_given_up(op)) {
fs/orangefs/devorangefs-req.c
73
list_del_init(&op->list);
fs/orangefs/devorangefs-req.c
75
return op;
fs/orangefs/dir.c
100
vfree(op->downcall.trailer_buf);
fs/orangefs/dir.c
107
vfree(op->downcall.trailer_buf);
fs/orangefs/dir.c
110
} else if (op->downcall.status) {
fs/orangefs/dir.c
111
vfree(op->downcall.trailer_buf);
fs/orangefs/dir.c
112
od->error = op->downcall.status;
fs/orangefs/dir.c
113
return op->downcall.status;
fs/orangefs/dir.c
120
if (op->downcall.trailer_size > PART_SIZE) {
fs/orangefs/dir.c
121
vfree(op->downcall.trailer_buf);
fs/orangefs/dir.c
127
op->downcall.trailer_buf;
fs/orangefs/dir.c
133
struct orangefs_kernel_op_s *op)
fs/orangefs/dir.c
148
new = (void *)op->downcall.trailer_buf;
fs/orangefs/dir.c
150
new->len = op->downcall.trailer_size -
fs/orangefs/dir.c
164
struct orangefs_kernel_op_s *op;
fs/orangefs/dir.c
167
op = op_alloc(ORANGEFS_VFS_OP_READDIR);
fs/orangefs/dir.c
168
if (!op) {
fs/orangefs/dir.c
172
r = do_readdir(od, inode, op);
fs/orangefs/dir.c
177
r = parse_readdir(od, op);
fs/orangefs/dir.c
185
op_release(op);
fs/orangefs/dir.c
63
struct orangefs_kernel_op_s *op)
fs/orangefs/dir.c
75
op->uses_shared_memory = 1;
fs/orangefs/dir.c
76
op->upcall.req.readdir.refn = oi->refn;
fs/orangefs/dir.c
77
op->upcall.req.readdir.token = od->token;
fs/orangefs/dir.c
78
op->upcall.req.readdir.max_dirent_count =
fs/orangefs/dir.c
88
op->upcall.req.readdir.buf_index = bufi;
fs/orangefs/dir.c
90
r = service_operation(op, "orangefs_readdir",
fs/orangefs/dir.c
95
if (op_state_purged(op)) {
fs/orangefs/dir.c
97
vfree(op->downcall.trailer_buf);
fs/orangefs/orangefs-cache.c
107
void orangefs_new_tag(struct orangefs_kernel_op_s *op)
fs/orangefs/orangefs-cache.c
110
op->tag = next_tag_value++;
fs/orangefs/orangefs-debugfs.c
491
new_op->upcall.req.param.op =
fs/orangefs/orangefs-kernel.h
137
#define set_op_state_waiting(op) ((op)->op_state = OP_VFS_STATE_WAITING)
fs/orangefs/orangefs-kernel.h
138
#define set_op_state_inprogress(op) ((op)->op_state = OP_VFS_STATE_INPROGR)
fs/orangefs/orangefs-kernel.h
139
#define set_op_state_given_up(op) ((op)->op_state = OP_VFS_STATE_GIVEN_UP)
fs/orangefs/orangefs-kernel.h
140
static inline void set_op_state_serviced(struct orangefs_kernel_op_s *op)
fs/orangefs/orangefs-kernel.h
142
op->op_state = OP_VFS_STATE_SERVICED;
fs/orangefs/orangefs-kernel.h
143
complete(&op->waitq);
fs/orangefs/orangefs-kernel.h
146
#define op_state_waiting(op) ((op)->op_state & OP_VFS_STATE_WAITING)
fs/orangefs/orangefs-kernel.h
147
#define op_state_in_progress(op) ((op)->op_state & OP_VFS_STATE_INPROGR)
fs/orangefs/orangefs-kernel.h
148
#define op_state_serviced(op) ((op)->op_state & OP_VFS_STATE_SERVICED)
fs/orangefs/orangefs-kernel.h
149
#define op_state_purged(op) ((op)->op_state & OP_VFS_STATE_PURGED)
fs/orangefs/orangefs-kernel.h
150
#define op_state_given_up(op) ((op)->op_state & OP_VFS_STATE_GIVEN_UP)
fs/orangefs/orangefs-kernel.h
151
#define op_is_cancel(op) ((op)->upcall.type == ORANGEFS_VFS_OP_CANCEL)
fs/orangefs/orangefs-kernel.h
153
void op_release(struct orangefs_kernel_op_s *op);
fs/orangefs/orangefs-kernel.h
156
static inline void put_cancel(struct orangefs_kernel_op_s *op)
fs/orangefs/orangefs-kernel.h
158
orangefs_bufmap_put(op->slot_to_free);
fs/orangefs/orangefs-kernel.h
159
op_release(op);
fs/orangefs/orangefs-kernel.h
162
static inline void set_op_state_purged(struct orangefs_kernel_op_s *op)
fs/orangefs/orangefs-kernel.h
164
spin_lock(&op->lock);
fs/orangefs/orangefs-kernel.h
165
if (unlikely(op_is_cancel(op))) {
fs/orangefs/orangefs-kernel.h
166
list_del_init(&op->list);
fs/orangefs/orangefs-kernel.h
167
spin_unlock(&op->lock);
fs/orangefs/orangefs-kernel.h
168
put_cancel(op);
fs/orangefs/orangefs-kernel.h
170
op->op_state |= OP_VFS_STATE_PURGED;
fs/orangefs/orangefs-kernel.h
171
complete(&op->waitq);
fs/orangefs/orangefs-kernel.h
172
spin_unlock(&op->lock);
fs/orangefs/orangefs-kernel.h
313
void orangefs_new_tag(struct orangefs_kernel_op_s *op);
fs/orangefs/orangefs-kernel.h
400
__s32 fsid_of_op(struct orangefs_kernel_op_s *op);
fs/orangefs/orangefs-kernel.h
422
bool orangefs_cancel_op_in_progress(struct orangefs_kernel_op_s *op);
fs/orangefs/orangefs-kernel.h
458
int service_operation(struct orangefs_kernel_op_s *op,
fs/orangefs/orangefs-mod.c
211
struct orangefs_kernel_op_s *op;
fs/orangefs/orangefs-mod.c
215
list_for_each_entry_safe(op,
fs/orangefs/orangefs-mod.c
219
set_op_state_purged(op);
fs/orangefs/orangefs-mod.c
223
get_opname_string(op),
fs/orangefs/orangefs-mod.c
224
op->op_state,
fs/orangefs/orangefs-sysfs.c
358
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
362
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
366
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
371
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
376
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
381
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
386
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
390
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
394
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
398
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
402
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
407
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
411
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
415
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
419
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
424
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
428
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
432
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
436
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
441
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
445
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
449
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
453
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
491
if (new_op->upcall.req.param.op ==
fs/orangefs/orangefs-sysfs.c
577
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
586
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
595
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
604
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
613
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
628
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
640
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
651
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
659
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
668
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
676
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
687
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
695
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
704
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
712
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
723
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
731
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
740
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
748
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
759
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
767
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
776
new_op->upcall.req.param.op =
fs/orangefs/orangefs-sysfs.c
784
new_op->upcall.req.param.op =
fs/orangefs/orangefs-utils.c
14
__s32 fsid_of_op(struct orangefs_kernel_op_s *op)
fs/orangefs/orangefs-utils.c
18
if (op) {
fs/orangefs/orangefs-utils.c
19
switch (op->upcall.type) {
fs/orangefs/orangefs-utils.c
21
fsid = op->upcall.req.io.refn.fs_id;
fs/orangefs/orangefs-utils.c
24
fsid = op->upcall.req.lookup.parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
27
fsid = op->upcall.req.create.parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
30
fsid = op->upcall.req.getattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
33
fsid = op->upcall.req.remove.parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
36
fsid = op->upcall.req.mkdir.parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
39
fsid = op->upcall.req.readdir.refn.fs_id;
fs/orangefs/orangefs-utils.c
42
fsid = op->upcall.req.setattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
45
fsid = op->upcall.req.sym.parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
48
fsid = op->upcall.req.rename.old_parent_refn.fs_id;
fs/orangefs/orangefs-utils.c
51
fsid = op->upcall.req.statfs.fs_id;
fs/orangefs/orangefs-utils.c
54
fsid = op->upcall.req.truncate.refn.fs_id;
fs/orangefs/orangefs-utils.c
57
fsid = op->upcall.req.ra_cache_flush.refn.fs_id;
fs/orangefs/orangefs-utils.c
60
fsid = op->upcall.req.fs_umount.fs_id;
fs/orangefs/orangefs-utils.c
63
fsid = op->upcall.req.getxattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
66
fsid = op->upcall.req.setxattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
69
fsid = op->upcall.req.listxattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
72
fsid = op->upcall.req.removexattr.refn.fs_id;
fs/orangefs/orangefs-utils.c
75
fsid = op->upcall.req.fsync.refn.fs_id;
fs/orangefs/super.c
384
struct orangefs_kernel_op_s *op;
fs/orangefs/super.c
386
op = op_alloc(ORANGEFS_VFS_OP_FS_UMOUNT);
fs/orangefs/super.c
387
if (!op)
fs/orangefs/super.c
389
op->upcall.req.fs_umount.id = id;
fs/orangefs/super.c
390
op->upcall.req.fs_umount.fs_id = fs_id;
fs/orangefs/super.c
391
strscpy(op->upcall.req.fs_umount.orangefs_config_server, devname);
fs/orangefs/super.c
392
r = service_operation(op, "orangefs_fs_umount", 0);
fs/orangefs/super.c
396
op_release(op);
fs/orangefs/upcall.h
191
enum orangefs_param_request_op op;
fs/orangefs/waitqueue.c
103
op->downcall.status = ret;
fs/orangefs/waitqueue.c
113
spin_lock(&op->lock);
fs/orangefs/waitqueue.c
114
set_op_state_waiting(op);
fs/orangefs/waitqueue.c
118
get_opname_string(op),
fs/orangefs/waitqueue.c
119
op->op_state,
fs/orangefs/waitqueue.c
123
list_add(&op->list, &orangefs_request_list);
fs/orangefs/waitqueue.c
125
list_add_tail(&op->list, &orangefs_request_list);
fs/orangefs/waitqueue.c
126
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
136
if (op->upcall.type == ORANGEFS_VFS_OP_FS_UMOUNT)
fs/orangefs/waitqueue.c
146
ret = wait_for_matching_downcall(op, timeout, flags);
fs/orangefs/waitqueue.c
151
op);
fs/orangefs/waitqueue.c
155
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
156
op->downcall.status =
fs/orangefs/waitqueue.c
157
orangefs_normalize_to_errno(op->downcall.status);
fs/orangefs/waitqueue.c
158
ret = op->downcall.status;
fs/orangefs/waitqueue.c
173
orangefs_clean_up_interrupted_operation(op);
fs/orangefs/waitqueue.c
175
op->downcall.status = ret;
fs/orangefs/waitqueue.c
178
op->attempts++;
fs/orangefs/waitqueue.c
183
llu(op->tag),
fs/orangefs/waitqueue.c
185
op->attempts);
fs/orangefs/waitqueue.c
192
if (!op->uses_shared_memory)
fs/orangefs/waitqueue.c
20
static int wait_for_matching_downcall(struct orangefs_kernel_op_s *op,
fs/orangefs/waitqueue.c
202
op);
fs/orangefs/waitqueue.c
207
bool orangefs_cancel_op_in_progress(struct orangefs_kernel_op_s *op)
fs/orangefs/waitqueue.c
209
u64 tag = op->tag;
fs/orangefs/waitqueue.c
210
if (!op_state_in_progress(op))
fs/orangefs/waitqueue.c
213
op->slot_to_free = op->upcall.req.io.buf_index;
fs/orangefs/waitqueue.c
214
memset(&op->upcall, 0, sizeof(op->upcall));
fs/orangefs/waitqueue.c
215
memset(&op->downcall, 0, sizeof(op->downcall));
fs/orangefs/waitqueue.c
216
op->upcall.type = ORANGEFS_VFS_OP_CANCEL;
fs/orangefs/waitqueue.c
217
op->upcall.req.cancel.op_tag = tag;
fs/orangefs/waitqueue.c
218
op->downcall.type = ORANGEFS_VFS_OP_INVALID;
fs/orangefs/waitqueue.c
219
op->downcall.status = -1;
fs/orangefs/waitqueue.c
220
orangefs_new_tag(op);
fs/orangefs/waitqueue.c
228
spin_lock(&op->lock);
fs/orangefs/waitqueue.c
229
set_op_state_waiting(op);
fs/orangefs/waitqueue.c
23
__acquires(op->lock);
fs/orangefs/waitqueue.c
233
get_opname_string(op),
fs/orangefs/waitqueue.c
234
op->op_state,
fs/orangefs/waitqueue.c
236
list_add(&op->list, &orangefs_request_list);
fs/orangefs/waitqueue.c
237
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
24
static void orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *op)
fs/orangefs/waitqueue.c
25
__releases(op->lock);
fs/orangefs/waitqueue.c
250
orangefs_clean_up_interrupted_operation(struct orangefs_kernel_op_s *op)
fs/orangefs/waitqueue.c
251
__releases(op->lock)
fs/orangefs/waitqueue.c
264
op->op_state |= OP_VFS_STATE_GIVEN_UP;
fs/orangefs/waitqueue.c
266
if (list_empty(&op->list)) {
fs/orangefs/waitqueue.c
268
BUG_ON(op_state_serviced(op));
fs/orangefs/waitqueue.c
269
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
270
wait_for_completion(&op->waitq);
fs/orangefs/waitqueue.c
271
} else if (op_state_waiting(op)) {
fs/orangefs/waitqueue.c
276
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
278
list_del_init(&op->list);
fs/orangefs/waitqueue.c
282
op);
fs/orangefs/waitqueue.c
283
} else if (op_state_in_progress(op)) {
fs/orangefs/waitqueue.c
285
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
287
list_del_init(&op->list);
fs/orangefs/waitqueue.c
292
op);
fs/orangefs/waitqueue.c
294
spin_unlock(&op->lock);
fs/orangefs/waitqueue.c
296
op->op_state);
fs/orangefs/waitqueue.c
298
reinit_completion(&op->waitq);
fs/orangefs/waitqueue.c
318
static int wait_for_matching_downcall(struct orangefs_kernel_op_s *op,
fs/orangefs/waitqueue.c
321
__acquires(op->lock)
fs/orangefs/waitqueue.c
334
n = wait_for_completion_io_timeout(&op->waitq, timeout);
fs/orangefs/waitqueue.c
336
n = wait_for_completion_interruptible_timeout(&op->waitq,
fs/orangefs/waitqueue.c
339
n = wait_for_completion_killable_timeout(&op->waitq, timeout);
fs/orangefs/waitqueue.c
341
spin_lock(&op->lock);
fs/orangefs/waitqueue.c
343
if (op_state_serviced(op))
fs/orangefs/waitqueue.c
350
llu(op->tag),
fs/orangefs/waitqueue.c
351
op);
fs/orangefs/waitqueue.c
354
if (op_state_purged(op)) {
fs/orangefs/waitqueue.c
358
llu(op->tag),
fs/orangefs/waitqueue.c
359
op,
fs/orangefs/waitqueue.c
36
struct orangefs_kernel_op_s *op, *tmp;
fs/orangefs/waitqueue.c
360
op->attempts);
fs/orangefs/waitqueue.c
361
return (op->attempts < ORANGEFS_PURGE_RETRY_COUNT) ?
fs/orangefs/waitqueue.c
369
llu(op->tag),
fs/orangefs/waitqueue.c
370
op,
fs/orangefs/waitqueue.c
371
op->attempts);
fs/orangefs/waitqueue.c
39
list_for_each_entry_safe(op, tmp, &orangefs_request_list, list) {
fs/orangefs/waitqueue.c
42
llu(op->tag),
fs/orangefs/waitqueue.c
43
get_opname_string(op));
fs/orangefs/waitqueue.c
44
set_op_state_purged(op);
fs/orangefs/waitqueue.c
48
get_opname_string(op),
fs/orangefs/waitqueue.c
49
op->op_state,
fs/orangefs/waitqueue.c
65
int service_operation(struct orangefs_kernel_op_s *op,
fs/orangefs/waitqueue.c
74
op->upcall.tgid = current->tgid;
fs/orangefs/waitqueue.c
75
op->upcall.pid = current->pid;
fs/orangefs/waitqueue.c
78
op->downcall.status = 0;
fs/orangefs/waitqueue.c
83
op,
fs/overlayfs/file.c
529
loff_t len, unsigned int flags, enum ovl_copyop op)
fs/overlayfs/file.c
536
if (op != OVL_DEDUPE) {
fs/overlayfs/file.c
555
switch (op) {
fs/overlayfs/file.c
595
enum ovl_copyop op;
fs/overlayfs/file.c
601
op = OVL_DEDUPE;
fs/overlayfs/file.c
603
op = OVL_CLONE;
fs/overlayfs/file.c
609
if (op == OVL_DEDUPE &&
fs/overlayfs/file.c
615
remap_flags, op);
fs/proc/base.c
160
union proc_op op;
fs/proc/base.c
169
.op = OP, \
fs/proc/base.c
1813
error = PROC_I(inode)->op.proc_get_link(dentry, &path);
fs/proc/base.c
1856
error = PROC_I(inode)->op.proc_get_link(dentry, &path);
fs/proc/base.c
2340
ei->op.proc_get_link = map_files_get_link;
fs/proc/base.c
2708
ei->op = p->op;
fs/proc/base.c
2786
length = security_getprocattr(task, PROC_I(inode)->op.lsmid,
fs/proc/base.c
2844
rv = security_setprocattr(PROC_I(inode)->op.lsmid,
fs/proc/base.c
812
ret = PROC_I(inode)->op.proc_show(m, ns, pid, task);
fs/proc/fd.c
219
ei->op.proc_get_link = proc_fd_link;
fs/proc/inode.c
62
ei->op.proc_get_link = NULL;
fs/proc/internal.h
120
union proc_op op;
fs/proc/page.c
63
enum kpage_operation op)
fs/proc/page.c
88
switch (op) {
fs/proc/proc_sysctl.c
425
static int test_perm(int mode, int op)
fs/proc/proc_sysctl.c
431
if ((op & ~mode & (MAY_READ|MAY_WRITE|MAY_EXEC)) == 0)
fs/proc/proc_sysctl.c
436
static int sysctl_perm(struct ctl_table_header *head, const struct ctl_table *table, int op)
fs/proc/proc_sysctl.c
446
return test_perm(mode, op);
fs/proc/task_mmu.c
159
if (m->op != &proc_pid_maps_op) {
fs/pstore/ftrace.c
28
struct ftrace_ops *op,
fs/pstore/inode.c
166
if (sf->op)
fs/read_write.c
500
static int warn_unsupported(struct file *file, const char *op)
fs/read_write.c
504
op, file, current->pid, current->comm);
fs/seq_file.c
107
p = m->op->start(m, &m->index);
fs/seq_file.c
112
error = m->op->show(m, p);
fs/seq_file.c
121
p = m->op->next(m, p, &m->index);
fs/seq_file.c
132
m->op->stop(m, p);
fs/seq_file.c
136
m->op->stop(m, p);
fs/seq_file.c
226
p = m->op->start(m, &m->index);
fs/seq_file.c
231
err = m->op->show(m, p);
fs/seq_file.c
237
p = m->op->next(m, p, &m->index);
fs/seq_file.c
243
m->op->stop(m, p);
fs/seq_file.c
249
p = m->op->start(m, &m->index);
fs/seq_file.c
252
m->op->stop(m, p);
fs/seq_file.c
263
p = m->op->next(m, p, &m->index);
fs/seq_file.c
266
m->op->next);
fs/seq_file.c
273
err = m->op->show(m, p);
fs/seq_file.c
281
m->op->stop(m, p);
fs/seq_file.c
576
struct seq_operations *op = kmalloc_obj(*op, GFP_KERNEL_ACCOUNT);
fs/seq_file.c
579
if (op) {
fs/seq_file.c
58
int seq_open(struct file *file, const struct seq_operations *op)
fs/seq_file.c
580
op->start = single_start;
fs/seq_file.c
581
op->next = single_next;
fs/seq_file.c
582
op->stop = single_stop;
fs/seq_file.c
583
op->show = show;
fs/seq_file.c
584
res = seq_open(file, op);
fs/seq_file.c
588
kfree(op);
fs/seq_file.c
614
const struct seq_operations *op = ((struct seq_file *)file->private_data)->op;
fs/seq_file.c
616
kfree(op);
fs/seq_file.c
71
p->op = op;
fs/smb/client/smb1ops.c
1156
__u8 op;
fs/smb/client/smb1ops.c
1158
op = !!((oplock & CIFS_CACHE_READ_FLG) || (sbflags & CIFS_MOUNT_RO_CACHE));
fs/smb/client/smb1ops.c
1160
LOCKING_ANDX_OPLOCK_RELEASE, false, op);
fs/smb/client/smb2ops.c
2693
__u8 op;
fs/smb/client/smb2ops.c
2699
op = !!((oplock & CIFS_CACHE_READ_FLG) || (sbflags & CIFS_MOUNT_RO_CACHE));
fs/smb/client/smb2ops.c
2700
return SMB2_oplock_break(0, tcon, persistent_fid, volatile_fid, op);
fs/smb/server/oplock.c
1365
struct oplock_info *op, *brk_op;
fs/smb/server/oplock.c
1374
op = opinfo_get(fp);
fs/smb/server/oplock.c
1408
if (op && op->is_lease && brk_op->is_lease &&
fs/smb/server/oplock.c
1411
!memcmp(op->o_lease->lease_key, brk_op->o_lease->lease_key,
fs/smb/server/oplock.c
1421
if (op)
fs/smb/server/oplock.c
1422
opinfo_put(op);
fs/smb/server/vfs_cache.c
1005
list_for_each_entry_rcu(op, &ci->m_op_list, op_entry) {
fs/smb/server/vfs_cache.c
1006
if (op->conn != conn)
fs/smb/server/vfs_cache.c
1008
if (op->conn && atomic_dec_and_test(&op->conn->refcnt))
fs/smb/server/vfs_cache.c
1009
kfree(op->conn);
fs/smb/server/vfs_cache.c
1010
op->conn = NULL;
fs/smb/server/vfs_cache.c
1092
struct oplock_info *op;
fs/smb/server/vfs_cache.c
1109
list_for_each_entry_rcu(op, &ci->m_op_list, op_entry) {
fs/smb/server/vfs_cache.c
1110
if (op->conn)
fs/smb/server/vfs_cache.c
1112
op->conn = fp->conn;
fs/smb/server/vfs_cache.c
1113
atomic_inc(&op->conn->refcnt);
fs/smb/server/vfs_cache.c
996
struct oplock_info *op;
fs/splice.c
920
static int warn_unsupported(struct file *file, const char *op)
fs/splice.c
924
op, file, current->pid, current->comm);
fs/xfs/libxfs/xfs_attr.c
1133
enum xfs_attr_update op,
fs/xfs/libxfs/xfs_attr.c
1145
switch (op) {
fs/xfs/libxfs/xfs_attr.c
1188
if (op != XFS_ATTRUPDATE_REMOVE || xfs_inode_hasattr(dp)) {
fs/xfs/libxfs/xfs_attr.c
1198
if (op == XFS_ATTRUPDATE_REMOVE) {
fs/xfs/libxfs/xfs_attr.c
1207
if (op == XFS_ATTRUPDATE_CREATE)
fs/xfs/libxfs/xfs_attr.c
1216
if (op == XFS_ATTRUPDATE_REMOVE)
fs/xfs/libxfs/xfs_attr.c
1220
if (op == XFS_ATTRUPDATE_REPLACE)
fs/xfs/libxfs/xfs_attr.h
562
int xfs_attr_set(struct xfs_da_args *args, enum xfs_attr_update op, bool rsvd);
fs/xfs/libxfs/xfs_rmap.c
2603
enum xfs_rmap_intent_type op,
fs/xfs/libxfs/xfs_rmap.c
2609
switch (op) {
fs/xfs/libxfs/xfs_rmap.c
909
enum xfs_rmap_intent_type op,
fs/xfs/libxfs/xfs_rmap.c
924
xfs_hooks_call(&xg->xg_rmap_update_hooks, op, &p);
fs/xfs/libxfs/xfs_rmap.h
197
enum xfs_rmap_intent_type op, xfs_agblock_t bno,
fs/xfs/scrub/trace.h
2766
TP_PROTO(const struct xfs_group *xg, unsigned int op,
fs/xfs/scrub/trace.h
2768
TP_ARGS(xg, op, p),
fs/xfs/scrub/trace.h
2773
__field(unsigned int, op)
fs/xfs/scrub/trace.h
2784
__entry->op = op;
fs/xfs/scrub/trace.h
2796
__entry->op,
fs/xfs/xfs_attr_item.c
1016
unsigned int op, i = 0;
fs/xfs/xfs_attr_item.c
1034
op = xfs_attr_log_item_op(attri_formatp);
fs/xfs/xfs_attr_item.c
1035
switch (op) {
fs/xfs/xfs_attr_item.c
550
unsigned int op = xfs_attr_log_item_op(attrp);
fs/xfs/xfs_attr_item.c
562
switch (op) {
fs/xfs/xfs_attr_item.c
868
enum xfs_attr_defer_op op)
fs/xfs/xfs_attr_item.c
886
switch (op) {
fs/xfs/xfs_attr_item.h
62
void xfs_attr_defer_add(struct xfs_da_args *args, enum xfs_attr_defer_op op);
fs/xfs/xfs_bio_io.c
18
enum req_op op)
fs/xfs/xfs_bio_io.c
25
op |= REQ_META | REQ_SYNC;
fs/xfs/xfs_bio_io.c
27
return bdev_rw_virt(bdev, sector, data, count, op);
fs/xfs/xfs_bio_io.c
29
bio = bio_alloc(bdev, bio_max_vecs(count), op, GFP_KERNEL);
fs/xfs/xfs_bio_io.c
50
if (op == REQ_OP_READ)
fs/xfs/xfs_buf.c
1316
blk_opf_t op;
fs/xfs/xfs_buf.c
1319
op = REQ_OP_WRITE;
fs/xfs/xfs_buf.c
1321
op = REQ_OP_READ;
fs/xfs/xfs_buf.c
1323
op |= REQ_RAHEAD;
fs/xfs/xfs_buf.c
1326
return op | REQ_META;
fs/xfs/xfs_log_recover.c
125
enum req_op op)
fs/xfs/xfs_log_recover.c
141
BBTOB(nbblks), data, op);
fs/xfs/xfs_log_recover.c
145
op == REQ_OP_WRITE ? "write" : "read",
fs/xfs/xfs_platform.h
225
char *data, enum req_op op);
fs/xfs/xfs_trace.h
3380
__field(int, op)
fs/xfs/xfs_trace.h
3394
__entry->op = ri->ri_type;
fs/xfs/xfs_trace.h
3398
__print_symbolic(__entry->op, XFS_RMAP_INTENT_STRINGS),
fs/xfs/xfs_trace.h
3447
__field(int, op)
fs/xfs/xfs_trace.h
3476
__entry->op = bi->bi_type;
fs/xfs/xfs_trace.h
3480
__print_symbolic(__entry->op, XFS_BMAP_INTENT_STRINGS),
fs/xfs/xfs_trace.h
3934
__field(int, op)
fs/xfs/xfs_trace.h
3942
__entry->op = refc->ri_type;
fs/xfs/xfs_trace.h
3949
__print_symbolic(__entry->op, XFS_REFCOUNT_INTENT_STRINGS),
fs/xfs/xfs_xattr.c
128
return xfs_attr_set(args, op,
fs/xfs/xfs_xattr.c
80
enum xfs_attr_update op)
fs/xfs/xfs_xattr.h
10
int xfs_attr_change(struct xfs_da_args *args, enum xfs_attr_update op);
fs/zonefs/file.c
203
enum req_op op;
fs/zonefs/file.c
215
op = REQ_OP_ZONE_RESET;
fs/zonefs/file.c
217
op = REQ_OP_ZONE_FINISH;
fs/zonefs/file.c
233
ret = zonefs_inode_zone_mgmt(inode, op);
fs/zonefs/super.c
100
struct zonefs_zone *z, enum req_op op)
fs/zonefs/super.c
112
if (op == REQ_OP_ZONE_CLOSE && !z->z_wpoffset)
fs/zonefs/super.c
113
op = REQ_OP_ZONE_RESET;
fs/zonefs/super.c
115
trace_zonefs_zone_mgmt(sb, z, op);
fs/zonefs/super.c
116
ret = blkdev_zone_mgmt(sb->s_bdev, op, z->z_sector,
fs/zonefs/super.c
121
blk_op_str(op), z->z_sector, ret);
fs/zonefs/super.c
128
int zonefs_inode_zone_mgmt(struct inode *inode, enum req_op op)
fs/zonefs/super.c
132
return zonefs_zone_mgmt(inode->i_sb, zonefs_inode_zone(inode), op);
fs/zonefs/trace.h
24
enum req_op op),
fs/zonefs/trace.h
25
TP_ARGS(sb, z, op),
fs/zonefs/trace.h
29
__field(enum req_op, op)
fs/zonefs/trace.h
37
__entry->op = op;
fs/zonefs/trace.h
43
blk_op_str(__entry->op), __entry->sector,
fs/zonefs/zonefs.h
260
int zonefs_inode_zone_mgmt(struct inode *inode, enum req_op op);
include/asm-generic/atomic.h
19
#define ATOMIC_OP(op, c_op) \
include/asm-generic/atomic.h
20
static inline void generic_atomic_##op(int i, atomic_t *v) \
include/asm-generic/atomic.h
29
#define ATOMIC_OP_RETURN(op, c_op) \
include/asm-generic/atomic.h
30
static inline int generic_atomic_##op##_return(int i, atomic_t *v) \
include/asm-generic/atomic.h
41
#define ATOMIC_FETCH_OP(op, c_op) \
include/asm-generic/atomic.h
42
static inline int generic_atomic_fetch_##op(int i, atomic_t *v) \
include/asm-generic/atomic.h
57
#define ATOMIC_OP(op, c_op) \
include/asm-generic/atomic.h
58
static inline void generic_atomic_##op(int i, atomic_t *v) \
include/asm-generic/atomic.h
67
#define ATOMIC_OP_RETURN(op, c_op) \
include/asm-generic/atomic.h
68
static inline int generic_atomic_##op##_return(int i, atomic_t *v) \
include/asm-generic/atomic.h
80
#define ATOMIC_FETCH_OP(op, c_op) \
include/asm-generic/atomic.h
81
static inline int generic_atomic_fetch_##op(int i, atomic_t *v) \
include/asm-generic/atomic64.h
21
#define ATOMIC64_OP(op) \
include/asm-generic/atomic64.h
22
extern void generic_atomic64_##op(s64 a, atomic64_t *v);
include/asm-generic/atomic64.h
24
#define ATOMIC64_OP_RETURN(op) \
include/asm-generic/atomic64.h
25
extern s64 generic_atomic64_##op##_return(s64 a, atomic64_t *v);
include/asm-generic/atomic64.h
27
#define ATOMIC64_FETCH_OP(op) \
include/asm-generic/atomic64.h
28
extern s64 generic_atomic64_fetch_##op(s64 a, atomic64_t *v);
include/asm-generic/atomic64.h
30
#define ATOMIC64_OPS(op) ATOMIC64_OP(op) ATOMIC64_OP_RETURN(op) ATOMIC64_FETCH_OP(op)
include/asm-generic/atomic64.h
36
#define ATOMIC64_OPS(op) ATOMIC64_OP(op) ATOMIC64_FETCH_OP(op)
include/asm-generic/futex.h
18
#define arch_futex_atomic_op_inuser(op, oparg, oval, uaddr) \
include/asm-generic/futex.h
19
futex_atomic_op_inuser_local(op, oparg, oval, uaddr)
include/asm-generic/futex.h
38
futex_atomic_op_inuser_local(int op, u32 oparg, int *oval, u32 __user *uaddr)
include/asm-generic/futex.h
52
switch (op) {
include/asm-generic/percpu.h
167
#define this_cpu_generic_to_op(pcp, val, op) \
include/asm-generic/percpu.h
171
raw_cpu_generic_to_op(pcp, val, op); \
include/asm-generic/percpu.h
85
#define raw_cpu_generic_to_op(pcp, val, op) \
include/asm-generic/percpu.h
87
*raw_cpu_ptr(&(pcp)) op val; \
include/crypto/engine.h
31
struct crypto_engine_op op;
include/crypto/engine.h
36
struct crypto_engine_op op;
include/crypto/engine.h
41
struct crypto_engine_op op;
include/crypto/engine.h
46
struct crypto_engine_op op;
include/crypto/engine.h
51
struct crypto_engine_op op;
include/crypto/if_alg.h
39
int op;
include/drm/drm_gpuvm.h
1031
#define drm_gpuva_for_each_op(op, ops) list_for_each_entry(op, &(ops)->list, entry)
include/drm/drm_gpuvm.h
1042
#define drm_gpuva_for_each_op_safe(op, next, ops) \
include/drm/drm_gpuvm.h
1043
list_for_each_entry_safe(op, next, &(ops)->list, entry)
include/drm/drm_gpuvm.h
1053
#define drm_gpuva_for_each_op_from_reverse(op, ops) \
include/drm/drm_gpuvm.h
1054
list_for_each_entry_from_reverse(op, &(ops)->list, entry)
include/drm/drm_gpuvm.h
1063
#define drm_gpuva_for_each_op_reverse(op, ops) \
include/drm/drm_gpuvm.h
1064
list_for_each_entry_reverse(op, &(ops)->list, entry)
include/drm/drm_gpuvm.h
1084
#define drm_gpuva_prev_op(op) list_prev_entry(op, entry)
include/drm/drm_gpuvm.h
1090
#define drm_gpuva_next_op(op) list_next_entry(op, entry)
include/drm/drm_gpuvm.h
1124
const struct drm_gpuva_op_map *op)
include/drm/drm_gpuvm.h
1126
va->va.addr = op->va.addr;
include/drm/drm_gpuvm.h
1127
va->va.range = op->va.range;
include/drm/drm_gpuvm.h
1128
va->gem.obj = op->gem.obj;
include/drm/drm_gpuvm.h
1129
va->gem.offset = op->gem.offset;
include/drm/drm_gpuvm.h
1170
void (*op_free)(struct drm_gpuva_op *op);
include/drm/drm_gpuvm.h
1217
int (*sm_step_map)(struct drm_gpuva_op *op, void *priv);
include/drm/drm_gpuvm.h
1234
int (*sm_step_remap)(struct drm_gpuva_op *op, void *priv);
include/drm/drm_gpuvm.h
1250
int (*sm_step_unmap)(struct drm_gpuva_op *op, void *priv);
include/drm/drm_gpuvm.h
1268
const struct drm_gpuva_op_map *op);
include/drm/drm_gpuvm.h
1272
const struct drm_gpuva_op_remap *op);
include/drm/drm_gpuvm.h
1274
void drm_gpuva_unmap(const struct drm_gpuva_op_unmap *op);
include/drm/drm_gpuvm.h
1288
drm_gpuva_op_remap_to_unmap_range(const struct drm_gpuva_op_remap *op,
include/drm/drm_gpuvm.h
1291
const u64 va_start = op->prev ?
include/drm/drm_gpuvm.h
1292
op->prev->va.addr + op->prev->va.range :
include/drm/drm_gpuvm.h
1293
op->unmap->va->va.addr;
include/drm/drm_gpuvm.h
1294
const u64 va_end = op->next ?
include/drm/drm_gpuvm.h
1295
op->next->va.addr :
include/drm/drm_gpuvm.h
1296
op->unmap->va->va.addr + op->unmap->va->va.range;
include/drm/drm_gpuvm.h
989
enum drm_gpuva_op_type op;
include/keys/request_key_auth-type.h
24
char op[8];
include/kunit/test.h
834
op, \
include/kunit/test.h
842
.operation = #op, \
include/kunit/test.h
848
if (likely(__left op __right)) \
include/kunit/test.h
865
op, \
include/kunit/test.h
873
left, op, right, \
include/kunit/test.h
880
op, \
include/kunit/test.h
888
left, op, right, \
include/kunit/test.h
895
op, \
include/kunit/test.h
903
.operation = #op, \
include/kunit/test.h
910
(strcmp(__left, __right) op 0))) \
include/kunit/test.h
928
op, \
include/kunit/test.h
938
.operation = #op, \
include/kunit/test.h
945
if (likely(memcmp(__left, __right, __size) op 0)) \
include/linux/amd-pmf-io.h
82
int amd_get_sfh_info(struct amd_sfh_info *sfh_info, enum sfh_message_type op);
include/linux/atomic.h
58
#define __atomic_op_acquire(op, args...) \
include/linux/atomic.h
60
typeof(op##_relaxed(args)) __ret = op##_relaxed(args); \
include/linux/atomic.h
65
#define __atomic_op_release(op, args...) \
include/linux/atomic.h
68
op##_relaxed(args); \
include/linux/atomic.h
71
#define __atomic_op_fence(op, args...) \
include/linux/atomic.h
73
typeof(op##_relaxed(args)) __ret; \
include/linux/atomic.h
75
__ret = op##_relaxed(args); \
include/linux/audit.h
315
extern void __audit_uring_entry(u8 op);
include/linux/audit.h
352
static inline void audit_uring_entry(u8 op)
include/linux/audit.h
360
__audit_uring_entry(op);
include/linux/audit.h
442
enum audit_nfcfgop op, gfp_t gfp);
include/linux/audit.h
586
enum audit_nfcfgop op, gfp_t gfp)
include/linux/audit.h
589
__audit_log_nfcfg(name, af, nentries, op, gfp);
include/linux/audit.h
601
static inline void audit_uring_entry(u8 op)
include/linux/audit.h
721
enum audit_nfcfgop op, gfp_t gfp)
include/linux/audit.h
76
u32 op;
include/linux/bio.h
409
bio_iov_bounce_nr_vecs(struct iov_iter *iter, blk_opf_t op)
include/linux/bio.h
418
if (op_is_write(op))
include/linux/bio.h
467
size_t len, enum req_op op);
include/linux/bitops.h
42
#define bitop(op, nr, addr) \
include/linux/bitops.h
47
const##op(nr, addr) : op(nr, addr))
include/linux/blk_types.h
459
static inline bool op_is_write(blk_opf_t op)
include/linux/blk_types.h
461
return !!(op & (__force blk_opf_t)1);
include/linux/blk_types.h
468
static inline bool op_is_flush(blk_opf_t op)
include/linux/blk_types.h
470
return op & (REQ_FUA | REQ_PREFLUSH);
include/linux/blk_types.h
478
static inline bool op_is_sync(blk_opf_t op)
include/linux/blk_types.h
480
return (op & REQ_OP_MASK) == REQ_OP_READ ||
include/linux/blk_types.h
481
(op & (REQ_SYNC | REQ_FUA | REQ_PREFLUSH));
include/linux/blk_types.h
484
static inline bool op_is_discard(blk_opf_t op)
include/linux/blk_types.h
486
return (op & REQ_OP_MASK) == REQ_OP_DISCARD;
include/linux/blk_types.h
492
static inline bool op_is_zone_mgmt(enum req_op op)
include/linux/blk_types.h
494
switch (op & REQ_OP_MASK) {
include/linux/blk_types.h
506
static inline int op_stat_group(enum req_op op)
include/linux/blk_types.h
508
if (op_is_discard(op))
include/linux/blk_types.h
510
return op_is_write(op);
include/linux/blkdev.h
1030
extern const char *blk_op_str(enum req_op op);
include/linux/blkdev.h
1704
unsigned long bdev_start_io_acct(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
1706
void bdev_end_io_acct(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
295
static inline bool blk_op_is_passthrough(blk_opf_t op)
include/linux/blkdev.h
297
op &= REQ_OP_MASK;
include/linux/blkdev.h
298
return op == REQ_OP_DRV_IN || op == REQ_OP_DRV_OUT;
include/linux/blkdev.h
447
int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
include/linux/blkdev.h
872
enum req_op op = bio_op(bio);
include/linux/blkdev.h
882
if (!op_is_write(op))
include/linux/blkdev.h
897
switch (op) {
include/linux/blkpg.h
15
compat_int_t op;
include/linux/bnxt/hsi.h
9025
__le16 op;
include/linux/ceph/ceph_fs.h
314
extern const char *ceph_session_op_name(int op);
include/linux/ceph/ceph_fs.h
317
__le32 op;
include/linux/ceph/ceph_fs.h
367
#define IS_CEPH_MDS_OP_NEWINODE(op) (op == CEPH_MDS_OP_CREATE || \
include/linux/ceph/ceph_fs.h
368
op == CEPH_MDS_OP_MKNOD || \
include/linux/ceph/ceph_fs.h
369
op == CEPH_MDS_OP_MKDIR || \
include/linux/ceph/ceph_fs.h
370
op == CEPH_MDS_OP_SYMLINK)
include/linux/ceph/ceph_fs.h
372
extern const char *ceph_mds_op_name(int op);
include/linux/ceph/ceph_fs.h
505
__le32 op; /* mds op code */
include/linux/ceph/ceph_fs.h
521
__le32 op; /* mds op code */
include/linux/ceph/ceph_fs.h
545
__le32 op;
include/linux/ceph/ceph_fs.h
764
extern const char *ceph_cap_op_name(int op);
include/linux/ceph/ceph_fs.h
775
__le32 op; /* CEPH_CAP_OP_* */
include/linux/ceph/ceph_fs.h
880
__le32 op; /* CEPH_SNAP_OP_* */
include/linux/ceph/osd_client.h
141
u16 op; /* CEPH_OSD_OP_* */
include/linux/ceph/osd_client.h
176
__u8 op; /* CEPH_OSD_WATCH_OP_ */
include/linux/ceph/osd_client.h
559
int __ceph_alloc_sparse_ext_map(struct ceph_osd_req_op *op, int cnt);
include/linux/ceph/osd_client.h
568
static inline int ceph_alloc_sparse_ext_map(struct ceph_osd_req_op *op, int cnt)
include/linux/ceph/osd_client.h
573
return __ceph_alloc_sparse_ext_map(op, cnt);
include/linux/ceph/osd_client.h
630
static inline u64 ceph_sparse_ext_map_end(struct ceph_osd_req_op *op)
include/linux/ceph/osd_client.h
635
if (op->extent.sparse_ext_cnt == 0)
include/linux/ceph/osd_client.h
638
ext = &op->extent.sparse_ext[op->extent.sparse_ext_cnt - 1];
include/linux/ceph/osd_client.h
640
return ext->off + ext->len - op->extent.offset;
include/linux/ceph/rados.h
328
static inline int ceph_osd_op_type_lock(int op)
include/linux/ceph/rados.h
330
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_LOCK;
include/linux/ceph/rados.h
332
static inline int ceph_osd_op_type_data(int op)
include/linux/ceph/rados.h
334
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_DATA;
include/linux/ceph/rados.h
336
static inline int ceph_osd_op_type_attr(int op)
include/linux/ceph/rados.h
338
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_ATTR;
include/linux/ceph/rados.h
340
static inline int ceph_osd_op_type_exec(int op)
include/linux/ceph/rados.h
342
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_EXEC;
include/linux/ceph/rados.h
344
static inline int ceph_osd_op_type_pg(int op)
include/linux/ceph/rados.h
346
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_PG;
include/linux/ceph/rados.h
348
static inline int ceph_osd_op_type_multi(int op)
include/linux/ceph/rados.h
350
return (op & CEPH_OSD_OP_TYPE) == CEPH_OSD_OP_TYPE_MULTI;
include/linux/ceph/rados.h
353
static inline int ceph_osd_op_mode_subop(int op)
include/linux/ceph/rados.h
355
return (op & CEPH_OSD_OP_MODE) == CEPH_OSD_OP_MODE_SUB;
include/linux/ceph/rados.h
357
static inline int ceph_osd_op_mode_read(int op)
include/linux/ceph/rados.h
359
return (op & CEPH_OSD_OP_MODE_RD) &&
include/linux/ceph/rados.h
360
op != CEPH_OSD_OP_CALL;
include/linux/ceph/rados.h
362
static inline int ceph_osd_op_mode_modify(int op)
include/linux/ceph/rados.h
364
return op & CEPH_OSD_OP_MODE_WR;
include/linux/ceph/rados.h
377
extern const char *ceph_osd_op_name(int op);
include/linux/ceph/rados.h
491
__le16 op; /* CEPH_OSD_OP_* */
include/linux/ceph/rados.h
520
__u8 op; /* CEPH_OSD_WATCH_OP_* */
include/linux/crush/crush.h
44
__u32 op;
include/linux/dma/qcom-gpi-dma.h
79
enum i2c_op op;
include/linux/eventpoll.h
60
int do_epoll_ctl(int epfd, int op, int fd, struct epoll_event *epds,
include/linux/eventpoll.h
64
static inline int ep_op_has_event(int op)
include/linux/eventpoll.h
66
return op != EPOLL_CTL_DEL;
include/linux/filter.h
1543
u8 op;
include/linux/firmware/imx/sm.h
80
int scmi_imx_lmm_operation(u32 lmid, enum scmi_imx_lmm_op op, u32 flags);
include/linux/firmware/imx/sm.h
84
static inline int scmi_imx_lmm_operation(u32 lmid, enum scmi_imx_lmm_op op, u32 flags)
include/linux/firmware/mediatek/mtk-adsp-ipc.h
57
int mtk_adsp_ipc_send(struct mtk_adsp_ipc *ipc, unsigned int idx, uint32_t op);
include/linux/fortify-string.h
624
p_size_field, q_size_field, op) ({ \
include/linux/fortify-string.h
634
__q_size_field, FORTIFY_FUNC_ ##op), \
include/linux/fortify-string.h
635
#op ": detected field-spanning write (size %zu) of single %s (size %zu)\n", \
include/linux/fortify-string.h
643
__underlying_##op(p, q, __copy_size); \
include/linux/ftrace.h
282
struct ftrace_ops *op, struct ftrace_regs *fregs);
include/linux/ftrace.h
408
typedef int (*ftrace_ops_func_t)(struct ftrace_ops *op, unsigned long ip, enum ftrace_ops_cmd cmd);
include/linux/ftrace.h
481
#define do_for_each_ftrace_op(op, list) \
include/linux/ftrace.h
482
op = rcu_dereference_raw_check(list); \
include/linux/ftrace.h
488
#define while_for_each_ftrace_op(op) \
include/linux/ftrace.h
489
while (likely(op = rcu_dereference_raw_check((op)->next)) && \
include/linux/ftrace.h
490
unlikely((op) != &ftrace_list_end))
include/linux/ftrace.h
514
struct ftrace_ops *op, struct ftrace_regs *fregs);
include/linux/ftrace.h
67
struct ftrace_ops *op, struct ftrace_regs *fregs);
include/linux/futex.h
101
static inline long do_futex(u32 __user *uaddr, int op, u32 val,
include/linux/futex.h
81
long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout,
include/linux/fwnode.h
187
#define fwnode_has_op(fwnode, op) \
include/linux/fwnode.h
188
(!IS_ERR_OR_NULL(fwnode) && (fwnode)->ops && (fwnode)->ops->op)
include/linux/fwnode.h
190
#define fwnode_call_int_op(fwnode, op, ...) \
include/linux/fwnode.h
191
(fwnode_has_op(fwnode, op) ? \
include/linux/fwnode.h
192
(fwnode)->ops->op(fwnode, ## __VA_ARGS__) : (IS_ERR_OR_NULL(fwnode) ? -EINVAL : -ENXIO))
include/linux/fwnode.h
194
#define fwnode_call_bool_op(fwnode, op, ...) \
include/linux/fwnode.h
195
(fwnode_has_op(fwnode, op) ? \
include/linux/fwnode.h
196
(fwnode)->ops->op(fwnode, ## __VA_ARGS__) : false)
include/linux/fwnode.h
198
#define fwnode_call_ptr_op(fwnode, op, ...) \
include/linux/fwnode.h
199
(fwnode_has_op(fwnode, op) ? \
include/linux/fwnode.h
200
(fwnode)->ops->op(fwnode, ## __VA_ARGS__) : NULL)
include/linux/fwnode.h
201
#define fwnode_call_void_op(fwnode, op, ...) \
include/linux/fwnode.h
203
if (fwnode_has_op(fwnode, op)) \
include/linux/fwnode.h
204
(fwnode)->ops->op(fwnode, ## __VA_ARGS__); \
include/linux/genl_magic_func.h
122
#define DPRINT_TLA(a, op, b) do {} while (0)
include/linux/genl_magic_func.h
96
#define DPRINT_TLA(a, op, b) pr_info("%s %s %s\n", a, op, b);
include/linux/hisi_acc_qm.h
573
bool op);
include/linux/iopoll.h
102
op; \
include/linux/iopoll.h
143
#define read_poll_timeout(op, val, cond, sleep_us, timeout_us, \
include/linux/iopoll.h
145
poll_timeout_us((val) = op(args), cond, sleep_us, timeout_us, sleep_before_read)
include/linux/iopoll.h
170
#define read_poll_timeout_atomic(op, val, cond, sleep_us, timeout_us, \
include/linux/iopoll.h
172
poll_timeout_us_atomic((val) = op(args), cond, sleep_us, timeout_us, sleep_before_read)
include/linux/iopoll.h
192
#define readx_poll_timeout(op, addr, val, cond, sleep_us, timeout_us) \
include/linux/iopoll.h
193
read_poll_timeout(op, val, cond, sleep_us, timeout_us, false, addr)
include/linux/iopoll.h
212
#define readx_poll_timeout_atomic(op, addr, val, cond, delay_us, timeout_us) \
include/linux/iopoll.h
213
read_poll_timeout_atomic(op, val, cond, delay_us, timeout_us, false, addr)
include/linux/iopoll.h
34
#define poll_timeout_us(op, cond, sleep_us, timeout_us, sleep_before_op) \
include/linux/iopoll.h
48
op; \
include/linux/iopoll.h
85
#define poll_timeout_us_atomic(op, cond, delay_us, timeout_us, \
include/linux/keyctl.h
39
enum kernel_pkey_operation op : 8;
include/linux/kprobes.h
346
extern int arch_check_optimized_kprobe(struct optimized_kprobe *op);
include/linux/kprobes.h
347
extern int arch_prepare_optimized_kprobe(struct optimized_kprobe *op,
include/linux/kprobes.h
349
extern void arch_remove_optimized_kprobe(struct optimized_kprobe *op);
include/linux/kprobes.h
353
extern void arch_unoptimize_kprobe(struct optimized_kprobe *op);
include/linux/kprobes.h
354
extern int arch_within_optimized_kprobe(struct optimized_kprobe *op,
include/linux/kprobes.h
362
bool optprobe_queued_unopt(struct optimized_kprobe *op);
include/linux/lsm_audit.h
96
struct lsm_ioctlop_audit *op;
include/linux/lsm_hook_defs.h
422
LSM_HOOK(int, 0, audit_rule_init, u32 field, u32 op, char *rulestr,
include/linux/lsm_hook_defs.h
425
LSM_HOOK(int, 0, audit_rule_match, struct lsm_prop *prop, u32 field, u32 op,
include/linux/mISDNif.h
400
int op;
include/linux/minmax.h
131
#define __careful_op3(op, x, y, z, ux, uy, uz) ({ \
include/linux/minmax.h
134
#op"3("#x", "#y", "#z") signedness error"); \
include/linux/minmax.h
135
__cmp(op, ux, __cmp(op, uy, uz)); })
include/linux/minmax.h
250
#define __minmax_array(op, array, len) ({ \
include/linux/minmax.h
255
__element = op(__element, __array[__len]); \
include/linux/minmax.h
83
#define __cmp(op, x, y) ((x) __cmp_op_##op (y) ? (x) : (y))
include/linux/minmax.h
85
#define __cmp_once_unique(op, type, x, y, ux, uy) \
include/linux/minmax.h
86
({ type ux = (x); type uy = (y); __cmp(op, ux, uy); })
include/linux/minmax.h
88
#define __cmp_once(op, type, x, y) \
include/linux/minmax.h
89
__cmp_once_unique(op, type, x, y, __UNIQUE_ID(x_), __UNIQUE_ID(y_))
include/linux/minmax.h
91
#define __careful_cmp_once(op, x, y, ux, uy) ({ \
include/linux/minmax.h
94
#op"("#x", "#y") signedness error"); \
include/linux/minmax.h
95
__cmp(op, ux, uy); })
include/linux/minmax.h
97
#define __careful_cmp(op, x, y) \
include/linux/minmax.h
98
__careful_cmp_once(op, x, y, __UNIQUE_ID(x_), __UNIQUE_ID(y_))
include/linux/mlx4/cmd.h
270
u16 op, unsigned long timeout, int native);
include/linux/mlx4/cmd.h
274
u8 op_modifier, u16 op, unsigned long timeout,
include/linux/mlx4/cmd.h
278
op_modifier, op, timeout, native);
include/linux/mlx4/cmd.h
283
u32 in_modifier, u8 op_modifier, u16 op,
include/linux/mlx4/cmd.h
287
op_modifier, op, timeout, native);
include/linux/mlx4/cmd.h
296
u32 in_modifier, u8 op_modifier, u16 op,
include/linux/mlx4/cmd.h
300
op_modifier, op, timeout, native);
include/linux/mlx5/driver.h
853
u16 op;
include/linux/mlx5/mlx5_ifc.h
10724
u8 op[0x4];
include/linux/mlx5/qp.h
500
__be32 op;
include/linux/mmc/host.h
274
int (*uhs2_control)(struct mmc_host *host, enum sd_uhs2_operation op);
include/linux/mtd/nand-ecc-mtk.h
28
enum mtk_ecc_operation op;
include/linux/mtd/rawnand.h
1037
const struct nand_operation *op, bool check_only);
include/linux/mtd/rawnand.h
1101
const struct nand_operation *op,
include/linux/nvme-fc-driver.h
693
u8 op;
include/linux/packing.h
102
enum packing_op op, u8 quirks);
include/linux/pds/pds_adminq.h
1198
__le32 op;
include/linux/pds/pds_adminq.h
1272
#define PDS_FWCTL_RPC_OPCODE_GET_CMD(op) FIELD_GET(PDS_FWCTL_RPC_OPCODE_CMD_MASK, op)
include/linux/pds/pds_adminq.h
1273
#define PDS_FWCTL_RPC_OPCODE_GET_VER(op) FIELD_GET(PDS_FWCTL_RPC_OPCODE_VER_MASK, op)
include/linux/pds/pds_adminq.h
1312
__le32 op;
include/linux/pds/pds_adminq.h
1395
__le32 op;
include/linux/percpu-defs.h
315
extern void __this_cpu_preempt_check(const char *op);
include/linux/percpu-defs.h
317
static __always_inline void __this_cpu_preempt_check(const char *op) { }
include/linux/phylink.h
774
#define __phylink_do_bit(op, bm, mode) \
include/linux/phylink.h
775
op(ETHTOOL_LINK_MODE_ ## mode ## _BIT, bm)
include/linux/platform_data/cros_ec_commands.h
1703
uint32_t op;
include/linux/reset/bcm63xx_pmb.h
34
u32 off, u32 op)
include/linux/reset/bcm63xx_pmb.h
39
cmd = (PMC_PMBM_START | op | (addr & 0xff) << 12 | off);
include/linux/security.h
2187
int security_audit_rule_init(u32 field, u32 op, char *rulestr, void **lsmrule,
include/linux/security.h
2190
int security_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op,
include/linux/security.h
2196
static inline int security_audit_rule_init(u32 field, u32 op, char *rulestr,
include/linux/security.h
2208
u32 op, void *lsmrule)
include/linux/seq_file.h
25
const struct seq_operations *op;
include/linux/signal.h
131
#define _SIG_SET_BINOP(name, op) \
include/linux/signal.h
140
r->sig[3] = op(a3, b3); \
include/linux/signal.h
141
r->sig[2] = op(a2, b2); \
include/linux/signal.h
145
r->sig[1] = op(a1, b1); \
include/linux/signal.h
149
r->sig[0] = op(a0, b0); \
include/linux/signal.h
170
#define _SIG_SET_OP(name, op) \
include/linux/signal.h
174
case 4: set->sig[3] = op(set->sig[3]); \
include/linux/signal.h
175
set->sig[2] = op(set->sig[2]); \
include/linux/signal.h
177
case 2: set->sig[1] = op(set->sig[1]); \
include/linux/signal.h
179
case 1: set->sig[0] = op(set->sig[0]); \
include/linux/skbuff.h
4532
#define __it(x, op) (x -= sizeof(u##op))
include/linux/skbuff.h
4533
#define __it_diff(a, b, op) (*(u##op *)__it(a, op)) ^ (*(u##op *)__it(b, op))
include/linux/soc/qcom/apr.h
125
typedef int (*gpr_port_cb) (const struct gpr_resp_pkt *d, void *priv, int op);
include/linux/spi/spi-mem.h
345
int (*adjust_op_size)(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
347
const struct spi_mem_op *op);
include/linux/spi/spi-mem.h
349
const struct spi_mem_op *op);
include/linux/spi/spi-mem.h
358
const struct spi_mem_op *op,
include/linux/spi/spi-mem.h
408
const struct spi_mem_op *op,
include/linux/spi/spi-mem.h
412
const struct spi_mem_op *op,
include/linux/spi/spi-mem.h
416
const struct spi_mem_op *op);
include/linux/spi/spi-mem.h
420
const struct spi_mem_op *op,
include/linux/spi/spi-mem.h
428
const struct spi_mem_op *op,
include/linux/spi/spi-mem.h
435
const struct spi_mem_op *op)
include/linux/spi/spi-mem.h
441
int spi_mem_adjust_op_size(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
442
void spi_mem_adjust_op_freq(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
443
u64 spi_mem_calc_op_duration(struct spi_mem *mem, struct spi_mem_op *op);
include/linux/spi/spi-mem.h
446
const struct spi_mem_op *op);
include/linux/spi/spi-mem.h
449
const struct spi_mem_op *op);
include/linux/spi/spi-mem.h
468
const struct spi_mem_op *op,
include/linux/syscalls.h
339
asmlinkage long sys_io_uring_register(unsigned int fd, unsigned int op,
include/linux/syscalls.h
384
asmlinkage long sys_epoll_ctl(int epfd, int op, int fd,
include/linux/syscalls.h
569
asmlinkage long sys_futex(u32 __user *uaddr, int op, u32 val,
include/linux/syscalls.h
572
asmlinkage long sys_futex_time32(u32 __user *uaddr, int op, u32 val,
include/linux/syscalls.h
934
asmlinkage long sys_seccomp(unsigned int op, unsigned int flags,
include/linux/sysrq.h
48
int register_sysrq_key(u8 key, const struct sysrq_key_op *op);
include/linux/sysrq.h
49
int unregister_sysrq_key(u8 key, const struct sysrq_key_op *op);
include/linux/sysrq.h
65
static inline int register_sysrq_key(u8 key, const struct sysrq_key_op *op)
include/linux/sysrq.h
70
static inline int unregister_sysrq_key(u8 key, const struct sysrq_key_op *op)
include/linux/tty_driver.h
584
const struct tty_operations *op)
include/linux/tty_driver.h
586
driver->ops = op;
include/linux/uaccess.h
629
#define unsafe_op_wrap(op, err) do { if (unlikely(op)) goto err; } while (0)
include/linux/virtio_vsock.h
179
u16 op;
include/linux/vt_kern.h
35
int con_font_op(struct vc_data *vc, struct console_font_op *op);
include/media/drv-intf/exynos-fimc.h
152
#define fimc_pipeline_call(ent, op, args...) \
include/media/drv-intf/exynos-fimc.h
154
(((ent)->pipe->ops && (ent)->pipe->ops->op) ? \
include/media/drv-intf/exynos-fimc.h
155
(ent)->pipe->ops->op(((ent)->pipe), ##args) : -ENOIOCTLCMD)) \
include/memory/renesas-rpc-if.h
76
void rpcif_prepare(struct device *dev, const struct rpcif_op *op, u64 *offs,
include/net/bluetooth/hci.h
3391
#define hci_opcode_ogf(op) (op >> 10)
include/net/bluetooth/hci.h
3392
#define hci_opcode_ocf(op) (op & 0x03ff)
include/net/flow_dissector.h
175
__u8 op;
include/net/genetlink.h
287
struct genl_split_ops op;
include/net/ip_tunnels.h
455
int ip_tunnel_encap_add_ops(const struct ip_tunnel_encap_ops *op,
include/net/ip_tunnels.h
457
int ip_tunnel_encap_del_ops(const struct ip_tunnel_encap_ops *op,
include/net/ip_tunnels.h
49
#define __ipt_flag_op(op, ...) \
include/net/ip_tunnels.h
50
op(__VA_ARGS__, __IP_TUNNEL_FLAG_NUM)
include/net/lwtunnel.h
114
int lwtunnel_encap_add_ops(const struct lwtunnel_encap_ops *op,
include/net/lwtunnel.h
116
int lwtunnel_encap_del_ops(const struct lwtunnel_encap_ops *op,
include/net/lwtunnel.h
190
static inline int lwtunnel_encap_add_ops(const struct lwtunnel_encap_ops *op,
include/net/lwtunnel.h
197
static inline int lwtunnel_encap_del_ops(const struct lwtunnel_encap_ops *op,
include/net/nfc/nci.h
184
#define nci_opcode_gid(op) (__u8)(((op)&0x0f00)>>8)
include/net/nfc/nci.h
185
#define nci_opcode_oid(op) (__u8)((op)&0x003f)
include/net/sock.h
1875
int sock_setsockopt(struct socket *sock, int level, int op,
include/net/sock.h
3113
void bpf_skops_tx_timestamping(struct sock *sk, struct sk_buff *skb, int op);
include/net/sock.h
3115
static inline void bpf_skops_tx_timestamping(struct sock *sk, struct sk_buff *skb, int op)
include/net/tcp.h
2822
static inline int tcp_call_bpf(struct sock *sk, int op, u32 nargs, u32 *args)
include/net/tcp.h
2835
sock_ops.op = op;
include/net/tcp.h
2847
static inline int tcp_call_bpf_2arg(struct sock *sk, int op, u32 arg1, u32 arg2)
include/net/tcp.h
2851
return tcp_call_bpf(sk, op, 2, args);
include/net/tcp.h
2854
static inline int tcp_call_bpf_3arg(struct sock *sk, int op, u32 arg1, u32 arg2,
include/net/tcp.h
2859
return tcp_call_bpf(sk, op, 3, args);
include/net/tcp.h
2863
static inline int tcp_call_bpf(struct sock *sk, int op, u32 nargs, u32 *args)
include/net/tcp.h
2868
static inline int tcp_call_bpf_2arg(struct sock *sk, int op, u32 arg1, u32 arg2)
include/net/tcp.h
2873
static inline int tcp_call_bpf_3arg(struct sock *sk, int op, u32 arg1, u32 arg2,
include/net/xfrm.h
809
static inline struct audit_buffer *xfrm_audit_start(const char *op)
include/net/xfrm.h
819
audit_log_format(audit_buf, "op=%s", op);
include/rdma/ib_pack.h
47
#define IB_OPCODE(transport, op) \
include/rdma/ib_pack.h
48
IB_OPCODE_ ## transport ## _ ## op = \
include/rdma/ib_pack.h
49
IB_OPCODE_ ## transport + IB_OPCODE_ ## op
include/rdma/rdma_netlink.h
66
int len, int client, int op, int flags);
include/scsi/libfc.h
486
struct fc_frame *, unsigned int op,
include/scsi/libfc.h
975
unsigned int op,
include/scsi/scsi_cmnd.h
263
static inline void scsi_set_prot_op(struct scsi_cmnd *scmd, unsigned char op)
include/scsi/scsi_cmnd.h
265
scmd->prot_op = op;
include/sound/opl3.h
236
struct fm_operator op[4];
include/trace/events/afs.h
1043
enum afs_edit_dir_op op,
include/trace/events/afs.h
1050
TP_ARGS(dvnode, why, op, block, slot, f_vnode, f_unique, name),
include/trace/events/afs.h
1056
__field(enum afs_edit_dir_op, op)
include/trace/events/afs.h
1070
__entry->op = op;
include/trace/events/afs.h
1082
__print_symbolic(__entry->op, afs_edit_dir_ops),
include/trace/events/afs.h
1270
TP_PROTO(struct afs_operation *op, struct afs_fid *fid, unsigned int index, s32 abort),
include/trace/events/afs.h
1272
TP_ARGS(op, fid, index, abort),
include/trace/events/afs.h
1276
__field(unsigned int, op)
include/trace/events/afs.h
1282
__entry->op = op->debug_id;
include/trace/events/afs.h
1289
__entry->op, __entry->index,
include/trace/events/afs.h
1368
enum afs_flock_operation op),
include/trace/events/afs.h
1370
TP_ARGS(vnode, fl, op),
include/trace/events/afs.h
1376
__field(enum afs_flock_operation, op)
include/trace/events/afs.h
1386
__entry->op = op;
include/trace/events/afs.h
1395
__print_symbolic(__entry->op, afs_flock_operations),
include/trace/events/afs.h
1738
TP_PROTO(struct afs_operation *op, enum afs_rotate_trace reason, unsigned int extra),
include/trace/events/afs.h
1740
TP_ARGS(op, reason, extra),
include/trace/events/afs.h
1743
__field(unsigned int, op)
include/trace/events/afs.h
1753
__entry->op = op->debug_id;
include/trace/events/afs.h
1754
__entry->flags = op->flags;
include/trace/events/afs.h
1755
__entry->iteration = op->nr_iterations;
include/trace/events/afs.h
1756
__entry->server_index = op->server_index;
include/trace/events/afs.h
1757
__entry->addr_index = op->addr_index;
include/trace/events/afs.h
1763
__entry->op,
include/trace/events/afs.h
1780
__field(enum afs_fs_operation, op)
include/trace/events/afs.h
1787
__entry->op = call->operation_ID;
include/trace/events/afs.h
1801
__print_symbolic(__entry->op, afs_vl_operations) :
include/trace/events/afs.h
1802
__print_symbolic(__entry->op, afs_fs_operations),
include/trace/events/afs.h
1809
TP_PROTO(const struct afs_operation *op, const struct afs_call *call),
include/trace/events/afs.h
1811
TP_ARGS(op, call),
include/trace/events/afs.h
1816
__field(unsigned int, op)
include/trace/events/afs.h
1823
__entry->op = op->debug_id;
include/trace/events/afs.h
1824
__entry->sreq = op->fetch.subreq->debug_index;
include/trace/events/afs.h
1825
__entry->rreq = op->fetch.subreq->rreq->debug_id;
include/trace/events/afs.h
1826
__entry->op_flags = op->flags;
include/trace/events/afs.h
1833
__entry->op,
include/trace/events/afs.h
670
__field(u32, op)
include/trace/events/afs.h
678
__entry->op = call->operation_ID;
include/trace/events/afs.h
687
__print_symbolic(__entry->op, yfs_cm_operations) :
include/trace/events/afs.h
688
__print_symbolic(__entry->op, afs_cm_operations),
include/trace/events/afs.h
695
TP_PROTO(unsigned int call_debug_id, enum afs_call_trace op,
include/trace/events/afs.h
698
TP_ARGS(call_debug_id, op, ref, outstanding, where),
include/trace/events/afs.h
702
__field(int, op)
include/trace/events/afs.h
710
__entry->op = op;
include/trace/events/afs.h
718
__print_symbolic(__entry->op, afs_call_traces),
include/trace/events/afs.h
731
__field(enum afs_fs_operation, op)
include/trace/events/afs.h
737
__entry->op = call->operation_ID;
include/trace/events/afs.h
752
__print_symbolic(__entry->op, afs_fs_operations))
include/trace/events/afs.h
764
__field(enum afs_fs_operation, op)
include/trace/events/afs.h
771
__entry->op = call->operation_ID;
include/trace/events/afs.h
786
__print_symbolic(__entry->op, afs_fs_operations),
include/trace/events/afs.h
798
__field(enum afs_fs_operation, op)
include/trace/events/afs.h
806
__entry->op = call->operation_ID;
include/trace/events/afs.h
823
__print_symbolic(__entry->op, afs_fs_operations),
include/trace/events/afs.h
835
__field(enum afs_fs_operation, op)
include/trace/events/afs.h
845
__entry->op = call->operation_ID;
include/trace/events/afs.h
864
__print_symbolic(__entry->op, afs_fs_operations),
include/trace/events/afs.h
876
__field(enum afs_vl_operation, op)
include/trace/events/afs.h
881
__entry->op = call->operation_ID;
include/trace/events/afs.h
886
__print_symbolic(__entry->op, afs_vl_operations))
include/trace/events/bcache.h
333
TP_PROTO(struct btree *b, struct bkey *k, unsigned op, unsigned status),
include/trace/events/bcache.h
334
TP_ARGS(b, k, op, status),
include/trace/events/bcache.h
343
__field(u8, op )
include/trace/events/bcache.h
354
__entry->op = op;
include/trace/events/bcache.h
359
__entry->status, __entry->op,
include/trace/events/btrfs.h
1659
u64 reserved, int op),
include/trace/events/btrfs.h
1661
TP_ARGS(inode, start, len, reserved, op),
include/trace/events/btrfs.h
1669
__field( int, op )
include/trace/events/btrfs.h
1678
__entry->op = op;
include/trace/events/btrfs.h
1684
__print_flags((unsigned long)__entry->op, "",
include/trace/events/btrfs.h
1692
u64 reserved, int op),
include/trace/events/btrfs.h
1694
TP_ARGS(inode, start, len, reserved, op)
include/trace/events/btrfs.h
1700
u64 reserved, int op),
include/trace/events/btrfs.h
1702
TP_ARGS(inode, start, len, reserved, op)
include/trace/events/ceph.h
106
__entry->op = req->r_op;
include/trace/events/ceph.h
113
ceph_mds_op_name(__entry->op),
include/trace/events/ceph.h
125
__field(int, op)
include/trace/events/ceph.h
130
__entry->op = req->r_op;
include/trace/events/ceph.h
135
ceph_mds_op_name(__entry->op))
include/trace/events/ceph.h
146
__field(int, op)
include/trace/events/ceph.h
152
__entry->op = req->r_op;
include/trace/events/ceph.h
158
ceph_mds_op_name(__entry->op),
include/trace/events/ceph.h
170
__field(int, op)
include/trace/events/ceph.h
177
__entry->op = req->r_op;
include/trace/events/ceph.h
184
ceph_mds_op_name(__entry->op),
include/trace/events/ceph.h
192
int op,
include/trace/events/ceph.h
197
TP_ARGS(mdsc, session, op, vino, inode, seq, mseq, issue_seq),
include/trace/events/ceph.h
201
__field(int, op)
include/trace/events/ceph.h
211
__entry->op = op;
include/trace/events/ceph.h
221
ceph_cap_op_name(__entry->op),
include/trace/events/ceph.h
60
__field(int, op)
include/trace/events/ceph.h
69
__entry->op = req->r_op;
include/trace/events/ceph.h
85
ceph_mds_op_name(__entry->op),
include/trace/events/ceph.h
99
__field(int, op)
include/trace/events/f2fs.h
1184
__field(enum req_op, op)
include/trace/events/f2fs.h
1196
__entry->op = fio->op;
include/trace/events/f2fs.h
1208
show_bio_type(__entry->op, __entry->op_flags),
include/trace/events/f2fs.h
1240
__field(enum req_op, op)
include/trace/events/f2fs.h
1250
__entry->op = bio_op(bio);
include/trace/events/f2fs.h
1260
show_bio_type(__entry->op, __entry->op_flags),
include/trace/events/f2fs.h
76
#define show_bio_type(op,op_flags) show_bio_op(op), \
include/trace/events/f2fs.h
79
#define show_bio_op(op) blk_op_str(op)
include/trace/events/fsi_master_ast_cf.h
11
TP_PROTO(const struct fsi_master_acf *master, uint32_t op),
include/trace/events/fsi_master_ast_cf.h
12
TP_ARGS(master, op),
include/trace/events/fsi_master_ast_cf.h
15
__field(uint32_t, op)
include/trace/events/fsi_master_ast_cf.h
19
__entry->op = op;
include/trace/events/fsi_master_ast_cf.h
22
__entry->master_idx, __entry->op
include/trace/events/rxrpc.h
683
TP_PROTO(unsigned int local_debug_id, enum rxrpc_local_trace op,
include/trace/events/rxrpc.h
686
TP_ARGS(local_debug_id, op, ref, usage),
include/trace/events/rxrpc.h
690
__field(int, op)
include/trace/events/rxrpc.h
697
__entry->op = op;
include/trace/events/rxrpc.h
704
__print_symbolic(__entry->op, rxrpc_local_traces),
include/trace/events/rxrpc.h
794
enum rxrpc_client_trace op),
include/trace/events/rxrpc.h
796
TP_ARGS(conn, channel, op),
include/trace/events/rxrpc.h
803
__field(enum rxrpc_client_trace, op)
include/trace/events/rxrpc.h
810
__entry->op = op;
include/trace/events/rxrpc.h
817
__print_symbolic(__entry->op, rxrpc_client_traces),
include/trace/events/spi-mem.h
20
TP_PROTO(struct spi_mem *mem, const struct spi_mem_op *op),
include/trace/events/spi-mem.h
21
TP_ARGS(mem, op),
include/trace/events/spi-mem.h
25
__dynamic_array(u8, op, 1 + op->addr.nbytes + op->dummy.nbytes)
include/trace/events/spi-mem.h
26
__dynamic_array(u8, data, op->data.dir == SPI_MEM_DATA_OUT ?
include/trace/events/spi-mem.h
27
min(op->data.nbytes, 64) : 0)
include/trace/events/spi-mem.h
43
__entry->max_freq = op->max_freq ?: mem->spi->max_speed_hz;
include/trace/events/spi-mem.h
45
__entry->cmd_buswidth = op->cmd.buswidth;
include/trace/events/spi-mem.h
46
__entry->cmd_dtr = op->cmd.dtr;
include/trace/events/spi-mem.h
47
*((u8 *)__get_dynamic_array(op)) = op->cmd.opcode;
include/trace/events/spi-mem.h
49
__entry->addr_buswidth = op->addr.buswidth;
include/trace/events/spi-mem.h
50
__entry->addr_dtr = op->addr.dtr;
include/trace/events/spi-mem.h
51
for (i = 0; i < op->addr.nbytes; i++)
include/trace/events/spi-mem.h
52
((u8 *)__get_dynamic_array(op))[i + 1] =
include/trace/events/spi-mem.h
53
op->addr.val >> (8 * (op->addr.nbytes - i - 1));
include/trace/events/spi-mem.h
55
memset(((u8 *)__get_dynamic_array(op)) + op->addr.nbytes + 1,
include/trace/events/spi-mem.h
56
0xff, op->dummy.nbytes);
include/trace/events/spi-mem.h
58
__entry->data_len = op->data.nbytes;
include/trace/events/spi-mem.h
59
__entry->data_buswidth = op->data.buswidth;
include/trace/events/spi-mem.h
60
__entry->data_dtr = op->data.dtr;
include/trace/events/spi-mem.h
61
if (op->data.dir == SPI_MEM_DATA_OUT)
include/trace/events/spi-mem.h
62
memcpy(__get_dynamic_array(data), op->data.buf.out,
include/trace/events/spi-mem.h
72
__get_dynamic_array_len(op), __get_dynamic_array(op),
include/trace/events/spi-mem.h
78
TP_PROTO(struct spi_mem *mem, const struct spi_mem_op *op),
include/trace/events/spi-mem.h
79
TP_ARGS(mem, op),
include/trace/events/spi-mem.h
83
__dynamic_array(u8, data, op->data.dir == SPI_MEM_DATA_IN ?
include/trace/events/spi-mem.h
84
min(op->data.nbytes, 64) : 0)
include/trace/events/spi-mem.h
90
__entry->data_len = op->data.nbytes;
include/trace/events/spi-mem.h
91
if (op->data.dir == SPI_MEM_DATA_IN)
include/trace/events/spi-mem.h
92
memcpy(__get_dynamic_array(data), op->data.buf.in,
include/trace/events/vsock_virtio_transport_common.h
105
op,
include/trace/events/vsock_virtio_transport_common.h
117
__field(__u16, op)
include/trace/events/vsock_virtio_transport_common.h
129
__entry->op = op;
include/trace/events/vsock_virtio_transport_common.h
140
show_op(__entry->op),
include/trace/events/vsock_virtio_transport_common.h
45
__u16 op,
include/trace/events/vsock_virtio_transport_common.h
54
op,
include/trace/events/vsock_virtio_transport_common.h
65
__field(__u16, op)
include/trace/events/vsock_virtio_transport_common.h
76
__entry->op = op;
include/trace/events/vsock_virtio_transport_common.h
85
show_op(__entry->op),
include/trace/events/vsock_virtio_transport_common.h
95
__u16 op,
include/trace/events/xen.h
114
TP_PROTO(unsigned long op, size_t args, enum xen_mc_extend_args res),
include/trace/events/xen.h
115
TP_ARGS(op, args, res),
include/trace/events/xen.h
117
__field(unsigned int, op)
include/trace/events/xen.h
121
TP_fast_assign(__entry->op = op;
include/trace/events/xen.h
125
__entry->op, xen_hypercall_name(__entry->op),
include/trace/events/xen.h
40
__field(unsigned int, op)
include/trace/events/xen.h
44
TP_fast_assign(__entry->op = mc->op;
include/trace/events/xen.h
50
__entry->op, xen_hypercall_name(__entry->op),
include/uapi/drm/amdgpu_drm.h
296
__u32 op;
include/uapi/drm/amdgpu_drm.h
355
__u32 op;
include/uapi/drm/amdgpu_drm.h
590
__u32 op;
include/uapi/drm/amdgpu_drm.h
610
__u32 op;
include/uapi/drm/amdgpu_drm.h
705
__u32 op;
include/uapi/drm/amdgpu_drm.h
826
__u32 op;
include/uapi/drm/asahi_drm.h
456
__u32 op;
include/uapi/drm/etnaviv_drm.h
119
__u32 op; /* in, mask of ETNA_PREP_x */
include/uapi/drm/habanalabs_accel.h
1378
__u32 op;
include/uapi/drm/habanalabs_accel.h
1415
__u32 op;
include/uapi/drm/habanalabs_accel.h
2022
__u32 op;
include/uapi/drm/habanalabs_accel.h
2182
__u32 op;
include/uapi/drm/lima_drm.h
136
__u32 op; /* in, CPU want to read/write this buffer */
include/uapi/drm/msm_drm.h
211
__u32 op; /* in, mask of MSM_PREP_x */
include/uapi/drm/msm_drm.h
361
__u32 op;
include/uapi/drm/msm_drm.h
409
struct drm_msm_vm_bind_op op;
include/uapi/drm/nouveau_drm.h
298
__u32 op;
include/uapi/drm/omap_drm.h
79
__u32 op; /* mask of omap_gem_op (in) */
include/uapi/drm/omap_drm.h
84
__u32 op; /* mask of omap_gem_op (in) */
include/uapi/drm/radeon_drm.h
921
__u32 op; /* RADEON_GEM_OP_* */
include/uapi/drm/vmwgfx_drm.h
1081
enum drm_vmw_synccpu_op op;
include/uapi/drm/xe_drm.h
1154
__u32 op;
include/uapi/drm/xe_drm.h
1573
__u16 op;
include/uapi/fwctl/pds.h
49
__u32 op;
include/uapi/linux/blkpg.h
12
int op;
include/uapi/linux/bpf.h
6899
__u32 op;
include/uapi/linux/futex.h
208
#define FUTEX_OP(op, oparg, cmp, cmparg) \
include/uapi/linux/futex.h
209
(((op & 0xf) << 28) | ((cmp & 0xf) << 24) \
include/uapi/linux/io_uring.h
1138
__u32 op; /* see enum zcrx_ctrl_op */
include/uapi/linux/io_uring.h
799
__u8 op;
include/uapi/linux/iommufd.h
342
__u16 op;
include/uapi/linux/iommufd.h
379
__u16 op;
include/uapi/linux/ipc.h
79
#define IPCCALL(version,op) ((version)<<16 | (op))
include/uapi/linux/kd.h
160
unsigned int op; /* operation code KD_FONT_OP_* */
include/uapi/linux/kfd_ioctl.h
1541
__u32 op;
include/uapi/linux/kfd_ioctl.h
717
__u32 op;
include/uapi/linux/kfd_ioctl.h
870
__u32 op;
include/uapi/linux/netfilter/ipset/ip_set.h
270
__u8 op;
include/uapi/linux/netfilter/ipset/ip_set.h
276
__u8 op;
include/uapi/linux/netfilter/ipset/ip_set.h
290
unsigned int op;
include/uapi/linux/netfilter/ipset/ip_set.h
300
unsigned int op;
include/uapi/linux/netfilter/ipset/ip_set.h
308
unsigned int op;
include/uapi/linux/ptrace.h
84
__u8 op; /* PTRACE_SYSCALL_INFO_* */
include/uapi/linux/rfkill.h
119
__u8 op;
include/uapi/linux/rfkill.h
96
__u8 op;
include/uapi/linux/target_core_user.h
91
static inline void tcmu_hdr_set_op(__u32 *len_op, enum tcmu_opcode op)
include/uapi/linux/target_core_user.h
94
*len_op |= (op & TCMU_OP_MASK);
include/uapi/linux/tc_ematch/tc_em_meta.h
20
__u8 op;
include/uapi/linux/tee.h
470
__u32 op;
include/uapi/linux/um_timetravel.h
27
__u32 op;
include/uapi/linux/vfio.h
1480
__u32 op;
include/uapi/linux/vfio.h
1790
__u32 op;
include/uapi/linux/virtio_crypto.h
179
__le32 op;
include/uapi/linux/virtio_crypto.h
42
#define VIRTIO_CRYPTO_OPCODE(service, op) (((service) << 8) | (op))
include/uapi/linux/virtio_crypto.h
95
__le32 op;
include/uapi/linux/virtio_pcidev.h
57
__u8 op;
include/uapi/linux/virtio_vsock.h
63
__le16 op; /* enum virtio_vsock_op */
include/uapi/linux/vsockmon.h
39
__le16 op; /* enum af_vsockmon_op */
include/uapi/rdma/rdma_netlink.h
24
#define RDMA_NL_GET_TYPE(client, op) ((client << 10) + op)
include/uapi/sound/asound_fm.h
28
unsigned char op; /* operator cell (0 or 1) */
include/uapi/xen/privcmd.h
42
__u64 op;
include/video/pm3fb.h
742
#define PM3LogicalOpMode_LogicOp(op) (((op) & 0xf) << 1)
include/video/pm3fb.h
747
#define PM3LogicalOpMode_Background_LogicOp(op) (((op) & 0xf) << 7)
include/xen/arm/hypercall.h
52
unsigned long HYPERVISOR_hvm_op(int op, void *arg);
include/xen/arm/hypercall.h
60
static inline int HYPERVISOR_platform_op(struct xen_platform_op *op)
include/xen/arm/hypercall.h
62
op->interface_version = XENPF_INTERFACE_VERSION;
include/xen/arm/hypercall.h
63
return HYPERVISOR_platform_op_raw(op);
include/xen/hvm.h
31
if (op >= ARRAY_SIZE(names))
include/xen/hvm.h
34
if (!names[op])
include/xen/hvm.h
37
return names[op];
include/xen/hvm.h
9
static const char *param_name(int op)
include/xen/interface/grant_table.h
516
uint32_t op;
include/xen/interface/io/displif.h
873
} op;
include/xen/interface/io/displif.h
884
} op;
include/xen/interface/io/displif.h
894
} op;
include/xen/interface/io/pciif.h
91
struct xen_pci_op op;
include/xen/interface/io/sndif.h
1007
} op;
include/xen/interface/io/sndif.h
1028
} op;
include/xen/interface/physdev.h
292
uint8_t op;
include/xen/interface/xen.h
473
xen_ulong_t op;
io_uring/epoll.c
18
int op;
io_uring/epoll.c
37
epoll->op = READ_ONCE(sqe->len);
io_uring/epoll.c
40
if (ep_op_has_event(epoll->op)) {
io_uring/epoll.c
57
ret = do_epoll_ctl(ie->epfd, ie->op, ie->fd, &ie->event, force_nonblock);
io_uring/openclose.c
126
struct open_flags op;
io_uring/openclose.c
133
ret = build_open_flags(&open->how, &op);
io_uring/openclose.c
136
nonblock_set = op.open_flag & O_NONBLOCK;
io_uring/openclose.c
140
op.lookup_flags |= LOOKUP_CACHED;
io_uring/openclose.c
141
op.open_flag |= O_NONBLOCK;
io_uring/openclose.c
150
file = do_file_open(open->dfd, name, &op);
io_uring/register.c
62
p->ops[i].op = i;
io_uring/zcrx.c
1183
switch (ctrl.op) {
kernel/audit.c
1638
static void audit_log_multicast(int group, const char *op, int err)
kernel/audit.c
1665
audit_log_format(ab, " nl-mcgrp=%d op=%s res=%d", group, op, !err);
kernel/audit.h
237
extern int audit_comparator(const u32 left, const u32 op, const u32 right);
kernel/audit.h
238
extern int audit_uid_comparator(kuid_t left, u32 op, kuid_t right);
kernel/audit.h
239
extern int audit_gid_comparator(kgid_t left, u32 op, kgid_t right);
kernel/audit.h
276
u32 op);
kernel/audit.h
298
extern int audit_make_tree(struct audit_krule *rule, char *pathname, u32 op);
kernel/audit.h
331
#define audit_make_tree(rule, str, op) -EINVAL
kernel/audit_fsnotify.c
116
static void audit_mark_log_rule_change(struct audit_fsnotify_mark *audit_mark, char *op)
kernel/audit_fsnotify.c
127
audit_log_format(ab, " op=%s path=", op);
kernel/audit_tree.c
725
int audit_make_tree(struct audit_krule *rule, char *pathname, u32 op)
kernel/audit_tree.c
731
op != Audit_equal ||
kernel/audit_watch.c
178
int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op)
kernel/audit_watch.c
188
op != Audit_equal ||
kernel/audit_watch.c
227
static void audit_watch_log_rule_change(struct audit_krule *r, struct audit_watch *w, char *op)
kernel/audit_watch.c
237
audit_log_format(ab, "op=%s path=", op);
kernel/auditfilter.c
1205
int audit_comparator(u32 left, u32 op, u32 right)
kernel/auditfilter.c
1207
switch (op) {
kernel/auditfilter.c
1229
int audit_uid_comparator(kuid_t left, u32 op, kuid_t right)
kernel/auditfilter.c
1231
switch (op) {
kernel/auditfilter.c
1251
int audit_gid_comparator(kgid_t left, u32 op, kgid_t right)
kernel/auditfilter.c
1253
switch (op) {
kernel/auditfilter.c
1354
result = audit_comparator(pid, f->op, f->val);
kernel/auditfilter.c
1357
result = audit_uid_comparator(current_uid(), f->op, f->uid);
kernel/auditfilter.c
1360
result = audit_gid_comparator(current_gid(), f->op, f->gid);
kernel/auditfilter.c
1364
f->op, f->uid);
kernel/auditfilter.c
1368
f->op, f->val);
kernel/auditfilter.c
1371
result = audit_comparator(msgtype, f->op, f->val);
kernel/auditfilter.c
1381
&prop, f->type, f->op,
kernel/auditfilter.c
1387
if (f->op == Audit_not_equal)
kernel/auditfilter.c
159
(f->op != Audit_equal && f->op != Audit_not_equal))
kernel/auditfilter.c
318
static u32 audit_to_op(u32 op)
kernel/auditfilter.c
321
for (n = Audit_equal; n < Audit_bad && audit_ops[n] != op; n++)
kernel/auditfilter.c
391
if (f->op == Audit_bitmask || f->op == Audit_bittest)
kernel/auditfilter.c
411
if (f->op != Audit_not_equal && f->op != Audit_equal)
kernel/auditfilter.c
471
f->op = audit_to_op(data->fieldflags[i]);
kernel/auditfilter.c
472
if (f->op == Audit_bad)
kernel/auditfilter.c
531
err = security_audit_rule_init(f->type, f->op, str,
kernel/auditfilter.c
549
err = audit_to_watch(&entry->rule, str, f_val, f->op);
kernel/auditfilter.c
562
err = audit_make_tree(&entry->rule, str, f->op);
kernel/auditfilter.c
608
if (entry->rule.inode_f && entry->rule.inode_f->op == Audit_not_equal)
kernel/auditfilter.c
653
data->fieldflags[i] = audit_ops[f->op];
kernel/auditfilter.c
718
a->fields[i].op != b->fields[i].op)
kernel/auditfilter.c
801
ret = security_audit_rule_init(df->type, df->op, df->lsm_str,
kernel/auditsc.c
122
enum audit_nfcfgop op;
kernel/auditsc.c
1876
void __audit_uring_entry(u8 op)
kernel/auditsc.c
1889
ctx->uring_op = op;
kernel/auditsc.c
2266
f->op, f->val)
kernel/auditsc.c
2384
f->op, f->val)
kernel/auditsc.c
2866
enum audit_nfcfgop op, gfp_t gfp)
kernel/auditsc.c
2875
name, af, nentries, audit_nfcfgs[op].s);
kernel/auditsc.c
341
rc = audit_uid_comparator(uid, f->op, name->uid);
kernel/auditsc.c
348
rc = audit_uid_comparator(uid, f->op, n->uid);
kernel/auditsc.c
365
rc = audit_gid_comparator(gid, f->op, name->gid);
kernel/auditsc.c
372
rc = audit_gid_comparator(gid, f->op, n->gid);
kernel/auditsc.c
408
return audit_uid_comparator(cred->uid, f->op,
kernel/auditsc.c
411
return audit_uid_comparator(cred->uid, f->op, cred->euid);
kernel/auditsc.c
413
return audit_uid_comparator(cred->uid, f->op, cred->suid);
kernel/auditsc.c
415
return audit_uid_comparator(cred->uid, f->op, cred->fsuid);
kernel/auditsc.c
418
return audit_uid_comparator(audit_get_loginuid(tsk), f->op,
kernel/auditsc.c
421
return audit_uid_comparator(audit_get_loginuid(tsk), f->op,
kernel/auditsc.c
424
return audit_uid_comparator(audit_get_loginuid(tsk), f->op,
kernel/auditsc.c
428
return audit_uid_comparator(cred->euid, f->op, cred->suid);
kernel/auditsc.c
430
return audit_uid_comparator(cred->euid, f->op, cred->fsuid);
kernel/auditsc.c
433
return audit_uid_comparator(cred->suid, f->op, cred->fsuid);
kernel/auditsc.c
436
return audit_gid_comparator(cred->gid, f->op, cred->egid);
kernel/auditsc.c
438
return audit_gid_comparator(cred->gid, f->op, cred->sgid);
kernel/auditsc.c
440
return audit_gid_comparator(cred->gid, f->op, cred->fsgid);
kernel/auditsc.c
443
return audit_gid_comparator(cred->egid, f->op, cred->sgid);
kernel/auditsc.c
445
return audit_gid_comparator(cred->egid, f->op, cred->fsgid);
kernel/auditsc.c
448
return audit_gid_comparator(cred->sgid, f->op, cred->fsgid);
kernel/auditsc.c
490
result = audit_comparator(pid, f->op, f->val);
kernel/auditsc.c
496
result = audit_comparator(ctx->ppid, f->op, f->val);
kernel/auditsc.c
501
if (f->op == Audit_not_equal)
kernel/auditsc.c
505
result = audit_uid_comparator(cred->uid, f->op, f->uid);
kernel/auditsc.c
508
result = audit_uid_comparator(cred->euid, f->op, f->uid);
kernel/auditsc.c
511
result = audit_uid_comparator(cred->suid, f->op, f->uid);
kernel/auditsc.c
514
result = audit_uid_comparator(cred->fsuid, f->op, f->uid);
kernel/auditsc.c
517
result = audit_gid_comparator(cred->gid, f->op, f->gid);
kernel/auditsc.c
518
if (f->op == Audit_equal) {
kernel/auditsc.c
521
} else if (f->op == Audit_not_equal) {
kernel/auditsc.c
527
result = audit_gid_comparator(cred->egid, f->op, f->gid);
kernel/auditsc.c
528
if (f->op == Audit_equal) {
kernel/auditsc.c
531
} else if (f->op == Audit_not_equal) {
kernel/auditsc.c
537
result = audit_gid_comparator(cred->sgid, f->op, f->gid);
kernel/auditsc.c
540
result = audit_gid_comparator(cred->fsgid, f->op, f->gid);
kernel/auditsc.c
544
result = audit_comparator(sessionid, f->op, f->val);
kernel/auditsc.c
547
result = audit_comparator(tsk->personality, f->op, f->val);
kernel/auditsc.c
551
result = audit_comparator(ctx->arch, f->op, f->val);
kernel/auditsc.c
556
result = audit_comparator(ctx->return_code, f->op, f->val);
kernel/auditsc.c
561
result = audit_comparator(ctx->return_valid, f->op, AUDITSC_SUCCESS);
kernel/auditsc.c
563
result = audit_comparator(ctx->return_valid, f->op, AUDITSC_FAILURE);
kernel/auditsc.c
568
if (audit_comparator(MAJOR(name->dev), f->op, f->val) ||
kernel/auditsc.c
569
audit_comparator(MAJOR(name->rdev), f->op, f->val))
kernel/auditsc.c
573
if (audit_comparator(MAJOR(n->dev), f->op, f->val) ||
kernel/auditsc.c
574
audit_comparator(MAJOR(n->rdev), f->op, f->val)) {
kernel/auditsc.c
583
if (audit_comparator(MINOR(name->dev), f->op, f->val) ||
kernel/auditsc.c
584
audit_comparator(MINOR(name->rdev), f->op, f->val))
kernel/auditsc.c
588
if (audit_comparator(MINOR(n->dev), f->op, f->val) ||
kernel/auditsc.c
589
audit_comparator(MINOR(n->rdev), f->op, f->val)) {
kernel/auditsc.c
598
result = audit_comparator(name->ino, f->op, f->val);
kernel/auditsc.c
601
if (audit_comparator(n->ino, f->op, f->val)) {
kernel/auditsc.c
610
result = audit_uid_comparator(name->uid, f->op, f->uid);
kernel/auditsc.c
613
if (audit_uid_comparator(n->uid, f->op, f->uid)) {
kernel/auditsc.c
622
result = audit_gid_comparator(name->gid, f->op, f->gid);
kernel/auditsc.c
625
if (audit_gid_comparator(n->gid, f->op, f->gid)) {
kernel/auditsc.c
637
if (f->op == Audit_not_equal)
kernel/auditsc.c
644
if (f->op == Audit_not_equal)
kernel/auditsc.c
650
f->op, f->uid);
kernel/auditsc.c
653
result = audit_comparator(audit_loginuid_set(tsk), f->op, f->val);
kernel/auditsc.c
658
f->op, f->val);
kernel/auditsc.c
687
f->op,
kernel/auditsc.c
704
f->op,
kernel/auditsc.c
711
f->op,
kernel/auditsc.c
722
f->type, f->op,
kernel/auditsc.c
732
result = audit_comparator(ctx->argv[f->type-AUDIT_ARG0], f->op, f->val);
kernel/auditsc.c
740
if (f->op == Audit_not_equal)
kernel/auditsc.c
745
if (f->op == Audit_not_equal)
kernel/auditsc.c
831
unsigned long op)
kernel/auditsc.c
837
if (audit_in_mask(&e->rule, op) &&
kernel/bpf/bpf_iter.c
127
p = seq->op->start(seq, &seq->index);
kernel/bpf/bpf_iter.c
132
seq->op->stop(seq, p);
kernel/bpf/bpf_iter.c
137
err = seq->op->show(seq, p);
kernel/bpf/bpf_iter.c
147
seq->op->stop(seq, p);
kernel/bpf/bpf_iter.c
158
p = seq->op->next(seq, p, &seq->index);
kernel/bpf/bpf_iter.c
162
seq->op->next);
kernel/bpf/bpf_iter.c
178
seq->op->stop(seq, p);
kernel/bpf/bpf_iter.c
184
err = seq->op->show(seq, p);
kernel/bpf/bpf_iter.c
193
seq->op->stop(seq, p);
kernel/bpf/bpf_iter.c
205
seq->op->stop(seq, NULL);
kernel/bpf/bpf_iter.c
210
seq->op->stop(seq, p);
kernel/bpf/btf.c
8303
static int btf_module_notify(struct notifier_block *nb, unsigned long op,
kernel/bpf/btf.c
8312
(op != MODULE_STATE_COMING && op != MODULE_STATE_LIVE &&
kernel/bpf/btf.c
8313
op != MODULE_STATE_GOING))
kernel/bpf/btf.c
8316
switch (op) {
kernel/bpf/helpers.c
1112
enum bpf_async_op op;
kernel/bpf/helpers.c
1444
static int bpf_async_schedule_op(struct bpf_async_cb *cb, enum bpf_async_op op,
kernel/bpf/helpers.c
1468
cmd->op = op;
kernel/bpf/helpers.c
1613
static void bpf_async_process_op(struct bpf_async_cb *cb, u32 op,
kernel/bpf/helpers.c
1620
switch (op) {
kernel/bpf/helpers.c
1633
switch (op) {
kernel/bpf/helpers.c
1662
bpf_async_process_op(cb, cmd->op, cmd->nsec, cmd->mode);
kernel/bpf/syscall.c
2309
static void bpf_audit_prog(const struct bpf_prog *prog, unsigned int op)
kernel/bpf/syscall.c
2314
if (WARN_ON_ONCE(op >= BPF_AUDIT_MAX))
kernel/bpf/syscall.c
2324
prog->aux->id, bpf_audit_str[op]);
kernel/bpf/task_iter.c
428
enum bpf_task_vma_iter_find_op op;
kernel/bpf/task_iter.c
485
op = task_vma_iter_find_vma;
kernel/bpf/task_iter.c
492
op = task_vma_iter_next_vma;
kernel/bpf/task_iter.c
504
op = task_vma_iter_first_vma;
kernel/bpf/task_iter.c
512
op = task_vma_iter_find_vma;
kernel/bpf/task_iter.c
525
switch (op) {
kernel/bpf/verifier.c
14760
const char *op = BPF_OP(insn->code) == BPF_ADD ? "add" : "sub";
kernel/bpf/verifier.c
14774
dst, op, err);
kernel/bpf/verifier.c
14778
dst, op, err);
kernel/bpf/verifier.c
22326
u8 op;
kernel/bpf/verifier.c
22328
op = BPF_OP(code);
kernel/bpf/verifier.c
22330
return op != BPF_JA;
kernel/bpf/verifier.c
22335
return op != BPF_JA && op != BPF_EXIT && op != BPF_CALL;
kernel/bpf/verifier.c
3799
u8 code, class, op;
kernel/bpf/verifier.c
3803
op = BPF_OP(code);
kernel/bpf/verifier.c
3808
if (op == BPF_EXIT)
kernel/bpf/verifier.c
3810
if (op == BPF_CALL) {
kernel/bpf/verifier.c
3828
if (class == BPF_ALU64 && op == BPF_END && (insn->imm == 16 || insn->imm == 32))
kernel/bpf/verifier.c
3832
(class == BPF_ALU && op == BPF_END && insn->imm == 64))
kernel/futex/futex.h
444
u32 __user *uaddr2, int nr_wake, int nr_wake2, int op);
kernel/futex/futex.h
45
static inline unsigned int futex_to_flags(unsigned int op)
kernel/futex/futex.h
49
if (!(op & FUTEX_PRIVATE_FLAG))
kernel/futex/futex.h
52
if (op & FUTEX_CLOCK_REALTIME)
kernel/futex/syscalls.c
112
long do_futex(u32 __user *uaddr, int op, u32 val, ktime_t *timeout,
kernel/futex/syscalls.c
115
unsigned int flags = futex_to_flags(op);
kernel/futex/syscalls.c
116
int cmd = op & FUTEX_CMD_MASK;
kernel/futex/syscalls.c
175
futex_init_timeout(u32 cmd, u32 op, struct timespec64 *ts, ktime_t *t)
kernel/futex/syscalls.c
183
else if (cmd != FUTEX_LOCK_PI && !(op & FUTEX_CLOCK_REALTIME))
kernel/futex/syscalls.c
188
SYSCALL_DEFINE6(futex, u32 __user *, uaddr, int, op, u32, val,
kernel/futex/syscalls.c
192
int ret, cmd = op & FUTEX_CMD_MASK;
kernel/futex/syscalls.c
197
if (unlikely(should_fail_futex(!(op & FUTEX_PRIVATE_FLAG))))
kernel/futex/syscalls.c
201
ret = futex_init_timeout(cmd, op, &ts, &t);
kernel/futex/syscalls.c
207
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
kernel/futex/syscalls.c
506
SYSCALL_DEFINE6(futex_time32, u32 __user *, uaddr, int, op, u32, val,
kernel/futex/syscalls.c
510
int ret, cmd = op & FUTEX_CMD_MASK;
kernel/futex/syscalls.c
517
ret = futex_init_timeout(cmd, op, &ts, &t);
kernel/futex/syscalls.c
523
return do_futex(uaddr, op, val, tp, uaddr2, (unsigned long)utime, val3);
kernel/futex/waitwake.c
204
unsigned int op = (encoded_op & 0x70000000) >> 28;
kernel/futex/waitwake.c
224
ret = arch_futex_atomic_op_inuser(op, oparg, &oldval, uaddr);
kernel/futex/waitwake.c
252
int nr_wake, int nr_wake2, int op)
kernel/futex/waitwake.c
273
op_ret = futex_atomic_op_inuser(op, uaddr2);
kernel/kcsan/core.c
1193
#define DEFINE_TSAN_ATOMIC_RMW(op, bits, suffix) \
kernel/kcsan/core.c
1194
u##bits __tsan_atomic##bits##_##op(u##bits *ptr, u##bits v, int memorder); \
kernel/kcsan/core.c
1195
u##bits __tsan_atomic##bits##_##op(u##bits *ptr, u##bits v, int memorder) \
kernel/kcsan/core.c
1203
return __atomic_##op##suffix(ptr, v, memorder); \
kernel/kcsan/core.c
1205
EXPORT_SYMBOL(__tsan_atomic##bits##_##op)
kernel/kprobes.c
443
struct optimized_kprobe *op;
kernel/kprobes.c
445
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
446
arch_remove_optimized_kprobe(op);
kernel/kprobes.c
448
kfree(op);
kernel/kprobes.c
454
struct optimized_kprobe *op;
kernel/kprobes.c
457
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
458
return arch_prepared_optinsn(&op->optinsn);
kernel/kprobes.c
467
struct optimized_kprobe *op;
kernel/kprobes.c
473
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
475
return kprobe_disabled(p) && list_empty(&op->list);
kernel/kprobes.c
481
struct optimized_kprobe *op;
kernel/kprobes.c
484
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
485
if (!list_empty(&op->list))
kernel/kprobes.c
499
struct optimized_kprobe *op;
kernel/kprobes.c
506
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
507
if (arch_within_optimized_kprobe(op, addr))
kernel/kprobes.c
566
struct optimized_kprobe *op, *tmp;
kernel/kprobes.c
576
list_for_each_entry_safe(op, tmp, &freeing_list, list) {
kernel/kprobes.c
578
op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
kernel/kprobes.c
580
if (kprobe_disabled(&op->kp) && !kprobe_gone(&op->kp))
kernel/kprobes.c
581
arch_disarm_kprobe(&op->kp);
kernel/kprobes.c
582
if (kprobe_unused(&op->kp)) {
kernel/kprobes.c
588
hlist_del_rcu(&op->kp.hlist);
kernel/kprobes.c
590
list_del_init(&op->list);
kernel/kprobes.c
597
struct optimized_kprobe *op, *tmp;
kernel/kprobes.c
599
list_for_each_entry_safe(op, tmp, &freeing_list, list) {
kernel/kprobes.c
600
list_del_init(&op->list);
kernel/kprobes.c
601
if (WARN_ON_ONCE(!kprobe_unused(&op->kp))) {
kernel/kprobes.c
615
struct kprobe *_p = get_optimized_kprobe(op->kp.addr);
kernel/kprobes.c
619
free_aggr_kprobe(&op->kp);
kernel/kprobes.c
734
bool optprobe_queued_unopt(struct optimized_kprobe *op)
kernel/kprobes.c
739
if (op == _op)
kernel/kprobes.c
749
struct optimized_kprobe *op;
kernel/kprobes.c
760
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
763
if (arch_check_optimized_kprobe(op) < 0)
kernel/kprobes.c
767
if (op->kp.flags & KPROBE_FLAG_OPTIMIZED) {
kernel/kprobes.c
768
if (optprobe_queued_unopt(op)) {
kernel/kprobes.c
770
list_del_init(&op->list);
kernel/kprobes.c
774
op->kp.flags |= KPROBE_FLAG_OPTIMIZED;
kernel/kprobes.c
780
if (WARN_ON_ONCE(!list_empty(&op->list)))
kernel/kprobes.c
783
list_add(&op->list, &optimizing_list);
kernel/kprobes.c
788
static void force_unoptimize_kprobe(struct optimized_kprobe *op)
kernel/kprobes.c
791
arch_unoptimize_kprobe(op);
kernel/kprobes.c
792
op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
kernel/kprobes.c
798
struct optimized_kprobe *op;
kernel/kprobes.c
803
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
807
if (!list_empty(&op->list)) {
kernel/kprobes.c
808
if (optprobe_queued_unopt(op)) {
kernel/kprobes.c
815
force_unoptimize_kprobe(op);
kernel/kprobes.c
816
list_move(&op->list, &freeing_list);
kernel/kprobes.c
820
list_del_init(&op->list);
kernel/kprobes.c
821
op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
kernel/kprobes.c
829
force_unoptimize_kprobe(op);
kernel/kprobes.c
831
list_add(&op->list, &unoptimizing_list);
kernel/kprobes.c
839
struct optimized_kprobe *op;
kernel/kprobes.c
845
op = container_of(ap, struct optimized_kprobe, kp);
kernel/kprobes.c
846
WARN_ON_ONCE(list_empty(&op->list));
kernel/kprobes.c
860
struct optimized_kprobe *op;
kernel/kprobes.c
862
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
863
if (!list_empty(&op->list))
kernel/kprobes.c
865
list_del_init(&op->list);
kernel/kprobes.c
866
op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
kernel/kprobes.c
874
if (optprobe_queued_unopt(op))
kernel/kprobes.c
875
list_move(&op->list, &freeing_list);
kernel/kprobes.c
879
arch_remove_optimized_kprobe(op);
kernel/kprobes.c
883
void __prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *p)
kernel/kprobes.c
886
arch_prepare_optimized_kprobe(op, p);
kernel/kprobes.c
892
struct optimized_kprobe *op;
kernel/kprobes.c
894
op = container_of(p, struct optimized_kprobe, kp);
kernel/kprobes.c
895
__prepare_optimized_kprobe(op, p);
kernel/kprobes.c
901
struct optimized_kprobe *op;
kernel/kprobes.c
903
op = kzalloc_obj(struct optimized_kprobe);
kernel/kprobes.c
904
if (!op)
kernel/kprobes.c
907
INIT_LIST_HEAD(&op->list);
kernel/kprobes.c
908
op->kp.addr = p->addr;
kernel/kprobes.c
909
__prepare_optimized_kprobe(op, p);
kernel/kprobes.c
911
return &op->kp;
kernel/kprobes.c
923
struct optimized_kprobe *op;
kernel/kprobes.c
938
op = container_of(ap, struct optimized_kprobe, kp);
kernel/kprobes.c
939
if (!arch_prepared_optinsn(&op->optinsn)) {
kernel/kprobes.c
941
arch_remove_optimized_kprobe(op);
kernel/kprobes.c
942
kfree(op);
kernel/liveupdate/luo_core.c
399
const struct luo_ioctl_op *op;
kernel/liveupdate/luo_core.c
414
op = &luo_ioctl_ops[nr - LIVEUPDATE_CMD_BASE];
kernel/liveupdate/luo_core.c
415
if (op->ioctl_num != cmd)
kernel/liveupdate/luo_core.c
417
if (ucmd.user_size < op->min_size)
kernel/liveupdate/luo_core.c
421
err = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer,
kernel/liveupdate/luo_core.c
426
return op->execute(&ucmd);
kernel/liveupdate/luo_session.c
328
const struct luo_ioctl_op *op;
kernel/liveupdate/luo_session.c
345
op = &luo_session_ioctl_ops[nr - LIVEUPDATE_CMD_SESSION_BASE];
kernel/liveupdate/luo_session.c
346
if (op->ioctl_num != cmd)
kernel/liveupdate/luo_session.c
348
if (ucmd.user_size < op->min_size)
kernel/liveupdate/luo_session.c
352
ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer,
kernel/liveupdate/luo_session.c
357
return op->execute(session, &ucmd);
kernel/locking/spinlock.c
100
_raw_##op##_lock_irqsave(lock); \
kernel/locking/spinlock.c
103
static void __lockfunc __raw_##op##_lock_bh(locktype##_t *lock) \
kernel/locking/spinlock.c
112
flags = _raw_##op##_lock_irqsave(lock); \
kernel/locking/spinlock.c
67
#define BUILD_LOCK_OPS(op, locktype) \
kernel/locking/spinlock.c
68
static void __lockfunc __raw_##op##_lock(locktype##_t *lock) \
kernel/locking/spinlock.c
72
if (likely(do_raw_##op##_trylock(lock))) \
kernel/locking/spinlock.c
76
arch_##op##_relax(&lock->raw_lock); \
kernel/locking/spinlock.c
80
static unsigned long __lockfunc __raw_##op##_lock_irqsave(locktype##_t *lock) \
kernel/locking/spinlock.c
87
if (likely(do_raw_##op##_trylock(lock))) \
kernel/locking/spinlock.c
92
arch_##op##_relax(&lock->raw_lock); \
kernel/locking/spinlock.c
98
static void __lockfunc __raw_##op##_lock_irq(locktype##_t *lock) \
kernel/printk/index.c
151
static int pi_module_notify(struct notifier_block *nb, unsigned long op,
kernel/printk/index.c
156
switch (op) {
kernel/ptrace.c
1007
switch (info.op) {
kernel/ptrace.c
1119
if (ptrace_get_syscall_info_op(child) != info.op)
kernel/ptrace.c
1122
switch (info.op) {
kernel/ptrace.c
999
.op = ptrace_get_syscall_info_op(child),
kernel/sched/ext.c
220
#define SCX_HAS_OP(sch, op) test_bit(SCX_OP_IDX(op), (sch)->has_op)
kernel/sched/ext.c
311
#define SCX_CALL_OP(sch, mask, op, rq, args...) \
kernel/sched/ext.c
317
(sch)->ops.op(args); \
kernel/sched/ext.c
320
(sch)->ops.op(args); \
kernel/sched/ext.c
326
#define SCX_CALL_OP_RET(sch, mask, op, rq, args...) \
kernel/sched/ext.c
328
__typeof__((sch)->ops.op(args)) __ret; \
kernel/sched/ext.c
334
__ret = (sch)->ops.op(args); \
kernel/sched/ext.c
337
__ret = (sch)->ops.op(args); \
kernel/sched/ext.c
355
#define SCX_CALL_OP_TASK(sch, mask, op, rq, task, args...) \
kernel/sched/ext.c
359
SCX_CALL_OP((sch), mask, op, rq, task, ##args); \
kernel/sched/ext.c
363
#define SCX_CALL_OP_TASK_RET(sch, mask, op, rq, task, args...) \
kernel/sched/ext.c
365
__typeof__((sch)->ops.op(task, ##args)) __ret; \
kernel/sched/ext.c
368
__ret = SCX_CALL_OP_RET((sch), mask, op, rq, task, ##args); \
kernel/sched/ext.c
373
#define SCX_CALL_OP_2TASKS_RET(sch, mask, op, rq, task0, task1, args...) \
kernel/sched/ext.c
375
__typeof__((sch)->ops.op(task0, task1, ##args)) __ret; \
kernel/sched/ext.c
379
__ret = SCX_CALL_OP_RET((sch), mask, op, rq, task0, task1, ##args); \
kernel/sched/ext_internal.h
8
#define SCX_OP_IDX(op) (offsetof(struct sched_ext_ops, op) / sizeof(void (*)(void)))
kernel/seccomp.c
2101
static long do_seccomp(unsigned int op, unsigned int flags,
kernel/seccomp.c
2104
switch (op) {
kernel/seccomp.c
2126
SYSCALL_DEFINE3(seccomp, unsigned int, op, unsigned int, flags,
kernel/seccomp.c
2129
return do_seccomp(op, flags, uargs);
kernel/seccomp.c
2141
unsigned int op;
kernel/seccomp.c
2146
op = SECCOMP_SET_MODE_STRICT;
kernel/seccomp.c
2155
op = SECCOMP_SET_MODE_FILTER;
kernel/seccomp.c
2163
return do_seccomp(op, 0, uargs);
kernel/trace/blktrace.c
324
const enum req_op op = opf & REQ_OP_MASK;
kernel/trace/blktrace.c
330
what |= ddir_act[op_is_write(op) ? WRITE : READ];
kernel/trace/blktrace.c
337
switch (op) {
kernel/trace/bpf_trace.c
2231
static int bpf_event_notify(struct notifier_block *nb, unsigned long op,
kernel/trace/bpf_trace.c
2239
(op != MODULE_STATE_COMING && op != MODULE_STATE_GOING))
kernel/trace/bpf_trace.c
2244
switch (op) {
kernel/trace/fgraph.c
1218
struct ftrace_ops *op;
kernel/trace/fgraph.c
1223
list_for_each_entry(op, &graph_ops.subop_list, list) {
kernel/trace/fgraph.c
1224
if (op->flags & FTRACE_OPS_FL_PID) {
kernel/trace/fgraph.c
1225
gops = container_of(op, struct fgraph_ops, ops);
kernel/trace/fgraph.c
1226
gops->entryfunc = ftrace_pids_enabled(op) ?
kernel/trace/ftrace.c
1105
struct ftrace_ops *op = NULL;
kernel/trace/ftrace.c
1113
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
1119
if (op->trampoline && op->trampoline_size)
kernel/trace/ftrace.c
1120
if (addr >= op->trampoline &&
kernel/trace/ftrace.c
1121
addr < op->trampoline + op->trampoline_size) {
kernel/trace/ftrace.c
1123
return op;
kernel/trace/ftrace.c
1125
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
130
struct ftrace_ops *op, struct ftrace_regs *fregs);
kernel/trace/ftrace.c
142
struct ftrace_ops *op,
kernel/trace/ftrace.c
173
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/ftrace.c
175
struct trace_array *tr = op->private;
kernel/trace/ftrace.c
187
op->saved_func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
1958
struct ftrace_ops *op;
kernel/trace/ftrace.c
1969
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
1971
if (op == ops)
kernel/trace/ftrace.c
1973
if (op->func_hash == &global_ops.local_hash)
kernel/trace/ftrace.c
1974
__ftrace_hash_rec_update(op, inc);
kernel/trace/ftrace.c
1975
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
2456
struct ftrace_ops *op;
kernel/trace/ftrace.c
2459
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
2461
if (!op->trampoline)
kernel/trace/ftrace.c
2464
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2465
return op;
kernel/trace/ftrace.c
2466
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
2474
struct ftrace_ops *op;
kernel/trace/ftrace.c
2477
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
2479
if (op == op_exclude || !op->trampoline)
kernel/trace/ftrace.c
2482
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2483
return op;
kernel/trace/ftrace.c
2484
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
2491
struct ftrace_ops *op)
kernel/trace/ftrace.c
2495
while_for_each_ftrace_op(op) {
kernel/trace/ftrace.c
2497
if (!op->trampoline)
kernel/trace/ftrace.c
2500
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2501
return op;
kernel/trace/ftrace.c
2510
struct ftrace_ops *op;
kernel/trace/ftrace.c
2542
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
2544
if (!op->trampoline)
kernel/trace/ftrace.c
2551
if (op->flags & FTRACE_OPS_FL_ADDING)
kernel/trace/ftrace.c
2560
if ((op->flags & FTRACE_OPS_FL_MODIFYING) &&
kernel/trace/ftrace.c
2561
hash_contains_ip(ip, &op->old_hash))
kernel/trace/ftrace.c
2562
return op;
kernel/trace/ftrace.c
2568
if (!(op->flags & FTRACE_OPS_FL_MODIFYING) &&
kernel/trace/ftrace.c
2569
hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2570
return op;
kernel/trace/ftrace.c
2572
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
2580
struct ftrace_ops *op;
kernel/trace/ftrace.c
2583
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
2585
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2586
return op;
kernel/trace/ftrace.c
2587
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
2595
struct ftrace_ops *op, *found = NULL;
kernel/trace/ftrace.c
2598
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
2600
if (hash_contains_ip(ip, op->func_hash)) {
kernel/trace/ftrace.c
2603
found = op;
kernel/trace/ftrace.c
2606
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
395
struct ftrace_ops *op;
kernel/trace/ftrace.c
401
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
402
if (op->flags & FTRACE_OPS_FL_PID) {
kernel/trace/ftrace.c
403
op->func = ftrace_pids_enabled(op) ?
kernel/trace/ftrace.c
404
ftrace_pid_func : op->saved_func;
kernel/trace/ftrace.c
405
ftrace_update_trampoline(op);
kernel/trace/ftrace.c
407
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
5004
struct ftrace_ops *op;
kernel/trace/ftrace.c
5022
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
5023
if (op->func_hash == &global_ops.local_hash &&
kernel/trace/ftrace.c
5024
op->flags & FTRACE_OPS_FL_ENABLED) {
kernel/trace/ftrace.c
5025
ftrace_run_modify_code(op, FTRACE_UPDATE_CALLS, old_hash);
kernel/trace/ftrace.c
5029
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
5048
struct ftrace_ops *op;
kernel/trace/ftrace.c
5051
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
5054
list_for_each_entry(subops, &op->subop_list, list) {
kernel/trace/ftrace.c
5060
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
5245
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/ftrace.c
5250
probe = container_of(op, struct ftrace_func_probe, ops);
kernel/trace/ftrace.c
7730
struct ftrace_ops *op;
kernel/trace/ftrace.c
7732
list_for_each_entry_rcu(op, &ftrace_ops_trampoline_list, list) {
kernel/trace/ftrace.c
7733
if (!op->trampoline || symnum--)
kernel/trace/ftrace.c
7735
*value = op->trampoline;
kernel/trace/ftrace.c
8475
struct ftrace_ops *op;
kernel/trace/ftrace.c
8487
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
8489
if (op->flags & FTRACE_OPS_FL_STUB)
kernel/trace/ftrace.c
8498
if ((!(op->flags & FTRACE_OPS_FL_RCU) || rcu_is_watching()) &&
kernel/trace/ftrace.c
8499
ftrace_ops_test(op, ip, regs)) {
kernel/trace/ftrace.c
8500
if (FTRACE_WARN_ON(!op->func)) {
kernel/trace/ftrace.c
8501
pr_warn("op=%p %pS\n", op, op);
kernel/trace/ftrace.c
8504
op->func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
8506
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
8529
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/ftrace.c
8548
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/ftrace.c
8556
if (!(op->flags & FTRACE_OPS_FL_RCU) || rcu_is_watching())
kernel/trace/ftrace.c
8557
op->func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
9049
struct ftrace_ops *op;
kernel/trace/ftrace.c
9065
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
9066
if (!(op->flags & FTRACE_OPS_FL_DIRECT))
kernel/trace/ftrace.c
9068
if (ops_references_ip(op, ip)) {
kernel/trace/ftrace.c
9072
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
9076
if (!op->ops_func)
kernel/trace/ftrace.c
9079
ret = op->ops_func(op, ip, FTRACE_OPS_CMD_ENABLE_SHARE_IPMODIFY_PEER);
kernel/trace/ftrace.c
9098
struct ftrace_ops *op;
kernel/trace/ftrace.c
9114
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
9115
if (!(op->flags & FTRACE_OPS_FL_DIRECT))
kernel/trace/ftrace.c
9117
if (ops_references_ip(op, ip)) {
kernel/trace/ftrace.c
9121
} while_for_each_ftrace_op(op);
kernel/trace/ftrace.c
9125
if (found_op && op->ops_func)
kernel/trace/ftrace.c
9126
op->ops_func(op, ip, FTRACE_OPS_CMD_DISABLE_SHARE_IPMODIFY_PEER);
kernel/trace/ftrace.c
9333
struct ftrace_ops *op;
kernel/trace/ftrace.c
9335
do_for_each_ftrace_op(op, ftrace_ops_list) {
kernel/trace/ftrace.c
9336
if (op->flags & FTRACE_OPS_FL_PERMANENT)
kernel/trace/ftrace.c
9338
} while_for_each_ftrace_op(op);
kernel/trace/rv/monitors/sleep/sleep.c
125
int op, cmd;
kernel/trace/rv/monitors/sleep/sleep.c
150
op = args[1];
kernel/trace/rv/monitors/sleep/sleep.c
151
cmd = op & FUTEX_CMD_MASK;
kernel/trace/trace_eprobe.c
374
switch (code->op) {
kernel/trace/trace_eprobe.c
406
switch (code->op) {
kernel/trace/trace_events.c
5011
struct ftrace_ops *op, struct ftrace_regs *regs)
kernel/trace/trace_events_filter.c
1009
return pred->op == OP_EQ ? ret : !ret;
kernel/trace/trace_events_filter.c
1130
if (pred->op == OP_GLOB) {
kernel/trace/trace_events_filter.c
1506
static enum filter_pred_fn select_comparison_fn(enum filter_op_ids op,
kernel/trace/trace_events_filter.c
1512
switch (op) {
kernel/trace/trace_events_filter.c
1517
if (WARN_ON_ONCE(op < PRED_FUNC_START))
kernel/trace/trace_events_filter.c
1519
pred_func_index = op - PRED_FUNC_START;
kernel/trace/trace_events_filter.c
1647
int op;
kernel/trace/trace_events_filter.c
1693
for (op = 0; ops[op]; op++) {
kernel/trace/trace_events_filter.c
1695
if (strncmp(str + i, ops[op], strlen(ops[op])) == 0)
kernel/trace/trace_events_filter.c
1699
if (!ops[op]) {
kernel/trace/trace_events_filter.c
1704
i += strlen(ops[op]);
kernel/trace/trace_events_filter.c
1717
pred->op = op;
kernel/trace/trace_events_filter.c
1727
switch (op) {
kernel/trace/trace_events_filter.c
1844
switch (op) {
kernel/trace/trace_events_filter.c
1916
if (pred->op == OP_BAND)
kernel/trace/trace_events_filter.c
1917
pred->op = OP_EQ;
kernel/trace/trace_events_filter.c
1924
if (pred->op == OP_BAND)
kernel/trace/trace_events_filter.c
1925
pred->op = OP_EQ;
kernel/trace/trace_events_filter.c
1927
pred->fn_num = select_comparison_fn(pred->op, field->size, false);
kernel/trace/trace_events_filter.c
1928
if (pred->op == OP_NE)
kernel/trace/trace_events_filter.c
1952
switch (op) {
kernel/trace/trace_events_filter.c
2032
if (op == OP_GLOB) {
kernel/trace/trace_events_filter.c
2069
pred->fn_num = select_comparison_fn(pred->op, field->size,
kernel/trace/trace_events_filter.c
2071
if (pred->op == OP_NE)
kernel/trace/trace_events_filter.c
2632
if ((pred->op != OP_EQ) && (pred->op != OP_NE))
kernel/trace/trace_events_filter.c
2651
return __ftrace_function_set_filter(pred->op == OP_EQ,
kernel/trace/trace_events_filter.c
653
do_filter_cpumask(int op, const struct cpumask *mask, const struct cpumask *cmp)
kernel/trace/trace_events_filter.c
655
switch (op) {
kernel/trace/trace_events_filter.c
669
do_filter_scalar_cpumask(int op, unsigned int cpu, const struct cpumask *mask)
kernel/trace/trace_events_filter.c
675
switch (op) {
kernel/trace/trace_events_filter.c
688
do_filter_cpumask_scalar(int op, const struct cpumask *mask, unsigned int cpu)
kernel/trace/trace_events_filter.c
690
switch (op) {
kernel/trace/trace_events_filter.c
716
switch (pred->op) { \
kernel/trace/trace_events_filter.c
756
return do_filter_scalar_cpumask(pred->op, cpu, pred->mask); \
kernel/trace/trace_events_filter.c
89
int op;
kernel/trace/trace_events_filter.c
943
switch (pred->op) {
kernel/trace/trace_events_filter.c
966
return do_filter_scalar_cpumask(pred->op, cpu, pred->mask);
kernel/trace/trace_events_filter.c
977
return do_filter_cpumask(pred->op, mask, cmp);
kernel/trace/trace_events_filter.c
988
return do_filter_cpumask_scalar(pred->op, mask, cpu);
kernel/trace/trace_fprobe.c
287
switch (code->op) {
kernel/trace/trace_fprobe.c
375
switch (code->op) {
kernel/trace/trace_functions.c
209
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_functions.c
211
struct trace_array *tr = op->private;
kernel/trace/trace_functions.c
233
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_functions.c
235
struct trace_array *tr = op->private;
kernel/trace/trace_functions.c
26
struct ftrace_ops *op, struct ftrace_regs *fregs);
kernel/trace/trace_functions.c
273
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_functions.c
275
struct trace_array *tr = op->private;
kernel/trace/trace_functions.c
29
struct ftrace_ops *op, struct ftrace_regs *fregs);
kernel/trace/trace_functions.c
300
if (ftrace_pids_enabled(op))
kernel/trace/trace_functions.c
32
struct ftrace_ops *op, struct ftrace_regs *fregs);
kernel/trace/trace_functions.c
342
struct ftrace_ops *op,
kernel/trace/trace_functions.c
346
struct trace_array *tr = op->private;
kernel/trace/trace_functions.c
35
struct ftrace_ops *op, struct ftrace_regs *fregs);
kernel/trace/trace_functions.c
38
struct ftrace_ops *op,
kernel/trace/trace_functions.c
383
struct ftrace_ops *op,
kernel/trace/trace_functions.c
387
struct trace_array *tr = op->private;
kernel/trace/trace_irqsoff.c
141
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_kprobe.c
1430
switch (code->op) {
kernel/trace/trace_probe.c
1001
code->op = FETCH_OP_STACK;
kernel/trace/trace_probe.c
1009
code->op = FETCH_OP_COMM;
kernel/trace/trace_probe.c
1027
code->op = FETCH_OP_ARG;
kernel/trace/trace_probe.c
1041
code->op = FETCH_OP_EDATA;
kernel/trace/trace_probe.c
1107
code->op = FETCH_OP_REG;
kernel/trace/trace_probe.c
1122
code->op = FETCH_OP_IMM;
kernel/trace/trace_probe.c
1136
code->op = FETCH_OP_FOFFS;
kernel/trace/trace_probe.c
1145
code->op = FETCH_NOP_SYMBOL;
kernel/trace/trace_probe.c
1153
code->op = FETCH_OP_IMM;
kernel/trace/trace_probe.c
1162
code->op = FETCH_OP_DEREF;
kernel/trace/trace_probe.c
1202
if (code->op == FETCH_OP_COMM ||
kernel/trace/trace_probe.c
1203
code->op == FETCH_OP_DATA) {
kernel/trace/trace_probe.c
1213
code->op = deref;
kernel/trace/trace_probe.c
1224
code->op = FETCH_OP_DATA;
kernel/trace/trace_probe.c
1231
code->op = FETCH_OP_IMM;
kernel/trace/trace_probe.c
1245
if (!ret && code->op == FETCH_OP_NOP) {
kernel/trace/trace_probe.c
1276
if (code->op != FETCH_OP_NOP)
kernel/trace/trace_probe.c
1280
code->op = FETCH_OP_MOD_BF;
kernel/trace/trace_probe.c
1366
if (code->op != FETCH_OP_REG && code->op != FETCH_OP_STACK &&
kernel/trace/trace_probe.c
1367
code->op != FETCH_OP_RETVAL && code->op != FETCH_OP_ARG &&
kernel/trace/trace_probe.c
1368
code->op != FETCH_OP_DEREF && code->op != FETCH_OP_TP_ARG) {
kernel/trace/trace_probe.c
1374
if (code->op != FETCH_OP_DEREF && code->op != FETCH_OP_UDEREF &&
kernel/trace/trace_probe.c
1375
code->op != FETCH_OP_IMM && code->op != FETCH_OP_COMM &&
kernel/trace/trace_probe.c
1376
code->op != FETCH_OP_DATA && code->op != FETCH_OP_TP_ARG) {
kernel/trace/trace_probe.c
1384
(code->op == FETCH_OP_IMM || code->op == FETCH_OP_COMM ||
kernel/trace/trace_probe.c
1385
code->op == FETCH_OP_DATA) || code->op == FETCH_OP_TP_ARG ||
kernel/trace/trace_probe.c
1396
if (code->op != FETCH_OP_NOP) {
kernel/trace/trace_probe.c
1404
code->op == FETCH_OP_UDEREF)
kernel/trace/trace_probe.c
1405
code->op = FETCH_OP_ST_USTRING;
kernel/trace/trace_probe.c
1407
code->op = FETCH_OP_ST_SYMSTR;
kernel/trace/trace_probe.c
1409
code->op = FETCH_OP_ST_STRING;
kernel/trace/trace_probe.c
1412
} else if (code->op == FETCH_OP_DEREF) {
kernel/trace/trace_probe.c
1413
code->op = FETCH_OP_ST_MEM;
kernel/trace/trace_probe.c
1415
} else if (code->op == FETCH_OP_UDEREF) {
kernel/trace/trace_probe.c
1416
code->op = FETCH_OP_ST_UMEM;
kernel/trace/trace_probe.c
1420
if (code->op != FETCH_OP_NOP) {
kernel/trace/trace_probe.c
1424
code->op = FETCH_OP_ST_RAW;
kernel/trace/trace_probe.c
1449
if (scode->op != FETCH_OP_ST_MEM &&
kernel/trace/trace_probe.c
1450
scode->op != FETCH_OP_ST_STRING &&
kernel/trace/trace_probe.c
1451
scode->op != FETCH_OP_ST_USTRING) {
kernel/trace/trace_probe.c
1456
if (code->op != FETCH_OP_NOP) {
kernel/trace/trace_probe.c
1460
code->op = FETCH_OP_LP_ARRAY;
kernel/trace/trace_probe.c
1466
code->op = FETCH_OP_END;
kernel/trace/trace_probe.c
1504
code[FETCH_INSN_MAX - 1].op = FETCH_OP_END;
kernel/trace/trace_probe.c
1542
if (code->op == FETCH_OP_END)
kernel/trace/trace_probe.c
1554
if (code->op == FETCH_NOP_SYMBOL ||
kernel/trace/trace_probe.c
1555
code->op == FETCH_OP_DATA)
kernel/trace/trace_probe.c
1649
while (code && code->op != FETCH_OP_END) {
kernel/trace/trace_probe.c
1650
if (code->op == FETCH_NOP_SYMBOL ||
kernel/trace/trace_probe.c
1651
code->op == FETCH_OP_DATA)
kernel/trace/trace_probe.c
1868
while (code && code->op != FETCH_OP_END) {
kernel/trace/trace_probe.c
1869
if (code->op == FETCH_NOP_SYMBOL) {
kernel/trace/trace_probe.c
1870
if (code[1].op != FETCH_OP_IMM)
kernel/trace/trace_probe.c
327
code->op = FETCH_OP_TP_ARG;
kernel/trace/trace_probe.c
396
if (code->op == FETCH_OP_END) {
kernel/trace/trace_probe.c
401
code->op = FETCH_OP_UDEREF;
kernel/trace/trace_probe.c
403
code->op = FETCH_OP_DEREF;
kernel/trace/trace_probe.c
619
code->op = FETCH_OP_DEREF; /* TODO: user deref support */
kernel/trace/trace_probe.c
657
code->op = FETCH_OP_RETVAL;
kernel/trace/trace_probe.c
689
code->op = FETCH_OP_ARG;
kernel/trace/trace_probe.c
695
code->op = FETCH_OP_EDATA;
kernel/trace/trace_probe.c
748
if (code->op != FETCH_OP_NOP) {
kernel/trace/trace_probe.c
754
code->op = FETCH_OP_MOD_BF;
kernel/trace/trace_probe.c
803
code[0].op = FETCH_OP_ARG;
kernel/trace/trace_probe.c
805
code[1].op = FETCH_OP_ST_EDATA;
kernel/trace/trace_probe.c
823
for (i = 0; i < earg->size - 1 && earg->code[i].op != FETCH_OP_END; i++) {
kernel/trace/trace_probe.c
824
if (earg->code[i].op == FETCH_OP_ST_EDATA)
kernel/trace/trace_probe.c
852
earg->code[i].op = FETCH_OP_END;
kernel/trace/trace_probe.c
871
for (i = 0; i < earg->size - 1 && earg->code[i].op != FETCH_OP_END; i += 2) {
kernel/trace/trace_probe.c
872
if (WARN_ON_ONCE(earg->code[i].op != FETCH_OP_ARG))
kernel/trace/trace_probe.c
878
if (WARN_ON_ONCE(earg->code[i + 1].op != FETCH_OP_ST_EDATA))
kernel/trace/trace_probe.c
889
if (WARN_ON_ONCE(earg->code[i - 1].op != FETCH_OP_ST_EDATA))
kernel/trace/trace_probe.c
921
switch (code->op) {
kernel/trace/trace_probe.c
962
code->op = FETCH_OP_COMM;
kernel/trace/trace_probe.c
977
code->op = FETCH_OP_RETVAL;
kernel/trace/trace_probe.c
987
code->op = FETCH_OP_STACKP;
kernel/trace/trace_probe.h
119
enum fetch_op op;
kernel/trace/trace_probe_tmpl.h
105
switch (code->op) {
kernel/trace/trace_probe_tmpl.h
134
if (code->op == FETCH_OP_DEREF) {
kernel/trace/trace_probe_tmpl.h
138
} else if (code->op == FETCH_OP_UDEREF) {
kernel/trace/trace_probe_tmpl.h
153
switch (code->op) {
kernel/trace/trace_probe_tmpl.h
171
switch (code->op) {
kernel/trace/trace_probe_tmpl.h
199
if (code->op == FETCH_OP_MOD_BF) {
kernel/trace/trace_probe_tmpl.h
206
if (code->op == FETCH_OP_LP_ARRAY) {
kernel/trace/trace_probe_tmpl.h
212
if (s3->op != FETCH_OP_ST_STRING &&
kernel/trace/trace_probe_tmpl.h
213
s3->op != FETCH_OP_ST_USTRING) {
kernel/trace/trace_probe_tmpl.h
230
return code->op == FETCH_OP_END ? ret : -EILSEQ;
kernel/trace/trace_sched_wakeup.c
228
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_selftest.c
110
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
119
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
128
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
137
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
146
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
451
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
466
struct ftrace_ops *op,
kernel/trace/trace_selftest.c
585
struct ftrace_ops *op,
kernel/trace/trace_stack.c
293
struct ftrace_ops *op, struct ftrace_regs *fregs)
kernel/trace/trace_uprobe.c
227
switch (code->op) {
kernel/user_namespace.c
716
return seq->op->start(seq, pos);
lib/842/842_decompress.c
234
u8 op = decomp_ops[o][i];
lib/842/842_decompress.c
236
pr_debug("op is %x\n", op);
lib/842/842_decompress.c
238
switch (op & OP_ACTION) {
lib/842/842_decompress.c
240
ret = do_data(p, op & OP_AMOUNT);
lib/842/842_decompress.c
243
ret = do_index(p, op & OP_AMOUNT);
lib/842/842_decompress.c
248
pr_err("Internal error, invalid op %x\n", op);
lib/842/842_decompress.c
282
u64 op, rep, tmp, bytes, total;
lib/842/842_decompress.c
297
ret = next_bits(&p, &op, OP_BITS);
lib/842/842_decompress.c
301
pr_debug("template is %lx\n", (unsigned long)op);
lib/842/842_decompress.c
303
switch (op) {
lib/842/842_decompress.c
367
ret = do_op(&p, op);
lib/842/842_decompress.c
372
} while (op != OP_END);
lib/asn1_decoder.c
175
enum asn1_opcode op;
lib/asn1_decoder.c
205
op = machine[pc];
lib/asn1_decoder.c
206
if (unlikely(pc + asn1_op_lengths[op] > machlen))
lib/asn1_decoder.c
212
if (op <= ASN1_OP__MATCHES_TAG) {
lib/asn1_decoder.c
216
if ((op & ASN1_OP_MATCH__COND && flags & FLAG_MATCHED) ||
lib/asn1_decoder.c
217
(op & ASN1_OP_MATCH__SKIP && dp == datalen)) {
lib/asn1_decoder.c
219
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
233
if (op & ASN1_OP_MATCH__ANY) {
lib/asn1_decoder.c
250
if (op & ASN1_OP_MATCH__SKIP) {
lib/asn1_decoder.c
251
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
311
switch (op) {
lib/asn1_decoder.c
339
if (op & ASN1_OP_MATCH__ACT) {
lib/asn1_decoder.c
342
if (op & ASN1_OP_MATCH__ANY)
lib/asn1_decoder.c
353
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
362
jump_stack[jsp++] = pc + asn1_op_lengths[op];
lib/asn1_decoder.c
369
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
407
if (op & ASN1_OP_END__OF) {
lib/asn1_decoder.c
420
if (dp < len && (op & ASN1_OP_END__OF)) {
lib/asn1_decoder.c
433
if (op & ASN1_OP_END__ACT) {
lib/asn1_decoder.c
435
if (op & ASN1_OP_END__OF)
lib/asn1_decoder.c
443
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
448
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
457
pc += asn1_op_lengths[op];
lib/asn1_decoder.c
473
op, pc);
lib/atomic64.c
101
EXPORT_SYMBOL(generic_atomic64_##op##_return);
lib/atomic64.c
103
#define ATOMIC64_FETCH_OP(op, c_op) \
lib/atomic64.c
104
s64 generic_atomic64_fetch_##op(s64 a, atomic64_t *v) \
lib/atomic64.c
118
EXPORT_SYMBOL(generic_atomic64_fetch_##op);
lib/atomic64.c
120
#define ATOMIC64_OPS(op, c_op) \
lib/atomic64.c
121
ATOMIC64_OP(op, c_op) \
lib/atomic64.c
122
ATOMIC64_OP_RETURN(op, c_op) \
lib/atomic64.c
123
ATOMIC64_FETCH_OP(op, c_op)
lib/atomic64.c
129
#define ATOMIC64_OPS(op, c_op) \
lib/atomic64.c
130
ATOMIC64_OP(op, c_op) \
lib/atomic64.c
131
ATOMIC64_FETCH_OP(op, c_op)
lib/atomic64.c
73
#define ATOMIC64_OP(op, c_op) \
lib/atomic64.c
74
void generic_atomic64_##op(s64 a, atomic64_t *v) \
lib/atomic64.c
85
EXPORT_SYMBOL(generic_atomic64_##op);
lib/atomic64.c
87
#define ATOMIC64_OP_RETURN(op, c_op) \
lib/atomic64.c
88
s64 generic_atomic64_##op##_return(s64 a, atomic64_t *v) \
lib/atomic64_test.c
20
#define TEST(bit, op, c_op, val) \
lib/atomic64_test.c
24
atomic##bit##_##op(val, &v); \
lib/atomic64_test.c
36
#define FAMILY_TEST(test, bit, op, args...) \
lib/atomic64_test.c
38
test(bit, op, ##args); \
lib/atomic64_test.c
39
test(bit, op##_acquire, ##args); \
lib/atomic64_test.c
40
test(bit, op##_release, ##args); \
lib/atomic64_test.c
41
test(bit, op##_relaxed, ##args); \
lib/atomic64_test.c
44
#define TEST_RETURN(bit, op, c_op, val) \
lib/atomic64_test.c
49
BUG_ON(atomic##bit##_##op(val, &v) != r); \
lib/atomic64_test.c
53
#define TEST_FETCH(bit, op, c_op, val) \
lib/atomic64_test.c
58
BUG_ON(atomic##bit##_##op(val, &v) != v0); \
lib/atomic64_test.c
62
#define RETURN_FAMILY_TEST(bit, op, c_op, val) \
lib/atomic64_test.c
64
FAMILY_TEST(TEST_RETURN, bit, op, c_op, val); \
lib/atomic64_test.c
67
#define FETCH_FAMILY_TEST(bit, op, c_op, val) \
lib/atomic64_test.c
69
FAMILY_TEST(TEST_FETCH, bit, op, c_op, val); \
lib/atomic64_test.c
72
#define TEST_ARGS(bit, op, init, ret, expect, args...) \
lib/atomic64_test.c
75
BUG_ON(atomic##bit##_##op(&v, ##args) != ret); \
lib/bootconfig.c
696
static int __init xbc_parse_kv(char **k, char *v, int op)
lib/bootconfig.c
721
if (op == '=')
lib/bootconfig.c
723
if (op == ':') {
lib/crc/mips/crc32.h
25
#define _ASM_UNSET_CRC(op, SZ, TYPE) ".purgem " #op "\n\t"
lib/crc/mips/crc32.h
27
#define _ASM_SET_CRC(op, SZ, TYPE) ".set\tcrc\n\t"
lib/crc/mips/crc32.h
28
#define _ASM_UNSET_CRC(op, SZ, TYPE)
lib/crc/mips/crc32.h
31
#define __CRC32(crc, value, op, SZ, TYPE) \
lib/crc/mips/crc32.h
35
_ASM_SET_CRC(op, SZ, TYPE) \
lib/crc/mips/crc32.h
36
#op " %0, %1, %0\n\t" \
lib/crc/mips/crc32.h
37
_ASM_UNSET_CRC(op, SZ, TYPE) \
lib/crc/mips/crc32.h
52
#define _CRC32(crc, value, size, op) \
lib/crc/mips/crc32.h
53
_CRC32_##op##size(crc, value)
lib/dynamic_debug.c
491
int op, i;
lib/dynamic_debug.c
497
op = *str++;
lib/dynamic_debug.c
503
v3pr_info("op='%c'\n", op);
lib/dynamic_debug.c
520
switch (op) {
lib/kunit/attributes.c
100
else if (!strncmp(op, "<", 1))
lib/kunit/attributes.c
102
else if (!strncmp(op, "=", 1))
lib/kunit/attributes.c
105
pr_err("kunit executor: invalid filter operation: %s\n", op);
lib/kunit/attributes.c
341
char op;
lib/kunit/attributes.c
363
op = input[op_index];
lib/kunit/attributes.c
374
input[op_index] = op;
lib/kunit/attributes.c
90
static int int_filter(long val, const char *op, int input, int *err)
lib/kunit/attributes.c
92
if (!strncmp(op, "<=", 2))
lib/kunit/attributes.c
94
else if (!strncmp(op, ">=", 2))
lib/kunit/attributes.c
96
else if (!strncmp(op, "!=", 2))
lib/kunit/attributes.c
98
else if (!strncmp(op, ">", 1))
lib/logic_iomem.c
172
#define MAKE_FALLBACK(op, sz) \
lib/logic_iomem.c
173
static u##sz real_raw_read ## op(const volatile void __iomem *addr) \
lib/logic_iomem.c
175
WARN(1, "Invalid read" #op " at address %llx\n", \
lib/logic_iomem.c
180
static void real_raw_write ## op(u ## sz val, \
lib/logic_iomem.c
183
WARN(1, "Invalid writeq" #op " of 0x%llx at address %llx\n", \
lib/logic_iomem.c
218
#define MAKE_OP(op, sz) \
lib/logic_iomem.c
219
u##sz __raw_read ## op(const volatile void __iomem *addr) \
lib/logic_iomem.c
224
return real_raw_read ## op(addr); \
lib/logic_iomem.c
230
EXPORT_SYMBOL(__raw_read ## op); \
lib/logic_iomem.c
232
void __raw_write ## op(u ## sz val, volatile void __iomem *addr) \
lib/logic_iomem.c
237
real_raw_write ## op(val, addr); \
lib/logic_iomem.c
245
EXPORT_SYMBOL(__raw_write ## op)
lib/lz4/lz4_compress.c
199
BYTE *op = (BYTE *) dest;
lib/lz4/lz4_compress.c
200
BYTE * const olimit = op + maxOutputSize;
lib/lz4/lz4_compress.c
303
token = op++;
lib/lz4/lz4_compress.c
307
(unlikely(op + litLength +
lib/lz4/lz4_compress.c
318
*op++ = 255;
lib/lz4/lz4_compress.c
319
*op++ = (BYTE)len;
lib/lz4/lz4_compress.c
324
LZ4_wildCopy(op, anchor, op + litLength);
lib/lz4/lz4_compress.c
325
op += litLength;
lib/lz4/lz4_compress.c
330
LZ4_writeLE16(op, (U16)(ip - match));
lib/lz4/lz4_compress.c
331
op += 2;
lib/lz4/lz4_compress.c
368
(unlikely(op +
lib/lz4/lz4_compress.c
376
LZ4_write32(op, 0xFFFFFFFF);
lib/lz4/lz4_compress.c
379
op += 4;
lib/lz4/lz4_compress.c
380
LZ4_write32(op, 0xFFFFFFFF);
lib/lz4/lz4_compress.c
384
op += matchCode / 255;
lib/lz4/lz4_compress.c
385
*op++ = (BYTE)(matchCode % 255);
lib/lz4/lz4_compress.c
418
token = op++;
lib/lz4/lz4_compress.c
434
((op - (BYTE *)dest) + lastRun + 1 +
lib/lz4/lz4_compress.c
440
*op++ = RUN_MASK << ML_BITS;
lib/lz4/lz4_compress.c
442
*op++ = 255;
lib/lz4/lz4_compress.c
443
*op++ = (BYTE) accumulator;
lib/lz4/lz4_compress.c
445
*op++ = (BYTE)(lastRun << ML_BITS);
lib/lz4/lz4_compress.c
448
LZ4_memcpy(op, anchor, lastRun);
lib/lz4/lz4_compress.c
450
op += lastRun;
lib/lz4/lz4_compress.c
454
return (int) (((char *)op) - dest);
lib/lz4/lz4_compress.c
537
BYTE *op = (BYTE *) dst;
lib/lz4/lz4_compress.c
538
BYTE * const oend = op + targetDstSize;
lib/lz4/lz4_compress.c
539
BYTE * const oMaxLit = op + targetDstSize - 2 /* offset */
lib/lz4/lz4_compress.c
541
BYTE * const oMaxMatch = op + targetDstSize
lib/lz4/lz4_compress.c
613
token = op++;
lib/lz4/lz4_compress.c
614
if (op + ((litLength + 240) / 255)
lib/lz4/lz4_compress.c
617
op--;
lib/lz4/lz4_compress.c
624
*op++ = 255;
lib/lz4/lz4_compress.c
625
*op++ = (BYTE)len;
lib/lz4/lz4_compress.c
630
LZ4_wildCopy(op, anchor, op + litLength);
lib/lz4/lz4_compress.c
631
op += litLength;
lib/lz4/lz4_compress.c
636
LZ4_writeLE16(op, (U16)(ip - match)); op += 2;
lib/lz4/lz4_compress.c
643
if (op + ((matchLength + 240)/255) > oMaxMatch) {
lib/lz4/lz4_compress.c
645
matchLength = (15 - 1) + (oMaxMatch - op) * 255;
lib/lz4/lz4_compress.c
654
*op++ = 255;
lib/lz4/lz4_compress.c
656
*op++ = (BYTE)matchLength;
lib/lz4/lz4_compress.c
666
if (op > oMaxSeq)
lib/lz4/lz4_compress.c
678
token = op++; *token = 0;
lib/lz4/lz4_compress.c
691
if (op + 1 /* token */
lib/lz4/lz4_compress.c
695
lastRunSize = (oend - op) - 1;
lib/lz4/lz4_compress.c
703
*op++ = RUN_MASK << ML_BITS;
lib/lz4/lz4_compress.c
705
*op++ = 255;
lib/lz4/lz4_compress.c
706
*op++ = (BYTE) accumulator;
lib/lz4/lz4_compress.c
708
*op++ = (BYTE)(lastRunSize<<ML_BITS);
lib/lz4/lz4_compress.c
710
LZ4_memcpy(op, anchor, lastRunSize);
lib/lz4/lz4_compress.c
711
op += lastRunSize;
lib/lz4/lz4_compress.c
716
return (int) (((char *)op) - dst);
lib/lz4/lz4_decompress.c
106
assert(lowPrefix <= op);
lib/lz4/lz4_decompress.c
153
(op <= shortoend))) {
lib/lz4/lz4_decompress.c
155
LZ4_memcpy(op, ip, endOnInput ? 16 : 8);
lib/lz4/lz4_decompress.c
156
op += length; ip += length;
lib/lz4/lz4_decompress.c
166
match = op - offset;
lib/lz4/lz4_decompress.c
167
assert(match <= op); /* check overflow */
lib/lz4/lz4_decompress.c
174
LZ4_memcpy(op + 0, match + 0, 8);
lib/lz4/lz4_decompress.c
175
LZ4_memcpy(op + 8, match + 8, 8);
lib/lz4/lz4_decompress.c
176
LZ4_memcpy(op + 16, match + 16, 2);
lib/lz4/lz4_decompress.c
177
op += length + MINMATCH;
lib/lz4/lz4_decompress.c
206
&& unlikely((uptrval)(op) +
lib/lz4/lz4_decompress.c
207
length < (uptrval)(op))) {
lib/lz4/lz4_decompress.c
220
cpy = op + length;
lib/lz4/lz4_decompress.c
233
length = oend - op;
lib/lz4/lz4_decompress.c
269
LZ4_memmove(op, ip, length);
lib/lz4/lz4_decompress.c
271
op += length;
lib/lz4/lz4_decompress.c
282
LZ4_wildCopy(op, ip, cpy);
lib/lz4/lz4_decompress.c
284
op = cpy;
lib/lz4/lz4_decompress.c
290
match = op - offset;
lib/lz4/lz4_decompress.c
307
assert(oend > op);
lib/lz4/lz4_decompress.c
308
assert(oend - op >= 4);
lib/lz4/lz4_decompress.c
310
LZ4_write32(op, (U32)offset);
lib/lz4/lz4_decompress.c
327
(uptrval)(op) + length < (uptrval)op)) {
lib/lz4/lz4_decompress.c
337
if (unlikely(op + length > oend - LASTLITERALS)) {
lib/lz4/lz4_decompress.c
341
length = min(length, (size_t)(oend - op));
lib/lz4/lz4_decompress.c
349
memmove(op, dictEnd - (lowPrefix - match),
lib/lz4/lz4_decompress.c
351
op += length;
lib/lz4/lz4_decompress.c
360
LZ4_memcpy(op, dictEnd - copySize, copySize);
lib/lz4/lz4_decompress.c
361
op += copySize;
lib/lz4/lz4_decompress.c
362
if (restSize > (size_t)(op - lowPrefix)) {
lib/lz4/lz4_decompress.c
364
BYTE * const endOfMatch = op + restSize;
lib/lz4/lz4_decompress.c
367
while (op < endOfMatch)
lib/lz4/lz4_decompress.c
368
*op++ = *copyFrom++;
lib/lz4/lz4_decompress.c
370
LZ4_memcpy(op, lowPrefix, restSize);
lib/lz4/lz4_decompress.c
371
op += restSize;
lib/lz4/lz4_decompress.c
378
cpy = op + length;
lib/lz4/lz4_decompress.c
384
assert(op <= oend);
lib/lz4/lz4_decompress.c
387
size_t const mlen = min(length, (size_t)(oend - op));
lib/lz4/lz4_decompress.c
389
BYTE * const copyEnd = op + mlen;
lib/lz4/lz4_decompress.c
391
if (matchEnd > op) {
lib/lz4/lz4_decompress.c
393
while (op < copyEnd)
lib/lz4/lz4_decompress.c
394
*op++ = *match++;
lib/lz4/lz4_decompress.c
396
LZ4_memcpy(op, match, mlen);
lib/lz4/lz4_decompress.c
398
op = copyEnd;
lib/lz4/lz4_decompress.c
399
if (op == oend)
lib/lz4/lz4_decompress.c
405
op[0] = match[0];
lib/lz4/lz4_decompress.c
406
op[1] = match[1];
lib/lz4/lz4_decompress.c
407
op[2] = match[2];
lib/lz4/lz4_decompress.c
408
op[3] = match[3];
lib/lz4/lz4_decompress.c
410
LZ4_memcpy(op + 4, match, 4);
lib/lz4/lz4_decompress.c
413
LZ4_copy8(op, match);
lib/lz4/lz4_decompress.c
417
op += 8;
lib/lz4/lz4_decompress.c
430
if (op < oCopyLimit) {
lib/lz4/lz4_decompress.c
431
LZ4_wildCopy(op, match, oCopyLimit);
lib/lz4/lz4_decompress.c
432
match += oCopyLimit - op;
lib/lz4/lz4_decompress.c
433
op = oCopyLimit;
lib/lz4/lz4_decompress.c
435
while (op < cpy)
lib/lz4/lz4_decompress.c
436
*op++ = *match++;
lib/lz4/lz4_decompress.c
438
LZ4_copy8(op, match);
lib/lz4/lz4_decompress.c
440
LZ4_wildCopy(op + 8, match + 8, cpy);
lib/lz4/lz4_decompress.c
442
op = cpy; /* wildcopy correction */
lib/lz4/lz4_decompress.c
448
return (int) (((char *)op) - dst);
lib/lz4/lz4_decompress.c
85
BYTE *op = (BYTE *) dst;
lib/lz4/lz4_decompress.c
86
BYTE * const oend = op + outputSize;
lib/lz4/lz4hc_compress.c
263
BYTE **op,
lib/lz4/lz4hc_compress.c
275
token = (*op)++;
lib/lz4/lz4hc_compress.c
278
&& ((*op + (length>>8)
lib/lz4/lz4hc_compress.c
289
*(*op)++ = 255;
lib/lz4/lz4hc_compress.c
290
*(*op)++ = (BYTE)len;
lib/lz4/lz4hc_compress.c
295
LZ4_wildCopy(*op, *anchor, (*op) + length);
lib/lz4/lz4hc_compress.c
296
*op += length;
lib/lz4/lz4hc_compress.c
299
LZ4_writeLE16(*op, (U16)(*ip - match));
lib/lz4/lz4hc_compress.c
300
*op += 2;
lib/lz4/lz4hc_compress.c
306
&& (*op + (length>>8)
lib/lz4/lz4hc_compress.c
317
*(*op)++ = 255;
lib/lz4/lz4hc_compress.c
318
*(*op)++ = 255;
lib/lz4/lz4hc_compress.c
323
*(*op)++ = 255;
lib/lz4/lz4hc_compress.c
326
*(*op)++ = (BYTE)length;
lib/lz4/lz4hc_compress.c
353
BYTE *op = (BYTE *) dest;
lib/lz4/lz4hc_compress.c
354
BYTE * const oend = op + maxOutputSize;
lib/lz4/lz4hc_compress.c
401
if (LZ4HC_encodeSequence(&ip, &op,
lib/lz4/lz4hc_compress.c
467
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
471
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
497
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
538
if (LZ4HC_encodeSequence(&ip, &op, &anchor, ml,
lib/lz4/lz4hc_compress.c
558
&& (((char *)op - dest) + lastRun + 1
lib/lz4/lz4hc_compress.c
565
*op++ = (RUN_MASK<<ML_BITS);
lib/lz4/lz4hc_compress.c
568
*op++ = 255;
lib/lz4/lz4hc_compress.c
569
*op++ = (BYTE) lastRun;
lib/lz4/lz4hc_compress.c
571
*op++ = (BYTE)(lastRun<<ML_BITS);
lib/lz4/lz4hc_compress.c
572
LZ4_memcpy(op, anchor, iend - anchor);
lib/lz4/lz4hc_compress.c
573
op += iend - anchor;
lib/lz4/lz4hc_compress.c
577
return (int) (((char *)op) - dest);
lib/lzo/lzo1x_compress.c
130
op[*state_offset] |= t;
lib/lzo/lzo1x_compress.c
132
COPY4(op, ii);
lib/lzo/lzo1x_compress.c
133
op += t;
lib/lzo/lzo1x_compress.c
136
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
137
COPY8(op, ii);
lib/lzo/lzo1x_compress.c
138
COPY8(op + 8, ii + 8);
lib/lzo/lzo1x_compress.c
139
op += t;
lib/lzo/lzo1x_compress.c
143
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
147
*op++ = 0;
lib/lzo/lzo1x_compress.c
151
*op++ = 0;
lib/lzo/lzo1x_compress.c
154
*op++ = tt;
lib/lzo/lzo1x_compress.c
158
COPY8(op, ii);
lib/lzo/lzo1x_compress.c
159
COPY8(op + 8, ii + 8);
lib/lzo/lzo1x_compress.c
160
op += 16;
lib/lzo/lzo1x_compress.c
165
*op++ = *ii++;
lib/lzo/lzo1x_compress.c
175
| (run_length & 0x7), op);
lib/lzo/lzo1x_compress.c
176
op += 4;
lib/lzo/lzo1x_compress.c
267
*op++ = (((m_len - 1) << 5) | ((m_off & 7) << 2));
lib/lzo/lzo1x_compress.c
268
*op++ = (m_off >> 3);
lib/lzo/lzo1x_compress.c
273
*op++ = (M3_MARKER | (m_len - 2));
lib/lzo/lzo1x_compress.c
276
*op++ = M3_MARKER | 0;
lib/lzo/lzo1x_compress.c
280
*op++ = 0;
lib/lzo/lzo1x_compress.c
283
*op++ = (m_len);
lib/lzo/lzo1x_compress.c
286
*op++ = (m_off << 2);
lib/lzo/lzo1x_compress.c
287
*op++ = (m_off >> 6);
lib/lzo/lzo1x_compress.c
292
*op++ = (M4_MARKER | ((m_off >> 11) & 8)
lib/lzo/lzo1x_compress.c
309
*op++ = (M4_MARKER | ((m_off >> 11) & 8));
lib/lzo/lzo1x_compress.c
313
*op++ = 0;
lib/lzo/lzo1x_compress.c
316
*op++ = (m_len);
lib/lzo/lzo1x_compress.c
319
*op++ = (m_off << 2);
lib/lzo/lzo1x_compress.c
320
*op++ = (m_off >> 6);
lib/lzo/lzo1x_compress.c
327
*out = op;
lib/lzo/lzo1x_compress.c
342
unsigned char *op = out;
lib/lzo/lzo1x_compress.c
352
*op++ = 17;
lib/lzo/lzo1x_compress.c
353
*op++ = bitstream_version;
lib/lzo/lzo1x_compress.c
359
data_start = op;
lib/lzo/lzo1x_compress.c
371
ip, ll, &op, op_end, &t, wrkmem,
lib/lzo/lzo1x_compress.c
383
if (op == data_start && t <= 238) {
lib/lzo/lzo1x_compress.c
385
*op++ = (17 + t);
lib/lzo/lzo1x_compress.c
387
op[state_offset] |= t;
lib/lzo/lzo1x_compress.c
39
unsigned char *op;
lib/lzo/lzo1x_compress.c
390
*op++ = (t - 3);
lib/lzo/lzo1x_compress.c
394
*op++ = 0;
lib/lzo/lzo1x_compress.c
398
*op++ = 0;
lib/lzo/lzo1x_compress.c
401
*op++ = tt;
lib/lzo/lzo1x_compress.c
405
COPY8(op, ii);
lib/lzo/lzo1x_compress.c
406
COPY8(op + 8, ii + 8);
lib/lzo/lzo1x_compress.c
407
op += 16;
lib/lzo/lzo1x_compress.c
412
*op++ = *ii++;
lib/lzo/lzo1x_compress.c
417
*op++ = M4_MARKER | 1;
lib/lzo/lzo1x_compress.c
418
*op++ = 0;
lib/lzo/lzo1x_compress.c
419
*op++ = 0;
lib/lzo/lzo1x_compress.c
421
*out_len = op - out;
lib/lzo/lzo1x_compress.c
46
op = *out;
lib/lzo/lzo1x_compress_safe.c
16
#define HAVE_OP(x) ((size_t)(op_end - op) >= (size_t)(x))
lib/lzo/lzo1x_decompress_safe.c
100
COPY8(op, ip);
lib/lzo/lzo1x_decompress_safe.c
101
op += 8;
lib/lzo/lzo1x_decompress_safe.c
103
COPY8(op, ip);
lib/lzo/lzo1x_decompress_safe.c
104
op += 8;
lib/lzo/lzo1x_decompress_safe.c
108
op = oe;
lib/lzo/lzo1x_decompress_safe.c
115
*op++ = *ip++;
lib/lzo/lzo1x_decompress_safe.c
122
m_pos = op - 1;
lib/lzo/lzo1x_decompress_safe.c
127
op[0] = m_pos[0];
lib/lzo/lzo1x_decompress_safe.c
128
op[1] = m_pos[1];
lib/lzo/lzo1x_decompress_safe.c
129
op += 2;
lib/lzo/lzo1x_decompress_safe.c
133
m_pos = op - (1 + M2_MAX_OFFSET);
lib/lzo/lzo1x_decompress_safe.c
140
m_pos = op - 1;
lib/lzo/lzo1x_decompress_safe.c
162
m_pos = op - 1;
lib/lzo/lzo1x_decompress_safe.c
178
memset(op, 0, t);
lib/lzo/lzo1x_decompress_safe.c
179
op += t;
lib/lzo/lzo1x_decompress_safe.c
184
m_pos = op;
lib/lzo/lzo1x_decompress_safe.c
207
if (m_pos == op)
lib/lzo/lzo1x_decompress_safe.c
214
if (op - m_pos >= 8) {
lib/lzo/lzo1x_decompress_safe.c
215
unsigned char *oe = op + t;
lib/lzo/lzo1x_decompress_safe.c
218
COPY8(op, m_pos);
lib/lzo/lzo1x_decompress_safe.c
219
op += 8;
lib/lzo/lzo1x_decompress_safe.c
221
COPY8(op, m_pos);
lib/lzo/lzo1x_decompress_safe.c
222
op += 8;
lib/lzo/lzo1x_decompress_safe.c
224
} while (op < oe);
lib/lzo/lzo1x_decompress_safe.c
225
op = oe;
lib/lzo/lzo1x_decompress_safe.c
228
COPY4(op, ip);
lib/lzo/lzo1x_decompress_safe.c
229
op += next;
lib/lzo/lzo1x_decompress_safe.c
236
*op++ = *m_pos++;
lib/lzo/lzo1x_decompress_safe.c
237
} while (op < oe);
lib/lzo/lzo1x_decompress_safe.c
24
#define HAVE_OP(x) ((size_t)(op_end - op) >= (size_t)(x))
lib/lzo/lzo1x_decompress_safe.c
242
unsigned char *oe = op + t;
lib/lzo/lzo1x_decompress_safe.c
244
op[0] = m_pos[0];
lib/lzo/lzo1x_decompress_safe.c
245
op[1] = m_pos[1];
lib/lzo/lzo1x_decompress_safe.c
246
op += 2;
lib/lzo/lzo1x_decompress_safe.c
249
*op++ = *m_pos++;
lib/lzo/lzo1x_decompress_safe.c
250
} while (op < oe);
lib/lzo/lzo1x_decompress_safe.c
257
COPY4(op, ip);
lib/lzo/lzo1x_decompress_safe.c
258
op += t;
lib/lzo/lzo1x_decompress_safe.c
266
*op++ = *ip++;
lib/lzo/lzo1x_decompress_safe.c
273
*out_len = op - out;
lib/lzo/lzo1x_decompress_safe.c
279
*out_len = op - out;
lib/lzo/lzo1x_decompress_safe.c
283
*out_len = op - out;
lib/lzo/lzo1x_decompress_safe.c
287
*out_len = op - out;
lib/lzo/lzo1x_decompress_safe.c
42
unsigned char *op;
lib/lzo/lzo1x_decompress_safe.c
52
op = out;
lib/lzo/lzo1x_decompress_safe.c
98
unsigned char *oe = op + t;
lib/packing.c
343
enum packing_op op, u8 quirks)
lib/packing.c
345
if (op == PACK)
lib/smp_processor_id.c
62
noinstr void __this_cpu_preempt_check(const char *op)
lib/smp_processor_id.c
64
check_preemption_disabled("__this_cpu_", op);
lib/test_bpf.c
1003
int op = *(int *)arg;
lib/test_bpf.c
1010
if (__bpf_alu_result(&res, (u32)dst, (u32)src, op)) {
lib/test_bpf.c
1014
insns[i++] = BPF_ALU32_REG(op, R1, R2);
lib/test_bpf.c
1022
static int __bpf_fill_alu64_imm(struct bpf_test *self, int op)
lib/test_bpf.c
1024
return __bpf_fill_pattern(self, &op, 64, 32,
lib/test_bpf.c
1029
static int __bpf_fill_alu32_imm(struct bpf_test *self, int op)
lib/test_bpf.c
1031
return __bpf_fill_pattern(self, &op, 64, 32,
lib/test_bpf.c
1036
static int __bpf_fill_alu64_reg(struct bpf_test *self, int op)
lib/test_bpf.c
1038
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
1043
static int __bpf_fill_alu32_reg(struct bpf_test *self, int op)
lib/test_bpf.c
1045
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
11934
#define BPF_TEST_CLOBBER_ALU(alu, op, dst, src) \
lib/test_bpf.c
11936
#alu "_" #op " to " #dst ": no clobbering", \
lib/test_bpf.c
11948
BPF_##alu(BPF_ ##op, dst, src), \
lib/test_bpf.c
12060
#define BPF_TEST_CLOBBER_ATOMIC(width, op) \
lib/test_bpf.c
12062
"Atomic_" #width " " #op ": no clobbering", \
lib/test_bpf.c
12075
(op) == BPF_CMPXCHG ? 0 : \
lib/test_bpf.c
12076
(op) & BPF_FETCH ? 1 : 0), \
lib/test_bpf.c
12077
BPF_ATOMIC_OP(width, op, R10, R1, -8), \
lib/test_bpf.c
12120
#define BPF_ALU32_SRC_ZEXT(op) \
lib/test_bpf.c
12122
"ALU32_" #op "_X: src preserved in zext", \
lib/test_bpf.c
12127
BPF_ALU32_REG(BPF_##op, R2, R1), \
lib/test_bpf.c
12149
#define BPF_ATOMIC32_SRC_ZEXT(op) \
lib/test_bpf.c
12151
"ATOMIC_W_" #op ": src preserved in zext", \
lib/test_bpf.c
12156
BPF_ATOMIC_OP(BPF_W, BPF_##op, R10, R1, -4), \
lib/test_bpf.c
12195
#define BPF_JMP32_IMM_ZEXT(op) \
lib/test_bpf.c
12197
"JMP32_" #op "_K: operand preserved in zext", \
lib/test_bpf.c
12201
BPF_JMP32_IMM(BPF_##op, R0, 1234, 1), \
lib/test_bpf.c
12226
#define BPF_JMP32_REG_ZEXT(op) \
lib/test_bpf.c
12228
"JMP32_" #op "_X: operands preserved in zext", \
lib/test_bpf.c
12234
BPF_JMP32_IMM(BPF_##op, R0, R1, 1), \
lib/test_bpf.c
1238
static int __bpf_fill_alu_imm_regs(struct bpf_test *self, u8 op, bool alu32)
lib/test_bpf.c
1258
if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH)
lib/test_bpf.c
1261
__bpf_alu_result(&res, dst, imm, op);
lib/test_bpf.c
1271
insns[i++] = BPF_ALU32_IMM(op, rd, imm);
lib/test_bpf.c
1273
insns[i++] = BPF_ALU64_IMM(op, rd, imm);
lib/test_bpf.c
1421
static int __bpf_fill_alu_reg_pairs(struct bpf_test *self, u8 op, bool alu32)
lib/test_bpf.c
1442
if (op == BPF_LSH || op == BPF_RSH || op == BPF_ARSH)
lib/test_bpf.c
1445
__bpf_alu_result(&res, dst, src, op);
lib/test_bpf.c
1446
__bpf_alu_result(&same, src, src, op);
lib/test_bpf.c
1462
insns[i++] = BPF_ALU32_REG(op, rd, rs);
lib/test_bpf.c
1464
insns[i++] = BPF_ALU64_REG(op, rd, rs);
lib/test_bpf.c
1617
int op = *(int *)arg;
lib/test_bpf.c
1624
switch (op) {
lib/test_bpf.c
1629
__bpf_alu_result(&res, dst, src, BPF_OP(op));
lib/test_bpf.c
1633
if (op & BPF_FETCH)
lib/test_bpf.c
1646
insns[i++] = BPF_ATOMIC_OP(BPF_DW, op, R10, R2, -8);
lib/test_bpf.c
1664
int op = *(int *)arg;
lib/test_bpf.c
1671
switch (op) {
lib/test_bpf.c
1676
__bpf_alu_result(&res, (u32)dst, (u32)src, BPF_OP(op));
lib/test_bpf.c
1680
if (op & BPF_FETCH)
lib/test_bpf.c
1693
insns[i++] = BPF_ATOMIC_OP(BPF_W, op, R10, R2, -4);
lib/test_bpf.c
1791
static int __bpf_fill_atomic64(struct bpf_test *self, int op)
lib/test_bpf.c
1793
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
1798
static int __bpf_fill_atomic32(struct bpf_test *self, int op)
lib/test_bpf.c
1800
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
1913
static int __bpf_fill_atomic_reg_pairs(struct bpf_test *self, u8 width, u8 op)
lib/test_bpf.c
1934
switch (op) {
lib/test_bpf.c
1942
__bpf_alu_result(&res, mem, upd, BPF_OP(op));
lib/test_bpf.c
1960
insn[i++] = BPF_ATOMIC_OP(width, op, rd, rs, -8);
lib/test_bpf.c
1961
if (op == BPF_CMPXCHG && width == BPF_W)
lib/test_bpf.c
1965
if (op == BPF_CMPXCHG)
lib/test_bpf.c
1984
if (rs == R0 && op == BPF_CMPXCHG)
lib/test_bpf.c
1986
else if (rs == rd && (op == BPF_CMPXCHG ||
lib/test_bpf.c
1987
!(op & BPF_FETCH)))
lib/test_bpf.c
1989
else if (op == BPF_CMPXCHG)
lib/test_bpf.c
1991
else if (op & BPF_FETCH)
lib/test_bpf.c
2008
if (!(rd == R0 && op == BPF_CMPXCHG) &&
lib/test_bpf.c
2009
!(rd == rs && (op & BPF_FETCH))) {
lib/test_bpf.c
2018
} else if (op == BPF_XCHG) { /* Aliased, XCHG */
lib/test_bpf.c
2020
} else if (op == BPF_CMPXCHG) { /* Aliased, CMPXCHG */
lib/test_bpf.c
2024
insn[i++] = BPF_ALU64_REG(BPF_OP(op), R1, R10);
lib/test_bpf.c
2293
static bool __bpf_match_jmp_cond(s64 v1, s64 v2, u8 op)
lib/test_bpf.c
2295
switch (op) {
lib/test_bpf.c
2325
int op = *(int *)arg;
lib/test_bpf.c
2328
bool match = __bpf_match_jmp_cond(dst, (s32)imm, op);
lib/test_bpf.c
2334
insns[i++] = BPF_JMP_IMM(op, R1, imm, 1);
lib/test_bpf.c
2348
int op = *(int *)arg;
lib/test_bpf.c
2351
bool match = __bpf_match_jmp_cond((s32)dst, (s32)imm, op);
lib/test_bpf.c
2355
insns[i++] = BPF_JMP32_IMM(op, R1, imm, 1);
lib/test_bpf.c
2369
int op = *(int *)arg;
lib/test_bpf.c
2372
bool match = __bpf_match_jmp_cond(dst, src, op);
lib/test_bpf.c
2377
insns[i++] = BPF_JMP_REG(op, R1, R2, 1);
lib/test_bpf.c
2391
int op = *(int *)arg;
lib/test_bpf.c
2394
bool match = __bpf_match_jmp_cond((s32)dst, (s32)src, op);
lib/test_bpf.c
2399
insns[i++] = BPF_JMP32_REG(op, R1, R2, 1);
lib/test_bpf.c
2410
static int __bpf_fill_jmp_imm(struct bpf_test *self, int op)
lib/test_bpf.c
2412
return __bpf_fill_pattern(self, &op, 64, 32,
lib/test_bpf.c
2417
static int __bpf_fill_jmp32_imm(struct bpf_test *self, int op)
lib/test_bpf.c
2419
return __bpf_fill_pattern(self, &op, 64, 32,
lib/test_bpf.c
2424
static int __bpf_fill_jmp_reg(struct bpf_test *self, int op)
lib/test_bpf.c
2426
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
2431
static int __bpf_fill_jmp32_reg(struct bpf_test *self, int op)
lib/test_bpf.c
2433
return __bpf_fill_pattern(self, &op, 64, 64,
lib/test_bpf.c
502
int op = ops[(i >> 1) % ARRAY_SIZE(ops)];
lib/test_bpf.c
505
insns[i++] = BPF_ALU32_REG(op, R0, R1);
lib/test_bpf.c
507
insns[i++] = BPF_ALU64_REG(op, R0, R1);
lib/test_bpf.c
563
static bool __bpf_alu_result(u64 *res, u64 v1, u64 v2, u8 op)
lib/test_bpf.c
566
switch (op) {
lib/test_bpf.c
614
static int __bpf_fill_alu_shift(struct bpf_test *self, u8 op,
lib/test_bpf.c
648
insn[i++] = BPF_ALU32_IMM(op, R1, imm);
lib/test_bpf.c
650
insn[i++] = BPF_ALU32_REG(op, R1, R2);
lib/test_bpf.c
652
if (op == BPF_ARSH)
lib/test_bpf.c
656
__bpf_alu_result(&val, reg, imm, op);
lib/test_bpf.c
660
insn[i++] = BPF_ALU64_IMM(op, R1, imm);
lib/test_bpf.c
662
insn[i++] = BPF_ALU64_REG(op, R1, R2);
lib/test_bpf.c
663
__bpf_alu_result(&val, reg, imm, op);
lib/test_bpf.c
753
static int __bpf_fill_alu_shift_same_reg(struct bpf_test *self, u8 op,
lib/test_bpf.c
774
insn[i++] = BPF_ALU32_REG(op, R1, R1);
lib/test_bpf.c
776
insn[i++] = BPF_ALU64_REG(op, R1, R1);
lib/test_bpf.c
779
__bpf_alu_result(&res, val, val, op);
lib/test_bpf.c
8964
#define BPF_ATOMIC_OP_TEST1(width, op, logic, old, update, result) \
lib/test_bpf.c
8966
"BPF_ATOMIC | " #width ", " #op ": Test: " \
lib/test_bpf.c
8971
BPF_ATOMIC_OP(width, op, R10, R5, -40), \
lib/test_bpf.c
8983
#define BPF_ATOMIC_OP_TEST2(width, op, logic, old, update, result) \
lib/test_bpf.c
8985
"BPF_ATOMIC | " #width ", " #op ": Test side effects, r10: " \
lib/test_bpf.c
8991
BPF_ATOMIC_OP(width, op, R10, R0, -40), \
lib/test_bpf.c
9004
#define BPF_ATOMIC_OP_TEST3(width, op, logic, old, update, result) \
lib/test_bpf.c
9006
"BPF_ATOMIC | " #width ", " #op ": Test side effects, r0: " \
lib/test_bpf.c
9012
BPF_ATOMIC_OP(width, op, R10, R1, -40), \
lib/test_bpf.c
9024
#define BPF_ATOMIC_OP_TEST4(width, op, logic, old, update, result) \
lib/test_bpf.c
9026
"BPF_ATOMIC | " #width ", " #op ": Test fetch: " \
lib/test_bpf.c
9031
BPF_ATOMIC_OP(width, op, R10, R3, -40), \
lib/test_bpf.c
9037
{ { 0, (op) & BPF_FETCH ? old : update } }, \
lib/test_bpf.c
939
int op = *(int *)arg;
lib/test_bpf.c
946
if (__bpf_alu_result(&res, dst, (s32)imm, op)) {
lib/test_bpf.c
949
insns[i++] = BPF_ALU64_IMM(op, R1, imm);
lib/test_bpf.c
960
int op = *(int *)arg;
lib/test_bpf.c
967
if (__bpf_alu_result(&res, (u32)dst, (u32)imm, op)) {
lib/test_bpf.c
970
insns[i++] = BPF_ALU32_IMM(op, R1, imm);
lib/test_bpf.c
981
int op = *(int *)arg;
lib/test_bpf.c
988
if (__bpf_alu_result(&res, dst, src, op)) {
lib/test_bpf.c
992
insns[i++] = BPF_ALU64_REG(op, R1, R2);
lib/test_context-analysis.c
30
#define TEST_SPINLOCK_COMMON(class, type, type_init, type_lock, type_unlock, type_trylock, op) \
lib/test_context-analysis.c
46
op(d->counter); \
lib/test_context-analysis.c
47
op(*d->pointer); \
lib/test_context-analysis.c
50
op(d->counter); \
lib/test_context-analysis.c
51
op(*d->pointer); \
lib/test_context-analysis.c
54
op(d->counter); \
lib/test_context-analysis.c
55
op(*d->pointer); \
lib/test_context-analysis.c
58
op(d->counter); \
lib/test_context-analysis.c
59
op(*d->pointer); \
lib/test_context-analysis.c
65
op(d->counter); \
lib/test_context-analysis.c
72
op(d->counter); \
lib/test_context-analysis.c
76
{ guard(class)(&d->lock); op(d->counter); } \
lib/test_context-analysis.c
77
{ guard(class##_irq)(&d->lock); op(d->counter); } \
lib/test_context-analysis.c
78
{ guard(class##_irqsave)(&d->lock); op(d->counter); } \
lib/tests/cpumask_kunit.c
26
#define EXPECT_FOR_EACH_CPU_OP_EQ(test, op, mask1, mask2) \
lib/tests/cpumask_kunit.c
32
cpumask_##op(&mask_tmp, m1, m2); \
lib/tests/cpumask_kunit.c
34
for_each_cpu_##op(cpu, mask1, mask2) \
lib/tests/overflow_kunit.c
253
#define check_one_op(t, fmt, op, sym, a, b, r, of) do { \
lib/tests/overflow_kunit.c
259
_of = check_ ## op ## _overflow(a, b, &_r); \
lib/tests/overflow_kunit.c
267
_of = check_ ## op ## _overflow(_a_orig++, _b_orig++, &_r); \
lib/tests/overflow_kunit.c
269
"Unexpected check " #op " macro side-effect!\n"); \
lib/tests/overflow_kunit.c
271
"Unexpected check " #op " macro side-effect!\n"); \
lib/tests/overflow_kunit.c
273
_r = wrapping_ ## op(t, a, b); \
lib/tests/overflow_kunit.c
280
_r = wrapping_ ## op(t, _a_orig++, _b_orig++); \
lib/tests/overflow_kunit.c
282
"Unexpected wrap " #op " macro side-effect!\n"); \
lib/tests/overflow_kunit.c
284
"Unexpected wrap " #op " macro side-effect!\n"); \
lib/tests/overflow_kunit.c
300
#define check_self_op(fmt, op, sym, a, b) do { \
lib/tests/overflow_kunit.c
310
_r = wrapping_ ## op(_a, _b); \
lib/tests/overflow_kunit.c
312
"expected "fmt" "#op" "fmt" == "fmt", got "fmt"\n", \
lib/tests/overflow_kunit.c
315
"expected "fmt" "#op" "fmt" == "fmt", got "fmt"\n", \
lib/tests/overflow_kunit.c
319
wrapping_ ## op(_a_orig[get_index()], _b_orig++); \
lib/tests/overflow_kunit.c
321
"Unexpected wrapping_" #op " macro side-effect on arg1!\n"); \
lib/tests/overflow_kunit.c
323
"Unexpected wrapping_" #op " macro side-effect on arg2!\n"); \
lib/ubsan.c
242
void *rhs, char op)
lib/ubsan.c
260
op,
lib/zlib_inflate/inffast.c
127
op = (unsigned)(this.bits);
lib/zlib_inflate/inffast.c
128
hold >>= op;
lib/zlib_inflate/inffast.c
129
bits -= op;
lib/zlib_inflate/inffast.c
130
op = (unsigned)(this.op);
lib/zlib_inflate/inffast.c
131
if (op == 0) { /* literal */
lib/zlib_inflate/inffast.c
134
else if (op & 16) { /* length base */
lib/zlib_inflate/inffast.c
136
op &= 15; /* number of extra bits */
lib/zlib_inflate/inffast.c
137
if (op) {
lib/zlib_inflate/inffast.c
138
if (bits < op) {
lib/zlib_inflate/inffast.c
142
len += (unsigned)hold & ((1U << op) - 1);
lib/zlib_inflate/inffast.c
143
hold >>= op;
lib/zlib_inflate/inffast.c
144
bits -= op;
lib/zlib_inflate/inffast.c
154
op = (unsigned)(this.bits);
lib/zlib_inflate/inffast.c
155
hold >>= op;
lib/zlib_inflate/inffast.c
156
bits -= op;
lib/zlib_inflate/inffast.c
157
op = (unsigned)(this.op);
lib/zlib_inflate/inffast.c
158
if (op & 16) { /* distance base */
lib/zlib_inflate/inffast.c
160
op &= 15; /* number of extra bits */
lib/zlib_inflate/inffast.c
161
if (bits < op) {
lib/zlib_inflate/inffast.c
164
if (bits < op) {
lib/zlib_inflate/inffast.c
169
dist += (unsigned)hold & ((1U << op) - 1);
lib/zlib_inflate/inffast.c
177
hold >>= op;
lib/zlib_inflate/inffast.c
178
bits -= op;
lib/zlib_inflate/inffast.c
179
op = (unsigned)(out - beg); /* max distance in output */
lib/zlib_inflate/inffast.c
180
if (dist > op) { /* see if copy from window */
lib/zlib_inflate/inffast.c
181
op = dist - op; /* distance back in window */
lib/zlib_inflate/inffast.c
182
if (op > whave) {
lib/zlib_inflate/inffast.c
189
from += wsize - op;
lib/zlib_inflate/inffast.c
190
if (op < len) { /* some from window */
lib/zlib_inflate/inffast.c
191
len -= op;
lib/zlib_inflate/inffast.c
194
} while (--op);
lib/zlib_inflate/inffast.c
198
else if (write < op) { /* wrap around window */
lib/zlib_inflate/inffast.c
199
from += wsize + write - op;
lib/zlib_inflate/inffast.c
200
op -= write;
lib/zlib_inflate/inffast.c
201
if (op < len) { /* some from end of window */
lib/zlib_inflate/inffast.c
202
len -= op;
lib/zlib_inflate/inffast.c
205
} while (--op);
lib/zlib_inflate/inffast.c
208
op = write;
lib/zlib_inflate/inffast.c
209
len -= op;
lib/zlib_inflate/inffast.c
212
} while (--op);
lib/zlib_inflate/inffast.c
218
from += write - op;
lib/zlib_inflate/inffast.c
219
if (op < len) { /* some from window */
lib/zlib_inflate/inffast.c
220
len -= op;
lib/zlib_inflate/inffast.c
223
} while (--op);
lib/zlib_inflate/inffast.c
285
else if ((op & 64) == 0) { /* 2nd level distance code */
lib/zlib_inflate/inffast.c
286
this = dcode[this.val + (hold & ((1U << op) - 1))];
lib/zlib_inflate/inffast.c
295
else if ((op & 64) == 0) { /* 2nd level length code */
lib/zlib_inflate/inffast.c
296
this = lcode[this.val + (hold & ((1U << op) - 1))];
lib/zlib_inflate/inffast.c
299
else if (op & 32) { /* end-of-block */
lib/zlib_inflate/inffast.c
89
unsigned op; /* code bits, operation, extra bits, or */
lib/zlib_inflate/inflate.c
591
if (this.op && (this.op & 0xf0) == 0) {
lib/zlib_inflate/inflate.c
595
(BITS(last.bits + last.op) >> last.bits)];
lib/zlib_inflate/inflate.c
603
if ((int)(this.op) == 0) {
lib/zlib_inflate/inflate.c
607
if (this.op & 32) {
lib/zlib_inflate/inflate.c
611
if (this.op & 64) {
lib/zlib_inflate/inflate.c
616
state->extra = (unsigned)(this.op) & 15;
lib/zlib_inflate/inflate.c
633
if ((this.op & 0xf0) == 0) {
lib/zlib_inflate/inflate.c
637
(BITS(last.bits + last.op) >> last.bits)];
lib/zlib_inflate/inflate.c
644
if (this.op & 64) {
lib/zlib_inflate/inflate.c
650
state->extra = (unsigned)(this.op) & 15;
lib/zlib_inflate/inftrees.c
104
this.op = (unsigned char)64; /* invalid code marker */
lib/zlib_inflate/inftrees.c
206
this.op = (unsigned char)0;
lib/zlib_inflate/inftrees.c
210
this.op = (unsigned char)(extra[work[sym]]);
lib/zlib_inflate/inftrees.c
214
this.op = (unsigned char)(32 + 64); /* end of block */
lib/zlib_inflate/inftrees.c
271
(*table)[low].op = (unsigned char)curr;
lib/zlib_inflate/inftrees.c
284
this.op = (unsigned char)64; /* invalid code marker */
lib/zlib_inflate/inftrees.h
28
unsigned char op; /* operation, extra bits, table bits */
lib/zstd/common/fse_decompress.c
180
BYTE* op = ostart;
lib/zstd/common/fse_decompress.c
181
BYTE* const omax = op + maxDstSize;
lib/zstd/common/fse_decompress.c
199
for ( ; (BIT_reloadDStream(&bitD)==BIT_DStream_unfinished) & (op<olimit) ; op+=4) {
lib/zstd/common/fse_decompress.c
200
op[0] = FSE_GETSYMBOL(&state1);
lib/zstd/common/fse_decompress.c
205
op[1] = FSE_GETSYMBOL(&state2);
lib/zstd/common/fse_decompress.c
208
{ if (BIT_reloadDStream(&bitD) > BIT_DStream_unfinished) { op+=2; break; } }
lib/zstd/common/fse_decompress.c
210
op[2] = FSE_GETSYMBOL(&state1);
lib/zstd/common/fse_decompress.c
215
op[3] = FSE_GETSYMBOL(&state2);
lib/zstd/common/fse_decompress.c
221
if (op>(omax-2)) return ERROR(dstSize_tooSmall);
lib/zstd/common/fse_decompress.c
222
*op++ = FSE_GETSYMBOL(&state1);
lib/zstd/common/fse_decompress.c
224
*op++ = FSE_GETSYMBOL(&state2);
lib/zstd/common/fse_decompress.c
228
if (op>(omax-2)) return ERROR(dstSize_tooSmall);
lib/zstd/common/fse_decompress.c
229
*op++ = FSE_GETSYMBOL(&state2);
lib/zstd/common/fse_decompress.c
231
*op++ = FSE_GETSYMBOL(&state1);
lib/zstd/common/fse_decompress.c
235
assert(op >= ostart);
lib/zstd/common/fse_decompress.c
236
return (size_t)(op-ostart);
lib/zstd/common/zstd_internal.h
214
BYTE* op = (BYTE*)dst;
lib/zstd/common/zstd_internal.h
215
BYTE* const oend = op + length;
lib/zstd/common/zstd_internal.h
220
COPY8(op, ip);
lib/zstd/common/zstd_internal.h
221
} while (op < oend);
lib/zstd/common/zstd_internal.h
230
ZSTD_copy16(op, ip);
lib/zstd/common/zstd_internal.h
232
op += 16;
lib/zstd/common/zstd_internal.h
235
COPY16(op, ip);
lib/zstd/common/zstd_internal.h
236
COPY16(op, ip);
lib/zstd/common/zstd_internal.h
238
while (op < oend);
lib/zstd/compress/huf_compress.c
1067
{ BYTE* op = ostart;
lib/zstd/compress/huf_compress.c
1068
size_t const initErr = HUF_initCStream(&bitC, op, (size_t)(oend-op));
lib/zstd/compress/huf_compress.c
1175
BYTE* op = ostart;
lib/zstd/compress/huf_compress.c
1179
op += 6; /* jumpTable */
lib/zstd/compress/huf_compress.c
1181
assert(op <= oend);
lib/zstd/compress/huf_compress.c
1182
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1185
op += cSize;
lib/zstd/compress/huf_compress.c
1189
assert(op <= oend);
lib/zstd/compress/huf_compress.c
1190
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1193
op += cSize;
lib/zstd/compress/huf_compress.c
1197
assert(op <= oend);
lib/zstd/compress/huf_compress.c
1198
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1201
op += cSize;
lib/zstd/compress/huf_compress.c
1205
assert(op <= oend);
lib/zstd/compress/huf_compress.c
1207
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, (size_t)(iend-ip), CTable, flags) );
lib/zstd/compress/huf_compress.c
1209
op += cSize;
lib/zstd/compress/huf_compress.c
1212
return (size_t)(op-ostart);
lib/zstd/compress/huf_compress.c
1223
BYTE* const ostart, BYTE* op, BYTE* const oend,
lib/zstd/compress/huf_compress.c
1228
HUF_compress1X_usingCTable_internal(op, (size_t)(oend - op), src, srcSize, CTable, flags) :
lib/zstd/compress/huf_compress.c
1229
HUF_compress4X_usingCTable_internal(op, (size_t)(oend - op), src, srcSize, CTable, flags);
lib/zstd/compress/huf_compress.c
1232
op += cSize;
lib/zstd/compress/huf_compress.c
1234
assert(op >= ostart);
lib/zstd/compress/huf_compress.c
1235
if ((size_t)(op-ostart) >= srcSize-1) { return 0; }
lib/zstd/compress/huf_compress.c
1236
return (size_t)(op-ostart);
lib/zstd/compress/huf_compress.c
1341
BYTE* op = ostart;
lib/zstd/compress/huf_compress.c
1358
return HUF_compressCTable_internal(ostart, op, oend,
lib/zstd/compress/huf_compress.c
1394
return HUF_compressCTable_internal(ostart, op, oend,
lib/zstd/compress/huf_compress.c
1410
{ CHECK_V_F(hSize, HUF_writeCTable_wksp(op, dstSize, table->CTable, maxSymbolValue, huffLog,
lib/zstd/compress/huf_compress.c
1417
return HUF_compressCTable_internal(ostart, op, oend,
lib/zstd/compress/huf_compress.c
1424
op += hSize;
lib/zstd/compress/huf_compress.c
1429
return HUF_compressCTable_internal(ostart, op, oend,
lib/zstd/compress/huf_compress.c
150
BYTE* op = ostart;
lib/zstd/compress/huf_compress.c
172
{ CHECK_V_F(hSize, FSE_writeNCount(op, (size_t)(oend-op), wksp->norm, maxSymbolValue, tableLog) );
lib/zstd/compress/huf_compress.c
173
op += hSize;
lib/zstd/compress/huf_compress.c
178
{ CHECK_V_F(cSize, FSE_compress_usingCTable(op, (size_t)(oend - op), weightTable, wtSize, wksp->CTable) );
lib/zstd/compress/huf_compress.c
180
op += cSize;
lib/zstd/compress/huf_compress.c
183
return (size_t)(op-ostart);
lib/zstd/compress/huf_compress.c
251
BYTE* op = (BYTE*)dst;
lib/zstd/compress/huf_compress.c
273
{ CHECK_V_F(hSize, HUF_compressWeights(op+1, maxDstSize-1, wksp->huffWeight, maxSymbolValue, &wksp->wksp, sizeof(wksp->wksp)) );
lib/zstd/compress/huf_compress.c
275
op[0] = (BYTE)hSize;
lib/zstd/compress/huf_compress.c
282
op[0] = (BYTE)(128 /*special case*/ + (maxSymbolValue-1));
lib/zstd/compress/huf_compress.c
285
op[(n/2)+1] = (BYTE)((wksp->huffWeight[n] << 4) + wksp->huffWeight[n+1]);
lib/zstd/compress/zstd_compress.c
2686
BYTE* op = ostart;
lib/zstd/compress/zstd_compress.c
2698
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2713
op, (size_t)(oend - op),
lib/zstd/compress/zstd_compress.c
2727
op += countSize;
lib/zstd/compress/zstd_compress.c
2728
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2745
op, (size_t)(oend - op),
lib/zstd/compress/zstd_compress.c
2759
op += countSize;
lib/zstd/compress/zstd_compress.c
2760
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2766
DEBUGLOG(5, "Building ML table (remaining space : %i)", (int)(oend-op));
lib/zstd/compress/zstd_compress.c
2775
op, (size_t)(oend - op),
lib/zstd/compress/zstd_compress.c
2789
op += countSize;
lib/zstd/compress/zstd_compress.c
2790
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2792
stats.size = (size_t)(op-ostart);
lib/zstd/compress/zstd_compress.c
2824
BYTE* op = ostart;
lib/zstd/compress/zstd_compress.c
2841
op, dstCapacity,
lib/zstd/compress/zstd_compress.c
2850
op += cSize;
lib/zstd/compress/zstd_compress.c
2854
RETURN_ERROR_IF((oend-op) < 3 /*max nbSeq Size*/ + 1 /*seqHead*/,
lib/zstd/compress/zstd_compress.c
2857
*op++ = (BYTE)nbSeq;
lib/zstd/compress/zstd_compress.c
2859
op[0] = (BYTE)((nbSeq>>8) + 0x80);
lib/zstd/compress/zstd_compress.c
2860
op[1] = (BYTE)nbSeq;
lib/zstd/compress/zstd_compress.c
2861
op+=2;
lib/zstd/compress/zstd_compress.c
2863
op[0]=0xFF;
lib/zstd/compress/zstd_compress.c
2864
MEM_writeLE16(op+1, (U16)(nbSeq - LONGNBSEQ));
lib/zstd/compress/zstd_compress.c
2865
op+=3;
lib/zstd/compress/zstd_compress.c
2867
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2871
return (size_t)(op - ostart);
lib/zstd/compress/zstd_compress.c
2873
{ BYTE* const seqHead = op++;
lib/zstd/compress/zstd_compress.c
2878
op, oend,
lib/zstd/compress/zstd_compress.c
2884
op += stats.size;
lib/zstd/compress/zstd_compress.c
2889
op, (size_t)(oend - op),
lib/zstd/compress/zstd_compress.c
2896
op += bitstreamSize;
lib/zstd/compress/zstd_compress.c
2897
assert(op <= oend);
lib/zstd/compress/zstd_compress.c
2915
DEBUGLOG(5, "compressed block size : %u", (unsigned)(op - ostart));
lib/zstd/compress/zstd_compress.c
2916
return (size_t)(op - ostart);
lib/zstd/compress/zstd_compress.c
3530
writeBlockHeader(void* op, size_t cSize, size_t blockSize, U32 lastBlock)
lib/zstd/compress/zstd_compress.c
3535
MEM_writeLE24(op, cBlockHeader);
lib/zstd/compress/zstd_compress.c
3685
BYTE* op = ostart;
lib/zstd/compress/zstd_compress.c
3693
prevEntropy, nextEntropy, op, oend,
lib/zstd/compress/zstd_compress.c
4027
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
4042
op + ZSTD_blockHeaderSize, dstCapacity - ZSTD_blockHeaderSize,
lib/zstd/compress/zstd_compress.c
4066
cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, srcSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4071
cSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, srcSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4077
writeBlockHeader(op, cSeqsSize, srcSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4182
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
4218
op, dstCapacity,
lib/zstd/compress/zstd_compress.c
4246
op, dstCapacity,
lib/zstd/compress/zstd_compress.c
4254
op += cSizeChunk;
lib/zstd/compress/zstd_compress.c
4308
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
4347
op[0] = ip[0];
lib/zstd/compress/zstd_compress.c
4514
BYTE* op = ostart;
lib/zstd/compress/zstd_compress.c
4551
cSize = ZSTD_compressBlock_targetCBlockSize(cctx, op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4556
cSize = ZSTD_compressBlock_splitBlock(cctx, op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4561
op+ZSTD_blockHeaderSize, dstCapacity-ZSTD_blockHeaderSize,
lib/zstd/compress/zstd_compress.c
4566
cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4572
MEM_writeLE24(op, cBlockHeader);
lib/zstd/compress/zstd_compress.c
4596
op += cSize;
lib/zstd/compress/zstd_compress.c
4604
if (lastFrameChunk && (op>ostart)) cctx->stage = ZSTDcs_ending;
lib/zstd/compress/zstd_compress.c
4605
return (size_t)(op-ostart);
lib/zstd/compress/zstd_compress.c
4613
BYTE* const op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
4634
op[pos++] = frameHeaderDescriptionByte;
lib/zstd/compress/zstd_compress.c
4635
if (!singleSegment) op[pos++] = windowLogByte;
lib/zstd/compress/zstd_compress.c
4642
case 1 : op[pos] = (BYTE)(dictID); pos++; break;
lib/zstd/compress/zstd_compress.c
4643
case 2 : MEM_writeLE16(op+pos, (U16)dictID); pos+=2; break;
lib/zstd/compress/zstd_compress.c
4644
case 3 : MEM_writeLE32(op+pos, dictID); pos+=4; break;
lib/zstd/compress/zstd_compress.c
4651
case 0 : if (singleSegment) op[pos++] = (BYTE)(pledgedSrcSize); break;
lib/zstd/compress/zstd_compress.c
4652
case 1 : MEM_writeLE16(op+pos, (U16)(pledgedSrcSize-256)); pos+=2; break;
lib/zstd/compress/zstd_compress.c
4653
case 2 : MEM_writeLE32(op+pos, (U32)(pledgedSrcSize)); pos+=4; break;
lib/zstd/compress/zstd_compress.c
4654
case 3 : MEM_writeLE64(op+pos, (U64)(pledgedSrcSize)); pos+=8; break;
lib/zstd/compress/zstd_compress.c
4667
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
4673
MEM_writeLE32(op, (U32)(ZSTD_MAGIC_SKIPPABLE_START + magicVariant));
lib/zstd/compress/zstd_compress.c
4674
MEM_writeLE32(op+4, (U32)srcSize);
lib/zstd/compress/zstd_compress.c
4675
ZSTD_memcpy(op+8, src, srcSize);
lib/zstd/compress/zstd_compress.c
5258
BYTE* op = ostart;
lib/zstd/compress/zstd_compress.c
5268
op += fhSize;
lib/zstd/compress/zstd_compress.c
5277
MEM_writeLE24(op, cBlockHeader24);
lib/zstd/compress/zstd_compress.c
5278
op += ZSTD_blockHeaderSize;
lib/zstd/compress/zstd_compress.c
5286
MEM_writeLE32(op, checksum);
lib/zstd/compress/zstd_compress.c
5287
op += 4;
lib/zstd/compress/zstd_compress.c
5291
return (size_t)(op-ostart);
lib/zstd/compress/zstd_compress.c
5999
char* op = (ostart != NULL) ? ostart + output->pos : ostart;
lib/zstd/compress/zstd_compress.c
6033
&& ( (size_t)(oend-op) >= ZSTD_compressBound((size_t)(iend-ip)) /* Enough output space */
lib/zstd/compress/zstd_compress.c
6038
op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress.c
6043
op += cSize;
lib/zstd/compress/zstd_compress.c
6086
size_t oSize = (size_t)(oend-op);
lib/zstd/compress/zstd_compress.c
6090
cDst = op; /* compress into output buffer, to skip flush stage */
lib/zstd/compress/zstd_compress.c
6122
if (cDst == op) { /* no need to flush */
lib/zstd/compress/zstd_compress.c
6123
op += cSize;
lib/zstd/compress/zstd_compress.c
6140
size_t const flushed = ZSTD_limitCopy(op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress.c
6143
(unsigned)toFlush, (unsigned)(oend-op), (unsigned)flushed);
lib/zstd/compress/zstd_compress.c
6145
op += flushed;
lib/zstd/compress/zstd_compress.c
6149
assert(op==oend);
lib/zstd/compress/zstd_compress.c
6170
output->pos = (size_t)(op - ostart);
lib/zstd/compress/zstd_compress.c
6746
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
6754
MEM_writeLE32(op, cBlockHeader24);
lib/zstd/compress/zstd_compress.c
6755
op += ZSTD_blockHeaderSize;
lib/zstd/compress/zstd_compress.c
6781
cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6786
op += cBlockSize;
lib/zstd/compress/zstd_compress.c
6796
op + ZSTD_blockHeaderSize /* Leave space for block header */, dstCapacity - ZSTD_blockHeaderSize,
lib/zstd/compress/zstd_compress.c
6815
cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6819
cBlockSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6831
MEM_writeLE24(op, cBlockHeader);
lib/zstd/compress/zstd_compress.c
6842
op += cBlockSize;
lib/zstd/compress/zstd_compress.c
6859
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
6868
{ size_t const frameHeaderSize = ZSTD_writeFrameHeader(op, dstCapacity,
lib/zstd/compress/zstd_compress.c
6870
op += frameHeaderSize;
lib/zstd/compress/zstd_compress.c
6881
op, dstCapacity,
lib/zstd/compress/zstd_compress.c
7276
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
7287
MEM_writeLE24(op, cBlockHeader24);
lib/zstd/compress/zstd_compress.c
7288
op += ZSTD_blockHeaderSize;
lib/zstd/compress/zstd_compress.c
7318
op + ZSTD_blockHeaderSize /* Leave space for block header */, dstCapacity - ZSTD_blockHeaderSize,
lib/zstd/compress/zstd_compress.c
7352
MEM_writeLE24(op, cBlockHeader);
lib/zstd/compress/zstd_compress.c
7358
op += cBlockSize;
lib/zstd/compress/zstd_compress.c
7382
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress.c
7404
{ size_t const frameHeaderSize = ZSTD_writeFrameHeader(op, dstCapacity,
lib/zstd/compress/zstd_compress.c
7406
op += frameHeaderSize;
lib/zstd/compress/zstd_compress.c
7414
op, dstCapacity,
lib/zstd/compress/zstd_compress_internal.h
656
BYTE* const op = (BYTE*)dst;
lib/zstd/compress/zstd_compress_internal.h
659
MEM_writeLE24(op, cBlockHeader);
lib/zstd/compress/zstd_compress_internal.h
660
op[3] = src;
lib/zstd/compress/zstd_compress_internal.h
698
ZSTD_safecopyLiterals(BYTE* op, BYTE const* ip, BYTE const* const iend, BYTE const* ilimit_w)
lib/zstd/compress/zstd_compress_internal.h
702
ZSTD_wildcopy(op, ip, ilimit_w - ip, ZSTD_no_overlap);
lib/zstd/compress/zstd_compress_internal.h
703
op += ilimit_w - ip;
lib/zstd/compress/zstd_compress_internal.h
706
while (ip < iend) *op++ = *ip++;
lib/zstd/compress/zstd_compress_sequences.c
252
BYTE* op = (BYTE*)dst;
lib/zstd/compress/zstd_compress_sequences.c
253
const BYTE* const oend = op + dstCapacity;
lib/zstd/compress/zstd_compress_sequences.c
260
*op = codeTable[0];
lib/zstd/compress/zstd_compress_sequences.c
280
assert(oend >= op);
lib/zstd/compress/zstd_compress_sequences.c
281
{ size_t const NCountSize = FSE_writeNCount(op, (size_t)(oend - op), wksp->norm, max, tableLog); /* overflow protected */
lib/zstd/compress/zstd_compress_superblock.c
125
DEBUGLOG(5, "Compressed literals: %u -> %u", (U32)litSize, (U32)(op-ostart));
lib/zstd/compress/zstd_compress_superblock.c
126
return (size_t)(op-ostart);
lib/zstd/compress/zstd_compress_superblock.c
175
BYTE* op = ostart;
lib/zstd/compress/zstd_compress_superblock.c
182
RETURN_ERROR_IF((oend-op) < 3 /*max nbSeq Size*/ + 1 /*seqHead*/,
lib/zstd/compress/zstd_compress_superblock.c
185
*op++ = (BYTE)nbSeq;
lib/zstd/compress/zstd_compress_superblock.c
187
op[0] = (BYTE)((nbSeq>>8) + 0x80), op[1] = (BYTE)nbSeq, op+=2;
lib/zstd/compress/zstd_compress_superblock.c
189
op[0]=0xFF, MEM_writeLE16(op+1, (U16)(nbSeq - LONGNBSEQ)), op+=3;
lib/zstd/compress/zstd_compress_superblock.c
191
return (size_t)(op - ostart);
lib/zstd/compress/zstd_compress_superblock.c
195
seqHead = op++;
lib/zstd/compress/zstd_compress_superblock.c
197
DEBUGLOG(5, "ZSTD_compressSubBlock_sequences (seqHeadSize=%u)", (unsigned)(op-ostart));
lib/zstd/compress/zstd_compress_superblock.c
205
ZSTD_memcpy(op, fseMetadata->fseTablesBuffer, fseMetadata->fseTablesSize);
lib/zstd/compress/zstd_compress_superblock.c
206
op += fseMetadata->fseTablesSize;
lib/zstd/compress/zstd_compress_superblock.c
213
op, (size_t)(oend - op),
lib/zstd/compress/zstd_compress_superblock.c
220
op += bitstreamSize;
lib/zstd/compress/zstd_compress_superblock.c
249
if (op-seqHead < 4) {
lib/zstd/compress/zstd_compress_superblock.c
257
return (size_t)(op - ostart);
lib/zstd/compress/zstd_compress_superblock.c
278
BYTE* op = ostart + ZSTD_blockHeaderSize;
lib/zstd/compress/zstd_compress_superblock.c
283
op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress_superblock.c
287
op += cLitSize;
lib/zstd/compress/zstd_compress_superblock.c
294
op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress_superblock.c
298
op += cSeqSize;
lib/zstd/compress/zstd_compress_superblock.c
301
{ size_t cSize = (size_t)(op-ostart) - ZSTD_blockHeaderSize;
lib/zstd/compress/zstd_compress_superblock.c
305
return (size_t)(op-ostart);
lib/zstd/compress/zstd_compress_superblock.c
502
BYTE* op = ostart;
lib/zstd/compress/zstd_compress_superblock.c
53
BYTE* op = ostart + lhSize;
lib/zstd/compress/zstd_compress_superblock.c
556
op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress_superblock.c
569
op += cSize;
lib/zstd/compress/zstd_compress_superblock.c
600
op, (size_t)(oend-op),
lib/zstd/compress/zstd_compress_superblock.c
613
op += cSize;
lib/zstd/compress/zstd_compress_superblock.c
644
size_t const cSize = ZSTD_noCompressBlock(op, (size_t)(oend - op), ip, rSize, lastBlock);
lib/zstd/compress/zstd_compress_superblock.c
648
op += cSize;
lib/zstd/compress/zstd_compress_superblock.c
662
(unsigned)(op-ostart));
lib/zstd/compress/zstd_compress_superblock.c
663
return (size_t)(op-ostart);
lib/zstd/compress/zstd_compress_superblock.c
73
ZSTD_memcpy(op, hufMetadata->hufDesBuffer, hufMetadata->hufDesSize);
lib/zstd/compress/zstd_compress_superblock.c
74
op += hufMetadata->hufDesSize;
lib/zstd/compress/zstd_compress_superblock.c
80
const size_t cSize = singleStream ? HUF_compress1X_usingCTable(op, (size_t)(oend-op), literals, litSize, hufTable, flags)
lib/zstd/compress/zstd_compress_superblock.c
81
: HUF_compress4X_usingCTable(op, (size_t)(oend-op), literals, litSize, hufTable, flags);
lib/zstd/compress/zstd_compress_superblock.c
82
op += cSize;
lib/zstd/decompress/huf_decompress.c
1263
HUF_decodeSymbolX2(void* op, BIT_DStream_t* DStream, const HUF_DEltX2* dt, const U32 dtLog)
lib/zstd/decompress/huf_decompress.c
1266
ZSTD_memcpy(op, &dt[val].sequence, 2);
lib/zstd/decompress/huf_decompress.c
1272
HUF_decodeLastSymbolX2(void* op, BIT_DStream_t* DStream, const HUF_DEltX2* dt, const U32 dtLog)
lib/zstd/decompress/huf_decompress.c
1275
ZSTD_memcpy(op, &dt[val].sequence, 1);
lib/zstd/decompress/huf_decompress.c
1525
BYTE* op[4];
lib/zstd/decompress/huf_decompress.c
1533
ZSTD_memcpy(&op, &args->op, sizeof(op));
lib/zstd/decompress/huf_decompress.c
1535
oend[0] = op[1];
lib/zstd/decompress/huf_decompress.c
1536
oend[1] = op[2];
lib/zstd/decompress/huf_decompress.c
1537
oend[2] = op[3];
lib/zstd/decompress/huf_decompress.c
1550
assert(op[stream] <= oend[stream]);
lib/zstd/decompress/huf_decompress.c
1570
size_t const oiters = (size_t)(oend[stream] - op[stream]) / 10;
lib/zstd/decompress/huf_decompress.c
1580
olimit = op[3] + (iters * 5);
lib/zstd/decompress/huf_decompress.c
1583
if (op[3] == olimit)
lib/zstd/decompress/huf_decompress.c
1607
MEM_write16(op[(_stream)], entry.sequence); \
lib/zstd/decompress/huf_decompress.c
1609
op[(_stream)] += (entry.length); \
lib/zstd/decompress/huf_decompress.c
1648
} while (op[3] < olimit);
lib/zstd/decompress/huf_decompress.c
1659
ZSTD_memcpy(&args->op, &op, sizeof(op));
lib/zstd/decompress/huf_decompress.c
1688
assert(args.op[3] <= oend);
lib/zstd/decompress/huf_decompress.c
1706
args.op[i] += HUF_decodeStreamX2(args.op[i], &bit, segmentEnd, (HUF_DEltX2 const*)dt, HUF_DECODER_FAST_TABLELOG);
lib/zstd/decompress/huf_decompress.c
1707
if (args.op[i] != segmentEnd)
lib/zstd/decompress/huf_decompress.c
172
BYTE* op[4];
lib/zstd/decompress/huf_decompress.c
244
args->op[0] = (BYTE*)dst;
lib/zstd/decompress/huf_decompress.c
245
args->op[1] = args->op[0] + (dstSize+3)/4;
lib/zstd/decompress/huf_decompress.c
246
args->op[2] = args->op[1] + (dstSize+3)/4;
lib/zstd/decompress/huf_decompress.c
247
args->op[3] = args->op[2] + (dstSize+3)/4;
lib/zstd/decompress/huf_decompress.c
250
if (args->op[3] >= oend)
lib/zstd/decompress/huf_decompress.c
281
if (args->op[stream] > segmentEnd)
lib/zstd/decompress/huf_decompress.c
577
BYTE* op = (BYTE*)dst;
lib/zstd/decompress/huf_decompress.c
578
BYTE* const oend = ZSTD_maybeNullPtrAdd(op, dstSize);
lib/zstd/decompress/huf_decompress.c
587
HUF_decodeStreamX1(op, &bitD, oend, dt, dtLog);
lib/zstd/decompress/huf_decompress.c
722
BYTE* op[4];
lib/zstd/decompress/huf_decompress.c
730
ZSTD_memcpy(&op, &args->op, sizeof(op));
lib/zstd/decompress/huf_decompress.c
742
assert(op[stream] <= (stream == 3 ? oend : op[stream + 1]));
lib/zstd/decompress/huf_decompress.c
749
size_t const oiters = (size_t)(oend - op[3]) / 5;
lib/zstd/decompress/huf_decompress.c
762
olimit = op[3] + symbols;
lib/zstd/decompress/huf_decompress.c
765
if (op[3] == olimit)
lib/zstd/decompress/huf_decompress.c
789
op[(_stream)][(_symbol)] = (BYTE)((entry >> 8) & 0xFF); \
lib/zstd/decompress/huf_decompress.c
797
op[(_stream)] += 5; \
lib/zstd/decompress/huf_decompress.c
816
} while (op[3] < olimit);
lib/zstd/decompress/huf_decompress.c
827
ZSTD_memcpy(&args->op, &op, sizeof(op));
lib/zstd/decompress/huf_decompress.c
864
assert(args.op[3] <= oend);
lib/zstd/decompress/huf_decompress.c
882
args.op[i] += HUF_decodeStreamX1(args.op[i], &bit, segmentEnd, (HUF_DEltX1 const*)dt, HUF_DECODER_FAST_TABLELOG);
lib/zstd/decompress/huf_decompress.c
883
if (args.op[i] != segmentEnd) return ERROR(corruption_detected);
lib/zstd/decompress/zstd_decompress.c
1002
RETURN_ERROR_IF((U64)(op-ostart) != dctx->fParams.frameContentSize,
lib/zstd/decompress/zstd_decompress.c
1016
ZSTD_DCtx_trace_end(dctx, (U64)(op-ostart), (U64)(ip-istart), /* streaming */ 0);
lib/zstd/decompress/zstd_decompress.c
1018
DEBUGLOG(4, "ZSTD_decompressFrame: decompressed frame of size %i, consuming %i bytes of input", (int)(op-ostart), (int)(ip - (const BYTE*)*srcPtr));
lib/zstd/decompress/zstd_decompress.c
1021
return (size_t)(op-ostart);
lib/zstd/decompress/zstd_decompress.c
1981
ZSTD_DStream* zds, char** op, char* oend,
lib/zstd/decompress/zstd_decompress.c
1997
size_t const dstSize = isSkipFrame ? 0 : (size_t)(oend - *op);
lib/zstd/decompress/zstd_decompress.c
1998
size_t const decodedSize = ZSTD_decompressContinue(zds, *op, dstSize, src, srcSize);
lib/zstd/decompress/zstd_decompress.c
2000
*op += decodedSize;
lib/zstd/decompress/zstd_decompress.c
2003
assert(*op <= oend);
lib/zstd/decompress/zstd_decompress.c
2018
char* op = ostart;
lib/zstd/decompress/zstd_decompress.c
2081
&& (U64)(size_t)(oend-op) >= zds->fParams.frameContentSize) {
lib/zstd/decompress/zstd_decompress.c
2085
size_t const decompressedSize = ZSTD_decompress_usingDDict(zds, op, (size_t)(oend-op), istart, cSize, ZSTD_getDDict(zds));
lib/zstd/decompress/zstd_decompress.c
2090
op = op ? op + decompressedSize : op; /* can occur if frameContentSize = 0 (empty frame) */
lib/zstd/decompress/zstd_decompress.c
2101
&& (U64)(size_t)(oend-op) < zds->fParams.frameContentSize) {
lib/zstd/decompress/zstd_decompress.c
2176
FORWARD_IF_ERROR(ZSTD_decompressContinueStream(zds, &op, oend, ip, neededInSize), "");
lib/zstd/decompress/zstd_decompress.c
2210
FORWARD_IF_ERROR(ZSTD_decompressContinueStream(zds, &op, oend, zds->inBuff, neededInSize), "");
lib/zstd/decompress/zstd_decompress.c
2217
size_t const flushedSize = ZSTD_limitCopy(op, (size_t)(oend-op), zds->outBuff + zds->outStart, toFlushSize);
lib/zstd/decompress/zstd_decompress.c
2219
op = op ? op + flushedSize : op;
lib/zstd/decompress/zstd_decompress.c
2244
output->pos = (size_t)(op - (char*)(output->dst));
lib/zstd/decompress/zstd_decompress.c
2249
if ((ip==istart) && (op==ostart)) { /* no forward progress */
lib/zstd/decompress/zstd_decompress.c
2252
RETURN_ERROR_IF(op==oend, noForwardProgress_destFull, "");
lib/zstd/decompress/zstd_decompress.c
917
BYTE* op = ostart;
lib/zstd/decompress/zstd_decompress.c
953
if (ip >= op && ip < oBlockEnd) {
lib/zstd/decompress/zstd_decompress.c
967
oBlockEnd = op + (ip - op);
lib/zstd/decompress/zstd_decompress.c
974
decodedSize = ZSTD_decompressBlock_internal(dctx, op, (size_t)(oBlockEnd-op), ip, cBlockSize, not_streaming);
lib/zstd/decompress/zstd_decompress.c
978
decodedSize = ZSTD_copyRawBlock(op, (size_t)(oend-op), ip, cBlockSize);
lib/zstd/decompress/zstd_decompress.c
981
decodedSize = ZSTD_setRleBlock(op, (size_t)(oBlockEnd-op), *ip, blockProperties.origSize);
lib/zstd/decompress/zstd_decompress.c
990
xxh64_update(&dctx->xxhState, op, decodedSize);
lib/zstd/decompress/zstd_decompress.c
993
op += decodedSize;
lib/zstd/decompress/zstd_decompress_block.c
1002
size_t ZSTD_execSequence(BYTE* op,
lib/zstd/decompress/zstd_decompress_block.c
1007
BYTE* const oLitEnd = op + sequence.litLength;
lib/zstd/decompress/zstd_decompress_block.c
1009
BYTE* const oMatchEnd = op + sequenceLength; /* risk : address space overflow (32-bits) */
lib/zstd/decompress/zstd_decompress_block.c
1014
assert(op != NULL /* Precondition */);
lib/zstd/decompress/zstd_decompress_block.c
1029
(MEM_32bits() && (size_t)(oend - op) < sequenceLength + WILDCOPY_OVERLENGTH)))
lib/zstd/decompress/zstd_decompress_block.c
1030
return ZSTD_execSequenceEnd(op, oend, sequence, litPtr, litLimit, prefixStart, virtualStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1033
assert(op <= oLitEnd /* No overflow */);
lib/zstd/decompress/zstd_decompress_block.c
1045
ZSTD_copy16(op, (*litPtr));
lib/zstd/decompress/zstd_decompress_block.c
1047
ZSTD_wildcopy(op + 16, (*litPtr) + 16, sequence.litLength - 16, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
1049
op = oLitEnd;
lib/zstd/decompress/zstd_decompress_block.c
1064
op = oLitEnd + length1;
lib/zstd/decompress/zstd_decompress_block.c
1070
assert(op <= oMatchEnd);
lib/zstd/decompress/zstd_decompress_block.c
1083
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
1089
ZSTD_overlapCopy8(&op, &match, sequence.offset);
lib/zstd/decompress/zstd_decompress_block.c
1093
assert(op < oMatchEnd);
lib/zstd/decompress/zstd_decompress_block.c
1094
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength - 8, ZSTD_overlap_src_before_dst);
lib/zstd/decompress/zstd_decompress_block.c
1101
size_t ZSTD_execSequenceSplitLitBuffer(BYTE* op,
lib/zstd/decompress/zstd_decompress_block.c
1106
BYTE* const oLitEnd = op + sequence.litLength;
lib/zstd/decompress/zstd_decompress_block.c
1108
BYTE* const oMatchEnd = op + sequenceLength; /* risk : address space overflow (32-bits) */
lib/zstd/decompress/zstd_decompress_block.c
1112
assert(op != NULL /* Precondition */);
lib/zstd/decompress/zstd_decompress_block.c
1122
(MEM_32bits() && (size_t)(oend - op) < sequenceLength + WILDCOPY_OVERLENGTH)))
lib/zstd/decompress/zstd_decompress_block.c
1123
return ZSTD_execSequenceEndSplitLitBuffer(op, oend, oend_w, sequence, litPtr, litLimit, prefixStart, virtualStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1126
assert(op <= oLitEnd /* No overflow */);
lib/zstd/decompress/zstd_decompress_block.c
1138
ZSTD_copy16(op, (*litPtr));
lib/zstd/decompress/zstd_decompress_block.c
1140
ZSTD_wildcopy(op+16, (*litPtr)+16, sequence.litLength-16, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
1142
op = oLitEnd;
lib/zstd/decompress/zstd_decompress_block.c
1157
op = oLitEnd + length1;
lib/zstd/decompress/zstd_decompress_block.c
1162
assert(op <= oMatchEnd);
lib/zstd/decompress/zstd_decompress_block.c
1175
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
1181
ZSTD_overlapCopy8(&op, &match, sequence.offset);
lib/zstd/decompress/zstd_decompress_block.c
1185
assert(op < oMatchEnd);
lib/zstd/decompress/zstd_decompress_block.c
1186
ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8, ZSTD_overlap_src_before_dst);
lib/zstd/decompress/zstd_decompress_block.c
1369
BYTE const* op, BYTE const* oend,
lib/zstd/decompress/zstd_decompress_block.c
1377
BYTE const* const oLitEnd = op + seq.litLength;
lib/zstd/decompress/zstd_decompress_block.c
1380
assert(op <= oend);
lib/zstd/decompress/zstd_decompress_block.c
1381
assert((size_t)(oend - op) >= sequenceSize);
lib/zstd/decompress/zstd_decompress_block.c
1394
(void)dctx, (void)op, (void)oend, (void)seq, (void)prefixStart, (void)virtualStart;
lib/zstd/decompress/zstd_decompress_block.c
1413
BYTE* op = ostart;
lib/zstd/decompress/zstd_decompress_block.c
1505
{ size_t const oneSeqSize = ZSTD_execSequenceSplitLitBuffer(op, oend, litPtr + sequence.litLength - WILDCOPY_OVERLENGTH, sequence, &litPtr, litBufferEnd, prefixStart, vBase, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1508
ZSTD_assertValidSequence(dctx, op, oend, sequence, prefixStart, vBase);
lib/zstd/decompress/zstd_decompress_block.c
1513
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1522
RETURN_ERROR_IF(leftoverLit > (size_t)(oend - op), dstSize_tooSmall, "remaining lit must fit within dstBuffer");
lib/zstd/decompress/zstd_decompress_block.c
1523
ZSTD_safecopyDstBeforeSrc(op, litPtr, leftoverLit);
lib/zstd/decompress/zstd_decompress_block.c
1525
op += leftoverLit;
lib/zstd/decompress/zstd_decompress_block.c
1530
{ size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequence, &litPtr, litBufferEnd, prefixStart, vBase, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1533
ZSTD_assertValidSequence(dctx, op, oend, sequence, prefixStart, vBase);
lib/zstd/decompress/zstd_decompress_block.c
1538
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1566
size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequence, &litPtr, litBufferEnd, prefixStart, vBase, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1569
ZSTD_assertValidSequence(dctx, op, oend, sequence, prefixStart, vBase);
lib/zstd/decompress/zstd_decompress_block.c
1574
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1592
RETURN_ERROR_IF(lastLLSize > (size_t)(oend - op), dstSize_tooSmall, "");
lib/zstd/decompress/zstd_decompress_block.c
1593
if (op != NULL) {
lib/zstd/decompress/zstd_decompress_block.c
1594
ZSTD_memmove(op, litPtr, lastLLSize);
lib/zstd/decompress/zstd_decompress_block.c
1595
op += lastLLSize;
lib/zstd/decompress/zstd_decompress_block.c
1604
RETURN_ERROR_IF(lastLLSize > (size_t)(oend-op), dstSize_tooSmall, "");
lib/zstd/decompress/zstd_decompress_block.c
1605
if (op != NULL) {
lib/zstd/decompress/zstd_decompress_block.c
1606
ZSTD_memcpy(op, litPtr, lastLLSize);
lib/zstd/decompress/zstd_decompress_block.c
1607
op += lastLLSize;
lib/zstd/decompress/zstd_decompress_block.c
1610
DEBUGLOG(6, "decoded block of size %u bytes", (U32)(op - ostart));
lib/zstd/decompress/zstd_decompress_block.c
1611
return (size_t)(op - ostart);
lib/zstd/decompress/zstd_decompress_block.c
1625
BYTE* op = ostart;
lib/zstd/decompress/zstd_decompress_block.c
1662
size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequence, &litPtr, litEnd, prefixStart, vBase, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1665
ZSTD_assertValidSequence(dctx, op, oend, sequence, prefixStart, vBase);
lib/zstd/decompress/zstd_decompress_block.c
1670
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1683
RETURN_ERROR_IF(lastLLSize > (size_t)(oend-op), dstSize_tooSmall, "");
lib/zstd/decompress/zstd_decompress_block.c
1684
if (op != NULL) {
lib/zstd/decompress/zstd_decompress_block.c
1685
ZSTD_memcpy(op, litPtr, lastLLSize);
lib/zstd/decompress/zstd_decompress_block.c
1686
op += lastLLSize;
lib/zstd/decompress/zstd_decompress_block.c
1689
DEBUGLOG(6, "decoded block of size %u bytes", (U32)(op - ostart));
lib/zstd/decompress/zstd_decompress_block.c
1690
return (size_t)(op - ostart);
lib/zstd/decompress/zstd_decompress_block.c
1744
BYTE* op = ostart;
lib/zstd/decompress/zstd_decompress_block.c
1760
size_t prefetchPos = (size_t)(op-prefixStart); /* track position relative to prefixStart */
lib/zstd/decompress/zstd_decompress_block.c
1789
RETURN_ERROR_IF(leftoverLit > (size_t)(oend - op), dstSize_tooSmall, "remaining lit must fit within dstBuffer");
lib/zstd/decompress/zstd_decompress_block.c
1790
ZSTD_safecopyDstBeforeSrc(op, litPtr, leftoverLit);
lib/zstd/decompress/zstd_decompress_block.c
1792
op += leftoverLit;
lib/zstd/decompress/zstd_decompress_block.c
1797
{ size_t const oneSeqSize = ZSTD_execSequence(op, oend, sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK], &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1800
ZSTD_assertValidSequence(dctx, op, oend, sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK], prefixStart, dictStart);
lib/zstd/decompress/zstd_decompress_block.c
1806
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1812
ZSTD_execSequenceSplitLitBuffer(op, oend, litPtr + sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK].litLength - WILDCOPY_OVERLENGTH, sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK], &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd) :
lib/zstd/decompress/zstd_decompress_block.c
1813
ZSTD_execSequence(op, oend, sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK], &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1816
ZSTD_assertValidSequence(dctx, op, oend, sequences[(seqNb - ADVANCED_SEQS) & STORED_SEQS_MASK], prefixStart, dictStart);
lib/zstd/decompress/zstd_decompress_block.c
1822
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1834
RETURN_ERROR_IF(leftoverLit > (size_t)(oend - op), dstSize_tooSmall, "remaining lit must fit within dstBuffer");
lib/zstd/decompress/zstd_decompress_block.c
1835
ZSTD_safecopyDstBeforeSrc(op, litPtr, leftoverLit);
lib/zstd/decompress/zstd_decompress_block.c
1837
op += leftoverLit;
lib/zstd/decompress/zstd_decompress_block.c
1842
{ size_t const oneSeqSize = ZSTD_execSequence(op, oend, *sequence, &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1845
ZSTD_assertValidSequence(dctx, op, oend, sequences[seqNb&STORED_SEQS_MASK], prefixStart, dictStart);
lib/zstd/decompress/zstd_decompress_block.c
1848
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1854
ZSTD_execSequenceSplitLitBuffer(op, oend, litPtr + sequence->litLength - WILDCOPY_OVERLENGTH, *sequence, &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd) :
lib/zstd/decompress/zstd_decompress_block.c
1855
ZSTD_execSequence(op, oend, *sequence, &litPtr, litBufferEnd, prefixStart, dictStart, dictEnd);
lib/zstd/decompress/zstd_decompress_block.c
1858
ZSTD_assertValidSequence(dctx, op, oend, sequences[seqNb&STORED_SEQS_MASK], prefixStart, dictStart);
lib/zstd/decompress/zstd_decompress_block.c
1861
op += oneSeqSize;
lib/zstd/decompress/zstd_decompress_block.c
1872
RETURN_ERROR_IF(lastLLSize > (size_t)(oend - op), dstSize_tooSmall, "");
lib/zstd/decompress/zstd_decompress_block.c
1873
if (op != NULL) {
lib/zstd/decompress/zstd_decompress_block.c
1874
ZSTD_memmove(op, litPtr, lastLLSize);
lib/zstd/decompress/zstd_decompress_block.c
1875
op += lastLLSize;
lib/zstd/decompress/zstd_decompress_block.c
1881
RETURN_ERROR_IF(lastLLSize > (size_t)(oend-op), dstSize_tooSmall, "");
lib/zstd/decompress/zstd_decompress_block.c
1882
if (op != NULL) {
lib/zstd/decompress/zstd_decompress_block.c
1883
ZSTD_memmove(op, litPtr, lastLLSize);
lib/zstd/decompress/zstd_decompress_block.c
1884
op += lastLLSize;
lib/zstd/decompress/zstd_decompress_block.c
1888
return (size_t)(op - ostart);
lib/zstd/decompress/zstd_decompress_block.c
1997
static size_t ZSTD_totalHistorySize(BYTE* op, BYTE const* virtualStart)
lib/zstd/decompress/zstd_decompress_block.c
1999
return (size_t)(op - virtualStart);
lib/zstd/decompress/zstd_decompress_block.c
805
HINT_INLINE void ZSTD_overlapCopy8(BYTE** op, BYTE const** ip, size_t offset) {
lib/zstd/decompress/zstd_decompress_block.c
806
assert(*ip <= *op);
lib/zstd/decompress/zstd_decompress_block.c
812
(*op)[0] = (*ip)[0];
lib/zstd/decompress/zstd_decompress_block.c
813
(*op)[1] = (*ip)[1];
lib/zstd/decompress/zstd_decompress_block.c
814
(*op)[2] = (*ip)[2];
lib/zstd/decompress/zstd_decompress_block.c
815
(*op)[3] = (*ip)[3];
lib/zstd/decompress/zstd_decompress_block.c
817
ZSTD_copy4(*op+4, *ip);
lib/zstd/decompress/zstd_decompress_block.c
820
ZSTD_copy8(*op, *ip);
lib/zstd/decompress/zstd_decompress_block.c
823
*op += 8;
lib/zstd/decompress/zstd_decompress_block.c
824
assert(*op - *ip >= 8);
lib/zstd/decompress/zstd_decompress_block.c
838
static void ZSTD_safecopy(BYTE* op, const BYTE* const oend_w, BYTE const* ip, ptrdiff_t length, ZSTD_overlap_e ovtype) {
lib/zstd/decompress/zstd_decompress_block.c
839
ptrdiff_t const diff = op - ip;
lib/zstd/decompress/zstd_decompress_block.c
840
BYTE* const oend = op + length;
lib/zstd/decompress/zstd_decompress_block.c
842
assert((ovtype == ZSTD_no_overlap && (diff <= -8 || diff >= 8 || op >= oend_w)) ||
lib/zstd/decompress/zstd_decompress_block.c
847
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
853
ZSTD_overlapCopy8(&op, &ip, diff);
lib/zstd/decompress/zstd_decompress_block.c
855
assert(op - ip >= 8);
lib/zstd/decompress/zstd_decompress_block.c
856
assert(op <= oend);
lib/zstd/decompress/zstd_decompress_block.c
861
ZSTD_wildcopy(op, ip, length, ovtype);
lib/zstd/decompress/zstd_decompress_block.c
864
if (op <= oend_w) {
lib/zstd/decompress/zstd_decompress_block.c
867
ZSTD_wildcopy(op, ip, oend_w - op, ovtype);
lib/zstd/decompress/zstd_decompress_block.c
868
ip += oend_w - op;
lib/zstd/decompress/zstd_decompress_block.c
869
op += oend_w - op;
lib/zstd/decompress/zstd_decompress_block.c
872
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
878
static void ZSTD_safecopyDstBeforeSrc(BYTE* op, const BYTE* ip, ptrdiff_t length) {
lib/zstd/decompress/zstd_decompress_block.c
879
ptrdiff_t const diff = op - ip;
lib/zstd/decompress/zstd_decompress_block.c
880
BYTE* const oend = op + length;
lib/zstd/decompress/zstd_decompress_block.c
884
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
888
if (op <= oend - WILDCOPY_OVERLENGTH && diff < -WILDCOPY_VECLEN) {
lib/zstd/decompress/zstd_decompress_block.c
889
ZSTD_wildcopy(op, ip, oend - WILDCOPY_OVERLENGTH - op, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
890
ip += oend - WILDCOPY_OVERLENGTH - op;
lib/zstd/decompress/zstd_decompress_block.c
891
op += oend - WILDCOPY_OVERLENGTH - op;
lib/zstd/decompress/zstd_decompress_block.c
895
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
908
size_t ZSTD_execSequenceEnd(BYTE* op,
lib/zstd/decompress/zstd_decompress_block.c
913
BYTE* const oLitEnd = op + sequence.litLength;
lib/zstd/decompress/zstd_decompress_block.c
920
RETURN_ERROR_IF(sequenceLength > (size_t)(oend - op), dstSize_tooSmall, "last match must fit within dstBuffer");
lib/zstd/decompress/zstd_decompress_block.c
922
assert(op < op + sequenceLength);
lib/zstd/decompress/zstd_decompress_block.c
923
assert(oLitEnd < op + sequenceLength);
lib/zstd/decompress/zstd_decompress_block.c
926
ZSTD_safecopy(op, oend_w, *litPtr, sequence.litLength, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
927
op = oLitEnd;
lib/zstd/decompress/zstd_decompress_block.c
942
op = oLitEnd + length1;
lib/zstd/decompress/zstd_decompress_block.c
947
ZSTD_safecopy(op, oend_w, match, sequence.matchLength, ZSTD_overlap_src_before_dst);
lib/zstd/decompress/zstd_decompress_block.c
956
size_t ZSTD_execSequenceEndSplitLitBuffer(BYTE* op,
lib/zstd/decompress/zstd_decompress_block.c
961
BYTE* const oLitEnd = op + sequence.litLength;
lib/zstd/decompress/zstd_decompress_block.c
968
RETURN_ERROR_IF(sequenceLength > (size_t)(oend - op), dstSize_tooSmall, "last match must fit within dstBuffer");
lib/zstd/decompress/zstd_decompress_block.c
970
assert(op < op + sequenceLength);
lib/zstd/decompress/zstd_decompress_block.c
971
assert(oLitEnd < op + sequenceLength);
lib/zstd/decompress/zstd_decompress_block.c
974
RETURN_ERROR_IF(op > *litPtr && op < *litPtr + sequence.litLength, dstSize_tooSmall, "output should not catch up to and overwrite literal buffer");
lib/zstd/decompress/zstd_decompress_block.c
975
ZSTD_safecopyDstBeforeSrc(op, *litPtr, sequence.litLength);
lib/zstd/decompress/zstd_decompress_block.c
976
op = oLitEnd;
lib/zstd/decompress/zstd_decompress_block.c
991
op = oLitEnd + length1;
lib/zstd/decompress/zstd_decompress_block.c
996
ZSTD_safecopy(op, oend_w, match, sequence.matchLength, ZSTD_overlap_src_before_dst);
mm/swap.c
204
#define folio_batch_add_and_move(folio, op) \
mm/swap.c
206
&cpu_fbatches.op, \
mm/swap.c
208
op, \
mm/swap.c
209
offsetof(struct cpu_fbatches, op) >= \
mm/vmscan.c
4310
static void lru_gen_rotate_memcg(struct lruvec *lruvec, int op)
mm/vmscan.c
4326
if (op == MEMCG_LRU_HEAD)
mm/vmscan.c
4328
else if (op == MEMCG_LRU_TAIL)
mm/vmscan.c
4330
else if (op == MEMCG_LRU_OLD)
mm/vmscan.c
4332
else if (op == MEMCG_LRU_YOUNG)
mm/vmscan.c
4342
if (op == MEMCG_LRU_HEAD || op == MEMCG_LRU_OLD)
mm/vmscan.c
4949
int op;
mm/vmscan.c
4961
op = 0;
mm/vmscan.c
4967
if (op) {
mm/vmscan.c
4968
lru_gen_rotate_memcg(lruvec, op);
mm/vmscan.c
4969
op = 0;
mm/vmscan.c
4989
op = shrink_one(lruvec, sc);
mm/vmscan.c
4999
if (op)
mm/vmscan.c
5000
lru_gen_rotate_memcg(lruvec, op);
net/batman-adv/distributed-arp-table.c
1470
__u8 op;
net/batman-adv/distributed-arp-table.c
1497
return dhcp_h->op;
net/batman-adv/distributed-arp-table.c
74
__u8 op;
net/bluetooth/hci_event.c
4065
.op = _op, \
net/bluetooth/hci_event.c
4078
u16 op;
net/bluetooth/hci_event.c
4254
cc->op, skb->len, cc->min_len);
net/bluetooth/hci_event.c
4264
cc->op, skb->len, cc->max_len);
net/bluetooth/hci_event.c
4266
data = hci_cc_skb_pull(hdev, skb, cc->op, cc->min_len);
net/bluetooth/hci_event.c
4286
if (hci_cc_table[i].op == *opcode) {
net/bluetooth/hci_event.c
4369
.op = _op, \
net/bluetooth/hci_event.c
4374
u16 op;
net/bluetooth/hci_event.c
4418
if (hci_cs_table[i].op == *opcode) {
net/bluetooth/hci_event.c
60
u16 op, size_t len)
net/bluetooth/hci_event.c
66
bt_dev_err(hdev, "Malformed Command Complete: 0x%4.4x", op);
net/bluetooth/mgmt.c
5413
void *data, u16 len, u16 op)
net/bluetooth/mgmt.c
5430
cmd = mgmt_pending_add(sk, op, hdev, data, len);
net/bluetooth/mgmt.c
5455
return mgmt_cmd_status(sk, hdev->id, op, status);
net/bluetooth/mgmt.c
5999
u16 op, void *data, u16 len)
net/bluetooth/mgmt.c
6011
err = mgmt_cmd_complete(sk, hdev->id, op,
net/bluetooth/mgmt.c
6019
err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY,
net/bluetooth/mgmt.c
6025
err = mgmt_cmd_complete(sk, hdev->id, op, status,
net/bluetooth/mgmt.c
6032
err = mgmt_cmd_complete(sk, hdev->id, op, MGMT_STATUS_BUSY,
net/bluetooth/mgmt.c
6044
if (op == MGMT_OP_START_LIMITED_DISCOVERY)
net/bluetooth/mgmt.c
6049
cmd = mgmt_pending_add(sk, op, hdev, data, len);
net/bridge/br_fdb.c
341
struct net_bridge_port *op;
net/bridge/br_fdb.c
345
list_for_each_entry(op, &br->port_list, list) {
net/bridge/br_fdb.c
346
vg = nbp_vlan_group(op);
net/bridge/br_fdb.c
347
if (op != p && ether_addr_equal(op->dev->dev_addr, addr) &&
net/bridge/br_fdb.c
349
f->dst = op;
net/can/bcm.c
1001
if (op->flags & SETTIMER)
net/can/bcm.c
1002
op->count = msg_head->count;
net/can/bcm.c
1004
spin_unlock_bh(&op->bcm_tx_lock);
net/can/bcm.c
1010
op = kzalloc(OPSIZ, GFP_KERNEL);
net/can/bcm.c
1011
if (!op)
net/can/bcm.c
1014
spin_lock_init(&op->bcm_tx_lock);
net/can/bcm.c
1015
op->can_id = msg_head->can_id;
net/can/bcm.c
1016
op->cfsiz = CFSIZ(msg_head->flags);
net/can/bcm.c
1017
op->flags = msg_head->flags;
net/can/bcm.c
1018
op->nframes = msg_head->nframes;
net/can/bcm.c
1020
if (op->flags & SETTIMER)
net/can/bcm.c
1021
op->count = msg_head->count;
net/can/bcm.c
1025
op->frames = kmalloc_array(msg_head->nframes,
net/can/bcm.c
1026
op->cfsiz,
net/can/bcm.c
1028
if (!op->frames) {
net/can/bcm.c
1029
kfree(op);
net/can/bcm.c
1033
op->frames = &op->sframe;
net/can/bcm.c
1037
cf = op->frames + op->cfsiz * i;
net/can/bcm.c
1038
err = memcpy_from_msg((u8 *)cf, msg, op->cfsiz);
net/can/bcm.c
1042
if (op->flags & CAN_FD_FRAME) {
net/can/bcm.c
1060
op->last_frames = NULL;
net/can/bcm.c
1063
op->sk = sk;
net/can/bcm.c
1064
op->ifindex = ifindex;
net/can/bcm.c
1067
hrtimer_setup(&op->timer, bcm_tx_timeout_handler, CLOCK_MONOTONIC,
net/can/bcm.c
1071
hrtimer_setup(&op->thrtimer, hrtimer_dummy_timeout, CLOCK_MONOTONIC,
net/can/bcm.c
1075
list_add(&op->list, &bo->tx_ops);
net/can/bcm.c
1079
if (op->flags & SETTIMER) {
net/can/bcm.c
1081
op->ival1 = msg_head->ival1;
net/can/bcm.c
1082
op->ival2 = msg_head->ival2;
net/can/bcm.c
1083
op->kt_ival1 = bcm_timeval_to_ktime(msg_head->ival1);
net/can/bcm.c
1084
op->kt_ival2 = bcm_timeval_to_ktime(msg_head->ival2);
net/can/bcm.c
1087
if (!op->kt_ival1 && !op->kt_ival2)
net/can/bcm.c
1088
hrtimer_cancel(&op->timer);
net/can/bcm.c
1091
if (op->flags & STARTTIMER) {
net/can/bcm.c
1092
hrtimer_cancel(&op->timer);
net/can/bcm.c
1094
op->flags |= TX_ANNOUNCE;
net/can/bcm.c
1097
if (op->flags & TX_ANNOUNCE)
net/can/bcm.c
1098
bcm_can_tx(op);
net/can/bcm.c
1100
if (op->flags & STARTTIMER)
net/can/bcm.c
1101
bcm_tx_start_timer(op);
net/can/bcm.c
1103
return msg_head->nframes * op->cfsiz + MHSIZ;
net/can/bcm.c
1106
if (op->frames != &op->sframe)
net/can/bcm.c
1107
kfree(op->frames);
net/can/bcm.c
1108
kfree(op);
net/can/bcm.c
1119
struct bcm_op *op;
net/can/bcm.c
1144
op = bcm_find_op(&bo->rx_ops, msg_head, ifindex);
net/can/bcm.c
1145
if (op) {
net/can/bcm.c
1153
if (msg_head->nframes > op->nframes)
net/can/bcm.c
1158
err = memcpy_from_msg(op->frames, msg,
net/can/bcm.c
1159
msg_head->nframes * op->cfsiz);
net/can/bcm.c
1164
memset(op->last_frames, 0, msg_head->nframes * op->cfsiz);
net/can/bcm.c
1167
op->nframes = msg_head->nframes;
net/can/bcm.c
1168
op->flags = msg_head->flags;
net/can/bcm.c
1175
op = kzalloc(OPSIZ, GFP_KERNEL);
net/can/bcm.c
1176
if (!op)
net/can/bcm.c
1179
spin_lock_init(&op->bcm_tx_lock);
net/can/bcm.c
1180
op->can_id = msg_head->can_id;
net/can/bcm.c
1181
op->nframes = msg_head->nframes;
net/can/bcm.c
1182
op->cfsiz = CFSIZ(msg_head->flags);
net/can/bcm.c
1183
op->flags = msg_head->flags;
net/can/bcm.c
1187
op->frames = kmalloc_array(msg_head->nframes,
net/can/bcm.c
1188
op->cfsiz,
net/can/bcm.c
1190
if (!op->frames) {
net/can/bcm.c
1191
kfree(op);
net/can/bcm.c
1196
op->last_frames = kcalloc(msg_head->nframes,
net/can/bcm.c
1197
op->cfsiz,
net/can/bcm.c
1199
if (!op->last_frames) {
net/can/bcm.c
1200
kfree(op->frames);
net/can/bcm.c
1201
kfree(op);
net/can/bcm.c
1206
op->frames = &op->sframe;
net/can/bcm.c
1207
op->last_frames = &op->last_sframe;
net/can/bcm.c
1211
err = memcpy_from_msg(op->frames, msg,
net/can/bcm.c
1212
msg_head->nframes * op->cfsiz);
net/can/bcm.c
1214
if (op->frames != &op->sframe)
net/can/bcm.c
1215
kfree(op->frames);
net/can/bcm.c
1216
if (op->last_frames != &op->last_sframe)
net/can/bcm.c
1217
kfree(op->last_frames);
net/can/bcm.c
1218
kfree(op);
net/can/bcm.c
1224
op->sk = sk;
net/can/bcm.c
1225
op->ifindex = ifindex;
net/can/bcm.c
1228
op->rx_ifindex = ifindex;
net/can/bcm.c
1231
hrtimer_setup(&op->timer, bcm_rx_timeout_handler, CLOCK_MONOTONIC,
net/can/bcm.c
1233
hrtimer_setup(&op->thrtimer, bcm_rx_thr_handler, CLOCK_MONOTONIC,
net/can/bcm.c
1237
list_add(&op->list, &bo->rx_ops);
net/can/bcm.c
1246
if (op->flags & RX_RTR_FRAME) {
net/can/bcm.c
1247
struct canfd_frame *frame0 = op->frames;
net/can/bcm.c
1250
hrtimer_cancel(&op->thrtimer);
net/can/bcm.c
1251
hrtimer_cancel(&op->timer);
net/can/bcm.c
1258
if ((op->flags & TX_CP_CAN_ID) ||
net/can/bcm.c
1259
(frame0->can_id == op->can_id))
net/can/bcm.c
1260
frame0->can_id = op->can_id & ~CAN_RTR_FLAG;
net/can/bcm.c
1263
if (op->flags & SETTIMER) {
net/can/bcm.c
1266
op->ival1 = msg_head->ival1;
net/can/bcm.c
1267
op->ival2 = msg_head->ival2;
net/can/bcm.c
1268
op->kt_ival1 = bcm_timeval_to_ktime(msg_head->ival1);
net/can/bcm.c
1269
op->kt_ival2 = bcm_timeval_to_ktime(msg_head->ival2);
net/can/bcm.c
1272
if (!op->kt_ival1)
net/can/bcm.c
1273
hrtimer_cancel(&op->timer);
net/can/bcm.c
1279
op->kt_lastmsg = 0;
net/can/bcm.c
1280
hrtimer_cancel(&op->thrtimer);
net/can/bcm.c
1281
bcm_rx_thr_flush(op);
net/can/bcm.c
1284
if ((op->flags & STARTTIMER) && op->kt_ival1)
net/can/bcm.c
1285
hrtimer_start(&op->timer, op->kt_ival1,
net/can/bcm.c
1297
op->can_id,
net/can/bcm.c
1298
REGMASK(op->can_id),
net/can/bcm.c
1299
bcm_rx_handler, op,
net/can/bcm.c
1302
op->rx_reg_dev = dev;
net/can/bcm.c
1307
err = can_rx_register(sock_net(sk), NULL, op->can_id,
net/can/bcm.c
1308
REGMASK(op->can_id),
net/can/bcm.c
1309
bcm_rx_handler, op, "bcm", sk);
net/can/bcm.c
1312
list_del_rcu(&op->list);
net/can/bcm.c
1313
bcm_remove_op(op);
net/can/bcm.c
1318
return msg_head->nframes * op->cfsiz + MHSIZ;
net/can/bcm.c
1492
struct bcm_op *op;
net/can/bcm.c
1504
list_for_each_entry(op, &bo->rx_ops, list)
net/can/bcm.c
1505
if (op->rx_reg_dev == dev)
net/can/bcm.c
1506
bcm_rx_unreg(dev, op);
net/can/bcm.c
1593
struct bcm_op *op, *next;
net/can/bcm.c
1620
list_for_each_entry_safe(op, next, &bo->tx_ops, list)
net/can/bcm.c
1621
bcm_remove_op(op);
net/can/bcm.c
1623
list_for_each_entry_safe(op, next, &bo->rx_ops, list) {
net/can/bcm.c
1628
if (op->ifindex) {
net/can/bcm.c
1634
if (op->rx_reg_dev) {
net/can/bcm.c
1637
dev = dev_get_by_index(net, op->ifindex);
net/can/bcm.c
1639
bcm_rx_unreg(dev, op);
net/can/bcm.c
1644
can_rx_unregister(net, NULL, op->can_id,
net/can/bcm.c
1645
REGMASK(op->can_id),
net/can/bcm.c
1646
bcm_rx_handler, op);
net/can/bcm.c
1652
list_for_each_entry_safe(op, next, &bo->rx_ops, list)
net/can/bcm.c
1653
bcm_remove_op(op);
net/can/bcm.c
214
struct bcm_op *op;
net/can/bcm.c
225
list_for_each_entry_rcu(op, &bo->rx_ops, list) {
net/can/bcm.c
230
if (!op->frames_abs)
net/can/bcm.c
233
seq_printf(m, "rx_op: %03X %-5s ", op->can_id,
net/can/bcm.c
234
bcm_proc_getifname(net, ifname, op->ifindex));
net/can/bcm.c
236
if (op->flags & CAN_FD_FRAME)
net/can/bcm.c
237
seq_printf(m, "(%u)", op->nframes);
net/can/bcm.c
239
seq_printf(m, "[%u]", op->nframes);
net/can/bcm.c
241
seq_printf(m, "%c ", (op->flags & RX_CHECK_DLC) ? 'd' : ' ');
net/can/bcm.c
243
if (op->kt_ival1)
net/can/bcm.c
245
(long long)ktime_to_us(op->kt_ival1));
net/can/bcm.c
247
if (op->kt_ival2)
net/can/bcm.c
249
(long long)ktime_to_us(op->kt_ival2));
net/can/bcm.c
252
op->frames_filtered, op->frames_abs);
net/can/bcm.c
254
reduction = 100 - (op->frames_filtered * 100) / op->frames_abs;
net/can/bcm.c
260
list_for_each_entry(op, &bo->tx_ops, list) {
net/can/bcm.c
262
seq_printf(m, "tx_op: %03X %s ", op->can_id,
net/can/bcm.c
263
bcm_proc_getifname(net, ifname, op->ifindex));
net/can/bcm.c
265
if (op->flags & CAN_FD_FRAME)
net/can/bcm.c
266
seq_printf(m, "(%u) ", op->nframes);
net/can/bcm.c
268
seq_printf(m, "[%u] ", op->nframes);
net/can/bcm.c
270
if (op->kt_ival1)
net/can/bcm.c
272
(long long)ktime_to_us(op->kt_ival1));
net/can/bcm.c
274
if (op->kt_ival2)
net/can/bcm.c
276
(long long)ktime_to_us(op->kt_ival2));
net/can/bcm.c
278
seq_printf(m, "# sent %ld\n", op->frames_abs);
net/can/bcm.c
292
static void bcm_can_tx(struct bcm_op *op)
net/can/bcm.c
301
if (!op->ifindex)
net/can/bcm.c
305
spin_lock_bh(&op->bcm_tx_lock);
net/can/bcm.c
306
cf = op->frames + op->cfsiz * op->currframe;
net/can/bcm.c
307
spin_unlock_bh(&op->bcm_tx_lock);
net/can/bcm.c
309
dev = dev_get_by_index(sock_net(op->sk), op->ifindex);
net/can/bcm.c
315
skb = alloc_skb(op->cfsiz, gfp_any());
net/can/bcm.c
327
skb_put_data(skb, cf, op->cfsiz);
net/can/bcm.c
331
can_skb_set_owner(skb, op->sk);
net/can/bcm.c
335
spin_lock_bh(&op->bcm_tx_lock);
net/can/bcm.c
338
op->frames_abs++;
net/can/bcm.c
340
op->currframe++;
net/can/bcm.c
343
if (op->currframe >= op->nframes)
net/can/bcm.c
344
op->currframe = 0;
net/can/bcm.c
346
if (op->count > 0)
net/can/bcm.c
347
op->count--;
net/can/bcm.c
349
spin_unlock_bh(&op->bcm_tx_lock);
net/can/bcm.c
358
static void bcm_send_to_user(struct bcm_op *op, struct bcm_msg_head *head,
net/can/bcm.c
364
struct sock *sk = op->sk;
net/can/bcm.c
365
unsigned int datalen = head->nframes * op->cfsiz;
net/can/bcm.c
408
skb->tstamp = op->rx_stamp;
net/can/bcm.c
421
addr->can_ifindex = op->rx_ifindex;
net/can/bcm.c
433
static bool bcm_tx_set_expiry(struct bcm_op *op, struct hrtimer *hrt)
net/can/bcm.c
437
if (op->kt_ival1 && op->count)
net/can/bcm.c
438
ival = op->kt_ival1;
net/can/bcm.c
439
else if (op->kt_ival2)
net/can/bcm.c
440
ival = op->kt_ival2;
net/can/bcm.c
448
static void bcm_tx_start_timer(struct bcm_op *op)
net/can/bcm.c
450
if (bcm_tx_set_expiry(op, &op->timer))
net/can/bcm.c
451
hrtimer_start_expires(&op->timer, HRTIMER_MODE_ABS_SOFT);
net/can/bcm.c
457
struct bcm_op *op = container_of(hrtimer, struct bcm_op, timer);
net/can/bcm.c
460
if (op->kt_ival1 && (op->count > 0)) {
net/can/bcm.c
461
bcm_can_tx(op);
net/can/bcm.c
462
if (!op->count && (op->flags & TX_COUNTEVT)) {
net/can/bcm.c
467
msg_head.flags = op->flags;
net/can/bcm.c
468
msg_head.count = op->count;
net/can/bcm.c
469
msg_head.ival1 = op->ival1;
net/can/bcm.c
470
msg_head.ival2 = op->ival2;
net/can/bcm.c
471
msg_head.can_id = op->can_id;
net/can/bcm.c
474
bcm_send_to_user(op, &msg_head, NULL, 0);
net/can/bcm.c
477
} else if (op->kt_ival2) {
net/can/bcm.c
478
bcm_can_tx(op);
net/can/bcm.c
481
return bcm_tx_set_expiry(op, &op->timer) ?
net/can/bcm.c
488
static void bcm_rx_changed(struct bcm_op *op, struct canfd_frame *data)
net/can/bcm.c
493
op->frames_filtered++;
net/can/bcm.c
496
if (op->frames_filtered > ULONG_MAX/100)
net/can/bcm.c
497
op->frames_filtered = op->frames_abs = 0;
net/can/bcm.c
504
head.flags = op->flags;
net/can/bcm.c
505
head.count = op->count;
net/can/bcm.c
506
head.ival1 = op->ival1;
net/can/bcm.c
507
head.ival2 = op->ival2;
net/can/bcm.c
508
head.can_id = op->can_id;
net/can/bcm.c
511
bcm_send_to_user(op, &head, data, 1);
net/can/bcm.c
519
static void bcm_rx_update_and_send(struct bcm_op *op,
net/can/bcm.c
524
memcpy(lastdata, rxdata, op->cfsiz);
net/can/bcm.c
533
if (!op->kt_ival2) {
net/can/bcm.c
535
bcm_rx_changed(op, lastdata);
net/can/bcm.c
540
if (hrtimer_active(&op->thrtimer))
net/can/bcm.c
544
if (!op->kt_lastmsg)
net/can/bcm.c
548
if (ktime_us_delta(ktime_get(), op->kt_lastmsg) <
net/can/bcm.c
549
ktime_to_us(op->kt_ival2)) {
net/can/bcm.c
551
hrtimer_start(&op->thrtimer,
net/can/bcm.c
552
ktime_add(op->kt_lastmsg, op->kt_ival2),
net/can/bcm.c
559
bcm_rx_changed(op, lastdata);
net/can/bcm.c
560
op->kt_lastmsg = ktime_get();
net/can/bcm.c
567
static void bcm_rx_cmp_to_index(struct bcm_op *op, unsigned int index,
net/can/bcm.c
571
struct canfd_frame *cf = op->frames + op->cfsiz * index;
net/can/bcm.c
572
struct canfd_frame *lcf = op->last_frames + op->cfsiz * index;
net/can/bcm.c
582
bcm_rx_update_and_send(op, lcf, rxdata, traffic_flags);
net/can/bcm.c
590
bcm_rx_update_and_send(op, lcf, rxdata, traffic_flags);
net/can/bcm.c
595
if (op->flags & RX_CHECK_DLC) {
net/can/bcm.c
598
bcm_rx_update_and_send(op, lcf, rxdata, traffic_flags);
net/can/bcm.c
607
static void bcm_rx_starttimer(struct bcm_op *op)
net/can/bcm.c
609
if (op->flags & RX_NO_AUTOTIMER)
net/can/bcm.c
612
if (op->kt_ival1)
net/can/bcm.c
613
hrtimer_start(&op->timer, op->kt_ival1, HRTIMER_MODE_REL_SOFT);
net/can/bcm.c
619
struct bcm_op *op = container_of(hrtimer, struct bcm_op, timer);
net/can/bcm.c
623
if ((op->flags & RX_ANNOUNCE_RESUME) && op->last_frames) {
net/can/bcm.c
625
memset(op->last_frames, 0, op->nframes * op->cfsiz);
net/can/bcm.c
631
msg_head.flags = op->flags;
net/can/bcm.c
632
msg_head.count = op->count;
net/can/bcm.c
633
msg_head.ival1 = op->ival1;
net/can/bcm.c
634
msg_head.ival2 = op->ival2;
net/can/bcm.c
635
msg_head.can_id = op->can_id;
net/can/bcm.c
638
bcm_send_to_user(op, &msg_head, NULL, 0);
net/can/bcm.c
646
static inline int bcm_rx_do_flush(struct bcm_op *op, unsigned int index)
net/can/bcm.c
648
struct canfd_frame *lcf = op->last_frames + op->cfsiz * index;
net/can/bcm.c
650
if ((op->last_frames) && (lcf->flags & RX_THR)) {
net/can/bcm.c
651
bcm_rx_changed(op, lcf);
net/can/bcm.c
660
static int bcm_rx_thr_flush(struct bcm_op *op)
net/can/bcm.c
664
if (op->nframes > 1) {
net/can/bcm.c
668
for (i = 1; i < op->nframes; i++)
net/can/bcm.c
669
updated += bcm_rx_do_flush(op, i);
net/can/bcm.c
673
updated += bcm_rx_do_flush(op, 0);
net/can/bcm.c
685
struct bcm_op *op = container_of(hrtimer, struct bcm_op, thrtimer);
net/can/bcm.c
687
if (bcm_rx_thr_flush(op)) {
net/can/bcm.c
688
hrtimer_forward_now(hrtimer, op->kt_ival2);
net/can/bcm.c
692
op->kt_lastmsg = 0;
net/can/bcm.c
702
struct bcm_op *op = (struct bcm_op *)data;
net/can/bcm.c
707
if (op->can_id != rxframe->can_id)
net/can/bcm.c
711
if (op->flags & CAN_FD_FRAME) {
net/can/bcm.c
720
hrtimer_cancel(&op->timer);
net/can/bcm.c
723
op->rx_stamp = skb->tstamp;
net/can/bcm.c
725
op->rx_ifindex = skb->dev->ifindex;
net/can/bcm.c
727
op->frames_abs++;
net/can/bcm.c
729
if (op->flags & RX_RTR_FRAME) {
net/can/bcm.c
731
bcm_can_tx(op);
net/can/bcm.c
739
if (skb->sk == op->sk)
net/can/bcm.c
743
if (op->flags & RX_FILTER_ID) {
net/can/bcm.c
745
bcm_rx_update_and_send(op, op->last_frames, rxframe,
net/can/bcm.c
750
if (op->nframes == 1) {
net/can/bcm.c
752
bcm_rx_cmp_to_index(op, 0, rxframe, traffic_flags);
net/can/bcm.c
756
if (op->nframes > 1) {
net/can/bcm.c
765
for (i = 1; i < op->nframes; i++) {
net/can/bcm.c
766
if ((get_u64(op->frames, 0) & get_u64(rxframe, 0)) ==
net/can/bcm.c
767
(get_u64(op->frames, 0) &
net/can/bcm.c
768
get_u64(op->frames + op->cfsiz * i, 0))) {
net/can/bcm.c
769
bcm_rx_cmp_to_index(op, i, rxframe,
net/can/bcm.c
777
bcm_rx_starttimer(op);
net/can/bcm.c
786
struct bcm_op *op;
net/can/bcm.c
788
list_for_each_entry(op, ops, list) {
net/can/bcm.c
789
if ((op->can_id == mh->can_id) && (op->ifindex == ifindex) &&
net/can/bcm.c
790
(op->flags & CAN_FD_FRAME) == (mh->flags & CAN_FD_FRAME))
net/can/bcm.c
791
return op;
net/can/bcm.c
799
struct bcm_op *op = container_of(rcu_head, struct bcm_op, rcu);
net/can/bcm.c
801
if ((op->frames) && (op->frames != &op->sframe))
net/can/bcm.c
802
kfree(op->frames);
net/can/bcm.c
804
if ((op->last_frames) && (op->last_frames != &op->last_sframe))
net/can/bcm.c
805
kfree(op->last_frames);
net/can/bcm.c
807
kfree(op);
net/can/bcm.c
810
static void bcm_remove_op(struct bcm_op *op)
net/can/bcm.c
812
hrtimer_cancel(&op->timer);
net/can/bcm.c
813
hrtimer_cancel(&op->thrtimer);
net/can/bcm.c
815
call_rcu(&op->rcu, bcm_free_op_rcu);
net/can/bcm.c
818
static void bcm_rx_unreg(struct net_device *dev, struct bcm_op *op)
net/can/bcm.c
820
if (op->rx_reg_dev == dev) {
net/can/bcm.c
821
can_rx_unregister(dev_net(dev), dev, op->can_id,
net/can/bcm.c
822
REGMASK(op->can_id), bcm_rx_handler, op);
net/can/bcm.c
825
op->rx_reg_dev = NULL;
net/can/bcm.c
828
"mismatch %p %p\n", op->rx_reg_dev, dev);
net/can/bcm.c
837
struct bcm_op *op, *n;
net/can/bcm.c
839
list_for_each_entry_safe(op, n, ops, list) {
net/can/bcm.c
840
if ((op->can_id == mh->can_id) && (op->ifindex == ifindex) &&
net/can/bcm.c
841
(op->flags & CAN_FD_FRAME) == (mh->flags & CAN_FD_FRAME)) {
net/can/bcm.c
844
op->flags |= RX_NO_AUTOTIMER;
net/can/bcm.c
851
if (op->ifindex) {
net/can/bcm.c
857
if (op->rx_reg_dev) {
net/can/bcm.c
860
dev = dev_get_by_index(sock_net(op->sk),
net/can/bcm.c
861
op->ifindex);
net/can/bcm.c
863
bcm_rx_unreg(dev, op);
net/can/bcm.c
868
can_rx_unregister(sock_net(op->sk), NULL,
net/can/bcm.c
869
op->can_id,
net/can/bcm.c
870
REGMASK(op->can_id),
net/can/bcm.c
871
bcm_rx_handler, op);
net/can/bcm.c
873
list_del_rcu(&op->list);
net/can/bcm.c
874
bcm_remove_op(op);
net/can/bcm.c
888
struct bcm_op *op, *n;
net/can/bcm.c
890
list_for_each_entry_safe(op, n, ops, list) {
net/can/bcm.c
891
if ((op->can_id == mh->can_id) && (op->ifindex == ifindex) &&
net/can/bcm.c
892
(op->flags & CAN_FD_FRAME) == (mh->flags & CAN_FD_FRAME)) {
net/can/bcm.c
893
list_del_rcu(&op->list);
net/can/bcm.c
894
bcm_remove_op(op);
net/can/bcm.c
908
struct bcm_op *op = bcm_find_op(ops, msg_head, ifindex);
net/can/bcm.c
910
if (!op)
net/can/bcm.c
914
msg_head->flags = op->flags;
net/can/bcm.c
915
msg_head->count = op->count;
net/can/bcm.c
916
msg_head->ival1 = op->ival1;
net/can/bcm.c
917
msg_head->ival2 = op->ival2;
net/can/bcm.c
918
msg_head->nframes = op->nframes;
net/can/bcm.c
920
bcm_send_to_user(op, msg_head, op->frames, 0);
net/can/bcm.c
932
struct bcm_op *op;
net/can/bcm.c
950
op = bcm_find_op(&bo->tx_ops, msg_head, ifindex);
net/can/bcm.c
951
if (op) {
net/can/bcm.c
959
if (msg_head->nframes > op->nframes)
net/can/bcm.c
965
cf = op->frames + op->cfsiz * i;
net/can/bcm.c
966
err = memcpy_from_msg((u8 *)cf, msg, op->cfsiz);
net/can/bcm.c
968
if (op->flags & CAN_FD_FRAME) {
net/can/bcm.c
984
op->flags = msg_head->flags;
net/can/bcm.c
987
if (op->nframes != msg_head->nframes ||
net/can/bcm.c
988
op->flags & TX_RESET_MULTI_IDX ||
net/can/bcm.c
989
op->flags & SETTIMER) {
net/can/bcm.c
991
spin_lock_bh(&op->bcm_tx_lock);
net/can/bcm.c
993
if (op->nframes != msg_head->nframes ||
net/can/bcm.c
994
op->flags & TX_RESET_MULTI_IDX) {
net/can/bcm.c
996
op->nframes = msg_head->nframes;
net/can/bcm.c
998
op->currframe = 0;
net/can/gw.c
154
#define MODFUNC(func, op) static void func(struct canfd_frame *cf, \
net/can/gw.c
155
struct cf_mod *mod) { op ; }
net/ceph/auth_x.c
564
head->op = cpu_to_le16(CEPHX_GET_AUTH_SESSION_KEY);
net/ceph/auth_x.c
740
int op;
net/ceph/auth_x.c
758
ceph_decode_16_safe(&p, end, op, e_inval);
net/ceph/auth_x.c
760
dout("handle_reply op %d result %d\n", op, result);
net/ceph/auth_x.c
761
switch (op) {
net/ceph/auth_x_protocol.h
58
__le16 op;
net/ceph/auth_x_protocol.h
62
__le16 op;
net/ceph/ceph_strings.c
49
const char *ceph_osd_op_name(int op)
net/ceph/ceph_strings.c
51
switch (op) {
net/ceph/ceph_strings.c
52
#define GENERATE_CASE(op, opcode, str) case CEPH_OSD_OP_##op: return (str);
net/ceph/crush/mapper.c
1001
curstep->op ==
net/ceph/crush/mapper.c
1093
curstep->op, step);
net/ceph/crush/mapper.c
945
switch (curstep->op) {
net/ceph/crush/mapper.c
999
curstep->op ==
net/ceph/debugfs.c
167
__u16 op;
net/ceph/debugfs.c
169
op = le16_to_cpu(req->request->hdr.type);
net/ceph/debugfs.c
170
if (op == CEPH_MSG_STATFS)
net/ceph/debugfs.c
172
else if (op == CEPH_MSG_MON_GET_VERSION)
net/ceph/debugfs.c
225
struct ceph_osd_req_op *op = &req->r_ops[i];
net/ceph/debugfs.c
228
ceph_osd_op_name(op->op));
net/ceph/debugfs.c
229
if (op->op == CEPH_OSD_OP_WATCH)
net/ceph/debugfs.c
231
ceph_osd_watch_op_name(op->watch.op));
net/ceph/debugfs.c
232
else if (op->op == CEPH_OSD_OP_CALL)
net/ceph/debugfs.c
233
seq_printf(s, "-%s/%s", op->cls.class_name,
net/ceph/debugfs.c
234
op->cls.method_name);
net/ceph/osd_client.c
1034
ceph_osd_op_name(src->op));
net/ceph/osd_client.c
1040
dst->op = cpu_to_le16(src->op);
net/ceph/osd_client.c
1151
int __ceph_alloc_sparse_ext_map(struct ceph_osd_req_op *op, int cnt)
net/ceph/osd_client.c
1153
WARN_ON(op->op != CEPH_OSD_OP_SPARSE_READ);
net/ceph/osd_client.c
1155
op->extent.sparse_ext_cnt = cnt;
net/ceph/osd_client.c
1156
op->extent.sparse_ext = kmalloc_objs(*op->extent.sparse_ext, cnt,
net/ceph/osd_client.c
1158
if (!op->extent.sparse_ext)
net/ceph/osd_client.c
2045
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
2051
for (op = req->r_ops; op != &req->r_ops[req->r_num_ops]; op++) {
net/ceph/osd_client.c
2052
switch (op->op) {
net/ceph/osd_client.c
2056
WARN_ON(op->indata_len != op->extent.length);
net/ceph/osd_client.c
2058
&op->extent.osd_data);
net/ceph/osd_client.c
2062
WARN_ON(op->indata_len != op->xattr.name_len +
net/ceph/osd_client.c
2063
op->xattr.value_len);
net/ceph/osd_client.c
2065
&op->xattr.osd_data);
net/ceph/osd_client.c
2069
&op->notify_ack.request_data);
net/ceph/osd_client.c
2073
&op->copy_from.osd_data);
net/ceph/osd_client.c
2079
&op->raw_data_in);
net/ceph/osd_client.c
2084
&op->extent.osd_data);
net/ceph/osd_client.c
2088
&op->list_watchers.response_data);
net/ceph/osd_client.c
2093
WARN_ON(op->indata_len != op->cls.class_len +
net/ceph/osd_client.c
2094
op->cls.method_len +
net/ceph/osd_client.c
2095
op->cls.indata_len);
net/ceph/osd_client.c
2097
&op->cls.request_info);
net/ceph/osd_client.c
2100
&op->cls.request_data);
net/ceph/osd_client.c
2103
&op->cls.response_data);
net/ceph/osd_client.c
2107
&op->notify.request_data);
net/ceph/osd_client.c
2109
&op->notify.response_data);
net/ceph/osd_client.c
3070
WARN_ON(req->r_ops[0].op != CEPH_OSD_OP_NOTIFY ||
net/ceph/osd_client.c
3702
struct ceph_osd_op *op = p;
net/ceph/osd_client.c
3704
m->outdata_len[i] = le32_to_cpu(op->payload_len);
net/ceph/osd_client.c
3705
p += sizeof(*op);
net/ceph/osd_client.c
374
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
377
op = &osd_req->r_ops[which];
net/ceph/osd_client.c
379
switch (op->op) {
net/ceph/osd_client.c
384
kfree(op->extent.sparse_ext);
net/ceph/osd_client.c
385
ceph_osd_data_release(&op->extent.osd_data);
net/ceph/osd_client.c
388
ceph_osd_data_release(&op->cls.request_info);
net/ceph/osd_client.c
389
ceph_osd_data_release(&op->cls.request_data);
net/ceph/osd_client.c
390
ceph_osd_data_release(&op->cls.response_data);
net/ceph/osd_client.c
394
ceph_osd_data_release(&op->xattr.osd_data);
net/ceph/osd_client.c
397
ceph_osd_data_release(&op->raw_data_in);
net/ceph/osd_client.c
400
ceph_osd_data_release(&op->notify_ack.request_data);
net/ceph/osd_client.c
403
ceph_osd_data_release(&op->notify.request_data);
net/ceph/osd_client.c
404
ceph_osd_data_release(&op->notify.response_data);
net/ceph/osd_client.c
407
ceph_osd_data_release(&op->list_watchers.response_data);
net/ceph/osd_client.c
410
ceph_osd_data_release(&op->copy_from.osd_data);
net/ceph/osd_client.c
4304
u8 op;
net/ceph/osd_client.c
4328
ceph_decode_8_safe(&p, end, m->op, e_inval);
net/ceph/osd_client.c
4537
switch (m.op) {
net/ceph/osd_client.c
4545
pr_err("%s osd%d unknown op %d\n", __func__, osd->o_osd, m.op);
net/ceph/osd_client.c
4848
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
4852
op = osd_req_op_init(req, which, CEPH_OSD_OP_NOTIFY_ACK, 0);
net/ceph/osd_client.c
4871
ceph_osd_data_pagelist_init(&op->notify_ack.request_data, pl);
net/ceph/osd_client.c
4872
op->indata_len = pl->length;
net/ceph/osd_client.c
5308
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
5316
op = osd_req_op_init(req, 0, CEPH_OSD_OP_COPY_FROM2,
net/ceph/osd_client.c
5318
op->copy_from.snapid = src_snapid;
net/ceph/osd_client.c
5319
op->copy_from.src_version = src_version;
net/ceph/osd_client.c
5320
op->copy_from.flags = copy_from_flags;
net/ceph/osd_client.c
5321
op->copy_from.src_fadvise_flags = src_fadvise_flags;
net/ceph/osd_client.c
5329
op->indata_len = PAGE_SIZE - (end - p);
net/ceph/osd_client.c
5331
ceph_osd_data_pages_init(&op->copy_from.osd_data, pages,
net/ceph/osd_client.c
5332
op->indata_len, 0, false, true);
net/ceph/osd_client.c
5396
struct ceph_osd_req_op *op = &req->r_ops[i];
net/ceph/osd_client.c
5398
if (op->op == CEPH_OSD_OP_SPARSE_READ)
net/ceph/osd_client.c
5399
len += op->extent.length;
net/ceph/osd_client.c
5722
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
5737
op = &req->r_ops[o->o_sparse_op_idx];
net/ceph/osd_client.c
5739
WARN_ON_ONCE(op->extent.sparse_ext);
net/ceph/osd_client.c
5742
op->extent.sparse_ext = sr->sr_extent;
net/ceph/osd_client.c
5744
op->extent.sparse_ext_cnt = sr->sr_count;
net/ceph/osd_client.c
5747
__func__, o->o_osd, op->extent.sparse_ext_cnt, cursor->resid);
net/ceph/osd_client.c
5749
end = ceph_sparse_ext_map_end(op);
net/ceph/osd_client.c
5758
op = &req->r_ops[o->o_sparse_op_idx];
net/ceph/osd_client.c
5759
if (op->op == CEPH_OSD_OP_SPARSE_READ)
net/ceph/osd_client.c
5768
sr->sr_req_off = op->extent.offset;
net/ceph/osd_client.c
5769
sr->sr_req_len = op->extent.length;
net/ceph/osd_client.c
5775
sr->sr_ext_len = op->extent.sparse_ext_cnt;
net/ceph/osd_client.c
5776
op->extent.sparse_ext_cnt = 0;
net/ceph/osd_client.c
5777
sr->sr_extent = op->extent.sparse_ext;
net/ceph/osd_client.c
5778
op->extent.sparse_ext = NULL;
net/ceph/osd_client.c
644
#define GENERATE_CASE(op, opcode, str) case CEPH_OSD_OP_##op: return true;
net/ceph/osd_client.c
656
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
661
for (op = req->r_ops; op != &req->r_ops[req->r_num_ops]; op++) {
net/ceph/osd_client.c
662
switch (op->op) {
net/ceph/osd_client.c
692
WARN_ON(!osd_req_opcode_valid(op->op));
net/ceph/osd_client.c
721
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
726
op = &osd_req->r_ops[which];
net/ceph/osd_client.c
727
memset(op, 0, sizeof (*op));
net/ceph/osd_client.c
728
op->op = opcode;
net/ceph/osd_client.c
729
op->flags = flags;
net/ceph/osd_client.c
731
return op;
net/ceph/osd_client.c
740
struct ceph_osd_req_op *op = osd_req_op_init(osd_req, which,
net/ceph/osd_client.c
748
op->extent.offset = offset;
net/ceph/osd_client.c
749
op->extent.length = length;
net/ceph/osd_client.c
750
op->extent.truncate_size = truncate_size;
net/ceph/osd_client.c
751
op->extent.truncate_seq = truncate_seq;
net/ceph/osd_client.c
755
op->indata_len = payload_len;
net/ceph/osd_client.c
762
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
766
op = &osd_req->r_ops[which];
net/ceph/osd_client.c
767
previous = op->extent.length;
net/ceph/osd_client.c
773
op->extent.length = length;
net/ceph/osd_client.c
774
if (op->op == CEPH_OSD_OP_WRITE || op->op == CEPH_OSD_OP_WRITEFULL)
net/ceph/osd_client.c
775
op->indata_len -= previous - length;
net/ceph/osd_client.c
782
struct ceph_osd_req_op *op, *prev_op;
net/ceph/osd_client.c
787
op = osd_req_op_init(osd_req, which + 1, prev_op->op, prev_op->flags);
net/ceph/osd_client.c
789
op->indata_len = prev_op->indata_len;
net/ceph/osd_client.c
790
op->outdata_len = prev_op->outdata_len;
net/ceph/osd_client.c
791
op->extent = prev_op->extent;
net/ceph/osd_client.c
793
op->extent.offset += offset_inc;
net/ceph/osd_client.c
794
op->extent.length -= offset_inc;
net/ceph/osd_client.c
796
if (op->op == CEPH_OSD_OP_WRITE || op->op == CEPH_OSD_OP_WRITEFULL)
net/ceph/osd_client.c
797
op->indata_len -= offset_inc;
net/ceph/osd_client.c
804
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
810
op = osd_req_op_init(osd_req, which, CEPH_OSD_OP_CALL, 0);
net/ceph/osd_client.c
816
op->cls.class_name = class;
net/ceph/osd_client.c
819
op->cls.class_len = size;
net/ceph/osd_client.c
825
op->cls.method_name = method;
net/ceph/osd_client.c
828
op->cls.method_len = size;
net/ceph/osd_client.c
835
op->indata_len = payload_len;
net/ceph/osd_client.c
848
struct ceph_osd_req_op *op = osd_req_op_init(osd_req, which,
net/ceph/osd_client.c
861
op->xattr.name_len = payload_len;
net/ceph/osd_client.c
866
op->xattr.value_len = size;
net/ceph/osd_client.c
872
op->xattr.cmp_op = cmp_op;
net/ceph/osd_client.c
873
op->xattr.cmp_mode = cmp_mode;
net/ceph/osd_client.c
875
ceph_osd_data_pagelist_init(&op->xattr.osd_data, pagelist);
net/ceph/osd_client.c
876
op->indata_len = payload_len;
net/ceph/osd_client.c
891
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
893
op = osd_req_op_init(req, which, CEPH_OSD_OP_WATCH, 0);
net/ceph/osd_client.c
894
op->watch.cookie = cookie;
net/ceph/osd_client.c
895
op->watch.op = watch_opcode;
net/ceph/osd_client.c
896
op->watch.gen = gen;
net/ceph/osd_client.c
906
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
908
op = osd_req_op_init(req, which, CEPH_OSD_OP_NOTIFY, 0);
net/ceph/osd_client.c
909
op->notify.cookie = cookie;
net/ceph/osd_client.c
911
ceph_osd_data_pagelist_init(&op->notify.request_data, request_pl);
net/ceph/osd_client.c
912
op->indata_len = request_pl->length;
net/ceph/osd_client.c
924
struct ceph_osd_req_op *op;
net/ceph/osd_client.c
926
op = osd_req_op_init(osd_req, which, CEPH_OSD_OP_SETALLOCHINT, 0);
net/ceph/osd_client.c
927
op->alloc_hint.expected_object_size = expected_object_size;
net/ceph/osd_client.c
928
op->alloc_hint.expected_write_size = expected_write_size;
net/ceph/osd_client.c
929
op->alloc_hint.flags = flags;
net/ceph/osd_client.c
936
op->flags |= CEPH_OSD_OP_FLAG_FAILOK;
net/ceph/osd_client.c
969
switch (src->op) {
net/ceph/osd_client.c
993
dst->watch.op = src->watch.op;
net/ceph/osdmap.c
601
r->steps[j].op = ceph_decode_32(p);
net/core/filter.c
10682
case offsetof(struct bpf_sock_ops, op):
net/core/filter.c
10684
op),
net/core/filter.c
10686
offsetof(struct bpf_sock_ops_kern, op));
net/core/filter.c
12327
if (skops->op != BPF_SOCK_OPS_TSTAMP_SENDMSG_CB)
net/core/filter.c
5608
return bpf_sock->op <= BPF_SOCK_OPS_WRITE_HDR_OPT_CB;
net/core/filter.c
7730
static const u8 *bpf_search_tcp_opt(const u8 *op, const u8 *opend,
net/core/filter.c
7738
while (op < opend) {
net/core/filter.c
7739
kind = op[0];
net/core/filter.c
7745
op++;
net/core/filter.c
7749
if (opend - op < 2 || opend - op < op[1] || op[1] < 2)
net/core/filter.c
7756
kind_len = op[1];
net/core/filter.c
7759
return op;
net/core/filter.c
7764
if (!memcmp(&op[2], magic, magic_len))
net/core/filter.c
7765
return op;
net/core/filter.c
7768
op += kind_len;
net/core/filter.c
7778
const u8 *op, *opend, *magic, *search = search_res;
net/core/filter.c
7813
ret = bpf_sock_ops_get_syn(bpf_sock, TCP_BPF_SYN, &op);
net/core/filter.c
7817
opend = op + ret;
net/core/filter.c
7818
op += sizeof(struct tcphdr);
net/core/filter.c
7821
bpf_sock->op == BPF_SOCK_OPS_HDR_OPT_LEN_CB)
net/core/filter.c
7826
op = bpf_sock->skb->data + sizeof(struct tcphdr);
net/core/filter.c
7829
op = bpf_search_tcp_opt(op, opend, search_kind, magic, magic_len,
net/core/filter.c
7831
if (IS_ERR(op))
net/core/filter.c
7832
return PTR_ERR(op);
net/core/filter.c
7834
copy_len = op[1];
net/core/filter.c
7841
memcpy(search_res, op, copy_len);
net/core/filter.c
7859
const u8 *op, *new_op, *magic = NULL;
net/core/filter.c
7863
if (bpf_sock->op != BPF_SOCK_OPS_WRITE_HDR_OPT_CB)
net/core/filter.c
7897
op = skb->data + sizeof(struct tcphdr);
net/core/filter.c
7900
op = bpf_search_tcp_opt(op, opend, new_kind, magic, magic_len,
net/core/filter.c
7902
if (!IS_ERR(op))
net/core/filter.c
7905
if (PTR_ERR(op) != -ENOMSG)
net/core/filter.c
7906
return PTR_ERR(op);
net/core/filter.c
7936
if (bpf_sock->op != BPF_SOCK_OPS_HDR_OPT_LEN_CB)
net/core/flow_dissector.c
620
key_arp->op = ntohs(arp->ar_op) & 0xff;
net/core/skbuff.c
5662
int op;
net/core/skbuff.c
5666
op = BPF_SOCK_OPS_TSTAMP_SCHED_CB;
net/core/skbuff.c
5670
op = BPF_SOCK_OPS_TSTAMP_SND_HW_CB;
net/core/skbuff.c
5673
op = BPF_SOCK_OPS_TSTAMP_SND_SW_CB;
net/core/skbuff.c
5677
op = BPF_SOCK_OPS_TSTAMP_ACK_CB;
net/core/skbuff.c
5683
bpf_skops_tx_timestamping(sk, skb, op);
net/core/sock.c
951
void bpf_skops_tx_timestamping(struct sock *sk, struct sk_buff *skb, int op)
net/core/sock.c
956
sock_ops.op = op;
net/core/sock_map.c
523
return ops->op == BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB ||
net/core/sock_map.c
524
ops->op == BPF_SOCK_OPS_ACTIVE_ESTABLISHED_CB ||
net/core/sock_map.c
525
ops->op == BPF_SOCK_OPS_TCP_LISTEN_CB;
net/ieee802154/nl802154.c
521
#define CMD(op, n) \
net/ieee802154/nl802154.c
523
if (rdev->ops->op) { \
net/ipv4/inet_diag.c
468
const struct inet_diag_bc_op *op = bc;
net/ipv4/inet_diag.c
470
switch (op->code) {
net/ipv4/inet_diag.c
477
yes = entry->sport == op[1].no;
net/ipv4/inet_diag.c
480
yes = entry->sport >= op[1].no;
net/ipv4/inet_diag.c
483
yes = entry->sport <= op[1].no;
net/ipv4/inet_diag.c
486
yes = entry->dport == op[1].no;
net/ipv4/inet_diag.c
489
yes = entry->dport >= op[1].no;
net/ipv4/inet_diag.c
492
yes = entry->dport <= op[1].no;
net/ipv4/inet_diag.c
502
cond = (const struct inet_diag_hostcond *)(op + 1);
net/ipv4/inet_diag.c
504
cond->port != (op->code == INET_DIAG_BC_S_COND ?
net/ipv4/inet_diag.c
510
if (op->code == INET_DIAG_BC_S_COND)
net/ipv4/inet_diag.c
541
ifindex = *((const u32 *)(op + 1));
net/ipv4/inet_diag.c
549
cond = (struct inet_diag_markcond *)(op + 1);
net/ipv4/inet_diag.c
558
cgroup_id = get_unaligned((const u64 *)(op + 1));
net/ipv4/inet_diag.c
567
len -= op->yes;
net/ipv4/inet_diag.c
568
bc += op->yes;
net/ipv4/inet_diag.c
570
len -= op->no;
net/ipv4/inet_diag.c
571
bc += op->no;
net/ipv4/inet_diag.c
633
const struct inet_diag_bc_op *op = bc;
net/ipv4/inet_diag.c
639
if (op->yes < 4 || op->yes & 3)
net/ipv4/inet_diag.c
641
len -= op->yes;
net/ipv4/inet_diag.c
642
bc += op->yes;
net/ipv4/inet_diag.c
648
static bool valid_devcond(const struct inet_diag_bc_op *op, int len,
net/ipv4/inet_diag.c
659
static bool valid_hostcond(const struct inet_diag_bc_op *op, int len,
net/ipv4/inet_diag.c
669
cond = (struct inet_diag_hostcond *)(op + 1);
net/ipv4/inet_diag.c
697
static bool valid_port_comparison(const struct inet_diag_bc_op *op,
net/ipv4/inet_diag.c
707
static bool valid_markcond(const struct inet_diag_bc_op *op, int len,
net/ipv4/inet_diag.c
715
static bool valid_cgroupcond(const struct inet_diag_bc_op *op, int len,
net/ipv4/inet_diag.c
743
const struct inet_diag_bc_op *op = bc;
net/ipv4/inet_diag.c
745
switch (op->code) {
net/ipv4/inet_diag.c
788
if (op->code != INET_DIAG_BC_NOP) {
net/ipv4/inet_diag.c
789
if (op->no < min_len || op->no > len + 4 || op->no & 3)
net/ipv4/inet_diag.c
791
if (op->no < len &&
net/ipv4/inet_diag.c
792
!valid_cc(bytecode, bytecode_len, len - op->no))
net/ipv4/inet_diag.c
796
if (op->yes < min_len || op->yes > len + 4 || op->yes & 3)
net/ipv4/inet_diag.c
798
bc += op->yes;
net/ipv4/inet_diag.c
799
len -= op->yes;
net/ipv4/ipconfig.c
1061
if (b->op != BOOTP_REPLY ||
net/ipv4/ipconfig.c
1064
d->dev->name, b->op, b->xid);
net/ipv4/ipconfig.c
630
u8 op; /* 1=request, 2=reply */
net/ipv4/ipconfig.c
854
b->op = BOOTP_REQUEST;
net/ipv4/tcp_input.c
173
sock_ops.op = BPF_SOCK_OPS_PARSE_HDR_OPT_CB;
net/ipv4/tcp_input.c
190
sock_ops.op = bpf_op;
net/ipv4/tcp_ipv4.c
3367
if (seq->op == &bpf_iter_tcp_seq_ops)
net/ipv4/tcp_output.c
490
sock_ops.op = BPF_SOCK_OPS_HDR_OPT_LEN_CB;
net/ipv4/tcp_output.c
549
sock_ops.op = BPF_SOCK_OPS_WRITE_HDR_OPT_CB;
net/ipv4/udp.c
3323
if (seq->op == &bpf_iter_udp_seq_ops)
net/ipv4/udp.c
3742
if (seq->op == &bpf_iter_udp_seq_ops)
net/ipv6/icmp.c
267
u8 _optval, *op;
net/ipv6/icmp.c
270
op = skb_header_pointer(skb, offset, sizeof(_optval), &_optval);
net/ipv6/icmp.c
271
if (!op)
net/ipv6/icmp.c
273
return (*op & 0xC0) == 0x80;
net/netfilter/ipset/ip_set_core.c
2244
unsigned int *op;
net/netfilter/ipset/ip_set_core.c
2264
op = data;
net/netfilter/ipset/ip_set_core.c
2266
if (*op < IP_SET_OP_VERSION) {
net/netfilter/ipset/ip_set_core.c
2281
switch (*op) {
net/netfilter/ipset/ip_set_core.c
589
ip_set_match_counter(u64 counter, u64 match, u8 op)
net/netfilter/ipset/ip_set_core.c
591
switch (op) {
net/netfilter/nf_conntrack_seqadj.c
137
unsigned char *op = skb->data + optoff;
net/netfilter/nf_conntrack_seqadj.c
139
switch (op[0]) {
net/netfilter/nf_conntrack_seqadj.c
148
optoff + op[1] > optend ||
net/netfilter/nf_conntrack_seqadj.c
149
op[1] < 2)
net/netfilter/nf_conntrack_seqadj.c
151
if (op[0] == TCPOPT_SACK &&
net/netfilter/nf_conntrack_seqadj.c
152
op[1] >= 2+TCPOLEN_SACK_PERBLOCK &&
net/netfilter/nf_conntrack_seqadj.c
153
((op[1] - 2) % TCPOLEN_SACK_PERBLOCK) == 0)
net/netfilter/nf_conntrack_seqadj.c
155
optoff+op[1],
net/netfilter/nf_conntrack_seqadj.c
157
optoff += op[1];
net/netfilter/nf_log_syslog.c
245
const u8 *op;
net/netfilter/nf_log_syslog.c
247
op = skb_header_pointer(skb, offset + sizeof(struct tcphdr),
net/netfilter/nf_log_syslog.c
249
if (!op) {
net/netfilter/nf_log_syslog.c
257
nf_log_buf_add(m, "%02X", op[i]);
net/netfilter/nf_log_syslog.c
346
const unsigned char *op;
net/netfilter/nf_log_syslog.c
350
op = skb_header_pointer(skb, iphoff + sizeof(_iph),
net/netfilter/nf_log_syslog.c
352
if (!op) {
net/netfilter/nf_log_syslog.c
360
nf_log_buf_add(m, "%02X", op[i]);
net/netfilter/nf_synproxy_core.c
204
unsigned char *op = skb->data + optoff;
net/netfilter/nf_synproxy_core.c
206
switch (op[0]) {
net/netfilter/nf_synproxy_core.c
214
optoff + op[1] > optend ||
net/netfilter/nf_synproxy_core.c
215
op[1] < 2)
net/netfilter/nf_synproxy_core.c
217
if (op[0] == TCPOPT_TIMESTAMP &&
net/netfilter/nf_synproxy_core.c
218
op[1] == TCPOLEN_TIMESTAMP) {
net/netfilter/nf_synproxy_core.c
220
ptr = (__be32 *)&op[2];
net/netfilter/nf_synproxy_core.c
225
ptr = (__be32 *)&op[6];
net/netfilter/nf_synproxy_core.c
234
optoff += op[1];
net/netfilter/nf_tables_api.c
10715
const struct nft_trans *trans, u32 op)
net/netfilter/nf_tables_api.c
10728
if (!adp->op || adp->op > op)
net/netfilter/nf_tables_api.c
10729
adp->op = op;
net/netfilter/nf_tables_api.c
10743
nft2audit_op[adp->op], GFP_KERNEL);
net/netfilter/nf_tables_api.c
79
int op;
net/netfilter/nft_bitwise.c
102
if (priv->op == NFT_BITWISE_LSHIFT) {
net/netfilter/nft_bitwise.c
106
if (priv->op == NFT_BITWISE_RSHIFT) {
net/netfilter/nft_bitwise.c
113
if (priv->op == NFT_BITWISE_AND) {
net/netfilter/nft_bitwise.c
117
if (priv->op == NFT_BITWISE_OR) {
net/netfilter/nft_bitwise.c
121
if (priv->op == NFT_BITWISE_XOR) {
net/netfilter/nft_bitwise.c
22
enum nft_bitwise_ops op:8;
net/netfilter/nft_bitwise.c
268
priv->op = ntohl(nla_get_be32(tb[NFTA_BITWISE_OP]));
net/netfilter/nft_bitwise.c
269
switch (priv->op) {
net/netfilter/nft_bitwise.c
281
priv->op = NFT_BITWISE_MASK_XOR;
net/netfilter/nft_bitwise.c
284
switch(priv->op) {
net/netfilter/nft_bitwise.c
352
if (nla_put_be32(skb, NFTA_BITWISE_OP, htonl(priv->op)))
net/netfilter/nft_bitwise.c
355
switch (priv->op) {
net/netfilter/nft_bitwise.c
382
if (priv->op != NFT_BITWISE_MASK_XOR)
net/netfilter/nft_bitwise.c
414
priv->op == bitwise->op &&
net/netfilter/nft_bitwise.c
98
if (priv->op == NFT_BITWISE_MASK_XOR) {
net/netfilter/nft_byteorder.c
112
priv->op = ntohl(nla_get_be32(tb[NFTA_BYTEORDER_OP]));
net/netfilter/nft_byteorder.c
113
switch (priv->op) {
net/netfilter/nft_byteorder.c
161
if (nla_put_be32(skb, NFTA_BYTEORDER_OP, htonl(priv->op)))
net/netfilter/nft_byteorder.c
21
enum nft_byteorder_ops op:8;
net/netfilter/nft_byteorder.c
44
switch (priv->op) {
net/netfilter/nft_byteorder.c
63
switch (priv->op) {
net/netfilter/nft_byteorder.c
75
switch (priv->op) {
net/netfilter/nft_cmp.c
102
if (nla_put_be32(skb, NFTA_CMP_OP, htonl(priv->op)))
net/netfilter/nft_cmp.c
149
if (priv->op != NFT_CMP_EQ || priv->len > reg->len)
net/netfilter/nft_cmp.c
23
enum nft_cmp_ops op:8;
net/netfilter/nft_cmp.c
251
.op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ,
net/netfilter/nft_cmp.c
261
enum nft_cmp_ops op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ;
net/netfilter/nft_cmp.c
266
if (nla_put_be32(skb, NFTA_CMP_OP, htonl(op)))
net/netfilter/nft_cmp.c
34
switch (priv->op) {
net/netfilter/nft_cmp.c
346
.op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ,
net/netfilter/nft_cmp.c
356
enum nft_cmp_ops op = priv->inv ? NFT_CMP_NEQ : NFT_CMP_EQ;
net/netfilter/nft_cmp.c
360
if (nla_put_be32(skb, NFTA_CMP_OP, htonl(op)))
net/netfilter/nft_cmp.c
391
enum nft_cmp_ops op;
net/netfilter/nft_cmp.c
400
op = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
net/netfilter/nft_cmp.c
401
switch (op) {
net/netfilter/nft_cmp.c
419
if (op == NFT_CMP_EQ || op == NFT_CMP_NEQ) {
net/netfilter/nft_cmp.c
90
priv->op = ntohl(nla_get_be32(tb[NFTA_CMP_OP]));
net/netfilter/nft_dynset.c
104
if (priv->op == NFT_DYNSET_OP_UPDATE &&
net/netfilter/nft_dynset.c
18
enum nft_dynset_ops op:8;
net/netfilter/nft_dynset.c
213
priv->op = ntohl(nla_get_be32(tb[NFTA_DYNSET_OP]));
net/netfilter/nft_dynset.c
214
if (priv->op > NFT_DYNSET_OP_DELETE)
net/netfilter/nft_dynset.c
387
if (nla_put_be32(skb, NFTA_DYNSET_OP, htonl(priv->op)))
net/netfilter/nft_dynset.c
97
if (priv->op == NFT_DYNSET_OP_DELETE) {
net/netfilter/nft_exthdr.c
23
u8 op;
net/netfilter/nft_exthdr.c
503
u32 offset, len, flags = 0, op = NFT_EXTHDR_OP_IPV6;
net/netfilter/nft_exthdr.c
530
err = nft_parse_u32_check(tb[NFTA_EXTHDR_OP], U8_MAX, &op);
net/netfilter/nft_exthdr.c
539
priv->op = op;
net/netfilter/nft_exthdr.c
551
u32 offset, len, flags = 0, op = NFT_EXTHDR_OP_IPV6;
net/netfilter/nft_exthdr.c
581
err = nft_parse_u32_check(tb[NFTA_EXTHDR_OP], U8_MAX, &op);
net/netfilter/nft_exthdr.c
589
priv->op = op;
net/netfilter/nft_exthdr.c
612
priv->op = NFT_EXTHDR_OP_TCPOPT;
net/netfilter/nft_exthdr.c
667
if (nla_put_be32(skb, NFTA_EXTHDR_OP, htonl(priv->op)))
net/netfilter/nft_exthdr.c
718
priv->op != exthdr->op ||
net/netfilter/nft_exthdr.c
801
u32 op;
net/netfilter/nft_exthdr.c
809
op = ntohl(nla_get_be32(tb[NFTA_EXTHDR_OP]));
net/netfilter/nft_exthdr.c
810
switch (op) {
net/netfilter/nft_range.c
104
priv->op = op;
net/netfilter/nft_range.c
121
if (nla_put_be32(skb, NFTA_RANGE_OP, htonl(priv->op)))
net/netfilter/nft_range.c
20
enum nft_range_ops op:8;
net/netfilter/nft_range.c
31
switch (priv->op) {
net/netfilter/nft_range.c
63
u32 op;
net/netfilter/nft_range.c
91
err = nft_parse_u32_check(tb[NFTA_RANGE_OP], U8_MAX, &op);
net/netfilter/nft_range.c
95
switch (op) {
net/netfilter/xt_dccp.c
41
const unsigned char *op;
net/netfilter/xt_dccp.c
53
op = skb_header_pointer(skb, protoff + optoff, optlen, dccp_optbuf);
net/netfilter/xt_dccp.c
54
if (op == NULL) {
net/netfilter/xt_dccp.c
60
if (op[i] == option) {
net/netfilter/xt_dccp.c
65
if (op[i] < 2 || i == optlen - 1)
net/netfilter/xt_dccp.c
68
i += op[i + 1] ? : 1;
net/netfilter/xt_set.c
173
info->packets.op, info->bytes.op);
net/netfilter/xt_set.c
175
if (info->packets.op != IPSET_COUNTER_NONE ||
net/netfilter/xt_set.c
176
info->bytes.op != IPSET_COUNTER_NONE)
net/netfilter/xt_set.c
196
info->packets.op, info->bytes.op);
net/netfilter/xt_set.c
198
if (info->packets.op != IPSET_COUNTER_NONE ||
net/netfilter/xt_set.c
199
info->bytes.op != IPSET_COUNTER_NONE)
net/netfilter/xt_tcpmss.c
31
const u_int8_t *op;
net/netfilter/xt_tcpmss.c
49
op = skb_header_pointer(skb, par->thoff + sizeof(*th), optlen, _opt);
net/netfilter/xt_tcpmss.c
50
if (op == NULL)
net/netfilter/xt_tcpmss.c
54
if (op[i] == TCPOPT_MSS
net/netfilter/xt_tcpmss.c
56
&& op[i+1] == TCPOLEN_MSS) {
net/netfilter/xt_tcpmss.c
59
mssval = (op[i+2] << 8) | op[i+3];
net/netfilter/xt_tcpmss.c
64
if (op[i] < 2 || i == optlen - 1)
net/netfilter/xt_tcpmss.c
67
i += op[i+1] ? : 1;
net/netfilter/xt_tcpudp.c
43
const u_int8_t *op;
net/netfilter/xt_tcpudp.c
53
op = skb_header_pointer(skb, protoff + sizeof(struct tcphdr),
net/netfilter/xt_tcpudp.c
55
if (op == NULL) {
net/netfilter/xt_tcpudp.c
61
if (op[i] == option) return !invert;
net/netfilter/xt_tcpudp.c
62
if (op[i] < 2 || i == optlen - 1)
net/netfilter/xt_tcpudp.c
65
i += op[i + 1] ? : 1;
net/netlink/genetlink.c
1019
const struct genl_split_ops *ops = &dump_info->op;
net/netlink/genetlink.c
1034
const struct genl_split_ops *ops = &dump_info->op;
net/netlink/genetlink.c
107
struct genl_ops *op)
net/netlink/genetlink.c
111
if (op->policy || op->cmd < family->resv_start_op)
net/netlink/genetlink.c
114
op->policy = genl_policy_reject_all;
net/netlink/genetlink.c
115
op->maxattr = 1;
net/netlink/genetlink.c
1162
struct genl_split_ops op;
net/netlink/genetlink.c
1179
if (genl_get_cmd(hdr->cmd, flags, family, &op))
net/netlink/genetlink.c
1182
if ((op.flags & GENL_ADMIN_PERM) &&
net/netlink/genetlink.c
1186
if ((op.flags & GENL_UNS_ADMIN_PERM) &&
net/netlink/genetlink.c
1192
&op, hdrlen, net);
net/netlink/genetlink.c
1195
&op, hdrlen, net);
net/netlink/genetlink.c
120
struct genl_split_ops *op)
net/netlink/genetlink.c
122
if (op->policy)
net/netlink/genetlink.c
125
op->policy = genl_policy_reject_all;
net/netlink/genetlink.c
126
op->maxattr = 1;
net/netlink/genetlink.c
1516
u32 op;
net/netlink/genetlink.c
1563
ctx->op = nla_get_u32(tb[CTRL_ATTR_OP]);
net/netlink/genetlink.c
1565
err = genl_get_cmd_both(ctx->op, rt, &doit, &dump);
net/netlink/genetlink.c
157
unsigned int i, struct genl_ops *op)
net/netlink/genetlink.c
159
*op = family->ops[i];
net/netlink/genetlink.c
161
if (!op->maxattr)
net/netlink/genetlink.c
162
op->maxattr = family->maxattr;
net/netlink/genetlink.c
163
if (!op->policy)
net/netlink/genetlink.c
164
op->policy = family->policy;
net/netlink/genetlink.c
166
genl_op_fill_in_reject_policy(family, op);
net/netlink/genetlink.c
170
struct genl_ops *op)
net/netlink/genetlink.c
1710
if (WARN_ON(genl_get_cmd_both(ctx->op, ctx->rt,
net/netlink/genetlink.c
176
genl_op_from_full(family, i, op);
net/netlink/genetlink.c
184
unsigned int i, struct genl_ops *op)
net/netlink/genetlink.c
186
memset(op, 0, sizeof(*op));
net/netlink/genetlink.c
187
op->doit = family->small_ops[i].doit;
net/netlink/genetlink.c
188
op->dumpit = family->small_ops[i].dumpit;
net/netlink/genetlink.c
189
op->cmd = family->small_ops[i].cmd;
net/netlink/genetlink.c
190
op->internal_flags = family->small_ops[i].internal_flags;
net/netlink/genetlink.c
191
op->flags = family->small_ops[i].flags;
net/netlink/genetlink.c
192
op->validate = family->small_ops[i].validate;
net/netlink/genetlink.c
194
op->maxattr = family->maxattr;
net/netlink/genetlink.c
195
op->policy = family->policy;
net/netlink/genetlink.c
197
genl_op_fill_in_reject_policy(family, op);
net/netlink/genetlink.c
201
struct genl_ops *op)
net/netlink/genetlink.c
207
genl_op_from_small(family, i, op);
net/netlink/genetlink.c
245
struct genl_split_ops *op)
net/netlink/genetlink.c
252
*op = family->split_ops[i];
net/netlink/genetlink.c
260
genl_cmd_full_to_split(struct genl_split_ops *op,
net/netlink/genetlink.c
266
memset(op, 0, sizeof(*op));
net/netlink/genetlink.c
271
op->start = full->start;
net/netlink/genetlink.c
272
op->dumpit = full->dumpit;
net/netlink/genetlink.c
273
op->done = full->done;
net/netlink/genetlink.c
275
op->pre_doit = family->pre_doit;
net/netlink/genetlink.c
276
op->doit = full->doit;
net/netlink/genetlink.c
277
op->post_doit = family->post_doit;
net/netlink/genetlink.c
282
op->policy = NULL;
net/netlink/genetlink.c
283
op->maxattr = 0;
net/netlink/genetlink.c
285
op->policy = full->policy;
net/netlink/genetlink.c
286
op->maxattr = full->maxattr;
net/netlink/genetlink.c
289
op->cmd = full->cmd;
net/netlink/genetlink.c
290
op->internal_flags = full->internal_flags;
net/netlink/genetlink.c
291
op->flags = full->flags;
net/netlink/genetlink.c
292
op->validate = full->validate;
net/netlink/genetlink.c
295
op->flags |= flags;
net/netlink/genetlink.c
303
struct genl_split_ops *op)
net/netlink/genetlink.c
313
return genl_cmd_full_to_split(op, family, &full, flags);
net/netlink/genetlink.c
315
err = genl_get_cmd_split(cmd, flags, family, op);
net/netlink/genetlink.c
317
memset(op, 0, sizeof(*op));
net/netlink/genetlink.c
354
struct genl_ops op;
net/netlink/genetlink.c
357
genl_op_from_full(family, iter->entry_idx, &op);
net/netlink/genetlink.c
360
&op);
net/netlink/genetlink.c
376
&op, GENL_CMD_CAP_DO);
net/netlink/genetlink.c
378
&op, GENL_CMD_CAP_DUMP);
net/netlink/genetlink.c
563
static bool genl_split_op_check(const struct genl_split_ops *op)
net/netlink/genetlink.c
565
if (WARN_ON(hweight8(op->flags & (GENL_CMD_CAP_DO |
net/netlink/genetlink.c
990
info->op = *ops;
net/nfc/nci/core.c
1427
const struct nci_driver_ops *op;
net/nfc/nci/core.c
1433
op = &ops[i];
net/nfc/nci/core.c
1434
if (op->opcode == opcode)
net/nfc/nci/core.c
1435
return op;
net/nfc/nci/core.c
1445
const struct nci_driver_ops *op;
net/nfc/nci/core.c
1447
op = ops_cmd_lookup(ops, n_ops, rsp_opcode);
net/nfc/nci/core.c
1448
if (!op || !op->rsp)
net/nfc/nci/core.c
1451
return op->rsp(ndev, skb);
net/nfc/nci/core.c
1458
const struct nci_driver_ops *op;
net/nfc/nci/core.c
1460
op = ops_cmd_lookup(ops, n_ops, ntf_opcode);
net/nfc/nci/core.c
1461
if (!op || !op->ntf)
net/nfc/nci/core.c
1464
return op->ntf(ndev, skb);
net/rds/ib.h
429
int rds_ib_xmit_rdma(struct rds_connection *conn, struct rm_rdma_op *op);
net/rds/ib.h
434
int rds_ib_xmit_atomic(struct rds_connection *conn, struct rm_atomic_op *op);
net/rds/ib_send.c
1000
prev->s_op = op;
net/rds/ib_send.c
1001
rds_message_addref(container_of(op, struct rds_message, rdma));
net/rds/ib_send.c
114
rds_ib_send_complete(container_of(op, struct rds_message, rdma),
net/rds/ib_send.c
117
if (op->op_write)
net/rds/ib_send.c
118
rds_stats_add(s_send_rdma_bytes, op->op_bytes);
net/rds/ib_send.c
120
rds_stats_add(s_recv_rdma_bytes, op->op_bytes);
net/rds/ib_send.c
124
struct rm_atomic_op *op,
net/rds/ib_send.c
128
if (op->op_mapped) {
net/rds/ib_send.c
129
ib_dma_unmap_sg(ic->i_cm_id->device, op->op_sg, 1,
net/rds/ib_send.c
131
op->op_mapped = 0;
net/rds/ib_send.c
134
rds_ib_send_complete(container_of(op, struct rds_message, atomic),
net/rds/ib_send.c
137
if (op->op_type == RDS_ATOMIC_TYPE_CSWP)
net/rds/ib_send.c
74
struct rm_data_op *op,
net/rds/ib_send.c
77
if (op->op_nents)
net/rds/ib_send.c
79
op->op_sg, op->op_nents,
net/rds/ib_send.c
792
int rds_ib_xmit_atomic(struct rds_connection *conn, struct rm_atomic_op *op)
net/rds/ib_send.c
813
if (op->op_type == RDS_ATOMIC_TYPE_CSWP) {
net/rds/ib_send.c
815
send->s_atomic_wr.compare_add = op->op_m_cswp.compare;
net/rds/ib_send.c
816
send->s_atomic_wr.swap = op->op_m_cswp.swap;
net/rds/ib_send.c
817
send->s_atomic_wr.compare_add_mask = op->op_m_cswp.compare_mask;
net/rds/ib_send.c
818
send->s_atomic_wr.swap_mask = op->op_m_cswp.swap_mask;
net/rds/ib_send.c
821
send->s_atomic_wr.compare_add = op->op_m_fadd.add;
net/rds/ib_send.c
823
send->s_atomic_wr.compare_add_mask = op->op_m_fadd.nocarry_mask;
net/rds/ib_send.c
827
nr_sig = rds_ib_set_wr_signal_state(ic, send, op->op_notify);
net/rds/ib_send.c
830
send->s_atomic_wr.remote_addr = op->op_remote_addr;
net/rds/ib_send.c
831
send->s_atomic_wr.rkey = op->op_rkey;
net/rds/ib_send.c
832
send->s_op = op;
net/rds/ib_send.c
836
ret = ib_dma_map_sg(ic->i_cm_id->device, op->op_sg, 1, DMA_FROM_DEVICE);
net/rds/ib_send.c
837
rdsdebug("ic %p mapping atomic op %p. mapped %d pg\n", ic, op, ret);
net/rds/ib_send.c
84
struct rm_rdma_op *op,
net/rds/ib_send.c
846
send->s_sge[0].addr = sg_dma_address(op->op_sg);
net/rds/ib_send.c
847
send->s_sge[0].length = sg_dma_len(op->op_sg);
net/rds/ib_send.c
850
rdsdebug("rva %Lx rpa %Lx len %u\n", op->op_remote_addr,
net/rds/ib_send.c
87
if (op->op_mapped) {
net/rds/ib_send.c
878
int rds_ib_xmit_rdma(struct rds_connection *conn, struct rm_rdma_op *op)
net/rds/ib_send.c
887
u64 remote_addr = op->op_remote_addr;
net/rds/ib_send.c
89
op->op_sg, op->op_nents,
net/rds/ib_send.c
897
u64 odp_addr = op->op_odp_addr;
net/rds/ib_send.c
90
op->op_write ? DMA_TO_DEVICE : DMA_FROM_DEVICE);
net/rds/ib_send.c
901
if (!op->op_odp_mr) {
net/rds/ib_send.c
902
if (!op->op_mapped) {
net/rds/ib_send.c
903
op->op_count =
net/rds/ib_send.c
904
ib_dma_map_sg(ic->i_cm_id->device, op->op_sg,
net/rds/ib_send.c
905
op->op_nents,
net/rds/ib_send.c
906
(op->op_write) ? DMA_TO_DEVICE :
net/rds/ib_send.c
908
rdsdebug("ic %p mapping op %p: %d\n", ic, op,
net/rds/ib_send.c
909
op->op_count);
net/rds/ib_send.c
91
op->op_mapped = 0;
net/rds/ib_send.c
910
if (op->op_count == 0) {
net/rds/ib_send.c
915
op->op_mapped = 1;
net/rds/ib_send.c
918
op->op_count = op->op_nents;
net/rds/ib_send.c
919
odp_lkey = rds_ib_get_lkey(op->op_odp_mr->r_trans_private);
net/rds/ib_send.c
926
i = DIV_ROUND_UP(op->op_count, max_sge);
net/rds/ib_send.c
939
scat = &op->op_sg[0];
net/rds/ib_send.c
941
num_sge = op->op_count;
net/rds/ib_send.c
943
for (i = 0; i < work_alloc && scat != &op->op_sg[op->op_count]; i++) {
net/rds/ib_send.c
948
if (!op->op_notify)
net/rds/ib_send.c
950
op->op_notify);
net/rds/ib_send.c
952
send->s_wr.opcode = op->op_write ? IB_WR_RDMA_WRITE : IB_WR_RDMA_READ;
net/rds/ib_send.c
954
send->s_rdma_wr.rkey = op->op_rkey;
net/rds/ib_send.c
969
scat != &op->op_sg[op->op_count]; j++) {
net/rds/ib_send.c
971
if (!op->op_odp_mr) {
net/rds/ib_send.c
999
if (scat == &op->op_sg[op->op_count]) {
net/rds/rdma.c
614
struct rm_rdma_op *op = &rm->rdma;
net/rds/rdma.c
661
op->op_write = !!(args->flags & RDS_RDMA_READWRITE);
net/rds/rdma.c
662
op->op_fence = !!(args->flags & RDS_RDMA_FENCE);
net/rds/rdma.c
663
op->op_notify = !!(args->flags & RDS_RDMA_NOTIFY_ME);
net/rds/rdma.c
664
op->op_silent = !!(args->flags & RDS_RDMA_SILENT);
net/rds/rdma.c
665
op->op_active = 1;
net/rds/rdma.c
666
op->op_recverr = rs->rs_recverr;
net/rds/rdma.c
667
op->op_odp_mr = NULL;
net/rds/rdma.c
670
op->op_sg = rds_message_alloc_sgs(rm, nr_pages);
net/rds/rdma.c
671
if (IS_ERR(op->op_sg)) {
net/rds/rdma.c
672
ret = PTR_ERR(op->op_sg);
net/rds/rdma.c
676
if (op->op_notify || op->op_recverr) {
net/rds/rdma.c
682
op->op_notifier = kmalloc_obj(struct rds_notifier);
net/rds/rdma.c
683
if (!op->op_notifier) {
net/rds/rdma.c
687
op->op_notifier->n_user_token = args->user_token;
net/rds/rdma.c
688
op->op_notifier->n_status = RDS_RDMA_SUCCESS;
net/rds/rdma.c
698
op->op_rkey = rds_rdma_cookie_key(args->cookie);
net/rds/rdma.c
699
op->op_remote_addr = args->remote_vec.addr + rds_rdma_cookie_offset(args->cookie);
net/rds/rdma.c
706
op->op_rkey);
net/rds/rdma.c
719
ret = rds_pin_pages(iov->addr, nr, pages, !op->op_write);
net/rds/rdma.c
754
op->op_odp_mr = local_odp_mr;
net/rds/rdma.c
755
op->op_odp_addr = iov->addr;
net/rds/rdma.c
767
sg = &op->op_sg[op->op_nents + j];
net/rds/rdma.c
780
op->op_nents += nr;
net/rds/rdma.c
790
op->op_bytes = nr_bytes;
net/rds/rdma.c
797
rds_rdma_free_op(op);
net/rds/rds.h
586
int (*xmit_rdma)(struct rds_connection *conn, struct rm_rdma_op *op);
net/rds/rds.h
587
int (*xmit_atomic)(struct rds_connection *conn, struct rm_atomic_op *op);
net/rfkill/core.c
1308
switch (ev.op) {
net/rfkill/core.c
264
enum rfkill_operation op)
net/rfkill/core.c
271
ev->op = op;
net/rfkill/core.c
290
static void rfkill_send_events(struct rfkill *rfkill, enum rfkill_operation op)
net/rfkill/core.c
299
if (rfkill_fill_event(ev, rfkill, data, op)) {
net/rfkill/input.c
102
enum rfkill_sched_op op = rfkill_op;
net/rfkill/input.c
108
__rfkill_handle_global_op(op);
net/rfkill/input.c
153
static void rfkill_schedule_global_op(enum rfkill_sched_op op)
net/rfkill/input.c
158
rfkill_op = op;
net/rfkill/input.c
160
if (op == RFKILL_GLOBAL_OP_EPO && !rfkill_is_epo_lock_active()) {
net/rfkill/input.c
54
static void __rfkill_handle_global_op(enum rfkill_sched_op op)
net/rfkill/input.c
58
switch (op) {
net/rfkill/input.c
78
op);
net/sched/bpf_qdisc.c
11
#define QDISC_OP_IDX(op) (offsetof(struct Qdisc_ops, op) / sizeof(void (*)(void)))
net/sched/cls_flower.c
2010
fl_set_key_val(tb, &key->arp.op, TCA_FLOWER_KEY_ARP_OP,
net/sched/cls_flower.c
2011
&mask->arp.op, TCA_FLOWER_KEY_ARP_OP_MASK,
net/sched/cls_flower.c
2012
sizeof(key->arp.op));
net/sched/cls_flower.c
3593
fl_dump_key_val(skb, &key->arp.op,
net/sched/cls_flower.c
3594
TCA_FLOWER_KEY_ARP_OP, &mask->arp.op,
net/sched/cls_flower.c
3596
sizeof(key->arp.op)) ||
net/sched/em_meta.c
859
switch (meta->lvalue.hdr.op) {
net/sched/sch_fq.c
842
struct rb_node *op, **np, *parent;
net/sched/sch_fq.c
850
while ((op = rb_first(oroot)) != NULL) {
net/sched/sch_fq.c
851
rb_erase(op, oroot);
net/sched/sch_fq.c
852
of = rb_entry(op, struct fq_flow, fq_node);
net/sctp/socket.c
1001
int addrs_size, int op)
net/sctp/socket.c
1011
__func__, sk, addr_buf, addrs_size, op);
net/sctp/socket.c
1035
switch (op) {
net/sunrpc/stats.c
207
static void _print_name(struct seq_file *seq, unsigned int op,
net/sunrpc/stats.c
210
if (procs[op].p_name)
net/sunrpc/stats.c
211
seq_printf(seq, "\t%12s: ", procs[op].p_name);
net/sunrpc/stats.c
212
else if (op == 0)
net/sunrpc/stats.c
215
seq_printf(seq, "\t%12u: ", op);
net/sunrpc/stats.c
232
int op, const struct rpc_procinfo *procs)
net/sunrpc/stats.c
234
_print_name(seq, op, procs);
net/sunrpc/stats.c
257
unsigned int op, maxproc = clnt->cl_maxproc;
net/sunrpc/stats.c
269
for (op = 0; op < maxproc; op++) {
net/sunrpc/stats.c
273
_add_rpc_iostats(&stats, &next->cl_metrics[op]);
net/sunrpc/stats.c
278
_print_rpc_iostats(seq, &stats, op, clnt->cl_procinfo);
net/vmw_vsock/virtio_transport_common.c
1075
.op = VIRTIO_VSOCK_OP_REQUEST,
net/vmw_vsock/virtio_transport_common.c
1087
.op = VIRTIO_VSOCK_OP_SHUTDOWN,
net/vmw_vsock/virtio_transport_common.c
1116
.op = VIRTIO_VSOCK_OP_RW,
net/vmw_vsock/virtio_transport_common.c
1155
.op = VIRTIO_VSOCK_OP_RST,
net/vmw_vsock/virtio_transport_common.c
1162
if (skb && le16_to_cpu(virtio_vsock_hdr(skb)->op) == VIRTIO_VSOCK_OP_RST)
net/vmw_vsock/virtio_transport_common.c
1180
.op = VIRTIO_VSOCK_OP_RST,
net/vmw_vsock/virtio_transport_common.c
1199
if (le16_to_cpu(hdr->op) == VIRTIO_VSOCK_OP_RST)
net/vmw_vsock/virtio_transport_common.c
131
hdr->op = cpu_to_le16(info->op);
net/vmw_vsock/virtio_transport_common.c
1336
switch (le16_to_cpu(hdr->op)) {
net/vmw_vsock/virtio_transport_common.c
1432
switch (le16_to_cpu(hdr->op)) {
net/vmw_vsock/virtio_transport_common.c
1482
if (le16_to_cpu(hdr->op) == VIRTIO_VSOCK_OP_RST)
net/vmw_vsock/virtio_transport_common.c
1492
.op = VIRTIO_VSOCK_OP_RESPONSE,
net/vmw_vsock/virtio_transport_common.c
1539
if (le16_to_cpu(hdr->op) != VIRTIO_VSOCK_OP_REQUEST) {
net/vmw_vsock/virtio_transport_common.c
1626
le16_to_cpu(hdr->op),
net/vmw_vsock/virtio_transport_common.c
196
switch (le16_to_cpu(pkt_hdr->op)) {
net/vmw_vsock/virtio_transport_common.c
199
hdr->op = cpu_to_le16(AF_VSOCK_OP_CONNECT);
net/vmw_vsock/virtio_transport_common.c
203
hdr->op = cpu_to_le16(AF_VSOCK_OP_DISCONNECT);
net/vmw_vsock/virtio_transport_common.c
206
hdr->op = cpu_to_le16(AF_VSOCK_OP_PAYLOAD);
net/vmw_vsock/virtio_transport_common.c
210
hdr->op = cpu_to_le16(AF_VSOCK_OP_CONTROL);
net/vmw_vsock/virtio_transport_common.c
213
hdr->op = cpu_to_le16(AF_VSOCK_OP_UNKNOWN);
net/vmw_vsock/virtio_transport_common.c
316
info->op,
net/vmw_vsock/virtio_transport_common.c
365
if (pkt_len == 0 && info->op == VIRTIO_VSOCK_OP_RW)
net/vmw_vsock/virtio_transport_common.c
405
skb_len == rest_len && info->op == VIRTIO_VSOCK_OP_RW) {
net/vmw_vsock/virtio_transport_common.c
527
.op = VIRTIO_VSOCK_OP_CREDIT_UPDATE,
net/wireless/nl80211.c
2208
#define CMD(op, n) \
net/wireless/nl80211.c
2210
if (rdev->ops->op) { \
samples/bpf/tcp_basertt_kern.c
30
int op;
samples/bpf/tcp_basertt_kern.c
32
op = (int) skops->op;
samples/bpf/tcp_basertt_kern.c
35
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_basertt_kern.c
46
switch (op) {
samples/bpf/tcp_bufs_kern.c
31
int op;
samples/bpf/tcp_bufs_kern.c
42
op = (int) skops->op;
samples/bpf/tcp_bufs_kern.c
51
switch (op) {
samples/bpf/tcp_clamp_kern.c
32
int op;
samples/bpf/tcp_clamp_kern.c
42
op = (int) skops->op;
samples/bpf/tcp_clamp_kern.c
45
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_clamp_kern.c
56
switch (op) {
samples/bpf/tcp_cong_kern.c
29
int op;
samples/bpf/tcp_cong_kern.c
40
op = (int) skops->op;
samples/bpf/tcp_cong_kern.c
43
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_cong_kern.c
54
switch (op) {
samples/bpf/tcp_dumpstats_kern.c
33
switch (ctx->op) {
samples/bpf/tcp_iw_kern.c
32
int op;
samples/bpf/tcp_iw_kern.c
43
op = (int) skops->op;
samples/bpf/tcp_iw_kern.c
46
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_iw_kern.c
52
switch (op) {
samples/bpf/tcp_rwnd_kern.c
28
int op;
samples/bpf/tcp_rwnd_kern.c
39
op = (int) skops->op;
samples/bpf/tcp_rwnd_kern.c
42
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_rwnd_kern.c
46
if (op == BPF_SOCK_OPS_RWND_INIT &&
samples/bpf/tcp_synrto_kern.c
28
int op;
samples/bpf/tcp_synrto_kern.c
39
op = (int) skops->op;
samples/bpf/tcp_synrto_kern.c
42
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_synrto_kern.c
46
if (op == BPF_SOCK_OPS_TIMEOUT_INIT &&
samples/bpf/tcp_tos_reflect_kern.c
33
int op;
samples/bpf/tcp_tos_reflect_kern.c
35
op = (int) skops->op;
samples/bpf/tcp_tos_reflect_kern.c
38
bpf_printk("BPF command: %d\n", op);
samples/bpf/tcp_tos_reflect_kern.c
40
switch (op) {
samples/ftrace/ftrace-ops.c
101
struct ftrace_ops *op,
samples/ftrace/ftrace-ops.c
108
struct ftrace_ops *op,
samples/ftrace/ftrace-ops.c
113
self = container_of(op, struct sample_ops, ops);
samples/seccomp/user-trap.c
26
static int seccomp(unsigned int op, unsigned int flags, void *args)
samples/seccomp/user-trap.c
29
return syscall(__NR_seccomp, op, flags, args);
scripts/gcc-plugins/latent_entropy_plugin.c
303
static enum tree_code op;
scripts/gcc-plugins/latent_entropy_plugin.c
308
switch (op) {
scripts/gcc-plugins/latent_entropy_plugin.c
310
op = PLUS_EXPR;
scripts/gcc-plugins/latent_entropy_plugin.c
315
op = LROTATE_EXPR;
scripts/gcc-plugins/latent_entropy_plugin.c
326
op = BIT_XOR_EXPR;
scripts/gcc-plugins/latent_entropy_plugin.c
331
return op;
scripts/gcc-plugins/latent_entropy_plugin.c
345
enum tree_code op;
scripts/gcc-plugins/latent_entropy_plugin.c
347
op = get_op(&rhs);
scripts/gcc-plugins/latent_entropy_plugin.c
348
assign = create_assign(op, local_entropy, local_entropy, rhs);
scripts/gcc-plugins/latent_entropy_plugin.c
359
enum tree_code op;
scripts/gcc-plugins/latent_entropy_plugin.c
372
op = get_op(NULL);
scripts/gcc-plugins/latent_entropy_plugin.c
373
assign = create_assign(op, temp, temp, local_entropy);
scripts/gcc-plugins/latent_entropy_plugin.c
435
enum tree_code op;
scripts/gcc-plugins/latent_entropy_plugin.c
471
op = get_op(NULL);
scripts/gcc-plugins/latent_entropy_plugin.c
472
assign = create_assign(op, local_entropy, local_entropy, rand_const);
scripts/recordmcount.c
237
unsigned char *op;
scripts/recordmcount.c
244
op = map + offset - 1;
scripts/recordmcount.c
245
if (*op != 0xe8)
scripts/unifdef.c
807
} op[5];
scripts/unifdef.c
915
const struct op *op;
scripts/unifdef.c
927
for (op = ops->op; op->str != NULL; op++)
scripts/unifdef.c
928
if (strncmp(cp, op->str, strlen(op->str)) == 0)
scripts/unifdef.c
930
if (op->str == NULL)
scripts/unifdef.c
932
cp += strlen(op->str);
scripts/unifdef.c
933
debug("eval%d %s", ops - eval_ops, op->str);
scripts/unifdef.c
937
lt = op->fn(valp, lt, *valp, rt, val);
security/apparmor/af_unix.c
240
return unix_fs_perm(ad->op, request, ad->subj_cred,
security/apparmor/af_unix.c
302
return unix_fs_perm(ad->op, AA_MAY_LISTEN,
security/apparmor/af_unix.c
33
static int unix_fs_perm(const char *op, u32 mask, const struct cred *subj_cred,
security/apparmor/af_unix.c
337
return unix_fs_perm(ad->op, AA_MAY_ACCEPT,
security/apparmor/af_unix.c
368
return unix_fs_perm(ad->op, request,
security/apparmor/af_unix.c
410
return unix_fs_perm(ad->op, request, ad->subj_cred,
security/apparmor/af_unix.c
413
return unix_fs_perm(ad->op, request, ad->subj_cred,
security/apparmor/af_unix.c
447
const char *op, u32 request, struct sock *sk,
security/apparmor/af_unix.c
452
DEFINE_AUDIT_SK(ad, op, subj_cred, sk);
security/apparmor/af_unix.c
462
int aa_unix_sock_perm(const char *op, u32 request, struct socket *sock)
security/apparmor/af_unix.c
468
error = aa_unix_label_sk_perm(current_cred(), label, op,
security/apparmor/af_unix.c
55
return aa_path_perm(op, subj_cred, label, path,
security/apparmor/af_unix.c
580
int aa_unix_opt_perm(const char *op, u32 request, struct socket *sock,
security/apparmor/af_unix.c
589
DEFINE_AUDIT_SK(ad, op, current_cred(), sock->sk);
security/apparmor/af_unix.c
601
struct aa_label *label, const char *op, u32 request,
security/apparmor/af_unix.c
607
DEFINE_AUDIT_SK(ad, op, subj_cred, sk);
security/apparmor/af_unix.c
624
struct aa_label *label, const char *op, u32 request,
security/apparmor/af_unix.c
638
return unix_peer_perm(subj_cred, label, op, request, sk,
security/apparmor/af_unix.c
710
const char *op, u32 request, struct file *file)
security/apparmor/af_unix.c
744
error = aa_unix_label_sk_perm(subj_cred, label, op,
security/apparmor/af_unix.c
758
unix_fs_perm(op, request, subj_cred, label,
security/apparmor/af_unix.c
771
xcheck(unix_peer_perm(subj_cred, label, op,
security/apparmor/af_unix.c
778
unix_peer_perm(file->f_cred, plabel, op,
security/apparmor/audit.c
101
if (ad->op)
security/apparmor/audit.c
102
audit_log_format(ab, " operation=\"%s\"", ad->op);
security/apparmor/audit.c
220
int aa_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule, gfp_t gfp)
security/apparmor/audit.c
226
if (op != Audit_equal && op != Audit_not_equal)
security/apparmor/audit.c
267
int aa_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op, void *vrule)
security/apparmor/audit.c
283
switch (op) {
security/apparmor/domain.c
1352
static int change_profile_perms_wrapper(const char *op, const char *name,
security/apparmor/domain.c
1367
error = aa_audit_file(subj_cred, profile, perms, op, request,
security/apparmor/domain.c
1401
char *op;
security/apparmor/domain.c
1425
op = OP_STACK_ONEXEC;
security/apparmor/domain.c
1427
op = OP_CHANGE_ONEXEC;
security/apparmor/domain.c
1431
op = OP_STACK;
security/apparmor/domain.c
1433
op = OP_CHANGE_PROFILE;
security/apparmor/domain.c
1452
aa_audit_file(subj_cred, profile, &perms, op,
security/apparmor/domain.c
1498
change_profile_perms_wrapper(op, auditname,
security/apparmor/domain.c
1572
profile, &perms, op, request, auditname,
security/apparmor/file.c
102
DEFINE_AUDIT_DATA(ad, LSM_AUDIT_DATA_TASK, AA_CLASS_FILE, op);
security/apparmor/file.c
149
static int path_name(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
167
profile, &nullperms, op, request, *name,
security/apparmor/file.c
226
int __aa_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
242
profile, perms, op, request, name, NULL, NULL,
security/apparmor/file.c
247
static int profile_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
259
error = path_name(op, subj_cred, &profile->label, path,
security/apparmor/file.c
264
return __aa_path_perm(op, subj_cred, profile, name, request, cond,
security/apparmor/file.c
280
int aa_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
296
profile_path_perm(op, subj_cred, profile, path, buffer,
security/apparmor/file.c
483
static int __file_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
511
profile_path_perm(op, subj_cred, profile,
security/apparmor/file.c
525
profile_path_perm(op, subj_cred,
security/apparmor/file.c
531
profile_path_perm(op, subj_cred,
security/apparmor/file.c
544
static int __file_sock_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
556
error = aa_sock_file_perm(subj_cred, label, op, request, file);
security/apparmor/file.c
560
last_error(error, aa_sock_file_perm(subj_cred, flabel, op,
security/apparmor/file.c
619
int aa_file_perm(const char *op, const struct cred *subj_cred,
security/apparmor/file.c
661
error = __file_path_perm(op, subj_cred, label, flabel, file,
security/apparmor/file.c
665
error = __file_sock_perm(op, subj_cred, label, flabel, file,
security/apparmor/file.c
97
const char *op, u32 request, const char *name,
security/apparmor/include/af_unix.h
36
struct aa_label *label, const char *op, u32 request,
security/apparmor/include/af_unix.h
39
int aa_unix_sock_perm(const char *op, u32 request, struct socket *sock);
security/apparmor/include/af_unix.h
48
int aa_unix_msg_perm(const char *op, u32 request, struct socket *sock,
security/apparmor/include/af_unix.h
50
int aa_unix_opt_perm(const char *op, u32 request, struct socket *sock, int level,
security/apparmor/include/af_unix.h
53
const char *op, u32 request, struct file *file);
security/apparmor/include/audit.h
115
const char *op;
security/apparmor/include/audit.h
180
.op = (X), \
security/apparmor/include/audit.h
208
int aa_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule, gfp_t gfp);
security/apparmor/include/audit.h
210
int aa_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op, void *vrule);
security/apparmor/include/file.h
76
const char *op, u32 request, const char *name,
security/apparmor/include/file.h
87
int __aa_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/include/file.h
91
int aa_path_perm(const char *op, const struct cred *subj_cred,
security/apparmor/include/file.h
99
int aa_file_perm(const char *op, const struct cred *subj_cred,
security/apparmor/include/net.h
100
const char *op, u32 request, u16 family,
security/apparmor/include/net.h
110
int aa_sk_perm(const char *op, u32 request, struct sock *sk);
security/apparmor/include/net.h
113
const char *op, u32 request,
security/apparmor/include/net.h
116
int apparmor_secmark_check(struct aa_label *label, char *op, u32 request,
security/apparmor/lsm.c
1412
static int aa_sock_msg_perm(const char *op, u32 request, struct socket *sock,
security/apparmor/lsm.c
1423
return aa_sk_perm(op, request, sock->sk);
security/apparmor/lsm.c
1439
static int aa_sock_perm(const char *op, u32 request, struct socket *sock)
security/apparmor/lsm.c
1446
return aa_unix_sock_perm(op, request, sock);
security/apparmor/lsm.c
1447
return aa_sk_perm(op, request, sock->sk);
security/apparmor/lsm.c
1461
static int aa_sock_opt_perm(const char *op, u32 request, struct socket *sock,
security/apparmor/lsm.c
1469
return aa_unix_opt_perm(op, request, sock, level, optname);
security/apparmor/lsm.c
1470
return aa_sk_perm(op, request, sock->sk);
security/apparmor/lsm.c
219
static int common_perm(const char *op, const struct path *path, u32 mask,
security/apparmor/lsm.c
228
error = aa_path_perm(op, current_cred(), label, path, 0, mask,
security/apparmor/lsm.c
243
static int common_perm_cond(const char *op, const struct path *path, u32 mask)
security/apparmor/lsm.c
255
return common_perm(op, path, mask, &cond);
security/apparmor/lsm.c
268
static int common_perm_dir_dentry(const char *op, const struct path *dir,
security/apparmor/lsm.c
274
return common_perm(op, &path, mask, cond);
security/apparmor/lsm.c
286
static int common_perm_rm(const char *op, const struct path *dir,
security/apparmor/lsm.c
300
return common_perm_dir_dentry(op, dir, dentry, mask, &cond);
security/apparmor/lsm.c
313
static int common_perm_create(const char *op, const struct path *dir,
security/apparmor/lsm.c
321
return common_perm_dir_dentry(op, dir, dentry, mask, &cond);
security/apparmor/lsm.c
523
static int common_file_perm(const char *op, struct file *file, u32 mask)
security/apparmor/lsm.c
529
error = aa_file_perm(op, current_cred(), label, file, mask, false);
security/apparmor/lsm.c
555
static int common_mmap(const char *op, struct file *file, unsigned long prot,
security/apparmor/lsm.c
574
return common_file_perm(op, file, mask);
security/apparmor/mount.c
135
struct aa_profile *profile, const char *op,
security/apparmor/mount.c
142
DEFINE_AUDIT_DATA(ad, LSM_AUDIT_DATA_NONE, AA_CLASS_MOUNT, op);
security/apparmor/net.c
273
const char *op, u32 request, u16 family, int type, int protocol)
security/apparmor/net.c
276
DEFINE_AUDIT_NET(ad, op, subj_cred, NULL, family, type, protocol);
security/apparmor/net.c
285
const char *op, u32 request,
security/apparmor/net.c
296
DEFINE_AUDIT_SK(ad, op, subj_cred, sk);
security/apparmor/net.c
306
int aa_sk_perm(const char *op, u32 request, struct sock *sk)
security/apparmor/net.c
316
error = aa_label_sk_perm(current_cred(), label, op, request, sk);
security/apparmor/net.c
324
const char *op, u32 request, struct file *file)
security/apparmor/net.c
335
return aa_unix_file_perm(subj_cred, label, op, request, file);
security/apparmor/net.c
336
return aa_label_sk_perm(subj_cred, label, op, request, sock->sk);
security/apparmor/net.c
395
int apparmor_secmark_check(struct aa_label *label, char *op, u32 request,
security/apparmor/net.c
399
DEFINE_AUDIT_SK(ad, op, NULL, sk);
security/apparmor/policy.c
1004
return audit_policy(label, op, NULL, NULL, "not policy admin",
security/apparmor/policy.c
1169
const char *op;
security/apparmor/policy.c
1173
op = mask & AA_MAY_REPLACE_POLICY ? OP_PROF_REPL : OP_PROF_LOAD;
security/apparmor/policy.c
1209
op = OP_PROF_LOAD;
security/apparmor/policy.c
1331
op = (!ent->old && !ent->rename) ? OP_PROF_LOAD : OP_PROF_REPL;
security/apparmor/policy.c
1336
audit_policy(label, op, ns_name, ent->new->base.hname,
security/apparmor/policy.c
1349
audit_policy(label, op, ns_name, ent->new->base.hname, NULL,
security/apparmor/policy.c
1386
op = (ent && !ent->old) ? OP_PROF_LOAD : OP_PROF_REPL;
security/apparmor/policy.c
1388
audit_policy(label, op, ns_name, ent ? ent->new->base.hname : NULL,
security/apparmor/policy.c
1398
op = (!tmp->old) ? OP_PROF_LOAD : OP_PROF_REPL;
security/apparmor/policy.c
1399
audit_policy(label, op, ns_name, tmp->new->base.hname, info,
security/apparmor/policy.c
842
static int audit_policy(struct aa_label *subj_label, const char *op,
security/apparmor/policy.c
846
DEFINE_AUDIT_DATA(ad, LSM_AUDIT_DATA_NONE, AA_CLASS_NONE, op);
security/apparmor/policy.c
984
const char *op;
security/apparmor/policy.c
987
op = OP_PROF_RM;
security/apparmor/policy.c
989
op = OP_PROF_REPL;
security/apparmor/policy.c
991
op = OP_PROF_LOAD;
security/apparmor/policy.c
995
return audit_policy(label, op, NULL, NULL, "policy_locked",
security/apparmor/policy.c
999
return audit_policy(label, op, NULL, NULL,
security/apparmor/procattr.c
77
static char *split_token_from_name(const char *op, char *args, u64 *token)
security/apparmor/procattr.c
83
AA_ERROR("%s: Invalid input '%s'", op, args);
security/integrity/ima/ima.h
263
const char *op, struct inode *inode,
security/integrity/ima/ima.h
272
struct ima_iint_cache *iint, const char *op,
security/integrity/ima/ima.h
561
static inline int ima_filter_rule_init(u32 field, u32 op, char *rulestr,
security/integrity/ima/ima.h
571
static inline int ima_filter_rule_match(struct lsm_prop *prop, u32 field, u32 op,
security/integrity/ima/ima_api.c
108
static const char op[] = "add_template_measure";
security/integrity/ima/ima_api.c
118
template_name, op,
security/integrity/ima/ima_api.c
124
result = ima_add_template_entry(entry, violation, op, inode, filename);
security/integrity/ima/ima_api.c
136
struct ima_iint_cache *iint, const char *op,
security/integrity/ima/ima_api.c
162
op, cause, result, 0);
security/integrity/ima/ima_api.c
356
static const char op[] = "add_template_measure";
security/integrity/ima/ima_api.c
381
op, audit_cause, result, 0);
security/integrity/ima/ima_appraise.c
487
static const char op[] = "appraise_data";
security/integrity/ima/ima_appraise.c
586
op, cause, rc, 0);
security/integrity/ima/ima_appraise.c
606
op, cause, rc, 0);
security/integrity/ima/ima_init.c
44
static const char op[] = "add_boot_aggregate";
security/integrity/ima/ima_init.c
98
integrity_audit_msg(AUDIT_INTEGRITY_PCR, NULL, boot_aggregate_name, op,
security/integrity/ima/ima_policy.c
1965
static const char op[] = "update_policy";
security/integrity/ima/ima_policy.c
1981
NULL, op, "-ENOMEM", -ENOMEM, audit_info);
security/integrity/ima/ima_policy.c
1991
NULL, op, "invalid-policy", result,
security/integrity/ima/ima_queue.c
167
const char *op, struct inode *inode,
security/integrity/ima/ima_queue.c
219
op, audit_cause, result, audit_info);
security/integrity/integrity.h
209
const unsigned char *fname, const char *op,
security/integrity/integrity.h
213
const unsigned char *fname, const char *op,
security/integrity/integrity.h
226
const char *op, const char *cause,
security/integrity/integrity.h
234
const char *op, const char *cause,
security/integrity/integrity_audit.c
29
const unsigned char *fname, const char *op,
security/integrity/integrity_audit.c
32
integrity_audit_message(audit_msgno, inode, fname, op, cause,
security/integrity/integrity_audit.c
37
const unsigned char *fname, const char *op,
security/integrity/integrity_audit.c
56
audit_log_format(ab, " op=%s cause=%s comm=", op, cause);
security/ipe/audit.c
132
const char *op = audit_op_names[ctx->op];
security/ipe/audit.c
146
op, audit_hook_names[ctx->hook], READ_ONCE(enforce),
security/ipe/audit.c
167
audit_log_format(ab, " rule=\"DEFAULT op=%s action=%s\"", op,
security/ipe/audit.c
201
void ipe_audit_policy_activation(const struct ipe_policy *const op,
security/ipe/audit.c
211
if (op) {
security/ipe/audit.c
212
audit_policy(ab, AUDIT_OLD_ACTIVE_POLICY_FMT, op);
security/ipe/audit.c
98
audit_log_format(ab, " rule=\"op=%s ", audit_op_names[r->op]);
security/ipe/audit.h
15
void ipe_audit_policy_activation(const struct ipe_policy *const op,
security/ipe/eval.c
100
ctx->op = op;
security/ipe/eval.c
334
if (ctx->op == IPE_OP_INVALID) {
security/ipe/eval.c
345
rules = &pol->parsed->rules[ctx->op];
security/ipe/eval.c
94
enum ipe_op_type op,
security/ipe/eval.h
41
enum ipe_op_type op;
security/ipe/eval.h
66
enum ipe_op_type op,
security/ipe/hooks.c
141
enum ipe_op_type op;
security/ipe/hooks.c
145
op = IPE_OP_FIRMWARE;
security/ipe/hooks.c
149
op = IPE_OP_KERNEL_MODULE;
security/ipe/hooks.c
152
op = IPE_OP_KEXEC_INITRAMFS;
security/ipe/hooks.c
155
op = IPE_OP_KEXEC_IMAGE;
security/ipe/hooks.c
158
op = IPE_OP_POLICY;
security/ipe/hooks.c
161
op = IPE_OP_X509;
security/ipe/hooks.c
164
op = IPE_OP_INVALID;
security/ipe/hooks.c
168
ipe_build_eval_ctx(&ctx, file, op, IPE_HOOK_KERNEL_READ);
security/ipe/hooks.c
187
enum ipe_op_type op;
security/ipe/hooks.c
191
op = IPE_OP_FIRMWARE;
security/ipe/hooks.c
194
op = IPE_OP_KERNEL_MODULE;
security/ipe/hooks.c
197
op = IPE_OP_KEXEC_INITRAMFS;
security/ipe/hooks.c
200
op = IPE_OP_KEXEC_IMAGE;
security/ipe/hooks.c
203
op = IPE_OP_POLICY;
security/ipe/hooks.c
206
op = IPE_OP_X509;
security/ipe/hooks.c
209
op = IPE_OP_INVALID;
security/ipe/hooks.c
213
ipe_build_eval_ctx(&ctx, NULL, op, IPE_HOOK_KERNEL_LOAD);
security/ipe/policy.h
54
enum ipe_op_type op;
security/ipe/policy_parser.c
365
enum ipe_op_type op = IPE_OP_INVALID;
security/ipe/policy_parser.c
390
op = parse_operation(t);
security/ipe/policy_parser.c
391
if (op == IPE_OP_INVALID)
security/ipe/policy_parser.c
414
} else if (op == IPE_OP_INVALID) {
security/ipe/policy_parser.c
420
if (p->rules[op].default_action != IPE_ACTION_INVALID)
security/ipe/policy_parser.c
423
p->rules[op].default_action = action;
security/ipe/policy_parser.c
425
} else if (op != IPE_OP_INVALID && action != IPE_ACTION_INVALID) {
security/ipe/policy_parser.c
426
r->op = op;
security/ipe/policy_parser.c
435
list_add_tail(&r->next, &p->rules[op].rules);
security/keys/internal.h
207
const char *op,
security/keys/internal.h
338
static inline long keyctl_pkey_e_d_s(int op,
security/keys/keyctl_pkey.c
115
int op,
security/keys/keyctl_pkey.c
136
switch (op) {
security/keys/keyctl_pkey.c
212
long keyctl_pkey_e_d_s(int op,
security/keys/keyctl_pkey.c
222
ret = keyctl_pkey_params_get_2(_params, _info, op, &params);
security/keys/keyctl_pkey.c
230
switch (op) {
security/keys/keyctl_pkey.c
232
params.op = kernel_pkey_encrypt;
security/keys/keyctl_pkey.c
235
params.op = kernel_pkey_decrypt;
security/keys/keyctl_pkey.c
238
params.op = kernel_pkey_sign;
security/keys/keyctl_pkey.c
318
params.op = kernel_pkey_verify;
security/keys/request_key.c
130
kenter("{%d},{%d},%s", key->serial, authkey->serial, rka->op);
security/keys/request_key.c
186
argv[i++] = (char *)rka->op;
security/keys/request_key_auth.c
161
struct key *request_key_auth_new(struct key *target, const char *op,
security/keys/request_key_auth.c
181
strscpy(rka->op, op, sizeof(rka->op));
security/landlock/fs.c
1737
.u.op = &(struct lsm_ioctlop_audit) {
security/lsm_audit.c
225
audit_log_d_path(ab, " path=", &a->u.op->path);
security/lsm_audit.c
227
inode = a->u.op->path.dentry->d_inode;
security/lsm_audit.c
234
audit_log_format(ab, " ioctlcmd=0x%hx", a->u.op->cmd);
security/security.c
5109
int security_audit_rule_init(u32 field, u32 op, char *rulestr, void **lsmrule,
security/security.c
5112
return call_int_hook(audit_rule_init, field, op, rulestr, lsmrule, gfp);
security/security.c
5154
int security_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op,
security/security.c
5157
return call_int_hook(audit_rule_match, prop, field, op, lsmrule);
security/selinux/hooks.c
3846
ad.u.op = &ioctl;
security/selinux/hooks.c
3847
ad.u.op->cmd = cmd;
security/selinux/hooks.c
3848
ad.u.op->path = file->f_path;
security/selinux/include/audit.h
39
int selinux_audit_rule_init(u32 field, u32 op, char *rulestr, void **rule,
security/selinux/include/audit.h
61
int selinux_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op,
security/selinux/ss/constraint.h
50
u32 op; /* operator */
security/selinux/ss/policydb.c
1270
e->op = le32_to_cpu(buf[2]);
security/selinux/ss/policydb.c
3060
buf[2] = cpu_to_le32(e->op);
security/selinux/ss/services.c
312
switch (e->op) {
security/selinux/ss/services.c
356
switch (e->op) {
security/selinux/ss/services.c
3590
int selinux_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule,
security/selinux/ss/services.c
3616
if (op != Audit_equal && op != Audit_not_equal)
security/selinux/ss/services.c
3716
int selinux_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op, void *vrule)
security/selinux/ss/services.c
3755
switch (op) {
security/selinux/ss/services.c
3766
switch (op) {
security/selinux/ss/services.c
3777
switch (op) {
security/selinux/ss/services.c
3793
switch (op) {
security/selinux/ss/services.c
382
switch (e->op) {
security/selinux/ss/services.c
418
switch (e->op) {
security/smack/smack_lsm.c
4777
static int smack_audit_rule_init(u32 field, u32 op, char *rulestr, void **vrule,
security/smack/smack_lsm.c
4787
if (op != Audit_equal && op != Audit_not_equal)
security/smack/smack_lsm.c
4832
static int smack_audit_rule_match(struct lsm_prop *prop, u32 field, u32 op,
security/smack/smack_lsm.c
4851
if (op == Audit_equal)
security/smack/smack_lsm.c
4853
if (op == Audit_not_equal)
sound/core/control.c
1799
int op;
sound/core/control.c
1811
if (op_flag == pairs[i].op && (vd->access & pairs[i].perm))
sound/drivers/opl3/opl3_drums.c
124
unsigned char op_offset = snd_opl3_regmap[data->voice][data->op];
sound/drivers/opl3/opl3_drums.c
34
int op;
sound/drivers/opl3/opl3_drums.c
70
unsigned char op_offset = snd_opl3_regmap[data->voice][data->op];
sound/drivers/opl3/opl3_midi.c
428
vol_op[i] = fm->op[i].ksl_level;
sound/drivers/opl3/opl3_midi.c
458
reg_val = fm->op[i].am_vib;
sound/drivers/opl3/opl3_midi.c
468
reg_val = fm->op[i].attack_decay;
sound/drivers/opl3/opl3_midi.c
473
reg_val = fm->op[i].sustain_release;
sound/drivers/opl3/opl3_midi.c
478
reg_val = fm->op[i].wave_select;
sound/drivers/opl3/opl3_synth.c
255
patch->inst.op[i].am_vib = data[AM_VIB + i];
sound/drivers/opl3/opl3_synth.c
256
patch->inst.op[i].ksl_level = data[KSL_LEVEL + i];
sound/drivers/opl3/opl3_synth.c
257
patch->inst.op[i].attack_decay = data[ATTACK_DECAY + i];
sound/drivers/opl3/opl3_synth.c
258
patch->inst.op[i].sustain_release = data[SUSTAIN_RELEASE + i];
sound/drivers/opl3/opl3_synth.c
259
patch->inst.op[i].wave_select = data[WAVE_SELECT + i];
sound/drivers/opl3/opl3_synth.c
265
patch->inst.op[i+2].am_vib =
sound/drivers/opl3/opl3_synth.c
267
patch->inst.op[i+2].ksl_level =
sound/drivers/opl3/opl3_synth.c
269
patch->inst.op[i+2].attack_decay =
sound/drivers/opl3/opl3_synth.c
271
patch->inst.op[i+2].sustain_release =
sound/drivers/opl3/opl3_synth.c
273
patch->inst.op[i+2].wave_select =
sound/drivers/opl3/opl3_synth.c
445
if (voice->op > 1)
sound/drivers/opl3/opl3_synth.c
465
voice_op = array_index_nospec(voice->op, 4);
sound/firewire/bebob/bebob_proc.c
160
void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b))
sound/firewire/bebob/bebob_proc.c
166
snd_info_set_text_ops(entry, bebob, op);
sound/firewire/dice/dice-proc.c
281
void (*op)(struct snd_info_entry *entry,
sound/firewire/dice/dice-proc.c
288
snd_info_set_text_ops(entry, dice, op);
sound/firewire/fireface/ff-proc.c
37
void (*op)(struct snd_info_entry *e,
sound/firewire/fireface/ff-proc.c
44
snd_info_set_text_ops(entry, ff, op);
sound/firewire/fireworks/fireworks_proc.c
196
void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b))
sound/firewire/fireworks/fireworks_proc.c
202
snd_info_set_text_ops(entry, efw, op);
sound/firewire/motu/motu-proc.c
83
void (*op)(struct snd_info_entry *e,
sound/firewire/motu/motu-proc.c
90
snd_info_set_text_ops(entry, motu, op);
sound/firewire/oxfw/oxfw-proc.c
79
void (*op)(struct snd_info_entry *e,
sound/firewire/oxfw/oxfw-proc.c
86
snd_info_set_text_ops(entry, oxfw, op);
sound/firewire/tascam/tascam-proc.c
54
void (*op)(struct snd_info_entry *e,
sound/firewire/tascam/tascam-proc.c
61
snd_info_set_text_ops(entry, tscm, op);
sound/pci/emu10k1/emufx.c
487
u32 op, u32 r, u32 a, u32 x, u32 y)
sound/pci/emu10k1/emufx.c
495
code[1] = ((op & 0x0f) << 20) | ((r & 0x3ff) << 10) | (a & 0x3ff);
sound/pci/emu10k1/emufx.c
499
#define OP(icode, ptr, op, r, a, x, y) \
sound/pci/emu10k1/emufx.c
500
snd_emu10k1_write_op(icode, ptr, op, r, a, x, y)
sound/pci/emu10k1/emufx.c
504
u32 op, u32 r, u32 a, u32 x, u32 y)
sound/pci/emu10k1/emufx.c
512
code[1] = ((op & 0x0f) << 24) | ((r & 0x7ff) << 12) | (a & 0x7ff);
sound/pci/emu10k1/emufx.c
516
#define A_OP(icode, ptr, op, r, a, x, y) \
sound/pci/emu10k1/emufx.c
517
snd_emu10k1_audigy_write_op(icode, ptr, op, r, a, x, y)
sound/pci/oxygen/xonar_dg.c
100
int cs4245_shadow_control(struct oxygen *chip, enum cs4245_shadow_operation op)
sound/pci/oxygen/xonar_dg.c
107
ret = (op == CS4245_SAVE_TO_SHADOW ?
sound/pci/oxygen/xonar_dg.h
41
int cs4245_shadow_control(struct oxygen *chip, enum cs4245_shadow_operation op);
sound/pci/rme96.c
1060
int op)
sound/pci/rme96.c
1062
if (op & RME96_TB_RESET_PLAYPOS)
sound/pci/rme96.c
1064
if (op & RME96_TB_RESET_CAPTUREPOS)
sound/pci/rme96.c
1066
if (op & RME96_TB_CLEAR_PLAYBACK_IRQ) {
sound/pci/rme96.c
1071
if (op & RME96_TB_CLEAR_CAPTURE_IRQ) {
sound/pci/rme96.c
1076
if (op & RME96_TB_START_PLAYBACK)
sound/pci/rme96.c
1078
if (op & RME96_TB_STOP_PLAYBACK)
sound/pci/rme96.c
1080
if (op & RME96_TB_START_CAPTURE)
sound/pci/rme96.c
1082
if (op & RME96_TB_STOP_CAPTURE)
sound/soc/codecs/wm_hubs.c
60
static void wait_for_dc_servo(struct snd_soc_component *component, unsigned int op)
sound/soc/codecs/wm_hubs.c
68
val = op | WM8993_DCS_ENA_CHAN_0 | WM8993_DCS_ENA_CHAN_1;
sound/soc/codecs/wm_hubs.c
91
} while (reg & op && count < timeout);
sound/soc/codecs/wm_hubs.c
93
if (reg & op)
sound/soc/codecs/wm_hubs.c
95
op);
sound/soc/fsl/mpc5200_dma.c
313
int mpc5200_audio_dma_create(struct platform_device *op)
sound/soc/fsl/mpc5200_dma.c
324
irq = irq_of_parse_and_map(op->dev.of_node, 0);
sound/soc/fsl/mpc5200_dma.c
325
if (of_address_to_resource(op->dev.of_node, 0, &res)) {
sound/soc/fsl/mpc5200_dma.c
326
dev_err(&op->dev, "Missing reg property\n");
sound/soc/fsl/mpc5200_dma.c
331
dev_err(&op->dev, "Could not map registers\n");
sound/soc/fsl/mpc5200_dma.c
343
prop = of_get_property(op->dev.of_node, "cell-index", &size);
sound/soc/fsl/mpc5200_dma.c
355
psc_dma->dev = &op->dev;
sound/soc/fsl/mpc5200_dma.c
369
dev_err(&op->dev, "Could not allocate bestcomm tasks\n");
sound/soc/fsl/mpc5200_dma.c
416
dev_set_drvdata(&op->dev, psc_dma);
sound/soc/fsl/mpc5200_dma.c
419
return devm_snd_soc_register_component(&op->dev,
sound/soc/fsl/mpc5200_dma.c
433
int mpc5200_audio_dma_destroy(struct platform_device *op)
sound/soc/fsl/mpc5200_dma.c
435
struct psc_dma *psc_dma = dev_get_drvdata(&op->dev);
sound/soc/fsl/mpc5200_dma.c
437
dev_dbg(&op->dev, "mpc5200_audio_dma_destroy()\n");
sound/soc/fsl/mpc5200_dma.c
449
dev_set_drvdata(&op->dev, NULL);
sound/soc/fsl/mpc5200_dma.h
85
int mpc5200_audio_dma_create(struct platform_device *op);
sound/soc/fsl/mpc5200_dma.h
86
int mpc5200_audio_dma_destroy(struct platform_device *op);
sound/soc/fsl/mpc5200_psc_ac97.c
275
static int psc_ac97_of_probe(struct platform_device *op)
sound/soc/fsl/mpc5200_psc_ac97.c
280
rc = mpc5200_audio_dma_create(op);
sound/soc/fsl/mpc5200_psc_ac97.c
286
dev_err(&op->dev, "Failed to set AC'97 ops: %d\n", rc);
sound/soc/fsl/mpc5200_psc_ac97.c
290
rc = snd_soc_register_component(&op->dev, &psc_ac97_component,
sound/soc/fsl/mpc5200_psc_ac97.c
293
dev_err(&op->dev, "Failed to register DAI\n");
sound/soc/fsl/mpc5200_psc_ac97.c
297
psc_dma = dev_get_drvdata(&op->dev);
sound/soc/fsl/mpc5200_psc_ac97.c
313
static void psc_ac97_of_remove(struct platform_device *op)
sound/soc/fsl/mpc5200_psc_ac97.c
315
mpc5200_audio_dma_destroy(op);
sound/soc/fsl/mpc5200_psc_ac97.c
316
snd_soc_unregister_component(&op->dev);
sound/soc/fsl/mpc5200_psc_i2s.c
159
static int psc_i2s_of_probe(struct platform_device *op)
sound/soc/fsl/mpc5200_psc_i2s.c
165
rc = mpc5200_audio_dma_create(op);
sound/soc/fsl/mpc5200_psc_i2s.c
169
rc = snd_soc_register_component(&op->dev, &psc_i2s_component,
sound/soc/fsl/mpc5200_psc_i2s.c
176
psc_dma = dev_get_drvdata(&op->dev);
sound/soc/fsl/mpc5200_psc_i2s.c
187
if (!of_property_present(op->dev.of_node, "codec-handle"))
sound/soc/fsl/mpc5200_psc_i2s.c
212
static void psc_i2s_of_remove(struct platform_device *op)
sound/soc/fsl/mpc5200_psc_i2s.c
214
mpc5200_audio_dma_destroy(op);
sound/soc/fsl/mpc5200_psc_i2s.c
215
snd_soc_unregister_component(&op->dev);
sound/soc/fsl/pcm030-audio-fabric.c
102
dev_err(&op->dev, "snd_soc_register_card() failed: %d\n", ret);
sound/soc/fsl/pcm030-audio-fabric.c
106
platform_set_drvdata(op, pdata);
sound/soc/fsl/pcm030-audio-fabric.c
111
static void pcm030_fabric_remove(struct platform_device *op)
sound/soc/fsl/pcm030-audio-fabric.c
113
struct pcm030_audio_data *pdata = platform_get_drvdata(op);
sound/soc/fsl/pcm030-audio-fabric.c
55
static int pcm030_fabric_probe(struct platform_device *op)
sound/soc/fsl/pcm030-audio-fabric.c
57
struct device_node *np = op->dev.of_node;
sound/soc/fsl/pcm030-audio-fabric.c
68
pdata = devm_kzalloc(&op->dev, sizeof(struct pcm030_audio_data),
sound/soc/fsl/pcm030-audio-fabric.c
73
card->dev = &op->dev;
sound/soc/fsl/pcm030-audio-fabric.c
79
dev_err(&op->dev, "ac97 not registered\n");
sound/soc/fsl/pcm030-audio-fabric.c
88
dev_err(&op->dev, "request_module returned: %d\n", ret);
sound/soc/fsl/pcm030-audio-fabric.c
92
dev_err(&op->dev, "platform_device_alloc() failed\n");
sound/soc/fsl/pcm030-audio-fabric.c
96
dev_err(&op->dev, "platform_device_add() failed: %d\n", ret);
sound/soc/intel/avs/avs.h
63
#define avs_dsp_op(adev, op, ...) \
sound/soc/intel/avs/avs.h
64
((adev)->spec->dsp_ops->op(adev, ## __VA_ARGS__))
sound/soc/intel/avs/pcm.c
1178
int (*op)(struct snd_soc_dai *, struct avs_dma_data *))
sound/soc/intel/avs/pcm.c
1190
ret = op(dai, data);
sound/soc/intel/avs/pcm.c
1203
ret = op(dai, data);
sound/soc/intel/avs/trace.h
129
TP_PROTO(const char *op, bool proceed, u64 header),
sound/soc/intel/avs/trace.h
13
TP_PROTO(unsigned int reg, unsigned int mask, const char *op, bool flag),
sound/soc/intel/avs/trace.h
131
TP_ARGS(op, proceed, header),
sound/soc/intel/avs/trace.h
134
__string(op, op )
sound/soc/intel/avs/trace.h
140
__assign_str(op);
sound/soc/intel/avs/trace.h
146
__entry->proceed ? "" : "ignore ", __get_str(op),
sound/soc/intel/avs/trace.h
15
TP_ARGS(reg, mask, op, flag),
sound/soc/intel/avs/trace.h
20
__string(op, op )
sound/soc/intel/avs/trace.h
27
__assign_str(op);
sound/soc/intel/avs/trace.h
32
__get_str(op), __entry->flag, __entry->mask, __entry->reg)
sound/soc/qcom/qdsp6/q6apm.c
493
static int graph_callback(const struct gpr_resp_pkt *data, void *priv, int op)
sound/soc/qcom/qdsp6/q6apm.c
768
static int apm_callback(const struct gpr_resp_pkt *data, void *priv, int op)
sound/soc/qcom/qdsp6/q6prm.c
178
static int prm_callback(const struct gpr_resp_pkt *data, void *priv, int op)
sound/soc/samsung/idma.c
109
static void idma_control(int op)
sound/soc/samsung/idma.c
115
switch (op) {
sound/soc/sof/sof-audio.c
675
enum sof_widget_op op)
sound/soc/sof/sof-audio.c
696
switch (op) {
sound/soc/sof/sof-audio.c
726
dev_err(sdev->dev, "Invalid widget op %d\n", op);
sound/sparc/amd7930.c
339
struct platform_device *op;
sound/sparc/amd7930.c
870
struct platform_device *op = amd->op;
sound/sparc/amd7930.c
878
of_iounmap(&op->resource[0], amd->regs,
sound/sparc/amd7930.c
879
resource_size(&op->resource[0]));
sound/sparc/amd7930.c
898
struct platform_device *op,
sound/sparc/amd7930.c
912
amd->op = op;
sound/sparc/amd7930.c
914
amd->regs = of_ioremap(&op->resource[0], 0,
sound/sparc/amd7930.c
915
resource_size(&op->resource[0]), "amd7930");
sound/sparc/amd7930.c
965
static int amd7930_sbus_probe(struct platform_device *op)
sound/sparc/amd7930.c
967
struct resource *rp = &op->resource[0];
sound/sparc/amd7930.c
973
irq = op->archdata.irqs[0];
sound/sparc/amd7930.c
982
err = snd_card_new(&op->dev, index[dev_num], id[dev_num],
sound/sparc/amd7930.c
995
if ((err = snd_amd7930_create(card, op,
sound/sparc/cs4231.c
114
struct platform_device *op;
sound/sparc/cs4231.c
1190
&chip->op->dev, 64 * 1024, 128 * 1024);
sound/sparc/cs4231.c
1483
static int cs4231_attach_begin(struct platform_device *op,
sound/sparc/cs4231.c
1500
err = snd_card_new(&op->dev, index[dev], id[dev], THIS_MODULE,
sound/sparc/cs4231.c
1536
dev_set_drvdata(&chip->op->dev, chip);
sound/sparc/cs4231.c
1683
struct platform_device *op = chip->op;
sound/sparc/cs4231.c
1689
of_iounmap(&op->resource[0], chip->port, chip->regs_size);
sound/sparc/cs4231.c
1706
struct platform_device *op,
sound/sparc/cs4231.c
1717
chip->op = op;
sound/sparc/cs4231.c
1718
chip->regs_size = resource_size(&op->resource[0]);
sound/sparc/cs4231.c
1722
chip->port = of_ioremap(&op->resource[0], 0,
sound/sparc/cs4231.c
1745
if (request_irq(op->archdata.irqs[0], snd_cs4231_sbus_interrupt,
sound/sparc/cs4231.c
1749
dev, op->archdata.irqs[0]);
sound/sparc/cs4231.c
1753
chip->irq[0] = op->archdata.irqs[0];
sound/sparc/cs4231.c
1771
static int cs4231_sbus_probe(struct platform_device *op)
sound/sparc/cs4231.c
1773
struct resource *rp = &op->resource[0];
sound/sparc/cs4231.c
1777
err = cs4231_attach_begin(op, &card);
sound/sparc/cs4231.c
1785
op->archdata.irqs[0]);
sound/sparc/cs4231.c
1787
err = snd_cs4231_sbus_create(card, op, dev);
sound/sparc/cs4231.c
1846
struct platform_device *op = chip->op;
sound/sparc/cs4231.c
1850
of_iounmap(&op->resource[2], chip->c_dma.ebus_info.regs, 0x10);
sound/sparc/cs4231.c
1854
of_iounmap(&op->resource[1], chip->p_dma.ebus_info.regs, 0x10);
sound/sparc/cs4231.c
1858
of_iounmap(&op->resource[0], chip->port, 0x10);
sound/sparc/cs4231.c
1875
struct platform_device *op,
sound/sparc/cs4231.c
1887
chip->op = op;
sound/sparc/cs4231.c
1894
chip->c_dma.ebus_info.irq = op->archdata.irqs[0];
sound/sparc/cs4231.c
1899
chip->p_dma.ebus_info.irq = op->archdata.irqs[1];
sound/sparc/cs4231.c
1911
chip->port = of_ioremap(&op->resource[0], 0, 0x10, "cs4231");
sound/sparc/cs4231.c
1913
of_ioremap(&op->resource[1], 0, 0x10, "cs4231_pdma");
sound/sparc/cs4231.c
1915
of_ioremap(&op->resource[2], 0, 0x10, "cs4231_cdma");
sound/sparc/cs4231.c
1969
static int cs4231_ebus_probe(struct platform_device *op)
sound/sparc/cs4231.c
1974
err = cs4231_attach_begin(op, &card);
sound/sparc/cs4231.c
1980
op->resource[0].start,
sound/sparc/cs4231.c
1981
op->archdata.irqs[0]);
sound/sparc/cs4231.c
1983
err = snd_cs4231_ebus_create(card, op, dev);
sound/sparc/cs4231.c
1993
static int cs4231_probe(struct platform_device *op)
sound/sparc/cs4231.c
1996
if (of_node_name_eq(op->dev.of_node->parent, "ebus"))
sound/sparc/cs4231.c
1997
return cs4231_ebus_probe(op);
sound/sparc/cs4231.c
2000
if (of_node_name_eq(op->dev.of_node->parent, "sbus") ||
sound/sparc/cs4231.c
2001
of_node_name_eq(op->dev.of_node->parent, "sbi"))
sound/sparc/cs4231.c
2002
return cs4231_sbus_probe(op);
sound/sparc/cs4231.c
2007
static void cs4231_remove(struct platform_device *op)
sound/sparc/cs4231.c
2009
struct snd_cs4231 *chip = dev_get_drvdata(&op->dev);
sound/sparc/dbri.c
2099
dma_map_single(&dbri->op->dev,
sound/sparc/dbri.c
2127
dma_unmap_single(&dbri->op->dev, info->dvma_buffer,
sound/sparc/dbri.c
2509
struct platform_device *op,
sound/sparc/dbri.c
2516
dbri->op = op;
sound/sparc/dbri.c
2519
dbri->dma = dma_alloc_coherent(&op->dev, sizeof(struct dbri_dma),
sound/sparc/dbri.c
2528
dbri->regs_size = resource_size(&op->resource[0]);
sound/sparc/dbri.c
2529
dbri->regs = of_ioremap(&op->resource[0], 0,
sound/sparc/dbri.c
2533
dma_free_coherent(&op->dev, sizeof(struct dbri_dma),
sound/sparc/dbri.c
2542
of_iounmap(&op->resource[0], dbri->regs, dbri->regs_size);
sound/sparc/dbri.c
2543
dma_free_coherent(&op->dev, sizeof(struct dbri_dma),
sound/sparc/dbri.c
2568
of_iounmap(&dbri->op->resource[0], dbri->regs, dbri->regs_size);
sound/sparc/dbri.c
2571
dma_free_coherent(&dbri->op->dev,
sound/sparc/dbri.c
2576
static int dbri_probe(struct platform_device *op)
sound/sparc/dbri.c
2592
irq = op->archdata.irqs[0];
sound/sparc/dbri.c
2598
err = snd_card_new(&op->dev, index[dev], id[dev], THIS_MODULE,
sound/sparc/dbri.c
2605
rp = &op->resource[0];
sound/sparc/dbri.c
2610
err = snd_dbri_create(card, op, irq, dev);
sound/sparc/dbri.c
2627
dev_set_drvdata(&op->dev, card);
sound/sparc/dbri.c
2635
dbri->irq, op->dev.of_node->name[9], dbri->mm.version);
sound/sparc/dbri.c
2646
static void dbri_remove(struct platform_device *op)
sound/sparc/dbri.c
2648
struct snd_card *card = dev_get_drvdata(&op->dev);
sound/sparc/dbri.c
304
struct platform_device *op; /* OF device info */
sound/xen/xen_snd_front.c
100
req->op.open.period_sz = period_sz;
sound/xen/xen_snd_front.c
101
req->op.open.gref_directory =
sound/xen/xen_snd_front.c
142
req->op.rw.length = count;
sound/xen/xen_snd_front.c
143
req->op.rw.offset = pos;
sound/xen/xen_snd_front.c
164
req->op.rw.length = count;
sound/xen/xen_snd_front.c
165
req->op.rw.offset = pos;
sound/xen/xen_snd_front.c
186
req->op.trigger.type = type;
sound/xen/xen_snd_front.c
69
req->op.hw_param = *hw_param_req;
sound/xen/xen_snd_front.c
96
req->op.open.pcm_format = format;
sound/xen/xen_snd_front.c
97
req->op.open.pcm_channels = channels;
sound/xen/xen_snd_front.c
98
req->op.open.pcm_rate = rate;
sound/xen/xen_snd_front.c
99
req->op.open.buffer_sz = buffer_sz;
sound/xen/xen_snd_front_evtchnl.c
118
event->op.cur_pos.position);
tools/arch/s390/include/uapi/asm/kvm.h
306
__u8 op; /* operation to perform */
tools/arch/s390/include/uapi/asm/kvm.h
65
__u32 op; /* type of operation */
tools/arch/x86/include/asm/rmwcc.h
15
#define GEN_UNARY_RMWcc(op, var, arg0, cc) \
tools/arch/x86/include/asm/rmwcc.h
16
__GEN_RMWcc(op " " arg0, var, cc)
tools/arch/x86/include/asm/rmwcc.h
18
#define GEN_BINARY_RMWcc(op, var, vcon, val, arg0, cc) \
tools/arch/x86/include/asm/rmwcc.h
19
__GEN_RMWcc(op " %1, " arg0, var, cc, vcon (val))
tools/arch/x86/lib/insn.c
280
insn_byte_t op;
tools/arch/x86/lib/insn.c
290
op = get_next(insn_byte_t, insn);
tools/arch/x86/lib/insn.c
291
insn_set_byte(opcode, 0, op);
tools/arch/x86/lib/insn.c
301
insn->attr = inat_get_xop_attribute(op, m);
tools/arch/x86/lib/insn.c
312
insn->attr = inat_get_avx_attribute(op, m, p);
tools/arch/x86/lib/insn.c
335
insn->attr = inat_get_escape_attribute(op, pfx_id, esc_attr);
tools/arch/x86/lib/insn.c
337
insn->attr = inat_get_opcode_attribute(op);
tools/arch/x86/lib/insn.c
342
insn->attr = inat_get_opcode_attribute(op);
tools/arch/x86/lib/insn.c
350
op = get_next(insn_byte_t, insn);
tools/arch/x86/lib/insn.c
351
opcode->bytes[opcode->nbytes++] = op;
tools/arch/x86/lib/insn.c
353
insn->attr = inat_get_escape_attribute(op, pfx_id, insn->attr);
tools/bpf/bpf_dbg.c
218
const char *op, *fmt;
tools/bpf/bpf_dbg.c
224
op = op_table[BPF_RET];
tools/bpf/bpf_dbg.c
228
op = op_table[BPF_RET];
tools/bpf/bpf_dbg.c
232
op = op_table[BPF_RET];
tools/bpf/bpf_dbg.c
236
op = op_table[BPF_MISC_TAX];
tools/bpf/bpf_dbg.c
240
op = op_table[BPF_MISC_TXA];
tools/bpf/bpf_dbg.c
244
op = op_table[BPF_ST];
tools/bpf/bpf_dbg.c
248
op = op_table[BPF_STX];
tools/bpf/bpf_dbg.c
252
op = op_table[BPF_LD_W];
tools/bpf/bpf_dbg.c
256
op = op_table[BPF_LD_H];
tools/bpf/bpf_dbg.c
260
op = op_table[BPF_LD_B];
tools/bpf/bpf_dbg.c
264
op = op_table[BPF_LD_W];
tools/bpf/bpf_dbg.c
268
op = op_table[BPF_LD_W];
tools/bpf/bpf_dbg.c
272
op = op_table[BPF_LD_H];
tools/bpf/bpf_dbg.c
276
op = op_table[BPF_LD_B];
tools/bpf/bpf_dbg.c
280
op = op_table[BPF_LD_W];
tools/bpf/bpf_dbg.c
284
op = op_table[BPF_LDX];
tools/bpf/bpf_dbg.c
288
op = op_table[BPF_LDX_B];
tools/bpf/bpf_dbg.c
292
op = op_table[BPF_LD_W];
tools/bpf/bpf_dbg.c
296
op = op_table[BPF_LDX];
tools/bpf/bpf_dbg.c
300
op = op_table[BPF_JMP_JA];
tools/bpf/bpf_dbg.c
305
op = op_table[BPF_JMP_JGT];
tools/bpf/bpf_dbg.c
309
op = op_table[BPF_JMP_JGT];
tools/bpf/bpf_dbg.c
313
op = op_table[BPF_JMP_JGE];
tools/bpf/bpf_dbg.c
317
op = op_table[BPF_JMP_JGE];
tools/bpf/bpf_dbg.c
321
op = op_table[BPF_JMP_JEQ];
tools/bpf/bpf_dbg.c
325
op = op_table[BPF_JMP_JEQ];
tools/bpf/bpf_dbg.c
329
op = op_table[BPF_JMP_JSET];
tools/bpf/bpf_dbg.c
333
op = op_table[BPF_JMP_JSET];
tools/bpf/bpf_dbg.c
337
op = op_table[BPF_ALU_NEG];
tools/bpf/bpf_dbg.c
341
op = op_table[BPF_ALU_LSH];
tools/bpf/bpf_dbg.c
345
op = op_table[BPF_ALU_LSH];
tools/bpf/bpf_dbg.c
349
op = op_table[BPF_ALU_RSH];
tools/bpf/bpf_dbg.c
353
op = op_table[BPF_ALU_RSH];
tools/bpf/bpf_dbg.c
357
op = op_table[BPF_ALU_ADD];
tools/bpf/bpf_dbg.c
361
op = op_table[BPF_ALU_ADD];
tools/bpf/bpf_dbg.c
365
op = op_table[BPF_ALU_SUB];
tools/bpf/bpf_dbg.c
369
op = op_table[BPF_ALU_SUB];
tools/bpf/bpf_dbg.c
373
op = op_table[BPF_ALU_MUL];
tools/bpf/bpf_dbg.c
377
op = op_table[BPF_ALU_MUL];
tools/bpf/bpf_dbg.c
381
op = op_table[BPF_ALU_DIV];
tools/bpf/bpf_dbg.c
385
op = op_table[BPF_ALU_DIV];
tools/bpf/bpf_dbg.c
389
op = op_table[BPF_ALU_MOD];
tools/bpf/bpf_dbg.c
393
op = op_table[BPF_ALU_MOD];
tools/bpf/bpf_dbg.c
397
op = op_table[BPF_ALU_AND];
tools/bpf/bpf_dbg.c
401
op = op_table[BPF_ALU_AND];
tools/bpf/bpf_dbg.c
405
op = op_table[BPF_ALU_OR];
tools/bpf/bpf_dbg.c
409
op = op_table[BPF_ALU_OR];
tools/bpf/bpf_dbg.c
413
op = op_table[BPF_ALU_XOR];
tools/bpf/bpf_dbg.c
417
op = op_table[BPF_ALU_XOR];
tools/bpf/bpf_dbg.c
421
op = "nosup";
tools/bpf/bpf_dbg.c
432
rl_printf("l%d:\t%s %s, l%d, l%d\n", i, op, buf,
tools/bpf/bpf_dbg.c
435
rl_printf("l%d:\t%s %s\n", i, op, buf);
tools/bpf/bpf_exp.y
43
static void bpf_set_curr_instr(uint16_t op, uint8_t jt, uint8_t jf, uint32_t k);
tools/bpf/bpftool/link.c
439
const char *hw_cache, *result, *op;
tools/bpf/bpftool/link.c
453
op = perf_event_name(evsel__hw_cache_op, (config >> 8) & 0xff);
tools/bpf/bpftool/link.c
454
if (op)
tools/bpf/bpftool/link.c
456
"%s-", op);
tools/build/feature/test-libdw.c
30
Dwarf_Op *op;
tools/build/feature/test-libdw.c
34
return (int)dwarf_getlocations(&attr, offset, &base, &start, &end, &op, &nops);
tools/hv/hv_kvp_daemon.c
1784
int op;
tools/hv/hv_kvp_daemon.c
1892
op = hv_msg->kvp_hdr.operation;
tools/hv/hv_kvp_daemon.c
1896
if ((in_hand_shake) && (op == KVP_OP_REGISTER1)) {
tools/hv/hv_kvp_daemon.c
1914
switch (op) {
tools/hv/hv_kvp_daemon.c
1973
if (op != KVP_OP_ENUMERATE)
tools/hv/hv_kvp_daemon.c
920
kvp_get_ip_info(int family, char *if_name, int op,
tools/hv/hv_kvp_daemon.c
937
if (op == KVP_OP_ENUMERATE) {
tools/hv/hv_kvp_daemon.c
989
if (op == KVP_OP_GET_IP_INFO) {
tools/hv/hv_vss_daemon.c
214
int op;
tools/hv/hv_vss_daemon.c
314
op = vss_msg->vss_hdr.operation;
tools/hv/hv_vss_daemon.c
317
switch (op) {
tools/hv/hv_vss_daemon.c
320
error = vss_operate(op);
tools/hv/hv_vss_daemon.c
322
op == VSS_OP_FREEZE ? "FREEZE" : "THAW",
tools/hv/hv_vss_daemon.c
327
syslog(LOG_ERR, "op=%d failed!", op);
tools/hv/hv_vss_daemon.c
336
syslog(LOG_ERR, "Illegal op:%d\n", op);
tools/include/io_uring/mini_liburing.h
259
static inline void io_uring_prep_cmd(struct io_uring_sqe *sqe, int op,
tools/include/io_uring/mini_liburing.h
268
sqe->cmd_op = op;
tools/include/linux/bitops.h
35
#define bitop(op, nr, addr) \
tools/include/linux/bitops.h
36
op(nr, addr)
tools/include/nolibc/sys/ptrace.h
22
long sys_ptrace(int op, pid_t pid, void *addr, void *data)
tools/include/nolibc/sys/ptrace.h
24
return my_syscall4(__NR_ptrace, op, pid, addr, data);
tools/include/nolibc/sys/ptrace.h
28
ssize_t ptrace(int op, pid_t pid, void *addr, void *data)
tools/include/nolibc/sys/ptrace.h
30
return __sysret(sys_ptrace(op, pid, addr, data));
tools/include/uapi/linux/bpf.h
6899
__u32 op;
tools/include/uapi/linux/io_uring.h
616
__u8 op;
tools/net/ynl/samples/netdev.c
19
static void netdev_print_device(struct netdev_dev_get_rsp *d, unsigned int op)
tools/net/ynl/samples/netdev.c
55
name = netdev_op_str(op);
tools/objtool/arch/loongarch/decode.c
129
struct stack_op *op)
tools/objtool/arch/loongarch/decode.c
136
ADD_OP(op) {
tools/objtool/arch/loongarch/decode.c
137
op->src.type = OP_SRC_ADD;
tools/objtool/arch/loongarch/decode.c
138
op->src.reg = inst.reg2i12_format.rj;
tools/objtool/arch/loongarch/decode.c
139
op->src.offset = insn->immediate;
tools/objtool/arch/loongarch/decode.c
140
op->dest.type = OP_DEST_REG;
tools/objtool/arch/loongarch/decode.c
141
op->dest.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
158
ADD_OP(op) {
tools/objtool/arch/loongarch/decode.c
159
op->src.type = OP_SRC_REG_INDIRECT;
tools/objtool/arch/loongarch/decode.c
160
op->src.reg = CFI_SP;
tools/objtool/arch/loongarch/decode.c
161
op->src.offset = insn->immediate;
tools/objtool/arch/loongarch/decode.c
162
op->dest.type = OP_DEST_REG;
tools/objtool/arch/loongarch/decode.c
163
op->dest.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
171
ADD_OP(op) {
tools/objtool/arch/loongarch/decode.c
172
op->src.type = OP_SRC_REG;
tools/objtool/arch/loongarch/decode.c
173
op->src.reg = inst.reg2i12_format.rd;
tools/objtool/arch/loongarch/decode.c
174
op->dest.type = OP_DEST_REG_INDIRECT;
tools/objtool/arch/loongarch/decode.c
175
op->dest.reg = CFI_SP;
tools/objtool/arch/loongarch/decode.c
176
op->dest.offset = insn->immediate;
tools/objtool/arch/loongarch/decode.c
197
struct stack_op *op)
tools/objtool/arch/loongarch/decode.c
204
ADD_OP(op) {
tools/objtool/arch/loongarch/decode.c
205
op->src.type = OP_SRC_REG_INDIRECT;
tools/objtool/arch/loongarch/decode.c
206
op->src.reg = CFI_SP;
tools/objtool/arch/loongarch/decode.c
207
op->src.offset = insn->immediate;
tools/objtool/arch/loongarch/decode.c
208
op->dest.type = OP_DEST_REG;
tools/objtool/arch/loongarch/decode.c
209
op->dest.reg = inst.reg2i14_format.rd;
tools/objtool/arch/loongarch/decode.c
222
ADD_OP(op) {
tools/objtool/arch/loongarch/decode.c
223
op->src.type = OP_SRC_REG;
tools/objtool/arch/loongarch/decode.c
224
op->src.reg = inst.reg2i14_format.rd;
tools/objtool/arch/loongarch/decode.c
225
op->dest.type = OP_DEST_REG_INDIRECT;
tools/objtool/arch/loongarch/decode.c
226
op->dest.reg = CFI_SP;
tools/objtool/arch/loongarch/decode.c
227
op->dest.offset = insn->immediate;
tools/objtool/arch/loongarch/decode.c
318
struct stack_op *op = NULL;
tools/objtool/arch/loongarch/decode.c
337
if (decode_insn_reg2i12_fomat(inst, insn, ops_list, op))
tools/objtool/arch/loongarch/decode.c
339
if (decode_insn_reg2i14_fomat(inst, insn, ops_list, op))
tools/objtool/arch/loongarch/decode.c
82
#define ADD_OP(op) \
tools/objtool/arch/loongarch/decode.c
83
if (!(op = calloc(1, sizeof(*op)))) \
tools/objtool/arch/loongarch/decode.c
85
else for (*ops_list = op, ops_list = &op->next; op; op = NULL)
tools/objtool/arch/x86/decode.c
169
#define ADD_OP(op) \
tools/objtool/arch/x86/decode.c
170
if (!(op = calloc(1, sizeof(*op)))) \
tools/objtool/arch/x86/decode.c
172
else for (*ops_list = op, ops_list = &op->next; op; op = NULL)
tools/objtool/arch/x86/decode.c
229
struct stack_op *op = NULL;
tools/objtool/arch/x86/decode.c
285
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
286
op->src.type = OP_SRC_ADD;
tools/objtool/arch/x86/decode.c
287
op->src.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
288
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
289
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
297
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
298
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
299
op->src.reg = (op1 & 0x7) + 8*rex_b;
tools/objtool/arch/x86/decode.c
300
op->dest.type = OP_DEST_PUSH;
tools/objtool/arch/x86/decode.c
308
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
309
op->src.type = OP_SRC_POP;
tools/objtool/arch/x86/decode.c
310
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
311
op->dest.reg = (op1 & 0x7) + 8*rex_b;
tools/objtool/arch/x86/decode.c
319
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
320
op->src.type = OP_SRC_CONST;
tools/objtool/arch/x86/decode.c
321
op->dest.type = OP_DEST_PUSH;
tools/objtool/arch/x86/decode.c
367
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
368
op->src.type = OP_SRC_ADD;
tools/objtool/arch/x86/decode.c
369
op->src.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
370
op->src.offset = imm;
tools/objtool/arch/x86/decode.c
371
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
372
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
378
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
379
op->src.type = OP_SRC_AND;
tools/objtool/arch/x86/decode.c
380
op->src.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
381
op->src.offset = ins.immediate.value;
tools/objtool/arch/x86/decode.c
382
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
383
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
400
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
401
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
402
op->src.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
403
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
404
op->dest.reg = modrm_rm;
tools/objtool/arch/x86/decode.c
422
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
423
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
424
op->src.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
425
op->dest.type = OP_DEST_REG_INDIRECT;
tools/objtool/arch/x86/decode.c
426
op->dest.reg = modrm_rm;
tools/objtool/arch/x86/decode.c
427
op->dest.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
440
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
441
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
442
op->src.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
443
op->dest.type = OP_DEST_REG_INDIRECT;
tools/objtool/arch/x86/decode.c
444
op->dest.reg = CFI_BP;
tools/objtool/arch/x86/decode.c
445
op->dest.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
453
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
454
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
455
op->src.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
456
op->dest.type = OP_DEST_REG_INDIRECT;
tools/objtool/arch/x86/decode.c
457
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
458
op->dest.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
472
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
473
op->src.type = OP_SRC_REG_INDIRECT;
tools/objtool/arch/x86/decode.c
474
op->src.reg = CFI_BP;
tools/objtool/arch/x86/decode.c
475
op->src.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
476
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
477
op->dest.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
485
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
486
op->src.type = OP_SRC_REG_INDIRECT;
tools/objtool/arch/x86/decode.c
487
op->src.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
488
op->src.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
489
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
490
op->dest.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
521
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
522
op->src.offset = ins.displacement.value;
tools/objtool/arch/x86/decode.c
523
if (!op->src.offset) {
tools/objtool/arch/x86/decode.c
525
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
528
op->src.type = OP_SRC_ADD;
tools/objtool/arch/x86/decode.c
530
op->src.reg = modrm_rm;
tools/objtool/arch/x86/decode.c
531
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
532
op->dest.reg = modrm_reg;
tools/objtool/arch/x86/decode.c
538
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
539
op->src.type = OP_SRC_POP;
tools/objtool/arch/x86/decode.c
540
op->dest.type = OP_DEST_MEM;
tools/objtool/arch/x86/decode.c
556
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
557
op->src.type = OP_SRC_CONST;
tools/objtool/arch/x86/decode.c
558
op->dest.type = OP_DEST_PUSHF;
tools/objtool/arch/x86/decode.c
564
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
565
op->src.type = OP_SRC_POPF;
tools/objtool/arch/x86/decode.c
566
op->dest.type = OP_DEST_MEM;
tools/objtool/arch/x86/decode.c
629
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
630
op->src.type = OP_SRC_CONST;
tools/objtool/arch/x86/decode.c
631
op->dest.type = OP_DEST_PUSH;
tools/objtool/arch/x86/decode.c
637
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
638
op->src.type = OP_SRC_POP;
tools/objtool/arch/x86/decode.c
639
op->dest.type = OP_DEST_MEM;
tools/objtool/arch/x86/decode.c
653
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
654
op->src.type = OP_SRC_REG;
tools/objtool/arch/x86/decode.c
655
op->src.reg = CFI_BP;
tools/objtool/arch/x86/decode.c
656
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
657
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
659
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
660
op->src.type = OP_SRC_POP;
tools/objtool/arch/x86/decode.c
661
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
662
op->dest.reg = CFI_BP;
tools/objtool/arch/x86/decode.c
727
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
729
op->src.type = OP_SRC_ADD;
tools/objtool/arch/x86/decode.c
730
op->src.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
731
op->src.offset = 5*8;
tools/objtool/arch/x86/decode.c
732
op->dest.type = OP_DEST_REG;
tools/objtool/arch/x86/decode.c
733
op->dest.reg = CFI_SP;
tools/objtool/arch/x86/decode.c
761
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
762
op->src.type = OP_SRC_CONST;
tools/objtool/arch/x86/decode.c
763
op->dest.type = OP_DEST_PUSH;
tools/objtool/arch/x86/decode.c
796
ADD_OP(op) {
tools/objtool/arch/x86/decode.c
797
op->src.type = OP_SRC_CONST;
tools/objtool/arch/x86/decode.c
798
op->dest.type = OP_DEST_PUSH;
tools/objtool/check.c
1374
struct stack_op *op, *next;
tools/objtool/check.c
1376
for (op = insn->stack_ops; op; op = next) {
tools/objtool/check.c
1377
next = op->next;
tools/objtool/check.c
1378
free(op);
tools/objtool/check.c
2768
struct stack_op *op)
tools/objtool/check.c
2776
if (op->dest.type == OP_DEST_PUSH || op->dest.type == OP_DEST_PUSHF)
tools/objtool/check.c
2780
if (op->src.type == OP_SRC_POP || op->src.type == OP_SRC_POPF)
tools/objtool/check.c
2784
if (op->dest.type == OP_DEST_REG && op->src.type == OP_SRC_ADD &&
tools/objtool/check.c
2785
op->dest.reg == CFI_SP && op->src.reg == CFI_SP)
tools/objtool/check.c
2786
cfa->offset -= op->src.offset;
tools/objtool/check.c
2861
struct cfi_state *cfi, struct stack_op *op)
tools/objtool/check.c
2881
return update_cfi_state_regs(insn, cfi, op);
tools/objtool/check.c
2883
switch (op->dest.type) {
tools/objtool/check.c
2886
switch (op->src.type) {
tools/objtool/check.c
2889
if (op->src.reg == CFI_SP && op->dest.reg == CFI_BP &&
tools/objtool/check.c
2894
cfa->base = op->dest.reg;
tools/objtool/check.c
2898
else if (op->src.reg == CFI_SP &&
tools/objtool/check.c
2899
op->dest.reg == CFI_BP && cfi->drap) {
tools/objtool/check.c
2907
else if (op->src.reg == CFI_SP && cfa->base == CFI_SP) {
tools/objtool/check.c
2919
cfi->vals[op->dest.reg].base = CFI_CFA;
tools/objtool/check.c
2920
cfi->vals[op->dest.reg].offset = -cfi->stack_size;
tools/objtool/check.c
2923
else if (op->src.reg == CFI_BP && op->dest.reg == CFI_SP &&
tools/objtool/check.c
2934
else if (op->dest.reg == cfa->base) {
tools/objtool/check.c
2938
cfi->vals[op->src.reg].base == CFI_CFA) {
tools/objtool/check.c
2948
cfa->offset = -cfi->vals[op->src.reg].offset;
tools/objtool/check.c
2952
cfi->vals[op->src.reg].base == CFI_SP_INDIRECT &&
tools/objtool/check.c
2953
cfi->vals[op->src.reg].offset == cfa->offset) {
tools/objtool/check.c
2989
else if (op->dest.reg == CFI_SP &&
tools/objtool/check.c
2990
cfi->vals[op->src.reg].base == CFI_SP_INDIRECT &&
tools/objtool/check.c
2991
cfi->vals[op->src.reg].offset == cfa->offset) {
tools/objtool/check.c
3002
else if (cfi->vals[op->src.reg].base == CFI_CFA) {
tools/objtool/check.c
3012
cfi->vals[op->dest.reg].base = CFI_CFA;
tools/objtool/check.c
3013
cfi->vals[op->dest.reg].offset = cfi->vals[op->src.reg].offset;
tools/objtool/check.c
3020
if (op->dest.reg == CFI_SP && op->src.reg == CFI_SP) {
tools/objtool/check.c
3023
cfi->stack_size -= op->src.offset;
tools/objtool/check.c
3025
cfa->offset -= op->src.offset;
tools/objtool/check.c
3029
if (op->dest.reg == CFI_BP && op->src.reg == CFI_SP &&
tools/objtool/check.c
3032
if (cfa->base == CFI_SP && cfa->offset == op->src.offset) {
tools/objtool/check.c
3039
if (op->dest.reg == CFI_SP && op->src.reg == CFI_BP) {
tools/objtool/check.c
3044
cfa->offset = -op->src.offset;
tools/objtool/check.c
3048
cfi->stack_size = -(op->src.offset + regs[CFI_BP].offset);
tools/objtool/check.c
3053
if (op->src.reg == CFI_SP && cfa->base == CFI_SP) {
tools/objtool/check.c
3056
cfi->drap_reg = op->dest.reg;
tools/objtool/check.c
3068
cfi->vals[op->dest.reg].base = CFI_CFA;
tools/objtool/check.c
3069
cfi->vals[op->dest.reg].offset = \
tools/objtool/check.c
3070
-cfi->stack_size + op->src.offset;
tools/objtool/check.c
3075
if (cfi->drap && op->dest.reg == CFI_SP &&
tools/objtool/check.c
3076
op->src.reg == cfi->drap_reg) {
tools/objtool/check.c
3080
cfa->offset = cfi->stack_size = -op->src.offset;
tools/objtool/check.c
3086
if (op->dest.reg == cfi->cfa.base && !(next_insn && next_insn->hint)) {
tools/objtool/check.c
3094
if (op->dest.reg != CFI_SP ||
tools/objtool/check.c
3117
if (op->dest.reg == CFI_SP && cfa->base == CFI_SP_INDIRECT) {
tools/objtool/check.c
3124
if (!cfi->drap && op->dest.reg == cfa->base) {
tools/objtool/check.c
3131
op->dest.reg == cfi->drap_reg &&
tools/objtool/check.c
3139
} else if (cfi->stack_size == -regs[op->dest.reg].offset) {
tools/objtool/check.c
3142
restore_reg(cfi, op->dest.reg);
tools/objtool/check.c
3152
if (!cfi->drap && op->dest.reg == cfa->base &&
tools/objtool/check.c
3153
op->dest.reg == CFI_BP) {
tools/objtool/check.c
3160
if (cfi->drap && op->src.reg == CFI_BP &&
tools/objtool/check.c
3161
op->src.offset == cfi->drap_offset) {
tools/objtool/check.c
3169
if (cfi->drap && op->src.reg == CFI_BP &&
tools/objtool/check.c
3170
op->src.offset == regs[op->dest.reg].offset) {
tools/objtool/check.c
3173
restore_reg(cfi, op->dest.reg);
tools/objtool/check.c
3175
} else if (op->src.reg == cfa->base &&
tools/objtool/check.c
3176
op->src.offset == regs[op->dest.reg].offset + cfa->offset) {
tools/objtool/check.c
3180
restore_reg(cfi, op->dest.reg);
tools/objtool/check.c
3182
} else if (op->src.reg == CFI_SP &&
tools/objtool/check.c
3183
op->src.offset == regs[op->dest.reg].offset + cfi->stack_size) {
tools/objtool/check.c
3186
restore_reg(cfi, op->dest.reg);
tools/objtool/check.c
3204
if (op->src.type != OP_SRC_REG)
tools/objtool/check.c
3208
if (op->src.reg == cfa->base && op->src.reg == cfi->drap_reg) {
tools/objtool/check.c
3217
} else if (op->src.reg == CFI_BP && cfa->base == cfi->drap_reg) {
tools/objtool/check.c
3225
save_reg(cfi, op->src.reg, CFI_BP, -cfi->stack_size);
tools/objtool/check.c
3231
save_reg(cfi, op->src.reg, CFI_CFA, -cfi->stack_size);
tools/objtool/check.c
3235
if (opts.stackval && insn_func(insn) && op->src.reg == CFI_BP &&
tools/objtool/check.c
3243
if (op->src.reg == cfa->base && op->src.reg == cfi->drap_reg) {
tools/objtool/check.c
3247
cfa->offset = op->dest.offset;
tools/objtool/check.c
3250
cfi->drap_offset = op->dest.offset;
tools/objtool/check.c
3254
save_reg(cfi, op->src.reg, CFI_BP, op->dest.offset);
tools/objtool/check.c
3257
} else if (op->dest.reg == cfa->base) {
tools/objtool/check.c
3261
save_reg(cfi, op->src.reg, CFI_CFA,
tools/objtool/check.c
3262
op->dest.offset - cfi->cfa.offset);
tools/objtool/check.c
3264
} else if (op->dest.reg == CFI_SP) {
tools/objtool/check.c
3267
save_reg(cfi, op->src.reg, CFI_CFA,
tools/objtool/check.c
3268
op->dest.offset - cfi->stack_size);
tools/objtool/check.c
3270
} else if (op->src.reg == CFI_SP && op->dest.offset == 0) {
tools/objtool/check.c
3273
cfi->vals[op->dest.reg].base = CFI_SP_INDIRECT;
tools/objtool/check.c
3274
cfi->vals[op->dest.reg].offset = cfa->offset;
tools/objtool/check.c
3280
if (op->src.type != OP_SRC_POP && op->src.type != OP_SRC_POPF) {
tools/objtool/check.c
3345
struct stack_op *op;
tools/objtool/check.c
3348
for (op = insn->stack_ops; op; op = op->next) {
tools/objtool/check.c
3350
ret = update_cfi_state(insn, next_insn, &state->cfi, op);
tools/objtool/check.c
3357
if (op->dest.type == OP_DEST_PUSHF) {
tools/objtool/check.c
3369
if (op->src.type == OP_SRC_POPF) {
tools/perf/arch/x86/tests/insn-x86.c
103
int op, branch, ret;
tools/perf/arch/x86/tests/insn-x86.c
119
op = get_op(dat->expected_op_str);
tools/perf/arch/x86/tests/insn-x86.c
127
if ((int)intel_pt_insn.op != op) {
tools/perf/arch/x86/tests/insn-x86.c
129
intel_pt_insn.op, op, dat->asm_rep);
tools/perf/arch/x86/util/cpuid.h
29
: "a"(op), "2"(op2));
tools/perf/arch/x86/util/cpuid.h
7
cpuid(unsigned int op, unsigned int op2, unsigned int *a, unsigned int *b,
tools/perf/bench/epoll-ctl.c
132
static inline void do_epoll_op(struct worker *w, int op, int fd)
tools/perf/bench/epoll-ctl.c
140
switch (op) {
tools/perf/bench/epoll-ctl.c
157
w->ops[op]++;
tools/perf/bench/epoll-ctl.c
163
int op, fd;
tools/perf/bench/epoll-ctl.c
166
op = rnd2 % EPOLL_NR_OPS;
tools/perf/bench/epoll-ctl.c
168
do_epoll_op(w, op, fd);
tools/perf/bench/futex.h
51
futex_syscall(volatile u_int32_t *uaddr, int op, u_int32_t val, struct timespec *timeout,
tools/perf/bench/futex.h
54
return syscall(SYS_futex, uaddr, op | opflags, val, timeout, uaddr2, val3);
tools/perf/bench/futex.h
58
futex_syscall_nr_requeue(volatile u_int32_t *uaddr, int op, u_int32_t val, int nr_requeue,
tools/perf/bench/futex.h
61
return syscall(SYS_futex, uaddr, op | opflags, val, nr_requeue, uaddr2, val3);
tools/perf/bench/sched-seccomp-notify.c
44
static int seccomp(unsigned int op, unsigned int flags, void *args)
tools/perf/bench/sched-seccomp-notify.c
46
return syscall(__NR_seccomp, op, flags, args);
tools/perf/builtin-script.c
1745
static int sample__fprintf_bpf_output(enum binary_printer_ops op,
tools/perf/builtin-script.c
1753
switch (op) {
tools/perf/builtin-trace.c
1179
.arg = { [1] = STRARRAY(op, epoll_ctl_ops), }, },
tools/perf/builtin-trace.c
3152
static int bpf_output__printer(enum binary_printer_ops op,
tools/perf/builtin-trace.c
3157
switch (op) {
tools/perf/tests/evsel-roundtrip-name.c
14
for (int op = 0; op < PERF_COUNT_HW_CACHE_OP_MAX; op++) {
tools/perf/tests/evsel-roundtrip-name.c
16
if (!evsel__is_cache_op_valid(type, op))
tools/perf/tests/evsel-roundtrip-name.c
29
__evsel__hw_cache_type_op_res_name(type, op, res,
tools/perf/trace/beauty/flock.c
27
int printed = 0, op = arg->val;
tools/perf/trace/beauty/flock.c
29
if (op == 0)
tools/perf/trace/beauty/flock.c
32
if ((op & LOCK_##cmd) == LOCK_##cmd) { \
tools/perf/trace/beauty/flock.c
34
op &= ~LOCK_##cmd; \
tools/perf/trace/beauty/flock.c
47
if (op)
tools/perf/trace/beauty/flock.c
48
printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", op);
tools/perf/trace/beauty/futex_op.c
32
int op = arg->val;
tools/perf/trace/beauty/futex_op.c
33
int cmd = op & FUTEX_CMD_MASK;
tools/perf/trace/beauty/futex_op.c
54
if (op & FUTEX_PRIVATE_FLAG)
tools/perf/trace/beauty/futex_op.c
57
if (op & FUTEX_CLOCK_REALTIME)
tools/perf/trace/beauty/seccomp.c
13
int op = arg->val;
tools/perf/trace/beauty/seccomp.c
16
switch (op) {
tools/perf/trace/beauty/seccomp.c
21
default: printed = scnprintf(bf, size, "%#x", op); break;
tools/perf/ui/browsers/hists.c
1287
__HPP_COLOR_MEM_STAT_FN(op, OP)
tools/perf/ui/hist.c
655
HPP_MEM_STAT_FNS(op, OP)
tools/perf/ui/hist.c
763
HPP__MEM_STAT_PRINT_FNS("Mem Op", op, OP),
tools/perf/util/annotate-arch/annotate-powerpc.c
9
#define PPC_OP(op) (((op) >> 26) & 0x3F)
tools/perf/util/annotate-data.c
1075
dloc->op->segment = INSN_SEG_X86_GS;
tools/perf/util/annotate-data.c
1076
dloc->op->imm = true;
tools/perf/util/annotate-data.c
1077
dloc->op->offset = 40;
tools/perf/util/annotate-data.c
1094
int reg = dloc->op->reg1;
tools/perf/util/annotate-data.c
1095
int offset = dloc->op->offset;
tools/perf/util/annotate-data.c
1126
if (dloc->op->offset < 0 && reg != state->stack_reg)
tools/perf/util/annotate-data.c
1136
dloc->type_offset = dloc->op->offset + state->regs[reg].offset;
tools/perf/util/annotate-data.c
1165
dloc->type_offset = dloc->op->offset + state->regs[reg].offset;
tools/perf/util/annotate-data.c
1184
dloc->type_offset = dloc->op->offset;
tools/perf/util/annotate-data.c
1208
u64 var_addr = dloc->op->offset;
tools/perf/util/annotate-data.c
1213
if (dloc->op->multi_regs) {
tools/perf/util/annotate-data.c
1214
int reg2 = dloc->op->reg2;
tools/perf/util/annotate-data.c
1216
if (dloc->op->reg2 == reg)
tools/perf/util/annotate-data.c
1217
reg2 = dloc->op->reg1;
tools/perf/util/annotate-data.c
1313
if (dloc->op->segment == INSN_SEG_X86_GS && dloc->op->imm &&
tools/perf/util/annotate-data.c
1317
addr = dloc->op->offset;
tools/perf/util/annotate-data.c
1328
if (dloc->op->offset < 0 && reg != state->stack_reg) {
tools/perf/util/annotate-data.c
1329
addr = (s64) dloc->op->offset;
tools/perf/util/annotate-data.c
1463
int offset = dloc->op->offset;
tools/perf/util/annotate-data.c
1471
if (dloc->op->multi_regs)
tools/perf/util/annotate-data.c
1473
dloc->op->reg1, dloc->op->reg2);
tools/perf/util/annotate-data.c
1475
snprintf(buf, sizeof(buf), "reg%d", dloc->op->reg1);
tools/perf/util/annotate-data.c
1498
struct annotated_op_loc *loc = dloc->op;
tools/perf/util/annotate-data.c
1513
if (dloc->op->multi_regs)
tools/perf/util/annotate-data.c
1514
snprintf(buf, sizeof(buf), "reg%d, reg%d", dloc->op->reg1, dloc->op->reg2);
tools/perf/util/annotate-data.c
1515
else if (dloc->op->reg1 == DWARF_REG_PC)
tools/perf/util/annotate-data.c
1518
snprintf(buf, sizeof(buf), "reg%d", dloc->op->reg1);
tools/perf/util/annotate-data.c
1715
dloc->type_offset = dloc->op->offset;
tools/perf/util/annotate-data.h
126
struct annotated_op_loc *op;
tools/perf/util/annotate.c
2837
.op = op_loc,
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
198
decoder->record.op |= ARM_SPE_OP_LDST;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
200
decoder->record.op |= ARM_SPE_OP_ST;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
202
decoder->record.op |= ARM_SPE_OP_LD;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
205
decoder->record.op |= ARM_SPE_OP_GP_REG;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
207
decoder->record.op |= ARM_SPE_OP_SIMD_FP;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
209
decoder->record.op |= ARM_SPE_OP_UNSPEC_REG;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
211
decoder->record.op |= ARM_SPE_OP_NV_SYSREG;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
213
decoder->record.op |= ARM_SPE_OP_MTE_TAG;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
216
decoder->record.op |= ARM_SPE_OP_AR;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
218
decoder->record.op |= ARM_SPE_OP_EXCL;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
220
decoder->record.op |= ARM_SPE_OP_ATOMIC;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
222
decoder->record.op |= ARM_SPE_OP_SVE;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
224
decoder->record.op |= ARM_SPE_OP_PRED;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
226
decoder->record.op |= ARM_SPE_OP_SG;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
228
decoder->record.op |= ARM_SPE_OP_MEMCPY;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
230
decoder->record.op |= ARM_SPE_OP_MEMSET;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
232
decoder->record.op |= ARM_SPE_OP_GCS;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
234
decoder->record.op |= ARM_SPE_OP_COMM;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
239
decoder->record.op |= ARM_SPE_OP_OTHER;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
241
decoder->record.op |= ARM_SPE_OP_SVE | ARM_SPE_OP_DP;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
243
decoder->record.op |= ARM_SPE_OP_FP;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
245
decoder->record.op |= ARM_SPE_OP_PRED;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
247
decoder->record.op |= ARM_SPE_OP_SME;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
249
decoder->record.op |= ARM_SPE_OP_FP;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
252
decoder->record.op |= ARM_SPE_OP_ASE;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
254
decoder->record.op |= ARM_SPE_OP_FP;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
256
decoder->record.op |= ARM_SPE_OP_COND;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
260
decoder->record.op |= ARM_SPE_OP_BRANCH_ERET;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
262
decoder->record.op |= ARM_SPE_OP_BR_COND;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
264
decoder->record.op |= ARM_SPE_OP_BR_INDIRECT;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
266
decoder->record.op |= ARM_SPE_OP_BR_GCS;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
268
decoder->record.op |= ARM_SPE_OP_BR_CR_BL;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
270
decoder->record.op |= ARM_SPE_OP_BR_CR_RET;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
272
decoder->record.op |= ARM_SPE_OP_BR_CR_NON_BL_RET;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.h
117
u32 op;
tools/perf/util/arm-spe.c
1000
if (!is_mem_op(record->op))
tools/perf/util/arm-spe.c
1003
if (record->op & ARM_SPE_OP_LD)
tools/perf/util/arm-spe.c
1005
else if (record->op & ARM_SPE_OP_ST)
tools/perf/util/arm-spe.c
1095
if (spe->sample_branch && (record->op & ARM_SPE_OP_BRANCH_ERET)) {
tools/perf/util/arm-spe.c
1109
if (spe->sample_memory && is_mem_op(record->op)) {
tools/perf/util/arm-spe.c
354
if (record->op & ARM_SPE_OP_SVE)
tools/perf/util/arm-spe.c
40
#define is_ldst_op(op) (!!((op) & ARM_SPE_OP_LDST))
tools/perf/util/arm-spe.c
415
if (record->op & ARM_SPE_OP_BR_CR_BL) {
tools/perf/util/arm-spe.c
416
if (record->op & ARM_SPE_OP_BR_COND)
tools/perf/util/arm-spe.c
42
#define is_simd_op(op) (!!((op) & (ARM_SPE_OP_SIMD_FP | ARM_SPE_OP_SVE | \
tools/perf/util/arm-spe.c
424
} else if (record->op & ARM_SPE_OP_BR_CR_RET ||
tools/perf/util/arm-spe.c
425
record->op & ARM_SPE_OP_BR_INDIRECT) {
tools/perf/util/arm-spe.c
426
if (record->op & ARM_SPE_OP_BR_COND)
tools/perf/util/arm-spe.c
430
} else if (record->op & ARM_SPE_OP_BR_CR_NON_BL_RET) {
tools/perf/util/arm-spe.c
431
if (record->op & ARM_SPE_OP_BR_COND)
tools/perf/util/arm-spe.c
436
if (record->op & ARM_SPE_OP_BR_COND)
tools/perf/util/arm-spe.c
45
#define is_mem_op(op) (is_ldst_op(op) || is_simd_op(op))
tools/perf/util/arm-spe.c
609
if (record->op & ARM_SPE_OP_BRANCH_ERET) {
tools/perf/util/arm-spe.c
621
if (record->op & ARM_SPE_OP_BR_COND)
tools/perf/util/arm-spe.c
624
if (record->op & ARM_SPE_OP_BR_CR_BL)
tools/perf/util/arm-spe.c
626
else if (record->op & ARM_SPE_OP_BR_CR_RET)
tools/perf/util/arm-spe.c
632
else if (record->op & ARM_SPE_OP_BR_INDIRECT)
tools/perf/util/arm-spe.c
652
if (record->op & ARM_SPE_OP_ST) {
tools/perf/util/arm-spe.c
747
common_record.op = record->op;
tools/perf/util/arm-spe.c
755
if (record->op & ARM_SPE_OP_ST) {
tools/perf/util/bpf-filter.c
134
if (expr->op == PBF_OP_GROUP_BEGIN) {
tools/perf/util/bpf-filter.c
169
entry[i].op = expr->op;
tools/perf/util/bpf-filter.c
175
if (expr->op == PBF_OP_GROUP_BEGIN) {
tools/perf/util/bpf-filter.c
182
entry[i].op = group->op;
tools/perf/util/bpf-filter.c
192
entry[i].op = PBF_OP_GROUP_END;
tools/perf/util/bpf-filter.c
199
entry[i].op = PBF_OP_DONE;
tools/perf/util/bpf-filter.c
614
enum perf_bpf_filter_op op,
tools/perf/util/bpf-filter.c
623
expr->op = op;
tools/perf/util/bpf-filter.h
13
enum perf_bpf_filter_op op;
tools/perf/util/bpf-filter.h
28
enum perf_bpf_filter_op op,
tools/perf/util/bpf-filter.y
38
enum perf_bpf_filter_op op;
tools/perf/util/bpf-filter.y
46
%type <op> BFT_OP
tools/perf/util/bpf-filter.y
68
if ($1->op == PBF_OP_GROUP_BEGIN) {
tools/perf/util/bpf_skel/sample-filter.h
60
enum perf_bpf_filter_op op;
tools/perf/util/bpf_skel/sample_filter.bpf.c
195
#define CHECK_RESULT(data, op, val) \
tools/perf/util/bpf_skel/sample_filter.bpf.c
196
if (!(data op val)) { \
tools/perf/util/bpf_skel/sample_filter.bpf.c
251
switch (entry[i].op) {
tools/perf/util/capstone.c
195
struct cs_x86_op *op = &insn->detail->x86.operands[0];
tools/perf/util/capstone.c
198
if (op->type == X86_OP_IMM &&
tools/perf/util/capstone.c
199
thread__find_symbol(thread, cpumode, op->imm, &al)) {
tools/perf/util/capstone.c
203
printed += fprintf(fp, " [%#" PRIx64 "]", op->imm);
tools/perf/util/capstone.c
266
struct cs_x86_op *op = &insn->detail->x86.operands[i];
tools/perf/util/capstone.c
269
if (op->type != X86_OP_MEM)
tools/perf/util/capstone.c
273
if (op->mem.base != X86_REG_RIP)
tools/perf/util/capstone.c
277
orig_addr = addr + insn->size + op->mem.disp;
tools/perf/util/cpu-set-sched.h
34
#define __CPU_OP_S(setsize, destset, srcset1, srcset2, op) \
tools/perf/util/cpu-set-sched.h
42
((__cpu_mask *) __dest->__bits)[__i] = __arr1[__i] op __arr2[__i]; \
tools/perf/util/debug.c
176
static int trace_event_printer(enum binary_printer_ops op,
tools/perf/util/debug.c
184
switch (op) {
tools/perf/util/disasm.c
568
static bool check_multi_regs(const struct arch *arch, const char *op)
tools/perf/util/disasm.c
576
op = strchr(op, arch->objdump.memory_ref_char);
tools/perf/util/disasm.c
577
if (op == NULL)
tools/perf/util/disasm.c
581
while ((op = strchr(op, arch->objdump.register_char)) != NULL) {
tools/perf/util/disasm.c
583
op++;
tools/perf/util/disasm.c
867
#define PPC_OP(op) (((op) >> 26) & 0x3F)
tools/perf/util/dwarf-aux.c
1185
static int reg_from_dwarf_op(Dwarf_Op *op)
tools/perf/util/dwarf-aux.c
1187
switch (op->atom) {
tools/perf/util/dwarf-aux.c
1189
return op->atom - DW_OP_reg0;
tools/perf/util/dwarf-aux.c
1191
return op->atom - DW_OP_breg0;
tools/perf/util/dwarf-aux.c
1194
return op->number;
tools/perf/util/dwarf-aux.c
1203
static int offset_from_dwarf_op(Dwarf_Op *op)
tools/perf/util/dwarf-aux.c
1205
switch (op->atom) {
tools/perf/util/dwarf-aux.c
1211
return op->number;
tools/perf/util/dwarf-aux.c
1213
return op->number2;
tools/perf/util/dwarf-aux.c
1326
Dwarf_Op *op;
tools/perf/util/dwarf-aux.c
1345
&start, &end, &op, &nops)) > 0) {
tools/perf/util/dwarf-regs-arch/dwarf-regs-powerpc.c
11
#define PPC_OP(op) (((op) >> 26) & 0x3F)
tools/perf/util/event.c
529
static int text_poke_printer(enum binary_printer_ops op, unsigned int val,
tools/perf/util/event.c
534
switch ((int)op) {
tools/perf/util/evsel.c
842
bool evsel__is_cache_op_valid(u8 type, u8 op)
tools/perf/util/evsel.c
844
if (evsel__hw_cache_stat[type] & COP(op))
tools/perf/util/evsel.c
850
int __evsel__hw_cache_type_op_res_name(u8 type, u8 op, u8 result, char *bf, size_t size)
tools/perf/util/evsel.c
854
evsel__hw_cache_op[op][0],
tools/perf/util/evsel.c
859
evsel__hw_cache_op[op][1]);
tools/perf/util/evsel.c
864
u8 op, result, type = (config >> 0) & 0xff;
tools/perf/util/evsel.c
870
op = (config >> 8) & 0xff;
tools/perf/util/evsel.c
872
if (op >= PERF_COUNT_HW_CACHE_OP_MAX)
tools/perf/util/evsel.c
881
if (!evsel__is_cache_op_valid(type, op))
tools/perf/util/evsel.c
884
return __evsel__hw_cache_type_op_res_name(type, op, result, bf, size);
tools/perf/util/evsel.h
296
bool evsel__is_cache_op_valid(u8 type, u8 op);
tools/perf/util/evsel.h
319
int __evsel__hw_cache_type_op_res_name(u8 type, u8 op, u8 result, char *bf, size_t size);
tools/perf/util/intel-bts.c
392
btsq->sample_flags = intel_pt_insn_type(btsq->intel_pt_insn.op);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1204
if (intel_pt_insn->op == INTEL_PT_OP_CALL) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1213
} else if (intel_pt_insn->op == INTEL_PT_OP_RET) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1256
decoder->state.insn_op = intel_pt_insn->op;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1647
if (intel_pt_insn.op == INTEL_PT_OP_RET) {
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
105
op = INTEL_PT_OP_RET;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
109
op = INTEL_PT_OP_IRET;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
113
op = INTEL_PT_OP_INT;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
117
op = INTEL_PT_OP_CALL;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
121
op = INTEL_PT_OP_CALL;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
125
op = INTEL_PT_OP_LOOP;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
129
op = INTEL_PT_OP_JCC;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
134
op = INTEL_PT_OP_JMP;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
138
op = INTEL_PT_OP_JMP;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
146
op = INTEL_PT_OP_CALL;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
151
op = INTEL_PT_OP_JMP;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
162
intel_pt_insn->op = op;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
258
const char *intel_pt_insn_name(enum intel_pt_insn_op op)
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
260
return branch_name[op];
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
270
intel_pt_insn_name(intel_pt_insn->op),
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
276
intel_pt_insn_name(intel_pt_insn->op));
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
28
enum intel_pt_insn_op op = INTEL_PT_OP_OTHER;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
283
int intel_pt_insn_type(enum intel_pt_insn_op op)
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
285
switch (op) {
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
36
intel_pt_insn->op = INTEL_PT_OP_OTHER;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
49
op = INTEL_PT_OP_VMENTRY;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
55
op = INTEL_PT_OP_ERETS;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
59
op = INTEL_PT_OP_ERETU;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
72
op = INTEL_PT_OP_SYSCALL;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
77
op = INTEL_PT_OP_SYSRET;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
81
op = INTEL_PT_OP_JCC;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
89
op = INTEL_PT_OP_JCC;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
94
intel_pt_insn->op = INTEL_PT_OP_JMP;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.h
40
enum intel_pt_insn_op op;
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.h
51
const char *intel_pt_insn_name(enum intel_pt_insn_op op);
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.h
56
int intel_pt_insn_type(enum intel_pt_insn_op op);
tools/perf/util/intel-pt.c
546
enum intel_pt_insn_op op;
tools/perf/util/intel-pt.c
633
e->op = intel_pt_insn->op;
tools/perf/util/intel-pt.c
769
intel_pt_insn->op = INTEL_PT_OP_OTHER;
tools/perf/util/intel-pt.c
842
intel_pt_insn->op = e->op;
tools/perf/util/intel-pt.c
904
intel_pt_insn->op = INTEL_PT_OP_OTHER;
tools/perf/util/mem-events.c
419
u64 op = PERF_MEM_LOCK_NA;
tools/perf/util/mem-events.c
423
op = mem_info__const_data_src(mem_info)->mem_op;
tools/perf/util/mem-events.c
425
if (op & PERF_MEM_OP_NA)
tools/perf/util/mem-events.c
427
else if (op & PERF_MEM_OP_LOAD)
tools/perf/util/mem-events.c
429
else if (op & PERF_MEM_OP_STORE)
tools/perf/util/mem-events.c
431
else if (op & PERF_MEM_OP_PFETCH)
tools/perf/util/mem-events.c
433
else if (op & PERF_MEM_OP_EXEC)
tools/perf/util/mem-events.c
630
u64 op = data_src->mem_op;
tools/perf/util/mem-events.c
668
if (op & P(OP, LOAD)) {
tools/perf/util/mem-events.c
736
} else if (op & P(OP, STORE)) {
tools/perf/util/print_binary.h
21
typedef int (*binary__fprintf_t)(enum binary_printer_ops op,
tools/perf/util/probe-file.c
763
char *op, *desc = strdup(arg), *new_op = NULL;
tools/perf/util/probe-file.c
777
op = strchr(desc, '@');
tools/perf/util/probe-file.c
778
if (op) {
tools/perf/util/probe-file.c
779
op[0] = '\0';
tools/perf/util/probe-file.c
780
op++;
tools/perf/util/probe-file.c
785
op = desc;
tools/perf/util/probe-file.c
788
ret = perf_sdt_arg_parse_op(EM_HOST, op, &new_op);
tools/perf/util/probe-finder.c
105
if (dwarf_getlocation_addr(&attr, addr, &op, &nops, 1) <= 0) {
tools/perf/util/probe-finder.c
126
if (dwarf_getlocation_addr(&attr, addr, &op,
tools/perf/util/probe-finder.c
137
if (op->atom == DW_OP_addr) {
tools/perf/util/probe-finder.c
154
if (op->atom == DW_OP_fbreg) {
tools/perf/util/probe-finder.c
158
offs = op->number;
tools/perf/util/probe-finder.c
159
op = &fb_ops[0];
tools/perf/util/probe-finder.c
162
if (op->atom >= DW_OP_breg0 && op->atom <= DW_OP_breg31) {
tools/perf/util/probe-finder.c
163
regn = op->atom - DW_OP_breg0;
tools/perf/util/probe-finder.c
164
offs += op->number;
tools/perf/util/probe-finder.c
166
} else if (op->atom >= DW_OP_reg0 && op->atom <= DW_OP_reg31) {
tools/perf/util/probe-finder.c
167
regn = op->atom - DW_OP_reg0;
tools/perf/util/probe-finder.c
168
} else if (op->atom == DW_OP_bregx) {
tools/perf/util/probe-finder.c
169
regn = op->number;
tools/perf/util/probe-finder.c
170
offs += op->number2;
tools/perf/util/probe-finder.c
172
} else if (op->atom == DW_OP_regx) {
tools/perf/util/probe-finder.c
173
regn = op->number;
tools/perf/util/probe-finder.c
175
pr_debug("DW_OP %x is not supported.\n", op->atom);
tools/perf/util/probe-finder.c
77
Dwarf_Op *op;
tools/perf/util/scripting-engines/trace-event-perl.c
243
if (strcmp(args->op.op, ":") == 0)
tools/perf/util/scripting-engines/trace-event-perl.c
245
define_event_symbols(event, ev_name, args->op.left);
tools/perf/util/scripting-engines/trace-event-perl.c
246
define_event_symbols(event, ev_name, args->op.right);
tools/perf/util/scripting-engines/trace-event-python.c
301
if (strcmp(args->op.op, ":") == 0)
tools/perf/util/scripting-engines/trace-event-python.c
303
define_event_symbols(event, ev_name, args->op.left);
tools/perf/util/scripting-engines/trace-event-python.c
304
define_event_symbols(event, ev_name, args->op.right);
tools/perf/util/strfilter.c
66
static struct strfilter_node *strfilter_node__alloc(const char *op,
tools/perf/util/strfilter.c
73
node->p = op;
tools/perf/util/trace-event-parse.c
162
field = find_arg_field(prev_state_field, arg->op.left);
tools/perf/util/trace-event-parse.c
165
field = find_arg_field(prev_state_field, arg->op.right);
tools/perf/util/trace-event-parse.c
193
sym = search_op(prev_state_field, arg->op.left);
tools/perf/util/trace-event-parse.c
197
sym = search_op(prev_state_field, arg->op.right);
tools/power/cpupower/utils/helpers/cpuid.c
25
unsigned int cpuid_##reg(unsigned int op) \
tools/power/cpupower/utils/helpers/cpuid.c
28
__cpuid(op, eax, ebx, ecx, edx); \
tools/power/cpupower/utils/helpers/helpers.h
160
unsigned int cpuid_eax(unsigned int op);
tools/power/cpupower/utils/helpers/helpers.h
161
unsigned int cpuid_ebx(unsigned int op);
tools/power/cpupower/utils/helpers/helpers.h
162
unsigned int cpuid_ecx(unsigned int op);
tools/power/cpupower/utils/helpers/helpers.h
163
unsigned int cpuid_edx(unsigned int op);
tools/power/cpupower/utils/helpers/helpers.h
207
static inline unsigned int cpuid_eax(unsigned int op) { return 0; };
tools/power/cpupower/utils/helpers/helpers.h
208
static inline unsigned int cpuid_ebx(unsigned int op) { return 0; };
tools/power/cpupower/utils/helpers/helpers.h
209
static inline unsigned int cpuid_ecx(unsigned int op) { return 0; };
tools/power/cpupower/utils/helpers/helpers.h
210
static inline unsigned int cpuid_edx(unsigned int op) { return 0; };
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
230
ctx.bench->bss->op = LPM_OP_NOOP;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
236
ctx.bench->bss->op = LPM_OP_BASELINE;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
242
ctx.bench->bss->op = LPM_OP_LOOKUP;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
248
ctx.bench->bss->op = LPM_OP_INSERT;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
254
ctx.bench->bss->op = LPM_OP_UPDATE;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
260
ctx.bench->bss->op = LPM_OP_DELETE;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
266
ctx.bench->bss->op = LPM_OP_FREE;
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
279
switch (ctx.bench->bss->op) {
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
290
ctx.bench->bss->op);
tools/testing/selftests/bpf/benchs/bench_lpm_trie_map.c
327
opts.retval, ctx.bench->bss->op);
tools/testing/selftests/bpf/bpf_experimental.h
251
#define __bpf_assert_check(LHS, op, RHS) \
tools/testing/selftests/bpf/bpf_experimental.h
257
#define __bpf_assert(LHS, op, cons, RHS, VAL) \
tools/testing/selftests/bpf/bpf_experimental.h
260
asm volatile ("if %[lhs] " op " %[rhs] goto +2; r1 = %[value]; call bpf_throw" \
tools/testing/selftests/bpf/bpf_experimental.h
264
#define __bpf_assert_op_sign(LHS, op, cons, RHS, VAL, supp_sign) \
tools/testing/selftests/bpf/bpf_experimental.h
266
__bpf_assert_check(LHS, op, RHS); \
tools/testing/selftests/bpf/bpf_experimental.h
268
__bpf_assert(LHS, "s" #op, cons, RHS, VAL); \
tools/testing/selftests/bpf/bpf_experimental.h
270
__bpf_assert(LHS, #op, cons, RHS, VAL); \
tools/testing/selftests/bpf/bpf_experimental.h
273
#define __bpf_assert_op(LHS, op, RHS, VAL, supp_sign) \
tools/testing/selftests/bpf/bpf_experimental.h
277
__bpf_assert_op_sign(LHS, op, "r", rhs_var, VAL, supp_sign); \
tools/testing/selftests/bpf/bpf_experimental.h
279
__bpf_assert_op_sign(LHS, op, "i", RHS, VAL, supp_sign); \
tools/testing/selftests/bpf/prog_tests/btf_dump.c
833
{ .op = 1, .args = { 1, 2, 3, 4}});
tools/testing/selftests/bpf/prog_tests/btf_dump.c
864
{ .op = 1, .skb_hwtstamp = 2});
tools/testing/selftests/bpf/prog_tests/btf_dump.c
868
{ .op = 1, .skb_hwtstamp = 0});
tools/testing/selftests/bpf/prog_tests/linked_list.c
32
#define TEST(test, op) \
tools/testing/selftests/bpf/prog_tests/linked_list.c
33
{ #test "_kptr_incorrect_lock_" #op, \
tools/testing/selftests/bpf/prog_tests/linked_list.c
36
{ #test "_global_incorrect_lock_" #op, \
tools/testing/selftests/bpf/prog_tests/linked_list.c
39
{ #test "_map_incorrect_lock_" #op, \
tools/testing/selftests/bpf/prog_tests/linked_list.c
42
{ #test "_inner_map_incorrect_lock_" #op, \
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1302
struct range x, struct range y, enum op op,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1312
enum op rev_op = complement_op(op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1351
*branch_taken = reg_state_branch_taken_op(cond_t, fr1, fr2, op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1360
reg_state_cond(cond_t, tr1, tr2, op, tr1, tr2, "TRUE");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1423
enum op op;
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1430
snappendf(sb, " (%s)%s ", t_str(t->cond_t), use_op ? op_str(t->op) : "<op>");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1440
struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1458
sim_case(init_t, cond_t, x, y, op, &fe1, &fe2, &te1, &te2, &branch_taken);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1460
err = load_range_cmp_prog(x, y, op, branch_taken, spec,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1522
for (sub.op = first_op; sub.op <= last_op; sub.op++) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1529
err = verify_case_op(init_t, cond_t, x, y, sub.op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
532
static enum op complement_op(enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
534
switch (op) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
545
static const char *op_str(enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
547
switch (op) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
563
static bool range_canbe_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
566
switch (op) { \
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
573
default: printf("range_canbe op %d\n", op); exit(1); \
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
592
static bool range_always_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
595
return !range_canbe_op(t, x, y, complement_op(op));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
603
static bool range_never_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
605
return !range_canbe_op(t, x, y, op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
613
static int range_branch_taken_op(enum num_t t, struct range x, struct range y, enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
615
if (range_always_op(t, x, y, op))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
617
if (range_never_op(t, x, y, op))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
630
enum op op, struct range *newx, struct range *newy)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
632
if (!range_canbe_op(t, x, y, op)) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
638
switch (op) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
780
static void reg_state_cond(enum num_t t, struct reg_state *x, struct reg_state *y, enum op op,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
789
if (op == OP_EQ || op == OP_NE) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
806
range_cond(t, z1, z2, op, &z1, &z2);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
825
enum op op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
827
if (op == OP_EQ || op == OP_NE) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
833
br_u = range_branch_taken_op(tu, x->r[tu], y->r[tu], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
834
br_s = range_branch_taken_op(ts, x->r[ts], y->r[ts], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
843
br = range_branch_taken_op(U32, x->r[U32], y->r[U32], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
847
if (op == OP_EQ && br == 0)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
852
if (op == OP_NE && br == 1)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
855
br = range_branch_taken_op(S32, x->r[S32], y->r[S32], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
856
if (op == OP_EQ && br == 0)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
858
if (op == OP_NE && br == 1)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
864
return range_branch_taken_op(t, x->r[t], y->r[t], op);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
891
static int load_range_cmp_prog(struct range x, struct range y, enum op op,
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
987
switch (op) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
995
printf("unrecognized op %d\n", op);
tools/testing/selftests/bpf/prog_tests/sock_addr.c
51
#define save_errno_do(op) ({ int __save = errno; op; errno = __save; })
tools/testing/selftests/bpf/prog_tests/sockopt.c
1010
io_uring_prep_cmd(sqe, op, fd, level, optname, optval, optlen);
tools/testing/selftests/bpf/prog_tests/sockopt.c
992
static int uring_sockopt(int op, int fd, int level, int optname,
tools/testing/selftests/bpf/prog_tests/test_bpf_smc.c
151
static bool smc_ueid(int op)
tools/testing/selftests/bpf/prog_tests/test_bpf_smc.c
177
NLM_F_REQUEST | NLM_F_ACK, op, SMC_NLA_EID_TABLE_ENTRY,
tools/testing/selftests/bpf/prog_tests/test_bpf_smc.c
189
switch (op) {
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
410
__u8 op;
tools/testing/selftests/bpf/prog_tests/uprobe_syscall.c
425
ASSERT_EQ(call->op, 0xe8, "call");
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
459
enum test_msg_op op, s64 operand_64,
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
473
msg->msg_op = op;
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
475
switch (op) {
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
485
PRINT_FAIL("Invalid operand %d\n", op);
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
547
enum test_msg_op op = i % TEST_MSG_OP_NUM_OPS;
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
551
err = send_test_message(user_ringbuf, op, operand_64, operand_32);
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
558
switch (op) {
tools/testing/selftests/bpf/prog_tests/user_ringbuf.c
572
PRINT_FAIL("Unexpected op %d\n", op);
tools/testing/selftests/bpf/progs/cgrp_ls_attach_cgroup.c
59
if (ctx->family != AF_INET6 || ctx->op != BPF_SOCK_OPS_TCP_CONNECT_CB)
tools/testing/selftests/bpf/progs/exceptions_assert.c
11
#define check_assert(type, op, name, value) \
tools/testing/selftests/bpf/progs/exceptions_assert.c
17
bpf_assert(bpf_cmp_unlikely(num, op, value)); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
105
#define CHECK_OP(op) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
106
CHECK(kptr_kptr, op, &f1->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
107
CHECK(kptr_global, op, &f1->lock, &ghead); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
108
CHECK(kptr_map, op, &f1->lock, &v->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
109
CHECK(kptr_inner_map, op, &f1->lock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
111
CHECK(global_global, op, &glock2, &ghead); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
112
CHECK(global_kptr, op, &glock, &f1->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
113
CHECK(global_map, op, &glock, &v->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
114
CHECK(global_inner_map, op, &glock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
116
CHECK(map_map, op, &v->lock, &v2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
117
CHECK(map_kptr, op, &v->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
118
CHECK(map_global, op, &v->lock, &ghead); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
119
CHECK(map_inner_map, op, &v->lock, &iv->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
121
CHECK(inner_map_inner_map, op, &iv->lock, &iv2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
122
CHECK(inner_map_kptr, op, &iv->lock, &f2->head); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
123
CHECK(inner_map_global, op, &iv->lock, &ghead); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
124
CHECK(inner_map_map, op, &iv->lock, &v->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
132
#define CHECK(test, op, lexpr, hexpr, nexpr) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
134
int test##_incorrect_lock_##op(void *ctx) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
138
bpf_list_##op(hexpr, nexpr); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
142
#define CHECK_OP(op) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
143
CHECK(kptr_kptr, op, &f1->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
144
CHECK(kptr_global, op, &f1->lock, &ghead, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
145
CHECK(kptr_map, op, &f1->lock, &v->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
146
CHECK(kptr_inner_map, op, &f1->lock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
148
CHECK(global_global, op, &glock2, &ghead, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
149
CHECK(global_kptr, op, &glock, &f1->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
150
CHECK(global_map, op, &glock, &v->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
151
CHECK(global_inner_map, op, &glock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
153
CHECK(map_map, op, &v->lock, &v2->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
154
CHECK(map_kptr, op, &v->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
155
CHECK(map_global, op, &v->lock, &ghead, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
156
CHECK(map_inner_map, op, &v->lock, &iv->head, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
158
CHECK(inner_map_inner_map, op, &iv->lock, &iv2->head, &f->node2);\
tools/testing/selftests/bpf/progs/linked_list_fail.c
159
CHECK(inner_map_kptr, op, &iv->lock, &f2->head, &b->node); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
160
CHECK(inner_map_global, op, &iv->lock, &ghead, &f->node2); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
161
CHECK(inner_map_map, op, &iv->lock, &v->head, &f->node2);
tools/testing/selftests/bpf/progs/linked_list_fail.c
47
#define CHECK(test, op, hexpr) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
49
int test##_missing_lock_##op(void *ctx) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
52
void (*p)(void *) = (void *)&bpf_list_##op; \
tools/testing/selftests/bpf/progs/linked_list_fail.c
578
int pop_ptr_off(void *(*op)(void *head))
tools/testing/selftests/bpf/progs/linked_list_fail.c
590
n = op(&p->head);
tools/testing/selftests/bpf/progs/linked_list_fail.c
71
#define CHECK(test, op, hexpr, nexpr) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
73
int test##_missing_lock_##op(void *ctx) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
76
bpf_list_##op(hexpr, nexpr); \
tools/testing/selftests/bpf/progs/linked_list_fail.c
94
#define CHECK(test, op, lexpr, hexpr) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
96
int test##_incorrect_lock_##op(void *ctx) \
tools/testing/selftests/bpf/progs/linked_list_fail.c
99
void (*p)(void *) = (void *)&bpf_list_##op; \
tools/testing/selftests/bpf/progs/linked_list_peek.c
84
#define TEST_FB(op, dolock) \
tools/testing/selftests/bpf/progs/linked_list_peek.c
87
long test_##op##_spinlock_##dolock(void *ctx) \
tools/testing/selftests/bpf/progs/linked_list_peek.c
94
l_n = bpf_list_##op(&ghead); \
tools/testing/selftests/bpf/progs/lpm_trie_bench.c
200
switch (op) {
tools/testing/selftests/bpf/progs/lpm_trie_bench.c
35
__u8 op;
tools/testing/selftests/bpf/progs/mptcp_sock.c
33
int op = (int)ctx->op;
tools/testing/selftests/bpf/progs/mptcp_sock.c
38
if (op != BPF_SOCK_OPS_TCP_CONNECT_CB)
tools/testing/selftests/bpf/progs/mptcp_sockmap.c
25
skops->op != BPF_SOCK_OPS_PASSIVE_ESTABLISHED_CB)
tools/testing/selftests/bpf/progs/mptcp_subflow.c
30
if (skops->op != BPF_SOCK_OPS_TCP_CONNECT_CB)
tools/testing/selftests/bpf/progs/net_timestamping.c
140
if (skops->op == BPF_SOCK_OPS_TSTAMP_SENDMSG_CB) {
tools/testing/selftests/bpf/progs/net_timestamping.c
161
switch (skops->op) {
tools/testing/selftests/bpf/progs/net_timestamping.c
182
if (skops->op == BPF_SOCK_OPS_TSTAMP_ACK_CB)
tools/testing/selftests/bpf/progs/net_timestamping.c
223
switch (skops->op) {
tools/testing/selftests/bpf/progs/netns_cookie_prog.c
46
switch (ctx->op) {
tools/testing/selftests/bpf/progs/rbtree_search.c
153
#define TEST_LR(op, dolock) \
tools/testing/selftests/bpf/progs/rbtree_search.c
156
long test_##op##_spinlock_##dolock(void *ctx) \
tools/testing/selftests/bpf/progs/rbtree_search.c
176
rb_n = bpf_rbtree_##op(&groot0, &n->r0); \
tools/testing/selftests/bpf/progs/setget_sockopt.c
404
switch (skops->op) {
tools/testing/selftests/bpf/progs/socket_cookie_prog.c
60
if (ctx->op != BPF_SOCK_OPS_TCP_CONNECT_CB)
tools/testing/selftests/bpf/progs/tcp_rtt.c
30
int op = (int) ctx->op;
tools/testing/selftests/bpf/progs/tcp_rtt.c
42
if (op == BPF_SOCK_OPS_TCP_CONNECT_CB) {
tools/testing/selftests/bpf/progs/tcp_rtt.c
47
if (op != BPF_SOCK_OPS_RTT_CB)
tools/testing/selftests/bpf/progs/test_misc_tcp_hdr_options.c
305
switch (skops->op) {
tools/testing/selftests/bpf/progs/test_sockmap_kern.h
191
int op, ret;
tools/testing/selftests/bpf/progs/test_sockmap_kern.h
193
op = (int) skops->op;
tools/testing/selftests/bpf/progs/test_sockmap_kern.h
195
switch (op) {
tools/testing/selftests/bpf/progs/test_tcp_hdr_options.c
599
switch (skops->op) {
tools/testing/selftests/bpf/progs/test_tcpbpf_kern.c
41
int op;
tools/testing/selftests/bpf/progs/test_tcpbpf_kern.c
51
: [op] "=r"(op)
tools/testing/selftests/bpf/progs/test_tcpbpf_kern.c
83
op = (int) skops->op;
tools/testing/selftests/bpf/progs/test_tcpbpf_kern.c
85
global.event_map |= (1 << op);
tools/testing/selftests/bpf/progs/test_tcpbpf_kern.c
87
switch (op) {
tools/testing/selftests/bpf/progs/test_tcpnotify_kern.c
34
int op;
tools/testing/selftests/bpf/progs/test_tcpnotify_kern.c
36
op = (int) skops->op;
tools/testing/selftests/bpf/progs/test_tcpnotify_kern.c
43
switch (op) {
tools/testing/selftests/bpf/progs/verifier_bswap.c
51
#define BSWAP_RANGE_TEST(name, op, in_value, out_value) \
tools/testing/selftests/bpf/progs/verifier_bswap.c
55
__msg("r0 = " op " r0 {{.*}}; R0=scalar({{.*}},var_off=(0x0; " #out_value "))") \
tools/testing/selftests/bpf/progs/verifier_bswap.c
61
"r0 = " op " r0;" \
tools/testing/selftests/bpf/progs/verifier_value_illegal_alu.c
188
#define DEFINE_BAD_OFFSET_TEST(name, op, off, imm) \
tools/testing/selftests/bpf/progs/verifier_value_illegal_alu.c
199
: __imm_insn(insn, BPF_RAW_INSN((op), 0, 0, (off), (imm))) \
tools/testing/selftests/bpf/verifier/atomic_fetch.c
104
BPF_ATOMIC_OP(BPF_DW, op, \
tools/testing/selftests/bpf/verifier/atomic_fetch.c
95
#define __ATOMIC_FETCH_OP_TEST(src_reg, dst_reg, operand1, op, operand2, expect) \
tools/testing/selftests/bpf/verifier/atomic_fetch.c
97
"atomic fetch " #op ", src=" #dst_reg " dst=" #dst_reg, \
tools/testing/selftests/bpf/verifier/atomic_invalid.c
1
#define __INVALID_ATOMIC_ACCESS_TEST(op) \
tools/testing/selftests/bpf/verifier/atomic_invalid.c
3
"atomic " #op " access through non-pointer ", \
tools/testing/selftests/bpf/verifier/atomic_invalid.c
7
BPF_ATOMIC_OP(BPF_DW, op, BPF_REG_1, BPF_REG_0, -8), \
tools/testing/selftests/bpf/veristat.c
153
enum operator_kind op;
tools/testing/selftests/bpf/veristat.c
2974
switch (f->op) {
tools/testing/selftests/bpf/veristat.c
2983
fprintf(stderr, "BUG: unknown filter op %d!\n", f->op);
tools/testing/selftests/bpf/veristat.c
3183
switch (f->op) {
tools/testing/selftests/bpf/veristat.c
3192
fprintf(stderr, "BUG: unknown filter op %d!\n", f->op);
tools/testing/selftests/bpf/veristat.c
665
f->op = operators[i].op_kind;
tools/testing/selftests/bpf/xdp_hw_metadata.c
569
static void hwtstamp_ioctl(int op, const char *ifname, struct hwtstamp_config *cfg)
tools/testing/selftests/bpf/xdp_hw_metadata.c
581
ret = ioctl(fd, op, &ifr);
tools/testing/selftests/bpf/xdp_hw_metadata.c
583
error(1, errno, "ioctl(%d)", op);
tools/testing/selftests/connector/proc_filter.c
268
enum proc_cn_mcast_op op = PROC_CN_MCAST_LISTEN;
tools/testing/selftests/connector/proc_filter.c
269
err = register_proc_netlink(&epoll_fd, (void*)&op);
tools/testing/selftests/connector/proc_filter.c
301
enum proc_cn_mcast_op op = PROC_CN_MCAST_IGNORE;
tools/testing/selftests/connector/proc_filter.c
302
send_message((void*)&op);
tools/testing/selftests/futex/include/futextest.h
92
#define futex(uaddr, op, val, timeout, uaddr2, val3, opflags) \
tools/testing/selftests/futex/include/futextest.h
93
syscall(SYS_futex, uaddr, op | opflags, val, timeout, uaddr2, val3)
tools/testing/selftests/iommu/iommufd.c
1065
.op = IOMMU_TEST_OP_ADD_RESERVED,
tools/testing/selftests/iommu/iommufd.c
1141
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd.c
1175
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd.c
1182
.op = IOMMU_TEST_OP_MD_CHECK_MAP,
tools/testing/selftests/iommu/iommufd.c
1247
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd.c
1276
.op = IOMMU_TEST_OP_ACCESS_RW,
tools/testing/selftests/iommu/iommufd.c
1445
.op = IOMMU_OPTION_OP_GET,
tools/testing/selftests/iommu/iommufd.c
1453
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
1457
cmd.op = IOMMU_OPTION_OP_GET;
tools/testing/selftests/iommu/iommufd.c
1462
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
1466
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
1491
.op = IOMMU_OPTION_OP_SET,
tools/testing/selftests/iommu/iommufd.c
168
.op = IOMMU_OPTION_OP_GET,
tools/testing/selftests/iommu/iommufd.c
1717
.op = IOMMU_TEST_OP_MD_CHECK_MAP, \
tools/testing/selftests/iommu/iommufd.c
177
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
183
cmd.op = IOMMU_OPTION_OP_GET;
tools/testing/selftests/iommu/iommufd.c
187
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
191
cmd.op = IOMMU_OPTION_OP_GET;
tools/testing/selftests/iommu/iommufd.c
194
cmd.op = IOMMU_OPTION_OP_SET;
tools/testing/selftests/iommu/iommufd.c
1948
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd.c
2102
.op = IOMMU_OPTION_OP_SET,
tools/testing/selftests/iommu/iommufd.c
2454
.op = IOMMU_VFIO_IOAS_GET,
tools/testing/selftests/iommu/iommufd.c
2462
vfio_ioas_cmd.op = IOMMU_VFIO_IOAS_SET;
tools/testing/selftests/iommu/iommufd.c
2471
vfio_ioas_cmd.op = IOMMU_VFIO_IOAS_GET;
tools/testing/selftests/iommu/iommufd.c
2476
vfio_ioas_cmd.op = IOMMU_VFIO_IOAS_CLEAR;
tools/testing/selftests/iommu/iommufd.c
2478
vfio_ioas_cmd.op = IOMMU_VFIO_IOAS_GET;
tools/testing/selftests/iommu/iommufd.c
2497
.op = IOMMU_VFIO_IOAS_SET,
tools/testing/selftests/iommu/iommufd.c
2537
.op = IOMMU_OPTION_OP_GET,
tools/testing/selftests/iommu/iommufd.c
261
.op = IOMMU_OPTION_OP_SET,
tools/testing/selftests/iommu/iommufd.c
885
.op = IOMMU_TEST_OP_ADD_RESERVED,
tools/testing/selftests/iommu/iommufd.c
973
.op = IOMMU_TEST_OP_ADD_RESERVED,
tools/testing/selftests/iommu/iommufd.c
999
.op = IOMMU_TEST_OP_ADD_RESERVED,
tools/testing/selftests/iommu/iommufd_fail_nth.c
456
.op = IOMMU_TEST_OP_ACCESS_RW,
tools/testing/selftests/iommu/iommufd_fail_nth.c
487
.op = IOMMU_TEST_OP_ACCESS_RW,
tools/testing/selftests/iommu/iommufd_fail_nth.c
536
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd_fail_nth.c
595
.op = IOMMU_TEST_OP_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd_utils.h
105
.op = IOMMU_TEST_OP_MOCK_DOMAIN,
tools/testing/selftests/iommu/iommufd_utils.h
1096
.op = IOMMU_TEST_OP_TRIGGER_VEVENT,
tools/testing/selftests/iommu/iommufd_utils.h
1179
.op = IOMMU_TEST_OP_PASID_ATTACH,
tools/testing/selftests/iommu/iommufd_utils.h
1205
.op = IOMMU_TEST_OP_PASID_REPLACE,
tools/testing/selftests/iommu/iommufd_utils.h
1230
.op = IOMMU_TEST_OP_PASID_DETACH,
tools/testing/selftests/iommu/iommufd_utils.h
1249
.op = IOMMU_TEST_OP_PASID_CHECK_HWPT,
tools/testing/selftests/iommu/iommufd_utils.h
136
.op = IOMMU_TEST_OP_MOCK_DOMAIN_FLAGS,
tools/testing/selftests/iommu/iommufd_utils.h
167
.op = IOMMU_TEST_OP_MOCK_DOMAIN_REPLACE,
tools/testing/selftests/iommu/iommufd_utils.h
261
.op = IOMMU_TEST_OP_MD_CHECK_IOTLB, \
tools/testing/selftests/iommu/iommufd_utils.h
285
.op = IOMMU_TEST_OP_DEV_CHECK_CACHE, \
tools/testing/selftests/iommu/iommufd_utils.h
373
.op = IOMMU_TEST_OP_ACCESS_REPLACE_IOAS,
tools/testing/selftests/iommu/iommufd_utils.h
436
.op = IOMMU_TEST_OP_DIRTY,
tools/testing/selftests/iommu/iommufd_utils.h
522
.op = IOMMU_TEST_OP_CREATE_ACCESS,
tools/testing/selftests/iommu/iommufd_utils.h
550
.op = IOMMU_TEST_OP_DESTROY_ACCESS_PAGES,
tools/testing/selftests/iommu/iommufd_utils.h
567
.op = IOMMU_TEST_OP_DMABUF_GET,
tools/testing/selftests/iommu/iommufd_utils.h
583
.op = IOMMU_TEST_OP_DMABUF_REVOKE,
tools/testing/selftests/iommu/iommufd_utils.h
781
.op = IOMMU_TEST_OP_SET_TEMP_MEMORY_LIMIT,
tools/testing/selftests/iommu/iommufd_utils.h
799
.op = IOMMU_TEST_OP_MD_CHECK_REFS,
tools/testing/selftests/iommu/iommufd_utils.h
89
.op = IOMMU_TEST_OP_MD_CHECK_REFS, \
tools/testing/selftests/iommu/iommufd_utils.h
932
.op = IOMMU_TEST_OP_TRIGGER_IOPF,
tools/testing/selftests/kvm/arm64/at.c
25
#define __at(op, addr) write_sysreg_s(addr, op)
tools/testing/selftests/kvm/arm64/at.c
27
#define test_at_insn(op, expect_fault) \
tools/testing/selftests/kvm/arm64/at.c
34
__at(OP_AT_##op, TEST_ADDR); \
tools/testing/selftests/kvm/arm64/at.c
42
"AT "#op": %sexpected fault (par: %lx)1", \
tools/testing/selftests/kvm/arm64/at.c
46
"AT "#op": expected access flag fault (par: %lx)", \
tools/testing/selftests/kvm/lib/sparsebit.c
2077
uint8_t op = get8() & 0xf;
tools/testing/selftests/kvm/lib/sparsebit.c
2081
operate(op, first, last);
tools/testing/selftests/kvm/s390/keyop.c
71
static int do_keyop(struct kvm_vcpu *vcpu, int op, unsigned long page_idx, unsigned char skey)
tools/testing/selftests/kvm/s390/keyop.c
76
.operation = op,
tools/testing/selftests/kvm/s390/memop.c
130
switch (ksmo->op) {
tools/testing/selftests/kvm/s390/memop.c
182
if (ksmo->op == KVM_S390_MEMOP_ABSOLUTE_CMPXCHG) {
tools/testing/selftests/kvm/s390/memop.c
71
ksmo.op = KVM_S390_MEMOP_LOGICAL_READ;
tools/testing/selftests/kvm/s390/memop.c
73
ksmo.op = KVM_S390_MEMOP_LOGICAL_WRITE;
tools/testing/selftests/kvm/s390/memop.c
77
ksmo.op = KVM_S390_MEMOP_SIDA_READ;
tools/testing/selftests/kvm/s390/memop.c
79
ksmo.op = KVM_S390_MEMOP_SIDA_WRITE;
tools/testing/selftests/kvm/s390/memop.c
83
ksmo.op = KVM_S390_MEMOP_ABSOLUTE_READ;
tools/testing/selftests/kvm/s390/memop.c
85
ksmo.op = KVM_S390_MEMOP_ABSOLUTE_WRITE;
tools/testing/selftests/kvm/s390/memop.c
87
ksmo.op = KVM_S390_MEMOP_ABSOLUTE_CMPXCHG;
tools/testing/selftests/kvm/s390/memop.c
93
ksmo.op = -1;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
203
enum ftracer_op op;
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
212
op = tracer->process_line(*lp);
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
222
if (op == FTRACER_LINE_DISCARD)
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
224
if (op == FTRACER_EXIT)
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
226
if (op != FTRACER_LINE_PRESERVE)
tools/testing/selftests/net/tcp_ao/lib/ftrace.c
227
test_error("unexpected tracer command %d", op);
tools/testing/selftests/powerpc/alignment/alignment_handler.c
106
#define LOAD_VSX_XFORM_TEST(op) TEST(op, op, stxvd2x, XFORM, 32, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
107
#define STORE_VSX_XFORM_TEST(op) TEST(op, lxvd2x, op, XFORM, 32, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
108
#define LOAD_VSX_DFORM_TEST(op) TEST(op, op, stxv, DFORM, 32, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
109
#define STORE_VSX_DFORM_TEST(op) TEST(op, lxv, op, DFORM, 32, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
110
#define LOAD_VMX_XFORM_TEST(op) TEST(op, op, stxvd2x, XFORM, 0, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
111
#define STORE_VMX_XFORM_TEST(op) TEST(op, lxvd2x, op, XFORM, 32, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
112
#define LOAD_VMX_DFORM_TEST(op) TEST(op, op, stxv, DFORM, 0, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
113
#define STORE_VMX_DFORM_TEST(op) TEST(op, lxv, op, DFORM, 32, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
115
#define LOAD_XFORM_TEST(op) TEST(op, op, stdx, XFORM, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
116
#define STORE_XFORM_TEST(op) TEST(op, ldx, op, XFORM, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
117
#define LOAD_DFORM_TEST(op) TEST(op, op, std, DFORM, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
118
#define STORE_DFORM_TEST(op) TEST(op, ld, op, DFORM, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
120
#define LOAD_FLOAT_DFORM_TEST(op) TEST(op, op, stfd, DFORM, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
121
#define STORE_FLOAT_DFORM_TEST(op) TEST(op, lfd, op, DFORM, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
122
#define LOAD_FLOAT_XFORM_TEST(op) TEST(op, op, stfdx, XFORM, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
123
#define STORE_FLOAT_XFORM_TEST(op) TEST(op, lfdx, op, XFORM, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
125
#define LOAD_MLS_PREFIX_TEST(op) TESTP(op, op, PSTD, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
126
#define STORE_MLS_PREFIX_TEST(op) TESTP(op, PLD, op, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
128
#define LOAD_8LS_PREFIX_TEST(op) TESTP(op, op, PSTD, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
129
#define STORE_8LS_PREFIX_TEST(op) TESTP(op, PLD, op, 31, 31)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
131
#define LOAD_FLOAT_MLS_PREFIX_TEST(op) TESTP(op, op, PSTFD, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
132
#define STORE_FLOAT_MLS_PREFIX_TEST(op) TESTP(op, PLFD, op, 0, 0)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
134
#define LOAD_VSX_8LS_PREFIX_TEST(op, tail) TESTP(op, op, PSTXV ## tail, 0, 32)
tools/testing/selftests/powerpc/alignment/alignment_handler.c
135
#define STORE_VSX_8LS_PREFIX_TEST(op, tail) TESTP(op, PLXV ## tail, op, 32, 0)
tools/testing/selftests/powerpc/benchmarks/context_switch.c
258
static long sys_futex(void *addr1, int op, int val1, struct timespec *timeout,
tools/testing/selftests/powerpc/benchmarks/context_switch.c
261
return syscall(SYS_futex, addr1, op, val1, timeout, addr2, val3);
tools/testing/selftests/powerpc/include/basic_asm.h
97
.macro OP_REGS op, reg_width, start_reg, end_reg, base_reg, base_reg_offset=0, skip=0
tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h
28
.macro OP_REGS op, width, start, end, base, offset
tools/testing/selftests/ptrace/get_syscall_info.c
114
.op = 0xff /* invalid PTRACE_SYSCALL_INFO_* op */
tools/testing/selftests/ptrace/get_syscall_info.c
163
ASSERT_EQ(PTRACE_SYSCALL_INFO_NONE, info.op) {
tools/testing/selftests/ptrace/get_syscall_info.c
191
ASSERT_EQ(PTRACE_SYSCALL_INFO_ENTRY, info.op) {
tools/testing/selftests/ptrace/get_syscall_info.c
231
ASSERT_EQ(PTRACE_SYSCALL_INFO_EXIT, info.op) {
tools/testing/selftests/ptrace/set_syscall_info.c
109
ASSERT_EQ(PTRACE_SYSCALL_INFO_EXIT, info->op) {
tools/testing/selftests/ptrace/set_syscall_info.c
369
.op = 0xff /* invalid PTRACE_SYSCALL_INFO_* op */
tools/testing/selftests/ptrace/set_syscall_info.c
452
info.op = 0xff;
tools/testing/selftests/ptrace/set_syscall_info.c
490
info.op = 0xff;
tools/testing/selftests/ptrace/set_syscall_info.c
80
ASSERT_EQ(PTRACE_SYSCALL_INFO_ENTRY, info->op) {
tools/testing/selftests/riscv/vector/vstate_ptrace.c
14
static long do_ptrace(enum __ptrace_request op, pid_t pid, long type, size_t size, void *data)
tools/testing/selftests/riscv/vector/vstate_ptrace.c
21
return ptrace(op, pid, type, &v_iovec);
tools/testing/selftests/seccomp/seccomp_bpf.c
305
int seccomp(unsigned int op, unsigned int flags, void *args)
tools/testing/selftests/seccomp/seccomp_bpf.c
308
return syscall(__NR_seccomp, op, flags, args);
tools/testing/selftests/sgx/main.c
262
#define ENCL_CALL(op, run, clobbered) \
tools/testing/selftests/sgx/main.c
266
ret = vdso_sgx_enter_enclave((unsigned long)(op), 0, 0, \
tools/testing/selftests/sgx/main.c
269
ret = sgx_enter_enclave((void *)(op), NULL, 0, EENTER, NULL, NULL, \
tools/testing/selftests/sgx/main.c
576
struct encl_op_header op;
tools/testing/selftests/sgx/main.c
583
op.type = ENCL_OP_NOP;
tools/testing/selftests/sgx/main.c
585
EXPECT_EQ(ENCL_CALL(&op, &self->run, true), 0);
tools/testing/selftests/sgx/main.c
595
EXPECT_EQ(ENCL_CALL(&op, &self->run, true), 0);
tools/testing/selftests/sgx/test_encl.c
101
struct encl_op_get_from_buf *op2 = op;
tools/testing/selftests/sgx/test_encl.c
108
struct encl_op_put_to_addr *op = _op;
tools/testing/selftests/sgx/test_encl.c
110
memcpy((void *)op->addr, &op->value, 8);
tools/testing/selftests/sgx/test_encl.c
115
struct encl_op_get_from_addr *op = _op;
tools/testing/selftests/sgx/test_encl.c
117
memcpy(&op->value, (void *)op->addr, 8);
tools/testing/selftests/sgx/test_encl.c
149
encl_op_t op;
tools/testing/selftests/sgx/test_encl.c
159
op = ((uint64_t)&__encl_base) + encl_op_array[header->type];
tools/testing/selftests/sgx/test_encl.c
161
(*op)(header);
tools/testing/selftests/sgx/test_encl.c
23
struct encl_op_emodpe *op = _op;
tools/testing/selftests/sgx/test_encl.c
25
secinfo.flags = op->flags;
tools/testing/selftests/sgx/test_encl.c
31
"c" (op->epc_addr)
tools/testing/selftests/sgx/test_encl.c
38
struct encl_op_eaccept *op = _op;
tools/testing/selftests/sgx/test_encl.c
41
secinfo.flags = op->flags;
tools/testing/selftests/sgx/test_encl.c
47
"c" (op->epc_addr)
tools/testing/selftests/sgx/test_encl.c
50
op->ret = rax;
tools/testing/selftests/sgx/test_encl.c
75
struct encl_op_init_tcs_page *op = _op;
tools/testing/selftests/sgx/test_encl.c
76
void *tcs = (void *)op->tcs_page;
tools/testing/selftests/sgx/test_encl.c
80
memcpy(tcs + 16, &op->ssa, 8); /* OSSA */
tools/testing/selftests/sgx/test_encl.c
84
memcpy(tcs + 32, &op->entry, 8); /* OENTRY */
tools/testing/selftests/sgx/test_encl.c
92
static void do_encl_op_put_to_buf(void *op)
tools/testing/selftests/sgx/test_encl.c
94
struct encl_op_put_to_buf *op2 = op;
tools/testing/selftests/sgx/test_encl.c
99
static void do_encl_op_get_from_buf(void *op)
tools/testing/selftests/syscall_user_dispatch/sud_test.c
106
int op;
tools/testing/selftests/syscall_user_dispatch/sud_test.c
109
op = -1;
tools/testing/selftests/syscall_user_dispatch/sud_test.c
110
prctl_invalid(_metadata, op, 0, 0, &sel, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
113
op = PR_SYS_DISPATCH_OFF;
tools/testing/selftests/syscall_user_dispatch/sud_test.c
116
prctl_invalid(_metadata, op, 0x1, 0x0, 0, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
119
prctl_invalid(_metadata, op, 0x0, 0xff, 0, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
122
prctl_invalid(_metadata, op, 0x0, 0x0, &sel, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
125
prctl_valid(_metadata, op, 0x0, 0x0, 0x0);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
128
op = PR_SYS_DISPATCH_EXCLUSIVE_ON;
tools/testing/selftests/syscall_user_dispatch/sud_test.c
131
prctl_invalid(_metadata, op, 0x1, 0x0, &sel, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
132
prctl_invalid(_metadata, op, -1L, 0x0, &sel, EINVAL);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
135
prctl_invalid(_metadata, op, 0x0, 0x1, (void *) -1, EFAULT);
tools/testing/selftests/syscall_user_dispatch/sud_test.c
350
unsigned long op, unsigned long off,
tools/testing/selftests/syscall_user_dispatch/sud_test.c
355
EXPECT_EQ(0, prctl(PR_SET_SYSCALL_USER_DISPATCH, op, off, size, &glob_sel));
tools/testing/selftests/syscall_user_dispatch/sud_test.c
89
unsigned long op, unsigned long off,
tools/testing/selftests/syscall_user_dispatch/sud_test.c
92
EXPECT_EQ(0, prctl(PR_SET_SYSCALL_USER_DISPATCH, op, off, size, sel));
tools/testing/selftests/syscall_user_dispatch/sud_test.c
96
unsigned long op, unsigned long off,
tools/testing/selftests/syscall_user_dispatch/sud_test.c
99
EXPECT_EQ(-1, prctl(PR_SET_SYSCALL_USER_DISPATCH, op, off, size, sel));
tools/testing/selftests/ublk/batch.c
230
struct io_uring_sqe *sqe, unsigned op,
tools/testing/selftests/ublk/batch.c
240
ublk_set_sqe_cmd_op(sqe, op);
tools/testing/selftests/ublk/batch.c
252
user_data = build_user_data(buf_idx, _IOC_NR(op), nr_elem, q_id, 0);
tools/testing/selftests/ublk/batch.c
260
__func__, t->idx, q_id, op, user_data,
tools/testing/selftests/ublk/batch.c
409
unsigned op)
tools/testing/selftests/ublk/batch.c
413
if (op == _IOC_NR(UBLK_U_IO_PREP_IO_CMDS))
tools/testing/selftests/ublk/batch.c
415
else if (op == _IOC_NR(UBLK_U_IO_COMMIT_IO_CMDS)) {
tools/testing/selftests/ublk/batch.c
428
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/batch.c
433
if (op == _IOC_NR(UBLK_U_IO_PREP_IO_CMDS) ||
tools/testing/selftests/ublk/batch.c
434
op == _IOC_NR(UBLK_U_IO_COMMIT_IO_CMDS)) {
tools/testing/selftests/ublk/batch.c
436
ublk_batch_compl_commit_cmd(t, cqe, op);
tools/testing/selftests/ublk/file_backed.c
134
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/file_backed.c
140
__func__, op, cqe->user_data);
tools/testing/selftests/ublk/file_backed.c
141
} else if (op != ublk_cmd_op_nr(UBLK_U_IO_UNREGISTER_IO_BUF)) {
tools/testing/selftests/ublk/file_backed.c
151
if (op == ublk_cmd_op_nr(UBLK_U_IO_REGISTER_IO_BUF))
tools/testing/selftests/ublk/file_backed.c
36
enum io_uring_op op = ublk_to_uring_op(iod, zc | auto_zc);
tools/testing/selftests/ublk/file_backed.c
47
io_uring_prep_rw(op, sqe[0], ublk_get_registered_fd(q, 2),
tools/testing/selftests/ublk/file_backed.c
61
io_uring_prep_rw(op, sqe[0], ublk_get_registered_fd(q, 1) /*fds[1]*/,
tools/testing/selftests/ublk/file_backed.c
80
io_uring_prep_rw(op, sqe[1], ublk_get_registered_fd(q, 1) /*fds[1]*/, 0,
tools/testing/selftests/ublk/kublk.h
333
static inline __u64 build_user_data(unsigned tag, unsigned op,
tools/testing/selftests/ublk/kublk.h
338
ublk_assert(!(tag >> 16) && !(op >> 8) && !(tgt_data >> 16) && !(q_id >> 7));
tools/testing/selftests/ublk/kublk.h
340
return tag | ((__u64)op << 16) | ((__u64)tgt_data << 24) |
tools/testing/selftests/ublk/kublk.h
364
static inline unsigned short ublk_cmd_op_nr(unsigned int op)
tools/testing/selftests/ublk/kublk.h
366
return _IOC_NR(op);
tools/testing/selftests/ublk/null.c
101
if (cqe->res < 0 || op != ublk_cmd_op_nr(UBLK_U_IO_UNREGISTER_IO_BUF)) {
tools/testing/selftests/ublk/null.c
106
__func__, op, cqe->user_data);
tools/testing/selftests/ublk/null.c
110
if (op == ublk_cmd_op_nr(UBLK_U_IO_REGISTER_IO_BUF))
tools/testing/selftests/ublk/null.c
98
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/stripe.c
132
enum io_uring_op op = stripe_to_uring_op(iod, zc | auto_zc);
tools/testing/selftests/ublk/stripe.c
155
io_uring_prep_rw(op, sqe[i],
tools/testing/selftests/ublk/stripe.c
239
unsigned op = user_data_to_op(cqe->user_data);
tools/testing/selftests/ublk/stripe.c
243
if (res < 0 || op != ublk_cmd_op_nr(UBLK_U_IO_UNREGISTER_IO_BUF)) {
tools/testing/selftests/ublk/stripe.c
251
if (op == ublk_cmd_op_nr(UBLK_U_IO_REGISTER_IO_BUF))
tools/testing/selftests/ublk/stripe.c
255
if (op == UBLK_IO_OP_READ || op == UBLK_IO_OP_WRITE) {
tools/testing/selftests/ublk/stripe.c
262
__func__, op, res, s->s[seq].vec->iov_len, tag);
tools/testing/selftests/vDSO/vdso_test_getrandom.c
280
if (info.op == PTRACE_SYSCALL_INFO_ENTRY && info.entry.nr == __NR_getrandom &&
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
119
unsigned int op:8;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
173
unsigned int op:8;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
217
unsigned int op:8;
tools/testing/selftests/vfio/lib/drivers/ioat/hw.h
86
unsigned int op:8;
tools/testing/selftests/vfio/lib/drivers/ioat/ioat.c
189
.ctl_f.op = IOAT_OP_COPY,
tools/testing/selftests/vfio/lib/vfio_pci_driver.c
34
static void vfio_check_driver_op(struct vfio_pci_driver *driver, void *op,
tools/testing/selftests/vfio/lib/vfio_pci_driver.c
38
VFIO_ASSERT_NOT_NULL(op, "Driver has no %s()\n", op_name);
tools/testing/selftests/vfio/lib/vfio_pci_driver.c
39
VFIO_ASSERT_EQ(driver->initialized, op != driver->ops->init);
tools/testing/selftests/vfio/lib/vfio_pci_driver.c
40
VFIO_ASSERT_EQ(driver->memcpy_in_progress, op == driver->ops->memcpy_wait);
tools/testing/vsock/util.c
108
bool vsock_ioctl_int(int fd, unsigned long op, int expected)
tools/testing/vsock/util.c
113
snprintf(name, sizeof(name), "ioctl(%lu)", op);
tools/testing/vsock/util.c
117
ret = ioctl(fd, op, &actual);
tools/testing/vsock/util.h
90
bool vsock_ioctl_int(int fd, unsigned long op, int expected);
tools/usb/ffs-test.c
421
const char *name, *op, *in_name, *out_name;
tools/usb/ffs-test.c
438
op = "write";
tools/usb/ffs-test.c
441
op = "read";
tools/usb/ffs-test.c
447
debug("%s: %s: EOF", name, op);
tools/usb/ffs-test.c
450
debug("%s: %s", name, op);
tools/usb/ffs-test.c
452
warn("%s: %s", name, op);