arch/riscv/include/asm/asm.h
32
#define REG_AMOSWAP_AQ __REG_SEL(amoswap.d.aq, amoswap.w.aq)
arch/riscv/include/asm/bitops.h
300
return __test_and_op_bit_ord(or, __NOP, nr, addr, .aq);
arch/riscv/net/bpf_jit.h
282
static inline u32 rv_amo_insn(u8 funct5, u8 aq, u8 rl, u8 rs2, u8 rs1,
arch/riscv/net/bpf_jit.h
285
u8 funct7 = (funct5 << 2) | (aq << 1) | rl;
arch/riscv/net/bpf_jit.h
565
static inline u32 rv_amoadd_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
567
return rv_amo_insn(0, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
570
static inline u32 rv_amoand_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
572
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
575
static inline u32 rv_amoor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
577
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
580
static inline u32 rv_amoxor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
582
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
585
static inline u32 rv_amoswap_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
587
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
590
static inline u32 rv_lr_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
592
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
595
static inline u32 rv_sc_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
597
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
757
static inline u32 rvzacas_amocas_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
759
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
762
static inline u32 rvzacas_amocas_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
764
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
899
static inline u32 rv_amoadd_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
901
return rv_amo_insn(0, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
904
static inline u32 rv_amoand_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
906
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
909
static inline u32 rv_amoor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
911
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
914
static inline u32 rv_amoxor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
916
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
919
static inline u32 rv_amoswap_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
921
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
924
static inline u32 rv_lr_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
926
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
929
static inline u32 rv_sc_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
931
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 3, rd, 0x2f);
drivers/base/devres.c
1035
va_list aq;
drivers/base/devres.c
1037
va_copy(aq, ap);
drivers/base/devres.c
1038
len = vsnprintf(NULL, 0, fmt, aq);
drivers/base/devres.c
1039
va_end(aq);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1474
struct aead_request *aq = req->aead_req.aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1476
return sec_cipher_map(ctx, req, aq->src, aq->dst);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1481
struct aead_request *aq = req->aead_req.aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1483
sec_cipher_unmap(ctx, req, aq->src, aq->dst);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1759
struct aead_request *aq = a_req->aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1760
struct crypto_aead *tfm = crypto_aead_reqtfm(aq);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1776
sec_sqe->type2.alen_ivllen = cpu_to_le32(aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1778
sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1787
struct aead_request *aq = a_req->aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1788
struct crypto_aead *tfm = crypto_aead_reqtfm(aq);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1804
sqe3->a_len_key = cpu_to_le32(aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1806
sqe3->cipher_src_offset = cpu_to_le16((u16)aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1815
struct aead_request *aq = a_req->aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1816
struct crypto_aead *tfm = crypto_aead_reqtfm(aq);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1836
sec_sqe->type2.alen_ivllen = cpu_to_le32(c_req->c_len + aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1838
sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1869
struct aead_request *aq = a_req->aead_req;
drivers/crypto/hisilicon/sec2/sec_crypto.c
1870
struct crypto_aead *tfm = crypto_aead_reqtfm(aq);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1891
sqe3->a_len_key = cpu_to_le32(c_req->c_len + aq->assoclen);
drivers/crypto/hisilicon/sec2/sec_crypto.c
1893
sqe3->cipher_src_offset = cpu_to_le16((u16)aq->assoclen);
drivers/infiniband/hw/efa/efa_com.c
1097
edev->aq.completion_timeout = timeout * 100000;
drivers/infiniband/hw/efa/efa_com.c
1099
edev->aq.completion_timeout = ADMIN_CMD_TIMEOUT_US;
drivers/infiniband/hw/efa/efa_com.c
1108
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
1123
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com.c
1142
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
1150
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com.c
130
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
131
struct efa_com_admin_sq *sq = &aq->sq;
drivers/infiniband/hw/efa/efa_com.c
132
u16 size = aq->depth * sizeof(*sq->entries);
drivers/infiniband/hw/efa/efa_com.c
138
dma_alloc_coherent(aq->dmadev, size, &sq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
156
EFA_SET(&aq_caps, EFA_REGS_AQ_CAPS_AQ_DEPTH, aq->depth);
drivers/infiniband/hw/efa/efa_com.c
167
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
168
struct efa_com_admin_cq *cq = &aq->cq;
drivers/infiniband/hw/efa/efa_com.c
169
u16 size = aq->depth * sizeof(*cq->entries);
drivers/infiniband/hw/efa/efa_com.c
175
dma_alloc_coherent(aq->dmadev, size, &cq->dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
190
EFA_SET(&acq_caps, EFA_REGS_ACQ_CAPS_ACQ_DEPTH, aq->depth);
drivers/infiniband/hw/efa/efa_com.c
194
aq->msix_vector_idx);
drivers/infiniband/hw/efa/efa_com.c
247
static u16 efa_com_alloc_ctx_id(struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
251
spin_lock(&aq->comp_ctx_lock);
drivers/infiniband/hw/efa/efa_com.c
252
ctx_id = aq->comp_ctx_pool[aq->comp_ctx_pool_next];
drivers/infiniband/hw/efa/efa_com.c
253
aq->comp_ctx_pool_next++;
drivers/infiniband/hw/efa/efa_com.c
254
spin_unlock(&aq->comp_ctx_lock);
drivers/infiniband/hw/efa/efa_com.c
259
static void efa_com_dealloc_ctx_id(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
262
spin_lock(&aq->comp_ctx_lock);
drivers/infiniband/hw/efa/efa_com.c
263
aq->comp_ctx_pool_next--;
drivers/infiniband/hw/efa/efa_com.c
264
aq->comp_ctx_pool[aq->comp_ctx_pool_next] = ctx_id;
drivers/infiniband/hw/efa/efa_com.c
265
spin_unlock(&aq->comp_ctx_lock);
drivers/infiniband/hw/efa/efa_com.c
268
static struct efa_comp_ctx *efa_com_alloc_comp_ctx(struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
273
ctx_id = efa_com_alloc_ctx_id(aq);
drivers/infiniband/hw/efa/efa_com.c
275
comp_ctx = &aq->comp_ctx[ctx_id];
drivers/infiniband/hw/efa/efa_com.c
277
efa_com_dealloc_ctx_id(aq, ctx_id);
drivers/infiniband/hw/efa/efa_com.c
278
ibdev_err_ratelimited(aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
285
ibdev_dbg(aq->efa_dev, "Take completion context[%u]\n", ctx_id);
drivers/infiniband/hw/efa/efa_com.c
289
static inline u16 efa_com_get_comp_ctx_id(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
292
return comp_ctx - aq->comp_ctx;
drivers/infiniband/hw/efa/efa_com.c
295
static inline void efa_com_dealloc_comp_ctx(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
298
u16 ctx_id = efa_com_get_comp_ctx_id(aq, comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
300
ibdev_dbg(aq->efa_dev, "Put completion context[%u]\n", ctx_id);
drivers/infiniband/hw/efa/efa_com.c
302
efa_com_dealloc_ctx_id(aq, ctx_id);
drivers/infiniband/hw/efa/efa_com.c
305
static inline struct efa_comp_ctx *efa_com_get_comp_ctx_by_cmd_id(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
308
u16 ctx_id = cmd_id & (aq->depth - 1);
drivers/infiniband/hw/efa/efa_com.c
310
return &aq->comp_ctx[ctx_id];
drivers/infiniband/hw/efa/efa_com.c
313
static void __efa_com_submit_admin_cmd(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
326
queue_size_mask = aq->depth - 1;
drivers/infiniband/hw/efa/efa_com.c
327
pi = aq->sq.pc & queue_size_mask;
drivers/infiniband/hw/efa/efa_com.c
328
ctx_id = efa_com_get_comp_ctx_id(aq, comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
332
cmd_id |= aq->sq.pc << ilog2(aq->depth);
drivers/infiniband/hw/efa/efa_com.c
337
EFA_ADMIN_AQ_COMMON_DESC_PHASE, aq->sq.phase);
drivers/infiniband/hw/efa/efa_com.c
347
aqe = &aq->sq.entries[pi];
drivers/infiniband/hw/efa/efa_com.c
351
aq->sq.pc++;
drivers/infiniband/hw/efa/efa_com.c
352
atomic64_inc(&aq->stats.submitted_cmd);
drivers/infiniband/hw/efa/efa_com.c
354
if ((aq->sq.pc & queue_size_mask) == 0)
drivers/infiniband/hw/efa/efa_com.c
355
aq->sq.phase = !aq->sq.phase;
drivers/infiniband/hw/efa/efa_com.c
358
writel(aq->sq.pc, aq->sq.db_addr);
drivers/infiniband/hw/efa/efa_com.c
361
static inline int efa_com_init_comp_ctxt(struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
363
size_t pool_size = aq->depth * sizeof(*aq->comp_ctx_pool);
drivers/infiniband/hw/efa/efa_com.c
364
size_t size = aq->depth * sizeof(struct efa_comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
368
aq->comp_ctx = devm_kzalloc(aq->dmadev, size, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
369
aq->comp_ctx_pool = devm_kzalloc(aq->dmadev, pool_size, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
370
if (!aq->comp_ctx || !aq->comp_ctx_pool) {
drivers/infiniband/hw/efa/efa_com.c
371
devm_kfree(aq->dmadev, aq->comp_ctx_pool);
drivers/infiniband/hw/efa/efa_com.c
372
devm_kfree(aq->dmadev, aq->comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
376
for (i = 0; i < aq->depth; i++) {
drivers/infiniband/hw/efa/efa_com.c
377
comp_ctx = &aq->comp_ctx[i];
drivers/infiniband/hw/efa/efa_com.c
381
aq->comp_ctx_pool[i] = i;
drivers/infiniband/hw/efa/efa_com.c
384
spin_lock_init(&aq->comp_ctx_lock);
drivers/infiniband/hw/efa/efa_com.c
386
aq->comp_ctx_pool_next = 0;
drivers/infiniband/hw/efa/efa_com.c
391
static int efa_com_submit_admin_cmd(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
398
spin_lock(&aq->sq.lock);
drivers/infiniband/hw/efa/efa_com.c
399
if (!test_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state)) {
drivers/infiniband/hw/efa/efa_com.c
400
ibdev_err_ratelimited(aq->efa_dev, "Admin queue is closed\n");
drivers/infiniband/hw/efa/efa_com.c
401
spin_unlock(&aq->sq.lock);
drivers/infiniband/hw/efa/efa_com.c
405
__efa_com_submit_admin_cmd(aq, comp_ctx, cmd, cmd_size_in_bytes, comp,
drivers/infiniband/hw/efa/efa_com.c
407
spin_unlock(&aq->sq.lock);
drivers/infiniband/hw/efa/efa_com.c
412
static int efa_com_handle_single_admin_completion(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
421
comp_ctx = efa_com_get_comp_ctx_by_cmd_id(aq, cmd_id);
drivers/infiniband/hw/efa/efa_com.c
423
ibdev_err(aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
425
cmd_id, comp_ctx->status, aq->sq.pc, aq->sq.cc,
drivers/infiniband/hw/efa/efa_com.c
426
aq->cq.cc);
drivers/infiniband/hw/efa/efa_com.c
433
if (!test_bit(EFA_AQ_STATE_POLLING_BIT, &aq->state))
drivers/infiniband/hw/efa/efa_com.c
439
static void efa_com_handle_admin_completion(struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
448
queue_size_mask = aq->depth - 1;
drivers/infiniband/hw/efa/efa_com.c
450
ci = aq->cq.cc & queue_size_mask;
drivers/infiniband/hw/efa/efa_com.c
451
phase = aq->cq.phase;
drivers/infiniband/hw/efa/efa_com.c
453
cqe = &aq->cq.entries[ci];
drivers/infiniband/hw/efa/efa_com.c
463
err = efa_com_handle_single_admin_completion(aq, cqe);
drivers/infiniband/hw/efa/efa_com.c
467
aq->cq.cc++;
drivers/infiniband/hw/efa/efa_com.c
469
if (ci == aq->depth) {
drivers/infiniband/hw/efa/efa_com.c
474
cqe = &aq->cq.entries[ci];
drivers/infiniband/hw/efa/efa_com.c
477
aq->cq.phase = phase;
drivers/infiniband/hw/efa/efa_com.c
478
aq->sq.cc += comp_cmds;
drivers/infiniband/hw/efa/efa_com.c
479
atomic64_add(comp_cmds, &aq->stats.completed_cmd);
drivers/infiniband/hw/efa/efa_com.c
502
struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
507
timeout = jiffies + usecs_to_jiffies(aq->completion_timeout);
drivers/infiniband/hw/efa/efa_com.c
510
spin_lock_irqsave(&aq->cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.c
511
efa_com_handle_admin_completion(aq);
drivers/infiniband/hw/efa/efa_com.c
512
spin_unlock_irqrestore(&aq->cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.c
519
aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
522
atomic64_inc(&aq->stats.no_completion);
drivers/infiniband/hw/efa/efa_com.c
524
clear_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
528
msleep(aq->poll_interval);
drivers/infiniband/hw/efa/efa_com.c
536
struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
541
usecs_to_jiffies(aq->completion_timeout));
drivers/infiniband/hw/efa/efa_com.c
550
spin_lock_irqsave(&aq->cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.c
551
efa_com_handle_admin_completion(aq);
drivers/infiniband/hw/efa/efa_com.c
552
spin_unlock_irqrestore(&aq->cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.c
554
atomic64_inc(&aq->stats.no_completion);
drivers/infiniband/hw/efa/efa_com.c
558
aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
562
comp_ctx->cmd_id, aq->sq.pc, aq->sq.cc,
drivers/infiniband/hw/efa/efa_com.c
563
aq->cq.cc);
drivers/infiniband/hw/efa/efa_com.c
566
aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
570
comp_ctx->cmd_id, aq->sq.pc, aq->sq.cc,
drivers/infiniband/hw/efa/efa_com.c
571
aq->cq.cc);
drivers/infiniband/hw/efa/efa_com.c
573
clear_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
590
struct efa_com_admin_queue *aq)
drivers/infiniband/hw/efa/efa_com.c
592
if (test_bit(EFA_AQ_STATE_POLLING_BIT, &aq->state))
drivers/infiniband/hw/efa/efa_com.c
593
return efa_com_wait_and_process_admin_cq_polling(comp_ctx, aq);
drivers/infiniband/hw/efa/efa_com.c
595
return efa_com_wait_and_process_admin_cq_interrupts(comp_ctx, aq);
drivers/infiniband/hw/efa/efa_com.c
611
int efa_com_cmd_exec(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com.c
623
down(&aq->avail_cmds);
drivers/infiniband/hw/efa/efa_com.c
625
ibdev_dbg(aq->efa_dev, "%s (opcode %d)\n",
drivers/infiniband/hw/efa/efa_com.c
629
comp_ctx = efa_com_alloc_comp_ctx(aq);
drivers/infiniband/hw/efa/efa_com.c
631
clear_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
632
up(&aq->avail_cmds);
drivers/infiniband/hw/efa/efa_com.c
636
err = efa_com_submit_admin_cmd(aq, comp_ctx, cmd, cmd_size, comp, comp_size);
drivers/infiniband/hw/efa/efa_com.c
639
aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
644
efa_com_dealloc_comp_ctx(aq, comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
645
up(&aq->avail_cmds);
drivers/infiniband/hw/efa/efa_com.c
646
atomic64_inc(&aq->stats.cmd_err);
drivers/infiniband/hw/efa/efa_com.c
650
err = efa_com_wait_and_process_admin_cq(comp_ctx, aq);
drivers/infiniband/hw/efa/efa_com.c
653
aq->efa_dev,
drivers/infiniband/hw/efa/efa_com.c
657
atomic64_inc(&aq->stats.cmd_err);
drivers/infiniband/hw/efa/efa_com.c
660
efa_com_dealloc_comp_ctx(aq, comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
661
up(&aq->avail_cmds);
drivers/infiniband/hw/efa/efa_com.c
672
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
674
struct efa_com_admin_cq *cq = &aq->cq;
drivers/infiniband/hw/efa/efa_com.c
675
struct efa_com_admin_sq *sq = &aq->sq;
drivers/infiniband/hw/efa/efa_com.c
678
clear_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
680
devm_kfree(edev->dmadev, aq->comp_ctx_pool);
drivers/infiniband/hw/efa/efa_com.c
681
devm_kfree(edev->dmadev, aq->comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
683
size = aq->depth * sizeof(*sq->entries);
drivers/infiniband/hw/efa/efa_com.c
686
size = aq->depth * sizeof(*cq->entries);
drivers/infiniband/hw/efa/efa_com.c
709
set_bit(EFA_AQ_STATE_POLLING_BIT, &edev->aq.state);
drivers/infiniband/hw/efa/efa_com.c
711
clear_bit(EFA_AQ_STATE_POLLING_BIT, &edev->aq.state);
drivers/infiniband/hw/efa/efa_com.c
716
atomic64_t *s = (atomic64_t *)&edev->aq.stats;
drivers/infiniband/hw/efa/efa_com.c
719
for (i = 0; i < sizeof(edev->aq.stats) / sizeof(*s); i++, s++)
drivers/infiniband/hw/efa/efa_com.c
736
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com.c
749
aq->depth = EFA_ADMIN_QUEUE_DEPTH;
drivers/infiniband/hw/efa/efa_com.c
751
aq->dmadev = edev->dmadev;
drivers/infiniband/hw/efa/efa_com.c
752
aq->efa_dev = edev->efa_dev;
drivers/infiniband/hw/efa/efa_com.c
753
set_bit(EFA_AQ_STATE_POLLING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
755
sema_init(&aq->avail_cmds, aq->depth);
drivers/infiniband/hw/efa/efa_com.c
759
err = efa_com_init_comp_ctxt(aq);
drivers/infiniband/hw/efa/efa_com.c
781
aq->completion_timeout = timeout * 100000;
drivers/infiniband/hw/efa/efa_com.c
783
aq->completion_timeout = ADMIN_CMD_TIMEOUT_US;
drivers/infiniband/hw/efa/efa_com.c
785
aq->poll_interval = EFA_POLL_INTERVAL_MS;
drivers/infiniband/hw/efa/efa_com.c
787
set_bit(EFA_AQ_STATE_RUNNING_BIT, &aq->state);
drivers/infiniband/hw/efa/efa_com.c
792
dma_free_coherent(edev->dmadev, aq->depth * sizeof(*aq->cq.entries),
drivers/infiniband/hw/efa/efa_com.c
793
aq->cq.entries, aq->cq.dma_addr);
drivers/infiniband/hw/efa/efa_com.c
795
dma_free_coherent(edev->dmadev, aq->depth * sizeof(*aq->sq.entries),
drivers/infiniband/hw/efa/efa_com.c
796
aq->sq.entries, aq->sq.dma_addr);
drivers/infiniband/hw/efa/efa_com.c
798
devm_kfree(edev->dmadev, aq->comp_ctx);
drivers/infiniband/hw/efa/efa_com.c
816
spin_lock_irqsave(&edev->aq.cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.c
817
efa_com_handle_admin_completion(&edev->aq);
drivers/infiniband/hw/efa/efa_com.c
818
spin_unlock_irqrestore(&edev->aq.cq.lock, flags);
drivers/infiniband/hw/efa/efa_com.h
107
struct efa_com_admin_queue aq;
drivers/infiniband/hw/efa/efa_com.h
171
int efa_com_cmd_exec(struct efa_com_admin_queue *aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
107
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
133
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
139
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
15
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
160
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
183
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
208
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
214
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
235
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
264
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
294
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
301
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
321
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
330
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
352
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
359
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
39
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
395
struct efa_com_admin_queue *aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
405
aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
419
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
568
struct efa_com_admin_queue *aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
578
aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
592
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
64
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
664
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
671
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
690
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
698
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
716
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
723
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
742
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
750
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
769
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_com_cmd.c
784
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
79
err = efa_com_cmd_exec(aq,
drivers/infiniband/hw/efa/efa_com_cmd.c
99
struct efa_com_admin_queue *aq = &edev->aq;
drivers/infiniband/hw/efa/efa_main.c
239
edev->aq.poll_interval = hw_hints->poll_interval;
drivers/infiniband/hw/efa/efa_main.c
242
edev->aq.completion_timeout =
drivers/infiniband/hw/efa/efa_main.c
611
edev->aq.msix_vector_idx = dev->admin_msix_vector_idx;
drivers/infiniband/hw/efa/efa_verbs.c
2204
struct efa_com_stats_admin *as = &dev->edev.aq.stats;
drivers/infiniband/hw/ionic/ionic_admin.c
105
lockdep_assert_held(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
1069
struct ionic_aq *aq;
drivers/infiniband/hw/ionic/ionic_admin.c
107
if (atomic_read(&aq->admin_state) == IONIC_ADMIN_KILLED) {
drivers/infiniband/hw/ionic/ionic_admin.c
108
list_for_each_entry_safe(wr, wr_next, &aq->wr_prod, aq_ent) {
drivers/infiniband/hw/ionic/ionic_admin.c
110
aq->q_wr[wr->status].wr = NULL;
drivers/infiniband/hw/ionic/ionic_admin.c
111
wr->status = atomic_read(&aq->admin_state);
drivers/infiniband/hw/ionic/ionic_admin.c
114
INIT_LIST_HEAD(&aq->wr_prod);
drivers/infiniband/hw/ionic/ionic_admin.c
1152
aq = ionic_create_rdma_adminq(dev, aq_i + dev->lif_cfg.aq_base,
drivers/infiniband/hw/ionic/ionic_admin.c
1154
if (IS_ERR(aq)) {
drivers/infiniband/hw/ionic/ionic_admin.c
1159
rc = PTR_ERR(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
116
list_for_each_entry_safe(wr, wr_next, &aq->wr_post, aq_ent) {
drivers/infiniband/hw/ionic/ionic_admin.c
1163
"failed to create aq %pe\n", aq);
drivers/infiniband/hw/ionic/ionic_admin.c
1169
aq_i, dev->lif_cfg.aq_count, aq);
drivers/infiniband/hw/ionic/ionic_admin.c
1173
vcq->ibcq.cq_context = aq;
drivers/infiniband/hw/ionic/ionic_admin.c
1174
aq->vcq = vcq;
drivers/infiniband/hw/ionic/ionic_admin.c
1176
atomic_set(&aq->admin_state, IONIC_ADMIN_ACTIVE);
drivers/infiniband/hw/ionic/ionic_admin.c
1177
dev->aq_vec[aq_i] = aq;
drivers/infiniband/hw/ionic/ionic_admin.c
118
wr->status = atomic_read(&aq->admin_state);
drivers/infiniband/hw/ionic/ionic_admin.c
1191
struct ionic_aq *aq;
drivers/infiniband/hw/ionic/ionic_admin.c
1204
aq = dev->aq_vec[--dev->lif_cfg.aq_count];
drivers/infiniband/hw/ionic/ionic_admin.c
1205
vcq = aq->vcq;
drivers/infiniband/hw/ionic/ionic_admin.c
1207
cancel_work_sync(&aq->work);
drivers/infiniband/hw/ionic/ionic_admin.c
1209
__ionic_destroy_rdma_adminq(dev, aq);
drivers/infiniband/hw/ionic/ionic_admin.c
121
INIT_LIST_HEAD(&aq->wr_post);
drivers/infiniband/hw/ionic/ionic_admin.c
139
if (unlikely(qid != aq->aqid)) {
drivers/infiniband/hw/ionic/ionic_admin.c
145
if (unlikely(be16_to_cpu(cqe->admin.cmd_idx) != aq->q.cons)) {
drivers/infiniband/hw/ionic/ionic_admin.c
149
aq->q.cons, qid);
drivers/infiniband/hw/ionic/ionic_admin.c
153
if (unlikely(ionic_queue_empty(&aq->q))) {
drivers/infiniband/hw/ionic/ionic_admin.c
159
wr = aq->q_wr[aq->q.cons].wr;
drivers/infiniband/hw/ionic/ionic_admin.c
161
aq->q_wr[aq->q.cons].wr = NULL;
drivers/infiniband/hw/ionic/ionic_admin.c
165
wr->status = atomic_read(&aq->admin_state);
drivers/infiniband/hw/ionic/ionic_admin.c
169
ionic_queue_consume_entries(&aq->q,
drivers/infiniband/hw/ionic/ionic_admin.c
170
aq->q_wr[aq->q.cons].wqe_strides);
drivers/infiniband/hw/ionic/ionic_admin.c
178
ionic_admin_reset_wdog(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
182
queue_work(ionic_evt_workq, &aq->work);
drivers/infiniband/hw/ionic/ionic_admin.c
183
} else if (!aq->armed) {
drivers/infiniband/hw/ionic/ionic_admin.c
184
aq->armed = true;
drivers/infiniband/hw/ionic/ionic_admin.c
189
queue_work(ionic_evt_workq, &aq->work);
drivers/infiniband/hw/ionic/ionic_admin.c
192
if (atomic_read(&aq->admin_state) != IONIC_ADMIN_ACTIVE)
drivers/infiniband/hw/ionic/ionic_admin.c
195
old_prod = aq->q.prod;
drivers/infiniband/hw/ionic/ionic_admin.c
197
if (ionic_queue_empty(&aq->q) && !list_empty(&aq->wr_post))
drivers/infiniband/hw/ionic/ionic_admin.c
198
ionic_admin_reset_wdog(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
200
if (list_empty(&aq->wr_post))
drivers/infiniband/hw/ionic/ionic_admin.c
208
wr = list_first_entry(&aq->wr_post, struct ionic_admin_wr,
drivers/infiniband/hw/ionic/ionic_admin.c
211
(ADMIN_WQE_STRIDE - 1)) >> aq->q.stride_log2;
drivers/infiniband/hw/ionic/ionic_admin.c
212
avlbl_strides = ionic_queue_length_remaining(&aq->q);
drivers/infiniband/hw/ionic/ionic_admin.c
217
list_move(&wr->aq_ent, &aq->wr_prod);
drivers/infiniband/hw/ionic/ionic_admin.c
218
wr->status = aq->q.prod;
drivers/infiniband/hw/ionic/ionic_admin.c
219
aq->q_wr[aq->q.prod].wr = wr;
drivers/infiniband/hw/ionic/ionic_admin.c
220
aq->q_wr[aq->q.prod].wqe_strides = wr_strides;
drivers/infiniband/hw/ionic/ionic_admin.c
226
memcpy(ionic_queue_at_prod(&aq->q), &wr->wqe,
drivers/infiniband/hw/ionic/ionic_admin.c
231
memcpy(ionic_queue_at_prod(&aq->q) + ADMIN_WQE_HDR_LEN,
drivers/infiniband/hw/ionic/ionic_admin.c
234
aq->q.prod, wr_strides);
drivers/infiniband/hw/ionic/ionic_admin.c
236
ionic_queue_at_prod(&aq->q),
drivers/infiniband/hw/ionic/ionic_admin.c
237
BIT(aq->q.stride_log2), true);
drivers/infiniband/hw/ionic/ionic_admin.c
238
ionic_queue_produce(&aq->q);
drivers/infiniband/hw/ionic/ionic_admin.c
247
memcpy(ionic_queue_at_prod(&aq->q), src + i,
drivers/infiniband/hw/ionic/ionic_admin.c
250
ionic_queue_at_prod(&aq->q),
drivers/infiniband/hw/ionic/ionic_admin.c
251
BIT(aq->q.stride_log2), true);
drivers/infiniband/hw/ionic/ionic_admin.c
252
ionic_queue_produce(&aq->q);
drivers/infiniband/hw/ionic/ionic_admin.c
254
} while (!list_empty(&aq->wr_post));
drivers/infiniband/hw/ionic/ionic_admin.c
256
if (old_prod != aq->q.prod)
drivers/infiniband/hw/ionic/ionic_admin.c
258
ionic_queue_dbell_val(&aq->q));
drivers/infiniband/hw/ionic/ionic_admin.c
265
struct ionic_aq *aq, *bad_aq = NULL;
drivers/infiniband/hw/ionic/ionic_admin.c
273
aq = dev->aq_vec[i];
drivers/infiniband/hw/ionic/ionic_admin.c
275
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
277
if (ionic_queue_empty(&aq->q))
drivers/infiniband/hw/ionic/ionic_admin.c
283
if (time_is_after_eq_jiffies(aq->stamp + IONIC_ADMIN_WARN))
drivers/infiniband/hw/ionic/ionic_admin.c
288
pos = aq->q.cons;
drivers/infiniband/hw/ionic/ionic_admin.c
289
ionic_admin_poll_locked(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
290
if (pos != aq->q.cons) {
drivers/infiniband/hw/ionic/ionic_admin.c
292
"missed event for acq %d\n", aq->cqid);
drivers/infiniband/hw/ionic/ionic_admin.c
296
if (time_is_after_eq_jiffies(aq->stamp +
drivers/infiniband/hw/ionic/ionic_admin.c
300
(u32)jiffies_to_msecs(jiffies - aq->stamp));
drivers/infiniband/hw/ionic/ionic_admin.c
305
bad_aq = aq;
drivers/infiniband/hw/ionic/ionic_admin.c
308
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
31
static void ionic_admin_timedout(struct ionic_aq *aq)
drivers/infiniband/hw/ionic/ionic_admin.c
321
struct ionic_aq *aq = container_of(ws, struct ionic_aq, work);
drivers/infiniband/hw/ionic/ionic_admin.c
324
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
325
ionic_admin_poll_locked(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
326
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
329
static void ionic_admin_post_aq(struct ionic_aq *aq, struct ionic_admin_wr *wr)
drivers/infiniband/hw/ionic/ionic_admin.c
33
struct ionic_ibdev *dev = aq->dev;
drivers/infiniband/hw/ionic/ionic_admin.c
335
wr->aq = aq;
drivers/infiniband/hw/ionic/ionic_admin.c
337
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
338
poll = list_empty(&aq->wr_post);
drivers/infiniband/hw/ionic/ionic_admin.c
339
list_add(&wr->aq_ent, &aq->wr_post);
drivers/infiniband/hw/ionic/ionic_admin.c
341
ionic_admin_poll_locked(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
342
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
356
struct ionic_aq *aq = wr->aq;
drivers/infiniband/hw/ionic/ionic_admin.c
359
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
364
aq->q_wr[wr->status].wr = NULL;
drivers/infiniband/hw/ionic/ionic_admin.c
367
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
37
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
372
struct ionic_aq *aq = wr->aq;
drivers/infiniband/hw/ionic/ionic_admin.c
38
if (ionic_queue_empty(&aq->q))
drivers/infiniband/hw/ionic/ionic_admin.c
382
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
383
ionic_admin_poll_locked(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
384
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
391
ionic_admin_timedout(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
42
if (atomic_read(&aq->admin_state) < IONIC_ADMIN_KILLED)
drivers/infiniband/hw/ionic/ionic_admin.c
46
aq->aqid, (u32)jiffies_to_msecs(jiffies - aq->stamp));
drivers/infiniband/hw/ionic/ionic_admin.c
48
pos = (aq->q.prod - 1) & aq->q.mask;
drivers/infiniband/hw/ionic/ionic_admin.c
49
if (pos == aq->q.cons)
drivers/infiniband/hw/ionic/ionic_admin.c
494
struct ionic_aq *aq = cq_context;
drivers/infiniband/hw/ionic/ionic_admin.c
497
spin_lock_irqsave(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
498
aq->armed = false;
drivers/infiniband/hw/ionic/ionic_admin.c
499
if (atomic_read(&aq->admin_state) < IONIC_ADMIN_KILLED)
drivers/infiniband/hw/ionic/ionic_admin.c
500
queue_work(ionic_evt_workq, &aq->work);
drivers/infiniband/hw/ionic/ionic_admin.c
501
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
506
struct ionic_aq *aq = cq_context;
drivers/infiniband/hw/ionic/ionic_admin.c
508
ibdev_err(&aq->dev->ibdev, "admincq event %d\n", event->event);
drivers/infiniband/hw/ionic/ionic_admin.c
54
ionic_queue_at(&aq->q, pos),
drivers/infiniband/hw/ionic/ionic_admin.c
55
BIT(aq->q.stride_log2), true);
drivers/infiniband/hw/ionic/ionic_admin.c
558
struct ionic_aq *aq;
drivers/infiniband/hw/ionic/ionic_admin.c
561
aq = kzalloc_obj(*aq);
drivers/infiniband/hw/ionic/ionic_admin.c
562
if (!aq)
drivers/infiniband/hw/ionic/ionic_admin.c
565
atomic_set(&aq->admin_state, IONIC_ADMIN_KILLED);
drivers/infiniband/hw/ionic/ionic_admin.c
566
aq->dev = dev;
drivers/infiniband/hw/ionic/ionic_admin.c
567
aq->aqid = aqid;
drivers/infiniband/hw/ionic/ionic_admin.c
568
aq->cqid = cqid;
drivers/infiniband/hw/ionic/ionic_admin.c
569
spin_lock_init(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
571
rc = ionic_queue_init(&aq->q, dev->lif_cfg.hwdev, IONIC_EQ_DEPTH,
drivers/infiniband/hw/ionic/ionic_admin.c
576
ionic_queue_dbell_init(&aq->q, aq->aqid);
drivers/infiniband/hw/ionic/ionic_admin.c
578
aq->q_wr = kzalloc_objs(*aq->q_wr, (u32)aq->q.mask + 1);
drivers/infiniband/hw/ionic/ionic_admin.c
579
if (!aq->q_wr) {
drivers/infiniband/hw/ionic/ionic_admin.c
58
spin_unlock_irqrestore(&aq->lock, irqflags);
drivers/infiniband/hw/ionic/ionic_admin.c
584
INIT_LIST_HEAD(&aq->wr_prod);
drivers/infiniband/hw/ionic/ionic_admin.c
585
INIT_LIST_HEAD(&aq->wr_post);
drivers/infiniband/hw/ionic/ionic_admin.c
587
INIT_WORK(&aq->work, ionic_admin_work);
drivers/infiniband/hw/ionic/ionic_admin.c
588
aq->armed = false;
drivers/infiniband/hw/ionic/ionic_admin.c
590
return aq;
drivers/infiniband/hw/ionic/ionic_admin.c
593
ionic_queue_destroy(&aq->q, dev->lif_cfg.hwdev);
drivers/infiniband/hw/ionic/ionic_admin.c
595
kfree(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
601
struct ionic_aq *aq)
drivers/infiniband/hw/ionic/ionic_admin.c
603
kfree(aq->q_wr);
drivers/infiniband/hw/ionic/ionic_admin.c
604
ionic_queue_destroy(&aq->q, dev->lif_cfg.hwdev);
drivers/infiniband/hw/ionic/ionic_admin.c
605
kfree(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
611
struct ionic_aq *aq;
drivers/infiniband/hw/ionic/ionic_admin.c
614
aq = __ionic_create_rdma_adminq(dev, aqid, cqid);
drivers/infiniband/hw/ionic/ionic_admin.c
615
if (IS_ERR(aq))
drivers/infiniband/hw/ionic/ionic_admin.c
616
return aq;
drivers/infiniband/hw/ionic/ionic_admin.c
618
rc = ionic_rdma_queue_devcmd(dev, &aq->q, aq->aqid, aq->cqid,
drivers/infiniband/hw/ionic/ionic_admin.c
623
return aq;
drivers/infiniband/hw/ionic/ionic_admin.c
626
__ionic_destroy_rdma_adminq(dev, aq);
drivers/infiniband/hw/ionic/ionic_admin.c
680
struct ionic_aq *aq = dev->aq_vec[i];
drivers/infiniband/hw/ionic/ionic_admin.c
682
spin_lock(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
683
if (atomic_read(&aq->admin_state) != IONIC_ADMIN_KILLED) {
drivers/infiniband/hw/ionic/ionic_admin.c
684
atomic_set(&aq->admin_state, IONIC_ADMIN_KILLED);
drivers/infiniband/hw/ionic/ionic_admin.c
686
ionic_admin_poll_locked(aq);
drivers/infiniband/hw/ionic/ionic_admin.c
689
spin_unlock(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
70
static void ionic_admin_reset_wdog(struct ionic_aq *aq)
drivers/infiniband/hw/ionic/ionic_admin.c
72
if (atomic_read(&aq->admin_state) == IONIC_ADMIN_KILLED)
drivers/infiniband/hw/ionic/ionic_admin.c
739
struct ionic_aq *aq = dev->aq_vec[i];
drivers/infiniband/hw/ionic/ionic_admin.c
741
spin_lock(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
743
if (atomic_read(&aq->admin_state) == IONIC_ADMIN_ACTIVE)
drivers/infiniband/hw/ionic/ionic_admin.c
744
atomic_set(&aq->admin_state, IONIC_ADMIN_PAUSED);
drivers/infiniband/hw/ionic/ionic_admin.c
745
spin_unlock(&aq->lock);
drivers/infiniband/hw/ionic/ionic_admin.c
75
aq->stamp = jiffies;
drivers/infiniband/hw/ionic/ionic_admin.c
76
ionic_admin_reset_dwork(aq->dev);
drivers/infiniband/hw/ionic/ionic_admin.c
94
static void ionic_admin_poll_locked(struct ionic_aq *aq)
drivers/infiniband/hw/ionic/ionic_admin.c
96
struct ionic_cq *cq = &aq->vcq->cq[0];
drivers/infiniband/hw/ionic/ionic_admin.c
98
struct ionic_ibdev *dev = aq->dev;
drivers/infiniband/hw/ionic/ionic_ibdev.h
143
struct ionic_aq *aq;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
121
struct ipu6_isys_queue *aq =
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
125
av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
127
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
130
list_add(&ib->head, &aq->active);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
132
list_add_tail(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
133
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
156
struct ipu6_isys_queue *aq;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
161
list_for_each_entry(aq, &stream->queues, node) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
162
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
165
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
166
list_for_each_entry_safe(ib, ib_safe, &aq->active, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
176
list_add(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
185
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
198
struct ipu6_isys_queue *aq;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
207
list_for_each_entry(aq, &stream->queues, node) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
210
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
211
if (list_empty(&aq->incoming)) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
212
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
218
ib = list_last_entry(&aq->incoming,
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
222
ipu6_isys_queue_to_video(aq)->vdev.name,
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
226
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
240
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
245
set->output_pins[aq->fw_output].addr = ivb->dma_addr;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
246
set->output_pins[aq->fw_output].out_buf_id = vb->index + 1;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
345
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
346
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
365
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
366
list_add(&ib->head, &aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
367
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
421
static int ipu6_isys_link_fmt_validate(struct ipu6_isys_queue *aq)
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
424
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
472
static void return_buffers(struct ipu6_isys_queue *aq,
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
475
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
480
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
481
while (!list_empty(&aq->incoming)) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
484
ib = list_first_entry(&aq->incoming, struct ipu6_isys_buffer,
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
488
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
492
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
500
while (!list_empty(&aq->active)) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
503
ib = list_first_entry(&aq->active, struct ipu6_isys_buffer,
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
508
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
512
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
516
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
534
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
535
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
568
ret = ipu6_isys_link_fmt_validate(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
593
list_add(&aq->node, &stream->queues);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
60
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
61
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
618
list_del(&aq->node);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
629
return_buffers(aq, VB2_BUF_STATE_QUEUED);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
636
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
637
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
647
list_del(&aq->node);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
656
return_buffers(aq, VB2_BUF_STATE_ERROR);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
709
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
710
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
748
struct ipu6_isys_queue *aq = stream->output_pins_queue[pin_id];
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
757
spin_lock_irqsave(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
758
if (list_empty(&aq->active)) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
759
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
764
list_for_each_entry_reverse(ib, &aq->active, head) {
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
795
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
806
spin_unlock_irqrestore(&aq->lock, flags);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
81
struct ipu6_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
82
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
828
int ipu6_isys_queue_init(struct ipu6_isys_queue *aq)
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
830
struct ipu6_isys *isys = ipu6_isys_queue_to_video(aq)->isys;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
831
struct ipu6_isys_video *av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
836
if (!aq->vbq.io_modes)
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
837
aq->vbq.io_modes = VB2_MMAP | VB2_DMABUF;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
839
aq->vbq.drv_priv = isys;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
840
aq->vbq.ops = &ipu6_isys_queue_ops;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
841
aq->vbq.lock = &av->mutex;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
842
aq->vbq.mem_ops = &vb2_dma_sg_memops;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
843
aq->vbq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
844
aq->vbq.min_queued_buffers = 1;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
845
aq->vbq.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
847
ret = vb2_queue_init(&aq->vbq);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
851
aq->vbq.dev = &adev->isp->pdev->dev;
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
852
spin_lock_init(&aq->lock);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
853
INIT_LIST_HEAD(&aq->active);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.c
854
INIT_LIST_HEAD(&aq->incoming);
drivers/media/pci/intel/ipu6/ipu6-isys-queue.h
69
int ipu6_isys_queue_init(struct ipu6_isys_queue *aq);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1263
ret = ipu6_isys_queue_init(&av->aq);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1279
av->vdev.queue = &av->aq.vbq;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1300
vb2_queue_release(&av->aq.vbq);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1317
if (av->aq.vbq.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1320
if (av->aq.vbq.type == V4L2_BUF_TYPE_META_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1328
if (av->aq.vbq.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1331
if (av->aq.vbq.type == V4L2_BUF_TYPE_META_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1339
if (av->aq.vbq.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1342
if (av->aq.vbq.type == V4L2_BUF_TYPE_META_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1350
if (av->aq.vbq.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1353
if (av->aq.vbq.type == V4L2_BUF_TYPE_META_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1361
if (av->aq.vbq.type == V4L2_BUF_TYPE_VIDEO_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1364
if (av->aq.vbq.type == V4L2_BUF_TYPE_META_CAPTURE)
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
293
if (vb2_is_busy(&av->aq.vbq))
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
338
if (vb2_is_busy(&av->aq.vbq))
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
353
av->aq.vbq.is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(p->type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
354
av->aq.vbq.is_output = V4L2_TYPE_IS_OUTPUT(p->type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
356
ret = vb2_queue_change_type(&av->aq.vbq, p->type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
369
av->aq.vbq.is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(p->format.type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
370
av->aq.vbq.is_output = V4L2_TYPE_IS_OUTPUT(p->format.type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
372
ret = vb2_queue_change_type(&av->aq.vbq, p->format.type);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
462
struct ipu6_isys_queue *aq = &av->aq;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
490
aq->fw_output = output_pins;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
491
stream->output_pins_queue[output_pins] = aq;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
528
struct ipu6_isys_queue *aq;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
542
list_for_each_entry(aq, &stream->queues, node) {
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
543
struct ipu6_isys_video *__av = ipu6_isys_queue_to_video(aq);
drivers/media/pci/intel/ipu6/ipu6-isys-video.h
100
container_of(__aq, struct ipu6_isys_video, aq)
drivers/media/pci/intel/ipu6/ipu6-isys-video.h
82
struct ipu6_isys_queue aq;
drivers/media/pci/intel/ipu6/ipu6-isys.c
251
av->aq.vbq.buf_struct_size =
drivers/net/ethernet/intel/i40e/i40e_adminq.c
100
for (i = 0; i < hw->aq.num_arq_entries; i++) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1008
mutex_lock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
101
bi = &hw->aq.arq.r.arq_bi[i];
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1014
*aq_status = hw->aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1015
mutex_unlock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
103
hw->aq.arq_buf_size,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1049
u16 ntc = hw->aq.arq.next_to_clean;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1062
mutex_lock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1064
if (hw->aq.arq.count == 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1080
desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1083
hw->aq.arq_last_status =
drivers/net/ethernet/intel/i40e/i40e_adminq.c
109
desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1091
hw->aq.arq_last_status);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1098
memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1103
hw->aq.arq_buf_size);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1109
bi = &hw->aq.arq.r.arq_bi[ntc];
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1113
if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
112
if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1123
if (ntc == hw->aq.num_arq_entries)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1125
hw->aq.arq.next_to_clean = ntc;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1126
hw->aq.arq.next_to_use = ntu;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1132
*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1134
mutex_unlock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1142
hw->aq.asq.next_to_use = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1143
hw->aq.asq.next_to_clean = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1147
hw->aq.arq.next_to_use = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1148
hw->aq.arq.next_to_clean = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
137
i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
138
i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
154
ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
155
(hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
158
hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
161
for (i = 0; i < hw->aq.num_asq_entries; i++) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
162
bi = &hw->aq.asq.r.asq_bi[i];
drivers/net/ethernet/intel/i40e/i40e_adminq.c
164
hw->aq.asq_buf_size,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
176
i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
177
i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
19
ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
191
for (i = 0; i < hw->aq.num_arq_entries; i++)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
192
i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
195
i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
198
i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
20
(hw->aq.num_asq_entries *
drivers/net/ethernet/intel/i40e/i40e_adminq.c
210
for (i = 0; i < hw->aq.num_asq_entries; i++)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
211
if (hw->aq.asq.r.asq_bi[i].pa)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
212
i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
215
i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
218
i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
221
i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
240
wr32(hw, I40E_PF_ATQLEN, (hw->aq.num_asq_entries |
drivers/net/ethernet/intel/i40e/i40e_adminq.c
242
wr32(hw, I40E_PF_ATQBAL, lower_32_bits(hw->aq.asq.desc_buf.pa));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
243
wr32(hw, I40E_PF_ATQBAH, upper_32_bits(hw->aq.asq.desc_buf.pa));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
247
if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa))
drivers/net/ethernet/intel/i40e/i40e_adminq.c
26
ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
269
wr32(hw, I40E_PF_ARQLEN, (hw->aq.num_arq_entries |
drivers/net/ethernet/intel/i40e/i40e_adminq.c
27
(hw->aq.num_asq_entries *
drivers/net/ethernet/intel/i40e/i40e_adminq.c
271
wr32(hw, I40E_PF_ARQBAL, lower_32_bits(hw->aq.arq.desc_buf.pa));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
272
wr32(hw, I40E_PF_ARQBAH, upper_32_bits(hw->aq.arq.desc_buf.pa));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
275
wr32(hw, I40E_PF_ARQT, hw->aq.num_arq_entries - 1);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
279
if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa))
drivers/net/ethernet/intel/i40e/i40e_adminq.c
30
i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
302
if (hw->aq.asq.count > 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
309
if ((hw->aq.num_asq_entries == 0) ||
drivers/net/ethernet/intel/i40e/i40e_adminq.c
310
(hw->aq.asq_buf_size == 0)) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
315
hw->aq.asq.next_to_use = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
316
hw->aq.asq.next_to_clean = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
334
hw->aq.asq.count = hw->aq.num_asq_entries;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
361
if (hw->aq.arq.count > 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
368
if ((hw->aq.num_arq_entries == 0) ||
drivers/net/ethernet/intel/i40e/i40e_adminq.c
369
(hw->aq.arq_buf_size == 0)) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
374
hw->aq.arq.next_to_use = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
375
hw->aq.arq.next_to_clean = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
393
hw->aq.arq.count = hw->aq.num_arq_entries;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
413
mutex_lock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
415
if (hw->aq.asq.count == 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
427
hw->aq.asq.count = 0; /* to indicate uninitialized queue */
drivers/net/ethernet/intel/i40e/i40e_adminq.c
433
mutex_unlock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
447
mutex_lock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
449
if (hw->aq.arq.count == 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
45
ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
46
(hw->aq.num_arq_entries *
drivers/net/ethernet/intel/i40e/i40e_adminq.c
461
hw->aq.arq.count = 0; /* to indicate uninitialized queue */
drivers/net/ethernet/intel/i40e/i40e_adminq.c
467
mutex_unlock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
582
if ((hw->aq.num_arq_entries == 0) ||
drivers/net/ethernet/intel/i40e/i40e_adminq.c
583
(hw->aq.num_asq_entries == 0) ||
drivers/net/ethernet/intel/i40e/i40e_adminq.c
584
(hw->aq.arq_buf_size == 0) ||
drivers/net/ethernet/intel/i40e/i40e_adminq.c
585
(hw->aq.asq_buf_size == 0)) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
591
hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
609
&hw->aq.fw_maj_ver,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
610
&hw->aq.fw_min_ver,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
611
&hw->aq.fw_build,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
612
&hw->aq.api_maj_ver,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
613
&hw->aq.api_min_ver,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
62
i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
691
struct i40e_adminq_ring *asq = &(hw->aq.asq);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
735
return rd32(hw, I40E_PF_ATQH) == hw->aq.asq.next_to_use;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
74
i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
767
if (hw->aq.asq.count == 0) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
774
hw->aq.asq_last_status = LIBIE_AQ_RC_OK;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
777
if (val >= hw->aq.num_asq_entries) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
784
details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
806
if (buff_size > hw->aq.asq_buf_size) {
drivers/net/ethernet/intel/i40e/i40e_adminq.c
839
desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
846
dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
864
(hw->aq.asq.next_to_use)++;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
865
if (hw->aq.asq.next_to_use == hw->aq.asq.count)
drivers/net/ethernet/intel/i40e/i40e_adminq.c
866
hw->aq.asq.next_to_use = 0;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
868
wr32(hw, I40E_PF_ATQT, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
889
} while (total_delay < hw->aq.asq_cmd_timeout);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
914
hw->aq.asq_last_status = (enum libie_aq_err)retval;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
93
ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
94
(hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
965
mutex_lock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
97
hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
970
mutex_unlock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_client.c
354
cdev->lan_info.fw_maj_ver = pf->hw.aq.fw_maj_ver;
drivers/net/ethernet/intel/i40e/i40e_client.c
355
cdev->lan_info.fw_min_ver = pf->hw.aq.fw_min_ver;
drivers/net/ethernet/intel/i40e/i40e_client.c
356
cdev->lan_info.fw_build = pf->hw.aq.fw_build;
drivers/net/ethernet/intel/i40e/i40e_client.c
547
err, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_client.c
685
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_client.c
712
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_common.c
1029
switch (hw->aq.asq_last_status) {
drivers/net/ethernet/intel/i40e/i40e_common.c
1043
} while ((hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN) &&
drivers/net/ethernet/intel/i40e/i40e_common.c
140
if (!hw->aq.asq.count)
drivers/net/ethernet/intel/i40e/i40e_common.c
2512
if (!status || hw->aq.asq_last_status == LIBIE_AQ_RC_EBUSY)
drivers/net/ethernet/intel/i40e/i40e_common.c
4540
if (hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN && retry) {
drivers/net/ethernet/intel/i40e/i40e_common.c
4602
if (hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN && retry) {
drivers/net/ethernet/intel/i40e/i40e_common.c
4856
struct i40e_profile_aq_section *aq)
drivers/net/ethernet/intel/i40e/i40e_common.c
4863
i40e_fill_default_direct_cmd_desc(&desc, aq->opcode);
drivers/net/ethernet/intel/i40e/i40e_common.c
4864
desc.flags |= cpu_to_le16(aq->flags);
drivers/net/ethernet/intel/i40e/i40e_common.c
4865
memcpy(desc.params.raw, aq->param, sizeof(desc.params.raw));
drivers/net/ethernet/intel/i40e/i40e_common.c
4867
msglen = aq->datalen;
drivers/net/ethernet/intel/i40e/i40e_common.c
4874
msg = &aq->data[0];
drivers/net/ethernet/intel/i40e/i40e_common.c
4882
aq->opcode, status);
drivers/net/ethernet/intel/i40e/i40e_common.c
4887
memcpy(aq->param, desc.params.raw, sizeof(desc.params.raw));
drivers/net/ethernet/intel/i40e/i40e_dcb.c
753
if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT)
drivers/net/ethernet/intel/i40e/i40e_dcb.c
802
if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT)
drivers/net/ethernet/intel/i40e/i40e_dcb.c
819
if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT)
drivers/net/ethernet/intel/i40e/i40e_dcb.c
928
} else if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT) {
drivers/net/ethernet/intel/i40e/i40e_dcb.c
932
} else if (hw->aq.asq_last_status == LIBIE_AQ_RC_EPERM) {
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
139
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
178
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
229
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_dcb_nl.c
294
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1003
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1116
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1256
desc->opcode, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1317
desc->opcode, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1352
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1362
__func__, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1379
__func__, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1387
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1411
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1439
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1458
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1469
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1512
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
1523
ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
448
ring = &(hw->aq.asq);
drivers/net/ethernet/intel/i40e/i40e_debugfs.c
461
ring = &(hw->aq.arq);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
55
struct i40e_adminq_info *aq = &hw->aq;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
57
snprintf(buf, len, "%u.%u", aq->fw_maj_ver, aq->fw_min_ver);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
62
struct i40e_adminq_info *aq = &hw->aq;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
64
snprintf(buf, len, "%05d", aq->fw_build);
drivers/net/ethernet/intel/i40e/i40e_devlink.c
69
struct i40e_adminq_info *aq = &hw->aq;
drivers/net/ethernet/intel/i40e/i40e_devlink.c
71
snprintf(buf, len, "%u.%u", aq->api_maj_ver, aq->api_min_ver);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1465
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1475
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1523
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1537
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1644
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1761
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1767
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1773
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1886
ret_val, hw->aq.asq_last_status, errno,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1904
ret_val, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1921
if (ret_val && hw->aq.asq_last_status == LIBIE_AQ_RC_EPERM) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1927
hw->aq.asq_last_status == LIBIE_AQ_RC_EACCES) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1935
offset, ret_val, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
1992
ret_val, hw->aq.asq_last_status, errno,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5371
if (ret && pf->hw.aq.asq_last_status != LIBIE_AQ_RC_ESRCH) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5375
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
5435
adq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_main.c
10296
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10307
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10330
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10341
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10474
if (pf->hw.aq.asq_last_status == LIBIE_AQ_RC_ENOMEM) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10477
} else if (pf->hw.aq.asq_last_status != LIBIE_AQ_RC_OK || err) {
drivers/net/ethernet/intel/i40e/i40e_main.c
10481
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10618
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10859
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
10970
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
11068
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
11099
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
12304
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
12317
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
12978
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
12997
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13887
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13916
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13935
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
13958
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14052
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14082
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14530
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14539
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14727
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14737
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14744
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14939
ret, libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
14984
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15007
if (ret && pf->hw.aq.asq_last_status != LIBIE_AQ_RC_ESRCH) {
drivers/net/ethernet/intel/i40e/i40e_main.c
15011
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15632
mutex_init(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
15633
mutex_init(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
15670
hw->aq.num_arq_entries = I40E_MIN_ARQ_LEN;
drivers/net/ethernet/intel/i40e/i40e_main.c
15671
hw->aq.num_asq_entries = I40E_MIN_ASQ_LEN;
drivers/net/ethernet/intel/i40e/i40e_main.c
15673
hw->aq.num_arq_entries = I40E_AQ_LEN;
drivers/net/ethernet/intel/i40e/i40e_main.c
15674
hw->aq.num_asq_entries = I40E_AQ_LEN;
drivers/net/ethernet/intel/i40e/i40e_main.c
15676
hw->aq.arq_buf_size = I40E_MAX_AQ_BUF_SIZE;
drivers/net/ethernet/intel/i40e/i40e_main.c
15677
hw->aq.asq_buf_size = I40E_MAX_AQ_BUF_SIZE;
drivers/net/ethernet/intel/i40e/i40e_main.c
15698
hw->aq.api_maj_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15699
hw->aq.api_min_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15714
hw->aq.fw_maj_ver, hw->aq.fw_min_ver, hw->aq.fw_build,
drivers/net/ethernet/intel/i40e/i40e_main.c
15715
hw->aq.api_maj_ver, hw->aq.api_min_ver, nvm_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15723
hw->aq.api_maj_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15724
hw->aq.api_min_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15730
hw->aq.api_maj_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15731
hw->aq.api_min_ver,
drivers/net/ethernet/intel/i40e/i40e_main.c
15911
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
15933
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16064
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16074
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16081
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
16251
mutex_destroy(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
16252
mutex_destroy(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/i40e/i40e_main.c
1842
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
1874
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
1886
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
2467
libie_aq_str(hw->aq.asq_last_status), vsi_name);
drivers/net/ethernet/intel/i40e/i40e_main.c
2508
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
2520
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
2530
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
2650
filter_list_len = hw->aq.asq_buf_size /
drivers/net/ethernet/intel/i40e/i40e_main.c
2711
filter_list_len = hw->aq.asq_buf_size /
drivers/net/ethernet/intel/i40e/i40e_main.c
2833
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
2838
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
2853
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
2859
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
3008
libie_aq_str(vsi->back->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
3042
libie_aq_str(vsi->back->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
3286
libie_aq_str(vsi->back->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5563
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5574
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5637
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
5764
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5820
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5887
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
5900
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6014
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6085
last_aq_status = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_main.c
6225
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6324
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6366
vsi->back->hw.aq.asq_last_status, ch->seid);
drivers/net/ethernet/intel/i40e/i40e_main.c
6567
if (ret && hw->aq.asq_last_status != LIBIE_AQ_RC_ESRCH)
drivers/net/ethernet/intel/i40e/i40e_main.c
6570
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6769
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6778
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6859
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6883
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
6922
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7038
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7077
ERR_PTR(ret), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7160
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7233
} else if (pf->hw.aq.asq_last_status == LIBIE_AQ_RC_EPERM) {
drivers/net/ethernet/intel/i40e/i40e_main.c
7239
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7495
ERR_PTR(err), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7506
ERR_PTR(err), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7550
ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
7723
*aq_err = hw->aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_main.c
7752
*aq_err = hw->aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_main.c
7962
ERR_PTR(ret), libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_main.c
8439
pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
8548
add ? "add" : "delete", ret, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
8933
return i40e_aq_rc_to_posix(err, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_main.c
9487
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_nvm.c
100
time_left, ret_code, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1033
hw->aq.asq_buf_size);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1056
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1057
*perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1197
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1208
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1222
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1239
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1256
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1272
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1276
*perrno = hw->aq.asq_last_status ?
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1278
hw->aq.asq_last_status) :
drivers/net/ethernet/intel/i40e/i40e_nvm.c
128
(total_delay < hw->aq.asq_cmd_timeout)) {
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1387
*perrno = hw->aq.asq_last_status ?
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1389
hw->aq.asq_last_status) :
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1403
*perrno = hw->aq.asq_last_status ?
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1405
hw->aq.asq_last_status) :
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1418
*perrno = hw->aq.asq_last_status ?
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1420
hw->aq.asq_last_status) :
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1445
if (status && hw->aq.asq_last_status == LIBIE_AQ_RC_EBUSY &&
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1447
u32 old_asq_status = hw->aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1461
hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1463
hw->aq.asq_last_status = old_asq_status;
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1547
mutex_lock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1585
mutex_unlock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
1605
if (hw->aq.arq_last_status) {
drivers/net/ethernet/intel/i40e/i40e_nvm.c
78
access, time_left, ret_code, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
896
__func__, status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
897
*perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
939
__func__, status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
940
*perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
979
__func__, status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_nvm.c
980
*perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_prototype.h
470
return (hw->aq.api_maj_ver > maj ||
drivers/net/ethernet/intel/i40e/i40e_prototype.h
471
(hw->aq.api_maj_ver == maj && hw->aq.api_min_ver >= min));
drivers/net/ethernet/intel/i40e/i40e_prototype.h
497
return (hw->aq.fw_maj_ver > maj ||
drivers/net/ethernet/intel/i40e/i40e_prototype.h
498
(hw->aq.fw_maj_ver == maj && hw->aq.fw_min_ver >= min));
drivers/net/ethernet/intel/i40e/i40e_prototype.h
524
return (hw->aq.fw_maj_ver == maj && hw->aq.fw_min_ver == min);
drivers/net/ethernet/intel/i40e/i40e_type.h
554
struct i40e_adminq_info aq;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1303
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1318
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1334
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
1349
int aq_err = pf->hw.aq.asq_last_status;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2007
vf->vf_id, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3770
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3872
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4018
libie_aq_str(pf->hw.aq.asq_last_status));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4593
vsi->back->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
4630
vsi->back->hw.aq.asq_last_status);
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
823
vf->vf_id, pf->hw.aq.asq_last_status);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
102
for (i = 0; i < hw->aq.num_arq_entries; i++) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
103
bi = &hw->aq.arq.r.arq_bi[i];
drivers/net/ethernet/intel/iavf/iavf_adminq.c
106
hw->aq.arq_buf_size,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
112
desc = IAVF_ADMINQ_DESC(hw->aq.arq, i);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
115
if (hw->aq.arq_buf_size > IAVF_AQ_LARGE_BUF)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
140
iavf_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
141
iavf_free_virt_mem(hw, &hw->aq.arq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
157
ret_code = iavf_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
158
(hw->aq.num_asq_entries *
drivers/net/ethernet/intel/iavf/iavf_adminq.c
162
hw->aq.asq.r.asq_bi = (struct iavf_dma_mem *)hw->aq.asq.dma_head.va;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
165
for (i = 0; i < hw->aq.num_asq_entries; i++) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
166
bi = &hw->aq.asq.r.asq_bi[i];
drivers/net/ethernet/intel/iavf/iavf_adminq.c
169
hw->aq.asq_buf_size,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
18
ret_code = iavf_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
181
iavf_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
182
iavf_free_virt_mem(hw, &hw->aq.asq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
196
for (i = 0; i < hw->aq.num_arq_entries; i++)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
197
iavf_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
20
(hw->aq.num_asq_entries *
drivers/net/ethernet/intel/iavf/iavf_adminq.c
200
iavf_free_dma_mem(hw, &hw->aq.arq.desc_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
203
iavf_free_virt_mem(hw, &hw->aq.arq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
215
for (i = 0; i < hw->aq.num_asq_entries; i++)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
216
if (hw->aq.asq.r.asq_bi[i].pa)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
217
iavf_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
220
iavf_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
223
iavf_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
226
iavf_free_virt_mem(hw, &hw->aq.asq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
245
wr32(hw, IAVF_VF_ATQLEN1, (hw->aq.num_asq_entries |
drivers/net/ethernet/intel/iavf/iavf_adminq.c
247
wr32(hw, IAVF_VF_ATQBAL1, lower_32_bits(hw->aq.asq.desc_buf.pa));
drivers/net/ethernet/intel/iavf/iavf_adminq.c
248
wr32(hw, IAVF_VF_ATQBAH1, upper_32_bits(hw->aq.asq.desc_buf.pa));
drivers/net/ethernet/intel/iavf/iavf_adminq.c
252
if (reg != lower_32_bits(hw->aq.asq.desc_buf.pa))
drivers/net/ethernet/intel/iavf/iavf_adminq.c
26
ret_code = iavf_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
27
(hw->aq.num_asq_entries *
drivers/net/ethernet/intel/iavf/iavf_adminq.c
274
wr32(hw, IAVF_VF_ARQLEN1, (hw->aq.num_arq_entries |
drivers/net/ethernet/intel/iavf/iavf_adminq.c
276
wr32(hw, IAVF_VF_ARQBAL1, lower_32_bits(hw->aq.arq.desc_buf.pa));
drivers/net/ethernet/intel/iavf/iavf_adminq.c
277
wr32(hw, IAVF_VF_ARQBAH1, upper_32_bits(hw->aq.arq.desc_buf.pa));
drivers/net/ethernet/intel/iavf/iavf_adminq.c
280
wr32(hw, IAVF_VF_ARQT1, hw->aq.num_arq_entries - 1);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
284
if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa))
drivers/net/ethernet/intel/iavf/iavf_adminq.c
30
iavf_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
308
if (hw->aq.asq.count > 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
315
if ((hw->aq.num_asq_entries == 0) ||
drivers/net/ethernet/intel/iavf/iavf_adminq.c
316
(hw->aq.asq_buf_size == 0)) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
321
hw->aq.asq.next_to_use = 0;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
322
hw->aq.asq.next_to_clean = 0;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
340
hw->aq.asq.count = hw->aq.num_asq_entries;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
344
for (i = 0; i < hw->aq.num_asq_entries; i++)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
345
iavf_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
346
iavf_free_virt_mem(hw, &hw->aq.asq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
373
if (hw->aq.arq.count > 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
380
if ((hw->aq.num_arq_entries == 0) ||
drivers/net/ethernet/intel/iavf/iavf_adminq.c
381
(hw->aq.arq_buf_size == 0)) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
386
hw->aq.arq.next_to_use = 0;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
387
hw->aq.arq.next_to_clean = 0;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
405
hw->aq.arq.count = hw->aq.num_arq_entries;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
409
for (i = 0; i < hw->aq.num_arq_entries; i++)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
410
iavf_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
411
iavf_free_virt_mem(hw, &hw->aq.arq.dma_head);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
429
mutex_lock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
431
if (hw->aq.asq.count == 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
443
hw->aq.asq.count = 0; /* to indicate uninitialized queue */
drivers/net/ethernet/intel/iavf/iavf_adminq.c
449
mutex_unlock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
45
ret_code = iavf_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
463
mutex_lock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
465
if (hw->aq.arq.count == 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
47
(hw->aq.num_arq_entries *
drivers/net/ethernet/intel/iavf/iavf_adminq.c
477
hw->aq.arq.count = 0; /* to indicate uninitialized queue */
drivers/net/ethernet/intel/iavf/iavf_adminq.c
483
mutex_unlock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
503
if ((hw->aq.num_arq_entries == 0) ||
drivers/net/ethernet/intel/iavf/iavf_adminq.c
504
(hw->aq.num_asq_entries == 0) ||
drivers/net/ethernet/intel/iavf/iavf_adminq.c
505
(hw->aq.arq_buf_size == 0) ||
drivers/net/ethernet/intel/iavf/iavf_adminq.c
506
(hw->aq.asq_buf_size == 0)) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
512
hw->aq.asq_cmd_timeout = IAVF_ASQ_CMD_TIMEOUT;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
558
struct iavf_adminq_ring *asq = &hw->aq.asq;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
603
return rd32(hw, IAVF_VF_ATQH1) == hw->aq.asq.next_to_use;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
63
iavf_free_dma_mem(hw, &hw->aq.asq.desc_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
631
mutex_lock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
633
if (hw->aq.asq.count == 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
640
hw->aq.asq_last_status = LIBIE_AQ_RC_OK;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
643
if (val >= hw->aq.num_asq_entries) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
650
details = IAVF_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
672
if (buff_size > hw->aq.asq_buf_size) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
705
desc_on_ring = IAVF_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
712
dma_buff = &hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use];
drivers/net/ethernet/intel/iavf/iavf_adminq.c
730
(hw->aq.asq.next_to_use)++;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
731
if (hw->aq.asq.next_to_use == hw->aq.asq.count)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
732
hw->aq.asq.next_to_use = 0;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
734
wr32(hw, IAVF_VF_ATQT1, hw->aq.asq.next_to_use);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
75
iavf_free_dma_mem(hw, &hw->aq.arq.desc_buf);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
750
} while (total_delay < hw->aq.asq_cmd_timeout);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
775
hw->aq.asq_last_status = (enum libie_aq_err)retval;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
801
mutex_unlock(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
834
u16 ntc = hw->aq.arq.next_to_clean;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
847
mutex_lock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
849
if (hw->aq.arq.count == 0) {
drivers/net/ethernet/intel/iavf/iavf_adminq.c
865
desc = IAVF_ADMINQ_DESC(hw->aq.arq, ntc);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
868
hw->aq.arq_last_status =
drivers/net/ethernet/intel/iavf/iavf_adminq.c
876
hw->aq.arq_last_status);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
883
memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
888
hw->aq.arq_buf_size);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
894
bi = &hw->aq.arq.r.arq_bi[ntc];
drivers/net/ethernet/intel/iavf/iavf_adminq.c
898
if (hw->aq.arq_buf_size > IAVF_AQ_LARGE_BUF)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
908
if (ntc == hw->aq.num_arq_entries)
drivers/net/ethernet/intel/iavf/iavf_adminq.c
910
hw->aq.arq.next_to_clean = ntc;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
911
hw->aq.arq.next_to_use = ntu;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
916
*pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
919
mutex_unlock(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
94
ret_code = iavf_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
95
(hw->aq.num_arq_entries *
drivers/net/ethernet/intel/iavf/iavf_adminq.c
99
hw->aq.arq.r.arq_bi = (struct iavf_dma_mem *)hw->aq.arq.dma_head.va;
drivers/net/ethernet/intel/iavf/iavf_common.c
224
if (!hw->aq.asq.count)
drivers/net/ethernet/intel/iavf/iavf_main.c
1673
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/iavf/iavf_main.c
1683
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/iavf/iavf_main.c
2348
hw->aq.num_arq_entries = IAVF_AQ_LEN;
drivers/net/ethernet/intel/iavf/iavf_main.c
2349
hw->aq.num_asq_entries = IAVF_AQ_LEN;
drivers/net/ethernet/intel/iavf/iavf_main.c
2350
hw->aq.arq_buf_size = IAVF_MAX_AQ_BUF_SIZE;
drivers/net/ethernet/intel/iavf/iavf_main.c
2351
hw->aq.asq_buf_size = IAVF_MAX_AQ_BUF_SIZE;
drivers/net/ethernet/intel/iavf/iavf_main.c
5392
mutex_init(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_main.c
5393
mutex_init(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_main.c
5599
if (hw->aq.asq.count)
drivers/net/ethernet/intel/iavf/iavf_main.c
5603
mutex_destroy(&hw->aq.arq_mutex);
drivers/net/ethernet/intel/iavf/iavf_main.c
5604
mutex_destroy(&hw->aq.asq_mutex);
drivers/net/ethernet/intel/iavf/iavf_type.h
174
struct iavf_adminq_info aq;
drivers/net/ethernet/intel/iavf/iavf_virtchnl.c
32
libie_aq_str(hw->aq.asq_last_status));
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2637
struct virtchnl2_add_queues aq = {};
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2645
aq.vport_id = cpu_to_le32(vport_id);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2646
aq.num_tx_q = cpu_to_le16(rsrc->num_txq);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2647
aq.num_tx_complq = cpu_to_le16(rsrc->num_complq);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2648
aq.num_rx_q = cpu_to_le16(rsrc->num_rxq);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2649
aq.num_rx_bufq = cpu_to_le16(rsrc->num_bufq);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2653
xn_params.send_buf.iov_base = &aq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2654
xn_params.send_buf.iov_len = sizeof(aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1201
void rvu_aq_free(struct rvu *rvu, struct admin_queue *aq)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1203
if (!aq)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1206
qmem_free(rvu->dev, aq->inst);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1207
qmem_free(rvu->dev, aq->res);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1208
devm_kfree(rvu->dev, aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1214
struct admin_queue *aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1217
*ad_queue = devm_kzalloc(rvu->dev, sizeof(*aq), GFP_KERNEL);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1220
aq = *ad_queue;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1223
err = qmem_alloc(rvu->dev, &aq->inst, qsize, inst_size);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1225
devm_kfree(rvu->dev, aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1230
err = qmem_alloc(rvu->dev, &aq->res, qsize, res_size);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1232
rvu_aq_free(rvu, aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1236
spin_lock_init(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
130
struct admin_queue *aq; /* NIX/NPA AQ */
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
934
void rvu_aq_free(struct rvu *rvu, struct admin_queue *aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1047
struct admin_queue *aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1056
aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1057
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1140
inst.res_addr = (u64)aq->res->iova;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1145
spin_lock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1148
memset(aq->res->base, 0, aq->res->entry_sz);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1150
ctx = aq->res->base + 128;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1152
mask = aq->res->base + 256;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1196
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1203
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1271
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4881
err = rvu_aq_alloc(rvu, &block->aq,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4889
NIX_AF_AQ_BASE, (u64)block->aq->inst->iova);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5138
rvu_aq_free(rvu, block->aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
969
struct admin_queue *aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
975
result = (struct nix_aq_res_s *)aq->res->base;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
981
memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)),
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
982
(void *)inst, aq->inst->entry_sz);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
101
inst.res_addr = (u64)aq->res->iova;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
106
spin_lock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
109
memset(aq->res->base, 0, aq->res->entry_sz);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
111
ctx = aq->res->base + 128;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
113
mask = aq->res->base + 256;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
153
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
160
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
18
struct admin_queue *aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
193
spin_unlock(&aq->lock);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
23
result = (struct npa_aq_res_s *)aq->res->base;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
29
memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)),
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
30
(void *)inst, aq->inst->entry_sz);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
512
err = rvu_aq_alloc(rvu, &block->aq,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
520
NPA_AF_AQ_BASE, (u64)block->aq->inst->iova);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
548
rvu_aq_free(rvu, block->aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
69
struct admin_queue *aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
83
aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
84
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
100
aq->sq.max_sqe_size = NIX_MAXSQESZ_W16; /* 128 byte */
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
101
aq->sq.cq_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
102
aq->sq.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
103
aq->sq.smq = otx2_get_smq_idx(pfvf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
104
aq->sq.smq_rr_weight = mtu_to_dwrr_weight(pfvf, pfvf->tx_max_pktlen);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
105
aq->sq.default_chan = pfvf->hw.tx_chan_base + chan_offset;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
106
aq->sq.sqe_stype = NIX_STYPE_STF; /* Cache SQB */
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
107
aq->sq.sqb_aura = sqb_aura;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
108
aq->sq.sq_int_ena = NIX_SQINT_BITS;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
109
aq->sq.qint_idx = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
113
aq->sq.cq_limit = ((SEND_CQ_SKID * 256) / (pfvf->qset.sqe_cnt));
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
116
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
117
aq->ctype = NIX_AQ_CTYPE_SQ;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
118
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
328
struct nix_cn10k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
330
aq = otx2_mbox_alloc_msg_nix_cn10k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
331
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
336
aq->rq.policer_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
338
aq->rq.policer_ena = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
339
aq->rq_mask.policer_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
341
aq->rq.band_prof_id = policer;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
342
aq->rq_mask.band_prof_id = GENMASK(9, 0);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
347
aq->rq.band_prof_id_h = policer >> 10;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
348
aq->rq_mask.band_prof_id_h = GENMASK(3, 0);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
351
aq->qidx = rq_idx;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
352
aq->ctype = NIX_AQ_CTYPE_RQ;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
353
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
396
struct nix_cn10k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
406
aq = otx2_mbox_alloc_msg_nix_cn10k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
407
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
411
aq->prof.icolor = 0x03;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
412
aq->prof_mask.icolor = 0x03;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
415
aq->prof.cir_exponent = rate_exp;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
416
aq->prof_mask.cir_exponent = 0x1F;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
418
aq->prof.cir_mantissa = rate_mantissa;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
419
aq->prof_mask.cir_mantissa = 0xFF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
421
aq->prof.cbs_exponent = burst_exp;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
422
aq->prof_mask.cbs_exponent = 0x1F;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
424
aq->prof.cbs_mantissa = burst_mantissa;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
425
aq->prof_mask.cbs_mantissa = 0xFF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
427
aq->prof.rdiv = rdiv;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
428
aq->prof_mask.rdiv = 0xF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
444
aq->prof.adjust_exponent = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
445
aq->prof_mask.adjust_exponent = 0x1F;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
447
aq->prof.adjust_mantissa = 384;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
448
aq->prof_mask.adjust_mantissa = 0x1FF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
450
aq->prof.lmode = 0x1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
451
aq->prof_mask.lmode = 0x1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
457
aq->prof.meter_algo = 2;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
458
aq->prof_mask.meter_algo = 0x3;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
460
aq->prof.rc_action = NIX_RX_BAND_PROF_ACTIONRESULT_DROP;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
461
aq->prof_mask.rc_action = 0x3;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
463
aq->prof.yc_action = NIX_RX_BAND_PROF_ACTIONRESULT_PASS;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
464
aq->prof_mask.yc_action = 0x3;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
466
aq->prof.gc_action = NIX_RX_BAND_PROF_ACTIONRESULT_PASS;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
467
aq->prof_mask.gc_action = 0x3;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
474
aq->prof.peir_exponent = 24;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
475
aq->prof_mask.peir_exponent = 0x1F;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
477
aq->prof.peir_mantissa = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
478
aq->prof_mask.peir_mantissa = 0xFF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
480
aq->prof.pebs_exponent = 24;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
481
aq->prof_mask.pebs_exponent = 0x1F;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
483
aq->prof.pebs_mantissa = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
484
aq->prof_mask.pebs_mantissa = 0xFF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
486
aq->prof.hl_en = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
487
aq->prof_mask.hl_en = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
490
aq->qidx = profile;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
491
aq->ctype = NIX_AQ_CTYPE_BANDPROF;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
492
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
91
struct nix_cn10k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
95
aq = otx2_mbox_alloc_msg_nix_cn10k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
96
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
99
aq->sq.cq = pfvf->hw.rx_queues + qidx;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
257
struct npa_cn20k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
274
aq = otx2_mbox_alloc_msg_npa_cn20k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
275
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
280
aq = otx2_mbox_alloc_msg_npa_cn20k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
281
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
285
aq->aura_id = aura_id;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
288
aq->aura.pool_addr = pool_id;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
289
aq->aura.pool_caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
290
aq->aura.shift = ilog2(numptrs) - 8;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
291
aq->aura.count = numptrs;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
292
aq->aura.limit = numptrs;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
293
aq->aura.avg_level = 255;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
294
aq->aura.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
295
aq->aura.fc_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
296
aq->aura.fc_addr = pool->fc_addr->iova;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
297
aq->aura.fc_hyst_bits = 0; /* Store count on all updates */
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
301
aq->aura.bp_ena = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
315
aq->aura.bp_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
318
aq->aura.bpid = pfvf->bpid[bpid_idx];
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
321
aq->aura.bp = RQ_BP_LVL_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
325
aq->ctype = NPA_AQ_CTYPE_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
326
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
336
struct npa_cn20k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
350
aq = otx2_mbox_alloc_msg_npa_cn20k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
351
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
358
aq = otx2_mbox_alloc_msg_npa_cn20k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
359
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
365
aq->aura_id = pool_id;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
366
aq->pool.stack_base = pool->stack->iova;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
367
aq->pool.stack_caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
368
aq->pool.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
369
aq->pool.buf_size = buf_size / 128;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
370
aq->pool.stack_max_pages = stack_pages;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
371
aq->pool.shift = ilog2(numptrs) - 8;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
372
aq->pool.ptr_start = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
373
aq->pool.ptr_end = ~0ULL;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
376
aq->ctype = NPA_AQ_CTYPE_POOL;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
377
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
402
struct nix_cn20k_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
406
aq = otx2_mbox_alloc_msg_nix_cn20k_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
407
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
410
aq->sq.cq = pfvf->hw.rx_queues + qidx;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
411
aq->sq.max_sqe_size = NIX_MAXSQESZ_W16; /* 128 byte */
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
412
aq->sq.cq_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
413
aq->sq.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
414
aq->sq.smq = otx2_get_smq_idx(pfvf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
415
aq->sq.smq_rr_weight = mtu_to_dwrr_weight(pfvf, pfvf->tx_max_pktlen);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
416
aq->sq.default_chan = pfvf->hw.tx_chan_base + chan_offset;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
417
aq->sq.sqe_stype = NIX_STYPE_STF; /* Cache SQB */
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
418
aq->sq.sqb_aura = sqb_aura;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
419
aq->sq.sq_int_ena = NIX_SQINT_BITS;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
420
aq->sq.qint_idx = 0;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
424
aq->sq.cq_limit = (SEND_CQ_SKID * 256) / (pfvf->qset.sqe_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
427
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
428
aq->ctype = NIX_AQ_CTYPE_SQ;
drivers/net/ethernet/marvell/octeontx2/nic/cn20k.c
429
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1055
struct nix_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1115
aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1116
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1119
aq->cq.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1120
aq->cq.qsize = Q_SIZE(cq->cqe_cnt, 4);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1121
aq->cq.caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1122
aq->cq.base = cq->cqe->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1123
aq->cq.cint_idx = cq->cint_idx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1124
aq->cq.cq_err_int_ena = NIX_CQERRINT_BITS;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1125
aq->cq.qint_idx = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1126
aq->cq.avg_level = 255;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1129
aq->cq.drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, cq->cqe_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1130
aq->cq.drop_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1134
aq->cq.bp_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1136
aq->cq.bpid = pfvf->bpid[pfvf->queue_to_pfc_map[qidx]];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1138
aq->cq.bpid = pfvf->bpid[0];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1142
aq->cq.bp = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1147
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1148
aq->ctype = NIX_AQ_CTYPE_CQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1149
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1379
struct npa_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1395
aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1396
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1401
aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1402
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1406
aq->aura_id = aura_id;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1408
aq->aura.pool_addr = pool_id;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1409
aq->aura.pool_caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1410
aq->aura.shift = ilog2(numptrs) - 8;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1411
aq->aura.count = numptrs;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1412
aq->aura.limit = numptrs;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1413
aq->aura.avg_level = 255;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1414
aq->aura.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1415
aq->aura.fc_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1416
aq->aura.fc_addr = pool->fc_addr->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1417
aq->aura.fc_hyst_bits = 0; /* Store count on all updates */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1421
aq->aura.bp_ena = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1435
aq->aura.bp_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1437
aq->aura.nix0_bpid = pfvf->bpid[pfvf->queue_to_pfc_map[aura_id]];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1439
aq->aura.nix0_bpid = pfvf->bpid[0];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1443
aq->aura.bp = RQ_BP_LVL_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1447
aq->ctype = NPA_AQ_CTYPE_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1448
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1465
struct npa_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1479
aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1480
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1487
aq = otx2_mbox_alloc_msg_npa_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1488
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1494
aq->aura_id = pool_id;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1495
aq->pool.stack_base = pool->stack->iova;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1496
aq->pool.stack_caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1497
aq->pool.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1498
aq->pool.buf_size = buf_size / 128;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1499
aq->pool.stack_max_pages = stack_pages;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1500
aq->pool.shift = ilog2(numptrs) - 8;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1501
aq->pool.ptr_start = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1502
aq->pool.ptr_end = ~0ULL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1505
aq->ctype = NPA_AQ_CTYPE_POOL;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
1506
aq->op = NPA_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
328
struct nix_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
339
aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
340
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
349
aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
350
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
356
aq->rss.rq = ind_tbl[idx];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
359
aq->qidx = index + idx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
360
aq->ctype = NIX_AQ_CTYPE_RSS;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
361
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
893
struct nix_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
896
aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
897
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
900
aq->rq.cq = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
901
aq->rq.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
902
aq->rq.pb_caching = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
903
aq->rq.lpb_aura = lpb_aura; /* Use large packet buffer aura */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
904
aq->rq.lpb_sizem1 = (DMA_BUFFER_LEN(pfvf->rbsize) / 8) - 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
905
aq->rq.xqe_imm_size = 0; /* Copying of packet to CQE not needed */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
906
aq->rq.flow_tagw = 32; /* Copy full 32bit flow_tag to CQE header */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
907
aq->rq.qint_idx = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
908
aq->rq.lpb_drop_ena = 1; /* Enable RED dropping for AURA */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
909
aq->rq.xqe_drop_ena = 1; /* Enable RED dropping for CQ/SSO */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
910
aq->rq.xqe_pass = RQ_PASS_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
911
aq->rq.xqe_drop = RQ_DROP_LVL_CQ(pfvf->hw.rq_skid, qset->rqe_cnt);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
912
aq->rq.lpb_aura_pass = RQ_PASS_LVL_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
913
aq->rq.lpb_aura_drop = RQ_DROP_LVL_AURA;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
916
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
917
aq->ctype = NIX_AQ_CTYPE_RQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
918
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
927
struct nix_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
932
aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
933
if (!aq)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
936
aq->sq.cq = pfvf->hw.rx_queues + qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
937
aq->sq.max_sqe_size = NIX_MAXSQESZ_W16; /* 128 byte */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
938
aq->sq.cq_ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
939
aq->sq.ena = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
940
aq->sq.smq = otx2_get_smq_idx(pfvf, qidx);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
941
aq->sq.smq_rr_quantum = mtu_to_dwrr_weight(pfvf, pfvf->tx_max_pktlen);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
942
aq->sq.default_chan = pfvf->hw.tx_chan_base + chan_offset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
943
aq->sq.sqe_stype = NIX_STYPE_STF; /* Cache SQB */
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
944
aq->sq.sqb_aura = sqb_aura;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
945
aq->sq.sq_int_ena = NIX_SQINT_BITS;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
946
aq->sq.qint_idx = 0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
950
aq->sq.cq_limit = ((SEND_CQ_SKID * 256) / (pfvf->qset.sqe_cnt));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
953
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
954
aq->ctype = NIX_AQ_CTYPE_SQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
955
aq->op = NIX_AQ_INSTOP_INIT;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
340
struct nix_aq_enq_req *aq;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
357
aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
358
if (!aq) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
363
aq->cq.bpid = pfvf->bpid[vlan_prio];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
364
aq->cq_mask.bpid = GENMASK(8, 0);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
367
aq->qidx = qidx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
368
aq->ctype = NIX_AQ_CTYPE_CQ;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_dcbnl.c
369
aq->op = NIX_AQ_INSTOP_WRITE;
drivers/s390/crypto/ap_bus.c
1075
struct ap_queue *aq;
drivers/s390/crypto/ap_bus.c
1078
hash_for_each(ap_queues, bkt, aq, hnode) {
drivers/s390/crypto/ap_bus.c
1079
if (aq->qid == qid) {
drivers/s390/crypto/ap_bus.c
1080
get_device(&aq->ap_dev.device);
drivers/s390/crypto/ap_bus.c
1082
return aq;
drivers/s390/crypto/ap_bus.c
1986
struct ap_queue *aq;
drivers/s390/crypto/ap_bus.c
2002
aq = dev ? to_ap_queue(dev) : NULL;
drivers/s390/crypto/ap_bus.c
2027
if (!aq) {
drivers/s390/crypto/ap_bus.c
2028
aq = ap_queue_create(qid, ac);
drivers/s390/crypto/ap_bus.c
2029
if (!aq) {
drivers/s390/crypto/ap_bus.c
2034
aq->config = !decfg;
drivers/s390/crypto/ap_bus.c
2035
aq->chkstop = chkstop;
drivers/s390/crypto/ap_bus.c
2036
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_bus.c
2037
dev = &aq->ap_dev.device;
drivers/s390/crypto/ap_bus.c
2058
ap_queue_init_state(aq);
drivers/s390/crypto/ap_bus.c
2065
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2067
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_bus.c
2069
if (chkstop && !aq->chkstop) {
drivers/s390/crypto/ap_bus.c
2071
aq->chkstop = true;
drivers/s390/crypto/ap_bus.c
2072
if (aq->dev_state > AP_DEV_STATE_UNINITIATED) {
drivers/s390/crypto/ap_bus.c
2073
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_bus.c
2074
aq->last_err_rc = AP_RESPONSE_CHECKSTOPPED;
drivers/s390/crypto/ap_bus.c
2076
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2080
ap_flush_queue(aq);
drivers/s390/crypto/ap_bus.c
2082
} else if (!chkstop && aq->chkstop) {
drivers/s390/crypto/ap_bus.c
2084
aq->chkstop = false;
drivers/s390/crypto/ap_bus.c
2085
if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
drivers/s390/crypto/ap_bus.c
2086
_ap_queue_init_state(aq);
drivers/s390/crypto/ap_bus.c
2087
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2093
if (decfg && aq->config) {
drivers/s390/crypto/ap_bus.c
2095
aq->config = false;
drivers/s390/crypto/ap_bus.c
2096
if (aq->dev_state > AP_DEV_STATE_UNINITIATED) {
drivers/s390/crypto/ap_bus.c
2097
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_bus.c
2098
aq->last_err_rc = AP_RESPONSE_DECONFIGURED;
drivers/s390/crypto/ap_bus.c
2100
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2103
ap_send_config_uevent(&aq->ap_dev, aq->config);
drivers/s390/crypto/ap_bus.c
2105
ap_flush_queue(aq);
drivers/s390/crypto/ap_bus.c
2107
} else if (!decfg && !aq->config) {
drivers/s390/crypto/ap_bus.c
2109
aq->config = true;
drivers/s390/crypto/ap_bus.c
2110
if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
drivers/s390/crypto/ap_bus.c
2111
_ap_queue_init_state(aq);
drivers/s390/crypto/ap_bus.c
2112
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2115
ap_send_config_uevent(&aq->ap_dev, aq->config);
drivers/s390/crypto/ap_bus.c
2119
if (!decfg && aq->dev_state == AP_DEV_STATE_ERROR) {
drivers/s390/crypto/ap_bus.c
2120
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
2122
ap_flush_queue(aq);
drivers/s390/crypto/ap_bus.c
2124
ap_queue_init_state(aq);
drivers/s390/crypto/ap_bus.c
2129
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
441
struct ap_queue *aq = timer_container_of(aq, t, timeout);
drivers/s390/crypto/ap_bus.c
443
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
444
ap_wait(ap_sm_event(aq, AP_SM_EVENT_TIMEOUT));
drivers/s390/crypto/ap_bus.c
445
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
481
struct ap_queue *aq;
drivers/s390/crypto/ap_bus.c
492
hash_for_each(ap_queues, bkt, aq, hnode) {
drivers/s390/crypto/ap_bus.c
493
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
494
wait = min(wait, ap_sm_event_loop(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_bus.c
495
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_bus.c
505
struct ap_queue *aq;
drivers/s390/crypto/ap_bus.c
508
hash_for_each(ap_queues, bkt, aq, hnode) {
drivers/s390/crypto/ap_bus.c
509
if (aq->queue_count == 0)
drivers/s390/crypto/ap_bus.c
692
struct ap_queue *aq = to_ap_queue(&ap_dev->device);
drivers/s390/crypto/ap_bus.c
695
if (aq->card->hwinfo.accel)
drivers/s390/crypto/ap_bus.c
697
else if (aq->card->hwinfo.cca)
drivers/s390/crypto/ap_bus.c
699
else if (aq->card->hwinfo.ep11)
drivers/s390/crypto/ap_bus.c
866
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_bus.c
867
struct ap_device *ap_dev = &aq->ap_dev;
drivers/s390/crypto/ap_bus.c
869
card = AP_QID_CARD(aq->qid);
drivers/s390/crypto/ap_bus.c
870
queue = AP_QID_QUEUE(aq->qid);
drivers/s390/crypto/ap_bus.c
921
struct ap_queue *aq;
drivers/s390/crypto/ap_bus.c
927
aq = ap_get_qdev(AP_MKQID(card, queue));
drivers/s390/crypto/ap_bus.c
928
if (aq) {
drivers/s390/crypto/ap_bus.c
929
const struct device_driver *drv = aq->ap_dev.device.driver;
drivers/s390/crypto/ap_bus.c
931
bool override = !!aq->ap_dev.driver_override;
drivers/s390/crypto/ap_bus.c
935
put_device(&aq->ap_dev.device);
drivers/s390/crypto/ap_card.c
105
hash_for_each(ap_queues, bkt, aq, hnode)
drivers/s390/crypto/ap_card.c
106
if (ac == aq->card)
drivers/s390/crypto/ap_card.c
107
reqq_cnt += aq->requestq_count;
drivers/s390/crypto/ap_card.c
118
struct ap_queue *aq;
drivers/s390/crypto/ap_card.c
124
hash_for_each(ap_queues, bkt, aq, hnode)
drivers/s390/crypto/ap_card.c
125
if (ac == aq->card)
drivers/s390/crypto/ap_card.c
126
penq_cnt += aq->pendingq_count;
drivers/s390/crypto/ap_card.c
80
struct ap_queue *aq;
drivers/s390/crypto/ap_card.c
84
hash_for_each(ap_queues, bkt, aq, hnode)
drivers/s390/crypto/ap_card.c
85
if (ac == aq->card)
drivers/s390/crypto/ap_card.c
86
aq->total_request_count = 0;
drivers/s390/crypto/ap_card.c
99
struct ap_queue *aq;
drivers/s390/crypto/ap_queue.c
1000
if (aq->sm_state < AP_SM_STATE_IDLE) {
drivers/s390/crypto/ap_queue.c
1006
status = ap_bapq(aq->qid);
drivers/s390/crypto/ap_queue.c
1010
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1014
aq->assoc_idx = ASSOC_IDX_INVALID;
drivers/s390/crypto/ap_queue.c
1017
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
1021
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1025
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
1030
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1037
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1041
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1050
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
1054
if (!ap_q_supports_assoc(aq))
drivers/s390/crypto/ap_queue.c
1057
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
1061
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1066
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1067
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
1068
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1072
if (aq->assoc_idx == ASSOC_IDX_INVALID) {
drivers/s390/crypto/ap_queue.c
1076
return sysfs_emit(buf, "associated %u\n", aq->assoc_idx);
drivers/s390/crypto/ap_queue.c
1078
if (aq->assoc_idx != ASSOC_IDX_INVALID)
drivers/s390/crypto/ap_queue.c
1090
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
1096
if (!ap_q_supports_assoc(aq))
drivers/s390/crypto/ap_queue.c
1107
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
1111
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1114
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1115
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
1119
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1125
if (aq->sm_state != AP_SM_STATE_IDLE) {
drivers/s390/crypto/ap_queue.c
1131
status = ap_aapq(aq->qid, value);
drivers/s390/crypto/ap_queue.c
1135
aq->sm_state = AP_SM_STATE_ASSOC_WAIT;
drivers/s390/crypto/ap_queue.c
1136
aq->assoc_idx = value;
drivers/s390/crypto/ap_queue.c
1137
ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_queue.c
1142
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
1150
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1173
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
1176
hash_del(&aq->hnode);
drivers/s390/crypto/ap_queue.c
1179
kfree(aq);
drivers/s390/crypto/ap_queue.c
1184
struct ap_queue *aq;
drivers/s390/crypto/ap_queue.c
1186
aq = kzalloc_obj(*aq);
drivers/s390/crypto/ap_queue.c
1187
if (!aq)
drivers/s390/crypto/ap_queue.c
1189
aq->card = ac;
drivers/s390/crypto/ap_queue.c
1190
aq->ap_dev.device.release = ap_queue_device_release;
drivers/s390/crypto/ap_queue.c
1191
aq->ap_dev.device.type = &ap_queue_type;
drivers/s390/crypto/ap_queue.c
1192
aq->ap_dev.device_type = ac->ap_dev.device_type;
drivers/s390/crypto/ap_queue.c
1194
if (ap_is_se_guest() && ap_q_supported_in_se(aq))
drivers/s390/crypto/ap_queue.c
1195
aq->ap_dev.device.groups = ap_queue_dev_sb_attr_groups;
drivers/s390/crypto/ap_queue.c
1196
aq->qid = qid;
drivers/s390/crypto/ap_queue.c
1197
spin_lock_init(&aq->lock);
drivers/s390/crypto/ap_queue.c
1198
INIT_LIST_HEAD(&aq->pendingq);
drivers/s390/crypto/ap_queue.c
1199
INIT_LIST_HEAD(&aq->requestq);
drivers/s390/crypto/ap_queue.c
1200
timer_setup(&aq->timeout, ap_request_timeout, 0);
drivers/s390/crypto/ap_queue.c
1202
return aq;
drivers/s390/crypto/ap_queue.c
1205
void ap_queue_init_reply(struct ap_queue *aq, struct ap_message *reply)
drivers/s390/crypto/ap_queue.c
1207
aq->reply = reply;
drivers/s390/crypto/ap_queue.c
1209
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
121
static enum ap_sm_wait ap_sm_nop(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1210
ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_queue.c
1211
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1220
int ap_queue_message(struct ap_queue *aq, struct ap_message *ap_msg)
drivers/s390/crypto/ap_queue.c
1227
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1230
if (aq->dev_state == AP_DEV_STATE_OPERATING) {
drivers/s390/crypto/ap_queue.c
1231
list_add_tail(&ap_msg->list, &aq->requestq);
drivers/s390/crypto/ap_queue.c
1232
aq->requestq_count++;
drivers/s390/crypto/ap_queue.c
1233
aq->total_request_count++;
drivers/s390/crypto/ap_queue.c
1234
atomic64_inc(&aq->card->total_request_count);
drivers/s390/crypto/ap_queue.c
1240
ap_wait(ap_sm_event_loop(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_queue.c
1242
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1256
bool ap_queue_usable(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1260
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1263
if (!aq->config || aq->chkstop) {
drivers/s390/crypto/ap_queue.c
1269
if (aq->dev_state != AP_DEV_STATE_OPERATING) {
drivers/s390/crypto/ap_queue.c
1276
if (!ap_q_supported_in_se(aq)) {
drivers/s390/crypto/ap_queue.c
1280
if (ap_q_needs_bind(aq) &&
drivers/s390/crypto/ap_queue.c
1281
!(aq->se_bstate == AP_BS_Q_USABLE ||
drivers/s390/crypto/ap_queue.c
1282
aq->se_bstate == AP_BS_Q_USABLE_NO_SECURE_KEY))
drivers/s390/crypto/ap_queue.c
1287
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1302
void ap_cancel_message(struct ap_queue *aq, struct ap_message *ap_msg)
drivers/s390/crypto/ap_queue.c
1306
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1308
list_for_each_entry(tmp, &aq->pendingq, list)
drivers/s390/crypto/ap_queue.c
1310
aq->pendingq_count--;
drivers/s390/crypto/ap_queue.c
1313
aq->requestq_count--;
drivers/s390/crypto/ap_queue.c
1317
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1327
static void __ap_flush_queue(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
133
static struct ap_queue_status ap_sm_recv(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1331
list_for_each_entry_safe(ap_msg, next, &aq->pendingq, list) {
drivers/s390/crypto/ap_queue.c
1333
aq->pendingq_count--;
drivers/s390/crypto/ap_queue.c
1335
ap_msg->receive(aq, ap_msg, NULL);
drivers/s390/crypto/ap_queue.c
1337
list_for_each_entry_safe(ap_msg, next, &aq->requestq, list) {
drivers/s390/crypto/ap_queue.c
1339
aq->requestq_count--;
drivers/s390/crypto/ap_queue.c
1341
ap_msg->receive(aq, ap_msg, NULL);
drivers/s390/crypto/ap_queue.c
1343
aq->queue_count = 0;
drivers/s390/crypto/ap_queue.c
1346
void ap_flush_queue(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1348
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1349
__ap_flush_queue(aq);
drivers/s390/crypto/ap_queue.c
1350
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1354
void ap_queue_prepare_remove(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1356
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1358
__ap_flush_queue(aq);
drivers/s390/crypto/ap_queue.c
1360
aq->dev_state = AP_DEV_STATE_SHUTDOWN;
drivers/s390/crypto/ap_queue.c
1361
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1362
timer_delete_sync(&aq->timeout);
drivers/s390/crypto/ap_queue.c
1365
void ap_queue_remove(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1373
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1374
ap_zapq(aq->qid, 0);
drivers/s390/crypto/ap_queue.c
1375
aq->dev_state = AP_DEV_STATE_UNINITIATED;
drivers/s390/crypto/ap_queue.c
1376
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1379
void _ap_queue_init_state(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1381
aq->dev_state = AP_DEV_STATE_OPERATING;
drivers/s390/crypto/ap_queue.c
1382
aq->sm_state = AP_SM_STATE_RESET_START;
drivers/s390/crypto/ap_queue.c
1383
aq->last_err_rc = 0;
drivers/s390/crypto/ap_queue.c
1384
aq->assoc_idx = ASSOC_IDX_INVALID;
drivers/s390/crypto/ap_queue.c
1385
ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_queue.c
1388
void ap_queue_init_state(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
1390
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
1391
_ap_queue_init_state(aq);
drivers/s390/crypto/ap_queue.c
1392
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
150
status = ap_dqap(aq->qid, &aq->reply->psmid,
drivers/s390/crypto/ap_queue.c
151
aq->reply->msg, aq->reply->bufsize,
drivers/s390/crypto/ap_queue.c
152
&aq->reply->len, &reslen, &resgr0);
drivers/s390/crypto/ap_queue.c
156
trace_s390_ap_dqap(AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid),
drivers/s390/crypto/ap_queue.c
157
status.value, aq->reply->psmid);
drivers/s390/crypto/ap_queue.c
162
aq->reply->msg, aq->reply->len, false);
drivers/s390/crypto/ap_queue.c
163
aq->queue_count = max_t(int, 0, aq->queue_count - 1);
drivers/s390/crypto/ap_queue.c
164
if (!status.queue_empty && !aq->queue_count)
drivers/s390/crypto/ap_queue.c
165
aq->queue_count++;
drivers/s390/crypto/ap_queue.c
166
if (aq->queue_count > 0)
drivers/s390/crypto/ap_queue.c
167
mod_timer(&aq->timeout,
drivers/s390/crypto/ap_queue.c
168
jiffies + aq->request_timeout);
drivers/s390/crypto/ap_queue.c
169
list_for_each_entry(ap_msg, &aq->pendingq, list) {
drivers/s390/crypto/ap_queue.c
170
if (ap_msg->psmid != aq->reply->psmid)
drivers/s390/crypto/ap_queue.c
173
aq->pendingq_count--;
drivers/s390/crypto/ap_queue.c
176
ap_msg->receive(aq, ap_msg, NULL);
drivers/s390/crypto/ap_queue.c
178
ap_msg->receive(aq, ap_msg, aq->reply);
drivers/s390/crypto/ap_queue.c
185
__func__, aq->reply->psmid,
drivers/s390/crypto/ap_queue.c
186
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
190
if (!status.queue_empty || aq->queue_count <= 0)
drivers/s390/crypto/ap_queue.c
193
aq->queue_count = 0;
drivers/s390/crypto/ap_queue.c
194
list_splice_init(&aq->pendingq, &aq->requestq);
drivers/s390/crypto/ap_queue.c
195
aq->requestq_count += aq->pendingq_count;
drivers/s390/crypto/ap_queue.c
197
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid),
drivers/s390/crypto/ap_queue.c
198
aq->pendingq_count, aq->requestq_count);
drivers/s390/crypto/ap_queue.c
199
aq->pendingq_count = 0;
drivers/s390/crypto/ap_queue.c
213
static enum ap_sm_wait ap_sm_read(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
217
if (!aq->reply)
drivers/s390/crypto/ap_queue.c
219
status = ap_sm_recv(aq);
drivers/s390/crypto/ap_queue.c
224
if (aq->queue_count > 0) {
drivers/s390/crypto/ap_queue.c
225
aq->sm_state = AP_SM_STATE_WORKING;
drivers/s390/crypto/ap_queue.c
228
aq->sm_state = AP_SM_STATE_IDLE;
drivers/s390/crypto/ap_queue.c
231
if (aq->queue_count > 0)
drivers/s390/crypto/ap_queue.c
234
aq->sm_state = AP_SM_STATE_IDLE;
drivers/s390/crypto/ap_queue.c
237
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
238
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
241
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
248
if (lsi_ptr && ap_queue_enable_irq(aq, lsi_ptr) == 0) {
drivers/s390/crypto/ap_queue.c
249
aq->sm_state = AP_SM_STATE_SETIRQ_WAIT;
drivers/s390/crypto/ap_queue.c
25
static void __ap_flush_queue(struct ap_queue *aq);
drivers/s390/crypto/ap_queue.c
262
static enum ap_sm_wait ap_sm_write(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
266
ap_qid_t qid = aq->qid;
drivers/s390/crypto/ap_queue.c
268
if (aq->requestq_count <= 0)
drivers/s390/crypto/ap_queue.c
272
ap_msg = list_entry(aq->requestq.next, struct ap_message, list);
drivers/s390/crypto/ap_queue.c
282
aq->queue_count = max_t(int, 1, aq->queue_count + 1);
drivers/s390/crypto/ap_queue.c
283
if (aq->queue_count == 1)
drivers/s390/crypto/ap_queue.c
284
mod_timer(&aq->timeout, jiffies + aq->request_timeout);
drivers/s390/crypto/ap_queue.c
285
list_move_tail(&ap_msg->list, &aq->pendingq);
drivers/s390/crypto/ap_queue.c
286
aq->requestq_count--;
drivers/s390/crypto/ap_queue.c
287
aq->pendingq_count++;
drivers/s390/crypto/ap_queue.c
288
if (aq->queue_count < aq->card->hwinfo.qd + 1) {
drivers/s390/crypto/ap_queue.c
289
aq->sm_state = AP_SM_STATE_WORKING;
drivers/s390/crypto/ap_queue.c
294
aq->sm_state = AP_SM_STATE_QUEUE_FULL;
drivers/s390/crypto/ap_queue.c
298
aq->sm_state = AP_SM_STATE_RESET_WAIT;
drivers/s390/crypto/ap_queue.c
306
aq->requestq_count--;
drivers/s390/crypto/ap_queue.c
308
ap_msg->receive(aq, ap_msg, NULL);
drivers/s390/crypto/ap_queue.c
31
static inline bool ap_q_supported_in_se(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
311
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
312
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
315
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
326
static enum ap_sm_wait ap_sm_read_write(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
328
return min(ap_sm_read(aq), ap_sm_write(aq));
drivers/s390/crypto/ap_queue.c
33
return aq->card->hwinfo.ep11 || aq->card->hwinfo.accel;
drivers/s390/crypto/ap_queue.c
337
static enum ap_sm_wait ap_sm_reset(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
341
status = ap_rapq(aq->qid, aq->rapq_fbit);
drivers/s390/crypto/ap_queue.c
347
aq->sm_state = AP_SM_STATE_RESET_WAIT;
drivers/s390/crypto/ap_queue.c
348
aq->rapq_fbit = 0;
drivers/s390/crypto/ap_queue.c
351
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
352
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
355
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
36
static inline bool ap_q_supports_bind(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
366
static enum ap_sm_wait ap_sm_reset_wait(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
373
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
377
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
379
if (lsi_ptr && ap_queue_enable_irq(aq, lsi_ptr) == 0)
drivers/s390/crypto/ap_queue.c
38
return aq->card->hwinfo.ep11 || aq->card->hwinfo.accel;
drivers/s390/crypto/ap_queue.c
380
aq->sm_state = AP_SM_STATE_SETIRQ_WAIT;
drivers/s390/crypto/ap_queue.c
382
aq->sm_state = (aq->queue_count > 0) ?
drivers/s390/crypto/ap_queue.c
392
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
393
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
396
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
407
static enum ap_sm_wait ap_sm_setirq_wait(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
41
static inline bool ap_q_supports_assoc(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
411
if (aq->queue_count > 0 && aq->reply)
drivers/s390/crypto/ap_queue.c
413
status = ap_sm_recv(aq);
drivers/s390/crypto/ap_queue.c
416
status = ap_tapq(aq->qid, NULL);
drivers/s390/crypto/ap_queue.c
420
aq->sm_state = (aq->queue_count > 0) ?
drivers/s390/crypto/ap_queue.c
426
if (aq->queue_count > 0)
drivers/s390/crypto/ap_queue.c
43
return aq->card->hwinfo.ep11;
drivers/s390/crypto/ap_queue.c
432
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
433
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
436
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
446
static enum ap_sm_wait ap_sm_assoc_wait(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
451
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
454
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
455
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
458
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
46
static inline bool ap_q_needs_bind(struct ap_queue *aq)
drivers/s390/crypto/ap_queue.c
462
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
463
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
466
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
471
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
477
aq->sm_state = AP_SM_STATE_IDLE;
drivers/s390/crypto/ap_queue.c
479
AP_QID_CARD(aq->qid),
drivers/s390/crypto/ap_queue.c
48
return ap_q_supports_bind(aq) && ap_sb_available();
drivers/s390/crypto/ap_queue.c
480
AP_QID_QUEUE(aq->qid), aq->assoc_idx);
drivers/s390/crypto/ap_queue.c
487
aq->assoc_idx = ASSOC_IDX_INVALID;
drivers/s390/crypto/ap_queue.c
488
aq->dev_state = AP_DEV_STATE_ERROR;
drivers/s390/crypto/ap_queue.c
489
aq->last_err_rc = status.response_code;
drivers/s390/crypto/ap_queue.c
492
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
531
enum ap_sm_wait ap_sm_event(struct ap_queue *aq, enum ap_sm_event event)
drivers/s390/crypto/ap_queue.c
533
if (aq->config && !aq->chkstop &&
drivers/s390/crypto/ap_queue.c
534
aq->dev_state > AP_DEV_STATE_UNINITIATED)
drivers/s390/crypto/ap_queue.c
535
return ap_jumptable[aq->sm_state][event](aq);
drivers/s390/crypto/ap_queue.c
540
enum ap_sm_wait ap_sm_event_loop(struct ap_queue *aq, enum ap_sm_event event)
drivers/s390/crypto/ap_queue.c
544
while ((wait = ap_sm_event(aq, event)) == AP_SM_WAIT_AGAIN)
drivers/s390/crypto/ap_queue.c
556
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
560
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
561
if (aq->dev_state > AP_DEV_STATE_UNINITIATED) {
drivers/s390/crypto/ap_queue.c
562
req_cnt = aq->total_request_count;
drivers/s390/crypto/ap_queue.c
565
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
577
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
579
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
580
aq->total_request_count = 0;
drivers/s390/crypto/ap_queue.c
581
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
591
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
594
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
595
if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
drivers/s390/crypto/ap_queue.c
596
reqq_cnt = aq->requestq_count;
drivers/s390/crypto/ap_queue.c
597
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
60
static int ap_queue_enable_irq(struct ap_queue *aq, void *ind)
drivers/s390/crypto/ap_queue.c
606
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
609
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
610
if (aq->dev_state > AP_DEV_STATE_UNINITIATED)
drivers/s390/crypto/ap_queue.c
611
penq_cnt = aq->pendingq_count;
drivers/s390/crypto/ap_queue.c
612
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
621
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
624
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
625
switch (aq->sm_state) {
drivers/s390/crypto/ap_queue.c
637
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
645
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
647
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
648
__ap_flush_queue(aq);
drivers/s390/crypto/ap_queue.c
649
aq->sm_state = AP_SM_STATE_RESET_START;
drivers/s390/crypto/ap_queue.c
650
ap_wait(ap_sm_event(aq, AP_SM_EVENT_POLL));
drivers/s390/crypto/ap_queue.c
651
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
654
__func__, AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
664
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
668
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
669
if (aq->sm_state == AP_SM_STATE_SETIRQ_WAIT) {
drivers/s390/crypto/ap_queue.c
67
status = ap_aqic(aq->qid, qirqctrl, virt_to_phys(ind));
drivers/s390/crypto/ap_queue.c
672
status = ap_tapq(aq->qid, NULL);
drivers/s390/crypto/ap_queue.c
678
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
688
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
691
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
692
rc = sysfs_emit(buf, "%d\n", aq->config ? 1 : 0);
drivers/s390/crypto/ap_queue.c
693
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
702
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
705
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
706
rc = sysfs_emit(buf, "%d\n", aq->chkstop ? 1 : 0);
drivers/s390/crypto/ap_queue.c
707
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
716
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
720
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
724
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
737
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
738
struct ap_device *ap_dev = &aq->ap_dev;
drivers/s390/crypto/ap_queue.c
755
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
756
struct ap_device *ap_dev = &aq->ap_dev;
drivers/s390/crypto/ap_queue.c
789
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
79
AP_QID_CARD(aq->qid),
drivers/s390/crypto/ap_queue.c
792
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
794
switch (aq->dev_state) {
drivers/s390/crypto/ap_queue.c
80
AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
811
if (aq->dev_state) {
drivers/s390/crypto/ap_queue.c
812
switch (aq->sm_state) {
drivers/s390/crypto/ap_queue.c
838
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
847
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
850
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
851
rc = aq->last_err_rc;
drivers/s390/crypto/ap_queue.c
852
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
922
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
926
if (!ap_q_supports_bind(aq))
drivers/s390/crypto/ap_queue.c
929
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
933
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
938
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
939
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
940
spin_unlock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
955
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/ap_queue.c
961
if (!ap_q_supports_bind(aq))
drivers/s390/crypto/ap_queue.c
971
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
972
__ap_flush_queue(aq);
drivers/s390/crypto/ap_queue.c
973
aq->rapq_fbit = 1;
drivers/s390/crypto/ap_queue.c
974
_ap_queue_init_state(aq);
drivers/s390/crypto/ap_queue.c
980
status = ap_test_queue(aq->qid, 1, &hwinfo);
drivers/s390/crypto/ap_queue.c
984
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/ap_queue.c
989
spin_lock_bh(&aq->lock);
drivers/s390/crypto/ap_queue.c
990
aq->se_bstate = hwinfo.bs;
drivers/s390/crypto/ap_queue.c
994
AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid));
drivers/s390/crypto/zcrypt_cex4.c
647
struct ap_queue *aq = to_ap_queue(&ap_dev->device);
drivers/s390/crypto/zcrypt_cex4.c
651
if (aq->card->hwinfo.accel) {
drivers/s390/crypto/zcrypt_cex4.c
652
zq = zcrypt_queue_alloc(aq->card->maxmsgsize);
drivers/s390/crypto/zcrypt_cex4.c
657
} else if (aq->card->hwinfo.cca) {
drivers/s390/crypto/zcrypt_cex4.c
658
zq = zcrypt_queue_alloc(aq->card->maxmsgsize);
drivers/s390/crypto/zcrypt_cex4.c
663
} else if (aq->card->hwinfo.ep11) {
drivers/s390/crypto/zcrypt_cex4.c
664
zq = zcrypt_queue_alloc(aq->card->maxmsgsize);
drivers/s390/crypto/zcrypt_cex4.c
673
zq->queue = aq;
drivers/s390/crypto/zcrypt_cex4.c
676
ap_queue_init_state(aq);
drivers/s390/crypto/zcrypt_cex4.c
677
ap_queue_init_reply(aq, &zq->reply);
drivers/s390/crypto/zcrypt_cex4.c
678
aq->request_timeout = CEX4_CLEANUP_TIME;
drivers/s390/crypto/zcrypt_cex4.c
686
if (aq->card->hwinfo.cca) {
drivers/s390/crypto/zcrypt_cex4.c
693
} else if (aq->card->hwinfo.ep11) {
drivers/s390/crypto/zcrypt_cex4.c
712
struct ap_queue *aq = to_ap_queue(&ap_dev->device);
drivers/s390/crypto/zcrypt_cex4.c
714
if (aq->card->hwinfo.cca)
drivers/s390/crypto/zcrypt_cex4.c
716
else if (aq->card->hwinfo.ep11)
drivers/s390/crypto/zcrypt_msgtype50.c
410
static void zcrypt_msgtype50_receive(struct ap_queue *aq,
drivers/s390/crypto/zcrypt_msgtype6.c
845
static void zcrypt_msgtype6_receive(struct ap_queue *aq,
drivers/s390/crypto/zcrypt_msgtype6.c
909
static void zcrypt_msgtype6_receive_ep11(struct ap_queue *aq,
drivers/s390/crypto/zcrypt_queue.c
44
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/zcrypt_queue.c
45
int online = aq->config && !aq->chkstop && zq->online ? 1 : 0;
drivers/s390/crypto/zcrypt_queue.c
55
struct ap_queue *aq = to_ap_queue(dev);
drivers/s390/crypto/zcrypt_queue.c
62
if (online && (!aq->config || !aq->card->config ||
drivers/s390/crypto/zcrypt_queue.c
63
aq->chkstop || aq->card->chkstop))
drivers/s390/crypto/zcrypt_queue.c
73
ap_send_online_uevent(&aq->ap_dev, online);
drivers/scsi/qla2xxx/qla_dbg.c
564
} aq, *aqp;
drivers/scsi/qla2xxx/qla_dbg.c
570
aqp = &aq;
drivers/spi/atmel-quadspi.c
1013
status = atmel_qspi_read(aq, QSPI_SR2);
drivers/spi/atmel-quadspi.c
1015
ret = atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
1018
atmel_qspi_write(QSPI_CR_QSPIDIS, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1031
aq, QSPI_PCALCFG);
drivers/spi/atmel-quadspi.c
1034
if (aq->caps->has_dllon)
drivers/spi/atmel-quadspi.c
1035
atmel_qspi_write(QSPI_CR_DLLON | QSPI_CR_STPCAL, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1038
atmel_qspi_write(QSPI_CR_STPCAL, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1044
if (aq->caps->has_dllon)
drivers/spi/atmel-quadspi.c
1045
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1050
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1056
aq->target_max_speed_hz / 1000),
drivers/spi/atmel-quadspi.c
1057
aq, QSPI_REFRESH);
drivers/spi/atmel-quadspi.c
1062
static int atmel_qspi_set_gclk(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
1068
if (aq->caps->has_dllon) {
drivers/spi/atmel-quadspi.c
1069
status = atmel_qspi_read(aq, QSPI_SR2);
drivers/spi/atmel-quadspi.c
1071
atmel_qspi_write(QSPI_CR_DLLOFF, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1072
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1079
if (aq->target_max_speed_hz > QSPI_DLLCFG_THRESHOLD_FREQ)
drivers/spi/atmel-quadspi.c
1080
atmel_qspi_write(QSPI_DLLCFG_RANGE, aq, QSPI_DLLCFG);
drivers/spi/atmel-quadspi.c
1082
atmel_qspi_write(0, aq, QSPI_DLLCFG);
drivers/spi/atmel-quadspi.c
1085
if (aq->caps->has_2xgclk)
drivers/spi/atmel-quadspi.c
1086
ret = clk_set_rate(aq->gclk, 2 * aq->target_max_speed_hz);
drivers/spi/atmel-quadspi.c
1088
ret = clk_set_rate(aq->gclk, aq->target_max_speed_hz);
drivers/spi/atmel-quadspi.c
1091
dev_err(&aq->pdev->dev, "Failed to set generic clock rate.\n");
drivers/spi/atmel-quadspi.c
1096
ret = clk_prepare_enable(aq->gclk);
drivers/spi/atmel-quadspi.c
1098
dev_err(&aq->pdev->dev, "Failed to enable generic clock.\n");
drivers/spi/atmel-quadspi.c
1103
static int atmel_qspi_sama7g5_init(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
1108
ret = atmel_qspi_set_gclk(aq);
drivers/spi/atmel-quadspi.c
1116
if (aq->caps->octal && aq->caps->has_padcalib) {
drivers/spi/atmel-quadspi.c
1117
ret = atmel_qspi_set_pad_calibration(aq);
drivers/spi/atmel-quadspi.c
1121
} else if (aq->caps->has_dllon) {
drivers/spi/atmel-quadspi.c
1122
atmel_qspi_write(QSPI_CR_DLLON, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1123
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1129
aq->mr |= QSPI_MR_DQSDLYEN;
drivers/spi/atmel-quadspi.c
1130
ret = atmel_qspi_set_serial_memory_mode(aq);
drivers/spi/atmel-quadspi.c
1135
atmel_qspi_write(QSPI_CR_QSPIEN, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1136
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1142
if (aq->caps->octal) {
drivers/spi/atmel-quadspi.c
1143
ret = readl_poll_timeout(aq->regs + QSPI_SR, val,
drivers/spi/atmel-quadspi.c
1148
atmel_qspi_write(QSPI_TOUT_TCNTM, aq, QSPI_TOUT);
drivers/spi/atmel-quadspi.c
1154
struct atmel_qspi *aq = spi_controller_get_devdata(spi->controller);
drivers/spi/atmel-quadspi.c
1157
aq->target_max_speed_hz = spi->max_speed_hz;
drivers/spi/atmel-quadspi.c
1159
return atmel_qspi_sama7g5_init(aq);
drivers/spi/atmel-quadspi.c
1165
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1176
if (aq->caps->has_gclk)
drivers/spi/atmel-quadspi.c
1179
src_rate = clk_get_rate(aq->pclk);
drivers/spi/atmel-quadspi.c
1192
aq->scr &= ~QSPI_SCR_SCBR_MASK;
drivers/spi/atmel-quadspi.c
1193
aq->scr |= QSPI_SCR_SCBR(scbr);
drivers/spi/atmel-quadspi.c
1194
atmel_qspi_write(aq->scr, aq, QSPI_SCR);
drivers/spi/atmel-quadspi.c
1204
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1212
clk_rate = clk_get_rate(aq->pclk);
drivers/spi/atmel-quadspi.c
1218
if (aq->mr & QSPI_MR_SMM) {
drivers/spi/atmel-quadspi.c
1220
dev_warn(&aq->pdev->dev,
drivers/spi/atmel-quadspi.c
1249
aq->scr &= ~QSPI_SCR_DLYBS_MASK;
drivers/spi/atmel-quadspi.c
1250
aq->scr |= QSPI_SCR_DLYBS(cs_setup);
drivers/spi/atmel-quadspi.c
1251
atmel_qspi_write(aq->scr, aq, QSPI_SCR);
drivers/spi/atmel-quadspi.c
1253
aq->mr &= ~(QSPI_MR_DLYBCT_MASK | QSPI_MR_DLYCS_MASK);
drivers/spi/atmel-quadspi.c
1254
aq->mr |= QSPI_MR_DLYBCT(cs_hold) | QSPI_MR_DLYCS(cs_inactive);
drivers/spi/atmel-quadspi.c
1255
atmel_qspi_write(aq->mr, aq, QSPI_MR);
drivers/spi/atmel-quadspi.c
1262
static int atmel_qspi_init(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
1266
if (aq->caps->has_gclk) {
drivers/spi/atmel-quadspi.c
1267
ret = atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
1270
atmel_qspi_write(QSPI_CR_SWRST, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1275
atmel_qspi_write(QSPI_CR_SWRST, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1278
ret = atmel_qspi_set_serial_memory_mode(aq);
drivers/spi/atmel-quadspi.c
1283
atmel_qspi_write(QSPI_CR_QSPIEN, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1289
struct atmel_qspi *aq = dev_id;
drivers/spi/atmel-quadspi.c
1292
status = atmel_qspi_read(aq, QSPI_SR);
drivers/spi/atmel-quadspi.c
1293
mask = atmel_qspi_read(aq, QSPI_IMR);
drivers/spi/atmel-quadspi.c
1299
aq->pending |= pending;
drivers/spi/atmel-quadspi.c
1300
if ((aq->pending & aq->irq_mask) == aq->irq_mask)
drivers/spi/atmel-quadspi.c
1301
complete(&aq->cmd_completion);
drivers/spi/atmel-quadspi.c
1308
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1311
aq->rx_chan = devm_dma_request_chan(&aq->pdev->dev, "rx");
drivers/spi/atmel-quadspi.c
1312
if (IS_ERR(aq->rx_chan)) {
drivers/spi/atmel-quadspi.c
1313
ret = dev_err_probe(&aq->pdev->dev, PTR_ERR(aq->rx_chan),
drivers/spi/atmel-quadspi.c
1315
aq->rx_chan = NULL;
drivers/spi/atmel-quadspi.c
1319
aq->tx_chan = devm_dma_request_chan(&aq->pdev->dev, "tx");
drivers/spi/atmel-quadspi.c
1320
if (IS_ERR(aq->tx_chan)) {
drivers/spi/atmel-quadspi.c
1321
ret = dev_err_probe(&aq->pdev->dev, PTR_ERR(aq->tx_chan),
drivers/spi/atmel-quadspi.c
1323
aq->rx_chan = NULL;
drivers/spi/atmel-quadspi.c
1324
aq->tx_chan = NULL;
drivers/spi/atmel-quadspi.c
1328
ctrl->dma_rx = aq->rx_chan;
drivers/spi/atmel-quadspi.c
1329
ctrl->dma_tx = aq->tx_chan;
drivers/spi/atmel-quadspi.c
1330
init_completion(&aq->dma_completion);
drivers/spi/atmel-quadspi.c
1332
dev_info(&aq->pdev->dev, "Using %s (tx) and %s (rx) for DMA transfers\n",
drivers/spi/atmel-quadspi.c
1333
dma_chan_name(aq->tx_chan), dma_chan_name(aq->rx_chan));
drivers/spi/atmel-quadspi.c
1351
struct atmel_qspi *aq;
drivers/spi/atmel-quadspi.c
1355
ctrl = devm_spi_alloc_host(&pdev->dev, sizeof(*aq));
drivers/spi/atmel-quadspi.c
1359
aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1361
aq->caps = of_device_get_match_data(&pdev->dev);
drivers/spi/atmel-quadspi.c
1362
if (!aq->caps) {
drivers/spi/atmel-quadspi.c
1367
init_completion(&aq->cmd_completion);
drivers/spi/atmel-quadspi.c
1368
aq->pdev = pdev;
drivers/spi/atmel-quadspi.c
1371
if (aq->caps->octal)
drivers/spi/atmel-quadspi.c
1374
if (aq->caps->has_gclk)
drivers/spi/atmel-quadspi.c
1375
aq->ops = &atmel_qspi_sama7g5_ops;
drivers/spi/atmel-quadspi.c
1377
aq->ops = &atmel_qspi_ops;
drivers/spi/atmel-quadspi.c
1379
ctrl->max_speed_hz = aq->caps->max_speed_hz;
drivers/spi/atmel-quadspi.c
1388
aq->regs = devm_platform_ioremap_resource_byname(pdev, "qspi_base");
drivers/spi/atmel-quadspi.c
1389
if (IS_ERR(aq->regs))
drivers/spi/atmel-quadspi.c
1390
return dev_err_probe(&pdev->dev, PTR_ERR(aq->regs),
drivers/spi/atmel-quadspi.c
1395
aq->mem = devm_ioremap_resource(&pdev->dev, res);
drivers/spi/atmel-quadspi.c
1396
if (IS_ERR(aq->mem))
drivers/spi/atmel-quadspi.c
1397
return dev_err_probe(&pdev->dev, PTR_ERR(aq->mem),
drivers/spi/atmel-quadspi.c
1400
aq->mmap_size = resource_size(res);
drivers/spi/atmel-quadspi.c
1401
aq->mmap_phys_base = (dma_addr_t)res->start;
drivers/spi/atmel-quadspi.c
1404
aq->pclk = devm_clk_get_enabled(&pdev->dev, "pclk");
drivers/spi/atmel-quadspi.c
1405
if (IS_ERR(aq->pclk))
drivers/spi/atmel-quadspi.c
1406
aq->pclk = devm_clk_get_enabled(&pdev->dev, NULL);
drivers/spi/atmel-quadspi.c
1408
if (IS_ERR(aq->pclk))
drivers/spi/atmel-quadspi.c
1409
return dev_err_probe(&pdev->dev, PTR_ERR(aq->pclk),
drivers/spi/atmel-quadspi.c
1412
if (aq->caps->has_qspick) {
drivers/spi/atmel-quadspi.c
1414
aq->qspick = devm_clk_get_enabled(&pdev->dev, "qspick");
drivers/spi/atmel-quadspi.c
1415
if (IS_ERR(aq->qspick)) {
drivers/spi/atmel-quadspi.c
1417
err = PTR_ERR(aq->qspick);
drivers/spi/atmel-quadspi.c
1421
} else if (aq->caps->has_gclk) {
drivers/spi/atmel-quadspi.c
1423
aq->gclk = devm_clk_get(&pdev->dev, "gclk");
drivers/spi/atmel-quadspi.c
1424
if (IS_ERR(aq->gclk)) {
drivers/spi/atmel-quadspi.c
1426
err = PTR_ERR(aq->gclk);
drivers/spi/atmel-quadspi.c
1431
if (aq->caps->has_dma) {
drivers/spi/atmel-quadspi.c
1443
0, dev_name(&pdev->dev), aq);
drivers/spi/atmel-quadspi.c
1452
err = atmel_qspi_init(aq);
drivers/spi/atmel-quadspi.c
1465
static int atmel_qspi_sama7g5_suspend(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
1470
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1477
atmel_qspi_write(QSPI_CR_QSPIDIS, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1478
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1484
clk_disable_unprepare(aq->gclk);
drivers/spi/atmel-quadspi.c
1486
if (aq->caps->has_dllon) {
drivers/spi/atmel-quadspi.c
1487
atmel_qspi_write(QSPI_CR_DLLOFF, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1488
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1495
if (aq->caps->has_padcalib)
drivers/spi/atmel-quadspi.c
1496
return readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
1505
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1512
if (aq->caps->has_gclk) {
drivers/spi/atmel-quadspi.c
1513
ret = atmel_qspi_sama7g5_suspend(aq);
drivers/spi/atmel-quadspi.c
1519
atmel_qspi_write(QSPI_CR_QSPIDIS, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1533
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1540
if (aq->caps->has_gclk) {
drivers/spi/atmel-quadspi.c
1541
ret = atmel_qspi_sama7g5_suspend(aq);
drivers/spi/atmel-quadspi.c
1542
clk_disable_unprepare(aq->pclk);
drivers/spi/atmel-quadspi.c
1546
atmel_qspi_write(QSPI_CR_QSPIDIS, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
1551
clk_unprepare(aq->qspick);
drivers/spi/atmel-quadspi.c
1552
clk_unprepare(aq->pclk);
drivers/spi/atmel-quadspi.c
1560
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1563
ret = clk_prepare(aq->pclk);
drivers/spi/atmel-quadspi.c
1567
ret = clk_prepare(aq->qspick);
drivers/spi/atmel-quadspi.c
1569
clk_unprepare(aq->pclk);
drivers/spi/atmel-quadspi.c
1573
if (aq->caps->has_gclk)
drivers/spi/atmel-quadspi.c
1574
return atmel_qspi_sama7g5_init(aq);
drivers/spi/atmel-quadspi.c
1580
atmel_qspi_init(aq);
drivers/spi/atmel-quadspi.c
1582
atmel_qspi_write(aq->scr, aq, QSPI_SCR);
drivers/spi/atmel-quadspi.c
1592
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1594
clk_disable(aq->qspick);
drivers/spi/atmel-quadspi.c
1595
clk_disable(aq->pclk);
drivers/spi/atmel-quadspi.c
1603
struct atmel_qspi *aq = spi_controller_get_devdata(ctrl);
drivers/spi/atmel-quadspi.c
1606
ret = clk_enable(aq->pclk);
drivers/spi/atmel-quadspi.c
1610
ret = clk_enable(aq->qspick);
drivers/spi/atmel-quadspi.c
1612
clk_disable(aq->pclk);
drivers/spi/atmel-quadspi.c
296
int (*set_cfg)(struct atmel_qspi *aq, const struct spi_mem_op *op,
drivers/spi/atmel-quadspi.c
395
static u32 atmel_qspi_read(struct atmel_qspi *aq, u32 offset)
drivers/spi/atmel-quadspi.c
397
u32 value = readl_relaxed(aq->regs + offset);
drivers/spi/atmel-quadspi.c
402
dev_vdbg(&aq->pdev->dev, "read 0x%08x from %s\n", value,
drivers/spi/atmel-quadspi.c
409
static void atmel_qspi_write(u32 value, struct atmel_qspi *aq, u32 offset)
drivers/spi/atmel-quadspi.c
414
dev_vdbg(&aq->pdev->dev, "write 0x%08x into %s\n", value,
drivers/spi/atmel-quadspi.c
418
writel_relaxed(value, aq->regs + offset);
drivers/spi/atmel-quadspi.c
421
static int atmel_qspi_reg_sync(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
426
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
432
static int atmel_qspi_update_config(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
436
ret = atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
439
atmel_qspi_write(QSPI_CR_UPDCFG, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
440
return atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
483
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
487
if (aq->caps->octal) {
drivers/spi/atmel-quadspi.c
509
static int atmel_qspi_set_serial_memory_mode(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
513
if (!(aq->mr & QSPI_MR_SMM)) {
drivers/spi/atmel-quadspi.c
514
aq->mr |= QSPI_MR_SMM;
drivers/spi/atmel-quadspi.c
515
atmel_qspi_write(aq->mr, aq, QSPI_MR);
drivers/spi/atmel-quadspi.c
517
if (aq->caps->has_gclk)
drivers/spi/atmel-quadspi.c
518
ret = atmel_qspi_update_config(aq);
drivers/spi/atmel-quadspi.c
524
static int atmel_qspi_set_cfg(struct atmel_qspi *aq,
drivers/spi/atmel-quadspi.c
603
mode = atmel_qspi_set_serial_memory_mode(aq);
drivers/spi/atmel-quadspi.c
608
(void)atmel_qspi_read(aq, QSPI_SR);
drivers/spi/atmel-quadspi.c
612
atmel_qspi_write(iar, aq, QSPI_IAR);
drivers/spi/atmel-quadspi.c
614
if (aq->caps->has_ricr) {
drivers/spi/atmel-quadspi.c
616
atmel_qspi_write(icr, aq, QSPI_RICR);
drivers/spi/atmel-quadspi.c
618
atmel_qspi_write(icr, aq, QSPI_WICR);
drivers/spi/atmel-quadspi.c
623
atmel_qspi_write(icr, aq, QSPI_ICR);
drivers/spi/atmel-quadspi.c
626
atmel_qspi_write(ifr, aq, QSPI_IFR);
drivers/spi/atmel-quadspi.c
631
static int atmel_qspi_wait_for_completion(struct atmel_qspi *aq, u32 irq_mask)
drivers/spi/atmel-quadspi.c
637
sr = atmel_qspi_read(aq, QSPI_SR);
drivers/spi/atmel-quadspi.c
642
reinit_completion(&aq->cmd_completion);
drivers/spi/atmel-quadspi.c
643
aq->pending = sr & irq_mask;
drivers/spi/atmel-quadspi.c
644
aq->irq_mask = irq_mask;
drivers/spi/atmel-quadspi.c
645
atmel_qspi_write(irq_mask, aq, QSPI_IER);
drivers/spi/atmel-quadspi.c
646
if (!wait_for_completion_timeout(&aq->cmd_completion,
drivers/spi/atmel-quadspi.c
649
atmel_qspi_write(irq_mask, aq, QSPI_IDR);
drivers/spi/atmel-quadspi.c
657
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
661
return atmel_qspi_wait_for_completion(aq,
drivers/spi/atmel-quadspi.c
665
(void)atmel_qspi_read(aq, QSPI_IFR);
drivers/spi/atmel-quadspi.c
669
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
675
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
683
atmel_qspi_write(QSPI_CR_LASTXFER, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
685
return atmel_qspi_wait_for_completion(aq, QSPI_SR_CMD_COMPLETED);
drivers/spi/atmel-quadspi.c
688
static int atmel_qspi_sama7g5_set_cfg(struct atmel_qspi *aq,
drivers/spi/atmel-quadspi.c
741
ret = atmel_qspi_set_serial_memory_mode(aq);
drivers/spi/atmel-quadspi.c
746
(void)atmel_qspi_read(aq, QSPI_SR);
drivers/spi/atmel-quadspi.c
750
atmel_qspi_write(iar, aq, QSPI_IAR);
drivers/spi/atmel-quadspi.c
753
atmel_qspi_write(icr, aq, QSPI_RICR);
drivers/spi/atmel-quadspi.c
755
atmel_qspi_write(icr, aq, QSPI_WICR);
drivers/spi/atmel-quadspi.c
759
aq, QSPI_WRACNT);
drivers/spi/atmel-quadspi.c
762
atmel_qspi_write(ifr, aq, QSPI_IFR);
drivers/spi/atmel-quadspi.c
764
return atmel_qspi_update_config(aq);
drivers/spi/atmel-quadspi.c
769
struct atmel_qspi *aq = param;
drivers/spi/atmel-quadspi.c
771
complete(&aq->dma_completion);
drivers/spi/atmel-quadspi.c
774
static int atmel_qspi_dma_xfer(struct atmel_qspi *aq, struct dma_chan *chan,
drivers/spi/atmel-quadspi.c
785
dev_err(&aq->pdev->dev, "device_prep_dma_memcpy error\n");
drivers/spi/atmel-quadspi.c
789
reinit_completion(&aq->dma_completion);
drivers/spi/atmel-quadspi.c
791
tx->callback_param = aq;
drivers/spi/atmel-quadspi.c
795
dev_err(&aq->pdev->dev, "dma_submit_error %d\n", cookie);
drivers/spi/atmel-quadspi.c
800
ret = wait_for_completion_timeout(&aq->dma_completion,
drivers/spi/atmel-quadspi.c
804
dev_err(&aq->pdev->dev, "DMA wait_for_completion_timeout\n");
drivers/spi/atmel-quadspi.c
815
struct atmel_qspi *aq =
drivers/spi/atmel-quadspi.c
822
dma_src = aq->mmap_phys_base + loff;
drivers/spi/atmel-quadspi.c
826
ret = atmel_qspi_dma_xfer(aq, aq->rx_chan, sg_dma_address(sg),
drivers/spi/atmel-quadspi.c
840
struct atmel_qspi *aq =
drivers/spi/atmel-quadspi.c
847
dma_dst = aq->mmap_phys_base + loff;
drivers/spi/atmel-quadspi.c
851
ret = atmel_qspi_dma_xfer(aq, aq->tx_chan, dma_dst,
drivers/spi/atmel-quadspi.c
885
struct atmel_qspi *aq =
drivers/spi/atmel-quadspi.c
892
ret = atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
895
atmel_qspi_write(QSPI_CR_STTFR, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
897
return atmel_qspi_wait_for_completion(aq, QSPI_SR_CSRA);
drivers/spi/atmel-quadspi.c
902
if (aq->rx_chan && op->addr.nbytes &&
drivers/spi/atmel-quadspi.c
908
memcpy_fromio(op->data.buf.in, aq->mem + offset,
drivers/spi/atmel-quadspi.c
913
ret = readl_poll_timeout(aq->regs + QSPI_SR2, val,
drivers/spi/atmel-quadspi.c
920
if (aq->tx_chan && op->addr.nbytes &&
drivers/spi/atmel-quadspi.c
926
memcpy_toio(aq->mem + offset, op->data.buf.out,
drivers/spi/atmel-quadspi.c
930
ret = atmel_qspi_wait_for_completion(aq, QSPI_SR_LWRA);
drivers/spi/atmel-quadspi.c
936
ret = atmel_qspi_reg_sync(aq);
drivers/spi/atmel-quadspi.c
939
atmel_qspi_write(QSPI_CR_LASTXFER, aq, QSPI_CR);
drivers/spi/atmel-quadspi.c
941
return atmel_qspi_wait_for_completion(aq, QSPI_SR_CSRA);
drivers/spi/atmel-quadspi.c
946
struct atmel_qspi *aq = spi_controller_get_devdata(mem->spi->controller);
drivers/spi/atmel-quadspi.c
955
if (op->addr.val + op->data.nbytes > aq->mmap_size)
drivers/spi/atmel-quadspi.c
961
err = pm_runtime_resume_and_get(&aq->pdev->dev);
drivers/spi/atmel-quadspi.c
965
err = aq->ops->set_cfg(aq, op, &offset);
drivers/spi/atmel-quadspi.c
969
err = aq->ops->transfer(mem, op, offset);
drivers/spi/atmel-quadspi.c
972
pm_runtime_put_autosuspend(&aq->pdev->dev);
drivers/spi/atmel-quadspi.c
987
static int atmel_qspi_set_pad_calibration(struct atmel_qspi *aq)
drivers/spi/atmel-quadspi.c
994
pclk_rate = clk_get_rate(aq->pclk);
drivers/staging/media/ipu7/ipu7-isys-queue.c
129
struct ipu7_isys_queue *aq =
drivers/staging/media/ipu7/ipu7-isys-queue.c
132
av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
133
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
136
list_add(&ib->head, &aq->active);
drivers/staging/media/ipu7/ipu7-isys-queue.c
138
list_add_tail(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
139
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
164
struct ipu7_isys_queue *aq;
drivers/staging/media/ipu7/ipu7-isys-queue.c
169
list_for_each_entry(aq, &stream->queues, node) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
170
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
174
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
175
list_for_each_entry_safe(ib, ib_safe, &aq->active, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
185
list_add(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
194
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
208
struct ipu7_isys_queue *aq;
drivers/staging/media/ipu7/ipu7-isys-queue.c
214
list_for_each_entry(aq, &stream->queues, node) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
217
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
218
if (list_empty(&aq->incoming)) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
219
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
225
ib = list_last_entry(&aq->incoming,
drivers/staging/media/ipu7/ipu7-isys-queue.c
229
ipu7_isys_queue_to_video(aq)->vdev.name,
drivers/staging/media/ipu7/ipu7-isys-queue.c
233
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
246
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/staging/media/ipu7/ipu7-isys-queue.c
251
set->output_pins[aq->fw_output].addr = ivb->dma_addr;
drivers/staging/media/ipu7/ipu7-isys-queue.c
252
set->output_pins[aq->fw_output].user_token = (uintptr_t)set;
drivers/staging/media/ipu7/ipu7-isys-queue.c
350
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/staging/media/ipu7/ipu7-isys-queue.c
351
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
372
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
373
list_add(&ib->head, &aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
374
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
437
static int ipu7_isys_link_fmt_validate(struct ipu7_isys_queue *aq)
drivers/staging/media/ipu7/ipu7-isys-queue.c
439
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
479
static void return_buffers(struct ipu7_isys_queue *aq,
drivers/staging/media/ipu7/ipu7-isys-queue.c
486
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
492
while (!list_empty(&aq->active)) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
493
ib = list_last_entry(&aq->active, struct ipu7_isys_buffer,
drivers/staging/media/ipu7/ipu7-isys-queue.c
498
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
502
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
505
while (!list_empty(&aq->incoming)) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
506
ib = list_last_entry(&aq->incoming, struct ipu7_isys_buffer,
drivers/staging/media/ipu7/ipu7-isys-queue.c
510
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
514
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
517
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
529
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/staging/media/ipu7/ipu7-isys-queue.c
530
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
549
ret = ipu7_isys_link_fmt_validate(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
572
list_add(&aq->node, &stream->queues);
drivers/staging/media/ipu7/ipu7-isys-queue.c
603
list_del(&aq->node);
drivers/staging/media/ipu7/ipu7-isys-queue.c
611
return_buffers(aq, VB2_BUF_STATE_QUEUED);
drivers/staging/media/ipu7/ipu7-isys-queue.c
618
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/staging/media/ipu7/ipu7-isys-queue.c
619
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
629
list_del(&aq->node);
drivers/staging/media/ipu7/ipu7-isys-queue.c
636
return_buffers(aq, VB2_BUF_STATE_ERROR);
drivers/staging/media/ipu7/ipu7-isys-queue.c
66
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(q);
drivers/staging/media/ipu7/ipu7-isys-queue.c
67
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
691
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/staging/media/ipu7/ipu7-isys-queue.c
692
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
729
struct ipu7_isys_queue *aq = stream->output_pins[info->pin_id].aq;
drivers/staging/media/ipu7/ipu7-isys-queue.c
740
ipu7_isys_queue_to_video(aq)->vdev.name, info->pin.addr,
drivers/staging/media/ipu7/ipu7-isys-queue.c
743
spin_lock_irqsave(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
744
if (list_empty(&aq->active)) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
745
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
750
list_for_each_entry_reverse(ib, &aq->active, head) {
drivers/staging/media/ipu7/ipu7-isys-queue.c
775
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
786
spin_unlock_irqrestore(&aq->lock, flags);
drivers/staging/media/ipu7/ipu7-isys-queue.c
799
int ipu7_isys_queue_init(struct ipu7_isys_queue *aq)
drivers/staging/media/ipu7/ipu7-isys-queue.c
801
struct ipu7_isys *isys = ipu7_isys_queue_to_video(aq)->isys;
drivers/staging/media/ipu7/ipu7-isys-queue.c
802
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
806
if (!aq->vbq.io_modes)
drivers/staging/media/ipu7/ipu7-isys-queue.c
807
aq->vbq.io_modes = VB2_MMAP | VB2_DMABUF;
drivers/staging/media/ipu7/ipu7-isys-queue.c
809
aq->vbq.drv_priv = isys;
drivers/staging/media/ipu7/ipu7-isys-queue.c
810
aq->vbq.ops = &ipu7_isys_queue_ops;
drivers/staging/media/ipu7/ipu7-isys-queue.c
811
aq->vbq.lock = &av->mutex;
drivers/staging/media/ipu7/ipu7-isys-queue.c
812
aq->vbq.mem_ops = &vb2_dma_sg_memops;
drivers/staging/media/ipu7/ipu7-isys-queue.c
813
aq->vbq.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/staging/media/ipu7/ipu7-isys-queue.c
814
aq->vbq.min_queued_buffers = 1;
drivers/staging/media/ipu7/ipu7-isys-queue.c
815
aq->vbq.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/staging/media/ipu7/ipu7-isys-queue.c
817
ret = vb2_queue_init(&aq->vbq);
drivers/staging/media/ipu7/ipu7-isys-queue.c
821
aq->dev = &adev->auxdev.dev;
drivers/staging/media/ipu7/ipu7-isys-queue.c
822
aq->vbq.dev = &adev->isp->pdev->dev;
drivers/staging/media/ipu7/ipu7-isys-queue.c
823
spin_lock_init(&aq->lock);
drivers/staging/media/ipu7/ipu7-isys-queue.c
824
INIT_LIST_HEAD(&aq->active);
drivers/staging/media/ipu7/ipu7-isys-queue.c
825
INIT_LIST_HEAD(&aq->incoming);
drivers/staging/media/ipu7/ipu7-isys-queue.c
87
struct ipu7_isys_queue *aq = vb2_queue_to_isys_queue(vb->vb2_queue);
drivers/staging/media/ipu7/ipu7-isys-queue.c
88
struct ipu7_isys_video *av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-queue.h
71
int ipu7_isys_queue_init(struct ipu7_isys_queue *aq);
drivers/staging/media/ipu7/ipu7-isys-video.c
1031
ret = ipu7_isys_queue_init(&av->aq);
drivers/staging/media/ipu7/ipu7-isys-video.c
1046
av->vdev.queue = &av->aq.vbq;
drivers/staging/media/ipu7/ipu7-isys-video.c
1065
vb2_queue_release(&av->aq.vbq);
drivers/staging/media/ipu7/ipu7-isys-video.c
229
if (vb2_is_busy(&av->aq.vbq))
drivers/staging/media/ipu7/ipu7-isys-video.c
254
av->aq.vbq.is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(p->type);
drivers/staging/media/ipu7/ipu7-isys-video.c
255
av->aq.vbq.is_output = V4L2_TYPE_IS_OUTPUT(p->type);
drivers/staging/media/ipu7/ipu7-isys-video.c
257
ret = vb2_queue_change_type(&av->aq.vbq, p->type);
drivers/staging/media/ipu7/ipu7-isys-video.c
270
av->aq.vbq.is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(p->format.type);
drivers/staging/media/ipu7/ipu7-isys-video.c
271
av->aq.vbq.is_output = V4L2_TYPE_IS_OUTPUT(p->format.type);
drivers/staging/media/ipu7/ipu7-isys-video.c
273
ret = vb2_queue_change_type(&av->aq.vbq, p->format.type);
drivers/staging/media/ipu7/ipu7-isys-video.c
362
struct ipu7_isys_queue *aq = &av->aq;
drivers/staging/media/ipu7/ipu7-isys-video.c
394
aq->fw_output = output_pins;
drivers/staging/media/ipu7/ipu7-isys-video.c
396
stream->output_pins[output_pins].aq = aq;
drivers/staging/media/ipu7/ipu7-isys-video.c
444
struct ipu7_isys_queue *aq;
drivers/staging/media/ipu7/ipu7-isys-video.c
461
list_for_each_entry(aq, &stream->queues, node) {
drivers/staging/media/ipu7/ipu7-isys-video.c
462
struct ipu7_isys_video *__av = ipu7_isys_queue_to_video(aq);
drivers/staging/media/ipu7/ipu7-isys-video.h
45
struct ipu7_isys_queue *aq;
drivers/staging/media/ipu7/ipu7-isys-video.h
80
struct ipu7_isys_queue aq;
drivers/staging/media/ipu7/ipu7-isys-video.h
95
container_of(__aq, struct ipu7_isys_video, aq)
drivers/staging/media/ipu7/ipu7-isys.c
287
av->aq.vbq.buf_struct_size =
lib/kasprintf.c
19
va_list aq;
lib/kasprintf.c
21
va_copy(aq, ap);
lib/kasprintf.c
22
first = vsnprintf(NULL, 0, fmt, aq);
lib/kasprintf.c
23
va_end(aq);
lib/tests/printf_kunit.c
45
va_list aq;
lib/tests/printf_kunit.c
51
va_copy(aq, ap);
lib/tests/printf_kunit.c
52
ret = vsnprintf(test_buffer, bufsize, fmt, aq);
lib/tests/printf_kunit.c
53
va_end(aq);
net/ceph/osdmap.c
2183
va_list aq;
net/ceph/osdmap.c
2186
va_copy(aq, ap);
net/ceph/osdmap.c
2187
len = oid_printf_vargs(oid, fmt, aq);
net/ceph/osdmap.c
2188
va_end(aq);