ern
fd = &msg->ern.fd;
req_fq->cb.ern = caam_fq_ern_cb;
u32 q_no, srn, ern;
ern = srn + oct->sriov_info.num_pf_rings;
for (q_no = srn; q_no < ern; q_no++) {
for (q_no = srn; q_no < ern; q_no++) {
u32 q_no, ern, srn;
ern = srn + oct->sriov_info.num_pf_rings;
for (q_no = 0; q_no < ern; q_no++) {
for (q_no = srn; q_no < ern; q_no++) {
u32 q_no, ern, srn;
ern = srn + oct->sriov_info.num_pf_rings;
for (q_no = srn; q_no < ern; q_no++) {
u32 srn, ern, q_no;
ern = srn + oct->num_iqs;
for (q_no = srn; q_no < ern; q_no++) {
for (q_no = srn; q_no < ern; q_no++) {
u32 srn, ern;
ern = srn + oct->num_iqs;
for (q_no = srn; q_no < ern; q_no++) {
for (q_no = srn; q_no < ern; q_no++) {
switch (msg->ern.rc & QM_MR_RC_MASK) {
const struct qm_fd *fd = &msg->ern.fd;
.egress_ern = { .cb = { .ern = egress_ern } }
e->verb, e->ern.rc, qm_fd_addr_get64(&e->ern.fd));
fq = tag_to_fq(be32_to_cpu(msg->ern.tag));
fq->cb.ern(p, fq, msg);
.cb.ern = cb_ern,
} __packed __aligned(64) ern;