Symbol: qp_info
drivers/crypto/hisilicon/qm.c
2668
struct hisi_qp_info qp_info;
drivers/crypto/hisilicon/qm.c
2688
if (copy_from_user(&qp_info, (void __user *)arg,
drivers/crypto/hisilicon/qm.c
2692
qp_info.sqe_size = qp->qm->sqe_size;
drivers/crypto/hisilicon/qm.c
2693
qp_info.sq_depth = qp->sq_depth;
drivers/crypto/hisilicon/qm.c
2694
qp_info.cq_depth = qp->cq_depth;
drivers/crypto/hisilicon/qm.c
2696
if (copy_to_user((void __user *)arg, &qp_info,
drivers/infiniband/core/mad.c
1022
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad.c
1030
qp_info = mad_send_wr->mad_agent_priv->qp_info;
drivers/infiniband/core/mad.c
1031
mad_send_wr->mad_list.mad_queue = &qp_info->send_queue;
drivers/infiniband/core/mad.c
1058
spin_lock_irqsave(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
1059
if (qp_info->send_queue.count < qp_info->send_queue.max_active) {
drivers/infiniband/core/mad.c
1060
trace_ib_mad_ib_send_mad(mad_send_wr, qp_info);
drivers/infiniband/core/mad.c
1063
list = &qp_info->send_queue.list;
drivers/infiniband/core/mad.c
1066
list = &qp_info->overflow_list;
drivers/infiniband/core/mad.c
1070
qp_info->send_queue.count++;
drivers/infiniband/core/mad.c
1073
spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
1487
port_priv = agent_priv->qp_info->port_priv;
drivers/infiniband/core/mad.c
1551
port_priv = agent_priv->qp_info->port_priv;
drivers/infiniband/core/mad.c
1653
port_priv = agent_priv->qp_info->port_priv;
drivers/infiniband/core/mad.c
1818
const struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
1822
u32 qp_num = qp_info->qp->qp_num;
drivers/infiniband/core/mad.c
2115
const struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
2156
qp_info->qp->qp_num,
drivers/infiniband/core/mad.c
2202
struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
2246
qp_info->qp->qp_num,
drivers/infiniband/core/mad.c
2258
struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
2269
return handle_opa_smi(port_priv, qp_info, wc, port_num, recv,
drivers/infiniband/core/mad.c
2272
return handle_ib_smi(port_priv, qp_info, wc, port_num, recv, response);
drivers/infiniband/core/mad.c
2280
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad.c
2301
qp_info = mad_list->mad_queue->qp_info;
drivers/infiniband/core/mad.c
2304
opa = rdma_cap_opa_mad(qp_info->port_priv->device,
drivers/infiniband/core/mad.c
2305
qp_info->port_priv->port_num);
drivers/infiniband/core/mad.c
2331
if (!validate_mad((const struct ib_mad_hdr *)recv->mad, qp_info, opa))
drivers/infiniband/core/mad.c
2334
trace_ib_mad_recv_done_handler(qp_info, wc,
drivers/infiniband/core/mad.c
2349
if (handle_smi(port_priv, qp_info, wc, port_num, recv,
drivers/infiniband/core/mad.c
2374
qp_info->qp->qp_num,
drivers/infiniband/core/mad.c
2394
qp_info->qp->qp_num, mad_size, opa);
drivers/infiniband/core/mad.c
2400
ib_mad_post_receive_mads(qp_info, response);
drivers/infiniband/core/mad.c
2403
ib_mad_post_receive_mads(qp_info, recv);
drivers/infiniband/core/mad.c
2424
mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
drivers/infiniband/core/mad.c
2441
mod_delayed_work(mad_agent_priv->qp_info->port_priv->wq,
drivers/infiniband/core/mad.c
2507
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad.c
2524
qp_info = send_queue->qp_info;
drivers/infiniband/core/mad.c
2542
mad_list = container_of(qp_info->overflow_list.next,
drivers/infiniband/core/mad.c
2557
trace_ib_mad_send_done_resend(queued_send_wr, qp_info);
drivers/infiniband/core/mad.c
2558
ret = ib_post_send(qp_info->qp, &queued_send_wr->send_wr.wr,
drivers/infiniband/core/mad.c
2570
static void mark_sends_for_retry(struct ib_mad_qp_info *qp_info)
drivers/infiniband/core/mad.c
2576
spin_lock_irqsave(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
2577
list_for_each_entry(mad_list, &qp_info->send_queue.list, list) {
drivers/infiniband/core/mad.c
2583
spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
2591
struct ib_mad_qp_info *qp_info = mad_list->mad_queue->qp_info;
drivers/infiniband/core/mad.c
2605
trace_ib_mad_error_handler(mad_send_wr, qp_info);
drivers/infiniband/core/mad.c
2606
ret = ib_post_send(qp_info->qp, &mad_send_wr->send_wr.wr,
drivers/infiniband/core/mad.c
2619
ret = ib_modify_qp(qp_info->qp, attr,
drivers/infiniband/core/mad.c
2627
mark_sends_for_retry(qp_info);
drivers/infiniband/core/mad.c
2762
opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
drivers/infiniband/core/mad.c
2763
mad_agent_priv->qp_info->port_priv->port_num);
drivers/infiniband/core/mad.c
2899
queue_delayed_work(mad_agent_priv->qp_info->
drivers/infiniband/core/mad.c
2926
static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
2933
struct ib_mad_queue *recv_queue = &qp_info->recv_queue;
drivers/infiniband/core/mad.c
2937
sg_list.lkey = qp_info->port_priv->pd->local_dma_lkey;
drivers/infiniband/core/mad.c
2950
mad_priv = alloc_mad_private(port_mad_size(qp_info->port_priv),
drivers/infiniband/core/mad.c
2956
sg_list.addr = ib_dma_map_single(qp_info->port_priv->device,
drivers/infiniband/core/mad.c
2960
if (unlikely(ib_dma_mapping_error(qp_info->port_priv->device,
drivers/infiniband/core/mad.c
2980
ret = ib_post_recv(qp_info->qp, &recv_wr, NULL);
drivers/infiniband/core/mad.c
2986
dev_err(&qp_info->port_priv->device->dev,
drivers/infiniband/core/mad.c
2992
ib_dma_unmap_single(qp_info->port_priv->device,
drivers/infiniband/core/mad.c
3003
static void cleanup_recv_queue(struct ib_mad_qp_info *qp_info)
drivers/infiniband/core/mad.c
3009
if (!qp_info->qp)
drivers/infiniband/core/mad.c
3012
while (!list_empty(&qp_info->recv_queue.list)) {
drivers/infiniband/core/mad.c
3014
mad_list = list_entry(qp_info->recv_queue.list.next,
drivers/infiniband/core/mad.c
3025
ib_dma_unmap_single(qp_info->port_priv->device,
drivers/infiniband/core/mad.c
3032
qp_info->recv_queue.count = 0;
drivers/infiniband/core/mad.c
3055
qp = port_priv->qp_info[i].qp;
drivers/infiniband/core/mad.c
3104
if (!port_priv->qp_info[i].qp)
drivers/infiniband/core/mad.c
3107
ret = ib_mad_post_receive_mads(&port_priv->qp_info[i], NULL);
drivers/infiniband/core/mad.c
3121
struct ib_mad_qp_info *qp_info = qp_context;
drivers/infiniband/core/mad.c
3124
dev_err(&qp_info->port_priv->device->dev,
drivers/infiniband/core/mad.c
3126
event->event, qp_info->qp->qp_num);
drivers/infiniband/core/mad.c
3129
static void init_mad_queue(struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
3132
mad_queue->qp_info = qp_info;
drivers/infiniband/core/mad.c
3139
struct ib_mad_qp_info *qp_info)
drivers/infiniband/core/mad.c
3141
qp_info->port_priv = port_priv;
drivers/infiniband/core/mad.c
3142
init_mad_queue(qp_info, &qp_info->send_queue);
drivers/infiniband/core/mad.c
3143
init_mad_queue(qp_info, &qp_info->recv_queue);
drivers/infiniband/core/mad.c
3144
INIT_LIST_HEAD(&qp_info->overflow_list);
drivers/infiniband/core/mad.c
3147
static int create_mad_qp(struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
3154
qp_init_attr.send_cq = qp_info->port_priv->cq;
drivers/infiniband/core/mad.c
3155
qp_init_attr.recv_cq = qp_info->port_priv->cq;
drivers/infiniband/core/mad.c
3162
qp_init_attr.port_num = qp_info->port_priv->port_num;
drivers/infiniband/core/mad.c
3163
qp_init_attr.qp_context = qp_info;
drivers/infiniband/core/mad.c
3165
qp_info->qp = ib_create_qp(qp_info->port_priv->pd, &qp_init_attr);
drivers/infiniband/core/mad.c
3166
if (IS_ERR(qp_info->qp)) {
drivers/infiniband/core/mad.c
3167
dev_err(&qp_info->port_priv->device->dev,
drivers/infiniband/core/mad.c
3170
ret = PTR_ERR(qp_info->qp);
drivers/infiniband/core/mad.c
3174
qp_info->send_queue.max_active = mad_sendq_size;
drivers/infiniband/core/mad.c
3175
qp_info->recv_queue.max_active = mad_recvq_size;
drivers/infiniband/core/mad.c
3182
static void destroy_mad_qp(struct ib_mad_qp_info *qp_info)
drivers/infiniband/core/mad.c
3184
if (!qp_info->qp)
drivers/infiniband/core/mad.c
3187
ib_destroy_qp(qp_info->qp);
drivers/infiniband/core/mad.c
3217
init_mad_qp(port_priv, &port_priv->qp_info[0]);
drivers/infiniband/core/mad.c
3218
init_mad_qp(port_priv, &port_priv->qp_info[1]);
drivers/infiniband/core/mad.c
3241
ret = create_mad_qp(&port_priv->qp_info[0], IB_QPT_SMI);
drivers/infiniband/core/mad.c
3247
ret = create_mad_qp(&port_priv->qp_info[1], IB_QPT_GSI);
drivers/infiniband/core/mad.c
3278
destroy_mad_qp(&port_priv->qp_info[1]);
drivers/infiniband/core/mad.c
3280
destroy_mad_qp(&port_priv->qp_info[0]);
drivers/infiniband/core/mad.c
3283
cleanup_recv_queue(&port_priv->qp_info[1]);
drivers/infiniband/core/mad.c
3284
cleanup_recv_queue(&port_priv->qp_info[0]);
drivers/infiniband/core/mad.c
3314
destroy_mad_qp(&port_priv->qp_info[1]);
drivers/infiniband/core/mad.c
3315
destroy_mad_qp(&port_priv->qp_info[0]);
drivers/infiniband/core/mad.c
3318
cleanup_recv_queue(&port_priv->qp_info[1]);
drivers/infiniband/core/mad.c
3319
cleanup_recv_queue(&port_priv->qp_info[0]);
drivers/infiniband/core/mad.c
381
if (!port_priv->qp_info[qpn].qp) {
drivers/infiniband/core/mad.c
404
mad_agent_priv->qp_info = &port_priv->qp_info[qpn];
drivers/infiniband/core/mad.c
411
mad_agent_priv->agent.qp = port_priv->qp_info[qpn].qp;
drivers/infiniband/core/mad.c
524
port_priv = mad_agent_priv->qp_info->port_priv;
drivers/infiniband/core/mad.c
59
struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad.c
635
size_t mad_size = port_mad_size(mad_agent_priv->qp_info->port_priv);
drivers/infiniband/core/mad.c
638
bool opa = rdma_cap_opa_mad(mad_agent_priv->qp_info->port_priv->device,
drivers/infiniband/core/mad.c
639
mad_agent_priv->qp_info->port_priv->port_num);
drivers/infiniband/core/mad.c
786
queue_work(mad_agent_priv->qp_info->port_priv->wq,
drivers/infiniband/core/mad.c
97
static int ib_mad_post_receive_mads(struct ib_mad_qp_info *qp_info,
drivers/infiniband/core/mad_priv.h
118
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad_priv.h
250
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad_priv.h
275
struct ib_mad_qp_info qp_info[IB_MAD_QPS_CORE];
drivers/infiniband/core/mad_priv.h
94
struct ib_mad_qp_info *qp_info;
drivers/infiniband/core/mad_rmpp.c
396
return max(agent->qp_info->recv_queue.max_active >> 3, 1);
drivers/infiniband/core/mad_rmpp.c
431
bool opa = rdma_cap_opa_mad(rmpp_recv->agent->qp_info->port_priv->device,
drivers/infiniband/core/mad_rmpp.c
432
rmpp_recv->agent->qp_info->port_priv->port_num);
drivers/infiniband/core/mad_rmpp.c
463
queue_delayed_work(rmpp_recv->agent->qp_info->port_priv->wq,
drivers/infiniband/core/mad_rmpp.c
556
queue_delayed_work(agent->qp_info->port_priv->wq,
drivers/infiniband/core/mad_rmpp.c
99
flush_workqueue(agent->qp_info->port_priv->wq);
drivers/infiniband/hw/irdma/utils.c
1065
struct irdma_create_qp_info *qp_info;
drivers/infiniband/hw/irdma/utils.c
1073
qp_info = &cqp_request->info.in.u.qp_create.info;
drivers/infiniband/hw/irdma/utils.c
1074
qp_info->cq_num_valid = true;
drivers/infiniband/hw/irdma/utils.c
1075
qp_info->next_iwarp_state = IRDMA_QP_STATE_RTS;
drivers/infiniband/hw/irdma/verbs.c
767
struct irdma_create_qp_info *qp_info;
drivers/infiniband/hw/irdma/verbs.c
775
qp_info = &cqp_request->info.in.u.qp_create.info;
drivers/infiniband/hw/irdma/verbs.c
776
qp_info->mac_valid = true;
drivers/infiniband/hw/irdma/verbs.c
777
qp_info->cq_num_valid = true;
drivers/infiniband/hw/irdma/verbs.c
778
qp_info->next_iwarp_state = IRDMA_QP_STATE_IDLE;
drivers/scsi/bnx2i/bnx2i.h
760
struct qp_info qp;
drivers/scsi/bnx2i/bnx2i_hwi.c
1960
struct qp_info *qp;
include/trace/events/ib_mad.h
109
struct ib_mad_qp_info *qp_info),
include/trace/events/ib_mad.h
110
TP_ARGS(wr, qp_info));
include/trace/events/ib_mad.h
113
struct ib_mad_qp_info *qp_info),
include/trace/events/ib_mad.h
114
TP_ARGS(wr, qp_info));
include/trace/events/ib_mad.h
117
struct ib_mad_qp_info *qp_info),
include/trace/events/ib_mad.h
118
TP_ARGS(wr, qp_info));
include/trace/events/ib_mad.h
146
__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
include/trace/events/ib_mad.h
184
TP_PROTO(struct ib_mad_qp_info *qp_info, struct ib_wc *wc,
include/trace/events/ib_mad.h
186
TP_ARGS(qp_info, wc, mad_hdr),
include/trace/events/ib_mad.h
19
struct ib_mad_qp_info *qp_info,
include/trace/events/ib_mad.h
209
__entry->dev_index = qp_info->port_priv->device->index;
include/trace/events/ib_mad.h
210
__entry->port_num = qp_info->port_priv->port_num;
include/trace/events/ib_mad.h
211
__entry->qp_num = qp_info->qp->qp_num;
include/trace/events/ib_mad.h
25
struct ib_mad_qp_info *qp_info),
include/trace/events/ib_mad.h
26
TP_ARGS(wr, qp_info),
include/trace/events/ib_mad.h
57
__entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
include/trace/events/ib_mad.h
83
create_mad_addr_info(wr, qp_info, __entry);