Symbol: send_queue
drivers/infiniband/core/mad.c
1031
mad_send_wr->mad_list.mad_queue = &qp_info->send_queue;
drivers/infiniband/core/mad.c
1058
spin_lock_irqsave(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
1059
if (qp_info->send_queue.count < qp_info->send_queue.max_active) {
drivers/infiniband/core/mad.c
1063
list = &qp_info->send_queue.list;
drivers/infiniband/core/mad.c
1070
qp_info->send_queue.count++;
drivers/infiniband/core/mad.c
1073
spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
2508
struct ib_mad_queue *send_queue;
drivers/infiniband/core/mad.c
2523
send_queue = mad_list->mad_queue;
drivers/infiniband/core/mad.c
2524
qp_info = send_queue->qp_info;
drivers/infiniband/core/mad.c
2537
spin_lock_irqsave(&send_queue->lock, flags);
drivers/infiniband/core/mad.c
2541
if (send_queue->count-- > send_queue->max_active) {
drivers/infiniband/core/mad.c
2547
list_move_tail(&mad_list->list, &send_queue->list);
drivers/infiniband/core/mad.c
2549
spin_unlock_irqrestore(&send_queue->lock, flags);
drivers/infiniband/core/mad.c
2576
spin_lock_irqsave(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
2577
list_for_each_entry(mad_list, &qp_info->send_queue.list, list) {
drivers/infiniband/core/mad.c
2583
spin_unlock_irqrestore(&qp_info->send_queue.lock, flags);
drivers/infiniband/core/mad.c
3142
init_mad_queue(qp_info, &qp_info->send_queue);
drivers/infiniband/core/mad.c
3174
qp_info->send_queue.max_active = mad_sendq_size;
drivers/infiniband/core/mad_priv.h
256
struct ib_mad_queue send_queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
1711
queue = &ctx->send_queue[ctx->queues - 1];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
132
queue = &ctx->send_queue[rule_attr.queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
15
return min(ctx->send_queue[queue_id].num_entries / 2,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
425
queue_full = mlx5hws_send_engine_full(&ctx->send_queue[queue_id]);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/context.h
49
struct mlx5hws_send_engine *send_queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
223
struct mlx5hws_send_engine *send_queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
230
send_queue = &ctx->send_queue[i];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
235
send_queue->used_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
236
send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
238
send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
239
send_queue->err,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
240
send_queue->completed.ci,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
241
send_queue->completed.pi,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
242
send_queue->completed.mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
244
send_ring = &send_queue->send_ring;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
313
ctx->send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
376
queue = &ctx->send_queue[ctx->queues - 1];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
312
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
417
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
455
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
528
if (unlikely(mlx5hws_send_engine_full(&ctx->send_queue[attr->queue_id])))
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
622
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1058
mlx5hws_send_queue_close(&ctx->send_queue[queues]);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1082
kfree(ctx->send_queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1135
ctx->send_queue = kzalloc_objs(*ctx->send_queue, ctx->queues);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1136
if (!ctx->send_queue) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1148
err = mlx5hws_send_queue_open(ctx, &ctx->send_queue[i], queue_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1158
kfree(ctx->send_queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1175
queue = &ctx->send_queue[queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1297
queue_id = queue - ctx->send_queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
678
return hws_send_engine_poll(&ctx->send_queue[queue_id], res, res_nb);
drivers/net/hamradio/yam.c
1098
skb_queue_head_init(&yp->send_queue);
drivers/net/hamradio/yam.c
125
struct sk_buff_head send_queue; /* Packets awaiting transmission */
drivers/net/hamradio/yam.c
588
skb_queue_tail(&yp->send_queue, skb);
drivers/net/hamradio/yam.c
608
skb_queue_empty(&yp->send_queue))
drivers/net/hamradio/yam.c
658
if (!(skb = skb_dequeue(&yp->send_queue))) {
drivers/net/hamradio/yam.c
702
if (skb_queue_empty(&yp->send_queue)) {
drivers/net/hamradio/yam.c
913
while ((skb = skb_dequeue(&yp->send_queue)))
drivers/net/virtio_net.c
1074
static void free_old_xmit(struct send_queue *sq, struct netdev_queue *txq,
drivers/net/virtio_net.c
1105
struct send_queue *sq)
drivers/net/virtio_net.c
1137
struct send_queue *sq)
drivers/net/virtio_net.c
1494
static int virtnet_xsk_xmit_one(struct send_queue *sq,
drivers/net/virtio_net.c
1515
static int virtnet_xsk_xmit_batch(struct send_queue *sq,
drivers/net/virtio_net.c
1547
static bool virtnet_xsk_xmit(struct send_queue *sq, struct xsk_buff_pool *pool,
drivers/net/virtio_net.c
1589
static void xsk_wakeup(struct send_queue *sq)
drivers/net/virtio_net.c
1602
struct send_queue *sq;
drivers/net/virtio_net.c
1616
static void virtnet_xsk_completed(struct send_queue *sq, int num)
drivers/net/virtio_net.c
1628
struct send_queue *sq,
drivers/net/virtio_net.c
1719
struct send_queue *sq;
drivers/net/virtio_net.c
2889
static void virtnet_napi_tx_enable(struct send_queue *sq)
drivers/net/virtio_net.c
2910
static void virtnet_napi_tx_disable(struct send_queue *sq)
drivers/net/virtio_net.c
3026
struct send_queue *sq = &vi->sq[index];
drivers/net/virtio_net.c
3074
struct send_queue *sq;
drivers/net/virtio_net.c
3208
struct send_queue *sq = container_of(napi, struct send_queue, napi);
drivers/net/virtio_net.c
3261
static int xmit_skb(struct send_queue *sq, struct sk_buff *skb, bool orphan)
drivers/net/virtio_net.c
3323
struct send_queue *sq = &vi->sq[qnum];
drivers/net/virtio_net.c
3444
static void virtnet_tx_pause(struct virtnet_info *vi, struct send_queue *sq)
drivers/net/virtio_net.c
3474
static void virtnet_tx_resume(struct virtnet_info *vi, struct send_queue *sq)
drivers/net/virtio_net.c
3493
static int virtnet_tx_resize(struct virtnet_info *vi, struct send_queue *sq,
drivers/net/virtio_net.c
3643
struct send_queue *sq = &vi->sq[i];
drivers/net/virtio_net.c
394
struct send_queue *sq;
drivers/net/virtio_net.c
4112
struct send_queue *sq;
drivers/net/virtio_net.c
5074
struct send_queue *sq = &vi->sq[i];
drivers/net/virtio_net.c
526
static void virtnet_xsk_completed(struct send_queue *sq, int num);
drivers/net/virtio_net.c
5612
struct send_queue *sq = &vi->sq[i];
drivers/net/virtio_net.c
566
static int virtnet_add_outbuf(struct send_queue *sq, int num, void *data,
drivers/net/virtio_net.c
5823
struct send_queue *sq,
drivers/net/virtio_net.c
585
static void __free_old_xmit(struct send_queue *sq, struct netdev_queue *txq,
drivers/net/virtio_net.c
5853
struct send_queue *sq;
drivers/net/virtio_net.c
5935
struct send_queue *sq;
drivers/net/virtio_net.c
6150
struct send_queue *sq = &priv->sq[txqueue];
drivers/net/virtio_net.c
629
static void virtnet_free_old_xmit(struct send_queue *sq,
drivers/net/virtio_net.c
6307
struct send_queue *sq;
drivers/net/virtio_net.c
760
struct send_queue *sq)
drivers/net/virtio_net.c
777
struct send_queue *sq = &vi->sq[index];
drivers/net/wireless/ath/ath6kl/htc_pipe.c
303
struct list_head send_queue; /* temp queue to hold packets */
drivers/net/wireless/ath/ath6kl/htc_pipe.c
315
INIT_LIST_HEAD(&send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
357
list_splice_tail_init(txq, &send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
373
list_move_tail(&packet->list, &send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
398
&send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
402
if (list_empty(&send_queue)) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
418
if (!list_empty(&send_queue)) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
420
list_splice_tail_init(&send_queue, &ep->txq);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
421
if (!list_empty(&send_queue)) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
426
INIT_LIST_HEAD(&send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
461
get_htc_packet_credit_based(target, ep, &send_queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
467
get_htc_packet(target, ep, &send_queue, tx_resources);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
470
if (get_queue_depth(&send_queue) == 0) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
481
htc_issue_packets(target, ep, &send_queue);
fs/dlm/midcomms.c
164
struct list_head send_queue;
fs/dlm/midcomms.c
310
list_for_each_entry_rcu(mh, &node->send_queue, list) {
fs/dlm/midcomms.c
362
INIT_LIST_HEAD(&node->send_queue);
fs/dlm/midcomms.c
455
list_for_each_entry_rcu(mh, &node->send_queue, list) {
fs/dlm/midcomms.c
466
list_for_each_entry_rcu(mh, &node->send_queue, list) {
fs/dlm/midcomms.c
941
list_for_each_entry_rcu(mh, &node->send_queue, list) {
fs/dlm/midcomms.c
971
list_add_tail_rcu(&mh->list, &mh->node->send_queue);
net/rxrpc/ar-internal.h
754
struct rxrpc_txqueue *send_queue; /* Queue that sendmsg is writing into */
net/rxrpc/sendmsg.c
245
struct rxrpc_txqueue *sq = call->send_queue;
net/rxrpc/sendmsg.c
271
call->send_queue = NULL;
net/rxrpc/sendmsg.c
293
if (call->send_queue) {
net/rxrpc/sendmsg.c
295
call->send_queue->next = tq;
net/rxrpc/sendmsg.c
296
call->send_queue = tq;
net/rxrpc/sendmsg.c
306
call->send_queue = tq;
net/rxrpc/sendmsg.c
397
if (!call->send_queue || !((call->send_top + 1) & RXRPC_TXQ_MASK)) {