arch/s390/include/uapi/asm/runtime_instr.h
22
__u32 qs : 1;
arch/s390/kernel/ptrace.c
822
cb->qs == 0 &&
drivers/md/dm-cache-policy-smq.c
258
struct ilist qs[MAX_LEVELS];
drivers/md/dm-cache-policy-smq.c
279
l_init(q->qs + i);
drivers/md/dm-cache-policy-smq.c
303
l_add_tail(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
313
l_add_head(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
323
l_add_before(q->es, q->qs + e->level, old, e);
drivers/md/dm-cache-policy-smq.c
328
l_del(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
344
for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) {
drivers/md/dm-cache-policy-smq.c
378
for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e))
drivers/md/dm-cache-policy-smq.c
380
l_del(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
438
l = q->qs + level;
drivers/md/dm-cache-policy-smq.c
458
l_above = q->qs + level + 1u;
drivers/md/dm-cache-policy-smq.c
481
for (de = l_head(q->es, q->qs + new_level); de && de->sentinel; de = l_next(q->es, de))
drivers/net/ethernet/brocade/bna/bfi_enet.h
494
struct bfi_enet_rxq qs; /* small/header buffers */
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1635
bfi_enet_datapath_q_init(&cfg_req->q_cfg[i].qs.q,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1637
cfg_req->q_cfg[i].qs.rx_buffer_size =
drivers/net/ethernet/cavium/thunder/nic.h
283
struct queue_set *qs;
drivers/net/ethernet/cavium/thunder/nic.h
587
struct qs_cfg_msg qs;
drivers/net/ethernet/cavium/thunder/nic_main.c
1026
nic_tx_channel_cfg(nic, mbx.qs.num, &mbx.sq);
drivers/net/ethernet/cavium/thunder/nic_main.c
981
(mbx.qs.num << NIC_QS_ID_SHIFT);
drivers/net/ethernet/cavium/thunder/nic_main.c
982
cfg = mbx.qs.cfg;
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
217
for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
225
for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
282
(nic->qs->rq_cnt + nic->qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
290
(snic->qs->rq_cnt + snic->qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
306
for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
309
*((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats)
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
313
for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
316
*((*data)++) = ((u64 *)&nic->qs->sq[qidx].stats)
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
475
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
478
ring->rx_pending = qs->cq_len;
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
480
ring->tx_pending = qs->sq_len;
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
489
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
504
if ((tx_count == qs->sq_len) && (rx_count == qs->cq_len))
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
508
qs->sq_len = rounddown_pow_of_two(tx_count);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
509
qs->cq_len = rounddown_pow_of_two(rx_count);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
752
nic->qs->rq_cnt = min_t(u8, nic->rx_queues, MAX_RCV_QUEUES_PER_QS);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
753
nic->qs->sq_cnt = min_t(u8, txq_count, MAX_SND_QUEUES_PER_QS);
drivers/net/ethernet/cavium/thunder/nicvf_ethtool.c
754
nic->qs->cq_cnt = max(nic->qs->rq_cnt, nic->qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1004
nicvf_cmp_queue_config(nic, qs, qidx, true);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1005
nicvf_sq_free_used_descs(nic->netdev, &qs->sq[qidx], qidx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1006
nicvf_sq_enable(nic, &qs->sq[qidx], qidx);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1068
for (qidx = 0; qidx < nic->qs->rbdr_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1138
for (irq = 0; irq < nic->qs->cq_cnt; irq++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1149
irq < (NICVF_INTR_ID_RBDR + nic->qs->rbdr_cnt); irq++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1288
sq = &snic->qs->sq[qid];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1315
for (qidx = 0; qidx < nic->qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1328
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1356
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1372
for (qidx = 0; qidx < nic->qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1453
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1467
for (qidx = 0; qidx < qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1543
for (qidx = 0; qidx < qs->cq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1547
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1567
for (qidx = 0; qidx < qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1656
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1707
for (qidx = 0; qidx < qs->rq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1709
for (qidx = 0; qidx < qs->sq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1818
nic->qs->rq_cnt = min_t(u8, nic->rx_queues, MAX_RCV_QUEUES_PER_QS);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1819
nic->qs->sq_cnt = min_t(u8, txq_count, MAX_SND_QUEUES_PER_QS);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1820
nic->qs->cq_cnt = max_t(u8, nic->qs->rq_cnt, nic->qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
319
mbx.cpi_cfg.rq_cnt = nic->qs->rq_cnt;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
441
nic->snicvf[sqs]->qs->rq_cnt = MAX_RCV_QUEUES_PER_QS;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
444
nic->snicvf[sqs]->qs->rq_cnt = rx_queues;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
449
nic->snicvf[sqs]->qs->sq_cnt = MAX_SND_QUEUES_PER_QS;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
452
nic->snicvf[sqs]->qs->sq_cnt = tx_queues;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
456
nic->snicvf[sqs]->qs->cq_cnt =
drivers/net/ethernet/cavium/thunder/nicvf_main.c
457
max(nic->snicvf[sqs]->qs->rq_cnt, nic->snicvf[sqs]->qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_main.c
667
sq = &nic->qs->sq[cqe_tx->sq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
852
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
853
struct cmp_queue *cq = &qs->cq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
856
struct snd_queue *sq = &qs->sq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
857
struct rcv_queue *rq = &qs->rq[cq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_main.c
988
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
995
for (qidx = 0; qidx < qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1002
for (qidx = 0; qidx < qs->sq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1003
nicvf_free_snd_queue(nic, &qs->sq[qidx]);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1009
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1012
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1013
if (nicvf_init_rbdr(nic, &qs->rbdr[qidx], qs->rbdr_len,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1019
for (qidx = 0; qidx < qs->sq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1020
if (nicvf_init_snd_queue(nic, &qs->sq[qidx], qs->sq_len, qidx))
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1025
for (qidx = 0; qidx < qs->cq_cnt; qidx++) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1026
if (nicvf_init_cmp_queue(nic, &qs->cq[qidx], qs->cq_len))
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1038
struct queue_set *qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1040
qs = devm_kzalloc(&nic->pdev->dev, sizeof(*qs), GFP_KERNEL);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1041
if (!qs)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1043
nic->qs = qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1046
qs->rbdr_cnt = DEFAULT_RBDR_CNT;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1047
qs->rq_cnt = min_t(u8, MAX_RCV_QUEUES_PER_QS, num_online_cpus());
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1048
qs->sq_cnt = min_t(u8, MAX_SND_QUEUES_PER_QS, num_online_cpus());
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1049
qs->cq_cnt = max_t(u8, qs->rq_cnt, qs->sq_cnt);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1052
qs->rbdr_len = RCV_BUF_COUNT;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1053
qs->sq_len = SND_QUEUE_LEN;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1054
qs->cq_len = CMP_QUEUE_LEN;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1056
nic->rx_queues = qs->rq_cnt;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1057
nic->tx_queues = qs->sq_cnt;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1066
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1067
struct queue_set *pqs = nic->pnicvf->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1070
if (!qs)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1078
qs->cq_len = pqs->cq_len;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1079
qs->sq_len = pqs->sq_len;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1086
for (qidx = 0; qidx < qs->sq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1087
nicvf_snd_queue_config(nic, qs, qidx, enable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1088
for (qidx = 0; qidx < qs->cq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1089
nicvf_cmp_queue_config(nic, qs, qidx, enable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1090
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1091
nicvf_rbdr_config(nic, qs, qidx, enable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1092
for (qidx = 0; qidx < qs->rq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1093
nicvf_rcv_queue_config(nic, qs, qidx, enable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1095
for (qidx = 0; qidx < qs->rq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1096
nicvf_rcv_queue_config(nic, qs, qidx, disable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1097
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1098
nicvf_rbdr_config(nic, qs, qidx, disable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1099
for (qidx = 0; qidx < qs->sq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1100
nicvf_snd_queue_config(nic, qs, qidx, disable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1101
for (qidx = 0; qidx < qs->cq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1102
nicvf_cmp_queue_config(nic, qs, qidx, disable);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1819
rq = &nic->qs->rq[rq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1832
sq = &nic->qs->sq[sq_idx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
381
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
382
int rbdr_idx = qs->rbdr_cnt;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
394
rbdr = &qs->rbdr[rbdr_idx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
403
if (qcount >= (qs->rbdr_len - 1))
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
406
refill_rb_cnt = qs->rbdr_len - qcount - 1;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
627
struct queue_set *qs, int qidx)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
639
struct queue_set *qs, int qidx)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
649
struct queue_set *qs, int qidx)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
744
static void nicvf_rcv_queue_config(struct nicvf *nic, struct queue_set *qs,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
751
rq = &qs->rq[qidx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
758
nicvf_reclaim_rcv_queue(nic, qs, qidx);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
763
rq->cq_qs = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
765
rq->start_rbdr_qs = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
766
rq->start_qs_rbdr_idx = qs->rbdr_cnt - 1;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
767
rq->cont_rbdr_qs = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
768
rq->cont_qs_rbdr_idx = qs->rbdr_cnt - 1;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
777
mbx.rq.qs_num = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
788
(qs->vnic_id << 0);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
817
void nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
823
cq = &qs->cq[qidx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
827
nicvf_reclaim_cmp_queue(nic, qs, qidx);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
847
cq_cfg.qsize = ilog2(qs->cq_len >> 10);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
858
static void nicvf_snd_queue_config(struct nicvf *nic, struct queue_set *qs,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
865
sq = &qs->sq[qidx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
869
nicvf_reclaim_snd_queue(nic, qs, qidx);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
876
sq->cq_qs = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
881
mbx.sq.qs_num = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
896
sq_cfg.qsize = ilog2(qs->sq_len >> 10);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
901
sq_cfg.cq_limit = (CMP_QUEUE_PIPELINE_RSVD * 256) / qs->cq_len;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
916
static void nicvf_rbdr_config(struct nicvf *nic, struct queue_set *qs,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
922
rbdr = &qs->rbdr[qidx];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
945
qidx, qs->rbdr_len - 1);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
956
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
959
if (!qs) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
965
qs->enable = enable;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
966
qs->vnic_id = nic->vf_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
969
mbx.qs.msg = NIC_MBOX_MSG_QS_CFG;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
970
mbx.qs.num = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
971
mbx.qs.sqs_count = nic->sqs_count;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
973
mbx.qs.cfg = 0;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
974
qs_cfg = (struct qs_cfg *)&mbx.qs.cfg;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
975
if (qs->enable) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
980
qs_cfg->vnic = qs->vnic_id;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
991
struct queue_set *qs = nic->qs;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
994
for (qidx = 0; qidx < qs->rbdr_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
995
nicvf_free_rbdr(nic, &qs->rbdr[qidx]);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
998
for (qidx = 0; qidx < qs->cq_cnt; qidx++)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
999
nicvf_free_cmp_queue(nic, &qs->cq[qidx]);
drivers/net/ethernet/cavium/thunder/nicvf_queues.h
335
void nicvf_cmp_queue_config(struct nicvf *nic, struct queue_set *qs,
drivers/net/ethernet/chelsio/cxgb3/adapter.h
216
struct sge_qset qs[SGE_QSETS];
drivers/net/ethernet/chelsio/cxgb3/adapter.h
325
void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
drivers/net/ethernet/chelsio/cxgb3/adapter.h
68
struct sge_qset *qs;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1170
struct sge_rspq *q = &adap->sge.qs[i].rspq;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1276
t3_intr_handler(adap, adap->sge.qs[0].rspq.polling),
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1657
tot += adapter->sge.qs[i].port_stats[idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2009
struct sge_qset *qs;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2017
qs = &adapter->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2019
t3_update_qset_coalesce(qs, qsp);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2218
struct sge_qset *qs =
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2219
&adapter->sge.qs[t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2222
t3_update_qset_coalesce(qs, q);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2620
struct sge_qset *qs = &adapter->sge.qs[qidx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2624
source = qs;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2628
t3_intr_handler(adapter, qs->rspq.polling) (0, source);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2768
struct sge_qset *qs = &adapter->sge.qs[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2777
qs->fl[i].empty += (v & 1);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2779
qs++;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
409
adap->sge.qs[qidx].
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
412
&adap->sge.qs[qidx]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
416
&adap->sge.qs[qidx]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
436
&adapter->sge.qs[i]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
446
while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
459
unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
595
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
597
if (qs->adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
599
t3_write_reg(adap, A_SG_KDOORBELL, F_SELEGRCNTX | V_EGRCNTX(qs->txq[j].cntxt_id));
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
608
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
610
if (qs->adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
611
netif_napi_add(qs->netdev, &qs->napi, qs->napi.poll);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
632
if (adap->sge.qs[i].adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
633
napi_disable(&adap->sge.qs[i].napi);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
640
if (adap->sge.qs[i].adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
641
napi_enable(&adap->sge.qs[i].napi);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
664
pi->qs = &adap->sge.qs[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1246
struct sge_qset *qs, struct sge_txq *q)
drivers/net/ethernet/chelsio/cxgb3/sge.c
1249
set_bit(TXQ_ETH, &qs->txq_stopped);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1267
struct sge_qset *qs;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1281
qs = &pi->qs[qidx];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1282
q = &qs->txq[TXQ_ETH];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1291
t3_stop_tx_queue(txq, qs, q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1308
t3_stop_tx_queue(txq, qs, q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1311
test_and_clear_bit(TXQ_ETH, &qs->txq_stopped)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1330
qs->port_stats[SGE_PSTAT_TX_CSUM]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1332
qs->port_stats[SGE_PSTAT_TSO]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1334
qs->port_stats[SGE_PSTAT_VLANINS]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1427
struct sge_qset *qs = txq_to_qset(q, qid);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1429
set_bit(qid, &qs->txq_stopped);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1433
test_and_clear_bit(qid, &qs->txq_stopped))
drivers/net/ethernet/chelsio/cxgb3/sge.c
1523
struct sge_qset *qs = container_of(w, struct sge_qset,
drivers/net/ethernet/chelsio/cxgb3/sge.c
1525
struct sge_txq *q = &qs->txq[TXQ_CTRL];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1543
set_bit(TXQ_CTRL, &qs->txq_stopped);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1547
test_and_clear_bit(TXQ_CTRL, &qs->txq_stopped))
drivers/net/ethernet/chelsio/cxgb3/sge.c
1554
t3_write_reg(qs->adap, A_SG_KDOORBELL,
drivers/net/ethernet/chelsio/cxgb3/sge.c
1565
ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1743
struct sge_qset *qs = container_of(w, struct sge_qset,
drivers/net/ethernet/chelsio/cxgb3/sge.c
1745
struct sge_txq *q = &qs->txq[TXQ_OFLD];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1746
const struct port_info *pi = netdev_priv(qs->netdev);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1758
set_bit(TXQ_OFLD, &qs->txq_stopped);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1762
test_and_clear_bit(TXQ_OFLD, &qs->txq_stopped))
drivers/net/ethernet/chelsio/cxgb3/sge.c
1836
struct sge_qset *qs = &adap->sge.qs[queue_set(skb)];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1839
return ctrl_xmit(adap, &qs->txq[TXQ_CTRL], skb);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1841
return ofld_xmit(adap, &qs->txq[TXQ_OFLD], skb);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1860
struct sge_qset *qs = rspq_to_qset(q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1862
napi_schedule(&qs->napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1898
struct sge_qset *qs = container_of(napi, struct sge_qset, napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1899
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1900
struct adapter *adapter = qs->adap;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1985
static void restart_tx(struct sge_qset *qs)
drivers/net/ethernet/chelsio/cxgb3/sge.c
1987
if (test_bit(TXQ_ETH, &qs->txq_stopped) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
1988
should_restart_tx(&qs->txq[TXQ_ETH]) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
1989
test_and_clear_bit(TXQ_ETH, &qs->txq_stopped)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1990
qs->txq[TXQ_ETH].restarts++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1991
if (netif_running(qs->netdev))
drivers/net/ethernet/chelsio/cxgb3/sge.c
1992
netif_tx_wake_queue(qs->tx_q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1995
if (test_bit(TXQ_OFLD, &qs->txq_stopped) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
1996
should_restart_tx(&qs->txq[TXQ_OFLD]) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
1997
test_and_clear_bit(TXQ_OFLD, &qs->txq_stopped)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1998
qs->txq[TXQ_OFLD].restarts++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2001
queue_work(cxgb3_wq, &qs->txq[TXQ_OFLD].qresume_task);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2003
if (test_bit(TXQ_CTRL, &qs->txq_stopped) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
2004
should_restart_tx(&qs->txq[TXQ_CTRL]) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
2005
test_and_clear_bit(TXQ_CTRL, &qs->txq_stopped)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2006
qs->txq[TXQ_CTRL].restarts++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2009
queue_work(cxgb3_wq, &qs->txq[TXQ_CTRL].qresume_task);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2088
struct sge_qset *qs = rspq_to_qset(rq);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2096
qs->port_stats[SGE_PSTAT_RX_CSUM_GOOD]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2100
skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2103
qs->port_stats[SGE_PSTAT_VLANEX]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2108
napi_gro_receive(&qs->napi, skb);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2134
static void lro_add_page(struct adapter *adap, struct sge_qset *qs,
drivers/net/ethernet/chelsio/cxgb3/sge.c
2138
struct port_info *pi = netdev_priv(qs->netdev);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2145
if (!qs->nomem) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2146
skb = napi_get_frags(&qs->napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2147
qs->nomem = !skb;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2164
qs->nomem = 0;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2173
cpl = qs->lro_va = sd->pg_chunk.va + 2;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2175
if ((qs->netdev->features & NETIF_F_RXCSUM) &&
drivers/net/ethernet/chelsio/cxgb3/sge.c
2178
qs->port_stats[SGE_PSTAT_RX_CSUM_GOOD]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2182
cpl = qs->lro_va;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2198
skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2201
qs->port_stats[SGE_PSTAT_VLANEX]++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2204
napi_gro_frags(&qs->napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2216
static inline void handle_rsp_cntrl_info(struct sge_qset *qs, u32 flags)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2222
clear_bit(TXQ_RUNNING, &qs->txq[TXQ_ETH].flags);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2227
qs->txq[TXQ_ETH].processed += credits;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2231
qs->txq[TXQ_CTRL].processed += credits;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2235
clear_bit(TXQ_RUNNING, &qs->txq[TXQ_OFLD].flags);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2239
qs->txq[TXQ_OFLD].processed += credits;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2252
static void check_ring_db(struct adapter *adap, struct sge_qset *qs,
drivers/net/ethernet/chelsio/cxgb3/sge.c
2256
struct sge_txq *txq = &qs->txq[TXQ_ETH];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2267
struct sge_txq *txq = &qs->txq[TXQ_OFLD];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2322
static int process_responses(struct adapter *adap, struct sge_qset *qs,
drivers/net/ethernet/chelsio/cxgb3/sge.c
2325
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2336
int lro = !!(qs->netdev->features & NETIF_F_GRO);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2373
fl = (len & F_RSPD_FLQ) ? &qs->fl[1] : &qs->fl[0];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2380
lro_add_page(adap, qs, fl,
drivers/net/ethernet/chelsio/cxgb3/sge.c
2408
handle_rsp_cntrl_info(qs, flags);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2450
check_ring_db(adap, qs, sleeping);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2453
if (unlikely(qs->txq_stopped != 0))
drivers/net/ethernet/chelsio/cxgb3/sge.c
2454
restart_tx(qs);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2476
struct sge_qset *qs = container_of(napi, struct sge_qset, napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2477
struct adapter *adap = qs->adap;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2478
int work_done = process_responses(adap, qs, budget);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2497
t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) |
drivers/net/ethernet/chelsio/cxgb3/sge.c
2498
V_NEWTIMER(qs->rspq.next_holdoff) |
drivers/net/ethernet/chelsio/cxgb3/sge.c
2499
V_NEWINDEX(qs->rspq.cidx));
drivers/net/ethernet/chelsio/cxgb3/sge.c
2518
static int process_pure_responses(struct adapter *adap, struct sge_qset *qs,
drivers/net/ethernet/chelsio/cxgb3/sge.c
2521
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2537
handle_rsp_cntrl_info(qs, flags);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2551
check_ring_db(adap, qs, sleeping);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2554
if (unlikely(qs->txq_stopped != 0))
drivers/net/ethernet/chelsio/cxgb3/sge.c
2555
restart_tx(qs);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2577
struct sge_qset *qs = rspq_to_qset(q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2583
if (is_pure_response(r) && process_pure_responses(adap, qs, r) == 0) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2588
napi_schedule(&qs->napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2598
struct sge_qset *qs = cookie;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2599
struct adapter *adap = qs->adap;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2600
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2603
if (process_responses(adap, qs, -1) == 0)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2617
struct sge_qset *qs = cookie;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2618
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2622
if (handle_responses(qs->adap, q) < 0)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2638
struct sge_rspq *q = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2642
if (process_responses(adap, &adap->sge.qs[0], -1)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2649
process_responses(adap, &adap->sge.qs[1], -1)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2650
struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2665
static int rspq_check_napi(struct sge_qset *qs)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2667
struct sge_rspq *q = &qs->rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2669
return is_new_response(&q->desc[q->cidx], q) && napi_schedule(&qs->napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2683
struct sge_rspq *q = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2687
new_packets = rspq_check_napi(&adap->sge.qs[0]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2689
new_packets += rspq_check_napi(&adap->sge.qs[1]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2721
struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2722
struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2759
struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2776
process_responses_gts(adap, &adap->sge.qs[1].rspq);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2793
struct sge_qset *qs0 = &adap->sge.qs[0];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2811
napi_schedule(&adap->sge.qs[1].napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2909
struct sge_qset *qs = timer_container_of(qs, t, tx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2910
struct port_info *pi = netdev_priv(qs->netdev);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2915
if (__netif_tx_trylock(qs->tx_q)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2916
tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH],
drivers/net/ethernet/chelsio/cxgb3/sge.c
2918
__netif_tx_unlock(qs->tx_q);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2921
if (spin_trylock(&qs->txq[TXQ_OFLD].lock)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2922
tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD],
drivers/net/ethernet/chelsio/cxgb3/sge.c
2924
spin_unlock(&qs->txq[TXQ_OFLD].lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2930
mod_timer(&qs->tx_reclaim_timer, jiffies + next_period);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2950
struct sge_qset *qs = timer_container_of(qs, t, rx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2951
struct port_info *pi = netdev_priv(qs->netdev);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2956
&qs->rspq.lock : &adap->sge.qs[0].rspq.lock;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2961
if (napi_is_scheduled(&qs->napi))
drivers/net/ethernet/chelsio/cxgb3/sge.c
2967
if (status & (1 << qs->rspq.cntxt_id)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2968
qs->rspq.starved++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2969
if (qs->rspq.credits) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2970
qs->rspq.credits--;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2971
refill_rspq(adap, &qs->rspq, 1);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2972
qs->rspq.restarted++;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2974
1 << qs->rspq.cntxt_id);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2979
if (qs->fl[0].credits < qs->fl[0].size)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2980
__refill_fl(adap, &qs->fl[0]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2981
if (qs->fl[1].credits < qs->fl[1].size)
drivers/net/ethernet/chelsio/cxgb3/sge.c
2982
__refill_fl(adap, &qs->fl[1]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2987
mod_timer(&qs->rx_reclaim_timer, jiffies + RX_RECLAIM_PERIOD);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2998
void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p)
drivers/net/ethernet/chelsio/cxgb3/sge.c
3000
qs->rspq.holdoff_tmr = max(p->coalesce_usecs * 10, 1U);/* can't be 0 */
drivers/net/ethernet/chelsio/cxgb3/sge.c
3001
qs->rspq.polling = p->polling;
drivers/net/ethernet/chelsio/cxgb3/sge.c
3002
qs->napi.poll = p->polling ? napi_rx_handler : ofld_poll;
drivers/net/ethernet/chelsio/cxgb3/sge.c
3027
struct sge_qset *q = &adapter->sge.qs[id];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3200
struct sge_qset *q = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3223
struct sge_qset *q = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3243
t3_free_qset(adap, &adap->sge.qs[i]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3292
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3294
cancel_work_sync(&qs->txq[TXQ_OFLD].qresume_task);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3295
cancel_work_sync(&qs->txq[TXQ_CTRL].qresume_task);
drivers/net/ethernet/chelsio/cxgb3/sge.c
726
static void init_qset_cntxt(struct sge_qset *qs, unsigned int id)
drivers/net/ethernet/chelsio/cxgb3/sge.c
728
qs->rspq.cntxt_id = id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
729
qs->fl[0].cntxt_id = 2 * id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
730
qs->fl[1].cntxt_id = 2 * id + 1;
drivers/net/ethernet/chelsio/cxgb3/sge.c
731
qs->txq[TXQ_ETH].cntxt_id = FW_TUNNEL_SGEEC_START + id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
732
qs->txq[TXQ_ETH].token = FW_TUNNEL_TID_START + id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
733
qs->txq[TXQ_OFLD].cntxt_id = FW_OFLD_SGEEC_START + id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
734
qs->txq[TXQ_CTRL].cntxt_id = FW_CTRL_SGEEC_START + id;
drivers/net/ethernet/chelsio/cxgb3/sge.c
735
qs->txq[TXQ_CTRL].token = FW_CTRL_TID_START + id;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1626
int qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1640
for (qs = pi->first_qset; qs < pi->first_qset + pi->nqsets; qs++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1641
s->ethrxq[qs].fl.size = rp->rx_pending + MIN_FL_RESID;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1642
s->ethrxq[qs].rspq.size = rp->rx_mini_pending;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1643
s->ethtxq[qs].q.size = rp->tx_pending;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1811
int qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1814
for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2045
int qs, r = (uintptr_t)v - 1;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2053
for (qs = 0; qs < n; ++qs) \
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2058
#define T(s, v) S3("u", s, txq[qs].v)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2059
#define R(s, v) S3("u", s, rxq[qs].v)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2068
(rxq[qs].rspq.netdev
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2069
? rxq[qs].rspq.netdev->name
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2072
(rxq[qs].rspq.netdev
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2074
netdev_priv(rxq[qs].rspq.netdev))->port_id
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2084
S3("u", "Intr delay:", qtimer_val(adapter, &rxq[qs].rspq));
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2086
adapter->sge.counter_val[rxq[qs].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2181
int qs, r = (uintptr_t)v - 1;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2189
for (qs = 0; qs < n; ++qs) \
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2195
#define T3(fmt, s, v) S3(fmt, s, txq[qs].v)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2198
#define R3(fmt, s, v) S3(fmt, s, rxq[qs].v)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2208
(rxq[qs].rspq.netdev
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2209
? rxq[qs].rspq.netdev->name
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2671
int q10g, n10g, qidx, pidx, qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2728
for (qs = 0; qs < s->max_ethqsets; qs++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2729
struct sge_eth_rxq *rxq = &s->ethrxq[qs];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2730
struct sge_eth_txq *txq = &s->ethtxq[qs];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
367
int qs, msi;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
369
for (qs = 0, msi = MSIX_IQFLINT; qs < pi->nqsets; qs++, msi++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
371
"%s-%d", dev->name, qs);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
634
int qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
636
for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
644
netdev_get_tx_queue(dev, qs),
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
649
rxq->rspq.idx = qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
665
int qs;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
667
for (qs = 0; qs < pi->nqsets; qs++, rxq++, txq++) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
707
int qs, err;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
709
for (qs = 0; qs < pi->nqsets; qs++)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
710
rss[qs] = rxq[qs].rspq.abs_id;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2568
int qs;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2570
for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) {
drivers/net/ethernet/hisilicon/hns/hnae.c
298
hnae_fini_queue(handle->qs[i]);
drivers/net/ethernet/hisilicon/hns/hnae.c
304
ret = hnae_init_queue(handle, handle->qs[i], handle->dev);
drivers/net/ethernet/hisilicon/hns/hnae.c
311
hnae_fini_queue(handle->qs[j]);
drivers/net/ethernet/hisilicon/hns/hnae.c
350
ret = hnae_init_queue(handle, handle->qs[i], dev);
drivers/net/ethernet/hisilicon/hns/hnae.c
363
hnae_fini_queue(handle->qs[j]);
drivers/net/ethernet/hisilicon/hns/hnae.c
377
hnae_fini_queue(h->qs[i]);
drivers/net/ethernet/hisilicon/hns/hnae.h
561
struct hnae_queue *qs[]; /* flexible array of all queues */
drivers/net/ethernet/hisilicon/hns/hnae.h
665
ring = &h->qs[i]->rx_ring;
drivers/net/ethernet/hisilicon/hns/hnae.h
680
ring = &h->qs[i]->rx_ring;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
111
ae_handle->qs[i] = &ring_pair_cb->q;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
112
ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i];
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
113
ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i];
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
143
hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
157
ret = hns_rcb_wait_tx_ring_clean(handle->qs[i]);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
189
hns_rcb_ring_enable_hw(handle->qs[i], val);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
315
q = handle->qs[i];
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
343
hns_rcb_int_clr_hw(handle->qs[k],
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
346
hns_rcbv2_int_clr_hw(handle->qs[k],
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
362
hns_rcb_wait_fbd_clean(handle->qs, handle->q_num, RCB_INT_FLAG_TX);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
373
hns_rcb_wait_fbd_clean(handle->qs, handle->q_num, RCB_INT_FLAG_RX);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
520
container_of(handle->qs[0], struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
532
container_of(handle->qs[0], struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
550
container_of(handle->qs[0], struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
561
container_of(handle->qs[0], struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
639
queue = handle->qs[idx];
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
718
hns_rcb_get_stats(handle->qs[idx], p);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
84
vf_cb = kzalloc_flex(*vf_cb, ae_handle.qs, qnum_per_vf);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
843
hns_rcb_get_ring_regs(handle->qs[i], p);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
36
void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag)
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
45
fbd_num += dsaf_read_dev(qs[i],
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
48
fbd_num += dsaf_read_dev(qs[i],
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
57
dev_err(qs[i]->handle->owner_dev,
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
61
int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs)
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
66
tail = dsaf_read_dev(&qs->tx_ring, RCB_REG_TAIL);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
69
head = dsaf_read_dev(&qs->tx_ring, RCB_REG_HEAD);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
77
dev_err(qs->dev->dev, "rcb wait timeout, head not equal to tail.\n");
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.h
133
void hns_rcb_wait_fbd_clean(struct hnae_queue **qs, int q_num, u32 flag);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.h
134
int hns_rcb_wait_tx_ring_clean(struct hnae_queue *qs);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1664
ring = &h->qs[i]->rx_ring;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1907
tx_bytes += h->qs[idx]->tx_ring.stats.tx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1908
tx_pkts += h->qs[idx]->tx_ring.stats.tx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1909
rx_bytes += h->qs[idx]->rx_ring.stats.rx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
1910
rx_pkts += h->qs[idx]->rx_ring.stats.rx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2008
i, h->qs[i]->tx_ring.next_to_clean);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2010
i, h->qs[i]->tx_ring.next_to_use);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2012
i, h->qs[i]->rx_ring.next_to_clean);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2014
i, h->qs[i]->rx_ring.next_to_use);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2123
rd->ring = &h->qs[i]->tx_ring;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
2134
rd->ring = &h->qs[i - h->q_num]->rx_ring;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
395
hnae_queue_xmit(priv->ae_handle->qs[skb->queue_mapping], buf_num);
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
676
queue = priv->ae_handle->qs[0];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1037
static void idpf_clean_queue_set(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1039
const struct idpf_q_vec_rsrc *rsrc = qs->qv_rsrc;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1041
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1042
const struct idpf_queue_ptr *q = &qs->qs[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1120
struct idpf_queue_set *qs;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1128
qs = idpf_alloc_queue_set(vport->adapter, &vport->dflt_qv_rsrc,
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1130
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1136
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_RX_BUFFER;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1137
qs->qs[num++].bufq = qv->bufq[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1141
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_RX;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1142
qs->qs[num++].rxq = qv->rx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1146
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1147
qs->qs[num++].txq = qv->tx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1151
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1152
qs->qs[num++].complq = qv->complq[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1162
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1163
qs->qs[num++].txq = vport->txqs[idx];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1165
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1166
qs->qs[num++].complq = vport->txqs[idx]->complq;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1170
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1171
qs->qs[num++].txq = qv->xsksq[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1173
qs->qs[num].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1174
qs->qs[num++].complq = qv->xsksq[i]->complq;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1179
if (num != qs->num) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1180
kfree(qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1184
return qs;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1188
const struct idpf_queue_set *qs, u32 qid)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1196
err = idpf_init_queue_set(vport, qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1211
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1212
const struct idpf_queue_ptr *q = &qs->qs[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1227
err = idpf_send_config_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1234
err = idpf_send_enable_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1250
const struct idpf_queue_set *qs, u32 qid)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1261
err = idpf_send_disable_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1268
idpf_clean_queue_set(qs);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1287
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1292
qs = idpf_vector_to_queue_set(q_vector);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1293
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1296
return en ? idpf_qp_enable(vport, qs, qid) :
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1297
idpf_qp_disable(vport, qs, qid);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
964
const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
969
splitq = idpf_is_queue_model_split(qs->qv_rsrc->rxq_model);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
971
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
972
const struct idpf_queue_ptr *q = &qs->qs[i];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1835
static int idpf_send_config_tx_queue_set_msg(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1839
.vport_id = qs->vport_id,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1846
qi = kzalloc_objs(*qi, qs->num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1852
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1853
if (qs->qs[i].type == VIRTCHNL2_QUEUE_TYPE_TX)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1854
idpf_fill_txq_config_chunk(qs->qv_rsrc, qs->qs[i].txq,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1856
else if (qs->qs[i].type == VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1857
idpf_fill_complq_config_chunk(qs->qv_rsrc,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1858
qs->qs[i].complq,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1862
return idpf_send_chunked_msg(qs->adapter, ¶ms);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1877
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1881
qs = idpf_alloc_queue_set(adapter, rsrc, vport_id, totqs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1882
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1890
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1891
qs->qs[k++].txq = tx_qgrp->txqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1895
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1896
qs->qs[k++].complq = tx_qgrp->complq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
1904
return idpf_send_config_tx_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2024
static int idpf_send_config_rx_queue_set_msg(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2028
.vport_id = qs->vport_id,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2035
qi = kzalloc_objs(*qi, qs->num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2041
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2042
if (qs->qs[i].type == VIRTCHNL2_QUEUE_TYPE_RX)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2043
idpf_fill_rxq_config_chunk(qs->qv_rsrc, qs->qs[i].rxq,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2045
else if (qs->qs[i].type == VIRTCHNL2_QUEUE_TYPE_RX_BUFFER)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2046
idpf_fill_bufq_config_chunk(qs->qv_rsrc, qs->qs[i].bufq,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2050
return idpf_send_chunked_msg(qs->adapter, ¶ms);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2066
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2070
qs = idpf_alloc_queue_set(adapter, rsrc, vport_id, totqs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2071
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2085
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_RX_BUFFER;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2086
qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2093
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_RX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2096
qs->qs[k++].rxq =
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2099
qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2107
return idpf_send_config_rx_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2148
static int idpf_send_ena_dis_queue_set_msg(const struct idpf_queue_set *qs,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2153
.vport_id = qs->vport_id,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2159
.num_chunks = qs->num,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2162
qc = kzalloc_objs(*qc, qs->num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2168
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2169
const struct idpf_queue_ptr *q = &qs->qs[i];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2195
return idpf_send_chunked_msg(qs->adapter, ¶ms);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2212
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2219
qs = idpf_alloc_queue_set(adapter, rsrc, vport_id, num_q);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2220
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2229
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2230
qs->qs[k++].txq = tx_qgrp->txqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2236
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2237
qs->qs[k++].complq = tx_qgrp->complq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2255
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_RX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2258
qs->qs[k++].rxq =
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2261
qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2268
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_RX_BUFFER;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2269
qs->qs[k++].bufq = &rx_qgrp->splitq.bufq_sets[j].bufq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2276
return idpf_send_ena_dis_queue_set_msg(qs, en);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2314
idpf_send_map_unmap_queue_set_vector_msg(const struct idpf_queue_set *qs,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2319
.vport_id = qs->vport_id,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2325
.num_chunks = qs->num,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2329
vqv = kzalloc_objs(*vqv, qs->num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2335
split = idpf_is_queue_model_split(qs->qv_rsrc->txq_model);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2337
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2338
const struct idpf_queue_ptr *q = &qs->qs[i];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2357
v_idx = qs->qv_rsrc->noirq_v_idx;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2377
v_idx = qs->qv_rsrc->noirq_v_idx;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2390
return idpf_send_chunked_msg(qs->adapter, ¶ms);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2407
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2411
qs = idpf_alloc_queue_set(adapter, rsrc, vport_id, num_q);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2412
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2419
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2420
qs->qs[k++].txq = tx_qgrp->txqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2437
qs->qs[k].type = VIRTCHNL2_QUEUE_TYPE_RX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2440
qs->qs[k++].rxq =
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2443
qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2450
return idpf_send_map_unmap_queue_set_vector_msg(qs, map);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2462
int idpf_send_enable_queue_set_msg(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2464
return idpf_send_ena_dis_queue_set_msg(qs, true);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2474
int idpf_send_disable_queue_set_msg(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2478
err = idpf_send_ena_dis_queue_set_msg(qs, false);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2482
return idpf_wait_for_marker_event_set(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2495
int idpf_send_config_queue_set_msg(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2499
err = idpf_send_config_tx_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2503
return idpf_send_config_rx_queue_set_msg(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
743
qp = kzalloc_flex(*qp, qs, num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
819
static int idpf_wait_for_marker_event_set(const struct idpf_queue_set *qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
825
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
826
switch (qs->qs[i].type) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
828
txq = qs->qs[i].txq;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
858
struct idpf_queue_set *qs __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
860
qs = idpf_alloc_queue_set(vport->adapter, &vport->dflt_qv_rsrc,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
862
if (!qs)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
865
for (u32 i = 0; i < qs->num; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
866
qs->qs[i].type = VIRTCHNL2_QUEUE_TYPE_TX;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
867
qs->qs[i].txq = vport->txqs[i];
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
870
return idpf_wait_for_marker_event_set(qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.h
146
struct idpf_queue_ptr qs[] __counted_by(num);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.h
153
int idpf_send_enable_queue_set_msg(const struct idpf_queue_set *qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.h
154
int idpf_send_disable_queue_set_msg(const struct idpf_queue_set *qs);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.h
155
int idpf_send_config_queue_set_msg(const struct idpf_queue_set *qs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
375
struct otx2_qset *qs = &pfvf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
378
ring->rx_pending = qs->rqe_cnt ? qs->rqe_cnt : Q_COUNT(Q_SIZE_256);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
380
ring->tx_pending = qs->sqe_cnt ? qs->sqe_cnt : Q_COUNT(Q_SIZE_4K);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
395
struct otx2_qset *qs = &pfvf->qset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
441
if (tx_count == qs->sqe_cnt && rx_count == qs->rqe_cnt &&
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
449
qs->sqe_cnt = tx_count;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_ethtool.c
450
qs->rqe_cnt = rx_count;
drivers/net/phy/sfp.c
591
static bool sfp_match(const char *qs, const char *str, size_t len)
drivers/net/phy/sfp.c
593
if (!qs)
drivers/net/phy/sfp.c
595
if (strlen(qs) != len)
drivers/net/phy/sfp.c
597
return !strncmp(qs, str, len);
drivers/s390/cio/qdio.h
116
: [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart),
drivers/s390/cio/qdio.h
97
: [ccq] "+&d" (_ccq), [qs] "+&d" (_queuestart)
drivers/scsi/elx/efct/efct_hw_queues.c
199
struct sli4_queue *qs[SLI4_MAX_CQ_SET_COUNT];
drivers/scsi/elx/efct/efct_hw_queues.c
217
qs[i] = cq->queue;
drivers/scsi/elx/efct/efct_hw_queues.c
222
if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) {
drivers/scsi/elx/efct/efct_hw_queues.c
315
struct sli4_queue *qs[SLI4_MAX_RQ_SET_COUNT * 2] = { NULL };
drivers/scsi/elx/efct/efct_hw_queues.c
342
qs[q_count] = rq->hdr;
drivers/scsi/elx/efct/efct_hw_queues.c
349
qs[q_count + 1] = rq->data;
drivers/scsi/elx/efct/efct_hw_queues.c
354
if (sli_fc_rq_set_alloc(&hw->sli, num_rq_pairs, qs,
drivers/scsi/elx/libefc_sli/sli4.c
428
struct sli4_queue *qs[], u32 base_cq_id,
drivers/scsi/elx/libefc_sli/sli4.c
439
page_count = sli_page_count(qs[0]->dma.size, SLI_PAGE_SIZE) * num_rqs;
drivers/scsi/elx/libefc_sli/sli4.c
463
num_pages = sli_page_count(qs[0]->dma.size, SLI_PAGE_SIZE);
drivers/scsi/elx/libefc_sli/sli4.c
465
req->rqe_count = cpu_to_le16(qs[0]->dma.size / SLI4_RQE_SIZE);
drivers/scsi/elx/libefc_sli/sli4.c
474
for (p = 0, addr = qs[i]->dma.phys; p < num_pages;
drivers/scsi/elx/libefc_sli/sli4.c
582
struct sli4_queue *qs[], u32 base_cq_id,
drivers/scsi/elx/libefc_sli/sli4.c
593
if (__sli_queue_init(sli4, qs[i], SLI4_QTYPE_RQ,
drivers/scsi/elx/libefc_sli/sli4.c
600
if (sli_cmd_rq_create_v2(sli4, num_rqs, qs, base_cq_id,
drivers/scsi/elx/libefc_sli/sli4.c
624
qs[i]->id = i + le16_to_cpu(rsp->q_id);
drivers/scsi/elx/libefc_sli/sli4.c
625
if ((qs[i]->id & 1) == 0)
drivers/scsi/elx/libefc_sli/sli4.c
626
qs[i]->u.flag |= SLI4_QUEUE_FLAG_HDR;
drivers/scsi/elx/libefc_sli/sli4.c
628
qs[i]->u.flag &= ~SLI4_QUEUE_FLAG_HDR;
drivers/scsi/elx/libefc_sli/sli4.c
630
qs[i]->db_regaddr = db_regaddr;
drivers/scsi/elx/libefc_sli/sli4.c
639
__sli_queue_destroy(sli4, qs[i]);
drivers/scsi/elx/libefc_sli/sli4.c
817
struct sli4_queue *qs[], u32 num_cqs,
drivers/scsi/elx/libefc_sli/sli4.c
829
n_cqe = qs[0]->dma.size / SLI4_CQE_BYTES;
drivers/scsi/elx/libefc_sli/sli4.c
845
num_pages_cq = sli_page_count(qs[0]->dma.size, page_bytes);
drivers/scsi/elx/libefc_sli/sli4.c
903
for (p = 0, addr = qs[i]->dma.phys; p < num_pages_cq;
drivers/scsi/elx/libefc_sli/sli4.c
917
sli_cq_alloc_set(struct sli4 *sli4, struct sli4_queue *qs[],
drivers/scsi/elx/libefc_sli/sli4.c
927
if (__sli_queue_init(sli4, qs[i], SLI4_QTYPE_CQ, SLI4_CQE_BYTES,
drivers/scsi/elx/libefc_sli/sli4.c
932
if (sli_cmd_cq_set_create(sli4, qs, num_cqs, eqs, &dma))
drivers/scsi/elx/libefc_sli/sli4.c
957
qs[i]->id = le16_to_cpu(res->q_id) + i;
drivers/scsi/elx/libefc_sli/sli4.c
958
qs[i]->db_regaddr = db_regaddr;
drivers/scsi/elx/libefc_sli/sli4.c
967
__sli_queue_destroy(sli4, qs[i]);
drivers/scsi/elx/libefc_sli/sli4.h
3992
sli_cq_alloc_set(struct sli4 *sli4, struct sli4_queue *qs[], u32 num_cqs,
drivers/scsi/lpfc/lpfc_init.c
10790
lpfc_sli4_release_queues(struct lpfc_queue ***qs, int max)
drivers/scsi/lpfc/lpfc_init.c
10794
if (*qs == NULL)
drivers/scsi/lpfc/lpfc_init.c
10798
__lpfc_sli4_release_queue(&(*qs)[idx]);
drivers/scsi/lpfc/lpfc_init.c
10800
kfree(*qs);
drivers/scsi/lpfc/lpfc_init.c
10801
*qs = NULL;
drivers/soc/qcom/socinfo.c
862
struct qcom_socinfo *qs;
drivers/soc/qcom/socinfo.c
873
qs = devm_kzalloc(&pdev->dev, sizeof(*qs), GFP_KERNEL);
drivers/soc/qcom/socinfo.c
874
if (!qs)
drivers/soc/qcom/socinfo.c
877
qs->attr.family = "Snapdragon";
drivers/soc/qcom/socinfo.c
878
qs->attr.machine = socinfo_machine(&pdev->dev,
drivers/soc/qcom/socinfo.c
880
qs->attr.soc_id = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u",
drivers/soc/qcom/socinfo.c
882
qs->attr.revision = devm_kasprintf(&pdev->dev, GFP_KERNEL, "%u.%u",
drivers/soc/qcom/socinfo.c
885
if (!qs->attr.soc_id || !qs->attr.revision)
drivers/soc/qcom/socinfo.c
889
qs->attr.serial_number = devm_kasprintf(&pdev->dev, GFP_KERNEL,
drivers/soc/qcom/socinfo.c
892
if (!qs->attr.serial_number)
drivers/soc/qcom/socinfo.c
896
qs->soc_dev = soc_device_register(&qs->attr);
drivers/soc/qcom/socinfo.c
897
if (IS_ERR(qs->soc_dev))
drivers/soc/qcom/socinfo.c
898
return PTR_ERR(qs->soc_dev);
drivers/soc/qcom/socinfo.c
900
socinfo_debugfs_init(qs, info, item_size);
drivers/soc/qcom/socinfo.c
905
platform_set_drvdata(pdev, qs);
drivers/soc/qcom/socinfo.c
912
struct qcom_socinfo *qs = platform_get_drvdata(pdev);
drivers/soc/qcom/socinfo.c
914
soc_device_unregister(qs->soc_dev);
drivers/soc/qcom/socinfo.c
916
socinfo_debugfs_exit(qs);
fs/ocfs2/cluster/quorum.c
100
lowest_reachable = test_bit(lowest_hb, qs->qs_conn_bm);
fs/ocfs2/cluster/quorum.c
103
"lowest: %d (%sreachable)\n", qs->qs_heartbeating,
fs/ocfs2/cluster/quorum.c
104
qs->qs_connected, lowest_hb, lowest_reachable ? "" : "un");
fs/ocfs2/cluster/quorum.c
106
if (!test_bit(o2nm_this_node(), qs->qs_hb_bm) ||
fs/ocfs2/cluster/quorum.c
107
qs->qs_heartbeating == 1)
fs/ocfs2/cluster/quorum.c
110
if (qs->qs_heartbeating & 1) {
fs/ocfs2/cluster/quorum.c
113
quorum = (qs->qs_heartbeating + 1)/2;
fs/ocfs2/cluster/quorum.c
114
if (qs->qs_connected < quorum) {
fs/ocfs2/cluster/quorum.c
118
qs->qs_connected, quorum,
fs/ocfs2/cluster/quorum.c
119
qs->qs_heartbeating);
fs/ocfs2/cluster/quorum.c
127
quorum = qs->qs_heartbeating / 2;
fs/ocfs2/cluster/quorum.c
128
if (qs->qs_connected < quorum) {
fs/ocfs2/cluster/quorum.c
132
qs->qs_connected, quorum,
fs/ocfs2/cluster/quorum.c
133
qs->qs_heartbeating);
fs/ocfs2/cluster/quorum.c
136
else if ((qs->qs_connected == quorum) &&
fs/ocfs2/cluster/quorum.c
141
"node %u\n", quorum, qs->qs_heartbeating,
fs/ocfs2/cluster/quorum.c
149
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
154
qs->qs_heartbeating, qs->qs_connected, lowest_hb,
fs/ocfs2/cluster/quorum.c
156
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
162
static void o2quo_set_hold(struct o2quo_state *qs, u8 node)
fs/ocfs2/cluster/quorum.c
164
assert_spin_locked(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
166
if (!test_and_set_bit(node, qs->qs_hold_bm)) {
fs/ocfs2/cluster/quorum.c
167
qs->qs_holds++;
fs/ocfs2/cluster/quorum.c
168
mlog_bug_on_msg(qs->qs_holds == O2NM_MAX_NODES,
fs/ocfs2/cluster/quorum.c
170
mlog(0, "node %u, %d total\n", node, qs->qs_holds);
fs/ocfs2/cluster/quorum.c
174
static void o2quo_clear_hold(struct o2quo_state *qs, u8 node)
fs/ocfs2/cluster/quorum.c
176
assert_spin_locked(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
178
if (test_and_clear_bit(node, qs->qs_hold_bm)) {
fs/ocfs2/cluster/quorum.c
179
mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1);
fs/ocfs2/cluster/quorum.c
180
if (--qs->qs_holds == 0) {
fs/ocfs2/cluster/quorum.c
181
if (qs->qs_pending) {
fs/ocfs2/cluster/quorum.c
182
qs->qs_pending = 0;
fs/ocfs2/cluster/quorum.c
183
schedule_work(&qs->qs_work);
fs/ocfs2/cluster/quorum.c
186
mlog_bug_on_msg(qs->qs_holds < 0, "node %u, holds %d\n",
fs/ocfs2/cluster/quorum.c
187
node, qs->qs_holds);
fs/ocfs2/cluster/quorum.c
197
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
199
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
201
qs->qs_heartbeating++;
fs/ocfs2/cluster/quorum.c
202
mlog_bug_on_msg(qs->qs_heartbeating == O2NM_MAX_NODES,
fs/ocfs2/cluster/quorum.c
204
mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node);
fs/ocfs2/cluster/quorum.c
205
set_bit(node, qs->qs_hb_bm);
fs/ocfs2/cluster/quorum.c
207
mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
fs/ocfs2/cluster/quorum.c
209
if (!test_bit(node, qs->qs_conn_bm))
fs/ocfs2/cluster/quorum.c
210
o2quo_set_hold(qs, node);
fs/ocfs2/cluster/quorum.c
212
o2quo_clear_hold(qs, node);
fs/ocfs2/cluster/quorum.c
214
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
221
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
223
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
225
qs->qs_heartbeating--;
fs/ocfs2/cluster/quorum.c
226
mlog_bug_on_msg(qs->qs_heartbeating < 0,
fs/ocfs2/cluster/quorum.c
228
node, qs->qs_heartbeating);
fs/ocfs2/cluster/quorum.c
229
mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node);
fs/ocfs2/cluster/quorum.c
230
clear_bit(node, qs->qs_hb_bm);
fs/ocfs2/cluster/quorum.c
232
mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
fs/ocfs2/cluster/quorum.c
234
o2quo_clear_hold(qs, node);
fs/ocfs2/cluster/quorum.c
236
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
246
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
248
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
252
qs->qs_pending = 1;
fs/ocfs2/cluster/quorum.c
253
o2quo_clear_hold(qs, node);
fs/ocfs2/cluster/quorum.c
255
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
265
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
267
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
269
qs->qs_connected++;
fs/ocfs2/cluster/quorum.c
270
mlog_bug_on_msg(qs->qs_connected == O2NM_MAX_NODES,
fs/ocfs2/cluster/quorum.c
272
mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node);
fs/ocfs2/cluster/quorum.c
273
set_bit(node, qs->qs_conn_bm);
fs/ocfs2/cluster/quorum.c
275
mlog(0, "node %u, %d total\n", node, qs->qs_connected);
fs/ocfs2/cluster/quorum.c
277
if (!test_bit(node, qs->qs_hb_bm))
fs/ocfs2/cluster/quorum.c
278
o2quo_set_hold(qs, node);
fs/ocfs2/cluster/quorum.c
280
o2quo_clear_hold(qs, node);
fs/ocfs2/cluster/quorum.c
282
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
291
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
293
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
295
if (test_bit(node, qs->qs_conn_bm)) {
fs/ocfs2/cluster/quorum.c
296
qs->qs_connected--;
fs/ocfs2/cluster/quorum.c
297
mlog_bug_on_msg(qs->qs_connected < 0,
fs/ocfs2/cluster/quorum.c
299
node, qs->qs_connected);
fs/ocfs2/cluster/quorum.c
301
clear_bit(node, qs->qs_conn_bm);
fs/ocfs2/cluster/quorum.c
303
if (test_bit(node, qs->qs_hb_bm))
fs/ocfs2/cluster/quorum.c
304
o2quo_set_hold(qs, node);
fs/ocfs2/cluster/quorum.c
307
mlog(0, "node %u, %d total\n", node, qs->qs_connected);
fs/ocfs2/cluster/quorum.c
310
spin_unlock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
315
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
317
spin_lock_init(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
318
INIT_WORK(&qs->qs_work, o2quo_make_decision);
fs/ocfs2/cluster/quorum.c
323
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
325
flush_work(&qs->qs_work);
fs/ocfs2/cluster/quorum.c
94
struct o2quo_state *qs = &o2quo_state;
fs/ocfs2/cluster/quorum.c
96
spin_lock_bh(&qs->qs_lock);
fs/ocfs2/cluster/quorum.c
98
lowest_hb = find_first_bit(qs->qs_hb_bm, O2NM_MAX_NODES);
fs/qnx4/inode.c
197
struct qnx4_sb_info *qs;
fs/qnx4/inode.c
200
qs = kzalloc_obj(struct qnx4_sb_info);
fs/qnx4/inode.c
201
if (!qs)
fs/qnx4/inode.c
203
s->s_fs_info = qs;
fs/qnx4/inode.c
259
struct qnx4_sb_info *qs = qnx4_sb(sb);
fs/qnx4/inode.c
261
if (qs) {
fs/qnx4/inode.c
262
kfree(qs->BitMap);
fs/qnx4/inode.c
263
kfree(qs);
fs/qnx4/inode.c
47
struct qnx4_sb_info *qs;
fs/qnx4/inode.c
50
qs = qnx4_sb(sb);
fs/qnx4/inode.c
51
qs->Version = QNX4_VERSION;
fs/qnx6/inode.c
299
struct qnx6_sb_info *qs;
fs/qnx6/inode.c
305
qs = kzalloc_obj(struct qnx6_sb_info);
fs/qnx6/inode.c
306
if (!qs)
fs/qnx6/inode.c
308
s->s_fs_info = qs;
fs/qnx6/inode.c
309
qs->s_mount_opt = ctx->s_mount_opts;
fs/qnx6/inode.c
317
if (qs->s_mount_opt == QNX6_MOUNT_MMI_FS) {
fs/qnx6/inode.c
469
kfree(qs);
fs/qnx6/inode.c
476
struct qnx6_sb_info *qs = QNX6_SB(sb);
fs/qnx6/inode.c
477
brelse(qs->sb_buf);
fs/qnx6/inode.c
478
iput(qs->longfile);
fs/qnx6/inode.c
479
iput(qs->inodes);
fs/qnx6/inode.c
480
kfree(qs);
include/asm-generic/vmlinux.lds.h
121
.text.sta[_0-9A-Za-qs-z]* .text.sta .text.sta.* \
kernel/bpf/queue_stack_maps.c
102
if (raw_res_spin_lock_irqsave(&qs->lock, flags))
kernel/bpf/queue_stack_maps.c
105
if (queue_stack_map_is_empty(qs)) {
kernel/bpf/queue_stack_maps.c
106
memset(value, 0, qs->map.value_size);
kernel/bpf/queue_stack_maps.c
111
ptr = &qs->elements[qs->tail * qs->map.value_size];
kernel/bpf/queue_stack_maps.c
112
memcpy(value, ptr, qs->map.value_size);
kernel/bpf/queue_stack_maps.c
115
if (unlikely(++qs->tail >= qs->size))
kernel/bpf/queue_stack_maps.c
116
qs->tail = 0;
kernel/bpf/queue_stack_maps.c
120
raw_res_spin_unlock_irqrestore(&qs->lock, flags);
kernel/bpf/queue_stack_maps.c
127
struct bpf_queue_stack *qs = bpf_queue_stack(map);
kernel/bpf/queue_stack_maps.c
133
if (raw_res_spin_lock_irqsave(&qs->lock, flags))
kernel/bpf/queue_stack_maps.c
136
if (queue_stack_map_is_empty(qs)) {
kernel/bpf/queue_stack_maps.c
137
memset(value, 0, qs->map.value_size);
kernel/bpf/queue_stack_maps.c
142
index = qs->head - 1;
kernel/bpf/queue_stack_maps.c
143
if (unlikely(index >= qs->size))
kernel/bpf/queue_stack_maps.c
144
index = qs->size - 1;
kernel/bpf/queue_stack_maps.c
146
ptr = &qs->elements[index * qs->map.value_size];
kernel/bpf/queue_stack_maps.c
147
memcpy(value, ptr, qs->map.value_size);
kernel/bpf/queue_stack_maps.c
150
qs->head = index;
kernel/bpf/queue_stack_maps.c
153
raw_res_spin_unlock_irqrestore(&qs->lock, flags);
kernel/bpf/queue_stack_maps.c
185
struct bpf_queue_stack *qs = bpf_queue_stack(map);
kernel/bpf/queue_stack_maps.c
199
if (raw_res_spin_lock_irqsave(&qs->lock, irq_flags))
kernel/bpf/queue_stack_maps.c
202
if (queue_stack_map_is_full(qs)) {
kernel/bpf/queue_stack_maps.c
208
if (unlikely(++qs->tail >= qs->size))
kernel/bpf/queue_stack_maps.c
209
qs->tail = 0;
kernel/bpf/queue_stack_maps.c
212
dst = &qs->elements[qs->head * qs->map.value_size];
kernel/bpf/queue_stack_maps.c
213
memcpy(dst, value, qs->map.value_size);
kernel/bpf/queue_stack_maps.c
215
if (unlikely(++qs->head >= qs->size))
kernel/bpf/queue_stack_maps.c
216
qs->head = 0;
kernel/bpf/queue_stack_maps.c
219
raw_res_spin_unlock_irqrestore(&qs->lock, irq_flags);
kernel/bpf/queue_stack_maps.c
31
static bool queue_stack_map_is_empty(struct bpf_queue_stack *qs)
kernel/bpf/queue_stack_maps.c
33
return qs->head == qs->tail;
kernel/bpf/queue_stack_maps.c
36
static bool queue_stack_map_is_full(struct bpf_queue_stack *qs)
kernel/bpf/queue_stack_maps.c
38
u32 head = qs->head + 1;
kernel/bpf/queue_stack_maps.c
40
if (unlikely(head >= qs->size))
kernel/bpf/queue_stack_maps.c
43
return head == qs->tail;
kernel/bpf/queue_stack_maps.c
68
struct bpf_queue_stack *qs;
kernel/bpf/queue_stack_maps.c
72
queue_size = sizeof(*qs) + size * attr->value_size;
kernel/bpf/queue_stack_maps.c
74
qs = bpf_map_area_alloc(queue_size, numa_node);
kernel/bpf/queue_stack_maps.c
75
if (!qs)
kernel/bpf/queue_stack_maps.c
78
bpf_map_init_from_attr(&qs->map, attr);
kernel/bpf/queue_stack_maps.c
80
qs->size = size;
kernel/bpf/queue_stack_maps.c
82
raw_res_spin_lock_init(&qs->lock);
kernel/bpf/queue_stack_maps.c
84
return &qs->map;
kernel/bpf/queue_stack_maps.c
90
struct bpf_queue_stack *qs = bpf_queue_stack(map);
kernel/bpf/queue_stack_maps.c
92
bpf_map_area_free(qs);
kernel/bpf/queue_stack_maps.c
97
struct bpf_queue_stack *qs = bpf_queue_stack(map);
net/sched/sch_cake.c
3120
struct gnet_stats_queue qs = { 0 };
net/sched/sch_cake.c
3135
qs.qlen++;
net/sched/sch_cake.c
3140
qs.backlog = b->backlogs[idx % CAKE_QUEUES];
net/sched/sch_cake.c
3141
qs.drops = flow->dropped;
net/sched/sch_cake.c
3143
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
net/sched/sch_fq_codel.c
649
struct gnet_stats_queue qs = { 0 };
net/sched/sch_fq_codel.c
676
qs.qlen++;
net/sched/sch_fq_codel.c
681
qs.backlog = q->backlogs[idx];
net/sched/sch_fq_codel.c
682
qs.drops = 0;
net/sched/sch_fq_codel.c
684
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
net/sched/sch_htb.c
1326
struct gnet_stats_queue qs = {
net/sched/sch_htb.c
1333
qdisc_qstats_qlen_backlog(cl->leaf.q, &qlen, &qs.backlog);
net/sched/sch_htb.c
1356
gnet_stats_copy_queue(d, NULL, &qs, qlen) < 0)
net/sched/sch_sfq.c
909
struct gnet_stats_queue qs = { 0 };
net/sched/sch_sfq.c
916
qs.qlen = slot->qlen;
net/sched/sch_sfq.c
917
qs.backlog = slot->backlog;
net/sched/sch_sfq.c
919
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
sound/pci/rme9652/hdspm.c
6133
levels->speed = qs;
tools/net/ynl/ynltool/qstats.c
100
jsonw_uint_field(json_wtr, "csum-none", qs->tx_csum_none);
tools/net/ynl/ynltool/qstats.c
101
if (qs->_present.tx_needs_csum)
tools/net/ynl/ynltool/qstats.c
102
jsonw_uint_field(json_wtr, "needs-csum", qs->tx_needs_csum);
tools/net/ynl/ynltool/qstats.c
103
if (qs->_present.tx_hw_gso_packets)
tools/net/ynl/ynltool/qstats.c
104
jsonw_uint_field(json_wtr, "hw-gso-packets", qs->tx_hw_gso_packets);
tools/net/ynl/ynltool/qstats.c
105
if (qs->_present.tx_hw_gso_bytes)
tools/net/ynl/ynltool/qstats.c
106
jsonw_uint_field(json_wtr, "hw-gso-bytes", qs->tx_hw_gso_bytes);
tools/net/ynl/ynltool/qstats.c
107
if (qs->_present.tx_hw_gso_wire_packets)
tools/net/ynl/ynltool/qstats.c
108
jsonw_uint_field(json_wtr, "hw-gso-wire-packets", qs->tx_hw_gso_wire_packets);
tools/net/ynl/ynltool/qstats.c
109
if (qs->_present.tx_hw_gso_wire_bytes)
tools/net/ynl/ynltool/qstats.c
110
jsonw_uint_field(json_wtr, "hw-gso-wire-bytes", qs->tx_hw_gso_wire_bytes);
tools/net/ynl/ynltool/qstats.c
111
if (qs->_present.tx_stop)
tools/net/ynl/ynltool/qstats.c
112
jsonw_uint_field(json_wtr, "stop", qs->tx_stop);
tools/net/ynl/ynltool/qstats.c
113
if (qs->_present.tx_wake)
tools/net/ynl/ynltool/qstats.c
114
jsonw_uint_field(json_wtr, "wake", qs->tx_wake);
tools/net/ynl/ynltool/qstats.c
151
ynl_dump_foreach(qstats, qs) {
tools/net/ynl/ynltool/qstats.c
156
name = if_indextoname(qs->ifindex, ifname);
tools/net/ynl/ynltool/qstats.c
160
printf("ifindex:%u", qs->ifindex);
tools/net/ynl/ynltool/qstats.c
162
if (qs->_present.queue_type && qs->_present.queue_id)
tools/net/ynl/ynltool/qstats.c
164
netdev_queue_type_str(qs->queue_type),
tools/net/ynl/ynltool/qstats.c
165
qs->queue_id);
tools/net/ynl/ynltool/qstats.c
172
print_one(qs->_present.rx_packets, "rx-packets", qs->rx_packets, &n);
tools/net/ynl/ynltool/qstats.c
173
print_one(qs->_present.rx_bytes, "rx-bytes", qs->rx_bytes, &n);
tools/net/ynl/ynltool/qstats.c
174
print_one(qs->_present.tx_packets, "tx-packets", qs->tx_packets, &n);
tools/net/ynl/ynltool/qstats.c
175
print_one(qs->_present.tx_bytes, "tx-bytes", qs->tx_bytes, &n);
tools/net/ynl/ynltool/qstats.c
178
print_one(qs->_present.rx_alloc_fail, "rx-alloc-fail",
tools/net/ynl/ynltool/qstats.c
179
qs->rx_alloc_fail, &n);
tools/net/ynl/ynltool/qstats.c
180
print_one(qs->_present.rx_hw_drops, "rx-hw-drops",
tools/net/ynl/ynltool/qstats.c
181
qs->rx_hw_drops, &n);
tools/net/ynl/ynltool/qstats.c
182
print_one(qs->_present.rx_hw_drop_overruns, "rx-hw-drop-overruns",
tools/net/ynl/ynltool/qstats.c
183
qs->rx_hw_drop_overruns, &n);
tools/net/ynl/ynltool/qstats.c
184
print_one(qs->_present.rx_hw_drop_ratelimits, "rx-hw-drop-ratelimits",
tools/net/ynl/ynltool/qstats.c
185
qs->rx_hw_drop_ratelimits, &n);
tools/net/ynl/ynltool/qstats.c
188
print_one(qs->_present.rx_csum_complete, "rx-csum-complete",
tools/net/ynl/ynltool/qstats.c
189
qs->rx_csum_complete, &n);
tools/net/ynl/ynltool/qstats.c
190
print_one(qs->_present.rx_csum_unnecessary, "rx-csum-unnecessary",
tools/net/ynl/ynltool/qstats.c
191
qs->rx_csum_unnecessary, &n);
tools/net/ynl/ynltool/qstats.c
192
print_one(qs->_present.rx_csum_none, "rx-csum-none",
tools/net/ynl/ynltool/qstats.c
193
qs->rx_csum_none, &n);
tools/net/ynl/ynltool/qstats.c
194
print_one(qs->_present.rx_csum_bad, "rx-csum-bad",
tools/net/ynl/ynltool/qstats.c
195
qs->rx_csum_bad, &n);
tools/net/ynl/ynltool/qstats.c
198
print_one(qs->_present.rx_hw_gro_packets, "rx-hw-gro-packets",
tools/net/ynl/ynltool/qstats.c
199
qs->rx_hw_gro_packets, &n);
tools/net/ynl/ynltool/qstats.c
200
print_one(qs->_present.rx_hw_gro_bytes, "rx-hw-gro-bytes",
tools/net/ynl/ynltool/qstats.c
201
qs->rx_hw_gro_bytes, &n);
tools/net/ynl/ynltool/qstats.c
202
print_one(qs->_present.rx_hw_gro_wire_packets, "rx-hw-gro-wire-packets",
tools/net/ynl/ynltool/qstats.c
203
qs->rx_hw_gro_wire_packets, &n);
tools/net/ynl/ynltool/qstats.c
204
print_one(qs->_present.rx_hw_gro_wire_bytes, "rx-hw-gro-wire-bytes",
tools/net/ynl/ynltool/qstats.c
205
qs->rx_hw_gro_wire_bytes, &n);
tools/net/ynl/ynltool/qstats.c
208
print_one(qs->_present.tx_hw_drops, "tx-hw-drops",
tools/net/ynl/ynltool/qstats.c
209
qs->tx_hw_drops, &n);
tools/net/ynl/ynltool/qstats.c
210
print_one(qs->_present.tx_hw_drop_errors, "tx-hw-drop-errors",
tools/net/ynl/ynltool/qstats.c
211
qs->tx_hw_drop_errors, &n);
tools/net/ynl/ynltool/qstats.c
212
print_one(qs->_present.tx_hw_drop_ratelimits, "tx-hw-drop-ratelimits",
tools/net/ynl/ynltool/qstats.c
213
qs->tx_hw_drop_ratelimits, &n);
tools/net/ynl/ynltool/qstats.c
216
print_one(qs->_present.tx_csum_none, "tx-csum-none",
tools/net/ynl/ynltool/qstats.c
217
qs->tx_csum_none, &n);
tools/net/ynl/ynltool/qstats.c
218
print_one(qs->_present.tx_needs_csum, "tx-needs-csum",
tools/net/ynl/ynltool/qstats.c
219
qs->tx_needs_csum, &n);
tools/net/ynl/ynltool/qstats.c
222
print_one(qs->_present.tx_hw_gso_packets, "tx-hw-gso-packets",
tools/net/ynl/ynltool/qstats.c
223
qs->tx_hw_gso_packets, &n);
tools/net/ynl/ynltool/qstats.c
224
print_one(qs->_present.tx_hw_gso_bytes, "tx-hw-gso-bytes",
tools/net/ynl/ynltool/qstats.c
225
qs->tx_hw_gso_bytes, &n);
tools/net/ynl/ynltool/qstats.c
226
print_one(qs->_present.tx_hw_gso_wire_packets, "tx-hw-gso-wire-packets",
tools/net/ynl/ynltool/qstats.c
227
qs->tx_hw_gso_wire_packets, &n);
tools/net/ynl/ynltool/qstats.c
228
print_one(qs->_present.tx_hw_gso_wire_bytes, "tx-hw-gso-wire-bytes",
tools/net/ynl/ynltool/qstats.c
229
qs->tx_hw_gso_wire_bytes, &n);
tools/net/ynl/ynltool/qstats.c
232
print_one(qs->_present.tx_stop, "tx-stop", qs->tx_stop, &n);
tools/net/ynl/ynltool/qstats.c
233
print_one(qs->_present.tx_wake, "tx-wake", qs->tx_wake, &n);
tools/net/ynl/ynltool/qstats.c
31
ynl_dump_foreach(qstats, qs) {
tools/net/ynl/ynltool/qstats.c
37
name = if_indextoname(qs->ifindex, ifname);
tools/net/ynl/ynltool/qstats.c
40
jsonw_uint_field(json_wtr, "ifindex", qs->ifindex);
tools/net/ynl/ynltool/qstats.c
42
if (qs->_present.queue_type)
tools/net/ynl/ynltool/qstats.c
436
ynl_dump_foreach(qstats, qs)
tools/net/ynl/ynltool/qstats.c
44
netdev_queue_type_str(qs->queue_type));
tools/net/ynl/ynltool/qstats.c
45
if (qs->_present.queue_id)
tools/net/ynl/ynltool/qstats.c
455
ynl_dump_foreach(qstats, qs)
tools/net/ynl/ynltool/qstats.c
456
sorted[i++] = qs;
tools/net/ynl/ynltool/qstats.c
46
jsonw_uint_field(json_wtr, "queue-id", qs->queue_id);
tools/net/ynl/ynltool/qstats.c
48
if (qs->_present.rx_packets || qs->_present.rx_bytes ||
tools/net/ynl/ynltool/qstats.c
49
qs->_present.rx_alloc_fail || qs->_present.rx_hw_drops ||
tools/net/ynl/ynltool/qstats.c
50
qs->_present.rx_csum_complete || qs->_present.rx_hw_gro_packets) {
tools/net/ynl/ynltool/qstats.c
53
if (qs->_present.rx_packets)
tools/net/ynl/ynltool/qstats.c
54
jsonw_uint_field(json_wtr, "packets", qs->rx_packets);
tools/net/ynl/ynltool/qstats.c
55
if (qs->_present.rx_bytes)
tools/net/ynl/ynltool/qstats.c
56
jsonw_uint_field(json_wtr, "bytes", qs->rx_bytes);
tools/net/ynl/ynltool/qstats.c
57
if (qs->_present.rx_alloc_fail)
tools/net/ynl/ynltool/qstats.c
58
jsonw_uint_field(json_wtr, "alloc-fail", qs->rx_alloc_fail);
tools/net/ynl/ynltool/qstats.c
587
ynl_dump_foreach(qstats, qs) {
tools/net/ynl/ynltool/qstats.c
59
if (qs->_present.rx_hw_drops)
tools/net/ynl/ynltool/qstats.c
592
if (!qs->_present.rx_packets ||
tools/net/ynl/ynltool/qstats.c
593
!qs->_present.rx_hw_gro_packets ||
tools/net/ynl/ynltool/qstats.c
594
!qs->_present.rx_hw_gro_wire_packets)
tools/net/ynl/ynltool/qstats.c
597
if (!qs->rx_packets)
tools/net/ynl/ynltool/qstats.c
60
jsonw_uint_field(json_wtr, "hw-drops", qs->rx_hw_drops);
tools/net/ynl/ynltool/qstats.c
601
savings = (double)(qs->rx_hw_gro_wire_packets -
tools/net/ynl/ynltool/qstats.c
602
qs->rx_hw_gro_packets) /
tools/net/ynl/ynltool/qstats.c
603
qs->rx_packets * 100.0;
tools/net/ynl/ynltool/qstats.c
605
name = if_indextoname(qs->ifindex, ifname);
tools/net/ynl/ynltool/qstats.c
609
jsonw_uint_field(json_wtr, "ifindex", qs->ifindex);
tools/net/ynl/ynltool/qstats.c
61
if (qs->_present.rx_hw_drop_overruns)
tools/net/ynl/ynltool/qstats.c
618
printf("ifindex:%u", qs->ifindex);
tools/net/ynl/ynltool/qstats.c
62
jsonw_uint_field(json_wtr, "hw-drop-overruns", qs->rx_hw_drop_overruns);
tools/net/ynl/ynltool/qstats.c
63
if (qs->_present.rx_hw_drop_ratelimits)
tools/net/ynl/ynltool/qstats.c
64
jsonw_uint_field(json_wtr, "hw-drop-ratelimits", qs->rx_hw_drop_ratelimits);
tools/net/ynl/ynltool/qstats.c
65
if (qs->_present.rx_csum_complete)
tools/net/ynl/ynltool/qstats.c
66
jsonw_uint_field(json_wtr, "csum-complete", qs->rx_csum_complete);
tools/net/ynl/ynltool/qstats.c
67
if (qs->_present.rx_csum_unnecessary)
tools/net/ynl/ynltool/qstats.c
68
jsonw_uint_field(json_wtr, "csum-unnecessary", qs->rx_csum_unnecessary);
tools/net/ynl/ynltool/qstats.c
69
if (qs->_present.rx_csum_none)
tools/net/ynl/ynltool/qstats.c
70
jsonw_uint_field(json_wtr, "csum-none", qs->rx_csum_none);
tools/net/ynl/ynltool/qstats.c
71
if (qs->_present.rx_csum_bad)
tools/net/ynl/ynltool/qstats.c
72
jsonw_uint_field(json_wtr, "csum-bad", qs->rx_csum_bad);
tools/net/ynl/ynltool/qstats.c
73
if (qs->_present.rx_hw_gro_packets)
tools/net/ynl/ynltool/qstats.c
74
jsonw_uint_field(json_wtr, "hw-gro-packets", qs->rx_hw_gro_packets);
tools/net/ynl/ynltool/qstats.c
75
if (qs->_present.rx_hw_gro_bytes)
tools/net/ynl/ynltool/qstats.c
76
jsonw_uint_field(json_wtr, "hw-gro-bytes", qs->rx_hw_gro_bytes);
tools/net/ynl/ynltool/qstats.c
77
if (qs->_present.rx_hw_gro_wire_packets)
tools/net/ynl/ynltool/qstats.c
78
jsonw_uint_field(json_wtr, "hw-gro-wire-packets", qs->rx_hw_gro_wire_packets);
tools/net/ynl/ynltool/qstats.c
79
if (qs->_present.rx_hw_gro_wire_bytes)
tools/net/ynl/ynltool/qstats.c
80
jsonw_uint_field(json_wtr, "hw-gro-wire-bytes", qs->rx_hw_gro_wire_bytes);
tools/net/ynl/ynltool/qstats.c
84
if (qs->_present.tx_packets || qs->_present.tx_bytes ||
tools/net/ynl/ynltool/qstats.c
85
qs->_present.tx_hw_drops || qs->_present.tx_csum_none ||
tools/net/ynl/ynltool/qstats.c
86
qs->_present.tx_hw_gso_packets) {
tools/net/ynl/ynltool/qstats.c
89
if (qs->_present.tx_packets)
tools/net/ynl/ynltool/qstats.c
90
jsonw_uint_field(json_wtr, "packets", qs->tx_packets);
tools/net/ynl/ynltool/qstats.c
91
if (qs->_present.tx_bytes)
tools/net/ynl/ynltool/qstats.c
92
jsonw_uint_field(json_wtr, "bytes", qs->tx_bytes);
tools/net/ynl/ynltool/qstats.c
93
if (qs->_present.tx_hw_drops)
tools/net/ynl/ynltool/qstats.c
94
jsonw_uint_field(json_wtr, "hw-drops", qs->tx_hw_drops);
tools/net/ynl/ynltool/qstats.c
95
if (qs->_present.tx_hw_drop_errors)
tools/net/ynl/ynltool/qstats.c
96
jsonw_uint_field(json_wtr, "hw-drop-errors", qs->tx_hw_drop_errors);
tools/net/ynl/ynltool/qstats.c
97
if (qs->_present.tx_hw_drop_ratelimits)
tools/net/ynl/ynltool/qstats.c
98
jsonw_uint_field(json_wtr, "hw-drop-ratelimits", qs->tx_hw_drop_ratelimits);
tools/net/ynl/ynltool/qstats.c
99
if (qs->_present.tx_csum_none)