Symbol: qlen
crypto/algapi.c
943
queue->qlen = 0;
crypto/algapi.c
953
if (unlikely(queue->qlen >= queue->max_qlen)) {
crypto/algapi.c
963
queue->qlen++;
crypto/algapi.c
974
if (unlikely(queue->qlen >= queue->max_qlen))
crypto/algapi.c
977
queue->qlen++;
crypto/algapi.c
986
if (unlikely(!queue->qlen))
crypto/algapi.c
989
queue->qlen--;
crypto/cryptd.c
127
BUG_ON(cpu_queue->queue.qlen);
crypto/cryptd.c
189
if (cpu_queue->queue.qlen)
crypto/crypto_engine.c
435
bool rt, int qlen)
crypto/crypto_engine.c
457
crypto_init_queue(&engine->queue, qlen);
drivers/bluetooth/hci_bcsp.c
303
if (bcsp->unack.qlen < BCSP_TXWINSIZE) {
drivers/bluetooth/hci_bcsp.c
699
BT_DBG("hu %p retransmitting %u pkts", hu, bcsp->unack.qlen);
drivers/bluetooth/hci_h5.c
180
BT_DBG("hu %p retransmitting %u pkts", hu, h5->unack.qlen);
drivers/bluetooth/hci_h5.c
807
if (h5->unack.qlen >= h5->tx_win)
drivers/crypto/cavium/cpt/cptpf.h
39
u32 qlen;
drivers/crypto/cavium/cpt/cptpf_mbox.c
117
vfx->qlen = mbx.data;
drivers/crypto/cavium/cpt/cptpf_mbox.c
118
cpt_cfg_qlen_for_vf(cpt, vf, vfx->qlen);
drivers/crypto/cavium/cpt/cptvf.h
86
u32 qlen; /* Queue length */
drivers/crypto/cavium/cpt/cptvf_main.c
103
static int alloc_pending_queues(struct pending_qinfo *pqinfo, u32 qlen,
drivers/crypto/cavium/cpt/cptvf_main.c
111
pqinfo->qlen = qlen;
drivers/crypto/cavium/cpt/cptvf_main.c
114
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/cavium/cpt/cptvf_main.c
136
static int init_pending_queues(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues)
drivers/crypto/cavium/cpt/cptvf_main.c
144
ret = alloc_pending_queues(&cptvf->pqinfo, qlen, nr_queues);
drivers/crypto/cavium/cpt/cptvf_main.c
203
u32 qlen)
drivers/crypto/cavium/cpt/cptvf_main.c
213
cptvf->qsize = min(qlen, cqinfo->qchunksize) *
drivers/crypto/cavium/cpt/cptvf_main.c
216
q_size = qlen * cqinfo->cmd_size;
drivers/crypto/cavium/cpt/cptvf_main.c
277
static int init_command_queues(struct cpt_vf *cptvf, u32 qlen)
drivers/crypto/cavium/cpt/cptvf_main.c
284
qlen);
drivers/crypto/cavium/cpt/cptvf_main.c
313
static int cptvf_sw_init(struct cpt_vf *cptvf, u32 qlen, u32 nr_queues)
drivers/crypto/cavium/cpt/cptvf_main.c
324
ret = init_command_queues(cptvf, qlen);
drivers/crypto/cavium/cpt/cptvf_main.c
331
ret = init_pending_queues(cptvf, qlen, nr_queues);
drivers/crypto/cavium/cpt/cptvf_main.c
99
pqinfo->qlen = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
16
int qlen)
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
27
if (unlikely(q->rear == qlen))
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
40
if (unlikely(queue->front == pqinfo->qlen))
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
485
pentry = get_free_pending_entry(pqueue, cptvf->pqinfo.qlen);
drivers/crypto/cavium/nitrox/nitrox_dev.h
249
u16 qlen;
drivers/crypto/cavium/nitrox/nitrox_hal.c
143
pkt_in_rsize.s.rsize = ndev->qlen;
drivers/crypto/cavium/nitrox/nitrox_hal.c
385
qsize.host_queue_size = ndev->qlen;
drivers/crypto/cavium/nitrox/nitrox_lib.c
29
cmdq->qsize = (ndev->qlen * cmdq->instr_size) + align_bytes;
drivers/crypto/cavium/nitrox/nitrox_main.c
47
static unsigned int qlen = DEFAULT_CMD_QLEN;
drivers/crypto/cavium/nitrox/nitrox_main.c
473
ndev->qlen = qlen;
drivers/crypto/cavium/nitrox/nitrox_main.c
48
module_param(qlen, uint, 0644);
drivers/crypto/cavium/nitrox/nitrox_main.c
49
MODULE_PARM_DESC(qlen, "Command queue length - default 2048");
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
266
static inline bool cmdq_full(struct nitrox_cmdq *cmdq, int qlen)
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
268
if (atomic_inc_return(&cmdq->pending_count) > qlen) {
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
310
cmdq->write_idx = incr_index(idx, 1, ndev->qlen);
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
331
if (unlikely(cmdq_full(cmdq, ndev->qlen))) {
drivers/crypto/cavium/nitrox/nitrox_reqmgr.c
357
if (unlikely(cmdq_full(cmdq, ndev->qlen))) {
drivers/crypto/hifn_795x.c
1806
dev->success, dev->queue.qlen, dev->queue.max_qlen,
drivers/crypto/hifn_795x.c
2062
if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
drivers/crypto/hifn_795x.c
2327
if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
103
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
112
queue->qlen = qlen;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
125
static int init_pending_queues(struct otx_cptvf *cptvf, u32 qlen,
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
134
ret = alloc_pending_queues(&cptvf->pqinfo, qlen, num_queues);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
188
u32 qlen)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
199
cptvf->qsize = min(qlen, cqinfo->qchunksize) *
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
202
q_size = qlen * OTX_CPT_INST_SIZE;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
263
static int init_command_queues(struct otx_cptvf *cptvf, u32 qlen)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
269
ret = alloc_command_queues(cptvf, &cptvf->cqinfo, qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
297
static int cptvf_sw_init(struct otx_cptvf *cptvf, u32 qlen, u32 num_queues)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
308
ret = init_command_queues(cptvf, qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
315
ret = init_pending_queues(cptvf, qlen, num_queues);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
88
queue->qlen = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
93
static int alloc_pending_queues(struct otx_cpt_pending_qinfo *pqinfo, u32 qlen,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
325
pentry = get_free_pending_entry(pqueue, pqueue->qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
331
pentry = get_free_pending_entry(pqueue, pqueue->qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
346
pqueue->pending_count > (pqueue->qlen - CPT_IQ_STOP_MARGIN)) {
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
567
resume_index = modulo_inc(pqueue->front, pqueue->qlen,
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
59
int qlen)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
593
pqueue->front = modulo_inc(pqueue->front, pqueue->qlen, 1);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
68
if (unlikely(q->rear == qlen))
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.h
153
u32 qlen; /* Queue length */
drivers/crypto/marvell/octeontx2/otx2_cpt_reqmgr.h
101
u32 qlen; /* Queue length */
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
189
lfs->lf[i].pqueue.qlen = OTX2_CPT_INST_QLEN_MSGS;
drivers/crypto/marvell/octeontx2/otx2_cptvf_main.c
190
size = lfs->lf[i].pqueue.qlen *
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
115
pentry = get_free_pending_entry(pqueue, pqueue->qlen);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
121
pentry = get_free_pending_entry(pqueue, pqueue->qlen);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
135
pqueue->pending_count > (pqueue->qlen - CPT_IQ_STOP_MARGIN)) {
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
349
resume_index = modulo_inc(pqueue->front, pqueue->qlen,
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
375
pqueue->front = modulo_inc(pqueue->front, pqueue->qlen, 1);
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
47
int qlen)
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
56
if (unlikely(q->rear == qlen))
drivers/firmware/samsung/exynos-acpm.c
149
unsigned int qlen;
drivers/firmware/samsung/exynos-acpm.c
279
i = (i + 1) % achan->qlen;
drivers/firmware/samsung/exynos-acpm.c
441
idx = (tx_front + 1) % achan->qlen;
drivers/firmware/samsung/exynos-acpm.c
482
achan->qlen = readl(&chan_shmem->qlen);
drivers/firmware/samsung/exynos-acpm.c
493
achan->id, achan->poll_completion, achan->mlen, achan->qlen);
drivers/firmware/samsung/exynos-acpm.c
83
u32 qlen;
drivers/input/misc/wistron_btns.c
1148
u8 qlen;
drivers/input/misc/wistron_btns.c
1152
qlen = CMOS_READ(cmos_address);
drivers/input/misc/wistron_btns.c
1153
if (qlen == 0)
drivers/md/dm-ps-queue-length.c
101
DMEMIT("%d ", atomic_read(&pi->qlen));
drivers/md/dm-ps-queue-length.c
153
atomic_set(&pi->qlen, 0);
drivers/md/dm-ps-queue-length.c
204
(atomic_read(&pi->qlen) < atomic_read(&best->qlen)))
drivers/md/dm-ps-queue-length.c
207
if (!atomic_read(&best->qlen))
drivers/md/dm-ps-queue-length.c
228
atomic_inc(&pi->qlen);
drivers/md/dm-ps-queue-length.c
238
atomic_dec(&pi->qlen);
drivers/md/dm-ps-queue-length.c
40
atomic_t qlen; /* the number of in-flight I/Os */
drivers/misc/hpilo.c
54
int qlen = len & 7 ? (len >> 3) + 1 : len >> 3;
drivers/misc/hpilo.c
55
return id << ENTRY_BITPOS_DESCRIPTOR | qlen << ENTRY_BITPOS_QWORDS;
drivers/net/caif/caif_serial.c
243
if (ser->head.qlen <= SEND_QUEUE_LOW &&
drivers/net/caif/caif_serial.c
261
if (ser->head.qlen > SEND_QUEUE_HIGH &&
drivers/net/ethernet/alacritech/slicoss.c
64
static inline int slic_next_queue_idx(unsigned int idx, unsigned int qlen)
drivers/net/ethernet/alacritech/slicoss.c
66
return (idx + 1) & (qlen - 1);
drivers/net/ethernet/alacritech/slicoss.c
71
unsigned int qlen)
drivers/net/ethernet/alacritech/slicoss.c
74
return (qlen - (put_idx - done_idx) - 1);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3070
if (q->sendq.qlen == 1)
drivers/net/ethernet/chelsio/cxgb4/srq.c
73
e->qlen = SRQT_QLEN_G(be32_to_cpu(rpl->qlen_qbase));
drivers/net/ethernet/chelsio/cxgb4/srq.h
46
u8 qlen;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
762
int qlen = skb_queue_len(&csk->txq);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
764
if (likely(qlen)) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
777
if (qlen == 1 &&
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
682
{I40E_HMC_STORE(i40e_hmc_obj_txq, qlen), 13, 33 + 128 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
700
{ I40E_HMC_STORE(i40e_hmc_obj_rxq, qlen), 13, 89 },
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
25
u16 qlen;
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
65
u16 qlen;
drivers/net/ethernet/intel/i40e/i40e_main.c
3469
tx_ctx.qlen = ring->count;
drivers/net/ethernet/intel/i40e/i40e_main.c
3628
rx_ctx.qlen = ring->count;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
663
tx_ctx.qlen = info->ring_len;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
733
rx_ctx.qlen = info->ring_len;
drivers/net/ethernet/intel/ice/ice_base.c
354
tlan_ctx->qlen = ring->count;
drivers/net/ethernet/intel/ice/ice_base.c
414
txtime_ctx->qlen = ring->count;
drivers/net/ethernet/intel/ice/ice_base.c
496
rlan_ctx.qlen = ring->count;
drivers/net/ethernet/intel/ice/ice_common.c
1389
ICE_CTX_STORE(ice_rlan_ctx, qlen, 13, 89),
drivers/net/ethernet/intel/ice/ice_common.c
1510
ICE_CTX_STORE(ice_tlan_ctx, qlen, 13, 135),
drivers/net/ethernet/intel/ice/ice_common.c
1711
ICE_CTX_STORE(ice_txtime_ctx, qlen, 13, 91),
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
387
u16 qlen;
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
565
u16 qlen;
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
600
u16 qlen;
drivers/net/ethernet/microsoft/mana/mana_en.c
2320
&tx_qp->txq.pending_skbs.qlen);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
466
qstats->qlen += new->backlog_pkts - old->backlog_pkts;
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
524
if (!critical && nn->mbox_cmsg.queue.qlen >= NFP_CCM_MAX_QLEN) {
drivers/net/ethernet/sun/niu.c
3700
int qlen, rcr_done = 0, work_done = 0;
drivers/net/ethernet/sun/niu.c
3706
qlen = nr64(RCRSTAT_A(rp->rx_channel)) & RCRSTAT_A_QLEN;
drivers/net/ethernet/sun/niu.c
3709
qlen = (le64_to_cpup(&mbox->rcrstat_a) & RCRSTAT_A_QLEN);
drivers/net/ethernet/sun/niu.c
3716
__func__, rp->rx_channel, (unsigned long long)stat, qlen);
drivers/net/ethernet/sun/niu.c
3719
qlen = min(qlen, budget);
drivers/net/ethernet/sun/niu.c
3720
while (work_done < qlen) {
drivers/net/ethernet/sun/niu.c
3740
if (qlen > 10)
drivers/net/ethernet/toshiba/tc35815.c
1786
int qlen = (lp->tfd_start + TX_FD_NUM
drivers/net/ethernet/toshiba/tc35815.c
1796
if (lp->lstats.max_tx_qlen < qlen)
drivers/net/ethernet/toshiba/tc35815.c
1797
lp->lstats.max_tx_qlen = qlen;
drivers/net/netdevsim/netdev.c
742
if (rq->skb_queue.qlen) {
drivers/net/netdevsim/netdev.c
744
dev_dstats_rx_dropped_add(dev, rq->skb_queue.qlen);
drivers/net/ppp/ppp_generic.c
1868
if (ppp->file.rq.qlen > PPP_MAX_RQLEN)
drivers/net/ppp/ppp_generic.c
2354
while (pch->file.rq.qlen > PPP_MAX_RQLEN &&
drivers/net/ppp/ppp_generic.c
2508
while (ppp->file.rq.qlen > PPP_MAX_RQLEN &&
drivers/net/usb/usbnet.c
1522
if (dev->txq.qlen >= TX_QLEN (dev))
drivers/net/usb/usbnet.c
1555
for (i = 0; i < 10 && dev->rxq.qlen < RX_QLEN(dev); i++) {
drivers/net/usb/usbnet.c
1620
if (dev->txq.qlen + dev->rxq.qlen + dev->done.qlen == 0)
drivers/net/usb/usbnet.c
1631
int temp = dev->rxq.qlen;
drivers/net/usb/usbnet.c
1636
if (temp != dev->rxq.qlen)
drivers/net/usb/usbnet.c
1639
temp, dev->rxq.qlen);
drivers/net/usb/usbnet.c
1640
if (dev->rxq.qlen < RX_QLEN(dev))
drivers/net/usb/usbnet.c
1643
if (dev->txq.qlen < TX_QLEN (dev))
drivers/net/usb/usbnet.c
1949
if (dev->txq.qlen && PMSG_IS_AUTO(message)) {
drivers/net/usb/usbnet.c
2016
if (!(dev->txq.qlen >= TX_QLEN(dev)))
drivers/net/usb/usbnet.c
464
if (dev->done.qlen == 1)
drivers/net/wireguard/send.c
337
peer->staged_packet_queue.qlen);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
394
if (!pktlist->qlen)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
426
target_list->qlen);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
573
addr, pktq->qlen);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
582
if (pktq->qlen == 1)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
644
brcmf_dbg(SDIO, "addr = 0x%x, size = %d\n", addr, pktq->qlen);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/bcmsdh.c
653
if (pktq->qlen == 1 || !sdiodev->sg_support) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1377
int qlen;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1390
qlen = brcmf_flowring_qlen(msgbuf->flow, flowid);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1391
if ((qlen > BRCMF_MSGBUF_TRICKLE_TXWORKER_THRS) ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1392
((qlen) && (atomic_read(&commonring->outstanding_tx) <
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
1711
bus->glom.qlen, pfirst, pfirst->data,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2220
if (bus->txglom && pktq->qlen > 1) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c
2322
bus->tx_seq = (bus->tx_seq + pktq->qlen) % SDPCM_SEQ_WRAP;
drivers/net/wireless/broadcom/brcm80211/brcmutil/utils.c
212
len += pq->q[prec].skblist.qlen;
drivers/net/wireless/broadcom/brcm80211/include/brcmu_utils.h
78
return pq->q[prec].skblist.qlen;
drivers/net/wireless/broadcom/brcm80211/include/brcmu_utils.h
83
return pq->q[prec].max - pq->q[prec].skblist.qlen;
drivers/net/wireless/broadcom/brcm80211/include/brcmu_utils.h
88
return pq->q[prec].skblist.qlen >= pq->q[prec].max;
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
365
u32 ctrl, val, qlen = 0;
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
379
qlen += mt76_get_field(dev, MT_PLE_FL_Q3_CTRL,
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
382
seq_printf(s, "AC%d%d: queued=%d\n", wmm_idx, acs, qlen);
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
853
u32 qlen, ctrl, val;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
864
qlen = mt76_get_field(dev, MT_FL_Q3_CTRL,
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
868
msta->vif->mt76.wmm_idx, ac, qlen);
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
75
u32 ctrl, val, qlen = 0;
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
86
qlen += mt76_get_field(dev, MT_PLE_FL_Q3_CTRL,
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
89
seq_printf(s, "AC%d: queued=%d\n", i, qlen);
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
653
u32 idx = msta_link->wcid.idx >> 5, qlen, ctrl, val;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
664
qlen = mt76_get_field(dev, MT_FL_Q3_CTRL,
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
668
mlink->wmm_idx, ac, qlen);
drivers/net/wireless/realtek/rtlwifi/usb.c
550
unsigned int qlen;
drivers/net/wireless/realtek/rtlwifi/usb.c
560
qlen = skb_queue_len(&rtlusb->rx_queue);
drivers/net/wireless/realtek/rtlwifi/usb.c
561
if (qlen >= __RX_SKB_MAX_QUEUED) {
drivers/net/wireless/realtek/rtlwifi/usb.c
563
qlen);
drivers/net/wireless/realtek/rtw89/pci.c
106
u32 qlen;
drivers/net/wireless/realtek/rtw89/pci.c
117
qlen = skb_queue_len(&rtwpci->h2c_release_queue);
drivers/net/wireless/realtek/rtw89/pci.c
119
qlen = qlen > RTW89_PCI_MULTITAG ? qlen - RTW89_PCI_MULTITAG : 0;
drivers/net/wireless/realtek/rtw89/pci.c
121
while (qlen--) {
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
1026
session->ul_list.qlen); /* nr_of_bytes */
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
1179
int qlevel = ul_list->qlen;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
1548
if_id, skb_queue_len(&session->ul_list), session->ul_list.qlen,
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
991
qlt->vfl.nr_of_bytes = cpu_to_le32(session->ul_list.qlen);
drivers/net/wwan/t7xx/t7xx_port_proxy.c
244
if (port->rx_skb_list.qlen >= port->rx_length_th) {
drivers/nfc/st-nci/ndlc.c
153
if (ndlc->rcv_q.qlen)
drivers/nfc/st-nci/ndlc.c
154
pr_debug("rcvQlen=%d\n", ndlc->rcv_q.qlen);
drivers/nfc/st-nci/ndlc.c
94
if (ndlc->send_q.qlen)
drivers/nfc/st-nci/ndlc.c
96
ndlc->send_q.qlen, ndlc->ack_pending_q.qlen);
drivers/nfc/st-nci/ndlc.c
98
while (ndlc->send_q.qlen) {
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
406
if (port->fcoe_pending_queue.qlen)
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
469
if (bg->fcoe_rx_list.qlen == 1)
drivers/scsi/fcoe/fcoe.c
551
if (port->fcoe_pending_queue.qlen)
drivers/scsi/fcoe/fcoe_transport.c
411
while (port->fcoe_pending_queue.qlen) {
drivers/scsi/fcoe/fcoe_transport.c
413
port->fcoe_pending_queue.qlen++;
drivers/scsi/fcoe/fcoe_transport.c
423
port->fcoe_pending_queue.qlen--;
drivers/scsi/fcoe/fcoe_transport.c
427
port->fcoe_pending_queue.qlen--;
drivers/scsi/fcoe/fcoe_transport.c
430
if (port->fcoe_pending_queue.qlen < port->min_queue_depth)
drivers/scsi/fcoe/fcoe_transport.c
432
if (port->fcoe_pending_queue.qlen && !timer_pending(&port->timer))
drivers/scsi/fcoe/fcoe_transport.c
436
if (port->fcoe_pending_queue.qlen > port->max_queue_depth)
drivers/tty/serial/jsm/jsm_neo.c
482
int qlen;
drivers/tty/serial/jsm/jsm_neo.c
522
qlen = kfifo_len(&tport->xmit_fifo);
drivers/tty/serial/jsm/jsm_neo.c
525
n = min(n, qlen);
drivers/tty/serial/men_z135_uart.c
301
int qlen;
drivers/tty/serial/men_z135_uart.c
313
qlen = kfifo_len(&tport->xmit_fifo);
drivers/tty/serial/men_z135_uart.c
314
if (qlen <= 0)
drivers/tty/serial/men_z135_uart.c
328
txfree, qlen);
drivers/tty/serial/men_z135_uart.c
335
if (align && qlen >= 3 && BYTES_TO_ALIGN(wptr))
drivers/tty/serial/men_z135_uart.c
337
else if (qlen > txfree)
drivers/tty/serial/men_z135_uart.c
340
n = qlen;
drivers/usb/gadget/function/f_hid.c
1075
for (i = 0; i < hidg->qlen && status == 0; i++) {
drivers/usb/gadget/function/f_hid.c
1649
hidg->qlen = 4;
drivers/usb/gadget/function/f_hid.c
84
unsigned int qlen;
drivers/usb/gadget/function/f_loopback.c
321
for (i = 0; i < loop->qlen && result == 0; i++) {
drivers/usb/gadget/function/f_loopback.c
34
unsigned qlen;
drivers/usb/gadget/function/f_loopback.c
439
loop->qlen = lb_opts->qlen;
drivers/usb/gadget/function/f_loopback.c
440
if (!loop->qlen)
drivers/usb/gadget/function/f_loopback.c
441
loop->qlen = 32;
drivers/usb/gadget/function/f_loopback.c
477
result = sprintf(page, "%d\n", opts->qlen);
drivers/usb/gadget/function/f_loopback.c
500
opts->qlen = num;
drivers/usb/gadget/function/f_loopback.c
507
CONFIGFS_ATTR(f_lb_opts_, qlen);
drivers/usb/gadget/function/f_loopback.c
577
lb_opts->qlen = GZERO_QLEN;
drivers/usb/gadget/function/f_midi.c
1234
F_MIDI_OPT(qlen, false, 0);
drivers/usb/gadget/function/f_midi.c
1291
opts->qlen = 32;
drivers/usb/gadget/function/f_midi.c
1382
midi->qlen = opts->qlen;
drivers/usb/gadget/function/f_midi.c
1386
status = kfifo_alloc(&midi->in_req_fifo, midi->qlen, GFP_KERNEL);
drivers/usb/gadget/function/f_midi.c
397
for (i = 0; i < midi->qlen && err == 0; i++) {
drivers/usb/gadget/function/f_midi.c
96
unsigned int buflen, qlen;
drivers/usb/gadget/function/f_sourcesink.c
588
int i, size, qlen, status = 0;
drivers/usb/gadget/function/f_sourcesink.c
607
qlen = ss->iso_qlen;
drivers/usb/gadget/function/f_sourcesink.c
610
qlen = ss->bulk_qlen;
drivers/usb/gadget/function/f_sourcesink.c
614
for (i = 0; i < qlen; i++) {
drivers/usb/gadget/function/g_zero.h
24
unsigned qlen;
drivers/usb/gadget/function/g_zero.h
54
unsigned qlen;
drivers/usb/gadget/function/u_ether.c
1164
result = alloc_requests(dev, link, qlen(dev->gadget,
drivers/usb/gadget/function/u_ether.c
1170
DBG(dev, "qlen %d\n", qlen(dev->gadget, dev->qmult));
drivers/usb/gadget/function/u_midi.h
26
unsigned int qlen;
drivers/usb/gadget/legacy/gmidi.c
157
midi_opts->qlen = qlen;
drivers/usb/gadget/legacy/gmidi.c
53
static unsigned int qlen = 32;
drivers/usb/gadget/legacy/gmidi.c
54
module_param(qlen, uint, S_IRUGO);
drivers/usb/gadget/legacy/gmidi.c
55
MODULE_PARM_DESC(qlen, "USB read and write request queue length");
drivers/usb/gadget/legacy/printer.c
51
static unsigned qlen = 10;
drivers/usb/gadget/legacy/printer.c
52
module_param(qlen, uint, S_IRUGO|S_IWUSR);
drivers/usb/gadget/legacy/printer.c
53
MODULE_PARM_DESC(qlen, "The number of 8k buffers to use per endpoint");
drivers/usb/gadget/legacy/printer.c
55
#define QLEN qlen
drivers/usb/gadget/legacy/zero.c
263
module_param_named(qlen, gzero_options.qlen, uint, S_IRUGO|S_IWUSR);
drivers/usb/gadget/legacy/zero.c
264
MODULE_PARM_DESC(qlen, "depth of loopback queue");
drivers/usb/gadget/legacy/zero.c
322
lb_opts->qlen = gzero_options.qlen;
drivers/usb/gadget/legacy/zero.c
66
.qlen = GZERO_QLEN,
drivers/usb/host/ohci-dbg.c
533
unsigned qlen = 0;
drivers/usb/host/ohci-dbg.c
537
qlen++;
drivers/usb/host/ohci-dbg.c
547
qlen,
fs/adfs/dir.c
333
static int __adfs_compare(const unsigned char *qstr, u32 qlen,
fs/adfs/dir.c
338
if (qlen != len)
fs/adfs/dir.c
341
for (i = 0; i < qlen; i++)
include/crypto/algapi.h
106
unsigned int qlen;
include/crypto/algapi.h
180
return queue->qlen;
include/crypto/engine.h
79
bool rt, int qlen);
include/linux/skbuff.h
2263
return list_->qlen;
include/linux/skbuff.h
2275
return READ_ONCE(list_->qlen);
include/linux/skbuff.h
2291
list->qlen = 0;
include/linux/skbuff.h
2332
WRITE_ONCE(list->qlen, list->qlen + 1);
include/linux/skbuff.h
2359
head->qlen += list->qlen;
include/linux/skbuff.h
2375
head->qlen += list->qlen;
include/linux/skbuff.h
2390
head->qlen += list->qlen;
include/linux/skbuff.h
2407
head->qlen += list->qlen;
include/linux/skbuff.h
2483
WRITE_ONCE(list->qlen, list->qlen - 1);
include/linux/skbuff.h
344
__u32 qlen;
include/linux/sunrpc/sched.h
202
unsigned int qlen; /* total # tasks waiting in queue */
include/net/gen_stats.h
62
struct gnet_stats_queue *q, __u32 qlen);
include/net/neighbour.h
87
u32 qlen;
include/net/request_sock.h
164
int qlen; /* # of pending (TCP_SYN_RECV) reqs */
include/net/request_sock.h
182
atomic_t qlen;
include/net/request_sock.h
222
atomic_dec(&queue->qlen);
include/net/request_sock.h
228
atomic_inc(&queue->qlen);
include/net/request_sock.h
233
return atomic_read(&queue->qlen);
include/net/sch_generic.h
1026
__u32 qlen = qdisc_qlen_sum(sch);
include/net/sch_generic.h
1028
return gnet_stats_copy_queue(d, sch->cpu_qstats, &sch->qstats, qlen);
include/net/sch_generic.h
1031
static inline void qdisc_qstats_qlen_backlog(struct Qdisc *sch, __u32 *qlen,
include/net/sch_generic.h
1037
*qlen = qstats.qlen + qdisc_qlen(sch);
include/net/sch_generic.h
1043
__u32 qlen, backlog;
include/net/sch_generic.h
1045
qdisc_qstats_qlen_backlog(sch, &qlen, &backlog);
include/net/sch_generic.h
1047
qdisc_tree_reduce_backlog(sch, qlen, backlog);
include/net/sch_generic.h
1063
qh->qlen++;
include/net/sch_generic.h
1081
qh->qlen++;
include/net/sch_generic.h
1090
qh->qlen--;
include/net/sch_generic.h
1105
sch->q.qlen--;
include/net/sch_generic.h
1237
sch->q.qlen++;
include/net/sch_generic.h
1254
sch->q.qlen--;
include/net/sch_generic.h
1266
sch->q.qlen++;
include/net/sch_generic.h
1282
sch->q.qlen--;
include/net/sch_generic.h
1298
if (qh->qlen) {
include/net/sch_generic.h
1303
qh->qlen = 0;
include/net/sch_generic.h
193
return !READ_ONCE(qdisc->q.qlen);
include/net/sch_generic.h
539
return q->q.qlen;
include/net/sch_generic.h
544
__u32 qlen = q->qstats.qlen;
include/net/sch_generic.h
549
qlen += per_cpu_ptr(q->cpu_qstats, i)->qlen;
include/net/sch_generic.h
551
qlen += q->q.qlen;
include/net/sch_generic.h
554
return qlen;
include/net/sch_generic.h
62
__u32 qlen;
include/net/sch_generic.h
981
this_cpu_inc(sch->cpu_qstats->qlen);
include/net/sch_generic.h
986
this_cpu_dec(sch->cpu_qstats->qlen);
include/sound/seq_midi_event.h
18
int qlen; /* queue length */
include/trace/events/rcu.h
514
TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen),
include/trace/events/rcu.h
516
TP_ARGS(rcuname, rhp, qlen),
include/trace/events/rcu.h
522
__field(long, qlen)
include/trace/events/rcu.h
529
__entry->qlen = qlen;
include/trace/events/rcu.h
534
__entry->qlen)
include/trace/events/rcu.h
572
TP_PROTO(const char *rcuname, long qlen, long blimit),
include/trace/events/rcu.h
574
TP_ARGS(rcuname, qlen, blimit),
include/trace/events/rcu.h
578
__field(long, qlen)
include/trace/events/rcu.h
584
__entry->qlen = qlen;
include/trace/events/rcu.h
589
__entry->rcuname, __entry->qlen, __entry->blimit)
include/uapi/linux/gen_stats.h
60
__u32 qlen;
include/uapi/linux/i2o-dev.h
92
unsigned int qlen; /* Length in bytes of query string buffer */
include/uapi/linux/pkt_sched.h
42
__u32 qlen;
net/atm/clip.c
378
if (entry->neigh->arp_queue.qlen < ATMARP_MAX_UNRES_PACKETS)
net/atm/lec.c
298
if (entry && (entry->tx_wait.qlen < LEC_UNRES_QUE_LEN)) {
net/caif/caif_dev.c
170
int err, high = 0, qlen = 0;
net/caif/caif_dev.c
203
if (!(sch->flags & TCQ_F_NOLOCK) && likely(sch->q.qlen < high))
net/caif/caif_dev.c
223
qlen, high);
net/core/dev.c
5303
static bool skb_flow_limit(struct sk_buff *skb, unsigned int qlen,
net/core/dev.c
5311
if (likely(qlen < (max_backlog >> 1)))
net/core/dev.c
5351
unsigned int qlen;
net/core/dev.c
5361
qlen = skb_queue_len_lockless(&sd->input_pkt_queue);
net/core/dev.c
5363
if (unlikely(qlen > max_backlog) ||
net/core/dev.c
5364
skb_flow_limit(skb, qlen, max_backlog))
net/core/dev.c
5367
qlen = skb_queue_len(&sd->input_pkt_queue);
net/core/dev.c
5368
if (likely(qlen <= max_backlog)) {
net/core/dev.c
5369
if (!qlen) {
net/core/gen_stats.c
348
qstats->qlen += qcpu->qlen;
net/core/gen_stats.c
363
qstats->qlen += q->qlen;
net/core/gen_stats.c
389
struct gnet_stats_queue *q, __u32 qlen)
net/core/gen_stats.c
394
qstats.qlen = qlen;
net/core/gen_stats.c
398
d->tc_stats.qlen = qstats.qlen;
net/core/neighbour.c
1730
if (p->qlen > NEIGH_VAR(p, PROXY_QLEN)) {
net/core/neighbour.c
1746
p->qlen++;
net/core/neighbour.c
1778
p->qlen = 0;
net/core/neighbour.c
1840
tbl->parms.qlen = 0;
net/core/neighbour.c
2281
.ndtc_proxy_qlen = READ_ONCE(tbl->proxy_queue.qlen),
net/core/neighbour.c
348
p->qlen--;
net/core/netpoll.c
234
while (READ_ONCE(skb_pool->qlen) < MAX_SKBS) {
net/ipv4/inet_connection_sock.c
1035
int max_syn_ack_retries, qlen, expire = 0, resend = 0;
net/ipv4/inet_connection_sock.c
1083
qlen = reqsk_queue_len(queue);
net/ipv4/inet_connection_sock.c
1084
if ((qlen << 1) > max(8U, READ_ONCE(sk_listener->sk_max_ack_backlog))) {
net/ipv4/inet_connection_sock.c
1088
if (qlen < young)
net/ipv4/inet_connection_sock.c
1317
queue->fastopenq.qlen = 0;
net/ipv4/inet_connection_sock.c
965
atomic_inc(&queue->qlen);
net/ipv4/ipmr.c
1176
if (c->_c.mfc_un.unres.unresolved.qlen > 3) {
net/ipv4/tcp_fastopen.c
341
queue->fastopenq.qlen++;
net/ipv4/tcp_fastopen.c
406
if (fastopenq->qlen >= max_qlen) {
net/ipv4/tcp_fastopen.c
417
fastopenq->qlen--;
net/ipv4/tcp_fastopen.c
61
fastopenq->qlen--;
net/ipv4/tcp_fastopen.c
89
fastopenq->qlen++;
net/ipv6/ioam6.c
804
__u32 qlen, backlog;
net/ipv6/ioam6.c
814
qdisc_qstats_qlen_backlog(qdisc, &qlen, &backlog);
net/ipv6/ip6mr.c
1216
if (c->_c.mfc_un.unres.unresolved.qlen > 3) {
net/kcm/kcmproc.c
119
kcm->sk.sk_receive_queue.qlen,
net/kcm/kcmproc.c
121
kcm->sk.sk_write_queue.qlen,
net/kcm/kcmproc.c
149
psock->sk->sk_receive_queue.qlen,
net/kcm/kcmproc.c
151
psock->sk->sk_write_queue.qlen,
net/kcm/kcmproc.c
167
if (psock->sk->sk_receive_queue.qlen) {
net/mac80211/cfg.c
2641
pinfo->frame_qlen = mpath->frame_queue.qlen;
net/mctp/test/route-test.c
1001
n = dev->pkts.qlen;
net/mctp/test/route-test.c
147
KUNIT_EXPECT_EQ(test, !!dev->pkts.qlen, params->input);
net/mctp/test/route-test.c
50
n = dev->pkts.qlen;
net/mctp/test/route-test.c
967
n = dev->pkts.qlen;
net/netfilter/nfnetlink_log.c
1097
inst->peer_portid, inst->qlen,
net/netfilter/nfnetlink_log.c
363
if (inst->qlen > 1) {
net/netfilter/nfnetlink_log.c
376
inst->qlen = 0;
net/netfilter/nfnetlink_log.c
65
unsigned int qlen; /* number of nlmsgs in skb */
net/netfilter/nfnetlink_log.c
799
inst->qlen++;
net/netfilter/nfnetlink_log.c
805
if (inst->qlen >= qthreshold)
net/nfc/hci/llc_shdlc.c
322
if (shdlc->send_q.qlen == 0) {
net/nfc/hci/llc_shdlc.c
462
if (shdlc->rcv_q.qlen)
net/nfc/hci/llc_shdlc.c
463
pr_debug("rcvQlen=%d\n", shdlc->rcv_q.qlen);
net/nfc/hci/llc_shdlc.c
518
if (shdlc->send_q.qlen)
net/nfc/hci/llc_shdlc.c
520
shdlc->send_q.qlen, shdlc->ns, shdlc->dnr,
net/nfc/hci/llc_shdlc.c
523
shdlc->ack_pending_q.qlen);
net/nfc/hci/llc_shdlc.c
525
while (shdlc->send_q.qlen && shdlc->ack_pending_q.qlen < shdlc->w &&
net/rxrpc/proc.c
426
local->rx_queue.qlen);
net/sched/act_api.c
1635
qstats.qlen) < 0)
net/sched/bpf_qdisc.c
62
case offsetof(struct Qdisc, q) + offsetof(struct qdisc_skb_head, qlen):
net/sched/bpf_qdisc.c
63
*end = offsetof(struct Qdisc, q) + offsetofend(struct qdisc_skb_head, qlen);
net/sched/em_meta.c
430
dst->value = sk->sk_receive_queue.qlen;
net/sched/em_meta.c
441
dst->value = sk->sk_write_queue.qlen;
net/sched/em_meta.c
516
dst->value = sk->sk_error_queue.qlen;
net/sched/sch_api.c
793
notify = !sch->q.qlen;
net/sched/sch_api.c
808
sch->q.qlen -= n;
net/sched/sch_api.c
910
__u32 qlen;
net/sched/sch_api.c
942
qlen = qdisc_qlen_sum(q);
net/sched/sch_api.c
962
gnet_stats_copy_queue(&d, cpu_qstats, &q->qstats, qlen) < 0)
net/sched/sch_cake.c
1608
sch->q.qlen--;
net/sched/sch_cake.c
1781
if (!sch->q.qlen) {
net/sched/sch_cake.c
1818
sch->q.qlen++;
net/sched/sch_cake.c
1857
sch->q.qlen++;
net/sched/sch_cake.c
1945
prev_qlen = sch->q.qlen;
net/sched/sch_cake.c
1955
prev_qlen -= sch->q.qlen;
net/sched/sch_cake.c
1983
sch->q.qlen--;
net/sched/sch_cake.c
2033
other_qlen = READ_ONCE(other_sch->q.qlen);
net/sched/sch_cake.c
2049
if (!sch->q.qlen)
net/sched/sch_cake.c
2261
if (ktime_after(q->time_next_packet, now) && sch->q.qlen) {
net/sched/sch_cake.c
2266
} else if (!sch->q.qlen) {
net/sched/sch_cake.c
3135
qs.qlen++;
net/sched/sch_cake.c
3143
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
net/sched/sch_cbs.c
100
sch->q.qlen++;
net/sched/sch_cbs.c
120
if (sch->q.qlen == 0 && q->credits > 0) {
net/sched/sch_cbs.c
171
sch->q.qlen--;
net/sched/sch_choke.c
126
--sch->q.qlen;
net/sched/sch_choke.c
220
q->vars.qavg = red_calc_qavg(p, &q->vars, sch->q.qlen);
net/sched/sch_choke.c
267
if (sch->q.qlen < q->limit) {
net/sched/sch_choke.c
270
++sch->q.qlen;
net/sched/sch_choke.c
297
--sch->q.qlen;
net/sched/sch_choke.c
380
unsigned int oqlen = sch->q.qlen, tail = 0;
net/sched/sch_choke.c
395
--sch->q.qlen;
net/sched/sch_choke.c
398
qdisc_tree_reduce_backlog(sch, oqlen - sch->q.qlen, dropped);
net/sched/sch_codel.c
145
while (sch->q.qlen > sch->limit) {
net/sched/sch_drr.c
267
__u32 qlen = qdisc_qlen_sum(cl->qdisc);
net/sched/sch_drr.c
272
if (qlen)
net/sched/sch_drr.c
277
gnet_stats_copy_queue(d, cl_q->cpu_qstats, &cl_q->qstats, qlen) < 0)
net/sched/sch_drr.c
369
sch->q.qlen++;
net/sched/sch_drr.c
396
if (cl->qdisc->q.qlen == 0)
net/sched/sch_drr.c
402
sch->q.qlen--;
net/sched/sch_drr.c
437
if (cl->qdisc->q.qlen)
net/sched/sch_dualpi2.c
420
++sch->q.qlen;
net/sched/sch_dualpi2.c
535
--sch->q.qlen;
net/sched/sch_etf.c
192
sch->q.qlen++;
net/sched/sch_etf.c
225
sch->q.qlen--;
net/sched/sch_etf.c
250
sch->q.qlen--;
net/sched/sch_etf.c
429
sch->q.qlen--;
net/sched/sch_ets.c
304
if (!ets_class_is_strict(q, cl) && sch->q.qlen)
net/sched/sch_ets.c
452
sch->q.qlen++;
net/sched/sch_ets.c
461
sch->q.qlen--;
net/sched/sch_ets.c
497
if (cl->qdisc->q.qlen == 0)
net/sched/sch_ets.c
666
if (q->classes[i].qdisc->q.qlen) {
net/sched/sch_ets.c
727
if (q->classes[band].qdisc->q.qlen)
net/sched/sch_fifo.c
32
if (likely(sch->q.qlen < READ_ONCE(sch->limit)))
net/sched/sch_fifo.c
46
if (likely(sch->q.qlen < READ_ONCE(sch->limit)))
net/sched/sch_fq.c
1141
while (sch->q.qlen > sch->limit) {
net/sched/sch_fq.c
319
if (sch->q.qlen != 0) {
net/sched/sch_fq.c
331
if (q->internal.qlen >= 8)
net/sched/sch_fq.c
500
sch->q.qlen--;
net/sched/sch_fq.c
579
if (unlikely(f->qlen >= q->flow_plimit)) {
net/sched/sch_fq.c
594
if (f->qlen == 0)
net/sched/sch_fq.c
598
f->qlen++;
net/sched/sch_fq.c
603
sch->q.qlen++;
net/sched/sch_fq.c
662
if (!sch->q.qlen)
net/sched/sch_fq.c
667
q->internal.qlen--;
net/sched/sch_fq.c
721
if (--f->qlen == 0)
net/sched/sch_fq.c
797
flow->qlen = 0;
net/sched/sch_fq.c
808
sch->q.qlen = 0;
net/sched/sch_fq.c
86
int qlen; /* number of packets in flow queue */
net/sched/sch_fq_codel.c
181
sch->q.qlen -= i;
net/sched/sch_fq_codel.c
218
if (++sch->q.qlen <= sch->limit && !memory_limited)
net/sched/sch_fq_codel.c
222
prev_qlen = sch->q.qlen;
net/sched/sch_fq_codel.c
233
prev_qlen -= sch->q.qlen;
net/sched/sch_fq_codel.c
268
sch->q.qlen--;
net/sched/sch_fq_codel.c
443
while (sch->q.qlen > sch->limit ||
net/sched/sch_fq_codel.c
676
qs.qlen++;
net/sched/sch_fq_codel.c
684
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
net/sched/sch_fq_pie.c
188
sch->q.qlen++;
net/sched/sch_fq_pie.c
194
sel_flow->qlen = 0;
net/sched/sch_fq_pie.c
197
sel_flow->qlen++;
net/sched/sch_fq_pie.c
266
sch->q.qlen--;
net/sched/sch_fq_pie.c
279
flow->qlen--;
net/sched/sch_fq_pie.c
367
while (sch->q.qlen > sch->limit) {
net/sched/sch_fq_pie.c
49
u32 qlen;
net/sched/sch_generic.c
1037
qdisc->q.qlen = 0;
net/sched/sch_generic.c
137
q->q.qlen++;
net/sched/sch_generic.c
166
q->q.qlen++;
net/sched/sch_generic.c
272
q->q.qlen--;
net/sched/sch_generic.c
365
dev->name, ret, q->q.qlen);
net/sched/sch_generic.c
675
.qlen = 0,
net/sched/sch_generic.c
681
.qlen = 0,
net/sched/sch_generic.c
837
q->qlen = 0;
net/sched/sch_generic.c
858
unsigned int qlen = qdisc_dev(qdisc)->tx_queue_len;
net/sched/sch_generic.c
863
if (!qlen)
net/sched/sch_generic.c
870
err = skb_array_init(q, qlen, GFP_KERNEL);
net/sched/sch_generic.c
96
q->q.qlen--;
net/sched/sch_gred.c
390
sch->qstats.qlen += hw_stats->stats.qstats[i].qlen;
net/sched/sch_hfsc.c
1338
__u32 qlen;
net/sched/sch_hfsc.c
1340
qdisc_qstats_qlen_backlog(cl->qdisc, &qlen, &cl->qstats.backlog);
net/sched/sch_hfsc.c
1348
gnet_stats_copy_queue(d, NULL, &cl->qstats, qlen) < 0)
net/sched/sch_hfsc.c
1553
first = !cl->qdisc->q.qlen;
net/sched/sch_hfsc.c
1564
sch->q.qlen++;
net/sched/sch_hfsc.c
1594
if (sch->q.qlen == 0)
net/sched/sch_hfsc.c
1632
if (cl->qdisc->q.qlen != 0) {
net/sched/sch_hfsc.c
1639
if (cl->qdisc->q.qlen != 0) {
net/sched/sch_hfsc.c
1653
sch->q.qlen--;
net/sched/sch_hfsc.c
756
if (cl->qdisc->q.qlen == 0 && cl->cl_flags & HFSC_FSC)
net/sched/sch_hfsc.c
984
if (cl->qdisc->q.qlen != 0)
net/sched/sch_hfsc.c
990
if (cl->qdisc->q.qlen != 0) {
net/sched/sch_hhf.c
362
sch->q.qlen--;
net/sched/sch_hhf.c
402
if (++sch->q.qlen <= sch->limit)
net/sched/sch_hhf.c
445
sch->q.qlen--;
net/sched/sch_hhf.c
564
while (sch->q.qlen > sch->limit) {
net/sched/sch_htb.c
1330
__u32 qlen = 0;
net/sched/sch_htb.c
1333
qdisc_qstats_qlen_backlog(cl->leaf.q, &qlen, &qs.backlog);
net/sched/sch_htb.c
1356
gnet_stats_copy_queue(d, NULL, &qs, qlen) < 0)
net/sched/sch_htb.c
629
if (q->direct_queue.qlen < q->direct_qlen) {
net/sched/sch_htb.c
654
sch->q.qlen++;
net/sched/sch_htb.c
895
if (unlikely(cl->leaf.q->q.qlen == 0)) {
net/sched/sch_htb.c
933
if (!cl->leaf.q->q.qlen)
net/sched/sch_htb.c
954
sch->q.qlen--;
net/sched/sch_htb.c
958
if (!sch->q.qlen)
net/sched/sch_mq.c
149
sch->q.qlen = 0;
net/sched/sch_mq.c
166
sch->q.qlen += qdisc_qlen(qdisc);
net/sched/sch_mqprio.c
561
sch->q.qlen = 0;
net/sched/sch_mqprio.c
578
sch->q.qlen += qdisc_qlen(qdisc);
net/sched/sch_mqprio.c
667
__u32 qlen;
net/sched/sch_mqprio.c
692
sch->q.qlen += qdisc_qlen(qdisc);
net/sched/sch_mqprio.c
696
qlen = qdisc_qlen(sch) + qstats.qlen;
net/sched/sch_mqprio.c
702
gnet_stats_copy_queue(d, NULL, &qstats, qlen) < 0)
net/sched/sch_multiq.c
109
sch->q.qlen--;
net/sched/sch_multiq.c
79
sch->q.qlen++;
net/sched/sch_netem.c
419
sch->q.qlen++;
net/sched/sch_netem.c
755
sch->q.qlen--;
net/sched/sch_netem.c
760
sch->q.qlen--;
net/sched/sch_netem.c
767
sch->q.qlen--;
net/sched/sch_netem.c
780
sch->q.qlen--;
net/sched/sch_pie.c
196
while (sch->q.qlen > sch->limit) {
net/sched/sch_prio.c
122
sch->q.qlen--;
net/sched/sch_prio.c
89
sch->q.qlen++;
net/sched/sch_qfq.c
1002
if (cl->qdisc->q.qlen == 0) /* no more packets, remove from list */
net/sched/sch_qfq.c
1140
} else if (sch->q.qlen == 0) { /* no aggregate to serve */
net/sched/sch_qfq.c
1155
sch->q.qlen--;
net/sched/sch_qfq.c
1160
sch->q.qlen++;
net/sched/sch_qfq.c
1268
++sch->q.qlen;
net/sched/sch_qfq.c
327
if (cl->qdisc->q.qlen > 0) { /* adding an active class */
net/sched/sch_red.c
136
sch->q.qlen++;
net/sched/sch_red.c
162
sch->q.qlen--;
net/sched/sch_red.c
304
if (!q->qdisc->q.qlen)
net/sched/sch_sfb.c
132
if (b[hash].qlen < 0xFFFF)
net/sched/sch_sfb.c
133
b[hash].qlen++;
net/sched/sch_sfb.c
161
if (b[hash].qlen > 0)
net/sched/sch_sfb.c
162
b[hash].qlen--;
net/sched/sch_sfb.c
201
u32 qlen = 0, prob = 0, totalpm = 0;
net/sched/sch_sfb.c
205
if (qlen < b->qlen)
net/sched/sch_sfb.c
206
qlen = b->qlen;
net/sched/sch_sfb.c
214
return qlen;
net/sched/sch_sfb.c
296
if (unlikely(sch->q.qlen >= q->limit)) {
net/sched/sch_sfb.c
336
if (b->qlen == 0)
net/sched/sch_sfb.c
338
else if (b->qlen >= q->bin_size)
net/sched/sch_sfb.c
340
if (minqlen > b->qlen)
net/sched/sch_sfb.c
341
minqlen = b->qlen;
net/sched/sch_sfb.c
369
if (b->qlen == 0)
net/sched/sch_sfb.c
371
else if (b->qlen >= q->bin_size)
net/sched/sch_sfb.c
40
u16 qlen; /* length of virtual queue */
net/sched/sch_sfb.c
410
sch->q.qlen++;
net/sched/sch_sfb.c
439
sch->q.qlen--;
net/sched/sch_sfq.c
200
int qlen = slot->qlen;
net/sched/sch_sfq.c
202
p = qlen + SFQ_MAX_FLOWS;
net/sched/sch_sfq.c
203
n = q->dep[qlen].next;
net/sched/sch_sfq.c
208
q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */
net/sched/sch_sfq.c
228
d = q->slots[x].qlen--;
net/sched/sch_sfq.c
241
d = ++q->slots[x].qlen;
net/sched/sch_sfq.c
303
sch->q.qlen--;
net/sched/sch_sfq.c
346
sfq_index x, qlen;
net/sched/sch_sfq.c
420
if (slot->qlen >= q->maxdepth) {
net/sched/sch_sfq.c
442
if (slot->qlen == 1) { /* The flow is new */
net/sched/sch_sfq.c
457
if (++sch->q.qlen <= q->limit)
net/sched/sch_sfq.c
460
qlen = slot->qlen;
net/sched/sch_sfq.c
465
if (qlen != slot->qlen) {
net/sched/sch_sfq.c
498
sch->q.qlen--;
net/sched/sch_sfq.c
502
if (slot->qlen == 0) {
net/sched/sch_sfq.c
545
if (!slot->qlen)
net/sched/sch_sfq.c
547
while (slot->qlen) {
net/sched/sch_sfq.c
577
if (slot->qlen >= q->maxdepth)
net/sched/sch_sfq.c
586
if (slot->qlen == 1) { /* The flow is new */
net/sched/sch_sfq.c
597
sch->q.qlen -= dropped;
net/sched/sch_sfq.c
633
unsigned int qlen, dropped = 0;
net/sched/sch_sfq.c
732
qlen = sch->q.qlen;
net/sched/sch_sfq.c
733
while (sch->q.qlen > q->limit) {
net/sched/sch_sfq.c
740
qdisc_tree_reduce_backlog(sch, qlen - sch->q.qlen, dropped);
net/sched/sch_sfq.c
916
qs.qlen = slot->qlen;
net/sched/sch_sfq.c
919
if (gnet_stats_copy_queue(d, NULL, &qs, qs.qlen) < 0)
net/sched/sch_sfq.c
97
sfq_index qlen; /* number of skbs in skblist */
net/sched/sch_skbprio.c
148
sch->q.qlen--;
net/sched/sch_skbprio.c
254
q->qstats[cl - 1].qlen) < 0)
net/sched/sch_skbprio.c
84
if (sch->q.qlen < READ_ONCE(sch->limit)) {
net/sched/sch_skbprio.c
96
sch->q.qlen++;
net/sched/sch_taprio.c
577
sch->q.qlen++;
net/sched/sch_taprio.c
758
sch->q.qlen--;
net/sched/sch_tbf.c
234
sch->q.qlen += nb;
net/sched/sch_tbf.c
267
sch->q.qlen++;
net/sched/sch_tbf.c
312
sch->q.qlen--;
net/sched/sch_teql.c
110
sch->q.qlen = dat->q.qlen + q->q.qlen;
net/sched/sch_teql.c
81
if (q->q.qlen < READ_ONCE(dev->tx_queue_len)) {
net/sunrpc/sched.c
215
queue->qlen++;
net/sunrpc/sched.c
240
queue->qlen--;
net/sunrpc/sched.c
252
queue->qlen = 0;
net/sunrpc/sysfs.c
235
xprt->min_reqs, xprt->num_reqs, xprt->binding.qlen,
net/sunrpc/sysfs.c
236
xprt->sending.qlen, xprt->pending.qlen,
net/sunrpc/sysfs.c
237
xprt->backlog.qlen, xprt->main, srcport,
net/sunrpc/xprt.c
1597
xprt->stat.bklog_u += xprt->backlog.qlen;
net/sunrpc/xprt.c
1598
xprt->stat.sending_u += xprt->sending.qlen;
net/sunrpc/xprt.c
1599
xprt->stat.pending_u += xprt->pending.qlen;
net/sunrpc/xprt.c
2210
if (!xprt->sending.qlen && !xprt->pending.qlen &&
net/sunrpc/xprt.c
2211
!xprt->backlog.qlen && !atomic_long_read(&xprt->queuelen))
net/tipc/link.c
1542
u32 qlen = skb_queue_len(&l->transmq);
net/tipc/link.c
1680
return qlen - skb_queue_len(&l->transmq);
net/unix/diag.c
77
sk->sk_receive_queue.qlen * sizeof(u32));
net/xfrm/xfrm_policy.c
2996
if (pq->hold_queue.qlen > XFRM_MAX_QUEUE_LEN) {
sound/core/seq/oss/seq_oss_ioctl.c
104
return put_user(dp->readq->qlen, p) ? -EFAULT : 0;
sound/core/seq/oss/seq_oss_readq.c
144
if (q->qlen >= q->maxlen - 1)
sound/core/seq/oss/seq_oss_readq.c
149
q->qlen++;
sound/core/seq/oss/seq_oss_readq.c
165
if (q->qlen == 0)
sound/core/seq/oss/seq_oss_readq.c
178
(q->qlen > 0 || q->head == q->tail),
sound/core/seq/oss/seq_oss_readq.c
189
if (q->qlen > 0) {
sound/core/seq/oss/seq_oss_readq.c
191
q->qlen--;
sound/core/seq/oss/seq_oss_readq.c
203
return q->qlen;
sound/core/seq/oss/seq_oss_readq.c
242
q->qlen, q->input_time);
sound/core/seq/oss/seq_oss_readq.c
48
q->qlen = 0;
sound/core/seq/oss/seq_oss_readq.c
76
if (q->qlen) {
sound/core/seq/oss/seq_oss_readq.c
77
q->qlen = 0;
sound/core/seq/oss/seq_oss_readq.h
20
int qlen;
sound/core/seq/seq_midi_event.c
141
dev->qlen = 0;
sound/core/seq/seq_midi_event.c
193
dev->qlen = status_event[dev->type].qlen;
sound/core/seq/seq_midi_event.c
195
if (dev->qlen > 0) {
sound/core/seq/seq_midi_event.c
199
dev->qlen--;
sound/core/seq/seq_midi_event.c
203
dev->qlen = status_event[dev->type].qlen - 1;
sound/core/seq/seq_midi_event.c
207
if (dev->qlen == 0) {
sound/core/seq/seq_midi_event.c
312
int qlen;
sound/core/seq/seq_midi_event.c
323
qlen = status_event[type].qlen + 1;
sound/core/seq/seq_midi_event.c
328
qlen = status_event[type].qlen;
sound/core/seq/seq_midi_event.c
330
if (count < qlen)
sound/core/seq/seq_midi_event.c
332
memcpy(buf, xbuf, qlen);
sound/core/seq/seq_midi_event.c
333
return qlen;
sound/core/seq/seq_midi_event.c
50
int qlen;
tools/include/uapi/linux/pkt_sched.h
41
__u32 qlen;
tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
219
int i, err, prog_fd, reuseport = 1, qlen = QLEN;
tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
262
&qlen, sizeof(qlen));
tools/testing/selftests/bpf/prog_tests/migrate_reuseport.c
269
err = listen(test_case->servers[i], qlen);
tools/testing/selftests/bpf/progs/bpf_qdisc_fifo.c
109
sch->q.qlen = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fifo.c
26
if (sch->q.qlen == sch->limit)
tools/testing/selftests/bpf/progs/bpf_qdisc_fifo.c
35
sch->q.qlen++;
tools/testing/selftests/bpf/progs/bpf_qdisc_fifo.c
72
sch->q.qlen--;
tools/testing/selftests/bpf/progs/bpf_qdisc_fifo.c
93
bpf_for(i, 0, sch->q.qlen) {
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
248
flow->qlen = 0,
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
317
if (sch->q.qlen >= sch->limit)
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
344
if (flow->qlen >= q.flow_plimit) {
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
381
flow->qlen++;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
384
sch->q.qlen++;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
537
flow->qlen--;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
595
if (!sch->q.qlen)
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
612
sch->q.qlen--;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
698
sch->q.qlen = 0;
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
78
u32 qlen;
tools/testing/selftests/net/mptcp/mptcp_connect.c
264
int qlen = 25;
tools/testing/selftests/net/mptcp/mptcp_connect.c
266
if (setsockopt(fd, IPPROTO_TCP, TCP_FASTOPEN, &qlen, sizeof(qlen)) == -1)
tools/testing/selftests/net/tcp_ao/lib/repair.c
42
static void test_sock_checkpoint_queue(int sk, int queue, int qlen,
tools/testing/selftests/net/tcp_ao/lib/repair.c
56
if (!qlen) {
tools/testing/selftests/net/tcp_ao/lib/repair.c
61
q->buf = malloc(qlen);
tools/testing/selftests/net/tcp_ao/lib/repair.c
64
ret = recv(sk, q->buf, qlen, MSG_PEEK | MSG_DONTWAIT);
tools/testing/selftests/net/tcp_ao/lib/repair.c
65
if (ret != qlen)
tools/testing/selftests/net/tcp_ao/lib/repair.c
66
test_error("recv(%d): %d", qlen, ret);
tools/testing/selftests/net/tcp_fastopen_backup_key.c
132
if (setsockopt(rcv_fds[i], SOL_TCP, TCP_FASTOPEN, &qlen,
tools/testing/selftests/net/tcp_fastopen_backup_key.c
133
sizeof(qlen)))
tools/testing/selftests/net/tcp_fastopen_backup_key.c
95
int qlen = 100;
tools/testing/selftests/net/tfo.c
48
unsigned long qlen = 32;
tools/testing/selftests/net/tfo.c
66
if (setsockopt(fd, SOL_TCP, TCP_FASTOPEN, &qlen, sizeof(qlen)) < 0)