Symbol: tx_queue
sys/arm64/broadcom/genet/if_genet.c
1097
struct tx_queue *q;
sys/arm64/broadcom/genet/if_genet.c
1102
q = &sc->tx_queue[DEF_TXQUEUE];
sys/arm64/broadcom/genet/if_genet.c
1362
gen_txintr(sc, &sc->tx_queue[DEF_TXQUEUE]);
sys/arm64/broadcom/genet/if_genet.c
1483
gen_txintr(struct gen_softc *sc, struct tx_queue *q)
sys/arm64/broadcom/genet/if_genet.c
192
struct tx_queue tx_queue[NTXQUEUE];
sys/arm64/broadcom/genet/if_genet.c
223
static void gen_txintr(struct gen_softc *sc, struct tx_queue *q);
sys/arm64/broadcom/genet/if_genet.c
725
struct tx_queue *q;
sys/arm64/broadcom/genet/if_genet.c
728
q = &sc->tx_queue[queue];
sys/arm64/broadcom/genet/if_genet.c
826
sc->tx_queue[i].queue = i;
sys/arm64/broadcom/genet/if_genet.c
836
sc->tx_queue[DEF_TXQUEUE].hwindex = GENET_DMA_DEFAULT_QUEUE;
sys/dev/bnxt/bnxt_en/hsi_struct_def.h
76136
uint8_t tx_queue;
sys/dev/bnxt/bnxt_en/hsi_struct_def.h
76287
uint8_t tx_queue;
sys/dev/bnxt/bnxt_en/hsi_struct_def.h
76776
uint8_t tx_queue;
sys/dev/bnxt/bnxt_en/hsi_struct_def.h
76980
uint8_t tx_queue;
sys/dev/enetc/if_enetc.c
1024
tx_queue = &sc->tx_queues[i];
sys/dev/enetc/if_enetc.c
1026
iflib_softirq_alloc_generic(ctx, &tx_queue->irq,
sys/dev/enetc/if_enetc.c
1027
IFLIB_INTR_TX, tx_queue, i, irq_name);
sys/dev/enetc/if_enetc.c
999
struct enetc_tx_queue *tx_queue;
sys/dev/iwx/if_iwxreg.h
6110
uint16_t tx_queue;
sys/dev/qcom_ess_edma/qcom_ess_edma.c
423
int tx_queue = intr->irq_rid;
sys/dev/qcom_ess_edma/qcom_ess_edma.c
430
(void) qcom_ess_edma_hw_intr_tx_intr_set_enable(sc, tx_queue,
sys/dev/qcom_ess_edma/qcom_ess_edma.c
436
taskqueue_enqueue(sc->sc_tx_state[tx_queue].completion_tq,
sys/dev/qcom_ess_edma/qcom_ess_edma.c
437
&sc->sc_tx_state[tx_queue].completion_task);
sys/dev/qcom_ess_edma/qcom_ess_edma_hw.c
301
qcom_ess_edma_hw_intr_tx_ack(struct qcom_ess_edma_softc *sc, int tx_queue)
sys/dev/qcom_ess_edma/qcom_ess_edma_hw.c
304
EDMA_RING_LOCK_ASSERT(&sc->sc_tx_ring[tx_queue]);
sys/dev/qcom_ess_edma/qcom_ess_edma_hw.c
305
EDMA_REG_WRITE(sc, EDMA_REG_TX_ISR, (1U << tx_queue));
sys/dev/qcom_ess_edma/qcom_ess_edma_hw.h
53
int tx_queue);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1018
struct ecore_ll2_tx_queue *p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1116
&p_ll2_info->tx_queue.txq_chain, OSAL_NULL);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1120
capacity = ecore_chain_get_capacity(&p_ll2_info->tx_queue.txq_chain);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1131
p_ll2_info->tx_queue.descq_array = p_descq;
sys/dev/qlnx/qlnxe/ecore_ll2.c
131
p_pkt = &p_ll2_conn->tx_queue.cur_completing_packet;
sys/dev/qlnx/qlnxe/ecore_ll2.c
133
if (!p_ll2_conn->tx_queue.b_completing_packet || !p_addr)
sys/dev/qlnx/qlnxe/ecore_ll2.c
1353
&p_ll2_info->tx_queue.tx_sb_index,
sys/dev/qlnx/qlnxe/ecore_ll2.c
1354
&p_ll2_info->tx_queue.p_fw_cons);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1355
p_ll2_info->tx_queue.b_cb_registred = true;
sys/dev/qlnx/qlnxe/ecore_ll2.c
136
if (p_ll2_conn->tx_queue.cur_completing_bd_idx == p_pkt->bd_used)
sys/dev/qlnx/qlnxe/ecore_ll2.c
140
cur_frag_idx = p_ll2_conn->tx_queue.cur_completing_bd_idx++;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1427
p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
144
p_ll2_conn->tx_queue.cur_completing_bd_idx;
sys/dev/qlnx/qlnxe/ecore_ll2.c
164
p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1661
struct ecore_chain *p_tx_chain = &p_ll2->tx_queue.txq_chain;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1723
if (p_ll2->tx_queue.cur_send_frag_num == pkt->num_of_bds)
sys/dev/qlnx/qlnxe/ecore_ll2.c
1727
for (frag_idx = p_ll2->tx_queue.cur_send_frag_num;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1743
bool b_notify = p_ll2_conn->tx_queue.cur_send_packet->notify_fw;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1744
struct ecore_ll2_tx_queue *p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1749
if (p_ll2_conn->tx_queue.cur_send_frag_num !=
sys/dev/qlnx/qlnxe/ecore_ll2.c
1750
p_ll2_conn->tx_queue.cur_send_packet->bd_used)
sys/dev/qlnx/qlnxe/ecore_ll2.c
1754
OSAL_LIST_PUSH_TAIL(&p_ll2_conn->tx_queue.cur_send_packet->list_entry,
sys/dev/qlnx/qlnxe/ecore_ll2.c
1755
&p_ll2_conn->tx_queue.sending_descq);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1756
p_ll2_conn->tx_queue.cur_send_packet = OSAL_NULL;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1757
p_ll2_conn->tx_queue.cur_send_frag_num = 0;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1763
bd_prod = ecore_chain_get_prod_idx(&p_ll2_conn->tx_queue.txq_chain);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1811
p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1865
if (!p_ll2_conn->tx_queue.cur_send_packet)
sys/dev/qlnx/qlnxe/ecore_ll2.c
1868
p_cur_send_packet = p_ll2_conn->tx_queue.cur_send_packet;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1869
cur_send_frag_num = p_ll2_conn->tx_queue.cur_send_frag_num;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1881
p_ll2_conn->tx_queue.cur_send_frag_num++;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1883
OSAL_SPIN_LOCK_IRQSAVE(&p_ll2_conn->tx_queue.lock, flags);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1885
OSAL_SPIN_UNLOCK_IRQSAVE(&p_ll2_conn->tx_queue.lock, flags);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1978
p_ll2_conn->tx_queue.b_cb_registred = false;
sys/dev/qlnx/qlnxe/ecore_ll2.c
1980
p_ll2_conn->tx_queue.tx_sb_index);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1983
OSAL_FREE(p_hwfn->p_dev, p_ll2_conn->tx_queue.descq_array);
sys/dev/qlnx/qlnxe/ecore_ll2.c
1984
ecore_chain_free(p_hwfn->p_dev, &p_ll2_conn->tx_queue.txq_chain);
sys/dev/qlnx/qlnxe/ecore_ll2.c
2024
if (OSAL_SPIN_LOCK_ALLOC(p_hwfn, &p_ll2_info[i].tx_queue.lock))
sys/dev/qlnx/qlnxe/ecore_ll2.c
2057
OSAL_SPIN_LOCK_DEALLOC(&p_hwfn->p_ll2_info[i].tx_queue.lock);
sys/dev/qlnx/qlnxe/ecore_ll2.c
213
struct ecore_ll2_tx_queue *p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
51
#define ECORE_LL2_TX_REGISTERED(ll2) ((ll2)->tx_queue.b_cb_registred)
sys/dev/qlnx/qlnxe/ecore_ll2.c
766
struct ecore_ll2_tx_queue *p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.c
895
struct ecore_ll2_tx_queue *p_tx = &p_ll2_conn->tx_queue;
sys/dev/qlnx/qlnxe/ecore_ll2.h
124
struct ecore_ll2_tx_queue tx_queue;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_proc.c
129
__u32 rx_queue, tx_queue;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_proc.c
147
tx_queue = sdp_sk(sk)->write_seq - sdp_sk(sk)->tx_ring.una_seq;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_proc.c
151
rx_queue, tx_queue, sk->sk_state);
usr.sbin/bhyve/pci_virtio_console.c
231
pci_vtcon_port_to_vq(struct pci_vtcon_port *port, bool tx_queue)
usr.sbin/bhyve/pci_virtio_console.c
235
qnum = tx_queue ? port->vsp_txq : port->vsp_rxq;