tx_desc
struct tx_desc *desc;
txd->tx_desc->addr = 0;
txd->tx_desc->len = 0;
txd->tx_desc->flags = 0;
txd->tx_desc = &rd->age_tx_ring[i];
struct tx_desc *age_tx_ring;
(sizeof(struct tx_desc) * AGE_TX_RING_CNT)
struct tx_desc *tx_desc;
struct tx_desc *desc;
#define ALC_TX_RING_ALIGN sizeof(struct tx_desc)
(sizeof(struct tx_desc) * ALC_TX_RING_CNT)
struct tx_desc *alc_tx_ring;
struct tx_desc *desc;
(sizeof(struct tx_desc) * ALE_TX_RING_CNT)
struct tx_desc *ale_tx_ring;
struct cgem_tx_desc *tx_desc;
tx_desc = (struct cgem_tx_desc *)(rx_desc + 1);
tx_desc->ctl = CGEM_TXDESC_USED | CGEM_TXDESC_WRAP;
struct tx_desc;
struct tx_desc *desc;
wr_gen2(struct tx_desc *d, unsigned int gen)
write_wr_hdr_sgl(unsigned int ndesc, struct tx_desc *txd, struct txq_state *txqs,
wr_gen2((struct tx_desc *)wp, ogen);
struct tx_desc *txd;
write_imm(struct tx_desc *d, caddr_t src,
struct tx_desc *d = &q->desc[pidx];
sizeof(struct tx_desc), sz,
struct tx_desc *txd;
struct tx_desc *txd;
memcpy(data, &qs->txq[qnum].desc[idx], sizeof(struct tx_desc));
return sizeof(struct tx_desc);
struct tx_desc *desc; /* KVA of descriptor ring */
struct tx_desc *desc;
tx_desc->mapped = true;
tx_desc->dma_addr = dma_addr;
tx_desc->tx_sg[0].addr = tx_desc->dma_addr;
tx_desc->tx_sg[0].length = ISER_HEADERS_LEN;
tx_desc->tx_sg[0].lkey = device->mr->lkey;
struct iser_tx_desc *tx_desc = &iser_pdu->desc;
if (likely(tx_desc->mapped)) {
ib_dma_unmap_single(device->ib_device, tx_desc->dma_addr,
tx_desc->mapped = false;
struct iser_tx_desc *tx_desc = &pdu->desc;
dma_addr = ib_dma_map_single(device->ib_device, (void *)tx_desc,
struct iser_tx_desc *tx_desc)
tx_desc->dma_addr, ISER_HEADERS_LEN, DMA_TO_DEVICE);
memset(&tx_desc->iser_header, 0, sizeof(struct iser_hdr));
tx_desc->iser_header.flags = ISER_VER;
tx_desc->num_sge = 1;
if (tx_desc->tx_sg[0].lkey != device->mr->lkey) {
tx_desc->tx_sg[0].lkey = device->mr->lkey;
ISER_DBG("sdesc %p lkey mismatch, fixing", tx_desc);
struct iser_tx_desc *tx_desc = &iser_pdu->desc;
tx_desc->type = ISCSI_TX_SCSI_COMMAND;
iser_create_send_desc(iser_conn, tx_desc);
err = iser_post_send(&iser_conn->ib_conn, tx_desc,
iser_snd_completion(struct iser_tx_desc *tx_desc,
struct icl_iser_pdu *iser_pdu = container_of(tx_desc, struct icl_iser_pdu, desc);
if (tx_desc && tx_desc->type == ISCSI_TX_CONTROL)
struct iser_tx_desc *tx_desc;
tx_desc = (struct iser_tx_desc *)(uintptr_t)wc->wr_id;
iser_snd_completion(tx_desc, ib_conn);
int iser_post_send(struct ib_conn *ib_conn, struct iser_tx_desc *tx_desc,
tx_desc->dma_addr, ISER_HEADERS_LEN,
send_wr.wr_id = (uintptr_t)tx_desc;
send_wr.sg_list = tx_desc->tx_sg;
send_wr.num_sge = tx_desc->num_sge;
desc = txd->tx_desc;
status = le32toh(txd->tx_desc->flags);
le32toh(txd->tx_desc->buflen) &
txd->tx_desc = &rd->jme_tx_ring[i];
struct jme_desc *tx_desc;
struct mlx4_en_tx_desc *tx_desc = (struct mlx4_en_tx_desc *)
volatile __be32 *ptr = (__be32 *)tx_desc;
volatile struct mlx4_en_tx_desc *tx_desc;
tx_desc = (volatile struct mlx4_en_tx_desc *)
dseg = &tx_desc->data;
tx_desc->ctrl.srcrb_flags = CTRL_FLAGS;
tx_desc->ctrl.srcrb_flags |= cpu_to_be32(MLX4_WQE_CTRL_IP_CSUM);
tx_desc->ctrl.srcrb_flags |= cpu_to_be32(MLX4_WQE_CTRL_TCP_UDP_CSUM);
if (likely(tx_desc->ctrl.srcrb_flags != CTRL_FLAGS)) {
tx_desc->ctrl.vlan_tag = cpu_to_be16(mb->m_pkthdr.ether_vtag);
tx_desc->ctrl.ins_vlan = MLX4_WQE_CTRL_INS_CVLAN;
tx_desc->ctrl.vlan_tag = 0;
tx_desc->ctrl.ins_vlan = 0;
m_copydata(mb, 0, 2, __DEVOLATILE(void *, &tx_desc->ctrl.srcrb_flags16[0]));
m_copydata(mb, 2, 4, __DEVOLATILE(void *, &tx_desc->ctrl.imm));
tx_desc->ctrl.imm = 0;
tx_desc->lso.mss_hdr_size = cpu_to_be32((mss << 16) | ihs);
ds_cnt = (dseg - ((volatile struct mlx4_wqe_data_seg *)tx_desc)) + nr_segs;
tx_desc->ctrl.fence_size = (ds_cnt & 0x3f);
(tx_desc->ctrl.ins_vlan != MLX4_WQE_CTRL_INS_CVLAN)) {
*(volatile __be32 *) (&tx_desc->ctrl.vlan_tag) |= cpu_to_be32(ring->doorbell_qpn);
tx_desc->ctrl.owner_opcode = opcode;
(volatile unsigned long *) &tx_desc->ctrl, bf_size);
tx_desc->ctrl.owner_opcode = opcode;
tx_desc->buffer_addr = htole64(buffer_info->dma); //XXX
q81_txb_desc_t *tx_desc;
tx_desc = ha->tx_ring[txr_idx].tx_buf[txr_next].oal_vaddr;
tx_desc->baddr = segs->ds_addr;
tx_desc->length = segs->ds_len;
tx_desc->flags = Q81_RXB_DESC_FLAGS_E;
tx_desc->flags = 0;
tx_desc++;
void *tx_desc = (uint8_t *)tx_ring->desc + sc->txdesc_len * i;
rtwn_pci_setup_tx_desc(pc, tx_desc, next_desc_addr);
rtwn_pci_copy_tx_desc(pc, txd, tx_desc);
uint8_t *tx_desc, int id)
rtwn_pci_copy_tx_desc(pc, txd, tx_desc);
struct mbuf *m, uint8_t *tx_desc, uint8_t type, int id)
error = rtwn_pci_tx_start_beacon(sc, m, tx_desc, id);
error = rtwn_pci_tx_start_frame(sc, ni, m, tx_desc, type);
struct mbuf *m, uint8_t *tx_desc, uint8_t type)
struct mbuf *m, uint8_t *tx_desc, uint8_t type, int id)
txd = (struct rtwn_tx_desc_common *)tx_desc;
rtwn_dump_tx_desc(sc, tx_desc);
memcpy(data->buf, tx_desc, sc->txdesc_len);
struct tsec_desc *tx_desc;
tx_desc = &sc->tsec_tx_vaddr[tx_idx];
if (tx_desc->flags & TSEC_TXBD_R) {
struct tsec_desc *tx_desc = sc->tsec_tx_vaddr;
tx_desc[i].bufptr = 0;
tx_desc[i].length = 0;
tx_desc[i].flags = ((i == TSEC_TX_NUM_DESC - 1) ?
struct tsec_desc *tx_desc;
tx_desc = &sc->tsec_tx_vaddr[tx_idx];
tx_desc->length = segs[i].ds_len;
tx_desc->bufptr = segs[i].ds_addr;
tx_desc->flags = (tx_idx == (TSEC_TX_NUM_DESC - 1) ?
txd->tx_desc = &rd->vge_tx_ring[i];
txd->tx_desc->vge_frag[0].vge_addrhi = 0;
txd->tx_desc->vge_sts = htole32(m->m_pkthdr.len << 16);
txd->tx_desc->vge_ctl = htole32(cflags | ((nsegs + 1) << 28) |
frag = &txd->tx_desc->vge_frag[i];
txd->tx_desc->vge_ctl |= htole32(VGE_TDCTL_TIC);
txd->tx_desc->vge_sts |= htole32(VGE_TDSTS_OWN);
txd->tx_desc->vge_frag[0].vge_addrhi |= htole32(VGE_TXDESC_Q);
struct vge_tx_desc *tx_desc;
txd->tx_desc->dtlen = htole16(VTE_TX_LEN(txsegs[0].ds_len));
txd->tx_desc->dtbp = htole32(txsegs[0].ds_addr);
txd->tx_desc->dtst = htole16(VTE_DTST_TX_OWN);
status = le16toh(txd->tx_desc->dtst);
txd->tx_desc = desc;
struct vte_tx_desc *tx_desc;