mt76_queue
struct mt76_queue *q = dev->phy.q_tx[i];
struct mt76_queue *q = &dev->q_rx[i];
struct mt76_queue *q = &dev->q_rx[i];
mt76_dma_queue_magic_cnt_init(struct mt76_dev *dev, struct mt76_queue *q)
mt76_dma_sync_idx(struct mt76_dev *dev, struct mt76_queue *q)
void mt76_dma_queue_reset(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_add_rx_buf(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_add_buf(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_tx_cleanup_idx(struct mt76_dev *dev, struct mt76_queue *q, int idx,
mt76_dma_kick_queue(struct mt76_dev *dev, struct mt76_queue *q)
mt76_dma_tx_cleanup(struct mt76_dev *dev, struct mt76_queue *q, bool flush)
mt76_dma_get_rxdmad_c_buf(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_get_buf(struct mt76_dev *dev, struct mt76_queue *q, int idx,
mt76_dma_dequeue(struct mt76_dev *dev, struct mt76_queue *q, bool flush,
mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_tx_queue_skb(struct mt76_phy *phy, struct mt76_queue *q,
mt76_dma_rx_fill_buf(struct mt76_dev *dev, struct mt76_queue *q,
int mt76_dma_rx_fill(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_alloc_queue(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_rx_cleanup(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q = &dev->q_rx[qid];
mt76_add_fragment(struct mt76_dev *dev, struct mt76_queue *q, void *data,
mt76_dma_rx_process(struct mt76_dev *dev, struct mt76_queue *q, int budget)
int mt76_dma_rx_fill(struct mt76_dev *dev, struct mt76_queue *q,
void mt76_dma_queue_reset(struct mt76_dev *dev, struct mt76_queue *q,
mt76_dma_reset_tx_queue(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *
struct mt76_queue *hwq;
int mt76_create_page_pool(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q;
int mt76_wed_dma_setup(struct mt76_dev *dev, struct mt76_queue *q, bool reset);
static inline int mt76_wed_dma_setup(struct mt76_dev *dev, struct mt76_queue *q,
struct mt76_queue *
struct mt76_queue *q;
struct mt76_queue *q;
int mt76_npu_dma_add_buf(struct mt76_phy *phy, struct mt76_queue *q,
int mt76_npu_rx_queue_init(struct mt76_dev *dev, struct mt76_queue *q);
int mt76_npu_fill_rx_queue(struct mt76_dev *dev, struct mt76_queue *q);
void mt76_npu_queue_cleanup(struct mt76_dev *dev, struct mt76_queue *q);
void mt76_npu_queue_setup(struct mt76_dev *dev, struct mt76_queue *q);
void mt76_npu_txdesc_cleanup(struct mt76_queue *q, int index);
struct mt76_queue *q,
struct mt76_queue *q)
struct mt76_queue *q)
struct mt76_queue *q)
static inline void mt76_npu_txdesc_cleanup(struct mt76_queue *q,
void mt76_queue_tx_complete(struct mt76_dev *dev, struct mt76_queue *q,
static inline bool mt76_queue_is_rx(struct mt76_dev *dev, struct mt76_queue *q)
static inline bool mt76_queue_is_wed_tx_free(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rro(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rro_ind(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rro_rxdmad_c(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rro_data(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rro_msdu_pg(struct mt76_queue *q)
static inline bool mt76_queue_is_wed_rx(struct mt76_queue *q)
static inline bool mt76_queue_is_emi(struct mt76_queue *q)
static inline bool mt76_queue_is_npu(struct mt76_queue *q)
static inline bool mt76_queue_is_npu_tx(struct mt76_queue *q)
static inline bool mt76_queue_is_npu_rx(struct mt76_queue *q)
int mt76_create_page_pool(struct mt76_dev *dev, struct mt76_queue *q);
mt76_get_page_pool_buf(struct mt76_queue *q, u32 *offset, u32 size)
int (*alloc)(struct mt76_dev *dev, struct mt76_queue *q,
int (*tx_queue_skb)(struct mt76_phy *phy, struct mt76_queue *q,
int (*tx_queue_skb_raw)(struct mt76_dev *dev, struct mt76_queue *q,
void *(*dequeue)(struct mt76_dev *dev, struct mt76_queue *q, bool flush,
void (*tx_cleanup)(struct mt76_dev *dev, struct mt76_queue *q,
void (*rx_cleanup)(struct mt76_dev *dev, struct mt76_queue *q);
void (*kick)(struct mt76_dev *dev, struct mt76_queue *q);
void (*reset_q)(struct mt76_dev *dev, struct mt76_queue *q,
int (*rx_rro_add_msdu_page)(struct mt76_dev *dev, struct mt76_queue *q,
struct mt76_queue *q_tx[__MT_TXQ_MAX];
struct mt76_queue *q_mcu[__MT_MCUQ_MAX];
struct mt76_queue q_rx[__MT_RXQ_MAX];
struct mt76_queue *q;
mt7603_init_rx_queue(struct mt7603_dev *dev, struct mt76_queue *q,
struct mt76_queue *q;
struct mt76_queue *q = dev->mphy.q_tx[i];
struct mt76_queue *q = dev->mphy.q_tx[qid];
struct mt76_queue *q;
struct mt76_queue *q = queue_map[i].q;
mt76x02_init_rx_queue(struct mt76x02_dev *dev, struct mt76_queue *q,
struct mt76_queue *q;
struct mt76_queue *q = dev->mphy.q_tx[MT_TXQ_PSD];
struct mt76_queue *q;
struct mt76_queue *q;
struct mt76_queue *q = queue_map[i].q;
mt7915_wed_check_ppe(struct mt7915_dev *dev, struct mt76_queue *q,
struct mt76_queue *q;
struct mt76_queue *q = queue_map[i].q;
struct mt76_queue *q;
struct mt76_queue *q = queue_map[i].q;
struct mt76_queue *q = &dev->mt76.q_rx[i];
int mt7996_rro_msdu_page_add(struct mt76_dev *mdev, struct mt76_queue *q,
struct mt76_queue *q;
mt7996_wed_check_ppe(struct mt7996_dev *dev, struct mt76_queue *q,
struct mt76_queue *q;
int mt7996_rro_msdu_page_add(struct mt76_dev *mdev, struct mt76_queue *q,
struct mt76_queue *q = &dev->mt76.q_rx[MT_RXQ_MAIN_WA];
int mt76_npu_fill_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q = q_instance;
int mt76_npu_dma_add_buf(struct mt76_phy *phy, struct mt76_queue *q,
void mt76_npu_txdesc_cleanup(struct mt76_queue *q, int index)
void mt76_npu_queue_setup(struct mt76_dev *dev, struct mt76_queue *q)
int mt76_npu_rx_queue_init(struct mt76_dev *dev, struct mt76_queue *q)
void mt76_npu_queue_cleanup(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q,
struct mt76_queue *q = &dev->q_rx[qid];
static struct mt76_queue *mt76s_alloc_tx_queue(struct mt76_dev *dev)
struct mt76_queue *q;
struct mt76_queue *q;
mt76s_get_next_rx_entry(struct mt76_queue *q)
mt76s_process_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
static int mt76s_process_tx_queue(struct mt76_dev *dev, struct mt76_queue *q)
mt76s_tx_queue_skb(struct mt76_phy *phy, struct mt76_queue *q,
mt76s_tx_queue_skb_raw(struct mt76_dev *dev, struct mt76_queue *q,
static void mt76s_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q = &dev->q_rx[i];
static int mt76s_tx_run_queue(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q;
struct mt76_queue *q = &dev->q_rx[qid];
struct mt76_queue *q;
struct mt76_queue *q = phy->q_tx[qid];
struct mt76_queue *hwq = phy->q_tx[MT_TXQ_PSD];
mt76_txq_stopped(struct mt76_queue *q)
mt76_txq_send_burst(struct mt76_phy *phy, struct mt76_queue *q,
struct mt76_queue *q;
struct mt76_queue *q;
struct mt76_queue *hwq;
void mt76_queue_tx_complete(struct mt76_dev *dev, struct mt76_queue *q,
struct mt76_queue *q = phy->q_tx[0];
struct mt76_queue *q;
mt76u_fill_rx_sg(struct mt76_dev *dev, struct mt76_queue *q, struct urb *urb,
mt76u_refill_rx(struct mt76_dev *dev, struct mt76_queue *q,
mt76u_rx_urb_alloc(struct mt76_dev *dev, struct mt76_queue *q,
mt76u_get_next_rx_entry(struct mt76_queue *q)
struct mt76_queue *q = urb->context;
mt76u_process_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q = &dev->q_rx[qid];
struct mt76_queue *q = &dev->q_rx[qid];
mt76u_free_rx_queue(struct mt76_dev *dev, struct mt76_queue *q)
struct mt76_queue *q = &dev->q_rx[i];
struct mt76_queue *q = &dev->q_rx[i];
struct mt76_queue *q;
mt76u_tx_queue_skb(struct mt76_phy *phy, struct mt76_queue *q,
static void mt76u_tx_kick(struct mt76_dev *dev, struct mt76_queue *q)
mt76u_ac_to_hwq(struct mt76_dev *dev, struct mt76_queue *q, u8 qid)
struct mt76_queue *q;
struct mt76_queue *q;
struct mt76_queue *q = &dev->q_rx[MT_RXQ_MAIN];
int mt76_wed_dma_setup(struct mt76_dev *dev, struct mt76_queue *q, bool reset)