queue_len
dma_addr_t *dma_handle, u16 *queue_len);
static inline int queue_ci_get(atomic_t *ci, u32 queue_len)
return atomic_read(ci) & ((queue_len << 1) - 1);
static inline int queue_free_slots(struct hl_hw_queue *q, u32 queue_len)
int delta = (q->pi - queue_ci_get(&q->ci, queue_len));
return (queue_len - delta);
return (abs(delta) - queue_len);
u16 *queue_len)
*queue_len = q->pq_size / QMAN_PQ_ENTRY_SIZE;
u32 ci, qm_ci_stream_off, queue_len;
queue_len = (q->queue_type == QUEUE_TYPE_INT) ?
ci = gaudi_queue_idx_dec(ci, queue_len);
ci = gaudi_queue_idx_dec(ci, queue_len);
dma_addr_t *dma_handle, u16 *queue_len)
*queue_len = MME_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
*queue_len = TPC_QMAN_LENGTH;
dma_addr_t *dma_handle, u16 *queue_len);
static DEVICE_ATTR_RW(queue_len);
static DEVICE_ATTR_RW(queue_len);
bcom_ata_init(int queue_len, int maxbufsize)
tsk = bcom_task_alloc(queue_len, sizeof(struct bcom_ata_bd), 0);
bcom_fec_tx_init(int queue_len, phys_addr_t fifo)
tsk = bcom_task_alloc(queue_len, sizeof(struct bcom_fec_bd),
bcom_fec_rx_init(int queue_len, phys_addr_t fifo, int maxbufsize)
tsk = bcom_task_alloc(queue_len, sizeof(struct bcom_fec_bd),
bcom_gen_bd_tx_init(int queue_len, phys_addr_t fifo,
tsk = bcom_task_alloc(queue_len, sizeof(struct bcom_gen_bd),
struct bcom_task * bcom_psc_gen_bd_rx_init(unsigned psc_num, int queue_len,
return bcom_gen_bd_rx_init(queue_len, fifo,
bcom_psc_gen_bd_tx_init(unsigned psc_num, int queue_len, phys_addr_t fifo)
return bcom_gen_bd_tx_init(queue_len, fifo,
bcom_gen_bd_rx_init(int queue_len, phys_addr_t fifo,
tsk = bcom_task_alloc(queue_len, sizeof(struct bcom_gen_bd),
int queue_len;
queue_len = skb_queue_len(&offload->skb_queue);
if (queue_len > offload->skb_queue_len_max / 8)
__func__, queue_len);
int queue_len;
queue_len = skb_queue_len(&offload->skb_queue);
if (queue_len > offload->skb_queue_len_max / 8)
__func__, queue_len);
adapter->rx_queue.queue_len = 1;
adapter->rx_queue.queue_len = sizeof(struct ibmveth_rx_q_entry) *
dma_alloc_coherent(dev, adapter->rx_queue.queue_len,
adapter->rx_queue.queue_len;
dma_free_coherent(dev, adapter->rx_queue.queue_len,
dma_free_coherent(dev, adapter->rx_queue.queue_len,
u32 queue_len;
int queue_len;
queue_len = skb_queue_len(queue);
if (!more || queue_len > NAPI_POLL_WEIGHT)
u32 queue_len;
queue_len = skb_queue_len(&ring->queue);
if (queue_len == 0 || queue_id == BEACON_QUEUE ||
u32 queue_len;
queue_len = skb_queue_len(&wl->tx_queue);
res = scnprintf(buf, sizeof(buf), "%u\n", queue_len);
int i, result_index, num_complete = 0, queue_len;
queue_len = skb_queue_len(&wl->tx_queue);
if ((num_complete > 0) && (queue_len > 0)) {
queue_len <= WL1251_TX_QUEUE_LOW_WATERMARK) {
u32 queue_len;
queue_len = wl1271_tx_total_queue_count(wl);
res = scnprintf(buf, sizeof(buf), "%u\n", queue_len);
__le32 queue_len;
__le32 queue_len;
extern struct bcom_task * bcom_ata_init(int queue_len, int maxbufsize);
bcom_fec_rx_init(int queue_len, phys_addr_t fifo, int maxbufsize);
bcom_fec_tx_init(int queue_len, phys_addr_t fifo);
bcom_gen_bd_rx_init(int queue_len, phys_addr_t fifo,
bcom_gen_bd_tx_init(int queue_len, phys_addr_t fifo,
struct bcom_task * bcom_psc_gen_bd_rx_init(unsigned psc_num, int queue_len,
struct bcom_task * bcom_psc_gen_bd_tx_init(unsigned psc_num, int queue_len,
u32 queue_len = atomic_read(&mrt->cache_resolve_queue_len);
nla_put_u32(skb, IPMRA_TABLE_CACHE_RES_QUEUE_LEN, queue_len) ||
int printed = 2, queue_len = 0;
queue_len = 0;
printed += queue_len;
queue_len = 0;
if (queue_len == context)
++queue_len;
int nr_free = QUEUE_SIZE - queue_len(q);
if (queue_len(&printqueue) > 0) {