NAPI_POLL_WEIGHT
ret = m_can_rx_handler(dev, NAPI_POLL_WEIGHT, ir);
NAPI_POLL_WEIGHT);
#define CMP_QUEUE_CQE_THRESH (NAPI_POLL_WEIGHT / 2)
#define MAX_RX_POST NAPI_POLL_WEIGHT /* Frags posted at a time */
static_assert(NAPI_POLL_WEIGHT >= MAX_TX_DESC_NEEDED);
to_do = min_t(u32, to_do, NAPI_POLL_WEIGHT);
1, NAPI_POLL_WEIGHT);
1, NAPI_POLL_WEIGHT);
pfvf->hw.cq_ecount_wait = (cur_moder.pkts > NAPI_POLL_WEIGHT) ?
NAPI_POLL_WEIGHT : cur_moder.pkts;
#define MLX4_EN_CACHE_SIZE (2 * NAPI_POLL_WEIGHT)
NAPI_POLL_WEIGHT);
#define TX_WORK_PER_LOOP NAPI_POLL_WEIGHT
#define RX_WORK_PER_LOOP NAPI_POLL_WEIGHT
thresh = tx_ring->count - NAPI_POLL_WEIGHT;
for (j = 0; j < NAPI_POLL_WEIGHT; j++)
if (j < NAPI_POLL_WEIGHT) {
if (cleaned_count++ == NAPI_POLL_WEIGHT) {
ch_budget = NAPI_POLL_WEIGHT / cpsw->tx_ch_num;
ch_budget = (consumed_rate * NAPI_POLL_WEIGHT) / max_rate;
ch_budget = (NAPI_POLL_WEIGHT - ch_budget) /
budget = NAPI_POLL_WEIGHT;
txv[i].budget = (ch_rate * NAPI_POLL_WEIGHT) / max_rate;
budget = NAPI_POLL_WEIGHT;
if (unlikely(filled > NAPI_POLL_WEIGHT)) {
if (!more || queue_len > NAPI_POLL_WEIGHT)
if (ec->rx_max_coalesced_frames > NAPI_POLL_WEIGHT)
tun->rx_batched = NAPI_POLL_WEIGHT;
static int napi_weight = NAPI_POLL_WEIGHT;
napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0;
napi_weight = ec->tx_max_coalesced_frames ? NAPI_POLL_WEIGHT : 0;
int budget = NAPI_POLL_WEIGHT;
if (epoll_params.busy_poll_budget > NAPI_POLL_WEIGHT &&
netif_napi_add_weight(dev, napi, poll, NAPI_POLL_WEIGHT);
netif_napi_add_weight_locked(dev, napi, poll, NAPI_POLL_WEIGHT);
netif_napi_add_weight_locked(dev, napi, poll, NAPI_POLL_WEIGHT);
netif_napi_add_tx_weight(dev, napi, poll, NAPI_POLL_WEIGHT);
if (packets >= NAPI_POLL_WEIGHT || empty) {
batch_size = NAPI_POLL_WEIGHT;
if (weight > NAPI_POLL_WEIGHT)