refill
if(!refill()) {
struct refill *refill = container_of(op, struct refill, op);
struct keybuf *buf = refill->buf;
if (bkey_cmp(k, refill->end) > 0) {
if (refill->pred(buf, k)) {
refill->nr_found++;
struct refill refill;
bch_btree_op_init(&refill.op, -1);
refill.nr_found = 0;
refill.buf = buf;
refill.end = end;
refill.pred = pred;
bch_btree_map_keys(&refill.op, c, &buf->last_scanned,
trace_bcache_keyscan(refill.nr_found,
refill:
goto refill;
static unsigned int geth_fill_freeq(struct gemini_ethernet *geth, bool refill)
pn = (refill ? rw.bits.wptr : rw.bits.rptr) >> fpp_order;
goto refill;
refill:
ring->desc_cb[i].refill = 0;
ring->desc_cb[i].refill = 1;
ring->desc_cb[i].refill = 1;
ring->desc_cb[i].refill = 1;
if (unlikely(ntc == ntu && !ring->desc_cb[ntc].refill))
ring->desc_cb[ring->next_to_clean].refill = 0;
u16 refill;
u16 *refill)
refill[tc] = tc_config[tc].path[direction].data_credits_refill;
static int ixgbe_ieee_credits(__u8 *bw, __u16 *refill,
u16 refill[MAX_TRAFFIC_CLASS];
ixgbe_dcb_unpack_refill(dcb_config, DCB_TX_CONFIG, refill);
return ixgbe_dcb_hw_config_82598(hw, pfc_en, refill, max,
return ixgbe_dcb_hw_config_82599(hw, pfc_en, refill, max,
__u16 refill[IEEE_8021QAZ_MAX_TCS], max[IEEE_8021QAZ_MAX_TCS];
ixgbe_ieee_credits(ets->tc_tx_bw, refill, max, max_frame);
return ixgbe_dcb_hw_ets_config(hw, refill, max,
u16 *refill, u16 *max, u8 *bwg_id,
ixgbe_dcb_config_rx_arbiter_82598(hw, refill, max,
ixgbe_dcb_config_tx_desc_arbiter_82598(hw, refill, max,
ixgbe_dcb_config_tx_data_arbiter_82598(hw, refill, max,
ixgbe_dcb_config_rx_arbiter_82599(hw, refill, max,
ixgbe_dcb_config_tx_desc_arbiter_82599(hw, refill, max,
ixgbe_dcb_config_tx_data_arbiter_82599(hw, refill, max, bwg_id,
refill[i] = val;
int ixgbe_dcb_hw_ets_config(struct ixgbe_hw *hw, u16 *refill, u16 *max,
int ixgbe_dcb_config_tx_data_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
reg = refill[i];
int ixgbe_dcb_config_rx_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_hw_config_82598(struct ixgbe_hw *hw, u8 pfc_en, u16 *refill,
ixgbe_dcb_config_rx_arbiter_82598(hw, refill, max, prio_type);
ixgbe_dcb_config_tx_desc_arbiter_82598(hw, refill, max,
ixgbe_dcb_config_tx_data_arbiter_82598(hw, refill, max,
credit_refill = refill[i];
int ixgbe_dcb_config_tx_desc_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
reg |= refill[i];
int ixgbe_dcb_config_rx_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_config_tx_desc_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_config_tx_data_arbiter_82598(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_hw_config_82598(struct ixgbe_hw *hw, u8 pfc_en, u16 *refill,
u16 *refill,
reg = refill[i];
u16 *refill,
int ixgbe_dcb_hw_config_82599(struct ixgbe_hw *hw, u8 pfc_en, u16 *refill,
ixgbe_dcb_config_rx_arbiter_82599(hw, refill, max, bwg_id,
ixgbe_dcb_config_tx_desc_arbiter_82599(hw, refill, max,
ixgbe_dcb_config_tx_data_arbiter_82599(hw, refill, max,
credit_refill = refill[i];
u16 *refill,
reg |= refill[i];
int ixgbe_dcb_config_rx_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_config_tx_desc_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_config_tx_data_arbiter_82599(struct ixgbe_hw *hw, u16 *refill,
int ixgbe_dcb_hw_config_82599(struct ixgbe_hw *hw, u8 pfc_en, u16 *refill,
u16 refill[MAX_TRAFFIC_CLASS], max[MAX_TRAFFIC_CLASS];
ixgbe_dcb_unpack_refill(dcb_cfg, DCB_TX_CONFIG, refill);
ixgbe_dcb_hw_ets_config(hw, refill, max, bwg_id,
int rx_proc = 0, rx_todo, refill, size = 0;
refill = mvneta_rx_refill_queue(pp, rxq);
mvneta_rxq_desc_num_update(pp, rxq, rx_proc, refill);
int refill;
refill = min_t(u16, rq->wqe.info.refill_unit, remaining);
mlx5e_free_rx_wqes(rq, ix + total_alloc, refill);
refill_alloc = mlx5e_alloc_rx_wqes(rq, ix + total_alloc, refill);
if (unlikely(refill_alloc != refill))
remaining -= refill;
for (int i = 0; i < total_alloc + refill; i++) {
rq_stats->pp_alloc_refill = stats.alloc_stats.refill;
goto refill;
refill:
goto refill;
goto refill;
goto refill;
goto refill;
goto refill;
refill:
bool refill)
if (refill)
writel(mmu_hw->refill, mmu_hw->base + MMU_REG_AXI_REFILL_IF_ID);
.refill = 0x00002524,
.refill = 0x00002726,
.refill = 0x00002322,
.refill = 0x00002120,
.refill = 0x00002120,
.refill = 0x00002726,
.refill = 0x00002524,
.refill = 0x00002120,
.refill = 0x00002120,
.refill = 0x00002322,
.refill = 0x00002928,
.refill = 0x00002726,
.refill = 0x00002524,
.refill = 0x00002322,
.refill = 0x00002120,
.refill = 0x00002322,
.refill = 0x00002726,
.refill = 0x00002322,
.refill = 0x00002524,
.refill = 0x00002120,
.refill = 0x00002726,
.refill = 0x00002726,
.refill = 0x00002524,
.refill = 0x00002120,
.refill = 0x00002322,
u32 refill;
u64 refill;
unsigned int refill = -tctx->cached_refs + IO_TCTX_REFS_CACHE_NR;
percpu_counter_add(&tctx->inflight, refill);
refcount_add(refill, ¤t->usage);
tctx->cached_refs += refill;
unsigned long refill;
refill = mas->node_request;
if (kmem_cache_sheaf_size(mas->sheaf) >= refill) {
if (mt_refill_sheaf(gfp, &mas->sheaf, refill))
refill:
goto refill;
goto refill;
*data++ = pool_stats->alloc_stats.refill;
alloc_stat_inc(pool, refill);
stats->alloc_stats.refill += pool->alloc_stats.refill;
stats.alloc_stats.refill) ||
refill:
goto refill;
int refill;
refill = kmem_cache_alloc_bulk(s, gfp, size - sheaf->size,
if (!refill)
sheaf->size += refill;