next_index
int next_index = 0; /* C1 is the default value */
next_index = idx;
return next_index;
uint64_t next_index:11;
uint64_t next_index:11;
uint64_t next_index:11;
uint64_t next_index:11;
unsigned short next_index,
if (likely(next_index <= 3)) {
loadseg(which, next_index);
if (likely(prev_index | next_index | prev_base))
loadseg(which, next_index);
if (prev_index != next_index)
loadseg(which, next_index);
loadseg(which, next_index);
unsigned int next_index;
next_index = sg->length;
while (next_index <= skip) {
index = next_index;
next_index += sg->length;
int next_index;
next_index = -1;
next_index = i;
if (next_index == -1)
if (initial_slack[next_index] > fair_pbn_alloc) {
vars[next_index].pbn += fair_pbn_alloc;
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
vars[next_index].bpp_x16 = bpp_x16_from_pbn(params[next_index], vars[next_index].pbn);
vars[next_index].pbn -= fair_pbn_alloc;
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
vars[next_index].pbn += initial_slack[next_index];
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
vars[next_index].bpp_x16 = params[next_index].bw_range.max_target_bpp_x16;
vars[next_index].pbn -= initial_slack[next_index];
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
bpp_increased[next_index] = true;
int next_index;
next_index = -1;
next_index = i;
if (next_index == -1)
DRM_DEBUG_DRIVER("MST_DSC index #%d, try no compression\n", next_index);
var_pbn = vars[next_index].pbn;
vars[next_index].pbn = kbps_to_peak_pbn(params[next_index].bw_range.stream_kbps, fec_overhead_multiplier_x1000);
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
__func__, __LINE__, next_index, ret);
vars[next_index].pbn = var_pbn;
DRM_DEBUG_DRIVER("MST_DSC index #%d, greedily disable dsc\n", next_index);
vars[next_index].dsc_enabled = false;
vars[next_index].bpp_x16 = 0;
DRM_DEBUG_DRIVER("MST_DSC index #%d, restore optimized pbn value\n", next_index);
vars[next_index].pbn = var_pbn;
params[next_index].port->mgr,
params[next_index].port,
vars[next_index].pbn);
__func__, __LINE__, next_index, ret);
tried[next_index] = true;
slab_block_number next_index = slab->search_cursor.index;
u8 *next_counter = &slab->counters[next_index];
zero_index = find_zero_byte_in_word(next_counter, next_index, end_index);
next_index += BYTES_PER_WORD;
zero_index = find_zero_byte_in_word(next_counter, next_index, end_index);
next_index += BYTES_PER_WORD;
buffer->next_index = 0;
event = &buffer->events[buffer->next_index];
buffer->next_index = (buffer->next_index + 1) % IBMASM_NUM_EVENTS;
index = buffer->next_index;
unsigned int next_index;
u32 next_index = tx_hash_table[index].next;
index = next_index;
u32 next_index = slave_info->head;
hash_table[hash_index].next = next_index;
if (next_index != TLB_NULL_INDEX)
hash_table[next_index].prev = hash_index;
u32 index, next_index;
for (; index != RLB_NULL_INDEX; index = next_index) {
next_index = rx_hash_table[index].used_next;
u32 next_index = bond_info->rx_hashtbl[index].used_next;
bond_info->rx_hashtbl_used_head = next_index;
bond_info->rx_hashtbl[prev_index].used_next = next_index;
if (next_index != RLB_NULL_INDEX)
bond_info->rx_hashtbl[next_index].used_prev = prev_index;
u32 next_index = bond_info->rx_hashtbl[index].src_next;
if (next_index != RLB_NULL_INDEX)
bond_info->rx_hashtbl[next_index].src_prev = prev_index;
bond_info->rx_hashtbl[prev_index].src_first = next_index;
bond_info->rx_hashtbl[prev_index].src_next = next_index;
u32 next_index = entry->src_next;
index = next_index;
u32 next_index = bond_info->rx_hashtbl[curr_index].used_next;
curr_index = next_index;
int next_index, i;
next_index = incr_ring(intf->tx_spb_index, DESC_RING_COUNT);
if (next_index == intf->tx_spb_clean_index)
s16 prev_index, next_index;
next_index = pkt->next;
list->head = next_index;
tx->dqo.pending_packets[prev_index].next = next_index;
if (next_index == -1) {
tx->dqo.pending_packets[next_index].prev = prev_index;
s16 next_index;
next_index = tx->dqo_compl.miss_completions.head;
while (next_index != -1) {
pending_packet = &tx->dqo.pending_packets[next_index];
next_index = pending_packet->next;
s16 next_index;
next_index = tx->dqo_compl.timed_out_completions.head;
while (next_index != -1) {
pending_packet = &tx->dqo.pending_packets[next_index];
next_index = pending_packet->next;
u32 next_index;
&next_table, &next_index,
rlen, next_table, next_index);
u8 next_index)
mlxsw_reg_qeec_next_element_index_set(payload, next_index);
enum mlxsw_reg_qeec_hr hr, u8 index, u8 next_index,
next_index);
u8 next_index, u32 maxrate, u8 burst_size)
next_index);
u8 next_index, u32 minrate)
next_index);
enum mlxsw_reg_qeec_hr hr, u8 index, u8 next_index,
u8 next_index, u32 maxrate, u8 burst_size);
u16 next_index = 0;
next_index = 1;
next_index = num_active_cmds >> ep->ec_shift;
if (next_index > num_active_cmds - event_coal_min)
next_index = num_active_cmds - event_coal_min;
if (!next_index)
next_index = 1;
cq_index = ep->qp.cqe_exp_seq_sn + next_index - 1;
return next_index;
pool->next_index = 0;
pool->next_index = 0;
pool->next_index = index;
u16 next_index;
index = pool->next_index;
if (index == pool->next_index)
pool->next_index = index == mp->pool_max_index ? 0 : index + 1;
u64 next_index;
next_index = key.offset;
next_index = found_key.offset + 1;
btrfs_set_first_dir_index_to_log(curr_inode, next_index);
u32 index = 0, next_index;
next_index = (index & ~(len - 1)) + len;
last = ((next_index >= hsize) ? 1 : 0);
index = next_index;
u32 table_index, next_index;
next_index = -1;
next_index = -1;
next_index =
next_index = le32_to_cpu(ldtentry->index);
free_index(tid, ip, table_index, next_index);
if (index >= jfs_ip->next_index) {
jfs_ip->next_index = 2;
if ((index < 2) || (index >= JFS_IP(inode)->next_index)) {
(dir_index >= JFS_IP(ip)->next_index)) {
if (loop_count++ > JFS_IP(ip)->next_index) {
JFS_IP(ip)->next_index)) {
if (jfs_ip->next_index < 2) {
jfs_ip->next_index);
jfs_ip->next_index = 2;
index = jfs_ip->next_index++;
jfs_ip->next_index--;
if (JFS_IP(ip)->next_index == DIREND) {
jfs_ip->next_index = le32_to_cpu(dip->di_next_index);
dip->di_next_index = cpu_to_le32(jfs_ip->next_index);
return (JFS_IP(inode)->next_index <= (MAX_INLINE_DIRTABLE_ENTRY + 1));
uint next_index; /* next available directory entry index */
jfs_inode->next_index = 0;
u64 *next_index, struct squashfs_page_actor *output)
if (next_index)
*next_index = index + length;
&entry->next_index, entry->actor);
*block = entry->next_index;
actor->next_index = start_index >> PAGE_SHIFT;
(actor->next_index != page_next_index(actor))) {
actor->next_index++;
actor->next_index++;
pgoff_t next_index;
u64 next_index;
block = entry->next_index;
if (!next_index)
uint32_t next_index;
next_index = 0;
next_index, XFS_DQ_LOOKUP_BATCH);
next_index = dqp->q_id + 1;
unsigned long next_index;
iter->next_index = start;
iter->next_index = iter->index;
return iter->next_index - iter->index;
iter->next_index = 0;
iter->next_index = __radix_tree_iter_add(iter, BITS_PER_LONG);
iter->next_index = iter->index;
index = iter->next_index;
iter->next_index = maxindex + 1;
iter->next_index = (index | node_maxindex(node)) + 1;
iter->index = iter->next_index;
unsigned long maxindex, start = iter->next_index;
start = next_index(start, node, offset);
iter->next_index = 1 + min(max, (start | node_maxindex(node)));
iter->next_index = 1;
unsigned long index, base, next_index, next_next_index;
next_index = round_down(base + nrpages, nrpages);
next_next_index = round_down(next_index + nrpages, nrpages);
XA_BUG_ON(xa, test_get_entry(xa, next_index) != NULL);
check_xa_multi_store_adv_add(xa, next_index, 0, &some_val_2);
XA_BUG_ON(xa, test_get_entry(xa, next_index) != &some_val_2);
check_xa_multi_store_adv_delete(xa, next_index, 0);
check_xa_multi_store_adv_add(xa, next_index, order, &some_val_2);
XA_BUG_ON(xa, test_get_entry(xa, next_index + i) != &some_val_2);
check_xa_multi_store_adv_delete(xa, next_index, order);
check_xa_multi_store_adv_add(xa, next_index, order, &some_val_2);
XA_BUG_ON(xa, test_get_entry(xa, next_index + i) != &some_val_2);
check_xa_multi_store_adv_delete(xa, next_index, order);
XA_BUG_ON(xa, test_get_entry(xa, next_index + i) != NULL);