first_entry
bool first_entry;
first_entry = list_first_entry(&hdev->cs_mirror_list,
first_entry && cs_needs_timeout(cs)) {
static void intel_gmch_gtt_insert_pages(unsigned int first_entry,
for (i = 0, j = first_entry; i < num_entries; i++, j++) {
void intel_gmch_gtt_clear_range(unsigned int first_entry, unsigned int num_entries)
for (i = first_entry; i < (first_entry + num_entries); i++) {
uint32_t tbuf_size, max_entries, num_entries, first_entry, i;
first_entry = num_entries % max_entries;
for (i = first_entry; i < num_entries; ++i) {
for (i = 0; i < first_entry; ++i) {
static void swap_table_entries(struct _vcs_dpi_voltage_scaling_st *first_entry,
struct _vcs_dpi_voltage_scaling_st temp_entry = *first_entry;
*first_entry = *second_entry;
static void swap_table_entries(struct _vcs_dpi_voltage_scaling_st *first_entry,
struct _vcs_dpi_voltage_scaling_st temp_entry = *first_entry;
*first_entry = *second_entry;
unsigned int first_entry = vma_res->start / I915_GTT_PAGE_SIZE;
unsigned int act_pt = first_entry / GEN6_PTES;
unsigned int act_pte = first_entry % GEN6_PTES;
const unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
unsigned int pde = first_entry / GEN6_PTES;
unsigned int pte = first_entry % GEN6_PTES;
unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
(gen8_pte_t __iomem *)ggtt->gsm + first_entry;
const int max_entries = ggtt_total_entries(ggtt) - first_entry;
first_entry, num_entries, max_entries))
unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
const int max_entries = ggtt_total_entries(ggtt) - first_entry;
first_entry, num_entries, max_entries))
if (should_update_ggtt_with_bind(ggtt) && gen8_ggtt_bind_ptes(ggtt, first_entry,
unsigned int first_entry = start / I915_GTT_PAGE_SIZE;
(gen6_pte_t __iomem *)ggtt->gsm + first_entry;
const int max_entries = ggtt_total_entries(ggtt) - first_entry;
first_entry, num_entries, max_entries))
entry[0].first_entry.ctrl = CIO2_FBPT_CTRL_VALID |
entry[0].first_entry.first_page_offset = 0;
entry[0].first_entry.first_page_offset = b->offset;
remaining = length + entry[0].first_entry.first_page_offset;
if (entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID) {
} while (!(entry->first_entry.ctrl & CIO2_FBPT_CTRL_VALID));
} first_entry;
union bfi_addr_u first_entry;
(bfi_q)->first_entry.a32.addr_lo = cur_q_addr.lsb; \
(bfi_q)->first_entry.a32.addr_hi = cur_q_addr.msb; \
__le32 first_entry;
int frag, first_entry = entry;
unwind_tx_frag_mapping(cp, skb, first_entry, entry);
txd = &cp->tx_ring[first_entry];
cp->tx_opts[first_entry] = ctrl;
first_entry, entry, skb->len);
unsigned int first_entry, tx_packets;
first_entry = tx_q->cur_tx;
WARN_ON(tx_q->tx_skbuff[first_entry]);
desc = &tx_q->dma_entx[first_entry].basic;
desc = &tx_q->dma_tx[first_entry];
is_last_segment = ((tx_q->cur_tx - first_entry) &
__func__, tx_q->cur_tx, tx_q->dirty_tx, first_entry,
unsigned int first_entry, tx_packets;
first_entry = entry;
WARN_ON(tx_q->tx_skbuff[first_entry]);
__func__, tx_q->cur_tx, tx_q->dirty_tx, first_entry,
tx_q->tx_skbuff_dma[first_entry].buf = des;
tx_q->tx_skbuff_dma[first_entry].buf_type = STMMAC_TXBUF_T_SKB;
tx_q->tx_skbuff_dma[first_entry].map_as_page = false;
tx_q->tx_skbuff_dma[first_entry].len = nopaged_len;
tx_q->tx_skbuff_dma[first_entry].last_segment = last_segment;
tbs_desc = &tx_q->dma_entx[first_entry];
int frag, first_entry = entry;
txd = &gp->init_block->txd[first_entry];
u32 first_len, u32 first_entry, u32 entry)
first_entry = NEXT_TX(first_entry);
while (first_entry != entry) {
struct happy_meal_txd *this = &txbase[first_entry];
int frag, first_entry = entry;
first_entry, entry);
hme_write_txd(hp, &hp->happy_block->happy_meal_txd[first_entry],
struct nilfs_dat_entry *entry, *first_entry;
first_entry = kmap_local_folio(entry_bh->b_folio, offset);
entry = (void *)first_entry +
kunmap_local(first_entry);
for (first_entry(ctl_dir, &h, &entry); h; next_entry(&h, &entry)) {
void intel_gmch_gtt_clear_range(unsigned int first_entry, unsigned int num_entries);
struct rpc_xprt *first_entry(struct list_head *head),
return first_entry(head);