list_last_entry
rdreg = list_last_entry(rd_regions,
last_entry = list_last_entry(&cs->staged_cs_node, struct hl_cs,
node = list_last_entry(&outcome_store->used_list,
node = list_last_entry(&outcome_store->free_list,
bo = list_last_entry(&dbc->xfer_list, struct qaic_bo, xfer_list);
page = list_last_entry(&qdev->bootlog, struct bootlog_page, node);
page = list_last_entry(&qdev->bootlog, struct bootlog_page, node);
dev_prv = list_last_entry(&drv->p->klist_devices.k_list,
dtc = list_last_entry(&info->channels,
dtc = list_last_entry(&info->channels, struct dmatest_chan, node);
vd = list_last_entry(&gchan->vc.desc_issued,
desc = list_last_entry(&chan->desc_freed, struct usb_dmac_desc,
desc = list_last_entry(&chan->active_list,
tail_desc = list_last_entry(&chan->pending_list,
tail_segment = list_last_entry(&tail_desc->segments,
desc = list_last_entry(&chan->done_list,
tail_desc = list_last_entry(&chan->pending_list,
tail_segment = list_last_entry(&tail_desc->segments,
tail_desc = list_last_entry(&chan->pending_list,
tail_segment = list_last_entry(&tail_desc->segments,
seg = list_last_entry(&desc->segments,
tail_desc = list_last_entry(&chan->pending_list,
tail_segment = list_last_entry(&tail_desc->segments,
cdma_tail_segment = list_last_entry(&tail_desc->segments,
axidma_tail_segment = list_last_entry(&tail_desc->segments,
list_last_entry(&tail_desc->segments,
segment = list_last_entry(&desc->segments,
segment = list_last_entry(&desc->segments,
segment = list_last_entry(&desc->segments,
segment = list_last_entry(&desc->segments,
desc = list_last_entry(&chan->pending_list,
desc = list_last_entry(&desc->tx_list,
prev = list_last_entry(&gpio_devices, struct gpio_device, list);
chunk = list_last_entry(&e->list, struct amdgpu_mux_chunk, entry);
chunk = list_last_entry(&e->list, struct amdgpu_mux_chunk, entry);
new_dev = list_last_entry(&topology_device_list, struct kfd_topology_device, list);
dev = list_last_entry(&topology_device_list,
dev = list_last_entry(&topology_device_list,
NULL : list_last_entry(zorder_list, typeof(*last), zlist_node);
msg = list_last_entry(&ctx->mt_queue, struct sii8620_mt_msg, node);
drm_bridge_get(list_last_entry(&encoder->bridge_chain,
block = list_last_entry(&allocated, typeof(*block), link);
while (master != top && master == list_last_entry(&master->lessor->lessees, struct drm_master, lessee_list))
rq = list_last_entry(&tl->requests, typeof(*rq), link);
struct intel_vgpu_mm *m = list_last_entry(&workload->lri_shadow_mm,
if (page == list_last_entry(&vma->page_list, typeof(*page), lru))
rq = list_last_entry(&tl->requests, typeof(*rq), link);
#define list_last_op(_ops) list_last_entry(_ops, struct bind_job_op, entry)
prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry);
return list_last_entry(&pstate->list, typeof(*cstate), head);
p = &list_last_entry(&ram->cfg, typeof(*cfg), head)->bios;
list_last_entry(&sched->job_list, typeof(*prev),
block = list_last_entry(&blocks, typeof(*block), link);
struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
struct vc4_bo *bo = list_last_entry(&vc4->bo_cache.time_list,
return list_last_entry(&vc4->render_job_list,
struct vc4_bo *bo = list_last_entry(&list,
operation = list_last_entry(&connection->operations,
req_node = list_last_entry(&req_list->list, struct request_list, list);
usage = list_last_entry(&application->mt_usages,
csdev = list_last_entry(&path->path_list, struct coresight_node, link)->csdev;
owner_info = list_last_entry(&cscfg_mgr->load_order_list,
load_list_item = list_last_entry(&cscfg_mgr->load_order_list,
struct msc_window *prev = list_last_entry(&msc->win_list,
buf = list_last_entry(&pbufl, struct irdma_puda_buf, list);
page = list_last_entry(&ent->mkeys_queue.pages_list,
last_page = list_last_entry(&ent->mkeys_queue.pages_list,
page = list_last_entry(&ent->mkeys_queue.pages_list,
struct cmdq_task *prev_task = list_last_entry(
block = list_last_entry(&journal->free_tail_blocks,
if (mddev == list_last_entry(&all_mddevs, struct mddev, all_mddevs))
struct stripe_head *sh_last = list_last_entry(
vid_cap_buf = list_last_entry(&data->vid_cap_active,
ib = list_last_entry(&aq->incoming,
msg = list_last_entry(&isys->framebuflist, struct isys_fw_msgs, head);
meta = list_last_entry(&ctx->buffer_meta_list, struct coda_buffer_meta,
dst_vb = list_last_entry(&dst_vq->done_list, struct vb2_buffer,
list_last_entry(&video->out_active,
buf = list_last_entry(&stream->post_bufs,
b = list_last_entry(&q_ctx->rdy_queue, struct v4l2_m2m_buffer, list);
last_e = list_last_entry(&gating_cfg->entries,
rx = list_last_entry(&rx_mod->rx_free_q, struct bna_rx, qe);
tx = list_last_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
last = list_last_entry(&tbl->matchers_list,
last = list_last_entry(&tbl->matchers_list,
tmp_matcher = list_last_entry(&tbl->matchers_list,
matcher = list_last_entry(&tbl->matchers_list, struct mlx5hws_matcher, list_node);
buff = list_last_entry(&dump_data->buff_list,
prev_nic_matcher = list_last_entry(&nic_tbl->nic_matcher_list,
last_ste = list_last_entry(miss_list, struct mlx5dr_ste, miss_list_node);
last_ste = list_last_entry(mlx5dr_ste_get_miss_list(ste),
last_nic_matcher = list_last_entry(&nic_tbl->nic_matcher_list,
vchunk = list_last_entry(&vregion->vchunk_list,
vregion = list_last_entry(&vgroup->vregion_list,
sublist = list_last_entry(&erif_list->erif_sublists,
mlxsw_sp_rt6 = list_last_entry(&fib6_entry->rt6_list,
mlxsw_sp_rt6 = list_last_entry(&fib6_entry->rt6_list,
elem = list_last_entry(&admin->rules, struct vcap_rule_internal,
list_last_entry(&(nfp_prog)->insns, struct nfp_insn_meta, l)
bf = list_last_entry(&bf_q, struct ath_buf, list);
if (ns != list_last_entry(&ctrl->namespaces, struct nvme_ns, list)) {
core1 = list_last_entry(&cluster->cores, struct k3_r5_core, elem);
list_last_entry(&cluster->cores, struct k3_r5_core, elem);
core = list_last_entry(&cluster->cores, struct k3_r5_core, elem);
core1 = list_last_entry(&cluster->cores, struct k3_r5_core, elem);
core1 = list_last_entry(&cluster->cores, struct k3_r5_core, elem);
allocated_sgl = list_last_entry(&lpfc_buf->dma_sgl_xtra_list,
allocated_buf = list_last_entry(&lpfc_buf->dma_cmd_rsp_list,
rdesc = list_last_entry(&sqi->bd_list_used, struct ring_desc, list);
if (last_xfer == list_last_entry(&msg->transfers,
if (xfer == list_last_entry(&msg->transfers, struct spi_transfer,
ib = list_last_entry(&aq->incoming,
ib = list_last_entry(&aq->active, struct ipu7_isys_buffer,
ib = list_last_entry(&aq->incoming, struct ipu7_isys_buffer,
msg = list_last_entry(&isys->framebuflist, struct isys_fw_msgs, head);
td = list_last_entry(&tz->trips_reached,
as = list_last_entry(list, struct async, asynclist);
usb_req = list_last_entry(&ep0->req_queue, struct vrequest, req_entry);
e = list_last_entry(&mkey->head, struct mlx5_vdpa_direct_mr, list);
vdev = list_last_entry(&dev_set->device_list,
last = list_last_entry(iova, struct vfio_iova, list);
node = list_last_entry(iova_copy, struct vfio_iova, list);
node = list_last_entry(iova_copy, struct vfio_iova, list);
last = list_last_entry(target_list,
last = list_last_entry(target_list,
range = list_last_entry(head, struct falloc_range, list);
root = list_last_entry(&fs_info->dead_roots,
root = list_last_entry(&fs_info->dead_roots,
found = list_last_entry(&table->stripe_cache,
tmp_ref = list_last_entry(list, struct recorded_ref, list);
curr = list_last_entry(delayed_ins_list, struct btrfs_delayed_item,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&ci->i_cap_flush_list,
req1 = list_last_entry(&ci->i_unsafe_dirops,
req2 = list_last_entry(&ci->i_unsafe_iops,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&ci->i_cap_snaps,
list_last_entry(&mdsc->cap_flush_list,
list_last_entry(&ci->i_cap_snaps, struct ceph_cap_snap,
e = list_last_entry(&con->writequeue, struct writequeue_entry, list);
return list_last_entry(&io_end->list_vec, struct ext4_io_end_vec, list);
jext = list_last_entry(&jd->extent_list, struct gfs2_journal_extent, list);
gh = list_last_entry(&gl->gl_holders, struct gfs2_holder, gh_list);
tr = list_last_entry(&sdp->sd_ail1_list,
last_iface = list_last_entry(&ses->iface_list,
last_iface = list_last_entry(&ses->iface_list, struct cifs_server_iface,
last_iface = list_last_entry(&ses->iface_list, struct cifs_server_iface,
last = list_last_entry(&batch->msg_list,
last = list_last_entry(&batch->msg_list,
last = list_last_entry(&send_ctx->msg_list,
last = list_last_entry(&send_ctx->msg_list,
msg = list_last_entry(&msg_list, struct smbdirect_rw_io, list);
return list_last_entry(&super_blocks, struct super_block, s_list);
wq = list_last_entry(&wqh->head, typeof(*wq), entry);
dfp = list_last_entry(&tp->t_dfops, struct xfs_defer_pending,
list_last_entry(&(ops)->list, struct drm_gpuva_op, entry)
return list_last_entry(&t->regions_list, struct damon_region, list);
list_last_entry(head, typeof(*(pos)), member) : list_prev_entry(pos, member))
for (pos = list_last_entry(head, typeof(*pos), member); \
for (pos = list_last_entry(head, typeof(*pos), member), \
list_last_entry(&(v4l2c)->links, struct v4l2_connector_link, head)
frag = list_last_entry(&xskb->pool->xskb_list, struct xdp_buff_xsk,
else if (list_last_entry(&slab_caches, struct kmem_cache, list) == prev)
struct gcov_fn_info *info = list_last_entry(¤t_info->functions,
luo_file = list_last_entry(&file_set->files_list,
luo_file = list_last_entry(&file_set->files_list,
w = list_last_entry(&lock->wait_list, struct mutex_waiter, list);
p = list_last_entry(tasks, struct task_struct, se.group_node);
worker = list_last_entry(&pool->idle_list, struct worker, entry);
worker = list_last_entry(&pool->idle_list, struct worker, entry);
res = list_last_entry(&test->resources,
return list_last_entry(&prio->item_list,
KUNIT_EXPECT_PTR_EQ(test, &test_struct2, list_last_entry(&list,
mrp = list_last_entry(&ranges,
last_filter = list_last_entry(filters, struct damos_filter, list);
page = list_last_entry(list, struct page, pcp_list);
end = list_last_entry(&vn->purge_list, struct vmap_area, list)->va_end;
end = max(end, list_last_entry(&vn->purge_list,
if (cur_instance == list_last_entry(&hdev->adv_instances,
session = list_last_entry(&jsk->sk_session_queue,
oldest = list_last_entry(head, struct sk_buff, list);
last = list_last_entry(&net_namespace_list, struct net, list);
rt = list_last_entry(&rt->fib6_siblings,
rt = list_last_entry(&rt->fib6_siblings,
last_sibling = list_last_entry(&rt->fib6_siblings,
tail = list_last_entry(&msk->backlog_list, struct sk_buff, list);
return list_last_entry(&msk->rtx_queue, struct mptcp_data_frag, list);
list_last_entry(&map->members, struct set_elem, list);
trans = list_last_entry(&nft_net->commit_list,
tail = list_last_entry(&nft_net->commit_list, struct nft_trans, list);
return list_last_entry(&peer->list, struct tipc_peer, list);
last = list_last_entry(&context->records_list,
ureq = list_last_entry(®_requests_list,
node = list_last_entry(&lru->list,
for (pos = list_last_entry(head, typeof(*pos), member); \
entry = list_last_entry(&priv->ctl_write_list, typeof(*entry), list);
if (mod == list_last_entry(&ppl->mod_list,
entry = list_last_entry(&adev->fw_list, typeof(*entry), node);
sdev = list_last_entry(&data->devices, struct snd_soc_usb_device, list);
sdev = list_last_entry(&data->devices, struct snd_soc_usb_device, list);
list_last_entry(&cfg->funcs, struct func_node, l)
list_last_entry(&func->bbs, struct bb_node, l)
(!list_empty(ptr) ? list_last_entry(ptr, type, member) : NULL)
for (pos = list_last_entry(head, typeof(*pos), member); \
for (pos = list_last_entry(head, typeof(*pos), member), \
cpu_tail = list_last_entry(&cpu_head, struct schedstat_cpu, cpu_list);
return list_last_entry(&entry->children, struct browser_entry, node);
last = list_last_entry(&browser->entries, struct browser_entry, node);
entry = list_last_entry(&browser->entries, typeof(*entry), node);
last_bb = list_last_entry(this_blocks, typeof(*last_bb), list);
if (curr == list_last_entry(sources, struct disasm_line, al.node))
next != list_last_entry(sources, struct disasm_line, al.node))
last_dl = list_last_entry(¬es->src->source,
cpb = list_last_entry(&cpr->blocks, struct call_path_block,