arch/arm/xen/p2m.c
162
p2m_entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys);
arch/arm/xen/p2m.c
42
entry = rb_entry(parent, struct xen_p2m_entry, rbnode_phys);
arch/arm/xen/p2m.c
73
entry = rb_entry(n, struct xen_p2m_entry, rbnode_phys);
arch/powerpc/kernel/eeh_cache.c
106
piar = rb_entry(n, struct pci_io_addr_range, rb_node);
arch/powerpc/kernel/eeh_cache.c
128
piar = rb_entry(parent, struct pci_io_addr_range, rb_node);
arch/powerpc/kernel/eeh_cache.c
221
piar = rb_entry(n, struct pci_io_addr_range, rb_node);
arch/powerpc/kernel/eeh_cache.c
271
piar = rb_entry(n, struct pci_io_addr_range, rb_node);
arch/powerpc/kernel/eeh_cache.c
60
piar = rb_entry(n, struct pci_io_addr_range, rb_node);
arch/powerpc/perf/hv-24x7.c
633
it = rb_entry(*new, struct event_uniq, node);
arch/sh/kernel/dwarf.c
323
cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node);
arch/sh/kernel/dwarf.c
359
fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node);
arch/sh/kernel/dwarf.c
840
cie_tmp = rb_entry(*rb_node, struct dwarf_cie, node);
arch/sh/kernel/dwarf.c
926
fde_tmp = rb_entry(*rb_node, struct dwarf_fde, node);
arch/x86/events/intel/uncore.c
1086
unit = rb_entry(node, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore.c
1369
unit = rb_entry(node, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
103
unit = rb_entry(b, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
126
unit = rb_entry(pos, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
132
unit = rb_entry(pos, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
160
a_node = rb_entry(a, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
161
b_node = rb_entry(b, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
183
unit = rb_entry(node, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
33
rb_entry((cur), struct intel_uncore_discovery_type, node)
arch/x86/events/intel/uncore_discovery.c
446
pos = rb_entry(node, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_discovery.c
753
type = rb_entry(node, struct intel_uncore_discovery_type, node);
arch/x86/events/intel/uncore_snbep.c
6118
pos = rb_entry(node, struct intel_uncore_discovery_unit, node);
arch/x86/events/intel/uncore_snbep.c
6346
unit = rb_entry(node, struct intel_uncore_discovery_unit, node);
block/bfq-iosched.c
2700
__bfqq = rb_entry(parent, struct bfq_queue, pos_node);
block/bfq-iosched.c
2711
__bfqq = rb_entry(node, struct bfq_queue, pos_node);
block/bfq-iosched.c
743
bfqq = rb_entry(parent, struct bfq_queue, pos_node);
block/bfq-wf2q.c
1303
entry = rb_entry(node, struct bfq_entity, rb_node);
block/bfq-wf2q.c
1309
entry = rb_entry(node->rb_left,
block/bfq-wf2q.c
27
return rb_entry(node, struct bfq_entity, rb_node);
block/bfq-wf2q.c
328
entity = rb_entry(node, struct bfq_entity, rb_node);
block/bfq-wf2q.c
387
entry = rb_entry(parent, struct bfq_entity, rb_node);
block/bfq-wf2q.c
416
child = rb_entry(node, struct bfq_entity, rb_node);
block/bfq-wf2q.c
432
struct bfq_entity *entity = rb_entry(node, struct bfq_entity, rb_node);
block/blk-throttle.c
29
#define rb_entry_tg(node) rb_entry((node), struct throtl_grp, rb_node)
block/elevator.c
227
__rq = rb_entry(parent, struct request, rb_node);
block/elevator.c
254
rq = rb_entry(n, struct request, rb_node);
block/elevator.h
213
#define rb_entry_rq(node) rb_entry((node), struct request, rb_node)
drivers/android/binder.c
1026
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
1051
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
1132
ref = rb_entry(parent, struct binder_ref, rb_node_node);
drivers/android/binder.c
1159
ref = rb_entry(parent, struct binder_ref, rb_node_desc);
drivers/android/binder.c
5257
thread = rb_entry(parent, struct binder_thread, rb_node);
drivers/android/binder.c
5594
struct binder_node *node = rb_entry(n, struct binder_node,
drivers/android/binder.c
5618
thread = rb_entry(n, struct binder_thread, rb_node);
drivers/android/binder.c
5635
node = rb_entry(n, struct binder_node, rb_node);
drivers/android/binder.c
588
thread = rb_entry(n, struct binder_thread, rb_node);
drivers/android/binder.c
6164
struct binder_thread *thread = rb_entry(n, struct binder_thread, rb_node);
drivers/android/binder.c
6295
thread = rb_entry(n, struct binder_thread, rb_node);
drivers/android/binder.c
6307
node = rb_entry(n, struct binder_node, rb_node);
drivers/android/binder.c
6327
ref = rb_entry(n, struct binder_ref, rb_node_desc);
drivers/android/binder.c
6635
print_binder_thread_ilocked(m, rb_entry(n, struct binder_thread,
drivers/android/binder.c
6639
struct binder_node *node = rb_entry(n, struct binder_node,
drivers/android/binder.c
6655
print_binder_ref_olocked(m, rb_entry(n,
drivers/android/binder.c
6821
struct binder_ref *ref = rb_entry(n, struct binder_ref,
drivers/android/binder.c
707
node = rb_entry(n, struct binder_node, rb_node);
drivers/android/binder.c
754
node = rb_entry(parent, struct binder_node, rb_node);
drivers/android/binder_alloc.c
1048
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
114
buffer = rb_entry(parent, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
136
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
438
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
447
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
488
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
537
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
563
buffer = rb_entry(best_fit, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
89
buffer = rb_entry(parent, struct binder_buffer, rb_node);
drivers/android/binder_alloc.c
974
buffer = rb_entry(n, struct binder_buffer, rb_node);
drivers/android/tests/binder_alloc_kunit.c
434
buf = rb_entry(n, struct binder_buffer, rb_node);
drivers/base/regmap/regcache-rbtree.c
107
rbnode_tmp = rb_entry(*new, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
150
n = rb_entry(node, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
213
rbtree_node = rb_entry(next, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
404
rbnode_tmp = rb_entry(node, struct regcache_rbtree_node,
drivers/base/regmap/regcache-rbtree.c
479
rbnode = rb_entry(node, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
521
rbnode = rb_entry(node, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache-rbtree.c
80
rbnode = rb_entry(node, struct regcache_rbtree_node, node);
drivers/base/regmap/regcache.c
453
rb_entry(node, struct regmap_range_node, node);
drivers/base/regmap/regmap-debugfs.c
642
range_node = rb_entry(next, struct regmap_range_node, node);
drivers/base/regmap/regmap.c
507
rb_entry(*new, struct regmap_range_node, node);
drivers/base/regmap/regmap.c
531
rb_entry(node, struct regmap_range_node, node);
drivers/base/regmap/regmap.c
551
range_node = rb_entry(next, struct regmap_range_node, node);
drivers/block/drbd/drbd_interval.c
12
struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb);
drivers/block/drbd/drbd_interval.c
128
rb_entry(node, struct drbd_interval, rb);
drivers/block/drbd/drbd_interval.c
157
i = rb_entry(node, struct drbd_interval, rb);
drivers/block/drbd/drbd_interval.c
34
rb_entry(*new, struct drbd_interval, rb);
drivers/block/drbd/drbd_interval.c
76
rb_entry(node, struct drbd_interval, rb);
drivers/dma-buf/sw_sync.c
274
other = rb_entry(parent, typeof(*pt), node);
drivers/dma-buf/sw_sync.c
295
parent ? &rb_entry(parent, typeof(*pt), node)->link : &obj->pt_list);
drivers/gpu/drm/drm_buddy.c
66
return node ? rb_entry(node, struct drm_buddy_block, rb) : NULL;
drivers/gpu/drm/drm_mm.c
178
parent = rb_entry(rb, struct drm_mm_node, rb);
drivers/gpu/drm/drm_mm.c
197
parent = rb_entry(rb, struct drm_mm_node, rb);
drivers/gpu/drm/drm_mm.c
218
return rb_entry(rb, struct drm_mm_node, rb_hole_size)->hole_size;
drivers/gpu/drm/drm_mm.c
254
parent = rb_entry(rb_parent, struct drm_mm_node, rb_hole_addr);
drivers/gpu/drm/drm_mm.c
313
rb_entry(rb, struct drm_mm_node, rb_hole_size);
drivers/gpu/drm/drm_prime.c
117
pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb);
drivers/gpu/drm/drm_prime.c
132
pos = rb_entry(rb, struct drm_prime_member, handle_rb);
drivers/gpu/drm/drm_prime.c
153
member = rb_entry(rb, struct drm_prime_member, handle_rb);
drivers/gpu/drm/drm_prime.c
175
member = rb_entry(rb, struct drm_prime_member, dmabuf_rb);
drivers/gpu/drm/drm_prime.c
198
member = rb_entry(rb, struct drm_prime_member, handle_rb);
drivers/gpu/drm/drm_vma_manager.c
153
node = rb_entry(iter, struct drm_mm_node, rb);
drivers/gpu/drm/drm_vma_manager.c
264
entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb);
drivers/gpu/drm/drm_vma_manager.c
368
entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb);
drivers/gpu/drm/drm_vma_manager.c
409
entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
667
rb_entry(rb, typeof(*mmo), offset);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
696
pos = rb_entry(rb, typeof(*pos), offset);
drivers/gpu/drm/i915/gt/intel_engine_user.c
25
rb_entry(p, typeof(*it), uabi_node);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1029
rb_entry(rb, typeof(*ve), nodes[engine->id].rb);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
259
return rb_entry(rb, struct i915_priolist, node);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
300
return rb ? rb_entry(rb, struct ve_node, rb)->prio : INT_MIN;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3200
rb_entry(rb, typeof(*ve), nodes[engine->id].rb);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3863
other = rb_entry(rb, typeof(*other), rb);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4114
struct i915_priolist *p = rb_entry(rb, typeof(*p), node);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4136
rb_entry(rb, typeof(*ve), nodes[engine->id].rb);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
406
return rb_entry(rb, struct i915_priolist, node);
drivers/gpu/drm/i915/gvt/kvmgt.c
217
itr = rb_entry(node, struct gvt_dma, dma_addr_node);
drivers/gpu/drm/i915/gvt/kvmgt.c
235
itr = rb_entry(node, struct gvt_dma, gfn_node);
drivers/gpu/drm/i915/gvt/kvmgt.c
267
itr = rb_entry(parent, struct gvt_dma, gfn_node);
drivers/gpu/drm/i915/gvt/kvmgt.c
282
itr = rb_entry(parent, struct gvt_dma, dma_addr_node);
drivers/gpu/drm/i915/gvt/kvmgt.c
317
dma = rb_entry(node, struct gvt_dma, gfn_node);
drivers/gpu/drm/i915/i915_active.c
305
node = rb_entry(parent, struct active_node, node);
drivers/gpu/drm/i915/i915_active.c
33
#define fetch_node(x) rb_entry(READ_ONCE(x), typeof(struct active_node), node)
drivers/gpu/drm/i915/i915_active.c
780
rb_entry(p, struct active_node, node);
drivers/gpu/drm/i915/i915_active.c
800
rb_entry(p, struct active_node, node);
drivers/gpu/drm/i915/i915_active.c
835
return rb_entry(p, struct active_node, node);
drivers/gpu/drm/i915/i915_active.c
947
it = rb_entry(parent, struct active_node, node);
drivers/gpu/drm/i915/i915_scheduler.c
36
return rb_entry(rb, struct i915_priolist, node);
drivers/gpu/drm/i915/i915_vma.c
243
pos = rb_entry(rb, struct i915_vma, obj_node);
drivers/gpu/drm/i915/i915_vma.c
295
struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node);
drivers/gpu/drm/nouveau/nvkm/core/object.c
39
object = rb_entry(node, typeof(*object), node);
drivers/gpu/drm/nouveau/nvkm/core/object.c
83
struct nvkm_object *this = rb_entry(*ptr, typeof(*this), node);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1064
struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1778
struct nvkm_vma *this = rb_entry(temp, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1794
struct nvkm_vma *this = rb_entry(node, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
880
struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
922
struct nvkm_vma *this = rb_entry(*ptr, typeof(*this), tree);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
942
struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree);
drivers/gpu/drm/scheduler/sched_main.c
139
struct drm_sched_entity *ent_a = rb_entry((a), struct drm_sched_entity, rb_tree_node);
drivers/gpu/drm/scheduler/sched_main.c
140
struct drm_sched_entity *ent_b = rb_entry((b), struct drm_sched_entity, rb_tree_node);
drivers/gpu/drm/scheduler/sched_main.c
322
entity = rb_entry(rb, struct drm_sched_entity, rb_tree_node);
drivers/gpu/drm/scheduler/sched_main.c
338
return rb ? rb_entry(rb, struct drm_sched_entity, rb_tree_node) : NULL;
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
876
res = rb_entry(rb_itr, struct vmw_resource, mob_node);
drivers/infiniband/core/cm.c
643
cur_cm_id_priv = rb_entry(parent, struct cm_id_private,
drivers/infiniband/core/cm.c
685
cm_id_priv = rb_entry(node, struct cm_id_private, service_node);
drivers/infiniband/core/cm.c
714
cur_timewait_info = rb_entry(parent, struct cm_timewait_info,
drivers/infiniband/core/cm.c
742
timewait_info = rb_entry(node, struct cm_timewait_info,
drivers/infiniband/core/cm.c
773
cur_timewait_info = rb_entry(parent, struct cm_timewait_info,
drivers/infiniband/core/cm.c
802
cur_cm_id_priv = rb_entry(parent, struct cm_id_private,
drivers/infiniband/core/multicast.c
137
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/core/multicast.c
161
cur_group = rb_entry(parent, struct mcast_group, node);
drivers/infiniband/core/multicast.c
780
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/core/uverbs_cmd.c
522
scan = rb_entry(parent, struct xrcd_table_entry, node);
drivers/infiniband/core/uverbs_cmd.c
547
entry = rb_entry(p, struct xrcd_table_entry, node);
drivers/infiniband/hw/hfi1/mmu_rb.c
101
rbnode = rb_entry(node, struct mmu_rb_node, node);
drivers/infiniband/hw/mlx4/cm.c
161
rb_entry(node, struct id_map_entry, node);
drivers/infiniband/hw/mlx4/cm.c
219
ent = rb_entry(parent, struct id_map_entry, node);
drivers/infiniband/hw/mlx4/cm.c
551
rb_entry(rb_first(sl_id_map),
drivers/infiniband/hw/mlx4/cm.c
563
rb_entry(nd, struct id_map_entry, node);
drivers/infiniband/hw/mlx4/mcg.c
1113
group = rb_entry(p, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
1242
group = rb_entry(p, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
170
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
193
cur_group = rb_entry(parent, struct mcast_group, node);
drivers/infiniband/hw/mlx5/mr.c
1000
ent = rb_entry(node, struct mlx5_cache_ent, node);
drivers/infiniband/hw/mlx5/mr.c
1028
ent = rb_entry(node, struct mlx5_cache_ent, node);
drivers/infiniband/hw/mlx5/mr.c
495
ent = rb_entry(node, struct mlx5_cache_ent, node);
drivers/infiniband/hw/mlx5/mr.c
677
cur = rb_entry(*new, struct mlx5_cache_ent, node);
drivers/infiniband/hw/mlx5/mr.c
708
cur = rb_entry(node, struct mlx5_cache_ent, node);
drivers/infiniband/hw/mlx5/mr.c
949
ent = rb_entry(node, struct mlx5_cache_ent, node);
drivers/infiniband/sw/rdmavt/mcast.c
119
mcast = rb_entry(n, struct rvt_mcast, rb_node);
drivers/infiniband/sw/rdmavt/mcast.c
165
tmcast = rb_entry(pn, struct rvt_mcast, rb_node);
drivers/infiniband/sw/rdmavt/mcast.c
325
mcast = rb_entry(n, struct rvt_mcast, rb_node);
drivers/infiniband/sw/rxe/rxe_mcast.c
134
mcg = rb_entry(node, struct rxe_mcg, node);
drivers/infiniband/sw/rxe/rxe_mcast.c
91
tmp = rb_entry(node, struct rxe_mcg, node);
drivers/infiniband/ulp/ipoib/ipoib_main.c
606
path = rb_entry(n, struct ipoib_path, rb_node);
drivers/infiniband/ulp/ipoib/ipoib_main.c
632
tpath = rb_entry(pn, struct ipoib_path, rb_node);
drivers/infiniband/ulp/ipoib/ipoib_main.c
703
path = rb_entry(n, struct ipoib_path, rb_node);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
1010
mcast = rb_entry(n, struct ipoib_mcast, rb_node);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
167
mcast = rb_entry(n, struct ipoib_mcast, rb_node);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
192
tmcast = rb_entry(pn, struct ipoib_mcast, rb_node);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1812
rb_entry(rhs, struct arm_smmu_stream, node);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1826
&rb_entry(lhs, struct arm_smmu_stream, node)->id, rhs);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1839
return rb_entry(node, struct arm_smmu_stream, node)->master;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3542
rb_entry(existing, struct arm_smmu_stream, node)
drivers/iommu/intel/iommu.c
107
rb_entry(lhs, struct device_domain_info, node);
drivers/iommu/intel/iommu.c
133
info = rb_entry(node, struct device_domain_info, node);
drivers/iommu/intel/iommu.c
92
rb_entry(node, struct device_domain_info, node);
drivers/iommu/iova.c
34
return rb_entry(node, struct iova, node);
drivers/macintosh/via-pmu.c
2133
struct rb_entry *rp = &pp->rb_buf[pp->rb_put];
drivers/macintosh/via-pmu.c
2188
struct rb_entry *rp = &pp->rb_buf[i];
drivers/md/dm-bio-prison-v1.c
149
rb_entry(*new, struct dm_bio_prison_cell, node);
drivers/md/dm-bio-prison-v2.c
125
rb_entry(*new, struct dm_bio_prison_cell_v2, node);
drivers/md/dm-crypt.c
1931
#define crypt_io_from_node(node) rb_entry((node), struct dm_crypt_io, rb_node)
drivers/md/dm-dust.c
300
bblk = rb_entry(node, struct badblock, node);
drivers/md/dm-dust.c
41
struct badblock *bblk = rb_entry(node, struct badblock, node);
drivers/md/dm-dust.c
62
bblk = rb_entry(parent, struct badblock, node);
drivers/md/dm-snap.c
1750
pe = rb_entry(next, struct dm_snap_pending_exception,
drivers/md/dm-snap.c
1766
pe2 = rb_entry(*p, struct dm_snap_pending_exception, out_of_order_node);
drivers/md/dm-thin.c
2115
#define thin_pbd(node) rb_entry((node), struct dm_thin_endio_hook, rb_node)
drivers/md/persistent-data/dm-transaction-manager.c
116
rb_entry(*node, struct shadow_info, node);
drivers/md/persistent-data/dm-transaction-manager.c
151
rb_entry(*node, struct shadow_info, node);
drivers/md/persistent-data/dm-transaction-manager.c
172
rb_entry(tm->buckets[i].rb_node, struct shadow_info, node);
drivers/mtd/mtdswap.c
211
cur = rb_entry(parent, struct swap_eb, rb);
drivers/mtd/mtdswap.c
438
median = rb_entry(medrb, struct swap_eb, rb)->erase_count;
drivers/mtd/mtdswap.c
576
eb = rb_entry(rb_first(clean_root), struct swap_eb, rb);
drivers/mtd/mtdswap.c
79
#define MTDSWAP_ECNT_MIN(rbroot) (rb_entry(rb_first(rbroot), struct swap_eb, \
drivers/mtd/mtdswap.c
81
#define MTDSWAP_ECNT_MAX(rbroot) (rb_entry(rb_last(rbroot), struct swap_eb, \
drivers/mtd/mtdswap.c
860
eb = rb_entry(rb_first(rp), struct swap_eb, rb);
drivers/mtd/ubi/attach.c
113
av = rb_entry(parent, struct ubi_ainf_volume, rb);
drivers/mtd/ubi/attach.c
1289
aeb = rb_entry(this, struct ubi_ainf_peb, u.rb);
drivers/mtd/ubi/attach.c
1346
av = rb_entry(rb, struct ubi_ainf_volume, rb);
drivers/mtd/ubi/attach.c
595
aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb);
drivers/mtd/ubi/eba.c
215
le = rb_entry(p, struct ubi_ltree_entry, rb);
drivers/mtd/ubi/eba.c
281
le1 = rb_entry(parent, struct ubi_ltree_entry, rb);
drivers/mtd/ubi/fastmap-wl.c
404
e = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/fastmap-wl.c
557
e = rb_entry(rb_next(root->rb_node),
drivers/mtd/ubi/fastmap.c
213
tmp_aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb);
drivers/mtd/ubi/fastmap.c
252
aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb);
drivers/mtd/ubi/wl.c
1061
e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
148
e1 = rb_entry(parent, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
1648
e = rb_entry(rb, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
250
e1 = rb_entry(p, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
333
e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
340
e1 = rb_entry(p, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
368
first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
369
last = rb_entry(rb_last(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
372
e = rb_entry(root->rb_node, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
741
e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
762
e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, u.rb);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
240
struct res_common *res = rb_entry(node, struct res_common,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
259
struct res_common *this = rb_entry(*new, struct res_common,
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
5031
fs_rule = rb_entry(p, struct res_fs_rule, com.node);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
142
tfp = rb_entry(parent, struct fw_page, rb_node);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
184
tfp = rb_entry(tmp, struct fw_page, rb_node);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
449
struct fw_page *fwp = rb_entry(p, struct fw_page, rb_node);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
510
fwp = rb_entry(p, struct fw_page, rb_node);
drivers/platform/surface/aggregator/controller.c
2576
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
2590
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
2620
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
330
entry = rb_entry(*link, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
391
entry = rb_entry(node, struct ssam_nf_refcount_entry, node);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
122
bo = rb_entry(n, struct hmm_buffer_object, node);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
147
bo = rb_entry(n, struct hmm_buffer_object, node);
drivers/staging/media/atomisp/pci/hmm/hmm_bo.c
69
this = rb_entry(node, struct hmm_buffer_object, node);
drivers/vfio/vfio_iommu_type1.c
1331
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
1493
dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
1505
dma = rb_entry(n,
drivers/vfio/vfio_iommu_type1.c
174
struct vfio_dma *dma = rb_entry(node, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
1826
dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
1898
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
1908
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
198
struct vfio_dma *dma = rb_entry(node, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
224
dma = rb_entry(parent, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
2434
vfio_remove_dma(iommu, rb_entry(node, struct vfio_dma, node));
drivers/vfio/vfio_iommu_type1.c
2446
dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
2450
struct vfio_pfn *vpfn = rb_entry(p, struct vfio_pfn,
drivers/vfio/vfio_iommu_type1.c
274
struct vfio_pfn *vpfn = rb_entry(p, struct vfio_pfn, node);
drivers/vfio/vfio_iommu_type1.c
286
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
297
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
305
struct vfio_dma *dma = rb_entry(p,
drivers/vfio/vfio_iommu_type1.c
322
struct vfio_dma *dma = rb_entry(n, struct vfio_dma, node);
drivers/vfio/vfio_iommu_type1.c
343
vpfn = rb_entry(node, struct vfio_pfn, node);
drivers/vfio/vfio_iommu_type1.c
369
vpfn = rb_entry(parent, struct vfio_pfn, node);
drivers/vfio/vfio_iommu_type1.c
655
vpfn = rb_entry(prev, struct vfio_pfn, node);
drivers/vfio/vfio_iommu_type1.c
662
vpfn = rb_entry(next, struct vfio_pfn, node);
drivers/xen/evtchn.c
117
this = rb_entry(*new, struct user_evtchn, node);
drivers/xen/evtchn.c
150
evtchn = rb_entry(node, struct user_evtchn, node);
drivers/xen/evtchn.c
676
evtchn = rb_entry(node, struct user_evtchn, node);
fs/afs/callback.c
128
volume = rb_entry(p, struct afs_volume, cell_node);
fs/afs/cell.c
293
cursor = rb_entry(parent, struct afs_cell, net_node);
fs/afs/cell.c
69
cell = rb_entry(p, struct afs_cell, net_node);
fs/afs/cell.c
909
struct afs_cell *cell = rb_entry(cursor, struct afs_cell, net_node);
fs/afs/server.c
47
server = rb_entry(p, struct afs_server, uuid_rb);
fs/afs/server.c
519
server = rb_entry(rb, struct afs_server, uuid_rb);
fs/afs/server.c
85
server = rb_entry(p, struct afs_server, uuid_rb);
fs/afs/vl_alias.c
147
pvol = afs_get_volume(rb_entry(p->volumes.rb_node,
fs/afs/volume.c
32
p = rb_entry(parent, struct afs_volume, cell_node);
fs/btrfs/backref.c
1591
ref = rb_entry(node, struct prelim_ref, rbnode);
fs/btrfs/backref.c
257
rb_entry(new, struct prelim_ref, rbnode);
fs/btrfs/backref.c
259
rb_entry(exist, struct prelim_ref, rbnode);
fs/btrfs/backref.c
301
struct prelim_ref *ref = rb_entry(exist, struct prelim_ref, rbnode);
fs/btrfs/backref.c
3256
upper = rb_entry(rb_node, struct btrfs_backref_node, rb_node);
fs/btrfs/backref.c
3423
upper = rb_entry(rb_node, struct btrfs_backref_node,
fs/btrfs/backref.c
454
ref = rb_entry(parent, struct prelim_ref, rbnode);
fs/btrfs/backref.c
755
ref = rb_entry(rnode, struct prelim_ref, rbnode);
fs/btrfs/backref.c
846
ref = rb_entry(node, struct prelim_ref, rbnode);
fs/btrfs/backref.c
898
node = rb_entry(n, struct btrfs_delayed_ref_node,
fs/btrfs/block-group.c
193
rb_entry(new, struct btrfs_block_group, cache_node);
fs/btrfs/block-group.c
195
rb_entry(exist, struct btrfs_block_group, cache_node);
fs/btrfs/block-group.c
241
cache = rb_entry(n, struct btrfs_block_group, cache_node);
fs/btrfs/block-group.c
2559
map = rb_entry(node, struct btrfs_chunk_map, rb_node);
fs/btrfs/block-group.c
304
cache = rb_entry(node, struct btrfs_block_group, cache_node);
fs/btrfs/block-group.c
4683
block_group = rb_entry(n, struct btrfs_block_group,
fs/btrfs/block-group.c
4908
bg = rb_entry(node_bg, struct btrfs_block_group, cache_node);
fs/btrfs/block-group.c
4909
map = rb_entry(node_chunk, struct btrfs_chunk_map, rb_node);
fs/btrfs/defrag.c
176
entry = rb_entry(parent, struct inode_defrag, rb_node);
fs/btrfs/defrag.c
66
const struct inode_defrag *new_defrag = rb_entry(new, struct inode_defrag, rb_node);
fs/btrfs/defrag.c
67
const struct inode_defrag *existing_defrag = rb_entry(existing, struct inode_defrag, rb_node);
fs/btrfs/defrag.c
89
entry = rb_entry(node, struct inode_defrag, rb_node);
fs/btrfs/delayed-inode.c
369
const struct btrfs_delayed_item *delayed_item = rb_entry(node,
fs/btrfs/delayed-inode.c
403
rb_entry(new, struct btrfs_delayed_item, rb_node);
fs/btrfs/delayed-ref.c
1253
entry = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
1295
ref = rb_entry(n, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
323
rb_entry(new, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
325
rb_entry(exist, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
397
next = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
452
ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/delayed-ref.c
569
ref = rb_entry(rb_first_cached(&head->ref_tree),
fs/btrfs/dev-replace.c
831
map = rb_entry(node, struct btrfs_chunk_map, rb_node);
fs/btrfs/disk-io.c
1188
root = rb_entry(node, struct btrfs_root, rb_node);
fs/btrfs/disk-io.c
691
const struct btrfs_root *a = rb_entry(a_node, struct btrfs_root, rb_node);
fs/btrfs/disk-io.c
692
const struct btrfs_root *b = rb_entry(b_node, struct btrfs_root, rb_node);
fs/btrfs/disk-io.c
700
const struct btrfs_root *root = rb_entry(node, struct btrfs_root, rb_node);
fs/btrfs/extent-io-tree.c
246
entry = rb_entry(prev, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
298
entry = rb_entry(*node, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
435
entry = rb_entry(parent, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
520
entry = rb_entry(parent, struct extent_state, rb_node);
fs/btrfs/extent-tree.c
2185
entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
fs/btrfs/extent-tree.c
2190
entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
fs/btrfs/extent-tree.c
2196
entry = rb_entry(n, struct btrfs_delayed_ref_node, rb_node);
fs/btrfs/extent-tree.c
2352
ref = rb_entry(node, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/extent-tree.c
2675
bg = rb_entry(leftmost, struct btrfs_block_group, cache_node);
fs/btrfs/extent_map.c
101
entry = rb_entry(parent, struct extent_map, rb_node);
fs/btrfs/extent_map.c
114
entry = rb_entry(parent, struct extent_map, rb_node);
fs/btrfs/extent_map.c
1166
em = rb_entry(node, struct extent_map, rb_node);
fs/btrfs/extent_map.c
121
entry = rb_entry(parent, struct extent_map, rb_node);
fs/btrfs/extent_map.c
124
entry = rb_entry(parent, struct extent_map, rb_node);
fs/btrfs/extent_map.c
151
entry = rb_entry(n, struct extent_map, rb_node);
fs/btrfs/extent_map.c
166
prev_entry = rb_entry(prev, struct extent_map, rb_node);
fs/btrfs/extent_map.c
179
prev_entry = rb_entry(prev, struct extent_map, rb_node);
fs/btrfs/extent_map.c
182
prev_entry = rb_entry(prev, struct extent_map, rb_node);
fs/btrfs/extent_map.c
539
em = rb_entry(rb_node, struct extent_map, rb_node);
fs/btrfs/extent_map.c
784
em = rb_entry(node, struct extent_map, rb_node);
fs/btrfs/free-space-cache.c
1099
e = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
1606
info = rb_entry(parent, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
1687
entry = rb_entry(node, struct btrfs_free_space, bytes_index);
fs/btrfs/free-space-cache.c
1688
exist = rb_entry(parent, struct btrfs_free_space, bytes_index);
fs/btrfs/free-space-cache.c
1710
entry = rb_entry(n, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
1736
entry = rb_entry(n, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
1750
prev = rb_entry(n, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
1768
entry = rb_entry(n, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
1782
prev = rb_entry(n, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
1800
entry = rb_entry(n, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
2031
entry = rb_entry(node, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
2034
entry = rb_entry(node, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
2194
bitmap_info = rb_entry(next, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
2347
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
2466
left_info = rb_entry(right_prev, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
2933
info = rb_entry(n, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
2998
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3073
info = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3248
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3259
entry = rb_entry(node, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
3273
entry = rb_entry(node, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
3447
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3457
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3488
entry = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
3752
entry = rb_entry(node, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
4166
block_group = rb_entry(node, struct btrfs_block_group, cache_node);
fs/btrfs/free-space-cache.c
4347
tmp = rb_entry(n, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
4361
tmp = rb_entry(n, struct btrfs_free_space,
fs/btrfs/free-space-cache.c
71
info = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
907
info = rb_entry(n, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-tree.c
1193
block_group = rb_entry(node, struct btrfs_block_group,
fs/btrfs/free-space-tree.c
1260
bg = rb_entry(node, struct btrfs_block_group, cache_node);
fs/btrfs/free-space-tree.c
1353
block_group = rb_entry(node, struct btrfs_block_group,
fs/btrfs/inode.c
10156
entry = rb_entry(parent, struct btrfs_swapfile_pin, node);
fs/btrfs/inode.c
10188
sp = rb_entry(node, struct btrfs_swapfile_pin, node);
fs/btrfs/inode.c
5559
state = rb_entry(node, struct extent_state, rb_node);
fs/btrfs/misc.h
133
entry = rb_entry(node, struct rb_simple_node, rb_node);
fs/btrfs/misc.h
161
entry = rb_entry(node, struct rb_simple_node, rb_node);
fs/btrfs/misc.h
182
struct rb_simple_node *new_entry = rb_entry(new, struct rb_simple_node, rb_node);
fs/btrfs/misc.h
183
struct rb_simple_node *existing_entry = rb_entry(existing, struct rb_simple_node, rb_node);
fs/btrfs/ordered-data.c
100
prev_entry = rb_entry(prev, struct btrfs_ordered_extent,
fs/btrfs/ordered-data.c
1005
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
1042
ordered = rb_entry(n, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
106
prev_entry = rb_entry(test, struct btrfs_ordered_extent,
fs/btrfs/ordered-data.c
1070
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
1105
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
1134
entry = rb_entry(prev, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
1139
entry = rb_entry(next, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
135
entry = rb_entry(inode->ordered_tree_last, struct btrfs_ordered_extent,
fs/btrfs/ordered-data.c
45
entry = rb_entry(parent, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
499
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
511
entry = rb_entry(node, struct btrfs_ordered_extent,
fs/btrfs/ordered-data.c
584
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
74
entry = rb_entry(n, struct btrfs_ordered_extent, rb_node);
fs/btrfs/ordered-data.c
92
prev_entry = rb_entry(test, struct btrfs_ordered_extent,
fs/btrfs/ordered-data.c
975
entry = rb_entry(node, struct btrfs_ordered_extent, rb_node);
fs/btrfs/qgroup.c
166
const struct btrfs_qgroup *qgroup = rb_entry(node, struct btrfs_qgroup, node);
fs/btrfs/qgroup.c
188
const struct btrfs_qgroup *new_qgroup = rb_entry(new, struct btrfs_qgroup, node);
fs/btrfs/qgroup.c
214
return rb_entry(node, struct btrfs_qgroup, node);
fs/btrfs/qgroup.c
4020
qgroup = rb_entry(n, struct btrfs_qgroup, node);
fs/btrfs/qgroup.c
4130
entry = rb_entry(node, struct ulist_node, rb_node);
fs/btrfs/qgroup.c
4138
entry = rb_entry(rb_prev(&entry->rb_node), struct ulist_node,
fs/btrfs/qgroup.c
4147
entry = rb_entry(node, struct ulist_node, rb_node);
fs/btrfs/qgroup.c
4708
const struct btrfs_qgroup_swapped_block *block = rb_entry(node,
fs/btrfs/qgroup.c
4721
const struct btrfs_qgroup_swapped_block *new_block = rb_entry(new,
fs/btrfs/qgroup.c
4799
entry = rb_entry(node, struct btrfs_qgroup_swapped_block, node);
fs/btrfs/qgroup.c
4861
block = rb_entry(node, struct btrfs_qgroup_swapped_block, node);
fs/btrfs/qgroup.c
650
qgroup = rb_entry(node, struct btrfs_qgroup, node);
fs/btrfs/qgroup.c
681
qgroup = rb_entry(n, struct btrfs_qgroup, node);
fs/btrfs/ref-verify.c
118
const struct root_entry *entry = rb_entry(node, struct root_entry, node);
fs/btrfs/ref-verify.c
130
const struct root_entry *new_entry = rb_entry(new, struct root_entry, node);
fs/btrfs/ref-verify.c
167
struct ref_entry *new_entry = rb_entry(new, struct ref_entry, node);
fs/btrfs/ref-verify.c
168
struct ref_entry *existing_entry = rb_entry(existing, struct ref_entry, node);
fs/btrfs/ref-verify.c
225
re = rb_entry(n, struct root_entry, node);
fs/btrfs/ref-verify.c
231
ref = rb_entry(n, struct ref_entry, node);
fs/btrfs/ref-verify.c
627
ref = rb_entry(n, struct ref_entry, node);
fs/btrfs/ref-verify.c
635
re = rb_entry(n, struct root_entry, node);
fs/btrfs/ref-verify.c
81
const struct block_entry *entry = rb_entry(node, struct block_entry, node);
fs/btrfs/ref-verify.c
898
be = rb_entry(n, struct block_entry, node);
fs/btrfs/ref-verify.c
918
entry = rb_entry(n, struct block_entry, node);
fs/btrfs/ref-verify.c
93
const struct block_entry *new_entry = rb_entry(new, struct block_entry, node);
fs/btrfs/ref-verify.c
945
be = rb_entry(n, struct block_entry, node);
fs/btrfs/relocation.c
1919
block = rb_entry(rb_node, struct tree_block, rb_node);
fs/btrfs/relocation.c
317
node = rb_entry(rb_node, struct mapping_node, rb_node);
fs/btrfs/relocation.c
524
node = rb_entry(rb_node, struct mapping_node, rb_node);
fs/btrfs/relocation.c
565
node = rb_entry(rb_node, struct mapping_node, rb_node);
fs/btrfs/send.c
3073
entry = rb_entry(parent, struct orphan_dir_info, node);
fs/btrfs/send.c
3106
entry = rb_entry(n, struct orphan_dir_info, node);
fs/btrfs/send.c
3297
entry = rb_entry(parent, struct waiting_dir_move, node);
fs/btrfs/send.c
3320
entry = rb_entry(n, struct waiting_dir_move, node);
fs/btrfs/send.c
3367
entry = rb_entry(parent, struct pending_dir_move, node);
fs/btrfs/send.c
3415
entry = rb_entry(n, struct pending_dir_move, node);
fs/btrfs/send.c
4101
const struct recorded_ref *ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4116
const struct recorded_ref *entry = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4618
const struct recorded_ref *ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4637
const struct recorded_ref *entry = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4702
ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
4731
ref = rb_entry(node, struct recorded_ref, node);
fs/btrfs/send.c
8249
pm = rb_entry(n, struct pending_dir_move, node);
fs/btrfs/send.c
8266
dm = rb_entry(n, struct waiting_dir_move, node);
fs/btrfs/send.c
8277
odi = rb_entry(n, struct orphan_dir_info, node);
fs/btrfs/tests/extent-io-tests.c
101
state = rb_entry(node, struct extent_state, rb_node);
fs/btrfs/tests/extent-map-tests.c
24
em = rb_entry(node, struct extent_map, rb_node);
fs/btrfs/tests/extent-map-tests.c
572
struct extent_map *entry = rb_entry(n, struct extent_map, rb_node);
fs/btrfs/tests/free-space-tests.c
861
entry = rb_entry(node, struct btrfs_free_space, bytes_index);
fs/btrfs/tests/free-space-tests.c
884
entry = rb_entry(node, struct btrfs_free_space, bytes_index);
fs/btrfs/tests/free-space-tests.c
939
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/tests/free-space-tests.c
964
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/tests/free-space-tests.c
979
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/tree-mod-log.c
127
tm = rb_entry(node, struct tree_mod_elem, node);
fs/btrfs/tree-mod-log.c
159
cur = rb_entry(*new, struct tree_mod_elem, node);
fs/btrfs/tree-mod-log.c
523
cur = rb_entry(node, struct tree_mod_elem, node);
fs/btrfs/tree-mod-log.c
932
tm = rb_entry(next, struct tree_mod_elem, node);
fs/btrfs/ulist.c
135
const struct ulist_node *unode = rb_entry(node, struct ulist_node, rb_node);
fs/btrfs/ulist.c
164
const struct ulist_node *unode = rb_entry(new, struct ulist_node, rb_node);
fs/btrfs/volumes.c
2010
map = rb_entry(n, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
3225
map = rb_entry(node, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
3245
prev_map = rb_entry(prev, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
3250
prev_map = rb_entry(prev, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
3253
prev_map = rb_entry(prev, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
5757
rb_entry(new, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
5759
rb_entry(exist, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
6135
map = rb_entry(node, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
8334
map = rb_entry(node, struct btrfs_chunk_map, rb_node);
fs/btrfs/volumes.c
8496
sp = rb_entry(node, struct btrfs_swapfile_pin, node);
fs/ceph/addr.c
2379
perm = rb_entry(*p, struct ceph_pool_perm, node);
fs/ceph/addr.c
2413
perm = rb_entry(parent, struct ceph_pool_perm, node);
fs/ceph/addr.c
2611
perm = rb_entry(n, struct ceph_pool_perm, node);
fs/ceph/caps.c
1093
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
1376
struct ceph_cap *cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
2128
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
448
cap = rb_entry(n, struct ceph_cap, ci_node);
fs/ceph/caps.c
481
cap = rb_entry(parent, struct ceph_cap, ci_node);
fs/ceph/caps.c
824
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
855
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
908
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/caps.c
935
cap = rb_entry(q, struct ceph_cap,
fs/ceph/caps.c
974
cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/debugfs.c
62
req = rb_entry(rp, struct ceph_mds_request, r_node);
fs/ceph/inode.c
314
frag = rb_entry(parent, struct ceph_inode_frag, node);
fs/ceph/inode.c
3205
struct ceph_cap *cap = rb_entry(p, struct ceph_cap, ci_node);
fs/ceph/inode.c
349
rb_entry(n, struct ceph_inode_frag, node);
fs/ceph/inode.c
532
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
562
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
593
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
775
frag = rb_entry(n, struct ceph_inode_frag, node);
fs/ceph/mds_client.c
1219
rb_entry(p, struct ceph_mds_request, r_node);
fs/ceph/mds_client.c
1415
cap = rb_entry(rb_first(&ci->i_caps), struct ceph_cap, ci_node);
fs/ceph/mds_client.c
1803
req = rb_entry(p, struct ceph_mds_request, r_node);
fs/ceph/mds_client.c
2617
return rb_entry(rb_first(&mdsc->request_tree),
fs/ceph/mds_client.c
3688
req = rb_entry(p, struct ceph_mds_request, r_node);
fs/ceph/mds_client.c
4505
req = rb_entry(p, struct ceph_mds_request, r_node);
fs/ceph/mds_client.c
4843
rb_entry(p, struct ceph_snap_realm, node);
fs/ceph/mds_client.c
5887
nextreq = rb_entry(n, struct ceph_mds_request, r_node);
fs/ceph/quota.c
191
qri = rb_entry(node, struct ceph_quotarealm_inode, node);
fs/ceph/snap.c
1200
exist = rb_entry(*p, struct ceph_snapid_map, node);
fs/ceph/snap.c
1239
exist = rb_entry(*p, struct ceph_snapid_map, node);
fs/ceph/snap.c
1326
sm = rb_entry(p, struct ceph_snapid_map, node);
fs/ceph/snap.c
160
r = rb_entry(n, struct ceph_snap_realm, node);
fs/ceph/snap.c
95
r = rb_entry(parent, struct ceph_snap_realm, node);
fs/ceph/xattr.c
583
xattr = rb_entry(parent, struct ceph_inode_xattr, node);
fs/ceph/xattr.c
682
xattr = rb_entry(parent, struct ceph_inode_xattr, node);
fs/ceph/xattr.c
748
xattr = rb_entry(p, struct ceph_inode_xattr, node);
fs/ceph/xattr.c
773
xattr = rb_entry(p, struct ceph_inode_xattr, node);
fs/ceph/xattr.c
925
xattr = rb_entry(p, struct ceph_inode_xattr, node);
fs/eventpoll.c
1079
struct epitem *epi = rb_entry(rbp, struct epitem, rbn);
fs/eventpoll.c
1183
epi = rb_entry(rbp, struct epitem, rbn);
fs/eventpoll.c
1205
epi = rb_entry(rbp, struct epitem, rbn);
fs/eventpoll.c
1394
epic = rb_entry(parent, struct epitem, rbn);
fs/eventpoll.c
2079
epi = rb_entry(rbp, struct epitem, rbn);
fs/eventpoll.c
914
epi = rb_entry(rbp, struct epitem, rbn);
fs/eventpoll.c
930
epi = rb_entry(rbp, struct epitem, rbn);
fs/ext2/balloc.c
1027
next_rsv = rb_entry(next, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
219
rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
306
rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
323
rsv = rb_entry(n, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
349
this = rb_entry(parent, struct ext2_reserve_window_node, rsv_node);
fs/ext2/balloc.c
792
rsv = rb_entry(next,struct ext2_reserve_window_node,rsv_node);
fs/ext4/block_validity.c
103
entry = rb_entry(node, struct ext4_system_zone, node);
fs/ext4/block_validity.c
115
entry = rb_entry(node, struct ext4_system_zone, node);
fs/ext4/block_validity.c
137
entry = rb_entry(node, struct ext4_system_zone, node);
fs/ext4/block_validity.c
321
entry = rb_entry(n, struct ext4_system_zone, node);
fs/ext4/block_validity.c
79
entry = rb_entry(parent, struct ext4_system_zone, node);
fs/ext4/dir.c
495
fname = rb_entry(parent, struct fname, rb_hash);
fs/ext4/dir.c
613
fname = rb_entry(info->curr_node, struct fname, rb_hash);
fs/ext4/dir.c
621
fname = rb_entry(info->curr_node, struct fname,
fs/ext4/extents_status.c
1123
es1 = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
1147
es1 = rb_entry(node, struct extent_status,
fs/ext4/extents_status.c
1206
rc->left_es = node ? rb_entry(node,
fs/ext4/extents_status.c
1316
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
1328
return node ? rb_entry(node, struct pending_reservation,
fs/ext4/extents_status.c
1390
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
1397
es = node ? rb_entry(node, struct extent_status,
fs/ext4/extents_status.c
1410
es = rb_entry(node, struct extent_status,
fs/ext4/extents_status.c
1455
pr = rb_entry(node, struct pending_reservation,
fs/ext4/extents_status.c
1567
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
1585
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
1934
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
1983
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
2005
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
2053
pr = rb_entry(node, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
2090
pr = rb_entry(parent, struct pending_reservation, rb_node);
fs/ext4/extents_status.c
222
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
273
es = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
287
return node ? rb_entry(node, struct extent_status, rb_node) :
fs/ext4/extents_status.c
341
es1 = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
634
es1 = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
658
es1 = rb_entry(node, struct extent_status, rb_node);
fs/ext4/extents_status.c
852
es = rb_entry(parent, struct extent_status, rb_node);
fs/ext4/mballoc.c
4354
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4402
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4426
right_pa = rb_entry(tmp,
fs/ext4/mballoc.c
4436
left_pa = rb_entry(tmp,
fs/ext4/mballoc.c
4452
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4472
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4918
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4932
tmp_pa = rb_entry(tmp, struct ext4_prealloc_space,
fs/ext4/mballoc.c
4959
tmp_pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
5235
iter_pa = rb_entry(*iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
5237
new_pa = rb_entry(new, struct ext4_prealloc_space,
fs/ext4/mballoc.c
5641
pa = rb_entry(iter, struct ext4_prealloc_space,
fs/ext4/mballoc.c
6417
prev = rb_entry(node, struct ext4_free_data, efd_node);
fs/ext4/mballoc.c
6433
next = rb_entry(node, struct ext4_free_data, efd_node);
fs/ext4/mballoc.c
6471
entry = rb_entry(parent, struct ext4_free_data, efd_node);
fs/f2fs/extent_cache.c
195
en = rb_entry(node, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
244
en = rb_entry(*pnode, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
259
en = rb_entry(parent, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
392
en = rb_entry(node, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
599
en = rb_entry(parent, struct extent_node, rb_node);
fs/f2fs/gc.c
441
cur_ve = rb_entry(cur, struct victim_entry, rb_node);
fs/f2fs/gc.c
442
next_ve = rb_entry(next, struct victim_entry, rb_node);
fs/f2fs/gc.c
464
ve = rb_entry(node, struct victim_entry, rb_node);
fs/f2fs/gc.c
504
ve = rb_entry(parent, struct victim_entry, rb_node);
fs/f2fs/gc.c
679
ve = rb_entry(stage == 0 ? rb_prev(&ve->rb_node) :
fs/f2fs/segment.c
1010
cur_dc = rb_entry(cur, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1011
next_dc = rb_entry(next, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1034
dc = rb_entry(node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1067
dc = rb_entry(*pnode, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1080
dc = rb_entry(parent, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1404
dc = rb_entry(parent, struct discard_cmd, rb_node);
fs/fuse/dev.c
2426
ff = rb_entry(p, struct fuse_file, polled_node);
fs/fuse/dir.c
145
cur = rb_entry(*p, struct fuse_dentry, node);
fs/fuse/dir.c
171
fd = rb_entry(node, struct fuse_dentry, node);
fs/fuse/file.c
2693
ff = rb_entry(last, struct fuse_file, polled_node);
fs/fuse/file.c
2789
ff = rb_entry(*link, struct fuse_file, polled_node);
fs/gfs2/rgrp.c
1501
rb_entry(*newn, struct gfs2_blkreserv, rs_node);
fs/gfs2/rgrp.c
1635
rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
fs/gfs2/rgrp.c
1651
rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
fs/gfs2/rgrp.c
2316
trs = rb_entry(n, struct gfs2_blkreserv, rs_node);
fs/gfs2/rgrp.c
532
cur = rb_entry(n, struct gfs2_rgrpd, rd_node);
fs/gfs2/rgrp.c
569
rgd = rb_entry(n, struct gfs2_rgrpd, rd_node);
fs/gfs2/rgrp.c
596
rgd = rb_entry(n, struct gfs2_rgrpd, rd_node);
fs/gfs2/rgrp.c
710
rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
fs/gfs2/rgrp.c
723
rgd = rb_entry(n, struct gfs2_rgrpd, rd_node);
fs/gfs2/rgrp.c
870
struct gfs2_rgrpd *cur = rb_entry(*newn, struct gfs2_rgrpd,
fs/jffs2/nodelist.c
131
base = rb_entry(parent, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.c
537
frag = rb_entry(next, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.h
334
return rb_entry(node, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.h
344
return rb_entry(node, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.h
347
#define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
348
#define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
349
#define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
350
#define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
351
#define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
354
#define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
355
#define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
356
#define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
357
#define tn_left(tn) rb_entry((tn)->rb.rb_left, struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
358
#define tn_right(tn) rb_entry((tn)->rb.rb_right, struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
360
#define tn_last(list) rb_entry(rb_last(list), struct jffs2_tmp_dnode_info, rb)
fs/jffs2/nodelist.h
361
#define tn_first(list) rb_entry(rb_first(list), struct jffs2_tmp_dnode_info, rb)
fs/jffs2/readinode.c
185
tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb);
fs/jffs2/readinode.c
344
insert_point = rb_entry(parent, struct jffs2_tmp_dnode_info, rb);
fs/jffs2/readinode.c
432
this_tn = rb_entry(parent, struct jffs2_tmp_dnode_info, rb);
fs/kernfs/dir.c
30
#define rb_to_kn(X) rb_entry((X), struct kernfs_node, rb)
fs/namespace.c
1072
return node ? rb_entry(node, struct mount, mnt_node) : NULL;
fs/nfs/blocklayout/extent_tree.c
16
return rb_entry(node, struct pnfs_block_extent, be_node);
fs/nfs/dir.c
2985
entry = rb_entry(n, struct nfs_access_entry, rb_node);
fs/nfs/dir.c
3054
rb_entry(n, struct nfs_access_entry, rb_node);
fs/nfs/dir.c
3193
entry = rb_entry(parent, struct nfs_access_entry, rb_node);
fs/nfs/nfs4state.c
169
sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
fs/nfs/nfs4state.c
1694
sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
fs/nfs/nfs4state.c
1746
sp = rb_entry(pos, struct nfs4_state_owner, so_server_node);
fs/nfs/nfs4state.c
1861
sp = rb_entry(pos,
fs/nfs/nfs4state.c
364
sp = rb_entry(parent, struct nfs4_state_owner, so_server_node);
fs/nfs/nfs4state.c
392
sp = rb_entry(parent, struct nfs4_state_owner, so_server_node);
fs/nfsd/nfs4state.c
3329
clp = rb_entry(*new, struct nfs4_client, cl_namenode);
fs/nfsd/nfs4state.c
3350
clp = rb_entry(node, struct nfs4_client, cl_namenode);
fs/nfsd/nfscache.c
420
rp = rb_entry(parent, struct nfsd_cacherep, c_node);
fs/nilfs2/the_nilfs.c
852
root = rb_entry(n, struct nilfs_root, rb_node);
fs/nilfs2/the_nilfs.c
891
root = rb_entry(parent, struct nilfs_root, rb_node);
fs/ntfs3/bitmap.c
1031
e = rb_entry(cr, struct e_node, start.node);
fs/ntfs3/bitmap.c
1046
e = pr ? rb_entry(pr, struct e_node, start.node) : NULL;
fs/ntfs3/bitmap.c
1077
e = rb_entry(rb_first(&wnd->count_tree), struct e_node, count.node);
fs/ntfs3/bitmap.c
137
rb_entry(node, struct e_node, start.node));
fs/ntfs3/bitmap.c
150
k = rb_entry(*p, struct rb_node_key, node);
fs/ntfs3/bitmap.c
176
rb_entry(parent = *p, struct e_node, count.node);
fs/ntfs3/bitmap.c
211
k = rb_entry(parent, struct e_node, start.node);
fs/ntfs3/bitmap.c
252
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
268
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
320
e = rb_entry(n, struct e_node, count.node);
fs/ntfs3/bitmap.c
328
e2 = rb_entry(n, struct e_node, count.node);
fs/ntfs3/bitmap.c
375
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
395
e3 = rb_entry(n3, struct e_node, start.node);
fs/ntfs3/bitmap.c
421
n3 ? rb_entry(n3, struct e_node, count.node)->count.key :
fs/ntfs3/bitmap.c
436
e3 = rb_entry(n3, struct e_node, count.node);
fs/ntfs3/bitmap.c
466
e = rb_entry(rb_last(&wnd->count_tree), struct e_node,
fs/ntfs3/bitmap.c
901
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
939
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/frecord.c
114
struct mft_inode *mi = rb_entry(node, struct mft_inode, node);
fs/ntfs3/frecord.c
1212
mi = rb_entry(node, struct mft_inode, node);
fs/ntfs3/frecord.c
1576
struct mft_inode *mi = rb_entry(node, struct mft_inode, node);
fs/ntfs3/frecord.c
2830
if (rb_entry(node, struct mft_inode, node)->dirty)
fs/ntfs3/frecord.c
30
mi = rb_entry(pr, struct mft_inode, node);
fs/ntfs3/frecord.c
3198
struct mft_inode *mi = rb_entry(node, struct mft_inode, node);
fs/ntfs3/frecord.c
44
return rb_entry(ins, struct mft_inode, node);
fs/ntfs3/frecord.c
945
mi = rb_entry(node, struct mft_inode, node);
fs/ocfs2/cluster/nodemanager.c
78
node = rb_entry(parent, struct o2nm_node, nd_ip_node);
fs/ocfs2/cluster/tcp.c
763
nmh = rb_entry(parent, struct o2net_msg_handler, nh_node);
fs/ocfs2/refcounttree.c
202
tree = rb_entry(n, struct ocfs2_refcount_tree, rf_node);
fs/ocfs2/refcounttree.c
227
tmp = rb_entry(parent, struct ocfs2_refcount_tree,
fs/ocfs2/refcounttree.c
537
tree = rb_entry(node, struct ocfs2_refcount_tree, rf_node);
fs/ocfs2/reservations.c
140
resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
fs/ocfs2/reservations.c
273
resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
fs/ocfs2/reservations.c
314
tmp = rb_entry(parent, struct ocfs2_alloc_reservation, r_node);
fs/ocfs2/reservations.c
365
resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
fs/ocfs2/reservations.c
523
next_resv = rb_entry(next, struct ocfs2_alloc_reservation,
fs/ocfs2/reservations.c
560
next_resv = rb_entry(next,
fs/ocfs2/reservations.c
603
prev_resv = rb_entry(prev, struct ocfs2_alloc_reservation,
fs/ocfs2/reservations.c
85
resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node);
fs/ocfs2/uptodate.c
136
item = rb_entry(node, struct ocfs2_meta_cache_item, c_node);
fs/ocfs2/uptodate.c
214
item = rb_entry(n, struct ocfs2_meta_cache_item, c_node);
fs/ocfs2/uptodate.c
318
tmp = rb_entry(parent, struct ocfs2_meta_cache_item, c_node);
fs/overlayfs/readdir.c
70
return rb_entry(n, struct ovl_cache_entry, node);
fs/proc/generic.c
75
struct proc_dir_entry *de = rb_entry(node,
fs/proc/generic.c
98
struct proc_dir_entry *this = rb_entry(*new,
fs/proc/nommu.c
76
return nommu_region_show(m, rb_entry(p, struct vm_region, vm_rb));
fs/proc/proc_sysctl.c
128
ctl_node = rb_entry(node, struct ctl_node, node);
fs/proc/proc_sysctl.c
162
parent_node = rb_entry(parent, struct ctl_node, node);
fs/proc/proc_sysctl.c
375
ctl_node = rb_entry(node, struct ctl_node, node);
fs/smb/client/cached_dir.c
550
tlink = rb_entry(node, struct tcon_link, tl_rbnode);
fs/smb/client/connect.c
3989
tlink = rb_entry(node, struct tcon_link, tl_rbnode);
fs/smb/client/connect.c
4313
tlink = rb_entry(node, struct tcon_link, tl_rbnode);
fs/smb/client/connect.c
4333
tlink = rb_entry(*new, struct tcon_link, tl_rbnode);
fs/smb/client/connect.c
4468
tlink = rb_entry(tmp, struct tcon_link, tl_rbnode);
fs/smb/client/misc.c
573
tlink = rb_entry(node, struct tcon_link, tl_rbnode);
fs/ubifs/debug.c
1834
fscki = rb_entry(parent, struct fsck_inode, rb);
fs/ubifs/debug.c
1912
fscki = rb_entry(p, struct fsck_inode, rb);
fs/ubifs/debug.c
2181
fscki = rb_entry(this, struct fsck_inode, rb);
fs/ubifs/debug.c
658
bud = rb_entry(rb, struct ubifs_bud, rb);
fs/ubifs/debug.c
746
bud = rb_entry(rb, struct ubifs_bud, rb);
fs/ubifs/log.c
122
b = rb_entry(parent, struct ubifs_bud, rb);
fs/ubifs/log.c
311
bud = rb_entry(p1, struct ubifs_bud, rb);
fs/ubifs/log.c
38
bud = rb_entry(p, struct ubifs_bud, rb);
fs/ubifs/log.c
568
dr = rb_entry(parent, struct done_ref, rb);
fs/ubifs/log.c
71
bud = rb_entry(p, struct ubifs_bud, rb);
fs/ubifs/orphan.c
104
o = rb_entry(p, struct ubifs_orphan, rb);
fs/ubifs/orphan.c
73
o = rb_entry(parent, struct ubifs_orphan, rb);
fs/ubifs/orphan.c
769
o = rb_entry(parent, struct check_orphan, rb);
fs/ubifs/orphan.c
791
o = rb_entry(p, struct check_orphan, rb);
fs/ubifs/recovery.c
1254
e = rb_entry(parent, struct size_entry, rb);
fs/ubifs/recovery.c
1287
e = rb_entry(p, struct size_entry, rb);
fs/ubifs/recovery.c
1534
e = rb_entry(this, struct size_entry, rb);
fs/ubifs/tnc.c
56
o = rb_entry(parent, struct ubifs_old_idx, rb);
fs/ubifs/tnc_commit.c
170
o = rb_entry(p, struct ubifs_old_idx, rb);
fs/xattr.c
1245
xattr = rb_entry(node, struct simple_xattr, rb_node);
fs/xattr.c
1263
xattr = rb_entry(new_node, struct simple_xattr, rb_node);
fs/xattr.c
1292
xattr = rb_entry(rbp, struct simple_xattr, rb_node);
fs/xattr.c
1362
old_xattr = rb_entry(*rbp, struct simple_xattr, rb_node);
fs/xattr.c
1469
xattr = rb_entry(rbp, struct simple_xattr, rb_node);
fs/xattr.c
1553
xattr = rb_entry(rbp, struct simple_xattr, rb_node);
fs/xfs/xfs_extent_busy.c
125
busyp = rb_entry(rbp, struct xfs_extent_busy, rb_node);
fs/xfs/xfs_extent_busy.c
322
rb_entry(rbp, struct xfs_extent_busy, rb_node);
fs/xfs/xfs_extent_busy.c
376
rb_entry(rbp, struct xfs_extent_busy, rb_node);
fs/xfs/xfs_extent_busy.c
58
busyp = rb_entry(parent, struct xfs_extent_busy, rb_node);
include/linux/ceph/libceph.h
199
type *cur = rb_entry(*n, type, nodefld); \
include/linux/ceph/libceph.h
239
type *cur = rb_entry(n, type, nodefld); \
include/linux/interval_tree_generic.h
107
node = rb_entry(node->ITRB.rb_right, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
136
node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
140
leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
161
ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
173
node = rb_entry(rb, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
48
parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \
include/linux/interval_tree_generic.h
89
ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \
include/linux/perf_event.h
866
struct list_head rb_entry;
include/linux/rbtree.h
107
____ptr ? rb_entry(____ptr, type, member) : NULL; \
include/linux/rbtree_augmented.h
106
RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
115
RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
116
RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
122
RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
123
RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
153
child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \
include/linux/rbtree_augmented.h
158
child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
ipc/mqueue.c
199
leaf = rb_entry(parent, struct posix_msg_tree_node, rb_node);
ipc/mqueue.c
269
leaf = rb_entry(parent, struct posix_msg_tree_node, rb_node);
kernel/bpf/range_tree.c
46
return rb_entry(rb, struct range_node, rb_range_size);
kernel/events/core.c
13308
INIT_LIST_HEAD(&event->rb_entry);
kernel/events/core.c
1825
rb_entry((node), struct perf_event, group_node)
kernel/events/core.c
6894
list_del_rcu(&event->rb_entry);
kernel/events/core.c
6908
list_add_rcu(&event->rb_entry, &rb->event_list);
kernel/events/core.c
6948
list_for_each_entry_rcu(event, &rb->event_list, rb_entry)
kernel/events/core.c
7078
list_for_each_entry_rcu(event, &rb->event_list, rb_entry) {
kernel/events/core.c
9070
list_for_each_entry_rcu(iter, &event->rb->event_list, rb_entry) {
kernel/events/uprobes.c
1504
struct uprobe *u = rb_entry(n, struct uprobe, rb_node);
kernel/events/uprobes.c
1543
u = rb_entry(t, struct uprobe, rb_node);
kernel/events/uprobes.c
1551
u = rb_entry(t, struct uprobe, rb_node);
kernel/events/uprobes.c
878
rb_entry((node), struct uprobe, rb_node)
kernel/locking/rtmutex.c
451
rb_entry((node), struct rt_mutex_waiter, tree.entry)
kernel/locking/rtmutex.c
500
rb_entry((node), struct rt_waiter_node, entry)
kernel/locking/rtmutex_common.h
126
return rb_entry(leftmost, struct rt_mutex_waiter, tree.entry) == waiter;
kernel/locking/rtmutex_common.h
137
w = rb_entry(leftmost, struct rt_mutex_waiter, tree.entry);
kernel/locking/rtmutex_common.h
152
return rb_entry(p->pi_waiters.rb_leftmost, struct rt_mutex_waiter,
kernel/locking/ww_mutex.h
108
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
117
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
126
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/locking/ww_mutex.h
99
return rb_entry(n, struct rt_mutex_waiter, tree.entry);
kernel/nstree.c
131
prev_node = rb_entry(prev, struct ns_tree_node, ns_node);
kernel/nstree.c
156
return rb_entry(node, struct ns_common, ns_tree_node.ns_node);
kernel/nstree.c
163
return rb_entry(node, struct ns_common, ns_unified_node.ns_node);
kernel/nstree.c
170
return rb_entry(node, struct ns_common, ns_owner_node.ns_node);
kernel/power/swap.c
136
ext = rb_entry(*new, struct swsusp_extent, node);
kernel/power/swap.c
199
ext = rb_entry(node, struct swsusp_extent, node);
kernel/power/wakelock.c
161
wl = rb_entry(*node, struct wakelock, node);
kernel/power/wakelock.c
47
wl = rb_entry(node, struct wakelock, node);
kernel/sched/core.c
281
#define __node_2_sc(node) rb_entry((node), struct task_struct, core_node)
kernel/sched/deadline.c
2026
rb_entry((node), struct sched_dl_entity, rb_node)
kernel/sched/deadline.c
562
rb_entry((node), struct task_struct, pushable_dl_tasks)
kernel/sched/fair.c
613
rb_entry((node), struct sched_entity, run_node)
lib/rbtree_test.c
121
parent = rb_entry(rb_parent, struct test_node, rb);
lib/rbtree_test.c
199
struct test_node *node = rb_entry(rb, struct test_node, rb);
lib/rbtree_test.c
225
struct test_node *node = rb_entry(rb, struct test_node, rb);
lib/rbtree_test.c
228
subtree = rb_entry(node->rb.rb_left, struct test_node,
lib/rbtree_test.c
234
subtree = rb_entry(node->rb.rb_right, struct test_node,
lib/rbtree_test.c
40
if (key < rb_entry(parent, struct test_node, rb)->key)
lib/rbtree_test.c
58
if (key < rb_entry(parent, struct test_node, rb)->key)
lib/rbtree_test.c
96
parent = rb_entry(rb_parent, struct test_node, rb);
lib/timerqueue.c
18
rb_entry((_n), struct timerqueue_node, node)
mm/backing-dev.c
1051
bdi = rb_entry(parent, struct backing_dev_info, rb_node);
mm/backing-dev.c
1081
bdi = rb_entry(*p, struct backing_dev_info, rb_node);
mm/interval_tree.c
42
parent = rb_entry(prev->shared.rb.rb_right,
mm/interval_tree.c
47
parent = rb_entry(parent->shared.rb.rb_left,
mm/kmemleak.c
430
object = rb_entry(rb, struct kmemleak_object, rb_node);
mm/kmemleak.c
738
parent = rb_entry(rb_parent, struct kmemleak_object, rb_node);
mm/ksm.c
1174
stable_node = rb_entry(root_stable_tree[nid].rb_node,
mm/ksm.c
1854
stable_node = rb_entry(*new, struct ksm_stable_node, node);
mm/ksm.c
2061
stable_node = rb_entry(*new, struct ksm_stable_node, node);
mm/ksm.c
2153
tree_rmap_item = rb_entry(*new, struct ksm_rmap_item, node);
mm/ksm.c
3354
stable_node = rb_entry(node, struct ksm_stable_node, node);
mm/memcontrol-v1.c
126
mz_node = rb_entry(parent, struct mem_cgroup_per_node,
mm/memcontrol-v1.c
246
mz = rb_entry(mctz->rb_rightmost,
mm/mempolicy.c
2896
struct sp_node *p = rb_entry(n, struct sp_node, nd);
mm/mempolicy.c
2912
w = rb_entry(prev, struct sp_node, nd);
mm/mempolicy.c
2917
return rb_entry(n, struct sp_node, nd);
mm/mempolicy.c
2932
nd = rb_entry(parent, struct sp_node, nd);
mm/mempolicy.c
3172
n = rb_entry(next, struct sp_node, nd);
mm/mempolicy.c
3280
n = rb_entry(next, struct sp_node, nd);
mm/nommu.c
1086
pregion = rb_entry(rb, struct vm_region, vm_rb);
mm/nommu.c
447
last = rb_entry(lastp, struct vm_region, vm_rb);
mm/nommu.c
452
region = rb_entry(p, struct vm_region, vm_rb);
mm/nommu.c
453
last = rb_entry(lastp, struct vm_region, vm_rb);
mm/nommu.c
482
pregion = rb_entry(parent, struct vm_region, vm_rb);
mm/shmem_quota.c
103
entry = rb_entry(node, struct quota_id, node);
mm/shmem_quota.c
129
entry = rb_entry(node, struct quota_id, node);
mm/shmem_quota.c
150
entry = rb_entry(node, struct quota_id, node);
mm/shmem_quota.c
182
entry = rb_entry(parent, struct quota_id, node);
mm/shmem_quota.c
281
entry = rb_entry(node, struct quota_id, node);
mm/swapfile.c
2528
struct swap_extent *se = rb_entry(rb, struct swap_extent, rb_node);
mm/swapfile.c
2568
se = rb_entry(parent, struct swap_extent, rb_node);
mm/swapfile.c
269
return rb_entry(rb, struct swap_extent, rb_node);
mm/swapfile.c
275
return rb ? rb_entry(rb, struct swap_extent, rb_node) : NULL;
mm/swapfile.c
323
se = rb_entry(rb, struct swap_extent, rb_node);
mm/vmalloc.c
1088
va = rb_entry(n, struct vmap_area, rb_node);
mm/vmalloc.c
1112
tmp = rb_entry(n, struct vmap_area, rb_node);
mm/vmalloc.c
1205
tmp_va = rb_entry(*link, struct vmap_area, rb_node);
mm/vmalloc.c
1242
list = &rb_entry(parent, struct vmap_area, rb_node)->list;
mm/vmalloc.c
1256
head = &rb_entry(parent, struct vmap_area, rb_node)->list;
mm/vmalloc.c
1584
va = rb_entry(node, struct vmap_area, rb_node);
mm/vmalloc.c
1610
va = rb_entry(node, struct vmap_area, rb_node);
mm/vmalloc.c
4816
tmp = rb_entry(n, struct vmap_area, rb_node);
net/802/garp.c
156
attr = rb_entry(parent, struct garp_attr, node);
net/802/garp.c
177
attr = rb_entry(parent, struct garp_attr, node);
net/802/garp.c
215
attr = rb_entry(node, struct garp_attr, node);
net/802/garp.c
402
attr = rb_entry(node, struct garp_attr, node);
net/802/mrp.c
245
attr = rb_entry(parent, struct mrp_attr, node);
net/802/mrp.c
266
attr = rb_entry(parent, struct mrp_attr, node);
net/802/mrp.c
304
attr = rb_entry(node, struct mrp_attr, node);
net/802/mrp.c
587
attr = rb_entry(node, struct mrp_attr, node);
net/bridge/br_multicast_eht.c
113
this = rb_entry(node, struct net_bridge_group_eht_set,
net/bridge/br_multicast_eht.c
185
set_h = rb_entry(node, struct net_bridge_group_eht_set_entry,
net/bridge/br_multicast_eht.c
202
eht_set = rb_entry(node, struct net_bridge_group_eht_set,
net/bridge/br_multicast_eht.c
254
this = rb_entry(*link, struct net_bridge_group_eht_host,
net/bridge/br_multicast_eht.c
299
this = rb_entry(*link, struct net_bridge_group_eht_set_entry,
net/bridge/br_multicast_eht.c
351
this = rb_entry(*link, struct net_bridge_group_eht_set,
net/bridge/br_multicast_eht.c
53
this = rb_entry(node, struct net_bridge_group_eht_host,
net/bridge/br_multicast_eht.c
89
this = rb_entry(node, struct net_bridge_group_eht_set_entry,
net/ceph/auth_x.c
159
th = rb_entry(parent, struct ceph_x_ticket_handler, node);
net/ceph/auth_x.c
991
rb_entry(p, struct ceph_x_ticket_handler, node);
net/ceph/ceph_common.c
622
rb_entry(n, struct crush_loc_node, cl_node);
net/ceph/debugfs.c
110
rb_entry(n, struct ceph_pg_mapping, node);
net/ceph/debugfs.c
117
rb_entry(n, struct ceph_pg_mapping, node);
net/ceph/debugfs.c
128
rb_entry(n, struct ceph_pg_mapping, node);
net/ceph/debugfs.c
168
req = rb_entry(rp, struct ceph_mon_generic_request, node);
net/ceph/debugfs.c
247
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/debugfs.c
273
rb_entry(n, struct ceph_osd_linger_request, node);
net/ceph/debugfs.c
334
rb_entry(n, struct ceph_osd_backoff, id_node);
net/ceph/debugfs.c
359
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/debugfs.c
367
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/debugfs.c
375
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/debugfs.c
76
rb_entry(n, struct ceph_pg_pool_info, node);
net/ceph/debugfs.c
99
rb_entry(n, struct ceph_pg_mapping, node);
net/ceph/mon_client.c
1071
req = rb_entry(p, struct ceph_mon_generic_request, node);
net/ceph/osd_client.c
1180
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/osd_client.c
1184
rb_entry(p, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
1194
rb_entry(p, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
1350
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
1360
rb_entry(n, struct ceph_osd_linger_request, node);
net/ceph/osd_client.c
1401
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
1497
rb_entry(n, struct ceph_pg_pool_info, node);
net/ceph/osd_client.c
1947
rb_entry(n, struct ceph_osd_backoff, spg_node);
net/ceph/osd_client.c
1975
rb_entry(rb_first(&osd->o_backoff_mappings),
net/ceph/osd_client.c
1980
rb_entry(rb_first(&spg->backoffs),
net/ceph/osd_client.c
3448
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/osd_client.c
3453
rb_entry(p, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
3471
rb_entry(p, struct ceph_osd_linger_request, node);
net/ceph/osd_client.c
3490
rb_entry(p, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
3887
rb_entry(n, struct ceph_pg_pool_info, node);
net/ceph/osd_client.c
3940
rb_entry(n, struct ceph_osd_linger_request, node);
net/ceph/osd_client.c
3976
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
4036
rb_entry(n, struct ceph_pg_pool_info, node);
net/ceph/osd_client.c
4061
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/osd_client.c
4088
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
4103
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
4255
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
4268
rb_entry(n, struct ceph_osd_linger_request, node);
net/ceph/osd_client.c
4497
rb_entry(n, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
4729
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/osd_client.c
4734
rb_entry(p, struct ceph_osd_request, r_node);
net/ceph/osd_client.c
5187
struct ceph_osd *osd = rb_entry(n, struct ceph_osd, o_node);
net/ceph/osd_client.c
5278
struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds),
net/ceph/osdmap.c
1144
rb_entry(rb_first(&map->pg_temp),
net/ceph/osdmap.c
1151
rb_entry(rb_first(&map->primary_temp),
net/ceph/osdmap.c
1158
rb_entry(rb_first(&map->pg_upmap),
net/ceph/osdmap.c
1165
rb_entry(rb_first(&map->pg_upmap_items),
net/ceph/osdmap.c
1172
rb_entry(rb_first(&map->pg_pools),
net/ceph/osdmap.c
223
rb_entry(rb_first(root), struct crush_name_node, cn_node);
net/ceph/osdmap.c
273
rb_entry(rb_first(&c->choose_args),
net/ceph/osdmap.c
2976
rb_entry(n1, struct crush_loc_node, cl_node);
net/ceph/osdmap.c
2978
rb_entry(n2, struct crush_loc_node, cl_node);
net/ceph/osdmap.c
2996
rb_entry(rb_first(locs), struct crush_loc_node, cl_node);
net/ceph/osdmap.c
767
rb_entry(rbp, struct ceph_pg_pool_info, node);
net/ceph/string_table.c
21
exist = rb_entry(*p, struct ceph_string, node);
net/ceph/string_table.c
56
exist = rb_entry(*p, struct ceph_string, node);
net/core/dev_addr_lists.c
175
struct netdev_hw_addr *ha = rb_entry(node, struct netdev_hw_addr, node);
net/core/dev_addr_lists.c
30
ha = rb_entry(*ins_point, struct netdev_hw_addr, node);
net/core/dev_addr_lists.c
87
ha = rb_entry(*ins_point, struct netdev_hw_addr, node);
net/core/skbuff.c
4038
struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode);
net/ipv4/inet_fragment.c
308
struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode);
net/ipv4/inetpeer.c
109
p = rb_entry(parent, struct inet_peer, rb_node);
net/ipv4/inetpeer.c
279
struct inet_peer *peer = rb_entry(p, struct inet_peer, rb_node);
net/ipv4/nexthop.c
2621
nh = rb_entry(parent, struct nexthop, rb_node);
net/ipv4/nexthop.c
2722
nh = rb_entry(node, struct nexthop, rb_node);
net/ipv4/nexthop.c
3563
nh = rb_entry(tmp, struct nexthop, rb_node);
net/ipv4/nexthop.c
3582
nh = rb_entry(node, struct nexthop, rb_node);
net/ipv4/nexthop.c
3977
nh = rb_entry(node, struct nexthop, rb_node);
net/ipv4/nexthop.c
602
nh = rb_entry(parent, struct nexthop, rb_node);
net/netfilter/nf_conncount.c
419
rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node);
net/netfilter/nf_conncount.c
506
rbconn = rb_entry(parent, struct nf_conncount_rb, node);
net/netfilter/nf_conncount.c
564
rbconn = rb_entry(node, struct nf_conncount_rb, node);
net/netfilter/nf_conncount.c
580
rbconn = rb_entry(node, struct nf_conncount_rb, node);
net/netfilter/nf_conncount.c
665
rbconn = rb_entry(node, struct nf_conncount_rb, node);
net/netfilter/nft_set_rbtree.c
1005
rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
1067
prev_rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
1088
prev_rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
1170
rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
221
rbe_prev = rb_entry(prev, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
231
rbe_prev = rb_entry(prev, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
246
first_elem = rb_entry(first, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
265
return rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
274
next_rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
357
rbe = rb_entry(parent, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
386
rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
523
rbe = rb_entry(parent, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
742
next_rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
806
rbe = rb_entry(parent, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
851
rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/netfilter/nft_set_rbtree.c
905
rbe = rb_entry(node, struct nft_rbtree_elem, node);
net/rds/cong.c
115
map = rb_entry(parent, struct rds_cong_map, m_rb_node);
net/rds/cong.c
406
map = rb_entry(node, struct rds_cong_map, m_rb_node);
net/rds/rdma.c
136
mr = rb_entry(node, struct rds_mr, r_rb_node);
net/rds/rdma.c
74
mr = rb_entry(parent, struct rds_mr, r_rb_node);
net/rxrpc/af_rxrpc.c
896
skb = rb_entry(rx->pending_oobq.rb_node, struct sk_buff, rbnode);
net/rxrpc/call_accept.c
114
xcall = rb_entry(parent, struct rxrpc_call, sock_node);
net/rxrpc/call_object.c
367
xcall = rb_entry(parent, struct rxrpc_call, sock_node);
net/rxrpc/call_object.c
97
call = rb_entry(p, struct rxrpc_call, sock_node);
net/rxrpc/conn_client.c
272
bundle = rb_entry(p, struct rxrpc_bundle, local_node);
net/rxrpc/conn_client.c
301
bundle = rb_entry(parent, struct rxrpc_bundle, local_node);
net/rxrpc/conn_service.c
43
conn = rb_entry(p, struct rxrpc_connection, service_node);
net/rxrpc/conn_service.c
77
cursor = rb_entry(parent,
net/rxrpc/oob.c
76
skb = rb_entry(p, struct sk_buff, rbnode);
net/sched/sch_fq.c
228
aux = rb_entry(parent, struct fq_flow, rate_node);
net/sched/sch_fq.c
269
f = rb_entry(parent, struct fq_flow, fq_node);
net/sched/sch_fq.c
410
f = rb_entry(parent, struct fq_flow, fq_node);
net/sched/sch_fq.c
629
struct fq_flow *f = rb_entry(p, struct fq_flow, rate_node);
net/sched/sch_fq.c
819
f = rb_entry(p, struct fq_flow, fq_node);
net/sched/sch_fq.c
852
of = rb_entry(op, struct fq_flow, fq_node);
net/sched/sch_fq.c
865
nf = rb_entry(parent, struct fq_flow, fq_node);
net/sched/sch_hfsc.c
198
cl1 = rb_entry(parent, struct hfsc_class, el_node);
net/sched/sch_hfsc.c
232
p = rb_entry(n, struct hfsc_class, el_node);
net/sched/sch_hfsc.c
250
return rb_entry(n, struct hfsc_class, el_node);
net/sched/sch_hfsc.c
266
cl1 = rb_entry(parent, struct hfsc_class, vt_node);
net/sched/sch_hfsc.c
296
p = rb_entry(n, struct hfsc_class, vt_node);
net/sched/sch_hfsc.c
335
cl1 = rb_entry(parent, struct hfsc_class, cf_node);
net/sched/sch_hfsc.c
667
p = rb_entry(n, struct hfsc_class, cf_node);
net/sched/sch_hfsc.c
690
max_cl = rb_entry(n, struct hfsc_class, vt_node);
net/sched/sch_htb.c
295
c = rb_entry(parent, struct htb_class, node[prio]);
net/sched/sch_htb.c
332
c = rb_entry(parent, struct htb_class, pq_node);
net/sched/sch_htb.c
765
cl = rb_entry(p, struct htb_class, pq_node);
net/sched/sch_htb.c
794
rb_entry(n, struct htb_class, node[prio]);
net/sched/sch_htb.c
858
cl = rb_entry(*sp->pptr, struct htb_class, node[prio]);
net/smc/smc_core.c
108
struct smc_connection *cur = rb_entry(*link,
net/smc/smc_core.c
1126
conn = rb_entry(node, struct smc_connection, alert_node);
net/smc/smc_core.c
1576
conn = rb_entry(node, struct smc_connection, alert_node);
net/smc/smc_core.h
469
struct smc_connection *cur = rb_entry(node,
net/sunrpc/xprt.c
1012
req = rb_entry(n, struct rpc_rqst, rq_recv);
net/sunrpc/xprt.c
1036
req = rb_entry(n, struct rpc_rqst, rq_recv);
net/wireless/scan.c
1661
tbss = rb_entry(parent, struct cfg80211_internal_bss, rbn);
net/wireless/scan.c
1691
bss = rb_entry(n, struct cfg80211_internal_bss, rbn);
net/xfrm/xfrm_policy.c
1027
node = rb_entry(rnode, struct xfrm_pol_inexact_node, node);
net/xfrm/xfrm_policy.c
1056
node = rb_entry(*p, struct xfrm_pol_inexact_node, node);
net/xfrm/xfrm_policy.c
1124
node = rb_entry(rn, struct xfrm_pol_inexact_node, node);
net/xfrm/xfrm_policy.c
2006
node = rb_entry(parent, struct xfrm_pol_inexact_node, node);
net/xfrm/xfrm_policy.c
973
node = rb_entry(*p, struct xfrm_pol_inexact_node, node);
security/apparmor/label.c
1963
struct aa_label *this = rb_entry(node, struct aa_label, node);
security/apparmor/label.c
1996
label = rb_entry(node, struct aa_label, node);
security/apparmor/label.c
690
struct aa_label *this = rb_entry(*new, struct aa_label, node);
security/apparmor/label.c
744
struct aa_label *this = rb_entry(node, struct aa_label, node);
security/keys/gc.c
218
key = rb_entry(cursor, struct key, serial_node);
security/keys/key.c
154
xkey = rb_entry(parent, struct key, serial_node);
security/keys/key.c
185
xkey = rb_entry(parent, struct key, serial_node);
security/keys/key.c
63
user = rb_entry(parent, struct key_user, node);
security/keys/key.c
683
key = rb_entry(n, struct key, serial_node);
security/keys/proc.c
108
minkey = rb_entry(n, struct key, serial_node);
security/keys/proc.c
131
struct key *key = rb_entry(n, struct key, serial_node);
security/keys/proc.c
156
struct key *key = rb_entry(_p, struct key, serial_node);
security/keys/proc.c
255
struct key_user *user = rb_entry(n, struct key_user, node);
security/keys/proc.c
306
struct key_user *user = rb_entry(_p, struct key_user, node);
security/keys/proc.c
70
struct key *key = rb_entry(n, struct key, serial_node);
security/keys/proc.c
85
struct key *key = rb_entry(n, struct key, serial_node);
security/landlock/ruleset.c
230
rb_entry(*walker_node, struct landlock_rule, node);
security/landlock/ruleset.c
602
rb_entry(node, struct landlock_rule, node);
tools/bpf/resolve_btfids/main.c
216
id = rb_entry(p, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
240
id = rb_entry(parent, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
363
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
726
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
756
id = rb_entry(next, struct btf_id, rb_node);
tools/bpf/resolve_btfids/main.c
964
set_id = rb_entry(next, struct btf_id, rb_node);
tools/include/linux/interval_tree_generic.h
107
node = rb_entry(node->ITRB.rb_right, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
136
node = rb_entry(root->rb_root.rb_node, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
140
leftmost = rb_entry(root->rb_leftmost, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
161
ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
173
node = rb_entry(rb, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
48
parent = rb_entry(rb_parent, ITSTRUCT, ITRB); \
tools/include/linux/interval_tree_generic.h
89
ITSTRUCT *left = rb_entry(node->ITRB.rb_left, \
tools/include/linux/rbtree.h
77
____ptr ? rb_entry(____ptr, type, member) : NULL; \
tools/include/linux/rbtree_augmented.h
129
child = rb_entry(node->RBFIELD.rb_left, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
134
child = rb_entry(node->RBFIELD.rb_right, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
82
RBSTRUCT *node = rb_entry(rb, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
91
RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
92
RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
98
RBSTRUCT *old = rb_entry(rb_old, RBSTRUCT, RBFIELD); \
tools/include/linux/rbtree_augmented.h
99
RBSTRUCT *new = rb_entry(rb_new, RBSTRUCT, RBFIELD); \
tools/objtool/elf.c
119
const struct symbol *s = rb_entry(node, struct symbol, node);
tools/objtool/elf.c
260
s = rb_entry(n, struct symbol, node);
tools/objtool/elf.c
501
entry = &rb_entry(pnode, struct symbol, node)->list;
tools/perf/builtin-annotate.c
422
struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/builtin-c2c.c
2391
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/builtin-c2c.c
2531
struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/builtin-c2c.c
2636
struct hist_entry *he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/builtin-diff.c
538
struct hist_entry *he = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/builtin-diff.c
591
struct hist_entry *he_pair = rb_entry(next, struct hist_entry,
tools/perf/builtin-diff.c
657
struct hist_entry *he = rb_entry(next, struct hist_entry,
tools/perf/builtin-diff.c
687
he = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/builtin-kmem.c
1014
struct alloc_stat *data = rb_entry(next, struct alloc_stat,
tools/perf/builtin-kmem.c
1087
data = rb_entry(next, struct page_stat, node);
tools/perf/builtin-kmem.c
1129
data = rb_entry(next, struct page_stat, node);
tools/perf/builtin-kmem.c
1274
this = rb_entry(*new, struct alloc_stat, node);
tools/perf/builtin-kmem.c
1305
data = rb_entry(node, struct alloc_stat, node);
tools/perf/builtin-kmem.c
1321
this = rb_entry(*new, struct page_stat, node);
tools/perf/builtin-kmem.c
1352
data = rb_entry(node, struct page_stat, node);
tools/perf/builtin-kmem.c
141
data = rb_entry(*node, struct alloc_stat, node);
tools/perf/builtin-kmem.c
233
data = rb_entry(node, struct alloc_stat, node);
tools/perf/builtin-kmem.c
472
data = rb_entry(*node, struct page_stat, node);
tools/perf/builtin-kmem.c
522
data = rb_entry(*node, struct page_stat, node);
tools/perf/builtin-kmem.c
577
data = rb_entry(*node, struct page_stat, node);
tools/perf/builtin-kmem.c
96
data = rb_entry(*node, struct alloc_stat, node);
tools/perf/builtin-kvm.c
1082
he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/builtin-kvm.c
523
struct hist_entry *he = rb_entry(nd, struct hist_entry,
tools/perf/builtin-kvm.c
655
he = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/builtin-kwork.c
1698
data = rb_entry(node, struct kwork_work, node);
tools/perf/builtin-kwork.c
1924
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2010
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2082
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
2129
data = rb_entry(node, struct kwork_work, node);
tools/perf/builtin-kwork.c
2161
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-kwork.c
555
work = rb_entry(next, struct kwork_work, node);
tools/perf/builtin-sched.c
1512
data = rb_entry(node, struct work_atoms, node);
tools/perf/builtin-sched.c
3061
chain = rb_entry(rb_node, struct callchain_node, rb_node);
tools/perf/builtin-sched.c
3473
data = rb_entry(node, struct work_atoms, node);
tools/perf/builtin-sched.c
3529
work_list = rb_entry(next, struct work_atoms, node);
tools/perf/builtin-sched.c
3548
data = rb_entry(next, struct work_atoms, node);
tools/perf/builtin-top.c
418
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/tests/code-reading.c
60
data = rb_entry(*node, struct tested_section, rb_node);
tools/perf/tests/code-reading.c
96
struct tested_section *ts = rb_entry(node,
tools/perf/tests/hists_common.c
179
he = rb_entry(node, struct hist_entry, rb_node_in);
tools/perf/tests/hists_common.c
208
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_cumulate.c
147
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_cumulate.c
217
node && (he = rb_entry(node, struct hist_entry, rb_node));
tools/perf/tests/hists_cumulate.c
238
cnode = rb_entry(rb_first(root), struct callchain_node, rb_node);
tools/perf/tests/hists_link.c
178
he = rb_entry(node, struct hist_entry, rb_node_in);
tools/perf/tests/hists_link.c
230
he = rb_entry(node, struct hist_entry, rb_node_in);
tools/perf/tests/hists_output.c
112
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
180
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
186
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
192
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
198
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
204
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
210
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
216
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
222
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
228
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
280
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
285
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
334
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
340
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
346
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
352
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
358
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
412
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
418
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
424
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
430
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
436
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
442
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
448
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
454
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
460
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
515
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
523
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
530
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
537
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
544
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
551
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
558
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
565
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
572
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/hists_output.c
579
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/tests/symbols.c
126
struct symbol *sym = rb_entry(nd, struct symbol, rb_node);
tools/perf/tests/vmlinux-kallsyms.c
291
sym = rb_entry(nd, struct symbol, rb_node);
tools/perf/ui/browsers/annotate.c
292
l = rb_entry(parent, struct annotation_line, rb_node);
tools/perf/ui/browsers/annotate.c
329
struct annotation_line * pos = rb_entry(nd, struct annotation_line, rb_node);
tools/perf/ui/browsers/annotate.c
860
al = rb_entry(browser->curr_hot, struct annotation_line, rb_node);
tools/perf/ui/browsers/hists.c
1029
struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
1118
struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
183
struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
1870
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
1912
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
1935
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
1983
h = rb_entry(browser->top, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2005
h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2027
h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2064
h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2073
h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2199
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
266
struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
290
child = rb_entry(node, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
2968
he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
334
struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
374
struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
517
struct callchain_node *child = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
558
struct callchain_node *node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
574
child = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
611
he = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
703
struct hist_entry *h = rb_entry(browser->b.top, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
72
rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/browsers/hists.c
905
child = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/browsers/hists.c
925
struct callchain_node *child = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/browsers/map.c
123
struct symbol *pos = rb_entry(nd, struct symbol, rb_node);
tools/perf/ui/browsers/map.c
27
struct symbol *sym = rb_entry(nd, struct symbol, rb_node);
tools/perf/ui/gtk/hists.c
109
node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/gtk/hists.c
173
node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/gtk/hists.c
233
node = rb_entry(nd, struct callchain_node, rb_node);
tools/perf/ui/gtk/hists.c
359
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/ui/gtk/hists.c
425
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/ui/stdio/hist.c
134
child = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/stdio/hist.c
217
cnode = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/stdio/hist.c
236
cnode = rb_entry(node, struct callchain_node, rb_node);
tools/perf/ui/stdio/hist.c
318
chain = rb_entry(rb_node, struct callchain_node, rb_node);
tools/perf/ui/stdio/hist.c
370
chain = rb_entry(rb_node, struct callchain_node, rb_node);
tools/perf/ui/stdio/hist.c
871
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/util/annotate-data.c
1774
pos = rb_entry(node, struct annotated_data_type, node);
tools/perf/util/annotate-data.c
204
type = rb_entry(node, struct annotated_data_type, node);
tools/perf/util/annotate-data.c
215
a = rb_entry(node_a, struct annotated_data_type, node);
tools/perf/util/annotate-data.c
216
b = rb_entry(node_b, struct annotated_data_type, node);
tools/perf/util/annotate-data.c
380
result = rb_entry(node, struct annotated_data_type, node);
tools/perf/util/annotate-data.c
650
gvar = rb_entry(node, struct global_var_entry, node);
tools/perf/util/annotate-data.c
661
gvar_a = rb_entry(node_a, struct global_var_entry, node);
tools/perf/util/annotate-data.c
662
gvar_b = rb_entry(node_b, struct global_var_entry, node);
tools/perf/util/annotate-data.c
676
return rb_entry(node, struct global_var_entry, node);
tools/perf/util/annotate-data.c
715
gvar = rb_entry(node, struct global_var_entry, node);
tools/perf/util/annotate.c
1072
iter = rb_entry(parent, struct annotation_line, rb_node);
tools/perf/util/annotate.c
1119
iter = rb_entry(parent, struct annotation_line, rb_node);
tools/perf/util/annotate.c
1140
al = rb_entry(node, struct annotation_line, rb_node);
tools/perf/util/annotate.c
1169
al = rb_entry(node, struct annotation_line, rb_node);
tools/perf/util/block-info.c
454
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/block-range.c
113
next = rb_entry(n, struct block_range, node);
tools/perf/util/block-range.c
19
struct block_range *entry = rb_entry(rb, struct block_range, node);
tools/perf/util/block-range.c
37
entry = rb_entry(parent, struct block_range, node);
tools/perf/util/block-range.c
86
entry = rb_entry(parent, struct block_range, node);
tools/perf/util/block-range.h
43
return rb_entry(n, struct block_range, node);
tools/perf/util/bpf_lock_contention.c
845
struct cgroup *cgrp = rb_entry(node, struct cgroup, node);
tools/perf/util/build-id.c
376
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/build-id.c
924
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/build-id.c
960
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/call-path.c
95
cp = rb_entry(node_parent, struct call_path, rb_node);
tools/perf/util/callchain.c
1309
child = rb_entry(n, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
1772
node = rb_entry(rb_node, struct callchain_node, rb_node);
tools/perf/util/callchain.c
1786
struct hist_entry *he = rb_entry(next, struct hist_entry,
tools/perf/util/callchain.c
402
rnode = rb_entry(parent, struct callchain_node, rb_node);
tools/perf/util/callchain.c
439
child = rb_entry(n, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
471
child = rb_entry(n, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
501
child = rb_entry(n, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
566
child = rb_entry(n, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
878
first = rb_entry(p, struct callchain_node, rb_node_in);
tools/perf/util/callchain.c
920
rnode = rb_entry(parent, struct callchain_node, rb_node_in);
tools/perf/util/cgroup.c
524
cgrp = rb_entry(parent, struct cgroup, node);
tools/perf/util/cgroup.c
592
cgrp = rb_entry(node, struct cgroup, node);
tools/perf/util/dso.c
1813
struct symbol *pos = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/dso.c
904
cache = rb_entry(next, struct dso_cache, rb_node);
tools/perf/util/dso.c
923
cache = rb_entry(parent, struct dso_cache, rb_node);
tools/perf/util/dso.c
951
cache = rb_entry(parent, struct dso_cache, rb_node);
tools/perf/util/env.c
102
(*cb)(rb_entry(node, struct bpf_prog_info_node, rb_node), data);
tools/perf/util/env.c
127
node = rb_entry(parent, struct btf_node, rb_node);
tools/perf/util/env.c
162
node = rb_entry(n, struct btf_node, rb_node);
tools/perf/util/env.c
187
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/env.c
203
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/env.c
50
node = rb_entry(parent, struct bpf_prog_info_node, rb_node);
tools/perf/util/env.c
77
node = rb_entry(n, struct bpf_prog_info_node, rb_node);
tools/perf/util/header.c
1051
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/header.c
1093
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/header.c
2001
node = rb_entry(next, struct bpf_prog_info_node, rb_node);
tools/perf/util/header.c
2028
node = rb_entry(next, struct btf_node, rb_node);
tools/perf/util/hist.c
1638
iter = rb_entry(parent, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
1761
iter = rb_entry(parent, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
1845
n = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
1918
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
1941
iter = rb_entry(parent, struct hist_entry, rb_node);
tools/perf/util/hist.c
1975
he = rb_entry(node, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2041
iter = rb_entry(parent, struct hist_entry, rb_node);
tools/perf/util/hist.c
2099
n = rb_entry(next, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2160
struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2164
he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2171
struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2190
struct hist_entry *he = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2213
child = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2217
child = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
2341
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/util/hist.c
2362
iter = rb_entry(parent, struct hist_entry, rb_node);
tools/perf/util/hist.c
2380
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/util/hist.c
2403
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/util/hist.c
2447
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
tools/perf/util/hist.c
2560
he = rb_entry(parent, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2604
he = rb_entry(parent, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2642
struct hist_entry *iter = rb_entry(n, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2665
iter = rb_entry(n, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2685
pos = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2716
pos = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
272
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
2733
pos = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2792
pos = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
2817
pos = rb_entry(nd, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
3035
he = rb_entry(node, struct hist_entry, rb_node_in);
tools/perf/util/hist.c
434
child = rb_entry(node, struct hist_entry, rb_node);
tools/perf/util/hist.c
477
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
493
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
507
n = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/hist.c
696
he = rb_entry(parent, struct hist_entry, rb_node_in);
tools/perf/util/intel-pt.c
328
v = rb_entry(parent, struct intel_pt_vmcs_info, rb_node);
tools/perf/util/intel-pt.c
370
v = rb_entry(n, struct intel_pt_vmcs_info, rb_node);
tools/perf/util/intlist.h
49
return rn ? rb_entry(rn, struct int_node, rb_node) : NULL;
tools/perf/util/intlist.h
57
return rn ? rb_entry(rn, struct int_node, rb_node) : NULL;
tools/perf/util/machine.c
1266
struct machine *pos = rb_entry(next, struct machine, rb_node);
tools/perf/util/machine.c
262
pos = rb_entry(parent, struct machine, rb_node);
tools/perf/util/machine.c
286
struct machine *machine = rb_entry(nd, struct machine, rb_node);
tools/perf/util/machine.c
304
machine = rb_entry(parent, struct machine, rb_node);
tools/perf/util/machine.c
3070
struct machine *machine = rb_entry(nd, struct machine, rb_node);
tools/perf/util/machine.c
454
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/machine.c
468
machine = rb_entry(node, struct machine, rb_node);
tools/perf/util/machine.c
908
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/machine.c
928
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/map.c
322
struct symbol *sym = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/map.c
335
struct symbol *sym = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/mem2node.c
127
entry = rb_entry(parent, struct phys_entry, rb_node);
tools/perf/util/mem2node.c
26
e = rb_entry(parent, struct phys_entry, rb_node);
tools/perf/util/session.c
2736
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/session.c
2952
struct machine *pos = rb_entry(nd, struct machine, rb_node);
tools/perf/util/srcline.c
381
i = rb_entry(parent, struct srcline_node, rb_node);
tools/perf/util/srcline.c
398
struct srcline_node *i = rb_entry(n, struct srcline_node,
tools/perf/util/srcline.c
418
pos = rb_entry(next, struct srcline_node, rb_node);
tools/perf/util/srcline.c
465
i = rb_entry(parent, struct inline_node, rb_node);
tools/perf/util/srcline.c
482
struct inline_node *i = rb_entry(n, struct inline_node,
tools/perf/util/srcline.c
502
pos = rb_entry(next, struct inline_node, rb_node);
tools/perf/util/stream.c
106
cnode = rb_entry(rb_node, struct callchain_node, rb_node);
tools/perf/util/stream.c
119
he = rb_entry(next, struct hist_entry, rb_node);
tools/perf/util/strlist.h
59
return rn ? rb_entry(rn, struct str_node, rb_node) : NULL;
tools/perf/util/strlist.h
67
return rn ? rb_entry(rn, struct str_node, rb_node) : NULL;
tools/perf/util/symbol.c
1047
m = rb_entry(parent, struct module_info, rb_node);
tools/perf/util/symbol.c
1063
mi = rb_entry(next, struct module_info, rb_node);
tools/perf/util/symbol.c
1080
m = rb_entry(n, struct module_info, rb_node);
tools/perf/util/symbol.c
1149
from_m = rb_entry(from_node, struct module_info, rb_node);
tools/perf/util/symbol.c
1150
to_m = rb_entry(to_node, struct module_info, rb_node);
tools/perf/util/symbol.c
218
curr = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/symbol.c
224
next = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/symbol.c
253
curr = rb_entry(prevnd, struct symbol, rb_node);
tools/perf/util/symbol.c
257
curr = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/symbol.c
353
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.c
382
s = rb_entry(parent, struct symbol, rb_node);
tools/perf/util/symbol.c
409
struct symbol *s = rb_entry(n, struct symbol, rb_node);
tools/perf/util/symbol.c
427
return rb_entry(n, struct symbol, rb_node);
tools/perf/util/symbol.c
437
return rb_entry(n, struct symbol, rb_node);
tools/perf/util/symbol.c
447
return rb_entry(n, struct symbol, rb_node);
tools/perf/util/symbol.c
474
struct symbol *pos = rb_entry(nd, struct symbol, rb_node);
tools/perf/util/symbol.c
821
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.c
880
pos = rb_entry(next, struct symbol, rb_node);
tools/perf/util/symbol.h
86
nd && (pos = rb_entry(nd, struct symbol, rb_node)); \
tools/testing/selftests/bpf/progs/rbtree_search.c
115
n = rb_entry(rb_n, struct node_data, r0);
tools/testing/selftests/bpf/progs/rbtree_search.c
129
bpf_obj_drop(rb_entry(rb_m, struct node_data, r1));
tools/testing/selftests/bpf/progs/rbtree_search.c
168
n = rb_entry(rb_n, struct node_data, r0); \
tools/testing/selftests/bpf/progs/rbtree_search.c
34
node_a = rb_entry(a, struct node_data, r0);
tools/testing/selftests/bpf/progs/rbtree_search.c
35
node_b = rb_entry(b, struct node_data, r0);
tools/testing/selftests/bpf/progs/rbtree_search.c
45
node_a = rb_entry(a, struct node_data, r1);
tools/testing/selftests/bpf/progs/rbtree_search.c
46
node_b = rb_entry(b, struct node_data, r1);
tools/testing/selftests/bpf/progs/rbtree_search.c
88
n = rb_entry(rb_n, struct node_data, r0);