Symbol: rb_entry_safe
drivers/gpu/drm/drm_mm.c
298
return rb_entry_safe(rb, struct drm_mm_node, rb_hole_size);
drivers/gpu/drm/drm_mm.c
303
return rb_entry_safe(rb, struct drm_mm_node, rb_hole_addr);
drivers/gpu/drm/i915/gt/intel_engine.h
382
rb_entry_safe(rb, struct intel_engine_cs, uabi_node)
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1002
engine = rb_entry_safe(rb_next(&engine->uabi_node),
fs/btrfs/backref.c
3173
while ((node = rb_entry_safe(rb_first(&cache->rb_root),
fs/btrfs/defrag.c
189
entry = rb_entry_safe(parent, struct inode_defrag, rb_node);
fs/btrfs/delayed-inode.c
396
return rb_entry_safe(node, struct btrfs_delayed_item, rb_node);
fs/btrfs/delayed-inode.c
479
return rb_entry_safe(p, struct btrfs_delayed_item, rb_node);
fs/btrfs/delayed-inode.c
487
return rb_entry_safe(p, struct btrfs_delayed_item, rb_node);
fs/btrfs/delayed-inode.c
495
return rb_entry_safe(p, struct btrfs_delayed_item, rb_node);
fs/btrfs/delayed-ref.c
336
return rb_entry_safe(exist, struct btrfs_delayed_ref_node, ref_node);
fs/btrfs/extent-io-tree.c
205
return rb_entry_safe(next, struct extent_state, rb_node);
fs/btrfs/extent-io-tree.c
212
return rb_entry_safe(next, struct extent_state, rb_node);
fs/btrfs/extent_map.c
376
merge = rb_entry_safe(rb, struct extent_map, rb_node);
fs/btrfs/extent_map.c
394
merge = rb_entry_safe(rb, struct extent_map, rb_node);
fs/btrfs/qgroup.c
183
return rb_entry_safe(node, struct btrfs_qgroup, node);
fs/btrfs/ref-verify.c
104
return rb_entry_safe(node, struct block_entry, node);
fs/btrfs/ref-verify.c
112
return rb_entry_safe(node, struct block_entry, node);
fs/btrfs/ref-verify.c
141
return rb_entry_safe(node, struct root_entry, node);
fs/btrfs/ref-verify.c
179
return rb_entry_safe(node, struct ref_entry, node);
fs/btrfs/ref-verify.c
187
return rb_entry_safe(node, struct root_entry, node);
fs/btrfs/ulist.c
150
return rb_entry_safe(node, struct ulist_node, rb_node);
fs/f2fs/extent_cache.c
263
*next_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
268
*prev_entry = rb_entry_safe(tmp_node, struct extent_node, rb_node);
fs/f2fs/extent_cache.c
275
*prev_entry = rb_entry_safe(tmp_node,
fs/f2fs/extent_cache.c
281
*next_entry = rb_entry_safe(tmp_node,
fs/f2fs/extent_cache.c
761
next_en = rb_entry_safe(node, struct extent_node,
fs/f2fs/gc.c
585
ve = rb_entry_safe(node, struct victim_entry, rb_node);
fs/f2fs/segment.c
1084
*next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1089
*prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1095
*prev_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1099
*next_entry = rb_entry_safe(tmp_node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1547
next_dc = rb_entry_safe(node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
1623
dc = rb_entry_safe(node, struct discard_cmd, rb_node);
fs/f2fs/segment.c
3468
dc = rb_entry_safe(node, struct discard_cmd, rb_node);
fs/proc/generic.c
58
return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry,
fs/proc/generic.c
64
return rb_entry_safe(rb_next(&dir->subdir_node), struct proc_dir_entry,
fs/xfs/scrub/bitmap.c
338
for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \
fs/xfs/scrub/bitmap.c
341
(bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
fs/xfs/scrub/bitmap.c
63
for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \
fs/xfs/scrub/bitmap.c
66
(bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
include/linux/rbtree.h
128
for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \
include/linux/rbtree.h
129
pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \
include/linux/skbuff.h
4127
#define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode)
include/linux/timerqueue.h
27
return rb_entry_safe(leftmost, struct timerqueue_node, node);
kernel/events/core.c
1958
for (event = rb_entry_safe(rb_first(&((groups)->tree)), \
kernel/events/core.c
1960
event = rb_entry_safe(rb_next(&event->group_node), \
mm/vmalloc.c
1059
va = rb_entry_safe(node, struct vmap_area, rb_node);
mm/vmalloc.c
4794
return rb_entry_safe(n, struct vmap_area, rb_node);
tools/include/linux/rbtree.h
98
for (pos = rb_entry_safe(rb_first_postorder(root), typeof(*pos), field); \
tools/include/linux/rbtree.h
99
pos && ({ n = rb_entry_safe(rb_next_postorder(&pos->field), \