arch/powerpc/kernel/eeh_cache.c
103
n = rb_first(&cache->rb_root);
arch/powerpc/kernel/eeh_cache.c
218
n = rb_first(&pci_io_addr_cache_root.rb_root);
arch/powerpc/kernel/eeh_cache.c
270
for (n = rb_first(&pci_io_addr_cache_root.rb_root); n; n = rb_next(n)) {
arch/x86/events/intel/uncore.c
1085
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
arch/x86/events/intel/uncore.c
1368
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_discovery.c
182
for (node = rb_first(root); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_discovery.c
445
node = rb_first(&type->units);
arch/x86/events/intel/uncore_discovery.c
752
for (node = rb_first(&discovery_tables); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_snbep.c
6117
node = rb_first(type->boxes);
arch/x86/events/intel/uncore_snbep.c
6345
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
block/bfq-cgroup.c
830
child_entity = bfq_entity_of(rb_first(child_active));
block/bfq-cgroup.c
855
while ((entity = bfq_entity_of(rb_first(active))))
block/bfq-iosched.c
1036
rbnext = rb_first(&bfqq->sort_list);
drivers/android/binder.c
1050
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n)) {
drivers/android/binder.c
5593
for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
5617
for (n = rb_first(&proc->threads); n; n = rb_next(n)) {
drivers/android/binder.c
5632
for (n = rb_first(&proc->nodes); n; n = rb_next(n)) {
drivers/android/binder.c
587
for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
6163
for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
6292
while ((n = rb_first(&proc->threads))) {
drivers/android/binder.c
6304
while ((n = rb_first(&proc->nodes))) {
drivers/android/binder.c
6324
while ((n = rb_first(&proc->refs_by_desc))) {
drivers/android/binder.c
6634
for (n = rb_first(&proc->threads); n; n = rb_next(n))
drivers/android/binder.c
6638
for (n = rb_first(&proc->nodes); n; n = rb_next(n)) {
drivers/android/binder.c
6654
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n))
drivers/android/binder.c
6798
for (n = rb_first(&proc->threads); n; n = rb_next(n))
drivers/android/binder.c
6812
for (n = rb_first(&proc->nodes); n; n = rb_next(n))
drivers/android/binder.c
6820
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
1047
for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
1105
for (n = rb_first(&alloc->allocated_buffers); n != NULL; n = rb_next(n))
drivers/android/binder_alloc.c
437
for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
446
for (n = rb_first(&alloc->free_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
486
for (n = rb_first(&alloc->allocated_buffers); n != NULL;
drivers/android/binder_alloc.c
973
while ((n = rb_first(&alloc->allocated_buffers))) {
drivers/android/tests/binder_alloc_kunit.c
430
n = rb_first(&alloc->allocated_buffers);
drivers/android/tests/binder_alloc_kunit.c
433
n = rb_first(&alloc->free_buffers);
drivers/base/regmap/regcache-rbtree.c
148
for (node = rb_first(&rbtree_ctx->root); node != NULL;
drivers/base/regmap/regcache-rbtree.c
211
next = rb_first(&rbtree_ctx->root);
drivers/base/regmap/regcache-rbtree.c
478
for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
drivers/base/regmap/regcache-rbtree.c
520
for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
drivers/base/regmap/regmap-debugfs.c
640
next = rb_first(&map->range_tree);
drivers/base/regmap/regmap.c
549
next = rb_first(&map->range_tree);
drivers/block/xen-blkback/blkback.c
151
for ((pos) = container_of(rb_first((rbtree)), typeof(*(pos)), node), \
drivers/fpga/dfl-afu-dma-region.c
206
struct rb_node *node = rb_first(&afu->dma_regions);
drivers/gpu/drm/i915/gt/intel_engine.h
385
for ((engine__) = rb_to_uabi_engine(rb_first(&(i915__)->uabi_engines));\
drivers/gpu/drm/i915/gvt/kvmgt.c
312
node = rb_first(&vgpu->gfn_cache);
drivers/gpu/drm/i915/i915_scheduler.c
48
rb_first(&sched_engine->queue.rb_root));
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1063
while ((node = rb_first(&vmm->root))) {
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
874
for (rb_itr = rb_first(&vbo->res_tree); rb_itr;
drivers/infiniband/core/multicast.c
779
for (node = rb_first(&port->table); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx4/cm.c
549
while (rb_first(sl_id_map)) {
drivers/infiniband/hw/mlx4/cm.c
551
rb_entry(rb_first(sl_id_map),
drivers/infiniband/hw/mlx4/cm.c
560
nd = rb_first(sl_id_map);
drivers/infiniband/hw/mlx4/mcg.c
1098
for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p))
drivers/infiniband/hw/mlx4/mcg.c
1112
while ((p = rb_first(&ctx->mcg_table)) != NULL) {
drivers/infiniband/hw/mlx4/mcg.c
1241
for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) {
drivers/infiniband/hw/mlx5/mr.c
1027
for (node = rb_first(root); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx5/mr.c
494
for (node = rb_first(&cache->rb_root); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx5/mr.c
947
node = rb_first(root);
drivers/infiniband/hw/mlx5/mr.c
999
for (node = rb_first(root); node; node = rb_next(node)) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
700
n = rb_first(&priv->path_tree);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
1007
n = rb_first(&priv->multicast_tree);
drivers/iommu/iova.c
528
for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) {
drivers/md/bcache/util.h
514
container_of_or_null(rb_first(root), type, member)
drivers/md/dm-crypt.c
1975
io = crypt_io_from_node(rb_first(&write_tree));
drivers/md/dm-dust.c
245
nnode = rb_first(tree);
drivers/md/dm-dust.c
299
for (node = rb_first(&badblocklist); node; node = rb_next(node)) {
drivers/md/dm-ioctl.c
349
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-ioctl.c
606
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-ioctl.c
631
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-pcache/cache_key.c
384
prev_node = rb_first(&cache_subtree->root);
drivers/md/dm-pcache/cache_key.c
663
node = rb_first(&cache_subtree->root);
drivers/md/dm-pcache/cache_key.c
873
node = rb_first(&cache_subtree->root);
drivers/md/dm-pcache/cache_writeback.c
156
node = rb_first(&cache_subtree->root);
drivers/md/dm-snap.c
1748
next = rb_first(&s->out_of_order_tree);
drivers/md/dm-thin.c
2147
for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) {
drivers/md/dm-writecache.c
2009
e = container_of(rb_first(&wc->tree), struct wc_entry, rb_node);
drivers/md/dm-writecache.c
731
next = rb_first(&wc->freetree);
drivers/mtd/mtdswap.c
241
p = rb_first(root);
drivers/mtd/mtdswap.c
576
eb = rb_entry(rb_first(clean_root), struct swap_eb, rb);
drivers/mtd/mtdswap.c
79
#define MTDSWAP_ECNT_MIN(rbroot) (rb_entry(rb_first(rbroot), struct swap_eb, \
drivers/mtd/mtdswap.c
860
eb = rb_entry(rb_first(rp), struct swap_eb, rb);
drivers/mtd/ubi/fastmap-wl.c
404
e = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/ubi.h
1054
for (rb = rb_first(root), \
drivers/mtd/ubi/wl.c
1061
e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
333
e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
368
first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
741
e1 = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
762
e1 = rb_entry(rb_first(&ubi->scrub), struct ubi_wl_entry, u.rb);
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
5030
for (p = rb_first(root); p; p = rb_next(p)) {
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
447
p = rb_first(root);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
508
p = rb_first(root);
drivers/platform/surface/aggregator/controller.c
2573
for (n = rb_first(&nf->refcount); n; n = rb_next(n)) {
drivers/platform/surface/aggregator/controller.c
2617
for (n = rb_first(&nf->refcount); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1330
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1820
n = rb_first(&iommu->dma_list);
drivers/vfio/vfio_iommu_type1.c
1897
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
2433
while ((node = rb_first(&iommu->dma_list)))
drivers/vfio/vfio_iommu_type1.c
2441
n = rb_first(&iommu->dma_list);
drivers/vfio/vfio_iommu_type1.c
2448
p = rb_first(&dma->pfn_list);
drivers/vfio/vfio_iommu_type1.c
273
for (p = rb_first(&dma->pfn_list); p; p = rb_next(p)) {
drivers/vfio/vfio_iommu_type1.c
285
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
296
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
321
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
fs/afs/cell.c
908
for (cursor = rb_first(&net->cells); cursor; cursor = rb_next(cursor)) {
fs/afs/server.c
518
for (rb = rb_first(&cell->fs_servers); rb; rb = rb_next(rb)) {
fs/btrfs/backref.c
3173
while ((node = rb_entry_safe(rb_first(&cache->rb_root),
fs/btrfs/extent-tree.c
2183
n = rb_first(root);
fs/btrfs/extent_map.c
1161
node = rb_first(&tree->root);
fs/btrfs/extent_map.c
779
node = rb_first(&tree->root);
fs/btrfs/extent_map.c
790
node = rb_first(&tree->root);
fs/btrfs/free-space-cache.c
1079
struct rb_node *node = rb_first(&ctl->free_space_offset);
fs/btrfs/free-space-cache.c
1091
node = rb_first(&cluster->root);
fs/btrfs/free-space-cache.c
1113
node = rb_first(&cluster->root);
fs/btrfs/free-space-cache.c
2341
node = rb_first(&cluster->root);
fs/btrfs/free-space-cache.c
2932
for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) {
fs/btrfs/free-space-cache.c
2994
node = rb_first(&cluster->root);
fs/btrfs/free-space-cache.c
3070
node = rb_first(&ctl->free_space_offset);
fs/btrfs/free-space-cache.c
3244
node = rb_first(&cluster->root);
fs/btrfs/free-space-cache.c
906
while (!ret && (n = rb_first(&ctl->free_space_offset)) != NULL) {
fs/btrfs/inode.c
10185
node = rb_first(&fs_info->swapfile_pins);
fs/btrfs/inode.c
5558
node = rb_first(&io_tree->state);
fs/btrfs/ordered-data.c
1039
for (n = rb_first(&inode->ordered_tree); n; n = rb_next(n)) {
fs/btrfs/qgroup.c
4019
for (n = rb_first(&fs_info->qgroup_tree); n; n = rb_next(n)) {
fs/btrfs/qgroup.c
646
for (node = rb_first(&fs_info->qgroup_tree); node; node = rb_next(node)) {
fs/btrfs/qgroup.c
680
while ((n = rb_first(&fs_info->qgroup_tree))) {
fs/btrfs/ref-verify.c
224
while ((n = rb_first(&be->roots))) {
fs/btrfs/ref-verify.c
230
while((n = rb_first(&be->refs))) {
fs/btrfs/ref-verify.c
626
for (n = rb_first(&be->refs); n; n = rb_next(n)) {
fs/btrfs/ref-verify.c
634
for (n = rb_first(&be->roots); n; n = rb_next(n)) {
fs/btrfs/ref-verify.c
897
while ((n = rb_first(&fs_info->block_tree))) {
fs/btrfs/relocation.c
1918
while ((rb_node = rb_first(blocks))) {
fs/btrfs/send.c
8248
n = rb_first(&sctx->pending_dir_moves);
fs/btrfs/send.c
8265
n = rb_first(&sctx->waiting_dir_moves);
fs/btrfs/send.c
8276
n = rb_first(&sctx->orphan_dirs);
fs/btrfs/tests/extent-io-tests.c
96
node = rb_first(&tree->state);
fs/btrfs/tests/extent-map-tests.c
23
node = rb_first(&em_tree->root);
fs/btrfs/tests/extent-map-tests.c
569
for (i = 0, n = rb_first(&em_tree->root);
fs/btrfs/tree-mod-log.c
125
for (node = rb_first(tm_root); node; node = next) {
fs/ceph/addr.c
2610
n = rb_first(&mdsc->pool_perm_tree);
fs/ceph/caps.c
1092
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
1374
p = rb_first(&ci->i_caps);
fs/ceph/caps.c
2124
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
823
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
854
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
907
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
933
for (q = rb_first(&ci->i_caps); q != p;
fs/ceph/caps.c
973
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/debugfs.c
61
for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) {
fs/ceph/inode.c
3203
p = rb_first(&ci->i_caps);
fs/ceph/inode.c
531
rb_node = rb_first(&ci->i_fragtree);
fs/ceph/inode.c
550
rb_node = rb_first(&ci->i_fragtree);
fs/ceph/inode.c
774
while ((n = rb_first(&ci->i_fragtree)) != NULL) {
fs/ceph/mds_client.c
1415
cap = rb_entry(rb_first(&ci->i_caps), struct ceph_cap, ci_node);
fs/ceph/mds_client.c
1801
p = rb_first(&mdsc->request_tree);
fs/ceph/mds_client.c
2617
return rb_entry(rb_first(&mdsc->request_tree),
fs/ceph/mds_client.c
3684
struct rb_node *p = rb_first(&mdsc->request_tree);
fs/ceph/mds_client.c
4503
p = rb_first(&mdsc->request_tree);
fs/ceph/mds_client.c
4841
for (p = rb_first(&mdsc->snap_realms); p; p = rb_next(p)) {
fs/ceph/quota.c
190
node = rb_first(&mdsc->quotarealms_inodes);
fs/ceph/snap.c
1325
while ((p = rb_first(&mdsc->snapid_map_tree))) {
fs/ceph/xattr.c
744
p = rb_first(&ci->i_xattrs.index);
fs/ceph/xattr.c
768
p = rb_first(&ci->i_xattrs.index);
fs/ceph/xattr.c
920
p = rb_first(&ci->i_xattrs.index);
fs/ext2/balloc.c
213
n = rb_first(root);
fs/ext4/block_validity.c
135
node = rb_first(&system_blks->root);
fs/ext4/dir.c
588
info->curr_node = rb_first(&info->root);
fs/ext4/dir.c
610
info->curr_node = rb_first(&info->root);
fs/ext4/extents_status.c
1981
node = rb_first(&tree->root);
fs/ext4/extents_status.c
2003
node = rb_first(&tree->root);
fs/ext4/extents_status.c
219
node = rb_first(&tree->root);
fs/ext4/mballoc.c
5639
for (iter = rb_first(&ei->i_prealloc_node); iter;
fs/fuse/dev.c
2422
p = rb_first(&fc->polled_files);
fs/fuse/dir.c
169
node = rb_first(&dentry_hash[i].tree);
fs/fuse/dir.c
187
node = rb_first(&dentry_hash[i].tree);
fs/gfs2/rgrp.c
2315
for (n = rb_first(&rgd->rd_rstree); n; n = rb_next(&trs->rs_node)) {
fs/gfs2/rgrp.c
568
n = rb_first(&sdp->sd_rindex_tree);
fs/gfs2/rgrp.c
590
n = rb_first(&sdp->sd_rindex_tree);
fs/gfs2/rgrp.c
709
while ((n = rb_first(&rgd->rd_rstree))) {
fs/gfs2/rgrp.c
722
while ((n = rb_first(&sdp->sd_rindex_tree))) {
fs/jffs2/nodelist.h
329
struct rb_node *node = rb_first(root);
fs/jffs2/nodelist.h
361
#define tn_first(list) rb_entry(rb_first(list), struct jffs2_tmp_dnode_info, rb)
fs/kernfs/dir.c
1352
rbn = rb_first(&pos->dir.children);
fs/nfs/blocklayout/extent_tree.c
22
struct rb_node *node = rb_first(root);
fs/nfs/dir.c
2984
while ((n = rb_first(root_node)) != NULL) {
fs/nfs/nfs4state.c
166
for (pos = rb_first(&server->state_owners);
fs/nfs/nfs4state.c
1691
for (pos = rb_first(&server->state_owners);
fs/nfs/nfs4state.c
1743
for (pos = rb_first(&server->state_owners);
fs/nfs/nfs4state.c
1858
for (pos = rb_first(&server->state_owners);
fs/ntfs3/bitmap.c
1077
e = rb_entry(rb_first(&wnd->count_tree), struct e_node, count.node);
fs/ntfs3/bitmap.c
131
node = rb_first(&wnd->start_tree);
fs/ntfs3/bitmap.c
250
n = rb_first(&wnd->start_tree);
fs/ntfs3/bitmap.c
419
n3 = rb_first(&wnd->count_tree);
fs/ntfs3/frecord.c
112
for (node = rb_first(&ni->mi_tree); node;) {
fs/ntfs3/frecord.c
1211
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ntfs3/frecord.c
1574
for (node = rb_first(&ni->mi_tree); node;) {
fs/ntfs3/frecord.c
2829
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ntfs3/frecord.c
3197
for (node = rb_first(&ni->mi_tree); node; node = next) {
fs/ntfs3/frecord.c
944
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ocfs2/reservations.c
138
node = rb_first(&resmap->m_reservations);
fs/ocfs2/reservations.c
363
node = rb_first(&resmap->m_reservations);
fs/ocfs2/reservations.c
522
next = rb_first(root);
fs/ocfs2/reservations.c
83
node = rb_first(&resmap->m_reservations);
fs/proc/generic.c
58
return rb_entry_safe(rb_first(&dir->subdir), struct proc_dir_entry,
fs/proc/nommu.c
86
for (p = rb_first(&nommu_region_tree); p; p = rb_next(p))
fs/proc/proc_sysctl.c
390
ctl_node = first_usable_entry(rb_first(&dir->root));
fs/smb/client/cached_dir.c
549
for (node = rb_first(root); node; node = rb_next(node)) {
fs/smb/client/connect.c
3988
while ((node = rb_first(root))) {
fs/smb/client/connect.c
4464
node = rb_first(root);
fs/smb/client/misc.c
572
for (node = rb_first(root); node; node = rb_next(node)) {
fs/ubifs/debug.c
2178
struct rb_node *this = rb_first(&fsckd->inodes);
fs/ubifs/debug.c
657
for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) {
fs/ubifs/debug.c
745
for (rb = rb_first((struct rb_root *)&c->buds); rb; rb = rb_next(rb)) {
fs/ubifs/log.c
304
p = rb_first(&c->buds);
fs/ubifs/recovery.c
1528
struct rb_node *this = rb_first(&c->size_tree);
fs/xattr.c
1468
for (rbp = rb_first(&xattrs->rb_root); rbp; rbp = rb_next(rbp)) {
fs/xattr.c
1547
rbp = rb_first(&xattrs->rb_root);
fs/xfs/scrub/bitmap.c
338
for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \
fs/xfs/scrub/bitmap.c
63
for ((bn) = rb_entry_safe(rb_first(&(bitmap)->xb_root.rb_root), \
include/linux/skbuff.h
4128
#define skb_rb_first(root) rb_to_skb(rb_first(root))
kernel/events/core.c
1958
for (event = rb_entry_safe(rb_first(&((groups)->tree)), \
kernel/locking/ww_mutex.h
96
struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root);
kernel/power/wakelock.c
46
for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) {
kernel/trace/trace_stat.c
185
node = rb_first(&session->stat_root);
kernel/trace/trace_stat.c
200
return rb_first(&session->stat_root);
lib/rbtree_test.c
198
for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) {
lib/rbtree_test.c
224
for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) {
lib/rbtree_test.c
291
for (node = rb_first(&root.rb_root); node; node = rb_next(node))
lib/rbtree_test.c
305
node = rb_first(&root.rb_root);
mm/ksm.c
3352
node = rb_first(root_stable_tree + nid);
mm/ksm.c
3359
node = rb_first(root_stable_tree + nid);
mm/mempolicy.c
3278
next = rb_first(&sp->root);
mm/nommu.c
1085
for (rb = rb_first(&nommu_region_tree); rb; rb = rb_next(rb)) {
mm/nommu.c
443
lastp = rb_first(&nommu_region_tree);
mm/shmem_quota.c
101
node = rb_first(root);
mm/swapfile.c
268
struct rb_node *rb = rb_first(&sis->swap_extent_root);
net/802/garp.c
212
for (node = rb_first(&app->gid);
net/802/garp.c
399
for (node = rb_first(&app->gid);
net/802/mrp.c
301
for (node = rb_first(&app->mad);
net/802/mrp.c
584
for (node = rb_first(&app->mad);
net/bridge/br_multicast_eht.c
184
while ((node = rb_first(&eht_set->entry_tree))) {
net/bridge/br_multicast_eht.c
201
while ((node = rb_first(&pg->eht_set_tree))) {
net/ceph/auth_x.c
989
while ((p = rb_first(&xi->ticket_handlers)) != NULL) {
net/ceph/ceph_common.c
620
for (n = rb_first(&opt->crush_locs); ; ) {
net/ceph/debugfs.c
108
for (n = rb_first(&map->primary_temp); n; n = rb_next(n)) {
net/ceph/debugfs.c
115
for (n = rb_first(&map->pg_upmap); n; n = rb_next(n)) {
net/ceph/debugfs.c
126
for (n = rb_first(&map->pg_upmap_items); n; n = rb_next(n)) {
net/ceph/debugfs.c
166
for (rp = rb_first(&monc->generic_request_tree); rp; rp = rb_next(rp)) {
net/ceph/debugfs.c
245
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/debugfs.c
271
for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) {
net/ceph/debugfs.c
332
for (n = rb_first(&osd->o_backoffs_by_id); n; n = rb_next(n)) {
net/ceph/debugfs.c
358
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
366
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
374
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
74
for (n = rb_first(&map->pg_pools); n; n = rb_next(n)) {
net/ceph/debugfs.c
97
for (n = rb_first(&map->pg_temp); n; n = rb_next(n)) {
net/ceph/mon_client.c
1070
for (p = rb_first(&monc->generic_request_tree); p; p = rb_next(p)) {
net/ceph/osd_client.c
1179
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
1182
for (p = rb_first(&osd->o_requests); p; ) {
net/ceph/osd_client.c
1192
for (p = rb_first(&osdc->homeless_osd.o_requests); p; ) {
net/ceph/osd_client.c
1348
for (n = rb_first(&osd->o_requests); n; ) {
net/ceph/osd_client.c
1358
for (n = rb_first(&osd->o_linger_requests); n; ) {
net/ceph/osd_client.c
1399
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
1495
for (n = rb_first(&osdc->osdmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
1975
rb_entry(rb_first(&osd->o_backoff_mappings),
net/ceph/osd_client.c
1980
rb_entry(rb_first(&spg->backoffs),
net/ceph/osd_client.c
3447
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
3451
for (p = rb_first(&osd->o_requests); p; ) {
net/ceph/osd_client.c
3469
for (p = rb_first(&osd->o_linger_requests); p; p = rb_next(p)) {
net/ceph/osd_client.c
3488
for (p = rb_first(&osdc->homeless_osd.o_requests); p; ) {
net/ceph/osd_client.c
3885
for (n = rb_first(&osdc->osdmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
3938
for (n = rb_first(&osd->o_linger_requests); n; ) {
net/ceph/osd_client.c
3974
for (n = rb_first(&osd->o_requests); n; ) {
net/ceph/osd_client.c
4034
for (n = rb_first(&newmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
4060
for (n = rb_first(&osdc->osds); n; ) {
net/ceph/osd_client.c
4086
for (n = rb_first(need_resend); n; ) {
net/ceph/osd_client.c
4101
for (n = rb_first(need_resend); n; ) {
net/ceph/osd_client.c
4253
for (n = rb_first(&osd->o_requests); n; ) {
net/ceph/osd_client.c
4266
for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
4495
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
4728
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
4732
for (p = rb_first(&osd->o_requests); p; p = rb_next(p)) {
net/ceph/osd_client.c
5186
for (n = rb_first(&osdc->osds); n; ) {
net/ceph/osd_client.c
5278
struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds),
net/ceph/osdmap.c
1144
rb_entry(rb_first(&map->pg_temp),
net/ceph/osdmap.c
1151
rb_entry(rb_first(&map->primary_temp),
net/ceph/osdmap.c
1158
rb_entry(rb_first(&map->pg_upmap),
net/ceph/osdmap.c
1165
rb_entry(rb_first(&map->pg_upmap_items),
net/ceph/osdmap.c
1172
rb_entry(rb_first(&map->pg_pools),
net/ceph/osdmap.c
223
rb_entry(rb_first(root), struct crush_name_node, cn_node);
net/ceph/osdmap.c
273
rb_entry(rb_first(&c->choose_args),
net/ceph/osdmap.c
2970
struct rb_node *n1 = rb_first(locs1);
net/ceph/osdmap.c
2971
struct rb_node *n2 = rb_first(locs2);
net/ceph/osdmap.c
2996
rb_entry(rb_first(locs), struct crush_loc_node, cl_node);
net/ceph/osdmap.c
765
for (rbp = rb_first(&map->pg_pools); rbp; rbp = rb_next(rbp)) {
net/core/skbuff.c
4037
struct rb_node *p = rb_first(root);
net/ipv4/inet_fragment.c
304
struct rb_node *p = rb_first(root);
net/ipv4/inetpeer.c
276
struct rb_node *p = rb_first(&base->rb_root);
net/ipv4/nexthop.c
2706
while ((node = rb_first(root))) {
net/ipv4/nexthop.c
3561
node = rb_first(root);
net/ipv4/nexthop.c
3959
for (node = rb_first(root); node; node = rb_next(node)) {
net/ipv4/tcp.c
3369
struct rb_node *p = rb_first(&sk->tcp_rtx_queue);
net/ipv4/tcp_input.c
5334
p = rb_first(&tp->out_of_order_queue);
net/mptcp/protocol.c
781
p = rb_first(&msk->out_of_order_queue);
net/netfilter/nf_conncount.c
563
for (node = rb_first(root); node != NULL; node = rb_next(node)) {
net/netfilter/nf_conncount.c
578
node = rb_first(root);
net/netfilter/nf_conncount.c
664
while ((node = rb_first(r)) != NULL) {
net/netfilter/nft_set_rbtree.c
377
first = rb_first(&priv->root);
net/netfilter/nft_set_rbtree.c
850
for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) {
net/netfilter/nft_set_rbtree.c
902
for (node = rb_first(&priv->root); node ; node = next) {
net/rds/cong.c
405
while ((node = rb_first(&rds_cong_tree))) {
net/rds/rdma.c
135
while ((node = rb_first(&rs->rs_rdma_keys))) {
net/sched/sch_fq.c
628
while ((p = rb_first(&q->delayed)) != NULL) {
net/sched/sch_fq.c
786
struct rb_node *p = rb_first(&flow->t_root);
net/sched/sch_fq.c
818
while ((p = rb_first(root)) != NULL) {
net/sched/sch_fq.c
850
while ((op = rb_first(oroot)) != NULL) {
net/sched/sch_hfsc.c
231
for (n = rb_first(&q->eligible); n != NULL; n = rb_next(n)) {
net/sched/sch_hfsc.c
247
n = rb_first(&q->eligible);
net/sched/sch_hfsc.c
295
for (n = rb_first(&cl->vt_tree); n != NULL; n = rb_next(n)) {
net/sched/sch_hfsc.c
660
struct rb_node *n = rb_first(&cl->cf_tree);
net/sched/sch_htb.c
760
struct rb_node *p = rb_first(wait_pq);
net/sched/sch_netem.c
375
struct rb_node *p = rb_first(&q->t_root);
net/smc/smc_core.c
1125
for (node = rb_first(&lgr->conns_all); node; node = rb_next(node)) {
net/smc/smc_core.c
1573
node = rb_first(&lgr->conns_all);
net/smc/smc_core.c
1584
node = rb_first(&lgr->conns_all);
net/tipc/group.c
269
for (n = rb_first(&grp->members); n; n = rb_next(n)) {
net/tipc/group.c
402
for (n = rb_first(&grp->members); n; n = rb_next(n)) {
net/tipc/name_table.c
1038
for (n = rb_first(&sc->ranges); n; n = rb_next(n)) {
net/tipc/name_table.c
746
for (n = rb_first(&sc->ranges); n; n = rb_next(n)) {
net/xfrm/xfrm_policy.c
1026
while ((rnode = rb_first(&v->root)) != NULL) {
net/xfrm/xfrm_policy.c
1121
struct rb_node *rn = rb_first(r);
rust/helpers/rbtree.c
14
return rb_first(root);
security/apparmor/include/label.h
79
for ((N) = rb_first(&(LS)->root); (N); (N) = rb_next(N))
security/apparmor/label.c
1962
for (node = rb_first(&ls->root); node; node = rb_first(&ls->root)) {
security/keys/gc.c
214
cursor = rb_first(&key_serial_tree);
security/keys/proc.c
270
struct rb_node *n = rb_first(r);
tools/bpf/resolve_btfids/main.c
361
next = rb_first(root);
tools/bpf/resolve_btfids/main.c
724
next = rb_first(root);
tools/bpf/resolve_btfids/main.c
749
next = rb_first(&obj->sets);
tools/bpf/resolve_btfids/main.c
963
for (next = rb_first(&obj->sets); next; next = rb_next(next)) {
tools/include/linux/rbtree.h
55
extern struct rb_node *rb_first(const struct rb_root *);
tools/perf/builtin-kmem.c
1011
next = rb_first(root);
tools/perf/builtin-kmem.c
1064
struct rb_node *next = rb_first(&page_alloc_sorted);
tools/perf/builtin-kmem.c
1113
struct rb_node *next = rb_first(&page_caller_sorted);
tools/perf/builtin-kmem.c
1300
node = rb_first(root);
tools/perf/builtin-kmem.c
1347
node = rb_first(root);
tools/perf/builtin-lock.c
1346
node = rb_first(&thread_stats);
tools/perf/tests/code-reading.c
95
struct rb_node *node = rb_first(root);
tools/perf/tests/hists_cumulate.c
216
for (node = rb_first(root), i = 0;
tools/perf/tests/hists_cumulate.c
238
cnode = rb_entry(rb_first(root), struct callchain_node, rb_node);
tools/perf/ui/browser.c
145
nd = rb_first(root);
tools/perf/ui/browser.c
174
browser->top = rb_first(browser->entries);
tools/perf/ui/browsers/annotate.c
951
nd = rb_first(&browser->entries);
tools/perf/ui/browsers/hists.c
1025
node = rb_first(root);
tools/perf/ui/browsers/hists.c
1114
node = rb_first(root);
tools/perf/ui/browsers/hists.c
182
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
1964
nd = hists__filter_entries(rb_first(browser->entries),
tools/perf/ui/browsers/hists.c
2194
struct rb_node *nd = hists__filter_entries(rb_first(browser->b.entries),
tools/perf/ui/browsers/hists.c
265
for (nd = rb_first(chain); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
331
struct rb_node *nd = rb_first(&node->rb_root);
tools/perf/ui/browsers/hists.c
333
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
370
struct rb_node *nd = rb_first(root);
tools/perf/ui/browsers/hists.c
373
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
516
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
557
for (nd = rb_first(chain); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
921
node = rb_first(root);
tools/perf/ui/browsers/map.c
122
for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
101
bool has_single_node = (rb_first(root) == rb_last(root));
tools/perf/ui/gtk/hists.c
103
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
165
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
224
bool has_single_node = (rb_first(root) == rb_last(root));
tools/perf/ui/gtk/hists.c
226
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/stdio/hist.c
129
node = rb_first(root);
tools/perf/ui/stdio/hist.c
234
node = rb_first(root);
tools/perf/ui/stdio/hist.c
315
struct rb_node *rb_node = rb_first(tree);
tools/perf/ui/stdio/hist.c
366
struct rb_node *rb_node = rb_first(tree);
tools/perf/util/annotate-data.c
1771
struct rb_node *node = rb_first(root);
tools/perf/util/annotate-data.c
712
struct rb_node *node = rb_first(root);
tools/perf/util/annotate.c
1136
node = rb_first(src_root);
tools/perf/util/annotate.c
1162
node = rb_first(root);
tools/perf/util/block-range.c
18
for (rb = rb_first(&block_ranges.root); rb; rb = rb_next(rb)) {
tools/perf/util/bpf_lock_contention.c
844
struct rb_node *node = rb_first(&con->cgroups);
tools/perf/util/callchain.c
1063
n = rb_first(&src->rb_root_in);
tools/perf/util/callchain.c
1307
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
1515
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
1540
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
1767
struct rb_node *rb_node = rb_first(root);
tools/perf/util/callchain.c
437
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
468
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
499
n = rb_first(&node->rb_root_in);
tools/perf/util/callchain.c
564
n = rb_first(&new->rb_root_in);
tools/perf/util/cgroup.c
591
node = rb_first(&env->cgroups.tree);
tools/perf/util/dso.c
898
struct rb_node *next = rb_first(root);
tools/perf/util/env.c
100
first = rb_first(&env->bpf_progs.infos);
tools/perf/util/env.c
182
next = rb_first(root);
tools/perf/util/env.c
198
next = rb_first(root);
tools/perf/util/header.c
1046
next = rb_first(root);
tools/perf/util/header.c
1089
next = rb_first(root);
tools/perf/util/header.c
1993
next = rb_first(root);
tools/perf/util/header.c
2020
next = rb_first(root);
tools/perf/util/intel-pt.c
368
n = rb_first(&pt->vmcs_info);
tools/perf/util/stream.c
102
struct rb_node *rb_node = rb_first(root);
tools/perf/util/symbol.c
1060
struct rb_node *next = rb_first(modules);
tools/perf/util/symbol.c
1143
from_node = rb_first(&from_modules);
tools/perf/util/symbol.c
1144
to_node = rb_first(&to_modules);