Symbol: rb_next
arch/powerpc/kernel/eeh_cache.c
111
n = rb_next(n);
arch/powerpc/kernel/eeh_cache.c
230
n = rb_next(n);
arch/powerpc/kernel/eeh_cache.c
270
for (n = rb_first(&pci_io_addr_cache_root.rb_root); n; n = rb_next(n)) {
arch/x86/events/intel/uncore.c
1085
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
arch/x86/events/intel/uncore.c
1368
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_discovery.c
131
for (; pos; pos = rb_next(pos)) {
arch/x86/events/intel/uncore_discovery.c
182
for (node = rb_first(root); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_discovery.c
752
for (node = rb_first(&discovery_tables); node; node = rb_next(node)) {
arch/x86/events/intel/uncore_snbep.c
6345
for (node = rb_first(type->boxes); node; node = rb_next(node)) {
block/bfq-iosched.c
1021
struct rb_node *rbnext = rb_next(&last->rb_node);
block/bfq-iosched.c
2705
node = rb_next(&__bfqq->pos_node);
block/bfq-wf2q.c
356
next = rb_next(&entity->rb_node);
block/bfq-wf2q.c
555
deepest = rb_next(node);
block/elevator.c
879
struct rb_node *rbnext = rb_next(&rq->rb_node);
drivers/android/binder.c
1050
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n)) {
drivers/android/binder.c
5593
for (n = rb_first(&proc->nodes); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
5617
for (n = rb_first(&proc->threads); n; n = rb_next(n)) {
drivers/android/binder.c
5632
for (n = rb_first(&proc->nodes); n; n = rb_next(n)) {
drivers/android/binder.c
587
for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
6163
for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) {
drivers/android/binder.c
6634
for (n = rb_first(&proc->threads); n; n = rb_next(n))
drivers/android/binder.c
6638
for (n = rb_first(&proc->nodes); n; n = rb_next(n)) {
drivers/android/binder.c
6654
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n))
drivers/android/binder.c
6798
for (n = rb_first(&proc->threads); n; n = rb_next(n))
drivers/android/binder.c
6812
for (n = rb_first(&proc->nodes); n; n = rb_next(n))
drivers/android/binder.c
6820
for (n = rb_first(&proc->refs_by_desc); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
1047
for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
1105
for (n = rb_first(&alloc->allocated_buffers); n != NULL; n = rb_next(n))
drivers/android/binder_alloc.c
437
for (n = rb_first(&alloc->allocated_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
446
for (n = rb_first(&alloc->free_buffers); n; n = rb_next(n)) {
drivers/android/binder_alloc.c
487
n = rb_next(n)) {
drivers/base/regmap/regcache-rbtree.c
149
node = rb_next(node)) {
drivers/base/regmap/regcache-rbtree.c
214
next = rb_next(&rbtree_node->node);
drivers/base/regmap/regcache-rbtree.c
478
for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
drivers/base/regmap/regcache-rbtree.c
520
for (node = rb_first(&rbtree_ctx->root); node; node = rb_next(node)) {
drivers/base/regmap/regmap-debugfs.c
649
next = rb_next(&range_node->node);
drivers/base/regmap/regmap.c
552
next = rb_next(&range_node->node);
drivers/block/drbd/drbd_interval.c
154
node = rb_next(&i->rb);
drivers/block/xen-blkback/blkback.c
152
(n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL; \
drivers/block/xen-blkback/blkback.c
155
(n) = (&(pos)->node != NULL) ? rb_next(&(pos)->node) : NULL)
drivers/block/xen-blkback/blkback.c
271
!rb_next(&persistent_gnt->node)) {
drivers/dma-buf/sw_sync.c
293
parent = rb_next(&pt->node);
drivers/fpga/dfl-afu-dma-region.c
225
node = rb_next(node);
drivers/gpu/drm/i915/gt/intel_engine.h
387
(engine__) = rb_to_uabi_engine(rb_next(&(engine__)->uabi_node)))
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4113
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4134
for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5511
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/i915_active.c
798
for (p = prev; p; p = rb_next(p)) {
drivers/gpu/drm/i915/i915_scheduler.c
51
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1002
engine = rb_entry_safe(rb_next(&engine->uabi_node),
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
1813
} while ((node = rb_next(node)));
drivers/gpu/drm/scheduler/sched_main.c
319
for (rb = rb_first_cached(&rq->rb_tree_root); rb; rb = rb_next(rb)) {
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
875
rb_itr = rb_next(rb_itr)) {
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1139
cur = rb_next(&found->mob_node);
drivers/infiniband/core/multicast.c
779
for (node = rb_first(&port->table); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx4/cm.c
564
nd = rb_next(nd);
drivers/infiniband/hw/mlx4/mcg.c
1098
for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p))
drivers/infiniband/hw/mlx4/mcg.c
1241
for (p = rb_first(&ctx->mcg_table); p; p = rb_next(p)) {
drivers/infiniband/hw/mlx5/mr.c
1027
for (node = rb_first(root); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx5/mr.c
494
for (node = rb_first(&cache->rb_root); node; node = rb_next(node)) {
drivers/infiniband/hw/mlx5/mr.c
950
node = rb_next(node);
drivers/infiniband/hw/mlx5/mr.c
999
for (node = rb_first(root); node; node = rb_next(node)) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
712
n = rb_next(n);
drivers/infiniband/ulp/ipoib/ipoib_multicast.c
1025
n = rb_next(n);
drivers/iommu/iova.c
528
for (node = rb_first(&iovad->rbroot); node; node = rb_next(node)) {
drivers/iommu/iova.c
89
iovad->cached32_node = rb_next(&free->node);
drivers/iommu/iova.c
96
iovad->cached_node = rb_next(&free->node);
drivers/md/bcache/util.h
520
container_of_or_null(rb_next(&(ptr)->member), typeof(*ptr), member)
drivers/md/dm-dust.c
253
nnode = rb_next(node);
drivers/md/dm-dust.c
299
for (node = rb_first(&badblocklist); node; node = rb_next(node)) {
drivers/md/dm-integrity.c
1373
next = rb_next(&node->node);
drivers/md/dm-integrity.c
1389
next = rb_next(&node->node);
drivers/md/dm-ioctl.c
349
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-ioctl.c
606
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-ioctl.c
631
for (n = rb_first(&name_rb_tree); n; n = rb_next(n)) {
drivers/md/dm-pcache/cache_key.c
335
node_tmp = rb_next(node_tmp);
drivers/md/dm-pcache/cache_key.c
666
node = rb_next(node);
drivers/md/dm-pcache/cache_key.c
876
node = rb_next(node);
drivers/md/dm-pcache/cache_writeback.c
159
node = rb_next(node);
drivers/md/dm-snap.c
1754
next = rb_next(next);
drivers/md/dm-thin.c
2147
for (node = rb_first(&tc->sort_bio_list); node; node = rb_next(node)) {
drivers/md/dm-writecache.c
1400
struct rb_node *next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
2038
next_node = rb_next(&f->rb_node);
drivers/md/dm-writecache.c
2073
next_node = rb_next(&f->rb_node);
drivers/md/dm-writecache.c
628
node = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
643
node = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
728
next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
890
struct rb_node *node = rb_next(&e->rb_node);
drivers/mtd/mtdswap.c
244
p = rb_next(p);
drivers/mtd/ubi/fastmap-wl.c
557
e = rb_entry(rb_next(root->rb_node),
drivers/mtd/ubi/ubi.h
1057
rb = rb_next(rb), \
drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
5030
for (p = rb_first(root); p; p = rb_next(p)) {
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
451
p = rb_next(p);
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
511
p = rb_next(p);
drivers/platform/surface/aggregator/controller.c
2573
for (n = rb_first(&nf->refcount); n; n = rb_next(n)) {
drivers/platform/surface/aggregator/controller.c
2617
for (n = rb_first(&nf->refcount); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1330
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1504
for (n = first_n; n != last_n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1517
n = rb_next(n);
drivers/vfio/vfio_iommu_type1.c
1541
n = rb_next(n);
drivers/vfio/vfio_iommu_type1.c
1822
for (; n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
1897
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
2442
for (; n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
2449
for (; p; p = rb_next(p)) {
drivers/vfio/vfio_iommu_type1.c
273
for (p = rb_first(&dma->pfn_list); p; p = rb_next(p)) {
drivers/vfio/vfio_iommu_type1.c
285
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
296
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
321
for (n = rb_first(&iommu->dma_list); n; n = rb_next(n)) {
drivers/vfio/vfio_iommu_type1.c
661
while ((next = rb_next(next))) {
fs/afs/cell.c
908
for (cursor = rb_first(&net->cells); cursor; cursor = rb_next(cursor)) {
fs/afs/server.c
518
for (rb = rb_first(&cell->fs_servers); rb; rb = rb_next(rb)) {
fs/btrfs/backref.c
1592
node = rb_next(&ref->rbnode);
fs/btrfs/backref.c
897
for (n = rb_first_cached(&head->ref_tree); n; n = rb_next(n)) {
fs/btrfs/block-group.c
2555
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
fs/btrfs/block-group.c
301
node = rb_next(&cache->cache_node);
fs/btrfs/block-group.c
4949
node_bg = rb_next(node_bg);
fs/btrfs/block-group.c
4950
node_chunk = rb_next(node_chunk);
fs/btrfs/defrag.c
188
parent = rb_next(parent);
fs/btrfs/delayed-inode.c
493
struct rb_node *p = rb_next(&item->rb_node);
fs/btrfs/delayed-ref.c
391
struct rb_node *node = rb_next(&ref->ref_node);
fs/btrfs/delayed-ref.c
398
node = rb_next(node);
fs/btrfs/delayed-ref.c
451
node = rb_next(node)) {
fs/btrfs/dev-replace.c
827
struct rb_node *next = rb_next(node);
fs/btrfs/extent-io-tree.c
203
struct rb_node *next = rb_next(&state->rb_node);
fs/btrfs/extent-tree.c
2348
node = rb_next(node)) {
fs/btrfs/extent_map.c
113
parent = rb_next(parent);
fs/btrfs/extent_map.c
1163
struct rb_node *next = rb_next(node);
fs/btrfs/extent_map.c
165
prev = rb_next(prev);
fs/btrfs/extent_map.c
393
rb = rb_next(&em->rb_node);
fs/btrfs/extent_map.c
632
next = rb_next(&em->rb_node);
fs/btrfs/extent_map.c
782
struct rb_node *next = rb_next(node);
fs/btrfs/free-space-cache.c
1111
node = rb_next(node);
fs/btrfs/free-space-cache.c
1733
n = rb_next(n);
fs/btrfs/free-space-cache.c
1797
n = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
2029
for (; node; node = rb_next(node)) {
fs/btrfs/free-space-cache.c
2080
struct rb_node *old_next = rb_next(node);
fs/btrfs/free-space-cache.c
2100
if (use_bytes_index && old_next != rb_next(node))
fs/btrfs/free-space-cache.c
2183
struct rb_node *next = rb_next(&bitmap_info->offset_index);
fs/btrfs/free-space-cache.c
2932
for (n = rb_first(&ctl->free_space_offset); n; n = rb_next(n)) {
fs/btrfs/free-space-cache.c
2999
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
3080
node = rb_next(node);
fs/btrfs/free-space-cache.c
3256
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
3270
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
3444
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
3455
for (node = rb_next(&entry->offset_index); node;
fs/btrfs/free-space-cache.c
3456
node = rb_next(&entry->offset_index)) {
fs/btrfs/free-space-cache.c
3489
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
3749
node = rb_next(&entry->offset_index);
fs/btrfs/free-space-cache.c
4170
node = rb_next(node);
fs/btrfs/free-space-cache.c
4359
n = rb_next(&info->offset_index);
fs/btrfs/free-space-cache.c
4366
n = rb_next(&tmp->offset_index);
fs/btrfs/free-space-tree.c
1201
node = rb_next(node);
fs/btrfs/free-space-tree.c
1262
node = rb_next(node);
fs/btrfs/free-space-tree.c
1373
node = rb_next(node);
fs/btrfs/inode.c
10187
next = rb_next(node);
fs/btrfs/ordered-data.c
1014
node = rb_next(node);
fs/btrfs/ordered-data.c
1039
for (n = rb_first(&inode->ordered_tree); n; n = rb_next(n)) {
fs/btrfs/ordered-data.c
1128
next = rb_next(cur);
fs/btrfs/ordered-data.c
507
node = rb_next(node);
fs/btrfs/ordered-data.c
89
test = rb_next(prev);
fs/btrfs/qgroup.c
4019
for (n = rb_first(&fs_info->qgroup_tree); n; n = rb_next(n)) {
fs/btrfs/qgroup.c
4115
for (node = start; node && ({ next = rb_next(node); 1;}); node = next)
fs/btrfs/qgroup.c
646
for (node = rb_first(&fs_info->qgroup_tree); node; node = rb_next(node)) {
fs/btrfs/ref-verify.c
626
for (n = rb_first(&be->refs); n; n = rb_next(n)) {
fs/btrfs/ref-verify.c
634
for (n = rb_first(&be->roots); n; n = rb_next(n)) {
fs/btrfs/ref-verify.c
946
n = rb_next(n);
fs/btrfs/tests/extent-io-tests.c
105
node = rb_next(node);
fs/btrfs/tests/extent-map-tests.c
571
i++, n = rb_next(n)) {
fs/btrfs/tests/free-space-tests.c
860
node = rb_next(node), i--) {
fs/btrfs/tests/free-space-tests.c
883
node = rb_next(node), i--) {
fs/btrfs/tree-mod-log.c
126
next = rb_next(node);
fs/btrfs/tree-mod-log.c
929
next = rb_next(&tm->node);
fs/btrfs/volumes.c
3244
prev = rb_next(prev);
fs/btrfs/volumes.c
8331
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
fs/ceph/caps.c
1092
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
1377
p = rb_next(p);
fs/ceph/caps.c
2124
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
823
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
854
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
907
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/caps.c
934
q = rb_next(q)) {
fs/ceph/caps.c
973
for (p = rb_first(&ci->i_caps); p; p = rb_next(p)) {
fs/ceph/debugfs.c
61
for (rp = rb_first(&mdsc->request_tree); rp; rp = rb_next(rp)) {
fs/ceph/inode.c
3207
p = rb_next(p);
fs/ceph/inode.c
533
if (frag->frag != ceph_frag_make(0, 0) || rb_next(rb_node))
fs/ceph/inode.c
567
rb_node = rb_next(rb_node);
fs/ceph/inode.c
570
rb_node = rb_next(rb_node);
fs/ceph/inode.c
594
rb_node = rb_next(rb_node);
fs/ceph/mds_client.c
1215
struct rb_node *p = rb_next(&req->r_node);
fs/ceph/mds_client.c
1224
p = rb_next(p);
fs/ceph/mds_client.c
1804
p = rb_next(p);
fs/ceph/mds_client.c
3689
p = rb_next(p);
fs/ceph/mds_client.c
4506
p = rb_next(p);
fs/ceph/mds_client.c
4841
for (p = rb_first(&mdsc->snap_realms); p; p = rb_next(p)) {
fs/ceph/mds_client.c
5885
n = rb_next(&req->r_node);
fs/ceph/xattr.c
756
p = rb_next(p);
fs/ceph/xattr.c
775
p = rb_next(tmp);
fs/ceph/xattr.c
934
p = rb_next(p);
fs/eventpoll.c
1078
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
1204
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
2078
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
913
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
929
next = rb_next(rbp);
fs/ext2/balloc.c
1022
next = rb_next(&my_rsv->rsv_node);
fs/ext2/balloc.c
241
n = rb_next(n);
fs/ext2/balloc.c
791
next = rb_next(&rsv->rsv_node);
fs/ext4/block_validity.c
113
node = rb_next(new_node);
fs/ext4/block_validity.c
141
node = rb_next(node);
fs/ext4/dir.c
619
info->curr_node = rb_next(info->curr_node);
fs/ext4/extents_status.c
1145
node = rb_next(&es1->rb_node);
fs/ext4/extents_status.c
1327
node = rb_next(&pr->rb_node);
fs/ext4/extents_status.c
1396
node = rb_next(&right_es->rb_node);
fs/ext4/extents_status.c
1407
node = rb_next(&es->rb_node);
fs/ext4/extents_status.c
1450
node = rb_next(&pr->rb_node);
fs/ext4/extents_status.c
1565
node = rb_next(&es->rb_node);
fs/ext4/extents_status.c
1578
node = rb_next(&es->rb_node);
fs/ext4/extents_status.c
1919
node = rb_next(&es->rb_node);
fs/ext4/extents_status.c
1984
node = rb_next(node);
fs/ext4/extents_status.c
2007
node = rb_next(node);
fs/ext4/extents_status.c
226
node = rb_next(node);
fs/ext4/extents_status.c
286
node = rb_next(&es->rb_node);
fs/ext4/extents_status.c
340
while ((node = rb_next(&es1->rb_node)) != NULL) {
fs/ext4/extents_status.c
654
node = rb_next(&es->rb_node);
fs/ext4/mballoc.c
4424
tmp = rb_next(&left_pa->pa_node.inode_node);
fs/ext4/mballoc.c
4466
iter = rb_next(iter)) {
fs/ext4/mballoc.c
5640
iter = rb_next(iter)) {
fs/ext4/mballoc.c
6429
node = rb_next(&entry->efd_node);
fs/f2fs/extent_cache.c
262
tmp_node = rb_next(parent);
fs/f2fs/extent_cache.c
280
tmp_node = rb_next(&en->rb_node);
fs/f2fs/extent_cache.c
391
next = rb_next(node);
fs/f2fs/extent_cache.c
759
struct rb_node *node = rb_next(&en->rb_node);
fs/f2fs/gc.c
437
next = rb_next(cur);
fs/f2fs/gc.c
616
node = rb_next(node);
fs/f2fs/gc.c
680
rb_next(&ve->rb_node),
fs/f2fs/segment.c
1006
next = rb_next(cur);
fs/f2fs/segment.c
1083
tmp_node = rb_next(parent);
fs/f2fs/segment.c
1098
tmp_node = rb_next(&dc->rb_node);
fs/f2fs/segment.c
1546
node = rb_next(&prev_dc->rb_node);
fs/f2fs/segment.c
1620
node = rb_next(&dc->rb_node);
fs/f2fs/segment.c
3465
node = rb_next(&dc->rb_node);
fs/fuse/dev.c
2429
p = rb_next(p);
fs/gfs2/rgrp.c
2315
for (n = rb_first(&rgd->rd_rstree); n; n = rb_next(&trs->rs_node)) {
fs/gfs2/rgrp.c
588
n = rb_next(&rgd->rd_node);
fs/jffs2/nodelist.h
347
#define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
354
#define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb)
fs/kernfs/dir.c
1389
rbn = rb_next(&pos->rb);
fs/kernfs/dir.c
1861
struct rb_node *node = rb_next(&pos->rb);
fs/kernfs/dir.c
1876
struct rb_node *node = rb_next(&pos->rb);
fs/mount.h
207
ns->mnt_first_node = rb_next(&mnt->mnt_node);
fs/namespace.c
1547
struct rb_node *node = rb_next(&mnt->mnt_node);
fs/namespace.c
5628
node = rb_next(&curr->mnt_node);
fs/nfs/blocklayout/extent_tree.c
36
struct rb_node *node = rb_next(&be->be_node);
fs/nfs/nfs4state.c
168
pos = rb_next(pos)) {
fs/nfs/nfs4state.c
1693
pos = rb_next(pos)) {
fs/nfs/nfs4state.c
1745
pos = rb_next(pos)) {
fs/nfs/nfs4state.c
1860
pos = rb_next(pos)) {
fs/ntfs3/bitmap.c
134
next = rb_next(node);
fs/ntfs3/bitmap.c
253
n = rb_next(n);
fs/ntfs3/bitmap.c
274
n = rb_next(n);
fs/ntfs3/bitmap.c
392
n3 = rb_next(n);
fs/ntfs3/bitmap.c
411
n3 = rb_next(n3);
fs/ntfs3/bitmap.c
431
n3 = rb_next(&e->count.node);
fs/ntfs3/frecord.c
113
struct rb_node *next = rb_next(node);
fs/ntfs3/frecord.c
1211
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ntfs3/frecord.c
1575
struct rb_node *next = rb_next(node);
fs/ntfs3/frecord.c
2829
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ntfs3/frecord.c
3201
next = rb_next(node);
fs/ntfs3/frecord.c
944
for (node = rb_first(&ni->mi_tree); node; node = rb_next(node)) {
fs/ocfs2/reservations.c
170
node = rb_next(node);
fs/ocfs2/reservations.c
377
node = rb_next(node);
fs/ocfs2/reservations.c
558
next = rb_next(prev);
fs/ocfs2/reservations.c
92
node = rb_next(node);
fs/proc/generic.c
64
return rb_entry_safe(rb_next(&dir->subdir_node), struct proc_dir_entry,
fs/proc/nommu.c
100
return rb_next((struct rb_node *) v);
fs/proc/nommu.c
86
for (p = rb_first(&nommu_region_tree); p; p = rb_next(p))
fs/proc/proc_sysctl.c
374
for (;node; node = rb_next(node)) {
fs/proc/proc_sysctl.c
409
ctl_node = first_usable_entry(rb_next(&ctl_node->node));
fs/smb/client/cached_dir.c
549
for (node = rb_first(root); node; node = rb_next(node)) {
fs/smb/client/connect.c
4467
node = rb_next(tmp);
fs/smb/client/misc.c
572
for (node = rb_first(root); node; node = rb_next(node)) {
fs/ubifs/debug.c
2182
this = rb_next(this);
fs/ubifs/debug.c
657
for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) {
fs/ubifs/debug.c
745
for (rb = rb_first((struct rb_root *)&c->buds); rb; rb = rb_next(rb)) {
fs/ubifs/log.c
310
p = rb_next(p);
fs/ubifs/recovery.c
1536
this = rb_next(this);
fs/xattr.c
1468
for (rbp = rb_first(&xattrs->rb_root); rbp; rbp = rb_next(rbp)) {
fs/xattr.c
1552
rbp_next = rb_next(rbp);
fs/xfs/scrub/bitmap.c
341
(bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
fs/xfs/scrub/bitmap.c
66
(bn) = rb_entry_safe(rb_next(&(bn)->bn_rbnode), \
include/linux/kvm_host.h
1126
iter->node = rb_next(iter->node);
include/linux/rbtree.h
152
leftmost = root->rb_leftmost = rb_next(node);
include/linux/rbtree.h
44
extern struct rb_node *rb_next(const struct rb_node *);
include/linux/rbtree.h
449
node = rb_next(node);
include/linux/rbtree_augmented.h
339
root->rb_leftmost = rb_next(node);
include/linux/skbuff.h
4130
#define skb_rb_next(skb) rb_to_skb(rb_next(&(skb)->rbnode))
kernel/events/core.c
1960
event = rb_entry_safe(rb_next(&event->group_node), \
kernel/events/uprobes.c
1550
for (t = n; (t = rb_next(t)); ) {
kernel/locking/ww_mutex.h
105
struct rb_node *n = rb_next(&w->tree.entry);
kernel/power/wakelock.c
46
for (node = rb_first(&wakelocks_tree); node; node = rb_next(node)) {
kernel/sched/core.c
351
node = rb_next(node);
kernel/sched/deadline.c
2706
next_node = rb_next(next_node);
kernel/trace/trace_stat.c
187
node = rb_next(node);
kernel/trace/trace_stat.c
202
return rb_next(node);
lib/rbtree.c
493
EXPORT_SYMBOL(rb_next);
lib/rbtree_test.c
198
for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) {
lib/rbtree_test.c
224
for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) {
lib/rbtree_test.c
291
for (node = rb_first(&root.rb_root); node; node = rb_next(node))
lib/timerqueue.c
79
next = rb_next(&node->node);
mm/ksm.c
3361
node = rb_next(node);
mm/mempolicy.c
3147
struct rb_node *next = rb_next(&n->nd);
mm/mempolicy.c
3281
next = rb_next(&n->nd);
mm/nommu.c
1085
for (rb = rb_first(&nommu_region_tree); rb; rb = rb_next(rb)) {
mm/nommu.c
451
while ((p = rb_next(lastp))) {
mm/shmem_quota.c
104
node = rb_next(&entry->node);
mm/shmem_quota.c
145
node = rb_next(&entry->node);
mm/swapfile.c
274
struct rb_node *rb = rb_next(&se->rb_node);
net/802/garp.c
213
next = node ? rb_next(node) : NULL, node != NULL;
net/802/garp.c
400
next = node ? rb_next(node) : NULL, node != NULL;
net/802/mrp.c
302
next = node ? rb_next(node) : NULL, node != NULL;
net/802/mrp.c
585
next = node ? rb_next(node) : NULL, node != NULL;
net/ceph/ceph_common.c
626
n = rb_next(n);
net/ceph/debugfs.c
108
for (n = rb_first(&map->primary_temp); n; n = rb_next(n)) {
net/ceph/debugfs.c
115
for (n = rb_first(&map->pg_upmap); n; n = rb_next(n)) {
net/ceph/debugfs.c
126
for (n = rb_first(&map->pg_upmap_items); n; n = rb_next(n)) {
net/ceph/debugfs.c
166
for (rp = rb_first(&monc->generic_request_tree); rp; rp = rb_next(rp)) {
net/ceph/debugfs.c
245
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/debugfs.c
271
for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) {
net/ceph/debugfs.c
332
for (n = rb_first(&osd->o_backoffs_by_id); n; n = rb_next(n)) {
net/ceph/debugfs.c
358
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
366
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
374
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/debugfs.c
74
for (n = rb_first(&map->pg_pools); n; n = rb_next(n)) {
net/ceph/debugfs.c
97
for (n = rb_first(&map->pg_temp); n; n = rb_next(n)) {
net/ceph/mon_client.c
1070
for (p = rb_first(&monc->generic_request_tree); p; p = rb_next(p)) {
net/ceph/osd_client.c
1179
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
1186
p = rb_next(p);
net/ceph/osd_client.c
1196
p = rb_next(p);
net/ceph/osd_client.c
1352
n = rb_next(n); /* unlink_request() */
net/ceph/osd_client.c
1362
n = rb_next(n); /* unlink_linger() */
net/ceph/osd_client.c
1399
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
1495
for (n = rb_first(&osdc->osdmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
3447
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
3455
p = rb_next(p); /* abort_request() */
net/ceph/osd_client.c
3469
for (p = rb_first(&osd->o_linger_requests); p; p = rb_next(p)) {
net/ceph/osd_client.c
3492
p = rb_next(p); /* abort_request() */
net/ceph/osd_client.c
3885
for (n = rb_first(&osdc->osdmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
3943
n = rb_next(n); /* recalc_linger_target() */
net/ceph/osd_client.c
3979
n = rb_next(n); /* unlink_request(), check_pool_dne() */
net/ceph/osd_client.c
4034
for (n = rb_first(&newmap->pg_pools); n; n = rb_next(n)) {
net/ceph/osd_client.c
4063
n = rb_next(n); /* close_osd() */
net/ceph/osd_client.c
4090
n = rb_next(n);
net/ceph/osd_client.c
4106
n = rb_next(n);
net/ceph/osd_client.c
4257
n = rb_next(n); /* cancel_linger_request() */
net/ceph/osd_client.c
4266
for (n = rb_first(&osd->o_linger_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
4495
for (n = rb_first(&osd->o_requests); n; n = rb_next(n)) {
net/ceph/osd_client.c
4728
for (n = rb_first(&osdc->osds); n; n = rb_next(n)) {
net/ceph/osd_client.c
4732
for (p = rb_first(&osd->o_requests); p; p = rb_next(p)) {
net/ceph/osd_client.c
5189
n = rb_next(n);
net/ceph/osdmap.c
2974
for ( ; n1 && n2; n1 = rb_next(n1), n2 = rb_next(n2)) {
net/ceph/osdmap.c
765
for (rbp = rb_first(&map->pg_pools); rbp; rbp = rb_next(rbp)) {
net/core/skbuff.c
4043
p = rb_next(p);
net/ipv4/inet_fragment.c
310
p = rb_next(p);
net/ipv4/inet_fragment.c
617
rbn = rb_next(&head->rbnode);
net/ipv4/inet_fragment.c
657
struct rb_node *rbnext = rb_next(rbn);
net/ipv4/inetpeer.c
281
p = rb_next(p);
net/ipv4/nexthop.c
3564
for (; node; node = rb_next(node)) {
net/ipv4/nexthop.c
3959
for (node = rb_first(root); node; node = rb_next(node)) {
net/ipv4/tcp.c
3375
p = rb_next(p);
net/ipv4/tcp_input.c
5347
p = rb_next(p);
net/mptcp/protocol.c
788
p = rb_next(p);
net/netfilter/nf_conncount.c
563
for (node = rb_first(root); node != NULL; node = rb_next(node)) {
net/netfilter/nf_conncount.c
581
node = rb_next(node);
net/netfilter/nft_set_rbtree.c
276
node = rb_next(node);
net/netfilter/nft_set_rbtree.c
289
return __nft_rbtree_next_active(rb_next(&rbe->node), genmask);
net/netfilter/nft_set_rbtree.c
384
next = rb_next(node);
net/netfilter/nft_set_rbtree.c
740
node = rb_next(&rbe->node);
net/netfilter/nft_set_rbtree.c
850
for (node = rb_first(&priv->root); node != NULL; node = rb_next(node)) {
net/netfilter/nft_set_rbtree.c
903
next = rb_next(node);
net/sched/sch_etf.c
425
p = rb_next(p);
net/sched/sch_fq.c
791
p = rb_next(p);
net/sched/sch_hfsc.c
231
for (n = rb_first(&q->eligible); n != NULL; n = rb_next(n)) {
net/sched/sch_hfsc.c
295
for (n = rb_first(&cl->vt_tree); n != NULL; n = rb_next(n)) {
net/sched/sch_htb.c
352
*n = rb_next(*n);
net/sched/sch_netem.c
380
p = rb_next(p);
net/smc/smc_core.c
1125
for (node = rb_first(&lgr->conns_all); node; node = rb_next(node)) {
net/tipc/group.c
269
for (n = rb_first(&grp->members); n; n = rb_next(n)) {
net/tipc/group.c
402
for (n = rb_first(&grp->members); n; n = rb_next(n)) {
net/tipc/name_table.c
1038
for (n = rb_first(&sc->ranges); n; n = rb_next(n)) {
net/tipc/name_table.c
746
for (n = rb_first(&sc->ranges); n; n = rb_next(n)) {
net/xfrm/xfrm_policy.c
1127
rn = rb_next(rn);
security/apparmor/include/label.h
79
for ((N) = rb_first(&(LS)->root); (N); (N) = rb_next(N))
security/keys/gc.c
219
cursor = rb_next(cursor);
security/keys/key.c
181
parent = rb_next(parent);
security/keys/proc.c
105
n = rb_next(&minkey->serial_node);
security/keys/proc.c
258
n = rb_next(n);
security/keys/proc.c
265
return __key_user_next(user_ns, rb_next(n));
security/keys/proc.c
68
n = rb_next(n);
security/keys/proc.c
73
n = rb_next(n);
tools/bpf/resolve_btfids/main.c
364
next = rb_next(&id->rb_node);
tools/bpf/resolve_btfids/main.c
731
next = rb_next(next);
tools/bpf/resolve_btfids/main.c
790
next = rb_next(next);
tools/bpf/resolve_btfids/main.c
963
for (next = rb_first(&obj->sets); next; next = rb_next(next)) {
tools/include/linux/rbtree.h
142
root->rb_leftmost = rb_next(node);
tools/include/linux/rbtree.h
328
node = rb_next(node);
tools/include/linux/rbtree.h
53
extern struct rb_node *rb_next(const struct rb_node *);
tools/include/linux/rbtree_augmented.h
304
root->rb_leftmost = rb_next(node);
tools/objtool/elf.c
252
n = rb_next(&hole.sym->node);
tools/perf/builtin-annotate.c
449
nd = rb_next(nd);
tools/perf/builtin-annotate.c
487
next = rb_next(nd);
tools/perf/builtin-annotate.c
521
nd = rb_next(nd);
tools/perf/builtin-annotate.c
532
next = rb_next(nd);
tools/perf/builtin-annotate.c
546
nd = rb_next(nd);
tools/perf/builtin-c2c.c
2395
next = rb_next(&he->rb_node);
tools/perf/builtin-c2c.c
2530
for (; nd; nd = rb_next(nd)) {
tools/perf/builtin-c2c.c
2641
nd = rb_next(nd);
tools/perf/builtin-diff.c
540
next = rb_next(&he->rb_node_in);
tools/perf/builtin-diff.c
594
next = rb_next(&he_pair->rb_node_in);
tools/perf/builtin-diff.c
661
next = rb_next(&he->rb_node_in);
tools/perf/builtin-diff.c
688
next = rb_next(&he->rb_node_in);
tools/perf/builtin-kmem.c
1044
next = rb_next(next);
tools/perf/builtin-kmem.c
1100
next = rb_next(next);
tools/perf/builtin-kmem.c
1142
next = rb_next(next);
tools/perf/builtin-kvm.c
1078
for (nd = rb_first_cached(&kvm_hists.hists.entries); nd; nd = rb_next(nd)) {
tools/perf/builtin-kvm.c
522
for (; nd; nd = rb_next(nd)) {
tools/perf/builtin-kvm.c
652
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/builtin-kwork.c
1934
next = rb_next(next);
tools/perf/builtin-kwork.c
2014
next = rb_next(next);
tools/perf/builtin-kwork.c
2096
next = rb_next(next);
tools/perf/builtin-kwork.c
2170
next = rb_next(next);
tools/perf/builtin-kwork.c
560
next = rb_next(next);
tools/perf/builtin-lock.c
1351
node = rb_next(node);
tools/perf/builtin-sched.c
3062
rb_node = rb_next(rb_node);
tools/perf/builtin-sched.c
3531
next = rb_next(next);
tools/perf/builtin-top.c
423
next = rb_next(&n->rb_node);
tools/perf/tests/hists_common.c
191
node = rb_next(node);
tools/perf/tests/hists_common.c
221
node = rb_next(node);
tools/perf/tests/hists_cumulate.c
218
node = rb_next(node), i++) {
tools/perf/tests/hists_link.c
191
node = rb_next(node);
tools/perf/tests/hists_link.c
248
node = rb_next(node);
tools/perf/tests/hists_output.c
185
node = rb_next(node);
tools/perf/tests/hists_output.c
191
node = rb_next(node);
tools/perf/tests/hists_output.c
197
node = rb_next(node);
tools/perf/tests/hists_output.c
203
node = rb_next(node);
tools/perf/tests/hists_output.c
209
node = rb_next(node);
tools/perf/tests/hists_output.c
215
node = rb_next(node);
tools/perf/tests/hists_output.c
221
node = rb_next(node);
tools/perf/tests/hists_output.c
227
node = rb_next(node);
tools/perf/tests/hists_output.c
284
node = rb_next(node);
tools/perf/tests/hists_output.c
339
node = rb_next(node);
tools/perf/tests/hists_output.c
345
node = rb_next(node);
tools/perf/tests/hists_output.c
351
node = rb_next(node);
tools/perf/tests/hists_output.c
357
node = rb_next(node);
tools/perf/tests/hists_output.c
417
node = rb_next(node);
tools/perf/tests/hists_output.c
423
node = rb_next(node);
tools/perf/tests/hists_output.c
429
node = rb_next(node);
tools/perf/tests/hists_output.c
435
node = rb_next(node);
tools/perf/tests/hists_output.c
441
node = rb_next(node);
tools/perf/tests/hists_output.c
447
node = rb_next(node);
tools/perf/tests/hists_output.c
453
node = rb_next(node);
tools/perf/tests/hists_output.c
459
node = rb_next(node);
tools/perf/tests/hists_output.c
522
node = rb_next(node);
tools/perf/tests/hists_output.c
529
node = rb_next(node);
tools/perf/tests/hists_output.c
536
node = rb_next(node);
tools/perf/tests/hists_output.c
543
node = rb_next(node);
tools/perf/tests/hists_output.c
550
node = rb_next(node);
tools/perf/tests/hists_output.c
557
node = rb_next(node);
tools/perf/tests/hists_output.c
564
node = rb_next(node);
tools/perf/tests/hists_output.c
571
node = rb_next(node);
tools/perf/tests/hists_output.c
578
node = rb_next(node);
tools/perf/tests/symbols.c
125
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
tools/perf/ui/browser.c
159
nd = rb_next(nd);
tools/perf/ui/browser.c
183
nd = rb_next(nd);
tools/perf/ui/browsers/annotate.c
949
nd = rb_next(nd);
tools/perf/ui/browsers/hists.c
1030
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/hists.c
1119
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/hists.c
182
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
1922
if (rb_next(nd))
tools/perf/ui/browsers/hists.c
1923
nd = rb_next(nd);
tools/perf/ui/browsers/hists.c
265
for (nd = rb_first(chain); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
300
node = rb_next(node);
tools/perf/ui/browsers/hists.c
333
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
371
bool has_sibling = nd && rb_next(nd);
tools/perf/ui/browsers/hists.c
373
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
516
for (nd = rb_first(&node->rb_root); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
557
for (nd = rb_first(chain); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
573
for (nd = rb_first_cached(&he->hroot_out); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
902
if (rb_next(node))
tools/perf/ui/browsers/hists.c
926
struct rb_node *next = rb_next(node);
tools/perf/ui/browsers/map.c
122
for (nd = rb_first(mb.b.entries); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
103
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
165
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
226
for (nd = rb_first(root); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
358
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
420
for (node = rb_first_cached(root); node; node = rb_next(node)) {
tools/perf/ui/stdio/hist.c
146
next = rb_next(node);
tools/perf/ui/stdio/hist.c
214
if (rb_next(node))
tools/perf/ui/stdio/hist.c
328
rb_node = rb_next(rb_node);
tools/perf/ui/stdio/hist.c
379
rb_node = rb_next(rb_node);
tools/perf/util/annotate.c
1141
next = rb_next(node);
tools/perf/util/annotate.c
1183
node = rb_next(node);
tools/perf/util/block-info.c
457
next = rb_next(&he->rb_node);
tools/perf/util/block-range.c
109
n = rb_next(n);
tools/perf/util/block-range.c
18
for (rb = rb_first(&block_ranges.root); rb; rb = rb_next(rb)) {
tools/perf/util/block-range.h
40
struct rb_node *n = rb_next(&br->node);
tools/perf/util/build-id.c
375
nd = rb_next(nd)) {
tools/perf/util/build-id.c
923
nd = rb_next(nd)) {
tools/perf/util/build-id.c
959
nd = rb_next(nd)) {
tools/perf/util/callchain.c
1066
n = rb_next(n);
tools/perf/util/callchain.c
1310
n = rb_next(n);
tools/perf/util/callchain.c
1518
n = rb_next(n);
tools/perf/util/callchain.c
1545
n = rb_next(n);
tools/perf/util/callchain.c
1586
chain->has_children = rb_prev(&node->rb_node) || rb_next(&node->rb_node);
tools/perf/util/callchain.c
1774
rb_node = rb_next(rb_node);
tools/perf/util/callchain.c
1790
next = rb_next(&he->rb_node);
tools/perf/util/callchain.c
440
n = rb_next(n);
tools/perf/util/callchain.c
472
n = rb_next(n);
tools/perf/util/callchain.c
502
n = rb_next(n);
tools/perf/util/callchain.c
568
n = rb_next(n);
tools/perf/util/dso.c
1812
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
tools/perf/util/dso.c
905
next = rb_next(&cache->rb_node);
tools/perf/util/env.c
101
for (struct rb_node *node = first; node != NULL; node = rb_next(node))
tools/perf/util/env.c
188
next = rb_next(&node->rb_node);
tools/perf/util/env.c
204
next = rb_next(&node->rb_node);
tools/perf/util/header.c
1052
next = rb_next(&node->rb_node);
tools/perf/util/header.c
1094
next = rb_next(&node->rb_node);
tools/perf/util/header.c
2002
next = rb_next(&node->rb_node);
tools/perf/util/header.c
2029
next = rb_next(&node->rb_node);
tools/perf/util/hist.c
1846
next = rb_next(&n->rb_node_in);
tools/perf/util/hist.c
1919
node = rb_next(node);
tools/perf/util/hist.c
1976
node = rb_next(node);
tools/perf/util/hist.c
2100
next = rb_next(&n->rb_node_in);
tools/perf/util/hist.c
2176
node = rb_next(node);
tools/perf/util/hist.c
2183
node = rb_next(&he->rb_node);
tools/perf/util/hist.c
2216
node = rb_next(node);
tools/perf/util/hist.c
2340
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2382
nd = rb_next(nd);
tools/perf/util/hist.c
2449
nd = rb_next(nd);
tools/perf/util/hist.c
2684
for (nd = rb_first_cached(leader_root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2715
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2732
for (nd = rb_first_cached(other_root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
275
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
2791
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2816
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
435
node = rb_next(node);
tools/perf/util/hist.c
478
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
494
next = rb_next(&n->rb_node);
tools/perf/util/hist.c
511
next = rb_next(&n->rb_node);
tools/perf/util/intel-pt.c
371
n = rb_next(n);
tools/perf/util/intlist.h
56
rn = rb_next(&in->rb_node);
tools/perf/util/machine.c
1268
next = rb_next(&pos->rb_node);
tools/perf/util/machine.c
285
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
3069
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
453
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
467
node = rb_next(node)) {
tools/perf/util/machine.c
907
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
927
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/print-events.c
312
next = rb_next(node);
tools/perf/util/python.c
1316
node = rb_next(node)) {
tools/perf/util/python.c
1423
node = rb_next(node)) {
tools/perf/util/python.c
2042
node = rb_next(node)) {
tools/perf/util/rblist.c
116
next = rb_next(pos);
tools/perf/util/rblist.c
134
node = rb_next(node)) {
tools/perf/util/session.c
2735
for (struct rb_node *nd = rb_first_cached(&session->machines.guests); nd; nd = rb_next(nd)) {
tools/perf/util/session.c
2951
nd = rb_next(nd)) {
tools/perf/util/srcline.c
419
next = rb_next(&pos->rb_node);
tools/perf/util/srcline.c
503
next = rb_next(&pos->rb_node);
tools/perf/util/stream.c
108
rb_node = rb_next(rb_node);
tools/perf/util/stream.c
121
next = rb_next(&he->rb_node);
tools/perf/util/strlist.h
66
rn = rb_next(&sn->rb_node);
tools/perf/util/symbol.c
1064
next = rb_next(&mi->rb_node);
tools/perf/util/symbol.c
1156
from_node = rb_next(from_node);
tools/perf/util/symbol.c
1157
to_node = rb_next(to_node);
tools/perf/util/symbol.c
220
nd = rb_next(&curr->rb_node);
tools/perf/util/symbol.c
237
nd = rb_next(&curr->rb_node);
tools/perf/util/symbol.c
255
for (nd = rb_next(prevnd); nd; nd = rb_next(nd)) {
tools/perf/util/symbol.c
354
next = rb_next(&pos->rb_node);
tools/perf/util/symbol.c
444
struct rb_node *n = rb_next(&sym->rb_node);
tools/perf/util/symbol.c
466
for (nd = rb_first_cached(source); nd; nd = rb_next(nd))
tools/perf/util/symbol.c
473
for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) {
tools/perf/util/symbol.c
822
next = rb_next(&pos->rb_node);
tools/perf/util/symbol.c
881
next = rb_next(&pos->rb_node);
tools/perf/util/symbol.h
87
nd = rb_next(nd))