Symbol: rb_first_cached
block/bfq-iosched.c
863
rb_first_cached(&bfqd->queue_weights_tree),
block/blk-throttle.c
371
n = rb_first_cached(&parent_sq->pending_tree);
drivers/gpu/drm/drm_mm.c
531
if (rb_to_hole_size_or_zero(rb_first_cached(&mm->holes_size)) < size)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1025
struct rb_node *rb = rb_first_cached(&el->virtual);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1036
rb = rb_first_cached(&el->virtual);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1488
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
289
rb = rb_first_cached(&sched_engine->queue);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
298
struct rb_node *rb = rb_first_cached(&el->virtual);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3179
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3198
while ((rb = rb_first_cached(&execlists->virtual))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3848
first = rb_first_cached(&sibling->execlists.virtual) ==
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4113
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4134
for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2002
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5511
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
953
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/i915_scheduler.c
47
GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
drivers/gpu/drm/i915/i915_scheduler.c
51
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/scheduler/sched_main.c
319
for (rb = rb_first_cached(&rq->rb_tree_root); rb; rb = rb_next(rb)) {
drivers/infiniband/hw/hfi1/mmu_rb.c
100
while ((node = rb_first_cached(&handler->root))) {
fs/btrfs/backref.c
1589
node = rb_first_cached(&preftrees.direct.root);
fs/btrfs/backref.c
751
while ((rnode = rb_first_cached(&preftrees->indirect.root))) {
fs/btrfs/backref.c
843
while ((node = rb_first_cached(&tree->root))) {
fs/btrfs/backref.c
897
for (n = rb_first_cached(&head->ref_tree); n; n = rb_next(n)) {
fs/btrfs/block-group.c
2555
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
fs/btrfs/block-group.c
4901
node_bg = rb_first_cached(&fs_info->block_group_cache_tree);
fs/btrfs/block-group.c
4902
node_chunk = rb_first_cached(&fs_info->mapping_tree);
fs/btrfs/delayed-inode.c
477
struct rb_node *p = rb_first_cached(&delayed_node->ins_root);
fs/btrfs/delayed-inode.c
485
struct rb_node *p = rb_first_cached(&delayed_node->del_root);
fs/btrfs/delayed-ref.c
1292
while ((n = rb_first_cached(&head->ref_tree)) != NULL) {
fs/btrfs/delayed-ref.c
450
for (node = rb_first_cached(&head->ref_tree); node;
fs/btrfs/delayed-ref.c
569
ref = rb_entry(rb_first_cached(&head->ref_tree),
fs/btrfs/dev-replace.c
825
node = rb_first_cached(&fs_info->mapping_tree);
fs/btrfs/extent-tree.c
2347
for (node = rb_first_cached(&head->ref_tree); node;
fs/btrfs/extent-tree.c
2671
leftmost = rb_first_cached(&fs_info->block_group_cache_tree);
fs/btrfs/free-space-cache.c
2020
node = rb_first_cached(&ctl->free_space_bytes);
fs/btrfs/free-space-cache.c
4162
node = rb_first_cached(&fs_info->block_group_cache_tree);
fs/btrfs/free-space-tree.c
1191
node = rb_first_cached(&fs_info->block_group_cache_tree);
fs/btrfs/free-space-tree.c
1256
node = rb_first_cached(&trans->fs_info->block_group_cache_tree);
fs/btrfs/free-space-tree.c
1349
node = rb_first_cached(&fs_info->block_group_cache_tree);
fs/btrfs/tests/free-space-tests.c
859
for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node;
fs/btrfs/tests/free-space-tests.c
882
for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node;
fs/btrfs/tests/free-space-tests.c
939
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/tests/free-space-tests.c
964
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/tests/free-space-tests.c
979
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
fs/btrfs/volumes.c
6134
node = rb_first_cached(&fs_info->mapping_tree);
fs/btrfs/volumes.c
8331
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
fs/eventpoll.c
1078
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
1204
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
2078
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
913
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
fs/eventpoll.c
928
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = next) {
fs/f2fs/extent_cache.c
388
node = rb_first_cached(&et->root);
fs/f2fs/gc.c
433
struct rb_node *cur = rb_first_cached(root), *next;
fs/f2fs/gc.c
583
node = rb_first_cached(root);
fs/f2fs/segment.c
1002
struct rb_node *cur = rb_first_cached(&dcc->root), *next;
include/linux/timerqueue.h
25
struct rb_node *leftmost = rb_first_cached(&head->rb_root);
kernel/bpf/helpers.c
2552
return (struct bpf_rb_node *)rb_first_cached(r);
kernel/locking/rtmutex_common.h
124
struct rb_node *leftmost = rb_first_cached(&lock->waiters);
kernel/locking/rtmutex_common.h
131
struct rb_node *leftmost = rb_first_cached(&lock->waiters);
kernel/sched/deadline.c
2055
struct rb_node *leftmost = rb_first_cached(&dl_rq->root);
kernel/sched/deadline.c
2590
struct rb_node *left = rb_first_cached(&dl_rq->root);
kernel/sched/deadline.c
2699
next_node = rb_first_cached(&rq->dl.pushable_dl_tasks_root);
kernel/sched/deadline.c
2809
p = __node_2_pdl(rb_first_cached(&rq->dl.pushable_dl_tasks_root));
kernel/sched/deadline.c
501
return rb_first_cached(&dl_rq->root) == &dl_se->rb_node;
kernel/sched/fair.c
942
struct rb_node *left = rb_first_cached(&cfs_rq->tasks_timeline);
lib/rbtree_test.c
317
node = rb_first_cached(&root);
net/sched/sch_etf.c
113
p = rb_first_cached(&q->head);
net/sched/sch_etf.c
420
struct rb_node *p = rb_first_cached(&q->head);
tools/objtool/elf.c
254
n = rb_first_cached(&sec->symbol_tree);
tools/perf/builtin-annotate.c
413
struct rb_node *nd = rb_first_cached(&hists->entries), *next;
tools/perf/builtin-c2c.c
2385
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/builtin-c2c.c
2528
nd = rb_first_cached(&c2c.hists.hists.entries);
tools/perf/builtin-c2c.c
2633
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
tools/perf/builtin-diff.c
536
next = rb_first_cached(root);
tools/perf/builtin-diff.c
587
struct rb_node *next = rb_first_cached(root);
tools/perf/builtin-diff.c
654
struct rb_node *next = rb_first_cached(root);
tools/perf/builtin-diff.c
680
next = rb_first_cached(root);
tools/perf/builtin-kvm.c
1078
for (nd = rb_first_cached(&kvm_hists.hists.entries); nd; nd = rb_next(nd)) {
tools/perf/builtin-kvm.c
519
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
tools/perf/builtin-kvm.c
652
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/builtin-kwork.c
1693
node = rb_first_cached(root);
tools/perf/builtin-kwork.c
1922
next = rb_first_cached(&kwork->sorted_work_root);
tools/perf/builtin-kwork.c
2008
next = rb_first_cached(&class->work_root);
tools/perf/builtin-kwork.c
2080
next = rb_first_cached(&class->work_root);
tools/perf/builtin-kwork.c
2124
node = rb_first_cached(&class->work_root);
tools/perf/builtin-kwork.c
2159
next = rb_first_cached(&kwork->sorted_work_root);
tools/perf/builtin-kwork.c
553
next = rb_first_cached(root);
tools/perf/builtin-sched.c
1507
node = rb_first_cached(root);
tools/perf/builtin-sched.c
3054
struct rb_node *rb_node = rb_first_cached(root);
tools/perf/builtin-sched.c
3471
while ((node = rb_first_cached(&sched->atom_root))) {
tools/perf/builtin-sched.c
3524
next = rb_first_cached(&sched->sorted_atom_root);
tools/perf/builtin-sched.c
3545
while ((next = rb_first_cached(&sched->sorted_atom_root))) {
tools/perf/builtin-top.c
416
next = rb_first_cached(&hists->entries);
tools/perf/tests/hists_common.c
175
node = rb_first_cached(root);
tools/perf/tests/hists_common.c
204
node = rb_first_cached(root);
tools/perf/tests/hists_cumulate.c
145
node = rb_first_cached(root_out);
tools/perf/tests/hists_link.c
174
node = rb_first_cached(root);
tools/perf/tests/hists_link.c
226
node = rb_first_cached(root);
tools/perf/tests/hists_output.c
110
node = rb_first_cached(root_out);
tools/perf/tests/hists_output.c
179
node = rb_first_cached(root);
tools/perf/tests/hists_output.c
279
node = rb_first_cached(root);
tools/perf/tests/hists_output.c
333
node = rb_first_cached(root);
tools/perf/tests/hists_output.c
411
node = rb_first_cached(root);
tools/perf/tests/hists_output.c
514
node = rb_first_cached(root);
tools/perf/tests/symbols.c
125
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
1852
browser->top = rb_first_cached(&hb->hists->entries);
tools/perf/ui/browsers/hists.c
286
node = rb_first_cached(&he->hroot_out);
tools/perf/ui/browsers/hists.c
2941
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
tools/perf/ui/browsers/hists.c
2961
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
tools/perf/ui/browsers/hists.c
573
for (nd = rb_first_cached(&he->hroot_out); nd; nd = rb_next(nd)) {
tools/perf/ui/browsers/hists.c
609
nd = rb_first_cached(&browser->hists->entries);
tools/perf/ui/browsers/hists.c
68
for (nd = rb_first_cached(&hists->entries);
tools/perf/ui/gtk/hists.c
358
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
tools/perf/ui/gtk/hists.c
420
for (node = rb_first_cached(root); node; node = rb_next(node)) {
tools/perf/ui/stdio/hist.c
869
for (nd = rb_first_cached(&hists->entries); nd;
tools/perf/util/block-info.c
443
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/build-id.c
374
for (nd = rb_first_cached(&session->machines.guests); nd;
tools/perf/util/build-id.c
922
for (nd = rb_first_cached(&machines->guests); nd;
tools/perf/util/build-id.c
958
for (nd = rb_first_cached(&session->machines.guests); nd;
tools/perf/util/callchain.c
1782
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/dso.c
1812
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
1840
next = rb_first_cached(root);
tools/perf/util/hist.c
1905
node = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
1972
node = rb_first_cached(root_in);
tools/perf/util/hist.c
2095
next = rb_first_cached(root);
tools/perf/util/hist.c
2174
node = rb_first_cached(&he->hroot_out);
tools/perf/util/hist.c
2212
node = rb_first_cached(&he->hroot_out);
tools/perf/util/hist.c
2340
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2378
nd = rb_first_cached(&he->hroot_out);
tools/perf/util/hist.c
2401
nd = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
2445
nd = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
265
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
2684
for (nd = rb_first_cached(leader_root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2715
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2732
for (nd = rb_first_cached(other_root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2791
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
2816
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
tools/perf/util/hist.c
3032
node = rb_first_cached(root);
tools/perf/util/hist.c
432
struct rb_node *node = rb_first_cached(&he->hroot_out);
tools/perf/util/hist.c
473
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
489
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/hist.c
502
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/intlist.h
48
struct rb_node *rn = rb_first_cached(&ilist->rblist.entries);
tools/perf/util/machine.c
1261
struct rb_node *next = rb_first_cached(&machines->guests);
tools/perf/util/machine.c
285
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
3069
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
453
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
466
for (node = rb_first_cached(&machines->guests); node;
tools/perf/util/machine.c
907
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/machine.c
927
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
tools/perf/util/map.c
319
struct rb_node *nd = rb_first_cached(symbols);
tools/perf/util/print-events.c
300
for (node = rb_first_cached(&groups.entries); node; node = next) {
tools/perf/util/python.c
1315
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries); node;
tools/perf/util/python.c
1421
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries);
tools/perf/util/python.c
2041
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries); node;
tools/perf/util/rblist.c
112
struct rb_node *pos, *next = rb_first_cached(&rblist->entries);
tools/perf/util/rblist.c
133
for (node = rb_first_cached(&rblist->entries); node;
tools/perf/util/session.c
2735
for (struct rb_node *nd = rb_first_cached(&session->machines.guests); nd; nd = rb_next(nd)) {
tools/perf/util/session.c
2950
for (nd = rb_first_cached(&session->machines.guests); nd;
tools/perf/util/srcline.c
415
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/srcline.c
499
struct rb_node *next = rb_first_cached(tree);
tools/perf/util/stream.c
114
struct rb_node *next = rb_first_cached(&hists->entries);
tools/perf/util/strlist.h
58
struct rb_node *rn = rb_first_cached(&slist->rblist.entries);
tools/perf/util/symbol.c
215
nd = rb_first_cached(symbols);
tools/perf/util/symbol.c
247
struct rb_node *nd, *prevnd = rb_first_cached(symbols);
tools/perf/util/symbol.c
350
struct rb_node *next = rb_first_cached(symbols);
tools/perf/util/symbol.c
424
struct rb_node *n = rb_first_cached(symbols);
tools/perf/util/symbol.c
466
for (nd = rb_first_cached(source); nd; nd = rb_next(nd))
tools/perf/util/symbol.c
473
for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) {
tools/perf/util/symbol.c
809
struct rb_node *next = rb_first_cached(root);
tools/perf/util/symbol.c
866
struct rb_node *next = rb_first_cached(root);
tools/perf/util/symbol.h
85
for (nd = rb_first_cached(symbols); \