rb_first_cached
rb_first_cached(&bfqd->queue_weights_tree),
n = rb_first_cached(&parent_sq->pending_tree);
if (rb_to_hole_size_or_zero(rb_first_cached(&mm->holes_size)) < size)
struct rb_node *rb = rb_first_cached(&el->virtual);
rb = rb_first_cached(&el->virtual);
while ((rb = rb_first_cached(&sched_engine->queue))) {
rb = rb_first_cached(&sched_engine->queue);
struct rb_node *rb = rb_first_cached(&el->virtual);
while ((rb = rb_first_cached(&sched_engine->queue))) {
while ((rb = rb_first_cached(&execlists->virtual))) {
first = rb_first_cached(&sibling->execlists.virtual) ==
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) {
while ((rb = rb_first_cached(&sched_engine->queue))) {
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
while ((rb = rb_first_cached(&sched_engine->queue))) {
GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(&rq->rb_tree_root); rb; rb = rb_next(rb)) {
while ((node = rb_first_cached(&handler->root))) {
node = rb_first_cached(&preftrees.direct.root);
while ((rnode = rb_first_cached(&preftrees->indirect.root))) {
while ((node = rb_first_cached(&tree->root))) {
for (n = rb_first_cached(&head->ref_tree); n; n = rb_next(n)) {
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
node_bg = rb_first_cached(&fs_info->block_group_cache_tree);
node_chunk = rb_first_cached(&fs_info->mapping_tree);
struct rb_node *p = rb_first_cached(&delayed_node->ins_root);
struct rb_node *p = rb_first_cached(&delayed_node->del_root);
while ((n = rb_first_cached(&head->ref_tree)) != NULL) {
for (node = rb_first_cached(&head->ref_tree); node;
ref = rb_entry(rb_first_cached(&head->ref_tree),
node = rb_first_cached(&fs_info->mapping_tree);
for (node = rb_first_cached(&head->ref_tree); node;
leftmost = rb_first_cached(&fs_info->block_group_cache_tree);
node = rb_first_cached(&ctl->free_space_bytes);
node = rb_first_cached(&fs_info->block_group_cache_tree);
node = rb_first_cached(&fs_info->block_group_cache_tree);
node = rb_first_cached(&trans->fs_info->block_group_cache_tree);
node = rb_first_cached(&fs_info->block_group_cache_tree);
for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node;
for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node;
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
entry = rb_entry(rb_first_cached(&ctl->free_space_bytes),
node = rb_first_cached(&fs_info->mapping_tree);
for (node = rb_first_cached(&fs_info->mapping_tree); node; node = rb_next(node)) {
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) {
for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = next) {
node = rb_first_cached(&et->root);
struct rb_node *cur = rb_first_cached(root), *next;
node = rb_first_cached(root);
struct rb_node *cur = rb_first_cached(&dcc->root), *next;
struct rb_node *leftmost = rb_first_cached(&head->rb_root);
return (struct bpf_rb_node *)rb_first_cached(r);
struct rb_node *leftmost = rb_first_cached(&lock->waiters);
struct rb_node *leftmost = rb_first_cached(&lock->waiters);
struct rb_node *leftmost = rb_first_cached(&dl_rq->root);
struct rb_node *left = rb_first_cached(&dl_rq->root);
next_node = rb_first_cached(&rq->dl.pushable_dl_tasks_root);
p = __node_2_pdl(rb_first_cached(&rq->dl.pushable_dl_tasks_root));
return rb_first_cached(&dl_rq->root) == &dl_se->rb_node;
struct rb_node *left = rb_first_cached(&cfs_rq->tasks_timeline);
node = rb_first_cached(&root);
p = rb_first_cached(&q->head);
struct rb_node *p = rb_first_cached(&q->head);
n = rb_first_cached(&sec->symbol_tree);
struct rb_node *nd = rb_first_cached(&hists->entries), *next;
struct rb_node *next = rb_first_cached(&hists->entries);
nd = rb_first_cached(&c2c.hists.hists.entries);
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
next = rb_first_cached(root);
struct rb_node *next = rb_first_cached(root);
struct rb_node *next = rb_first_cached(root);
next = rb_first_cached(root);
for (nd = rb_first_cached(&kvm_hists.hists.entries); nd; nd = rb_next(nd)) {
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
node = rb_first_cached(root);
next = rb_first_cached(&kwork->sorted_work_root);
next = rb_first_cached(&class->work_root);
next = rb_first_cached(&class->work_root);
node = rb_first_cached(&class->work_root);
next = rb_first_cached(&kwork->sorted_work_root);
next = rb_first_cached(root);
node = rb_first_cached(root);
struct rb_node *rb_node = rb_first_cached(root);
while ((node = rb_first_cached(&sched->atom_root))) {
next = rb_first_cached(&sched->sorted_atom_root);
while ((next = rb_first_cached(&sched->sorted_atom_root))) {
next = rb_first_cached(&hists->entries);
node = rb_first_cached(root);
node = rb_first_cached(root);
node = rb_first_cached(root_out);
node = rb_first_cached(root);
node = rb_first_cached(root);
node = rb_first_cached(root_out);
node = rb_first_cached(root);
node = rb_first_cached(root);
node = rb_first_cached(root);
node = rb_first_cached(root);
node = rb_first_cached(root);
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
browser->top = rb_first_cached(&hb->hists->entries);
node = rb_first_cached(&he->hroot_out);
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
struct rb_node *nd = rb_first_cached(&hb->hists->entries);
for (nd = rb_first_cached(&he->hroot_out); nd; nd = rb_next(nd)) {
nd = rb_first_cached(&browser->hists->entries);
for (nd = rb_first_cached(&hists->entries);
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
for (node = rb_first_cached(root); node; node = rb_next(node)) {
for (nd = rb_first_cached(&hists->entries); nd;
struct rb_node *next = rb_first_cached(&hists->entries);
for (nd = rb_first_cached(&session->machines.guests); nd;
for (nd = rb_first_cached(&machines->guests); nd;
for (nd = rb_first_cached(&session->machines.guests); nd;
struct rb_node *next = rb_first_cached(&hists->entries);
for (nd = rb_first_cached(dso__symbols(dso)); nd; nd = rb_next(nd)) {
next = rb_first_cached(root);
node = rb_first_cached(&hists->entries);
node = rb_first_cached(root_in);
next = rb_first_cached(root);
node = rb_first_cached(&he->hroot_out);
node = rb_first_cached(&he->hroot_out);
for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) {
nd = rb_first_cached(&he->hroot_out);
nd = rb_first_cached(&hists->entries);
nd = rb_first_cached(&hists->entries);
struct rb_node *next = rb_first_cached(&hists->entries);
for (nd = rb_first_cached(leader_root); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(other_root); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(root); nd; nd = rb_next(nd)) {
node = rb_first_cached(root);
struct rb_node *node = rb_first_cached(&he->hroot_out);
struct rb_node *next = rb_first_cached(&hists->entries);
struct rb_node *next = rb_first_cached(&hists->entries);
struct rb_node *next = rb_first_cached(&hists->entries);
struct rb_node *rn = rb_first_cached(&ilist->rblist.entries);
struct rb_node *next = rb_first_cached(&machines->guests);
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
for (node = rb_first_cached(&machines->guests); node;
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) {
struct rb_node *nd = rb_first_cached(symbols);
for (node = rb_first_cached(&groups.entries); node; node = next) {
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries); node;
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries);
for (node = rb_first_cached(&pevlist->evlist.metric_events.entries); node;
struct rb_node *pos, *next = rb_first_cached(&rblist->entries);
for (node = rb_first_cached(&rblist->entries); node;
for (struct rb_node *nd = rb_first_cached(&session->machines.guests); nd; nd = rb_next(nd)) {
for (nd = rb_first_cached(&session->machines.guests); nd;
struct rb_node *next = rb_first_cached(tree);
struct rb_node *next = rb_first_cached(tree);
struct rb_node *next = rb_first_cached(&hists->entries);
struct rb_node *rn = rb_first_cached(&slist->rblist.entries);
nd = rb_first_cached(symbols);
struct rb_node *nd, *prevnd = rb_first_cached(symbols);
struct rb_node *next = rb_first_cached(symbols);
struct rb_node *n = rb_first_cached(symbols);
for (nd = rb_first_cached(source); nd; nd = rb_next(nd))
for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) {
struct rb_node *next = rb_first_cached(root);
struct rb_node *next = rb_first_cached(root);
for (nd = rb_first_cached(symbols); \