rb_first_cached
for (rb = rb_first_cached(root); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(root); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(root); rb; rb = rb_next(rb)) {
if (rb_to_hole_size_or_zero(rb_first_cached(&mm->holes_size)) < size)
struct rb_node *rb = rb_first_cached(&el->virtual);
rb = rb_first_cached(&el->virtual);
while ((rb = rb_first_cached(&sched_engine->queue))) {
rb = rb_first_cached(&sched_engine->queue);
struct rb_node *rb = rb_first_cached(&el->virtual);
while ((rb = rb_first_cached(&sched_engine->queue))) {
while ((rb = rb_first_cached(&execlists->virtual))) {
first = rb_first_cached(&sibling->execlists.virtual) ==
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) {
while ((rb = rb_first_cached(&sched_engine->queue))) {
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
while ((rb = rb_first_cached(&sched_engine->queue))) {
GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(root); rb; rb = rb_next(rb)) {
for (rb = rb_first_cached(&rq->rb_tree_root); rb; rb = rb_next(rb)) {