rcu_dereference_all
return __rht_ptr(rcu_dereference_all(*bkt), bkt, freq);
pos = rcu_dereference_all(pos->next))
pos = rcu_dereference_all(pos->next))
for (pos = list; pos; pos = rcu_dereference_all(pos->next))
pos = rcu_dereference_all(pos->next))
sch = rcu_dereference_all(scx_root);
sd_share = rcu_dereference_all(per_cpu(sd_llc_shared, env->dst_cpu));
rcu_dereference_all(env->dst_rq->rd->pd))
sd = rcu_dereference_all(rq->sd);
sd = rcu_dereference_all(per_cpu(sd_asym_packing, cpu));
sd = rcu_dereference_all(per_cpu(sd_asym_cpucapacity, cpu));
sds = rcu_dereference_all(per_cpu(sd_llc_shared, cpu));
sd = rcu_dereference_all(per_cpu(sd_llc, cpu));
sd = rcu_dereference_all(per_cpu(sd_llc, cpu));
ng = rcu_dereference_all(p->numa_group);
ng = rcu_dereference_all(p->numa_group);
ng = rcu_dereference_all(p->numa_group);
cur = rcu_dereference_all(dst_rq->curr);
cur_ng = rcu_dereference_all(cur->numa_group);
sd = rcu_dereference_all(per_cpu(sd_numa, env.src_cpu));
grp = rcu_dereference_all(tsk->numa_group);
is_idle = is_idle_task(rcu_dereference_all(rq->curr));
sds = rcu_dereference_all(per_cpu(sd_llc_shared, cpu));
sds = rcu_dereference_all(per_cpu(sd_llc_shared, cpu));
sd_share = rcu_dereference_all(per_cpu(sd_llc_shared, target));
sd = rcu_dereference_all(per_cpu(sd_asym_cpucapacity, target));
sd = rcu_dereference_all(per_cpu(sd_llc, target));
pd = rcu_dereference_all(rd->pd);
sd = rcu_dereference_all(*this_cpu_ptr(&sd_asym_cpucapacity));
struct numa_group *numa_group = rcu_dereference_all(p->numa_group);