Symbol: rcu_data
kernel/rcu/tree.c
1016
static bool rcu_start_this_gp(struct rcu_node *rnp_start, struct rcu_data *rdp,
kernel/rcu/tree.c
1094
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
1143
static bool rcu_accelerate_cbs(struct rcu_node *rnp, struct rcu_data *rdp)
kernel/rcu/tree.c
1190
struct rcu_data *rdp)
kernel/rcu/tree.c
1219
static bool rcu_advance_cbs(struct rcu_node *rnp, struct rcu_data *rdp)
kernel/rcu/tree.c
1243
struct rcu_data *rdp)
kernel/rcu/tree.c
1273
static bool __note_gp_changes(struct rcu_node *rnp, struct rcu_data *rdp)
kernel/rcu/tree.c
1322
static void note_gp_changes(struct rcu_data *rdp)
kernel/rcu/tree.c
162
static void rcu_report_exp_rdp(struct rcu_data *rdp);
kernel/rcu/tree.c
163
static void rcu_report_qs_rdp(struct rcu_data *rdp);
kernel/rcu/tree.c
164
static void check_cb_ovld_locked(struct rcu_data *rdp, struct rcu_node *rnp);
kernel/rcu/tree.c
165
static bool rcu_rdp_is_offloaded(struct rcu_data *rdp);
kernel/rcu/tree.c
166
static bool rcu_rdp_cpu_online(struct rcu_data *rdp);
kernel/rcu/tree.c
1809
struct rcu_data *rdp;
kernel/rcu/tree.c
1957
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
1995
rcu_report_qs_rdp(this_cpu_ptr(&rcu_data));
kernel/rcu/tree.c
2157
struct rcu_data *rdp;
kernel/rcu/tree.c
2198
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
2206
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
2225
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
231
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
2443
rcu_report_qs_rdp(struct rcu_data *rdp)
kernel/rcu/tree.c
2498
rcu_check_quiescent_state(struct rcu_data *rdp)
kernel/rcu/tree.c
2540
static void rcu_do_batch(struct rcu_data *rdp)
kernel/rcu/tree.c
2702
WARN_ON_ONCE(time_before(j, __this_cpu_read(rcu_data.last_sched_clock)));
kernel/rcu/tree.c
2703
__this_cpu_write(rcu_data.last_sched_clock, j);
kernel/rcu/tree.c
2707
raw_cpu_inc(rcu_data.ticks_this_gp);
kernel/rcu/tree.c
2709
if (smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) {
kernel/rcu/tree.c
2713
__this_cpu_write(rcu_data.rcu_urgent_qs, false);
kernel/rcu/tree.c
2732
static void force_qs_rnp(int (*f)(struct rcu_data *rdp))
kernel/rcu/tree.c
2762
struct rcu_data *rdp;
kernel/rcu/tree.c
2765
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
2801
rnp = raw_cpu_read(rcu_data.mynode);
kernel/rcu/tree.c
2837
struct rcu_data *rdp = raw_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
2905
__this_cpu_write(rcu_data.rcu_cpu_has_work, 1);
kernel/rcu/tree.c
2906
t = __this_cpu_read(rcu_data.rcu_cpu_kthread_task);
kernel/rcu/tree.c
2908
rcu_wake_cond(t, __this_cpu_read(rcu_data.rcu_cpu_kthread_status));
kernel/rcu/tree.c
2927
per_cpu(rcu_data.rcu_cpu_kthread_status, cpu) = RCU_KTHREAD_OFFCPU;
kernel/rcu/tree.c
2932
return __this_cpu_read(rcu_data.rcu_cpu_has_work);
kernel/rcu/tree.c
2942
unsigned int *statusp = this_cpu_ptr(&rcu_data.rcu_cpu_kthread_status);
kernel/rcu/tree.c
2943
char work, *workp = this_cpu_ptr(&rcu_data.rcu_cpu_has_work);
kernel/rcu/tree.c
2944
unsigned long *j = this_cpu_ptr(&rcu_data.rcuc_activity);
kernel/rcu/tree.c
2974
.store = &rcu_data.rcu_cpu_kthread_task,
kernel/rcu/tree.c
2990
per_cpu(rcu_data.rcu_cpu_has_work, cpu) = 0;
kernel/rcu/tree.c
2998
static void rcutree_enqueue(struct rcu_data *rdp, struct rcu_head *head, rcu_callback_t func)
kernel/rcu/tree.c
3009
static void call_rcu_core(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree.c
3065
static void check_cb_ovld_locked(struct rcu_data *rdp, struct rcu_node *rnp)
kernel/rcu/tree.c
3088
static void check_cb_ovld(struct rcu_data *rdp)
kernel/rcu/tree.c
3107
struct rcu_data *rdp;
kernel/rcu/tree.c
312
static bool rcu_watching_snap_stopped_since(struct rcu_data *rdp, int snap)
kernel/rcu/tree.c
3134
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
3382
for (rnp = this_cpu_ptr(&rcu_data)->mynode; rnp; rnp = rnp->parent)
kernel/rcu/tree.c
3462
struct rcu_data *rdp;
kernel/rcu/tree.c
3466
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
362
raw_cpu_write(rcu_data.rcu_need_heavy_qs, false);
kernel/rcu/tree.c
3673
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
3754
static void rcu_barrier_entrain(struct rcu_data *rdp)
kernel/rcu/tree.c
3794
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
3822
struct rcu_data *rdp;
kernel/rcu/tree.c
3861
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
3903
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4005
static bool rcu_rdp_cpu_online(struct rcu_data *rdp)
kernel/rcu/tree.c
4012
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4034
struct rcu_data *rdp;
kernel/rcu/tree.c
4040
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
4144
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4244
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4293
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4305
struct rcu_data *rdp;
kernel/rcu/tree.c
4308
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4338
struct rcu_data *rdp;
kernel/rcu/tree.c
4343
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4396
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
4450
struct rcu_data *my_rdp;
kernel/rcu/tree.c
4452
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4466
my_rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
4520
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4536
struct rcu_data *rdp;
kernel/rcu/tree.c
4539
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
4584
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
4728
per_cpu_ptr(&rcu_data, i)->mynode = rnp;
kernel/rcu/tree.c
4729
per_cpu_ptr(&rcu_data, i)->barrier_head.next =
kernel/rcu/tree.c
4730
&per_cpu_ptr(&rcu_data, i)->barrier_head;
kernel/rcu/tree.c
515
static void force_qs_rnp(int (*f)(struct rcu_data *rdp));
kernel/rcu/tree.c
601
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
664
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree.c
712
return !rcu_segcblist_empty(&this_cpu_ptr(&rcu_data)->cblist) &&
kernel/rcu/tree.c
713
!rcu_rdp_is_offloaded(this_cpu_ptr(&rcu_data));
kernel/rcu/tree.c
721
static void rcu_disable_urgency_upon_qs(struct rcu_data *rdp)
kernel/rcu/tree.c
773
smp_store_release(per_cpu_ptr(&rcu_data.rcu_urgent_qs, cpu), true);
kernel/rcu/tree.c
80
static DEFINE_PER_CPU_SHARED_ALIGNED(struct rcu_data, rcu_data) = {
kernel/rcu/tree.c
802
static void rcu_gpnum_ovf(struct rcu_node *rnp, struct rcu_data *rdp)
kernel/rcu/tree.c
819
static int rcu_watching_snap_save(struct rcu_data *rdp)
kernel/rcu/tree.c
855
static int rcu_watching_snap_recheck(struct rcu_data *rdp)
kernel/rcu/tree.c
86
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree.c
992
static void trace_rcu_this_gp(struct rcu_node *rnp, struct rcu_data *rdp,
kernel/rcu/tree.h
266
struct rcu_data *nocb_toggling_rdp; /* rdp queued for (de-)offloading */
kernel/rcu/tree.h
269
struct rcu_data *nocb_gp_rdp ____cacheline_internodealigned_in_smp;
kernel/rcu/tree.h
490
static void rcu_preempt_deferred_qs_init(struct rcu_data *rdp);
kernel/rcu/tree.h
493
static bool rcu_is_callbacks_kthread(struct rcu_data *rdp);
kernel/rcu/tree.h
498
static void zero_cpu_stall_ticks(struct rcu_data *rdp);
kernel/rcu/tree.h
502
static bool wake_nocb_gp(struct rcu_data *rdp);
kernel/rcu/tree.h
503
static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
kernel/rcu/tree.h
505
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree.h
507
static void __maybe_unused __call_rcu_nocb_wake(struct rcu_data *rdp, bool was_empty,
kernel/rcu/tree.h
509
static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level);
kernel/rcu/tree.h
510
static bool do_nocb_deferred_wakeup(struct rcu_data *rdp);
kernel/rcu/tree.h
511
static void rcu_boot_init_nocb_percpu_data(struct rcu_data *rdp);
kernel/rcu/tree.h
513
static void show_rcu_nocb_state(struct rcu_data *rdp);
kernel/rcu/tree.h
514
static void rcu_nocb_lock(struct rcu_data *rdp);
kernel/rcu/tree.h
515
static void rcu_nocb_unlock(struct rcu_data *rdp);
kernel/rcu/tree.h
516
static void rcu_nocb_unlock_irqrestore(struct rcu_data *rdp,
kernel/rcu/tree.h
518
static void rcu_lockdep_assert_cblist_protected(struct rcu_data *rdp);
kernel/rcu/tree.h
542
static void check_cpu_stall(struct rcu_data *rdp);
kernel/rcu/tree.h
543
static void rcu_check_gp_start_stall(struct rcu_node *rnp, struct rcu_data *rdp,
kernel/rcu/tree_exp.h
1026
struct rcu_data *rdp;
kernel/rcu/tree_exp.h
1031
rdp = per_cpu_ptr(&rcu_data, raw_smp_processor_id());
kernel/rcu/tree_exp.h
245
struct rcu_data *rdp;
kernel/rcu/tree_exp.h
255
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
267
static void rcu_report_exp_rdp(struct rcu_data *rdp)
kernel/rcu/tree_exp.h
303
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, raw_smp_processor_id());
kernel/rcu/tree_exp.h
372
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
407
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
575
struct rcu_data *rdp;
kernel/rcu/tree_exp.h
581
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
632
struct rcu_data *rdp;
kernel/rcu/tree_exp.h
646
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_exp.h
733
ASSERT_EXCLUSIVE_WRITER_SCOPED(*this_cpu_ptr(&rcu_data.cpu_no_qs.b.exp));
kernel/rcu/tree_exp.h
734
__this_cpu_write(rcu_data.cpu_no_qs.b.exp, true);
kernel/rcu/tree_exp.h
736
smp_store_release(this_cpu_ptr(&rcu_data.rcu_urgent_qs), true);
kernel/rcu/tree_exp.h
753
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_exp.h
867
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_exp.h
873
__this_cpu_read(rcu_data.cpu_no_qs.b.exp))
kernel/rcu/tree_exp.h
877
rcu_report_exp_rdp(this_cpu_ptr(&rcu_data));
kernel/rcu/tree_nocb.h
1013
static bool rcu_nocb_rdp_deoffload_wait_cond(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1030
static int rcu_nocb_rdp_deoffload(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1034
struct rcu_data *rdp_gp = rdp->nocb_gp_rdp;
kernel/rcu/tree_nocb.h
1086
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
110
static bool rcu_nocb_bypass_trylock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1108
static bool rcu_nocb_rdp_offload_wait_cond(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1120
static int rcu_nocb_rdp_offload(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1154
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
119
static void rcu_nocb_bypass_unlock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1192
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1227
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1265
struct rcu_data *rdp;
kernel/rcu/tree_nocb.h
130
static void rcu_nocb_lock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1319
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1328
static void __init rcu_boot_init_nocb_percpu_data(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1349
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1350
struct rcu_data *rdp_gp;
kernel/rcu/tree_nocb.h
142
static void rcu_nocb_unlock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1427
struct rcu_data *rdp;
kernel/rcu/tree_nocb.h
1428
struct rcu_data *rdp_gp = NULL; /* Suppress misguided gcc warn. */
kernel/rcu/tree_nocb.h
1443
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_nocb.h
1501
static void show_rcu_nocb_gp_state(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1525
static void show_rcu_nocb_state(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1532
struct rcu_data *nocb_next_rdp;
kernel/rcu/tree_nocb.h
154
static void rcu_nocb_unlock_irqrestore(struct rcu_data *rdp,
kernel/rcu/tree_nocb.h
1599
static void rcu_nocb_lock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1604
static void rcu_nocb_unlock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1609
static void rcu_nocb_unlock_irqrestore(struct rcu_data *rdp,
kernel/rcu/tree_nocb.h
1616
static void rcu_lockdep_assert_cblist_protected(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1634
static bool wake_nocb_gp(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1639
static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
kernel/rcu/tree_nocb.h
1645
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree_nocb.h
1651
static void __call_rcu_nocb_wake(struct rcu_data *rdp, bool was_empty,
kernel/rcu/tree_nocb.h
1657
static void __init rcu_boot_init_nocb_percpu_data(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
166
static void rcu_lockdep_assert_cblist_protected(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1661
static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level)
kernel/rcu/tree_nocb.h
1666
static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
1675
static void show_rcu_nocb_state(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
194
static void nocb_defer_wakeup_cancel(struct rcu_data *rdp_gp)
kernel/rcu/tree_nocb.h
20
static inline bool rcu_current_is_nocb_kthread(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
202
static bool __wake_nocb_gp(struct rcu_data *rdp_gp,
kernel/rcu/tree_nocb.h
203
struct rcu_data *rdp,
kernel/rcu/tree_nocb.h
234
static bool wake_nocb_gp(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
237
struct rcu_data *rdp_gp = rdp->nocb_gp_rdp;
kernel/rcu/tree_nocb.h
272
static void wake_nocb_gp_defer(struct rcu_data *rdp, int waketype,
kernel/rcu/tree_nocb.h
276
struct rcu_data *rdp_gp = rdp->nocb_gp_rdp;
kernel/rcu/tree_nocb.h
314
static bool rcu_nocb_do_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp_in,
kernel/rcu/tree_nocb.h
358
static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
kernel/rcu/tree_nocb.h
372
static void rcu_nocb_try_flush_bypass(struct rcu_data *rdp, unsigned long j)
kernel/rcu/tree_nocb.h
399
static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
kernel/rcu/tree_nocb.h
530
static void __call_rcu_nocb_wake(struct rcu_data *rdp, bool was_alldone,
kernel/rcu/tree_nocb.h
575
static void call_rcu_nocb(struct rcu_data *rdp, struct rcu_head *head,
kernel/rcu/tree_nocb.h
587
static void nocb_gp_toggle_rdp(struct rcu_data *rdp_gp, struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
616
static void nocb_gp_sleep(struct rcu_data *my_rdp, int cpu)
kernel/rcu/tree_nocb.h
628
static void nocb_gp_wait(struct rcu_data *my_rdp)
kernel/rcu/tree_nocb.h
640
struct rcu_data *rdp, *rdp_toggling = NULL;
kernel/rcu/tree_nocb.h
835
struct rcu_data *rdp = arg;
kernel/rcu/tree_nocb.h
845
static inline bool nocb_cb_wait_cond(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
854
static void nocb_cb_wait(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
92
static void rcu_nocb_bypass_lock(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
923
struct rcu_data *rdp = arg;
kernel/rcu/tree_nocb.h
935
static int rcu_nocb_need_deferred_wakeup(struct rcu_data *rdp, int level)
kernel/rcu/tree_nocb.h
941
static bool do_nocb_deferred_wakeup_common(struct rcu_data *rdp_gp,
kernel/rcu/tree_nocb.h
942
struct rcu_data *rdp, int level,
kernel/rcu/tree_nocb.h
963
struct rcu_data *rdp = timer_container_of(rdp, t, nocb_timer);
kernel/rcu/tree_nocb.h
977
static bool do_nocb_deferred_wakeup(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
980
struct rcu_data *rdp_gp = rdp->nocb_gp_rdp;
kernel/rcu/tree_nocb.h
991
do_nocb_deferred_wakeup(this_cpu_ptr(&rcu_data));
kernel/rcu/tree_nocb.h
995
static int rcu_nocb_queue_toggle_rdp(struct rcu_data *rdp)
kernel/rcu/tree_nocb.h
997
struct rcu_data *rdp_gp = rdp->nocb_gp_rdp;
kernel/rcu/tree_plugin.h
1000
if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs)))
kernel/rcu/tree_plugin.h
1002
this_cpu_write(rcu_data.rcu_urgent_qs, false);
kernel/rcu/tree_plugin.h
1003
if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs)))
kernel/rcu/tree_plugin.h
1046
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_plugin.h
1105
static void rcu_preempt_deferred_qs_init(struct rcu_data *rdp) { }
kernel/rcu/tree_plugin.h
1114
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_plugin.h
1125
static bool rcu_is_callbacks_nocb_kthread(struct rcu_data *rdp)
kernel/rcu/tree_plugin.h
1138
static bool rcu_is_callbacks_kthread(struct rcu_data *rdp)
kernel/rcu/tree_plugin.h
16
static bool rcu_rdp_is_offloaded(struct rcu_data *rdp)
kernel/rcu/tree_plugin.h
162
static void rcu_preempt_ctxt_queue(struct rcu_node *rnp, struct rcu_data *rdp)
kernel/rcu/tree_plugin.h
301
if (__this_cpu_read(rcu_data.cpu_no_qs.b.norm)) {
kernel/rcu/tree_plugin.h
303
__this_cpu_read(rcu_data.gp_seq),
kernel/rcu/tree_plugin.h
305
__this_cpu_write(rcu_data.cpu_no_qs.b.norm, false);
kernel/rcu/tree_plugin.h
327
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_plugin.h
33
rdp == this_cpu_ptr(&rcu_data)) ||
kernel/rcu/tree_plugin.h
485
struct rcu_data *rdp;
kernel/rcu/tree_plugin.h
489
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_plugin.h
602
return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) ||
kernel/rcu/tree_plugin.h
629
struct rcu_data *rdp;
kernel/rcu/tree_plugin.h
632
rdp = container_of(iwp, struct rcu_data, defer_qs_iw);
kernel/rcu/tree_plugin.h
669
struct rcu_data *rdp,
kernel/rcu/tree_plugin.h
740
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_plugin.h
830
__this_cpu_read(rcu_data.core_needs_qs) &&
kernel/rcu/tree_plugin.h
831
__this_cpu_read(rcu_data.cpu_no_qs.b.norm) &&
kernel/rcu/tree_plugin.h
872
struct rcu_data *rdp;
kernel/rcu/tree_plugin.h
894
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_plugin.h
902
static void rcu_preempt_deferred_qs_init(struct rcu_data *rdp)
kernel/rcu/tree_plugin.h
915
struct rcu_data *rdp;
kernel/rcu/tree_plugin.h
928
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_plugin.h
953
if (!__this_cpu_read(rcu_data.cpu_no_qs.s))
kernel/rcu/tree_plugin.h
956
__this_cpu_read(rcu_data.gp_seq), TPS("cpuqs"));
kernel/rcu/tree_plugin.h
957
__this_cpu_write(rcu_data.cpu_no_qs.b.norm, false);
kernel/rcu/tree_plugin.h
958
if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp))
kernel/rcu/tree_plugin.h
959
rcu_report_exp_rdp(this_cpu_ptr(&rcu_data));
kernel/rcu/tree_plugin.h
973
if (!raw_cpu_read(rcu_data.rcu_urgent_qs))
kernel/rcu/tree_plugin.h
977
if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) {
kernel/rcu/tree_plugin.h
981
this_cpu_write(rcu_data.rcu_urgent_qs, false);
kernel/rcu/tree_plugin.h
982
if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) {
kernel/rcu/tree_stall.h
1002
static void rcu_check_gp_start_stall(struct rcu_node *rnp, struct rcu_data *rdp,
kernel/rcu/tree_stall.h
1067
struct rcu_data *rdp;
kernel/rcu/tree_stall.h
1077
rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_stall.h
216
static void zero_cpu_stall_ticks(struct rcu_data *rdp)
kernel/rcu/tree_stall.h
251
struct rcu_data *rdp;
kernel/rcu/tree_stall.h
254
rdp = container_of(iwp, struct rcu_data, rcu_iw);
kernel/rcu/tree_stall.h
457
static bool rcu_is_rcuc_kthread_starving(struct rcu_data *rdp, unsigned long *jp)
kernel/rcu/tree_stall.h
481
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
522
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
585
struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
713
struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
kernel/rcu/tree_stall.h
772
static void check_cpu_stall(struct rcu_data *rdp)
kernel/rcu/tree_stall.h
943
struct rcu_data *rdp;
kernel/rcu/tree_stall.h
979
rdp = per_cpu_ptr(&rcu_data, cpu);
kernel/rcu/tree_stall.h
989
rdp = per_cpu_ptr(&rcu_data, cpu);