softirq_count
int softirq_count;
TI_SOFTIRQ != offsetof(struct thread_info, softirq_count) ||
return !softirq_count();
# define irq_count() ((preempt_count() & (NMI_MASK | HARDIRQ_MASK)) | softirq_count())
#define in_serving_softirq() (softirq_count() & SOFTIRQ_OFFSET)
#define in_softirq() (softirq_count())
if (softirq_count()) {
int softirq_count; /* # softirqs during this sample */
__field( int, softirq_count )
__entry->softirq_count = s->softirq_count;
__entry->softirq_count,
curr->softirq_context, softirq_count() >> SOFTIRQ_SHIFT,
lockdep_softirq_context(curr), softirq_count() >> SOFTIRQ_SHIFT,
DEBUG_LOCKS_WARN_ON(!softirq_count());
if (softirq_count()) {
!softirq_count(), ROEC_ARGS);
softirq_count(), ROEC_ARGS);
if (IS_ENABLED(CONFIG_PREEMPT_RT) && softirq_count())
if ((preempt_count() & HARDIRQ_MASK) || softirq_count())
((!(IS_ENABLED(CONFIG_PREEMPT_COUNT) && preemptible()) || softirq_count()) &&
if (softirq_count() == cnt)
if (softirq_count() == (cnt & SOFTIRQ_MASK))
if (softirq_count() == (cnt & SOFTIRQ_MASK))
if (softirq_count() == SOFTIRQ_DISABLE_OFFSET)
if (softirq_count() >> (SOFTIRQ_SHIFT + 1))
__field( unsigned int, softirq_count )
__entry->softirq_count,
s->softirq_count = osn_var->softirq.count;
s->softirq_count = osn_var->softirq.count - s->softirq_count;
entry->softirq_count = sample->softirq_count;
trace_seq_printf(s, " %6u", field->softirq_count);
field->softirq_count,
if (softirq_count())
lockdep_assert_once(hardirq_count() | softirq_count());
bool need_bh_off = !(hardirq_count() | softirq_count());
WARN_ON_ONCE(softirq_count() == 0);
WARN_ON_ONCE(softirq_count() == 0);
WARN_ON_ONCE(softirq_count() == 0);
update_sum(&cpu_data->softirq_count, &val);
trace_seq_printf(s, "%12llu ", cpu_data->softirq_count);
unsigned long long softirq_count;