in_nmi
!in_hardirq() && !in_nmi();
in_nmi() ? "NMI" : "hardirq");
((task == current) && in_nmi()) \
if (!in_nmi())
static volatile int in_nmi;
if (in_nmi)
in_nmi = 1;
in_nmi = 0;
WRITE_ONCE(in_nmi, 0);
static int in_nmi;
if (READ_ONCE(in_nmi))
WRITE_ONCE(in_nmi, 1);
u16 in_nmi; /* In nmi handler */
OFFSET(PACA_IN_NMI, paca_struct, in_nmi);
WARN_ON_ONCE(in_nmi());
get_paca()->in_nmi = 1;
get_paca()->in_nmi = 0;
p->irq_soft_mask, p->in_mce, p->in_nmi);
BUG_ON(get_paca()->in_nmi == 0);
if (get_paca()->in_nmi > 1)
if (in_nmi())
DUMP(p, in_nmi, "%#-*x");
if (in_hardirq() || in_nmi())
if (in_nmi())
if (in_nmi())
if (in_nmi())
WARN_ON(in_nmi());
WARN_ON(in_nmi());
WARN_ON(in_nmi());
(!!in_nmi() == ((vcpu)->arch.handling_intr_from_guest == KVM_HANDLING_NMI)))
atomic_t in_nmi; /* flag this node in UV NMI IRQ */
if (WARN_ON_ONCE(in_nmi()))
if (in_nmi())
if (in_nmi())
WARN(in_nmi(),
if (!in_nmi()) {
VM_WARN_ON(in_nmi() || preemptible());
int first = atomic_add_unless(&hub_nmi->in_nmi, 1, 1);
nmi = atomic_read(&hub_nmi->in_nmi);
nmi = atomic_read(&hub_nmi->in_nmi);
atomic_set(&hub_nmi->in_nmi, 0);
if (irq == -1 || in_nmi())
if (!in_nmi())
if (!in_nmi())
if (!in_nmi())
if (WARN_ON_ONCE(in_nmi()))
WARN_ON(in_nmi());
WARN_ON(in_nmi());
if (in_nmi())
if (in_nmi()) {
if (in_nmi())
in_nmi() ? "NMI" : why);
BUG_ON(in_nmi() == NMI_MASK); \
BUG_ON(!in_nmi()); \
if (in_nmi() | in_hardirq()) { \
(!in_softirq() || in_hardirq() || in_nmi())); \
WARN_ON_ONCE(IS_ENABLED(CONFIG_PROVE_RCU) && in_nmi());
WARN_ON_ONCE(in_nmi());
WARN_ON_ONCE(IS_ENABLED(CONFIG_PROVE_RCU) && in_nmi());
WARN_ON_ONCE(in_nmi());
if (unlikely(idx >= _Q_MAX_NODES || (in_nmi() && idx > 0))) {
!IS_ENABLED(CONFIG_ARCH_HAS_NMI_SAFE_THIS_CPU_OPS)) && in_nmi())
if (!in_nmi())
if (!in_nmi())
} else if (!in_nmi()) {
else if (in_nmi())
notify_mode = in_nmi() ? TWA_NMI_CURRENT : TWA_RESUME;
WARN_ON_ONCE(!in_nmi());
WARN_ON_ONCE(in_nmi());
return in_serving_softirq() && !in_hardirq() && !in_nmi();
if (unlikely(in_nmi()))
if (unlikely(in_nmi())) {
if (in_nmi()) {
if (!in_nmi())
if (in_nmi()) {
if (in_nmi())
in_nmi() ||
if (!IS_ENABLED(CONFIG_RCU_TORTURE_TEST_CHK_RDR_STATE) || in_nmi())
WARN_ON_ONCE(in_nmi());
read_flavor != SRCU_READ_FLAVOR_FAST && in_nmi());
if (in_nmi() || !rcu_scheduler_fully_active)
if (in_nmi())
if (in_nmi())
local_timers_pending_force_th() && !(in_nmi() | in_hardirq()))
if (!WARN_ON_ONCE(in_nmi()))
if (unlikely(in_nmi()))
(unlikely(in_nmi()))) {
if (likely(!in_nmi())) {
} else if (in_nmi()) {
if (unlikely(in_nmi()))
if (unlikely(in_nmi()))
if (in_nmi()) {
if (unlikely(in_nmi()))
if (in_nmi()) \
if (in_nmi())
if (in_nmi()) {
BUG_ON(in_nmi());
BUG_ON(in_nmi());
if (in_nmi() || !allow_spin) {
if (in_hardirq() || in_serving_softirq() || in_nmi()) {
if (in_nmi())
if (likely(!in_nmi())) {
if (IS_ENABLED(CONFIG_MEMCG_NMI_UNSAFE) && in_nmi())
if (likely(!in_nmi())) {
&& (in_nmi() || in_hardirq()))) {
if (IS_ENABLED(CONFIG_PREEMPT_RT) && (in_nmi() || in_hardirq()))
if (IS_ENABLED(CONFIG_PREEMPT_RT) && (in_nmi() || in_hardirq()))
BUG_ON(in_nmi());
BUG_ON(in_nmi());
if (in_hardirq() || in_nmi())
if (in_hardirq() || in_nmi())