is_idle_task
if (is_idle_task(prev))
if (!is_idle_task(current)) {
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current));
if (is_idle_task(p)) {
(is_idle_task(current) || arch_in_rcu_eqs())) {
if (!(event->attr.exclude_idle && is_idle_task(current)))
!is_idle_task(t) && READ_ONCE(t->rcu_tasks_idle_cpu) >= 0)) {
t, ".I"[is_idle_task(t)],
if (is_idle_task(t))
if (unlikely(is_idle_task(current))) {
if (unlikely(is_idle_task(current))) {
need_resched(), is_idle_task(current),
if (count >= bl && (need_resched() || !is_idle_task(current)))
is_idle_task(current), rcu_is_callbacks_kthread(rdp));
if (t && (status != RCU_KTHREAD_YIELDING || is_idle_task(current)))
RCU_LOCKDEP_WARN(!in_task() || !is_idle_task(current),
if (is_idle_task(curr) && tif == TIF_NEED_RESCHED_LAZY)
if (is_idle_task(rcu_dereference(rq->curr))) {
if (is_idle_task(rq->curr))
if (!is_idle_task(curr)) {
!is_idle_task(current) && !current->non_block_count) ||
is_idle_task(tsk)) {
if (is_idle_task(current))
return !is_idle_task(rq->curr) && !(rq->scx.flags & SCX_RQ_IN_BALANCE);
if (do_notify || is_idle_task(rq->curr))
is_idle = is_idle_task(rcu_dereference_all(rq->curr));
if (unlikely(is_idle_task(rq->curr))) {
(is_idle_task(current) && (irq_count() == HARDIRQ_OFFSET)))
if (is_idle_task(current))
is_idle_task(current)) { \