sched_feat
if (sched_feat(TTWU_QUEUE) && ttwu_queue_cond(p, cpu)) {
if (sched_feat(LATENCY_WARN))
if (sched_feat(LATENCY_WARN) && resched_latency)
if (sched_feat(HRTICK) || sched_feat(HRTICK_DL))
if ((irq_delta + steal) && sched_feat(NONTASK_CAPACITY))
if (sched_feat(WARN_DOUBLE_CLOCK))
if (sched_feat(PICK_BUDDY) &&
if (!sched_feat(SIS_UTIL) || env->idle == CPU_NEWLY_IDLE)
if (sched_feat(NI_RANDOM)) {
update_load_avg(cfs_rq, se, sched_feat(ATTACH_AGE_LOAD) ? 0 : SKIP_AGE_LOAD);
if (!sched_feat(ATTACH_AGE_LOAD))
if (!sched_feat(UTIL_EST))
if (!sched_feat(UTIL_EST))
if (!sched_feat(UTIL_EST))
if (sched_feat(PLACE_LAG) && cfs_rq->nr_queued && se->vlag) {
if (sched_feat(PLACE_DEADLINE_INITIAL) && (flags & ENQUEUE_INITIAL))
if (sched_feat(DELAY_ZERO) && se->vlag > 0)
if (sched_feat(DELAY_DEQUEUE) && delay &&
if (sched_feat(PLACE_REL_DEADLINE) && !sleep) {
if (sched_feat(DELAY_ZERO)) {
if (sched_feat(WA_BIAS))
if (sched_feat(WA_BIAS))
if (sched_feat(WA_IDLE))
if (sched_feat(WA_WEIGHT) && target == nr_cpumask_bits)
if (sched_feat(SIS_UTIL)) {
if (sched_feat(UTIL_EST)) {
if (!sched_feat(WAKEUP_PREEMPTION))
if (sched_feat(PREEMPT_SHORT) && (pse->slice < se->slice)) {
if (sched_feat(NEXT_BUDDY) &&
if (sched_feat(RUN_TO_PARITY))
if (sched_feat(CACHE_HOT_BUDDY) && env->dst_rq->nr_running &&
if (sched_feat(PLACE_LAG) && dst_cfs_rq->nr_queued &&
if (sched_feat(RUN_TO_PARITY))
if (sched_feat(LB_MIN) &&
if (!sched_feat(UTIL_EST))
if (sched_feat(RT_PUSH_IPI)) {
if (!sched_feat(RT_RUNTIME_SHARE))
if (!sched_feat(RT_RUNTIME_SHARE) && rt_rq->rt_runtime != RUNTIME_INF)
if (!sched_feat(HRTICK))
if (!sched_feat(HRTICK_DL))