@@ -2558,6 +2558,7 @@ static inline void update_cpu_concurrency(struct rq *rq);
*/
void idle_enter_fair(struct rq *this_rq)
{
+ update_cpu_concurrency(this_rq);
}
/*
@@ -2567,6 +2568,7 @@ void idle_enter_fair(struct rq *this_rq)
*/
void idle_exit_fair(struct rq *this_rq)
{
+ update_cpu_concurrency(this_rq);
}
static int idle_balance(struct rq *this_rq);
@@ -3877,8 +3879,10 @@ enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags)
update_entity_load_avg(se, 1);
}
- if (!se)
+ if (!se) {
inc_nr_running(rq);
+ update_cpu_concurrency(rq);
+ }
hrtick_update(rq);
}
@@ -3937,8 +3941,10 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags)
update_entity_load_avg(se, 1);
}
- if (!se)
+ if (!se) {
dec_nr_running(rq);
+ update_cpu_concurrency(rq);
+ }
hrtick_update(rq);
}
@@ -5381,6 +5387,8 @@ static void update_blocked_averages(int cpu)
__update_blocked_averages_cpu(cfs_rq->tg, rq->cpu);
}
+ update_cpu_concurrency(rq);
+
raw_spin_unlock_irqrestore(&rq->lock, flags);
}
@@ -7233,6 +7241,8 @@ static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued)
if (numabalancing_enabled)
task_tick_numa(rq, curr);
+
+ update_cpu_concurrency(rq);
}
/*
CC can only be modified when enqueue and dequeue the CPU rq. We also update it in scheduler tick, load balancing, and idle enter/exit in case we may not have enqueue and dequeue for a long time. Therefore, we update/track CC in and only in these points: we update cpu concurrency at: 1) enqueue task, which increases concurrency 2) dequeue task, which decreases concurrency 3) periodic scheduler tick, in case no en/dequeue for long 4) enter and exit idle 5) update_blocked_averages Signed-off-by: Yuyang Du <yuyang.du@intel.com> --- kernel/sched/fair.c | 14 ++++++++++++-- 1 file changed, 12 insertions(+), 2 deletions(-)