|
@@ -2712,6 +2712,9 @@ static inline void update_load_avg(struct sched_entity *se, int update_tg)
|
|
|
|
|
|
static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se)
|
|
|
{
|
|
|
+ if (!sched_feat(ATTACH_AGE_LOAD))
|
|
|
+ goto skip_aging;
|
|
|
+
|
|
|
/*
|
|
|
* If we got migrated (either between CPUs or between cgroups) we'll
|
|
|
* have aged the average right before clearing @last_update_time.
|
|
@@ -2726,6 +2729,7 @@ static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *s
|
|
|
*/
|
|
|
}
|
|
|
|
|
|
+skip_aging:
|
|
|
se->avg.last_update_time = cfs_rq->avg.last_update_time;
|
|
|
cfs_rq->avg.load_avg += se->avg.load_avg;
|
|
|
cfs_rq->avg.load_sum += se->avg.load_sum;
|