|
@@ -2879,20 +2879,21 @@ static inline int update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
|
|
{
|
|
{
|
|
struct sched_avg *sa = &cfs_rq->avg;
|
|
struct sched_avg *sa = &cfs_rq->avg;
|
|
struct rq *rq = rq_of(cfs_rq);
|
|
struct rq *rq = rq_of(cfs_rq);
|
|
- int decayed, removed = 0;
|
|
|
|
|
|
+ int decayed, removed_load = 0, removed_util = 0;
|
|
int cpu = cpu_of(rq);
|
|
int cpu = cpu_of(rq);
|
|
|
|
|
|
if (atomic_long_read(&cfs_rq->removed_load_avg)) {
|
|
if (atomic_long_read(&cfs_rq->removed_load_avg)) {
|
|
s64 r = atomic_long_xchg(&cfs_rq->removed_load_avg, 0);
|
|
s64 r = atomic_long_xchg(&cfs_rq->removed_load_avg, 0);
|
|
sa->load_avg = max_t(long, sa->load_avg - r, 0);
|
|
sa->load_avg = max_t(long, sa->load_avg - r, 0);
|
|
sa->load_sum = max_t(s64, sa->load_sum - r * LOAD_AVG_MAX, 0);
|
|
sa->load_sum = max_t(s64, sa->load_sum - r * LOAD_AVG_MAX, 0);
|
|
- removed = 1;
|
|
|
|
|
|
+ removed_load = 1;
|
|
}
|
|
}
|
|
|
|
|
|
if (atomic_long_read(&cfs_rq->removed_util_avg)) {
|
|
if (atomic_long_read(&cfs_rq->removed_util_avg)) {
|
|
long r = atomic_long_xchg(&cfs_rq->removed_util_avg, 0);
|
|
long r = atomic_long_xchg(&cfs_rq->removed_util_avg, 0);
|
|
sa->util_avg = max_t(long, sa->util_avg - r, 0);
|
|
sa->util_avg = max_t(long, sa->util_avg - r, 0);
|
|
sa->util_sum = max_t(s32, sa->util_sum - r * LOAD_AVG_MAX, 0);
|
|
sa->util_sum = max_t(s32, sa->util_sum - r * LOAD_AVG_MAX, 0);
|
|
|
|
+ removed_util = 1;
|
|
}
|
|
}
|
|
|
|
|
|
decayed = __update_load_avg(now, cpu, sa,
|
|
decayed = __update_load_avg(now, cpu, sa,
|
|
@@ -2903,7 +2904,8 @@ static inline int update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
|
|
cfs_rq->load_last_update_time_copy = sa->last_update_time;
|
|
cfs_rq->load_last_update_time_copy = sa->last_update_time;
|
|
#endif
|
|
#endif
|
|
|
|
|
|
- if (cpu == smp_processor_id() && &rq->cfs == cfs_rq) {
|
|
|
|
|
|
+ if (cpu == smp_processor_id() && &rq->cfs == cfs_rq &&
|
|
|
|
+ (decayed || removed_util)) {
|
|
unsigned long max = rq->cpu_capacity_orig;
|
|
unsigned long max = rq->cpu_capacity_orig;
|
|
|
|
|
|
/*
|
|
/*
|
|
@@ -2926,7 +2928,7 @@ static inline int update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq)
|
|
min(sa->util_avg, max), max);
|
|
min(sa->util_avg, max), max);
|
|
}
|
|
}
|
|
|
|
|
|
- return decayed || removed;
|
|
|
|
|
|
+ return decayed || removed_load;
|
|
}
|
|
}
|
|
|
|
|
|
/* Update task and its cfs_rq load average */
|
|
/* Update task and its cfs_rq load average */
|