|
@@ -369,8 +369,9 @@ static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
}
|
|
|
|
|
|
/* Iterate thr' all leaf cfs_rq's on a runqueue */
|
|
|
-#define for_each_leaf_cfs_rq(rq, cfs_rq) \
|
|
|
- list_for_each_entry_rcu(cfs_rq, &rq->leaf_cfs_rq_list, leaf_cfs_rq_list)
|
|
|
+#define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \
|
|
|
+ list_for_each_entry_safe(cfs_rq, pos, &rq->leaf_cfs_rq_list, \
|
|
|
+ leaf_cfs_rq_list)
|
|
|
|
|
|
/* Do the two (enqueued) entities belong to the same group ? */
|
|
|
static inline struct cfs_rq *
|
|
@@ -463,8 +464,8 @@ static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq)
|
|
|
{
|
|
|
}
|
|
|
|
|
|
-#define for_each_leaf_cfs_rq(rq, cfs_rq) \
|
|
|
- for (cfs_rq = &rq->cfs; cfs_rq; cfs_rq = NULL)
|
|
|
+#define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \
|
|
|
+ for (cfs_rq = &rq->cfs, pos = NULL; cfs_rq; cfs_rq = pos)
|
|
|
|
|
|
static inline struct sched_entity *parent_entity(struct sched_entity *se)
|
|
|
{
|
|
@@ -6953,10 +6954,28 @@ static void attach_tasks(struct lb_env *env)
|
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_FAIR_GROUP_SCHED
|
|
|
+
|
|
|
+static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq)
|
|
|
+{
|
|
|
+ if (cfs_rq->load.weight)
|
|
|
+ return false;
|
|
|
+
|
|
|
+ if (cfs_rq->avg.load_sum)
|
|
|
+ return false;
|
|
|
+
|
|
|
+ if (cfs_rq->avg.util_sum)
|
|
|
+ return false;
|
|
|
+
|
|
|
+ if (cfs_rq->runnable_load_sum)
|
|
|
+ return false;
|
|
|
+
|
|
|
+ return true;
|
|
|
+}
|
|
|
+
|
|
|
static void update_blocked_averages(int cpu)
|
|
|
{
|
|
|
struct rq *rq = cpu_rq(cpu);
|
|
|
- struct cfs_rq *cfs_rq;
|
|
|
+ struct cfs_rq *cfs_rq, *pos;
|
|
|
struct rq_flags rf;
|
|
|
|
|
|
rq_lock_irqsave(rq, &rf);
|
|
@@ -6966,7 +6985,7 @@ static void update_blocked_averages(int cpu)
|
|
|
* Iterates the task_group tree in a bottom up fashion, see
|
|
|
* list_add_leaf_cfs_rq() for details.
|
|
|
*/
|
|
|
- for_each_leaf_cfs_rq(rq, cfs_rq) {
|
|
|
+ for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) {
|
|
|
struct sched_entity *se;
|
|
|
|
|
|
/* throttled entities do not contribute to load */
|
|
@@ -6980,6 +6999,13 @@ static void update_blocked_averages(int cpu)
|
|
|
se = cfs_rq->tg->se[cpu];
|
|
|
if (se && !skip_blocked_update(se))
|
|
|
update_load_avg(se, 0);
|
|
|
+
|
|
|
+ /*
|
|
|
+ * There can be a lot of idle CPU cgroups. Don't let fully
|
|
|
+ * decayed cfs_rqs linger on the list.
|
|
|
+ */
|
|
|
+ if (cfs_rq_is_decayed(cfs_rq))
|
|
|
+ list_del_leaf_cfs_rq(cfs_rq);
|
|
|
}
|
|
|
rq_unlock_irqrestore(rq, &rf);
|
|
|
}
|
|
@@ -9503,10 +9529,10 @@ const struct sched_class fair_sched_class = {
|
|
|
#ifdef CONFIG_SCHED_DEBUG
|
|
|
void print_cfs_stats(struct seq_file *m, int cpu)
|
|
|
{
|
|
|
- struct cfs_rq *cfs_rq;
|
|
|
+ struct cfs_rq *cfs_rq, *pos;
|
|
|
|
|
|
rcu_read_lock();
|
|
|
- for_each_leaf_cfs_rq(cpu_rq(cpu), cfs_rq)
|
|
|
+ for_each_leaf_cfs_rq_safe(cpu_rq(cpu), cfs_rq, pos)
|
|
|
print_cfs_rq(m, cpu, cfs_rq);
|
|
|
rcu_read_unlock();
|
|
|
}
|