|
@@ -6199,6 +6199,13 @@ static inline int find_idlest_cpu(struct sched_domain *sd, struct task_struct *p
|
|
|
if (!cpumask_intersects(sched_domain_span(sd), &p->cpus_allowed))
|
|
|
return prev_cpu;
|
|
|
|
|
|
+ /*
|
|
|
+ * We need task's util for capacity_spare_wake, sync it up to prev_cpu's
|
|
|
+ * last_update_time.
|
|
|
+ */
|
|
|
+ if (!(sd_flag & SD_BALANCE_FORK))
|
|
|
+ sync_entity_load_avg(&p->se);
|
|
|
+
|
|
|
while (sd) {
|
|
|
struct sched_group *group;
|
|
|
struct sched_domain *tmp;
|
|
@@ -6651,15 +6658,6 @@ select_task_rq_fair(struct task_struct *p, int prev_cpu, int sd_flag, int wake_f
|
|
|
|
|
|
if (unlikely(sd)) {
|
|
|
/* Slow path */
|
|
|
-
|
|
|
- /*
|
|
|
- * We're going to need the task's util for capacity_spare_wake
|
|
|
- * in find_idlest_group. Sync it up to prev_cpu's
|
|
|
- * last_update_time.
|
|
|
- */
|
|
|
- if (!(sd_flag & SD_BALANCE_FORK))
|
|
|
- sync_entity_load_avg(&p->se);
|
|
|
-
|
|
|
new_cpu = find_idlest_cpu(sd, p, cpu, prev_cpu, sd_flag);
|
|
|
} else if (sd_flag & SD_BALANCE_WAKE) { /* XXX always ? */
|
|
|
/* Fast path */
|