|
@@ -8041,13 +8041,12 @@ static void task_move_group_fair(struct task_struct *p, int queued)
|
|
se->vruntime -= cfs_rq_of(se)->min_vruntime;
|
|
se->vruntime -= cfs_rq_of(se)->min_vruntime;
|
|
set_task_rq(p, task_cpu(p));
|
|
set_task_rq(p, task_cpu(p));
|
|
se->depth = se->parent ? se->parent->depth + 1 : 0;
|
|
se->depth = se->parent ? se->parent->depth + 1 : 0;
|
|
- if (!queued) {
|
|
|
|
- cfs_rq = cfs_rq_of(se);
|
|
|
|
|
|
+ cfs_rq = cfs_rq_of(se);
|
|
|
|
+ if (!queued)
|
|
se->vruntime += cfs_rq->min_vruntime;
|
|
se->vruntime += cfs_rq->min_vruntime;
|
|
|
|
|
|
- /* Virtually synchronize task with its new cfs_rq */
|
|
|
|
- attach_entity_load_avg(cfs_rq, se);
|
|
|
|
- }
|
|
|
|
|
|
+ /* Virtually synchronize task with its new cfs_rq */
|
|
|
|
+ attach_entity_load_avg(cfs_rq, se);
|
|
}
|
|
}
|
|
|
|
|
|
void free_fair_sched_group(struct task_group *tg)
|
|
void free_fair_sched_group(struct task_group *tg)
|