|
@@ -6073,7 +6073,6 @@ static void update_cpu_capacity(struct sched_domain *sd, int cpu)
|
|
|
capacity >>= SCHED_CAPACITY_SHIFT;
|
|
|
|
|
|
cpu_rq(cpu)->cpu_capacity_orig = capacity;
|
|
|
- sdg->sgc->capacity_orig = capacity;
|
|
|
|
|
|
capacity *= scale_rt_capacity(cpu);
|
|
|
capacity >>= SCHED_CAPACITY_SHIFT;
|
|
@@ -6089,7 +6088,7 @@ void update_group_capacity(struct sched_domain *sd, int cpu)
|
|
|
{
|
|
|
struct sched_domain *child = sd->child;
|
|
|
struct sched_group *group, *sdg = sd->groups;
|
|
|
- unsigned long capacity, capacity_orig;
|
|
|
+ unsigned long capacity;
|
|
|
unsigned long interval;
|
|
|
|
|
|
interval = msecs_to_jiffies(sd->balance_interval);
|
|
@@ -6101,7 +6100,7 @@ void update_group_capacity(struct sched_domain *sd, int cpu)
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
- capacity_orig = capacity = 0;
|
|
|
+ capacity = 0;
|
|
|
|
|
|
if (child->flags & SD_OVERLAP) {
|
|
|
/*
|
|
@@ -6121,19 +6120,15 @@ void update_group_capacity(struct sched_domain *sd, int cpu)
|
|
|
* Use capacity_of(), which is set irrespective of domains
|
|
|
* in update_cpu_capacity().
|
|
|
*
|
|
|
- * This avoids capacity/capacity_orig from being 0 and
|
|
|
+ * This avoids capacity from being 0 and
|
|
|
* causing divide-by-zero issues on boot.
|
|
|
- *
|
|
|
- * Runtime updates will correct capacity_orig.
|
|
|
*/
|
|
|
if (unlikely(!rq->sd)) {
|
|
|
- capacity_orig += capacity_orig_of(cpu);
|
|
|
capacity += capacity_of(cpu);
|
|
|
continue;
|
|
|
}
|
|
|
|
|
|
sgc = rq->sd->groups->sgc;
|
|
|
- capacity_orig += sgc->capacity_orig;
|
|
|
capacity += sgc->capacity;
|
|
|
}
|
|
|
} else {
|
|
@@ -6144,13 +6139,11 @@ void update_group_capacity(struct sched_domain *sd, int cpu)
|
|
|
|
|
|
group = child->groups;
|
|
|
do {
|
|
|
- capacity_orig += group->sgc->capacity_orig;
|
|
|
capacity += group->sgc->capacity;
|
|
|
group = group->next;
|
|
|
} while (group != child->groups);
|
|
|
}
|
|
|
|
|
|
- sdg->sgc->capacity_orig = capacity_orig;
|
|
|
sdg->sgc->capacity = capacity;
|
|
|
}
|
|
|
|