|
@@ -2224,8 +2224,8 @@ static __always_inline u64 decay_load(u64 val, u64 n)
|
|
|
|
|
|
/*
|
|
|
* As y^PERIOD = 1/2, we can combine
|
|
|
- * y^n = 1/2^(n/PERIOD) * k^(n%PERIOD)
|
|
|
- * With a look-up table which covers k^n (n<PERIOD)
|
|
|
+ * y^n = 1/2^(n/PERIOD) * y^(n%PERIOD)
|
|
|
+ * With a look-up table which covers y^n (n<PERIOD)
|
|
|
*
|
|
|
* To achieve constant time decay_load.
|
|
|
*/
|
|
@@ -6410,7 +6410,7 @@ static struct sched_group *find_busiest_group(struct lb_env *env)
|
|
|
goto force_balance;
|
|
|
|
|
|
/*
|
|
|
- * If the local group is more busy than the selected busiest group
|
|
|
+ * If the local group is busier than the selected busiest group
|
|
|
* don't try and pull any tasks.
|
|
|
*/
|
|
|
if (local->avg_load >= busiest->avg_load)
|