|
@@ -717,17 +717,12 @@ static u64 sched_vslice(struct cfs_rq *cfs_rq, struct sched_entity *se)
|
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
+
|
|
|
+#include "sched-pelt.h"
|
|
|
+
|
|
|
static int select_idle_sibling(struct task_struct *p, int prev_cpu, int cpu);
|
|
|
static unsigned long task_h_load(struct task_struct *p);
|
|
|
|
|
|
-/*
|
|
|
- * We choose a half-life close to 1 scheduling period.
|
|
|
- * Note: The tables runnable_avg_yN_inv and runnable_avg_yN_sum are
|
|
|
- * dependent on this value.
|
|
|
- */
|
|
|
-#define LOAD_AVG_PERIOD 32
|
|
|
-#define LOAD_AVG_MAX 47742 /* maximum possible load avg */
|
|
|
-
|
|
|
/* Give new sched_entity start runnable values to heavy its load in infant time */
|
|
|
void init_entity_runnable_average(struct sched_entity *se)
|
|
|
{
|
|
@@ -2732,16 +2727,6 @@ static inline void update_cfs_shares(struct sched_entity *se)
|
|
|
#endif /* CONFIG_FAIR_GROUP_SCHED */
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
-/* Precomputed fixed inverse multiplies for multiplication by y^n */
|
|
|
-static const u32 runnable_avg_yN_inv[] = {
|
|
|
- 0xffffffff, 0xfa83b2da, 0xf5257d14, 0xefe4b99a, 0xeac0c6e6, 0xe5b906e6,
|
|
|
- 0xe0ccdeeb, 0xdbfbb796, 0xd744fcc9, 0xd2a81d91, 0xce248c14, 0xc9b9bd85,
|
|
|
- 0xc5672a10, 0xc12c4cc9, 0xbd08a39e, 0xb8fbaf46, 0xb504f333, 0xb123f581,
|
|
|
- 0xad583ee9, 0xa9a15ab4, 0xa5fed6a9, 0xa2704302, 0x9ef5325f, 0x9b8d39b9,
|
|
|
- 0x9837f050, 0x94f4efa8, 0x91c3d373, 0x8ea4398a, 0x8b95c1e3, 0x88980e80,
|
|
|
- 0x85aac367, 0x82cd8698,
|
|
|
-};
|
|
|
-
|
|
|
/*
|
|
|
* Approximate:
|
|
|
* val * y^n, where y^32 ~= 0.5 (~1 scheduling period)
|