|
@@ -6531,7 +6531,7 @@ out:
|
|
|
* idle_balance is called by schedule() if this_cpu is about to become
|
|
|
* idle. Attempts to pull tasks from other CPUs.
|
|
|
*/
|
|
|
-void idle_balance(struct rq *this_rq)
|
|
|
+int idle_balance(struct rq *this_rq)
|
|
|
{
|
|
|
struct sched_domain *sd;
|
|
|
int pulled_task = 0;
|
|
@@ -6539,10 +6539,8 @@ void idle_balance(struct rq *this_rq)
|
|
|
u64 curr_cost = 0;
|
|
|
int this_cpu = this_rq->cpu;
|
|
|
|
|
|
- this_rq->idle_stamp = rq_clock(this_rq);
|
|
|
-
|
|
|
if (this_rq->avg_idle < sysctl_sched_migration_cost)
|
|
|
- return;
|
|
|
+ return 0;
|
|
|
|
|
|
/*
|
|
|
* Drop the rq->lock, but keep IRQ/preempt disabled.
|
|
@@ -6580,10 +6578,8 @@ void idle_balance(struct rq *this_rq)
|
|
|
interval = msecs_to_jiffies(sd->balance_interval);
|
|
|
if (time_after(next_balance, sd->last_balance + interval))
|
|
|
next_balance = sd->last_balance + interval;
|
|
|
- if (pulled_task) {
|
|
|
- this_rq->idle_stamp = 0;
|
|
|
+ if (pulled_task)
|
|
|
break;
|
|
|
- }
|
|
|
}
|
|
|
rcu_read_unlock();
|
|
|
|
|
@@ -6594,7 +6590,7 @@ void idle_balance(struct rq *this_rq)
|
|
|
* A task could have be enqueued in the meantime
|
|
|
*/
|
|
|
if (this_rq->nr_running && !pulled_task)
|
|
|
- return;
|
|
|
+ return 1;
|
|
|
|
|
|
if (pulled_task || time_after(jiffies, this_rq->next_balance)) {
|
|
|
/*
|
|
@@ -6606,6 +6602,8 @@ void idle_balance(struct rq *this_rq)
|
|
|
|
|
|
if (curr_cost > this_rq->max_idle_balance_cost)
|
|
|
this_rq->max_idle_balance_cost = curr_cost;
|
|
|
+
|
|
|
+ return pulled_task;
|
|
|
}
|
|
|
|
|
|
/*
|