|
@@ -1340,7 +1340,12 @@ select_task_rq_rt(struct task_struct *p, int cpu, int sd_flag, int flags)
|
|
|
curr->prio <= p->prio)) {
|
|
|
int target = find_lowest_rq(p);
|
|
|
|
|
|
- if (target != -1)
|
|
|
+ /*
|
|
|
+ * Don't bother moving it if the destination CPU is
|
|
|
+ * not running a lower priority task.
|
|
|
+ */
|
|
|
+ if (target != -1 &&
|
|
|
+ p->prio < cpu_rq(target)->rt.highest_prio.curr)
|
|
|
cpu = target;
|
|
|
}
|
|
|
rcu_read_unlock();
|
|
@@ -1617,6 +1622,16 @@ static struct rq *find_lock_lowest_rq(struct task_struct *task, struct rq *rq)
|
|
|
|
|
|
lowest_rq = cpu_rq(cpu);
|
|
|
|
|
|
+ if (lowest_rq->rt.highest_prio.curr <= task->prio) {
|
|
|
+ /*
|
|
|
+ * Target rq has tasks of equal or higher priority,
|
|
|
+ * retrying does not release any lock and is unlikely
|
|
|
+ * to yield a different result.
|
|
|
+ */
|
|
|
+ lowest_rq = NULL;
|
|
|
+ break;
|
|
|
+ }
|
|
|
+
|
|
|
/* if the prio of this runqueue changed, try again */
|
|
|
if (double_lock_balance(rq, lowest_rq)) {
|
|
|
/*
|