|
|
@@ -493,6 +493,11 @@ static void sched_rt_rq_dequeue(struct rt_rq *rt_rq)
|
|
|
dequeue_rt_entity(rt_se);
|
|
|
}
|
|
|
|
|
|
+static inline int rt_rq_throttled(struct rt_rq *rt_rq)
|
|
|
+{
|
|
|
+ return rt_rq->rt_throttled && !rt_rq->rt_nr_boosted;
|
|
|
+}
|
|
|
+
|
|
|
static int rt_se_boosted(struct sched_rt_entity *rt_se)
|
|
|
{
|
|
|
struct rt_rq *rt_rq = group_rt_rq(rt_se);
|
|
|
@@ -569,6 +574,11 @@ static inline void sched_rt_rq_dequeue(struct rt_rq *rt_rq)
|
|
|
dequeue_top_rt_rq(rt_rq);
|
|
|
}
|
|
|
|
|
|
+static inline int rt_rq_throttled(struct rt_rq *rt_rq)
|
|
|
+{
|
|
|
+ return rt_rq->rt_throttled;
|
|
|
+}
|
|
|
+
|
|
|
static inline const struct cpumask *sched_rt_period_mask(void)
|
|
|
{
|
|
|
return cpu_online_mask;
|