|
@@ -581,68 +581,6 @@ hrtimer_force_reprogram(struct hrtimer_cpu_base *cpu_base, int skip_equal)
|
|
tick_program_event(cpu_base->expires_next, 1);
|
|
tick_program_event(cpu_base->expires_next, 1);
|
|
}
|
|
}
|
|
|
|
|
|
-/*
|
|
|
|
- * When a timer is enqueued and expires earlier than the already enqueued
|
|
|
|
- * timers, we have to check, whether it expires earlier than the timer for
|
|
|
|
- * which the clock event device was armed.
|
|
|
|
- *
|
|
|
|
- * Called with interrupts disabled and base->cpu_base.lock held
|
|
|
|
- */
|
|
|
|
-static void hrtimer_reprogram(struct hrtimer *timer,
|
|
|
|
- struct hrtimer_clock_base *base)
|
|
|
|
-{
|
|
|
|
- struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
|
|
|
|
- ktime_t expires = ktime_sub(hrtimer_get_expires(timer), base->offset);
|
|
|
|
-
|
|
|
|
- WARN_ON_ONCE(hrtimer_get_expires_tv64(timer) < 0);
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * If the timer is not on the current cpu, we cannot reprogram
|
|
|
|
- * the other cpus clock event device.
|
|
|
|
- */
|
|
|
|
- if (base->cpu_base != cpu_base)
|
|
|
|
- return;
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * If the hrtimer interrupt is running, then it will
|
|
|
|
- * reevaluate the clock bases and reprogram the clock event
|
|
|
|
- * device. The callbacks are always executed in hard interrupt
|
|
|
|
- * context so we don't need an extra check for a running
|
|
|
|
- * callback.
|
|
|
|
- */
|
|
|
|
- if (cpu_base->in_hrtirq)
|
|
|
|
- return;
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * CLOCK_REALTIME timer might be requested with an absolute
|
|
|
|
- * expiry time which is less than base->offset. Set it to 0.
|
|
|
|
- */
|
|
|
|
- if (expires < 0)
|
|
|
|
- expires = 0;
|
|
|
|
-
|
|
|
|
- if (expires >= cpu_base->expires_next)
|
|
|
|
- return;
|
|
|
|
-
|
|
|
|
- /* Update the pointer to the next expiring timer */
|
|
|
|
- cpu_base->next_timer = timer;
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * If a hang was detected in the last timer interrupt then we
|
|
|
|
- * do not schedule a timer which is earlier than the expiry
|
|
|
|
- * which we enforced in the hang detection. We want the system
|
|
|
|
- * to make progress.
|
|
|
|
- */
|
|
|
|
- if (cpu_base->hang_detected)
|
|
|
|
- return;
|
|
|
|
-
|
|
|
|
- /*
|
|
|
|
- * Program the timer hardware. We enforce the expiry for
|
|
|
|
- * events which are already in the past.
|
|
|
|
- */
|
|
|
|
- cpu_base->expires_next = expires;
|
|
|
|
- tick_program_event(expires, 1);
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
/*
|
|
/*
|
|
* Retrigger next event is called after clock was set
|
|
* Retrigger next event is called after clock was set
|
|
*
|
|
*
|
|
@@ -703,15 +641,72 @@ static inline int hrtimer_is_hres_enabled(void) { return 0; }
|
|
static inline void hrtimer_switch_to_hres(void) { }
|
|
static inline void hrtimer_switch_to_hres(void) { }
|
|
static inline void
|
|
static inline void
|
|
hrtimer_force_reprogram(struct hrtimer_cpu_base *base, int skip_equal) { }
|
|
hrtimer_force_reprogram(struct hrtimer_cpu_base *base, int skip_equal) { }
|
|
-static inline int hrtimer_reprogram(struct hrtimer *timer,
|
|
|
|
- struct hrtimer_clock_base *base)
|
|
|
|
-{
|
|
|
|
- return 0;
|
|
|
|
-}
|
|
|
|
static inline void retrigger_next_event(void *arg) { }
|
|
static inline void retrigger_next_event(void *arg) { }
|
|
|
|
|
|
#endif /* CONFIG_HIGH_RES_TIMERS */
|
|
#endif /* CONFIG_HIGH_RES_TIMERS */
|
|
|
|
|
|
|
|
+/*
|
|
|
|
+ * When a timer is enqueued and expires earlier than the already enqueued
|
|
|
|
+ * timers, we have to check, whether it expires earlier than the timer for
|
|
|
|
+ * which the clock event device was armed.
|
|
|
|
+ *
|
|
|
|
+ * Called with interrupts disabled and base->cpu_base.lock held
|
|
|
|
+ */
|
|
|
|
+static void hrtimer_reprogram(struct hrtimer *timer,
|
|
|
|
+ struct hrtimer_clock_base *base)
|
|
|
|
+{
|
|
|
|
+ struct hrtimer_cpu_base *cpu_base = this_cpu_ptr(&hrtimer_bases);
|
|
|
|
+ ktime_t expires = ktime_sub(hrtimer_get_expires(timer), base->offset);
|
|
|
|
+
|
|
|
|
+ WARN_ON_ONCE(hrtimer_get_expires_tv64(timer) < 0);
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * If the timer is not on the current cpu, we cannot reprogram
|
|
|
|
+ * the other cpus clock event device.
|
|
|
|
+ */
|
|
|
|
+ if (base->cpu_base != cpu_base)
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * If the hrtimer interrupt is running, then it will
|
|
|
|
+ * reevaluate the clock bases and reprogram the clock event
|
|
|
|
+ * device. The callbacks are always executed in hard interrupt
|
|
|
|
+ * context so we don't need an extra check for a running
|
|
|
|
+ * callback.
|
|
|
|
+ */
|
|
|
|
+ if (cpu_base->in_hrtirq)
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * CLOCK_REALTIME timer might be requested with an absolute
|
|
|
|
+ * expiry time which is less than base->offset. Set it to 0.
|
|
|
|
+ */
|
|
|
|
+ if (expires < 0)
|
|
|
|
+ expires = 0;
|
|
|
|
+
|
|
|
|
+ if (expires >= cpu_base->expires_next)
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ /* Update the pointer to the next expiring timer */
|
|
|
|
+ cpu_base->next_timer = timer;
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * If a hang was detected in the last timer interrupt then we
|
|
|
|
+ * do not schedule a timer which is earlier than the expiry
|
|
|
|
+ * which we enforced in the hang detection. We want the system
|
|
|
|
+ * to make progress.
|
|
|
|
+ */
|
|
|
|
+ if (cpu_base->hang_detected)
|
|
|
|
+ return;
|
|
|
|
+
|
|
|
|
+ /*
|
|
|
|
+ * Program the timer hardware. We enforce the expiry for
|
|
|
|
+ * events which are already in the past.
|
|
|
|
+ */
|
|
|
|
+ cpu_base->expires_next = expires;
|
|
|
|
+ tick_program_event(expires, 1);
|
|
|
|
+}
|
|
|
|
+
|
|
/*
|
|
/*
|
|
* Clock realtime was set
|
|
* Clock realtime was set
|
|
*
|
|
*
|