Browse Source

sched/clock, x86/tsc: Rework the x86 'unstable' sched_clock() interface

Wanpeng Li reported that since the following commit:

  acb04058de49 ("sched/clock: Fix hotplug crash")

... KVM always runs with unstable sched-clock even though KVM's
kvm_clock _is_ stable.

The problem is that we've tied clear_sched_clock_stable() to the TSC
state, and overlooked that sched_clock() is a paravirt function.

Solve this by doing two things:

 - tie the sched_clock() stable state more clearly to the TSC stable
   state for the normal (!paravirt) case.

 - only call clear_sched_clock_stable() when we mark TSC unstable
   when we use native_sched_clock().

The first means we can actually run with stable sched_clock in more
situations then before, which is good. And since commit:

  12907fbb1a69 ("sched/clock, clocksource: Add optional cs::mark_unstable() method")

... this should be reliable. Since any detection of TSC fail now results
in marking the TSC unstable.

Reported-by: Wanpeng Li <kernellwp@gmail.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Paolo Bonzini <pbonzini@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-kernel@vger.kernel.org
Fixes: acb04058de49 ("sched/clock: Fix hotplug crash")
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Peter Zijlstra 8 years ago
parent
commit
f94c8d1169

+ 0 - 4
arch/x86/kernel/cpu/amd.c

@@ -555,10 +555,6 @@ static void early_init_amd(struct cpuinfo_x86 *c)
 	if (c->x86_power & (1 << 8)) {
 	if (c->x86_power & (1 << 8)) {
 		set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
 		set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
 		set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
 		set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
-		if (check_tsc_unstable())
-			clear_sched_clock_stable();
-	} else {
-		clear_sched_clock_stable();
 	}
 	}
 
 
 	/* Bit 12 of 8000_0007 edx is accumulated power mechanism. */
 	/* Bit 12 of 8000_0007 edx is accumulated power mechanism. */

+ 0 - 2
arch/x86/kernel/cpu/centaur.c

@@ -104,8 +104,6 @@ static void early_init_centaur(struct cpuinfo_x86 *c)
 #ifdef CONFIG_X86_64
 #ifdef CONFIG_X86_64
 	set_cpu_cap(c, X86_FEATURE_SYSENTER32);
 	set_cpu_cap(c, X86_FEATURE_SYSENTER32);
 #endif
 #endif
-
-	clear_sched_clock_stable();
 }
 }
 
 
 static void init_centaur(struct cpuinfo_x86 *c)
 static void init_centaur(struct cpuinfo_x86 *c)

+ 0 - 3
arch/x86/kernel/cpu/common.c

@@ -86,7 +86,6 @@ static void default_init(struct cpuinfo_x86 *c)
 			strcpy(c->x86_model_id, "386");
 			strcpy(c->x86_model_id, "386");
 	}
 	}
 #endif
 #endif
-	clear_sched_clock_stable();
 }
 }
 
 
 static const struct cpu_dev default_cpu = {
 static const struct cpu_dev default_cpu = {
@@ -1075,8 +1074,6 @@ static void identify_cpu(struct cpuinfo_x86 *c)
 	 */
 	 */
 	if (this_cpu->c_init)
 	if (this_cpu->c_init)
 		this_cpu->c_init(c);
 		this_cpu->c_init(c);
-	else
-		clear_sched_clock_stable();
 
 
 	/* Disable the PN if appropriate */
 	/* Disable the PN if appropriate */
 	squash_the_stupid_serial_number(c);
 	squash_the_stupid_serial_number(c);

+ 0 - 1
arch/x86/kernel/cpu/cyrix.c

@@ -184,7 +184,6 @@ static void early_init_cyrix(struct cpuinfo_x86 *c)
 		set_cpu_cap(c, X86_FEATURE_CYRIX_ARR);
 		set_cpu_cap(c, X86_FEATURE_CYRIX_ARR);
 		break;
 		break;
 	}
 	}
-	clear_sched_clock_stable();
 }
 }
 
 
 static void init_cyrix(struct cpuinfo_x86 *c)
 static void init_cyrix(struct cpuinfo_x86 *c)

+ 0 - 4
arch/x86/kernel/cpu/intel.c

@@ -161,10 +161,6 @@ static void early_init_intel(struct cpuinfo_x86 *c)
 	if (c->x86_power & (1 << 8)) {
 	if (c->x86_power & (1 << 8)) {
 		set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
 		set_cpu_cap(c, X86_FEATURE_CONSTANT_TSC);
 		set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
 		set_cpu_cap(c, X86_FEATURE_NONSTOP_TSC);
-		if (check_tsc_unstable())
-			clear_sched_clock_stable();
-	} else {
-		clear_sched_clock_stable();
 	}
 	}
 
 
 	/* Penwell and Cloverview have the TSC which doesn't sleep on S3 */
 	/* Penwell and Cloverview have the TSC which doesn't sleep on S3 */

+ 0 - 2
arch/x86/kernel/cpu/transmeta.c

@@ -15,8 +15,6 @@ static void early_init_transmeta(struct cpuinfo_x86 *c)
 		if (xlvl >= 0x80860001)
 		if (xlvl >= 0x80860001)
 			c->x86_capability[CPUID_8086_0001_EDX] = cpuid_edx(0x80860001);
 			c->x86_capability[CPUID_8086_0001_EDX] = cpuid_edx(0x80860001);
 	}
 	}
-
-	clear_sched_clock_stable();
 }
 }
 
 
 static void init_transmeta(struct cpuinfo_x86 *c)
 static void init_transmeta(struct cpuinfo_x86 *c)

+ 23 - 12
arch/x86/kernel/tsc.c

@@ -326,9 +326,16 @@ unsigned long long sched_clock(void)
 {
 {
 	return paravirt_sched_clock();
 	return paravirt_sched_clock();
 }
 }
+
+static inline bool using_native_sched_clock(void)
+{
+	return pv_time_ops.sched_clock == native_sched_clock;
+}
 #else
 #else
 unsigned long long
 unsigned long long
 sched_clock(void) __attribute__((alias("native_sched_clock")));
 sched_clock(void) __attribute__((alias("native_sched_clock")));
+
+static inline bool using_native_sched_clock(void) { return true; }
 #endif
 #endif
 
 
 int check_tsc_unstable(void)
 int check_tsc_unstable(void)
@@ -1111,8 +1118,10 @@ static void tsc_cs_mark_unstable(struct clocksource *cs)
 {
 {
 	if (tsc_unstable)
 	if (tsc_unstable)
 		return;
 		return;
+
 	tsc_unstable = 1;
 	tsc_unstable = 1;
-	clear_sched_clock_stable();
+	if (using_native_sched_clock())
+		clear_sched_clock_stable();
 	disable_sched_clock_irqtime();
 	disable_sched_clock_irqtime();
 	pr_info("Marking TSC unstable due to clocksource watchdog\n");
 	pr_info("Marking TSC unstable due to clocksource watchdog\n");
 }
 }
@@ -1134,18 +1143,20 @@ static struct clocksource clocksource_tsc = {
 
 
 void mark_tsc_unstable(char *reason)
 void mark_tsc_unstable(char *reason)
 {
 {
-	if (!tsc_unstable) {
-		tsc_unstable = 1;
+	if (tsc_unstable)
+		return;
+
+	tsc_unstable = 1;
+	if (using_native_sched_clock())
 		clear_sched_clock_stable();
 		clear_sched_clock_stable();
-		disable_sched_clock_irqtime();
-		pr_info("Marking TSC unstable due to %s\n", reason);
-		/* Change only the rating, when not registered */
-		if (clocksource_tsc.mult)
-			clocksource_mark_unstable(&clocksource_tsc);
-		else {
-			clocksource_tsc.flags |= CLOCK_SOURCE_UNSTABLE;
-			clocksource_tsc.rating = 0;
-		}
+	disable_sched_clock_irqtime();
+	pr_info("Marking TSC unstable due to %s\n", reason);
+	/* Change only the rating, when not registered */
+	if (clocksource_tsc.mult) {
+		clocksource_mark_unstable(&clocksource_tsc);
+	} else {
+		clocksource_tsc.flags |= CLOCK_SOURCE_UNSTABLE;
+		clocksource_tsc.rating = 0;
 	}
 	}
 }
 }