|
@@ -363,4 +363,14 @@ static inline int arch_read_trylock(arch_rwlock_t *rw)
|
|
|
#define arch_read_relax(lock) cpu_relax()
|
|
|
#define arch_write_relax(lock) cpu_relax()
|
|
|
|
|
|
+/*
|
|
|
+ * Accesses appearing in program order before a spin_lock() operation
|
|
|
+ * can be reordered with accesses inside the critical section, by virtue
|
|
|
+ * of arch_spin_lock being constructed using acquire semantics.
|
|
|
+ *
|
|
|
+ * In cases where this is problematic (e.g. try_to_wake_up), an
|
|
|
+ * smp_mb__before_spinlock() can restore the required ordering.
|
|
|
+ */
|
|
|
+#define smp_mb__before_spinlock() smp_mb()
|
|
|
+
|
|
|
#endif /* __ASM_SPINLOCK_H */
|