|
@@ -37,25 +37,6 @@ do { __asm__ __volatile__("ba,pt %%xcc, 1f\n\t" \
|
|
|
#define rmb() __asm__ __volatile__("":::"memory")
|
|
|
#define wmb() __asm__ __volatile__("":::"memory")
|
|
|
|
|
|
-#define dma_rmb() rmb()
|
|
|
-#define dma_wmb() wmb()
|
|
|
-
|
|
|
-#define smp_store_mb(__var, __value) \
|
|
|
- do { WRITE_ONCE(__var, __value); membar_safe("#StoreLoad"); } while(0)
|
|
|
-
|
|
|
-#ifdef CONFIG_SMP
|
|
|
-#define smp_mb() mb()
|
|
|
-#define smp_rmb() rmb()
|
|
|
-#define smp_wmb() wmb()
|
|
|
-#else
|
|
|
-#define smp_mb() __asm__ __volatile__("":::"memory")
|
|
|
-#define smp_rmb() __asm__ __volatile__("":::"memory")
|
|
|
-#define smp_wmb() __asm__ __volatile__("":::"memory")
|
|
|
-#endif
|
|
|
-
|
|
|
-#define read_barrier_depends() do { } while (0)
|
|
|
-#define smp_read_barrier_depends() do { } while (0)
|
|
|
-
|
|
|
#define smp_store_release(p, v) \
|
|
|
do { \
|
|
|
compiletime_assert_atomic_type(*p); \
|
|
@@ -74,4 +55,6 @@ do { \
|
|
|
#define smp_mb__before_atomic() barrier()
|
|
|
#define smp_mb__after_atomic() barrier()
|
|
|
|
|
|
+#include <asm-generic/barrier.h>
|
|
|
+
|
|
|
#endif /* !(__SPARC64_BARRIER_H) */
|