|
@@ -87,7 +87,6 @@ static inline void arch_spin_unlock(arch_spinlock_t *lp)
|
|
|
{
|
|
|
typecheck(unsigned int, lp->lock);
|
|
|
asm volatile(
|
|
|
- __ASM_BARRIER
|
|
|
"st %1,%0\n"
|
|
|
: "+Q" (lp->lock)
|
|
|
: "d" (0)
|
|
@@ -169,7 +168,6 @@ static inline int arch_write_trylock_once(arch_rwlock_t *rw)
|
|
|
\
|
|
|
typecheck(unsigned int *, ptr); \
|
|
|
asm volatile( \
|
|
|
- "bcr 14,0\n" \
|
|
|
op_string " %0,%2,%1\n" \
|
|
|
: "=d" (old_val), "+Q" (*ptr) \
|
|
|
: "d" (op_val) \
|
|
@@ -243,7 +241,6 @@ static inline void arch_write_unlock(arch_rwlock_t *rw)
|
|
|
|
|
|
rw->owner = 0;
|
|
|
asm volatile(
|
|
|
- __ASM_BARRIER
|
|
|
"st %1,%0\n"
|
|
|
: "+Q" (rw->lock)
|
|
|
: "d" (0)
|