|
|
@@ -75,7 +75,7 @@ extern void __add_wrong_size(void)
|
|
|
* use "asm volatile" and "memory" clobbers to prevent gcc from moving
|
|
|
* information around.
|
|
|
*/
|
|
|
-#define xchg(ptr, v) __xchg_op((ptr), (v), xchg, "")
|
|
|
+#define arch_xchg(ptr, v) __xchg_op((ptr), (v), xchg, "")
|
|
|
|
|
|
/*
|
|
|
* Atomic compare and exchange. Compare OLD with MEM, if identical,
|