|
@@ -259,10 +259,10 @@ static __inline__ int atomic_add_unless(atomic_t *v, int a, int u)
|
|
|
|
|
|
|
|
#define ATOMIC64_INIT(i) ((atomic64_t) { (i) })
|
|
#define ATOMIC64_INIT(i) ((atomic64_t) { (i) })
|
|
|
|
|
|
|
|
-static __inline__ int
|
|
|
|
|
|
|
+static __inline__ s64
|
|
|
__atomic64_add_return(s64 i, atomic64_t *v)
|
|
__atomic64_add_return(s64 i, atomic64_t *v)
|
|
|
{
|
|
{
|
|
|
- int ret;
|
|
|
|
|
|
|
+ s64 ret;
|
|
|
unsigned long flags;
|
|
unsigned long flags;
|
|
|
_atomic_spin_lock_irqsave(v, flags);
|
|
_atomic_spin_lock_irqsave(v, flags);
|
|
|
|
|
|