|
@@ -352,7 +352,7 @@ static __always_inline int atomic_fetch_add_unless(atomic_t *v, int a, int u)
|
|
#define atomic_fetch_add_unless atomic_fetch_add_unless
|
|
#define atomic_fetch_add_unless atomic_fetch_add_unless
|
|
|
|
|
|
#ifndef CONFIG_GENERIC_ATOMIC64
|
|
#ifndef CONFIG_GENERIC_ATOMIC64
|
|
-static __always_inline long __atomic64_add_unless(atomic64_t *v, long a, long u)
|
|
|
|
|
|
+static __always_inline long atomic64_fetch_add_unless(atomic64_t *v, long a, long u)
|
|
{
|
|
{
|
|
long prev, rc;
|
|
long prev, rc;
|
|
|
|
|
|
@@ -369,11 +369,7 @@ static __always_inline long __atomic64_add_unless(atomic64_t *v, long a, long u)
|
|
: "memory");
|
|
: "memory");
|
|
return prev;
|
|
return prev;
|
|
}
|
|
}
|
|
-
|
|
|
|
-static __always_inline int atomic64_add_unless(atomic64_t *v, long a, long u)
|
|
|
|
-{
|
|
|
|
- return __atomic64_add_unless(v, a, u) != u;
|
|
|
|
-}
|
|
|
|
|
|
+#define atomic64_fetch_add_unless atomic64_fetch_add_unless
|
|
#endif
|
|
#endif
|
|
|
|
|
|
/*
|
|
/*
|