@@ -4,7 +4,6 @@
#include <linux/percpu.h>
#include <linux/atomic.h>
#include <asm/types.h>
-#include <asm-generic/local_wrap.h>
/*
* A signed long type for operations which are atomic for a single CPU.
@@ -48,23 +47,6 @@ typedef struct
#define local_add_unless(l, _a, u) atomic_long_add_unless((&(l)->a), (_a), (u))
#define local_inc_not_zero(l) atomic_long_inc_not_zero(&(l)->a)
-#define local_read_wrap(l) atomic_long_read_wrap(&(l)->a)
-#define local_set_wrap(l,i) atomic_long_set_wrap((&(l)->a),(i))
-#define local_inc_wrap(l) atomic_long_inc_wrap(&(l)->a)
-#define local_inc_return_wrap(l) atomic_long_return_wrap(&(l)->a)
-#define local_inc_and_test_wrap(l) atomic_long_inc_and_test_wrap(&(l)->a)
-#define local_dec_wrap(l) atomic_long_dec_wrap(&(l)->a)
-#define local_dec_return_wrap(l) atomic_long_dec_return_wrap(&(l)->a)
-#define local_dec_and_test_wrap(l) atomic_long_dec_and_test_wrap(&(l)->a)
-#define local_add_wrap(i,l) atomic_long_add_wrap((i),(&(l)->a))
-#define local_add_return_wrap(i, l) atomic_long_add_return_wrap((i), (&(l)->a))
-#define local_sub_wrap(i,l) atomic_long_sub_wrap((i),(&(l)->a))
-#define local_sub_return_wrap(i, l) atomic_long_sub_return_wrap((i), (&(l)->a))
-#define local_sub_and_test_wrap(i, l) atomic_long_sub_and_test_wrap((i), (&(l)->a))
-#define local_cmpxchg_wrap(l, o, n) atomic_long_cmpxchg_wrap((&(l)->a), (o), (n))
-#define local_add_unless_wrap(l, _a, u) atomic_long_add_unless_wrap((&(l)->a), (_a), (u))
-#define local_add_negative_wrap(i, l) atomic_long_add_negative_wrap((i), (&(l)->a))
-
/* Non-atomic variants, ie. preemption disabled and won't be touched
* in interrupt, etc. Some archs can optimize this case well. */
#define __local_inc(l) local_set((l), local_read(l) + 1)
@@ -72,4 +54,6 @@ typedef struct
#define __local_add(i,l) local_set((l), local_read(l) + (i))
#define __local_sub(i,l) local_set((l), local_read(l) - (i))
+#include <asm-generic/local_wrap.h>
+
#endif /* _ASM_GENERIC_LOCAL_H */