summaryrefslogtreecommitdiff
path: root/c_src/include/linux/atomic.h
diff options
context:
space:
mode:
authorKent Overstreet <kent.overstreet@linux.dev>2024-01-16 17:00:02 -0500
committerKent Overstreet <kent.overstreet@linux.dev>2024-01-16 17:17:23 -0500
commitb5fd066153c40a70a29caa1ea7987723ab687763 (patch)
tree6d43a8b0a90d549a54c65565ac96c92b3e84b594 /c_src/include/linux/atomic.h
parent06ff8b55b70fda44d91b31b5511fafd1680a8934 (diff)
Move c_src dirs back to toplevel
We just wanted c sourcefiles out of the top level, not c source directories. Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
Diffstat (limited to 'c_src/include/linux/atomic.h')
-rw-r--r--c_src/include/linux/atomic.h349
1 files changed, 0 insertions, 349 deletions
diff --git a/c_src/include/linux/atomic.h b/c_src/include/linux/atomic.h
deleted file mode 100644
index 5313f850..00000000
--- a/c_src/include/linux/atomic.h
+++ /dev/null
@@ -1,349 +0,0 @@
-#ifndef __TOOLS_LINUX_ATOMIC_H
-#define __TOOLS_LINUX_ATOMIC_H
-
-#include <linux/compiler.h>
-#include <linux/types.h>
-
-typedef struct {
- int counter;
-} atomic_t;
-
-typedef struct {
- long counter;
-} atomic_long_t;
-
-typedef struct {
- u64 counter;
-} atomic64_t;
-
-#ifndef C11_ATOMICS
-
-#include <urcu/uatomic.h>
-
-#if (CAA_BITS_PER_LONG != 64)
-#define ATOMIC64_SPINLOCK
-#endif
-
-#define __ATOMIC_READ(p) uatomic_read(p)
-#define __ATOMIC_SET(p, v) uatomic_set(p, v)
-#define __ATOMIC_ADD_RETURN(v, p) uatomic_add_return(p, v)
-#define __ATOMIC_SUB_RETURN(v, p) uatomic_sub_return(p, v)
-#define __ATOMIC_ADD(v, p) uatomic_add(p, v)
-#define __ATOMIC_SUB(v, p) uatomic_sub(p, v)
-#define __ATOMIC_INC(p) uatomic_inc(p)
-#define __ATOMIC_DEC(p) uatomic_dec(p)
-#define __ATOMIC_AND(v, p) uatomic_and(p, v)
-#define __ATOMIC_OR(v, p) uatomic_or(p, v)
-
-#define xchg(p, v) uatomic_xchg(p, v)
-#define xchg_acquire(p, v) uatomic_xchg(p, v)
-#define cmpxchg(p, old, new) uatomic_cmpxchg(p, old, new)
-#define cmpxchg_acquire(p, old, new) uatomic_cmpxchg(p, old, new)
-#define cmpxchg_release(p, old, new) uatomic_cmpxchg(p, old, new)
-
-#define smp_mb__before_atomic() cmm_smp_mb__before_uatomic_add()
-#define smp_mb__after_atomic() cmm_smp_mb__after_uatomic_add()
-#define smp_wmb() cmm_smp_wmb()
-#define smp_rmb() cmm_smp_rmb()
-#define smp_mb() cmm_smp_mb()
-#define smp_read_barrier_depends() cmm_smp_read_barrier_depends()
-#define smp_acquire__after_ctrl_dep() cmm_smp_mb()
-
-#else /* C11_ATOMICS */
-
-#define __ATOMIC_READ(p) __atomic_load_n(p, __ATOMIC_RELAXED)
-#define __ATOMIC_SET(p, v) __atomic_store_n(p, v, __ATOMIC_RELAXED)
-#define __ATOMIC_ADD_RETURN(v, p) __atomic_add_fetch(p, v, __ATOMIC_RELAXED)
-#define __ATOMIC_ADD_RETURN_RELEASE(v, p) \
- __atomic_add_fetch(p, v, __ATOMIC_RELEASE)
-#define __ATOMIC_SUB_RETURN(v, p) __atomic_sub_fetch(p, v, __ATOMIC_RELAXED)
-#define __ATOMIC_SUB_RETURN_RELEASE(v, p) \
- __atomic_sub_fetch(p, v, __ATOMIC_RELEASE)
-#define __ATOMIC_AND(p) __atomic_and_fetch(p, v, __ATOMIC_RELAXED)
-#define __ATOMIC_OR(p) __atomic_or_fetch(p, v, __ATOMIC_RELAXED)
-
-#define xchg(p, v) __atomic_exchange_n(p, v, __ATOMIC_SEQ_CST)
-#define xchg_acquire(p, v) __atomic_exchange_n(p, v, __ATOMIC_ACQUIRE)
-
-#define cmpxchg(p, old, new) \
-({ \
- typeof(*(p)) __old = (old); \
- \
- __atomic_compare_exchange_n((p), &__old, new, false, \
- __ATOMIC_SEQ_CST, \
- __ATOMIC_SEQ_CST); \
- __old; \
-})
-
-#define cmpxchg_acquire(p, old, new) \
-({ \
- typeof(*(p)) __old = (old); \
- \
- __atomic_compare_exchange_n((p), &__old, new, false, \
- __ATOMIC_ACQUIRE, \
- __ATOMIC_ACQUIRE); \
- __old; \
-})
-
-#define cmpxchg_release(p, old, new) \
-({ \
- typeof(*(p)) __old = (old); \
- \
- __atomic_compare_exchange_n((p), &__old, new, false, \
- __ATOMIC_RELEASE, \
- __ATOMIC_RELEASE); \
- __old; \
-})
-
-#define smp_mb__before_atomic() __atomic_thread_fence(__ATOMIC_SEQ_CST)
-#define smp_mb__after_atomic() __atomic_thread_fence(__ATOMIC_SEQ_CST)
-#define smp_wmb() __atomic_thread_fence(__ATOMIC_SEQ_CST)
-#define smp_rmb() __atomic_thread_fence(__ATOMIC_SEQ_CST)
-#define smp_mb() __atomic_thread_fence(__ATOMIC_SEQ_CST)
-#define smp_read_barrier_depends()
-
-#endif
-
-#define smp_store_mb(var, value) do { WRITE_ONCE(var, value); smp_mb(); } while (0)
-
-#define smp_load_acquire(p) \
-({ \
- typeof(*p) ___p1 = READ_ONCE(*p); \
- smp_mb(); \
- ___p1; \
-})
-
-#define smp_store_release(p, v) \
-do { \
- smp_mb(); \
- WRITE_ONCE(*p, v); \
-} while (0)
-
-/* atomic interface: */
-
-#ifndef __ATOMIC_ADD
-#define __ATOMIC_ADD(i, v) __ATOMIC_ADD_RETURN(i, v)
-#endif
-
-#ifndef __ATOMIC_ADD_RETURN_RELEASE
-#define __ATOMIC_ADD_RETURN_RELEASE(i, v) \
- ({ smp_mb__before_atomic(); __ATOMIC_ADD_RETURN(i, v); })
-#endif
-
-#ifndef __ATOMIC_SUB_RETURN_RELEASE
-#define __ATOMIC_SUB_RETURN_RELEASE(i, v) \
- ({ smp_mb__before_atomic(); __ATOMIC_SUB_RETURN(i, v); })
-#endif
-
-#ifndef __ATOMIC_SUB
-#define __ATOMIC_SUB(i, v) __ATOMIC_SUB_RETURN(i, v)
-#endif
-
-#ifndef __ATOMIC_INC_RETURN
-#define __ATOMIC_INC_RETURN(v) __ATOMIC_ADD_RETURN(1, v)
-#endif
-
-#ifndef __ATOMIC_DEC_RETURN
-#define __ATOMIC_DEC_RETURN(v) __ATOMIC_SUB_RETURN(1, v)
-#endif
-
-#ifndef __ATOMIC_INC
-#define __ATOMIC_INC(v) __ATOMIC_ADD(1, v)
-#endif
-
-#ifndef __ATOMIC_DEC
-#define __ATOMIC_DEC(v) __ATOMIC_SUB(1, v)
-#endif
-
-#define DEF_ATOMIC_OPS(a_type, i_type) \
-static inline i_type a_type##_read(const a_type##_t *v) \
-{ \
- return __ATOMIC_READ(&v->counter); \
-} \
- \
-static inline i_type a_type##_read_acquire(const a_type##_t *v) \
-{ \
- i_type ret = __ATOMIC_READ(&v->counter); \
- smp_mb__after_atomic(); \
- return ret; \
-} \
- \
-static inline void a_type##_set(a_type##_t *v, i_type i) \
-{ \
- return __ATOMIC_SET(&v->counter, i); \
-} \
- \
-static inline i_type a_type##_add_return(i_type i, a_type##_t *v) \
-{ \
- return __ATOMIC_ADD_RETURN(i, &v->counter); \
-} \
- \
-static inline i_type a_type##_add_return_release(i_type i, a_type##_t *v)\
-{ \
- return __ATOMIC_ADD_RETURN_RELEASE(i, &v->counter); \
-} \
- \
-static inline i_type a_type##_sub_return_release(i_type i, a_type##_t *v)\
-{ \
- return __ATOMIC_SUB_RETURN_RELEASE(i, &v->counter); \
-} \
- \
-static inline i_type a_type##_sub_return(i_type i, a_type##_t *v) \
-{ \
- return __ATOMIC_SUB_RETURN(i, &v->counter); \
-} \
- \
-static inline void a_type##_add(i_type i, a_type##_t *v) \
-{ \
- __ATOMIC_ADD(i, &v->counter); \
-} \
- \
-static inline void a_type##_sub(i_type i, a_type##_t *v) \
-{ \
- __ATOMIC_SUB(i, &v->counter); \
-} \
- \
-static inline i_type a_type##_inc_return(a_type##_t *v) \
-{ \
- return __ATOMIC_INC_RETURN(&v->counter); \
-} \
- \
-static inline i_type a_type##_dec_return(a_type##_t *v) \
-{ \
- return __ATOMIC_DEC_RETURN(&v->counter); \
-} \
- \
-static inline i_type a_type##_dec_return_release(a_type##_t *v) \
-{ \
- return __ATOMIC_SUB_RETURN_RELEASE(1, &v->counter); \
-} \
- \
-static inline void a_type##_inc(a_type##_t *v) \
-{ \
- __ATOMIC_INC(&v->counter); \
-} \
- \
-static inline void a_type##_dec(a_type##_t *v) \
-{ \
- __ATOMIC_DEC(&v->counter); \
-} \
- \
-static inline bool a_type##_add_negative(i_type i, a_type##_t *v) \
-{ \
- return __ATOMIC_ADD_RETURN(i, &v->counter) < 0; \
-} \
- \
-static inline bool a_type##_sub_and_test(i_type i, a_type##_t *v) \
-{ \
- return __ATOMIC_SUB_RETURN(i, &v->counter) == 0; \
-} \
- \
-static inline bool a_type##_inc_and_test(a_type##_t *v) \
-{ \
- return __ATOMIC_INC_RETURN(&v->counter) == 0; \
-} \
- \
-static inline bool a_type##_dec_and_test(a_type##_t *v) \
-{ \
- return __ATOMIC_DEC_RETURN(&v->counter) == 0; \
-} \
- \
-static inline i_type a_type##_add_unless(a_type##_t *v, i_type a, i_type u)\
-{ \
- i_type old, c = __ATOMIC_READ(&v->counter); \
- while (c != u && (old = cmpxchg(&v->counter, c, c + a)) != c) \
- c = old; \
- return c; \
-} \
- \
-static inline bool a_type##_inc_not_zero(a_type##_t *v) \
-{ \
- return a_type##_add_unless(v, 1, 0); \
-} \
- \
-static inline void a_type##_and(i_type a, a_type##_t *v) \
-{ \
- __ATOMIC_AND(a, v); \
-} \
- \
-static inline void a_type##_or(i_type a, a_type##_t *v) \
-{ \
- __ATOMIC_OR(a, v); \
-} \
- \
-static inline i_type a_type##_xchg(a_type##_t *v, i_type i) \
-{ \
- return xchg(&v->counter, i); \
-} \
- \
-static inline i_type a_type##_cmpxchg(a_type##_t *v, i_type old, i_type new)\
-{ \
- return cmpxchg(&v->counter, old, new); \
-} \
- \
-static inline i_type a_type##_cmpxchg_acquire(a_type##_t *v, i_type old, i_type new)\
-{ \
- return cmpxchg_acquire(&v->counter, old, new); \
-} \
- \
-static inline bool a_type##_try_cmpxchg_acquire(a_type##_t *v, i_type *old, i_type new)\
-{ \
- i_type prev = *old; \
- *old = cmpxchg_acquire(&v->counter, *old, new); \
- return prev == *old; \
-}
-
-DEF_ATOMIC_OPS(atomic, int)
-DEF_ATOMIC_OPS(atomic_long, long)
-
-#ifndef ATOMIC64_SPINLOCK
-DEF_ATOMIC_OPS(atomic64, s64)
-#else
-s64 atomic64_read(const atomic64_t *v);
-static inline s64 atomic64_read_acquire(const atomic64_t *v)
-{
- s64 ret = atomic64_read(v);
- smp_mb__after_atomic();
- return ret;
-}
-
-void atomic64_set(atomic64_t *v, s64);
-
-s64 atomic64_add_return(s64, atomic64_t *);
-s64 atomic64_sub_return(s64, atomic64_t *);
-void atomic64_add(s64, atomic64_t *);
-void atomic64_sub(s64, atomic64_t *);
-
-s64 atomic64_xchg(atomic64_t *, s64);
-s64 atomic64_cmpxchg(atomic64_t *, s64, s64);
-
-#define atomic64_add_negative(a, v) (atomic64_add_return((a), (v)) < 0)
-#define atomic64_inc(v) atomic64_add(1LL, (v))
-#define atomic64_inc_return(v) atomic64_add_return(1LL, (v))
-#define atomic64_inc_and_test(v) (atomic64_inc_return(v) == 0)
-#define atomic64_sub_and_test(a, v) (atomic64_sub_return((a), (v)) == 0)
-#define atomic64_dec(v) atomic64_sub(1LL, (v))
-#define atomic64_dec_return(v) atomic64_sub_return(1LL, (v))
-#define atomic64_dec_and_test(v) (atomic64_dec_return((v)) == 0)
-#define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1LL, 0LL)
-
-static inline s64 atomic64_add_return_release(s64 i, atomic64_t *v)
-{
- smp_mb__before_atomic();
- return atomic64_add_return(i, v);
-}
-
-static inline s64 atomic64_cmpxchg_acquire(atomic64_t *v, s64 old, s64 new)
-{
- return atomic64_cmpxchg(v, old, new);
-}
-
-static inline s64 atomic64_sub_return_release(s64 i, atomic64_t *v)
-{
- smp_mb__before_atomic();
- return atomic64_sub_return(i, v);
-}
-
-#endif
-
-#endif /* __TOOLS_LINUX_ATOMIC_H */