Lines Matching full:s64
125 static __always_inline s64 arch_atomic64_read(const atomic64_t *v) in arch_atomic64_read()
131 static __always_inline void arch_atomic64_set(atomic64_t *v, s64 i) in arch_atomic64_set()
137 static __always_inline s64 arch_atomic64_add_return(s64 i, atomic64_t *v) in arch_atomic64_add_return()
143 static __always_inline s64 arch_atomic64_fetch_add(s64 i, atomic64_t *v) in arch_atomic64_fetch_add()
149 static __always_inline void arch_atomic64_add(s64 i, atomic64_t *v) in arch_atomic64_add()
167 static __always_inline bool arch_atomic64_sub_and_test(s64 i, atomic64_t *v) in arch_atomic64_sub_and_test()
185 static __always_inline s64 arch_atomic64_xchg(atomic64_t *v, s64 new) in arch_atomic64_xchg()
191 static __always_inline s64 arch_atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) in arch_atomic64_cmpxchg()
197 static __always_inline bool arch_atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) in arch_atomic64_try_cmpxchg()
204 static __always_inline void arch_atomic64_##op(s64 i, atomic64_t *v) \
208 static __always_inline long arch_atomic64_fetch_##op(s64 i, atomic64_t *v) \
226 #define arch_atomic64_sub_return(_i, _v) arch_atomic64_add_return(-(s64)(_i), _v)
227 #define arch_atomic64_fetch_sub(_i, _v) arch_atomic64_fetch_add(-(s64)(_i), _v)
228 #define arch_atomic64_sub(_i, _v) arch_atomic64_add(-(s64)(_i), _v)