//===-- A simple equivalent of std::atomic ----------------------*- C++ -*-===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #ifndef LLVM_LIBC_SRC___SUPPORT_CPP_ATOMIC_H #define LLVM_LIBC_SRC___SUPPORT_CPP_ATOMIC_H #include "src/__support/macros/attributes.h" #include "src/__support/macros/config.h" #include "src/__support/macros/properties/architectures.h" #include "type_traits.h" namespace LIBC_NAMESPACE_DECL { namespace cpp { enum class MemoryOrder : int { RELAXED = __ATOMIC_RELAXED, CONSUME = __ATOMIC_CONSUME, ACQUIRE = __ATOMIC_ACQUIRE, RELEASE = __ATOMIC_RELEASE, ACQ_REL = __ATOMIC_ACQ_REL, SEQ_CST = __ATOMIC_SEQ_CST }; // These are a clang extension, see the clang documenation for more information: // https://clang.llvm.org/docs/LanguageExtensions.html#scoped-atomic-builtins. enum class MemoryScope : int { #if defined(__MEMORY_SCOPE_SYSTEM) && defined(__MEMORY_SCOPE_DEVICE) SYSTEM = __MEMORY_SCOPE_SYSTEM, DEVICE = __MEMORY_SCOPE_DEVICE, #else SYSTEM = 0, DEVICE = 0, #endif }; template struct Atomic { // For now, we will restrict to only arithmetic types. static_assert(is_arithmetic_v, "Only arithmetic types can be atomic."); private: // The value stored should be appropriately aligned so that // hardware instructions used to perform atomic operations work // correctly. static constexpr int ALIGNMENT = sizeof(T) > alignof(T) ? sizeof(T) : alignof(T); public: using value_type = T; // We keep the internal value public so that it can be addressable. // This is useful in places like the Linux futex operations where // we need pointers to the memory of the atomic values. Load and store // operations should be performed using the atomic methods however. alignas(ALIGNMENT) value_type val; constexpr Atomic() = default; // Intializes the value without using atomic operations. constexpr Atomic(value_type v) : val(v) {} Atomic(const Atomic &) = delete; Atomic &operator=(const Atomic &) = delete; // Atomic load. operator T() { return __atomic_load_n(&val, int(MemoryOrder::SEQ_CST)); } T load(MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_load_n) return __scoped_atomic_load_n(&val, int(mem_ord), (int)(mem_scope)); #else return __atomic_load_n(&val, int(mem_ord)); #endif } // Atomic store. T operator=(T rhs) { __atomic_store_n(&val, rhs, int(MemoryOrder::SEQ_CST)); return rhs; } void store(T rhs, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_store_n) __scoped_atomic_store_n(&val, rhs, int(mem_ord), (int)(mem_scope)); #else __atomic_store_n(&val, rhs, int(mem_ord)); #endif } // Atomic compare exchange bool compare_exchange_strong( T &expected, T desired, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { return __atomic_compare_exchange_n(&val, &expected, desired, false, int(mem_ord), int(mem_ord)); } // Atomic compare exchange (separate success and failure memory orders) bool compare_exchange_strong( T &expected, T desired, MemoryOrder success_order, MemoryOrder failure_order, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { return __atomic_compare_exchange_n(&val, &expected, desired, false, static_cast(success_order), static_cast(failure_order)); } // Atomic compare exchange (weak version) bool compare_exchange_weak( T &expected, T desired, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { return __atomic_compare_exchange_n(&val, &expected, desired, true, static_cast(mem_ord), static_cast(mem_ord)); } // Atomic compare exchange (weak version with separate success and failure // memory orders) bool compare_exchange_weak( T &expected, T desired, MemoryOrder success_order, MemoryOrder failure_order, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { return __atomic_compare_exchange_n(&val, &expected, desired, true, static_cast(success_order), static_cast(failure_order)); } T exchange(T desired, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_exchange_n) return __scoped_atomic_exchange_n(&val, desired, int(mem_ord), (int)(mem_scope)); #else return __atomic_exchange_n(&val, desired, int(mem_ord)); #endif } T fetch_add(T increment, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_fetch_add) return __scoped_atomic_fetch_add(&val, increment, int(mem_ord), (int)(mem_scope)); #else return __atomic_fetch_add(&val, increment, int(mem_ord)); #endif } T fetch_or(T mask, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_fetch_or) return __scoped_atomic_fetch_or(&val, mask, int(mem_ord), (int)(mem_scope)); #else return __atomic_fetch_or(&val, mask, int(mem_ord)); #endif } T fetch_and(T mask, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_fetch_and) return __scoped_atomic_fetch_and(&val, mask, int(mem_ord), (int)(mem_scope)); #else return __atomic_fetch_and(&val, mask, int(mem_ord)); #endif } T fetch_sub(T decrement, MemoryOrder mem_ord = MemoryOrder::SEQ_CST, [[maybe_unused]] MemoryScope mem_scope = MemoryScope::DEVICE) { #if __has_builtin(__scoped_atomic_fetch_sub) return __scoped_atomic_fetch_sub(&val, decrement, int(mem_ord), (int)(mem_scope)); #else return __atomic_fetch_sub(&val, decrement, int(mem_ord)); #endif } // Set the value without using an atomic operation. This is useful // in initializing atomic values without a constructor. void set(T rhs) { val = rhs; } }; // Issue a thread fence with the given memory ordering. LIBC_INLINE void atomic_thread_fence([[maybe_unused]] MemoryOrder mem_ord) { // The NVPTX backend currently does not support atomic thread fences so we use a // full system fence instead. #ifdef LIBC_TARGET_ARCH_IS_NVPTX __nvvm_membar_sys(); #else __atomic_thread_fence(static_cast(mem_ord)); #endif } // Establishes memory synchronization ordering of non-atomic and relaxed atomic // accesses, as instructed by order, between a thread and a signal handler // executed on the same thread. This is equivalent to atomic_thread_fence, // except no instructions for memory ordering are issued. Only reordering of // the instructions by the compiler is suppressed as order instructs. LIBC_INLINE void atomic_signal_fence([[maybe_unused]] MemoryOrder mem_ord) { #if __has_builtin(__atomic_signal_fence) __atomic_signal_fence(static_cast(mem_ord)); #else // if the builtin is not ready, use asm as a full compiler barrier. asm volatile("" ::: "memory"); #endif } } // namespace cpp } // namespace LIBC_NAMESPACE_DECL #endif // LLVM_LIBC_SRC___SUPPORT_CPP_ATOMIC_H