aboutsummaryrefslogtreecommitdiff
path: root/system/include/libcxx/atomic
diff options
context:
space:
mode:
Diffstat (limited to 'system/include/libcxx/atomic')
-rw-r--r--system/include/libcxx/atomic750
1 files changed, 379 insertions, 371 deletions
diff --git a/system/include/libcxx/atomic b/system/include/libcxx/atomic
index 31385a23..db67e762 100644
--- a/system/include/libcxx/atomic
+++ b/system/include/libcxx/atomic
@@ -29,10 +29,11 @@ typedef enum memory_order
memory_order_seq_cst // store-release load-acquire
} memory_order;
-template <class T> T kill_dependency(T y);
+template <class T> T kill_dependency(T y) noexcept;
// lock-free property
+#define ATOMIC_BOOL_LOCK_FREE unspecified
#define ATOMIC_CHAR_LOCK_FREE unspecified
#define ATOMIC_CHAR16_T_LOCK_FREE unspecified
#define ATOMIC_CHAR32_T_LOCK_FREE unspecified
@@ -41,45 +42,46 @@ template <class T> T kill_dependency(T y);
#define ATOMIC_INT_LOCK_FREE unspecified
#define ATOMIC_LONG_LOCK_FREE unspecified
#define ATOMIC_LLONG_LOCK_FREE unspecified
+#define ATOMIC_POINTER_LOCK_FREE unspecified
// flag type and operations
typedef struct atomic_flag
{
- bool test_and_set(memory_order m = memory_order_seq_cst) volatile;
- bool test_and_set(memory_order m = memory_order_seq_cst);
- void clear(memory_order m = memory_order_seq_cst) volatile;
- void clear(memory_order m = memory_order_seq_cst);
- atomic_flag() = default;
+ bool test_and_set(memory_order m = memory_order_seq_cst) volatile noexcept;
+ bool test_and_set(memory_order m = memory_order_seq_cst) noexcept;
+ void clear(memory_order m = memory_order_seq_cst) volatile noexcept;
+ void clear(memory_order m = memory_order_seq_cst) noexcept;
+ atomic_flag() noexcept = default;
atomic_flag(const atomic_flag&) = delete;
atomic_flag& operator=(const atomic_flag&) = delete;
atomic_flag& operator=(const atomic_flag&) volatile = delete;
} atomic_flag;
bool
- atomic_flag_test_and_set(volatile atomic_flag* obj);
+ atomic_flag_test_and_set(volatile atomic_flag* obj) noexcept;
bool
- atomic_flag_test_and_set(atomic_flag* obj);
+ atomic_flag_test_and_set(atomic_flag* obj) noexcept;
bool
atomic_flag_test_and_set_explicit(volatile atomic_flag* obj,
- memory_order m);
+ memory_order m) noexcept;
bool
- atomic_flag_test_and_set_explicit(atomic_flag* obj, memory_order m);
+ atomic_flag_test_and_set_explicit(atomic_flag* obj, memory_order m) noexcept;
void
- atomic_flag_clear(volatile atomic_flag* obj);
+ atomic_flag_clear(volatile atomic_flag* obj) noexcept;
void
- atomic_flag_clear(atomic_flag* obj);
+ atomic_flag_clear(atomic_flag* obj) noexcept;
void
- atomic_flag_clear_explicit(volatile atomic_flag* obj, memory_order m);
+ atomic_flag_clear_explicit(volatile atomic_flag* obj, memory_order m) noexcept;
void
- atomic_flag_clear_explicit(atomic_flag* obj, memory_order m);
+ atomic_flag_clear_explicit(atomic_flag* obj, memory_order m) noexcept;
#define ATOMIC_FLAG_INIT see below
#define ATOMIC_VAR_INIT(value) see below
@@ -87,391 +89,392 @@ void
template <class T>
struct atomic
{
- bool is_lock_free() const volatile;
- bool is_lock_free() const;
- void store(T desr, memory_order m = memory_order_seq_cst) volatile;
- void store(T desr, memory_order m = memory_order_seq_cst);
- T load(memory_order m = memory_order_seq_cst) const volatile;
- T load(memory_order m = memory_order_seq_cst) const;
- operator T() const volatile;
- operator T() const;
- T exchange(T desr, memory_order m = memory_order_seq_cst) volatile;
- T exchange(T desr, memory_order m = memory_order_seq_cst);
+ bool is_lock_free() const volatile noexcept;
+ bool is_lock_free() const noexcept;
+ void store(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+ void store(T desr, memory_order m = memory_order_seq_cst) noexcept;
+ T load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+ T load(memory_order m = memory_order_seq_cst) const noexcept;
+ operator T() const volatile noexcept;
+ operator T() const noexcept;
+ T exchange(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+ T exchange(T desr, memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_weak(T& expc, T desr,
- memory_order s, memory_order f) volatile;
- bool compare_exchange_weak(T& expc, T desr, memory_order s, memory_order f);
+ memory_order s, memory_order f) volatile noexcept;
+ bool compare_exchange_weak(T& expc, T desr, memory_order s, memory_order f) noexcept;
bool compare_exchange_strong(T& expc, T desr,
- memory_order s, memory_order f) volatile;
+ memory_order s, memory_order f) volatile noexcept;
bool compare_exchange_strong(T& expc, T desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
bool compare_exchange_weak(T& expc, T desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_weak(T& expc, T desr,
- memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_strong(T& expc, T desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_strong(T& expc, T desr,
- memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) noexcept;
- atomic() = default;
- constexpr atomic(T desr);
+ atomic() noexcept = default;
+ constexpr atomic(T desr) noexcept;
atomic(const atomic&) = delete;
atomic& operator=(const atomic&) = delete;
atomic& operator=(const atomic&) volatile = delete;
- T operator=(T) volatile;
- T operator=(T);
+ T operator=(T) volatile noexcept;
+ T operator=(T) noexcept;
};
template <>
struct atomic<integral>
{
- bool is_lock_free() const volatile;
- bool is_lock_free() const;
- void store(integral desr, memory_order m = memory_order_seq_cst) volatile;
- void store(integral desr, memory_order m = memory_order_seq_cst);
- integral load(memory_order m = memory_order_seq_cst) const volatile;
- integral load(memory_order m = memory_order_seq_cst) const;
- operator integral() const volatile;
- operator integral() const;
+ bool is_lock_free() const volatile noexcept;
+ bool is_lock_free() const noexcept;
+ void store(integral desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+ void store(integral desr, memory_order m = memory_order_seq_cst) noexcept;
+ integral load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+ integral load(memory_order m = memory_order_seq_cst) const noexcept;
+ operator integral() const volatile noexcept;
+ operator integral() const noexcept;
integral exchange(integral desr,
- memory_order m = memory_order_seq_cst) volatile;
- integral exchange(integral desr, memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral exchange(integral desr, memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_weak(integral& expc, integral desr,
- memory_order s, memory_order f) volatile;
+ memory_order s, memory_order f) volatile noexcept;
bool compare_exchange_weak(integral& expc, integral desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
bool compare_exchange_strong(integral& expc, integral desr,
- memory_order s, memory_order f) volatile;
+ memory_order s, memory_order f) volatile noexcept;
bool compare_exchange_strong(integral& expc, integral desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
bool compare_exchange_weak(integral& expc, integral desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_weak(integral& expc, integral desr,
- memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_strong(integral& expc, integral desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_strong(integral& expc, integral desr,
- memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) noexcept;
integral
- fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile;
- integral fetch_add(integral op, memory_order m = memory_order_seq_cst);
+ fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral fetch_add(integral op, memory_order m = memory_order_seq_cst) noexcept;
integral
- fetch_sub(integral op, memory_order m = memory_order_seq_cst) volatile;
- integral fetch_sub(integral op, memory_order m = memory_order_seq_cst);
+ fetch_sub(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral fetch_sub(integral op, memory_order m = memory_order_seq_cst) noexcept;
integral
- fetch_and(integral op, memory_order m = memory_order_seq_cst) volatile;
- integral fetch_and(integral op, memory_order m = memory_order_seq_cst);
+ fetch_and(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral fetch_and(integral op, memory_order m = memory_order_seq_cst) noexcept;
integral
- fetch_or(integral op, memory_order m = memory_order_seq_cst) volatile;
- integral fetch_or(integral op, memory_order m = memory_order_seq_cst);
+ fetch_or(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral fetch_or(integral op, memory_order m = memory_order_seq_cst) noexcept;
integral
- fetch_xor(integral op, memory_order m = memory_order_seq_cst) volatile;
- integral fetch_xor(integral op, memory_order m = memory_order_seq_cst);
+ fetch_xor(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ integral fetch_xor(integral op, memory_order m = memory_order_seq_cst) noexcept;
- atomic() = default;
- constexpr atomic(integral desr);
+ atomic() noexcept = default;
+ constexpr atomic(integral desr) noexcept;
atomic(const atomic&) = delete;
atomic& operator=(const atomic&) = delete;
atomic& operator=(const atomic&) volatile = delete;
- integral operator=(integral desr) volatile;
- integral operator=(integral desr);
-
- integral operator++(int) volatile;
- integral operator++(int);
- integral operator--(int) volatile;
- integral operator--(int);
- integral operator++() volatile;
- integral operator++();
- integral operator--() volatile;
- integral operator--();
- integral operator+=(integral op) volatile;
- integral operator+=(integral op);
- integral operator-=(integral op) volatile;
- integral operator-=(integral op);
- integral operator&=(integral op) volatile;
- integral operator&=(integral op);
- integral operator|=(integral op) volatile;
- integral operator|=(integral op);
- integral operator^=(integral op) volatile;
- integral operator^=(integral op);
+ integral operator=(integral desr) volatile noexcept;
+ integral operator=(integral desr) noexcept;
+
+ integral operator++(int) volatile noexcept;
+ integral operator++(int) noexcept;
+ integral operator--(int) volatile noexcept;
+ integral operator--(int) noexcept;
+ integral operator++() volatile noexcept;
+ integral operator++() noexcept;
+ integral operator--() volatile noexcept;
+ integral operator--() noexcept;
+ integral operator+=(integral op) volatile noexcept;
+ integral operator+=(integral op) noexcept;
+ integral operator-=(integral op) volatile noexcept;
+ integral operator-=(integral op) noexcept;
+ integral operator&=(integral op) volatile noexcept;
+ integral operator&=(integral op) noexcept;
+ integral operator|=(integral op) volatile noexcept;
+ integral operator|=(integral op) noexcept;
+ integral operator^=(integral op) volatile noexcept;
+ integral operator^=(integral op) noexcept;
};
template <class T>
struct atomic<T*>
{
- bool is_lock_free() const volatile;
- bool is_lock_free() const;
- void store(T* desr, memory_order m = memory_order_seq_cst) volatile;
- void store(T* desr, memory_order m = memory_order_seq_cst);
- T* load(memory_order m = memory_order_seq_cst) const volatile;
- T* load(memory_order m = memory_order_seq_cst) const;
- operator T*() const volatile;
- operator T*() const;
- T* exchange(T* desr, memory_order m = memory_order_seq_cst) volatile;
- T* exchange(T* desr, memory_order m = memory_order_seq_cst);
+ bool is_lock_free() const volatile noexcept;
+ bool is_lock_free() const noexcept;
+ void store(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+ void store(T* desr, memory_order m = memory_order_seq_cst) noexcept;
+ T* load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+ T* load(memory_order m = memory_order_seq_cst) const noexcept;
+ operator T*() const volatile noexcept;
+ operator T*() const noexcept;
+ T* exchange(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+ T* exchange(T* desr, memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_weak(T*& expc, T* desr,
- memory_order s, memory_order f) volatile;
+ memory_order s, memory_order f) volatile noexcept;
bool compare_exchange_weak(T*& expc, T* desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
bool compare_exchange_strong(T*& expc, T* desr,
- memory_order s, memory_order f) volatile;
+ memory_order s, memory_order f) volatile noexcept;
bool compare_exchange_strong(T*& expc, T* desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
bool compare_exchange_weak(T*& expc, T* desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_weak(T*& expc, T* desr,
- memory_order m = memory_order_seq_cst);
+ memory_order m = memory_order_seq_cst) noexcept;
bool compare_exchange_strong(T*& expc, T* desr,
- memory_order m = memory_order_seq_cst) volatile;
+ memory_order m = memory_order_seq_cst) volatile noexcept;
bool compare_exchange_strong(T*& expc, T* desr,
- memory_order m = memory_order_seq_cst);
- T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile;
- T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst);
- T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile;
- T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst);
-
- atomic() = default;
- constexpr atomic(T* desr);
+ memory_order m = memory_order_seq_cst) noexcept;
+ T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
+ T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
+ T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
+
+ atomic() noexcept = default;
+ constexpr atomic(T* desr) noexcept;
atomic(const atomic&) = delete;
atomic& operator=(const atomic&) = delete;
atomic& operator=(const atomic&) volatile = delete;
- T* operator=(T*) volatile;
- T* operator=(T*);
- T* operator++(int) volatile;
- T* operator++(int);
- T* operator--(int) volatile;
- T* operator--(int);
- T* operator++() volatile;
- T* operator++();
- T* operator--() volatile;
- T* operator--();
- T* operator+=(ptrdiff_t op) volatile;
- T* operator+=(ptrdiff_t op);
- T* operator-=(ptrdiff_t op) volatile;
- T* operator-=(ptrdiff_t op);
+ T* operator=(T*) volatile noexcept;
+ T* operator=(T*) noexcept;
+ T* operator++(int) volatile noexcept;
+ T* operator++(int) noexcept;
+ T* operator--(int) volatile noexcept;
+ T* operator--(int) noexcept;
+ T* operator++() volatile noexcept;
+ T* operator++() noexcept;
+ T* operator--() volatile noexcept;
+ T* operator--() noexcept;
+ T* operator+=(ptrdiff_t op) volatile noexcept;
+ T* operator+=(ptrdiff_t op) noexcept;
+ T* operator-=(ptrdiff_t op) volatile noexcept;
+ T* operator-=(ptrdiff_t op) noexcept;
};
template <class T>
bool
- atomic_is_lock_free(const volatile atomic<T>* obj);
+ atomic_is_lock_free(const volatile atomic<T>* obj) noexcept;
template <class T>
bool
- atomic_is_lock_free(const atomic<T>* obj);
+ atomic_is_lock_free(const atomic<T>* obj) noexcept;
template <class T>
void
- atomic_init(volatile atomic<T>* obj, T desr);
+ atomic_init(volatile atomic<T>* obj, T desr) noexcept;
template <class T>
void
- atomic_init(atomic<T>* obj, T desr);
+ atomic_init(atomic<T>* obj, T desr) noexcept;
template <class T>
void
- atomic_store(volatile atomic<T>* obj, T desr);
+ atomic_store(volatile atomic<T>* obj, T desr) noexcept;
template <class T>
void
- atomic_store(atomic<T>* obj, T desr);
+ atomic_store(atomic<T>* obj, T desr) noexcept;
template <class T>
void
- atomic_store_explicit(volatile atomic<T>* obj, T desr, memory_order m);
+ atomic_store_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
template <class T>
void
- atomic_store_explicit(atomic<T>* obj, T desr, memory_order m);
+ atomic_store_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
template <class T>
T
- atomic_load(const volatile atomic<T>* obj);
+ atomic_load(const volatile atomic<T>* obj) noexcept;
template <class T>
T
- atomic_load(const atomic<T>* obj);
+ atomic_load(const atomic<T>* obj) noexcept;
template <class T>
T
- atomic_load_explicit(const volatile atomic<T>* obj, memory_order m);
+ atomic_load_explicit(const volatile atomic<T>* obj, memory_order m) noexcept;
template <class T>
T
- atomic_load_explicit(const atomic<T>* obj, memory_order m);
+ atomic_load_explicit(const atomic<T>* obj, memory_order m) noexcept;
template <class T>
T
- atomic_exchange(volatile atomic<T>* obj, T desr);
+ atomic_exchange(volatile atomic<T>* obj, T desr) noexcept;
template <class T>
T
- atomic_exchange(atomic<T>* obj, T desr);
+ atomic_exchange(atomic<T>* obj, T desr) noexcept;
template <class T>
T
- atomic_exchange_explicit(volatile atomic<T>* obj, T desr, memory_order m);
+ atomic_exchange_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
template <class T>
T
- atomic_exchange_explicit(atomic<T>* obj, T desr, memory_order m);
+ atomic_exchange_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
template <class T>
bool
- atomic_compare_exchange_weak(volatile atomic<T>* obj, T* expc, T desr);
+ atomic_compare_exchange_weak(volatile atomic<T>* obj, T* expc, T desr) noexcept;
template <class T>
bool
- atomic_compare_exchange_weak(atomic<T>* obj, T* expc, T desr);
+ atomic_compare_exchange_weak(atomic<T>* obj, T* expc, T desr) noexcept;
template <class T>
bool
- atomic_compare_exchange_strong(volatile atomic<T>* obj, T* expc, T desr);
+ atomic_compare_exchange_strong(volatile atomic<T>* obj, T* expc, T desr) noexcept;
template <class T>
bool
- atomic_compare_exchange_strong(atomic<T>* obj, T* expc, T desr);
+ atomic_compare_exchange_strong(atomic<T>* obj, T* expc, T desr) noexcept;
template <class T>
bool
atomic_compare_exchange_weak_explicit(volatile atomic<T>* obj, T* expc,
T desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
template <class T>
bool
atomic_compare_exchange_weak_explicit(atomic<T>* obj, T* expc, T desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
template <class T>
bool
atomic_compare_exchange_strong_explicit(volatile atomic<T>* obj,
T* expc, T desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
template <class T>
bool
atomic_compare_exchange_strong_explicit(atomic<T>* obj, T* expc,
T desr,
- memory_order s, memory_order f);
+ memory_order s, memory_order f) noexcept;
template <class Integral>
Integral
- atomic_fetch_add(volatile atomic<Integral>* obj, Integral op);
+ atomic_fetch_add(volatile atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
- atomic_fetch_add(atomic<Integral>* obj, Integral op);
+ atomic_fetch_add(atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
atomic_fetch_add_explicit(volatile atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
atomic_fetch_add_explicit(atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
- atomic_fetch_sub(volatile atomic<Integral>* obj, Integral op);
+ atomic_fetch_sub(volatile atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
- atomic_fetch_sub(atomic<Integral>* obj, Integral op);
+ atomic_fetch_sub(atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
atomic_fetch_sub_explicit(volatile atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
atomic_fetch_sub_explicit(atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
- atomic_fetch_and(volatile atomic<Integral>* obj, Integral op);
+ atomic_fetch_and(volatile atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
- atomic_fetch_and(atomic<Integral>* obj, Integral op);
+ atomic_fetch_and(atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
atomic_fetch_and_explicit(volatile atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
atomic_fetch_and_explicit(atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
- atomic_fetch_or(volatile atomic<Integral>* obj, Integral op);
+ atomic_fetch_or(volatile atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
- atomic_fetch_or(atomic<Integral>* obj, Integral op);
+ atomic_fetch_or(atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
atomic_fetch_or_explicit(volatile atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
atomic_fetch_or_explicit(atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
- atomic_fetch_xor(volatile atomic<Integral>* obj, Integral op);
+ atomic_fetch_xor(volatile atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
- atomic_fetch_xor(atomic<Integral>* obj, Integral op);
+ atomic_fetch_xor(atomic<Integral>* obj, Integral op) noexcept;
template <class Integral>
Integral
atomic_fetch_xor_explicit(volatile atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class Integral>
Integral
atomic_fetch_xor_explicit(atomic<Integral>* obj, Integral op,
- memory_order m);
+ memory_order m) noexcept;
template <class T>
T*
- atomic_fetch_add(volatile atomic<T*>* obj, ptrdiff_t op);
+ atomic_fetch_add(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
template <class T>
T*
- atomic_fetch_add(atomic<T*>* obj, ptrdiff_t op);
+ atomic_fetch_add(atomic<T*>* obj, ptrdiff_t op) noexcept;
template <class T>
T*
atomic_fetch_add_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
- memory_order m);
+ memory_order m) noexcept;
template <class T>
T*
- atomic_fetch_add_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m);
+ atomic_fetch_add_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
template <class T>
T*
- atomic_fetch_sub(volatile atomic<T*>* obj, ptrdiff_t op);
+ atomic_fetch_sub(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
template <class T>
T*
- atomic_fetch_sub(atomic<T*>* obj, ptrdiff_t op);
+ atomic_fetch_sub(atomic<T*>* obj, ptrdiff_t op) noexcept;
template <class T>
T*
atomic_fetch_sub_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
- memory_order m);
+ memory_order m) noexcept;
template <class T>
T*
- atomic_fetch_sub_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m);
+ atomic_fetch_sub_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
// Atomics for standard typedef types
+typedef atomic<bool> atomic_bool;
typedef atomic<char> atomic_char;
typedef atomic<signed char> atomic_schar;
typedef atomic<unsigned char> atomic_uchar;
@@ -514,8 +517,8 @@ typedef atomic<uintmax_t> atomic_uintmax_t;
// fences
-void atomic_thread_fence(memory_order m);
-void atomic_signal_fence(memory_order m);
+void atomic_thread_fence(memory_order m) noexcept;
+void atomic_signal_fence(memory_order m) noexcept;
} // std
@@ -526,7 +529,9 @@ void atomic_signal_fence(memory_order m);
#include <cstdint>
#include <type_traits>
+#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER)
#pragma GCC system_header
+#endif
_LIBCPP_BEGIN_NAMESPACE_STD
@@ -543,7 +548,7 @@ typedef enum memory_order
template <class _Tp>
inline _LIBCPP_INLINE_VISIBILITY
_Tp
-kill_dependency(_Tp __y)
+kill_dependency(_Tp __y) _NOEXCEPT
{
return __y;
}
@@ -553,73 +558,73 @@ kill_dependency(_Tp __y)
template <class _Tp, bool = is_integral<_Tp>::value && !is_same<_Tp, bool>::value>
struct __atomic_base // false
{
- _Tp __a_;
+ mutable _Atomic(_Tp) __a_;
_LIBCPP_INLINE_VISIBILITY
- bool is_lock_free() const volatile
- {return __atomic_is_lock_free(_Tp());}
+ bool is_lock_free() const volatile _NOEXCEPT
+ {return __c11_atomic_is_lock_free(sizeof(_Tp));}
_LIBCPP_INLINE_VISIBILITY
- bool is_lock_free() const
- {return __atomic_is_lock_free(_Tp());}
+ bool is_lock_free() const _NOEXCEPT
+ {return __c11_atomic_is_lock_free(sizeof(_Tp));}
_LIBCPP_INLINE_VISIBILITY
- void store(_Tp __d, memory_order __m = memory_order_seq_cst) volatile
- {__atomic_store(&__a_, __d, __m);}
+ void store(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {__c11_atomic_store(&__a_, __d, __m);}
_LIBCPP_INLINE_VISIBILITY
- void store(_Tp __d, memory_order __m = memory_order_seq_cst)
- {__atomic_store(&__a_, __d, __m);}
+ void store(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {__c11_atomic_store(&__a_, __d, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp load(memory_order __m = memory_order_seq_cst) const volatile
- {return __atomic_load(&__a_, __m);}
+ _Tp load(memory_order __m = memory_order_seq_cst) const volatile _NOEXCEPT
+ {return __c11_atomic_load(&__a_, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp load(memory_order __m = memory_order_seq_cst) const
- {return __atomic_load(&__a_, __m);}
+ _Tp load(memory_order __m = memory_order_seq_cst) const _NOEXCEPT
+ {return __c11_atomic_load(&__a_, __m);}
_LIBCPP_INLINE_VISIBILITY
- operator _Tp() const volatile {return load();}
+ operator _Tp() const volatile _NOEXCEPT {return load();}
_LIBCPP_INLINE_VISIBILITY
- operator _Tp() const {return load();}
+ operator _Tp() const _NOEXCEPT {return load();}
_LIBCPP_INLINE_VISIBILITY
- _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_exchange(&__a_, __d, __m);}
+ _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_exchange(&__a_, __d, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst)
- {return __atomic_exchange(&__a_, __d, __m);}
+ _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_exchange(&__a_, __d, __m);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_weak(_Tp& __e, _Tp __d,
- memory_order __s, memory_order __f) volatile
- {return __atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
+ memory_order __s, memory_order __f) volatile _NOEXCEPT
+ {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_weak(_Tp& __e, _Tp __d,
- memory_order __s, memory_order __f)
- {return __atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
+ memory_order __s, memory_order __f) _NOEXCEPT
+ {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_strong(_Tp& __e, _Tp __d,
- memory_order __s, memory_order __f) volatile
- {return __atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
+ memory_order __s, memory_order __f) volatile _NOEXCEPT
+ {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_strong(_Tp& __e, _Tp __d,
- memory_order __s, memory_order __f)
- {return __atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
+ memory_order __s, memory_order __f) _NOEXCEPT
+ {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_weak(_Tp& __e, _Tp __d,
- memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
+ memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_weak(_Tp& __e, _Tp __d,
- memory_order __m = memory_order_seq_cst)
- {return __atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
+ memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_strong(_Tp& __e, _Tp __d,
- memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
+ memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
_LIBCPP_INLINE_VISIBILITY
bool compare_exchange_strong(_Tp& __e, _Tp __d,
- memory_order __m = memory_order_seq_cst)
- {return __atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
+ memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
_LIBCPP_INLINE_VISIBILITY
- __atomic_base() {} // = default;
+ __atomic_base() _NOEXCEPT {} // = default;
_LIBCPP_INLINE_VISIBILITY
- /*constexpr*/ __atomic_base(_Tp __d) : __a_(__d) {}
+ _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __a_(__d) {}
#ifndef _LIBCPP_HAS_NO_DELETED_FUNCTIONS
__atomic_base(const __atomic_base&) = delete;
__atomic_base& operator=(const __atomic_base&) = delete;
@@ -640,77 +645,77 @@ struct __atomic_base<_Tp, true>
{
typedef __atomic_base<_Tp, false> __base;
_LIBCPP_INLINE_VISIBILITY
- __atomic_base() {} // = default;
+ __atomic_base() _NOEXCEPT {} // = default;
_LIBCPP_INLINE_VISIBILITY
- /*constexpr*/ __atomic_base(_Tp __d) : __base(__d) {}
+ _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __base(__d) {}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_fetch_add(&this->__a_, __op, __m);}
+ _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst)
- {return __atomic_fetch_add(&this->__a_, __op, __m);}
+ _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_fetch_sub(&this->__a_, __op, __m);}
+ _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst)
- {return __atomic_fetch_sub(&this->__a_, __op, __m);}
+ _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_fetch_and(&this->__a_, __op, __m);}
+ _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst)
- {return __atomic_fetch_and(&this->__a_, __op, __m);}
+ _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_fetch_or(&this->__a_, __op, __m);}
+ _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst)
- {return __atomic_fetch_or(&this->__a_, __op, __m);}
+ _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
- {return __atomic_fetch_xor(&this->__a_, __op, __m);}
+ _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
+ {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst)
- {return __atomic_fetch_xor(&this->__a_, __op, __m);}
+ _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
+ {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
_LIBCPP_INLINE_VISIBILITY
- _Tp operator++(int) volatile {return fetch_add(_Tp(1));}
+ _Tp operator++(int) volatile _NOEXCEPT {return fetch_add(_Tp(1));}
_LIBCPP_INLINE_VISIBILITY
- _Tp operator++(int) {return fetch_add(_Tp(1));}
+ _Tp operator++(int) _NOEXCEPT {return fetch_add(_Tp(1));}
_LIBCPP_INLINE_VISIBILITY
- _Tp operator--(int) volatile {return fetc