[cfe-commits] [libcxx] r154526 - /libcxx/trunk/include/atomic

Howard Hinnant hhinnant at apple.com
Wed Apr 11 13:14:21 PDT 2012


Author: hhinnant
Date: Wed Apr 11 15:14:21 2012
New Revision: 154526

URL: http://llvm.org/viewvc/llvm-project?rev=154526&view=rev
Log:
Apply noexcept and constexpr to <atomic>.

Modified:
    libcxx/trunk/include/atomic

Modified: libcxx/trunk/include/atomic
URL: http://llvm.org/viewvc/llvm-project/libcxx/trunk/include/atomic?rev=154526&r1=154525&r2=154526&view=diff
==============================================================================
--- libcxx/trunk/include/atomic (original)
+++ libcxx/trunk/include/atomic Wed Apr 11 15:14:21 2012
@@ -29,7 +29,7 @@
     memory_order_seq_cst   // store-release load-acquire
 } memory_order;
 
-template <class T> T kill_dependency(T y);
+template <class T> T kill_dependency(T y) noexcept;
 
 // lock-free property
 
@@ -46,40 +46,40 @@
 
 typedef struct atomic_flag
 {
-    bool test_and_set(memory_order m = memory_order_seq_cst) volatile;
-    bool test_and_set(memory_order m = memory_order_seq_cst);
-    void clear(memory_order m = memory_order_seq_cst) volatile;
-    void clear(memory_order m = memory_order_seq_cst);
-    atomic_flag() = default;
+    bool test_and_set(memory_order m = memory_order_seq_cst) volatile noexcept;
+    bool test_and_set(memory_order m = memory_order_seq_cst) noexcept;
+    void clear(memory_order m = memory_order_seq_cst) volatile noexcept;
+    void clear(memory_order m = memory_order_seq_cst) noexcept;
+    atomic_flag()  noexcept = default;
     atomic_flag(const atomic_flag&) = delete;
     atomic_flag& operator=(const atomic_flag&) = delete;
     atomic_flag& operator=(const atomic_flag&) volatile = delete;
 } atomic_flag;
 
 bool
-    atomic_flag_test_and_set(volatile atomic_flag* obj);
+    atomic_flag_test_and_set(volatile atomic_flag* obj) noexcept;
 
 bool
-    atomic_flag_test_and_set(atomic_flag* obj);
+    atomic_flag_test_and_set(atomic_flag* obj) noexcept;
 
 bool
     atomic_flag_test_and_set_explicit(volatile atomic_flag* obj,
-                                      memory_order m);
+                                      memory_order m) noexcept;
 
 bool
-    atomic_flag_test_and_set_explicit(atomic_flag* obj, memory_order m);
+    atomic_flag_test_and_set_explicit(atomic_flag* obj, memory_order m) noexcept;
 
 void
-    atomic_flag_clear(volatile atomic_flag* obj);
+    atomic_flag_clear(volatile atomic_flag* obj) noexcept;
 
 void
-    atomic_flag_clear(atomic_flag* obj);
+    atomic_flag_clear(atomic_flag* obj) noexcept;
 
 void
-    atomic_flag_clear_explicit(volatile atomic_flag* obj, memory_order m);
+    atomic_flag_clear_explicit(volatile atomic_flag* obj, memory_order m) noexcept;
 
 void
-    atomic_flag_clear_explicit(atomic_flag* obj, memory_order m);
+    atomic_flag_clear_explicit(atomic_flag* obj, memory_order m) noexcept;
 
 #define ATOMIC_FLAG_INIT see below
 #define ATOMIC_VAR_INIT(value) see below
@@ -87,388 +87,388 @@
 template <class T>
 struct atomic
 {
-    bool is_lock_free() const volatile;
-    bool is_lock_free() const;
-    void store(T desr, memory_order m = memory_order_seq_cst) volatile;
-    void store(T desr, memory_order m = memory_order_seq_cst);
-    T load(memory_order m = memory_order_seq_cst) const volatile;
-    T load(memory_order m = memory_order_seq_cst) const;
-    operator T() const volatile;
-    operator T() const;
-    T exchange(T desr, memory_order m = memory_order_seq_cst) volatile;
-    T exchange(T desr, memory_order m = memory_order_seq_cst);
+    bool is_lock_free() const volatile noexcept;
+    bool is_lock_free() const noexcept;
+    void store(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+    void store(T desr, memory_order m = memory_order_seq_cst) noexcept;
+    T load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+    T load(memory_order m = memory_order_seq_cst) const noexcept;
+    operator T() const volatile noexcept;
+    operator T() const noexcept;
+    T exchange(T desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+    T exchange(T desr, memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_weak(T& expc, T desr,
-                               memory_order s, memory_order f) volatile;
-    bool compare_exchange_weak(T& expc, T desr, memory_order s, memory_order f);
+                               memory_order s, memory_order f) volatile noexcept;
+    bool compare_exchange_weak(T& expc, T desr, memory_order s, memory_order f) noexcept;
     bool compare_exchange_strong(T& expc, T desr,
-                                 memory_order s, memory_order f) volatile;
+                                 memory_order s, memory_order f) volatile noexcept;
     bool compare_exchange_strong(T& expc, T desr,
-                                 memory_order s, memory_order f);
+                                 memory_order s, memory_order f) noexcept;
     bool compare_exchange_weak(T& expc, T desr,
-                               memory_order m = memory_order_seq_cst) volatile;
+                               memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_weak(T& expc, T desr,
-                               memory_order m = memory_order_seq_cst);
+                               memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_strong(T& expc, T desr,
-                                memory_order m = memory_order_seq_cst) volatile;
+                                memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_strong(T& expc, T desr,
-                                 memory_order m = memory_order_seq_cst);
+                                 memory_order m = memory_order_seq_cst) noexcept;
 
-    atomic() = default;
-    constexpr atomic(T desr);
+    atomic() noexcept = default;
+    constexpr atomic(T desr) noexcept;
     atomic(const atomic&) = delete;
     atomic& operator=(const atomic&) = delete;
     atomic& operator=(const atomic&) volatile = delete;
-    T operator=(T) volatile;
-    T operator=(T);
+    T operator=(T) volatile noexcept;
+    T operator=(T) noexcept;
 };
 
 template <>
 struct atomic<integral>
 {
-    bool is_lock_free() const volatile;
-    bool is_lock_free() const;
-    void store(integral desr, memory_order m = memory_order_seq_cst) volatile;
-    void store(integral desr, memory_order m = memory_order_seq_cst);
-    integral load(memory_order m = memory_order_seq_cst) const volatile;
-    integral load(memory_order m = memory_order_seq_cst) const;
-    operator integral() const volatile;
-    operator integral() const;
+    bool is_lock_free() const volatile noexcept;
+    bool is_lock_free() const noexcept;
+    void store(integral desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+    void store(integral desr, memory_order m = memory_order_seq_cst) noexcept;
+    integral load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+    integral load(memory_order m = memory_order_seq_cst) const noexcept;
+    operator integral() const volatile noexcept;
+    operator integral() const noexcept;
     integral exchange(integral desr,
-                      memory_order m = memory_order_seq_cst) volatile;
-    integral exchange(integral desr, memory_order m = memory_order_seq_cst);
+                      memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral exchange(integral desr, memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_weak(integral& expc, integral desr,
-                               memory_order s, memory_order f) volatile;
+                               memory_order s, memory_order f) volatile noexcept;
     bool compare_exchange_weak(integral& expc, integral desr,
-                               memory_order s, memory_order f);
+                               memory_order s, memory_order f) noexcept;
     bool compare_exchange_strong(integral& expc, integral desr,
-                                 memory_order s, memory_order f) volatile;
+                                 memory_order s, memory_order f) volatile noexcept;
     bool compare_exchange_strong(integral& expc, integral desr,
-                                 memory_order s, memory_order f);
+                                 memory_order s, memory_order f) noexcept;
     bool compare_exchange_weak(integral& expc, integral desr,
-                               memory_order m = memory_order_seq_cst) volatile;
+                               memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_weak(integral& expc, integral desr,
-                               memory_order m = memory_order_seq_cst);
+                               memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_strong(integral& expc, integral desr,
-                                memory_order m = memory_order_seq_cst) volatile;
+                                memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_strong(integral& expc, integral desr,
-                                 memory_order m = memory_order_seq_cst);
+                                 memory_order m = memory_order_seq_cst) noexcept;
 
     integral
-        fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile;
-    integral fetch_add(integral op, memory_order m = memory_order_seq_cst);
+        fetch_add(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral fetch_add(integral op, memory_order m = memory_order_seq_cst) noexcept;
     integral
-        fetch_sub(integral op, memory_order m = memory_order_seq_cst) volatile;
-    integral fetch_sub(integral op, memory_order m = memory_order_seq_cst);
+        fetch_sub(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral fetch_sub(integral op, memory_order m = memory_order_seq_cst) noexcept;
     integral
-        fetch_and(integral op, memory_order m = memory_order_seq_cst) volatile;
-    integral fetch_and(integral op, memory_order m = memory_order_seq_cst);
+        fetch_and(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral fetch_and(integral op, memory_order m = memory_order_seq_cst) noexcept;
     integral
-        fetch_or(integral op, memory_order m = memory_order_seq_cst) volatile;
-    integral fetch_or(integral op, memory_order m = memory_order_seq_cst);
+        fetch_or(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral fetch_or(integral op, memory_order m = memory_order_seq_cst) noexcept;
     integral
-        fetch_xor(integral op, memory_order m = memory_order_seq_cst) volatile;
-    integral fetch_xor(integral op, memory_order m = memory_order_seq_cst);
+        fetch_xor(integral op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    integral fetch_xor(integral op, memory_order m = memory_order_seq_cst) noexcept;
 
-    atomic() = default;
-    constexpr atomic(integral desr);
+    atomic() noexcept = default;
+    constexpr atomic(integral desr) noexcept;
     atomic(const atomic&) = delete;
     atomic& operator=(const atomic&) = delete;
     atomic& operator=(const atomic&) volatile = delete;
-    integral operator=(integral desr) volatile;
-    integral operator=(integral desr);
+    integral operator=(integral desr) volatile noexcept;
+    integral operator=(integral desr) noexcept;
 
-    integral operator++(int) volatile;
-    integral operator++(int);
-    integral operator--(int) volatile;
-    integral operator--(int);
-    integral operator++() volatile;
-    integral operator++();
-    integral operator--() volatile;
-    integral operator--();
-    integral operator+=(integral op) volatile;
-    integral operator+=(integral op);
-    integral operator-=(integral op) volatile;
-    integral operator-=(integral op);
-    integral operator&=(integral op) volatile;
-    integral operator&=(integral op);
-    integral operator|=(integral op) volatile;
-    integral operator|=(integral op);
-    integral operator^=(integral op) volatile;
-    integral operator^=(integral op);
+    integral operator++(int) volatile noexcept;
+    integral operator++(int) noexcept;
+    integral operator--(int) volatile noexcept;
+    integral operator--(int) noexcept;
+    integral operator++() volatile noexcept;
+    integral operator++() noexcept;
+    integral operator--() volatile noexcept;
+    integral operator--() noexcept;
+    integral operator+=(integral op) volatile noexcept;
+    integral operator+=(integral op) noexcept;
+    integral operator-=(integral op) volatile noexcept;
+    integral operator-=(integral op) noexcept;
+    integral operator&=(integral op) volatile noexcept;
+    integral operator&=(integral op) noexcept;
+    integral operator|=(integral op) volatile noexcept;
+    integral operator|=(integral op) noexcept;
+    integral operator^=(integral op) volatile noexcept;
+    integral operator^=(integral op) noexcept;
 };
 
 template <class T>
 struct atomic<T*>
 {
-    bool is_lock_free() const volatile;
-    bool is_lock_free() const;
-    void store(T* desr, memory_order m = memory_order_seq_cst) volatile;
-    void store(T* desr, memory_order m = memory_order_seq_cst);
-    T* load(memory_order m = memory_order_seq_cst) const volatile;
-    T* load(memory_order m = memory_order_seq_cst) const;
-    operator T*() const volatile;
-    operator T*() const;
-    T* exchange(T* desr, memory_order m = memory_order_seq_cst) volatile;
-    T* exchange(T* desr, memory_order m = memory_order_seq_cst);
+    bool is_lock_free() const volatile noexcept;
+    bool is_lock_free() const noexcept;
+    void store(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+    void store(T* desr, memory_order m = memory_order_seq_cst) noexcept;
+    T* load(memory_order m = memory_order_seq_cst) const volatile noexcept;
+    T* load(memory_order m = memory_order_seq_cst) const noexcept;
+    operator T*() const volatile noexcept;
+    operator T*() const noexcept;
+    T* exchange(T* desr, memory_order m = memory_order_seq_cst) volatile noexcept;
+    T* exchange(T* desr, memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_weak(T*& expc, T* desr,
-                               memory_order s, memory_order f) volatile;
+                               memory_order s, memory_order f) volatile noexcept;
     bool compare_exchange_weak(T*& expc, T* desr,
-                               memory_order s, memory_order f);
+                               memory_order s, memory_order f) noexcept;
     bool compare_exchange_strong(T*& expc, T* desr,
-                                 memory_order s, memory_order f) volatile;
+                                 memory_order s, memory_order f) volatile noexcept;
     bool compare_exchange_strong(T*& expc, T* desr,
-                                 memory_order s, memory_order f);
+                                 memory_order s, memory_order f) noexcept;
     bool compare_exchange_weak(T*& expc, T* desr,
-                               memory_order m = memory_order_seq_cst) volatile;
+                               memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_weak(T*& expc, T* desr,
-                               memory_order m = memory_order_seq_cst);
+                               memory_order m = memory_order_seq_cst) noexcept;
     bool compare_exchange_strong(T*& expc, T* desr,
-                                memory_order m = memory_order_seq_cst) volatile;
+                                memory_order m = memory_order_seq_cst) volatile noexcept;
     bool compare_exchange_strong(T*& expc, T* desr,
-                                 memory_order m = memory_order_seq_cst);
-    T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile;
-    T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst);
-    T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile;
-    T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst);
+                                 memory_order m = memory_order_seq_cst) noexcept;
+    T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    T* fetch_add(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
+    T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) volatile noexcept;
+    T* fetch_sub(ptrdiff_t op, memory_order m = memory_order_seq_cst) noexcept;
 
-    atomic() = default;
-    constexpr atomic(T* desr);
+    atomic() noexcept = default;
+    constexpr atomic(T* desr) noexcept;
     atomic(const atomic&) = delete;
     atomic& operator=(const atomic&) = delete;
     atomic& operator=(const atomic&) volatile = delete;
 
-    T* operator=(T*) volatile;
-    T* operator=(T*);
-    T* operator++(int) volatile;
-    T* operator++(int);
-    T* operator--(int) volatile;
-    T* operator--(int);
-    T* operator++() volatile;
-    T* operator++();
-    T* operator--() volatile;
-    T* operator--();
-    T* operator+=(ptrdiff_t op) volatile;
-    T* operator+=(ptrdiff_t op);
-    T* operator-=(ptrdiff_t op) volatile;
-    T* operator-=(ptrdiff_t op);
+    T* operator=(T*) volatile noexcept;
+    T* operator=(T*) noexcept;
+    T* operator++(int) volatile noexcept;
+    T* operator++(int) noexcept;
+    T* operator--(int) volatile noexcept;
+    T* operator--(int) noexcept;
+    T* operator++() volatile noexcept;
+    T* operator++() noexcept;
+    T* operator--() volatile noexcept;
+    T* operator--() noexcept;
+    T* operator+=(ptrdiff_t op) volatile noexcept;
+    T* operator+=(ptrdiff_t op) noexcept;
+    T* operator-=(ptrdiff_t op) volatile noexcept;
+    T* operator-=(ptrdiff_t op) noexcept;
 };
 
 
 template <class T>
     bool
-    atomic_is_lock_free(const volatile atomic<T>* obj);
+    atomic_is_lock_free(const volatile atomic<T>* obj) noexcept;
 
 template <class T>
     bool
-    atomic_is_lock_free(const atomic<T>* obj);
+    atomic_is_lock_free(const atomic<T>* obj) noexcept;
 
 template <class T>
     void
-    atomic_init(volatile atomic<T>* obj, T desr);
+    atomic_init(volatile atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     void
-    atomic_init(atomic<T>* obj, T desr);
+    atomic_init(atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     void
-    atomic_store(volatile atomic<T>* obj, T desr);
+    atomic_store(volatile atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     void
-    atomic_store(atomic<T>* obj, T desr);
+    atomic_store(atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     void
-    atomic_store_explicit(volatile atomic<T>* obj, T desr, memory_order m);
+    atomic_store_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
 
 template <class T>
     void
-    atomic_store_explicit(atomic<T>* obj, T desr, memory_order m);
+    atomic_store_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
 
 template <class T>
     T
-    atomic_load(const volatile atomic<T>* obj);
+    atomic_load(const volatile atomic<T>* obj) noexcept;
 
 template <class T>
     T
-    atomic_load(const atomic<T>* obj);
+    atomic_load(const atomic<T>* obj) noexcept;
 
 template <class T>
     T
-    atomic_load_explicit(const volatile atomic<T>* obj, memory_order m);
+    atomic_load_explicit(const volatile atomic<T>* obj, memory_order m) noexcept;
 
 template <class T>
     T
-    atomic_load_explicit(const atomic<T>* obj, memory_order m);
+    atomic_load_explicit(const atomic<T>* obj, memory_order m) noexcept;
 
 template <class T>
     T
-    atomic_exchange(volatile atomic<T>* obj, T desr);
+    atomic_exchange(volatile atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     T
-    atomic_exchange(atomic<T>* obj, T desr);
+    atomic_exchange(atomic<T>* obj, T desr) noexcept;
 
 template <class T>
     T
-    atomic_exchange_explicit(volatile atomic<T>* obj, T desr, memory_order m);
+    atomic_exchange_explicit(volatile atomic<T>* obj, T desr, memory_order m) noexcept;
 
 template <class T>
     T
-    atomic_exchange_explicit(atomic<T>* obj, T desr, memory_order m);
+    atomic_exchange_explicit(atomic<T>* obj, T desr, memory_order m) noexcept;
 
 template <class T>
     bool
-    atomic_compare_exchange_weak(volatile atomic<T>* obj, T* expc, T desr);
+    atomic_compare_exchange_weak(volatile atomic<T>* obj, T* expc, T desr) noexcept;
 
 template <class T>
     bool
-    atomic_compare_exchange_weak(atomic<T>* obj, T* expc, T desr);
+    atomic_compare_exchange_weak(atomic<T>* obj, T* expc, T desr) noexcept;
 
 template <class T>
     bool
-    atomic_compare_exchange_strong(volatile atomic<T>* obj, T* expc, T desr);
+    atomic_compare_exchange_strong(volatile atomic<T>* obj, T* expc, T desr) noexcept;
 
 template <class T>
     bool
-    atomic_compare_exchange_strong(atomic<T>* obj, T* expc, T desr);
+    atomic_compare_exchange_strong(atomic<T>* obj, T* expc, T desr) noexcept;
 
 template <class T>
     bool
     atomic_compare_exchange_weak_explicit(volatile atomic<T>* obj, T* expc,
                                           T desr,
-                                          memory_order s, memory_order f);
+                                          memory_order s, memory_order f) noexcept;
 
 template <class T>
     bool
     atomic_compare_exchange_weak_explicit(atomic<T>* obj, T* expc, T desr,
-                                          memory_order s, memory_order f);
+                                          memory_order s, memory_order f) noexcept;
 
 template <class T>
     bool
     atomic_compare_exchange_strong_explicit(volatile atomic<T>* obj,
                                             T* expc, T desr,
-                                            memory_order s, memory_order f);
+                                            memory_order s, memory_order f) noexcept;
 
 template <class T>
     bool
     atomic_compare_exchange_strong_explicit(atomic<T>* obj, T* expc,
                                             T desr,
-                                            memory_order s, memory_order f);
+                                            memory_order s, memory_order f) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_add(volatile atomic<Integral>* obj, Integral op);
+    atomic_fetch_add(volatile atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_add(atomic<Integral>* obj, Integral op);
+    atomic_fetch_add(atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
     atomic_fetch_add_explicit(volatile atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
     atomic_fetch_add_explicit(atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
-    atomic_fetch_sub(volatile atomic<Integral>* obj, Integral op);
+    atomic_fetch_sub(volatile atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_sub(atomic<Integral>* obj, Integral op);
+    atomic_fetch_sub(atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
     atomic_fetch_sub_explicit(volatile atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
     atomic_fetch_sub_explicit(atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
-    atomic_fetch_and(volatile atomic<Integral>* obj, Integral op);
+    atomic_fetch_and(volatile atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_and(atomic<Integral>* obj, Integral op);
+    atomic_fetch_and(atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
     atomic_fetch_and_explicit(volatile atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
     atomic_fetch_and_explicit(atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
-    atomic_fetch_or(volatile atomic<Integral>* obj, Integral op);
+    atomic_fetch_or(volatile atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_or(atomic<Integral>* obj, Integral op);
+    atomic_fetch_or(atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
     atomic_fetch_or_explicit(volatile atomic<Integral>* obj, Integral op,
-                             memory_order m);
+                             memory_order m) noexcept;
 template <class Integral>
     Integral
     atomic_fetch_or_explicit(atomic<Integral>* obj, Integral op,
-                             memory_order m);
+                             memory_order m) noexcept;
 template <class Integral>
     Integral
-    atomic_fetch_xor(volatile atomic<Integral>* obj, Integral op);
+    atomic_fetch_xor(volatile atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
-    atomic_fetch_xor(atomic<Integral>* obj, Integral op);
+    atomic_fetch_xor(atomic<Integral>* obj, Integral op) noexcept;
 
 template <class Integral>
     Integral
     atomic_fetch_xor_explicit(volatile atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class Integral>
     Integral
     atomic_fetch_xor_explicit(atomic<Integral>* obj, Integral op,
-                              memory_order m);
+                              memory_order m) noexcept;
 
 template <class T>
     T*
-    atomic_fetch_add(volatile atomic<T*>* obj, ptrdiff_t op);
+    atomic_fetch_add(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
 
 template <class T>
     T*
-    atomic_fetch_add(atomic<T*>* obj, ptrdiff_t op);
+    atomic_fetch_add(atomic<T*>* obj, ptrdiff_t op) noexcept;
 
 template <class T>
     T*
     atomic_fetch_add_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class T>
     T*
-    atomic_fetch_add_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m);
+    atomic_fetch_add_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
 
 template <class T>
     T*
-    atomic_fetch_sub(volatile atomic<T*>* obj, ptrdiff_t op);
+    atomic_fetch_sub(volatile atomic<T*>* obj, ptrdiff_t op) noexcept;
 
 template <class T>
     T*
-    atomic_fetch_sub(atomic<T*>* obj, ptrdiff_t op);
+    atomic_fetch_sub(atomic<T*>* obj, ptrdiff_t op) noexcept;
 
 template <class T>
     T*
     atomic_fetch_sub_explicit(volatile atomic<T*>* obj, ptrdiff_t op,
-                              memory_order m);
+                              memory_order m) noexcept;
 template <class T>
     T*
-    atomic_fetch_sub_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m);
+    atomic_fetch_sub_explicit(atomic<T*>* obj, ptrdiff_t op, memory_order m) noexcept;
 
 // Atomics for standard typedef types
 
@@ -514,8 +514,8 @@
 
 // fences
 
-void atomic_thread_fence(memory_order m);
-void atomic_signal_fence(memory_order m);
+void atomic_thread_fence(memory_order m) noexcept;
+void atomic_signal_fence(memory_order m) noexcept;
 
 }  // std
 
@@ -545,7 +545,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-kill_dependency(_Tp __y)
+kill_dependency(_Tp __y) _NOEXCEPT
 {
     return __y;
 }
@@ -558,70 +558,70 @@
     _Atomic(_Tp) __a_;
 
     _LIBCPP_INLINE_VISIBILITY
-    bool is_lock_free() const volatile
+    bool is_lock_free() const volatile _NOEXCEPT
         {return __c11_atomic_is_lock_free(sizeof(_Tp));}
     _LIBCPP_INLINE_VISIBILITY
-    bool is_lock_free() const
+    bool is_lock_free() const _NOEXCEPT
         {return __c11_atomic_is_lock_free(sizeof(_Tp));}
     _LIBCPP_INLINE_VISIBILITY
-    void store(_Tp __d, memory_order __m = memory_order_seq_cst) volatile
+    void store(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {__c11_atomic_store(&__a_, __d, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    void store(_Tp __d, memory_order __m = memory_order_seq_cst)
+    void store(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {__c11_atomic_store(&__a_, __d, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp load(memory_order __m = memory_order_seq_cst) const volatile
+    _Tp load(memory_order __m = memory_order_seq_cst) const volatile _NOEXCEPT
         {return __c11_atomic_load(&__a_, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp load(memory_order __m = memory_order_seq_cst) const
+    _Tp load(memory_order __m = memory_order_seq_cst) const _NOEXCEPT
         {return __c11_atomic_load(&__a_, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    operator _Tp() const volatile {return load();}
+    operator _Tp() const volatile _NOEXCEPT {return load();}
     _LIBCPP_INLINE_VISIBILITY
-    operator _Tp() const          {return load();}
+    operator _Tp() const _NOEXCEPT          {return load();}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) volatile
+    _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_exchange(&__a_, __d, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst)
+    _Tp exchange(_Tp __d, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_exchange(&__a_, __d, __m);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_weak(_Tp& __e, _Tp __d,
-                               memory_order __s, memory_order __f) volatile
+                               memory_order __s, memory_order __f) volatile _NOEXCEPT
         {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_weak(_Tp& __e, _Tp __d,
-                               memory_order __s, memory_order __f)
+                               memory_order __s, memory_order __f) _NOEXCEPT
         {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __s, __f);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_strong(_Tp& __e, _Tp __d,
-                                 memory_order __s, memory_order __f) volatile
+                                 memory_order __s, memory_order __f) volatile _NOEXCEPT
         {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_strong(_Tp& __e, _Tp __d,
-                                 memory_order __s, memory_order __f)
+                                 memory_order __s, memory_order __f) _NOEXCEPT
         {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __s, __f);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_weak(_Tp& __e, _Tp __d,
-                              memory_order __m = memory_order_seq_cst) volatile
+                              memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_weak(_Tp& __e, _Tp __d,
-                               memory_order __m = memory_order_seq_cst)
+                               memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_compare_exchange_weak(&__a_, &__e, __d, __m, __m);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_strong(_Tp& __e, _Tp __d,
-                              memory_order __m = memory_order_seq_cst) volatile
+                              memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
     _LIBCPP_INLINE_VISIBILITY
     bool compare_exchange_strong(_Tp& __e, _Tp __d,
-                                 memory_order __m = memory_order_seq_cst)
+                                 memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_compare_exchange_strong(&__a_, &__e, __d, __m, __m);}
 
     _LIBCPP_INLINE_VISIBILITY
-    __atomic_base() {} // = default;
+    __atomic_base() _NOEXCEPT {} // = default;
     _LIBCPP_INLINE_VISIBILITY
-    /*constexpr*/ __atomic_base(_Tp __d) : __a_(__d) {}
+    _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __a_(__d) {}
 #ifndef _LIBCPP_HAS_NO_DELETED_FUNCTIONS
     __atomic_base(const __atomic_base&) = delete;
     __atomic_base& operator=(const __atomic_base&) = delete;
@@ -642,77 +642,77 @@
 {
     typedef __atomic_base<_Tp, false> __base;
     _LIBCPP_INLINE_VISIBILITY
-    __atomic_base() {} // = default;
+    __atomic_base() _NOEXCEPT {} // = default;
     _LIBCPP_INLINE_VISIBILITY
-    /*constexpr*/ __atomic_base(_Tp __d) : __base(__d) {}
+    _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __base(__d) {}
 
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
+    _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst)
+    _Tp fetch_add(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
+    _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst)
+    _Tp fetch_sub(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
+    _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst)
+    _Tp fetch_and(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_and(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
+    _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst)
+    _Tp fetch_or(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_or(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) volatile
+    _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst)
+    _Tp fetch_xor(_Tp __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_xor(&this->__a_, __op, __m);}
 
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator++(int) volatile      {return fetch_add(_Tp(1));}
+    _Tp operator++(int) volatile _NOEXCEPT      {return fetch_add(_Tp(1));}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator++(int)               {return fetch_add(_Tp(1));}
+    _Tp operator++(int) _NOEXCEPT               {return fetch_add(_Tp(1));}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator--(int) volatile      {return fetch_sub(_Tp(1));}
+    _Tp operator--(int) volatile _NOEXCEPT      {return fetch_sub(_Tp(1));}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator--(int)               {return fetch_sub(_Tp(1));}
+    _Tp operator--(int) _NOEXCEPT               {return fetch_sub(_Tp(1));}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator++() volatile         {return fetch_add(_Tp(1)) + _Tp(1);}
+    _Tp operator++() volatile _NOEXCEPT         {return fetch_add(_Tp(1)) + _Tp(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator++()                  {return fetch_add(_Tp(1)) + _Tp(1);}
+    _Tp operator++() _NOEXCEPT                  {return fetch_add(_Tp(1)) + _Tp(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator--() volatile         {return fetch_sub(_Tp(1)) - _Tp(1);}
+    _Tp operator--() volatile _NOEXCEPT         {return fetch_sub(_Tp(1)) - _Tp(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator--()                  {return fetch_sub(_Tp(1)) - _Tp(1);}
+    _Tp operator--() _NOEXCEPT                  {return fetch_sub(_Tp(1)) - _Tp(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator+=(_Tp __op) volatile {return fetch_add(__op) + __op;}
+    _Tp operator+=(_Tp __op) volatile _NOEXCEPT {return fetch_add(__op) + __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator+=(_Tp __op)          {return fetch_add(__op) + __op;}
+    _Tp operator+=(_Tp __op) _NOEXCEPT          {return fetch_add(__op) + __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator-=(_Tp __op) volatile {return fetch_sub(__op) - __op;}
+    _Tp operator-=(_Tp __op) volatile _NOEXCEPT {return fetch_sub(__op) - __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator-=(_Tp __op)          {return fetch_sub(__op) - __op;}
+    _Tp operator-=(_Tp __op) _NOEXCEPT          {return fetch_sub(__op) - __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator&=(_Tp __op) volatile {return fetch_and(__op) & __op;}
+    _Tp operator&=(_Tp __op) volatile _NOEXCEPT {return fetch_and(__op) & __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator&=(_Tp __op)          {return fetch_and(__op) & __op;}
+    _Tp operator&=(_Tp __op) _NOEXCEPT          {return fetch_and(__op) & __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator|=(_Tp __op) volatile {return fetch_or(__op) | __op;}
+    _Tp operator|=(_Tp __op) volatile _NOEXCEPT {return fetch_or(__op) | __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator|=(_Tp __op)          {return fetch_or(__op) | __op;}
+    _Tp operator|=(_Tp __op) _NOEXCEPT          {return fetch_or(__op) | __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator^=(_Tp __op) volatile {return fetch_xor(__op) ^ __op;}
+    _Tp operator^=(_Tp __op) volatile _NOEXCEPT {return fetch_xor(__op) ^ __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator^=(_Tp __op)          {return fetch_xor(__op) ^ __op;}
+    _Tp operator^=(_Tp __op) _NOEXCEPT          {return fetch_xor(__op) ^ __op;}
 };
 
 // atomic<T>
@@ -723,15 +723,15 @@
 {
     typedef __atomic_base<_Tp> __base;
     _LIBCPP_INLINE_VISIBILITY
-    atomic() {} // = default;
+    atomic() _NOEXCEPT {} // = default;
     _LIBCPP_INLINE_VISIBILITY
-    /*constexpr*/ atomic(_Tp __d) : __base(__d) {}
+    _LIBCPP_CONSTEXPR atomic(_Tp __d) _NOEXCEPT : __base(__d) {}
 
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator=(_Tp __d) volatile
+    _Tp operator=(_Tp __d) volatile _NOEXCEPT
         {__base::store(__d); return __d;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp operator=(_Tp __d)
+    _Tp operator=(_Tp __d) _NOEXCEPT
         {__base::store(__d); return __d;}
 };
 
@@ -743,56 +743,56 @@
 {
     typedef __atomic_base<_Tp*> __base;
     _LIBCPP_INLINE_VISIBILITY
-    atomic() {} // = default;
+    atomic() _NOEXCEPT {} // = default;
     _LIBCPP_INLINE_VISIBILITY
-    /*constexpr*/ atomic(_Tp* __d) : __base(__d) {}
+    _LIBCPP_CONSTEXPR atomic(_Tp* __d) _NOEXCEPT : __base(__d) {}
 
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator=(_Tp* __d) volatile
+    _Tp* operator=(_Tp* __d) volatile _NOEXCEPT
         {__base::store(__d); return __d;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator=(_Tp* __d)
+    _Tp* operator=(_Tp* __d) _NOEXCEPT
         {__base::store(__d); return __d;}
 
     _LIBCPP_INLINE_VISIBILITY
     _Tp* fetch_add(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
-                                                                        volatile
+                                                                        volatile _NOEXCEPT
         {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* fetch_add(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
+    _Tp* fetch_add(ptrdiff_t __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_add(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
     _Tp* fetch_sub(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
-                                                                        volatile
+                                                                        volatile _NOEXCEPT
         {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* fetch_sub(ptrdiff_t __op, memory_order __m = memory_order_seq_cst)
+    _Tp* fetch_sub(ptrdiff_t __op, memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_fetch_sub(&this->__a_, __op, __m);}
 
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator++(int) volatile            {return fetch_add(1);}
+    _Tp* operator++(int) volatile _NOEXCEPT            {return fetch_add(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator++(int)                     {return fetch_add(1);}
+    _Tp* operator++(int) _NOEXCEPT                     {return fetch_add(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator--(int) volatile            {return fetch_sub(1);}
+    _Tp* operator--(int) volatile _NOEXCEPT            {return fetch_sub(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator--(int)                     {return fetch_sub(1);}
+    _Tp* operator--(int) _NOEXCEPT                     {return fetch_sub(1);}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator++() volatile               {return fetch_add(1) + 1;}
+    _Tp* operator++() volatile _NOEXCEPT               {return fetch_add(1) + 1;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator++()                        {return fetch_add(1) + 1;}
+    _Tp* operator++() _NOEXCEPT                        {return fetch_add(1) + 1;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator--() volatile               {return fetch_sub(1) - 1;}
+    _Tp* operator--() volatile _NOEXCEPT               {return fetch_sub(1) - 1;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator--()                        {return fetch_sub(1) - 1;}
+    _Tp* operator--() _NOEXCEPT                        {return fetch_sub(1) - 1;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator+=(ptrdiff_t __op) volatile {return fetch_add(__op) + __op;}
+    _Tp* operator+=(ptrdiff_t __op) volatile _NOEXCEPT {return fetch_add(__op) + __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator+=(ptrdiff_t __op)          {return fetch_add(__op) + __op;}
+    _Tp* operator+=(ptrdiff_t __op) _NOEXCEPT          {return fetch_add(__op) + __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator-=(ptrdiff_t __op) volatile {return fetch_sub(__op) - __op;}
+    _Tp* operator-=(ptrdiff_t __op) volatile _NOEXCEPT {return fetch_sub(__op) - __op;}
     _LIBCPP_INLINE_VISIBILITY
-    _Tp* operator-=(ptrdiff_t __op)          {return fetch_sub(__op) - __op;}
+    _Tp* operator-=(ptrdiff_t __op) _NOEXCEPT          {return fetch_sub(__op) - __op;}
 };
 
 // atomic_is_lock_free
@@ -800,7 +800,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_is_lock_free(const volatile atomic<_Tp>* __o)
+atomic_is_lock_free(const volatile atomic<_Tp>* __o) _NOEXCEPT
 {
     return __o->is_lock_free();
 }
@@ -808,7 +808,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_is_lock_free(const atomic<_Tp>* __o)
+atomic_is_lock_free(const atomic<_Tp>* __o) _NOEXCEPT
 {
     return __o->is_lock_free();
 }
@@ -818,7 +818,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_init(volatile atomic<_Tp>* __o, _Tp __d)
+atomic_init(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     __c11_atomic_init(&__o->__a_, __d);
 }
@@ -826,7 +826,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_init(atomic<_Tp>* __o, _Tp __d)
+atomic_init(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     __c11_atomic_init(&__o->__a_, __d);
 }
@@ -836,7 +836,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_store(volatile atomic<_Tp>* __o, _Tp __d)
+atomic_store(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     __o->store(__d);
 }
@@ -844,7 +844,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_store(atomic<_Tp>* __o, _Tp __d)
+atomic_store(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     __o->store(__d);
 }
@@ -854,7 +854,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_store_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m)
+atomic_store_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
 {
     __o->store(__d, __m);
 }
@@ -862,7 +862,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_store_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m)
+atomic_store_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
 {
     __o->store(__d, __m);
 }
@@ -872,7 +872,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_load(const volatile atomic<_Tp>* __o)
+atomic_load(const volatile atomic<_Tp>* __o) _NOEXCEPT
 {
     return __o->load();
 }
@@ -880,7 +880,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_load(const atomic<_Tp>* __o)
+atomic_load(const atomic<_Tp>* __o) _NOEXCEPT
 {
     return __o->load();
 }
@@ -890,7 +890,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_load_explicit(const volatile atomic<_Tp>* __o, memory_order __m)
+atomic_load_explicit(const volatile atomic<_Tp>* __o, memory_order __m) _NOEXCEPT
 {
     return __o->load(__m);
 }
@@ -898,7 +898,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_load_explicit(const atomic<_Tp>* __o, memory_order __m)
+atomic_load_explicit(const atomic<_Tp>* __o, memory_order __m) _NOEXCEPT
 {
     return __o->load(__m);
 }
@@ -908,7 +908,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_exchange(volatile atomic<_Tp>* __o, _Tp __d)
+atomic_exchange(volatile atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     return __o->exchange(__d);
 }
@@ -916,7 +916,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_exchange(atomic<_Tp>* __o, _Tp __d)
+atomic_exchange(atomic<_Tp>* __o, _Tp __d) _NOEXCEPT
 {
     return __o->exchange(__d);
 }
@@ -926,7 +926,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_exchange_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m)
+atomic_exchange_explicit(volatile atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
 {
     return __o->exchange(__d, __m);
 }
@@ -934,7 +934,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp
-atomic_exchange_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m)
+atomic_exchange_explicit(atomic<_Tp>* __o, _Tp __d, memory_order __m) _NOEXCEPT
 {
     return __o->exchange(__d, __m);
 }
@@ -944,7 +944,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_compare_exchange_weak(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d)
+atomic_compare_exchange_weak(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
 {
     return __o->compare_exchange_weak(*__e, __d);
 }
@@ -952,7 +952,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_compare_exchange_weak(atomic<_Tp>* __o, _Tp* __e, _Tp __d)
+atomic_compare_exchange_weak(atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
 {
     return __o->compare_exchange_weak(*__e, __d);
 }
@@ -962,7 +962,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_compare_exchange_strong(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d)
+atomic_compare_exchange_strong(volatile atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
 {
     return __o->compare_exchange_strong(*__e, __d);
 }
@@ -970,7 +970,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_compare_exchange_strong(atomic<_Tp>* __o, _Tp* __e, _Tp __d)
+atomic_compare_exchange_strong(atomic<_Tp>* __o, _Tp* __e, _Tp __d) _NOEXCEPT
 {
     return __o->compare_exchange_strong(*__e, __d);
 }
@@ -982,7 +982,7 @@
 bool
 atomic_compare_exchange_weak_explicit(volatile atomic<_Tp>* __o, _Tp* __e,
                                       _Tp __d,
-                                      memory_order __s, memory_order __f)
+                                      memory_order __s, memory_order __f) _NOEXCEPT
 {
     return __o->compare_exchange_weak(*__e, __d, __s, __f);
 }
@@ -991,7 +991,7 @@
 inline _LIBCPP_INLINE_VISIBILITY
 bool
 atomic_compare_exchange_weak_explicit(atomic<_Tp>* __o, _Tp* __e, _Tp __d,
-                                      memory_order __s, memory_order __f)
+                                      memory_order __s, memory_order __f) _NOEXCEPT
 {
     return __o->compare_exchange_weak(*__e, __d, __s, __f);
 }
@@ -1003,7 +1003,7 @@
 bool
 atomic_compare_exchange_strong_explicit(volatile atomic<_Tp>* __o,
                                         _Tp* __e, _Tp __d,
-                                        memory_order __s, memory_order __f)
+                                        memory_order __s, memory_order __f) _NOEXCEPT
 {
     return __o->compare_exchange_strong(*__e, __d, __s, __f);
 }
@@ -1013,7 +1013,7 @@
 bool
 atomic_compare_exchange_strong_explicit(atomic<_Tp>* __o, _Tp* __e,
                                         _Tp __d,
-                                        memory_order __s, memory_order __f)
+                                        memory_order __s, memory_order __f) _NOEXCEPT
 {
     return __o->compare_exchange_strong(*__e, __d, __s, __f);
 }
@@ -1027,7 +1027,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_add(volatile atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_add(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_add(__op);
 }
@@ -1039,7 +1039,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_add(atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_add(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_add(__op);
 }
@@ -1047,7 +1047,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_add(volatile atomic<_Tp*>* __o, ptrdiff_t __op)
+atomic_fetch_add(volatile atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
 {
     return __o->fetch_add(__op);
 }
@@ -1055,7 +1055,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_add(atomic<_Tp*>* __o, ptrdiff_t __op)
+atomic_fetch_add(atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
 {
     return __o->fetch_add(__op);
 }
@@ -1069,7 +1069,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_add_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_add_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_add(__op, __m);
 }
@@ -1081,7 +1081,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_add_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_add_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_add(__op, __m);
 }
@@ -1090,7 +1090,7 @@
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
 atomic_fetch_add_explicit(volatile atomic<_Tp*>* __o, ptrdiff_t __op,
-                          memory_order __m)
+                          memory_order __m) _NOEXCEPT
 {
     return __o->fetch_add(__op, __m);
 }
@@ -1098,7 +1098,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_add_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m)
+atomic_fetch_add_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_add(__op, __m);
 }
@@ -1112,7 +1112,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_sub(volatile atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_sub(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_sub(__op);
 }
@@ -1124,7 +1124,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_sub(atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_sub(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_sub(__op);
 }
@@ -1132,7 +1132,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_sub(volatile atomic<_Tp*>* __o, ptrdiff_t __op)
+atomic_fetch_sub(volatile atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
 {
     return __o->fetch_sub(__op);
 }
@@ -1140,7 +1140,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_sub(atomic<_Tp*>* __o, ptrdiff_t __op)
+atomic_fetch_sub(atomic<_Tp*>* __o, ptrdiff_t __op) _NOEXCEPT
 {
     return __o->fetch_sub(__op);
 }
@@ -1154,7 +1154,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_sub_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_sub_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_sub(__op, __m);
 }
@@ -1166,7 +1166,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_sub_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_sub_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_sub(__op, __m);
 }
@@ -1175,7 +1175,7 @@
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
 atomic_fetch_sub_explicit(volatile atomic<_Tp*>* __o, ptrdiff_t __op,
-                          memory_order __m)
+                          memory_order __m) _NOEXCEPT
 {
     return __o->fetch_sub(__op, __m);
 }
@@ -1183,7 +1183,7 @@
 template <class _Tp>
 inline _LIBCPP_INLINE_VISIBILITY
 _Tp*
-atomic_fetch_sub_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m)
+atomic_fetch_sub_explicit(atomic<_Tp*>* __o, ptrdiff_t __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_sub(__op, __m);
 }
@@ -1197,7 +1197,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_and(volatile atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_and(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_and(__op);
 }
@@ -1209,7 +1209,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_and(atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_and(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_and(__op);
 }
@@ -1223,7 +1223,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_and_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_and_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_and(__op, __m);
 }
@@ -1235,7 +1235,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_and_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_and_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_and(__op, __m);
 }
@@ -1249,7 +1249,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_or(volatile atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_or(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_or(__op);
 }
@@ -1261,7 +1261,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_or(atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_or(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_or(__op);
 }
@@ -1275,7 +1275,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_or_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_or_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_or(__op, __m);
 }
@@ -1287,7 +1287,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_or_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_or_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_or(__op, __m);
 }
@@ -1301,7 +1301,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_xor(volatile atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_xor(volatile atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_xor(__op);
 }
@@ -1313,7 +1313,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_xor(atomic<_Tp>* __o, _Tp __op)
+atomic_fetch_xor(atomic<_Tp>* __o, _Tp __op) _NOEXCEPT
 {
     return __o->fetch_xor(__op);
 }
@@ -1327,7 +1327,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_xor_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_xor_explicit(volatile atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_xor(__op, __m);
 }
@@ -1339,7 +1339,7 @@
     is_integral<_Tp>::value && !is_same<_Tp, bool>::value,
     _Tp
 >::type
-atomic_fetch_xor_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m)
+atomic_fetch_xor_explicit(atomic<_Tp>* __o, _Tp __op, memory_order __m) _NOEXCEPT
 {
     return __o->fetch_xor(__op, __m);
 }
@@ -1351,22 +1351,22 @@
     _Atomic(bool) __a_;
 
     _LIBCPP_INLINE_VISIBILITY
-    bool test_and_set(memory_order __m = memory_order_seq_cst) volatile
+    bool test_and_set(memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {return __c11_atomic_exchange(&__a_, true, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    bool test_and_set(memory_order __m = memory_order_seq_cst)
+    bool test_and_set(memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {return __c11_atomic_exchange(&__a_, true, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    void clear(memory_order __m = memory_order_seq_cst) volatile
+    void clear(memory_order __m = memory_order_seq_cst) volatile _NOEXCEPT
         {__c11_atomic_store(&__a_, false, __m);}
     _LIBCPP_INLINE_VISIBILITY
-    void clear(memory_order __m = memory_order_seq_cst)
+    void clear(memory_order __m = memory_order_seq_cst) _NOEXCEPT
         {__c11_atomic_store(&__a_, false, __m);}
 
     _LIBCPP_INLINE_VISIBILITY
-    atomic_flag() {} // = default;
+    atomic_flag() _NOEXCEPT {} // = default;
     _LIBCPP_INLINE_VISIBILITY
-    atomic_flag(bool __b) : __a_(__b) {}
+    atomic_flag(bool __b) _NOEXCEPT : __a_(__b) {}
 
 #ifndef _LIBCPP_HAS_NO_DELETED_FUNCTIONS
     atomic_flag(const atomic_flag&) = delete;
@@ -1382,56 +1382,56 @@
 
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_flag_test_and_set(volatile atomic_flag* __o)
+atomic_flag_test_and_set(volatile atomic_flag* __o) _NOEXCEPT
 {
     return __o->test_and_set();
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_flag_test_and_set(atomic_flag* __o)
+atomic_flag_test_and_set(atomic_flag* __o) _NOEXCEPT
 {
     return __o->test_and_set();
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_flag_test_and_set_explicit(volatile atomic_flag* __o, memory_order __m)
+atomic_flag_test_and_set_explicit(volatile atomic_flag* __o, memory_order __m) _NOEXCEPT
 {
     return __o->test_and_set(__m);
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 bool
-atomic_flag_test_and_set_explicit(atomic_flag* __o, memory_order __m)
+atomic_flag_test_and_set_explicit(atomic_flag* __o, memory_order __m) _NOEXCEPT
 {
     return __o->test_and_set(__m);
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_flag_clear(volatile atomic_flag* __o)
+atomic_flag_clear(volatile atomic_flag* __o) _NOEXCEPT
 {
     __o->clear();
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_flag_clear(atomic_flag* __o)
+atomic_flag_clear(atomic_flag* __o) _NOEXCEPT
 {
     __o->clear();
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_flag_clear_explicit(volatile atomic_flag* __o, memory_order __m)
+atomic_flag_clear_explicit(volatile atomic_flag* __o, memory_order __m) _NOEXCEPT
 {
     __o->clear(__m);
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_flag_clear_explicit(atomic_flag* __o, memory_order __m)
+atomic_flag_clear_explicit(atomic_flag* __o, memory_order __m) _NOEXCEPT
 {
     __o->clear(__m);
 }
@@ -1440,14 +1440,14 @@
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_thread_fence(memory_order __m)
+atomic_thread_fence(memory_order __m) _NOEXCEPT
 {
     __c11_atomic_thread_fence(__m);
 }
 
 inline _LIBCPP_INLINE_VISIBILITY
 void
-atomic_signal_fence(memory_order __m)
+atomic_signal_fence(memory_order __m) _NOEXCEPT
 {
     __c11_atomic_signal_fence(__m);
 }





More information about the cfe-commits mailing list