}
/// @}
+#if __cplusplus >= 202002L
+# define __cpp_lib_atomic_shared_ptr 201711L
+ template<typename _Tp>
+ class atomic;
+
+ template<typename _Up>
+ static constexpr bool __is_shared_ptr = false;
+ template<typename _Up>
+ static constexpr bool __is_shared_ptr<shared_ptr<_Up>> = true;
+
+ template<typename _Tp>
+ class _Sp_atomic
+ {
+ using value_type = _Tp;
+
+ friend class atomic<_Tp>;
+
+ // An atomic version of __shared_count<> and __weak_count<>.
+ // Stores a _Sp_counted_base<>* but uses the LSB as a lock.
+ struct _Atomic_count
+ {
+ // Either __shared_count<> or __weak_count<>
+ using __count_type = decltype(_Tp::_M_refcount);
+
+ // _Sp_counted_base<>*
+ using pointer = decltype(__count_type::_M_pi);
+
+ // Ensure we can use the LSB as the lock bit.
+ static_assert(alignof(remove_pointer_t<pointer>) > 1);
+
+ _Atomic_count() : _M_val(0) { }
+
+ explicit
+ _Atomic_count(__count_type&& __c) noexcept
+ : _M_val(reinterpret_cast<uintptr_t>(__c._M_pi))
+ {
+ __c._M_pi = nullptr;
+ }
+
+ ~_Atomic_count()
+ {
+ auto __val = _M_val.load(memory_order_relaxed);
+ __glibcxx_assert(!(__val & _S_lock_bit));
+ if (auto __pi = reinterpret_cast<pointer>(__val))
+ {
+ if constexpr (__is_shared_ptr<_Tp>)
+ __pi->_M_release();
+ else
+ __pi->_M_weak_release();
+ }
+ }
+
+ _Atomic_count(const _Atomic_count&) = delete;
+ _Atomic_count& operator=(const _Atomic_count&) = delete;
+
+ // Precondition: Caller does not hold lock!
+ // Returns the raw pointer value without the lock bit set.
+ pointer
+ lock(memory_order __o) const noexcept
+ {
+ // To acquire the lock we flip the LSB from 0 to 1.
+
+ auto __current = _M_val.load(memory_order_relaxed);
+ while (__current & _S_lock_bit)
+ {
+ __detail::__thread_relax();
+ __current = _M_val.load(memory_order_relaxed);
+ }
+
+ while (!_M_val.compare_exchange_strong(__current,
+ __current | _S_lock_bit,
+ __o,
+ memory_order_relaxed))
+ {
+ __detail::__thread_relax();
+ __current = __current & ~_S_lock_bit;
+ }
+ return reinterpret_cast<pointer>(__current);
+ }
+
+ // Precondition: caller holds lock!
+ void
+ unlock(memory_order __o) const noexcept
+ {
+ _M_val.fetch_sub(1, __o);
+ }
+
+ // Swaps the values of *this and __c, and unlocks *this.
+ // Precondition: caller holds lock!
+ void
+ _M_swap_unlock(__count_type& __c, memory_order __o) noexcept
+ {
+ if (__o != memory_order_seq_cst)
+ __o = memory_order_release;
+ auto __x = reinterpret_cast<uintptr_t>(__c._M_pi);
+ __x = _M_val.exchange(__x, __o);
+ __c._M_pi = reinterpret_cast<pointer>(__x & ~_S_lock_bit);
+ }
+
+#if __cpp_lib_atomic_wait
+ // Precondition: caller holds lock!
+ void
+ _M_wait_unlock(memory_order __o) const noexcept
+ {
+ auto __v = _M_val.fetch_sub(1, memory_order_relaxed);
+ _M_val.wait(__v & ~_S_lock_bit, __o);
+ }
+
+ void
+ notify_one() noexcept
+ {
+ _M_val.notify_one();
+ }
+
+ void
+ notify_all() noexcept
+ {
+ _M_val.notify_all();
+ }
+#endif
+
+ private:
+ mutable __atomic_base<uintptr_t> _M_val{0};
+ static constexpr uintptr_t _S_lock_bit{1};
+ };
+
+ typename _Tp::element_type* _M_ptr;
+ _Atomic_count _M_refcount;
+
+ static _Atomic_count::pointer
+ _S_add_ref(_Atomic_count::pointer __p)
+ {
+ if (__p)
+ {
+ if constexpr (__is_shared_ptr<_Tp>)
+ __p->_M_add_ref_copy();
+ else
+ __p->_M_weak_add_ref();
+ }
+ return __p;
+ }
+
+ constexpr _Sp_atomic() noexcept = default;
+
+ explicit
+ _Sp_atomic(value_type __r) noexcept
+ : _M_ptr(__r._M_ptr), _M_refcount(std::move(__r._M_refcount))
+ { }
+
+ ~_Sp_atomic() = default;
+
+ _Sp_atomic(const _Sp_atomic&) = delete;
+ void operator=(const _Sp_atomic&) = delete;
+
+ value_type
+ load(memory_order __o) const noexcept
+ {
+ __glibcxx_assert(__o != memory_order_release
+ && __o != memory_order_acq_rel);
+ // Ensure that the correct value of _M_ptr is visible after locking.,
+ // by upgrading relaxed or consume to acquire.
+ if (__o != memory_order_seq_cst)
+ __o = memory_order_acquire;
+
+ value_type __ret;
+ auto __pi = _M_refcount.lock(__o);
+ __ret._M_ptr = _M_ptr;
+ __ret._M_refcount._M_pi = _S_add_ref(__pi);
+ _M_refcount.unlock(memory_order_relaxed);
+ return __ret;
+ }
+
+ void
+ swap(value_type& __r, memory_order __o) noexcept
+ {
+ _M_refcount.lock(memory_order_acquire);
+ std::swap(_M_ptr, __r._M_ptr);
+ _M_refcount._M_swap_unlock(__r._M_refcount, __o);
+ }
+
+ bool
+ compare_exchange_strong(value_type& __expected, value_type __desired,
+ memory_order __o, memory_order __o2) noexcept
+ {
+ bool __result = true;
+ auto __pi = _M_refcount.lock(memory_order_acquire);
+ if (_M_ptr == __expected._M_ptr
+ && __pi == __expected._M_refcount._M_pi)
+ {
+ _M_ptr = __desired._M_ptr;
+ _M_refcount._M_swap_unlock(__desired._M_refcount, __o);
+ }
+ else
+ {
+ _Tp __sink = std::move(__expected);
+ __expected._M_ptr = _M_ptr;
+ __expected._M_refcount._M_pi = _S_add_ref(__pi);
+ _M_refcount.unlock(__o2);
+ __result = false;
+ }
+ return __result;
+ }
+
+#if __cpp_lib_atomic_wait
+ void
+ wait(value_type __old, memory_order __o) const noexcept
+ {
+ auto __pi = _M_refcount.lock(memory_order_acquire);
+ if (_M_ptr == __old._M_ptr && __pi == __old._M_refcount._M_pi)
+ _M_refcount._M_wait_unlock(__o);
+ else
+ _M_refcount.unlock(memory_order_relaxed);
+ }
+
+ void
+ notify_one() noexcept
+ {
+ _M_refcount.notify_one();
+ }
+
+ void
+ notify_all() noexcept
+ {
+ _M_refcount.notify_all();
+ }
+#endif
+ };
+
+ template<typename _Tp>
+ class atomic<shared_ptr<_Tp>>
+ {
+ public:
+ using value_type = shared_ptr<_Tp>;
+
+ static constexpr bool is_always_lock_free = false;
+
+ bool
+ is_lock_free() const noexcept
+ { return false; }
+
+ constexpr atomic() noexcept = default;
+
+ atomic(shared_ptr<_Tp> __r) noexcept
+ : _M_impl(std::move(__r))
+ { }
+
+ atomic(const atomic&) = delete;
+ void operator=(const atomic&) = delete;
+
+ shared_ptr<_Tp>
+ load(memory_order __o = memory_order_seq_cst) const noexcept
+ { return _M_impl.load(__o); }
+
+ operator shared_ptr<_Tp>() const noexcept
+ { return _M_impl.load(memory_order_seq_cst); }
+
+ void
+ store(shared_ptr<_Tp> __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ { _M_impl.swap(__desired, __o); }
+
+ void
+ operator=(shared_ptr<_Tp> __desired) noexcept
+ { _M_impl.swap(__desired, memory_order_seq_cst); }
+
+ shared_ptr<_Tp>
+ exchange(shared_ptr<_Tp> __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ _M_impl.swap(__desired, __o);
+ return __desired;
+ }
+
+ bool
+ compare_exchange_strong(shared_ptr<_Tp>& __expected,
+ shared_ptr<_Tp> __desired,
+ memory_order __o, memory_order __o2) noexcept
+ {
+ return _M_impl.compare_exchange_strong(__expected, __desired, __o, __o2);
+ }
+
+ bool
+ compare_exchange_strong(value_type& __expected, value_type __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ memory_order __o2;
+ switch (__o)
+ {
+ case memory_order_acq_rel:
+ __o2 = memory_order_acquire;
+ break;
+ case memory_order_release:
+ __o2 = memory_order_relaxed;
+ break;
+ default:
+ __o2 = __o;
+ }
+ return compare_exchange_strong(__expected, std::move(__desired),
+ __o, __o2);
+ }
+
+ bool
+ compare_exchange_weak(value_type& __expected, value_type __desired,
+ memory_order __o, memory_order __o2) noexcept
+ {
+ return compare_exchange_strong(__expected, std::move(__desired),
+ __o, __o2);
+ }
+
+ bool
+ compare_exchange_weak(value_type& __expected, value_type __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ return compare_exchange_strong(__expected, std::move(__desired), __o);
+ }
+
+#if __cpp_lib_atomic_wait
+ void
+ wait(value_type __old,
+ memory_order __o = memory_order_seq_cst) const noexcept
+ {
+ _M_impl.wait(std::move(__old), __o);
+ }
+
+ void
+ notify_one() noexcept
+ {
+ _M_impl.notify_one();
+ }
+
+ void
+ notify_all() noexcept
+ {
+ _M_impl.notify_all();
+ }
+#endif
+
+ private:
+ _Sp_atomic<shared_ptr<_Tp>> _M_impl;
+ };
+
+ template<typename _Tp>
+ class atomic<weak_ptr<_Tp>>
+ {
+ public:
+ using value_type = weak_ptr<_Tp>;
+
+ static constexpr bool is_always_lock_free = false;
+
+ bool
+ is_lock_free() const noexcept
+ { return false; }
+
+ constexpr atomic() noexcept = default;
+
+ atomic(weak_ptr<_Tp> __r) noexcept
+ : _M_impl(move(__r))
+ { }
+
+ atomic(const atomic&) = delete;
+ void operator=(const atomic&) = delete;
+
+ weak_ptr<_Tp>
+ load(memory_order __o = memory_order_seq_cst) const noexcept
+ { return _M_impl.load(__o); }
+
+ operator weak_ptr<_Tp>() const noexcept
+ { return _M_impl.load(memory_order_seq_cst); }
+
+ void
+ store(weak_ptr<_Tp> __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ { _M_impl.swap(__desired, __o); }
+
+ void
+ operator=(weak_ptr<_Tp> __desired) noexcept
+ { _M_impl.swap(__desired, memory_order_seq_cst); }
+
+ weak_ptr<_Tp>
+ exchange(weak_ptr<_Tp> __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ _M_impl.swap(__desired, __o);
+ return __desired;
+ }
+
+ bool
+ compare_exchange_strong(weak_ptr<_Tp>& __expected,
+ weak_ptr<_Tp> __desired,
+ memory_order __o, memory_order __o2) noexcept
+ {
+ return _M_impl.compare_exchange_strong(__expected, __desired, __o, __o2);
+ }
+
+ bool
+ compare_exchange_strong(value_type& __expected, value_type __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ memory_order __o2;
+ switch (__o)
+ {
+ case memory_order_acq_rel:
+ __o2 = memory_order_acquire;
+ break;
+ case memory_order_release:
+ __o2 = memory_order_relaxed;
+ break;
+ default:
+ __o2 = __o;
+ }
+ return compare_exchange_strong(__expected, std::move(__desired),
+ __o, __o2);
+ }
+
+ bool
+ compare_exchange_weak(value_type& __expected, value_type __desired,
+ memory_order __o, memory_order __o2) noexcept
+ {
+ return compare_exchange_strong(__expected, std::move(__desired),
+ __o, __o2);
+ }
+
+ bool
+ compare_exchange_weak(value_type& __expected, value_type __desired,
+ memory_order __o = memory_order_seq_cst) noexcept
+ {
+ return compare_exchange_strong(__expected, std::move(__desired), __o);
+ }
+
+#if __cpp_lib_atomic_wait
+ void
+ wait(value_type __old,
+ memory_order __o = memory_order_seq_cst) const noexcept
+ {
+ _M_impl.wait(std::move(__old), __o);
+ }
+
+ void
+ notify_one() noexcept
+ {
+ _M_impl.notify_one();
+ }
+
+ void
+ notify_all() noexcept
+ {
+ _M_impl.notify_all();
+ }
+#endif
+
+ private:
+ _Sp_atomic<weak_ptr<_Tp>> _M_impl;
+ };
+#endif // C++20
+
/// @} relates shared_ptr
/// @} group pointer_abstractions
--- /dev/null
+// { dg-options "-std=gnu++20" }
+// { dg-do run { target c++20 } }
+// { dg-require-effective-target gthreads }
+// { dg-additional-options "-pthread" { target pthread } }
+// { dg-add-options libatomic }
+
+#include <memory>
+
+#ifndef __cpp_lib_atomic_shared_ptr
+# error "Feature-test macro for atomic<shared_ptr<T>> missing in <memory>"
+#elif __cpp_lib_atomic_shared_ptr != 201711L
+# error "Feature-test macro for atomic<shared_ptr<T>> has wrong value in <memory>"
+#endif
+
+#include <thread>
+
+#include <testsuite_hooks.h>
+
+void
+test_is_lock_free()
+{
+ using test_type = std::atomic<std::shared_ptr<int>>;
+ static_assert( test_type::is_always_lock_free == false );
+
+ test_type p;
+ VERIFY( p.is_lock_free() == false );
+}
+
+void
+test_atomic_shared_ptr()
+{
+ struct A { int a; int b; };
+
+ auto a = std::make_shared<A>( 0, 42 );
+ using ptr_t = std::shared_ptr<A>;
+ {
+ std::atomic<ptr_t> p{ };
+ VERIFY( p.load().get() == nullptr );
+ }
+
+ std::atomic<ptr_t> p{ a };
+ VERIFY( p.load().get() == a.get() );
+ auto b = std::make_shared<A>( 42, 0 );
+ p.store(b);
+ VERIFY( p.load().get() != a.get() );
+ VERIFY( p.load().get() == b.get() );
+ p.exchange(a);
+ VERIFY( p.load().get() != b.get() );
+ VERIFY( p.load().get() == a.get() );
+
+ {
+ ptr_t aa{ a };
+ VERIFY( p.compare_exchange_strong(aa, b,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == true );
+ ptr_t bb{ a };
+ VERIFY( p.compare_exchange_strong(bb, b,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == false );
+ VERIFY( bb.get() == b.get() );
+ }
+
+ {
+ ptr_t bb{ b };
+ VERIFY( p.compare_exchange_weak(bb, a,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == true );
+ ptr_t aa{ b };
+ VERIFY( p.compare_exchange_weak(aa, a,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == false );
+ VERIFY( aa.get() == a.get() );
+ }
+}
+
+void
+test_wait_notify()
+{
+ std::atomic<std::shared_ptr<int>> p;
+ std::shared_ptr<int> a = std::make_shared<int>();
+ std::shared_ptr<int> b = std::make_shared<int>();
+
+ p.store(a);
+ p.wait(b);
+ std::thread t([&]
+ {
+ p.store(b);
+ p.notify_one();
+ });
+ p.wait(a);
+ t.join();
+}
+
+int counter = 0;
+
+void
+test_counting()
+{
+ struct X
+ {
+ ~X() { ++counter; }
+ };
+
+ {
+ std::atomic<std::shared_ptr<X>> p{ std::make_shared<X>() };
+ std::shared_ptr<X> a = p.load();
+ VERIFY( a.use_count() == 2 ); // p, a
+ p.store({});
+ VERIFY( a.use_count() == 1 ); // a
+ p.store(a);
+ VERIFY( a.use_count() == 2 ); // p, a
+ std::shared_ptr<X> b = std::make_shared<X>();
+ std::shared_ptr<X> c = p.exchange(b);
+ VERIFY( a.use_count() == 2 ); // a, c
+ VERIFY( c == a );
+ VERIFY( b.use_count() == 2 ); // p, b
+ std::atomic<std::shared_ptr<X>> p2{a};
+ VERIFY( a.use_count() == 3 ); // p2, a, c
+ VERIFY( p2.compare_exchange_strong(a, b) );
+ VERIFY( a.use_count() == 2 ); // a, c
+ VERIFY( b.use_count() == 3 ); // p, p2, b
+ VERIFY ( ! p2.compare_exchange_strong(a, b) );
+ VERIFY( a == b );
+ VERIFY( a.use_count() == 4 ); // p, p2, a, b
+ VERIFY( b.use_count() == 4 );
+ VERIFY( c.use_count() == 1 ); // c
+ VERIFY( p.compare_exchange_weak(b, c) );
+ VERIFY( b.use_count() == 3 ); // p2, a, b
+ VERIFY( c.use_count() == 2 ); // p, c
+ VERIFY( ! p.compare_exchange_weak(a, b) );
+ VERIFY( a == c );
+ VERIFY( a.use_count() == 3 ); // p, a, c
+ VERIFY( b.use_count() == 2 ); // p2, b
+ VERIFY( c.use_count() == 3 ); // p, a, c
+ a.reset();
+ b.reset();
+ c.reset();
+ VERIFY( counter == 0 );
+ }
+ VERIFY( counter == 2 );
+}
+
+int
+main()
+{
+ test_is_lock_free();
+ test_atomic_shared_ptr();
+ test_wait_notify();
+ test_counting();
+}
--- /dev/null
+// { dg-options "-std=gnu++20" }
+// { dg-do run { target c++20 } }
+// { dg-require-effective-target gthreads }
+// { dg-additional-options "-pthread" { target pthread } }
+// { dg-add-options libatomic }
+
+#include <memory>
+#include <thread>
+#include <testsuite_hooks.h>
+
+void
+test_is_lock_free()
+{
+ using test_type = std::atomic<std::weak_ptr<int>>;
+ static_assert( test_type::is_always_lock_free == false );
+
+ test_type p;
+ VERIFY( p.is_lock_free() == false );
+}
+
+void
+test_atomic_weak_ptr()
+{
+ struct A { int a; int b; };
+
+ auto a = std::make_shared<A>( 0, 42 );
+ using ptr_t = std::weak_ptr<A>;
+ ptr_t wa{ a };
+ {
+ std::atomic<ptr_t> p{ };
+ VERIFY( p.load().lock().get() == nullptr );
+ }
+
+ std::atomic<ptr_t> p{ wa };
+ VERIFY( p.load().lock().get() == a.get() );
+
+ auto b = std::make_shared<A>( 42, 0 );
+ ptr_t wb{ b };
+ p.store(wb);
+ VERIFY( p.load().lock().get() != a.get() );
+ VERIFY( p.load().lock().get() == b.get() );
+ p.exchange(wa);
+ VERIFY( p.load().lock().get() != b.get() );
+ VERIFY( p.load().lock().get() == a.get() );
+
+ {
+ ptr_t aa{ a };
+ VERIFY( p.compare_exchange_strong(aa, b,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == true );
+ ptr_t bb{ a };
+ VERIFY( p.compare_exchange_strong(bb, b,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == false );
+ VERIFY( bb.lock().get() == b.get() );
+ }
+
+ {
+ ptr_t bb{ b };
+ VERIFY( p.compare_exchange_weak(bb, a,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == true );
+ ptr_t aa{ b };
+ VERIFY( p.compare_exchange_weak(aa, a,
+ std::memory_order_seq_cst,
+ std::memory_order_seq_cst) == false );
+ VERIFY( aa.lock().get() == a.get() );
+ }
+}
+
+void
+test_wait_notify()
+{
+ std::atomic<std::weak_ptr<int>> p;
+ std::weak_ptr<int> a = std::make_shared<int>();
+ std::weak_ptr<int> b = std::make_shared<int>();
+
+ p.store(a);
+ p.wait(b);
+ std::thread t([&]
+ {
+ p.store(b);
+ p.notify_one();
+ });
+ p.wait(a);
+ t.join();
+}
+
+int
+main()
+{
+ test_is_lock_free();
+ test_atomic_weak_ptr();
+ test_wait_notify();
+}