123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368 |
- #ifndef BOOST_ATOMIC_DETAIL_GCC_ALPHA_HPP
- #define BOOST_ATOMIC_DETAIL_GCC_ALPHA_HPP
- // Copyright (c) 2009 Helge Bahmann
- // Copyright (c) 2013 Tim Blechmann
- //
- // Distributed under the Boost Software License, Version 1.0.
- // See accompanying file LICENSE_1_0.txt or copy at
- // http://www.boost.org/LICENSE_1_0.txt)
- #include <boost/atomic/detail/config.hpp>
- #include <boost/atomic/detail/base.hpp>
- #include <boost/atomic/detail/builder.hpp>
- #ifdef BOOST_HAS_PRAGMA_ONCE
- #pragma once
- #endif
- /*
- Refer to http://h71000.www7.hp.com/doc/82final/5601/5601pro_004.html
- (HP OpenVMS systems documentation) and the alpha reference manual.
- */
- /*
- NB: The most natural thing would be to write the increment/decrement
- operators along the following lines:
- __asm__ __volatile__(
- "1: ldl_l %0,%1 \n"
- "addl %0,1,%0 \n"
- "stl_c %0,%1 \n"
- "beq %0,1b\n"
- : "=&b" (tmp)
- : "m" (value)
- : "cc"
- );
- However according to the comments on the HP website and matching
- comments in the Linux kernel sources this defies branch prediction,
- as the cpu assumes that backward branches are always taken; so
- instead copy the trick from the Linux kernel, introduce a forward
- branch and back again.
- I have, however, had a hard time measuring the difference between
- the two versions in microbenchmarks -- I am leaving it in nevertheless
- as it apparently does not hurt either.
- */
- namespace boost {
- namespace atomics {
- namespace detail {
- inline void fence_before(memory_order order)
- {
- switch(order) {
- case memory_order_consume:
- case memory_order_release:
- case memory_order_acq_rel:
- case memory_order_seq_cst:
- __asm__ __volatile__ ("mb" ::: "memory");
- default:;
- }
- }
- inline void fence_after(memory_order order)
- {
- switch(order) {
- case memory_order_acquire:
- case memory_order_acq_rel:
- case memory_order_seq_cst:
- __asm__ __volatile__ ("mb" ::: "memory");
- default:;
- }
- }
- template<>
- inline void platform_atomic_thread_fence(memory_order order)
- {
- switch(order) {
- case memory_order_acquire:
- case memory_order_consume:
- case memory_order_release:
- case memory_order_acq_rel:
- case memory_order_seq_cst:
- __asm__ __volatile__ ("mb" ::: "memory");
- default:;
- }
- }
- template<typename T>
- class atomic_alpha_32
- {
- public:
- typedef T integral_type;
- BOOST_CONSTEXPR atomic_alpha_32(T v) BOOST_NOEXCEPT: i(v) {}
- atomic_alpha_32() {}
- T load(memory_order order=memory_order_seq_cst) const volatile BOOST_NOEXCEPT
- {
- T v=*reinterpret_cast<volatile const int *>(&i);
- fence_after(order);
- return v;
- }
- void store(T v, memory_order order=memory_order_seq_cst) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- *reinterpret_cast<volatile int *>(&i)=(int)v;
- }
- bool compare_exchange_weak(
- T &expected,
- T desired,
- memory_order success_order,
- memory_order failure_order) volatile BOOST_NOEXCEPT
- {
- fence_before(success_order);
- int current, success;
- __asm__ __volatile__(
- "1: ldl_l %2, %4\n"
- "cmpeq %2, %0, %3\n"
- "mov %2, %0\n"
- "beq %3, 3f\n"
- "stl_c %1, %4\n"
- "2:\n"
- ".subsection 2\n"
- "3: mov %3, %1\n"
- "br 2b\n"
- ".previous\n"
- : "+&r" (expected), "+&r" (desired), "=&r"(current), "=&r"(success)
- : "m" (i)
- :
- );
- if (desired) fence_after(success_order);
- else fence_after(failure_order);
- return desired;
- }
- bool is_lock_free(void) const volatile BOOST_NOEXCEPT {return true;}
- protected:
- inline T fetch_add_var(T c, memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- T original, modified;
- __asm__ __volatile__(
- "1: ldl_l %0, %2\n"
- "addl %0, %3, %1\n"
- "stl_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i), "r" (c)
- :
- );
- fence_after(order);
- return original;
- }
- inline T fetch_inc(memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- int original, modified;
- __asm__ __volatile__(
- "1: ldl_l %0, %2\n"
- "addl %0, 1, %1\n"
- "stl_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i)
- :
- );
- fence_after(order);
- return original;
- }
- inline T fetch_dec(memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- int original, modified;
- __asm__ __volatile__(
- "1: ldl_l %0, %2\n"
- "subl %0, 1, %1\n"
- "stl_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i)
- :
- );
- fence_after(order);
- return original;
- }
- private:
- T i;
- };
- template<typename T>
- class atomic_alpha_64
- {
- public:
- typedef T integral_type;
- BOOST_CONSTEXPR atomic_alpha_64(T v) BOOST_NOEXCEPT: i(v) {}
- atomic_alpha_64() {}
- T load(memory_order order=memory_order_seq_cst) const volatile
- {
- T v=*reinterpret_cast<volatile const T *>(&i);
- fence_after(order);
- return v;
- }
- void store(T v, memory_order order=memory_order_seq_cst) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- *reinterpret_cast<volatile T *>(&i)=v;
- }
- bool compare_exchange_weak(
- T &expected,
- T desired,
- memory_order success_order,
- memory_order failure_order) volatile BOOST_NOEXCEPT
- {
- fence_before(success_order);
- int current, success;
- __asm__ __volatile__(
- "1: ldq_l %2, %4\n"
- "cmpeq %2, %0, %3\n"
- "mov %2, %0\n"
- "beq %3, 3f\n"
- "stq_c %1, %4\n"
- "2:\n"
- ".subsection 2\n"
- "3: mov %3, %1\n"
- "br 2b\n"
- ".previous\n"
- : "+&r" (expected), "+&r" (desired), "=&r"(current), "=&r"(success)
- : "m" (i)
- :
- );
- if (desired) fence_after(success_order);
- else fence_after(failure_order);
- return desired;
- }
- bool is_lock_free(void) const volatile BOOST_NOEXCEPT {return true;}
- protected:
- inline T fetch_add_var(T c, memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- T original, modified;
- __asm__ __volatile__(
- "1: ldq_l %0, %2\n"
- "addq %0, %3, %1\n"
- "stq_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i), "r" (c)
- :
- );
- fence_after(order);
- return original;
- }
- inline T fetch_inc(memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- T original, modified;
- __asm__ __volatile__(
- "1: ldq_l %0, %2\n"
- "addq %0, 1, %1\n"
- "stq_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i)
- :
- );
- fence_after(order);
- return original;
- }
- inline T fetch_dec(memory_order order) volatile BOOST_NOEXCEPT
- {
- fence_before(order);
- T original, modified;
- __asm__ __volatile__(
- "1: ldq_l %0, %2\n"
- "subq %0, 1, %1\n"
- "stq_c %1, %2\n"
- "beq %1, 2f\n"
- ".subsection 2\n"
- "2: br 1b\n"
- ".previous\n"
- : "=&r" (original), "=&r" (modified)
- : "m" (i)
- :
- );
- fence_after(order);
- return original;
- }
- private:
- T i;
- };
- template<typename T>
- class platform_atomic_integral<T, 4> :
- public build_atomic_from_typical<build_exchange<atomic_alpha_32<T> > >
- {
- public:
- typedef build_atomic_from_typical<build_exchange<atomic_alpha_32<T> > > super;
- BOOST_CONSTEXPR platform_atomic_integral(T v) BOOST_NOEXCEPT: super(v) {}
- platform_atomic_integral(void) {}
- };
- template<typename T>
- class platform_atomic_integral<T, 8> :
- public build_atomic_from_typical<build_exchange<atomic_alpha_64<T> > >
- {
- public:
- typedef build_atomic_from_typical<build_exchange<atomic_alpha_64<T> > > super;
- BOOST_CONSTEXPR platform_atomic_integral(T v) BOOST_NOEXCEPT: super(v) {}
- platform_atomic_integral(void) {}
- };
- template<typename T>
- class platform_atomic_integral<T, 1> :
- public build_atomic_from_larger_type<atomic_alpha_32<uint32_t>, T>
- {
- public:
- typedef build_atomic_from_larger_type<atomic_alpha_32<uint32_t>, T> super;
- BOOST_CONSTEXPR platform_atomic_integral(T v) BOOST_NOEXCEPT: super(v) {}
- platform_atomic_integral(void) {}
- };
- template<typename T>
- class platform_atomic_integral<T, 2> :
- public build_atomic_from_larger_type<atomic_alpha_32<uint32_t>, T>
- {
- public:
- typedef build_atomic_from_larger_type<atomic_alpha_32<uint32_t>, T> super;
- BOOST_CONSTEXPR platform_atomic_integral(T v) BOOST_NOEXCEPT: super(v) {}
- platform_atomic_integral(void) {}
- };
- }
- }
- }
- #endif
|