2
0
mirror of https://github.com/boostorg/math.git synced 2026-02-24 04:02:18 +00:00

Remove boost.atomic dependency

This commit is contained in:
Matt Borland
2021-02-27 21:49:20 +03:00
parent 5531bf750e
commit eb8097db49
2 changed files with 32 additions and 71 deletions

View File

@@ -9,7 +9,7 @@
#include <sstream>
#include <algorithm>
#include <vector>
#include <boost/atomic.hpp>
#include <atomic>
#include <functional>
#include <future>
#include <thread>
@@ -161,9 +161,9 @@ public:
Real inv_denom = 1/static_cast<Real>(((gen.max)()-(gen.min)()));
m_num_threads = (std::max)(m_num_threads, (uint64_t) 1);
m_thread_calls.reset(new boost::atomic<uint64_t>[threads]);
m_thread_Ss.reset(new boost::atomic<Real>[threads]);
m_thread_averages.reset(new boost::atomic<Real>[threads]);
m_thread_calls.reset(new std::atomic<uint64_t>[threads]);
m_thread_Ss.reset(new std::atomic<Real>[threads]);
m_thread_averages.reset(new std::atomic<Real>[threads]);
Real avg = 0;
for (uint64_t i = 0; i < m_num_threads; ++i)
@@ -306,27 +306,27 @@ private:
uint64_t total_calls = 0;
for (uint64_t i = 0; i < m_num_threads; ++i)
{
uint64_t t_calls = m_thread_calls[i].load(boost::memory_order::consume);
uint64_t t_calls = m_thread_calls[i].load(std::memory_order_consume);
total_calls += t_calls;
}
Real variance = 0;
Real avg = 0;
for (uint64_t i = 0; i < m_num_threads; ++i)
{
uint64_t t_calls = m_thread_calls[i].load(boost::memory_order::consume);
uint64_t t_calls = m_thread_calls[i].load(std::memory_order_consume);
// Will this overflow? Not hard to remove . . .
avg += m_thread_averages[i].load(boost::memory_order::relaxed)*((Real)t_calls / (Real)total_calls);
variance += m_thread_Ss[i].load(boost::memory_order::relaxed);
avg += m_thread_averages[i].load(std::memory_order_relaxed)*((Real)t_calls / (Real)total_calls);
variance += m_thread_Ss[i].load(std::memory_order_relaxed);
}
m_avg.store(avg, boost::memory_order::release);
m_variance.store(variance / (total_calls - 1), boost::memory_order::release);
m_avg.store(avg, std::memory_order_release);
m_variance.store(variance / (total_calls - 1), std::memory_order_release);
m_total_calls = total_calls; // relaxed store, it's just for user feedback
// Allow cancellation:
if (m_done) // relaxed load
{
break;
}
} while (m_total_calls < 2048 || this->current_error_estimate() > m_error_goal.load(boost::memory_order::consume));
} while (m_total_calls < 2048 || this->current_error_estimate() > m_error_goal.load(std::memory_order_consume));
// Error bound met; signal the threads:
m_done = true; // relaxed store, threads will get the message in the end
std::for_each(threads.begin(), threads.end(),
@@ -339,7 +339,7 @@ private:
uint64_t total_calls = 0;
for (uint64_t i = 0; i < m_num_threads; ++i)
{
uint64_t t_calls = m_thread_calls[i].load(boost::memory_order::consume);
uint64_t t_calls = m_thread_calls[i].load(std::memory_order_consume);
total_calls += t_calls;
}
Real variance = 0;
@@ -347,13 +347,13 @@ private:
for (uint64_t i = 0; i < m_num_threads; ++i)
{
uint64_t t_calls = m_thread_calls[i].load(boost::memory_order::consume);
uint64_t t_calls = m_thread_calls[i].load(std::memory_order_consume);
// Averages weighted by the number of calls the thread made:
avg += m_thread_averages[i].load(boost::memory_order::relaxed)*((Real)t_calls / (Real)total_calls);
variance += m_thread_Ss[i].load(boost::memory_order::relaxed);
avg += m_thread_averages[i].load(std::memory_order_relaxed)*((Real)t_calls / (Real)total_calls);
variance += m_thread_Ss[i].load(std::memory_order_relaxed);
}
m_avg.store(avg, boost::memory_order::release);
m_variance.store(variance / (total_calls - 1), boost::memory_order::release);
m_avg.store(avg, std::memory_order_release);
m_variance.store(variance / (total_calls - 1), std::memory_order_release);
m_total_calls = total_calls; // relaxed store, this is just user feedback
// Sometimes, the master will observe the variance at a very "good" (or bad?) moment,
@@ -362,7 +362,7 @@ private:
}
while ((--max_repeat_tries >= 0) && (this->current_error_estimate() > m_error_goal));
return m_avg.load(boost::memory_order::consume);
return m_avg.load(std::memory_order_consume);
}
void m_thread_monte(uint64_t thread_index, uint64_t seed)
@@ -373,14 +373,14 @@ private:
std::vector<Real> x(m_lbs.size());
RandomNumberGenerator gen(seed);
Real inv_denom = (Real) 1/(Real)( (gen.max)() - (gen.min)() );
Real M1 = m_thread_averages[thread_index].load(boost::memory_order::consume);
Real S = m_thread_Ss[thread_index].load(boost::memory_order::consume);
Real M1 = m_thread_averages[thread_index].load(std::memory_order_consume);
Real S = m_thread_Ss[thread_index].load(std::memory_order_consume);
// Kahan summation is required or the value of the integrand will go on a random walk during long computations.
// See the implementation discussion.
// The idea is that the unstabilized additions have error sigma(f)/sqrt(N) + epsilon*N, which diverges faster than it converges!
// Kahan summation turns this to sigma(f)/sqrt(N) + epsilon^2*N, and the random walk occurs on a timescale of 10^14 years (on current hardware)
Real compensator = 0;
uint64_t k = m_thread_calls[thread_index].load(boost::memory_order::consume);
uint64_t k = m_thread_calls[thread_index].load(std::memory_order_consume);
while (!m_done) // relaxed load
{
int j = 0;
@@ -418,9 +418,9 @@ private:
S += (f - M1)*(f - M2);
M1 = M2;
}
m_thread_averages[thread_index].store(M1, boost::memory_order::release);
m_thread_Ss[thread_index].store(S, boost::memory_order::release);
m_thread_calls[thread_index].store(k, boost::memory_order::release);
m_thread_averages[thread_index].store(M1, std::memory_order_release);
m_thread_Ss[thread_index].store(S, std::memory_order_release);
m_thread_calls[thread_index].store(k, std::memory_order_release);
}
}
catch (...)
@@ -434,20 +434,20 @@ private:
std::function<Real(std::vector<Real> &)> m_integrand;
uint64_t m_num_threads;
uint64_t m_seed;
boost::atomic<Real> m_error_goal;
boost::atomic<bool> m_done;
std::atomic<Real> m_error_goal;
std::atomic<bool> m_done;
std::vector<Real> m_lbs;
std::vector<Real> m_dxs;
std::vector<detail::limit_classification> m_limit_types;
Real m_volume;
boost::atomic<uint64_t> m_total_calls;
std::atomic<uint64_t> m_total_calls;
// I wanted these to be vectors rather than maps,
// but you can't resize a vector of atomics.
std::unique_ptr<boost::atomic<uint64_t>[]> m_thread_calls;
boost::atomic<Real> m_variance;
std::unique_ptr<boost::atomic<Real>[]> m_thread_Ss;
boost::atomic<Real> m_avg;
std::unique_ptr<boost::atomic<Real>[]> m_thread_averages;
std::unique_ptr<std::atomic<uint64_t>[]> m_thread_calls;
std::atomic<Real> m_variance;
std::unique_ptr<std::atomic<Real>[]> m_thread_Ss;
std::atomic<Real> m_avg;
std::unique_ptr<std::atomic<Real>[]> m_thread_averages;
std::chrono::time_point<std::chrono::system_clock> m_start;
std::exception_ptr m_exception;
};

View File

@@ -43,45 +43,6 @@ namespace boost {
#endif
}
}}
#else // BOOST_NO_CXX11_HDR_ATOMIC
//
// We need Boost.Atomic, but on any platform that supports auto-linking we do
// not need to link against a separate library:
//
#define BOOST_ATOMIC_NO_LIB
#include <boost/atomic.hpp>
# define BOOST_MATH_ATOMIC_NS boost
namespace boost{ namespace math{ namespace detail{
//
// We need a type to use as an atomic counter:
//
#if BOOST_ATOMIC_INT_LOCK_FREE == 2
typedef boost::atomic<int> atomic_counter_type;
typedef boost::atomic<unsigned> atomic_unsigned_type;
typedef int atomic_integer_type;
typedef unsigned atomic_unsigned_integer_type;
#elif BOOST_ATOMIC_SHORT_LOCK_FREE == 2
typedef boost::atomic<short> atomic_counter_type;
typedef boost::atomic<unsigned short> atomic_unsigned_type;
typedef short atomic_integer_type;
typedef unsigned short atomic_unsigned_integer_type;
#elif BOOST_ATOMIC_LONG_LOCK_FREE == 2
typedef boost::atomic<long> atomic_counter_type;
typedef boost::atomic<unsigned long> atomic_unsigned_type;
typedef long atomic_integer_type;
typedef unsigned long atomic_unsigned_integer_type;
#elif BOOST_ATOMIC_LLONG_LOCK_FREE == 2
typedef boost::atomic<long long> atomic_counter_type;
typedef boost::atomic<unsigned long long> atomic_unsigned_type;
typedef long long atomic_integer_type;
typedef unsigned long long atomic_unsigned_integer_type;
#else
# define BOOST_MATH_NO_ATOMIC_INT
#endif
}}} // namespaces
#endif // BOOST_NO_CXX11_HDR_ATOMIC