31#ifndef ETL_SPSC_QUEUE_ATOMIC_INCLUDED
32#define ETL_SPSC_QUEUE_ATOMIC_INCLUDED
50 template <
size_t Memory_Model = etl::memory_model::MEMORY_MODEL_LARGE>
51 class queue_spsc_atomic_base
56 typedef typename etl::size_type_lookup<Memory_Model>::type size_type;
65 return read.load(etl::memory_order_acquire) ==
write.load(etl::memory_order_acquire);
75 size_type next_index = get_next_index(
write.load(etl::memory_order_acquire), Reserved);
77 return (next_index ==
read.load(etl::memory_order_acquire));
84 size_type
size()
const
86 size_type write_index =
write.load(etl::memory_order_acquire);
87 size_type read_index =
read.load(etl::memory_order_acquire);
91 if (write_index >= read_index)
93 n = write_index - read_index;
97 n = Reserved - read_index + write_index;
107 size_type available()
const
109 return Reserved -
size() - 1;
115 size_type capacity()
const
130 queue_spsc_atomic_base(size_type reserved_)
140 static size_type get_next_index(size_type index, size_type maximum)
144 if (index == maximum) ETL_UNLIKELY
152 etl::atomic<size_type>
write;
153 etl::atomic<size_type>
read;
154 const size_type Reserved;
161#if defined(ETL_POLYMORPHIC_SPSC_QUEUE_ATOMIC) || defined(ETL_POLYMORPHIC_CONTAINERS)
163 virtual ~queue_spsc_atomic_base()
168 ~queue_spsc_atomic_base()
185 template <
typename T, const
size_t Memory_Model = etl::memory_model::MEMORY_MODEL_LARGE>
186 class iqueue_spsc_atomic :
public queue_spsc_atomic_base<Memory_Model>
190 typedef typename etl::queue_spsc_atomic_base<Memory_Model> base_t;
194 typedef T value_type;
195 typedef T& reference;
196 typedef const T& const_reference;
198 typedef T&& rvalue_reference;
200 typedef typename base_t::size_type size_type;
204 using base_t::Reserved;
205 using base_t::get_next_index;
210 bool push(const_reference value)
212 size_type write_index =
write.load(etl::memory_order_relaxed);
213 size_type next_index = get_next_index(write_index, Reserved);
215 if (next_index !=
read.load(etl::memory_order_acquire))
217 ::new (&p_buffer[write_index]) T(value);
219 write.store(next_index, etl::memory_order_release);
228#if ETL_USING_CPP11 && ETL_NOT_USING_STLPORT && !defined(ETL_QUEUE_ATOMIC_FORCE_CPP03_IMPLEMENTATION)
232 bool push(rvalue_reference value)
234 size_type write_index =
write.load(etl::memory_order_relaxed);
235 size_type next_index = get_next_index(write_index, Reserved);
237 if (next_index !=
read.load(etl::memory_order_acquire))
239 ::new (&p_buffer[write_index]) T(etl::move(value));
241 write.store(next_index, etl::memory_order_release);
251#if ETL_USING_CPP11 && ETL_NOT_USING_STLPORT && !defined(ETL_QUEUE_ATOMIC_FORCE_CPP03_IMPLEMENTATION)
256 template <
typename ... Args>
257 bool emplace(Args&&... args)
259 size_type write_index =
write.load(etl::memory_order_relaxed);
260 size_type next_index = get_next_index(write_index, Reserved);
262 if (next_index !=
read.load(etl::memory_order_acquire))
264 ::new (&p_buffer[write_index]) T(etl::forward<Args>(args)...);
266 write.store(next_index, etl::memory_order_release);
281 size_type write_index =
write.load(etl::memory_order_relaxed);
282 size_type next_index = get_next_index(write_index, Reserved);
284 if (next_index !=
read.load(etl::memory_order_acquire))
286 ::new (&p_buffer[write_index]) T();
288 write.store(next_index, etl::memory_order_release);
301 template <
typename T1>
302 bool emplace(
const T1& value1)
304 size_type write_index =
write.load(etl::memory_order_relaxed);
305 size_type next_index = get_next_index(write_index, Reserved);
307 if (next_index !=
read.load(etl::memory_order_acquire))
309 ::new (&p_buffer[write_index]) T(value1);
311 write.store(next_index, etl::memory_order_release);
324 template <
typename T1,
typename T2>
325 bool emplace(
const T1& value1,
const T2& value2)
327 size_type write_index =
write.load(etl::memory_order_relaxed);
328 size_type next_index = get_next_index(write_index, Reserved);
330 if (next_index !=
read.load(etl::memory_order_acquire))
332 ::new (&p_buffer[write_index]) T(value1, value2);
334 write.store(next_index, etl::memory_order_release);
347 template <
typename T1,
typename T2,
typename T3>
348 bool emplace(
const T1& value1,
const T2& value2,
const T3& value3)
350 size_type write_index =
write.load(etl::memory_order_relaxed);
351 size_type next_index = get_next_index(write_index, Reserved);
353 if (next_index !=
read.load(etl::memory_order_acquire))
355 ::new (&p_buffer[write_index]) T(value1, value2, value3);
357 write.store(next_index, etl::memory_order_release);
370 template <
typename T1,
typename T2,
typename T3,
typename T4>
371 bool emplace(
const T1& value1,
const T2& value2,
const T3& value3,
const T4& value4)
373 size_type write_index =
write.load(etl::memory_order_relaxed);
374 size_type next_index = get_next_index(write_index, Reserved);
376 if (next_index !=
read.load(etl::memory_order_acquire))
378 ::new (&p_buffer[write_index]) T(value1, value2, value3, value4);
380 write.store(next_index, etl::memory_order_release);
393 bool front(reference value)
395 size_type read_index =
read.load(etl::memory_order_relaxed);
397 if (read_index ==
write.load(etl::memory_order_acquire))
403 value = p_buffer[read_index];
411 bool pop(reference value)
413 size_type read_index =
read.load(etl::memory_order_relaxed);
415 if (read_index ==
write.load(etl::memory_order_acquire))
421 size_type next_index = get_next_index(read_index, Reserved);
423#if ETL_USING_CPP11 && ETL_NOT_USING_STLPORT && !defined(ETL_QUEUE_LOCKABLE_FORCE_CPP03_IMPLEMENTATION)
424 value = etl::move(p_buffer[read_index]);
426 value = p_buffer[read_index];
429 p_buffer[read_index].~T();
431 read.store(next_index, etl::memory_order_release);
441 size_type read_index =
read.load(etl::memory_order_relaxed);
443 if (read_index ==
write.load(etl::memory_order_acquire))
449 size_type next_index = get_next_index(read_index, Reserved);
451 p_buffer[read_index].~T();
453 read.store(next_index, etl::memory_order_release);
463 size_type read_index =
read.load(etl::memory_order_relaxed);
465 return p_buffer[read_index];
471 const_reference front()
const
473 size_type read_index =
read.load(etl::memory_order_relaxed);
475 return p_buffer[read_index];
485 if ETL_IF_CONSTEXPR(etl::is_trivially_destructible<T>::value)
504 iqueue_spsc_atomic(T* p_buffer_, size_type reserved_)
513 iqueue_spsc_atomic(
const iqueue_spsc_atomic&) ETL_DELETE;
514 iqueue_spsc_atomic& operator =(
const iqueue_spsc_atomic&) ETL_DELETE;
517 iqueue_spsc_atomic(iqueue_spsc_atomic&&) =
delete;
518 iqueue_spsc_atomic& operator =(iqueue_spsc_atomic&&) =
delete;
532 template <
typename T,
size_t Size, const
size_t Memory_Model = etl::memory_model::MEMORY_MODEL_LARGE>
533 class queue_spsc_atomic :
public iqueue_spsc_atomic<T, Memory_Model>
537 typedef typename etl::iqueue_spsc_atomic<T, Memory_Model> base_t;
541 typedef typename base_t::size_type size_type;
545 static ETL_CONSTANT size_type Reserved_Size = size_type(Size + 1);
549 ETL_STATIC_ASSERT((Size <= (etl::integral_limits<size_type>::max - 1)),
"Size too large for memory model");
551 static ETL_CONSTANT size_type MAX_SIZE = size_type(Size);
557 : base_t(reinterpret_cast<T*>(&buffer[0]), Reserved_Size)
572 typename etl::aligned_storage<
sizeof(T), etl::alignment_of<T>::value>::type buffer[Reserved_Size];
575 template <
typename T,
size_t Size, const
size_t Memory_Model>
576 ETL_CONSTANT
typename queue_spsc_atomic<T, Size, Memory_Model>::size_type queue_spsc_atomic<T, Size, Memory_Model>::MAX_SIZE;
bitset_ext
Definition absolute.h:39
size_t max_size() const
Returns the maximum number of items in the variant_pool.
Definition variant_pool_generator.h:395
etl::optional< T > read(etl::bit_stream_reader &stream)
Read a checked type from a stream.
Definition bit_stream.h:1379
ETL_CONSTEXPR TContainer::size_type size(const TContainer &container)
Definition iterator.h:1187
bool write(etl::bit_stream_writer &stream, bool value)
Definition bit_stream.h:997