#pragma once #include #include #include #include #include #include "libipc/def.h" #include "libipc/platform/detail.h" #include "libipc/circ/elem_def.h" #include "libipc/utility/log.h" #include "libipc/utility/utility.h" namespace ipc { //////////////////////////////////////////////////////////////// /// producer-consumer implementation //////////////////////////////////////////////////////////////// template struct prod_cons_impl; template <> struct prod_cons_impl> { template struct elem_t { std::aligned_storage_t data_ {}; }; alignas(cache_line_size) std::atomic rd_; // read index alignas(cache_line_size) std::atomic wt_; // write index constexpr circ::u2_t cursor() const noexcept { return 0; } template bool push(W* /*wrapper*/, F&& f, E* elems) { auto cur_wt = circ::index_of(wt_.load(std::memory_order_relaxed)); if (cur_wt == circ::index_of(rd_.load(std::memory_order_acquire) - 1)) { return false; // full } std::forward(f)(&(elems[cur_wt].data_)); wt_.fetch_add(1, std::memory_order_release); return true; } /** * In single-single-unicast, 'force_push' means 'no reader' or 'the only one reader is dead'. * So we could just disconnect all connections of receiver, and return false. */ template bool force_push(W* wrapper, F&&, E*) { wrapper->elems()->disconnect_receiver(~static_cast(0u)); return false; } template bool pop(W* /*wrapper*/, circ::u2_t& /*cur*/, F&& f, R&& out, E* elems) { auto cur_rd = circ::index_of(rd_.load(std::memory_order_relaxed)); if (cur_rd == circ::index_of(wt_.load(std::memory_order_acquire))) { return false; // empty } std::forward(f)(&(elems[cur_rd].data_)); std::forward(out)(true); rd_.fetch_add(1, std::memory_order_release); return true; } }; template <> struct prod_cons_impl> : prod_cons_impl> { template bool force_push(W* wrapper, F&&, E*) { wrapper->elems()->disconnect_receiver(1); return false; } template class E, std::size_t DS, std::size_t AS> bool pop(W* /*wrapper*/, circ::u2_t& /*cur*/, F&& f, R&& out, E* elems) { byte_t buff[DS]; for (unsigned k = 0;;) { auto cur_rd = rd_.load(std::memory_order_relaxed); if (circ::index_of(cur_rd) == circ::index_of(wt_.load(std::memory_order_acquire))) { return false; // empty } std::memcpy(buff, &(elems[circ::index_of(cur_rd)].data_), sizeof(buff)); if (rd_.compare_exchange_weak(cur_rd, cur_rd + 1, std::memory_order_release)) { std::forward(f)(buff); std::forward(out)(true); return true; } ipc::yield(k); } } }; template <> struct prod_cons_impl> : prod_cons_impl> { using flag_t = std::uint64_t; template struct elem_t { std::aligned_storage_t data_ {}; std::atomic f_ct_ { 0 }; // commit flag }; alignas(cache_line_size) std::atomic ct_; // commit index template bool push(W* /*wrapper*/, F&& f, E* elems) { circ::u2_t cur_ct, nxt_ct; for (unsigned k = 0;;) { cur_ct = ct_.load(std::memory_order_relaxed); if (circ::index_of(nxt_ct = cur_ct + 1) == circ::index_of(rd_.load(std::memory_order_acquire))) { return false; // full } if (ct_.compare_exchange_weak(cur_ct, nxt_ct, std::memory_order_acq_rel)) { break; } ipc::yield(k); } auto* el = elems + circ::index_of(cur_ct); std::forward(f)(&(el->data_)); // set flag & try update wt el->f_ct_.store(~static_cast(cur_ct), std::memory_order_release); while (1) { auto cac_ct = el->f_ct_.load(std::memory_order_acquire); if (cur_ct != wt_.load(std::memory_order_relaxed)) { return true; } if ((~cac_ct) != cur_ct) { return true; } if (!el->f_ct_.compare_exchange_strong(cac_ct, 0, std::memory_order_relaxed)) { return true; } wt_.store(nxt_ct, std::memory_order_release); cur_ct = nxt_ct; nxt_ct = cur_ct + 1; el = elems + circ::index_of(cur_ct); } return true; } template bool force_push(W* wrapper, F&&, E*) { wrapper->elems()->disconnect_receiver(1); return false; } template class E, std::size_t DS, std::size_t AS> bool pop(W* /*wrapper*/, circ::u2_t& /*cur*/, F&& f, R&& out, E* elems) { byte_t buff[DS]; for (unsigned k = 0;;) { auto cur_rd = rd_.load(std::memory_order_relaxed); auto cur_wt = wt_.load(std::memory_order_acquire); auto id_rd = circ::index_of(cur_rd); auto id_wt = circ::index_of(cur_wt); if (id_rd == id_wt) { auto* el = elems + id_wt; auto cac_ct = el->f_ct_.load(std::memory_order_acquire); if ((~cac_ct) != cur_wt) { return false; // empty } if (el->f_ct_.compare_exchange_weak(cac_ct, 0, std::memory_order_relaxed)) { wt_.store(cur_wt + 1, std::memory_order_release); } k = 0; } else { std::memcpy(buff, &(elems[circ::index_of(cur_rd)].data_), sizeof(buff)); if (rd_.compare_exchange_weak(cur_rd, cur_rd + 1, std::memory_order_release)) { std::forward(f)(buff); std::forward(out)(true); return true; } ipc::yield(k); } } } }; template <> struct prod_cons_impl> { using rc_t = std::uint64_t; enum : rc_t { ep_mask = 0x00000000ffffffffull, ep_incr = 0x0000000100000000ull }; template struct elem_t { std::aligned_storage_t data_ {}; std::atomic rc_ { 0 }; // read-counter }; alignas(cache_line_size) std::atomic wt_; // write index alignas(cache_line_size) rc_t epoch_ { 0 }; // only one writer circ::u2_t cursor() const noexcept { return wt_.load(std::memory_order_acquire); } template bool push(W* wrapper, F&& f, E* elems) { E* el; for (unsigned k = 0;;) { circ::cc_t cc = wrapper->elems()->connections(std::memory_order_relaxed); if (cc == 0) return false; // no reader el = elems + circ::index_of(wt_.load(std::memory_order_relaxed)); // check all consumers have finished reading this element auto cur_rc = el->rc_.load(std::memory_order_acquire); circ::cc_t rem_cc = cur_rc & ep_mask; if ((cc & rem_cc) && ((cur_rc & ~ep_mask) == epoch_)) { return false; // has not finished yet } // consider rem_cc to be 0 here if (el->rc_.compare_exchange_weak( cur_rc, epoch_ | static_cast(cc), std::memory_order_release)) { break; } ipc::yield(k); } std::forward(f)(&(el->data_)); wt_.fetch_add(1, std::memory_order_release); return true; } template bool force_push(W* wrapper, F&& f, E* elems) { E* el; epoch_ += ep_incr; for (unsigned k = 0;;) { circ::cc_t cc = wrapper->elems()->connections(std::memory_order_relaxed); if (cc == 0) return false; // no reader el = elems + circ::index_of(wt_.load(std::memory_order_relaxed)); // check all consumers have finished reading this element auto cur_rc = el->rc_.load(std::memory_order_acquire); circ::cc_t rem_cc = cur_rc & ep_mask; if (cc & rem_cc) { ipc::log("force_push: k = %u, cc = %u, rem_cc = %u\n", k, cc, rem_cc); cc = wrapper->elems()->disconnect_receiver(rem_cc); // disconnect all invalid readers if (cc == 0) return false; // no reader } // just compare & exchange if (el->rc_.compare_exchange_weak( cur_rc, epoch_ | static_cast(cc), std::memory_order_release)) { break; } ipc::yield(k); } std::forward(f)(&(el->data_)); wt_.fetch_add(1, std::memory_order_release); return true; } template bool pop(W* wrapper, circ::u2_t& cur, F&& f, R&& out, E* elems) { if (cur == cursor()) return false; // acquire auto* el = elems + circ::index_of(cur++); std::forward(f)(&(el->data_)); for (unsigned k = 0;;) { auto cur_rc = el->rc_.load(std::memory_order_acquire); if ((cur_rc & ep_mask) == 0) { std::forward(out)(true); return true; } auto nxt_rc = cur_rc & ~static_cast(wrapper->connected_id()); if (el->rc_.compare_exchange_weak(cur_rc, nxt_rc, std::memory_order_release)) { std::forward(out)((nxt_rc & ep_mask) == 0); return true; } ipc::yield(k); } } }; template <> struct prod_cons_impl> { using rc_t = std::uint64_t; using flag_t = std::uint64_t; enum : rc_t { rc_mask = 0x00000000ffffffffull, ep_mask = 0x00ffffffffffffffull, ep_incr = 0x0100000000000000ull, ic_mask = 0xff000000ffffffffull, ic_incr = 0x0000000100000000ull }; template struct elem_t { std::aligned_storage_t data_ {}; std::atomic rc_ { 0 }; // read-counter std::atomic f_ct_ { 0 }; // commit flag }; alignas(cache_line_size) std::atomic ct_; // commit index alignas(cache_line_size) std::atomic epoch_ { 0 }; circ::u2_t cursor() const noexcept { return ct_.load(std::memory_order_acquire); } constexpr static rc_t inc_rc(rc_t rc) noexcept { return (rc & ic_mask) | ((rc + ic_incr) & ~ic_mask); } constexpr static rc_t inc_mask(rc_t rc) noexcept { return inc_rc(rc) & ~rc_mask; } template bool push(W* wrapper, F&& f, E* elems) { E* el; circ::u2_t cur_ct; rc_t epoch = epoch_.load(std::memory_order_acquire); for (unsigned k = 0;;) { circ::cc_t cc = wrapper->elems()->connections(std::memory_order_relaxed); if (cc == 0) return false; // no reader el = elems + circ::index_of(cur_ct = ct_.load(std::memory_order_relaxed)); // check all consumers have finished reading this element auto cur_rc = el->rc_.load(std::memory_order_relaxed); circ::cc_t rem_cc = cur_rc & rc_mask; if ((cc & rem_cc) && ((cur_rc & ~ep_mask) == epoch)) { return false; // has not finished yet } else if (!rem_cc) { auto cur_fl = el->f_ct_.load(std::memory_order_acquire); if ((cur_fl != cur_ct) && cur_fl) { return false; // full } } // consider rem_cc to be 0 here if (el->rc_.compare_exchange_weak( cur_rc, inc_mask(epoch | (cur_rc & ep_mask)) | static_cast(cc), std::memory_order_relaxed) && epoch_.compare_exchange_weak(epoch, epoch, std::memory_order_acq_rel)) { break; } ipc::yield(k); } // only one thread/process would touch here at one time ct_.store(cur_ct + 1, std::memory_order_release); std::forward(f)(&(el->data_)); // set flag & try update wt el->f_ct_.store(~static_cast(cur_ct), std::memory_order_release); return true; } template bool force_push(W* wrapper, F&& f, E* elems) { E* el; circ::u2_t cur_ct; rc_t epoch = epoch_.fetch_add(ep_incr, std::memory_order_release) + ep_incr; for (unsigned k = 0;;) { circ::cc_t cc = wrapper->elems()->connections(std::memory_order_relaxed); if (cc == 0) return false; // no reader el = elems + circ::index_of(cur_ct = ct_.load(std::memory_order_relaxed)); // check all consumers have finished reading this element auto cur_rc = el->rc_.load(std::memory_order_acquire); circ::cc_t rem_cc = cur_rc & rc_mask; if (cc & rem_cc) { ipc::log("force_push: k = %u, cc = %u, rem_cc = %u\n", k, cc, rem_cc); cc = wrapper->elems()->disconnect_receiver(rem_cc); // disconnect all invalid readers if (cc == 0) return false; // no reader } // just compare & exchange if (el->rc_.compare_exchange_weak( cur_rc, inc_mask(epoch | (cur_rc & ep_mask)) | static_cast(cc), std::memory_order_relaxed)) { if (epoch == epoch_.load(std::memory_order_acquire)) { break; } else if (push(wrapper, std::forward(f), elems)) { return true; } epoch = epoch_.fetch_add(ep_incr, std::memory_order_release) + ep_incr; } ipc::yield(k); } // only one thread/process would touch here at one time ct_.store(cur_ct + 1, std::memory_order_release); std::forward(f)(&(el->data_)); // set flag & try update wt el->f_ct_.store(~static_cast(cur_ct), std::memory_order_release); return true; } template bool pop(W* wrapper, circ::u2_t& cur, F&& f, R&& out, E(& elems)[N]) { auto* el = elems + circ::index_of(cur); auto cur_fl = el->f_ct_.load(std::memory_order_acquire); if (cur_fl != ~static_cast(cur)) { return false; // empty } ++cur; std::forward(f)(&(el->data_)); for (unsigned k = 0;;) { auto cur_rc = el->rc_.load(std::memory_order_acquire); if ((cur_rc & rc_mask) == 0) { std::forward(out)(true); el->f_ct_.store(cur + N - 1, std::memory_order_release); return true; } auto nxt_rc = inc_rc(cur_rc) & ~static_cast(wrapper->connected_id()); bool last_one = false; if ((last_one = (nxt_rc & rc_mask) == 0)) { el->f_ct_.store(cur + N - 1, std::memory_order_release); } if (el->rc_.compare_exchange_weak(cur_rc, nxt_rc, std::memory_order_release)) { std::forward(out)(last_one); return true; } ipc::yield(k); } } }; } // namespace ipc