mmdeploy/csrc/execution/when_all.h

176 lines
6.0 KiB
C
Raw Normal View History

[Feature] New pipeline & executor for SDK (#497) * executor prototype * add split/when_all * fix GCC build * WIP let_value * fix let_value * WIP ensure_started * ensure_started & start_detached * fix let_value + when_all combo on MSVC 142 * fix static thread pool * generic just, then, let_value, sync_wait * minor * generic split and when_all * fully generic sender adapters * when_all: workaround for GCC7 * support legacy spdlog * fix memleak * bulk * static detector * fix bulk & first pipeline * bulk for static thread pools * fix on MSVC * WIP async batch submission * WIP collation * async batch * fix detector * fix async detector * fix * fix * debug * fix cuda allocator * WIP type erased executor * better type erasure * simplify C API impl * Expand & type erase TC * deduction guide for type erased senders * fix GCC build * when_all for arrays of Value senders * WIP pipeline v2 * WIP pipeline parser * WIP timed batch operation * add registry * experiment * fix pipeline * naming * fix mem-leak * fix deferred batch operation * WIP * WIP configurable scheduler * WIP configurable scheduler * add comment * parse scheduler config * force link schedulers * WIP pipeable sender * WIP CPO * ADL isolation and dismantle headers * type erase single thread context * fix MSVC build * CPO * replace decay_t with remove_cvref_t * structure adjustment * structure adjustment * apply CPOs & C API rework * refine C API * detector async C API * adjust detector async C API * # Conflicts: # csrc/apis/c/detector.cpp * fix when_all for type erased senders * support void return for Then * async detector * fix some CPOs * minor * WIP rework capture mechanism for type erased types * minor fix * fix MSVC build * move expand.h to execution * make `Expand` pipeable * fix type erased * un-templatize `_TypeErasedOperation` * re-work C API * remove async_detector C API * fix pipeline * add flatten & unflatten * fix flatten & unflatten * add aync OCR demo * config executor for nodes & better executor API * working async OCR example * minor * dynamic batch via scheduler * dynamic batch on `Value` * fix MSVC build * type erase dynamic batch scheduler * sender as Python Awaitable * naming * naming * add docs * minor * merge tmp branch * unify C APIs * fix ocr * unify APIs * fix typo * update async OCR demo * add v3 API text recognizer * fix v3 API * fix lint * add license info & reformat * add demo async_ocr_v2 * revert files * revert files * resolve link issues * fix scheduler linkage for shared libs * fix license header * add docs for `mmdeploy_executor_split` * add missing `mmdeploy_executor_transfer_just` and `mmdeploy_executor_execute` * make `TimedSingleThreadContext` header only * fix lint * simplify type-erased sender
2022-06-01 14:10:43 +08:00
// Copyright (c) OpenMMLab. All rights reserved.
// Modified from
// https://github.com/brycelelbach/wg21_p2300_std_execution/blob/main/include/execution.hpp
#ifndef MMDEPLOY_CSRC_EXPERIMENTAL_EXECUTION_WHEN_ALL_H_
#define MMDEPLOY_CSRC_EXPERIMENTAL_EXECUTION_WHEN_ALL_H_
#include "utility.h"
namespace mmdeploy {
namespace __when_all {
template <typename... Senders>
using __concat_t = decltype(std::tuple_cat(std::declval<completion_signatures_of_t<Senders>>()...));
template <typename CvrefReceiver, typename... Senders>
struct _Operation {
struct type;
};
template <typename CvrefReceiver, typename... Senders>
using Operation = typename _Operation<CvrefReceiver, Senders...>::type;
template <typename CvrefReceiver, size_t Index, typename... Senders>
struct _Receiver {
struct type;
};
template <typename CvrefReceiver, size_t Index, typename... Senders>
using receiver_t = typename _Receiver<CvrefReceiver, Index, Senders...>::type;
template <typename CvrefReceiver, size_t Index, typename... Senders>
struct _Receiver<CvrefReceiver, Index, Senders...>::type {
using Receiver = remove_cvref_t<CvrefReceiver>;
Operation<CvrefReceiver, Senders...>* op_state_;
template <typename... As>
friend void tag_invoke(set_value_t, type&& self, As&&... as) noexcept {
std::get<Index>(self.op_state_->vals_).emplace((As &&) as...);
self.op_state_->_Arrive();
}
};
template <typename CvrefReceiver, typename... Senders>
struct _Operation<CvrefReceiver, Senders...>::type {
using Receiver = remove_cvref_t<CvrefReceiver>;
template <size_t Index>
using _receiver_t = receiver_t<CvrefReceiver, Index, Senders...>;
template <typename Sender, size_t Index>
using _ChildOpState = connect_result_t<_copy_cvref_t<CvrefReceiver, Sender>, _receiver_t<Index>>;
using _Indices = std::index_sequence_for<Senders...>;
// workaround for a bug in GCC7 that `Is` in a lambda is treated as unexpanded parameter pack
template <typename Sender, typename Receiver>
static auto _Connect1(Sender&& sender, Receiver&& receiver) {
return __conv{[&]() mutable { return Connect((Sender &&) sender, (Receiver &&) receiver); }};
}
template <size_t... Is, typename... _Senders>
static auto _ConnectChildren(type* self, std::index_sequence<Is...>, _Senders&&... senders)
-> std::tuple<_ChildOpState<Senders, Is>...> {
return {_Connect1((_Senders &&) senders, _receiver_t<Is>{self})...};
}
using _ChildOpStates = decltype(_ConnectChildren(
nullptr, _Indices{}, std::declval<_copy_cvref_t<CvrefReceiver, Senders>>()...));
using _ChildValueTuple = std::tuple<std::optional<completion_signatures_of_t<Senders>>...>;
void _Arrive() noexcept {
if (0 == --count_) {
_Complete();
}
}
void _Complete() noexcept {
std::apply(
[this](auto&... opt_vals) -> void {
std::apply(
[this](auto&... all_vals) -> void {
SetValue((Receiver &&) receiver_, std::move(all_vals)...);
},
std::tuple_cat(
std::apply([](auto&... vals) { return std::tie(vals...); }, *opt_vals)...));
},
vals_);
}
template <typename... _Senders>
explicit type(Receiver&& receiver, _Senders&&... senders)
: child_states_{_ConnectChildren(this, _Indices{}, (_Senders &&) senders...)},
receiver_(std::move(receiver)) {}
friend void tag_invoke(start_t, type& self) noexcept {
std::apply([](auto&&... child_ops) noexcept -> void { (Start(child_ops), ...); },
self.child_states_);
}
type(const type&) = delete;
type(type&&) = delete;
type& operator=(const type&) = delete;
type& operator=(type&&) = delete;
_ChildOpStates child_states_;
Receiver receiver_;
std::atomic<size_t> count_{sizeof...(Senders)};
_ChildValueTuple vals_;
};
template <typename... Senders>
struct _Sender {
struct type;
};
template <typename... Senders>
using Sender = typename _Sender<remove_cvref_t<Senders>...>::type;
template <typename... Senders>
struct _Sender<Senders...>::type {
using value_types = __concat_t<Senders...>;
template <typename Receiver>
using operation_t = Operation<Receiver, Senders...>;
template <typename Self, typename Receiver, _decays_to<Self, type, int> = 0>
friend auto tag_invoke(connect_t, Self&& self, Receiver&& receiver)
-> operation_t<_copy_cvref_t<Self, remove_cvref_t<Receiver>>> { // cvref encoded in receiver
// type
return std::apply(
[&](auto&&... senders) {
// MSVC v142 doesn't recognize operation_t here
return Operation<_copy_cvref_t<Self, remove_cvref_t<Receiver>>, Senders...>(
(Receiver &&) receiver, (decltype(senders)&&)senders...);
},
((Self &&) self).senders_);
}
std::tuple<Senders...> senders_;
};
struct when_all_t {
template <typename... Senders,
std::enable_if_t<(_is_sender<Senders> && ...) && (sizeof...(Senders) > 0) &&
tag_invocable<when_all_t, Senders...>,
int> = 0>
auto operator()(Senders&&... senders) const {
return tag_invoke(when_all_t{}, (Senders &&) senders...);
}
template <
typename Range, typename ValueType = typename remove_cvref_t<Range>::value_type,
std::enable_if_t<
_is_range_v<Range> && _is_sender<ValueType> && tag_invocable<when_all_t, Range>, int> = 0>
auto operator()(Range&& range) const {
return tag_invoke(when_all_t{}, (Range &&) range);
}
template <typename... Senders,
std::enable_if_t<(_is_sender<Senders> && ...) && (sizeof...(Senders) > 0) &&
!tag_invocable<when_all_t, Senders...>,
int> = 0>
Sender<Senders...> operator()(Senders&&... senders) const {
return {{(Senders &&) senders...}};
}
};
} // namespace __when_all
using __when_all::when_all_t;
inline constexpr when_all_t WhenAll{};
} // namespace mmdeploy
#endif // MMDEPLOY_CSRC_EXPERIMENTAL_EXECUTION_WHEN_ALL_H_