endif()
set(the_description "OpenCV G-API Core Module")
+
ocv_add_module(gapi opencv_imgproc)
file(GLOB gapi_ext_hdrs
src/api/kernels_core.cpp
src/api/kernels_imgproc.cpp
src/api/render.cpp
+ src/api/ginfer.cpp
# Compiler part
src/compiler/gmodel.cpp
src/backends/ocl/goclimgproc.cpp
src/backends/ocl/goclcore.cpp
+ # IE Backend. FIXME: should be included by CMake
+ # if and only if IE support is enabled
+ src/backends/ie/giebackend.cpp
+
# Compound
src/backends/common/gcompoundbackend.cpp
src/backends/common/gcompoundkernel.cpp
# Note `ade` is not a module name but link dependency for ${the_module}
# (which is opencv_gapi)
-ocv_create_module(ade)
-
-ocv_add_accuracy_tests()
+ocv_create_module(ade ${INF_ENGINE_TARGET})
+ocv_add_accuracy_tests(${INF_ENGINE_TARGET})
# FIXME: test binary is linked with ADE directly since ADE symbols
# are not exported from libopencv_gapi.so in any form - thus
# there're two copies of ADE code in memory when tests run (!)
#include <opencv2/gapi/util/throw.hpp>
#include <opencv2/gapi/own/assert.hpp>
+#include <opencv2/gapi/gmat.hpp> // flatten_g only!
+#include <opencv2/gapi/gscalar.hpp> // flatten_g only!
+
namespace cv
{
// Forward declaration; GNode and GOrigin are an internal
return m_ref->m_desc;
}
};
+
+ // Helper (FIXME: work-around?)
+ // stripping G types to their host types
+ // like cv::GArray<GMat> would still map to std::vector<cv::Mat>
+ // but not to std::vector<cv::GMat>
+#if defined(GAPI_STANDALONE)
+# define FLATTEN_NS cv::gapi::own
+#else
+# define FLATTEN_NS cv
+#endif
+ template<class T> struct flatten_g;
+ template<> struct flatten_g<cv::GMat> { using type = FLATTEN_NS::Mat; };
+ template<> struct flatten_g<cv::GScalar> { using type = FLATTEN_NS::Scalar; };
+ template<class T> struct flatten_g { using type = T; };
+#undef FLATTEN_NS
+ // FIXME: the above mainly duplicates "ProtoToParam" thing from gtyped.hpp
+ // but I decided not to include gtyped here - probably worth moving that stuff
+ // to some common place? (DM)
} // namespace detail
/** \addtogroup gapi_data_objects
detail::GArrayU strip() const { return m_ref; }
private:
- static void VCTor(detail::VectorRef& vref) { vref.reset<T>(); }
+ // Host type (or Flat type) - the type this GArray is actually
+ // specified to.
+ using HT = typename detail::flatten_g<typename std::decay<T>::type>::type;
+
+ static void VCTor(detail::VectorRef& vref) {
+ vref.reset<HT>();
+ }
void putDetails() {
m_ref.setConstructFcn(&VCTor);
- m_ref.specifyType<T>();
+ m_ref.specifyType<HT>();
}
detail::GArrayU m_ref;
using M = std::function<GMetaArgs(const GMetaArgs &, const GArgs &)>;
const std::string name; // kernel ID, defined by its API (signature)
+ const std::string tag; // some (implementation-specific) tag
const M outMeta; // generic adaptor to API::outMeta(...)
- const GShapes outShapes; // types (shapes) kernel's outputs
+ const GShapes outShapes; // types (shapes) kernel's outputs
};
// GKernelImpl describes particular kernel implementation to the system
}
};
+ ////////////////////////////////////////////////////////////////////////////
+ // Helper class to introduce tags to calls. By default there's no tag
+ struct NoTag {
+ static constexpr const char *tag() { return ""; }
+ };
+
} // namespace detail
// GKernelType and GKernelTypeM are base classes which implement typed ::on()
// GKernelTypeM respectively.
template<typename K, typename... R, typename... Args>
-class GKernelTypeM<K, std::function<std::tuple<R...>(Args...)> >:
- public detail::MetaHelper<K, std::tuple<Args...>, std::tuple<R...>>
+class GKernelTypeM<K, std::function<std::tuple<R...>(Args...)> >
+ : public detail::MetaHelper<K, std::tuple<Args...>, std::tuple<R...>>
+ , public detail::NoTag
{
template<int... IIs>
static std::tuple<R...> yield(cv::GCall &call, detail::Seq<IIs...>)
static std::tuple<R...> on(Args... args)
{
- cv::GCall call(GKernel{K::id(), &K::getOutMeta, {detail::GTypeTraits<R>::shape...}});
+ cv::GCall call(GKernel{K::id(), K::tag(), &K::getOutMeta, {detail::GTypeTraits<R>::shape...}});
call.pass(args...);
return yield(call, typename detail::MkSeq<sizeof...(R)>::type());
}
template<typename, typename> class GKernelType;
template<typename K, typename R, typename... Args>
-class GKernelType<K, std::function<R(Args...)> >:
- public detail::MetaHelper<K, std::tuple<Args...>, R>
+class GKernelType<K, std::function<R(Args...)> >
+ : public detail::MetaHelper<K, std::tuple<Args...>, R>
+ , public detail::NoTag
{
public:
using InArgs = std::tuple<Args...>;
static R on(Args... args)
{
- cv::GCall call(GKernel{K::id(), &K::getOutMeta, {detail::GTypeTraits<R>::shape}});
+ cv::GCall call(GKernel{K::id(), K::tag(), &K::getOutMeta, {detail::GTypeTraits<R>::shape}});
call.pass(args...);
return detail::Yield<R>::yield(call, 0);
}
public detail::G_ID_HELPER_CLASS(Class)
// {body} is to be defined by user
+#define G_API_OP G_TYPED_KERNEL
+#define G_API_OP_M G_TYPED_KERNEL_M
+
namespace cv
{
namespace gapi
return includesAPI(KAPI::id());
}
+ // FIXME: The below comment is wrong, and who needs this function?
/**
* @brief Find a kernel (by its API)
*
int chan;
cv::gapi::own::Size size; // NB.: no multi-dimensional cases covered yet
bool planar;
+ std::vector<int> dims; // FIXME: Maybe it's real questionable to have it here
GMatDesc(int d, int c, cv::gapi::own::Size s, bool p = false)
: depth(d), chan(c), size(s), planar(p) {}
+ GMatDesc(int d, const std::vector<int> &dd)
+ : depth(d), chan(-1), size{-1,-1}, planar(false), dims(dd) {}
+
+ GMatDesc(int d, std::vector<int> &&dd)
+ : depth(d), chan(-1), size{-1,-1}, planar(false), dims(std::move(dd)) {}
+
GMatDesc() : GMatDesc(-1, -1, {-1,-1}) {}
inline bool operator== (const GMatDesc &rhs) const
{
- return depth == rhs.depth && chan == rhs.chan && size == rhs.size && planar == rhs.planar;
+ return depth == rhs.depth
+ && chan == rhs.chan
+ && size == rhs.size
+ && planar == rhs.planar
+ && dims == rhs.dims;
}
inline bool operator!= (const GMatDesc &rhs) const
return !(*this == rhs);
}
+ bool isND() const { return !dims.empty(); }
+
// Checks if the passed mat can be described by this descriptor
// (it handles the case when
// 1-channel mat can be reinterpreted as is (1-channel mat)
} // namespace detail
+// Note: descr_of(std::vector<..>) returns a GArrayDesc, while
+// descrs_of(std::vector<..>) returns an array of Meta args!
class Mat;
class UMat;
-GAPI_EXPORTS cv::GMetaArgs descr_of(const std::vector<cv::Mat> &vec);
-GAPI_EXPORTS cv::GMetaArgs descr_of(const std::vector<cv::UMat> &vec);
+GAPI_EXPORTS cv::GMetaArgs descrs_of(const std::vector<cv::Mat> &vec);
+GAPI_EXPORTS cv::GMetaArgs descrs_of(const std::vector<cv::UMat> &vec);
namespace gapi { namespace own {
class Mat;
- GAPI_EXPORTS cv::GMetaArgs descr_of(const std::vector<Mat> &vec);
+ GAPI_EXPORTS cv::GMetaArgs descrs_of(const std::vector<Mat> &vec);
}} // namespace gapi::own
} // namespace cv
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2019 Intel Corporation
+
+
+#ifndef OPENCV_GAPI_INFER_HPP
+#define OPENCV_GAPI_INFER_HPP
+
+// FIXME: Inference API is currently only available in full mode
+#if !defined(GAPI_STANDALONE)
+
+#include <functional>
+#include <string> // string
+#include <utility> // tuple
+
+#include <opencv2/gapi/util/any.hpp> // any<>
+#include <opencv2/gapi/gkernel.hpp> // GKernelType[M], GBackend
+#include <opencv2/gapi/garg.hpp> // GArg
+#include <opencv2/gapi/gcommon.hpp> // CompileArgTag
+#include <opencv2/gapi/gmetaarg.hpp> // GMetaArg
+
+namespace cv {
+
+namespace detail {
+ // This tiny class eliminates the semantic difference between
+ // GKernelType and GKernelTypeM.
+ // FIXME: Something similar can be reused for regular kernels
+ template<typename, typename>
+ struct KernelTypeMedium;
+
+ template<class K, typename... R, typename... Args>
+ struct KernelTypeMedium<K, std::function<std::tuple<R...>(Args...)> >:
+ public GKernelTypeM<K, std::function<std::tuple<R...>(Args...)> > {};
+
+ template<class K, typename R, typename... Args>
+ struct KernelTypeMedium<K, std::function<R(Args...)> >:
+ public GKernelType<K, std::function<R(Args...)> > {};
+
+} // namespace detail
+
+template<typename, typename> class GNetworkType;
+
+// TODO: maybe tuple_wrap_helper from util.hpp may help with this.
+// Multiple-return-value network definition (specialized base class)
+template<typename K, typename... R, typename... Args>
+class GNetworkType<K, std::function<std::tuple<R...>(Args...)> >
+{
+public:
+ using InArgs = std::tuple<Args...>;
+ using OutArgs = std::tuple<R...>;
+
+ using Result = OutArgs;
+ using API = std::function<Result(Args...)>;
+
+ using ResultL = std::tuple< cv::GArray<R>... >;
+ using APIList = std::function<ResultL(cv::GArray<cv::Rect>, Args...)>;
+};
+
+// Single-return-value network definition (specialized base class)
+template<typename K, typename R, typename... Args>
+class GNetworkType<K, std::function<R(Args...)> >
+{
+public:
+ using InArgs = std::tuple<Args...>;
+ using OutArgs = std::tuple<R>;
+
+ using Result = R;
+ using API = std::function<R(Args...)>;
+
+ using ResultL = cv::GArray<R>;
+ using APIList = std::function<ResultL(cv::GArray<cv::Rect>, Args...)>;
+};
+
+// Base "Infer" kernel. Note - for whatever network, kernel ID
+// is always the same. Different inference calls are distinguished by
+// network _tag_ (an extra field in GCall)
+//
+// getOutMeta is a stub callback collected by G-API kernel subsystem
+// automatically. This is a rare case when this callback is defined by
+// a particular backend, not by a network itself.
+struct GInferBase {
+ static constexpr const char * id() {
+ return "org.opencv.dnn.infer"; // Universal stub
+ }
+ static GMetaArgs getOutMeta(const GMetaArgs &, const GArgs &) {
+ return GMetaArgs{}; // One more universal stub
+ }
+};
+
+
+// Base "Infer list" kernel.
+// All notes from "Infer" kernel apply here as well.
+struct GInferListBase {
+ static constexpr const char * id() {
+ return "org.opencv.dnn.infer-roi"; // Universal stub
+ }
+ static GMetaArgs getOutMeta(const GMetaArgs &, const GArgs &) {
+ return GMetaArgs{}; // One more universal stub
+ }
+};
+
+// A generic inference kernel. API (::on()) is fully defined by the Net
+// template parameter.
+// Acts as a regular kernel in graph (via KernelTypeMedium).
+template<typename Net>
+struct GInfer final
+ : public GInferBase
+ , public detail::KernelTypeMedium< GInfer<Net>
+ , typename Net::API > {
+ using GInferBase::getOutMeta; // FIXME: name lookup conflict workaround?
+
+ static constexpr const char* tag() { return Net::tag(); }
+};
+
+// A generic roi-list inference kernel. API (::on()) is derived from
+// the Net template parameter (see more in infer<> overload).
+template<typename Net>
+struct GInferList final
+ : public GInferListBase
+ , public detail::KernelTypeMedium< GInferList<Net>
+ , typename Net::APIList > {
+ using GInferListBase::getOutMeta; // FIXME: name lookup conflict workaround?
+
+ static constexpr const char* tag() { return Net::tag(); }
+};
+
+} // namespace cv
+
+// FIXME: Probably the <API> signature makes a function/tuple/function round-trip
+#define G_API_NET(Class, API, Tag) \
+ struct Class final: public cv::GNetworkType<Class, std::function API> { \
+ static constexpr const char * tag() { return Tag; } \
+ }
+
+namespace cv {
+namespace gapi {
+
+
+/** @brief Calculates responses for the specified network (template
+ * parameter) for every region in the source image.
+ *
+ * @tparam A network type defined with G_API_NET() macro.
+ * @param roi a list of rectangles describing regions of interest
+ * in the source image. Usually an output of object detector or tracker.
+ * @param args network's input parameters as specified in G_API_NET() macro.
+ * NOTE: verified to work reliably with 1-input topologies only.
+ * @return a list of objects of return type as defined in G_API_NET().
+ * If a network has multiple return values (defined with a tuple), a tuple of
+ * GArray<> objects is returned with the appropriate types inside.
+ * @sa G_API_NET()
+ */
+template<typename Net, typename... Args>
+typename Net::ResultL infer(cv::GArray<cv::Rect> roi, Args&&... args) {
+ return GInferList<Net>::on(roi, std::forward<Args>(args)...);
+}
+
+/**
+ * @brief Calculates response for the specified network (template
+ * parameter) given the input data.
+ *
+ * @tparam A network type defined with G_API_NET() macro.
+ * @param args network's input parameters as specified in G_API_NET() macro.
+ * @return an object of return type as defined in G_API_NET().
+ * If a network has multiple return values (defined with a tuple), a tuple of
+ * objects of apprpriate type is returned.
+ * @sa G_API_NET()
+ */
+template<typename Net, typename... Args>
+typename Net::Result infer(Args&&... args) {
+ return GInfer<Net>::on(std::forward<Args>(args)...);
+}
+
+
+} // namespace gapi
+} // namespace cv
+
+#endif // GAPI_STANDALONE
+
+namespace cv {
+namespace gapi {
+
+// Note: the below code _is_ part of STANDALONE build,
+// just to make our compiler code compileable.
+
+// A type-erased form of network parameters.
+// Similar to how a type-erased GKernel is represented and used.
+struct GAPI_EXPORTS GNetParam {
+ std::string tag; // FIXME: const?
+ GBackend backend; // Specifies the execution model
+ util::any params; // Backend-interpreted parameter structure
+};
+
+/**
+ * @brief A container class for network configurations. Similar to
+ * GKernelPackage.Use cv::gapi::networks() to construct this object.
+ *
+ * @sa cv::gapi::networks
+ */
+struct GAPI_EXPORTS GNetPackage {
+ explicit GNetPackage(std::initializer_list<GNetParam> &&ii = {});
+ std::vector<GBackend> backends() const;
+ std::vector<GNetParam> networks;
+};
+} // namespace gapi
+
+namespace detail {
+template<typename T>
+gapi::GNetParam strip(T&& t) {
+ return gapi::GNetParam { t.tag()
+ , t.backend()
+ , t.params()
+ };
+}
+
+template<> struct CompileArgTag<cv::gapi::GNetPackage> {
+ static const char* tag() { return "gapi.net_package"; }
+};
+
+} // namespace cv::detail
+
+namespace gapi {
+template<typename... Args>
+cv::gapi::GNetPackage networks(Args&&... args) {
+ return cv::gapi::GNetPackage({ cv::detail::strip(args)... });
+}
+} // namespace gapi
+} // namespace cv
+
+#endif // OPENCV_GAPI_INFER_HPP
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2019 Intel Corporation
+
+#ifndef OPENCV_GAPI_INFER_IE_HPP
+#define OPENCV_GAPI_INFER_IE_HPP
+
+#ifdef HAVE_INF_ENGINE
+
+#include <unordered_map>
+#include <string>
+#include <array>
+#include <tuple> // tuple, tuple_size
+
+#include <opencv2/gapi/opencv_includes.hpp>
+#include <opencv2/gapi/util/any.hpp>
+
+namespace cv {
+namespace gapi {
+// FIXME: introduce a new sub-namespace for NN?
+namespace ie {
+
+GAPI_EXPORTS cv::gapi::GBackend backend();
+
+namespace detail {
+ struct ParamDesc {
+ std::string model_path;
+ std::string weights_path;
+ std::string device_id;
+
+ // NB: Here order follows the `Net` API
+ std::vector<std::string> input_names;
+ std::vector<std::string> output_names;
+
+ std::unordered_map<std::string, cv::Mat> const_inputs;
+
+ // NB: nun_* may differ from topology's real input/output port numbers
+ // (e.g. topology's partial execution)
+ std::size_t num_in; // How many inputs are defined in the operation
+ std::size_t num_out; // How many outputs are defined in the operation
+ };
+} // namespace detail
+
+// FIXME: this is probably a shared (reusable) thing
+template<typename Net>
+struct PortCfg {
+ using In = std::array
+ < std::string
+ , std::tuple_size<typename Net::InArgs>::value >;
+ using Out = std::array
+ < std::string
+ , std::tuple_size<typename Net::OutArgs>::value >;
+};
+
+template<typename Net> class Params {
+public:
+ Params(const std::string &model,
+ const std::string &weights,
+ const std::string &device)
+ : desc{ model, weights, device, {}, {}, {}
+ , std::tuple_size<typename Net::InArgs>::value
+ , std::tuple_size<typename Net::OutArgs>::value
+ } {
+ };
+
+ Params<Net>& cfgInputLayers(const typename PortCfg<Net>::In &ll) {
+ desc.input_names.clear();
+ desc.input_names.reserve(ll.size());
+ std::copy(ll.begin(), ll.end(),
+ std::back_inserter(desc.input_names));
+ return *this;
+ }
+
+ Params<Net>& cfgOutputLayers(const typename PortCfg<Net>::Out &ll) {
+ desc.output_names.clear();
+ desc.output_names.reserve(ll.size());
+ std::copy(ll.begin(), ll.end(),
+ std::back_inserter(desc.output_names));
+ return *this;
+ }
+
+ Params<Net>& constInput(const std::string &layer_name,
+ const cv::Mat &data) {
+ desc.const_inputs[layer_name] = data;
+ return *this;
+ }
+
+ // BEGIN(G-API's network parametrization API)
+ GBackend backend() const { return cv::gapi::ie::backend(); }
+ std::string tag() const { return Net::tag(); }
+ cv::util::any params() const { return { desc }; }
+ // END(G-API's network parametrization API)
+
+protected:
+ detail::ParamDesc desc;
+};
+
+} // namespace ie
+} // namespace gapi
+} // namespace cv
+
+#endif // HAVE_INF_ENGINE
+
+#endif // OPENCV_GAPI_INFER_HPP
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2019 Intel Corporation
+
+#ifndef OPENCV_GAPI_INFER_IE_UTIL_HPP
+#define OPENCV_GAPI_INFER_IE_UTIL_HPP
+
+#ifdef HAVE_INF_ENGINE
+
+// NOTE: This file is not included by default in infer/ie.hpp
+// and won't be. infer/ie.hpp doesn't depend on IE headers itself.
+// This file does -- so needs to be included separately by those who care.
+
+#include "inference_engine.hpp"
+
+namespace cv {
+namespace gapi {
+namespace ie {
+namespace util {
+
+GAPI_EXPORTS std::vector<int> to_ocv(const InferenceEngine::SizeVector &dims);
+
+GAPI_EXPORTS cv::Mat to_ocv(InferenceEngine::Blob::Ptr blob);
+GAPI_EXPORTS InferenceEngine::Blob::Ptr to_ie(cv::Mat &blob);
+
+}}}}
+
+#endif // HAVE_INF_ENGINE
+#endif // OPENCV_GAPI_INFER_IE_UTIL_HPP
namespace cv
{
- inline cv::gapi::own::Mat to_own(Mat const& m) { return {m.rows, m.cols, m.type(), m.data, m.step};};
+ template<typename T>
+ std::vector<T> to_own(const cv::MatSize &sz) {
+ std::vector<T> result(sz.dims());
+ for (int i = 0; i < sz.dims(); i++) {
+ // Note: cv::MatSize is not iterable
+ result[i] = static_cast<T>(sz[i]);
+ }
+ return result;
+ }
+
cv::gapi::own::Mat to_own(Mat&&) = delete;
+ inline cv::gapi::own::Mat to_own(Mat const& m) {
+ return (m.dims == 2)
+ ? cv::gapi::own::Mat{m.rows, m.cols, m.type(), m.data, m.step}
+ : cv::gapi::own::Mat{to_own<int>(m.size), m.type(), m.data};
+ };
+
inline cv::gapi::own::Scalar to_own(const cv::Scalar& s) { return {s[0], s[1], s[2], s[3]}; };
{
namespace own
{
- inline cv::Mat to_ocv(Mat const& m) { return {m.rows, m.cols, m.type(), m.data, m.step};};
+ inline cv::Mat to_ocv(Mat const& m) {
+ return m.dims.empty()
+ ? cv::Mat{m.rows, m.cols, m.type(), m.data, m.step}
+ : cv::Mat{m.dims, m.type(), m.data};
+ }
cv::Mat to_ocv(Mat&&) = delete;
inline cv::Scalar to_ocv(const Scalar& s) { return {s[0], s[1], s[2], s[3]}; };
#include <memory> //std::shared_ptr
#include <cstring> //std::memcpy
+#include <numeric> //std::accumulate
#include <opencv2/gapi/util/throw.hpp>
namespace cv { namespace gapi { namespace own {
: flags((type & TYPE_MASK)), rows(_rows), cols(_cols), data((uchar*)_data), step(_step == AUTO_STEP ? detail::default_step(type, _cols) : _step)
{}
+ MatHeader(const std::vector<int> &_dims, int type, void* _data)
+ : flags((type & TYPE_MASK)), data((uchar*)_data), step(0), dims(_dims)
+ {}
+
MatHeader(const MatHeader& ) = default;
MatHeader(MatHeader&& src) : MatHeader(src) // reuse copy constructor here
{
//! pointer to the data
uchar* data = nullptr;
size_t step = 0;
+ //! dimensions (ND-case)
+ std::vector<int> dims;
};
- }
+ } // namespace detail
//concise version of cv::Mat suitable for GAPI needs (used when no dependence on OpenCV is required)
class Mat : public detail::MatHeader{
public:
: MatHeader (_rows, _cols, _type, _data, _step)
{}
+ Mat(const std::vector<int> &_dims, int _type, void* _data)
+ : MatHeader (_dims, _type, _data)
+ {}
+
+ Mat(std::vector<int> &&_dims, int _type, void* _data)
+ : MatHeader (std::move(_dims), _type, _data)
+ {}
+
Mat(Mat const& src, const Rect& roi )
: Mat(src)
{
Mat& operator = (const Scalar& s)
{
constexpr unsigned max_channels = 4; //Scalar can't fit more than 4
- const auto channels = static_cast<unsigned int>(this->channels());
- GAPI_Assert(channels <= max_channels);
-
using func_p_t = void (*)(void*, int, Scalar const&);
using detail::assign_row;
#define TABLE_ENTRY(type) {assign_row<type, 1>, assign_row<type, 2>, assign_row<type, 3>, assign_row<type, 4>}
const auto depth = static_cast<unsigned int>(this->depth());
GAPI_Assert(depth < sizeof(func_tbl)/sizeof(func_tbl[0]));
- for (int r = 0; r < rows; ++r)
+ if (dims.empty())
+ {
+ const auto channels = static_cast<unsigned int>(this->channels());
+ GAPI_Assert(channels <= max_channels);
+
+ auto* f = func_tbl[depth][channels - 1];
+ for (int r = 0; r < rows; ++r)
+ {
+ (*f)(static_cast<void *>(ptr(r)), cols, s );
+ }
+ }
+ else
{
- auto* f = func_tbl[depth][channels -1];
- (*f)(static_cast<void *>(ptr(r)), cols, s );
+ auto* f = func_tbl[depth][0];
+ // FIXME: better to refactor assign_row to use std::size_t by default
+ (*f)(static_cast<void *>(data), static_cast<int>(total()), s);
}
return *this;
}
/** @brief Returns the number of matrix channels.
The method returns the number of matrix channels.
+ If matrix is N-dimensional, -1 is returned.
*/
- int channels() const {return CV_MAT_CN(flags);}
+ int channels() const {return dims.empty() ? CV_MAT_CN(flags) : -1;}
/**
@param _rows New number of rows.
*/
void create(int _rows, int _cols, int _type)
{
- create({_cols, _rows}, _type);
+ create(Size{_cols, _rows}, _type);
}
/** @overload
@param _size Alternative new matrix size specification: Size(cols, rows)
}
}
+ void create(const std::vector<int> &_dims, int _type)
+ {
+ // FIXME: make a proper reallocation-on-demands
+ // WARNING: no tensor views, so no strides
+ Mat tmp{_dims, _type, nullptr};
+ // FIXME: this accumulate duplicates a lot
+ const auto sz = std::accumulate(_dims.begin(), _dims.end(), 1, std::multiplies<int>());
+ tmp.memory.reset(new uchar[CV_ELEM_SIZE(_type)*sz], [](uchar * p){delete[] p;});
+ tmp.data = tmp.memory.get();
+ *this = std::move(tmp);
+ }
+
/** @brief Copies the matrix to another one.
The method copies the matrix data to another matrix. Before copying the data, the method invokes :
*/
void copyTo(Mat& dst) const
{
- dst.create(rows, cols, type());
- for (int r = 0; r < rows; ++r)
+ if (dims.empty())
{
- std::copy_n(ptr(r), detail::default_step(type(),cols), dst.ptr(r));
+ dst.create(rows, cols, type());
+ for (int r = 0; r < rows; ++r)
+ {
+ std::copy_n(ptr(r), detail::default_step(type(),cols), dst.ptr(r));
+ }
+ }
+ else
+ {
+ dst.create(dims, depth());
+ std::copy_n(data, total()*elemSize(), data);
}
}
*/
size_t total() const
{
- return static_cast<size_t>(rows * cols);
+ return static_cast<std::size_t>
+ (dims.empty()
+ ? (rows * cols)
+ : std::accumulate(dims.begin(), dims.end(), 1, std::multiplies<int>()));
}
-
/** @overload
@param roi Extracted submatrix specified as a rectangle.
*/
#include "precomp.hpp"
#include <memory> // unique_ptr
+#include <functional> // multiplies
#include <opencv2/gapi/gkernel.hpp>
#include <opencv2/gapi/own/convert.hpp>
} // namespace magazine
-void createMat(const cv::GMatDesc desc, cv::gapi::own::Mat& mat)
+void createMat(const cv::GMatDesc &desc, cv::gapi::own::Mat& mat)
{
- const auto type = desc.planar ? desc.depth : CV_MAKETYPE(desc.depth, desc.chan);
- const auto size = desc.planar ? cv::gapi::own::Size{desc.size.width, desc.size.height*desc.chan}
- : desc.size;
- mat.create(size, type);
+ // FIXME: Refactor (probably start supporting N-Dimensional blobs natively
+ if (desc.dims.empty())
+ {
+ const auto type = desc.planar ? desc.depth : CV_MAKETYPE(desc.depth, desc.chan);
+ const auto size = desc.planar ? cv::gapi::own::Size{desc.size.width, desc.size.height*desc.chan}
+ : desc.size;
+ mat.create(size, type);
+ }
+ else
+ {
+ GAPI_Assert(!desc.planar);
+ mat.create(desc.dims, desc.depth);
+ }
}
#if !defined(GAPI_STANDALONE)
-void createMat(const cv::GMatDesc desc, cv::Mat& mat)
+void createMat(const cv::GMatDesc &desc, cv::Mat& mat)
{
- const auto type = desc.planar ? desc.depth : CV_MAKETYPE(desc.depth, desc.chan);
- const auto size = desc.planar ? cv::Size{desc.size.width, desc.size.height*desc.chan}
- : cv::gapi::own::to_ocv(desc.size);
- mat.create(size, type);
+ // FIXME: Refactor (probably start supporting N-Dimensional blobs natively
+ if (desc.dims.empty())
+ {
+ const auto type = desc.planar ? desc.depth : CV_MAKETYPE(desc.depth, desc.chan);
+ const auto size = desc.planar ? cv::Size{desc.size.width, desc.size.height*desc.chan}
+ : cv::gapi::own::to_ocv(desc.size);
+ mat.create(size, type);
+ }
+ else
+ {
+ GAPI_Assert(!desc.planar);
+ mat.create(desc.dims, desc.depth);
+ }
}
#endif
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2018-2019 Intel Corporation
+
+
+#include "precomp.hpp"
+
+#include <functional> // hash
+#include <numeric> // accumulate
+#include <unordered_set>
+#include <iterator>
+
+#include <ade/util/algorithm.hpp>
+
+#include <opencv2/gapi/infer.hpp>
+
+cv::gapi::GNetPackage::GNetPackage(std::initializer_list<GNetParam> &&ii)
+ : networks(std::move(ii)) {
+}
+
+std::vector<cv::gapi::GBackend> cv::gapi::GNetPackage::backends() const {
+ std::unordered_set<cv::gapi::GBackend> unique_set;
+ for (const auto &nn : networks) unique_set.insert(nn.backend);
+ return std::vector<cv::gapi::GBackend>(unique_set.begin(), unique_set.end());
+}
#include "precomp.hpp"
+
+#include <ade/util/iota_range.hpp>
+#include <ade/util/algorithm.hpp>
+
#include <opencv2/gapi/opencv_includes.hpp>
#include <opencv2/gapi/own/mat.hpp> //gapi::own::Mat
#include <opencv2/gapi/gmat.hpp>
#if !defined(GAPI_STANDALONE)
cv::GMatDesc cv::descr_of(const cv::Mat &mat)
{
- return GMatDesc{mat.depth(), mat.channels(), {mat.cols, mat.rows}};
+ const auto mat_dims = mat.size.dims();
+
+ if (mat_dims == 2)
+ return GMatDesc{mat.depth(), mat.channels(), {mat.cols, mat.rows}};
+
+ std::vector<int> dims(mat_dims);
+ for (auto i : ade::util::iota(mat_dims)) {
+ // Note: cv::MatSize is not iterable
+ dims[i] = mat.size[i];
+ }
+ return GMatDesc{mat.depth(), std::move(dims)};
}
cv::GMatDesc cv::descr_of(const cv::UMat &mat)
{
+ GAPI_Assert(mat.size.dims() == 2);
return GMatDesc{ mat.depth(), mat.channels(),{ mat.cols, mat.rows } };
}
-cv::GMetaArgs cv::descr_of(const std::vector<cv::Mat> &vec)
+cv::GMetaArgs cv::descrs_of(const std::vector<cv::Mat> &vec)
{
return vec_descr_of(vec);
}
-cv::GMetaArgs cv::descr_of(const std::vector<cv::UMat> &vec)
+cv::GMetaArgs cv::descrs_of(const std::vector<cv::UMat> &vec)
{
return vec_descr_of(vec);
}
cv::GMatDesc cv::gapi::own::descr_of(const cv::gapi::own::Mat &mat)
{
- return GMatDesc{mat.depth(), mat.channels(), {mat.cols, mat.rows}};
+ return (mat.dims.empty())
+ ? GMatDesc{mat.depth(), mat.channels(), {mat.cols, mat.rows}}
+ : GMatDesc{mat.depth(), mat.dims};
}
-cv::GMetaArgs cv::gapi::own::descr_of(const std::vector<cv::gapi::own::Mat> &vec)
+cv::GMetaArgs cv::gapi::own::descrs_of(const std::vector<cv::gapi::own::Mat> &vec)
{
return vec_descr_of(vec);
}
return cv::util::optional<T>();
}
-void createMat(const cv::GMatDesc desc, cv::gapi::own::Mat& mat);
+void createMat(const cv::GMatDesc& desc, cv::gapi::own::Mat& mat);
#if !defined(GAPI_STANDALONE)
-void createMat(const cv::GMatDesc desc, cv::Mat& mat);
+void createMat(const cv::GMatDesc& desc, cv::Mat& mat);
#endif
}} // cv::gimpl
#include "precomp.hpp"
-#include <functional>
-#include <unordered_set>
-
#include <ade/util/algorithm.hpp>
#include <ade/util/range.hpp>
#include "compiler/gmodel.hpp"
#include "backends/cpu/gcpubackend.hpp"
-#include <opencv2/gapi/cpu/imgproc.hpp>
-#include <opencv2/gapi/cpu/core.hpp>
#include "api/gbackend_priv.hpp" // FIXME: Make it part of Backend SDK!
return this_backend;
}
-// GCPUExcecutable implementation //////////////////////////////////////////////
+// GCPUExecutable implementation //////////////////////////////////////////////
cv::gimpl::GCPUExecutable::GCPUExecutable(const ade::Graph &g,
const std::vector<ade::NodeHandle> &nodes)
: m_g(g), m_gm(m_g)
{
m_dataNodes.push_back(nh);
const auto &desc = m_gm.metadata(nh).get<Data>();
- if (desc.storage == Data::Storage::CONST)
+ if (desc.storage == Data::Storage::CONST_VAL)
{
auto rc = RcDesc{desc.rc, desc.shape, desc.ctor};
magazine::bindInArg(m_res, rc, m_gm.metadata(nh).get<ConstValue>().arg);
}}
-#endif // OPENCV_GAPI_GBACKEND_HPP
+#endif // OPENCV_GAPI_GCPUBACKEND_HPP
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2018 Intel Corporation
+
+#include "precomp.hpp"
+
+#ifdef HAVE_INF_ENGINE
+
+#if INF_ENGINE_RELEASE <= 2018050000
+# error G-API IE module supports only OpenVINO IE >= 2019 R1
+#endif
+
+#include <functional>
+#include <unordered_set>
+
+#include <ade/util/algorithm.hpp>
+
+#include <ade/util/range.hpp>
+#include <ade/util/zip_range.hpp>
+#include <ade/util/chain_range.hpp>
+#include <ade/typed_graph.hpp>
+
+#include <opencv2/gapi/gcommon.hpp>
+#include <opencv2/gapi/garray.hpp>
+#include <opencv2/gapi/util/any.hpp>
+#include <opencv2/gapi/gtype_traits.hpp>
+
+#include <opencv2/gapi/infer.hpp>
+#include <opencv2/gapi/infer/ie/util.hpp>
+
+#include "compiler/gobjref.hpp"
+#include "compiler/gmodel.hpp"
+
+#include "backends/ie/giebackend.hpp"
+
+#include "api/gbackend_priv.hpp" // FIXME: Make it part of Backend SDK!
+
+namespace IE = InferenceEngine;
+
+namespace {
+
+inline IE::ROI toIE(const cv::Rect &rc) {
+ return IE::ROI
+ { 0u
+ , static_cast<std::size_t>(rc.x)
+ , static_cast<std::size_t>(rc.y)
+ , static_cast<std::size_t>(rc.width)
+ , static_cast<std::size_t>(rc.height)
+ };
+}
+
+inline IE::SizeVector toIE(const cv::MatSize &sz) {
+ return cv::to_own<IE::SizeVector::value_type>(sz);
+}
+inline std::vector<int> toCV(const IE::SizeVector &vsz) {
+ std::vector<int> result;
+ result.reserve(vsz.size());
+ for (auto sz : vsz) {
+ result.push_back(ade::util::checked_cast<int>(sz));
+ }
+ return result;
+}
+
+inline IE::Precision toIE(int depth) {
+ switch (depth) {
+ case CV_8U: return IE::Precision::U8;
+ case CV_32F: return IE::Precision::FP32;
+ default: GAPI_Assert(false && "Unsupported data type");
+ }
+ return IE::Precision::UNSPECIFIED;
+}
+inline int toCV(IE::Precision prec) {
+ switch (prec) {
+ case IE::Precision::U8: return CV_8U;
+ case IE::Precision::FP32: return CV_32F;
+ default: GAPI_Assert(false && "Unsupported data type");
+ }
+ return -1;
+}
+
+inline IE::TensorDesc toIE(const cv::Mat &mat) {
+ const auto &sz = mat.size;
+
+ // NB: For some reason RGB image is 2D image
+ // (since channel component is not counted here).
+ if (sz.dims() == 2) {
+ // NB: This logic is mainly taken from IE samples
+ const size_t channels = mat.channels();
+ const size_t height = mat.size().height;
+ const size_t width = mat.size().width;
+
+ const size_t strideH = mat.step.buf[0];
+ const size_t strideW = mat.step.buf[1];
+
+ const bool is_dense =
+ strideW == channels &&
+ strideH == channels * width;
+
+ if (!is_dense)
+ cv::util::throw_error(std::logic_error("Doesn't support conversion"
+ " from non-dense cv::Mat"));
+
+ return IE::TensorDesc(toIE(mat.depth()),
+ IE::SizeVector{1, channels, height, width},
+ IE::Layout::NHWC);
+ }
+
+ GAPI_Assert(sz.dims() == 4); // NB: Will relax when needed (to known use)
+ return IE::TensorDesc(toIE(mat.depth()), toIE(sz), IE::Layout::NCHW);
+}
+
+inline IE::Blob::Ptr wrapIE(const cv::Mat &mat) {
+ const auto tDesc = toIE(mat);
+ switch (mat.depth()) {
+ // NB: Seems there's no way to create an untyped (T-less) Blob::Ptr
+ // in IE given only precision via TensorDesc. So we have to do this:
+#define HANDLE(E,T) \
+ case CV_##E: return IE::make_shared_blob<T>(tDesc, const_cast<T*>(mat.ptr<T>()))
+ HANDLE(8U, uint8_t);
+ HANDLE(32F, float);
+#undef HANDLE
+ default: GAPI_Assert(false && "Unsupported data type");
+ }
+ return IE::Blob::Ptr{};
+}
+
+template<class MatType>
+inline void copyFromIE(const IE::Blob::Ptr &blob, MatType &mat) {
+ switch (blob->getTensorDesc().getPrecision()) {
+#define HANDLE(E,T) \
+ case IE::Precision::E: std::copy_n(blob->buffer().as<T*>(), \
+ mat.total(), \
+ reinterpret_cast<T*>(mat.data)); \
+ break;
+ HANDLE(U8, uint8_t);
+ HANDLE(FP32, float);
+#undef HANDLE
+ default: GAPI_Assert(false && "Unsupported data type");
+ }
+}
+
+// IE-specific metadata, represents a network with its parameters
+struct IEUnit {
+ static const char *name() { return "IEModelConfig"; }
+
+ cv::gapi::ie::detail::ParamDesc params;
+ IE::CNNNetwork net;
+ IE::InputsDataMap inputs;
+ IE::OutputsDataMap outputs;
+
+ explicit IEUnit(const cv::gapi::ie::detail::ParamDesc &pp)
+ : params(pp) {
+
+ IE::CNNNetReader reader;
+ reader.ReadNetwork(params.model_path);
+ reader.ReadWeights(params.weights_path);
+ net = reader.getNetwork();
+ inputs = net.getInputsInfo();
+ outputs = net.getOutputsInfo();
+
+ // The practice shows that not all inputs and not all outputs
+ // are mandatory to specify in IE model.
+ // So what we're concerned here about is:
+ // if opeation's (not topology's) input/output number is
+ // greater than 1, then we do care about input/output layer
+ // names. Otherwise, names are picked up automatically.
+ // TODO: Probably this check could be done at the API entry point? (gnet)
+ if (params.num_in > 1u && params.num_in != params.input_names.size()) {
+ cv::util::throw_error(std::logic_error("Please specify input layer names for "
+ + params.model_path));
+ }
+ if (params.num_out > 1u && params.num_out != params.output_names.size()) {
+ cv::util::throw_error(std::logic_error("Please specify output layer names for "
+ + params.model_path));
+ }
+ if (params.num_in == 1u && params.input_names.empty()) {
+ params.input_names = { inputs.begin()->first };
+ }
+ if (params.num_out == 1u && params.output_names.empty()) {
+ params.output_names = { outputs.begin()->first };
+ }
+ }
+
+ // This method is [supposed to be] called at Island compilation stage
+ cv::gimpl::ie::IECompiled compile() const {
+ auto this_plugin = IE::PluginDispatcher().getPluginByDevice(params.device_id);
+ auto this_network = this_plugin.LoadNetwork(net, {}); // FIXME: 2nd parameter to be
+ // configurable via the API
+ auto this_request = this_network.CreateInferRequest();
+
+ // Bind const data to infer request
+ for (auto &&p : params.const_inputs) {
+ this_request.SetBlob(p.first, wrapIE(p.second));
+ }
+
+ return {this_plugin, this_network, this_request};
+ }
+};
+
+struct IECallContext
+{
+ // Input parameters passed to an inference operation.
+ std::vector<cv::GArg> args;
+
+ //FIXME: avoid conversion of arguments from internal representaion to OpenCV one on each call
+ //to OCV kernel. (This can be achieved by a two single time conversions in GCPUExecutable::run,
+ //once on enter for input and output arguments, and once before return for output arguments only
+ //FIXME: check if the above applies to this backend (taken from CPU)
+ std::unordered_map<std::size_t, cv::GRunArgP> results;
+
+ // Generic accessor API
+ template<typename T>
+ const T& inArg(std::size_t input) { return args.at(input).get<T>(); }
+
+ // Syntax sugar
+ const cv::gapi::own::Mat& inMat(std::size_t input) {
+ return inArg<cv::gapi::own::Mat>(input);
+ }
+ cv::gapi::own::Mat& outMatR(std::size_t output) {
+ return *cv::util::get<cv::gapi::own::Mat*>(results.at(output));
+ }
+
+ template<typename T> std::vector<T>& outVecR(std::size_t output) { // FIXME: the same issue
+ return outVecRef(output).wref<T>();
+ }
+ cv::detail::VectorRef& outVecRef(std::size_t output) {
+ return cv::util::get<cv::detail::VectorRef>(results.at(output));
+ }
+};
+
+struct IECallable {
+ static const char *name() { return "IERequestCallable"; }
+ // FIXME: Make IECallContext manage them all? (3->1)
+ using Run = std::function<void(cv::gimpl::ie::IECompiled &, const IEUnit &, IECallContext &)>;
+ Run run;
+};
+
+struct KImpl {
+ cv::gimpl::CustomMetaFunction::CM customMetaFunc;
+ IECallable::Run run;
+};
+
+// FIXME: Is there a way to take a typed graph (our GModel),
+// and create a new typed graph _ATOP_ of that (by extending with a couple of
+// new types?).
+// Alternatively, is there a way to compose types graphs?
+//
+// If not, we need to introduce that!
+using GIEModel = ade::TypedGraph
+ < cv::gimpl::Protocol
+ , cv::gimpl::Op
+ , cv::gimpl::NetworkParams
+ , cv::gimpl::CustomMetaFunction
+ , IEUnit
+ , IECallable
+ >;
+
+// FIXME: Same issue with Typed and ConstTyped
+using GConstGIEModel = ade::ConstTypedGraph
+ < cv::gimpl::Protocol
+ , cv::gimpl::Op
+ , cv::gimpl::NetworkParams
+ , cv::gimpl::CustomMetaFunction
+ , IEUnit
+ , IECallable
+ >;
+} // anonymous namespace
+
+// GCPUExcecutable implementation //////////////////////////////////////////////
+cv::gimpl::ie::GIEExecutable::GIEExecutable(const ade::Graph &g,
+ const std::vector<ade::NodeHandle> &nodes)
+ : m_g(g), m_gm(m_g) {
+ // FIXME: Currently this backend is capable to run a single inference node only.
+ // Need to extend our island fusion with merge/not-to-merge decision making parametrization
+ GConstGIEModel iem(g);
+
+ for (auto &nh : nodes) {
+ switch (m_gm.metadata(nh).get<NodeType>().t) {
+ case NodeType::OP:
+ if (this_nh == nullptr) {
+ this_nh = nh;
+ this_iec = iem.metadata(this_nh).get<IEUnit>().compile();
+ }
+ else
+ util::throw_error(std::logic_error("Multi-node inference is not supported!"));
+ break;
+
+ case NodeType::DATA: {
+ m_dataNodes.push_back(nh);
+ const auto &desc = m_gm.metadata(nh).get<Data>();
+ if (desc.storage == Data::Storage::CONST_VAL) {
+ util::throw_error(std::logic_error("No const data please!"));
+ }
+ if (desc.storage == Data::Storage::INTERNAL) {
+ util::throw_error(std::logic_error("No internal data please!"));
+ }
+ break;
+ }
+ default: util::throw_error(std::logic_error("Unsupported NodeType type"));
+ }
+ }
+}
+
+// FIXME: Document what it does
+cv::GArg cv::gimpl::ie::GIEExecutable::packArg(const cv::GArg &arg) {
+ // No API placeholders allowed at this point
+ // FIXME: this check has to be done somewhere in compilation stage.
+ GAPI_Assert( arg.kind != cv::detail::ArgKind::GMAT
+ && arg.kind != cv::detail::ArgKind::GSCALAR
+ && arg.kind != cv::detail::ArgKind::GARRAY);
+
+ if (arg.kind != cv::detail::ArgKind::GOBJREF) {
+ util::throw_error(std::logic_error("Inference supports G-types ONLY!"));
+ }
+ GAPI_Assert(arg.kind == cv::detail::ArgKind::GOBJREF);
+
+ // Wrap associated CPU object (either host or an internal one)
+ // FIXME: object can be moved out!!! GExecutor faced that.
+ const cv::gimpl::RcDesc &ref = arg.get<cv::gimpl::RcDesc>();
+ switch (ref.shape)
+ {
+ case GShape::GMAT: return GArg(m_res.slot<cv::gapi::own::Mat>()[ref.id]);
+
+ // Note: .at() is intentional for GArray as object MUST be already there
+ // (and constructed by either bindIn/Out or resetInternal)
+ case GShape::GARRAY: return GArg(m_res.slot<cv::detail::VectorRef>().at(ref.id));
+
+ default:
+ util::throw_error(std::logic_error("Unsupported GShape type"));
+ break;
+ }
+}
+
+void cv::gimpl::ie::GIEExecutable::run(std::vector<InObj> &&input_objs,
+ std::vector<OutObj> &&output_objs) {
+ // Update resources with run-time information - what this Island
+ // has received from user (or from another Island, or mix...)
+ // FIXME: Check input/output objects against GIsland protocol
+
+ for (auto& it : input_objs) magazine::bindInArg (m_res, it.first, it.second);
+ for (auto& it : output_objs) magazine::bindOutArg(m_res, it.first, it.second);
+
+ // FIXME: Running just a single node now.
+ // Not sure if need to support many of them, though
+ // FIXME: Make this island-unmergeable?
+ const auto &op = m_gm.metadata(this_nh).get<Op>();
+
+ // Initialize kernel's execution context:
+ // - Input parameters
+ IECallContext context;
+ context.args.reserve(op.args.size());
+ using namespace std::placeholders;
+ ade::util::transform(op.args,
+ std::back_inserter(context.args),
+ std::bind(&GIEExecutable::packArg, this, _1));
+
+ // - Output parameters.
+ for (const auto &out_it : ade::util::indexed(op.outs)) {
+ // FIXME: Can the same GArg type resolution mechanism be reused here?
+ const auto out_port = ade::util::index(out_it);
+ const auto out_desc = ade::util::value(out_it);
+ context.results[out_port] = magazine::getObjPtr(m_res, out_desc);
+ }
+
+ // And now trigger the execution
+ GConstGIEModel giem(m_g);
+ const auto &uu = giem.metadata(this_nh).get<IEUnit>();
+ const auto &kk = giem.metadata(this_nh).get<IECallable>();
+ kk.run(this_iec, uu, context);
+
+ for (auto &it : output_objs) magazine::writeBack(m_res, it.first, it.second);
+}
+
+namespace cv {
+namespace gimpl {
+namespace ie {
+
+struct Infer: public cv::detail::KernelTag {
+ using API = cv::GInferBase;
+ static cv::gapi::GBackend backend() { return cv::gapi::ie::backend(); }
+ static KImpl kernel() { return KImpl{outMeta, run}; }
+
+ static cv::GMetaArgs outMeta(const ade::Graph &gr,
+ const ade::NodeHandle &nh,
+ const cv::GMetaArgs &in_metas,
+ const cv::GArgs &/*in_args*/) {
+ // Specify network's output layer metadata to the framework
+ // Also specify the input information to the IE from the framework
+ // NB: Have no clue if network's input [dimensions] may ever define
+ // its output dimensions. It seems possible with OpenCV DNN APIs
+
+ cv::GMetaArgs result;
+
+ GConstGIEModel gm(gr);
+ const auto &uu = gm.metadata(nh).get<IEUnit>();
+
+ // Initialize input information
+ // Note our input layers list order matches the API order and so
+ // meta order.
+ GAPI_Assert(uu.params.input_names.size() == in_metas.size()
+ && "Known input layers count doesn't match input meta count");
+
+ for (auto &&it : ade::util::zip(ade::util::toRange(uu.params.input_names),
+ ade::util::toRange(in_metas))) {
+ auto &&ii = uu.inputs.at(std::get<0>(it));
+ const auto & mm = std::get<1>(it);
+
+ GAPI_Assert(util::holds_alternative<cv::GMatDesc>(mm)
+ && "Non-GMat inputs are not supported");
+
+ const auto &meta = util::get<cv::GMatDesc>(mm);
+ ii->setPrecision(toIE(meta.depth));
+ ii->setLayout(meta.isND() ? IE::Layout::NCHW : IE::Layout::NHWC);
+ ii->getPreProcess().setResizeAlgorithm(IE::RESIZE_BILINEAR);
+ }
+
+ // FIXME: It would be nice here to have an exact number of network's
+ // input/output parameters. Probably GCall should store it here for us.
+ // It doesn't, as far as I know..
+ for (const auto &out_name : uu.params.output_names) {
+ // NOTE: our output_names vector follows the API order
+ // of this operation's outputs
+ const IE::DataPtr& ie_out = uu.outputs.at(out_name);
+ const IE::SizeVector dims = ie_out->getTensorDesc().getDims();
+
+ cv::GMatDesc outm(toCV(ie_out->getPrecision()),
+ toCV(ie_out->getTensorDesc().getDims()));
+ result.emplace_back(outm);
+ }
+ return result;
+ }
+
+ static void run(IECompiled &iec, const IEUnit &uu, IECallContext &ctx) {
+ // non-generic version for now:
+ // - assumes all inputs/outputs are always Mats
+ for (auto i : ade::util::iota(uu.params.num_in)) {
+ // TODO: Ideally we shouldn't do SetBlob() but GetBlob() instead,
+ // and redirect our data producers to this memory
+ // (A memory dialog comes to the picture again)
+
+ const cv::Mat this_mat = to_ocv(ctx.inMat(i));
+ IE::Blob::Ptr this_blob = wrapIE(this_mat);
+ iec.this_request.SetBlob(uu.params.input_names[i], this_blob);
+ }
+ iec.this_request.Infer();
+ for (auto i : ade::util::iota(uu.params.num_out)) {
+ // TODO: Think on avoiding copying here.
+ // Either we should ask IE to use our memory (what is not always the
+ // best policy) or use IE-allocated buffer inside (and pass it to the graph).
+ // Not a <very> big deal for classifiers and detectors,
+ // but may be critical to segmentation.
+
+ cv::gapi::own::Mat& out_mat = ctx.outMatR(i);
+ IE::Blob::Ptr this_blob = iec.this_request.GetBlob(uu.params.output_names[i]);
+ copyFromIE(this_blob, out_mat);
+ }
+ }
+};
+
+struct InferList: public cv::detail::KernelTag {
+ using API = cv::GInferListBase;
+ static cv::gapi::GBackend backend() { return cv::gapi::ie::backend(); }
+ static KImpl kernel() { return KImpl{outMeta, run}; }
+
+ static cv::GMetaArgs outMeta(const ade::Graph &gr,
+ const ade::NodeHandle &nh,
+ const cv::GMetaArgs &in_metas,
+ const cv::GArgs &/*in_args*/) {
+ // Specify the input information to the IE from the framework
+ // NB: Have no clue if network's input [dimensions] may ever define
+ // its output dimensions. It seems possible with OpenCV DNN APIs
+
+ GConstGIEModel gm(gr);
+ const auto &uu = gm.metadata(nh).get<IEUnit>();
+
+ // Initialize input information
+ // Note our input layers list order matches the API order and so
+ // meta order.
+ GAPI_Assert(uu.params.input_names.size() == (in_metas.size() - 1u)
+ && "Known input layers count doesn't match input meta count");
+
+ std::size_t idx = 1u;
+ for (auto &&input_name : uu.params.input_names) {
+ auto &&ii = uu.inputs.at(input_name);
+ const auto & mm = in_metas[idx++];
+
+ GAPI_Assert(util::holds_alternative<cv::GMatDesc>(mm)
+ && "Non-GMat inputs are not supported");
+
+ const auto &meta = util::get<cv::GMatDesc>(mm);
+ ii->setPrecision(toIE(meta.depth));
+ ii->setLayout(meta.isND() ? IE::Layout::NCHW : IE::Layout::NHWC);
+ ii->getPreProcess().setResizeAlgorithm(IE::RESIZE_BILINEAR);
+ }
+
+ // roi-list version is much easier at the moment.
+ // All our outputs are vectors which don't have
+ // metadata at the moment - so just create a vector of
+ // "empty" array metadatas of the required size.
+ return cv::GMetaArgs(uu.params.output_names.size(),
+ cv::GMetaArg{cv::empty_array_desc()});
+ }
+
+ static void run(IECompiled &iec, const IEUnit &uu, IECallContext &ctx) {
+ // non-generic version for now:
+ // - assumes zero input is always ROI list
+ // - assumes all inputs/outputs are always Mats
+ GAPI_Assert(uu.params.num_in == 1); // roi list is not counted in net's inputs
+
+ const auto& in_roi_vec = ctx.inArg<cv::detail::VectorRef>(0u).rref<cv::Rect>();
+ const cv::Mat this_mat = to_ocv(ctx.inMat(1u));
+ IE::Blob::Ptr this_blob = wrapIE(this_mat);
+
+ // FIXME: This could be done ONCE at graph compile stage!
+ std::vector< std::vector<int> > cached_dims(uu.params.num_out);
+ for (auto i : ade::util::iota(uu.params.num_out)) {
+ const IE::DataPtr& ie_out = uu.outputs.at(uu.params.output_names[i]);
+ cached_dims[i] = toCV(ie_out->getTensorDesc().getDims());
+ ctx.outVecR<cv::Mat>(i).clear();
+ // FIXME: Isn't this should be done automatically
+ // by some resetInternalData(), etc? (Probably at the GExecutor level)
+ }
+
+ for (const auto &rc : in_roi_vec) {
+ // FIXME: Assumed only 1 input
+ IE::Blob::Ptr roi_blob = IE::make_shared_blob(this_blob, toIE(rc));
+ iec.this_request.SetBlob(uu.params.input_names[0u], roi_blob);
+ iec.this_request.Infer();
+
+ // While input is fixed to be 1,
+ // there may be still multiple outputs
+ for (auto i : ade::util::iota(uu.params.num_out)) {
+ std::vector<cv::Mat> &out_vec = ctx.outVecR<cv::Mat>(i);
+
+ IE::Blob::Ptr out_blob = iec.this_request.GetBlob(uu.params.output_names[i]);
+
+ cv::Mat out_mat(cached_dims[i], toCV(out_blob->getTensorDesc().getPrecision()));
+ copyFromIE(out_blob, out_mat); // FIXME: Avoid data copy. Not sure if it is possible though
+ out_vec.push_back(std::move(out_mat));
+ }
+ }
+ }
+};
+
+} // namespace ie
+} // namespace gapi
+} // namespace cv
+
+
+// IE backend implementation of GBackend::Priv ///////////////////////
+namespace {
+ class GIEBackendImpl final: public cv::gapi::GBackend::Priv {
+ virtual void unpackKernel(ade::Graph &gr,
+ const ade::NodeHandle &nh,
+ const cv::GKernelImpl &ii) override {
+ using namespace cv::gimpl;
+ // FIXME: Introduce a DNNBackend interface which'd specify
+ // the framework for this???
+ GIEModel gm(gr);
+ const auto &np = gm.metadata(nh).get<NetworkParams>();
+ const auto &pp = cv::util::any_cast<cv::gapi::ie::detail::ParamDesc>(np.opaque);
+ const auto &ki = cv::util::any_cast<KImpl>(ii.opaque);
+ gm.metadata(nh).set(IEUnit{pp});
+ gm.metadata(nh).set(IECallable{ki.run});
+ gm.metadata(nh).set(CustomMetaFunction{ki.customMetaFunc});
+ }
+
+ virtual EPtr compile(const ade::Graph &graph,
+ const cv::GCompileArgs &,
+ const std::vector<ade::NodeHandle> &nodes) const override {
+ return EPtr{new cv::gimpl::ie::GIEExecutable(graph, nodes)};
+ }
+
+ virtual cv::gapi::GKernelPackage auxiliaryKernels() const override {
+ return cv::gapi::kernels< cv::gimpl::ie::Infer
+ , cv::gimpl::ie::InferList
+ >();
+ }
+ };
+}
+
+cv::gapi::GBackend cv::gapi::ie::backend() {
+ static cv::gapi::GBackend this_backend(std::make_shared<GIEBackendImpl>());
+ return this_backend;
+}
+
+cv::Mat cv::gapi::ie::util::to_ocv(InferenceEngine::Blob::Ptr blob) {
+ const auto& tdesc = blob->getTensorDesc();
+ return cv::Mat(toCV(tdesc.getDims()),
+ toCV(tdesc.getPrecision()),
+ blob->buffer().as<uint8_t*>());
+}
+
+std::vector<int> cv::gapi::ie::util::to_ocv(const InferenceEngine::SizeVector &dims) {
+ return toCV(dims);
+}
+
+InferenceEngine::Blob::Ptr cv::gapi::ie::util::to_ie(cv::Mat &blob) {
+ return wrapIE(blob);
+}
+
+#endif // HAVE_INF_ENGINE
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2018 Intel Corporation
+
+#ifndef OPENCV_GAPI_GIEBACKEND_HPP
+#define OPENCV_GAPI_GIEBACKEND_HPP
+
+#ifdef HAVE_INF_ENGINE
+
+#include <ade/util/algorithm.hpp> // type_list_index
+
+////////////////////////////////////////////////////////////////////////////////
+// FIXME: Suppress deprecation warnings for OpenVINO 2019R2+
+// BEGIN {{{
+#if defined(__GNUC__)
+#pragma GCC diagnostic ignored "-Wdeprecated-declarations"
+#endif
+#ifdef _MSC_VER
+#pragma warning(disable: 4996) // was declared deprecated
+#endif
+
+#if defined(__GNUC__)
+#pragma GCC visibility push(default)
+#endif
+
+#include <inference_engine.hpp>
+
+#if defined(__GNUC__)
+#pragma GCC visibility pop
+#endif
+// END }}}
+////////////////////////////////////////////////////////////////////////////////
+
+#include <opencv2/gapi/garg.hpp>
+#include <opencv2/gapi/gproto.hpp>
+#include <opencv2/gapi/infer/ie.hpp>
+
+#include "api/gorigin.hpp"
+#include "backends/common/gbackend.hpp"
+#include "compiler/gislandmodel.hpp"
+
+namespace cv {
+namespace gimpl {
+namespace ie {
+
+struct IECompiled {
+ InferenceEngine::InferencePlugin this_plugin;
+ InferenceEngine::ExecutableNetwork this_network;
+ InferenceEngine::InferRequest this_request;
+};
+
+class GIEExecutable final: public GIslandExecutable
+{
+ const ade::Graph &m_g;
+ GModel::ConstGraph m_gm;
+
+ // The only executable stuff in this graph
+ // (assuming it is always single-op)
+ ade::NodeHandle this_nh;
+ IECompiled this_iec;
+
+ // List of all resources in graph (both internal and external)
+ std::vector<ade::NodeHandle> m_dataNodes;
+
+ // Actual data of all resources in graph (both internal and external)
+ Mag m_res;
+
+ // Execution helpers
+ GArg packArg(const GArg &arg);
+
+public:
+ GIEExecutable(const ade::Graph &graph,
+ const std::vector<ade::NodeHandle> &nodes);
+
+ virtual inline bool canReshape() const override { return false; }
+ virtual inline void reshape(ade::Graph&, const GCompileArgs&) override {
+ GAPI_Assert(false); // Not implemented yet
+ }
+
+ virtual void run(std::vector<InObj> &&input_objs,
+ std::vector<OutObj> &&output_objs) override;
+};
+
+}}}
+
+#endif // HAVE_INF_ENGINE
+#endif // OPENCV_GAPI_GIEBACKEND_HPP
#include "precomp.hpp"
-#include <functional>
-#include <unordered_set>
-
#include <ade/util/algorithm.hpp>
#include <ade/util/range.hpp>
#include "compiler/gmodel.hpp"
#include "backends/ocl/goclbackend.hpp"
-#include "backends/ocl/goclimgproc.hpp"
-#include "backends/ocl/goclcore.hpp"
#include "api/gbackend_priv.hpp" // FIXME: Make it part of Backend SDK!
{
m_dataNodes.push_back(nh);
const auto &desc = m_gm.metadata(nh).get<Data>();
- if (desc.storage == Data::Storage::CONST)
+ if (desc.storage == Data::Storage::CONST_VAL)
{
auto rc = RcDesc{desc.rc, desc.shape, desc.ctor};
magazine::bindInArg(m_res, rc, m_gm.metadata(nh).get<ConstValue>().arg);
return combine(ocv_pkg, user_pkg_with_aux);
}
+ cv::gapi::GNetPackage getNetworkPackage(cv::GCompileArgs &args)
+ {
+ return cv::gimpl::getCompileArg<cv::gapi::GNetPackage>(args)
+ .value_or(cv::gapi::GNetPackage{});
+ }
+
cv::util::optional<std::string> getGraphDumpDirectory(cv::GCompileArgs& args)
{
auto dump_info = cv::gimpl::getCompileArg<cv::graph_dump_path>(args);
return cv::util::make_optional(dump_info.value().m_dump_path);
}
}
+
+ template<typename C>
+ cv::gapi::GKernelPackage auxKernelsFrom(const C& c) {
+ cv::gapi::GKernelPackage result;
+ for (const auto &b : c) {
+ result = cv::gapi::combine(result, b.priv().auxiliaryKernels());
+ }
+ return result;
+ }
+
} // anonymous namespace
: m_c(c), m_metas(std::move(metas)), m_args(std::move(args))
{
using namespace std::placeholders;
- m_all_kernels = getKernelPackage(m_args);
- auto dump_path = getGraphDumpDirectory(m_args);
+
+ auto kernels_to_use = getKernelPackage(m_args);
+ auto networks_to_use = getNetworkPackage(m_args);
+ std::unordered_set<cv::gapi::GBackend> all_backends;
+ const auto take = [&](std::vector<cv::gapi::GBackend> &&v) {
+ all_backends.insert(v.begin(), v.end());
+ };
+ take(kernels_to_use.backends());
+ take(networks_to_use.backends());
+ m_all_kernels = cv::gapi::combine(kernels_to_use,
+ auxKernelsFrom(all_backends));
+ // NB: The expectation in the line above is that
+ // NN backends (present here via network package) always add their
+ // inference kernels via auxiliary...()
+
+ auto dump_path = getGraphDumpDirectory(m_args);
m_e.addPassStage("init");
m_e.addPass("init", "check_cycles", ade::passes::CheckCycles());
- m_e.addPass("init", "expand_kernels", std::bind(passes::expandKernels, _1,
- m_all_kernels)); // NB: package is copied
+ m_e.addPass("init", "expand_kernels",
+ std::bind(passes::expandKernels, _1,
+ m_all_kernels)); // NB: package is copied
m_e.addPass("init", "topo_sort", ade::passes::TopologicalSort());
m_e.addPass("init", "init_islands", passes::initIslands);
m_e.addPass("init", "check_islands", passes::checkIslands);
m_all_kernels.remove(cv::gapi::compound::backend());
m_e.addPassStage("kernels");
- m_e.addPass("kernels", "resolve_kernels", std::bind(passes::resolveKernels, _1,
- std::ref(m_all_kernels))); // NB: and not copied here
+ m_e.addPass("kernels", "bind_net_params",
+ std::bind(passes::bindNetParams, _1,
+ networks_to_use));
+ m_e.addPass("kernels", "resolve_kernels",
+ std::bind(passes::resolveKernels, _1,
+ std::ref(m_all_kernels))); // NB: and not copied here
+ // (no compound backend present here)
m_e.addPass("kernels", "check_islands_content", passes::checkIslandsContent);
m_e.addPassStage("meta");
dump_path.value()));
}
- // Process backends at the last moment (after all G-API passes are added).
+ // FIXME: This should be called for "ActiveBackends" only (see metadata).
+ // However, ActiveBackends are known only after passes are actually executed.
+ // At these stage, they are not executed yet.
ade::ExecutionEngineSetupContext ectx(m_e);
auto backends = m_all_kernels.backends();
for (auto &b : backends)
#include <opencv2/gapi/gcommon.hpp>
#include <opencv2/gapi/gkernel.hpp>
+#include <opencv2/gapi/infer.hpp>
#include <opencv2/gapi/gcomputation.hpp>
#include <ade/execution_engine/execution_engine.hpp>
ade::ExecutionEngine m_e;
cv::gapi::GKernelPackage m_all_kernels;
+ cv::gapi::GNetPackage m_all_networks;
void validateInputMeta();
void validateOutProtoArgs();
{
auto value = value_of(origin);
meta = descr_of(value);
- storage = Data::Storage::CONST;
+ storage = Data::Storage::CONST_VAL;
g.metadata(data_h).set(ConstValue{value});
}
g.metadata(data_h).set(Data{origin.shape, id, meta, origin.ctor, storage});
// This part of the system is API-unaware by its design.
//
+#include <opencv2/gapi/util/any.hpp>
+
#include <opencv2/gapi/garg.hpp>
#include <opencv2/gapi/gkernel.hpp>
INTERNAL, // data object is not listed in GComputation protocol
INPUT, // data object is listed in GComputation protocol as Input
OUTPUT, // data object is listed in GComputation protocol as Output
- CONST, // data object is constant
+ CONST_VAL, // data object is constant.
+ // Note: CONST is sometimes defined in Win sys headers
};
Storage storage;
};
std::unordered_set<cv::gapi::GBackend> backends;
};
+// Backend-specific inference parameters for a neural network.
+// Since these parameters are set on compilation stage (not
+// on a construction stage), these parameters are bound lately
+// to the operation node.
+// NB: These parameters are not included into GModel by default
+// since it is not used regularly by all parties.
+struct NetworkParams
+{
+ static const char *name() { return "NetworkParams"; }
+ cv::util::any opaque;
+};
+
+// This is a custom metadata handling operator.
+// Sometimes outMeta() can't be bound to input parameters only
+// so several backends (today -- mainly inference) may find this useful.
+// If provided, the meta inference pass uses this function instead of
+// OP.k.outMeta.
+struct CustomMetaFunction
+{
+ static const char *name() { return "CustomMetaFunction"; }
+ using CM = std::function< cv::GMetaArgs( const ade::Graph &,
+ const ade::NodeHandle &,
+ const cv::GMetaArgs &,
+ const cv::GArgs &)>;
+ CM customOutMeta;
+};
+
namespace GModel
{
using Graph = ade::TypedGraph
, DataObjectCounter
, IslandModel
, ActiveBackends
+ , CustomMetaFunction
>;
// FIXME: How to define it based on GModel???
, DataObjectCounter
, IslandModel
, ActiveBackends
+ , CustomMetaFunction
>;
// FIXME:
#include <ade/passes/check_cycles.hpp>
#include <opencv2/gapi/gcompoundkernel.hpp> // compound::backend()
+#include <opencv2/gapi/gkernel.hpp> // GKernelPackage
+#include <opencv2/gapi/infer.hpp> // GNetPackage
#include "compiler/gmodel.hpp"
#include "compiler/passes/passes.hpp"
gr.erase(subgr_out_nh);
}
}
+} // anonymous namespace
+
+// This pass, given the network package, associates every infer[list] node
+// with particular inference backend and its parameters.
+void cv::gimpl::passes::bindNetParams(ade::passes::PassContext &ctx,
+ const gapi::GNetPackage &pkg)
+{
+ GModel::Graph gr(ctx.graph);
+ ade::TypedGraph<NetworkParams> pgr(ctx.graph);
+
+ for (const auto &nh : gr.nodes())
+ {
+ if (gr.metadata(nh).get<NodeType>().t == NodeType::OP)
+ {
+ auto &op = gr.metadata(nh).get<Op>();
+ if (op.k.tag.empty())
+ continue;
+
+ // FIXME: What if there's more than one???
+ const auto it = ade::util::find_if(pkg.networks,
+ [&](const cv::gapi::GNetParam &p) {
+ return p.tag == op.k.tag;
+ });
+ if (it == std::end(pkg.networks))
+ continue;
+
+ pgr.metadata(nh).set(NetworkParams{it->params});
+ }
+ }
}
+
// This pass, given the kernel package, selects a kernel implementation
// for every operation in the graph
void cv::gimpl::passes::resolveKernels(ade::passes::PassContext &ctx,
// Prepare operation's input metadata vector
// Note that it's size is usually different from nh.inEdges.size(),
- // and its element count is equal to operation's arguments count.
+ // and its element count is equal to operation's arguments count
+ // (which may contain graph-construction-time parameters like integers, etc)
GMetaArgs input_meta_args(op.args.size());
// Iterate through input edges, update input_meta_args's slots
{
// No meta in an input argument - a fatal error
// (note graph is traversed here in topoligcal order)
- util::throw_error(std::logic_error("Fatal: input object's metadata "
- "not found!"));
+ util::throw_error(std::logic_error("Fatal: input object's metadata "
+ "not found!"));
// FIXME: Add more details!!!
}
input_meta_args.at(input_port) = input_meta;
}
+
// Now ask kernel for it's output meta.
// Resulting out_args may have a larger size than op.outs, since some
// outputs could stay unused (unconnected)
- const auto out_metas = op.k.outMeta(input_meta_args, op.args);
+ const auto out_metas = gr.metadata(nh).contains<CustomMetaFunction>()
+ ? gr.metadata(nh).get<CustomMetaFunction>().customOutMeta(ctx.graph,
+ nh,
+ input_meta_args,
+ op.args)
+ : op.k.outMeta(input_meta_args, op.args);
// Walk through operation's outputs, update meta of output objects
// appropriately
namespace cv {
+// Forward declarations - internal
+namespace gapi {
+ class GKernelPackage;
+ struct GNetPackage;
+} // namespace gapi
+
namespace gimpl { namespace passes {
void dumpDot(const ade::Graph &g, std::ostream& os);
void expandKernels(ade::passes::PassContext &ctx,
const gapi::GKernelPackage& kernels);
+void bindNetParams(ade::passes::PassContext &ctx,
+ const gapi::GNetPackage &networks);
+
void resolveKernels(ade::passes::PassContext &ctx,
const gapi::GKernelPackage &kernels);
const Data &d = m_gm.metadata(orig_nh).get<Data>();
if ( d.storage != Data::Storage::INTERNAL
- && d.storage != Data::Storage::CONST)
+ && d.storage != Data::Storage::CONST_VAL)
return;
// INTERNALS+CONST only! no need to allocate/reset output objects
break;
case GShape::GSCALAR:
- if (d.storage == Data::Storage::CONST)
+ if (d.storage == Data::Storage::CONST_VAL)
{
auto rc = RcDesc{d.rc, d.shape, d.ctor};
magazine::bindInArg(m_res, rc, m_gm.metadata(orig_nh).get<ConstValue>().arg);
EXPECT_EQ(1, desc1.chan);
EXPECT_EQ(320, desc1.size.width);
EXPECT_EQ(240, desc1.size.height);
+ EXPECT_FALSE(desc1.isND());
cv::Mat m2(480, 640, CV_8UC3);
const auto desc2 = cv::descr_of(m2);
EXPECT_EQ(3, desc2.chan);
EXPECT_EQ(640, desc2.size.width);
EXPECT_EQ(480, desc2.size.height);
+ EXPECT_FALSE(desc2.isND());
+}
+
+TEST(GAPI_MetaDesc, MatDescND)
+{
+ std::vector<int> dims = {1,3,299,299};
+ cv::Mat m(dims, CV_32F);
+ const auto desc = cv::descr_of(m);
+ EXPECT_EQ(CV_32F, desc.depth);
+ EXPECT_EQ(-1, desc.chan);
+ EXPECT_EQ(1, desc.dims[0]);
+ EXPECT_EQ(3, desc.dims[1]);
+ EXPECT_EQ(299, desc.dims[2]);
+ EXPECT_EQ(299, desc.dims[3]);
+ EXPECT_TRUE(desc.isND());
}
TEST(GAPI_MetaDesc, VecMatDesc)
std::vector<cv::Mat> vec1 = {
cv::Mat(240, 320, CV_8U)};
- const auto desc1 = cv::descr_of(vec1);
+ const auto desc1 = cv::descrs_of(vec1);
EXPECT_EQ((GMatDesc{CV_8U, 1, {320, 240}}), get<GMatDesc>(desc1[0]));
std::vector<cv::UMat> vec2 = {
cv::UMat(480, 640, CV_8UC3)};
- const auto desc2 = cv::descr_of(vec2);
+ const auto desc2 = cv::descrs_of(vec2);
EXPECT_EQ((GMatDesc{CV_8U, 3, {640, 480}}), get<GMatDesc>(desc2[0]));
}
cv::gapi::own::Mat(240, 320, CV_8U, nullptr),
cv::gapi::own::Mat(480, 640, CV_8UC3, nullptr)};
- const auto desc = cv::gapi::own::descr_of(vec);
+ const auto desc = cv::gapi::own::descrs_of(vec);
EXPECT_EQ((GMatDesc{CV_8U, 1, {320, 240}}), get<GMatDesc>(desc[0]));
EXPECT_EQ((GMatDesc{CV_8U, 3, {640, 480}}), get<GMatDesc>(desc[1]));
}
cv::gapi::own::Mat(240, 320, CV_8U, nullptr),
cv::gapi::own::Mat(480, 640, CV_8UC3, nullptr)};
- const auto desc = descr_of(vec);
+ const auto desc = descrs_of(vec);
EXPECT_EQ((GMatDesc{CV_8U, 1, {320, 240}}), get<GMatDesc>(desc[0]));
EXPECT_EQ((GMatDesc{CV_8U, 3, {640, 480}}), get<GMatDesc>(desc[1]));
}
EXPECT_TRUE(desc1 != desc2);
}
+TEST(GAPI_MetaDesc, Compare_Equal_MatDesc_ND)
+{
+ const auto desc1 = cv::GMatDesc{CV_8U, {1,3,224,224}};
+ const auto desc2 = cv::GMatDesc{CV_8U, {1,3,224,224}};
+
+ EXPECT_TRUE(desc1 == desc2);
+}
+
+TEST(GAPI_MetaDesc, Compare_Not_Equal_MatDesc_ND_1)
+{
+ const auto desc1 = cv::GMatDesc{CV_8U, {1,1000}};
+ const auto desc2 = cv::GMatDesc{CV_32F, {1,1000}};
+
+ EXPECT_TRUE(desc1 != desc2);
+}
+
+TEST(GAPI_MetaDesc, Compare_Not_Equal_MatDesc_ND_2)
+{
+ const auto desc1 = cv::GMatDesc{CV_8U, {1,1000}};
+ const auto desc2 = cv::GMatDesc{CV_8U, {1,1400}};
+
+ EXPECT_TRUE(desc1 != desc2);
+}
+
+TEST(GAPI_MetaDesc, Compare_Not_Equal_MatDesc_ND_3)
+{
+ const auto desc1 = cv::GMatDesc{CV_8U, {1,1000}};
+ const auto desc2 = cv::GMatDesc{CV_8U, 1, {32,32}};
+
+ EXPECT_TRUE(desc1 != desc2);
+}
+
TEST(GAPI_MetaDesc, Compile_MatchMetaNumber_1)
{
cv::GMat in;
--- /dev/null
+// This file is part of OpenCV project.
+// It is subject to the license terms in the LICENSE file found in the top-level directory
+// of this distribution and at http://opencv.org/license.html.
+//
+// Copyright (C) 2019 Intel Corporation
+
+#include "../test_precomp.hpp"
+
+#ifdef HAVE_INF_ENGINE
+
+#include <stdexcept>
+
+////////////////////////////////////////////////////////////////////////////////
+// FIXME: Suppress deprecation warnings for OpenVINO 2019R2+
+// BEGIN {{{
+#if defined(__GNUC__)
+#pragma GCC diagnostic ignored "-Wdeprecated-declarations"
+#endif
+#ifdef _MSC_VER
+#pragma warning(disable: 4996) // was declared deprecated
+#endif
+
+#if defined(__GNUC__)
+#pragma GCC visibility push(default)
+#endif
+
+#include <inference_engine.hpp>
+
+#if defined(__GNUC__)
+#pragma GCC visibility pop
+#endif
+// END }}}
+////////////////////////////////////////////////////////////////////////////////
+
+#include <ade/util/iota_range.hpp>
+
+#include <opencv2/gapi/infer/ie.hpp>
+#include <opencv2/gapi/infer/ie/util.hpp>
+
+namespace opencv_test
+{
+namespace {
+
+// FIXME: taken from DNN module
+static void initDLDTDataPath()
+{
+#ifndef WINRT
+ static bool initialized = false;
+ if (!initialized)
+ {
+ const char* omzDataPath = getenv("OPENCV_OPEN_MODEL_ZOO_DATA_PATH");
+ if (omzDataPath)
+ cvtest::addDataSearchPath(omzDataPath);
+ const char* dnnDataPath = getenv("OPENCV_DNN_TEST_DATA_PATH");
+ if (dnnDataPath) {
+ // Add the dnnDataPath itself - G-API is using some images there directly
+ cvtest::addDataSearchPath(dnnDataPath);
+ cvtest::addDataSearchPath(dnnDataPath + std::string("/omz_intel_models"));
+ }
+ initialized = true;
+ }
+#endif // WINRT
+}
+
+// FIXME: taken from the DNN module
+void normAssert(cv::InputArray ref, cv::InputArray test,
+ const char *comment /*= ""*/,
+ double l1 = 0.00001, double lInf = 0.0001)
+{
+ double normL1 = cvtest::norm(ref, test, cv::NORM_L1) / ref.getMat().total();
+ EXPECT_LE(normL1, l1) << comment;
+
+ double normInf = cvtest::norm(ref, test, cv::NORM_INF);
+ EXPECT_LE(normInf, lInf) << comment;
+}
+
+} // anonymous namespace
+
+// TODO: Probably DNN/IE part can be further parametrized with a template
+// NOTE: here ".." is used to leave the default "gapi/" search scope
+TEST(TestAgeGenderIE, InferBasicTensor)
+{
+ initDLDTDataPath();
+
+ const std::string path = "Retail/object_attributes/age_gender/dldt/age-gender-recognition-retail-0013";
+ const auto topology_path = findDataFile(path + ".xml", false);
+ const auto weights_path = findDataFile(path + ".bin", false);
+
+ // Load IE network, initialize input data using that.
+ namespace IE = InferenceEngine;
+ cv::Mat in_mat;
+ cv::Mat gapi_age, gapi_gender;
+
+ IE::Blob::Ptr ie_age, ie_gender;
+ {
+ IE::CNNNetReader reader;
+ reader.ReadNetwork(topology_path);
+ reader.ReadWeights(weights_path);
+ auto net = reader.getNetwork();
+
+ const auto &iedims = net.getInputsInfo().begin()->second->getDims();
+ auto cvdims = cv::gapi::ie::util::to_ocv(iedims);
+ std::reverse(cvdims.begin(), cvdims.end());
+ in_mat.create(cvdims, CV_32F);
+ cv::randu(in_mat, -1, 1);
+
+ auto plugin = IE::PluginDispatcher().getPluginByDevice("CPU");
+ auto plugin_net = plugin.LoadNetwork(net, {});
+ auto infer_request = plugin_net.CreateInferRequest();
+
+ infer_request.SetBlob("data", cv::gapi::ie::util::to_ie(in_mat));
+ infer_request.Infer();
+ ie_age = infer_request.GetBlob("age_conv3");
+ ie_gender = infer_request.GetBlob("prob");
+ }
+
+ // Configure & run G-API
+ using AGInfo = std::tuple<cv::GMat, cv::GMat>;
+ G_API_NET(AgeGender, <AGInfo(cv::GMat)>, "test-age-gender");
+
+ cv::GMat in;
+ cv::GMat age, gender;
+ std::tie(age, gender) = cv::gapi::infer<AgeGender>(in);
+ cv::GComputation comp(cv::GIn(in), cv::GOut(age, gender));
+
+ auto pp = cv::gapi::ie::Params<AgeGender> {
+ topology_path, weights_path, "CPU"
+ }.cfgOutputLayers({ "age_conv3", "prob" });
+ comp.apply(cv::gin(in_mat), cv::gout(gapi_age, gapi_gender),
+ cv::compile_args(cv::gapi::networks(pp)));
+
+ // Validate with IE itself (avoid DNN module dependency here)
+ normAssert(cv::gapi::ie::util::to_ocv(ie_age), gapi_age, "Test age output" );
+ normAssert(cv::gapi::ie::util::to_ocv(ie_gender), gapi_gender, "Test gender output");
+}
+
+TEST(TestAgeGenderIE, InferBasicImage)
+{
+ initDLDTDataPath();
+
+ const std::string path = "Retail/object_attributes/age_gender/dldt/age-gender-recognition-retail-0013";
+ const auto topology_path = findDataFile(path + ".xml", false);
+ const auto weights_path = findDataFile(path + ".bin", false);
+
+ // FIXME: Ideally it should be an image from disk
+ // cv::Mat in_mat = cv::imread(findDataFile("grace_hopper_227.png"));
+ cv::Mat in_mat(cv::Size(320, 240), CV_8UC3);
+ cv::randu(in_mat, 0, 255);
+
+ cv::Mat gapi_age, gapi_gender;
+
+ // Load & run IE network
+ namespace IE = InferenceEngine;
+ IE::Blob::Ptr ie_age, ie_gender;
+ {
+ IE::CNNNetReader reader;
+ reader.ReadNetwork(topology_path);
+ reader.ReadWeights(weights_path);
+ auto net = reader.getNetwork();
+ auto &ii = net.getInputsInfo().at("data");
+ ii->setPrecision(IE::Precision::U8);
+ ii->setLayout(IE::Layout::NHWC);
+ ii->getPreProcess().setResizeAlgorithm(IE::RESIZE_BILINEAR);
+
+ auto plugin = IE::PluginDispatcher().getPluginByDevice("CPU");
+ auto plugin_net = plugin.LoadNetwork(net, {});
+ auto infer_request = plugin_net.CreateInferRequest();
+
+ infer_request.SetBlob("data", cv::gapi::ie::util::to_ie(in_mat));
+ infer_request.Infer();
+ ie_age = infer_request.GetBlob("age_conv3");
+ ie_gender = infer_request.GetBlob("prob");
+ }
+
+ // Configure & run G-API
+ using AGInfo = std::tuple<cv::GMat, cv::GMat>;
+ G_API_NET(AgeGender, <AGInfo(cv::GMat)>, "test-age-gender");
+
+ cv::GMat in;
+ cv::GMat age, gender;
+ std::tie(age, gender) = cv::gapi::infer<AgeGender>(in);
+ cv::GComputation comp(cv::GIn(in), cv::GOut(age, gender));
+
+ auto pp = cv::gapi::ie::Params<AgeGender> {
+ topology_path, weights_path, "CPU"
+ }.cfgOutputLayers({ "age_conv3", "prob" });
+ comp.apply(cv::gin(in_mat), cv::gout(gapi_age, gapi_gender),
+ cv::compile_args(cv::gapi::networks(pp)));
+
+ // Validate with IE itself (avoid DNN module dependency here)
+ normAssert(cv::gapi::ie::util::to_ocv(ie_age), gapi_age, "Test age output" );
+ normAssert(cv::gapi::ie::util::to_ocv(ie_gender), gapi_gender, "Test gender output");
+}
+
+TEST(TestAgeGenderIE, InferROIList)
+{
+ initDLDTDataPath();
+
+ const std::string path = "Retail/object_attributes/age_gender/dldt/age-gender-recognition-retail-0013";
+ const auto topology_path = findDataFile(path + ".xml", false);
+ const auto weights_path = findDataFile(path + ".bin", false);
+
+ // FIXME: Ideally it should be an image from disk
+ // cv::Mat in_mat = cv::imread(findDataFile("grace_hopper_227.png"));
+ cv::Mat in_mat(cv::Size(640, 480), CV_8UC3);
+ cv::randu(in_mat, 0, 255);
+
+ std::vector<cv::Rect> rois = {
+ cv::Rect(cv::Point{ 0, 0}, cv::Size{80, 120}),
+ cv::Rect(cv::Point{50, 100}, cv::Size{96, 160}),
+ };
+
+ std::vector<cv::Mat> gapi_age, gapi_gender;
+
+ // Load & run IE network
+ namespace IE = InferenceEngine;
+ std::vector<cv::Mat> ie_age, ie_gender;
+ {
+ IE::CNNNetReader reader;
+ reader.ReadNetwork(topology_path);
+ reader.ReadWeights(weights_path);
+ auto net = reader.getNetwork();
+ auto &ii = net.getInputsInfo().at("data");
+ ii->setPrecision(IE::Precision::U8);
+ ii->setLayout(IE::Layout::NHWC);
+ ii->getPreProcess().setResizeAlgorithm(IE::RESIZE_BILINEAR);
+
+ auto plugin = IE::PluginDispatcher().getPluginByDevice("CPU");
+ auto plugin_net = plugin.LoadNetwork(net, {});
+ auto infer_request = plugin_net.CreateInferRequest();
+ auto frame_blob = cv::gapi::ie::util::to_ie(in_mat);
+
+ for (auto &&rc : rois) {
+ const auto ie_rc = IE::ROI {
+ 0u
+ , static_cast<std::size_t>(rc.x)
+ , static_cast<std::size_t>(rc.y)
+ , static_cast<std::size_t>(rc.width)
+ , static_cast<std::size_t>(rc.height)
+ };
+ infer_request.SetBlob("data", IE::make_shared_blob(frame_blob, ie_rc));
+ infer_request.Infer();
+
+ using namespace cv::gapi::ie::util;
+ ie_age.push_back(to_ocv(infer_request.GetBlob("age_conv3")).clone());
+ ie_gender.push_back(to_ocv(infer_request.GetBlob("prob")).clone());
+ }
+ }
+
+ // Configure & run G-API
+ using AGInfo = std::tuple<cv::GMat, cv::GMat>;
+ G_API_NET(AgeGender, <AGInfo(cv::GMat)>, "test-age-gender");
+
+ cv::GArray<cv::Rect> rr;
+ cv::GMat in;
+ cv::GArray<cv::GMat> age, gender;
+ std::tie(age, gender) = cv::gapi::infer<AgeGender>(rr, in);
+ cv::GComputation comp(cv::GIn(in, rr), cv::GOut(age, gender));
+
+ auto pp = cv::gapi::ie::Params<AgeGender> {
+ topology_path, weights_path, "CPU"
+ }.cfgOutputLayers({ "age_conv3", "prob" });
+ comp.apply(cv::gin(in_mat, rois), cv::gout(gapi_age, gapi_gender),
+ cv::compile_args(cv::gapi::networks(pp)));
+
+ // Validate with IE itself (avoid DNN module dependency here)
+ ASSERT_EQ(2u, ie_age.size() );
+ ASSERT_EQ(2u, ie_gender.size());
+ ASSERT_EQ(2u, gapi_age.size() );
+ ASSERT_EQ(2u, gapi_gender.size());
+
+ normAssert(ie_age [0], gapi_age [0], "0: Test age output");
+ normAssert(ie_gender[0], gapi_gender[0], "0: Test gender output");
+ normAssert(ie_age [1], gapi_age [1], "1: Test age output");
+ normAssert(ie_gender[1], gapi_gender[1], "1: Test gender output");
+}
+
+
+} // namespace opencv_test
+
+#endif // HAVE_INF_ENGINE
{
cv::GMat unaryOp(cv::GMat m)
{
- return cv::GCall(cv::GKernel{"gapi.test.unaryop", nullptr, { GShape::GMAT } }).pass(m).yield(0);
+ return cv::GCall(cv::GKernel{"gapi.test.unaryop", "", nullptr, { GShape::GMAT } }).pass(m).yield(0);
}
cv::GMat binaryOp(cv::GMat m1, cv::GMat m2)
{
- return cv::GCall(cv::GKernel{"gapi.test.binaryOp", nullptr, { GShape::GMAT } }).pass(m1, m2).yield(0);
+ return cv::GCall(cv::GKernel{"gapi.test.binaryOp", "", nullptr, { GShape::GMAT } }).pass(m1, m2).yield(0);
}
std::vector<ade::NodeHandle> collectOperations(const cv::gimpl::GModel::Graph& gr)
namespace opencv_test
{
using Mat = cv::gapi::own::Mat;
+using Dims = std::vector<int>;
TEST(OwnMat, DefaultConstruction)
{
ASSERT_EQ(m.cols, 0);
ASSERT_EQ(m.type(), 0);
ASSERT_EQ(m.depth(), 0);
+ ASSERT_TRUE(m.dims.empty());
}
TEST(OwnMat, Create)
ASSERT_EQ(m.channels(), 1);
ASSERT_EQ(m.elemSize(), sizeof(uint8_t));
ASSERT_EQ(m.step, sizeof(uint8_t) * m.cols);
+ ASSERT_TRUE(m.dims.empty());
+}
+
+TEST(OwnMat, CreateND)
+{
+ Dims dims = {1,1,32,32};
+ Mat m;
+ m.create(dims, CV_32F);
+
+ ASSERT_NE(nullptr , m.data );
+ ASSERT_EQ((cv::gapi::own::Size{0,0}), (cv::gapi::own::Size{m.cols, m.rows}));
+
+ ASSERT_EQ(static_cast<size_t>(dims[0]*dims[1]*dims[2]*dims[3]), m.total());
+ ASSERT_EQ(CV_32F , m.type() );
+ ASSERT_EQ(CV_32F , m.depth() );
+ ASSERT_EQ(-1 , m.channels());
+ ASSERT_EQ(sizeof(float) , m.elemSize());
+ ASSERT_EQ(0u , m.step );
+ ASSERT_EQ(dims , m.dims );
}
TEST(OwnMat, CreateOverload)
ASSERT_EQ(m.channels(), 1);
ASSERT_EQ(m.elemSize(), sizeof(uint8_t));
ASSERT_EQ(m.step, sizeof(uint8_t) * m.cols);
+ ASSERT_TRUE(m.dims.empty());
}
+
TEST(OwnMat, Create3chan)
{
auto size = cv::Size{32,16};
ASSERT_EQ(m.channels(), 3);
ASSERT_EQ(m.elemSize(), 3 * sizeof(uint8_t));
ASSERT_EQ(m.step, 3* sizeof(uint8_t) * m.cols);
+ ASSERT_TRUE(m.dims.empty());
}
struct NonEmptyMat {
cv::Size{mat.cols, mat.rows},
mat.type(),
mat.depth(),
- mat.channels()
+ mat.channels(),
+ mat.dims
);
};
<< (cvMat != cvMatFromOwn);
}
+TEST(OwnMatConversion, WithND)
+{
+ const Dims dims = {1,3,8,8};
+ std::vector<uint8_t> data(dims[0]*dims[1]*dims[2]*dims[3]);
+ for (size_t i = 0u; i < data.size(); i++)
+ {
+ data[i] = static_cast<uint8_t>(i);
+ }
+ cv::Mat cvMat(dims, CV_32S, data.data());
+ auto ownMat = to_own(cvMat);
+ auto cvMatFromOwn = cv::gapi::own::to_ocv(ownMat);
+
+ EXPECT_EQ(0, cv::countNonZero(cvMat != cvMatFromOwn))
+ << cvMat << std::endl
+ << (cvMat != cvMatFromOwn);
+}
+
TEST(OwnMat, PtrWithStep)
{
constexpr int width = 8;
<< (to_ocv(mat) != to_ocv(dst));
}
+TEST(OwnMat, AssignNDtoRegular)
+{
+ const auto sz = cv::gapi::own::Size{32,32};
+ const auto dims = Dims{1,3,224,224};
+
+ Mat a;
+ a.create(sz, CV_8U);
+ const auto *old_ptr = a.data;
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_EQ(sz , (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(sz.width*sz.height), a.total());
+ ASSERT_EQ(CV_8U , a.type());
+ ASSERT_EQ(CV_8U , a.depth());
+ ASSERT_EQ(1 , a.channels());
+ ASSERT_EQ(sizeof(uint8_t), a.elemSize());
+ ASSERT_EQ(static_cast<size_t>(sz.width), a.step);
+ ASSERT_TRUE(a.dims.empty());
+
+ Mat b;
+ b.create(dims, CV_32F);
+ a = b;
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_NE(old_ptr , a.data);
+ ASSERT_EQ((cv::gapi::own::Size{0,0}), (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(dims[0]*dims[1]*dims[2]*dims[3]), a.total());
+ ASSERT_EQ(CV_32F , a.type());
+ ASSERT_EQ(CV_32F , a.depth());
+ ASSERT_EQ(-1 , a.channels());
+ ASSERT_EQ(sizeof(float), a.elemSize());
+ ASSERT_EQ(0u , a.step);
+ ASSERT_EQ(dims , a.dims);
+}
+
+TEST(OwnMat, AssignRegularToND)
+{
+ const auto sz = cv::gapi::own::Size{32,32};
+ const auto dims = Dims{1,3,224,224};
+
+ Mat a;
+ a.create(dims, CV_32F);
+ const auto *old_ptr = a.data;
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_EQ((cv::gapi::own::Size{0,0}), (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(dims[0]*dims[1]*dims[2]*dims[3]), a.total());
+ ASSERT_EQ(CV_32F , a.type());
+ ASSERT_EQ(CV_32F , a.depth());
+ ASSERT_EQ(-1 , a.channels());
+ ASSERT_EQ(sizeof(float), a.elemSize());
+ ASSERT_EQ(0u , a.step);
+ ASSERT_EQ(dims , a.dims);
+
+ Mat b;
+ b.create(sz, CV_8U);
+ a = b;
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_NE(old_ptr , a.data);
+ ASSERT_EQ(sz , (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(sz.width*sz.height), a.total());
+ ASSERT_EQ(CV_8U , a.type());
+ ASSERT_EQ(CV_8U , a.depth());
+ ASSERT_EQ(1 , a.channels());
+ ASSERT_EQ(sizeof(uint8_t), a.elemSize());
+ ASSERT_EQ(static_cast<size_t>(sz.width), a.step);
+ ASSERT_TRUE(a.dims.empty());
+}
+
+TEST(OwnMat, CopyNDtoRegular)
+{
+ const auto sz = cv::gapi::own::Size{32,32};
+ const auto dims = Dims{1,3,224,224};
+
+ Mat a;
+ a.create(sz, CV_8U);
+ const auto *old_ptr = a.data;
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_EQ(sz , (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(sz.width*sz.height), a.total());
+ ASSERT_EQ(CV_8U , a.type());
+ ASSERT_EQ(CV_8U , a.depth());
+ ASSERT_EQ(1 , a.channels());
+ ASSERT_EQ(sizeof(uint8_t), a.elemSize());
+ ASSERT_EQ(static_cast<size_t>(sz.width), a.step);
+ ASSERT_TRUE(a.dims.empty());
+
+ Mat b;
+ b.create(dims, CV_32F);
+ b.copyTo(a);
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_NE(old_ptr , a.data);
+ ASSERT_NE(b.data , a.data);
+ ASSERT_EQ((cv::gapi::own::Size{0,0}), (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(dims[0]*dims[1]*dims[2]*dims[3]), a.total());
+ ASSERT_EQ(CV_32F , a.type());
+ ASSERT_EQ(CV_32F , a.depth());
+ ASSERT_EQ(-1 , a.channels());
+ ASSERT_EQ(sizeof(float), a.elemSize());
+ ASSERT_EQ(0u , a.step);
+ ASSERT_EQ(dims , a.dims);
+}
+
+TEST(OwnMat, CopyRegularToND)
+{
+ const auto sz = cv::gapi::own::Size{32,32};
+ const auto dims = Dims{1,3,224,224};
+
+ Mat a;
+ a.create(dims, CV_32F);
+ const auto *old_ptr = a.data;
+
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_EQ((cv::gapi::own::Size{0,0}), (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(dims[0]*dims[1]*dims[2]*dims[3]), a.total());
+ ASSERT_EQ(CV_32F , a.type());
+ ASSERT_EQ(CV_32F , a.depth());
+ ASSERT_EQ(-1 , a.channels());
+ ASSERT_EQ(sizeof(float), a.elemSize());
+ ASSERT_EQ(0u , a.step);
+ ASSERT_EQ(dims , a.dims);
+
+ Mat b;
+ b.create(sz, CV_8U);
+ b.copyTo(a);
+
+ ASSERT_NE(nullptr , a.data);
+ ASSERT_NE(old_ptr , a.data);
+ ASSERT_NE(b.data , a.data);
+ ASSERT_EQ(sz , (cv::gapi::own::Size{a.cols, a.rows}));
+ ASSERT_EQ(static_cast<size_t>(sz.width*sz.height), a.total());
+ ASSERT_EQ(CV_8U , a.type());
+ ASSERT_EQ(CV_8U , a.depth());
+ ASSERT_EQ(1 , a.channels());
+ ASSERT_EQ(sizeof(uint8_t), a.elemSize());
+ ASSERT_EQ(static_cast<size_t>(sz.width), a.step);
+ ASSERT_TRUE(a.dims.empty());
+}
+
TEST(OwnMat, ScalarAssign32SC1)
{
constexpr int width = 8;
<< cmp_result_mat << std::endl;
}
+TEST(OwnMat, ScalarAssignND)
+{
+ std::vector<int> dims = {1,1000};
+ Mat m;
+ m.create(dims, CV_32F);
+ m = cv::gapi::own::Scalar{-1};
+ const float *ptr = reinterpret_cast<float*>(m.data);
+
+ for (auto i = 0u; i < m.total(); i++) {
+ EXPECT_EQ(-1.f, ptr[i]);
+ }
+}
+
TEST(OwnMat, ScalarAssign8UC3)
{
constexpr auto cv_type = CV_8SC3;
#include <vector>
#include <opencv2/ts.hpp>
+
#include <opencv2/gapi.hpp>
#include <opencv2/gapi/imgproc.hpp>
#include <opencv2/gapi/core.hpp>
#include <opencv2/gapi/operators.hpp>
#include <opencv2/gapi/fluid/imgproc.hpp>
#include <opencv2/gapi/fluid/core.hpp>
+#include <opencv2/gapi/infer.hpp>
#endif // __OPENCV_GAPI_TEST_PRECOMP_HPP__