{
}
- Net Net::readFromModelOptimizer(const String& xml, const String& bin)
+ #ifdef HAVE_INF_ENGINE
+ /*static*/
+ Net Net::Impl::createNetworkFromModelOptimizer(InferenceEngine::CNNNetwork& ieNet)
{
- #ifndef HAVE_INF_ENGINE
- CV_Error(Error::StsError, "Build OpenCV with Inference Engine to enable loading models from Model Optimizer.");
- #else
-
- #if INF_ENGINE_VER_MAJOR_LE(INF_ENGINE_RELEASE_2019R3)
- InferenceEngine::CNNNetReader reader;
- reader.ReadNetwork(xml);
- reader.ReadWeights(bin);
-
- InferenceEngine::CNNNetwork ieNet = reader.getNetwork();
- #else
- InferenceEngine::Core& ie = getCore();
- InferenceEngine::CNNNetwork ieNet = ie.ReadNetwork(xml, bin);
- #endif
+ CV_TRACE_FUNCTION();
std::vector<String> inputsNames;
+ std::vector<MatShape> inp_shapes;
for (auto& it : ieNet.getInputsInfo())
{
inputsNames.push_back(it.first);