std::vector<std::string> models;
+ inference_model_format_e model_format = (inference_model_format_e)key->second;
+
// Push model file information to models vector properly according to detected model format.
- switch (key->second) {
+ switch (model_format) {
case INFERENCE_MODEL_CAFFE:
case INFERENCE_MODEL_TF:
case INFERENCE_MODEL_DARKNET:
}
// Request model loading to backend engine.
- ret = mBackend->Load(models, (inference_model_format_e)key->second);
+ ret = mBackend->Load(models, model_format);
if (ret != INFERENCE_ENGINE_ERROR_NONE) {
delete mBackend;
LOGE("Fail to load model");