mSupportedInferenceBackend.insert(std::make_pair(
MV_INFERENCE_BACKEND_ONE, std::make_pair("mlapi", false)));
mSupportedInferenceBackend.insert(std::make_pair(
- MV_INFERENCE_BACKEND_DLDT, std::make_pair("dldt", false)));
+ MV_INFERENCE_BACKEND_DLDT, std::make_pair("dldt", true)));
CheckSupportedInferenceBackend();
if (mConfig.mTargetTypes & MV_INFERENCE_TARGET_DEVICE_GPU) {
if (!(mBackendCapacity.supported_accel_devices &
INFERENCE_TARGET_GPU)) {
- LOGE("Backend doesn't support CPU device as an accelerator.");
+ LOGE("Backend doesn't support GPU device as an accelerator.");
return MEDIA_VISION_ERROR_INVALID_PARAMETER;
}
}
if (mConfig.mTargetTypes & MV_INFERENCE_TARGET_DEVICE_CUSTOM) {
if (!(mBackendCapacity.supported_accel_devices &
INFERENCE_TARGET_CUSTOM)) {
- LOGE("Backend doesn't support CPU device as an accelerator.");
+ LOGE("Backend doesn't support CUSTOM device as an accelerator.");
return MEDIA_VISION_ERROR_INVALID_PARAMETER;
}
}
case INFERENCE_MODEL_CAFFE:
case INFERENCE_MODEL_TF:
case INFERENCE_MODEL_DARKNET:
- case INFERENCE_MODEL_DLDT:
case INFERENCE_MODEL_ONNX:
case INFERENCE_MODEL_VIVANTE:
models.push_back(mConfig.mWeightFilePath);
models.push_back(mConfig.mConfigFilePath);
break;
+ case INFERENCE_MODEL_DLDT:
+ models.push_back(mConfig.mConfigFilePath);
+ models.push_back(mConfig.mWeightFilePath);
+ break;
case INFERENCE_MODEL_TFLITE:
case INFERENCE_MODEL_TORCH:
models.push_back(mConfig.mWeightFilePath);
MV_INFERENCE_MODEL_WEIGHT_FILE_PATH,
PE_DLDT_1_WEIGHT_PATH);
- mv_engine_config_set_int_attribute(mv_engine_cfg,
+ mv_engine_config_set_string_attribute(mv_engine_cfg,
+ MV_INFERENCE_MODEL_CONFIGURATION_FILE_PATH,
+ PE_DLDT_1_XML_PATH);
+
+ mv_engine_config_set_int_attribute(mv_engine_cfg,
MV_INFERENCE_INPUT_DATA_TYPE,
MV_INFERENCE_DATA_FLOAT32);
MV_INFERENCE_MODEL_WEIGHT_FILE_PATH,
PE_DLDT_2_WEIGHT_PATH);
+ mv_engine_config_set_string_attribute(mv_engine_cfg,
+ MV_INFERENCE_MODEL_CONFIGURATION_FILE_PATH,
+ PE_DLDT_2_XML_PATH);
+
+
mv_engine_config_set_int_attribute(mv_engine_cfg,
MV_INFERENCE_INPUT_DATA_TYPE,
MV_INFERENCE_DATA_FLOAT32);
start = clock();
mv_inference_prepare(hp_mv_infer2);
end = clock();
- printf("time: %2.3f\n", (double)(end - start)/CLOCKS_PER_SEC);
+ printf("time 2: %2.3f\n", (double)(end - start)/CLOCKS_PER_SEC);
mv_create_source(&mv_src_p2);
}