int convertOutputDataTypeToFloat();
int preprocess(std::vector<mv_source_h> &mv_sources, std::vector<cv::Mat> &cv_sources);
inference_tensor_data_type_e convertToIE(int given_type);
- int prepareTenosrBuffers(void);
+ int prepareTensorBuffers(void);
void cleanupTensorBuffers(void);
int setUserFile(std::string filename);
};
LOGI("LEAVE");
}
-int Inference::prepareTenosrBuffers(void)
+int Inference::prepareTensorBuffers(void)
{
// If there are input and output tensor buffers allocated before then release the buffers.
// They will be allocated again according to a new model file to be loaded.
std::vector<std::string>().swap(models);
// Prepare input and output tensor buffers.
- ret = prepareTenosrBuffers();
+ ret = prepareTensorBuffers();
if (ret != INFERENCE_ENGINE_ERROR_NONE) {
LOGE("Fail to prepare buffer");
mCanRun = false;