Change-Id: Iaecaebe2eb1219a4903c06ff06dde3eb7e69e683
Signed-off-by: Inki Dae <inki.dae@samsung.com>
return INFERENCE_ENGINE_ERROR_NONE;
}
-int InferenceARMNN::GetInferenceResult(tensor_t& results)
-{
- LOGI("ENTER");
-
- LOGI("LEAVE");
-
- return INFERENCE_ENGINE_ERROR_NONE;
-}
-
extern "C"
{
class IInferenceEngineCommon* EngineCommonInit(void)
int Run(std::vector<inference_engine_tensor_buffer> &input_buffers,
std::vector<inference_engine_tensor_buffer> &output_buffers) override;
- int GetInferenceResult(tensor_t& results) override;
-
private:
int CreateTfLiteNetwork(std::string model_path);
int CreateNetwork(std::vector<std::string> model_paths, inference_model_format_e model_format);