From 914ffafdb80497e83874f103ae90972926f6f8e9 Mon Sep 17 00:00:00 2001 From: Jaeyun Jung Date: Thu, 1 Jun 2023 14:53:49 +0900 Subject: [PATCH] [Test] feature for tf-lite Enable testcase if tensorflow-lite ver2.x is available. Signed-off-by: Jaeyun Jung --- tests/capi/unittest_capi_inference_latency.cc | 2 +- tests/capi/unittest_capi_inference_nnfw_runtime.cc | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/tests/capi/unittest_capi_inference_latency.cc b/tests/capi/unittest_capi_inference_latency.cc index 1beaa3b..5dc21c1 100644 --- a/tests/capi/unittest_capi_inference_latency.cc +++ b/tests/capi/unittest_capi_inference_latency.cc @@ -248,7 +248,7 @@ class nnstreamer_capi_singleshot_latency : public ::testing::Test float single_invoke_duration_f, direct_invoke_duration_f; }; -#if defined(ENABLE_TENSORFLOW_LITE) +#if defined(ENABLE_TENSORFLOW_LITE) || defined(ENABLE_TENSORFLOW2_LITE) /** * @brief Measure latency for NNStreamer single shot (tensorflow-lite) * @note Measure the invoke latency added by NNStreamer single shot diff --git a/tests/capi/unittest_capi_inference_nnfw_runtime.cc b/tests/capi/unittest_capi_inference_nnfw_runtime.cc index ada1f30..34cbd67 100644 --- a/tests/capi/unittest_capi_inference_nnfw_runtime.cc +++ b/tests/capi/unittest_capi_inference_nnfw_runtime.cc @@ -707,7 +707,7 @@ TEST_F (MLAPIInferenceNNFW, multimodel_01_p) MLAPIInferenceNNFW::wait_for_sink (&call_cnt2, 1); } -#ifdef ENABLE_TENSORFLOW_LITE +#if defined(ENABLE_TENSORFLOW_LITE) || defined(ENABLE_TENSORFLOW2_LITE) /** * @brief Test nnfw subplugin multi-model (pipeline, ML-API) * @detail Invoke two models which have different framework via Pipeline API, sharing a single input stream -- 2.7.4