From: Jaeyun Date: Fri, 25 Sep 2020 10:46:08 +0000 (+0900) Subject: [Filter/TF-lite] nnapi option X-Git-Tag: accepted/tizen/unified/20201027.124817~33 X-Git-Url: http://review.tizen.org/git/?a=commitdiff_plain;h=f12730778e8abd82b54b22034edf5e80008aa138;p=platform%2Fupstream%2Fnnstreamer.git [Filter/TF-lite] nnapi option Add option structure to open tf-lite model, set NNAPI false as default value. Using NNAPI in tf-lite may make lower performance and unexpected result for some model. Let the developer select NNAPI option when constructing the pipeline. Signed-off-by: Jaeyun --- diff --git a/ext/nnstreamer/tensor_filter/tensor_filter_tensorflow_lite.cc b/ext/nnstreamer/tensor_filter/tensor_filter_tensorflow_lite.cc index 34a45e0..ff30b2c 100644 --- a/ext/nnstreamer/tensor_filter/tensor_filter_tensorflow_lite.cc +++ b/ext/nnstreamer/tensor_filter/tensor_filter_tensorflow_lite.cc @@ -52,6 +52,19 @@ #define DBG FALSE #endif +/** + * @brief Option to open tf-lite model. + */ +typedef struct +{ + const gchar *model_file; /**< path to tensorflow-lite model file */ + const gchar *accelerators; /**< accelerators set for this subplugin */ + gboolean use_nnapi; /**< flag to use NNAPI */ +} tflite_option_s; + +/** + * @brief Possible accelerators. + */ static const gchar *tflite_accl_support[] = { ACCL_CPU_NEON_STR, ACCL_CPU_SIMD_STR, @@ -132,7 +145,7 @@ private: class TFLiteCore { public: - TFLiteCore (const char *_model_path, const char *accelerators); + TFLiteCore (tflite_option_s * option); int init (); int loadModel (); @@ -154,7 +167,7 @@ private: TFLiteInterpreter interpreter; TFLiteInterpreter interpreter_sub; - void setAccelerator (const char * accelerators); + void setAccelerator (const char * accelerators, bool nnapi); }; extern "C" { /* accessed by android api */ @@ -537,31 +550,30 @@ fail_exit: /** * @brief TFLiteCore creator - * @param _model_path : the logical path to '{model_name}.tflite' file - * @param accelerators : the accelerators property set for this subplugin + * @param option options to initialize tf-lite model * @note the model of _model_path will be loaded simultaneously * @return Nothing */ -TFLiteCore::TFLiteCore (const char * _model_path, const char * accelerators) +TFLiteCore::TFLiteCore (tflite_option_s * option) { - interpreter.setModelPath (_model_path); + interpreter.setModelPath (option->model_file); - setAccelerator (accelerators); - if (accelerators != NULL) { - g_message ("nnapi = %d, accl = %s", use_nnapi, get_accl_hw_str (accelerator)); - } + setAccelerator (option->accelerators, option->use_nnapi); + g_message ("nnapi = %d, accl = %s", use_nnapi, get_accl_hw_str (accelerator)); } /** * @brief Set the accelerator for the tf engine */ -void TFLiteCore::setAccelerator (const char * accelerators) +void TFLiteCore::setAccelerator (const char * accelerators, bool nnapi) { - use_nnapi = TRUE; + use_nnapi = nnapi; accelerator = parse_accl_hw (accelerators, tflite_accl_support, tflite_accl_auto, tflite_accl_default); - if (accelerators == NULL || accelerator == ACCL_NONE) + if (accelerators == NULL || accelerator == ACCL_NONE) { + g_warning ("Try to get nnapi flag from the configuration."); goto use_nnapi_ini; + } return; @@ -828,6 +840,51 @@ TFLiteCore::cacheInOutTensorPtr () } /** + * @brief Internal function to get the option for tf-lite model. + */ +static int +tflite_parseCustomOption (const GstTensorFilterProperties * prop, + tflite_option_s * option) +{ + if (prop->num_models != 1 || prop->model_files[0] == NULL) + return -1; + + option->model_file = prop->model_files[0]; + option->accelerators = prop->accl_str; + option->use_nnapi = FALSE; + + if (prop->custom_properties) { + gchar **strv; + guint i, len; + + strv = g_strsplit (prop->custom_properties, ",", -1); + len = g_strv_length (strv); + + for (i = 0; i < len; ++i) { + gchar **pair = g_strsplit (strv[i], ":", -1); + + if (g_strv_length (pair) > 1) { + g_strstrip (pair[0]); + g_strstrip (pair[1]); + + if (g_ascii_strcasecmp (pair[0], "UseNNAPI") == 0) { + if (g_ascii_strcasecmp (pair[1], "true") == 0) + option->use_nnapi = TRUE; + } else { + g_warning ("Unknown option (%s).", strv[i]); + } + } + + g_strfreev (pair); + } + + g_strfreev (strv); + } + + return 0; +} + +/** * @brief Free privateData and move on. */ static void @@ -855,25 +912,23 @@ tflite_loadModelFile (const GstTensorFilterProperties * prop, void **private_data) { TFLiteCore *core; - const gchar *model_file; + tflite_option_s option = { 0, }; - if (prop->num_models != 1) + if (tflite_parseCustomOption (prop, &option) != 0) { + g_printerr ("Failed to parse options to initialize tensorflow-lite model."); return -1; + } core = static_cast(*private_data); - model_file = prop->model_files[0]; - - if (model_file == NULL) - return -1; if (core != NULL) { - if (core->compareModelPath (model_file)) + if (core->compareModelPath (option.model_file)) return 1; /* skipped */ tflite_close (prop, private_data); } - core = new TFLiteCore (model_file, prop->accl_str); + core = new TFLiteCore (&option); if (core == NULL) { g_printerr ("Failed to allocate memory for filter subplugin."); return -1; diff --git a/tests/nnstreamer_filter_tensorflow2_lite/runTest.sh b/tests/nnstreamer_filter_tensorflow2_lite/runTest.sh index 45a17c5..8ae326d 100644 --- a/tests/nnstreamer_filter_tensorflow2_lite/runTest.sh +++ b/tests/nnstreamer_filter_tensorflow2_lite/runTest.sh @@ -92,7 +92,7 @@ gstTest "--gst-plugin-path=${PATH_TO_PLUGIN} videotestsrc num_buffers=4 ! videoc # Test the backend setting done with tensorflow2-lite # This also performs tests for generic backend configuration parsing function run_pipeline() { - gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter framework=tensorflow2-lite model=${PATH_TO_MODEL} accelerator=$1 ! filesink location=tensorfilter.out.log 2>info + gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter framework=tensorflow2-lite model=${PATH_TO_MODEL} accelerator=$1 custom=UseNNAPI:true ! filesink location=tensorfilter.out.log 2>info } arch=$(uname -m) diff --git a/tests/nnstreamer_filter_tensorflow_lite/runTest.sh b/tests/nnstreamer_filter_tensorflow_lite/runTest.sh index 27081e3..2b100bf 100644 --- a/tests/nnstreamer_filter_tensorflow_lite/runTest.sh +++ b/tests/nnstreamer_filter_tensorflow_lite/runTest.sh @@ -92,7 +92,7 @@ gstTest "--gst-plugin-path=${PATH_TO_PLUGIN} videotestsrc num_buffers=4 ! videoc # Test the backend setting done with tensorflow-lite # This also performs tests for generic backend configuration parsing function run_pipeline() { - gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter framework=tensorflow-lite model=${PATH_TO_MODEL} accelerator=$1 ! filesink location=tensorfilter.out.log 2>info + gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter framework=tensorflow-lite model=${PATH_TO_MODEL} accelerator=$1 custom=UseNNAPI:true ! filesink location=tensorfilter.out.log 2>info } arch=$(uname -m) @@ -190,7 +190,7 @@ cat info | grep "nnapi = 1, accl = ${auto_accl}$" testResult $? 2-17 "NNAPI activation test" 0 1 # Property reading test for nnapi before setting the framework (analogous test is 2-3) -gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter accelerator=true:!npu,gpu framework=tensorflow-lite model=${PATH_TO_MODEL} ! filesink location=tensorfilter.out.log 2>info +gst-launch-1.0 --gst-plugin-path=${PATH_TO_PLUGIN} filesrc location=${PATH_TO_IMAGE} ! pngdec ! videoscale ! imagefreeze ! videoconvert ! video/x-raw,format=RGB,framerate=0/1 ! tensor_converter ! tensor_filter accelerator=true:!npu,gpu framework=tensorflow-lite model=${PATH_TO_MODEL} custom=UseNNAPI:true ! filesink location=tensorfilter.out.log 2>info cat info | grep "nnapi = 1, accl = gpu$" testResult $? 2-18 "NNAPI activation test" 0 1