Add snpe lite2_fp32 model test 68/279668/1
authorKwanghoon Son <k.son@samsung.com>
Tue, 16 Aug 2022 01:45:51 +0000 (21:45 -0400)
committerKwanghoon Son <k.son@samsung.com>
Tue, 16 Aug 2022 01:46:12 +0000 (21:46 -0400)
Change-Id: Id9f25f655cc85de348195fe83f54277a3d2abf9b
Signed-off-by: Kwanghoon Son <k.son@samsung.com>
test/src/inference_engine_tc.cpp
test/src/inference_engine_test_common.cpp

index cb6dfc405e3f939db62e2b0df1c02e2229fc8f34..c06543343f489a20d526d02300c50bcb7a3398f8 100644 (file)
@@ -673,7 +673,10 @@ INSTANTIATE_TEST_CASE_P(
                                // ML Single API for NNStreamer.
                                ParamType_Three(
                                                "one", INFERENCE_TARGET_CPU,
-                                               { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" })
+                                               { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" }),
+                               ParamType_Three(
+                                               "snpe", INFERENCE_TARGET_CUSTOM,
+                                               { "/usr/share/capi-media-vision/models/OD/snpe/lite2_fp32.dlc" })
                                /* TODO */
                                ));
 
index 67036d228da17e8e0113085e5a5016a22938623b..206b983910bf7f820b0be0aadc1bb32d6640e4e6 100644 (file)
@@ -38,7 +38,7 @@ static std::map<std::string, int> Model_Formats = {
        { "caffemodel", INFERENCE_MODEL_CAFFE }, { "pb", INFERENCE_MODEL_TF },
        { "tflite", INFERENCE_MODEL_TFLITE },   { "t7", INFERENCE_MODEL_TORCH },
        { "weights", INFERENCE_MODEL_DARKNET },  { "xml", INFERENCE_MODEL_DLDT },
-       { "onnx", INFERENCE_MODEL_ONNX }
+       { "onnx", INFERENCE_MODEL_ONNX }, { "dlc", INFERENCE_MODEL_SNPE }
 };
 
 static MachineCapacity MachineCap[] = {
@@ -126,6 +126,7 @@ int GetModelInfo(std::vector<std::string> &model_paths,
                break;
        case INFERENCE_MODEL_TFLITE:
        case INFERENCE_MODEL_TORCH:
+       case INFERENCE_MODEL_SNPE:
                models.push_back(model_paths[0]);
                break;
        default: