#define INFERENCE_ITERATION 10
+// Macros for tflite inference test cases.
+#define PARAM_TYPE_TFLITE_IC_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_IMAGE_CLASSIFICATION, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/image_classification.bin" }, 224,\
+ 224, 3, { "input_2" }, { "dense_3/Softmax" }, \
+ { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" }, \
+ { 3 }
+
+#define PARAM_TYPE_TFLITE_IC_Q_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_IMAGE_CLASSIFICATION, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_UINT8, \
+ { "/opt/usr/images/image_classification_q.bin" }, \
+ 224, 224, 3, { "input" }, \
+ { "MobilenetV1/Predictions/Reshape_1" }, \
+ { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" }, \
+ { 955 }
+
+#define PARAM_TYPE_TFLITE_OD_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_OBJECT_DETECTION, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/object_detection.bin" }, \
+ 300, 300, 3, \
+ { "normalized_input_image_tensor" }, \
+ { "TFLite_Detection_PostProcess", \
+ "TFLite_Detection_PostProcess:1", \
+ "TFLite_Detection_PostProcess:2", \
+ "TFLite_Detection_PostProcess:3" }, \
+ { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" }, \
+ { 451, 474, 714, 969 }
+
+#define PARAM_TYPE_TFLITE_FD_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_FACE_DETECTION, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/face_detection.bin" }, \
+ 300, 300, 3, \
+ { "normalized_input_image_tensor" }, \
+ { "TFLite_Detection_PostProcess", \
+ "TFLite_Detection_PostProcess:1", \
+ "TFLite_Detection_PostProcess:2", \
+ "TFLite_Detection_PostProcess:3" }, \
+ { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" }, \
+ { 727, 225, 960, 555 }
+
+#define PARAM_TYPE_TFLITE_PE_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_POSE_ESTIMATION, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/pose_estimation.bin" }, \
+ 192, 192, 3, { "image" }, \
+ { "Convolutional_Pose_Machine/stage_5_out" }, \
+ { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" }, \
+ { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351, \
+ 351, 382, 382, 382, 76, 146, 170, 193, 216, 146, \
+ 123, 99, 287, 381, 451, 287, 381, 475 }
+
+#define PARAM_TYPE_TFLITE_AICHG_1_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_AIC_HAND_GESTURE_1, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/hand.bin" }, 224, 224, 3, \
+ { "input" }, \
+ { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" }, \
+ { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, \
+ { 0 }
+
+#define PARAM_TYPE_TFLITE_AICHG_2_INFER(backend, device, iter) \
+ backend, device, TEST_MODEL_AIC_HAND_GESTURE_2, \
+ iter, INFERENCE_TENSOR_DATA_TYPE_FLOAT32, \
+ { "/opt/usr/images/hand.bin" }, 56, 56, 21, \
+ { "input" }, \
+ { "mobilenetv2/coord_refine", "mobilenetv2/gesture" }, \
+ { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" }, \
+ { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39, \
+ 78, 36, 82, 42, 82, 44, 83, 45, 35, 37, \
+ 61, 36, 59, 36, 52, 39, 35, 32, 40, 34, \
+ 62, 39, 70, 40, 58, 41, 34, 42, 34, 41, \
+ 38, 38, 12 }
+
+
+// Macros for tflite model based cltuner test cases.
+#define PARAM_TYPE_TFLITE_IC_CLTUNER(active, update, mode, backend, device, iter) \
+ active, update, mode, \
+ PARAM_TYPE_TFLITE_IC_INFER(backend, device, iter)
+
typedef std::tuple<std::string, int, int, int, int, std::vector<std::string>,
int, int, int, std::vector<std::string>,
std::vector<std::string>, std::vector<std::string>,
ParamType_Infer;
-typedef std::tuple<std::string, int, bool, bool, inference_engine_cltuner_mode_e,
+typedef std::tuple<bool, bool, inference_engine_cltuner_mode_e, std::string, int,
int, int, int, std::vector<std::string>,
int, int, int, std::vector<std::string>,
std::vector<std::string>, std::vector<std::string>,
TEST_P(InferenceEngineTfliteCLTunerTest, Inference)
{
- std::string backend_name;
- int target_devices;
bool active;
bool update;
inference_engine_cltuner_mode_e tuning_mode;
+ std::string backend_name;
+ int target_devices;
int test_type;
int iteration;
int tensor_type;
std::vector<std::string> model_paths;
std::vector<int> answers;
- std::tie(backend_name, target_devices, active, update, tuning_mode, test_type,
+ std::tie(active, update, tuning_mode, backend_name, target_devices, test_type,
iteration, tensor_type, image_paths, height, width, ch, input_layers,
output_layers, model_paths, answers) = GetParam();
// mobilenet based image classification test
// ARMNN.
ParamType_Infer(
- "armnn", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_CPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_CPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_CPU, TEST_MODEL_POSE_ESTIMATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("armnn", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("armnn", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("armnn", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
+
// mobilenet based image classification test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_GPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_GPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "armnn", INFERENCE_TARGET_GPU, TEST_MODEL_POSE_ESTIMATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("armnn", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("armnn", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
/*********************************************************************************/
// parameter order : backend name, target device, input image path/s, height, width, channel count, input layer names, output layer names, model path/s, inference result
// mobilenet based image classification test
// TFLITE.
ParamType_Infer(
- "tflite", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_CPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_CPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_CPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("tflite", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("tflite", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("tflite", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// mobilenet based image classification test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_GPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_GPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "tflite", INFERENCE_TARGET_GPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("tflite", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("tflite", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 })
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("tflite", INFERENCE_TARGET_GPU, INFERENCE_ITERATION))
/* TODO */
- ));
+ ));
INSTANTIATE_TEST_CASE_P(
Inhouse, InferenceEngineTfliteTest,
// mobilenet based image classification test
// ONE via MLAPI.
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
/*********************************************************************************/
// parameter order : backend name, target device, input image path/s, height, width, channel count, input layer names, output layer names, model path/s, inference result
// mobilenet based image classification test
// TFLITE via MLAPI.
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
+ PARAM_TYPE_TFLITE_IC_Q_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
+ PARAM_TYPE_TFLITE_OD_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
+ PARAM_TYPE_TFLITE_FD_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "one", INFERENCE_TARGET_CPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_CPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("one", INFERENCE_TARGET_CPU, INFERENCE_ITERATION)),
// mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
-
+ PARAM_TYPE_TFLITE_IC_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// quantized mobilenet based image classification test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_UINT8,
- { "/opt/usr/images/image_classification_q.bin" }, 224,
- 224, 3, { "input" },
- { "MobilenetV1/Predictions/Reshape_1" },
- { "/usr/share/capi-media-vision/models/IC_Q/tflite/ic_tflite_q_model.tflite" },
- { 955 }),
-
+ PARAM_TYPE_TFLITE_IC_Q_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// object detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_OBJECT_DETECTION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/object_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/OD/tflite/od_tflite_model.tflite" },
- { 451, 474, 714, 969 }),
-
+ PARAM_TYPE_TFLITE_OD_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// face detection test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_FACE_DETECTION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/face_detection.bin" }, 300, 300, 3,
- { "normalized_input_image_tensor" },
- { "TFLite_Detection_PostProcess",
- "TFLite_Detection_PostProcess:1",
- "TFLite_Detection_PostProcess:2",
- "TFLite_Detection_PostProcess:3" },
- { "/usr/share/capi-media-vision/models/FD/tflite/fd_tflite_model1.tflite" },
- { 727, 225, 960, 555 }),
-
+ PARAM_TYPE_TFLITE_FD_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// pose estimation test
ParamType_Infer(
- "one", INFERENCE_TARGET_GPU, TEST_MODEL_POSE_ESTIMATION,
- INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/pose_estimation.bin" }, 192, 192, 3,
- { "image" },
- { "Convolutional_Pose_Machine/stage_5_out" },
- { "/usr/share/capi-media-vision/models/PE/tflite/ped_tflite_model.tflite" },
- { 382, 351, 320, 257, 226, 414, 414, 445, 351, 351,
- 351, 382, 382, 382, 76, 146, 170, 193, 216, 146,
- 123, 99, 287, 381, 451, 287, 381, 475 }),
+ PARAM_TYPE_TFLITE_PE_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 1 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_1, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 224, 224, 3, { "input" }, { "mobilenetv2/boundingbox", "mobilenetv2/heatmap" },
- { "/usr/share/capi-media-vision/models/PE_1/tflite/posenet1_lite_224.tflite" }, { 0 }),
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_1_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
// Hand gesture model 2 from AIC
- ParamType_Infer("one", INFERENCE_TARGET_GPU, TEST_MODEL_AIC_HAND_GESTURE_2, INFERENCE_ITERATION, INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/hand.bin" }, 56, 56, 21, { "input" }, { "mobilenetv2/coord_refine", "mobilenetv2/gesture" },
- { "/usr/share/capi-media-vision/models/PE_2/tflite/posenet2_lite_224.tflite" },
- { 55, 39, 51, 40, 50, 42, 61, 43, 71, 39,
- 78, 36, 82, 42, 82, 44, 83, 45, 35, 37,
- 61, 36, 59, 36, 52, 39, 35, 32, 40, 34,
- 62, 39, 70, 40, 58, 41, 34, 42, 34, 41,
- 38, 38, 12 })
+ ParamType_Infer(
+ PARAM_TYPE_TFLITE_AICHG_2_INFER("one", INFERENCE_TARGET_GPU, INFERENCE_ITERATION))
/* TODO */
));
// mobilenet based image classification test
// ARMNN.
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, true, INFERENCE_ENGINE_CLTUNER_RAPID,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, true, INFERENCE_ENGINE_CLTUNER_RAPID, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, false, INFERENCE_ENGINE_CLTUNER_READ,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, false, INFERENCE_ENGINE_CLTUNER_READ, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, true, INFERENCE_ENGINE_CLTUNER_NORMAL,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, true, INFERENCE_ENGINE_CLTUNER_NORMAL, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, false, INFERENCE_ENGINE_CLTUNER_READ,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, false, INFERENCE_ENGINE_CLTUNER_READ, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, true, INFERENCE_ENGINE_CLTUNER_EXHAUSTIVE,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 }),
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, true, INFERENCE_ENGINE_CLTUNER_EXHAUSTIVE, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION)),
ParamType_CLTuner(
- "armnn", INFERENCE_TARGET_GPU, true, false, INFERENCE_ENGINE_CLTUNER_READ,
- TEST_MODEL_IMAGE_CLASSIFICATION, INFERENCE_ITERATION,
- INFERENCE_TENSOR_DATA_TYPE_FLOAT32,
- { "/opt/usr/images/image_classification.bin" }, 224,
- 224, 3, { "input_2" }, { "dense_3/Softmax" },
- { "/usr/share/capi-media-vision/models/IC/tflite/ic_tflite_model.tflite" },
- { 3 })
+ PARAM_TYPE_TFLITE_IC_CLTUNER(true, false, INFERENCE_ENGINE_CLTUNER_READ, "armnn", INFERENCE_TARGET_GPU, INFERENCE_ITERATION))
/* TODO */
));