From 1f638e0f656a1058eb594c2efc2d5c1fc01aafbc Mon Sep 17 00:00:00 2001 From: Tae-Young Chung Date: Tue, 8 Sep 2020 10:51:03 +0900 Subject: [PATCH] Fix description for html link '#' is used for html link of enum item Change-Id: I3472999c294a504fe37c0223b26b4eafc91b0010 Signed-off-by: Tae-Young Chung --- include/mv_inference_type.h | 10 +++++----- 1 file changed, 5 insertions(+), 5 deletions(-) diff --git a/include/mv_inference_type.h b/include/mv_inference_type.h index 97a56ea..a259b77 100644 --- a/include/mv_inference_type.h +++ b/include/mv_inference_type.h @@ -36,21 +36,21 @@ extern "C" { /** * @brief Enumeration for inference backend. - * @MV_INFERENCE_BACKEND_OPENCV An open source computer vision and machine learning + * #MV_INFERENCE_BACKEND_OPENCV An open source computer vision and machine learning * software library. * (https://opencv.org/about/) - * @MV_INFERENCE_BACKEND_TFLITE Google-introduced open source inference engine for embedded systems, + * #MV_INFERENCE_BACKEND_TFLITE Google-introduced open source inference engine for embedded systems, * which runs Tensorflow Lite model. * (https://www.tensorflow.org/lite/guide/get_started) - * @MV_INFERENCE_BACKEND_ARMNN ARM-introduced open source inference engine for CPUs, GPUs and NPUs, which + * #MV_INFERENCE_BACKEND_ARMNN ARM-introduced open source inference engine for CPUs, GPUs and NPUs, which * enables efficient translation of existing neural network frameworks * such as TensorFlow, TensorFlow Lite and Caffes, allowing them to * run efficiently without modification on Embedded hardware. * (https://developer.arm.com/ip-products/processors/machine-learning/arm-nn) - * @MV_INFERENCE_BACKEND_MLAPI Samsung-introduced open source ML single API framework of NNStreamer, which + * #MV_INFERENCE_BACKEND_MLAPI Samsung-introduced open source ML single API framework of NNStreamer, which * runs various NN models via tensor filters of NNStreamer. * (https://github.com/nnstreamer/nnstreamer) - * @MV_INFERENCE_BACKEND_ONE Samsung-introduced open source inference engine called On-device Neural Engine, which + * #MV_INFERENCE_BACKEND_ONE Samsung-introduced open source inference engine called On-device Neural Engine, which * performs inference of a given NN model on various devices such as CPU, GPU, DSP and NPU. * (https://github.com/Samsung/ONE) * -- 2.7.4