From 94d187e269170a212bcfd9b7fe14da920a0597fd Mon Sep 17 00:00:00 2001 From: Daniel Mallia Date: Tue, 19 May 2020 18:59:28 -0400 Subject: [PATCH] Add next and previous navigation links to all tutorials --- doc/tutorials/calib3d/camera_calibration/camera_calibration.markdown | 4 ++++ .../camera_calibration_pattern/camera_calibration_pattern.markdown | 3 +++ .../camera_calibration_square_chess.markdown | 4 ++++ .../calib3d/interactive_calibration/interactive_calibration.markdown | 3 +++ doc/tutorials/calib3d/real_time_pose/real_time_pose.markdown | 4 ++++ doc/tutorials/dnn/dnn_android/dnn_android.markdown | 3 +++ doc/tutorials/dnn/dnn_custom_layers/dnn_custom_layers.md | 2 ++ doc/tutorials/dnn/dnn_googlenet/dnn_googlenet.markdown | 2 ++ doc/tutorials/dnn/dnn_halide/dnn_halide.markdown | 3 +++ .../dnn/dnn_halide_scheduling/dnn_halide_scheduling.markdown | 3 +++ doc/tutorials/dnn/dnn_javascript/dnn_javascript.markdown | 3 +++ doc/tutorials/dnn/dnn_yolo/dnn_yolo.markdown | 3 +++ doc/tutorials/features2d/akaze_matching/akaze_matching.markdown | 3 +++ doc/tutorials/features2d/akaze_tracking/akaze_tracking.markdown | 3 +++ .../detection_of_planar_objects/detection_of_planar_objects.markdown | 4 ++++ .../features2d/feature_description/feature_description.markdown | 3 +++ doc/tutorials/features2d/feature_detection/feature_detection.markdown | 3 +++ .../features2d/feature_flann_matcher/feature_flann_matcher.markdown | 3 +++ .../features2d/feature_homography/feature_homography.markdown | 3 +++ doc/tutorials/features2d/homography/homography.markdown | 2 ++ .../trackingmotion/corner_subpixels/corner_subpixels.markdown | 3 +++ .../generic_corner_detector/generic_corner_detector.markdown | 4 ++++ .../good_features_to_track/good_features_to_track.markdown | 3 +++ .../trackingmotion/harris_detector/harris_detector.markdown | 2 ++ .../gpu/gpu-basics-similarity/gpu_basics_similarity.markdown | 2 ++ doc/tutorials/gpu/gpu-thrust-interop/gpu_thrust_interop.markdown | 2 ++ doc/tutorials/introduction/android_binary_package/O4A_SDK.markdown | 4 ++++ .../introduction/android_binary_package/android_dev_intro.markdown | 4 ++++ .../introduction/android_binary_package/android_ocl_intro.markdown | 4 ++++ .../android_binary_package/dev_with_OCV_on_Android.markdown | 4 ++++ .../introduction/building_tegra_cuda/building_tegra_cuda.markdown | 4 ++++ .../introduction/clojure_dev_intro/clojure_dev_intro.markdown | 4 ++++ .../cross_referencing/tutorial_cross_referencing.markdown | 3 +++ .../crosscompilation/arm_crosscompile_with_cmake.markdown | 4 ++++ doc/tutorials/introduction/desktop_java/java_dev_intro.markdown | 4 ++++ doc/tutorials/introduction/display_image/display_image.markdown | 3 +++ .../introduction/documenting_opencv/documentation_tutorial.markdown | 4 ++++ doc/tutorials/introduction/ios_install/ios_install.markdown | 3 +++ doc/tutorials/introduction/java_eclipse/java_eclipse.markdown | 4 ++++ doc/tutorials/introduction/linux_eclipse/linux_eclipse.markdown | 3 +++ doc/tutorials/introduction/linux_gcc_cmake/linux_gcc_cmake.markdown | 4 ++++ doc/tutorials/introduction/linux_install/linux_install.markdown | 3 +++ doc/tutorials/introduction/macos_install/macos_install.markdown | 4 ++++ doc/tutorials/introduction/transition_guide/transition_guide.markdown | 4 ++++ doc/tutorials/introduction/windows_install/windows_install.markdown | 4 ++++ .../windows_visual_studio_image_watch.markdown | 4 ++++ .../windows_visual_studio_opencv.markdown | 4 ++++ doc/tutorials/ios/hello/hello.markdown | 2 ++ doc/tutorials/ios/image_manipulation/image_manipulation.markdown | 3 +++ doc/tutorials/ios/video_processing/video_processing.markdown | 3 +++ doc/tutorials/ml/introduction_to_pca/introduction_to_pca.markdown | 2 ++ doc/tutorials/ml/introduction_to_svm/introduction_to_svm.markdown | 2 ++ doc/tutorials/ml/non_linear_svms/non_linear_svms.markdown | 3 +++ .../objdetect/cascade_classifier/cascade_classifier.markdown | 2 ++ doc/tutorials/objdetect/traincascade.markdown | 2 ++ .../video/background_subtraction/background_subtraction.markdown | 2 ++ doc/tutorials/video/meanshift/meanshift.markdown | 3 +++ doc/tutorials/video/optical_flow/optical_flow.markdown | 2 ++ doc/tutorials/videoio/intelperc.markdown | 3 +++ doc/tutorials/videoio/kinect_openni.markdown | 4 ++++ .../videoio/video-input-psnr-ssim/video_input_psnr_ssim.markdown | 2 ++ doc/tutorials/videoio/video-write/video_write.markdown | 3 +++ doc/tutorials/viz/creating_widgets/creating_widgets.markdown | 3 +++ doc/tutorials/viz/histo3D/histo3D.markdown | 2 ++ doc/tutorials/viz/launching_viz/launching_viz.markdown | 2 ++ doc/tutorials/viz/transformations/transformations.markdown | 3 +++ doc/tutorials/viz/widget_pose/widget_pose.markdown | 3 +++ 67 files changed, 207 insertions(+) diff --git a/doc/tutorials/calib3d/camera_calibration/camera_calibration.markdown b/doc/tutorials/calib3d/camera_calibration/camera_calibration.markdown index 6ce791c..1c13f9b 100644 --- a/doc/tutorials/calib3d/camera_calibration/camera_calibration.markdown +++ b/doc/tutorials/calib3d/camera_calibration/camera_calibration.markdown @@ -1,6 +1,10 @@ Camera calibration With OpenCV {#tutorial_camera_calibration} ============================== +@prev_tutorial{tutorial_camera_calibration_square_chess} +@next_tutorial{tutorial_real_time_pose} + + Cameras have been around for a long-long time. However, with the introduction of the cheap *pinhole* cameras in the late 20th century, they became a common occurrence in our everyday life. Unfortunately, this cheapness comes with its price: significant distortion. Luckily, these are diff --git a/doc/tutorials/calib3d/camera_calibration_pattern/camera_calibration_pattern.markdown b/doc/tutorials/calib3d/camera_calibration_pattern/camera_calibration_pattern.markdown index fba4f5a..754e75b 100644 --- a/doc/tutorials/calib3d/camera_calibration_pattern/camera_calibration_pattern.markdown +++ b/doc/tutorials/calib3d/camera_calibration_pattern/camera_calibration_pattern.markdown @@ -1,6 +1,9 @@ Create calibration pattern {#tutorial_camera_calibration_pattern} ========================================= +@next_tutorial{tutorial_camera_calibration_square_chess} + + The goal of this tutorial is to learn how to create calibration pattern. You can find a chessboard pattern in https://github.com/opencv/opencv/blob/3.4/doc/pattern.png diff --git a/doc/tutorials/calib3d/camera_calibration_square_chess/camera_calibration_square_chess.markdown b/doc/tutorials/calib3d/camera_calibration_square_chess/camera_calibration_square_chess.markdown index bf1eadc..51b0a5e 100644 --- a/doc/tutorials/calib3d/camera_calibration_square_chess/camera_calibration_square_chess.markdown +++ b/doc/tutorials/calib3d/camera_calibration_square_chess/camera_calibration_square_chess.markdown @@ -1,6 +1,10 @@ Camera calibration with square chessboard {#tutorial_camera_calibration_square_chess} ========================================= +@prev_tutorial{tutorial_camera_calibration_pattern} +@next_tutorial{tutorial_camera_calibration} + + The goal of this tutorial is to learn how to calibrate a camera given a set of chessboard images. *Test data*: use images in your data/chess folder. diff --git a/doc/tutorials/calib3d/interactive_calibration/interactive_calibration.markdown b/doc/tutorials/calib3d/interactive_calibration/interactive_calibration.markdown index 8280262..36e19e0 100644 --- a/doc/tutorials/calib3d/interactive_calibration/interactive_calibration.markdown +++ b/doc/tutorials/calib3d/interactive_calibration/interactive_calibration.markdown @@ -1,6 +1,9 @@ Interactive camera calibration application {#tutorial_interactive_calibration} ============================== +@prev_tutorial{tutorial_real_time_pose} + + According to classical calibration technique user must collect all data first and when run @ref cv::calibrateCamera function to obtain camera parameters. If average re-projection error is huge or if estimated parameters seems to be wrong, process of selection or collecting data and starting of @ref cv::calibrateCamera repeats. diff --git a/doc/tutorials/calib3d/real_time_pose/real_time_pose.markdown b/doc/tutorials/calib3d/real_time_pose/real_time_pose.markdown index f61fa7a..98a44b2 100644 --- a/doc/tutorials/calib3d/real_time_pose/real_time_pose.markdown +++ b/doc/tutorials/calib3d/real_time_pose/real_time_pose.markdown @@ -1,6 +1,10 @@ Real Time pose estimation of a textured object {#tutorial_real_time_pose} ============================================== +@prev_tutorial{tutorial_camera_calibration} +@next_tutorial{tutorial_interactive_calibration} + + Nowadays, augmented reality is one of the top research topic in computer vision and robotics fields. The most elemental problem in augmented reality is the estimation of the camera pose respect of an object in the case of computer vision area to do later some 3D rendering or in the case of robotics diff --git a/doc/tutorials/dnn/dnn_android/dnn_android.markdown b/doc/tutorials/dnn/dnn_android/dnn_android.markdown index 1a3f922..cd81672 100644 --- a/doc/tutorials/dnn/dnn_android/dnn_android.markdown +++ b/doc/tutorials/dnn/dnn_android/dnn_android.markdown @@ -1,5 +1,8 @@ # How to run deep networks on Android device {#tutorial_dnn_android} +@prev_tutorial{tutorial_dnn_halide_scheduling} +@next_tutorial{tutorial_dnn_yolo} + ## Introduction In this tutorial you'll know how to run deep learning networks on Android device using OpenCV deep learning module. diff --git a/doc/tutorials/dnn/dnn_custom_layers/dnn_custom_layers.md b/doc/tutorials/dnn/dnn_custom_layers/dnn_custom_layers.md index 0486b31..a3f521d 100644 --- a/doc/tutorials/dnn/dnn_custom_layers/dnn_custom_layers.md +++ b/doc/tutorials/dnn/dnn_custom_layers/dnn_custom_layers.md @@ -1,5 +1,7 @@ # Custom deep learning layers support {#tutorial_dnn_custom_layers} +@prev_tutorial{tutorial_dnn_javascript} + ## Introduction Deep learning is a fast growing area. The new approaches to build neural networks usually introduce new types of layers. They could be modifications of existing diff --git a/doc/tutorials/dnn/dnn_googlenet/dnn_googlenet.markdown b/doc/tutorials/dnn/dnn_googlenet/dnn_googlenet.markdown index 4c5ebf9..c5992c3 100644 --- a/doc/tutorials/dnn/dnn_googlenet/dnn_googlenet.markdown +++ b/doc/tutorials/dnn/dnn_googlenet/dnn_googlenet.markdown @@ -1,6 +1,8 @@ Load Caffe framework models {#tutorial_dnn_googlenet} =========================== +@next_tutorial{tutorial_dnn_halide} + Introduction ------------ diff --git a/doc/tutorials/dnn/dnn_halide/dnn_halide.markdown b/doc/tutorials/dnn/dnn_halide/dnn_halide.markdown index 7271d08..5ab7774 100644 --- a/doc/tutorials/dnn/dnn_halide/dnn_halide.markdown +++ b/doc/tutorials/dnn/dnn_halide/dnn_halide.markdown @@ -1,5 +1,8 @@ # How to enable Halide backend for improve efficiency {#tutorial_dnn_halide} +@prev_tutorial{tutorial_dnn_googlenet} +@next_tutorial{tutorial_dnn_halide_scheduling} + ## Introduction This tutorial guidelines how to run your models in OpenCV deep learning module using Halide language backend. Halide is an open-source project that let us diff --git a/doc/tutorials/dnn/dnn_halide_scheduling/dnn_halide_scheduling.markdown b/doc/tutorials/dnn/dnn_halide_scheduling/dnn_halide_scheduling.markdown index a43ab63..b825da7 100644 --- a/doc/tutorials/dnn/dnn_halide_scheduling/dnn_halide_scheduling.markdown +++ b/doc/tutorials/dnn/dnn_halide_scheduling/dnn_halide_scheduling.markdown @@ -1,5 +1,8 @@ # How to schedule your network for Halide backend {#tutorial_dnn_halide_scheduling} +@prev_tutorial{tutorial_dnn_halide} +@next_tutorial{tutorial_dnn_android} + ## Introduction Halide code is the same for every device we use. But for achieving the satisfied efficiency we should schedule computations properly. In this tutorial we describe diff --git a/doc/tutorials/dnn/dnn_javascript/dnn_javascript.markdown b/doc/tutorials/dnn/dnn_javascript/dnn_javascript.markdown index 0a9caea..9ad632f 100644 --- a/doc/tutorials/dnn/dnn_javascript/dnn_javascript.markdown +++ b/doc/tutorials/dnn/dnn_javascript/dnn_javascript.markdown @@ -1,5 +1,8 @@ # How to run deep networks in browser {#tutorial_dnn_javascript} +@prev_tutorial{tutorial_dnn_yolo} +@next_tutorial{tutorial_dnn_custom_layers} + ## Introduction This tutorial will show us how to run deep learning models using OpenCV.js right in a browser. Tutorial refers a sample of face detection and face recognition diff --git a/doc/tutorials/dnn/dnn_yolo/dnn_yolo.markdown b/doc/tutorials/dnn/dnn_yolo/dnn_yolo.markdown index 76bf0ed..443f58c 100644 --- a/doc/tutorials/dnn/dnn_yolo/dnn_yolo.markdown +++ b/doc/tutorials/dnn/dnn_yolo/dnn_yolo.markdown @@ -1,6 +1,9 @@ YOLO DNNs {#tutorial_dnn_yolo} =============================== +@prev_tutorial{tutorial_dnn_android} +@next_tutorial{tutorial_dnn_javascript} + Introduction ------------ diff --git a/doc/tutorials/features2d/akaze_matching/akaze_matching.markdown b/doc/tutorials/features2d/akaze_matching/akaze_matching.markdown index d0f7724..406a1e6 100644 --- a/doc/tutorials/features2d/akaze_matching/akaze_matching.markdown +++ b/doc/tutorials/features2d/akaze_matching/akaze_matching.markdown @@ -1,6 +1,9 @@ AKAZE local features matching {#tutorial_akaze_matching} ============================= +@prev_tutorial{tutorial_detection_of_planar_objects} +@next_tutorial{tutorial_akaze_tracking} + Introduction ------------ diff --git a/doc/tutorials/features2d/akaze_tracking/akaze_tracking.markdown b/doc/tutorials/features2d/akaze_tracking/akaze_tracking.markdown index 518cada..58071ff 100644 --- a/doc/tutorials/features2d/akaze_tracking/akaze_tracking.markdown +++ b/doc/tutorials/features2d/akaze_tracking/akaze_tracking.markdown @@ -1,6 +1,9 @@ AKAZE and ORB planar tracking {#tutorial_akaze_tracking} ============================= +@prev_tutorial{tutorial_akaze_matching} +@next_tutorial{tutorial_homography} + Introduction ------------ diff --git a/doc/tutorials/features2d/detection_of_planar_objects/detection_of_planar_objects.markdown b/doc/tutorials/features2d/detection_of_planar_objects/detection_of_planar_objects.markdown index 4112b18..9febdb7 100644 --- a/doc/tutorials/features2d/detection_of_planar_objects/detection_of_planar_objects.markdown +++ b/doc/tutorials/features2d/detection_of_planar_objects/detection_of_planar_objects.markdown @@ -1,6 +1,10 @@ Detection of planar objects {#tutorial_detection_of_planar_objects} =========================== +@prev_tutorial{tutorial_feature_homography} +@next_tutorial{tutorial_akaze_matching} + + The goal of this tutorial is to learn how to use *features2d* and *calib3d* modules for detecting known planar objects in scenes. diff --git a/doc/tutorials/features2d/feature_description/feature_description.markdown b/doc/tutorials/features2d/feature_description/feature_description.markdown index 01ec0de..22badd6 100644 --- a/doc/tutorials/features2d/feature_description/feature_description.markdown +++ b/doc/tutorials/features2d/feature_description/feature_description.markdown @@ -1,6 +1,9 @@ Feature Description {#tutorial_feature_description} =================== +@prev_tutorial{tutorial_feature_detection} +@next_tutorial{tutorial_feature_flann_matcher} + Goal ---- diff --git a/doc/tutorials/features2d/feature_detection/feature_detection.markdown b/doc/tutorials/features2d/feature_detection/feature_detection.markdown index bb26586..686fade 100644 --- a/doc/tutorials/features2d/feature_detection/feature_detection.markdown +++ b/doc/tutorials/features2d/feature_detection/feature_detection.markdown @@ -1,6 +1,9 @@ Feature Detection {#tutorial_feature_detection} ================= +@prev_tutorial{tutorial_corner_subpixels} +@next_tutorial{tutorial_feature_description} + Goal ---- diff --git a/doc/tutorials/features2d/feature_flann_matcher/feature_flann_matcher.markdown b/doc/tutorials/features2d/feature_flann_matcher/feature_flann_matcher.markdown index 47ba7a0..35cdba9 100644 --- a/doc/tutorials/features2d/feature_flann_matcher/feature_flann_matcher.markdown +++ b/doc/tutorials/features2d/feature_flann_matcher/feature_flann_matcher.markdown @@ -1,6 +1,9 @@ Feature Matching with FLANN {#tutorial_feature_flann_matcher} =========================== +@prev_tutorial{tutorial_feature_description} +@next_tutorial{tutorial_feature_homography} + Goal ---- diff --git a/doc/tutorials/features2d/feature_homography/feature_homography.markdown b/doc/tutorials/features2d/feature_homography/feature_homography.markdown index 908f2c6..5380471 100644 --- a/doc/tutorials/features2d/feature_homography/feature_homography.markdown +++ b/doc/tutorials/features2d/feature_homography/feature_homography.markdown @@ -1,6 +1,9 @@ Features2D + Homography to find a known object {#tutorial_feature_homography} ============================================== +@prev_tutorial{tutorial_feature_flann_matcher} +@next_tutorial{tutorial_detection_of_planar_objects} + Goal ---- diff --git a/doc/tutorials/features2d/homography/homography.markdown b/doc/tutorials/features2d/homography/homography.markdown index 80c1984..3d8f97c 100644 --- a/doc/tutorials/features2d/homography/homography.markdown +++ b/doc/tutorials/features2d/homography/homography.markdown @@ -1,6 +1,8 @@ Basic concepts of the homography explained with code {#tutorial_homography} ==================================================== +@prev_tutorial{tutorial_akaze_tracking} + @tableofcontents Introduction {#tutorial_homography_Introduction} diff --git a/doc/tutorials/features2d/trackingmotion/corner_subpixels/corner_subpixels.markdown b/doc/tutorials/features2d/trackingmotion/corner_subpixels/corner_subpixels.markdown index 514fd33..db0a75b 100644 --- a/doc/tutorials/features2d/trackingmotion/corner_subpixels/corner_subpixels.markdown +++ b/doc/tutorials/features2d/trackingmotion/corner_subpixels/corner_subpixels.markdown @@ -1,6 +1,9 @@ Detecting corners location in subpixels {#tutorial_corner_subpixels} ======================================= +@prev_tutorial{tutorial_generic_corner_detector} +@next_tutorial{tutorial_feature_detection} + Goal ---- diff --git a/doc/tutorials/features2d/trackingmotion/generic_corner_detector/generic_corner_detector.markdown b/doc/tutorials/features2d/trackingmotion/generic_corner_detector/generic_corner_detector.markdown index aded0d1..bb19112 100644 --- a/doc/tutorials/features2d/trackingmotion/generic_corner_detector/generic_corner_detector.markdown +++ b/doc/tutorials/features2d/trackingmotion/generic_corner_detector/generic_corner_detector.markdown @@ -1,6 +1,10 @@ Creating your own corner detector {#tutorial_generic_corner_detector} ================================= +@prev_tutorial{tutorial_good_features_to_track} +@next_tutorial{tutorial_corner_subpixels} + + Goal ---- diff --git a/doc/tutorials/features2d/trackingmotion/good_features_to_track/good_features_to_track.markdown b/doc/tutorials/features2d/trackingmotion/good_features_to_track/good_features_to_track.markdown index 1c5c702..a91872f 100644 --- a/doc/tutorials/features2d/trackingmotion/good_features_to_track/good_features_to_track.markdown +++ b/doc/tutorials/features2d/trackingmotion/good_features_to_track/good_features_to_track.markdown @@ -1,6 +1,9 @@ Shi-Tomasi corner detector {#tutorial_good_features_to_track} ========================== +@prev_tutorial{tutorial_harris_detector} +@next_tutorial{tutorial_generic_corner_detector} + Goal ---- diff --git a/doc/tutorials/features2d/trackingmotion/harris_detector/harris_detector.markdown b/doc/tutorials/features2d/trackingmotion/harris_detector/harris_detector.markdown index a59a2ad..6855940 100644 --- a/doc/tutorials/features2d/trackingmotion/harris_detector/harris_detector.markdown +++ b/doc/tutorials/features2d/trackingmotion/harris_detector/harris_detector.markdown @@ -1,6 +1,8 @@ Harris corner detector {#tutorial_harris_detector} ====================== +@next_tutorial{tutorial_good_features_to_track} + Goal ---- diff --git a/doc/tutorials/gpu/gpu-basics-similarity/gpu_basics_similarity.markdown b/doc/tutorials/gpu/gpu-basics-similarity/gpu_basics_similarity.markdown index b511221..adc35fc 100644 --- a/doc/tutorials/gpu/gpu-basics-similarity/gpu_basics_similarity.markdown +++ b/doc/tutorials/gpu/gpu-basics-similarity/gpu_basics_similarity.markdown @@ -2,6 +2,8 @@ Similarity check (PNSR and SSIM) on the GPU {#tutorial_gpu_basics_similarity} =========================================== @todo update this tutorial +@next_tutorial{tutorial_gpu_thrust_interop} + Goal ---- diff --git a/doc/tutorials/gpu/gpu-thrust-interop/gpu_thrust_interop.markdown b/doc/tutorials/gpu/gpu-thrust-interop/gpu_thrust_interop.markdown index 0332808..7067e6d 100644 --- a/doc/tutorials/gpu/gpu-thrust-interop/gpu_thrust_interop.markdown +++ b/doc/tutorials/gpu/gpu-thrust-interop/gpu_thrust_interop.markdown @@ -1,6 +1,8 @@ Using a cv::cuda::GpuMat with thrust {#tutorial_gpu_thrust_interop} =========================================== +@prev_tutorial{tutorial_gpu_basics_similarity} + Goal ---- diff --git a/doc/tutorials/introduction/android_binary_package/O4A_SDK.markdown b/doc/tutorials/introduction/android_binary_package/O4A_SDK.markdown index 57e4c3f..d386d18 100644 --- a/doc/tutorials/introduction/android_binary_package/O4A_SDK.markdown +++ b/doc/tutorials/introduction/android_binary_package/O4A_SDK.markdown @@ -1,6 +1,10 @@ OpenCV4Android SDK {#tutorial_O4A_SDK} ================== +@prev_tutorial{tutorial_android_dev_intro} +@next_tutorial{tutorial_dev_with_OCV_on_Android} + + This tutorial was designed to help you with installation and configuration of OpenCV4Android SDK. This guide was written with MS Windows 7 in mind, though it should work with GNU Linux and Apple Mac diff --git a/doc/tutorials/introduction/android_binary_package/android_dev_intro.markdown b/doc/tutorials/introduction/android_binary_package/android_dev_intro.markdown index 5d43b22..dc153d3 100644 --- a/doc/tutorials/introduction/android_binary_package/android_dev_intro.markdown +++ b/doc/tutorials/introduction/android_binary_package/android_dev_intro.markdown @@ -1,6 +1,10 @@ Introduction into Android Development {#tutorial_android_dev_intro} ===================================== +@prev_tutorial{tutorial_clojure_dev_intro} +@next_tutorial{tutorial_O4A_SDK} + + This guide was designed to help you in learning Android development basics and setting up your working environment quickly. It was written with Windows 7 in mind, though it would work with Linux (Ubuntu), Mac OS X and any other OS supported by Android SDK. diff --git a/doc/tutorials/introduction/android_binary_package/android_ocl_intro.markdown b/doc/tutorials/introduction/android_binary_package/android_ocl_intro.markdown index 4bd1bc6..3b37e54 100644 --- a/doc/tutorials/introduction/android_binary_package/android_ocl_intro.markdown +++ b/doc/tutorials/introduction/android_binary_package/android_ocl_intro.markdown @@ -1,6 +1,10 @@ Use OpenCL in Android camera preview based CV application {#tutorial_android_ocl_intro} ===================================== +@prev_tutorial{tutorial_dev_with_OCV_on_Android} +@next_tutorial{tutorial_macos_install} + + This guide was designed to help you in use of [OpenCL ™](https://www.khronos.org/opencl/) in Android camera preview based CV application. It was written for [Eclipse-based ADT tools](http://developer.android.com/tools/help/adt.html) (deprecated by Google now), but it easily can be reproduced with [Android Studio](http://developer.android.com/tools/studio/index.html). diff --git a/doc/tutorials/introduction/android_binary_package/dev_with_OCV_on_Android.markdown b/doc/tutorials/introduction/android_binary_package/dev_with_OCV_on_Android.markdown index cd24514..6f9e6bd 100644 --- a/doc/tutorials/introduction/android_binary_package/dev_with_OCV_on_Android.markdown +++ b/doc/tutorials/introduction/android_binary_package/dev_with_OCV_on_Android.markdown @@ -1,6 +1,10 @@ Android Development with OpenCV {#tutorial_dev_with_OCV_on_Android} =============================== +@prev_tutorial{tutorial_O4A_SDK} +@next_tutorial{tutorial_android_ocl_intro} + + This tutorial has been created to help you use OpenCV library within your Android project. This guide was written with Windows 7 in mind, though it should work with any other OS supported by diff --git a/doc/tutorials/introduction/building_tegra_cuda/building_tegra_cuda.markdown b/doc/tutorials/introduction/building_tegra_cuda/building_tegra_cuda.markdown index fcee70c..983d012 100644 --- a/doc/tutorials/introduction/building_tegra_cuda/building_tegra_cuda.markdown +++ b/doc/tutorials/introduction/building_tegra_cuda/building_tegra_cuda.markdown @@ -1,6 +1,10 @@ Building OpenCV for Tegra with CUDA {#tutorial_building_tegra_cuda} =================================== +@prev_tutorial{tutorial_arm_crosscompile_with_cmake} +@next_tutorial{tutorial_display_image} + + @tableofcontents OpenCV with CUDA for Tegra diff --git a/doc/tutorials/introduction/clojure_dev_intro/clojure_dev_intro.markdown b/doc/tutorials/introduction/clojure_dev_intro/clojure_dev_intro.markdown index 1dd440c..a187da9 100644 --- a/doc/tutorials/introduction/clojure_dev_intro/clojure_dev_intro.markdown +++ b/doc/tutorials/introduction/clojure_dev_intro/clojure_dev_intro.markdown @@ -1,6 +1,10 @@ Introduction to OpenCV Development with Clojure {#tutorial_clojure_dev_intro} =============================================== +@prev_tutorial{tutorial_java_eclipse} +@next_tutorial{tutorial_android_dev_intro} + + As of OpenCV 2.4.4, OpenCV supports desktop Java development using nearly the same interface as for Android development. diff --git a/doc/tutorials/introduction/cross_referencing/tutorial_cross_referencing.markdown b/doc/tutorials/introduction/cross_referencing/tutorial_cross_referencing.markdown index c64c012..17bf776 100644 --- a/doc/tutorials/introduction/cross_referencing/tutorial_cross_referencing.markdown +++ b/doc/tutorials/introduction/cross_referencing/tutorial_cross_referencing.markdown @@ -1,6 +1,9 @@ Cross referencing OpenCV from other Doxygen projects {#tutorial_cross_referencing} ==================================================== +@prev_tutorial{tutorial_transition_guide} + + Cross referencing OpenCV ------------------------ diff --git a/doc/tutorials/introduction/crosscompilation/arm_crosscompile_with_cmake.markdown b/doc/tutorials/introduction/crosscompilation/arm_crosscompile_with_cmake.markdown index 3ef08f8..8d9c666 100644 --- a/doc/tutorials/introduction/crosscompilation/arm_crosscompile_with_cmake.markdown +++ b/doc/tutorials/introduction/crosscompilation/arm_crosscompile_with_cmake.markdown @@ -1,6 +1,10 @@ Cross compilation for ARM based Linux systems {#tutorial_arm_crosscompile_with_cmake} ============================================= +@prev_tutorial{tutorial_ios_install} +@next_tutorial{tutorial_building_tegra_cuda} + + This steps are tested on Ubuntu Linux 12.04, but should work for other Linux distributions. I case of other distributions package names and names of cross compilation tools may differ. There are several popular EABI versions that are used on ARM platform. This tutorial is written for *gnueabi* diff --git a/doc/tutorials/introduction/desktop_java/java_dev_intro.markdown b/doc/tutorials/introduction/desktop_java/java_dev_intro.markdown index c6e8c06..8511304 100644 --- a/doc/tutorials/introduction/desktop_java/java_dev_intro.markdown +++ b/doc/tutorials/introduction/desktop_java/java_dev_intro.markdown @@ -1,6 +1,10 @@ Introduction to Java Development {#tutorial_java_dev_intro} ================================ +@prev_tutorial{tutorial_windows_visual_studio_image_watch} +@next_tutorial{tutorial_java_eclipse} + + As of OpenCV 2.4.4, OpenCV supports desktop Java development using nearly the same interface as for Android development. This guide will help you to create your first Java (or Scala) application using OpenCV. We will use either [Apache Ant](http://ant.apache.org/) or [Simple Build Tool diff --git a/doc/tutorials/introduction/display_image/display_image.markdown b/doc/tutorials/introduction/display_image/display_image.markdown index ee6c200..a3cbc96 100644 --- a/doc/tutorials/introduction/display_image/display_image.markdown +++ b/doc/tutorials/introduction/display_image/display_image.markdown @@ -1,6 +1,9 @@ Getting Started with Images {#tutorial_display_image} =========================== +@prev_tutorial{tutorial_building_tegra_cuda} +@next_tutorial{tutorial_documentation} + Goal ---- diff --git a/doc/tutorials/introduction/documenting_opencv/documentation_tutorial.markdown b/doc/tutorials/introduction/documenting_opencv/documentation_tutorial.markdown index 058f392..96d39b4 100644 --- a/doc/tutorials/introduction/documenting_opencv/documentation_tutorial.markdown +++ b/doc/tutorials/introduction/documenting_opencv/documentation_tutorial.markdown @@ -1,6 +1,10 @@ Writing documentation for OpenCV {#tutorial_documentation} ================================ +@prev_tutorial{tutorial_display_image} +@next_tutorial{tutorial_transition_guide} + + @tableofcontents Doxygen overview {#tutorial_documentation_overview} diff --git a/doc/tutorials/introduction/ios_install/ios_install.markdown b/doc/tutorials/introduction/ios_install/ios_install.markdown index 1aaec1f..f893c7f 100644 --- a/doc/tutorials/introduction/ios_install/ios_install.markdown +++ b/doc/tutorials/introduction/ios_install/ios_install.markdown @@ -1,6 +1,9 @@ Installation in iOS {#tutorial_ios_install} =================== +@prev_tutorial{tutorial_macos_install} +@next_tutorial{tutorial_arm_crosscompile_with_cmake} + Required Packages ----------------- diff --git a/doc/tutorials/introduction/java_eclipse/java_eclipse.markdown b/doc/tutorials/introduction/java_eclipse/java_eclipse.markdown index ed1f527..f9ef84f 100644 --- a/doc/tutorials/introduction/java_eclipse/java_eclipse.markdown +++ b/doc/tutorials/introduction/java_eclipse/java_eclipse.markdown @@ -1,6 +1,10 @@ Using OpenCV Java with Eclipse {#tutorial_java_eclipse} ============================== +@prev_tutorial{tutorial_java_dev_intro} +@next_tutorial{tutorial_clojure_dev_intro} + + Since version 2.4.4 [OpenCV supports Java](http://opencv.org/opencv-java-api.html). In this tutorial I will explain how to setup development environment for using OpenCV Java with Eclipse in **Windows**, so you can enjoy the benefits of garbage collected, very refactorable (rename variable, diff --git a/doc/tutorials/introduction/linux_eclipse/linux_eclipse.markdown b/doc/tutorials/introduction/linux_eclipse/linux_eclipse.markdown index 4aca7ee..66ca510 100644 --- a/doc/tutorials/introduction/linux_eclipse/linux_eclipse.markdown +++ b/doc/tutorials/introduction/linux_eclipse/linux_eclipse.markdown @@ -1,6 +1,9 @@ Using OpenCV with Eclipse (plugin CDT) {#tutorial_linux_eclipse} ====================================== +@prev_tutorial{tutorial_linux_gcc_cmake} +@next_tutorial{tutorial_windows_install} + Prerequisites ------------- Two ways, one by forming a project directly, and another by CMake Prerequisites diff --git a/doc/tutorials/introduction/linux_gcc_cmake/linux_gcc_cmake.markdown b/doc/tutorials/introduction/linux_gcc_cmake/linux_gcc_cmake.markdown index fd44730..0f4d1c8 100644 --- a/doc/tutorials/introduction/linux_gcc_cmake/linux_gcc_cmake.markdown +++ b/doc/tutorials/introduction/linux_gcc_cmake/linux_gcc_cmake.markdown @@ -1,6 +1,10 @@ Using OpenCV with gcc and CMake {#tutorial_linux_gcc_cmake} =============================== +@prev_tutorial{tutorial_linux_install} +@next_tutorial{tutorial_linux_eclipse} + + @note We assume that you have successfully installed OpenCV in your workstation. - The easiest way of using OpenCV in your code is to use [CMake](http://www.cmake.org/). A few diff --git a/doc/tutorials/introduction/linux_install/linux_install.markdown b/doc/tutorials/introduction/linux_install/linux_install.markdown index 9f691c2..7d37495 100644 --- a/doc/tutorials/introduction/linux_install/linux_install.markdown +++ b/doc/tutorials/introduction/linux_install/linux_install.markdown @@ -1,6 +1,9 @@ Installation in Linux {#tutorial_linux_install} ===================== +@next_tutorial{tutorial_linux_gcc_cmake} + + The following steps have been tested for Ubuntu 10.04 but should work with other distros as well. Required Packages diff --git a/doc/tutorials/introduction/macos_install/macos_install.markdown b/doc/tutorials/introduction/macos_install/macos_install.markdown index 016e32a..4db1cbf 100644 --- a/doc/tutorials/introduction/macos_install/macos_install.markdown +++ b/doc/tutorials/introduction/macos_install/macos_install.markdown @@ -1,6 +1,10 @@ Installation in MacOS {#tutorial_macos_install} ===================== +@prev_tutorial{tutorial_android_ocl_intro} +@next_tutorial{tutorial_ios_install} + + The following steps have been tested for MacOSX (Mavericks) but should work with other versions as well. Required Packages diff --git a/doc/tutorials/introduction/transition_guide/transition_guide.markdown b/doc/tutorials/introduction/transition_guide/transition_guide.markdown index c5c4320..535ee46 100644 --- a/doc/tutorials/introduction/transition_guide/transition_guide.markdown +++ b/doc/tutorials/introduction/transition_guide/transition_guide.markdown @@ -1,6 +1,10 @@ Transition guide {#tutorial_transition_guide} ================ +@prev_tutorial{tutorial_documentation} +@next_tutorial{tutorial_cross_referencing} + + @tableofcontents Changes overview {#tutorial_transition_overview} diff --git a/doc/tutorials/introduction/windows_install/windows_install.markdown b/doc/tutorials/introduction/windows_install/windows_install.markdown index b30085c..8eca529 100644 --- a/doc/tutorials/introduction/windows_install/windows_install.markdown +++ b/doc/tutorials/introduction/windows_install/windows_install.markdown @@ -1,6 +1,10 @@ Installation in Windows {#tutorial_windows_install} ======================= +@prev_tutorial{tutorial_linux_eclipse} +@next_tutorial{tutorial_windows_visual_studio_opencv} + + The description here was tested on Windows 7 SP1. Nevertheless, it should also work on any other relatively modern version of Windows OS. If you encounter errors after following the steps described below, feel free to contact us via our [OpenCV Q&A forum](http://answers.opencv.org). We'll do our diff --git a/doc/tutorials/introduction/windows_visual_studio_image_watch/windows_visual_studio_image_watch.markdown b/doc/tutorials/introduction/windows_visual_studio_image_watch/windows_visual_studio_image_watch.markdown index 8bc2b32..770a9c3 100644 --- a/doc/tutorials/introduction/windows_visual_studio_image_watch/windows_visual_studio_image_watch.markdown +++ b/doc/tutorials/introduction/windows_visual_studio_image_watch/windows_visual_studio_image_watch.markdown @@ -1,6 +1,10 @@ Image Watch: viewing in-memory images in the Visual Studio debugger {#tutorial_windows_visual_studio_image_watch} =================================================================== +@prev_tutorial{tutorial_windows_visual_studio_opencv} +@next_tutorial{tutorial_java_dev_intro} + + Image Watch is a plug-in for Microsoft Visual Studio that lets you to visualize in-memory images (*cv::Mat* or *IplImage_* objects, for example) while debugging an application. This can be helpful for tracking down bugs, or for simply understanding what a given piece of code is doing. diff --git a/doc/tutorials/introduction/windows_visual_studio_opencv/windows_visual_studio_opencv.markdown b/doc/tutorials/introduction/windows_visual_studio_opencv/windows_visual_studio_opencv.markdown index caf7e57..d1f907a 100644 --- a/doc/tutorials/introduction/windows_visual_studio_opencv/windows_visual_studio_opencv.markdown +++ b/doc/tutorials/introduction/windows_visual_studio_opencv/windows_visual_studio_opencv.markdown @@ -1,6 +1,10 @@ How to build applications with OpenCV inside the "Microsoft Visual Studio" {#tutorial_windows_visual_studio_opencv} ========================================================================== +@prev_tutorial{tutorial_windows_install} +@next_tutorial{tutorial_windows_visual_studio_image_watch} + + Everything I describe here will apply to the `C\C++` interface of OpenCV. I start out from the assumption that you have read and completed with success the @ref tutorial_windows_install tutorial. Therefore, before you go any further make sure you have an OpenCV directory that contains the OpenCV diff --git a/doc/tutorials/ios/hello/hello.markdown b/doc/tutorials/ios/hello/hello.markdown index 44f14ad..fc6992c 100644 --- a/doc/tutorials/ios/hello/hello.markdown +++ b/doc/tutorials/ios/hello/hello.markdown @@ -1,6 +1,8 @@ OpenCV iOS Hello {#tutorial_hello} ================ +@next_tutorial{tutorial_image_manipulation} + Goal ---- diff --git a/doc/tutorials/ios/image_manipulation/image_manipulation.markdown b/doc/tutorials/ios/image_manipulation/image_manipulation.markdown index 42fefa5..61590d8 100644 --- a/doc/tutorials/ios/image_manipulation/image_manipulation.markdown +++ b/doc/tutorials/ios/image_manipulation/image_manipulation.markdown @@ -1,6 +1,9 @@ OpenCV iOS - Image Processing {#tutorial_image_manipulation} ============================= +@prev_tutorial{tutorial_hello} +@next_tutorial{tutorial_video_processing} + Goal ---- diff --git a/doc/tutorials/ios/video_processing/video_processing.markdown b/doc/tutorials/ios/video_processing/video_processing.markdown index 43b0b96..04bdd14 100644 --- a/doc/tutorials/ios/video_processing/video_processing.markdown +++ b/doc/tutorials/ios/video_processing/video_processing.markdown @@ -1,6 +1,9 @@ OpenCV iOS - Video Processing {#tutorial_video_processing} ============================= +@prev_tutorial{tutorial_image_manipulation} + + This tutorial explains how to process video frames using the iPhone's camera and OpenCV. Prerequisites: diff --git a/doc/tutorials/ml/introduction_to_pca/introduction_to_pca.markdown b/doc/tutorials/ml/introduction_to_pca/introduction_to_pca.markdown index 1908e96..55d9623 100644 --- a/doc/tutorials/ml/introduction_to_pca/introduction_to_pca.markdown +++ b/doc/tutorials/ml/introduction_to_pca/introduction_to_pca.markdown @@ -1,6 +1,8 @@ Introduction to Principal Component Analysis (PCA) {#tutorial_introduction_to_pca} ======================================= +@prev_tutorial{tutorial_non_linear_svms} + Goal ---- diff --git a/doc/tutorials/ml/introduction_to_svm/introduction_to_svm.markdown b/doc/tutorials/ml/introduction_to_svm/introduction_to_svm.markdown index d8bf40e..485070d 100644 --- a/doc/tutorials/ml/introduction_to_svm/introduction_to_svm.markdown +++ b/doc/tutorials/ml/introduction_to_svm/introduction_to_svm.markdown @@ -1,6 +1,8 @@ Introduction to Support Vector Machines {#tutorial_introduction_to_svm} ======================================= +@next_tutorial{tutorial_non_linear_svms} + Goal ---- diff --git a/doc/tutorials/ml/non_linear_svms/non_linear_svms.markdown b/doc/tutorials/ml/non_linear_svms/non_linear_svms.markdown index 7ad1266..8941b8c 100644 --- a/doc/tutorials/ml/non_linear_svms/non_linear_svms.markdown +++ b/doc/tutorials/ml/non_linear_svms/non_linear_svms.markdown @@ -1,6 +1,9 @@ Support Vector Machines for Non-Linearly Separable Data {#tutorial_non_linear_svms} ======================================================= +@prev_tutorial{tutorial_introduction_to_svm} +@next_tutorial{tutorial_introduction_to_pca} + Goal ---- diff --git a/doc/tutorials/objdetect/cascade_classifier/cascade_classifier.markdown b/doc/tutorials/objdetect/cascade_classifier/cascade_classifier.markdown index 138f43b..445f154 100644 --- a/doc/tutorials/objdetect/cascade_classifier/cascade_classifier.markdown +++ b/doc/tutorials/objdetect/cascade_classifier/cascade_classifier.markdown @@ -1,6 +1,8 @@ Cascade Classifier {#tutorial_cascade_classifier} ================== +@next_tutorial{tutorial_traincascade} + Goal ---- diff --git a/doc/tutorials/objdetect/traincascade.markdown b/doc/tutorials/objdetect/traincascade.markdown index 2a0b1ae..d7acdca 100644 --- a/doc/tutorials/objdetect/traincascade.markdown +++ b/doc/tutorials/objdetect/traincascade.markdown @@ -1,6 +1,8 @@ Cascade Classifier Training {#tutorial_traincascade} =========================== +@prev_tutorial{tutorial_cascade_classifier} + Introduction ------------ diff --git a/doc/tutorials/video/background_subtraction/background_subtraction.markdown b/doc/tutorials/video/background_subtraction/background_subtraction.markdown index 2e480f5..91dbd02 100644 --- a/doc/tutorials/video/background_subtraction/background_subtraction.markdown +++ b/doc/tutorials/video/background_subtraction/background_subtraction.markdown @@ -1,6 +1,8 @@ How to Use Background Subtraction Methods {#tutorial_background_subtraction} ========================================= +@next_tutorial{tutorial_meanshift} + - Background subtraction (BS) is a common and widely used technique for generating a foreground mask (namely, a binary image containing the pixels belonging to moving objects in the scene) by using static cameras. diff --git a/doc/tutorials/video/meanshift/meanshift.markdown b/doc/tutorials/video/meanshift/meanshift.markdown index 5b2d789..ffd61ad 100644 --- a/doc/tutorials/video/meanshift/meanshift.markdown +++ b/doc/tutorials/video/meanshift/meanshift.markdown @@ -1,6 +1,9 @@ Meanshift and Camshift {#tutorial_meanshift} ====================== +@prev_tutorial{tutorial_background_subtraction} +@next_tutorial{tutorial_optical_flow} + Goal ---- diff --git a/doc/tutorials/video/optical_flow/optical_flow.markdown b/doc/tutorials/video/optical_flow/optical_flow.markdown index d7a6896..45bbfa4 100644 --- a/doc/tutorials/video/optical_flow/optical_flow.markdown +++ b/doc/tutorials/video/optical_flow/optical_flow.markdown @@ -1,6 +1,8 @@ Optical Flow {#tutorial_optical_flow} ============ +@prev_tutorial{tutorial_meanshift} + Goal ---- diff --git a/doc/tutorials/videoio/intelperc.markdown b/doc/tutorials/videoio/intelperc.markdown index 4822c41..288d4b6 100644 --- a/doc/tutorials/videoio/intelperc.markdown +++ b/doc/tutorials/videoio/intelperc.markdown @@ -1,6 +1,9 @@ Using Creative Senz3D and other Intel Perceptual Computing SDK compatible depth sensors {#tutorial_intelperc} ======================================================================================= +@prev_tutorial{tutorial_kinect_openni} + + Depth sensors compatible with Intel Perceptual Computing SDK are supported through VideoCapture class. Depth map, RGB image and some other formats of output can be retrieved by using familiar interface of VideoCapture. diff --git a/doc/tutorials/videoio/kinect_openni.markdown b/doc/tutorials/videoio/kinect_openni.markdown index 8e28286..96d169e 100644 --- a/doc/tutorials/videoio/kinect_openni.markdown +++ b/doc/tutorials/videoio/kinect_openni.markdown @@ -1,6 +1,10 @@ Using Kinect and other OpenNI compatible depth sensors {#tutorial_kinect_openni} ====================================================== +@prev_tutorial{tutorial_video_write} +@next_tutorial{tutorial_intelperc} + + Depth sensors compatible with OpenNI (Kinect, XtionPRO, ...) are supported through VideoCapture class. Depth map, BGR image and some other formats of output can be retrieved by using familiar interface of VideoCapture. diff --git a/doc/tutorials/videoio/video-input-psnr-ssim/video_input_psnr_ssim.markdown b/doc/tutorials/videoio/video-input-psnr-ssim/video_input_psnr_ssim.markdown index 311ffe4..2cd0382 100644 --- a/doc/tutorials/videoio/video-input-psnr-ssim/video_input_psnr_ssim.markdown +++ b/doc/tutorials/videoio/video-input-psnr-ssim/video_input_psnr_ssim.markdown @@ -1,6 +1,8 @@ Video Input with OpenCV and similarity measurement {#tutorial_video_input_psnr_ssim} ================================================== +@next_tutorial{tutorial_video_write} + Goal ---- diff --git a/doc/tutorials/videoio/video-write/video_write.markdown b/doc/tutorials/videoio/video-write/video_write.markdown index dff53c1..feafc44 100644 --- a/doc/tutorials/videoio/video-write/video_write.markdown +++ b/doc/tutorials/videoio/video-write/video_write.markdown @@ -1,6 +1,9 @@ Creating a video with OpenCV {#tutorial_video_write} ============================ +@prev_tutorial{tutorial_video_input_psnr_ssim} +@next_tutorial{tutorial_kinect_openni} + Goal ---- diff --git a/doc/tutorials/viz/creating_widgets/creating_widgets.markdown b/doc/tutorials/viz/creating_widgets/creating_widgets.markdown index 3023e9b..c8b46d4 100644 --- a/doc/tutorials/viz/creating_widgets/creating_widgets.markdown +++ b/doc/tutorials/viz/creating_widgets/creating_widgets.markdown @@ -1,6 +1,9 @@ Creating Widgets {#tutorial_creating_widgets} ================ +@prev_tutorial{tutorial_transformations} +@next_tutorial{tutorial_histo3D} + Goal ---- diff --git a/doc/tutorials/viz/histo3D/histo3D.markdown b/doc/tutorials/viz/histo3D/histo3D.markdown index 10e9996..147e501 100644 --- a/doc/tutorials/viz/histo3D/histo3D.markdown +++ b/doc/tutorials/viz/histo3D/histo3D.markdown @@ -1,6 +1,8 @@ Creating a 3D histogram {#tutorial_histo3D} ================ +@prev_tutorial{tutorial_creating_widgets} + Goal ---- diff --git a/doc/tutorials/viz/launching_viz/launching_viz.markdown b/doc/tutorials/viz/launching_viz/launching_viz.markdown index 6a02b9b..498f50b 100644 --- a/doc/tutorials/viz/launching_viz/launching_viz.markdown +++ b/doc/tutorials/viz/launching_viz/launching_viz.markdown @@ -1,6 +1,8 @@ Launching Viz {#tutorial_launching_viz} ============= +@next_tutorial{tutorial_widget_pose} + Goal ---- diff --git a/doc/tutorials/viz/transformations/transformations.markdown b/doc/tutorials/viz/transformations/transformations.markdown index c10b90b..e6523d1 100644 --- a/doc/tutorials/viz/transformations/transformations.markdown +++ b/doc/tutorials/viz/transformations/transformations.markdown @@ -1,6 +1,9 @@ Transformations {#tutorial_transformations} =============== +@prev_tutorial{tutorial_widget_pose} +@next_tutorial{tutorial_creating_widgets} + Goal ---- diff --git a/doc/tutorials/viz/widget_pose/widget_pose.markdown b/doc/tutorials/viz/widget_pose/widget_pose.markdown index ea3b93e..d4e4f23 100644 --- a/doc/tutorials/viz/widget_pose/widget_pose.markdown +++ b/doc/tutorials/viz/widget_pose/widget_pose.markdown @@ -1,6 +1,9 @@ Pose of a widget {#tutorial_widget_pose} ================ +@prev_tutorial{tutorial_launching_viz} +@next_tutorial{tutorial_transformations} + Goal ---- -- 2.7.4