set(ALIASOF_libavcodec_VERSION 53.61.100)
set(ALIASOF_libavformat_VERSION 53.32.100)
set(ALIASOF_libavutil_VERSION 51.35.100)
-set(ALIASOF_libswscale_VERSION 2.1.100)
\ No newline at end of file
+set(ALIASOF_libswscale_VERSION 2.1.100)
set path=c:\dev\msys32\bin;%path% & gcc -Wall -shared -o opencv_ffmpeg.dll -O2 -x c++ -I../include -I../include/ffmpeg_ -I../../modules/highgui/src ffopencv.c -L../lib -lavformat -lavcodec -lavdevice -lswscale -lavutil -lwsock32
-set path=c:\dev\msys64\bin;%path% & gcc -m64 -Wall -shared -o opencv_ffmpeg_64.dll -O2 -x c++ -I../include -I../include/ffmpeg_ -I../../modules/highgui/src ffopencv.c -L../lib -lavformat64 -lavcodec64 -lavdevice64 -lswscale64 -lavutil64 -lavcore64 -lwsock32 -lws2_32
\ No newline at end of file
+set path=c:\dev\msys64\bin;%path% & gcc -m64 -Wall -shared -o opencv_ffmpeg_64.dll -O2 -x c++ -I../include -I../include/ffmpeg_ -I../../modules/highgui/src ffopencv.c -L../lib -lavformat64 -lavcodec64 -lavdevice64 -lswscale64 -lavutil64 -lavcore64 -lwsock32 -lws2_32
8. Then, go to <opencv>\3rdparty\ffmpeg, edit make.bat
(change paths to the actual paths to your msys32 and msys64 distributions) and then run make.bat
-
-
if(NOT BUILD_SHARED_LIBS)
install(TARGETS ${JASPER_LIBRARY} ARCHIVE DESTINATION ${OPENCV_3P_LIB_INSTALL_PATH} COMPONENT main)
endif()
-
/* Support Deflate compression */
#define ZIP_SUPPORT 1
-
-tbb*.tgz
\ No newline at end of file
+tbb*.tgz
if("${CMAKE_CURRENT_SOURCE_DIR}" STREQUAL "${CMAKE_CURRENT_BINARY_DIR}")
message(WARNING "The source directory is the same as binary directory. \"make clean\" may damage the source tree")
endif()
-
set_target_properties(opencv_haartraining PROPERTIES FOLDER "applications")
set_target_properties(opencv_haartraining_engine PROPERTIES FOLDER "applications")
endif()
-
void icvSave( const CvArr* ptr, const char* filename, int line );
#endif /* __CVCOMMON_H_ */
-
endif()
install(TARGETS ${the_target} RUNTIME DESTINATION bin COMPONENT main)
-
string(REPLACE "/Zi" "" CMAKE_C_FLAGS_DEBUG "${CMAKE_C_FLAGS_DEBUG}")
string(REPLACE "/Zi" "" CMAKE_CXX_FLAGS_DEBUG "${CMAKE_C_FLAGS_DEBUG}")
endif()
-
if(NOT ENABLE_NOISY_WARNINGS)
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} /wd4251") #class 'std::XXX' needs to have dll-interface to be used by clients of YYY
endif()
-endif()
\ No newline at end of file
+endif()
set(OpenCV_FOUND FALSE CACHE BOOL "" FORCE)
set(OPENCV_FOUND FALSE CACHE BOOL "" FORCE)
endif()
-
# See http://code.opencv.org/issues/1906 for additional details
set(MSV_NTDLL "ntdll")
set(IPP_LIBRARIES ${IPP_LIBRARIES} ${MSV_NTDLL}${IPP_LIB_SUFFIX})
-endif()
\ No newline at end of file
+endif()
mark_as_advanced(FORCE XIMEA_FOUND)
mark_as_advanced(FORCE XIMEA_PATH)
-mark_as_advanced(FORCE XIMEA_LIBRARY_DIR)
\ No newline at end of file
+mark_as_advanced(FORCE XIMEA_LIBRARY_DIR)
set(BUILD_PNG ${OPENCV_BUILD_3RDPARTY_LIBS} CACHE BOOL "Set via depricated OPENCV_BUILD_3RDPARTY_LIBS" FORCE)
unset(OPENCV_BUILD_3RDPARTY_LIBS CACHE)
endif()
-
cmake_parse_arguments(OCV_SOURCE_GROUP "" "" "GLOB" ${ARGN})
file(GLOB srcs ${OCV_SOURCE_GROUP_GLOB})
source_group(${group} FILES ${srcs})
-endfunction()
\ No newline at end of file
+endfunction()
printf("%d.%d ", prop.major, prop.minor);
}
return 0;
-}
\ No newline at end of file
+}
AVIFileInit();
AVIFileExit();
return 0;
-}
\ No newline at end of file
+}
int main(int, char**)
{
return 0;
-}
\ No newline at end of file
+}
file(APPEND ${OUTPUT} "const char* ${cl_filename}=\"${lines};\n")
endforeach()
-file(APPEND ${OUTPUT} "}\n}\n")
\ No newline at end of file
+file(APPEND ${OUTPUT} "}\n}\n")
MESSAGE(STATUS "File \"$ENV{DESTDIR}${file}\" does not exist.")
ENDIF(EXISTS "$ENV{DESTDIR}${file}")
ENDFOREACH(file)
-
-
/* Define to 1 if your processor stores words with the most significant byte
first (like Motorola and SPARC, unlike Intel and VAX). */
#cmakedefine WORDS_BIGENDIAN
-
*
*/
-@OPENCV_MODULE_DEFINITIONS_CONFIGMAKE@
\ No newline at end of file
+@OPENCV_MODULE_DEFINITIONS_CONFIGMAKE@
install(FILES "${f}" DESTINATION "${OPENCV_DOC_INSTALL_PATH}" OPTIONAL)
endforeach()
-endif()
\ No newline at end of file
+endif()
element.parentNode.replaceChild(iframe, element);
}
}
-
div.sphinxsidebar #searchbox input[type="submit"] {
width:auto;
-}
\ No newline at end of file
+}
bodyfont = sans-serif
headfont = 'Trebuchet MS', sans-serif
guifont = "Lucida Sans","Lucida Sans Unicode","Lucida Grande",Verdana,Arial,Helvetica,sans-serif
-lang = none
\ No newline at end of file
+lang = none
for m in opencv_module_list:
print "\n\n*************************** " + m + " *************************\n"
p.check_module_docs(m)
-
-
#7 & #8 & #9
\end{bmatrix}
}
-
output.append(LineAxis(x, start, x, end, start, end, bars, False, False, **self.attr).SVG(trans))
return output
-
#.
Calculate reprojection error like it is done in ``calibration`` sample (see ``opencv/samples/cpp/calibration.cpp``, function ``computeReprojectionErrors``).
-Question: how to calculate the distance from the camera origin to any of the corners?
\ No newline at end of file
+Question: how to calculate the distance from the camera origin to any of the corners?
<div align="center">
<iframe title="File Input and Output using XML and YAML files in OpenCV" width="560" height="349" src="http://www.youtube.com/embed/A4yqVnByMMM?rel=0&loop=1" frameborder="0" allowfullscreen align="middle"></iframe>
</div>
-
<div align="center">
<iframe title="Interoperability with OpenCV 1" width="560" height="349" src="http://www.youtube.com/embed/qckm-zvo31w?rel=0&loop=1" frameborder="0" allowfullscreen align="middle"></iframe>
</div>
-
-
-
../random_generator_and_text/random_generator_and_text
../discrete_fourier_transform/discrete_fourier_transform
../file_input_output_with_xml_yml/file_input_output_with_xml_yml
- ../interoperability_with_OpenCV_1/interoperability_with_OpenCV_1
\ No newline at end of file
+ ../interoperability_with_OpenCV_1/interoperability_with_OpenCV_1
-Include in this directory only defintion files. None of the reST files entered here will be parsed by the Sphinx Builder.
\ No newline at end of file
+Include in this directory only defintion files. None of the reST files entered here will be parsed by the Sphinx Builder.
.. note::
- Unfortunetly we have no tutorials into this section. And you can help us with that, since OpenCV is a community effort. If you have a tutorial suggestion or you have written a tutorial yourself (or coded a sample code) that you would like to see here, please contact follow these instructions: :ref:`howToWriteTutorial` and :how_to_contribute:`How to contribute <>`.
\ No newline at end of file
+ Unfortunetly we have no tutorials into this section. And you can help us with that, since OpenCV is a community effort. If you have a tutorial suggestion or you have written a tutorial yourself (or coded a sample code) that you would like to see here, please contact follow these instructions: :ref:`howToWriteTutorial` and :how_to_contribute:`How to contribute <>`.
.. image:: images/Feature_Description_BruteForce_Result.jpg
:align: center
:height: 200pt
-
-
-
.. image:: images/Feature_Detection_Result_b.jpg
:align: center
:height: 200pt
-
.. image:: images/Feature_Homography_Result.jpg
:align: center
:height: 200pt
-
../feature_flann_matcher/feature_flann_matcher
../feature_homography/feature_homography
../detection_of_planar_objects/detection_of_planar_objects
-
.. image:: images/Corner_Subpixeles_Result.jpg
:align: center
-
.. image:: images/My_Shi_Tomasi_corner_detector_Result.jpg
:align: center
-
.. image:: images/Feature_Detection_Result_a.jpg
:align: center
-
-
.. image:: images/Harris_Detector_Result.jpg
:align: center
-
-
.. raw:: latex
\pagebreak
-
../trackbar/trackbar
../video-input-psnr-ssim/video-input-psnr-ssim
- ../video-write/video-write
\ No newline at end of file
+ ../video-write/video-write
.. image:: images/Adding_Trackbars_Tutorial_Result_1.jpg
:alt: Adding Trackbars - Lena
:align: center
-
-
-
-
-
.. image:: images/Histogram_Calculation_Result.jpg
:align: center
-
.. image:: images/Template_Matching_Image_Result.jpg
:align: center
-
:align: center
* Notice how the image is superposed to the black background on the edge regions.
-
-
-
:align: center
You may observe that the number of lines detected vary while you change the *threshold*. The explanation is sort of evident: If you establish a higher threshold, fewer lines will be detected (since you will need more points to declare a line detected).
-
:alt: Result 0 for remapping
:width: 250pt
:align: center
-
:alt: Original image
:width: 250pt
:align: center
-
.. image:: images/Morphology_2_Tutorial_Cover.jpg
:alt: Morphology 2: Result sample
:align: center
-
.. image:: images/Pyramids_Tutorial_PyrUp_Result.jpg
:alt: Pyramids: PyrUp Result
:align: center
-
-
.. |BRC_1| image:: images/Bounding_Rects_Circles_Result.jpg
:align: middle
-
.. |BRE_1| image:: images/Bounding_Rotated_Ellipses_Result.jpg
:align: middle
-
.. |contour_1| image:: images/Find_Contours_Result.jpg
:align: middle
-
.. |Hull_1| image:: images/Hull_Result.jpg
:align: middle
-
.. |MU_2| image:: images/Moments_Result2.jpg
:width: 250pt
:align: middle
-
.. |PPT_1| image:: images/Point_Polygon_Test_Result.jpg
:align: middle
-
../shapedescriptors/bounding_rotated_ellipses/bounding_rotated_ellipses
../shapedescriptors/moments/moments
../shapedescriptors/point_polygon_test/point_polygon_test
-
-
-
however if you submitted it, be sure that eventually it will show up.
If you have any questions or advices relating to this tutorial you can contact us at
--delete-admin@-delete-opencv.org (delete the -delete- parts of that email address).
\ No newline at end of file
+-delete-admin@-delete-opencv.org (delete the -delete- parts of that email address).
a. You can also optionally modify the ``Build command:`` from ``make`` to something like ``make VERBOSE=1 -j4`` which tells the compiler to produce detailed symbol files for debugging and also to compile in 4 parallel threads.
#. Done!
-
-
-
.. note::
If the size of the created library is a critical issue (like in case of an Android build) you can use the ``install/strip`` command to get the smallest size as possible. The *stripped* version appears to be twice as small. However, we do not recommend using this unless those extra megabytes do really matter.
-
.. image:: images/output.png
:alt: output
:align: center
-
<div align="center">
<iframe width="560" height="350" src="http://www.youtube.com/embed/Ko3K_xdhJ1I" frameborder="0" allowfullscreen></iframe>
- </div>
\ No newline at end of file
+ </div>
.. image:: images/result.png
:alt: The seperated planes
:align: center
-
<div align="center">
<iframe title="Support Vector Machines for Non-Linearly Separable Data" width="560" height="349" src="http://www.youtube.com/embed/vFv2yPcSo-Q?rel=0&loop=1" frameborder="0" allowfullscreen align="middle"></iframe>
- </div>
\ No newline at end of file
+ </div>
.. image:: images/Cascade_Classifier_Tutorial_Result_LBP.jpg
:align: center
:height: 300pt
-
install(FILES "opencv2/opencv.hpp"
DESTINATION ${OPENCV_INCLUDE_INSTALL_PATH}/opencv2
COMPONENT main)
-
-
#endif //__cplusplus
#endif // __OPENCV_OLD_CV_H_
-
#include "opencv2/core/eigen.hpp"
#endif
-
void setCameraPropertyC(void* camera, int propIdx, double value);
void applyCameraPropertiesC(void** camera);
}
-
int frameHeight;
};
-#endif
\ No newline at end of file
+#endif
:maxdepth: 2
camera_calibration_and_3d_reconstruction
-
};
#endif // _CV_MODEL_EST_H_
-
pX[i] = (pb[i] - sum) / A2[i];
}
}
-
}
TEST(Calib3d_EstimateAffineTransform, accuracy) { CV_Affine3D_EstTest test; test.safe_run(); }
-
TEST(Calib3d_CalibrateCamera_C, badarg) { CV_CameraCalibrationBadArgTest test; test.safe_run(); }
TEST(Calib3d_Rodrigues_C, badarg) { CV_Rodrigues2BadArgTest test; test.safe_run(); }
TEST(Calib3d_ProjectPoints_C, badarg) { CV_ProjectPoints2BadArgTest test; test.safe_run(); }
-
-
return generateChessBoard(bg, camMat, distCoeffs, zero, pb1, pb2,
squareSize.width, squareSize.height, pts3d, corners);
}
-
};
TEST(Calib3d_ComposeRT, accuracy) { CV_composeRT_Test test; test.safe_run(); }
-
};
TEST(Calib3d_ReprojectImageTo3D, accuracy) { CV_ReprojectImageTo3DTest test; test.safe_run(); }
-
EXPECT_LT(tnorm, 1e-6);
}
-#endif
\ No newline at end of file
+#endif
}
}
-TEST(Calib3d_Undistort, accuracy) { CV_UndistortTest test; test.safe_run(); }
\ No newline at end of file
+TEST(Calib3d_Undistort, accuracy) { CV_UndistortTest test; test.safe_run(); }
.. literalinclude:: etc/at.txt
:language: none
:linenos:
-
-
* :ref:`genindex`
* :ref:`modindex`
* :ref:`search`
-
add_executable(facerec_lbph facerec_lbph.cpp)
target_link_libraries(facerec_lbph opencv_contrib opencv_core opencv_imgproc opencv_highgui)
-
+---------------------------------+----------------------------------------------------------------------------+
| 0.2 (20%), 0.2 (20%), (70,70) | .. image:: ../img/tutorial/gender_classification/arnie_20_20_70_70.jpg |
+---------------------------------+----------------------------------------------------------------------------+
-
-
.. image:: ../img/eigenface_reconstruction_opencv.png
:align: center
-
+---------------------------------+----------------------------------------------------------------------------+
| 0.2 (20%), 0.2 (20%), (70,70) | .. image:: ../img/tutorial/gender_classification/arnie_20_20_70_70.jpg |
+---------------------------------+----------------------------------------------------------------------------+
-
-
**Note**:
The method is not constant, so you should not use the same ``StereoVar`` instance from different threads simultaneously.
-
-
}
#endif /* __OPENCV_CONTRIB_RETINA_HPP__ */
-
{
return prev_center;
}
-
-
//M*/
#include "precomp.hpp"
-
center.y = (float)(prev_center.y + prev_trackwindow.height/2.0);
return center;
}
-
searchMode = tsTracking;
}
};
-
Rect CvHybridTracker::getTrackingWindow() {
return prev_window;
}
-
return (*_magnoYOutput);
}
}
-
-
}
#endif /*MagnoRetinaFilter_H_*/
-
-
void Retina::activateContoursProcessing(const bool activate){_retinaFilter->activateContoursProcessing(activate);}
} // end of namespace cv
-
}
#endif /*RETINACOLOR_HPP_*/
-
-
}
#endif /*RETINACLASSES_H_*/
-
-
-
-
u.release();
}
-} // namespace
\ No newline at end of file
+} // namespace
#include <iostream>
#endif
-
* Add public virtual method ``AlgorithmInfo* info() const;`` to your class.
* Add constructor function, ``AlgorithmInfo`` instance and implement the ``info()`` method. The simplest way is to take http://code.opencv.org/projects/opencv/repository/revisions/master/entry/modules/ml/src/ml_init.cpp as the reference and modify it according to the list of your parameters.
* Add some public function (e.g. ``initModule_<mymodule>()``) that calls info() of your algorithm and put it into the same source file as ``info()`` implementation. This is to force C++ linker to include this object file into the target application. See ``Algorithm::create()`` for details.
-
clustering
utility_and_system_functions_and_macros
opengl_interop
-
Symbols that cannot be rendered using the specified font are
replaced by question marks. See
:ocv:func:`getTextSize` for a text rendering code example.
-
:param storage: Container for the sequence
The function puts pointers of all nodes reachable from ``first`` into a single sequence. The pointers are written sequentially in the depth-first order.
-
(480, 640, 3) 1
.. note:: In the new Python wrappers (**cv2** module) the function is not needed, since cv2 can process Numpy arrays (and this is the only supported array type).
-
TEST_CYCLE_MULTIRUN(runs) merge( (vector<Mat> &)mv, dst );
SANITY_CHECK(dst, 1e-12);
-}
\ No newline at end of file
+}
cv::normalize( src, dst, a, b, norm_type, dst.type(), mask );
}
-/* End of file. */
\ No newline at end of file
+/* End of file. */
cv::SVD::backSubst(w, u, v, rhs, dst);
CV_Assert( dst.data == dst0.data );
-}
\ No newline at end of file
+}
ASSERT_EQ(4, cvRound(3.5));
ASSERT_EQ(-2, cvRound(-2.5));
ASSERT_EQ(-4, cvRound(-3.5));
-}
\ No newline at end of file
+}
TEST(Core_DS_Set, basic_operations) { Core_SetTest test; test.safe_run(); }
TEST(Core_DS_Graph, basic_operations) { Core_GraphTest test; test.safe_run(); }
TEST(Core_DS_Graph, scan) { Core_GraphScanTest test; test.safe_run(); }
-
-
};
TEST(Core_DFT, complex_output) { Core_DFTComplexOutputTest test; test.safe_run(); }
-
-
set(the_description "2D Features Framework")
ocv_define_module(features2d opencv_imgproc opencv_flann OPTIONAL opencv_highgui)
-
VectorDescriptorMatcher matcher( new SurfDescriptorExtractor,
new BruteForceMatcher<L2<float> > );
-
-
:param color: Color of keypoints.
:param flags: Flags setting drawing features. Possible ``flags`` bit values are defined by ``DrawMatchesFlags``. See details above in :ocv:func:`drawMatches` .
-
Returns an image descriptor type.
.. ocv:function:: int BOWImgDescriptorExtractor::descriptorType() const
-
SANITY_CHECK_KEYPOINTS(points);
}
-
}
TEST(Features2d_BRISK, regression) { CV_BRISKTest test; test.safe_run(); }
-
}
TEST(Features2d_FAST, regression) { CV_FastTest test; test.safe_run(); }
-
CV_FeatureDetectorKeypointsTest test(Algorithm::create<FeatureDetector>("Feature2D.Dense"));
test.safe_run();
}
-
-
}
TEST(Features2d_MSER, DISABLED_regression) { CV_MserTest test; test.safe_run(); }
-
set(the_description "Clustering and Search in Multi-Dimensional Spaces")
ocv_define_module(flann opencv_core)
-
}
#endif //OPENCV_FLANN_RANDOM_H
-
-
}
#endif //OPENCV_FLANN_RESULTSET_H
-
}
void dummyfunc() {}
-}
\ No newline at end of file
+}
#include "opencv2/flann/flann_base.hpp"
#endif
-
{
CV_EXPORTS static cudaStream_t getStream(const Stream& stream);
};
-
%
performance increase comparing to the single Fermi GPU. For a source code example, see
http://code.opencv.org/projects/opencv/repository/revisions/master/entry/samples/gpu/.
-
}}}
#endif /* __OPENCV_GPU_DEVICE_BLOCK_HPP__ */
-
-
outputFile = open(sys.argv[2], 'w')
outputFile.writelines(lines)
outputFile.close()
-
set_target_properties(${the_target} PROPERTIES LINK_FLAGS "/NODEFAULTLIB:atlthunk.lib /NODEFAULTLIB:atlsd.lib /DEBUG")
endif()
endif()
-
createButton("button6",callbackButton2,NULL,CV_PUSH_BUTTON,1);
..
-
-
:param image: The written frame
The functions/methods write the specified image to video file. It must have the same size as has been specified when opening the video writer.
-
UIImage* MatToUIImage(const cv::Mat& image);
void UIImageToMat(const UIImage* image,
- cv::Mat& m, bool alphaExist = false);
\ No newline at end of file
+ cv::Mat& m, bool alphaExist = false);
SANITY_CHECK(dummy);
}
-#endif // BUILD_WITH_VIDEO_INPUT_SUPPORT
\ No newline at end of file
+#endif // BUILD_WITH_VIDEO_INPUT_SUPPORT
SANITY_CHECK(dummy);
}
-#endif // BUILD_WITH_VIDEO_OUTPUT_SUPPORT
\ No newline at end of file
+#endif // BUILD_WITH_VIDEO_OUTPUT_SUPPORT
return NULL;
}
-#endif
\ No newline at end of file
+#endif
You may not use, or allow anyone else to use the icons to create pornographic, libelous, obscene, or defamatory material.
-All icon files are provided "as is". You agree not to hold IconEden.com liable for any damages that may occur due to use, or inability to use, icons or image data from IconEden.com.
\ No newline at end of file
+All icon files are provided "as is". You agree not to hold IconEden.com liable for any damages that may occur due to use, or inability to use, icons or image data from IconEden.com.
image.CGImage);
CGContextRelease(contextRef);
CGColorSpaceRelease(colorSpace);
-}
\ No newline at end of file
+}
EXPECT_EQ(fourcc, (unsigned long)(unsigned)fourccFromParam);
}
-}
\ No newline at end of file
+}
#include "test_precomp.hpp"
CV_TEST_MAIN("highgui")
-
#if BUILD_WITH_VIDEO_INPUT_SUPPORT && defined HAVE_FFMPEG
TEST (Highgui_Video, seek_progressive) { CV_VideoProgressivePositioningTest test; test.safe_run(); }
TEST (Highgui_Video, seek_random) { CV_VideoRandomPositioningTest test; test.safe_run(); }
-#endif
\ No newline at end of file
+#endif
.. seealso::
:ocv:func:`cartToPolar`
-
where ``undistort()`` is an approximate iterative algorithm that estimates the normalized original point coordinates out of the normalized distorted point coordinates ("normalized" means that the coordinates do not depend on the camera matrix).
The function can be used for both a stereo camera head or a monocular camera (when R is empty).
-
-
* An example using the GrabCut algorithm can be found at opencv_source_code/samples/cpp/grabcut.cpp
- * (Python) An example using the GrabCut algorithm can be found at opencv_source_code/samples/python2/grabcut.py
\ No newline at end of file
+ * (Python) An example using the GrabCut algorithm can be found at opencv_source_code/samples/python2/grabcut.py
.. note::
- * (Python) An example on how to match mouse selected regions in an image can be found at opencv_source_code/samples/python2/mouse_and_match.py
\ No newline at end of file
+ * (Python) An example on how to match mouse selected regions in an image can be found at opencv_source_code/samples/python2/mouse_and_match.py
extractChannel(dst, l1, 0);
SANITY_CHECK(l1, 2e-5);
-}
\ No newline at end of file
+}
TEST_CYCLE() cornerHarris(src, dst, blockSize, apertureSize, k, borderType);
SANITY_CHECK(dst, 2e-5);
-}
\ No newline at end of file
+}
SANITY_CHECK(filteredImage, 1e-3);
}
-
-
SANITY_CHECK(transformCoefficient, 1e-5);
}
-
box.angle = (float)(box.angle*180/CV_PI);
return box;
}
-
}
TEST(Imgproc_ApproxPoly, accuracy) { CV_ApproxPolyTest test; test.safe_run(); }
-
TEST(Imgproc_FitEllipse, small) { CV_FitEllipseSmallTest test; test.safe_run(); }
/* End of file. */
-
TEST(Imgproc_DistanceTransform, accuracy) { CV_DisTransTest test; test.safe_run(); }
-
-
};
TEST(Imgproc_Filtering, supportedFormats) { CV_FilterSupportedFormatsTest test; test.safe_run(); }
-
TEST(Imgproc_HoughLines, regression) { CV_StandartHoughLinesTest test; test.safe_run(); }
TEST(Imgproc_HoughLinesP, regression) { CV_ProbabilisticHoughLinesTest test; test.safe_run(); }
-
}
TEST(Imgproc_Threshold, accuracy) { CV_ThreshTest test; test.safe_run(); }
-
}
TEST(Imgproc_Watershed, regression) { CV_WatershedTest test; test.safe_run(); }
-
local.properties
proguard-project.txt
project.properties
-default.properties
\ No newline at end of file
+default.properties
assertMatEqual(truth, markers);
}
-}
\ No newline at end of file
+}
print "Not tested methods found:", len(funcs)
print "Total tests found:", parser.get_tests_count()
print "Empty test stubs found:", parser.get_empty_stubs_count()
-
#print "Generating module '" + module + "' from headers:\n\t" + "\n\t".join(srcfiles)
generator = JavaWrapperGenerator()
generator.gen(srcfiles, module, dstdir)
-
# pragma warning(disable:4800 4244)
#endif
-#endif //__JAVA_COMMON_H__
\ No newline at end of file
+#endif //__JAVA_COMMON_H__
cv::redirectError(0);
else
cv::redirectError((cv::ErrorCallback)quietCallback);
-}
\ No newline at end of file
+}
//do nothing
}
-} // extern "C"
\ No newline at end of file
+} // extern "C"
} // extern "C"
-#endif //__ANDROID__
\ No newline at end of file
+#endif //__ANDROID__
import shutil
for f in os.listdir("."):
- shutil.copyfile(f, os.path.join("../../../../../../modules/java/generator/src/java/", "android+" + f));
\ No newline at end of file
+ shutil.copyfile(f, os.path.join("../../../../../../modules/java/generator/src/java/", "android+" + f));
* @return Returns names of OpenCV libraries, separated by semicolon.
*/
String getLibraryList(String version);
-}
\ No newline at end of file
+}
virtual int descriptorType() const;
protected:
...
- }
\ No newline at end of file
+ }
:param fs: A file storage with the trained model.
:param node: The parent map. If it is NULL, the function searches a node with parameters in all the top-level nodes (streams), starting with the first one.
-
..
The macros ``GetHistValue`` return a pointer to the specified bin of the 1D, 2D, 3D, or N-D histogram. In case of a sparse histogram, the function creates a new bin and sets it to 0, unless it exists already.
-
:param vely: Vertical component of the optical flow of the same size as input images, 32-bit floating-point, single-channel
The function computes the flow for every pixel of the first input image using the Lucas and Kanade algorithm [Lucas81]_. The function is obsolete. To track sparse features, use :ocv:func:`calcOpticalFlowPyrLK`. To track all the pixels, use :ocv:func:`calcOpticalFlowFarneback`.
-
-
:param pt: Inserted point.
The function inserts a single point into a subdivision and modifies the subdivision topology appropriately. If a point with the same coordinates exists already, no new point is added. The function returns a pointer to the allocated point. No virtual point coordinates are calculated at this stage.
-
/*======================================================================================*/
#endif/*_CV_VM_H_*/
-
ppT[0]->Release();
ppT[0] = 0;
}
-
}
/* ======================== Analyser modules ============================= */
-
return (CvBlobTrackAnalysis*)pIOR;
}/* cvCreateCvBlobTrackAnalysisIOR */
/* ======================== Analyser modules ============================= */
-
CvBlobTrackAnalysis* cvCreateModuleBlobTrackAnalysisTrackDist()
{return (CvBlobTrackAnalysis*) new CvBlobTrackAnalysisTrackDist;}
-
{
return (CvBlobTrackGen*) new CvBlobTrackGen1(0);
}
-
-
{
return (CvBlobTrackGen*) new CvBlobTrackGenYML;
}
-
-
TIME_END("TrackAnalysis",m_BlobList.GetBlobNum())
} /* CvBlobTrackerAuto1::Process */
-
{
return cvCreateBlobTrackerList(cvCreateBlobTrackerOneMSPF);
}
-
{
return cvCreateBlobTrackerList(cvCreateBlobTrackerOneMSFGS);
}
-
return (CvBlobTrackPredictor*) new CvBlobTrackPredictKalman;
}
/*======================= KALMAN PREDICTOR =========================*/
-
}//end cvBronKerbosch
#endif
-
#endif
// End of file decomppoly.cpp
-
return result;
} /* cvDetectNewBlob */
-
-
/* Blob detector constructor: */
//CvBlobDetector* cvCreateBlobDetectorReal(CvTestSeq* pTestSeq){return new CvBlobDetectorReal(pTestSeq);}
-
*/
#endif /* WIN32 */
-
-
-
-
}
}// inline void RFace::ResizeRect(CvRect * lpRect,long lDir,long lD)
-
m_pCurElem = m_pCurElem->m_pNext;
return m_pCurElem->m_pFace;
}//Face * FaceDetectionList::GetData()
-
-
#endif//__FACETEMPLATE_H__
-
return lpSeq;
}//cvPostBoostingFindFace(IplImage * Image)
-
IPPI_CALL( icvFindHandRegionA( points, count, indexs, line, size, jc,
center, storage, numbers ));
}
-
PrPoint->y = PointO->y - VectorAO.y + alfa*VectorAB.y;
return;
}//end of _cvProjectionPointToSegment
-
-
-
-
return error;
} /* icvPoint7 */
-
second_corr /* correspond information for the 2nd seq */
));
}
-
cvReleaseImage(&pImgAdd);
} /* cvAddNoise */
-
dbLMS = dbXtXt + dbYtYt - ((double)pow(dbXtXs + dbYtYs,2) + (double)pow(dbXtYs - dbYtXs,2)) / del;
return dbLMS;
}
-
};
TEST(Legacy_BruteForceMatcher, accuracy) { CV_BruteForceMatcherTest test; test.safe_run(); }
-
TEST(Legacy_OpticalFlow, accuracy) { CV_OptFlowTest test; test.safe_run(); }
-
-
TEST(Legacy_Subdiv, correctness) { CV_SubdivTest test; test.safe_run(); }
/* End of file. */
-
.. [Breiman84] Breiman, L., Friedman, J. Olshen, R. and Stone, C. (1984), *Classification and Regression Trees*, Wadsworth.
-
cvReleaseMat( &trainData );
return 0;
}
-
-
.. ocv:function:: double* CvANN_MLP::get_weights(int layer)
:param layer: Index of the particular layer.
-
The method is used to predict the response for a new sample. In case of a classification, the method returns the class label. In case of a regression, the method returns the output function value. The input sample must have as many components as the ``train_data`` passed to ``train`` contains. If the ``var_idx`` parameter is passed to ``train``, it is remembered and then is used to extract only the necessary components from the input sample in the method ``predict``.
The suffix ``const`` means that prediction does not affect the internal model state, so the method can be safely called from within different threads.
-
#endif
/* End of file. */
-
TEST(ML_EM, accuracy) { CV_EMTest test; test.safe_run(); }
TEST(ML_EM, save_load) { CV_EMTest_SaveLoad test; test.safe_run(); }
TEST(ML_EM, classification) { CV_EMTest_Classification test; test.safe_run(); }
-
.. note::
- * OCL : An example of the SURF detector can be found at opencv_source_code/samples/ocl/surf_matcher.cpp
\ No newline at end of file
+ * OCL : An example of the SURF detector can be found at opencv_source_code/samples/ocl/surf_matcher.cpp
}
}
-#endif //__OPENCV_NONFREE_OCL_HPP__
\ No newline at end of file
+#endif //__OPENCV_NONFREE_OCL_HPP__
SANITY_CHECK(cpu_dp, 1);
}
-#endif // HAVE_OPENCV_OCL
\ No newline at end of file
+#endif // HAVE_OPENCV_OCL
TEST(Features2d_Detectors, regression) { CV_DetectorsTest test; test.safe_run(); }
-
-
-
FeatureDetectorUsingMaskTest test(Algorithm::create<FeatureDetector>("Feature2D.SURF"));
test.safe_run();
}
-
.. [Felzenszwalb2010] Felzenszwalb, P. F. and Girshick, R. B. and McAllester, D. and Ramanan, D. *Object Detection with Discriminatively Trained Part Based Models*. PAMI, vol. 32, no. 9, pp. 1627-1645, September 2010
-
-
int *kPoints,
CvPoint ****partsDisplacement);
-#endif
\ No newline at end of file
+#endif
}
return LATENT_SVM_OK;
-}
\ No newline at end of file
+}
free(x_outTmp);
return FFT_OK;
}
-
return LATENT_SVM_OK;
};
#endif
-
All the 3-channel matrix(i.e. RGB image) are represented by 4-channel matrix in oclMat. It means 3-channel image have 4-channel space with the last channel unused. We provide a transparent interface to handle the difference between OpenCV Mat and oclMat.
-For example: If a oclMat has 3 channels, channels() returns 3 and oclchannels() returns 4
\ No newline at end of file
+For example: If a oclMat has 3 channels, channels() returns 3 and oclchannels() returns 4
* **DESCR_FORMAT_COL_BY_COL** - Column-major order.
-The function is mainly used to learn the classifier.
\ No newline at end of file
+The function is mainly used to learn the classifier.
:param weights2: Weights for second image. Must have tha same size as ``img2`` . Supports only ``CV_32F`` type.
- :param result: Destination image.
\ No newline at end of file
+ :param result: Destination image.
:param m: The Source image of all depth
-Counts the squared sum of matrix elements for each channel.
\ No newline at end of file
+Counts the squared sum of matrix elements for each channel.
.. ocv:function:: void* ocl::getoclCommandQueue()
-Thefunction are used to get opencl command queue so that opencv can interactive with other opencl program.
\ No newline at end of file
+Thefunction are used to get opencl command queue so that opencv can interactive with other opencl program.
TestSystem::instance().run();
return 0;
-}
\ No newline at end of file
+}
}
}
-}
\ No newline at end of file
+}
TestSystem::instance().ExpectedMatNear(dst, ocl_dst, 1.f);
}
}
-}
\ No newline at end of file
+}
else
TestSystem::instance().setAccurate(0, diff);
}
-}
\ No newline at end of file
+}
TestSystem::instance().setAccurate(-1, 0.);
}
-
-
-
-
-
-
-
-
-
GPU_FULL_OFF;
TestSystem::instance().ExceptedMatSimilar(edges, ocl_edges, 2e-2);
-}
\ No newline at end of file
+}
}
-}
\ No newline at end of file
+}
}
}
-}
\ No newline at end of file
+}
TestSystem::instance().ExpectedMatNear(ocl_dst, dst, src1.cols * src1.rows * 1e-4);
}
-}
\ No newline at end of file
+}
faceCascade.detectMultiScale(d_img, oclfaces,
1.1, 2, 0 | CV_HAAR_SCALE_IMAGE, Size(30, 30));
GPU_FULL_OFF;
-}
\ No newline at end of file
+}
d_src.upload(src);
ocl_hog.detectMultiScale(d_src, found_locations);
GPU_FULL_OFF;
-}
\ No newline at end of file
+}
}
}
}
-}
\ No newline at end of file
+}
}
}
-}
\ No newline at end of file
+}
ocl::norm(d_src1, d_src2, NORM_INF);
GPU_FULL_OFF;
}
-}
\ No newline at end of file
+}
matchTemplate(m1, m2, diff, CV_TM_CCORR_NORMED);
return std::abs(diff.at<float>(0, 0) - 1.f);
}
-
-
-
-
TestSystem::instance().ExpectedMatNear(dst, ocl_dst, (src.depth() == CV_32F ? 1e-4f : 1.0));
}
}
-}
\ No newline at end of file
+}
openCLExecuteKernel(ctx, &blend_linear, kernelName, globalSize, localSize, args, channels, depth);
}
-}
\ No newline at end of file
+}
openCLExecuteKernel(clCxt, &imgproc_columnsum, kernelName, globalThreads, localThreads, args, src.channels(), src.depth());
-}
\ No newline at end of file
+}
openCLExecuteKernel(clCxt, &objdetect_hog, kernelName, globalThreads,
localThreads, args, -1, -1);
-}
\ No newline at end of file
+}
}
texture = bindTexture(img);
}
-
return best_compactness;
}
-
}
}
-*/
\ No newline at end of file
+*/
*((__global float *)((__global char *)dst + dst_index)) = tmp;
}
-}
\ No newline at end of file
+}
}//end for(int grploop=grpidx;grploop<totalgrp;grploop+=grpnumx)
}//end for(int scalei = 0; scalei <loopcount; scalei++)
}
-
-
newnode[counter].alpha[1] = t1.alpha[1];
newnode[counter].alpha[2] = t1.alpha[2];
}
-
for (int j=0; j< same; j++)
corners[pos + j] = pt1;
}
-
}
barrier(CLK_LOCAL_MEM_FENCE);
}
-}
\ No newline at end of file
+}
// + m03 ( = m03' + 3*y*m02' + 3*y*y*m01' + y*y*y*m00' )
*(dst_m + mad24(DST_ROW_03 * blocky, dst_step, mad24(wgidy, dst_cols, wgidx))) = mom[9] + y * (3. * mom[5] + y * (3. * mom[2] + ym));
}
-}
\ No newline at end of file
+}
pyrdown_run(src, dst);
}
-
openCLExecuteKernel(clCxt, &pyr_up, kernelName, globalThreads, localThreads, args, src.oclchannels(), src.depth());
}
}
-}
\ No newline at end of file
+}
{
cv::ocl::stereoBM::operator_(minSSD, leBuf, riBuf, preset, ndisp, winSize, avergeTexThreshold, left, right, disparity);
}
-
::StereoBeliefPropagationImpl impl(*this, u, d, l, r, u2, d2, l2, r2, datas, out);
impl(data, disp);
}
-
releaseTexture(I1_tex);
releaseTexture(I1x_tex);
releaseTexture(I1y_tex);
-}
\ No newline at end of file
+}
DIFFERENT_SIZES,
testing::Values(MatType(CV_8UC1), MatType(CV_8UC3), MatType(CV_8UC4), MatType(CV_32FC1), MatType(CV_32FC4))
));
-#endif
\ No newline at end of file
+#endif
INSTANTIATE_TEST_CASE_P(OCL_ImgProc, Canny, testing::Combine(
testing::Values(AppertureSize(3), AppertureSize(5)),
testing::Values(L2gradient(false), L2gradient(true))));
-#endif
\ No newline at end of file
+#endif
Combine(Values(CV_HAAR_SCALE_IMAGE, 0),
Values(cascade_frontalface_alt/*, cascade_frontalface_alt2*/)));
-#endif //HAVE_OPENCL
\ No newline at end of file
+#endif //HAVE_OPENCL
testing::Values(UseInitFlow(false), UseInitFlow(true))));
#endif // HAVE_OPENCL
-
//#include "add_test_info.h"
#endif
-
}
return final_test_result;
}
-
:param hForColorComponents: The same as h but for color components.
The function converts images to CIELAB colorspace and then separately denoise L and AB components with given h parameters using ``fastNlMeansDenoisingMulti`` function.
-
* An example using the inpainting technique can be found at opencv_source_code/samples/cpp/inpaint.cpp
- * (Python) An example using the inpainting technique can be found at opencv_source_code/samples/python2/inpaint.py
\ No newline at end of file
+ * (Python) An example using the inpainting technique can be found at opencv_source_code/samples/python2/inpaint.py
cvtColor(dst_lab, dst, CV_Lab2LBGR);
}
-
-
#include "pyopencv_generated_const_reg.h"
}
-
cv.ShowImage("H-S Histogram", hs_histogram(src))
cv.WaitKey(0)
-
set(the_description "Images stitching")
ocv_define_module(stitching opencv_imgproc opencv_features2d opencv_calib3d opencv_objdetect OPTIONAL opencv_gpu opencv_nonfree)
-
:param pairwise_matches: Matches between all image pairs.
:param focals: Estimated focal lengths for each camera.
-
};
.. seealso:: :ocv:class:`detail::ExposureCompensator`
-
};
.. seealso:: :ocv:class:`detail::SphericalWarper`
-
seam_estimation
exposure_compensation
blenders
-
-
.. ocv:function:: void detail::CylindricalWarper::CylindricalWarper(float scale)
- :param scale: Projected image scale multiplier
\ No newline at end of file
+ :param scale: Projected image scale multiplier
} // namespace detail
} // namespace cv
-
} // namespace detail
} // namespace cv
-
ASSERT_EQ(bad_count, 0);
}
-#endif
\ No newline at end of file
+#endif
{
return new BTVL1_OCL;
}
-#endif
\ No newline at end of file
+#endif
dst[y * dst_step + 4 * x + 3] = dstVal.w;
}
}
-}
\ No newline at end of file
+}
return new FarneBack_OCL;
}
-#endif
\ No newline at end of file
+#endif
if not options.failedOnly:
print '\nOverall time: %.2f min\n' % overall_time
tbl.consolePrintTable(sys.stdout)
- print 2 * '\n'
\ No newline at end of file
+ print 2 * '\n'
tbl.htmlPrintTable(sys.stdout)
htmlPrintFooter(sys.stdout)
else:
- tbl.consolePrintTable(sys.stdout)
\ No newline at end of file
+ tbl.consolePrintTable(sys.stdout)
}// namespace detail
}// namespace cv
-
cv::calcOpticalFlowFarneback( prev, next, flow, pyr_scale, levels,
winsize, iterations, poly_n, poly_sigma, flags );
}
-
set(the_description "Video stabilization")
ocv_define_module(videostab opencv_imgproc opencv_features2d opencv_video opencv_photo opencv_calib3d opencv_highgui OPTIONAL opencv_gpu)
-
install(FILES "${APK_NAME}" DESTINATION "apk/" COMPONENT main)
add_custom_target(android_package ALL SOURCES "${APK_NAME}" )
-add_dependencies(android_package opencv_java)
\ No newline at end of file
+add_dependencies(android_package opencv_java)
:maxdepth: 2
service/doc/index.rst
- java.rst
\ No newline at end of file
+ java.rst
os.system(PushCommand)
print("Testing on device \"%s\"" % Name)
TestCommand = "%s \"%s\" \"%s\" 2>&1" % (os.path.join(HomeDir, "test_native.py"), Arch, Name)
- os.system(TestCommand)
\ No newline at end of file
+ os.system(TestCommand)
armeabi;15c000000581404;
x86;0123456789ABCDEF;
-mips;Novo7 Basic;
\ No newline at end of file
+mips;Novo7 Basic;
Architecture of OpenCV Manager
------------------------------
-.. image:: img/Structure.png
\ No newline at end of file
+.. image:: img/Structure.png
.. data:: INIT_FAILED
- OpenCV library initialization failed
\ No newline at end of file
+ OpenCV library initialization failed
OpenCV Manager and OpenCV library has been already installed.
-.. image:: img/LibInstalled.png
\ No newline at end of file
+.. image:: img/LibInstalled.png
for filename in os.listdir("."):
if ("dia" == filename[-3:]):
- os.system("%s --export %s %s" % (DiaPath, os.path.join(TARGET_PATH, filename[0:len(filename)-4] + ".png"), filename))
\ No newline at end of file
+ os.system("%s --export %s %s" % (DiaPath, os.path.join(TARGET_PATH, filename[0:len(filename)-4] + ".png"), filename))
# Native test application
#---------------------------------------------------------------------
-#include $(LOCAL_PATH)/Tests/Tests.mk
\ No newline at end of file
+#include $(LOCAL_PATH)/Tests/Tests.mk
APP_STL := stlport_static
APP_CPPFLAGS := -fno-rtti -fno-exceptions
NDK_TOOLCHAIN_VERSION=4.4.3
-#APP_OPTIM := debug
\ No newline at end of file
+#APP_OPTIM := debug
}
return android::NO_ERROR;
-}
\ No newline at end of file
+}
};
-#endif
\ No newline at end of file
+#endif
return static_cast<bool>(reply.readInt32());
}
-IMPLEMENT_META_INTERFACE(OpenCVEngine, OPECV_ENGINE_CLASSNAME)
\ No newline at end of file
+IMPLEMENT_META_INTERFACE(OpenCVEngine, OPECV_ENGINE_CLASSNAME)
virtual bool InstallVersion(android::String16 version);
};
-#endif
\ No newline at end of file
+#endif
{
return PLATFORM_UNKNOWN;
}
-}
\ No newline at end of file
+}
std::string GetPlatformName();
int GetCpuID();
-#endif
\ No newline at end of file
+#endif
// public part
std::map<std::string, std::string> GetCpuInfo();
-#endif
\ No newline at end of file
+#endif
}
return result;
-}
\ No newline at end of file
+}
bool ParseString(const std::string& src, std::string& key, std::string& value);
std::vector<std::string> SplitStringVector(const std::string& src, const char seporator);
-#endif
\ No newline at end of file
+#endif
}
return result;
-}
\ No newline at end of file
+}
int DetectTegra();
-#endif
\ No newline at end of file
+#endif
JNIEXPORT jint JNICALL Java_org_opencv_engine_HardwareDetector_DetectKnownPlatforms(JNIEnv* , jclass)
{
return DetectKnownPlatforms();
-}
\ No newline at end of file
+}
printf("OpenCVEngine version %d started", EngineVersion);
return 0;
-}
\ No newline at end of file
+}
NativePackageManager::~NativePackageManager()
{
-}
\ No newline at end of file
+}
virtual std::vector<PackageInfo> GetInstalledPackages();
};
-#endif
\ No newline at end of file
+#endif
delete PackageManager;
return 0;
-}
\ No newline at end of file
+}
PackageManagerStub::~PackageManagerStub()
{
-}
\ No newline at end of file
+}
virtual std::vector<PackageInfo> GetInstalledPackages();
};
-#endif
\ No newline at end of file
+#endif
{
::testing::InitGoogleTest(&argc, argv);
return RUN_ALL_TESTS();
-}
\ No newline at end of file
+}
LOCAL_SHARED_LIBRARIES += libOpenCVEngine
-include $(BUILD_EXECUTABLE)
\ No newline at end of file
+include $(BUILD_EXECUTABLE)
typedef const char* (*InfoFunctionType)();
-#endif
\ No newline at end of file
+#endif
virtual bool InstallVersion(android::String16 version) = 0;
};
-#endif
\ No newline at end of file
+#endif
* @return Returns OpenCV libraries names seporated by symbol ";" in loading order
*/
String getLibraryList(String version);
-}
\ No newline at end of file
+}
os.system("adb %s push ./engine/libs/%s/OpenCVEngineNativeService %s" % (DEVICE_STR, DEVICE_ARCH, TARGET_DEVICE_PATH))
os.system("adb %s push ./engine/libs/%s/OpenCVEngineTest %s" % (DEVICE_STR, DEVICE_ARCH, TARGET_DEVICE_PATH))
os.system("adb %s push ./engine/libs/%s/OpenCVEngineTestApp %s" % (DEVICE_STR, DEVICE_ARCH, TARGET_DEVICE_PATH))
- os.system("adb %s push ./engine/libs/%s/libOpenCVEngine_jni.so %s" % (DEVICE_STR, DEVICE_ARCH, TARGET_DEVICE_PATH))
\ No newline at end of file
+ os.system("adb %s push ./engine/libs/%s/libOpenCVEngine_jni.so %s" % (DEVICE_STR, DEVICE_ARCH, TARGET_DEVICE_PATH))
print "Usage:\n\t./build_framework.py <outputdir>\n\n"
sys.exit(0)
- build_framework(os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), "../..")), os.path.abspath(sys.argv[1]))
\ No newline at end of file
+ build_framework(os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), "../..")), os.path.abspath(sys.argv[1]))
# Project install destination.
"${CMAKE_INSTALL_PREFIX}"
-)
\ No newline at end of file
+)
SET(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)
SET(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)
-message (STATUS "iPhoneOS toolchain loaded")
\ No newline at end of file
+message (STATUS "iPhoneOS toolchain loaded")
SET(CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY)
SET(CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY)
-message (STATUS "iPhoneSimulator toolchain loaded")
\ No newline at end of file
+message (STATUS "iPhoneSimulator toolchain loaded")
cd ~/<my_working_directory>
python opencv/platforms/ios/build_framework.py ios
-If everything's fine, a few minutes later you will get ~/<my_working_directory>/ios/opencv2.framework. You can add this framework to your Xcode projects.
\ No newline at end of file
+If everything's fine, a few minutes later you will get ~/<my_working_directory>/ios/opencv2.framework. You can add this framework to your Xcode projects.
This folder contains toolchains and additional files that are needed for cross compilation.
For more information see introduction tutorials for target platform in documentation:
-http://docs.opencv.org/doc/tutorials/introduction/table_of_content_introduction/table_of_content_introduction.html#table-of-content-introduction
\ No newline at end of file
+http://docs.opencv.org/doc/tutorials/introduction/table_of_content_introduction/table_of_content_introduction.html#table-of-content-introduction
call "%msvc_path%\Common7\Tools\VsDevCmd.bat"
call "%msvc_path%\VC\bin\x86_arm\vcvarsx86_arm.bat"
-cmake.exe -GNinja -DCMAKE_BUILD_TYPE=Release -DENABLE_WINRT_MODE=ON -DWITH_FFMPEG=OFF -DWITH_MSMF=OFF -DWITH_DSHOW=OFF -DWITH_VFW=OFF -DWITH_TIFF=OFF -DWITH_OPENEXR=OFF -DWITH_CUDA=OFF -DBUILD_opencv_gpu=OFF -DBUILD_opencv_python=OFF -DBUILD_opencv_java=OFF -DCMAKE_TOOLCHAIN_FILE=..\winrt\arm.winrt.toolchain.cmake %* ..\..
\ No newline at end of file
+cmake.exe -GNinja -DCMAKE_BUILD_TYPE=Release -DENABLE_WINRT_MODE=ON -DWITH_FFMPEG=OFF -DWITH_MSMF=OFF -DWITH_DSHOW=OFF -DWITH_VFW=OFF -DWITH_TIFF=OFF -DWITH_OPENEXR=OFF -DWITH_CUDA=OFF -DBUILD_opencv_gpu=OFF -DBUILD_opencv_python=OFF -DBUILD_opencv_java=OFF -DCMAKE_TOOLCHAIN_FILE=..\winrt\arm.winrt.toolchain.cmake %* ..\..
set(CMAKE_FIND_ROOT_PATH "${CMAKE_SOURCE_DIR}/platforms/winrt")
set(CMAKE_REQUIRED_DEFINITIONS -D_ARM_WINAPI_PARTITION_DESKTOP_SDK_AVAILABLE)
-add_definitions(-D_ARM_WINAPI_PARTITION_DESKTOP_SDK_AVAILABLE)
\ No newline at end of file
+add_definitions(-D_ARM_WINAPI_PARTITION_DESKTOP_SDK_AVAILABLE)
local.properties
proguard-project.txt
project.properties
-default.properties
\ No newline at end of file
+default.properties
if(TARGET ${sample})
add_dependencies(opencv_android_examples ${sample})
endif()
-
return new Scalar(pointMatRgba.get(0, 0));
}
-}
\ No newline at end of file
+}
if(TARGET ${sample})
add_dependencies(opencv_android_examples ${sample})
endif()
-
LOCAL_MODULE := detection_based_tracker
-include $(BUILD_SHARED_LIBRARY)
\ No newline at end of file
+include $(BUILD_SHARED_LIBRARY)
jenv->ThrowNew(je, "Unknown exception in JNI code {highgui::VideoCapture_n_1VideoCapture__()}");
}
LOGD("Java_org_opencv_samples_facedetect_DetectionBasedTracker_nativeDetect exit");
-}
\ No newline at end of file
+}
super.onResume();
OpenCVLoader.initAsync(OpenCVLoader.OPENCV_VERSION_2_4_3, this, mLoaderCallback);
}
-}
\ No newline at end of file
+}
if(TARGET ${sample})
add_dependencies(opencv_android_examples ${sample})
endif()
-
if(TARGET ${sample})
add_dependencies(opencv_android_examples ${sample})
endif()
-
}
}
-}
\ No newline at end of file
+}
DESTINATION share/OpenCV/samples/c
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ)
endif ()
-
return 0;
} /* main() */
-
-
-
TARGET_LINK_LIBRARIES(opencv_example ${OpenCV_LIBS})
#MESSAGE(STATUS "OpenCV_LIBS: ${OpenCV_LIBS}")
-
-
-
$ cmake-gui <OPENCV_SRC_PATH>/samples/c/example_cmake/
And pick the correct value for OpenCV_DIR.
-
-
-
-
-
#ifdef _EiC
main(1,"convexhull.c");
#endif
-
DESTINATION share/OpenCV/samples/cpp
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ)
endif()
-
return generageChessBoard(bg, camMat, distCoeffs, zero, pb1, pb2, sqWidth, sqHeight, pts3d, corners);
}
-
}
return 0;
-}
\ No newline at end of file
+}
StereoCalib(imagelist, boardSize, true, showRectified);
return 0;
}
-
LOGLN("Finished, total time: " << ((getTickCount() - app_start_time) / getTickFrequency()) << " sec");
return 0;
}
-
-
morphologyEx( src, dst, operation, element );
imshow( window_name, dst );
}
-
-
return 0;
}
-
-
-
-
-
-
-
circle( drawing, mc[i], 4, color, -1, 8, 0 );
}
}
-
waitKey(0);
return(0);
}
-
-
}
imshow( myHarris_window, myHarris_copy );
}
-
for( size_t i = 0; i < corners.size(); i++ )
{ cout<<" -- Refined Corner ["<<i<<"] ("<<corners[i].x<<","<<corners[i].y<<")"<<endl; }
}
-
namedWindow( source_window, CV_WINDOW_AUTOSIZE );
imshow( source_window, copy );
}
-
thickness,
lineType );
}
-
-
waitKey();
return 0;
-}
\ No newline at end of file
+}
<< "Tip: Open up " << filename << " with a text editor to see the serialized data." << endl;
return 0;
-}
\ No newline at end of file
+}
}
return I;
-}
\ No newline at end of file
+}
Result.row(Result.rows-1).setTo(Scalar(0));
Result.col(0).setTo(Scalar(0));
Result.col(Result.cols-1).setTo(Scalar(0));
-}
\ No newline at end of file
+}
cout << "A vector of 2D Points = " << vPoints << endl << endl;
return 0;
-}
\ No newline at end of file
+}
waitKey(0); // Wait for a keystroke in the window
return 0;
-}
\ No newline at end of file
+}
Scalar mssim = mean( ssim_map ); // mssim = average of ssim map
return mssim;
-}
\ No newline at end of file
+}
imwrite("result.png", I); // save the Image
imshow("SVM for Non-Linear Training Data", I); // show it to the user
waitKey(0);
-}
\ No newline at end of file
+}
DESTINATION share/OpenCV/samples/${project}
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ)
endif()
-
row[j][3] = 255;
}
}
-}
\ No newline at end of file
+}
ss << work_fps;
return ss.str();
}
-
System.out.println(String.format("Writing %s", filename));
Highgui.imwrite(filename, image);
}
-}
\ No newline at end of file
+}
println(s"Writing ${filename}")
assert(Highgui.imwrite(filename, correspondenceImage))
}
-}
\ No newline at end of file
+}
println(s"Writing ${filename}")
assert(Highgui.imwrite(filename, image))
}
-}
\ No newline at end of file
+}
DESTINATION share/opencv/samples/${project}
PERMISSIONS OWNER_READ GROUP_READ WORLD_READ)
endif()
-
}
return final_test_result;
}
-
waitKey();
return 0;
-}
\ No newline at end of file
+}
cv.AddWeighted(wshed, 0.5, img_gray, 0.5, 0, wshed)
cv.ShowImage("watershed transform", wshed)
cv.DestroyAllWindows()
-
match_and_draw('affine find_obj')
cv2.waitKey()
cv2.destroyAllWindows()
-
print "camera matrix:\n", camera_matrix
print "distortion coefficients: ", dist_coefs.ravel()
cv2.destroyAllWindows()
-
except: video_src = 0
print __doc__
App(video_src).run()
-
for kp in keypoints:
x, y = kp.pt
cv2.circle(vis, (int(x), int(y)), 2, color)
-
if ch == ord(' '):
defocus = not defocus
update(None)
-
if __name__ == '__main__':
App().run()
-
if need_update:
update()
cv2.destroyAllWindows()
-
if ch == 27:
break
cv2.destroyAllWindows()
-
if 0xFF & cv2.waitKey(5) == 27:
break
cv2.destroyAllWindows()
-
cv2.destroyAllWindows()
break
cv2.destroyAllWindows()
-
mark[:] = 0
sketch.show()
cv2.destroyAllWindows()
-
if cv2.waitKey(1) == 27:
break
-
-
cur_glitch = img.copy()
print 'glitch is', ['off', 'on'][show_glitch]
cv2.destroyAllWindows()
-
<Import Project="$(VCTargetsPath)\Microsoft.Cpp.targets" />
<ImportGroup Label="ExtensionTargets">
</ImportGroup>
-</Project>
\ No newline at end of file
+</Project>
IMFMediaBuffer *m_pBuffer;
IMF2DBuffer *m_p2DBuffer;
};
-
return hr;
}
-#pragma warning( pop )
\ No newline at end of file
+#pragma warning( pop )
runtimeclass OcvImageManipulations
{
}
-}
\ No newline at end of file
+}
EXPORTS
DllCanUnloadNow PRIVATE
DllGetActivationFactory PRIVATE
- DllGetClassObject PRIVATE
\ No newline at end of file
+ DllGetClassObject PRIVATE
IMFAttributes *m_pAttributes;
};
-#endif
\ No newline at end of file
+#endif
</Target>
<ImportGroup Label="ExtensionTargets">
</ImportGroup>
-</Project>
\ No newline at end of file
+</Project>
</InProcessServer>
</Extension>
</Extensions>
-</Package>
\ No newline at end of file
+</Package>