diff --git a/doc/py_tutorials/py_core/py_basic_ops/py_basic_ops.markdown b/doc/py_tutorials/py_core/py_basic_ops/py_basic_ops.markdown index a9b4905022..19992fc95b 100644 --- a/doc/py_tutorials/py_core/py_basic_ops/py_basic_ops.markdown +++ b/doc/py_tutorials/py_core/py_basic_ops/py_basic_ops.markdown @@ -173,7 +173,7 @@ from matplotlib import pyplot as plt BLUE = [255,0,0] -img1 = cv2.imread('opencv_logo.png') +img1 = cv2.imread('opencv-logo.png') replicate = cv2.copyMakeBorder(img1,10,10,10,10,cv2.BORDER_REPLICATE) reflect = cv2.copyMakeBorder(img1,10,10,10,10,cv2.BORDER_REFLECT) diff --git a/doc/py_tutorials/py_core/py_image_arithmetics/py_image_arithmetics.markdown b/doc/py_tutorials/py_core/py_image_arithmetics/py_image_arithmetics.markdown index 609e610959..8995c93422 100644 --- a/doc/py_tutorials/py_core/py_image_arithmetics/py_image_arithmetics.markdown +++ b/doc/py_tutorials/py_core/py_image_arithmetics/py_image_arithmetics.markdown @@ -51,7 +51,7 @@ is given 0.3. cv2.addWeighted() applies following equation on the image. Here \f$\gamma\f$ is taken as zero. @code{.py} img1 = cv2.imread('ml.png') -img2 = cv2.imread('opencv_logo.jpg') +img2 = cv2.imread('opencv-logo.png') dst = cv2.addWeighted(img1,0.7,img2,0.3,0) @@ -77,7 +77,7 @@ bitwise operations as below: @code{.py} # Load two images img1 = cv2.imread('messi5.jpg') -img2 = cv2.imread('opencv_logo.png') +img2 = cv2.imread('opencv-logo.png') # I want to put logo on top-left corner, So I create a ROI rows,cols,channels = img2.shape diff --git a/doc/py_tutorials/py_imgproc/py_filtering/py_filtering.markdown b/doc/py_tutorials/py_imgproc/py_filtering/py_filtering.markdown index 95bfef213f..bc3e87bc57 100644 --- a/doc/py_tutorials/py_imgproc/py_filtering/py_filtering.markdown +++ b/doc/py_tutorials/py_imgproc/py_filtering/py_filtering.markdown @@ -69,7 +69,7 @@ import cv2 import numpy as np from matplotlib import pyplot as plt -img = cv2.imread('opencv_logo.png') +img = cv2.imread('opencv-logo-white.png') blur = cv2.blur(img,(5,5)) diff --git a/doc/py_tutorials/py_imgproc/py_geometric_transformations/py_geometric_transformations.markdown b/doc/py_tutorials/py_imgproc/py_geometric_transformations/py_geometric_transformations.markdown index bce33743d4..a71878f820 100644 --- a/doc/py_tutorials/py_imgproc/py_geometric_transformations/py_geometric_transformations.markdown +++ b/doc/py_tutorials/py_imgproc/py_geometric_transformations/py_geometric_transformations.markdown @@ -135,7 +135,7 @@ matrix. See the code below: @code{.py} -img = cv2.imread('sudokusmall.png') +img = cv2.imread('sudoku.png') rows,cols,ch = img.shape pts1 = np.float32([[56,65],[368,52],[28,387],[389,390]]) diff --git a/doc/py_tutorials/py_imgproc/py_houghcircles/py_houghcircles.markdown b/doc/py_tutorials/py_imgproc/py_houghcircles/py_houghcircles.markdown index 159e47953d..50fc05c235 100644 --- a/doc/py_tutorials/py_imgproc/py_houghcircles/py_houghcircles.markdown +++ b/doc/py_tutorials/py_imgproc/py_houghcircles/py_houghcircles.markdown @@ -23,7 +23,7 @@ explained in the documentation. So we directly go to the code. import cv2 import numpy as np -img = cv2.imread('opencv_logo.png',0) +img = cv2.imread('opencv-logo-white.png',0) img = cv2.medianBlur(img,5) cimg = cv2.cvtColor(img,cv2.COLOR_GRAY2BGR) diff --git a/doc/py_tutorials/py_imgproc/py_houghlines/py_houghlines.markdown b/doc/py_tutorials/py_imgproc/py_houghlines/py_houghlines.markdown index 881cf2a29e..5b569ed074 100644 --- a/doc/py_tutorials/py_imgproc/py_houghlines/py_houghlines.markdown +++ b/doc/py_tutorials/py_imgproc/py_houghlines/py_houghlines.markdown @@ -73,7 +73,7 @@ represents the minimum length of line that should be detected. import cv2 import numpy as np -img = cv2.imread('dave.jpg') +img = cv2.imread('sudoku.png') gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY) edges = cv2.Canny(gray,50,150,apertureSize = 3) @@ -121,7 +121,7 @@ the parameters of lines, and you had to find all the points. Here, everything is import cv2 import numpy as np -img = cv2.imread('dave.jpg') +img = cv2.imread('sudoku.png') gray = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY) edges = cv2.Canny(gray,50,150,apertureSize = 3) lines = cv2.HoughLinesP(edges,1,np.pi/180,100,minLineLength=100,maxLineGap=10) diff --git a/doc/py_tutorials/py_imgproc/py_thresholding/py_thresholding.markdown b/doc/py_tutorials/py_imgproc/py_thresholding/py_thresholding.markdown index b9cc0296a8..5fc6240238 100644 --- a/doc/py_tutorials/py_imgproc/py_thresholding/py_thresholding.markdown +++ b/doc/py_tutorials/py_imgproc/py_thresholding/py_thresholding.markdown @@ -87,7 +87,7 @@ import cv2 import numpy as np from matplotlib import pyplot as plt -img = cv2.imread('dave.jpg',0) +img = cv2.imread('sudoku.png',0) img = cv2.medianBlur(img,5) ret,th1 = cv2.threshold(img,127,255,cv2.THRESH_BINARY) diff --git a/samples/cpp/peopledetect.cpp b/samples/cpp/peopledetect.cpp index aca05fd0ca..5fa2382f0f 100644 --- a/samples/cpp/peopledetect.cpp +++ b/samples/cpp/peopledetect.cpp @@ -16,7 +16,7 @@ const char* keys = "{ help h | | print help message }" "{ image i | | specify input image}" "{ camera c | | enable camera capturing }" - "{ video v | ../data/768x576.avi | use video as input }" + "{ video v | ../data/vtest.avi | use video as input }" "{ directory d | | images directory}" }; @@ -79,7 +79,7 @@ int main(int argc, char** argv) namedWindow("people detector", 1); string pattern_glob = ""; - string video_filename = "../data/768x576.avi"; + string video_filename = "../data/vtest.avi"; int camera_id = -1; if (parser.has("directory")) { diff --git a/samples/data/apple.jpg b/samples/data/apple.jpg new file mode 100644 index 0000000000..2ff27ee0e1 Binary files /dev/null and b/samples/data/apple.jpg differ diff --git a/samples/data/gradient.png b/samples/data/gradient.png new file mode 100644 index 0000000000..255e77f945 Binary files /dev/null and b/samples/data/gradient.png differ diff --git a/samples/data/ml.png b/samples/data/ml.png new file mode 100644 index 0000000000..6420b44b2c Binary files /dev/null and b/samples/data/ml.png differ diff --git a/samples/data/opencv-logo-white.png b/samples/data/opencv-logo-white.png new file mode 100644 index 0000000000..3c7098459e Binary files /dev/null and b/samples/data/opencv-logo-white.png differ diff --git a/samples/data/opencv-logo.png b/samples/data/opencv-logo.png index 8cd8155c1a..bc71a2ae50 100644 Binary files a/samples/data/opencv-logo.png and b/samples/data/opencv-logo.png differ diff --git a/samples/data/orange.jpg b/samples/data/orange.jpg new file mode 100644 index 0000000000..773a846d9f Binary files /dev/null and b/samples/data/orange.jpg differ diff --git a/samples/data/sudoku.png b/samples/data/sudoku.png new file mode 100644 index 0000000000..763e518e2d Binary files /dev/null and b/samples/data/sudoku.png differ diff --git a/samples/data/768x576.avi b/samples/data/vtest.avi similarity index 100% rename from samples/data/768x576.avi rename to samples/data/vtest.avi diff --git a/samples/gpu/bgfg_segm.cpp b/samples/gpu/bgfg_segm.cpp index 00bb59e248..5383db7f0f 100644 --- a/samples/gpu/bgfg_segm.cpp +++ b/samples/gpu/bgfg_segm.cpp @@ -23,10 +23,10 @@ enum Method int main(int argc, const char** argv) { cv::CommandLineParser cmd(argc, argv, - "{ c camera | | use camera }" - "{ f file | ../data/768x576.avi | input video file }" - "{ m method | mog | method (mog, mog2, gmg, fgd) }" - "{ h help | | print help message }"); + "{ c camera | | use camera }" + "{ f file | ../data/vtest.avi | input video file }" + "{ m method | mog | method (mog, mog2, gmg, fgd) }" + "{ h help | | print help message }"); if (cmd.has("help") || !cmd.check()) { diff --git a/samples/gpu/performance/tests.cpp b/samples/gpu/performance/tests.cpp index b4bf4cfbec..aad40bf09f 100644 --- a/samples/gpu/performance/tests.cpp +++ b/samples/gpu/performance/tests.cpp @@ -1191,10 +1191,10 @@ TEST(GoodFeaturesToTrack) TEST(MOG) { - const std::string inputFile = abspath("../data/768x576.avi"); + const std::string inputFile = abspath("../data/vtest.avi"); cv::VideoCapture cap(inputFile); - if (!cap.isOpened()) throw runtime_error("can't open ../data/768x576.avi"); + if (!cap.isOpened()) throw runtime_error("can't open ../data/vtest.avi"); cv::Mat frame; cap >> frame; diff --git a/samples/tapi/bgfg_segm.cpp b/samples/tapi/bgfg_segm.cpp index 8b944d184a..ff097dc208 100644 --- a/samples/tapi/bgfg_segm.cpp +++ b/samples/tapi/bgfg_segm.cpp @@ -17,11 +17,11 @@ using namespace cv; int main(int argc, const char** argv) { CommandLineParser cmd(argc, argv, - "{ c camera | | use camera }" - "{ f file | ../data/768x576.avi | input video file }" - "{ t type | mog2 | method's type (knn, mog2) }" - "{ h help | | print help message }" - "{ m cpu_mode | false | press 'm' to switch OpenCL<->CPU}"); + "{ c camera | | use camera }" + "{ f file | ../data/vtest.avi | input video file }" + "{ t type | mog2 | method's type (knn, mog2) }" + "{ h help | | print help message }" + "{ m cpu_mode | false | press 'm' to switch OpenCL<->CPU}"); if (cmd.has("help")) { diff --git a/samples/tapi/hog.cpp b/samples/tapi/hog.cpp index db31396106..1ead0d947f 100644 --- a/samples/tapi/hog.cpp +++ b/samples/tapi/hog.cpp @@ -71,7 +71,7 @@ int main(int argc, char** argv) "{ h help | | print help message }" "{ i input | | specify input image}" "{ c camera | -1 | enable camera capturing }" - "{ v video | ../data/768x576.avi | use video as input }" + "{ v video | ../data/vtest.avi | use video as input }" "{ g gray | | convert image to gray one or not}" "{ s scale | 1.0 | resize the image before detect}" "{ o output | | specify output path when input is images}";