SamFC10
6111935835
Added exp layer
4 years ago
Tsukasa Sugiura
107f233626
Merge pull request #19484 from UnaNancyOwen:fix_highlevelapi
...
* [dnn] fix high level api for python
* [dnn] add test_textdetection_model_db
* [dnn] fix textdetection test only check type and shape
4 years ago
Alexander Alekhin
83aa711346
dnn: rename clamp() => normalize_axis()
4 years ago
Alexander Alekhin
d1b5a78171
build warnings
...
- GCC 4.8.5 / CentOS 7
4 years ago
Wenqing Zhang
22d64ae08f
Merge pull request #17570 from HannibalAPE:text_det_recog_demo
...
[GSoC] High Level API and Samples for Scene Text Detection and Recognition
* APIs and samples for scene text detection and recognition
* update APIs and tutorial for Text Detection and Recognition
* API updates:
(1) put decodeType into struct Voc
(2) optimize the post-processing of DB
* sample update:
(1) add transformation into scene_text_spotting.cpp
(2) modify text_detection.cpp with API update
* update tutorial
* simplify text recognition API
update tutorial
* update impl usage in recognize() and detect()
* dnn: refactoring public API of TextRecognitionModel/TextDetectionModel
* update provided models
update opencv.bib
* dnn: adjust text rectangle angle
* remove points ordering operation in model.cpp
* update gts of DB test in test_model.cpp
* dnn: ensure to keep text rectangle angle
- avoid 90/180 degree turns
* dnn(text): use quadrangle result in TextDetectionModel API
* dnn: update Text Detection API
(1) keep points' order consistent with (bl, tl, tr, br) in unclip
(2) update contourScore with boundingRect
4 years ago
Daniel Cauchi
9d37cdaa66
Merge pull request #18891 from CowKeyMan:NMS_boxes_with_different_labels
...
Add option for NMS for boxes with different labels
* DetectionModel impl
* Add option for NMS for boxes with different labels
In the detect function in modules/dnn/include/opencv2/dnn/dnn.hpp, whose implementation can be found at modules/dnn/src/model.cpp, the Non Max Suppression (NMS) is applied only for objects of the same label. Thus, a flag
was added with the purpose to allow developers to choose if they want to keep the default implementation or wether they would like NMS to be applied to all the boxes, regardless of label.
The flag is called nmsDifferentLabels, and is given a default value of false, which applies the current default implementation, thus allowing existing projects to update opencv without disruption
Solves issue opencv#18832
* Change return type of set & Add default constr
* Add assertions due to default constructor
4 years ago
Sergei Slashchinin
f4f462c50b
Merge pull request #18862 from sl-sergei:support_pool1d
...
Support for Pool1d layer for OpenCV and OpenCL targets
* Initial version of Pool1d support
* Fix variable naming
* Fix 1d pooling for OpenCL
* Change support logic, remove unnecessary variable, split the tests
* Remove other depricated variables
* Fix warning. Check tests
* Change support check logic
* Change support check logic, 2
4 years ago
Alexander Alekhin
9485113923
pre: OpenCV 3.4.13 (version++)
4 years ago
Omar Alzaibaq
a316b11aaa
Merge pull request #18220 from Omar-AE:hddl-supported
...
* added HDDL VPU support
* changed to return True in one line if any device connected
* dnn: use releaseHDDLPlugin()
* dnn(hddl): fix conditions
4 years ago
Alexander Alekhin
23baf1a75e
dnn: fix High-Level public API (cv::dnn::Model class)
...
- proxy selected Net methods only (don't derive from Net directly)
- default Model ctor is protected
4 years ago
Sergey Slashchinin
32e7ef8a3d
Add fixes and tests for different layers
4 years ago
Alexander Alekhin
a12ceb04bb
pre: OpenCV 4.5.0 (version++)
4 years ago
Alexander Alekhin
50ff40d684
pre: OpenCV 3.4.12 (version++)
4 years ago
kadi soheib
6bed5c181b
Corrected Comment as requested by reviewer.
4 years ago
kadi soheib
17c430da88
Updated comment.
4 years ago
kadi soheib
96a501c08b
Adding comment from source code to documentation.
4 years ago
Alexander Alekhin
5f3012fc9a
pre: OpenCV 4.4.0 (version++)
5 years ago
Alexander Alekhin
a43e3bebe6
pre: OpenCV 3.4.11 (version++)
5 years ago
Liubov Batanina
d3aaf2d3a3
Merge pull request #17371 from l-bat:nms_model
...
* Fix NMS bug in DetectionModel
* Fixed comments
* Refactoring
5 years ago
Liubov Batanina
d991c22090
Merge pull request #16575 from l-bat:flownet2
...
Support FlowNet2 model
* Support DataAugmentation layer
* Fix warnings
* Fix comments
* Support Correlation layer
* TEST
* Support Correlation layer
* Supported Accum and FlowWarp layers
* Supported ChannelNorm layer
* Supported Resample with inputs.size() > 1
* Fixed comments
* Refactoring
* Added tests
* Add resample test
* Added asserts in resize layer
* Updated DataAugmentation layer
* Update convolution layer
* Refactoring
* Fix data augmentation layer
* Fix caffe importer
* Fix resize
* Switch to Mat ptr
* Remove useless resize type
* Used ResizeLayer in Accum
* Split ChannelNormLayer
* Delete duplicate assert
* Add sample
* Fix sample
* Added colormap
5 years ago
Pavel Rojtberg
66cf55ea1f
dnn: expose only float variant of NMSBoxes for bindings
...
the float variant was always shadowed by the int version as
Rect2d is implicitly convertible to Rect.
This swaps things which is fine, as the vector of boxes was always
copied and the computation was done in double.
5 years ago
Alexander Alekhin
db95aec4a7
dnn(ie): switch to nGraph backend by default
5 years ago
Alexander Alekhin
01048e5603
Merge pull request #16616 from alalek:dnn_fix_input_shape
...
* dnn: fix processing of input shapes
- importer: avoid using of .setInput() => .setInputShape()
- setInput: shape limitation check (partial)
* dnn(test): test .setInput() in readNet()
5 years ago
Liubov Batanina
4a19ac5aca
Move instruction
5 years ago
Liubov Batanina
7e5b5390ba
Fix comments
5 years ago
Liubov Batanina
832ca0734d
Refactoring
5 years ago
Julien
886220b9be
Merge pull request #16273 from JulienMaille:wrapper_available_target
...
* add a wrapper for getAvailableTargets
* add java wrapper on Target enum
5 years ago
Liubov Batanina
7eba3a7c96
Add pack description
5 years ago
Liubov Batanina
752653c70b
Update global pooling
5 years ago
Brian Wignall
f9c514b391
Fix spelling typos
...
backport commit 659ffaddb4
5 years ago
Brian Wignall
659ffaddb4
Fix spelling typos
5 years ago
Liubov Batanina
543e0302d3
Support global pooling by axis
5 years ago
antalzsiroscandid
aa80f754f4
dnn: reading IR models from buffer
5 years ago
Diego
5b0b59ecfb
Merge pull request #15189 from dvd42:keypoints_module
...
Keypoints module
5 years ago
Alexander Alekhin
5ee7abbe3c
Merge pull request #16088 from alalek:dnn_eltwise_layer_different_src_channels
...
dnn(eltwise): fix handling of different number of channels
* dnn(test): reproducer for Eltwise layer issue from PR16063
* dnn(eltwise): rework support for inputs with different channels
* dnn(eltwise): get rid of finalize(), variableChannels
* dnn(eltwise): update input sorting by number of channels
- do not swap inputs if number of channels are same after truncation
* dnn(test): skip "shortcut" with batch size 2 on MYRIAD targets
5 years ago
Lubov Batanina
7523c777c5
Merge pull request #15537 from l-bat:ngraph
...
* Support nGraph
* Fix resize
5 years ago
Manjunath Bhat
78c5e41c23
Merge pull request #15808 from thebhatman:Mish_swish
...
* Added Swish and Mish activations
* Fixed whitespace errors
* Kernel implementation done
* Added function for launching kernel
* Changed type of 1.0
* Attempt to add test for Swish and Mish
* Resolving type mismatch for log
* exp from device
* Use log1pexp instead of adding 1
* Added openCL kernels
5 years ago
thebhatman
8a18d132fc
Port Swish and Mish layers
5 years ago
Yashas Samaga B L
613c12e590
Merge pull request #14827 from YashasSamaga:cuda4dnn-csl-low
...
CUDA backend for the DNN module
* stub cuda4dnn design
* minor fixes for tests and doxygen
* add csl public api directory to module headers
* add low-level CSL components
* add high-level CSL components
* integrate csl::Tensor into backbone code
* switch to CPU iff unsupported; otherwise, fail on error
* add fully connected layer
* add softmax layer
* add activation layers
* support arbitary rank TensorDescriptor
* pass input wrappers to `initCUDA()`
* add 1d/2d/3d-convolution
* add pooling layer
* reorganize and refactor code
* fixes for gcc, clang and doxygen; remove cxx14/17 code
* add blank_layer
* add LRN layer
* add rounding modes for pooling layer
* split tensor.hpp into tensor.hpp and tensor_ops.hpp
* add concat layer
* add scale layer
* add batch normalization layer
* split math.cu into activations.cu and math.hpp
* add eltwise layer
* add flatten layer
* add tensor transform api
* add asymmetric padding support for convolution layer
* add reshape layer
* fix rebase issues
* add permute layer
* add padding support for concat layer
* refactor and reorganize code
* add normalize layer
* optimize bias addition in scale layer
* add prior box layer
* fix and optimize normalize layer
* add asymmetric padding support for pooling layer
* add event API
* improve pooling performance for some padding scenarios
* avoid over-allocation of compute resources to kernels
* improve prior box performance
* enable layer fusion
* add const layer
* add resize layer
* add slice layer
* add padding layer
* add deconvolution layer
* fix channelwise ReLU initialization
* add vector traits
* add vectorized versions of relu, clipped_relu, power
* add vectorized concat kernels
* improve concat_with_offsets performance
* vectorize scale and bias kernels
* add support for multi-billion element tensors
* vectorize prior box kernels
* fix address alignment check
* improve bias addition performance of conv/deconv/fc layers
* restructure code for supporting multiple targets
* add DNN_TARGET_CUDA_FP64
* add DNN_TARGET_FP16
* improve vectorization
* add region layer
* improve tensor API, add dynamic ranks
1. use ManagedPtr instead of a Tensor in backend wrapper
2. add new methods to tensor classes
- size_range: computes the combined size of for a given axis range
- tensor span/view can be constructed from a raw pointer and shape
3. the tensor classes can change their rank at runtime (previously rank was fixed at compile-time)
4. remove device code from tensor classes (as they are unused)
5. enforce strict conditions on tensor class APIs to improve debugging ability
* fix parametric relu activation
* add squeeze/unsqueeze tensor API
* add reorg layer
* optimize permute and enable 2d permute
* enable 1d and 2d slice
* add split layer
* add shuffle channel layer
* allow tensors of different ranks in reshape primitive
* patch SliceOp to allow Crop Layer
* allow extra shape inputs in reshape layer
* use `std::move_backward` instead of `std::move` for insert in resizable_static_array
* improve workspace management
* add spatial LRN
* add nms (cpu) to region layer
* add max pooling with argmax ( and a fix to limits.hpp)
* add max unpooling layer
* rename DNN_TARGET_CUDA_FP32 to DNN_TARGET_CUDA
* update supportBackend to be more rigorous
* remove stray include from preventing non-cuda build
* include op_cuda.hpp outside condition #if
* refactoring, fixes and many optimizations
* drop DNN_TARGET_CUDA_FP64
* fix gcc errors
* increase max. tensor rank limit to six
* add Interp layer
* drop custom layers; use BackendNode
* vectorize activation kernels
* fixes for gcc
* remove wrong assertion
* fix broken assertion in unpooling primitive
* fix build errors in non-CUDA build
* completely remove workspace from public API
* fix permute layer
* enable accuracy and perf. tests for DNN_TARGET_CUDA
* add asynchronous forward
* vectorize eltwise ops
* vectorize fill kernel
* fixes for gcc
* remove CSL headers from public API
* remove csl header source group from cmake
* update min. cudnn version in cmake
* add numerically stable FP32 log1pexp
* refactor code
* add FP16 specialization to cudnn based tensor addition
* vectorize scale1 and bias1 + minor refactoring
* fix doxygen build
* fix invalid alignment assertion
* clear backend wrappers before allocateLayers
* ignore memory lock failures
* do not allocate internal blobs
* integrate NVTX
* add numerically stable half precision log1pexp
* fix indentation, following coding style, improve docs
* remove accidental modification of IE code
* Revert "add asynchronous forward"
This reverts commit 1154b9da9da07e9b52f8a81bdcea48cf31c56f70.
* [cmake] throw error for unsupported CC versions
* fix rebase issues
* add more docs, refactor code, fix bugs
* minor refactoring and fixes
* resolve warnings/errors from clang
* remove haveCUDA() checks from supportBackend()
* remove NVTX integration
* changes based on review comments
* avoid exception when no CUDA device is present
* add color code for CUDA in Net::dump
5 years ago
Alexander Alekhin
48d41ab088
dnn: bump API version
5 years ago
Alexander Alekhin
70dfae31a2
experimental version++
5 years ago
luz.paz
fcc7d8dd4e
Fix modules/ typos
...
Found using `codespell -q 3 -S ./3rdparty -L activ,amin,ang,atleast,childs,dof,endwhile,halfs,hist,iff,nd,od,uint`
backporting of commit: ec43292e1e
5 years ago
luz.paz
ec43292e1e
Fix modules/ typos
...
Found using `codespell -q 3 -S ./3rdparty -L activ,amin,ang,atleast,childs,dof,endwhile,halfs,hist,iff,nd,od,uint`
5 years ago
Diego
f7f2438478
Merge pull request #15082 from dvd42:segmentation-module
...
Segmentation module (#15082 )
5 years ago
Dmitry Kurtaev
a9839af903
Add preprocessing warps for separate parameters
5 years ago
Lubov Batanina
778f42ad34
Add high level API (Merge pull request #14780 )
...
* Add high level API
* Fix Model
* Add DetectionModel
* Add ClassificationModel
* Fix classify
* Add python test
* Fix pytest
* Fix comments to review
* Fix detect
* Fix docs
* Modify DetectionOutput postprocessing
* Fix test
* Extract ref boxes
* Fix draw rect
* fix test
* Add rect wrap
* Fix wrap
* Fix detect
* Fix Rect wrap
* Fix OCL_FP16
* Fix MyriadX
* Fix nms
* Fix NMS
* Fix coords
5 years ago
Lubov Batanina
12fdaf895e
Merge pull request #15057 from l-bat:fix_vizualizer
...
* Fix dumpToFile
* Add test
* Fix test
5 years ago
Liubov Batanina
0d2bc7b5fd
Fix TF Split layer
5 years ago
Alexander Alekhin
f84b08ff08
dnn: bump API version
6 years ago
Alexander Alekhin
f5ddaef46b
experimental version++
6 years ago