Fixing dnn Resize layer for variable input size
* Fix onnx loading of resize/upsample layers for different opset
* group all DynamicResize tests
* cleaned up scales checks
* Simplify branching
Fix loading issue for Faster RCNN model from #16783
* Add a reproducer with multi-output Gather
* Fix an issue with ONNX graph simplifier
* fix build
* Move checks to correct class
* Minor changes for better code appearence
Add support for Conv1D on OpenCV backend
* Add support for Conv1D on OpenCV backend
* disable tests on other targets/backends
* Fix formatting
* Restore comment
* Remove unnecessary flag and fix test logic
* Fix perf test
* fix braces
* Fix indentation, assert check and remove unnecessary condition
* Remove unnecessary changes
* Add test cases for variable weights and bias
* dnn(conv): fallback on OpenCV+CPU instead of failures
* coding style
Fix loading of ONNX models with Resize operation with Opset 11 for newer versions of Pytorch
* Add reproducer for Resize operation from newer versions of Pytorch
* Fix loading of scales parameter for Resize layer
* Change check type for better diagnostic messages
* Fix ONNX loading in issues opencv#17516, opencv#17531
* Add tests for Linear and Matmul layers
* Disable tests for IE versions lower than 20.4
* Skip unstable tests with OpenCL FP16 on Intel GPU
* Add correct test filtering for OpenCL FP16 tests
revise default proto to match the filename in documentations
fix a bug
beautify python codes
fix bug
beautify codes
add test samples with larger/smaller size
remove unless code
using bytearray without creating tmp file
remove useless codes
* hopefully, eliminated compile warnings, errors, as well as failure in one test
* * fixed a few typos
* decreased buffer size in some cases
* added more optimal im2row branch in the case of 1x1 convolutions
* tuned fastConv to reduce the number of passes over arrays
backport of commit 77b01deb80
add relu option
add relu as activation option in darknet
simplify the setParams if-else ladder
add relu as activation option in darknet
correct activation_param type
format
format
add relu as activation option in darknet
spacing
spacing
add relu as activation option in darknet
DNN: OpenCL/slice update
* dnn(ocl/slice): make slice kernel VTune friendly
- more unique names
- inline code of copy functions
* dnn(ocl/slice): prefer to spawn more work groups
- even in case with 1D copy
- perf improvement up to 2x of kernel time (due to changed configuration 128x1x1 => 128x32x1)
* dnn(ocl/slice): cache kernel exec info