Alexander Alekhin
e8c20bfae6
dnn: fix support drop for IE<2018R3
...
- fix build with IE 2018R3
6 years ago
Dmitry Kurtaev
9cfd219d70
Fix Mobilenet v2 from TensorFlow slim
6 years ago
Kohei Yoshida
212420516f
Add a test case for handling output_shape property for deconvolution layer.
6 years ago
Kohei Yoshida
3bcff7f941
Make use of explicit output shape for the de-convolution layer.
...
This is relevant to cases where the net is loaded from an ONNX model
containing ConvTranspose layers, which may include the output_shape
attribute.
c.f. https://github.com/onnx/onnx/blob/master/docs/Operators.md#ConvTranspose
6 years ago
Dmitry Kurtaev
714b38360f
Extend LRN layer support with IE backend
6 years ago
zuoshaobo
8b52eabd48
fix the region layer's output anchor size normalization error
6 years ago
Giles Payne
55f02d8c66
Merge pull request #14011 from komakai:debug-build
...
* Add Android SDK debug build flag
* Change LOG_TAG define for consistency with other modules
6 years ago
Easton Liu
fcfb29766b
Add ability to read thresh and nms_threshold from YOLO layer in YOLOV3 cfg file.
...
Currently the thresh is hard-coded to be 0.2 and nms_threshold as 0.4.
6 years ago
Alexander Nesterov
74574dfae4
Added optimization fuse
6 years ago
Alexander Alekhin
80d37ba698
dnn: fix usage of CV_LOG_VERBOSE macro
6 years ago
Dmitry Kurtaev
20400aa9f7
Import Upsample and Unsqueeze from ONNX
6 years ago
Dmitry Kurtaev
4cbd09c41c
Add extra limitations for LRN from Inference Engine backend
6 years ago
Dmitry Kurtaev
ed710eaa1c
Make Inference Engine R3 as a minimal supported version
6 years ago
Dmitry Kurtaev
bfd663c281
Add a test for grouped deconvolution from ONNX
6 years ago
Ayush Pandey
5c7fe0fe05
Fix Issue #13479
6 years ago
Dmitry Kurtaev
715f881dda
Replace default confidence threshold for dnn detections from -FLT_MAX to 0
6 years ago
Dmitry Kurtaev
ca5976e3d4
Fix IE backend considering future changes.
6 years ago
Liubov Batanina
183c0fcab1
Changed condition for resize and lrn layers
6 years ago
Dmitry Kurtaev
1606137df2
Read raw floats data from Caffe models
6 years ago
Dmitry Kurtaev
0711dab09d
Fix Intel's Inference Engine backend from future. Second try.
6 years ago
klemens
5d9c6723ee
spelling fixes
...
backport 997b7b18af
6 years ago
Liubov Batanina
6b4becfd03
Enabled tests on IE backend
6 years ago
klemens
997b7b18af
spelling fixes
6 years ago
Alexander Nesterov
9cbdb48d6d
Fix change step
6 years ago
Liubov Batanina
b068d26fad
Using IE backend for normalize layer tests
6 years ago
Dmitry Kurtaev
bc4e471847
Add a mutex for shared Inference Engine plugins
6 years ago
Alexander Alekhin
eab6744ac7
dnn(ocl): use compile-time LOCAL_SIZE parameter
...
instead of get_local_size(0) and dynamic local memory allocation
6 years ago
Dmitry Kurtaev
c918ac298c
Fix IE tests
6 years ago
Dmitry Kurtaev
ac262f5b5d
Clone convolution layer weights only for fusion
6 years ago
Dmitry Kurtaev
3c3c5ef2b6
Fix a dnn bug with retrieving all the output blobs
6 years ago
Dmitry Kurtaev
ff775b2e54
Remove ASSERT_ANY_THROW checks fpr Myriad plugin and FP32 networks
6 years ago
Alexander Nesterov
97c3bcb1b7
Added fix for other size
6 years ago
Lee Jaehwan
3721c8bb06
Merge pull request #13586 from eightco:Core_bugfix3
...
* Add Operator override for multi-channel Mat with literal constant.
* simple test
* Operator overloading channel constraint for primitive types
* fix some test for #13586
6 years ago
Dmitry Kurtaev
f0ddf302b2
Move Inference Engine to new API
6 years ago
Raphael Graf
82c77fa244
dnn: remove malloc.h include
6 years ago
Dmitry Kurtaev
d0504c95f4
Add a text message for Convolution layer's input channels check
6 years ago
WuZhiwen
3d44e9ad92
Merge pull request #13520 from wzw-intel:hang
...
* dnn/Vulkan: fix GPU hang for heavy convolution tasks
Intel i915 driver will declare GPU hang if the compute shader
takes too long to complete. See
https://bugs.freedesktop.org/show_bug.cgi?id=108947 for details.
The idea in this commit is to divide heavy task into several light
ones and run compute shader multiple times to make each run take
short time enough.
TODO: Add more efficient compute shader
Signed-off-by: Wu Zhiwen <zhiwen.wu@intel.com>
* dnn/Vulkan: add a more efficient conv shader
6 years ago
Wu Zhiwen
dae03273cd
dnn: fixup missing vkcom/vulkan combination of backend/target in dnn test
6 years ago
Wu Zhiwen
be6a837e15
dnn: add Vulkan device check for BackendRegistry
6 years ago
Dmitry Kurtaev
840c892abd
Batch normalization in training phase from Torch
6 years ago
Dmitry Kurtaev
59ce1d80a5
Fix dnn tests for Inference Engine R5
6 years ago
Dmitry Kurtaev
257f60582a
Add serialize method for IE net wrapper
...
backport 4ba4901ca9
6 years ago
Alexander Alekhin
bbdc987fc6
dnn: add OpenVINO 2018R5 defines
...
https://software.intel.com/en-us/openvino-toolkit
6 years ago
Dmitry Kurtaev
e71758cfdf
Operate with shapes in ONNX models
6 years ago
Alexander Alekhin
d8583b2c7a
dnn: fix vulkan backend builds with Clang
6 years ago
Dmitry Kurtaev
53f6198f27
Minor fixes in IE backend tests
6 years ago
Dmitry Kurtaev
8422dda2c7
Element-wise subtraction from TensorFlow
6 years ago
Dmitry Kurtaev
93971a53d9
Exclude Input layer from list of outputs for IE networks
6 years ago
Dmitry Kurtaev
3868cb44f1
Exclude Input layer from list of outputs for IE networks
6 years ago
Alexander Alekhin
9ff1c39daa
dnn: fixup available backends/targets
6 years ago