Dmitry Kurtaev
62d079fa69
Fix Normalize layer for Mac
6 years ago
Dmitry Kurtaev
92cbec6fb1
Fix dnn with IE from GitHub
6 years ago
Dmitry Kurtaev
0cfd95c097
Fix TensorFlow's Squeeze and a new fusion for SoftMax from slim backend
6 years ago
Lubov Batanina
60a841c797
Merge pull request #14255 from l-bat:networks_visualization
...
* Add networks visualization
* Disable CXX11
* Fixed multy inputs support
* Added output shapes
* Added color for DLIE/CPU
* Fixed graph colors
6 years ago
Dmitry Kurtaev
a2bbfa1db5
Enable some tests for Inference Engine 2019R1
6 years ago
Dmitry Kurtaev
7ce2397b62
Fix a bug in FusedBatchNorm (TensorFlow) layer importer
6 years ago
Dmitry Kurtaev
ec41a4897a
Remove Switch and Merge nodes from TensorFlow networks
6 years ago
BALACHANDAR S
aa167434e6
Merge pull request #14252 from balachandarsv:master-mac-openvino-support
...
* Mac support for op inference engine
Adding condition to check for mac and add corresponding libraries
* Adding mac support in test cases
6 years ago
103yiran
4bb6edf176
Merge pull request #14117 from 103yiran:103yiran-patch-dnn
...
* Postpone variable definitions
* dnn: reduce scope of 'Mat image' variable
6 years ago
Alexander Alekhin
8483801eab
dnn: use OpenVINO 2019R1 defines
6 years ago
Dmitry Kurtaev
e3286c9055
Enable 1x1 convolution optimization
6 years ago
zuoshaobo
a34c02a001
add support for tf.add_n operation
6 years ago
Lubov Batanina
7d3d6bc4e2
Merge pull request #13932 from l-bat:MyriadX_master_dldt
...
* Fix precision in tests for MyriadX
* Fix ONNX tests
* Add output range in ONNX tests
* Skip tests on Myriad OpenVINO 2018R5
* Add detect MyriadX
* Add detect MyriadX on OpenVINO R5
* Skip tests on Myriad next version of OpenVINO
* dnn(ie): VPU type from environment variable
* dnn(test): validate VPU type
* dnn(test): update DLIE test skip conditions
6 years ago
Alexander Alekhin
e8c20bfae6
dnn: fix support drop for IE<2018R3
...
- fix build with IE 2018R3
6 years ago
Dmitry Kurtaev
9cfd219d70
Fix Mobilenet v2 from TensorFlow slim
6 years ago
Kohei Yoshida
3bcff7f941
Make use of explicit output shape for the de-convolution layer.
...
This is relevant to cases where the net is loaded from an ONNX model
containing ConvTranspose layers, which may include the output_shape
attribute.
c.f. https://github.com/onnx/onnx/blob/master/docs/Operators.md#ConvTranspose
6 years ago
Dmitry Kurtaev
714b38360f
Extend LRN layer support with IE backend
6 years ago
zuoshaobo
8b52eabd48
fix the region layer's output anchor size normalization error
6 years ago
Easton Liu
fcfb29766b
Add ability to read thresh and nms_threshold from YOLO layer in YOLOV3 cfg file.
...
Currently the thresh is hard-coded to be 0.2 and nms_threshold as 0.4.
6 years ago
Alexander Nesterov
74574dfae4
Added optimization fuse
6 years ago
Alexander Alekhin
80d37ba698
dnn: fix usage of CV_LOG_VERBOSE macro
6 years ago
Dmitry Kurtaev
20400aa9f7
Import Upsample and Unsqueeze from ONNX
6 years ago
Dmitry Kurtaev
4cbd09c41c
Add extra limitations for LRN from Inference Engine backend
6 years ago
Dmitry Kurtaev
ed710eaa1c
Make Inference Engine R3 as a minimal supported version
6 years ago
Dmitry Kurtaev
bfd663c281
Add a test for grouped deconvolution from ONNX
6 years ago
Ayush Pandey
5c7fe0fe05
Fix Issue #13479
6 years ago
Dmitry Kurtaev
715f881dda
Replace default confidence threshold for dnn detections from -FLT_MAX to 0
6 years ago
Dmitry Kurtaev
ca5976e3d4
Fix IE backend considering future changes.
6 years ago
Liubov Batanina
183c0fcab1
Changed condition for resize and lrn layers
6 years ago
Dmitry Kurtaev
1606137df2
Read raw floats data from Caffe models
6 years ago
Dmitry Kurtaev
0711dab09d
Fix Intel's Inference Engine backend from future. Second try.
6 years ago
klemens
5d9c6723ee
spelling fixes
...
backport 997b7b18af
6 years ago
Liubov Batanina
6b4becfd03
Enabled tests on IE backend
6 years ago
klemens
997b7b18af
spelling fixes
6 years ago
Alexander Nesterov
9cbdb48d6d
Fix change step
6 years ago
Liubov Batanina
b068d26fad
Using IE backend for normalize layer tests
6 years ago
Dmitry Kurtaev
bc4e471847
Add a mutex for shared Inference Engine plugins
6 years ago
Alexander Alekhin
eab6744ac7
dnn(ocl): use compile-time LOCAL_SIZE parameter
...
instead of get_local_size(0) and dynamic local memory allocation
6 years ago
Dmitry Kurtaev
c918ac298c
Fix IE tests
6 years ago
Dmitry Kurtaev
ac262f5b5d
Clone convolution layer weights only for fusion
6 years ago
Dmitry Kurtaev
3c3c5ef2b6
Fix a dnn bug with retrieving all the output blobs
6 years ago
Dmitry Kurtaev
ff775b2e54
Remove ASSERT_ANY_THROW checks fpr Myriad plugin and FP32 networks
6 years ago
Alexander Nesterov
97c3bcb1b7
Added fix for other size
6 years ago
Dmitry Kurtaev
f0ddf302b2
Move Inference Engine to new API
6 years ago
Raphael Graf
82c77fa244
dnn: remove malloc.h include
6 years ago
Dmitry Kurtaev
d0504c95f4
Add a text message for Convolution layer's input channels check
6 years ago
WuZhiwen
3d44e9ad92
Merge pull request #13520 from wzw-intel:hang
...
* dnn/Vulkan: fix GPU hang for heavy convolution tasks
Intel i915 driver will declare GPU hang if the compute shader
takes too long to complete. See
https://bugs.freedesktop.org/show_bug.cgi?id=108947 for details.
The idea in this commit is to divide heavy task into several light
ones and run compute shader multiple times to make each run take
short time enough.
TODO: Add more efficient compute shader
Signed-off-by: Wu Zhiwen <zhiwen.wu@intel.com>
* dnn/Vulkan: add a more efficient conv shader
6 years ago
Wu Zhiwen
be6a837e15
dnn: add Vulkan device check for BackendRegistry
6 years ago
Dmitry Kurtaev
840c892abd
Batch normalization in training phase from Torch
6 years ago
Dmitry Kurtaev
59ce1d80a5
Fix dnn tests for Inference Engine R5
6 years ago