From 083616d05f5efe574517cae3f9c91cc8f91938d9 Mon Sep 17 00:00:00 2001 From: Anna Petrovicheva Date: Fri, 22 Jul 2016 16:15:45 +0300 Subject: [PATCH] Fixed warnings --- modules/dnn/samples/ssd_object_detection.cpp | 2 +- modules/dnn/src/layers/concat_layer.cpp | 5 ++--- modules/dnn/src/layers/prior_box_layer.cpp | 2 ++ 3 files changed, 5 insertions(+), 4 deletions(-) diff --git a/modules/dnn/samples/ssd_object_detection.cpp b/modules/dnn/samples/ssd_object_detection.cpp index e2660b1b1..fc0d501dc 100644 --- a/modules/dnn/samples/ssd_object_detection.cpp +++ b/modules/dnn/samples/ssd_object_detection.cpp @@ -115,7 +115,7 @@ int main(int argc, char** argv) Mat detectionMat(detection.rows(), detection.cols(), CV_32F, detection.ptrf()); float confidenceThreshold = parser.get("min_confidence"); - for(size_t i = 0; i < detectionMat.rows; i++) + for(int i = 0; i < detectionMat.rows; i++) { float confidence = detectionMat.at(i, 2); diff --git a/modules/dnn/src/layers/concat_layer.cpp b/modules/dnn/src/layers/concat_layer.cpp index 05a701f26..d4c2a9e39 100644 --- a/modules/dnn/src/layers/concat_layer.cpp +++ b/modules/dnn/src/layers/concat_layer.cpp @@ -84,7 +84,7 @@ void ConcatLayer::allocate(const std::vector &inputs, std::vector void ConcatLayer::forward(std::vector &inputs, std::vector &outputs) { - // In case when Blob shape used in allocation and inner matrix shape do not match, this layer did not work in previous implementation. This implementation is just a fix and needs to be rewritten more optimally. + // In case when Blob shape used in allocation and inner matrix shape do not match, this layer did not work in previous implementation. This implementation needs to be rewritten more optimally. if (inputs.size() == 1) { @@ -96,7 +96,7 @@ void ConcatLayer::forward(std::vector &inputs, std::vector &output size_t outputStride = outputs[0].total(axis); size_t offset = 0; - for (int i = 0; i < inputs.size(); ++i) + for (size_t i = 0; i < inputs.size(); ++i) { size_t inputSliceSize = inputs[i]->total(axis); const float* inputData = inputs[i]->ptrf(); @@ -105,7 +105,6 @@ void ConcatLayer::forward(std::vector &inputs, std::vector &output { const float* src = inputData + n * inputSliceSize; float* dst = outputData + n * outputStride + offset; -// memcpy(dst, src, inputSliceSize); for(size_t k = 0; k < inputSliceSize; k++) { dst[k] = src[k]; diff --git a/modules/dnn/src/layers/prior_box_layer.cpp b/modules/dnn/src/layers/prior_box_layer.cpp index 629752e58..be4227c76 100644 --- a/modules/dnn/src/layers/prior_box_layer.cpp +++ b/modules/dnn/src/layers/prior_box_layer.cpp @@ -211,6 +211,8 @@ void PriorBoxLayer::allocate(const std::vector &inputs, std::vector void PriorBoxLayer::forward(std::vector &inputs, std::vector &outputs) { + (void)inputs; // to suppress unused parameter warning + float* outputPtr = outputs[0].ptrf(); // first prior: aspect_ratio = 1, size = min_size