Merge pull request #21772 from luzpaz:typo/dnn

pull/21781/head
Alexander Alekhin 3 years ago
commit 78bc11465b
  1. 4
      modules/dnn/include/opencv2/dnn/dnn.hpp
  2. 2
      modules/dnn/src/cuda/concat.cu
  3. 2
      modules/dnn/src/cuda/kernel_dispatcher.hpp
  4. 4
      modules/dnn/src/cuda/permute.cu
  5. 2
      modules/dnn/src/cuda/slice.cu
  6. 2
      modules/dnn/src/cuda4dnn/csl/memory.hpp
  7. 4
      modules/dnn/src/cuda4dnn/csl/pointer.hpp
  8. 2
      modules/dnn/src/darknet/darknet_io.cpp
  9. 2
      modules/dnn/src/model.cpp
  10. 2
      modules/dnn/src/tengine4dnn/src/tengine_graph_convolution.cpp
  11. 4
      modules/dnn/test/test_ie_models.cpp
  12. 4
      samples/dnn/speech_recognition.cpp
  13. 6
      samples/dnn/speech_recognition.py

@ -389,7 +389,7 @@ CV__DNN_INLINE_NS_BEGIN
/**
* @brief "Deattaches" all the layers, attached to particular layer.
* @brief "Detaches" all the layers, attached to particular layer.
*/
virtual void unsetAttached();
@ -1579,7 +1579,7 @@ public:
* - top-right
* - bottom-right
*
* Use cv::getPerspectiveTransform function to retrive image region without perspective transformations.
* Use cv::getPerspectiveTransform function to retrieve image region without perspective transformations.
*
* @note If DL model doesn't support that kind of output then result may be derived from detectTextRectangles() output.
*

@ -100,7 +100,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace kernels {
CV_Assert(output.rank() == input.rank());
CV_Assert(output_axis_offset < output.get_axis_size(axis));
/* if axes preceeding the concat axis are all singleton, the concat blocks are contiguous
/* if axes preceding the concat axis are all singleton, the concat blocks are contiguous
* in the output and we can copy each block directly
*/
if (output.size_range(0, axis) == 1)

@ -33,7 +33,7 @@
* template <class T, std::size_t Rank>
* void launch_some_kernel(...);
*
* // creates the dispatcher named "some_dispatcher" which invokves the correct instantiation of "launch_some_kernel"
* // creates the dispatcher named "some_dispatcher" which invokes the correct instantiation of "launch_some_kernel"
* GENERATE_KERNEL_DISPATCHER(some_dispatcher, launch_some_kernel);
*
* // internal API function

@ -72,7 +72,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace kernels {
__syncthreads();
/* We interchange `threadIdx.x` and `threadIdx.y` so that consecutive output indices map to
* consecutive threads. This would allow writes across threds in a warp to be coalesced.
* consecutive threads. This would allow writes across threads in a warp to be coalesced.
*/
const index_type out_x = blockIdx.y * TILE_SIZE + threadIdx.x;
const index_type out_y_begin = blockIdx.x * TILE_SIZE + threadIdx.y;
@ -156,7 +156,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace kernels {
* tensor indices be [o1, o2, ...]. The permutation operation essentially copies items
* from the input tensor to new locations in the output tensor as dictated by the indices.
*
* If the size of the nth axis (say i2) of the input is one the input and output indicies for
* If the size of the nth axis (say i2) of the input is one the input and output indices for
* all the elements will be of the form be [i1, 0, ...] and [..., 0, ...] respectively.
* The index does not contribute to the element's address calculation and hence would give
* identical result if it weren't there.

@ -159,7 +159,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace kernels {
/* We can do a copy if the reduced rank is two and only the first axis is sliced.
* The general requirement is that only one axis is sliced and all the axes that
* preceed the sliced axis are singleton. However, the reductions above will remove
* precede the sliced axis are singleton. However, the reductions above will remove
* all the leading singleton axes and merge the trailing unsliced axes into one, or
* zero if there are no trailing unsliced axes. The latter is handled separately.
*/

@ -68,7 +68,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace csl {
}
}
});
/* std::shared_ptr<T>::reset invokves the deleter if an exception occurs; hence, we don't
/* std::shared_ptr<T>::reset invokes the deleter if an exception occurs; hence, we don't
* need to have a try-catch block to free the allocated device memory
*/

@ -147,7 +147,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace csl {
/* host const void pointer to const void device pointer */
CUDA4DNN_HOST_DEVICE explicit DevicePtr(pointer ptr_) noexcept : ptr{ ptr_ } { }
/* allow any device pointer to be implicitly convereted to void device pointer */
/* allow any device pointer to be implicitly converted to void device pointer */
template <class T>
CUDA4DNN_HOST_DEVICE DevicePtr(DevicePtr<T> ptr_) noexcept : ptr{ ptr_.get() } { }
@ -199,7 +199,7 @@ namespace cv { namespace dnn { namespace cuda4dnn { namespace csl {
/* host pointer to device pointer */
CUDA4DNN_HOST_DEVICE explicit DevicePtr(pointer ptr_) noexcept : ptr{ ptr_ } { }
/* allow any device pointer to mutable memory to be implicitly convereted to void device pointer */
/* allow any device pointer to mutable memory to be implicitly converted to void device pointer */
template <class T, typename std::enable_if<!std::is_const<T>::value, bool>::type = false>
CUDA4DNN_HOST_DEVICE DevicePtr(DevicePtr<T> ptr_) noexcept : ptr { ptr_.get() } { }

@ -791,7 +791,7 @@ namespace cv {
if (layers_vec.size() > 1)
{
// layer ids in layers_vec - inputs of Slice layers
// after adding offset to layers_vec: layer ids - ouputs of Slice layers
// after adding offset to layers_vec: layer ids - outputs of Slice layers
for (size_t k = 0; k < layers_vec.size(); ++k)
layers_vec[k] += layers_vec.size();

@ -799,7 +799,7 @@ struct TextRecognitionModel_Impl : public Model::Impl
virtual
std::string ctcPrefixBeamSearchDecode(const Mat& prediction) {
// CTC prefix beam seach decode.
// CTC prefix beam search decode.
// For more detail, refer to:
// https://distill.pub/2017/ctc/#inference
// https://gist.github.com/awni/56369a90d03953e370f3964c826ed4b0i

@ -331,7 +331,7 @@ teng_graph_t tengine_init(const char* layer_name, float* input_, int inch, int g
teg_weight = kernel_;
}
/* initial the resoruce of tengine */
/* initial the resource of tengine */
if(false == tengine_init_flag)
{
init_tengine();

@ -290,7 +290,7 @@ void runIE(Target target, const std::string& xmlPath, const std::string& binPath
if (cvtest::debugLevel > 0)
{
const std::vector<size_t>& dims = desc.getDims();
std::cout << "Input: '" << it.first << "' precison=" << desc.getPrecision() << " dims=" << dims.size() << " [";
std::cout << "Input: '" << it.first << "' precision=" << desc.getPrecision() << " dims=" << dims.size() << " [";
for (auto d : dims)
std::cout << " " << d;
std::cout << "] ocv_mat=" << inputsMap[it.first].size << " of " << typeToString(inputsMap[it.first].type()) << std::endl;
@ -308,7 +308,7 @@ void runIE(Target target, const std::string& xmlPath, const std::string& binPath
if (cvtest::debugLevel > 0)
{
const std::vector<size_t>& dims = desc.getDims();
std::cout << "Output: '" << it.first << "' precison=" << desc.getPrecision() << " dims=" << dims.size() << " [";
std::cout << "Output: '" << it.first << "' precision=" << desc.getPrecision() << " dims=" << dims.size() << " [";
for (auto d : dims)
std::cout << " " << d;
std::cout << "] ocv_mat=" << outputsMap[it.first].size << " of " << typeToString(outputsMap[it.first].type()) << std::endl;

@ -33,7 +33,7 @@ private:
double highfreq = sample_rate / 2;
public:
// Mel filterbanks preperation
// Mel filterbanks preparation
double hz_to_mel(double frequencies)
{
//Converts frequencies from hz to mel scale
@ -149,7 +149,7 @@ public:
return weights;
}
// STFT preperation
// STFT preparation
vector<double> pad_window_center(vector<double>&data, int size)
{
// Pad the window out to n_fft size

@ -44,7 +44,7 @@ import os
model.graph.initializer.insert(i,init)
```
6. Add an additional reshape node to handle the inconsistant input from python and c++ of openCV.
6. Add an additional reshape node to handle the inconsistent input from python and c++ of openCV.
see https://github.com/opencv/opencv/issues/19091
Make & insert a new node with 'Reshape' operation & required initializer
```
@ -256,7 +256,7 @@ class FilterbankFeatures:
weights *= enorm[:, np.newaxis]
return weights
# STFT preperation
# STFT preparation
def pad_window_center(self, data, size, axis=-1, **kwargs):
'''
Centers the data and pads.
@ -329,7 +329,7 @@ class FilterbankFeatures:
then padded with zeros to match n_fft
fft_window : a vector or array of length `n_fft` having values computed by a
window function
pad_mode : mode while padding the singnal
pad_mode : mode while padding the signal
return_complex : returns array with complex data type if `True`
return : Matrix of short-term Fourier transform coefficients.
'''

Loading…
Cancel
Save