diff --git a/modules/core/src/count_non_zero.cpp b/modules/core/src/count_non_zero.cpp index 368dcfc3a6..142a4983c1 100644 --- a/modules/core/src/count_non_zero.cpp +++ b/modules/core/src/count_non_zero.cpp @@ -25,51 +25,34 @@ static int countNonZero_(const T* src, int len ) static int countNonZero8u( const uchar* src, int len ) { int i=0, nz = 0; -#if CV_SSE2 - if(USE_SSE2)//5x-6x - { - __m128i v_zero = _mm_setzero_si128(); - __m128i sum = _mm_setzero_si128(); - - for (; i<=len-16; i+=16) - { - __m128i r0 = _mm_loadu_si128((const __m128i*)(src+i)); - sum = _mm_add_epi32(sum, _mm_sad_epu8(_mm_sub_epi8(v_zero, _mm_cmpeq_epi8(r0, v_zero)), v_zero)); - } - nz = i - _mm_cvtsi128_si32(_mm_add_epi32(sum, _mm_unpackhi_epi64(sum, sum))); - } -#elif CV_NEON - int len0 = len & -16, blockSize1 = (1 << 8) - 16, blockSize0 = blockSize1 << 6; - uint32x4_t v_nz = vdupq_n_u32(0u); - uint8x16_t v_zero = vdupq_n_u8(0), v_1 = vdupq_n_u8(1); - const uchar * src0 = src; +#if CV_SIMD + int len0 = len & -v_uint8::nlanes; + v_uint8 v_zero = vx_setzero_u8(); + v_uint8 v_one = vx_setall_u8(1); - while( i < len0 ) + v_uint32 v_sum32 = vx_setzero_u32(); + while (i < len0) { - int blockSizei = std::min(len0 - i, blockSize0), j = 0; - - while (j < blockSizei) + v_uint16 v_sum16 = vx_setzero_u16(); + int j = i; + while (j < std::min(len0, i + 65280 * v_uint16::nlanes)) { - int blockSizej = std::min(blockSizei - j, blockSize1), k = 0; - uint8x16_t v_pz = v_zero; - - for( ; k <= blockSizej - 16; k += 16 ) - v_pz = vaddq_u8(v_pz, vandq_u8(vceqq_u8(vld1q_u8(src0 + k), v_zero), v_1)); - - uint16x8_t v_p1 = vmovl_u8(vget_low_u8(v_pz)), v_p2 = vmovl_u8(vget_high_u8(v_pz)); - v_nz = vaddq_u32(vaddl_u16(vget_low_u16(v_p1), vget_high_u16(v_p1)), v_nz); - v_nz = vaddq_u32(vaddl_u16(vget_low_u16(v_p2), vget_high_u16(v_p2)), v_nz); - - src0 += blockSizej; - j += blockSizej; + v_uint8 v_sum8 = vx_setzero_u8(); + int k = j; + for (; k < std::min(len0, j + 255 * v_uint8::nlanes); k += v_uint8::nlanes) + v_sum8 += v_one & (vx_load(src + k) == v_zero); + v_uint16 part1, part2; + v_expand(v_sum8, part1, part2); + v_sum16 += part1 + part2; + j = k; } - - i += blockSizei; + v_uint32 part1, part2; + v_expand(v_sum16, part1, part2); + v_sum32 += part1 + part2; + i = j; } - - CV_DECL_ALIGNED(16) unsigned int buf[4]; - vst1q_u32(buf, v_nz); - nz += i - saturate_cast(buf[0] + buf[1] + buf[2] + buf[3]); + nz = i - v_reduce_sum(v_sum32); + v_cleanup(); #endif for( ; i < len; i++ ) nz += src[i] != 0; @@ -79,159 +62,112 @@ static int countNonZero8u( const uchar* src, int len ) static int countNonZero16u( const ushort* src, int len ) { int i = 0, nz = 0; -#if CV_SSE2 - if (USE_SSE2) - { - __m128i v_zero = _mm_setzero_si128 (); - __m128i sum = _mm_setzero_si128(); +#if CV_SIMD + int len0 = len & -v_int8::nlanes; + v_uint16 v_zero = vx_setzero_u16(); + v_int8 v_one = vx_setall_s8(1); - for ( ; i <= len - 8; i += 8) - { - __m128i r0 = _mm_loadu_si128((const __m128i*)(src + i)); - sum = _mm_add_epi32(sum, _mm_sad_epu8(_mm_sub_epi8(v_zero, _mm_cmpeq_epi16(r0, v_zero)), v_zero)); - } - - nz = i - (_mm_cvtsi128_si32(_mm_add_epi32(sum, _mm_unpackhi_epi64(sum, sum))) >> 1); - src += i; - } -#elif CV_NEON - int len0 = len & -8, blockSize1 = (1 << 15), blockSize0 = blockSize1 << 6; - uint32x4_t v_nz = vdupq_n_u32(0u); - uint16x8_t v_zero = vdupq_n_u16(0), v_1 = vdupq_n_u16(1); - - while( i < len0 ) + v_int32 v_sum32 = vx_setzero_s32(); + while (i < len0) { - int blockSizei = std::min(len0 - i, blockSize0), j = 0; - - while (j < blockSizei) + v_int16 v_sum16 = vx_setzero_s16(); + int j = i; + while (j < std::min(len0, i + 32766 * v_int16::nlanes)) { - int blockSizej = std::min(blockSizei - j, blockSize1), k = 0; - uint16x8_t v_pz = v_zero; - - for( ; k <= blockSizej - 8; k += 8 ) - v_pz = vaddq_u16(v_pz, vandq_u16(vceqq_u16(vld1q_u16(src + k), v_zero), v_1)); - - v_nz = vaddq_u32(vaddl_u16(vget_low_u16(v_pz), vget_high_u16(v_pz)), v_nz); - - src += blockSizej; - j += blockSizej; + v_int8 v_sum8 = vx_setzero_s8(); + int k = j; + for (; k < std::min(len0, j + 127 * v_int8::nlanes); k += v_int8::nlanes) + v_sum8 += v_one & v_pack(v_reinterpret_as_s16(vx_load(src + k) == v_zero), v_reinterpret_as_s16(vx_load(src + k + v_uint16::nlanes) == v_zero)); + v_int16 part1, part2; + v_expand(v_sum8, part1, part2); + v_sum16 += part1 + part2; + j = k; } - - i += blockSizei; + v_int32 part1, part2; + v_expand(v_sum16, part1, part2); + v_sum32 += part1 + part2; + i = j; } - - CV_DECL_ALIGNED(16) unsigned int buf[4]; - vst1q_u32(buf, v_nz); - nz += i - saturate_cast(buf[0] + buf[1] + buf[2] + buf[3]); + nz = i - v_reduce_sum(v_sum32); + v_cleanup(); #endif - return nz + countNonZero_(src, len - i); + return nz + countNonZero_(src + i, len - i); } static int countNonZero32s( const int* src, int len ) { int i = 0, nz = 0; -#if CV_SSE2 - if (USE_SSE2) - { - __m128i v_zero = _mm_setzero_si128 (); - __m128i sum = _mm_setzero_si128(); +#if CV_SIMD + int len0 = len & -v_int8::nlanes; + v_int32 v_zero = vx_setzero_s32(); + v_int8 v_one = vx_setall_s8(1); - for ( ; i <= len - 4; i += 4) - { - __m128i r0 = _mm_loadu_si128((const __m128i*)(src + i)); - sum = _mm_add_epi32(sum, _mm_sad_epu8(_mm_sub_epi8(v_zero, _mm_cmpeq_epi32(r0, v_zero)), v_zero)); - } - - nz = i - (_mm_cvtsi128_si32(_mm_add_epi32(sum, _mm_unpackhi_epi64(sum, sum))) >> 2); - src += i; - } -#elif CV_NEON - int len0 = len & -8, blockSize1 = (1 << 15), blockSize0 = blockSize1 << 6; - uint32x4_t v_nz = vdupq_n_u32(0u); - int32x4_t v_zero = vdupq_n_s32(0.0f); - uint16x8_t v_1 = vdupq_n_u16(1u), v_zerou = vdupq_n_u16(0u); - - while( i < len0 ) + v_int32 v_sum32 = vx_setzero_s32(); + while (i < len0) { - int blockSizei = std::min(len0 - i, blockSize0), j = 0; - - while (j < blockSizei) + v_int16 v_sum16 = vx_setzero_s16(); + int j = i; + while (j < std::min(len0, i + 32766 * v_int16::nlanes)) { - int blockSizej = std::min(blockSizei - j, blockSize1), k = 0; - uint16x8_t v_pz = v_zerou; - - for( ; k <= blockSizej - 8; k += 8 ) - v_pz = vaddq_u16(v_pz, vandq_u16(vcombine_u16(vmovn_u32(vceqq_s32(vld1q_s32(src + k), v_zero)), - vmovn_u32(vceqq_s32(vld1q_s32(src + k + 4), v_zero))), v_1)); - - v_nz = vaddq_u32(vaddl_u16(vget_low_u16(v_pz), vget_high_u16(v_pz)), v_nz); - - src += blockSizej; - j += blockSizej; + v_int8 v_sum8 = vx_setzero_s8(); + int k = j; + for (; k < std::min(len0, j + 127 * v_int8::nlanes); k += v_int8::nlanes) + v_sum8 += v_one & v_pack( + v_pack(vx_load(src + k ) == v_zero, vx_load(src + k + v_int32::nlanes) == v_zero), + v_pack(vx_load(src + k + 2*v_int32::nlanes) == v_zero, vx_load(src + k + 3*v_int32::nlanes) == v_zero) + ); + v_int16 part1, part2; + v_expand(v_sum8, part1, part2); + v_sum16 += part1 + part2; + j = k; } - - i += blockSizei; + v_int32 part1, part2; + v_expand(v_sum16, part1, part2); + v_sum32 += part1 + part2; + i = j; } - - CV_DECL_ALIGNED(16) unsigned int buf[4]; - vst1q_u32(buf, v_nz); - nz += i - saturate_cast(buf[0] + buf[1] + buf[2] + buf[3]); + nz = i - v_reduce_sum(v_sum32); + v_cleanup(); #endif - return nz + countNonZero_(src, len - i); + return nz + countNonZero_(src + i, len - i); } static int countNonZero32f( const float* src, int len ) { int i = 0, nz = 0; -#if CV_SSE2 - if (USE_SSE2) - { - __m128 v_zero_f = _mm_setzero_ps(); - __m128i v_zero = _mm_setzero_si128 (); - __m128i sum = _mm_setzero_si128(); +#if CV_SIMD + int len0 = len & -v_int8::nlanes; + v_float32 v_zero = vx_setzero_f32(); + v_int8 v_one = vx_setall_s8(1); - for ( ; i <= len - 4; i += 4) - { - __m128 r0 = _mm_loadu_ps(src + i); - sum = _mm_add_epi32(sum, _mm_sad_epu8(_mm_sub_epi8(v_zero, _mm_castps_si128(_mm_cmpeq_ps(r0, v_zero_f))), v_zero)); - } - - nz = i - (_mm_cvtsi128_si32(_mm_add_epi32(sum, _mm_unpackhi_epi64(sum, sum))) >> 2); - src += i; - } -#elif CV_NEON - int len0 = len & -8, blockSize1 = (1 << 15), blockSize0 = blockSize1 << 6; - uint32x4_t v_nz = vdupq_n_u32(0u); - float32x4_t v_zero = vdupq_n_f32(0.0f); - uint16x8_t v_1 = vdupq_n_u16(1u), v_zerou = vdupq_n_u16(0u); - - while( i < len0 ) + v_int32 v_sum32 = vx_setzero_s32(); + while (i < len0) { - int blockSizei = std::min(len0 - i, blockSize0), j = 0; - - while (j < blockSizei) + v_int16 v_sum16 = vx_setzero_s16(); + int j = i; + while (j < std::min(len0, i + 32766 * v_int16::nlanes)) { - int blockSizej = std::min(blockSizei - j, blockSize1), k = 0; - uint16x8_t v_pz = v_zerou; - - for( ; k <= blockSizej - 8; k += 8 ) - v_pz = vaddq_u16(v_pz, vandq_u16(vcombine_u16(vmovn_u32(vceqq_f32(vld1q_f32(src + k), v_zero)), - vmovn_u32(vceqq_f32(vld1q_f32(src + k + 4), v_zero))), v_1)); - - v_nz = vaddq_u32(vaddl_u16(vget_low_u16(v_pz), vget_high_u16(v_pz)), v_nz); - - src += blockSizej; - j += blockSizej; + v_int8 v_sum8 = vx_setzero_s8(); + int k = j; + for (; k < std::min(len0, j + 127 * v_int8::nlanes); k += v_int8::nlanes) + v_sum8 += v_one & v_pack( + v_pack(v_reinterpret_as_s32(vx_load(src + k ) == v_zero), v_reinterpret_as_s32(vx_load(src + k + v_float32::nlanes) == v_zero)), + v_pack(v_reinterpret_as_s32(vx_load(src + k + 2*v_float32::nlanes) == v_zero), v_reinterpret_as_s32(vx_load(src + k + 3*v_float32::nlanes) == v_zero)) + ); + v_int16 part1, part2; + v_expand(v_sum8, part1, part2); + v_sum16 += part1 + part2; + j = k; } - - i += blockSizei; + v_int32 part1, part2; + v_expand(v_sum16, part1, part2); + v_sum32 += part1 + part2; + i = j; } - - CV_DECL_ALIGNED(16) unsigned int buf[4]; - vst1q_u32(buf, v_nz); - nz += i - saturate_cast(buf[0] + buf[1] + buf[2] + buf[3]); + nz = i - v_reduce_sum(v_sum32); + v_cleanup(); #endif - return nz + countNonZero_(src, len - i); + return nz + countNonZero_(src + i, len - i); } static int countNonZero64f( const double* src, int len )