|
|
|
;******************************************************************************
|
|
|
|
;* Pixel utilities SIMD
|
|
|
|
;*
|
|
|
|
;* Copyright (C) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
|
|
|
|
;* Copyright (C) 2014 Clément Bœsch <u pkh me>
|
|
|
|
;*
|
|
|
|
;* This file is part of FFmpeg.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
|
|
;* License as published by the Free Software Foundation; either
|
|
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
;* Lesser General Public License for more details.
|
|
|
|
;*
|
|
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
;******************************************************************************
|
|
|
|
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
|
|
|
|
SECTION .text
|
|
|
|
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_8x8_mmxext(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
INIT_MMX mmxext
|
|
|
|
cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2
|
|
|
|
pxor m2, m2
|
|
|
|
%rep 4
|
|
|
|
mova m0, [src1q]
|
|
|
|
mova m1, [src1q + stride1q]
|
|
|
|
psadbw m0, [src2q]
|
|
|
|
psadbw m1, [src2q + stride2q]
|
|
|
|
paddw m2, m0
|
|
|
|
paddw m2, m1
|
|
|
|
lea src1q, [src1q + 2*stride1q]
|
|
|
|
lea src2q, [src2q + 2*stride2q]
|
|
|
|
%endrep
|
|
|
|
movd eax, m2
|
avutil/x86/pixelutils: Empty MMX state in ff_pixelutils_sad_8x8_mmxext
We currently mostly do not empty the MMX state in our MMX
DSP functions; instead we only do so before code that might
be using x87 code. This is a violation of the System V i386 ABI
(and maybe of other ABIs, too):
"The CPU shall be in x87 mode upon entry to a function. Therefore,
every function that uses the MMX registers is required to issue an
emms or femms instruction after using MMX registers, before returning
or calling another function." (See 2.2.1 in [1])
This patch does not intend to change all these functions to abide
by the ABI; it only does so for ff_pixelutils_sad_8x8_mmxext, as this
function can by called by external users, because it is exported
via the pixelutils API. Without this, the following fragment will
assert (on x86/x64):
uint8_t src1[8 * 8], src2[8 * 8];
av_pixelutils_sad_fn fn = av_pixelutils_get_sad_fn(3, 3, 0, NULL);
fn(src1, 8, src2, 8);
av_assert0_fpu();
[1]: https://raw.githubusercontent.com/wiki/hjl-tools/x86-psABI/intel386-psABI-1.1.pdf
Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
1 year ago
|
|
|
emms
|
|
|
|
RET
|
|
|
|
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal pixelutils_sad_16x16, 4,4,5, src1, stride1, src2, stride2
|
|
|
|
movu m4, [src1q]
|
|
|
|
movu m2, [src2q]
|
|
|
|
movu m1, [src1q + stride1q]
|
|
|
|
movu m3, [src2q + stride2q]
|
|
|
|
psadbw m4, m2
|
|
|
|
psadbw m1, m3
|
|
|
|
paddw m4, m1
|
|
|
|
%rep 7
|
|
|
|
lea src1q, [src1q + 2*stride1q]
|
|
|
|
lea src2q, [src2q + 2*stride2q]
|
|
|
|
movu m0, [src1q]
|
|
|
|
movu m2, [src2q]
|
|
|
|
movu m1, [src1q + stride1q]
|
|
|
|
movu m3, [src2q + stride2q]
|
|
|
|
psadbw m0, m2
|
|
|
|
psadbw m1, m3
|
|
|
|
paddw m4, m0
|
|
|
|
paddw m4, m1
|
|
|
|
%endrep
|
|
|
|
movhlps m0, m4
|
|
|
|
paddw m4, m0
|
|
|
|
movd eax, m4
|
|
|
|
RET
|
|
|
|
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_[au]_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
%macro SAD_XMM_16x16 1
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal pixelutils_sad_%1_16x16, 4,4,3, src1, stride1, src2, stride2
|
|
|
|
mov%1 m2, [src2q]
|
|
|
|
psadbw m2, [src1q]
|
|
|
|
mov%1 m1, [src2q + stride2q]
|
|
|
|
psadbw m1, [src1q + stride1q]
|
|
|
|
paddw m2, m1
|
|
|
|
%rep 7
|
|
|
|
lea src1q, [src1q + 2*stride1q]
|
|
|
|
lea src2q, [src2q + 2*stride2q]
|
|
|
|
mov%1 m0, [src2q]
|
|
|
|
psadbw m0, [src1q]
|
|
|
|
mov%1 m1, [src2q + stride2q]
|
|
|
|
psadbw m1, [src1q + stride1q]
|
|
|
|
paddw m2, m0
|
|
|
|
paddw m2, m1
|
|
|
|
%endrep
|
|
|
|
movhlps m0, m2
|
|
|
|
paddw m2, m0
|
|
|
|
movd eax, m2
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
SAD_XMM_16x16 a
|
|
|
|
SAD_XMM_16x16 u
|
|
|
|
|
|
|
|
|
|
|
|
%macro PROCESS_SAD_32x4_U 0
|
|
|
|
movu m1, [r2]
|
|
|
|
movu m2, [r2 + 16]
|
|
|
|
movu m3, [r0]
|
|
|
|
movu m4, [r0 + 16]
|
|
|
|
psadbw m1, m3
|
|
|
|
psadbw m2, m4
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
movu m1, [r2]
|
|
|
|
movu m2, [r2 + 16]
|
|
|
|
movu m3, [r0]
|
|
|
|
movu m4, [r0 + 16]
|
|
|
|
psadbw m1, m3
|
|
|
|
psadbw m2, m4
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
movu m1, [r2]
|
|
|
|
movu m2, [r2 + 16]
|
|
|
|
movu m3, [r0]
|
|
|
|
movu m4, [r0 + 16]
|
|
|
|
psadbw m1, m3
|
|
|
|
psadbw m2, m4
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
movu m1, [r2]
|
|
|
|
movu m2, [r2 + 16]
|
|
|
|
movu m3, [r0]
|
|
|
|
movu m4, [r0 + 16]
|
|
|
|
psadbw m1, m3
|
|
|
|
psadbw m2, m4
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
%macro PROCESS_SAD_32x4 1
|
|
|
|
mov%1 m1, [r2]
|
|
|
|
mov%1 m2, [r2 + 16]
|
|
|
|
psadbw m1, [r0]
|
|
|
|
psadbw m2, [r0 + 16]
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
mov%1 m1, [r2]
|
|
|
|
mov%1 m2, [r2 + 16]
|
|
|
|
psadbw m1, [r0]
|
|
|
|
psadbw m2, [r0 + 16]
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
mov%1 m1, [r2]
|
|
|
|
mov%1 m2, [r2 + 16]
|
|
|
|
psadbw m1, [r0]
|
|
|
|
psadbw m2, [r0 + 16]
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
|
|
|
|
mov%1 m1, [r2]
|
|
|
|
mov%1 m2, [r2 + 16]
|
|
|
|
psadbw m1, [r0]
|
|
|
|
psadbw m2, [r0 + 16]
|
|
|
|
paddd m1, m2
|
|
|
|
paddd m0, m1
|
|
|
|
lea r2, [r2 + r3]
|
|
|
|
lea r0, [r0 + r1]
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
;-----------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-----------------------------------------------------------------------------
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal pixelutils_sad_32x32, 4,5,5, src1, stride1, src2, stride2
|
|
|
|
pxor m0, m0
|
|
|
|
mov r4d, 4
|
|
|
|
.loop:
|
|
|
|
PROCESS_SAD_32x4_U
|
|
|
|
PROCESS_SAD_32x4_U
|
|
|
|
dec r4d
|
|
|
|
jnz .loop
|
|
|
|
|
|
|
|
movhlps m1, m0
|
|
|
|
paddd m0, m1
|
|
|
|
movd eax, m0
|
|
|
|
RET
|
|
|
|
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_[au]_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
%macro SAD_XMM_32x32 1
|
|
|
|
INIT_XMM sse2
|
|
|
|
cglobal pixelutils_sad_%1_32x32, 4,5,3, src1, stride1, src2, stride2
|
|
|
|
pxor m0, m0
|
|
|
|
mov r4d, 4
|
|
|
|
.loop:
|
|
|
|
PROCESS_SAD_32x4 %1
|
|
|
|
PROCESS_SAD_32x4 %1
|
|
|
|
dec r4d
|
|
|
|
jnz .loop
|
|
|
|
|
|
|
|
movhlps m1, m0
|
|
|
|
paddd m0, m1
|
|
|
|
movd eax, m0
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
SAD_XMM_32x32 a
|
|
|
|
SAD_XMM_32x32 u
|
|
|
|
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
INIT_YMM avx2
|
|
|
|
cglobal pixelutils_sad_32x32, 4,7,5, src1, stride1, src2, stride2
|
|
|
|
pxor m0, m0
|
|
|
|
mov r4d, 32/4
|
|
|
|
lea r5, [stride1q * 3]
|
|
|
|
lea r6, [stride2q * 3]
|
|
|
|
|
|
|
|
.loop:
|
|
|
|
movu m1, [src1q] ; row 0 of pix0
|
|
|
|
movu m2, [src2q] ; row 0 of pix1
|
|
|
|
movu m3, [src1q + stride1q] ; row 1 of pix0
|
|
|
|
movu m4, [src2q + stride2q] ; row 1 of pix1
|
|
|
|
|
|
|
|
psadbw m1, m2
|
|
|
|
psadbw m3, m4
|
|
|
|
paddd m0, m1
|
|
|
|
paddd m0, m3
|
|
|
|
|
|
|
|
movu m1, [src1q + 2 * stride1q] ; row 2 of pix0
|
|
|
|
movu m2, [src2q + 2 * stride2q] ; row 2 of pix1
|
|
|
|
movu m3, [src1q + r5] ; row 3 of pix0
|
|
|
|
movu m4, [src2q + r6] ; row 3 of pix1
|
|
|
|
|
|
|
|
psadbw m1, m2
|
|
|
|
psadbw m3, m4
|
|
|
|
paddd m0, m1
|
|
|
|
paddd m0, m3
|
|
|
|
|
|
|
|
lea src2q, [src2q + 4 * stride2q]
|
|
|
|
lea src1q, [src1q + 4 * stride1q]
|
|
|
|
|
|
|
|
dec r4d
|
|
|
|
jnz .loop
|
|
|
|
|
|
|
|
vextracti128 xm1, m0, 1
|
|
|
|
paddd xm0, xm1
|
|
|
|
pshufd xm1, xm0, 2
|
|
|
|
paddd xm0, xm1
|
|
|
|
movd eax, xm0
|
|
|
|
RET
|
|
|
|
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
; int ff_pixelutils_sad_[au]_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
|
|
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
|
|
;-------------------------------------------------------------------------------
|
|
|
|
%macro SAD_AVX2_32x32 1
|
|
|
|
INIT_YMM avx2
|
|
|
|
cglobal pixelutils_sad_%1_32x32, 4,7,3, src1, stride1, src2, stride2
|
|
|
|
pxor m0, m0
|
|
|
|
mov r4d, 32/4
|
|
|
|
lea r5, [stride1q * 3]
|
|
|
|
lea r6, [stride2q * 3]
|
|
|
|
|
|
|
|
.loop:
|
|
|
|
mov%1 m1, [src2q] ; row 0 of pix1
|
|
|
|
psadbw m1, [src1q]
|
|
|
|
mov%1 m2, [src2q + stride2q] ; row 1 of pix1
|
|
|
|
psadbw m2, [src1q + stride1q]
|
|
|
|
|
|
|
|
paddd m0, m1
|
|
|
|
paddd m0, m2
|
|
|
|
|
|
|
|
mov%1 m1, [src2q + 2 * stride2q] ; row 2 of pix1
|
|
|
|
psadbw m1, [src1q + 2 * stride1q]
|
|
|
|
mov%1 m2, [src2q + r6] ; row 3 of pix1
|
|
|
|
psadbw m2, [src1q + r5]
|
|
|
|
|
|
|
|
paddd m0, m1
|
|
|
|
paddd m0, m2
|
|
|
|
|
|
|
|
lea src2q, [src2q + 4 * stride2q]
|
|
|
|
lea src1q, [src1q + 4 * stride1q]
|
|
|
|
|
|
|
|
dec r4d
|
|
|
|
jnz .loop
|
|
|
|
|
|
|
|
vextracti128 xm1, m0, 1
|
|
|
|
paddd xm0, xm1
|
|
|
|
pshufd xm1, xm0, 2
|
|
|
|
paddd xm0, xm1
|
|
|
|
movd eax, xm0
|
|
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
SAD_AVX2_32x32 a
|
|
|
|
SAD_AVX2_32x32 u
|
|
|
|
%endif
|