mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
375 lines
8.3 KiB
375 lines
8.3 KiB
;***************************************************************************** |
|
;* SSE2-optimized weighted prediction code |
|
;***************************************************************************** |
|
;* Copyright (c) 2004-2005 Michael Niedermayer, Loren Merritt |
|
;* Copyright (C) 2010 Eli Friedman <eli.friedman@gmail.com> |
|
;* |
|
;* This file is part of Libav. |
|
;* |
|
;* Libav is free software; you can redistribute it and/or |
|
;* modify it under the terms of the GNU Lesser General Public |
|
;* License as published by the Free Software Foundation; either |
|
;* version 2.1 of the License, or (at your option) any later version. |
|
;* |
|
;* Libav is distributed in the hope that it will be useful, |
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
;* Lesser General Public License for more details. |
|
;* |
|
;* You should have received a copy of the GNU Lesser General Public |
|
;* License along with Libav; if not, write to the Free Software |
|
;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
;****************************************************************************** |
|
|
|
%include "x86inc.asm" |
|
|
|
SECTION .text |
|
|
|
;----------------------------------------------------------------------------- |
|
; biweight pred: |
|
; |
|
; void h264_biweight_16x16_sse2(uint8_t *dst, uint8_t *src, int stride, |
|
; int log2_denom, int weightd, int weights, |
|
; int offset); |
|
; and |
|
; void h264_weight_16x16_sse2(uint8_t *dst, int stride, |
|
; int log2_denom, int weight, |
|
; int offset); |
|
;----------------------------------------------------------------------------- |
|
|
|
%macro WEIGHT_SETUP 0 |
|
add r4, r4 |
|
inc r4 |
|
movd m3, r3d |
|
movd m5, r4d |
|
movd m6, r2d |
|
pslld m5, m6 |
|
psrld m5, 1 |
|
%if mmsize == 16 |
|
pshuflw m3, m3, 0 |
|
pshuflw m5, m5, 0 |
|
punpcklqdq m3, m3 |
|
punpcklqdq m5, m5 |
|
%else |
|
pshufw m3, m3, 0 |
|
pshufw m5, m5, 0 |
|
%endif |
|
pxor m7, m7 |
|
%endmacro |
|
|
|
%macro WEIGHT_OP 2 |
|
movh m0, [r0+%1] |
|
movh m1, [r0+%2] |
|
punpcklbw m0, m7 |
|
punpcklbw m1, m7 |
|
pmullw m0, m3 |
|
pmullw m1, m3 |
|
paddsw m0, m5 |
|
paddsw m1, m5 |
|
psraw m0, m6 |
|
psraw m1, m6 |
|
packuswb m0, m1 |
|
%endmacro |
|
|
|
%macro WEIGHT_FUNC_DBL_MM 1 |
|
cglobal h264_weight_16x%1_mmx2, 5, 5, 0 |
|
WEIGHT_SETUP |
|
mov r2, %1 |
|
%if %1 == 16 |
|
.nextrow |
|
WEIGHT_OP 0, 4 |
|
mova [r0 ], m0 |
|
WEIGHT_OP 8, 12 |
|
mova [r0+8], m0 |
|
add r0, r1 |
|
dec r2 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_weight_16x16_mmx2.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
WEIGHT_FUNC_DBL_MM 16 |
|
WEIGHT_FUNC_DBL_MM 8 |
|
|
|
%macro WEIGHT_FUNC_MM 4 |
|
cglobal h264_weight_%1x%2_%4, 7, 7, %3 |
|
WEIGHT_SETUP |
|
mov r2, %2 |
|
%if %2 == 16 |
|
.nextrow |
|
WEIGHT_OP 0, mmsize/2 |
|
mova [r0], m0 |
|
add r0, r1 |
|
dec r2 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_weight_%1x16_%4.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
WEIGHT_FUNC_MM 8, 16, 0, mmx2 |
|
WEIGHT_FUNC_MM 8, 8, 0, mmx2 |
|
WEIGHT_FUNC_MM 8, 4, 0, mmx2 |
|
INIT_XMM |
|
WEIGHT_FUNC_MM 16, 16, 8, sse2 |
|
WEIGHT_FUNC_MM 16, 8, 8, sse2 |
|
|
|
%macro WEIGHT_FUNC_HALF_MM 5 |
|
cglobal h264_weight_%1x%2_%5, 5, 5, %4 |
|
WEIGHT_SETUP |
|
mov r2, %2/2 |
|
lea r3, [r1*2] |
|
%if %2 == mmsize |
|
.nextrow |
|
WEIGHT_OP 0, r1 |
|
movh [r0], m0 |
|
%if mmsize == 16 |
|
movhps [r0+r1], m0 |
|
%else |
|
psrlq m0, 32 |
|
movh [r0+r1], m0 |
|
%endif |
|
add r0, r3 |
|
dec r2 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_weight_%1x%3_%5.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
WEIGHT_FUNC_HALF_MM 4, 8, 8, 0, mmx2 |
|
WEIGHT_FUNC_HALF_MM 4, 4, 8, 0, mmx2 |
|
WEIGHT_FUNC_HALF_MM 4, 2, 8, 0, mmx2 |
|
INIT_XMM |
|
WEIGHT_FUNC_HALF_MM 8, 16, 16, 8, sse2 |
|
WEIGHT_FUNC_HALF_MM 8, 8, 16, 8, sse2 |
|
WEIGHT_FUNC_HALF_MM 8, 4, 16, 8, sse2 |
|
|
|
%macro BIWEIGHT_SETUP 0 |
|
add r6, 1 |
|
or r6, 1 |
|
add r3, 1 |
|
movd m3, r4d |
|
movd m4, r5d |
|
movd m5, r6d |
|
movd m6, r3d |
|
pslld m5, m6 |
|
psrld m5, 1 |
|
%if mmsize == 16 |
|
pshuflw m3, m3, 0 |
|
pshuflw m4, m4, 0 |
|
pshuflw m5, m5, 0 |
|
punpcklqdq m3, m3 |
|
punpcklqdq m4, m4 |
|
punpcklqdq m5, m5 |
|
%else |
|
pshufw m3, m3, 0 |
|
pshufw m4, m4, 0 |
|
pshufw m5, m5, 0 |
|
%endif |
|
pxor m7, m7 |
|
%endmacro |
|
|
|
%macro BIWEIGHT_STEPA 3 |
|
movh m%1, [r0+%3] |
|
movh m%2, [r1+%3] |
|
punpcklbw m%1, m7 |
|
punpcklbw m%2, m7 |
|
pmullw m%1, m3 |
|
pmullw m%2, m4 |
|
paddsw m%1, m%2 |
|
%endmacro |
|
|
|
%macro BIWEIGHT_STEPB 0 |
|
paddsw m0, m5 |
|
paddsw m1, m5 |
|
psraw m0, m6 |
|
psraw m1, m6 |
|
packuswb m0, m1 |
|
%endmacro |
|
|
|
%macro BIWEIGHT_FUNC_DBL_MM 1 |
|
cglobal h264_biweight_16x%1_mmx2, 7, 7, 0 |
|
BIWEIGHT_SETUP |
|
mov r3, %1 |
|
%if %1 == 16 |
|
.nextrow |
|
BIWEIGHT_STEPA 0, 1, 0 |
|
BIWEIGHT_STEPA 1, 2, 4 |
|
BIWEIGHT_STEPB |
|
mova [r0], m0 |
|
BIWEIGHT_STEPA 0, 1, 8 |
|
BIWEIGHT_STEPA 1, 2, 12 |
|
BIWEIGHT_STEPB |
|
mova [r0+8], m0 |
|
add r0, r2 |
|
add r1, r2 |
|
dec r3 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_biweight_16x16_mmx2.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
BIWEIGHT_FUNC_DBL_MM 16 |
|
BIWEIGHT_FUNC_DBL_MM 8 |
|
|
|
%macro BIWEIGHT_FUNC_MM 4 |
|
cglobal h264_biweight_%1x%2_%4, 7, 7, %3 |
|
BIWEIGHT_SETUP |
|
mov r3, %2 |
|
%if %2 == 16 |
|
.nextrow |
|
BIWEIGHT_STEPA 0, 1, 0 |
|
BIWEIGHT_STEPA 1, 2, mmsize/2 |
|
BIWEIGHT_STEPB |
|
mova [r0], m0 |
|
add r0, r2 |
|
add r1, r2 |
|
dec r3 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_biweight_%1x16_%4.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
BIWEIGHT_FUNC_MM 8, 16, 0, mmx2 |
|
BIWEIGHT_FUNC_MM 8, 8, 0, mmx2 |
|
BIWEIGHT_FUNC_MM 8, 4, 0, mmx2 |
|
INIT_XMM |
|
BIWEIGHT_FUNC_MM 16, 16, 8, sse2 |
|
BIWEIGHT_FUNC_MM 16, 8, 8, sse2 |
|
|
|
%macro BIWEIGHT_FUNC_HALF_MM 5 |
|
cglobal h264_biweight_%1x%2_%5, 7, 7, %4 |
|
BIWEIGHT_SETUP |
|
mov r3, %2/2 |
|
lea r4, [r2*2] |
|
%if %2 == mmsize |
|
.nextrow |
|
BIWEIGHT_STEPA 0, 1, 0 |
|
BIWEIGHT_STEPA 1, 2, r2 |
|
BIWEIGHT_STEPB |
|
movh [r0], m0 |
|
%if mmsize == 16 |
|
movhps [r0+r2], m0 |
|
%else |
|
psrlq m0, 32 |
|
movh [r0+r2], m0 |
|
%endif |
|
add r0, r4 |
|
add r1, r4 |
|
dec r3 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_biweight_%1x%3_%5.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_MMX |
|
BIWEIGHT_FUNC_HALF_MM 4, 8, 8, 0, mmx2 |
|
BIWEIGHT_FUNC_HALF_MM 4, 4, 8, 0, mmx2 |
|
BIWEIGHT_FUNC_HALF_MM 4, 2, 8, 0, mmx2 |
|
INIT_XMM |
|
BIWEIGHT_FUNC_HALF_MM 8, 16, 16, 8, sse2 |
|
BIWEIGHT_FUNC_HALF_MM 8, 8, 16, 8, sse2 |
|
BIWEIGHT_FUNC_HALF_MM 8, 4, 16, 8, sse2 |
|
|
|
%macro BIWEIGHT_SSSE3_SETUP 0 |
|
add r6, 1 |
|
or r6, 1 |
|
add r3, 1 |
|
movd m4, r4d |
|
movd m0, r5d |
|
movd m5, r6d |
|
movd m6, r3d |
|
pslld m5, m6 |
|
psrld m5, 1 |
|
punpcklbw m4, m0 |
|
pshuflw m4, m4, 0 |
|
pshuflw m5, m5, 0 |
|
punpcklqdq m4, m4 |
|
punpcklqdq m5, m5 |
|
%endmacro |
|
|
|
%macro BIWEIGHT_SSSE3_OP 0 |
|
pmaddubsw m0, m4 |
|
pmaddubsw m2, m4 |
|
paddsw m0, m5 |
|
paddsw m2, m5 |
|
psraw m0, m6 |
|
psraw m2, m6 |
|
packuswb m0, m2 |
|
%endmacro |
|
|
|
%macro BIWEIGHT_SSSE3_16 1 |
|
cglobal h264_biweight_16x%1_ssse3, 7, 7, 8 |
|
BIWEIGHT_SSSE3_SETUP |
|
mov r3, %1 |
|
|
|
%if %1 == 16 |
|
.nextrow |
|
movh m0, [r0] |
|
movh m2, [r0+8] |
|
movh m3, [r1+8] |
|
punpcklbw m0, [r1] |
|
punpcklbw m2, m3 |
|
BIWEIGHT_SSSE3_OP |
|
mova [r0], m0 |
|
add r0, r2 |
|
add r1, r2 |
|
dec r3 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_biweight_16x16_ssse3.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_XMM |
|
BIWEIGHT_SSSE3_16 16 |
|
BIWEIGHT_SSSE3_16 8 |
|
|
|
%macro BIWEIGHT_SSSE3_8 1 |
|
cglobal h264_biweight_8x%1_ssse3, 7, 7, 8 |
|
BIWEIGHT_SSSE3_SETUP |
|
mov r3, %1/2 |
|
lea r4, [r2*2] |
|
|
|
%if %1 == 16 |
|
.nextrow |
|
movh m0, [r0] |
|
movh m1, [r1] |
|
movh m2, [r0+r2] |
|
movh m3, [r1+r2] |
|
punpcklbw m0, m1 |
|
punpcklbw m2, m3 |
|
BIWEIGHT_SSSE3_OP |
|
movh [r0], m0 |
|
movhps [r0+r2], m0 |
|
add r0, r4 |
|
add r1, r4 |
|
dec r3 |
|
jnz .nextrow |
|
REP_RET |
|
%else |
|
jmp mangle(ff_h264_biweight_8x16_ssse3.nextrow) |
|
%endif |
|
%endmacro |
|
|
|
INIT_XMM |
|
BIWEIGHT_SSSE3_8 16 |
|
BIWEIGHT_SSSE3_8 8 |
|
BIWEIGHT_SSSE3_8 4
|
|
|