mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
429 lines
13 KiB
429 lines
13 KiB
;****************************************************************************** |
|
;* x86-SIMD-optimized IDCT for prores |
|
;* this is identical to "simple" IDCT except for the clip range |
|
;* |
|
;* Copyright (c) 2011 Ronald S. Bultje <rsbultje@gmail.com> |
|
;* |
|
;* This file is part of Libav. |
|
;* |
|
;* Libav is free software; you can redistribute it and/or |
|
;* modify it under the terms of the GNU Lesser General Public |
|
;* License as published by the Free Software Foundation; either |
|
;* version 2.1 of the License, or (at your option) any later version. |
|
;* |
|
;* Libav is distributed in the hope that it will be useful, |
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
;* Lesser General Public License for more details. |
|
;* |
|
;* You should have received a copy of the GNU Lesser General Public |
|
;* License along with Libav; if not, write to the Free Software |
|
;* 51, Inc., Foundation Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
;****************************************************************************** |
|
|
|
%include "x86inc.asm" |
|
%include "x86util.asm" |
|
|
|
%define W1sh2 22725 ; W1 = 90901 = 22725<<2 + 1 |
|
%define W2sh2 21407 ; W2 = 85627 = 21407<<2 - 1 |
|
%define W3sh2 19265 ; W3 = 77062 = 19265<<2 + 2 |
|
%define W4sh2 16384 ; W4 = 65535 = 16384<<2 - 1 |
|
%define W5sh2 12873 ; W5 = 51491 = 12873<<2 - 1 |
|
%define W6sh2 8867 ; W6 = 35468 = 8867<<2 |
|
%define W7sh2 4520 ; W7 = 18081 = 4520<<2 + 1 |
|
|
|
%if ARCH_X86_64 |
|
|
|
SECTION_RODATA |
|
|
|
w4_plus_w2: times 4 dw W4sh2, +W2sh2 |
|
w4_min_w2: times 4 dw W4sh2, -W2sh2 |
|
w4_plus_w6: times 4 dw W4sh2, +W6sh2 |
|
w4_min_w6: times 4 dw W4sh2, -W6sh2 |
|
w1_plus_w3: times 4 dw W1sh2, +W3sh2 |
|
w3_min_w1: times 4 dw W3sh2, -W1sh2 |
|
w7_plus_w3: times 4 dw W7sh2, +W3sh2 |
|
w3_min_w7: times 4 dw W3sh2, -W7sh2 |
|
w1_plus_w5: times 4 dw W1sh2, +W5sh2 |
|
w5_min_w1: times 4 dw W5sh2, -W1sh2 |
|
w5_plus_w7: times 4 dw W5sh2, +W7sh2 |
|
w7_min_w5: times 4 dw W7sh2, -W5sh2 |
|
row_round: times 8 dw (1<<14) |
|
|
|
cextern pw_4 |
|
cextern pw_8 |
|
cextern pw_512 |
|
cextern pw_1019 |
|
|
|
section .text align=16 |
|
|
|
; interleave data while maintaining source |
|
; %1=type, %2=dstlo, %3=dsthi, %4=src, %5=interleave |
|
%macro SBUTTERFLY3 5 |
|
punpckl%1 m%2, m%4, m%5 |
|
punpckh%1 m%3, m%4, m%5 |
|
%endmacro |
|
|
|
; %1/%2=src1/dst1, %3/%4=dst2, %5/%6=src2, %7=shift |
|
; action: %3/%4 = %1/%2 - %5/%6; %1/%2 += %5/%6 |
|
; %1/%2/%3/%4 >>= %7; dword -> word (in %1/%3) |
|
%macro SUMSUB_SHPK 7 |
|
psubd %3, %1, %5 ; { a0 - b0 }[0-3] |
|
psubd %4, %2, %6 ; { a0 - b0 }[4-7] |
|
paddd %1, %5 ; { a0 + b0 }[0-3] |
|
paddd %2, %6 ; { a0 + b0 }[4-7] |
|
psrad %1, %7 |
|
psrad %2, %7 |
|
psrad %3, %7 |
|
psrad %4, %7 |
|
packssdw %1, %2 ; row[0] |
|
packssdw %3, %4 ; row[7] |
|
%endmacro |
|
|
|
; %1 = row or col (for rounding variable) |
|
; %2 = number of bits to shift at the end |
|
%macro IDCT_1D 2 |
|
; a0 = (W4 * row[0]) + (1 << (15 - 1)); |
|
; a1 = a0; |
|
; a2 = a0; |
|
; a3 = a0; |
|
; a0 += W2 * row[2]; |
|
; a1 += W6 * row[2]; |
|
; a2 -= W6 * row[2]; |
|
; a3 -= W2 * row[2]; |
|
%ifidn %1, col |
|
paddw m10,[pw_8] |
|
%endif |
|
SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[0], row[2] }[0-3]/[4-7] |
|
%ifidn %1, row |
|
psubw m10,[row_round] |
|
%endif |
|
SIGNEXTEND m8, m9, m14 ; { row[2] }[0-3] / [4-7] |
|
SIGNEXTEND m10, m11, m14 ; { row[0] }[0-3] / [4-7] |
|
pmaddwd m2, m0, [w4_plus_w6] |
|
pmaddwd m3, m1, [w4_plus_w6] |
|
pmaddwd m4, m0, [w4_min_w6] |
|
pmaddwd m5, m1, [w4_min_w6] |
|
pmaddwd m6, m0, [w4_min_w2] |
|
pmaddwd m7, m1, [w4_min_w2] |
|
pmaddwd m0, [w4_plus_w2] |
|
pmaddwd m1, [w4_plus_w2] |
|
pslld m2, 2 |
|
pslld m3, 2 |
|
pslld m4, 2 |
|
pslld m5, 2 |
|
pslld m6, 2 |
|
pslld m7, 2 |
|
pslld m0, 2 |
|
pslld m1, 2 |
|
|
|
; a0: -1*row[0]-1*row[2] |
|
; a1: -1*row[0] |
|
; a2: -1*row[0] |
|
; a3: -1*row[0]+1*row[2] |
|
psubd m2, m10 ; a1[0-3] |
|
psubd m3, m11 ; a1[4-7] |
|
psubd m4, m10 ; a2[0-3] |
|
psubd m5, m11 ; a2[4-7] |
|
psubd m0, m10 |
|
psubd m1, m11 |
|
psubd m6, m10 |
|
psubd m7, m11 |
|
psubd m0, m8 ; a0[0-3] |
|
psubd m1, m9 ; a0[4-7] |
|
paddd m6, m8 ; a3[0-3] |
|
paddd m7, m9 ; a3[4-7] |
|
|
|
; a0 += W4*row[4] + W6*row[6]; i.e. -1*row[4] |
|
; a1 -= W4*row[4] + W2*row[6]; i.e. -1*row[4]-1*row[6] |
|
; a2 -= W4*row[4] - W2*row[6]; i.e. -1*row[4]+1*row[6] |
|
; a3 += W4*row[4] - W6*row[6]; i.e. -1*row[4] |
|
SBUTTERFLY3 wd, 8, 9, 13, 12 ; { row[4], row[6] }[0-3]/[4-7] |
|
SIGNEXTEND m13, m14, m10 ; { row[4] }[0-3] / [4-7] |
|
pmaddwd m10, m8, [w4_plus_w6] |
|
pmaddwd m11, m9, [w4_plus_w6] |
|
pslld m10, 2 |
|
pslld m11, 2 |
|
psubd m10, m13 |
|
psubd m11, m14 |
|
paddd m0, m10 ; a0[0-3] |
|
paddd m1, m11 ; a0[4-7] |
|
pmaddwd m10, m8, [w4_min_w6] |
|
pmaddwd m11, m9, [w4_min_w6] |
|
pslld m10, 2 |
|
pslld m11, 2 |
|
psubd m10, m13 |
|
psubd m11, m14 |
|
paddd m6, m10 ; a3[0-3] |
|
paddd m7, m11 ; a3[4-7] |
|
pmaddwd m10, m8, [w4_min_w2] |
|
pmaddwd m11, m9, [w4_min_w2] |
|
pmaddwd m8, [w4_plus_w2] |
|
pmaddwd m9, [w4_plus_w2] |
|
pslld m10, 2 |
|
pslld m11, 2 |
|
pslld m8, 2 |
|
pslld m9, 2 |
|
psubd m10, m13 |
|
psubd m11, m14 |
|
psubd m8, m13 |
|
psubd m9, m14 |
|
psubd m4, m10 ; a2[0-3] intermediate |
|
psubd m5, m11 ; a2[4-7] intermediate |
|
psubd m2, m8 ; a1[0-3] intermediate |
|
psubd m3, m9 ; a1[4-7] intermediate |
|
SIGNEXTEND m12, m13, m10 ; { row[6] }[0-3] / [4-7] |
|
psubd m4, m12 ; a2[0-3] |
|
psubd m5, m13 ; a2[4-7] |
|
paddd m2, m12 ; a1[0-3] |
|
paddd m3, m13 ; a1[4-7] |
|
|
|
; load/store |
|
mova [r2+ 0], m0 |
|
mova [r2+ 32], m2 |
|
mova [r2+ 64], m4 |
|
mova [r2+ 96], m6 |
|
mova m10,[r2+ 16] ; { row[1] }[0-7] |
|
mova m8, [r2+ 48] ; { row[3] }[0-7] |
|
mova m13,[r2+ 80] ; { row[5] }[0-7] |
|
mova m14,[r2+112] ; { row[7] }[0-7] |
|
mova [r2+ 16], m1 |
|
mova [r2+ 48], m3 |
|
mova [r2+ 80], m5 |
|
mova [r2+112], m7 |
|
%ifidn %1, row |
|
pmullw m10,[r3+ 16] |
|
pmullw m8, [r3+ 48] |
|
pmullw m13,[r3+ 80] |
|
pmullw m14,[r3+112] |
|
%endif |
|
|
|
; b0 = MUL(W1, row[1]); |
|
; MAC(b0, W3, row[3]); |
|
; b1 = MUL(W3, row[1]); |
|
; MAC(b1, -W7, row[3]); |
|
; b2 = MUL(W5, row[1]); |
|
; MAC(b2, -W1, row[3]); |
|
; b3 = MUL(W7, row[1]); |
|
; MAC(b3, -W5, row[3]); |
|
SBUTTERFLY3 wd, 0, 1, 10, 8 ; { row[1], row[3] }[0-3]/[4-7] |
|
SIGNEXTEND m10, m11, m12 ; { row[1] }[0-3] / [4-7] |
|
SIGNEXTEND m8, m9, m12 ; { row[3] }[0-3] / [4-7] |
|
pmaddwd m2, m0, [w3_min_w7] |
|
pmaddwd m3, m1, [w3_min_w7] |
|
pmaddwd m4, m0, [w5_min_w1] |
|
pmaddwd m5, m1, [w5_min_w1] |
|
pmaddwd m6, m0, [w7_min_w5] |
|
pmaddwd m7, m1, [w7_min_w5] |
|
pmaddwd m0, [w1_plus_w3] |
|
pmaddwd m1, [w1_plus_w3] |
|
pslld m2, 2 |
|
pslld m3, 2 |
|
pslld m4, 2 |
|
pslld m5, 2 |
|
pslld m6, 2 |
|
pslld m7, 2 |
|
pslld m0, 2 |
|
pslld m1, 2 |
|
|
|
; b0: +1*row[1]+2*row[3] |
|
; b1: +2*row[1]-1*row[3] |
|
; b2: -1*row[1]-1*row[3] |
|
; b3: +1*row[1]+1*row[3] |
|
psubd m2, m8 |
|
psubd m3, m9 |
|
paddd m0, m8 |
|
paddd m1, m9 |
|
paddd m8, m10 ; { row[1] + row[3] }[0-3] |
|
paddd m9, m11 ; { row[1] + row[3] }[4-7] |
|
paddd m10, m10 |
|
paddd m11, m11 |
|
paddd m0, m8 ; b0[0-3] |
|
paddd m1, m9 ; b0[4-7] |
|
paddd m2, m10 ; b1[0-3] |
|
paddd m3, m11 ; b2[4-7] |
|
psubd m4, m8 ; b2[0-3] |
|
psubd m5, m9 ; b2[4-7] |
|
paddd m6, m8 ; b3[0-3] |
|
paddd m7, m9 ; b3[4-7] |
|
|
|
; MAC(b0, W5, row[5]); |
|
; MAC(b0, W7, row[7]); |
|
; MAC(b1, -W1, row[5]); |
|
; MAC(b1, -W5, row[7]); |
|
; MAC(b2, W7, row[5]); |
|
; MAC(b2, W3, row[7]); |
|
; MAC(b3, W3, row[5]); |
|
; MAC(b3, -W1, row[7]); |
|
SBUTTERFLY3 wd, 8, 9, 13, 14 ; { row[5], row[7] }[0-3]/[4-7] |
|
SIGNEXTEND m13, m12, m11 ; { row[5] }[0-3] / [4-7] |
|
SIGNEXTEND m14, m11, m10 ; { row[7] }[0-3] / [4-7] |
|
|
|
; b0: -1*row[5]+1*row[7] |
|
; b1: -1*row[5]+1*row[7] |
|
; b2: +1*row[5]+2*row[7] |
|
; b3: +2*row[5]-1*row[7] |
|
paddd m4, m13 |
|
paddd m5, m12 |
|
paddd m6, m13 |
|
paddd m7, m12 |
|
psubd m13, m14 ; { row[5] - row[7] }[0-3] |
|
psubd m12, m11 ; { row[5] - row[7] }[4-7] |
|
paddd m14, m14 |
|
paddd m11, m11 |
|
psubd m0, m13 |
|
psubd m1, m12 |
|
psubd m2, m13 |
|
psubd m3, m12 |
|
paddd m4, m14 |
|
paddd m5, m11 |
|
paddd m6, m13 |
|
paddd m7, m12 |
|
|
|
pmaddwd m10, m8, [w1_plus_w5] |
|
pmaddwd m11, m9, [w1_plus_w5] |
|
pmaddwd m12, m8, [w5_plus_w7] |
|
pmaddwd m13, m9, [w5_plus_w7] |
|
pslld m10, 2 |
|
pslld m11, 2 |
|
pslld m12, 2 |
|
pslld m13, 2 |
|
psubd m2, m10 ; b1[0-3] |
|
psubd m3, m11 ; b1[4-7] |
|
paddd m0, m12 ; b0[0-3] |
|
paddd m1, m13 ; b0[4-7] |
|
pmaddwd m12, m8, [w7_plus_w3] |
|
pmaddwd m13, m9, [w7_plus_w3] |
|
pmaddwd m8, [w3_min_w1] |
|
pmaddwd m9, [w3_min_w1] |
|
pslld m12, 2 |
|
pslld m13, 2 |
|
pslld m8, 2 |
|
pslld m9, 2 |
|
paddd m4, m12 ; b2[0-3] |
|
paddd m5, m13 ; b2[4-7] |
|
paddd m6, m8 ; b3[0-3] |
|
paddd m7, m9 ; b3[4-7] |
|
|
|
; row[0] = (a0 + b0) >> 15; |
|
; row[7] = (a0 - b0) >> 15; |
|
; row[1] = (a1 + b1) >> 15; |
|
; row[6] = (a1 - b1) >> 15; |
|
; row[2] = (a2 + b2) >> 15; |
|
; row[5] = (a2 - b2) >> 15; |
|
; row[3] = (a3 + b3) >> 15; |
|
; row[4] = (a3 - b3) >> 15; |
|
mova m8, [r2+ 0] ; a0[0-3] |
|
mova m9, [r2+16] ; a0[4-7] |
|
SUMSUB_SHPK m8, m9, m10, m11, m0, m1, %2 |
|
mova m0, [r2+32] ; a1[0-3] |
|
mova m1, [r2+48] ; a1[4-7] |
|
SUMSUB_SHPK m0, m1, m9, m11, m2, m3, %2 |
|
mova m1, [r2+64] ; a2[0-3] |
|
mova m2, [r2+80] ; a2[4-7] |
|
SUMSUB_SHPK m1, m2, m11, m3, m4, m5, %2 |
|
mova m2, [r2+96] ; a3[0-3] |
|
mova m3, [r2+112] ; a3[4-7] |
|
SUMSUB_SHPK m2, m3, m4, m5, m6, m7, %2 |
|
%endmacro |
|
|
|
; void prores_idct_put_10_<opt>(uint8_t *pixels, int stride, |
|
; DCTELEM *block, const int16_t *qmat); |
|
%macro idct_put_fn 1 |
|
cglobal prores_idct_put_10, 4, 4, %1 |
|
movsxd r1, r1d |
|
pxor m15, m15 ; zero |
|
|
|
; for (i = 0; i < 8; i++) |
|
; idctRowCondDC(block + i*8); |
|
mova m10,[r2+ 0] ; { row[0] }[0-7] |
|
mova m8, [r2+32] ; { row[2] }[0-7] |
|
mova m13,[r2+64] ; { row[4] }[0-7] |
|
mova m12,[r2+96] ; { row[6] }[0-7] |
|
|
|
pmullw m10,[r3+ 0] |
|
pmullw m8, [r3+32] |
|
pmullw m13,[r3+64] |
|
pmullw m12,[r3+96] |
|
|
|
IDCT_1D row, 17 |
|
|
|
; transpose for second part of IDCT |
|
TRANSPOSE8x8W 8, 0, 1, 2, 4, 11, 9, 10, 3 |
|
mova [r2+ 16], m0 |
|
mova [r2+ 48], m2 |
|
mova [r2+ 80], m11 |
|
mova [r2+112], m10 |
|
SWAP 8, 10 |
|
SWAP 1, 8 |
|
SWAP 4, 13 |
|
SWAP 9, 12 |
|
|
|
; for (i = 0; i < 8; i++) |
|
; idctSparseColAdd(dest + i, line_size, block + i); |
|
IDCT_1D col, 20 |
|
|
|
; clip/store |
|
mova m6, [pw_512] |
|
mova m3, [pw_4] |
|
mova m5, [pw_1019] |
|
paddw m8, m6 |
|
paddw m0, m6 |
|
paddw m1, m6 |
|
paddw m2, m6 |
|
paddw m4, m6 |
|
paddw m11, m6 |
|
paddw m9, m6 |
|
paddw m10, m6 |
|
pmaxsw m8, m3 |
|
pmaxsw m0, m3 |
|
pmaxsw m1, m3 |
|
pmaxsw m2, m3 |
|
pmaxsw m4, m3 |
|
pmaxsw m11, m3 |
|
pmaxsw m9, m3 |
|
pmaxsw m10, m3 |
|
pminsw m8, m5 |
|
pminsw m0, m5 |
|
pminsw m1, m5 |
|
pminsw m2, m5 |
|
pminsw m4, m5 |
|
pminsw m11, m5 |
|
pminsw m9, m5 |
|
pminsw m10, m5 |
|
|
|
lea r2, [r1*3] |
|
mova [r0 ], m8 |
|
mova [r0+r1 ], m0 |
|
mova [r0+r1*2], m1 |
|
mova [r0+r2 ], m2 |
|
lea r0, [r0+r1*4] |
|
mova [r0 ], m4 |
|
mova [r0+r1 ], m11 |
|
mova [r0+r1*2], m9 |
|
mova [r0+r2 ], m10 |
|
RET |
|
%endmacro |
|
|
|
%macro SIGNEXTEND 2-3 |
|
%if cpuflag(sse4) ; dstlow, dsthigh |
|
movhlps %2, %1 |
|
pmovsxwd %1, %1 |
|
pmovsxwd %2, %2 |
|
%elif cpuflag(sse2) ; dstlow, dsthigh, tmp |
|
pxor %3, %3 |
|
pcmpgtw %3, %1 |
|
mova %2, %1 |
|
punpcklwd %1, %3 |
|
punpckhwd %2, %3 |
|
%endif |
|
%endmacro |
|
|
|
INIT_XMM sse2 |
|
idct_put_fn 16 |
|
INIT_XMM sse4 |
|
idct_put_fn 16 |
|
INIT_XMM avx |
|
idct_put_fn 16 |
|
|
|
%endif
|
|
|