|
|
|
;******************************************************************************
|
|
|
|
;*
|
|
|
|
;* Copyright (c) 2000-2001 Fabrice Bellard <fabrice@bellard.org>
|
|
|
|
;* Copyright (c) Nick Kurshev <nickols_k@mail.ru>
|
|
|
|
;* Copyright (c) 2002 Michael Niedermayer <michaelni@gmx.at>
|
|
|
|
;* Copyright (c) 2002 Zdenek Kabelac <kabi@informatics.muni.cz>
|
|
|
|
;* Copyright (c) 2013 Daniel Kang
|
|
|
|
;*
|
|
|
|
;* MMX optimized hpel functions
|
|
|
|
;*
|
|
|
|
;* This file is part of FFmpeg.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
|
|
;* License as published by the Free Software Foundation; either
|
|
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
;*
|
|
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
;* Lesser General Public License for more details.
|
|
|
|
;*
|
|
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
;******************************************************************************
|
|
|
|
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
|
|
|
|
SECTION_RODATA
|
|
|
|
cextern pb_1
|
|
|
|
|
|
|
|
SECTION_TEXT
|
|
|
|
|
|
|
|
; put_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_PIXELS8_X2 0
|
|
|
|
cglobal put_pixels8_x2, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
.loop:
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
add r1, r4
|
|
|
|
add r0, r4
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
add r1, r4
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_PIXELS8_X2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_PIXELS8_X2
|
|
|
|
|
|
|
|
|
|
|
|
; put_pixels16_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_PIXELS_16 0
|
|
|
|
cglobal put_pixels16_x2, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
.loop:
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m2, [r1+8]
|
|
|
|
mova m3, [r1+r2+8]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
PAVGB m2, [r1+9]
|
|
|
|
PAVGB m3, [r1+r2+9]
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
mova [r0+8], m2
|
|
|
|
mova [r0+r2+8], m3
|
|
|
|
add r1, r4
|
|
|
|
add r0, r4
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m2, [r1+8]
|
|
|
|
mova m3, [r1+r2+8]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
PAVGB m2, [r1+9]
|
|
|
|
PAVGB m3, [r1+r2+9]
|
|
|
|
add r1, r4
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
mova [r0+8], m2
|
|
|
|
mova [r0+r2+8], m3
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_PIXELS_16
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_PIXELS_16
|
|
|
|
|
|
|
|
|
|
|
|
; put_no_rnd_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_NO_RND_PIXELS8_X2 0
|
|
|
|
cglobal put_no_rnd_pixels8_x2, 4,5
|
|
|
|
mova m6, [pb_1]
|
|
|
|
lea r4, [r2*2]
|
|
|
|
.loop:
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
mova m1, [r1+1]
|
|
|
|
mova m3, [r1+r2+1]
|
|
|
|
add r1, r4
|
|
|
|
psubusb m0, m6
|
|
|
|
psubusb m2, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m2, m3
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m1, [r1+1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
mova m3, [r1+r2+1]
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
psubusb m0, m6
|
|
|
|
psubusb m2, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m2, m3
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_NO_RND_PIXELS8_X2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_NO_RND_PIXELS8_X2
|
|
|
|
|
|
|
|
|
|
|
|
; put_no_rnd_pixels8_x2_exact(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_NO_RND_PIXELS8_X2_EXACT 0
|
|
|
|
cglobal put_no_rnd_pixels8_x2_exact, 4,5
|
|
|
|
lea r4, [r2*3]
|
|
|
|
pcmpeqb m6, m6
|
|
|
|
.loop:
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
mova m1, [r1+1]
|
|
|
|
mova m3, [r1+r2+1]
|
|
|
|
pxor m0, m6
|
|
|
|
pxor m2, m6
|
|
|
|
pxor m1, m6
|
|
|
|
pxor m3, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m2, m3
|
|
|
|
pxor m0, m6
|
|
|
|
pxor m2, m6
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova m0, [r1+r2*2]
|
|
|
|
mova m1, [r1+r2*2+1]
|
|
|
|
mova m2, [r1+r4]
|
|
|
|
mova m3, [r1+r4+1]
|
|
|
|
pxor m0, m6
|
|
|
|
pxor m1, m6
|
|
|
|
pxor m2, m6
|
|
|
|
pxor m3, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m2, m3
|
|
|
|
pxor m0, m6
|
|
|
|
pxor m2, m6
|
|
|
|
mova [r0+r2*2], m0
|
|
|
|
mova [r0+r4], m2
|
|
|
|
lea r1, [r1+r2*4]
|
|
|
|
lea r0, [r0+r2*4]
|
|
|
|
sub r3d, 4
|
|
|
|
jg .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_NO_RND_PIXELS8_X2_EXACT
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_NO_RND_PIXELS8_X2_EXACT
|
|
|
|
|
|
|
|
|
|
|
|
; put_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_PIXELS8_Y2 0
|
|
|
|
cglobal put_pixels8_y2, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
mova m0, [r1]
|
|
|
|
sub r0, r2
|
|
|
|
.loop:
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m2, [r1+r4]
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m1, m2
|
|
|
|
mova [r0+r2], m0
|
|
|
|
mova [r0+r4], m1
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m0, [r1+r4]
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m2, m1
|
|
|
|
PAVGB m1, m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova [r0+r4], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_PIXELS8_Y2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_PIXELS8_Y2
|
|
|
|
|
|
|
|
|
|
|
|
; put_no_rnd_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_NO_RND_PIXELS8_Y2 0
|
|
|
|
cglobal put_no_rnd_pixels8_y2, 4,5
|
|
|
|
mova m6, [pb_1]
|
|
|
|
lea r4, [r2+r2]
|
|
|
|
mova m0, [r1]
|
|
|
|
sub r0, r2
|
|
|
|
.loop:
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m2, [r1+r4]
|
|
|
|
add r1, r4
|
|
|
|
psubusb m1, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m1, m2
|
|
|
|
mova [r0+r2], m0
|
|
|
|
mova [r0+r4], m1
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m0, [r1+r4]
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
psubusb m1, m6
|
|
|
|
PAVGB m2, m1
|
|
|
|
PAVGB m1, m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova [r0+r4], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_NO_RND_PIXELS8_Y2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_NO_RND_PIXELS8_Y2
|
|
|
|
|
|
|
|
|
|
|
|
; put_no_rnd_pixels8_y2_exact(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro PUT_NO_RND_PIXELS8_Y2_EXACT 0
|
|
|
|
cglobal put_no_rnd_pixels8_y2_exact, 4,5
|
|
|
|
lea r4, [r2*3]
|
|
|
|
mova m0, [r1]
|
|
|
|
pcmpeqb m6, m6
|
|
|
|
add r1, r2
|
|
|
|
pxor m0, m6
|
|
|
|
.loop:
|
|
|
|
mova m1, [r1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
pxor m1, m6
|
|
|
|
pxor m2, m6
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m1, m2
|
|
|
|
pxor m0, m6
|
|
|
|
pxor m1, m6
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
mova m1, [r1+r2*2]
|
|
|
|
mova m0, [r1+r4]
|
|
|
|
pxor m1, m6
|
|
|
|
pxor m0, m6
|
|
|
|
PAVGB m2, m1
|
|
|
|
PAVGB m1, m0
|
|
|
|
pxor m2, m6
|
|
|
|
pxor m1, m6
|
|
|
|
mova [r0+r2*2], m2
|
|
|
|
mova [r0+r4], m1
|
|
|
|
lea r1, [r1+r2*4]
|
|
|
|
lea r0, [r0+r2*4]
|
|
|
|
sub r3d, 4
|
|
|
|
jg .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
PUT_NO_RND_PIXELS8_Y2_EXACT
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
PUT_NO_RND_PIXELS8_Y2_EXACT
|
|
|
|
|
|
|
|
|
|
|
|
; avg_pixels8(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro AVG_PIXELS8 0
|
|
|
|
cglobal avg_pixels8, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
.loop:
|
|
|
|
mova m0, [r0]
|
|
|
|
mova m1, [r0+r2]
|
|
|
|
PAVGB m0, [r1]
|
|
|
|
PAVGB m1, [r1+r2]
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
add r1, r4
|
|
|
|
add r0, r4
|
|
|
|
mova m0, [r0]
|
|
|
|
mova m1, [r0+r2]
|
|
|
|
PAVGB m0, [r1]
|
|
|
|
PAVGB m1, [r1+r2]
|
|
|
|
add r1, r4
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
AVG_PIXELS8
|
|
|
|
|
|
|
|
|
|
|
|
; avg_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro AVG_PIXELS8_X2 0
|
|
|
|
cglobal avg_pixels8_x2, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
.loop:
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m2, [r1+r2+1]
|
|
|
|
PAVGB m0, [r0]
|
|
|
|
PAVGB m2, [r0+r2]
|
|
|
|
add r1, r4
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova m0, [r1]
|
|
|
|
mova m2, [r1+r2]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
PAVGB m2, [r1+r2+1]
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m0, [r0]
|
|
|
|
PAVGB m2, [r0+r2]
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m2
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
AVG_PIXELS8_X2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
AVG_PIXELS8_X2
|
|
|
|
|
|
|
|
|
|
|
|
; avg_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro AVG_PIXELS8_Y2 0
|
|
|
|
cglobal avg_pixels8_y2, 4,5
|
|
|
|
lea r4, [r2*2]
|
|
|
|
mova m0, [r1]
|
|
|
|
sub r0, r2
|
|
|
|
.loop:
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m2, [r1+r4]
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m1, m2
|
|
|
|
mova m3, [r0+r2]
|
|
|
|
mova m4, [r0+r4]
|
|
|
|
PAVGB m0, m3
|
|
|
|
PAVGB m1, m4
|
|
|
|
mova [r0+r2], m0
|
|
|
|
mova [r0+r4], m1
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m0, [r1+r4]
|
|
|
|
PAVGB m2, m1
|
|
|
|
PAVGB m1, m0
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
mova m3, [r0+r2]
|
|
|
|
mova m4, [r0+r4]
|
|
|
|
PAVGB m2, m3
|
|
|
|
PAVGB m1, m4
|
|
|
|
mova [r0+r2], m2
|
|
|
|
mova [r0+r4], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
AVG_PIXELS8_Y2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
AVG_PIXELS8_Y2
|
|
|
|
|
|
|
|
|
|
|
|
; avg_pixels8_xy2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
|
|
%macro AVG_PIXELS8_XY2 0
|
|
|
|
cglobal avg_pixels8_xy2, 4,5
|
|
|
|
mova m6, [pb_1]
|
|
|
|
lea r4, [r2*2]
|
|
|
|
mova m0, [r1]
|
|
|
|
PAVGB m0, [r1+1]
|
|
|
|
.loop:
|
|
|
|
mova m2, [r1+r4]
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
psubusb m2, m6
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
PAVGB m2, [r1+r4+1]
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m0, m1
|
|
|
|
PAVGB m1, m2
|
|
|
|
PAVGB m0, [r0]
|
|
|
|
PAVGB m1, [r0+r2]
|
|
|
|
mova [r0], m0
|
|
|
|
mova [r0+r2], m1
|
|
|
|
mova m1, [r1+r2]
|
|
|
|
mova m0, [r1+r4]
|
|
|
|
PAVGB m1, [r1+r2+1]
|
|
|
|
PAVGB m0, [r1+r4+1]
|
|
|
|
add r0, r4
|
|
|
|
add r1, r4
|
|
|
|
PAVGB m2, m1
|
|
|
|
PAVGB m1, m0
|
|
|
|
PAVGB m2, [r0]
|
|
|
|
PAVGB m1, [r0+r2]
|
|
|
|
mova [r0], m2
|
|
|
|
mova [r0+r2], m1
|
|
|
|
add r0, r4
|
|
|
|
sub r3d, 4
|
|
|
|
jne .loop
|
|
|
|
REP_RET
|
|
|
|
%endmacro
|
|
|
|
|
|
|
|
INIT_MMX mmxext
|
|
|
|
AVG_PIXELS8_XY2
|
|
|
|
INIT_MMX 3dnow
|
|
|
|
AVG_PIXELS8_XY2
|