mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
300 lines
7.9 KiB
300 lines
7.9 KiB
;****************************************************************************** |
|
;* Copyright Nick Kurshev |
|
;* Copyright Michael (michaelni@gmx.at) |
|
;* Copyright 2018 Jokyo Images |
|
;* Copyright Ivo van Poorten |
|
;* |
|
;* This file is part of FFmpeg. |
|
;* |
|
;* FFmpeg is free software; you can redistribute it and/or |
|
;* modify it under the terms of the GNU Lesser General Public |
|
;* License as published by the Free Software Foundation; either |
|
;* version 2.1 of the License, or (at your option) any later version. |
|
;* |
|
;* FFmpeg is distributed in the hope that it will be useful, |
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
;* Lesser General Public License for more details. |
|
;* |
|
;* You should have received a copy of the GNU Lesser General Public |
|
;* License along with FFmpeg; if not, write to the Free Software |
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
;****************************************************************************** |
|
|
|
%include "libavutil/x86/x86util.asm" |
|
|
|
SECTION_RODATA |
|
|
|
pb_mask_shuffle2103_mmx times 8 dw 255 |
|
pb_shuffle2103: db 2, 1, 0, 3, 6, 5, 4, 7, 10, 9, 8, 11, 14, 13, 12, 15 |
|
pb_shuffle0321: db 0, 3, 2, 1, 4, 7, 6, 5, 8, 11, 10, 9, 12, 15, 14, 13 |
|
pb_shuffle1230: db 1, 2, 3, 0, 5, 6, 7, 4, 9, 10, 11, 8, 13, 14, 15, 12 |
|
pb_shuffle3012: db 3, 0, 1, 2, 7, 4, 5, 6, 11, 8, 9, 10, 15, 12, 13, 14 |
|
pb_shuffle3210: db 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 |
|
|
|
SECTION .text |
|
|
|
%macro RSHIFT_COPY 3 |
|
; %1 dst ; %2 src ; %3 shift |
|
%if cpuflag(avx) |
|
psrldq %1, %2, %3 |
|
%else |
|
mova %1, %2 |
|
RSHIFT %1, %3 |
|
%endif |
|
%endmacro |
|
|
|
;------------------------------------------------------------------------------ |
|
; shuffle_bytes_2103_mmext (const uint8_t *src, uint8_t *dst, int src_size) |
|
;------------------------------------------------------------------------------ |
|
INIT_MMX mmxext |
|
cglobal shuffle_bytes_2103, 3, 5, 8, src, dst, w, tmp, x |
|
mova m6, [pb_mask_shuffle2103_mmx] |
|
mova m7, m6 |
|
psllq m7, 8 |
|
|
|
movsxdifnidn wq, wd |
|
mov xq, wq |
|
|
|
add srcq, wq |
|
add dstq, wq |
|
neg wq |
|
|
|
;calc scalar loop |
|
and xq, mmsize*2 -4 |
|
je .loop_simd |
|
|
|
.loop_scalar: |
|
mov tmpb, [srcq + wq + 2] |
|
mov [dstq+wq + 0], tmpb |
|
mov tmpb, [srcq + wq + 1] |
|
mov [dstq+wq + 1], tmpb |
|
mov tmpb, [srcq + wq + 0] |
|
mov [dstq+wq + 2], tmpb |
|
mov tmpb, [srcq + wq + 3] |
|
mov [dstq+wq + 3], tmpb |
|
add wq, 4 |
|
sub xq, 4 |
|
jg .loop_scalar |
|
|
|
;check if src_size < mmsize * 2 |
|
cmp wq, 0 |
|
jge .end |
|
|
|
.loop_simd: |
|
movu m0, [srcq+wq] |
|
movu m1, [srcq+wq+8] |
|
|
|
pshufw m3, m0, 177 |
|
pshufw m5, m1, 177 |
|
|
|
pand m0, m7 |
|
pand m3, m6 |
|
|
|
pand m1, m7 |
|
pand m5, m6 |
|
|
|
por m0, m3 |
|
por m1, m5 |
|
|
|
movu [dstq+wq], m0 |
|
movu [dstq+wq + 8], m1 |
|
|
|
add wq, mmsize*2 |
|
jl .loop_simd |
|
|
|
.end: |
|
RET |
|
|
|
;------------------------------------------------------------------------------ |
|
; shuffle_bytes_## (const uint8_t *src, uint8_t *dst, int src_size) |
|
;------------------------------------------------------------------------------ |
|
; %1-4 index shuffle |
|
%macro SHUFFLE_BYTES 4 |
|
cglobal shuffle_bytes_%1%2%3%4, 3, 5, 2, src, dst, w, tmp, x |
|
VBROADCASTI128 m0, [pb_shuffle%1%2%3%4] |
|
movsxdifnidn wq, wd |
|
mov xq, wq |
|
|
|
add srcq, wq |
|
add dstq, wq |
|
neg wq |
|
|
|
;calc scalar loop |
|
and xq, mmsize-4 |
|
je .loop_simd |
|
|
|
.loop_scalar: |
|
mov tmpb, [srcq + wq + %1] |
|
mov [dstq+wq + 0], tmpb |
|
mov tmpb, [srcq + wq + %2] |
|
mov [dstq+wq + 1], tmpb |
|
mov tmpb, [srcq + wq + %3] |
|
mov [dstq+wq + 2], tmpb |
|
mov tmpb, [srcq + wq + %4] |
|
mov [dstq+wq + 3], tmpb |
|
add wq, 4 |
|
sub xq, 4 |
|
jg .loop_scalar |
|
|
|
;check if src_size < mmsize |
|
cmp wq, 0 |
|
jge .end |
|
|
|
.loop_simd: |
|
movu m1, [srcq+wq] |
|
pshufb m1, m0 |
|
movu [dstq+wq], m1 |
|
add wq, mmsize |
|
jl .loop_simd |
|
|
|
.end: |
|
RET |
|
%endmacro |
|
|
|
INIT_XMM ssse3 |
|
SHUFFLE_BYTES 2, 1, 0, 3 |
|
SHUFFLE_BYTES 0, 3, 2, 1 |
|
SHUFFLE_BYTES 1, 2, 3, 0 |
|
SHUFFLE_BYTES 3, 0, 1, 2 |
|
SHUFFLE_BYTES 3, 2, 1, 0 |
|
|
|
;----------------------------------------------------------------------------------------------- |
|
; uyvytoyuv422(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
; const uint8_t *src, int width, int height, |
|
; int lumStride, int chromStride, int srcStride) |
|
;----------------------------------------------------------------------------------------------- |
|
%macro UYVY_TO_YUV422 0 |
|
cglobal uyvytoyuv422, 9, 14, 8, ydst, udst, vdst, src, w, h, lum_stride, chrom_stride, src_stride, wtwo, whalf, tmp, x, back_w |
|
pxor m0, m0 |
|
pcmpeqw m1, m1 |
|
psrlw m1, 8 |
|
|
|
movsxdifnidn wq, wd |
|
movsxdifnidn lum_strideq, lum_strided |
|
movsxdifnidn chrom_strideq, chrom_strided |
|
movsxdifnidn src_strideq, src_strided |
|
|
|
mov back_wq, wq |
|
mov whalfq, wq |
|
shr whalfq, 1 ; whalf = width / 2 |
|
|
|
lea srcq, [srcq + wq * 2] |
|
add ydstq, wq |
|
add udstq, whalfq |
|
add vdstq, whalfq |
|
|
|
.loop_line: |
|
mov xq, wq |
|
mov wtwoq, wq |
|
add wtwoq, wtwoq ; wtwo = width * 2 |
|
|
|
neg wq |
|
neg wtwoq |
|
neg whalfq |
|
|
|
;calc scalar loop count |
|
and xq, mmsize * 2 - 1 |
|
je .loop_simd |
|
|
|
.loop_scalar: |
|
mov tmpb, [srcq + wtwoq + 0] |
|
mov [udstq + whalfq], tmpb |
|
|
|
mov tmpb, [srcq + wtwoq + 1] |
|
mov [ydstq + wq], tmpb |
|
|
|
mov tmpb, [srcq + wtwoq + 2] |
|
mov [vdstq + whalfq], tmpb |
|
|
|
mov tmpb, [srcq + wtwoq + 3] |
|
mov [ydstq + wq + 1], tmpb |
|
|
|
add wq, 2 |
|
add wtwoq, 4 |
|
add whalfq, 1 |
|
sub xq, 2 |
|
jg .loop_scalar |
|
|
|
; check if simd loop is need |
|
cmp wq, 0 |
|
jge .end_line |
|
|
|
.loop_simd: |
|
movu m2, [srcq + wtwoq ] |
|
movu m3, [srcq + wtwoq + mmsize ] |
|
movu m4, [srcq + wtwoq + mmsize * 2] |
|
movu m5, [srcq + wtwoq + mmsize * 3] |
|
|
|
; extract y part 1 |
|
RSHIFT_COPY m6, m2, 1 ; UYVY UYVY -> YVYU YVY... |
|
pand m6, m1; YxYx YxYx... |
|
|
|
RSHIFT_COPY m7, m3, 1 ; UYVY UYVY -> YVYU YVY... |
|
pand m7, m1 ; YxYx YxYx... |
|
|
|
packuswb m6, m7 ; YYYY YYYY... |
|
movu [ydstq + wq], m6 |
|
|
|
; extract y part 2 |
|
RSHIFT_COPY m6, m4, 1 ; UYVY UYVY -> YVYU YVY... |
|
pand m6, m1; YxYx YxYx... |
|
|
|
RSHIFT_COPY m7, m5, 1 ; UYVY UYVY -> YVYU YVY... |
|
pand m7, m1 ; YxYx YxYx... |
|
|
|
packuswb m6, m7 ; YYYY YYYY... |
|
movu [ydstq + wq + mmsize], m6 |
|
|
|
; extract uv |
|
pand m2, m1 ; UxVx... |
|
pand m3, m1 ; UxVx... |
|
pand m4, m1 ; UxVx... |
|
pand m5, m1 ; UxVx... |
|
|
|
packuswb m2, m3 ; UVUV... |
|
packuswb m4, m5 ; UVUV... |
|
|
|
; U |
|
pand m6, m2, m1 ; UxUx... |
|
pand m7, m4, m1 ; UxUx... |
|
|
|
packuswb m6, m7 ; UUUU |
|
movu [udstq + whalfq], m6 |
|
|
|
|
|
; V |
|
psrlw m2, 8 ; VxVx... |
|
psrlw m4, 8 ; VxVx... |
|
packuswb m2, m4 ; VVVV |
|
movu [vdstq + whalfq], m2 |
|
|
|
add whalfq, mmsize |
|
add wtwoq, mmsize * 4 |
|
add wq, mmsize * 2 |
|
jl .loop_simd |
|
|
|
.end_line: |
|
add srcq, src_strideq |
|
add ydstq, lum_strideq |
|
add udstq, chrom_strideq |
|
add vdstq, chrom_strideq |
|
|
|
;restore initial state of line variable |
|
mov wq, back_wq |
|
mov xq, wq |
|
mov whalfq, wq |
|
shr whalfq, 1 ; whalf = width / 2 |
|
sub hd, 1 |
|
jg .loop_line |
|
|
|
RET |
|
%endmacro |
|
|
|
%if ARCH_X86_64 |
|
INIT_XMM sse2 |
|
UYVY_TO_YUV422 |
|
|
|
INIT_XMM avx |
|
UYVY_TO_YUV422 |
|
%endif
|
|
|