mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
359 lines
13 KiB
359 lines
13 KiB
/* |
|
* Copyright (C) 2001-2003 Michael Niedermayer <michaelni@gmx.at> |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include "../swscale_internal.h" |
|
#include "libavutil/x86/asm.h" |
|
#include "libavutil/x86/cpu.h" |
|
|
|
#define RET 0xC3 // near return opcode for x86 |
|
#define PREFETCH "prefetchnta" |
|
|
|
#if HAVE_INLINE_ASM |
|
av_cold int ff_init_hscaler_mmxext(int dstW, int xInc, uint8_t *filterCode, |
|
int16_t *filter, int32_t *filterPos, |
|
int numSplits) |
|
{ |
|
uint8_t *fragmentA; |
|
x86_reg imm8OfPShufW1A; |
|
x86_reg imm8OfPShufW2A; |
|
x86_reg fragmentLengthA; |
|
uint8_t *fragmentB; |
|
x86_reg imm8OfPShufW1B; |
|
x86_reg imm8OfPShufW2B; |
|
x86_reg fragmentLengthB; |
|
int fragmentPos; |
|
|
|
int xpos, i; |
|
|
|
// create an optimized horizontal scaling routine |
|
/* This scaler is made of runtime-generated MMXEXT code using specially tuned |
|
* pshufw instructions. For every four output pixels, if four input pixels |
|
* are enough for the fast bilinear scaling, then a chunk of fragmentB is |
|
* used. If five input pixels are needed, then a chunk of fragmentA is used. |
|
*/ |
|
|
|
// code fragment |
|
|
|
__asm__ volatile ( |
|
"jmp 9f \n\t" |
|
// Begin |
|
"0: \n\t" |
|
"movq (%%"REG_d", %%"REG_a"), %%mm3 \n\t" |
|
"movd (%%"REG_c", %%"REG_S"), %%mm0 \n\t" |
|
"movd 1(%%"REG_c", %%"REG_S"), %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm0 \n\t" |
|
"pshufw $0xFF, %%mm1, %%mm1 \n\t" |
|
"1: \n\t" |
|
"pshufw $0xFF, %%mm0, %%mm0 \n\t" |
|
"2: \n\t" |
|
"psubw %%mm1, %%mm0 \n\t" |
|
"movl 8(%%"REG_b", %%"REG_a"), %%esi \n\t" |
|
"pmullw %%mm3, %%mm0 \n\t" |
|
"psllw $7, %%mm1 \n\t" |
|
"paddw %%mm1, %%mm0 \n\t" |
|
|
|
"movq %%mm0, (%%"REG_D", %%"REG_a") \n\t" |
|
|
|
"add $8, %%"REG_a" \n\t" |
|
// End |
|
"9: \n\t" |
|
"lea " LOCAL_MANGLE(0b) ", %0 \n\t" |
|
"lea " LOCAL_MANGLE(1b) ", %1 \n\t" |
|
"lea " LOCAL_MANGLE(2b) ", %2 \n\t" |
|
"dec %1 \n\t" |
|
"dec %2 \n\t" |
|
"sub %0, %1 \n\t" |
|
"sub %0, %2 \n\t" |
|
"lea " LOCAL_MANGLE(9b) ", %3 \n\t" |
|
"sub %0, %3 \n\t" |
|
|
|
|
|
: "=r" (fragmentA), "=r" (imm8OfPShufW1A), "=r" (imm8OfPShufW2A), |
|
"=r" (fragmentLengthA) |
|
); |
|
|
|
__asm__ volatile ( |
|
"jmp 9f \n\t" |
|
// Begin |
|
"0: \n\t" |
|
"movq (%%"REG_d", %%"REG_a"), %%mm3 \n\t" |
|
"movd (%%"REG_c", %%"REG_S"), %%mm0 \n\t" |
|
"punpcklbw %%mm7, %%mm0 \n\t" |
|
"pshufw $0xFF, %%mm0, %%mm1 \n\t" |
|
"1: \n\t" |
|
"pshufw $0xFF, %%mm0, %%mm0 \n\t" |
|
"2: \n\t" |
|
"psubw %%mm1, %%mm0 \n\t" |
|
"movl 8(%%"REG_b", %%"REG_a"), %%esi \n\t" |
|
"pmullw %%mm3, %%mm0 \n\t" |
|
"psllw $7, %%mm1 \n\t" |
|
"paddw %%mm1, %%mm0 \n\t" |
|
|
|
"movq %%mm0, (%%"REG_D", %%"REG_a") \n\t" |
|
|
|
"add $8, %%"REG_a" \n\t" |
|
// End |
|
"9: \n\t" |
|
"lea " LOCAL_MANGLE(0b) ", %0 \n\t" |
|
"lea " LOCAL_MANGLE(1b) ", %1 \n\t" |
|
"lea " LOCAL_MANGLE(2b) ", %2 \n\t" |
|
"dec %1 \n\t" |
|
"dec %2 \n\t" |
|
"sub %0, %1 \n\t" |
|
"sub %0, %2 \n\t" |
|
"lea " LOCAL_MANGLE(9b) ", %3 \n\t" |
|
"sub %0, %3 \n\t" |
|
|
|
|
|
: "=r" (fragmentB), "=r" (imm8OfPShufW1B), "=r" (imm8OfPShufW2B), |
|
"=r" (fragmentLengthB) |
|
); |
|
|
|
xpos = 0; // lumXInc/2 - 0x8000; // difference between pixel centers |
|
fragmentPos = 0; |
|
|
|
for (i = 0; i < dstW / numSplits; i++) { |
|
int xx = xpos >> 16; |
|
|
|
if ((i & 3) == 0) { |
|
int a = 0; |
|
int b = ((xpos + xInc) >> 16) - xx; |
|
int c = ((xpos + xInc * 2) >> 16) - xx; |
|
int d = ((xpos + xInc * 3) >> 16) - xx; |
|
int inc = (d + 1 < 4); |
|
uint8_t *fragment = inc ? fragmentB : fragmentA; |
|
x86_reg imm8OfPShufW1 = inc ? imm8OfPShufW1B : imm8OfPShufW1A; |
|
x86_reg imm8OfPShufW2 = inc ? imm8OfPShufW2B : imm8OfPShufW2A; |
|
x86_reg fragmentLength = inc ? fragmentLengthB : fragmentLengthA; |
|
int maxShift = 3 - (d + inc); |
|
int shift = 0; |
|
|
|
if (filterCode) { |
|
filter[i] = ((xpos & 0xFFFF) ^ 0xFFFF) >> 9; |
|
filter[i + 1] = (((xpos + xInc) & 0xFFFF) ^ 0xFFFF) >> 9; |
|
filter[i + 2] = (((xpos + xInc * 2) & 0xFFFF) ^ 0xFFFF) >> 9; |
|
filter[i + 3] = (((xpos + xInc * 3) & 0xFFFF) ^ 0xFFFF) >> 9; |
|
filterPos[i / 2] = xx; |
|
|
|
memcpy(filterCode + fragmentPos, fragment, fragmentLength); |
|
|
|
filterCode[fragmentPos + imm8OfPShufW1] = (a + inc) | |
|
((b + inc) << 2) | |
|
((c + inc) << 4) | |
|
((d + inc) << 6); |
|
filterCode[fragmentPos + imm8OfPShufW2] = a | (b << 2) | |
|
(c << 4) | |
|
(d << 6); |
|
|
|
if (i + 4 - inc >= dstW) |
|
shift = maxShift; // avoid overread |
|
else if ((filterPos[i / 2] & 3) <= maxShift) |
|
shift = filterPos[i / 2] & 3; // align |
|
|
|
if (shift && i >= shift) { |
|
filterCode[fragmentPos + imm8OfPShufW1] += 0x55 * shift; |
|
filterCode[fragmentPos + imm8OfPShufW2] += 0x55 * shift; |
|
filterPos[i / 2] -= shift; |
|
} |
|
} |
|
|
|
fragmentPos += fragmentLength; |
|
|
|
if (filterCode) |
|
filterCode[fragmentPos] = RET; |
|
} |
|
xpos += xInc; |
|
} |
|
if (filterCode) |
|
filterPos[((i / 2) + 1) & (~1)] = xpos >> 16; // needed to jump to the next part |
|
|
|
return fragmentPos + 1; |
|
} |
|
|
|
void ff_hyscale_fast_mmxext(SwsContext *c, int16_t *dst, |
|
int dstWidth, const uint8_t *src, |
|
int srcW, int xInc) |
|
{ |
|
int32_t *filterPos = c->hLumFilterPos; |
|
int16_t *filter = c->hLumFilter; |
|
void *mmxextFilterCode = c->lumMmxextFilterCode; |
|
int i; |
|
#if ARCH_X86_64 |
|
uint64_t retsave; |
|
#else |
|
#if defined(PIC) |
|
uint64_t ebxsave; |
|
#endif |
|
#endif |
|
|
|
__asm__ volatile( |
|
#if ARCH_X86_64 |
|
"mov -8(%%rsp), %%"REG_a" \n\t" |
|
"mov %%"REG_a", %5 \n\t" // retsave |
|
#else |
|
#if defined(PIC) |
|
"mov %%"REG_b", %5 \n\t" // ebxsave |
|
#endif |
|
#endif |
|
"pxor %%mm7, %%mm7 \n\t" |
|
"mov %0, %%"REG_c" \n\t" |
|
"mov %1, %%"REG_D" \n\t" |
|
"mov %2, %%"REG_d" \n\t" |
|
"mov %3, %%"REG_b" \n\t" |
|
"xor %%"REG_a", %%"REG_a" \n\t" // i |
|
PREFETCH" (%%"REG_c") \n\t" |
|
PREFETCH" 32(%%"REG_c") \n\t" |
|
PREFETCH" 64(%%"REG_c") \n\t" |
|
|
|
#if ARCH_X86_64 |
|
#define CALL_MMXEXT_FILTER_CODE \ |
|
"movl (%%"REG_b"), %%esi \n\t"\ |
|
"call *%4 \n\t"\ |
|
"movl (%%"REG_b", %%"REG_a"), %%esi \n\t"\ |
|
"add %%"REG_S", %%"REG_c" \n\t"\ |
|
"add %%"REG_a", %%"REG_D" \n\t"\ |
|
"xor %%"REG_a", %%"REG_a" \n\t"\ |
|
|
|
#else |
|
#define CALL_MMXEXT_FILTER_CODE \ |
|
"movl (%%"REG_b"), %%esi \n\t"\ |
|
"call *%4 \n\t"\ |
|
"addl (%%"REG_b", %%"REG_a"), %%"REG_c" \n\t"\ |
|
"add %%"REG_a", %%"REG_D" \n\t"\ |
|
"xor %%"REG_a", %%"REG_a" \n\t"\ |
|
|
|
#endif /* ARCH_X86_64 */ |
|
|
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
|
|
#if ARCH_X86_64 |
|
"mov %5, %%"REG_a" \n\t" |
|
"mov %%"REG_a", -8(%%rsp) \n\t" |
|
#else |
|
#if defined(PIC) |
|
"mov %5, %%"REG_b" \n\t" |
|
#endif |
|
#endif |
|
:: "m" (src), "m" (dst), "m" (filter), "m" (filterPos), |
|
"m" (mmxextFilterCode) |
|
#if ARCH_X86_64 |
|
,"m"(retsave) |
|
#else |
|
#if defined(PIC) |
|
,"m" (ebxsave) |
|
#endif |
|
#endif |
|
: "%"REG_a, "%"REG_c, "%"REG_d, "%"REG_S, "%"REG_D |
|
#if ARCH_X86_64 || !defined(PIC) |
|
,"%"REG_b |
|
#endif |
|
); |
|
|
|
for (i=dstWidth-1; (i*xInc)>>16 >=srcW-1; i--) |
|
dst[i] = src[srcW-1]*128; |
|
} |
|
|
|
void ff_hcscale_fast_mmxext(SwsContext *c, int16_t *dst1, int16_t *dst2, |
|
int dstWidth, const uint8_t *src1, |
|
const uint8_t *src2, int srcW, int xInc) |
|
{ |
|
int32_t *filterPos = c->hChrFilterPos; |
|
int16_t *filter = c->hChrFilter; |
|
void *mmxextFilterCode = c->chrMmxextFilterCode; |
|
int i; |
|
#if ARCH_X86_64 |
|
DECLARE_ALIGNED(8, uint64_t, retsave); |
|
#else |
|
#if defined(PIC) |
|
DECLARE_ALIGNED(8, uint64_t, ebxsave); |
|
#endif |
|
#endif |
|
__asm__ volatile( |
|
#if ARCH_X86_64 |
|
"mov -8(%%rsp), %%"REG_a" \n\t" |
|
"mov %%"REG_a", %7 \n\t" // retsave |
|
#else |
|
#if defined(PIC) |
|
"mov %%"REG_b", %7 \n\t" // ebxsave |
|
#endif |
|
#endif |
|
"pxor %%mm7, %%mm7 \n\t" |
|
"mov %0, %%"REG_c" \n\t" |
|
"mov %1, %%"REG_D" \n\t" |
|
"mov %2, %%"REG_d" \n\t" |
|
"mov %3, %%"REG_b" \n\t" |
|
"xor %%"REG_a", %%"REG_a" \n\t" // i |
|
PREFETCH" (%%"REG_c") \n\t" |
|
PREFETCH" 32(%%"REG_c") \n\t" |
|
PREFETCH" 64(%%"REG_c") \n\t" |
|
|
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
"xor %%"REG_a", %%"REG_a" \n\t" // i |
|
"mov %5, %%"REG_c" \n\t" // src2 |
|
"mov %6, %%"REG_D" \n\t" // dst2 |
|
PREFETCH" (%%"REG_c") \n\t" |
|
PREFETCH" 32(%%"REG_c") \n\t" |
|
PREFETCH" 64(%%"REG_c") \n\t" |
|
|
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
CALL_MMXEXT_FILTER_CODE |
|
|
|
#if ARCH_X86_64 |
|
"mov %7, %%"REG_a" \n\t" |
|
"mov %%"REG_a", -8(%%rsp) \n\t" |
|
#else |
|
#if defined(PIC) |
|
"mov %7, %%"REG_b" \n\t" |
|
#endif |
|
#endif |
|
:: "m" (src1), "m" (dst1), "m" (filter), "m" (filterPos), |
|
"m" (mmxextFilterCode), "m" (src2), "m"(dst2) |
|
#if ARCH_X86_64 |
|
,"m"(retsave) |
|
#else |
|
#if defined(PIC) |
|
,"m" (ebxsave) |
|
#endif |
|
#endif |
|
: "%"REG_a, "%"REG_c, "%"REG_d, "%"REG_S, "%"REG_D |
|
#if ARCH_X86_64 || !defined(PIC) |
|
,"%"REG_b |
|
#endif |
|
); |
|
|
|
for (i=dstWidth-1; (i*xInc)>>16 >=srcW-1; i--) { |
|
dst1[i] = src1[srcW-1]*128; |
|
dst2[i] = src2[srcW-1]*128; |
|
} |
|
} |
|
#endif //HAVE_INLINE_ASM
|
|
|