mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1365 lines
35 KiB
1365 lines
35 KiB
/* |
|
* |
|
* rgb2rgb.c, Software RGB to RGB convertor |
|
* pluralize by Software PAL8 to RGB convertor |
|
* Software YUV to YUV convertor |
|
* Software YUV to RGB convertor |
|
* Written by Nick Kurshev. |
|
* palette & yuv & runtime cpu stuff by Michael (michaelni@gmx.at) (under GPL) |
|
*/ |
|
|
|
#undef PREFETCH |
|
#undef MOVNTQ |
|
#undef EMMS |
|
#undef SFENCE |
|
#undef MMREG_SIZE |
|
#undef PREFETCHW |
|
#undef PAVGB |
|
|
|
#ifdef HAVE_SSE2 |
|
#define MMREG_SIZE 16 |
|
#else |
|
#define MMREG_SIZE 8 |
|
#endif |
|
|
|
#ifdef HAVE_3DNOW |
|
#define PREFETCH "prefetch" |
|
#define PREFETCHW "prefetchw" |
|
#define PAVGB "pavgusb" |
|
#elif defined ( HAVE_MMX2 ) |
|
#define PREFETCH "prefetchnta" |
|
#define PREFETCHW "prefetcht0" |
|
#define PAVGB "pavgb" |
|
#else |
|
#define PREFETCH "/nop" |
|
#define PREFETCHW "/nop" |
|
#endif |
|
|
|
#ifdef HAVE_3DNOW |
|
/* On K6 femms is faster of emms. On K7 femms is directly mapped on emms. */ |
|
#define EMMS "femms" |
|
#else |
|
#define EMMS "emms" |
|
#endif |
|
|
|
#ifdef HAVE_MMX2 |
|
#define MOVNTQ "movntq" |
|
#define SFENCE "sfence" |
|
#else |
|
#define MOVNTQ "movq" |
|
#define SFENCE "/nop" |
|
#endif |
|
|
|
static inline void RENAME(rgb24to32)(const uint8_t *src,uint8_t *dst,unsigned src_size) |
|
{ |
|
uint8_t *dest = dst; |
|
const uint8_t *s = src; |
|
const uint8_t *end; |
|
#ifdef HAVE_MMX |
|
uint8_t *mm_end; |
|
#endif |
|
end = s + src_size; |
|
#ifdef HAVE_MMX |
|
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*4))*(MMREG_SIZE*4)); |
|
__asm __volatile("movq %0, %%mm7"::"m"(mask32):"memory"); |
|
if(mm_end == end) mm_end -= MMREG_SIZE*4; |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movd %1, %%mm0\n\t" |
|
"punpckldq 3%1, %%mm0\n\t" |
|
"movd 6%1, %%mm1\n\t" |
|
"punpckldq 9%1, %%mm1\n\t" |
|
"movd 12%1, %%mm2\n\t" |
|
"punpckldq 15%1, %%mm2\n\t" |
|
"movd 18%1, %%mm3\n\t" |
|
"punpckldq 21%1, %%mm3\n\t" |
|
"pand %%mm7, %%mm0\n\t" |
|
"pand %%mm7, %%mm1\n\t" |
|
"pand %%mm7, %%mm2\n\t" |
|
"pand %%mm7, %%mm3\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
MOVNTQ" %%mm1, 8%0\n\t" |
|
MOVNTQ" %%mm2, 16%0\n\t" |
|
MOVNTQ" %%mm3, 24%0" |
|
:"=m"(*dest) |
|
:"m"(*s) |
|
:"memory"); |
|
dest += 32; |
|
s += 24; |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#endif |
|
while(s < end) |
|
{ |
|
*dest++ = *s++; |
|
*dest++ = *s++; |
|
*dest++ = *s++; |
|
*dest++ = 0; |
|
} |
|
} |
|
|
|
static inline void RENAME(rgb32to24)(const uint8_t *src,uint8_t *dst,unsigned src_size) |
|
{ |
|
uint8_t *dest = dst; |
|
const uint8_t *s = src; |
|
const uint8_t *end; |
|
#ifdef HAVE_MMX |
|
uint8_t *mm_end; |
|
#endif |
|
end = s + src_size; |
|
#ifdef HAVE_MMX |
|
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*4))*(MMREG_SIZE*4)); |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movq %1, %%mm0\n\t" |
|
"movq 8%1, %%mm1\n\t" |
|
"movq 16%1, %%mm4\n\t" |
|
"movq 24%1, %%mm5\n\t" |
|
"movq %%mm0, %%mm2\n\t" |
|
"movq %%mm1, %%mm3\n\t" |
|
"movq %%mm4, %%mm6\n\t" |
|
"movq %%mm5, %%mm7\n\t" |
|
"psrlq $8, %%mm2\n\t" |
|
"psrlq $8, %%mm3\n\t" |
|
"psrlq $8, %%mm6\n\t" |
|
"psrlq $8, %%mm7\n\t" |
|
"pand %2, %%mm0\n\t" |
|
"pand %2, %%mm1\n\t" |
|
"pand %2, %%mm4\n\t" |
|
"pand %2, %%mm5\n\t" |
|
"pand %3, %%mm2\n\t" |
|
"pand %3, %%mm3\n\t" |
|
"pand %3, %%mm6\n\t" |
|
"pand %3, %%mm7\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"por %%mm3, %%mm1\n\t" |
|
"por %%mm6, %%mm4\n\t" |
|
"por %%mm7, %%mm5\n\t" |
|
|
|
"movq %%mm1, %%mm2\n\t" |
|
"movq %%mm4, %%mm3\n\t" |
|
"psllq $48, %%mm2\n\t" |
|
"psllq $32, %%mm3\n\t" |
|
"pand %4, %%mm2\n\t" |
|
"pand %5, %%mm3\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"psrlq $16, %%mm1\n\t" |
|
"psrlq $32, %%mm4\n\t" |
|
"psllq $16, %%mm5\n\t" |
|
"por %%mm3, %%mm1\n\t" |
|
"pand %6, %%mm5\n\t" |
|
"por %%mm5, %%mm4\n\t" |
|
|
|
MOVNTQ" %%mm0, %0\n\t" |
|
MOVNTQ" %%mm1, 8%0\n\t" |
|
MOVNTQ" %%mm4, 16%0" |
|
:"=m"(*dest) |
|
:"m"(*s),"m"(mask24l), |
|
"m"(mask24h),"m"(mask24hh),"m"(mask24hhh),"m"(mask24hhhh) |
|
:"memory"); |
|
dest += 24; |
|
s += 32; |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#endif |
|
while(s < end) |
|
{ |
|
*dest++ = *s++; |
|
*dest++ = *s++; |
|
*dest++ = *s++; |
|
s++; |
|
} |
|
} |
|
|
|
/* |
|
Original by Strepto/Astral |
|
ported to gcc & bugfixed : A'rpi |
|
MMX2, 3DNOW optimization by Nick Kurshev |
|
32bit c version, and and&add trick by Michael Niedermayer |
|
*/ |
|
static inline void RENAME(rgb15to16)(const uint8_t *src,uint8_t *dst,unsigned src_size) |
|
{ |
|
#ifdef HAVE_MMX |
|
register const char* s=src+src_size; |
|
register char* d=dst+src_size; |
|
register int offs=-src_size; |
|
__asm __volatile(PREFETCH" %0"::"m"(*(s+offs))); |
|
__asm __volatile( |
|
"movq %0, %%mm4\n\t" |
|
::"m"(mask15s)); |
|
while(offs<0) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movq %1, %%mm0\n\t" |
|
"movq 8%1, %%mm2\n\t" |
|
"movq %%mm0, %%mm1\n\t" |
|
"movq %%mm2, %%mm3\n\t" |
|
"pand %%mm4, %%mm0\n\t" |
|
"pand %%mm4, %%mm2\n\t" |
|
"paddw %%mm1, %%mm0\n\t" |
|
"paddw %%mm3, %%mm2\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
MOVNTQ" %%mm2, 8%0" |
|
:"=m"(*(d+offs)) |
|
:"m"(*(s+offs)) |
|
); |
|
offs+=16; |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
#if 0 |
|
const uint16_t *s1=( uint16_t * )src; |
|
uint16_t *d1=( uint16_t * )dst; |
|
uint16_t *e=((uint8_t *)s1)+src_size; |
|
while( s1<e ){ |
|
register int x=*( s1++ ); |
|
/* rrrrrggggggbbbbb |
|
0rrrrrgggggbbbbb |
|
0111 1111 1110 0000=0x7FE0 |
|
00000000000001 1111=0x001F */ |
|
*( d1++ )=( x&0x001F )|( ( x&0x7FE0 )<<1 ); |
|
} |
|
#else |
|
const unsigned *s1=( unsigned * )src; |
|
unsigned *d1=( unsigned * )dst; |
|
int i; |
|
int size= src_size>>2; |
|
for(i=0; i<size; i++) |
|
{ |
|
register int x= s1[i]; |
|
// d1[i] = x + (x&0x7FE07FE0); //faster but need msbit =0 which might not allways be true |
|
d1[i] = (x&0x7FFF7FFF) + (x&0x7FE07FE0); |
|
|
|
} |
|
#endif |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb32to16)(const uint8_t *src, uint8_t *dst, unsigned src_size) |
|
{ |
|
#ifdef HAVE_MMX |
|
const uint8_t *s = src; |
|
const uint8_t *end,*mm_end; |
|
uint16_t *d = (uint16_t *)dst; |
|
end = s + src_size; |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); |
|
__asm __volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
__asm __volatile( |
|
"movq %0, %%mm7\n\t" |
|
"movq %1, %%mm6\n\t" |
|
::"m"(red_16mask),"m"(green_16mask)); |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movd %1, %%mm0\n\t" |
|
"movd 4%1, %%mm3\n\t" |
|
"punpckldq 8%1, %%mm0\n\t" |
|
"punpckldq 12%1, %%mm3\n\t" |
|
"movq %%mm0, %%mm1\n\t" |
|
"movq %%mm0, %%mm2\n\t" |
|
"movq %%mm3, %%mm4\n\t" |
|
"movq %%mm3, %%mm5\n\t" |
|
"psrlq $3, %%mm0\n\t" |
|
"psrlq $3, %%mm3\n\t" |
|
"pand %2, %%mm0\n\t" |
|
"pand %2, %%mm3\n\t" |
|
"psrlq $5, %%mm1\n\t" |
|
"psrlq $5, %%mm4\n\t" |
|
"pand %%mm6, %%mm1\n\t" |
|
"pand %%mm6, %%mm4\n\t" |
|
"psrlq $8, %%mm2\n\t" |
|
"psrlq $8, %%mm5\n\t" |
|
"pand %%mm7, %%mm2\n\t" |
|
"pand %%mm7, %%mm5\n\t" |
|
"por %%mm1, %%mm0\n\t" |
|
"por %%mm4, %%mm3\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"por %%mm5, %%mm3\n\t" |
|
"psllq $16, %%mm3\n\t" |
|
"por %%mm3, %%mm0\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
:"=m"(*d):"m"(*s),"m"(blue_16mask):"memory"); |
|
d += 4; |
|
s += 16; |
|
} |
|
while(s < end) |
|
{ |
|
const int b= *s++; |
|
const int g= *s++; |
|
const int r= *s++; |
|
*d++ = (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
unsigned j,i,num_pixels=src_size/4; |
|
uint16_t *d = (uint16_t *)dst; |
|
for(i=0,j=0; j<num_pixels; i+=4,j++) |
|
{ |
|
const int b= src[i+0]; |
|
const int g= src[i+1]; |
|
const int r= src[i+2]; |
|
|
|
d[j]= (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
} |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb32to15)(const uint8_t *src, uint8_t *dst, unsigned src_size) |
|
{ |
|
#ifdef HAVE_MMX |
|
const uint8_t *s = src; |
|
const uint8_t *end,*mm_end; |
|
uint16_t *d = (uint16_t *)dst; |
|
end = s + src_size; |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); |
|
__asm __volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
__asm __volatile( |
|
"movq %0, %%mm7\n\t" |
|
"movq %1, %%mm6\n\t" |
|
::"m"(red_15mask),"m"(green_15mask)); |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movd %1, %%mm0\n\t" |
|
"movd 4%1, %%mm3\n\t" |
|
"punpckldq 8%1, %%mm0\n\t" |
|
"punpckldq 12%1, %%mm3\n\t" |
|
"movq %%mm0, %%mm1\n\t" |
|
"movq %%mm0, %%mm2\n\t" |
|
"movq %%mm3, %%mm4\n\t" |
|
"movq %%mm3, %%mm5\n\t" |
|
"psrlq $3, %%mm0\n\t" |
|
"psrlq $3, %%mm3\n\t" |
|
"pand %2, %%mm0\n\t" |
|
"pand %2, %%mm3\n\t" |
|
"psrlq $6, %%mm1\n\t" |
|
"psrlq $6, %%mm4\n\t" |
|
"pand %%mm6, %%mm1\n\t" |
|
"pand %%mm6, %%mm4\n\t" |
|
"psrlq $9, %%mm2\n\t" |
|
"psrlq $9, %%mm5\n\t" |
|
"pand %%mm7, %%mm2\n\t" |
|
"pand %%mm7, %%mm5\n\t" |
|
"por %%mm1, %%mm0\n\t" |
|
"por %%mm4, %%mm3\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"por %%mm5, %%mm3\n\t" |
|
"psllq $16, %%mm3\n\t" |
|
"por %%mm3, %%mm0\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
:"=m"(*d):"m"(*s),"m"(blue_15mask):"memory"); |
|
d += 4; |
|
s += 16; |
|
} |
|
while(s < end) |
|
{ |
|
const int b= *s++; |
|
const int g= *s++; |
|
const int r= *s++; |
|
*d++ = (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
unsigned j,i,num_pixels=src_size/4; |
|
uint16_t *d = (uint16_t *)dst; |
|
for(i=0,j=0; j<num_pixels; i+=4,j++) |
|
{ |
|
const int b= src[i+0]; |
|
const int g= src[i+1]; |
|
const int r= src[i+2]; |
|
|
|
d[j]= (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
} |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb24to16)(const uint8_t *src, uint8_t *dst, unsigned src_size) |
|
{ |
|
#ifdef HAVE_MMX |
|
const uint8_t *s = src; |
|
const uint8_t *end,*mm_end; |
|
uint16_t *d = (uint16_t *)dst; |
|
end = s + src_size; |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); |
|
__asm __volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
__asm __volatile( |
|
"movq %0, %%mm7\n\t" |
|
"movq %1, %%mm6\n\t" |
|
::"m"(red_16mask),"m"(green_16mask)); |
|
if(mm_end == end) mm_end -= MMREG_SIZE*2; |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movd %1, %%mm0\n\t" |
|
"movd 3%1, %%mm3\n\t" |
|
"punpckldq 6%1, %%mm0\n\t" |
|
"punpckldq 9%1, %%mm3\n\t" |
|
"movq %%mm0, %%mm1\n\t" |
|
"movq %%mm0, %%mm2\n\t" |
|
"movq %%mm3, %%mm4\n\t" |
|
"movq %%mm3, %%mm5\n\t" |
|
"psrlq $3, %%mm0\n\t" |
|
"psrlq $3, %%mm3\n\t" |
|
"pand %2, %%mm0\n\t" |
|
"pand %2, %%mm3\n\t" |
|
"psrlq $5, %%mm1\n\t" |
|
"psrlq $5, %%mm4\n\t" |
|
"pand %%mm6, %%mm1\n\t" |
|
"pand %%mm6, %%mm4\n\t" |
|
"psrlq $8, %%mm2\n\t" |
|
"psrlq $8, %%mm5\n\t" |
|
"pand %%mm7, %%mm2\n\t" |
|
"pand %%mm7, %%mm5\n\t" |
|
"por %%mm1, %%mm0\n\t" |
|
"por %%mm4, %%mm3\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"por %%mm5, %%mm3\n\t" |
|
"psllq $16, %%mm3\n\t" |
|
"por %%mm3, %%mm0\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
:"=m"(*d):"m"(*s),"m"(blue_16mask):"memory"); |
|
d += 4; |
|
s += 12; |
|
} |
|
while(s < end) |
|
{ |
|
const int b= *s++; |
|
const int g= *s++; |
|
const int r= *s++; |
|
*d++ = (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
unsigned j,i,num_pixels=src_size/3; |
|
uint16_t *d = (uint16_t *)dst; |
|
for(i=0,j=0; j<num_pixels; i+=3,j++) |
|
{ |
|
const int b= src[i+0]; |
|
const int g= src[i+1]; |
|
const int r= src[i+2]; |
|
|
|
d[j]= (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
} |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb24to15)(const uint8_t *src, uint8_t *dst, unsigned src_size) |
|
{ |
|
#ifdef HAVE_MMX |
|
const uint8_t *s = src; |
|
const uint8_t *end,*mm_end; |
|
uint16_t *d = (uint16_t *)dst; |
|
end = s + src_size; |
|
mm_end = (uint8_t*)((((unsigned long)end)/(MMREG_SIZE*2))*(MMREG_SIZE*2)); |
|
__asm __volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
__asm __volatile( |
|
"movq %0, %%mm7\n\t" |
|
"movq %1, %%mm6\n\t" |
|
::"m"(red_15mask),"m"(green_15mask)); |
|
if(mm_end == end) mm_end -= MMREG_SIZE*2; |
|
while(s < mm_end) |
|
{ |
|
__asm __volatile( |
|
PREFETCH" 32%1\n\t" |
|
"movd %1, %%mm0\n\t" |
|
"movd 3%1, %%mm3\n\t" |
|
"punpckldq 6%1, %%mm0\n\t" |
|
"punpckldq 9%1, %%mm3\n\t" |
|
"movq %%mm0, %%mm1\n\t" |
|
"movq %%mm0, %%mm2\n\t" |
|
"movq %%mm3, %%mm4\n\t" |
|
"movq %%mm3, %%mm5\n\t" |
|
"psrlq $3, %%mm0\n\t" |
|
"psrlq $3, %%mm3\n\t" |
|
"pand %2, %%mm0\n\t" |
|
"pand %2, %%mm3\n\t" |
|
"psrlq $6, %%mm1\n\t" |
|
"psrlq $6, %%mm4\n\t" |
|
"pand %%mm6, %%mm1\n\t" |
|
"pand %%mm6, %%mm4\n\t" |
|
"psrlq $9, %%mm2\n\t" |
|
"psrlq $9, %%mm5\n\t" |
|
"pand %%mm7, %%mm2\n\t" |
|
"pand %%mm7, %%mm5\n\t" |
|
"por %%mm1, %%mm0\n\t" |
|
"por %%mm4, %%mm3\n\t" |
|
"por %%mm2, %%mm0\n\t" |
|
"por %%mm5, %%mm3\n\t" |
|
"psllq $16, %%mm3\n\t" |
|
"por %%mm3, %%mm0\n\t" |
|
MOVNTQ" %%mm0, %0\n\t" |
|
:"=m"(*d):"m"(*s),"m"(blue_15mask):"memory"); |
|
d += 4; |
|
s += 12; |
|
} |
|
while(s < end) |
|
{ |
|
const int b= *s++; |
|
const int g= *s++; |
|
const int r= *s++; |
|
*d++ = (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
} |
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
unsigned j,i,num_pixels=src_size/3; |
|
uint16_t *d = (uint16_t *)dst; |
|
for(i=0,j=0; j<num_pixels; i+=3,j++) |
|
{ |
|
const int b= src[i+0]; |
|
const int g= src[i+1]; |
|
const int r= src[i+2]; |
|
|
|
d[j]= (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
} |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb32tobgr32)(const uint8_t *src, uint8_t *dst, unsigned int src_size) |
|
{ |
|
int num_pixels= src_size >> 2; |
|
#ifdef HAVE_MMX |
|
asm volatile ( |
|
"xorl %%eax, %%eax \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 32(%0, %%eax) \n\t" |
|
"movq (%0, %%eax), %%mm0 \n\t" |
|
"movq %%mm0, %%mm1 \n\t" |
|
"movq %%mm0, %%mm2 \n\t" |
|
"pslld $16, %%mm0 \n\t" |
|
"psrld $16, %%mm1 \n\t" |
|
"pand "MANGLE(mask32r)", %%mm0 \n\t" |
|
"pand "MANGLE(mask32g)", %%mm2 \n\t" |
|
"pand "MANGLE(mask32b)", %%mm1 \n\t" |
|
"por %%mm0, %%mm2 \n\t" |
|
"por %%mm1, %%mm2 \n\t" |
|
MOVNTQ" %%mm2, (%1, %%eax) \n\t" |
|
"addl $2, %%eax \n\t" |
|
"cmpl %2, %%eax \n\t" |
|
" jb 1b \n\t" |
|
:: "r" (src), "r"(dst), "r" (num_pixels) |
|
: "%eax" |
|
); |
|
|
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
#else |
|
int i; |
|
for(i=0; i<num_pixels; i++) |
|
{ |
|
dst[4*i + 0] = src[4*i + 2]; |
|
dst[4*i + 1] = src[4*i + 1]; |
|
dst[4*i + 2] = src[4*i + 0]; |
|
} |
|
#endif |
|
} |
|
|
|
static inline void RENAME(rgb24tobgr24)(const uint8_t *src, uint8_t *dst, unsigned int src_size) |
|
{ |
|
int i; |
|
#ifdef HAVE_MMX |
|
int mmx_size= 23 - src_size; |
|
asm volatile ( |
|
"movq "MANGLE(mask24r)", %%mm5 \n\t" |
|
"movq "MANGLE(mask24g)", %%mm6 \n\t" |
|
"movq "MANGLE(mask24b)", %%mm7 \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 32(%1, %%eax) \n\t" |
|
"movq (%1, %%eax), %%mm0 \n\t" // BGR BGR BG |
|
"movq (%1, %%eax), %%mm1 \n\t" // BGR BGR BG |
|
"movq 2(%1, %%eax), %%mm2 \n\t" // R BGR BGR B |
|
"psllq $16, %%mm0 \n\t" // 00 BGR BGR |
|
"pand %%mm5, %%mm0 \n\t" |
|
"pand %%mm6, %%mm1 \n\t" |
|
"pand %%mm7, %%mm2 \n\t" |
|
"por %%mm0, %%mm1 \n\t" |
|
"por %%mm2, %%mm1 \n\t" |
|
"movq 6(%1, %%eax), %%mm0 \n\t" // BGR BGR BG |
|
MOVNTQ" %%mm1, (%2, %%eax) \n\t" // RGB RGB RG |
|
"movq 8(%1, %%eax), %%mm1 \n\t" // R BGR BGR B |
|
"movq 10(%1, %%eax), %%mm2 \n\t" // GR BGR BGR |
|
"pand %%mm7, %%mm0 \n\t" |
|
"pand %%mm5, %%mm1 \n\t" |
|
"pand %%mm6, %%mm2 \n\t" |
|
"por %%mm0, %%mm1 \n\t" |
|
"por %%mm2, %%mm1 \n\t" |
|
"movq 14(%1, %%eax), %%mm0 \n\t" // R BGR BGR B |
|
MOVNTQ" %%mm1, 8(%2, %%eax) \n\t" // B RGB RGB R |
|
"movq 16(%1, %%eax), %%mm1 \n\t" // GR BGR BGR |
|
"movq 18(%1, %%eax), %%mm2 \n\t" // BGR BGR BG |
|
"pand %%mm6, %%mm0 \n\t" |
|
"pand %%mm7, %%mm1 \n\t" |
|
"pand %%mm5, %%mm2 \n\t" |
|
"por %%mm0, %%mm1 \n\t" |
|
"por %%mm2, %%mm1 \n\t" |
|
MOVNTQ" %%mm1, 16(%2, %%eax) \n\t" |
|
"addl $24, %%eax \n\t" |
|
" js 1b \n\t" |
|
: "+a" (mmx_size) |
|
: "r" (src-mmx_size), "r"(dst-mmx_size) |
|
); |
|
|
|
__asm __volatile(SFENCE:::"memory"); |
|
__asm __volatile(EMMS:::"memory"); |
|
|
|
if(!mmx_size) return; //finihsed, was multiple of 8 |
|
|
|
src+= src_size; |
|
dst+= src_size; |
|
src_size= 24-mmx_size; |
|
src-= src_size; |
|
dst-= src_size; |
|
#endif |
|
for(i=0; i<src_size; i+=3) |
|
{ |
|
register int x; |
|
x = src[i + 2]; |
|
dst[i + 1] = src[i + 1]; |
|
dst[i + 2] = src[i + 0]; |
|
dst[i + 0] = x; |
|
} |
|
} |
|
|
|
static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *usrc, const uint8_t *vsrc, uint8_t *dst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int dstStride, int vertLumPerChroma) |
|
{ |
|
int y; |
|
const int chromWidth= width>>1; |
|
for(y=0; y<height; y++) |
|
{ |
|
#ifdef HAVE_MMX |
|
//FIXME handle 2 lines a once (fewer prefetch, reuse some chrom, but very likely limited by mem anyway) |
|
asm volatile( |
|
"xorl %%eax, %%eax \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 32(%1, %%eax, 2) \n\t" |
|
PREFETCH" 32(%2, %%eax) \n\t" |
|
PREFETCH" 32(%3, %%eax) \n\t" |
|
"movq (%2, %%eax), %%mm0 \n\t" // U(0) |
|
"movq %%mm0, %%mm2 \n\t" // U(0) |
|
"movq (%3, %%eax), %%mm1 \n\t" // V(0) |
|
"punpcklbw %%mm1, %%mm0 \n\t" // UVUV UVUV(0) |
|
"punpckhbw %%mm1, %%mm2 \n\t" // UVUV UVUV(8) |
|
|
|
"movq (%1, %%eax,2), %%mm3 \n\t" // Y(0) |
|
"movq 8(%1, %%eax,2), %%mm5 \n\t" // Y(8) |
|
"movq %%mm3, %%mm4 \n\t" // Y(0) |
|
"movq %%mm5, %%mm6 \n\t" // Y(8) |
|
"punpcklbw %%mm0, %%mm3 \n\t" // YUYV YUYV(0) |
|
"punpckhbw %%mm0, %%mm4 \n\t" // YUYV YUYV(4) |
|
"punpcklbw %%mm2, %%mm5 \n\t" // YUYV YUYV(8) |
|
"punpckhbw %%mm2, %%mm6 \n\t" // YUYV YUYV(12) |
|
|
|
MOVNTQ" %%mm3, (%0, %%eax, 4) \n\t" |
|
MOVNTQ" %%mm4, 8(%0, %%eax, 4) \n\t" |
|
MOVNTQ" %%mm5, 16(%0, %%eax, 4) \n\t" |
|
MOVNTQ" %%mm6, 24(%0, %%eax, 4) \n\t" |
|
|
|
"addl $8, %%eax \n\t" |
|
"cmpl %4, %%eax \n\t" |
|
" jb 1b \n\t" |
|
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "r" (chromWidth) |
|
: "%eax" |
|
); |
|
#else |
|
int i; |
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
dst[4*i+0] = ysrc[2*i+0]; |
|
dst[4*i+1] = usrc[i]; |
|
dst[4*i+2] = ysrc[2*i+1]; |
|
dst[4*i+3] = vsrc[i]; |
|
} |
|
#endif |
|
if((y&(vertLumPerChroma-1))==(vertLumPerChroma-1) ) |
|
{ |
|
usrc += chromStride; |
|
vsrc += chromStride; |
|
} |
|
ysrc += lumStride; |
|
dst += dstStride; |
|
} |
|
#ifdef HAVE_MMX |
|
asm( EMMS" \n\t" |
|
SFENCE" \n\t" |
|
:::"memory"); |
|
#endif |
|
} |
|
|
|
/** |
|
* |
|
* height should be a multiple of 2 and width should be a multiple of 16 (if this is a |
|
* problem for anyone then tell me, and ill fix it) |
|
*/ |
|
static inline void RENAME(yv12toyuy2)(const uint8_t *ysrc, const uint8_t *usrc, const uint8_t *vsrc, uint8_t *dst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int dstStride) |
|
{ |
|
//FIXME interpolate chroma |
|
RENAME(yuvPlanartoyuy2)(ysrc, usrc, vsrc, dst, width, height, lumStride, chromStride, dstStride, 2); |
|
} |
|
|
|
/** |
|
* |
|
* width should be a multiple of 16 |
|
*/ |
|
static inline void RENAME(yuv422ptoyuy2)(const uint8_t *ysrc, const uint8_t *usrc, const uint8_t *vsrc, uint8_t *dst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int dstStride) |
|
{ |
|
RENAME(yuvPlanartoyuy2)(ysrc, usrc, vsrc, dst, width, height, lumStride, chromStride, dstStride, 1); |
|
} |
|
|
|
/** |
|
* |
|
* height should be a multiple of 2 and width should be a multiple of 16 (if this is a |
|
* problem for anyone then tell me, and ill fix it) |
|
*/ |
|
static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int srcStride) |
|
{ |
|
int y; |
|
const int chromWidth= width>>1; |
|
for(y=0; y<height; y+=2) |
|
{ |
|
#ifdef HAVE_MMX |
|
asm volatile( |
|
"xorl %%eax, %%eax \n\t" |
|
"pcmpeqw %%mm7, %%mm7 \n\t" |
|
"psrlw $8, %%mm7 \n\t" // FF,00,FF,00... |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%eax, 4) \n\t" |
|
"movq (%0, %%eax, 4), %%mm0 \n\t" // YUYV YUYV(0) |
|
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(4) |
|
"movq %%mm0, %%mm2 \n\t" // YUYV YUYV(0) |
|
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(4) |
|
"psrlw $8, %%mm0 \n\t" // U0V0 U0V0(0) |
|
"psrlw $8, %%mm1 \n\t" // U0V0 U0V0(4) |
|
"pand %%mm7, %%mm2 \n\t" // Y0Y0 Y0Y0(0) |
|
"pand %%mm7, %%mm3 \n\t" // Y0Y0 Y0Y0(4) |
|
"packuswb %%mm1, %%mm0 \n\t" // UVUV UVUV(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(0) |
|
|
|
MOVNTQ" %%mm2, (%1, %%eax, 2) \n\t" |
|
|
|
"movq 16(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(8) |
|
"movq 24(%0, %%eax, 4), %%mm2 \n\t" // YUYV YUYV(12) |
|
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(8) |
|
"movq %%mm2, %%mm4 \n\t" // YUYV YUYV(12) |
|
"psrlw $8, %%mm1 \n\t" // U0V0 U0V0(8) |
|
"psrlw $8, %%mm2 \n\t" // U0V0 U0V0(12) |
|
"pand %%mm7, %%mm3 \n\t" // Y0Y0 Y0Y0(8) |
|
"pand %%mm7, %%mm4 \n\t" // Y0Y0 Y0Y0(12) |
|
"packuswb %%mm2, %%mm1 \n\t" // UVUV UVUV(8) |
|
"packuswb %%mm4, %%mm3 \n\t" // YYYY YYYY(8) |
|
|
|
MOVNTQ" %%mm3, 8(%1, %%eax, 2) \n\t" |
|
|
|
"movq %%mm0, %%mm2 \n\t" // UVUV UVUV(0) |
|
"movq %%mm1, %%mm3 \n\t" // UVUV UVUV(8) |
|
"psrlw $8, %%mm0 \n\t" // V0V0 V0V0(0) |
|
"psrlw $8, %%mm1 \n\t" // V0V0 V0V0(8) |
|
"pand %%mm7, %%mm2 \n\t" // U0U0 U0U0(0) |
|
"pand %%mm7, %%mm3 \n\t" // U0U0 U0U0(8) |
|
"packuswb %%mm1, %%mm0 \n\t" // VVVV VVVV(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // UUUU UUUU(0) |
|
|
|
MOVNTQ" %%mm0, (%3, %%eax) \n\t" |
|
MOVNTQ" %%mm2, (%2, %%eax) \n\t" |
|
|
|
"addl $8, %%eax \n\t" |
|
"cmpl %4, %%eax \n\t" |
|
" jb 1b \n\t" |
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth) |
|
: "memory", "%eax" |
|
); |
|
|
|
ydst += lumStride; |
|
src += srcStride; |
|
|
|
asm volatile( |
|
"xorl %%eax, %%eax \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%eax, 4) \n\t" |
|
"movq (%0, %%eax, 4), %%mm0 \n\t" // YUYV YUYV(0) |
|
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(4) |
|
"movq 16(%0, %%eax, 4), %%mm2 \n\t" // YUYV YUYV(8) |
|
"movq 24(%0, %%eax, 4), %%mm3 \n\t" // YUYV YUYV(12) |
|
"pand %%mm7, %%mm0 \n\t" // Y0Y0 Y0Y0(0) |
|
"pand %%mm7, %%mm1 \n\t" // Y0Y0 Y0Y0(4) |
|
"pand %%mm7, %%mm2 \n\t" // Y0Y0 Y0Y0(8) |
|
"pand %%mm7, %%mm3 \n\t" // Y0Y0 Y0Y0(12) |
|
"packuswb %%mm1, %%mm0 \n\t" // YYYY YYYY(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(8) |
|
|
|
MOVNTQ" %%mm0, (%1, %%eax, 2) \n\t" |
|
MOVNTQ" %%mm2, 8(%1, %%eax, 2) \n\t" |
|
|
|
"addl $8, %%eax \n\t" |
|
"cmpl %4, %%eax \n\t" |
|
" jb 1b \n\t" |
|
|
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth) |
|
: "memory", "%eax" |
|
); |
|
#else |
|
int i; |
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
ydst[2*i+0] = src[4*i+0]; |
|
udst[i] = src[4*i+1]; |
|
ydst[2*i+1] = src[4*i+2]; |
|
vdst[i] = src[4*i+3]; |
|
} |
|
ydst += lumStride; |
|
src += srcStride; |
|
|
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
ydst[2*i+0] = src[4*i+0]; |
|
ydst[2*i+1] = src[4*i+2]; |
|
} |
|
#endif |
|
udst += chromStride; |
|
vdst += chromStride; |
|
ydst += lumStride; |
|
src += srcStride; |
|
} |
|
#ifdef HAVE_MMX |
|
asm volatile( EMMS" \n\t" |
|
SFENCE" \n\t" |
|
:::"memory"); |
|
#endif |
|
} |
|
|
|
/** |
|
* |
|
* height should be a multiple of 2 and width should be a multiple of 16 (if this is a |
|
* problem for anyone then tell me, and ill fix it) |
|
* chrominance data is only taken from every secound line others are ignored FIXME write HQ version |
|
*/ |
|
static inline void RENAME(uyvytoyv12)(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int srcStride) |
|
{ |
|
int y; |
|
const int chromWidth= width>>1; |
|
for(y=0; y<height; y+=2) |
|
{ |
|
#ifdef HAVE_MMX |
|
asm volatile( |
|
"xorl %%eax, %%eax \n\t" |
|
"pcmpeqw %%mm7, %%mm7 \n\t" |
|
"psrlw $8, %%mm7 \n\t" // FF,00,FF,00... |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%eax, 4) \n\t" |
|
"movq (%0, %%eax, 4), %%mm0 \n\t" // UYVY UYVY(0) |
|
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // UYVY UYVY(4) |
|
"movq %%mm0, %%mm2 \n\t" // UYVY UYVY(0) |
|
"movq %%mm1, %%mm3 \n\t" // UYVY UYVY(4) |
|
"pand %%mm7, %%mm0 \n\t" // U0V0 U0V0(0) |
|
"pand %%mm7, %%mm1 \n\t" // U0V0 U0V0(4) |
|
"psrlw $8, %%mm2 \n\t" // Y0Y0 Y0Y0(0) |
|
"psrlw $8, %%mm3 \n\t" // Y0Y0 Y0Y0(4) |
|
"packuswb %%mm1, %%mm0 \n\t" // UVUV UVUV(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(0) |
|
|
|
MOVNTQ" %%mm2, (%1, %%eax, 2) \n\t" |
|
|
|
"movq 16(%0, %%eax, 4), %%mm1 \n\t" // UYVY UYVY(8) |
|
"movq 24(%0, %%eax, 4), %%mm2 \n\t" // UYVY UYVY(12) |
|
"movq %%mm1, %%mm3 \n\t" // UYVY UYVY(8) |
|
"movq %%mm2, %%mm4 \n\t" // UYVY UYVY(12) |
|
"pand %%mm7, %%mm1 \n\t" // U0V0 U0V0(8) |
|
"pand %%mm7, %%mm2 \n\t" // U0V0 U0V0(12) |
|
"psrlw $8, %%mm3 \n\t" // Y0Y0 Y0Y0(8) |
|
"psrlw $8, %%mm4 \n\t" // Y0Y0 Y0Y0(12) |
|
"packuswb %%mm2, %%mm1 \n\t" // UVUV UVUV(8) |
|
"packuswb %%mm4, %%mm3 \n\t" // YYYY YYYY(8) |
|
|
|
MOVNTQ" %%mm3, 8(%1, %%eax, 2) \n\t" |
|
|
|
"movq %%mm0, %%mm2 \n\t" // UVUV UVUV(0) |
|
"movq %%mm1, %%mm3 \n\t" // UVUV UVUV(8) |
|
"psrlw $8, %%mm0 \n\t" // V0V0 V0V0(0) |
|
"psrlw $8, %%mm1 \n\t" // V0V0 V0V0(8) |
|
"pand %%mm7, %%mm2 \n\t" // U0U0 U0U0(0) |
|
"pand %%mm7, %%mm3 \n\t" // U0U0 U0U0(8) |
|
"packuswb %%mm1, %%mm0 \n\t" // VVVV VVVV(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // UUUU UUUU(0) |
|
|
|
MOVNTQ" %%mm0, (%3, %%eax) \n\t" |
|
MOVNTQ" %%mm2, (%2, %%eax) \n\t" |
|
|
|
"addl $8, %%eax \n\t" |
|
"cmpl %4, %%eax \n\t" |
|
" jb 1b \n\t" |
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth) |
|
: "memory", "%eax" |
|
); |
|
|
|
ydst += lumStride; |
|
src += srcStride; |
|
|
|
asm volatile( |
|
"xorl %%eax, %%eax \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%eax, 4) \n\t" |
|
"movq (%0, %%eax, 4), %%mm0 \n\t" // YUYV YUYV(0) |
|
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(4) |
|
"movq 16(%0, %%eax, 4), %%mm2 \n\t" // YUYV YUYV(8) |
|
"movq 24(%0, %%eax, 4), %%mm3 \n\t" // YUYV YUYV(12) |
|
"psrlw $8, %%mm0 \n\t" // Y0Y0 Y0Y0(0) |
|
"psrlw $8, %%mm1 \n\t" // Y0Y0 Y0Y0(4) |
|
"psrlw $8, %%mm2 \n\t" // Y0Y0 Y0Y0(8) |
|
"psrlw $8, %%mm3 \n\t" // Y0Y0 Y0Y0(12) |
|
"packuswb %%mm1, %%mm0 \n\t" // YYYY YYYY(0) |
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(8) |
|
|
|
MOVNTQ" %%mm0, (%1, %%eax, 2) \n\t" |
|
MOVNTQ" %%mm2, 8(%1, %%eax, 2) \n\t" |
|
|
|
"addl $8, %%eax \n\t" |
|
"cmpl %4, %%eax \n\t" |
|
" jb 1b \n\t" |
|
|
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "r" (chromWidth) |
|
: "memory", "%eax" |
|
); |
|
#else |
|
int i; |
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
udst[i] = src[4*i+0]; |
|
ydst[2*i+0] = src[4*i+1]; |
|
vdst[i] = src[4*i+2]; |
|
ydst[2*i+1] = src[4*i+3]; |
|
} |
|
ydst += lumStride; |
|
src += srcStride; |
|
|
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
ydst[2*i+0] = src[4*i+1]; |
|
ydst[2*i+1] = src[4*i+3]; |
|
} |
|
#endif |
|
udst += chromStride; |
|
vdst += chromStride; |
|
ydst += lumStride; |
|
src += srcStride; |
|
} |
|
#ifdef HAVE_MMX |
|
asm volatile( EMMS" \n\t" |
|
SFENCE" \n\t" |
|
:::"memory"); |
|
#endif |
|
} |
|
|
|
/** |
|
* |
|
* height should be a multiple of 2 and width should be a multiple of 2 (if this is a |
|
* problem for anyone then tell me, and ill fix it) |
|
* chrominance data is only taken from every secound line others are ignored in the C version FIXME write HQ version |
|
*/ |
|
static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
unsigned int width, unsigned int height, |
|
unsigned int lumStride, unsigned int chromStride, unsigned int srcStride) |
|
{ |
|
int y; |
|
const int chromWidth= width>>1; |
|
#ifdef HAVE_MMX |
|
for(y=0; y<height-2; y+=2) |
|
{ |
|
int i; |
|
for(i=0; i<2; i++) |
|
{ |
|
asm volatile( |
|
"movl %2, %%eax \n\t" |
|
"movq "MANGLE(bgr2YCoeff)", %%mm6 \n\t" |
|
"movq "MANGLE(w1111)", %%mm5 \n\t" |
|
"pxor %%mm7, %%mm7 \n\t" |
|
"leal (%%eax, %%eax, 2), %%ebx \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%ebx) \n\t" |
|
"movd (%0, %%ebx), %%mm0 \n\t" |
|
"movd 3(%0, %%ebx), %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm0 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"movd 6(%0, %%ebx), %%mm2 \n\t" |
|
"movd 9(%0, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"pmaddwd %%mm6, %%mm0 \n\t" |
|
"pmaddwd %%mm6, %%mm1 \n\t" |
|
"pmaddwd %%mm6, %%mm2 \n\t" |
|
"pmaddwd %%mm6, %%mm3 \n\t" |
|
#ifndef FAST_BGR2YV12 |
|
"psrad $8, %%mm0 \n\t" |
|
"psrad $8, %%mm1 \n\t" |
|
"psrad $8, %%mm2 \n\t" |
|
"psrad $8, %%mm3 \n\t" |
|
#endif |
|
"packssdw %%mm1, %%mm0 \n\t" |
|
"packssdw %%mm3, %%mm2 \n\t" |
|
"pmaddwd %%mm5, %%mm0 \n\t" |
|
"pmaddwd %%mm5, %%mm2 \n\t" |
|
"packssdw %%mm2, %%mm0 \n\t" |
|
"psraw $7, %%mm0 \n\t" |
|
|
|
"movd 12(%0, %%ebx), %%mm4 \n\t" |
|
"movd 15(%0, %%ebx), %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm4 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"movd 18(%0, %%ebx), %%mm2 \n\t" |
|
"movd 21(%0, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"pmaddwd %%mm6, %%mm4 \n\t" |
|
"pmaddwd %%mm6, %%mm1 \n\t" |
|
"pmaddwd %%mm6, %%mm2 \n\t" |
|
"pmaddwd %%mm6, %%mm3 \n\t" |
|
#ifndef FAST_BGR2YV12 |
|
"psrad $8, %%mm4 \n\t" |
|
"psrad $8, %%mm1 \n\t" |
|
"psrad $8, %%mm2 \n\t" |
|
"psrad $8, %%mm3 \n\t" |
|
#endif |
|
"packssdw %%mm1, %%mm4 \n\t" |
|
"packssdw %%mm3, %%mm2 \n\t" |
|
"pmaddwd %%mm5, %%mm4 \n\t" |
|
"pmaddwd %%mm5, %%mm2 \n\t" |
|
"addl $24, %%ebx \n\t" |
|
"packssdw %%mm2, %%mm4 \n\t" |
|
"psraw $7, %%mm4 \n\t" |
|
|
|
"packuswb %%mm4, %%mm0 \n\t" |
|
"paddusb "MANGLE(bgr2YOffset)", %%mm0 \n\t" |
|
|
|
MOVNTQ" %%mm0, (%1, %%eax) \n\t" |
|
"addl $8, %%eax \n\t" |
|
" js 1b \n\t" |
|
: : "r" (src+width*3), "r" (ydst+width), "g" (-width) |
|
: "%eax", "%ebx" |
|
); |
|
ydst += lumStride; |
|
src += srcStride; |
|
} |
|
src -= srcStride*2; |
|
asm volatile( |
|
"movl %4, %%eax \n\t" |
|
"movq "MANGLE(w1111)", %%mm5 \n\t" |
|
"movq "MANGLE(bgr2UCoeff)", %%mm6 \n\t" |
|
"pxor %%mm7, %%mm7 \n\t" |
|
"leal (%%eax, %%eax, 2), %%ebx \n\t" |
|
"addl %%ebx, %%ebx \n\t" |
|
".balign 16 \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%0, %%ebx) \n\t" |
|
PREFETCH" 64(%1, %%ebx) \n\t" |
|
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW) |
|
"movq (%0, %%ebx), %%mm0 \n\t" |
|
"movq (%1, %%ebx), %%mm1 \n\t" |
|
"movq 6(%0, %%ebx), %%mm2 \n\t" |
|
"movq 6(%1, %%ebx), %%mm3 \n\t" |
|
PAVGB" %%mm1, %%mm0 \n\t" |
|
PAVGB" %%mm3, %%mm2 \n\t" |
|
"movq %%mm0, %%mm1 \n\t" |
|
"movq %%mm2, %%mm3 \n\t" |
|
"psrlq $24, %%mm0 \n\t" |
|
"psrlq $24, %%mm2 \n\t" |
|
PAVGB" %%mm1, %%mm0 \n\t" |
|
PAVGB" %%mm3, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm0 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
#else |
|
"movd (%0, %%ebx), %%mm0 \n\t" |
|
"movd (%1, %%ebx), %%mm1 \n\t" |
|
"movd 3(%0, %%ebx), %%mm2 \n\t" |
|
"movd 3(%1, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm0 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"paddw %%mm1, %%mm0 \n\t" |
|
"paddw %%mm3, %%mm2 \n\t" |
|
"paddw %%mm2, %%mm0 \n\t" |
|
"movd 6(%0, %%ebx), %%mm4 \n\t" |
|
"movd 6(%1, %%ebx), %%mm1 \n\t" |
|
"movd 9(%0, %%ebx), %%mm2 \n\t" |
|
"movd 9(%1, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm4 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"paddw %%mm1, %%mm4 \n\t" |
|
"paddw %%mm3, %%mm2 \n\t" |
|
"paddw %%mm4, %%mm2 \n\t" |
|
"psrlw $2, %%mm0 \n\t" |
|
"psrlw $2, %%mm2 \n\t" |
|
#endif |
|
"movq "MANGLE(bgr2VCoeff)", %%mm1 \n\t" |
|
"movq "MANGLE(bgr2VCoeff)", %%mm3 \n\t" |
|
|
|
"pmaddwd %%mm0, %%mm1 \n\t" |
|
"pmaddwd %%mm2, %%mm3 \n\t" |
|
"pmaddwd %%mm6, %%mm0 \n\t" |
|
"pmaddwd %%mm6, %%mm2 \n\t" |
|
#ifndef FAST_BGR2YV12 |
|
"psrad $8, %%mm0 \n\t" |
|
"psrad $8, %%mm1 \n\t" |
|
"psrad $8, %%mm2 \n\t" |
|
"psrad $8, %%mm3 \n\t" |
|
#endif |
|
"packssdw %%mm2, %%mm0 \n\t" |
|
"packssdw %%mm3, %%mm1 \n\t" |
|
"pmaddwd %%mm5, %%mm0 \n\t" |
|
"pmaddwd %%mm5, %%mm1 \n\t" |
|
"packssdw %%mm1, %%mm0 \n\t" // V1 V0 U1 U0 |
|
"psraw $7, %%mm0 \n\t" |
|
|
|
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW) |
|
"movq 12(%0, %%ebx), %%mm4 \n\t" |
|
"movq 12(%1, %%ebx), %%mm1 \n\t" |
|
"movq 18(%0, %%ebx), %%mm2 \n\t" |
|
"movq 18(%1, %%ebx), %%mm3 \n\t" |
|
PAVGB" %%mm1, %%mm4 \n\t" |
|
PAVGB" %%mm3, %%mm2 \n\t" |
|
"movq %%mm4, %%mm1 \n\t" |
|
"movq %%mm2, %%mm3 \n\t" |
|
"psrlq $24, %%mm4 \n\t" |
|
"psrlq $24, %%mm2 \n\t" |
|
PAVGB" %%mm1, %%mm4 \n\t" |
|
PAVGB" %%mm3, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm4 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
#else |
|
"movd 12(%0, %%ebx), %%mm4 \n\t" |
|
"movd 12(%1, %%ebx), %%mm1 \n\t" |
|
"movd 15(%0, %%ebx), %%mm2 \n\t" |
|
"movd 15(%1, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm4 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"paddw %%mm1, %%mm4 \n\t" |
|
"paddw %%mm3, %%mm2 \n\t" |
|
"paddw %%mm2, %%mm4 \n\t" |
|
"movd 18(%0, %%ebx), %%mm5 \n\t" |
|
"movd 18(%1, %%ebx), %%mm1 \n\t" |
|
"movd 21(%0, %%ebx), %%mm2 \n\t" |
|
"movd 21(%1, %%ebx), %%mm3 \n\t" |
|
"punpcklbw %%mm7, %%mm5 \n\t" |
|
"punpcklbw %%mm7, %%mm1 \n\t" |
|
"punpcklbw %%mm7, %%mm2 \n\t" |
|
"punpcklbw %%mm7, %%mm3 \n\t" |
|
"paddw %%mm1, %%mm5 \n\t" |
|
"paddw %%mm3, %%mm2 \n\t" |
|
"paddw %%mm5, %%mm2 \n\t" |
|
"movq "MANGLE(w1111)", %%mm5 \n\t" |
|
"psrlw $2, %%mm4 \n\t" |
|
"psrlw $2, %%mm2 \n\t" |
|
#endif |
|
"movq "MANGLE(bgr2VCoeff)", %%mm1 \n\t" |
|
"movq "MANGLE(bgr2VCoeff)", %%mm3 \n\t" |
|
|
|
"pmaddwd %%mm4, %%mm1 \n\t" |
|
"pmaddwd %%mm2, %%mm3 \n\t" |
|
"pmaddwd %%mm6, %%mm4 \n\t" |
|
"pmaddwd %%mm6, %%mm2 \n\t" |
|
#ifndef FAST_BGR2YV12 |
|
"psrad $8, %%mm4 \n\t" |
|
"psrad $8, %%mm1 \n\t" |
|
"psrad $8, %%mm2 \n\t" |
|
"psrad $8, %%mm3 \n\t" |
|
#endif |
|
"packssdw %%mm2, %%mm4 \n\t" |
|
"packssdw %%mm3, %%mm1 \n\t" |
|
"pmaddwd %%mm5, %%mm4 \n\t" |
|
"pmaddwd %%mm5, %%mm1 \n\t" |
|
"addl $24, %%ebx \n\t" |
|
"packssdw %%mm1, %%mm4 \n\t" // V3 V2 U3 U2 |
|
"psraw $7, %%mm4 \n\t" |
|
|
|
"movq %%mm0, %%mm1 \n\t" |
|
"punpckldq %%mm4, %%mm0 \n\t" |
|
"punpckhdq %%mm4, %%mm1 \n\t" |
|
"packsswb %%mm1, %%mm0 \n\t" |
|
"paddb "MANGLE(bgr2UVOffset)", %%mm0 \n\t" |
|
|
|
"movd %%mm0, (%2, %%eax) \n\t" |
|
"punpckhdq %%mm0, %%mm0 \n\t" |
|
"movd %%mm0, (%3, %%eax) \n\t" |
|
"addl $4, %%eax \n\t" |
|
" js 1b \n\t" |
|
: : "r" (src+width*6), "r" (src+srcStride+width*6), "r" (udst+width), "r" (vdst+width), "g" (-width) |
|
: "%eax", "%ebx" |
|
); |
|
|
|
udst += chromStride; |
|
vdst += chromStride; |
|
src += srcStride*2; |
|
} |
|
|
|
asm volatile( EMMS" \n\t" |
|
SFENCE" \n\t" |
|
:::"memory"); |
|
#else |
|
y=0; |
|
#endif |
|
for(; y<height; y+=2) |
|
{ |
|
int i; |
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
unsigned int b= src[6*i+0]; |
|
unsigned int g= src[6*i+1]; |
|
unsigned int r= src[6*i+2]; |
|
|
|
unsigned int Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
unsigned int V = ((RV*r + GV*g + BV*b)>>RGB2YUV_SHIFT) + 128; |
|
unsigned int U = ((RU*r + GU*g + BU*b)>>RGB2YUV_SHIFT) + 128; |
|
|
|
udst[i] = U; |
|
vdst[i] = V; |
|
ydst[2*i] = Y; |
|
|
|
b= src[6*i+3]; |
|
g= src[6*i+4]; |
|
r= src[6*i+5]; |
|
|
|
Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
ydst[2*i+1] = Y; |
|
} |
|
ydst += lumStride; |
|
src += srcStride; |
|
|
|
for(i=0; i<chromWidth; i++) |
|
{ |
|
unsigned int b= src[6*i+0]; |
|
unsigned int g= src[6*i+1]; |
|
unsigned int r= src[6*i+2]; |
|
|
|
unsigned int Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
|
|
ydst[2*i] = Y; |
|
|
|
b= src[6*i+3]; |
|
g= src[6*i+4]; |
|
r= src[6*i+5]; |
|
|
|
Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
ydst[2*i+1] = Y; |
|
} |
|
udst += chromStride; |
|
vdst += chromStride; |
|
ydst += lumStride; |
|
src += srcStride; |
|
} |
|
} |
|
|
|
void RENAME(interleaveBytes)(uint8_t *src1, uint8_t *src2, uint8_t *dest, |
|
int width, int height, int src1Stride, int src2Stride, int dstStride){ |
|
int h; |
|
|
|
for(h=0; h < height; h++) |
|
{ |
|
int w; |
|
|
|
#ifdef HAVE_MMX |
|
#ifdef HAVE_SSE2 |
|
asm( |
|
"xorl %%eax, %%eax \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%1, %%eax) \n\t" |
|
PREFETCH" 64(%2, %%eax) \n\t" |
|
"movdqa (%1, %%eax), %%xmm0 \n\t" |
|
"movdqa (%1, %%eax), %%xmm1 \n\t" |
|
"movdqa (%2, %%eax), %%xmm2 \n\t" |
|
"punpcklbw %%xmm2, %%xmm0 \n\t" |
|
"punpckhbw %%xmm2, %%xmm1 \n\t" |
|
"movntdq %%xmm0, (%0, %%eax, 2) \n\t" |
|
"movntdq %%xmm1, 16(%0, %%eax, 2)\n\t" |
|
"addl $16, %%eax \n\t" |
|
"cmpl %3, %%eax \n\t" |
|
" jb 1b \n\t" |
|
::"r"(dest), "r"(src1), "r"(src2), "r" (width-15) |
|
: "memory", "%eax" |
|
); |
|
#else |
|
asm( |
|
"xorl %%eax, %%eax \n\t" |
|
"1: \n\t" |
|
PREFETCH" 64(%1, %%eax) \n\t" |
|
PREFETCH" 64(%2, %%eax) \n\t" |
|
"movq (%1, %%eax), %%mm0 \n\t" |
|
"movq 8(%1, %%eax), %%mm2 \n\t" |
|
"movq %%mm0, %%mm1 \n\t" |
|
"movq %%mm2, %%mm3 \n\t" |
|
"movq (%2, %%eax), %%mm4 \n\t" |
|
"movq 8(%2, %%eax), %%mm5 \n\t" |
|
"punpcklbw %%mm4, %%mm0 \n\t" |
|
"punpckhbw %%mm4, %%mm1 \n\t" |
|
"punpcklbw %%mm5, %%mm2 \n\t" |
|
"punpckhbw %%mm5, %%mm3 \n\t" |
|
MOVNTQ" %%mm0, (%0, %%eax, 2) \n\t" |
|
MOVNTQ" %%mm1, 8(%0, %%eax, 2) \n\t" |
|
MOVNTQ" %%mm2, 16(%0, %%eax, 2) \n\t" |
|
MOVNTQ" %%mm3, 24(%0, %%eax, 2) \n\t" |
|
"addl $16, %%eax \n\t" |
|
"cmpl %3, %%eax \n\t" |
|
" jb 1b \n\t" |
|
::"r"(dest), "r"(src1), "r"(src2), "r" (width-15) |
|
: "memory", "%eax" |
|
); |
|
#endif |
|
for(w= (width&(~15)); w < width; w++) |
|
{ |
|
dest[2*w+0] = src1[w]; |
|
dest[2*w+1] = src2[w]; |
|
} |
|
#else |
|
for(w=0; w < width; w++) |
|
{ |
|
dest[2*w+0] = src1[w]; |
|
dest[2*w+1] = src2[w]; |
|
} |
|
#endif |
|
dest += dstStride; |
|
src1 += src1Stride; |
|
src2 += src2Stride; |
|
} |
|
#ifdef HAVE_MMX |
|
asm( |
|
EMMS" \n\t" |
|
SFENCE" \n\t" |
|
::: "memory" |
|
); |
|
#endif |
|
}
|
|
|