|
|
|
@ -69,11 +69,8 @@ static inline void RENAME(rgb24tobgr32)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
uint8_t *dest = dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
mm_end = end - 23; |
|
|
|
|
__asm__ volatile("movq %0, %%mm7"::"m"(mask32a):"memory"); |
|
|
|
@ -104,21 +101,11 @@ static inline void RENAME(rgb24tobgr32)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
/* RGB24 (= R,G,B) -> RGB32 (= A,B,G,R) */ |
|
|
|
|
*dest++ = 255; |
|
|
|
|
*dest++ = s[2]; |
|
|
|
|
*dest++ = s[1]; |
|
|
|
|
*dest++ = s[0]; |
|
|
|
|
s+=3; |
|
|
|
|
#else |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = 255; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -164,11 +151,8 @@ static inline void RENAME(rgb32tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
uint8_t *dest = dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
mm_end = end - 31; |
|
|
|
|
while (s < mm_end) { |
|
|
|
@ -191,21 +175,11 @@ static inline void RENAME(rgb32tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
/* RGB32 (= A,B,G,R) -> RGB24 (= R,G,B) */ |
|
|
|
|
s++; |
|
|
|
|
dest[2] = *s++; |
|
|
|
|
dest[1] = *s++; |
|
|
|
|
dest[0] = *s++; |
|
|
|
|
dest += 3; |
|
|
|
|
#else |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
s++; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -222,7 +196,6 @@ static inline void RENAME(rgb15to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
register const uint8_t *end; |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s)); |
|
|
|
|
__asm__ volatile("movq %0, %%mm4"::"m"(mask15s)); |
|
|
|
|
mm_end = end - 15; |
|
|
|
@ -247,7 +220,6 @@ static inline void RENAME(rgb15to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
mm_end = end - 3; |
|
|
|
|
while (s < mm_end) { |
|
|
|
|
register unsigned x= *((const uint32_t *)s); |
|
|
|
@ -268,7 +240,6 @@ static inline void RENAME(rgb16to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
register const uint8_t *end; |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s)); |
|
|
|
|
__asm__ volatile("movq %0, %%mm7"::"m"(mask15rg)); |
|
|
|
|
__asm__ volatile("movq %0, %%mm6"::"m"(mask15b)); |
|
|
|
@ -298,7 +269,6 @@ static inline void RENAME(rgb16to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
mm_end = end - 3; |
|
|
|
|
while (s < mm_end) { |
|
|
|
|
register uint32_t x= *((const uint32_t*)s); |
|
|
|
@ -316,12 +286,9 @@ static inline void RENAME(rgb32to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
mm_end = end - 15; |
|
|
|
|
#if 1 //is faster only if multiplies are reasonably fast (FIXME figure out on which CPUs this is faster, on Athlon it is slightly faster)
|
|
|
|
|
__asm__ volatile( |
|
|
|
@ -401,7 +368,6 @@ static inline void RENAME(rgb32to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
#endif |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xFF)>>3) + ((rgb&0xFC00)>>5) + ((rgb&0xF80000)>>8); |
|
|
|
@ -412,12 +378,9 @@ static inline void RENAME(rgb32tobgr16)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -460,7 +423,6 @@ static inline void RENAME(rgb32tobgr16)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xF8)<<8) + ((rgb&0xFC00)>>5) + ((rgb&0xF80000)>>19); |
|
|
|
@ -471,12 +433,9 @@ static inline void RENAME(rgb32to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
mm_end = end - 15; |
|
|
|
|
#if 1 //is faster only if multiplies are reasonably fast (FIXME figure out on which CPUs this is faster, on Athlon it is slightly faster)
|
|
|
|
|
__asm__ volatile( |
|
|
|
@ -556,7 +515,6 @@ static inline void RENAME(rgb32to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
#endif |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xFF)>>3) + ((rgb&0xF800)>>6) + ((rgb&0xF80000)>>9); |
|
|
|
@ -567,12 +525,9 @@ static inline void RENAME(rgb32tobgr15)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -615,7 +570,6 @@ static inline void RENAME(rgb32tobgr15)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xF8)<<7) + ((rgb&0xF800)>>6) + ((rgb&0xF80000)>>19); |
|
|
|
@ -626,12 +580,9 @@ static inline void RENAME(rgb24tobgr16)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -674,7 +625,6 @@ static inline void RENAME(rgb24tobgr16)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
const int b = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
@ -687,12 +637,9 @@ static inline void RENAME(rgb24to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -735,7 +682,6 @@ static inline void RENAME(rgb24to16)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
const int r = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
@ -748,12 +694,9 @@ static inline void RENAME(rgb24tobgr15)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -796,7 +739,6 @@ static inline void RENAME(rgb24tobgr15)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
const int b = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
@ -809,12 +751,9 @@ static inline void RENAME(rgb24to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint8_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
end = s + src_size; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*src):"memory"); |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"movq %0, %%mm7 \n\t" |
|
|
|
@ -857,7 +796,6 @@ static inline void RENAME(rgb24to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
const int r = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
@ -890,13 +828,10 @@ static inline void RENAME(rgb24to15)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
static inline void RENAME(rgb15tobgr24)(const uint8_t *src, uint8_t *dst, long src_size) |
|
|
|
|
{ |
|
|
|
|
const uint16_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint16_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint8_t *d = dst; |
|
|
|
|
const uint16_t *s = (const uint16_t*)src; |
|
|
|
|
end = s + src_size/2; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
mm_end = end - 7; |
|
|
|
|
while (s < mm_end) { |
|
|
|
@ -984,7 +919,6 @@ static inline void RENAME(rgb15tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr; |
|
|
|
|
bgr = *s++; |
|
|
|
@ -997,13 +931,10 @@ static inline void RENAME(rgb15tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
static inline void RENAME(rgb16tobgr24)(const uint8_t *src, uint8_t *dst, long src_size) |
|
|
|
|
{ |
|
|
|
|
const uint16_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint16_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint8_t *d = (uint8_t *)dst; |
|
|
|
|
const uint16_t *s = (const uint16_t *)src; |
|
|
|
|
end = s + src_size/2; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
mm_end = end - 7; |
|
|
|
|
while (s < mm_end) { |
|
|
|
@ -1090,7 +1021,6 @@ static inline void RENAME(rgb16tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr; |
|
|
|
|
bgr = *s++; |
|
|
|
@ -1122,13 +1052,10 @@ static inline void RENAME(rgb16tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
static inline void RENAME(rgb15to32)(const uint8_t *src, uint8_t *dst, long src_size) |
|
|
|
|
{ |
|
|
|
|
const uint16_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint16_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint8_t *d = dst; |
|
|
|
|
const uint16_t *s = (const uint16_t *)src; |
|
|
|
|
end = s + src_size/2; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
__asm__ volatile("pxor %%mm7,%%mm7 \n\t":::"memory"); |
|
|
|
|
__asm__ volatile("pcmpeqd %%mm6,%%mm6 \n\t":::"memory"); |
|
|
|
@ -1154,34 +1081,23 @@ static inline void RENAME(rgb15to32)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr; |
|
|
|
|
bgr = *s++; |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*d++ = 255; |
|
|
|
|
*d++ = (bgr&0x7C00)>>7; |
|
|
|
|
*d++ = (bgr&0x3E0)>>2; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
#else |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x3E0)>>2; |
|
|
|
|
*d++ = (bgr&0x7C00)>>7; |
|
|
|
|
*d++ = 255; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void RENAME(rgb16to32)(const uint8_t *src, uint8_t *dst, long src_size) |
|
|
|
|
{ |
|
|
|
|
const uint16_t *end; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
const uint16_t *mm_end; |
|
|
|
|
#endif |
|
|
|
|
uint8_t *d = dst; |
|
|
|
|
const uint16_t *s = (const uint16_t*)src; |
|
|
|
|
end = s + src_size/2; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(PREFETCH" %0"::"m"(*s):"memory"); |
|
|
|
|
__asm__ volatile("pxor %%mm7,%%mm7 \n\t":::"memory"); |
|
|
|
|
__asm__ volatile("pcmpeqd %%mm6,%%mm6 \n\t":::"memory"); |
|
|
|
@ -1207,21 +1123,13 @@ static inline void RENAME(rgb16to32)(const uint8_t *src, uint8_t *dst, long src_ |
|
|
|
|
} |
|
|
|
|
__asm__ volatile(SFENCE:::"memory"); |
|
|
|
|
__asm__ volatile(EMMS:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr; |
|
|
|
|
bgr = *s++; |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*d++ = 255; |
|
|
|
|
*d++ = (bgr&0xF800)>>8; |
|
|
|
|
*d++ = (bgr&0x7E0)>>3; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
#else |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x7E0)>>3; |
|
|
|
|
*d++ = (bgr&0xF800)>>8; |
|
|
|
|
*d++ = 255; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -1230,7 +1138,6 @@ static inline void RENAME(shuffle_bytes_2103)(const uint8_t *src, uint8_t *dst, |
|
|
|
|
x86_reg idx = 15 - src_size; |
|
|
|
|
const uint8_t *s = src-idx; |
|
|
|
|
uint8_t *d = dst-idx; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"test %0, %0 \n\t" |
|
|
|
|
"jns 2f \n\t" |
|
|
|
@ -1281,7 +1188,6 @@ static inline void RENAME(shuffle_bytes_2103)(const uint8_t *src, uint8_t *dst, |
|
|
|
|
: "+&r"(idx) |
|
|
|
|
: "r" (s), "r" (d), "m" (mask32b), "m" (mask32r), "m" (mmx_one) |
|
|
|
|
: "memory"); |
|
|
|
|
#endif |
|
|
|
|
for (; idx<15; idx+=4) { |
|
|
|
|
register int v = *(const uint32_t *)&s[idx], g = v & 0xff00ff00; |
|
|
|
|
v &= 0xff00ff; |
|
|
|
@ -1292,7 +1198,6 @@ static inline void RENAME(shuffle_bytes_2103)(const uint8_t *src, uint8_t *dst, |
|
|
|
|
static inline void RENAME(rgb24tobgr24)(const uint8_t *src, uint8_t *dst, long src_size) |
|
|
|
|
{ |
|
|
|
|
unsigned i; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
x86_reg mmx_size= 23 - src_size; |
|
|
|
|
__asm__ volatile ( |
|
|
|
|
"test %%"REG_a", %%"REG_a" \n\t" |
|
|
|
@ -1348,7 +1253,6 @@ static inline void RENAME(rgb24tobgr24)(const uint8_t *src, uint8_t *dst, long s |
|
|
|
|
src_size= 23-mmx_size; |
|
|
|
|
src-= src_size; |
|
|
|
|
dst-= src_size; |
|
|
|
|
#endif |
|
|
|
|
for (i=0; i<src_size; i+=3) { |
|
|
|
|
register uint8_t x; |
|
|
|
|
x = src[i + 2]; |
|
|
|
@ -1365,7 +1269,6 @@ static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
long y; |
|
|
|
|
const x86_reg chromWidth= width>>1; |
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
//FIXME handle 2 lines at once (fewer prefetches, reuse some chroma, but very likely memory-limited anyway)
|
|
|
|
|
__asm__ volatile( |
|
|
|
|
"xor %%"REG_a", %%"REG_a" \n\t" |
|
|
|
@ -1400,88 +1303,6 @@ static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" (chromWidth) |
|
|
|
|
: "%"REG_a |
|
|
|
|
); |
|
|
|
|
#else |
|
|
|
|
|
|
|
|
|
#if ARCH_ALPHA && HAVE_MVI |
|
|
|
|
#define pl2yuy2(n) \ |
|
|
|
|
y1 = yc[n]; \
|
|
|
|
|
y2 = yc2[n]; \
|
|
|
|
|
u = uc[n]; \
|
|
|
|
|
v = vc[n]; \
|
|
|
|
|
__asm__("unpkbw %1, %0" : "=r"(y1) : "r"(y1)); \
|
|
|
|
|
__asm__("unpkbw %1, %0" : "=r"(y2) : "r"(y2)); \
|
|
|
|
|
__asm__("unpkbl %1, %0" : "=r"(u) : "r"(u)); \
|
|
|
|
|
__asm__("unpkbl %1, %0" : "=r"(v) : "r"(v)); \
|
|
|
|
|
yuv1 = (u << 8) + (v << 24); \
|
|
|
|
|
yuv2 = yuv1 + y2; \
|
|
|
|
|
yuv1 += y1; \
|
|
|
|
|
qdst[n] = yuv1; \
|
|
|
|
|
qdst2[n] = yuv2; |
|
|
|
|
|
|
|
|
|
int i; |
|
|
|
|
uint64_t *qdst = (uint64_t *) dst; |
|
|
|
|
uint64_t *qdst2 = (uint64_t *) (dst + dstStride); |
|
|
|
|
const uint32_t *yc = (uint32_t *) ysrc; |
|
|
|
|
const uint32_t *yc2 = (uint32_t *) (ysrc + lumStride); |
|
|
|
|
const uint16_t *uc = (uint16_t*) usrc, *vc = (uint16_t*) vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i += 8) { |
|
|
|
|
uint64_t y1, y2, yuv1, yuv2; |
|
|
|
|
uint64_t u, v; |
|
|
|
|
/* Prefetch */ |
|
|
|
|
__asm__("ldq $31,64(%0)" :: "r"(yc)); |
|
|
|
|
__asm__("ldq $31,64(%0)" :: "r"(yc2)); |
|
|
|
|
__asm__("ldq $31,64(%0)" :: "r"(uc)); |
|
|
|
|
__asm__("ldq $31,64(%0)" :: "r"(vc)); |
|
|
|
|
|
|
|
|
|
pl2yuy2(0); |
|
|
|
|
pl2yuy2(1); |
|
|
|
|
pl2yuy2(2); |
|
|
|
|
pl2yuy2(3); |
|
|
|
|
|
|
|
|
|
yc += 4; |
|
|
|
|
yc2 += 4; |
|
|
|
|
uc += 4; |
|
|
|
|
vc += 4; |
|
|
|
|
qdst += 4; |
|
|
|
|
qdst2 += 4; |
|
|
|
|
} |
|
|
|
|
y++; |
|
|
|
|
ysrc += lumStride; |
|
|
|
|
dst += dstStride; |
|
|
|
|
|
|
|
|
|
#elif HAVE_FAST_64BIT |
|
|
|
|
int i; |
|
|
|
|
uint64_t *ldst = (uint64_t *) dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i += 2) { |
|
|
|
|
uint64_t k, l; |
|
|
|
|
k = yc[0] + (uc[0] << 8) + |
|
|
|
|
(yc[1] << 16) + (vc[0] << 24); |
|
|
|
|
l = yc[2] + (uc[1] << 8) + |
|
|
|
|
(yc[3] << 16) + (vc[1] << 24); |
|
|
|
|
*ldst++ = k + (l << 32); |
|
|
|
|
yc += 4; |
|
|
|
|
uc += 2; |
|
|
|
|
vc += 2; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
#else |
|
|
|
|
int i, *idst = (int32_t *) dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*idst++ = (yc[0] << 24)+ (uc[0] << 16) + |
|
|
|
|
(yc[1] << 8) + (vc[0] << 0); |
|
|
|
|
#else |
|
|
|
|
*idst++ = yc[0] + (uc[0] << 8) + |
|
|
|
|
(yc[1] << 16) + (vc[0] << 24); |
|
|
|
|
#endif |
|
|
|
|
yc += 2; |
|
|
|
|
uc++; |
|
|
|
|
vc++; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
#endif |
|
|
|
|
if ((y&(vertLumPerChroma-1)) == vertLumPerChroma-1) { |
|
|
|
|
usrc += chromStride; |
|
|
|
|
vsrc += chromStride; |
|
|
|
@ -1489,11 +1310,9 @@ static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
ysrc += lumStride; |
|
|
|
|
dst += dstStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
@ -1515,7 +1334,6 @@ static inline void RENAME(yuvPlanartouyvy)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
long y; |
|
|
|
|
const x86_reg chromWidth= width>>1; |
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
//FIXME handle 2 lines at once (fewer prefetches, reuse some chroma, but very likely memory-limited anyway)
|
|
|
|
|
__asm__ volatile( |
|
|
|
|
"xor %%"REG_a", %%"REG_a" \n\t" |
|
|
|
@ -1550,42 +1368,6 @@ static inline void RENAME(yuvPlanartouyvy)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" (chromWidth) |
|
|
|
|
: "%"REG_a |
|
|
|
|
); |
|
|
|
|
#else |
|
|
|
|
//FIXME adapt the Alpha ASM code from yv12->yuy2
|
|
|
|
|
|
|
|
|
|
#if HAVE_FAST_64BIT |
|
|
|
|
int i; |
|
|
|
|
uint64_t *ldst = (uint64_t *) dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i += 2) { |
|
|
|
|
uint64_t k, l; |
|
|
|
|
k = uc[0] + (yc[0] << 8) + |
|
|
|
|
(vc[0] << 16) + (yc[1] << 24); |
|
|
|
|
l = uc[1] + (yc[2] << 8) + |
|
|
|
|
(vc[1] << 16) + (yc[3] << 24); |
|
|
|
|
*ldst++ = k + (l << 32); |
|
|
|
|
yc += 4; |
|
|
|
|
uc += 2; |
|
|
|
|
vc += 2; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
#else |
|
|
|
|
int i, *idst = (int32_t *) dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*idst++ = (uc[0] << 24)+ (yc[0] << 16) + |
|
|
|
|
(vc[0] << 8) + (yc[1] << 0); |
|
|
|
|
#else |
|
|
|
|
*idst++ = uc[0] + (yc[0] << 8) + |
|
|
|
|
(vc[0] << 16) + (yc[1] << 24); |
|
|
|
|
#endif |
|
|
|
|
yc += 2; |
|
|
|
|
uc++; |
|
|
|
|
vc++; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
#endif |
|
|
|
|
if ((y&(vertLumPerChroma-1)) == vertLumPerChroma-1) { |
|
|
|
|
usrc += chromStride; |
|
|
|
|
vsrc += chromStride; |
|
|
|
@ -1593,11 +1375,9 @@ static inline void RENAME(yuvPlanartouyvy)(const uint8_t *ysrc, const uint8_t *u |
|
|
|
|
ysrc += lumStride; |
|
|
|
|
dst += dstStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
@ -1643,7 +1423,6 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t |
|
|
|
|
long y; |
|
|
|
|
const x86_reg chromWidth= width>>1; |
|
|
|
|
for (y=0; y<height; y+=2) { |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"xor %%"REG_a", %%"REG_a" \n\t" |
|
|
|
|
"pcmpeqw %%mm7, %%mm7 \n\t" |
|
|
|
@ -1725,32 +1504,14 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t |
|
|
|
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" (chromWidth) |
|
|
|
|
: "memory", "%"REG_a |
|
|
|
|
); |
|
|
|
|
#else |
|
|
|
|
long i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+0]; |
|
|
|
|
udst[i] = src[4*i+1]; |
|
|
|
|
ydst[2*i+1] = src[4*i+2]; |
|
|
|
|
vdst[i] = src[4*i+3]; |
|
|
|
|
} |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
|
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+0]; |
|
|
|
|
ydst[2*i+1] = src[4*i+2]; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void RENAME(planar2x)(const uint8_t *src, uint8_t *dst, long srcWidth, long srcHeight, long srcStride, long dstStride) |
|
|
|
@ -1859,11 +1620,9 @@ static inline void RENAME(planar2x)(const uint8_t *src, uint8_t *dst, long srcWi |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
|
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
@ -1879,7 +1638,6 @@ static inline void RENAME(uyvytoyv12)(const uint8_t *src, uint8_t *ydst, uint8_t |
|
|
|
|
long y; |
|
|
|
|
const x86_reg chromWidth= width>>1; |
|
|
|
|
for (y=0; y<height; y+=2) { |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile( |
|
|
|
|
"xor %%"REG_a", %%"REG_a" \n\t" |
|
|
|
|
"pcmpeqw %%mm7, %%mm7 \n\t" |
|
|
|
@ -1961,32 +1719,14 @@ static inline void RENAME(uyvytoyv12)(const uint8_t *src, uint8_t *ydst, uint8_t |
|
|
|
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" (chromWidth) |
|
|
|
|
: "memory", "%"REG_a |
|
|
|
|
); |
|
|
|
|
#else |
|
|
|
|
long i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
udst[i] = src[4*i+0]; |
|
|
|
|
ydst[2*i+0] = src[4*i+1]; |
|
|
|
|
vdst[i] = src[4*i+2]; |
|
|
|
|
ydst[2*i+1] = src[4*i+3]; |
|
|
|
|
} |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
|
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+1]; |
|
|
|
|
ydst[2*i+1] = src[4*i+3]; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
@ -2002,7 +1742,6 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_ |
|
|
|
|
{ |
|
|
|
|
long y; |
|
|
|
|
const x86_reg chromWidth= width>>1; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
for (y=0; y<height-2; y+=2) { |
|
|
|
|
long i; |
|
|
|
|
for (i=0; i<2; i++) { |
|
|
|
@ -2236,9 +1975,7 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_ |
|
|
|
|
__asm__ volatile(EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
:::"memory"); |
|
|
|
|
#else |
|
|
|
|
y=0; |
|
|
|
|
#endif |
|
|
|
|
|
|
|
|
|
for (; y<height; y+=2) { |
|
|
|
|
long i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
@ -2296,7 +2033,6 @@ static void RENAME(interleaveBytes)(const uint8_t *src1, const uint8_t *src2, ui |
|
|
|
|
for (h=0; h < height; h++) { |
|
|
|
|
long w; |
|
|
|
|
|
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
#if COMPILE_TEMPLATE_SSE2 |
|
|
|
|
__asm__( |
|
|
|
|
"xor %%"REG_a", %%"REG_a" \n\t" |
|
|
|
@ -2347,23 +2083,15 @@ static void RENAME(interleaveBytes)(const uint8_t *src1, const uint8_t *src2, ui |
|
|
|
|
dest[2*w+0] = src1[w]; |
|
|
|
|
dest[2*w+1] = src2[w]; |
|
|
|
|
} |
|
|
|
|
#else |
|
|
|
|
for (w=0; w < width; w++) { |
|
|
|
|
dest[2*w+0] = src1[w]; |
|
|
|
|
dest[2*w+1] = src2[w]; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
dest += dstStride; |
|
|
|
|
src1 += src1Stride; |
|
|
|
|
src2 += src2Stride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void RENAME(vu9_to_vu12)(const uint8_t *src1, const uint8_t *src2, |
|
|
|
@ -2375,17 +2103,14 @@ static inline void RENAME(vu9_to_vu12)(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
x86_reg y; |
|
|
|
|
long x,w,h; |
|
|
|
|
w=width/2; h=height/2; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__ volatile( |
|
|
|
|
PREFETCH" %0 \n\t" |
|
|
|
|
PREFETCH" %1 \n\t" |
|
|
|
|
::"m"(*(src1+srcStride1)),"m"(*(src2+srcStride2)):"memory"); |
|
|
|
|
#endif |
|
|
|
|
for (y=0;y<h;y++) { |
|
|
|
|
const uint8_t* s1=src1+srcStride1*(y>>1); |
|
|
|
|
uint8_t* d=dst1+dstStride1*y; |
|
|
|
|
x=0; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
for (;x<w-31;x+=32) { |
|
|
|
|
__asm__ volatile( |
|
|
|
|
PREFETCH" 32%1 \n\t" |
|
|
|
@ -2417,14 +2142,12 @@ static inline void RENAME(vu9_to_vu12)(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
:"m"(s1[x]) |
|
|
|
|
:"memory"); |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
for (;x<w;x++) d[2*x]=d[2*x+1]=s1[x]; |
|
|
|
|
} |
|
|
|
|
for (y=0;y<h;y++) { |
|
|
|
|
const uint8_t* s2=src2+srcStride2*(y>>1); |
|
|
|
|
uint8_t* d=dst2+dstStride2*y; |
|
|
|
|
x=0; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
for (;x<w-31;x+=32) { |
|
|
|
|
__asm__ volatile( |
|
|
|
|
PREFETCH" 32%1 \n\t" |
|
|
|
@ -2456,16 +2179,13 @@ static inline void RENAME(vu9_to_vu12)(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
:"m"(s2[x]) |
|
|
|
|
:"memory"); |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
for (;x<w;x++) d[2*x]=d[2*x+1]=s2[x]; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void RENAME(yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2, const uint8_t *src3, |
|
|
|
@ -2483,7 +2203,6 @@ static inline void RENAME(yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2 |
|
|
|
|
const uint8_t* vp=src3+srcStride3*(y>>2); |
|
|
|
|
uint8_t* d=dst+dstStride*y; |
|
|
|
|
x=0; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
for (;x<w-7;x+=8) { |
|
|
|
|
__asm__ volatile( |
|
|
|
|
PREFETCH" 32(%1, %0) \n\t" |
|
|
|
@ -2536,7 +2255,6 @@ static inline void RENAME(yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2 |
|
|
|
|
: "r"(yp), "r" (up), "r"(vp), "r"(d) |
|
|
|
|
:"memory"); |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
for (; x<w; x++) { |
|
|
|
|
const long x2 = x<<2; |
|
|
|
|
d[8*x+0] = yp[x2]; |
|
|
|
@ -2549,13 +2267,11 @@ static inline void RENAME(yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2 |
|
|
|
|
d[8*x+7] = vp[x]; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void RENAME(extract_even)(const uint8_t *src, uint8_t *dst, x86_reg count) |
|
|
|
@ -2564,7 +2280,6 @@ static void RENAME(extract_even)(const uint8_t *src, uint8_t *dst, x86_reg count |
|
|
|
|
src += 2*count; |
|
|
|
|
count= - count; |
|
|
|
|
|
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
if(count <= -16) { |
|
|
|
|
count += 15; |
|
|
|
|
__asm__ volatile( |
|
|
|
@ -2590,7 +2305,6 @@ static void RENAME(extract_even)(const uint8_t *src, uint8_t *dst, x86_reg count |
|
|
|
|
); |
|
|
|
|
count -= 15; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
while(count<0) { |
|
|
|
|
dst[count]= src[2*count]; |
|
|
|
|
count++; |
|
|
|
@ -2603,7 +2317,6 @@ static void RENAME(extract_even2)(const uint8_t *src, uint8_t *dst0, uint8_t *ds |
|
|
|
|
dst1+= count; |
|
|
|
|
src += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
if(count <= -8) { |
|
|
|
|
count += 7; |
|
|
|
|
__asm__ volatile( |
|
|
|
@ -2637,7 +2350,6 @@ static void RENAME(extract_even2)(const uint8_t *src, uint8_t *dst0, uint8_t *ds |
|
|
|
|
); |
|
|
|
|
count -= 7; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= src[4*count+0]; |
|
|
|
|
dst1[count]= src[4*count+2]; |
|
|
|
@ -2704,7 +2416,6 @@ static void RENAME(extract_odd2)(const uint8_t *src, uint8_t *dst0, uint8_t *dst |
|
|
|
|
dst1+= count; |
|
|
|
|
src += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
if(count <= -8) { |
|
|
|
|
count += 7; |
|
|
|
|
__asm__ volatile( |
|
|
|
@ -2738,7 +2449,6 @@ static void RENAME(extract_odd2)(const uint8_t *src, uint8_t *dst0, uint8_t *dst |
|
|
|
|
); |
|
|
|
|
count -= 7; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
src++; |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= src[4*count+0]; |
|
|
|
@ -2820,13 +2530,11 @@ static void RENAME(yuyvtoyuv420)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, co |
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void RENAME(yuyvtoyuv422)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, const uint8_t *src, |
|
|
|
@ -2845,13 +2553,11 @@ static void RENAME(yuyvtoyuv422)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, co |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void RENAME(uyvytoyuv420)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, const uint8_t *src, |
|
|
|
@ -2872,13 +2578,11 @@ static void RENAME(uyvytoyuv420)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, co |
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void RENAME(uyvytoyuv422)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, const uint8_t *src, |
|
|
|
@ -2897,13 +2601,11 @@ static void RENAME(uyvytoyuv422)(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, co |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
} |
|
|
|
|
#if COMPILE_TEMPLATE_MMX |
|
|
|
|
__asm__( |
|
|
|
|
EMMS" \n\t" |
|
|
|
|
SFENCE" \n\t" |
|
|
|
|
::: "memory" |
|
|
|
|
); |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void RENAME(rgb2rgb_init)(void) |
|
|
|
|