|
|
|
@ -26,7 +26,8 @@ |
|
|
|
|
|
|
|
|
|
#include <stddef.h> |
|
|
|
|
|
|
|
|
|
static inline void rgb24tobgr32_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb24tobgr32_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
uint8_t *dest = dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
@ -34,12 +35,12 @@ static inline void rgb24tobgr32_c(const uint8_t *src, uint8_t *dst, int src_size |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
/* RGB24 (= R,G,B) -> RGB32 (= A,B,G,R) */ |
|
|
|
|
/* RGB24 (= R, G, B) -> RGB32 (= A, B, G, R) */ |
|
|
|
|
*dest++ = 255; |
|
|
|
|
*dest++ = s[2]; |
|
|
|
|
*dest++ = s[1]; |
|
|
|
|
*dest++ = s[0]; |
|
|
|
|
s+=3; |
|
|
|
|
s += 3; |
|
|
|
|
#else |
|
|
|
|
*dest++ = *s++; |
|
|
|
|
*dest++ = *s++; |
|
|
|
@ -49,7 +50,8 @@ static inline void rgb24tobgr32_c(const uint8_t *src, uint8_t *dst, int src_size |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb32tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb32tobgr24_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
uint8_t *dest = dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
@ -57,7 +59,7 @@ static inline void rgb32tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
/* RGB32 (= A,B,G,R) -> RGB24 (= R,G,B) */ |
|
|
|
|
/* RGB32 (= A, B, G, R) -> RGB24 (= R, G, B) */ |
|
|
|
|
s++; |
|
|
|
|
dest[2] = *s++; |
|
|
|
|
dest[1] = *s++; |
|
|
|
@ -73,189 +75,207 @@ static inline void rgb32tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
original by Strepto/Astral |
|
|
|
|
ported to gcc & bugfixed: A'rpi |
|
|
|
|
MMX2, 3DNOW optimization by Nick Kurshev |
|
|
|
|
32-bit C version, and and&add trick by Michael Niedermayer |
|
|
|
|
*/ |
|
|
|
|
* original by Strepto/Astral |
|
|
|
|
* ported to gcc & bugfixed: A'rpi |
|
|
|
|
* MMX2, 3DNOW optimization by Nick Kurshev |
|
|
|
|
* 32-bit C version, and and&add trick by Michael Niedermayer |
|
|
|
|
*/ |
|
|
|
|
static inline void rgb15to16_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
register const uint8_t* s=src; |
|
|
|
|
register uint8_t* d=dst; |
|
|
|
|
register uint8_t *d = dst; |
|
|
|
|
register const uint8_t *s = src; |
|
|
|
|
register const uint8_t *end = s + src_size; |
|
|
|
|
const uint8_t *mm_end = end - 3; |
|
|
|
|
|
|
|
|
|
while (s < mm_end) { |
|
|
|
|
register unsigned x= *((const uint32_t *)s); |
|
|
|
|
*((uint32_t *)d) = (x&0x7FFF7FFF) + (x&0x7FE07FE0); |
|
|
|
|
d+=4; |
|
|
|
|
s+=4; |
|
|
|
|
register unsigned x = *((const uint32_t *)s); |
|
|
|
|
*((uint32_t *)d) = (x & 0x7FFF7FFF) + (x & 0x7FE07FE0); |
|
|
|
|
d += 4; |
|
|
|
|
s += 4; |
|
|
|
|
} |
|
|
|
|
if (s < end) { |
|
|
|
|
register unsigned short x= *((const uint16_t *)s); |
|
|
|
|
*((uint16_t *)d) = (x&0x7FFF) + (x&0x7FE0); |
|
|
|
|
register unsigned short x = *((const uint16_t *)s); |
|
|
|
|
*((uint16_t *)d) = (x & 0x7FFF) + (x & 0x7FE0); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb16to15_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
register const uint8_t* s=src; |
|
|
|
|
register uint8_t* d=dst; |
|
|
|
|
register uint8_t *d = dst; |
|
|
|
|
register const uint8_t *s = src; |
|
|
|
|
register const uint8_t *end = s + src_size; |
|
|
|
|
const uint8_t *mm_end = end - 3; |
|
|
|
|
|
|
|
|
|
while (s < mm_end) { |
|
|
|
|
register uint32_t x= *((const uint32_t*)s); |
|
|
|
|
*((uint32_t *)d) = ((x>>1)&0x7FE07FE0) | (x&0x001F001F); |
|
|
|
|
s+=4; |
|
|
|
|
d+=4; |
|
|
|
|
register uint32_t x = *((const uint32_t *)s); |
|
|
|
|
*((uint32_t *)d) = ((x >> 1) & 0x7FE07FE0) | (x & 0x001F001F); |
|
|
|
|
s += 4; |
|
|
|
|
d += 4; |
|
|
|
|
} |
|
|
|
|
if (s < end) { |
|
|
|
|
register uint16_t x= *((const uint16_t*)s); |
|
|
|
|
*((uint16_t *)d) = ((x>>1)&0x7FE0) | (x&0x001F); |
|
|
|
|
register uint16_t x = *((const uint16_t *)s); |
|
|
|
|
*((uint16_t *)d) = ((x >> 1) & 0x7FE0) | (x & 0x001F); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb32to16_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xFF)>>3) + ((rgb&0xFC00)>>5) + ((rgb&0xF80000)>>8); |
|
|
|
|
register int rgb = *(const uint32_t *)s; |
|
|
|
|
s += 4; |
|
|
|
|
*d++ = ((rgb & 0xFF) >> 3) + |
|
|
|
|
((rgb & 0xFC00) >> 5) + |
|
|
|
|
((rgb & 0xF80000) >> 8); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb32tobgr16_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb32tobgr16_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xF8)<<8) + ((rgb&0xFC00)>>5) + ((rgb&0xF80000)>>19); |
|
|
|
|
register int rgb = *(const uint32_t *)s; |
|
|
|
|
s += 4; |
|
|
|
|
*d++ = ((rgb & 0xF8) << 8) + |
|
|
|
|
((rgb & 0xFC00) >> 5) + |
|
|
|
|
((rgb & 0xF80000) >> 19); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb32to15_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xFF)>>3) + ((rgb&0xF800)>>6) + ((rgb&0xF80000)>>9); |
|
|
|
|
register int rgb = *(const uint32_t *)s; |
|
|
|
|
s += 4; |
|
|
|
|
*d++ = ((rgb & 0xFF) >> 3) + |
|
|
|
|
((rgb & 0xF800) >> 6) + |
|
|
|
|
((rgb & 0xF80000) >> 9); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb32tobgr15_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb32tobgr15_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register int rgb = *(const uint32_t*)s; s += 4; |
|
|
|
|
*d++ = ((rgb&0xF8)<<7) + ((rgb&0xF800)>>6) + ((rgb&0xF80000)>>19); |
|
|
|
|
register int rgb = *(const uint32_t *)s; |
|
|
|
|
s += 4; |
|
|
|
|
*d++ = ((rgb & 0xF8) << 7) + |
|
|
|
|
((rgb & 0xF800) >> 6) + |
|
|
|
|
((rgb & 0xF80000) >> 19); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb24tobgr16_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb24tobgr16_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
const int b = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
|
const int r = *s++; |
|
|
|
|
*d++ = (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
|
|
|
*d++ = (b >> 3) | ((g & 0xFC) << 3) | ((r & 0xF8) << 8); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb24to16_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
const int r = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
|
const int b = *s++; |
|
|
|
|
*d++ = (b>>3) | ((g&0xFC)<<3) | ((r&0xF8)<<8); |
|
|
|
|
*d++ = (b >> 3) | ((g & 0xFC) << 3) | ((r & 0xF8) << 8); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb24tobgr15_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb24tobgr15_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
const int b = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
|
const int r = *s++; |
|
|
|
|
*d++ = (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
|
|
|
*d++ = (b >> 3) | ((g & 0xF8) << 2) | ((r & 0xF8) << 7); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb24to15_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
uint16_t *d = (uint16_t *)dst; |
|
|
|
|
const uint8_t *s = src; |
|
|
|
|
const uint8_t *end = s + src_size; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
const int r = *s++; |
|
|
|
|
const int g = *s++; |
|
|
|
|
const int b = *s++; |
|
|
|
|
*d++ = (b>>3) | ((g&0xF8)<<2) | ((r&0xF8)<<7); |
|
|
|
|
*d++ = (b >> 3) | ((g & 0xF8) << 2) | ((r & 0xF8) << 7); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
I use less accurate approximation here by simply left-shifting the input |
|
|
|
|
value and filling the low order bits with zeroes. This method improves PNG |
|
|
|
|
compression but this scheme cannot reproduce white exactly, since it does |
|
|
|
|
not generate an all-ones maximum value; the net effect is to darken the |
|
|
|
|
image slightly. |
|
|
|
|
|
|
|
|
|
The better method should be "left bit replication": |
|
|
|
|
|
|
|
|
|
4 3 2 1 0 |
|
|
|
|
--------- |
|
|
|
|
1 1 0 1 1 |
|
|
|
|
|
|
|
|
|
7 6 5 4 3 2 1 0 |
|
|
|
|
---------------- |
|
|
|
|
1 1 0 1 1 1 1 0 |
|
|
|
|
|=======| |===| |
|
|
|
|
| leftmost bits repeated to fill open bits |
|
|
|
|
| |
|
|
|
|
original bits |
|
|
|
|
*/ |
|
|
|
|
static inline void rgb15tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
* I use less accurate approximation here by simply left-shifting the input |
|
|
|
|
* value and filling the low order bits with zeroes. This method improves PNG |
|
|
|
|
* compression but this scheme cannot reproduce white exactly, since it does |
|
|
|
|
* not generate an all-ones maximum value; the net effect is to darken the |
|
|
|
|
* image slightly. |
|
|
|
|
* |
|
|
|
|
* The better method should be "left bit replication": |
|
|
|
|
* |
|
|
|
|
* 4 3 2 1 0 |
|
|
|
|
* --------- |
|
|
|
|
* 1 1 0 1 1 |
|
|
|
|
* |
|
|
|
|
* 7 6 5 4 3 2 1 0 |
|
|
|
|
* ---------------- |
|
|
|
|
* 1 1 0 1 1 1 1 0 |
|
|
|
|
* |=======| |===| |
|
|
|
|
* | leftmost bits repeated to fill open bits |
|
|
|
|
* | |
|
|
|
|
* original bits |
|
|
|
|
*/ |
|
|
|
|
static inline void rgb15tobgr24_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
uint8_t *d = dst; |
|
|
|
|
const uint16_t *s = (const uint16_t*)src; |
|
|
|
|
const uint16_t *s = (const uint16_t *)src; |
|
|
|
|
const uint16_t *end = s + src_size / 2; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr = *s++; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x3E0)>>2; |
|
|
|
|
*d++ = (bgr&0x7C00)>>7; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
*d++ = (bgr & 0x3E0) >> 2; |
|
|
|
|
*d++ = (bgr & 0x7C00) >> 7; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb16tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb16tobgr24_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
uint8_t *d = (uint8_t *)dst; |
|
|
|
|
const uint16_t *s = (const uint16_t *)src; |
|
|
|
@ -263,9 +283,9 @@ static inline void rgb16tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr = *s++; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x7E0)>>3; |
|
|
|
|
*d++ = (bgr&0xF800)>>8; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
*d++ = (bgr & 0x7E0) >> 3; |
|
|
|
|
*d++ = (bgr & 0xF800) >> 8; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -279,13 +299,13 @@ static inline void rgb15to32_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
register uint16_t bgr = *s++; |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*d++ = 255; |
|
|
|
|
*d++ = (bgr&0x7C00)>>7; |
|
|
|
|
*d++ = (bgr&0x3E0)>>2; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr & 0x7C00) >> 7; |
|
|
|
|
*d++ = (bgr & 0x3E0) >> 2; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
#else |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x3E0)>>2; |
|
|
|
|
*d++ = (bgr&0x7C00)>>7; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
*d++ = (bgr & 0x3E0) >> 2; |
|
|
|
|
*d++ = (bgr & 0x7C00) >> 7; |
|
|
|
|
*d++ = 255; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
@ -294,41 +314,44 @@ static inline void rgb15to32_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void rgb16to32_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
uint8_t *d = dst; |
|
|
|
|
const uint16_t *s = (const uint16_t*)src; |
|
|
|
|
const uint16_t *s = (const uint16_t *)src; |
|
|
|
|
const uint16_t *end = s + src_size / 2; |
|
|
|
|
|
|
|
|
|
while (s < end) { |
|
|
|
|
register uint16_t bgr = *s++; |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*d++ = 255; |
|
|
|
|
*d++ = (bgr&0xF800)>>8; |
|
|
|
|
*d++ = (bgr&0x7E0)>>3; |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr & 0xF800) >> 8; |
|
|
|
|
*d++ = (bgr & 0x7E0) >> 3; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
#else |
|
|
|
|
*d++ = (bgr&0x1F)<<3; |
|
|
|
|
*d++ = (bgr&0x7E0)>>3; |
|
|
|
|
*d++ = (bgr&0xF800)>>8; |
|
|
|
|
*d++ = (bgr & 0x1F) << 3; |
|
|
|
|
*d++ = (bgr & 0x7E0) >> 3; |
|
|
|
|
*d++ = (bgr & 0xF800) >> 8; |
|
|
|
|
*d++ = 255; |
|
|
|
|
#endif |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void shuffle_bytes_2103_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
static inline void shuffle_bytes_2103_c(const uint8_t *src, uint8_t *dst, |
|
|
|
|
int src_size) |
|
|
|
|
{ |
|
|
|
|
int idx = 15 - src_size; |
|
|
|
|
const uint8_t *s = src-idx; |
|
|
|
|
uint8_t *d = dst-idx; |
|
|
|
|
for (; idx<15; idx+=4) { |
|
|
|
|
const uint8_t *s = src - idx; |
|
|
|
|
uint8_t *d = dst - idx; |
|
|
|
|
|
|
|
|
|
for (; idx < 15; idx += 4) { |
|
|
|
|
register int v = *(const uint32_t *)&s[idx], g = v & 0xff00ff00; |
|
|
|
|
v &= 0xff00ff; |
|
|
|
|
*(uint32_t *)&d[idx] = (v>>16) + g + (v<<16); |
|
|
|
|
*(uint32_t *)&d[idx] = (v >> 16) + g + (v << 16); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static inline void rgb24tobgr24_c(const uint8_t *src, uint8_t *dst, int src_size) |
|
|
|
|
{ |
|
|
|
|
unsigned i; |
|
|
|
|
for (i=0; i<src_size; i+=3) { |
|
|
|
|
|
|
|
|
|
for (i = 0; i < src_size; i += 3) { |
|
|
|
|
register uint8_t x = src[i + 2]; |
|
|
|
|
dst[i + 1] = src[i + 1]; |
|
|
|
|
dst[i + 2] = src[i + 0]; |
|
|
|
@ -344,9 +367,10 @@ static inline void yuvPlanartoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
{ |
|
|
|
|
int y, i; |
|
|
|
|
const int chromWidth = width >> 1; |
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
|
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
#if HAVE_FAST_64BIT |
|
|
|
|
uint64_t *ldst = (uint64_t *) dst; |
|
|
|
|
uint64_t *ldst = (uint64_t *)dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i += 2) { |
|
|
|
|
uint64_t k = yc[0] + (uc[0] << 8) + |
|
|
|
@ -360,11 +384,12 @@ static inline void yuvPlanartoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
#else |
|
|
|
|
int *idst = (int32_t *) dst; |
|
|
|
|
int *idst = (int32_t *)dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
|
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*idst++ = (yc[0] << 24)+ (uc[0] << 16) + |
|
|
|
|
*idst++ = (yc[0] << 24) + (uc[0] << 16) + |
|
|
|
|
(yc[1] << 8) + (vc[0] << 0); |
|
|
|
|
#else |
|
|
|
|
*idst++ = yc[0] + (uc[0] << 8) + |
|
|
|
@ -375,7 +400,7 @@ static inline void yuvPlanartoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
vc++; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
if ((y&(vertLumPerChroma-1)) == vertLumPerChroma-1) { |
|
|
|
|
if ((y & (vertLumPerChroma - 1)) == vertLumPerChroma - 1) { |
|
|
|
|
usrc += chromStride; |
|
|
|
|
vsrc += chromStride; |
|
|
|
|
} |
|
|
|
@ -390,9 +415,8 @@ static inline void yuvPlanartoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
*/ |
|
|
|
|
static inline void yv12toyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
const uint8_t *vsrc, uint8_t *dst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int dstStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
//FIXME interpolate chroma
|
|
|
|
|
yuvPlanartoyuy2_c(ysrc, usrc, vsrc, dst, width, height, lumStride, |
|
|
|
@ -407,9 +431,10 @@ static inline void yuvPlanartouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
{ |
|
|
|
|
int y, i; |
|
|
|
|
const int chromWidth = width >> 1; |
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
|
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
#if HAVE_FAST_64BIT |
|
|
|
|
uint64_t *ldst = (uint64_t *) dst; |
|
|
|
|
uint64_t *ldst = (uint64_t *)dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
for (i = 0; i < chromWidth; i += 2) { |
|
|
|
|
uint64_t k = uc[0] + (yc[0] << 8) + |
|
|
|
@ -423,11 +448,12 @@ static inline void yuvPlanartouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
#else |
|
|
|
|
int *idst = (int32_t *) dst; |
|
|
|
|
int *idst = (int32_t *)dst; |
|
|
|
|
const uint8_t *yc = ysrc, *uc = usrc, *vc = vsrc; |
|
|
|
|
|
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
#if HAVE_BIGENDIAN |
|
|
|
|
*idst++ = (uc[0] << 24)+ (yc[0] << 16) + |
|
|
|
|
*idst++ = (uc[0] << 24) + (yc[0] << 16) + |
|
|
|
|
(vc[0] << 8) + (yc[1] << 0); |
|
|
|
|
#else |
|
|
|
|
*idst++ = uc[0] + (yc[0] << 8) + |
|
|
|
@ -438,7 +464,7 @@ static inline void yuvPlanartouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
vc++; |
|
|
|
|
} |
|
|
|
|
#endif |
|
|
|
|
if ((y&(vertLumPerChroma-1)) == vertLumPerChroma-1) { |
|
|
|
|
if ((y & (vertLumPerChroma - 1)) == vertLumPerChroma - 1) { |
|
|
|
|
usrc += chromStride; |
|
|
|
|
vsrc += chromStride; |
|
|
|
|
} |
|
|
|
@ -453,9 +479,8 @@ static inline void yuvPlanartouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
*/ |
|
|
|
|
static inline void yv12touyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
const uint8_t *vsrc, uint8_t *dst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int dstStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
//FIXME interpolate chroma
|
|
|
|
|
yuvPlanartouyvy_c(ysrc, usrc, vsrc, dst, width, height, lumStride, |
|
|
|
@ -467,9 +492,8 @@ static inline void yv12touyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
*/ |
|
|
|
|
static inline void yuv422ptouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
const uint8_t *vsrc, uint8_t *dst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int dstStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
yuvPlanartouyvy_c(ysrc, usrc, vsrc, dst, width, height, lumStride, |
|
|
|
|
chromStride, dstStride, 1); |
|
|
|
@ -480,9 +504,8 @@ static inline void yuv422ptouyvy_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
*/ |
|
|
|
|
static inline void yuv422ptoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
const uint8_t *vsrc, uint8_t *dst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int dstStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
yuvPlanartoyuy2_c(ysrc, usrc, vsrc, dst, width, height, lumStride, |
|
|
|
|
chromStride, dstStride, 1); |
|
|
|
@ -494,26 +517,26 @@ static inline void yuv422ptoyuy2_c(const uint8_t *ysrc, const uint8_t *usrc, |
|
|
|
|
*/ |
|
|
|
|
static inline void yuy2toyv12_c(const uint8_t *src, uint8_t *ydst, |
|
|
|
|
uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int srcStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth = width >> 1; |
|
|
|
|
for (y=0; y<height; y+=2) { |
|
|
|
|
|
|
|
|
|
for (y = 0; y < height; y += 2) { |
|
|
|
|
int i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+0]; |
|
|
|
|
udst[i] = src[4*i+1]; |
|
|
|
|
ydst[2*i+1] = src[4*i+2]; |
|
|
|
|
vdst[i] = src[4*i+3]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
ydst[2 * i + 0] = src[4 * i + 0]; |
|
|
|
|
udst[i] = src[4 * i + 1]; |
|
|
|
|
ydst[2 * i + 1] = src[4 * i + 2]; |
|
|
|
|
vdst[i] = src[4 * i + 3]; |
|
|
|
|
} |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
|
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+0]; |
|
|
|
|
ydst[2*i+1] = src[4*i+2]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
ydst[2 * i + 0] = src[4 * i + 0]; |
|
|
|
|
ydst[2 * i + 1] = src[4 * i + 2]; |
|
|
|
|
} |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
@ -525,46 +548,46 @@ static inline void yuy2toyv12_c(const uint8_t *src, uint8_t *ydst, |
|
|
|
|
static inline void planar2x_c(const uint8_t *src, uint8_t *dst, int srcWidth, |
|
|
|
|
int srcHeight, int srcStride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
int x,y; |
|
|
|
|
int x, y; |
|
|
|
|
|
|
|
|
|
dst[0]= src[0]; |
|
|
|
|
dst[0] = src[0]; |
|
|
|
|
|
|
|
|
|
// first line
|
|
|
|
|
for (x=0; x<srcWidth-1; x++) { |
|
|
|
|
dst[2*x+1]= (3*src[x] + src[x+1])>>2; |
|
|
|
|
dst[2*x+2]= ( src[x] + 3*src[x+1])>>2; |
|
|
|
|
for (x = 0; x < srcWidth - 1; x++) { |
|
|
|
|
dst[2 * x + 1] = (3 * src[x] + src[x + 1]) >> 2; |
|
|
|
|
dst[2 * x + 2] = (src[x] + 3 * src[x + 1]) >> 2; |
|
|
|
|
} |
|
|
|
|
dst[2*srcWidth-1]= src[srcWidth-1]; |
|
|
|
|
dst[2 * srcWidth - 1] = src[srcWidth - 1]; |
|
|
|
|
|
|
|
|
|
dst+= dstStride; |
|
|
|
|
dst += dstStride; |
|
|
|
|
|
|
|
|
|
for (y=1; y<srcHeight; y++) { |
|
|
|
|
for (y = 1; y < srcHeight; y++) { |
|
|
|
|
const int mmxSize = 1; |
|
|
|
|
|
|
|
|
|
dst[0 ]= (3*src[0] + src[srcStride])>>2; |
|
|
|
|
dst[dstStride]= ( src[0] + 3*src[srcStride])>>2; |
|
|
|
|
dst[0] = (src[0] * 3 + src[srcStride]) >> 2; |
|
|
|
|
dst[dstStride] = (src[0] + 3 * src[srcStride]) >> 2; |
|
|
|
|
|
|
|
|
|
for (x=mmxSize-1; x<srcWidth-1; x++) { |
|
|
|
|
dst[2*x +1]= (3*src[x+0] + src[x+srcStride+1])>>2; |
|
|
|
|
dst[2*x+dstStride+2]= ( src[x+0] + 3*src[x+srcStride+1])>>2; |
|
|
|
|
dst[2*x+dstStride+1]= ( src[x+1] + 3*src[x+srcStride ])>>2; |
|
|
|
|
dst[2*x +2]= (3*src[x+1] + src[x+srcStride ])>>2; |
|
|
|
|
for (x = mmxSize - 1; x < srcWidth - 1; x++) { |
|
|
|
|
dst[2 * x + 1] = (src[x + 0] * 3 + src[x + srcStride + 1]) >> 2; |
|
|
|
|
dst[2 * x + dstStride + 2] = (src[x + 0] + 3 * src[x + srcStride + 1]) >> 2; |
|
|
|
|
dst[2 * x + dstStride + 1] = (src[x + 1] + 3 * src[x + srcStride]) >> 2; |
|
|
|
|
dst[2 * x + 2] = (src[x + 1] * 3 + src[x + srcStride]) >> 2; |
|
|
|
|
} |
|
|
|
|
dst[srcWidth*2 -1 ]= (3*src[srcWidth-1] + src[srcWidth-1 + srcStride])>>2; |
|
|
|
|
dst[srcWidth*2 -1 + dstStride]= ( src[srcWidth-1] + 3*src[srcWidth-1 + srcStride])>>2; |
|
|
|
|
dst[srcWidth * 2 - 1] = (src[srcWidth - 1] * 3 + src[srcWidth - 1 + srcStride]) >> 2; |
|
|
|
|
dst[srcWidth * 2 - 1 + dstStride] = (src[srcWidth - 1] + 3 * src[srcWidth - 1 + srcStride]) >> 2; |
|
|
|
|
|
|
|
|
|
dst+=dstStride*2; |
|
|
|
|
src+=srcStride; |
|
|
|
|
dst += dstStride * 2; |
|
|
|
|
src += srcStride; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// last line
|
|
|
|
|
dst[0]= src[0]; |
|
|
|
|
dst[0] = src[0]; |
|
|
|
|
|
|
|
|
|
for (x=0; x<srcWidth-1; x++) { |
|
|
|
|
dst[2*x+1]= (3*src[x] + src[x+1])>>2; |
|
|
|
|
dst[2*x+2]= ( src[x] + 3*src[x+1])>>2; |
|
|
|
|
for (x = 0; x < srcWidth - 1; x++) { |
|
|
|
|
dst[2 * x + 1] = (src[x] * 3 + src[x + 1]) >> 2; |
|
|
|
|
dst[2 * x + 2] = (src[x] + 3 * src[x + 1]) >> 2; |
|
|
|
|
} |
|
|
|
|
dst[2*srcWidth-1]= src[srcWidth-1]; |
|
|
|
|
dst[2 * srcWidth - 1] = src[srcWidth - 1]; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/**
|
|
|
|
@ -575,26 +598,26 @@ static inline void planar2x_c(const uint8_t *src, uint8_t *dst, int srcWidth, |
|
|
|
|
*/ |
|
|
|
|
static inline void uyvytoyv12_c(const uint8_t *src, uint8_t *ydst, |
|
|
|
|
uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int width, int height, |
|
|
|
|
int lumStride, int chromStride, |
|
|
|
|
int srcStride) |
|
|
|
|
int width, int height, int lumStride, |
|
|
|
|
int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth = width >> 1; |
|
|
|
|
for (y=0; y<height; y+=2) { |
|
|
|
|
|
|
|
|
|
for (y = 0; y < height; y += 2) { |
|
|
|
|
int i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
udst[i] = src[4*i+0]; |
|
|
|
|
ydst[2*i+0] = src[4*i+1]; |
|
|
|
|
vdst[i] = src[4*i+2]; |
|
|
|
|
ydst[2*i+1] = src[4*i+3]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
udst[i] = src[4 * i + 0]; |
|
|
|
|
ydst[2 * i + 0] = src[4 * i + 1]; |
|
|
|
|
vdst[i] = src[4 * i + 2]; |
|
|
|
|
ydst[2 * i + 1] = src[4 * i + 3]; |
|
|
|
|
} |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
|
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
ydst[2*i+0] = src[4*i+1]; |
|
|
|
|
ydst[2*i+1] = src[4*i+3]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
ydst[2 * i + 0] = src[4 * i + 1]; |
|
|
|
|
ydst[2 * i + 1] = src[4 * i + 3]; |
|
|
|
|
} |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
@ -619,44 +642,44 @@ void rgb24toyv12_c(const uint8_t *src, uint8_t *ydst, uint8_t *udst, |
|
|
|
|
|
|
|
|
|
for (y = 0; y < height; y += 2) { |
|
|
|
|
int i; |
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
unsigned int b = src[6*i+0]; |
|
|
|
|
unsigned int g = src[6*i+1]; |
|
|
|
|
unsigned int r = src[6*i+2]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
unsigned int b = src[6 * i + 0]; |
|
|
|
|
unsigned int g = src[6 * i + 1]; |
|
|
|
|
unsigned int r = src[6 * i + 2]; |
|
|
|
|
|
|
|
|
|
unsigned int Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
|
|
|
unsigned int V = ((RV*r + GV*g + BV*b)>>RGB2YUV_SHIFT) + 128; |
|
|
|
|
unsigned int U = ((RU*r + GU*g + BU*b)>>RGB2YUV_SHIFT) + 128; |
|
|
|
|
unsigned int Y = ((RY * r + GY * g + BY * b) >> RGB2YUV_SHIFT) + 16; |
|
|
|
|
unsigned int V = ((RV * r + GV * g + BV * b) >> RGB2YUV_SHIFT) + 128; |
|
|
|
|
unsigned int U = ((RU * r + GU * g + BU * b) >> RGB2YUV_SHIFT) + 128; |
|
|
|
|
|
|
|
|
|
udst[i] = U; |
|
|
|
|
vdst[i] = V; |
|
|
|
|
ydst[2*i] = Y; |
|
|
|
|
ydst[2 * i] = Y; |
|
|
|
|
|
|
|
|
|
b = src[6*i+3]; |
|
|
|
|
g = src[6*i+4]; |
|
|
|
|
r = src[6*i+5]; |
|
|
|
|
b = src[6 * i + 3]; |
|
|
|
|
g = src[6 * i + 4]; |
|
|
|
|
r = src[6 * i + 5]; |
|
|
|
|
|
|
|
|
|
Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
|
|
|
ydst[2*i+1] = Y; |
|
|
|
|
Y = ((RY * r + GY * g + BY * b) >> RGB2YUV_SHIFT) + 16; |
|
|
|
|
ydst[2 * i + 1] = Y; |
|
|
|
|
} |
|
|
|
|
ydst += lumStride; |
|
|
|
|
src += srcStride; |
|
|
|
|
|
|
|
|
|
for (i=0; i<chromWidth; i++) { |
|
|
|
|
unsigned int b = src[6*i+0]; |
|
|
|
|
unsigned int g = src[6*i+1]; |
|
|
|
|
unsigned int r = src[6*i+2]; |
|
|
|
|
for (i = 0; i < chromWidth; i++) { |
|
|
|
|
unsigned int b = src[6 * i + 0]; |
|
|
|
|
unsigned int g = src[6 * i + 1]; |
|
|
|
|
unsigned int r = src[6 * i + 2]; |
|
|
|
|
|
|
|
|
|
unsigned int Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
|
|
|
unsigned int Y = ((RY * r + GY * g + BY * b) >> RGB2YUV_SHIFT) + 16; |
|
|
|
|
|
|
|
|
|
ydst[2*i] = Y; |
|
|
|
|
ydst[2 * i] = Y; |
|
|
|
|
|
|
|
|
|
b = src[6*i+3]; |
|
|
|
|
g = src[6*i+4]; |
|
|
|
|
r = src[6*i+5]; |
|
|
|
|
b = src[6 * i + 3]; |
|
|
|
|
g = src[6 * i + 4]; |
|
|
|
|
r = src[6 * i + 5]; |
|
|
|
|
|
|
|
|
|
Y = ((RY*r + GY*g + BY*b)>>RGB2YUV_SHIFT) + 16; |
|
|
|
|
ydst[2*i+1] = Y; |
|
|
|
|
Y = ((RY * r + GY * g + BY * b) >> RGB2YUV_SHIFT) + 16; |
|
|
|
|
ydst[2 * i + 1] = Y; |
|
|
|
|
} |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
@ -666,17 +689,16 @@ void rgb24toyv12_c(const uint8_t *src, uint8_t *ydst, uint8_t *udst, |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void interleaveBytes_c(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
uint8_t *dest, int width, |
|
|
|
|
int height, int src1Stride, |
|
|
|
|
int src2Stride, int dstStride) |
|
|
|
|
uint8_t *dest, int width, int height, |
|
|
|
|
int src1Stride, int src2Stride, int dstStride) |
|
|
|
|
{ |
|
|
|
|
int h; |
|
|
|
|
|
|
|
|
|
for (h=0; h < height; h++) { |
|
|
|
|
for (h = 0; h < height; h++) { |
|
|
|
|
int w; |
|
|
|
|
for (w=0; w < width; w++) { |
|
|
|
|
dest[2*w+0] = src1[w]; |
|
|
|
|
dest[2*w+1] = src2[w]; |
|
|
|
|
for (w = 0; w < width; w++) { |
|
|
|
|
dest[2 * w + 0] = src1[w]; |
|
|
|
|
dest[2 * w + 1] = src2[w]; |
|
|
|
|
} |
|
|
|
|
dest += dstStride; |
|
|
|
|
src1 += src1Stride; |
|
|
|
@ -694,15 +716,15 @@ static inline void vu9_to_vu12_c(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
int w = width / 2; |
|
|
|
|
int h = height / 2; |
|
|
|
|
|
|
|
|
|
for (y=0;y<h;y++) { |
|
|
|
|
const uint8_t* s1=src1+srcStride1*(y>>1); |
|
|
|
|
uint8_t* d=dst1+dstStride1*y; |
|
|
|
|
for (y = 0; y < h; y++) { |
|
|
|
|
const uint8_t *s1 = src1 + srcStride1 * (y >> 1); |
|
|
|
|
uint8_t *d = dst1 + dstStride1 * y; |
|
|
|
|
for (x = 0; x < w; x++) |
|
|
|
|
d[2 * x] = d[2 * x + 1] = s1[x]; |
|
|
|
|
} |
|
|
|
|
for (y=0;y<h;y++) { |
|
|
|
|
const uint8_t* s2=src2+srcStride2*(y>>1); |
|
|
|
|
uint8_t* d=dst2+dstStride2*y; |
|
|
|
|
for (y = 0; y < h; y++) { |
|
|
|
|
const uint8_t *s2 = src2 + srcStride2 * (y >> 1); |
|
|
|
|
uint8_t *d = dst2 + dstStride2 * y; |
|
|
|
|
for (x = 0; x < w; x++) |
|
|
|
|
d[2 * x] = d[2 * x + 1] = s2[x]; |
|
|
|
|
} |
|
|
|
@ -718,21 +740,21 @@ static inline void yvu9_to_yuy2_c(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
int w = width / 2; |
|
|
|
|
int h = height; |
|
|
|
|
|
|
|
|
|
for (y=0;y<h;y++) { |
|
|
|
|
const uint8_t* yp=src1+srcStride1*y; |
|
|
|
|
const uint8_t* up=src2+srcStride2*(y>>2); |
|
|
|
|
const uint8_t* vp=src3+srcStride3*(y>>2); |
|
|
|
|
uint8_t* d=dst+dstStride*y; |
|
|
|
|
for (y = 0; y < h; y++) { |
|
|
|
|
const uint8_t *yp = src1 + srcStride1 * y; |
|
|
|
|
const uint8_t *up = src2 + srcStride2 * (y >> 2); |
|
|
|
|
const uint8_t *vp = src3 + srcStride3 * (y >> 2); |
|
|
|
|
uint8_t *d = dst + dstStride * y; |
|
|
|
|
for (x = 0; x < w; x++) { |
|
|
|
|
const int x2 = x<<2; |
|
|
|
|
d[8*x+0] = yp[x2]; |
|
|
|
|
d[8*x+1] = up[x]; |
|
|
|
|
d[8*x+2] = yp[x2+1]; |
|
|
|
|
d[8*x+3] = vp[x]; |
|
|
|
|
d[8*x+4] = yp[x2+2]; |
|
|
|
|
d[8*x+5] = up[x]; |
|
|
|
|
d[8*x+6] = yp[x2+3]; |
|
|
|
|
d[8*x+7] = vp[x]; |
|
|
|
|
const int x2 = x << 2; |
|
|
|
|
d[8 * x + 0] = yp[x2]; |
|
|
|
|
d[8 * x + 1] = up[x]; |
|
|
|
|
d[8 * x + 2] = yp[x2 + 1]; |
|
|
|
|
d[8 * x + 3] = vp[x]; |
|
|
|
|
d[8 * x + 4] = yp[x2 + 2]; |
|
|
|
|
d[8 * x + 5] = up[x]; |
|
|
|
|
d[8 * x + 6] = yp[x2 + 3]; |
|
|
|
|
d[8 * x + 7] = vp[x]; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -740,11 +762,10 @@ static inline void yvu9_to_yuy2_c(const uint8_t *src1, const uint8_t *src2, |
|
|
|
|
static void extract_even_c(const uint8_t *src, uint8_t *dst, int count) |
|
|
|
|
{ |
|
|
|
|
dst += count; |
|
|
|
|
src += 2*count; |
|
|
|
|
count= - count; |
|
|
|
|
|
|
|
|
|
while(count<0) { |
|
|
|
|
dst[count]= src[2*count]; |
|
|
|
|
src += count * 2; |
|
|
|
|
count = -count; |
|
|
|
|
while (count < 0) { |
|
|
|
|
dst[count] = src[2 * count]; |
|
|
|
|
count++; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -752,13 +773,13 @@ static void extract_even_c(const uint8_t *src, uint8_t *dst, int count) |
|
|
|
|
static void extract_even2_c(const uint8_t *src, uint8_t *dst0, uint8_t *dst1, |
|
|
|
|
int count) |
|
|
|
|
{ |
|
|
|
|
dst0+= count; |
|
|
|
|
dst1+= count; |
|
|
|
|
src += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= src[4*count+0]; |
|
|
|
|
dst1[count]= src[4*count+2]; |
|
|
|
|
dst0 += count; |
|
|
|
|
dst1 += count; |
|
|
|
|
src += count * 4; |
|
|
|
|
count = -count; |
|
|
|
|
while (count < 0) { |
|
|
|
|
dst0[count] = src[4 * count + 0]; |
|
|
|
|
dst1[count] = src[4 * count + 2]; |
|
|
|
|
count++; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -768,12 +789,12 @@ static void extract_even2avg_c(const uint8_t *src0, const uint8_t *src1, |
|
|
|
|
{ |
|
|
|
|
dst0 += count; |
|
|
|
|
dst1 += count; |
|
|
|
|
src0 += 4*count; |
|
|
|
|
src1 += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= (src0[4*count+0]+src1[4*count+0])>>1; |
|
|
|
|
dst1[count]= (src0[4*count+2]+src1[4*count+2])>>1; |
|
|
|
|
src0 += count * 4; |
|
|
|
|
src1 += count * 4; |
|
|
|
|
count = -count; |
|
|
|
|
while (count < 0) { |
|
|
|
|
dst0[count] = (src0[4 * count + 0] + src1[4 * count + 0]) >> 1; |
|
|
|
|
dst1[count] = (src0[4 * count + 2] + src1[4 * count + 2]) >> 1; |
|
|
|
|
count++; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -781,14 +802,14 @@ static void extract_even2avg_c(const uint8_t *src0, const uint8_t *src1, |
|
|
|
|
static void extract_odd2_c(const uint8_t *src, uint8_t *dst0, uint8_t *dst1, |
|
|
|
|
int count) |
|
|
|
|
{ |
|
|
|
|
dst0+= count; |
|
|
|
|
dst1+= count; |
|
|
|
|
src += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
dst0 += count; |
|
|
|
|
dst1 += count; |
|
|
|
|
src += count * 4; |
|
|
|
|
count = -count; |
|
|
|
|
src++; |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= src[4*count+0]; |
|
|
|
|
dst1[count]= src[4*count+2]; |
|
|
|
|
while (count < 0) { |
|
|
|
|
dst0[count] = src[4 * count + 0]; |
|
|
|
|
dst1[count] = src[4 * count + 2]; |
|
|
|
|
count++; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -798,14 +819,14 @@ static void extract_odd2avg_c(const uint8_t *src0, const uint8_t *src1, |
|
|
|
|
{ |
|
|
|
|
dst0 += count; |
|
|
|
|
dst1 += count; |
|
|
|
|
src0 += 4*count; |
|
|
|
|
src1 += 4*count; |
|
|
|
|
count= - count; |
|
|
|
|
src0 += count * 4; |
|
|
|
|
src1 += count * 4; |
|
|
|
|
count = -count; |
|
|
|
|
src0++; |
|
|
|
|
src1++; |
|
|
|
|
while(count<0) { |
|
|
|
|
dst0[count]= (src0[4*count+0]+src1[4*count+0])>>1; |
|
|
|
|
dst1[count]= (src0[4*count+2]+src1[4*count+2])>>1; |
|
|
|
|
while (count < 0) { |
|
|
|
|
dst0[count] = (src0[4 * count + 0] + src1[4 * count + 0]) >> 1; |
|
|
|
|
dst1[count] = (src0[4 * count + 2] + src1[4 * count + 2]) >> 1; |
|
|
|
|
count++; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -815,18 +836,18 @@ static void yuyvtoyuv420_c(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int lumStride, int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth= -((-width)>>1); |
|
|
|
|
const int chromWidth = -((-width) >> 1); |
|
|
|
|
|
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
extract_even_c(src, ydst, width); |
|
|
|
|
if(y&1) { |
|
|
|
|
if (y & 1) { |
|
|
|
|
extract_odd2avg_c(src - srcStride, src, udst, vdst, chromWidth); |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -835,16 +856,16 @@ static void yuyvtoyuv422_c(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int lumStride, int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth= -((-width)>>1); |
|
|
|
|
const int chromWidth = -((-width) >> 1); |
|
|
|
|
|
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
extract_even_c(src, ydst, width); |
|
|
|
|
extract_odd2_c(src, udst, vdst, chromWidth); |
|
|
|
|
|
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -853,18 +874,18 @@ static void uyvytoyuv420_c(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int lumStride, int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth= -((-width)>>1); |
|
|
|
|
const int chromWidth = -((-width) >> 1); |
|
|
|
|
|
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
extract_even_c(src + 1, ydst, width); |
|
|
|
|
if(y&1) { |
|
|
|
|
if (y & 1) { |
|
|
|
|
extract_even2avg_c(src - srcStride, src, udst, vdst, chromWidth); |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
@ -873,16 +894,16 @@ static void uyvytoyuv422_c(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, |
|
|
|
|
int lumStride, int chromStride, int srcStride) |
|
|
|
|
{ |
|
|
|
|
int y; |
|
|
|
|
const int chromWidth= -((-width)>>1); |
|
|
|
|
const int chromWidth = -((-width) >> 1); |
|
|
|
|
|
|
|
|
|
for (y=0; y<height; y++) { |
|
|
|
|
for (y = 0; y < height; y++) { |
|
|
|
|
extract_even_c(src + 1, ydst, width); |
|
|
|
|
extract_even2_c(src, udst, vdst, chromWidth); |
|
|
|
|
|
|
|
|
|
src += srcStride; |
|
|
|
|
ydst+= lumStride; |
|
|
|
|
udst+= chromStride; |
|
|
|
|
vdst+= chromStride; |
|
|
|
|
ydst += lumStride; |
|
|
|
|
udst += chromStride; |
|
|
|
|
vdst += chromStride; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|