|
|
|
/*
|
|
|
|
* Copyright (c) 2020 Martin Storsjo
|
|
|
|
* Copyright (c) 2024 Ramiro Polla
|
|
|
|
*
|
|
|
|
* This file is part of FFmpeg.
|
|
|
|
*
|
|
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
|
|
* License as published by the Free Software Foundation; either
|
|
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
|
|
*
|
|
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
* Lesser General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "libavutil/aarch64/asm.S"
|
|
|
|
|
|
|
|
#define RGB2YUV_COEFFS 16*4+16*32
|
|
|
|
#define BY v0.h[0]
|
|
|
|
#define GY v0.h[1]
|
|
|
|
#define RY v0.h[2]
|
|
|
|
#define BU v1.h[0]
|
|
|
|
#define GU v1.h[1]
|
|
|
|
#define RU v1.h[2]
|
|
|
|
#define BV v2.h[0]
|
|
|
|
#define GV v2.h[1]
|
|
|
|
#define RV v2.h[2]
|
|
|
|
#define Y_OFFSET v22
|
|
|
|
#define UV_OFFSET v23
|
|
|
|
|
|
|
|
// convert rgb to 16-bit y, u, or v
|
|
|
|
// uses v3 and v4
|
|
|
|
.macro rgbconv16 dst, b, g, r, bc, gc, rc
|
|
|
|
smull v3.4s, \b\().4h, \bc
|
|
|
|
smlal v3.4s, \g\().4h, \gc
|
|
|
|
smlal v3.4s, \r\().4h, \rc
|
|
|
|
smull2 v4.4s, \b\().8h, \bc
|
|
|
|
smlal2 v4.4s, \g\().8h, \gc
|
|
|
|
smlal2 v4.4s, \r\().8h, \rc // v3:v4 = b * bc + g * gc + r * rc (32-bit)
|
|
|
|
shrn \dst\().4h, v3.4s, #7
|
|
|
|
shrn2 \dst\().8h, v4.4s, #7 // dst = b * bc + g * gc + r * rc (16-bit)
|
|
|
|
.endm
|
|
|
|
|
|
|
|
// void ff_rgb24toyv12_neon(const uint8_t *src, uint8_t *ydst, uint8_t *udst,
|
|
|
|
// uint8_t *vdst, int width, int height, int lumStride,
|
|
|
|
// int chromStride, int srcStride, int32_t *rgb2yuv);
|
|
|
|
function ff_rgb24toyv12_neon, export=1
|
|
|
|
// x0 const uint8_t *src
|
|
|
|
// x1 uint8_t *ydst
|
|
|
|
// x2 uint8_t *udst
|
|
|
|
// x3 uint8_t *vdst
|
|
|
|
// w4 int width
|
|
|
|
// w5 int height
|
|
|
|
// w6 int lumStride
|
|
|
|
// w7 int chromStride
|
|
|
|
ldrsw x14, [sp]
|
|
|
|
ldr x15, [sp, #8]
|
|
|
|
// x14 int srcStride
|
|
|
|
// x15 int32_t *rgb2yuv
|
|
|
|
|
|
|
|
// extend width and stride parameters
|
|
|
|
uxtw x4, w4
|
|
|
|
sxtw x6, w6
|
|
|
|
sxtw x7, w7
|
|
|
|
|
|
|
|
// src1 = x0
|
|
|
|
// src2 = x10
|
|
|
|
add x10, x0, x14 // x10 = src + srcStride
|
|
|
|
lsl x14, x14, #1 // srcStride *= 2
|
|
|
|
add x11, x4, x4, lsl #1 // x11 = 3 * width
|
|
|
|
sub x14, x14, x11 // srcPadding = (2 * srcStride) - (3 * width)
|
|
|
|
|
|
|
|
// ydst1 = x1
|
|
|
|
// ydst2 = x11
|
|
|
|
add x11, x1, x6 // x11 = ydst + lumStride
|
|
|
|
lsl x6, x6, #1 // lumStride *= 2
|
|
|
|
sub x6, x6, x4 // lumPadding = (2 * lumStride) - width
|
|
|
|
|
|
|
|
sub x7, x7, x4, lsr #1 // chromPadding = chromStride - (width / 2)
|
|
|
|
|
|
|
|
// load rgb2yuv coefficients into v0, v1, and v2
|
|
|
|
add x15, x15, #RGB2YUV_COEFFS
|
|
|
|
ld1 {v0.8h-v2.8h}, [x15] // load 24 values
|
|
|
|
|
|
|
|
// load offset constants
|
|
|
|
movi Y_OFFSET.8h, #0x10, lsl #8
|
|
|
|
movi UV_OFFSET.8h, #0x80, lsl #8
|
|
|
|
|
|
|
|
1:
|
|
|
|
mov w15, w4 // w15 = width
|
|
|
|
|
|
|
|
2:
|
|
|
|
// load first line
|
|
|
|
ld3 {v26.16b, v27.16b, v28.16b}, [x0], #48
|
|
|
|
|
|
|
|
// widen first line to 16-bit
|
|
|
|
uxtl v16.8h, v26.8b // v16 = B11
|
|
|
|
uxtl v17.8h, v27.8b // v17 = G11
|
|
|
|
uxtl v18.8h, v28.8b // v18 = R11
|
|
|
|
uxtl2 v19.8h, v26.16b // v19 = B12
|
|
|
|
uxtl2 v20.8h, v27.16b // v20 = G12
|
|
|
|
uxtl2 v21.8h, v28.16b // v21 = R12
|
|
|
|
|
|
|
|
// calculate Y values for first line
|
|
|
|
rgbconv16 v24, v16, v17, v18, BY, GY, RY // v24 = Y11
|
|
|
|
rgbconv16 v25, v19, v20, v21, BY, GY, RY // v25 = Y12
|
|
|
|
|
|
|
|
// load second line
|
|
|
|
ld3 {v26.16b, v27.16b, v28.16b}, [x10], #48
|
|
|
|
|
|
|
|
// pairwise add and save rgb values to calculate average
|
|
|
|
addp v5.8h, v16.8h, v19.8h
|
|
|
|
addp v6.8h, v17.8h, v20.8h
|
|
|
|
addp v7.8h, v18.8h, v21.8h
|
|
|
|
|
|
|
|
// widen second line to 16-bit
|
|
|
|
uxtl v16.8h, v26.8b // v16 = B21
|
|
|
|
uxtl v17.8h, v27.8b // v17 = G21
|
|
|
|
uxtl v18.8h, v28.8b // v18 = R21
|
|
|
|
uxtl2 v19.8h, v26.16b // v19 = B22
|
|
|
|
uxtl2 v20.8h, v27.16b // v20 = G22
|
|
|
|
uxtl2 v21.8h, v28.16b // v21 = R22
|
|
|
|
|
|
|
|
// calculate Y values for second line
|
|
|
|
rgbconv16 v26, v16, v17, v18, BY, GY, RY // v26 = Y21
|
|
|
|
rgbconv16 v27, v19, v20, v21, BY, GY, RY // v27 = Y22
|
|
|
|
|
|
|
|
// pairwise add rgb values to calculate average
|
|
|
|
addp v16.8h, v16.8h, v19.8h
|
|
|
|
addp v17.8h, v17.8h, v20.8h
|
|
|
|
addp v18.8h, v18.8h, v21.8h
|
|
|
|
|
|
|
|
// calculate average
|
|
|
|
add v16.8h, v16.8h, v5.8h
|
|
|
|
add v17.8h, v17.8h, v6.8h
|
|
|
|
add v18.8h, v18.8h, v7.8h
|
|
|
|
ushr v16.8h, v16.8h, #2
|
|
|
|
ushr v17.8h, v17.8h, #2
|
|
|
|
ushr v18.8h, v18.8h, #2
|
|
|
|
|
|
|
|
// calculate U and V values
|
|
|
|
rgbconv16 v28, v16, v17, v18, BU, GU, RU // v28 = U
|
|
|
|
rgbconv16 v29, v16, v17, v18, BV, GV, RV // v29 = V
|
|
|
|
|
|
|
|
// add offsets and narrow all values
|
|
|
|
addhn v24.8b, v24.8h, Y_OFFSET.8h
|
|
|
|
addhn v25.8b, v25.8h, Y_OFFSET.8h
|
|
|
|
addhn v26.8b, v26.8h, Y_OFFSET.8h
|
|
|
|
addhn v27.8b, v27.8h, Y_OFFSET.8h
|
|
|
|
addhn v28.8b, v28.8h, UV_OFFSET.8h
|
|
|
|
addhn v29.8b, v29.8h, UV_OFFSET.8h
|
|
|
|
|
|
|
|
subs w15, w15, #16
|
|
|
|
|
|
|
|
// store output
|
|
|
|
st1 {v24.8b, v25.8b}, [x1], #16 // store ydst1
|
|
|
|
st1 {v26.8b, v27.8b}, [x11], #16 // store ydst2
|
|
|
|
st1 {v28.8b}, [x2], #8 // store udst
|
|
|
|
st1 {v29.8b}, [x3], #8 // store vdst
|
|
|
|
|
|
|
|
b.gt 2b
|
|
|
|
|
|
|
|
subs w5, w5, #2
|
|
|
|
|
|
|
|
// row += 2
|
|
|
|
add x0, x0, x14 // src1 += srcPadding
|
|
|
|
add x10, x10, x14 // src2 += srcPadding
|
|
|
|
add x1, x1, x6 // ydst1 += lumPadding
|
|
|
|
add x11, x11, x6 // ydst2 += lumPadding
|
|
|
|
add x2, x2, x7 // udst += chromPadding
|
|
|
|
add x3, x3, x7 // vdst += chromPadding
|
|
|
|
b.gt 1b
|
|
|
|
|
|
|
|
ret
|
|
|
|
endfunc
|
|
|
|
|
|
|
|
// void ff_interleave_bytes_neon(const uint8_t *src1, const uint8_t *src2,
|
|
|
|
// uint8_t *dest, int width, int height,
|
|
|
|
// int src1Stride, int src2Stride, int dstStride);
|
|
|
|
function ff_interleave_bytes_neon, export=1
|
|
|
|
sub w5, w5, w3
|
|
|
|
sub w6, w6, w3
|
|
|
|
sub w7, w7, w3, lsl #1
|
|
|
|
1:
|
|
|
|
ands w8, w3, #0xfffffff0 // & ~15
|
|
|
|
b.eq 3f
|
|
|
|
2:
|
|
|
|
ld1 {v0.16b}, [x0], #16
|
|
|
|
ld1 {v1.16b}, [x1], #16
|
|
|
|
subs w8, w8, #16
|
|
|
|
st2 {v0.16b, v1.16b}, [x2], #32
|
|
|
|
b.gt 2b
|
|
|
|
|
|
|
|
tst w3, #15
|
|
|
|
b.eq 9f
|
|
|
|
|
|
|
|
3:
|
|
|
|
tst w3, #8
|
|
|
|
b.eq 4f
|
|
|
|
ld1 {v0.8b}, [x0], #8
|
|
|
|
ld1 {v1.8b}, [x1], #8
|
|
|
|
st2 {v0.8b, v1.8b}, [x2], #16
|
|
|
|
4:
|
|
|
|
tst w3, #4
|
|
|
|
b.eq 5f
|
|
|
|
|
|
|
|
ld1 {v0.s}[0], [x0], #4
|
|
|
|
ld1 {v1.s}[0], [x1], #4
|
|
|
|
zip1 v0.8b, v0.8b, v1.8b
|
|
|
|
st1 {v0.8b}, [x2], #8
|
|
|
|
|
|
|
|
5:
|
|
|
|
ands w8, w3, #3
|
|
|
|
b.eq 9f
|
|
|
|
6:
|
|
|
|
ldrb w9, [x0], #1
|
|
|
|
ldrb w10, [x1], #1
|
|
|
|
subs w8, w8, #1
|
|
|
|
bfi w9, w10, #8, #8
|
|
|
|
strh w9, [x2], #2
|
|
|
|
b.gt 6b
|
|
|
|
|
|
|
|
9:
|
|
|
|
subs w4, w4, #1
|
|
|
|
b.eq 0f
|
|
|
|
add x0, x0, w5, sxtw
|
|
|
|
add x1, x1, w6, sxtw
|
|
|
|
add x2, x2, w7, sxtw
|
|
|
|
b 1b
|
|
|
|
|
|
|
|
0:
|
|
|
|
ret
|
|
|
|
endfunc
|
|
|
|
|
|
|
|
// void ff_deinterleave_bytes_neon(const uint8_t *src, uint8_t *dst1, uint8_t *dst2,
|
|
|
|
// int width, int height, int srcStride,
|
|
|
|
// int dst1Stride, int dst2Stride);
|
|
|
|
function ff_deinterleave_bytes_neon, export=1
|
|
|
|
sub w5, w5, w3, lsl #1
|
|
|
|
sub w6, w6, w3
|
|
|
|
sub w7, w7, w3
|
|
|
|
1:
|
|
|
|
ands w8, w3, #0xfffffff0 // & ~15
|
|
|
|
b.eq 3f
|
|
|
|
2:
|
|
|
|
ld2 {v0.16b, v1.16b}, [x0], #32
|
|
|
|
subs w8, w8, #16
|
|
|
|
st1 {v0.16b}, [x1], #16
|
|
|
|
st1 {v1.16b}, [x2], #16
|
|
|
|
b.gt 2b
|
|
|
|
|
|
|
|
tst w3, #15
|
|
|
|
b.eq 9f
|
|
|
|
|
|
|
|
3:
|
|
|
|
tst w3, #8
|
|
|
|
b.eq 4f
|
|
|
|
ld2 {v0.8b, v1.8b}, [x0], #16
|
|
|
|
st1 {v0.8b}, [x1], #8
|
|
|
|
st1 {v1.8b}, [x2], #8
|
|
|
|
4:
|
|
|
|
tst w3, #4
|
|
|
|
b.eq 5f
|
|
|
|
|
|
|
|
ld1 {v0.8b}, [x0], #8
|
|
|
|
shrn v1.8b, v0.8h, #8
|
|
|
|
xtn v0.8b, v0.8h
|
|
|
|
st1 {v0.s}[0], [x1], #4
|
|
|
|
st1 {v1.s}[0], [x2], #4
|
|
|
|
|
|
|
|
5:
|
|
|
|
ands w8, w3, #3
|
|
|
|
b.eq 9f
|
|
|
|
6:
|
|
|
|
ldrh w9, [x0], #2
|
|
|
|
subs w8, w8, #1
|
|
|
|
ubfx w10, w9, #8, #8
|
|
|
|
strb w9, [x1], #1
|
|
|
|
strb w10, [x2], #1
|
|
|
|
b.gt 6b
|
|
|
|
|
|
|
|
9:
|
|
|
|
subs w4, w4, #1
|
|
|
|
b.eq 0f
|
|
|
|
add x0, x0, w5, sxtw
|
|
|
|
add x1, x1, w6, sxtw
|
|
|
|
add x2, x2, w7, sxtw
|
|
|
|
b 1b
|
|
|
|
|
|
|
|
0:
|
|
|
|
ret
|
|
|
|
endfunc
|