mirror of https://github.com/FFmpeg/FFmpeg.git
This work is sponsored by, and copyright, Google.
These are ported from the ARM version; it is essentially a 1:1
port with no extra added features, but with some hand tuning
(especially for the plain copy/avg functions). The ARM version
isn't very register starved to begin with, so there's not much
to be gained from having more spare registers here - we only
avoid having to clobber callee-saved registers.
Examples of runtimes vs the 32 bit version, on a Cortex A53:
ARM AArch64
vp9_avg4_neon: 27.2 23.7
vp9_avg8_neon: 56.5 54.7
vp9_avg16_neon: 169.9 167.4
vp9_avg32_neon: 585.8 585.2
vp9_avg64_neon: 2460.3 2294.7
vp9_avg_8tap_smooth_4h_neon: 132.7 125.2
vp9_avg_8tap_smooth_4hv_neon: 478.8 442.0
vp9_avg_8tap_smooth_4v_neon: 126.0 93.7
vp9_avg_8tap_smooth_8h_neon: 241.7 234.2
vp9_avg_8tap_smooth_8hv_neon: 690.9 646.5
vp9_avg_8tap_smooth_8v_neon: 245.0 205.5
vp9_avg_8tap_smooth_64h_neon: 11273.2 11280.1
vp9_avg_8tap_smooth_64hv_neon: 22980.6 22184.1
vp9_avg_8tap_smooth_64v_neon: 11549.7 10781.1
vp9_put4_neon: 18.0 17.2
vp9_put8_neon: 40.2 37.7
vp9_put16_neon: 97.4 99.5
vp9_put32_neon/armv8: 346.0 307.4
vp9_put64_neon/armv8: 1319.0 1107.5
vp9_put_8tap_smooth_4h_neon: 126.7 118.2
vp9_put_8tap_smooth_4hv_neon: 465.7 434.0
vp9_put_8tap_smooth_4v_neon: 113.0 86.5
vp9_put_8tap_smooth_8h_neon: 229.7 221.6
vp9_put_8tap_smooth_8hv_neon: 658.9 621.3
vp9_put_8tap_smooth_8v_neon: 215.0 187.5
vp9_put_8tap_smooth_64h_neon: 10636.7 10627.8
vp9_put_8tap_smooth_64hv_neon: 21076.8 21026.9
vp9_put_8tap_smooth_64v_neon: 9635.0 9632.4
These are generally about as fast as the corresponding ARM
routines on the same CPU (at least on the A53), in most cases
marginally faster.
The speedup vs C code is pretty much the same as for the 32 bit
case; on the A53 it's around 6-13x for ther larger 8tap filters.
The exact speedup varies a little, since the C versions generally
don't end up exactly as slow/fast as on 32 bit.
This is an adapted cherry-pick from libav commit
383d96aa22
.
Signed-off-by: Ronald S. Bultje <rsbultje@gmail.com>
pull/241/head
parent
7fe898dbb9
commit
1f7801c2bc
6 changed files with 840 additions and 4 deletions
@ -0,0 +1,156 @@ |
||||
/*
|
||||
* Copyright (c) 2016 Google Inc. |
||||
* |
||||
* This file is part of FFmpeg. |
||||
* |
||||
* FFmpeg is free software; you can redistribute it and/or |
||||
* modify it under the terms of the GNU Lesser General Public |
||||
* License as published by the Free Software Foundation; either |
||||
* version 2.1 of the License, or (at your option) any later version. |
||||
* |
||||
* FFmpeg is distributed in the hope that it will be useful, |
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
||||
* Lesser General Public License for more details. |
||||
* |
||||
* You should have received a copy of the GNU Lesser General Public |
||||
* License along with FFmpeg; if not, write to the Free Software |
||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
||||
*/ |
||||
|
||||
#include <stdint.h> |
||||
|
||||
#include "libavutil/attributes.h" |
||||
#include "libavutil/aarch64/cpu.h" |
||||
#include "libavcodec/vp9dsp.h" |
||||
|
||||
#define declare_fpel(type, sz) \ |
||||
void ff_vp9_##type##sz##_neon(uint8_t *dst, ptrdiff_t dst_stride, \
|
||||
const uint8_t *src, ptrdiff_t src_stride, \
|
||||
int h, int mx, int my) |
||||
|
||||
#define declare_copy_avg(sz) \ |
||||
declare_fpel(copy, sz); \
|
||||
declare_fpel(avg , sz) |
||||
|
||||
#define decl_mc_func(op, filter, dir, sz) \ |
||||
void ff_vp9_##op##_##filter##sz##_##dir##_neon(uint8_t *dst, ptrdiff_t dst_stride, \
|
||||
const uint8_t *src, ptrdiff_t src_stride, \
|
||||
int h, int mx, int my) |
||||
|
||||
#define define_8tap_2d_fn(op, filter, sz) \ |
||||
static void op##_##filter##sz##_hv_neon(uint8_t *dst, ptrdiff_t dst_stride, \
|
||||
const uint8_t *src, ptrdiff_t src_stride, \
|
||||
int h, int mx, int my) \
|
||||
{ \
|
||||
LOCAL_ALIGNED_16(uint8_t, temp, [((1 + (sz < 64)) * sz + 8) * sz]); \
|
||||
/* We only need h + 7 lines, but the horizontal filter assumes an \
|
||||
* even number of rows, so filter h + 8 lines here. */ \
|
||||
ff_vp9_put_##filter##sz##_h_neon(temp, sz, \
|
||||
src - 3 * src_stride, src_stride, \
|
||||
h + 8, mx, 0); \
|
||||
ff_vp9_##op##_##filter##sz##_v_neon(dst, dst_stride, \
|
||||
temp + 3 * sz, sz, \
|
||||
h, 0, my); \
|
||||
} |
||||
|
||||
#define decl_filter_funcs(op, dir, sz) \ |
||||
decl_mc_func(op, regular, dir, sz); \
|
||||
decl_mc_func(op, sharp, dir, sz); \
|
||||
decl_mc_func(op, smooth, dir, sz) |
||||
|
||||
#define decl_mc_funcs(sz) \ |
||||
decl_filter_funcs(put, h, sz); \
|
||||
decl_filter_funcs(avg, h, sz); \
|
||||
decl_filter_funcs(put, v, sz); \
|
||||
decl_filter_funcs(avg, v, sz); \
|
||||
decl_filter_funcs(put, hv, sz); \
|
||||
decl_filter_funcs(avg, hv, sz) |
||||
|
||||
#define ff_vp9_copy32_neon ff_vp9_copy32_aarch64 |
||||
#define ff_vp9_copy64_neon ff_vp9_copy64_aarch64 |
||||
|
||||
declare_copy_avg(64); |
||||
declare_copy_avg(32); |
||||
declare_copy_avg(16); |
||||
declare_copy_avg(8); |
||||
declare_copy_avg(4); |
||||
|
||||
decl_mc_funcs(64); |
||||
decl_mc_funcs(32); |
||||
decl_mc_funcs(16); |
||||
decl_mc_funcs(8); |
||||
decl_mc_funcs(4); |
||||
|
||||
#define define_8tap_2d_funcs(sz) \ |
||||
define_8tap_2d_fn(put, regular, sz) \
|
||||
define_8tap_2d_fn(put, sharp, sz) \
|
||||
define_8tap_2d_fn(put, smooth, sz) \
|
||||
define_8tap_2d_fn(avg, regular, sz) \
|
||||
define_8tap_2d_fn(avg, sharp, sz) \
|
||||
define_8tap_2d_fn(avg, smooth, sz) |
||||
|
||||
define_8tap_2d_funcs(64) |
||||
define_8tap_2d_funcs(32) |
||||
define_8tap_2d_funcs(16) |
||||
define_8tap_2d_funcs(8) |
||||
define_8tap_2d_funcs(4) |
||||
|
||||
av_cold void ff_vp9dsp_init_aarch64(VP9DSPContext *dsp, int bpp) |
||||
{ |
||||
int cpu_flags = av_get_cpu_flags(); |
||||
|
||||
if (bpp != 8) |
||||
return; |
||||
|
||||
#define init_fpel(idx1, idx2, sz, type, suffix) \ |
||||
dsp->mc[idx1][FILTER_8TAP_SMOOTH ][idx2][0][0] = \
|
||||
dsp->mc[idx1][FILTER_8TAP_REGULAR][idx2][0][0] = \
|
||||
dsp->mc[idx1][FILTER_8TAP_SHARP ][idx2][0][0] = \
|
||||
dsp->mc[idx1][FILTER_BILINEAR ][idx2][0][0] = ff_vp9_##type##sz##suffix |
||||
|
||||
#define init_copy(idx, sz, suffix) \ |
||||
init_fpel(idx, 0, sz, copy, suffix) |
||||
|
||||
#define init_avg(idx, sz, suffix) \ |
||||
init_fpel(idx, 1, sz, avg, suffix) |
||||
|
||||
#define init_copy_avg(idx, sz) \ |
||||
init_copy(idx, sz, _neon); \
|
||||
init_avg (idx, sz, _neon) |
||||
|
||||
if (have_armv8(cpu_flags)) { |
||||
init_copy(0, 64, _aarch64); |
||||
init_copy(1, 32, _aarch64); |
||||
} |
||||
|
||||
if (have_neon(cpu_flags)) { |
||||
#define init_mc_func(idx1, idx2, op, filter, fname, dir, mx, my, sz, pfx) \ |
||||
dsp->mc[idx1][filter][idx2][mx][my] = pfx##op##_##fname##sz##_##dir##_neon |
||||
|
||||
#define init_mc_funcs(idx, dir, mx, my, sz, pfx) \ |
||||
init_mc_func(idx, 0, put, FILTER_8TAP_REGULAR, regular, dir, mx, my, sz, pfx); \
|
||||
init_mc_func(idx, 0, put, FILTER_8TAP_SHARP, sharp, dir, mx, my, sz, pfx); \
|
||||
init_mc_func(idx, 0, put, FILTER_8TAP_SMOOTH, smooth, dir, mx, my, sz, pfx); \
|
||||
init_mc_func(idx, 1, avg, FILTER_8TAP_REGULAR, regular, dir, mx, my, sz, pfx); \
|
||||
init_mc_func(idx, 1, avg, FILTER_8TAP_SHARP, sharp, dir, mx, my, sz, pfx); \
|
||||
init_mc_func(idx, 1, avg, FILTER_8TAP_SMOOTH, smooth, dir, mx, my, sz, pfx) |
||||
|
||||
#define init_mc_funcs_dirs(idx, sz) \ |
||||
init_mc_funcs(idx, h, 1, 0, sz, ff_vp9_); \
|
||||
init_mc_funcs(idx, v, 0, 1, sz, ff_vp9_); \
|
||||
init_mc_funcs(idx, hv, 1, 1, sz,) |
||||
|
||||
init_avg(0, 64, _neon); |
||||
init_avg(1, 32, _neon); |
||||
init_copy_avg(2, 16); |
||||
init_copy_avg(3, 8); |
||||
init_copy_avg(4, 4); |
||||
|
||||
init_mc_funcs_dirs(0, 64); |
||||
init_mc_funcs_dirs(1, 32); |
||||
init_mc_funcs_dirs(2, 16); |
||||
init_mc_funcs_dirs(3, 8); |
||||
init_mc_funcs_dirs(4, 4); |
||||
} |
||||
} |
@ -0,0 +1,676 @@ |
||||
/* |
||||
* Copyright (c) 2016 Google Inc. |
||||
* |
||||
* This file is part of FFmpeg. |
||||
* |
||||
* FFmpeg is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU Lesser General Public |
||||
* License as published by the Free Software Foundation; either
|
||||
* version 2.1 of the License, or (at your option) any later version. |
||||
* |
||||
* FFmpeg is distributed in the hope that it will be useful, |
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
||||
* Lesser General Public License for more details. |
||||
* |
||||
* You should have received a copy of the GNU Lesser General Public |
||||
* License along with FFmpeg; if not, write to the Free Software
|
||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
||||
*/ |
||||
|
||||
#include "libavutil/aarch64/asm.S" |
||||
|
||||
// All public functions in this file have the following signature: |
||||
// typedef void (*vp9_mc_func)(uint8_t *dst, ptrdiff_t dst_stride, |
||||
// const uint8_t *ref, ptrdiff_t ref_stride, |
||||
// int h, int mx, int my);
|
||||
|
||||
function ff_vp9_copy64_aarch64, export=1 |
||||
1: |
||||
ldp x5, x6, [x2] |
||||
ldp x7, x8, [x2, #16] |
||||
stp x5, x6, [x0] |
||||
ldp x9, x10, [x2, #32] |
||||
stp x7, x8, [x0, #16] |
||||
subs w4, w4, #1 |
||||
ldp x11, x12, [x2, #48] |
||||
stp x9, x10, [x0, #32] |
||||
stp x11, x12, [x0, #48] |
||||
add x2, x2, x3 |
||||
add x0, x0, x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_avg64_neon, export=1 |
||||
mov x5, x0 |
||||
1: |
||||
ld1 {v4.16b, v5.16b, v6.16b, v7.16b}, [x2], x3 |
||||
ld1 {v0.16b, v1.16b, v2.16b, v3.16b}, [x0], x1 |
||||
ld1 {v20.16b, v21.16b, v22.16b, v23.16b}, [x2], x3 |
||||
urhadd v0.16b, v0.16b, v4.16b |
||||
urhadd v1.16b, v1.16b, v5.16b |
||||
ld1 {v16.16b, v17.16b, v18.16b, v19.16b}, [x0], x1 |
||||
urhadd v2.16b, v2.16b, v6.16b |
||||
urhadd v3.16b, v3.16b, v7.16b |
||||
subs w4, w4, #2 |
||||
urhadd v16.16b, v16.16b, v20.16b |
||||
urhadd v17.16b, v17.16b, v21.16b |
||||
st1 {v0.16b, v1.16b, v2.16b, v3.16b}, [x5], x1 |
||||
urhadd v18.16b, v18.16b, v22.16b |
||||
urhadd v19.16b, v19.16b, v23.16b |
||||
st1 {v16.16b, v17.16b, v18.16b, v19.16b}, [x5], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_copy32_aarch64, export=1 |
||||
1: |
||||
ldp x5, x6, [x2] |
||||
ldp x7, x8, [x2, #16] |
||||
stp x5, x6, [x0] |
||||
subs w4, w4, #1 |
||||
stp x7, x8, [x0, #16] |
||||
add x2, x2, x3 |
||||
add x0, x0, x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_avg32_neon, export=1 |
||||
1: |
||||
ld1 {v2.16b, v3.16b}, [x2], x3 |
||||
ld1 {v0.16b, v1.16b}, [x0] |
||||
urhadd v0.16b, v0.16b, v2.16b |
||||
urhadd v1.16b, v1.16b, v3.16b |
||||
subs w4, w4, #1 |
||||
st1 {v0.16b, v1.16b}, [x0], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_copy16_neon, export=1 |
||||
add x5, x0, x1 |
||||
lsl x1, x1, #1 |
||||
add x6, x2, x3 |
||||
lsl x3, x3, #1 |
||||
1: |
||||
ld1 {v0.16b}, [x2], x3 |
||||
ld1 {v1.16b}, [x6], x3 |
||||
ld1 {v2.16b}, [x2], x3 |
||||
ld1 {v3.16b}, [x6], x3 |
||||
subs w4, w4, #4 |
||||
st1 {v0.16b}, [x0], x1 |
||||
st1 {v1.16b}, [x5], x1 |
||||
st1 {v2.16b}, [x0], x1 |
||||
st1 {v3.16b}, [x5], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_avg16_neon, export=1 |
||||
mov x5, x0 |
||||
1: |
||||
ld1 {v2.16b}, [x2], x3 |
||||
ld1 {v0.16b}, [x0], x1 |
||||
ld1 {v3.16b}, [x2], x3 |
||||
urhadd v0.16b, v0.16b, v2.16b |
||||
ld1 {v1.16b}, [x0], x1 |
||||
urhadd v1.16b, v1.16b, v3.16b |
||||
subs w4, w4, #2 |
||||
st1 {v0.16b}, [x5], x1 |
||||
st1 {v1.16b}, [x5], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_copy8_neon, export=1 |
||||
1: |
||||
ld1 {v0.8b}, [x2], x3 |
||||
ld1 {v1.8b}, [x2], x3 |
||||
subs w4, w4, #2 |
||||
st1 {v0.8b}, [x0], x1 |
||||
st1 {v1.8b}, [x0], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_avg8_neon, export=1 |
||||
mov x5, x0 |
||||
1: |
||||
ld1 {v2.8b}, [x2], x3 |
||||
ld1 {v0.8b}, [x0], x1 |
||||
ld1 {v3.8b}, [x2], x3 |
||||
urhadd v0.8b, v0.8b, v2.8b |
||||
ld1 {v1.8b}, [x0], x1 |
||||
urhadd v1.8b, v1.8b, v3.8b |
||||
subs w4, w4, #2 |
||||
st1 {v0.8b}, [x5], x1 |
||||
st1 {v1.8b}, [x5], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_copy4_neon, export=1 |
||||
1: |
||||
ld1 {v0.s}[0], [x2], x3 |
||||
ld1 {v1.s}[0], [x2], x3 |
||||
st1 {v0.s}[0], [x0], x1 |
||||
ld1 {v2.s}[0], [x2], x3 |
||||
st1 {v1.s}[0], [x0], x1 |
||||
ld1 {v3.s}[0], [x2], x3 |
||||
subs w4, w4, #4 |
||||
st1 {v2.s}[0], [x0], x1 |
||||
st1 {v3.s}[0], [x0], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
function ff_vp9_avg4_neon, export=1 |
||||
mov x5, x0 |
||||
1: |
||||
ld1 {v2.s}[0], [x2], x3 |
||||
ld1 {v0.s}[0], [x0], x1 |
||||
ld1 {v2.s}[1], [x2], x3 |
||||
ld1 {v0.s}[1], [x0], x1 |
||||
ld1 {v3.s}[0], [x2], x3 |
||||
ld1 {v1.s}[0], [x0], x1 |
||||
ld1 {v3.s}[1], [x2], x3 |
||||
ld1 {v1.s}[1], [x0], x1 |
||||
subs w4, w4, #4 |
||||
urhadd v0.8b, v0.8b, v2.8b |
||||
urhadd v1.8b, v1.8b, v3.8b |
||||
st1 {v0.s}[0], [x5], x1 |
||||
st1 {v0.s}[1], [x5], x1 |
||||
st1 {v1.s}[0], [x5], x1 |
||||
st1 {v1.s}[1], [x5], x1 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
|
||||
|
||||
// Extract a vector from src1-src2 and src4-src5 (src1-src3 and src4-src6 |
||||
// for size >= 16), and multiply-accumulate into dst1 and dst3 (or |
||||
// dst1-dst2 and dst3-dst4 for size >= 16) |
||||
.macro extmla dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size |
||||
ext v20.16b, \src1, \src2, #(2*\offset) |
||||
ext v22.16b, \src4, \src5, #(2*\offset) |
||||
.if \size >= 16 |
||||
mla \dst1, v20.8h, v0.h[\offset] |
||||
ext v21.16b, \src2, \src3, #(2*\offset) |
||||
mla \dst3, v22.8h, v0.h[\offset] |
||||
ext v23.16b, \src5, \src6, #(2*\offset) |
||||
mla \dst2, v21.8h, v0.h[\offset] |
||||
mla \dst4, v23.8h, v0.h[\offset] |
||||
.else |
||||
mla \dst1, v20.8h, v0.h[\offset] |
||||
mla \dst3, v22.8h, v0.h[\offset] |
||||
.endif |
||||
.endm |
||||
// The same as above, but don't accumulate straight into the |
||||
// destination, but use a temp register and accumulate with saturation. |
||||
.macro extmulqadd dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size |
||||
ext v20.16b, \src1, \src2, #(2*\offset) |
||||
ext v22.16b, \src4, \src5, #(2*\offset) |
||||
.if \size >= 16 |
||||
mul v20.8h, v20.8h, v0.h[\offset] |
||||
ext v21.16b, \src2, \src3, #(2*\offset) |
||||
mul v22.8h, v22.8h, v0.h[\offset] |
||||
ext v23.16b, \src5, \src6, #(2*\offset) |
||||
mul v21.8h, v21.8h, v0.h[\offset] |
||||
mul v23.8h, v23.8h, v0.h[\offset] |
||||
.else |
||||
mul v20.8h, v20.8h, v0.h[\offset] |
||||
mul v22.8h, v22.8h, v0.h[\offset] |
||||
.endif |
||||
sqadd \dst1, \dst1, v20.8h |
||||
sqadd \dst3, \dst3, v22.8h |
||||
.if \size >= 16 |
||||
sqadd \dst2, \dst2, v21.8h |
||||
sqadd \dst4, \dst4, v23.8h |
||||
.endif |
||||
.endm |
||||
|
||||
|
||||
// Instantiate a horizontal filter function for the given size. |
||||
// This can work on 4, 8 or 16 pixels in parallel; for larger
|
||||
// widths it will do 16 pixels at a time and loop horizontally. |
||||
// The actual width is passed in x5, the height in w4 and the |
||||
// filter coefficients in x9. idx2 is the index of the largest |
||||
// filter coefficient (3 or 4) and idx1 is the other one of them. |
||||
.macro do_8tap_h type, size, idx1, idx2 |
||||
function \type\()_8tap_\size\()h_\idx1\idx2 |
||||
sub x2, x2, #3 |
||||
add x6, x0, x1 |
||||
add x7, x2, x3 |
||||
add x1, x1, x1 |
||||
add x3, x3, x3 |
||||
// Only size >= 16 loops horizontally and needs |
||||
// reduced dst stride |
||||
.if \size >= 16 |
||||
sub x1, x1, x5 |
||||
.endif |
||||
// size >= 16 loads two qwords and increments r2, |
||||
// for size 4/8 it's enough with one qword and no |
||||
// postincrement |
||||
.if \size >= 16 |
||||
sub x3, x3, x5 |
||||
sub x3, x3, #8 |
||||
.endif |
||||
// Load the filter vector |
||||
ld1 {v0.8h}, [x9] |
||||
1: |
||||
.if \size >= 16 |
||||
mov x9, x5 |
||||
.endif |
||||
// Load src |
||||
.if \size >= 16 |
||||
ld1 {v4.8b, v5.8b, v6.8b}, [x2], #24 |
||||
ld1 {v16.8b, v17.8b, v18.8b}, [x7], #24 |
||||
.else |
||||
ld1 {v4.8b, v5.8b}, [x2] |
||||
ld1 {v16.8b, v17.8b}, [x7] |
||||
.endif |
||||
uxtl v4.8h, v4.8b |
||||
uxtl v5.8h, v5.8b |
||||
uxtl v16.8h, v16.8b |
||||
uxtl v17.8h, v17.8b |
||||
.if \size >= 16 |
||||
uxtl v6.8h, v6.8b |
||||
uxtl v18.8h, v18.8b |
||||
.endif |
||||
2: |
||||
|
||||
// Accumulate, adding idx2 last with a separate |
||||
// saturating add. The positive filter coefficients |
||||
// for all indices except idx2 must add up to less |
||||
// than 127 for this not to overflow. |
||||
mul v1.8h, v4.8h, v0.h[0] |
||||
mul v24.8h, v16.8h, v0.h[0] |
||||
.if \size >= 16 |
||||
mul v2.8h, v5.8h, v0.h[0] |
||||
mul v25.8h, v17.8h, v0.h[0] |
||||
.endif |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, 1, \size |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, 2, \size |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, \idx1, \size |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, 5, \size |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, 6, \size |
||||
extmla v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, 7, \size |
||||
extmulqadd v1.8h, v2.8h, v24.8h, v25.8h, v4.16b, v5.16b, v6.16b, v16.16b, v17.16b, v18.16b, \idx2, \size |
||||
|
||||
// Round, shift and saturate |
||||
sqrshrun v1.8b, v1.8h, #7 |
||||
sqrshrun v24.8b, v24.8h, #7 |
||||
.if \size >= 16 |
||||
sqrshrun2 v1.16b, v2.8h, #7 |
||||
sqrshrun2 v24.16b, v25.8h, #7 |
||||
.endif |
||||
// Average |
||||
.ifc \type,avg |
||||
.if \size >= 16 |
||||
ld1 {v2.16b}, [x0] |
||||
ld1 {v3.16b}, [x6] |
||||
urhadd v1.16b, v1.16b, v2.16b |
||||
urhadd v24.16b, v24.16b, v3.16b |
||||
.elseif \size == 8 |
||||
ld1 {v2.8b}, [x0] |
||||
ld1 {v3.8b}, [x6] |
||||
urhadd v1.8b, v1.8b, v2.8b |
||||
urhadd v24.8b, v24.8b, v3.8b |
||||
.else |
||||
ld1 {v2.s}[0], [x0] |
||||
ld1 {v3.s}[0], [x6] |
||||
urhadd v1.8b, v1.8b, v2.8b |
||||
urhadd v24.8b, v24.8b, v3.8b |
||||
.endif |
||||
.endif |
||||
// Store and loop horizontally (for size >= 16) |
||||
.if \size >= 16 |
||||
subs x9, x9, #16 |
||||
st1 {v1.16b}, [x0], #16 |
||||
st1 {v24.16b}, [x6], #16 |
||||
beq 3f |
||||
mov v4.16b, v6.16b |
||||
mov v16.16b, v18.16b |
||||
ld1 {v6.16b}, [x2], #16 |
||||
ld1 {v18.16b}, [x7], #16 |
||||
uxtl v5.8h, v6.8b |
||||
uxtl2 v6.8h, v6.16b |
||||
uxtl v17.8h, v18.8b |
||||
uxtl2 v18.8h, v18.16b |
||||
b 2b |
||||
.elseif \size == 8 |
||||
st1 {v1.8b}, [x0] |
||||
st1 {v24.8b}, [x6] |
||||
.else // \size == 4 |
||||
st1 {v1.s}[0], [x0] |
||||
st1 {v24.s}[0], [x6] |
||||
.endif |
||||
3: |
||||
// Loop vertically |
||||
add x0, x0, x1 |
||||
add x6, x6, x1 |
||||
add x2, x2, x3 |
||||
add x7, x7, x3 |
||||
subs w4, w4, #2 |
||||
b.ne 1b |
||||
ret |
||||
endfunc |
||||
.endm |
||||
|
||||
.macro do_8tap_h_size size |
||||
do_8tap_h put, \size, 3, 4 |
||||
do_8tap_h avg, \size, 3, 4 |
||||
do_8tap_h put, \size, 4, 3 |
||||
do_8tap_h avg, \size, 4, 3 |
||||
.endm |
||||
|
||||
do_8tap_h_size 4 |
||||
do_8tap_h_size 8 |
||||
do_8tap_h_size 16 |
||||
|
||||
.macro do_8tap_h_func type, filter, offset, size |
||||
function ff_vp9_\type\()_\filter\()\size\()_h_neon, export=1 |
||||
movrel x6, X(ff_vp9_subpel_filters), 256*\offset |
||||
cmp w5, #8 |
||||
add x9, x6, w5, uxtw #4 |
||||
mov x5, #\size |
||||
.if \size >= 16 |
||||
bge \type\()_8tap_16h_34 |
||||
b \type\()_8tap_16h_43 |
||||
.else |
||||
bge \type\()_8tap_\size\()h_34 |
||||
b \type\()_8tap_\size\()h_43 |
||||
.endif |
||||
endfunc |
||||
.endm |
||||
|
||||
.macro do_8tap_h_filters size |
||||
do_8tap_h_func put, regular, 1, \size |
||||
do_8tap_h_func avg, regular, 1, \size |
||||
do_8tap_h_func put, sharp, 2, \size |
||||
do_8tap_h_func avg, sharp, 2, \size |
||||
do_8tap_h_func put, smooth, 0, \size |
||||
do_8tap_h_func avg, smooth, 0, \size |
||||
.endm |
||||
|
||||
do_8tap_h_filters 64 |
||||
do_8tap_h_filters 32 |
||||
do_8tap_h_filters 16 |
||||
do_8tap_h_filters 8 |
||||
do_8tap_h_filters 4 |
||||
|
||||
|
||||
// Vertical filters |
||||
|
||||
// Round, shift and saturate and store reg1-reg2 over 4 lines |
||||
.macro do_store4 reg1, reg2, tmp1, tmp2, type |
||||
sqrshrun \reg1\().8b, \reg1\().8h, #7 |
||||
sqrshrun \reg2\().8b, \reg2\().8h, #7 |
||||
.ifc \type,avg |
||||
ld1 {\tmp1\().s}[0], [x7], x1 |
||||
ld1 {\tmp2\().s}[0], [x7], x1 |
||||
ld1 {\tmp1\().s}[1], [x7], x1 |
||||
ld1 {\tmp2\().s}[1], [x7], x1 |
||||
urhadd \reg1\().8b, \reg1\().8b, \tmp1\().8b |
||||
urhadd \reg2\().8b, \reg2\().8b, \tmp2\().8b |
||||
.endif |
||||
st1 {\reg1\().s}[0], [x0], x1 |
||||
st1 {\reg2\().s}[0], [x0], x1 |
||||
st1 {\reg1\().s}[1], [x0], x1 |
||||
st1 {\reg2\().s}[1], [x0], x1 |
||||
.endm |
||||
|
||||
// Round, shift and saturate and store reg1-4 |
||||
.macro do_store reg1, reg2, reg3, reg4, tmp1, tmp2, tmp3, tmp4, type |
||||
sqrshrun \reg1\().8b, \reg1\().8h, #7 |
||||
sqrshrun \reg2\().8b, \reg2\().8h, #7 |
||||
sqrshrun \reg3\().8b, \reg3\().8h, #7 |
||||
sqrshrun \reg4\().8b, \reg4\().8h, #7 |
||||
.ifc \type,avg |
||||
ld1 {\tmp1\().8b}, [x7], x1 |
||||
ld1 {\tmp2\().8b}, [x7], x1 |
||||
ld1 {\tmp3\().8b}, [x7], x1 |
||||
ld1 {\tmp4\().8b}, [x7], x1 |
||||
urhadd \reg1\().8b, \reg1\().8b, \tmp1\().8b |
||||
urhadd \reg2\().8b, \reg2\().8b, \tmp2\().8b |
||||
urhadd \reg3\().8b, \reg3\().8b, \tmp3\().8b |
||||
urhadd \reg4\().8b, \reg4\().8b, \tmp4\().8b |
||||
.endif |
||||
st1 {\reg1\().8b}, [x0], x1 |
||||
st1 {\reg2\().8b}, [x0], x1 |
||||
st1 {\reg3\().8b}, [x0], x1 |
||||
st1 {\reg4\().8b}, [x0], x1 |
||||
.endm |
||||
|
||||
// Evaluate the filter twice in parallel, from the inputs src1-src9 into dst1-dst2 |
||||
// (src1-src8 into dst1, src2-src9 into dst2), adding idx2 separately |
||||
// at the end with saturation. Indices 0 and 7 always have negative or zero |
||||
// coefficients, so they can be accumulated into tmp1-tmp2 together with the |
||||
// largest coefficient. |
||||
.macro convolve dst1, dst2, src1, src2, src3, src4, src5, src6, src7, src8, src9, idx1, idx2, tmp1, tmp2 |
||||
mul \dst1\().8h, \src2\().8h, v0.h[1] |
||||
mul \dst2\().8h, \src3\().8h, v0.h[1] |
||||
mul \tmp1\().8h, \src1\().8h, v0.h[0] |
||||
mul \tmp2\().8h, \src2\().8h, v0.h[0] |
||||
mla \dst1\().8h, \src3\().8h, v0.h[2] |
||||
mla \dst2\().8h, \src4\().8h, v0.h[2] |
||||
.if \idx1 == 3 |
||||
mla \dst1\().8h, \src4\().8h, v0.h[3] |
||||
mla \dst2\().8h, \src5\().8h, v0.h[3] |
||||
.else |
||||
mla \dst1\().8h, \src5\().8h, v0.h[4] |
||||
mla \dst2\().8h, \src6\().8h, v0.h[4] |
||||
.endif |
||||
mla \dst1\().8h, \src6\().8h, v0.h[5] |
||||
mla \dst2\().8h, \src7\().8h, v0.h[5] |
||||
mla \tmp1\().8h, \src8\().8h, v0.h[7] |
||||
mla \tmp2\().8h, \src9\().8h, v0.h[7] |
||||
mla \dst1\().8h, \src7\().8h, v0.h[6] |
||||
mla \dst2\().8h, \src8\().8h, v0.h[6] |
||||
.if \idx2 == 3 |
||||
mla \tmp1\().8h, \src4\().8h, v0.h[3] |
||||
mla \tmp2\().8h, \src5\().8h, v0.h[3] |
||||
.else |
||||
mla \tmp1\().8h, \src5\().8h, v0.h[4] |
||||
mla \tmp2\().8h, \src6\().8h, v0.h[4] |
||||
.endif |
||||
sqadd \dst1\().8h, \dst1\().8h, \tmp1\().8h |
||||
sqadd \dst2\().8h, \dst2\().8h, \tmp2\().8h |
||||
.endm |
||||
|
||||
// Load pixels and extend them to 16 bit |
||||
.macro loadl dst1, dst2, dst3, dst4 |
||||
ld1 {v1.8b}, [x2], x3 |
||||
ld1 {v2.8b}, [x2], x3 |
||||
ld1 {v3.8b}, [x2], x3 |
||||
.ifnb \dst4 |
||||
ld1 {v4.8b}, [x2], x3 |
||||
.endif |
||||
uxtl \dst1\().8h, v1.8b |
||||
uxtl \dst2\().8h, v2.8b |
||||
uxtl \dst3\().8h, v3.8b |
||||
.ifnb \dst4 |
||||
uxtl \dst4\().8h, v4.8b |
||||
.endif |
||||
.endm |
||||
|
||||
// Instantiate a vertical filter function for filtering 8 pixels at a time. |
||||
// The height is passed in x4, the width in x5 and the filter coefficients |
||||
// in x6. idx2 is the index of the largest filter coefficient (3 or 4) |
||||
// and idx1 is the other one of them. |
||||
.macro do_8tap_8v type, idx1, idx2 |
||||
function \type\()_8tap_8v_\idx1\idx2 |
||||
sub x2, x2, x3, lsl #1 |
||||
sub x2, x2, x3 |
||||
ld1 {v0.8h}, [x6] |
||||
1: |
||||
.ifc \type,avg |
||||
mov x7, x0 |
||||
.endif |
||||
mov x6, x4 |
||||
|
||||
loadl v17, v18, v19 |
||||
|
||||
loadl v20, v21, v22, v23 |
||||
2: |
||||
loadl v24, v25, v26, v27 |
||||
convolve v1, v2, v17, v18, v19, v20, v21, v22, v23, v24, v25, \idx1, \idx2, v5, v6 |
||||
convolve v3, v4, v19, v20, v21, v22, v23, v24, v25, v26, v27, \idx1, \idx2, v5, v6 |
||||
do_store v1, v2, v3, v4, v5, v6, v7, v28, \type |
||||
|
||||
subs x6, x6, #4 |
||||
b.eq 8f |
||||
|
||||
loadl v16, v17, v18, v19 |
||||
convolve v1, v2, v21, v22, v23, v24, v25, v26, v27, v16, v17, \idx1, \idx2, v5, v6 |
||||
convolve v3, v4, v23, v24, v25, v26, v27, v16, v17, v18, v19, \idx1, \idx2, v5, v6 |
||||
do_store v1, v2, v3, v4, v5, v6, v7, v28, \type |
||||
|
||||
subs x6, x6, #4 |
||||
b.eq 8f |
||||
|
||||
loadl v20, v21, v22, v23 |
||||
convolve v1, v2, v25, v26, v27, v16, v17, v18, v19, v20, v21, \idx1, \idx2, v5, v6 |
||||
convolve v3, v4, v27, v16, v17, v18, v19, v20, v21, v22, v23, \idx1, \idx2, v5, v6 |
||||
do_store v1, v2, v3, v4, v5, v6, v7, v28, \type |
||||
|
||||
subs x6, x6, #4 |
||||
b.ne 2b |
||||
|
||||
8: |
||||
subs x5, x5, #8 |
||||
b.eq 9f |
||||
// x0 -= h * dst_stride |
||||
msub x0, x1, x4, x0 |
||||
// x2 -= h * src_stride |
||||
msub x2, x3, x4, x2 |
||||
// x2 -= 8 * src_stride |
||||
sub x2, x2, x3, lsl #3 |
||||
// x2 += 1 * src_stride |
||||
add x2, x2, x3 |
||||
add x2, x2, #8 |
||||
add x0, x0, #8 |
||||
b 1b |
||||
9: |
||||
ret |
||||
endfunc |
||||
.endm |
||||
|
||||
do_8tap_8v put, 3, 4 |
||||
do_8tap_8v put, 4, 3 |
||||
do_8tap_8v avg, 3, 4 |
||||
do_8tap_8v avg, 4, 3 |
||||
|
||||
|
||||
// Instantiate a vertical filter function for filtering a 4 pixels wide |
||||
// slice. The first half of the registers contain one row, while the second |
||||
// half of a register contains the second-next row (also stored in the first |
||||
// half of the register two steps ahead). The convolution does two outputs |
||||
// at a time; the output of v17-v24 into one, and v18-v25 into another one.
|
||||
// The first half of first output is the first output row, the first half |
||||
// of the other output is the second output row. The second halves of the |
||||
// registers are rows 3 and 4. |
||||
// This only is designed to work for 4 or 8 output lines. |
||||
.macro do_8tap_4v type, idx1, idx2 |
||||
function \type\()_8tap_4v_\idx1\idx2 |
||||
sub x2, x2, x3, lsl #1 |
||||
sub x2, x2, x3 |
||||
ld1 {v0.8h}, [x6] |
||||
.ifc \type,avg |
||||
mov x7, x0 |
||||
.endif |
||||
|
||||
ld1 {v1.s}[0], [x2], x3 |
||||
ld1 {v2.s}[0], [x2], x3 |
||||
ld1 {v3.s}[0], [x2], x3 |
||||
ld1 {v4.s}[0], [x2], x3 |
||||
ld1 {v5.s}[0], [x2], x3 |
||||
ld1 {v6.s}[0], [x2], x3 |
||||
trn1 v1.2s, v1.2s, v3.2s |
||||
ld1 {v7.s}[0], [x2], x3 |
||||
trn1 v2.2s, v2.2s, v4.2s |
||||
ld1 {v26.s}[0], [x2], x3 |
||||
uxtl v17.8h, v1.8b |
||||
trn1 v3.2s, v3.2s, v5.2s |
||||
ld1 {v27.s}[0], [x2], x3 |
||||
uxtl v18.8h, v2.8b |
||||
trn1 v4.2s, v4.2s, v6.2s |
||||
ld1 {v28.s}[0], [x2], x3 |
||||
uxtl v19.8h, v3.8b |
||||
trn1 v5.2s, v5.2s, v7.2s |
||||
ld1 {v29.s}[0], [x2], x3 |
||||
uxtl v20.8h, v4.8b |
||||
trn1 v6.2s, v6.2s, v26.2s |
||||
uxtl v21.8h, v5.8b |
||||
trn1 v7.2s, v7.2s, v27.2s |
||||
uxtl v22.8h, v6.8b |
||||
trn1 v26.2s, v26.2s, v28.2s |
||||
uxtl v23.8h, v7.8b |
||||
trn1 v27.2s, v27.2s, v29.2s |
||||
uxtl v24.8h, v26.8b |
||||
uxtl v25.8h, v27.8b |
||||
|
||||
convolve v1, v2, v17, v18, v19, v20, v21, v22, v23, v24, v25, \idx1, \idx2, v3, v4 |
||||
do_store4 v1, v2, v5, v6, \type |
||||
|
||||
subs x4, x4, #4 |
||||
b.eq 9f |
||||
|
||||
ld1 {v1.s}[0], [x2], x3 |
||||
ld1 {v2.s}[0], [x2], x3 |
||||
trn1 v28.2s, v28.2s, v1.2s |
||||
trn1 v29.2s, v29.2s, v2.2s |
||||
ld1 {v1.s}[1], [x2], x3 |
||||
uxtl v26.8h, v28.8b |
||||
ld1 {v2.s}[1], [x2], x3 |
||||
uxtl v27.8h, v29.8b |
||||
uxtl v28.8h, v1.8b |
||||
uxtl v29.8h, v2.8b |
||||
|
||||
convolve v1, v2, v21, v22, v23, v24, v25, v26, v27, v28, v29, \idx1, \idx2, v3, v4 |
||||
do_store4 v1, v2, v5, v6, \type |
||||
|
||||
9: |
||||
ret |
||||
endfunc |
||||
.endm |
||||
|
||||
do_8tap_4v put, 3, 4 |
||||
do_8tap_4v put, 4, 3 |
||||
do_8tap_4v avg, 3, 4 |
||||
do_8tap_4v avg, 4, 3 |
||||
|
||||
|
||||
.macro do_8tap_v_func type, filter, offset, size |
||||
function ff_vp9_\type\()_\filter\()\size\()_v_neon, export=1 |
||||
uxtw x4, w4 |
||||
movrel x5, X(ff_vp9_subpel_filters), 256*\offset |
||||
cmp w6, #8 |
||||
add x6, x5, w6, uxtw #4 |
||||
mov x5, #\size |
||||
.if \size >= 8 |
||||
b.ge \type\()_8tap_8v_34 |
||||
b \type\()_8tap_8v_43 |
||||
.else |
||||
b.ge \type\()_8tap_4v_34 |
||||
b \type\()_8tap_4v_43 |
||||
.endif |
||||
endfunc |
||||
.endm |
||||
|
||||
.macro do_8tap_v_filters size |
||||
do_8tap_v_func put, regular, 1, \size |
||||
do_8tap_v_func avg, regular, 1, \size |
||||
do_8tap_v_func put, sharp, 2, \size |
||||
do_8tap_v_func avg, sharp, 2, \size |
||||
do_8tap_v_func put, smooth, 0, \size |
||||
do_8tap_v_func avg, smooth, 0, \size |
||||
.endm |
||||
|
||||
do_8tap_v_filters 64 |
||||
do_8tap_v_filters 32 |
||||
do_8tap_v_filters 16 |
||||
do_8tap_v_filters 8 |
||||
do_8tap_v_filters 4 |
Loading…
Reference in new issue