mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
393 lines
12 KiB
393 lines
12 KiB
/* |
|
* SPDX-License-Identifier: BSD-2-Clause |
|
* |
|
* Copyright © 2024 Rémi Denis-Courmont. |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions are met: |
|
* |
|
* 1. Redistributions of source code must retain the above copyright notice, |
|
* this list of conditions and the following disclaimer. |
|
* |
|
* 2. Redistributions in binary form must reproduce the above copyright notice, |
|
* this list of conditions and the following disclaimer in the documentation |
|
* and/or other materials provided with the distribution. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
|
* POSSIBILITY OF SUCH DAMAGE. |
|
*/ |
|
|
|
#include "libavutil/riscv/asm.S" |
|
|
|
func ff_h264_weight_pixels_simple_8_rvv, zve32x |
|
csrwi vxrm, 0 |
|
sll a5, a5, a3 |
|
1: |
|
vsetvli zero, a6, e16, m2, ta, ma |
|
vle8.v v8, (a0) |
|
addi a2, a2, -1 |
|
vzext.vf2 v24, v8 |
|
vmul.vx v16, v24, a4 |
|
vsadd.vx v16, v16, a5 |
|
vmax.vx v16, v16, zero |
|
vsetvli zero, zero, e8, m1, ta, ma |
|
vnclipu.wx v8, v16, a3 |
|
vse8.v v8, (a0) |
|
add a0, a0, a1 |
|
bnez a2, 1b |
|
|
|
ret |
|
endfunc |
|
|
|
.variant_cc ff_h264_biweight_pixels_simple_8_rvv |
|
func ff_h264_biweight_pixels_simple_8_rvv, zve32x |
|
csrwi vxrm, 2 |
|
addi a7, a7, 1 |
|
ori a7, a7, 1 |
|
sll a7, a7, a4 |
|
addi a4, a4, 1 |
|
1: |
|
vsetvli zero, t6, e16, m2, ta, ma |
|
vle8.v v8, (a0) |
|
addi a3, a3, -1 |
|
vle8.v v12, (a1) |
|
add a1, a1, a2 |
|
vmv.v.x v16, a7 |
|
vsetvli zero, zero, e8, m1, ta, ma |
|
vwmaccsu.vx v16, a5, v8 |
|
vwmaccsu.vx v16, a6, v12 |
|
vsetvli zero, zero, e16, m2, ta, ma |
|
vmax.vx v16, v16, zero |
|
vsetvli zero, zero, e8, m1, ta, ma |
|
vnclipu.wx v8, v16, a4 |
|
vse8.v v8, (a0) |
|
add a0, a0, a2 |
|
bnez a3, 1b |
|
|
|
ret |
|
endfunc |
|
|
|
func ff_h264_weight_pixels_8_rvv, zve32x |
|
csrwi vxrm, 0 |
|
sll a5, a5, a3 |
|
1: |
|
mv t0, a0 |
|
mv t6, a6 |
|
2: |
|
vsetvli t2, a2, e16, m8, ta, ma |
|
vlsseg2e8.v v0, (t0), a1 |
|
addi t6, t6, -2 |
|
vzext.vf2 v16, v0 |
|
vzext.vf2 v24, v4 |
|
vmul.vx v16, v16, a4 |
|
vmul.vx v24, v24, a4 |
|
vsadd.vx v16, v16, a5 |
|
vsadd.vx v24, v24, a5 |
|
vmax.vx v16, v16, zero |
|
vmax.vx v24, v24, zero |
|
vsetvli zero, zero, e8, m4, ta, ma |
|
vnclipu.wx v0, v16, a3 |
|
vnclipu.wx v4, v24, a3 |
|
vssseg2e8.v v0, (t0), a1 |
|
addi t0, t0, 2 |
|
bnez t6, 2b |
|
|
|
mul t3, a1, t2 |
|
sub a2, a2, t2 |
|
add a0, a0, t3 |
|
bnez a2, 1b |
|
|
|
ret |
|
endfunc |
|
|
|
.variant_cc ff_h264_biweight_pixels_8_rvv |
|
func ff_h264_biweight_pixels_8_rvv, zve32x |
|
csrwi vxrm, 2 |
|
addi a7, a7, 1 |
|
ori a7, a7, 1 |
|
sll a7, a7, a4 |
|
addi a4, a4, 1 |
|
1: |
|
mv t0, a0 |
|
mv t1, a1 |
|
mv t5, t6 |
|
2: |
|
vsetvli t2, a3, e16, m8, ta, ma |
|
vlsseg2e8.v v0, (t0), a2 |
|
vlsseg2e8.v v8, (t1), a2 |
|
addi t5, t5, -2 |
|
vmv.v.x v16, a7 |
|
vmv.v.x v24, a7 |
|
vsetvli zero, zero, e8, m4, ta, ma |
|
vwmaccsu.vx v16, a5, v0 |
|
vwmaccsu.vx v24, a5, v4 |
|
vwmaccsu.vx v16, a6, v8 |
|
vwmaccsu.vx v24, a6, v12 |
|
vsetvli zero, zero, e16, m8, ta, ma |
|
vmax.vx v16, v16, zero |
|
vmax.vx v24, v24, zero |
|
vsetvli zero, zero, e8, m4, ta, ma |
|
vnclipu.wx v0, v16, a4 |
|
vnclipu.wx v4, v24, a4 |
|
vssseg2e8.v v0, (t0), a2 |
|
addi t0, t0, 2 |
|
addi t1, t1, 2 |
|
bnez t5, 2b |
|
|
|
mul t3, a2, t2 |
|
sub a3, a3, t2 |
|
add a0, a0, t3 |
|
add a1, a1, t3 |
|
bnez a3, 1b |
|
|
|
ret |
|
endfunc |
|
|
|
.irp w, 16, 8, 4, 2 |
|
func ff_h264_weight_pixels\w\()_8_rvv, zve32x |
|
lpad 0 |
|
li a6, \w |
|
.if \w == 16 |
|
j ff_h264_weight_pixels_simple_8_rvv |
|
.else |
|
j ff_h264_weight_pixels_8_rvv |
|
.endif |
|
endfunc |
|
|
|
func ff_h264_biweight_pixels\w\()_8_rvv, zve32x |
|
lpad 0 |
|
li t6, \w |
|
.if \w == 16 |
|
j ff_h264_biweight_pixels_simple_8_rvv |
|
.else |
|
j ff_h264_biweight_pixels_8_rvv |
|
.endif |
|
endfunc |
|
.endr |
|
|
|
.global ff_h264_weight_funcs_8_rvv |
|
.hidden ff_h264_weight_funcs_8_rvv |
|
const ff_h264_weight_funcs_8_rvv, relocate=1 |
|
.irp w, 16, 8, 4, 2 |
|
#if __riscv_xlen == 32 |
|
.word ff_h264_weight_pixels\w\()_8_rvv |
|
.word ff_h264_biweight_pixels\w\()_8_rvv |
|
#elif __riscv_xlen == 64 |
|
.dword ff_h264_weight_pixels\w\()_8_rvv |
|
.dword ff_h264_biweight_pixels\w\()_8_rvv |
|
#else |
|
.qword ff_h264_weight_pixels\w\()_8_rvv |
|
.qword ff_h264_biweight_pixels\w\()_8_rvv |
|
#endif |
|
.endr |
|
endconst |
|
|
|
.macro loop_filter type, inners, e8mul, e16mul |
|
.variant_cc ff_h264_loop_filter_\type\()_8_rvv |
|
func ff_h264_loop_filter_\type\()_8_rvv, zve32x |
|
# p2: v8, p1: v9, p0: v10, q0: v11, q1: v12, q2: v13 |
|
# alpha: a2, beta: a3, tc_orig: v6 |
|
csrwi vxrm, 0 |
|
.ifc \type, luma |
|
vaaddu.vv v14, v10, v11 # (p0 + q0 + 1) / 2 |
|
.endif |
|
vwsubu.vv v16, v9, v12 |
|
.ifc \type, luma |
|
vwaddu.vv v18, v8, v14 |
|
vwaddu.vv v20, v13, v14 |
|
.endif |
|
vnsra.wi v24, v16, 2 # (p1 - q1) / 4 |
|
.ifc \type, luma |
|
vnsrl.wi v14, v18, 1 |
|
vnsrl.wi v15, v20, 1 |
|
vneg.v v5, v6 # -tc_orig |
|
.endif |
|
vwsubu.vv v22, v11, v10 # q0 - p0 |
|
.ifc \type, luma |
|
vwsubu.vv v18, v14, v9 |
|
vwsubu.vv v20, v15, v12 |
|
.endif |
|
vwadd.wv v16, v22, v24 |
|
vmsge.vi v7, v6, 0 # tc_orig >= 0 |
|
.ifc \type, luma |
|
vnclip.wi v14, v18, 0 |
|
vnclip.wi v15, v20, 0 |
|
.endif |
|
vnclip.wi v16, v16, 1 # clip8((q0 - p0 + (p1 - q1) / 4 + 1) >> 1) |
|
.ifc \type, luma |
|
vmin.vv v14, v14, v6 |
|
vmin.vv v15, v15, v6 |
|
vmax.vv v14, v14, v5 # clip(p2 + ... - p1, +/-tc_orig) |
|
vmax.vv v15, v15, v5 # clip(q2 + ... - q1, +/-tc_orig) |
|
.endif |
|
vwsubu.vv v20, v10, v11 |
|
vwsubu.vv v24, v9, v10 |
|
vwsubu.vv v26, v10, v9 |
|
vwsubu.vv v28, v12, v11 |
|
vwsubu.vv v30, v11, v12 |
|
.ifc \type, luma |
|
vwsubu.vv v0, v8, v10 |
|
vwsubu.vv v2, v10, v8 |
|
vwsubu.vv v4, v13, v11 |
|
vwsubu.vv v18, v11, v13 |
|
.endif |
|
vsetvli zero, zero, e16, \e16mul, ta, ma |
|
vmax.vv v20, v20, v22 # abs(p0 - q0) |
|
vmax.vv v24, v24, v26 # abs(p1 - p0) |
|
vmax.vv v28, v28, v30 # abs(q1 - q0) |
|
.ifc \type, luma |
|
vmax.vv v22, v0, v2 # abs(p2 - p0) |
|
vmax.vv v26, v4, v18 # abs(q2 - q0) |
|
.endif |
|
vmslt.vx v1, v20, a2 |
|
vmslt.vx v2, v24, a3 |
|
vmand.mm v7, v7, v1 |
|
vmslt.vx v3, v28, a3 |
|
vmand.mm v7, v7, v2 |
|
.ifc \type, luma |
|
vmslt.vx v0, v22, a3 |
|
vmand.mm v7, v7, v3 # whether to update p0 and q0 |
|
vmslt.vx v1, v26, a3 |
|
vmand.mm v0, v0, v7 |
|
.else |
|
vmand.mm v0, v7, v3 # whether to update p0 and q0 |
|
.endif |
|
vsetvli zero, zero, e8, \e8mul, ta, mu |
|
.ifc \type, luma |
|
vadd.vi v6, v6, 1, v0.t # tc++ |
|
vadd.vv v9, v9, v14, v0.t # p1' |
|
vmand.mm v0, v1, v7 |
|
vadd.vi v6, v6, 1, v0.t # tc++ |
|
vadd.vv v12, v12, v15, v0.t # q1' |
|
vmmv.m v0, v7 |
|
.endif |
|
vneg.v v5, v6 # -tc |
|
vmin.vv v16, v16, v6 |
|
vwcvtu.x.x.v v18, v10 |
|
vmax.vv v16, v16, v5 |
|
vwcvtu.x.x.v v20, v11 |
|
vwadd.wv v18, v18, v16 |
|
vwsub.wv v20, v20, v16 |
|
vsetvli zero, zero, e16, \e16mul, ta, ma |
|
vmax.vx v18, v18, zero |
|
vmax.vx v20, v20, zero |
|
vsetvli zero, zero, e8, \e8mul, ta, mu |
|
vnclipu.wi v10, v18, 0, v0.t # p0' |
|
vnclipu.wi v11, v20, 0, v0.t # q0' |
|
jr t0 |
|
endfunc |
|
|
|
func ff_h264_v_loop_filter_\type\()_8_rvv, zve32x |
|
lpad 0 |
|
.ifc \type, luma |
|
vsetivli zero, 4, e32, m1, ta, ma |
|
vle8.v v4, (a4) |
|
li t0, 0x01010101 |
|
vzext.vf4 v6, v4 |
|
.else |
|
vsetivli zero, 4, e16, mf2, ta, ma |
|
vle8.v v4, (a4) |
|
li t0, 0x0101 |
|
vzext.vf2 v6, v4 |
|
.endif |
|
sub t3, a0, a1 |
|
vmul.vx v6, v6, t0 |
|
vsetivli zero, 4 * \inners, e8, \e8mul, ta, ma |
|
vle8.v v11, (a0) |
|
sub t2, t3, a1 |
|
vid.v v0 |
|
vle8.v v10, (t3) |
|
add t5, a0, a1 |
|
vle8.v v9, (t2) |
|
.ifc \type, luma |
|
sub t1, t2, a1 |
|
.endif |
|
vle8.v v12, (t5) |
|
.ifc \type, luma |
|
add t6, t5, a1 |
|
vle8.v v8, (t1) |
|
vle8.v v13, (t6) |
|
.endif |
|
jal t0, ff_h264_loop_filter_\type\()_8_rvv |
|
.ifc \type, luma |
|
vse8.v v9, (t2) |
|
vse8.v v12, (t5) |
|
.endif |
|
vse8.v v10, (t3) |
|
vse8.v v11, (a0) |
|
ret |
|
endfunc |
|
.endm |
|
|
|
loop_filter luma, 4, m1, m2 |
|
loop_filter chroma, 2, mf2, m1 |
|
|
|
func ff_h264_h_loop_filter_luma_8_rvv, zve32x |
|
lpad 0 |
|
vsetivli zero, 4, e32, m1, ta, ma |
|
vle8.v v4, (a4) |
|
li t0, 0x01010101 |
|
vzext.vf4 v6, v4 |
|
addi a0, a0, -3 |
|
vmul.vx v6, v6, t0 |
|
vsetivli zero, 16, e8, m1, ta, ma |
|
vlsseg6e8.v v8, (a0), a1 |
|
addi a0, a0, 1 |
|
jal t0, ff_h264_loop_filter_luma_8_rvv |
|
vssseg4e8.v v9, (a0), a1 |
|
ret |
|
endfunc |
|
|
|
func ff_h264_h_loop_filter_luma_mbaff_8_rvv, zve32x |
|
lpad 0 |
|
vsetivli zero, 4, e16, mf2, ta, ma |
|
vle8.v v4, (a4) |
|
li t0, 0x0101 |
|
vzext.vf2 v6, v4 |
|
addi a0, a0, -3 |
|
vmul.vx v6, v6, t0 # tc_orig |
|
vsetivli zero, 8, e8, m1, ta, ma |
|
vlsseg6e8.v v8, (a0), a1 |
|
addi a0, a0, 1 |
|
jal t0, ff_h264_loop_filter_luma_8_rvv |
|
vssseg4e8.v v9, (a0), a1 |
|
ret |
|
endfunc |
|
|
|
func ff_h264_h_loop_filter_chroma_8_rvv, zve32x |
|
lpad 0 |
|
vsetivli zero, 4, e16, mf2, ta, ma |
|
vle8.v v4, (a4) |
|
li t0, 0x0101 |
|
vzext.vf2 v6, v4 |
|
addi a0, a0, -2 |
|
vmul.vx v6, v6, t0 |
|
vsetivli zero, 8, e8, mf2, ta, ma |
|
vlsseg4e8.v v9, (a0), a1 |
|
addi a0, a0, 1 |
|
jal t0, ff_h264_loop_filter_chroma_8_rvv |
|
vssseg2e8.v v10, (a0), a1 |
|
ret |
|
endfunc |
|
|
|
func ff_h264_h_loop_filter_chroma_mbaff_8_rvv, zve32x |
|
lpad 0 |
|
vsetivli zero, 4, e8, mf4, ta, ma |
|
vle8.v v6, (a4) |
|
addi a0, a0, -2 |
|
vsetivli zero, 4, e8, mf2, ta, ma |
|
vlsseg4e8.v v9, (a0), a1 |
|
addi a0, a0, 1 |
|
jal t0, ff_h264_loop_filter_chroma_8_rvv |
|
vssseg2e8.v v10, (a0), a1 |
|
ret |
|
endfunc
|
|
|