mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
527 lines
15 KiB
527 lines
15 KiB
/* |
|
* Copyright (c) 2024 Institue of Software Chinese Academy of Sciences (ISCAS). |
|
* |
|
* This file is part of FFmpeg. |
|
* |
|
* FFmpeg is free software; you can redistribute it and/or |
|
* modify it under the terms of the GNU Lesser General Public |
|
* License as published by the Free Software Foundation; either |
|
* version 2.1 of the License, or (at your option) any later version. |
|
* |
|
* FFmpeg is distributed in the hope that it will be useful, |
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
* Lesser General Public License for more details. |
|
* |
|
* You should have received a copy of the GNU Lesser General Public |
|
* License along with FFmpeg; if not, write to the Free Software |
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
*/ |
|
|
|
#include "libavutil/riscv/asm.S" |
|
|
|
.macro pix_abs_ret |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
vmv.x.s a0, v0 |
|
ret |
|
.endm |
|
|
|
func ff_pix_abs16_rvv, zve32x |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 16, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
add a1, a1, a3 |
|
vwsubu.vv v20, v12, v4 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
vmax.vv v16, v16, v20 |
|
add a2, a2, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_pix_abs8_rvv, zve32x |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 8, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
add a1, a1, a3 |
|
vwsubu.vv v20, v12, v4 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
vmax.vv v16, v16, v20 |
|
add a2, a2, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_pix_abs16_x2_rvv, zve32x |
|
csrwi vxrm, 0 |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
li t5, 1 |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 17, e8, m2, tu, ma |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vslide1down.vx v24, v12, t5 |
|
vsetivli zero, 16, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vaaddu.vv v12, v12, v24 |
|
vwsubu.vv v16, v4, v12 |
|
add a1, a1, a3 |
|
vwsubu.vv v20, v12, v4 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
vmax.vv v16, v16, v20 |
|
add a2, a2, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_pix_abs8_x2_rvv, zve32x |
|
csrwi vxrm, 0 |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
li t5, 1 |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 9, e8, m1, tu, ma |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vslide1down.vx v24, v12, t5 |
|
vsetivli zero, 8, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vaaddu.vv v12, v12, v24 |
|
vwsubu.vv v16, v4, v12 |
|
add a1, a1, a3 |
|
vwsubu.vv v20, v12, v4 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
vmax.vv v16, v16, v20 |
|
add a2, a2, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_pix_abs16_y2_rvv, zve32x |
|
csrwi vxrm, 0 |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
add t1, a2, a3 |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 16, e8, m1, tu, ma |
|
vle8.v v12, (a2) |
|
vle8.v v24, (t1) |
|
addi a4, a4, -1 |
|
vle8.v v4, (a1) |
|
vaaddu.vv v12, v12, v24 |
|
vwsubu.vv v16, v4, v12 |
|
vwsubu.vv v20, v12, v4 |
|
add a1, a1, a3 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
add a2, a2, a3 |
|
vmax.vv v16, v16, v20 |
|
add t1, t1, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_pix_abs8_y2_rvv, zve32x |
|
csrwi vxrm, 0 |
|
vsetivli zero, 1, e32, m1, ta, ma |
|
add t1, a2, a3 |
|
vmv.s.x v0, zero |
|
1: |
|
vsetivli zero, 8, e8, mf2, tu, ma |
|
vle8.v v12, (a2) |
|
vle8.v v24, (t1) |
|
addi a4, a4, -1 |
|
vle8.v v4, (a1) |
|
vaaddu.vv v12, v12, v24 |
|
vwsubu.vv v16, v4, v12 |
|
vwsubu.vv v20, v12, v4 |
|
add a1, a1, a3 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
add a2, a2, a3 |
|
vmax.vv v16, v16, v20 |
|
add t1, t1, a3 |
|
vwredsum.vs v0, v16, v0 |
|
bnez a4, 1b |
|
|
|
pix_abs_ret |
|
endfunc |
|
|
|
func ff_sse16_rvv, zve32x |
|
vsetivli t0, 16, e32, m4, ta, ma |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
vwmacc.vv v24, v16, v16 |
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m4, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
endfunc |
|
|
|
func ff_sse8_rvv, zve32x |
|
vsetivli t0, 8, e32, m2, ta, ma |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
vwmacc.vv v24, v16, v16 |
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m2, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
endfunc |
|
|
|
func ff_sse4_rvv, zve32x |
|
vsetivli t0, 4, e32, m1, ta, ma |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, mf4, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, mf2, tu, ma |
|
vwmacc.vv v24, v16, v16 |
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m1, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
endfunc |
|
|
|
.macro vabsaddu dst src tmp |
|
vneg.v \tmp, \src |
|
vmax.vv \tmp, \src, \tmp |
|
vwaddu.wv \dst, \dst, \tmp |
|
.endm |
|
|
|
.macro vsad_vsse16 type |
|
vsetivli t0, 16, e32, m4, ta, ma |
|
addi a4, a4, -1 |
|
add t1, a1, a3 |
|
add t2, a2, a3 |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v8, (t1) |
|
vle8.v v12, (a2) |
|
vle8.v v16, (t2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v28, v4, v12 |
|
vwsubu.wv v12, v28, v8 |
|
vwaddu.wv v28, v12, v16 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
|
|
.ifc \type,abs |
|
vabsaddu v24, v28, v12 |
|
.endif |
|
.ifc \type,square |
|
vwmacc.vv v24, v28, v28 |
|
.endif |
|
|
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
add t1, t1, a3 |
|
add t2, t2, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m4, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
.endm |
|
|
|
.macro vsad_vsse8 type |
|
vsetivli t0, 8, e32, m2, ta, ma |
|
addi a4, a4, -1 |
|
add t1, a1, a3 |
|
add t2, a2, a3 |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v8, (t1) |
|
vle8.v v12, (a2) |
|
vle8.v v16, (t2) |
|
addi a4, a4, -1 |
|
vwsubu.vv v28, v4, v12 |
|
vwsubu.wv v12, v28, v8 |
|
vwaddu.wv v28, v12, v16 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
|
|
.ifc \type,abs |
|
vabsaddu v24, v28, v12 |
|
.endif |
|
.ifc \type,square |
|
vwmacc.vv v24, v28, v28 |
|
.endif |
|
|
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
add t1, t1, a3 |
|
add t2, t2, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m2, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
.endm |
|
|
|
.macro vsad_vsse_intra16 type |
|
vsetivli t0, 16, e32, m4, ta, ma |
|
addi a4, a4, -1 |
|
add t1, a1, a3 |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (t1) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
|
|
.ifc \type,abs |
|
vabsaddu v24, v16, v12 |
|
.endif |
|
.ifc \type,square |
|
vwmacc.vv v24, v16, v16 |
|
.endif |
|
|
|
add a1, a1, a3 |
|
add t1, t1, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m4, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
.endm |
|
|
|
.macro vsad_vsse_intra8 type |
|
vsetivli t0, 8, e32, m2, ta, ma |
|
addi a4, a4, -1 |
|
add t1, a1, a3 |
|
vmv.v.x v24, zero |
|
vmv.s.x v0, zero |
|
1: |
|
vsetvli zero, zero, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (t1) |
|
addi a4, a4, -1 |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
|
|
.ifc \type,abs |
|
vabsaddu v24, v16, v12 |
|
.endif |
|
.ifc \type,square |
|
vwmacc.vv v24, v16, v16 |
|
.endif |
|
|
|
add a1, a1, a3 |
|
add t1, t1, a3 |
|
bnez a4, 1b |
|
|
|
vsetvli zero, zero, e32, m2, tu, ma |
|
vredsum.vs v0, v24, v0 |
|
vmv.x.s a0, v0 |
|
ret |
|
.endm |
|
|
|
func ff_vsse16_rvv, zve32x |
|
vsad_vsse16 square |
|
endfunc |
|
|
|
func ff_vsse8_rvv, zve32x |
|
vsad_vsse8 square |
|
endfunc |
|
|
|
func ff_vsse_intra16_rvv, zve32x |
|
vsad_vsse_intra16 square |
|
endfunc |
|
|
|
func ff_vsse_intra8_rvv, zve32x |
|
vsad_vsse_intra8 square |
|
endfunc |
|
|
|
func ff_vsad16_rvv, zve32x |
|
vsad_vsse16 abs |
|
endfunc |
|
|
|
func ff_vsad8_rvv, zve32x |
|
vsad_vsse8 abs |
|
endfunc |
|
|
|
func ff_vsad_intra16_rvv, zve32x |
|
vsad_vsse_intra16 abs |
|
endfunc |
|
|
|
func ff_vsad_intra8_rvv, zve32x |
|
vsad_vsse_intra8 abs |
|
endfunc |
|
|
|
func ff_nsse16_rvv, zve32x |
|
.macro squarediff16 |
|
vsetivli zero, 16, e8, m1, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m2, tu, ma |
|
vwmacc.vv v24, v16, v16 |
|
.endm |
|
|
|
.macro gradiff16 srcx srcv |
|
vsetivli zero, 16, e8, m1, tu, ma |
|
vle8.v v8, (\srcx) |
|
vslide1down.vx v0, \srcv, t5 |
|
vslide1down.vx v16, v8, t5 |
|
vwsubu.vv v20, \srcv, v0 |
|
vwsubu.wv v0, v20, v8 |
|
vwaddu.wv v20, v0, v16 |
|
vsetivli zero, 15, e16, m2, tu, ma |
|
vneg.v v0, v20 |
|
vmax.vv v0, v20, v0 |
|
.endm |
|
|
|
csrwi vxrm, 0 |
|
vsetivli t0, 16, e32, m4, ta, ma |
|
addi a4, a4, -1 |
|
li t5, 1 |
|
vmv.v.x v24, zero |
|
vmv.v.x v28, zero |
|
1: |
|
add t1, a1, a3 |
|
add t2, a2, a3 |
|
addi a4, a4, -1 |
|
squarediff16 |
|
gradiff16 t1, v4 |
|
vwaddu.wv v28, v28, v0 |
|
gradiff16 t2, v12 |
|
vwsubu.wv v28, v28, v0 |
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
bnez a4, 1b |
|
|
|
squarediff16 |
|
vsetivli zero, 16, e32, m4, tu, ma |
|
vmv.s.x v0, zero |
|
vmv.s.x v4, zero |
|
vredsum.vs v0, v24, v0 |
|
vredsum.vs v4, v28, v4 |
|
vmv.x.s t1, v0 |
|
vmv.x.s t2, v4 |
|
srai t3, t2, 31 |
|
xor t2, t3, t2 |
|
sub t2, t2, t3 |
|
mul t2, t2, a0 |
|
add a0, t2, t1 |
|
|
|
ret |
|
endfunc |
|
|
|
func ff_nsse8_rvv, zve32x |
|
.macro squarediff8 |
|
vsetivli zero, 8, e8, mf2, tu, ma |
|
vle8.v v4, (a1) |
|
vle8.v v12, (a2) |
|
vwsubu.vv v16, v4, v12 |
|
vsetvli zero, zero, e16, m1, tu, ma |
|
vwmacc.vv v24, v16, v16 |
|
.endm |
|
|
|
.macro gradiff8 srcx srcv |
|
vsetivli zero, 8, e8, mf2, tu, ma |
|
vle8.v v8, (\srcx) |
|
vslide1down.vx v0, \srcv, t5 |
|
vslide1down.vx v16, v8, t5 |
|
vwsubu.vv v20, \srcv, v0 |
|
vwsubu.wv v0, v20, v8 |
|
vwaddu.wv v20, v0, v16 |
|
vsetivli zero, 7, e16, m1, tu, ma |
|
vneg.v v0, v20 |
|
vmax.vv v0, v20, v0 |
|
.endm |
|
|
|
csrwi vxrm, 0 |
|
vsetivli t0, 8, e32, m2, ta, ma |
|
addi a4, a4, -1 |
|
li t5, 1 |
|
vmv.v.x v24, zero |
|
vmv.v.x v28, zero |
|
1: |
|
add t1, a1, a3 |
|
add t2, a2, a3 |
|
addi a4, a4, -1 |
|
squarediff8 |
|
gradiff8 t1, v4 |
|
vwaddu.wv v28, v28, v0 |
|
gradiff8 t2, v12 |
|
vwsubu.wv v28, v28, v0 |
|
add a1, a1, a3 |
|
add a2, a2, a3 |
|
bnez a4, 1b |
|
|
|
squarediff8 |
|
vsetivli zero, 8, e32, m2, tu, ma |
|
vmv.s.x v0, zero |
|
vmv.s.x v4, zero |
|
vredsum.vs v0, v24, v0 |
|
vredsum.vs v4, v28, v4 |
|
vmv.x.s t1, v0 |
|
vmv.x.s t2, v4 |
|
srai t3, t2, 31 |
|
xor t2, t3, t2 |
|
sub t2, t2, t3 |
|
mul t2, t2, a0 |
|
add a0, t2, t1 |
|
|
|
ret |
|
endfunc
|
|
|