You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1869 lines
70 KiB

/*
* Loongson LSX optimized swscale
*
* Copyright (c) 2023 Loongson Technology Corporation Limited
* Contributed by Lu Wang <wanglu@loongson.cn>
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavcodec/loongarch/loongson_asm.S"
/* void ff_hscale_8_to_15_lsx(SwsContext *c, int16_t *dst, int dstW,
* const uint8_t *src, const int16_t *filter,
* const int32_t *filterPos, int filterSize)
*/
function ff_hscale_8_to_15_lsx
addi.d sp, sp, -72
st.d s0, sp, 0
st.d s1, sp, 8
st.d s2, sp, 16
st.d s3, sp, 24
st.d s4, sp, 32
st.d s5, sp, 40
st.d s6, sp, 48
st.d s7, sp, 56
st.d s8, sp, 64
li.w t0, 32767
li.w t8, 8
li.w t7, 4
vldi vr0, 0
vreplgr2vr.w vr20, t0
beq a6, t7, .LOOP_DSTW4
beq a6, t8, .LOOP_DSTW8
blt t8, a6, .LOOP_START
b .END_DSTW4
.LOOP_START:
li.w t1, 0
li.w s1, 0
li.w s2, 0
li.w s3, 0
li.w s4, 0
li.w s5, 0
vldi vr22, 0
addi.w s0, a6, -7
slli.w s7, a6, 1
slli.w s8, a6, 2
add.w t6, s7, s8
.LOOP_DSTW:
ld.w t2, a5, 0
ld.w t3, a5, 4
ld.w t4, a5, 8
ld.w t5, a5, 12
fldx.d f1, a3, t2
fldx.d f2, a3, t3
fldx.d f3, a3, t4
fldx.d f4, a3, t5
vld vr9, a4, 0
vldx vr10, a4, s7
vldx vr11, a4, s8
vldx vr12, a4, t6
vilvl.b vr1, vr0, vr1
vilvl.b vr2, vr0, vr2
vilvl.b vr3, vr0, vr3
vilvl.b vr4, vr0, vr4
vdp2.w.h vr17, vr1, vr9
vdp2.w.h vr18, vr2, vr10
vdp2.w.h vr19, vr3, vr11
vdp2.w.h vr21, vr4, vr12
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.d vr1, vr3, vr1
vadd.w vr22, vr22, vr1
addi.w s1, s1, 8
addi.d a3, a3, 8
addi.d a4, a4, 16
blt s1, s0, .LOOP_DSTW
blt s1, a6, .DSTWA
b .END_FILTER
.DSTWA:
ld.w t2, a5, 0
li.w t3, 0
move s6, s1
.FILTERSIZEA:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s2, s2, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERSIZEA
ld.w t2, a5, 4
li.w t3, 0
move s6, s1
addi.w t1, t1, 1
.FILTERSIZEB:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s3, s3, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERSIZEB
ld.w t2, a5, 8
addi.w t1, t1, 1
li.w t3, 0
move s6, s1
.FILTERSIZEC:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s4, s4, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERSIZEC
ld.w t2, a5, 12
addi.w t1, t1, 1
move s6, s1
li.w t3, 0
.FILTERSIZED:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s5, s5, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERSIZED
.END_FILTER:
vpickve2gr.w t1, vr22, 0
vpickve2gr.w t2, vr22, 1
vpickve2gr.w t3, vr22, 2
vpickve2gr.w t4, vr22, 3
add.w s2, s2, t1
add.w s3, s3, t2
add.w s4, s4, t3
add.w s5, s5, t4
srai.w s2, s2, 7
srai.w s3, s3, 7
srai.w s4, s4, 7
srai.w s5, s5, 7
slt t1, s2, t0
slt t2, s3, t0
slt t3, s4, t0
slt t4, s5, t0
maskeqz s2, s2, t1
maskeqz s3, s3, t2
maskeqz s4, s4, t3
maskeqz s5, s5, t4
masknez t1, t0, t1
masknez t2, t0, t2
masknez t3, t0, t3
masknez t4, t0, t4
or s2, s2, t1
or s3, s3, t2
or s4, s4, t3
or s5, s5, t4
st.h s2, a1, 0
st.h s3, a1, 2
st.h s4, a1, 4
st.h s5, a1, 6
addi.d a1, a1, 8
sub.d a3, a3, s1
addi.d a5, a5, 16
slli.d t3, a6, 3
add.d a4, a4, t3
sub.d a4, a4, s1
sub.d a4, a4, s1
addi.d a2, a2, -4
bge a2, t7, .LOOP_START
blt zero, a2, .RES
b .END_LOOP
.RES:
li.w t1, 0
.DSTW:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTERSIZE:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTERSIZE
srai.w t8, t8, 7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DSTW
b .END_LOOP
.LOOP_DSTW8:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
fldx.d f1, a3, t1
fldx.d f2, a3, t2
fldx.d f3, a3, t3
fldx.d f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
fldx.d f5, a3, t1
fldx.d f6, a3, t2
fldx.d f7, a3, t3
fldx.d f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vld vr13, a4, 64
vld vr14, a4, 80
vld vr15, a4, 96
vld vr16, a4, 112
vilvl.b vr1, vr0, vr1
vilvl.b vr2, vr0, vr2
vilvl.b vr3, vr0, vr3
vilvl.b vr4, vr0, vr4
vilvl.b vr5, vr0, vr5
vilvl.b vr6, vr0, vr6
vilvl.b vr7, vr0, vr7
vilvl.b vr8, vr0, vr8
vdp2.w.h vr17, vr1, vr9
vdp2.w.h vr18, vr2, vr10
vdp2.w.h vr19, vr3, vr11
vdp2.w.h vr21, vr4, vr12
vdp2.w.h vr1, vr5, vr13
vdp2.w.h vr2, vr6, vr14
vdp2.w.h vr3, vr7, vr15
vdp2.w.h vr4, vr8, vr16
vhaddw.d.w vr5, vr1, vr1
vhaddw.d.w vr6, vr2, vr2
vhaddw.d.w vr7, vr3, vr3
vhaddw.d.w vr8, vr4, vr4
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vhaddw.q.d vr5, vr5, vr5
vhaddw.q.d vr6, vr6, vr6
vhaddw.q.d vr7, vr7, vr7
vhaddw.q.d vr8, vr8, vr8
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.d vr1, vr3, vr1
vilvl.d vr5, vr7, vr5
vsrai.w vr1, vr1, 7
vsrai.w vr5, vr5, 7
vmin.w vr1, vr1, vr20
vmin.w vr5, vr5, vr20
vpickev.h vr1, vr5, vr1
vst vr1, a1, 0
addi.d a1, a1, 16
addi.d a5, a5, 32
addi.d a4, a4, 128
addi.d a2, a2, -8
bge a2, t8, .LOOP_DSTW8
blt zero, a2, .RES8
b .END_LOOP
.RES8:
li.w t1, 0
.DSTW8:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTERSIZE8:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTERSIZE8
srai.w t8, t8, 7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DSTW8
b .END_LOOP
.LOOP_DSTW4:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
fldx.s f1, a3, t1
fldx.s f2, a3, t2
fldx.s f3, a3, t3
fldx.s f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
fldx.s f5, a3, t1
fldx.s f6, a3, t2
fldx.s f7, a3, t3
fldx.s f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.b vr1, vr0, vr1
vilvl.b vr3, vr0, vr3
vilvl.b vr5, vr0, vr5
vilvl.b vr7, vr0, vr7
vdp2.w.h vr13, vr1, vr9
vdp2.w.h vr14, vr3, vr10
vdp2.w.h vr15, vr5, vr11
vdp2.w.h vr16, vr7, vr12
vhaddw.d.w vr13, vr13, vr13
vhaddw.d.w vr14, vr14, vr14
vhaddw.d.w vr15, vr15, vr15
vhaddw.d.w vr16, vr16, vr16
vpickev.w vr13, vr14, vr13
vpickev.w vr15, vr16, vr15
vsrai.w vr13, vr13, 7
vsrai.w vr15, vr15, 7
vmin.w vr13, vr13, vr20
vmin.w vr15, vr15, vr20
vpickev.h vr13, vr15, vr13
vst vr13, a1, 0
addi.d a1, a1, 16
addi.d a5, a5, 32
addi.d a4, a4, 64
addi.d a2, a2, -8
bge a2, t8, .LOOP_DSTW4
blt zero, a2, .RES4
b .END_LOOP
.RES4:
li.w t1, 0
.DSTW4:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTERSIZE4:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTERSIZE4
srai.w t8, t8, 7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DSTW4
b .END_LOOP
.END_DSTW4:
li.w t1, 0
.LOOP_DSTW1:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTERSIZE1:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTERSIZE1
srai.w t8, t8, 7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .LOOP_DSTW1
b .END_LOOP
.END_LOOP:
ld.d s0, sp, 0
ld.d s1, sp, 8
ld.d s2, sp, 16
ld.d s3, sp, 24
ld.d s4, sp, 32
ld.d s5, sp, 40
ld.d s6, sp, 48
ld.d s7, sp, 56
ld.d s8, sp, 64
addi.d sp, sp, 72
endfunc
/* void ff_hscale_8_to_19_lsx(SwsContext *c, int16_t *dst, int dstW,
* const uint8_t *src, const int16_t *filter,
* const int32_t *filterPos, int filterSize)
*/
function ff_hscale_8_to_19_lsx
addi.d sp, sp, -72
st.d s0, sp, 0
st.d s1, sp, 8
st.d s2, sp, 16
st.d s3, sp, 24
st.d s4, sp, 32
st.d s5, sp, 40
st.d s6, sp, 48
st.d s7, sp, 56
st.d s8, sp, 64
li.w t0, 524287
li.w t8, 8
li.w t7, 4
vldi vr0, 0
vreplgr2vr.w vr20, t0
beq a6, t7, .LOOP_DST4
beq a6, t8, .LOOP_DST8
blt t8, a6, .LOOP
b .END_DST4
.LOOP:
li.w t1, 0
li.w s1, 0
li.w s2, 0
li.w s3, 0
li.w s4, 0
li.w s5, 0
vldi vr22, 0
addi.w s0, a6, -7
slli.w s7, a6, 1
slli.w s8, a6, 2
add.w t6, s7, s8
.LOOP_DST:
ld.w t2, a5, 0
ld.w t3, a5, 4
ld.w t4, a5, 8
ld.w t5, a5, 12
fldx.d f1, a3, t2
fldx.d f2, a3, t3
fldx.d f3, a3, t4
fldx.d f4, a3, t5
vld vr9, a4, 0
vldx vr10, a4, s7
vldx vr11, a4, s8
vldx vr12, a4, t6
vilvl.b vr1, vr0, vr1
vilvl.b vr2, vr0, vr2
vilvl.b vr3, vr0, vr3
vilvl.b vr4, vr0, vr4
vdp2.w.h vr17, vr1, vr9
vdp2.w.h vr18, vr2, vr10
vdp2.w.h vr19, vr3, vr11
vdp2.w.h vr21, vr4, vr12
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.d vr1, vr3, vr1
vadd.w vr22, vr22, vr1
addi.w s1, s1, 8
addi.d a3, a3, 8
addi.d a4, a4, 16
blt s1, s0, .LOOP_DST
blt s1, a6, .DSTA
b .END_FILTERA
.DSTA:
ld.w t2, a5, 0
li.w t3, 0
move s6, s1
.FILTERA:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s2, s2, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERA
ld.w t2, a5, 4
li.w t3, 0
move s6, s1
addi.w t1, t1, 1
.FILTERB:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s3, s3, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERB
ld.w t2, a5, 8
addi.w t1, t1, 1
li.w t3, 0
move s6, s1
.FILTERC:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s4, s4, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERC
ld.w t2, a5, 12
addi.w t1, t1, 1
move s6, s1
li.w t3, 0
.FILTERD:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s5, s5, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .FILTERD
.END_FILTERA:
vpickve2gr.w t1, vr22, 0
vpickve2gr.w t2, vr22, 1
vpickve2gr.w t3, vr22, 2
vpickve2gr.w t4, vr22, 3
add.w s2, s2, t1
add.w s3, s3, t2
add.w s4, s4, t3
add.w s5, s5, t4
srai.w s2, s2, 3
srai.w s3, s3, 3
srai.w s4, s4, 3
srai.w s5, s5, 3
slt t1, s2, t0
slt t2, s3, t0
slt t3, s4, t0
slt t4, s5, t0
maskeqz s2, s2, t1
maskeqz s3, s3, t2
maskeqz s4, s4, t3
maskeqz s5, s5, t4
masknez t1, t0, t1
masknez t2, t0, t2
masknez t3, t0, t3
masknez t4, t0, t4
or s2, s2, t1
or s3, s3, t2
or s4, s4, t3
or s5, s5, t4
st.w s2, a1, 0
st.w s3, a1, 4
st.w s4, a1, 8
st.w s5, a1, 12
addi.d a1, a1, 16
sub.d a3, a3, s1
addi.d a5, a5, 16
slli.d t3, a6, 3
add.d a4, a4, t3
sub.d a4, a4, s1
sub.d a4, a4, s1
addi.d a2, a2, -4
bge a2, t7, .LOOP
blt zero, a2, .RESA
b .END
.RESA:
li.w t1, 0
.DST:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTER:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTER
srai.w t8, t8, 3
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DST
b .END
.LOOP_DST8:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
fldx.d f1, a3, t1
fldx.d f2, a3, t2
fldx.d f3, a3, t3
fldx.d f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
fldx.d f5, a3, t1
fldx.d f6, a3, t2
fldx.d f7, a3, t3
fldx.d f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vld vr13, a4, 64
vld vr14, a4, 80
vld vr15, a4, 96
vld vr16, a4, 112
vilvl.b vr1, vr0, vr1
vilvl.b vr2, vr0, vr2
vilvl.b vr3, vr0, vr3
vilvl.b vr4, vr0, vr4
vilvl.b vr5, vr0, vr5
vilvl.b vr6, vr0, vr6
vilvl.b vr7, vr0, vr7
vilvl.b vr8, vr0, vr8
vdp2.w.h vr17, vr1, vr9
vdp2.w.h vr18, vr2, vr10
vdp2.w.h vr19, vr3, vr11
vdp2.w.h vr21, vr4, vr12
vdp2.w.h vr1, vr5, vr13
vdp2.w.h vr2, vr6, vr14
vdp2.w.h vr3, vr7, vr15
vdp2.w.h vr4, vr8, vr16
vhaddw.d.w vr5, vr1, vr1
vhaddw.d.w vr6, vr2, vr2
vhaddw.d.w vr7, vr3, vr3
vhaddw.d.w vr8, vr4, vr4
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vhaddw.q.d vr5, vr5, vr5
vhaddw.q.d vr6, vr6, vr6
vhaddw.q.d vr7, vr7, vr7
vhaddw.q.d vr8, vr8, vr8
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.d vr1, vr3, vr1
vilvl.d vr5, vr7, vr5
vsrai.w vr1, vr1, 3
vsrai.w vr5, vr5, 3
vmin.w vr1, vr1, vr20
vmin.w vr5, vr5, vr20
vst vr1, a1, 0
vst vr5, a1, 16
addi.d a1, a1, 32
addi.d a5, a5, 32
addi.d a4, a4, 128
addi.d a2, a2, -8
bge a2, t8, .LOOP_DST8
blt zero, a2, .REST8
b .END
.REST8:
li.w t1, 0
.DST8:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTER8:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTER8
srai.w t8, t8, 3
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DST8
b .END
.LOOP_DST4:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
fldx.s f1, a3, t1
fldx.s f2, a3, t2
fldx.s f3, a3, t3
fldx.s f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
fldx.s f5, a3, t1
fldx.s f6, a3, t2
fldx.s f7, a3, t3
fldx.s f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.b vr1, vr0, vr1
vilvl.b vr3, vr0, vr3
vilvl.b vr5, vr0, vr5
vilvl.b vr7, vr0, vr7
vdp2.w.h vr13, vr1, vr9
vdp2.w.h vr14, vr3, vr10
vdp2.w.h vr15, vr5, vr11
vdp2.w.h vr16, vr7, vr12
vhaddw.d.w vr13, vr13, vr13
vhaddw.d.w vr14, vr14, vr14
vhaddw.d.w vr15, vr15, vr15
vhaddw.d.w vr16, vr16, vr16
vpickev.w vr13, vr14, vr13
vpickev.w vr15, vr16, vr15
vsrai.w vr13, vr13, 3
vsrai.w vr15, vr15, 3
vmin.w vr13, vr13, vr20
vmin.w vr15, vr15, vr20
vst vr13, a1, 0
vst vr15, a1, 16
addi.d a1, a1, 32
addi.d a5, a5, 32
addi.d a4, a4, 64
addi.d a2, a2, -8
bge a2, t8, .LOOP_DST4
blt zero, a2, .REST4
b .END
.REST4:
li.w t1, 0
.DST4:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTER4:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTER4
srai.w t8, t8, 3
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .DST4
b .END
.END_DST4:
li.w t1, 0
.LOOP_DST1:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.FILTER1:
add.w t4, t2, t3
ldx.bu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .FILTER1
srai.w t8, t8, 3
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .LOOP_DST1
b .END
.END:
ld.d s0, sp, 0
ld.d s1, sp, 8
ld.d s2, sp, 16
ld.d s3, sp, 24
ld.d s4, sp, 32
ld.d s5, sp, 40
ld.d s6, sp, 48
ld.d s7, sp, 56
ld.d s8, sp, 64
addi.d sp, sp, 72
endfunc
/* void ff_hscale_16_to_15_sub_lsx(SwsContext *c, int16_t *dst, int dstW,
* const uint8_t *src, const int16_t *filter,
* const int32_t *filterPos, int filterSize, int sh)
*/
function ff_hscale_16_to_15_sub_lsx
addi.d sp, sp, -72
st.d s0, sp, 0
st.d s1, sp, 8
st.d s2, sp, 16
st.d s3, sp, 24
st.d s4, sp, 32
st.d s5, sp, 40
st.d s6, sp, 48
st.d s7, sp, 56
st.d s8, sp, 64
li.w t0, 32767
li.w t8, 8
li.w t7, 4
vreplgr2vr.w vr20, t0
vreplgr2vr.w vr0, a7
beq a6, t7, .LOOP_HS15_DST4
beq a6, t8, .LOOP_HS15_DST8
blt t8, a6, .LOOP_HS15
b .END_HS15_DST4
.LOOP_HS15:
li.w t1, 0
li.w s1, 0
li.w s2, 0
li.w s3, 0
li.w s4, 0
li.w s5, 0
vldi vr22, 0
addi.w s0, a6, -7
slli.w s7, a6, 1
slli.w s8, a6, 2
add.w t6, s7, s8
.LOOP_HS15_DST:
ld.w t2, a5, 0
ld.w t3, a5, 4
ld.w t4, a5, 8
ld.w t5, a5, 12
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
slli.w t5, t5, 1
vldx vr1, a3, t2
vldx vr2, a3, t3
vldx vr3, a3, t4
vldx vr4, a3, t5
vld vr9, a4, 0
vldx vr10, a4, s7
vldx vr11, a4, s8
vldx vr12, a4, t6
vmulwev.w.hu.h vr17, vr1, vr9
vmulwev.w.hu.h vr18, vr2, vr10
vmulwev.w.hu.h vr19, vr3, vr11
vmulwev.w.hu.h vr21, vr4, vr12
vmaddwod.w.hu.h vr17, vr1, vr9
vmaddwod.w.hu.h vr18, vr2, vr10
vmaddwod.w.hu.h vr19, vr3, vr11
vmaddwod.w.hu.h vr21, vr4, vr12
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.d vr1, vr3, vr1
vadd.w vr22, vr22, vr1
addi.w s1, s1, 8
addi.d a3, a3, 16
addi.d a4, a4, 16
blt s1, s0, .LOOP_HS15_DST
blt s1, a6, .HS15_DSTA
b .END_HS15_FILTERA
.HS15_DSTA:
ld.w t2, a5, 0
li.w t3, 0
move s6, s1
.HS15_FILTERA:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s2, s2, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS15_FILTERA
ld.w t2, a5, 4
li.w t3, 0
move s6, s1
addi.w t1, t1, 1
.HS15_FILTERB:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s3, s3, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS15_FILTERB
ld.w t2, a5, 8
addi.w t1, t1, 1
li.w t3, 0
move s6, s1
.HS15_FILTERC:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s4, s4, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS15_FILTERC
ld.w t2, a5, 12
addi.w t1, t1, 1
move s6, s1
li.w t3, 0
.HS15_FILTERD:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s5, s5, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS15_FILTERD
.END_HS15_FILTERA:
vpickve2gr.w t1, vr22, 0
vpickve2gr.w t2, vr22, 1
vpickve2gr.w t3, vr22, 2
vpickve2gr.w t4, vr22, 3
add.w s2, s2, t1
add.w s3, s3, t2
add.w s4, s4, t3
add.w s5, s5, t4
sra.w s2, s2, a7
sra.w s3, s3, a7
sra.w s4, s4, a7
sra.w s5, s5, a7
slt t1, s2, t0
slt t2, s3, t0
slt t3, s4, t0
slt t4, s5, t0
maskeqz s2, s2, t1
maskeqz s3, s3, t2
maskeqz s4, s4, t3
maskeqz s5, s5, t4
masknez t1, t0, t1
masknez t2, t0, t2
masknez t3, t0, t3
masknez t4, t0, t4
or s2, s2, t1
or s3, s3, t2
or s4, s4, t3
or s5, s5, t4
st.h s2, a1, 0
st.h s3, a1, 2
st.h s4, a1, 4
st.h s5, a1, 6
addi.d a1, a1, 8
sub.d a3, a3, s1
sub.d a3, a3, s1
addi.d a5, a5, 16
slli.d t3, a6, 3
add.d a4, a4, t3
sub.d a4, a4, s1
sub.d a4, a4, s1
addi.d a2, a2, -4
bge a2, t7, .LOOP_HS15
blt zero, a2, .HS15_RESA
b .HS15_END
.HS15_RESA:
li.w t1, 0
.HS15_DST:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS15_FILTER:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS15_FILTER
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS15_DST
b .HS15_END
.LOOP_HS15_DST8:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
vldx vr1, a3, t1
vldx vr2, a3, t2
vldx vr3, a3, t3
vldx vr4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
vldx vr5, a3, t1
vldx vr6, a3, t2
vldx vr7, a3, t3
vldx vr8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vld vr13, a4, 64
vld vr14, a4, 80
vld vr15, a4, 96
vld vr16, a4, 112
vmulwev.w.hu.h vr17, vr1, vr9
vmulwev.w.hu.h vr18, vr2, vr10
vmulwev.w.hu.h vr19, vr3, vr11
vmulwev.w.hu.h vr21, vr4, vr12
vmaddwod.w.hu.h vr17, vr1, vr9
vmaddwod.w.hu.h vr18, vr2, vr10
vmaddwod.w.hu.h vr19, vr3, vr11
vmaddwod.w.hu.h vr21, vr4, vr12
vmulwev.w.hu.h vr1, vr5, vr13
vmulwev.w.hu.h vr2, vr6, vr14
vmulwev.w.hu.h vr3, vr7, vr15
vmulwev.w.hu.h vr4, vr8, vr16
vmaddwod.w.hu.h vr1, vr5, vr13
vmaddwod.w.hu.h vr2, vr6, vr14
vmaddwod.w.hu.h vr3, vr7, vr15
vmaddwod.w.hu.h vr4, vr8, vr16
vhaddw.d.w vr5, vr1, vr1
vhaddw.d.w vr6, vr2, vr2
vhaddw.d.w vr7, vr3, vr3
vhaddw.d.w vr8, vr4, vr4
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vhaddw.q.d vr5, vr5, vr5
vhaddw.q.d vr6, vr6, vr6
vhaddw.q.d vr7, vr7, vr7
vhaddw.q.d vr8, vr8, vr8
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.d vr1, vr3, vr1
vilvl.d vr5, vr7, vr5
vsra.w vr1, vr1, vr0
vsra.w vr5, vr5, vr0
vmin.w vr1, vr1, vr20
vmin.w vr5, vr5, vr20
vpickev.h vr1, vr5, vr1
vst vr1, a1, 0
addi.d a1, a1, 16
addi.d a5, a5, 32
addi.d a4, a4, 128
addi.d a2, a2, -8
bge a2, t8, .LOOP_HS15_DST8
blt zero, a2, .HS15_REST8
b .HS15_END
.HS15_REST8:
li.w t1, 0
.HS15_DST8:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS15_FILTER8:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS15_FILTER8
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS15_DST8
b .HS15_END
.LOOP_HS15_DST4:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
fldx.d f1, a3, t1
fldx.d f2, a3, t2
fldx.d f3, a3, t3
fldx.d f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
fldx.d f5, a3, t1
fldx.d f6, a3, t2
fldx.d f7, a3, t3
fldx.d f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vilvl.d vr1, vr2, vr1
vilvl.d vr3, vr4, vr3
vilvl.d vr5, vr6, vr5
vilvl.d vr7, vr8, vr7
vmulwev.w.hu.h vr13, vr1, vr9
vmulwev.w.hu.h vr14, vr3, vr10
vmulwev.w.hu.h vr15, vr5, vr11
vmulwev.w.hu.h vr16, vr7, vr12
vmaddwod.w.hu.h vr13, vr1, vr9
vmaddwod.w.hu.h vr14, vr3, vr10
vmaddwod.w.hu.h vr15, vr5, vr11
vmaddwod.w.hu.h vr16, vr7, vr12
vhaddw.d.w vr13, vr13, vr13
vhaddw.d.w vr14, vr14, vr14
vhaddw.d.w vr15, vr15, vr15
vhaddw.d.w vr16, vr16, vr16
vpickev.w vr13, vr14, vr13
vpickev.w vr15, vr16, vr15
vsra.w vr13, vr13, vr0
vsra.w vr15, vr15, vr0
vmin.w vr13, vr13, vr20
vmin.w vr15, vr15, vr20
vpickev.h vr13, vr15, vr13
vst vr13, a1, 0
addi.d a1, a1, 16
addi.d a5, a5, 32
addi.d a4, a4, 64
addi.d a2, a2, -8
bge a2, t8, .LOOP_HS15_DST4
blt zero, a2, .HS15_REST4
b .HS15_END
.HS15_REST4:
li.w t1, 0
.HS15_DST4:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS15_FILTER4:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS15_FILTER4
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS15_DST4
b .HS15_END
.END_HS15_DST4:
li.w t1, 0
.LOOP_HS15_DST1:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS15_FILTER1:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS15_FILTER1
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 1
stx.h t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .LOOP_HS15_DST1
b .HS15_END
.HS15_END:
ld.d s0, sp, 0
ld.d s1, sp, 8
ld.d s2, sp, 16
ld.d s3, sp, 24
ld.d s4, sp, 32
ld.d s5, sp, 40
ld.d s6, sp, 48
ld.d s7, sp, 56
ld.d s8, sp, 64
addi.d sp, sp, 72
endfunc
/* void ff_hscale_16_to_19_sub_lsx(SwsContext *c, int16_t *dst, int dstW,
* const uint8_t *src, const int16_t *filter,
* const int32_t *filterPos, int filterSize, int sh)
*/
function ff_hscale_16_to_19_sub_lsx
addi.d sp, sp, -72
st.d s0, sp, 0
st.d s1, sp, 8
st.d s2, sp, 16
st.d s3, sp, 24
st.d s4, sp, 32
st.d s5, sp, 40
st.d s6, sp, 48
st.d s7, sp, 56
st.d s8, sp, 64
li.w t0, 524287
li.w t8, 8
li.w t7, 4
vreplgr2vr.w vr20, t0
vreplgr2vr.w vr0, a7
beq a6, t7, .LOOP_HS19_DST4
beq a6, t8, .LOOP_HS19_DST8
blt t8, a6, .LOOP_HS19
b .END_HS19_DST4
.LOOP_HS19:
li.w t1, 0
li.w s1, 0
li.w s2, 0
li.w s3, 0
li.w s4, 0
li.w s5, 0
vldi vr22, 0
addi.w s0, a6, -7
slli.w s7, a6, 1
slli.w s8, a6, 2
add.w t6, s7, s8
.LOOP_HS19_DST:
ld.w t2, a5, 0
ld.w t3, a5, 4
ld.w t4, a5, 8
ld.w t5, a5, 12
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
slli.w t5, t5, 1
vldx vr1, a3, t2
vldx vr2, a3, t3
vldx vr3, a3, t4
vldx vr4, a3, t5
vld vr9, a4, 0
vldx vr10, a4, s7
vldx vr11, a4, s8
vldx vr12, a4, t6
vmulwev.w.hu.h vr17, vr1, vr9
vmulwev.w.hu.h vr18, vr2, vr10
vmulwev.w.hu.h vr19, vr3, vr11
vmulwev.w.hu.h vr21, vr4, vr12
vmaddwod.w.hu.h vr17, vr1, vr9
vmaddwod.w.hu.h vr18, vr2, vr10
vmaddwod.w.hu.h vr19, vr3, vr11
vmaddwod.w.hu.h vr21, vr4, vr12
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.d vr1, vr3, vr1
vadd.w vr22, vr22, vr1
addi.w s1, s1, 8
addi.d a3, a3, 16
addi.d a4, a4, 16
blt s1, s0, .LOOP_HS19_DST
blt s1, a6, .HS19_DSTA
b .END_HS19_FILTERA
.HS19_DSTA:
ld.w t2, a5, 0
li.w t3, 0
move s6, s1
.HS19_FILTERA:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s2, s2, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS19_FILTERA
ld.w t2, a5, 4
li.w t3, 0
move s6, s1
addi.w t1, t1, 1
.HS19_FILTERB:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s3, s3, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS19_FILTERB
ld.w t2, a5, 8
addi.w t1, t1, 1
li.w t3, 0
move s6, s1
.HS19_FILTERC:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s4, s4, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS19_FILTERC
ld.w t2, a5, 12
addi.w t1, t1, 1
move s6, s1
li.w t3, 0
.HS19_FILTERD:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t6, t6, 1
ldx.h t6, a4, t6
mul.w t6, t5, t6
add.w s5, s5, t6
addi.w t3, t3, 1
addi.w s6, s6, 1
blt s6, a6, .HS19_FILTERD
.END_HS19_FILTERA:
vpickve2gr.w t1, vr22, 0
vpickve2gr.w t2, vr22, 1
vpickve2gr.w t3, vr22, 2
vpickve2gr.w t4, vr22, 3
add.w s2, s2, t1
add.w s3, s3, t2
add.w s4, s4, t3
add.w s5, s5, t4
sra.w s2, s2, a7
sra.w s3, s3, a7
sra.w s4, s4, a7
sra.w s5, s5, a7
slt t1, s2, t0
slt t2, s3, t0
slt t3, s4, t0
slt t4, s5, t0
maskeqz s2, s2, t1
maskeqz s3, s3, t2
maskeqz s4, s4, t3
maskeqz s5, s5, t4
masknez t1, t0, t1
masknez t2, t0, t2
masknez t3, t0, t3
masknez t4, t0, t4
or s2, s2, t1
or s3, s3, t2
or s4, s4, t3
or s5, s5, t4
st.w s2, a1, 0
st.w s3, a1, 4
st.w s4, a1, 8
st.w s5, a1, 12
addi.d a1, a1, 16
sub.d a3, a3, s1
sub.d a3, a3, s1
addi.d a5, a5, 16
slli.d t3, a6, 3
add.d a4, a4, t3
sub.d a4, a4, s1
sub.d a4, a4, s1
addi.d a2, a2, -4
bge a2, t7, .LOOP_HS19
blt zero, a2, .HS19_RESA
b .HS19_END
.HS19_RESA:
li.w t1, 0
.HS19_DST:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS19_FILTER:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS19_FILTER
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS19_DST
b .HS19_END
.LOOP_HS19_DST8:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
vldx vr1, a3, t1
vldx vr2, a3, t2
vldx vr3, a3, t3
vldx vr4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
vldx vr5, a3, t1
vldx vr6, a3, t2
vldx vr7, a3, t3
vldx vr8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vld vr13, a4, 64
vld vr14, a4, 80
vld vr15, a4, 96
vld vr16, a4, 112
vmulwev.w.hu.h vr17, vr1, vr9
vmulwev.w.hu.h vr18, vr2, vr10
vmulwev.w.hu.h vr19, vr3, vr11
vmulwev.w.hu.h vr21, vr4, vr12
vmaddwod.w.hu.h vr17, vr1, vr9
vmaddwod.w.hu.h vr18, vr2, vr10
vmaddwod.w.hu.h vr19, vr3, vr11
vmaddwod.w.hu.h vr21, vr4, vr12
vmulwev.w.hu.h vr1, vr5, vr13
vmulwev.w.hu.h vr2, vr6, vr14
vmulwev.w.hu.h vr3, vr7, vr15
vmulwev.w.hu.h vr4, vr8, vr16
vmaddwod.w.hu.h vr1, vr5, vr13
vmaddwod.w.hu.h vr2, vr6, vr14
vmaddwod.w.hu.h vr3, vr7, vr15
vmaddwod.w.hu.h vr4, vr8, vr16
vhaddw.d.w vr5, vr1, vr1
vhaddw.d.w vr6, vr2, vr2
vhaddw.d.w vr7, vr3, vr3
vhaddw.d.w vr8, vr4, vr4
vhaddw.d.w vr1, vr17, vr17
vhaddw.d.w vr2, vr18, vr18
vhaddw.d.w vr3, vr19, vr19
vhaddw.d.w vr4, vr21, vr21
vhaddw.q.d vr1, vr1, vr1
vhaddw.q.d vr2, vr2, vr2
vhaddw.q.d vr3, vr3, vr3
vhaddw.q.d vr4, vr4, vr4
vhaddw.q.d vr5, vr5, vr5
vhaddw.q.d vr6, vr6, vr6
vhaddw.q.d vr7, vr7, vr7
vhaddw.q.d vr8, vr8, vr8
vilvl.w vr1, vr2, vr1
vilvl.w vr3, vr4, vr3
vilvl.w vr5, vr6, vr5
vilvl.w vr7, vr8, vr7
vilvl.d vr1, vr3, vr1
vilvl.d vr5, vr7, vr5
vsra.w vr1, vr1, vr0
vsra.w vr5, vr5, vr0
vmin.w vr1, vr1, vr20
vmin.w vr5, vr5, vr20
vst vr1, a1, 0
vst vr5, a1, 16
addi.d a1, a1, 32
addi.d a5, a5, 32
addi.d a4, a4, 128
addi.d a2, a2, -8
bge a2, t8, .LOOP_HS19_DST8
blt zero, a2, .HS19_REST8
b .HS19_END
.HS19_REST8:
li.w t1, 0
.HS19_DST8:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS19_FILTER8:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS19_FILTER8
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS19_DST8
b .HS19_END
.LOOP_HS19_DST4:
ld.w t1, a5, 0
ld.w t2, a5, 4
ld.w t3, a5, 8
ld.w t4, a5, 12
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
fldx.d f1, a3, t1
fldx.d f2, a3, t2
fldx.d f3, a3, t3
fldx.d f4, a3, t4
ld.w t1, a5, 16
ld.w t2, a5, 20
ld.w t3, a5, 24
ld.w t4, a5, 28
slli.w t1, t1, 1
slli.w t2, t2, 1
slli.w t3, t3, 1
slli.w t4, t4, 1
fldx.d f5, a3, t1
fldx.d f6, a3, t2
fldx.d f7, a3, t3
fldx.d f8, a3, t4
vld vr9, a4, 0
vld vr10, a4, 16
vld vr11, a4, 32
vld vr12, a4, 48
vilvl.d vr1, vr2, vr1
vilvl.d vr3, vr4, vr3
vilvl.d vr5, vr6, vr5
vilvl.d vr7, vr8, vr7
vmulwev.w.hu.h vr13, vr1, vr9
vmulwev.w.hu.h vr14, vr3, vr10
vmulwev.w.hu.h vr15, vr5, vr11
vmulwev.w.hu.h vr16, vr7, vr12
vmaddwod.w.hu.h vr13, vr1, vr9
vmaddwod.w.hu.h vr14, vr3, vr10
vmaddwod.w.hu.h vr15, vr5, vr11
vmaddwod.w.hu.h vr16, vr7, vr12
vhaddw.d.w vr13, vr13, vr13
vhaddw.d.w vr14, vr14, vr14
vhaddw.d.w vr15, vr15, vr15
vhaddw.d.w vr16, vr16, vr16
vpickev.w vr13, vr14, vr13
vpickev.w vr15, vr16, vr15
vsra.w vr13, vr13, vr0
vsra.w vr15, vr15, vr0
vmin.w vr13, vr13, vr20
vmin.w vr15, vr15, vr20
vst vr13, a1, 0
vst vr15, a1, 16
addi.d a1, a1, 32
addi.d a5, a5, 32
addi.d a4, a4, 64
addi.d a2, a2, -8
bge a2, t8, .LOOP_HS19_DST4
blt zero, a2, .HS19_REST4
b .HS19_END
.HS19_REST4:
li.w t1, 0
.HS19_DST4:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS19_FILTER4:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS19_FILTER4
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .HS19_DST4
b .HS19_END
.END_HS19_DST4:
li.w t1, 0
.LOOP_HS19_DST1:
slli.w t2, t1, 2
ldx.w t2, a5, t2
li.w t3, 0
li.w t8, 0
.HS19_FILTER1:
add.w t4, t2, t3
slli.w t4, t4, 1
ldx.hu t5, a3, t4
mul.w t6, a6, t1
add.w t6, t6, t3
slli.w t7, t6, 1
ldx.h t7, a4, t7
mul.w t7, t5, t7
add.w t8, t8, t7
addi.w t3, t3, 1
blt t3, a6, .HS19_FILTER1
sra.w t8, t8, a7
slt t5, t8, t0
maskeqz t8, t8, t5
masknez t5, t0, t5
or t8, t8, t5
slli.w t4, t1, 2
stx.w t8, a1, t4
addi.w t1, t1, 1
blt t1, a2, .LOOP_HS19_DST1
b .HS19_END
.HS19_END:
ld.d s0, sp, 0
ld.d s1, sp, 8
ld.d s2, sp, 16
ld.d s3, sp, 24
ld.d s4, sp, 32
ld.d s5, sp, 40
ld.d s6, sp, 48
ld.d s7, sp, 56
ld.d s8, sp, 64
addi.d sp, sp, 72
endfunc