mirror of https://github.com/FFmpeg/FFmpeg.git
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
726 lines
20 KiB
726 lines
20 KiB
@ |
|
@ ARMv4 optimized DSP utils |
|
@ Copyright (c) 2004 AGAWA Koji <i (AT) atty (DOT) jp> |
|
@ |
|
@ This file is part of FFmpeg. |
|
@ |
|
@ FFmpeg is free software; you can redistribute it and/or |
|
@ modify it under the terms of the GNU Lesser General Public |
|
@ License as published by the Free Software Foundation; either |
|
@ version 2.1 of the License, or (at your option) any later version. |
|
@ |
|
@ FFmpeg is distributed in the hope that it will be useful, |
|
@ but WITHOUT ANY WARRANTY; without even the implied warranty of |
|
@ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
|
@ Lesser General Public License for more details. |
|
@ |
|
@ You should have received a copy of the GNU Lesser General Public |
|
@ License along with FFmpeg; if not, write to the Free Software |
|
@ Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
|
@ |
|
|
|
#include "config.h" |
|
#include "asm.S" |
|
|
|
preserve8 |
|
|
|
#if !HAVE_PLD |
|
.macro pld reg |
|
.endm |
|
#endif |
|
|
|
#if HAVE_ARMV5TE |
|
function ff_prefetch_arm, export=1 |
|
subs r2, r2, #1 |
|
pld [r0] |
|
add r0, r0, r1 |
|
bne ff_prefetch_arm |
|
bx lr |
|
.endfunc |
|
#endif |
|
|
|
.macro ADJ_ALIGN_QUADWORD_D shift, Rd0, Rd1, Rd2, Rd3, Rn0, Rn1, Rn2, Rn3, Rn4 |
|
mov \Rd0, \Rn0, lsr #(\shift * 8) |
|
mov \Rd1, \Rn1, lsr #(\shift * 8) |
|
mov \Rd2, \Rn2, lsr #(\shift * 8) |
|
mov \Rd3, \Rn3, lsr #(\shift * 8) |
|
orr \Rd0, \Rd0, \Rn1, lsl #(32 - \shift * 8) |
|
orr \Rd1, \Rd1, \Rn2, lsl #(32 - \shift * 8) |
|
orr \Rd2, \Rd2, \Rn3, lsl #(32 - \shift * 8) |
|
orr \Rd3, \Rd3, \Rn4, lsl #(32 - \shift * 8) |
|
.endm |
|
.macro ADJ_ALIGN_DOUBLEWORD shift, R0, R1, R2 |
|
mov \R0, \R0, lsr #(\shift * 8) |
|
orr \R0, \R0, \R1, lsl #(32 - \shift * 8) |
|
mov \R1, \R1, lsr #(\shift * 8) |
|
orr \R1, \R1, \R2, lsl #(32 - \shift * 8) |
|
.endm |
|
.macro ADJ_ALIGN_DOUBLEWORD_D shift, Rdst0, Rdst1, Rsrc0, Rsrc1, Rsrc2 |
|
mov \Rdst0, \Rsrc0, lsr #(\shift * 8) |
|
mov \Rdst1, \Rsrc1, lsr #(\shift * 8) |
|
orr \Rdst0, \Rdst0, \Rsrc1, lsl #(32 - (\shift * 8)) |
|
orr \Rdst1, \Rdst1, \Rsrc2, lsl #(32 - (\shift * 8)) |
|
.endm |
|
|
|
.macro RND_AVG32 Rd0, Rd1, Rn0, Rn1, Rm0, Rm1, Rmask |
|
@ Rd = (Rn | Rm) - (((Rn ^ Rm) & ~0x01010101) >> 1) |
|
@ Rmask = 0xFEFEFEFE |
|
@ Rn = destroy |
|
eor \Rd0, \Rn0, \Rm0 |
|
eor \Rd1, \Rn1, \Rm1 |
|
orr \Rn0, \Rn0, \Rm0 |
|
orr \Rn1, \Rn1, \Rm1 |
|
and \Rd0, \Rd0, \Rmask |
|
and \Rd1, \Rd1, \Rmask |
|
sub \Rd0, \Rn0, \Rd0, lsr #1 |
|
sub \Rd1, \Rn1, \Rd1, lsr #1 |
|
.endm |
|
|
|
.macro NO_RND_AVG32 Rd0, Rd1, Rn0, Rn1, Rm0, Rm1, Rmask |
|
@ Rd = (Rn & Rm) - (((Rn ^ Rm) & ~0x01010101) >> 1) |
|
@ Rmask = 0xFEFEFEFE |
|
@ Rn = destroy |
|
eor \Rd0, \Rn0, \Rm0 |
|
eor \Rd1, \Rn1, \Rm1 |
|
and \Rn0, \Rn0, \Rm0 |
|
and \Rn1, \Rn1, \Rm1 |
|
and \Rd0, \Rd0, \Rmask |
|
and \Rd1, \Rd1, \Rmask |
|
add \Rd0, \Rn0, \Rd0, lsr #1 |
|
add \Rd1, \Rn1, \Rd1, lsr #1 |
|
.endm |
|
|
|
.macro JMP_ALIGN tmp, reg |
|
ands \tmp, \reg, #3 |
|
bic \reg, \reg, #3 |
|
beq 1f |
|
subs \tmp, \tmp, #1 |
|
beq 2f |
|
subs \tmp, \tmp, #1 |
|
beq 3f |
|
b 4f |
|
.endm |
|
|
|
@ ---------------------------------------------------------------- |
|
.align 5 |
|
function put_pixels16_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r11, lr} @ R14 is also called LR |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r7} |
|
add r1, r1, r2 |
|
stmia r0, {r4-r7} |
|
pld [r1] |
|
subs r3, r3, #1 |
|
add r0, r0, r2 |
|
bne 1b |
|
ldmfd sp!, {r4-r11, pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r8} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_QUADWORD_D 1, r9, r10, r11, r12, r4, r5, r6, r7, r8 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r9-r12} |
|
add r0, r0, r2 |
|
bne 2b |
|
ldmfd sp!, {r4-r11, pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r8} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_QUADWORD_D 2, r9, r10, r11, r12, r4, r5, r6, r7, r8 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r9-r12} |
|
add r0, r0, r2 |
|
bne 3b |
|
ldmfd sp!, {r4-r11, pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r8} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_QUADWORD_D 3, r9, r10, r11, r12, r4, r5, r6, r7, r8 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r9-r12} |
|
add r0, r0, r2 |
|
bne 4b |
|
ldmfd sp!, {r4-r11,pc} |
|
.endfunc |
|
|
|
@ ---------------------------------------------------------------- |
|
.align 5 |
|
function put_pixels8_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r5,lr} @ R14 is also called LR |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r5} |
|
add r1, r1, r2 |
|
subs r3, r3, #1 |
|
pld [r1] |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 1b |
|
ldmfd sp!, {r4-r5,pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r5, r12} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r12 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 2b |
|
ldmfd sp!, {r4-r5,pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r5, r12} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r12 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 3b |
|
ldmfd sp!, {r4-r5,pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r5, r12} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r12 |
|
pld [r1] |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 4b |
|
ldmfd sp!, {r4-r5,pc} |
|
.endfunc |
|
|
|
@ ---------------------------------------------------------------- |
|
.align 5 |
|
function put_pixels8_x2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r10,lr} @ R14 is also called LR |
|
ldr r12, =0xfefefefe |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 |
|
pld [r1] |
|
RND_AVG32 r8, r9, r4, r5, r6, r7, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 1b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r8, r9, r4, r5, r10 |
|
pld [r1] |
|
RND_AVG32 r4, r5, r6, r7, r8, r9, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 2b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r4, r5, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r8, r9, r4, r5, r10 |
|
pld [r1] |
|
RND_AVG32 r4, r5, r6, r7, r8, r9, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 3b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r4, r5, r10 |
|
pld [r1] |
|
RND_AVG32 r8, r9, r6, r7, r5, r10, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 4b |
|
ldmfd sp!, {r4-r10,pc} @@ update PC with LR content. |
|
.endfunc |
|
|
|
.align 5 |
|
function put_no_rnd_pixels8_x2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r10,lr} @ R14 is also called LR |
|
ldr r12, =0xfefefefe |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 |
|
pld [r1] |
|
NO_RND_AVG32 r8, r9, r4, r5, r6, r7, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 1b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r6, r7, r4, r5, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r8, r9, r4, r5, r10 |
|
pld [r1] |
|
NO_RND_AVG32 r4, r5, r6, r7, r8, r9, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 2b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r4, r5, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r8, r9, r4, r5, r10 |
|
pld [r1] |
|
NO_RND_AVG32 r4, r5, r6, r7, r8, r9, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bne 3b |
|
ldmfd sp!, {r4-r10,pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r5, r10} |
|
add r1, r1, r2 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r4, r5, r10 |
|
pld [r1] |
|
NO_RND_AVG32 r8, r9, r6, r7, r5, r10, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 4b |
|
ldmfd sp!, {r4-r10,pc} @@ update PC with LR content. |
|
.endfunc |
|
|
|
|
|
@ ---------------------------------------------------------------- |
|
.align 5 |
|
function put_pixels8_y2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r11,lr} @ R14 is also called LR |
|
mov r3, r3, lsr #1 |
|
ldr r12, =0xfefefefe |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r5} |
|
add r1, r1, r2 |
|
6: ldmia r1, {r6-r7} |
|
add r1, r1, r2 |
|
pld [r1] |
|
RND_AVG32 r8, r9, r4, r5, r6, r7, r12 |
|
ldmia r1, {r4-r5} |
|
add r1, r1, r2 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
pld [r1] |
|
RND_AVG32 r8, r9, r6, r7, r4, r5, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r7, r8, r9 |
|
RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 |
|
subs r3, r3, #1 |
|
RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r7, r8, r9 |
|
RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 |
|
subs r3, r3, #1 |
|
RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r7, r8, r9 |
|
RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 |
|
subs r3, r3, #1 |
|
RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.endfunc |
|
|
|
.align 5 |
|
function put_no_rnd_pixels8_y2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r11,lr} @ R14 is also called LR |
|
mov r3, r3, lsr #1 |
|
ldr r12, =0xfefefefe |
|
JMP_ALIGN r5, r1 |
|
1: |
|
ldmia r1, {r4-r5} |
|
add r1, r1, r2 |
|
6: ldmia r1, {r6-r7} |
|
add r1, r1, r2 |
|
pld [r1] |
|
NO_RND_AVG32 r8, r9, r4, r5, r6, r7, r12 |
|
ldmia r1, {r4-r5} |
|
add r1, r1, r2 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
pld [r1] |
|
NO_RND_AVG32 r8, r9, r6, r7, r4, r5, r12 |
|
subs r3, r3, #1 |
|
stmia r0, {r8-r9} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
2: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r7, r8, r9 |
|
NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 1, r4, r5, r6 |
|
subs r3, r3, #1 |
|
NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
3: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r7, r8, r9 |
|
NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 2, r4, r5, r6 |
|
subs r3, r3, #1 |
|
NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.align 5 |
|
4: |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 |
|
6: ldmia r1, {r7-r9} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r7, r8, r9 |
|
NO_RND_AVG32 r10, r11, r4, r5, r7, r8, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
ldmia r1, {r4-r6} |
|
add r1, r1, r2 |
|
pld [r1] |
|
ADJ_ALIGN_DOUBLEWORD 3, r4, r5, r6 |
|
subs r3, r3, #1 |
|
NO_RND_AVG32 r10, r11, r7, r8, r4, r5, r12 |
|
stmia r0, {r10-r11} |
|
add r0, r0, r2 |
|
bne 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.endfunc |
|
|
|
.ltorg |
|
|
|
@ ---------------------------------------------------------------- |
|
.macro RND_XY2_IT align, rnd |
|
@ l1= (a & 0x03030303) + (b & 0x03030303) ?(+ 0x02020202) |
|
@ h1= ((a & 0xFCFCFCFCUL) >> 2) + ((b & 0xFCFCFCFCUL) >> 2) |
|
.if \align == 0 |
|
ldmia r1, {r6-r8} |
|
.elseif \align == 3 |
|
ldmia r1, {r5-r7} |
|
.else |
|
ldmia r1, {r8-r10} |
|
.endif |
|
add r1, r1, r2 |
|
pld [r1] |
|
.if \align == 0 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r4, r5, r6, r7, r8 |
|
.elseif \align == 1 |
|
ADJ_ALIGN_DOUBLEWORD_D 1, r4, r5, r8, r9, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r6, r7, r8, r9, r10 |
|
.elseif \align == 2 |
|
ADJ_ALIGN_DOUBLEWORD_D 2, r4, r5, r8, r9, r10 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r6, r7, r8, r9, r10 |
|
.elseif \align == 3 |
|
ADJ_ALIGN_DOUBLEWORD_D 3, r4, r5, r5, r6, r7 |
|
.endif |
|
ldr r14, =0x03030303 |
|
tst r3, #1 |
|
and r8, r4, r14 |
|
and r9, r5, r14 |
|
and r10, r6, r14 |
|
and r11, r7, r14 |
|
andeq r14, r14, r14, \rnd #1 |
|
add r8, r8, r10 |
|
add r9, r9, r11 |
|
ldr r12, =0xfcfcfcfc >> 2 |
|
addeq r8, r8, r14 |
|
addeq r9, r9, r14 |
|
and r4, r12, r4, lsr #2 |
|
and r5, r12, r5, lsr #2 |
|
and r6, r12, r6, lsr #2 |
|
and r7, r12, r7, lsr #2 |
|
add r10, r4, r6 |
|
add r11, r5, r7 |
|
subs r3, r3, #1 |
|
.endm |
|
|
|
.macro RND_XY2_EXPAND align, rnd |
|
RND_XY2_IT \align, \rnd |
|
6: stmfd sp!, {r8-r11} |
|
RND_XY2_IT \align, \rnd |
|
ldmfd sp!, {r4-r7} |
|
add r4, r4, r8 |
|
add r5, r5, r9 |
|
ldr r14, =0x0f0f0f0f |
|
add r6, r6, r10 |
|
add r7, r7, r11 |
|
and r4, r14, r4, lsr #2 |
|
and r5, r14, r5, lsr #2 |
|
add r4, r4, r6 |
|
add r5, r5, r7 |
|
stmia r0, {r4-r5} |
|
add r0, r0, r2 |
|
bge 6b |
|
ldmfd sp!, {r4-r11,pc} |
|
.endm |
|
|
|
.align 5 |
|
function put_pixels8_xy2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r11,lr} @ R14 is also called LR |
|
JMP_ALIGN r5, r1 |
|
1: |
|
RND_XY2_EXPAND 0, lsl |
|
|
|
.align 5 |
|
2: |
|
RND_XY2_EXPAND 1, lsl |
|
|
|
.align 5 |
|
3: |
|
RND_XY2_EXPAND 2, lsl |
|
|
|
.align 5 |
|
4: |
|
RND_XY2_EXPAND 3, lsl |
|
.endfunc |
|
|
|
.align 5 |
|
function put_no_rnd_pixels8_xy2_arm, export=1 |
|
@ void func(uint8_t *block, const uint8_t *pixels, int line_size, int h) |
|
@ block = word aligned, pixles = unaligned |
|
pld [r1] |
|
stmfd sp!, {r4-r11,lr} @ R14 is also called LR |
|
JMP_ALIGN r5, r1 |
|
1: |
|
RND_XY2_EXPAND 0, lsr |
|
|
|
.align 5 |
|
2: |
|
RND_XY2_EXPAND 1, lsr |
|
|
|
.align 5 |
|
3: |
|
RND_XY2_EXPAND 2, lsr |
|
|
|
.align 5 |
|
4: |
|
RND_XY2_EXPAND 3, lsr |
|
.endfunc |
|
|
|
.align 5 |
|
@ void ff_add_pixels_clamped_ARM(int16_t *block, uint8_t *dest, int stride) |
|
function ff_add_pixels_clamped_ARM, export=1 |
|
push {r4-r10} |
|
mov r10, #8 |
|
1: |
|
ldr r4, [r1] /* load dest */ |
|
/* block[0] and block[1]*/ |
|
ldrsh r5, [r0] |
|
ldrsh r7, [r0, #2] |
|
and r6, r4, #0xFF |
|
and r8, r4, #0xFF00 |
|
add r6, r5, r6 |
|
add r8, r7, r8, lsr #8 |
|
mvn r5, r5 |
|
mvn r7, r7 |
|
tst r6, #0x100 |
|
movne r6, r5, lsr #24 |
|
tst r8, #0x100 |
|
movne r8, r7, lsr #24 |
|
mov r9, r6 |
|
ldrsh r5, [r0, #4] /* moved form [A] */ |
|
orr r9, r9, r8, lsl #8 |
|
/* block[2] and block[3] */ |
|
/* [A] */ |
|
ldrsh r7, [r0, #6] |
|
and r6, r4, #0xFF0000 |
|
and r8, r4, #0xFF000000 |
|
add r6, r5, r6, lsr #16 |
|
add r8, r7, r8, lsr #24 |
|
mvn r5, r5 |
|
mvn r7, r7 |
|
tst r6, #0x100 |
|
movne r6, r5, lsr #24 |
|
tst r8, #0x100 |
|
movne r8, r7, lsr #24 |
|
orr r9, r9, r6, lsl #16 |
|
ldr r4, [r1, #4] /* moved form [B] */ |
|
orr r9, r9, r8, lsl #24 |
|
/* store dest */ |
|
ldrsh r5, [r0, #8] /* moved form [C] */ |
|
str r9, [r1] |
|
|
|
/* load dest */ |
|
/* [B] */ |
|
/* block[4] and block[5] */ |
|
/* [C] */ |
|
ldrsh r7, [r0, #10] |
|
and r6, r4, #0xFF |
|
and r8, r4, #0xFF00 |
|
add r6, r5, r6 |
|
add r8, r7, r8, lsr #8 |
|
mvn r5, r5 |
|
mvn r7, r7 |
|
tst r6, #0x100 |
|
movne r6, r5, lsr #24 |
|
tst r8, #0x100 |
|
movne r8, r7, lsr #24 |
|
mov r9, r6 |
|
ldrsh r5, [r0, #12] /* moved from [D] */ |
|
orr r9, r9, r8, lsl #8 |
|
/* block[6] and block[7] */ |
|
/* [D] */ |
|
ldrsh r7, [r0, #14] |
|
and r6, r4, #0xFF0000 |
|
and r8, r4, #0xFF000000 |
|
add r6, r5, r6, lsr #16 |
|
add r8, r7, r8, lsr #24 |
|
mvn r5, r5 |
|
mvn r7, r7 |
|
tst r6, #0x100 |
|
movne r6, r5, lsr #24 |
|
tst r8, #0x100 |
|
movne r8, r7, lsr #24 |
|
orr r9, r9, r6, lsl #16 |
|
add r0, r0, #16 /* moved from [E] */ |
|
orr r9, r9, r8, lsl #24 |
|
subs r10, r10, #1 /* moved from [F] */ |
|
/* store dest */ |
|
str r9, [r1, #4] |
|
|
|
/* [E] */ |
|
/* [F] */ |
|
add r1, r1, r2 |
|
bne 1b |
|
|
|
pop {r4-r10} |
|
bx lr |
|
.endfunc
|
|
|