mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-21 10:55:51 +02:00
e54e6f25cf
Change the size specifiers to match the actual element sizes of the data. This makes no practical difference with strict alignment checking disabled (the default) other than somewhat documenting the code. With strict alignment checking on, it avoids trapping the unaligned loads. Signed-off-by: Mans Rullgard <mans@mansr.com>
912 lines
30 KiB
ArmAsm
912 lines
30 KiB
ArmAsm
/*
|
|
* ARM NEON optimised DSP functions
|
|
* Copyright (c) 2008 Mans Rullgard <mans@mansr.com>
|
|
*
|
|
* This file is part of Libav.
|
|
*
|
|
* Libav is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* Libav is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with Libav; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "config.h"
|
|
#include "asm.S"
|
|
|
|
preserve8
|
|
|
|
function ff_clear_block_neon, export=1
|
|
vmov.i16 q0, #0
|
|
.rept 8
|
|
vst1.16 {q0}, [r0,:128]!
|
|
.endr
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_clear_blocks_neon, export=1
|
|
vmov.i16 q0, #0
|
|
.rept 8*6
|
|
vst1.16 {q0}, [r0,:128]!
|
|
.endr
|
|
bx lr
|
|
endfunc
|
|
|
|
.macro pixels16 rnd=1, avg=0
|
|
.if \avg
|
|
mov r12, r0
|
|
.endif
|
|
1: vld1.8 {q0}, [r1], r2
|
|
vld1.8 {q1}, [r1], r2
|
|
vld1.8 {q2}, [r1], r2
|
|
pld [r1, r2, lsl #2]
|
|
vld1.8 {q3}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
pld [r1, r2, lsl #1]
|
|
.if \avg
|
|
vld1.8 {q8}, [r12,:128], r2
|
|
vrhadd.u8 q0, q0, q8
|
|
vld1.8 {q9}, [r12,:128], r2
|
|
vrhadd.u8 q1, q1, q9
|
|
vld1.8 {q10}, [r12,:128], r2
|
|
vrhadd.u8 q2, q2, q10
|
|
vld1.8 {q11}, [r12,:128], r2
|
|
vrhadd.u8 q3, q3, q11
|
|
.endif
|
|
subs r3, r3, #4
|
|
vst1.64 {q0}, [r0,:128], r2
|
|
vst1.64 {q1}, [r0,:128], r2
|
|
vst1.64 {q2}, [r0,:128], r2
|
|
vst1.64 {q3}, [r0,:128], r2
|
|
bne 1b
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels16_x2 rnd=1, avg=0
|
|
1: vld1.8 {d0-d2}, [r1], r2
|
|
vld1.8 {d4-d6}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
subs r3, r3, #2
|
|
vext.8 q1, q0, q1, #1
|
|
avg q0, q0, q1
|
|
vext.8 q3, q2, q3, #1
|
|
avg q2, q2, q3
|
|
.if \avg
|
|
vld1.8 {q1}, [r0,:128], r2
|
|
vld1.8 {q3}, [r0,:128]
|
|
vrhadd.u8 q0, q0, q1
|
|
vrhadd.u8 q2, q2, q3
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {q0}, [r0,:128], r2
|
|
vst1.8 {q2}, [r0,:128], r2
|
|
bne 1b
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels16_y2 rnd=1, avg=0
|
|
sub r3, r3, #2
|
|
vld1.8 {q0}, [r1], r2
|
|
vld1.8 {q1}, [r1], r2
|
|
1: subs r3, r3, #2
|
|
avg q2, q0, q1
|
|
vld1.8 {q0}, [r1], r2
|
|
avg q3, q0, q1
|
|
vld1.8 {q1}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
.if \avg
|
|
vld1.8 {q8}, [r0,:128], r2
|
|
vld1.8 {q9}, [r0,:128]
|
|
vrhadd.u8 q2, q2, q8
|
|
vrhadd.u8 q3, q3, q9
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {q2}, [r0,:128], r2
|
|
vst1.8 {q3}, [r0,:128], r2
|
|
bne 1b
|
|
|
|
avg q2, q0, q1
|
|
vld1.8 {q0}, [r1], r2
|
|
avg q3, q0, q1
|
|
.if \avg
|
|
vld1.8 {q8}, [r0,:128], r2
|
|
vld1.8 {q9}, [r0,:128]
|
|
vrhadd.u8 q2, q2, q8
|
|
vrhadd.u8 q3, q3, q9
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {q2}, [r0,:128], r2
|
|
vst1.8 {q3}, [r0,:128], r2
|
|
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels16_xy2 rnd=1, avg=0
|
|
sub r3, r3, #2
|
|
vld1.8 {d0-d2}, [r1], r2
|
|
vld1.8 {d4-d6}, [r1], r2
|
|
NRND vmov.i16 q13, #1
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vext.8 q1, q0, q1, #1
|
|
vext.8 q3, q2, q3, #1
|
|
vaddl.u8 q8, d0, d2
|
|
vaddl.u8 q10, d1, d3
|
|
vaddl.u8 q9, d4, d6
|
|
vaddl.u8 q11, d5, d7
|
|
1: subs r3, r3, #2
|
|
vld1.8 {d0-d2}, [r1], r2
|
|
vadd.u16 q12, q8, q9
|
|
pld [r1]
|
|
NRND vadd.u16 q12, q12, q13
|
|
vext.8 q15, q0, q1, #1
|
|
vadd.u16 q1 , q10, q11
|
|
shrn d28, q12, #2
|
|
NRND vadd.u16 q1, q1, q13
|
|
shrn d29, q1, #2
|
|
.if \avg
|
|
vld1.8 {q8}, [r0,:128]
|
|
vrhadd.u8 q14, q14, q8
|
|
.endif
|
|
vaddl.u8 q8, d0, d30
|
|
vld1.8 {d2-d4}, [r1], r2
|
|
vaddl.u8 q10, d1, d31
|
|
vst1.8 {q14}, [r0,:128], r2
|
|
vadd.u16 q12, q8, q9
|
|
pld [r1, r2]
|
|
NRND vadd.u16 q12, q12, q13
|
|
vext.8 q2, q1, q2, #1
|
|
vadd.u16 q0, q10, q11
|
|
shrn d30, q12, #2
|
|
NRND vadd.u16 q0, q0, q13
|
|
shrn d31, q0, #2
|
|
.if \avg
|
|
vld1.8 {q9}, [r0,:128]
|
|
vrhadd.u8 q15, q15, q9
|
|
.endif
|
|
vaddl.u8 q9, d2, d4
|
|
vaddl.u8 q11, d3, d5
|
|
vst1.8 {q15}, [r0,:128], r2
|
|
bgt 1b
|
|
|
|
vld1.8 {d0-d2}, [r1], r2
|
|
vadd.u16 q12, q8, q9
|
|
NRND vadd.u16 q12, q12, q13
|
|
vext.8 q15, q0, q1, #1
|
|
vadd.u16 q1 , q10, q11
|
|
shrn d28, q12, #2
|
|
NRND vadd.u16 q1, q1, q13
|
|
shrn d29, q1, #2
|
|
.if \avg
|
|
vld1.8 {q8}, [r0,:128]
|
|
vrhadd.u8 q14, q14, q8
|
|
.endif
|
|
vaddl.u8 q8, d0, d30
|
|
vaddl.u8 q10, d1, d31
|
|
vst1.8 {q14}, [r0,:128], r2
|
|
vadd.u16 q12, q8, q9
|
|
NRND vadd.u16 q12, q12, q13
|
|
vadd.u16 q0, q10, q11
|
|
shrn d30, q12, #2
|
|
NRND vadd.u16 q0, q0, q13
|
|
shrn d31, q0, #2
|
|
.if \avg
|
|
vld1.8 {q9}, [r0,:128]
|
|
vrhadd.u8 q15, q15, q9
|
|
.endif
|
|
vst1.8 {q15}, [r0,:128], r2
|
|
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels8 rnd=1, avg=0
|
|
1: vld1.8 {d0}, [r1], r2
|
|
vld1.8 {d1}, [r1], r2
|
|
vld1.8 {d2}, [r1], r2
|
|
pld [r1, r2, lsl #2]
|
|
vld1.8 {d3}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
pld [r1, r2, lsl #1]
|
|
.if \avg
|
|
vld1.8 {d4}, [r0,:64], r2
|
|
vrhadd.u8 d0, d0, d4
|
|
vld1.8 {d5}, [r0,:64], r2
|
|
vrhadd.u8 d1, d1, d5
|
|
vld1.8 {d6}, [r0,:64], r2
|
|
vrhadd.u8 d2, d2, d6
|
|
vld1.8 {d7}, [r0,:64], r2
|
|
vrhadd.u8 d3, d3, d7
|
|
sub r0, r0, r2, lsl #2
|
|
.endif
|
|
subs r3, r3, #4
|
|
vst1.8 {d0}, [r0,:64], r2
|
|
vst1.8 {d1}, [r0,:64], r2
|
|
vst1.8 {d2}, [r0,:64], r2
|
|
vst1.8 {d3}, [r0,:64], r2
|
|
bne 1b
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels8_x2 rnd=1, avg=0
|
|
1: vld1.8 {q0}, [r1], r2
|
|
vext.8 d1, d0, d1, #1
|
|
vld1.8 {q1}, [r1], r2
|
|
vext.8 d3, d2, d3, #1
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
subs r3, r3, #2
|
|
vswp d1, d2
|
|
avg q0, q0, q1
|
|
.if \avg
|
|
vld1.8 {d4}, [r0,:64], r2
|
|
vld1.8 {d5}, [r0,:64]
|
|
vrhadd.u8 q0, q0, q2
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {d0}, [r0,:64], r2
|
|
vst1.8 {d1}, [r0,:64], r2
|
|
bne 1b
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels8_y2 rnd=1, avg=0
|
|
sub r3, r3, #2
|
|
vld1.8 {d0}, [r1], r2
|
|
vld1.8 {d1}, [r1], r2
|
|
1: subs r3, r3, #2
|
|
avg d4, d0, d1
|
|
vld1.8 {d0}, [r1], r2
|
|
avg d5, d0, d1
|
|
vld1.8 {d1}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
.if \avg
|
|
vld1.8 {d2}, [r0,:64], r2
|
|
vld1.8 {d3}, [r0,:64]
|
|
vrhadd.u8 q2, q2, q1
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {d4}, [r0,:64], r2
|
|
vst1.8 {d5}, [r0,:64], r2
|
|
bne 1b
|
|
|
|
avg d4, d0, d1
|
|
vld1.8 {d0}, [r1], r2
|
|
avg d5, d0, d1
|
|
.if \avg
|
|
vld1.8 {d2}, [r0,:64], r2
|
|
vld1.8 {d3}, [r0,:64]
|
|
vrhadd.u8 q2, q2, q1
|
|
sub r0, r0, r2
|
|
.endif
|
|
vst1.8 {d4}, [r0,:64], r2
|
|
vst1.8 {d5}, [r0,:64], r2
|
|
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixels8_xy2 rnd=1, avg=0
|
|
sub r3, r3, #2
|
|
vld1.8 {q0}, [r1], r2
|
|
vld1.8 {q1}, [r1], r2
|
|
NRND vmov.i16 q11, #1
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vext.8 d4, d0, d1, #1
|
|
vext.8 d6, d2, d3, #1
|
|
vaddl.u8 q8, d0, d4
|
|
vaddl.u8 q9, d2, d6
|
|
1: subs r3, r3, #2
|
|
vld1.8 {q0}, [r1], r2
|
|
pld [r1]
|
|
vadd.u16 q10, q8, q9
|
|
vext.8 d4, d0, d1, #1
|
|
NRND vadd.u16 q10, q10, q11
|
|
vaddl.u8 q8, d0, d4
|
|
shrn d5, q10, #2
|
|
vld1.8 {q1}, [r1], r2
|
|
vadd.u16 q10, q8, q9
|
|
pld [r1, r2]
|
|
.if \avg
|
|
vld1.8 {d7}, [r0,:64]
|
|
vrhadd.u8 d5, d5, d7
|
|
.endif
|
|
NRND vadd.u16 q10, q10, q11
|
|
vst1.8 {d5}, [r0,:64], r2
|
|
shrn d7, q10, #2
|
|
.if \avg
|
|
vld1.8 {d5}, [r0,:64]
|
|
vrhadd.u8 d7, d7, d5
|
|
.endif
|
|
vext.8 d6, d2, d3, #1
|
|
vaddl.u8 q9, d2, d6
|
|
vst1.8 {d7}, [r0,:64], r2
|
|
bgt 1b
|
|
|
|
vld1.8 {q0}, [r1], r2
|
|
vadd.u16 q10, q8, q9
|
|
vext.8 d4, d0, d1, #1
|
|
NRND vadd.u16 q10, q10, q11
|
|
vaddl.u8 q8, d0, d4
|
|
shrn d5, q10, #2
|
|
vadd.u16 q10, q8, q9
|
|
.if \avg
|
|
vld1.8 {d7}, [r0,:64]
|
|
vrhadd.u8 d5, d5, d7
|
|
.endif
|
|
NRND vadd.u16 q10, q10, q11
|
|
vst1.8 {d5}, [r0,:64], r2
|
|
shrn d7, q10, #2
|
|
.if \avg
|
|
vld1.8 {d5}, [r0,:64]
|
|
vrhadd.u8 d7, d7, d5
|
|
.endif
|
|
vst1.8 {d7}, [r0,:64], r2
|
|
|
|
bx lr
|
|
.endm
|
|
|
|
.macro pixfunc pfx, name, suf, rnd=1, avg=0
|
|
.if \rnd
|
|
.macro avg rd, rn, rm
|
|
vrhadd.u8 \rd, \rn, \rm
|
|
.endm
|
|
.macro shrn rd, rn, rm
|
|
vrshrn.u16 \rd, \rn, \rm
|
|
.endm
|
|
.macro NRND insn:vararg
|
|
.endm
|
|
.else
|
|
.macro avg rd, rn, rm
|
|
vhadd.u8 \rd, \rn, \rm
|
|
.endm
|
|
.macro shrn rd, rn, rm
|
|
vshrn.u16 \rd, \rn, \rm
|
|
.endm
|
|
.macro NRND insn:vararg
|
|
\insn
|
|
.endm
|
|
.endif
|
|
function ff_\pfx\name\suf\()_neon, export=1
|
|
\name \rnd, \avg
|
|
endfunc
|
|
.purgem avg
|
|
.purgem shrn
|
|
.purgem NRND
|
|
.endm
|
|
|
|
.macro pixfunc2 pfx, name, avg=0
|
|
pixfunc \pfx, \name, rnd=1, avg=\avg
|
|
pixfunc \pfx, \name, _no_rnd, rnd=0, avg=\avg
|
|
.endm
|
|
|
|
function ff_put_h264_qpel16_mc00_neon, export=1
|
|
mov r3, #16
|
|
endfunc
|
|
|
|
pixfunc put_, pixels16, avg=0
|
|
pixfunc2 put_, pixels16_x2, avg=0
|
|
pixfunc2 put_, pixels16_y2, avg=0
|
|
pixfunc2 put_, pixels16_xy2, avg=0
|
|
|
|
function ff_avg_h264_qpel16_mc00_neon, export=1
|
|
mov r3, #16
|
|
endfunc
|
|
|
|
pixfunc avg_, pixels16, avg=1
|
|
pixfunc2 avg_, pixels16_x2, avg=1
|
|
pixfunc2 avg_, pixels16_y2, avg=1
|
|
pixfunc2 avg_, pixels16_xy2, avg=1
|
|
|
|
function ff_put_h264_qpel8_mc00_neon, export=1
|
|
mov r3, #8
|
|
endfunc
|
|
|
|
pixfunc put_, pixels8, avg=0
|
|
pixfunc2 put_, pixels8_x2, avg=0
|
|
pixfunc2 put_, pixels8_y2, avg=0
|
|
pixfunc2 put_, pixels8_xy2, avg=0
|
|
|
|
function ff_avg_h264_qpel8_mc00_neon, export=1
|
|
mov r3, #8
|
|
endfunc
|
|
|
|
pixfunc avg_, pixels8, avg=1
|
|
pixfunc2 avg_, pixels8_x2, avg=1
|
|
pixfunc2 avg_, pixels8_y2, avg=1
|
|
pixfunc2 avg_, pixels8_xy2, avg=1
|
|
|
|
function ff_put_pixels_clamped_neon, export=1
|
|
vld1.16 {d16-d19}, [r0,:128]!
|
|
vqmovun.s16 d0, q8
|
|
vld1.16 {d20-d23}, [r0,:128]!
|
|
vqmovun.s16 d1, q9
|
|
vld1.16 {d24-d27}, [r0,:128]!
|
|
vqmovun.s16 d2, q10
|
|
vld1.16 {d28-d31}, [r0,:128]!
|
|
vqmovun.s16 d3, q11
|
|
vst1.8 {d0}, [r1,:64], r2
|
|
vqmovun.s16 d4, q12
|
|
vst1.8 {d1}, [r1,:64], r2
|
|
vqmovun.s16 d5, q13
|
|
vst1.8 {d2}, [r1,:64], r2
|
|
vqmovun.s16 d6, q14
|
|
vst1.8 {d3}, [r1,:64], r2
|
|
vqmovun.s16 d7, q15
|
|
vst1.8 {d4}, [r1,:64], r2
|
|
vst1.8 {d5}, [r1,:64], r2
|
|
vst1.8 {d6}, [r1,:64], r2
|
|
vst1.8 {d7}, [r1,:64], r2
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_put_signed_pixels_clamped_neon, export=1
|
|
vmov.u8 d31, #128
|
|
vld1.16 {d16-d17}, [r0,:128]!
|
|
vqmovn.s16 d0, q8
|
|
vld1.16 {d18-d19}, [r0,:128]!
|
|
vqmovn.s16 d1, q9
|
|
vld1.16 {d16-d17}, [r0,:128]!
|
|
vqmovn.s16 d2, q8
|
|
vld1.16 {d18-d19}, [r0,:128]!
|
|
vadd.u8 d0, d0, d31
|
|
vld1.16 {d20-d21}, [r0,:128]!
|
|
vadd.u8 d1, d1, d31
|
|
vld1.16 {d22-d23}, [r0,:128]!
|
|
vadd.u8 d2, d2, d31
|
|
vst1.8 {d0}, [r1,:64], r2
|
|
vqmovn.s16 d3, q9
|
|
vst1.8 {d1}, [r1,:64], r2
|
|
vqmovn.s16 d4, q10
|
|
vst1.8 {d2}, [r1,:64], r2
|
|
vqmovn.s16 d5, q11
|
|
vld1.16 {d24-d25}, [r0,:128]!
|
|
vadd.u8 d3, d3, d31
|
|
vld1.16 {d26-d27}, [r0,:128]!
|
|
vadd.u8 d4, d4, d31
|
|
vadd.u8 d5, d5, d31
|
|
vst1.8 {d3}, [r1,:64], r2
|
|
vqmovn.s16 d6, q12
|
|
vst1.8 {d4}, [r1,:64], r2
|
|
vqmovn.s16 d7, q13
|
|
vst1.8 {d5}, [r1,:64], r2
|
|
vadd.u8 d6, d6, d31
|
|
vadd.u8 d7, d7, d31
|
|
vst1.8 {d6}, [r1,:64], r2
|
|
vst1.8 {d7}, [r1,:64], r2
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_add_pixels_clamped_neon, export=1
|
|
mov r3, r1
|
|
vld1.8 {d16}, [r1,:64], r2
|
|
vld1.16 {d0-d1}, [r0,:128]!
|
|
vaddw.u8 q0, q0, d16
|
|
vld1.8 {d17}, [r1,:64], r2
|
|
vld1.16 {d2-d3}, [r0,:128]!
|
|
vqmovun.s16 d0, q0
|
|
vld1.8 {d18}, [r1,:64], r2
|
|
vaddw.u8 q1, q1, d17
|
|
vld1.16 {d4-d5}, [r0,:128]!
|
|
vaddw.u8 q2, q2, d18
|
|
vst1.8 {d0}, [r3,:64], r2
|
|
vqmovun.s16 d2, q1
|
|
vld1.8 {d19}, [r1,:64], r2
|
|
vld1.16 {d6-d7}, [r0,:128]!
|
|
vaddw.u8 q3, q3, d19
|
|
vqmovun.s16 d4, q2
|
|
vst1.8 {d2}, [r3,:64], r2
|
|
vld1.8 {d16}, [r1,:64], r2
|
|
vqmovun.s16 d6, q3
|
|
vld1.16 {d0-d1}, [r0,:128]!
|
|
vaddw.u8 q0, q0, d16
|
|
vst1.8 {d4}, [r3,:64], r2
|
|
vld1.8 {d17}, [r1,:64], r2
|
|
vld1.16 {d2-d3}, [r0,:128]!
|
|
vaddw.u8 q1, q1, d17
|
|
vst1.8 {d6}, [r3,:64], r2
|
|
vqmovun.s16 d0, q0
|
|
vld1.8 {d18}, [r1,:64], r2
|
|
vld1.16 {d4-d5}, [r0,:128]!
|
|
vaddw.u8 q2, q2, d18
|
|
vst1.8 {d0}, [r3,:64], r2
|
|
vqmovun.s16 d2, q1
|
|
vld1.8 {d19}, [r1,:64], r2
|
|
vqmovun.s16 d4, q2
|
|
vld1.16 {d6-d7}, [r0,:128]!
|
|
vaddw.u8 q3, q3, d19
|
|
vst1.8 {d2}, [r3,:64], r2
|
|
vqmovun.s16 d6, q3
|
|
vst1.8 {d4}, [r3,:64], r2
|
|
vst1.8 {d6}, [r3,:64], r2
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_vector_fmul_neon, export=1
|
|
subs r3, r3, #8
|
|
vld1.32 {d0-d3}, [r1,:128]!
|
|
vld1.32 {d4-d7}, [r2,:128]!
|
|
vmul.f32 q8, q0, q2
|
|
vmul.f32 q9, q1, q3
|
|
beq 3f
|
|
bics ip, r3, #15
|
|
beq 2f
|
|
1: subs ip, ip, #16
|
|
vld1.32 {d0-d1}, [r1,:128]!
|
|
vld1.32 {d4-d5}, [r2,:128]!
|
|
vmul.f32 q10, q0, q2
|
|
vld1.32 {d2-d3}, [r1,:128]!
|
|
vld1.32 {d6-d7}, [r2,:128]!
|
|
vmul.f32 q11, q1, q3
|
|
vst1.32 {d16-d19},[r0,:128]!
|
|
vld1.32 {d0-d1}, [r1,:128]!
|
|
vld1.32 {d4-d5}, [r2,:128]!
|
|
vmul.f32 q8, q0, q2
|
|
vld1.32 {d2-d3}, [r1,:128]!
|
|
vld1.32 {d6-d7}, [r2,:128]!
|
|
vmul.f32 q9, q1, q3
|
|
vst1.32 {d20-d23},[r0,:128]!
|
|
bne 1b
|
|
ands r3, r3, #15
|
|
beq 3f
|
|
2: vld1.32 {d0-d1}, [r1,:128]!
|
|
vld1.32 {d4-d5}, [r2,:128]!
|
|
vst1.32 {d16-d17},[r0,:128]!
|
|
vmul.f32 q8, q0, q2
|
|
vld1.32 {d2-d3}, [r1,:128]!
|
|
vld1.32 {d6-d7}, [r2,:128]!
|
|
vst1.32 {d18-d19},[r0,:128]!
|
|
vmul.f32 q9, q1, q3
|
|
3: vst1.32 {d16-d19},[r0,:128]!
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_vector_fmul_window_neon, export=1
|
|
push {r4,r5,lr}
|
|
ldr lr, [sp, #12]
|
|
sub r2, r2, #8
|
|
sub r5, lr, #2
|
|
add r2, r2, r5, lsl #2
|
|
add r4, r3, r5, lsl #3
|
|
add ip, r0, r5, lsl #3
|
|
mov r5, #-16
|
|
vld1.32 {d0,d1}, [r1,:128]!
|
|
vld1.32 {d2,d3}, [r2,:128], r5
|
|
vld1.32 {d4,d5}, [r3,:128]!
|
|
vld1.32 {d6,d7}, [r4,:128], r5
|
|
1: subs lr, lr, #4
|
|
vmul.f32 d22, d0, d4
|
|
vrev64.32 q3, q3
|
|
vmul.f32 d23, d1, d5
|
|
vrev64.32 q1, q1
|
|
vmul.f32 d20, d0, d7
|
|
vmul.f32 d21, d1, d6
|
|
beq 2f
|
|
vmla.f32 d22, d3, d7
|
|
vld1.32 {d0,d1}, [r1,:128]!
|
|
vmla.f32 d23, d2, d6
|
|
vld1.32 {d18,d19},[r2,:128], r5
|
|
vmls.f32 d20, d3, d4
|
|
vld1.32 {d24,d25},[r3,:128]!
|
|
vmls.f32 d21, d2, d5
|
|
vld1.32 {d6,d7}, [r4,:128], r5
|
|
vmov q1, q9
|
|
vrev64.32 q11, q11
|
|
vmov q2, q12
|
|
vswp d22, d23
|
|
vst1.32 {d20,d21},[r0,:128]!
|
|
vst1.32 {d22,d23},[ip,:128], r5
|
|
b 1b
|
|
2: vmla.f32 d22, d3, d7
|
|
vmla.f32 d23, d2, d6
|
|
vmls.f32 d20, d3, d4
|
|
vmls.f32 d21, d2, d5
|
|
vrev64.32 q11, q11
|
|
vswp d22, d23
|
|
vst1.32 {d20,d21},[r0,:128]!
|
|
vst1.32 {d22,d23},[ip,:128], r5
|
|
pop {r4,r5,pc}
|
|
endfunc
|
|
|
|
#if CONFIG_VORBIS_DECODER
|
|
function ff_vorbis_inverse_coupling_neon, export=1
|
|
vmov.i32 q10, #1<<31
|
|
subs r2, r2, #4
|
|
mov r3, r0
|
|
mov r12, r1
|
|
beq 3f
|
|
|
|
vld1.32 {d24-d25},[r1,:128]!
|
|
vld1.32 {d22-d23},[r0,:128]!
|
|
vcle.s32 q8, q12, #0
|
|
vand q9, q11, q10
|
|
veor q12, q12, q9
|
|
vand q2, q12, q8
|
|
vbic q3, q12, q8
|
|
vadd.f32 q12, q11, q2
|
|
vsub.f32 q11, q11, q3
|
|
1: vld1.32 {d2-d3}, [r1,:128]!
|
|
vld1.32 {d0-d1}, [r0,:128]!
|
|
vcle.s32 q8, q1, #0
|
|
vand q9, q0, q10
|
|
veor q1, q1, q9
|
|
vst1.32 {d24-d25},[r3, :128]!
|
|
vst1.32 {d22-d23},[r12,:128]!
|
|
vand q2, q1, q8
|
|
vbic q3, q1, q8
|
|
vadd.f32 q1, q0, q2
|
|
vsub.f32 q0, q0, q3
|
|
subs r2, r2, #8
|
|
ble 2f
|
|
vld1.32 {d24-d25},[r1,:128]!
|
|
vld1.32 {d22-d23},[r0,:128]!
|
|
vcle.s32 q8, q12, #0
|
|
vand q9, q11, q10
|
|
veor q12, q12, q9
|
|
vst1.32 {d2-d3}, [r3, :128]!
|
|
vst1.32 {d0-d1}, [r12,:128]!
|
|
vand q2, q12, q8
|
|
vbic q3, q12, q8
|
|
vadd.f32 q12, q11, q2
|
|
vsub.f32 q11, q11, q3
|
|
b 1b
|
|
|
|
2: vst1.32 {d2-d3}, [r3, :128]!
|
|
vst1.32 {d0-d1}, [r12,:128]!
|
|
it lt
|
|
bxlt lr
|
|
|
|
3: vld1.32 {d2-d3}, [r1,:128]
|
|
vld1.32 {d0-d1}, [r0,:128]
|
|
vcle.s32 q8, q1, #0
|
|
vand q9, q0, q10
|
|
veor q1, q1, q9
|
|
vand q2, q1, q8
|
|
vbic q3, q1, q8
|
|
vadd.f32 q1, q0, q2
|
|
vsub.f32 q0, q0, q3
|
|
vst1.32 {d2-d3}, [r0,:128]!
|
|
vst1.32 {d0-d1}, [r1,:128]!
|
|
bx lr
|
|
endfunc
|
|
#endif
|
|
|
|
function ff_vector_fmul_scalar_neon, export=1
|
|
VFP len .req r2
|
|
NOVFP len .req r3
|
|
VFP vdup.32 q8, d0[0]
|
|
NOVFP vdup.32 q8, r2
|
|
bics r12, len, #15
|
|
beq 3f
|
|
vld1.32 {q0},[r1,:128]!
|
|
vld1.32 {q1},[r1,:128]!
|
|
1: vmul.f32 q0, q0, q8
|
|
vld1.32 {q2},[r1,:128]!
|
|
vmul.f32 q1, q1, q8
|
|
vld1.32 {q3},[r1,:128]!
|
|
vmul.f32 q2, q2, q8
|
|
vst1.32 {q0},[r0,:128]!
|
|
vmul.f32 q3, q3, q8
|
|
vst1.32 {q1},[r0,:128]!
|
|
subs r12, r12, #16
|
|
beq 2f
|
|
vld1.32 {q0},[r1,:128]!
|
|
vst1.32 {q2},[r0,:128]!
|
|
vld1.32 {q1},[r1,:128]!
|
|
vst1.32 {q3},[r0,:128]!
|
|
b 1b
|
|
2: vst1.32 {q2},[r0,:128]!
|
|
vst1.32 {q3},[r0,:128]!
|
|
ands len, len, #15
|
|
it eq
|
|
bxeq lr
|
|
3: vld1.32 {q0},[r1,:128]!
|
|
vmul.f32 q0, q0, q8
|
|
vst1.32 {q0},[r0,:128]!
|
|
subs len, len, #4
|
|
bgt 3b
|
|
bx lr
|
|
.unreq len
|
|
endfunc
|
|
|
|
function ff_vector_fmac_scalar_neon, export=1
|
|
VFP len .req r2
|
|
VFP acc .req r3
|
|
NOVFP len .req r3
|
|
NOVFP acc .req r2
|
|
VFP vdup.32 q15, d0[0]
|
|
NOVFP vdup.32 q15, r2
|
|
bics r12, len, #15
|
|
mov acc, r0
|
|
beq 3f
|
|
vld1.32 {q0}, [r1,:128]!
|
|
vld1.32 {q8}, [acc,:128]!
|
|
vld1.32 {q1}, [r1,:128]!
|
|
vld1.32 {q9}, [acc,:128]!
|
|
1: vmla.f32 q8, q0, q15
|
|
vld1.32 {q2}, [r1,:128]!
|
|
vld1.32 {q10}, [acc,:128]!
|
|
vmla.f32 q9, q1, q15
|
|
vld1.32 {q3}, [r1,:128]!
|
|
vld1.32 {q11}, [acc,:128]!
|
|
vmla.f32 q10, q2, q15
|
|
vst1.32 {q8}, [r0,:128]!
|
|
vmla.f32 q11, q3, q15
|
|
vst1.32 {q9}, [r0,:128]!
|
|
subs r12, r12, #16
|
|
beq 2f
|
|
vld1.32 {q0}, [r1,:128]!
|
|
vld1.32 {q8}, [acc,:128]!
|
|
vst1.32 {q10}, [r0,:128]!
|
|
vld1.32 {q1}, [r1,:128]!
|
|
vld1.32 {q9}, [acc,:128]!
|
|
vst1.32 {q11}, [r0,:128]!
|
|
b 1b
|
|
2: vst1.32 {q10}, [r0,:128]!
|
|
vst1.32 {q11}, [r0,:128]!
|
|
ands len, len, #15
|
|
it eq
|
|
bxeq lr
|
|
3: vld1.32 {q0}, [r1,:128]!
|
|
vld1.32 {q8}, [acc,:128]!
|
|
vmla.f32 q8, q0, q15
|
|
vst1.32 {q8}, [r0,:128]!
|
|
subs len, len, #4
|
|
bgt 3b
|
|
bx lr
|
|
.unreq len
|
|
endfunc
|
|
|
|
function ff_butterflies_float_neon, export=1
|
|
1: vld1.32 {q0},[r0,:128]
|
|
vld1.32 {q1},[r1,:128]
|
|
vsub.f32 q2, q0, q1
|
|
vadd.f32 q1, q0, q1
|
|
vst1.32 {q2},[r1,:128]!
|
|
vst1.32 {q1},[r0,:128]!
|
|
subs r2, r2, #4
|
|
bgt 1b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_scalarproduct_float_neon, export=1
|
|
vmov.f32 q2, #0.0
|
|
1: vld1.32 {q0},[r0,:128]!
|
|
vld1.32 {q1},[r1,:128]!
|
|
vmla.f32 q2, q0, q1
|
|
subs r2, r2, #4
|
|
bgt 1b
|
|
vadd.f32 d0, d4, d5
|
|
vpadd.f32 d0, d0, d0
|
|
NOVFP vmov.32 r0, d0[0]
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_vector_fmul_reverse_neon, export=1
|
|
add r2, r2, r3, lsl #2
|
|
sub r2, r2, #32
|
|
mov r12, #-32
|
|
vld1.32 {q0-q1}, [r1,:128]!
|
|
vld1.32 {q2-q3}, [r2,:128], r12
|
|
1: pld [r1, #32]
|
|
vrev64.32 q3, q3
|
|
vmul.f32 d16, d0, d7
|
|
vmul.f32 d17, d1, d6
|
|
pld [r2, #-32]
|
|
vrev64.32 q2, q2
|
|
vmul.f32 d18, d2, d5
|
|
vmul.f32 d19, d3, d4
|
|
subs r3, r3, #8
|
|
beq 2f
|
|
vld1.32 {q0-q1}, [r1,:128]!
|
|
vld1.32 {q2-q3}, [r2,:128], r12
|
|
vst1.32 {q8-q9}, [r0,:128]!
|
|
b 1b
|
|
2: vst1.32 {q8-q9}, [r0,:128]!
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_vector_fmul_add_neon, export=1
|
|
ldr r12, [sp]
|
|
vld1.32 {q0-q1}, [r1,:128]!
|
|
vld1.32 {q8-q9}, [r2,:128]!
|
|
vld1.32 {q2-q3}, [r3,:128]!
|
|
vmul.f32 q10, q0, q8
|
|
vmul.f32 q11, q1, q9
|
|
1: vadd.f32 q12, q2, q10
|
|
vadd.f32 q13, q3, q11
|
|
pld [r1, #16]
|
|
pld [r2, #16]
|
|
pld [r3, #16]
|
|
subs r12, r12, #8
|
|
beq 2f
|
|
vld1.32 {q0}, [r1,:128]!
|
|
vld1.32 {q8}, [r2,:128]!
|
|
vmul.f32 q10, q0, q8
|
|
vld1.32 {q1}, [r1,:128]!
|
|
vld1.32 {q9}, [r2,:128]!
|
|
vmul.f32 q11, q1, q9
|
|
vld1.32 {q2-q3}, [r3,:128]!
|
|
vst1.32 {q12-q13},[r0,:128]!
|
|
b 1b
|
|
2: vst1.32 {q12-q13},[r0,:128]!
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_vector_clipf_neon, export=1
|
|
VFP vdup.32 q1, d0[1]
|
|
VFP vdup.32 q0, d0[0]
|
|
NOVFP vdup.32 q0, r2
|
|
NOVFP vdup.32 q1, r3
|
|
NOVFP ldr r2, [sp]
|
|
vld1.f32 {q2},[r1,:128]!
|
|
vmin.f32 q10, q2, q1
|
|
vld1.f32 {q3},[r1,:128]!
|
|
vmin.f32 q11, q3, q1
|
|
1: vmax.f32 q8, q10, q0
|
|
vmax.f32 q9, q11, q0
|
|
subs r2, r2, #8
|
|
beq 2f
|
|
vld1.f32 {q2},[r1,:128]!
|
|
vmin.f32 q10, q2, q1
|
|
vld1.f32 {q3},[r1,:128]!
|
|
vmin.f32 q11, q3, q1
|
|
vst1.f32 {q8},[r0,:128]!
|
|
vst1.f32 {q9},[r0,:128]!
|
|
b 1b
|
|
2: vst1.f32 {q8},[r0,:128]!
|
|
vst1.f32 {q9},[r0,:128]!
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_apply_window_int16_neon, export=1
|
|
push {r4,lr}
|
|
add r4, r1, r3, lsl #1
|
|
add lr, r0, r3, lsl #1
|
|
sub r4, r4, #16
|
|
sub lr, lr, #16
|
|
mov r12, #-16
|
|
1:
|
|
vld1.16 {q0}, [r1,:128]!
|
|
vld1.16 {q2}, [r2,:128]!
|
|
vld1.16 {q1}, [r4,:128], r12
|
|
vrev64.16 q3, q2
|
|
vqrdmulh.s16 q0, q0, q2
|
|
vqrdmulh.s16 d2, d2, d7
|
|
vqrdmulh.s16 d3, d3, d6
|
|
vst1.16 {q0}, [r0,:128]!
|
|
vst1.16 {q1}, [lr,:128], r12
|
|
subs r3, r3, #16
|
|
bgt 1b
|
|
|
|
pop {r4,pc}
|
|
endfunc
|
|
|
|
function ff_vector_clip_int32_neon, export=1
|
|
vdup.32 q0, r2
|
|
vdup.32 q1, r3
|
|
ldr r2, [sp]
|
|
1:
|
|
vld1.32 {q2-q3}, [r1,:128]!
|
|
vmin.s32 q2, q2, q1
|
|
vmin.s32 q3, q3, q1
|
|
vmax.s32 q2, q2, q0
|
|
vmax.s32 q3, q3, q0
|
|
vst1.32 {q2-q3}, [r0,:128]!
|
|
subs r2, r2, #8
|
|
bgt 1b
|
|
bx lr
|
|
endfunc
|