mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-11-26 19:01:44 +02:00
2912e87a6c
Signed-off-by: Mans Rullgard <mans@mansr.com>
119 lines
4.0 KiB
ArmAsm
119 lines
4.0 KiB
ArmAsm
/*
|
|
* ARM NEON optimised integer operations
|
|
* Copyright (c) 2009 Kostya Shishkov
|
|
*
|
|
* This file is part of Libav.
|
|
*
|
|
* Libav is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* Libav is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with Libav; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "asm.S"
|
|
|
|
preserve8
|
|
.fpu neon
|
|
.text
|
|
|
|
function ff_scalarproduct_int16_neon, export=1
|
|
vmov.i16 q0, #0
|
|
vmov.i16 q1, #0
|
|
vmov.i16 q2, #0
|
|
vmov.i16 q3, #0
|
|
negs r3, r3
|
|
beq 2f
|
|
|
|
vdup.s32 q12, r3
|
|
1: vld1.16 {d16-d17}, [r0]!
|
|
vld1.16 {d20-d21}, [r1,:128]!
|
|
vmull.s16 q12, d16, d20
|
|
vld1.16 {d18-d19}, [r0]!
|
|
vmull.s16 q13, d17, d21
|
|
vld1.16 {d22-d23}, [r1,:128]!
|
|
vmull.s16 q14, d18, d22
|
|
vmull.s16 q15, d19, d23
|
|
vshl.s32 q8, q12, q12
|
|
vshl.s32 q9, q13, q12
|
|
vadd.s32 q0, q0, q8
|
|
vshl.s32 q10, q14, q12
|
|
vadd.s32 q1, q1, q9
|
|
vshl.s32 q11, q15, q12
|
|
vadd.s32 q2, q2, q10
|
|
vadd.s32 q3, q3, q11
|
|
subs r2, r2, #16
|
|
bne 1b
|
|
b 3f
|
|
|
|
2: vld1.16 {d16-d17}, [r0]!
|
|
vld1.16 {d20-d21}, [r1,:128]!
|
|
vmlal.s16 q0, d16, d20
|
|
vld1.16 {d18-d19}, [r0]!
|
|
vmlal.s16 q1, d17, d21
|
|
vld1.16 {d22-d23}, [r1,:128]!
|
|
vmlal.s16 q2, d18, d22
|
|
vmlal.s16 q3, d19, d23
|
|
subs r2, r2, #16
|
|
bne 2b
|
|
|
|
3: vpadd.s32 d16, d0, d1
|
|
vpadd.s32 d17, d2, d3
|
|
vpadd.s32 d10, d4, d5
|
|
vpadd.s32 d11, d6, d7
|
|
vpadd.s32 d0, d16, d17
|
|
vpadd.s32 d1, d10, d11
|
|
vpadd.s32 d2, d0, d1
|
|
vpaddl.s32 d3, d2
|
|
vmov.32 r0, d3[0]
|
|
bx lr
|
|
endfunc
|
|
|
|
@ scalarproduct_and_madd_int16(/*aligned*/v0,v1,v2,order,mul)
|
|
function ff_scalarproduct_and_madd_int16_neon, export=1
|
|
vld1.16 {d28[],d29[]}, [sp]
|
|
vmov.i16 q0, #0
|
|
vmov.i16 q1, #0
|
|
vmov.i16 q2, #0
|
|
vmov.i16 q3, #0
|
|
mov r12, r0
|
|
|
|
1: vld1.16 {d16-d17}, [r0,:128]!
|
|
vld1.16 {d18-d19}, [r1]!
|
|
vld1.16 {d20-d21}, [r2]!
|
|
vld1.16 {d22-d23}, [r0,:128]!
|
|
vld1.16 {d24-d25}, [r1]!
|
|
vld1.16 {d26-d27}, [r2]!
|
|
vmul.s16 q10, q10, q14
|
|
vmul.s16 q13, q13, q14
|
|
vmlal.s16 q0, d16, d18
|
|
vmlal.s16 q1, d17, d19
|
|
vadd.s16 q10, q8, q10
|
|
vadd.s16 q13, q11, q13
|
|
vmlal.s16 q2, d22, d24
|
|
vmlal.s16 q3, d23, d25
|
|
vst1.16 {q10}, [r12,:128]!
|
|
subs r3, r3, #16
|
|
vst1.16 {q13}, [r12,:128]!
|
|
bne 1b
|
|
|
|
vpadd.s32 d16, d0, d1
|
|
vpadd.s32 d17, d2, d3
|
|
vpadd.s32 d10, d4, d5
|
|
vpadd.s32 d11, d6, d7
|
|
vpadd.s32 d0, d16, d17
|
|
vpadd.s32 d1, d10, d11
|
|
vpadd.s32 d2, d0, d1
|
|
vpaddl.s32 d3, d2
|
|
vmov.32 r0, d3[0]
|
|
bx lr
|
|
endfunc
|