mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-28 20:53:54 +02:00
2589060b92
This unbreaks the fate-checkasm-hevc_pel test on arm targets. The assembly assumed that the width passed to the DSP functions is a multiple of 8, while the checkasm test used other widths too. This wasn't noticed before, because the hevc_pel checkasm tests (that were added in9c513edb79
in January) weren't run as part of fate until inb492cacffd
in August. As this hasn't been an issue in practice with actual full decoding tests, it seems like the actual decoder doesn't call these functions with such widths. Therefore, we could alternatively fix the test to only test things that the real decoder does, and this modification could be reverted. Signed-off-by: Martin Storsjö <martin@martin.st>
1000 lines
27 KiB
ArmAsm
1000 lines
27 KiB
ArmAsm
/*
|
|
* Copyright (c) 2014 - 2015 Seppo Tomperi <seppo.tomperi@vtt.fi>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/arm/asm.S"
|
|
#include "neon.S"
|
|
|
|
#define MAX_PB_SIZE #64
|
|
|
|
.macro regshuffle_d8
|
|
vmov d16, d17
|
|
vmov d17, d18
|
|
vmov d18, d19
|
|
vmov d19, d20
|
|
vmov d20, d21
|
|
vmov d21, d22
|
|
vmov d22, d23
|
|
.endm
|
|
|
|
.macro regshuffle_q8
|
|
vmov q0, q1
|
|
vmov q1, q2
|
|
vmov q2, q3
|
|
vmov q3, q4
|
|
vmov q4, q5
|
|
vmov q5, q6
|
|
vmov q6, q7
|
|
.endm
|
|
|
|
.macro vextin8
|
|
pld [r2]
|
|
vld1.8 {q11}, [r2], r3
|
|
vext.8 d16, d22, d23, #1
|
|
vext.8 d17, d22, d23, #2
|
|
vext.8 d18, d22, d23, #3
|
|
vext.8 d19, d22, d23, #4
|
|
vext.8 d20, d22, d23, #5
|
|
vext.8 d21, d22, d23, #6
|
|
vext.8 d22, d22, d23, #7
|
|
.endm
|
|
|
|
.macro loadin8
|
|
pld [r2]
|
|
vld1.8 {d16}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d17}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d18}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d19}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d20}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d21}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d22}, [r2], r3
|
|
pld [r2]
|
|
vld1.8 {d23}, [r2], r3
|
|
.endm
|
|
|
|
.macro qpel_filter_1_32b
|
|
vmov.i16 d16, #58
|
|
vmov.i16 d17, #10
|
|
vmull.s16 q9, d6, d16 // 58 * d0
|
|
vmull.s16 q10, d7, d16 // 58 * d1
|
|
vmov.i16 d16, #17
|
|
vmull.s16 q11, d4, d17 // 10 * c0
|
|
vmull.s16 q12, d5, d17 // 10 * c1
|
|
vmov.i16 d17, #5
|
|
vmull.s16 q13, d8, d16 // 17 * e0
|
|
vmull.s16 q14, d9, d16 // 17 * e1
|
|
vmull.s16 q15, d10, d17 // 5 * f0
|
|
vmull.s16 q8, d11, d17 // 5 * f1
|
|
vsub.s32 q9, q11 // 58 * d0 - 10 * c0
|
|
vsub.s32 q10, q12 // 58 * d1 - 10 * c1
|
|
vshll.s16 q11, d2, #2 // 4 * b0
|
|
vshll.s16 q12, d3, #2 // 4 * b1
|
|
vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0
|
|
vadd.s32 q10, q14 // 58 * d1 - 10 * c1 + 17 * e1
|
|
vsubl.s16 q13, d12, d0 // g0 - a0
|
|
vsubl.s16 q14, d13, d1 // g1 - a1
|
|
vadd.s32 q9, q11 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0
|
|
vadd.s32 q10, q12 // 58 * d1 - 10 * c1 + 17 * e1 + 4 * b1
|
|
vsub.s32 q13, q15 // g0 - a0 - 5 * f0
|
|
vsub.s32 q14, q8 // g1 - a1 - 5 * f1
|
|
vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0 + g0 - a0 - 5 * f0
|
|
vadd.s32 q10, q14 // 58 * d1 - 10 * c1 + 17 * e1 + 4 * b1 + g1 - a1 - 5 * f1
|
|
vqshrn.s32 d16, q9, #6
|
|
vqshrn.s32 d17, q10, #6
|
|
.endm
|
|
|
|
// input q0 - q7
|
|
// output q8
|
|
.macro qpel_filter_2_32b
|
|
vmov.i32 q8, #11
|
|
vaddl.s16 q9, d6, d8 // d0 + e0
|
|
vaddl.s16 q10, d7, d9 // d1 + e1
|
|
vaddl.s16 q11, d4, d10 // c0 + f0
|
|
vaddl.s16 q12, d5, d11 // c1 + f1
|
|
vmul.s32 q11, q8 // 11 * (c0 + f0)
|
|
vmul.s32 q12, q8 // 11 * (c1 + f1)
|
|
vmov.i32 q8, #40
|
|
vaddl.s16 q15, d2, d12 // b0 + g0
|
|
vmul.s32 q9, q8 // 40 * (d0 + e0)
|
|
vmul.s32 q10, q8 // 40 * (d1 + e1)
|
|
vaddl.s16 q8, d3, d13 // b1 + g1
|
|
vaddl.s16 q13, d0, d14 // a0 + h0
|
|
vaddl.s16 q14, d1, d15 // a1 + h1
|
|
vshl.s32 q15, #2 // 4*(b0+g0)
|
|
vshl.s32 q8, #2 // 4*(b1+g1)
|
|
vadd.s32 q11, q13 // 11 * (c0 + f0) + a0 + h0
|
|
vadd.s32 q12, q14 // 11 * (c1 + f1) + a1 + h1
|
|
vadd.s32 q9, q15 // 40 * (d0 + e0) + 4*(b0+g0)
|
|
vadd.s32 q10, q8 // 40 * (d1 + e1) + 4*(b1+g1)
|
|
vsub.s32 q9, q11 // 40 * (d0 + e0) + 4*(b0+g0) - (11 * (c0 + f0) + a0 + h0)
|
|
vsub.s32 q10, q12 // 40 * (d1 + e1) + 4*(b1+g1) - (11 * (c1 + f1) + a1 + h1)
|
|
vqshrn.s32 d16, q9, #6
|
|
vqshrn.s32 d17, q10, #6
|
|
.endm
|
|
|
|
.macro qpel_filter_3_32b
|
|
vmov.i16 d16, #58
|
|
vmov.i16 d17, #10
|
|
vmull.s16 q9, d8, d16 // 58 * d0
|
|
vmull.s16 q10, d9, d16 // 58 * d1
|
|
vmov.i16 d16, #17
|
|
vmull.s16 q11, d10, d17 // 10 * c0
|
|
vmull.s16 q12, d11, d17 // 10 * c1
|
|
vmov.i16 d17, #5
|
|
vmull.s16 q13, d6, d16 // 17 * e0
|
|
vmull.s16 q14, d7, d16 // 17 * e1
|
|
vmull.s16 q15, d4, d17 // 5 * f0
|
|
vmull.s16 q8, d5, d17 // 5 * f1
|
|
vsub.s32 q9, q11 // 58 * d0 - 10 * c0
|
|
vsub.s32 q10, q12 // 58 * d1 - 10 * c1
|
|
vshll.s16 q11, d12, #2 // 4 * b0
|
|
vshll.s16 q12, d13, #2 // 4 * b1
|
|
vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0
|
|
vadd.s32 q10, q14 // 58 * d1 - 10 * c1 + 17 * e1
|
|
vsubl.s16 q13, d2, d14 // g0 - a0
|
|
vsubl.s16 q14, d3, d15 // g1 - a1
|
|
vadd.s32 q9, q11 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0
|
|
vadd.s32 q10, q12 // 58 * d1 - 10 * c1 + 17 * e1 + 4 * b1
|
|
vsub.s32 q13, q15 // g0 - a0 - 5 * f0
|
|
vsub.s32 q14, q8 // g1 - a1 - 5 * f1
|
|
vadd.s32 q9, q13 // 58 * d0 - 10 * c0 + 17 * e0 + 4 * b0 + g0 - a0 - 5 * f0
|
|
vadd.s32 q10, q14 // 58 * d1 - 10 * c1 + 17 * e1 + 4 * b1 + g1 - a1 - 5 * f1
|
|
vqshrn.s32 d16, q9, #6
|
|
vqshrn.s32 d17, q10, #6
|
|
.endm
|
|
|
|
.macro qpel_filter_1 out=q7
|
|
vmov.u8 d24, #58
|
|
vmov.u8 d25, #10
|
|
vshll.u8 q13, d20, #4 // 16*e
|
|
vshll.u8 q14, d21, #2 // 4*f
|
|
vmull.u8 \out, d19, d24 // 58*d
|
|
vaddw.u8 q13, q13, d20 // 17*e
|
|
vmull.u8 q15, d18, d25 // 10*c
|
|
vaddw.u8 q14, q14, d21 // 5*f
|
|
vsubl.u8 q12, d22, d16 // g - a
|
|
vadd.u16 \out, q13 // 58d + 17e
|
|
vshll.u8 q13, d17, #2 // 4*b
|
|
vadd.u16 q15, q14 // 10*c + 5*f
|
|
vadd.s16 q13, q12 // - a + 4*b + g
|
|
vsub.s16 \out, q15 // -10*c + 58*d + 17*e -5*f
|
|
vadd.s16 \out, q13 // -a + 4*b -10*c + 58*d + 17*e -5*f
|
|
.endm
|
|
|
|
.macro qpel_filter_2 out=q7
|
|
vmov.i16 q12, #10
|
|
vmov.i16 q14, #11
|
|
vaddl.u8 q13, d19, d20 // d + e
|
|
vaddl.u8 q15, d18, d21 // c + f
|
|
vmul.u16 q13, q12 // 10 * (d+e)
|
|
vmul.u16 q15, q14 // 11 * ( c + f)
|
|
vaddl.u8 \out, d17, d22 // b + g
|
|
vaddl.u8 q12, d16, d23 // a + h
|
|
vadd.u16 \out, q13 // b + 10 * (d + e) + g
|
|
vadd.s16 q12, q15
|
|
vshl.u16 \out, #2 // 4 * (b + 10 * (d + e) + g)
|
|
vsub.s16 \out, q12
|
|
.endm
|
|
|
|
.macro qpel_filter_3 out=q7
|
|
vmov.u8 d24, #58
|
|
vmov.u8 d25, #10
|
|
vshll.u8 q13, d19, #4 // 16*e
|
|
vshll.u8 q14, d18, #2 // 4*f
|
|
vmull.u8 \out, d20, d24 // 58*d
|
|
vaddw.u8 q13, q13, d19 // 17*e
|
|
vmull.u8 q15, d21, d25 // 10*c
|
|
vaddw.u8 q14, q14, d18 // 5*f
|
|
vsubl.u8 q12, d17, d23 // g - a
|
|
vadd.u16 \out, q13 // 58d + 17e
|
|
vshll.u8 q13, d22, #2 // 4*b
|
|
vadd.u16 q15, q14 // 10*c + 5*f
|
|
vadd.s16 q13, q12 // - a + 4*b + g
|
|
vsub.s16 \out, q15 // -10*c + 58*d + 17*e -5*f
|
|
vadd.s16 \out, q13 // -a + 4*b -10*c + 58*d + 17*e -5*f
|
|
.endm
|
|
|
|
.macro hevc_put_qpel_vX_neon_8 filter
|
|
push {r4, r5, r6, r7}
|
|
ldr r4, [sp, #16] // height
|
|
ldr r5, [sp, #20] // width
|
|
vpush {d8-d15}
|
|
sub r2, r2, r3, lsl #1
|
|
sub r2, r3
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
lsl r1, #1
|
|
0: loadin8
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filter
|
|
vst1.16 {q7}, [r0], r1
|
|
regshuffle_d8
|
|
vld1.8 {d23}, [r2], r3
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #16
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filter
|
|
vst1.16 d14, [r0], r1
|
|
regshuffle_d8
|
|
vld1.32 {d23[0]}, [r2], r3
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4, r5, r6, r7}
|
|
bx lr
|
|
.endm
|
|
|
|
.macro hevc_put_qpel_uw_vX_neon_8 filter
|
|
push {r4-r10}
|
|
ldr r5, [sp, #28] // width
|
|
ldr r4, [sp, #32] // height
|
|
ldr r8, [sp, #36] // src2
|
|
ldr r9, [sp, #40] // src2stride
|
|
vpush {d8-d15}
|
|
sub r2, r2, r3, lsl #1
|
|
sub r2, r3
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
cmp r8, #0
|
|
bne .Lbi\@
|
|
0: loadin8
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filter
|
|
vqrshrun.s16 d0, q7, #6
|
|
vst1.8 d0, [r0], r1
|
|
regshuffle_d8
|
|
vld1.8 {d23}, [r2], r3
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filter
|
|
vqrshrun.s16 d0, q7, #6
|
|
vst1.32 d0[0], [r0], r1
|
|
regshuffle_d8
|
|
vld1.32 {d23[0]}, [r2], r3
|
|
bne 4b
|
|
b 99f
|
|
.Lbi\@: lsl r9, #1
|
|
mov r10, r8
|
|
0: loadin8
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filter
|
|
vld1.16 {q0}, [r8], r9
|
|
vqadd.s16 q0, q7
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.8 d0, [r0], r1
|
|
regshuffle_d8
|
|
vld1.8 {d23}, [r2], r3
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
mov r0, r6
|
|
add r10, #16
|
|
mov r8, r10
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filter
|
|
vld1.16 d0, [r8], r9
|
|
vqadd.s16 d0, d14
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.32 d0[0], [r0], r1
|
|
regshuffle_d8
|
|
vld1.32 {d23[0]}, [r2], r3
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4-r10}
|
|
bx lr
|
|
.endm
|
|
|
|
function ff_hevc_put_qpel_v1_neon_8, export=1
|
|
hevc_put_qpel_vX_neon_8 qpel_filter_1
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_v2_neon_8, export=1
|
|
hevc_put_qpel_vX_neon_8 qpel_filter_2
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_v3_neon_8, export=1
|
|
hevc_put_qpel_vX_neon_8 qpel_filter_3
|
|
endfunc
|
|
|
|
|
|
function ff_hevc_put_qpel_uw_v1_neon_8, export=1
|
|
hevc_put_qpel_uw_vX_neon_8 qpel_filter_1
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_v2_neon_8, export=1
|
|
hevc_put_qpel_uw_vX_neon_8 qpel_filter_2
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_v3_neon_8, export=1
|
|
hevc_put_qpel_uw_vX_neon_8 qpel_filter_3
|
|
endfunc
|
|
|
|
.macro hevc_put_qpel_hX_neon_8 filter
|
|
push {r4, r5, r6, r7}
|
|
ldr r4, [sp, #16] // height
|
|
ldr r5, [sp, #20] // width
|
|
|
|
vpush {d8-d15}
|
|
sub r2, #4
|
|
lsl r1, #1
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vst1.16 {q7}, [r0], r1
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #16
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
cmp r5, #4
|
|
bne 8b
|
|
4: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vst1.16 d14, [r0], r1
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4, r5, r6, r7}
|
|
bx lr
|
|
.endm
|
|
|
|
.macro hevc_put_qpel_uw_hX_neon_8 filter
|
|
push {r4-r10}
|
|
ldr r5, [sp, #28] // width
|
|
ldr r4, [sp, #32] // height
|
|
ldr r8, [sp, #36] // src2
|
|
ldr r9, [sp, #40] // src2stride
|
|
vpush {d8-d15}
|
|
sub r2, #4
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
cmp r8, #0
|
|
bne .Lbi\@
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vqrshrun.s16 d0, q7, #6
|
|
vst1.8 d0, [r0], r1
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
cmp r5, #4
|
|
bne 8b
|
|
4: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vqrshrun.s16 d0, q7, #6
|
|
vst1.32 d0[0], [r0], r1
|
|
bne 4b
|
|
b 99f
|
|
.Lbi\@:
|
|
lsl r9, #1
|
|
cmp r5, #4
|
|
beq 4f
|
|
mov r10, r8
|
|
8: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vld1.16 {q0}, [r8], r9
|
|
vqadd.s16 q0, q7
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.8 d0, [r0], r1
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
add r10, #16
|
|
mov r8, r10
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
cmp r5, #4
|
|
bne 8b
|
|
4: subs r4, #1
|
|
vextin8
|
|
\filter
|
|
vld1.16 d0, [r8], r9
|
|
vqadd.s16 d0, d14
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.32 d0[0], [r0], r1
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4-r10}
|
|
bx lr
|
|
.endm
|
|
|
|
function ff_hevc_put_qpel_h1_neon_8, export=1
|
|
hevc_put_qpel_hX_neon_8 qpel_filter_1
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h2_neon_8, export=1
|
|
hevc_put_qpel_hX_neon_8 qpel_filter_2
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h3_neon_8, export=1
|
|
hevc_put_qpel_hX_neon_8 qpel_filter_3
|
|
endfunc
|
|
|
|
|
|
function ff_hevc_put_qpel_uw_h1_neon_8, export=1
|
|
hevc_put_qpel_uw_hX_neon_8 qpel_filter_1
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h2_neon_8, export=1
|
|
hevc_put_qpel_uw_hX_neon_8 qpel_filter_2
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h3_neon_8, export=1
|
|
hevc_put_qpel_uw_hX_neon_8 qpel_filter_3
|
|
endfunc
|
|
|
|
.macro hevc_put_qpel_hXvY_neon_8 filterh filterv
|
|
push {r4, r5, r6, r7}
|
|
ldr r4, [sp, #16] // height
|
|
ldr r5, [sp, #20] // width
|
|
|
|
vpush {d8-d15}
|
|
sub r2, #4
|
|
sub r2, r2, r3, lsl #1
|
|
sub r2, r3 // extra_before 3
|
|
lsl r1, #1
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
0: vextin8
|
|
\filterh q0
|
|
vextin8
|
|
\filterh q1
|
|
vextin8
|
|
\filterh q2
|
|
vextin8
|
|
\filterh q3
|
|
vextin8
|
|
\filterh q4
|
|
vextin8
|
|
\filterh q5
|
|
vextin8
|
|
\filterh q6
|
|
vextin8
|
|
\filterh q7
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filterv
|
|
vst1.16 {q8}, [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #16
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filterv
|
|
vst1.16 d16, [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4, r5, r6, r7}
|
|
bx lr
|
|
.endm
|
|
|
|
.macro hevc_put_qpel_uw_hXvY_neon_8 filterh filterv
|
|
push {r4-r10}
|
|
ldr r5, [sp, #28] // width
|
|
ldr r4, [sp, #32] // height
|
|
ldr r8, [sp, #36] // src2
|
|
ldr r9, [sp, #40] // src2stride
|
|
vpush {d8-d15}
|
|
sub r2, #4
|
|
sub r2, r2, r3, lsl #1
|
|
sub r2, r3 // extra_before 3
|
|
mov r12, r4
|
|
mov r6, r0
|
|
mov r7, r2
|
|
cmp r8, #0
|
|
bne .Lbi\@
|
|
0: vextin8
|
|
\filterh q0
|
|
vextin8
|
|
\filterh q1
|
|
vextin8
|
|
\filterh q2
|
|
vextin8
|
|
\filterh q3
|
|
vextin8
|
|
\filterh q4
|
|
vextin8
|
|
\filterh q5
|
|
vextin8
|
|
\filterh q6
|
|
vextin8
|
|
\filterh q7
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filterv
|
|
vqrshrun.s16 d0, q8, #6
|
|
vst1.8 d0, [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
mov r0, r6
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filterv
|
|
vqrshrun.s16 d0, q8, #6
|
|
vst1.32 d0[0], [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 4b
|
|
b 99f
|
|
.Lbi\@: lsl r9, #1
|
|
mov r10, r8
|
|
0: vextin8
|
|
\filterh q0
|
|
vextin8
|
|
\filterh q1
|
|
vextin8
|
|
\filterh q2
|
|
vextin8
|
|
\filterh q3
|
|
vextin8
|
|
\filterh q4
|
|
vextin8
|
|
\filterh q5
|
|
vextin8
|
|
\filterh q6
|
|
vextin8
|
|
\filterh q7
|
|
cmp r5, #4
|
|
beq 4f
|
|
8: subs r4, #1
|
|
\filterv
|
|
vld1.16 {q0}, [r8], r9
|
|
vqadd.s16 q0, q8
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.8 d0, [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 8b
|
|
subs r5, #8
|
|
ble 99f
|
|
mov r4, r12
|
|
add r6, #8
|
|
mov r0, r6
|
|
add r10, #16
|
|
mov r8, r10
|
|
add r7, #8
|
|
mov r2, r7
|
|
b 0b
|
|
4: subs r4, #1
|
|
\filterv
|
|
vld1.16 d0, [r8], r9
|
|
vqadd.s16 d0, d16
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.32 d0[0], [r0], r1
|
|
regshuffle_q8
|
|
vextin8
|
|
\filterh q7
|
|
bne 4b
|
|
99: vpop {d8-d15}
|
|
pop {r4-r10}
|
|
bx lr
|
|
.endm
|
|
|
|
|
|
function ff_hevc_put_qpel_h1v1_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_1, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h2v1_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_2, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h3v1_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_3, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h1v2_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_1, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h2v2_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_2, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h3v2_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_3, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h1v3_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_1, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h2v3_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_2, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_h3v3_neon_8, export=1
|
|
hevc_put_qpel_hXvY_neon_8 qpel_filter_3, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
|
|
function ff_hevc_put_qpel_uw_h1v1_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_1, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h2v1_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_2, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h3v1_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_3, qpel_filter_1_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h1v2_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_1, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h2v2_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_2, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h3v2_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_3, qpel_filter_2_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h1v3_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_1, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h2v3_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_2, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_h3v3_neon_8, export=1
|
|
hevc_put_qpel_uw_hXvY_neon_8 qpel_filter_3, qpel_filter_3_32b
|
|
endfunc
|
|
|
|
.macro init_put_pixels
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
mov r12, MAX_PB_SIZE
|
|
lsl r12, #1
|
|
.endm
|
|
|
|
function ff_hevc_put_pixels_w2_neon_8, export=1
|
|
init_put_pixels
|
|
vmov.u8 d5, #255
|
|
vshr.u64 d5, #32
|
|
0: subs r3, #1
|
|
vld1.32 {d0[0]}, [r1], r2
|
|
pld [r1]
|
|
vld1.32 d6, [r0]
|
|
vshll.u8 q0, d0, #6
|
|
vbit d6, d0, d5
|
|
vst1.32 d6, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w4_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #2
|
|
vld1.32 {d0[0]}, [r1], r2
|
|
vld1.32 {d0[1]}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vshll.u8 q0, d0, #6
|
|
vst1.64 {d0}, [r0], r12
|
|
vst1.64 {d1}, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w6_neon_8, export=1
|
|
init_put_pixels
|
|
vmov.u8 q10, #255
|
|
vshr.u64 d21, #32
|
|
0: subs r3, #1
|
|
vld1.16 {d0}, [r1], r2
|
|
pld [r1]
|
|
vshll.u8 q0, d0, #6
|
|
vld1.8 {q12}, [r0]
|
|
vbit q12, q0, q10
|
|
vst1.8 {q12}, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w8_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #2
|
|
vld1.8 {d0}, [r1], r2
|
|
vld1.8 {d2}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vshll.u8 q0, d0, #6
|
|
vshll.u8 q1, d2, #6
|
|
vst1.16 {q0}, [r0], r12
|
|
vst1.16 {q1}, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w12_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #2
|
|
vld1.64 {d0}, [r1]
|
|
add r1, #8
|
|
vld1.32 {d1[0]}, [r1], r2
|
|
sub r1, #8
|
|
vld1.64 {d2}, [r1]
|
|
add r1, #8
|
|
vld1.32 {d1[1]}, [r1], r2
|
|
sub r1, #8
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vshll.u8 q8, d0, #6
|
|
vshll.u8 q9, d1, #6
|
|
vshll.u8 q10, d2, #6
|
|
vmov d22, d19
|
|
vst1.64 {d16, d17, d18}, [r0], r12
|
|
vst1.64 {d20, d21, d22}, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w16_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #2
|
|
vld1.8 {q0}, [r1], r2
|
|
vld1.8 {q1}, [r1], r2
|
|
pld [r1]
|
|
pld [r1, r2]
|
|
vshll.u8 q8, d0, #6
|
|
vshll.u8 q9, d1, #6
|
|
vshll.u8 q10, d2, #6
|
|
vshll.u8 q11, d3, #6
|
|
vst1.8 {q8, q9}, [r0], r12
|
|
vst1.8 {q10, q11}, [r0], r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w24_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #1
|
|
vld1.8 {d0, d1, d2}, [r1], r2
|
|
pld [r1]
|
|
vshll.u8 q10, d0, #6
|
|
vshll.u8 q11, d1, #6
|
|
vshll.u8 q12, d2, #6
|
|
vstm r0, {q10, q11, q12}
|
|
add r0, r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w32_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #1
|
|
vld1.8 {q0, q1}, [r1], r2
|
|
pld [r1]
|
|
vshll.u8 q8, d0, #6
|
|
vshll.u8 q9, d1, #6
|
|
vshll.u8 q10, d2, #6
|
|
vshll.u8 q11, d3, #6
|
|
vstm r0, {q8, q9, q10, q11}
|
|
add r0, r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w48_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #1
|
|
vld1.8 {q0, q1}, [r1]
|
|
add r1, #32
|
|
vld1.8 {q2}, [r1], r2
|
|
sub r1, #32
|
|
pld [r1]
|
|
vshll.u8 q8, d0, #6
|
|
vshll.u8 q9, d1, #6
|
|
vshll.u8 q10, d2, #6
|
|
vshll.u8 q11, d3, #6
|
|
vshll.u8 q12, d4, #6
|
|
vshll.u8 q13, d5, #6
|
|
vstm r0, {q8, q9, q10, q11, q12, q13}
|
|
add r0, r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_pixels_w64_neon_8, export=1
|
|
init_put_pixels
|
|
0: subs r3, #1
|
|
vld1.8 {q0, q1}, [r1]
|
|
add r1, #32
|
|
vld1.8 {q2, q3}, [r1], r2
|
|
sub r1, #32
|
|
pld [r1]
|
|
vshll.u8 q8, d0, #6
|
|
vshll.u8 q9, d1, #6
|
|
vshll.u8 q10, d2, #6
|
|
vshll.u8 q11, d3, #6
|
|
vshll.u8 q12, d4, #6
|
|
vshll.u8 q13, d5, #6
|
|
vshll.u8 q14, d6, #6
|
|
vshll.u8 q15, d7, #6
|
|
vstm r0, {q8, q9, q10, q11, q12, q13, q14, q15}
|
|
add r0, r12
|
|
bne 0b
|
|
bx lr
|
|
endfunc
|
|
|
|
function ff_hevc_put_qpel_uw_pixels_neon_8, export=1
|
|
push {r4-r9}
|
|
ldr r5, [sp, #24] // width
|
|
ldr r4, [sp, #28] // height
|
|
ldr r8, [sp, #32] // src2
|
|
ldr r9, [sp, #36] // src2stride
|
|
vpush {d8-d15}
|
|
cmp r8, #0
|
|
bne 2f
|
|
1: subs r4, #1
|
|
vld1.8 {d0}, [r2], r3
|
|
vst1.8 d0, [r0], r1
|
|
bne 1b
|
|
vpop {d8-d15}
|
|
pop {r4-r9}
|
|
bx lr
|
|
2: subs r4, #1
|
|
vld1.8 {d0}, [r2], r3
|
|
vld1.16 {q1}, [r8], r9
|
|
vshll.u8 q0, d0, #6
|
|
vqadd.s16 q0, q1
|
|
vqrshrun.s16 d0, q0, #7
|
|
vst1.8 d0, [r0], r1
|
|
bne 2b
|
|
vpop {d8-d15}
|
|
pop {r4-r9}
|
|
bx lr
|
|
endfunc
|
|
|
|
.macro put_qpel_uw_pixels width, regs, regs2, regs3, regs4
|
|
function ff_hevc_put_qpel_uw_pixels_w\width\()_neon_8, export=1
|
|
ldr r12, [sp] // height
|
|
1: subs r12, #4
|
|
vld1.32 {\regs} , [r2], r3
|
|
vld1.32 {\regs2} , [r2], r3
|
|
vld1.32 {\regs3} , [r2], r3
|
|
vld1.32 {\regs4} , [r2], r3
|
|
vst1.32 {\regs} , [r0], r1
|
|
vst1.32 {\regs2} , [r0], r1
|
|
vst1.32 {\regs3} , [r0], r1
|
|
vst1.32 {\regs4} , [r0], r1
|
|
bne 1b
|
|
bx lr
|
|
endfunc
|
|
.endm
|
|
|
|
.macro put_qpel_uw_pixels_m width, regs, regs2, regs3, regs4
|
|
function ff_hevc_put_qpel_uw_pixels_w\width\()_neon_8, export=1
|
|
push {r4-r5}
|
|
ldr r12, [sp, #8] // height
|
|
1: subs r12, #2
|
|
mov r4, r2
|
|
vld1.32 {\regs} , [r2]!
|
|
vld1.32 {\regs2} , [r2]
|
|
add r2, r4, r3
|
|
mov r4, r2
|
|
vld1.32 {\regs3} , [r2]!
|
|
vld1.32 {\regs4} , [r2]
|
|
add r2, r4, r3
|
|
mov r5, r0
|
|
vst1.32 {\regs} , [r0]!
|
|
vst1.32 {\regs2} , [r0]
|
|
add r0, r5, r1
|
|
mov r5, r0
|
|
vst1.32 {\regs3} , [r0]!
|
|
vst1.32 {\regs4} , [r0]
|
|
add r0, r5, r1
|
|
bne 1b
|
|
pop {r4-r5}
|
|
bx lr
|
|
endfunc
|
|
.endm
|
|
|
|
put_qpel_uw_pixels 4, d0[0], d0[1], d1[0], d1[1]
|
|
put_qpel_uw_pixels 8, d0, d1, d2, d3
|
|
put_qpel_uw_pixels_m 12, d0, d1[0], d2, d3[0]
|
|
put_qpel_uw_pixels 16, q0, q1, q2, q3
|
|
put_qpel_uw_pixels 24, d0-d2, d3-d5, d16-d18, d19-d21
|
|
put_qpel_uw_pixels 32, q0-q1, q2-q3, q8-q9, q10-q11
|
|
put_qpel_uw_pixels_m 48, q0-q1, q2, q8-q9, q10
|
|
put_qpel_uw_pixels_m 64, q0-q1, q2-q3, q8-q9, q10-q11
|