mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
3a2d1465c8
* commit '2d60444331fca1910510038dd3817bea885c2367': dsputil: Split motion estimation compare bits off into their own context Conflicts: configure libavcodec/Makefile libavcodec/arm/Makefile libavcodec/dvenc.c libavcodec/error_resilience.c libavcodec/h264.h libavcodec/h264_slice.c libavcodec/me_cmp.c libavcodec/me_cmp.h libavcodec/motion_est.c libavcodec/motion_est_template.c libavcodec/mpeg4videoenc.c libavcodec/mpegvideo.c libavcodec/mpegvideo_enc.c libavcodec/x86/Makefile libavcodec/x86/me_cmp_init.c Merged-by: Michael Niedermayer <michaelni@gmx.at>
245 lines
8.3 KiB
ArmAsm
245 lines
8.3 KiB
ArmAsm
/*
|
|
* Copyright (c) 2009 Mans Rullgard <mans@mansr.com>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/arm/asm.S"
|
|
|
|
function ff_pix_abs16_armv6, export=1
|
|
ldr r0, [sp]
|
|
push {r4-r9, lr}
|
|
mov r12, #0
|
|
mov lr, #0
|
|
ldm r1, {r4-r7}
|
|
ldr r8, [r2]
|
|
1:
|
|
ldr r9, [r2, #4]
|
|
pld [r1, r3]
|
|
usada8 r12, r4, r8, r12
|
|
ldr r8, [r2, #8]
|
|
pld [r2, r3]
|
|
usada8 lr, r5, r9, lr
|
|
ldr r9, [r2, #12]
|
|
usada8 r12, r6, r8, r12
|
|
subs r0, r0, #1
|
|
usada8 lr, r7, r9, lr
|
|
beq 2f
|
|
add r1, r1, r3
|
|
ldm r1, {r4-r7}
|
|
add r2, r2, r3
|
|
ldr r8, [r2]
|
|
b 1b
|
|
2:
|
|
add r0, r12, lr
|
|
pop {r4-r9, pc}
|
|
endfunc
|
|
|
|
function ff_pix_abs16_x2_armv6, export=1
|
|
ldr r12, [sp]
|
|
push {r4-r11, lr}
|
|
mov r0, #0
|
|
mov lr, #1
|
|
orr lr, lr, lr, lsl #8
|
|
orr lr, lr, lr, lsl #16
|
|
1:
|
|
ldr r8, [r2]
|
|
ldr r9, [r2, #4]
|
|
lsr r10, r8, #8
|
|
ldr r4, [r1]
|
|
lsr r6, r9, #8
|
|
orr r10, r10, r9, lsl #24
|
|
ldr r5, [r2, #8]
|
|
eor r11, r8, r10
|
|
uhadd8 r7, r8, r10
|
|
orr r6, r6, r5, lsl #24
|
|
and r11, r11, lr
|
|
uadd8 r7, r7, r11
|
|
ldr r8, [r1, #4]
|
|
usada8 r0, r4, r7, r0
|
|
eor r7, r9, r6
|
|
lsr r10, r5, #8
|
|
and r7, r7, lr
|
|
uhadd8 r4, r9, r6
|
|
ldr r6, [r2, #12]
|
|
uadd8 r4, r4, r7
|
|
pld [r1, r3]
|
|
orr r10, r10, r6, lsl #24
|
|
usada8 r0, r8, r4, r0
|
|
ldr r4, [r1, #8]
|
|
eor r11, r5, r10
|
|
ldrb r7, [r2, #16]
|
|
and r11, r11, lr
|
|
uhadd8 r8, r5, r10
|
|
ldr r5, [r1, #12]
|
|
uadd8 r8, r8, r11
|
|
pld [r2, r3]
|
|
lsr r10, r6, #8
|
|
usada8 r0, r4, r8, r0
|
|
orr r10, r10, r7, lsl #24
|
|
subs r12, r12, #1
|
|
eor r11, r6, r10
|
|
add r1, r1, r3
|
|
uhadd8 r9, r6, r10
|
|
and r11, r11, lr
|
|
uadd8 r9, r9, r11
|
|
add r2, r2, r3
|
|
usada8 r0, r5, r9, r0
|
|
bgt 1b
|
|
|
|
pop {r4-r11, pc}
|
|
endfunc
|
|
|
|
.macro usad_y2 p0, p1, p2, p3, n0, n1, n2, n3
|
|
ldr \n0, [r2]
|
|
eor \n1, \p0, \n0
|
|
uhadd8 \p0, \p0, \n0
|
|
and \n1, \n1, lr
|
|
ldr \n2, [r1]
|
|
uadd8 \p0, \p0, \n1
|
|
ldr \n1, [r2, #4]
|
|
usada8 r0, \p0, \n2, r0
|
|
pld [r1, r3]
|
|
eor \n3, \p1, \n1
|
|
uhadd8 \p1, \p1, \n1
|
|
and \n3, \n3, lr
|
|
ldr \p0, [r1, #4]
|
|
uadd8 \p1, \p1, \n3
|
|
ldr \n2, [r2, #8]
|
|
usada8 r0, \p1, \p0, r0
|
|
pld [r2, r3]
|
|
eor \p0, \p2, \n2
|
|
uhadd8 \p2, \p2, \n2
|
|
and \p0, \p0, lr
|
|
ldr \p1, [r1, #8]
|
|
uadd8 \p2, \p2, \p0
|
|
ldr \n3, [r2, #12]
|
|
usada8 r0, \p2, \p1, r0
|
|
eor \p1, \p3, \n3
|
|
uhadd8 \p3, \p3, \n3
|
|
and \p1, \p1, lr
|
|
ldr \p0, [r1, #12]
|
|
uadd8 \p3, \p3, \p1
|
|
add r1, r1, r3
|
|
usada8 r0, \p3, \p0, r0
|
|
add r2, r2, r3
|
|
.endm
|
|
|
|
function ff_pix_abs16_y2_armv6, export=1
|
|
pld [r1]
|
|
pld [r2]
|
|
ldr r12, [sp]
|
|
push {r4-r11, lr}
|
|
mov r0, #0
|
|
mov lr, #1
|
|
orr lr, lr, lr, lsl #8
|
|
orr lr, lr, lr, lsl #16
|
|
ldr r4, [r2]
|
|
ldr r5, [r2, #4]
|
|
ldr r6, [r2, #8]
|
|
ldr r7, [r2, #12]
|
|
add r2, r2, r3
|
|
1:
|
|
usad_y2 r4, r5, r6, r7, r8, r9, r10, r11
|
|
subs r12, r12, #2
|
|
usad_y2 r8, r9, r10, r11, r4, r5, r6, r7
|
|
bgt 1b
|
|
|
|
pop {r4-r11, pc}
|
|
endfunc
|
|
|
|
function ff_pix_abs8_armv6, export=1
|
|
pld [r2, r3]
|
|
ldr r12, [sp]
|
|
push {r4-r9, lr}
|
|
mov r0, #0
|
|
mov lr, #0
|
|
ldrd_post r4, r5, r1, r3
|
|
1:
|
|
subs r12, r12, #2
|
|
ldr r7, [r2, #4]
|
|
ldr_post r6, r2, r3
|
|
ldrd_post r8, r9, r1, r3
|
|
usada8 r0, r4, r6, r0
|
|
pld [r2, r3]
|
|
usada8 lr, r5, r7, lr
|
|
ldr r7, [r2, #4]
|
|
ldr_post r6, r2, r3
|
|
beq 2f
|
|
ldrd_post r4, r5, r1, r3
|
|
usada8 r0, r8, r6, r0
|
|
pld [r2, r3]
|
|
usada8 lr, r9, r7, lr
|
|
b 1b
|
|
2:
|
|
usada8 r0, r8, r6, r0
|
|
usada8 lr, r9, r7, lr
|
|
add r0, r0, lr
|
|
pop {r4-r9, pc}
|
|
endfunc
|
|
|
|
function ff_sse16_armv6, export=1
|
|
ldr r12, [sp]
|
|
push {r4-r9, lr}
|
|
mov r0, #0
|
|
1:
|
|
ldrd r4, r5, [r1]
|
|
ldr r8, [r2]
|
|
uxtb16 lr, r4
|
|
uxtb16 r4, r4, ror #8
|
|
uxtb16 r9, r8
|
|
uxtb16 r8, r8, ror #8
|
|
ldr r7, [r2, #4]
|
|
usub16 lr, lr, r9
|
|
usub16 r4, r4, r8
|
|
smlad r0, lr, lr, r0
|
|
uxtb16 r6, r5
|
|
uxtb16 lr, r5, ror #8
|
|
uxtb16 r8, r7
|
|
uxtb16 r9, r7, ror #8
|
|
smlad r0, r4, r4, r0
|
|
ldrd r4, r5, [r1, #8]
|
|
usub16 r6, r6, r8
|
|
usub16 r8, lr, r9
|
|
ldr r7, [r2, #8]
|
|
smlad r0, r6, r6, r0
|
|
uxtb16 lr, r4
|
|
uxtb16 r4, r4, ror #8
|
|
uxtb16 r9, r7
|
|
uxtb16 r7, r7, ror #8
|
|
smlad r0, r8, r8, r0
|
|
ldr r8, [r2, #12]
|
|
usub16 lr, lr, r9
|
|
usub16 r4, r4, r7
|
|
smlad r0, lr, lr, r0
|
|
uxtb16 r6, r5
|
|
uxtb16 r5, r5, ror #8
|
|
uxtb16 r9, r8
|
|
uxtb16 r8, r8, ror #8
|
|
smlad r0, r4, r4, r0
|
|
usub16 r6, r6, r9
|
|
usub16 r5, r5, r8
|
|
smlad r0, r6, r6, r0
|
|
add r1, r1, r3
|
|
add r2, r2, r3
|
|
subs r12, r12, #1
|
|
smlad r0, r5, r5, r0
|
|
bgt 1b
|
|
|
|
pop {r4-r9, pc}
|
|
endfunc
|