mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
7fd60d1e7a
* commit 'ac6b95dbc0b53b3ea461bd5e5e7f7f31d2983733': aarch64: add ',' between assembler macro arguments where missing Merged-by: Michael Niedermayer <michaelni@gmx.at>
227 lines
7.3 KiB
ArmAsm
227 lines
7.3 KiB
ArmAsm
/*
|
|
* Copyright (c) 2014 Janne Grunau <janne-libav@jannau.net>
|
|
*
|
|
* This file is part of FFmpeg.
|
|
*
|
|
* FFmpeg is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* FFmpeg is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with FFmpeg; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
|
|
#include "libavutil/aarch64/asm.S"
|
|
|
|
#define FRAC_BITS 23 // fractional bits for sb_samples and dct
|
|
#define WFRAC_BITS 16 // fractional bits for window
|
|
#define OUT_SHIFT (WFRAC_BITS + FRAC_BITS - 15)
|
|
|
|
const tbl_rev128.s, align=4
|
|
.byte 12, 13, 14, 15
|
|
.byte 8, 9, 10, 11
|
|
.byte 4, 5, 6, 7
|
|
.byte 0, 1, 2, 3
|
|
endconst
|
|
|
|
.macro apply_window type, st
|
|
function ff_mpadsp_apply_window_\type\()_neon, export=1
|
|
mov x7, x0
|
|
sxtw x4, w4 // incr
|
|
add x8, x0, #512<<2
|
|
ld1 {v0.4s,v1.4s,v2.4s,v3.4s}, [x7], #64
|
|
ld1 {v4.4s,v5.4s,v6.4s,v7.4s}, [x7], #64
|
|
st1 {v0.4s,v1.4s,v2.4s,v3.4s}, [x8], #64
|
|
st1 {v4.4s,v5.4s,v6.4s,v7.4s}, [x8], #64
|
|
movrel x15, tbl_rev128.s
|
|
ld1 {v27.4s}, [x15]
|
|
.ifc \type, fixed
|
|
lsl x4, x4, #1
|
|
.else
|
|
lsl x4, x4, #2
|
|
.endif
|
|
add x10, x0, #45<<2
|
|
add x0, x0, #16<<2
|
|
add x1, x1, #16<<2
|
|
add x5, x3, x4, lsl #5
|
|
sub x5, x5, x4 // samples2
|
|
neg x13, x4 // -incr
|
|
mov x9, #64<<2
|
|
.ifc \type, fixed
|
|
ld1r {v16.2s}, [x2] // dither_state
|
|
sxtl v16.2d, v16.2s
|
|
movi v29.2d, #0
|
|
movi v30.2d, #(1<<OUT_SHIFT)-1
|
|
trn1 v31.2d, v29.2d, v30.2d
|
|
trn2 v30.2d, v30.2d, v29.2d
|
|
trn1 v16.2d, v16.2d, v29.2d
|
|
.else
|
|
movi v16.4s, #0
|
|
movi v28.4s, #0
|
|
.endif
|
|
mov x14, #4
|
|
1:
|
|
mov x8, x0
|
|
sub x7, x1, #3<<2
|
|
sub x6, x1, x14, lsl #4
|
|
add x7, x7, x14, lsl #4
|
|
add x11, x6, #(32)<<2 // w + 32
|
|
add x12, x7, #(32)<<2 // w2 + 32
|
|
mov x15, #8
|
|
movi v17.2d, #0
|
|
movi v18.2d, #0
|
|
movi v19.2d, #0
|
|
2:
|
|
subs x15, x15, #1
|
|
ld1 {v0.4s}, [x8], x9
|
|
ld1 {v1.4s}, [x10], x9
|
|
ld1 {v2.4s}, [x6], x9
|
|
ld1 {v3.4s}, [x7], x9
|
|
tbl v6.16b, {v0.16b}, v27.16b
|
|
tbl v7.16b, {v1.16b}, v27.16b
|
|
ld1 {v4.4s}, [x11], x9
|
|
ld1 {v5.4s}, [x12], x9
|
|
MLA v16, v2, v0
|
|
MLA2 v17, v2, v0
|
|
MLS v18, v3, v6
|
|
MLS2 v19, v3, v6
|
|
MLS v16, v4, v7
|
|
MLS2 v17, v4, v7
|
|
MLS v18, v5, v1
|
|
MLS2 v19, v5, v1
|
|
b.gt 2b
|
|
|
|
cmp x14, #4
|
|
sub x10, x10, #64<<5 // 64 * 8 * sizeof(int32_t)
|
|
|
|
.ifc \type, fixed
|
|
and v28.16b, v16.16b, v30.16b
|
|
ext v28.16b, v29.16b, v28.16b, #8
|
|
|
|
b.eq 4f
|
|
round_sample v19, 1, 1
|
|
4:
|
|
round_sample v16, 1, 0
|
|
shrn v16.2s, v16.2d, #OUT_SHIFT
|
|
round_sample v19, 0, 0
|
|
shrn v19.2s, v19.2d, #OUT_SHIFT
|
|
round_sample v17, 0, 1
|
|
round_sample v18, 1, 1
|
|
round_sample v17, 1, 0
|
|
shrn2 v16.4s, v17.2d, #OUT_SHIFT
|
|
round_sample v18, 0, 0
|
|
shrn2 v19.4s, v18.2d, #OUT_SHIFT
|
|
sqxtn v16.4h, v16.4s
|
|
sqxtn v18.4h, v19.4s
|
|
.else
|
|
ext v18.16b, v18.16b, v18.16b, #8
|
|
.endif
|
|
|
|
st1 {v16.\st\()}[0], [x3], x4
|
|
b.eq 4f
|
|
st1 {v18.\st\()}[1], [x5], x13
|
|
4:
|
|
st1 {v16.\st\()}[1], [x3], x4
|
|
st1 {v18.\st\()}[0], [x5], x13
|
|
st1 {v16.\st\()}[2], [x3], x4
|
|
st1 {v18.\st\()}[3], [x5], x13
|
|
st1 {v16.\st\()}[3], [x3], x4
|
|
st1 {v18.\st\()}[2], [x5], x13
|
|
|
|
mov v16.16b, v28.16b
|
|
|
|
subs x14, x14, #1
|
|
add x0, x0, #4<<2
|
|
sub x10, x10, #4<<2
|
|
b.gt 1b
|
|
|
|
// comuting samples[16]
|
|
add x6, x1, #32<<2
|
|
ld1 {v0.2s}, [x6], x9
|
|
ld1 {v1.2s}, [x0], x9
|
|
.rept 3
|
|
ld1 {v2.2s}, [x6], x9
|
|
ld1 {v3.2s}, [x0], x9
|
|
MLS v16, v0, v1
|
|
ld1 {v0.2s}, [x6], x9
|
|
ld1 {v1.2s}, [x0], x9
|
|
MLS v16, v2, v3
|
|
.endr
|
|
ld1 {v2.2s}, [x6], x9
|
|
ld1 {v3.2s}, [x0], x9
|
|
MLS v16, v0, v1
|
|
MLS v16, v2, v3
|
|
|
|
.ifc \type, fixed
|
|
and v28.16b, v16.16b, v30.16b
|
|
shrn v20.2s, v16.2d, #OUT_SHIFT
|
|
xtn v28.2s, v28.2d
|
|
sqxtn v20.4h, v20.4s
|
|
st1 {v28.s}[0], [x2] // save dither_state
|
|
st1 {v20.h}[0], [x3]
|
|
.else
|
|
st1 {v16.s}[0], [x3]
|
|
.endif
|
|
|
|
ret
|
|
endfunc
|
|
.purgem round_sample
|
|
.purgem MLA
|
|
.purgem MLA2
|
|
.purgem MLS
|
|
.purgem MLS2
|
|
.endm
|
|
|
|
|
|
.macro round_sample r, idx, next
|
|
add \r\().2d, \r\().2d, v28.2d
|
|
.if \idx == 0
|
|
and v28.16b, \r\().16b, v30.16b
|
|
.else // \idx == 1
|
|
and v28.16b, \r\().16b, v31.16b
|
|
.endif
|
|
.if \idx != \next
|
|
.if \next == 0
|
|
ext v28.16b, v28.16b, v29.16b, #8
|
|
.else
|
|
ext v28.16b, v29.16b, v28.16b, #8
|
|
.endif
|
|
.endif
|
|
.endm
|
|
.macro MLA d, s1, s2
|
|
smlal \d\().2d, \s1\().2s, \s2\().2s
|
|
.endm
|
|
.macro MLA2 d, s1, s2
|
|
smlal2 \d\().2d, \s1\().4s, \s2\().4s
|
|
.endm
|
|
.macro MLS d, s1, s2
|
|
smlsl \d\().2d, \s1\().2s, \s2\().2s
|
|
.endm
|
|
.macro MLS2 d, s1, s2
|
|
smlsl2 \d\().2d, \s1\().4s, \s2\().4s
|
|
.endm
|
|
apply_window fixed, h
|
|
|
|
|
|
// nothing to do for round_sample and ML{A,S}2
|
|
.macro round_sample r, idx, next
|
|
.endm
|
|
.macro MLA2 d, s1, s2
|
|
.endm
|
|
.macro MLS2 d, s1, s2
|
|
.endm
|
|
.macro MLA d, s1, s2
|
|
fmla \d\().4s, \s1\().4s, \s2\().4s
|
|
.endm
|
|
.macro MLS d, s1, s2
|
|
fmls \d\().4s, \s1\().4s, \s2\().4s
|
|
.endm
|
|
apply_window float, s
|