mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-28 20:53:54 +02:00
ddea3b7106
And use the x86util ones instead, which are optimized for mmxext/sse2. About ~1% increase in performance on pre SSSE3 processors. Signed-off-by: James Almer <jamrial@gmail.com> Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
256 lines
5.8 KiB
NASM
256 lines
5.8 KiB
NASM
;*****************************************************************************
|
|
;* x86-optimized functions for yadif filter
|
|
;*
|
|
;* Copyright (C) 2006 Michael Niedermayer <michaelni@gmx.at>
|
|
;* Copyright (c) 2013 Daniel Kang <daniel.d.kang@gmail.com>
|
|
;* Copyright (c) 2011-2013 James Darnley <james.darnley@gmail.com>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA
|
|
|
|
pw_1: times 8 dw 1
|
|
|
|
SECTION .text
|
|
|
|
%macro PMAXUW 2
|
|
%if cpuflag(sse4)
|
|
pmaxuw %1, %2
|
|
%else
|
|
psubusw %1, %2
|
|
paddusw %1, %2
|
|
%endif
|
|
%endmacro
|
|
|
|
%macro CHECK 2
|
|
movu m2, [curq+t1+%1*2]
|
|
movu m3, [curq+t0+%2*2]
|
|
mova m4, m2
|
|
mova m5, m2
|
|
pxor m4, m3
|
|
pavgw m5, m3
|
|
pand m4, [pw_1]
|
|
psubusw m5, m4
|
|
RSHIFT m5, 2
|
|
mova m4, m2
|
|
psubusw m2, m3
|
|
psubusw m3, m4
|
|
PMAXUW m2, m3
|
|
mova m3, m2
|
|
mova m4, m2
|
|
RSHIFT m3, 2
|
|
RSHIFT m4, 4
|
|
paddw m2, m3
|
|
paddw m2, m4
|
|
%endmacro
|
|
|
|
%macro CHECK1 0
|
|
mova m3, m0
|
|
pcmpgtw m3, m2
|
|
pminsw m0, m2
|
|
mova m6, m3
|
|
pand m5, m3
|
|
pandn m3, m1
|
|
por m3, m5
|
|
mova m1, m3
|
|
%endmacro
|
|
|
|
; %macro CHECK2 0
|
|
; paddw m6, [pw_1]
|
|
; psllw m6, 14
|
|
; paddsw m2, m6
|
|
; mova m3, m0
|
|
; pcmpgtw m3, m2
|
|
; pminsw m0, m2
|
|
; pand m5, m3
|
|
; pandn m3, m1
|
|
; por m3, m5
|
|
; mova m1, m3
|
|
; %endmacro
|
|
|
|
; This version of CHECK2 is required for 14-bit samples. The left-shift trick
|
|
; in the old code is not large enough to correctly select pixels or scores.
|
|
|
|
%macro CHECK2 0
|
|
mova m3, m0
|
|
pcmpgtw m0, m2
|
|
pand m0, m6
|
|
mova m6, m0
|
|
pand m5, m6
|
|
pand m2, m0
|
|
pandn m6, m1
|
|
pandn m0, m3
|
|
por m6, m5
|
|
por m0, m2
|
|
mova m1, m6
|
|
%endmacro
|
|
|
|
%macro LOAD 2
|
|
movu %1, %2
|
|
%endmacro
|
|
|
|
%macro FILTER 3
|
|
.loop%1:
|
|
pxor m7, m7
|
|
LOAD m0, [curq+t1]
|
|
LOAD m1, [curq+t0]
|
|
LOAD m2, [%2]
|
|
LOAD m3, [%3]
|
|
mova m4, m3
|
|
paddw m3, m2
|
|
psraw m3, 1
|
|
mova [rsp+ 0], m0
|
|
mova [rsp+16], m3
|
|
mova [rsp+32], m1
|
|
psubw m2, m4
|
|
ABS1 m2, m4
|
|
LOAD m3, [prevq+t1]
|
|
LOAD m4, [prevq+t0]
|
|
psubw m3, m0
|
|
psubw m4, m1
|
|
ABS2 m3, m4, m5, m6
|
|
paddw m3, m4
|
|
psrlw m2, 1
|
|
psrlw m3, 1
|
|
pmaxsw m2, m3
|
|
LOAD m3, [nextq+t1]
|
|
LOAD m4, [nextq+t0]
|
|
psubw m3, m0
|
|
psubw m4, m1
|
|
ABS2 m3, m4, m5, m6
|
|
paddw m3, m4
|
|
psrlw m3, 1
|
|
pmaxsw m2, m3
|
|
mova [rsp+48], m2
|
|
|
|
paddw m1, m0
|
|
paddw m0, m0
|
|
psubw m0, m1
|
|
psrlw m1, 1
|
|
ABS1 m0, m2
|
|
|
|
movu m2, [curq+t1-1*2]
|
|
movu m3, [curq+t0-1*2]
|
|
mova m4, m2
|
|
psubusw m2, m3
|
|
psubusw m3, m4
|
|
PMAXUW m2, m3
|
|
mova m3, m2
|
|
RSHIFT m3, 4
|
|
paddw m0, m2
|
|
paddw m0, m3
|
|
psubw m0, [pw_1]
|
|
|
|
CHECK -2, 0
|
|
CHECK1
|
|
CHECK -3, 1
|
|
CHECK2
|
|
CHECK 0, -2
|
|
CHECK1
|
|
CHECK 1, -3
|
|
CHECK2
|
|
|
|
mova m6, [rsp+48]
|
|
cmp DWORD r8m, 2
|
|
jge .end%1
|
|
LOAD m2, [%2+t1*2]
|
|
LOAD m4, [%3+t1*2]
|
|
LOAD m3, [%2+t0*2]
|
|
LOAD m5, [%3+t0*2]
|
|
paddw m2, m4
|
|
paddw m3, m5
|
|
psrlw m2, 1
|
|
psrlw m3, 1
|
|
mova m4, [rsp+ 0]
|
|
mova m5, [rsp+16]
|
|
mova m7, [rsp+32]
|
|
psubw m2, m4
|
|
psubw m3, m7
|
|
mova m0, m5
|
|
psubw m5, m4
|
|
psubw m0, m7
|
|
mova m4, m2
|
|
pminsw m2, m3
|
|
pmaxsw m3, m4
|
|
pmaxsw m2, m5
|
|
pminsw m3, m5
|
|
pmaxsw m2, m0
|
|
pminsw m3, m0
|
|
pxor m4, m4
|
|
pmaxsw m6, m3
|
|
psubw m4, m2
|
|
pmaxsw m6, m4
|
|
|
|
.end%1:
|
|
mova m2, [rsp+16]
|
|
mova m3, m2
|
|
psubw m2, m6
|
|
paddw m3, m6
|
|
pmaxsw m1, m2
|
|
pminsw m1, m3
|
|
|
|
movu [dstq], m1
|
|
add dstq, mmsize-4
|
|
add prevq, mmsize-4
|
|
add curq, mmsize-4
|
|
add nextq, mmsize-4
|
|
sub DWORD r4m, mmsize/2-2
|
|
jg .loop%1
|
|
%endmacro
|
|
|
|
%macro YADIF 0
|
|
%if ARCH_X86_32
|
|
cglobal yadif_filter_line_10bit, 4, 6, 8, 80, dst, prev, cur, next, w, \
|
|
prefs, mrefs, parity, mode
|
|
%else
|
|
cglobal yadif_filter_line_10bit, 4, 7, 8, 80, dst, prev, cur, next, w, \
|
|
prefs, mrefs, parity, mode
|
|
%endif
|
|
%if ARCH_X86_32
|
|
mov r4, r5mp
|
|
mov r5, r6mp
|
|
DECLARE_REG_TMP 4,5
|
|
%else
|
|
movsxd r5, DWORD r5m
|
|
movsxd r6, DWORD r6m
|
|
DECLARE_REG_TMP 5,6
|
|
%endif
|
|
|
|
cmp DWORD paritym, 0
|
|
je .parity0
|
|
FILTER 1, prevq, curq
|
|
jmp .ret
|
|
|
|
.parity0:
|
|
FILTER 0, curq, nextq
|
|
|
|
.ret:
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
YADIF
|
|
INIT_XMM sse2
|
|
YADIF
|
|
%if ARCH_X86_32
|
|
INIT_MMX mmxext
|
|
YADIF
|
|
%endif
|