mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-12 19:18:44 +02:00
4011a76494
x64 always has MMX, MMXEXT, SSE and SSE2 and this means that some functions for MMX, MMXEXT and 3dnow are always overridden by other functions (unless one e.g. explicitly disables SSE2) for x64. So given that the only systems that benefit from these functions are truely ancient 32bit x86s they are removed. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
855 lines
20 KiB
NASM
855 lines
20 KiB
NASM
;*****************************************************************************
|
|
;* MMX/SSE2/SSSE3-optimized H.264 QPEL code
|
|
;*****************************************************************************
|
|
;* Copyright (c) 2004-2005 Michael Niedermayer, Loren Merritt
|
|
;* Copyright (C) 2012 Daniel Kang
|
|
;*
|
|
;* Authors: Daniel Kang <daniel.d.kang@gmail.com>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA 32
|
|
|
|
cextern pw_16
|
|
cextern pw_5
|
|
cextern pb_0
|
|
|
|
SECTION .text
|
|
|
|
|
|
%macro op_avgh 3
|
|
movh %3, %2
|
|
pavgb %1, %3
|
|
movh %2, %1
|
|
%endmacro
|
|
|
|
%macro op_avg 2-3
|
|
pavgb %1, %2
|
|
mova %2, %1
|
|
%endmacro
|
|
|
|
%macro op_puth 2-3
|
|
movh %2, %1
|
|
%endmacro
|
|
|
|
%macro op_put 2-3
|
|
mova %2, %1
|
|
%endmacro
|
|
|
|
%macro QPEL4_H_LOWPASS_OP 1
|
|
cglobal %1_h264_qpel4_h_lowpass, 4,5 ; dst, src, dstStride, srcStride
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
pxor m7, m7
|
|
mova m4, [pw_5]
|
|
mova m5, [pw_16]
|
|
mov r4d, 4
|
|
.loop:
|
|
movh m1, [r1-1]
|
|
movh m2, [r1+0]
|
|
movh m3, [r1+1]
|
|
movh m0, [r1+2]
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m0, m7
|
|
paddw m1, m0
|
|
paddw m2, m3
|
|
movh m0, [r1-2]
|
|
movh m3, [r1+3]
|
|
punpcklbw m0, m7
|
|
punpcklbw m3, m7
|
|
paddw m0, m3
|
|
psllw m2, 2
|
|
psubw m2, m1
|
|
pmullw m2, m4
|
|
paddw m0, m5
|
|
paddw m0, m2
|
|
psraw m0, 5
|
|
packuswb m0, m0
|
|
op_%1h m0, [r0], m6
|
|
add r0, r2
|
|
add r1, r3
|
|
dec r4d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL4_H_LOWPASS_OP put
|
|
QPEL4_H_LOWPASS_OP avg
|
|
|
|
%macro QPEL8_H_LOWPASS_OP 1
|
|
cglobal %1_h264_qpel8_h_lowpass, 4,5 ; dst, src, dstStride, srcStride
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
mov r4d, 8
|
|
pxor m7, m7
|
|
mova m6, [pw_5]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+1]
|
|
mova m1, m0
|
|
mova m3, m2
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
paddw m0, m2
|
|
paddw m1, m3
|
|
psllw m0, 2
|
|
psllw m1, 2
|
|
mova m2, [r1-1]
|
|
mova m4, [r1+2]
|
|
mova m3, m2
|
|
mova m5, m4
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
punpcklbw m4, m7
|
|
punpckhbw m5, m7
|
|
paddw m2, m4
|
|
paddw m5, m3
|
|
psubw m0, m2
|
|
psubw m1, m5
|
|
pmullw m0, m6
|
|
pmullw m1, m6
|
|
movd m2, [r1-2]
|
|
movd m5, [r1+7]
|
|
punpcklbw m2, m7
|
|
punpcklbw m5, m7
|
|
paddw m2, m3
|
|
paddw m4, m5
|
|
mova m5, [pw_16]
|
|
paddw m2, m5
|
|
paddw m4, m5
|
|
paddw m0, m2
|
|
paddw m1, m4
|
|
psraw m0, 5
|
|
psraw m1, 5
|
|
packuswb m0, m1
|
|
op_%1 m0, [r0], m4
|
|
add r0, r2
|
|
add r1, r3
|
|
dec r4d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL8_H_LOWPASS_OP put
|
|
QPEL8_H_LOWPASS_OP avg
|
|
|
|
%macro QPEL8_H_LOWPASS_OP_XMM 1
|
|
cglobal %1_h264_qpel8_h_lowpass, 4,5,8 ; dst, src, dstStride, srcStride
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
mov r4d, 8
|
|
pxor m7, m7
|
|
mova m6, [pw_5]
|
|
.loop:
|
|
movu m1, [r1-2]
|
|
mova m0, m1
|
|
punpckhbw m1, m7
|
|
punpcklbw m0, m7
|
|
mova m2, m1
|
|
mova m3, m1
|
|
mova m4, m1
|
|
mova m5, m1
|
|
palignr m4, m0, 2
|
|
palignr m3, m0, 4
|
|
palignr m2, m0, 6
|
|
palignr m1, m0, 8
|
|
palignr m5, m0, 10
|
|
paddw m0, m5
|
|
paddw m2, m3
|
|
paddw m1, m4
|
|
psllw m2, 2
|
|
psubw m2, m1
|
|
paddw m0, [pw_16]
|
|
pmullw m2, m6
|
|
paddw m2, m0
|
|
psraw m2, 5
|
|
packuswb m2, m2
|
|
op_%1h m2, [r0], m4
|
|
add r1, r3
|
|
add r0, r2
|
|
dec r4d
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
QPEL8_H_LOWPASS_OP_XMM put
|
|
QPEL8_H_LOWPASS_OP_XMM avg
|
|
|
|
|
|
%macro QPEL4_H_LOWPASS_L2_OP 1
|
|
cglobal %1_h264_qpel4_h_lowpass_l2, 5,6 ; dst, src, src2, dstStride, srcStride
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
pxor m7, m7
|
|
mova m4, [pw_5]
|
|
mova m5, [pw_16]
|
|
mov r5d, 4
|
|
.loop:
|
|
movh m1, [r1-1]
|
|
movh m2, [r1+0]
|
|
movh m3, [r1+1]
|
|
movh m0, [r1+2]
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m0, m7
|
|
paddw m1, m0
|
|
paddw m2, m3
|
|
movh m0, [r1-2]
|
|
movh m3, [r1+3]
|
|
punpcklbw m0, m7
|
|
punpcklbw m3, m7
|
|
paddw m0, m3
|
|
psllw m2, 2
|
|
psubw m2, m1
|
|
pmullw m2, m4
|
|
paddw m0, m5
|
|
paddw m0, m2
|
|
movh m3, [r2]
|
|
psraw m0, 5
|
|
packuswb m0, m0
|
|
pavgb m0, m3
|
|
op_%1h m0, [r0], m6
|
|
add r0, r3
|
|
add r1, r3
|
|
add r2, r4
|
|
dec r5d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL4_H_LOWPASS_L2_OP put
|
|
QPEL4_H_LOWPASS_L2_OP avg
|
|
|
|
|
|
%macro QPEL8_H_LOWPASS_L2_OP 1
|
|
cglobal %1_h264_qpel8_h_lowpass_l2, 5,6 ; dst, src, src2, dstStride, srcStride
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
mov r5d, 8
|
|
pxor m7, m7
|
|
mova m6, [pw_5]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+1]
|
|
mova m1, m0
|
|
mova m3, m2
|
|
punpcklbw m0, m7
|
|
punpckhbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
paddw m0, m2
|
|
paddw m1, m3
|
|
psllw m0, 2
|
|
psllw m1, 2
|
|
mova m2, [r1-1]
|
|
mova m4, [r1+2]
|
|
mova m3, m2
|
|
mova m5, m4
|
|
punpcklbw m2, m7
|
|
punpckhbw m3, m7
|
|
punpcklbw m4, m7
|
|
punpckhbw m5, m7
|
|
paddw m2, m4
|
|
paddw m5, m3
|
|
psubw m0, m2
|
|
psubw m1, m5
|
|
pmullw m0, m6
|
|
pmullw m1, m6
|
|
movd m2, [r1-2]
|
|
movd m5, [r1+7]
|
|
punpcklbw m2, m7
|
|
punpcklbw m5, m7
|
|
paddw m2, m3
|
|
paddw m4, m5
|
|
mova m5, [pw_16]
|
|
paddw m2, m5
|
|
paddw m4, m5
|
|
paddw m0, m2
|
|
paddw m1, m4
|
|
psraw m0, 5
|
|
psraw m1, 5
|
|
mova m4, [r2]
|
|
packuswb m0, m1
|
|
pavgb m0, m4
|
|
op_%1 m0, [r0], m4
|
|
add r0, r3
|
|
add r1, r3
|
|
add r2, r4
|
|
dec r5d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL8_H_LOWPASS_L2_OP put
|
|
QPEL8_H_LOWPASS_L2_OP avg
|
|
|
|
|
|
%macro QPEL8_H_LOWPASS_L2_OP_XMM 1
|
|
cglobal %1_h264_qpel8_h_lowpass_l2, 5,6,8 ; dst, src, src2, dstStride, src2Stride
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
mov r5d, 8
|
|
pxor m7, m7
|
|
mova m6, [pw_5]
|
|
.loop:
|
|
lddqu m1, [r1-2]
|
|
mova m0, m1
|
|
punpckhbw m1, m7
|
|
punpcklbw m0, m7
|
|
mova m2, m1
|
|
mova m3, m1
|
|
mova m4, m1
|
|
mova m5, m1
|
|
palignr m4, m0, 2
|
|
palignr m3, m0, 4
|
|
palignr m2, m0, 6
|
|
palignr m1, m0, 8
|
|
palignr m5, m0, 10
|
|
paddw m0, m5
|
|
paddw m2, m3
|
|
paddw m1, m4
|
|
psllw m2, 2
|
|
movh m3, [r2]
|
|
psubw m2, m1
|
|
paddw m0, [pw_16]
|
|
pmullw m2, m6
|
|
paddw m2, m0
|
|
psraw m2, 5
|
|
packuswb m2, m2
|
|
pavgb m2, m3
|
|
op_%1h m2, [r0], m4
|
|
add r1, r3
|
|
add r0, r3
|
|
add r2, r4
|
|
dec r5d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
QPEL8_H_LOWPASS_L2_OP_XMM put
|
|
QPEL8_H_LOWPASS_L2_OP_XMM avg
|
|
|
|
|
|
; All functions that call this are required to have function arguments of
|
|
; dst, src, dstStride, srcStride
|
|
%macro FILT_V 1
|
|
mova m6, m2
|
|
movh m5, [r1]
|
|
paddw m6, m3
|
|
psllw m6, 2
|
|
psubw m6, m1
|
|
psubw m6, m4
|
|
punpcklbw m5, m7
|
|
pmullw m6, [pw_5]
|
|
paddw m0, [pw_16]
|
|
add r1, r3
|
|
paddw m0, m5
|
|
paddw m6, m0
|
|
psraw m6, 5
|
|
packuswb m6, m6
|
|
op_%1h m6, [r0], m0 ; 1
|
|
add r0, r2
|
|
SWAP 0, 1, 2, 3, 4, 5
|
|
%endmacro
|
|
|
|
%macro QPEL4_V_LOWPASS_OP 1
|
|
cglobal %1_h264_qpel4_v_lowpass, 4,4 ; dst, src, dstStride, srcStride
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
sub r1, r3
|
|
sub r1, r3
|
|
pxor m7, m7
|
|
movh m0, [r1]
|
|
movh m1, [r1+r3]
|
|
lea r1, [r1+2*r3]
|
|
movh m2, [r1]
|
|
movh m3, [r1+r3]
|
|
lea r1, [r1+2*r3]
|
|
movh m4, [r1]
|
|
add r1, r3
|
|
punpcklbw m0, m7
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m4, m7
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL4_V_LOWPASS_OP put
|
|
QPEL4_V_LOWPASS_OP avg
|
|
|
|
|
|
|
|
%macro QPEL8OR16_V_LOWPASS_OP 1
|
|
%if cpuflag(sse2)
|
|
cglobal %1_h264_qpel8or16_v_lowpass, 5,5,8 ; dst, src, dstStride, srcStride, h
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
sub r1, r3
|
|
sub r1, r3
|
|
%else
|
|
cglobal %1_h264_qpel8or16_v_lowpass_op, 5,5,8 ; dst, src, dstStride, srcStride, h
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
%endif
|
|
pxor m7, m7
|
|
movh m0, [r1]
|
|
movh m1, [r1+r3]
|
|
lea r1, [r1+2*r3]
|
|
movh m2, [r1]
|
|
movh m3, [r1+r3]
|
|
lea r1, [r1+2*r3]
|
|
movh m4, [r1]
|
|
add r1, r3
|
|
punpcklbw m0, m7
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m4, m7
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
cmp r4d, 16
|
|
jne .end
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
FILT_V %1
|
|
.end:
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
QPEL8OR16_V_LOWPASS_OP put
|
|
QPEL8OR16_V_LOWPASS_OP avg
|
|
|
|
|
|
; All functions that use this are required to have args:
|
|
; src, tmp, srcSize
|
|
%macro FILT_HV 1 ; offset
|
|
mova m6, m2
|
|
movh m5, [r0]
|
|
paddw m6, m3
|
|
psllw m6, 2
|
|
paddw m0, [pw_16]
|
|
psubw m6, m1
|
|
psubw m6, m4
|
|
punpcklbw m5, m7
|
|
pmullw m6, [pw_5]
|
|
paddw m0, m5
|
|
add r0, r2
|
|
paddw m6, m0
|
|
mova [r1+%1], m6
|
|
SWAP 0, 1, 2, 3, 4, 5
|
|
%endmacro
|
|
|
|
%macro QPEL4_HV1_LOWPASS_OP 1
|
|
cglobal %1_h264_qpel4_hv_lowpass_v, 3,3 ; src, tmp, srcStride
|
|
movsxdifnidn r2, r2d
|
|
pxor m7, m7
|
|
movh m0, [r0]
|
|
movh m1, [r0+r2]
|
|
lea r0, [r0+2*r2]
|
|
movh m2, [r0]
|
|
movh m3, [r0+r2]
|
|
lea r0, [r0+2*r2]
|
|
movh m4, [r0]
|
|
add r0, r2
|
|
punpcklbw m0, m7
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m4, m7
|
|
FILT_HV 0*24
|
|
FILT_HV 1*24
|
|
FILT_HV 2*24
|
|
FILT_HV 3*24
|
|
RET
|
|
|
|
cglobal %1_h264_qpel4_hv_lowpass_h, 3,4 ; tmp, dst, dstStride
|
|
movsxdifnidn r2, r2d
|
|
mov r3d, 4
|
|
.loop:
|
|
mova m0, [r0]
|
|
paddw m0, [r0+10]
|
|
mova m1, [r0+2]
|
|
paddw m1, [r0+8]
|
|
mova m2, [r0+4]
|
|
paddw m2, [r0+6]
|
|
psubw m0, m1
|
|
psraw m0, 2
|
|
psubw m0, m1
|
|
paddsw m0, m2
|
|
psraw m0, 2
|
|
paddw m0, m2
|
|
psraw m0, 6
|
|
packuswb m0, m0
|
|
op_%1h m0, [r1], m7
|
|
add r0, 24
|
|
add r1, r2
|
|
dec r3d
|
|
jnz .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL4_HV1_LOWPASS_OP put
|
|
QPEL4_HV1_LOWPASS_OP avg
|
|
|
|
%macro QPEL8OR16_HV1_LOWPASS_OP 1
|
|
cglobal %1_h264_qpel8or16_hv1_lowpass_op, 4,4,8 ; src, tmp, srcStride, size
|
|
movsxdifnidn r2, r2d
|
|
pxor m7, m7
|
|
movh m0, [r0]
|
|
movh m1, [r0+r2]
|
|
lea r0, [r0+2*r2]
|
|
movh m2, [r0]
|
|
movh m3, [r0+r2]
|
|
lea r0, [r0+2*r2]
|
|
movh m4, [r0]
|
|
add r0, r2
|
|
punpcklbw m0, m7
|
|
punpcklbw m1, m7
|
|
punpcklbw m2, m7
|
|
punpcklbw m3, m7
|
|
punpcklbw m4, m7
|
|
FILT_HV 0*48
|
|
FILT_HV 1*48
|
|
FILT_HV 2*48
|
|
FILT_HV 3*48
|
|
FILT_HV 4*48
|
|
FILT_HV 5*48
|
|
FILT_HV 6*48
|
|
FILT_HV 7*48
|
|
cmp r3d, 16
|
|
jne .end
|
|
FILT_HV 8*48
|
|
FILT_HV 9*48
|
|
FILT_HV 10*48
|
|
FILT_HV 11*48
|
|
FILT_HV 12*48
|
|
FILT_HV 13*48
|
|
FILT_HV 14*48
|
|
FILT_HV 15*48
|
|
.end:
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
QPEL8OR16_HV1_LOWPASS_OP put
|
|
|
|
|
|
|
|
%macro QPEL8OR16_HV2_LOWPASS_OP 1
|
|
; unused is to match ssse3 and mmxext args
|
|
cglobal %1_h264_qpel8or16_hv2_lowpass_op, 5,5 ; dst, tmp, dstStride, unused, h
|
|
movsxdifnidn r2, r2d
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m3, [r1+8]
|
|
mova m1, [r1+2]
|
|
mova m4, [r1+10]
|
|
paddw m0, m4
|
|
paddw m1, m3
|
|
paddw m3, [r1+18]
|
|
paddw m4, [r1+16]
|
|
mova m2, [r1+4]
|
|
mova m5, [r1+12]
|
|
paddw m2, [r1+6]
|
|
paddw m5, [r1+14]
|
|
psubw m0, m1
|
|
psubw m3, m4
|
|
psraw m0, 2
|
|
psraw m3, 2
|
|
psubw m0, m1
|
|
psubw m3, m4
|
|
paddsw m0, m2
|
|
paddsw m3, m5
|
|
psraw m0, 2
|
|
psraw m3, 2
|
|
paddw m0, m2
|
|
paddw m3, m5
|
|
psraw m0, 6
|
|
psraw m3, 6
|
|
packuswb m0, m3
|
|
op_%1 m0, [r0], m7
|
|
add r1, 48
|
|
add r0, r2
|
|
dec r4d
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
QPEL8OR16_HV2_LOWPASS_OP put
|
|
QPEL8OR16_HV2_LOWPASS_OP avg
|
|
|
|
%macro QPEL8OR16_HV2_LOWPASS_OP_XMM 1
|
|
cglobal %1_h264_qpel8or16_hv2_lowpass, 5,5,8 ; dst, tmp, dstStride, tmpStride, size
|
|
movsxdifnidn r2, r2d
|
|
movsxdifnidn r3, r3d
|
|
cmp r4d, 16
|
|
je .op16
|
|
.loop8:
|
|
mova m1, [r1+16]
|
|
mova m0, [r1]
|
|
mova m2, m1
|
|
mova m3, m1
|
|
mova m4, m1
|
|
mova m5, m1
|
|
palignr m5, m0, 10
|
|
palignr m4, m0, 8
|
|
palignr m3, m0, 6
|
|
palignr m2, m0, 4
|
|
palignr m1, m0, 2
|
|
paddw m0, m5
|
|
paddw m1, m4
|
|
paddw m2, m3
|
|
psubw m0, m1
|
|
psraw m0, 2
|
|
psubw m0, m1
|
|
paddw m0, m2
|
|
psraw m0, 2
|
|
paddw m0, m2
|
|
psraw m0, 6
|
|
packuswb m0, m0
|
|
op_%1h m0, [r0], m7
|
|
add r1, 48
|
|
add r0, r2
|
|
dec r4d
|
|
jne .loop8
|
|
jmp .done
|
|
.op16:
|
|
mova m4, [r1+32]
|
|
mova m5, [r1+16]
|
|
mova m7, [r1]
|
|
mova m3, m4
|
|
mova m2, m4
|
|
mova m1, m4
|
|
mova m0, m4
|
|
palignr m0, m5, 10
|
|
palignr m1, m5, 8
|
|
palignr m2, m5, 6
|
|
palignr m3, m5, 4
|
|
palignr m4, m5, 2
|
|
paddw m0, m5
|
|
paddw m1, m4
|
|
paddw m2, m3
|
|
mova m6, m5
|
|
mova m4, m5
|
|
mova m3, m5
|
|
palignr m4, m7, 8
|
|
palignr m6, m7, 2
|
|
palignr m3, m7, 10
|
|
paddw m4, m6
|
|
mova m6, m5
|
|
palignr m5, m7, 6
|
|
palignr m6, m7, 4
|
|
paddw m3, m7
|
|
paddw m5, m6
|
|
psubw m0, m1
|
|
psubw m3, m4
|
|
psraw m0, 2
|
|
psraw m3, 2
|
|
psubw m0, m1
|
|
psubw m3, m4
|
|
paddw m0, m2
|
|
paddw m3, m5
|
|
psraw m0, 2
|
|
psraw m3, 2
|
|
paddw m0, m2
|
|
paddw m3, m5
|
|
psraw m0, 6
|
|
psraw m3, 6
|
|
packuswb m3, m0
|
|
op_%1 m3, [r0], m7
|
|
add r1, 48
|
|
add r0, r2
|
|
dec r4d
|
|
jne .op16
|
|
.done:
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
QPEL8OR16_HV2_LOWPASS_OP_XMM put
|
|
QPEL8OR16_HV2_LOWPASS_OP_XMM avg
|
|
|
|
|
|
%macro PIXELS4_L2_SHIFT5 1
|
|
cglobal %1_pixels4_l2_shift5,6,6 ; dst, src16, src8, dstStride, src8Stride, h
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
mova m0, [r1]
|
|
mova m1, [r1+24]
|
|
psraw m0, 5
|
|
psraw m1, 5
|
|
packuswb m0, m0
|
|
packuswb m1, m1
|
|
pavgb m0, [r2]
|
|
pavgb m1, [r2+r4]
|
|
op_%1h m0, [r0], m4
|
|
op_%1h m1, [r0+r3], m5
|
|
lea r2, [r2+r4*2]
|
|
lea r0, [r0+r3*2]
|
|
mova m0, [r1+48]
|
|
mova m1, [r1+72]
|
|
psraw m0, 5
|
|
psraw m1, 5
|
|
packuswb m0, m0
|
|
packuswb m1, m1
|
|
pavgb m0, [r2]
|
|
pavgb m1, [r2+r4]
|
|
op_%1h m0, [r0], m4
|
|
op_%1h m1, [r0+r3], m5
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PIXELS4_L2_SHIFT5 put
|
|
PIXELS4_L2_SHIFT5 avg
|
|
|
|
|
|
%macro PIXELS8_L2_SHIFT5 1
|
|
cglobal %1_pixels8_l2_shift5, 6, 6 ; dst, src16, src8, dstStride, src8Stride, h
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, [r1+8]
|
|
mova m2, [r1+48]
|
|
mova m3, [r1+48+8]
|
|
psraw m0, 5
|
|
psraw m1, 5
|
|
psraw m2, 5
|
|
psraw m3, 5
|
|
packuswb m0, m1
|
|
packuswb m2, m3
|
|
pavgb m0, [r2]
|
|
pavgb m2, [r2+r4]
|
|
op_%1 m0, [r0], m4
|
|
op_%1 m2, [r0+r3], m5
|
|
lea r2, [r2+2*r4]
|
|
add r1, 48*2
|
|
lea r0, [r0+2*r3]
|
|
sub r5d, 2
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PIXELS8_L2_SHIFT5 put
|
|
PIXELS8_L2_SHIFT5 avg
|
|
|
|
|
|
%if ARCH_X86_64
|
|
%macro QPEL16_H_LOWPASS_L2_OP 1
|
|
cglobal %1_h264_qpel16_h_lowpass_l2, 5, 6, 16 ; dst, src, src2, dstStride, src2Stride
|
|
movsxdifnidn r3, r3d
|
|
movsxdifnidn r4, r4d
|
|
mov r5d, 16
|
|
pxor m15, m15
|
|
mova m14, [pw_5]
|
|
mova m13, [pw_16]
|
|
.loop:
|
|
lddqu m1, [r1+6]
|
|
lddqu m7, [r1-2]
|
|
mova m0, m1
|
|
punpckhbw m1, m15
|
|
punpcklbw m0, m15
|
|
punpcklbw m7, m15
|
|
mova m2, m1
|
|
mova m6, m0
|
|
mova m3, m1
|
|
mova m8, m0
|
|
mova m4, m1
|
|
mova m9, m0
|
|
mova m12, m0
|
|
mova m11, m1
|
|
palignr m11, m0, 10
|
|
palignr m12, m7, 10
|
|
palignr m4, m0, 2
|
|
palignr m9, m7, 2
|
|
palignr m3, m0, 4
|
|
palignr m8, m7, 4
|
|
palignr m2, m0, 6
|
|
palignr m6, m7, 6
|
|
paddw m11, m0
|
|
palignr m1, m0, 8
|
|
palignr m0, m7, 8
|
|
paddw m7, m12
|
|
paddw m2, m3
|
|
paddw m6, m8
|
|
paddw m1, m4
|
|
paddw m0, m9
|
|
psllw m2, 2
|
|
psllw m6, 2
|
|
psubw m2, m1
|
|
psubw m6, m0
|
|
paddw m11, m13
|
|
paddw m7, m13
|
|
pmullw m2, m14
|
|
pmullw m6, m14
|
|
lddqu m3, [r2]
|
|
paddw m2, m11
|
|
paddw m6, m7
|
|
psraw m2, 5
|
|
psraw m6, 5
|
|
packuswb m6, m2
|
|
pavgb m6, m3
|
|
op_%1 m6, [r0], m11
|
|
add r1, r3
|
|
add r0, r3
|
|
add r2, r4
|
|
dec r5d
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_XMM ssse3
|
|
QPEL16_H_LOWPASS_L2_OP put
|
|
QPEL16_H_LOWPASS_L2_OP avg
|
|
%endif
|