mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-12 19:18:44 +02:00
ea043cc53e
x64 always has MMX, MMXEXT, SSE and SSE2 and this means that some functions for MMX, MMXEXT, SSE and 3dnow are always overridden by other functions (unless one e.g. explicitly disables SSE2). So given that the only systems which benefit from the 8x8 MMX (overridden by MMXEXT) or the 16x16 MMXEXT (overridden by SSE2) are truely ancient 32bit x86s they are removed. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
329 lines
9.6 KiB
NASM
329 lines
9.6 KiB
NASM
;******************************************************************************
|
|
;* Pixel utilities SIMD
|
|
;*
|
|
;* Copyright (C) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
|
|
;* Copyright (C) 2014 Clément Bœsch <u pkh me>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION .text
|
|
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_8x8_mmxext(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
INIT_MMX mmxext
|
|
cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2
|
|
pxor m2, m2
|
|
%rep 4
|
|
mova m0, [src1q]
|
|
mova m1, [src1q + stride1q]
|
|
psadbw m0, [src2q]
|
|
psadbw m1, [src2q + stride2q]
|
|
paddw m2, m0
|
|
paddw m2, m1
|
|
lea src1q, [src1q + 2*stride1q]
|
|
lea src2q, [src2q + 2*stride2q]
|
|
%endrep
|
|
movd eax, m2
|
|
RET
|
|
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
INIT_XMM sse2
|
|
cglobal pixelutils_sad_16x16, 4,4,5, src1, stride1, src2, stride2
|
|
movu m4, [src1q]
|
|
movu m2, [src2q]
|
|
movu m1, [src1q + stride1q]
|
|
movu m3, [src2q + stride2q]
|
|
psadbw m4, m2
|
|
psadbw m1, m3
|
|
paddw m4, m1
|
|
%rep 7
|
|
lea src1q, [src1q + 2*stride1q]
|
|
lea src2q, [src2q + 2*stride2q]
|
|
movu m0, [src1q]
|
|
movu m2, [src2q]
|
|
movu m1, [src1q + stride1q]
|
|
movu m3, [src2q + stride2q]
|
|
psadbw m0, m2
|
|
psadbw m1, m3
|
|
paddw m4, m0
|
|
paddw m4, m1
|
|
%endrep
|
|
movhlps m0, m4
|
|
paddw m4, m0
|
|
movd eax, m4
|
|
RET
|
|
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_[au]_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
%macro SAD_XMM_16x16 1
|
|
INIT_XMM sse2
|
|
cglobal pixelutils_sad_%1_16x16, 4,4,3, src1, stride1, src2, stride2
|
|
mov%1 m2, [src2q]
|
|
psadbw m2, [src1q]
|
|
mov%1 m1, [src2q + stride2q]
|
|
psadbw m1, [src1q + stride1q]
|
|
paddw m2, m1
|
|
%rep 7
|
|
lea src1q, [src1q + 2*stride1q]
|
|
lea src2q, [src2q + 2*stride2q]
|
|
mov%1 m0, [src2q]
|
|
psadbw m0, [src1q]
|
|
mov%1 m1, [src2q + stride2q]
|
|
psadbw m1, [src1q + stride1q]
|
|
paddw m2, m0
|
|
paddw m2, m1
|
|
%endrep
|
|
movhlps m0, m2
|
|
paddw m2, m0
|
|
movd eax, m2
|
|
RET
|
|
%endmacro
|
|
|
|
SAD_XMM_16x16 a
|
|
SAD_XMM_16x16 u
|
|
|
|
|
|
%macro PROCESS_SAD_32x4_U 0
|
|
movu m1, [r2]
|
|
movu m2, [r2 + 16]
|
|
movu m3, [r0]
|
|
movu m4, [r0 + 16]
|
|
psadbw m1, m3
|
|
psadbw m2, m4
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
movu m1, [r2]
|
|
movu m2, [r2 + 16]
|
|
movu m3, [r0]
|
|
movu m4, [r0 + 16]
|
|
psadbw m1, m3
|
|
psadbw m2, m4
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
movu m1, [r2]
|
|
movu m2, [r2 + 16]
|
|
movu m3, [r0]
|
|
movu m4, [r0 + 16]
|
|
psadbw m1, m3
|
|
psadbw m2, m4
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
movu m1, [r2]
|
|
movu m2, [r2 + 16]
|
|
movu m3, [r0]
|
|
movu m4, [r0 + 16]
|
|
psadbw m1, m3
|
|
psadbw m2, m4
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
%endmacro
|
|
|
|
%macro PROCESS_SAD_32x4 1
|
|
mov%1 m1, [r2]
|
|
mov%1 m2, [r2 + 16]
|
|
psadbw m1, [r0]
|
|
psadbw m2, [r0 + 16]
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
mov%1 m1, [r2]
|
|
mov%1 m2, [r2 + 16]
|
|
psadbw m1, [r0]
|
|
psadbw m2, [r0 + 16]
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
mov%1 m1, [r2]
|
|
mov%1 m2, [r2 + 16]
|
|
psadbw m1, [r0]
|
|
psadbw m2, [r0 + 16]
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
|
|
mov%1 m1, [r2]
|
|
mov%1 m2, [r2 + 16]
|
|
psadbw m1, [r0]
|
|
psadbw m2, [r0 + 16]
|
|
paddd m1, m2
|
|
paddd m0, m1
|
|
lea r2, [r2 + r3]
|
|
lea r0, [r0 + r1]
|
|
%endmacro
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-----------------------------------------------------------------------------
|
|
INIT_XMM sse2
|
|
cglobal pixelutils_sad_32x32, 4,5,5, src1, stride1, src2, stride2
|
|
pxor m0, m0
|
|
mov r4d, 4
|
|
.loop:
|
|
PROCESS_SAD_32x4_U
|
|
PROCESS_SAD_32x4_U
|
|
dec r4d
|
|
jnz .loop
|
|
|
|
movhlps m1, m0
|
|
paddd m0, m1
|
|
movd eax, m0
|
|
RET
|
|
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_[au]_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
%macro SAD_XMM_32x32 1
|
|
INIT_XMM sse2
|
|
cglobal pixelutils_sad_%1_32x32, 4,5,3, src1, stride1, src2, stride2
|
|
pxor m0, m0
|
|
mov r4d, 4
|
|
.loop:
|
|
PROCESS_SAD_32x4 %1
|
|
PROCESS_SAD_32x4 %1
|
|
dec r4d
|
|
jnz .loop
|
|
|
|
movhlps m1, m0
|
|
paddd m0, m1
|
|
movd eax, m0
|
|
RET
|
|
%endmacro
|
|
|
|
SAD_XMM_32x32 a
|
|
SAD_XMM_32x32 u
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
INIT_YMM avx2
|
|
cglobal pixelutils_sad_32x32, 4,7,5, src1, stride1, src2, stride2
|
|
pxor m0, m0
|
|
mov r4d, 32/4
|
|
lea r5, [stride1q * 3]
|
|
lea r6, [stride2q * 3]
|
|
|
|
.loop:
|
|
movu m1, [src1q] ; row 0 of pix0
|
|
movu m2, [src2q] ; row 0 of pix1
|
|
movu m3, [src1q + stride1q] ; row 1 of pix0
|
|
movu m4, [src2q + stride2q] ; row 1 of pix1
|
|
|
|
psadbw m1, m2
|
|
psadbw m3, m4
|
|
paddd m0, m1
|
|
paddd m0, m3
|
|
|
|
movu m1, [src1q + 2 * stride1q] ; row 2 of pix0
|
|
movu m2, [src2q + 2 * stride2q] ; row 2 of pix1
|
|
movu m3, [src1q + r5] ; row 3 of pix0
|
|
movu m4, [src2q + r6] ; row 3 of pix1
|
|
|
|
psadbw m1, m2
|
|
psadbw m3, m4
|
|
paddd m0, m1
|
|
paddd m0, m3
|
|
|
|
lea src2q, [src2q + 4 * stride2q]
|
|
lea src1q, [src1q + 4 * stride1q]
|
|
|
|
dec r4d
|
|
jnz .loop
|
|
|
|
vextracti128 xm1, m0, 1
|
|
paddd xm0, xm1
|
|
pshufd xm1, xm0, 2
|
|
paddd xm0, xm1
|
|
movd eax, xm0
|
|
RET
|
|
|
|
;-------------------------------------------------------------------------------
|
|
; int ff_pixelutils_sad_[au]_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
|
|
; const uint8_t *src2, ptrdiff_t stride2);
|
|
;-------------------------------------------------------------------------------
|
|
%macro SAD_AVX2_32x32 1
|
|
INIT_YMM avx2
|
|
cglobal pixelutils_sad_%1_32x32, 4,7,3, src1, stride1, src2, stride2
|
|
pxor m0, m0
|
|
mov r4d, 32/4
|
|
lea r5, [stride1q * 3]
|
|
lea r6, [stride2q * 3]
|
|
|
|
.loop:
|
|
mov%1 m1, [src2q] ; row 0 of pix1
|
|
psadbw m1, [src1q]
|
|
mov%1 m2, [src2q + stride2q] ; row 1 of pix1
|
|
psadbw m2, [src1q + stride1q]
|
|
|
|
paddd m0, m1
|
|
paddd m0, m2
|
|
|
|
mov%1 m1, [src2q + 2 * stride2q] ; row 2 of pix1
|
|
psadbw m1, [src1q + 2 * stride1q]
|
|
mov%1 m2, [src2q + r6] ; row 3 of pix1
|
|
psadbw m2, [src1q + r5]
|
|
|
|
paddd m0, m1
|
|
paddd m0, m2
|
|
|
|
lea src2q, [src2q + 4 * stride2q]
|
|
lea src1q, [src1q + 4 * stride1q]
|
|
|
|
dec r4d
|
|
jnz .loop
|
|
|
|
vextracti128 xm1, m0, 1
|
|
paddd xm0, xm1
|
|
pshufd xm1, xm0, 2
|
|
paddd xm0, xm1
|
|
movd eax, xm0
|
|
RET
|
|
%endmacro
|
|
|
|
SAD_AVX2_32x32 a
|
|
SAD_AVX2_32x32 u
|
|
%endif
|