1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-21 10:55:51 +02:00
FFmpeg/libavutil/x86/pixelutils.asm
James Almer d5b3077ecf x86/pixelutils: add missing preprocessor wrapper to the AVX2 functions
Should fix compilation with old yasm/nasm

Signed-off-by: James Almer <jamrial@gmail.com>
2018-07-31 22:14:42 -03:00

387 lines
11 KiB
NASM

;******************************************************************************
;* Pixel utilities SIMD
;*
;* Copyright (C) 2002-2004 Michael Niedermayer <michaelni@gmx.at>
;* Copyright (C) 2014 Clément Bœsch <u pkh me>
;*
;* This file is part of FFmpeg.
;*
;* FFmpeg is free software; you can redistribute it and/or
;* modify it under the terms of the GNU Lesser General Public
;* License as published by the Free Software Foundation; either
;* version 2.1 of the License, or (at your option) any later version.
;*
;* FFmpeg is distributed in the hope that it will be useful,
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
;* Lesser General Public License for more details.
;*
;* You should have received a copy of the GNU Lesser General Public
;* License along with FFmpeg; if not, write to the Free Software
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
;******************************************************************************
%include "x86util.asm"
SECTION .text
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_8x8_mmx(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
INIT_MMX mmx
cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2
pxor m7, m7
pxor m6, m6
%rep 4
mova m0, [src1q]
mova m2, [src1q + stride1q]
mova m1, [src2q]
mova m3, [src2q + stride2q]
psubusb m4, m0, m1
psubusb m5, m2, m3
psubusb m1, m0
psubusb m3, m2
por m1, m4
por m3, m5
punpcklbw m0, m1, m7
punpcklbw m2, m3, m7
punpckhbw m1, m7
punpckhbw m3, m7
paddw m0, m1
paddw m2, m3
paddw m0, m2
paddw m6, m0
lea src1q, [src1q + 2*stride1q]
lea src2q, [src2q + 2*stride2q]
%endrep
psrlq m0, m6, 32
paddw m6, m0
psrlq m0, m6, 16
paddw m6, m0
movd eax, m6
movzx eax, ax
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_8x8_mmxext(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
INIT_MMX mmxext
cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2
pxor m2, m2
%rep 4
mova m0, [src1q]
mova m1, [src1q + stride1q]
psadbw m0, [src2q]
psadbw m1, [src2q + stride2q]
paddw m2, m0
paddw m2, m1
lea src1q, [src1q + 2*stride1q]
lea src2q, [src2q + 2*stride2q]
%endrep
movd eax, m2
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_16x16_mmxext(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
INIT_MMX mmxext
cglobal pixelutils_sad_16x16, 4,4,0, src1, stride1, src2, stride2
pxor m2, m2
%rep 16
mova m0, [src1q]
mova m1, [src1q + 8]
psadbw m0, [src2q]
psadbw m1, [src2q + 8]
paddw m2, m0
paddw m2, m1
add src1q, stride1q
add src2q, stride2q
%endrep
movd eax, m2
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
INIT_XMM sse2
cglobal pixelutils_sad_16x16, 4,4,5, src1, stride1, src2, stride2
movu m4, [src1q]
movu m2, [src2q]
movu m1, [src1q + stride1q]
movu m3, [src2q + stride2q]
psadbw m4, m2
psadbw m1, m3
paddw m4, m1
%rep 7
lea src1q, [src1q + 2*stride1q]
lea src2q, [src2q + 2*stride2q]
movu m0, [src1q]
movu m2, [src2q]
movu m1, [src1q + stride1q]
movu m3, [src2q + stride2q]
psadbw m0, m2
psadbw m1, m3
paddw m4, m0
paddw m4, m1
%endrep
movhlps m0, m4
paddw m4, m0
movd eax, m4
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_[au]_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
%macro SAD_XMM_16x16 1
INIT_XMM sse2
cglobal pixelutils_sad_%1_16x16, 4,4,3, src1, stride1, src2, stride2
mov%1 m2, [src2q]
psadbw m2, [src1q]
mov%1 m1, [src2q + stride2q]
psadbw m1, [src1q + stride1q]
paddw m2, m1
%rep 7
lea src1q, [src1q + 2*stride1q]
lea src2q, [src2q + 2*stride2q]
mov%1 m0, [src2q]
psadbw m0, [src1q]
mov%1 m1, [src2q + stride2q]
psadbw m1, [src1q + stride1q]
paddw m2, m0
paddw m2, m1
%endrep
movhlps m0, m2
paddw m2, m0
movd eax, m2
RET
%endmacro
SAD_XMM_16x16 a
SAD_XMM_16x16 u
%macro PROCESS_SAD_32x4_U 0
movu m1, [r2]
movu m2, [r2 + 16]
movu m3, [r0]
movu m4, [r0 + 16]
psadbw m1, m3
psadbw m2, m4
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
movu m1, [r2]
movu m2, [r2 + 16]
movu m3, [r0]
movu m4, [r0 + 16]
psadbw m1, m3
psadbw m2, m4
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
movu m1, [r2]
movu m2, [r2 + 16]
movu m3, [r0]
movu m4, [r0 + 16]
psadbw m1, m3
psadbw m2, m4
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
movu m1, [r2]
movu m2, [r2 + 16]
movu m3, [r0]
movu m4, [r0 + 16]
psadbw m1, m3
psadbw m2, m4
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
%endmacro
%macro PROCESS_SAD_32x4 1
mov%1 m1, [r2]
mov%1 m2, [r2 + 16]
psadbw m1, [r0]
psadbw m2, [r0 + 16]
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
mov%1 m1, [r2]
mov%1 m2, [r2 + 16]
psadbw m1, [r0]
psadbw m2, [r0 + 16]
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
mov%1 m1, [r2]
mov%1 m2, [r2 + 16]
psadbw m1, [r0]
psadbw m2, [r0 + 16]
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
mov%1 m1, [r2]
mov%1 m2, [r2 + 16]
psadbw m1, [r0]
psadbw m2, [r0 + 16]
paddd m1, m2
paddd m0, m1
lea r2, [r2 + r3]
lea r0, [r0 + r1]
%endmacro
;-----------------------------------------------------------------------------
; int ff_pixelutils_sad_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-----------------------------------------------------------------------------
INIT_XMM sse2
cglobal pixelutils_sad_32x32, 4,5,5, src1, stride1, src2, stride2
pxor m0, m0
mov r4d, 4
.loop:
PROCESS_SAD_32x4_U
PROCESS_SAD_32x4_U
dec r4d
jnz .loop
movhlps m1, m0
paddd m0, m1
movd eax, m0
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_[au]_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
%macro SAD_XMM_32x32 1
INIT_XMM sse2
cglobal pixelutils_sad_%1_32x32, 4,5,3, src1, stride1, src2, stride2
pxor m0, m0
mov r4d, 4
.loop:
PROCESS_SAD_32x4 %1
PROCESS_SAD_32x4 %1
dec r4d
jnz .loop
movhlps m1, m0
paddd m0, m1
movd eax, m0
RET
%endmacro
SAD_XMM_32x32 a
SAD_XMM_32x32 u
%if HAVE_AVX2_EXTERNAL
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
INIT_YMM avx2
cglobal pixelutils_sad_32x32, 4,7,5, src1, stride1, src2, stride2
pxor m0, m0
mov r4d, 32/4
lea r5, [stride1q * 3]
lea r6, [stride2q * 3]
.loop:
movu m1, [src1q] ; row 0 of pix0
movu m2, [src2q] ; row 0 of pix1
movu m3, [src1q + stride1q] ; row 1 of pix0
movu m4, [src2q + stride2q] ; row 1 of pix1
psadbw m1, m2
psadbw m3, m4
paddd m0, m1
paddd m0, m3
movu m1, [src1q + 2 * stride1q] ; row 2 of pix0
movu m2, [src2q + 2 * stride2q] ; row 2 of pix1
movu m3, [src1q + r5] ; row 3 of pix0
movu m4, [src2q + r6] ; row 3 of pix1
psadbw m1, m2
psadbw m3, m4
paddd m0, m1
paddd m0, m3
lea src2q, [src2q + 4 * stride2q]
lea src1q, [src1q + 4 * stride1q]
dec r4d
jnz .loop
vextracti128 xm1, m0, 1
paddd xm0, xm1
pshufd xm1, xm0, 2
paddd xm0, xm1
movd eax, xm0
RET
;-------------------------------------------------------------------------------
; int ff_pixelutils_sad_[au]_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1,
; const uint8_t *src2, ptrdiff_t stride2);
;-------------------------------------------------------------------------------
%macro SAD_AVX2_32x32 1
INIT_YMM avx2
cglobal pixelutils_sad_%1_32x32, 4,7,3, src1, stride1, src2, stride2
pxor m0, m0
mov r4d, 32/4
lea r5, [stride1q * 3]
lea r6, [stride2q * 3]
.loop:
mov%1 m1, [src2q] ; row 0 of pix1
psadbw m1, [src1q]
mov%1 m2, [src2q + stride2q] ; row 1 of pix1
psadbw m2, [src1q + stride1q]
paddd m0, m1
paddd m0, m2
mov%1 m1, [src2q + 2 * stride2q] ; row 2 of pix1
psadbw m1, [src1q + 2 * stride1q]
mov%1 m2, [src2q + r6] ; row 3 of pix1
psadbw m2, [src1q + r5]
paddd m0, m1
paddd m0, m2
lea src2q, [src2q + 4 * stride2q]
lea src1q, [src1q + 4 * stride1q]
dec r4d
jnz .loop
vextracti128 xm1, m0, 1
paddd xm0, xm1
pshufd xm1, xm0, 2
paddd xm0, xm1
movd eax, xm0
RET
%endmacro
SAD_AVX2_32x32 a
SAD_AVX2_32x32 u
%endif