mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-12 19:18:44 +02:00
074155360d
There are instructions pavgb and pavgusb. Both instructions do the same
operation but they have different enconding. Pavgb exists in SSE (or
MMXEXT) instruction set and pavgusb exists in 3D-NOW instruction set.
livavcodec uses the macro PAVGB to select the proper instruction. However,
the function avg_pixels8_xy2 doesn't use this macro, it uses pavgb
directly.
As a consequence, the function avg_pixels8_xy2 crashes on AMD K6-2 and
K6-3 processors, because they have pavgusb, but not pavgb.
This bug seems to be introduced by commit
71155d7b41
, "dsputil: x86: Convert mpeg4
qpel and dsputil avg to yasm"
Signed-off-by: Mikulas Patocka <mikulas@artax.karlin.mff.cuni.cz>
Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
462 lines
11 KiB
NASM
462 lines
11 KiB
NASM
;******************************************************************************
|
|
;*
|
|
;* Copyright (c) 2000-2001 Fabrice Bellard <fabrice@bellard.org>
|
|
;* Copyright (c) Nick Kurshev <nickols_k@mail.ru>
|
|
;* Copyright (c) 2002 Michael Niedermayer <michaelni@gmx.at>
|
|
;* Copyright (c) 2002 Zdenek Kabelac <kabi@informatics.muni.cz>
|
|
;* Copyright (c) 2013 Daniel Kang
|
|
;*
|
|
;* MMX optimized hpel functions
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA
|
|
cextern pb_1
|
|
|
|
SECTION_TEXT
|
|
|
|
; put_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_PIXELS8_X2 0
|
|
cglobal put_pixels8_x2, 4,5
|
|
lea r4, [r2*2]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, [r1+r2]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m1, [r1+r2+1]
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
add r1, r4
|
|
add r0, r4
|
|
mova m0, [r1]
|
|
mova m1, [r1+r2]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m1, [r1+r2+1]
|
|
add r1, r4
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_PIXELS8_X2
|
|
INIT_MMX 3dnow
|
|
PUT_PIXELS8_X2
|
|
|
|
|
|
; put_pixels16_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_PIXELS_16 0
|
|
cglobal put_pixels16_x2, 4,5
|
|
lea r4, [r2*2]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m1, [r1+r2]
|
|
mova m2, [r1+8]
|
|
mova m3, [r1+r2+8]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m1, [r1+r2+1]
|
|
PAVGB m2, [r1+9]
|
|
PAVGB m3, [r1+r2+9]
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
mova [r0+8], m2
|
|
mova [r0+r2+8], m3
|
|
add r1, r4
|
|
add r0, r4
|
|
mova m0, [r1]
|
|
mova m1, [r1+r2]
|
|
mova m2, [r1+8]
|
|
mova m3, [r1+r2+8]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m1, [r1+r2+1]
|
|
PAVGB m2, [r1+9]
|
|
PAVGB m3, [r1+r2+9]
|
|
add r1, r4
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
mova [r0+8], m2
|
|
mova [r0+r2+8], m3
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_PIXELS_16
|
|
INIT_MMX 3dnow
|
|
PUT_PIXELS_16
|
|
|
|
|
|
; put_no_rnd_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_NO_RND_PIXELS8_X2 0
|
|
cglobal put_no_rnd_pixels8_x2, 4,5
|
|
mova m6, [pb_1]
|
|
lea r4, [r2*2]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+r2]
|
|
mova m1, [r1+1]
|
|
mova m3, [r1+r2+1]
|
|
add r1, r4
|
|
psubusb m0, m6
|
|
psubusb m2, m6
|
|
PAVGB m0, m1
|
|
PAVGB m2, m3
|
|
mova [r0], m0
|
|
mova [r0+r2], m2
|
|
mova m0, [r1]
|
|
mova m1, [r1+1]
|
|
mova m2, [r1+r2]
|
|
mova m3, [r1+r2+1]
|
|
add r0, r4
|
|
add r1, r4
|
|
psubusb m0, m6
|
|
psubusb m2, m6
|
|
PAVGB m0, m1
|
|
PAVGB m2, m3
|
|
mova [r0], m0
|
|
mova [r0+r2], m2
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS8_X2
|
|
INIT_MMX 3dnow
|
|
PUT_NO_RND_PIXELS8_X2
|
|
|
|
|
|
; put_no_rnd_pixels8_x2_exact(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_NO_RND_PIXELS8_X2_EXACT 0
|
|
cglobal put_no_rnd_pixels8_x2_exact, 4,5
|
|
lea r4, [r2*3]
|
|
pcmpeqb m6, m6
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+r2]
|
|
mova m1, [r1+1]
|
|
mova m3, [r1+r2+1]
|
|
pxor m0, m6
|
|
pxor m2, m6
|
|
pxor m1, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m1
|
|
PAVGB m2, m3
|
|
pxor m0, m6
|
|
pxor m2, m6
|
|
mova [r0], m0
|
|
mova [r0+r2], m2
|
|
mova m0, [r1+r2*2]
|
|
mova m1, [r1+r2*2+1]
|
|
mova m2, [r1+r4]
|
|
mova m3, [r1+r4+1]
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
pxor m3, m6
|
|
PAVGB m0, m1
|
|
PAVGB m2, m3
|
|
pxor m0, m6
|
|
pxor m2, m6
|
|
mova [r0+r2*2], m0
|
|
mova [r0+r4], m2
|
|
lea r1, [r1+r2*4]
|
|
lea r0, [r0+r2*4]
|
|
sub r3d, 4
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS8_X2_EXACT
|
|
INIT_MMX 3dnow
|
|
PUT_NO_RND_PIXELS8_X2_EXACT
|
|
|
|
|
|
; put_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_PIXELS8_Y2 0
|
|
cglobal put_pixels8_y2, 4,5
|
|
lea r4, [r2*2]
|
|
mova m0, [r1]
|
|
sub r0, r2
|
|
.loop:
|
|
mova m1, [r1+r2]
|
|
mova m2, [r1+r4]
|
|
add r1, r4
|
|
PAVGB m0, m1
|
|
PAVGB m1, m2
|
|
mova [r0+r2], m0
|
|
mova [r0+r4], m1
|
|
mova m1, [r1+r2]
|
|
mova m0, [r1+r4]
|
|
add r0, r4
|
|
add r1, r4
|
|
PAVGB m2, m1
|
|
PAVGB m1, m0
|
|
mova [r0+r2], m2
|
|
mova [r0+r4], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_PIXELS8_Y2
|
|
INIT_MMX 3dnow
|
|
PUT_PIXELS8_Y2
|
|
|
|
|
|
; put_no_rnd_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_NO_RND_PIXELS8_Y2 0
|
|
cglobal put_no_rnd_pixels8_y2, 4,5
|
|
mova m6, [pb_1]
|
|
lea r4, [r2+r2]
|
|
mova m0, [r1]
|
|
sub r0, r2
|
|
.loop:
|
|
mova m1, [r1+r2]
|
|
mova m2, [r1+r4]
|
|
add r1, r4
|
|
psubusb m1, m6
|
|
PAVGB m0, m1
|
|
PAVGB m1, m2
|
|
mova [r0+r2], m0
|
|
mova [r0+r4], m1
|
|
mova m1, [r1+r2]
|
|
mova m0, [r1+r4]
|
|
add r0, r4
|
|
add r1, r4
|
|
psubusb m1, m6
|
|
PAVGB m2, m1
|
|
PAVGB m1, m0
|
|
mova [r0+r2], m2
|
|
mova [r0+r4], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS8_Y2
|
|
INIT_MMX 3dnow
|
|
PUT_NO_RND_PIXELS8_Y2
|
|
|
|
|
|
; put_no_rnd_pixels8_y2_exact(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro PUT_NO_RND_PIXELS8_Y2_EXACT 0
|
|
cglobal put_no_rnd_pixels8_y2_exact, 4,5
|
|
lea r4, [r2*3]
|
|
mova m0, [r1]
|
|
pcmpeqb m6, m6
|
|
add r1, r2
|
|
pxor m0, m6
|
|
.loop:
|
|
mova m1, [r1]
|
|
mova m2, [r1+r2]
|
|
pxor m1, m6
|
|
pxor m2, m6
|
|
PAVGB m0, m1
|
|
PAVGB m1, m2
|
|
pxor m0, m6
|
|
pxor m1, m6
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
mova m1, [r1+r2*2]
|
|
mova m0, [r1+r4]
|
|
pxor m1, m6
|
|
pxor m0, m6
|
|
PAVGB m2, m1
|
|
PAVGB m1, m0
|
|
pxor m2, m6
|
|
pxor m1, m6
|
|
mova [r0+r2*2], m2
|
|
mova [r0+r4], m1
|
|
lea r1, [r1+r2*4]
|
|
lea r0, [r0+r2*4]
|
|
sub r3d, 4
|
|
jg .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
PUT_NO_RND_PIXELS8_Y2_EXACT
|
|
INIT_MMX 3dnow
|
|
PUT_NO_RND_PIXELS8_Y2_EXACT
|
|
|
|
|
|
; avg_pixels8(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro AVG_PIXELS8 0
|
|
cglobal avg_pixels8, 4,5
|
|
lea r4, [r2*2]
|
|
.loop:
|
|
mova m0, [r0]
|
|
mova m1, [r0+r2]
|
|
PAVGB m0, [r1]
|
|
PAVGB m1, [r1+r2]
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
add r1, r4
|
|
add r0, r4
|
|
mova m0, [r0]
|
|
mova m1, [r0+r2]
|
|
PAVGB m0, [r1]
|
|
PAVGB m1, [r1+r2]
|
|
add r1, r4
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX 3dnow
|
|
AVG_PIXELS8
|
|
|
|
|
|
; avg_pixels8_x2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro AVG_PIXELS8_X2 0
|
|
cglobal avg_pixels8_x2, 4,5
|
|
lea r4, [r2*2]
|
|
.loop:
|
|
mova m0, [r1]
|
|
mova m2, [r1+r2]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m2, [r1+r2+1]
|
|
PAVGB m0, [r0]
|
|
PAVGB m2, [r0+r2]
|
|
add r1, r4
|
|
mova [r0], m0
|
|
mova [r0+r2], m2
|
|
mova m0, [r1]
|
|
mova m2, [r1+r2]
|
|
PAVGB m0, [r1+1]
|
|
PAVGB m2, [r1+r2+1]
|
|
add r0, r4
|
|
add r1, r4
|
|
PAVGB m0, [r0]
|
|
PAVGB m2, [r0+r2]
|
|
mova [r0], m0
|
|
mova [r0+r2], m2
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
AVG_PIXELS8_X2
|
|
INIT_MMX 3dnow
|
|
AVG_PIXELS8_X2
|
|
|
|
|
|
; avg_pixels8_y2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro AVG_PIXELS8_Y2 0
|
|
cglobal avg_pixels8_y2, 4,5
|
|
lea r4, [r2*2]
|
|
mova m0, [r1]
|
|
sub r0, r2
|
|
.loop:
|
|
mova m1, [r1+r2]
|
|
mova m2, [r1+r4]
|
|
add r1, r4
|
|
PAVGB m0, m1
|
|
PAVGB m1, m2
|
|
mova m3, [r0+r2]
|
|
mova m4, [r0+r4]
|
|
PAVGB m0, m3
|
|
PAVGB m1, m4
|
|
mova [r0+r2], m0
|
|
mova [r0+r4], m1
|
|
mova m1, [r1+r2]
|
|
mova m0, [r1+r4]
|
|
PAVGB m2, m1
|
|
PAVGB m1, m0
|
|
add r0, r4
|
|
add r1, r4
|
|
mova m3, [r0+r2]
|
|
mova m4, [r0+r4]
|
|
PAVGB m2, m3
|
|
PAVGB m1, m4
|
|
mova [r0+r2], m2
|
|
mova [r0+r4], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
AVG_PIXELS8_Y2
|
|
INIT_MMX 3dnow
|
|
AVG_PIXELS8_Y2
|
|
|
|
|
|
; avg_pixels8_xy2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h)
|
|
%macro AVG_PIXELS8_XY2 0
|
|
cglobal avg_pixels8_xy2, 4,5
|
|
mova m6, [pb_1]
|
|
lea r4, [r2*2]
|
|
mova m0, [r1]
|
|
PAVGB m0, [r1+1]
|
|
.loop:
|
|
mova m2, [r1+r4]
|
|
mova m1, [r1+r2]
|
|
psubusb m2, m6
|
|
PAVGB m1, [r1+r2+1]
|
|
PAVGB m2, [r1+r4+1]
|
|
add r1, r4
|
|
PAVGB m0, m1
|
|
PAVGB m1, m2
|
|
PAVGB m0, [r0]
|
|
PAVGB m1, [r0+r2]
|
|
mova [r0], m0
|
|
mova [r0+r2], m1
|
|
mova m1, [r1+r2]
|
|
mova m0, [r1+r4]
|
|
PAVGB m1, [r1+r2+1]
|
|
PAVGB m0, [r1+r4+1]
|
|
add r0, r4
|
|
add r1, r4
|
|
PAVGB m2, m1
|
|
PAVGB m1, m0
|
|
PAVGB m2, [r0]
|
|
PAVGB m1, [r0+r2]
|
|
mova [r0], m2
|
|
mova [r0+r2], m1
|
|
add r0, r4
|
|
sub r3d, 4
|
|
jne .loop
|
|
REP_RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
AVG_PIXELS8_XY2
|
|
INIT_MMX 3dnow
|
|
AVG_PIXELS8_XY2
|