mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2025-03-23 04:24:35 +02:00
avcodec/x86/vc1dsp_init: Remove obsolete 3dnow, MMX(EXT) functions
x64 always has MMX, MMXEXT, SSE and SSE2 and this means that some functions for MMX, MMXEXT and 3dnow are always overridden by other functions (unless one e.g. explicitly disables SSE2) for x64. So given that the only systems that benefit from these functions are truely ancient 32bit x86s they are removed. Signed-off-by: Andreas Rheinhardt <andreas.rheinhardt@outlook.com>
This commit is contained in:
parent
c02dd59cd3
commit
e02ffed004
libavcodec/x86
@ -448,7 +448,6 @@ chroma_mc2_mmx_func avg, h264
|
||||
|
||||
INIT_MMX 3dnow
|
||||
chroma_mc8_mmx_func avg, h264, _rnd
|
||||
chroma_mc8_mmx_func avg, vc1, _nornd
|
||||
chroma_mc8_mmx_func avg, rv40
|
||||
chroma_mc4_mmx_func avg, h264
|
||||
chroma_mc4_mmx_func avg, rv40
|
||||
|
@ -33,9 +33,10 @@
|
||||
#include "vc1dsp.h"
|
||||
#include "config.h"
|
||||
|
||||
#define LOOP_FILTER(EXT) \
|
||||
#define LOOP_FILTER4(EXT) \
|
||||
void ff_vc1_v_loop_filter4_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq); \
|
||||
void ff_vc1_h_loop_filter4_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq); \
|
||||
void ff_vc1_h_loop_filter4_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq);
|
||||
#define LOOP_FILTER816(EXT) \
|
||||
void ff_vc1_v_loop_filter8_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq); \
|
||||
void ff_vc1_h_loop_filter8_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq); \
|
||||
\
|
||||
@ -52,9 +53,10 @@ static void vc1_h_loop_filter16_ ## EXT(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
}
|
||||
|
||||
#if HAVE_X86ASM
|
||||
LOOP_FILTER(mmxext)
|
||||
LOOP_FILTER(sse2)
|
||||
LOOP_FILTER(ssse3)
|
||||
LOOP_FILTER4(mmxext)
|
||||
LOOP_FILTER816(sse2)
|
||||
LOOP_FILTER4(ssse3)
|
||||
LOOP_FILTER816(ssse3)
|
||||
|
||||
void ff_vc1_h_loop_filter8_sse4(uint8_t *src, ptrdiff_t stride, int pq);
|
||||
|
||||
@ -72,11 +74,7 @@ static void vc1_h_loop_filter16_sse4(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
}
|
||||
|
||||
DECLARE_FUNCTION(put_, 8, _mmx)
|
||||
DECLARE_FUNCTION(put_, 16, _mmx)
|
||||
DECLARE_FUNCTION(avg_, 8, _mmx)
|
||||
DECLARE_FUNCTION(avg_, 16, _mmx)
|
||||
DECLARE_FUNCTION(avg_, 8, _mmxext)
|
||||
DECLARE_FUNCTION(avg_, 16, _mmxext)
|
||||
DECLARE_FUNCTION(put_, 16, _sse2)
|
||||
DECLARE_FUNCTION(avg_, 16, _sse2)
|
||||
|
||||
@ -86,8 +84,6 @@ void ff_put_vc1_chroma_mc8_nornd_mmx (uint8_t *dst, uint8_t *src,
|
||||
ptrdiff_t stride, int h, int x, int y);
|
||||
void ff_avg_vc1_chroma_mc8_nornd_mmxext(uint8_t *dst, uint8_t *src,
|
||||
ptrdiff_t stride, int h, int x, int y);
|
||||
void ff_avg_vc1_chroma_mc8_nornd_3dnow(uint8_t *dst, uint8_t *src,
|
||||
ptrdiff_t stride, int h, int x, int y);
|
||||
void ff_put_vc1_chroma_mc8_nornd_ssse3(uint8_t *dst, uint8_t *src,
|
||||
ptrdiff_t stride, int h, int x, int y);
|
||||
void ff_avg_vc1_chroma_mc8_nornd_ssse3(uint8_t *dst, uint8_t *src,
|
||||
@ -114,9 +110,10 @@ av_cold void ff_vc1dsp_init_x86(VC1DSPContext *dsp)
|
||||
if (EXTERNAL_MMXEXT(cpu_flags))
|
||||
ff_vc1dsp_init_mmxext(dsp);
|
||||
|
||||
#define ASSIGN_LF(EXT) \
|
||||
#define ASSIGN_LF4(EXT) \
|
||||
dsp->vc1_v_loop_filter4 = ff_vc1_v_loop_filter4_ ## EXT; \
|
||||
dsp->vc1_h_loop_filter4 = ff_vc1_h_loop_filter4_ ## EXT; \
|
||||
dsp->vc1_h_loop_filter4 = ff_vc1_h_loop_filter4_ ## EXT
|
||||
#define ASSIGN_LF816(EXT) \
|
||||
dsp->vc1_v_loop_filter8 = ff_vc1_v_loop_filter8_ ## EXT; \
|
||||
dsp->vc1_h_loop_filter8 = ff_vc1_h_loop_filter8_ ## EXT; \
|
||||
dsp->vc1_v_loop_filter16 = vc1_v_loop_filter16_ ## EXT; \
|
||||
@ -127,19 +124,12 @@ av_cold void ff_vc1dsp_init_x86(VC1DSPContext *dsp)
|
||||
dsp->put_no_rnd_vc1_chroma_pixels_tab[0] = ff_put_vc1_chroma_mc8_nornd_mmx;
|
||||
|
||||
dsp->put_vc1_mspel_pixels_tab[1][0] = put_vc1_mspel_mc00_8_mmx;
|
||||
dsp->put_vc1_mspel_pixels_tab[0][0] = put_vc1_mspel_mc00_16_mmx;
|
||||
dsp->avg_vc1_mspel_pixels_tab[1][0] = avg_vc1_mspel_mc00_8_mmx;
|
||||
dsp->avg_vc1_mspel_pixels_tab[0][0] = avg_vc1_mspel_mc00_16_mmx;
|
||||
}
|
||||
if (EXTERNAL_AMD3DNOW(cpu_flags)) {
|
||||
dsp->avg_no_rnd_vc1_chroma_pixels_tab[0] = ff_avg_vc1_chroma_mc8_nornd_3dnow;
|
||||
}
|
||||
if (EXTERNAL_MMXEXT(cpu_flags)) {
|
||||
ASSIGN_LF(mmxext);
|
||||
ASSIGN_LF4(mmxext);
|
||||
dsp->avg_no_rnd_vc1_chroma_pixels_tab[0] = ff_avg_vc1_chroma_mc8_nornd_mmxext;
|
||||
|
||||
dsp->avg_vc1_mspel_pixels_tab[1][0] = avg_vc1_mspel_mc00_8_mmxext;
|
||||
dsp->avg_vc1_mspel_pixels_tab[0][0] = avg_vc1_mspel_mc00_16_mmxext;
|
||||
|
||||
dsp->vc1_inv_trans_8x8_dc = ff_vc1_inv_trans_8x8_dc_mmxext;
|
||||
dsp->vc1_inv_trans_4x8_dc = ff_vc1_inv_trans_4x8_dc_mmxext;
|
||||
@ -147,16 +137,14 @@ av_cold void ff_vc1dsp_init_x86(VC1DSPContext *dsp)
|
||||
dsp->vc1_inv_trans_4x4_dc = ff_vc1_inv_trans_4x4_dc_mmxext;
|
||||
}
|
||||
if (EXTERNAL_SSE2(cpu_flags)) {
|
||||
dsp->vc1_v_loop_filter8 = ff_vc1_v_loop_filter8_sse2;
|
||||
dsp->vc1_h_loop_filter8 = ff_vc1_h_loop_filter8_sse2;
|
||||
dsp->vc1_v_loop_filter16 = vc1_v_loop_filter16_sse2;
|
||||
dsp->vc1_h_loop_filter16 = vc1_h_loop_filter16_sse2;
|
||||
ASSIGN_LF816(sse2);
|
||||
|
||||
dsp->put_vc1_mspel_pixels_tab[0][0] = put_vc1_mspel_mc00_16_sse2;
|
||||
dsp->avg_vc1_mspel_pixels_tab[0][0] = avg_vc1_mspel_mc00_16_sse2;
|
||||
}
|
||||
if (EXTERNAL_SSSE3(cpu_flags)) {
|
||||
ASSIGN_LF(ssse3);
|
||||
ASSIGN_LF4(ssse3);
|
||||
ASSIGN_LF816(ssse3);
|
||||
dsp->put_no_rnd_vc1_chroma_pixels_tab[0] = ff_put_vc1_chroma_mc8_nornd_ssse3;
|
||||
dsp->avg_no_rnd_vc1_chroma_pixels_tab[0] = ff_avg_vc1_chroma_mc8_nornd_ssse3;
|
||||
}
|
||||
|
@ -228,48 +228,20 @@ SECTION .text
|
||||
imul r2, 0x01010101
|
||||
%endmacro
|
||||
|
||||
%macro VC1_LF 0
|
||||
cglobal vc1_v_loop_filter_internal
|
||||
VC1_V_LOOP_FILTER 4, d
|
||||
ret
|
||||
|
||||
cglobal vc1_h_loop_filter_internal
|
||||
VC1_H_LOOP_FILTER 4, r4
|
||||
ret
|
||||
|
||||
; void ff_vc1_v_loop_filter4_mmxext(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
INIT_MMX mmxext
|
||||
cglobal vc1_v_loop_filter4, 3,5,0
|
||||
START_V_FILTER
|
||||
call vc1_v_loop_filter_internal
|
||||
VC1_V_LOOP_FILTER 4, d
|
||||
RET
|
||||
|
||||
; void ff_vc1_h_loop_filter4_mmxext(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
INIT_MMX mmxext
|
||||
cglobal vc1_h_loop_filter4, 3,5,0
|
||||
START_H_FILTER 4
|
||||
call vc1_h_loop_filter_internal
|
||||
VC1_H_LOOP_FILTER 4, r4
|
||||
RET
|
||||
|
||||
; void ff_vc1_v_loop_filter8_mmxext(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
cglobal vc1_v_loop_filter8, 3,5,0
|
||||
START_V_FILTER
|
||||
call vc1_v_loop_filter_internal
|
||||
add r4, 4
|
||||
add r0, 4
|
||||
call vc1_v_loop_filter_internal
|
||||
RET
|
||||
|
||||
; void ff_vc1_h_loop_filter8_mmxext(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
cglobal vc1_h_loop_filter8, 3,5,0
|
||||
START_H_FILTER 4
|
||||
call vc1_h_loop_filter_internal
|
||||
lea r0, [r0+4*r1]
|
||||
call vc1_h_loop_filter_internal
|
||||
RET
|
||||
%endmacro
|
||||
|
||||
INIT_MMX mmxext
|
||||
VC1_LF
|
||||
|
||||
INIT_XMM sse2
|
||||
; void ff_vc1_v_loop_filter8_sse2(uint8_t *src, ptrdiff_t stride, int pq)
|
||||
cglobal vc1_v_loop_filter8, 3,5,8
|
||||
|
Loading…
x
Reference in New Issue
Block a user