mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-23 12:43:46 +02:00
sse2 h264 motion compensation. not new code, just separate out the cases that didn't need ssse3.
Originally committed as revision 11877 to svn://svn.ffmpeg.org/ffmpeg/trunk
This commit is contained in:
parent
20d565be6d
commit
1d67b037f7
@ -3523,20 +3523,48 @@ void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx)
|
||||
c->avg_h264_chroma_pixels_tab[1]= avg_h264_chroma_mc4_3dnow;
|
||||
}
|
||||
|
||||
/* FIXME works in most codecs, but crashes svq1 due to unaligned chroma
|
||||
|
||||
#define H264_QPEL_FUNCS(x, y, CPU)\
|
||||
c->put_h264_qpel_pixels_tab[0][x+y*4] = put_h264_qpel16_mc##x##y##_##CPU;\
|
||||
c->put_h264_qpel_pixels_tab[1][x+y*4] = put_h264_qpel8_mc##x##y##_##CPU;\
|
||||
c->avg_h264_qpel_pixels_tab[0][x+y*4] = avg_h264_qpel16_mc##x##y##_##CPU;\
|
||||
c->avg_h264_qpel_pixels_tab[1][x+y*4] = avg_h264_qpel8_mc##x##y##_##CPU;
|
||||
if((mm_flags & MM_SSE2) && !(mm_flags & MM_3DNOW)){
|
||||
// these functions are slower than mmx on AMD, but faster on Intel
|
||||
/* FIXME works in most codecs, but crashes svq1 due to unaligned chroma
|
||||
c->put_pixels_tab[0][0] = put_pixels16_sse2;
|
||||
c->avg_pixels_tab[0][0] = avg_pixels16_sse2;
|
||||
}
|
||||
*/
|
||||
|
||||
H264_QPEL_FUNCS(0, 0, sse2);
|
||||
}
|
||||
if(mm_flags & MM_SSE2){
|
||||
H264_QPEL_FUNCS(0, 1, sse2);
|
||||
H264_QPEL_FUNCS(0, 2, sse2);
|
||||
H264_QPEL_FUNCS(0, 3, sse2);
|
||||
H264_QPEL_FUNCS(1, 1, sse2);
|
||||
H264_QPEL_FUNCS(1, 2, sse2);
|
||||
H264_QPEL_FUNCS(1, 3, sse2);
|
||||
H264_QPEL_FUNCS(2, 1, sse2);
|
||||
H264_QPEL_FUNCS(2, 2, sse2);
|
||||
H264_QPEL_FUNCS(2, 3, sse2);
|
||||
H264_QPEL_FUNCS(3, 1, sse2);
|
||||
H264_QPEL_FUNCS(3, 2, sse2);
|
||||
H264_QPEL_FUNCS(3, 3, sse2);
|
||||
}
|
||||
#ifdef HAVE_SSSE3
|
||||
if(mm_flags & MM_SSSE3){
|
||||
SET_QPEL_FUNCS(put_h264_qpel, 0, 16, ssse3);
|
||||
SET_QPEL_FUNCS(put_h264_qpel, 1, 8, ssse3);
|
||||
SET_QPEL_FUNCS(avg_h264_qpel, 0, 16, ssse3);
|
||||
SET_QPEL_FUNCS(avg_h264_qpel, 1, 8, ssse3);
|
||||
H264_QPEL_FUNCS(1, 0, ssse3);
|
||||
H264_QPEL_FUNCS(1, 1, ssse3);
|
||||
H264_QPEL_FUNCS(1, 2, ssse3);
|
||||
H264_QPEL_FUNCS(1, 3, ssse3);
|
||||
H264_QPEL_FUNCS(2, 0, ssse3);
|
||||
H264_QPEL_FUNCS(2, 1, ssse3);
|
||||
H264_QPEL_FUNCS(2, 2, ssse3);
|
||||
H264_QPEL_FUNCS(2, 3, ssse3);
|
||||
H264_QPEL_FUNCS(3, 0, ssse3);
|
||||
H264_QPEL_FUNCS(3, 1, ssse3);
|
||||
H264_QPEL_FUNCS(3, 2, ssse3);
|
||||
H264_QPEL_FUNCS(3, 3, ssse3);
|
||||
}
|
||||
#endif
|
||||
|
||||
|
@ -1111,8 +1111,7 @@ static av_noinline void OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(uint8_t *dst,
|
||||
dst += 4-h*dstStride;\
|
||||
}\
|
||||
}\
|
||||
static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride, int size){\
|
||||
int h = size;\
|
||||
static av_always_inline void OPNAME ## h264_qpel8or16_hv1_lowpass_ ## MMX(int16_t *tmp, uint8_t *src, int tmpStride, int srcStride, int size){\
|
||||
int w = (size+8)>>2;\
|
||||
src -= 2*srcStride+2;\
|
||||
while(w--){\
|
||||
@ -1163,10 +1162,11 @@ static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst
|
||||
tmp += 4;\
|
||||
src += 4 - (size+5)*srcStride;\
|
||||
}\
|
||||
tmp -= size+8;\
|
||||
w = size>>4;\
|
||||
}\
|
||||
static av_always_inline void OPNAME ## h264_qpel8or16_hv2_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, int dstStride, int tmpStride, int size){\
|
||||
int w = size>>4;\
|
||||
do{\
|
||||
h = size;\
|
||||
int h = size;\
|
||||
asm volatile(\
|
||||
"1: \n\t"\
|
||||
"movq (%0), %%mm0 \n\t"\
|
||||
@ -1218,7 +1218,7 @@ static av_noinline void OPNAME ## h264_qpel16_v_lowpass_ ## MMX(uint8_t *dst, ui
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride, 16);\
|
||||
}\
|
||||
\
|
||||
static av_noinline void OPNAME ## h264_qpel16_h_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
static void OPNAME ## h264_qpel16_h_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst , src , dstStride, srcStride);\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride);\
|
||||
src += 8*srcStride;\
|
||||
@ -1237,6 +1237,10 @@ static av_noinline void OPNAME ## h264_qpel16_h_lowpass_l2_ ## MMX(uint8_t *dst,
|
||||
OPNAME ## h264_qpel8_h_lowpass_l2_ ## MMX(dst+8, src+8, src2+8, dstStride, src2Stride);\
|
||||
}\
|
||||
\
|
||||
static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride, int size){\
|
||||
put_h264_qpel8or16_hv1_lowpass_ ## MMX(tmp, src, tmpStride, srcStride, size);\
|
||||
OPNAME ## h264_qpel8or16_hv2_lowpass_ ## MMX(dst, tmp, dstStride, tmpStride, size);\
|
||||
}\
|
||||
static void OPNAME ## h264_qpel8_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(dst , tmp , src , dstStride, tmpStride, srcStride, 8);\
|
||||
}\
|
||||
@ -1306,8 +1310,9 @@ static void OPNAME ## pixels16_l2_shift5_ ## MMX(uint8_t *dst, int16_t *src16, u
|
||||
OPNAME ## pixels8_l2_shift5_ ## MMX(dst+8, src16+8, src8+8, dstStride, src8Stride, h);\
|
||||
}\
|
||||
|
||||
|
||||
#ifdef ARCH_X86_64
|
||||
#define QPEL_H264_HL2_XMM(OPNAME, OP, MMX)\
|
||||
#define QPEL_H264_H16_XMM(OPNAME, OP, MMX)\
|
||||
static av_noinline void OPNAME ## h264_qpel16_h_lowpass_l2_ ## MMX(uint8_t *dst, uint8_t *src, uint8_t *src2, int dstStride, int src2Stride){\
|
||||
int h=16;\
|
||||
asm volatile(\
|
||||
@ -1373,7 +1378,7 @@ static av_noinline void OPNAME ## h264_qpel16_h_lowpass_l2_ ## MMX(uint8_t *dst,
|
||||
);\
|
||||
}
|
||||
#else // ARCH_X86_64
|
||||
#define QPEL_H264_HL2_XMM(OPNAME, OP, MMX)\
|
||||
#define QPEL_H264_H16_XMM(OPNAME, OP, MMX)\
|
||||
static av_noinline void OPNAME ## h264_qpel16_h_lowpass_l2_ ## MMX(uint8_t *dst, uint8_t *src, uint8_t *src2, int dstStride, int src2Stride){\
|
||||
OPNAME ## h264_qpel8_h_lowpass_l2_ ## MMX(dst , src , src2 , dstStride, src2Stride);\
|
||||
OPNAME ## h264_qpel8_h_lowpass_l2_ ## MMX(dst+8, src+8, src2+8, dstStride, src2Stride);\
|
||||
@ -1385,7 +1390,7 @@ static av_noinline void OPNAME ## h264_qpel16_h_lowpass_l2_ ## MMX(uint8_t *dst,
|
||||
}
|
||||
#endif // ARCH_X86_64
|
||||
|
||||
#define QPEL_H264_XMM(OPNAME, OP, MMX)\
|
||||
#define QPEL_H264_H_XMM(OPNAME, OP, MMX)\
|
||||
static av_noinline void OPNAME ## h264_qpel8_h_lowpass_l2_ ## MMX(uint8_t *dst, uint8_t *src, uint8_t *src2, int dstStride, int src2Stride){\
|
||||
int h=8;\
|
||||
asm volatile(\
|
||||
@ -1431,8 +1436,58 @@ static av_noinline void OPNAME ## h264_qpel8_h_lowpass_l2_ ## MMX(uint8_t *dst,
|
||||
);\
|
||||
}while(--h);\
|
||||
}\
|
||||
QPEL_H264_HL2_XMM(OPNAME, OP, MMX)\
|
||||
QPEL_H264_H16_XMM(OPNAME, OP, MMX)\
|
||||
\
|
||||
static av_noinline void OPNAME ## h264_qpel8_h_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
int h=8;\
|
||||
asm volatile(\
|
||||
"pxor %%xmm7, %%xmm7 \n\t"\
|
||||
"movdqa %5, %%xmm6 \n\t"\
|
||||
"1: \n\t"\
|
||||
"lddqu -5(%0), %%xmm1 \n\t"\
|
||||
"movdqa %%xmm1, %%xmm0 \n\t"\
|
||||
"punpckhbw %%xmm7, %%xmm1 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm0 \n\t"\
|
||||
"movdqa %%xmm1, %%xmm2 \n\t"\
|
||||
"movdqa %%xmm1, %%xmm3 \n\t"\
|
||||
"movdqa %%xmm1, %%xmm4 \n\t"\
|
||||
"movdqa %%xmm1, %%xmm5 \n\t"\
|
||||
"palignr $6, %%xmm0, %%xmm5 \n\t"\
|
||||
"palignr $8, %%xmm0, %%xmm4 \n\t"\
|
||||
"palignr $10,%%xmm0, %%xmm3 \n\t"\
|
||||
"paddw %%xmm1, %%xmm5 \n\t"\
|
||||
"palignr $12,%%xmm0, %%xmm2 \n\t"\
|
||||
"palignr $14,%%xmm0, %%xmm1 \n\t"\
|
||||
"paddw %%xmm3, %%xmm2 \n\t"\
|
||||
"paddw %%xmm4, %%xmm1 \n\t"\
|
||||
"psllw $2, %%xmm2 \n\t"\
|
||||
"psubw %%xmm1, %%xmm2 \n\t"\
|
||||
"paddw %6, %%xmm5 \n\t"\
|
||||
"pmullw %%xmm6, %%xmm2 \n\t"\
|
||||
"paddw %%xmm5, %%xmm2 \n\t"\
|
||||
"psraw $5, %%xmm2 \n\t"\
|
||||
"packuswb %%xmm2, %%xmm2 \n\t"\
|
||||
OP(%%xmm2, (%1), %%xmm4, q)\
|
||||
"add %3, %0 \n\t"\
|
||||
"add %4, %1 \n\t"\
|
||||
"decl %2 \n\t"\
|
||||
" jnz 1b \n\t"\
|
||||
: "+a"(src), "+c"(dst), "+g"(h)\
|
||||
: "D"((long)srcStride), "S"((long)dstStride),\
|
||||
"m"(ff_pw_5), "m"(ff_pw_16)\
|
||||
: "memory"\
|
||||
);\
|
||||
}\
|
||||
static void OPNAME ## h264_qpel16_h_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst , src , dstStride, srcStride);\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride);\
|
||||
src += 8*srcStride;\
|
||||
dst += 8*dstStride;\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst , src , dstStride, srcStride);\
|
||||
OPNAME ## h264_qpel8_h_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride);\
|
||||
}\
|
||||
|
||||
#define QPEL_H264_V_XMM(OPNAME, OP, MMX)\
|
||||
static av_noinline void OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride, int h){\
|
||||
src -= 2*srcStride;\
|
||||
\
|
||||
@ -1483,61 +1538,71 @@ static av_noinline void OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(uint8_t *dst,
|
||||
);\
|
||||
}\
|
||||
}\
|
||||
static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride, int size){\
|
||||
static void OPNAME ## h264_qpel8_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst , src , dstStride, srcStride, 8);\
|
||||
}\
|
||||
static av_noinline void OPNAME ## h264_qpel16_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst , src , dstStride, srcStride, 16);\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride, 16);\
|
||||
}
|
||||
|
||||
static av_always_inline void put_h264_qpel8or16_hv1_lowpass_sse2(int16_t *tmp, uint8_t *src, int tmpStride, int srcStride, int size){
|
||||
int w = (size+8)>>3;
|
||||
src -= 2*srcStride+2;
|
||||
while(w--){
|
||||
asm volatile(
|
||||
"pxor %%xmm7, %%xmm7 \n\t"
|
||||
"movq (%0), %%xmm0 \n\t"
|
||||
"add %2, %0 \n\t"
|
||||
"movq (%0), %%xmm1 \n\t"
|
||||
"add %2, %0 \n\t"
|
||||
"movq (%0), %%xmm2 \n\t"
|
||||
"add %2, %0 \n\t"
|
||||
"movq (%0), %%xmm3 \n\t"
|
||||
"add %2, %0 \n\t"
|
||||
"movq (%0), %%xmm4 \n\t"
|
||||
"add %2, %0 \n\t"
|
||||
"punpcklbw %%xmm7, %%xmm0 \n\t"
|
||||
"punpcklbw %%xmm7, %%xmm1 \n\t"
|
||||
"punpcklbw %%xmm7, %%xmm2 \n\t"
|
||||
"punpcklbw %%xmm7, %%xmm3 \n\t"
|
||||
"punpcklbw %%xmm7, %%xmm4 \n\t"
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 0*48)
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 1*48)
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 2*48)
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 3*48)
|
||||
QPEL_H264HV_XMM(%%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, 4*48)
|
||||
QPEL_H264HV_XMM(%%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, 5*48)
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 6*48)
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 7*48)
|
||||
: "+a"(src)
|
||||
: "c"(tmp), "S"((long)srcStride), "m"(ff_pw_5), "m"(ff_pw_16)
|
||||
: "memory"
|
||||
);
|
||||
if(size==16){
|
||||
asm volatile(
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 8*48)
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 9*48)
|
||||
QPEL_H264HV_XMM(%%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, 10*48)
|
||||
QPEL_H264HV_XMM(%%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, 11*48)
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 12*48)
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 13*48)
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 14*48)
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 15*48)
|
||||
: "+a"(src)
|
||||
: "c"(tmp), "S"((long)srcStride), "m"(ff_pw_5), "m"(ff_pw_16)
|
||||
: "memory"
|
||||
);
|
||||
}
|
||||
tmp += 8;
|
||||
src += 8 - (size+5)*srcStride;
|
||||
}
|
||||
}
|
||||
|
||||
#define QPEL_H264_HV2_XMM(OPNAME, OP, MMX)\
|
||||
static av_always_inline void OPNAME ## h264_qpel8or16_hv2_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, int dstStride, int tmpStride, int size){\
|
||||
int h = size;\
|
||||
int w = (size+8)>>3;\
|
||||
src -= 2*srcStride+2;\
|
||||
while(w--){\
|
||||
asm volatile(\
|
||||
"pxor %%xmm7, %%xmm7 \n\t"\
|
||||
"movq (%0), %%xmm0 \n\t"\
|
||||
"add %2, %0 \n\t"\
|
||||
"movq (%0), %%xmm1 \n\t"\
|
||||
"add %2, %0 \n\t"\
|
||||
"movq (%0), %%xmm2 \n\t"\
|
||||
"add %2, %0 \n\t"\
|
||||
"movq (%0), %%xmm3 \n\t"\
|
||||
"add %2, %0 \n\t"\
|
||||
"movq (%0), %%xmm4 \n\t"\
|
||||
"add %2, %0 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm0 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm1 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm2 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm3 \n\t"\
|
||||
"punpcklbw %%xmm7, %%xmm4 \n\t"\
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 0*48)\
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 1*48)\
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 2*48)\
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 3*48)\
|
||||
QPEL_H264HV_XMM(%%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, 4*48)\
|
||||
QPEL_H264HV_XMM(%%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, 5*48)\
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 6*48)\
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 7*48)\
|
||||
: "+a"(src)\
|
||||
: "c"(tmp), "S"((long)srcStride), "m"(ff_pw_5), "m"(ff_pw_16)\
|
||||
: "memory"\
|
||||
);\
|
||||
if(size==16){\
|
||||
asm volatile(\
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 8*48)\
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 9*48)\
|
||||
QPEL_H264HV_XMM(%%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, 10*48)\
|
||||
QPEL_H264HV_XMM(%%xmm5, %%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, 11*48)\
|
||||
QPEL_H264HV_XMM(%%xmm0, %%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, 12*48)\
|
||||
QPEL_H264HV_XMM(%%xmm1, %%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, 13*48)\
|
||||
QPEL_H264HV_XMM(%%xmm2, %%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, 14*48)\
|
||||
QPEL_H264HV_XMM(%%xmm3, %%xmm4, %%xmm5, %%xmm0, %%xmm1, %%xmm2, 15*48)\
|
||||
: "+a"(src)\
|
||||
: "c"(tmp), "S"((long)srcStride), "m"(ff_pw_5), "m"(ff_pw_16)\
|
||||
: "memory"\
|
||||
);\
|
||||
}\
|
||||
tmp += 8;\
|
||||
src += 8 - (size+5)*srcStride;\
|
||||
}\
|
||||
tmp -= size+8;\
|
||||
h = size;\
|
||||
if(size==16){\
|
||||
if(size == 16){\
|
||||
asm volatile(\
|
||||
"1: \n\t"\
|
||||
"movdqa 32(%0), %%xmm4 \n\t"\
|
||||
@ -1627,47 +1692,72 @@ static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst
|
||||
: "memory"\
|
||||
);\
|
||||
}\
|
||||
}\
|
||||
\
|
||||
static void OPNAME ## h264_qpel8_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst , src , dstStride, srcStride, 8);\
|
||||
}\
|
||||
static av_noinline void OPNAME ## h264_qpel16_v_lowpass_ ## MMX(uint8_t *dst, uint8_t *src, int dstStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst , src , dstStride, srcStride, 16);\
|
||||
OPNAME ## h264_qpel8or16_v_lowpass_ ## MMX(dst+8, src+8, dstStride, srcStride, 16);\
|
||||
}
|
||||
|
||||
#define QPEL_H264_HV_XMM(OPNAME, OP, MMX)\
|
||||
static av_noinline void OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride, int size){\
|
||||
put_h264_qpel8or16_hv1_lowpass_sse2(tmp, src, tmpStride, srcStride, size);\
|
||||
OPNAME ## h264_qpel8or16_hv2_lowpass_ ## MMX(dst, tmp, dstStride, tmpStride, size);\
|
||||
}\
|
||||
static void OPNAME ## h264_qpel8_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(dst , tmp , src , dstStride, tmpStride, srcStride, 8);\
|
||||
OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(dst, tmp, src, dstStride, tmpStride, srcStride, 8);\
|
||||
}\
|
||||
static void OPNAME ## h264_qpel16_hv_lowpass_ ## MMX(uint8_t *dst, int16_t *tmp, uint8_t *src, int dstStride, int tmpStride, int srcStride){\
|
||||
OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(dst , tmp , src , dstStride, tmpStride, srcStride, 16);\
|
||||
OPNAME ## h264_qpel8or16_hv_lowpass_ ## MMX(dst, tmp, src, dstStride, tmpStride, srcStride, 16);\
|
||||
}\
|
||||
|
||||
#define put_pixels8_ssse3 put_pixels8_mmx2
|
||||
#define put_pixels16_ssse3 put_pixels16_sse2
|
||||
#define avg_pixels8_ssse3 avg_pixels8_mmx2
|
||||
#define avg_pixels16_ssse3 avg_pixels16_sse2
|
||||
|
||||
#define put_pixels8_l2_sse2 put_pixels8_l2_mmx2
|
||||
#define avg_pixels8_l2_sse2 avg_pixels8_l2_mmx2
|
||||
#define put_pixels16_l2_sse2 put_pixels16_l2_mmx2
|
||||
#define avg_pixels16_l2_sse2 avg_pixels16_l2_mmx2
|
||||
#define put_pixels8_l2_ssse3 put_pixels8_l2_mmx2
|
||||
#define put_pixels16_l2_ssse3 put_pixels16_l2_mmx2
|
||||
#define avg_pixels8_l2_ssse3 avg_pixels8_l2_mmx2
|
||||
#define put_pixels16_l2_ssse3 put_pixels16_l2_mmx2
|
||||
#define avg_pixels16_l2_ssse3 avg_pixels16_l2_mmx2
|
||||
|
||||
#define put_pixels8_l2_shift5_sse2 put_pixels8_l2_shift5_mmx2
|
||||
#define avg_pixels8_l2_shift5_sse2 avg_pixels8_l2_shift5_mmx2
|
||||
#define put_pixels16_l2_shift5_sse2 put_pixels16_l2_shift5_mmx2
|
||||
#define avg_pixels16_l2_shift5_sse2 avg_pixels16_l2_shift5_mmx2
|
||||
#define put_pixels8_l2_shift5_ssse3 put_pixels8_l2_shift5_mmx2
|
||||
#define put_pixels16_l2_shift5_ssse3 put_pixels16_l2_shift5_mmx2
|
||||
#define avg_pixels8_l2_shift5_ssse3 avg_pixels8_l2_shift5_mmx2
|
||||
#define put_pixels16_l2_shift5_ssse3 put_pixels16_l2_shift5_mmx2
|
||||
#define avg_pixels16_l2_shift5_ssse3 avg_pixels16_l2_shift5_mmx2
|
||||
|
||||
#define put_h264_qpel8_h_lowpass_ssse3 put_h264_qpel8_h_lowpass_mmx2
|
||||
#define put_h264_qpel16_h_lowpass_ssse3 put_h264_qpel16_h_lowpass_mmx2
|
||||
#define avg_h264_qpel8_h_lowpass_ssse3 avg_h264_qpel8_h_lowpass_mmx2
|
||||
#define avg_h264_qpel16_h_lowpass_ssse3 avg_h264_qpel16_h_lowpass_mmx2
|
||||
#define put_h264_qpel8_h_lowpass_l2_sse2 put_h264_qpel8_h_lowpass_l2_mmx2
|
||||
#define avg_h264_qpel8_h_lowpass_l2_sse2 avg_h264_qpel8_h_lowpass_l2_mmx2
|
||||
#define put_h264_qpel16_h_lowpass_l2_sse2 put_h264_qpel16_h_lowpass_l2_mmx2
|
||||
#define avg_h264_qpel16_h_lowpass_l2_sse2 avg_h264_qpel16_h_lowpass_l2_mmx2
|
||||
|
||||
#define put_h264_qpel8_v_lowpass_ssse3 put_h264_qpel8_v_lowpass_sse2
|
||||
#define avg_h264_qpel8_v_lowpass_ssse3 avg_h264_qpel8_v_lowpass_sse2
|
||||
#define put_h264_qpel16_v_lowpass_ssse3 put_h264_qpel16_v_lowpass_sse2
|
||||
#define avg_h264_qpel16_v_lowpass_ssse3 avg_h264_qpel16_v_lowpass_sse2
|
||||
|
||||
#define put_h264_qpel8or16_hv2_lowpass_sse2 put_h264_qpel8or16_hv2_lowpass_mmx2
|
||||
#define avg_h264_qpel8or16_hv2_lowpass_sse2 avg_h264_qpel8or16_hv2_lowpass_mmx2
|
||||
|
||||
#define H264_MC(OPNAME, SIZE, MMX, ALIGN) \
|
||||
H264_MC_C(OPNAME, SIZE, MMX, ALIGN)\
|
||||
H264_MC_V(OPNAME, SIZE, MMX, ALIGN)\
|
||||
H264_MC_H(OPNAME, SIZE, MMX, ALIGN)\
|
||||
H264_MC_HV(OPNAME, SIZE, MMX, ALIGN)\
|
||||
|
||||
static void put_h264_qpel16_mc00_sse2 (uint8_t *dst, uint8_t *src, int stride){
|
||||
put_pixels16_sse2(dst, src, stride, 16);
|
||||
}
|
||||
static void avg_h264_qpel16_mc00_sse2 (uint8_t *dst, uint8_t *src, int stride){
|
||||
avg_pixels16_sse2(dst, src, stride, 16);
|
||||
}
|
||||
#define put_h264_qpel8_mc00_sse2 put_h264_qpel8_mc00_mmx2
|
||||
#define avg_h264_qpel8_mc00_sse2 avg_h264_qpel8_mc00_mmx2
|
||||
|
||||
#define H264_MC_C(OPNAME, SIZE, MMX, ALIGN) \
|
||||
static void OPNAME ## h264_qpel ## SIZE ## _mc00_ ## MMX (uint8_t *dst, uint8_t *src, int stride){\
|
||||
OPNAME ## pixels ## SIZE ## _ ## MMX(dst, src, stride, SIZE);\
|
||||
}\
|
||||
\
|
||||
|
||||
#define H264_MC_H(OPNAME, SIZE, MMX, ALIGN) \
|
||||
static void OPNAME ## h264_qpel ## SIZE ## _mc10_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
||||
OPNAME ## h264_qpel ## SIZE ## _h_lowpass_l2_ ## MMX(dst, src, src, stride, stride);\
|
||||
}\
|
||||
@ -1679,7 +1769,8 @@ static void OPNAME ## h264_qpel ## SIZE ## _mc20_ ## MMX(uint8_t *dst, uint8_t *
|
||||
static void OPNAME ## h264_qpel ## SIZE ## _mc30_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
||||
OPNAME ## h264_qpel ## SIZE ## _h_lowpass_l2_ ## MMX(dst, src, src+1, stride, stride);\
|
||||
}\
|
||||
\
|
||||
|
||||
#define H264_MC_V(OPNAME, SIZE, MMX, ALIGN) \
|
||||
static void OPNAME ## h264_qpel ## SIZE ## _mc01_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
||||
DECLARE_ALIGNED(ALIGN, uint8_t, temp[SIZE*SIZE]);\
|
||||
put_h264_qpel ## SIZE ## _v_lowpass_ ## MMX(temp, src, SIZE, stride);\
|
||||
@ -1695,7 +1786,8 @@ static void OPNAME ## h264_qpel ## SIZE ## _mc03_ ## MMX(uint8_t *dst, uint8_t *
|
||||
put_h264_qpel ## SIZE ## _v_lowpass_ ## MMX(temp, src, SIZE, stride);\
|
||||
OPNAME ## pixels ## SIZE ## _l2_ ## MMX(dst, src+stride, temp, stride, stride, SIZE);\
|
||||
}\
|
||||
\
|
||||
|
||||
#define H264_MC_HV(OPNAME, SIZE, MMX, ALIGN) \
|
||||
static void OPNAME ## h264_qpel ## SIZE ## _mc11_ ## MMX(uint8_t *dst, uint8_t *src, int stride){\
|
||||
DECLARE_ALIGNED(ALIGN, uint8_t, temp[SIZE*SIZE]);\
|
||||
put_h264_qpel ## SIZE ## _v_lowpass_ ## MMX(temp, src, SIZE, stride);\
|
||||
@ -1761,6 +1853,20 @@ static void OPNAME ## h264_qpel ## SIZE ## _mc32_ ## MMX(uint8_t *dst, uint8_t *
|
||||
OPNAME ## pixels ## SIZE ## _l2_shift5_ ## MMX(dst, halfV+3, halfHV, stride, SIZE, SIZE);\
|
||||
}\
|
||||
|
||||
#define H264_MC_4816(MMX)\
|
||||
H264_MC(put_, 4, MMX, 8)\
|
||||
H264_MC(put_, 8, MMX, 8)\
|
||||
H264_MC(put_, 16,MMX, 8)\
|
||||
H264_MC(avg_, 4, MMX, 8)\
|
||||
H264_MC(avg_, 8, MMX, 8)\
|
||||
H264_MC(avg_, 16,MMX, 8)\
|
||||
|
||||
#define H264_MC_816(QPEL, XMM)\
|
||||
QPEL(put_, 8, XMM, 16)\
|
||||
QPEL(put_, 16,XMM, 16)\
|
||||
QPEL(avg_, 8, XMM, 16)\
|
||||
QPEL(avg_, 16,XMM, 16)\
|
||||
|
||||
|
||||
#define AVG_3DNOW_OP(a,b,temp, size) \
|
||||
"mov" #size " " #b ", " #temp " \n\t"\
|
||||
@ -1778,38 +1884,29 @@ QPEL_H264(avg_, AVG_3DNOW_OP, 3dnow)
|
||||
#define PAVGB "pavgb"
|
||||
QPEL_H264(put_, PUT_OP, mmx2)
|
||||
QPEL_H264(avg_, AVG_MMX2_OP, mmx2)
|
||||
QPEL_H264_V_XMM(put_, PUT_OP, sse2)
|
||||
QPEL_H264_V_XMM(avg_, AVG_MMX2_OP, sse2)
|
||||
QPEL_H264_HV_XMM(put_, PUT_OP, sse2)
|
||||
QPEL_H264_HV_XMM(avg_, AVG_MMX2_OP, sse2)
|
||||
#ifdef HAVE_SSSE3
|
||||
QPEL_H264_XMM(put_, PUT_OP, ssse3)
|
||||
QPEL_H264_XMM(avg_, AVG_MMX2_OP, ssse3)
|
||||
QPEL_H264_H_XMM(put_, PUT_OP, ssse3)
|
||||
QPEL_H264_H_XMM(avg_, AVG_MMX2_OP, ssse3)
|
||||
QPEL_H264_HV2_XMM(put_, PUT_OP, ssse3)
|
||||
QPEL_H264_HV2_XMM(avg_, AVG_MMX2_OP, ssse3)
|
||||
QPEL_H264_HV_XMM(put_, PUT_OP, ssse3)
|
||||
QPEL_H264_HV_XMM(avg_, AVG_MMX2_OP, ssse3)
|
||||
#endif
|
||||
#undef PAVGB
|
||||
|
||||
H264_MC(put_, 4, 3dnow, 8)
|
||||
H264_MC(put_, 8, 3dnow, 8)
|
||||
H264_MC(put_, 16,3dnow, 8)
|
||||
H264_MC(avg_, 4, 3dnow, 8)
|
||||
H264_MC(avg_, 8, 3dnow, 8)
|
||||
H264_MC(avg_, 16,3dnow, 8)
|
||||
H264_MC(put_, 4, mmx2, 8)
|
||||
H264_MC(put_, 8, mmx2, 8)
|
||||
H264_MC(put_, 16,mmx2, 8)
|
||||
H264_MC(avg_, 4, mmx2, 8)
|
||||
H264_MC(avg_, 8, mmx2, 8)
|
||||
H264_MC(avg_, 16,mmx2, 8)
|
||||
H264_MC_4816(3dnow)
|
||||
H264_MC_4816(mmx2)
|
||||
H264_MC_816(H264_MC_V, sse2)
|
||||
H264_MC_816(H264_MC_HV, sse2)
|
||||
#ifdef HAVE_SSSE3
|
||||
// FIXME some of these only require sse2
|
||||
H264_MC(put_, 8, ssse3, 16)
|
||||
H264_MC(put_, 16,ssse3, 16)
|
||||
H264_MC(avg_, 8, ssse3, 16)
|
||||
H264_MC(avg_, 16,ssse3, 16)
|
||||
H264_MC_816(H264_MC_H, ssse3)
|
||||
H264_MC_816(H264_MC_HV, ssse3)
|
||||
#endif
|
||||
|
||||
// no mc20 because it wasn't significantly faster than mmx2
|
||||
#define put_h264_qpel8_mc20_ssse3 put_h264_qpel8_mc20_mmx2
|
||||
#define put_h264_qpel16_mc20_ssse3 put_h264_qpel16_mc20_mmx2
|
||||
#define avg_h264_qpel8_mc20_ssse3 avg_h264_qpel8_mc20_mmx2
|
||||
#define avg_h264_qpel16_mc20_ssse3 avg_h264_qpel16_mc20_mmx2
|
||||
|
||||
|
||||
#define H264_CHROMA_OP(S,D)
|
||||
#define H264_CHROMA_OP4(S,D,T)
|
||||
|
Loading…
Reference in New Issue
Block a user