mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2025-01-29 22:00:58 +02:00
adapting existing mmx/mmx2/sse/3dnow optimizations so they work on x86_64
Originally committed as revision 13721 to svn://svn.mplayerhq.hu/mplayer/trunk/postproc
This commit is contained in:
parent
ed861c6bd2
commit
6e1c66bc06
@ -11,6 +11,7 @@
|
|||||||
#include "../config.h"
|
#include "../config.h"
|
||||||
#include "rgb2rgb.h"
|
#include "rgb2rgb.h"
|
||||||
#include "swscale.h"
|
#include "swscale.h"
|
||||||
|
#include "../cpudetect.h"
|
||||||
#include "../mangle.h"
|
#include "../mangle.h"
|
||||||
#include "../bswap.h"
|
#include "../bswap.h"
|
||||||
#include "../libvo/fastmemcpy.h"
|
#include "../libvo/fastmemcpy.h"
|
||||||
@ -68,7 +69,7 @@ void (*yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2, const uint8_t *sr
|
|||||||
int srcStride1, int srcStride2,
|
int srcStride1, int srcStride2,
|
||||||
int srcStride3, int dstStride);
|
int srcStride3, int dstStride);
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
static const uint64_t mmx_null __attribute__((aligned(8))) = 0x0000000000000000ULL;
|
static const uint64_t mmx_null __attribute__((aligned(8))) = 0x0000000000000000ULL;
|
||||||
static const uint64_t mmx_one __attribute__((aligned(8))) = 0xFFFFFFFFFFFFFFFFULL;
|
static const uint64_t mmx_one __attribute__((aligned(8))) = 0xFFFFFFFFFFFFFFFFULL;
|
||||||
static const uint64_t mask32b attribute_used __attribute__((aligned(8))) = 0x000000FF000000FFULL;
|
static const uint64_t mask32b attribute_used __attribute__((aligned(8))) = 0x000000FF000000FFULL;
|
||||||
@ -152,7 +153,7 @@ static uint64_t __attribute__((aligned(8))) dither8[2]={
|
|||||||
#define RENAME(a) a ## _C
|
#define RENAME(a) a ## _C
|
||||||
#include "rgb2rgb_template.c"
|
#include "rgb2rgb_template.c"
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
|
|
||||||
//MMX versions
|
//MMX versions
|
||||||
#undef RENAME
|
#undef RENAME
|
||||||
@ -181,7 +182,7 @@ static uint64_t __attribute__((aligned(8))) dither8[2]={
|
|||||||
#define RENAME(a) a ## _3DNOW
|
#define RENAME(a) a ## _3DNOW
|
||||||
#include "rgb2rgb_template.c"
|
#include "rgb2rgb_template.c"
|
||||||
|
|
||||||
#endif //ARCH_X86
|
#endif //ARCH_X86 || ARCH_X86_64
|
||||||
|
|
||||||
/*
|
/*
|
||||||
rgb15->rgb16 Original by Strepto/Astral
|
rgb15->rgb16 Original by Strepto/Astral
|
||||||
@ -191,7 +192,7 @@ static uint64_t __attribute__((aligned(8))) dither8[2]={
|
|||||||
*/
|
*/
|
||||||
|
|
||||||
void sws_rgb2rgb_init(int flags){
|
void sws_rgb2rgb_init(int flags){
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
if(flags & SWS_CPU_CAPS_MMX2){
|
if(flags & SWS_CPU_CAPS_MMX2){
|
||||||
rgb15to16= rgb15to16_MMX2;
|
rgb15to16= rgb15to16_MMX2;
|
||||||
rgb15to24= rgb15to24_MMX2;
|
rgb15to24= rgb15to24_MMX2;
|
||||||
|
@ -349,9 +349,9 @@ static inline void RENAME(rgb32to16)(const uint8_t *src, uint8_t *dst, unsigned
|
|||||||
"pslld $11, %%mm3 \n\t"
|
"pslld $11, %%mm3 \n\t"
|
||||||
"por %%mm3, %%mm0 \n\t"
|
"por %%mm3, %%mm0 \n\t"
|
||||||
MOVNTQ" %%mm0, (%0) \n\t"
|
MOVNTQ" %%mm0, (%0) \n\t"
|
||||||
"addl $16, %1 \n\t"
|
"add $16, %1 \n\t"
|
||||||
"addl $8, %0 \n\t"
|
"add $8, %0 \n\t"
|
||||||
"cmpl %2, %1 \n\t"
|
"cmp %2, %1 \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
: "+r" (d), "+r"(s)
|
: "+r" (d), "+r"(s)
|
||||||
: "r" (mm_end), "m" (mask3216g), "m" (mask3216br), "m" (mul3216)
|
: "r" (mm_end), "m" (mask3216g), "m" (mask3216br), "m" (mul3216)
|
||||||
@ -509,9 +509,9 @@ static inline void RENAME(rgb32to15)(const uint8_t *src, uint8_t *dst, unsigned
|
|||||||
"pslld $10, %%mm3 \n\t"
|
"pslld $10, %%mm3 \n\t"
|
||||||
"por %%mm3, %%mm0 \n\t"
|
"por %%mm3, %%mm0 \n\t"
|
||||||
MOVNTQ" %%mm0, (%0) \n\t"
|
MOVNTQ" %%mm0, (%0) \n\t"
|
||||||
"addl $16, %1 \n\t"
|
"add $16, %1 \n\t"
|
||||||
"addl $8, %0 \n\t"
|
"add $8, %0 \n\t"
|
||||||
"cmpl %2, %1 \n\t"
|
"cmp %2, %1 \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
: "+r" (d), "+r"(s)
|
: "+r" (d), "+r"(s)
|
||||||
: "r" (mm_end), "m" (mask3215g), "m" (mask3216br), "m" (mul3215)
|
: "r" (mm_end), "m" (mask3215g), "m" (mask3216br), "m" (mul3215)
|
||||||
@ -1345,11 +1345,11 @@ static inline void RENAME(rgb32tobgr32)(const uint8_t *src, uint8_t *dst, unsign
|
|||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
/* TODO: unroll this loop */
|
/* TODO: unroll this loop */
|
||||||
asm volatile (
|
asm volatile (
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 32(%0, %%eax) \n\t"
|
PREFETCH" 32(%0, %%"REG_a") \n\t"
|
||||||
"movq (%0, %%eax), %%mm0 \n\t"
|
"movq (%0, %%"REG_a"), %%mm0 \n\t"
|
||||||
"movq %%mm0, %%mm1 \n\t"
|
"movq %%mm0, %%mm1 \n\t"
|
||||||
"movq %%mm0, %%mm2 \n\t"
|
"movq %%mm0, %%mm2 \n\t"
|
||||||
"pslld $16, %%mm0 \n\t"
|
"pslld $16, %%mm0 \n\t"
|
||||||
@ -1359,12 +1359,12 @@ static inline void RENAME(rgb32tobgr32)(const uint8_t *src, uint8_t *dst, unsign
|
|||||||
"pand "MANGLE(mask32b)", %%mm1 \n\t"
|
"pand "MANGLE(mask32b)", %%mm1 \n\t"
|
||||||
"por %%mm0, %%mm2 \n\t"
|
"por %%mm0, %%mm2 \n\t"
|
||||||
"por %%mm1, %%mm2 \n\t"
|
"por %%mm1, %%mm2 \n\t"
|
||||||
MOVNTQ" %%mm2, (%1, %%eax) \n\t"
|
MOVNTQ" %%mm2, (%1, %%"REG_a") \n\t"
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
"cmpl %2, %%eax \n\t"
|
"cmp %2, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
:: "r" (src), "r"(dst), "r" (src_size-7)
|
:: "r" (src), "r"(dst), "r" ((long)src_size-7)
|
||||||
: "%eax"
|
: "%"REG_a
|
||||||
);
|
);
|
||||||
|
|
||||||
__asm __volatile(SFENCE:::"memory");
|
__asm __volatile(SFENCE:::"memory");
|
||||||
@ -1391,43 +1391,43 @@ static inline void RENAME(rgb24tobgr24)(const uint8_t *src, uint8_t *dst, unsign
|
|||||||
{
|
{
|
||||||
unsigned i;
|
unsigned i;
|
||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
int mmx_size= 23 - src_size;
|
long mmx_size= 23 - src_size;
|
||||||
asm volatile (
|
asm volatile (
|
||||||
"movq "MANGLE(mask24r)", %%mm5 \n\t"
|
"movq "MANGLE(mask24r)", %%mm5 \n\t"
|
||||||
"movq "MANGLE(mask24g)", %%mm6 \n\t"
|
"movq "MANGLE(mask24g)", %%mm6 \n\t"
|
||||||
"movq "MANGLE(mask24b)", %%mm7 \n\t"
|
"movq "MANGLE(mask24b)", %%mm7 \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 32(%1, %%eax) \n\t"
|
PREFETCH" 32(%1, %%"REG_a") \n\t"
|
||||||
"movq (%1, %%eax), %%mm0 \n\t" // BGR BGR BG
|
"movq (%1, %%"REG_a"), %%mm0 \n\t" // BGR BGR BG
|
||||||
"movq (%1, %%eax), %%mm1 \n\t" // BGR BGR BG
|
"movq (%1, %%"REG_a"), %%mm1 \n\t" // BGR BGR BG
|
||||||
"movq 2(%1, %%eax), %%mm2 \n\t" // R BGR BGR B
|
"movq 2(%1, %%"REG_a"), %%mm2 \n\t" // R BGR BGR B
|
||||||
"psllq $16, %%mm0 \n\t" // 00 BGR BGR
|
"psllq $16, %%mm0 \n\t" // 00 BGR BGR
|
||||||
"pand %%mm5, %%mm0 \n\t"
|
"pand %%mm5, %%mm0 \n\t"
|
||||||
"pand %%mm6, %%mm1 \n\t"
|
"pand %%mm6, %%mm1 \n\t"
|
||||||
"pand %%mm7, %%mm2 \n\t"
|
"pand %%mm7, %%mm2 \n\t"
|
||||||
"por %%mm0, %%mm1 \n\t"
|
"por %%mm0, %%mm1 \n\t"
|
||||||
"por %%mm2, %%mm1 \n\t"
|
"por %%mm2, %%mm1 \n\t"
|
||||||
"movq 6(%1, %%eax), %%mm0 \n\t" // BGR BGR BG
|
"movq 6(%1, %%"REG_a"), %%mm0 \n\t" // BGR BGR BG
|
||||||
MOVNTQ" %%mm1, (%2, %%eax) \n\t" // RGB RGB RG
|
MOVNTQ" %%mm1, (%2, %%"REG_a")\n\t" // RGB RGB RG
|
||||||
"movq 8(%1, %%eax), %%mm1 \n\t" // R BGR BGR B
|
"movq 8(%1, %%"REG_a"), %%mm1 \n\t" // R BGR BGR B
|
||||||
"movq 10(%1, %%eax), %%mm2 \n\t" // GR BGR BGR
|
"movq 10(%1, %%"REG_a"), %%mm2 \n\t" // GR BGR BGR
|
||||||
"pand %%mm7, %%mm0 \n\t"
|
"pand %%mm7, %%mm0 \n\t"
|
||||||
"pand %%mm5, %%mm1 \n\t"
|
"pand %%mm5, %%mm1 \n\t"
|
||||||
"pand %%mm6, %%mm2 \n\t"
|
"pand %%mm6, %%mm2 \n\t"
|
||||||
"por %%mm0, %%mm1 \n\t"
|
"por %%mm0, %%mm1 \n\t"
|
||||||
"por %%mm2, %%mm1 \n\t"
|
"por %%mm2, %%mm1 \n\t"
|
||||||
"movq 14(%1, %%eax), %%mm0 \n\t" // R BGR BGR B
|
"movq 14(%1, %%"REG_a"), %%mm0 \n\t" // R BGR BGR B
|
||||||
MOVNTQ" %%mm1, 8(%2, %%eax) \n\t" // B RGB RGB R
|
MOVNTQ" %%mm1, 8(%2, %%"REG_a")\n\t" // B RGB RGB R
|
||||||
"movq 16(%1, %%eax), %%mm1 \n\t" // GR BGR BGR
|
"movq 16(%1, %%"REG_a"), %%mm1 \n\t" // GR BGR BGR
|
||||||
"movq 18(%1, %%eax), %%mm2 \n\t" // BGR BGR BG
|
"movq 18(%1, %%"REG_a"), %%mm2 \n\t" // BGR BGR BG
|
||||||
"pand %%mm6, %%mm0 \n\t"
|
"pand %%mm6, %%mm0 \n\t"
|
||||||
"pand %%mm7, %%mm1 \n\t"
|
"pand %%mm7, %%mm1 \n\t"
|
||||||
"pand %%mm5, %%mm2 \n\t"
|
"pand %%mm5, %%mm2 \n\t"
|
||||||
"por %%mm0, %%mm1 \n\t"
|
"por %%mm0, %%mm1 \n\t"
|
||||||
"por %%mm2, %%mm1 \n\t"
|
"por %%mm2, %%mm1 \n\t"
|
||||||
MOVNTQ" %%mm1, 16(%2, %%eax) \n\t"
|
MOVNTQ" %%mm1, 16(%2, %%"REG_a")\n\t"
|
||||||
"addl $24, %%eax \n\t"
|
"add $24, %%"REG_a" \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
: "+a" (mmx_size)
|
: "+a" (mmx_size)
|
||||||
: "r" (src-mmx_size), "r"(dst-mmx_size)
|
: "r" (src-mmx_size), "r"(dst-mmx_size)
|
||||||
@ -1465,20 +1465,20 @@ static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *u
|
|||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
//FIXME handle 2 lines a once (fewer prefetch, reuse some chrom, but very likely limited by mem anyway)
|
//FIXME handle 2 lines a once (fewer prefetch, reuse some chrom, but very likely limited by mem anyway)
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 32(%1, %%eax, 2) \n\t"
|
PREFETCH" 32(%1, %%"REG_a", 2) \n\t"
|
||||||
PREFETCH" 32(%2, %%eax) \n\t"
|
PREFETCH" 32(%2, %%"REG_a") \n\t"
|
||||||
PREFETCH" 32(%3, %%eax) \n\t"
|
PREFETCH" 32(%3, %%"REG_a") \n\t"
|
||||||
"movq (%2, %%eax), %%mm0 \n\t" // U(0)
|
"movq (%2, %%"REG_a"), %%mm0 \n\t" // U(0)
|
||||||
"movq %%mm0, %%mm2 \n\t" // U(0)
|
"movq %%mm0, %%mm2 \n\t" // U(0)
|
||||||
"movq (%3, %%eax), %%mm1 \n\t" // V(0)
|
"movq (%3, %%"REG_a"), %%mm1 \n\t" // V(0)
|
||||||
"punpcklbw %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
"punpcklbw %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
||||||
"punpckhbw %%mm1, %%mm2 \n\t" // UVUV UVUV(8)
|
"punpckhbw %%mm1, %%mm2 \n\t" // UVUV UVUV(8)
|
||||||
|
|
||||||
"movq (%1, %%eax,2), %%mm3 \n\t" // Y(0)
|
"movq (%1, %%"REG_a",2), %%mm3 \n\t" // Y(0)
|
||||||
"movq 8(%1, %%eax,2), %%mm5 \n\t" // Y(8)
|
"movq 8(%1, %%"REG_a",2), %%mm5 \n\t" // Y(8)
|
||||||
"movq %%mm3, %%mm4 \n\t" // Y(0)
|
"movq %%mm3, %%mm4 \n\t" // Y(0)
|
||||||
"movq %%mm5, %%mm6 \n\t" // Y(8)
|
"movq %%mm5, %%mm6 \n\t" // Y(8)
|
||||||
"punpcklbw %%mm0, %%mm3 \n\t" // YUYV YUYV(0)
|
"punpcklbw %%mm0, %%mm3 \n\t" // YUYV YUYV(0)
|
||||||
@ -1486,16 +1486,16 @@ static inline void RENAME(yuvPlanartoyuy2)(const uint8_t *ysrc, const uint8_t *u
|
|||||||
"punpcklbw %%mm2, %%mm5 \n\t" // YUYV YUYV(8)
|
"punpcklbw %%mm2, %%mm5 \n\t" // YUYV YUYV(8)
|
||||||
"punpckhbw %%mm2, %%mm6 \n\t" // YUYV YUYV(12)
|
"punpckhbw %%mm2, %%mm6 \n\t" // YUYV YUYV(12)
|
||||||
|
|
||||||
MOVNTQ" %%mm3, (%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm3, (%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm4, 8(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm4, 8(%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm5, 16(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm5, 16(%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm6, 24(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm6, 24(%0, %%"REG_a", 4)\n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
"cmpl %4, %%eax \n\t"
|
"cmp %4, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" (chromWidth)
|
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" ((long)chromWidth)
|
||||||
: "%eax"
|
: "%"REG_a
|
||||||
);
|
);
|
||||||
#else
|
#else
|
||||||
|
|
||||||
@ -1618,20 +1618,20 @@ static inline void RENAME(yuvPlanartouyvy)(const uint8_t *ysrc, const uint8_t *u
|
|||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
//FIXME handle 2 lines a once (fewer prefetch, reuse some chrom, but very likely limited by mem anyway)
|
//FIXME handle 2 lines a once (fewer prefetch, reuse some chrom, but very likely limited by mem anyway)
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 32(%1, %%eax, 2) \n\t"
|
PREFETCH" 32(%1, %%"REG_a", 2) \n\t"
|
||||||
PREFETCH" 32(%2, %%eax) \n\t"
|
PREFETCH" 32(%2, %%"REG_a") \n\t"
|
||||||
PREFETCH" 32(%3, %%eax) \n\t"
|
PREFETCH" 32(%3, %%"REG_a") \n\t"
|
||||||
"movq (%2, %%eax), %%mm0 \n\t" // U(0)
|
"movq (%2, %%"REG_a"), %%mm0 \n\t" // U(0)
|
||||||
"movq %%mm0, %%mm2 \n\t" // U(0)
|
"movq %%mm0, %%mm2 \n\t" // U(0)
|
||||||
"movq (%3, %%eax), %%mm1 \n\t" // V(0)
|
"movq (%3, %%"REG_a"), %%mm1 \n\t" // V(0)
|
||||||
"punpcklbw %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
"punpcklbw %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
||||||
"punpckhbw %%mm1, %%mm2 \n\t" // UVUV UVUV(8)
|
"punpckhbw %%mm1, %%mm2 \n\t" // UVUV UVUV(8)
|
||||||
|
|
||||||
"movq (%1, %%eax,2), %%mm3 \n\t" // Y(0)
|
"movq (%1, %%"REG_a",2), %%mm3 \n\t" // Y(0)
|
||||||
"movq 8(%1, %%eax,2), %%mm5 \n\t" // Y(8)
|
"movq 8(%1, %%"REG_a",2), %%mm5 \n\t" // Y(8)
|
||||||
"movq %%mm0, %%mm4 \n\t" // Y(0)
|
"movq %%mm0, %%mm4 \n\t" // Y(0)
|
||||||
"movq %%mm2, %%mm6 \n\t" // Y(8)
|
"movq %%mm2, %%mm6 \n\t" // Y(8)
|
||||||
"punpcklbw %%mm3, %%mm0 \n\t" // YUYV YUYV(0)
|
"punpcklbw %%mm3, %%mm0 \n\t" // YUYV YUYV(0)
|
||||||
@ -1639,16 +1639,16 @@ static inline void RENAME(yuvPlanartouyvy)(const uint8_t *ysrc, const uint8_t *u
|
|||||||
"punpcklbw %%mm5, %%mm2 \n\t" // YUYV YUYV(8)
|
"punpcklbw %%mm5, %%mm2 \n\t" // YUYV YUYV(8)
|
||||||
"punpckhbw %%mm5, %%mm6 \n\t" // YUYV YUYV(12)
|
"punpckhbw %%mm5, %%mm6 \n\t" // YUYV YUYV(12)
|
||||||
|
|
||||||
MOVNTQ" %%mm0, (%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm0, (%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm4, 8(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm4, 8(%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm2, 16(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm2, 16(%0, %%"REG_a", 4)\n\t"
|
||||||
MOVNTQ" %%mm6, 24(%0, %%eax, 4) \n\t"
|
MOVNTQ" %%mm6, 24(%0, %%"REG_a", 4)\n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
"cmpl %4, %%eax \n\t"
|
"cmp %4, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" (chromWidth)
|
::"r"(dst), "r"(ysrc), "r"(usrc), "r"(vsrc), "g" ((long)chromWidth)
|
||||||
: "%eax"
|
: "%"REG_a
|
||||||
);
|
);
|
||||||
#else
|
#else
|
||||||
//FIXME adapt the alpha asm code from yv12->yuy2
|
//FIXME adapt the alpha asm code from yv12->yuy2
|
||||||
@ -1740,14 +1740,14 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t
|
|||||||
{
|
{
|
||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
"pcmpeqw %%mm7, %%mm7 \n\t"
|
"pcmpeqw %%mm7, %%mm7 \n\t"
|
||||||
"psrlw $8, %%mm7 \n\t" // FF,00,FF,00...
|
"psrlw $8, %%mm7 \n\t" // FF,00,FF,00...
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%0, %%eax, 4) \n\t"
|
PREFETCH" 64(%0, %%"REG_a", 4) \n\t"
|
||||||
"movq (%0, %%eax, 4), %%mm0 \n\t" // YUYV YUYV(0)
|
"movq (%0, %%"REG_a", 4), %%mm0 \n\t" // YUYV YUYV(0)
|
||||||
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(4)
|
"movq 8(%0, %%"REG_a", 4), %%mm1\n\t" // YUYV YUYV(4)
|
||||||
"movq %%mm0, %%mm2 \n\t" // YUYV YUYV(0)
|
"movq %%mm0, %%mm2 \n\t" // YUYV YUYV(0)
|
||||||
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(4)
|
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(4)
|
||||||
"psrlw $8, %%mm0 \n\t" // U0V0 U0V0(0)
|
"psrlw $8, %%mm0 \n\t" // U0V0 U0V0(0)
|
||||||
@ -1757,10 +1757,10 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t
|
|||||||
"packuswb %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
"packuswb %%mm1, %%mm0 \n\t" // UVUV UVUV(0)
|
||||||
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(0)
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(0)
|
||||||
|
|
||||||
MOVNTQ" %%mm2, (%1, %%eax, 2) \n\t"
|
MOVNTQ" %%mm2, (%1, %%"REG_a", 2)\n\t"
|
||||||
|
|
||||||
"movq 16(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(8)
|
"movq 16(%0, %%"REG_a", 4), %%mm1\n\t" // YUYV YUYV(8)
|
||||||
"movq 24(%0, %%eax, 4), %%mm2 \n\t" // YUYV YUYV(12)
|
"movq 24(%0, %%"REG_a", 4), %%mm2\n\t" // YUYV YUYV(12)
|
||||||
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(8)
|
"movq %%mm1, %%mm3 \n\t" // YUYV YUYV(8)
|
||||||
"movq %%mm2, %%mm4 \n\t" // YUYV YUYV(12)
|
"movq %%mm2, %%mm4 \n\t" // YUYV YUYV(12)
|
||||||
"psrlw $8, %%mm1 \n\t" // U0V0 U0V0(8)
|
"psrlw $8, %%mm1 \n\t" // U0V0 U0V0(8)
|
||||||
@ -1770,7 +1770,7 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t
|
|||||||
"packuswb %%mm2, %%mm1 \n\t" // UVUV UVUV(8)
|
"packuswb %%mm2, %%mm1 \n\t" // UVUV UVUV(8)
|
||||||
"packuswb %%mm4, %%mm3 \n\t" // YYYY YYYY(8)
|
"packuswb %%mm4, %%mm3 \n\t" // YYYY YYYY(8)
|
||||||
|
|
||||||
MOVNTQ" %%mm3, 8(%1, %%eax, 2) \n\t"
|
MOVNTQ" %%mm3, 8(%1, %%"REG_a", 2)\n\t"
|
||||||
|
|
||||||
"movq %%mm0, %%mm2 \n\t" // UVUV UVUV(0)
|
"movq %%mm0, %%mm2 \n\t" // UVUV UVUV(0)
|
||||||
"movq %%mm1, %%mm3 \n\t" // UVUV UVUV(8)
|
"movq %%mm1, %%mm3 \n\t" // UVUV UVUV(8)
|
||||||
@ -1781,28 +1781,28 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t
|
|||||||
"packuswb %%mm1, %%mm0 \n\t" // VVVV VVVV(0)
|
"packuswb %%mm1, %%mm0 \n\t" // VVVV VVVV(0)
|
||||||
"packuswb %%mm3, %%mm2 \n\t" // UUUU UUUU(0)
|
"packuswb %%mm3, %%mm2 \n\t" // UUUU UUUU(0)
|
||||||
|
|
||||||
MOVNTQ" %%mm0, (%3, %%eax) \n\t"
|
MOVNTQ" %%mm0, (%3, %%"REG_a") \n\t"
|
||||||
MOVNTQ" %%mm2, (%2, %%eax) \n\t"
|
MOVNTQ" %%mm2, (%2, %%"REG_a") \n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
"cmpl %4, %%eax \n\t"
|
"cmp %4, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" (chromWidth)
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" ((long)chromWidth)
|
||||||
: "memory", "%eax"
|
: "memory", "%"REG_a
|
||||||
);
|
);
|
||||||
|
|
||||||
ydst += lumStride;
|
ydst += lumStride;
|
||||||
src += srcStride;
|
src += srcStride;
|
||||||
|
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%0, %%eax, 4) \n\t"
|
PREFETCH" 64(%0, %%"REG_a", 4) \n\t"
|
||||||
"movq (%0, %%eax, 4), %%mm0 \n\t" // YUYV YUYV(0)
|
"movq (%0, %%"REG_a", 4), %%mm0 \n\t" // YUYV YUYV(0)
|
||||||
"movq 8(%0, %%eax, 4), %%mm1 \n\t" // YUYV YUYV(4)
|
"movq 8(%0, %%"REG_a", 4), %%mm1\n\t" // YUYV YUYV(4)
|
||||||
"movq 16(%0, %%eax, 4), %%mm2 \n\t" // YUYV YUYV(8)
|
"movq 16(%0, %%"REG_a", 4), %%mm2\n\t" // YUYV YUYV(8)
|
||||||
"movq 24(%0, %%eax, 4), %%mm3 \n\t" // YUYV YUYV(12)
|
"movq 24(%0, %%"REG_a", 4), %%mm3\n\t" // YUYV YUYV(12)
|
||||||
"pand %%mm7, %%mm0 \n\t" // Y0Y0 Y0Y0(0)
|
"pand %%mm7, %%mm0 \n\t" // Y0Y0 Y0Y0(0)
|
||||||
"pand %%mm7, %%mm1 \n\t" // Y0Y0 Y0Y0(4)
|
"pand %%mm7, %%mm1 \n\t" // Y0Y0 Y0Y0(4)
|
||||||
"pand %%mm7, %%mm2 \n\t" // Y0Y0 Y0Y0(8)
|
"pand %%mm7, %%mm2 \n\t" // Y0Y0 Y0Y0(8)
|
||||||
@ -1810,15 +1810,15 @@ static inline void RENAME(yuy2toyv12)(const uint8_t *src, uint8_t *ydst, uint8_t
|
|||||||
"packuswb %%mm1, %%mm0 \n\t" // YYYY YYYY(0)
|
"packuswb %%mm1, %%mm0 \n\t" // YYYY YYYY(0)
|
||||||
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(8)
|
"packuswb %%mm3, %%mm2 \n\t" // YYYY YYYY(8)
|
||||||
|
|
||||||
MOVNTQ" %%mm0, (%1, %%eax, 2) \n\t"
|
MOVNTQ" %%mm0, (%1, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm2, 8(%1, %%eax, 2) \n\t"
|
MOVNTQ" %%mm2, 8(%1, %%"REG_a", 2)\n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
"cmpl %4, %%eax \n\t"
|
"cmp %4, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
|
|
||||||
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" (chromWidth)
|
::"r"(src), "r"(ydst), "r"(udst), "r"(vdst), "g" ((long)chromWidth)
|
||||||
: "memory", "%eax"
|
: "memory", "%"REG_a
|
||||||
);
|
);
|
||||||
#else
|
#else
|
||||||
unsigned i;
|
unsigned i;
|
||||||
@ -1877,16 +1877,16 @@ static inline void RENAME(planar2x)(const uint8_t *src, uint8_t *dst, int srcWid
|
|||||||
|
|
||||||
for(y=1; y<srcHeight; y++){
|
for(y=1; y<srcHeight; y++){
|
||||||
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
||||||
const int mmxSize= srcWidth&~15;
|
const long mmxSize= srcWidth&~15;
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"movl %4, %%eax \n\t"
|
"mov %4, %%"REG_a" \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
"movq (%0, %%eax), %%mm0 \n\t"
|
"movq (%0, %%"REG_a"), %%mm0 \n\t"
|
||||||
"movq (%1, %%eax), %%mm1 \n\t"
|
"movq (%1, %%"REG_a"), %%mm1 \n\t"
|
||||||
"movq 1(%0, %%eax), %%mm2 \n\t"
|
"movq 1(%0, %%"REG_a"), %%mm2 \n\t"
|
||||||
"movq 1(%1, %%eax), %%mm3 \n\t"
|
"movq 1(%1, %%"REG_a"), %%mm3 \n\t"
|
||||||
"movq -1(%0, %%eax), %%mm4 \n\t"
|
"movq -1(%0, %%"REG_a"), %%mm4 \n\t"
|
||||||
"movq -1(%1, %%eax), %%mm5 \n\t"
|
"movq -1(%1, %%"REG_a"), %%mm5 \n\t"
|
||||||
PAVGB" %%mm0, %%mm5 \n\t"
|
PAVGB" %%mm0, %%mm5 \n\t"
|
||||||
PAVGB" %%mm0, %%mm3 \n\t"
|
PAVGB" %%mm0, %%mm3 \n\t"
|
||||||
PAVGB" %%mm0, %%mm5 \n\t"
|
PAVGB" %%mm0, %%mm5 \n\t"
|
||||||
@ -1902,22 +1902,22 @@ static inline void RENAME(planar2x)(const uint8_t *src, uint8_t *dst, int srcWid
|
|||||||
"punpcklbw %%mm2, %%mm4 \n\t"
|
"punpcklbw %%mm2, %%mm4 \n\t"
|
||||||
"punpckhbw %%mm2, %%mm6 \n\t"
|
"punpckhbw %%mm2, %%mm6 \n\t"
|
||||||
#if 1
|
#if 1
|
||||||
MOVNTQ" %%mm5, (%2, %%eax, 2) \n\t"
|
MOVNTQ" %%mm5, (%2, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm7, 8(%2, %%eax, 2) \n\t"
|
MOVNTQ" %%mm7, 8(%2, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm4, (%3, %%eax, 2) \n\t"
|
MOVNTQ" %%mm4, (%3, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm6, 8(%3, %%eax, 2) \n\t"
|
MOVNTQ" %%mm6, 8(%3, %%"REG_a", 2)\n\t"
|
||||||
#else
|
#else
|
||||||
"movq %%mm5, (%2, %%eax, 2) \n\t"
|
"movq %%mm5, (%2, %%"REG_a", 2) \n\t"
|
||||||
"movq %%mm7, 8(%2, %%eax, 2) \n\t"
|
"movq %%mm7, 8(%2, %%"REG_a", 2)\n\t"
|
||||||
"movq %%mm4, (%3, %%eax, 2) \n\t"
|
"movq %%mm4, (%3, %%"REG_a", 2) \n\t"
|
||||||
"movq %%mm6, 8(%3, %%eax, 2) \n\t"
|
"movq %%mm6, 8(%3, %%"REG_a", 2)\n\t"
|
||||||
#endif
|
#endif
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
:: "r" (src + mmxSize ), "r" (src + srcStride + mmxSize ),
|
:: "r" (src + mmxSize ), "r" (src + srcStride + mmxSize ),
|
||||||
"r" (dst + mmxSize*2), "r" (dst + dstStride + mmxSize*2),
|
"r" (dst + mmxSize*2), "r" (dst + dstStride + mmxSize*2),
|
||||||
"g" (-mmxSize)
|
"g" (-mmxSize)
|
||||||
: "%eax"
|
: "%"REG_a
|
||||||
|
|
||||||
);
|
);
|
||||||
#else
|
#else
|
||||||
@ -2107,20 +2107,20 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
for(i=0; i<2; i++)
|
for(i=0; i<2; i++)
|
||||||
{
|
{
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"movl %2, %%eax \n\t"
|
"mov %2, %%"REG_a" \n\t"
|
||||||
"movq "MANGLE(bgr2YCoeff)", %%mm6 \n\t"
|
"movq "MANGLE(bgr2YCoeff)", %%mm6 \n\t"
|
||||||
"movq "MANGLE(w1111)", %%mm5 \n\t"
|
"movq "MANGLE(w1111)", %%mm5 \n\t"
|
||||||
"pxor %%mm7, %%mm7 \n\t"
|
"pxor %%mm7, %%mm7 \n\t"
|
||||||
"leal (%%eax, %%eax, 2), %%ebx \n\t"
|
"lea (%%"REG_a", %%"REG_a", 2), %%"REG_b"\n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%0, %%ebx) \n\t"
|
PREFETCH" 64(%0, %%"REG_b") \n\t"
|
||||||
"movd (%0, %%ebx), %%mm0 \n\t"
|
"movd (%0, %%"REG_b"), %%mm0 \n\t"
|
||||||
"movd 3(%0, %%ebx), %%mm1 \n\t"
|
"movd 3(%0, %%"REG_b"), %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm0 \n\t"
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"movd 6(%0, %%ebx), %%mm2 \n\t"
|
"movd 6(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 9(%0, %%ebx), %%mm3 \n\t"
|
"movd 9(%0, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm3 \n\t"
|
"punpcklbw %%mm7, %%mm3 \n\t"
|
||||||
"pmaddwd %%mm6, %%mm0 \n\t"
|
"pmaddwd %%mm6, %%mm0 \n\t"
|
||||||
@ -2140,12 +2140,12 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"packssdw %%mm2, %%mm0 \n\t"
|
"packssdw %%mm2, %%mm0 \n\t"
|
||||||
"psraw $7, %%mm0 \n\t"
|
"psraw $7, %%mm0 \n\t"
|
||||||
|
|
||||||
"movd 12(%0, %%ebx), %%mm4 \n\t"
|
"movd 12(%0, %%"REG_b"), %%mm4 \n\t"
|
||||||
"movd 15(%0, %%ebx), %%mm1 \n\t"
|
"movd 15(%0, %%"REG_b"), %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm4 \n\t"
|
"punpcklbw %%mm7, %%mm4 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"movd 18(%0, %%ebx), %%mm2 \n\t"
|
"movd 18(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 21(%0, %%ebx), %%mm3 \n\t"
|
"movd 21(%0, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm3 \n\t"
|
"punpcklbw %%mm7, %%mm3 \n\t"
|
||||||
"pmaddwd %%mm6, %%mm4 \n\t"
|
"pmaddwd %%mm6, %%mm4 \n\t"
|
||||||
@ -2162,39 +2162,39 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"packssdw %%mm3, %%mm2 \n\t"
|
"packssdw %%mm3, %%mm2 \n\t"
|
||||||
"pmaddwd %%mm5, %%mm4 \n\t"
|
"pmaddwd %%mm5, %%mm4 \n\t"
|
||||||
"pmaddwd %%mm5, %%mm2 \n\t"
|
"pmaddwd %%mm5, %%mm2 \n\t"
|
||||||
"addl $24, %%ebx \n\t"
|
"add $24, %%"REG_b" \n\t"
|
||||||
"packssdw %%mm2, %%mm4 \n\t"
|
"packssdw %%mm2, %%mm4 \n\t"
|
||||||
"psraw $7, %%mm4 \n\t"
|
"psraw $7, %%mm4 \n\t"
|
||||||
|
|
||||||
"packuswb %%mm4, %%mm0 \n\t"
|
"packuswb %%mm4, %%mm0 \n\t"
|
||||||
"paddusb "MANGLE(bgr2YOffset)", %%mm0 \n\t"
|
"paddusb "MANGLE(bgr2YOffset)", %%mm0 \n\t"
|
||||||
|
|
||||||
MOVNTQ" %%mm0, (%1, %%eax) \n\t"
|
MOVNTQ" %%mm0, (%1, %%"REG_a") \n\t"
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
: : "r" (src+width*3), "r" (ydst+width), "g" (-width)
|
: : "r" (src+width*3), "r" (ydst+width), "g" ((long)-width)
|
||||||
: "%eax", "%ebx"
|
: "%"REG_a, "%"REG_b
|
||||||
);
|
);
|
||||||
ydst += lumStride;
|
ydst += lumStride;
|
||||||
src += srcStride;
|
src += srcStride;
|
||||||
}
|
}
|
||||||
src -= srcStride*2;
|
src -= srcStride*2;
|
||||||
asm volatile(
|
asm volatile(
|
||||||
"movl %4, %%eax \n\t"
|
"mov %4, %%"REG_a" \n\t"
|
||||||
"movq "MANGLE(w1111)", %%mm5 \n\t"
|
"movq "MANGLE(w1111)", %%mm5 \n\t"
|
||||||
"movq "MANGLE(bgr2UCoeff)", %%mm6 \n\t"
|
"movq "MANGLE(bgr2UCoeff)", %%mm6 \n\t"
|
||||||
"pxor %%mm7, %%mm7 \n\t"
|
"pxor %%mm7, %%mm7 \n\t"
|
||||||
"leal (%%eax, %%eax, 2), %%ebx \n\t"
|
"lea (%%"REG_a", %%"REG_a", 2), %%"REG_b"\n\t"
|
||||||
"addl %%ebx, %%ebx \n\t"
|
"add %%"REG_b", %%"REG_b" \n\t"
|
||||||
".balign 16 \n\t"
|
".balign 16 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%0, %%ebx) \n\t"
|
PREFETCH" 64(%0, %%"REG_b") \n\t"
|
||||||
PREFETCH" 64(%1, %%ebx) \n\t"
|
PREFETCH" 64(%1, %%"REG_b") \n\t"
|
||||||
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
||||||
"movq (%0, %%ebx), %%mm0 \n\t"
|
"movq (%0, %%"REG_b"), %%mm0 \n\t"
|
||||||
"movq (%1, %%ebx), %%mm1 \n\t"
|
"movq (%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movq 6(%0, %%ebx), %%mm2 \n\t"
|
"movq 6(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movq 6(%1, %%ebx), %%mm3 \n\t"
|
"movq 6(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
PAVGB" %%mm1, %%mm0 \n\t"
|
PAVGB" %%mm1, %%mm0 \n\t"
|
||||||
PAVGB" %%mm3, %%mm2 \n\t"
|
PAVGB" %%mm3, %%mm2 \n\t"
|
||||||
"movq %%mm0, %%mm1 \n\t"
|
"movq %%mm0, %%mm1 \n\t"
|
||||||
@ -2206,10 +2206,10 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"punpcklbw %%mm7, %%mm0 \n\t"
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
#else
|
#else
|
||||||
"movd (%0, %%ebx), %%mm0 \n\t"
|
"movd (%0, %%"REG_b"), %%mm0 \n\t"
|
||||||
"movd (%1, %%ebx), %%mm1 \n\t"
|
"movd (%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movd 3(%0, %%ebx), %%mm2 \n\t"
|
"movd 3(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 3(%1, %%ebx), %%mm3 \n\t"
|
"movd 3(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm0 \n\t"
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
@ -2217,10 +2217,10 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"paddw %%mm1, %%mm0 \n\t"
|
"paddw %%mm1, %%mm0 \n\t"
|
||||||
"paddw %%mm3, %%mm2 \n\t"
|
"paddw %%mm3, %%mm2 \n\t"
|
||||||
"paddw %%mm2, %%mm0 \n\t"
|
"paddw %%mm2, %%mm0 \n\t"
|
||||||
"movd 6(%0, %%ebx), %%mm4 \n\t"
|
"movd 6(%0, %%"REG_b"), %%mm4 \n\t"
|
||||||
"movd 6(%1, %%ebx), %%mm1 \n\t"
|
"movd 6(%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movd 9(%0, %%ebx), %%mm2 \n\t"
|
"movd 9(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 9(%1, %%ebx), %%mm3 \n\t"
|
"movd 9(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm4 \n\t"
|
"punpcklbw %%mm7, %%mm4 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
@ -2252,10 +2252,10 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"psraw $7, %%mm0 \n\t"
|
"psraw $7, %%mm0 \n\t"
|
||||||
|
|
||||||
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
#if defined (HAVE_MMX2) || defined (HAVE_3DNOW)
|
||||||
"movq 12(%0, %%ebx), %%mm4 \n\t"
|
"movq 12(%0, %%"REG_b"), %%mm4 \n\t"
|
||||||
"movq 12(%1, %%ebx), %%mm1 \n\t"
|
"movq 12(%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movq 18(%0, %%ebx), %%mm2 \n\t"
|
"movq 18(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movq 18(%1, %%ebx), %%mm3 \n\t"
|
"movq 18(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
PAVGB" %%mm1, %%mm4 \n\t"
|
PAVGB" %%mm1, %%mm4 \n\t"
|
||||||
PAVGB" %%mm3, %%mm2 \n\t"
|
PAVGB" %%mm3, %%mm2 \n\t"
|
||||||
"movq %%mm4, %%mm1 \n\t"
|
"movq %%mm4, %%mm1 \n\t"
|
||||||
@ -2267,10 +2267,10 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"punpcklbw %%mm7, %%mm4 \n\t"
|
"punpcklbw %%mm7, %%mm4 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
#else
|
#else
|
||||||
"movd 12(%0, %%ebx), %%mm4 \n\t"
|
"movd 12(%0, %%"REG_b"), %%mm4 \n\t"
|
||||||
"movd 12(%1, %%ebx), %%mm1 \n\t"
|
"movd 12(%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movd 15(%0, %%ebx), %%mm2 \n\t"
|
"movd 15(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 15(%1, %%ebx), %%mm3 \n\t"
|
"movd 15(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm4 \n\t"
|
"punpcklbw %%mm7, %%mm4 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
@ -2278,10 +2278,10 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"paddw %%mm1, %%mm4 \n\t"
|
"paddw %%mm1, %%mm4 \n\t"
|
||||||
"paddw %%mm3, %%mm2 \n\t"
|
"paddw %%mm3, %%mm2 \n\t"
|
||||||
"paddw %%mm2, %%mm4 \n\t"
|
"paddw %%mm2, %%mm4 \n\t"
|
||||||
"movd 18(%0, %%ebx), %%mm5 \n\t"
|
"movd 18(%0, %%"REG_b"), %%mm5 \n\t"
|
||||||
"movd 18(%1, %%ebx), %%mm1 \n\t"
|
"movd 18(%1, %%"REG_b"), %%mm1 \n\t"
|
||||||
"movd 21(%0, %%ebx), %%mm2 \n\t"
|
"movd 21(%0, %%"REG_b"), %%mm2 \n\t"
|
||||||
"movd 21(%1, %%ebx), %%mm3 \n\t"
|
"movd 21(%1, %%"REG_b"), %%mm3 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm5 \n\t"
|
"punpcklbw %%mm7, %%mm5 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm2 \n\t"
|
"punpcklbw %%mm7, %%mm2 \n\t"
|
||||||
@ -2310,7 +2310,7 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"packssdw %%mm3, %%mm1 \n\t"
|
"packssdw %%mm3, %%mm1 \n\t"
|
||||||
"pmaddwd %%mm5, %%mm4 \n\t"
|
"pmaddwd %%mm5, %%mm4 \n\t"
|
||||||
"pmaddwd %%mm5, %%mm1 \n\t"
|
"pmaddwd %%mm5, %%mm1 \n\t"
|
||||||
"addl $24, %%ebx \n\t"
|
"add $24, %%"REG_b" \n\t"
|
||||||
"packssdw %%mm1, %%mm4 \n\t" // V3 V2 U3 U2
|
"packssdw %%mm1, %%mm4 \n\t" // V3 V2 U3 U2
|
||||||
"psraw $7, %%mm4 \n\t"
|
"psraw $7, %%mm4 \n\t"
|
||||||
|
|
||||||
@ -2319,14 +2319,13 @@ static inline void RENAME(rgb24toyv12)(const uint8_t *src, uint8_t *ydst, uint8_
|
|||||||
"punpckhdq %%mm4, %%mm1 \n\t"
|
"punpckhdq %%mm4, %%mm1 \n\t"
|
||||||
"packsswb %%mm1, %%mm0 \n\t"
|
"packsswb %%mm1, %%mm0 \n\t"
|
||||||
"paddb "MANGLE(bgr2UVOffset)", %%mm0 \n\t"
|
"paddb "MANGLE(bgr2UVOffset)", %%mm0 \n\t"
|
||||||
|
"movd %%mm0, (%2, %%"REG_a") \n\t"
|
||||||
"movd %%mm0, (%2, %%eax) \n\t"
|
|
||||||
"punpckhdq %%mm0, %%mm0 \n\t"
|
"punpckhdq %%mm0, %%mm0 \n\t"
|
||||||
"movd %%mm0, (%3, %%eax) \n\t"
|
"movd %%mm0, (%3, %%"REG_a") \n\t"
|
||||||
"addl $4, %%eax \n\t"
|
"add $4, %%"REG_a" \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
: : "r" (src+chromWidth*6), "r" (src+srcStride+chromWidth*6), "r" (udst+chromWidth), "r" (vdst+chromWidth), "g" (-chromWidth)
|
: : "r" (src+chromWidth*6), "r" (src+srcStride+chromWidth*6), "r" (udst+chromWidth), "r" (vdst+chromWidth), "g" ((long)-chromWidth)
|
||||||
: "%eax", "%ebx"
|
: "%"REG_a, "%"REG_b
|
||||||
);
|
);
|
||||||
|
|
||||||
udst += chromStride;
|
udst += chromStride;
|
||||||
@ -2403,48 +2402,48 @@ void RENAME(interleaveBytes)(uint8_t *src1, uint8_t *src2, uint8_t *dest,
|
|||||||
#ifdef HAVE_MMX
|
#ifdef HAVE_MMX
|
||||||
#ifdef HAVE_SSE2
|
#ifdef HAVE_SSE2
|
||||||
asm(
|
asm(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%1, %%eax) \n\t"
|
PREFETCH" 64(%1, %%"REG_a") \n\t"
|
||||||
PREFETCH" 64(%2, %%eax) \n\t"
|
PREFETCH" 64(%2, %%"REG_a") \n\t"
|
||||||
"movdqa (%1, %%eax), %%xmm0 \n\t"
|
"movdqa (%1, %%"REG_a"), %%xmm0 \n\t"
|
||||||
"movdqa (%1, %%eax), %%xmm1 \n\t"
|
"movdqa (%1, %%"REG_a"), %%xmm1 \n\t"
|
||||||
"movdqa (%2, %%eax), %%xmm2 \n\t"
|
"movdqa (%2, %%"REG_a"), %%xmm2 \n\t"
|
||||||
"punpcklbw %%xmm2, %%xmm0 \n\t"
|
"punpcklbw %%xmm2, %%xmm0 \n\t"
|
||||||
"punpckhbw %%xmm2, %%xmm1 \n\t"
|
"punpckhbw %%xmm2, %%xmm1 \n\t"
|
||||||
"movntdq %%xmm0, (%0, %%eax, 2) \n\t"
|
"movntdq %%xmm0, (%0, %%"REG_a", 2)\n\t"
|
||||||
"movntdq %%xmm1, 16(%0, %%eax, 2)\n\t"
|
"movntdq %%xmm1, 16(%0, %%"REG_a", 2)\n\t"
|
||||||
"addl $16, %%eax \n\t"
|
"add $16, %%"REG_a" \n\t"
|
||||||
"cmpl %3, %%eax \n\t"
|
"cmp %3, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
::"r"(dest), "r"(src1), "r"(src2), "r" (width-15)
|
::"r"(dest), "r"(src1), "r"(src2), "r" ((long)width-15)
|
||||||
: "memory", "%eax"
|
: "memory", "%"REG_a""
|
||||||
);
|
);
|
||||||
#else
|
#else
|
||||||
asm(
|
asm(
|
||||||
"xorl %%eax, %%eax \n\t"
|
"xor %%"REG_a", %%"REG_a" \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
PREFETCH" 64(%1, %%eax) \n\t"
|
PREFETCH" 64(%1, %%"REG_a") \n\t"
|
||||||
PREFETCH" 64(%2, %%eax) \n\t"
|
PREFETCH" 64(%2, %%"REG_a") \n\t"
|
||||||
"movq (%1, %%eax), %%mm0 \n\t"
|
"movq (%1, %%"REG_a"), %%mm0 \n\t"
|
||||||
"movq 8(%1, %%eax), %%mm2 \n\t"
|
"movq 8(%1, %%"REG_a"), %%mm2 \n\t"
|
||||||
"movq %%mm0, %%mm1 \n\t"
|
"movq %%mm0, %%mm1 \n\t"
|
||||||
"movq %%mm2, %%mm3 \n\t"
|
"movq %%mm2, %%mm3 \n\t"
|
||||||
"movq (%2, %%eax), %%mm4 \n\t"
|
"movq (%2, %%"REG_a"), %%mm4 \n\t"
|
||||||
"movq 8(%2, %%eax), %%mm5 \n\t"
|
"movq 8(%2, %%"REG_a"), %%mm5 \n\t"
|
||||||
"punpcklbw %%mm4, %%mm0 \n\t"
|
"punpcklbw %%mm4, %%mm0 \n\t"
|
||||||
"punpckhbw %%mm4, %%mm1 \n\t"
|
"punpckhbw %%mm4, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm5, %%mm2 \n\t"
|
"punpcklbw %%mm5, %%mm2 \n\t"
|
||||||
"punpckhbw %%mm5, %%mm3 \n\t"
|
"punpckhbw %%mm5, %%mm3 \n\t"
|
||||||
MOVNTQ" %%mm0, (%0, %%eax, 2) \n\t"
|
MOVNTQ" %%mm0, (%0, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm1, 8(%0, %%eax, 2) \n\t"
|
MOVNTQ" %%mm1, 8(%0, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm2, 16(%0, %%eax, 2) \n\t"
|
MOVNTQ" %%mm2, 16(%0, %%"REG_a", 2)\n\t"
|
||||||
MOVNTQ" %%mm3, 24(%0, %%eax, 2) \n\t"
|
MOVNTQ" %%mm3, 24(%0, %%"REG_a", 2)\n\t"
|
||||||
"addl $16, %%eax \n\t"
|
"add $16, %%"REG_a" \n\t"
|
||||||
"cmpl %3, %%eax \n\t"
|
"cmp %3, %%"REG_a" \n\t"
|
||||||
" jb 1b \n\t"
|
" jb 1b \n\t"
|
||||||
::"r"(dest), "r"(src1), "r"(src2), "r" (width-15)
|
::"r"(dest), "r"(src1), "r"(src2), "r" ((long)width-15)
|
||||||
: "memory", "%eax"
|
: "memory", "%"REG_a
|
||||||
);
|
);
|
||||||
#endif
|
#endif
|
||||||
for(w= (width&(~15)); w < width; w++)
|
for(w= (width&(~15)); w < width; w++)
|
||||||
@ -2582,7 +2581,7 @@ static inline void RENAME(yvu9_to_yuy2)(const uint8_t *src1, const uint8_t *src2
|
|||||||
int srcStride1, int srcStride2,
|
int srcStride1, int srcStride2,
|
||||||
int srcStride3, int dstStride)
|
int srcStride3, int dstStride)
|
||||||
{
|
{
|
||||||
unsigned y,x,w,h;
|
unsigned long y,x,w,h;
|
||||||
w=width/2; h=height;
|
w=width/2; h=height;
|
||||||
for(y=0;y<h;y++){
|
for(y=0;y<h;y++){
|
||||||
const uint8_t* yp=src1+srcStride1*y;
|
const uint8_t* yp=src1+srcStride1*y;
|
||||||
|
@ -104,7 +104,7 @@ static void doTest(uint8_t *ref[3], int refStride[3], int w, int h, int srcForma
|
|||||||
sws_scale(dstContext, src, srcStride, 0, srcH, dst, dstStride);
|
sws_scale(dstContext, src, srcStride, 0, srcH, dst, dstStride);
|
||||||
sws_scale(outContext, dst, dstStride, 0, dstH, out, refStride);
|
sws_scale(outContext, dst, dstStride, 0, dstH, out, refStride);
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
asm volatile ("emms\n\t");
|
asm volatile ("emms\n\t");
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
@ -199,14 +199,14 @@ int main(int argc, char **argv){
|
|||||||
rgb_data[ x + y*4*W]= random();
|
rgb_data[ x + y*4*W]= random();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
sws_rgb2rgb_init(SWS_CPU_CAPS_MMX*0);
|
sws_rgb2rgb_init(SWS_CPU_CAPS_MMX*0);
|
||||||
#else
|
#else
|
||||||
sws_rgb2rgb_init(0);
|
sws_rgb2rgb_init(0);
|
||||||
#endif
|
#endif
|
||||||
sws_scale(sws, rgb_src, rgb_stride, 0, H , src, stride);
|
sws_scale(sws, rgb_src, rgb_stride, 0, H , src, stride);
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
asm volatile ("emms\n\t");
|
asm volatile ("emms\n\t");
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
@ -145,7 +145,7 @@ write special BGR->BGR scaler
|
|||||||
#define MIN(a,b) ((a) > (b) ? (b) : (a))
|
#define MIN(a,b) ((a) > (b) ? (b) : (a))
|
||||||
#define MAX(a,b) ((a) < (b) ? (b) : (a))
|
#define MAX(a,b) ((a) < (b) ? (b) : (a))
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
static uint64_t attribute_used __attribute__((aligned(8))) bF8= 0xF8F8F8F8F8F8F8F8LL;
|
static uint64_t attribute_used __attribute__((aligned(8))) bF8= 0xF8F8F8F8F8F8F8F8LL;
|
||||||
static uint64_t attribute_used __attribute__((aligned(8))) bFC= 0xFCFCFCFCFCFCFCFCLL;
|
static uint64_t attribute_used __attribute__((aligned(8))) bFC= 0xFCFCFCFCFCFCFCFCLL;
|
||||||
static uint64_t __attribute__((aligned(8))) w10= 0x0010001000100010LL;
|
static uint64_t __attribute__((aligned(8))) w10= 0x0010001000100010LL;
|
||||||
@ -204,7 +204,7 @@ extern const uint8_t dither_8x8_32[8][8];
|
|||||||
extern const uint8_t dither_8x8_73[8][8];
|
extern const uint8_t dither_8x8_73[8][8];
|
||||||
extern const uint8_t dither_8x8_220[8][8];
|
extern const uint8_t dither_8x8_220[8][8];
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
void in_asm_used_var_warning_killer()
|
void in_asm_used_var_warning_killer()
|
||||||
{
|
{
|
||||||
volatile int i= bF8+bFC+w10+
|
volatile int i= bF8+bFC+w10+
|
||||||
@ -679,7 +679,7 @@ static inline void yuv2packedXinC(SwsContext *c, int16_t *lumFilter, int16_t **l
|
|||||||
#endif //HAVE_ALTIVEC
|
#endif //HAVE_ALTIVEC
|
||||||
#endif //ARCH_POWERPC
|
#endif //ARCH_POWERPC
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
|
|
||||||
#if (defined (HAVE_MMX) && !defined (HAVE_3DNOW) && !defined (HAVE_MMX2)) || defined (RUNTIME_CPUDETECT)
|
#if (defined (HAVE_MMX) && !defined (HAVE_3DNOW) && !defined (HAVE_MMX2)) || defined (RUNTIME_CPUDETECT)
|
||||||
#define COMPILE_MMX
|
#define COMPILE_MMX
|
||||||
@ -692,7 +692,7 @@ static inline void yuv2packedXinC(SwsContext *c, int16_t *lumFilter, int16_t **l
|
|||||||
#if (defined (HAVE_3DNOW) && !defined (HAVE_MMX2)) || defined (RUNTIME_CPUDETECT)
|
#if (defined (HAVE_3DNOW) && !defined (HAVE_MMX2)) || defined (RUNTIME_CPUDETECT)
|
||||||
#define COMPILE_3DNOW
|
#define COMPILE_3DNOW
|
||||||
#endif
|
#endif
|
||||||
#endif //ARCH_X86
|
#endif //ARCH_X86 || ARCH_X86_64
|
||||||
|
|
||||||
#undef HAVE_MMX
|
#undef HAVE_MMX
|
||||||
#undef HAVE_MMX2
|
#undef HAVE_MMX2
|
||||||
@ -716,7 +716,7 @@ static inline void yuv2packedXinC(SwsContext *c, int16_t *lumFilter, int16_t **l
|
|||||||
#endif
|
#endif
|
||||||
#endif //ARCH_POWERPC
|
#endif //ARCH_POWERPC
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
|
|
||||||
//X86 versions
|
//X86 versions
|
||||||
/*
|
/*
|
||||||
@ -758,7 +758,7 @@ static inline void yuv2packedXinC(SwsContext *c, int16_t *lumFilter, int16_t **l
|
|||||||
#include "swscale_template.c"
|
#include "swscale_template.c"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#endif //ARCH_X86
|
#endif //ARCH_X86 || ARCH_X86_64
|
||||||
|
|
||||||
// minor note: the HAVE_xyz is messed up after that line so don't use it
|
// minor note: the HAVE_xyz is messed up after that line so don't use it
|
||||||
|
|
||||||
@ -783,7 +783,7 @@ static inline void initFilter(int16_t **outFilter, int16_t **filterPos, int *out
|
|||||||
int minFilterSize;
|
int minFilterSize;
|
||||||
double *filter=NULL;
|
double *filter=NULL;
|
||||||
double *filter2=NULL;
|
double *filter2=NULL;
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
if(flags & SWS_CPU_CAPS_MMX)
|
if(flags & SWS_CPU_CAPS_MMX)
|
||||||
asm volatile("emms\n\t"::: "memory"); //FIXME this shouldnt be required but it IS (even for non mmx versions)
|
asm volatile("emms\n\t"::: "memory"); //FIXME this shouldnt be required but it IS (even for non mmx versions)
|
||||||
#endif
|
#endif
|
||||||
@ -1142,17 +1142,17 @@ static inline void initFilter(int16_t **outFilter, int16_t **filterPos, int *out
|
|||||||
free(filter);
|
free(filter);
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *filter, int32_t *filterPos, int numSplits)
|
static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *filter, int32_t *filterPos, int numSplits)
|
||||||
{
|
{
|
||||||
uint8_t *fragmentA;
|
uint8_t *fragmentA;
|
||||||
int imm8OfPShufW1A;
|
long imm8OfPShufW1A;
|
||||||
int imm8OfPShufW2A;
|
long imm8OfPShufW2A;
|
||||||
int fragmentLengthA;
|
long fragmentLengthA;
|
||||||
uint8_t *fragmentB;
|
uint8_t *fragmentB;
|
||||||
int imm8OfPShufW1B;
|
long imm8OfPShufW1B;
|
||||||
int imm8OfPShufW2B;
|
long imm8OfPShufW2B;
|
||||||
int fragmentLengthB;
|
long fragmentLengthB;
|
||||||
int fragmentPos;
|
int fragmentPos;
|
||||||
|
|
||||||
int xpos, i;
|
int xpos, i;
|
||||||
@ -1165,9 +1165,9 @@ static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *fil
|
|||||||
"jmp 9f \n\t"
|
"jmp 9f \n\t"
|
||||||
// Begin
|
// Begin
|
||||||
"0: \n\t"
|
"0: \n\t"
|
||||||
"movq (%%edx, %%eax), %%mm3 \n\t"
|
"movq (%%"REG_d", %%"REG_a"), %%mm3\n\t"
|
||||||
"movd (%%ecx, %%esi), %%mm0 \n\t"
|
"movd (%%"REG_c", %%"REG_S"), %%mm0\n\t"
|
||||||
"movd 1(%%ecx, %%esi), %%mm1 \n\t"
|
"movd 1(%%"REG_c", %%"REG_S"), %%mm1\n\t"
|
||||||
"punpcklbw %%mm7, %%mm1 \n\t"
|
"punpcklbw %%mm7, %%mm1 \n\t"
|
||||||
"punpcklbw %%mm7, %%mm0 \n\t"
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
||||||
"pshufw $0xFF, %%mm1, %%mm1 \n\t"
|
"pshufw $0xFF, %%mm1, %%mm1 \n\t"
|
||||||
@ -1175,26 +1175,26 @@ static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *fil
|
|||||||
"pshufw $0xFF, %%mm0, %%mm0 \n\t"
|
"pshufw $0xFF, %%mm0, %%mm0 \n\t"
|
||||||
"2: \n\t"
|
"2: \n\t"
|
||||||
"psubw %%mm1, %%mm0 \n\t"
|
"psubw %%mm1, %%mm0 \n\t"
|
||||||
"movl 8(%%ebx, %%eax), %%esi \n\t"
|
"mov 8(%%"REG_b", %%"REG_a"), %%"REG_S"\n\t"
|
||||||
"pmullw %%mm3, %%mm0 \n\t"
|
"pmullw %%mm3, %%mm0 \n\t"
|
||||||
"psllw $7, %%mm1 \n\t"
|
"psllw $7, %%mm1 \n\t"
|
||||||
"paddw %%mm1, %%mm0 \n\t"
|
"paddw %%mm1, %%mm0 \n\t"
|
||||||
|
|
||||||
"movq %%mm0, (%%edi, %%eax) \n\t"
|
"movq %%mm0, (%%"REG_D", %%"REG_a")\n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
// End
|
// End
|
||||||
"9: \n\t"
|
"9: \n\t"
|
||||||
// "int $3\n\t"
|
// "int $3\n\t"
|
||||||
"leal 0b, %0 \n\t"
|
"lea 0b, %0 \n\t"
|
||||||
"leal 1b, %1 \n\t"
|
"lea 1b, %1 \n\t"
|
||||||
"leal 2b, %2 \n\t"
|
"lea 2b, %2 \n\t"
|
||||||
"decl %1 \n\t"
|
"dec %1 \n\t"
|
||||||
"decl %2 \n\t"
|
"dec %2 \n\t"
|
||||||
"subl %0, %1 \n\t"
|
"sub %0, %1 \n\t"
|
||||||
"subl %0, %2 \n\t"
|
"sub %0, %2 \n\t"
|
||||||
"leal 9b, %3 \n\t"
|
"lea 9b, %3 \n\t"
|
||||||
"subl %0, %3 \n\t"
|
"sub %0, %3 \n\t"
|
||||||
|
|
||||||
|
|
||||||
:"=r" (fragmentA), "=r" (imm8OfPShufW1A), "=r" (imm8OfPShufW2A),
|
:"=r" (fragmentA), "=r" (imm8OfPShufW1A), "=r" (imm8OfPShufW2A),
|
||||||
@ -1205,34 +1205,34 @@ static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *fil
|
|||||||
"jmp 9f \n\t"
|
"jmp 9f \n\t"
|
||||||
// Begin
|
// Begin
|
||||||
"0: \n\t"
|
"0: \n\t"
|
||||||
"movq (%%edx, %%eax), %%mm3 \n\t"
|
"movq (%%"REG_d", %%"REG_a"), %%mm3\n\t"
|
||||||
"movd (%%ecx, %%esi), %%mm0 \n\t"
|
"movd (%%"REG_c", %%"REG_S"), %%mm0\n\t"
|
||||||
"punpcklbw %%mm7, %%mm0 \n\t"
|
"punpcklbw %%mm7, %%mm0 \n\t"
|
||||||
"pshufw $0xFF, %%mm0, %%mm1 \n\t"
|
"pshufw $0xFF, %%mm0, %%mm1 \n\t"
|
||||||
"1: \n\t"
|
"1: \n\t"
|
||||||
"pshufw $0xFF, %%mm0, %%mm0 \n\t"
|
"pshufw $0xFF, %%mm0, %%mm0 \n\t"
|
||||||
"2: \n\t"
|
"2: \n\t"
|
||||||
"psubw %%mm1, %%mm0 \n\t"
|
"psubw %%mm1, %%mm0 \n\t"
|
||||||
"movl 8(%%ebx, %%eax), %%esi \n\t"
|
"mov 8(%%"REG_b", %%"REG_a"), %%"REG_S"\n\t"
|
||||||
"pmullw %%mm3, %%mm0 \n\t"
|
"pmullw %%mm3, %%mm0 \n\t"
|
||||||
"psllw $7, %%mm1 \n\t"
|
"psllw $7, %%mm1 \n\t"
|
||||||
"paddw %%mm1, %%mm0 \n\t"
|
"paddw %%mm1, %%mm0 \n\t"
|
||||||
|
|
||||||
"movq %%mm0, (%%edi, %%eax) \n\t"
|
"movq %%mm0, (%%"REG_D", %%"REG_a")\n\t"
|
||||||
|
|
||||||
"addl $8, %%eax \n\t"
|
"add $8, %%"REG_a" \n\t"
|
||||||
// End
|
// End
|
||||||
"9: \n\t"
|
"9: \n\t"
|
||||||
// "int $3\n\t"
|
// "int $3\n\t"
|
||||||
"leal 0b, %0 \n\t"
|
"lea 0b, %0 \n\t"
|
||||||
"leal 1b, %1 \n\t"
|
"lea 1b, %1 \n\t"
|
||||||
"leal 2b, %2 \n\t"
|
"lea 2b, %2 \n\t"
|
||||||
"decl %1 \n\t"
|
"dec %1 \n\t"
|
||||||
"decl %2 \n\t"
|
"dec %2 \n\t"
|
||||||
"subl %0, %1 \n\t"
|
"sub %0, %1 \n\t"
|
||||||
"subl %0, %2 \n\t"
|
"sub %0, %2 \n\t"
|
||||||
"leal 9b, %3 \n\t"
|
"lea 9b, %3 \n\t"
|
||||||
"subl %0, %3 \n\t"
|
"sub %0, %3 \n\t"
|
||||||
|
|
||||||
|
|
||||||
:"=r" (fragmentB), "=r" (imm8OfPShufW1B), "=r" (imm8OfPShufW2B),
|
:"=r" (fragmentB), "=r" (imm8OfPShufW1B), "=r" (imm8OfPShufW2B),
|
||||||
@ -1313,7 +1313,7 @@ static void initMMX2HScaler(int dstW, int xInc, uint8_t *funnyCode, int16_t *fil
|
|||||||
}
|
}
|
||||||
filterPos[i/2]= xpos>>16; // needed to jump to the next part
|
filterPos[i/2]= xpos>>16; // needed to jump to the next part
|
||||||
}
|
}
|
||||||
#endif // ARCH_X86
|
#endif // ARCH_X86 || ARCH_X86_64
|
||||||
|
|
||||||
static void globalInit(){
|
static void globalInit(){
|
||||||
// generating tables:
|
// generating tables:
|
||||||
@ -1327,7 +1327,7 @@ static void globalInit(){
|
|||||||
static SwsFunc getSwsFunc(int flags){
|
static SwsFunc getSwsFunc(int flags){
|
||||||
|
|
||||||
#ifdef RUNTIME_CPUDETECT
|
#ifdef RUNTIME_CPUDETECT
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
// ordered per speed fasterst first
|
// ordered per speed fasterst first
|
||||||
if(flags & SWS_CPU_CAPS_MMX2)
|
if(flags & SWS_CPU_CAPS_MMX2)
|
||||||
return swScale_MMX2;
|
return swScale_MMX2;
|
||||||
@ -1755,7 +1755,7 @@ SwsContext *sws_getContext(int srcW, int srcH, int origSrcFormat, int dstW, int
|
|||||||
int unscaled, needsDither;
|
int unscaled, needsDither;
|
||||||
int srcFormat, dstFormat;
|
int srcFormat, dstFormat;
|
||||||
SwsFilter dummyFilter= {NULL, NULL, NULL, NULL};
|
SwsFilter dummyFilter= {NULL, NULL, NULL, NULL};
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
if(flags & SWS_CPU_CAPS_MMX)
|
if(flags & SWS_CPU_CAPS_MMX)
|
||||||
asm volatile("emms\n\t"::: "memory");
|
asm volatile("emms\n\t"::: "memory");
|
||||||
#endif
|
#endif
|
||||||
@ -1995,7 +1995,7 @@ SwsContext *sws_getContext(int srcW, int srcH, int origSrcFormat, int dstW, int
|
|||||||
(flags&SWS_BICUBLIN) ? (flags|SWS_BILINEAR) : flags,
|
(flags&SWS_BICUBLIN) ? (flags|SWS_BILINEAR) : flags,
|
||||||
srcFilter->chrH, dstFilter->chrH, c->param);
|
srcFilter->chrH, dstFilter->chrH, c->param);
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
// can't downscale !!!
|
// can't downscale !!!
|
||||||
if(c->canMMX2BeUsed && (flags & SWS_FAST_BILINEAR))
|
if(c->canMMX2BeUsed && (flags & SWS_FAST_BILINEAR))
|
||||||
{
|
{
|
||||||
@ -2136,7 +2136,7 @@ SwsContext *sws_getContext(int srcW, int srcH, int origSrcFormat, int dstW, int
|
|||||||
}
|
}
|
||||||
else
|
else
|
||||||
{
|
{
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
MSG_V("SwScaler: using X86-Asm scaler for horizontal scaling\n");
|
MSG_V("SwScaler: using X86-Asm scaler for horizontal scaling\n");
|
||||||
#else
|
#else
|
||||||
if(flags & SWS_FAST_BILINEAR)
|
if(flags & SWS_FAST_BILINEAR)
|
||||||
|
File diff suppressed because it is too large
Load Diff
@ -156,7 +156,7 @@ const uint8_t __attribute__((aligned(8))) dither_8x8_220[8][8]={
|
|||||||
};
|
};
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
|
|
||||||
/* hope these constant values are cache line aligned */
|
/* hope these constant values are cache line aligned */
|
||||||
uint64_t attribute_used __attribute__((aligned(8))) mmx_00ffw = 0x00ff00ff00ff00ffULL;
|
uint64_t attribute_used __attribute__((aligned(8))) mmx_00ffw = 0x00ff00ff00ff00ffULL;
|
||||||
@ -183,14 +183,12 @@ uint64_t __attribute__((aligned(8))) dither8[2]={
|
|||||||
0x0004000400040004LL,};
|
0x0004000400040004LL,};
|
||||||
|
|
||||||
#undef HAVE_MMX
|
#undef HAVE_MMX
|
||||||
#undef ARCH_X86
|
|
||||||
|
|
||||||
//MMX versions
|
//MMX versions
|
||||||
#undef RENAME
|
#undef RENAME
|
||||||
#define HAVE_MMX
|
#define HAVE_MMX
|
||||||
#undef HAVE_MMX2
|
#undef HAVE_MMX2
|
||||||
#undef HAVE_3DNOW
|
#undef HAVE_3DNOW
|
||||||
#define ARCH_X86
|
|
||||||
#define RENAME(a) a ## _MMX
|
#define RENAME(a) a ## _MMX
|
||||||
#include "yuv2rgb_template.c"
|
#include "yuv2rgb_template.c"
|
||||||
|
|
||||||
@ -199,7 +197,6 @@ uint64_t __attribute__((aligned(8))) dither8[2]={
|
|||||||
#define HAVE_MMX
|
#define HAVE_MMX
|
||||||
#define HAVE_MMX2
|
#define HAVE_MMX2
|
||||||
#undef HAVE_3DNOW
|
#undef HAVE_3DNOW
|
||||||
#define ARCH_X86
|
|
||||||
#define RENAME(a) a ## _MMX2
|
#define RENAME(a) a ## _MMX2
|
||||||
#include "yuv2rgb_template.c"
|
#include "yuv2rgb_template.c"
|
||||||
|
|
||||||
@ -583,7 +580,7 @@ EPILOG(1)
|
|||||||
|
|
||||||
SwsFunc yuv2rgb_get_func_ptr (SwsContext *c)
|
SwsFunc yuv2rgb_get_func_ptr (SwsContext *c)
|
||||||
{
|
{
|
||||||
#ifdef ARCH_X86
|
#if defined(ARCH_X86) || defined(ARCH_X86_64)
|
||||||
if(c->flags & SWS_CPU_CAPS_MMX2){
|
if(c->flags & SWS_CPU_CAPS_MMX2){
|
||||||
switch(c->dstFormat){
|
switch(c->dstFormat){
|
||||||
case IMGFMT_BGR32: return yuv420_rgb32_MMX2;
|
case IMGFMT_BGR32: return yuv420_rgb32_MMX2;
|
||||||
|
@ -143,7 +143,7 @@ static inline int RENAME(yuv420_rgb16)(SwsContext *c, uint8_t* src[], int srcStr
|
|||||||
uint8_t *_py = src[0] + y*srcStride[0];
|
uint8_t *_py = src[0] + y*srcStride[0];
|
||||||
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
||||||
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
||||||
int index= -h_size/2;
|
long index= -h_size/2;
|
||||||
|
|
||||||
b5Dither= dither8[y&1];
|
b5Dither= dither8[y&1];
|
||||||
g6Dither= dither4[y&1];
|
g6Dither= dither4[y&1];
|
||||||
@ -204,8 +204,8 @@ YUV2RGB
|
|||||||
|
|
||||||
MOVNTQ " %%mm5, 8 (%1);" /* store pixel 4-7 */
|
MOVNTQ " %%mm5, 8 (%1);" /* store pixel 4-7 */
|
||||||
|
|
||||||
"addl $16, %1 \n\t"
|
"add $16, %1 \n\t"
|
||||||
"addl $4, %0 \n\t"
|
"add $4, %0 \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
|
|
||||||
: "+r" (index), "+r" (_image)
|
: "+r" (index), "+r" (_image)
|
||||||
@ -238,7 +238,7 @@ static inline int RENAME(yuv420_rgb15)(SwsContext *c, uint8_t* src[], int srcStr
|
|||||||
uint8_t *_py = src[0] + y*srcStride[0];
|
uint8_t *_py = src[0] + y*srcStride[0];
|
||||||
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
||||||
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
||||||
int index= -h_size/2;
|
long index= -h_size/2;
|
||||||
|
|
||||||
b5Dither= dither8[y&1];
|
b5Dither= dither8[y&1];
|
||||||
g6Dither= dither4[y&1];
|
g6Dither= dither4[y&1];
|
||||||
@ -295,8 +295,8 @@ YUV2RGB
|
|||||||
|
|
||||||
MOVNTQ " %%mm5, 8 (%1);" /* store pixel 4-7 */
|
MOVNTQ " %%mm5, 8 (%1);" /* store pixel 4-7 */
|
||||||
|
|
||||||
"addl $16, %1 \n\t"
|
"add $16, %1 \n\t"
|
||||||
"addl $4, %0 \n\t"
|
"add $4, %0 \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
: "+r" (index), "+r" (_image)
|
: "+r" (index), "+r" (_image)
|
||||||
: "r" (_pu - index), "r" (_pv - index), "r"(&c->redDither), "r" (_py - 2*index)
|
: "r" (_pu - index), "r" (_pv - index), "r"(&c->redDither), "r" (_py - 2*index)
|
||||||
@ -326,7 +326,7 @@ static inline int RENAME(yuv420_rgb24)(SwsContext *c, uint8_t* src[], int srcStr
|
|||||||
uint8_t *_py = src[0] + y*srcStride[0];
|
uint8_t *_py = src[0] + y*srcStride[0];
|
||||||
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
||||||
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
||||||
int index= -h_size/2;
|
long index= -h_size/2;
|
||||||
|
|
||||||
/* this mmx assembly code deals with SINGLE scan line at a time, it convert 8
|
/* this mmx assembly code deals with SINGLE scan line at a time, it convert 8
|
||||||
pixels in each iteration */
|
pixels in each iteration */
|
||||||
@ -440,8 +440,8 @@ YUV2RGB
|
|||||||
"pxor %%mm4, %%mm4 \n\t"
|
"pxor %%mm4, %%mm4 \n\t"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
"addl $24, %1 \n\t"
|
"add $24, %1 \n\t"
|
||||||
"addl $4, %0 \n\t"
|
"add $4, %0 \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
|
|
||||||
: "+r" (index), "+r" (_image)
|
: "+r" (index), "+r" (_image)
|
||||||
@ -472,7 +472,7 @@ static inline int RENAME(yuv420_rgb32)(SwsContext *c, uint8_t* src[], int srcStr
|
|||||||
uint8_t *_py = src[0] + y*srcStride[0];
|
uint8_t *_py = src[0] + y*srcStride[0];
|
||||||
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
uint8_t *_pu = src[1] + (y>>1)*srcStride[1];
|
||||||
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
uint8_t *_pv = src[2] + (y>>1)*srcStride[2];
|
||||||
int index= -h_size/2;
|
long index= -h_size/2;
|
||||||
|
|
||||||
/* this mmx assembly code deals with SINGLE scan line at a time, it convert 8
|
/* this mmx assembly code deals with SINGLE scan line at a time, it convert 8
|
||||||
pixels in each iteration */
|
pixels in each iteration */
|
||||||
@ -526,8 +526,8 @@ YUV2RGB
|
|||||||
"pxor %%mm4, %%mm4;" /* zero mm4 */
|
"pxor %%mm4, %%mm4;" /* zero mm4 */
|
||||||
"movq 8 (%5, %0, 2), %%mm6;" /* Load 8 Y Y7 Y6 Y5 Y4 Y3 Y2 Y1 Y0 */
|
"movq 8 (%5, %0, 2), %%mm6;" /* Load 8 Y Y7 Y6 Y5 Y4 Y3 Y2 Y1 Y0 */
|
||||||
|
|
||||||
"addl $32, %1 \n\t"
|
"add $32, %1 \n\t"
|
||||||
"addl $4, %0 \n\t"
|
"add $4, %0 \n\t"
|
||||||
" js 1b \n\t"
|
" js 1b \n\t"
|
||||||
|
|
||||||
: "+r" (index), "+r" (_image)
|
: "+r" (index), "+r" (_image)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user