You've already forked FFmpeg
							
							
				mirror of
				https://github.com/FFmpeg/FFmpeg.git
				synced 2025-10-30 23:18:11 +02:00 
			
		
		
		
	x86/qpel: move fullpel and l2 functions to separate file.
This way, they can be shared between mpeg4qpel and h264qpel without requiring either one to be compiled unconditionally. Signed-off-by: Michael Niedermayer <michaelni@gmx.at>
This commit is contained in:
		
				
					committed by
					
						 Michael Niedermayer
						Michael Niedermayer
					
				
			
			
				
	
			
			
			
						parent
						
							9d40f7106d
						
					
				
				
					commit
					22cc8a103c
				
			| @@ -65,7 +65,8 @@ YASM-OBJS-$(CONFIG_H264DSP)            += x86/h264_deblock.o            \ | ||||
| YASM-OBJS-$(CONFIG_H264PRED)           += x86/h264_intrapred.o          \ | ||||
|                                           x86/h264_intrapred_10bit.o | ||||
| YASM-OBJS-$(CONFIG_H264QPEL)           += x86/h264_qpel_8bit.o          \ | ||||
|                                           x86/h264_qpel_10bit.o | ||||
|                                           x86/h264_qpel_10bit.o         \ | ||||
|                                           x86/qpelbase.o | ||||
| YASM-OBJS-$(CONFIG_MPEGAUDIODSP)       += x86/imdct36.o | ||||
| YASM-OBJS-$(CONFIG_PNG_DECODER)        += x86/pngdsp.o | ||||
| YASM-OBJS-$(CONFIG_PRORES_DECODER)     += x86/proresdsp.o | ||||
| @@ -86,3 +87,4 @@ YASM-OBJS                              += x86/dsputil.o                 \ | ||||
|                                           x86/fmtconvert.o              \ | ||||
|                                           x86/hpeldsp.o                 \ | ||||
|                                           x86/mpeg4qpel.o               \ | ||||
|                                           x86/qpelbase.o | ||||
|   | ||||
| @@ -650,195 +650,3 @@ BSWAP32_BUF | ||||
|  | ||||
| INIT_XMM ssse3 | ||||
| BSWAP32_BUF | ||||
|  | ||||
|  | ||||
| ; FIXME: All of the code below should be put back in h264_qpel_8bit.asm. | ||||
| ; Unfortunately it is unconditionally used from dsputil_mmx.c since 71155d7 .. | ||||
|  | ||||
| %macro op_avgh 3 | ||||
|     movh   %3, %2 | ||||
|     pavgb  %1, %3 | ||||
|     movh   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_avg 2 | ||||
|     pavgb  %1, %2 | ||||
|     mova   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_puth 2-3 | ||||
|     movh   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_put 2 | ||||
|     mova   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| ; void pixels4_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS4_L2 1 | ||||
| %define OP op_%1h | ||||
| cglobal %1_pixels4_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     movd         m0, [r1] | ||||
|     movd         m1, [r2] | ||||
|     add          r1, r4 | ||||
|     add          r2, 4 | ||||
|     pavgb        m0, m1 | ||||
|     OP           m0, [r0], m3 | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+4] | ||||
|     OP           m0, [r0], m3 | ||||
|     OP           m1, [r0+r3], m3 | ||||
|     lea          r0, [r0+2*r3] | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2+8] | ||||
|     pavgb        m1, [r2+12] | ||||
|     OP           m0, [r0], m3 | ||||
|     OP           m1, [r0+r3], m3 | ||||
|     lea          r0, [r0+2*r3] | ||||
|     add          r2, 16 | ||||
|     sub         r5d, 4 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS4_L2 put | ||||
| PIXELS4_L2 avg | ||||
|  | ||||
| ; void pixels8_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS8_L2 1 | ||||
| %define OP op_%1 | ||||
| cglobal %1_pixels8_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r2] | ||||
|     add          r1, r4 | ||||
|     add          r2, 8 | ||||
|     pavgb        m0, m1 | ||||
|     OP           m0, [r0] | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+r3] | ||||
|     lea          r0, [r0+2*r3] | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2+16] | ||||
|     pavgb        m1, [r2+24] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+r3] | ||||
|     lea          r0, [r0+2*r3] | ||||
|     add          r2, 32 | ||||
|     sub         r5d, 4 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS8_L2 put | ||||
| PIXELS8_L2 avg | ||||
|  | ||||
| ; void pixels16_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS16_L2 1 | ||||
| %define OP op_%1 | ||||
| cglobal %1_pixels16_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     add          r1, r4 | ||||
|     add          r2, 16 | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     add          r1, r4 | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     add          r1, r4 | ||||
|     pavgb        m0, [r2+16] | ||||
|     pavgb        m1, [r2+24] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     add          r2, 32 | ||||
|     sub         r5d, 2 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS16_L2 put | ||||
| PIXELS16_L2 avg | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| ; void pixels(uint8_t *block, const uint8_t *pixels, int line_size, int h) | ||||
| %macro PIXELS48 2 | ||||
| %if %2 == 4 | ||||
| %define OP movh | ||||
| %else | ||||
| %define OP mova | ||||
| %endif | ||||
| cglobal %1_pixels%2, 4,5 | ||||
|     movsxdifnidn r2, r2d | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     OP           m0, [r1] | ||||
|     OP           m1, [r1+r2] | ||||
|     OP           m2, [r1+r2*2] | ||||
|     OP           m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
| %ifidn %1, avg | ||||
|     pavgb        m0, [r0] | ||||
|     pavgb        m1, [r0+r2] | ||||
|     pavgb        m2, [r0+r2*2] | ||||
|     pavgb        m3, [r0+r4] | ||||
| %endif | ||||
|     OP         [r0], m0 | ||||
|     OP      [r0+r2], m1 | ||||
|     OP    [r0+r2*2], m2 | ||||
|     OP      [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jne       .loop | ||||
|     RET | ||||
| %endmacro | ||||
|  | ||||
| PIXELS48 put, 4 | ||||
| PIXELS48 avg, 4 | ||||
| PIXELS48 put, 8 | ||||
| PIXELS48 avg, 8 | ||||
|   | ||||
| @@ -459,44 +459,3 @@ INIT_MMX mmxext | ||||
| AVG_PIXELS8_XY2 | ||||
| INIT_MMX 3dnow | ||||
| AVG_PIXELS8_XY2 | ||||
|  | ||||
| INIT_XMM sse2 | ||||
| ; void put_pixels16_sse2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h) | ||||
| cglobal put_pixels16, 4,5,4 | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     movu         m0, [r1] | ||||
|     movu         m1, [r1+r2] | ||||
|     movu         m2, [r1+r2*2] | ||||
|     movu         m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
|     mova       [r0], m0 | ||||
|     mova    [r0+r2], m1 | ||||
|     mova  [r0+r2*2], m2 | ||||
|     mova    [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jnz       .loop | ||||
|     REP_RET | ||||
|  | ||||
| ; void avg_pixels16_sse2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h) | ||||
| cglobal avg_pixels16, 4,5,4 | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     movu         m0, [r1] | ||||
|     movu         m1, [r1+r2] | ||||
|     movu         m2, [r1+r2*2] | ||||
|     movu         m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
|     pavgb        m0, [r0] | ||||
|     pavgb        m1, [r0+r2] | ||||
|     pavgb        m2, [r0+r2*2] | ||||
|     pavgb        m3, [r0+r4] | ||||
|     mova       [r0], m0 | ||||
|     mova    [r0+r2], m1 | ||||
|     mova  [r0+r2*2], m2 | ||||
|     mova    [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jnz       .loop | ||||
|     REP_RET | ||||
|   | ||||
							
								
								
									
										256
									
								
								libavcodec/x86/qpelbase.asm
									
									
									
									
									
										Normal file
									
								
							
							
						
						
									
										256
									
								
								libavcodec/x86/qpelbase.asm
									
									
									
									
									
										Normal file
									
								
							| @@ -0,0 +1,256 @@ | ||||
| ;****************************************************************************** | ||||
| ;* MMX optimized DSP utils | ||||
| ;* Copyright (c) 2008 Loren Merritt | ||||
| ;* Copyright (c) 2003-2013 Michael Niedermayer | ||||
| ;* Copyright (c) 2013 Daniel Kang | ||||
| ;* | ||||
| ;* This file is part of FFmpeg. | ||||
| ;* | ||||
| ;* FFmpeg is free software; you can redistribute it and/or | ||||
| ;* modify it under the terms of the GNU Lesser General Public | ||||
| ;* License as published by the Free Software Foundation; either | ||||
| ;* version 2.1 of the License, or (at your option) any later version. | ||||
| ;* | ||||
| ;* FFmpeg is distributed in the hope that it will be useful, | ||||
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU | ||||
| ;* Lesser General Public License for more details. | ||||
| ;* | ||||
| ;* You should have received a copy of the GNU Lesser General Public | ||||
| ;* License along with FFmpeg; if not, write to the Free Software | ||||
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | ||||
| ;****************************************************************************** | ||||
|  | ||||
| %include "libavutil/x86/x86util.asm" | ||||
|  | ||||
| SECTION .text | ||||
|  | ||||
| %macro op_avgh 3 | ||||
|     movh   %3, %2 | ||||
|     pavgb  %1, %3 | ||||
|     movh   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_avg 2 | ||||
|     pavgb  %1, %2 | ||||
|     mova   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_puth 2-3 | ||||
|     movh   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| %macro op_put 2 | ||||
|     mova   %2, %1 | ||||
| %endmacro | ||||
|  | ||||
| ; void pixels4_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS4_L2 1 | ||||
| %define OP op_%1h | ||||
| cglobal %1_pixels4_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     movd         m0, [r1] | ||||
|     movd         m1, [r2] | ||||
|     add          r1, r4 | ||||
|     add          r2, 4 | ||||
|     pavgb        m0, m1 | ||||
|     OP           m0, [r0], m3 | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+4] | ||||
|     OP           m0, [r0], m3 | ||||
|     OP           m1, [r0+r3], m3 | ||||
|     lea          r0, [r0+2*r3] | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2+8] | ||||
|     pavgb        m1, [r2+12] | ||||
|     OP           m0, [r0], m3 | ||||
|     OP           m1, [r0+r3], m3 | ||||
|     lea          r0, [r0+2*r3] | ||||
|     add          r2, 16 | ||||
|     sub         r5d, 4 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS4_L2 put | ||||
| PIXELS4_L2 avg | ||||
|  | ||||
| ; void pixels8_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS8_L2 1 | ||||
| %define OP op_%1 | ||||
| cglobal %1_pixels8_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r2] | ||||
|     add          r1, r4 | ||||
|     add          r2, 8 | ||||
|     pavgb        m0, m1 | ||||
|     OP           m0, [r0] | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+r3] | ||||
|     lea          r0, [r0+2*r3] | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+r4] | ||||
|     lea          r1, [r1+2*r4] | ||||
|     pavgb        m0, [r2+16] | ||||
|     pavgb        m1, [r2+24] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+r3] | ||||
|     lea          r0, [r0+2*r3] | ||||
|     add          r2, 32 | ||||
|     sub         r5d, 4 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS8_L2 put | ||||
| PIXELS8_L2 avg | ||||
|  | ||||
| ; void pixels16_l2_mmxext(uint8_t *dst, uint8_t *src1, uint8_t *src2, int dstStride, int src1Stride, int h) | ||||
| %macro PIXELS16_L2 1 | ||||
| %define OP op_%1 | ||||
| cglobal %1_pixels16_l2, 6,6 | ||||
|     movsxdifnidn r3, r3d | ||||
|     movsxdifnidn r4, r4d | ||||
|     test        r5d, 1 | ||||
|     je        .loop | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     add          r1, r4 | ||||
|     add          r2, 16 | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     dec         r5d | ||||
| .loop: | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     add          r1, r4 | ||||
|     pavgb        m0, [r2] | ||||
|     pavgb        m1, [r2+8] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     mova         m0, [r1] | ||||
|     mova         m1, [r1+8] | ||||
|     add          r1, r4 | ||||
|     pavgb        m0, [r2+16] | ||||
|     pavgb        m1, [r2+24] | ||||
|     OP           m0, [r0] | ||||
|     OP           m1, [r0+8] | ||||
|     add          r0, r3 | ||||
|     add          r2, 32 | ||||
|     sub         r5d, 2 | ||||
|     jne       .loop | ||||
|     REP_RET | ||||
| %endmacro | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| PIXELS16_L2 put | ||||
| PIXELS16_L2 avg | ||||
|  | ||||
| INIT_MMX mmxext | ||||
| ; void pixels(uint8_t *block, const uint8_t *pixels, int line_size, int h) | ||||
| %macro PIXELS48 2 | ||||
| %if %2 == 4 | ||||
| %define OP movh | ||||
| %else | ||||
| %define OP mova | ||||
| %endif | ||||
| cglobal %1_pixels%2, 4,5 | ||||
|     movsxdifnidn r2, r2d | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     OP           m0, [r1] | ||||
|     OP           m1, [r1+r2] | ||||
|     OP           m2, [r1+r2*2] | ||||
|     OP           m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
| %ifidn %1, avg | ||||
|     pavgb        m0, [r0] | ||||
|     pavgb        m1, [r0+r2] | ||||
|     pavgb        m2, [r0+r2*2] | ||||
|     pavgb        m3, [r0+r4] | ||||
| %endif | ||||
|     OP         [r0], m0 | ||||
|     OP      [r0+r2], m1 | ||||
|     OP    [r0+r2*2], m2 | ||||
|     OP      [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jne       .loop | ||||
|     RET | ||||
| %endmacro | ||||
|  | ||||
| PIXELS48 put, 4 | ||||
| PIXELS48 avg, 4 | ||||
| PIXELS48 put, 8 | ||||
| PIXELS48 avg, 8 | ||||
|  | ||||
|  | ||||
| INIT_XMM sse2 | ||||
| ; void put_pixels16_sse2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h) | ||||
| cglobal put_pixels16, 4,5,4 | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     movu         m0, [r1] | ||||
|     movu         m1, [r1+r2] | ||||
|     movu         m2, [r1+r2*2] | ||||
|     movu         m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
|     mova       [r0], m0 | ||||
|     mova    [r0+r2], m1 | ||||
|     mova  [r0+r2*2], m2 | ||||
|     mova    [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jnz       .loop | ||||
|     REP_RET | ||||
|  | ||||
| ; void avg_pixels16_sse2(uint8_t *block, const uint8_t *pixels, ptrdiff_t line_size, int h) | ||||
| cglobal avg_pixels16, 4,5,4 | ||||
|     lea          r4, [r2*3] | ||||
| .loop: | ||||
|     movu         m0, [r1] | ||||
|     movu         m1, [r1+r2] | ||||
|     movu         m2, [r1+r2*2] | ||||
|     movu         m3, [r1+r4] | ||||
|     lea          r1, [r1+r2*4] | ||||
|     pavgb        m0, [r0] | ||||
|     pavgb        m1, [r0+r2] | ||||
|     pavgb        m2, [r0+r2*2] | ||||
|     pavgb        m3, [r0+r4] | ||||
|     mova       [r0], m0 | ||||
|     mova    [r0+r2], m1 | ||||
|     mova  [r0+r2*2], m2 | ||||
|     mova    [r0+r4], m3 | ||||
|     sub         r3d, 4 | ||||
|     lea          r0, [r0+r2*4] | ||||
|     jnz       .loop | ||||
|     REP_RET | ||||
		Reference in New Issue
	
	Block a user