mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2024-12-12 19:18:44 +02:00
bbe95f7353
From x86inc: > On AMD cpus <=K10, an ordinary ret is slow if it immediately follows either > a branch or a branch target. So switch to a 2-byte form of ret in that case. > We can automatically detect "follows a branch", but not a branch target. > (SSSE3 is a sufficient condition to know that your cpu doesn't have this problem.) x86inc can automatically determine whether to use REP_RET rather than REP in most of these cases, so impact is minimal. Additionally, a few REP_RETs were used unnecessary, despite the return being nowhere near a branch. The only CPUs affected were AMD K10s, made between 2007 and 2011, 16 years ago and 12 years ago, respectively. In the future, everyone involved with x86inc should consider dropping REP_RETs altogether.
1043 lines
26 KiB
NASM
1043 lines
26 KiB
NASM
;*****************************************************************************
|
|
;* MMX/SSE2/AVX-optimized 10-bit H.264 deblocking code
|
|
;*****************************************************************************
|
|
;* Copyright (C) 2005-2011 x264 project
|
|
;*
|
|
;* Authors: Oskar Arvidsson <oskar@irock.se>
|
|
;* Loren Merritt <lorenm@u.washington.edu>
|
|
;* Fiona Glaser <fiona@x264.com>
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION .text
|
|
|
|
cextern pw_2
|
|
cextern pw_3
|
|
cextern pw_4
|
|
cextern pw_1023
|
|
%define pw_pixel_max pw_1023
|
|
|
|
; out: %4 = |%1-%2|-%3
|
|
; clobbers: %5
|
|
%macro ABS_SUB 5
|
|
psubusw %5, %2, %1
|
|
psubusw %4, %1, %2
|
|
por %4, %5
|
|
psubw %4, %3
|
|
%endmacro
|
|
|
|
; out: %4 = |%1-%2|<%3
|
|
%macro DIFF_LT 5
|
|
psubusw %4, %2, %1
|
|
psubusw %5, %1, %2
|
|
por %5, %4 ; |%1-%2|
|
|
pxor %4, %4
|
|
psubw %5, %3 ; |%1-%2|-%3
|
|
pcmpgtw %4, %5 ; 0 > |%1-%2|-%3
|
|
%endmacro
|
|
|
|
%macro LOAD_AB 4
|
|
movd %1, %3
|
|
movd %2, %4
|
|
SPLATW %1, %1
|
|
SPLATW %2, %2
|
|
%endmacro
|
|
|
|
; in: %2=tc reg
|
|
; out: %1=splatted tc
|
|
%macro LOAD_TC 2
|
|
movd %1, [%2]
|
|
punpcklbw %1, %1
|
|
%if mmsize == 8
|
|
pshufw %1, %1, 0
|
|
%else
|
|
pshuflw %1, %1, 01010000b
|
|
pshufd %1, %1, 01010000b
|
|
%endif
|
|
psraw %1, 6
|
|
%endmacro
|
|
|
|
; in: %1=p1, %2=p0, %3=q0, %4=q1
|
|
; %5=alpha, %6=beta, %7-%9=tmp
|
|
; out: %7=mask
|
|
%macro LOAD_MASK 9
|
|
ABS_SUB %2, %3, %5, %8, %7 ; |p0-q0| - alpha
|
|
ABS_SUB %1, %2, %6, %9, %7 ; |p1-p0| - beta
|
|
pand %8, %9
|
|
ABS_SUB %3, %4, %6, %9, %7 ; |q1-q0| - beta
|
|
pxor %7, %7
|
|
pand %8, %9
|
|
pcmpgtw %7, %8
|
|
%endmacro
|
|
|
|
; in: %1=p0, %2=q0, %3=p1, %4=q1, %5=mask, %6=tmp, %7=tmp
|
|
; out: %1=p0', m2=q0'
|
|
%macro DEBLOCK_P0_Q0 7
|
|
psubw %3, %4
|
|
pxor %7, %7
|
|
paddw %3, [pw_4]
|
|
psubw %7, %5
|
|
psubw %6, %2, %1
|
|
psllw %6, 2
|
|
paddw %3, %6
|
|
psraw %3, 3
|
|
mova %6, [pw_pixel_max]
|
|
CLIPW %3, %7, %5
|
|
pxor %7, %7
|
|
paddw %1, %3
|
|
psubw %2, %3
|
|
CLIPW %1, %7, %6
|
|
CLIPW %2, %7, %6
|
|
%endmacro
|
|
|
|
; in: %1=x2, %2=x1, %3=p0, %4=q0 %5=mask&tc, %6=tmp
|
|
%macro LUMA_Q1 6
|
|
pavgw %6, %3, %4 ; (p0+q0+1)>>1
|
|
paddw %1, %6
|
|
pxor %6, %6
|
|
psraw %1, 1
|
|
psubw %6, %5
|
|
psubw %1, %2
|
|
CLIPW %1, %6, %5
|
|
paddw %1, %2
|
|
%endmacro
|
|
|
|
%macro LUMA_DEBLOCK_ONE 3
|
|
DIFF_LT m5, %1, bm, m4, m6
|
|
pxor m6, m6
|
|
mova %3, m4
|
|
pcmpgtw m6, tcm
|
|
pand m4, tcm
|
|
pandn m6, m7
|
|
pand m4, m6
|
|
LUMA_Q1 m5, %2, m1, m2, m4, m6
|
|
%endmacro
|
|
|
|
%macro LUMA_H_STORE 2
|
|
%if mmsize == 8
|
|
movq [r0-4], m0
|
|
movq [r0+r1-4], m1
|
|
movq [r0+r1*2-4], m2
|
|
movq [r0+%2-4], m3
|
|
%else
|
|
movq [r0-4], m0
|
|
movhps [r0+r1-4], m0
|
|
movq [r0+r1*2-4], m1
|
|
movhps [%1-4], m1
|
|
movq [%1+r1-4], m2
|
|
movhps [%1+r1*2-4], m2
|
|
movq [%1+%2-4], m3
|
|
movhps [%1+r1*4-4], m3
|
|
%endif
|
|
%endmacro
|
|
|
|
%macro DEBLOCK_LUMA 0
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_v_luma_10(uint16_t *pix, int stride, int alpha, int beta,
|
|
; int8_t *tc0)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_v_luma_10, 5,5,8*(mmsize/16)
|
|
%assign pad 5*mmsize+12-(stack_offset&15)
|
|
%define tcm [rsp]
|
|
%define ms1 [rsp+mmsize]
|
|
%define ms2 [rsp+mmsize*2]
|
|
%define am [rsp+mmsize*3]
|
|
%define bm [rsp+mmsize*4]
|
|
SUB rsp, pad
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
LOAD_AB m4, m5, r2d, r3d
|
|
mov r3, 32/mmsize
|
|
mov r2, r0
|
|
sub r0, r1
|
|
mova am, m4
|
|
sub r0, r1
|
|
mova bm, m5
|
|
sub r0, r1
|
|
.loop:
|
|
mova m0, [r0+r1]
|
|
mova m1, [r0+r1*2]
|
|
mova m2, [r2]
|
|
mova m3, [r2+r1]
|
|
|
|
LOAD_MASK m0, m1, m2, m3, am, bm, m7, m4, m6
|
|
LOAD_TC m6, r4
|
|
mova tcm, m6
|
|
|
|
mova m5, [r0]
|
|
LUMA_DEBLOCK_ONE m1, m0, ms1
|
|
mova [r0+r1], m5
|
|
|
|
mova m5, [r2+r1*2]
|
|
LUMA_DEBLOCK_ONE m2, m3, ms2
|
|
mova [r2+r1], m5
|
|
|
|
pxor m5, m5
|
|
mova m6, tcm
|
|
pcmpgtw m5, tcm
|
|
psubw m6, ms1
|
|
pandn m5, m7
|
|
psubw m6, ms2
|
|
pand m5, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m5, m7, m6
|
|
mova [r0+r1*2], m1
|
|
mova [r2], m2
|
|
|
|
add r0, mmsize
|
|
add r2, mmsize
|
|
add r4, mmsize/8
|
|
dec r3
|
|
jg .loop
|
|
ADD rsp, pad
|
|
RET
|
|
|
|
cglobal deblock_h_luma_10, 5,6,8*(mmsize/16)
|
|
%assign pad 7*mmsize+12-(stack_offset&15)
|
|
%define tcm [rsp]
|
|
%define ms1 [rsp+mmsize]
|
|
%define ms2 [rsp+mmsize*2]
|
|
%define p1m [rsp+mmsize*3]
|
|
%define p2m [rsp+mmsize*4]
|
|
%define am [rsp+mmsize*5]
|
|
%define bm [rsp+mmsize*6]
|
|
SUB rsp, pad
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
LOAD_AB m4, m5, r2d, r3d
|
|
mov r3, r1
|
|
mova am, m4
|
|
add r3, r1
|
|
mov r5, 32/mmsize
|
|
mova bm, m5
|
|
add r3, r1
|
|
%if mmsize == 16
|
|
mov r2, r0
|
|
add r2, r3
|
|
%endif
|
|
.loop:
|
|
%if mmsize == 8
|
|
movq m2, [r0-8] ; y q2 q1 q0
|
|
movq m7, [r0+0]
|
|
movq m5, [r0+r1-8]
|
|
movq m3, [r0+r1+0]
|
|
movq m0, [r0+r1*2-8]
|
|
movq m6, [r0+r1*2+0]
|
|
movq m1, [r0+r3-8]
|
|
TRANSPOSE4x4W 2, 5, 0, 1, 4
|
|
SWAP 2, 7
|
|
movq m7, [r0+r3]
|
|
TRANSPOSE4x4W 2, 3, 6, 7, 4
|
|
%else
|
|
movu m5, [r0-8] ; y q2 q1 q0 p0 p1 p2 x
|
|
movu m0, [r0+r1-8]
|
|
movu m2, [r0+r1*2-8]
|
|
movu m3, [r2-8]
|
|
TRANSPOSE4x4W 5, 0, 2, 3, 6
|
|
mova tcm, m3
|
|
|
|
movu m4, [r2+r1-8]
|
|
movu m1, [r2+r1*2-8]
|
|
movu m3, [r2+r3-8]
|
|
movu m7, [r2+r1*4-8]
|
|
TRANSPOSE4x4W 4, 1, 3, 7, 6
|
|
|
|
mova m6, tcm
|
|
punpcklqdq m6, m7
|
|
punpckhqdq m5, m4
|
|
SBUTTERFLY qdq, 0, 1, 7
|
|
SBUTTERFLY qdq, 2, 3, 7
|
|
%endif
|
|
|
|
mova p2m, m6
|
|
LOAD_MASK m0, m1, m2, m3, am, bm, m7, m4, m6
|
|
LOAD_TC m6, r4
|
|
mova tcm, m6
|
|
|
|
LUMA_DEBLOCK_ONE m1, m0, ms1
|
|
mova p1m, m5
|
|
|
|
mova m5, p2m
|
|
LUMA_DEBLOCK_ONE m2, m3, ms2
|
|
mova p2m, m5
|
|
|
|
pxor m5, m5
|
|
mova m6, tcm
|
|
pcmpgtw m5, tcm
|
|
psubw m6, ms1
|
|
pandn m5, m7
|
|
psubw m6, ms2
|
|
pand m5, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m5, m7, m6
|
|
mova m0, p1m
|
|
mova m3, p2m
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
LUMA_H_STORE r2, r3
|
|
|
|
add r4, mmsize/8
|
|
lea r0, [r0+r1*(mmsize/2)]
|
|
lea r2, [r2+r1*(mmsize/2)]
|
|
dec r5
|
|
jg .loop
|
|
ADD rsp, pad
|
|
RET
|
|
%endmacro
|
|
|
|
%if ARCH_X86_64
|
|
; in: m0=p1, m1=p0, m2=q0, m3=q1, m8=p2, m9=q2
|
|
; m12=alpha, m13=beta
|
|
; out: m0=p1', m3=q1', m1=p0', m2=q0'
|
|
; clobbers: m4, m5, m6, m7, m10, m11, m14
|
|
%macro DEBLOCK_LUMA_INTER_SSE2 0
|
|
LOAD_MASK m0, m1, m2, m3, m12, m13, m7, m4, m6
|
|
LOAD_TC m6, r4
|
|
DIFF_LT m8, m1, m13, m10, m4
|
|
DIFF_LT m9, m2, m13, m11, m4
|
|
pand m6, m7
|
|
|
|
mova m14, m6
|
|
pxor m4, m4
|
|
pcmpgtw m6, m4
|
|
pand m6, m14
|
|
|
|
mova m5, m10
|
|
pand m5, m6
|
|
LUMA_Q1 m8, m0, m1, m2, m5, m4
|
|
|
|
mova m5, m11
|
|
pand m5, m6
|
|
LUMA_Q1 m9, m3, m1, m2, m5, m4
|
|
|
|
pxor m4, m4
|
|
psubw m6, m10
|
|
pcmpgtw m4, m14
|
|
pandn m4, m7
|
|
psubw m6, m11
|
|
pand m4, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m4, m5, m6
|
|
|
|
SWAP 0, 8
|
|
SWAP 3, 9
|
|
%endmacro
|
|
|
|
%macro DEBLOCK_LUMA_64 0
|
|
cglobal deblock_v_luma_10, 5,5,15
|
|
%define p2 m8
|
|
%define p1 m0
|
|
%define p0 m1
|
|
%define q0 m2
|
|
%define q1 m3
|
|
%define q2 m9
|
|
%define mask0 m7
|
|
%define mask1 m10
|
|
%define mask2 m11
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
LOAD_AB m12, m13, r2d, r3d
|
|
mov r2, r0
|
|
sub r0, r1
|
|
sub r0, r1
|
|
sub r0, r1
|
|
mov r3, 2
|
|
.loop:
|
|
mova p2, [r0]
|
|
mova p1, [r0+r1]
|
|
mova p0, [r0+r1*2]
|
|
mova q0, [r2]
|
|
mova q1, [r2+r1]
|
|
mova q2, [r2+r1*2]
|
|
DEBLOCK_LUMA_INTER_SSE2
|
|
mova [r0+r1], p1
|
|
mova [r0+r1*2], p0
|
|
mova [r2], q0
|
|
mova [r2+r1], q1
|
|
add r0, mmsize
|
|
add r2, mmsize
|
|
add r4, 2
|
|
dec r3
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal deblock_h_luma_10, 5,7,15
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
LOAD_AB m12, m13, r2d, r3d
|
|
mov r2, r1
|
|
add r2, r1
|
|
add r2, r1
|
|
mov r5, r0
|
|
add r5, r2
|
|
mov r6, 2
|
|
.loop:
|
|
movu m8, [r0-8] ; y q2 q1 q0 p0 p1 p2 x
|
|
movu m0, [r0+r1-8]
|
|
movu m2, [r0+r1*2-8]
|
|
movu m9, [r5-8]
|
|
movu m5, [r5+r1-8]
|
|
movu m1, [r5+r1*2-8]
|
|
movu m3, [r5+r2-8]
|
|
movu m7, [r5+r1*4-8]
|
|
|
|
TRANSPOSE4x4W 8, 0, 2, 9, 10
|
|
TRANSPOSE4x4W 5, 1, 3, 7, 10
|
|
|
|
punpckhqdq m8, m5
|
|
SBUTTERFLY qdq, 0, 1, 10
|
|
SBUTTERFLY qdq, 2, 3, 10
|
|
punpcklqdq m9, m7
|
|
|
|
DEBLOCK_LUMA_INTER_SSE2
|
|
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
LUMA_H_STORE r5, r2
|
|
add r4, 2
|
|
lea r0, [r0+r1*8]
|
|
lea r5, [r5+r1*8]
|
|
dec r6
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DEBLOCK_LUMA_64
|
|
%if HAVE_AVX_EXTERNAL
|
|
INIT_XMM avx
|
|
DEBLOCK_LUMA_64
|
|
%endif
|
|
%endif
|
|
|
|
%macro SWAPMOVA 2
|
|
%ifid %1
|
|
SWAP %1, %2
|
|
%else
|
|
mova %1, %2
|
|
%endif
|
|
%endmacro
|
|
|
|
; in: t0-t2: tmp registers
|
|
; %1=p0 %2=p1 %3=p2 %4=p3 %5=q0 %6=q1 %7=mask0
|
|
; %8=mask1p %9=2 %10=p0' %11=p1' %12=p2'
|
|
%macro LUMA_INTRA_P012 12 ; p0..p3 in memory
|
|
%if ARCH_X86_64
|
|
paddw t0, %3, %2
|
|
mova t2, %4
|
|
paddw t2, %3
|
|
%else
|
|
mova t0, %3
|
|
mova t2, %4
|
|
paddw t0, %2
|
|
paddw t2, %3
|
|
%endif
|
|
paddw t0, %1
|
|
paddw t2, t2
|
|
paddw t0, %5
|
|
paddw t2, %9
|
|
paddw t0, %9 ; (p2 + p1 + p0 + q0 + 2)
|
|
paddw t2, t0 ; (2*p3 + 3*p2 + p1 + p0 + q0 + 4)
|
|
|
|
psrlw t2, 3
|
|
psrlw t1, t0, 2
|
|
psubw t2, %3
|
|
psubw t1, %2
|
|
pand t2, %8
|
|
pand t1, %8
|
|
paddw t2, %3
|
|
paddw t1, %2
|
|
SWAPMOVA %11, t1
|
|
|
|
psubw t1, t0, %3
|
|
paddw t0, t0
|
|
psubw t1, %5
|
|
psubw t0, %3
|
|
paddw t1, %6
|
|
paddw t1, %2
|
|
paddw t0, %6
|
|
psrlw t1, 2 ; (2*p1 + p0 + q1 + 2)/4
|
|
psrlw t0, 3 ; (p2 + 2*p1 + 2*p0 + 2*q0 + q1 + 4)>>3
|
|
|
|
pxor t0, t1
|
|
pxor t1, %1
|
|
pand t0, %8
|
|
pand t1, %7
|
|
pxor t0, t1
|
|
pxor t0, %1
|
|
SWAPMOVA %10, t0
|
|
SWAPMOVA %12, t2
|
|
%endmacro
|
|
|
|
%macro LUMA_INTRA_INIT 1
|
|
%xdefine pad %1*mmsize+((gprsize*3) % mmsize)-(stack_offset&15)
|
|
%define t0 m4
|
|
%define t1 m5
|
|
%define t2 m6
|
|
%define t3 m7
|
|
%assign i 4
|
|
%rep %1
|
|
CAT_XDEFINE t, i, [rsp+mmsize*(i-4)]
|
|
%assign i i+1
|
|
%endrep
|
|
SUB rsp, pad
|
|
%endmacro
|
|
|
|
; in: %1-%3=tmp, %4=p2, %5=q2
|
|
%macro LUMA_INTRA_INTER 5
|
|
LOAD_AB t0, t1, r2d, r3d
|
|
mova %1, t0
|
|
LOAD_MASK m0, m1, m2, m3, %1, t1, t0, t2, t3
|
|
%if ARCH_X86_64
|
|
mova %2, t0 ; mask0
|
|
psrlw t3, %1, 2
|
|
%else
|
|
mova t3, %1
|
|
mova %2, t0 ; mask0
|
|
psrlw t3, 2
|
|
%endif
|
|
paddw t3, [pw_2] ; alpha/4+2
|
|
DIFF_LT m1, m2, t3, t2, t0 ; t2 = |p0-q0| < alpha/4+2
|
|
pand t2, %2
|
|
mova t3, %5 ; q2
|
|
mova %1, t2 ; mask1
|
|
DIFF_LT t3, m2, t1, t2, t0 ; t2 = |q2-q0| < beta
|
|
pand t2, %1
|
|
mova t3, %4 ; p2
|
|
mova %3, t2 ; mask1q
|
|
DIFF_LT t3, m1, t1, t2, t0 ; t2 = |p2-p0| < beta
|
|
pand t2, %1
|
|
mova %1, t2 ; mask1p
|
|
%endmacro
|
|
|
|
%macro LUMA_H_INTRA_LOAD 0
|
|
%if mmsize == 8
|
|
movu t0, [r0-8]
|
|
movu t1, [r0+r1-8]
|
|
movu m0, [r0+r1*2-8]
|
|
movu m1, [r0+r4-8]
|
|
TRANSPOSE4x4W 4, 5, 0, 1, 2
|
|
mova t4, t0 ; p3
|
|
mova t5, t1 ; p2
|
|
|
|
movu m2, [r0]
|
|
movu m3, [r0+r1]
|
|
movu t0, [r0+r1*2]
|
|
movu t1, [r0+r4]
|
|
TRANSPOSE4x4W 2, 3, 4, 5, 6
|
|
mova t6, t0 ; q2
|
|
mova t7, t1 ; q3
|
|
%else
|
|
movu t0, [r0-8]
|
|
movu t1, [r0+r1-8]
|
|
movu m0, [r0+r1*2-8]
|
|
movu m1, [r0+r5-8]
|
|
movu m2, [r4-8]
|
|
movu m3, [r4+r1-8]
|
|
movu t2, [r4+r1*2-8]
|
|
movu t3, [r4+r5-8]
|
|
TRANSPOSE8x8W 4, 5, 0, 1, 2, 3, 6, 7, t4, t5
|
|
mova t4, t0 ; p3
|
|
mova t5, t1 ; p2
|
|
mova t6, t2 ; q2
|
|
mova t7, t3 ; q3
|
|
%endif
|
|
%endmacro
|
|
|
|
; in: %1=q3 %2=q2' %3=q1' %4=q0' %5=p0' %6=p1' %7=p2' %8=p3 %9=tmp
|
|
%macro LUMA_H_INTRA_STORE 9
|
|
%if mmsize == 8
|
|
TRANSPOSE4x4W %1, %2, %3, %4, %9
|
|
movq [r0-8], m%1
|
|
movq [r0+r1-8], m%2
|
|
movq [r0+r1*2-8], m%3
|
|
movq [r0+r4-8], m%4
|
|
movq m%1, %8
|
|
TRANSPOSE4x4W %5, %6, %7, %1, %9
|
|
movq [r0], m%5
|
|
movq [r0+r1], m%6
|
|
movq [r0+r1*2], m%7
|
|
movq [r0+r4], m%1
|
|
%else
|
|
TRANSPOSE2x4x4W %1, %2, %3, %4, %9
|
|
movq [r0-8], m%1
|
|
movq [r0+r1-8], m%2
|
|
movq [r0+r1*2-8], m%3
|
|
movq [r0+r5-8], m%4
|
|
movhps [r4-8], m%1
|
|
movhps [r4+r1-8], m%2
|
|
movhps [r4+r1*2-8], m%3
|
|
movhps [r4+r5-8], m%4
|
|
%ifnum %8
|
|
SWAP %1, %8
|
|
%else
|
|
mova m%1, %8
|
|
%endif
|
|
TRANSPOSE2x4x4W %5, %6, %7, %1, %9
|
|
movq [r0], m%5
|
|
movq [r0+r1], m%6
|
|
movq [r0+r1*2], m%7
|
|
movq [r0+r5], m%1
|
|
movhps [r4], m%5
|
|
movhps [r4+r1], m%6
|
|
movhps [r4+r1*2], m%7
|
|
movhps [r4+r5], m%1
|
|
%endif
|
|
%endmacro
|
|
|
|
%if ARCH_X86_64
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_v_luma_intra_10(uint16_t *pix, int stride, int alpha,
|
|
; int beta)
|
|
;-----------------------------------------------------------------------------
|
|
%macro DEBLOCK_LUMA_INTRA_64 0
|
|
cglobal deblock_v_luma_intra_10, 4,7,16
|
|
%define t0 m1
|
|
%define t1 m2
|
|
%define t2 m4
|
|
%define p2 m8
|
|
%define p1 m9
|
|
%define p0 m10
|
|
%define q0 m11
|
|
%define q1 m12
|
|
%define q2 m13
|
|
%define aa m5
|
|
%define bb m14
|
|
lea r4, [r1*4]
|
|
lea r5, [r1*3] ; 3*stride
|
|
neg r4
|
|
add r4, r0 ; pix-4*stride
|
|
mov r6, 2
|
|
mova m0, [pw_2]
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
LOAD_AB aa, bb, r2d, r3d
|
|
.loop:
|
|
mova p2, [r4+r1]
|
|
mova p1, [r4+2*r1]
|
|
mova p0, [r4+r5]
|
|
mova q0, [r0]
|
|
mova q1, [r0+r1]
|
|
mova q2, [r0+2*r1]
|
|
|
|
LOAD_MASK p1, p0, q0, q1, aa, bb, m3, t0, t1
|
|
mova t2, aa
|
|
psrlw t2, 2
|
|
paddw t2, m0 ; alpha/4+2
|
|
DIFF_LT p0, q0, t2, m6, t0 ; m6 = |p0-q0| < alpha/4+2
|
|
DIFF_LT p2, p0, bb, t1, t0 ; m7 = |p2-p0| < beta
|
|
DIFF_LT q2, q0, bb, m7, t0 ; t1 = |q2-q0| < beta
|
|
pand m6, m3
|
|
pand m7, m6
|
|
pand m6, t1
|
|
LUMA_INTRA_P012 p0, p1, p2, [r4], q0, q1, m3, m6, m0, [r4+r5], [r4+2*r1], [r4+r1]
|
|
LUMA_INTRA_P012 q0, q1, q2, [r0+r5], p0, p1, m3, m7, m0, [r0], [r0+r1], [r0+2*r1]
|
|
add r0, mmsize
|
|
add r4, mmsize
|
|
dec r6
|
|
jg .loop
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_h_luma_intra_10(uint16_t *pix, int stride, int alpha,
|
|
; int beta)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_h_luma_intra_10, 4,7,16
|
|
%define t0 m15
|
|
%define t1 m14
|
|
%define t2 m2
|
|
%define q3 m5
|
|
%define q2 m8
|
|
%define q1 m9
|
|
%define q0 m10
|
|
%define p0 m11
|
|
%define p1 m12
|
|
%define p2 m13
|
|
%define p3 m4
|
|
%define spill [rsp]
|
|
%assign pad 24-(stack_offset&15)
|
|
SUB rsp, pad
|
|
lea r4, [r1*4]
|
|
lea r5, [r1*3] ; 3*stride
|
|
add r4, r0 ; pix+4*stride
|
|
mov r6, 2
|
|
mova m0, [pw_2]
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
.loop:
|
|
movu q3, [r0-8]
|
|
movu q2, [r0+r1-8]
|
|
movu q1, [r0+r1*2-8]
|
|
movu q0, [r0+r5-8]
|
|
movu p0, [r4-8]
|
|
movu p1, [r4+r1-8]
|
|
movu p2, [r4+r1*2-8]
|
|
movu p3, [r4+r5-8]
|
|
TRANSPOSE8x8W 5, 8, 9, 10, 11, 12, 13, 4, 1
|
|
|
|
LOAD_AB m1, m2, r2d, r3d
|
|
LOAD_MASK q1, q0, p0, p1, m1, m2, m3, t0, t1
|
|
psrlw m1, 2
|
|
paddw m1, m0 ; alpha/4+2
|
|
DIFF_LT p0, q0, m1, m6, t0 ; m6 = |p0-q0| < alpha/4+2
|
|
DIFF_LT q2, q0, m2, t1, t0 ; t1 = |q2-q0| < beta
|
|
DIFF_LT p0, p2, m2, m7, t0 ; m7 = |p2-p0| < beta
|
|
pand m6, m3
|
|
pand m7, m6
|
|
pand m6, t1
|
|
|
|
mova spill, q3
|
|
LUMA_INTRA_P012 q0, q1, q2, q3, p0, p1, m3, m6, m0, m5, m1, q2
|
|
LUMA_INTRA_P012 p0, p1, p2, p3, q0, q1, m3, m7, m0, p0, m6, p2
|
|
mova m7, spill
|
|
|
|
LUMA_H_INTRA_STORE 7, 8, 1, 5, 11, 6, 13, 4, 14
|
|
|
|
lea r0, [r0+r1*8]
|
|
lea r4, [r4+r1*8]
|
|
dec r6
|
|
jg .loop
|
|
ADD rsp, pad
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DEBLOCK_LUMA_INTRA_64
|
|
%if HAVE_AVX_EXTERNAL
|
|
INIT_XMM avx
|
|
DEBLOCK_LUMA_INTRA_64
|
|
%endif
|
|
|
|
%endif
|
|
|
|
%macro DEBLOCK_LUMA_INTRA 0
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_v_luma_intra_10(uint16_t *pix, int stride, int alpha,
|
|
; int beta)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_v_luma_intra_10, 4,7,8*(mmsize/16)
|
|
LUMA_INTRA_INIT 3
|
|
lea r4, [r1*4]
|
|
lea r5, [r1*3]
|
|
neg r4
|
|
add r4, r0
|
|
mov r6, 32/mmsize
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
.loop:
|
|
mova m0, [r4+r1*2] ; p1
|
|
mova m1, [r4+r5] ; p0
|
|
mova m2, [r0] ; q0
|
|
mova m3, [r0+r1] ; q1
|
|
LUMA_INTRA_INTER t4, t5, t6, [r4+r1], [r0+r1*2]
|
|
LUMA_INTRA_P012 m1, m0, t3, [r4], m2, m3, t5, t4, [pw_2], [r4+r5], [r4+2*r1], [r4+r1]
|
|
mova t3, [r0+r1*2] ; q2
|
|
LUMA_INTRA_P012 m2, m3, t3, [r0+r5], m1, m0, t5, t6, [pw_2], [r0], [r0+r1], [r0+2*r1]
|
|
add r0, mmsize
|
|
add r4, mmsize
|
|
dec r6
|
|
jg .loop
|
|
ADD rsp, pad
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_h_luma_intra_10(uint16_t *pix, int stride, int alpha,
|
|
; int beta)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_h_luma_intra_10, 4,7,8*(mmsize/16)
|
|
LUMA_INTRA_INIT 8
|
|
%if mmsize == 8
|
|
lea r4, [r1*3]
|
|
mov r5, 32/mmsize
|
|
%else
|
|
lea r4, [r1*4]
|
|
lea r5, [r1*3] ; 3*stride
|
|
add r4, r0 ; pix+4*stride
|
|
mov r6, 32/mmsize
|
|
%endif
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
.loop:
|
|
LUMA_H_INTRA_LOAD
|
|
LUMA_INTRA_INTER t8, t9, t10, t5, t6
|
|
|
|
LUMA_INTRA_P012 m1, m0, t3, t4, m2, m3, t9, t8, [pw_2], t8, t5, t11
|
|
mova t3, t6 ; q2
|
|
LUMA_INTRA_P012 m2, m3, t3, t7, m1, m0, t9, t10, [pw_2], m4, t6, m5
|
|
|
|
mova m2, t4
|
|
mova m0, t11
|
|
mova m1, t5
|
|
mova m3, t8
|
|
mova m6, t6
|
|
|
|
LUMA_H_INTRA_STORE 2, 0, 1, 3, 4, 6, 5, t7, 7
|
|
|
|
lea r0, [r0+r1*(mmsize/2)]
|
|
%if mmsize == 8
|
|
dec r5
|
|
%else
|
|
lea r4, [r4+r1*(mmsize/2)]
|
|
dec r6
|
|
%endif
|
|
jg .loop
|
|
ADD rsp, pad
|
|
RET
|
|
%endmacro
|
|
|
|
%if ARCH_X86_64 == 0
|
|
%if HAVE_ALIGNED_STACK == 0
|
|
INIT_MMX mmxext
|
|
DEBLOCK_LUMA
|
|
DEBLOCK_LUMA_INTRA
|
|
%endif
|
|
INIT_XMM sse2
|
|
DEBLOCK_LUMA
|
|
DEBLOCK_LUMA_INTRA
|
|
%if HAVE_AVX_EXTERNAL
|
|
INIT_XMM avx
|
|
DEBLOCK_LUMA
|
|
DEBLOCK_LUMA_INTRA
|
|
%endif
|
|
%endif
|
|
|
|
; in: %1=p0, %2=q0, %3=p1, %4=q1, %5=mask, %6=tmp, %7=tmp
|
|
; out: %1=p0', %2=q0'
|
|
%macro CHROMA_DEBLOCK_P0_Q0_INTRA 7
|
|
mova %6, [pw_2]
|
|
paddw %6, %3
|
|
paddw %6, %4
|
|
paddw %7, %6, %2
|
|
paddw %6, %1
|
|
paddw %6, %3
|
|
paddw %7, %4
|
|
psraw %6, 2
|
|
psraw %7, 2
|
|
psubw %6, %1
|
|
psubw %7, %2
|
|
pand %6, %5
|
|
pand %7, %5
|
|
paddw %1, %6
|
|
paddw %2, %7
|
|
%endmacro
|
|
|
|
%macro CHROMA_V_LOAD 1
|
|
mova m0, [r0] ; p1
|
|
mova m1, [r0+r1] ; p0
|
|
mova m2, [%1] ; q0
|
|
mova m3, [%1+r1] ; q1
|
|
%endmacro
|
|
|
|
%macro CHROMA_V_STORE 0
|
|
mova [r0+1*r1], m1
|
|
mova [r0+2*r1], m2
|
|
%endmacro
|
|
|
|
; in: 8 rows of 4 words in %4..%11
|
|
; out: 4 rows of 8 words in m0..m3
|
|
%macro TRANSPOSE4x8W_LOAD 8
|
|
movq m0, %1
|
|
movq m2, %2
|
|
movq m1, %3
|
|
movq m3, %4
|
|
|
|
punpcklwd m0, m2
|
|
punpcklwd m1, m3
|
|
punpckhdq m2, m0, m1
|
|
punpckldq m0, m1
|
|
|
|
movq m4, %5
|
|
movq m6, %6
|
|
movq m5, %7
|
|
movq m3, %8
|
|
|
|
punpcklwd m4, m6
|
|
punpcklwd m5, m3
|
|
punpckhdq m6, m4, m5
|
|
punpckldq m4, m5
|
|
|
|
punpckhqdq m1, m0, m4
|
|
punpcklqdq m0, m4
|
|
punpckhqdq m3, m2, m6
|
|
punpcklqdq m2, m6
|
|
%endmacro
|
|
|
|
; in: 4 rows of 8 words in m0..m3
|
|
; out: 8 rows of 4 words in %1..%8
|
|
%macro TRANSPOSE8x4W_STORE 8
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
movq %1, m0
|
|
movhps %2, m0
|
|
movq %3, m1
|
|
movhps %4, m1
|
|
movq %5, m2
|
|
movhps %6, m2
|
|
movq %7, m3
|
|
movhps %8, m3
|
|
%endmacro
|
|
|
|
; %1 = base + 3*stride
|
|
; %2 = 3*stride (unused on mmx)
|
|
; %3, %4 = place to store p1 and q1 values
|
|
%macro CHROMA_H_LOAD 4
|
|
%if mmsize == 8
|
|
movq m0, [pix_q - 4]
|
|
movq m1, [pix_q + stride_q - 4]
|
|
movq m2, [pix_q + 2*stride_q - 4]
|
|
movq m3, [%1 - 4]
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
%else
|
|
TRANSPOSE4x8W_LOAD PASS8ROWS(pix_q-4, %1-4, stride_q, %2)
|
|
%endif
|
|
mova %3, m0
|
|
mova %4, m3
|
|
%endmacro
|
|
|
|
; %1 = base + 3*stride
|
|
; %2 = 3*stride (unused on mmx)
|
|
; %3, %4 = place to load p1 and q1 values
|
|
%macro CHROMA_H_STORE 4
|
|
mova m0, %3
|
|
mova m3, %4
|
|
%if mmsize == 8
|
|
TRANSPOSE4x4W 0, 1, 2, 3, 4
|
|
movq [pix_q - 4], m0
|
|
movq [pix_q + stride_q - 4], m1
|
|
movq [pix_q + 2*stride_q - 4], m2
|
|
movq [%1 - 4], m3
|
|
%else
|
|
TRANSPOSE8x4W_STORE PASS8ROWS(pix_q-4, %1-4, stride_q, %2)
|
|
%endif
|
|
%endmacro
|
|
|
|
%macro CHROMA_V_LOAD_TC 2
|
|
movd %1, [%2]
|
|
punpcklbw %1, %1
|
|
punpcklwd %1, %1
|
|
psraw %1, 6
|
|
%endmacro
|
|
|
|
%macro DEBLOCK_CHROMA 0
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_v_chroma_10(uint16_t *pix, int stride, int alpha, int beta,
|
|
; int8_t *tc0)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_v_chroma_10, 5,7-(mmsize/16),8*(mmsize/16)
|
|
mov r5, r0
|
|
sub r0, r1
|
|
sub r0, r1
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
CHROMA_V_LOAD r5
|
|
LOAD_AB m4, m5, r2d, r3d
|
|
LOAD_MASK m0, m1, m2, m3, m4, m5, m7, m6, m4
|
|
pxor m4, m4
|
|
CHROMA_V_LOAD_TC m6, r4
|
|
psubw m6, [pw_3]
|
|
pmaxsw m6, m4
|
|
pand m7, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m7, m5, m6
|
|
CHROMA_V_STORE
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_v_chroma_intra_10(uint16_t *pix, int stride, int alpha,
|
|
; int beta)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_v_chroma_intra_10, 4,6-(mmsize/16),8*(mmsize/16)
|
|
mov r4, r0
|
|
sub r0, r1
|
|
sub r0, r1
|
|
shl r2d, 2
|
|
shl r3d, 2
|
|
CHROMA_V_LOAD r4
|
|
LOAD_AB m4, m5, r2d, r3d
|
|
LOAD_MASK m0, m1, m2, m3, m4, m5, m7, m6, m4
|
|
CHROMA_DEBLOCK_P0_Q0_INTRA m1, m2, m0, m3, m7, m5, m6
|
|
CHROMA_V_STORE
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_h_chroma_10(uint16_t *pix, int stride, int alpha, int beta,
|
|
; int8_t *tc0)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_h_chroma_10, 5, 7, 8, 0-2*mmsize, pix_, stride_, alpha_, beta_, tc0_
|
|
shl alpha_d, 2
|
|
shl beta_d, 2
|
|
mov r5, pix_q
|
|
lea r6, [3*stride_q]
|
|
add r5, r6
|
|
|
|
CHROMA_H_LOAD r5, r6, [rsp], [rsp + mmsize]
|
|
LOAD_AB m4, m5, alpha_d, beta_d
|
|
LOAD_MASK m0, m1, m2, m3, m4, m5, m7, m6, m4
|
|
pxor m4, m4
|
|
CHROMA_V_LOAD_TC m6, tc0_q
|
|
psubw m6, [pw_3]
|
|
pmaxsw m6, m4
|
|
pand m7, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m7, m5, m6
|
|
CHROMA_H_STORE r5, r6, [rsp], [rsp + mmsize]
|
|
|
|
RET
|
|
|
|
;-----------------------------------------------------------------------------
|
|
; void ff_deblock_h_chroma422_10(uint16_t *pix, int stride, int alpha, int beta,
|
|
; int8_t *tc0)
|
|
;-----------------------------------------------------------------------------
|
|
cglobal deblock_h_chroma422_10, 5, 7, 8, 0-3*mmsize, pix_, stride_, alpha_, beta_, tc0_
|
|
shl alpha_d, 2
|
|
shl beta_d, 2
|
|
|
|
movd m0, [tc0_q]
|
|
punpcklbw m0, m0
|
|
psraw m0, 6
|
|
movq [rsp], m0
|
|
|
|
mov r5, pix_q
|
|
lea r6, [3*stride_q]
|
|
add r5, r6
|
|
|
|
mov r4, -8
|
|
.loop:
|
|
|
|
CHROMA_H_LOAD r5, r6, [rsp + 1*mmsize], [rsp + 2*mmsize]
|
|
LOAD_AB m4, m5, alpha_d, beta_d
|
|
LOAD_MASK m0, m1, m2, m3, m4, m5, m7, m6, m4
|
|
pxor m4, m4
|
|
movd m6, [rsp + r4 + 8]
|
|
punpcklwd m6, m6
|
|
punpcklwd m6, m6
|
|
psubw m6, [pw_3]
|
|
pmaxsw m6, m4
|
|
pand m7, m6
|
|
DEBLOCK_P0_Q0 m1, m2, m0, m3, m7, m5, m6
|
|
CHROMA_H_STORE r5, r6, [rsp + 1*mmsize], [rsp + 2*mmsize]
|
|
|
|
lea pix_q, [pix_q + (mmsize/2)*stride_q]
|
|
lea r5, [r5 + (mmsize/2)*stride_q]
|
|
add r4, (mmsize/4)
|
|
jl .loop
|
|
RET
|
|
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DEBLOCK_CHROMA
|
|
%if HAVE_AVX_EXTERNAL
|
|
INIT_XMM avx
|
|
DEBLOCK_CHROMA
|
|
%endif
|