1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2024-11-26 19:01:44 +02:00
FFmpeg/libavcodec/riscv/me_cmp_rvv.S
sunyuechi a7ad76fbbf lavc/me_cmp: R-V V nsse
C908:
nsse_0_c: 1990.0
nsse_0_rvv_i32: 572.0
nsse_1_c: 910.0
nsse_1_rvv_i32: 456.0

Signed-off-by: Rémi Denis-Courmont <remi@remlab.net>
2024-02-27 20:31:30 +02:00

528 lines
15 KiB
ArmAsm

/*
* Copyright (c) 2024 Institue of Software Chinese Academy of Sciences (ISCAS).
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/riscv/asm.S"
.macro pix_abs_ret
vsetivli zero, 1, e32, m1, ta, ma
vmv.x.s a0, v0
ret
.endm
func ff_pix_abs16_rvv, zve32x
vsetivli zero, 1, e32, m1, ta, ma
vmv.s.x v0, zero
1:
vsetivli zero, 16, e8, m1, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
add a1, a1, a3
vwsubu.vv v20, v12, v4
vsetvli zero, zero, e16, m2, tu, ma
vmax.vv v16, v16, v20
add a2, a2, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_pix_abs8_rvv, zve32x
vsetivli zero, 1, e32, m1, ta, ma
vmv.s.x v0, zero
1:
vsetivli zero, 8, e8, mf2, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
add a1, a1, a3
vwsubu.vv v20, v12, v4
vsetvli zero, zero, e16, m1, tu, ma
vmax.vv v16, v16, v20
add a2, a2, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_pix_abs16_x2_rvv, zve32x
csrwi vxrm, 0
vsetivli zero, 1, e32, m1, ta, ma
li t5, 1
vmv.s.x v0, zero
1:
vsetivli zero, 17, e8, m2, tu, ma
vle8.v v12, (a2)
addi a4, a4, -1
vslide1down.vx v24, v12, t5
vsetivli zero, 16, e8, m1, tu, ma
vle8.v v4, (a1)
vaaddu.vv v12, v12, v24
vwsubu.vv v16, v4, v12
add a1, a1, a3
vwsubu.vv v20, v12, v4
vsetvli zero, zero, e16, m2, tu, ma
vmax.vv v16, v16, v20
add a2, a2, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_pix_abs8_x2_rvv, zve32x
csrwi vxrm, 0
vsetivli zero, 1, e32, m1, ta, ma
li t5, 1
vmv.s.x v0, zero
1:
vsetivli zero, 9, e8, m1, tu, ma
vle8.v v12, (a2)
addi a4, a4, -1
vslide1down.vx v24, v12, t5
vsetivli zero, 8, e8, mf2, tu, ma
vle8.v v4, (a1)
vaaddu.vv v12, v12, v24
vwsubu.vv v16, v4, v12
add a1, a1, a3
vwsubu.vv v20, v12, v4
vsetvli zero, zero, e16, m1, tu, ma
vmax.vv v16, v16, v20
add a2, a2, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_pix_abs16_y2_rvv, zve32x
csrwi vxrm, 0
vsetivli zero, 1, e32, m1, ta, ma
add t1, a2, a3
vmv.s.x v0, zero
1:
vsetivli zero, 16, e8, m1, tu, ma
vle8.v v12, (a2)
vle8.v v24, (t1)
addi a4, a4, -1
vle8.v v4, (a1)
vaaddu.vv v12, v12, v24
vwsubu.vv v16, v4, v12
vwsubu.vv v20, v12, v4
add a1, a1, a3
vsetvli zero, zero, e16, m2, tu, ma
add a2, a2, a3
vmax.vv v16, v16, v20
add t1, t1, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_pix_abs8_y2_rvv, zve32x
csrwi vxrm, 0
vsetivli zero, 1, e32, m1, ta, ma
add t1, a2, a3
vmv.s.x v0, zero
1:
vsetivli zero, 8, e8, mf2, tu, ma
vle8.v v12, (a2)
vle8.v v24, (t1)
addi a4, a4, -1
vle8.v v4, (a1)
vaaddu.vv v12, v12, v24
vwsubu.vv v16, v4, v12
vwsubu.vv v20, v12, v4
add a1, a1, a3
vsetvli zero, zero, e16, m1, tu, ma
add a2, a2, a3
vmax.vv v16, v16, v20
add t1, t1, a3
vwredsum.vs v0, v16, v0
bnez a4, 1b
pix_abs_ret
endfunc
func ff_sse16_rvv, zve32x
vsetivli t0, 16, e32, m4, ta, ma
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, m1, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m2, tu, ma
vwmacc.vv v24, v16, v16
add a1, a1, a3
add a2, a2, a3
bnez a4, 1b
vsetvli zero, zero, e32, m4, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
endfunc
func ff_sse8_rvv, zve32x
vsetivli t0, 8, e32, m2, ta, ma
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, mf2, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m1, tu, ma
vwmacc.vv v24, v16, v16
add a1, a1, a3
add a2, a2, a3
bnez a4, 1b
vsetvli zero, zero, e32, m2, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
endfunc
func ff_sse4_rvv, zve32x
vsetivli t0, 4, e32, m1, ta, ma
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, mf4, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, mf2, tu, ma
vwmacc.vv v24, v16, v16
add a1, a1, a3
add a2, a2, a3
bnez a4, 1b
vsetvli zero, zero, e32, m1, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
endfunc
.macro vabsaddu dst src tmp
vneg.v \tmp, \src
vmax.vv \tmp, \src, \tmp
vwaddu.wv \dst, \dst, \tmp
.endm
.macro vsad_vsse16 type
vsetivli t0, 16, e32, m4, ta, ma
addi a4, a4, -1
add t1, a1, a3
add t2, a2, a3
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, m1, tu, ma
vle8.v v4, (a1)
vle8.v v8, (t1)
vle8.v v12, (a2)
vle8.v v16, (t2)
addi a4, a4, -1
vwsubu.vv v28, v4, v12
vwsubu.wv v12, v28, v8
vwaddu.wv v28, v12, v16
vsetvli zero, zero, e16, m2, tu, ma
.ifc \type,abs
vabsaddu v24, v28, v12
.endif
.ifc \type,square
vwmacc.vv v24, v28, v28
.endif
add a1, a1, a3
add a2, a2, a3
add t1, t1, a3
add t2, t2, a3
bnez a4, 1b
vsetvli zero, zero, e32, m4, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
.endm
.macro vsad_vsse8 type
vsetivli t0, 8, e32, m2, ta, ma
addi a4, a4, -1
add t1, a1, a3
add t2, a2, a3
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, mf2, tu, ma
vle8.v v4, (a1)
vle8.v v8, (t1)
vle8.v v12, (a2)
vle8.v v16, (t2)
addi a4, a4, -1
vwsubu.vv v28, v4, v12
vwsubu.wv v12, v28, v8
vwaddu.wv v28, v12, v16
vsetvli zero, zero, e16, m1, tu, ma
.ifc \type,abs
vabsaddu v24, v28, v12
.endif
.ifc \type,square
vwmacc.vv v24, v28, v28
.endif
add a1, a1, a3
add a2, a2, a3
add t1, t1, a3
add t2, t2, a3
bnez a4, 1b
vsetvli zero, zero, e32, m2, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
.endm
.macro vsad_vsse_intra16 type
vsetivli t0, 16, e32, m4, ta, ma
addi a4, a4, -1
add t1, a1, a3
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, m1, tu, ma
vle8.v v4, (a1)
vle8.v v12, (t1)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m2, tu, ma
.ifc \type,abs
vabsaddu v24, v16, v12
.endif
.ifc \type,square
vwmacc.vv v24, v16, v16
.endif
add a1, a1, a3
add t1, t1, a3
bnez a4, 1b
vsetvli zero, zero, e32, m4, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
.endm
.macro vsad_vsse_intra8 type
vsetivli t0, 8, e32, m2, ta, ma
addi a4, a4, -1
add t1, a1, a3
vmv.v.x v24, zero
vmv.s.x v0, zero
1:
vsetvli zero, zero, e8, mf2, tu, ma
vle8.v v4, (a1)
vle8.v v12, (t1)
addi a4, a4, -1
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m1, tu, ma
.ifc \type,abs
vabsaddu v24, v16, v12
.endif
.ifc \type,square
vwmacc.vv v24, v16, v16
.endif
add a1, a1, a3
add t1, t1, a3
bnez a4, 1b
vsetvli zero, zero, e32, m2, tu, ma
vredsum.vs v0, v24, v0
vmv.x.s a0, v0
ret
.endm
func ff_vsse16_rvv, zve32x
vsad_vsse16 square
endfunc
func ff_vsse8_rvv, zve32x
vsad_vsse8 square
endfunc
func ff_vsse_intra16_rvv, zve32x
vsad_vsse_intra16 square
endfunc
func ff_vsse_intra8_rvv, zve32x
vsad_vsse_intra8 square
endfunc
func ff_vsad16_rvv, zve32x
vsad_vsse16 abs
endfunc
func ff_vsad8_rvv, zve32x
vsad_vsse8 abs
endfunc
func ff_vsad_intra16_rvv, zve32x
vsad_vsse_intra16 abs
endfunc
func ff_vsad_intra8_rvv, zve32x
vsad_vsse_intra8 abs
endfunc
func ff_nsse16_rvv, zve32x
.macro squarediff16
vsetivli zero, 16, e8, m1, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m2, tu, ma
vwmacc.vv v24, v16, v16
.endm
.macro gradiff16 srcx srcv
vsetivli zero, 16, e8, m1, tu, ma
vle8.v v8, (\srcx)
vslide1down.vx v0, \srcv, t5
vslide1down.vx v16, v8, t5
vwsubu.vv v20, \srcv, v0
vwsubu.wv v0, v20, v8
vwaddu.wv v20, v0, v16
vsetivli zero, 15, e16, m2, tu, ma
vneg.v v0, v20
vmax.vv v0, v20, v0
.endm
csrwi vxrm, 0
vsetivli t0, 16, e32, m4, ta, ma
addi a4, a4, -1
li t5, 1
vmv.v.x v24, zero
vmv.v.x v28, zero
1:
add t1, a1, a3
add t2, a2, a3
addi a4, a4, -1
squarediff16
gradiff16 t1, v4
vwaddu.wv v28, v28, v0
gradiff16 t2, v12
vwsubu.wv v28, v28, v0
add a1, a1, a3
add a2, a2, a3
bnez a4, 1b
squarediff16
vsetivli zero, 16, e32, m4, tu, ma
vmv.s.x v0, zero
vmv.s.x v4, zero
vredsum.vs v0, v24, v0
vredsum.vs v4, v28, v4
vmv.x.s t1, v0
vmv.x.s t2, v4
srai t3, t2, 31
xor t2, t3, t2
sub t2, t2, t3
mul t2, t2, a0
add a0, t2, t1
ret
endfunc
func ff_nsse8_rvv, zve32x
.macro squarediff8
vsetivli zero, 8, e8, mf2, tu, ma
vle8.v v4, (a1)
vle8.v v12, (a2)
vwsubu.vv v16, v4, v12
vsetvli zero, zero, e16, m1, tu, ma
vwmacc.vv v24, v16, v16
.endm
.macro gradiff8 srcx srcv
vsetivli zero, 8, e8, mf2, tu, ma
vle8.v v8, (\srcx)
vslide1down.vx v0, \srcv, t5
vslide1down.vx v16, v8, t5
vwsubu.vv v20, \srcv, v0
vwsubu.wv v0, v20, v8
vwaddu.wv v20, v0, v16
vsetivli zero, 7, e16, m1, tu, ma
vneg.v v0, v20
vmax.vv v0, v20, v0
.endm
csrwi vxrm, 0
vsetivli t0, 8, e32, m2, ta, ma
addi a4, a4, -1
li t5, 1
vmv.v.x v24, zero
vmv.v.x v28, zero
1:
add t1, a1, a3
add t2, a2, a3
addi a4, a4, -1
squarediff8
gradiff8 t1, v4
vwaddu.wv v28, v28, v0
gradiff8 t2, v12
vwsubu.wv v28, v28, v0
add a1, a1, a3
add a2, a2, a3
bnez a4, 1b
squarediff8
vsetivli zero, 8, e32, m2, tu, ma
vmv.s.x v0, zero
vmv.s.x v4, zero
vredsum.vs v0, v24, v0
vredsum.vs v4, v28, v4
vmv.x.s t1, v0
vmv.x.s t2, v4
srai t3, t2, 31
xor t2, t3, t2
sub t2, t2, t3
mul t2, t2, a0
add a0, t2, t1
ret
endfunc