1
0
mirror of https://github.com/FFmpeg/FFmpeg.git synced 2025-01-29 22:00:58 +02:00

lavc/vc1dsp: unify R-V V DC bypass functions

This commit is contained in:
Rémi Denis-Courmont 2024-07-27 18:34:58 +03:00
parent bd0c3edb13
commit 784a72a116

View File

@ -21,101 +21,45 @@
#include "libavutil/riscv/asm.S" #include "libavutil/riscv/asm.S"
func ff_vc1_inv_trans_8x8_dc_rvv, zve64x, zba .macro inv_trans_dc rows, cols, w, mat_lmul, row_lmul
func ff_vc1_inv_trans_\cols\()x\rows\()_dc_rvv, zve64x, zba
lpad 0 lpad 0
lh t2, (a2) lh t2, (a2)
vsetivli zero, 8, e8, mf2, ta, ma li a4, 22 - (5 * \cols) / 4
vlse64.v v0, (a0), a1 mul t2, t2, a4
sh1add t2, t2, t2 vsetivli zero, \rows, e8, m\row_lmul, ta, ma
addi t2, t2, 1 vlse\w\().v v0, (a0), a1
srai t2, t2, 1 addi t2, t2, 4
sh1add t2, t2, t2 li a5, 22 - (5 * \rows) / 4
addi t2, t2, 16 srai t2, t2, 3
srai t2, t2, 5 mul t2, t2, a5
li t0, 8*8 .if \cols * \rows >= 32
vsetvli zero, t0, e16, m8, ta, ma li t0, \cols * \rows
vzext.vf2 v8, v0 .endif
vadd.vx v8, v8, t2 addi t2, t2, 64
vmax.vx v8, v8, zero srai t2, t2, 7
vsetvli zero, zero, e8, m4, ta, ma .if \rows * \cols == 64
vnclipu.wi v0, v8, 0 vsetvli zero, t0, e16, m8, ta, ma
vsetivli zero, 8, e8, mf2, ta, ma .elseif \rows * \cols == 32
vsse64.v v0, (a0), a1 vsetvli zero, t0, e16, m4, ta, ma
.else
vsetivli zero, \rows * \cols, e16, m2, ta, ma
.endif
vzext.vf2 v8, v0
vadd.vx v8, v8, t2
vmax.vx v8, v8, zero
vsetvli zero, zero, e8, m\mat_lmul, ta, ma
vnclipu.wi v0, v8, 0
vsetivli zero, \rows, e8, m\row_lmul, ta, ma
vsse\w\().v v0, (a0), a1
ret ret
endfunc endfunc
.endm
func ff_vc1_inv_trans_4x8_dc_rvv, zve32x, zba inv_trans_dc 8, 8, 64, 4, f2
lpad 0 inv_trans_dc 4, 8, 64, 2, f4
lh t2, (a2) inv_trans_dc 8, 4, 32, 2, f2
vsetivli zero, 8, e8, mf2, ta, ma inv_trans_dc 4, 4, 32, 1, f4
vlse32.v v0, (a0), a1
slli t1, t2, 4
add t2, t2, t1
addi t2, t2, 4
srai t2, t2, 3
sh1add t2, t2, t2
slli t2, t2, 2
addi t2, t2, 64
srai t2, t2, 7
li t0, 4*8
vsetvli zero, t0, e16, m4, ta, ma
vzext.vf2 v4, v0
vadd.vx v4, v4, t2
vmax.vx v4, v4, zero
vsetvli zero, zero, e8, m2, ta, ma
vnclipu.wi v0, v4, 0
vsetivli zero, 8, e8, mf2, ta, ma
vsse32.v v0, (a0), a1
ret
endfunc
func ff_vc1_inv_trans_8x4_dc_rvv, zve64x, zba
lpad 0
lh t2, (a2)
vsetivli zero, 4, e8, mf4, ta, ma
vlse64.v v0, (a0), a1
sh1add t2, t2, t2
addi t2, t2, 1
srai t2, t2, 1
slli t1, t2, 4
add t2, t2, t1
addi t2, t2, 64
srai t2, t2, 7
li t0, 8*4
vsetvli zero, t0, e16, m4, ta, ma
vzext.vf2 v4, v0
vadd.vx v4, v4, t2
vmax.vx v4, v4, zero
vsetvli zero, zero, e8, m2, ta, ma
vnclipu.wi v0, v4, 0
vsetivli zero, 4, e8, mf4, ta, ma
vsse64.v v0, (a0), a1
ret
endfunc
func ff_vc1_inv_trans_4x4_dc_rvv, zve32x
lpad 0
lh t2, (a2)
vsetivli zero, 4, e8, mf4, ta, ma
vlse32.v v0, (a0), a1
slli t1, t2, 4
add t2, t2, t1
addi t2, t2, 4
srai t2, t2, 3
slli t1, t2, 4
add t2, t2, t1
addi t2, t2, 64
srai t2, t2, 7
vsetivli zero, 4*4, e16, m2, ta, ma
vzext.vf2 v2, v0
vadd.vx v2, v2, t2
vmax.vx v2, v2, zero
vsetvli zero, zero, e8, m1, ta, ma
vnclipu.wi v0, v2, 0
vsetivli zero, 4, e8, mf4, ta, ma
vsse32.v v0, (a0), a1
ret
endfunc
.variant_cc ff_vc1_inv_trans_8_rvv .variant_cc ff_vc1_inv_trans_8_rvv
func ff_vc1_inv_trans_8_rvv, zve32x func ff_vc1_inv_trans_8_rvv, zve32x