diff mbox series

[FFmpeg-devel,PATCHv2,2/4] lavc/vc1dsp: R-V V vc1_inv_trans_8x4

Message ID 20240604185908.22309-2-remi@remlab.net
State Accepted
Commit a169f3bca5937cf71d61de1fb85b522bd69c6cfc
Headers show
Series [FFmpeg-devel,PATCHv2,1/4] lavc/vc1dsp: R-V V vc1_inv_trans_8x8 | expand

Checks

Context Check Description
yinshiyou/make_loongarch64 success Make finished
yinshiyou/make_fate_loongarch64 success Make fate finished
andriy/make_x86 success Make finished
andriy/make_fate_x86 success Make fate finished

Commit Message

Rémi Denis-Courmont June 4, 2024, 6:59 p.m. UTC
T-Head C908 (cycles):
vc1dsp.vc1_inv_trans_8x4_c:       626.2
vc1dsp.vc1_inv_trans_8x4_rvv_i32: 215.2

---
Changes since version 1:
- Properly set VXRM (callee-clobbered).

---
 libavcodec/riscv/vc1dsp_init.c |  2 +
 libavcodec/riscv/vc1dsp_rvv.S  | 73 ++++++++++++++++++++++++++++++++++
 2 files changed, 75 insertions(+)
diff mbox series

Patch

diff --git a/libavcodec/riscv/vc1dsp_init.c b/libavcodec/riscv/vc1dsp_init.c
index b8a1015ce5..e63870ad44 100644
--- a/libavcodec/riscv/vc1dsp_init.c
+++ b/libavcodec/riscv/vc1dsp_init.c
@@ -29,6 +29,7 @@  void ff_vc1_inv_trans_8x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block
 void ff_vc1_inv_trans_8x8_rvv(int16_t block[64]);
 void ff_vc1_inv_trans_4x8_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
 void ff_vc1_inv_trans_8x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
+void ff_vc1_inv_trans_8x4_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
 void ff_vc1_inv_trans_4x4_dc_rvv(uint8_t *dest, ptrdiff_t stride, int16_t *block);
 void ff_put_pixels16x16_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd);
 void ff_put_pixels8x8_rvi(uint8_t *dst, const uint8_t *src, ptrdiff_t line_size, int rnd);
@@ -55,6 +56,7 @@  av_cold void ff_vc1dsp_init_riscv(VC1DSPContext *dsp)
     if (flags & AV_CPU_FLAG_RVV_I32) {
         if (ff_rv_vlen_least(128)) {
             dsp->vc1_inv_trans_8x8 = ff_vc1_inv_trans_8x8_rvv;
+            dsp->vc1_inv_trans_8x4 = ff_vc1_inv_trans_8x4_rvv;
             dsp->vc1_inv_trans_4x8_dc = ff_vc1_inv_trans_4x8_dc_rvv;
             dsp->vc1_inv_trans_4x4_dc = ff_vc1_inv_trans_4x4_dc_rvv;
             dsp->avg_vc1_mspel_pixels_tab[0][0] = ff_avg_pixels16x16_rvv;
diff --git a/libavcodec/riscv/vc1dsp_rvv.S b/libavcodec/riscv/vc1dsp_rvv.S
index e15783d113..d003185ade 100644
--- a/libavcodec/riscv/vc1dsp_rvv.S
+++ b/libavcodec/riscv/vc1dsp_rvv.S
@@ -173,6 +173,31 @@  func ff_vc1_inv_trans_8_rvv, zve32x
         jr      t0
 endfunc
 
+        .variant_cc ff_vc1_inv_trans_4_rvv
+func ff_vc1_inv_trans_4_rvv, zve32x
+        li      t3, 17
+        vmul.vx v8, v0, t3
+        li      t4, 22
+        vmul.vx v10, v2, t3
+        li      t2, 10
+        vmul.vx v14, v1, t4
+        vadd.vv v24, v8, v10  # t1
+        vsub.vv v25, v8, v10  # t2
+        vmul.vx v16, v3, t2
+        vmul.vx v18, v3, t4
+        vmul.vx v20, v1, t2
+        vadd.vv v26, v14, v16 # t3
+        vsub.vv v27, v18, v20 # t4
+        vadd.vv v0, v24, v26
+        vsub.vv v1, v25, v27
+        vadd.vv v2, v25, v27
+        vsub.vv v3, v24, v26
+        .irp    n,0,1,2,3
+        vssra.vx v\n, v\n, t1 # + 4 >> 3 or + 64 >> 7
+        .endr
+        jr      t0
+endfunc
+
 func ff_vc1_inv_trans_8x8_rvv, zve32x
         csrwi    vxrm, 0
         vsetivli zero, 8, e16, m1, ta, ma
@@ -223,6 +248,54 @@  func ff_vc1_inv_trans_8x8_rvv, zve32x
         ret
 endfunc
 
+func ff_vc1_inv_trans_8x4_rvv, zve32x
+        csrwi       vxrm, 0
+        vsetivli    zero, 4, e16, mf2, ta, ma
+        vlseg8e16.v v0, (a2)
+        jal         t0, ff_vc1_inv_trans_8_rvv
+        vsseg8e16.v v0, (a2)
+        addi        a3, a2, 1 * 8 * 2
+        vsetivli    zero, 8, e16, m1, ta, ma
+        vle16.v     v0, (a2)
+        addi        a4, a2, 2 * 8 * 2
+        vle16.v     v1, (a3)
+        addi        a5, a2, 3 * 8 * 2
+        vle16.v     v2, (a4)
+        vle16.v     v3, (a5)
+        .irp    n,0,1,2,3
+        # shift 4 vectors of 8 elems after transpose instead of 8 of 4
+        vssra.vi    v\n, v\n, 3
+        .endr
+        li          t1, 7
+        jal         t0, ff_vc1_inv_trans_4_rvv
+        add         a3, a1, a0
+        vle8.v      v8, (a0)
+        add         a4, a1, a3
+        vle8.v      v9, (a3)
+        add         a5, a1, a4
+        vle8.v      v10, (a4)
+        vle8.v      v11, (a5)
+        vsetvli     zero, zero, e8, mf2, ta, ma
+        vwaddu.wv   v0, v0, v8
+        vwaddu.wv   v1, v1, v9
+        vwaddu.wv   v2, v2, v10
+        vwaddu.wv   v3, v3, v11
+        vsetvli     zero, zero, e16, m1, ta, ma
+        .irp    n,0,1,2,3
+        vmax.vx     v\n, v\n, zero
+        .endr
+        vsetvli     zero, zero, e8, mf2, ta, ma
+        vnclipu.wi  v8, v0, 0
+        vnclipu.wi  v9, v1, 0
+        vse8.v      v8, (a0)
+        vnclipu.wi  v10, v2, 0
+        vse8.v      v9, (a3)
+        vnclipu.wi  v11, v3, 0
+        vse8.v      v10, (a4)
+        vse8.v      v11, (a5)
+        ret
+endfunc
+
 .macro mspel_op op pos n1 n2
         add           t1, \pos, a2
         v\op\()e8.v   v\n1, (\pos)