From: sunyuechi <sunyue...@iscas.ac.cn> C908 X60 vp8_loop_filter8uv_inner_v_c : 12.5 11.0 vp8_loop_filter8uv_inner_v_rvv_i32 : 7.7 6.2 vp8_loop_filter16y_inner_h_c : 11.7 10.2 vp8_loop_filter16y_inner_h_rvv_i32 : 8.5 6.5 vp8_loop_filter16y_inner_v_c : 11.5 10.7 vp8_loop_filter16y_inner_v_rvv_i32 : 5.0 3.5 --- libavcodec/riscv/vp8dsp_init.c | 5 +- libavcodec/riscv/vp8dsp_rvv.S | 104 +++++++++++++++++++++++++++++++++ 2 files changed, 108 insertions(+), 1 deletion(-)
diff --git a/libavcodec/riscv/vp8dsp_init.c b/libavcodec/riscv/vp8dsp_init.c index 8c5b2c8b04..94f78cd84b 100644 --- a/libavcodec/riscv/vp8dsp_init.c +++ b/libavcodec/riscv/vp8dsp_init.c @@ -154,7 +154,10 @@ av_cold void ff_vp8dsp_init_riscv(VP8DSPContext *c) #define init_loop_filter(vlen) \ c->vp8_v_loop_filter_simple = ff_vp8_v_loop_filter16_simple_rvv##vlen; \ - c->vp8_h_loop_filter_simple = ff_vp8_h_loop_filter16_simple_rvv##vlen; + c->vp8_h_loop_filter_simple = ff_vp8_h_loop_filter16_simple_rvv##vlen; \ + c->vp8_v_loop_filter16y_inner = ff_vp8_v_loop_filter16_inner_rvv##vlen; \ + c->vp8_h_loop_filter16y_inner = ff_vp8_h_loop_filter16_inner_rvv##vlen; \ + c->vp8_v_loop_filter8uv_inner = ff_vp8_v_loop_filter8uv_inner_rvv##vlen; int flags = av_get_cpu_flags(); diff --git a/libavcodec/riscv/vp8dsp_rvv.S b/libavcodec/riscv/vp8dsp_rvv.S index b5f8bb31b4..ed789ec4fd 100644 --- a/libavcodec/riscv/vp8dsp_rvv.S +++ b/libavcodec/riscv/vp8dsp_rvv.S @@ -275,6 +275,13 @@ func ff_vp78_idct_dc_add4uv_rvv, zve64x ret endfunc +.macro filter_abs dst, diff, fI + vneg.v v8, \diff + vmax.vv \dst, v8, \diff + vmsleu.vx v8, \dst, \fI + vmand.mm v27, v27, v8 +.endm + .macro filter_fmin len, vlen, a, f1, p0f2, q0f1 vsetvlstatic16 \len, \vlen vsext.vf2 \q0f1, \a @@ -300,6 +307,16 @@ endfunc vle8.v v11, (t4) vle8.v v17, (t1) vle8.v v22, (\dst) + .if \normal + sub t3, t2, a6 + sub t0, t1, a6 + add t6, \dst, a6 + add a7, t4, a6 + vle8.v v2, (t3) + vle8.v v15, (t0) + vle8.v v10, (t6) + vle8.v v14, (a7) + .endif .else addi t1, \dst, -1 addi a6, \dst, -2 @@ -308,9 +325,27 @@ endfunc vlse8.v v11, (t4), \stride vlse8.v v17, (t1), \stride vlse8.v v22, (\dst), \stride + .if \normal + addi t5, \dst, -4 + addi t0, \dst, -3 + addi t6, \dst, 2 + addi a7, \dst, 3 + vlse8.v v2, (t5), \stride + vlse8.v v15, (t0), \stride + vlse8.v v10, (t6), \stride + vlse8.v v14, (a7), \stride + .endif .endif vwsubu.vv v12, v1, v11 // p1-q1 vwsubu.vv v24, v22, v17 // q0-p0 +.if \normal + vwsubu.vv v30, v1, v17 + vwsubu.vv v20, v11, v22 + vwsubu.vv v28, v1, v15 + vwsubu.vv v4, v2, v15 + vwsubu.vv v6, v10, v11 + vwsubu.vv v2, v14, v10 +.endif vnclip.wi v23, v12, 0 vsetvlstatic16 \len, \vlen // vp8_simple_limit(dst + i, stride, flim) @@ -322,6 +357,25 @@ endfunc vsrl.vi v18, v18, 1 vmacc.vx v18, a7, v8 vmsleu.vx v0, v18, \fE +.if \normal + vneg.v v18, v30 + vmax.vv v30, v18, v30 + vmsleu.vx v27, v30, \fI + filter_abs v18, v28, \fI + filter_abs v18, v4, \fI + filter_abs v18, v6, \fI + filter_abs v18, v2, \fI + filter_abs v20, v20, \fI + vmand.mm v27, v0, v27 // vp8_simple_limit && normal + + vmsgtu.vx v20, v20, \thresh // hev + vmsgtu.vx v3, v30, \thresh + vmor.mm v3, v3, v20 // v3 = hev: > thresh + vzext.vf2 v18, v1 // v18 = p1 + vmand.mm v0, v27, v3 // v0 = normal && hev + vzext.vf2 v20, v11 // v12 = q1 + vmnot.m v3, v3 // v3 = !hv +.endif li t5, 3 li a7, 124 @@ -346,6 +400,37 @@ endfunc vsse8.v v6, (\dst), \stride, v0.t .endif +.if \normal + vmand.mm v0, v27, v3 // vp8_normal_limit & !hv + + .if \inner + vnclip.wi v30, v30, 0 + filter_fmin \len, \vlen, v30, v24, v4, v6 + vadd.vi v24, v24, 1 + vsra.vi v24, v24, 1 // (f1 + 1) >> 1; + vadd.vv v8, v18, v24 + vsub.vv v10, v20, v24 + .endif + + vmax.vx v8, v8, zero + vmax.vx v10, v10, zero + vsetvlstatic8 \len, \vlen + vnclipu.wi v4, v4, 0 + vnclipu.wi v5, v6, 0 + vnclipu.wi v6, v8, 0 + vnclipu.wi v7, v10, 0 + .ifc \type,v + vse8.v v4, (t1), v0.t + vse8.v v5, (\dst), v0.t + vse8.v v6, (t2), v0.t + vse8.v v7, (t4), v0.t + .else + vsse8.v v4, (t1), \stride, v0.t + vsse8.v v5, (\dst), \stride, v0.t + vsse8.v v6, (a6), \stride, v0.t + vsse8.v v7, (t4), \stride, v0.t + .endif +.endif .endm .irp vlen,256,128 @@ -360,6 +445,25 @@ func ff_vp8_h_loop_filter16_simple_rvv\vlen, zve32x filter 16, \vlen, h, 0, 0, a0, a1, a2, a3, a4 ret endfunc + +func ff_vp8_h_loop_filter16_inner_rvv\vlen, zve32x + vsetvlstatic8 16, \vlen + filter 16, \vlen, h, 1, 1, a0, a1, a2, a3, a4 + ret +endfunc + +func ff_vp8_v_loop_filter16_inner_rvv\vlen, zve32x + vsetvlstatic8 16, \vlen + filter 16, \vlen, v, 1, 1, a0, a1, a2, a3, a4 + ret +endfunc + +func ff_vp8_v_loop_filter8uv_inner_rvv\vlen, zve32x + vsetvlstatic8 8, \vlen + filter 8, \vlen, v, 1, 1, a0, a2, a3, a4, a5 + filter 8, \vlen, v, 1, 1, a1, a2, a3, a4, a5 + ret +endfunc .endr .macro bilin_load_h dst mn -- 2.45.2 _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".