Le 6 décembre 2024 13:19:29 GMT+01:00, Niklas Haas <ffm...@haasn.xyz> a écrit : >On Thu, 05 Dec 2024 21:41:51 +0800 uk7b-at-foxmail....@ffmpeg.org wrote: >> From: sunyuechi <sunyue...@iscas.ac.cn> >> >> --- >> libavcodec/riscv/rv40dsp_rvv.S | 116 ++++++++++++++++++++++----------- >> 1 file changed, 78 insertions(+), 38 deletions(-) >> >> diff --git a/libavcodec/riscv/rv40dsp_rvv.S b/libavcodec/riscv/rv40dsp_rvv.S >> index ca431eb8ab..d4beb7f1e9 100644 >> --- a/libavcodec/riscv/rv40dsp_rvv.S >> +++ b/libavcodec/riscv/rv40dsp_rvv.S >> @@ -20,15 +20,30 @@ >> >> #include "libavutil/riscv/asm.S" >> >> -.macro manual_avg dst src1 src2 >> - vadd.vv \dst, \src1, \src2 >> - vadd.vi \dst, \dst, 1 >> - vsrl.vi \dst, \dst, 1 >> -.endm >> +const rv40_bias >> + .byte 0, 16, 32, 16 >> + .byte 32, 28, 32, 28 >> + .byte 0, 32, 16, 32 >> + .byte 32, 28, 32, 28 >> +endconst >> >> .macro do_chroma_mc type unroll >> - csrwi vxrm, 2 >> + csrwi vxrm, 0 >> + addi sp, sp, -16 >> +#if __riscv_xlen == 32 >> + sw s2, (sp) >> +#elif __riscv_xlen == 64 >> + sd s2, (sp) >> +#else >> + sq s2, (sp) >> +#endif > >You can use the macro `sx` from libavcodec/riscv/h264qpel_rvv.S > >(I suggest to move them into some common include path)
As we need to backport this, should we merge this to stable and the sx/lx version to master? Or should we merge this to master so that we can cherry-pick the exact patch to master? > >> + lla t4, rv40_bias >> + srli t5, a5, 1 >> + sh2add t4, t5, t4 >> + srli t5, a4, 1 >> + add t5, t4, t5 >> slli t2, a5, 3 >> + lb s2, (t5) >> mul t1, a5, a4 >> sh3add a5, a4, t2 >> slli a4, a4, 3 >> @@ -80,17 +95,19 @@ >> vwmulu.vx v12, v14, a6 >> vwmaccu.vx v10, t1, v15 >> vwmaccu.vx v12, a7, v15 >> - vnclipu.wi v15, v8, 6 >> + vwaddu.wx v20, v8, s2 >> + vnsrl.wi v15, v20, 6 >> .ifc \type,avg >> vle8.v v9, (a0) >> - manual_avg v15, v15, v9 >> + vaaddu.vv v15, v15, v9 >> .endif >> vse8.v v15, (a0) >> add a0, a0, a2 >> - vnclipu.wi v8, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v9, (a0) >> - manual_avg v8, v8, v9 >> + vaaddu.vv v8, v8, v9 >> .endif >> add t4, t4, t3 >> vse8.v v8, (a0) >> @@ -115,17 +132,19 @@ >> vslide1down.vx v14, v14, t5 >> vsetvli zero, t6, e8, m1, ta, ma >> vwmaccu.vx v16, t1, v14 >> - vnclipu.wi v8, v12, 6 >> + vwaddu.wx v20, v12, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v9, (a0) >> - manual_avg v8, v8, v9 >> + vaaddu.vv v8, v8, v9 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> - vnclipu.wi v8, v16, 6 >> + vwaddu.wx v20, v16, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v9, (a0) >> - manual_avg v8, v8, v9 >> + vaaddu.vv v8, v8, v9 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> @@ -159,18 +178,20 @@ >> vwmaccu.vx v10, t0, v8 >> add a4, a4, a7 >> vwmaccu.vx v12, t0, v9 >> - vnclipu.wi v15, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v15, v20, 6 >> vwmulu.vx v10, v9, a6 >> - vnclipu.wi v9, v12, 6 >> + vwaddu.wx v20, v12, s2 >> + vnsrl.wi v9, v20, 6 >> .ifc \type,avg >> vle8.v v16, (a0) >> - manual_avg v15, v15, v16 >> + vaaddu.vv v15, v15, v16 >> .endif >> vse8.v v15, (a0) >> add a0, a0, a2 >> .ifc \type,avg >> vle8.v v16, (a0) >> - manual_avg v9, v9, v16 >> + vaaddu.vv v9, v9, v16 >> .endif >> vse8.v v9, (a0) >> add a0, a0, a2 >> @@ -179,18 +200,20 @@ >> vle8.v v14, (a5) >> vwmaccu.vx v10, t0, v8 >> vwmulu.vx v12, v8, a6 >> - vnclipu.wi v8, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v8, v20, 6 >> vwmaccu.vx v12, t0, v14 >> .ifc \type,avg >> vle8.v v16, (a0) >> - manual_avg v8, v8, v16 >> + vaaddu.vv v8, v8, v16 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> - vnclipu.wi v8, v12, 6 >> + vwaddu.wx v20, v12, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v16, (a0) >> - manual_avg v8, v8, v16 >> + vaaddu.vv v8, v8, v16 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> @@ -226,17 +249,19 @@ >> vsetvli zero, t6, e8, m1, ta, ma >> vwmulu.vx v12, v8, a6 >> vwmaccu.vx v12, a7, v9 >> - vnclipu.wi v16, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v16, v20, 6 >> .ifc \type,avg >> vle8.v v18, (a0) >> - manual_avg v16, v16, v18 >> + vaaddu.vv v16, v16, v18 >> .endif >> vse8.v v16, (a0) >> add a0, a0, a2 >> - vnclipu.wi v10, v12, 6 >> + vwaddu.wx v20, v12, s2 >> + vnsrl.wi v10, v20, 6 >> .ifc \type,avg >> vle8.v v18, (a0) >> - manual_avg v10, v10, v18 >> + vaaddu.vv v10, v10, v18 >> .endif >> add a4, a4, t1 >> vse8.v v10, (a0) >> @@ -254,18 +279,20 @@ >> vslide1down.vx v9, v8, t5 >> vsetvli zero, t6, e8, m1, ta, ma >> vwmulu.vx v12, v8, a6 >> - vnclipu.wi v8, v14, 6 >> + vwaddu.wx v20, v14, s2 >> + vnsrl.wi v8, v20, 6 >> vwmaccu.vx v12, a7, v9 >> .ifc \type,avg >> vle8.v v18, (a0) >> - manual_avg v8, v8, v18 >> + vaaddu.vv v8, v8, v18 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> - vnclipu.wi v8, v12, 6 >> + vwaddu.wx v20, v12, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v18, (a0) >> - manual_avg v8, v8, v18 >> + vaaddu.vv v8, v8, v18 >> .endif >> vse8.v v8, (a0) >> add a0, a0, a2 >> @@ -293,18 +320,20 @@ >> vwmulu.vx v10, v8, a6 >> vle8.v v8, (t0) >> add t0, t1, a2 >> - vnclipu.wi v13, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v13, v20, 6 >> vwmulu.vx v10, v8, a6 >> .ifc \type,avg >> vle8.v v18, (a5) >> - manual_avg v13, v13, v18 >> + vaaddu.vv v13, v13, v18 >> .endif >> vse8.v v13, (a5) >> add a5, a5, a2 >> - vnclipu.wi v8, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v18, (a5) >> - manual_avg v8, v8, v18 >> + vaaddu.vv v8, v8, v18 >> .endif >> vse8.v v8, (a5) >> add a5, a5, a2 >> @@ -312,23 +341,34 @@ >> vle8.v v9, (t1) >> vle8.v v12, (t0) >> vwmulu.vx v10, v9, a6 >> - vnclipu.wi v8, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v8, v20, 6 >> vwmulu.vx v10, v12, a6 >> .ifc \type,avg >> vle8.v v18, (a5) >> - manual_avg v8, v8, v18 >> + vaaddu.vv v8, v8, v18 >> .endif >> vse8.v v8, (a5) >> add a5, a5, a2 >> - vnclipu.wi v8, v10, 6 >> + vwaddu.wx v20, v10, s2 >> + vnsrl.wi v8, v20, 6 >> .ifc \type,avg >> vle8.v v18, (a5) >> - manual_avg v8, v8, v18 >> + vaaddu.vv v8, v8, v18 >> .endif >> vse8.v v8, (a5) >> .endif >> blt t2, a3, 7b >> 8: >> +#if __riscv_xlen == 32 >> + lw s2, (sp) >> +#elif __riscv_xlen == 64 >> + ld s2, (sp) >> +#else >> + lq s2, (sp) >> +#endif > >Ditto for `lx` > >> + addi sp, sp, 16 >> + >> ret >> .endm >> >> -- >> 2.47.1 >> >> _______________________________________________ >> ffmpeg-devel mailing list >> ffmpeg-devel@ffmpeg.org >> https://ffmpeg.org/mailman/listinfo/ffmpeg-devel >> >> To unsubscribe, visit link above, or email >> ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe". >_______________________________________________ >ffmpeg-devel mailing list >ffmpeg-devel@ffmpeg.org >https://ffmpeg.org/mailman/listinfo/ffmpeg-devel > >To unsubscribe, visit link above, or email >ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe". > _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".