This patch includes: - VSRLRN.{B.H/H.W/W.D}; - VSRARN.{B.H/H.W/W.D}; - VSRLRNI.{B.H/H.W/W.D/D.Q}; - VSRARNI.{B.H/H.W/W.D/D.Q}.
Signed-off-by: Song Gao <gaos...@loongson.cn> --- target/loongarch/disas.c | 16 +++ target/loongarch/helper.h | 16 +++ target/loongarch/insn_trans/trans_lsx.c.inc | 16 +++ target/loongarch/insns.decode | 16 +++ target/loongarch/lsx_helper.c | 108 ++++++++++++++++++++ 5 files changed, 172 insertions(+) diff --git a/target/loongarch/disas.c b/target/loongarch/disas.c index e6f4411b43..507f34feaa 100644 --- a/target/loongarch/disas.c +++ b/target/loongarch/disas.c @@ -1153,3 +1153,19 @@ INSN_LSX(vsrani_b_h, vv_i) INSN_LSX(vsrani_h_w, vv_i) INSN_LSX(vsrani_w_d, vv_i) INSN_LSX(vsrani_d_q, vv_i) + +INSN_LSX(vsrlrn_b_h, vvv) +INSN_LSX(vsrlrn_h_w, vvv) +INSN_LSX(vsrlrn_w_d, vvv) +INSN_LSX(vsrarn_b_h, vvv) +INSN_LSX(vsrarn_h_w, vvv) +INSN_LSX(vsrarn_w_d, vvv) + +INSN_LSX(vsrlrni_b_h, vv_i) +INSN_LSX(vsrlrni_h_w, vv_i) +INSN_LSX(vsrlrni_w_d, vv_i) +INSN_LSX(vsrlrni_d_q, vv_i) +INSN_LSX(vsrarni_b_h, vv_i) +INSN_LSX(vsrarni_h_w, vv_i) +INSN_LSX(vsrarni_w_d, vv_i) +INSN_LSX(vsrarni_d_q, vv_i) diff --git a/target/loongarch/helper.h b/target/loongarch/helper.h index eccfbfbb3e..bb868961d1 100644 --- a/target/loongarch/helper.h +++ b/target/loongarch/helper.h @@ -508,3 +508,19 @@ DEF_HELPER_4(vsrani_b_h, void, env, i32, i32, i32) DEF_HELPER_4(vsrani_h_w, void, env, i32, i32, i32) DEF_HELPER_4(vsrani_w_d, void, env, i32, i32, i32) DEF_HELPER_4(vsrani_d_q, void, env, i32, i32, i32) + +DEF_HELPER_4(vsrlrn_b_h, void, env, i32, i32, i32) +DEF_HELPER_4(vsrlrn_h_w, void, env, i32, i32, i32) +DEF_HELPER_4(vsrlrn_w_d, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarn_b_h, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarn_h_w, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarn_w_d, void, env, i32, i32, i32) + +DEF_HELPER_4(vsrlrni_b_h, void, env, i32, i32, i32) +DEF_HELPER_4(vsrlrni_h_w, void, env, i32, i32, i32) +DEF_HELPER_4(vsrlrni_w_d, void, env, i32, i32, i32) +DEF_HELPER_4(vsrlrni_d_q, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarni_b_h, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarni_h_w, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarni_w_d, void, env, i32, i32, i32) +DEF_HELPER_4(vsrarni_d_q, void, env, i32, i32, i32) diff --git a/target/loongarch/insn_trans/trans_lsx.c.inc b/target/loongarch/insn_trans/trans_lsx.c.inc index 5b4410852d..d3ab0a4a6a 100644 --- a/target/loongarch/insn_trans/trans_lsx.c.inc +++ b/target/loongarch/insn_trans/trans_lsx.c.inc @@ -424,3 +424,19 @@ TRANS(vsrani_b_h, gen_vv_i, gen_helper_vsrani_b_h) TRANS(vsrani_h_w, gen_vv_i, gen_helper_vsrani_h_w) TRANS(vsrani_w_d, gen_vv_i, gen_helper_vsrani_w_d) TRANS(vsrani_d_q, gen_vv_i, gen_helper_vsrani_d_q) + +TRANS(vsrlrn_b_h, gen_vvv, gen_helper_vsrlrn_b_h) +TRANS(vsrlrn_h_w, gen_vvv, gen_helper_vsrlrn_h_w) +TRANS(vsrlrn_w_d, gen_vvv, gen_helper_vsrlrn_w_d) +TRANS(vsrarn_b_h, gen_vvv, gen_helper_vsrarn_b_h) +TRANS(vsrarn_h_w, gen_vvv, gen_helper_vsrarn_h_w) +TRANS(vsrarn_w_d, gen_vvv, gen_helper_vsrarn_w_d) + +TRANS(vsrlrni_b_h, gen_vv_i, gen_helper_vsrlrni_b_h) +TRANS(vsrlrni_h_w, gen_vv_i, gen_helper_vsrlrni_h_w) +TRANS(vsrlrni_w_d, gen_vv_i, gen_helper_vsrlrni_w_d) +TRANS(vsrlrni_d_q, gen_vv_i, gen_helper_vsrlrni_d_q) +TRANS(vsrarni_b_h, gen_vv_i, gen_helper_vsrarni_b_h) +TRANS(vsrarni_h_w, gen_vv_i, gen_helper_vsrarni_h_w) +TRANS(vsrarni_w_d, gen_vv_i, gen_helper_vsrarni_w_d) +TRANS(vsrarni_d_q, gen_vv_i, gen_helper_vsrarni_d_q) diff --git a/target/loongarch/insns.decode b/target/loongarch/insns.decode index 859def6752..0b30175f6b 100644 --- a/target/loongarch/insns.decode +++ b/target/loongarch/insns.decode @@ -882,3 +882,19 @@ vsrani_b_h 0111 00110101 10000 1 .... ..... ..... @vv_ui4 vsrani_h_w 0111 00110101 10001 ..... ..... ..... @vv_ui5 vsrani_w_d 0111 00110101 1001 ...... ..... ..... @vv_ui6 vsrani_d_q 0111 00110101 101 ....... ..... ..... @vv_ui7 + +vsrlrn_b_h 0111 00001111 10001 ..... ..... ..... @vvv +vsrlrn_h_w 0111 00001111 10010 ..... ..... ..... @vvv +vsrlrn_w_d 0111 00001111 10011 ..... ..... ..... @vvv +vsrarn_b_h 0111 00001111 10101 ..... ..... ..... @vvv +vsrarn_h_w 0111 00001111 10110 ..... ..... ..... @vvv +vsrarn_w_d 0111 00001111 10111 ..... ..... ..... @vvv + +vsrlrni_b_h 0111 00110100 01000 1 .... ..... ..... @vv_ui4 +vsrlrni_h_w 0111 00110100 01001 ..... ..... ..... @vv_ui5 +vsrlrni_w_d 0111 00110100 0101 ...... ..... ..... @vv_ui6 +vsrlrni_d_q 0111 00110100 011 ....... ..... ..... @vv_ui7 +vsrarni_b_h 0111 00110101 11000 1 .... ..... ..... @vv_ui4 +vsrarni_h_w 0111 00110101 11001 ..... ..... ..... @vv_ui5 +vsrarni_w_d 0111 00110101 1101 ...... ..... ..... @vv_ui6 +vsrarni_d_q 0111 00110101 111 ....... ..... ..... @vv_ui7 diff --git a/target/loongarch/lsx_helper.c b/target/loongarch/lsx_helper.c index 30b8da837a..8ccfa75fe3 100644 --- a/target/loongarch/lsx_helper.c +++ b/target/loongarch/lsx_helper.c @@ -2516,3 +2516,111 @@ DO_HELPER_VV_I(vsrani_b_h, 16, helper_vv_ni_c, do_vsrani) DO_HELPER_VV_I(vsrani_h_w, 32, helper_vv_ni_c, do_vsrani) DO_HELPER_VV_I(vsrani_w_d, 64, helper_vv_ni_c, do_vsrani) DO_HELPER_VV_I(vsrani_d_q, 128, helper_vv_ni_c, do_vsrani) + +static void do_vsrlrn(vec_t *Vd, vec_t *Vj, vec_t *Vk, int bit, int n) +{ + switch (bit) { + case 16: + Vd->B[n] = vsrlr((uint16_t)Vj->H[n], Vk->H[n], bit); + break; + case 32: + Vd->H[n] = vsrlr((uint32_t)Vj->W[n], Vk->W[n], bit); + break; + case 64: + Vd->W[n] = vsrlr((uint64_t)Vj->D[n], Vk->D[n], bit); + break; + default: + g_assert_not_reached(); + } +} + +static void do_vsrarn(vec_t *Vd, vec_t *Vj, vec_t *Vk, int bit, int n) +{ + switch (bit) { + case 16: + Vd->B[n] = vsrar(Vj->H[n], Vk->H[n], bit); + break; + case 32: + Vd->H[n] = vsrar(Vj->W[n], Vk->W[n], bit); + break; + case 64: + Vd->W[n] = vsrar(Vj->D[n], Vk->D[n], bit); + break; + default: + g_assert_not_reached(); + } +} + +DO_HELPER_VVV(vsrlrn_b_h, 16, helper_vvv_hz, do_vsrlrn) +DO_HELPER_VVV(vsrlrn_h_w, 32, helper_vvv_hz, do_vsrlrn) +DO_HELPER_VVV(vsrlrn_w_d, 64, helper_vvv_hz, do_vsrlrn) +DO_HELPER_VVV(vsrarn_b_h, 16, helper_vvv_hz, do_vsrarn) +DO_HELPER_VVV(vsrarn_h_w, 32, helper_vvv_hz, do_vsrarn) +DO_HELPER_VVV(vsrarn_w_d, 64, helper_vvv_hz, do_vsrarn) + +static __int128_t vsrlrn(__int128_t s1, uint32_t imm) +{ + if (imm == 0) { + return s1; + } else { + __uint128_t t1 = (__uint128_t)1 << (imm -1); + return (s1 + t1) >> imm; + } +} + +static void do_vsrlrni(vec_t *dest, vec_t *Vd, vec_t *Vj, uint32_t imm, int bit, int n) +{ + switch (bit) { + case 16: + dest->B[n] = vsrlrn((uint16_t)Vj->H[n], imm); + dest->B[n + 128 / bit] = vsrlrn((uint16_t)Vd->H[n], imm); + break; + case 32: + dest->H[n] = vsrlrn((uint32_t)Vj->W[n], imm); + dest->H[n + 128 / bit] = vsrlrn((uint32_t)Vd->W[n], imm); + break; + case 64: + dest->W[n] = vsrlrn((uint64_t)Vj->D[n], imm); + dest->W[n + 128 / bit] = vsrlrn((uint64_t)Vd->D[n], imm); + break; + case 128: + dest->D[n] = vsrlrn((__uint128_t)Vj->Q[n], imm); + dest->D[n + 128 / bit] = vsrlrn((__uint128_t)Vd->Q[n], imm); + break; + default: + g_assert_not_reached(); + } +} + +static void do_vsrarni(vec_t *dest, vec_t *Vd, vec_t *Vj, uint32_t imm, int bit, int n) +{ + switch (bit) { + case 16: + dest->B[n] = vsrlrn(Vj->H[n], imm); + dest->B[n + 128 / bit] = vsrlrn(Vd->H[n], imm); + break; + case 32: + dest->H[n] = vsrlrn(Vj->W[n], imm); + dest->H[n + 128 / bit] = vsrlrn(Vd->W[n], imm); + break; + case 64: + dest->W[n] = vsrlrn(Vj->D[n], imm); + dest->W[n + 128 / bit] = vsrlrn(Vd->D[n], imm); + break; + case 128: + dest->D[n] = vsrlrn(Vj->Q[n], imm); + dest->D[n + 128 / bit] = vsrlrn(Vd->Q[n], imm); + break; + default: + g_assert_not_reached(); + } +} + +DO_HELPER_VV_I(vsrlrni_b_h, 16, helper_vv_ni_c, do_vsrlrni) +DO_HELPER_VV_I(vsrlrni_h_w, 32, helper_vv_ni_c, do_vsrlrni) +DO_HELPER_VV_I(vsrlrni_w_d, 64, helper_vv_ni_c, do_vsrlrni) +DO_HELPER_VV_I(vsrlrni_d_q, 128, helper_vv_ni_c, do_vsrlrni) +DO_HELPER_VV_I(vsrarni_b_h, 16, helper_vv_ni_c, do_vsrarni) +DO_HELPER_VV_I(vsrarni_h_w, 32, helper_vv_ni_c, do_vsrarni) +DO_HELPER_VV_I(vsrarni_w_d, 64, helper_vv_ni_c, do_vsrarni) +DO_HELPER_VV_I(vsrarni_d_q, 128, helper_vv_ni_c, do_vsrarni) -- 2.31.1