This patch includes: - VBITCLR[I].{B/H/W/D}; - VBITSET[I].{B/H/W/D}; - VBITREV[I].{B/H/W/D}.
Signed-off-by: Song Gao <gaos...@loongson.cn> --- target/loongarch/disas.c | 25 +++ target/loongarch/helper.h | 25 +++ target/loongarch/insn_trans/trans_lsx.c.inc | 25 +++ target/loongarch/insns.decode | 25 +++ target/loongarch/lsx_helper.c | 162 ++++++++++++++++++++ 5 files changed, 262 insertions(+) diff --git a/target/loongarch/disas.c b/target/loongarch/disas.c index e3d4d105fe..7212f86eb0 100644 --- a/target/loongarch/disas.c +++ b/target/loongarch/disas.c @@ -1243,3 +1243,28 @@ INSN_LSX(vpcnt_b, vv) INSN_LSX(vpcnt_h, vv) INSN_LSX(vpcnt_w, vv) INSN_LSX(vpcnt_d, vv) + +INSN_LSX(vbitclr_b, vvv) +INSN_LSX(vbitclr_h, vvv) +INSN_LSX(vbitclr_w, vvv) +INSN_LSX(vbitclr_d, vvv) +INSN_LSX(vbitclri_b, vv_i) +INSN_LSX(vbitclri_h, vv_i) +INSN_LSX(vbitclri_w, vv_i) +INSN_LSX(vbitclri_d, vv_i) +INSN_LSX(vbitset_b, vvv) +INSN_LSX(vbitset_h, vvv) +INSN_LSX(vbitset_w, vvv) +INSN_LSX(vbitset_d, vvv) +INSN_LSX(vbitseti_b, vv_i) +INSN_LSX(vbitseti_h, vv_i) +INSN_LSX(vbitseti_w, vv_i) +INSN_LSX(vbitseti_d, vv_i) +INSN_LSX(vbitrev_b, vvv) +INSN_LSX(vbitrev_h, vvv) +INSN_LSX(vbitrev_w, vvv) +INSN_LSX(vbitrev_d, vvv) +INSN_LSX(vbitrevi_b, vv_i) +INSN_LSX(vbitrevi_h, vv_i) +INSN_LSX(vbitrevi_w, vv_i) +INSN_LSX(vbitrevi_d, vv_i) diff --git a/target/loongarch/helper.h b/target/loongarch/helper.h index 6869b05105..d1983d9404 100644 --- a/target/loongarch/helper.h +++ b/target/loongarch/helper.h @@ -598,3 +598,28 @@ DEF_HELPER_3(vpcnt_b, void, env, i32, i32) DEF_HELPER_3(vpcnt_h, void, env, i32, i32) DEF_HELPER_3(vpcnt_w, void, env, i32, i32) DEF_HELPER_3(vpcnt_d, void, env, i32, i32) + +DEF_HELPER_4(vbitclr_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclr_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclr_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclr_d, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclri_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclri_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclri_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitclri_d, void, env, i32, i32, i32) +DEF_HELPER_4(vbitset_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitset_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitset_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitset_d, void, env, i32, i32, i32) +DEF_HELPER_4(vbitseti_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitseti_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitseti_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitseti_d, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrev_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrev_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrev_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrev_d, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrevi_b, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrevi_h, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrevi_w, void, env, i32, i32, i32) +DEF_HELPER_4(vbitrevi_d, void, env, i32, i32, i32) diff --git a/target/loongarch/insn_trans/trans_lsx.c.inc b/target/loongarch/insn_trans/trans_lsx.c.inc index 38493c98b0..141d7474dc 100644 --- a/target/loongarch/insn_trans/trans_lsx.c.inc +++ b/target/loongarch/insn_trans/trans_lsx.c.inc @@ -514,3 +514,28 @@ TRANS(vpcnt_b, gen_vv, gen_helper_vpcnt_b) TRANS(vpcnt_h, gen_vv, gen_helper_vpcnt_h) TRANS(vpcnt_w, gen_vv, gen_helper_vpcnt_w) TRANS(vpcnt_d, gen_vv, gen_helper_vpcnt_d) + +TRANS(vbitclr_b, gen_vvv, gen_helper_vbitclr_b) +TRANS(vbitclr_h, gen_vvv, gen_helper_vbitclr_h) +TRANS(vbitclr_w, gen_vvv, gen_helper_vbitclr_w) +TRANS(vbitclr_d, gen_vvv, gen_helper_vbitclr_d) +TRANS(vbitclri_b, gen_vv_i, gen_helper_vbitclri_b) +TRANS(vbitclri_h, gen_vv_i, gen_helper_vbitclri_h) +TRANS(vbitclri_w, gen_vv_i, gen_helper_vbitclri_w) +TRANS(vbitclri_d, gen_vv_i, gen_helper_vbitclri_d) +TRANS(vbitset_b, gen_vvv, gen_helper_vbitset_b) +TRANS(vbitset_h, gen_vvv, gen_helper_vbitset_h) +TRANS(vbitset_w, gen_vvv, gen_helper_vbitset_w) +TRANS(vbitset_d, gen_vvv, gen_helper_vbitset_d) +TRANS(vbitseti_b, gen_vv_i, gen_helper_vbitseti_b) +TRANS(vbitseti_h, gen_vv_i, gen_helper_vbitseti_h) +TRANS(vbitseti_w, gen_vv_i, gen_helper_vbitseti_w) +TRANS(vbitseti_d, gen_vv_i, gen_helper_vbitseti_d) +TRANS(vbitrev_b, gen_vvv, gen_helper_vbitrev_b) +TRANS(vbitrev_h, gen_vvv, gen_helper_vbitrev_h) +TRANS(vbitrev_w, gen_vvv, gen_helper_vbitrev_w) +TRANS(vbitrev_d, gen_vvv, gen_helper_vbitrev_d) +TRANS(vbitrevi_b, gen_vv_i, gen_helper_vbitrevi_b) +TRANS(vbitrevi_h, gen_vv_i, gen_helper_vbitrevi_h) +TRANS(vbitrevi_w, gen_vv_i, gen_helper_vbitrevi_w) +TRANS(vbitrevi_d, gen_vv_i, gen_helper_vbitrevi_d) diff --git a/target/loongarch/insns.decode b/target/loongarch/insns.decode index 812262ff78..74667ae6e0 100644 --- a/target/loongarch/insns.decode +++ b/target/loongarch/insns.decode @@ -972,3 +972,28 @@ vpcnt_b 0111 00101001 11000 01000 ..... ..... @vv vpcnt_h 0111 00101001 11000 01001 ..... ..... @vv vpcnt_w 0111 00101001 11000 01010 ..... ..... @vv vpcnt_d 0111 00101001 11000 01011 ..... ..... @vv + +vbitclr_b 0111 00010000 11000 ..... ..... ..... @vvv +vbitclr_h 0111 00010000 11001 ..... ..... ..... @vvv +vbitclr_w 0111 00010000 11010 ..... ..... ..... @vvv +vbitclr_d 0111 00010000 11011 ..... ..... ..... @vvv +vbitclri_b 0111 00110001 00000 01 ... ..... ..... @vv_ui3 +vbitclri_h 0111 00110001 00000 1 .... ..... ..... @vv_ui4 +vbitclri_w 0111 00110001 00001 ..... ..... ..... @vv_ui5 +vbitclri_d 0111 00110001 0001 ...... ..... ..... @vv_ui6 +vbitset_b 0111 00010000 11100 ..... ..... ..... @vvv +vbitset_h 0111 00010000 11101 ..... ..... ..... @vvv +vbitset_w 0111 00010000 11110 ..... ..... ..... @vvv +vbitset_d 0111 00010000 11111 ..... ..... ..... @vvv +vbitseti_b 0111 00110001 01000 01 ... ..... ..... @vv_ui3 +vbitseti_h 0111 00110001 01000 1 .... ..... ..... @vv_ui4 +vbitseti_w 0111 00110001 01001 ..... ..... ..... @vv_ui5 +vbitseti_d 0111 00110001 0101 ...... ..... ..... @vv_ui6 +vbitrev_b 0111 00010001 00000 ..... ..... ..... @vvv +vbitrev_h 0111 00010001 00001 ..... ..... ..... @vvv +vbitrev_w 0111 00010001 00010 ..... ..... ..... @vvv +vbitrev_d 0111 00010001 00011 ..... ..... ..... @vvv +vbitrevi_b 0111 00110001 10000 01 ... ..... ..... @vv_ui3 +vbitrevi_h 0111 00110001 10000 1 .... ..... ..... @vv_ui4 +vbitrevi_w 0111 00110001 10001 ..... ..... ..... @vv_ui5 +vbitrevi_d 0111 00110001 1001 ...... ..... ..... @vv_ui6 diff --git a/target/loongarch/lsx_helper.c b/target/loongarch/lsx_helper.c index c9913dec54..f88719908a 100644 --- a/target/loongarch/lsx_helper.c +++ b/target/loongarch/lsx_helper.c @@ -3236,3 +3236,165 @@ DO_HELPER_VV(vpcnt_b, 8, helper_vv, do_vpcnt) DO_HELPER_VV(vpcnt_h, 16, helper_vv, do_vpcnt) DO_HELPER_VV(vpcnt_w, 32, helper_vv, do_vpcnt) DO_HELPER_VV(vpcnt_d, 64, helper_vv, do_vpcnt) + +static int64_t vbitclr(int64_t s1, int64_t imm, int bit) +{ + return (s1 & (~(1LL << (imm % bit)))) & MAKE_64BIT_MASK(0, bit); +} + +static void do_vbitclr(vec_t *Vd, vec_t *Vj, vec_t *Vk, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitclr(Vj->B[n], Vk->B[n], bit); + break; + case 16: + Vd->H[n] = vbitclr(Vj->H[n], Vk->H[n], bit); + break; + case 32: + Vd->W[n] = vbitclr(Vj->W[n], Vk->W[n], bit); + break; + case 64: + Vd->D[n] = vbitclr(Vj->D[n], Vk->D[n], bit); + break; + default: + g_assert_not_reached(); + } +} + +static void do_vbitclr_i(vec_t *Vd, vec_t *Vj, uint32_t imm, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitclr(Vj->B[n], imm, bit); + break; + case 16: + Vd->H[n] = vbitclr(Vj->H[n], imm, bit); + break; + case 32: + Vd->W[n] = vbitclr(Vj->W[n], imm, bit); + break; + case 64: + Vd->D[n] = vbitclr(Vj->D[n], imm, bit); + break; + default: + g_assert_not_reached(); + } +} + +DO_HELPER_VVV(vbitclr_b, 8, helper_vvv, do_vbitclr) +DO_HELPER_VVV(vbitclr_h, 16, helper_vvv, do_vbitclr) +DO_HELPER_VVV(vbitclr_w, 32, helper_vvv, do_vbitclr) +DO_HELPER_VVV(vbitclr_d, 64, helper_vvv, do_vbitclr) +DO_HELPER_VV_I(vbitclri_b, 8, helper_vv_i, do_vbitclr_i) +DO_HELPER_VV_I(vbitclri_h, 16, helper_vv_i, do_vbitclr_i) +DO_HELPER_VV_I(vbitclri_w, 32, helper_vv_i, do_vbitclr_i) +DO_HELPER_VV_I(vbitclri_d, 64, helper_vv_i, do_vbitclr_i) + +static int64_t vbitset(int64_t s1, int64_t imm, int bit) +{ + return (s1 | (1LL << (imm % bit))) & MAKE_64BIT_MASK(0, bit); +} + +static void do_vbitset(vec_t *Vd, vec_t *Vj, vec_t *Vk, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitset(Vj->B[n], Vk->B[n], bit); + break; + case 16: + Vd->H[n] = vbitset(Vj->H[n], Vk->H[n], bit); + break; + case 32: + Vd->W[n] = vbitset(Vj->W[n], Vk->W[n], bit); + break; + case 64: + Vd->D[n] = vbitset(Vj->D[n], Vk->D[n], bit); + break; + default: + g_assert_not_reached(); + } +} + +static void do_vbitset_i(vec_t *Vd, vec_t *Vj, uint32_t imm, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitset(Vj->B[n], imm, bit); + break; + case 16: + Vd->H[n] = vbitset(Vj->H[n], imm, bit); + break; + case 32: + Vd->W[n] = vbitset(Vj->W[n], imm, bit); + break; + case 64: + Vd->D[n] = vbitset(Vj->D[n], imm, bit); + break; + default: + g_assert_not_reached(); + } +} + +DO_HELPER_VVV(vbitset_b, 8, helper_vvv, do_vbitset) +DO_HELPER_VVV(vbitset_h, 16, helper_vvv, do_vbitset) +DO_HELPER_VVV(vbitset_w, 32, helper_vvv, do_vbitset) +DO_HELPER_VVV(vbitset_d, 64, helper_vvv, do_vbitset) +DO_HELPER_VV_I(vbitseti_b, 8, helper_vv_i, do_vbitset_i) +DO_HELPER_VV_I(vbitseti_h, 16, helper_vv_i, do_vbitset_i) +DO_HELPER_VV_I(vbitseti_w, 32, helper_vv_i, do_vbitset_i) +DO_HELPER_VV_I(vbitseti_d, 64, helper_vv_i, do_vbitset_i) + +static int64_t vbitrev(int64_t s1, int64_t imm, int bit) +{ + return (s1 ^ (1LL << (imm % bit))) & MAKE_64BIT_MASK(0, bit); +} + +static void do_vbitrev(vec_t *Vd, vec_t *Vj, vec_t *Vk, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitrev(Vj->B[n], Vk->B[n], bit); + break; + case 16: + Vd->H[n] = vbitrev(Vj->H[n], Vk->H[n], bit); + break; + case 32: + Vd->W[n] = vbitrev(Vj->W[n], Vk->W[n], bit); + break; + case 64: + Vd->D[n] = vbitrev(Vj->D[n], Vk->D[n], bit); + break; + default: + g_assert_not_reached(); + } +} + +static void do_vbitrev_i(vec_t *Vd, vec_t *Vj, uint32_t imm, int bit, int n) +{ + switch (bit) { + case 8: + Vd->B[n] = vbitrev(Vj->B[n], imm, bit); + break; + case 16: + Vd->H[n] = vbitrev(Vj->H[n], imm, bit); + break; + case 32: + Vd->W[n] = vbitrev(Vj->W[n], imm, bit); + break; + case 64: + Vd->D[n] = vbitrev(Vj->D[n], imm, bit); + break; + default: + g_assert_not_reached(); + } +} + +DO_HELPER_VVV(vbitrev_b, 8, helper_vvv, do_vbitrev) +DO_HELPER_VVV(vbitrev_h, 16, helper_vvv, do_vbitrev) +DO_HELPER_VVV(vbitrev_w, 32, helper_vvv, do_vbitrev) +DO_HELPER_VVV(vbitrev_d, 64, helper_vvv, do_vbitrev) +DO_HELPER_VV_I(vbitrevi_b, 8, helper_vv_i, do_vbitrev_i) +DO_HELPER_VV_I(vbitrevi_h, 16, helper_vv_i, do_vbitrev_i) +DO_HELPER_VV_I(vbitrevi_w, 32, helper_vv_i, do_vbitrev_i) +DO_HELPER_VV_I(vbitrevi_d, 64, helper_vv_i, do_vbitrev_i) -- 2.31.1