On Mon, Dec 17, 2018 at 10:38:49PM -0800, Richard Henderson wrote: > From: Mark Cave-Ayland <mark.cave-ayl...@ilande.co.uk> > > These helpers allow us to move AVR register values to/from the specified > TCGv_i64 > argument. > > To prevent VMX helpers accessing the cpu_avr{l,h} arrays directly, add extra > TCG > temporaries as required. > > Signed-off-by: Mark Cave-Ayland <mark.cave-ayl...@ilande.co.uk> > Reviewed-by: Richard Henderson <richard.hender...@linaro.org>
Acked-by: David Gibson <da...@gibson.dropbear.id.au> > Message-Id: <20181217122405.18732-3-mark.cave-ayl...@ilande.co.uk> > --- > target/ppc/translate.c | 10 +++ > target/ppc/translate/vmx-impl.inc.c | 128 ++++++++++++++++++++++------ > 2 files changed, 110 insertions(+), 28 deletions(-) > > diff --git a/target/ppc/translate.c b/target/ppc/translate.c > index 1d4bf624a3..fa3e8dc114 100644 > --- a/target/ppc/translate.c > +++ b/target/ppc/translate.c > @@ -6704,6 +6704,16 @@ static inline void set_fpr(int regno, TCGv_i64 src) > tcg_gen_mov_i64(cpu_fpr[regno], src); > } > > +static inline void get_avr64(TCGv_i64 dst, int regno, bool high) > +{ > + tcg_gen_mov_i64(dst, (high ? cpu_avrh : cpu_avrl)[regno]); > +} > + > +static inline void set_avr64(int regno, TCGv_i64 src, bool high) > +{ > + tcg_gen_mov_i64((high ? cpu_avrh : cpu_avrl)[regno], src); > +} > + > #include "translate/fp-impl.inc.c" > > #include "translate/vmx-impl.inc.c" > diff --git a/target/ppc/translate/vmx-impl.inc.c > b/target/ppc/translate/vmx-impl.inc.c > index 3cb6fc2926..30046c6e31 100644 > --- a/target/ppc/translate/vmx-impl.inc.c > +++ b/target/ppc/translate/vmx-impl.inc.c > @@ -18,52 +18,66 @@ static inline TCGv_ptr gen_avr_ptr(int reg) > static void glue(gen_, name)(DisasContext *ctx) > \ > { > \ > TCGv EA; > \ > + TCGv_i64 avr; > \ > if (unlikely(!ctx->altivec_enabled)) { > \ > gen_exception(ctx, POWERPC_EXCP_VPU); > \ > return; > \ > } > \ > gen_set_access_type(ctx, ACCESS_INT); > \ > + avr = tcg_temp_new_i64(); > \ > EA = tcg_temp_new(); > \ > gen_addr_reg_index(ctx, EA); > \ > tcg_gen_andi_tl(EA, EA, ~0xf); > \ > /* We only need to swap high and low halves. gen_qemu_ld64_i64 does > \ > necessary 64-bit byteswap already. */ > \ > if (ctx->le_mode) { > \ > - gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA); > \ > + gen_qemu_ld64_i64(ctx, avr, EA); > \ > + set_avr64(rD(ctx->opcode), avr, false); > \ > tcg_gen_addi_tl(EA, EA, 8); > \ > - gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA); > \ > + gen_qemu_ld64_i64(ctx, avr, EA); > \ > + set_avr64(rD(ctx->opcode), avr, true); > \ > } else { > \ > - gen_qemu_ld64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA); > \ > + gen_qemu_ld64_i64(ctx, avr, EA); > \ > + set_avr64(rD(ctx->opcode), avr, true); > \ > tcg_gen_addi_tl(EA, EA, 8); > \ > - gen_qemu_ld64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA); > \ > + gen_qemu_ld64_i64(ctx, avr, EA); > \ > + set_avr64(rD(ctx->opcode), avr, false); > \ > } > \ > tcg_temp_free(EA); > \ > + tcg_temp_free_i64(avr); > \ > } > > #define GEN_VR_STX(name, opc2, opc3) > \ > static void gen_st##name(DisasContext *ctx) > \ > { > \ > TCGv EA; > \ > + TCGv_i64 avr; > \ > if (unlikely(!ctx->altivec_enabled)) { > \ > gen_exception(ctx, POWERPC_EXCP_VPU); > \ > return; > \ > } > \ > gen_set_access_type(ctx, ACCESS_INT); > \ > + avr = tcg_temp_new_i64(); > \ > EA = tcg_temp_new(); > \ > gen_addr_reg_index(ctx, EA); > \ > tcg_gen_andi_tl(EA, EA, ~0xf); > \ > /* We only need to swap high and low halves. gen_qemu_st64_i64 does > \ > necessary 64-bit byteswap already. */ > \ > if (ctx->le_mode) { > \ > - gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA); > \ > + get_avr64(avr, rD(ctx->opcode), false); > \ > + gen_qemu_st64_i64(ctx, avr, EA); > \ > tcg_gen_addi_tl(EA, EA, 8); > \ > - gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA); > \ > + get_avr64(avr, rD(ctx->opcode), true); > \ > + gen_qemu_st64_i64(ctx, avr, EA); > \ > } else { > \ > - gen_qemu_st64_i64(ctx, cpu_avrh[rD(ctx->opcode)], EA); > \ > + get_avr64(avr, rD(ctx->opcode), true); > \ > + gen_qemu_st64_i64(ctx, avr, EA); > \ > tcg_gen_addi_tl(EA, EA, 8); > \ > - gen_qemu_st64_i64(ctx, cpu_avrl[rD(ctx->opcode)], EA); > \ > + get_avr64(avr, rD(ctx->opcode), false); > \ > + gen_qemu_st64_i64(ctx, avr, EA); > \ > } > \ > tcg_temp_free(EA); > \ > + tcg_temp_free_i64(avr); > \ > } > > #define GEN_VR_LVE(name, opc2, opc3, size) \ > @@ -159,15 +173,20 @@ static void gen_lvsr(DisasContext *ctx) > static void gen_mfvscr(DisasContext *ctx) > { > TCGv_i32 t; > + TCGv_i64 avr; > if (unlikely(!ctx->altivec_enabled)) { > gen_exception(ctx, POWERPC_EXCP_VPU); > return; > } > - tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0); > + avr = tcg_temp_new_i64(); > + tcg_gen_movi_i64(avr, 0); > + set_avr64(rD(ctx->opcode), avr, true); > t = tcg_temp_new_i32(); > tcg_gen_ld_i32(t, cpu_env, offsetof(CPUPPCState, vscr)); > - tcg_gen_extu_i32_i64(cpu_avrl[rD(ctx->opcode)], t); > + tcg_gen_extu_i32_i64(avr, t); > + set_avr64(rD(ctx->opcode), avr, false); > tcg_temp_free_i32(t); > + tcg_temp_free_i64(avr); > } > > static void gen_mtvscr(DisasContext *ctx) > @@ -188,6 +207,7 @@ static void glue(gen_, name)(DisasContext *ctx) > \ > TCGv_i64 t0 = tcg_temp_new_i64(); \ > TCGv_i64 t1 = tcg_temp_new_i64(); \ > TCGv_i64 t2 = tcg_temp_new_i64(); \ > + TCGv_i64 avr = tcg_temp_new_i64(); \ > TCGv_i64 ten, z; \ > \ > if (unlikely(!ctx->altivec_enabled)) { \ > @@ -199,26 +219,35 @@ static void glue(gen_, name)(DisasContext *ctx) > \ > z = tcg_const_i64(0); \ > \ > if (add_cin) { \ > - tcg_gen_mulu2_i64(t0, t1, cpu_avrl[rA(ctx->opcode)], ten); \ > - tcg_gen_andi_i64(t2, cpu_avrl[rB(ctx->opcode)], 0xF); \ > - tcg_gen_add2_i64(cpu_avrl[rD(ctx->opcode)], t2, t0, t1, t2, z); \ > + get_avr64(avr, rA(ctx->opcode), false); \ > + tcg_gen_mulu2_i64(t0, t1, avr, ten); \ > + get_avr64(avr, rB(ctx->opcode), false); \ > + tcg_gen_andi_i64(t2, avr, 0xF); \ > + tcg_gen_add2_i64(avr, t2, t0, t1, t2, z); \ > + set_avr64(rD(ctx->opcode), avr, false); \ > } else { \ > - tcg_gen_mulu2_i64(cpu_avrl[rD(ctx->opcode)], t2, \ > - cpu_avrl[rA(ctx->opcode)], ten); \ > + get_avr64(avr, rA(ctx->opcode), false); \ > + tcg_gen_mulu2_i64(avr, t2, avr, ten); \ > + set_avr64(rD(ctx->opcode), avr, false); \ > } \ > \ > if (ret_carry) { \ > - tcg_gen_mulu2_i64(t0, t1, cpu_avrh[rA(ctx->opcode)], ten); \ > - tcg_gen_add2_i64(t0, cpu_avrl[rD(ctx->opcode)], t0, t1, t2, z); \ > - tcg_gen_movi_i64(cpu_avrh[rD(ctx->opcode)], 0); \ > + get_avr64(avr, rA(ctx->opcode), true); \ > + tcg_gen_mulu2_i64(t0, t1, avr, ten); \ > + tcg_gen_add2_i64(t0, avr, t0, t1, t2, z); \ > + set_avr64(rD(ctx->opcode), avr, false); \ > + set_avr64(rD(ctx->opcode), z, true); \ > } else { \ > - tcg_gen_mul_i64(t0, cpu_avrh[rA(ctx->opcode)], ten); \ > - tcg_gen_add_i64(cpu_avrh[rD(ctx->opcode)], t0, t2); \ > + get_avr64(avr, rA(ctx->opcode), true); \ > + tcg_gen_mul_i64(t0, avr, ten); \ > + tcg_gen_add_i64(avr, t0, t2); \ > + set_avr64(rD(ctx->opcode), avr, true); \ > } \ > \ > tcg_temp_free_i64(t0); \ > tcg_temp_free_i64(t1); \ > tcg_temp_free_i64(t2); \ > + tcg_temp_free_i64(avr); \ > tcg_temp_free_i64(ten); \ > tcg_temp_free_i64(z); \ > } \ > @@ -232,12 +261,27 @@ GEN_VX_VMUL10(vmul10ecuq, 1, 1); > #define GEN_VX_LOGICAL(name, tcg_op, opc2, opc3) \ > static void glue(gen_, name)(DisasContext *ctx) > \ > { \ > + TCGv_i64 t0 = tcg_temp_new_i64(); \ > + TCGv_i64 t1 = tcg_temp_new_i64(); \ > + TCGv_i64 avr = tcg_temp_new_i64(); \ > + \ > if (unlikely(!ctx->altivec_enabled)) { \ > gen_exception(ctx, POWERPC_EXCP_VPU); \ > return; \ > } \ > - tcg_op(cpu_avrh[rD(ctx->opcode)], cpu_avrh[rA(ctx->opcode)], > cpu_avrh[rB(ctx->opcode)]); \ > - tcg_op(cpu_avrl[rD(ctx->opcode)], cpu_avrl[rA(ctx->opcode)], > cpu_avrl[rB(ctx->opcode)]); \ > + get_avr64(t0, rA(ctx->opcode), true); \ > + get_avr64(t1, rB(ctx->opcode), true); \ > + tcg_op(avr, t0, t1); \ > + set_avr64(rD(ctx->opcode), avr, true); \ > + \ > + get_avr64(t0, rA(ctx->opcode), false); \ > + get_avr64(t1, rB(ctx->opcode), false); \ > + tcg_op(avr, t0, t1); \ > + set_avr64(rD(ctx->opcode), avr, false); \ > + \ > + tcg_temp_free_i64(t0); \ > + tcg_temp_free_i64(t1); \ > + tcg_temp_free_i64(avr); \ > } > > GEN_VX_LOGICAL(vand, tcg_gen_and_i64, 2, 16); > @@ -406,6 +450,7 @@ GEN_VXFORM(vmrglw, 6, 6); > static void gen_vmrgew(DisasContext *ctx) > { > TCGv_i64 tmp; > + TCGv_i64 avr; > int VT, VA, VB; > if (unlikely(!ctx->altivec_enabled)) { > gen_exception(ctx, POWERPC_EXCP_VPU); > @@ -415,15 +460,28 @@ static void gen_vmrgew(DisasContext *ctx) > VA = rA(ctx->opcode); > VB = rB(ctx->opcode); > tmp = tcg_temp_new_i64(); > - tcg_gen_shri_i64(tmp, cpu_avrh[VB], 32); > - tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VA], tmp, 0, 32); > - tcg_gen_shri_i64(tmp, cpu_avrl[VB], 32); > - tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VA], tmp, 0, 32); > + avr = tcg_temp_new_i64(); > + > + get_avr64(avr, VB, true); > + tcg_gen_shri_i64(tmp, avr, 32); > + get_avr64(avr, VA, true); > + tcg_gen_deposit_i64(avr, avr, tmp, 0, 32); > + set_avr64(VT, avr, true); > + > + get_avr64(avr, VB, false); > + tcg_gen_shri_i64(tmp, avr, 32); > + get_avr64(avr, VA, false); > + tcg_gen_deposit_i64(avr, avr, tmp, 0, 32); > + set_avr64(VT, avr, false); > + > tcg_temp_free_i64(tmp); > + tcg_temp_free_i64(avr); > } > > static void gen_vmrgow(DisasContext *ctx) > { > + TCGv_i64 t0, t1; > + TCGv_i64 avr; > int VT, VA, VB; > if (unlikely(!ctx->altivec_enabled)) { > gen_exception(ctx, POWERPC_EXCP_VPU); > @@ -432,9 +490,23 @@ static void gen_vmrgow(DisasContext *ctx) > VT = rD(ctx->opcode); > VA = rA(ctx->opcode); > VB = rB(ctx->opcode); > + t0 = tcg_temp_new_i64(); > + t1 = tcg_temp_new_i64(); > + avr = tcg_temp_new_i64(); > > - tcg_gen_deposit_i64(cpu_avrh[VT], cpu_avrh[VB], cpu_avrh[VA], 32, 32); > - tcg_gen_deposit_i64(cpu_avrl[VT], cpu_avrl[VB], cpu_avrl[VA], 32, 32); > + get_avr64(t0, VB, true); > + get_avr64(t1, VA, true); > + tcg_gen_deposit_i64(avr, t0, t1, 32, 32); > + set_avr64(VT, avr, true); > + > + get_avr64(t0, VB, false); > + get_avr64(t1, VA, false); > + tcg_gen_deposit_i64(avr, t0, t1, 32, 32); > + set_avr64(VT, avr, false); > + > + tcg_temp_free_i64(t0); > + tcg_temp_free_i64(t1); > + tcg_temp_free_i64(avr); > } > > GEN_VXFORM(vmuloub, 4, 0); -- David Gibson | I'll have my music baroque, and my code david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_ | _way_ _around_! http://www.ozlabs.org/~dgibson
signature.asc
Description: PGP signature