On 17 February 2018 at 18:22, Richard Henderson <richard.hender...@linaro.org> wrote: > Signed-off-by: Richard Henderson <richard.hender...@linaro.org> > --- > target/arm/translate-sve.c | 132 > +++++++++++++++++++++++++++++++++++++++++++++ > target/arm/sve.decode | 22 +++++++- > 2 files changed, 153 insertions(+), 1 deletion(-) > > diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c > index 50cf2a1fdd..c0cccfda6f 100644 > --- a/target/arm/translate-sve.c > +++ b/target/arm/translate-sve.c > @@ -46,6 +46,19 @@ typedef void GVecGen3Fn(unsigned, uint32_t, uint32_t, > * Implement all of the translator functions referenced by the decoder. > */ > > +/* Return the offset info CPUARMState of the predicate vector register Pn. > + * Note for this purpose, FFR is P16. */
Missing newline before */. > +/* > + *** SVE Memory - 32-bit Gather and Unsized Contiguous Group > + */ > + > +/* Subroutine loading a vector register at VOFS of LEN bytes. > + * The load should begin at the address Rn + IMM. > + */ > + > +#if UINTPTR_MAX == UINT32_MAX > +# define ptr i32 > +#else > +# define ptr i64 > +#endif I think that rather than doing this we should improve the provision that tcg/tcg-op.h has for *_ptr() wrappers, so from the target's point of view it has a proper tcg_const_local_ptr() and tcg_gen_brcondi_ptr(), same as for _i32, _i64 and _tl. > + > +static void do_ldr(DisasContext *s, uint32_t vofs, uint32_t len, > + int rn, int imm) > +{ > + uint32_t len_align = QEMU_ALIGN_DOWN(len, 8); > + uint32_t len_remain = len % 8; > + uint32_t nparts = len / 8 + ctpop8(len_remain); > + int midx = get_mem_index(s); > + TCGv_i64 addr, t0, t1; > + > + addr = tcg_temp_new_i64(); > + t0 = tcg_temp_new_i64(); > + > + /* Note that unpredicated load/store of vector/predicate registers > + * are defined as a stream of bytes, which equates to little-endian > + * operations on larger quantities. There is no nice way to force > + * a little-endian load for aarch64_be-linux-user out of line. > + * > + * Attempt to keep code expansion to a minimum by limiting the > + * amount of unrolling done. > + */ > + if (nparts <= 4) { > + int i; > + > + for (i = 0; i < len_align; i += 8) { > + tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + i); > + tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ); > + tcg_gen_st_i64(t0, cpu_env, vofs + i); > + } > + } else { > + TCGLabel *loop = gen_new_label(); > + TCGv_ptr i = TCGV_NAT_TO_PTR(glue(tcg_const_local_, ptr)(0)); > + TCGv_ptr dest; > + > + gen_set_label(loop); > + > + /* Minimize the number of local temps that must be re-read from > + * the stack each iteration. Instead, re-compute values other > + * than the loop counter. > + */ > + dest = tcg_temp_new_ptr(); > + tcg_gen_addi_ptr(dest, i, imm); > +#if UINTPTR_MAX == UINT32_MAX > + tcg_gen_extu_i32_i64(addr, TCGV_PTR_TO_NAT(dest)); > + tcg_gen_add_i64(addr, addr, cpu_reg_sp(s, rn)); > +#else > + tcg_gen_add_i64(addr, TCGV_PTR_TO_NAT(dest), cpu_reg_sp(s, rn)); > +#endif Can we avoid the ifdef by creating a tcg_gen_ext_ptr_i64() (similar to but opposite in effect to the existing tcg_gen_ext_i32_ptr()) ? > + > + tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEQ); > + > + tcg_gen_add_ptr(dest, cpu_env, i); > + tcg_gen_addi_ptr(i, i, 8); > + tcg_gen_st_i64(t0, dest, vofs); > + tcg_temp_free_ptr(dest); > + > + glue(tcg_gen_brcondi_, ptr)(TCG_COND_LTU, TCGV_PTR_TO_NAT(i), > + len_align, loop); > + tcg_temp_free_ptr(i); > + } > + > + /* Predicate register loads can be any multiple of 2. > + * Note that we still store the entire 64-bit unit into cpu_env. > + */ > + if (len_remain) { > + tcg_gen_addi_i64(addr, cpu_reg_sp(s, rn), imm + len_align); > + > + switch (len_remain) { > + case 2: > + case 4: > + case 8: > + tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LE | ctz32(len_remain)); > + break; > + > + case 6: > + t1 = tcg_temp_new_i64(); > + tcg_gen_qemu_ld_i64(t0, addr, midx, MO_LEUL); > + tcg_gen_addi_i64(addr, addr, 4); > + tcg_gen_qemu_ld_i64(t1, addr, midx, MO_LEUW); > + tcg_gen_deposit_i64(t0, t0, t1, 32, 32); > + tcg_temp_free_i64(t1); > + break; > + > + default: > + g_assert_not_reached(); > + } > + tcg_gen_st_i64(t0, cpu_env, vofs + len_align); > + } > + tcg_temp_free_i64(addr); > + tcg_temp_free_i64(t0); > +} > + > +#undef ptr > > +&rri rd rn imm > &rrr_esz rd rn rm esz > > ########################################################################### > @@ -31,7 +37,13 @@ > # reduce the amount of duplication between instruction patterns. > > # Three operand with unused vector element size > -@rd_rn_rm_e0 ........ ... rm:5 ... ... rn:5 rd:5 &rrr_esz esz=0 > +@rd_rn_rm_e0 ........ ... rm:5 ... ... rn:5 rd:5 &rrr_esz esz=0 This change looks like it should be squashed into a previous patch? > + > +# Basic Load/Store with 9-bit immediate offset > +@pd_rn_i9 ........ ........ ...... rn:5 . rd:4 \ > + &rri imm=%imm9_16_10 > +@rd_rn_i9 ........ ........ ...... rn:5 rd:5 \ > + &rri imm=%imm9_16_10 > thanks -- PMM