Jonathan Wright <jonathan.wri...@arm.com> writes: > Hi, > > This patch declares unsigned type-qualified builtins and uses them to > implement the vector reduction Neon intrinsics. This removes the need > for many casts in arm_neon.h. > > Regression tested and bootstrapped on aarch64-none-linux-gnu - no > issues. > > Ok for master? > > Thanks, > Jonathan > > --- > > gcc/ChangeLog: > > 2021-11-09 Jonathan Wright <jonathan.wri...@arm.com> > > * config/aarch64/aarch64-simd-builtins.def: Declare unsigned > builtins for vector reduction. > * config/aarch64/arm_neon.h (vaddv_u8): Use type-qualified > builtin and remove casts. > (vaddv_u16): Likewise. > (vaddv_u32): Likewise. > (vaddvq_u8): Likewise. > (vaddvq_u16): Likewise. > (vaddvq_u32): Likewise. > (vaddvq_u64): Likewise.
OK, thanks. Richard > > diff --git a/gcc/config/aarch64/aarch64-simd-builtins.def > b/gcc/config/aarch64/aarch64-simd-builtins.def > index > 7d6de6728cf7c63872e09850a394101f7abf21d4..35a099e1fb8dd1acb9e35583d1267df257d961b0 > 100644 > --- a/gcc/config/aarch64/aarch64-simd-builtins.def > +++ b/gcc/config/aarch64/aarch64-simd-builtins.def > @@ -513,6 +513,7 @@ > > /* Implemented by aarch64_reduc_plus_<mode>. */ > BUILTIN_VALL (UNOP, reduc_plus_scal_, 10, NONE) > + BUILTIN_VDQ_I (UNOPU, reduc_plus_scal_, 10, NONE) > > /* Implemented by reduc_<maxmin_uns>_scal_<mode> (producing scalar). */ > BUILTIN_VDQIF_F16 (UNOP, reduc_smax_scal_, 10, NONE) > diff --git a/gcc/config/aarch64/arm_neon.h b/gcc/config/aarch64/arm_neon.h > index > ab46897d784b81bec9654d87557640ca4c1e5681..3c03432b5b6c6cd0f349671366615925d38121e5 > 100644 > --- a/gcc/config/aarch64/arm_neon.h > +++ b/gcc/config/aarch64/arm_neon.h > @@ -9695,21 +9695,21 @@ __extension__ extern __inline uint8_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddv_u8 (uint8x8_t __a) > { > - return (uint8_t) __builtin_aarch64_reduc_plus_scal_v8qi ((int8x8_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v8qi_uu (__a); > } > > __extension__ extern __inline uint16_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddv_u16 (uint16x4_t __a) > { > - return (uint16_t) __builtin_aarch64_reduc_plus_scal_v4hi ((int16x4_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v4hi_uu (__a); > } > > __extension__ extern __inline uint32_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddv_u32 (uint32x2_t __a) > { > - return (int32_t) __builtin_aarch64_reduc_plus_scal_v2si ((int32x2_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v2si_uu (__a); > } > > __extension__ extern __inline int8_t > @@ -9744,28 +9744,28 @@ __extension__ extern __inline uint8_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddvq_u8 (uint8x16_t __a) > { > - return (uint8_t) __builtin_aarch64_reduc_plus_scal_v16qi ((int8x16_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v16qi_uu (__a); > } > > __extension__ extern __inline uint16_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddvq_u16 (uint16x8_t __a) > { > - return (uint16_t) __builtin_aarch64_reduc_plus_scal_v8hi ((int16x8_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v8hi_uu (__a); > } > > __extension__ extern __inline uint32_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddvq_u32 (uint32x4_t __a) > { > - return (uint32_t) __builtin_aarch64_reduc_plus_scal_v4si ((int32x4_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v4si_uu (__a); > } > > __extension__ extern __inline uint64_t > __attribute__ ((__always_inline__, __gnu_inline__, __artificial__)) > vaddvq_u64 (uint64x2_t __a) > { > - return (uint64_t) __builtin_aarch64_reduc_plus_scal_v2di ((int64x2_t) __a); > + return __builtin_aarch64_reduc_plus_scal_v2di_uu (__a); > } > > __extension__ extern __inline float32_t