On Fri, Apr 14, 2023 at 12:03 AM Lou Knauer via Gcc-patches <gcc-patches@gcc.gnu.org> wrote: > > This adds support for the -mveclibabi option to the AArch64 backend of GCC by > implementing the builtin_vectorized_function target hook for AArch64. > The SLEEF Vectorized Math Library's GNUABI interface is used, and > NEON/Advanced SIMD as well as SVE are supported. > > This was tested on the gcc testsuite and the llvm-test-suite on a AArch64 > host for NEON and SVE as well as on hand-written benchmarks. Where the > vectorization of builtins was applied successfully in loops bound by the > calls to those, significant (>2) performance gains can be observed.
This is so wrong and it is better if you actually just used a header file instead. Specifically the openmp vect pragmas. Thanks, Andrew Pinski > > gcc/ChangeLog: > > * config/aarch64/aarch64.opt: Add -mveclibabi option. > * config/aarch64/aarch64-opts.h: Add aarch64_veclibabi enum. > * config/aarch64/aarch64-protos.h: Add > aarch64_builtin_vectorized_function declaration. > * config/aarch64/aarch64.cc: Handle -mveclibabi option and pure > scalable type info for scalable vectors without "SVE type" attributes. > * config/aarch64/aarch64-builtins.cc: Add > aarch64_builtin_vectorized_function definition. > * doc/invoke.texi: Document -mveclibabi for AArch64 targets. > > gcc/testsuite/ChangeLog: > > * gcc.target/aarch64/vect-vecabi-sleefgnu-neon.c: New testcase. > * gcc.target/aarch64/vect-vecabi-sleefgnu-sve.c: New testcase. > --- > gcc/config/aarch64/aarch64-builtins.cc | 113 ++++++++++++++++++ > gcc/config/aarch64/aarch64-opts.h | 5 + > gcc/config/aarch64/aarch64-protos.h | 3 + > gcc/config/aarch64/aarch64.cc | 66 ++++++++++ > gcc/config/aarch64/aarch64.opt | 15 +++ > gcc/doc/invoke.texi | 15 +++ > .../aarch64/vect-vecabi-sleefgnu-neon.c | 16 +++ > .../aarch64/vect-vecabi-sleefgnu-sve.c | 16 +++ > 8 files changed, 249 insertions(+) > create mode 100644 > gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-neon.c > create mode 100644 > gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-sve.c > > diff --git a/gcc/config/aarch64/aarch64-builtins.cc > b/gcc/config/aarch64/aarch64-builtins.cc > index cc6b7c01fd1..f53fa91b8d0 100644 > --- a/gcc/config/aarch64/aarch64-builtins.cc > +++ b/gcc/config/aarch64/aarch64-builtins.cc > @@ -47,6 +47,7 @@ > #include "stringpool.h" > #include "attribs.h" > #include "gimple-fold.h" > +#include "builtins.h" > > #define v8qi_UP E_V8QImode > #define v8di_UP E_V8DImode > @@ -3450,6 +3451,118 @@ aarch64_resolve_overloaded_builtin_general > (location_t loc, tree function, > return NULL_TREE; > } > > +/* The vector library abi to use, if any. */ > +extern enum aarch64_veclibabi aarch64_selected_veclibabi; > + > +/* Returns a function declaration for a vectorized version of the combined > + function with combined_fn code FN and the result vector type TYPE. > + NULL_TREE is returned if there is none available. */ > +tree > +aarch64_builtin_vectorized_function (unsigned int fn_code, > + tree type_out, tree type_in) > +{ > + if (TREE_CODE (type_out) != VECTOR_TYPE > + || TREE_CODE (type_in) != VECTOR_TYPE > + || aarch64_selected_veclibabi != aarch64_veclibabi_type_sleefgnu > + || !flag_unsafe_math_optimizations) > + return NULL_TREE; > + > + machine_mode mode = TYPE_MODE (TREE_TYPE (type_out)); > + poly_uint64 n = TYPE_VECTOR_SUBPARTS (type_out); > + if (mode != TYPE_MODE (TREE_TYPE (type_in)) > + || !known_eq (n, TYPE_VECTOR_SUBPARTS (type_in))) > + return NULL_TREE; > + > + bool is_scalable = !n.is_constant (); > + if (is_scalable) > + { > + /* SVE is needed for scalable vectors, a SVE register's size is > + always a multiple of 128. */ > + if (!TARGET_SVE > + || (mode == DFmode && !known_eq (n, poly_uint64 (2, 2))) > + || (mode == SFmode && !known_eq (n, poly_uint64 (4, 4)))) > + return NULL_TREE; > + } > + else > + { > + /* A NEON register can hold two doubles or one float. */ > + if (!TARGET_SIMD > + || (mode == DFmode && n.to_constant () != 2) > + || (mode == SFmode && n.to_constant () != 4)) > + return NULL_TREE; > + } > + > + tree fntype; > + combined_fn fn = combined_fn (fn_code); > + const char *argencoding; > + switch (fn) > + { > + CASE_CFN_EXP: > + CASE_CFN_LOG: > + CASE_CFN_LOG10: > + CASE_CFN_TANH: > + CASE_CFN_TAN: > + CASE_CFN_ATAN: > + CASE_CFN_ATANH: > + CASE_CFN_CBRT: > + CASE_CFN_SINH: > + CASE_CFN_SIN: > + CASE_CFN_ASINH: > + CASE_CFN_ASIN: > + CASE_CFN_COSH: > + CASE_CFN_COS: > + CASE_CFN_ACOSH: > + CASE_CFN_ACOS: > + fntype = build_function_type_list (type_out, type_in, NULL); > + argencoding = "v"; > + break; > + > + CASE_CFN_POW: > + CASE_CFN_ATAN2: > + fntype = build_function_type_list (type_out, type_in, type_in, NULL); > + argencoding = "vv"; > + break; > + > + default: > + return NULL_TREE; > + } > + > + tree fndecl = mathfn_built_in (mode == DFmode > + ? double_type_node : float_type_node, fn); > + const char *scalar_name = IDENTIFIER_POINTER (DECL_NAME (fndecl)); > + /* Builtins will always be prefixed with '__builtin_'. */ > + gcc_assert (strncmp (scalar_name, "__builtin_", 10) == 0); > + scalar_name += 10; > + > + char vectorized_name[32]; > + if (is_scalable) > + { > + /* SVE ISA */ > + int n = snprintf (vectorized_name, sizeof (vectorized_name), > + "_ZGVsNx%s_%s", argencoding, scalar_name); > + if (n < 0 || n > sizeof (vectorized_name)) > + return NULL_TREE; > + } > + else > + { > + /* NEON ISA */ > + int n = snprintf (vectorized_name, sizeof (vectorized_name), > + "_ZGVnN%d%s_%s", mode == SFmode ? 4 : 2, > + argencoding, scalar_name); > + if (n < 0 || n > sizeof (vectorized_name)) > + return NULL_TREE; > + } > + > + tree new_fndecl = build_decl (BUILTINS_LOCATION, FUNCTION_DECL, > + get_identifier (vectorized_name), fntype); > + TREE_PUBLIC (new_fndecl) = 1; > + TREE_READONLY (new_fndecl) = 1; > + DECL_EXTERNAL (new_fndecl) = 1; > + DECL_IS_NOVOPS (new_fndecl) = 1; > + > + return new_fndecl; > +} > + > #undef AARCH64_CHECK_BUILTIN_MODE > #undef AARCH64_FIND_FRINT_VARIANT > #undef CF0 > diff --git a/gcc/config/aarch64/aarch64-opts.h > b/gcc/config/aarch64/aarch64-opts.h > index a9f3e2715ca..d12871b893c 100644 > --- a/gcc/config/aarch64/aarch64-opts.h > +++ b/gcc/config/aarch64/aarch64-opts.h > @@ -98,4 +98,9 @@ enum aarch64_key_type { > AARCH64_KEY_B > }; > > +enum aarch64_veclibabi { > + aarch64_veclibabi_type_none, > + aarch64_veclibabi_type_sleefgnu > +}; > + > #endif > diff --git a/gcc/config/aarch64/aarch64-protos.h > b/gcc/config/aarch64/aarch64-protos.h > index 63339fa47df..53c6e455da8 100644 > --- a/gcc/config/aarch64/aarch64-protos.h > +++ b/gcc/config/aarch64/aarch64-protos.h > @@ -1066,4 +1066,7 @@ extern bool aarch64_harden_sls_blr_p (void); > > extern void aarch64_output_patchable_area (unsigned int, bool); > > +extern tree aarch64_builtin_vectorized_function (unsigned int fn, > + tree type_out, tree type_in); > + > #endif /* GCC_AARCH64_PROTOS_H */ > diff --git a/gcc/config/aarch64/aarch64.cc b/gcc/config/aarch64/aarch64.cc > index 42617ced73a..50ac37ff01e 100644 > --- a/gcc/config/aarch64/aarch64.cc > +++ b/gcc/config/aarch64/aarch64.cc > @@ -84,6 +84,7 @@ > #include "aarch64-feature-deps.h" > #include "config/arm/aarch-common.h" > #include "config/arm/aarch-common-protos.h" > +#include "print-tree.h" > > /* This file should be included last. */ > #include "target-def.h" > @@ -2951,6 +2952,62 @@ pure_scalable_type_info::analyze (const_tree type) > return IS_PST; > } > > + /* Only functions and types that are part of the ARM C Language > + Extensions (arm_sve.h) have the SVE type attributes. > + The auto-vectorizer does not annotate the vector types it creates with > + those attributes. With the support of vectorized libm function > + builtins for SVE, scalable vectors without special attributes > + have to be treated as well. */ > + if (TREE_CODE (type) == VECTOR_TYPE > + && !TYPE_VECTOR_SUBPARTS (type).is_constant ()) > + { > + /* Boolean vectors are special because they are used by > + the vectorizer as masks that must go into the > + predicate registers. */ > + if (TREE_CODE (TREE_TYPE (type)) == BOOLEAN_TYPE) > + { > + p.num_zr = 0; > + p.num_pr = 1; > + p.mode = p.orig_mode = TYPE_MODE (type); > + add_piece (p); > + return IS_PST; > + } > + > + static const struct { > + machine_mode mode; > + unsigned int element_size; > + poly_uint64 vector_size; > + } valid_vectors[] = { > + { VNx8BFmode, 16, poly_uint64 (8, 8) }, /* svbfloat16_t */ > + { VNx8HFmode, 16, poly_uint64 (8, 8) }, /* svfloat16_t */ > + { VNx4SFmode, 32, poly_uint64 (4, 4) }, /* svfloat32_t */ > + { VNx2DFmode, 64, poly_uint64 (2, 2) }, /* svfloat64_t */ > + { VNx16BImode, 8, poly_uint64 (16, 16) }, /* sv[u]int8_t */ > + { VNx8HImode, 16, poly_uint64 (8, 8) }, /* sv[u]int16_t */ > + { VNx4SImode, 32, poly_uint64 (4, 4) }, /* sv[u]int32_t */ > + { VNx2DImode, 64, poly_uint64 (2, 2) }, /* sv[u]int64_t */ > + }; > + > + machine_mode elm_mode = TYPE_MODE (TREE_TYPE (type)); > + unsigned int elm_size = GET_MODE_BITSIZE (elm_mode).to_constant (); > + for (unsigned i = 0; > + i < sizeof (valid_vectors) / sizeof (valid_vectors[0]); i++) > + if (valid_vectors[i].element_size == elm_size > + && valid_vectors[i].mode == TYPE_MODE (type) > + && known_eq (valid_vectors[i].vector_size, > + TYPE_VECTOR_SUBPARTS (type))) > + { > + p.num_zr = 1; > + p.num_pr = 0; > + p.mode = p.orig_mode = valid_vectors[i].mode; > + add_piece (p); > + return IS_PST; > + } > + > + fatal_error (input_location, "unsupported vector type %qT" > + " as function parameter without SVE attributes", type); > + } > + > /* Check for user-defined PSTs. */ > if (TREE_CODE (type) == ARRAY_TYPE) > return analyze_array (type); > @@ -17851,6 +17908,8 @@ aarch64_override_options_after_change_1 (struct > gcc_options *opts) > flag_mrecip_low_precision_sqrt = true; > } > > +enum aarch64_veclibabi aarch64_selected_veclibabi = > aarch64_veclibabi_type_none; > + > /* 'Unpack' up the internal tuning structs and update the options > in OPTS. The caller must have set up selected_tune and selected_arch > as all the other target-specific codegen decisions are > @@ -18031,6 +18090,9 @@ aarch64_override_options_internal (struct gcc_options > *opts) > && opts->x_optimize >= aarch64_tune_params.prefetch->default_opt_level) > opts->x_flag_prefetch_loop_arrays = 1; > > + if (opts->x_aarch64_veclibabi_type == aarch64_veclibabi_type_sleefgnu) > + aarch64_selected_veclibabi = aarch64_veclibabi_type_sleefgnu; > + > aarch64_override_options_after_change_1 (opts); > } > > @@ -28085,6 +28147,10 @@ aarch64_libgcc_floating_mode_supported_p > #undef TARGET_CONST_ANCHOR > #define TARGET_CONST_ANCHOR 0x1000000 > > +#undef TARGET_VECTORIZE_BUILTIN_VECTORIZED_FUNCTION > +#define TARGET_VECTORIZE_BUILTIN_VECTORIZED_FUNCTION \ > + aarch64_builtin_vectorized_function > + > struct gcc_target targetm = TARGET_INITIALIZER; > > #include "gt-aarch64.h" > diff --git a/gcc/config/aarch64/aarch64.opt b/gcc/config/aarch64/aarch64.opt > index 1d7967db9c0..76013dacdea 100644 > --- a/gcc/config/aarch64/aarch64.opt > +++ b/gcc/config/aarch64/aarch64.opt > @@ -302,3 +302,18 @@ Constant memset size in bytes from which to start using > MOPS sequence. > -param=aarch64-vect-unroll-limit= > Target Joined UInteger Var(aarch64_vect_unroll_limit) Init(4) Param > Limit how much the autovectorizer may unroll a loop. > + > +;; -mveclibabi= > +TargetVariable > +enum aarch64_veclibabi aarch64_veclibabi_type = aarch64_veclibabi_type_none > + > +mveclibabi= > +Target RejectNegative Joined Var(aarch64_veclibabi_type) > Enum(aarch64_veclibabi) Init(aarch64_veclibabi_type_none) > +Vector library ABI to use. > + > +Enum > +Name(aarch64_veclibabi) Type(enum aarch64_veclibabi) > +Known vectorization library ABIs (for use with the -mveclibabi= option): > + > +EnumValue > +Enum(aarch64_veclibabi) String(sleefgnu) > Value(aarch64_veclibabi_type_sleefgnu) > diff --git a/gcc/doc/invoke.texi b/gcc/doc/invoke.texi > index a38547f53e5..71fbbf27522 100644 > --- a/gcc/doc/invoke.texi > +++ b/gcc/doc/invoke.texi > @@ -20383,6 +20383,21 @@ across releases. > > This option is only intended to be useful when developing GCC. > > +@opindex mveclibabi > +@item -mveclibabi=@var{type} > +Specifies the ABI type to use for vectorizing intrinsics using an > +external library. The only type supported at present is @samp{sleefgnu}, > +which specifies to use the GNU ABI variant of the Sleef Vectorized > +Math Library. This flag can be used for both, Advanced SIMD (NEON) and SVE. > + > +GCC currently emits vectorized calls to @code{exp}, @code{log}, @code{log10}, > +@code{tanh}, @code{tan}, @code{atan}, @code{atanh}, @code{cbrt}, @code{sinh}, > +@code{sin}, @code{asinh} and @code{asin} when possible and profitable > +on AArch64. > + > +Both @option{-ftree-vectorize} and @option{-funsafe-math-optimizations} > +must also be enabled. The libsleefgnu must be specified at link time. > + > @opindex mverbose-cost-dump > @item -mverbose-cost-dump > Enable verbose cost model dumping in the debug dump files. This option is > diff --git a/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-neon.c > b/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-neon.c > new file mode 100644 > index 00000000000..e9f6078cd12 > --- /dev/null > +++ b/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-neon.c > @@ -0,0 +1,16 @@ > +/* { dg-do compile } */ > +/* { dg-options "-O3 -march=armv8-a+simd -ftree-vectorize > -mveclibabi=sleefgnu -ffast-math" } */ > + > +extern float sinf(float); > + > +float x[256]; > + > +void foo(void) > +{ > + int i; > + > + for (i=0; i<256; ++i) > + x[i] = sinf(x[i]); > +} > + > +/* { dg-final { scan-assembler "_ZGVnN4v_sinf" } } */ > diff --git a/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-sve.c > b/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-sve.c > new file mode 100644 > index 00000000000..8319ae420e1 > --- /dev/null > +++ b/gcc/testsuite/gcc.target/aarch64/vect-vecabi-sleefgnu-sve.c > @@ -0,0 +1,16 @@ > +/* { dg-do compile } */ > +/* { dg-options "-O3 -march=armv8-a+sve -ftree-vectorize > -mveclibabi=sleefgnu -ffast-math" } */ > + > +extern float sinf(float); > + > +float x[256]; > + > +void foo(void) > +{ > + int i; > + > + for (i=0; i<256; ++i) > + x[i] = sinf(x[i]); > +} > + > +/* { dg-final { scan-assembler "_ZGVsNxv_sinf" } } */ > -- > 2.25.1 >