On Thu, 14 Dec 2023, Richard Sandiford wrote:

> Richard Biener <rguent...@suse.de> writes:
> > The following changes the unsigned group_size argument to a poly_uint64
> > one to avoid too much special-casing in callers for VLA vectors when
> > passing down the effective maximum desirable vector size to vector
> > type query routines.  The intent is to be able to pass down
> > the vectorization factor (times the SLP group size) eventually.
> >
> >     * tree-vectorizer.h (get_vectype_for_scalar_type,
> >     get_mask_type_for_scalar_type, vect_get_vector_types_for_stmt):
> >     Change group_size argument to poly_uint64 type.
> >     (vect_get_mask_type_for_stmt): Remove prototype for no longer
> >     existing function.
> >     * tree-vect-stmts.cc (get_vectype_for_scalar_type): Change
> >     group_size argument to poly_uint64.
> >     (get_mask_type_for_scalar_type): Likewise.
> >     (vect_get_vector_types_for_stmt): Likewise.
> 
> LGTM FWIW, although...
> 
> > ---
> >  gcc/tree-vect-stmts.cc | 25 ++++++++++++++-----------
> >  gcc/tree-vectorizer.h  |  7 +++----
> >  2 files changed, 17 insertions(+), 15 deletions(-)
> >
> > diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
> > index 88401a2a00b..a5e26b746fb 100644
> > --- a/gcc/tree-vect-stmts.cc
> > +++ b/gcc/tree-vect-stmts.cc
> > @@ -13297,14 +13297,14 @@ get_related_vectype_for_scalar_type (machine_mode 
> > prevailing_mode,
> >  
> >  tree
> >  get_vectype_for_scalar_type (vec_info *vinfo, tree scalar_type,
> > -                        unsigned int group_size)
> > +                        poly_uint64 group_size)
> >  {
> >    /* For BB vectorization, we should always have a group size once we've
> >       constructed the SLP tree; the only valid uses of zero GROUP_SIZEs
> >       are tentative requests during things like early data reference
> >       analysis and pattern recognition.  */
> >    if (is_a <bb_vec_info> (vinfo))
> > -    gcc_assert (vinfo->slp_instances.is_empty () || group_size != 0);
> > +    gcc_assert (vinfo->slp_instances.is_empty () || known_ne (group_size, 
> > 0));
> >    else
> >      group_size = 0;
> >  
> > @@ -13320,9 +13320,11 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree 
> > scalar_type,
> >  
> >    /* If the natural choice of vector type doesn't satisfy GROUP_SIZE,
> >       try again with an explicit number of elements.  */
> > +  uint64_t cst_group_size;
> >    if (vectype
> > -      && group_size
> > -      && maybe_ge (TYPE_VECTOR_SUBPARTS (vectype), group_size))
> > +      && group_size.is_constant (&cst_group_size)
> > +      && cst_group_size != 0
> > +      && maybe_ge (TYPE_VECTOR_SUBPARTS (vectype), cst_group_size))
> >      {
> >        /* Start with the biggest number of units that fits within
> >      GROUP_SIZE and halve it until we find a valid vector type.
> 
> ...it feels like this makes sense for VLA too in some form, if we
> plan to keep it longer-term.  It's not a trivial adaption though,
> so would definitely be a separate patch.

I did wonder if it's even possible to have multiple VLA vector sizes
but then you do have these x2 and x4 modes which are exactly that,
VLA vectors with a multiple number of elements?

So practically the goal is to find a vector mode suitable for a given
vectorization factor rather than a "size".

Richard.


> Thanks,
> Richard
> 
> > @@ -13336,7 +13338,7 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree 
> > scalar_type,
> >      even though the group is not a multiple of that vector size.
> >      The BB vectorizer will then try to carve up the group into
> >      smaller pieces.  */
> > -      unsigned int nunits = 1 << floor_log2 (group_size);
> > +      unsigned int nunits = 1 << floor_log2 (cst_group_size);
> >        do
> >     {
> >       vectype = get_related_vectype_for_scalar_type (vinfo->vector_mode,
> > @@ -13372,7 +13374,7 @@ get_vectype_for_scalar_type (vec_info *vinfo, tree 
> > scalar_type, slp_tree node)
> >  
> >  tree
> >  get_mask_type_for_scalar_type (vec_info *vinfo, tree scalar_type,
> > -                          unsigned int group_size)
> > +                          poly_uint64 group_size)
> >  {
> >    tree vectype = get_vectype_for_scalar_type (vinfo, scalar_type, 
> > group_size);
> >  
> > @@ -14243,7 +14245,7 @@ opt_result
> >  vect_get_vector_types_for_stmt (vec_info *vinfo, stmt_vec_info stmt_info,
> >                             tree *stmt_vectype_out,
> >                             tree *nunits_vectype_out,
> > -                           unsigned int group_size)
> > +                           poly_uint64 group_size)
> >  {
> >    gimple *stmt = stmt_info->stmt;
> >  
> > @@ -14252,7 +14254,7 @@ vect_get_vector_types_for_stmt (vec_info *vinfo, 
> > stmt_vec_info stmt_info,
> >       are tentative requests during things like early data reference
> >       analysis and pattern recognition.  */
> >    if (is_a <bb_vec_info> (vinfo))
> > -    gcc_assert (vinfo->slp_instances.is_empty () || group_size != 0);
> > +    gcc_assert (vinfo->slp_instances.is_empty () || known_ne (group_size, 
> > 0));
> >    else
> >      group_size = 0;
> >  
> > @@ -14281,7 +14283,7 @@ vect_get_vector_types_for_stmt (vec_info *vinfo, 
> > stmt_vec_info stmt_info,
> >  
> >    tree vectype;
> >    tree scalar_type = NULL_TREE;
> > -  if (group_size == 0 && STMT_VINFO_VECTYPE (stmt_info))
> > +  if (known_eq (group_size, 0U) && STMT_VINFO_VECTYPE (stmt_info))
> >      {
> >        vectype = STMT_VINFO_VECTYPE (stmt_info);
> >        if (dump_enabled_p ())
> > @@ -14310,10 +14312,11 @@ vect_get_vector_types_for_stmt (vec_info *vinfo, 
> > stmt_vec_info stmt_info,
> >  
> >        if (dump_enabled_p ())
> >     {
> > -     if (group_size)
> > +     if (known_ne (group_size, 0U))
> >         dump_printf_loc (MSG_NOTE, vect_location,
> >                          "get vectype for scalar type (group size %d):"
> > -                        " %T\n", group_size, scalar_type);
> > +                        " %T\n", (int)constant_lower_bound (group_size),
> > +                        scalar_type);
> >       else
> >         dump_printf_loc (MSG_NOTE, vect_location,
> >                          "get vectype for scalar type: %T\n", scalar_type);
> > diff --git a/gcc/tree-vectorizer.h b/gcc/tree-vectorizer.h
> > index a2bab8676af..95240504d18 100644
> > --- a/gcc/tree-vectorizer.h
> > +++ b/gcc/tree-vectorizer.h
> > @@ -2204,9 +2204,9 @@ extern edge vec_init_loop_exit_info (class loop *);
> >  /* In tree-vect-stmts.cc.  */
> >  extern tree get_related_vectype_for_scalar_type (machine_mode, tree,
> >                                              poly_uint64 = 0);
> > -extern tree get_vectype_for_scalar_type (vec_info *, tree, unsigned int = 
> > 0);
> > +extern tree get_vectype_for_scalar_type (vec_info *, tree, poly_uint64 = 
> > 0);
> >  extern tree get_vectype_for_scalar_type (vec_info *, tree, slp_tree);
> > -extern tree get_mask_type_for_scalar_type (vec_info *, tree, unsigned int 
> > = 0);
> > +extern tree get_mask_type_for_scalar_type (vec_info *, tree, poly_uint64 = 
> > 0);
> >  extern tree get_mask_type_for_scalar_type (vec_info *, tree, slp_tree);
> >  extern tree get_same_sized_vectype (tree, tree);
> >  extern bool vect_chooses_same_modes_p (vec_info *, machine_mode);
> > @@ -2295,8 +2295,7 @@ extern tree vect_gen_while (gimple_seq *, tree, tree, 
> > tree,
> >  extern tree vect_gen_while_not (gimple_seq *, tree, tree, tree);
> >  extern opt_result vect_get_vector_types_for_stmt (vec_info *,
> >                                               stmt_vec_info, tree *,
> > -                                             tree *, unsigned int = 0);
> > -extern opt_tree vect_get_mask_type_for_stmt (stmt_vec_info, unsigned int = 
> > 0);
> > +                                             tree *, poly_uint64 = 0);
> >  
> >  /* In tree-vect-data-refs.cc.  */
> >  extern bool vect_can_force_dr_alignment_p (const_tree, poly_uint64);
> 

-- 
Richard Biener <rguent...@suse.de>
SUSE Software Solutions Germany GmbH,
Frankenstrasse 146, 90461 Nuernberg, Germany;
GF: Ivo Totev, Andrew McDonald, Werner Knoblich; (HRB 36809, AG Nuernberg)

Reply via email to