On Fri, 28 Jul 2023, Juzhe-Zhong wrote:

> Hi, Richard and Richi.
> 
> Base on the suggestions from Richard:
> https://gcc.gnu.org/pipermail/gcc-patches/2023-July/625396.html
> 
> This patch choose (1) approach that Richard provided, meaning:
> 
> RVV implements cond_* optabs as expanders.  RVV therefore supports
> both IFN_COND_ADD and IFN_COND_LEN_ADD.  No dummy length arguments
> are needed at the gimple level.
> 
> Such approach can make codes much cleaner and reasonable.
> 
> Consider this following case:
> void foo (float * __restrict a, float * __restrict b, int * __restrict cond, 
> int n)
> {
>   for (int i = 0; i < n; i++)
>     if (cond[i])
>       a[i] = b[i] + a[i];
> }
> 
> 
> Output of RISC-V (32-bits) gcc (trunk) (Compiler #3)
> <source>:5:21: missed: couldn't vectorize loop
> <source>:5:21: missed: not vectorized: control flow in loop.
> 
> ARM SVE:
> 
> ...
> mask__27.10_51 = vect__4.9_49 != { 0, ... };
> ...
> vec_mask_and_55 = loop_mask_49 & mask__27.10_51;
> ...
> vect__9.17_62 = .COND_ADD (vec_mask_and_55, vect__6.13_56, vect__8.16_60, 
> vect__6.13_56);
> 
> For RVV, we want IR as follows:
> 
> ...
> _68 = .SELECT_VL (ivtmp_66, POLY_INT_CST [4, 4]);
> ...
> mask__27.10_51 = vect__4.9_49 != { 0, ... };
> ...
> vect__9.17_60 = .COND_LEN_ADD (mask__27.10_51, vect__6.13_55, vect__8.16_59, 
> vect__6.13_55, _68, 0);
> ...
> 
> Both len and mask of COND_LEN_ADD are real not dummy.
> 
> This patch has been fully tested in RISC-V port with supporting both COND_* 
> and COND_LEN_*.
> 
> And also, Bootstrap and Regression on X86 passed.
> 
> OK for trunk?
> 
> gcc/ChangeLog:
> 
>         * internal-fn.cc (FOR_EACH_LEN_FN_PAIR): New macro.
>         (get_len_internal_fn): New function.
>         (CASE): Ditto.
>         * internal-fn.h (get_len_internal_fn): Ditto.
>         * tree-vect-stmts.cc (vectorizable_call): Support CALL vectorization 
> with COND_LEN_*.
> 
> ---
>  gcc/internal-fn.cc     | 46 ++++++++++++++++++++++
>  gcc/internal-fn.h      |  1 +
>  gcc/tree-vect-stmts.cc | 87 +++++++++++++++++++++++++++++++++++++-----
>  3 files changed, 125 insertions(+), 9 deletions(-)
> 
> diff --git a/gcc/internal-fn.cc b/gcc/internal-fn.cc
> index 8e294286388..379220bebc7 100644
> --- a/gcc/internal-fn.cc
> +++ b/gcc/internal-fn.cc
> @@ -4443,6 +4443,52 @@ get_conditional_internal_fn (internal_fn fn)
>      }
>  }
>  
> +/* Invoke T(IFN) for each internal function IFN that also has an
> +   IFN_COND_LEN_* or IFN_MASK_LEN_* form.  */
> +#define FOR_EACH_LEN_FN_PAIR(T)                                              
>   \
> +  T (MASK_LOAD, MASK_LEN_LOAD)                                               
>   \
> +  T (MASK_STORE, MASK_LEN_STORE)                                             
>   \
> +  T (MASK_GATHER_LOAD, MASK_LEN_GATHER_LOAD)                                 
>   \
> +  T (MASK_SCATTER_STORE, MASK_LEN_SCATTER_STORE)                             
>   \
> +  T (COND_ADD, COND_LEN_ADD)                                                 
>   \
> +  T (COND_SUB, COND_LEN_SUB)                                                 
>   \
> +  T (COND_MUL, COND_LEN_MUL)                                                 
>   \
> +  T (COND_DIV, COND_LEN_DIV)                                                 
>   \
> +  T (COND_MOD, COND_LEN_MOD)                                                 
>   \
> +  T (COND_RDIV, COND_LEN_RDIV)                                               
>   \
> +  T (COND_FMIN, COND_LEN_FMIN)                                               
>   \
> +  T (COND_FMAX, COND_LEN_FMAX)                                               
>   \
> +  T (COND_MIN, COND_LEN_MIN)                                                 
>   \
> +  T (COND_MAX, COND_LEN_MAX)                                                 
>   \
> +  T (COND_AND, COND_LEN_AND)                                                 
>   \
> +  T (COND_IOR, COND_LEN_IOR)                                                 
>   \
> +  T (COND_XOR, COND_LEN_XOR)                                                 
>   \
> +  T (COND_SHL, COND_LEN_SHL)                                                 
>   \
> +  T (COND_SHR, COND_LEN_SHR)                                                 
>   \
> +  T (COND_NEG, COND_LEN_NEG)                                                 
>   \
> +  T (COND_FMA, COND_LEN_FMA)                                                 
>   \
> +  T (COND_FMS, COND_LEN_FMS)                                                 
>   \
> +  T (COND_FNMA, COND_LEN_FNMA)                                               
>   \
> +  T (COND_FNMS, COND_LEN_FNMS)
> +
> +/* If there exists an internal function like IFN that operates on vectors,
> +   but with additional length and bias parameters, return the internal_fn
> +   for that function, otherwise return IFN_LAST.  */
> +internal_fn
> +get_len_internal_fn (internal_fn fn)
> +{
> +  switch (fn)
> +    {
> +#define CASE(NAME, LEN_NAME)                                                 
>   \
> +  case IFN_##NAME:                                                           
>   \
> +    return IFN_##LEN_NAME;
> +      FOR_EACH_LEN_FN_PAIR (CASE)
> +#undef CASE
> +    default:
> +      return IFN_LAST;
> +    }
> +}
> +
>  /* If IFN implements the conditional form of an unconditional internal
>     function, return that unconditional function, otherwise return IFN_LAST.  
> */
>  
> diff --git a/gcc/internal-fn.h b/gcc/internal-fn.h
> index a5c3f4765ff..410c1b623d6 100644
> --- a/gcc/internal-fn.h
> +++ b/gcc/internal-fn.h
> @@ -224,6 +224,7 @@ extern bool set_edom_supported_p (void);
>  
>  extern internal_fn get_conditional_internal_fn (tree_code);
>  extern internal_fn get_conditional_internal_fn (internal_fn);
> +extern internal_fn get_len_internal_fn (internal_fn);
>  extern internal_fn get_conditional_len_internal_fn (tree_code);
>  extern tree_code conditional_internal_fn_code (internal_fn);
>  extern internal_fn get_unconditional_internal_fn (internal_fn);
> diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
> index 6a4e8fce126..ae5b0b09c08 100644
> --- a/gcc/tree-vect-stmts.cc
> +++ b/gcc/tree-vect-stmts.cc
> @@ -3540,7 +3540,10 @@ vectorizable_call (vec_info *vinfo,
>  
>    int reduc_idx = STMT_VINFO_REDUC_IDX (stmt_info);
>    internal_fn cond_fn = get_conditional_internal_fn (ifn);
> +  internal_fn cond_len_fn = get_len_internal_fn (ifn);
> +  int len_opno = internal_fn_len_index (cond_len_fn);
>    vec_loop_masks *masks = (loop_vinfo ? &LOOP_VINFO_MASKS (loop_vinfo) : 
> NULL);
> +  vec_loop_lens *lens = (loop_vinfo ? &LOOP_VINFO_LENS (loop_vinfo) : NULL);
>    if (!vec_stmt) /* transformation not required.  */
>      {
>        if (slp_node)
> @@ -3586,8 +3589,14 @@ vectorizable_call (vec_info *vinfo,

Above for reduc_idx >= 0 there's a check whether cond_fn is supported,
don't you need to amend that with a check for cond_len_fn?

>             tree scalar_mask = NULL_TREE;
>             if (mask_opno >= 0)
>               scalar_mask = gimple_call_arg (stmt_info->stmt, mask_opno);
> -           vect_record_loop_mask (loop_vinfo, masks, nvectors,
> -                                  vectype_out, scalar_mask);
> +           if (cond_len_fn != IFN_LAST
> +               && direct_internal_fn_supported_p (cond_len_fn, vectype_out,
> +                                                  OPTIMIZE_FOR_SPEED))
> +             vect_record_loop_len (loop_vinfo, lens, nvectors, vectype_out,
> +                                   1);
> +           else
> +             vect_record_loop_mask (loop_vinfo, masks, nvectors, vectype_out,
> +                                    scalar_mask);
>           }
>       }
>        return true;
> @@ -3603,8 +3612,24 @@ vectorizable_call (vec_info *vinfo,
>    vec_dest = vect_create_destination_var (scalar_dest, vectype_out);
>  
>    bool masked_loop_p = loop_vinfo && LOOP_VINFO_FULLY_MASKED_P (loop_vinfo);
> +  bool len_loop_p = loop_vinfo && LOOP_VINFO_FULLY_WITH_LENGTH_P 
> (loop_vinfo);
>    unsigned int vect_nargs = nargs;
> -  if (masked_loop_p && reduc_idx >= 0)
> +  if (len_loop_p)
> +    {
> +      if (len_opno >= 0)
> +     {
> +       ifn = cond_len_fn;
> +       /* COND_* -> COND_LEN_* takes 2 extra arguments:LEN,BIAS.  */
> +       vect_nargs += 2;
> +     }
> +      else if (reduc_idx >= 0)
> +     {
> +       /* FMA -> COND_LEN_FMA takes 4 extra arguments:MASK,ELSE,LEN,BIAS.  */
> +       ifn = get_len_internal_fn (cond_fn);
> +       vect_nargs += 4;

I'm a bit confused (but also by the existing mask code), whether
vect_nargs needs adjustment depends on the IFN in the IL we analyze.
If if-conversion recognizes a .COND_ADD then we need to add nothing
for masking (that is, ifn == cond_fn already).  In your code above
you either use cond_len_fn or get_len_internal_fn (cond_fn) but
isn't that the very same?!  So how come you in one case add two
and in the other add four args?

Please make sure to place gcc_unreachable () in each arm and check
you have test coverage.  I believe that the else arm is unreachable
but when you vectorize .FMA you will need to add 4 and when you
vectorize .COND_FMA you will need to add two arguments (as said,
no idea why we special case reduc_idx >= 0 at the moment).

Otherwise the patch looks OK to me.

Thanks,
Richard.

> +     }
> +    }
> +  else if (masked_loop_p && reduc_idx >= 0)
>      {
>        ifn = cond_fn;
>        vect_nargs += 2;
> @@ -3629,7 +3654,18 @@ vectorizable_call (vec_info *vinfo,
>             FOR_EACH_VEC_ELT (vec_oprnds0, i, vec_oprnd0)
>               {
>                 int varg = 0;
> -               if (masked_loop_p && reduc_idx >= 0)
> +               if (len_loop_p && reduc_idx >= 0)
> +                 {
> +                   /* Always true for SLP.  */
> +                   gcc_assert (ncopies == 1);
> +                   /* For COND_LEN_* operations used by reduction of
> +                      CALL vectorization, the LEN argument is the real
> +                      loop len produced by SELECT_VL or MIN wheras the
> +                      MASK argument here is the dummy mask.  */
> +                   vargs[varg++]
> +                     = build_minus_one_cst (truth_type_for (vectype_out));
> +                 }
> +               else if (masked_loop_p && reduc_idx >= 0)
>                   {
>                     unsigned int vec_num = vec_oprnds0.length ();
>                     /* Always true for SLP.  */
> @@ -3644,7 +3680,7 @@ vectorizable_call (vec_info *vinfo,
>                     vec<tree> vec_oprndsk = vec_defs[k];
>                     vargs[varg++] = vec_oprndsk[i];
>                   }
> -               if (masked_loop_p && reduc_idx >= 0)
> +               if ((masked_loop_p || len_loop_p) && reduc_idx >= 0)
>                   vargs[varg++] = vargs[reduc_idx + 1];
>                 gimple *new_stmt;
>                 if (modifier == NARROW)
> @@ -3671,7 +3707,21 @@ vectorizable_call (vec_info *vinfo,
>                   }
>                 else
>                   {
> -                   if (mask_opno >= 0 && masked_loop_p)
> +                   if (len_opno >= 0 && len_loop_p)
> +                     {
> +                       unsigned int vec_num = vec_oprnds0.length ();
> +                       /* Always true for SLP.  */
> +                       gcc_assert (ncopies == 1);
> +                       tree len
> +                         = vect_get_loop_len (loop_vinfo, gsi, lens, vec_num,
> +                                              vectype_out, i, 1);
> +                       signed char biasval
> +                         = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
> +                       tree bias = build_int_cst (intQI_type_node, biasval);
> +                       vargs[len_opno] = len;
> +                       vargs[len_opno + 1] = bias;
> +                     }
> +                   else if (mask_opno >= 0 && masked_loop_p)
>                       {
>                         unsigned int vec_num = vec_oprnds0.length ();
>                         /* Always true for SLP.  */
> @@ -3701,7 +3751,16 @@ vectorizable_call (vec_info *vinfo,
>           }
>  
>         int varg = 0;
> -       if (masked_loop_p && reduc_idx >= 0)
> +       if (len_loop_p && reduc_idx >= 0)
> +         {
> +           /* For COND_LEN_* operations used by reduction of
> +              CALL vectorization, the LEN argument is the real
> +              loop len produced by SELECT_VL or MIN wheras the
> +              MASK argument here is the dummy mask.  */
> +           vargs[varg++]
> +             = build_minus_one_cst (truth_type_for (vectype_out));
> +         }
> +       else if (masked_loop_p && reduc_idx >= 0)
>           vargs[varg++] = vect_get_loop_mask (loop_vinfo, gsi, masks, ncopies,
>                                               vectype_out, j);
>         for (i = 0; i < nargs; i++)
> @@ -3716,10 +3775,20 @@ vectorizable_call (vec_info *vinfo,
>               }
>             vargs[varg++] = vec_defs[i][j];
>           }
> -       if (masked_loop_p && reduc_idx >= 0)
> +       if ((masked_loop_p || len_loop_p) && reduc_idx >= 0)
>           vargs[varg++] = vargs[reduc_idx + 1];
>  
> -       if (mask_opno >= 0 && masked_loop_p)
> +       if (len_opno >= 0 && len_loop_p)
> +         {
> +           tree len = vect_get_loop_len (loop_vinfo, gsi, lens, ncopies,
> +                                         vectype_out, j, 1);
> +           signed char biasval
> +             = LOOP_VINFO_PARTIAL_LOAD_STORE_BIAS (loop_vinfo);
> +           tree bias = build_int_cst (intQI_type_node, biasval);
> +           vargs[len_opno] = len;
> +           vargs[len_opno + 1] = bias;
> +         }
> +       else if (mask_opno >= 0 && masked_loop_p)
>           {
>             tree mask = vect_get_loop_mask (loop_vinfo, gsi, masks, ncopies,
>                                             vectype_out, j);
> 

-- 
Richard Biener <rguent...@suse.de>
SUSE Software Solutions Germany GmbH,
Frankenstrasse 146, 90461 Nuernberg, Germany;
GF: Ivo Totev, Andrew McDonald, Werner Knoblich; (HRB 36809, AG Nuernberg)

Reply via email to