On Thu, Sep 14, 2023 at 5:12 AM Kewen Lin <li...@linux.ibm.com> wrote:
>
> This patch adjusts the cost handling on VMAT_CONTIGUOUS_PERMUTE
> in function vectorizable_store.  We don't call function
> vect_model_store_cost for it any more.  It's the case of
> interleaving stores, so it skips all stmts excepting for
> first_stmt_info, consider the whole group when costing
> first_stmt_info.  This patch shouldn't have any functional
> changes.

OK.

> gcc/ChangeLog:
>
>         * tree-vect-stmts.cc (vect_model_store_cost): Assert it will never
>         get VMAT_CONTIGUOUS_PERMUTE and remove VMAT_CONTIGUOUS_PERMUTE related
>         handlings.
>         (vectorizable_store): Adjust the cost handling on
>         VMAT_CONTIGUOUS_PERMUTE without calling vect_model_store_cost.
> ---
>  gcc/tree-vect-stmts.cc | 128 ++++++++++++++++++++++++-----------------
>  1 file changed, 74 insertions(+), 54 deletions(-)
>
> diff --git a/gcc/tree-vect-stmts.cc b/gcc/tree-vect-stmts.cc
> index fbd16b8a487..e3ba8077091 100644
> --- a/gcc/tree-vect-stmts.cc
> +++ b/gcc/tree-vect-stmts.cc
> @@ -967,10 +967,10 @@ vect_model_store_cost (vec_info *vinfo, stmt_vec_info 
> stmt_info, int ncopies,
>    gcc_assert (memory_access_type != VMAT_GATHER_SCATTER
>               && memory_access_type != VMAT_ELEMENTWISE
>               && memory_access_type != VMAT_STRIDED_SLP
> -             && memory_access_type != VMAT_LOAD_STORE_LANES);
> +             && memory_access_type != VMAT_LOAD_STORE_LANES
> +             && memory_access_type != VMAT_CONTIGUOUS_PERMUTE);
> +
>    unsigned int inside_cost = 0, prologue_cost = 0;
> -  stmt_vec_info first_stmt_info = stmt_info;
> -  bool grouped_access_p = STMT_VINFO_GROUPED_ACCESS (stmt_info);
>
>    /* ???  Somehow we need to fix this at the callers.  */
>    if (slp_node)
> @@ -983,35 +983,6 @@ vect_model_store_cost (vec_info *vinfo, stmt_vec_info 
> stmt_info, int ncopies,
>                                            stmt_info, 0, vect_prologue);
>      }
>
> -  /* Grouped stores update all elements in the group at once,
> -     so we want the DR for the first statement.  */
> -  if (!slp_node && grouped_access_p)
> -    first_stmt_info = DR_GROUP_FIRST_ELEMENT (stmt_info);
> -
> -  /* True if we should include any once-per-group costs as well as
> -     the cost of the statement itself.  For SLP we only get called
> -     once per group anyhow.  */
> -  bool first_stmt_p = (first_stmt_info == stmt_info);
> -
> -  /* We assume that the cost of a single store-lanes instruction is
> -     equivalent to the cost of DR_GROUP_SIZE separate stores.  If a grouped
> -     access is instead being provided by a permute-and-store operation,
> -     include the cost of the permutes.  */
> -  if (first_stmt_p
> -      && memory_access_type == VMAT_CONTIGUOUS_PERMUTE)
> -    {
> -      /* Uses a high and low interleave or shuffle operations for each
> -        needed permute.  */
> -      int group_size = DR_GROUP_SIZE (first_stmt_info);
> -      int nstmts = ncopies * ceil_log2 (group_size) * group_size;
> -      inside_cost = record_stmt_cost (cost_vec, nstmts, vec_perm,
> -                                     stmt_info, 0, vect_body);
> -
> -      if (dump_enabled_p ())
> -        dump_printf_loc (MSG_NOTE, vect_location,
> -                         "vect_model_store_cost: strided group_size = %d 
> .\n",
> -                         group_size);
> -    }
>
>    /* Costs of the stores.  */
>    vect_get_store_cost (vinfo, stmt_info, ncopies, alignment_support_scheme,
> @@ -8408,9 +8379,7 @@ vectorizable_store (vec_info *vinfo,
>          costing, use the first one instead.  */
>        if (grouped_store
>           && !slp
> -         && first_stmt_info != stmt_info
> -         && (memory_access_type == VMAT_ELEMENTWISE
> -             || memory_access_type == VMAT_LOAD_STORE_LANES))
> +         && first_stmt_info != stmt_info)
>         return true;
>      }
>    gcc_assert (memory_access_type == STMT_VINFO_MEMORY_ACCESS_TYPE 
> (stmt_info));
> @@ -9254,14 +9223,15 @@ vectorizable_store (vec_info *vinfo,
>        return true;
>      }
>
> +  unsigned inside_cost = 0, prologue_cost = 0;
>    auto_vec<tree> result_chain (group_size);
>    auto_vec<tree, 1> vec_oprnds;
>    for (j = 0; j < ncopies; j++)
>      {
>        gimple *new_stmt;
> -      if (j == 0 && !costing_p)
> +      if (j == 0)
>         {
> -         if (slp)
> +         if (slp && !costing_p)
>             {
>               /* Get vectorized arguments for SLP_NODE.  */
>               vect_get_vec_defs (vinfo, stmt_info, slp_node, 1, op,
> @@ -9287,13 +9257,20 @@ vectorizable_store (vec_info *vinfo,
>                      that there is no interleaving, DR_GROUP_SIZE is 1,
>                      and only one iteration of the loop will be executed.  */
>                   op = vect_get_store_rhs (next_stmt_info);
> -                 vect_get_vec_defs_for_operand (vinfo, next_stmt_info, 
> ncopies,
> -                                                op, gvec_oprnds[i]);
> -                 vec_oprnd = (*gvec_oprnds[i])[0];
> -                 dr_chain.quick_push (vec_oprnd);
> +                 if (costing_p
> +                     && memory_access_type == VMAT_CONTIGUOUS_PERMUTE)
> +                   update_prologue_cost (&prologue_cost, op);
> +                 else if (!costing_p)
> +                   {
> +                     vect_get_vec_defs_for_operand (vinfo, next_stmt_info,
> +                                                    ncopies, op,
> +                                                    gvec_oprnds[i]);
> +                     vec_oprnd = (*gvec_oprnds[i])[0];
> +                     dr_chain.quick_push (vec_oprnd);
> +                   }
>                   next_stmt_info = DR_GROUP_NEXT_ELEMENT (next_stmt_info);
>                 }
> -             if (mask)
> +             if (mask && !costing_p)
>                 {
>                   vect_get_vec_defs_for_operand (vinfo, stmt_info, ncopies,
>                                                  mask, &vec_masks,
> @@ -9303,11 +9280,13 @@ vectorizable_store (vec_info *vinfo,
>             }
>
>           /* We should have catched mismatched types earlier.  */
> -         gcc_assert (useless_type_conversion_p (vectype,
> -                                                TREE_TYPE (vec_oprnd)));
> +         gcc_assert (costing_p
> +                     || useless_type_conversion_p (vectype,
> +                                                   TREE_TYPE (vec_oprnd)));
>           bool simd_lane_access_p
>             = STMT_VINFO_SIMD_LANE_ACCESS_P (stmt_info) != 0;
> -         if (simd_lane_access_p
> +         if (!costing_p
> +             && simd_lane_access_p
>               && !loop_masks
>               && TREE_CODE (DR_BASE_ADDRESS (first_dr_info->dr)) == ADDR_EXPR
>               && VAR_P (TREE_OPERAND (DR_BASE_ADDRESS (first_dr_info->dr), 0))
> @@ -9319,7 +9298,7 @@ vectorizable_store (vec_info *vinfo,
>               dataref_ptr = unshare_expr (DR_BASE_ADDRESS 
> (first_dr_info->dr));
>               dataref_offset = build_int_cst (ref_type, 0);
>             }
> -         else
> +         else if (!costing_p)
>             dataref_ptr
>               = vect_create_data_ref_ptr (vinfo, first_stmt_info, aggr_type,
>                                           simd_lane_access_p ? loop : NULL,
> @@ -9347,16 +9326,46 @@ vectorizable_store (vec_info *vinfo,
>         }
>
>        new_stmt = NULL;
> -      if (!costing_p && grouped_store)
> -       /* Permute.  */
> -       vect_permute_store_chain (vinfo, dr_chain, group_size, stmt_info, gsi,
> -                                 &result_chain);
> +      if (grouped_store)
> +       {
> +         /* Permute.  */
> +         gcc_assert (memory_access_type == VMAT_CONTIGUOUS_PERMUTE);
> +         if (costing_p)
> +           {
> +             int group_size = DR_GROUP_SIZE (first_stmt_info);
> +             int nstmts = ceil_log2 (group_size) * group_size;
> +             inside_cost += record_stmt_cost (cost_vec, nstmts, vec_perm,
> +                                              stmt_info, 0, vect_body);
> +             if (dump_enabled_p ())
> +               dump_printf_loc (MSG_NOTE, vect_location,
> +                                "vect_model_store_cost: "
> +                                "strided group_size = %d .\n",
> +                                group_size);
> +           }
> +         else
> +           vect_permute_store_chain (vinfo, dr_chain, group_size, stmt_info,
> +                                     gsi, &result_chain);
> +       }
>
>        stmt_vec_info next_stmt_info = first_stmt_info;
>        for (i = 0; i < vec_num; i++)
>         {
>           if (costing_p)
> -           continue;
> +           {
> +             if (memory_access_type == VMAT_CONTIGUOUS_PERMUTE)
> +               vect_get_store_cost (vinfo, stmt_info, 1,
> +                                    alignment_support_scheme, misalignment,
> +                                    &inside_cost, cost_vec);
> +
> +             if (!slp)
> +               {
> +                 next_stmt_info = DR_GROUP_NEXT_ELEMENT (next_stmt_info);
> +                 if (!next_stmt_info)
> +                   break;
> +               }
> +
> +             continue;
> +           }
>           unsigned misalign;
>           unsigned HOST_WIDE_INT align;
>
> @@ -9540,9 +9549,20 @@ vectorizable_store (vec_info *vinfo,
>      }
>
>    if (costing_p)
> -    vect_model_store_cost (vinfo, stmt_info, ncopies, memory_access_type,
> -                          alignment_support_scheme, misalignment, vls_type,
> -                          slp_node, cost_vec);
> +    {
> +      if (memory_access_type == VMAT_CONTIGUOUS_PERMUTE)
> +       {
> +         if (dump_enabled_p ())
> +           dump_printf_loc (MSG_NOTE, vect_location,
> +                            "vect_model_store_cost: inside_cost = %d, "
> +                            "prologue_cost = %d .\n",
> +                            inside_cost, prologue_cost);
> +       }
> +      else
> +       vect_model_store_cost (vinfo, stmt_info, ncopies, memory_access_type,
> +                              alignment_support_scheme, misalignment, 
> vls_type,
> +                              slp_node, cost_vec);
> +    }
>
>    return true;
>  }
> --
> 2.31.1
>

Reply via email to