On Wed, 19 May 2021 at 13:13, Richard Biener <rguent...@suse.de> wrote:
>
> On Wed, 19 May 2021, Jakub Jelinek wrote:
>
> > On Wed, May 19, 2021 at 11:09:19AM +0200, Jakub Jelinek via Gcc-patches 
> > wrote:
> > > On Wed, May 19, 2021 at 10:15:53AM +0200, Christophe Lyon via Gcc-patches 
> > > wrote:
> > > > After this update, the test fails on arm and aarch64: according to the
> > > > logs, the optimization is still performed 14 times.
> > >
> > > Seems this is because
> > >               if (change
> > >                   && !flag_syntax_only
> > >                   && (load_extend_op (TYPE_MODE (TREE_TYPE (and0)))
> > >                       == ZERO_EXTEND))
> > >                 {
> > >                   tree uns = unsigned_type_for (TREE_TYPE (and0));
> > >                   and0 = fold_convert_loc (loc, uns, and0);
> > >                   and1 = fold_convert_loc (loc, uns, and1);
> > >                 }
> > > in fold-const.c adds on these targets extra casts that prevent the
> > > optimizations.
> >
> > This patch seems to fix it (but I don't have an easy way to test on aarch64
> > or arm on the trunk and 11 branch would need numerous backports).
>
> OK if somebody manages to test on arm/aarch64.
>
I confirm it fixes the problem on arm. (aarch64 in progress)

Thanks!

> Richard.
>
> > 2021-05-19  Jakub Jelinek  <ja...@redhat.com>
> >
> >       PR tree-optimization/94589
> >       * match.pd ((X & Y) == X -> (X & ~Y) == 0): Simplify even in presence
> >       of integral conversions.
> >
> > --- gcc/match.pd.jj   2021-05-15 10:10:28.000000000 +0200
> > +++ gcc/match.pd      2021-05-19 11:34:42.130624557 +0200
> > @@ -4769,6 +4769,16 @@ (define_operator_list COND_TERNARY
> >   (simplify
> >    (cmp:c (bit_and:c @0 @1) @0)
> >    (cmp (bit_and @0 (bit_not! @1)) { build_zero_cst (TREE_TYPE (@0)); }))
> > + (simplify
> > +  (cmp:c (convert@3 (bit_and (convert@2 @0) INTEGER_CST@1)) (convert @0))
> > +  (if (INTEGRAL_TYPE_P (TREE_TYPE (@0))
> > +       && INTEGRAL_TYPE_P (TREE_TYPE (@2))
> > +       && INTEGRAL_TYPE_P (TREE_TYPE (@3))
> > +       && TYPE_PRECISION (TREE_TYPE (@2)) == TYPE_PRECISION (TREE_TYPE 
> > (@0))
> > +       && TYPE_PRECISION (TREE_TYPE (@3)) > TYPE_PRECISION (TREE_TYPE (@2))
> > +       && !wi::neg_p (wi::to_wide (@1)))
> > +   (cmp (bit_and @0 (convert (bit_not @1)))
> > +     { build_zero_cst (TREE_TYPE (@0)); })))
> >
> >   /* (X | Y) == Y becomes (X & ~Y) == 0.  */
> >   (simplify
> >
> >
> >       Jakub

Reply via email to