On Thu, 22 Aug 2019 at 16:44, Richard Biener <richard.guent...@gmail.com> wrote:
>
> On Wed, Aug 21, 2019 at 8:24 PM Prathamesh Kulkarni
> <prathamesh.kulka...@linaro.org> wrote:
> >
> > On Thu, 15 Aug 2019 at 01:50, Richard Sandiford
> > <richard.sandif...@arm.com> wrote:
> > >
> > > Richard Biener <richard.guent...@gmail.com> writes:
> > > > On Wed, Aug 14, 2019 at 6:49 PM Richard Biener
> > > > <richard.guent...@gmail.com> wrote:
> > > >>
> > > >> On Wed, Aug 14, 2019 at 5:06 PM Prathamesh Kulkarni
> > > >> <prathamesh.kulka...@linaro.org> wrote:
> > > >> >
> > > >> > Hi,
> > > >> > The attached patch tries to fix PR86753.
> > > >> >
> > > >> > For following test:
> > > >> > void
> > > >> > f1 (int *restrict x, int *restrict y, int *restrict z)
> > > >> > {
> > > >> >   for (int i = 0; i < 100; ++i)
> > > >> >     x[i] = y[i] ? z[i] : 10;
> > > >> > }
> > > >> >
> > > >> > vect dump shows:
> > > >> >   vect_cst__42 = { 0, ... };
> > > >> >   vect_cst__48 = { 0, ... };
> > > >> >
> > > >> >   vect__4.7_41 = .MASK_LOAD (vectp_y.5_38, 4B, loop_mask_40);
> > > >> >   _4 = *_3;
> > > >> >   _5 = z_12(D) + _2;
> > > >> >   mask__35.8_43 = vect__4.7_41 != vect_cst__42;
> > > >> >   _35 = _4 != 0;
> > > >> >   vec_mask_and_46 = mask__35.8_43 & loop_mask_40;
> > > >> >   vect_iftmp.11_47 = .MASK_LOAD (vectp_z.9_44, 4B, vec_mask_and_46);
> > > >> >   iftmp.0_13 = 0;
> > > >> >   vect_iftmp.12_50 = VEC_COND_EXPR <vect__4.7_41 != vect_cst__48,
> > > >> > vect_iftmp.11_47, vect_cst__49>;
> > > >> >
> > > >> > and following code-gen:
> > > >> > L2:
> > > >> >         ld1w    z0.s, p2/z, [x1, x3, lsl 2]
> > > >> >         cmpne   p1.s, p3/z, z0.s, #0
> > > >> >         cmpne   p0.s, p2/z, z0.s, #0
> > > >> >         ld1w    z0.s, p0/z, [x2, x3, lsl 2]
> > > >> >         sel     z0.s, p1, z0.s, z1.s
> > > >> >
> > > >> > We could reuse vec_mask_and_46 in vec_cond_expr since the conditions
> > > >> > vect__4.7_41 != vect_cst__48 and vect__4.7_41 != vect_cst__42
> > > >> > are equivalent, and vect_iftmp.11_47 depends on vect__4.7_41 != 
> > > >> > vect_cst__48.
> > > >> >
> > > >> > I suppose in general for vec_cond_expr <C, T, E> if T comes from 
> > > >> > masked load,
> > > >> > which is conditional on C, then we could reuse the mask used in load,
> > > >> > in vec_cond_expr ?
> > > >> >
> > > >> > The patch maintains a hash_map cond_to_vec_mask
> > > >> > from <cond, loop_mask -> vec_mask (with loop predicate applied).
> > > >> > In prepare_load_store_mask, we record <cond, loop_mask> -> vec_mask 
> > > >> > & loop_mask,
> > > >> > and in vectorizable_condition, we check if <cond, loop_mask> exists 
> > > >> > in
> > > >> > cond_to_vec_mask
> > > >> > and if found, the corresponding vec_mask is used as 1st operand of
> > > >> > vec_cond_expr.
> > > >> >
> > > >> > <cond, loop_mask> is represented with cond_vmask_key, and the patch
> > > >> > adds tree_cond_ops to represent condition operator and operands 
> > > >> > coming
> > > >> > either from cond_expr
> > > >> > or a gimple comparison stmt. If the stmt is not comparison, it 
> > > >> > returns
> > > >> > <ne_expr, lhs, 0> and inserts that into cond_to_vec_mask.
> > > >> >
> > > >> > With patch, the redundant p1 is eliminated and sel uses p0 for above 
> > > >> > test.
> > > >> >
> > > >> > For following test:
> > > >> > void
> > > >> > f2 (int *restrict x, int *restrict y, int *restrict z, int fallback)
> > > >> > {
> > > >> >   for (int i = 0; i < 100; ++i)
> > > >> >     x[i] = y[i] ? z[i] : fallback;
> > > >> > }
> > > >> >
> > > >> > input to vectorizer has operands swapped in cond_expr:
> > > >> >   _36 = _4 != 0;
> > > >> >   iftmp.0_14 = .MASK_LOAD (_5, 32B, _36);
> > > >> >   iftmp.0_8 = _4 == 0 ? fallback_12(D) : iftmp.0_14;
> > > >> >
> > > >> > So we need to check for inverted condition in cond_to_vec_mask,
> > > >> > and swap the operands.
> > > >> > Does the patch look OK so far ?
> > > >> >
> > > >> > One major issue remaining with the patch is value  numbering.
> > > >> > Currently, it does value numbering for entire function using sccvn
> > > >> > during start of vect pass, which is too expensive since we only need
> > > >> > block based VN. I am looking into that.
> > > >>
> > > >> Why do you need it at all?  We run VN on the if-converted loop bodies 
> > > >> btw.
> > >
> > > This was my suggestion, but with the idea being to do the numbering
> > > per-statement as we vectorise.  We'll then see pattern statements too.
> > >
> > > That's important because we use pattern statements to set the right
> > > vector boolean type (e.g. vect_recog_mask_conversion_pattern).
> > > So some of the masks we care about don't exist after if converison.
> > >
> > > > Also I can't trivially see the equality of the masks and probably so
> > > > can't VN.  Is it that we just don't bother to apply loop_mask to
> > > > VEC_COND but there's no harm if we do?
> > >
> > > Yeah.  The idea of the optimisation is to decide when it's more profitable
> > > to apply the loop mask, even though doing so isn't necessary.  It would
> > > be hard to do after vectorisation because the masks aren't equivalent.
> > > We're relying on knowledge of how the vectoriser uses the result.
> > Hi,
> > Sorry for late response. This is an updated patch, that integrates
> > block-based VN into vect pass.
> > The patch
> > (a) Exports visit_stmt (renamed to vn_visit_stmt), vn_bb_init to
> > initialize VN state, and vn_bb_free to free it.
> > (b) Calls vn_visit_stmt in vect_transform_stmt for value numbering
> > stmts. We're only interested in obtaining
> > value numbers, not eliminating redundancies.
> > Does it look in the right direction ?
>
> It looks a bit odd to me.  I'd have expected it to work by generating
> the stmts as before in the vectorizer and then on the stmts we care
> invoke vn_visit_stmt that does both value-numbering and elimination.
> Alternatively you could ask the VN state to generate the stmt for
> you via vn_nary_build_or_lookup () (certainly that needs a bit more
> work).  One complication might be availability if you don't value-number
> all stmts in the block, but well.  I'm not sure constraining to a single
> block is necessary - I've thought of having a "CSE"ing gimple_build
> for some time (add & CSE new stmts onto a sequence), so one
> should keep this mode in mind when designing the one working on
> an existing BB.  Note as you write it it depends on visiting the
> stmts in proper order - is that guaranteed when for example
> vectorizing SLP?
Hi,
Indeed, I wrote the function with assumption that, stmts would be
visited in proper order.
This doesn't affect SLP currently, because call to vn_visit_stmt in
vect_transform_stmt is
conditional on cond_to_vec_mask, which is only allocated inside
vect_transform_loop.
But I agree we could make it more general.
AFAIU, the idea of constraining VN to single block was to avoid using defs from
non-dominating scalar stmts during outer-loop vectorization.

* fmla_2.c regression with patch:
This happens because with patch, forwprop4 is able to convert all 3
vec_cond_expr's
to .cond_fma(), which results in 3 calls to fmla, regressing the
test-case. If matching with
inverted condition is disabled in patch in vectorizable_condition,
then the old behavior gets preserved.

Thanks,
Prathamesh
>
> > I am not sure if the initialization in vn_bb_init is entirely correct.
> >
> > PS: The patch seems to regress fmla_2.c. I am looking into it.
> >
> > Thanks,
> > Prathamesh
> > >
> > > Thanks,
> > > Richard

Reply via email to