When generating code for x86 targets, this is able to simplify XOR+SETcc sequences. SETcc generates a setcond+deposit pair of TCG opcodes which used to become setcond+ext32u after optimization; now TCG recognizes that the output of setcond is itself already zero extended and turns the deposit into just a mov.
There are similar cases in fold_movcond and fold_setcond_zmask, but I couldn't trigger them and they require moving around functions to avoid forward references[1], so I am leaving them aside for now. [1] I assume the lack of forward references is intentional in order to avoid possible mutual recursion Signed-off-by: Paolo Bonzini <pbonz...@redhat.com> --- tcg/optimize.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/tcg/optimize.c b/tcg/optimize.c index e9ef16b3c6b..e0fdaeb5500 100644 --- a/tcg/optimize.c +++ b/tcg/optimize.c @@ -1620,7 +1620,7 @@ static bool fold_deposit(OptContext *ctx, TCGOp *op) op->args[1] = op->args[2]; op->args[2] = arg_new_constant(ctx, mask); ctx->z_mask = mask & arg_info(op->args[1])->z_mask; - return false; + return fold_and(ctx, op); } /* Inserting zero into a value. */ @@ -1630,7 +1630,7 @@ static bool fold_deposit(OptContext *ctx, TCGOp *op) op->opc = and_opc; op->args[2] = arg_new_constant(ctx, mask); ctx->z_mask = mask & arg_info(op->args[1])->z_mask; - return false; + return fold_and(ctx, op); } ctx->z_mask = deposit64(arg_info(op->args[1])->z_mask, -- 2.47.0