On Mon, Mar 7, 2022 at 11:04 AM Roger Sayle <ro...@nextmovesoftware.com> wrote: > > > This patch is the main middle-end piece of a fix for PR tree-opt/98335, > which is a code-quality regression affecting mainline. The issue occurs > in DSE's (dead store elimination's) compute_trims function that determines > where a store to memory can be trimmed. In the testcase given in the > PR, this function notices that the first byte of a DImode store is dead, > and replaces the 8-byte store at (aligned) offset zero, with a 7-byte store > at (unaligned) offset one. Most architectures can store a power-of-two > bytes (up to a maximum) in single instruction, so writing 7 bytes requires > more instructions than writing 8 bytes. This patch follows Jakub Jelinek's > suggestion in comment 5, that compute_trims needs improved heuristics. > > In this patch, decision of whether and how to align trim_head is based > on the number of bytes being written, the alignment of the start of the > object and where within the object the first byte is written. The first > tests check whether we're already writing to the start of the object, > and that we're writing three or more bytes. If we're only writing one > or two bytes, there's no benefit from providing additional alignment. > Then we determine the alignment of the object, which is either 1, 2, > 4, 8 or 16 byte aligned (capping at 16 guarantees that we never write > more than 7 bytes beyond the minimum required). If the buffer is only > 1 or 2 byte aligned there's no benefit from additional alignment. For > the remaining cases, alignment of trim_head is based upon where within > each aligned block (word) the first byte is written. For example, > storing the last byte (or last half-word) of a word can be performed > with a single insn. > > On x86_64-pc-linux-gnu with -O2 the new test case in the PR goes from: > > movl $0, -24(%rsp) > movabsq $72057594037927935, %rdx > movl $0, -21(%rsp) > andq -24(%rsp), %rdx > movq %rdx, %rax > salq $8, %rax > movb c(%rip), %al > ret > > to > > xorl %eax, %eax > movb c(%rip), %al > ret > > This patch has been tested on x86_64-pc-linux-gnu with make bootstrap > and make -k check with no new failures. I've also added new testcases > for the original motivating PR tree-optimization/86010, to ensure that > those remain optimized (in future). Ok for mainline?
diff --git a/gcc/tree-ssa-dse.cc b/gcc/tree-ssa-dse.cc index 2b22a61..080e406 100644 --- a/gcc/tree-ssa-dse.cc +++ b/gcc/tree-ssa-dse.cc @@ -405,10 +405,36 @@ compute_trims (ao_ref *ref, sbitmap live, int *trim_head, int *trim_tail, int first_live = bitmap_first_set_bit (live); *trim_head = first_live - first_orig; - /* If more than a word remains, then make sure to keep the - starting point at least word aligned. */ - if (last_live - first_live > UNITS_PER_WORD) - *trim_head &= ~(UNITS_PER_WORD - 1); + /* If REF is aligned, try to maintain this alignment if it reduces + the number of (power-of-two sized aligned) writes to memory. + First check that we're writing >= 3 bytes at a non-zero offset. */ + if (first_live + && last_live - first_live >= 2) + { + unsigned int align = TYPE_ALIGN_UNIT (TREE_TYPE (ref->base)); you can't simply use TYPE_ALIGN_* on ref->base. You can use get_object_alignment on ref->ref, but ref->ref can be NULL in case the ref was initialized from a builtin call like memcpy. Also ref->base is offsetted by ref->offset which you don't seem to account for. In theory one could export get_object_alignment_2 and if ref->ref is NULL, use that on ref->base, passing addr_p = true, and then adjust the resulting bitpos by ref->offset and fix align accordingly (trimming might also align an access if the original access was offsetted from known alignment). That said, a helper like ao_ref_alignment () might be useful here. I wonder if we can apply good heuristics to compute_trims without taking into account context, like maybe_trimp_complex_store is already limiting itself to useful subsets and the constructor and memstar cases will only benefit if they end up being expanded inline via *_by_pieces, not if expanded as a call. You don't seem to adjust *trim_tail at all, if an aligned 16 byte region is trimmed there by 3 that will result in two extra stores as well, no? + if (DECL_P (ref->base) && DECL_ALIGN_UNIT (ref->base) > align) + align = DECL_ALIGN_UNIT (ref->base); + if (align > UNITS_PER_WORD) + align = UNITS_PER_WORD; + if (align > 16) + align = 16; + if (align > 2) + { + /* ALIGN is 4, 8 or 16. */ + unsigned int low = first_live & (align - 1); + if (low * 2 < align) + { + if (align == 16 && low >= 4 && last_live < 15) + *trim_head &= ~3; + else + *trim_head &= ~(align - 1); + } + else if (low + 3 == align) + *trim_head &= ~1; + else if (low > 8 && low < 12) + *trim_head &= ~3; + } + } if ((*trim_head || *trim_tail) && dump_file && (dump_flags & TDF_DETAILS)) > 2022-03-07 Roger Sayle <ro...@nextmovesoftware.com> > > gcc/ChangeLog > PR tree-optimization/98335 > * tree-ssa-dse.cc (compute_trims): Improve logic deciding whether > to align trim_head, writing more bytes by using fewer instructions. > > gcc/testsuite/ChangeLog > PR tree-optimization/98335 > * g++.dg/pr98335.C: New test case. > * gcc.dg/pr86010.c: New test case. > * gcc.dg/pr86010-2.c: New test case. > > Thanks in advance, > Roger > -- >