Eric Botcazou <ebotca...@adacore.com> writes:
>> This patch replaces:
>> 
>>       /* Stop if the mode is wider than the alignment of the containing
>>       object.
>> 
>>       It is tempting to omit the following line unless STRICT_ALIGNMENT
>>       is true.  But that is incorrect, since if the bitfield uses part
>>       of 3 bytes and we use a 4-byte mode, we could get a spurious segv
>>       if the extra 4th byte is past the end of memory.
>>       (Though at least one Unix compiler ignores this problem:
>>       that on the Sequent 386 machine.  */
>>       if (unit > align_)
>>      break;
>> 
>> with two checks: one for whether the final byte of the mode is known
>> to be mapped, and one for whether the bitfield is sufficiently aligned.
>> Later patches in the series rely on this in order not to pessimise
>> memory handling.
>> 
>> However, as described in the patch, I found that extending this
>> behaviour to get_best_mode affected code generation for x86_64-linux-gnu
>> and powerpc64-linux-gnu, and led to a failure in bb-slp-5.c on both.
>> I therefore chickened out and added the original check back to
>> get_best_mode.
>> 
>> I'm certainly interested in any feedback on the comment, but at the
>> same time I'd like this series to be a no-op on targets that keep
>> the traditional .md patterns.  Any change to get_best_mode is probably
>> best done as a separate patch.
>
> I agree with your conservative approach here.
>
>> gcc/
>>      * stor-layout.c (bit_field_mode_iterator::bit_field_mode_iterator):
>>      Set up a default value of bitregion_end_.
>>      (bit_field_mode_iterator::next_mode): Always apply bitregion_end_
>>      check.  Include SLOW_UNALIGNED_ACCESS in the alignment check.
>>      (get_best_mode): Ignore modes that are wider than the alignment.
>
> Fine with me, modulo:
>
>> @@ -2754,6 +2753,62 @@ get_best_mode (int bitsize, int bitpos,
>>    enum machine_mode widest_mode = VOIDmode;
>>    enum machine_mode mode;
>>    while (iter.next_mode (&mode)
>> +     /* ??? For compatiblity, reject modes that are wider than the
>> +        alignment.  This has both advantages and disadvantages.
>
> "For compatibility" is a bit vague (compatibility with what?).  I'd write 
> "For 
> historical reasons" or something along these lines.

Yeah, that's better.  Here's what I committed after updating to the patch
for the inclusive bitregion_end_.

Richard


gcc/
        * stor-layout.c (bit_field_mode_iterator::bit_field_mode_iterator):
        Set up a default value of bitregion_end_.
        (bit_field_mode_iterator::next_mode): Always apply bitregion_end_
        check.  Include SLOW_UNALIGNED_ACCESS in the alignment check.
        (get_best_mode): Ignore modes that are wider than the alignment.

Index: gcc/stor-layout.c
===================================================================
--- gcc/stor-layout.c   2012-11-18 17:26:31.159273478 +0000
+++ gcc/stor-layout.c   2012-11-18 17:28:17.481177253 +0000
@@ -2646,6 +2646,13 @@ fixup_unsigned_type (tree type)
   bitregion_end_ (bitregion_end), align_ (MIN (align, BIGGEST_ALIGNMENT)),
   volatilep_ (volatilep), count_ (0)
 {
+  if (!bitregion_end_)
+    {
+      /* We can assume that any aligned chunk of ALIGN_ bits that overlaps
+        the bitfield is mapped and won't trap.  */
+      bitregion_end_ = bitpos + bitsize + align_ - 1;
+      bitregion_end_ -= bitregion_end_ % align_ + 1;
+    }
 }
 
 /* Calls to this function return successively larger modes that can be used
@@ -2676,23 +2683,15 @@ bit_field_mode_iterator::next_mode (enum
       if (count_ > 0 && unit > BITS_PER_WORD)
        break;
 
-      /* Stop if the mode is wider than the alignment of the containing
-        object.
-
-        It is tempting to omit the following line unless STRICT_ALIGNMENT
-        is true.  But that is incorrect, since if the bitfield uses part
-        of 3 bytes and we use a 4-byte mode, we could get a spurious segv
-        if the extra 4th byte is past the end of memory.
-        (Though at least one Unix compiler ignores this problem:
-        that on the Sequent 386 machine.  */
-      if (unit > align_)
-       break;
-
       /* Stop if the mode goes outside the bitregion.  */
       HOST_WIDE_INT start = bitpos_ - (bitpos_ % unit);
       if (bitregion_start_ && start < bitregion_start_)
        break;
-      if (bitregion_end_ && start + unit > bitregion_end_ + 1)
+      if (start + unit > bitregion_end_ + 1)
+       break;
+
+      /* Stop if the mode requires too much alignment.  */
+      if (unit > align_ && SLOW_UNALIGNED_ACCESS (mode_, align_))
        break;
 
       *out_mode = mode_;
@@ -2751,6 +2750,62 @@ get_best_mode (int bitsize, int bitpos,
   enum machine_mode widest_mode = VOIDmode;
   enum machine_mode mode;
   while (iter.next_mode (&mode)
+        /* ??? For historical reasons, reject modes that are wider than
+           the alignment.  This has both advantages and disadvantages.
+           Removing this check means that something like:
+
+              struct s { unsigned int x; unsigned int y; };
+              int f (struct s *s) { return s->x == 0 && s->y == 0; }
+
+           can be implemented using a single load and compare on
+           64-bit machines that have no alignment restrictions.
+           For example, on powerpc64-linux-gnu, we would generate:
+
+                   ld 3,0(3)
+                   cntlzd 3,3
+                   srdi 3,3,6
+                   blr
+
+           rather than:
+
+                   lwz 9,0(3)
+                   cmpwi 7,9,0
+                   bne 7,.L3
+                   lwz 3,4(3)
+                   cntlzw 3,3
+                   srwi 3,3,5
+                   extsw 3,3
+                   blr
+                   .p2align 4,,15
+           .L3:
+                   li 3,0
+                   blr
+
+           However, accessing more than one field can make life harder
+           for the gimple optimizers.  For example, gcc.dg/vect/bb-slp-5.c
+           has a series of unsigned short copies followed by a series of
+           unsigned short comparisons.  With this check, both the copies
+           and comparisons remain 16-bit accesses and FRE is able
+           to eliminate the latter.  Without the check, the comparisons
+           can be done using 2 64-bit operations, which FRE isn't able
+           to handle in the same way.
+
+           Either way, it would probably be worth disabling this check
+           during expand.  One particular example where removing the
+           check would help is the get_best_mode call in store_bit_field.
+           If we are given a memory bitregion of 128 bits that is aligned
+           to a 64-bit boundary, and the bitfield we want to modify is
+           in the second half of the bitregion, this check causes
+           store_bitfield to turn the memory into a 64-bit reference
+           to the _first_ half of the region.  We later use
+           adjust_bitfield_address to get a reference to the correct half,
+           but doing so looks to adjust_bitfield_address as though we are
+           moving past the end of the original object, so it drops the
+           associated MEM_EXPR and MEM_OFFSET.  Removing the check
+           causes store_bit_field to keep a 128-bit memory reference,
+           so that the final bitfield reference still has a MEM_EXPR
+           and MEM_OFFSET.  */
+        && GET_MODE_BITSIZE (mode) <= align
         && (largest_mode == VOIDmode
             || GET_MODE_SIZE (mode) <= GET_MODE_SIZE (largest_mode)))
     {

Reply via email to