https://bugs.llvm.org/show_bug.cgi?id=43381

            Bug ID: 43381
           Summary: Suboptimal shift+mask code gen with BZHI
           Product: libraries
           Version: trunk
          Hardware: PC
                OS: All
            Status: NEW
          Severity: enhancement
          Priority: P
         Component: Backend: X86
          Assignee: unassignedb...@nondot.org
          Reporter: d...@znu.io
                CC: craig.top...@gmail.com, llvm-bugs@lists.llvm.org,
                    llvm-...@redking.me.uk, spatel+l...@rotateright.com

The two C examples at the end of this report should generate the following
assembly when BZHI is available:
```
        movb    $60, %cl
        bzhiq   %rcx, (%rdi), %rax
        shrq    $23, %rax
```

In practice they generate:
```
        movq    (%rdi), %rax
        shrq    $23, %rax
        movb    $37, %cl
        bzhiq   %rcx, %rax, %rax
```

The example C code:
```
unsigned long example1(unsigned long *mem) {
        unsigned long temp = *mem & ((1ul << 60) - 1);
        return temp >> 23;
}

unsigned long example2(unsigned long *mem) {
        unsigned long temp = *mem >> 23;
        return temp & ((1ul << 37) - 1);
}
```

It seems to me that LLVM and/or the backend is biased towards "shift then mask"
but sometimes "mask then shift" generates better code.

-- 
You are receiving this mail because:
You are on the CC list for the bug.
_______________________________________________
llvm-bugs mailing list
llvm-bugs@lists.llvm.org
https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-bugs

Reply via email to