On Fri, 17 Aug 2007, Nick Piggin wrote:
> 
> I'm surprised too. Numbers were from the "...use asm() like the other
> atomic operations already do" thread. According to them,
> 
>   text    data     bss     dec     hex filename
> 3434150  249176  176128 3859454  3ae3fe atomic_normal/vmlinux
> 3436203  249176  176128 3861507  3aec03 atomic_volatile/vmlinux
> 
> The first one is a stock kenel, the second is with atomic_read/set
> cast to volatile. gcc-4.1 -- maybe if you have an earlier gcc it
> won't optimise as much?

No, see my earlier reply. "volatile" really *is* an incredible piece of 
crap.

Just try it yourself:

        volatile int i;
        int j;

        int testme(void)
        {
                return i <= 1;
        }

        int testme2(void)
        {
                return j <= 1;
        }

and compile with all the optimizations you can.

I get:

        testme:
                movl    i(%rip), %eax
                subl    $1, %eax
                setle   %al
                movzbl  %al, %eax
                ret

vs

        testme2:
                xorl    %eax, %eax
                cmpl    $1, j(%rip)
                setle   %al
                ret

(now, whether that "xorl + setle" is better than "setle + movzbl", I don't 
really know - maybe it is. But that's not thepoint. The point is the 
difference between

                movl    i(%rip), %eax
                subl    $1, %eax

and

                cmpl    $1, j(%rip)

and imagine this being done for *every* single volatile access.

Just do a 

        git grep atomic_read

to see how atomics are actually used. A lot of them are exactly the above 
kind of "compare against a value".

                        Linus
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to