I have another idea for sse, and this one is far safer: only use sse prefetch, leave the string operations for the actual copy. The prefetch operations only prefetch, don't touch the sse registers, thus neither any reentency nor interrupt problems. I tried the attached hack^H^H^H^Hpatch, and read(fd, buf, 4000000) from user space got 7% faster (from 264768842 cycles to 246303748 cycles, single cpu, noacpi, 'linux -b', fastest time from several thousand runs). The reason why this works is simple: Intel Pentium III and P 4 have hardcoded "fast stringcopy" operations that invalidate whole cachelines during write (documented in the most obvious place: multiprocessor management, memory ordering) The result is a very fast write, but the read is still slow. -- Manfred
--- 2.4/mm/filemap.c Wed Feb 14 10:51:42 2001 +++ build-2.4/mm/filemap.c Wed Feb 14 22:11:44 2001 @@ -1248,6 +1248,20 @@ size = count; kaddr = kmap(page); + if (size > 128) { + int i; + __asm__ __volatile__( + "mov %1, %0\n\t" + : "=r" (i) + : "r" (kaddr+offset)); /* load tlb entry */ + for(i=0;i<size;i+=64) { + __asm__ __volatile__( + "prefetchnta (%1, %0)\n\t" + "prefetchnta 32(%1, %0)\n\t" + : /* no output */ + : "r" (i), "r" (kaddr+offset)); + } + } left = __copy_to_user(desc->buf, kaddr + offset, size); kunmap(page);