clear_user() used MOVQ+MOVB and if MOVQ faults, code simply exits and
honestly returns remaining length. In case of unaligned area, unaligned
remainder would count towards return value (correctly) but not cleared
(lazy code at least):

        clear_user(p + 4096 - 4, 8) = 8

No one would have noticed but REP MOVSB addition to clear_user()
repertoire creates a problem: REP MOVSB does everything correctly,
clears and counts to the last possible byte, but REP STOSQ and MOVQ
variants DO NOT:

        MOVQ            clear_user(p + 4096 - 4, 8) = 8
        REP STOSQ       clear_user(p + 4096 - 4, 8) = 8
        REP STOSB       clear_user(p + 4096 - 4, 8) = 4

Patch fixes incomplete clear on 32-bit and 64-bit REP STOSQ, MOVQ.

Signed-off-by: Alexey Dobriyan <adobri...@gmail.com>
---

 arch/x86/lib/clear_user_64.S |    9 ++++++---
 arch/x86/lib/usercopy_32.c   |    3 ++-
 2 files changed, 8 insertions(+), 4 deletions(-)

--- a/arch/x86/lib/clear_user_64.S
+++ b/arch/x86/lib/clear_user_64.S
@@ -26,7 +26,8 @@ ENTRY(__clear_user)
 
        .section .fixup,"ax"
 4:     lea     (%rsi,%rcx,8),%rcx
-       jmp     3b
+       # Fill as much as possible with byte stores.
+       jmp     2b
        .previous
 
        _ASM_EXTABLE(1b,4b)
@@ -57,7 +58,8 @@ ENTRY(__clear_user_movq)
 3:
        movb    $0, (%rdi)
        add     $1, %rdi
-       sub     $1, %ecx
+       # Unaligned area and 4GB+ tail after recovery require RCX here.
+       sub     $1, %rcx
        jnz     3b
 4:
        mov     %rcx, %rax
@@ -66,7 +68,8 @@ ENTRY(__clear_user_movq)
 
        .section .fixup,"ax"
 5:     lea     (%rsi,%rcx,8),%rcx
-       jmp     4b
+       # Fill as much as possible with byte stores.
+       jmp     3b
        .previous
 
        _ASM_EXTABLE(1b,5b)
--- a/arch/x86/lib/usercopy_32.c
+++ b/arch/x86/lib/usercopy_32.c
@@ -49,7 +49,8 @@ do {                                                          
        \
                "2: " ASM_CLAC "\n"                                     \
                ".section .fixup,\"ax\"\n"                              \
                "3:     lea 0(%2,%0,4),%0\n"                            \
-               "       jmp 2b\n"                                       \
+               /* Fill as much as possible with byte stores. */        \
+               "       jmp 1b\n"                                       \
                ".previous\n"                                           \
                _ASM_EXTABLE(0b,3b)                                     \
                _ASM_EXTABLE(1b,2b)                                     \
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to