The barrier_nospec() in 64-bit __get_user() is slow.  Instead use
pointer masking to force the user pointer to all 1's if a previous
access_ok() mispredicted true for an invalid address.

Note that for safety on some AMD CPUs, this relies on recent commit
86e6b1547b3d ("x86: fix user address masking non-canonical speculation
issue").

Signed-off-by: Josh Poimboeuf <jpoim...@kernel.org>
---
 arch/x86/lib/getuser.S | 24 ++++++++++++++++++++----
 1 file changed, 20 insertions(+), 4 deletions(-)

diff --git a/arch/x86/lib/getuser.S b/arch/x86/lib/getuser.S
index 4357ec2a0bfc..998d5be6b794 100644
--- a/arch/x86/lib/getuser.S
+++ b/arch/x86/lib/getuser.S
@@ -112,8 +112,12 @@ EXPORT_SYMBOL(__get_user_8)
 
 /* .. and the same for __get_user, just without the range checks */
 SYM_FUNC_START(__get_user_nocheck_1)
-       ASM_STAC
+#ifdef CONFIG_X86_64
+       check_range size=1
+#else
        ASM_BARRIER_NOSPEC
+#endif
+       ASM_STAC
        UACCESS movzbl (%_ASM_AX),%edx
        xor %eax,%eax
        ASM_CLAC
@@ -122,8 +126,12 @@ SYM_FUNC_END(__get_user_nocheck_1)
 EXPORT_SYMBOL(__get_user_nocheck_1)
 
 SYM_FUNC_START(__get_user_nocheck_2)
-       ASM_STAC
+#ifdef CONFIG_X86_64
+       check_range size=2
+#else
        ASM_BARRIER_NOSPEC
+#endif
+       ASM_STAC
        UACCESS movzwl (%_ASM_AX),%edx
        xor %eax,%eax
        ASM_CLAC
@@ -132,8 +140,12 @@ SYM_FUNC_END(__get_user_nocheck_2)
 EXPORT_SYMBOL(__get_user_nocheck_2)
 
 SYM_FUNC_START(__get_user_nocheck_4)
-       ASM_STAC
+#ifdef CONFIG_X86_64
+       check_range size=4
+#else
        ASM_BARRIER_NOSPEC
+#endif
+       ASM_STAC
        UACCESS movl (%_ASM_AX),%edx
        xor %eax,%eax
        ASM_CLAC
@@ -142,8 +154,12 @@ SYM_FUNC_END(__get_user_nocheck_4)
 EXPORT_SYMBOL(__get_user_nocheck_4)
 
 SYM_FUNC_START(__get_user_nocheck_8)
-       ASM_STAC
+#ifdef CONFIG_X86_64
+       check_range size=8
+#else
        ASM_BARRIER_NOSPEC
+#endif
+       ASM_STAC
 #ifdef CONFIG_X86_64
        UACCESS movq (%_ASM_AX),%rdx
 #else
-- 
2.47.0


Reply via email to