[PATCH 2/2] x86: fix incomplete clear by clear_user()
From: Alexey Dobriyan
Date: Sat Jun 20 2015 - 17:47:28 EST
clear_user() used MOVQ+MOVB and if MOVQ faults, code simply exits and
honestly returns remaining length. In case of unaligned area, unaligned
remainder would count towards return value (correctly) but not cleared
(lazy code at least):
clear_user(p + 4096 - 4, 8) = 8
No one would have noticed but REP MOVSB addition to clear_user()
repertoire creates a problem: REP MOVSB does everything correctly,
clears and counts to the last possible byte, but REP STOSQ and MOVQ
variants DO NOT:
MOVQ clear_user(p + 4096 - 4, 8) = 8
REP STOSQ clear_user(p + 4096 - 4, 8) = 8
REP STOSB clear_user(p + 4096 - 4, 8) = 4
Patch fixes incomplete clear on 32-bit and 64-bit REP STOSQ, MOVQ.
Signed-off-by: Alexey Dobriyan <adobriyan@xxxxxxxxx>
---
arch/x86/lib/clear_user_64.S | 9 ++++++---
arch/x86/lib/usercopy_32.c | 3 ++-
2 files changed, 8 insertions(+), 4 deletions(-)
--- a/arch/x86/lib/clear_user_64.S
+++ b/arch/x86/lib/clear_user_64.S
@@ -26,7 +26,8 @@ ENTRY(__clear_user)
.section .fixup,"ax"
4: lea (%rsi,%rcx,8),%rcx
- jmp 3b
+ # Fill as much as possible with byte stores.
+ jmp 2b
.previous
_ASM_EXTABLE(1b,4b)
@@ -57,7 +58,8 @@ ENTRY(__clear_user_movq)
3:
movb $0, (%rdi)
add $1, %rdi
- sub $1, %ecx
+ # Unaligned area and 4GB+ tail after recovery require RCX here.
+ sub $1, %rcx
jnz 3b
4:
mov %rcx, %rax
@@ -66,7 +68,8 @@ ENTRY(__clear_user_movq)
.section .fixup,"ax"
5: lea (%rsi,%rcx,8),%rcx
- jmp 4b
+ # Fill as much as possible with byte stores.
+ jmp 3b
.previous
_ASM_EXTABLE(1b,5b)
--- a/arch/x86/lib/usercopy_32.c
+++ b/arch/x86/lib/usercopy_32.c
@@ -49,7 +49,8 @@ do { \
"2: " ASM_CLAC "\n" \
".section .fixup,\"ax\"\n" \
"3: lea 0(%2,%0,4),%0\n" \
- " jmp 2b\n" \
+ /* Fill as much as possible with byte stores. */ \
+ " jmp 1b\n" \
".previous\n" \
_ASM_EXTABLE(0b,3b) \
_ASM_EXTABLE(1b,2b) \
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
Please read the FAQ at http://www.tux.org/lkml/