diff options
| author | Tony Luck <tony.luck@intel.com> | 2016-02-18 11:47:26 -0800 | 
|---|---|---|
| committer | Ingo Molnar <mingo@kernel.org> | 2016-03-08 17:54:38 +0100 | 
| commit | 92b0729c34cab1f46d89aace3e66015f0bb4a682 (patch) | |
| tree | c1509bb8d1dc7044bb2d68a280799e4ccdca2da9 /arch/x86/lib | |
| parent | ea2ca36b658cfc6081ee454e97593c81f646806e (diff) | |
x86/mm, x86/mce: Add memcpy_mcsafe()
Make use of the EXTABLE_FAULT exception table entries to write
a kernel copy routine that doesn't crash the system if it
encounters a machine check. Prime use case for this is to copy
from large arrays of non-volatile memory used as storage.
We have to use an unrolled copy loop for now because current
hardware implementations treat a machine check in "rep mov"
as fatal. When that is fixed we can simplify.
Return type is a "bool". True means that we copied OK, false means
that it didn't.
Signed-off-by: Tony Luck <tony.luck@intel.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Andy Lutomirski <luto@amacapital.net>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Tony Luck <tony.luck@gmail.com>
Link: http://lkml.kernel.org/r/a44e1055efc2d2a9473307b22c91caa437aa3f8b.1456439214.git.tony.luck@intel.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'arch/x86/lib')
| -rw-r--r-- | arch/x86/lib/memcpy_64.S | 117 | 
1 files changed, 117 insertions, 0 deletions
diff --git a/arch/x86/lib/memcpy_64.S b/arch/x86/lib/memcpy_64.S index 16698bba87de..7d37641ada5b 100644 --- a/arch/x86/lib/memcpy_64.S +++ b/arch/x86/lib/memcpy_64.S @@ -177,3 +177,120 @@ ENTRY(memcpy_orig)  .Lend:  	retq  ENDPROC(memcpy_orig) + +#ifndef CONFIG_UML +/* + * memcpy_mcsafe - memory copy with machine check exception handling + * Note that we only catch machine checks when reading the source addresses. + * Writes to target are posted and don't generate machine checks. + */ +ENTRY(memcpy_mcsafe) +	cmpl $8, %edx +	/* Less than 8 bytes? Go to byte copy loop */ +	jb .L_no_whole_words + +	/* Check for bad alignment of source */ +	testl $7, %esi +	/* Already aligned */ +	jz .L_8byte_aligned + +	/* Copy one byte at a time until source is 8-byte aligned */ +	movl %esi, %ecx +	andl $7, %ecx +	subl $8, %ecx +	negl %ecx +	subl %ecx, %edx +.L_copy_leading_bytes: +	movb (%rsi), %al +	movb %al, (%rdi) +	incq %rsi +	incq %rdi +	decl %ecx +	jnz .L_copy_leading_bytes + +.L_8byte_aligned: +	/* Figure out how many whole cache lines (64-bytes) to copy */ +	movl %edx, %ecx +	andl $63, %edx +	shrl $6, %ecx +	jz .L_no_whole_cache_lines + +	/* Loop copying whole cache lines */ +.L_cache_w0: movq (%rsi), %r8 +.L_cache_w1: movq 1*8(%rsi), %r9 +.L_cache_w2: movq 2*8(%rsi), %r10 +.L_cache_w3: movq 3*8(%rsi), %r11 +	movq %r8, (%rdi) +	movq %r9, 1*8(%rdi) +	movq %r10, 2*8(%rdi) +	movq %r11, 3*8(%rdi) +.L_cache_w4: movq 4*8(%rsi), %r8 +.L_cache_w5: movq 5*8(%rsi), %r9 +.L_cache_w6: movq 6*8(%rsi), %r10 +.L_cache_w7: movq 7*8(%rsi), %r11 +	movq %r8, 4*8(%rdi) +	movq %r9, 5*8(%rdi) +	movq %r10, 6*8(%rdi) +	movq %r11, 7*8(%rdi) +	leaq 64(%rsi), %rsi +	leaq 64(%rdi), %rdi +	decl %ecx +	jnz .L_cache_w0 + +	/* Are there any trailing 8-byte words? */ +.L_no_whole_cache_lines: +	movl %edx, %ecx +	andl $7, %edx +	shrl $3, %ecx +	jz .L_no_whole_words + +	/* Copy trailing words */ +.L_copy_trailing_words: +	movq (%rsi), %r8 +	mov %r8, (%rdi) +	leaq 8(%rsi), %rsi +	leaq 8(%rdi), %rdi +	decl %ecx +	jnz .L_copy_trailing_words + +	/* Any trailing bytes? */ +.L_no_whole_words: +	andl %edx, %edx +	jz .L_done_memcpy_trap + +	/* Copy trailing bytes */ +	movl %edx, %ecx +.L_copy_trailing_bytes: +	movb (%rsi), %al +	movb %al, (%rdi) +	incq %rsi +	incq %rdi +	decl %ecx +	jnz .L_copy_trailing_bytes + +	/* Copy successful. Return true */ +.L_done_memcpy_trap: +	xorq %rax, %rax +	ret +ENDPROC(memcpy_mcsafe) + +	.section .fixup, "ax" +	/* Return false for any failure */ +.L_memcpy_mcsafe_fail: +	mov	$1, %rax +	ret + +	.previous + +	_ASM_EXTABLE_FAULT(.L_copy_leading_bytes, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w0, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w1, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w3, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w3, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w4, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w5, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w6, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_cache_w7, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_copy_trailing_words, .L_memcpy_mcsafe_fail) +	_ASM_EXTABLE_FAULT(.L_copy_trailing_bytes, .L_memcpy_mcsafe_fail) +#endif  | 
