summaryrefslogtreecommitdiffstats
path: root/arch/x86/crypto/aegis128l-aesni-asm.S
diff options
context:
space:
mode:
authorLinus Torvalds2018-08-13 22:35:26 +0200
committerLinus Torvalds2018-08-13 22:35:26 +0200
commitf24d6f2654d39355cdf8285e21409ed8d56d4284 (patch)
treee6d2c683e61f30147bf73eba8d9fbf2c05865f03 /arch/x86/crypto/aegis128l-aesni-asm.S
parentMerge branch 'x86-boot-for-linus' of git://git.kernel.org/pub/scm/linux/kerne... (diff)
parentx86/entry/64: Add two more instruction suffixes (diff)
downloadkernel-qcow2-linux-f24d6f2654d39355cdf8285e21409ed8d56d4284.tar.gz
kernel-qcow2-linux-f24d6f2654d39355cdf8285e21409ed8d56d4284.tar.xz
kernel-qcow2-linux-f24d6f2654d39355cdf8285e21409ed8d56d4284.zip
Merge branch 'x86-asm-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86 asm updates from Thomas Gleixner: "The lowlevel and ASM code updates for x86: - Make stack trace unwinding more reliable - ASM instruction updates for better code generation - Various cleanups" * 'x86-asm-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86/entry/64: Add two more instruction suffixes x86/asm/64: Use 32-bit XOR to zero registers x86/build/vdso: Simplify 'cmd_vdso2c' x86/build/vdso: Remove unused vdso-syms.lds x86/stacktrace: Enable HAVE_RELIABLE_STACKTRACE for the ORC unwinder x86/unwind/orc: Detect the end of the stack x86/stacktrace: Do not fail for ORC with regs on stack x86/stacktrace: Clarify the reliable success paths x86/stacktrace: Remove STACKTRACE_DUMP_ONCE x86/stacktrace: Do not unwind after user regs x86/asm: Use CC_SET/CC_OUT in percpu_cmpxchg8b_double() to micro-optimize code generation
Diffstat (limited to 'arch/x86/crypto/aegis128l-aesni-asm.S')
-rw-r--r--arch/x86/crypto/aegis128l-aesni-asm.S2
1 files changed, 1 insertions, 1 deletions
diff --git a/arch/x86/crypto/aegis128l-aesni-asm.S b/arch/x86/crypto/aegis128l-aesni-asm.S
index 4eda2b8db9e1..491dd61c845c 100644
--- a/arch/x86/crypto/aegis128l-aesni-asm.S
+++ b/arch/x86/crypto/aegis128l-aesni-asm.S
@@ -66,7 +66,7 @@
* %r9
*/
__load_partial:
- xor %r9, %r9
+ xor %r9d, %r9d
pxor MSG0, MSG0
pxor MSG1, MSG1