Backport fixes for correct register constraints in cmpxchg.h (rhbz 809014)

This commit is contained in:
Josh Boyer 2012-04-10 09:11:21 -04:00
parent 446e33ca6d
commit 961ef7f50d
3 changed files with 79 additions and 1 deletions

View File

@ -42,7 +42,7 @@ Summary: The Linux kernel
# When changing real_sublevel below, reset this by hand to 1
# (or to 0 and then use rpmdev-bumpspec).
#
%global baserelease 3
%global baserelease 4
%global fedora_build %{baserelease}
# real_sublevel is the 3.x kernel version we're starting with
@ -705,6 +705,10 @@ Patch21380: wimax-i2400m-prevent-a-possible-kernel-bug-due-to-mi.patch
#rhbz 806676 807632
Patch21385: libata-disable-runtime-pm-for-hotpluggable-port.patch
#rhbz 809014
Patch21390: x86-Use-correct-byte-sized-register-constraint-in-__xchg_op.patch
Patch21391: x86-Use-correct-byte-sized-register-constraint-in-__add.patch
Patch21501: nfs-Fix-length-of-buffer-copied-in-__nfs4_get_acl_uncached.patch
#rhbz 808207 CVE-2012-1601
@ -1314,6 +1318,10 @@ ApplyPatch wimax-i2400m-prevent-a-possible-kernel-bug-due-to-mi.patch
#rhbz 806676 807632
ApplyPatch libata-disable-runtime-pm-for-hotpluggable-port.patch
#rhbz 809014
ApplyPatch x86-Use-correct-byte-sized-register-constraint-in-__xchg_op.patch
ApplyPatch x86-Use-correct-byte-sized-register-constraint-in-__add.patch
# END OF PATCH APPLICATIONS
%endif
@ -1961,6 +1969,9 @@ fi
# and build.
%changelog
* Tue Apr 10 2012 Josh Boyer <jwboyer@redhat.com>
- Backport fixes for correct register constraints in cmpxchg.h (rhbz 809014)
* Thu Apr 05 2012 Dave Jones <davej@redhat.com>
- Better watermark the number of pages used by hibernation I/O (Bojan Smojver) (rhbz 785384)

View File

@ -0,0 +1,35 @@
From: H. Peter Anvin <hpa@zytor.com>
Date: Fri, 6 Apr 2012 16:30:57 +0000 (-0700)
Subject: x86: Use correct byte-sized register constraint in __add()
X-Git-Url: http://git.kernel.org/?p=linux%2Fkernel%2Fgit%2Ftip%2Ftip.git;a=commitdiff_plain;h=8c91c5325e107ec17e40a59a47c6517387d64eb7
x86: Use correct byte-sized register constraint in __add()
Similar to:
2ca052a x86: Use correct byte-sized register constraint in __xchg_op()
... the __add() macro also needs to use a "q" constraint in the
byte-sized case, lest we try to generate an illegal register.
Link: http://lkml.kernel.org/r/4F7A3315.501@goop.org
Signed-off-by: H. Peter Anvin <hpa@zytor.com>
Cc: Jeremy Fitzhardinge <jeremy@goop.org>
Cc: Leigh Scott <leigh123linux@googlemail.com>
Cc: Thomas Reitmayr <treitmayr@devbase.at>
Cc: <stable@vger.kernel.org> v3.3
---
diff --git a/arch/x86/include/asm/cmpxchg.h b/arch/x86/include/asm/cmpxchg.h
index bc18d0e..99480e5 100644
--- a/arch/x86/include/asm/cmpxchg.h
+++ b/arch/x86/include/asm/cmpxchg.h
@@ -173,7 +173,7 @@ extern void __add_wrong_size(void)
switch (sizeof(*(ptr))) { \
case __X86_CASE_B: \
asm volatile (lock "addb %b1, %0\n" \
- : "+m" (*(ptr)) : "ri" (inc) \
+ : "+m" (*(ptr)) : "qi" (inc) \
: "memory", "cc"); \
break; \
case __X86_CASE_W: \

View File

@ -0,0 +1,32 @@
From: Jeremy Fitzhardinge <jeremy@goop.org>
Date: Mon, 2 Apr 2012 23:15:33 +0000 (-0700)
Subject: x86: Use correct byte-sized register constraint in __xchg_op()
X-Git-Url: http://git.kernel.org/?p=linux%2Fkernel%2Fgit%2Ftip%2Ftip.git;a=commitdiff_plain;h=2ca052a3710fac208eee690faefdeb8bbd4586a1
x86: Use correct byte-sized register constraint in __xchg_op()
x86-64 can access the low half of any register, but i386 can only do
it with a subset of registers. 'r' causes compilation failures on i386,
but 'q' expresses the constraint properly.
Signed-off-by: Jeremy Fitzhardinge <jeremy@goop.org>
Link: http://lkml.kernel.org/r/4F7A3315.501@goop.org
Reported-by: Leigh Scott <leigh123linux@googlemail.com>
Tested-by: Thomas Reitmayr <treitmayr@devbase.at>
Signed-off-by: H. Peter Anvin <hpa@zytor.com>
Cc: <stable@vger.kernel.org> v3.3
---
diff --git a/arch/x86/include/asm/cmpxchg.h b/arch/x86/include/asm/cmpxchg.h
index b3b7332..bc18d0e 100644
--- a/arch/x86/include/asm/cmpxchg.h
+++ b/arch/x86/include/asm/cmpxchg.h
@@ -43,7 +43,7 @@ extern void __add_wrong_size(void)
switch (sizeof(*(ptr))) { \
case __X86_CASE_B: \
asm volatile (lock #op "b %b0, %1\n" \
- : "+r" (__ret), "+m" (*(ptr)) \
+ : "+q" (__ret), "+m" (*(ptr)) \
: : "memory", "cc"); \
break; \
case __X86_CASE_W: \