summaryrefslogtreecommitdiffstats
path: root/arch/x86/boot/compressed/head_32.S
diff options
context:
space:
mode:
authorH. Peter Anvin <hpa@zytor.com>2009-05-08 16:45:15 -0700
committerH. Peter Anvin <hpa@zytor.com>2009-05-08 17:34:52 -0700
commit36d3793c947f1ef7ba3d24eeeddc1be41adc5ab4 (patch)
treebf5987e4caacb7aab5a78ca683a3179bd8eba717 /arch/x86/boot/compressed/head_32.S
parent97541912785369925723b6255438ad9fce2ddf04 (diff)
downloadlinux-36d3793c947f1ef7ba3d24eeeddc1be41adc5ab4.tar.gz
linux-36d3793c947f1ef7ba3d24eeeddc1be41adc5ab4.tar.bz2
linux-36d3793c947f1ef7ba3d24eeeddc1be41adc5ab4.zip
x86, boot: use appropriate rep string for move and clear
In the pre-decompression code, use the appropriate largest possible rep movs and rep stos to move code and clear bss, respectively. For reverse copy, do note that the initial values are supposed to be the address of the first (highest) copy datum, not one byte beyond the end of the buffer. rep strings are not necessarily the fastest way to perform these operations on all current processors, but are likely to be in the future, and perhaps more importantly, we want to encourage the architecturally right thing to do here. This also fixes a couple of trivial inefficiencies on 64 bits. [ Impact: trivial performance enhancement, increase code similarity ] Signed-off-by: H. Peter Anvin <hpa@zytor.com>
Diffstat (limited to 'arch/x86/boot/compressed/head_32.S')
-rw-r--r--arch/x86/boot/compressed/head_32.S11
1 files changed, 6 insertions, 5 deletions
diff --git a/arch/x86/boot/compressed/head_32.S b/arch/x86/boot/compressed/head_32.S
index d02a4f02be13..6710dc78ac59 100644
--- a/arch/x86/boot/compressed/head_32.S
+++ b/arch/x86/boot/compressed/head_32.S
@@ -100,11 +100,12 @@ ENTRY(startup_32)
* where decompression in place becomes safe.
*/
pushl %esi
- leal _bss(%ebp), %esi
- leal _bss(%ebx), %edi
+ leal (_bss-4)(%ebp), %esi
+ leal (_bss-4)(%ebx), %edi
movl $(_bss - startup_32), %ecx
+ shrl $2, %ecx
std
- rep movsb
+ rep movsl
cld
popl %esi
@@ -135,8 +136,8 @@ relocated:
leal _bss(%ebx), %edi
leal _ebss(%ebx), %ecx
subl %edi, %ecx
- cld
- rep stosb
+ shrl $2, %ecx
+ rep stosl
/*
* Do the decompression, and jump to the new kernel..