summaryrefslogtreecommitdiff
path: root/sysdeps/i386/i586/lshift.S
diff options
context:
space:
mode:
Diffstat (limited to 'sysdeps/i386/i586/lshift.S')
-rw-r--r--sysdeps/i386/i586/lshift.S54
1 files changed, 30 insertions, 24 deletions
diff --git a/sysdeps/i386/i586/lshift.S b/sysdeps/i386/i586/lshift.S
index bf9b223a0b..1d72fc901e 100644
--- a/sysdeps/i386/i586/lshift.S
+++ b/sysdeps/i386/i586/lshift.S
@@ -1,5 +1,5 @@
/* Pentium optimized __mpn_lshift --
- Copyright (C) 1992, 1994, 1995, 1996 Free Software Foundation, Inc.
+ Copyright (C) 1992, 1994, 1995, 1996, 1997 Free Software Foundation, Inc.
This file is part of the GNU C Library.
The GNU C Library is free software; you can redistribute it and/or modify
@@ -44,15 +44,15 @@ C_SYMBOL_NAME(__mpn_lshift:)
/* We can use faster code for shift-by-1 under certain conditions. */
cmp $1,%ecx
- jne Lnormal
+ jne L(normal)
leal 4(%esi),%eax
cmpl %edi,%eax
- jnc Lspecial /* jump if s_ptr + 1 >= res_ptr */
+ jnc L(special) /* jump if s_ptr + 1 >= res_ptr */
leal (%esi,%ebp,4),%eax
cmpl %eax,%edi
- jnc Lspecial /* jump if res_ptr >= s_ptr + size */
+ jnc L(special) /* jump if res_ptr >= s_ptr + size */
-Lnormal:
+L(normal):
leal -4(%edi,%ebp,4),%edi
leal -4(%esi,%ebp,4),%esi
@@ -65,12 +65,12 @@ Lnormal:
decl %ebp
pushl %ebp
shrl $3,%ebp
- jz Lend
+ jz L(end)
movl (%edi),%eax /* fetch destination cache line */
ALIGN (2)
-Loop: movl -28(%edi),%eax /* fetch destination cache line */
+L(oop): movl -28(%edi),%eax /* fetch destination cache line */
movl %edx,%ebx
movl (%esi),%eax
@@ -104,21 +104,23 @@ Loop: movl -28(%edi),%eax /* fetch destination cache line */
subl $32,%esi
subl $32,%edi
decl %ebp
- jnz Loop
+ jnz L(oop)
-Lend: popl %ebp
+L(end): popl %ebp
andl $7,%ebp
- jz Lend2
-Loop2: movl (%esi),%eax
+ jz L(end2)
+L(oop2):
+ movl (%esi),%eax
shldl %cl,%eax,%edx
movl %edx,(%edi)
movl %eax,%edx
subl $4,%esi
subl $4,%edi
decl %ebp
- jnz Loop2
+ jnz L(oop2)
-Lend2: shll %cl,%edx /* compute least significant limb */
+L(end2):
+ shll %cl,%edx /* compute least significant limb */
movl %edx,(%edi) /* store it */
popl %eax /* pop carry limb */
@@ -134,7 +136,7 @@ Lend2: shll %cl,%edx /* compute least significant limb */
function is documented to work for overlapping source and destination.
*/
-Lspecial:
+L(special):
movl (%esi),%edx
addl $4,%esi
@@ -145,12 +147,13 @@ Lspecial:
addl %edx,%edx
incl %ebp
decl %ebp
- jz LLend
+ jz L(Lend)
movl (%edi),%eax /* fetch destination cache line */
ALIGN (2)
-LLoop: movl 28(%edi),%eax /* fetch destination cache line */
+L(Loop):
+ movl 28(%edi),%eax /* fetch destination cache line */
movl %edx,%ebx
movl (%esi),%eax
@@ -184,14 +187,16 @@ LLoop: movl 28(%edi),%eax /* fetch destination cache line */
leal 32(%esi),%esi /* use leal not to clobber carry */
leal 32(%edi),%edi
decl %ebp
- jnz LLoop
+ jnz L(Loop)
-LLend: popl %ebp
+L(Lend):
+ popl %ebp
sbbl %eax,%eax /* save carry in %eax */
andl $7,%ebp
- jz LLend2
+ jz L(Lend2)
addl %eax,%eax /* restore carry from eax */
-LLoop2: movl %edx,%ebx
+L(Loop2):
+ movl %edx,%ebx
movl (%esi),%edx
adcl %edx,%edx
movl %ebx,(%edi)
@@ -199,11 +204,12 @@ LLoop2: movl %edx,%ebx
leal 4(%esi),%esi /* use leal not to clobber carry */
leal 4(%edi),%edi
decl %ebp
- jnz LLoop2
+ jnz L(Loop2)
- jmp LL1
-LLend2: addl %eax,%eax /* restore carry from eax */
-LL1: movl %edx,(%edi) /* store last limb */
+ jmp L(L1)
+L(Lend2):
+ addl %eax,%eax /* restore carry from eax */
+L(L1): movl %edx,(%edi) /* store last limb */
sbbl %eax,%eax
negl %eax