Index: gcc/gmp/mpn/powerpc64/mode64/mul_basecase.asm |
diff --git a/gcc/gmp/mpn/powerpc64/mode64/mul_basecase.asm b/gcc/gmp/mpn/powerpc64/mode64/mul_basecase.asm |
deleted file mode 100644 |
index cea5417eb2fd92504ceb3a162e25f77820516768..0000000000000000000000000000000000000000 |
--- a/gcc/gmp/mpn/powerpc64/mode64/mul_basecase.asm |
+++ /dev/null |
@@ -1,698 +0,0 @@ |
-dnl PowerPC-64 mpn_basecase. |
- |
-dnl Copyright 1999, 2000, 2001, 2003, 2004, 2005, 2006, 2008 Free Software |
-dnl Foundation, Inc. |
- |
-dnl This file is part of the GNU MP Library. |
- |
-dnl The GNU MP Library is free software; you can redistribute it and/or modify |
-dnl it under the terms of the GNU Lesser General Public License as published |
-dnl by the Free Software Foundation; either version 3 of the License, or (at |
-dnl your option) any later version. |
- |
-dnl The GNU MP Library is distributed in the hope that it will be useful, but |
-dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
-dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public |
-dnl License for more details. |
- |
-dnl You should have received a copy of the GNU Lesser General Public License |
-dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. |
- |
-include(`../config.m4') |
- |
-C cycles/limb |
-C POWER3/PPC630: 6-18 |
-C POWER4/PPC970: 8 |
-C POWER5: 8 |
- |
- |
-C INPUT PARAMETERS |
-define(`rp', `r3') |
-define(`up', `r4') |
-define(`un', `r5') |
-define(`vp', `r6') |
-define(`vn', `r7') |
- |
-define(`v0', `r25') |
-define(`outer_rp', `r22') |
-define(`outer_up', `r23') |
- |
-ASM_START() |
-PROLOGUE(mpn_mul_basecase) |
- |
-C Special code for un <= 2, for efficiency of these important cases, |
-C and since it simplifies the default code. |
- cmpdi cr0, un, 2 |
- bgt cr0, L(un_gt2) |
- cmpdi cr6, vn, 1 |
- ld r7, 0(vp) |
- ld r5, 0(up) |
- mulld r8, r5, r7 C weight 0 |
- mulhdu r9, r5, r7 C weight 1 |
- std r8, 0(rp) |
- beq cr0, L(2x) |
- std r9, 8(rp) |
- blr |
- ALIGN(16) |
-L(2x): ld r0, 8(up) |
- mulld r8, r0, r7 C weight 1 |
- mulhdu r10, r0, r7 C weight 2 |
- addc r9, r9, r8 |
- addze r10, r10 |
- bne cr6, L(2x2) |
- std r9, 8(rp) |
- std r10, 16(rp) |
- blr |
- ALIGN(16) |
-L(2x2): ld r6, 8(vp) |
- nop |
- mulld r8, r5, r6 C weight 1 |
- mulhdu r11, r5, r6 C weight 2 |
- addc r9, r9, r8 |
- std r9, 8(rp) |
- adde r11, r11, r10 |
- mulld r12, r0, r6 C weight 2 |
- mulhdu r0, r0, r6 C weight 3 |
- addze r0, r0 |
- addc r11, r11, r12 |
- addze r0, r0 |
- std r11, 16(rp) |
- std r0, 24(rp) |
- blr |
- |
-L(un_gt2): |
- std r31, -8(r1) |
- std r30, -16(r1) |
- std r29, -24(r1) |
- std r28, -32(r1) |
- std r27, -40(r1) |
- std r26, -48(r1) |
- std r25, -56(r1) |
- std r24, -64(r1) |
- std r23, -72(r1) |
- std r22, -80(r1) |
- |
- mr outer_rp, rp |
- mr outer_up, up |
- |
- ld v0, 0(vp) C new v limb |
- addi vp, vp, 8 |
- ld r26, 0(up) |
- |
- rldicl. r0, un, 0,62 C r0 = n & 3, set cr0 |
- cmpdi cr6, r0, 2 |
- addi un, un, 1 C compute count... |
- srdi un, un, 2 C ...for ctr |
- mtctr un C copy inner loop count into ctr |
- beq cr0, L(b0) |
- blt cr6, L(b1) |
- beq cr6, L(b2) |
- |
- |
- ALIGN(16) |
-L(b3): mulld r0, r26, v0 |
- mulhdu r12, r26, v0 |
- addic r0, r0, 0 |
- std r0, 0(rp) |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_m_3) |
- |
- ALIGN(16) |
-L(lo_m_3): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 24(up) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r27, 32(up) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r26, 40(up) |
- nop |
- mulld r11, r27, v0 |
- mulhdu r12, r27, v0 |
- ld r27, 48(up) |
- std r0, 8(rp) |
- adde r9, r9, r8 |
- std r24, 16(rp) |
- adde r11, r11, r10 |
- std r9, 24(rp) |
- addi up, up, 32 |
- std r11, 32(rp) |
- addi rp, rp, 32 |
- bdnz L(lo_m_3) |
- |
- ALIGN(16) |
-L(end_m_3): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- |
- std r0, 8(rp) |
- std r24, 16(rp) |
- addze r8, r8 |
- std r8, 24(rp) |
- addic. vn, vn, -1 |
- beq L(ret) |
- |
- ALIGN(16) |
-L(outer_lo_3): |
- mtctr un C copy inner loop count into ctr |
- addi rp, outer_rp, 8 |
- mr up, outer_up |
- addi outer_rp, outer_rp, 8 |
- ld v0, 0(vp) C new v limb |
- addi vp, vp, 8 |
- ld r26, 0(up) |
- ld r28, 0(rp) |
- mulld r0, r26, v0 |
- mulhdu r12, r26, v0 |
- addc r0, r0, r28 |
- std r0, 0(rp) |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_3) |
- |
- ALIGN(16) C registers dying |
-L(lo_3): |
- mulld r0, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 24(up) C |
- ld r28, 8(rp) C |
- mulld r24, r27, v0 C |
- mulhdu r8, r27, v0 C 27 |
- ld r27, 32(up) C |
- ld r29, 16(rp) C |
- adde r0, r0, r12 C 0 12 |
- adde r24, r24, r10 C 24 10 |
- mulld r9, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 40(up) C |
- ld r30, 24(rp) C |
- mulld r11, r27, v0 C |
- mulhdu r12, r27, v0 C 27 |
- ld r27, 48(up) C |
- ld r31, 32(rp) C |
- adde r9, r9, r8 C 8 9 |
- adde r11, r11, r10 C 10 11 |
- addze r12, r12 C 12 |
- addc r0, r0, r28 C 0 28 |
- std r0, 8(rp) C 0 |
- adde r24, r24, r29 C 7 29 |
- std r24, 16(rp) C 7 |
- adde r9, r9, r30 C 9 30 |
- std r9, 24(rp) C 9 |
- adde r11, r11, r31 C 11 31 |
- std r11, 32(rp) C 11 |
- addi up, up, 32 C |
- addi rp, rp, 32 C |
- bdnz L(lo_3) C |
- |
- ALIGN(16) |
-L(end_3): |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r28, 8(rp) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r29, 16(rp) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r10 |
- addze r8, r8 |
- addc r0, r0, r28 |
- std r0, 8(rp) |
- adde r24, r24, r29 |
- std r24, 16(rp) |
- addze r8, r8 |
- std r8, 24(rp) |
- |
- addic. vn, vn, -1 |
- bne L(outer_lo_3) |
- b L(ret) |
- |
- |
- ALIGN(16) |
-L(b0): ld r27, 8(up) |
- addi up, up, 8 |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- addc r24, r24, r10 |
- addze r12, r8 |
- std r0, 0(rp) |
- std r24, 8(rp) |
- addi rp, rp, 8 |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_m_0) |
- |
- ALIGN(16) |
-L(lo_m_0): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 24(up) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r27, 32(up) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r26, 40(up) |
- nop |
- mulld r11, r27, v0 |
- mulhdu r12, r27, v0 |
- ld r27, 48(up) |
- std r0, 8(rp) |
- adde r9, r9, r8 |
- std r24, 16(rp) |
- adde r11, r11, r10 |
- std r9, 24(rp) |
- addi up, up, 32 |
- std r11, 32(rp) |
- addi rp, rp, 32 |
- bdnz L(lo_m_0) |
- |
- ALIGN(16) |
-L(end_m_0): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- |
- std r0, 8(rp) |
- addze r8, r8 |
- std r24, 16(rp) |
- addic. vn, vn, -1 |
- std r8, 24(rp) |
- nop |
- beq L(ret) |
- |
- ALIGN(16) |
-L(outer_lo_0): |
- mtctr un C copy inner loop count into ctr |
- addi rp, outer_rp, 16 |
- addi up, outer_up, 8 |
- addi outer_rp, outer_rp, 8 |
- ld v0, 0(vp) C new v limb |
- addi vp, vp, 8 |
- ld r26, -8(up) |
- ld r27, 0(up) |
- ld r28, -8(rp) |
- ld r29, 0(rp) |
- nop |
- nop |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- addc r24, r24, r10 |
- addze r12, r8 |
- addc r0, r0, r28 |
- std r0, -8(rp) |
- adde r24, r24, r29 |
- std r24, 0(rp) |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_0) |
- |
- ALIGN(16) C registers dying |
-L(lo_0): |
- mulld r0, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 24(up) C |
- ld r28, 8(rp) C |
- mulld r24, r27, v0 C |
- mulhdu r8, r27, v0 C 27 |
- ld r27, 32(up) C |
- ld r29, 16(rp) C |
- adde r0, r0, r12 C 0 12 |
- adde r24, r24, r10 C 24 10 |
- mulld r9, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 40(up) C |
- ld r30, 24(rp) C |
- mulld r11, r27, v0 C |
- mulhdu r12, r27, v0 C 27 |
- ld r27, 48(up) C |
- ld r31, 32(rp) C |
- adde r9, r9, r8 C 8 9 |
- adde r11, r11, r10 C 10 11 |
- addze r12, r12 C 12 |
- addc r0, r0, r28 C 0 28 |
- std r0, 8(rp) C 0 |
- adde r24, r24, r29 C 7 29 |
- std r24, 16(rp) C 7 |
- adde r9, r9, r30 C 9 30 |
- std r9, 24(rp) C 9 |
- adde r11, r11, r31 C 11 31 |
- std r11, 32(rp) C 11 |
- addi up, up, 32 C |
- addi rp, rp, 32 C |
- bdnz L(lo_0) C |
- |
- ALIGN(16) |
-L(end_0): |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r28, 8(rp) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r29, 16(rp) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r10 |
- addze r8, r8 |
- addic. vn, vn, -1 |
- addc r0, r0, r28 |
- std r0, 8(rp) |
- adde r24, r24, r29 |
- std r24, 16(rp) |
- addze r8, r8 |
- std r8, 24(rp) |
- bne L(outer_lo_0) |
- b L(ret) |
- |
- |
- ALIGN(16) |
-L(b1): ld r27, 8(up) |
- nop |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 16(up) |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- addc r24, r24, r31 |
- adde r9, r9, r8 |
- addze r12, r10 |
- std r0, 0(rp) |
- std r24, 8(rp) |
- std r9, 16(rp) |
- addi up, up, 16 |
- addi rp, rp, 16 |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_m_1) |
- |
- ALIGN(16) |
-L(lo_m_1): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 24(up) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r27, 32(up) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r26, 40(up) |
- nop |
- mulld r11, r27, v0 |
- mulhdu r12, r27, v0 |
- ld r27, 48(up) |
- std r0, 8(rp) |
- adde r9, r9, r8 |
- std r24, 16(rp) |
- adde r11, r11, r10 |
- std r9, 24(rp) |
- addi up, up, 32 |
- std r11, 32(rp) |
- addi rp, rp, 32 |
- bdnz L(lo_m_1) |
- |
- ALIGN(16) |
-L(end_m_1): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- |
- std r0, 8(rp) |
- addze r8, r8 |
- std r24, 16(rp) |
- addic. vn, vn, -1 |
- std r8, 24(rp) |
- nop |
- beq L(ret) |
- |
- ALIGN(16) |
-L(outer_lo_1): |
- mtctr un C copy inner loop count into ctr |
- addi rp, outer_rp, 24 |
- addi up, outer_up, 16 |
- addi outer_rp, outer_rp, 8 |
- ld v0, 0(vp) C new v limb |
- addi vp, vp, 8 |
- ld r26, -16(up) |
- ld r27, -8(up) |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 0(up) |
- ld r28, -16(rp) |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r29, -8(rp) |
- ld r30, 0(rp) |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- addc r24, r24, r31 |
- adde r9, r9, r8 |
- addze r12, r10 |
- addc r0, r0, r28 |
- std r0, -16(rp) |
- adde r24, r24, r29 |
- std r24, -8(rp) |
- adde r9, r9, r30 |
- std r9, 0(rp) |
- ld r26, 8(up) |
- ld r27, 16(up) |
- bdz L(end_1) |
- |
- ALIGN(16) C registers dying |
-L(lo_1): |
- mulld r0, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 24(up) C |
- ld r28, 8(rp) C |
- mulld r24, r27, v0 C |
- mulhdu r8, r27, v0 C 27 |
- ld r27, 32(up) C |
- ld r29, 16(rp) C |
- adde r0, r0, r12 C 0 12 |
- adde r24, r24, r10 C 24 10 |
- mulld r9, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 40(up) C |
- ld r30, 24(rp) C |
- mulld r11, r27, v0 C |
- mulhdu r12, r27, v0 C 27 |
- ld r27, 48(up) C |
- ld r31, 32(rp) C |
- adde r9, r9, r8 C 8 9 |
- adde r11, r11, r10 C 10 11 |
- addze r12, r12 C 12 |
- addc r0, r0, r28 C 0 28 |
- std r0, 8(rp) C 0 |
- adde r24, r24, r29 C 7 29 |
- std r24, 16(rp) C 7 |
- adde r9, r9, r30 C 9 30 |
- std r9, 24(rp) C 9 |
- adde r11, r11, r31 C 11 31 |
- std r11, 32(rp) C 11 |
- addi up, up, 32 C |
- addi rp, rp, 32 C |
- bdnz L(lo_1) C |
- |
- ALIGN(16) |
-L(end_1): |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r28, 8(rp) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r29, 16(rp) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r10 |
- addze r8, r8 |
- addic. vn, vn, -1 |
- addc r0, r0, r28 |
- std r0, 8(rp) |
- adde r24, r24, r29 |
- std r24, 16(rp) |
- addze r8, r8 |
- std r8, 24(rp) |
- bne L(outer_lo_1) |
- b L(ret) |
- |
- |
- ALIGN(16) |
-L(b2): ld r27, 8(up) |
- addi up, up, -8 |
- addi rp, rp, -8 |
- li r12, 0 |
- addic r12, r12, 0 |
- |
- ALIGN(16) |
-L(lo_m_2): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- ld r26, 24(up) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r27, 32(up) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- mulld r9, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r26, 40(up) |
- nop |
- mulld r11, r27, v0 |
- mulhdu r12, r27, v0 |
- ld r27, 48(up) |
- std r0, 8(rp) |
- adde r9, r9, r8 |
- std r24, 16(rp) |
- adde r11, r11, r10 |
- std r9, 24(rp) |
- addi up, up, 32 |
- std r11, 32(rp) |
- |
- addi rp, rp, 32 |
- bdnz L(lo_m_2) |
- |
- ALIGN(16) |
-L(end_m_2): |
- mulld r0, r26, v0 |
- mulhdu r31, r26, v0 |
- |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- |
- adde r0, r0, r12 |
- adde r24, r24, r31 |
- |
- std r0, 8(rp) |
- addze r8, r8 |
- std r24, 16(rp) |
- addic. vn, vn, -1 |
- std r8, 24(rp) |
- nop |
- beq L(ret) |
- |
- ALIGN(16) |
-L(outer_lo_2): |
- mtctr un C copy inner loop count into ctr |
- addi rp, outer_rp, 0 |
- addi up, outer_up, -8 |
- addi outer_rp, outer_rp, 8 |
- ld v0, 0(vp) C new v limb |
- addi vp, vp, 8 |
- ld r26, 8(up) |
- ld r27, 16(up) |
- li r12, 0 |
- addic r12, r12, 0 |
- |
- ALIGN(16) C registers dying |
-L(lo_2): |
- mulld r0, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 24(up) C |
- ld r28, 8(rp) C |
- mulld r24, r27, v0 C |
- mulhdu r8, r27, v0 C 27 |
- ld r27, 32(up) C |
- ld r29, 16(rp) C |
- adde r0, r0, r12 C 0 12 |
- adde r24, r24, r10 C 24 10 |
- mulld r9, r26, v0 C |
- mulhdu r10, r26, v0 C 26 |
- ld r26, 40(up) C |
- ld r30, 24(rp) C |
- mulld r11, r27, v0 C |
- mulhdu r12, r27, v0 C 27 |
- ld r27, 48(up) C |
- ld r31, 32(rp) C |
- adde r9, r9, r8 C 8 9 |
- adde r11, r11, r10 C 10 11 |
- addze r12, r12 C 12 |
- addc r0, r0, r28 C 0 28 |
- std r0, 8(rp) C 0 |
- adde r24, r24, r29 C 7 29 |
- std r24, 16(rp) C 7 |
- adde r9, r9, r30 C 9 30 |
- std r9, 24(rp) C 9 |
- adde r11, r11, r31 C 11 31 |
- std r11, 32(rp) C 11 |
- addi up, up, 32 C |
- addi rp, rp, 32 C |
- bdnz L(lo_2) C |
- |
- ALIGN(16) |
-L(end_2): |
- mulld r0, r26, v0 |
- mulhdu r10, r26, v0 |
- ld r28, 8(rp) |
- nop |
- mulld r24, r27, v0 |
- mulhdu r8, r27, v0 |
- ld r29, 16(rp) |
- nop |
- adde r0, r0, r12 |
- adde r24, r24, r10 |
- addze r8, r8 |
- addic. vn, vn, -1 |
- addc r0, r0, r28 |
- std r0, 8(rp) |
- adde r24, r24, r29 |
- std r24, 16(rp) |
- addze r8, r8 |
- std r8, 24(rp) |
- bne L(outer_lo_2) |
- b L(ret) |
- |
- |
-L(ret): ld r31, -8(r1) |
- ld r30, -16(r1) |
- ld r29, -24(r1) |
- ld r28, -32(r1) |
- ld r27, -40(r1) |
- ld r26, -48(r1) |
- ld r25, -56(r1) |
- ld r24, -64(r1) |
- ld r23, -72(r1) |
- ld r22, -80(r1) |
- blr |
-EPILOGUE() |