Index: gcc/gmp/mpn/x86_64/addaddmul_1msb0.asm |
diff --git a/gcc/gmp/mpn/x86_64/addaddmul_1msb0.asm b/gcc/gmp/mpn/x86_64/addaddmul_1msb0.asm |
deleted file mode 100644 |
index 89e7bed9800431eabc31891ae9ffe2ae4546735c..0000000000000000000000000000000000000000 |
--- a/gcc/gmp/mpn/x86_64/addaddmul_1msb0.asm |
+++ /dev/null |
@@ -1,155 +0,0 @@ |
-dnl AMD64 mpn_addaddmul_1msb0, R = Au + Bv, u,v < 2^63. |
- |
-dnl Copyright 2008 Free Software Foundation, Inc. |
- |
-dnl This file is part of the GNU MP Library. |
- |
-dnl The GNU MP Library is free software; you can redistribute it and/or modify |
-dnl it under the terms of the GNU Lesser General Public License as published |
-dnl by the Free Software Foundation; either version 3 of the License, or (at |
-dnl your option) any later version. |
- |
-dnl The GNU MP Library is distributed in the hope that it will be useful, but |
-dnl WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
-dnl or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public |
-dnl License for more details. |
- |
-dnl You should have received a copy of the GNU Lesser General Public License |
-dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. |
- |
-include(`../config.m4') |
- |
-C cycles/limb |
-C K8: 2.167 |
-C P4: 12.0 |
-C P6-15: 4.0 |
- |
-C TODO |
-C * Perhaps handle various n mod 3 sizes better. The code now is too large. |
- |
-C INPUT PARAMETERS |
-define(`rp', `%rdi') |
-define(`ap', `%rsi') |
-define(`bp_param', `%rdx') |
-define(`n', `%rcx') |
-define(`u0', `%r8') |
-define(`v0', `%r9') |
- |
- |
-define(`bp', `%rbp') |
- |
-ASM_START() |
- TEXT |
- ALIGN(16) |
-PROLOGUE(mpn_addaddmul_1msb0) |
- push %r12 |
- push %rbp |
- |
- lea (ap,n,8), ap |
- lea (bp_param,n,8), bp |
- lea (rp,n,8), rp |
- neg n |
- |
- mov (ap,n,8), %rax |
- mul %r8 |
- mov %rax, %r12 |
- mov (bp,n,8), %rax |
- mov %rdx, %r10 |
- add $3, n |
- jns L(end) |
- |
- ALIGN(16) |
-L(top): mul %r9 |
- add %rax, %r12 |
- mov -16(ap,n,8), %rax |
- adc %rdx, %r10 |
- mov %r12, -24(rp,n,8) |
- mul %r8 |
- add %rax, %r10 |
- mov -16(bp,n,8), %rax |
- mov $0, %r11d |
- adc %rdx, %r11 |
- mul %r9 |
- add %rax, %r10 |
- mov -8(ap,n,8), %rax |
- adc %rdx, %r11 |
- mov %r10, -16(rp,n,8) |
- mul %r8 |
- add %rax, %r11 |
- mov -8(bp,n,8), %rax |
- mov $0, %r12d |
- adc %rdx, %r12 |
- mul %r9 |
- add %rax, %r11 |
- adc %rdx, %r12 |
- mov (ap,n,8), %rax |
- mul %r8 |
- add %rax, %r12 |
- mov %r11, -8(rp,n,8) |
- mov (bp,n,8), %rax |
- mov $0, %r10d |
- adc %rdx, %r10 |
- add $3, n |
- js L(top) |
- |
-L(end): cmp $1, R32(n) |
- ja 2f |
- jz 1f |
- |
- mul %r9 |
- add %rax, %r12 |
- mov -16(ap), %rax |
- adc %rdx, %r10 |
- mov %r12, -24(rp) |
- mul %r8 |
- add %rax, %r10 |
- mov -16(bp), %rax |
- mov $0, %r11d |
- adc %rdx, %r11 |
- mul %r9 |
- add %rax, %r10 |
- mov -8(ap), %rax |
- adc %rdx, %r11 |
- mov %r10, -16(rp) |
- mul %r8 |
- add %rax, %r11 |
- mov -8(bp), %rax |
- mov $0, %r12d |
- adc %rdx, %r12 |
- mul %r9 |
- add %rax, %r11 |
- adc %rdx, %r12 |
- mov %r11, -8(rp) |
- mov %r12, %rax |
- pop %rbp |
- pop %r12 |
- ret |
- |
-1: mul %r9 |
- add %rax, %r12 |
- mov -8(ap), %rax |
- adc %rdx, %r10 |
- mov %r12, -16(rp) |
- mul %r8 |
- add %rax, %r10 |
- mov -8(bp), %rax |
- mov $0, %r11d |
- adc %rdx, %r11 |
- mul %r9 |
- add %rax, %r10 |
- adc %rdx, %r11 |
- mov %r10, -8(rp) |
- mov %r11, %rax |
- pop %rbp |
- pop %r12 |
- ret |
- |
-2: mul %r9 |
- add %rax, %r12 |
- mov %r12, -8(rp) |
- adc %rdx, %r10 |
- mov %r10, %rax |
- pop %rbp |
- pop %r12 |
- ret |
-EPILOGUE() |