Index: gcc/gmp/mpn/x86/k7/mmx/lshift.asm |
diff --git a/gcc/gmp/mpn/x86/k7/mmx/lshift.asm b/gcc/gmp/mpn/x86/k7/mmx/lshift.asm |
deleted file mode 100644 |
index b3bff8ffd16ecef6980a8e7f67a85e4b245c7e69..0000000000000000000000000000000000000000 |
--- a/gcc/gmp/mpn/x86/k7/mmx/lshift.asm |
+++ /dev/null |
@@ -1,470 +0,0 @@ |
-dnl AMD K7 mpn_lshift -- mpn left shift. |
- |
-dnl Copyright 1999, 2000, 2001, 2002 Free Software Foundation, Inc. |
-dnl |
-dnl This file is part of the GNU MP Library. |
-dnl |
-dnl The GNU MP Library is free software; you can redistribute it and/or |
-dnl modify it under the terms of the GNU Lesser General Public License as |
-dnl published by the Free Software Foundation; either version 3 of the |
-dnl License, or (at your option) any later version. |
-dnl |
-dnl The GNU MP Library is distributed in the hope that it will be useful, |
-dnl but WITHOUT ANY WARRANTY; without even the implied warranty of |
-dnl MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
-dnl Lesser General Public License for more details. |
-dnl |
-dnl You should have received a copy of the GNU Lesser General Public License |
-dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/. |
- |
-include(`../config.m4') |
- |
- |
-C K7: 1.21 cycles/limb (at 16 limbs/loop). |
- |
- |
- |
-dnl K7: UNROLL_COUNT cycles/limb |
-dnl 4 1.51 |
-dnl 8 1.26 |
-dnl 16 1.21 |
-dnl 32 1.2 |
-dnl Maximum possible with the current code is 64. |
- |
-deflit(UNROLL_COUNT, 16) |
- |
- |
-C mp_limb_t mpn_lshift (mp_ptr dst, mp_srcptr src, mp_size_t size, |
-C unsigned shift); |
-C |
-C Shift src,size left by shift many bits and store the result in dst,size. |
-C Zeros are shifted in at the right. The bits shifted out at the left are |
-C the return value. |
-C |
-C The comments in mpn_rshift apply here too. |
- |
-ifdef(`PIC',` |
-deflit(UNROLL_THRESHOLD, 10) |
-',` |
-deflit(UNROLL_THRESHOLD, 10) |
-') |
- |
-defframe(PARAM_SHIFT,16) |
-defframe(PARAM_SIZE, 12) |
-defframe(PARAM_SRC, 8) |
-defframe(PARAM_DST, 4) |
- |
-defframe(SAVE_EDI, -4) |
-defframe(SAVE_ESI, -8) |
-defframe(SAVE_EBX, -12) |
-deflit(SAVE_SIZE, 12) |
- |
- TEXT |
- ALIGN(32) |
- |
-PROLOGUE(mpn_lshift) |
-deflit(`FRAME',0) |
- |
- movl PARAM_SIZE, %eax |
- movl PARAM_SRC, %edx |
- subl $SAVE_SIZE, %esp |
-deflit(`FRAME',SAVE_SIZE) |
- |
- movl PARAM_SHIFT, %ecx |
- movl %edi, SAVE_EDI |
- |
- movl PARAM_DST, %edi |
- decl %eax |
- jnz L(more_than_one_limb) |
- |
- movl (%edx), %edx |
- |
- shldl( %cl, %edx, %eax) C eax was decremented to zero |
- |
- shll %cl, %edx |
- |
- movl %edx, (%edi) |
- movl SAVE_EDI, %edi |
- addl $SAVE_SIZE, %esp |
- |
- ret |
- |
- |
-C ----------------------------------------------------------------------------- |
-L(more_than_one_limb): |
- C eax size-1 |
- C ebx |
- C ecx shift |
- C edx src |
- C esi |
- C edi dst |
- C ebp |
- |
- movd PARAM_SHIFT, %mm6 |
- movd (%edx,%eax,4), %mm5 C src high limb |
- cmp $UNROLL_THRESHOLD-1, %eax |
- |
- jae L(unroll) |
- negl %ecx |
- movd (%edx), %mm4 C src low limb |
- |
- addl $32, %ecx |
- |
- movd %ecx, %mm7 |
- |
-L(simple_top): |
- C eax loop counter, limbs |
- C ebx |
- C ecx |
- C edx src |
- C esi |
- C edi dst |
- C ebp |
- C |
- C mm0 scratch |
- C mm4 src low limb |
- C mm5 src high limb |
- C mm6 shift |
- C mm7 32-shift |
- |
- movq -4(%edx,%eax,4), %mm0 |
- decl %eax |
- |
- psrlq %mm7, %mm0 |
- |
- movd %mm0, 4(%edi,%eax,4) |
- jnz L(simple_top) |
- |
- |
- psllq %mm6, %mm5 |
- psllq %mm6, %mm4 |
- |
- psrlq $32, %mm5 |
- movd %mm4, (%edi) C dst low limb |
- |
- movd %mm5, %eax C return value |
- |
- movl SAVE_EDI, %edi |
- addl $SAVE_SIZE, %esp |
- emms |
- |
- ret |
- |
- |
-C ----------------------------------------------------------------------------- |
- ALIGN(16) |
-L(unroll): |
- C eax size-1 |
- C ebx (saved) |
- C ecx shift |
- C edx src |
- C esi |
- C edi dst |
- C ebp |
- C |
- C mm5 src high limb, for return value |
- C mm6 lshift |
- |
- movl %esi, SAVE_ESI |
- movl %ebx, SAVE_EBX |
- leal -4(%edx,%eax,4), %edx C &src[size-2] |
- |
- testb $4, %dl |
- movq (%edx), %mm1 C src high qword |
- |
- jz L(start_src_aligned) |
- |
- |
- C src isn't aligned, process high limb (marked xxx) separately to |
- C make it so |
- C |
- C source -4(edx,%eax,4) |
- C | |
- C +-------+-------+-------+-- |
- C | xxx | |
- C +-------+-------+-------+-- |
- C 0mod8 4mod8 0mod8 |
- C |
- C dest -4(edi,%eax,4) |
- C | |
- C +-------+-------+-- |
- C | xxx | | |
- C +-------+-------+-- |
- |
- psllq %mm6, %mm1 |
- subl $4, %edx |
- movl %eax, PARAM_SIZE C size-1 |
- |
- psrlq $32, %mm1 |
- decl %eax C size-2 is new size-1 |
- |
- movd %mm1, 4(%edi,%eax,4) |
- movq (%edx), %mm1 C new src high qword |
-L(start_src_aligned): |
- |
- |
- leal -4(%edi,%eax,4), %edi C &dst[size-2] |
- psllq %mm6, %mm5 |
- |
- testl $4, %edi |
- psrlq $32, %mm5 C return value |
- |
- jz L(start_dst_aligned) |
- |
- |
- C dst isn't aligned, subtract 4 bytes to make it so, and pretend the |
- C shift is 32 bits extra. High limb of dst (marked xxx) handled |
- C here separately. |
- C |
- C source %edx |
- C +-------+-------+-- |
- C | mm1 | |
- C +-------+-------+-- |
- C 0mod8 4mod8 |
- C |
- C dest %edi |
- C +-------+-------+-------+-- |
- C | xxx | |
- C +-------+-------+-------+-- |
- C 0mod8 4mod8 0mod8 |
- |
- movq %mm1, %mm0 |
- psllq %mm6, %mm1 |
- addl $32, %ecx C shift+32 |
- |
- psrlq $32, %mm1 |
- |
- movd %mm1, 4(%edi) |
- movq %mm0, %mm1 |
- subl $4, %edi |
- |
- movd %ecx, %mm6 C new lshift |
-L(start_dst_aligned): |
- |
- decl %eax C size-2, two last limbs handled at end |
- movq %mm1, %mm2 C copy of src high qword |
- negl %ecx |
- |
- andl $-2, %eax C round size down to even |
- addl $64, %ecx |
- |
- movl %eax, %ebx |
- negl %eax |
- |
- andl $UNROLL_MASK, %eax |
- decl %ebx |
- |
- shll %eax |
- |
- movd %ecx, %mm7 C rshift = 64-lshift |
- |
-ifdef(`PIC',` |
- call L(pic_calc) |
-L(here): |
-',` |
- leal L(entry) (%eax,%eax,4), %esi |
-') |
- shrl $UNROLL_LOG2, %ebx C loop counter |
- |
- leal ifelse(UNROLL_BYTES,256,128) -8(%edx,%eax,2), %edx |
- leal ifelse(UNROLL_BYTES,256,128) (%edi,%eax,2), %edi |
- movl PARAM_SIZE, %eax C for use at end |
- jmp *%esi |
- |
- |
-ifdef(`PIC',` |
-L(pic_calc): |
- C See mpn/x86/README about old gas bugs |
- leal (%eax,%eax,4), %esi |
- addl $L(entry)-L(here), %esi |
- addl (%esp), %esi |
- |
- ret_internal |
-') |
- |
- |
-C ----------------------------------------------------------------------------- |
- ALIGN(32) |
-L(top): |
- C eax size (for use at end) |
- C ebx loop counter |
- C ecx rshift |
- C edx src |
- C esi computed jump |
- C edi dst |
- C ebp |
- C |
- C mm0 scratch |
- C mm1 \ carry (alternating, mm2 first) |
- C mm2 / |
- C mm6 lshift |
- C mm7 rshift |
- C |
- C 10 code bytes/limb |
- C |
- C The two chunks differ in whether mm1 or mm2 hold the carry. |
- C The computed jump puts the initial carry in both mm1 and mm2. |
- |
-L(entry): |
-deflit(CHUNK_COUNT, 4) |
-forloop(i, 0, UNROLL_COUNT/CHUNK_COUNT-1, ` |
- deflit(`disp0', eval(-i*CHUNK_COUNT*4 ifelse(UNROLL_BYTES,256,-128))) |
- deflit(`disp1', eval(disp0 - 8)) |
- |
-Zdisp( movq, disp0,(%edx), %mm0) |
- psllq %mm6, %mm2 |
- |
- movq %mm0, %mm1 |
- psrlq %mm7, %mm0 |
- |
- por %mm2, %mm0 |
-Zdisp( movq, %mm0, disp0,(%edi)) |
- |
- |
-Zdisp( movq, disp1,(%edx), %mm0) |
- psllq %mm6, %mm1 |
- |
- movq %mm0, %mm2 |
- psrlq %mm7, %mm0 |
- |
- por %mm1, %mm0 |
-Zdisp( movq, %mm0, disp1,(%edi)) |
-') |
- |
- subl $UNROLL_BYTES, %edx |
- subl $UNROLL_BYTES, %edi |
- decl %ebx |
- |
- jns L(top) |
- |
- |
- |
-define(`disp', `m4_empty_if_zero(eval($1 ifelse(UNROLL_BYTES,256,-128)))') |
- |
-L(end): |
- testb $1, %al |
- movl SAVE_EBX, %ebx |
- psllq %mm6, %mm2 C wanted left shifted in all cases below |
- |
- movd %mm5, %eax |
- |
- movl SAVE_ESI, %esi |
- jz L(end_even) |
- |
- |
-L(end_odd): |
- |
- C Size odd, destination was aligned. |
- C |
- C source edx+8 edx+4 |
- C --+---------------+-------+ |
- C | mm2 | | |
- C --+---------------+-------+ |
- C |
- C dest edi |
- C --+---------------+---------------+-------+ |
- C | written | | | |
- C --+---------------+---------------+-------+ |
- C |
- C mm6 = shift |
- C mm7 = ecx = 64-shift |
- |
- |
- C Size odd, destination was unaligned. |
- C |
- C source edx+8 edx+4 |
- C --+---------------+-------+ |
- C | mm2 | | |
- C --+---------------+-------+ |
- C |
- C dest edi |
- C --+---------------+---------------+ |
- C | written | | |
- C --+---------------+---------------+ |
- C |
- C mm6 = shift+32 |
- C mm7 = ecx = 64-(shift+32) |
- |
- |
- C In both cases there's one extra limb of src to fetch and combine |
- C with mm2 to make a qword at (%edi), and in the aligned case |
- C there's an extra limb of dst to be formed from that extra src limb |
- C left shifted. |
- |
- movd disp(4) (%edx), %mm0 |
- testb $32, %cl |
- |
- movq %mm0, %mm1 |
- psllq $32, %mm0 |
- |
- psrlq %mm7, %mm0 |
- psllq %mm6, %mm1 |
- |
- por %mm2, %mm0 |
- |
- movq %mm0, disp(0) (%edi) |
- jz L(end_odd_unaligned) |
- movd %mm1, disp(-4) (%edi) |
-L(end_odd_unaligned): |
- |
- movl SAVE_EDI, %edi |
- addl $SAVE_SIZE, %esp |
- emms |
- |
- ret |
- |
- |
-L(end_even): |
- |
- C Size even, destination was aligned. |
- C |
- C source edx+8 |
- C --+---------------+ |
- C | mm2 | |
- C --+---------------+ |
- C |
- C dest edi |
- C --+---------------+---------------+ |
- C | written | | |
- C --+---------------+---------------+ |
- C |
- C mm6 = shift |
- C mm7 = ecx = 64-shift |
- |
- |
- C Size even, destination was unaligned. |
- C |
- C source edx+8 |
- C --+---------------+ |
- C | mm2 | |
- C --+---------------+ |
- C |
- C dest edi+4 |
- C --+---------------+-------+ |
- C | written | | |
- C --+---------------+-------+ |
- C |
- C mm6 = shift+32 |
- C mm7 = ecx = 64-(shift+32) |
- |
- |
- C The movq for the aligned case overwrites the movd for the |
- C unaligned case. |
- |
- movq %mm2, %mm0 |
- psrlq $32, %mm2 |
- |
- testb $32, %cl |
- movd %mm2, disp(4) (%edi) |
- |
- jz L(end_even_unaligned) |
- movq %mm0, disp(0) (%edi) |
-L(end_even_unaligned): |
- |
- movl SAVE_EDI, %edi |
- addl $SAVE_SIZE, %esp |
- emms |
- |
- ret |
- |
-EPILOGUE() |