| Index: gcc/gmp/mpn/x86/pentium4/sse2/divrem_1.asm
|
| diff --git a/gcc/gmp/mpn/x86/pentium4/sse2/divrem_1.asm b/gcc/gmp/mpn/x86/pentium4/sse2/divrem_1.asm
|
| deleted file mode 100644
|
| index 7f973dbf983861af7115994fbc43284888d72757..0000000000000000000000000000000000000000
|
| --- a/gcc/gmp/mpn/x86/pentium4/sse2/divrem_1.asm
|
| +++ /dev/null
|
| @@ -1,635 +0,0 @@
|
| -dnl Intel Pentium-4 mpn_divrem_1 -- mpn by limb division.
|
| -
|
| -dnl Copyright 1999, 2000, 2001, 2002, 2003, 2004 Free Software Foundation,
|
| -dnl Inc.
|
| -dnl
|
| -dnl This file is part of the GNU MP Library.
|
| -dnl
|
| -dnl The GNU MP Library is free software; you can redistribute it and/or
|
| -dnl modify it under the terms of the GNU Lesser General Public License as
|
| -dnl published by the Free Software Foundation; either version 3 of the
|
| -dnl License, or (at your option) any later version.
|
| -dnl
|
| -dnl The GNU MP Library is distributed in the hope that it will be useful,
|
| -dnl but WITHOUT ANY WARRANTY; without even the implied warranty of
|
| -dnl MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
| -dnl Lesser General Public License for more details.
|
| -dnl
|
| -dnl You should have received a copy of the GNU Lesser General Public License
|
| -dnl along with the GNU MP Library. If not, see http://www.gnu.org/licenses/.
|
| -
|
| -include(`../config.m4')
|
| -
|
| -
|
| -C P4: 32 cycles/limb integer part, 30 cycles/limb fraction part.
|
| -
|
| -
|
| -C mp_limb_t mpn_divrem_1 (mp_ptr dst, mp_size_t xsize,
|
| -C mp_srcptr src, mp_size_t size,
|
| -C mp_limb_t divisor);
|
| -C mp_limb_t mpn_divrem_1c (mp_ptr dst, mp_size_t xsize,
|
| -C mp_srcptr src, mp_size_t size,
|
| -C mp_limb_t divisor, mp_limb_t carry);
|
| -C mp_limb_t mpn_preinv_divrem_1 (mp_ptr dst, mp_size_t xsize,
|
| -C mp_srcptr src, mp_size_t size,
|
| -C mp_limb_t divisor, mp_limb_t inverse,
|
| -C unsigned shift);
|
| -C
|
| -C Algorithm:
|
| -C
|
| -C The method and nomenclature follow part 8 of "Division by Invariant
|
| -C Integers using Multiplication" by Granlund and Montgomery, reference in
|
| -C gmp.texi.
|
| -C
|
| -C "m" is written for what is m' in the paper, and "d" for d_norm, which
|
| -C won't cause any confusion since it's only the normalized divisor that's of
|
| -C any use in the code. "b" is written for 2^N, the size of a limb, N being
|
| -C 32 here.
|
| -C
|
| -C The step "sdword dr = n - 2^N*d + (2^N-1-q1) * d" is instead done as
|
| -C "n-d - q1*d". This rearrangement gives the same two-limb answer but lets
|
| -C us have just a psubq on the dependent chain.
|
| -C
|
| -C For reference, the way the k7 code uses "n-(q1+1)*d" would not suit here,
|
| -C detecting an overflow of q1+1 when q1=0xFFFFFFFF would cost too much.
|
| -C
|
| -C Notes:
|
| -C
|
| -C mpn_divrem_1 and mpn_preinv_divrem_1 avoid one division if the src high
|
| -C limb is less than the divisor. mpn_divrem_1c doesn't check for a zero
|
| -C carry, since in normal circumstances that will be a very rare event.
|
| -C
|
| -C The test for skipping a division is branch free (once size>=1 is tested).
|
| -C The store to the destination high limb is 0 when a divide is skipped, or
|
| -C if it's not skipped then a copy of the src high limb is stored. The
|
| -C latter is in case src==dst.
|
| -C
|
| -C There's a small bias towards expecting xsize==0, by having code for
|
| -C xsize==0 in a straight line and xsize!=0 under forward jumps.
|
| -C
|
| -C Enhancements:
|
| -C
|
| -C The loop measures 32 cycles, but the dependent chain would suggest it
|
| -C could be done with 30. Not sure where to start looking for the extras.
|
| -C
|
| -C Alternatives:
|
| -C
|
| -C If the divisor is normalized (high bit set) then a division step can
|
| -C always be skipped, since the high destination limb is always 0 or 1 in
|
| -C that case. It doesn't seem worth checking for this though, since it
|
| -C probably occurs infrequently.
|
| -
|
| -
|
| -dnl MUL_THRESHOLD is the value of xsize+size at which the multiply by
|
| -dnl inverse method is used, rather than plain "divl"s. Minimum value 1.
|
| -dnl
|
| -dnl The inverse takes about 80-90 cycles to calculate, but after that the
|
| -dnl multiply is 32 c/l versus division at about 58 c/l.
|
| -dnl
|
| -dnl At 4 limbs the div is a touch faster than the mul (and of course
|
| -dnl simpler), so start the mul from 5 limbs.
|
| -
|
| -deflit(MUL_THRESHOLD, 5)
|
| -
|
| -
|
| -defframe(PARAM_PREINV_SHIFT, 28) dnl mpn_preinv_divrem_1
|
| -defframe(PARAM_PREINV_INVERSE, 24) dnl mpn_preinv_divrem_1
|
| -defframe(PARAM_CARRY, 24) dnl mpn_divrem_1c
|
| -defframe(PARAM_DIVISOR,20)
|
| -defframe(PARAM_SIZE, 16)
|
| -defframe(PARAM_SRC, 12)
|
| -defframe(PARAM_XSIZE, 8)
|
| -defframe(PARAM_DST, 4)
|
| -
|
| -dnl re-use parameter space
|
| -define(SAVE_ESI,`PARAM_SIZE')
|
| -define(SAVE_EBP,`PARAM_SRC')
|
| -define(SAVE_EDI,`PARAM_DIVISOR')
|
| -define(SAVE_EBX,`PARAM_DST')
|
| -
|
| - TEXT
|
| -
|
| - ALIGN(16)
|
| -PROLOGUE(mpn_preinv_divrem_1)
|
| -deflit(`FRAME',0)
|
| -
|
| - movl PARAM_SIZE, %ecx
|
| - xorl %edx, %edx C carry if can't skip a div
|
| -
|
| - movl %esi, SAVE_ESI
|
| - movl PARAM_SRC, %esi
|
| -
|
| - movl %ebp, SAVE_EBP
|
| - movl PARAM_DIVISOR, %ebp
|
| -
|
| - movl %edi, SAVE_EDI
|
| - movl PARAM_DST, %edi
|
| -
|
| - movl -4(%esi,%ecx,4), %eax C src high limb
|
| -
|
| - movl %ebx, SAVE_EBX
|
| - movl PARAM_XSIZE, %ebx
|
| -
|
| - movd PARAM_PREINV_INVERSE, %mm4
|
| -
|
| - movd PARAM_PREINV_SHIFT, %mm7 C l
|
| - cmpl %ebp, %eax C high cmp divisor
|
| -
|
| - cmovc( %eax, %edx) C high is carry if high<divisor
|
| - movd %edx, %mm0 C carry
|
| -
|
| - movd %edx, %mm1 C carry
|
| - movl $0, %edx
|
| -
|
| - movd %ebp, %mm5 C d
|
| - cmovnc( %eax, %edx) C 0 if skip div, src high if not
|
| - C (the latter in case src==dst)
|
| - leal -4(%edi,%ebx,4), %edi C &dst[xsize-1]
|
| -
|
| - movl %edx, (%edi,%ecx,4) C dst high limb
|
| - sbbl $0, %ecx C skip one division if high<divisor
|
| - movl $32, %eax
|
| -
|
| - subl PARAM_PREINV_SHIFT, %eax
|
| - psllq %mm7, %mm5 C d normalized
|
| - leal (%edi,%ecx,4), %edi C &dst[xsize+size-1]
|
| - leal -4(%esi,%ecx,4), %esi C &src[size-1]
|
| -
|
| - movd %eax, %mm6 C 32-l
|
| - jmp L(start_preinv)
|
| -
|
| -EPILOGUE()
|
| -
|
| -
|
| - ALIGN(16)
|
| -PROLOGUE(mpn_divrem_1c)
|
| -deflit(`FRAME',0)
|
| -
|
| - movl PARAM_CARRY, %edx
|
| -
|
| - movl PARAM_SIZE, %ecx
|
| -
|
| - movl %esi, SAVE_ESI
|
| - movl PARAM_SRC, %esi
|
| -
|
| - movl %ebp, SAVE_EBP
|
| - movl PARAM_DIVISOR, %ebp
|
| -
|
| - movl %edi, SAVE_EDI
|
| - movl PARAM_DST, %edi
|
| -
|
| - movl %ebx, SAVE_EBX
|
| - movl PARAM_XSIZE, %ebx
|
| -
|
| - leal -4(%edi,%ebx,4), %edi C &dst[xsize-1]
|
| - jmp L(start_1c)
|
| -
|
| -EPILOGUE()
|
| -
|
| -
|
| - ALIGN(16)
|
| -PROLOGUE(mpn_divrem_1)
|
| -deflit(`FRAME',0)
|
| -
|
| - movl PARAM_SIZE, %ecx
|
| - xorl %edx, %edx C initial carry (if can't skip a div)
|
| -
|
| - movl %esi, SAVE_ESI
|
| - movl PARAM_SRC, %esi
|
| -
|
| - movl %ebp, SAVE_EBP
|
| - movl PARAM_DIVISOR, %ebp
|
| -
|
| - movl %edi, SAVE_EDI
|
| - movl PARAM_DST, %edi
|
| -
|
| - movl %ebx, SAVE_EBX
|
| - movl PARAM_XSIZE, %ebx
|
| - leal -4(%edi,%ebx,4), %edi C &dst[xsize-1]
|
| -
|
| - orl %ecx, %ecx C size
|
| - jz L(no_skip_div) C if size==0
|
| - movl -4(%esi,%ecx,4), %eax C src high limb
|
| -
|
| - cmpl %ebp, %eax C high cmp divisor
|
| -
|
| - cmovnc( %eax, %edx) C 0 if skip div, src high if not
|
| - movl %edx, (%edi,%ecx,4) C dst high limb
|
| -
|
| - movl $0, %edx
|
| - cmovc( %eax, %edx) C high is carry if high<divisor
|
| -
|
| - sbbl $0, %ecx C size-1 if high<divisor
|
| -L(no_skip_div):
|
| -
|
| -
|
| -L(start_1c):
|
| - C eax
|
| - C ebx xsize
|
| - C ecx size
|
| - C edx carry
|
| - C esi src
|
| - C edi &dst[xsize-1]
|
| - C ebp divisor
|
| -
|
| - leal (%ebx,%ecx), %eax C size+xsize
|
| - leal -4(%esi,%ecx,4), %esi C &src[size-1]
|
| - leal (%edi,%ecx,4), %edi C &dst[size+xsize-1]
|
| -
|
| - cmpl $MUL_THRESHOLD, %eax
|
| - jae L(mul_by_inverse)
|
| -
|
| -
|
| - orl %ecx, %ecx
|
| - jz L(divide_no_integer) C if size==0
|
| -
|
| -L(divide_integer):
|
| - C eax scratch (quotient)
|
| - C ebx xsize
|
| - C ecx counter
|
| - C edx carry
|
| - C esi src, decrementing
|
| - C edi dst, decrementing
|
| - C ebp divisor
|
| -
|
| - movl (%esi), %eax
|
| - subl $4, %esi
|
| -
|
| - divl %ebp
|
| -
|
| - movl %eax, (%edi)
|
| - subl $4, %edi
|
| -
|
| - subl $1, %ecx
|
| - jnz L(divide_integer)
|
| -
|
| -
|
| -L(divide_no_integer):
|
| - orl %ebx, %ebx
|
| - jnz L(divide_fraction) C if xsize!=0
|
| -
|
| -L(divide_done):
|
| - movl SAVE_ESI, %esi
|
| - movl SAVE_EDI, %edi
|
| - movl SAVE_EBX, %ebx
|
| - movl SAVE_EBP, %ebp
|
| - movl %edx, %eax
|
| - ret
|
| -
|
| -
|
| -L(divide_fraction):
|
| - C eax scratch (quotient)
|
| - C ebx counter
|
| - C ecx
|
| - C edx carry
|
| - C esi
|
| - C edi dst, decrementing
|
| - C ebp divisor
|
| -
|
| - movl $0, %eax
|
| -
|
| - divl %ebp
|
| -
|
| - movl %eax, (%edi)
|
| - subl $4, %edi
|
| -
|
| - subl $1, %ebx
|
| - jnz L(divide_fraction)
|
| -
|
| - jmp L(divide_done)
|
| -
|
| -
|
| -
|
| -C -----------------------------------------------------------------------------
|
| -
|
| -L(mul_by_inverse):
|
| - C eax
|
| - C ebx xsize
|
| - C ecx size
|
| - C edx carry
|
| - C esi &src[size-1]
|
| - C edi &dst[size+xsize-1]
|
| - C ebp divisor
|
| -
|
| - bsrl %ebp, %eax C 31-l
|
| - movd %edx, %mm0 C carry
|
| - movd %edx, %mm1 C carry
|
| - movl %ecx, %edx C size
|
| - movl $31, %ecx
|
| -
|
| - C
|
| -
|
| - xorl %eax, %ecx C l = leading zeros on d
|
| - addl $1, %eax
|
| -
|
| - shll %cl, %ebp C d normalized
|
| - movd %ecx, %mm7 C l
|
| - movl %edx, %ecx C size
|
| -
|
| - movd %eax, %mm6 C 32-l
|
| - movl $-1, %edx
|
| - movl $-1, %eax
|
| -
|
| - C
|
| -
|
| - subl %ebp, %edx C (b-d)-1 so edx:eax = b*(b-d)-1
|
| -
|
| - divl %ebp C floor (b*(b-d)-1 / d)
|
| - movd %ebp, %mm5 C d
|
| -
|
| - C
|
| -
|
| - movd %eax, %mm4 C m
|
| -
|
| -
|
| -L(start_preinv):
|
| - C eax inverse
|
| - C ebx xsize
|
| - C ecx size
|
| - C edx
|
| - C esi &src[size-1]
|
| - C edi &dst[size+xsize-1]
|
| - C ebp
|
| - C
|
| - C mm0 carry
|
| - C mm1 carry
|
| - C mm2
|
| - C mm4 m
|
| - C mm5 d
|
| - C mm6 31-l
|
| - C mm7 l
|
| -
|
| - psllq %mm7, %mm0 C n2 = carry << l, for size==0
|
| -
|
| - subl $1, %ecx
|
| - jb L(integer_none)
|
| -
|
| - movd (%esi), %mm0 C src high limb
|
| - punpckldq %mm1, %mm0
|
| - psrlq %mm6, %mm0 C n2 = high (carry:srchigh << l)
|
| - jz L(integer_last)
|
| -
|
| -
|
| -C The dependent chain here consists of
|
| -C
|
| -C 2 paddd n1+n2
|
| -C 8 pmuludq m*(n1+n2)
|
| -C 2 paddq n2:nadj + m*(n1+n2)
|
| -C 2 psrlq q1
|
| -C 8 pmuludq d*q1
|
| -C 2 psubq (n-d)-q1*d
|
| -C 2 psrlq high n-(q1+1)*d mask
|
| -C 2 pand d masked
|
| -C 2 paddd n2+d addback
|
| -C --
|
| -C 30
|
| -C
|
| -C But it seems to run at 32 cycles, so presumably there's something else
|
| -C going on.
|
| -
|
| - ALIGN(16)
|
| -L(integer_top):
|
| - C eax
|
| - C ebx
|
| - C ecx counter, size-1 to 0
|
| - C edx
|
| - C esi src, decrementing
|
| - C edi dst, decrementing
|
| - C
|
| - C mm0 n2
|
| - C mm4 m
|
| - C mm5 d
|
| - C mm6 32-l
|
| - C mm7 l
|
| -
|
| - ASSERT(b,`C n2<d
|
| - movd %mm0, %eax
|
| - movd %mm5, %edx
|
| - cmpl %edx, %eax')
|
| -
|
| - movd -4(%esi), %mm1 C next src limbs
|
| - movd (%esi), %mm2
|
| - leal -4(%esi), %esi
|
| -
|
| - punpckldq %mm2, %mm1
|
| - psrlq %mm6, %mm1 C n10
|
| -
|
| - movq %mm1, %mm2 C n10
|
| - movq %mm1, %mm3 C n10
|
| - psrad $31, %mm1 C -n1
|
| - pand %mm5, %mm1 C -n1 & d
|
| - paddd %mm2, %mm1 C nadj = n10+(-n1&d), ignore overflow
|
| -
|
| - psrld $31, %mm2 C n1
|
| - paddd %mm0, %mm2 C n2+n1
|
| - punpckldq %mm0, %mm1 C n2:nadj
|
| -
|
| - pmuludq %mm4, %mm2 C m*(n2+n1)
|
| -
|
| - C
|
| -
|
| - paddq %mm2, %mm1 C n2:nadj + m*(n2+n1)
|
| - pxor %mm2, %mm2 C break dependency, saves 4 cycles
|
| - pcmpeqd %mm2, %mm2 C FF...FF
|
| - psrlq $63, %mm2 C 1
|
| -
|
| - psrlq $32, %mm1 C q1 = high(n2:nadj + m*(n2+n1))
|
| -
|
| - paddd %mm1, %mm2 C q1+1
|
| - pmuludq %mm5, %mm1 C q1*d
|
| -
|
| - punpckldq %mm0, %mm3 C n = n2:n10
|
| - pxor %mm0, %mm0
|
| -
|
| - psubq %mm5, %mm3 C n - d
|
| -
|
| - C
|
| -
|
| - psubq %mm1, %mm3 C n - (q1+1)*d
|
| -
|
| - por %mm3, %mm0 C copy remainder -> new n2
|
| - psrlq $32, %mm3 C high n - (q1+1)*d, 0 or -1
|
| -
|
| - ASSERT(be,`C 0 or -1
|
| - movd %mm3, %eax
|
| - addl $1, %eax
|
| - cmpl $1, %eax')
|
| -
|
| - paddd %mm3, %mm2 C q
|
| - pand %mm5, %mm3 C mask & d
|
| -
|
| - paddd %mm3, %mm0 C addback if necessary
|
| - movd %mm2, (%edi)
|
| - leal -4(%edi), %edi
|
| -
|
| - subl $1, %ecx
|
| - ja L(integer_top)
|
| -
|
| -
|
| -L(integer_last):
|
| - C eax
|
| - C ebx xsize
|
| - C ecx
|
| - C edx
|
| - C esi &src[0]
|
| - C edi &dst[xsize]
|
| - C
|
| - C mm0 n2
|
| - C mm4 m
|
| - C mm5 d
|
| - C mm6
|
| - C mm7 l
|
| -
|
| - ASSERT(b,`C n2<d
|
| - movd %mm0, %eax
|
| - movd %mm5, %edx
|
| - cmpl %edx, %eax')
|
| -
|
| - movd (%esi), %mm1 C src[0]
|
| - psllq %mm7, %mm1 C n10
|
| -
|
| - movq %mm1, %mm2 C n10
|
| - movq %mm1, %mm3 C n10
|
| - psrad $31, %mm1 C -n1
|
| - pand %mm5, %mm1 C -n1 & d
|
| - paddd %mm2, %mm1 C nadj = n10+(-n1&d), ignore overflow
|
| -
|
| - psrld $31, %mm2 C n1
|
| - paddd %mm0, %mm2 C n2+n1
|
| - punpckldq %mm0, %mm1 C n2:nadj
|
| -
|
| - pmuludq %mm4, %mm2 C m*(n2+n1)
|
| -
|
| - C
|
| -
|
| - paddq %mm2, %mm1 C n2:nadj + m*(n2+n1)
|
| - pcmpeqd %mm2, %mm2 C FF...FF
|
| - psrlq $63, %mm2 C 1
|
| -
|
| - psrlq $32, %mm1 C q1 = high(n2:nadj + m*(n2+n1))
|
| - paddd %mm1, %mm2 C q1
|
| -
|
| - pmuludq %mm5, %mm1 C q1*d
|
| - punpckldq %mm0, %mm3 C n
|
| - psubq %mm5, %mm3 C n - d
|
| - pxor %mm0, %mm0
|
| -
|
| - C
|
| -
|
| - psubq %mm1, %mm3 C n - (q1+1)*d
|
| -
|
| - por %mm3, %mm0 C remainder -> n2
|
| - psrlq $32, %mm3 C high n - (q1+1)*d, 0 or -1
|
| -
|
| - ASSERT(be,`C 0 or -1
|
| - movd %mm3, %eax
|
| - addl $1, %eax
|
| - cmpl $1, %eax')
|
| -
|
| - paddd %mm3, %mm2 C q
|
| - pand %mm5, %mm3 C mask & d
|
| -
|
| - paddd %mm3, %mm0 C addback if necessary
|
| - movd %mm2, (%edi)
|
| - leal -4(%edi), %edi
|
| -
|
| -
|
| -L(integer_none):
|
| - C eax
|
| - C ebx xsize
|
| -
|
| - orl %ebx, %ebx
|
| - jnz L(fraction_some) C if xsize!=0
|
| -
|
| -
|
| -L(fraction_done):
|
| - movl SAVE_EBP, %ebp
|
| - psrld %mm7, %mm0 C remainder
|
| -
|
| - movl SAVE_EDI, %edi
|
| - movd %mm0, %eax
|
| -
|
| - movl SAVE_ESI, %esi
|
| - movl SAVE_EBX, %ebx
|
| - emms
|
| - ret
|
| -
|
| -
|
| -
|
| -C -----------------------------------------------------------------------------
|
| -C
|
| -
|
| -L(fraction_some):
|
| - C eax
|
| - C ebx xsize
|
| - C ecx
|
| - C edx
|
| - C esi
|
| - C edi &dst[xsize-1]
|
| - C ebp
|
| -
|
| -
|
| -L(fraction_top):
|
| - C eax
|
| - C ebx counter, xsize iterations
|
| - C ecx
|
| - C edx
|
| - C esi src, decrementing
|
| - C edi dst, decrementing
|
| - C
|
| - C mm0 n2
|
| - C mm4 m
|
| - C mm5 d
|
| - C mm6 32-l
|
| - C mm7 l
|
| -
|
| - ASSERT(b,`C n2<d
|
| - movd %mm0, %eax
|
| - movd %mm5, %edx
|
| - cmpl %edx, %eax')
|
| -
|
| - movq %mm0, %mm1 C n2
|
| - pmuludq %mm4, %mm0 C m*n2
|
| -
|
| - pcmpeqd %mm2, %mm2
|
| - psrlq $63, %mm2
|
| -
|
| - C
|
| -
|
| - psrlq $32, %mm0 C high(m*n2)
|
| -
|
| - paddd %mm1, %mm0 C q1 = high(n2:0 + m*n2)
|
| -
|
| - paddd %mm0, %mm2 C q1+1
|
| - pmuludq %mm5, %mm0 C q1*d
|
| -
|
| - psllq $32, %mm1 C n = n2:0
|
| - psubq %mm5, %mm1 C n - d
|
| -
|
| - C
|
| -
|
| - psubq %mm0, %mm1 C r = n - (q1+1)*d
|
| - pxor %mm0, %mm0
|
| -
|
| - por %mm1, %mm0 C r -> n2
|
| - psrlq $32, %mm1 C high n - (q1+1)*d, 0 or -1
|
| -
|
| - ASSERT(be,`C 0 or -1
|
| - movd %mm1, %eax
|
| - addl $1, %eax
|
| - cmpl $1, %eax')
|
| -
|
| - paddd %mm1, %mm2 C q
|
| - pand %mm5, %mm1 C mask & d
|
| -
|
| - paddd %mm1, %mm0 C addback if necessary
|
| - movd %mm2, (%edi)
|
| - leal -4(%edi), %edi
|
| -
|
| - subl $1, %ebx
|
| - jne L(fraction_top)
|
| -
|
| -
|
| - jmp L(fraction_done)
|
| -
|
| -EPILOGUE()
|
|
|