Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(509)

Side by Side Diff: openssl/crypto/bn/asm/modexp512-x86_64.pl

Issue 2072073002: Delete bundled copy of OpenSSL and replace with README. (Closed) Base URL: https://chromium.googlesource.com/chromium/deps/openssl@master
Patch Set: Delete bundled copy of OpenSSL and replace with README. Created 4 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « openssl/crypto/bn/asm/modexp512-x86_64.S ('k') | openssl/crypto/bn/asm/pa-risc2.s » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
(Empty)
1 #!/usr/bin/env perl
2 #
3 # Copyright (c) 2010-2011 Intel Corp.
4 # Author: Vinodh.Gopal@intel.com
5 # Jim Guilford
6 # Erdinc.Ozturk@intel.com
7 # Maxim.Perminov@intel.com
8 #
9 # More information about algorithm used can be found at:
10 # http://www.cse.buffalo.edu/srds2009/escs2009_submission_Gopal.pdf
11 #
12 # ====================================================================
13 # Copyright (c) 2011 The OpenSSL Project. All rights reserved.
14 #
15 # Redistribution and use in source and binary forms, with or without
16 # modification, are permitted provided that the following conditions
17 # are met:
18 #
19 # 1. Redistributions of source code must retain the above copyright
20 # notice, this list of conditions and the following disclaimer.
21 #
22 # 2. Redistributions in binary form must reproduce the above copyright
23 # notice, this list of conditions and the following disclaimer in
24 # the documentation and/or other materials provided with the
25 # distribution.
26 #
27 # 3. All advertising materials mentioning features or use of this
28 # software must display the following acknowledgment:
29 # "This product includes software developed by the OpenSSL Project
30 # for use in the OpenSSL Toolkit. (http://www.OpenSSL.org/)"
31 #
32 # 4. The names "OpenSSL Toolkit" and "OpenSSL Project" must not be used to
33 # endorse or promote products derived from this software without
34 # prior written permission. For written permission, please contact
35 # licensing@OpenSSL.org.
36 #
37 # 5. Products derived from this software may not be called "OpenSSL"
38 # nor may "OpenSSL" appear in their names without prior written
39 # permission of the OpenSSL Project.
40 #
41 # 6. Redistributions of any form whatsoever must retain the following
42 # acknowledgment:
43 # "This product includes software developed by the OpenSSL Project
44 # for use in the OpenSSL Toolkit (http://www.OpenSSL.org/)"
45 #
46 # THIS SOFTWARE IS PROVIDED BY THE OpenSSL PROJECT ``AS IS'' AND ANY
47 # EXPRESSED OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
48 # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
49 # PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE OpenSSL PROJECT OR
50 # ITS CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
51 # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
52 # NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
53 # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
54 # HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
55 # STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
56 # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
57 # OF THE POSSIBILITY OF SUCH DAMAGE.
58 # ====================================================================
59
60 $flavour = shift;
61 $output = shift;
62 if ($flavour =~ /\./) { $output = $flavour; undef $flavour; }
63
64 my $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/);
65
66 $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1;
67 ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or
68 ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or
69 die "can't locate x86_64-xlate.pl";
70
71 open OUT,"| \"$^X\" $xlate $flavour $output";
72 *STDOUT=*OUT;
73
74 use strict;
75 my $code=".text\n\n";
76 my $m=0;
77
78 #
79 # Define x512 macros
80 #
81
82 #MULSTEP_512_ADD MACRO x7, x6, x5, x4, x3, x2, x1, x0, dst, src1, src2, add_src, tmp1, tmp2
83 #
84 # uses rax, rdx, and args
85 sub MULSTEP_512_ADD
86 {
87 my ($x, $DST, $SRC2, $ASRC, $OP, $TMP)=@_;
88 my @X=@$x; # make a copy
89 $code.=<<___;
90 mov (+8*0)($SRC2), %rax
91 mul $OP # rdx:rax = %OP * [0]
92 mov ($ASRC), $X[0]
93 add %rax, $X[0]
94 adc \$0, %rdx
95 mov $X[0], $DST
96 ___
97 for(my $i=1;$i<8;$i++) {
98 $code.=<<___;
99 mov %rdx, $TMP
100
101 mov (+8*$i)($SRC2), %rax
102 mul $OP # rdx:rax = %OP * [$i]
103 mov (+8*$i)($ASRC), $X[$i]
104 add %rax, $X[$i]
105 adc \$0, %rdx
106 add $TMP, $X[$i]
107 adc \$0, %rdx
108 ___
109 }
110 $code.=<<___;
111 mov %rdx, $X[0]
112 ___
113 }
114
115 #MULSTEP_512 MACRO x7, x6, x5, x4, x3, x2, x1, x0, dst, src2, src1_val, tmp
116 #
117 # uses rax, rdx, and args
118 sub MULSTEP_512
119 {
120 my ($x, $DST, $SRC2, $OP, $TMP)=@_;
121 my @X=@$x; # make a copy
122 $code.=<<___;
123 mov (+8*0)($SRC2), %rax
124 mul $OP # rdx:rax = %OP * [0]
125 add %rax, $X[0]
126 adc \$0, %rdx
127 mov $X[0], $DST
128 ___
129 for(my $i=1;$i<8;$i++) {
130 $code.=<<___;
131 mov %rdx, $TMP
132
133 mov (+8*$i)($SRC2), %rax
134 mul $OP # rdx:rax = %OP * [$i]
135 add %rax, $X[$i]
136 adc \$0, %rdx
137 add $TMP, $X[$i]
138 adc \$0, %rdx
139 ___
140 }
141 $code.=<<___;
142 mov %rdx, $X[0]
143 ___
144 }
145
146 #
147 # Swizzle Macros
148 #
149
150 # macro to copy data from flat space to swizzled table
151 #MACRO swizzle pDst, pSrc, tmp1, tmp2
152 # pDst and pSrc are modified
153 sub swizzle
154 {
155 my ($pDst, $pSrc, $cnt, $d0)=@_;
156 $code.=<<___;
157 mov \$8, $cnt
158 loop_$m:
159 mov ($pSrc), $d0
160 mov $d0#w, ($pDst)
161 shr \$16, $d0
162 mov $d0#w, (+64*1)($pDst)
163 shr \$16, $d0
164 mov $d0#w, (+64*2)($pDst)
165 shr \$16, $d0
166 mov $d0#w, (+64*3)($pDst)
167 lea 8($pSrc), $pSrc
168 lea 64*4($pDst), $pDst
169 dec $cnt
170 jnz loop_$m
171 ___
172
173 $m++;
174 }
175
176 # macro to copy data from swizzled table to flat space
177 #MACRO unswizzle pDst, pSrc, tmp*3
178 sub unswizzle
179 {
180 my ($pDst, $pSrc, $cnt, $d0, $d1)=@_;
181 $code.=<<___;
182 mov \$4, $cnt
183 loop_$m:
184 movzxw (+64*3+256*0)($pSrc), $d0
185 movzxw (+64*3+256*1)($pSrc), $d1
186 shl \$16, $d0
187 shl \$16, $d1
188 mov (+64*2+256*0)($pSrc), $d0#w
189 mov (+64*2+256*1)($pSrc), $d1#w
190 shl \$16, $d0
191 shl \$16, $d1
192 mov (+64*1+256*0)($pSrc), $d0#w
193 mov (+64*1+256*1)($pSrc), $d1#w
194 shl \$16, $d0
195 shl \$16, $d1
196 mov (+64*0+256*0)($pSrc), $d0#w
197 mov (+64*0+256*1)($pSrc), $d1#w
198 mov $d0, (+8*0)($pDst)
199 mov $d1, (+8*1)($pDst)
200 lea 256*2($pSrc), $pSrc
201 lea 8*2($pDst), $pDst
202 sub \$1, $cnt
203 jnz loop_$m
204 ___
205
206 $m++;
207 }
208
209 #
210 # Data Structures
211 #
212
213 # Reduce Data
214 #
215 #
216 # Offset Value
217 # 0C0 Carries
218 # 0B8 X2[10]
219 # 0B0 X2[9]
220 # 0A8 X2[8]
221 # 0A0 X2[7]
222 # 098 X2[6]
223 # 090 X2[5]
224 # 088 X2[4]
225 # 080 X2[3]
226 # 078 X2[2]
227 # 070 X2[1]
228 # 068 X2[0]
229 # 060 X1[12] P[10]
230 # 058 X1[11] P[9] Z[8]
231 # 050 X1[10] P[8] Z[7]
232 # 048 X1[9] P[7] Z[6]
233 # 040 X1[8] P[6] Z[5]
234 # 038 X1[7] P[5] Z[4]
235 # 030 X1[6] P[4] Z[3]
236 # 028 X1[5] P[3] Z[2]
237 # 020 X1[4] P[2] Z[1]
238 # 018 X1[3] P[1] Z[0]
239 # 010 X1[2] P[0] Y[2]
240 # 008 X1[1] Q[1] Y[1]
241 # 000 X1[0] Q[0] Y[0]
242
243 my $X1_offset = 0; # 13 qwords
244 my $X2_offset = $X1_offset + 13*8; # 11 qwords
245 my $Carries_offset = $X2_offset + 11*8; # 1 qword
246 my $Q_offset = 0; # 2 qwords
247 my $P_offset = $Q_offset + 2*8; # 11 qwords
248 my $Y_offset = 0; # 3 qwords
249 my $Z_offset = $Y_offset + 3*8; # 9 qwords
250
251 my $Red_Data_Size = $Carries_offset + 1*8; # (25 qw ords)
252
253 #
254 # Stack Frame
255 #
256 #
257 # offset value
258 # ... <old stack contents>
259 # ...
260 # 280 Garray
261
262 # 278 tmp16[15]
263 # ... ...
264 # 200 tmp16[0]
265
266 # 1F8 tmp[7]
267 # ... ...
268 # 1C0 tmp[0]
269
270 # 1B8 GT[7]
271 # ... ...
272 # 180 GT[0]
273
274 # 178 Reduce Data
275 # ... ...
276 # 0B8 Reduce Data
277 # 0B0 reserved
278 # 0A8 reserved
279 # 0A0 reserved
280 # 098 reserved
281 # 090 reserved
282 # 088 reduce result addr
283 # 080 exp[8]
284
285 # ...
286 # 048 exp[1]
287 # 040 exp[0]
288
289 # 038 reserved
290 # 030 loop_idx
291 # 028 pg
292 # 020 i
293 # 018 pData ; arg 4
294 # 010 pG ; arg 2
295 # 008 pResult ; arg 1
296 # 000 rsp ; stack pointer before subtract
297
298 my $rsp_offset = 0;
299 my $pResult_offset = 8*1 + $rsp_offset;
300 my $pG_offset = 8*1 + $pResult_offset;
301 my $pData_offset = 8*1 + $pG_offset;
302 my $i_offset = 8*1 + $pData_offset;
303 my $pg_offset = 8*1 + $i_offset;
304 my $loop_idx_offset = 8*1 + $pg_offset;
305 my $reserved1_offset = 8*1 + $loop_idx_offset;
306 my $exp_offset = 8*1 + $reserved1_offset;
307 my $red_result_addr_offset= 8*9 + $exp_offset;
308 my $reserved2_offset = 8*1 + $red_result_addr_offset;
309 my $Reduce_Data_offset = 8*5 + $reserved2_offset;
310 my $GT_offset = $Red_Data_Size + $Reduce_Data_offset;
311 my $tmp_offset = 8*8 + $GT_offset;
312 my $tmp16_offset = 8*8 + $tmp_offset;
313 my $garray_offset = 8*16 + $tmp16_offset;
314 my $mem_size = 8*8*32 + $garray_offset;
315
316 #
317 # Offsets within Reduce Data
318 #
319 #
320 # struct MODF_2FOLD_MONT_512_C1_DATA {
321 # UINT64 t[8][8];
322 # UINT64 m[8];
323 # UINT64 m1[8]; /* 2^768 % m */
324 # UINT64 m2[8]; /* 2^640 % m */
325 # UINT64 k1[2]; /* (- 1/m) % 2^128 */
326 # };
327
328 my $T = 0;
329 my $M = 512; # = 8 * 8 * 8
330 my $M1 = 576; # = 8 * 8 * 9 /* += 8 * 8 */
331 my $M2 = 640; # = 8 * 8 * 10 /* += 8 * 8 */
332 my $K1 = 704; # = 8 * 8 * 11 /* += 8 * 8 */
333
334 #
335 # FUNCTIONS
336 #
337
338 {{{
339 #
340 # MULADD_128x512 : Function to multiply 128-bits (2 qwords) by 512-bits (8 qword s)
341 # and add 512-bits (8 qwords)
342 # to get 640 bits (10 qwords)
343 # Input: 128-bit mul source: [rdi+8*1], rbp
344 # 512-bit mul source: [rsi+8*n]
345 # 512-bit add source: r15, r14, ..., r9, r8
346 # Output: r9, r8, r15, r14, r13, r12, r11, r10, [rcx+8*1], [rcx+8*0]
347 # Clobbers all regs except: rcx, rsi, rdi
348 $code.=<<___;
349 .type MULADD_128x512,\@abi-omnipotent
350 .align 16
351 MULADD_128x512:
352 ___
353 &MULSTEP_512([map("%r$_",(8..15))], "(+8*0)(%rcx)", "%rsi", "%rbp", "%rb x");
354 $code.=<<___;
355 mov (+8*1)(%rdi), %rbp
356 ___
357 &MULSTEP_512([map("%r$_",(9..15,8))], "(+8*1)(%rcx)", "%rsi", "%rbp", "% rbx");
358 $code.=<<___;
359 ret
360 .size MULADD_128x512,.-MULADD_128x512
361 ___
362 }}}
363
364 {{{
365 #MULADD_256x512 MACRO pDst, pA, pB, OP, TMP, X7, X6, X5, X4, X3, X2, X1, X0
366 #
367 # Inputs: pDst: Destination (768 bits, 12 qwords)
368 # pA: Multiplicand (1024 bits, 16 qwords)
369 # pB: Multiplicand (512 bits, 8 qwords)
370 # Dst = Ah * B + Al
371 # where Ah is (in qwords) A[15:12] (256 bits) and Al is A[7:0] (512 bits)
372 # Results in X3 X2 X1 X0 X7 X6 X5 X4 Dst[3:0]
373 # Uses registers: arguments, RAX, RDX
374 sub MULADD_256x512
375 {
376 my ($pDst, $pA, $pB, $OP, $TMP, $X)=@_;
377 $code.=<<___;
378 mov (+8*12)($pA), $OP
379 ___
380 &MULSTEP_512_ADD($X, "(+8*0)($pDst)", $pB, $pA, $OP, $TMP);
381 push(@$X,shift(@$X));
382
383 $code.=<<___;
384 mov (+8*13)($pA), $OP
385 ___
386 &MULSTEP_512($X, "(+8*1)($pDst)", $pB, $OP, $TMP);
387 push(@$X,shift(@$X));
388
389 $code.=<<___;
390 mov (+8*14)($pA), $OP
391 ___
392 &MULSTEP_512($X, "(+8*2)($pDst)", $pB, $OP, $TMP);
393 push(@$X,shift(@$X));
394
395 $code.=<<___;
396 mov (+8*15)($pA), $OP
397 ___
398 &MULSTEP_512($X, "(+8*3)($pDst)", $pB, $OP, $TMP);
399 push(@$X,shift(@$X));
400 }
401
402 #
403 # mont_reduce(UINT64 *x, /* 1024 bits, 16 qwords */
404 # UINT64 *m, /* 512 bits, 8 qwords */
405 # MODF_2FOLD_MONT_512_C1_DATA *data,
406 # UINT64 *r) /* 512 bits, 8 qwords */
407 # Input: x (number to be reduced): tmp16 (Implicit)
408 # m (modulus): [pM] (Implicit)
409 # data (reduce data): [pData] (Implicit)
410 # Output: r (result): Address in [red_res_addr]
411 # result also in: r9, r8, r15, r14, r13, r12, r11, r10
412
413 my @X=map("%r$_",(8..15));
414
415 $code.=<<___;
416 .type mont_reduce,\@abi-omnipotent
417 .align 16
418 mont_reduce:
419 ___
420
421 my $STACK_DEPTH = 8;
422 #
423 # X1 = Xh * M1 + Xl
424 $code.=<<___;
425 lea (+$Reduce_Data_offset+$X1_offset+$STACK_DEPTH)(%rsp), %rdi # pX1 (Dst) 769 bits, 13 qwords
426 mov (+$pData_offset+$STACK_DEPTH)(%rsp), %rsi # pM1 (Bsrc) 512 bits, 8 qwords
427 add \$$M1, %rsi
428 lea (+$tmp16_offset+$STACK_DEPTH)(%rsp), %rcx # X (Asrc) 1024 bits, 16 qwords
429
430 ___
431
432 &MULADD_256x512("%rdi", "%rcx", "%rsi", "%rbp", "%rbx", \@X); # rotate s @X 4 times
433 # results in r11, r10, r9, r8, r15, r14, r13, r12, X1[3:0]
434
435 $code.=<<___;
436 xor %rax, %rax
437 # X1 += xl
438 add (+8*8)(%rcx), $X[4]
439 adc (+8*9)(%rcx), $X[5]
440 adc (+8*10)(%rcx), $X[6]
441 adc (+8*11)(%rcx), $X[7]
442 adc \$0, %rax
443 # X1 is now rax, r11-r8, r15-r12, tmp16[3:0]
444
445 #
446 # check for carry ;; carry stored in rax
447 mov $X[4], (+8*8)(%rdi) # rdi points to X1
448 mov $X[5], (+8*9)(%rdi)
449 mov $X[6], %rbp
450 mov $X[7], (+8*11)(%rdi)
451
452 mov %rax, (+$Reduce_Data_offset+$Carries_offset+$STACK_DEPTH)(%rsp)
453
454 mov (+8*0)(%rdi), $X[4]
455 mov (+8*1)(%rdi), $X[5]
456 mov (+8*2)(%rdi), $X[6]
457 mov (+8*3)(%rdi), $X[7]
458
459 # X1 is now stored in: X1[11], rbp, X1[9:8], r15-r8
460 # rdi -> X1
461 # rsi -> M1
462
463 #
464 # X2 = Xh * M2 + Xl
465 # do first part (X2 = Xh * M2)
466 add \$8*10, %rdi # rdi -> pXh ; 128 bits, 2 qword s
467 # Xh is actually { [rdi+8*1], rbp }
468 add \$`$M2-$M1`, %rsi # rsi -> M2
469 lea (+$Reduce_Data_offset+$X2_offset+$STACK_DEPTH)(%rsp), %rcx # rcx -> pX2 ; 641 bits, 11 qwords
470 ___
471 unshift(@X,pop(@X)); unshift(@X,pop(@X));
472 $code.=<<___;
473
474 call MULADD_128x512 # args in rcx, rdi / rbp, rsi, r 15-r8
475 # result in r9, r8, r15, r14, r13, r12, r11, r10, X2[1:0]
476 mov (+$Reduce_Data_offset+$Carries_offset+$STACK_DEPTH)(%rsp), %rax
477
478 # X2 += Xl
479 add (+8*8-8*10)(%rdi), $X[6] # (-8*10) is to adjust r di -> Xh to Xl
480 adc (+8*9-8*10)(%rdi), $X[7]
481 mov $X[6], (+8*8)(%rcx)
482 mov $X[7], (+8*9)(%rcx)
483
484 adc %rax, %rax
485 mov %rax, (+$Reduce_Data_offset+$Carries_offset+$STACK_DEPTH)(%rsp)
486
487 lea (+$Reduce_Data_offset+$Q_offset+$STACK_DEPTH)(%rsp), %rdi # rdi -> pQ ; 128 bits, 2 qwords
488 add \$`$K1-$M2`, %rsi # rsi -> pK1 ; 128 bits, 2 qwords
489
490 # MUL_128x128t128 rdi, rcx, rsi ; Q = X2 * K1 (bottom half)
491 # B1:B0 = rsi[1:0] = K1[1:0]
492 # A1:A0 = rcx[1:0] = X2[1:0]
493 # Result = rdi[1],rbp = Q[1],rbp
494 mov (%rsi), %r8 # B0
495 mov (+8*1)(%rsi), %rbx # B1
496
497 mov (%rcx), %rax # A0
498 mul %r8 # B0
499 mov %rax, %rbp
500 mov %rdx, %r9
501
502 mov (+8*1)(%rcx), %rax # A1
503 mul %r8 # B0
504 add %rax, %r9
505
506 mov (%rcx), %rax # A0
507 mul %rbx # B1
508 add %rax, %r9
509
510 mov %r9, (+8*1)(%rdi)
511 # end MUL_128x128t128
512
513 sub \$`$K1-$M`, %rsi
514
515 mov (%rcx), $X[6]
516 mov (+8*1)(%rcx), $X[7] # r9:r8 = X2[1:0]
517
518 call MULADD_128x512 # args in rcx, rdi / rbp, rsi, r 15-r8
519 # result in r9, r8, r15, r14, r13, r12, r11, r10, X2[1:0]
520
521 # load first half of m to rdx, rdi, rbx, rax
522 # moved this here for efficiency
523 mov (+8*0)(%rsi), %rax
524 mov (+8*1)(%rsi), %rbx
525 mov (+8*2)(%rsi), %rdi
526 mov (+8*3)(%rsi), %rdx
527
528 # continue with reduction
529 mov (+$Reduce_Data_offset+$Carries_offset+$STACK_DEPTH)(%rsp), %rbp
530
531 add (+8*8)(%rcx), $X[6]
532 adc (+8*9)(%rcx), $X[7]
533
534 #accumulate the final carry to rbp
535 adc %rbp, %rbp
536
537 # Add in overflow corrections: R = (X2>>128) += T[overflow]
538 # R = {r9, r8, r15, r14, ..., r10}
539 shl \$3, %rbp
540 mov (+$pData_offset+$STACK_DEPTH)(%rsp), %rcx # rsi -> Data (and points to T)
541 add %rcx, %rbp # pT ; 512 bits, 8 qwords, sprea d out
542
543 # rsi will be used to generate a mask after the addition
544 xor %rsi, %rsi
545
546 add (+8*8*0)(%rbp), $X[0]
547 adc (+8*8*1)(%rbp), $X[1]
548 adc (+8*8*2)(%rbp), $X[2]
549 adc (+8*8*3)(%rbp), $X[3]
550 adc (+8*8*4)(%rbp), $X[4]
551 adc (+8*8*5)(%rbp), $X[5]
552 adc (+8*8*6)(%rbp), $X[6]
553 adc (+8*8*7)(%rbp), $X[7]
554
555 # if there is a carry: rsi = 0xFFFFFFFFFFFFFFFF
556 # if carry is clear: rsi = 0x0000000000000000
557 sbb \$0, %rsi
558
559 # if carry is clear, subtract 0. Otherwise, subtract 256 bits of m
560 and %rsi, %rax
561 and %rsi, %rbx
562 and %rsi, %rdi
563 and %rsi, %rdx
564
565 mov \$1, %rbp
566 sub %rax, $X[0]
567 sbb %rbx, $X[1]
568 sbb %rdi, $X[2]
569 sbb %rdx, $X[3]
570
571 # if there is a borrow: rbp = 0
572 # if there is no borrow: rbp = 1
573 # this is used to save the borrows in between the first half and the 2nd half of the subtraction of m
574 sbb \$0, %rbp
575
576 #load second half of m to rdx, rdi, rbx, rax
577
578 add \$$M, %rcx
579 mov (+8*4)(%rcx), %rax
580 mov (+8*5)(%rcx), %rbx
581 mov (+8*6)(%rcx), %rdi
582 mov (+8*7)(%rcx), %rdx
583
584 # use the rsi mask as before
585 # if carry is clear, subtract 0. Otherwise, subtract 256 bits of m
586 and %rsi, %rax
587 and %rsi, %rbx
588 and %rsi, %rdi
589 and %rsi, %rdx
590
591 # if rbp = 0, there was a borrow before, it is moved to the carry flag
592 # if rbp = 1, there was not a borrow before, carry flag is cleared
593 sub \$1, %rbp
594
595 sbb %rax, $X[4]
596 sbb %rbx, $X[5]
597 sbb %rdi, $X[6]
598 sbb %rdx, $X[7]
599
600 # write R back to memory
601
602 mov (+$red_result_addr_offset+$STACK_DEPTH)(%rsp), %rsi
603 mov $X[0], (+8*0)(%rsi)
604 mov $X[1], (+8*1)(%rsi)
605 mov $X[2], (+8*2)(%rsi)
606 mov $X[3], (+8*3)(%rsi)
607 mov $X[4], (+8*4)(%rsi)
608 mov $X[5], (+8*5)(%rsi)
609 mov $X[6], (+8*6)(%rsi)
610 mov $X[7], (+8*7)(%rsi)
611
612 ret
613 .size mont_reduce,.-mont_reduce
614 ___
615 }}}
616
617 {{{
618 #MUL_512x512 MACRO pDst, pA, pB, x7, x6, x5, x4, x3, x2, x1, x0, tmp*2
619 #
620 # Inputs: pDst: Destination (1024 bits, 16 qwords)
621 # pA: Multiplicand (512 bits, 8 qwords)
622 # pB: Multiplicand (512 bits, 8 qwords)
623 # Uses registers rax, rdx, args
624 # B operand in [pB] and also in x7...x0
625 sub MUL_512x512
626 {
627 my ($pDst, $pA, $pB, $x, $OP, $TMP, $pDst_o)=@_;
628 my ($pDst, $pDst_o) = ($pDst =~ m/([^+]*)\+?(.*)?/);
629 my @X=@$x; # make a copy
630
631 $code.=<<___;
632 mov (+8*0)($pA), $OP
633
634 mov $X[0], %rax
635 mul $OP # rdx:rax = %OP * [0]
636 mov %rax, (+$pDst_o+8*0)($pDst)
637 mov %rdx, $X[0]
638 ___
639 for(my $i=1;$i<8;$i++) {
640 $code.=<<___;
641 mov $X[$i], %rax
642 mul $OP # rdx:rax = %OP * [$i]
643 add %rax, $X[$i-1]
644 adc \$0, %rdx
645 mov %rdx, $X[$i]
646 ___
647 }
648
649 for(my $i=1;$i<8;$i++) {
650 $code.=<<___;
651 mov (+8*$i)($pA), $OP
652 ___
653
654 &MULSTEP_512(\@X, "(+$pDst_o+8*$i)($pDst)", $pB, $OP, $TMP);
655 push(@X,shift(@X));
656 }
657
658 $code.=<<___;
659 mov $X[0], (+$pDst_o+8*8)($pDst)
660 mov $X[1], (+$pDst_o+8*9)($pDst)
661 mov $X[2], (+$pDst_o+8*10)($pDst)
662 mov $X[3], (+$pDst_o+8*11)($pDst)
663 mov $X[4], (+$pDst_o+8*12)($pDst)
664 mov $X[5], (+$pDst_o+8*13)($pDst)
665 mov $X[6], (+$pDst_o+8*14)($pDst)
666 mov $X[7], (+$pDst_o+8*15)($pDst)
667 ___
668 }
669
670 #
671 # mont_mul_a3b : subroutine to compute (Src1 * Src2) % M (all 512-bits)
672 # Input: src1: Address of source 1: rdi
673 # src2: Address of source 2: rsi
674 # Output: dst: Address of destination: [red_res_addr]
675 # src2 and result also in: r9, r8, r15, r14, r13, r12, r11, r10
676 # Temp: Clobbers [tmp16], all registers
677 $code.=<<___;
678 .type mont_mul_a3b,\@abi-omnipotent
679 .align 16
680 mont_mul_a3b:
681 #
682 # multiply tmp = src1 * src2
683 # For multiply: dst = rcx, src1 = rdi, src2 = rsi
684 # stack depth is extra 8 from call
685 ___
686 &MUL_512x512("%rsp+$tmp16_offset+8", "%rdi", "%rsi", [map("%r$_",(10..15 ,8..9))], "%rbp", "%rbx");
687 $code.=<<___;
688 #
689 # Dst = tmp % m
690 # Call reduce(tmp, m, data, dst)
691
692 # tail recursion optimization: jmp to mont_reduce and return from there
693 jmp mont_reduce
694 # call mont_reduce
695 # ret
696 .size mont_mul_a3b,.-mont_mul_a3b
697 ___
698 }}}
699
700 {{{
701 #SQR_512 MACRO pDest, pA, x7, x6, x5, x4, x3, x2, x1, x0, tmp*4
702 #
703 # Input in memory [pA] and also in x7...x0
704 # Uses all argument registers plus rax and rdx
705 #
706 # This version computes all of the off-diagonal terms into memory,
707 # and then it adds in the diagonal terms
708
709 sub SQR_512
710 {
711 my ($pDst, $pA, $x, $A, $tmp, $x7, $x6, $pDst_o)=@_;
712 my ($pDst, $pDst_o) = ($pDst =~ m/([^+]*)\+?(.*)?/);
713 my @X=@$x; # make a copy
714 $code.=<<___;
715 # ------------------
716 # first pass 01...07
717 # ------------------
718 mov $X[0], $A
719
720 mov $X[1],%rax
721 mul $A
722 mov %rax, (+$pDst_o+8*1)($pDst)
723 ___
724 for(my $i=2;$i<8;$i++) {
725 $code.=<<___;
726 mov %rdx, $X[$i-2]
727 mov $X[$i],%rax
728 mul $A
729 add %rax, $X[$i-2]
730 adc \$0, %rdx
731 ___
732 }
733 $code.=<<___;
734 mov %rdx, $x7
735
736 mov $X[0], (+$pDst_o+8*2)($pDst)
737
738 # ------------------
739 # second pass 12...17
740 # ------------------
741
742 mov (+8*1)($pA), $A
743
744 mov (+8*2)($pA),%rax
745 mul $A
746 add %rax, $X[1]
747 adc \$0, %rdx
748 mov $X[1], (+$pDst_o+8*3)($pDst)
749
750 mov %rdx, $X[0]
751 mov (+8*3)($pA),%rax
752 mul $A
753 add %rax, $X[2]
754 adc \$0, %rdx
755 add $X[0], $X[2]
756 adc \$0, %rdx
757 mov $X[2], (+$pDst_o+8*4)($pDst)
758
759 mov %rdx, $X[0]
760 mov (+8*4)($pA),%rax
761 mul $A
762 add %rax, $X[3]
763 adc \$0, %rdx
764 add $X[0], $X[3]
765 adc \$0, %rdx
766
767 mov %rdx, $X[0]
768 mov (+8*5)($pA),%rax
769 mul $A
770 add %rax, $X[4]
771 adc \$0, %rdx
772 add $X[0], $X[4]
773 adc \$0, %rdx
774
775 mov %rdx, $X[0]
776 mov $X[6],%rax
777 mul $A
778 add %rax, $X[5]
779 adc \$0, %rdx
780 add $X[0], $X[5]
781 adc \$0, %rdx
782
783 mov %rdx, $X[0]
784 mov $X[7],%rax
785 mul $A
786 add %rax, $x7
787 adc \$0, %rdx
788 add $X[0], $x7
789 adc \$0, %rdx
790
791 mov %rdx, $X[1]
792
793 # ------------------
794 # third pass 23...27
795 # ------------------
796 mov (+8*2)($pA), $A
797
798 mov (+8*3)($pA),%rax
799 mul $A
800 add %rax, $X[3]
801 adc \$0, %rdx
802 mov $X[3], (+$pDst_o+8*5)($pDst)
803
804 mov %rdx, $X[0]
805 mov (+8*4)($pA),%rax
806 mul $A
807 add %rax, $X[4]
808 adc \$0, %rdx
809 add $X[0], $X[4]
810 adc \$0, %rdx
811 mov $X[4], (+$pDst_o+8*6)($pDst)
812
813 mov %rdx, $X[0]
814 mov (+8*5)($pA),%rax
815 mul $A
816 add %rax, $X[5]
817 adc \$0, %rdx
818 add $X[0], $X[5]
819 adc \$0, %rdx
820
821 mov %rdx, $X[0]
822 mov $X[6],%rax
823 mul $A
824 add %rax, $x7
825 adc \$0, %rdx
826 add $X[0], $x7
827 adc \$0, %rdx
828
829 mov %rdx, $X[0]
830 mov $X[7],%rax
831 mul $A
832 add %rax, $X[1]
833 adc \$0, %rdx
834 add $X[0], $X[1]
835 adc \$0, %rdx
836
837 mov %rdx, $X[2]
838
839 # ------------------
840 # fourth pass 34...37
841 # ------------------
842
843 mov (+8*3)($pA), $A
844
845 mov (+8*4)($pA),%rax
846 mul $A
847 add %rax, $X[5]
848 adc \$0, %rdx
849 mov $X[5], (+$pDst_o+8*7)($pDst)
850
851 mov %rdx, $X[0]
852 mov (+8*5)($pA),%rax
853 mul $A
854 add %rax, $x7
855 adc \$0, %rdx
856 add $X[0], $x7
857 adc \$0, %rdx
858 mov $x7, (+$pDst_o+8*8)($pDst)
859
860 mov %rdx, $X[0]
861 mov $X[6],%rax
862 mul $A
863 add %rax, $X[1]
864 adc \$0, %rdx
865 add $X[0], $X[1]
866 adc \$0, %rdx
867
868 mov %rdx, $X[0]
869 mov $X[7],%rax
870 mul $A
871 add %rax, $X[2]
872 adc \$0, %rdx
873 add $X[0], $X[2]
874 adc \$0, %rdx
875
876 mov %rdx, $X[5]
877
878 # ------------------
879 # fifth pass 45...47
880 # ------------------
881 mov (+8*4)($pA), $A
882
883 mov (+8*5)($pA),%rax
884 mul $A
885 add %rax, $X[1]
886 adc \$0, %rdx
887 mov $X[1], (+$pDst_o+8*9)($pDst)
888
889 mov %rdx, $X[0]
890 mov $X[6],%rax
891 mul $A
892 add %rax, $X[2]
893 adc \$0, %rdx
894 add $X[0], $X[2]
895 adc \$0, %rdx
896 mov $X[2], (+$pDst_o+8*10)($pDst)
897
898 mov %rdx, $X[0]
899 mov $X[7],%rax
900 mul $A
901 add %rax, $X[5]
902 adc \$0, %rdx
903 add $X[0], $X[5]
904 adc \$0, %rdx
905
906 mov %rdx, $X[1]
907
908 # ------------------
909 # sixth pass 56...57
910 # ------------------
911 mov (+8*5)($pA), $A
912
913 mov $X[6],%rax
914 mul $A
915 add %rax, $X[5]
916 adc \$0, %rdx
917 mov $X[5], (+$pDst_o+8*11)($pDst)
918
919 mov %rdx, $X[0]
920 mov $X[7],%rax
921 mul $A
922 add %rax, $X[1]
923 adc \$0, %rdx
924 add $X[0], $X[1]
925 adc \$0, %rdx
926 mov $X[1], (+$pDst_o+8*12)($pDst)
927
928 mov %rdx, $X[2]
929
930 # ------------------
931 # seventh pass 67
932 # ------------------
933 mov $X[6], $A
934
935 mov $X[7],%rax
936 mul $A
937 add %rax, $X[2]
938 adc \$0, %rdx
939 mov $X[2], (+$pDst_o+8*13)($pDst)
940
941 mov %rdx, (+$pDst_o+8*14)($pDst)
942
943 # start finalize (add in squares, and double off-terms)
944 mov (+$pDst_o+8*1)($pDst), $X[0]
945 mov (+$pDst_o+8*2)($pDst), $X[1]
946 mov (+$pDst_o+8*3)($pDst), $X[2]
947 mov (+$pDst_o+8*4)($pDst), $X[3]
948 mov (+$pDst_o+8*5)($pDst), $X[4]
949 mov (+$pDst_o+8*6)($pDst), $X[5]
950
951 mov (+8*3)($pA), %rax
952 mul %rax
953 mov %rax, $x6
954 mov %rdx, $X[6]
955
956 add $X[0], $X[0]
957 adc $X[1], $X[1]
958 adc $X[2], $X[2]
959 adc $X[3], $X[3]
960 adc $X[4], $X[4]
961 adc $X[5], $X[5]
962 adc \$0, $X[6]
963
964 mov (+8*0)($pA), %rax
965 mul %rax
966 mov %rax, (+$pDst_o+8*0)($pDst)
967 mov %rdx, $A
968
969 mov (+8*1)($pA), %rax
970 mul %rax
971
972 add $A, $X[0]
973 adc %rax, $X[1]
974 adc \$0, %rdx
975
976 mov %rdx, $A
977 mov $X[0], (+$pDst_o+8*1)($pDst)
978 mov $X[1], (+$pDst_o+8*2)($pDst)
979
980 mov (+8*2)($pA), %rax
981 mul %rax
982
983 add $A, $X[2]
984 adc %rax, $X[3]
985 adc \$0, %rdx
986
987 mov %rdx, $A
988
989 mov $X[2], (+$pDst_o+8*3)($pDst)
990 mov $X[3], (+$pDst_o+8*4)($pDst)
991
992 xor $tmp, $tmp
993 add $A, $X[4]
994 adc $x6, $X[5]
995 adc \$0, $tmp
996
997 mov $X[4], (+$pDst_o+8*5)($pDst)
998 mov $X[5], (+$pDst_o+8*6)($pDst)
999
1000 # %%tmp has 0/1 in column 7
1001 # %%A6 has a full value in column 7
1002
1003 mov (+$pDst_o+8*7)($pDst), $X[0]
1004 mov (+$pDst_o+8*8)($pDst), $X[1]
1005 mov (+$pDst_o+8*9)($pDst), $X[2]
1006 mov (+$pDst_o+8*10)($pDst), $X[3]
1007 mov (+$pDst_o+8*11)($pDst), $X[4]
1008 mov (+$pDst_o+8*12)($pDst), $X[5]
1009 mov (+$pDst_o+8*13)($pDst), $x6
1010 mov (+$pDst_o+8*14)($pDst), $x7
1011
1012 mov $X[7], %rax
1013 mul %rax
1014 mov %rax, $X[7]
1015 mov %rdx, $A
1016
1017 add $X[0], $X[0]
1018 adc $X[1], $X[1]
1019 adc $X[2], $X[2]
1020 adc $X[3], $X[3]
1021 adc $X[4], $X[4]
1022 adc $X[5], $X[5]
1023 adc $x6, $x6
1024 adc $x7, $x7
1025 adc \$0, $A
1026
1027 add $tmp, $X[0]
1028
1029 mov (+8*4)($pA), %rax
1030 mul %rax
1031
1032 add $X[6], $X[0]
1033 adc %rax, $X[1]
1034 adc \$0, %rdx
1035
1036 mov %rdx, $tmp
1037
1038 mov $X[0], (+$pDst_o+8*7)($pDst)
1039 mov $X[1], (+$pDst_o+8*8)($pDst)
1040
1041 mov (+8*5)($pA), %rax
1042 mul %rax
1043
1044 add $tmp, $X[2]
1045 adc %rax, $X[3]
1046 adc \$0, %rdx
1047
1048 mov %rdx, $tmp
1049
1050 mov $X[2], (+$pDst_o+8*9)($pDst)
1051 mov $X[3], (+$pDst_o+8*10)($pDst)
1052
1053 mov (+8*6)($pA), %rax
1054 mul %rax
1055
1056 add $tmp, $X[4]
1057 adc %rax, $X[5]
1058 adc \$0, %rdx
1059
1060 mov $X[4], (+$pDst_o+8*11)($pDst)
1061 mov $X[5], (+$pDst_o+8*12)($pDst)
1062
1063 add %rdx, $x6
1064 adc $X[7], $x7
1065 adc \$0, $A
1066
1067 mov $x6, (+$pDst_o+8*13)($pDst)
1068 mov $x7, (+$pDst_o+8*14)($pDst)
1069 mov $A, (+$pDst_o+8*15)($pDst)
1070 ___
1071 }
1072
1073 #
1074 # sqr_reduce: subroutine to compute Result = reduce(Result * Result)
1075 #
1076 # input and result also in: r9, r8, r15, r14, r13, r12, r11, r10
1077 #
1078 $code.=<<___;
1079 .type sqr_reduce,\@abi-omnipotent
1080 .align 16
1081 sqr_reduce:
1082 mov (+$pResult_offset+8)(%rsp), %rcx
1083 ___
1084 &SQR_512("%rsp+$tmp16_offset+8", "%rcx", [map("%r$_",(10..15,8..9))], "% rbx", "%rbp", "%rsi", "%rdi");
1085 $code.=<<___;
1086 # tail recursion optimization: jmp to mont_reduce and return from there
1087 jmp mont_reduce
1088 # call mont_reduce
1089 # ret
1090 .size sqr_reduce,.-sqr_reduce
1091 ___
1092 }}}
1093
1094 #
1095 # MAIN FUNCTION
1096 #
1097
1098 #mod_exp_512(UINT64 *result, /* 512 bits, 8 qwords */
1099 # UINT64 *g, /* 512 bits, 8 qwords */
1100 # UINT64 *exp, /* 512 bits, 8 qwords */
1101 # struct mod_ctx_512 *data)
1102
1103 # window size = 5
1104 # table size = 2^5 = 32
1105 #table_entries equ 32
1106 #table_size equ table_entries * 8
1107 $code.=<<___;
1108 .globl mod_exp_512
1109 .type mod_exp_512,\@function,4
1110 mod_exp_512:
1111 push %rbp
1112 push %rbx
1113 push %r12
1114 push %r13
1115 push %r14
1116 push %r15
1117
1118 # adjust stack down and then align it with cache boundary
1119 mov %rsp, %r8
1120 sub \$$mem_size, %rsp
1121 and \$-64, %rsp
1122
1123 # store previous stack pointer and arguments
1124 mov %r8, (+$rsp_offset)(%rsp)
1125 mov %rdi, (+$pResult_offset)(%rsp)
1126 mov %rsi, (+$pG_offset)(%rsp)
1127 mov %rcx, (+$pData_offset)(%rsp)
1128 .Lbody:
1129 # transform g into montgomery space
1130 # GT = reduce(g * C2) = reduce(g * (2^256))
1131 # reduce expects to have the input in [tmp16]
1132 pxor %xmm4, %xmm4
1133 movdqu (+16*0)(%rsi), %xmm0
1134 movdqu (+16*1)(%rsi), %xmm1
1135 movdqu (+16*2)(%rsi), %xmm2
1136 movdqu (+16*3)(%rsi), %xmm3
1137 movdqa %xmm4, (+$tmp16_offset+16*0)(%rsp)
1138 movdqa %xmm4, (+$tmp16_offset+16*1)(%rsp)
1139 movdqa %xmm4, (+$tmp16_offset+16*6)(%rsp)
1140 movdqa %xmm4, (+$tmp16_offset+16*7)(%rsp)
1141 movdqa %xmm0, (+$tmp16_offset+16*2)(%rsp)
1142 movdqa %xmm1, (+$tmp16_offset+16*3)(%rsp)
1143 movdqa %xmm2, (+$tmp16_offset+16*4)(%rsp)
1144 movdqa %xmm3, (+$tmp16_offset+16*5)(%rsp)
1145
1146 # load pExp before rdx gets blown away
1147 movdqu (+16*0)(%rdx), %xmm0
1148 movdqu (+16*1)(%rdx), %xmm1
1149 movdqu (+16*2)(%rdx), %xmm2
1150 movdqu (+16*3)(%rdx), %xmm3
1151
1152 lea (+$GT_offset)(%rsp), %rbx
1153 mov %rbx, (+$red_result_addr_offset)(%rsp)
1154 call mont_reduce
1155
1156 # Initialize tmp = C
1157 lea (+$tmp_offset)(%rsp), %rcx
1158 xor %rax, %rax
1159 mov %rax, (+8*0)(%rcx)
1160 mov %rax, (+8*1)(%rcx)
1161 mov %rax, (+8*3)(%rcx)
1162 mov %rax, (+8*4)(%rcx)
1163 mov %rax, (+8*5)(%rcx)
1164 mov %rax, (+8*6)(%rcx)
1165 mov %rax, (+8*7)(%rcx)
1166 mov %rax, (+$exp_offset+8*8)(%rsp)
1167 movq \$1, (+8*2)(%rcx)
1168
1169 lea (+$garray_offset)(%rsp), %rbp
1170 mov %rcx, %rsi # pTmp
1171 mov %rbp, %rdi # Garray[][0]
1172 ___
1173
1174 &swizzle("%rdi", "%rcx", "%rax", "%rbx");
1175
1176 # for (rax = 31; rax != 0; rax--) {
1177 # tmp = reduce(tmp * G)
1178 # swizzle(pg, tmp);
1179 # pg += 2; }
1180 $code.=<<___;
1181 mov \$31, %rax
1182 mov %rax, (+$i_offset)(%rsp)
1183 mov %rbp, (+$pg_offset)(%rsp)
1184 # rsi -> pTmp
1185 mov %rsi, (+$red_result_addr_offset)(%rsp)
1186 mov (+8*0)(%rsi), %r10
1187 mov (+8*1)(%rsi), %r11
1188 mov (+8*2)(%rsi), %r12
1189 mov (+8*3)(%rsi), %r13
1190 mov (+8*4)(%rsi), %r14
1191 mov (+8*5)(%rsi), %r15
1192 mov (+8*6)(%rsi), %r8
1193 mov (+8*7)(%rsi), %r9
1194 init_loop:
1195 lea (+$GT_offset)(%rsp), %rdi
1196 call mont_mul_a3b
1197 lea (+$tmp_offset)(%rsp), %rsi
1198 mov (+$pg_offset)(%rsp), %rbp
1199 add \$2, %rbp
1200 mov %rbp, (+$pg_offset)(%rsp)
1201 mov %rsi, %rcx # rcx = rsi = addr of tmp
1202 ___
1203
1204 &swizzle("%rbp", "%rcx", "%rax", "%rbx");
1205 $code.=<<___;
1206 mov (+$i_offset)(%rsp), %rax
1207 sub \$1, %rax
1208 mov %rax, (+$i_offset)(%rsp)
1209 jne init_loop
1210
1211 #
1212 # Copy exponent onto stack
1213 movdqa %xmm0, (+$exp_offset+16*0)(%rsp)
1214 movdqa %xmm1, (+$exp_offset+16*1)(%rsp)
1215 movdqa %xmm2, (+$exp_offset+16*2)(%rsp)
1216 movdqa %xmm3, (+$exp_offset+16*3)(%rsp)
1217
1218
1219 #
1220 # Do exponentiation
1221 # Initialize result to G[exp{511:507}]
1222 mov (+$exp_offset+62)(%rsp), %eax
1223 mov %rax, %rdx
1224 shr \$11, %rax
1225 and \$0x07FF, %edx
1226 mov %edx, (+$exp_offset+62)(%rsp)
1227 lea (+$garray_offset)(%rsp,%rax,2), %rsi
1228 mov (+$pResult_offset)(%rsp), %rdx
1229 ___
1230
1231 &unswizzle("%rdx", "%rsi", "%rbp", "%rbx", "%rax");
1232
1233 #
1234 # Loop variables
1235 # rcx = [loop_idx] = index: 510-5 to 0 by 5
1236 $code.=<<___;
1237 movq \$505, (+$loop_idx_offset)(%rsp)
1238
1239 mov (+$pResult_offset)(%rsp), %rcx
1240 mov %rcx, (+$red_result_addr_offset)(%rsp)
1241 mov (+8*0)(%rcx), %r10
1242 mov (+8*1)(%rcx), %r11
1243 mov (+8*2)(%rcx), %r12
1244 mov (+8*3)(%rcx), %r13
1245 mov (+8*4)(%rcx), %r14
1246 mov (+8*5)(%rcx), %r15
1247 mov (+8*6)(%rcx), %r8
1248 mov (+8*7)(%rcx), %r9
1249 jmp sqr_2
1250
1251 main_loop_a3b:
1252 call sqr_reduce
1253 call sqr_reduce
1254 call sqr_reduce
1255 sqr_2:
1256 call sqr_reduce
1257 call sqr_reduce
1258
1259 #
1260 # Do multiply, first look up proper value in Garray
1261 mov (+$loop_idx_offset)(%rsp), %rcx # bit index
1262 mov %rcx, %rax
1263 shr \$4, %rax # rax is word pointer
1264 mov (+$exp_offset)(%rsp,%rax,2), %edx
1265 and \$15, %rcx
1266 shrq %cl, %rdx
1267 and \$0x1F, %rdx
1268
1269 lea (+$garray_offset)(%rsp,%rdx,2), %rsi
1270 lea (+$tmp_offset)(%rsp), %rdx
1271 mov %rdx, %rdi
1272 ___
1273
1274 &unswizzle("%rdx", "%rsi", "%rbp", "%rbx", "%rax");
1275 # rdi = tmp = pG
1276
1277 #
1278 # Call mod_mul_a1(pDst, pSrc1, pSrc2, pM, pData)
1279 # result result pG M Data
1280 $code.=<<___;
1281 mov (+$pResult_offset)(%rsp), %rsi
1282 call mont_mul_a3b
1283
1284 #
1285 # finish loop
1286 mov (+$loop_idx_offset)(%rsp), %rcx
1287 sub \$5, %rcx
1288 mov %rcx, (+$loop_idx_offset)(%rsp)
1289 jge main_loop_a3b
1290
1291 #
1292
1293 end_main_loop_a3b:
1294 # transform result out of Montgomery space
1295 # result = reduce(result)
1296 mov (+$pResult_offset)(%rsp), %rdx
1297 pxor %xmm4, %xmm4
1298 movdqu (+16*0)(%rdx), %xmm0
1299 movdqu (+16*1)(%rdx), %xmm1
1300 movdqu (+16*2)(%rdx), %xmm2
1301 movdqu (+16*3)(%rdx), %xmm3
1302 movdqa %xmm4, (+$tmp16_offset+16*4)(%rsp)
1303 movdqa %xmm4, (+$tmp16_offset+16*5)(%rsp)
1304 movdqa %xmm4, (+$tmp16_offset+16*6)(%rsp)
1305 movdqa %xmm4, (+$tmp16_offset+16*7)(%rsp)
1306 movdqa %xmm0, (+$tmp16_offset+16*0)(%rsp)
1307 movdqa %xmm1, (+$tmp16_offset+16*1)(%rsp)
1308 movdqa %xmm2, (+$tmp16_offset+16*2)(%rsp)
1309 movdqa %xmm3, (+$tmp16_offset+16*3)(%rsp)
1310 call mont_reduce
1311
1312 # If result > m, subract m
1313 # load result into r15:r8
1314 mov (+$pResult_offset)(%rsp), %rax
1315 mov (+8*0)(%rax), %r8
1316 mov (+8*1)(%rax), %r9
1317 mov (+8*2)(%rax), %r10
1318 mov (+8*3)(%rax), %r11
1319 mov (+8*4)(%rax), %r12
1320 mov (+8*5)(%rax), %r13
1321 mov (+8*6)(%rax), %r14
1322 mov (+8*7)(%rax), %r15
1323
1324 # subtract m
1325 mov (+$pData_offset)(%rsp), %rbx
1326 add \$$M, %rbx
1327
1328 sub (+8*0)(%rbx), %r8
1329 sbb (+8*1)(%rbx), %r9
1330 sbb (+8*2)(%rbx), %r10
1331 sbb (+8*3)(%rbx), %r11
1332 sbb (+8*4)(%rbx), %r12
1333 sbb (+8*5)(%rbx), %r13
1334 sbb (+8*6)(%rbx), %r14
1335 sbb (+8*7)(%rbx), %r15
1336
1337 # if Carry is clear, replace result with difference
1338 mov (+8*0)(%rax), %rsi
1339 mov (+8*1)(%rax), %rdi
1340 mov (+8*2)(%rax), %rcx
1341 mov (+8*3)(%rax), %rdx
1342 cmovnc %r8, %rsi
1343 cmovnc %r9, %rdi
1344 cmovnc %r10, %rcx
1345 cmovnc %r11, %rdx
1346 mov %rsi, (+8*0)(%rax)
1347 mov %rdi, (+8*1)(%rax)
1348 mov %rcx, (+8*2)(%rax)
1349 mov %rdx, (+8*3)(%rax)
1350
1351 mov (+8*4)(%rax), %rsi
1352 mov (+8*5)(%rax), %rdi
1353 mov (+8*6)(%rax), %rcx
1354 mov (+8*7)(%rax), %rdx
1355 cmovnc %r12, %rsi
1356 cmovnc %r13, %rdi
1357 cmovnc %r14, %rcx
1358 cmovnc %r15, %rdx
1359 mov %rsi, (+8*4)(%rax)
1360 mov %rdi, (+8*5)(%rax)
1361 mov %rcx, (+8*6)(%rax)
1362 mov %rdx, (+8*7)(%rax)
1363
1364 mov (+$rsp_offset)(%rsp), %rsi
1365 mov 0(%rsi),%r15
1366 mov 8(%rsi),%r14
1367 mov 16(%rsi),%r13
1368 mov 24(%rsi),%r12
1369 mov 32(%rsi),%rbx
1370 mov 40(%rsi),%rbp
1371 lea 48(%rsi),%rsp
1372 .Lepilogue:
1373 ret
1374 .size mod_exp_512, . - mod_exp_512
1375 ___
1376
1377 if ($win64) {
1378 # EXCEPTION_DISPOSITION handler (EXCEPTION_RECORD *rec,ULONG64 frame,
1379 # CONTEXT *context,DISPATCHER_CONTEXT *disp)
1380 my $rec="%rcx";
1381 my $frame="%rdx";
1382 my $context="%r8";
1383 my $disp="%r9";
1384
1385 $code.=<<___;
1386 .extern __imp_RtlVirtualUnwind
1387 .type mod_exp_512_se_handler,\@abi-omnipotent
1388 .align 16
1389 mod_exp_512_se_handler:
1390 push %rsi
1391 push %rdi
1392 push %rbx
1393 push %rbp
1394 push %r12
1395 push %r13
1396 push %r14
1397 push %r15
1398 pushfq
1399 sub \$64,%rsp
1400
1401 mov 120($context),%rax # pull context->Rax
1402 mov 248($context),%rbx # pull context->Rip
1403
1404 lea .Lbody(%rip),%r10
1405 cmp %r10,%rbx # context->Rip<prologue label
1406 jb .Lin_prologue
1407
1408 mov 152($context),%rax # pull context->Rsp
1409
1410 lea .Lepilogue(%rip),%r10
1411 cmp %r10,%rbx # context->Rip>=epilogue label
1412 jae .Lin_prologue
1413
1414 mov $rsp_offset(%rax),%rax # pull saved Rsp
1415
1416 mov 32(%rax),%rbx
1417 mov 40(%rax),%rbp
1418 mov 24(%rax),%r12
1419 mov 16(%rax),%r13
1420 mov 8(%rax),%r14
1421 mov 0(%rax),%r15
1422 lea 48(%rax),%rax
1423 mov %rbx,144($context) # restore context->Rbx
1424 mov %rbp,160($context) # restore context->Rbp
1425 mov %r12,216($context) # restore context->R12
1426 mov %r13,224($context) # restore context->R13
1427 mov %r14,232($context) # restore context->R14
1428 mov %r15,240($context) # restore context->R15
1429
1430 .Lin_prologue:
1431 mov 8(%rax),%rdi
1432 mov 16(%rax),%rsi
1433 mov %rax,152($context) # restore context->Rsp
1434 mov %rsi,168($context) # restore context->Rsi
1435 mov %rdi,176($context) # restore context->Rdi
1436
1437 mov 40($disp),%rdi # disp->ContextRecord
1438 mov $context,%rsi # context
1439 mov \$154,%ecx # sizeof(CONTEXT)
1440 .long 0xa548f3fc # cld; rep movsq
1441
1442 mov $disp,%rsi
1443 xor %rcx,%rcx # arg1, UNW_FLAG_NHANDLER
1444 mov 8(%rsi),%rdx # arg2, disp->ImageBase
1445 mov 0(%rsi),%r8 # arg3, disp->ControlPc
1446 mov 16(%rsi),%r9 # arg4, disp->FunctionEntry
1447 mov 40(%rsi),%r10 # disp->ContextRecord
1448 lea 56(%rsi),%r11 # &disp->HandlerData
1449 lea 24(%rsi),%r12 # &disp->EstablisherFrame
1450 mov %r10,32(%rsp) # arg5
1451 mov %r11,40(%rsp) # arg6
1452 mov %r12,48(%rsp) # arg7
1453 mov %rcx,56(%rsp) # arg8, (NULL)
1454 call *__imp_RtlVirtualUnwind(%rip)
1455
1456 mov \$1,%eax # ExceptionContinueSearch
1457 add \$64,%rsp
1458 popfq
1459 pop %r15
1460 pop %r14
1461 pop %r13
1462 pop %r12
1463 pop %rbp
1464 pop %rbx
1465 pop %rdi
1466 pop %rsi
1467 ret
1468 .size mod_exp_512_se_handler,.-mod_exp_512_se_handler
1469
1470 .section .pdata
1471 .align 4
1472 .rva .LSEH_begin_mod_exp_512
1473 .rva .LSEH_end_mod_exp_512
1474 .rva .LSEH_info_mod_exp_512
1475
1476 .section .xdata
1477 .align 8
1478 .LSEH_info_mod_exp_512:
1479 .byte 9,0,0,0
1480 .rva mod_exp_512_se_handler
1481 ___
1482 }
1483
1484 sub reg_part {
1485 my ($reg,$conv)=@_;
1486 if ($reg =~ /%r[0-9]+/) { $reg .= $conv; }
1487 elsif ($conv eq "b") { $reg =~ s/%[er]([^x]+)x?/%$1l/; }
1488 elsif ($conv eq "w") { $reg =~ s/%[er](.+)/%$1/; }
1489 elsif ($conv eq "d") { $reg =~ s/%[er](.+)/%e$1/; }
1490 return $reg;
1491 }
1492
1493 $code =~ s/(%[a-z0-9]+)#([bwd])/reg_part($1,$2)/gem;
1494 $code =~ s/\`([^\`]*)\`/eval $1/gem;
1495 $code =~ s/(\(\+[^)]+\))/eval $1/gem;
1496 print $code;
1497 close STDOUT;
OLDNEW
« no previous file with comments | « openssl/crypto/bn/asm/modexp512-x86_64.S ('k') | openssl/crypto/bn/asm/pa-risc2.s » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698