Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(260)

Side by Side Diff: openssl/crypto/modes/asm/ghash-x86_64.S

Issue 2072073002: Delete bundled copy of OpenSSL and replace with README. (Closed) Base URL: https://chromium.googlesource.com/chromium/deps/openssl@master
Patch Set: Delete bundled copy of OpenSSL and replace with README. Created 4 years, 6 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « openssl/crypto/modes/asm/ghash-x86-mac.S ('k') | openssl/crypto/modes/asm/ghash-x86_64.pl » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
(Empty)
1 .text
2
3 .globl gcm_gmult_4bit
4 .type gcm_gmult_4bit,@function
5 .align 16
6 gcm_gmult_4bit:
7 pushq %rbx
8 pushq %rbp
9 pushq %r12
10 .Lgmult_prologue:
11
12 movzbq 15(%rdi),%r8
13 leaq .Lrem_4bit(%rip),%r11
14 xorq %rax,%rax
15 xorq %rbx,%rbx
16 movb %r8b,%al
17 movb %r8b,%bl
18 shlb $4,%al
19 movq $14,%rcx
20 movq 8(%rsi,%rax,1),%r8
21 movq (%rsi,%rax,1),%r9
22 andb $240,%bl
23 movq %r8,%rdx
24 jmp .Loop1
25
26 .align 16
27 .Loop1:
28 shrq $4,%r8
29 andq $15,%rdx
30 movq %r9,%r10
31 movb (%rdi,%rcx,1),%al
32 shrq $4,%r9
33 xorq 8(%rsi,%rbx,1),%r8
34 shlq $60,%r10
35 xorq (%rsi,%rbx,1),%r9
36 movb %al,%bl
37 xorq (%r11,%rdx,8),%r9
38 movq %r8,%rdx
39 shlb $4,%al
40 xorq %r10,%r8
41 decq %rcx
42 js .Lbreak1
43
44 shrq $4,%r8
45 andq $15,%rdx
46 movq %r9,%r10
47 shrq $4,%r9
48 xorq 8(%rsi,%rax,1),%r8
49 shlq $60,%r10
50 xorq (%rsi,%rax,1),%r9
51 andb $240,%bl
52 xorq (%r11,%rdx,8),%r9
53 movq %r8,%rdx
54 xorq %r10,%r8
55 jmp .Loop1
56
57 .align 16
58 .Lbreak1:
59 shrq $4,%r8
60 andq $15,%rdx
61 movq %r9,%r10
62 shrq $4,%r9
63 xorq 8(%rsi,%rax,1),%r8
64 shlq $60,%r10
65 xorq (%rsi,%rax,1),%r9
66 andb $240,%bl
67 xorq (%r11,%rdx,8),%r9
68 movq %r8,%rdx
69 xorq %r10,%r8
70
71 shrq $4,%r8
72 andq $15,%rdx
73 movq %r9,%r10
74 shrq $4,%r9
75 xorq 8(%rsi,%rbx,1),%r8
76 shlq $60,%r10
77 xorq (%rsi,%rbx,1),%r9
78 xorq %r10,%r8
79 xorq (%r11,%rdx,8),%r9
80
81 bswapq %r8
82 bswapq %r9
83 movq %r8,8(%rdi)
84 movq %r9,(%rdi)
85
86 movq 16(%rsp),%rbx
87 leaq 24(%rsp),%rsp
88 .Lgmult_epilogue:
89 .byte 0xf3,0xc3
90 .size gcm_gmult_4bit,.-gcm_gmult_4bit
91 .globl gcm_ghash_4bit
92 .type gcm_ghash_4bit,@function
93 .align 16
94 gcm_ghash_4bit:
95 pushq %rbx
96 pushq %rbp
97 pushq %r12
98 pushq %r13
99 pushq %r14
100 pushq %r15
101 subq $280,%rsp
102 .Lghash_prologue:
103 movq %rdx,%r14
104 movq %rcx,%r15
105 subq $-128,%rsi
106 leaq 16+128(%rsp),%rbp
107 xorl %edx,%edx
108 movq 0+0-128(%rsi),%r8
109 movq 0+8-128(%rsi),%rax
110 movb %al,%dl
111 shrq $4,%rax
112 movq %r8,%r10
113 shrq $4,%r8
114 movq 16+0-128(%rsi),%r9
115 shlb $4,%dl
116 movq 16+8-128(%rsi),%rbx
117 shlq $60,%r10
118 movb %dl,0(%rsp)
119 orq %r10,%rax
120 movb %bl,%dl
121 shrq $4,%rbx
122 movq %r9,%r10
123 shrq $4,%r9
124 movq %r8,0(%rbp)
125 movq 32+0-128(%rsi),%r8
126 shlb $4,%dl
127 movq %rax,0-128(%rbp)
128 movq 32+8-128(%rsi),%rax
129 shlq $60,%r10
130 movb %dl,1(%rsp)
131 orq %r10,%rbx
132 movb %al,%dl
133 shrq $4,%rax
134 movq %r8,%r10
135 shrq $4,%r8
136 movq %r9,8(%rbp)
137 movq 48+0-128(%rsi),%r9
138 shlb $4,%dl
139 movq %rbx,8-128(%rbp)
140 movq 48+8-128(%rsi),%rbx
141 shlq $60,%r10
142 movb %dl,2(%rsp)
143 orq %r10,%rax
144 movb %bl,%dl
145 shrq $4,%rbx
146 movq %r9,%r10
147 shrq $4,%r9
148 movq %r8,16(%rbp)
149 movq 64+0-128(%rsi),%r8
150 shlb $4,%dl
151 movq %rax,16-128(%rbp)
152 movq 64+8-128(%rsi),%rax
153 shlq $60,%r10
154 movb %dl,3(%rsp)
155 orq %r10,%rbx
156 movb %al,%dl
157 shrq $4,%rax
158 movq %r8,%r10
159 shrq $4,%r8
160 movq %r9,24(%rbp)
161 movq 80+0-128(%rsi),%r9
162 shlb $4,%dl
163 movq %rbx,24-128(%rbp)
164 movq 80+8-128(%rsi),%rbx
165 shlq $60,%r10
166 movb %dl,4(%rsp)
167 orq %r10,%rax
168 movb %bl,%dl
169 shrq $4,%rbx
170 movq %r9,%r10
171 shrq $4,%r9
172 movq %r8,32(%rbp)
173 movq 96+0-128(%rsi),%r8
174 shlb $4,%dl
175 movq %rax,32-128(%rbp)
176 movq 96+8-128(%rsi),%rax
177 shlq $60,%r10
178 movb %dl,5(%rsp)
179 orq %r10,%rbx
180 movb %al,%dl
181 shrq $4,%rax
182 movq %r8,%r10
183 shrq $4,%r8
184 movq %r9,40(%rbp)
185 movq 112+0-128(%rsi),%r9
186 shlb $4,%dl
187 movq %rbx,40-128(%rbp)
188 movq 112+8-128(%rsi),%rbx
189 shlq $60,%r10
190 movb %dl,6(%rsp)
191 orq %r10,%rax
192 movb %bl,%dl
193 shrq $4,%rbx
194 movq %r9,%r10
195 shrq $4,%r9
196 movq %r8,48(%rbp)
197 movq 128+0-128(%rsi),%r8
198 shlb $4,%dl
199 movq %rax,48-128(%rbp)
200 movq 128+8-128(%rsi),%rax
201 shlq $60,%r10
202 movb %dl,7(%rsp)
203 orq %r10,%rbx
204 movb %al,%dl
205 shrq $4,%rax
206 movq %r8,%r10
207 shrq $4,%r8
208 movq %r9,56(%rbp)
209 movq 144+0-128(%rsi),%r9
210 shlb $4,%dl
211 movq %rbx,56-128(%rbp)
212 movq 144+8-128(%rsi),%rbx
213 shlq $60,%r10
214 movb %dl,8(%rsp)
215 orq %r10,%rax
216 movb %bl,%dl
217 shrq $4,%rbx
218 movq %r9,%r10
219 shrq $4,%r9
220 movq %r8,64(%rbp)
221 movq 160+0-128(%rsi),%r8
222 shlb $4,%dl
223 movq %rax,64-128(%rbp)
224 movq 160+8-128(%rsi),%rax
225 shlq $60,%r10
226 movb %dl,9(%rsp)
227 orq %r10,%rbx
228 movb %al,%dl
229 shrq $4,%rax
230 movq %r8,%r10
231 shrq $4,%r8
232 movq %r9,72(%rbp)
233 movq 176+0-128(%rsi),%r9
234 shlb $4,%dl
235 movq %rbx,72-128(%rbp)
236 movq 176+8-128(%rsi),%rbx
237 shlq $60,%r10
238 movb %dl,10(%rsp)
239 orq %r10,%rax
240 movb %bl,%dl
241 shrq $4,%rbx
242 movq %r9,%r10
243 shrq $4,%r9
244 movq %r8,80(%rbp)
245 movq 192+0-128(%rsi),%r8
246 shlb $4,%dl
247 movq %rax,80-128(%rbp)
248 movq 192+8-128(%rsi),%rax
249 shlq $60,%r10
250 movb %dl,11(%rsp)
251 orq %r10,%rbx
252 movb %al,%dl
253 shrq $4,%rax
254 movq %r8,%r10
255 shrq $4,%r8
256 movq %r9,88(%rbp)
257 movq 208+0-128(%rsi),%r9
258 shlb $4,%dl
259 movq %rbx,88-128(%rbp)
260 movq 208+8-128(%rsi),%rbx
261 shlq $60,%r10
262 movb %dl,12(%rsp)
263 orq %r10,%rax
264 movb %bl,%dl
265 shrq $4,%rbx
266 movq %r9,%r10
267 shrq $4,%r9
268 movq %r8,96(%rbp)
269 movq 224+0-128(%rsi),%r8
270 shlb $4,%dl
271 movq %rax,96-128(%rbp)
272 movq 224+8-128(%rsi),%rax
273 shlq $60,%r10
274 movb %dl,13(%rsp)
275 orq %r10,%rbx
276 movb %al,%dl
277 shrq $4,%rax
278 movq %r8,%r10
279 shrq $4,%r8
280 movq %r9,104(%rbp)
281 movq 240+0-128(%rsi),%r9
282 shlb $4,%dl
283 movq %rbx,104-128(%rbp)
284 movq 240+8-128(%rsi),%rbx
285 shlq $60,%r10
286 movb %dl,14(%rsp)
287 orq %r10,%rax
288 movb %bl,%dl
289 shrq $4,%rbx
290 movq %r9,%r10
291 shrq $4,%r9
292 movq %r8,112(%rbp)
293 shlb $4,%dl
294 movq %rax,112-128(%rbp)
295 shlq $60,%r10
296 movb %dl,15(%rsp)
297 orq %r10,%rbx
298 movq %r9,120(%rbp)
299 movq %rbx,120-128(%rbp)
300 addq $-128,%rsi
301 movq 8(%rdi),%r8
302 movq 0(%rdi),%r9
303 addq %r14,%r15
304 leaq .Lrem_8bit(%rip),%r11
305 jmp .Louter_loop
306 .align 16
307 .Louter_loop:
308 xorq (%r14),%r9
309 movq 8(%r14),%rdx
310 leaq 16(%r14),%r14
311 xorq %r8,%rdx
312 movq %r9,(%rdi)
313 movq %rdx,8(%rdi)
314 shrq $32,%rdx
315 xorq %rax,%rax
316 roll $8,%edx
317 movb %dl,%al
318 movzbl %dl,%ebx
319 shlb $4,%al
320 shrl $4,%ebx
321 roll $8,%edx
322 movq 8(%rsi,%rax,1),%r8
323 movq (%rsi,%rax,1),%r9
324 movb %dl,%al
325 movzbl %dl,%ecx
326 shlb $4,%al
327 movzbq (%rsp,%rbx,1),%r12
328 shrl $4,%ecx
329 xorq %r8,%r12
330 movq %r9,%r10
331 shrq $8,%r8
332 movzbq %r12b,%r12
333 shrq $8,%r9
334 xorq -128(%rbp,%rbx,8),%r8
335 shlq $56,%r10
336 xorq (%rbp,%rbx,8),%r9
337 roll $8,%edx
338 xorq 8(%rsi,%rax,1),%r8
339 xorq (%rsi,%rax,1),%r9
340 movb %dl,%al
341 xorq %r10,%r8
342 movzwq (%r11,%r12,2),%r12
343 movzbl %dl,%ebx
344 shlb $4,%al
345 movzbq (%rsp,%rcx,1),%r13
346 shrl $4,%ebx
347 shlq $48,%r12
348 xorq %r8,%r13
349 movq %r9,%r10
350 xorq %r12,%r9
351 shrq $8,%r8
352 movzbq %r13b,%r13
353 shrq $8,%r9
354 xorq -128(%rbp,%rcx,8),%r8
355 shlq $56,%r10
356 xorq (%rbp,%rcx,8),%r9
357 roll $8,%edx
358 xorq 8(%rsi,%rax,1),%r8
359 xorq (%rsi,%rax,1),%r9
360 movb %dl,%al
361 xorq %r10,%r8
362 movzwq (%r11,%r13,2),%r13
363 movzbl %dl,%ecx
364 shlb $4,%al
365 movzbq (%rsp,%rbx,1),%r12
366 shrl $4,%ecx
367 shlq $48,%r13
368 xorq %r8,%r12
369 movq %r9,%r10
370 xorq %r13,%r9
371 shrq $8,%r8
372 movzbq %r12b,%r12
373 movl 8(%rdi),%edx
374 shrq $8,%r9
375 xorq -128(%rbp,%rbx,8),%r8
376 shlq $56,%r10
377 xorq (%rbp,%rbx,8),%r9
378 roll $8,%edx
379 xorq 8(%rsi,%rax,1),%r8
380 xorq (%rsi,%rax,1),%r9
381 movb %dl,%al
382 xorq %r10,%r8
383 movzwq (%r11,%r12,2),%r12
384 movzbl %dl,%ebx
385 shlb $4,%al
386 movzbq (%rsp,%rcx,1),%r13
387 shrl $4,%ebx
388 shlq $48,%r12
389 xorq %r8,%r13
390 movq %r9,%r10
391 xorq %r12,%r9
392 shrq $8,%r8
393 movzbq %r13b,%r13
394 shrq $8,%r9
395 xorq -128(%rbp,%rcx,8),%r8
396 shlq $56,%r10
397 xorq (%rbp,%rcx,8),%r9
398 roll $8,%edx
399 xorq 8(%rsi,%rax,1),%r8
400 xorq (%rsi,%rax,1),%r9
401 movb %dl,%al
402 xorq %r10,%r8
403 movzwq (%r11,%r13,2),%r13
404 movzbl %dl,%ecx
405 shlb $4,%al
406 movzbq (%rsp,%rbx,1),%r12
407 shrl $4,%ecx
408 shlq $48,%r13
409 xorq %r8,%r12
410 movq %r9,%r10
411 xorq %r13,%r9
412 shrq $8,%r8
413 movzbq %r12b,%r12
414 shrq $8,%r9
415 xorq -128(%rbp,%rbx,8),%r8
416 shlq $56,%r10
417 xorq (%rbp,%rbx,8),%r9
418 roll $8,%edx
419 xorq 8(%rsi,%rax,1),%r8
420 xorq (%rsi,%rax,1),%r9
421 movb %dl,%al
422 xorq %r10,%r8
423 movzwq (%r11,%r12,2),%r12
424 movzbl %dl,%ebx
425 shlb $4,%al
426 movzbq (%rsp,%rcx,1),%r13
427 shrl $4,%ebx
428 shlq $48,%r12
429 xorq %r8,%r13
430 movq %r9,%r10
431 xorq %r12,%r9
432 shrq $8,%r8
433 movzbq %r13b,%r13
434 shrq $8,%r9
435 xorq -128(%rbp,%rcx,8),%r8
436 shlq $56,%r10
437 xorq (%rbp,%rcx,8),%r9
438 roll $8,%edx
439 xorq 8(%rsi,%rax,1),%r8
440 xorq (%rsi,%rax,1),%r9
441 movb %dl,%al
442 xorq %r10,%r8
443 movzwq (%r11,%r13,2),%r13
444 movzbl %dl,%ecx
445 shlb $4,%al
446 movzbq (%rsp,%rbx,1),%r12
447 shrl $4,%ecx
448 shlq $48,%r13
449 xorq %r8,%r12
450 movq %r9,%r10
451 xorq %r13,%r9
452 shrq $8,%r8
453 movzbq %r12b,%r12
454 movl 4(%rdi),%edx
455 shrq $8,%r9
456 xorq -128(%rbp,%rbx,8),%r8
457 shlq $56,%r10
458 xorq (%rbp,%rbx,8),%r9
459 roll $8,%edx
460 xorq 8(%rsi,%rax,1),%r8
461 xorq (%rsi,%rax,1),%r9
462 movb %dl,%al
463 xorq %r10,%r8
464 movzwq (%r11,%r12,2),%r12
465 movzbl %dl,%ebx
466 shlb $4,%al
467 movzbq (%rsp,%rcx,1),%r13
468 shrl $4,%ebx
469 shlq $48,%r12
470 xorq %r8,%r13
471 movq %r9,%r10
472 xorq %r12,%r9
473 shrq $8,%r8
474 movzbq %r13b,%r13
475 shrq $8,%r9
476 xorq -128(%rbp,%rcx,8),%r8
477 shlq $56,%r10
478 xorq (%rbp,%rcx,8),%r9
479 roll $8,%edx
480 xorq 8(%rsi,%rax,1),%r8
481 xorq (%rsi,%rax,1),%r9
482 movb %dl,%al
483 xorq %r10,%r8
484 movzwq (%r11,%r13,2),%r13
485 movzbl %dl,%ecx
486 shlb $4,%al
487 movzbq (%rsp,%rbx,1),%r12
488 shrl $4,%ecx
489 shlq $48,%r13
490 xorq %r8,%r12
491 movq %r9,%r10
492 xorq %r13,%r9
493 shrq $8,%r8
494 movzbq %r12b,%r12
495 shrq $8,%r9
496 xorq -128(%rbp,%rbx,8),%r8
497 shlq $56,%r10
498 xorq (%rbp,%rbx,8),%r9
499 roll $8,%edx
500 xorq 8(%rsi,%rax,1),%r8
501 xorq (%rsi,%rax,1),%r9
502 movb %dl,%al
503 xorq %r10,%r8
504 movzwq (%r11,%r12,2),%r12
505 movzbl %dl,%ebx
506 shlb $4,%al
507 movzbq (%rsp,%rcx,1),%r13
508 shrl $4,%ebx
509 shlq $48,%r12
510 xorq %r8,%r13
511 movq %r9,%r10
512 xorq %r12,%r9
513 shrq $8,%r8
514 movzbq %r13b,%r13
515 shrq $8,%r9
516 xorq -128(%rbp,%rcx,8),%r8
517 shlq $56,%r10
518 xorq (%rbp,%rcx,8),%r9
519 roll $8,%edx
520 xorq 8(%rsi,%rax,1),%r8
521 xorq (%rsi,%rax,1),%r9
522 movb %dl,%al
523 xorq %r10,%r8
524 movzwq (%r11,%r13,2),%r13
525 movzbl %dl,%ecx
526 shlb $4,%al
527 movzbq (%rsp,%rbx,1),%r12
528 shrl $4,%ecx
529 shlq $48,%r13
530 xorq %r8,%r12
531 movq %r9,%r10
532 xorq %r13,%r9
533 shrq $8,%r8
534 movzbq %r12b,%r12
535 movl 0(%rdi),%edx
536 shrq $8,%r9
537 xorq -128(%rbp,%rbx,8),%r8
538 shlq $56,%r10
539 xorq (%rbp,%rbx,8),%r9
540 roll $8,%edx
541 xorq 8(%rsi,%rax,1),%r8
542 xorq (%rsi,%rax,1),%r9
543 movb %dl,%al
544 xorq %r10,%r8
545 movzwq (%r11,%r12,2),%r12
546 movzbl %dl,%ebx
547 shlb $4,%al
548 movzbq (%rsp,%rcx,1),%r13
549 shrl $4,%ebx
550 shlq $48,%r12
551 xorq %r8,%r13
552 movq %r9,%r10
553 xorq %r12,%r9
554 shrq $8,%r8
555 movzbq %r13b,%r13
556 shrq $8,%r9
557 xorq -128(%rbp,%rcx,8),%r8
558 shlq $56,%r10
559 xorq (%rbp,%rcx,8),%r9
560 roll $8,%edx
561 xorq 8(%rsi,%rax,1),%r8
562 xorq (%rsi,%rax,1),%r9
563 movb %dl,%al
564 xorq %r10,%r8
565 movzwq (%r11,%r13,2),%r13
566 movzbl %dl,%ecx
567 shlb $4,%al
568 movzbq (%rsp,%rbx,1),%r12
569 shrl $4,%ecx
570 shlq $48,%r13
571 xorq %r8,%r12
572 movq %r9,%r10
573 xorq %r13,%r9
574 shrq $8,%r8
575 movzbq %r12b,%r12
576 shrq $8,%r9
577 xorq -128(%rbp,%rbx,8),%r8
578 shlq $56,%r10
579 xorq (%rbp,%rbx,8),%r9
580 roll $8,%edx
581 xorq 8(%rsi,%rax,1),%r8
582 xorq (%rsi,%rax,1),%r9
583 movb %dl,%al
584 xorq %r10,%r8
585 movzwq (%r11,%r12,2),%r12
586 movzbl %dl,%ebx
587 shlb $4,%al
588 movzbq (%rsp,%rcx,1),%r13
589 shrl $4,%ebx
590 shlq $48,%r12
591 xorq %r8,%r13
592 movq %r9,%r10
593 xorq %r12,%r9
594 shrq $8,%r8
595 movzbq %r13b,%r13
596 shrq $8,%r9
597 xorq -128(%rbp,%rcx,8),%r8
598 shlq $56,%r10
599 xorq (%rbp,%rcx,8),%r9
600 roll $8,%edx
601 xorq 8(%rsi,%rax,1),%r8
602 xorq (%rsi,%rax,1),%r9
603 movb %dl,%al
604 xorq %r10,%r8
605 movzwq (%r11,%r13,2),%r13
606 movzbl %dl,%ecx
607 shlb $4,%al
608 movzbq (%rsp,%rbx,1),%r12
609 andl $240,%ecx
610 shlq $48,%r13
611 xorq %r8,%r12
612 movq %r9,%r10
613 xorq %r13,%r9
614 shrq $8,%r8
615 movzbq %r12b,%r12
616 movl -4(%rdi),%edx
617 shrq $8,%r9
618 xorq -128(%rbp,%rbx,8),%r8
619 shlq $56,%r10
620 xorq (%rbp,%rbx,8),%r9
621 movzwq (%r11,%r12,2),%r12
622 xorq 8(%rsi,%rax,1),%r8
623 xorq (%rsi,%rax,1),%r9
624 shlq $48,%r12
625 xorq %r10,%r8
626 xorq %r12,%r9
627 movzbq %r8b,%r13
628 shrq $4,%r8
629 movq %r9,%r10
630 shlb $4,%r13b
631 shrq $4,%r9
632 xorq 8(%rsi,%rcx,1),%r8
633 movzwq (%r11,%r13,2),%r13
634 shlq $60,%r10
635 xorq (%rsi,%rcx,1),%r9
636 xorq %r10,%r8
637 shlq $48,%r13
638 bswapq %r8
639 xorq %r13,%r9
640 bswapq %r9
641 cmpq %r15,%r14
642 jb .Louter_loop
643 movq %r8,8(%rdi)
644 movq %r9,(%rdi)
645
646 leaq 280(%rsp),%rsi
647 movq 0(%rsi),%r15
648 movq 8(%rsi),%r14
649 movq 16(%rsi),%r13
650 movq 24(%rsi),%r12
651 movq 32(%rsi),%rbp
652 movq 40(%rsi),%rbx
653 leaq 48(%rsi),%rsp
654 .Lghash_epilogue:
655 .byte 0xf3,0xc3
656 .size gcm_ghash_4bit,.-gcm_ghash_4bit
657 .globl gcm_init_clmul
658 .type gcm_init_clmul,@function
659 .align 16
660 gcm_init_clmul:
661 movdqu (%rsi),%xmm2
662 pshufd $78,%xmm2,%xmm2
663
664
665 pshufd $255,%xmm2,%xmm4
666 movdqa %xmm2,%xmm3
667 psllq $1,%xmm2
668 pxor %xmm5,%xmm5
669 psrlq $63,%xmm3
670 pcmpgtd %xmm4,%xmm5
671 pslldq $8,%xmm3
672 por %xmm3,%xmm2
673
674
675 pand .L0x1c2_polynomial(%rip),%xmm5
676 pxor %xmm5,%xmm2
677
678
679 movdqa %xmm2,%xmm0
680 movdqa %xmm0,%xmm1
681 pshufd $78,%xmm0,%xmm3
682 pshufd $78,%xmm2,%xmm4
683 pxor %xmm0,%xmm3
684 pxor %xmm2,%xmm4
685 .byte 102,15,58,68,194,0
686 .byte 102,15,58,68,202,17
687 .byte 102,15,58,68,220,0
688 pxor %xmm0,%xmm3
689 pxor %xmm1,%xmm3
690
691 movdqa %xmm3,%xmm4
692 psrldq $8,%xmm3
693 pslldq $8,%xmm4
694 pxor %xmm3,%xmm1
695 pxor %xmm4,%xmm0
696
697 movdqa %xmm0,%xmm3
698 psllq $1,%xmm0
699 pxor %xmm3,%xmm0
700 psllq $5,%xmm0
701 pxor %xmm3,%xmm0
702 psllq $57,%xmm0
703 movdqa %xmm0,%xmm4
704 pslldq $8,%xmm0
705 psrldq $8,%xmm4
706 pxor %xmm3,%xmm0
707 pxor %xmm4,%xmm1
708
709
710 movdqa %xmm0,%xmm4
711 psrlq $5,%xmm0
712 pxor %xmm4,%xmm0
713 psrlq $1,%xmm0
714 pxor %xmm4,%xmm0
715 pxor %xmm1,%xmm4
716 psrlq $1,%xmm0
717 pxor %xmm4,%xmm0
718 movdqu %xmm2,(%rdi)
719 movdqu %xmm0,16(%rdi)
720 .byte 0xf3,0xc3
721 .size gcm_init_clmul,.-gcm_init_clmul
722 .globl gcm_gmult_clmul
723 .type gcm_gmult_clmul,@function
724 .align 16
725 gcm_gmult_clmul:
726 movdqu (%rdi),%xmm0
727 movdqa .Lbswap_mask(%rip),%xmm5
728 movdqu (%rsi),%xmm2
729 .byte 102,15,56,0,197
730 movdqa %xmm0,%xmm1
731 pshufd $78,%xmm0,%xmm3
732 pshufd $78,%xmm2,%xmm4
733 pxor %xmm0,%xmm3
734 pxor %xmm2,%xmm4
735 .byte 102,15,58,68,194,0
736 .byte 102,15,58,68,202,17
737 .byte 102,15,58,68,220,0
738 pxor %xmm0,%xmm3
739 pxor %xmm1,%xmm3
740
741 movdqa %xmm3,%xmm4
742 psrldq $8,%xmm3
743 pslldq $8,%xmm4
744 pxor %xmm3,%xmm1
745 pxor %xmm4,%xmm0
746
747 movdqa %xmm0,%xmm3
748 psllq $1,%xmm0
749 pxor %xmm3,%xmm0
750 psllq $5,%xmm0
751 pxor %xmm3,%xmm0
752 psllq $57,%xmm0
753 movdqa %xmm0,%xmm4
754 pslldq $8,%xmm0
755 psrldq $8,%xmm4
756 pxor %xmm3,%xmm0
757 pxor %xmm4,%xmm1
758
759
760 movdqa %xmm0,%xmm4
761 psrlq $5,%xmm0
762 pxor %xmm4,%xmm0
763 psrlq $1,%xmm0
764 pxor %xmm4,%xmm0
765 pxor %xmm1,%xmm4
766 psrlq $1,%xmm0
767 pxor %xmm4,%xmm0
768 .byte 102,15,56,0,197
769 movdqu %xmm0,(%rdi)
770 .byte 0xf3,0xc3
771 .size gcm_gmult_clmul,.-gcm_gmult_clmul
772 .globl gcm_ghash_clmul
773 .type gcm_ghash_clmul,@function
774 .align 16
775 gcm_ghash_clmul:
776 movdqa .Lbswap_mask(%rip),%xmm5
777
778 movdqu (%rdi),%xmm0
779 movdqu (%rsi),%xmm2
780 .byte 102,15,56,0,197
781
782 subq $16,%rcx
783 jz .Lodd_tail
784
785 movdqu 16(%rsi),%xmm8
786
787
788
789
790
791 movdqu (%rdx),%xmm3
792 movdqu 16(%rdx),%xmm6
793 .byte 102,15,56,0,221
794 .byte 102,15,56,0,245
795 pxor %xmm3,%xmm0
796 movdqa %xmm6,%xmm7
797 pshufd $78,%xmm6,%xmm3
798 pshufd $78,%xmm2,%xmm4
799 pxor %xmm6,%xmm3
800 pxor %xmm2,%xmm4
801 .byte 102,15,58,68,242,0
802 .byte 102,15,58,68,250,17
803 .byte 102,15,58,68,220,0
804 pxor %xmm6,%xmm3
805 pxor %xmm7,%xmm3
806
807 movdqa %xmm3,%xmm4
808 psrldq $8,%xmm3
809 pslldq $8,%xmm4
810 pxor %xmm3,%xmm7
811 pxor %xmm4,%xmm6
812 movdqa %xmm0,%xmm1
813 pshufd $78,%xmm0,%xmm3
814 pshufd $78,%xmm8,%xmm4
815 pxor %xmm0,%xmm3
816 pxor %xmm8,%xmm4
817
818 leaq 32(%rdx),%rdx
819 subq $32,%rcx
820 jbe .Leven_tail
821
822 .Lmod_loop:
823 .byte 102,65,15,58,68,192,0
824 .byte 102,65,15,58,68,200,17
825 .byte 102,15,58,68,220,0
826 pxor %xmm0,%xmm3
827 pxor %xmm1,%xmm3
828
829 movdqa %xmm3,%xmm4
830 psrldq $8,%xmm3
831 pslldq $8,%xmm4
832 pxor %xmm3,%xmm1
833 pxor %xmm4,%xmm0
834 movdqu (%rdx),%xmm3
835 pxor %xmm6,%xmm0
836 pxor %xmm7,%xmm1
837
838 movdqu 16(%rdx),%xmm6
839 .byte 102,15,56,0,221
840 .byte 102,15,56,0,245
841
842 movdqa %xmm6,%xmm7
843 pshufd $78,%xmm6,%xmm9
844 pshufd $78,%xmm2,%xmm10
845 pxor %xmm6,%xmm9
846 pxor %xmm2,%xmm10
847 pxor %xmm3,%xmm1
848
849 movdqa %xmm0,%xmm3
850 psllq $1,%xmm0
851 pxor %xmm3,%xmm0
852 psllq $5,%xmm0
853 pxor %xmm3,%xmm0
854 .byte 102,15,58,68,242,0
855 psllq $57,%xmm0
856 movdqa %xmm0,%xmm4
857 pslldq $8,%xmm0
858 psrldq $8,%xmm4
859 pxor %xmm3,%xmm0
860 pxor %xmm4,%xmm1
861
862 .byte 102,15,58,68,250,17
863 movdqa %xmm0,%xmm4
864 psrlq $5,%xmm0
865 pxor %xmm4,%xmm0
866 psrlq $1,%xmm0
867 pxor %xmm4,%xmm0
868 pxor %xmm1,%xmm4
869 psrlq $1,%xmm0
870 pxor %xmm4,%xmm0
871
872 .byte 102,69,15,58,68,202,0
873 movdqa %xmm0,%xmm1
874 pshufd $78,%xmm0,%xmm3
875 pshufd $78,%xmm8,%xmm4
876 pxor %xmm0,%xmm3
877 pxor %xmm8,%xmm4
878
879 pxor %xmm6,%xmm9
880 pxor %xmm7,%xmm9
881 movdqa %xmm9,%xmm10
882 psrldq $8,%xmm9
883 pslldq $8,%xmm10
884 pxor %xmm9,%xmm7
885 pxor %xmm10,%xmm6
886
887 leaq 32(%rdx),%rdx
888 subq $32,%rcx
889 ja .Lmod_loop
890
891 .Leven_tail:
892 .byte 102,65,15,58,68,192,0
893 .byte 102,65,15,58,68,200,17
894 .byte 102,15,58,68,220,0
895 pxor %xmm0,%xmm3
896 pxor %xmm1,%xmm3
897
898 movdqa %xmm3,%xmm4
899 psrldq $8,%xmm3
900 pslldq $8,%xmm4
901 pxor %xmm3,%xmm1
902 pxor %xmm4,%xmm0
903 pxor %xmm6,%xmm0
904 pxor %xmm7,%xmm1
905
906 movdqa %xmm0,%xmm3
907 psllq $1,%xmm0
908 pxor %xmm3,%xmm0
909 psllq $5,%xmm0
910 pxor %xmm3,%xmm0
911 psllq $57,%xmm0
912 movdqa %xmm0,%xmm4
913 pslldq $8,%xmm0
914 psrldq $8,%xmm4
915 pxor %xmm3,%xmm0
916 pxor %xmm4,%xmm1
917
918
919 movdqa %xmm0,%xmm4
920 psrlq $5,%xmm0
921 pxor %xmm4,%xmm0
922 psrlq $1,%xmm0
923 pxor %xmm4,%xmm0
924 pxor %xmm1,%xmm4
925 psrlq $1,%xmm0
926 pxor %xmm4,%xmm0
927 testq %rcx,%rcx
928 jnz .Ldone
929
930 .Lodd_tail:
931 movdqu (%rdx),%xmm3
932 .byte 102,15,56,0,221
933 pxor %xmm3,%xmm0
934 movdqa %xmm0,%xmm1
935 pshufd $78,%xmm0,%xmm3
936 pshufd $78,%xmm2,%xmm4
937 pxor %xmm0,%xmm3
938 pxor %xmm2,%xmm4
939 .byte 102,15,58,68,194,0
940 .byte 102,15,58,68,202,17
941 .byte 102,15,58,68,220,0
942 pxor %xmm0,%xmm3
943 pxor %xmm1,%xmm3
944
945 movdqa %xmm3,%xmm4
946 psrldq $8,%xmm3
947 pslldq $8,%xmm4
948 pxor %xmm3,%xmm1
949 pxor %xmm4,%xmm0
950
951 movdqa %xmm0,%xmm3
952 psllq $1,%xmm0
953 pxor %xmm3,%xmm0
954 psllq $5,%xmm0
955 pxor %xmm3,%xmm0
956 psllq $57,%xmm0
957 movdqa %xmm0,%xmm4
958 pslldq $8,%xmm0
959 psrldq $8,%xmm4
960 pxor %xmm3,%xmm0
961 pxor %xmm4,%xmm1
962
963
964 movdqa %xmm0,%xmm4
965 psrlq $5,%xmm0
966 pxor %xmm4,%xmm0
967 psrlq $1,%xmm0
968 pxor %xmm4,%xmm0
969 pxor %xmm1,%xmm4
970 psrlq $1,%xmm0
971 pxor %xmm4,%xmm0
972 .Ldone:
973 .byte 102,15,56,0,197
974 movdqu %xmm0,(%rdi)
975 .byte 0xf3,0xc3
976 .LSEH_end_gcm_ghash_clmul:
977 .size gcm_ghash_clmul,.-gcm_ghash_clmul
978 .align 64
979 .Lbswap_mask:
980 .byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
981 .L0x1c2_polynomial:
982 .byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
983 .align 64
984 .type .Lrem_4bit,@object
985 .Lrem_4bit:
986 .long 0,0,0,471859200,0,943718400,0,610271232
987 .long 0,1887436800,0,1822425088,0,1220542464,0,1423966208
988 .long 0,3774873600,0,4246732800,0,3644850176,0,3311403008
989 .long 0,2441084928,0,2376073216,0,2847932416,0,3051356160
990 .type .Lrem_8bit,@object
991 .Lrem_8bit:
992 .value 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E
993 .value 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E
994 .value 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E
995 .value 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E
996 .value 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E
997 .value 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E
998 .value 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E
999 .value 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E
1000 .value 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE
1001 .value 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE
1002 .value 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE
1003 .value 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE
1004 .value 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E
1005 .value 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E
1006 .value 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE
1007 .value 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE
1008 .value 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E
1009 .value 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E
1010 .value 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E
1011 .value 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E
1012 .value 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E
1013 .value 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E
1014 .value 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E
1015 .value 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E
1016 .value 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE
1017 .value 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE
1018 .value 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE
1019 .value 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE
1020 .value 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E
1021 .value 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E
1022 .value 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE
1023 .value 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE
1024
1025 .byte 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84 ,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,10 8,46,111,114,103,62,0
1026 .align 64
OLDNEW
« no previous file with comments | « openssl/crypto/modes/asm/ghash-x86-mac.S ('k') | openssl/crypto/modes/asm/ghash-x86_64.pl » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698