Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(565)

Side by Side Diff: third_party/boringssl/linux-x86_64/crypto/modes/ghash-x86_64.S

Issue 2354623003: Pull boringssl generated source from boringssl_gen (Closed)
Patch Set: . Created 4 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
(Empty)
1 #if defined(__x86_64__)
2 .text
3 .extern OPENSSL_ia32cap_P
4 .hidden OPENSSL_ia32cap_P
5
6 .globl gcm_gmult_4bit
7 .hidden gcm_gmult_4bit
8 .type gcm_gmult_4bit,@function
9 .align 16
10 gcm_gmult_4bit:
11 pushq %rbx
12 pushq %rbp
13 pushq %r12
14 .Lgmult_prologue:
15
16 movzbq 15(%rdi),%r8
17 leaq .Lrem_4bit(%rip),%r11
18 xorq %rax,%rax
19 xorq %rbx,%rbx
20 movb %r8b,%al
21 movb %r8b,%bl
22 shlb $4,%al
23 movq $14,%rcx
24 movq 8(%rsi,%rax,1),%r8
25 movq (%rsi,%rax,1),%r9
26 andb $0xf0,%bl
27 movq %r8,%rdx
28 jmp .Loop1
29
30 .align 16
31 .Loop1:
32 shrq $4,%r8
33 andq $0xf,%rdx
34 movq %r9,%r10
35 movb (%rdi,%rcx,1),%al
36 shrq $4,%r9
37 xorq 8(%rsi,%rbx,1),%r8
38 shlq $60,%r10
39 xorq (%rsi,%rbx,1),%r9
40 movb %al,%bl
41 xorq (%r11,%rdx,8),%r9
42 movq %r8,%rdx
43 shlb $4,%al
44 xorq %r10,%r8
45 decq %rcx
46 js .Lbreak1
47
48 shrq $4,%r8
49 andq $0xf,%rdx
50 movq %r9,%r10
51 shrq $4,%r9
52 xorq 8(%rsi,%rax,1),%r8
53 shlq $60,%r10
54 xorq (%rsi,%rax,1),%r9
55 andb $0xf0,%bl
56 xorq (%r11,%rdx,8),%r9
57 movq %r8,%rdx
58 xorq %r10,%r8
59 jmp .Loop1
60
61 .align 16
62 .Lbreak1:
63 shrq $4,%r8
64 andq $0xf,%rdx
65 movq %r9,%r10
66 shrq $4,%r9
67 xorq 8(%rsi,%rax,1),%r8
68 shlq $60,%r10
69 xorq (%rsi,%rax,1),%r9
70 andb $0xf0,%bl
71 xorq (%r11,%rdx,8),%r9
72 movq %r8,%rdx
73 xorq %r10,%r8
74
75 shrq $4,%r8
76 andq $0xf,%rdx
77 movq %r9,%r10
78 shrq $4,%r9
79 xorq 8(%rsi,%rbx,1),%r8
80 shlq $60,%r10
81 xorq (%rsi,%rbx,1),%r9
82 xorq %r10,%r8
83 xorq (%r11,%rdx,8),%r9
84
85 bswapq %r8
86 bswapq %r9
87 movq %r8,8(%rdi)
88 movq %r9,(%rdi)
89
90 movq 16(%rsp),%rbx
91 leaq 24(%rsp),%rsp
92 .Lgmult_epilogue:
93 .byte 0xf3,0xc3
94 .size gcm_gmult_4bit,.-gcm_gmult_4bit
95 .globl gcm_ghash_4bit
96 .hidden gcm_ghash_4bit
97 .type gcm_ghash_4bit,@function
98 .align 16
99 gcm_ghash_4bit:
100 pushq %rbx
101 pushq %rbp
102 pushq %r12
103 pushq %r13
104 pushq %r14
105 pushq %r15
106 subq $280,%rsp
107 .Lghash_prologue:
108 movq %rdx,%r14
109 movq %rcx,%r15
110 subq $-128,%rsi
111 leaq 16+128(%rsp),%rbp
112 xorl %edx,%edx
113 movq 0+0-128(%rsi),%r8
114 movq 0+8-128(%rsi),%rax
115 movb %al,%dl
116 shrq $4,%rax
117 movq %r8,%r10
118 shrq $4,%r8
119 movq 16+0-128(%rsi),%r9
120 shlb $4,%dl
121 movq 16+8-128(%rsi),%rbx
122 shlq $60,%r10
123 movb %dl,0(%rsp)
124 orq %r10,%rax
125 movb %bl,%dl
126 shrq $4,%rbx
127 movq %r9,%r10
128 shrq $4,%r9
129 movq %r8,0(%rbp)
130 movq 32+0-128(%rsi),%r8
131 shlb $4,%dl
132 movq %rax,0-128(%rbp)
133 movq 32+8-128(%rsi),%rax
134 shlq $60,%r10
135 movb %dl,1(%rsp)
136 orq %r10,%rbx
137 movb %al,%dl
138 shrq $4,%rax
139 movq %r8,%r10
140 shrq $4,%r8
141 movq %r9,8(%rbp)
142 movq 48+0-128(%rsi),%r9
143 shlb $4,%dl
144 movq %rbx,8-128(%rbp)
145 movq 48+8-128(%rsi),%rbx
146 shlq $60,%r10
147 movb %dl,2(%rsp)
148 orq %r10,%rax
149 movb %bl,%dl
150 shrq $4,%rbx
151 movq %r9,%r10
152 shrq $4,%r9
153 movq %r8,16(%rbp)
154 movq 64+0-128(%rsi),%r8
155 shlb $4,%dl
156 movq %rax,16-128(%rbp)
157 movq 64+8-128(%rsi),%rax
158 shlq $60,%r10
159 movb %dl,3(%rsp)
160 orq %r10,%rbx
161 movb %al,%dl
162 shrq $4,%rax
163 movq %r8,%r10
164 shrq $4,%r8
165 movq %r9,24(%rbp)
166 movq 80+0-128(%rsi),%r9
167 shlb $4,%dl
168 movq %rbx,24-128(%rbp)
169 movq 80+8-128(%rsi),%rbx
170 shlq $60,%r10
171 movb %dl,4(%rsp)
172 orq %r10,%rax
173 movb %bl,%dl
174 shrq $4,%rbx
175 movq %r9,%r10
176 shrq $4,%r9
177 movq %r8,32(%rbp)
178 movq 96+0-128(%rsi),%r8
179 shlb $4,%dl
180 movq %rax,32-128(%rbp)
181 movq 96+8-128(%rsi),%rax
182 shlq $60,%r10
183 movb %dl,5(%rsp)
184 orq %r10,%rbx
185 movb %al,%dl
186 shrq $4,%rax
187 movq %r8,%r10
188 shrq $4,%r8
189 movq %r9,40(%rbp)
190 movq 112+0-128(%rsi),%r9
191 shlb $4,%dl
192 movq %rbx,40-128(%rbp)
193 movq 112+8-128(%rsi),%rbx
194 shlq $60,%r10
195 movb %dl,6(%rsp)
196 orq %r10,%rax
197 movb %bl,%dl
198 shrq $4,%rbx
199 movq %r9,%r10
200 shrq $4,%r9
201 movq %r8,48(%rbp)
202 movq 128+0-128(%rsi),%r8
203 shlb $4,%dl
204 movq %rax,48-128(%rbp)
205 movq 128+8-128(%rsi),%rax
206 shlq $60,%r10
207 movb %dl,7(%rsp)
208 orq %r10,%rbx
209 movb %al,%dl
210 shrq $4,%rax
211 movq %r8,%r10
212 shrq $4,%r8
213 movq %r9,56(%rbp)
214 movq 144+0-128(%rsi),%r9
215 shlb $4,%dl
216 movq %rbx,56-128(%rbp)
217 movq 144+8-128(%rsi),%rbx
218 shlq $60,%r10
219 movb %dl,8(%rsp)
220 orq %r10,%rax
221 movb %bl,%dl
222 shrq $4,%rbx
223 movq %r9,%r10
224 shrq $4,%r9
225 movq %r8,64(%rbp)
226 movq 160+0-128(%rsi),%r8
227 shlb $4,%dl
228 movq %rax,64-128(%rbp)
229 movq 160+8-128(%rsi),%rax
230 shlq $60,%r10
231 movb %dl,9(%rsp)
232 orq %r10,%rbx
233 movb %al,%dl
234 shrq $4,%rax
235 movq %r8,%r10
236 shrq $4,%r8
237 movq %r9,72(%rbp)
238 movq 176+0-128(%rsi),%r9
239 shlb $4,%dl
240 movq %rbx,72-128(%rbp)
241 movq 176+8-128(%rsi),%rbx
242 shlq $60,%r10
243 movb %dl,10(%rsp)
244 orq %r10,%rax
245 movb %bl,%dl
246 shrq $4,%rbx
247 movq %r9,%r10
248 shrq $4,%r9
249 movq %r8,80(%rbp)
250 movq 192+0-128(%rsi),%r8
251 shlb $4,%dl
252 movq %rax,80-128(%rbp)
253 movq 192+8-128(%rsi),%rax
254 shlq $60,%r10
255 movb %dl,11(%rsp)
256 orq %r10,%rbx
257 movb %al,%dl
258 shrq $4,%rax
259 movq %r8,%r10
260 shrq $4,%r8
261 movq %r9,88(%rbp)
262 movq 208+0-128(%rsi),%r9
263 shlb $4,%dl
264 movq %rbx,88-128(%rbp)
265 movq 208+8-128(%rsi),%rbx
266 shlq $60,%r10
267 movb %dl,12(%rsp)
268 orq %r10,%rax
269 movb %bl,%dl
270 shrq $4,%rbx
271 movq %r9,%r10
272 shrq $4,%r9
273 movq %r8,96(%rbp)
274 movq 224+0-128(%rsi),%r8
275 shlb $4,%dl
276 movq %rax,96-128(%rbp)
277 movq 224+8-128(%rsi),%rax
278 shlq $60,%r10
279 movb %dl,13(%rsp)
280 orq %r10,%rbx
281 movb %al,%dl
282 shrq $4,%rax
283 movq %r8,%r10
284 shrq $4,%r8
285 movq %r9,104(%rbp)
286 movq 240+0-128(%rsi),%r9
287 shlb $4,%dl
288 movq %rbx,104-128(%rbp)
289 movq 240+8-128(%rsi),%rbx
290 shlq $60,%r10
291 movb %dl,14(%rsp)
292 orq %r10,%rax
293 movb %bl,%dl
294 shrq $4,%rbx
295 movq %r9,%r10
296 shrq $4,%r9
297 movq %r8,112(%rbp)
298 shlb $4,%dl
299 movq %rax,112-128(%rbp)
300 shlq $60,%r10
301 movb %dl,15(%rsp)
302 orq %r10,%rbx
303 movq %r9,120(%rbp)
304 movq %rbx,120-128(%rbp)
305 addq $-128,%rsi
306 movq 8(%rdi),%r8
307 movq 0(%rdi),%r9
308 addq %r14,%r15
309 leaq .Lrem_8bit(%rip),%r11
310 jmp .Louter_loop
311 .align 16
312 .Louter_loop:
313 xorq (%r14),%r9
314 movq 8(%r14),%rdx
315 leaq 16(%r14),%r14
316 xorq %r8,%rdx
317 movq %r9,(%rdi)
318 movq %rdx,8(%rdi)
319 shrq $32,%rdx
320 xorq %rax,%rax
321 roll $8,%edx
322 movb %dl,%al
323 movzbl %dl,%ebx
324 shlb $4,%al
325 shrl $4,%ebx
326 roll $8,%edx
327 movq 8(%rsi,%rax,1),%r8
328 movq (%rsi,%rax,1),%r9
329 movb %dl,%al
330 movzbl %dl,%ecx
331 shlb $4,%al
332 movzbq (%rsp,%rbx,1),%r12
333 shrl $4,%ecx
334 xorq %r8,%r12
335 movq %r9,%r10
336 shrq $8,%r8
337 movzbq %r12b,%r12
338 shrq $8,%r9
339 xorq -128(%rbp,%rbx,8),%r8
340 shlq $56,%r10
341 xorq (%rbp,%rbx,8),%r9
342 roll $8,%edx
343 xorq 8(%rsi,%rax,1),%r8
344 xorq (%rsi,%rax,1),%r9
345 movb %dl,%al
346 xorq %r10,%r8
347 movzwq (%r11,%r12,2),%r12
348 movzbl %dl,%ebx
349 shlb $4,%al
350 movzbq (%rsp,%rcx,1),%r13
351 shrl $4,%ebx
352 shlq $48,%r12
353 xorq %r8,%r13
354 movq %r9,%r10
355 xorq %r12,%r9
356 shrq $8,%r8
357 movzbq %r13b,%r13
358 shrq $8,%r9
359 xorq -128(%rbp,%rcx,8),%r8
360 shlq $56,%r10
361 xorq (%rbp,%rcx,8),%r9
362 roll $8,%edx
363 xorq 8(%rsi,%rax,1),%r8
364 xorq (%rsi,%rax,1),%r9
365 movb %dl,%al
366 xorq %r10,%r8
367 movzwq (%r11,%r13,2),%r13
368 movzbl %dl,%ecx
369 shlb $4,%al
370 movzbq (%rsp,%rbx,1),%r12
371 shrl $4,%ecx
372 shlq $48,%r13
373 xorq %r8,%r12
374 movq %r9,%r10
375 xorq %r13,%r9
376 shrq $8,%r8
377 movzbq %r12b,%r12
378 movl 8(%rdi),%edx
379 shrq $8,%r9
380 xorq -128(%rbp,%rbx,8),%r8
381 shlq $56,%r10
382 xorq (%rbp,%rbx,8),%r9
383 roll $8,%edx
384 xorq 8(%rsi,%rax,1),%r8
385 xorq (%rsi,%rax,1),%r9
386 movb %dl,%al
387 xorq %r10,%r8
388 movzwq (%r11,%r12,2),%r12
389 movzbl %dl,%ebx
390 shlb $4,%al
391 movzbq (%rsp,%rcx,1),%r13
392 shrl $4,%ebx
393 shlq $48,%r12
394 xorq %r8,%r13
395 movq %r9,%r10
396 xorq %r12,%r9
397 shrq $8,%r8
398 movzbq %r13b,%r13
399 shrq $8,%r9
400 xorq -128(%rbp,%rcx,8),%r8
401 shlq $56,%r10
402 xorq (%rbp,%rcx,8),%r9
403 roll $8,%edx
404 xorq 8(%rsi,%rax,1),%r8
405 xorq (%rsi,%rax,1),%r9
406 movb %dl,%al
407 xorq %r10,%r8
408 movzwq (%r11,%r13,2),%r13
409 movzbl %dl,%ecx
410 shlb $4,%al
411 movzbq (%rsp,%rbx,1),%r12
412 shrl $4,%ecx
413 shlq $48,%r13
414 xorq %r8,%r12
415 movq %r9,%r10
416 xorq %r13,%r9
417 shrq $8,%r8
418 movzbq %r12b,%r12
419 shrq $8,%r9
420 xorq -128(%rbp,%rbx,8),%r8
421 shlq $56,%r10
422 xorq (%rbp,%rbx,8),%r9
423 roll $8,%edx
424 xorq 8(%rsi,%rax,1),%r8
425 xorq (%rsi,%rax,1),%r9
426 movb %dl,%al
427 xorq %r10,%r8
428 movzwq (%r11,%r12,2),%r12
429 movzbl %dl,%ebx
430 shlb $4,%al
431 movzbq (%rsp,%rcx,1),%r13
432 shrl $4,%ebx
433 shlq $48,%r12
434 xorq %r8,%r13
435 movq %r9,%r10
436 xorq %r12,%r9
437 shrq $8,%r8
438 movzbq %r13b,%r13
439 shrq $8,%r9
440 xorq -128(%rbp,%rcx,8),%r8
441 shlq $56,%r10
442 xorq (%rbp,%rcx,8),%r9
443 roll $8,%edx
444 xorq 8(%rsi,%rax,1),%r8
445 xorq (%rsi,%rax,1),%r9
446 movb %dl,%al
447 xorq %r10,%r8
448 movzwq (%r11,%r13,2),%r13
449 movzbl %dl,%ecx
450 shlb $4,%al
451 movzbq (%rsp,%rbx,1),%r12
452 shrl $4,%ecx
453 shlq $48,%r13
454 xorq %r8,%r12
455 movq %r9,%r10
456 xorq %r13,%r9
457 shrq $8,%r8
458 movzbq %r12b,%r12
459 movl 4(%rdi),%edx
460 shrq $8,%r9
461 xorq -128(%rbp,%rbx,8),%r8
462 shlq $56,%r10
463 xorq (%rbp,%rbx,8),%r9
464 roll $8,%edx
465 xorq 8(%rsi,%rax,1),%r8
466 xorq (%rsi,%rax,1),%r9
467 movb %dl,%al
468 xorq %r10,%r8
469 movzwq (%r11,%r12,2),%r12
470 movzbl %dl,%ebx
471 shlb $4,%al
472 movzbq (%rsp,%rcx,1),%r13
473 shrl $4,%ebx
474 shlq $48,%r12
475 xorq %r8,%r13
476 movq %r9,%r10
477 xorq %r12,%r9
478 shrq $8,%r8
479 movzbq %r13b,%r13
480 shrq $8,%r9
481 xorq -128(%rbp,%rcx,8),%r8
482 shlq $56,%r10
483 xorq (%rbp,%rcx,8),%r9
484 roll $8,%edx
485 xorq 8(%rsi,%rax,1),%r8
486 xorq (%rsi,%rax,1),%r9
487 movb %dl,%al
488 xorq %r10,%r8
489 movzwq (%r11,%r13,2),%r13
490 movzbl %dl,%ecx
491 shlb $4,%al
492 movzbq (%rsp,%rbx,1),%r12
493 shrl $4,%ecx
494 shlq $48,%r13
495 xorq %r8,%r12
496 movq %r9,%r10
497 xorq %r13,%r9
498 shrq $8,%r8
499 movzbq %r12b,%r12
500 shrq $8,%r9
501 xorq -128(%rbp,%rbx,8),%r8
502 shlq $56,%r10
503 xorq (%rbp,%rbx,8),%r9
504 roll $8,%edx
505 xorq 8(%rsi,%rax,1),%r8
506 xorq (%rsi,%rax,1),%r9
507 movb %dl,%al
508 xorq %r10,%r8
509 movzwq (%r11,%r12,2),%r12
510 movzbl %dl,%ebx
511 shlb $4,%al
512 movzbq (%rsp,%rcx,1),%r13
513 shrl $4,%ebx
514 shlq $48,%r12
515 xorq %r8,%r13
516 movq %r9,%r10
517 xorq %r12,%r9
518 shrq $8,%r8
519 movzbq %r13b,%r13
520 shrq $8,%r9
521 xorq -128(%rbp,%rcx,8),%r8
522 shlq $56,%r10
523 xorq (%rbp,%rcx,8),%r9
524 roll $8,%edx
525 xorq 8(%rsi,%rax,1),%r8
526 xorq (%rsi,%rax,1),%r9
527 movb %dl,%al
528 xorq %r10,%r8
529 movzwq (%r11,%r13,2),%r13
530 movzbl %dl,%ecx
531 shlb $4,%al
532 movzbq (%rsp,%rbx,1),%r12
533 shrl $4,%ecx
534 shlq $48,%r13
535 xorq %r8,%r12
536 movq %r9,%r10
537 xorq %r13,%r9
538 shrq $8,%r8
539 movzbq %r12b,%r12
540 movl 0(%rdi),%edx
541 shrq $8,%r9
542 xorq -128(%rbp,%rbx,8),%r8
543 shlq $56,%r10
544 xorq (%rbp,%rbx,8),%r9
545 roll $8,%edx
546 xorq 8(%rsi,%rax,1),%r8
547 xorq (%rsi,%rax,1),%r9
548 movb %dl,%al
549 xorq %r10,%r8
550 movzwq (%r11,%r12,2),%r12
551 movzbl %dl,%ebx
552 shlb $4,%al
553 movzbq (%rsp,%rcx,1),%r13
554 shrl $4,%ebx
555 shlq $48,%r12
556 xorq %r8,%r13
557 movq %r9,%r10
558 xorq %r12,%r9
559 shrq $8,%r8
560 movzbq %r13b,%r13
561 shrq $8,%r9
562 xorq -128(%rbp,%rcx,8),%r8
563 shlq $56,%r10
564 xorq (%rbp,%rcx,8),%r9
565 roll $8,%edx
566 xorq 8(%rsi,%rax,1),%r8
567 xorq (%rsi,%rax,1),%r9
568 movb %dl,%al
569 xorq %r10,%r8
570 movzwq (%r11,%r13,2),%r13
571 movzbl %dl,%ecx
572 shlb $4,%al
573 movzbq (%rsp,%rbx,1),%r12
574 shrl $4,%ecx
575 shlq $48,%r13
576 xorq %r8,%r12
577 movq %r9,%r10
578 xorq %r13,%r9
579 shrq $8,%r8
580 movzbq %r12b,%r12
581 shrq $8,%r9
582 xorq -128(%rbp,%rbx,8),%r8
583 shlq $56,%r10
584 xorq (%rbp,%rbx,8),%r9
585 roll $8,%edx
586 xorq 8(%rsi,%rax,1),%r8
587 xorq (%rsi,%rax,1),%r9
588 movb %dl,%al
589 xorq %r10,%r8
590 movzwq (%r11,%r12,2),%r12
591 movzbl %dl,%ebx
592 shlb $4,%al
593 movzbq (%rsp,%rcx,1),%r13
594 shrl $4,%ebx
595 shlq $48,%r12
596 xorq %r8,%r13
597 movq %r9,%r10
598 xorq %r12,%r9
599 shrq $8,%r8
600 movzbq %r13b,%r13
601 shrq $8,%r9
602 xorq -128(%rbp,%rcx,8),%r8
603 shlq $56,%r10
604 xorq (%rbp,%rcx,8),%r9
605 roll $8,%edx
606 xorq 8(%rsi,%rax,1),%r8
607 xorq (%rsi,%rax,1),%r9
608 movb %dl,%al
609 xorq %r10,%r8
610 movzwq (%r11,%r13,2),%r13
611 movzbl %dl,%ecx
612 shlb $4,%al
613 movzbq (%rsp,%rbx,1),%r12
614 andl $240,%ecx
615 shlq $48,%r13
616 xorq %r8,%r12
617 movq %r9,%r10
618 xorq %r13,%r9
619 shrq $8,%r8
620 movzbq %r12b,%r12
621 movl -4(%rdi),%edx
622 shrq $8,%r9
623 xorq -128(%rbp,%rbx,8),%r8
624 shlq $56,%r10
625 xorq (%rbp,%rbx,8),%r9
626 movzwq (%r11,%r12,2),%r12
627 xorq 8(%rsi,%rax,1),%r8
628 xorq (%rsi,%rax,1),%r9
629 shlq $48,%r12
630 xorq %r10,%r8
631 xorq %r12,%r9
632 movzbq %r8b,%r13
633 shrq $4,%r8
634 movq %r9,%r10
635 shlb $4,%r13b
636 shrq $4,%r9
637 xorq 8(%rsi,%rcx,1),%r8
638 movzwq (%r11,%r13,2),%r13
639 shlq $60,%r10
640 xorq (%rsi,%rcx,1),%r9
641 xorq %r10,%r8
642 shlq $48,%r13
643 bswapq %r8
644 xorq %r13,%r9
645 bswapq %r9
646 cmpq %r15,%r14
647 jb .Louter_loop
648 movq %r8,8(%rdi)
649 movq %r9,(%rdi)
650
651 leaq 280(%rsp),%rsi
652 movq 0(%rsi),%r15
653 movq 8(%rsi),%r14
654 movq 16(%rsi),%r13
655 movq 24(%rsi),%r12
656 movq 32(%rsi),%rbp
657 movq 40(%rsi),%rbx
658 leaq 48(%rsi),%rsp
659 .Lghash_epilogue:
660 .byte 0xf3,0xc3
661 .size gcm_ghash_4bit,.-gcm_ghash_4bit
662 .globl gcm_init_clmul
663 .hidden gcm_init_clmul
664 .type gcm_init_clmul,@function
665 .align 16
666 gcm_init_clmul:
667 .L_init_clmul:
668 movdqu (%rsi),%xmm2
669 pshufd $78,%xmm2,%xmm2
670
671
672 pshufd $255,%xmm2,%xmm4
673 movdqa %xmm2,%xmm3
674 psllq $1,%xmm2
675 pxor %xmm5,%xmm5
676 psrlq $63,%xmm3
677 pcmpgtd %xmm4,%xmm5
678 pslldq $8,%xmm3
679 por %xmm3,%xmm2
680
681
682 pand .L0x1c2_polynomial(%rip),%xmm5
683 pxor %xmm5,%xmm2
684
685
686 pshufd $78,%xmm2,%xmm6
687 movdqa %xmm2,%xmm0
688 pxor %xmm2,%xmm6
689 movdqa %xmm0,%xmm1
690 pshufd $78,%xmm0,%xmm3
691 pxor %xmm0,%xmm3
692 .byte 102,15,58,68,194,0
693 .byte 102,15,58,68,202,17
694 .byte 102,15,58,68,222,0
695 pxor %xmm0,%xmm3
696 pxor %xmm1,%xmm3
697
698 movdqa %xmm3,%xmm4
699 psrldq $8,%xmm3
700 pslldq $8,%xmm4
701 pxor %xmm3,%xmm1
702 pxor %xmm4,%xmm0
703
704 movdqa %xmm0,%xmm4
705 movdqa %xmm0,%xmm3
706 psllq $5,%xmm0
707 pxor %xmm0,%xmm3
708 psllq $1,%xmm0
709 pxor %xmm3,%xmm0
710 psllq $57,%xmm0
711 movdqa %xmm0,%xmm3
712 pslldq $8,%xmm0
713 psrldq $8,%xmm3
714 pxor %xmm4,%xmm0
715 pxor %xmm3,%xmm1
716
717
718 movdqa %xmm0,%xmm4
719 psrlq $1,%xmm0
720 pxor %xmm4,%xmm1
721 pxor %xmm0,%xmm4
722 psrlq $5,%xmm0
723 pxor %xmm4,%xmm0
724 psrlq $1,%xmm0
725 pxor %xmm1,%xmm0
726 pshufd $78,%xmm2,%xmm3
727 pshufd $78,%xmm0,%xmm4
728 pxor %xmm2,%xmm3
729 movdqu %xmm2,0(%rdi)
730 pxor %xmm0,%xmm4
731 movdqu %xmm0,16(%rdi)
732 .byte 102,15,58,15,227,8
733 movdqu %xmm4,32(%rdi)
734 movdqa %xmm0,%xmm1
735 pshufd $78,%xmm0,%xmm3
736 pxor %xmm0,%xmm3
737 .byte 102,15,58,68,194,0
738 .byte 102,15,58,68,202,17
739 .byte 102,15,58,68,222,0
740 pxor %xmm0,%xmm3
741 pxor %xmm1,%xmm3
742
743 movdqa %xmm3,%xmm4
744 psrldq $8,%xmm3
745 pslldq $8,%xmm4
746 pxor %xmm3,%xmm1
747 pxor %xmm4,%xmm0
748
749 movdqa %xmm0,%xmm4
750 movdqa %xmm0,%xmm3
751 psllq $5,%xmm0
752 pxor %xmm0,%xmm3
753 psllq $1,%xmm0
754 pxor %xmm3,%xmm0
755 psllq $57,%xmm0
756 movdqa %xmm0,%xmm3
757 pslldq $8,%xmm0
758 psrldq $8,%xmm3
759 pxor %xmm4,%xmm0
760 pxor %xmm3,%xmm1
761
762
763 movdqa %xmm0,%xmm4
764 psrlq $1,%xmm0
765 pxor %xmm4,%xmm1
766 pxor %xmm0,%xmm4
767 psrlq $5,%xmm0
768 pxor %xmm4,%xmm0
769 psrlq $1,%xmm0
770 pxor %xmm1,%xmm0
771 movdqa %xmm0,%xmm5
772 movdqa %xmm0,%xmm1
773 pshufd $78,%xmm0,%xmm3
774 pxor %xmm0,%xmm3
775 .byte 102,15,58,68,194,0
776 .byte 102,15,58,68,202,17
777 .byte 102,15,58,68,222,0
778 pxor %xmm0,%xmm3
779 pxor %xmm1,%xmm3
780
781 movdqa %xmm3,%xmm4
782 psrldq $8,%xmm3
783 pslldq $8,%xmm4
784 pxor %xmm3,%xmm1
785 pxor %xmm4,%xmm0
786
787 movdqa %xmm0,%xmm4
788 movdqa %xmm0,%xmm3
789 psllq $5,%xmm0
790 pxor %xmm0,%xmm3
791 psllq $1,%xmm0
792 pxor %xmm3,%xmm0
793 psllq $57,%xmm0
794 movdqa %xmm0,%xmm3
795 pslldq $8,%xmm0
796 psrldq $8,%xmm3
797 pxor %xmm4,%xmm0
798 pxor %xmm3,%xmm1
799
800
801 movdqa %xmm0,%xmm4
802 psrlq $1,%xmm0
803 pxor %xmm4,%xmm1
804 pxor %xmm0,%xmm4
805 psrlq $5,%xmm0
806 pxor %xmm4,%xmm0
807 psrlq $1,%xmm0
808 pxor %xmm1,%xmm0
809 pshufd $78,%xmm5,%xmm3
810 pshufd $78,%xmm0,%xmm4
811 pxor %xmm5,%xmm3
812 movdqu %xmm5,48(%rdi)
813 pxor %xmm0,%xmm4
814 movdqu %xmm0,64(%rdi)
815 .byte 102,15,58,15,227,8
816 movdqu %xmm4,80(%rdi)
817 .byte 0xf3,0xc3
818 .size gcm_init_clmul,.-gcm_init_clmul
819 .globl gcm_gmult_clmul
820 .hidden gcm_gmult_clmul
821 .type gcm_gmult_clmul,@function
822 .align 16
823 gcm_gmult_clmul:
824 .L_gmult_clmul:
825 movdqu (%rdi),%xmm0
826 movdqa .Lbswap_mask(%rip),%xmm5
827 movdqu (%rsi),%xmm2
828 movdqu 32(%rsi),%xmm4
829 .byte 102,15,56,0,197
830 movdqa %xmm0,%xmm1
831 pshufd $78,%xmm0,%xmm3
832 pxor %xmm0,%xmm3
833 .byte 102,15,58,68,194,0
834 .byte 102,15,58,68,202,17
835 .byte 102,15,58,68,220,0
836 pxor %xmm0,%xmm3
837 pxor %xmm1,%xmm3
838
839 movdqa %xmm3,%xmm4
840 psrldq $8,%xmm3
841 pslldq $8,%xmm4
842 pxor %xmm3,%xmm1
843 pxor %xmm4,%xmm0
844
845 movdqa %xmm0,%xmm4
846 movdqa %xmm0,%xmm3
847 psllq $5,%xmm0
848 pxor %xmm0,%xmm3
849 psllq $1,%xmm0
850 pxor %xmm3,%xmm0
851 psllq $57,%xmm0
852 movdqa %xmm0,%xmm3
853 pslldq $8,%xmm0
854 psrldq $8,%xmm3
855 pxor %xmm4,%xmm0
856 pxor %xmm3,%xmm1
857
858
859 movdqa %xmm0,%xmm4
860 psrlq $1,%xmm0
861 pxor %xmm4,%xmm1
862 pxor %xmm0,%xmm4
863 psrlq $5,%xmm0
864 pxor %xmm4,%xmm0
865 psrlq $1,%xmm0
866 pxor %xmm1,%xmm0
867 .byte 102,15,56,0,197
868 movdqu %xmm0,(%rdi)
869 .byte 0xf3,0xc3
870 .size gcm_gmult_clmul,.-gcm_gmult_clmul
871 .globl gcm_ghash_clmul
872 .hidden gcm_ghash_clmul
873 .type gcm_ghash_clmul,@function
874 .align 32
875 gcm_ghash_clmul:
876 .L_ghash_clmul:
877 movdqa .Lbswap_mask(%rip),%xmm10
878
879 movdqu (%rdi),%xmm0
880 movdqu (%rsi),%xmm2
881 movdqu 32(%rsi),%xmm7
882 .byte 102,65,15,56,0,194
883
884 subq $0x10,%rcx
885 jz .Lodd_tail
886
887 movdqu 16(%rsi),%xmm6
888 movl OPENSSL_ia32cap_P+4(%rip),%eax
889 cmpq $0x30,%rcx
890 jb .Lskip4x
891
892 andl $71303168,%eax
893 cmpl $4194304,%eax
894 je .Lskip4x
895
896 subq $0x30,%rcx
897 movq $0xA040608020C0E000,%rax
898 movdqu 48(%rsi),%xmm14
899 movdqu 64(%rsi),%xmm15
900
901
902
903
904 movdqu 48(%rdx),%xmm3
905 movdqu 32(%rdx),%xmm11
906 .byte 102,65,15,56,0,218
907 .byte 102,69,15,56,0,218
908 movdqa %xmm3,%xmm5
909 pshufd $78,%xmm3,%xmm4
910 pxor %xmm3,%xmm4
911 .byte 102,15,58,68,218,0
912 .byte 102,15,58,68,234,17
913 .byte 102,15,58,68,231,0
914
915 movdqa %xmm11,%xmm13
916 pshufd $78,%xmm11,%xmm12
917 pxor %xmm11,%xmm12
918 .byte 102,68,15,58,68,222,0
919 .byte 102,68,15,58,68,238,17
920 .byte 102,68,15,58,68,231,16
921 xorps %xmm11,%xmm3
922 xorps %xmm13,%xmm5
923 movups 80(%rsi),%xmm7
924 xorps %xmm12,%xmm4
925
926 movdqu 16(%rdx),%xmm11
927 movdqu 0(%rdx),%xmm8
928 .byte 102,69,15,56,0,218
929 .byte 102,69,15,56,0,194
930 movdqa %xmm11,%xmm13
931 pshufd $78,%xmm11,%xmm12
932 pxor %xmm8,%xmm0
933 pxor %xmm11,%xmm12
934 .byte 102,69,15,58,68,222,0
935 movdqa %xmm0,%xmm1
936 pshufd $78,%xmm0,%xmm8
937 pxor %xmm0,%xmm8
938 .byte 102,69,15,58,68,238,17
939 .byte 102,68,15,58,68,231,0
940 xorps %xmm11,%xmm3
941 xorps %xmm13,%xmm5
942
943 leaq 64(%rdx),%rdx
944 subq $0x40,%rcx
945 jc .Ltail4x
946
947 jmp .Lmod4_loop
948 .align 32
949 .Lmod4_loop:
950 .byte 102,65,15,58,68,199,0
951 xorps %xmm12,%xmm4
952 movdqu 48(%rdx),%xmm11
953 .byte 102,69,15,56,0,218
954 .byte 102,65,15,58,68,207,17
955 xorps %xmm3,%xmm0
956 movdqu 32(%rdx),%xmm3
957 movdqa %xmm11,%xmm13
958 .byte 102,68,15,58,68,199,16
959 pshufd $78,%xmm11,%xmm12
960 xorps %xmm5,%xmm1
961 pxor %xmm11,%xmm12
962 .byte 102,65,15,56,0,218
963 movups 32(%rsi),%xmm7
964 xorps %xmm4,%xmm8
965 .byte 102,68,15,58,68,218,0
966 pshufd $78,%xmm3,%xmm4
967
968 pxor %xmm0,%xmm8
969 movdqa %xmm3,%xmm5
970 pxor %xmm1,%xmm8
971 pxor %xmm3,%xmm4
972 movdqa %xmm8,%xmm9
973 .byte 102,68,15,58,68,234,17
974 pslldq $8,%xmm8
975 psrldq $8,%xmm9
976 pxor %xmm8,%xmm0
977 movdqa .L7_mask(%rip),%xmm8
978 pxor %xmm9,%xmm1
979 .byte 102,76,15,110,200
980
981 pand %xmm0,%xmm8
982 .byte 102,69,15,56,0,200
983 pxor %xmm0,%xmm9
984 .byte 102,68,15,58,68,231,0
985 psllq $57,%xmm9
986 movdqa %xmm9,%xmm8
987 pslldq $8,%xmm9
988 .byte 102,15,58,68,222,0
989 psrldq $8,%xmm8
990 pxor %xmm9,%xmm0
991 pxor %xmm8,%xmm1
992 movdqu 0(%rdx),%xmm8
993
994 movdqa %xmm0,%xmm9
995 psrlq $1,%xmm0
996 .byte 102,15,58,68,238,17
997 xorps %xmm11,%xmm3
998 movdqu 16(%rdx),%xmm11
999 .byte 102,69,15,56,0,218
1000 .byte 102,15,58,68,231,16
1001 xorps %xmm13,%xmm5
1002 movups 80(%rsi),%xmm7
1003 .byte 102,69,15,56,0,194
1004 pxor %xmm9,%xmm1
1005 pxor %xmm0,%xmm9
1006 psrlq $5,%xmm0
1007
1008 movdqa %xmm11,%xmm13
1009 pxor %xmm12,%xmm4
1010 pshufd $78,%xmm11,%xmm12
1011 pxor %xmm9,%xmm0
1012 pxor %xmm8,%xmm1
1013 pxor %xmm11,%xmm12
1014 .byte 102,69,15,58,68,222,0
1015 psrlq $1,%xmm0
1016 pxor %xmm1,%xmm0
1017 movdqa %xmm0,%xmm1
1018 .byte 102,69,15,58,68,238,17
1019 xorps %xmm11,%xmm3
1020 pshufd $78,%xmm0,%xmm8
1021 pxor %xmm0,%xmm8
1022
1023 .byte 102,68,15,58,68,231,0
1024 xorps %xmm13,%xmm5
1025
1026 leaq 64(%rdx),%rdx
1027 subq $0x40,%rcx
1028 jnc .Lmod4_loop
1029
1030 .Ltail4x:
1031 .byte 102,65,15,58,68,199,0
1032 .byte 102,65,15,58,68,207,17
1033 .byte 102,68,15,58,68,199,16
1034 xorps %xmm12,%xmm4
1035 xorps %xmm3,%xmm0
1036 xorps %xmm5,%xmm1
1037 pxor %xmm0,%xmm1
1038 pxor %xmm4,%xmm8
1039
1040 pxor %xmm1,%xmm8
1041 pxor %xmm0,%xmm1
1042
1043 movdqa %xmm8,%xmm9
1044 psrldq $8,%xmm8
1045 pslldq $8,%xmm9
1046 pxor %xmm8,%xmm1
1047 pxor %xmm9,%xmm0
1048
1049 movdqa %xmm0,%xmm4
1050 movdqa %xmm0,%xmm3
1051 psllq $5,%xmm0
1052 pxor %xmm0,%xmm3
1053 psllq $1,%xmm0
1054 pxor %xmm3,%xmm0
1055 psllq $57,%xmm0
1056 movdqa %xmm0,%xmm3
1057 pslldq $8,%xmm0
1058 psrldq $8,%xmm3
1059 pxor %xmm4,%xmm0
1060 pxor %xmm3,%xmm1
1061
1062
1063 movdqa %xmm0,%xmm4
1064 psrlq $1,%xmm0
1065 pxor %xmm4,%xmm1
1066 pxor %xmm0,%xmm4
1067 psrlq $5,%xmm0
1068 pxor %xmm4,%xmm0
1069 psrlq $1,%xmm0
1070 pxor %xmm1,%xmm0
1071 addq $0x40,%rcx
1072 jz .Ldone
1073 movdqu 32(%rsi),%xmm7
1074 subq $0x10,%rcx
1075 jz .Lodd_tail
1076 .Lskip4x:
1077
1078
1079
1080
1081
1082 movdqu (%rdx),%xmm8
1083 movdqu 16(%rdx),%xmm3
1084 .byte 102,69,15,56,0,194
1085 .byte 102,65,15,56,0,218
1086 pxor %xmm8,%xmm0
1087
1088 movdqa %xmm3,%xmm5
1089 pshufd $78,%xmm3,%xmm4
1090 pxor %xmm3,%xmm4
1091 .byte 102,15,58,68,218,0
1092 .byte 102,15,58,68,234,17
1093 .byte 102,15,58,68,231,0
1094
1095 leaq 32(%rdx),%rdx
1096 nop
1097 subq $0x20,%rcx
1098 jbe .Leven_tail
1099 nop
1100 jmp .Lmod_loop
1101
1102 .align 32
1103 .Lmod_loop:
1104 movdqa %xmm0,%xmm1
1105 movdqa %xmm4,%xmm8
1106 pshufd $78,%xmm0,%xmm4
1107 pxor %xmm0,%xmm4
1108
1109 .byte 102,15,58,68,198,0
1110 .byte 102,15,58,68,206,17
1111 .byte 102,15,58,68,231,16
1112
1113 pxor %xmm3,%xmm0
1114 pxor %xmm5,%xmm1
1115 movdqu (%rdx),%xmm9
1116 pxor %xmm0,%xmm8
1117 .byte 102,69,15,56,0,202
1118 movdqu 16(%rdx),%xmm3
1119
1120 pxor %xmm1,%xmm8
1121 pxor %xmm9,%xmm1
1122 pxor %xmm8,%xmm4
1123 .byte 102,65,15,56,0,218
1124 movdqa %xmm4,%xmm8
1125 psrldq $8,%xmm8
1126 pslldq $8,%xmm4
1127 pxor %xmm8,%xmm1
1128 pxor %xmm4,%xmm0
1129
1130 movdqa %xmm3,%xmm5
1131
1132 movdqa %xmm0,%xmm9
1133 movdqa %xmm0,%xmm8
1134 psllq $5,%xmm0
1135 pxor %xmm0,%xmm8
1136 .byte 102,15,58,68,218,0
1137 psllq $1,%xmm0
1138 pxor %xmm8,%xmm0
1139 psllq $57,%xmm0
1140 movdqa %xmm0,%xmm8
1141 pslldq $8,%xmm0
1142 psrldq $8,%xmm8
1143 pxor %xmm9,%xmm0
1144 pshufd $78,%xmm5,%xmm4
1145 pxor %xmm8,%xmm1
1146 pxor %xmm5,%xmm4
1147
1148 movdqa %xmm0,%xmm9
1149 psrlq $1,%xmm0
1150 .byte 102,15,58,68,234,17
1151 pxor %xmm9,%xmm1
1152 pxor %xmm0,%xmm9
1153 psrlq $5,%xmm0
1154 pxor %xmm9,%xmm0
1155 leaq 32(%rdx),%rdx
1156 psrlq $1,%xmm0
1157 .byte 102,15,58,68,231,0
1158 pxor %xmm1,%xmm0
1159
1160 subq $0x20,%rcx
1161 ja .Lmod_loop
1162
1163 .Leven_tail:
1164 movdqa %xmm0,%xmm1
1165 movdqa %xmm4,%xmm8
1166 pshufd $78,%xmm0,%xmm4
1167 pxor %xmm0,%xmm4
1168
1169 .byte 102,15,58,68,198,0
1170 .byte 102,15,58,68,206,17
1171 .byte 102,15,58,68,231,16
1172
1173 pxor %xmm3,%xmm0
1174 pxor %xmm5,%xmm1
1175 pxor %xmm0,%xmm8
1176 pxor %xmm1,%xmm8
1177 pxor %xmm8,%xmm4
1178 movdqa %xmm4,%xmm8
1179 psrldq $8,%xmm8
1180 pslldq $8,%xmm4
1181 pxor %xmm8,%xmm1
1182 pxor %xmm4,%xmm0
1183
1184 movdqa %xmm0,%xmm4
1185 movdqa %xmm0,%xmm3
1186 psllq $5,%xmm0
1187 pxor %xmm0,%xmm3
1188 psllq $1,%xmm0
1189 pxor %xmm3,%xmm0
1190 psllq $57,%xmm0
1191 movdqa %xmm0,%xmm3
1192 pslldq $8,%xmm0
1193 psrldq $8,%xmm3
1194 pxor %xmm4,%xmm0
1195 pxor %xmm3,%xmm1
1196
1197
1198 movdqa %xmm0,%xmm4
1199 psrlq $1,%xmm0
1200 pxor %xmm4,%xmm1
1201 pxor %xmm0,%xmm4
1202 psrlq $5,%xmm0
1203 pxor %xmm4,%xmm0
1204 psrlq $1,%xmm0
1205 pxor %xmm1,%xmm0
1206 testq %rcx,%rcx
1207 jnz .Ldone
1208
1209 .Lodd_tail:
1210 movdqu (%rdx),%xmm8
1211 .byte 102,69,15,56,0,194
1212 pxor %xmm8,%xmm0
1213 movdqa %xmm0,%xmm1
1214 pshufd $78,%xmm0,%xmm3
1215 pxor %xmm0,%xmm3
1216 .byte 102,15,58,68,194,0
1217 .byte 102,15,58,68,202,17
1218 .byte 102,15,58,68,223,0
1219 pxor %xmm0,%xmm3
1220 pxor %xmm1,%xmm3
1221
1222 movdqa %xmm3,%xmm4
1223 psrldq $8,%xmm3
1224 pslldq $8,%xmm4
1225 pxor %xmm3,%xmm1
1226 pxor %xmm4,%xmm0
1227
1228 movdqa %xmm0,%xmm4
1229 movdqa %xmm0,%xmm3
1230 psllq $5,%xmm0
1231 pxor %xmm0,%xmm3
1232 psllq $1,%xmm0
1233 pxor %xmm3,%xmm0
1234 psllq $57,%xmm0
1235 movdqa %xmm0,%xmm3
1236 pslldq $8,%xmm0
1237 psrldq $8,%xmm3
1238 pxor %xmm4,%xmm0
1239 pxor %xmm3,%xmm1
1240
1241
1242 movdqa %xmm0,%xmm4
1243 psrlq $1,%xmm0
1244 pxor %xmm4,%xmm1
1245 pxor %xmm0,%xmm4
1246 psrlq $5,%xmm0
1247 pxor %xmm4,%xmm0
1248 psrlq $1,%xmm0
1249 pxor %xmm1,%xmm0
1250 .Ldone:
1251 .byte 102,65,15,56,0,194
1252 movdqu %xmm0,(%rdi)
1253 .byte 0xf3,0xc3
1254 .size gcm_ghash_clmul,.-gcm_ghash_clmul
1255 .globl gcm_init_avx
1256 .hidden gcm_init_avx
1257 .type gcm_init_avx,@function
1258 .align 32
1259 gcm_init_avx:
1260 jmp .L_init_clmul
1261 .size gcm_init_avx,.-gcm_init_avx
1262 .globl gcm_gmult_avx
1263 .hidden gcm_gmult_avx
1264 .type gcm_gmult_avx,@function
1265 .align 32
1266 gcm_gmult_avx:
1267 jmp .L_gmult_clmul
1268 .size gcm_gmult_avx,.-gcm_gmult_avx
1269 .globl gcm_ghash_avx
1270 .hidden gcm_ghash_avx
1271 .type gcm_ghash_avx,@function
1272 .align 32
1273 gcm_ghash_avx:
1274 jmp .L_ghash_clmul
1275 .size gcm_ghash_avx,.-gcm_ghash_avx
1276 .align 64
1277 .Lbswap_mask:
1278 .byte 15,14,13,12,11,10,9,8,7,6,5,4,3,2,1,0
1279 .L0x1c2_polynomial:
1280 .byte 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0xc2
1281 .L7_mask:
1282 .long 7,0,7,0
1283 .L7_mask_poly:
1284 .long 7,0,450,0
1285 .align 64
1286 .type .Lrem_4bit,@object
1287 .Lrem_4bit:
1288 .long 0,0,0,471859200,0,943718400,0,610271232
1289 .long 0,1887436800,0,1822425088,0,1220542464,0,1423966208
1290 .long 0,3774873600,0,4246732800,0,3644850176,0,3311403008
1291 .long 0,2441084928,0,2376073216,0,2847932416,0,3051356160
1292 .type .Lrem_8bit,@object
1293 .Lrem_8bit:
1294 .value 0x0000,0x01C2,0x0384,0x0246,0x0708,0x06CA,0x048C,0x054E
1295 .value 0x0E10,0x0FD2,0x0D94,0x0C56,0x0918,0x08DA,0x0A9C,0x0B5E
1296 .value 0x1C20,0x1DE2,0x1FA4,0x1E66,0x1B28,0x1AEA,0x18AC,0x196E
1297 .value 0x1230,0x13F2,0x11B4,0x1076,0x1538,0x14FA,0x16BC,0x177E
1298 .value 0x3840,0x3982,0x3BC4,0x3A06,0x3F48,0x3E8A,0x3CCC,0x3D0E
1299 .value 0x3650,0x3792,0x35D4,0x3416,0x3158,0x309A,0x32DC,0x331E
1300 .value 0x2460,0x25A2,0x27E4,0x2626,0x2368,0x22AA,0x20EC,0x212E
1301 .value 0x2A70,0x2BB2,0x29F4,0x2836,0x2D78,0x2CBA,0x2EFC,0x2F3E
1302 .value 0x7080,0x7142,0x7304,0x72C6,0x7788,0x764A,0x740C,0x75CE
1303 .value 0x7E90,0x7F52,0x7D14,0x7CD6,0x7998,0x785A,0x7A1C,0x7BDE
1304 .value 0x6CA0,0x6D62,0x6F24,0x6EE6,0x6BA8,0x6A6A,0x682C,0x69EE
1305 .value 0x62B0,0x6372,0x6134,0x60F6,0x65B8,0x647A,0x663C,0x67FE
1306 .value 0x48C0,0x4902,0x4B44,0x4A86,0x4FC8,0x4E0A,0x4C4C,0x4D8E
1307 .value 0x46D0,0x4712,0x4554,0x4496,0x41D8,0x401A,0x425C,0x439E
1308 .value 0x54E0,0x5522,0x5764,0x56A6,0x53E8,0x522A,0x506C,0x51AE
1309 .value 0x5AF0,0x5B32,0x5974,0x58B6,0x5DF8,0x5C3A,0x5E7C,0x5FBE
1310 .value 0xE100,0xE0C2,0xE284,0xE346,0xE608,0xE7CA,0xE58C,0xE44E
1311 .value 0xEF10,0xEED2,0xEC94,0xED56,0xE818,0xE9DA,0xEB9C,0xEA5E
1312 .value 0xFD20,0xFCE2,0xFEA4,0xFF66,0xFA28,0xFBEA,0xF9AC,0xF86E
1313 .value 0xF330,0xF2F2,0xF0B4,0xF176,0xF438,0xF5FA,0xF7BC,0xF67E
1314 .value 0xD940,0xD882,0xDAC4,0xDB06,0xDE48,0xDF8A,0xDDCC,0xDC0E
1315 .value 0xD750,0xD692,0xD4D4,0xD516,0xD058,0xD19A,0xD3DC,0xD21E
1316 .value 0xC560,0xC4A2,0xC6E4,0xC726,0xC268,0xC3AA,0xC1EC,0xC02E
1317 .value 0xCB70,0xCAB2,0xC8F4,0xC936,0xCC78,0xCDBA,0xCFFC,0xCE3E
1318 .value 0x9180,0x9042,0x9204,0x93C6,0x9688,0x974A,0x950C,0x94CE
1319 .value 0x9F90,0x9E52,0x9C14,0x9DD6,0x9898,0x995A,0x9B1C,0x9ADE
1320 .value 0x8DA0,0x8C62,0x8E24,0x8FE6,0x8AA8,0x8B6A,0x892C,0x88EE
1321 .value 0x83B0,0x8272,0x8034,0x81F6,0x84B8,0x857A,0x873C,0x86FE
1322 .value 0xA9C0,0xA802,0xAA44,0xAB86,0xAEC8,0xAF0A,0xAD4C,0xAC8E
1323 .value 0xA7D0,0xA612,0xA454,0xA596,0xA0D8,0xA11A,0xA35C,0xA29E
1324 .value 0xB5E0,0xB422,0xB664,0xB7A6,0xB2E8,0xB32A,0xB16C,0xB0AE
1325 .value 0xBBF0,0xBA32,0xB874,0xB9B6,0xBCF8,0xBD3A,0xBF7C,0xBEBE
1326
1327 .byte 71,72,65,83,72,32,102,111,114,32,120,56,54,95,54,52,44,32,67,82,89,80,84 ,79,71,65,77,83,32,98,121,32,60,97,112,112,114,111,64,111,112,101,110,115,115,10 8,46,111,114,103,62,0
1328 .align 64
1329 #endif
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698