OLD | NEW |
---|---|
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 480 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
491 COMPARE(vcmp(d0, d1), | 491 COMPARE(vcmp(d0, d1), |
492 "eeb40b41 vcmp.f64 d0, d1"); | 492 "eeb40b41 vcmp.f64 d0, d1"); |
493 COMPARE(vcmp(d0, 0.0), | 493 COMPARE(vcmp(d0, 0.0), |
494 "eeb50b40 vcmp.f64 d0, #0.0"); | 494 "eeb50b40 vcmp.f64 d0, #0.0"); |
495 | 495 |
496 COMPARE(vsqrt(d0, d0), | 496 COMPARE(vsqrt(d0, d0), |
497 "eeb10bc0 vsqrt.f64 d0, d0"); | 497 "eeb10bc0 vsqrt.f64 d0, d0"); |
498 COMPARE(vsqrt(d2, d3, ne), | 498 COMPARE(vsqrt(d2, d3, ne), |
499 "1eb12bc3 vsqrtne.f64 d2, d3"); | 499 "1eb12bc3 vsqrtne.f64 d2, d3"); |
500 | 500 |
501 COMPARE(vsqrt(d0, d0), "eeb10bc0 vsqrt.f64 d0, d0"); | |
Rodolph Perfetta (ARM)
2014/10/28 19:41:41
nit: copy/paste error.
sigurds
2014/10/29 11:47:15
Done.
| |
502 COMPARE(vsqrt(d2, d3, ne), "1eb12bc3 vsqrtne.f64 d2, d3"); | |
503 | |
504 COMPARE(vrinta(d0, d0), "feb80b40 vrinta.f64.f64 d0, d0"); | |
505 COMPARE(vrinta(d2, d3), "feb82b43 vrinta.f64.f64 d2, d3"); | |
506 | |
507 COMPARE(vrintp(d0, d0), "feba0b40 vrintp.f64.f64 d0, d0"); | |
508 COMPARE(vrintp(d2, d3), "feba2b43 vrintp.f64.f64 d2, d3"); | |
509 | |
510 COMPARE(vrintn(d0, d0), "feb90b40 vrintn.f64.f64 d0, d0"); | |
511 COMPARE(vrintn(d2, d3), "feb92b43 vrintn.f64.f64 d2, d3"); | |
512 | |
513 COMPARE(vrintm(d0, d0), "febb0b40 vrintm.f64.f64 d0, d0"); | |
514 COMPARE(vrintm(d2, d3), "febb2b43 vrintm.f64.f64 d2, d3"); | |
515 | |
516 COMPARE(vrintz(d0, d0), "eeb60bc0 vrintz.f64.f64 d0, d0"); | |
517 COMPARE(vrintz(d2, d3), "eeb62bc3 vrintz.f64.f64 d2, d3"); | |
518 | |
501 COMPARE(vmov(d0, 1.0), | 519 COMPARE(vmov(d0, 1.0), |
502 "eeb70b00 vmov.f64 d0, #1"); | 520 "eeb70b00 vmov.f64 d0, #1"); |
503 COMPARE(vmov(d2, -13.0), | 521 COMPARE(vmov(d2, -13.0), |
504 "eeba2b0a vmov.f64 d2, #-13"); | 522 "eeba2b0a vmov.f64 d2, #-13"); |
505 | 523 |
506 COMPARE(vmov(d0, VmovIndexLo, r0), | 524 COMPARE(vmov(d0, VmovIndexLo, r0), |
507 "ee000b10 vmov.32 d0[0], r0"); | 525 "ee000b10 vmov.32 d0[0], r0"); |
508 COMPARE(vmov(d0, VmovIndexHi, r0), | 526 COMPARE(vmov(d0, VmovIndexHi, r0), |
509 "ee200b10 vmov.32 d0[1], r0"); | 527 "ee200b10 vmov.32 d0[1], r0"); |
510 | 528 |
(...skipping 392 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
903 "e16dc7ff strd ip, [sp, #-127]!"); | 921 "e16dc7ff strd ip, [sp, #-127]!"); |
904 | 922 |
905 COMPARE(pld(MemOperand(r1, 0)), | 923 COMPARE(pld(MemOperand(r1, 0)), |
906 "f5d1f000 pld [r1]"); | 924 "f5d1f000 pld [r1]"); |
907 COMPARE(pld(MemOperand(r2, 128)), | 925 COMPARE(pld(MemOperand(r2, 128)), |
908 "f5d2f080 pld [r2, #+128]"); | 926 "f5d2f080 pld [r2, #+128]"); |
909 } | 927 } |
910 | 928 |
911 VERIFY_RUN(); | 929 VERIFY_RUN(); |
912 } | 930 } |
OLD | NEW |