OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 976 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
987 COMPARE(vswp(q0, q15), | 987 COMPARE(vswp(q0, q15), |
988 "f3b2006e vswp q0, q15"); | 988 "f3b2006e vswp q0, q15"); |
989 COMPARE(vswp(q8, q9), | 989 COMPARE(vswp(q8, q9), |
990 "f3f20062 vswp q8, q9"); | 990 "f3f20062 vswp q8, q9"); |
991 COMPARE(vdup(Neon8, q0, r0), | 991 COMPARE(vdup(Neon8, q0, r0), |
992 "eee00b10 vdup.8 q0, r0"); | 992 "eee00b10 vdup.8 q0, r0"); |
993 COMPARE(vdup(Neon16, q1, r4), | 993 COMPARE(vdup(Neon16, q1, r4), |
994 "eea24b30 vdup.16 q1, r4"); | 994 "eea24b30 vdup.16 q1, r4"); |
995 COMPARE(vdup(Neon32, q15, r1), | 995 COMPARE(vdup(Neon32, q15, r1), |
996 "eeae1b90 vdup.32 q15, r1"); | 996 "eeae1b90 vdup.32 q15, r1"); |
997 COMPARE(vdup(q0, s3), | 997 COMPARE(vdup(Neon32, q0, d1, 1), |
998 "f3bc0c41 vdup q0, d1[1]"); | 998 "f3bc0c41 vdup.32 q0, d1[1]"); |
999 COMPARE(vdup(q15, s2), | 999 COMPARE(vdup(Neon32, q15, d1, 0), |
1000 "f3f4ec41 vdup q15, d1[0]"); | 1000 "f3f4ec41 vdup.32 q15, d1[0]"); |
| 1001 COMPARE(vdup(Neon16, q7, d8, 3), |
| 1002 "f3beec48 vdup.16 q7, d8[3]"); |
| 1003 COMPARE(vdup(Neon32, d0, d30, 0), |
| 1004 "f3b40c2e vdup.32 d0, d30[0]"); |
1001 COMPARE(vcvt_f32_s32(q15, q1), | 1005 COMPARE(vcvt_f32_s32(q15, q1), |
1002 "f3fbe642 vcvt.f32.s32 q15, q1"); | 1006 "f3fbe642 vcvt.f32.s32 q15, q1"); |
1003 COMPARE(vcvt_f32_u32(q8, q9), | 1007 COMPARE(vcvt_f32_u32(q8, q9), |
1004 "f3fb06e2 vcvt.f32.u32 q8, q9"); | 1008 "f3fb06e2 vcvt.f32.u32 q8, q9"); |
1005 COMPARE(vcvt_s32_f32(q15, q1), | 1009 COMPARE(vcvt_s32_f32(q15, q1), |
1006 "f3fbe742 vcvt.s32.f32 q15, q1"); | 1010 "f3fbe742 vcvt.s32.f32 q15, q1"); |
1007 COMPARE(vcvt_u32_f32(q8, q9), | 1011 COMPARE(vcvt_u32_f32(q8, q9), |
1008 "f3fb07e2 vcvt.u32.f32 q8, q9"); | 1012 "f3fb07e2 vcvt.u32.f32 q8, q9"); |
1009 COMPARE(vabs(q0, q1), | 1013 COMPARE(vabs(q0, q1), |
1010 "f3b90742 vabs.f32 q0, q1"); | 1014 "f3b90742 vabs.f32 q0, q1"); |
(...skipping 88 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1099 COMPARE(vshl(NeonU16, q15, q0, 10), | 1103 COMPARE(vshl(NeonU16, q15, q0, 10), |
1100 "f2dae550 vshl.i16 q15, q0, #10"); | 1104 "f2dae550 vshl.i16 q15, q0, #10"); |
1101 COMPARE(vshl(NeonS32, q15, q0, 17), | 1105 COMPARE(vshl(NeonS32, q15, q0, 17), |
1102 "f2f1e550 vshl.i32 q15, q0, #17"); | 1106 "f2f1e550 vshl.i32 q15, q0, #17"); |
1103 COMPARE(vshr(NeonS8, q15, q0, 6), | 1107 COMPARE(vshr(NeonS8, q15, q0, 6), |
1104 "f2cae050 vshr.s8 q15, q0, #6"); | 1108 "f2cae050 vshr.s8 q15, q0, #6"); |
1105 COMPARE(vshr(NeonU16, q15, q0, 10), | 1109 COMPARE(vshr(NeonU16, q15, q0, 10), |
1106 "f3d6e050 vshr.u16 q15, q0, #10"); | 1110 "f3d6e050 vshr.u16 q15, q0, #10"); |
1107 COMPARE(vshr(NeonS32, q15, q0, 17), | 1111 COMPARE(vshr(NeonS32, q15, q0, 17), |
1108 "f2efe050 vshr.s32 q15, q0, #17"); | 1112 "f2efe050 vshr.s32 q15, q0, #17"); |
| 1113 COMPARE(vsli(Neon64, d2, d0, 32), |
| 1114 "f3a02590 vsli.64 d2, d0, #32"); |
| 1115 COMPARE(vsli(Neon32, d7, d8, 17), |
| 1116 "f3b17518 vsli.32 d7, d8, #17"); |
| 1117 COMPARE(vsri(Neon64, d2, d0, 32), |
| 1118 "f3a02490 vsri.64 d2, d0, #32"); |
| 1119 COMPARE(vsri(Neon16, d7, d8, 8), |
| 1120 "f3987418 vsri.16 d7, d8, #8"); |
1109 COMPARE(vrecpe(q15, q0), | 1121 COMPARE(vrecpe(q15, q0), |
1110 "f3fbe540 vrecpe.f32 q15, q0"); | 1122 "f3fbe540 vrecpe.f32 q15, q0"); |
1111 COMPARE(vrecps(q15, q0, q8), | 1123 COMPARE(vrecps(q15, q0, q8), |
1112 "f240ef70 vrecps.f32 q15, q0, q8"); | 1124 "f240ef70 vrecps.f32 q15, q0, q8"); |
1113 COMPARE(vrsqrte(q15, q0), | 1125 COMPARE(vrsqrte(q15, q0), |
1114 "f3fbe5c0 vrsqrte.f32 q15, q0"); | 1126 "f3fbe5c0 vrsqrte.f32 q15, q0"); |
1115 COMPARE(vrsqrts(q15, q0, q8), | 1127 COMPARE(vrsqrts(q15, q0, q8), |
1116 "f260ef70 vrsqrts.f32 q15, q0, q8"); | 1128 "f260ef70 vrsqrts.f32 q15, q0, q8"); |
1117 COMPARE(vtst(Neon8, q0, q1, q2), | 1129 COMPARE(vtst(Neon8, q0, q1, q2), |
1118 "f2020854 vtst.i8 q0, q1, q2"); | 1130 "f2020854 vtst.i8 q0, q1, q2"); |
(...skipping 377 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1496 | 1508 |
1497 COMPARE(ldrexb(r0, r1), "e1d10f9f ldrexb r0, [r1]"); | 1509 COMPARE(ldrexb(r0, r1), "e1d10f9f ldrexb r0, [r1]"); |
1498 COMPARE(strexb(r0, r1, r2), "e1c20f91 strexb r0, r1, [r2]"); | 1510 COMPARE(strexb(r0, r1, r2), "e1c20f91 strexb r0, r1, [r2]"); |
1499 COMPARE(ldrexh(r0, r1), "e1f10f9f ldrexh r0, [r1]"); | 1511 COMPARE(ldrexh(r0, r1), "e1f10f9f ldrexh r0, [r1]"); |
1500 COMPARE(strexh(r0, r1, r2), "e1e20f91 strexh r0, r1, [r2]"); | 1512 COMPARE(strexh(r0, r1, r2), "e1e20f91 strexh r0, r1, [r2]"); |
1501 COMPARE(ldrex(r0, r1), "e1910f9f ldrex r0, [r1]"); | 1513 COMPARE(ldrex(r0, r1), "e1910f9f ldrex r0, [r1]"); |
1502 COMPARE(strex(r0, r1, r2), "e1820f91 strex r0, r1, [r2]"); | 1514 COMPARE(strex(r0, r1, r2), "e1820f91 strex r0, r1, [r2]"); |
1503 | 1515 |
1504 VERIFY_RUN(); | 1516 VERIFY_RUN(); |
1505 } | 1517 } |
OLD | NEW |