| OLD | NEW |
| 1 // Copyright (c) 1994-2006 Sun Microsystems Inc. | 1 // Copyright (c) 1994-2006 Sun Microsystems Inc. |
| 2 // All Rights Reserved. | 2 // All Rights Reserved. |
| 3 // | 3 // |
| 4 // Redistribution and use in source and binary forms, with or without | 4 // Redistribution and use in source and binary forms, with or without |
| 5 // modification, are permitted provided that the following conditions are | 5 // modification, are permitted provided that the following conditions are |
| 6 // met: | 6 // met: |
| 7 // | 7 // |
| 8 // - Redistributions of source code must retain the above copyright notice, | 8 // - Redistributions of source code must retain the above copyright notice, |
| 9 // this list of conditions and the following disclaimer. | 9 // this list of conditions and the following disclaimer. |
| 10 // | 10 // |
| (...skipping 1293 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1304 } | 1304 } |
| 1305 void vmaxsd(XMMRegister dst, XMMRegister src1, const Operand& src2) { | 1305 void vmaxsd(XMMRegister dst, XMMRegister src1, const Operand& src2) { |
| 1306 vsd(0x5f, dst, src1, src2); | 1306 vsd(0x5f, dst, src1, src2); |
| 1307 } | 1307 } |
| 1308 void vminsd(XMMRegister dst, XMMRegister src1, XMMRegister src2) { | 1308 void vminsd(XMMRegister dst, XMMRegister src1, XMMRegister src2) { |
| 1309 vsd(0x5d, dst, src1, src2); | 1309 vsd(0x5d, dst, src1, src2); |
| 1310 } | 1310 } |
| 1311 void vminsd(XMMRegister dst, XMMRegister src1, const Operand& src2) { | 1311 void vminsd(XMMRegister dst, XMMRegister src1, const Operand& src2) { |
| 1312 vsd(0x5d, dst, src1, src2); | 1312 vsd(0x5d, dst, src1, src2); |
| 1313 } | 1313 } |
| 1314 void vcvtlsi2sd(XMMRegister dst, XMMRegister src1, Register src2) { | |
| 1315 XMMRegister isrc2 = {src2.code()}; | |
| 1316 vsd(0x2a, dst, src1, isrc2); | |
| 1317 } | |
| 1318 void vcvtlsi2sd(XMMRegister dst, XMMRegister src1, const Operand& src2) { | |
| 1319 vsd(0x2a, dst, src1, src2); | |
| 1320 } | |
| 1321 void vucomisd(XMMRegister dst, XMMRegister src); | 1314 void vucomisd(XMMRegister dst, XMMRegister src); |
| 1322 void vucomisd(XMMRegister dst, const Operand& src); | 1315 void vucomisd(XMMRegister dst, const Operand& src); |
| 1323 void vsd(byte op, XMMRegister dst, XMMRegister src1, XMMRegister src2); | 1316 void vsd(byte op, XMMRegister dst, XMMRegister src1, XMMRegister src2); |
| 1324 void vsd(byte op, XMMRegister dst, XMMRegister src1, const Operand& src2); | 1317 void vsd(byte op, XMMRegister dst, XMMRegister src1, const Operand& src2); |
| 1325 | 1318 |
| 1326 void vaddss(XMMRegister dst, XMMRegister src1, XMMRegister src2) { | 1319 void vaddss(XMMRegister dst, XMMRegister src1, XMMRegister src2) { |
| 1327 vss(0x58, dst, src1, src2); | 1320 vss(0x58, dst, src1, src2); |
| 1328 } | 1321 } |
| 1329 void vaddss(XMMRegister dst, XMMRegister src1, const Operand& src2) { | 1322 void vaddss(XMMRegister dst, XMMRegister src1, const Operand& src2) { |
| 1330 vss(0x58, dst, src1, src2); | 1323 vss(0x58, dst, src1, src2); |
| (...skipping 788 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2119 Assembler* assembler_; | 2112 Assembler* assembler_; |
| 2120 #ifdef DEBUG | 2113 #ifdef DEBUG |
| 2121 int space_before_; | 2114 int space_before_; |
| 2122 #endif | 2115 #endif |
| 2123 }; | 2116 }; |
| 2124 | 2117 |
| 2125 } // namespace internal | 2118 } // namespace internal |
| 2126 } // namespace v8 | 2119 } // namespace v8 |
| 2127 | 2120 |
| 2128 #endif // V8_X64_ASSEMBLER_X64_H_ | 2121 #endif // V8_X64_ASSEMBLER_X64_H_ |
| OLD | NEW |