OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 332 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
343 // Implementation of Assembler. | 343 // Implementation of Assembler. |
344 | 344 |
345 #ifdef GENERATED_CODE_COVERAGE | 345 #ifdef GENERATED_CODE_COVERAGE |
346 static void InitCoverageLog(); | 346 static void InitCoverageLog(); |
347 #endif | 347 #endif |
348 | 348 |
349 Assembler::Assembler(Isolate* arg_isolate, void* buffer, int buffer_size) | 349 Assembler::Assembler(Isolate* arg_isolate, void* buffer, int buffer_size) |
350 : AssemblerBase(arg_isolate), | 350 : AssemblerBase(arg_isolate), |
351 code_targets_(100), | 351 code_targets_(100), |
352 positions_recorder_(this), | 352 positions_recorder_(this), |
353 emit_debug_code_(FLAG_debug_code), | 353 emit_debug_code_(FLAG_debug_code) { |
354 predictable_code_size_(false) { | |
355 if (buffer == NULL) { | 354 if (buffer == NULL) { |
356 // Do our own buffer management. | 355 // Do our own buffer management. |
357 if (buffer_size <= kMinimalBufferSize) { | 356 if (buffer_size <= kMinimalBufferSize) { |
358 buffer_size = kMinimalBufferSize; | 357 buffer_size = kMinimalBufferSize; |
359 | 358 |
360 if (isolate() != NULL && isolate()->assembler_spare_buffer() != NULL) { | 359 if (isolate() != NULL && isolate()->assembler_spare_buffer() != NULL) { |
361 buffer = isolate()->assembler_spare_buffer(); | 360 buffer = isolate()->assembler_spare_buffer(); |
362 isolate()->set_assembler_spare_buffer(NULL); | 361 isolate()->set_assembler_spare_buffer(NULL); |
363 } | 362 } |
364 } | 363 } |
(...skipping 866 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1231 EnsureSpace ensure_space(this); | 1230 EnsureSpace ensure_space(this); |
1232 ASSERT(is_uint4(cc)); | 1231 ASSERT(is_uint4(cc)); |
1233 if (L->is_bound()) { | 1232 if (L->is_bound()) { |
1234 const int short_size = 2; | 1233 const int short_size = 2; |
1235 const int long_size = 6; | 1234 const int long_size = 6; |
1236 int offs = L->pos() - pc_offset(); | 1235 int offs = L->pos() - pc_offset(); |
1237 ASSERT(offs <= 0); | 1236 ASSERT(offs <= 0); |
1238 // Determine whether we can use 1-byte offsets for backwards branches, | 1237 // Determine whether we can use 1-byte offsets for backwards branches, |
1239 // which have a max range of 128 bytes. | 1238 // which have a max range of 128 bytes. |
1240 | 1239 |
1241 // We also need to check the predictable_code_size_ flag here, because | 1240 // We also need to check predictable_code_size() flag here, because on x64, |
1242 // on x64, when the full code generator recompiles code for debugging, some | 1241 // when the full code generator recompiles code for debugging, some places |
1243 // places need to be padded out to a certain size. The debugger is keeping | 1242 // need to be padded out to a certain size. The debugger is keeping track of |
1244 // track of how often it did this so that it can adjust return addresses on | 1243 // how often it did this so that it can adjust return addresses on the |
1245 // the stack, but if the size of jump instructions can also change, that's | 1244 // stack, but if the size of jump instructions can also change, that's not |
1246 // not enough and the calculated offsets would be incorrect. | 1245 // enough and the calculated offsets would be incorrect. |
1247 if (is_int8(offs - short_size) && !predictable_code_size_) { | 1246 if (is_int8(offs - short_size) && !predictable_code_size()) { |
1248 // 0111 tttn #8-bit disp. | 1247 // 0111 tttn #8-bit disp. |
1249 emit(0x70 | cc); | 1248 emit(0x70 | cc); |
1250 emit((offs - short_size) & 0xFF); | 1249 emit((offs - short_size) & 0xFF); |
1251 } else { | 1250 } else { |
1252 // 0000 1111 1000 tttn #32-bit disp. | 1251 // 0000 1111 1000 tttn #32-bit disp. |
1253 emit(0x0F); | 1252 emit(0x0F); |
1254 emit(0x80 | cc); | 1253 emit(0x80 | cc); |
1255 emitl(offs - long_size); | 1254 emitl(offs - long_size); |
1256 } | 1255 } |
1257 } else if (distance == Label::kNear) { | 1256 } else if (distance == Label::kNear) { |
(...skipping 36 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1294 } | 1293 } |
1295 | 1294 |
1296 | 1295 |
1297 void Assembler::jmp(Label* L, Label::Distance distance) { | 1296 void Assembler::jmp(Label* L, Label::Distance distance) { |
1298 EnsureSpace ensure_space(this); | 1297 EnsureSpace ensure_space(this); |
1299 const int short_size = sizeof(int8_t); | 1298 const int short_size = sizeof(int8_t); |
1300 const int long_size = sizeof(int32_t); | 1299 const int long_size = sizeof(int32_t); |
1301 if (L->is_bound()) { | 1300 if (L->is_bound()) { |
1302 int offs = L->pos() - pc_offset() - 1; | 1301 int offs = L->pos() - pc_offset() - 1; |
1303 ASSERT(offs <= 0); | 1302 ASSERT(offs <= 0); |
1304 if (is_int8(offs - short_size) && !predictable_code_size_) { | 1303 if (is_int8(offs - short_size) && !predictable_code_size()) { |
1305 // 1110 1011 #8-bit disp. | 1304 // 1110 1011 #8-bit disp. |
1306 emit(0xEB); | 1305 emit(0xEB); |
1307 emit((offs - short_size) & 0xFF); | 1306 emit((offs - short_size) & 0xFF); |
1308 } else { | 1307 } else { |
1309 // 1110 1001 #32-bit disp. | 1308 // 1110 1001 #32-bit disp. |
1310 emit(0xE9); | 1309 emit(0xE9); |
1311 emitl(offs - long_size); | 1310 emitl(offs - long_size); |
1312 } | 1311 } |
1313 } else if (distance == Label::kNear) { | 1312 } else if (distance == Label::kNear) { |
1314 emit(0xEB); | 1313 emit(0xEB); |
(...skipping 1739 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3054 bool RelocInfo::IsCodedSpecially() { | 3053 bool RelocInfo::IsCodedSpecially() { |
3055 // The deserializer needs to know whether a pointer is specially coded. Being | 3054 // The deserializer needs to know whether a pointer is specially coded. Being |
3056 // specially coded on x64 means that it is a relative 32 bit address, as used | 3055 // specially coded on x64 means that it is a relative 32 bit address, as used |
3057 // by branch instructions. | 3056 // by branch instructions. |
3058 return (1 << rmode_) & kApplyMask; | 3057 return (1 << rmode_) & kApplyMask; |
3059 } | 3058 } |
3060 | 3059 |
3061 } } // namespace v8::internal | 3060 } } // namespace v8::internal |
3062 | 3061 |
3063 #endif // V8_TARGET_ARCH_X64 | 3062 #endif // V8_TARGET_ARCH_X64 |
OLD | NEW |