OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 23 matching lines...) Expand all Loading... |
34 #include "full-codegen.h" | 34 #include "full-codegen.h" |
35 #include "global-handles.h" | 35 #include "global-handles.h" |
36 #include "macro-assembler.h" | 36 #include "macro-assembler.h" |
37 #include "prettyprinter.h" | 37 #include "prettyprinter.h" |
38 | 38 |
39 | 39 |
40 namespace v8 { | 40 namespace v8 { |
41 namespace internal { | 41 namespace internal { |
42 | 42 |
43 DeoptimizerData::DeoptimizerData() { | 43 DeoptimizerData::DeoptimizerData() { |
44 eager_deoptimization_entry_code_ = NULL; | 44 eager_deoptimization_entry_code_entries_ = -1; |
45 lazy_deoptimization_entry_code_ = NULL; | 45 lazy_deoptimization_entry_code_entries_ = -1; |
| 46 size_t deopt_table_size = Deoptimizer::GetMaxDeoptTableSize(); |
| 47 eager_deoptimization_entry_code_ = new VirtualMemory(deopt_table_size); |
| 48 lazy_deoptimization_entry_code_ = new VirtualMemory(deopt_table_size); |
46 current_ = NULL; | 49 current_ = NULL; |
47 deoptimizing_code_list_ = NULL; | 50 deoptimizing_code_list_ = NULL; |
48 #ifdef ENABLE_DEBUGGER_SUPPORT | 51 #ifdef ENABLE_DEBUGGER_SUPPORT |
49 deoptimized_frame_info_ = NULL; | 52 deoptimized_frame_info_ = NULL; |
50 #endif | 53 #endif |
51 } | 54 } |
52 | 55 |
53 | 56 |
54 DeoptimizerData::~DeoptimizerData() { | 57 DeoptimizerData::~DeoptimizerData() { |
55 if (eager_deoptimization_entry_code_ != NULL) { | 58 delete eager_deoptimization_entry_code_; |
56 Isolate::Current()->memory_allocator()->Free( | 59 eager_deoptimization_entry_code_ = NULL; |
57 eager_deoptimization_entry_code_); | 60 delete lazy_deoptimization_entry_code_; |
58 eager_deoptimization_entry_code_ = NULL; | 61 lazy_deoptimization_entry_code_ = NULL; |
59 } | 62 |
60 if (lazy_deoptimization_entry_code_ != NULL) { | |
61 Isolate::Current()->memory_allocator()->Free( | |
62 lazy_deoptimization_entry_code_); | |
63 lazy_deoptimization_entry_code_ = NULL; | |
64 } | |
65 DeoptimizingCodeListNode* current = deoptimizing_code_list_; | 63 DeoptimizingCodeListNode* current = deoptimizing_code_list_; |
66 while (current != NULL) { | 64 while (current != NULL) { |
67 DeoptimizingCodeListNode* prev = current; | 65 DeoptimizingCodeListNode* prev = current; |
68 current = current->next(); | 66 current = current->next(); |
69 delete prev; | 67 delete prev; |
70 } | 68 } |
71 deoptimizing_code_list_ = NULL; | 69 deoptimizing_code_list_ = NULL; |
72 } | 70 } |
73 | 71 |
74 | 72 |
(...skipping 21 matching lines...) Expand all Loading... |
96 bailout_id, | 94 bailout_id, |
97 from, | 95 from, |
98 fp_to_sp_delta, | 96 fp_to_sp_delta, |
99 NULL); | 97 NULL); |
100 ASSERT(isolate->deoptimizer_data()->current_ == NULL); | 98 ASSERT(isolate->deoptimizer_data()->current_ == NULL); |
101 isolate->deoptimizer_data()->current_ = deoptimizer; | 99 isolate->deoptimizer_data()->current_ = deoptimizer; |
102 return deoptimizer; | 100 return deoptimizer; |
103 } | 101 } |
104 | 102 |
105 | 103 |
| 104 // No larger than 2K on all platforms |
| 105 static const int kDeoptTableMaxEpilogueCodeSize = 2 * KB; |
| 106 |
| 107 |
| 108 size_t Deoptimizer::GetMaxDeoptTableSize() { |
| 109 size_t entries_size = |
| 110 Deoptimizer::kMaxNumberOfEntries * Deoptimizer::table_entry_size_; |
| 111 int page_count = ((kDeoptTableMaxEpilogueCodeSize + entries_size - 1) / |
| 112 OS::CommitPageSize()) + 1; |
| 113 return OS::CommitPageSize() * page_count; |
| 114 } |
| 115 |
| 116 |
106 Deoptimizer* Deoptimizer::Grab(Isolate* isolate) { | 117 Deoptimizer* Deoptimizer::Grab(Isolate* isolate) { |
107 ASSERT(isolate == Isolate::Current()); | 118 ASSERT(isolate == Isolate::Current()); |
108 Deoptimizer* result = isolate->deoptimizer_data()->current_; | 119 Deoptimizer* result = isolate->deoptimizer_data()->current_; |
109 ASSERT(result != NULL); | 120 ASSERT(result != NULL); |
110 result->DeleteFrameDescriptions(); | 121 result->DeleteFrameDescriptions(); |
111 isolate->deoptimizer_data()->current_ = NULL; | 122 isolate->deoptimizer_data()->current_ = NULL; |
112 return result; | 123 return result; |
113 } | 124 } |
114 | 125 |
115 | 126 |
(...skipping 338 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
454 for (int i = 0; i < output_count_; ++i) { | 465 for (int i = 0; i < output_count_; ++i) { |
455 if (output_[i] != input_) delete output_[i]; | 466 if (output_[i] != input_) delete output_[i]; |
456 } | 467 } |
457 delete[] output_; | 468 delete[] output_; |
458 input_ = NULL; | 469 input_ = NULL; |
459 output_ = NULL; | 470 output_ = NULL; |
460 ASSERT(!HEAP->allow_allocation(true)); | 471 ASSERT(!HEAP->allow_allocation(true)); |
461 } | 472 } |
462 | 473 |
463 | 474 |
464 Address Deoptimizer::GetDeoptimizationEntry(int id, BailoutType type) { | 475 Address Deoptimizer::GetDeoptimizationEntry(int id, |
| 476 BailoutType type, |
| 477 GetEntryMode mode) { |
465 ASSERT(id >= 0); | 478 ASSERT(id >= 0); |
466 if (id >= kNumberOfEntries) return NULL; | 479 if (id >= kMaxNumberOfEntries) return NULL; |
467 MemoryChunk* base = NULL; | 480 VirtualMemory* base = NULL; |
468 DeoptimizerData* data = Isolate::Current()->deoptimizer_data(); | 481 if (mode == ENSURE_ENTRY_CODE) { |
469 if (type == EAGER) { | 482 EnsureCodeForDeoptimizationEntry(type, id); |
470 if (data->eager_deoptimization_entry_code_ == NULL) { | |
471 data->eager_deoptimization_entry_code_ = CreateCode(type); | |
472 } | |
473 base = data->eager_deoptimization_entry_code_; | |
474 } else { | 483 } else { |
475 if (data->lazy_deoptimization_entry_code_ == NULL) { | 484 ASSERT(mode == CALCULATE_ENTRY_ADDRESS); |
476 data->lazy_deoptimization_entry_code_ = CreateCode(type); | |
477 } | |
478 base = data->lazy_deoptimization_entry_code_; | |
479 } | 485 } |
480 return | |
481 static_cast<Address>(base->area_start()) + (id * table_entry_size_); | |
482 } | |
483 | |
484 | |
485 int Deoptimizer::GetDeoptimizationId(Address addr, BailoutType type) { | |
486 MemoryChunk* base = NULL; | |
487 DeoptimizerData* data = Isolate::Current()->deoptimizer_data(); | 486 DeoptimizerData* data = Isolate::Current()->deoptimizer_data(); |
488 if (type == EAGER) { | 487 if (type == EAGER) { |
489 base = data->eager_deoptimization_entry_code_; | 488 base = data->eager_deoptimization_entry_code_; |
490 } else { | 489 } else { |
491 base = data->lazy_deoptimization_entry_code_; | 490 base = data->lazy_deoptimization_entry_code_; |
492 } | 491 } |
| 492 return |
| 493 static_cast<Address>(base->address()) + (id * table_entry_size_); |
| 494 } |
| 495 |
| 496 |
| 497 int Deoptimizer::GetDeoptimizationId(Address addr, BailoutType type) { |
| 498 VirtualMemory* base = NULL; |
| 499 DeoptimizerData* data = Isolate::Current()->deoptimizer_data(); |
| 500 if (type == EAGER) { |
| 501 base = data->eager_deoptimization_entry_code_; |
| 502 } else { |
| 503 base = data->lazy_deoptimization_entry_code_; |
| 504 } |
| 505 Address base_casted = reinterpret_cast<Address>(base->address()); |
493 if (base == NULL || | 506 if (base == NULL || |
494 addr < base->area_start() || | 507 addr < base->address() || |
495 addr >= base->area_start() + | 508 addr >= base_casted + (kMaxNumberOfEntries * table_entry_size_)) { |
496 (kNumberOfEntries * table_entry_size_)) { | |
497 return kNotDeoptimizationEntry; | 509 return kNotDeoptimizationEntry; |
498 } | 510 } |
499 ASSERT_EQ(0, | 511 ASSERT_EQ(0, |
500 static_cast<int>(addr - base->area_start()) % table_entry_size_); | 512 static_cast<int>(addr - base_casted) % table_entry_size_); |
501 return static_cast<int>(addr - base->area_start()) / table_entry_size_; | 513 return static_cast<int>(addr - base_casted) / table_entry_size_; |
502 } | 514 } |
503 | 515 |
504 | 516 |
505 int Deoptimizer::GetOutputInfo(DeoptimizationOutputData* data, | 517 int Deoptimizer::GetOutputInfo(DeoptimizationOutputData* data, |
506 BailoutId id, | 518 BailoutId id, |
507 SharedFunctionInfo* shared) { | 519 SharedFunctionInfo* shared) { |
508 // TODO(kasperl): For now, we do a simple linear search for the PC | 520 // TODO(kasperl): For now, we do a simple linear search for the PC |
509 // offset associated with the given node id. This should probably be | 521 // offset associated with the given node id. This should probably be |
510 // changed to a binary search. | 522 // changed to a binary search. |
511 int length = data->DeoptPoints(); | 523 int length = data->DeoptPoints(); |
(...skipping 865 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1377 } | 1389 } |
1378 | 1390 |
1379 | 1391 |
1380 void Deoptimizer::AddDoubleValue(intptr_t slot_address, double value) { | 1392 void Deoptimizer::AddDoubleValue(intptr_t slot_address, double value) { |
1381 HeapNumberMaterializationDescriptor value_desc( | 1393 HeapNumberMaterializationDescriptor value_desc( |
1382 reinterpret_cast<Address>(slot_address), value); | 1394 reinterpret_cast<Address>(slot_address), value); |
1383 deferred_heap_numbers_.Add(value_desc); | 1395 deferred_heap_numbers_.Add(value_desc); |
1384 } | 1396 } |
1385 | 1397 |
1386 | 1398 |
1387 MemoryChunk* Deoptimizer::CreateCode(BailoutType type) { | 1399 void Deoptimizer::EnsureCodeForDeoptimizationEntry(BailoutType type, |
| 1400 int max_entry_id) { |
1388 // We cannot run this if the serializer is enabled because this will | 1401 // We cannot run this if the serializer is enabled because this will |
1389 // cause us to emit relocation information for the external | 1402 // cause us to emit relocation information for the external |
1390 // references. This is fine because the deoptimizer's code section | 1403 // references. This is fine because the deoptimizer's code section |
1391 // isn't meant to be serialized at all. | 1404 // isn't meant to be serialized at all. |
1392 ASSERT(!Serializer::enabled()); | 1405 ASSERT(!Serializer::enabled()); |
1393 | 1406 |
| 1407 ASSERT(type == EAGER || type == LAZY); |
| 1408 DeoptimizerData* data = Isolate::Current()->deoptimizer_data(); |
| 1409 int entry_count = (type == EAGER) |
| 1410 ? data->eager_deoptimization_entry_code_entries_ |
| 1411 : data->lazy_deoptimization_entry_code_entries_; |
| 1412 if (max_entry_id < entry_count) return; |
| 1413 entry_count = Min(Max(entry_count * 2, Deoptimizer::kMinNumberOfEntries), |
| 1414 Deoptimizer::kMaxNumberOfEntries); |
| 1415 |
1394 MacroAssembler masm(Isolate::Current(), NULL, 16 * KB); | 1416 MacroAssembler masm(Isolate::Current(), NULL, 16 * KB); |
1395 masm.set_emit_debug_code(false); | 1417 masm.set_emit_debug_code(false); |
1396 GenerateDeoptimizationEntries(&masm, kNumberOfEntries, type); | 1418 GenerateDeoptimizationEntries(&masm, entry_count, type); |
1397 CodeDesc desc; | 1419 CodeDesc desc; |
1398 masm.GetCode(&desc); | 1420 masm.GetCode(&desc); |
1399 ASSERT(desc.reloc_size == 0); | 1421 ASSERT(desc.reloc_size == 0); |
1400 | 1422 |
1401 MemoryChunk* chunk = | 1423 VirtualMemory* memory = type == EAGER |
1402 Isolate::Current()->memory_allocator()->AllocateChunk(desc.instr_size, | 1424 ? data->eager_deoptimization_entry_code_ |
1403 EXECUTABLE, | 1425 : data->lazy_deoptimization_entry_code_; |
1404 NULL); | 1426 size_t table_size = Deoptimizer::GetMaxDeoptTableSize(); |
1405 ASSERT(chunk->area_size() >= desc.instr_size); | 1427 ASSERT(static_cast<int>(table_size) >= desc.instr_size); |
1406 if (chunk == NULL) { | 1428 memory->Commit(memory->address(), table_size, true); |
1407 V8::FatalProcessOutOfMemory("Not enough memory for deoptimization table"); | 1429 memcpy(memory->address(), desc.buffer, desc.instr_size); |
| 1430 CPU::FlushICache(memory->address(), desc.instr_size); |
| 1431 |
| 1432 if (type == EAGER) { |
| 1433 data->eager_deoptimization_entry_code_entries_ = entry_count; |
| 1434 } else { |
| 1435 data->lazy_deoptimization_entry_code_entries_ = entry_count; |
1408 } | 1436 } |
1409 memcpy(chunk->area_start(), desc.buffer, desc.instr_size); | |
1410 CPU::FlushICache(chunk->area_start(), desc.instr_size); | |
1411 return chunk; | |
1412 } | 1437 } |
1413 | 1438 |
1414 | 1439 |
1415 Code* Deoptimizer::FindDeoptimizingCodeFromAddress(Address addr) { | 1440 Code* Deoptimizer::FindDeoptimizingCodeFromAddress(Address addr) { |
1416 DeoptimizingCodeListNode* node = | 1441 DeoptimizingCodeListNode* node = |
1417 Isolate::Current()->deoptimizer_data()->deoptimizing_code_list_; | 1442 Isolate::Current()->deoptimizer_data()->deoptimizing_code_list_; |
1418 while (node != NULL) { | 1443 while (node != NULL) { |
1419 if (node->code()->contains(addr)) return *node->code(); | 1444 if (node->code()->contains(addr)) return *node->code(); |
1420 node = node->next(); | 1445 node = node->next(); |
1421 } | 1446 } |
(...skipping 574 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1996 | 2021 |
1997 void DeoptimizedFrameInfo::Iterate(ObjectVisitor* v) { | 2022 void DeoptimizedFrameInfo::Iterate(ObjectVisitor* v) { |
1998 v->VisitPointer(BitCast<Object**>(&function_)); | 2023 v->VisitPointer(BitCast<Object**>(&function_)); |
1999 v->VisitPointers(parameters_, parameters_ + parameters_count_); | 2024 v->VisitPointers(parameters_, parameters_ + parameters_count_); |
2000 v->VisitPointers(expression_stack_, expression_stack_ + expression_count_); | 2025 v->VisitPointers(expression_stack_, expression_stack_ + expression_count_); |
2001 } | 2026 } |
2002 | 2027 |
2003 #endif // ENABLE_DEBUGGER_SUPPORT | 2028 #endif // ENABLE_DEBUGGER_SUPPORT |
2004 | 2029 |
2005 } } // namespace v8::internal | 2030 } } // namespace v8::internal |
OLD | NEW |