OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 3139 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3150 return result; | 3150 return result; |
3151 } | 3151 } |
3152 | 3152 |
3153 | 3153 |
3154 MaybeObject* Heap::CreateCode(const CodeDesc& desc, | 3154 MaybeObject* Heap::CreateCode(const CodeDesc& desc, |
3155 Code::Flags flags, | 3155 Code::Flags flags, |
3156 Handle<Object> self_reference, | 3156 Handle<Object> self_reference, |
3157 bool immovable) { | 3157 bool immovable) { |
3158 // Allocate ByteArray before the Code object, so that we do not risk | 3158 // Allocate ByteArray before the Code object, so that we do not risk |
3159 // leaving uninitialized Code object (and breaking the heap). | 3159 // leaving uninitialized Code object (and breaking the heap). |
3160 Object* reloc_info; | 3160 ByteArray* reloc_info; |
3161 { MaybeObject* maybe_reloc_info = AllocateByteArray(desc.reloc_size, TENURED); | 3161 MaybeObject* maybe_reloc_info = AllocateByteArray(desc.reloc_size, TENURED); |
3162 if (!maybe_reloc_info->ToObject(&reloc_info)) return maybe_reloc_info; | 3162 if (!maybe_reloc_info->To(&reloc_info)) return maybe_reloc_info; |
3163 } | |
3164 | 3163 |
3165 // Compute size. | 3164 // Compute size. |
3166 int body_size = RoundUp(desc.instr_size, kObjectAlignment); | 3165 int body_size = RoundUp(desc.instr_size, kObjectAlignment); |
3167 int obj_size = Code::SizeFor(body_size); | 3166 int obj_size = Code::SizeFor(body_size); |
3168 ASSERT(IsAligned(static_cast<intptr_t>(obj_size), kCodeAlignment)); | 3167 ASSERT(IsAligned(static_cast<intptr_t>(obj_size), kCodeAlignment)); |
3169 MaybeObject* maybe_result; | 3168 MaybeObject* maybe_result; |
3170 // Large code objects and code objects which should stay at a fixed address | 3169 // Large code objects and code objects which should stay at a fixed address |
3171 // are allocated in large object space. | 3170 // are allocated in large object space. |
3172 if (obj_size > MaxObjectSizeInPagedSpace() || immovable) { | 3171 if (obj_size > MaxObjectSizeInPagedSpace() || immovable) { |
3173 maybe_result = lo_space_->AllocateRaw(obj_size, EXECUTABLE); | 3172 maybe_result = lo_space_->AllocateRaw(obj_size, EXECUTABLE); |
3174 } else { | 3173 } else { |
3175 maybe_result = code_space_->AllocateRaw(obj_size); | 3174 maybe_result = code_space_->AllocateRaw(obj_size); |
3176 } | 3175 } |
3177 | 3176 |
3178 Object* result; | 3177 Object* result; |
3179 if (!maybe_result->ToObject(&result)) return maybe_result; | 3178 if (!maybe_result->ToObject(&result)) return maybe_result; |
3180 | 3179 |
3181 // Initialize the object | 3180 // Initialize the object |
3182 HeapObject::cast(result)->set_map(code_map()); | 3181 HeapObject::cast(result)->set_map(code_map()); |
3183 Code* code = Code::cast(result); | 3182 Code* code = Code::cast(result); |
3184 ASSERT(!isolate_->code_range()->exists() || | 3183 ASSERT(!isolate_->code_range()->exists() || |
3185 isolate_->code_range()->contains(code->address())); | 3184 isolate_->code_range()->contains(code->address())); |
3186 code->set_instruction_size(desc.instr_size); | 3185 code->set_instruction_size(desc.instr_size); |
3187 code->set_relocation_info(ByteArray::cast(reloc_info)); | 3186 code->set_relocation_info(reloc_info); |
3188 code->set_flags(flags); | 3187 code->set_flags(flags); |
3189 if (code->is_call_stub() || code->is_keyed_call_stub()) { | 3188 if (code->is_call_stub() || code->is_keyed_call_stub()) { |
3190 code->set_check_type(RECEIVER_MAP_CHECK); | 3189 code->set_check_type(RECEIVER_MAP_CHECK); |
3191 } | 3190 } |
3192 code->set_deoptimization_data(empty_fixed_array()); | 3191 code->set_deoptimization_data(empty_fixed_array()); |
| 3192 code->set_handler_table(empty_fixed_array()); |
3193 code->set_next_code_flushing_candidate(undefined_value()); | 3193 code->set_next_code_flushing_candidate(undefined_value()); |
3194 // Allow self references to created code object by patching the handle to | 3194 // Allow self references to created code object by patching the handle to |
3195 // point to the newly allocated Code object. | 3195 // point to the newly allocated Code object. |
3196 if (!self_reference.is_null()) { | 3196 if (!self_reference.is_null()) { |
3197 *(self_reference.location()) = code; | 3197 *(self_reference.location()) = code; |
3198 } | 3198 } |
3199 // Migrate generated code. | 3199 // Migrate generated code. |
3200 // The generated code can contain Object** values (typically from handles) | 3200 // The generated code can contain Object** values (typically from handles) |
3201 // that are dereferenced during the copy to point directly to the actual heap | 3201 // that are dereferenced during the copy to point directly to the actual heap |
3202 // objects. These pointers can include references to the code object itself, | 3202 // objects. These pointers can include references to the code object itself, |
(...skipping 3253 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6456 isolate_->heap()->store_buffer()->Compact(); | 6456 isolate_->heap()->store_buffer()->Compact(); |
6457 isolate_->heap()->store_buffer()->Filter(MemoryChunk::ABOUT_TO_BE_FREED); | 6457 isolate_->heap()->store_buffer()->Filter(MemoryChunk::ABOUT_TO_BE_FREED); |
6458 for (chunk = chunks_queued_for_free_; chunk != NULL; chunk = next) { | 6458 for (chunk = chunks_queued_for_free_; chunk != NULL; chunk = next) { |
6459 next = chunk->next_chunk(); | 6459 next = chunk->next_chunk(); |
6460 isolate_->memory_allocator()->Free(chunk); | 6460 isolate_->memory_allocator()->Free(chunk); |
6461 } | 6461 } |
6462 chunks_queued_for_free_ = NULL; | 6462 chunks_queued_for_free_ = NULL; |
6463 } | 6463 } |
6464 | 6464 |
6465 } } // namespace v8::internal | 6465 } } // namespace v8::internal |
OLD | NEW |