OLD | NEW |
1 // Copyright 2011 the V8 project authors. All rights reserved. | 1 // Copyright 2011 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 148 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
159 | 159 |
160 // Setup the object header. | 160 // Setup the object header. |
161 __ LoadRoot(a2, Heap::kFunctionContextMapRootIndex); | 161 __ LoadRoot(a2, Heap::kFunctionContextMapRootIndex); |
162 __ sw(a2, FieldMemOperand(v0, HeapObject::kMapOffset)); | 162 __ sw(a2, FieldMemOperand(v0, HeapObject::kMapOffset)); |
163 __ li(a2, Operand(Smi::FromInt(length))); | 163 __ li(a2, Operand(Smi::FromInt(length))); |
164 __ sw(a2, FieldMemOperand(v0, FixedArray::kLengthOffset)); | 164 __ sw(a2, FieldMemOperand(v0, FixedArray::kLengthOffset)); |
165 | 165 |
166 // Setup the fixed slots. | 166 // Setup the fixed slots. |
167 __ li(a1, Operand(Smi::FromInt(0))); | 167 __ li(a1, Operand(Smi::FromInt(0))); |
168 __ sw(a3, MemOperand(v0, Context::SlotOffset(Context::CLOSURE_INDEX))); | 168 __ sw(a3, MemOperand(v0, Context::SlotOffset(Context::CLOSURE_INDEX))); |
169 __ sw(v0, MemOperand(v0, Context::SlotOffset(Context::FCONTEXT_INDEX))); | |
170 __ sw(cp, MemOperand(v0, Context::SlotOffset(Context::PREVIOUS_INDEX))); | 169 __ sw(cp, MemOperand(v0, Context::SlotOffset(Context::PREVIOUS_INDEX))); |
171 __ sw(a1, MemOperand(v0, Context::SlotOffset(Context::EXTENSION_INDEX))); | 170 __ sw(a1, MemOperand(v0, Context::SlotOffset(Context::EXTENSION_INDEX))); |
172 | 171 |
173 // Copy the global object from the previous context. | 172 // Copy the global object from the previous context. |
174 __ lw(a1, MemOperand(cp, Context::SlotOffset(Context::GLOBAL_INDEX))); | 173 __ lw(a1, MemOperand(cp, Context::SlotOffset(Context::GLOBAL_INDEX))); |
175 __ sw(a1, MemOperand(v0, Context::SlotOffset(Context::GLOBAL_INDEX))); | 174 __ sw(a1, MemOperand(v0, Context::SlotOffset(Context::GLOBAL_INDEX))); |
176 | 175 |
177 // Initialize the rest of the slots to undefined. | 176 // Initialize the rest of the slots to undefined. |
178 __ LoadRoot(a1, Heap::kUndefinedValueRootIndex); | 177 __ LoadRoot(a1, Heap::kUndefinedValueRootIndex); |
179 for (int i = Context::MIN_CONTEXT_SLOTS; i < length; i++) { | 178 for (int i = Context::MIN_CONTEXT_SLOTS; i < length; i++) { |
(...skipping 6707 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6887 __ mov(result, zero_reg); | 6886 __ mov(result, zero_reg); |
6888 __ Ret(); | 6887 __ Ret(); |
6889 } | 6888 } |
6890 | 6889 |
6891 | 6890 |
6892 #undef __ | 6891 #undef __ |
6893 | 6892 |
6894 } } // namespace v8::internal | 6893 } } // namespace v8::internal |
6895 | 6894 |
6896 #endif // V8_TARGET_ARCH_MIPS | 6895 #endif // V8_TARGET_ARCH_MIPS |
OLD | NEW |