OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include "src/v8.h" | 5 #include "src/v8.h" |
6 | 6 |
7 #if V8_TARGET_ARCH_MIPS64 | 7 #if V8_TARGET_ARCH_MIPS64 |
8 | 8 |
9 #include "src/bootstrapper.h" | 9 #include "src/bootstrapper.h" |
10 #include "src/code-stubs.h" | 10 #include "src/code-stubs.h" |
(...skipping 1115 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1126 __ InvokeBuiltin(native, JUMP_FUNCTION); | 1126 __ InvokeBuiltin(native, JUMP_FUNCTION); |
1127 | 1127 |
1128 __ bind(&miss); | 1128 __ bind(&miss); |
1129 GenerateMiss(masm); | 1129 GenerateMiss(masm); |
1130 } | 1130 } |
1131 | 1131 |
1132 | 1132 |
1133 void StoreRegistersStateStub::Generate(MacroAssembler* masm) { | 1133 void StoreRegistersStateStub::Generate(MacroAssembler* masm) { |
1134 __ mov(t9, ra); | 1134 __ mov(t9, ra); |
1135 __ pop(ra); | 1135 __ pop(ra); |
1136 if (save_doubles_ == kSaveFPRegs) { | 1136 __ PushSafepointRegisters(); |
1137 __ PushSafepointRegistersAndDoubles(); | |
1138 } else { | |
1139 __ PushSafepointRegisters(); | |
1140 } | |
1141 __ Jump(t9); | 1137 __ Jump(t9); |
1142 } | 1138 } |
1143 | 1139 |
1144 | 1140 |
1145 void RestoreRegistersStateStub::Generate(MacroAssembler* masm) { | 1141 void RestoreRegistersStateStub::Generate(MacroAssembler* masm) { |
1146 __ mov(t9, ra); | 1142 __ mov(t9, ra); |
1147 __ pop(ra); | 1143 __ pop(ra); |
1148 if (save_doubles_ == kSaveFPRegs) { | 1144 __ PopSafepointRegisters(); |
1149 __ PopSafepointRegistersAndDoubles(); | |
1150 } else { | |
1151 __ PopSafepointRegisters(); | |
1152 } | |
1153 __ Jump(t9); | 1145 __ Jump(t9); |
1154 } | 1146 } |
1155 | 1147 |
1156 | 1148 |
1157 void StoreBufferOverflowStub::Generate(MacroAssembler* masm) { | 1149 void StoreBufferOverflowStub::Generate(MacroAssembler* masm) { |
1158 // We don't allow a GC during a store buffer overflow so there is no need to | 1150 // We don't allow a GC during a store buffer overflow so there is no need to |
1159 // store the registers in any particular way, but we do have to store and | 1151 // store the registers in any particular way, but we do have to store and |
1160 // restore them. | 1152 // restore them. |
1161 __ MultiPush(kJSCallerSaved | ra.bit()); | 1153 __ MultiPush(kJSCallerSaved | ra.bit()); |
1162 if (save_doubles_ == kSaveFPRegs) { | 1154 if (save_doubles_ == kSaveFPRegs) { |
(...skipping 245 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1408 StubFailureTrampolineStub::GenerateAheadOfTime(isolate); | 1400 StubFailureTrampolineStub::GenerateAheadOfTime(isolate); |
1409 ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate); | 1401 ArrayConstructorStubBase::GenerateStubsAheadOfTime(isolate); |
1410 CreateAllocationSiteStub::GenerateAheadOfTime(isolate); | 1402 CreateAllocationSiteStub::GenerateAheadOfTime(isolate); |
1411 BinaryOpICStub::GenerateAheadOfTime(isolate); | 1403 BinaryOpICStub::GenerateAheadOfTime(isolate); |
1412 StoreRegistersStateStub::GenerateAheadOfTime(isolate); | 1404 StoreRegistersStateStub::GenerateAheadOfTime(isolate); |
1413 RestoreRegistersStateStub::GenerateAheadOfTime(isolate); | 1405 RestoreRegistersStateStub::GenerateAheadOfTime(isolate); |
1414 BinaryOpICWithAllocationSiteStub::GenerateAheadOfTime(isolate); | 1406 BinaryOpICWithAllocationSiteStub::GenerateAheadOfTime(isolate); |
1415 } | 1407 } |
1416 | 1408 |
1417 | 1409 |
1418 void StoreRegistersStateStub::GenerateAheadOfTime( | 1410 void StoreRegistersStateStub::GenerateAheadOfTime(Isolate* isolate) { |
1419 Isolate* isolate) { | 1411 StoreRegistersStateStub stub(isolate); |
1420 StoreRegistersStateStub stub1(isolate, kDontSaveFPRegs); | 1412 stub.GetCode(); |
1421 stub1.GetCode(); | |
1422 // Hydrogen code stubs need stub2 at snapshot time. | |
1423 StoreRegistersStateStub stub2(isolate, kSaveFPRegs); | |
1424 stub2.GetCode(); | |
1425 } | 1413 } |
1426 | 1414 |
1427 | 1415 |
1428 void RestoreRegistersStateStub::GenerateAheadOfTime( | 1416 void RestoreRegistersStateStub::GenerateAheadOfTime(Isolate* isolate) { |
1429 Isolate* isolate) { | 1417 RestoreRegistersStateStub stub(isolate); |
1430 RestoreRegistersStateStub stub1(isolate, kDontSaveFPRegs); | 1418 stub.GetCode(); |
1431 stub1.GetCode(); | |
1432 // Hydrogen code stubs need stub2 at snapshot time. | |
1433 RestoreRegistersStateStub stub2(isolate, kSaveFPRegs); | |
1434 stub2.GetCode(); | |
1435 } | 1419 } |
1436 | 1420 |
1437 | 1421 |
1438 void CodeStub::GenerateFPStubs(Isolate* isolate) { | 1422 void CodeStub::GenerateFPStubs(Isolate* isolate) { |
1439 SaveFPRegsMode mode = kSaveFPRegs; | 1423 SaveFPRegsMode mode = kSaveFPRegs; |
1440 CEntryStub save_doubles(isolate, 1, mode); | 1424 CEntryStub save_doubles(isolate, 1, mode); |
1441 StoreBufferOverflowStub stub(isolate, mode); | 1425 StoreBufferOverflowStub stub(isolate, mode); |
1442 // These stubs might already be in the snapshot, detect that and don't | 1426 // These stubs might already be in the snapshot, detect that and don't |
1443 // regenerate, which would lead to code stub initialization state being messed | 1427 // regenerate, which would lead to code stub initialization state being messed |
1444 // up. | 1428 // up. |
(...skipping 3879 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5324 MemOperand(fp, 6 * kPointerSize), | 5308 MemOperand(fp, 6 * kPointerSize), |
5325 NULL); | 5309 NULL); |
5326 } | 5310 } |
5327 | 5311 |
5328 | 5312 |
5329 #undef __ | 5313 #undef __ |
5330 | 5314 |
5331 } } // namespace v8::internal | 5315 } } // namespace v8::internal |
5332 | 5316 |
5333 #endif // V8_TARGET_ARCH_MIPS64 | 5317 #endif // V8_TARGET_ARCH_MIPS64 |
OLD | NEW |