| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #if V8_TARGET_ARCH_MIPS | 5 #if V8_TARGET_ARCH_MIPS |
| 6 | 6 |
| 7 #include "src/code-stubs.h" | 7 #include "src/code-stubs.h" |
| 8 #include "src/api-arguments.h" | 8 #include "src/api-arguments.h" |
| 9 #include "src/base/bits.h" | 9 #include "src/base/bits.h" |
| 10 #include "src/bootstrapper.h" | 10 #include "src/bootstrapper.h" |
| (...skipping 3248 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 3259 __ Ret(USE_DELAY_SLOT); | 3259 __ Ret(USE_DELAY_SLOT); |
| 3260 __ Addu(sp, sp, a1); | 3260 __ Addu(sp, sp, a1); |
| 3261 } | 3261 } |
| 3262 | 3262 |
| 3263 void CallICTrampolineStub::Generate(MacroAssembler* masm) { | 3263 void CallICTrampolineStub::Generate(MacroAssembler* masm) { |
| 3264 __ EmitLoadTypeFeedbackVector(a2); | 3264 __ EmitLoadTypeFeedbackVector(a2); |
| 3265 CallICStub stub(isolate(), state()); | 3265 CallICStub stub(isolate(), state()); |
| 3266 __ Jump(stub.GetCode(), RelocInfo::CODE_TARGET); | 3266 __ Jump(stub.GetCode(), RelocInfo::CODE_TARGET); |
| 3267 } | 3267 } |
| 3268 | 3268 |
| 3269 | |
| 3270 static void HandleArrayCases(MacroAssembler* masm, Register feedback, | |
| 3271 Register receiver_map, Register scratch1, | |
| 3272 Register scratch2, bool is_polymorphic, | |
| 3273 Label* miss) { | |
| 3274 // feedback initially contains the feedback array | |
| 3275 Label next_loop, prepare_next; | |
| 3276 Label start_polymorphic; | |
| 3277 | |
| 3278 Register cached_map = scratch1; | |
| 3279 | |
| 3280 __ lw(cached_map, | |
| 3281 FieldMemOperand(feedback, FixedArray::OffsetOfElementAt(0))); | |
| 3282 __ lw(cached_map, FieldMemOperand(cached_map, WeakCell::kValueOffset)); | |
| 3283 __ Branch(&start_polymorphic, ne, receiver_map, Operand(cached_map)); | |
| 3284 // found, now call handler. | |
| 3285 Register handler = feedback; | |
| 3286 __ lw(handler, FieldMemOperand(feedback, FixedArray::OffsetOfElementAt(1))); | |
| 3287 __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); | |
| 3288 __ Jump(t9); | |
| 3289 | |
| 3290 | |
| 3291 Register length = scratch2; | |
| 3292 __ bind(&start_polymorphic); | |
| 3293 __ lw(length, FieldMemOperand(feedback, FixedArray::kLengthOffset)); | |
| 3294 if (!is_polymorphic) { | |
| 3295 // If the IC could be monomorphic we have to make sure we don't go past the | |
| 3296 // end of the feedback array. | |
| 3297 __ Branch(miss, eq, length, Operand(Smi::FromInt(2))); | |
| 3298 } | |
| 3299 | |
| 3300 Register too_far = length; | |
| 3301 Register pointer_reg = feedback; | |
| 3302 | |
| 3303 // +-----+------+------+-----+-----+ ... ----+ | |
| 3304 // | map | len | wm0 | h0 | wm1 | hN | | |
| 3305 // +-----+------+------+-----+-----+ ... ----+ | |
| 3306 // 0 1 2 len-1 | |
| 3307 // ^ ^ | |
| 3308 // | | | |
| 3309 // pointer_reg too_far | |
| 3310 // aka feedback scratch2 | |
| 3311 // also need receiver_map | |
| 3312 // use cached_map (scratch1) to look in the weak map values. | |
| 3313 __ Lsa(too_far, feedback, length, kPointerSizeLog2 - kSmiTagSize); | |
| 3314 __ Addu(too_far, too_far, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); | |
| 3315 __ Addu(pointer_reg, feedback, | |
| 3316 Operand(FixedArray::OffsetOfElementAt(2) - kHeapObjectTag)); | |
| 3317 | |
| 3318 __ bind(&next_loop); | |
| 3319 __ lw(cached_map, MemOperand(pointer_reg)); | |
| 3320 __ lw(cached_map, FieldMemOperand(cached_map, WeakCell::kValueOffset)); | |
| 3321 __ Branch(&prepare_next, ne, receiver_map, Operand(cached_map)); | |
| 3322 __ lw(handler, MemOperand(pointer_reg, kPointerSize)); | |
| 3323 __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); | |
| 3324 __ Jump(t9); | |
| 3325 | |
| 3326 __ bind(&prepare_next); | |
| 3327 __ Addu(pointer_reg, pointer_reg, Operand(kPointerSize * 2)); | |
| 3328 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); | |
| 3329 | |
| 3330 // We exhausted our array of map handler pairs. | |
| 3331 __ jmp(miss); | |
| 3332 } | |
| 3333 | |
| 3334 | |
| 3335 static void HandleMonomorphicCase(MacroAssembler* masm, Register receiver, | |
| 3336 Register receiver_map, Register feedback, | |
| 3337 Register vector, Register slot, | |
| 3338 Register scratch, Label* compare_map, | |
| 3339 Label* load_smi_map, Label* try_array) { | |
| 3340 __ JumpIfSmi(receiver, load_smi_map); | |
| 3341 __ lw(receiver_map, FieldMemOperand(receiver, HeapObject::kMapOffset)); | |
| 3342 __ bind(compare_map); | |
| 3343 Register cached_map = scratch; | |
| 3344 // Move the weak map into the weak_cell register. | |
| 3345 __ lw(cached_map, FieldMemOperand(feedback, WeakCell::kValueOffset)); | |
| 3346 __ Branch(try_array, ne, cached_map, Operand(receiver_map)); | |
| 3347 Register handler = feedback; | |
| 3348 | |
| 3349 __ Lsa(handler, vector, slot, kPointerSizeLog2 - kSmiTagSize); | |
| 3350 __ lw(handler, | |
| 3351 FieldMemOperand(handler, FixedArray::kHeaderSize + kPointerSize)); | |
| 3352 __ Addu(t9, handler, Operand(Code::kHeaderSize - kHeapObjectTag)); | |
| 3353 __ Jump(t9); | |
| 3354 } | |
| 3355 | |
| 3356 void KeyedStoreICTrampolineStub::Generate(MacroAssembler* masm) { | |
| 3357 __ EmitLoadTypeFeedbackVector(StoreWithVectorDescriptor::VectorRegister()); | |
| 3358 KeyedStoreICStub stub(isolate(), state()); | |
| 3359 stub.GenerateForTrampoline(masm); | |
| 3360 } | |
| 3361 | |
| 3362 void KeyedStoreICStub::Generate(MacroAssembler* masm) { | |
| 3363 GenerateImpl(masm, false); | |
| 3364 } | |
| 3365 | |
| 3366 void KeyedStoreICStub::GenerateForTrampoline(MacroAssembler* masm) { | |
| 3367 GenerateImpl(masm, true); | |
| 3368 } | |
| 3369 | |
| 3370 | |
| 3371 static void HandlePolymorphicStoreCase(MacroAssembler* masm, Register feedback, | |
| 3372 Register receiver_map, Register scratch1, | |
| 3373 Register scratch2, Label* miss) { | |
| 3374 // feedback initially contains the feedback array | |
| 3375 Label next_loop, prepare_next; | |
| 3376 Label start_polymorphic; | |
| 3377 Label transition_call; | |
| 3378 | |
| 3379 Register cached_map = scratch1; | |
| 3380 Register too_far = scratch2; | |
| 3381 Register pointer_reg = feedback; | |
| 3382 __ lw(too_far, FieldMemOperand(feedback, FixedArray::kLengthOffset)); | |
| 3383 | |
| 3384 // +-----+------+------+-----+-----+-----+ ... ----+ | |
| 3385 // | map | len | wm0 | wt0 | h0 | wm1 | hN | | |
| 3386 // +-----+------+------+-----+-----+ ----+ ... ----+ | |
| 3387 // 0 1 2 len-1 | |
| 3388 // ^ ^ | |
| 3389 // | | | |
| 3390 // pointer_reg too_far | |
| 3391 // aka feedback scratch2 | |
| 3392 // also need receiver_map | |
| 3393 // use cached_map (scratch1) to look in the weak map values. | |
| 3394 __ Lsa(too_far, feedback, too_far, kPointerSizeLog2 - kSmiTagSize); | |
| 3395 __ Addu(too_far, too_far, Operand(FixedArray::kHeaderSize - kHeapObjectTag)); | |
| 3396 __ Addu(pointer_reg, feedback, | |
| 3397 Operand(FixedArray::OffsetOfElementAt(0) - kHeapObjectTag)); | |
| 3398 | |
| 3399 __ bind(&next_loop); | |
| 3400 __ lw(cached_map, MemOperand(pointer_reg)); | |
| 3401 __ lw(cached_map, FieldMemOperand(cached_map, WeakCell::kValueOffset)); | |
| 3402 __ Branch(&prepare_next, ne, receiver_map, Operand(cached_map)); | |
| 3403 // Is it a transitioning store? | |
| 3404 __ lw(too_far, MemOperand(pointer_reg, kPointerSize)); | |
| 3405 __ LoadRoot(at, Heap::kUndefinedValueRootIndex); | |
| 3406 __ Branch(&transition_call, ne, too_far, Operand(at)); | |
| 3407 __ lw(pointer_reg, MemOperand(pointer_reg, kPointerSize * 2)); | |
| 3408 __ Addu(t9, pointer_reg, Operand(Code::kHeaderSize - kHeapObjectTag)); | |
| 3409 __ Jump(t9); | |
| 3410 | |
| 3411 __ bind(&transition_call); | |
| 3412 __ lw(too_far, FieldMemOperand(too_far, WeakCell::kValueOffset)); | |
| 3413 __ JumpIfSmi(too_far, miss); | |
| 3414 | |
| 3415 __ lw(receiver_map, MemOperand(pointer_reg, kPointerSize * 2)); | |
| 3416 | |
| 3417 // Load the map into the correct register. | |
| 3418 DCHECK(feedback.is(StoreTransitionDescriptor::MapRegister())); | |
| 3419 __ mov(feedback, too_far); | |
| 3420 | |
| 3421 __ Addu(t9, receiver_map, Operand(Code::kHeaderSize - kHeapObjectTag)); | |
| 3422 __ Jump(t9); | |
| 3423 | |
| 3424 __ bind(&prepare_next); | |
| 3425 __ Addu(pointer_reg, pointer_reg, Operand(kPointerSize * 3)); | |
| 3426 __ Branch(&next_loop, lt, pointer_reg, Operand(too_far)); | |
| 3427 | |
| 3428 // We exhausted our array of map handler pairs. | |
| 3429 __ jmp(miss); | |
| 3430 } | |
| 3431 | |
| 3432 void KeyedStoreICStub::GenerateImpl(MacroAssembler* masm, bool in_frame) { | |
| 3433 Register receiver = StoreWithVectorDescriptor::ReceiverRegister(); // a1 | |
| 3434 Register key = StoreWithVectorDescriptor::NameRegister(); // a2 | |
| 3435 Register vector = StoreWithVectorDescriptor::VectorRegister(); // a3 | |
| 3436 Register slot = StoreWithVectorDescriptor::SlotRegister(); // t0 | |
| 3437 DCHECK(StoreWithVectorDescriptor::ValueRegister().is(a0)); // a0 | |
| 3438 Register feedback = t1; | |
| 3439 Register receiver_map = t2; | |
| 3440 Register scratch1 = t5; | |
| 3441 | |
| 3442 __ Lsa(feedback, vector, slot, kPointerSizeLog2 - kSmiTagSize); | |
| 3443 __ lw(feedback, FieldMemOperand(feedback, FixedArray::kHeaderSize)); | |
| 3444 | |
| 3445 // Try to quickly handle the monomorphic case without knowing for sure | |
| 3446 // if we have a weak cell in feedback. We do know it's safe to look | |
| 3447 // at WeakCell::kValueOffset. | |
| 3448 Label try_array, load_smi_map, compare_map; | |
| 3449 Label not_array, miss; | |
| 3450 HandleMonomorphicCase(masm, receiver, receiver_map, feedback, vector, slot, | |
| 3451 scratch1, &compare_map, &load_smi_map, &try_array); | |
| 3452 | |
| 3453 __ bind(&try_array); | |
| 3454 // Is it a fixed array? | |
| 3455 __ lw(scratch1, FieldMemOperand(feedback, HeapObject::kMapOffset)); | |
| 3456 __ LoadRoot(at, Heap::kFixedArrayMapRootIndex); | |
| 3457 __ Branch(¬_array, ne, scratch1, Operand(at)); | |
| 3458 | |
| 3459 // We have a polymorphic element handler. | |
| 3460 Label polymorphic, try_poly_name; | |
| 3461 __ bind(&polymorphic); | |
| 3462 | |
| 3463 Register scratch2 = t4; | |
| 3464 | |
| 3465 HandlePolymorphicStoreCase(masm, feedback, receiver_map, scratch1, scratch2, | |
| 3466 &miss); | |
| 3467 | |
| 3468 __ bind(¬_array); | |
| 3469 // Is it generic? | |
| 3470 __ LoadRoot(at, Heap::kmegamorphic_symbolRootIndex); | |
| 3471 __ Branch(&try_poly_name, ne, feedback, Operand(at)); | |
| 3472 Handle<Code> megamorphic_stub = | |
| 3473 KeyedStoreIC::ChooseMegamorphicStub(masm->isolate(), GetExtraICState()); | |
| 3474 __ Jump(megamorphic_stub, RelocInfo::CODE_TARGET); | |
| 3475 | |
| 3476 __ bind(&try_poly_name); | |
| 3477 // We might have a name in feedback, and a fixed array in the next slot. | |
| 3478 __ Branch(&miss, ne, key, Operand(feedback)); | |
| 3479 // If the name comparison succeeded, we know we have a fixed array with | |
| 3480 // at least one map/handler pair. | |
| 3481 __ Lsa(feedback, vector, slot, kPointerSizeLog2 - kSmiTagSize); | |
| 3482 __ lw(feedback, | |
| 3483 FieldMemOperand(feedback, FixedArray::kHeaderSize + kPointerSize)); | |
| 3484 HandleArrayCases(masm, feedback, receiver_map, scratch1, scratch2, false, | |
| 3485 &miss); | |
| 3486 | |
| 3487 __ bind(&miss); | |
| 3488 KeyedStoreIC::GenerateMiss(masm); | |
| 3489 | |
| 3490 __ bind(&load_smi_map); | |
| 3491 __ Branch(USE_DELAY_SLOT, &compare_map); | |
| 3492 __ LoadRoot(receiver_map, Heap::kHeapNumberMapRootIndex); // In delay slot. | |
| 3493 } | |
| 3494 | |
| 3495 | |
| 3496 void ProfileEntryHookStub::MaybeCallEntryHook(MacroAssembler* masm) { | 3269 void ProfileEntryHookStub::MaybeCallEntryHook(MacroAssembler* masm) { |
| 3497 if (masm->isolate()->function_entry_hook() != NULL) { | 3270 if (masm->isolate()->function_entry_hook() != NULL) { |
| 3498 ProfileEntryHookStub stub(masm->isolate()); | 3271 ProfileEntryHookStub stub(masm->isolate()); |
| 3499 __ push(ra); | 3272 __ push(ra); |
| 3500 __ CallStub(&stub); | 3273 __ CallStub(&stub); |
| 3501 __ pop(ra); | 3274 __ pop(ra); |
| 3502 } | 3275 } |
| 3503 } | 3276 } |
| 3504 | 3277 |
| 3505 | 3278 |
| (...skipping 1260 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 4766 kStackUnwindSpace, kInvalidStackOffset, | 4539 kStackUnwindSpace, kInvalidStackOffset, |
| 4767 return_value_operand, NULL); | 4540 return_value_operand, NULL); |
| 4768 } | 4541 } |
| 4769 | 4542 |
| 4770 #undef __ | 4543 #undef __ |
| 4771 | 4544 |
| 4772 } // namespace internal | 4545 } // namespace internal |
| 4773 } // namespace v8 | 4546 } // namespace v8 |
| 4774 | 4547 |
| 4775 #endif // V8_TARGET_ARCH_MIPS | 4548 #endif // V8_TARGET_ARCH_MIPS |
| OLD | NEW |