OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #if V8_TARGET_ARCH_MIPS64 | 5 #if V8_TARGET_ARCH_MIPS64 |
6 | 6 |
7 #include "src/bootstrapper.h" | 7 #include "src/bootstrapper.h" |
8 #include "src/code-stubs.h" | 8 #include "src/code-stubs.h" |
9 #include "src/codegen.h" | 9 #include "src/codegen.h" |
10 #include "src/ic/handler-compiler.h" | 10 #include "src/ic/handler-compiler.h" |
(...skipping 2508 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2519 CallStubInRecordCallTarget(masm, &weak_cell_stub, is_super); | 2519 CallStubInRecordCallTarget(masm, &weak_cell_stub, is_super); |
2520 __ bind(&done); | 2520 __ bind(&done); |
2521 } | 2521 } |
2522 | 2522 |
2523 | 2523 |
2524 static void EmitContinueIfStrictOrNative(MacroAssembler* masm, Label* cont) { | 2524 static void EmitContinueIfStrictOrNative(MacroAssembler* masm, Label* cont) { |
2525 __ ld(a3, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); | 2525 __ ld(a3, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); |
2526 | 2526 |
2527 // Do not transform the receiver for strict mode functions. | 2527 // Do not transform the receiver for strict mode functions. |
2528 int32_t strict_mode_function_mask = | 2528 int32_t strict_mode_function_mask = |
2529 1 << SharedFunctionInfo::kStrictModeBitWithinByte ; | 2529 1 << SharedFunctionInfo::kStrictModeBitWithinByte; |
2530 // Do not transform the receiver for native (Compilerhints already in a3). | 2530 // Do not transform the receiver for native (Compilerhints already in a3). |
2531 int32_t native_mask = 1 << SharedFunctionInfo::kNativeBitWithinByte; | 2531 int32_t native_mask = 1 << SharedFunctionInfo::kNativeBitWithinByte; |
2532 | 2532 |
2533 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kStrictModeByteOffset)); | 2533 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kStrictModeByteOffset)); |
2534 __ And(at, a4, Operand(strict_mode_function_mask)); | 2534 __ And(at, a4, Operand(strict_mode_function_mask)); |
2535 __ Branch(cont, ne, at, Operand(zero_reg)); | 2535 __ Branch(cont, ne, at, Operand(zero_reg)); |
2536 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kNativeByteOffset)); | 2536 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kNativeByteOffset)); |
2537 __ And(at, a4, Operand(native_mask)); | 2537 __ And(at, a4, Operand(native_mask)); |
2538 __ Branch(cont, ne, at, Operand(zero_reg)); | 2538 __ Branch(cont, ne, at, Operand(zero_reg)); |
2539 } | 2539 } |
(...skipping 3223 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5763 MemOperand(fp, 6 * kPointerSize), NULL); | 5763 MemOperand(fp, 6 * kPointerSize), NULL); |
5764 } | 5764 } |
5765 | 5765 |
5766 | 5766 |
5767 #undef __ | 5767 #undef __ |
5768 | 5768 |
5769 } // namespace internal | 5769 } // namespace internal |
5770 } // namespace v8 | 5770 } // namespace v8 |
5771 | 5771 |
5772 #endif // V8_TARGET_ARCH_MIPS64 | 5772 #endif // V8_TARGET_ARCH_MIPS64 |
OLD | NEW |