| OLD | NEW |
| 1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #if V8_TARGET_ARCH_MIPS64 | 5 #if V8_TARGET_ARCH_MIPS64 |
| 6 | 6 |
| 7 #include "src/bootstrapper.h" | 7 #include "src/bootstrapper.h" |
| 8 #include "src/code-stubs.h" | 8 #include "src/code-stubs.h" |
| 9 #include "src/codegen.h" | 9 #include "src/codegen.h" |
| 10 #include "src/ic/handler-compiler.h" | 10 #include "src/ic/handler-compiler.h" |
| (...skipping 2508 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2519 CallStubInRecordCallTarget(masm, &weak_cell_stub, is_super); | 2519 CallStubInRecordCallTarget(masm, &weak_cell_stub, is_super); |
| 2520 __ bind(&done); | 2520 __ bind(&done); |
| 2521 } | 2521 } |
| 2522 | 2522 |
| 2523 | 2523 |
| 2524 static void EmitContinueIfStrictOrNative(MacroAssembler* masm, Label* cont) { | 2524 static void EmitContinueIfStrictOrNative(MacroAssembler* masm, Label* cont) { |
| 2525 __ ld(a3, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); | 2525 __ ld(a3, FieldMemOperand(a1, JSFunction::kSharedFunctionInfoOffset)); |
| 2526 | 2526 |
| 2527 // Do not transform the receiver for strict mode functions. | 2527 // Do not transform the receiver for strict mode functions. |
| 2528 int32_t strict_mode_function_mask = | 2528 int32_t strict_mode_function_mask = |
| 2529 1 << SharedFunctionInfo::kStrictModeBitWithinByte ; | 2529 1 << SharedFunctionInfo::kStrictModeBitWithinByte; |
| 2530 // Do not transform the receiver for native (Compilerhints already in a3). | 2530 // Do not transform the receiver for native (Compilerhints already in a3). |
| 2531 int32_t native_mask = 1 << SharedFunctionInfo::kNativeBitWithinByte; | 2531 int32_t native_mask = 1 << SharedFunctionInfo::kNativeBitWithinByte; |
| 2532 | 2532 |
| 2533 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kStrictModeByteOffset)); | 2533 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kStrictModeByteOffset)); |
| 2534 __ And(at, a4, Operand(strict_mode_function_mask)); | 2534 __ And(at, a4, Operand(strict_mode_function_mask)); |
| 2535 __ Branch(cont, ne, at, Operand(zero_reg)); | 2535 __ Branch(cont, ne, at, Operand(zero_reg)); |
| 2536 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kNativeByteOffset)); | 2536 __ lbu(a4, FieldMemOperand(a3, SharedFunctionInfo::kNativeByteOffset)); |
| 2537 __ And(at, a4, Operand(native_mask)); | 2537 __ And(at, a4, Operand(native_mask)); |
| 2538 __ Branch(cont, ne, at, Operand(zero_reg)); | 2538 __ Branch(cont, ne, at, Operand(zero_reg)); |
| 2539 } | 2539 } |
| (...skipping 3223 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 5763 MemOperand(fp, 6 * kPointerSize), NULL); | 5763 MemOperand(fp, 6 * kPointerSize), NULL); |
| 5764 } | 5764 } |
| 5765 | 5765 |
| 5766 | 5766 |
| 5767 #undef __ | 5767 #undef __ |
| 5768 | 5768 |
| 5769 } // namespace internal | 5769 } // namespace internal |
| 5770 } // namespace v8 | 5770 } // namespace v8 |
| 5771 | 5771 |
| 5772 #endif // V8_TARGET_ARCH_MIPS64 | 5772 #endif // V8_TARGET_ARCH_MIPS64 |
| OLD | NEW |