Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(217)

Side by Side Diff: src/x64/code-stubs-x64.cc

Issue 11108011: Ensure that we do not accidentally emit debug code. (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: Created 8 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch | Annotate | Revision Log
OLDNEW
1 // Copyright 2012 the V8 project authors. All rights reserved. 1 // Copyright 2012 the V8 project authors. All rights reserved.
2 // Redistribution and use in source and binary forms, with or without 2 // Redistribution and use in source and binary forms, with or without
3 // modification, are permitted provided that the following conditions are 3 // modification, are permitted provided that the following conditions are
4 // met: 4 // met:
5 // 5 //
6 // * Redistributions of source code must retain the above copyright 6 // * Redistributions of source code must retain the above copyright
7 // notice, this list of conditions and the following disclaimer. 7 // notice, this list of conditions and the following disclaimer.
8 // * Redistributions in binary form must reproduce the above 8 // * Redistributions in binary form must reproduce the above
9 // copyright notice, this list of conditions and the following 9 // copyright notice, this list of conditions and the following
10 // disclaimer in the documentation and/or other materials provided 10 // disclaimer in the documentation and/or other materials provided
(...skipping 1279 matching lines...) Expand 10 before | Expand all | Expand 10 after
1290 // Allocate heap number in new space. 1290 // Allocate heap number in new space.
1291 // Not using AllocateHeapNumber macro in order to reuse 1291 // Not using AllocateHeapNumber macro in order to reuse
1292 // already loaded heap_number_map. 1292 // already loaded heap_number_map.
1293 __ AllocateInNewSpace(HeapNumber::kSize, 1293 __ AllocateInNewSpace(HeapNumber::kSize,
1294 rax, 1294 rax,
1295 rdx, 1295 rdx,
1296 no_reg, 1296 no_reg,
1297 &allocation_failed, 1297 &allocation_failed,
1298 TAG_OBJECT); 1298 TAG_OBJECT);
1299 // Set the map. 1299 // Set the map.
1300 if (FLAG_debug_code) { 1300 __ AssertRootValue(heap_number_map,
Michael Starzinger 2012/10/12 11:00:22 Indentation is off.
Sven Panne 2012/10/12 11:05:36 Done.
1301 __ AbortIfNotRootValue(heap_number_map, 1301 Heap::kHeapNumberMapRootIndex,
1302 Heap::kHeapNumberMapRootIndex, 1302 "HeapNumberMap register clobbered.");
1303 "HeapNumberMap register clobbered.");
1304 }
1305 __ movq(FieldOperand(rax, HeapObject::kMapOffset), 1303 __ movq(FieldOperand(rax, HeapObject::kMapOffset),
1306 heap_number_map); 1304 heap_number_map);
1307 __ cvtqsi2sd(xmm0, rbx); 1305 __ cvtqsi2sd(xmm0, rbx);
1308 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm0); 1306 __ movsd(FieldOperand(rax, HeapNumber::kValueOffset), xmm0);
1309 __ Ret(); 1307 __ Ret();
1310 1308
1311 __ bind(&allocation_failed); 1309 __ bind(&allocation_failed);
1312 // We need tagged values in rdx and rax for the following code, 1310 // We need tagged values in rdx and rax for the following code,
1313 // not int32 in rax and rcx. 1311 // not int32 in rax and rcx.
1314 __ Integer32ToSmi(rax, rcx); 1312 __ Integer32ToSmi(rax, rcx);
(...skipping 727 matching lines...) Expand 10 before | Expand all | Expand 10 after
2042 // Check if conversion was successful by converting back and 2040 // Check if conversion was successful by converting back and
2043 // comparing to the original double's bits. 2041 // comparing to the original double's bits.
2044 __ cvtlsi2sd(xmm1, smi_result); 2042 __ cvtlsi2sd(xmm1, smi_result);
2045 __ movq(kScratchRegister, xmm1); 2043 __ movq(kScratchRegister, xmm1);
2046 __ cmpq(scratch2, kScratchRegister); 2044 __ cmpq(scratch2, kScratchRegister);
2047 __ j(not_equal, on_not_smis); 2045 __ j(not_equal, on_not_smis);
2048 __ Integer32ToSmi(first, smi_result); 2046 __ Integer32ToSmi(first, smi_result);
2049 2047
2050 __ JumpIfSmi(second, (on_success != NULL) ? on_success : &done); 2048 __ JumpIfSmi(second, (on_success != NULL) ? on_success : &done);
2051 __ bind(&first_smi); 2049 __ bind(&first_smi);
2052 if (FLAG_debug_code) { 2050 __ AssertNotSmi(second);
2053 // Second should be non-smi if we get here.
2054 __ AbortIfSmi(second);
2055 }
2056 __ cmpq(FieldOperand(second, HeapObject::kMapOffset), heap_number_map); 2051 __ cmpq(FieldOperand(second, HeapObject::kMapOffset), heap_number_map);
2057 __ j(not_equal, on_not_smis); 2052 __ j(not_equal, on_not_smis);
2058 // Convert second to smi, if possible. 2053 // Convert second to smi, if possible.
2059 __ movsd(xmm0, FieldOperand(second, HeapNumber::kValueOffset)); 2054 __ movsd(xmm0, FieldOperand(second, HeapNumber::kValueOffset));
2060 __ movq(scratch2, xmm0); 2055 __ movq(scratch2, xmm0);
2061 __ cvttsd2siq(smi_result, xmm0); 2056 __ cvttsd2siq(smi_result, xmm0);
2062 __ cvtlsi2sd(xmm1, smi_result); 2057 __ cvtlsi2sd(xmm1, smi_result);
2063 __ movq(kScratchRegister, xmm1); 2058 __ movq(kScratchRegister, xmm1);
2064 __ cmpq(scratch2, kScratchRegister); 2059 __ cmpq(scratch2, kScratchRegister);
2065 __ j(not_equal, on_not_smis); 2060 __ j(not_equal, on_not_smis);
(...skipping 3856 matching lines...) Expand 10 before | Expand all | Expand 10 after
5922 Label* done, 5917 Label* done,
5923 Register elements, 5918 Register elements,
5924 Register name, 5919 Register name,
5925 Register r0, 5920 Register r0,
5926 Register r1) { 5921 Register r1) {
5927 ASSERT(!elements.is(r0)); 5922 ASSERT(!elements.is(r0));
5928 ASSERT(!elements.is(r1)); 5923 ASSERT(!elements.is(r1));
5929 ASSERT(!name.is(r0)); 5924 ASSERT(!name.is(r0));
5930 ASSERT(!name.is(r1)); 5925 ASSERT(!name.is(r1));
5931 5926
5932 // Assert that name contains a string. 5927 __ AssertString(name);
5933 if (FLAG_debug_code) __ AbortIfNotString(name);
5934 5928
5935 __ SmiToInteger32(r0, FieldOperand(elements, kCapacityOffset)); 5929 __ SmiToInteger32(r0, FieldOperand(elements, kCapacityOffset));
5936 __ decl(r0); 5930 __ decl(r0);
5937 5931
5938 for (int i = 0; i < kInlinedProbes; i++) { 5932 for (int i = 0; i < kInlinedProbes; i++) {
5939 // Compute the masked index: (hash + i + i * i) & mask. 5933 // Compute the masked index: (hash + i + i * i) & mask.
5940 __ movl(r1, FieldOperand(name, String::kHashFieldOffset)); 5934 __ movl(r1, FieldOperand(name, String::kHashFieldOffset));
5941 __ shrl(r1, Immediate(String::kHashShift)); 5935 __ shrl(r1, Immediate(String::kHashShift));
5942 if (i > 0) { 5936 if (i > 0) {
5943 __ addl(r1, Immediate(StringDictionary::GetProbeOffset(i))); 5937 __ addl(r1, Immediate(StringDictionary::GetProbeOffset(i)));
(...skipping 561 matching lines...) Expand 10 before | Expand all | Expand 10 after
6505 #endif 6499 #endif
6506 6500
6507 __ Ret(); 6501 __ Ret();
6508 } 6502 }
6509 6503
6510 #undef __ 6504 #undef __
6511 6505
6512 } } // namespace v8::internal 6506 } } // namespace v8::internal
6513 6507
6514 #endif // V8_TARGET_ARCH_X64 6508 #endif // V8_TARGET_ARCH_X64
OLDNEW

Powered by Google App Engine
This is Rietveld 408576698