Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(568)

Side by Side Diff: src/spaces-inl.h

Issue 40083002: Make top and limit field in AllocationInfo private, assert on non-aligned setting of these fields, … (Closed) Base URL: https://v8.googlecode.com/svn/branches/bleeding_edge
Patch Set: Created 7 years, 1 month ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch | Annotate | Revision Log
« src/spaces.h ('K') | « src/spaces.cc ('k') | no next file » | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
1 // Copyright 2011 the V8 project authors. All rights reserved. 1 // Copyright 2011 the V8 project authors. All rights reserved.
2 // Redistribution and use in source and binary forms, with or without 2 // Redistribution and use in source and binary forms, with or without
3 // modification, are permitted provided that the following conditions are 3 // modification, are permitted provided that the following conditions are
4 // met: 4 // met:
5 // 5 //
6 // * Redistributions of source code must retain the above copyright 6 // * Redistributions of source code must retain the above copyright
7 // notice, this list of conditions and the following disclaimer. 7 // notice, this list of conditions and the following disclaimer.
8 // * Redistributions in binary form must reproduce the above 8 // * Redistributions in binary form must reproduce the above
9 // copyright notice, this list of conditions and the following 9 // copyright notice, this list of conditions and the following
10 // disclaimer in the documentation and/or other materials provided 10 // disclaimer in the documentation and/or other materials provided
(...skipping 246 matching lines...) Expand 10 before | Expand all | Expand 10 after
257 ASSERT(page->owner() == owner()); 257 ASSERT(page->owner() == owner());
258 set_prev_chunk(page); 258 set_prev_chunk(page);
259 } 259 }
260 260
261 261
262 // Try linear allocation in the page of alloc_info's allocation top. Does 262 // Try linear allocation in the page of alloc_info's allocation top. Does
263 // not contain slow case logic (e.g. move to the next page or try free list 263 // not contain slow case logic (e.g. move to the next page or try free list
264 // allocation) so it can be used by all the allocation functions and for all 264 // allocation) so it can be used by all the allocation functions and for all
265 // the paged spaces. 265 // the paged spaces.
266 HeapObject* PagedSpace::AllocateLinearly(int size_in_bytes) { 266 HeapObject* PagedSpace::AllocateLinearly(int size_in_bytes) {
267 Address current_top = allocation_info_.top; 267 Address current_top = allocation_info_.top();
268 Address new_top = current_top + size_in_bytes; 268 Address new_top = current_top + size_in_bytes;
269 if (new_top > allocation_info_.limit) return NULL; 269 if (new_top > allocation_info_.limit()) return NULL;
270 270
271 allocation_info_.top = new_top; 271 allocation_info_.set_top(new_top);
272 return HeapObject::FromAddress(current_top); 272 return HeapObject::FromAddress(current_top);
273 } 273 }
274 274
275 275
276 // Raw allocation. 276 // Raw allocation.
277 MaybeObject* PagedSpace::AllocateRaw(int size_in_bytes, 277 MaybeObject* PagedSpace::AllocateRaw(int size_in_bytes,
278 AllocationType event) { 278 AllocationType event) {
279 HeapProfiler* profiler = heap()->isolate()->heap_profiler(); 279 HeapProfiler* profiler = heap()->isolate()->heap_profiler();
280 280
281 HeapObject* object = AllocateLinearly(size_in_bytes); 281 HeapObject* object = AllocateLinearly(size_in_bytes);
(...skipping 35 matching lines...) Expand 10 before | Expand all | Expand 10 after
317 317
318 return Failure::RetryAfterGC(identity()); 318 return Failure::RetryAfterGC(identity());
319 } 319 }
320 320
321 321
322 // ----------------------------------------------------------------------------- 322 // -----------------------------------------------------------------------------
323 // NewSpace 323 // NewSpace
324 324
325 325
326 MaybeObject* NewSpace::AllocateRaw(int size_in_bytes) { 326 MaybeObject* NewSpace::AllocateRaw(int size_in_bytes) {
327 Address old_top = allocation_info_.top; 327 Address old_top = allocation_info_.top();
328 #ifdef DEBUG 328 #ifdef DEBUG
329 // If we are stressing compaction we waste some memory in new space 329 // If we are stressing compaction we waste some memory in new space
330 // in order to get more frequent GCs. 330 // in order to get more frequent GCs.
331 if (FLAG_stress_compaction && !heap()->linear_allocation()) { 331 if (FLAG_stress_compaction && !heap()->linear_allocation()) {
332 if (allocation_info_.limit - old_top >= size_in_bytes * 4) { 332 if (allocation_info_.limit() - old_top >= size_in_bytes * 4) {
333 int filler_size = size_in_bytes * 4; 333 int filler_size = size_in_bytes * 4;
334 for (int i = 0; i < filler_size; i += kPointerSize) { 334 for (int i = 0; i < filler_size; i += kPointerSize) {
335 *(reinterpret_cast<Object**>(old_top + i)) = 335 *(reinterpret_cast<Object**>(old_top + i)) =
336 heap()->one_pointer_filler_map(); 336 heap()->one_pointer_filler_map();
337 } 337 }
338 old_top += filler_size; 338 old_top += filler_size;
339 allocation_info_.top += filler_size; 339 Address new_top = allocation_info_.top() + filler_size;
Michael Starzinger 2013/10/25 08:34:56 nit: Only one white-space after assignment operato
Hannes Payer (out of office) 2013/11/20 09:58:47 Done.
340 allocation_info_.set_top(new_top);
Michael Starzinger 2013/10/25 08:34:56 nit: Maybe inline calculation instead of local var
Hannes Payer (out of office) 2013/11/20 09:58:47 Done.
340 } 341 }
341 } 342 }
342 #endif 343 #endif
343 344
344 if (allocation_info_.limit - old_top < size_in_bytes) { 345 if (allocation_info_.limit() - old_top < size_in_bytes) {
345 return SlowAllocateRaw(size_in_bytes); 346 return SlowAllocateRaw(size_in_bytes);
346 } 347 }
347 348
348 HeapObject* obj = HeapObject::FromAddress(old_top); 349 HeapObject* obj = HeapObject::FromAddress(old_top);
349 allocation_info_.top += size_in_bytes; 350 Address new_top = allocation_info_.top() + size_in_bytes;
351 allocation_info_.set_top(new_top);
Michael Starzinger 2013/10/25 08:34:56 nit: Maybe inline calculation instead of local var
Hannes Payer (out of office) 2013/11/20 09:58:47 Done.
350 ASSERT_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_); 352 ASSERT_SEMISPACE_ALLOCATION_INFO(allocation_info_, to_space_);
351 353
352 HeapProfiler* profiler = heap()->isolate()->heap_profiler(); 354 HeapProfiler* profiler = heap()->isolate()->heap_profiler();
353 if (profiler != NULL && profiler->is_tracking_allocations()) { 355 if (profiler != NULL && profiler->is_tracking_allocations()) {
354 profiler->NewObjectEvent(obj->address(), size_in_bytes); 356 profiler->NewObjectEvent(obj->address(), size_in_bytes);
355 } 357 }
356 358
357 return obj; 359 return obj;
358 } 360 }
359 361
(...skipping 13 matching lines...) Expand all
373 Map* map = object->map(); 375 Map* map = object->map();
374 Heap* heap = object->GetHeap(); 376 Heap* heap = object->GetHeap();
375 return map == heap->raw_unchecked_free_space_map() 377 return map == heap->raw_unchecked_free_space_map()
376 || map == heap->raw_unchecked_one_pointer_filler_map() 378 || map == heap->raw_unchecked_one_pointer_filler_map()
377 || map == heap->raw_unchecked_two_pointer_filler_map(); 379 || map == heap->raw_unchecked_two_pointer_filler_map();
378 } 380 }
379 381
380 } } // namespace v8::internal 382 } } // namespace v8::internal
381 383
382 #endif // V8_SPACES_INL_H_ 384 #endif // V8_SPACES_INL_H_
OLDNEW
« src/spaces.h ('K') | « src/spaces.cc ('k') | no next file » | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698