| OLD | NEW |
| 1 // Copyright 2015 The Chromium Authors. All rights reserved. | 1 // Copyright 2015 The Chromium Authors. All rights reserved. |
| 2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
| 3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
| 4 | 4 |
| 5 #include "base/trace_event/heap_profiler_allocation_context_tracker.h" | 5 #include "base/trace_event/heap_profiler_allocation_context_tracker.h" |
| 6 | 6 |
| 7 #include <algorithm> | 7 #include <algorithm> |
| 8 #include <iterator> | 8 #include <iterator> |
| 9 | 9 |
| 10 #include "base/atomicops.h" | 10 #include "base/atomicops.h" |
| 11 #include "base/threading/thread_local_storage.h" | 11 #include "base/threading/thread_local_storage.h" |
| 12 #include "base/trace_event/heap_profiler_allocation_context.h" | 12 #include "base/trace_event/heap_profiler_allocation_context.h" |
| 13 | 13 |
| 14 namespace base { | 14 namespace base { |
| 15 namespace trace_event { | 15 namespace trace_event { |
| 16 | 16 |
| 17 subtle::Atomic32 AllocationContextTracker::capture_enabled_ = 0; | 17 subtle::Atomic32 AllocationContextTracker::capture_enabled_ = 0; |
| 18 | 18 |
| 19 namespace { | 19 namespace { |
| 20 | 20 |
| 21 const size_t kMaxStackDepth = 128u; | |
| 22 AllocationContextTracker* const kInitializingSentinel = | |
| 23 reinterpret_cast<AllocationContextTracker*>(-1); | |
| 24 | |
| 25 ThreadLocalStorage::StaticSlot g_tls_alloc_ctx_tracker = TLS_INITIALIZER; | 21 ThreadLocalStorage::StaticSlot g_tls_alloc_ctx_tracker = TLS_INITIALIZER; |
| 26 | 22 |
| 27 // This function is added to the TLS slot to clean up the instance when the | 23 // This function is added to the TLS slot to clean up the instance when the |
| 28 // thread exits. | 24 // thread exits. |
| 29 void DestructAllocationContextTracker(void* alloc_ctx_tracker) { | 25 void DestructAllocationContextTracker(void* alloc_ctx_tracker) { |
| 30 delete static_cast<AllocationContextTracker*>(alloc_ctx_tracker); | 26 delete static_cast<AllocationContextTracker*>(alloc_ctx_tracker); |
| 31 } | 27 } |
| 32 | 28 |
| 33 } // namespace | 29 } // namespace |
| 34 | 30 |
| 31 AllocationContextTracker::AllocationContextTracker() {} |
| 32 AllocationContextTracker::~AllocationContextTracker() {} |
| 33 |
| 35 // static | 34 // static |
| 36 AllocationContextTracker* | 35 AllocationContextTracker* AllocationContextTracker::GetThreadLocalTracker() { |
| 37 AllocationContextTracker::GetInstanceForCurrentThread() { | 36 auto tracker = |
| 38 AllocationContextTracker* tracker = | |
| 39 static_cast<AllocationContextTracker*>(g_tls_alloc_ctx_tracker.Get()); | 37 static_cast<AllocationContextTracker*>(g_tls_alloc_ctx_tracker.Get()); |
| 40 if (tracker == kInitializingSentinel) | |
| 41 return nullptr; // Re-entrancy case. | |
| 42 | 38 |
| 43 if (!tracker) { | 39 if (!tracker) { |
| 44 g_tls_alloc_ctx_tracker.Set(kInitializingSentinel); | |
| 45 tracker = new AllocationContextTracker(); | 40 tracker = new AllocationContextTracker(); |
| 46 g_tls_alloc_ctx_tracker.Set(tracker); | 41 g_tls_alloc_ctx_tracker.Set(tracker); |
| 47 } | 42 } |
| 48 | 43 |
| 49 return tracker; | 44 return tracker; |
| 50 } | 45 } |
| 51 | 46 |
| 52 AllocationContextTracker::AllocationContextTracker() { | |
| 53 pseudo_stack_.reserve(kMaxStackDepth); | |
| 54 } | |
| 55 AllocationContextTracker::~AllocationContextTracker() {} | |
| 56 | |
| 57 // static | 47 // static |
| 58 void AllocationContextTracker::SetCaptureEnabled(bool enabled) { | 48 void AllocationContextTracker::SetCaptureEnabled(bool enabled) { |
| 59 // When enabling capturing, also initialize the TLS slot. This does not create | 49 // When enabling capturing, also initialize the TLS slot. This does not create |
| 60 // a TLS instance yet. | 50 // a TLS instance yet. |
| 61 if (enabled && !g_tls_alloc_ctx_tracker.initialized()) | 51 if (enabled && !g_tls_alloc_ctx_tracker.initialized()) |
| 62 g_tls_alloc_ctx_tracker.Initialize(DestructAllocationContextTracker); | 52 g_tls_alloc_ctx_tracker.Initialize(DestructAllocationContextTracker); |
| 63 | 53 |
| 64 // Release ordering ensures that when a thread observes |capture_enabled_| to | 54 // Release ordering ensures that when a thread observes |capture_enabled_| to |
| 65 // be true through an acquire load, the TLS slot has been initialized. | 55 // be true through an acquire load, the TLS slot has been initialized. |
| 66 subtle::Release_Store(&capture_enabled_, enabled); | 56 subtle::Release_Store(&capture_enabled_, enabled); |
| 67 } | 57 } |
| 68 | 58 |
| 59 // static |
| 69 void AllocationContextTracker::PushPseudoStackFrame(StackFrame frame) { | 60 void AllocationContextTracker::PushPseudoStackFrame(StackFrame frame) { |
| 61 auto tracker = AllocationContextTracker::GetThreadLocalTracker(); |
| 62 |
| 70 // Impose a limit on the height to verify that every push is popped, because | 63 // Impose a limit on the height to verify that every push is popped, because |
| 71 // in practice the pseudo stack never grows higher than ~20 frames. | 64 // in practice the pseudo stack never grows higher than ~20 frames. |
| 72 if (pseudo_stack_.size() < kMaxStackDepth) | 65 DCHECK_LT(tracker->pseudo_stack_.size(), 128u); |
| 73 pseudo_stack_.push_back(frame); | 66 tracker->pseudo_stack_.push_back(frame); |
| 74 else | |
| 75 NOTREACHED(); | |
| 76 } | 67 } |
| 77 | 68 |
| 78 // static | 69 // static |
| 79 void AllocationContextTracker::PopPseudoStackFrame(StackFrame frame) { | 70 void AllocationContextTracker::PopPseudoStackFrame(StackFrame frame) { |
| 71 auto tracker = AllocationContextTracker::GetThreadLocalTracker(); |
| 72 |
| 80 // Guard for stack underflow. If tracing was started with a TRACE_EVENT in | 73 // Guard for stack underflow. If tracing was started with a TRACE_EVENT in |
| 81 // scope, the frame was never pushed, so it is possible that pop is called | 74 // scope, the frame was never pushed, so it is possible that pop is called |
| 82 // on an empty stack. | 75 // on an empty stack. |
| 83 if (pseudo_stack_.empty()) | 76 if (tracker->pseudo_stack_.empty()) |
| 84 return; | 77 return; |
| 85 | 78 |
| 86 // Assert that pushes and pops are nested correctly. This DCHECK can be | 79 // Assert that pushes and pops are nested correctly. This DCHECK can be |
| 87 // hit if some TRACE_EVENT macro is unbalanced (a TRACE_EVENT_END* call | 80 // hit if some TRACE_EVENT macro is unbalanced (a TRACE_EVENT_END* call |
| 88 // without a corresponding TRACE_EVENT_BEGIN). | 81 // without a corresponding TRACE_EVENT_BEGIN). |
| 89 DCHECK_EQ(frame, pseudo_stack_.back()) | 82 DCHECK_EQ(frame, tracker->pseudo_stack_.back()) |
| 90 << "Encountered an unmatched TRACE_EVENT_END"; | 83 << "Encountered an unmatched TRACE_EVENT_END"; |
| 91 | 84 |
| 92 pseudo_stack_.pop_back(); | 85 tracker->pseudo_stack_.pop_back(); |
| 93 } | 86 } |
| 94 | 87 |
| 95 // static | 88 // static |
| 96 AllocationContext AllocationContextTracker::GetContextSnapshot() { | 89 AllocationContext AllocationContextTracker::GetContextSnapshot() { |
| 90 AllocationContextTracker* tracker = GetThreadLocalTracker(); |
| 97 AllocationContext ctx; | 91 AllocationContext ctx; |
| 98 | 92 |
| 99 // Fill the backtrace. | 93 // Fill the backtrace. |
| 100 { | 94 { |
| 101 auto src = pseudo_stack_.begin(); | 95 auto src = tracker->pseudo_stack_.begin(); |
| 102 auto dst = std::begin(ctx.backtrace.frames); | 96 auto dst = std::begin(ctx.backtrace.frames); |
| 103 auto src_end = pseudo_stack_.end(); | 97 auto src_end = tracker->pseudo_stack_.end(); |
| 104 auto dst_end = std::end(ctx.backtrace.frames); | 98 auto dst_end = std::end(ctx.backtrace.frames); |
| 105 | 99 |
| 106 // Copy as much of the bottom of the pseudo stack into the backtrace as | 100 // Copy as much of the bottom of the pseudo stack into the backtrace as |
| 107 // possible. | 101 // possible. |
| 108 for (; src != src_end && dst != dst_end; src++, dst++) | 102 for (; src != src_end && dst != dst_end; src++, dst++) |
| 109 *dst = *src; | 103 *dst = *src; |
| 110 | 104 |
| 111 // If there is room for more, fill the remaining slots with empty frames. | 105 // If there is room for more, fill the remaining slots with empty frames. |
| 112 std::fill(dst, dst_end, nullptr); | 106 std::fill(dst, dst_end, nullptr); |
| 113 } | 107 } |
| 114 | 108 |
| 115 ctx.type_name = nullptr; | 109 ctx.type_name = nullptr; |
| 116 | 110 |
| 117 return ctx; | 111 return ctx; |
| 118 } | 112 } |
| 119 | 113 |
| 120 } // namespace trace_event | 114 } // namespace trace_event |
| 121 } // namespace base | 115 } // namespace base |
| OLD | NEW |