| Index: base/debug/activity_tracker.cc
|
| diff --git a/base/debug/activity_tracker.cc b/base/debug/activity_tracker.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..ea9e1258e870823df406c94b184f5615f7c19edc
|
| --- /dev/null
|
| +++ b/base/debug/activity_tracker.cc
|
| @@ -0,0 +1,780 @@
|
| +// Copyright 2016 The Chromium Authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include "base/debug/activity_tracker.h"
|
| +
|
| +#include "base/debug/stack_trace.h"
|
| +#include "base/files/file.h"
|
| +#include "base/files/file_path.h"
|
| +#include "base/files/memory_mapped_file.h"
|
| +#include "base/logging.h"
|
| +#include "base/memory/ptr_util.h"
|
| +#include "base/metrics/field_trial.h"
|
| +#include "base/metrics/histogram_macros.h"
|
| +#include "base/pending_task.h"
|
| +#include "base/process/process.h"
|
| +#include "base/process/process_handle.h"
|
| +#include "base/stl_util.h"
|
| +#include "base/strings/string_util.h"
|
| +#include "base/threading/platform_thread.h"
|
| +
|
| +namespace base {
|
| +namespace debug {
|
| +
|
| +namespace {
|
| +
|
| +// A number that identifies the memory as having been initialized. It's
|
| +// arbitrary but happens to be the first 8 bytes of SHA1(ThreadActivityTracker).
|
| +// A version number is added on so that major structure changes won't try to
|
| +// read an older version (since the cookie won't match).
|
| +const uint64_t kHeaderCookie = 0xC0029B240D4A3092ULL + 1; // v1
|
| +
|
| +// The minimum depth a stack should support.
|
| +const int kMinStackDepth = 2;
|
| +
|
| +} // namespace
|
| +
|
| +
|
| +// This information is kept for every thread that is tracked. It is filled
|
| +// the very first time the thread is seen. All fields must be of exact sizes
|
| +// so there is no issue moving between 32 and 64-bit builds.
|
| +struct ThreadActivityTracker::Header {
|
| + // This unique number indicates a valid initialization of the memory.
|
| + uint64_t cookie;
|
| +
|
| + // The process-id and thread-id to which this data belongs. These identifiers
|
| + // are not guaranteed to mean anything but are unique, in combination, among
|
| + // all active trackers. It would be nice to always have the process_id be a
|
| + // 64-bit value but the necessity of having it atomic (for the memory barriers
|
| + // it provides) limits it to the natural word size of the machine.
|
| +#ifdef ARCH_CPU_64_BITS
|
| + std::atomic<int64_t> process_id;
|
| +#else
|
| + std::atomic<int32_t> process_id;
|
| + int32_t process_id_padding;
|
| +#endif
|
| +
|
| + union {
|
| + int64_t as_id;
|
| +#if defined(OS_WIN)
|
| + // On Windows, the handle itself is often a pseudo-handle with a common
|
| + // value meaning "this thread" and so the thread-id is used. The former
|
| + // can be converted to a thread-id with a system call.
|
| + PlatformThreadId as_tid;
|
| +#elif defined(OS_POSIX)
|
| + // On Posix, the handle is always a unique identifier so no conversion
|
| + // needs to be done. However, it's value is officially opaque so there
|
| + // is no one correct way to convert it to a numerical identifier.
|
| + PlatformThreadHandle::Handle as_handle;
|
| +#endif
|
| + } thread_ref;
|
| +
|
| + // The start-time and start-ticks when the data was created. Each activity
|
| + // record has a |time_internal| value that can be converted to a "wall time"
|
| + // with these two values.
|
| + int64_t start_time;
|
| + int64_t start_ticks;
|
| +
|
| + // The number of Activity slots in the data.
|
| + uint32_t stack_slots;
|
| +
|
| + // The current depth of the stack. This may be greater than the number of
|
| + // slots. If the depth exceeds the number of slots, the newest entries
|
| + // won't be recorded.
|
| + std::atomic<uint32_t> current_depth;
|
| +
|
| + // A memory location used to indicate if changes have been made to the stack
|
| + // that would invalidate an in-progress read of its contents. The active
|
| + // tracker will zero the value whenever something gets popped from the
|
| + // stack. A monitoring tracker can write a non-zero value here, copy the
|
| + // stack contents, and read the value to know, if it is still non-zero, that
|
| + // the contents didn't change while being copied. This can handle concurrent
|
| + // snapshot operations only if each snapshot writes a different bit (which
|
| + // is not the current implementation so no parallel snapshots allowed).
|
| + std::atomic<uint32_t> stack_unchanged;
|
| +
|
| + // The name of the thread (up to a maximum length). Dynamic-length names
|
| + // are not practical since the memory has to come from the same persistent
|
| + // allocator that holds this structure and to which this object has no
|
| + // reference.
|
| + char thread_name[32];
|
| +};
|
| +
|
| +// It doesn't matter what is contained in this (though it will be all zeros)
|
| +// as only the address of it is important.
|
| +const ThreadActivityTracker::ActivityData
|
| + ThreadActivityTracker::kNullActivityData = {};
|
| +
|
| +ThreadActivityTracker::ActivityData
|
| +ThreadActivityTracker::ActivityData::ForThread(
|
| + const PlatformThreadHandle& handle) {
|
| + // Header already has a conversion union; reuse that.
|
| + ThreadActivityTracker::Header header;
|
| + header.thread_ref.as_id = 0; // Zero the union in case other is smaller.
|
| +#if defined(OS_WIN)
|
| + header.thread_ref.as_tid = ::GetThreadId(handle.platform_handle());
|
| +#elif defined(OS_POSIX)
|
| + header.thread_ref.as_handle = handle.platform_handle();
|
| +#endif
|
| + return ForThread(header.thread_ref.as_id);
|
| +}
|
| +
|
| +ThreadActivityTracker::ActivitySnapshot::ActivitySnapshot() {}
|
| +ThreadActivityTracker::ActivitySnapshot::~ActivitySnapshot() {}
|
| +
|
| +
|
| +ThreadActivityTracker::ThreadActivityTracker(void* base, size_t size)
|
| + : header_(static_cast<Header*>(base)),
|
| + stack_(reinterpret_cast<Activity*>(reinterpret_cast<char*>(base) +
|
| + sizeof(Header))),
|
| + stack_slots_(
|
| + static_cast<uint32_t>((size - sizeof(Header)) / sizeof(Activity))) {
|
| + DCHECK(thread_checker_.CalledOnValidThread());
|
| +
|
| + // Verify the parameters but fail gracefully if they're not valid so that
|
| + // production code based on external inputs will not crash. IsValid() will
|
| + // return false in this case.
|
| + if (!base ||
|
| + // Ensure there is enough space for the header and at least a few records.
|
| + size < sizeof(Header) + kMinStackDepth * sizeof(Activity) ||
|
| + // Ensure that the |stack_slots_| calculation didn't overflow.
|
| + (size - sizeof(Header)) / sizeof(Activity) >
|
| + std::numeric_limits<uint32_t>::max()) {
|
| + NOTREACHED();
|
| + return;
|
| + }
|
| +
|
| + // Ensure that the thread reference doesn't exceed the size of the ID number.
|
| + // This won't compile at the global scope because Header is a private struct.
|
| + static_assert(
|
| + sizeof(header_->thread_ref) == sizeof(header_->thread_ref.as_id),
|
| + "PlatformThreadHandle::Handle is too big to hold in 64-bit ID");
|
| +
|
| + // Ensure that the alignment of Activity.data is properly aligned to a
|
| + // 64-bit boundary so there are no interoperability-issues across cpu
|
| + // architectures.
|
| + static_assert(offsetof(Activity, data) % sizeof(uint64_t) == 0,
|
| + "ActivityData.data is not 64-bit aligned");
|
| +
|
| + // Provided memory should either be completely initialized or all zeros.
|
| + if (header_->cookie == 0) {
|
| + // This is a new file. Double-check other fields and then initialize.
|
| + DCHECK_EQ(0, header_->process_id.load(std::memory_order_relaxed));
|
| + DCHECK_EQ(0, header_->thread_ref.as_id);
|
| + DCHECK_EQ(0, header_->start_time);
|
| + DCHECK_EQ(0, header_->start_ticks);
|
| + DCHECK_EQ(0U, header_->stack_slots);
|
| + DCHECK_EQ(0U, header_->current_depth.load(std::memory_order_relaxed));
|
| + DCHECK_EQ(0U, header_->stack_unchanged.load(std::memory_order_relaxed));
|
| + DCHECK_EQ(0, stack_[0].time_internal);
|
| + DCHECK_EQ(0U, stack_[0].origin_address);
|
| + DCHECK_EQ(0U, stack_[0].call_stack[0]);
|
| + DCHECK_EQ(0U, stack_[0].data.task.sequence_id);
|
| +
|
| +#if defined(OS_WIN)
|
| + header_->thread_ref.as_tid = PlatformThread::CurrentId();
|
| +#elif defined(OS_POSIX)
|
| + header_->thread_ref.as_handle =
|
| + PlatformThread::CurrentHandle().platform_handle();
|
| +#endif
|
| + header_->start_time = base::Time::Now().ToInternalValue();
|
| + header_->start_ticks = base::TimeTicks::Now().ToInternalValue();
|
| + header_->stack_slots = stack_slots_;
|
| + strlcpy(header_->thread_name, PlatformThread::GetName(),
|
| + sizeof(header_->thread_name));
|
| + header_->cookie = kHeaderCookie;
|
| +
|
| + // This is done last so as to guarantee that everything above is "released"
|
| + // by the time this value gets written.
|
| + header_->process_id.store(GetCurrentProcId(), std::memory_order_release);
|
| +
|
| + valid_ = true;
|
| + DCHECK(IsValid());
|
| + } else {
|
| + // This is a file with existing data. Perform basic consistency checks.
|
| + valid_ = true;
|
| + valid_ = IsValid();
|
| + }
|
| +}
|
| +
|
| +ThreadActivityTracker::~ThreadActivityTracker() {}
|
| +
|
| +void ThreadActivityTracker::PushActivity(const void* origin,
|
| + ActivityType type,
|
| + const ActivityData& data) {
|
| + // A thread-checker creates a lock to check the thread-id which means
|
| + // re-entry into this code if lock acquisitions are being tracked.
|
| + DCHECK(type == ACT_LOCK_ACQUIRE || thread_checker_.CalledOnValidThread());
|
| +
|
| + // Get the current depth of the stack. No access to other memory guarded
|
| + // by this variable is done here so a "relaxed" load is acceptable.
|
| + uint32_t depth = header_->current_depth.load(std::memory_order_relaxed);
|
| +
|
| + // Handle the case where the stack depth has exceeded the storage capacity.
|
| + // Extra entries will be lost leaving only the base of the stack.
|
| + if (depth >= stack_slots_) {
|
| + // Since no other threads modify the data, no compare/exchange is needed.
|
| + // Since no other memory is being modified, a "relaxed" store is acceptable.
|
| + header_->current_depth.store(depth + 1, std::memory_order_relaxed);
|
| + return;
|
| + }
|
| +
|
| + // Get a pointer to the next activity and load it. No atomicity is required
|
| + // here because the memory is known only to this thread. It will be made
|
| + // known to other threads once the depth is incremented.
|
| + Activity* activity = &stack_[depth];
|
| + activity->time_internal = base::TimeTicks::Now().ToInternalValue();
|
| + activity->origin_address = reinterpret_cast<uintptr_t>(origin);
|
| + activity->activity_type = type;
|
| + activity->data = data;
|
| +
|
| +#if defined(SYZYASAN)
|
| + // Create a stacktrace from the current location and get the addresses.
|
| + StackTrace stack_trace;
|
| + size_t stack_depth;
|
| + const void* const* stack_addrs = stack_trace.Addresses(&stack_depth);
|
| + // Copy the stack addresses, ignoring the first one (here).
|
| + size_t i;
|
| + for (i = 1; i < stack_depth && i < kActivityCallStackSize; ++i) {
|
| + activity->call_stack[i - 1] = reinterpret_cast<uintptr_t>(stack_addrs[i]);
|
| + }
|
| + activity->call_stack[i - 1] = 0;
|
| +#else
|
| + // Since the memory was initially zero and nothing ever overwrites it in
|
| + // this "else" case, there is no need to write even the null terminator.
|
| + //activity->call_stack[0] = 0;
|
| +#endif
|
| +
|
| + // Save the incremented depth. Because this guards |activity| memory filled
|
| + // above that may be read by another thread once the recorded depth changes,
|
| + // a "release" store is required.
|
| + header_->current_depth.store(depth + 1, std::memory_order_release);
|
| +}
|
| +
|
| +void ThreadActivityTracker::ChangeActivity(ActivityType type,
|
| + const ActivityData& data) {
|
| + DCHECK(thread_checker_.CalledOnValidThread());
|
| + DCHECK(type != ACT_NULL || &data != &kNullActivityData);
|
| +
|
| + // Get the current depth of the stack and acquire the data held there.
|
| + uint32_t depth = header_->current_depth.load(std::memory_order_acquire);
|
| + DCHECK_LT(0U, depth);
|
| +
|
| + // Update the information if it is being recorded (i.e. within slot limit).
|
| + if (depth <= stack_slots_) {
|
| + Activity* activity = &stack_[depth - 1];
|
| +
|
| + if (type != ACT_NULL) {
|
| + DCHECK_EQ(activity->activity_type & ACT_CATEGORY_MASK,
|
| + type & ACT_CATEGORY_MASK);
|
| + activity->activity_type = type;
|
| + }
|
| +
|
| + if (&data != &kNullActivityData)
|
| + activity->data = data;
|
| + }
|
| +}
|
| +
|
| +void ThreadActivityTracker::PopActivity() {
|
| + // Do an atomic decrement of the depth. No changes to stack entries guarded
|
| + // by this variable are done here so a "relaxed" operation is acceptable.
|
| + // |depth| will receive the value BEFORE it was modified.
|
| + uint32_t depth =
|
| + header_->current_depth.fetch_sub(1, std::memory_order_relaxed);
|
| +
|
| + // Validate that everything is running correctly.
|
| + DCHECK_LT(0U, depth);
|
| +
|
| + // A thread-checker creates a lock to check the thread-id which means
|
| + // re-entry into this code if lock acquisitions are being tracked.
|
| + DCHECK(stack_[depth - 1].activity_type == ACT_LOCK_ACQUIRE ||
|
| + thread_checker_.CalledOnValidThread());
|
| +
|
| + // The stack has shrunk meaning that some other thread trying to copy the
|
| + // contents for reporting purposes could get bad data. That thread would
|
| + // have written a non-zero value into |stack_unchanged|; clearing it here
|
| + // will let that thread detect that something did change. This needs to
|
| + // happen after the atomic |depth| operation above so a "release" store
|
| + // is required.
|
| + header_->stack_unchanged.store(0, std::memory_order_release);
|
| +}
|
| +
|
| +bool ThreadActivityTracker::IsValid() const {
|
| + if (header_->cookie != kHeaderCookie ||
|
| + header_->process_id.load(std::memory_order_relaxed) == 0 ||
|
| + header_->thread_ref.as_id == 0 ||
|
| + header_->start_time == 0 ||
|
| + header_->start_ticks == 0 ||
|
| + header_->stack_slots != stack_slots_ ||
|
| + header_->thread_name[sizeof(header_->thread_name) - 1] != '\0') {
|
| + return false;
|
| + }
|
| +
|
| + return valid_;
|
| +}
|
| +
|
| +bool ThreadActivityTracker::Snapshot(ActivitySnapshot* output_snapshot) const {
|
| + DCHECK(output_snapshot);
|
| +
|
| + // There is no "called on valid thread" check for this method as it can be
|
| + // called from other threads or even other processes. It is also the reason
|
| + // why atomic operations must be used in certain places above.
|
| +
|
| + // It's possible for the data to change while reading it in such a way that it
|
| + // invalidates the read. Make several attempts but don't try forever.
|
| + const int kMaxAttempts = 10;
|
| + uint32_t depth;
|
| +
|
| + // Stop here if the data isn't valid.
|
| + if (!IsValid())
|
| + return false;
|
| +
|
| + // Allocate the maximum size for the stack so it doesn't have to be done
|
| + // during the time-sensitive snapshot operation. It is shrunk once the
|
| + // actual size is known.
|
| + output_snapshot->activity_stack.reserve(stack_slots_);
|
| +
|
| + for (int attempt = 0; attempt < kMaxAttempts; ++attempt) {
|
| + // Remember the process and thread IDs to ensure they aren't replaced
|
| + // during the snapshot operation. Use "acquire" to ensure that all the
|
| + // non-atomic fields of the structure are valid (at least at the current
|
| + // moment in time).
|
| + const int64_t starting_process_id =
|
| + header_->process_id.load(std::memory_order_acquire);
|
| + const int64_t starting_thread_id = header_->thread_ref.as_id;
|
| +
|
| + // Write a non-zero value to |stack_unchanged| so it's possible to detect
|
| + // at the end that nothing has changed since copying the data began. A
|
| + // "cst" operation is required to ensure it occurs before everything else.
|
| + // Using "cst" memory ordering is relatively expensive but this is only
|
| + // done during analysis so doesn't directly affect the worker threads.
|
| + header_->stack_unchanged.store(1, std::memory_order_seq_cst);
|
| +
|
| + // Fetching the current depth also "acquires" the contents of the stack.
|
| + depth = header_->current_depth.load(std::memory_order_acquire);
|
| + uint32_t count = std::min(depth, stack_slots_);
|
| + output_snapshot->activity_stack.resize(count);
|
| + if (count > 0) {
|
| + // Copy the existing contents. Memcpy is used for speed.
|
| + memcpy(&output_snapshot->activity_stack[0], stack_,
|
| + count * sizeof(Activity));
|
| + }
|
| +
|
| + // Retry if something changed during the copy. A "cst" operation ensures
|
| + // it must happen after all the above operations.
|
| + if (!header_->stack_unchanged.load(std::memory_order_seq_cst))
|
| + continue;
|
| +
|
| + // Stack copied. Record it's full depth.
|
| + output_snapshot->activity_stack_depth = depth;
|
| +
|
| + // TODO(bcwhite): Snapshot other things here.
|
| +
|
| + // Get the general thread information. Loading of "process_id" is guaranteed
|
| + // to be last so that it's possible to detect below if any content has
|
| + // changed while reading it. It's technically possible for a thread to end,
|
| + // have its data cleared, a new thread get created with the same IDs, and
|
| + // it perform an action which starts tracking all in the time since the
|
| + // ID reads above but the chance is so unlikely that it's not worth the
|
| + // effort and complexity of protecting against it (perhaps with an
|
| + // "unchanged" field like is done for the stack).
|
| + output_snapshot->thread_name =
|
| + std::string(header_->thread_name, sizeof(header_->thread_name) - 1);
|
| + output_snapshot->thread_id = header_->thread_ref.as_id;
|
| + output_snapshot->process_id =
|
| + header_->process_id.load(std::memory_order_seq_cst);
|
| +
|
| + // All characters of the thread-name buffer were copied so as to not break
|
| + // if the trailing NUL were missing. Now limit the length if the actual
|
| + // name is shorter.
|
| + output_snapshot->thread_name.resize(
|
| + strlen(output_snapshot->thread_name.c_str()));
|
| +
|
| + // If the process or thread ID has changed then the tracker has exited and
|
| + // the memory reused by a new one. Try again.
|
| + if (output_snapshot->process_id != starting_process_id ||
|
| + output_snapshot->thread_id != starting_thread_id) {
|
| + continue;
|
| + }
|
| +
|
| + // Only successful if the data is still valid once everything is done since
|
| + // it's possible for the thread to end somewhere in the middle and all its
|
| + // values become garbage.
|
| + if (!IsValid())
|
| + return false;
|
| +
|
| + // Change all the timestamps in the activities from "ticks" to "wall" time.
|
| + const Time start_time = Time::FromInternalValue(header_->start_time);
|
| + const int64_t start_ticks = header_->start_ticks;
|
| + for (Activity& activity : output_snapshot->activity_stack) {
|
| + activity.time_internal =
|
| + (start_time +
|
| + TimeDelta::FromInternalValue(activity.time_internal - start_ticks))
|
| + .ToInternalValue();
|
| + }
|
| +
|
| + // Success!
|
| + return true;
|
| + }
|
| +
|
| + // Too many attempts.
|
| + return false;
|
| +}
|
| +
|
| +// static
|
| +size_t ThreadActivityTracker::SizeForStackDepth(int stack_depth) {
|
| + return static_cast<size_t>(stack_depth) * sizeof(Activity) + sizeof(Header);
|
| +}
|
| +
|
| +
|
| +GlobalActivityTracker* GlobalActivityTracker::g_tracker_ = nullptr;
|
| +
|
| +GlobalActivityTracker::ManagedActivityTracker::ManagedActivityTracker(
|
| + PersistentMemoryAllocator::Reference mem_reference,
|
| + void* base,
|
| + size_t size)
|
| + : ThreadActivityTracker(base, size),
|
| + mem_reference_(mem_reference),
|
| + mem_base_(base) {}
|
| +
|
| +GlobalActivityTracker::ManagedActivityTracker::~ManagedActivityTracker() {
|
| + // The global |g_tracker_| must point to the owner of this class since all
|
| + // objects of this type must be destructed before |g_tracker_| can be changed
|
| + // (something that only occurs in tests).
|
| + DCHECK(g_tracker_);
|
| + g_tracker_->ReturnTrackerMemory(this);
|
| +}
|
| +
|
| +void GlobalActivityTracker::CreateWithAllocator(
|
| + std::unique_ptr<PersistentMemoryAllocator> allocator,
|
| + int stack_depth) {
|
| + // There's no need to do anything with the result. It is self-managing.
|
| + GlobalActivityTracker* global_tracker =
|
| + new GlobalActivityTracker(std::move(allocator), stack_depth);
|
| + // Create a tracker for this thread since it is known.
|
| + global_tracker->CreateTrackerForCurrentThread();
|
| +}
|
| +
|
| +#if !defined(OS_NACL)
|
| +// static
|
| +void GlobalActivityTracker::CreateWithFile(const FilePath& file_path,
|
| + size_t size,
|
| + uint64_t id,
|
| + StringPiece name,
|
| + int stack_depth) {
|
| + DCHECK(!file_path.empty());
|
| + DCHECK_GE(static_cast<uint64_t>(std::numeric_limits<int64_t>::max()), size);
|
| +
|
| + // Create and map the file into memory and make it globally available.
|
| + std::unique_ptr<MemoryMappedFile> mapped_file(new MemoryMappedFile());
|
| + bool success =
|
| + mapped_file->Initialize(File(file_path,
|
| + File::FLAG_CREATE_ALWAYS | File::FLAG_READ |
|
| + File::FLAG_WRITE | File::FLAG_SHARE_DELETE),
|
| + {0, static_cast<int64_t>(size)},
|
| + MemoryMappedFile::READ_WRITE_EXTEND);
|
| + DCHECK(success);
|
| + CreateWithAllocator(WrapUnique(new FilePersistentMemoryAllocator(
|
| + std::move(mapped_file), size, id, name, false)),
|
| + stack_depth);
|
| +}
|
| +#endif // !defined(OS_NACL)
|
| +
|
| +// static
|
| +void GlobalActivityTracker::CreateWithLocalMemory(size_t size,
|
| + uint64_t id,
|
| + StringPiece name,
|
| + int stack_depth) {
|
| + CreateWithAllocator(
|
| + WrapUnique(new LocalPersistentMemoryAllocator(size, id, name)),
|
| + stack_depth);
|
| +}
|
| +
|
| +ThreadActivityTracker* GlobalActivityTracker::CreateTrackerForCurrentThread() {
|
| + DCHECK(!this_thread_tracker_.Get());
|
| +
|
| + PersistentMemoryAllocator::Reference mem_reference = 0;
|
| + void* mem_base = nullptr;
|
| +
|
| + // Get the current count of available memories, acquiring the array values.
|
| + int count = available_memories_count_.load(std::memory_order_acquire);
|
| + while (count > 0) {
|
| + // There is a memory block that was previously released (and zeroed) so
|
| + // just re-use that rather than allocating a new one. Use "relaxed" because
|
| + // the value is guarded by the |count| "acquire". A zero reference replaces
|
| + // the existing value so that it can't be used by another thread that
|
| + // manages to interrupt this one before the count can be decremented.
|
| + // A zero reference is also required for the "push" operation to work
|
| + // once the count finally does get decremented.
|
| + mem_reference =
|
| + available_memories_[count - 1].exchange(0, std::memory_order_relaxed);
|
| +
|
| + // If the reference is zero, it's already been taken but count hasn't yet
|
| + // been decremented. Give that other thread a chance to finish then reload
|
| + // the "count" value and try again.
|
| + if (!mem_reference) {
|
| + PlatformThread::YieldCurrentThread();
|
| + count = available_memories_count_.load(std::memory_order_acquire);
|
| + continue;
|
| + }
|
| +
|
| + // Decrement the count indicating that the value has been taken. If this
|
| + // fails then another thread has pushed something new and incremented the
|
| + // count.
|
| + // NOTE: |oldcount| will be loaded with the existing value.
|
| + int oldcount = count;
|
| + if (!available_memories_count_.compare_exchange_strong(
|
| + oldcount, count - 1, std::memory_order_acquire,
|
| + std::memory_order_acquire)) {
|
| + DCHECK_LT(count, oldcount);
|
| +
|
| + // Restore the reference that was zeroed above and try again.
|
| + available_memories_[count - 1].store(mem_reference,
|
| + std::memory_order_relaxed);
|
| + count = oldcount;
|
| + continue;
|
| + }
|
| +
|
| + // Turn the reference back into one of the activity-tracker type.
|
| + mem_base = allocator_->GetAsObject<char>(mem_reference,
|
| + kTypeIdActivityTrackerFree);
|
| + DCHECK(mem_base);
|
| + DCHECK_LE(stack_memory_size_, allocator_->GetAllocSize(mem_reference));
|
| + bool changed = allocator_->ChangeType(mem_reference, kTypeIdActivityTracker,
|
| + kTypeIdActivityTrackerFree);
|
| + DCHECK(changed);
|
| +
|
| + // Success.
|
| + break;
|
| + }
|
| +
|
| + // Handle the case where no previously-used memories are available.
|
| + if (count == 0) {
|
| + // Allocate a block of memory from the persistent segment.
|
| + mem_reference =
|
| + allocator_->Allocate(stack_memory_size_, kTypeIdActivityTracker);
|
| + if (mem_reference) {
|
| + // Success. Convert the reference to an actual memory address.
|
| + mem_base =
|
| + allocator_->GetAsObject<char>(mem_reference, kTypeIdActivityTracker);
|
| + // Make the allocation iterable so it can be found by other processes.
|
| + allocator_->MakeIterable(mem_reference);
|
| + } else {
|
| + // Failure. This shouldn't happen.
|
| + NOTREACHED();
|
| + // But if it does, probably because the allocator wasn't given enough
|
| + // memory to satisfy all possible requests, handle it gracefully by
|
| + // allocating the required memory from the heap.
|
| + mem_base = new char[stack_memory_size_];
|
| + memset(mem_base, 0, stack_memory_size_);
|
| + // Report the thread-count at which the allocator was full so that the
|
| + // failure can be seen and underlying memory resized appropriately.
|
| + UMA_HISTOGRAM_COUNTS_1000(
|
| + "ActivityTracker.ThreadTrackers.MemLimitTrackerCount",
|
| + thread_tracker_count_.load(std::memory_order_relaxed));
|
| + }
|
| + }
|
| +
|
| + // Create a tracker with the acquired memory and set it as the tracker
|
| + // for this particular thread in thread-local-storage.
|
| + DCHECK(mem_base);
|
| + ManagedActivityTracker* tracker =
|
| + new ManagedActivityTracker(mem_reference, mem_base, stack_memory_size_);
|
| + DCHECK(tracker->IsValid());
|
| + this_thread_tracker_.Set(tracker);
|
| + int old_count = thread_tracker_count_.fetch_add(1, std::memory_order_relaxed);
|
| +
|
| + UMA_HISTOGRAM_ENUMERATION("ActivityTracker.ThreadTrackers.Count",
|
| + old_count + 1, kMaxThreadCount);
|
| + return tracker;
|
| +}
|
| +
|
| +void GlobalActivityTracker::ReleaseTrackerForCurrentThreadForTesting() {
|
| + ThreadActivityTracker* tracker =
|
| + reinterpret_cast<ThreadActivityTracker*>(this_thread_tracker_.Get());
|
| + if (tracker) {
|
| + this_thread_tracker_.Free();
|
| + delete tracker;
|
| + }
|
| +}
|
| +
|
| +GlobalActivityTracker::GlobalActivityTracker(
|
| + std::unique_ptr<PersistentMemoryAllocator> allocator,
|
| + int stack_depth)
|
| + : allocator_(std::move(allocator)),
|
| + stack_memory_size_(ThreadActivityTracker::SizeForStackDepth(stack_depth)),
|
| + this_thread_tracker_(&OnTLSDestroy),
|
| + thread_tracker_count_(0),
|
| + available_memories_count_(0) {
|
| + // Clear the available-memories array.
|
| + memset(available_memories_, 0, sizeof(available_memories_));
|
| +
|
| + // Ensure the passed memory is valid and empty (iterator finds nothing).
|
| + uint32_t type;
|
| + DCHECK(!PersistentMemoryAllocator::Iterator(allocator_.get()).GetNext(&type));
|
| +
|
| + // Ensure that there is no other global object and then make this one such.
|
| + DCHECK(!g_tracker_);
|
| + g_tracker_ = this;
|
| +}
|
| +
|
| +GlobalActivityTracker::~GlobalActivityTracker() {
|
| + DCHECK_EQ(g_tracker_, this);
|
| + DCHECK_EQ(0, thread_tracker_count_.load(std::memory_order_relaxed));
|
| + g_tracker_ = nullptr;
|
| +}
|
| +
|
| +void GlobalActivityTracker::ReturnTrackerMemory(
|
| + ManagedActivityTracker* tracker) {
|
| + PersistentMemoryAllocator::Reference mem_reference = tracker->mem_reference_;
|
| + void* mem_base = tracker->mem_base_;
|
| +
|
| + // Zero the memory so that it is ready for use if needed again later. It's
|
| + // better to clear the memory now, when a thread is exiting, than to do it
|
| + // when it is first needed by a thread doing actual work.
|
| + memset(mem_base, 0, stack_memory_size_);
|
| +
|
| + // Remove the destructed tracker from the set of known ones.
|
| + DCHECK_LE(1, thread_tracker_count_.load(std::memory_order_relaxed));
|
| + thread_tracker_count_.fetch_sub(1, std::memory_order_relaxed);
|
| +
|
| + // Deal with the memory that was used by the tracker.
|
| + if (mem_reference) {
|
| + // The memory was within the persistent memory allocator. Change its type
|
| + // so that iteration won't find it.
|
| + allocator_->ChangeType(mem_reference, kTypeIdActivityTrackerFree,
|
| + kTypeIdActivityTracker);
|
| + // There is no way to free memory from a persistent allocator so instead
|
| + // push it on the internal list of available memory blocks.
|
| + while (true) {
|
| + // Get the existing count of available memories and ensure we won't
|
| + // burst the array. Acquire the values in the array.
|
| + int count = available_memories_count_.load(std::memory_order_acquire);
|
| + if (count >= kMaxThreadCount) {
|
| + NOTREACHED();
|
| + // Storage is full. Just forget about this memory. It won't be re-used
|
| + // but there's no real loss.
|
| + break;
|
| + }
|
| +
|
| + // Write the reference of the memory being returned to this slot in the
|
| + // array. Empty slots have a value of zero so do an atomic compare-and-
|
| + // exchange to ensure that a race condition doesn't exist with another
|
| + // thread doing the same.
|
| + PersistentMemoryAllocator::Reference mem_expected = 0;
|
| + if (!available_memories_[count].compare_exchange_strong(
|
| + mem_expected, mem_reference, std::memory_order_release,
|
| + std::memory_order_relaxed)) {
|
| + PlatformThread::YieldCurrentThread();
|
| + continue; // Try again.
|
| + }
|
| +
|
| + // Increment the count, releasing the value written to the array. This
|
| + // could fail if a simultaneous "pop" operation decremented the counter.
|
| + // If that happens, clear the array slot and start over. Do a "strong"
|
| + // exchange to avoid spurious retries that can occur with a "weak" one.
|
| + int expected = count; // Updated by compare/exchange.
|
| + if (!available_memories_count_.compare_exchange_strong(
|
| + expected, count + 1, std::memory_order_release,
|
| + std::memory_order_relaxed)) {
|
| + available_memories_[count].store(0, std::memory_order_relaxed);
|
| + continue;
|
| + }
|
| +
|
| + // Count was successfully incremented to reflect the newly added value.
|
| + break;
|
| + }
|
| + } else {
|
| + // The memory was allocated from the process heap. This shouldn't happen
|
| + // because the persistent memory segment should be big enough for all
|
| + // thread stacks but it's better to support falling back to allocation
|
| + // from the heap rather than crash. Everything will work as normal but
|
| + // the data won't be persisted.
|
| + delete[] reinterpret_cast<char*>(mem_base);
|
| + }
|
| +}
|
| +
|
| +// static
|
| +void GlobalActivityTracker::OnTLSDestroy(void* value) {
|
| + delete reinterpret_cast<ManagedActivityTracker*>(value);
|
| +}
|
| +
|
| +
|
| +ScopedActivity::ScopedActivity(const tracked_objects::Location& location,
|
| + uint8_t action,
|
| + uint32_t id,
|
| + int32_t info)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + location.program_counter(),
|
| + static_cast<ThreadActivityTracker::ActivityType>(
|
| + ThreadActivityTracker::ACT_GENERIC | action),
|
| + ThreadActivityTracker::ActivityData::ForGeneric(id, info),
|
| + /*lock_allowed=*/true),
|
| + id_(id) {
|
| + // The action must not affect the category bits of the activity type.
|
| + DCHECK_EQ(0, action & ThreadActivityTracker::ACT_CATEGORY_MASK);
|
| +}
|
| +
|
| +void ScopedActivity::ChangeAction(uint8_t action) {
|
| + DCHECK_EQ(0, action & ThreadActivityTracker::ACT_CATEGORY_MASK);
|
| + ChangeTypeAndData(static_cast<ThreadActivityTracker::ActivityType>(
|
| + ThreadActivityTracker::ACT_GENERIC | action),
|
| + ThreadActivityTracker::kNullActivityData);
|
| +}
|
| +
|
| +void ScopedActivity::ChangeInfo(int32_t info) {
|
| + ChangeTypeAndData(ThreadActivityTracker::ACT_NULL,
|
| + ThreadActivityTracker::ActivityData::ForGeneric(id_, info));
|
| +}
|
| +
|
| +void ScopedActivity::ChangeActionAndInfo(uint8_t action, int32_t info) {
|
| + DCHECK_EQ(0, action & ThreadActivityTracker::ACT_CATEGORY_MASK);
|
| + ChangeTypeAndData(static_cast<ThreadActivityTracker::ActivityType>(
|
| + ThreadActivityTracker::ACT_GENERIC | action),
|
| + ThreadActivityTracker::ActivityData::ForGeneric(id_, info));
|
| +}
|
| +
|
| +ScopedTaskRunActivity::ScopedTaskRunActivity(const base::PendingTask& task)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + task.posted_from.program_counter(),
|
| + ThreadActivityTracker::ACT_TASK_RUN,
|
| + ThreadActivityTracker::ActivityData::ForTask(task.sequence_num),
|
| + /*lock_allowed=*/true) {}
|
| +
|
| +ScopedLockAcquireActivity::ScopedLockAcquireActivity(
|
| + const base::internal::LockImpl* lock)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + nullptr,
|
| + ThreadActivityTracker::ACT_LOCK_ACQUIRE,
|
| + ThreadActivityTracker::ActivityData::ForLock(lock),
|
| + /*lock_allowed=*/false) {}
|
| +
|
| +ScopedEventWaitActivity::ScopedEventWaitActivity(
|
| + const base::WaitableEvent* event)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + nullptr,
|
| + ThreadActivityTracker::ACT_EVENT_WAIT,
|
| + ThreadActivityTracker::ActivityData::ForEvent(event),
|
| + /*lock_allowed=*/true) {}
|
| +
|
| +ScopedThreadJoinActivity::ScopedThreadJoinActivity(
|
| + const base::PlatformThreadHandle* thread)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + nullptr,
|
| + ThreadActivityTracker::ACT_THREAD_JOIN,
|
| + ThreadActivityTracker::ActivityData::ForThread(*thread),
|
| + /*lock_allowed=*/true) {}
|
| +
|
| +#if !defined(OS_NACL) && !defined(OS_IOS)
|
| +ScopedProcessWaitActivity::ScopedProcessWaitActivity(
|
| + const base::Process* process)
|
| + : GlobalActivityTracker::ScopedThreadActivity(
|
| + nullptr,
|
| + ThreadActivityTracker::ACT_PROCESS_WAIT,
|
| + ThreadActivityTracker::ActivityData::ForProcess(process->Pid()),
|
| + /*lock_allowed=*/true) {}
|
| +#endif
|
| +
|
| +} // namespace debug
|
| +} // namespace base
|
|
|