| Index: src/profiler/sampler.cc
|
| diff --git a/src/profiler/sampler.cc b/src/profiler/sampler.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..ae47dca1504ed094448329f7b163450ee96016d0
|
| --- /dev/null
|
| +++ b/src/profiler/sampler.cc
|
| @@ -0,0 +1,828 @@
|
| +// Copyright 2013 the V8 project authors. All rights reserved.
|
| +// Use of this source code is governed by a BSD-style license that can be
|
| +// found in the LICENSE file.
|
| +
|
| +#include "src/profiler/sampler.h"
|
| +
|
| +#if V8_OS_POSIX && !V8_OS_CYGWIN
|
| +
|
| +#define USE_SIGNALS
|
| +
|
| +#include <errno.h>
|
| +#include <pthread.h>
|
| +#include <signal.h>
|
| +#include <sys/time.h>
|
| +
|
| +#if !V8_OS_QNX && !V8_OS_NACL && !V8_OS_AIX
|
| +#include <sys/syscall.h> // NOLINT
|
| +#endif
|
| +
|
| +#if V8_OS_MACOSX
|
| +#include <mach/mach.h>
|
| +// OpenBSD doesn't have <ucontext.h>. ucontext_t lives in <signal.h>
|
| +// and is a typedef for struct sigcontext. There is no uc_mcontext.
|
| +#elif(!V8_OS_ANDROID || defined(__BIONIC_HAVE_UCONTEXT_T)) && \
|
| + !V8_OS_OPENBSD && !V8_OS_NACL
|
| +#include <ucontext.h>
|
| +#endif
|
| +
|
| +#include <unistd.h>
|
| +
|
| +// GLibc on ARM defines mcontext_t has a typedef for 'struct sigcontext'.
|
| +// Old versions of the C library <signal.h> didn't define the type.
|
| +#if V8_OS_ANDROID && !defined(__BIONIC_HAVE_UCONTEXT_T) && \
|
| + (defined(__arm__) || defined(__aarch64__)) && \
|
| + !defined(__BIONIC_HAVE_STRUCT_SIGCONTEXT)
|
| +#include <asm/sigcontext.h> // NOLINT
|
| +#endif
|
| +
|
| +#elif V8_OS_WIN || V8_OS_CYGWIN
|
| +
|
| +#include "src/base/win32-headers.h"
|
| +
|
| +#endif
|
| +
|
| +#include "src/base/atomic-utils.h"
|
| +#include "src/base/platform/platform.h"
|
| +#include "src/profiler/cpu-profiler-inl.h"
|
| +#include "src/profiler/tick-sample.h"
|
| +#include "src/simulator.h"
|
| +#include "src/v8threads.h"
|
| +
|
| +
|
| +#if V8_OS_ANDROID && !defined(__BIONIC_HAVE_UCONTEXT_T)
|
| +
|
| +// Not all versions of Android's C library provide ucontext_t.
|
| +// Detect this and provide custom but compatible definitions. Note that these
|
| +// follow the GLibc naming convention to access register values from
|
| +// mcontext_t.
|
| +//
|
| +// See http://code.google.com/p/android/issues/detail?id=34784
|
| +
|
| +#if defined(__arm__)
|
| +
|
| +typedef struct sigcontext mcontext_t;
|
| +
|
| +typedef struct ucontext {
|
| + uint32_t uc_flags;
|
| + struct ucontext* uc_link;
|
| + stack_t uc_stack;
|
| + mcontext_t uc_mcontext;
|
| + // Other fields are not used by V8, don't define them here.
|
| +} ucontext_t;
|
| +
|
| +#elif defined(__aarch64__)
|
| +
|
| +typedef struct sigcontext mcontext_t;
|
| +
|
| +typedef struct ucontext {
|
| + uint64_t uc_flags;
|
| + struct ucontext *uc_link;
|
| + stack_t uc_stack;
|
| + mcontext_t uc_mcontext;
|
| + // Other fields are not used by V8, don't define them here.
|
| +} ucontext_t;
|
| +
|
| +#elif defined(__mips__)
|
| +// MIPS version of sigcontext, for Android bionic.
|
| +typedef struct {
|
| + uint32_t regmask;
|
| + uint32_t status;
|
| + uint64_t pc;
|
| + uint64_t gregs[32];
|
| + uint64_t fpregs[32];
|
| + uint32_t acx;
|
| + uint32_t fpc_csr;
|
| + uint32_t fpc_eir;
|
| + uint32_t used_math;
|
| + uint32_t dsp;
|
| + uint64_t mdhi;
|
| + uint64_t mdlo;
|
| + uint32_t hi1;
|
| + uint32_t lo1;
|
| + uint32_t hi2;
|
| + uint32_t lo2;
|
| + uint32_t hi3;
|
| + uint32_t lo3;
|
| +} mcontext_t;
|
| +
|
| +typedef struct ucontext {
|
| + uint32_t uc_flags;
|
| + struct ucontext* uc_link;
|
| + stack_t uc_stack;
|
| + mcontext_t uc_mcontext;
|
| + // Other fields are not used by V8, don't define them here.
|
| +} ucontext_t;
|
| +
|
| +#elif defined(__i386__)
|
| +// x86 version for Android.
|
| +typedef struct {
|
| + uint32_t gregs[19];
|
| + void* fpregs;
|
| + uint32_t oldmask;
|
| + uint32_t cr2;
|
| +} mcontext_t;
|
| +
|
| +typedef uint32_t kernel_sigset_t[2]; // x86 kernel uses 64-bit signal masks
|
| +typedef struct ucontext {
|
| + uint32_t uc_flags;
|
| + struct ucontext* uc_link;
|
| + stack_t uc_stack;
|
| + mcontext_t uc_mcontext;
|
| + // Other fields are not used by V8, don't define them here.
|
| +} ucontext_t;
|
| +enum { REG_EBP = 6, REG_ESP = 7, REG_EIP = 14 };
|
| +
|
| +#elif defined(__x86_64__)
|
| +// x64 version for Android.
|
| +typedef struct {
|
| + uint64_t gregs[23];
|
| + void* fpregs;
|
| + uint64_t __reserved1[8];
|
| +} mcontext_t;
|
| +
|
| +typedef struct ucontext {
|
| + uint64_t uc_flags;
|
| + struct ucontext *uc_link;
|
| + stack_t uc_stack;
|
| + mcontext_t uc_mcontext;
|
| + // Other fields are not used by V8, don't define them here.
|
| +} ucontext_t;
|
| +enum { REG_RBP = 10, REG_RSP = 15, REG_RIP = 16 };
|
| +#endif
|
| +
|
| +#endif // V8_OS_ANDROID && !defined(__BIONIC_HAVE_UCONTEXT_T)
|
| +
|
| +
|
| +namespace v8 {
|
| +namespace internal {
|
| +
|
| +namespace {
|
| +
|
| +class PlatformDataCommon : public Malloced {
|
| + public:
|
| + PlatformDataCommon() : profiled_thread_id_(ThreadId::Current()) {}
|
| + ThreadId profiled_thread_id() { return profiled_thread_id_; }
|
| +
|
| + protected:
|
| + ~PlatformDataCommon() {}
|
| +
|
| + private:
|
| + ThreadId profiled_thread_id_;
|
| +};
|
| +
|
| +
|
| +typedef List<Sampler*> SamplerList;
|
| +
|
| +#if defined(USE_SIGNALS)
|
| +class AtomicGuard {
|
| + public:
|
| + explicit AtomicGuard(base::AtomicValue<int>* atomic, bool is_block = true)
|
| + : atomic_(atomic),
|
| + is_success_(false) {
|
| + do {
|
| + // Use Acquire_Load to gain mutual exclusion.
|
| + USE(atomic_->Value());
|
| + is_success_ = atomic_->TrySetValue(0, 1);
|
| + } while (is_block && !is_success_);
|
| + }
|
| +
|
| + bool is_success() { return is_success_; }
|
| +
|
| + ~AtomicGuard() {
|
| + if (is_success_) {
|
| + atomic_->SetValue(0);
|
| + }
|
| + atomic_ = NULL;
|
| + }
|
| +
|
| + private:
|
| + base::AtomicValue<int>* atomic_;
|
| + bool is_success_;
|
| +};
|
| +
|
| +
|
| +// Returns key for hash map.
|
| +void* ThreadKey(pthread_t thread_id) {
|
| + return reinterpret_cast<void*>(thread_id);
|
| +}
|
| +
|
| +
|
| +// Returns hash value for hash map.
|
| +uint32_t ThreadHash(pthread_t thread_id) {
|
| +#if V8_OS_MACOSX
|
| + return static_cast<uint32_t>(reinterpret_cast<intptr_t>(thread_id));
|
| +#else
|
| + return static_cast<uint32_t>(thread_id);
|
| +#endif
|
| +}
|
| +#endif // USE_SIGNALS
|
| +
|
| +} // namespace
|
| +
|
| +#if defined(USE_SIGNALS)
|
| +
|
| +class Sampler::PlatformData : public PlatformDataCommon {
|
| + public:
|
| + PlatformData() : vm_tid_(pthread_self()) {}
|
| + pthread_t vm_tid() const { return vm_tid_; }
|
| +
|
| + private:
|
| + pthread_t vm_tid_;
|
| +};
|
| +
|
| +#elif V8_OS_WIN || V8_OS_CYGWIN
|
| +
|
| +// ----------------------------------------------------------------------------
|
| +// Win32 profiler support. On Cygwin we use the same sampler implementation as
|
| +// on Win32.
|
| +
|
| +class Sampler::PlatformData : public PlatformDataCommon {
|
| + public:
|
| + // Get a handle to the calling thread. This is the thread that we are
|
| + // going to profile. We need to make a copy of the handle because we are
|
| + // going to use it in the sampler thread. Using GetThreadHandle() will
|
| + // not work in this case. We're using OpenThread because DuplicateHandle
|
| + // for some reason doesn't work in Chrome's sandbox.
|
| + PlatformData()
|
| + : profiled_thread_(OpenThread(THREAD_GET_CONTEXT |
|
| + THREAD_SUSPEND_RESUME |
|
| + THREAD_QUERY_INFORMATION,
|
| + false,
|
| + GetCurrentThreadId())) {}
|
| +
|
| + ~PlatformData() {
|
| + if (profiled_thread_ != NULL) {
|
| + CloseHandle(profiled_thread_);
|
| + profiled_thread_ = NULL;
|
| + }
|
| + }
|
| +
|
| + HANDLE profiled_thread() { return profiled_thread_; }
|
| +
|
| + private:
|
| + HANDLE profiled_thread_;
|
| +};
|
| +#endif
|
| +
|
| +
|
| +#if defined(USE_SIGNALS)
|
| +
|
| +class SignalHandler : public AllStatic {
|
| + public:
|
| + static void SetUp() { if (!mutex_) mutex_ = new base::Mutex(); }
|
| + static void TearDown() { delete mutex_; mutex_ = NULL; }
|
| +
|
| + static void IncreaseSamplerCount() {
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| + if (++client_count_ == 1) Install();
|
| + }
|
| +
|
| + static void DecreaseSamplerCount() {
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| + if (--client_count_ == 0) Restore();
|
| + }
|
| +
|
| + static bool Installed() {
|
| + return signal_handler_installed_;
|
| + }
|
| +
|
| +#if !V8_OS_NACL
|
| + static void CollectSample(void* context, Sampler* sampler);
|
| +#endif
|
| +
|
| + private:
|
| + static void Install() {
|
| +#if !V8_OS_NACL
|
| + struct sigaction sa;
|
| + sa.sa_sigaction = &HandleProfilerSignal;
|
| + sigemptyset(&sa.sa_mask);
|
| +#if V8_OS_QNX
|
| + sa.sa_flags = SA_SIGINFO;
|
| +#else
|
| + sa.sa_flags = SA_RESTART | SA_SIGINFO;
|
| +#endif
|
| + signal_handler_installed_ =
|
| + (sigaction(SIGPROF, &sa, &old_signal_handler_) == 0);
|
| +#endif
|
| + }
|
| +
|
| + static void Restore() {
|
| +#if !V8_OS_NACL
|
| + if (signal_handler_installed_) {
|
| + sigaction(SIGPROF, &old_signal_handler_, 0);
|
| + signal_handler_installed_ = false;
|
| + }
|
| +#endif
|
| + }
|
| +
|
| +#if !V8_OS_NACL
|
| + static void HandleProfilerSignal(int signal, siginfo_t* info, void* context);
|
| +#endif
|
| + // Protects the process wide state below.
|
| + static base::Mutex* mutex_;
|
| + static int client_count_;
|
| + static bool signal_handler_installed_;
|
| + static struct sigaction old_signal_handler_;
|
| +};
|
| +
|
| +
|
| +base::Mutex* SignalHandler::mutex_ = NULL;
|
| +int SignalHandler::client_count_ = 0;
|
| +struct sigaction SignalHandler::old_signal_handler_;
|
| +bool SignalHandler::signal_handler_installed_ = false;
|
| +
|
| +
|
| +// As Native Client does not support signal handling, profiling is disabled.
|
| +#if !V8_OS_NACL
|
| +void SignalHandler::CollectSample(void* context, Sampler* sampler) {
|
| + if (sampler == NULL || (!sampler->IsProfiling() &&
|
| + !sampler->IsRegistered())) {
|
| + return;
|
| + }
|
| + Isolate* isolate = sampler->isolate();
|
| +
|
| + // We require a fully initialized and entered isolate.
|
| + if (isolate == NULL || !isolate->IsInUse()) return;
|
| +
|
| + if (v8::Locker::IsActive() &&
|
| + !isolate->thread_manager()->IsLockedByCurrentThread()) {
|
| + return;
|
| + }
|
| +
|
| + v8::RegisterState state;
|
| +
|
| +#if defined(USE_SIMULATOR)
|
| + if (!SimulatorHelper::FillRegisters(isolate, &state)) return;
|
| +#else
|
| + // Extracting the sample from the context is extremely machine dependent.
|
| + ucontext_t* ucontext = reinterpret_cast<ucontext_t*>(context);
|
| +#if !(V8_OS_OPENBSD || (V8_OS_LINUX && (V8_HOST_ARCH_PPC || V8_HOST_ARCH_S390)))
|
| + mcontext_t& mcontext = ucontext->uc_mcontext;
|
| +#endif
|
| +#if V8_OS_LINUX
|
| +#if V8_HOST_ARCH_IA32
|
| + state.pc = reinterpret_cast<Address>(mcontext.gregs[REG_EIP]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[REG_ESP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[REG_EBP]);
|
| +#elif V8_HOST_ARCH_X64
|
| + state.pc = reinterpret_cast<Address>(mcontext.gregs[REG_RIP]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[REG_RSP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[REG_RBP]);
|
| +#elif V8_HOST_ARCH_ARM
|
| +#if V8_LIBC_GLIBC && !V8_GLIBC_PREREQ(2, 4)
|
| + // Old GLibc ARM versions used a gregs[] array to access the register
|
| + // values from mcontext_t.
|
| + state.pc = reinterpret_cast<Address>(mcontext.gregs[R15]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[R13]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[R11]);
|
| +#else
|
| + state.pc = reinterpret_cast<Address>(mcontext.arm_pc);
|
| + state.sp = reinterpret_cast<Address>(mcontext.arm_sp);
|
| + state.fp = reinterpret_cast<Address>(mcontext.arm_fp);
|
| +#endif // V8_LIBC_GLIBC && !V8_GLIBC_PREREQ(2, 4)
|
| +#elif V8_HOST_ARCH_ARM64
|
| + state.pc = reinterpret_cast<Address>(mcontext.pc);
|
| + state.sp = reinterpret_cast<Address>(mcontext.sp);
|
| + // FP is an alias for x29.
|
| + state.fp = reinterpret_cast<Address>(mcontext.regs[29]);
|
| +#elif V8_HOST_ARCH_MIPS
|
| + state.pc = reinterpret_cast<Address>(mcontext.pc);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[29]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[30]);
|
| +#elif V8_HOST_ARCH_MIPS64
|
| + state.pc = reinterpret_cast<Address>(mcontext.pc);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[29]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[30]);
|
| +#elif V8_HOST_ARCH_PPC
|
| + state.pc = reinterpret_cast<Address>(ucontext->uc_mcontext.regs->nip);
|
| + state.sp = reinterpret_cast<Address>(ucontext->uc_mcontext.regs->gpr[PT_R1]);
|
| + state.fp = reinterpret_cast<Address>(ucontext->uc_mcontext.regs->gpr[PT_R31]);
|
| +#elif V8_HOST_ARCH_S390
|
| +#if V8_TARGET_ARCH_32_BIT
|
| + // 31-bit target will have bit 0 (MSB) of the PSW set to denote addressing
|
| + // mode. This bit needs to be masked out to resolve actual address.
|
| + state.pc =
|
| + reinterpret_cast<Address>(ucontext->uc_mcontext.psw.addr & 0x7FFFFFFF);
|
| +#else
|
| + state.pc = reinterpret_cast<Address>(ucontext->uc_mcontext.psw.addr);
|
| +#endif // V8_TARGET_ARCH_32_BIT
|
| + state.sp = reinterpret_cast<Address>(ucontext->uc_mcontext.gregs[15]);
|
| + state.fp = reinterpret_cast<Address>(ucontext->uc_mcontext.gregs[11]);
|
| +#endif // V8_HOST_ARCH_*
|
| +#elif V8_OS_MACOSX
|
| +#if V8_HOST_ARCH_X64
|
| +#if __DARWIN_UNIX03
|
| + state.pc = reinterpret_cast<Address>(mcontext->__ss.__rip);
|
| + state.sp = reinterpret_cast<Address>(mcontext->__ss.__rsp);
|
| + state.fp = reinterpret_cast<Address>(mcontext->__ss.__rbp);
|
| +#else // !__DARWIN_UNIX03
|
| + state.pc = reinterpret_cast<Address>(mcontext->ss.rip);
|
| + state.sp = reinterpret_cast<Address>(mcontext->ss.rsp);
|
| + state.fp = reinterpret_cast<Address>(mcontext->ss.rbp);
|
| +#endif // __DARWIN_UNIX03
|
| +#elif V8_HOST_ARCH_IA32
|
| +#if __DARWIN_UNIX03
|
| + state.pc = reinterpret_cast<Address>(mcontext->__ss.__eip);
|
| + state.sp = reinterpret_cast<Address>(mcontext->__ss.__esp);
|
| + state.fp = reinterpret_cast<Address>(mcontext->__ss.__ebp);
|
| +#else // !__DARWIN_UNIX03
|
| + state.pc = reinterpret_cast<Address>(mcontext->ss.eip);
|
| + state.sp = reinterpret_cast<Address>(mcontext->ss.esp);
|
| + state.fp = reinterpret_cast<Address>(mcontext->ss.ebp);
|
| +#endif // __DARWIN_UNIX03
|
| +#endif // V8_HOST_ARCH_IA32
|
| +#elif V8_OS_FREEBSD
|
| +#if V8_HOST_ARCH_IA32
|
| + state.pc = reinterpret_cast<Address>(mcontext.mc_eip);
|
| + state.sp = reinterpret_cast<Address>(mcontext.mc_esp);
|
| + state.fp = reinterpret_cast<Address>(mcontext.mc_ebp);
|
| +#elif V8_HOST_ARCH_X64
|
| + state.pc = reinterpret_cast<Address>(mcontext.mc_rip);
|
| + state.sp = reinterpret_cast<Address>(mcontext.mc_rsp);
|
| + state.fp = reinterpret_cast<Address>(mcontext.mc_rbp);
|
| +#elif V8_HOST_ARCH_ARM
|
| + state.pc = reinterpret_cast<Address>(mcontext.mc_r15);
|
| + state.sp = reinterpret_cast<Address>(mcontext.mc_r13);
|
| + state.fp = reinterpret_cast<Address>(mcontext.mc_r11);
|
| +#endif // V8_HOST_ARCH_*
|
| +#elif V8_OS_NETBSD
|
| +#if V8_HOST_ARCH_IA32
|
| + state.pc = reinterpret_cast<Address>(mcontext.__gregs[_REG_EIP]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.__gregs[_REG_ESP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.__gregs[_REG_EBP]);
|
| +#elif V8_HOST_ARCH_X64
|
| + state.pc = reinterpret_cast<Address>(mcontext.__gregs[_REG_RIP]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.__gregs[_REG_RSP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.__gregs[_REG_RBP]);
|
| +#endif // V8_HOST_ARCH_*
|
| +#elif V8_OS_OPENBSD
|
| +#if V8_HOST_ARCH_IA32
|
| + state.pc = reinterpret_cast<Address>(ucontext->sc_eip);
|
| + state.sp = reinterpret_cast<Address>(ucontext->sc_esp);
|
| + state.fp = reinterpret_cast<Address>(ucontext->sc_ebp);
|
| +#elif V8_HOST_ARCH_X64
|
| + state.pc = reinterpret_cast<Address>(ucontext->sc_rip);
|
| + state.sp = reinterpret_cast<Address>(ucontext->sc_rsp);
|
| + state.fp = reinterpret_cast<Address>(ucontext->sc_rbp);
|
| +#endif // V8_HOST_ARCH_*
|
| +#elif V8_OS_SOLARIS
|
| + state.pc = reinterpret_cast<Address>(mcontext.gregs[REG_PC]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.gregs[REG_SP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.gregs[REG_FP]);
|
| +#elif V8_OS_QNX
|
| +#if V8_HOST_ARCH_IA32
|
| + state.pc = reinterpret_cast<Address>(mcontext.cpu.eip);
|
| + state.sp = reinterpret_cast<Address>(mcontext.cpu.esp);
|
| + state.fp = reinterpret_cast<Address>(mcontext.cpu.ebp);
|
| +#elif V8_HOST_ARCH_ARM
|
| + state.pc = reinterpret_cast<Address>(mcontext.cpu.gpr[ARM_REG_PC]);
|
| + state.sp = reinterpret_cast<Address>(mcontext.cpu.gpr[ARM_REG_SP]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.cpu.gpr[ARM_REG_FP]);
|
| +#endif // V8_HOST_ARCH_*
|
| +#elif V8_OS_AIX
|
| + state.pc = reinterpret_cast<Address>(mcontext.jmp_context.iar);
|
| + state.sp = reinterpret_cast<Address>(mcontext.jmp_context.gpr[1]);
|
| + state.fp = reinterpret_cast<Address>(mcontext.jmp_context.gpr[31]);
|
| +#endif // V8_OS_AIX
|
| +#endif // USE_SIMULATOR
|
| + sampler->SampleStack(state);
|
| +}
|
| +#endif // V8_OS_NACL
|
| +
|
| +#endif // USE_SIGNALS
|
| +
|
| +
|
| +class SamplerThread : public base::Thread {
|
| + public:
|
| + static const int kSamplerThreadStackSize = 64 * KB;
|
| +
|
| + explicit SamplerThread(int interval)
|
| + : Thread(base::Thread::Options("SamplerThread", kSamplerThreadStackSize)),
|
| + interval_(interval) {}
|
| +
|
| + static void SetUp() { if (!mutex_) mutex_ = new base::Mutex(); }
|
| + static void TearDown() { delete mutex_; mutex_ = NULL; }
|
| +
|
| + static void AddActiveSampler(Sampler* sampler) {
|
| + bool need_to_start = false;
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| + if (instance_ == NULL) {
|
| + // Start a thread that will send SIGPROF signal to VM threads,
|
| + // when CPU profiling will be enabled.
|
| + instance_ = new SamplerThread(sampler->interval());
|
| + need_to_start = true;
|
| + }
|
| +
|
| + DCHECK(sampler->IsActive());
|
| + DCHECK(instance_->interval_ == sampler->interval());
|
| +
|
| +#if defined(USE_SIGNALS)
|
| + AddSampler(sampler);
|
| +#else
|
| + DCHECK(!instance_->active_samplers_.Contains(sampler));
|
| + instance_->active_samplers_.Add(sampler);
|
| +#endif // USE_SIGNALS
|
| +
|
| + if (need_to_start) instance_->StartSynchronously();
|
| + }
|
| +
|
| + static void RemoveSampler(Sampler* sampler) {
|
| + SamplerThread* instance_to_remove = NULL;
|
| + {
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| +
|
| + DCHECK(sampler->IsActive() || sampler->IsRegistered());
|
| +#if defined(USE_SIGNALS)
|
| + {
|
| + AtomicGuard atomic_guard(&sampler_list_access_counter_);
|
| + // Remove sampler from map.
|
| + pthread_t thread_id = sampler->platform_data()->vm_tid();
|
| + void* thread_key = ThreadKey(thread_id);
|
| + uint32_t thread_hash = ThreadHash(thread_id);
|
| + HashMap::Entry* entry =
|
| + thread_id_to_samplers_.Get().Lookup(thread_key, thread_hash);
|
| + DCHECK(entry != NULL);
|
| + SamplerList* samplers = reinterpret_cast<SamplerList*>(entry->value);
|
| + samplers->RemoveElement(sampler);
|
| + if (samplers->is_empty()) {
|
| + thread_id_to_samplers_.Pointer()->Remove(thread_key, thread_hash);
|
| + delete samplers;
|
| + }
|
| + if (thread_id_to_samplers_.Get().occupancy() == 0) {
|
| + instance_to_remove = instance_;
|
| + instance_ = NULL;
|
| + }
|
| + }
|
| +#else
|
| + bool removed = instance_->active_samplers_.RemoveElement(sampler);
|
| + DCHECK(removed);
|
| + USE(removed);
|
| +
|
| + // We cannot delete the instance immediately as we need to Join() the
|
| + // thread but we are holding mutex_ and the thread may try to acquire it.
|
| + if (instance_->active_samplers_.is_empty()) {
|
| + instance_to_remove = instance_;
|
| + instance_ = NULL;
|
| + }
|
| +#endif // USE_SIGNALS
|
| + }
|
| +
|
| + if (!instance_to_remove) return;
|
| + instance_to_remove->Join();
|
| + delete instance_to_remove;
|
| + }
|
| +
|
| + // Unlike AddActiveSampler, this method only adds a sampler,
|
| + // but won't start the sampler thread.
|
| + static void RegisterSampler(Sampler* sampler) {
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| +#if defined(USE_SIGNALS)
|
| + AddSampler(sampler);
|
| +#endif // USE_SIGNALS
|
| + }
|
| +
|
| + // Implement Thread::Run().
|
| + virtual void Run() {
|
| + while (true) {
|
| + {
|
| + base::LockGuard<base::Mutex> lock_guard(mutex_);
|
| +#if defined(USE_SIGNALS)
|
| + if (thread_id_to_samplers_.Get().occupancy() == 0) break;
|
| + if (SignalHandler::Installed()) {
|
| + for (HashMap::Entry *p = thread_id_to_samplers_.Get().Start();
|
| + p != NULL; p = thread_id_to_samplers_.Get().Next(p)) {
|
| +#if V8_OS_AIX && V8_TARGET_ARCH_PPC64
|
| + // on AIX64, cannot cast (void *) to pthread_t which is
|
| + // of type unsigned int (4bytes)
|
| + pthread_t thread_id = reinterpret_cast<intptr_t>(p->key);
|
| +#else
|
| + pthread_t thread_id = reinterpret_cast<pthread_t>(p->key);
|
| +#endif
|
| + pthread_kill(thread_id, SIGPROF);
|
| + }
|
| + }
|
| +#else
|
| + if (active_samplers_.is_empty()) break;
|
| + // When CPU profiling is enabled both JavaScript and C++ code is
|
| + // profiled. We must not suspend.
|
| + for (int i = 0; i < active_samplers_.length(); ++i) {
|
| + Sampler* sampler = active_samplers_.at(i);
|
| + if (!sampler->IsProfiling()) continue;
|
| + sampler->DoSample();
|
| + }
|
| +#endif // USE_SIGNALS
|
| + }
|
| + base::OS::Sleep(base::TimeDelta::FromMilliseconds(interval_));
|
| + }
|
| + }
|
| +
|
| + private:
|
| + // Protects the process wide state below.
|
| + static base::Mutex* mutex_;
|
| + static SamplerThread* instance_;
|
| +
|
| + const int interval_;
|
| +
|
| +#if defined(USE_SIGNALS)
|
| + struct HashMapCreateTrait {
|
| + static void Construct(HashMap* allocated_ptr) {
|
| + new (allocated_ptr) HashMap(HashMap::PointersMatch);
|
| + }
|
| + };
|
| + friend class SignalHandler;
|
| + static base::LazyInstance<HashMap, HashMapCreateTrait>::type
|
| + thread_id_to_samplers_;
|
| + static base::AtomicValue<int> sampler_list_access_counter_;
|
| + static void AddSampler(Sampler* sampler) {
|
| + AtomicGuard atomic_guard(&sampler_list_access_counter_);
|
| + // Add sampler into map if needed.
|
| + pthread_t thread_id = sampler->platform_data()->vm_tid();
|
| + HashMap::Entry *entry =
|
| + thread_id_to_samplers_.Pointer()->LookupOrInsert(ThreadKey(thread_id),
|
| + ThreadHash(thread_id));
|
| + if (entry->value == NULL) {
|
| + SamplerList* samplers = new SamplerList();
|
| + samplers->Add(sampler);
|
| + entry->value = samplers;
|
| + } else {
|
| + SamplerList* samplers = reinterpret_cast<SamplerList*>(entry->value);
|
| + if (!samplers->Contains(sampler)) {
|
| + samplers->Add(sampler);
|
| + }
|
| + }
|
| + }
|
| +#else
|
| + SamplerList active_samplers_;
|
| +#endif // USE_SIGNALS
|
| +
|
| + DISALLOW_COPY_AND_ASSIGN(SamplerThread);
|
| +};
|
| +
|
| +
|
| +base::Mutex* SamplerThread::mutex_ = NULL;
|
| +SamplerThread* SamplerThread::instance_ = NULL;
|
| +#if defined(USE_SIGNALS)
|
| +base::LazyInstance<HashMap, SamplerThread::HashMapCreateTrait>::type
|
| + SamplerThread::thread_id_to_samplers_ = LAZY_INSTANCE_INITIALIZER;
|
| +base::AtomicValue<int> SamplerThread::sampler_list_access_counter_(0);
|
| +
|
| +// As Native Client does not support signal handling, profiling is disabled.
|
| +#if !V8_OS_NACL
|
| +void SignalHandler::HandleProfilerSignal(int signal, siginfo_t* info,
|
| + void* context) {
|
| + USE(info);
|
| + if (signal != SIGPROF) return;
|
| + AtomicGuard atomic_guard(&SamplerThread::sampler_list_access_counter_, false);
|
| + if (!atomic_guard.is_success()) return;
|
| + pthread_t thread_id = pthread_self();
|
| + HashMap::Entry* entry =
|
| + SamplerThread::thread_id_to_samplers_.Pointer()->Lookup(
|
| + ThreadKey(thread_id), ThreadHash(thread_id));
|
| + if (entry == NULL)
|
| + return;
|
| + SamplerList* samplers = reinterpret_cast<SamplerList*>(entry->value);
|
| + for (int i = 0; i < samplers->length(); ++i) {
|
| + Sampler* sampler = samplers->at(i);
|
| + CollectSample(context, sampler);
|
| + }
|
| +}
|
| +#endif // !V8_OS_NACL
|
| +#endif // USE_SIGNALs
|
| +
|
| +
|
| +void Sampler::SetUp() {
|
| +#if defined(USE_SIGNALS)
|
| + SignalHandler::SetUp();
|
| +#endif
|
| + SamplerThread::SetUp();
|
| +}
|
| +
|
| +
|
| +void Sampler::TearDown() {
|
| + SamplerThread::TearDown();
|
| +#if defined(USE_SIGNALS)
|
| + SignalHandler::TearDown();
|
| +#endif
|
| +}
|
| +
|
| +Sampler::Sampler(Isolate* isolate, int interval)
|
| + : isolate_(isolate),
|
| + interval_(interval),
|
| + profiling_(false),
|
| + has_processing_thread_(false),
|
| + active_(false),
|
| + registered_(false),
|
| + is_counting_samples_(false),
|
| + js_sample_count_(0),
|
| + external_sample_count_(0) {
|
| + data_ = new PlatformData;
|
| +}
|
| +
|
| +Sampler::~Sampler() {
|
| + DCHECK(!IsActive());
|
| + if (IsRegistered()) {
|
| + SamplerThread::RemoveSampler(this);
|
| + }
|
| + delete data_;
|
| +}
|
| +
|
| +void Sampler::Start() {
|
| + DCHECK(!IsActive());
|
| + SetActive(true);
|
| + SamplerThread::AddActiveSampler(this);
|
| +}
|
| +
|
| +
|
| +void Sampler::Stop() {
|
| + DCHECK(IsActive());
|
| + SamplerThread::RemoveSampler(this);
|
| + SetActive(false);
|
| + SetRegistered(false);
|
| +}
|
| +
|
| +
|
| +void Sampler::IncreaseProfilingDepth() {
|
| + base::NoBarrier_AtomicIncrement(&profiling_, 1);
|
| +#if defined(USE_SIGNALS)
|
| + SignalHandler::IncreaseSamplerCount();
|
| +#endif
|
| +}
|
| +
|
| +
|
| +void Sampler::DecreaseProfilingDepth() {
|
| +#if defined(USE_SIGNALS)
|
| + SignalHandler::DecreaseSamplerCount();
|
| +#endif
|
| + base::NoBarrier_AtomicIncrement(&profiling_, -1);
|
| +}
|
| +
|
| +
|
| +void Sampler::SampleStack(const v8::RegisterState& state) {
|
| + TickSample* sample = isolate_->cpu_profiler()->StartTickSample();
|
| + TickSample sample_obj;
|
| + if (sample == NULL) sample = &sample_obj;
|
| + sample->Init(isolate_, state, TickSample::kIncludeCEntryFrame, true);
|
| + if (is_counting_samples_ && !sample->timestamp.IsNull()) {
|
| + if (sample->state == JS) ++js_sample_count_;
|
| + if (sample->state == EXTERNAL) ++external_sample_count_;
|
| + }
|
| + Tick(sample);
|
| + if (sample != &sample_obj) {
|
| + isolate_->cpu_profiler()->FinishTickSample();
|
| + }
|
| +}
|
| +
|
| +
|
| +#if defined(USE_SIGNALS)
|
| +
|
| +void Sampler::DoSample() {
|
| + if (!SignalHandler::Installed()) return;
|
| + if (!IsActive() && !IsRegistered()) {
|
| + SamplerThread::RegisterSampler(this);
|
| + SetRegistered(true);
|
| + }
|
| + pthread_kill(platform_data()->vm_tid(), SIGPROF);
|
| +}
|
| +
|
| +#elif V8_OS_WIN || V8_OS_CYGWIN
|
| +
|
| +void Sampler::DoSample() {
|
| + HANDLE profiled_thread = platform_data()->profiled_thread();
|
| + if (profiled_thread == NULL) return;
|
| +
|
| + const DWORD kSuspendFailed = static_cast<DWORD>(-1);
|
| + if (SuspendThread(profiled_thread) == kSuspendFailed) return;
|
| +
|
| + // Context used for sampling the register state of the profiled thread.
|
| + CONTEXT context;
|
| + memset(&context, 0, sizeof(context));
|
| + context.ContextFlags = CONTEXT_FULL;
|
| + if (GetThreadContext(profiled_thread, &context) != 0) {
|
| + v8::RegisterState state;
|
| +#if defined(USE_SIMULATOR)
|
| + if (!SimulatorHelper::FillRegisters(isolate(), &state)) {
|
| + ResumeThread(profiled_thread);
|
| + return;
|
| + }
|
| +#else
|
| +#if V8_HOST_ARCH_X64
|
| + state.pc = reinterpret_cast<Address>(context.Rip);
|
| + state.sp = reinterpret_cast<Address>(context.Rsp);
|
| + state.fp = reinterpret_cast<Address>(context.Rbp);
|
| +#else
|
| + state.pc = reinterpret_cast<Address>(context.Eip);
|
| + state.sp = reinterpret_cast<Address>(context.Esp);
|
| + state.fp = reinterpret_cast<Address>(context.Ebp);
|
| +#endif
|
| +#endif // USE_SIMULATOR
|
| + SampleStack(state);
|
| + }
|
| + ResumeThread(profiled_thread);
|
| +}
|
| +
|
| +#endif // USE_SIGNALS
|
| +
|
| +
|
| +} // namespace internal
|
| +} // namespace v8
|
|
|