| OLD | NEW |
| 1 // Copyright 2010 the V8 project authors. All rights reserved. | 1 // Copyright 2010 the V8 project authors. All rights reserved. |
| 2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
| 3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
| 4 // met: | 4 // met: |
| 5 // | 5 // |
| 6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
| 7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
| 8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
| 9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
| 10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
| (...skipping 85 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 96 public: | 96 public: |
| 97 Address alias; | 97 Address alias; |
| 98 Address start; | 98 Address start; |
| 99 | 99 |
| 100 INLINE(void UpdateCodeMap(CodeMap* code_map)); | 100 INLINE(void UpdateCodeMap(CodeMap* code_map)); |
| 101 }; | 101 }; |
| 102 | 102 |
| 103 | 103 |
| 104 class TickSampleEventRecord BASE_EMBEDDED { | 104 class TickSampleEventRecord BASE_EMBEDDED { |
| 105 public: | 105 public: |
| 106 // In memory, the first machine word of a TickSampleEventRecord will be the | 106 // The first machine word of a TickSampleEventRecord must not ever |
| 107 // first entry of TickSample, that is -- the VM state field. | 107 // become equal to SamplingCircularQueue::kClear. As both order and |
| 108 // TickSample is put first, because 'order' can become equal to | 108 // TickSample's first field are not reliable in this sense (order |
| 109 // SamplingCircularQueue::kClear, while VM state can't, see | 109 // can overflow, TickSample can have all fields reset), we are |
| 110 // the definition of 'enum StateTag'. | 110 // forced to use an artificial filler field. |
| 111 int filler; |
| 112 unsigned order; |
| 111 TickSample sample; | 113 TickSample sample; |
| 112 unsigned order; | |
| 113 | 114 |
| 114 static TickSampleEventRecord* cast(void* value) { | 115 static TickSampleEventRecord* cast(void* value) { |
| 115 return reinterpret_cast<TickSampleEventRecord*>(value); | 116 return reinterpret_cast<TickSampleEventRecord*>(value); |
| 116 } | 117 } |
| 117 | 118 |
| 119 INLINE(static TickSampleEventRecord* init(void* value)); |
| 120 |
| 118 private: | 121 private: |
| 119 DISALLOW_IMPLICIT_CONSTRUCTORS(TickSampleEventRecord); | 122 DISALLOW_IMPLICIT_CONSTRUCTORS(TickSampleEventRecord); |
| 120 }; | 123 }; |
| 121 | 124 |
| 122 | 125 |
| 123 // This class implements both the profile events processor thread and | 126 // This class implements both the profile events processor thread and |
| 124 // methods called by event producers: VM and stack sampler threads. | 127 // methods called by event producers: VM and stack sampler threads. |
| 125 class ProfilerEventsProcessor : public Thread { | 128 class ProfilerEventsProcessor : public Thread { |
| 126 public: | 129 public: |
| 127 explicit ProfilerEventsProcessor(ProfileGenerator* generator); | 130 explicit ProfilerEventsProcessor(ProfileGenerator* generator); |
| (...skipping 121 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 249 void StartCollectingProfile(String* title); | 252 void StartCollectingProfile(String* title); |
| 250 void StartProcessorIfNotStarted(); | 253 void StartProcessorIfNotStarted(); |
| 251 CpuProfile* StopCollectingProfile(const char* title); | 254 CpuProfile* StopCollectingProfile(const char* title); |
| 252 CpuProfile* StopCollectingProfile(String* title); | 255 CpuProfile* StopCollectingProfile(String* title); |
| 253 void StopProcessorIfLastProfile(); | 256 void StopProcessorIfLastProfile(); |
| 254 | 257 |
| 255 CpuProfilesCollection* profiles_; | 258 CpuProfilesCollection* profiles_; |
| 256 unsigned next_profile_uid_; | 259 unsigned next_profile_uid_; |
| 257 ProfileGenerator* generator_; | 260 ProfileGenerator* generator_; |
| 258 ProfilerEventsProcessor* processor_; | 261 ProfilerEventsProcessor* processor_; |
| 262 int saved_logging_nesting_; |
| 259 | 263 |
| 260 static CpuProfiler* singleton_; | 264 static CpuProfiler* singleton_; |
| 261 | 265 |
| 262 #else | 266 #else |
| 263 static INLINE(bool is_profiling()) { return false; } | 267 static INLINE(bool is_profiling()) { return false; } |
| 264 #endif // ENABLE_CPP_PROFILES_PROCESSOR | 268 #endif // ENABLE_CPP_PROFILES_PROCESSOR |
| 265 | 269 |
| 266 private: | 270 private: |
| 267 DISALLOW_COPY_AND_ASSIGN(CpuProfiler); | 271 DISALLOW_COPY_AND_ASSIGN(CpuProfiler); |
| 268 }; | 272 }; |
| 269 | 273 |
| 270 } } // namespace v8::internal | 274 } } // namespace v8::internal |
| 271 | 275 |
| 272 | 276 |
| 273 #endif // V8_CPU_PROFILER_H_ | 277 #endif // V8_CPU_PROFILER_H_ |
| OLD | NEW |