OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 27 matching lines...) Expand all Loading... |
38 #include "global-handles.h" | 38 #include "global-handles.h" |
39 #include "isolate-inl.h" | 39 #include "isolate-inl.h" |
40 #include "mark-compact.h" | 40 #include "mark-compact.h" |
41 #include "platform.h" | 41 #include "platform.h" |
42 #include "scopeinfo.h" | 42 #include "scopeinfo.h" |
43 | 43 |
44 namespace v8 { | 44 namespace v8 { |
45 namespace internal { | 45 namespace internal { |
46 | 46 |
47 | 47 |
48 // Optimization sampler constants. | |
49 static const int kSamplerFrameCount = 2; | |
50 | |
51 // Constants for statistical profiler. | |
52 static const int kSamplerFrameWeight[kSamplerFrameCount] = { 2, 1 }; | |
53 | |
54 static const int kSamplerTicksBetweenThresholdAdjustment = 32; | |
55 | |
56 static const int kSamplerThresholdInit = 3; | |
57 static const int kSamplerThresholdMin = 1; | |
58 static const int kSamplerThresholdDelta = 1; | |
59 | |
60 static const int kSamplerThresholdSizeFactorInit = 3; | |
61 | |
62 static const int kSizeLimit = 1500; | |
63 | |
64 // Constants for counter based profiler. | |
65 | |
66 // Number of times a function has to be seen on the stack before it is | 48 // Number of times a function has to be seen on the stack before it is |
67 // optimized. | 49 // optimized. |
68 static const int kProfilerTicksBeforeOptimization = 2; | 50 static const int kProfilerTicksBeforeOptimization = 2; |
69 // If the function optimization was disabled due to high deoptimization count, | 51 // If the function optimization was disabled due to high deoptimization count, |
70 // but the function is hot and has been seen on the stack this number of times, | 52 // but the function is hot and has been seen on the stack this number of times, |
71 // then we try to reenable optimization for this function. | 53 // then we try to reenable optimization for this function. |
72 static const int kProfilerTicksBeforeReenablingOptimization = 250; | 54 static const int kProfilerTicksBeforeReenablingOptimization = 250; |
73 // If a function does not have enough type info (according to | 55 // If a function does not have enough type info (according to |
74 // FLAG_type_info_threshold), but has seen a huge number of ticks, | 56 // FLAG_type_info_threshold), but has seen a huge number of ticks, |
75 // optimize it as it is. | 57 // optimize it as it is. |
(...skipping 11 matching lines...) Expand all Loading... |
87 3 * FullCodeGenerator::kCodeSizeMultiplier; | 69 3 * FullCodeGenerator::kCodeSizeMultiplier; |
88 | 70 |
89 // Maximum size in bytes of generated code for a function to be optimized | 71 // Maximum size in bytes of generated code for a function to be optimized |
90 // the very first time it is seen on the stack. | 72 // the very first time it is seen on the stack. |
91 static const int kMaxSizeEarlyOpt = | 73 static const int kMaxSizeEarlyOpt = |
92 5 * FullCodeGenerator::kCodeSizeMultiplier; | 74 5 * FullCodeGenerator::kCodeSizeMultiplier; |
93 | 75 |
94 | 76 |
95 RuntimeProfiler::RuntimeProfiler(Isolate* isolate) | 77 RuntimeProfiler::RuntimeProfiler(Isolate* isolate) |
96 : isolate_(isolate), | 78 : isolate_(isolate), |
97 sampler_threshold_(kSamplerThresholdInit), | 79 any_ic_changed_(false) { |
98 sampler_threshold_size_factor_(kSamplerThresholdSizeFactorInit), | |
99 sampler_ticks_until_threshold_adjustment_( | |
100 kSamplerTicksBetweenThresholdAdjustment), | |
101 sampler_window_position_(0), | |
102 any_ic_changed_(false), | |
103 code_generated_(false) { | |
104 ClearSampleBuffer(); | |
105 } | 80 } |
106 | 81 |
107 | 82 |
108 static void GetICCounts(Code* shared_code, | 83 static void GetICCounts(Code* shared_code, |
109 int* ic_with_type_info_count, | 84 int* ic_with_type_info_count, |
110 int* ic_total_count, | 85 int* ic_total_count, |
111 int* percentage) { | 86 int* percentage) { |
112 *ic_total_count = 0; | 87 *ic_total_count = 0; |
113 *ic_with_type_info_count = 0; | 88 *ic_with_type_info_count = 0; |
114 Object* raw_info = shared_code->type_feedback_info(); | 89 Object* raw_info = shared_code->type_feedback_info(); |
(...skipping 67 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
182 if (FLAG_trace_osr) { | 157 if (FLAG_trace_osr) { |
183 PrintF("[OSR - patching back edges in "); | 158 PrintF("[OSR - patching back edges in "); |
184 function->PrintName(); | 159 function->PrintName(); |
185 PrintF("]\n"); | 160 PrintF("]\n"); |
186 } | 161 } |
187 | 162 |
188 BackEdgeTable::Patch(isolate_, shared->code()); | 163 BackEdgeTable::Patch(isolate_, shared->code()); |
189 } | 164 } |
190 | 165 |
191 | 166 |
192 void RuntimeProfiler::ClearSampleBuffer() { | |
193 memset(sampler_window_, 0, sizeof(sampler_window_)); | |
194 memset(sampler_window_weight_, 0, sizeof(sampler_window_weight_)); | |
195 } | |
196 | |
197 | |
198 int RuntimeProfiler::LookupSample(JSFunction* function) { | |
199 int weight = 0; | |
200 for (int i = 0; i < kSamplerWindowSize; i++) { | |
201 Object* sample = sampler_window_[i]; | |
202 if (sample != NULL) { | |
203 bool fits = FLAG_lookup_sample_by_shared | |
204 ? (function->shared() == JSFunction::cast(sample)->shared()) | |
205 : (function == JSFunction::cast(sample)); | |
206 if (fits) { | |
207 weight += sampler_window_weight_[i]; | |
208 } | |
209 } | |
210 } | |
211 return weight; | |
212 } | |
213 | |
214 | |
215 void RuntimeProfiler::AddSample(JSFunction* function, int weight) { | |
216 ASSERT(IsPowerOf2(kSamplerWindowSize)); | |
217 sampler_window_[sampler_window_position_] = function; | |
218 sampler_window_weight_[sampler_window_position_] = weight; | |
219 sampler_window_position_ = (sampler_window_position_ + 1) & | |
220 (kSamplerWindowSize - 1); | |
221 } | |
222 | |
223 | |
224 void RuntimeProfiler::OptimizeNow() { | 167 void RuntimeProfiler::OptimizeNow() { |
225 HandleScope scope(isolate_); | 168 HandleScope scope(isolate_); |
226 | 169 |
227 if (isolate_->DebuggerHasBreakPoints()) return; | 170 if (isolate_->DebuggerHasBreakPoints()) return; |
228 | 171 |
229 DisallowHeapAllocation no_gc; | 172 DisallowHeapAllocation no_gc; |
230 | 173 |
231 // Run through the JavaScript frames and collect them. If we already | 174 // Run through the JavaScript frames and collect them. If we already |
232 // have a sample of the function, we mark it for optimizations | 175 // have a sample of the function, we mark it for optimizations |
233 // (eagerly or lazily). | 176 // (eagerly or lazily). |
234 JSFunction* samples[kSamplerFrameCount]; | |
235 int sample_count = 0; | |
236 int frame_count = 0; | 177 int frame_count = 0; |
237 int frame_count_limit = FLAG_watch_ic_patching ? FLAG_frame_count | 178 int frame_count_limit = FLAG_frame_count; |
238 : kSamplerFrameCount; | |
239 for (JavaScriptFrameIterator it(isolate_); | 179 for (JavaScriptFrameIterator it(isolate_); |
240 frame_count++ < frame_count_limit && !it.done(); | 180 frame_count++ < frame_count_limit && !it.done(); |
241 it.Advance()) { | 181 it.Advance()) { |
242 JavaScriptFrame* frame = it.frame(); | 182 JavaScriptFrame* frame = it.frame(); |
243 JSFunction* function = frame->function(); | 183 JSFunction* function = frame->function(); |
244 | 184 |
245 if (!FLAG_watch_ic_patching) { | |
246 // Adjust threshold each time we have processed | |
247 // a certain number of ticks. | |
248 if (sampler_ticks_until_threshold_adjustment_ > 0) { | |
249 sampler_ticks_until_threshold_adjustment_--; | |
250 if (sampler_ticks_until_threshold_adjustment_ <= 0) { | |
251 // If the threshold is not already at the minimum | |
252 // modify and reset the ticks until next adjustment. | |
253 if (sampler_threshold_ > kSamplerThresholdMin) { | |
254 sampler_threshold_ -= kSamplerThresholdDelta; | |
255 sampler_ticks_until_threshold_adjustment_ = | |
256 kSamplerTicksBetweenThresholdAdjustment; | |
257 } | |
258 } | |
259 } | |
260 } | |
261 | |
262 SharedFunctionInfo* shared = function->shared(); | 185 SharedFunctionInfo* shared = function->shared(); |
263 Code* shared_code = shared->code(); | 186 Code* shared_code = shared->code(); |
264 | 187 |
265 if (shared_code->kind() != Code::FUNCTION) continue; | 188 if (shared_code->kind() != Code::FUNCTION) continue; |
266 if (function->IsInRecompileQueue()) continue; | 189 if (function->IsInRecompileQueue()) continue; |
267 | 190 |
268 if (FLAG_always_osr && | 191 if (FLAG_always_osr && |
269 shared_code->allow_osr_at_loop_nesting_level() == 0) { | 192 shared_code->allow_osr_at_loop_nesting_level() == 0) { |
270 // Testing mode: always try an OSR compile for every function. | 193 // Testing mode: always try an OSR compile for every function. |
271 for (int i = 0; i < Code::kMaxLoopNestingMarker; i++) { | 194 for (int i = 0; i < Code::kMaxLoopNestingMarker; i++) { |
(...skipping 43 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
315 shared_code->set_profiler_ticks(0); | 238 shared_code->set_profiler_ticks(0); |
316 shared->TryReenableOptimization(); | 239 shared->TryReenableOptimization(); |
317 } else { | 240 } else { |
318 shared_code->set_profiler_ticks(ticks + 1); | 241 shared_code->set_profiler_ticks(ticks + 1); |
319 } | 242 } |
320 } | 243 } |
321 continue; | 244 continue; |
322 } | 245 } |
323 if (!function->IsOptimizable()) continue; | 246 if (!function->IsOptimizable()) continue; |
324 | 247 |
325 if (FLAG_watch_ic_patching) { | 248 int ticks = shared_code->profiler_ticks(); |
326 int ticks = shared_code->profiler_ticks(); | |
327 | 249 |
328 if (ticks >= kProfilerTicksBeforeOptimization) { | 250 if (ticks >= kProfilerTicksBeforeOptimization) { |
329 int typeinfo, total, percentage; | 251 int typeinfo, total, percentage; |
330 GetICCounts(shared_code, &typeinfo, &total, &percentage); | 252 GetICCounts(shared_code, &typeinfo, &total, &percentage); |
331 if (percentage >= FLAG_type_info_threshold) { | 253 if (percentage >= FLAG_type_info_threshold) { |
332 // If this particular function hasn't had any ICs patched for enough | 254 // If this particular function hasn't had any ICs patched for enough |
333 // ticks, optimize it now. | 255 // ticks, optimize it now. |
334 Optimize(function, "hot and stable"); | 256 Optimize(function, "hot and stable"); |
335 } else if (ticks >= kTicksWhenNotEnoughTypeInfo) { | 257 } else if (ticks >= kTicksWhenNotEnoughTypeInfo) { |
336 Optimize(function, "not much type info but very hot"); | 258 Optimize(function, "not much type info but very hot"); |
337 } else { | |
338 shared_code->set_profiler_ticks(ticks + 1); | |
339 if (FLAG_trace_opt_verbose) { | |
340 PrintF("[not yet optimizing "); | |
341 function->PrintName(); | |
342 PrintF(", not enough type info: %d/%d (%d%%)]\n", | |
343 typeinfo, total, percentage); | |
344 } | |
345 } | |
346 } else if (!any_ic_changed_ && | |
347 shared_code->instruction_size() < kMaxSizeEarlyOpt) { | |
348 // If no IC was patched since the last tick and this function is very | |
349 // small, optimistically optimize it now. | |
350 Optimize(function, "small function"); | |
351 } else { | 259 } else { |
352 shared_code->set_profiler_ticks(ticks + 1); | 260 shared_code->set_profiler_ticks(ticks + 1); |
| 261 if (FLAG_trace_opt_verbose) { |
| 262 PrintF("[not yet optimizing "); |
| 263 function->PrintName(); |
| 264 PrintF(", not enough type info: %d/%d (%d%%)]\n", |
| 265 typeinfo, total, percentage); |
| 266 } |
353 } | 267 } |
354 } else { // !FLAG_watch_ic_patching | 268 } else if (!any_ic_changed_ && |
355 samples[sample_count++] = function; | 269 shared_code->instruction_size() < kMaxSizeEarlyOpt) { |
356 | 270 // If no IC was patched since the last tick and this function is very |
357 int function_size = function->shared()->SourceSize(); | 271 // small, optimistically optimize it now. |
358 int threshold_size_factor = (function_size > kSizeLimit) | 272 Optimize(function, "small function"); |
359 ? sampler_threshold_size_factor_ | 273 } else { |
360 : 1; | 274 shared_code->set_profiler_ticks(ticks + 1); |
361 | |
362 int threshold = sampler_threshold_ * threshold_size_factor; | |
363 | |
364 if (LookupSample(function) >= threshold) { | |
365 Optimize(function, "sampler window lookup"); | |
366 } | |
367 } | 275 } |
368 } | 276 } |
369 if (FLAG_watch_ic_patching) { | 277 any_ic_changed_ = false; |
370 any_ic_changed_ = false; | |
371 } else { // !FLAG_watch_ic_patching | |
372 // Add the collected functions as samples. It's important not to do | |
373 // this as part of collecting them because this will interfere with | |
374 // the sample lookup in case of recursive functions. | |
375 for (int i = 0; i < sample_count; i++) { | |
376 AddSample(samples[i], kSamplerFrameWeight[i]); | |
377 } | |
378 } | |
379 } | 278 } |
380 | 279 |
381 | 280 |
382 void RuntimeProfiler::SetUp() { | |
383 if (!FLAG_watch_ic_patching) { | |
384 ClearSampleBuffer(); | |
385 } | |
386 } | |
387 | |
388 | |
389 void RuntimeProfiler::Reset() { | |
390 if (!FLAG_watch_ic_patching) { | |
391 sampler_threshold_ = kSamplerThresholdInit; | |
392 sampler_threshold_size_factor_ = kSamplerThresholdSizeFactorInit; | |
393 sampler_ticks_until_threshold_adjustment_ = | |
394 kSamplerTicksBetweenThresholdAdjustment; | |
395 } | |
396 } | |
397 | |
398 | |
399 void RuntimeProfiler::TearDown() { | |
400 // Nothing to do. | |
401 } | |
402 | |
403 | |
404 // Update the pointers in the sampler window after a GC. | |
405 void RuntimeProfiler::UpdateSamplesAfterScavenge() { | |
406 for (int i = 0; i < kSamplerWindowSize; i++) { | |
407 Object* function = sampler_window_[i]; | |
408 if (function != NULL && isolate_->heap()->InNewSpace(function)) { | |
409 MapWord map_word = HeapObject::cast(function)->map_word(); | |
410 if (map_word.IsForwardingAddress()) { | |
411 sampler_window_[i] = map_word.ToForwardingAddress(); | |
412 } else { | |
413 sampler_window_[i] = NULL; | |
414 } | |
415 } | |
416 } | |
417 } | |
418 | |
419 | |
420 void RuntimeProfiler::RemoveDeadSamples() { | |
421 for (int i = 0; i < kSamplerWindowSize; i++) { | |
422 Object* function = sampler_window_[i]; | |
423 if (function != NULL && | |
424 !Marking::MarkBitFrom(HeapObject::cast(function)).Get()) { | |
425 sampler_window_[i] = NULL; | |
426 } | |
427 } | |
428 } | |
429 | |
430 | |
431 void RuntimeProfiler::UpdateSamplesAfterCompact(ObjectVisitor* visitor) { | |
432 for (int i = 0; i < kSamplerWindowSize; i++) { | |
433 visitor->VisitPointer(&sampler_window_[i]); | |
434 } | |
435 } | |
436 | |
437 | |
438 } } // namespace v8::internal | 281 } } // namespace v8::internal |
OLD | NEW |