Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(614)

Unified Diff: tools/perf/metrics/gpu_timeline.py

Issue 854833003: Added GPU performance metrics. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Removed gpu_device check in gpu_times_unittest Created 5 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « tools/perf/measurements/gpu_times_unittest.py ('k') | tools/perf/metrics/gpu_timeline_unittest.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: tools/perf/metrics/gpu_timeline.py
diff --git a/tools/perf/metrics/gpu_timeline.py b/tools/perf/metrics/gpu_timeline.py
new file mode 100644
index 0000000000000000000000000000000000000000..920079fb6bbee089db960a72e526acf9b899e5a1
--- /dev/null
+++ b/tools/perf/metrics/gpu_timeline.py
@@ -0,0 +1,209 @@
+# Copyright 2015 The Chromium Authors. All rights reserved.
+# Use of this source code is governed by a BSD-style license that can be
+# found in the LICENSE file.
+import collections
+import math
+
+from telemetry.timeline import async_slice as async_slice_module
+from telemetry.timeline import slice as slice_module
+from telemetry.value import scalar
+from telemetry.web_perf.metrics import timeline_based_metric
+
+TOPLEVEL_GL_CATEGORY = 'gpu_toplevel'
+TOPLEVEL_SERVICE_CATEGORY = 'disabled-by-default-gpu.service'
+TOPLEVEL_DEVICE_CATEGORY = 'disabled-by-default-gpu.device'
+
+FRAME_END_MARKER = ('gpu', 'GLES2DecoderImpl::DoSwapBuffers')
epenner 2015/01/16 01:17:59 Two things on this trace: First, is there no GPU-
vmiura 2015/01/16 01:41:15 We should be able to add a "DoSwapBuffers" gpu tra
David Yen 2015/01/16 22:58:04 Done here: https://codereview.chromium.org/7997530
David Yen 2015/01/20 19:37:26 Done.
+
+TRACKED_NAMES = { 'RenderCompositor': 'render_compositor',
+ 'Compositor': 'compositor' }
+
+GPU_SERVICE_DEVICE_VARIANCE = 5
epenner 2015/01/16 01:17:59 See below, I'm not sure this threshold is really n
vmiura 2015/01/16 01:41:14 Perhaps if we have the "swap" trace on the GPU we
David Yen 2015/01/16 22:58:04 Yes, once the Swap trace is plumbed through we won
David Yen 2015/01/20 19:37:26 Done.
+
+
+class GPUTimelineMetric(timeline_based_metric.TimelineBasedMetric):
+ """Computes GPU based metrics."""
+
+ def __init__(self):
+ super(GPUTimelineMetric, self).__init__()
+
+ def AddResults(self, model, _, interaction_records, results):
+ service_times = self._CalculateGPUTimelineData(model)
+ for name, durations in service_times.iteritems():
+ count = len(durations)
+ avg = 0.0
+ stddev = 0.0
+ maximum = 0.0
+ if count:
+ avg = sum(durations) / count
+ stddev = math.sqrt(sum((d - avg) ** 2 for d in durations) / count)
+ maximum = max(durations)
+
+ results.AddValue(scalar.ScalarValue(results.current_page,
+ name + '_max', 'ms', maximum))
+ results.AddValue(scalar.ScalarValue(results.current_page,
+ name + '_avg', 'ms', avg))
+ results.AddValue(scalar.ScalarValue(results.current_page,
+ name + '_stddev', 'ms', stddev))
+
+ def _CalculateGPUTimelineData(self, model):
+ """Uses the model and calculates the times for various values for each
+ frame. The return value will be a dictionary of the following format:
+ {
+ EVENT_NAME1: [FRAME0_TIME, FRAME1_TIME...etc.],
+ EVENT_NAME2: [FRAME0_TIME, FRAME1_TIME...etc.],
+ }
+
+ Event Names:
+ total_frame - Total time each frame is calculated to be.
+ total_gpu_service: Total time the GPU service took per frame.
+ total_gpu_device: Total time the GPU device took per frame.
+ TRACKED_NAMES_service: Using the TRACKED_NAMES dictionary, we include
epenner 2015/01/16 01:17:59 It took me a minute to parse what these mean. Does
vmiura 2015/01/16 01:41:14 Currently the 'gpu.service' traces are using the t
David Yen 2015/01/16 22:58:04 I've changed the Traces to use normal traces inste
David Yen 2015/01/20 19:37:26 Done.
+ service traces per frame for the tracked name.
+ TRACKED_NAMES_device: Using the TRACKED_NAMES dictionary, we include
+ device traces per frame for the tracked name.
+ """
+ service_events = []
+ device_events = []
+ buffer_swap_events = []
+
+ for event in model.IterAllEvents():
+ if isinstance(event, slice_module.Slice):
+ if (event.category, event.name) == FRAME_END_MARKER:
+ buffer_swap_events.append(event)
+ elif isinstance(event, async_slice_module.AsyncSlice):
+ if event.thread_start:
epenner 2015/01/16 01:17:59 This is the only use of thread_start, are you sure
David Yen 2015/01/16 22:58:04 I was going to look into this later, but the Async
+ if event.args.get('gl_category', None) == TOPLEVEL_GL_CATEGORY:
+ if event.category == TOPLEVEL_SERVICE_CATEGORY:
+ service_events.append(event)
+ elif event.category == TOPLEVEL_DEVICE_CATEGORY:
+ device_events.append(event)
+
+ # Some platforms do not support GPU device tracing, fill in empty values.
+ no_device_traces = False
+ if service_events and not device_events:
+ device_events = [async_slice_module.AsyncSlice(TOPLEVEL_DEVICE_CATEGORY,
+ event.name, 0)
+ for event in service_events]
+ no_device_traces = True
+
+ # Allow some variance in the number of service and device events, depending
+ # on when the tracing stopped the device trace could not have come back yet.
epenner 2015/01/16 01:17:59 Another reason for this could be that the device e
David Yen 2015/01/16 22:58:04 This is no longer relevant and will be removed onc
David Yen 2015/01/20 19:37:26 Done.
+ if len(service_events) > len(device_events):
+ event_difference = len(service_events) - len(device_events)
+ if event_difference <= GPU_SERVICE_DEVICE_VARIANCE:
+ service_events = service_events[:-event_difference]
+
+ # Group together GPU events and validate that the markers match.
+ assert len(service_events) == len(device_events), (
epenner 2015/01/16 01:17:59 It seems like we should either have a hard '==' in
David Yen 2015/01/16 22:58:04 Before we were using the BufferSwap trace that onl
David Yen 2015/01/20 19:37:26 Done.
+ 'Mismatching number of GPU Service (%s) and Device events (%s).' %
+ (len(service_events), len(device_events)))
+
+ service_events_dict = collections.defaultdict(list)
+ for event in service_events:
+ service_events_dict[event.name].append(event)
+
+ device_events_dict = collections.defaultdict(list)
+ for event in device_events:
+ device_events_dict[event.name].append(event)
+
+ assert set(service_events_dict.keys()) == set(device_events_dict.keys()), (
+ 'Mismatching event names between GPU Service and Device events.')
+
+ gpu_events = []
+ for event_name in service_events_dict:
+ service_events_list = service_events_dict[event_name]
+ device_events_list = device_events_dict[event_name]
+ assert len(service_events_list) == len(device_events_list), (
+ 'GPU service event (%s) does not correspond with all device events.' %
+ (event_name))
+
+ gpu_events.extend(zip(service_events_list, device_events_list))
+
+ gpu_events.sort(key=lambda events: events[0].start)
+
+ # Utilize Swap Buffer event to separate out gpu events by frames.
+ gpu_events_by_frame = []
+ gpu_event_iter = iter(gpu_events)
+ current_frame = []
+ for buffer_swap_event in buffer_swap_events:
epenner 2015/01/16 01:17:59 Couple minor things: Firstly, this looks to be M*
David Yen 2015/01/16 22:58:04 Before I only had SwapBuffer traces on the CPU sid
David Yen 2015/01/20 19:37:26 This is all removed now.
epenner 2015/01/27 22:06:38 I also just noticed that you calculate stats on th
+ for gpu_event in gpu_event_iter:
+ service_event, device_event = gpu_event
+ if service_event.end <= buffer_swap_event.end:
+ current_frame.append(gpu_event)
+ else:
+ if current_frame:
+ gpu_events_by_frame.append(current_frame)
+ current_frame = [gpu_event]
+ break
+
+ current_frame.extend([gpu_event for gpu_event in gpu_event_iter])
+ if current_frame:
+ gpu_events_by_frame.append(current_frame)
+
+ # Calculate service times that we care about.
+ total_frame_times = []
+ gpu_service_times = []
+ gpu_device_times = []
+ tracked_times = {}
+
+ tracked_times.update(dict([(value + "_service", [])
+ for value in TRACKED_NAMES.itervalues()]))
+ tracked_times.update(dict([(value + "_device", [])
+ for value in TRACKED_NAMES.itervalues()]))
+
+ if gpu_events:
+ first_service_event, _ = gpu_events[0]
+ prev_frame_end = first_service_event.start
+ else:
+ prev_frame_end = 0
+
+ for frame_gpu_events in gpu_events_by_frame:
+ last_service_in_frame, _ = frame_gpu_events[-1]
+
+ total_frame_time = last_service_in_frame.end - prev_frame_end
+ prev_frame_end = last_service_in_frame.end
+
+ total_gpu_service_time = 0
+ total_gpu_device_time = 0
+ tracked_markers = collections.defaultdict(lambda : 0)
+ for service_event, device_event in frame_gpu_events:
+ service_time = service_event.end - service_event.start
+ device_time = device_event.end - device_event.start
+ total_gpu_service_time += service_time
+ total_gpu_device_time += device_time
+
+ base_name = service_event.name
+ dash_index = base_name.rfind('-')
+ if dash_index != -1:
+ base_name = base_name[:dash_index]
+
+ tracked_name = TRACKED_NAMES.get(base_name, None)
+ if tracked_name:
+ tracked_markers[tracked_name + '_service'] += service_time
+ tracked_markers[tracked_name + '_device'] += device_time
+
+ total_frame_times.append(total_frame_time)
+ gpu_service_times.append(total_gpu_service_time)
+ gpu_device_times.append(total_gpu_device_time)
+
+ for tracked_name in TRACKED_NAMES.values():
+ service_name = tracked_name + '_service'
+ device_name = tracked_name + '_device'
+ tracked_times[service_name].append(tracked_markers[service_name])
+ tracked_times[device_name].append(tracked_markers[device_name])
+
+ # Create the service times dictionary.
+ service_times = { 'total_frame': total_frame_times,
+ 'total_gpu_service': gpu_service_times,
+ 'total_gpu_device': gpu_device_times }
+ service_times.update(tracked_times)
+
+ # Remove device metrics if no device traces were found.
+ if no_device_traces:
+ for device_name in [name
+ for name in service_times.iterkeys()
+ if name.endswith('_device')]:
+ service_times.pop(device_name)
+
+ return service_times
« no previous file with comments | « tools/perf/measurements/gpu_times_unittest.py ('k') | tools/perf/metrics/gpu_timeline_unittest.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698