Index: tools/perf/benchmarks/canvasmark.py |
diff --git a/tools/perf/benchmarks/canvasmark.py b/tools/perf/benchmarks/canvasmark.py |
index 55a219b338c60d4fa30c70fd8e6589683cfb98a7..dacac775aeeb54dc9b45b0bcc4253509dc8e83c5 100644 |
--- a/tools/perf/benchmarks/canvasmark.py |
+++ b/tools/perf/benchmarks/canvasmark.py |
@@ -11,13 +11,13 @@ fills, shadows and text functions. |
import os |
+from measurements import PageTestMeasurement |
from telemetry import benchmark |
from telemetry.page import page_set |
-from telemetry.page import page_test |
from telemetry.value import scalar |
-class _CanvasMarkMeasurement(page_test.PageTest): |
+class _CanvasMarkMeasurement(PageTestMeasurement): |
def WillNavigateToPage(self, page, tab): |
page.script_to_evaluate_on_commit = """ |
@@ -29,7 +29,7 @@ class _CanvasMarkMeasurement(page_test.PageTest): |
} |
""" |
- def ValidateAndMeasurePage(self, _, tab, results): |
+ def ValidateAndMeasurePage(self, page, tab, results): |
tab.WaitForJavaScriptExpression('__results.length == 8', 300) |
results_log = tab.EvaluateJavaScript('__results') |
total = 0 |
@@ -48,6 +48,8 @@ class _CanvasMarkMeasurement(page_test.PageTest): |
results.AddValue(scalar.ScalarValue( |
results.current_page, 'Score', 'score', total)) |
+ super(_CanvasMarkMeasurement, self).ValidateAndMeasurePage( |
+ page, tab, results) |
@benchmark.Disabled |
class CanvasMark(benchmark.Benchmark): |