| Index: tools/perf/benchmarks/canvasmark.py
|
| diff --git a/tools/perf/benchmarks/canvasmark.py b/tools/perf/benchmarks/canvasmark.py
|
| index 55a219b338c60d4fa30c70fd8e6589683cfb98a7..dacac775aeeb54dc9b45b0bcc4253509dc8e83c5 100644
|
| --- a/tools/perf/benchmarks/canvasmark.py
|
| +++ b/tools/perf/benchmarks/canvasmark.py
|
| @@ -11,13 +11,13 @@ fills, shadows and text functions.
|
|
|
| import os
|
|
|
| +from measurements import PageTestMeasurement
|
| from telemetry import benchmark
|
| from telemetry.page import page_set
|
| -from telemetry.page import page_test
|
| from telemetry.value import scalar
|
|
|
|
|
| -class _CanvasMarkMeasurement(page_test.PageTest):
|
| +class _CanvasMarkMeasurement(PageTestMeasurement):
|
|
|
| def WillNavigateToPage(self, page, tab):
|
| page.script_to_evaluate_on_commit = """
|
| @@ -29,7 +29,7 @@ class _CanvasMarkMeasurement(page_test.PageTest):
|
| }
|
| """
|
|
|
| - def ValidateAndMeasurePage(self, _, tab, results):
|
| + def ValidateAndMeasurePage(self, page, tab, results):
|
| tab.WaitForJavaScriptExpression('__results.length == 8', 300)
|
| results_log = tab.EvaluateJavaScript('__results')
|
| total = 0
|
| @@ -48,6 +48,8 @@ class _CanvasMarkMeasurement(page_test.PageTest):
|
| results.AddValue(scalar.ScalarValue(
|
| results.current_page, 'Score', 'score', total))
|
|
|
| + super(_CanvasMarkMeasurement, self).ValidateAndMeasurePage(
|
| + page, tab, results)
|
|
|
| @benchmark.Disabled
|
| class CanvasMark(benchmark.Benchmark):
|
|
|