| Index: tools/telemetry/telemetry/benchmark.py
|
| diff --git a/tools/telemetry/telemetry/benchmark.py b/tools/telemetry/telemetry/benchmark.py
|
| index e10b518f7157c28cba20331464bf57ef1198f1fc..2a8ae336c1febe25fdfcf462cf8a81f6598ae7b9 100644
|
| --- a/tools/telemetry/telemetry/benchmark.py
|
| +++ b/tools/telemetry/telemetry/benchmark.py
|
| @@ -2,6 +2,7 @@
|
| # Use of this source code is governed by a BSD-style license that can be
|
| # found in the LICENSE file.
|
|
|
| +import json
|
| import logging
|
| import optparse
|
| import os
|
| @@ -17,6 +18,7 @@ from telemetry.page import page_runner
|
| from telemetry.page import page_set
|
| from telemetry.page import page_test
|
| from telemetry.page import test_expectations
|
| +from telemetry.results import chart_json
|
| from telemetry.results import results_options
|
| from telemetry.util import cloud_storage
|
|
|
| @@ -89,13 +91,20 @@ class Benchmark(command_line.Command):
|
|
|
| self._DownloadGeneratedProfileArchive(finder_options)
|
|
|
| - results = results_options.CreateResults(self.GetMetadata(), finder_options)
|
| + benchmark_metadata = self.GetMetadata()
|
| + results = results_options.CreateResults(benchmark_metadata, finder_options)
|
| try:
|
| page_runner.Run(pt, ps, expectations, finder_options, results)
|
| except page_test.TestNotSupportedOnPlatformFailure as failure:
|
| logging.warning(str(failure))
|
|
|
| - results.PrintSummary()
|
| + if finder_options.chartjson:
|
| + print json.dumps(chart_json.ResultsAsChartDict(
|
| + benchmark_metadata,
|
| + results.all_page_specific_values,
|
| + results.all_summary_values))
|
| + else:
|
| + results.PrintSummary()
|
| return len(results.failures)
|
|
|
| def _DownloadGeneratedProfileArchive(self, options):
|
|
|