| Index: telemetry/telemetry/benchmark_runner.py
|
| diff --git a/telemetry/telemetry/benchmark_runner.py b/telemetry/telemetry/benchmark_runner.py
|
| index d7116194848c9cf51aebd1528703fcae02ce0077..598bc5315e1937707c3cb2461f00a39586ab008f 100644
|
| --- a/telemetry/telemetry/benchmark_runner.py
|
| +++ b/telemetry/telemetry/benchmark_runner.py
|
| @@ -24,6 +24,7 @@ from telemetry.internal.util import command_line
|
| from telemetry.internal.util import ps_util
|
| from telemetry.util import matching
|
|
|
| +
|
| # Right now, we only have one of each of our power perf bots. This means that
|
| # all eligible Telemetry benchmarks are run unsharded, which results in very
|
| # long (12h) cycle times. We'd like to reduce the number of tests that we run
|
| @@ -37,6 +38,7 @@ GOOD_POWER_PERF_BOT_WHITELIST = [
|
| "Mac Power Low-End Perf"
|
| ]
|
|
|
| +
|
| DEFAULT_LOG_FORMAT = (
|
| '(%(levelname)s) %(asctime)s %(module)s.%(funcName)s:%(lineno)d '
|
| '%(message)s')
|
| @@ -375,7 +377,7 @@ def _GetJsonBenchmarkList(possible_browser, possible_reference_browser,
|
|
|
|
|
| def main(environment, extra_commands=None, **log_config_kwargs):
|
| - # the log level is set in browser_options
|
| + # The log level is set in browser_options.
|
| log_config_kwargs.pop('level', None)
|
| log_config_kwargs.setdefault('format', DEFAULT_LOG_FORMAT)
|
| logging.basicConfig(**log_config_kwargs)
|
|
|