Index: telemetry/telemetry/benchmark_runner.py |
diff --git a/telemetry/telemetry/benchmark_runner.py b/telemetry/telemetry/benchmark_runner.py |
index d7116194848c9cf51aebd1528703fcae02ce0077..598bc5315e1937707c3cb2461f00a39586ab008f 100644 |
--- a/telemetry/telemetry/benchmark_runner.py |
+++ b/telemetry/telemetry/benchmark_runner.py |
@@ -24,6 +24,7 @@ from telemetry.internal.util import command_line |
from telemetry.internal.util import ps_util |
from telemetry.util import matching |
+ |
# Right now, we only have one of each of our power perf bots. This means that |
# all eligible Telemetry benchmarks are run unsharded, which results in very |
# long (12h) cycle times. We'd like to reduce the number of tests that we run |
@@ -37,6 +38,7 @@ GOOD_POWER_PERF_BOT_WHITELIST = [ |
"Mac Power Low-End Perf" |
] |
+ |
DEFAULT_LOG_FORMAT = ( |
'(%(levelname)s) %(asctime)s %(module)s.%(funcName)s:%(lineno)d ' |
'%(message)s') |
@@ -375,7 +377,7 @@ def _GetJsonBenchmarkList(possible_browser, possible_reference_browser, |
def main(environment, extra_commands=None, **log_config_kwargs): |
- # the log level is set in browser_options |
+ # The log level is set in browser_options. |
log_config_kwargs.pop('level', None) |
log_config_kwargs.setdefault('format', DEFAULT_LOG_FORMAT) |
logging.basicConfig(**log_config_kwargs) |