Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(474)

Unified Diff: sky/tools/webkitpy/tool/commands/rebaseline.py

Issue 946753002: Delete a bunch of dead python code. (Closed) Base URL: git@github.com:domokit/mojo.git@master
Patch Set: Created 5 years, 10 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: sky/tools/webkitpy/tool/commands/rebaseline.py
diff --git a/sky/tools/webkitpy/tool/commands/rebaseline.py b/sky/tools/webkitpy/tool/commands/rebaseline.py
deleted file mode 100644
index fa25236e093238bc791256fd3d479fa0a7ee0cb6..0000000000000000000000000000000000000000
--- a/sky/tools/webkitpy/tool/commands/rebaseline.py
+++ /dev/null
@@ -1,945 +0,0 @@
-# Copyright (c) 2010 Google Inc. All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions are
-# met:
-#
-# * Redistributions of source code must retain the above copyright
-# notice, this list of conditions and the following disclaimer.
-# * Redistributions in binary form must reproduce the above
-# copyright notice, this list of conditions and the following disclaimer
-# in the documentation and/or other materials provided with the
-# distribution.
-# * Neither the name of Google Inc. nor the names of its
-# contributors may be used to endorse or promote products derived from
-# this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR/ OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-import Queue
-import json
-import logging
-import optparse
-import re
-import sys
-import threading
-import time
-import traceback
-import urllib
-import urllib2
-
-from webkitpy.common.checkout.baselineoptimizer import BaselineOptimizer
-from webkitpy.common.memoized import memoized
-from webkitpy.common.system.executive import ScriptError
-from webkitpy.layout_tests.controllers.test_result_writer import TestResultWriter
-from webkitpy.layout_tests.models import test_failures
-from webkitpy.layout_tests.models.test_expectations import TestExpectations, BASELINE_SUFFIX_LIST, SKIP
-from webkitpy.layout_tests.port import builders
-from webkitpy.layout_tests.port import factory
-from webkitpy.tool.multicommandtool import AbstractDeclarativeCommand
-
-
-_log = logging.getLogger(__name__)
-
-
-# FIXME: Should TestResultWriter know how to compute this string?
-def _baseline_name(fs, test_name, suffix):
- return fs.splitext(test_name)[0] + TestResultWriter.FILENAME_SUFFIX_EXPECTED + "." + suffix
-
-
-class AbstractRebaseliningCommand(AbstractDeclarativeCommand):
- # not overriding execute() - pylint: disable=W0223
-
- no_optimize_option = optparse.make_option('--no-optimize', dest='optimize', action='store_false', default=True,
- help=('Do not optimize/de-dup the expectations after rebaselining (default is to de-dup automatically). '
- 'You can use "webkit-patch optimize-baselines" to optimize separately.'))
-
- platform_options = factory.platform_options(use_globs=True)
-
- results_directory_option = optparse.make_option("--results-directory", help="Local results directory to use")
-
- suffixes_option = optparse.make_option("--suffixes", default=','.join(BASELINE_SUFFIX_LIST), action="store",
- help="Comma-separated-list of file types to rebaseline")
-
- def __init__(self, options=None):
- super(AbstractRebaseliningCommand, self).__init__(options=options)
- self._baseline_suffix_list = BASELINE_SUFFIX_LIST
- self._scm_changes = {'add': [], 'delete': [], 'remove-lines': []}
-
- def _add_to_scm_later(self, path):
- self._scm_changes['add'].append(path)
-
- def _delete_from_scm_later(self, path):
- self._scm_changes['delete'].append(path)
-
-
-class BaseInternalRebaselineCommand(AbstractRebaseliningCommand):
- def __init__(self):
- super(BaseInternalRebaselineCommand, self).__init__(options=[
- self.results_directory_option,
- self.suffixes_option,
- optparse.make_option("--builder", help="Builder to pull new baselines from"),
- optparse.make_option("--test", help="Test to rebaseline"),
- ])
-
- def _baseline_directory(self, builder_name):
- port = self._tool.port_factory.get_from_builder_name(builder_name)
- override_dir = builders.rebaseline_override_dir(builder_name)
- if override_dir:
- return self._tool.filesystem.join(port.layout_tests_dir(), 'platform', override_dir)
- return port.baseline_version_dir()
-
- def _test_root(self, test_name):
- return self._tool.filesystem.splitext(test_name)[0]
-
- def _file_name_for_actual_result(self, test_name, suffix):
- return "%s-actual.%s" % (self._test_root(test_name), suffix)
-
- def _file_name_for_expected_result(self, test_name, suffix):
- return "%s-expected.%s" % (self._test_root(test_name), suffix)
-
-
-class CopyExistingBaselinesInternal(BaseInternalRebaselineCommand):
- name = "copy-existing-baselines-internal"
- help_text = "Copy existing baselines down one level in the baseline order to ensure new baselines don't break existing passing platforms."
-
- @memoized
- def _immediate_predecessors_in_fallback(self, path_to_rebaseline):
- port_names = self._tool.port_factory.all_port_names()
- immediate_predecessors_in_fallback = []
- for port_name in port_names:
- port = self._tool.port_factory.get(port_name)
- if not port.buildbot_archives_baselines():
- continue
- baseline_search_path = port.baseline_search_path()
- try:
- index = baseline_search_path.index(path_to_rebaseline)
- if index:
- immediate_predecessors_in_fallback.append(self._tool.filesystem.basename(baseline_search_path[index - 1]))
- except ValueError:
- # index throw's a ValueError if the item isn't in the list.
- pass
- return immediate_predecessors_in_fallback
-
- def _port_for_primary_baseline(self, baseline):
- for port in [self._tool.port_factory.get(port_name) for port_name in self._tool.port_factory.all_port_names()]:
- if self._tool.filesystem.basename(port.baseline_version_dir()) == baseline:
- return port
- raise Exception("Failed to find port for primary baseline %s." % baseline)
-
- def _copy_existing_baseline(self, builder_name, test_name, suffix):
- baseline_directory = self._baseline_directory(builder_name)
- ports = [self._port_for_primary_baseline(baseline) for baseline in self._immediate_predecessors_in_fallback(baseline_directory)]
-
- old_baselines = []
- new_baselines = []
-
- # Need to gather all the baseline paths before modifying the filesystem since
- # the modifications can affect the results of port.expected_filename.
- for port in ports:
- old_baseline = port.expected_filename(test_name, "." + suffix)
- if not self._tool.filesystem.exists(old_baseline):
- _log.debug("No existing baseline for %s." % test_name)
- continue
-
- new_baseline = self._tool.filesystem.join(port.baseline_path(), self._file_name_for_expected_result(test_name, suffix))
- if self._tool.filesystem.exists(new_baseline):
- _log.debug("Existing baseline at %s, not copying over it." % new_baseline)
- continue
-
- expectations = TestExpectations(port, [test_name])
- if SKIP in expectations.get_expectations(test_name):
- _log.debug("%s is skipped on %s." % (test_name, port.name()))
- continue
-
- old_baselines.append(old_baseline)
- new_baselines.append(new_baseline)
-
- for i in range(len(old_baselines)):
- old_baseline = old_baselines[i]
- new_baseline = new_baselines[i]
-
- _log.debug("Copying baseline from %s to %s." % (old_baseline, new_baseline))
- self._tool.filesystem.maybe_make_directory(self._tool.filesystem.dirname(new_baseline))
- self._tool.filesystem.copyfile(old_baseline, new_baseline)
- if not self._tool.scm().exists(new_baseline):
- self._add_to_scm_later(new_baseline)
-
- def execute(self, options, args, tool):
- for suffix in options.suffixes.split(','):
- self._copy_existing_baseline(options.builder, options.test, suffix)
- print json.dumps(self._scm_changes)
-
-
-class RebaselineTest(BaseInternalRebaselineCommand):
- name = "rebaseline-test-internal"
- help_text = "Rebaseline a single test from a buildbot. Only intended for use by other webkit-patch commands."
-
- def _results_url(self, builder_name):
- return self._tool.buildbot_for_builder_name(builder_name).builder_with_name(builder_name).latest_layout_test_results_url()
-
- def _save_baseline(self, data, target_baseline, baseline_directory, test_name, suffix):
- if not data:
- _log.debug("No baseline data to save.")
- return
-
- filesystem = self._tool.filesystem
- filesystem.maybe_make_directory(filesystem.dirname(target_baseline))
- filesystem.write_binary_file(target_baseline, data)
- if not self._tool.scm().exists(target_baseline):
- self._add_to_scm_later(target_baseline)
-
- def _rebaseline_test(self, builder_name, test_name, suffix, results_url):
- baseline_directory = self._baseline_directory(builder_name)
-
- source_baseline = "%s/%s" % (results_url, self._file_name_for_actual_result(test_name, suffix))
- target_baseline = self._tool.filesystem.join(baseline_directory, self._file_name_for_expected_result(test_name, suffix))
-
- _log.debug("Retrieving %s." % source_baseline)
- self._save_baseline(self._tool.web.get_binary(source_baseline, convert_404_to_None=True), target_baseline, baseline_directory, test_name, suffix)
-
- def _rebaseline_test_and_update_expectations(self, options):
- port = self._tool.port_factory.get_from_builder_name(options.builder)
- if (port.reference_files(options.test)):
- _log.warning("Cannot rebaseline reftest: %s", options.test)
- return
-
- if options.results_directory:
- results_url = 'file://' + options.results_directory
- else:
- results_url = self._results_url(options.builder)
- self._baseline_suffix_list = options.suffixes.split(',')
-
- for suffix in self._baseline_suffix_list:
- self._rebaseline_test(options.builder, options.test, suffix, results_url)
- self._scm_changes['remove-lines'].append({'builder': options.builder, 'test': options.test})
-
- def execute(self, options, args, tool):
- self._rebaseline_test_and_update_expectations(options)
- print json.dumps(self._scm_changes)
-
-
-class OptimizeBaselines(AbstractRebaseliningCommand):
- name = "optimize-baselines"
- help_text = "Reshuffles the baselines for the given tests to use as litte space on disk as possible."
- show_in_main_help = True
- argument_names = "TEST_NAMES"
-
- def __init__(self):
- super(OptimizeBaselines, self).__init__(options=[
- self.suffixes_option,
- optparse.make_option('--no-modify-scm', action='store_true', default=False, help='Dump SCM commands as JSON instead of '),
- ] + self.platform_options)
-
- def _optimize_baseline(self, optimizer, test_name):
- files_to_delete = []
- files_to_add = []
- for suffix in self._baseline_suffix_list:
- baseline_name = _baseline_name(self._tool.filesystem, test_name, suffix)
- succeeded, more_files_to_delete, more_files_to_add = optimizer.optimize(baseline_name)
- if not succeeded:
- print "Heuristics failed to optimize %s" % baseline_name
- files_to_delete.extend(more_files_to_delete)
- files_to_add.extend(more_files_to_add)
- return files_to_delete, files_to_add
-
- def execute(self, options, args, tool):
- self._baseline_suffix_list = options.suffixes.split(',')
- port_names = tool.port_factory.all_port_names(options.platform)
- if not port_names:
- print "No port names match '%s'" % options.platform
- return
-
- optimizer = BaselineOptimizer(tool, port_names, skip_scm_commands=options.no_modify_scm)
- port = tool.port_factory.get(port_names[0])
- for test_name in port.tests(args):
- _log.info("Optimizing %s" % test_name)
- files_to_delete, files_to_add = self._optimize_baseline(optimizer, test_name)
- for path in files_to_delete:
- self._delete_from_scm_later(path)
- for path in files_to_add:
- self._add_to_scm_later(path)
-
- print json.dumps(self._scm_changes)
-
-
-class AnalyzeBaselines(AbstractRebaseliningCommand):
- name = "analyze-baselines"
- help_text = "Analyzes the baselines for the given tests and prints results that are identical."
- show_in_main_help = True
- argument_names = "TEST_NAMES"
-
- def __init__(self):
- super(AnalyzeBaselines, self).__init__(options=[
- self.suffixes_option,
- optparse.make_option('--missing', action='store_true', default=False, help='show missing baselines as well'),
- ] + self.platform_options)
- self._optimizer_class = BaselineOptimizer # overridable for testing
- self._baseline_optimizer = None
- self._port = None
-
- def _write(self, msg):
- print msg
-
- def _analyze_baseline(self, options, test_name):
- for suffix in self._baseline_suffix_list:
- baseline_name = _baseline_name(self._tool.filesystem, test_name, suffix)
- results_by_directory = self._baseline_optimizer.read_results_by_directory(baseline_name)
- if results_by_directory:
- self._write("%s:" % baseline_name)
- self._baseline_optimizer.write_by_directory(results_by_directory, self._write, " ")
- elif options.missing:
- self._write("%s: (no baselines found)" % baseline_name)
-
- def execute(self, options, args, tool):
- self._baseline_suffix_list = options.suffixes.split(',')
- port_names = tool.port_factory.all_port_names(options.platform)
- if not port_names:
- print "No port names match '%s'" % options.platform
- return
-
- self._baseline_optimizer = self._optimizer_class(tool, port_names, skip_scm_commands=False)
- self._port = tool.port_factory.get(port_names[0])
- for test_name in self._port.tests(args):
- self._analyze_baseline(options, test_name)
-
-
-class AbstractParallelRebaselineCommand(AbstractRebaseliningCommand):
- # not overriding execute() - pylint: disable=W0223
-
- def __init__(self, options=None):
- super(AbstractParallelRebaselineCommand, self).__init__(options=options)
- self._builder_data = {}
-
- def builder_data(self):
- if not self._builder_data:
- for builder_name in self._release_builders():
- builder = self._tool.buildbot_for_builder_name(builder_name).builder_with_name(builder_name)
- self._builder_data[builder_name] = builder.latest_layout_test_results()
- return self._builder_data
-
- # The release builders cycle much faster than the debug ones and cover all the platforms.
- def _release_builders(self):
- release_builders = []
- for builder_name in builders.all_builder_names():
- if builder_name.find('ASAN') != -1:
- continue
- port = self._tool.port_factory.get_from_builder_name(builder_name)
- if port.test_configuration().build_type == 'release':
- release_builders.append(builder_name)
- return release_builders
-
- def _run_webkit_patch(self, args, verbose):
- try:
- verbose_args = ['--verbose'] if verbose else []
- stderr = self._tool.executive.run_command([self._tool.path()] + verbose_args + args, cwd=self._tool.scm().checkout_root, return_stderr=True)
- for line in stderr.splitlines():
- _log.warning(line)
- except ScriptError, e:
- _log.error(e)
-
- def _builders_to_fetch_from(self, builders_to_check):
- # This routine returns the subset of builders that will cover all of the baseline search paths
- # used in the input list. In particular, if the input list contains both Release and Debug
- # versions of a configuration, we *only* return the Release version (since we don't save
- # debug versions of baselines).
- release_builders = set()
- debug_builders = set()
- builders_to_fallback_paths = {}
- for builder in builders_to_check:
- port = self._tool.port_factory.get_from_builder_name(builder)
- if port.test_configuration().build_type == 'release':
- release_builders.add(builder)
- else:
- debug_builders.add(builder)
- for builder in list(release_builders) + list(debug_builders):
- port = self._tool.port_factory.get_from_builder_name(builder)
- fallback_path = port.baseline_search_path()
- if fallback_path not in builders_to_fallback_paths.values():
- builders_to_fallback_paths[builder] = fallback_path
- return builders_to_fallback_paths.keys()
-
- def _rebaseline_commands(self, test_prefix_list, options):
- path_to_webkit_patch = self._tool.path()
- cwd = self._tool.scm().checkout_root
- copy_baseline_commands = []
- rebaseline_commands = []
- lines_to_remove = {}
- port = self._tool.port_factory.get()
-
- for test_prefix in test_prefix_list:
- for test in port.tests([test_prefix]):
- for builder in self._builders_to_fetch_from(test_prefix_list[test_prefix]):
- actual_failures_suffixes = self._suffixes_for_actual_failures(test, builder, test_prefix_list[test_prefix][builder])
- if not actual_failures_suffixes:
- # If we're not going to rebaseline the test because it's passing on this
- # builder, we still want to remove the line from TestExpectations.
- if test not in lines_to_remove:
- lines_to_remove[test] = []
- lines_to_remove[test].append(builder)
- continue
-
- suffixes = ','.join(actual_failures_suffixes)
- cmd_line = ['--suffixes', suffixes, '--builder', builder, '--test', test]
- if options.results_directory:
- cmd_line.extend(['--results-directory', options.results_directory])
- if options.verbose:
- cmd_line.append('--verbose')
- copy_baseline_commands.append(tuple([[self._tool.executable, path_to_webkit_patch, 'copy-existing-baselines-internal'] + cmd_line, cwd]))
- rebaseline_commands.append(tuple([[self._tool.executable, path_to_webkit_patch, 'rebaseline-test-internal'] + cmd_line, cwd]))
- return copy_baseline_commands, rebaseline_commands, lines_to_remove
-
- def _serial_commands(self, command_results):
- files_to_add = set()
- files_to_delete = set()
- lines_to_remove = {}
- for output in [result[1].split('\n') for result in command_results]:
- file_added = False
- for line in output:
- try:
- if line:
- parsed_line = json.loads(line)
- if 'add' in parsed_line:
- files_to_add.update(parsed_line['add'])
- if 'delete' in parsed_line:
- files_to_delete.update(parsed_line['delete'])
- if 'remove-lines' in parsed_line:
- for line_to_remove in parsed_line['remove-lines']:
- test = line_to_remove['test']
- builder = line_to_remove['builder']
- if test not in lines_to_remove:
- lines_to_remove[test] = []
- lines_to_remove[test].append(builder)
- file_added = True
- except ValueError:
- _log.debug('"%s" is not a JSON object, ignoring' % line)
-
- if not file_added:
- _log.debug('Could not add file based off output "%s"' % output)
-
- return list(files_to_add), list(files_to_delete), lines_to_remove
-
- def _optimize_baselines(self, test_prefix_list, verbose=False):
- optimize_commands = []
- for test in test_prefix_list:
- all_suffixes = set()
- for builder in self._builders_to_fetch_from(test_prefix_list[test]):
- all_suffixes.update(self._suffixes_for_actual_failures(test, builder, test_prefix_list[test][builder]))
-
- # FIXME: We should propagate the platform options as well.
- cmd_line = ['--no-modify-scm', '--suffixes', ','.join(all_suffixes), test]
- if verbose:
- cmd_line.append('--verbose')
-
- path_to_webkit_patch = self._tool.path()
- cwd = self._tool.scm().checkout_root
- optimize_commands.append(tuple([[self._tool.executable, path_to_webkit_patch, 'optimize-baselines'] + cmd_line, cwd]))
- return optimize_commands
-
- def _update_expectations_files(self, lines_to_remove):
- # FIXME: This routine is way too expensive. We're creating O(n ports) TestExpectations objects.
- # This is slow and uses a lot of memory.
- tests = lines_to_remove.keys()
- to_remove = []
-
- # This is so we remove lines for builders that skip this test, e.g. Android skips most
- # tests and we don't want to leave stray [ Android ] lines in TestExpectations..
- # This is only necessary for "webkit-patch rebaseline" and for rebaselining expected
- # failures from garden-o-matic. rebaseline-expectations and auto-rebaseline will always
- # pass the exact set of ports to rebaseline.
- for port_name in self._tool.port_factory.all_port_names():
- port = self._tool.port_factory.get(port_name)
- generic_expectations = TestExpectations(port, tests=tests, include_overrides=False)
- full_expectations = TestExpectations(port, tests=tests, include_overrides=True)
- for test in tests:
- if self._port_skips_test(port, test, generic_expectations, full_expectations):
- for test_configuration in port.all_test_configurations():
- if test_configuration.version == port.test_configuration().version:
- to_remove.append((test, test_configuration))
-
- for test in lines_to_remove:
- for builder in lines_to_remove[test]:
- port = self._tool.port_factory.get_from_builder_name(builder)
- for test_configuration in port.all_test_configurations():
- if test_configuration.version == port.test_configuration().version:
- to_remove.append((test, test_configuration))
-
- port = self._tool.port_factory.get()
- expectations = TestExpectations(port, include_overrides=False)
- expectationsString = expectations.remove_configurations(to_remove)
- path = port.path_to_generic_test_expectations_file()
- self._tool.filesystem.write_text_file(path, expectationsString)
-
- def _port_skips_test(self, port, test, generic_expectations, full_expectations):
- fs = port.host.filesystem
- if port.default_smoke_test_only():
- smoke_test_filename = fs.join(port.layout_tests_dir(), 'SmokeTests')
- if fs.exists(smoke_test_filename) and test not in fs.read_text_file(smoke_test_filename):
- return True
-
- return (SKIP in full_expectations.get_expectations(test) and
- SKIP not in generic_expectations.get_expectations(test))
-
- def _run_in_parallel_and_update_scm(self, commands):
- command_results = self._tool.executive.run_in_parallel(commands)
- log_output = '\n'.join(result[2] for result in command_results).replace('\n\n', '\n')
- for line in log_output.split('\n'):
- if line:
- print >> sys.stderr, line # FIXME: Figure out how to log properly.
-
- files_to_add, files_to_delete, lines_to_remove = self._serial_commands(command_results)
- if files_to_delete:
- self._tool.scm().delete_list(files_to_delete)
- if files_to_add:
- self._tool.scm().add_list(files_to_add)
- return lines_to_remove
-
- def _rebaseline(self, options, test_prefix_list):
- for test, builders_to_check in sorted(test_prefix_list.items()):
- _log.info("Rebaselining %s" % test)
- for builder, suffixes in sorted(builders_to_check.items()):
- _log.debug(" %s: %s" % (builder, ",".join(suffixes)))
-
- copy_baseline_commands, rebaseline_commands, extra_lines_to_remove = self._rebaseline_commands(test_prefix_list, options)
- lines_to_remove = {}
-
- if copy_baseline_commands:
- self._run_in_parallel_and_update_scm(copy_baseline_commands)
- if rebaseline_commands:
- lines_to_remove = self._run_in_parallel_and_update_scm(rebaseline_commands)
-
- for test in extra_lines_to_remove:
- if test in lines_to_remove:
- lines_to_remove[test] = lines_to_remove[test] + extra_lines_to_remove[test]
- else:
- lines_to_remove[test] = extra_lines_to_remove[test]
-
- if lines_to_remove:
- self._update_expectations_files(lines_to_remove)
-
- if options.optimize:
- self._run_in_parallel_and_update_scm(self._optimize_baselines(test_prefix_list, options.verbose))
-
- def _suffixes_for_actual_failures(self, test, builder_name, existing_suffixes):
- actual_results = self.builder_data()[builder_name].actual_results(test)
- if not actual_results:
- return set()
- return set(existing_suffixes) & TestExpectations.suffixes_for_actual_expectations_string(actual_results)
-
-
-class RebaselineJson(AbstractParallelRebaselineCommand):
- name = "rebaseline-json"
- help_text = "Rebaseline based off JSON passed to stdin. Intended to only be called from other scripts."
-
- def __init__(self,):
- super(RebaselineJson, self).__init__(options=[
- self.no_optimize_option,
- self.results_directory_option,
- ])
-
- def execute(self, options, args, tool):
- self._rebaseline(options, json.loads(sys.stdin.read()))
-
-
-class RebaselineExpectations(AbstractParallelRebaselineCommand):
- name = "rebaseline-expectations"
- help_text = "Rebaselines the tests indicated in TestExpectations."
- show_in_main_help = True
-
- def __init__(self):
- super(RebaselineExpectations, self).__init__(options=[
- self.no_optimize_option,
- ] + self.platform_options)
- self._test_prefix_list = None
-
- def _tests_to_rebaseline(self, port):
- tests_to_rebaseline = {}
- for path, value in port.expectations_dict().items():
- expectations = TestExpectations(port, include_overrides=False, expectations_dict={path: value})
- for test in expectations.get_rebaselining_failures():
- suffixes = TestExpectations.suffixes_for_expectations(expectations.get_expectations(test))
- tests_to_rebaseline[test] = suffixes or BASELINE_SUFFIX_LIST
- return tests_to_rebaseline
-
- def _add_tests_to_rebaseline_for_port(self, port_name):
- builder_name = builders.builder_name_for_port_name(port_name)
- if not builder_name:
- return
- tests = self._tests_to_rebaseline(self._tool.port_factory.get(port_name)).items()
-
- if tests:
- _log.info("Retrieving results for %s from %s." % (port_name, builder_name))
-
- for test_name, suffixes in tests:
- _log.info(" %s (%s)" % (test_name, ','.join(suffixes)))
- if test_name not in self._test_prefix_list:
- self._test_prefix_list[test_name] = {}
- self._test_prefix_list[test_name][builder_name] = suffixes
-
- def execute(self, options, args, tool):
- options.results_directory = None
- self._test_prefix_list = {}
- port_names = tool.port_factory.all_port_names(options.platform)
- for port_name in port_names:
- self._add_tests_to_rebaseline_for_port(port_name)
- if not self._test_prefix_list:
- _log.warning("Did not find any tests marked Rebaseline.")
- return
-
- self._rebaseline(options, self._test_prefix_list)
-
-
-class Rebaseline(AbstractParallelRebaselineCommand):
- name = "rebaseline"
- help_text = "Rebaseline tests with results from the build bots. Shows the list of failing tests on the builders if no test names are provided."
- show_in_main_help = True
- argument_names = "[TEST_NAMES]"
-
- def __init__(self):
- super(Rebaseline, self).__init__(options=[
- self.no_optimize_option,
- # FIXME: should we support the platform options in addition to (or instead of) --builders?
- self.suffixes_option,
- self.results_directory_option,
- optparse.make_option("--builders", default=None, action="append", help="Comma-separated-list of builders to pull new baselines from (can also be provided multiple times)"),
- ])
-
- def _builders_to_pull_from(self):
- chosen_names = self._tool.user.prompt_with_list("Which builder to pull results from:", self._release_builders(), can_choose_multiple=True)
- return [self._builder_with_name(name) for name in chosen_names]
-
- def _builder_with_name(self, name):
- return self._tool.buildbot_for_builder_name(name).builder_with_name(name)
-
- def execute(self, options, args, tool):
- if not args:
- _log.error("Must list tests to rebaseline.")
- return
-
- if options.builders:
- builders_to_check = []
- for builder_names in options.builders:
- builders_to_check += [self._builder_with_name(name) for name in builder_names.split(",")]
- else:
- builders_to_check = self._builders_to_pull_from()
-
- test_prefix_list = {}
- suffixes_to_update = options.suffixes.split(",")
-
- for builder in builders_to_check:
- for test in args:
- if test not in test_prefix_list:
- test_prefix_list[test] = {}
- test_prefix_list[test][builder.name()] = suffixes_to_update
-
- if options.verbose:
- _log.debug("rebaseline-json: " + str(test_prefix_list))
-
- self._rebaseline(options, test_prefix_list)
-
-
-class AutoRebaseline(AbstractParallelRebaselineCommand):
- name = "auto-rebaseline"
- help_text = "Rebaselines any NeedsRebaseline lines in TestExpectations that have cycled through all the bots."
- AUTO_REBASELINE_BRANCH_NAME = "auto-rebaseline-temporary-branch"
-
- # Rietveld uploader stinks. Limit the number of rebaselines in a given patch to keep upload from failing.
- # FIXME: http://crbug.com/263676 Obviously we should fix the uploader here.
- MAX_LINES_TO_REBASELINE = 200
-
- SECONDS_BEFORE_GIVING_UP = 300
-
- def __init__(self):
- super(AutoRebaseline, self).__init__(options=[
- # FIXME: Remove this option.
- self.no_optimize_option,
- # FIXME: Remove this option.
- self.results_directory_option,
- ])
-
- def bot_revision_data(self):
- revisions = []
- for result in self.builder_data().values():
- if result.run_was_interrupted():
- _log.error("Can't rebaseline because the latest run on %s exited early." % result.builder_name())
- return []
- revisions.append({
- "builder": result.builder_name(),
- "revision": result.blink_revision(),
- })
- return revisions
-
- def tests_to_rebaseline(self, tool, min_revision, print_revisions):
- port = tool.port_factory.get()
- expectations_file_path = port.path_to_generic_test_expectations_file()
-
- tests = set()
- revision = None
- author = None
- bugs = set()
- has_any_needs_rebaseline_lines = False
-
- for line in tool.scm().blame(expectations_file_path).split("\n"):
- comment_index = line.find("#")
- if comment_index == -1:
- comment_index = len(line)
- line_without_comments = re.sub(r"\s+", " ", line[:comment_index].strip())
-
- if "NeedsRebaseline" not in line_without_comments:
- continue
-
- has_any_needs_rebaseline_lines = True
-
- parsed_line = re.match("^(\S*)[^(]*\((\S*).*?([^ ]*)\ \[[^[]*$", line_without_comments)
-
- commit_hash = parsed_line.group(1)
- svn_revision = tool.scm().svn_revision_from_git_commit(commit_hash)
-
- test = parsed_line.group(3)
- if print_revisions:
- _log.info("%s is waiting for r%s" % (test, svn_revision))
-
- if not svn_revision or svn_revision > min_revision:
- continue
-
- if revision and svn_revision != revision:
- continue
-
- if not revision:
- revision = svn_revision
- author = parsed_line.group(2)
-
- bugs.update(re.findall("crbug\.com\/(\d+)", line_without_comments))
- tests.add(test)
-
- if len(tests) >= self.MAX_LINES_TO_REBASELINE:
- _log.info("Too many tests to rebaseline in one patch. Doing the first %d." % self.MAX_LINES_TO_REBASELINE)
- break
-
- return tests, revision, author, bugs, has_any_needs_rebaseline_lines
-
- def link_to_patch(self, revision):
- return "http://src.chromium.org/viewvc/blink?view=revision&revision=" + str(revision)
-
- def commit_message(self, author, revision, bugs):
- bug_string = ""
- if bugs:
- bug_string = "BUG=%s\n" % ",".join(bugs)
-
- return """Auto-rebaseline for r%s
-
-%s
-
-%sTBR=%s
-""" % (revision, self.link_to_patch(revision), bug_string, author)
-
- def get_test_prefix_list(self, tests):
- test_prefix_list = {}
- lines_to_remove = {}
-
- for builder_name in self._release_builders():
- port_name = builders.port_name_for_builder_name(builder_name)
- port = self._tool.port_factory.get(port_name)
- expectations = TestExpectations(port, include_overrides=True)
- for test in expectations.get_needs_rebaseline_failures():
- if test not in tests:
- continue
-
- if test not in test_prefix_list:
- lines_to_remove[test] = []
- test_prefix_list[test] = {}
- lines_to_remove[test].append(builder_name)
- test_prefix_list[test][builder_name] = BASELINE_SUFFIX_LIST
-
- return test_prefix_list, lines_to_remove
-
- def _run_git_cl_command(self, options, command):
- subprocess_command = ['git', 'cl'] + command
- if options.verbose:
- subprocess_command.append('--verbose')
-
- process = self._tool.executive.popen(subprocess_command, stdout=self._tool.executive.PIPE, stderr=self._tool.executive.STDOUT)
- last_output_time = time.time()
-
- # git cl sometimes completely hangs. Bail if we haven't gotten any output to stdout/stderr in a while.
- while process.poll() == None and time.time() < last_output_time + self.SECONDS_BEFORE_GIVING_UP:
- # FIXME: This doesn't make any sense. readline blocks, so all this code to
- # try and bail is useless. Instead, we should do the readline calls on a
- # subthread. Then the rest of this code would make sense.
- out = process.stdout.readline().rstrip('\n')
- if out:
- last_output_time = time.time()
- _log.info(out)
-
- if process.poll() == None:
- _log.error('Command hung: %s' % subprocess_command)
- return False
- return True
-
- # FIXME: Move this somewhere more general.
- def tree_status(self):
- blink_tree_status_url = "http://blink-status.appspot.com/status"
- status = urllib2.urlopen(blink_tree_status_url).read().lower()
- if status.find('closed') != -1 or status == "0":
- return 'closed'
- elif status.find('open') != -1 or status == "1":
- return 'open'
- return 'unknown'
-
- def execute(self, options, args, tool):
- if tool.scm().executable_name == "svn":
- _log.error("Auto rebaseline only works with a git checkout.")
- return
-
- if tool.scm().has_working_directory_changes():
- _log.error("Cannot proceed with working directory changes. Clean working directory first.")
- return
-
- revision_data = self.bot_revision_data()
- if not revision_data:
- return
-
- min_revision = int(min([item["revision"] for item in revision_data]))
- tests, revision, author, bugs, has_any_needs_rebaseline_lines = self.tests_to_rebaseline(tool, min_revision, print_revisions=options.verbose)
-
- if options.verbose:
- _log.info("Min revision across all bots is %s." % min_revision)
- for item in revision_data:
- _log.info("%s: r%s" % (item["builder"], item["revision"]))
-
- if not tests:
- _log.debug('No tests to rebaseline.')
- return
-
- if self.tree_status() == 'closed':
- _log.info('Cannot proceed. Tree is closed.')
- return
-
- _log.info('Rebaselining %s for r%s by %s.' % (list(tests), revision, author))
-
- test_prefix_list, lines_to_remove = self.get_test_prefix_list(tests)
-
- did_finish = False
- try:
- old_branch_name = tool.scm().current_branch()
- tool.scm().delete_branch(self.AUTO_REBASELINE_BRANCH_NAME)
- tool.scm().create_clean_branch(self.AUTO_REBASELINE_BRANCH_NAME)
-
- # If the tests are passing everywhere, then this list will be empty. We don't need
- # to rebaseline, but we'll still need to update TestExpectations.
- if test_prefix_list:
- self._rebaseline(options, test_prefix_list)
-
- tool.scm().commit_locally_with_message(self.commit_message(author, revision, bugs))
-
- # FIXME: It would be nice if we could dcommit the patch without uploading, but still
- # go through all the precommit hooks. For rebaselines with lots of files, uploading
- # takes a long time and sometimes fails, but we don't want to commit if, e.g. the
- # tree is closed.
- did_finish = self._run_git_cl_command(options, ['upload', '-f'])
-
- if did_finish:
- # Uploading can take a very long time. Do another pull to make sure TestExpectations is up to date,
- # so the dcommit can go through.
- # FIXME: Log the pull and dcommit stdout/stderr to the log-server.
- tool.executive.run_command(['git', 'pull'])
-
- self._run_git_cl_command(options, ['dcommit', '-f'])
- except Exception as e:
- _log.error(e)
- finally:
- if did_finish:
- self._run_git_cl_command(options, ['set_close'])
- tool.scm().ensure_cleanly_tracking_remote_master()
- tool.scm().checkout_branch(old_branch_name)
- tool.scm().delete_branch(self.AUTO_REBASELINE_BRANCH_NAME)
-
-
-class RebaselineOMatic(AbstractDeclarativeCommand):
- name = "rebaseline-o-matic"
- help_text = "Calls webkit-patch auto-rebaseline in a loop."
- show_in_main_help = True
-
- SLEEP_TIME_IN_SECONDS = 30
- LOG_SERVER = 'blinkrebaseline.appspot.com'
- QUIT_LOG = '##QUIT##'
-
- # Uploaded log entries append to the existing entry unless the
- # newentry flag is set. In that case it starts a new entry to
- # start appending to.
- def _log_to_server(self, log='', is_new_entry=False):
- query = {
- 'log': log,
- }
- if is_new_entry:
- query['newentry'] = 'on'
- try:
- urllib2.urlopen("http://" + self.LOG_SERVER + "/updatelog", data=urllib.urlencode(query))
- except:
- traceback.print_exc(file=sys.stderr)
-
- def _log_to_server_thread(self):
- is_new_entry = True
- while True:
- messages = [self._log_queue.get()]
- while not self._log_queue.empty():
- messages.append(self._log_queue.get())
- self._log_to_server('\n'.join(messages), is_new_entry=is_new_entry)
- is_new_entry = False
- if self.QUIT_LOG in messages:
- return
-
- def _post_log_to_server(self, log):
- self._log_queue.put(log)
-
- def _log_line(self, handle):
- out = handle.readline().rstrip('\n')
- if out:
- if self._verbose:
- print out
- self._post_log_to_server(out)
- return out
-
- def _run_logged_command(self, command):
- process = self._tool.executive.popen(command, stdout=self._tool.executive.PIPE, stderr=self._tool.executive.STDOUT)
-
- out = self._log_line(process.stdout)
- while out:
- # FIXME: This should probably batch up lines if they're available and log to the server once.
- out = self._log_line(process.stdout)
-
- def _do_one_rebaseline(self):
- self._log_queue = Queue.Queue(256)
- log_thread = threading.Thread(name='LogToServer', target=self._log_to_server_thread)
- log_thread.start()
- try:
- old_branch_name = self._tool.scm().current_branch()
- self._run_logged_command(['git', 'pull'])
- rebaseline_command = [self._tool.filesystem.join(self._tool.scm().checkout_root, 'tools', 'webkit-patch'), 'auto-rebaseline']
- if self._verbose:
- rebaseline_command.append('--verbose')
- self._run_logged_command(rebaseline_command)
- except:
- self._log_queue.put(self.QUIT_LOG)
- traceback.print_exc(file=sys.stderr)
- # Sometimes git crashes and leaves us on a detached head.
- self._tool.scm().checkout_branch(old_branch_name)
- else:
- self._log_queue.put(self.QUIT_LOG)
- log_thread.join()
-
- def execute(self, options, args, tool):
- self._verbose = options.verbose
- while True:
- self._do_one_rebaseline()
- time.sleep(self.SLEEP_TIME_IN_SECONDS)
« no previous file with comments | « sky/tools/webkitpy/tool/commands/queries_unittest.py ('k') | sky/tools/webkitpy/tool/commands/rebaseline_unittest.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698