Index: tools/gen_bench_expectations_from_codereview.py |
diff --git a/tools/gen_bench_expectations_from_codereview.py b/tools/gen_bench_expectations_from_codereview.py |
deleted file mode 100644 |
index f0463fea15b6e24495cb3cc62bb8ce85bb0c7dbc..0000000000000000000000000000000000000000 |
--- a/tools/gen_bench_expectations_from_codereview.py |
+++ /dev/null |
@@ -1,236 +0,0 @@ |
-#!/usr/bin/python |
- |
-# Copyright (c) 2014 The Chromium Authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
- |
- |
-"""Generate new bench expectations from results of trybots on a code review.""" |
- |
- |
-import collections |
-import compare_codereview |
-import json |
-import os |
-import re |
-import shutil |
-import subprocess |
-import sys |
-import urllib2 |
- |
- |
-BENCH_DATA_URL = 'gs://chromium-skia-gm/perfdata/%s/%s/bench_*_data_*' |
-BUILD_STATUS_SUCCESS = 0 |
-BUILD_STATUS_WARNINGS = 1 |
-CHECKOUT_PATH = os.path.realpath(os.path.join( |
- os.path.dirname(os.path.abspath(__file__)), os.pardir)) |
-TMP_BENCH_DATA_DIR = os.path.join(CHECKOUT_PATH, '.bench_data') |
- |
- |
-TryBuild = collections.namedtuple( |
- 'TryBuild', ['builder_name', 'build_number', 'is_finished', 'json_url']) |
- |
- |
-def find_all_builds(codereview_url): |
- """Finds and returns information about trybot runs for a code review. |
- |
- Args: |
- codereview_url: URL of the codereview in question. |
- |
- Returns: |
- List of NamedTuples: (builder_name, build_number, is_finished) |
- """ |
- results = compare_codereview.CodeReviewHTMLParser().parse(codereview_url) |
- try_builds = [] |
- for builder, data in results.iteritems(): |
- if builder.startswith('Perf'): |
- build_num = None |
- json_url = None |
- if data.url: |
- split_url = data.url.split('/') |
- build_num = split_url[-1] |
- split_url.insert(split_url.index('builders'), 'json') |
- json_url = '/'.join(split_url) |
- is_finished = (data.status not in ('pending', 'try-pending') and |
- build_num is not None) |
- try_builds.append(TryBuild(builder_name=builder, |
- build_number=build_num, |
- is_finished=is_finished, |
- json_url=json_url)) |
- return try_builds |
- |
- |
-def _all_trybots_finished(try_builds): |
- """Return True iff all of the given try jobs have finished. |
- |
- Args: |
- try_builds: list of TryBuild instances. |
- |
- Returns: |
- True if all of the given try jobs have finished, otherwise False. |
- """ |
- for try_build in try_builds: |
- if not try_build.is_finished: |
- return False |
- return True |
- |
- |
-def all_trybots_finished(codereview_url): |
- """Return True iff all of the try jobs on the given codereview have finished. |
- |
- Args: |
- codereview_url: string; URL of the codereview. |
- |
- Returns: |
- True if all of the try jobs have finished, otherwise False. |
- """ |
- return _all_trybots_finished(find_all_builds(codereview_url)) |
- |
- |
-def get_bench_data(builder, build_num, dest_dir): |
- """Download the bench data for the given builder at the given build_num. |
- |
- Args: |
- builder: string; name of the builder. |
- build_num: string; build number. |
- dest_dir: string; destination directory for the bench data. |
- """ |
- url = BENCH_DATA_URL % (builder, build_num) |
- subprocess.check_call(['gsutil', 'cp', '-R', url, dest_dir]) |
- |
- |
-def find_revision_from_downloaded_data(dest_dir): |
- """Finds the revision at which the downloaded data was generated. |
- |
- Args: |
- dest_dir: string; directory holding the downloaded data. |
- |
- Returns: |
- The revision (git commit hash) at which the downloaded data was |
- generated, or None if no revision can be found. |
- """ |
- for data_file in os.listdir(dest_dir): |
- match = re.match('bench_(?P<revision>[0-9a-fA-F]{2,40})_data.*', data_file) |
- if match: |
- return match.group('revision') |
- return None |
- |
- |
-class TrybotNotFinishedError(Exception): |
- pass |
- |
- |
-def _step_succeeded(try_build, step_name): |
- """Return True if the given step succeeded and False otherwise. |
- |
- This function talks to the build master's JSON interface, which is slow. |
- |
- TODO(borenet): There are now a few places which talk to the master's JSON |
- interface. Maybe it'd be worthwhile to create a module which does this. |
- |
- Args: |
- try_build: TryBuild instance; the build we're concerned about. |
- step_name: string; name of the step we're concerned about. |
- """ |
- step_url = '/'.join((try_build.json_url, 'steps', step_name)) |
- step_data = json.load(urllib2.urlopen(step_url)) |
- # step_data['results'] may not be present if the step succeeded. If present, |
- # it is a list whose first element is a result code, per the documentation: |
- # http://docs.buildbot.net/latest/developer/results.html |
- result = step_data.get('results', [BUILD_STATUS_SUCCESS])[0] |
- if result in (BUILD_STATUS_SUCCESS, BUILD_STATUS_WARNINGS): |
- return True |
- return False |
- |
- |
-def gen_bench_expectations_from_codereview(codereview_url, |
- error_on_unfinished=True, |
- error_on_try_failure=True): |
- """Generate bench expectations from a code review. |
- |
- Scans the given code review for Perf trybot runs. Downloads the results of |
- finished trybots and uses them to generate new expectations for their |
- waterfall counterparts. |
- |
- Args: |
- url: string; URL of the code review. |
- error_on_unfinished: bool; throw an error if any trybot has not finished. |
- error_on_try_failure: bool; throw an error if any trybot failed an |
- important step. |
- """ |
- try_builds = find_all_builds(codereview_url) |
- |
- # Verify that all trybots have finished running. |
- if error_on_unfinished and not _all_trybots_finished(try_builds): |
- raise TrybotNotFinishedError('Not all trybots have finished.') |
- |
- failed_run = [] |
- failed_data_pull = [] |
- failed_gen_expectations = [] |
- |
- # Don't even try to do anything if BenchPictures, PostBench, or |
- # UploadBenchResults failed. |
- for try_build in try_builds: |
- for step in ('BenchPictures', 'PostBench', 'UploadBenchResults'): |
- if not _step_succeeded(try_build, step): |
- msg = '%s failed on %s!' % (step, try_build.builder_name) |
- if error_on_try_failure: |
- raise Exception(msg) |
- print 'WARNING: %s Skipping.' % msg |
- failed_run.append(try_build.builder_name) |
- |
- if os.path.isdir(TMP_BENCH_DATA_DIR): |
- shutil.rmtree(TMP_BENCH_DATA_DIR) |
- |
- for try_build in try_builds: |
- try_builder = try_build.builder_name |
- |
- # Even if we're not erroring out on try failures, we can't generate new |
- # expectations for failed bots. |
- if try_builder in failed_run: |
- continue |
- |
- builder = try_builder.replace('-Trybot', '') |
- |
- # Download the data. |
- dest_dir = os.path.join(TMP_BENCH_DATA_DIR, builder) |
- os.makedirs(dest_dir) |
- try: |
- get_bench_data(try_builder, try_build.build_number, dest_dir) |
- except subprocess.CalledProcessError: |
- failed_data_pull.append(try_builder) |
- continue |
- |
- # Find the revision at which the data was generated. |
- revision = find_revision_from_downloaded_data(dest_dir) |
- if not revision: |
- # If we can't find a revision, then something is wrong with the data we |
- # downloaded. Skip this builder. |
- failed_data_pull.append(try_builder) |
- continue |
- |
- # Generate new expectations. |
- output_file = os.path.join(CHECKOUT_PATH, 'expectations', 'bench', |
- 'bench_expectations_%s.txt' % builder) |
- try: |
- subprocess.check_call(['python', |
- os.path.join(CHECKOUT_PATH, 'bench', |
- 'gen_bench_expectations.py'), |
- '-b', builder, '-o', output_file, |
- '-d', dest_dir, '-r', revision]) |
- except subprocess.CalledProcessError: |
- failed_gen_expectations.append(builder) |
- |
- failure = '' |
- if failed_data_pull: |
- failure += 'Failed to load data for: %s\n\n' % ','.join(failed_data_pull) |
- if failed_gen_expectations: |
- failure += 'Failed to generate expectations for: %s\n\n' % ','.join( |
- failed_gen_expectations) |
- if failure: |
- raise Exception(failure) |
- |
- |
-if __name__ == '__main__': |
- gen_bench_expectations_from_codereview(sys.argv[1]) |
- |