Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(406)

Unified Diff: bench/gen_bench_expectations.py

Issue 2125953002: Remove obsolete bench analysis scripts (Closed) Base URL: https://skia.googlesource.com/skia.git@master
Patch Set: Created 4 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « bench/bench_util.py ('k') | bench/tile_analyze.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: bench/gen_bench_expectations.py
diff --git a/bench/gen_bench_expectations.py b/bench/gen_bench_expectations.py
deleted file mode 100644
index 4edc38c09d236954a37edaf165231d59e9eeb7c5..0000000000000000000000000000000000000000
--- a/bench/gen_bench_expectations.py
+++ /dev/null
@@ -1,223 +0,0 @@
-#!/usr/bin/env python
-# Copyright (c) 2014 The Chromium Authors. All rights reserved.
-# Use of this source code is governed by a BSD-style license that can be
-# found in the LICENSE file.
-
-""" Generate bench_expectations file from a given set of bench data files. """
-
-import argparse
-import bench_util
-import json
-import os
-import re
-import sys
-import urllib2
-
-# Parameters for calculating bench ranges.
-RANGE_RATIO_UPPER = 1.5 # Ratio of range for upper bounds.
-RANGE_RATIO_LOWER = 2.0 # Ratio of range for lower bounds.
-ERR_RATIO = 0.08 # Further widens the range by the ratio of average value.
-ERR_UB = 1.0 # Adds an absolute upper error to cope with small benches.
-ERR_LB = 1.5
-
-# List of bench configs to monitor. Ignore all other configs.
-CONFIGS_TO_INCLUDE = ['simple_viewport_1000x1000',
- 'simple_viewport_1000x1000_angle',
- 'simple_viewport_1000x1000_gpu',
- 'simple_viewport_1000x1000_scalar_1.100000',
- 'simple_viewport_1000x1000_scalar_1.100000_gpu',
- ]
-
-# List of flaky entries that should be excluded. Each entry is defined by a list
-# of 3 strings, corresponding to the substrings of [bench, config, builder] to
-# search for. A bench expectations line is excluded when each of the 3 strings
-# in the list is a substring of the corresponding element of the given line. For
-# instance, ['desk_yahooanswers', 'gpu', 'Ubuntu'] will skip expectation entries
-# of SKP benchs whose name contains 'desk_yahooanswers' on all gpu-related
-# configs of all Ubuntu builders.
-ENTRIES_TO_EXCLUDE = [
- ]
-
-_GS_CLOUD_FORMAT = 'http://storage.googleapis.com/chromium-skia-gm/perfdata/%s/%s'
-
-def compute_ranges(benches, more_benches=None):
- """Given a list of bench numbers, calculate the alert range.
-
- Args:
- benches: a list of float bench values.
- more_benches: a tuple of lists of additional bench values.
- The first value of each tuple is the number of commits before the current
- one that set of values is at, and the second value is a list of
- bench results.
-
- Returns:
- a list of float [lower_bound, upper_bound].
- """
- avg = sum(benches)/len(benches)
- minimum = min(benches)
- maximum = max(benches)
- diff = maximum - minimum
-
- return [minimum - diff*RANGE_RATIO_LOWER - avg*ERR_RATIO - ERR_LB,
- maximum + diff*RANGE_RATIO_UPPER + avg*ERR_RATIO + ERR_UB]
-
-
-def create_expectations_dict(revision_data_points, builder, extra_data=None):
- """Convert list of bench data points into a dictionary of expectations data.
-
- Args:
- revision_data_points: a list of BenchDataPoint objects.
- builder: string of the corresponding buildbot builder name.
-
- Returns:
- a dictionary of this form:
- keys = tuple of (config, bench) strings.
- values = list of float [expected, lower_bound, upper_bound] for the key.
- """
- bench_dict = {}
- for point in revision_data_points:
- if (point.time_type or # Not walltime which has time_type ''
- not point.config in CONFIGS_TO_INCLUDE):
- continue
- to_skip = False
- for bench_substr, config_substr, builder_substr in ENTRIES_TO_EXCLUDE:
- if (bench_substr in point.bench and config_substr in point.config and
- builder_substr in builder):
- to_skip = True
- break
- if to_skip:
- continue
- key = (point.config, point.bench)
-
- extras = []
- for idx, dataset in extra_data:
- for data in dataset:
- if (data.bench == point.bench and data.config == point.config and
- data.time_type == point.time_type and data.per_iter_time):
- extras.append((idx, data.per_iter_time))
-
- if key in bench_dict:
- raise Exception('Duplicate bench entry: ' + str(key))
- bench_dict[key] = [point.time] + compute_ranges(point.per_iter_time, extras)
-
- return bench_dict
-
-
-def get_parent_commits(start_hash, num_back):
- """Returns a list of commits that are the parent of the commit passed in."""
- list_commits = urllib2.urlopen(
- 'https://skia.googlesource.com/skia/+log/%s?format=json&n=%d' %
- (start_hash, num_back))
- # NOTE: Very brittle. Removes the four extraneous characters
- # so json can be read successfully
- trunc_list = list_commits.read()[4:]
- json_data = json.loads(trunc_list)
- return [revision['commit'] for revision in json_data['log']]
-
-
-def get_file_suffixes(commit_hash, directory):
- """Gets all the suffixes available in the directory"""
- possible_files = os.listdir(directory)
- prefix = 'bench_' + commit_hash + '_data_'
- return [name[len(prefix):] for name in possible_files
- if name.startswith(prefix)]
-
-
-def download_bench_data(builder, commit_hash, suffixes, directory):
- """Downloads data, returns the number successfully downloaded"""
- cur_files = os.listdir(directory)
- count = 0
- for suffix in suffixes:
- file_name = 'bench_'+commit_hash+'_data_'+suffix
- if file_name in cur_files:
- continue
- try:
- src = urllib2.urlopen(_GS_CLOUD_FORMAT % (builder, file_name))
- with open(os.path.join(directory, file_name), 'w') as dest:
- dest.writelines(src)
- count += 1
- except urllib2.HTTPError:
- pass
- return count
-
-
-def main():
- """Reads bench data points, then calculate and export expectations.
- """
- parser = argparse.ArgumentParser()
- parser.add_argument(
- '-a', '--representation_alg', default='25th',
- help='bench representation algorithm to use, see bench_util.py.')
- parser.add_argument(
- '-b', '--builder', required=True,
- help='name of the builder whose bench ranges we are computing.')
- parser.add_argument(
- '-d', '--input_dir', required=True,
- help='a directory containing bench data files.')
- parser.add_argument(
- '-o', '--output_file', required=True,
- help='file path and name for storing the output bench expectations.')
- parser.add_argument(
- '-r', '--git_revision', required=True,
- help='the git hash to indicate the revision of input data to use.')
- parser.add_argument(
- '-t', '--back_track', required=False, default=10,
- help='the number of commit hashes backwards to look to include' +
- 'in the calculations.')
- parser.add_argument(
- '-m', '--max_commits', required=False, default=1,
- help='the number of commit hashes to include in the calculations.')
- args = parser.parse_args()
-
- builder = args.builder
-
- data_points = bench_util.parse_skp_bench_data(
- args.input_dir, args.git_revision, args.representation_alg)
-
- parent_commits = get_parent_commits(args.git_revision, args.back_track)
- print "Using commits: {}".format(parent_commits)
- suffixes = get_file_suffixes(args.git_revision, args.input_dir)
- print "Using suffixes: {}".format(suffixes)
-
- # TODO(kelvinly): Find a better approach to than directly copying from
- # the GS server?
- downloaded_commits = []
- for idx, commit in enumerate(parent_commits):
- num_downloaded = download_bench_data(
- builder, commit, suffixes, args.input_dir)
- if num_downloaded > 0:
- downloaded_commits.append((num_downloaded, idx, commit))
-
- if len(downloaded_commits) < args.max_commits:
- print ('Less than desired number of commits found. Please increase'
- '--back_track in later runs')
- trunc_commits = sorted(downloaded_commits, reverse=True)[:args.max_commits]
- extra_data = []
- for _, idx, commit in trunc_commits:
- extra_data.append((idx, bench_util.parse_skp_bench_data(
- args.input_dir, commit, args.representation_alg)))
-
- expectations_dict = create_expectations_dict(data_points, builder,
- extra_data)
-
- out_lines = []
- keys = expectations_dict.keys()
- keys.sort()
- for (config, bench) in keys:
- (expected, lower_bound, upper_bound) = expectations_dict[(config, bench)]
- out_lines.append('%(bench)s_%(config)s_,%(builder)s-%(representation)s,'
- '%(expected)s,%(lower_bound)s,%(upper_bound)s' % {
- 'bench': bench,
- 'config': config,
- 'builder': builder,
- 'representation': args.representation_alg,
- 'expected': expected,
- 'lower_bound': lower_bound,
- 'upper_bound': upper_bound})
-
- with open(args.output_file, 'w') as file_handle:
- file_handle.write('\n'.join(out_lines))
-
-
-if __name__ == "__main__":
- main()
« no previous file with comments | « bench/bench_util.py ('k') | bench/tile_analyze.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698