Index: trunk/bench/check_bench_regressions.py |
=================================================================== |
--- trunk/bench/check_bench_regressions.py (working copy) |
+++ trunk/bench/check_bench_regressions.py (working copy) |
@@ -15,54 +15,33 @@ |
import urllib2 |
import xml.sax.saxutils |
-# We throw out any measurement outside this range, and log a warning. |
-MIN_REASONABLE_TIME = 0 |
-MAX_REASONABLE_TIME = 99999 |
+# Maximum expected number of characters we expect in an svn revision. |
+MAX_SVN_REV_LENGTH = 5 |
-# Constants for prefixes in output title used in buildbot. |
-TITLE_PREAMBLE = 'Bench_Performance_for_' |
-TITLE_PREAMBLE_LENGTH = len(TITLE_PREAMBLE) |
- |
def usage(): |
"""Prints simple usage information.""" |
- print '-a <url> the url to use for adding bench values to app engine app.' |
- print ' Example: "https://skiadash.appspot.com/add_point".' |
- print ' If not set, will skip this step.' |
- print '-b <bench> the bench to show.' |
- print '-c <config> the config to show (GPU, 8888, 565, etc).' |
- print '-d <dir> a directory containing bench_r<revision>_<scalar> files.' |
+ print '-a <representation_alg> bench representation algorithm to use. ' |
+ print ' Defaults to "25th". See bench_util.py for details.' |
+ print '-b <builder> name of the builder whose bench data we are checking.' |
+ print '-d <dir> a directory containing bench_<revision>_<scalar> files.' |
print '-e <file> file containing expected bench builder values/ranges.' |
print ' Will raise exception if actual bench values are out of range.' |
print ' See bench_expectations_<builder>.txt for data format / examples.' |
- print '-f <revision>[:<revision>] the revisions to use for fitting.' |
- print ' Negative <revision> is taken as offset from most recent revision.' |
- print '-i <time> the time to ignore (w, c, g, etc).' |
- print ' The flag is ignored when -t is set; otherwise we plot all the' |
- print ' times except the one specified here.' |
- print '-l <title> title to use for the output graph' |
- print '-m <representation> representation of bench value.' |
- print ' See _ListAlgorithm class in bench_util.py.' |
- print '-o <path> path to which to write output.' |
- print '-r <revision>[:<revision>] the revisions to show.' |
- print ' Negative <revision> is taken as offset from most recent revision.' |
- print '-s <setting>[=<value>] a setting to show (alpha, scalar, etc).' |
- print '-t <time> the time to show (w, c, g, etc).' |
- print '-x <int> the desired width of the svg.' |
- print '-y <int> the desired height of the svg.' |
- print '--default-setting <setting>[=<value>] setting for those without.' |
- |
+ print '-r <revision> the git commit hash or svn revision for checking ' |
+ print ' bench values.' |
+ |
class Label: |
"""The information in a label. |
- |
+ |
(str, str, str, str, {str:str})""" |
def __init__(self, bench, config, time_type, settings): |
self.bench = bench |
self.config = config |
self.time_type = time_type |
self.settings = settings |
- |
+ |
def __repr__(self): |
return "Label(%s, %s, %s, %s)" % ( |
str(self.bench), |
@@ -70,7 +49,7 @@ |
str(self.time_type), |
str(self.settings), |
) |
- |
+ |
def __str__(self): |
return "%s_%s_%s_%s" % ( |
str(self.bench), |
@@ -78,417 +57,138 @@ |
str(self.time_type), |
str(self.settings), |
) |
- |
+ |
def __eq__(self, other): |
return (self.bench == other.bench and |
self.config == other.config and |
self.time_type == other.time_type and |
self.settings == other.settings) |
- |
+ |
def __hash__(self): |
return (hash(self.bench) ^ |
hash(self.config) ^ |
hash(self.time_type) ^ |
hash(frozenset(self.settings.iteritems()))) |
-def get_latest_revision(directory): |
- """Returns the latest revision number found within this directory. |
+def parse_dir(directory, default_settings, revision, rep): |
+ """Parses bench data from bench logs files. |
+ revision can be either svn revision or git commit hash. |
""" |
- latest_revision_found = -1 |
- for bench_file in os.listdir(directory): |
- file_name_match = re.match('bench_r(\d+)_(\S+)', bench_file) |
- if (file_name_match is None): |
- continue |
- revision = int(file_name_match.group(1)) |
- if revision > latest_revision_found: |
- latest_revision_found = revision |
- if latest_revision_found < 0: |
- return None |
- else: |
- return latest_revision_found |
- |
-def parse_dir(directory, default_settings, oldest_revision, newest_revision, |
- rep): |
- """Parses bench data from files like bench_r<revision>_<scalar>. |
- |
- (str, {str, str}, Number, Number) -> {int:[BenchDataPoints]}""" |
- revision_data_points = {} # {revision : [BenchDataPoints]} |
+ revision_data_points = [] # list of BenchDataPoint |
file_list = os.listdir(directory) |
file_list.sort() |
for bench_file in file_list: |
- file_name_match = re.match('bench_r(\d+)_(\S+)', bench_file) |
- if (file_name_match is None): |
+ scalar_type = None |
+ # Scalar type, if any, is in the bench filename after revision |
+ if (len(revision) > MAX_SVN_REV_LENGTH and |
+ bench_file.startswith('bench_' + revision + '_')): |
+ # The revision is GIT commit hash. |
+ scalar_type = bench_file[len(revision) + len('bench_') + 1:] |
+ elif (bench_file.startswith('bench_r' + revision + '_') and |
+ revision.isdigit()): |
+ # The revision is SVN number |
+ scalar_type = bench_file[len(revision) + len('bench_r') + 1:] |
+ else: |
continue |
- revision = int(file_name_match.group(1)) |
- scalar_type = file_name_match.group(2) |
- |
- if (revision < oldest_revision or revision > newest_revision): |
- continue |
- |
file_handle = open(directory + '/' + bench_file, 'r') |
- if (revision not in revision_data_points): |
- revision_data_points[revision] = [] |
default_settings['scalar'] = scalar_type |
- revision_data_points[revision].extend( |
+ revision_data_points.extend( |
bench_util.parse(default_settings, file_handle, rep)) |
file_handle.close() |
return revision_data_points |
-def add_to_revision_data_points(new_point, revision, revision_data_points): |
- """Add new_point to set of revision_data_points we are building up. |
- """ |
- if (revision not in revision_data_points): |
- revision_data_points[revision] = [] |
- revision_data_points[revision].append(new_point) |
+def create_bench_dict(revision_data_points): |
+ """Convert current revision data into a dictionary of line data. |
-def filter_data_points(unfiltered_revision_data_points): |
- """Filter out any data points that are utterly bogus. |
- |
- Returns (allowed_revision_data_points, ignored_revision_data_points): |
- allowed_revision_data_points: points that survived the filter |
- ignored_revision_data_points: points that did NOT survive the filter |
- """ |
- allowed_revision_data_points = {} # {revision : [BenchDataPoints]} |
- ignored_revision_data_points = {} # {revision : [BenchDataPoints]} |
- revisions = unfiltered_revision_data_points.keys() |
- revisions.sort() |
- for revision in revisions: |
- for point in unfiltered_revision_data_points[revision]: |
- if point.time < MIN_REASONABLE_TIME or point.time > MAX_REASONABLE_TIME: |
- add_to_revision_data_points(point, revision, ignored_revision_data_points) |
- else: |
- add_to_revision_data_points(point, revision, allowed_revision_data_points) |
- return (allowed_revision_data_points, ignored_revision_data_points) |
- |
-def get_abs_path(relative_path): |
- """My own implementation of os.path.abspath() that better handles paths |
- which approach Window's 260-character limit. |
- See https://code.google.com/p/skia/issues/detail?id=674 |
- |
- This implementation adds path components one at a time, resolving the |
- absolute path each time, to take advantage of any chdirs into outer |
- directories that will shorten the total path length. |
- |
- TODO: share a single implementation with upload_to_bucket.py, instead |
- of pasting this same code into both files.""" |
- if os.path.isabs(relative_path): |
- return relative_path |
- path_parts = relative_path.split(os.sep) |
- abs_path = os.path.abspath('.') |
- for path_part in path_parts: |
- abs_path = os.path.abspath(os.path.join(abs_path, path_part)) |
- return abs_path |
- |
-def redirect_stdout(output_path): |
- """Redirect all following stdout to a file. |
- |
- You may be asking yourself, why redirect stdout within Python rather than |
- redirecting the script's output in the calling shell? |
- The answer lies in https://code.google.com/p/skia/issues/detail?id=674 |
- ('buildbot: windows GenerateBenchGraphs step fails due to filename length'): |
- On Windows, we need to generate the absolute path within Python to avoid |
- the operating system's 260-character pathname limit, including chdirs.""" |
- abs_path = get_abs_path(output_path) |
- sys.stdout = open(abs_path, 'w') |
- |
-def create_lines(revision_data_points, settings |
- , bench_of_interest, config_of_interest, time_of_interest |
- , time_to_ignore): |
- """Convert revision data into a dictionary of line data. |
- |
Args: |
- revision_data_points: a dictionary with integer keys (revision #) and a |
- list of bench data points as values |
- settings: a dictionary of setting names to value |
- bench_of_interest: optional filter parameters: which bench type is of |
- interest. If None, process them all. |
- config_of_interest: optional filter parameters: which config type is of |
- interest. If None, process them all. |
- time_of_interest: optional filter parameters: which timer type is of |
- interest. If None, process them all. |
- time_to_ignore: optional timer type to ignore |
+ revision_data_points: a list of bench data points |
Returns: |
a dictionary of this form: |
keys = Label objects |
- values = a list of (x, y) tuples sorted such that x values increase |
- monotonically |
+ values = the corresponding bench value |
""" |
- revisions = revision_data_points.keys() |
- revisions.sort() |
- lines = {} # {Label:[(x,y)] | x[n] <= x[n+1]} |
- for revision in revisions: |
- for point in revision_data_points[revision]: |
- if (bench_of_interest is not None and |
- not bench_of_interest == point.bench): |
- continue |
- |
- if (config_of_interest is not None and |
- not config_of_interest == point.config): |
- continue |
- |
- if (time_of_interest is not None and |
- not time_of_interest == point.time_type): |
- continue |
- elif (time_to_ignore is not None and |
- time_to_ignore == point.time_type): |
- continue |
- |
- skip = False |
- for key, value in settings.items(): |
- if key in point.settings and point.settings[key] != value: |
- skip = True |
- break |
- if skip: |
- continue |
- |
- line_name = Label(point.bench |
- , point.config |
- , point.time_type |
- , point.settings) |
- |
- if line_name not in lines: |
- lines[line_name] = [] |
- |
- lines[line_name].append((revision, point.time)) |
- |
- return lines |
+ bench_dict = {} |
+ for point in revision_data_points: |
+ point_name = Label(point.bench,point.config,point.time_type, |
+ point.settings) |
+ if point_name not in bench_dict: |
+ bench_dict[point_name] = point.time |
+ else: |
+ raise Exception('Duplicate expectation entry: ' + str(point_name)) |
-def bounds(lines): |
- """Finds the bounding rectangle for the lines. |
- |
- {Label:[(x,y)]} -> ((min_x, min_y),(max_x,max_y))""" |
- min_x = bench_util.Max |
- min_y = bench_util.Max |
- max_x = bench_util.Min |
- max_y = bench_util.Min |
- |
- for line in lines.itervalues(): |
- for x, y in line: |
- min_x = min(min_x, x) |
- min_y = min(min_y, y) |
- max_x = max(max_x, x) |
- max_y = max(max_y, y) |
- |
- return ((min_x, min_y), (max_x, max_y)) |
+ return bench_dict |
-def create_regressions(lines, start_x, end_x): |
- """Creates regression data from line segments. |
- |
- ({Label:[(x,y)] | [n].x <= [n+1].x}, Number, Number) |
- -> {Label:LinearRegression}""" |
- regressions = {} # {Label : LinearRegression} |
- |
- for label, line in lines.iteritems(): |
- regression_line = [p for p in line if start_x <= p[0] <= end_x] |
- |
- if (len(regression_line) < 2): |
+def read_expectations(expectations, filename): |
+ """Reads expectations data from file and put in expectations dict.""" |
+ for expectation in open(filename).readlines(): |
+ elements = expectation.strip().split(',') |
+ if not elements[0] or elements[0].startswith('#'): |
continue |
- regression = bench_util.LinearRegression(regression_line) |
- regressions[label] = regression |
- |
- return regressions |
+ if len(elements) != 5: |
+ raise Exception("Invalid expectation line format: %s" % |
+ expectation) |
+ bench_entry = elements[0] + ',' + elements[1] |
+ if bench_entry in expectations: |
+ raise Exception("Dup entries for bench expectation %s" % |
+ bench_entry) |
+ # [<Bench_BmpConfig_TimeType>,<Platform-Alg>] -> (LB, UB) |
+ expectations[bench_entry] = (float(elements[-2]), |
+ float(elements[-1])) |
-def bounds_slope(regressions): |
- """Finds the extreme up and down slopes of a set of linear regressions. |
- |
- ({Label:LinearRegression}) -> (max_up_slope, min_down_slope)""" |
- max_up_slope = 0 |
- min_down_slope = 0 |
- for regression in regressions.itervalues(): |
- min_slope = regression.find_min_slope() |
- max_up_slope = max(max_up_slope, min_slope) |
- min_down_slope = min(min_down_slope, min_slope) |
- |
- return (max_up_slope, min_down_slope) |
+def check_expectations(lines, expectations, revision, key_suffix): |
+ """Check if there are benches in the given revising out of range. |
+ """ |
+ # The platform for this bot, to pass to the dashboard plot. |
+ platform = key_suffix[ : key_suffix.rfind('-')] |
+ exceptions = [] |
+ for line in lines: |
+ line_str = str(line) |
+ line_str = line_str[ : line_str.find('_{')] |
+ bench_platform_key = line_str + ',' + key_suffix |
+ if bench_platform_key not in expectations: |
+ continue |
+ this_bench_value = lines[line] |
+ this_min, this_max = expectations[bench_platform_key] |
+ if this_bench_value < this_min or this_bench_value > this_max: |
+ exception = 'Bench %s value %s out of range [%s, %s].' % ( |
+ bench_platform_key, this_bench_value, this_min, this_max) |
+ exceptions.append(exception) |
+ if exceptions: |
+ raise Exception('Bench values out of range:\n' + |
+ '\n'.join(exceptions)) |
def main(): |
- """Parses command line and writes output.""" |
- |
+ """Parses command line and checks bench expectations.""" |
try: |
- opts, _ = getopt.getopt(sys.argv[1:] |
- , "a:b:c:d:e:f:i:l:m:o:r:s:t:x:y:" |
- , "default-setting=") |
+ opts, _ = getopt.getopt(sys.argv[1:], |
+ "a:b:d:e:r:", |
+ "default-setting=") |
except getopt.GetoptError, err: |
- print str(err) |
+ print str(err) |
usage() |
sys.exit(2) |
- |
+ |
directory = None |
- config_of_interest = None |
- bench_of_interest = None |
- time_of_interest = None |
- time_to_ignore = None |
- output_path = None |
bench_expectations = {} |
- appengine_url = None # used for adding data to appengine datastore |
- rep = None # bench representation algorithm |
- revision_range = '0:' |
- regression_range = '0:' |
- latest_revision = None |
- requested_height = None |
- requested_width = None |
- title = 'Bench graph' |
- settings = {} |
- default_settings = {} |
+ rep = '25th' # bench representation algorithm, default to 25th |
+ rev = None # git commit hash or svn revision number |
+ bot = None |
- def parse_range(range): |
- """Takes '<old>[:<new>]' as a string and returns (old, new). |
- Any revision numbers that are dependent on the latest revision number |
- will be filled in based on latest_revision. |
- """ |
- old, _, new = range.partition(":") |
- old = int(old) |
- if old < 0: |
- old += latest_revision; |
- if not new: |
- new = latest_revision; |
- new = int(new) |
- if new < 0: |
- new += latest_revision; |
- return (old, new) |
- |
- def add_setting(settings, setting): |
- """Takes <key>[=<value>] adds {key:value} or {key:True} to settings.""" |
- name, _, value = setting.partition('=') |
- if not value: |
- settings[name] = True |
- else: |
- settings[name] = value |
- |
- def read_expectations(expectations, filename): |
- """Reads expectations data from file and put in expectations dict.""" |
- for expectation in open(filename).readlines(): |
- elements = expectation.strip().split(',') |
- if not elements[0] or elements[0].startswith('#'): |
- continue |
- if len(elements) != 5: |
- raise Exception("Invalid expectation line format: %s" % |
- expectation) |
- bench_entry = elements[0] + ',' + elements[1] |
- if bench_entry in expectations: |
- raise Exception("Dup entries for bench expectation %s" % |
- bench_entry) |
- # [<Bench_BmpConfig_TimeType>,<Platform-Alg>] -> (LB, UB) |
- expectations[bench_entry] = (float(elements[-2]), |
- float(elements[-1])) |
- |
- def check_expectations(lines, expectations, newest_revision, key_suffix): |
- """Check if there are benches in latest rev outside expected range. |
- For exceptions, also outputs URL link for the dashboard plot. |
- The link history token format here only works for single-line plots. |
- """ |
- # The platform for this bot, to pass to the dashboard plot. |
- platform = key_suffix[ : key_suffix.rfind('-')] |
- # Starting revision for the dashboard plot. |
- start_rev = str(newest_revision - 100) # Displays about 100 revisions. |
- exceptions = [] |
- for line in lines: |
- line_str = str(line) |
- line_str = line_str[ : line_str.find('_{')] |
- bench_platform_key = line_str + ',' + key_suffix |
- this_revision, this_bench_value = lines[line][-1] |
- if (this_revision != newest_revision or |
- bench_platform_key not in expectations): |
- # Skip benches without value for latest revision. |
- continue |
- this_min, this_max = expectations[bench_platform_key] |
- if this_bench_value < this_min or this_bench_value > this_max: |
- link = '' |
- # For skp benches out of range, create dashboard plot link. |
- if line_str.find('.skp_') > 0: |
- # Extract bench and config for dashboard plot. |
- bench, config = line_str.strip('_').split('.skp_') |
- link = ' <a href="' |
- link += 'http://go/skpdash/SkpDash.html#%s~%s~%s~%s" ' % ( |
- start_rev, bench, platform, config) |
- link += 'target="_blank">graph</a>' |
- exception = 'Bench %s value %s out of range [%s, %s].%s' % ( |
- bench_platform_key, this_bench_value, this_min, this_max, |
- link) |
- exceptions.append(exception) |
- if exceptions: |
- raise Exception('Bench values out of range:\n' + |
- '\n'.join(exceptions)) |
- |
- def write_to_appengine(line_data_dict, url, newest_revision, bot): |
- """Writes latest bench values to appengine datastore. |
- line_data_dict: dictionary from create_lines. |
- url: the appengine url used to send bench values to write |
- newest_revision: the latest revision that this script reads |
- bot: the bot platform the bench is run on |
- """ |
- config_data_dic = {} |
- for label in line_data_dict.iterkeys(): |
- if not label.bench.endswith('.skp') or label.time_type: |
- # filter out non-picture and non-walltime benches |
- continue |
- config = label.config |
- rev, val = line_data_dict[label][-1] |
- # This assumes that newest_revision is >= the revision of the last |
- # data point we have for each line. |
- if rev != newest_revision: |
- continue |
- if config not in config_data_dic: |
- config_data_dic[config] = [] |
- config_data_dic[config].append(label.bench.replace('.skp', '') + |
- ':%.2f' % val) |
- for config in config_data_dic: |
- if config_data_dic[config]: |
- data = {'master': 'Skia', 'bot': bot, 'test': config, |
- 'revision': newest_revision, |
- 'benches': ','.join(config_data_dic[config])} |
- req = urllib2.Request(appengine_url, |
- urllib.urlencode({'data': json.dumps(data)})) |
- try: |
- urllib2.urlopen(req) |
- except urllib2.HTTPError, e: |
- sys.stderr.write("HTTPError for JSON data %s: %s\n" % ( |
- data, e)) |
- except urllib2.URLError, e: |
- sys.stderr.write("URLError for JSON data %s: %s\n" % ( |
- data, e)) |
- except httplib.HTTPException, e: |
- sys.stderr.write("HTTPException for JSON data %s: %s\n" % ( |
- data, e)) |
- |
try: |
for option, value in opts: |
if option == "-a": |
- appengine_url = value |
+ rep = value |
elif option == "-b": |
- bench_of_interest = value |
- elif option == "-c": |
- config_of_interest = value |
+ bot = value |
elif option == "-d": |
directory = value |
elif option == "-e": |
read_expectations(bench_expectations, value) |
- elif option == "-f": |
- regression_range = value |
- elif option == "-i": |
- time_to_ignore = value |
- elif option == "-l": |
- title = value |
- elif option == "-m": |
- rep = value |
- elif option == "-o": |
- output_path = value |
- redirect_stdout(output_path) |
elif option == "-r": |
- revision_range = value |
- elif option == "-s": |
- add_setting(settings, value) |
- elif option == "-t": |
- time_of_interest = value |
- elif option == "-x": |
- requested_width = int(value) |
- elif option == "-y": |
- requested_height = int(value) |
- elif option == "--default-setting": |
- add_setting(default_settings, value) |
+ rev = value |
else: |
usage() |
assert False, "unhandled option" |
@@ -496,567 +196,23 @@ |
usage() |
sys.exit(2) |
- if directory is None: |
+ if directory is None or bot is None or rev is None: |
usage() |
sys.exit(2) |
- if time_of_interest: |
- time_to_ignore = None |
+ platform_and_alg = bot + '-' + rep |
- # The title flag (-l) provided in buildbot slave is in the format |
- # Bench_Performance_for_<platform>, and we want to extract <platform> |
- # for use in platform_and_alg to track matching benches later. If title flag |
- # is not in this format, there may be no matching benches in the file |
- # provided by the expectation_file flag (-e). |
- bot = title # To store the platform as bot name |
- platform_and_alg = title |
- if platform_and_alg.startswith(TITLE_PREAMBLE): |
- bot = platform_and_alg[TITLE_PREAMBLE_LENGTH:] |
- platform_and_alg = bot + '-' + rep |
- title += ' [representation: %s]' % rep |
+ data_points = parse_dir(directory, |
+ {}, # Sets default settings to empty. |
+ rev, |
+ rep) |
- latest_revision = get_latest_revision(directory) |
- oldest_revision, newest_revision = parse_range(revision_range) |
- oldest_regression, newest_regression = parse_range(regression_range) |
+ bench_dict = create_bench_dict(data_points) |
- unfiltered_revision_data_points = parse_dir(directory |
- , default_settings |
- , oldest_revision |
- , newest_revision |
- , rep) |
- |
- # Filter out any data points that are utterly bogus... make sure to report |
- # that we did so later! |
- (allowed_revision_data_points, ignored_revision_data_points) = filter_data_points( |
- unfiltered_revision_data_points) |
- |
- # Update oldest_revision and newest_revision based on the data we could find |
- all_revision_numbers = allowed_revision_data_points.keys() |
- oldest_revision = min(all_revision_numbers) |
- newest_revision = max(all_revision_numbers) |
- |
- lines = create_lines(allowed_revision_data_points |
- , settings |
- , bench_of_interest |
- , config_of_interest |
- , time_of_interest |
- , time_to_ignore) |
- |
- regressions = create_regressions(lines |
- , oldest_regression |
- , newest_regression) |
- |
- if output_path: |
- output_xhtml(lines, oldest_revision, newest_revision, |
- ignored_revision_data_points, regressions, requested_width, |
- requested_height, title) |
- |
- if appengine_url: |
- write_to_appengine(lines, appengine_url, newest_revision, bot) |
- |
if bench_expectations: |
- check_expectations(lines, bench_expectations, newest_revision, |
+ check_expectations(bench_dict, bench_expectations, rev, |
platform_and_alg) |
-def qa(out): |
- """Stringify input and quote as an xml attribute.""" |
- return xml.sax.saxutils.quoteattr(str(out)) |
-def qe(out): |
- """Stringify input and escape as xml data.""" |
- return xml.sax.saxutils.escape(str(out)) |
-def create_select(qualifier, lines, select_id=None): |
- """Output select with options showing lines which qualifier maps to it. |
- |
- ((Label) -> str, {Label:_}, str?) -> _""" |
- options = {} #{ option : [Label]} |
- for label in lines.keys(): |
- option = qualifier(label) |
- if (option not in options): |
- options[option] = [] |
- options[option].append(label) |
- option_list = list(options.keys()) |
- option_list.sort() |
- print '<select class="lines"', |
- if select_id is not None: |
- print 'id=%s' % qa(select_id) |
- print 'multiple="true" size="10" onchange="updateSvg();">' |
- for option in option_list: |
- print '<option value=' + qa('[' + |
- reduce(lambda x,y:x+json.dumps(str(y))+',',options[option],"")[0:-1] |
- + ']') + '>'+qe(option)+'</option>' |
- print '</select>' |
- |
-def output_ignored_data_points_warning(ignored_revision_data_points): |
- """Write description of ignored_revision_data_points to stdout as xhtml. |
- """ |
- num_ignored_points = 0 |
- description = '' |
- revisions = ignored_revision_data_points.keys() |
- if revisions: |
- revisions.sort() |
- revisions.reverse() |
- for revision in revisions: |
- num_ignored_points += len(ignored_revision_data_points[revision]) |
- points_at_this_revision = [] |
- for point in ignored_revision_data_points[revision]: |
- points_at_this_revision.append(point.bench) |
- points_at_this_revision.sort() |
- description += 'r%d: %s\n' % (revision, points_at_this_revision) |
- if num_ignored_points == 0: |
- print 'Did not discard any data points; all were within the range [%d-%d]' % ( |
- MIN_REASONABLE_TIME, MAX_REASONABLE_TIME) |
- else: |
- print '<table width="100%" bgcolor="ff0000"><tr><td align="center">' |
- print 'Discarded %d data points outside of range [%d-%d]' % ( |
- num_ignored_points, MIN_REASONABLE_TIME, MAX_REASONABLE_TIME) |
- print '</td></tr><tr><td width="100%" align="center">' |
- print ('<textarea rows="4" style="width:97%" readonly="true" wrap="off">' |
- + qe(description) + '</textarea>') |
- print '</td></tr></table>' |
- |
-def output_xhtml(lines, oldest_revision, newest_revision, ignored_revision_data_points, |
- regressions, requested_width, requested_height, title): |
- """Outputs an svg/xhtml view of the data.""" |
- print '<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN"', |
- print '"http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">' |
- print '<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en">' |
- print '<head>' |
- print '<title>%s</title>' % qe(title) |
- print '</head>' |
- print '<body>' |
- |
- output_svg(lines, regressions, requested_width, requested_height) |
- |
- #output the manipulation controls |
- print """ |
-<script type="text/javascript">//<![CDATA[ |
- function getElementsByClass(node, searchClass, tag) { |
- var classElements = new Array(); |
- var elements = node.getElementsByTagName(tag); |
- var pattern = new RegExp("^|\\s"+searchClass+"\\s|$"); |
- for (var i = 0, elementsFound = 0; i < elements.length; ++i) { |
- if (pattern.test(elements[i].className)) { |
- classElements[elementsFound] = elements[i]; |
- ++elementsFound; |
- } |
- } |
- return classElements; |
- } |
- function getAllLines() { |
- var selectElem = document.getElementById('benchSelect'); |
- var linesObj = {}; |
- for (var i = 0; i < selectElem.options.length; ++i) { |
- var lines = JSON.parse(selectElem.options[i].value); |
- for (var j = 0; j < lines.length; ++j) { |
- linesObj[lines[j]] = true; |
- } |
- } |
- return linesObj; |
- } |
- function getOptions(selectElem) { |
- var linesSelectedObj = {}; |
- for (var i = 0; i < selectElem.options.length; ++i) { |
- if (!selectElem.options[i].selected) continue; |
- |
- var linesSelected = JSON.parse(selectElem.options[i].value); |
- for (var j = 0; j < linesSelected.length; ++j) { |
- linesSelectedObj[linesSelected[j]] = true; |
- } |
- } |
- return linesSelectedObj; |
- } |
- function objectEmpty(obj) { |
- for (var p in obj) { |
- return false; |
- } |
- return true; |
- } |
- function markSelectedLines(selectElem, allLines) { |
- var linesSelected = getOptions(selectElem); |
- if (!objectEmpty(linesSelected)) { |
- for (var line in allLines) { |
- allLines[line] &= (linesSelected[line] == true); |
- } |
- } |
- } |
- function updateSvg() { |
- var allLines = getAllLines(); |
- |
- var selects = getElementsByClass(document, 'lines', 'select'); |
- for (var i = 0; i < selects.length; ++i) { |
- markSelectedLines(selects[i], allLines); |
- } |
- |
- for (var line in allLines) { |
- var svgLine = document.getElementById(line); |
- var display = (allLines[line] ? 'inline' : 'none'); |
- svgLine.setAttributeNS(null,'display', display); |
- } |
- } |
- |
- function mark(markerId) { |
- for (var line in getAllLines()) { |
- var svgLineGroup = document.getElementById(line); |
- var display = svgLineGroup.getAttributeNS(null,'display'); |
- if (display == null || display == "" || display != "none") { |
- var svgLine = document.getElementById(line+'_line'); |
- if (markerId == null) { |
- svgLine.removeAttributeNS(null,'marker-mid'); |
- } else { |
- svgLine.setAttributeNS(null,'marker-mid', markerId); |
- } |
- } |
- } |
- } |
-//]]></script>""" |
- |
- all_settings = {} |
- variant_settings = set() |
- for label in lines.keys(): |
- for key, value in label.settings.items(): |
- if key not in all_settings: |
- all_settings[key] = value |
- elif all_settings[key] != value: |
- variant_settings.add(key) |
- |
- print '<table border="0" width="%s">' % requested_width |
- #output column headers |
- print """ |
-<tr valign="top"><td width="50%"> |
-<table border="0" width="100%"> |
-<tr><td align="center"><table border="0"> |
-<form> |
-<tr valign="bottom" align="center"> |
-<td width="1">Bench Type</td> |
-<td width="1">Bitmap Config</td> |
-<td width="1">Timer Type (Cpu/Gpu/wall)</td> |
-""" |
- |
- for k in variant_settings: |
- print '<td width="1">%s</td>' % qe(k) |
- |
- print '<td width="1"><!--buttons--></td></tr>' |
- |
- #output column contents |
- print '<tr valign="top" align="center">' |
- print '<td width="1">' |
- create_select(lambda l: l.bench, lines, 'benchSelect') |
- print '</td><td width="1">' |
- create_select(lambda l: l.config, lines) |
- print '</td><td width="1">' |
- create_select(lambda l: l.time_type, lines) |
- |
- for k in variant_settings: |
- print '</td><td width="1">' |
- create_select(lambda l: l.settings.get(k, " "), lines) |
- |
- print '</td><td width="1"><button type="button"', |
- print 'onclick=%s' % qa("mark('url(#circleMark)'); return false;"), |
- print '>Mark Points</button>' |
- print '<button type="button" onclick="mark(null);">Clear Points</button>' |
- print '</td>' |
- print """ |
-</tr> |
-</form> |
-</table></td></tr> |
-<tr><td align="center"> |
-<hr /> |
-""" |
- |
- output_ignored_data_points_warning(ignored_revision_data_points) |
- print '</td></tr></table>' |
- print '</td><td width="2%"><!--gutter--></td>' |
- |
- print '<td><table border="0">' |
- print '<tr><td align="center">%s<br></br>revisions r%s - r%s</td></tr>' % ( |
- qe(title), |
- bench_util.CreateRevisionLink(oldest_revision), |
- bench_util.CreateRevisionLink(newest_revision)) |
- print """ |
-<tr><td align="left"> |
-<p>Brighter red indicates tests that have gotten worse; brighter green |
-indicates tests that have gotten better.</p> |
-<p>To highlight individual tests, hold down CONTROL and mouse over |
-graph lines.</p> |
-<p>To highlight revision numbers, hold down SHIFT and mouse over |
-the graph area.</p> |
-<p>To only show certain tests on the graph, select any combination of |
-tests in the selectors at left. (To show all, select all.)</p> |
-<p>Use buttons at left to mark/clear points on the lines for selected |
-benchmarks.</p> |
-</td></tr> |
-</table> |
- |
-</td> |
-</tr> |
-</table> |
-</body> |
-</html>""" |
- |
-def compute_size(requested_width, requested_height, rev_width, time_height): |
- """Converts potentially empty requested size into a concrete size. |
- |
- (Number?, Number?) -> (Number, Number)""" |
- pic_width = 0 |
- pic_height = 0 |
- if (requested_width is not None and requested_height is not None): |
- pic_height = requested_height |
- pic_width = requested_width |
- |
- elif (requested_width is not None): |
- pic_width = requested_width |
- pic_height = pic_width * (float(time_height) / rev_width) |
- |
- elif (requested_height is not None): |
- pic_height = requested_height |
- pic_width = pic_height * (float(rev_width) / time_height) |
- |
- else: |
- pic_height = 800 |
- pic_width = max(rev_width*3 |
- , pic_height * (float(rev_width) / time_height)) |
- |
- return (pic_width, pic_height) |
- |
-def output_svg(lines, regressions, requested_width, requested_height): |
- """Outputs an svg view of the data.""" |
- |
- (global_min_x, _), (global_max_x, global_max_y) = bounds(lines) |
- max_up_slope, min_down_slope = bounds_slope(regressions) |
- |
- #output |
- global_min_y = 0 |
- x = global_min_x |
- y = global_min_y |
- w = global_max_x - global_min_x |
- h = global_max_y - global_min_y |
- font_size = 16 |
- line_width = 2 |
- |
- # If there is nothing to see, don't try to draw anything. |
- if w == 0 or h == 0: |
- return |
- |
- pic_width, pic_height = compute_size(requested_width, requested_height |
- , w, h) |
- |
- def cw(w1): |
- """Converts a revision difference to display width.""" |
- return (pic_width / float(w)) * w1 |
- def cx(x): |
- """Converts a revision to a horizontal display position.""" |
- return cw(x - global_min_x) |
- |
- def ch(h1): |
- """Converts a time difference to a display height.""" |
- return -(pic_height / float(h)) * h1 |
- def cy(y): |
- """Converts a time to a vertical display position.""" |
- return pic_height + ch(y - global_min_y) |
- |
- print '<!--Picture height %.2f corresponds to bench value %.2f.-->' % ( |
- pic_height, h) |
- print '<svg', |
- print 'width=%s' % qa(str(pic_width)+'px') |
- print 'height=%s' % qa(str(pic_height)+'px') |
- print 'viewBox="0 0 %s %s"' % (str(pic_width), str(pic_height)) |
- print 'onclick=%s' % qa( |
- "var event = arguments[0] || window.event;" |
- " if (event.shiftKey) { highlightRevision(null); }" |
- " if (event.ctrlKey) { highlight(null); }" |
- " return false;") |
- print 'xmlns="http://www.w3.org/2000/svg"' |
- print 'xmlns:xlink="http://www.w3.org/1999/xlink">' |
- |
- print """ |
-<defs> |
- <marker id="circleMark" |
- viewBox="0 0 2 2" refX="1" refY="1" |
- markerUnits="strokeWidth" |
- markerWidth="2" markerHeight="2" |
- orient="0"> |
- <circle cx="1" cy="1" r="1"/> |
- </marker> |
-</defs>""" |
- |
- #output the revisions |
- print """ |
-<script type="text/javascript">//<![CDATA[ |
- var previousRevision; |
- var previousRevisionFill; |
- var previousRevisionStroke |
- function highlightRevision(id) { |
- if (previousRevision == id) return; |
- |
- document.getElementById('revision').firstChild.nodeValue = 'r' + id; |
- document.getElementById('rev_link').setAttribute('xlink:href', |
- 'http://code.google.com/p/skia/source/detail?r=' + id); |
- |
- var preRevision = document.getElementById(previousRevision); |
- if (preRevision) { |
- preRevision.setAttributeNS(null,'fill', previousRevisionFill); |
- preRevision.setAttributeNS(null,'stroke', previousRevisionStroke); |
- } |
- |
- var revision = document.getElementById(id); |
- previousRevision = id; |
- if (revision) { |
- previousRevisionFill = revision.getAttributeNS(null,'fill'); |
- revision.setAttributeNS(null,'fill','rgb(100%, 95%, 95%)'); |
- |
- previousRevisionStroke = revision.getAttributeNS(null,'stroke'); |
- revision.setAttributeNS(null,'stroke','rgb(100%, 90%, 90%)'); |
- } |
- } |
-//]]></script>""" |
- |
- def print_rect(x, y, w, h, revision): |
- """Outputs a revision rectangle in display space, |
- taking arguments in revision space.""" |
- disp_y = cy(y) |
- disp_h = ch(h) |
- if disp_h < 0: |
- disp_y += disp_h |
- disp_h = -disp_h |
- |
- print '<rect id=%s x=%s y=%s' % (qa(revision), qa(cx(x)), qa(disp_y),), |
- print 'width=%s height=%s' % (qa(cw(w)), qa(disp_h),), |
- print 'fill="white"', |
- print 'stroke="rgb(98%%,98%%,88%%)" stroke-width=%s' % qa(line_width), |
- print 'onmouseover=%s' % qa( |
- "var event = arguments[0] || window.event;" |
- " if (event.shiftKey) {" |
- " highlightRevision('"+str(revision)+"');" |
- " return false;" |
- " }"), |
- print ' />' |
- |
- xes = set() |
- for line in lines.itervalues(): |
- for point in line: |
- xes.add(point[0]) |
- revisions = list(xes) |
- revisions.sort() |
- |
- left = x |
- current_revision = revisions[0] |
- for next_revision in revisions[1:]: |
- width = (((next_revision - current_revision) / 2.0) |
- + (current_revision - left)) |
- print_rect(left, y, width, h, current_revision) |
- left += width |
- current_revision = next_revision |
- print_rect(left, y, x+w - left, h, current_revision) |
- |
- #output the lines |
- print """ |
-<script type="text/javascript">//<![CDATA[ |
- var previous; |
- var previousColor; |
- var previousOpacity; |
- function highlight(id) { |
- if (previous == id) return; |
- |
- document.getElementById('label').firstChild.nodeValue = id; |
- |
- var preGroup = document.getElementById(previous); |
- if (preGroup) { |
- var preLine = document.getElementById(previous+'_line'); |
- preLine.setAttributeNS(null,'stroke', previousColor); |
- preLine.setAttributeNS(null,'opacity', previousOpacity); |
- |
- var preSlope = document.getElementById(previous+'_linear'); |
- if (preSlope) { |
- preSlope.setAttributeNS(null,'visibility', 'hidden'); |
- } |
- } |
- |
- var group = document.getElementById(id); |
- previous = id; |
- if (group) { |
- group.parentNode.appendChild(group); |
- |
- var line = document.getElementById(id+'_line'); |
- previousColor = line.getAttributeNS(null,'stroke'); |
- previousOpacity = line.getAttributeNS(null,'opacity'); |
- line.setAttributeNS(null,'stroke', 'blue'); |
- line.setAttributeNS(null,'opacity', '1'); |
- |
- var slope = document.getElementById(id+'_linear'); |
- if (slope) { |
- slope.setAttributeNS(null,'visibility', 'visible'); |
- } |
- } |
- } |
-//]]></script>""" |
- |
- # Add a new element to each item in the 'lines' list: the label in string |
- # form. Then use that element to sort the list. |
- sorted_lines = [] |
- for label, line in lines.items(): |
- sorted_lines.append([str(label), label, line]) |
- sorted_lines.sort() |
- |
- for label_as_string, label, line in sorted_lines: |
- print '<g id=%s>' % qa(label_as_string) |
- r = 128 |
- g = 128 |
- b = 128 |
- a = .10 |
- if label in regressions: |
- regression = regressions[label] |
- min_slope = regression.find_min_slope() |
- if min_slope < 0: |
- d = max(0, (min_slope / min_down_slope)) |
- g += int(d*128) |
- a += d*0.9 |
- elif min_slope > 0: |
- d = max(0, (min_slope / max_up_slope)) |
- r += int(d*128) |
- a += d*0.9 |
- |
- slope = regression.slope |
- intercept = regression.intercept |
- min_x = regression.min_x |
- max_x = regression.max_x |
- print '<polyline id=%s' % qa(str(label)+'_linear'), |
- print 'fill="none" stroke="yellow"', |
- print 'stroke-width=%s' % qa(abs(ch(regression.serror*2))), |
- print 'opacity="0.5" pointer-events="none" visibility="hidden"', |
- print 'points="', |
- print '%s,%s' % (str(cx(min_x)), str(cy(slope*min_x + intercept))), |
- print '%s,%s' % (str(cx(max_x)), str(cy(slope*max_x + intercept))), |
- print '"/>' |
- |
- print '<polyline id=%s' % qa(str(label)+'_line'), |
- print 'onmouseover=%s' % qa( |
- "var event = arguments[0] || window.event;" |
- " if (event.ctrlKey) {" |
- " highlight('"+str(label).replace("'", "\\'")+"');" |
- " return false;" |
- " }"), |
- print 'fill="none" stroke="rgb(%s,%s,%s)"' % (str(r), str(g), str(b)), |
- print 'stroke-width=%s' % qa(line_width), |
- print 'opacity=%s' % qa(a), |
- print 'points="', |
- for point in line: |
- print '%s,%s' % (str(cx(point[0])), str(cy(point[1]))), |
- print '"/>' |
- |
- print '</g>' |
- |
- #output the labels |
- print '<text id="label" x="0" y=%s' % qa(font_size), |
- print 'font-size=%s> </text>' % qa(font_size) |
- |
- print '<a id="rev_link" xlink:href="" target="_top">' |
- print '<text id="revision" x="0" y=%s style="' % qa(font_size*2) |
- print 'font-size: %s; ' % qe(font_size) |
- print 'stroke: #0000dd; text-decoration: underline; ' |
- print '"> </text></a>' |
- |
- print '</svg>' |
- |
if __name__ == "__main__": |
main() |