Index: tools/run_benchmarks.py |
diff --git a/tools/run_benchmarks.py b/tools/run_benchmarks.py |
deleted file mode 100755 |
index cc0bb2c5b87f15ae9e386101e37e6fbdec58d3ca..0000000000000000000000000000000000000000 |
--- a/tools/run_benchmarks.py |
+++ /dev/null |
@@ -1,488 +0,0 @@ |
-#!/usr/bin/env python |
-# Copyright 2014 the V8 project authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
- |
-""" |
-Performance runner for d8. |
- |
-Call e.g. with tools/run-benchmarks.py --arch ia32 some_suite.json |
- |
-The suite json format is expected to be: |
-{ |
- "path": <relative path chunks to benchmark resources and main file>, |
- "name": <optional suite name, file name is default>, |
- "archs": [<architecture name for which this suite is run>, ...], |
- "binary": <name of binary to run, default "d8">, |
- "flags": [<flag to d8>, ...], |
- "run_count": <how often will this suite run (optional)>, |
- "run_count_XXX": <how often will this suite run for arch XXX (optional)>, |
- "resources": [<js file to be loaded before main>, ...] |
- "main": <main js benchmark runner file>, |
- "results_regexp": <optional regexp>, |
- "results_processor": <optional python results processor script>, |
- "units": <the unit specification for the performance dashboard>, |
- "benchmarks": [ |
- { |
- "name": <name of the benchmark>, |
- "results_regexp": <optional more specific regexp>, |
- "results_processor": <optional python results processor script>, |
- "units": <the unit specification for the performance dashboard>, |
- }, ... |
- ] |
-} |
- |
-The benchmarks field can also nest other suites in arbitrary depth. A suite |
-with a "main" file is a leaf suite that can contain one more level of |
-benchmarks. |
- |
-A suite's results_regexp is expected to have one string place holder |
-"%s" for the benchmark name. A benchmark's results_regexp overwrites suite |
-defaults. |
- |
-A suite's results_processor may point to an optional python script. If |
-specified, it is called after running the benchmarks like this (with a path |
-relatve to the suite level's path): |
-<results_processor file> <same flags as for d8> <suite level name> <output> |
- |
-The <output> is a temporary file containing d8 output. The results_regexp will |
-be applied to the output of this script. |
- |
-A suite without "benchmarks" is considered a benchmark itself. |
- |
-Full example (suite with one runner): |
-{ |
- "path": ["."], |
- "flags": ["--expose-gc"], |
- "archs": ["ia32", "x64"], |
- "run_count": 5, |
- "run_count_ia32": 3, |
- "main": "run.js", |
- "results_regexp": "^%s: (.+)$", |
- "units": "score", |
- "benchmarks": [ |
- {"name": "Richards"}, |
- {"name": "DeltaBlue"}, |
- {"name": "NavierStokes", |
- "results_regexp": "^NavierStokes: (.+)$"} |
- ] |
-} |
- |
-Full example (suite with several runners): |
-{ |
- "path": ["."], |
- "flags": ["--expose-gc"], |
- "archs": ["ia32", "x64"], |
- "run_count": 5, |
- "units": "score", |
- "benchmarks": [ |
- {"name": "Richards", |
- "path": ["richards"], |
- "main": "run.js", |
- "run_count": 3, |
- "results_regexp": "^Richards: (.+)$"}, |
- {"name": "NavierStokes", |
- "path": ["navier_stokes"], |
- "main": "run.js", |
- "results_regexp": "^NavierStokes: (.+)$"} |
- ] |
-} |
- |
-Path pieces are concatenated. D8 is always run with the suite's path as cwd. |
-""" |
- |
-import json |
-import math |
-import optparse |
-import os |
-import re |
-import sys |
- |
-from testrunner.local import commands |
-from testrunner.local import utils |
- |
-ARCH_GUESS = utils.DefaultArch() |
-SUPPORTED_ARCHS = ["android_arm", |
- "android_arm64", |
- "android_ia32", |
- "arm", |
- "ia32", |
- "mips", |
- "mipsel", |
- "nacl_ia32", |
- "nacl_x64", |
- "x64", |
- "arm64"] |
- |
-GENERIC_RESULTS_RE = re.compile( |
- r"^Trace\(([^\)]+)\), Result\(([^\)]+)\), StdDev\(([^\)]+)\)$") |
- |
- |
-def GeometricMean(values): |
- """Returns the geometric mean of a list of values. |
- |
- The mean is calculated using log to avoid overflow. |
- """ |
- values = map(float, values) |
- return str(math.exp(sum(map(math.log, values)) / len(values))) |
- |
- |
-class Results(object): |
- """Place holder for result traces.""" |
- def __init__(self, traces=None, errors=None): |
- self.traces = traces or [] |
- self.errors = errors or [] |
- |
- def ToDict(self): |
- return {"traces": self.traces, "errors": self.errors} |
- |
- def WriteToFile(self, file_name): |
- with open(file_name, "w") as f: |
- f.write(json.dumps(self.ToDict())) |
- |
- def __add__(self, other): |
- self.traces += other.traces |
- self.errors += other.errors |
- return self |
- |
- def __str__(self): # pragma: no cover |
- return str(self.ToDict()) |
- |
- |
-class Node(object): |
- """Represents a node in the benchmark suite tree structure.""" |
- def __init__(self, *args): |
- self._children = [] |
- |
- def AppendChild(self, child): |
- self._children.append(child) |
- |
- |
-class DefaultSentinel(Node): |
- """Fake parent node with all default values.""" |
- def __init__(self): |
- super(DefaultSentinel, self).__init__() |
- self.binary = "d8" |
- self.run_count = 10 |
- self.path = [] |
- self.graphs = [] |
- self.flags = [] |
- self.resources = [] |
- self.results_regexp = None |
- self.stddev_regexp = None |
- self.units = "score" |
- self.total = False |
- |
- |
-class Graph(Node): |
- """Represents a benchmark suite definition. |
- |
- Can either be a leaf or an inner node that provides default values. |
- """ |
- def __init__(self, suite, parent, arch): |
- super(Graph, self).__init__() |
- self._suite = suite |
- |
- assert isinstance(suite.get("path", []), list) |
- assert isinstance(suite["name"], basestring) |
- assert isinstance(suite.get("flags", []), list) |
- assert isinstance(suite.get("resources", []), list) |
- |
- # Accumulated values. |
- self.path = parent.path[:] + suite.get("path", []) |
- self.graphs = parent.graphs[:] + [suite["name"]] |
- self.flags = parent.flags[:] + suite.get("flags", []) |
- self.resources = parent.resources[:] + suite.get("resources", []) |
- |
- # Descrete values (with parent defaults). |
- self.binary = suite.get("binary", parent.binary) |
- self.run_count = suite.get("run_count", parent.run_count) |
- self.run_count = suite.get("run_count_%s" % arch, self.run_count) |
- self.units = suite.get("units", parent.units) |
- self.total = suite.get("total", parent.total) |
- |
- # A regular expression for results. If the parent graph provides a |
- # regexp and the current suite has none, a string place holder for the |
- # suite name is expected. |
- # TODO(machenbach): Currently that makes only sense for the leaf level. |
- # Multiple place holders for multiple levels are not supported. |
- if parent.results_regexp: |
- regexp_default = parent.results_regexp % re.escape(suite["name"]) |
- else: |
- regexp_default = None |
- self.results_regexp = suite.get("results_regexp", regexp_default) |
- |
- # A similar regular expression for the standard deviation (optional). |
- if parent.stddev_regexp: |
- stddev_default = parent.stddev_regexp % re.escape(suite["name"]) |
- else: |
- stddev_default = None |
- self.stddev_regexp = suite.get("stddev_regexp", stddev_default) |
- |
- |
-class Trace(Graph): |
- """Represents a leaf in the benchmark suite tree structure. |
- |
- Handles collection of measurements. |
- """ |
- def __init__(self, suite, parent, arch): |
- super(Trace, self).__init__(suite, parent, arch) |
- assert self.results_regexp |
- self.results = [] |
- self.errors = [] |
- self.stddev = "" |
- |
- def ConsumeOutput(self, stdout): |
- try: |
- self.results.append( |
- re.search(self.results_regexp, stdout, re.M).group(1)) |
- except: |
- self.errors.append("Regexp \"%s\" didn't match for benchmark %s." |
- % (self.results_regexp, self.graphs[-1])) |
- |
- try: |
- if self.stddev_regexp and self.stddev: |
- self.errors.append("Benchmark %s should only run once since a stddev " |
- "is provided by the benchmark." % self.graphs[-1]) |
- if self.stddev_regexp: |
- self.stddev = re.search(self.stddev_regexp, stdout, re.M).group(1) |
- except: |
- self.errors.append("Regexp \"%s\" didn't match for benchmark %s." |
- % (self.stddev_regexp, self.graphs[-1])) |
- |
- def GetResults(self): |
- return Results([{ |
- "graphs": self.graphs, |
- "units": self.units, |
- "results": self.results, |
- "stddev": self.stddev, |
- }], self.errors) |
- |
- |
-class Runnable(Graph): |
- """Represents a runnable benchmark suite definition (i.e. has a main file). |
- """ |
- @property |
- def main(self): |
- return self._suite.get("main", "") |
- |
- def ChangeCWD(self, suite_path): |
- """Changes the cwd to to path defined in the current graph. |
- |
- The benchmarks are supposed to be relative to the suite configuration. |
- """ |
- suite_dir = os.path.abspath(os.path.dirname(suite_path)) |
- bench_dir = os.path.normpath(os.path.join(*self.path)) |
- os.chdir(os.path.join(suite_dir, bench_dir)) |
- |
- def GetCommand(self, shell_dir): |
- # TODO(machenbach): This requires +.exe if run on windows. |
- return ( |
- [os.path.join(shell_dir, self.binary)] + |
- self.flags + |
- self.resources + |
- [self.main] |
- ) |
- |
- def Run(self, runner): |
- """Iterates over several runs and handles the output for all traces.""" |
- for stdout in runner(): |
- for trace in self._children: |
- trace.ConsumeOutput(stdout) |
- res = reduce(lambda r, t: r + t.GetResults(), self._children, Results()) |
- |
- if not res.traces or not self.total: |
- return res |
- |
- # Assume all traces have the same structure. |
- if len(set(map(lambda t: len(t["results"]), res.traces))) != 1: |
- res.errors.append("Not all traces have the same number of results.") |
- return res |
- |
- # Calculate the geometric means for all traces. Above we made sure that |
- # there is at least one trace and that the number of results is the same |
- # for each trace. |
- n_results = len(res.traces[0]["results"]) |
- total_results = [GeometricMean(t["results"][i] for t in res.traces) |
- for i in range(0, n_results)] |
- res.traces.append({ |
- "graphs": self.graphs + ["Total"], |
- "units": res.traces[0]["units"], |
- "results": total_results, |
- "stddev": "", |
- }) |
- return res |
- |
-class RunnableTrace(Trace, Runnable): |
- """Represents a runnable benchmark suite definition that is a leaf.""" |
- def __init__(self, suite, parent, arch): |
- super(RunnableTrace, self).__init__(suite, parent, arch) |
- |
- def Run(self, runner): |
- """Iterates over several runs and handles the output.""" |
- for stdout in runner(): |
- self.ConsumeOutput(stdout) |
- return self.GetResults() |
- |
- |
-class RunnableGeneric(Runnable): |
- """Represents a runnable benchmark suite definition with generic traces.""" |
- def __init__(self, suite, parent, arch): |
- super(RunnableGeneric, self).__init__(suite, parent, arch) |
- |
- def Run(self, runner): |
- """Iterates over several runs and handles the output.""" |
- traces = {} |
- for stdout in runner(): |
- for line in stdout.strip().splitlines(): |
- match = GENERIC_RESULTS_RE.match(line) |
- if match: |
- trace = match.group(1) |
- result = match.group(2) |
- stddev = match.group(3) |
- trace_result = traces.setdefault(trace, Results([{ |
- "graphs": self.graphs + [trace], |
- "units": self.units, |
- "results": [], |
- "stddev": "", |
- }], [])) |
- trace_result.traces[0]["results"].append(result) |
- trace_result.traces[0]["stddev"] = stddev |
- |
- return reduce(lambda r, t: r + t, traces.itervalues(), Results()) |
- |
- |
-def MakeGraph(suite, arch, parent): |
- """Factory method for making graph objects.""" |
- if isinstance(parent, Runnable): |
- # Below a runnable can only be traces. |
- return Trace(suite, parent, arch) |
- elif suite.get("main"): |
- # A main file makes this graph runnable. |
- if suite.get("benchmarks"): |
- # This graph has subbenchmarks (traces). |
- return Runnable(suite, parent, arch) |
- else: |
- # This graph has no subbenchmarks, it's a leaf. |
- return RunnableTrace(suite, parent, arch) |
- elif suite.get("generic"): |
- # This is a generic suite definition. It is either a runnable executable |
- # or has a main js file. |
- return RunnableGeneric(suite, parent, arch) |
- elif suite.get("benchmarks"): |
- # This is neither a leaf nor a runnable. |
- return Graph(suite, parent, arch) |
- else: # pragma: no cover |
- raise Exception("Invalid benchmark suite configuration.") |
- |
- |
-def BuildGraphs(suite, arch, parent=None): |
- """Builds a tree structure of graph objects that corresponds to the suite |
- configuration. |
- """ |
- parent = parent or DefaultSentinel() |
- |
- # TODO(machenbach): Implement notion of cpu type? |
- if arch not in suite.get("archs", ["ia32", "x64"]): |
- return None |
- |
- graph = MakeGraph(suite, arch, parent) |
- for subsuite in suite.get("benchmarks", []): |
- BuildGraphs(subsuite, arch, graph) |
- parent.AppendChild(graph) |
- return graph |
- |
- |
-def FlattenRunnables(node): |
- """Generator that traverses the tree structure and iterates over all |
- runnables. |
- """ |
- if isinstance(node, Runnable): |
- yield node |
- elif isinstance(node, Node): |
- for child in node._children: |
- for result in FlattenRunnables(child): |
- yield result |
- else: # pragma: no cover |
- raise Exception("Invalid benchmark suite configuration.") |
- |
- |
-# TODO: Implement results_processor. |
-def Main(args): |
- parser = optparse.OptionParser() |
- parser.add_option("--arch", |
- help=("The architecture to run tests for, " |
- "'auto' or 'native' for auto-detect"), |
- default="x64") |
- parser.add_option("--buildbot", |
- help="Adapt to path structure used on buildbots", |
- default=False, action="store_true") |
- parser.add_option("--json-test-results", |
- help="Path to a file for storing json results.") |
- parser.add_option("--outdir", help="Base directory with compile output", |
- default="out") |
- (options, args) = parser.parse_args(args) |
- |
- if len(args) == 0: # pragma: no cover |
- parser.print_help() |
- return 1 |
- |
- if options.arch in ["auto", "native"]: # pragma: no cover |
- options.arch = ARCH_GUESS |
- |
- if not options.arch in SUPPORTED_ARCHS: # pragma: no cover |
- print "Unknown architecture %s" % options.arch |
- return 1 |
- |
- workspace = os.path.abspath(os.path.join(os.path.dirname(__file__), "..")) |
- |
- if options.buildbot: |
- shell_dir = os.path.join(workspace, options.outdir, "Release") |
- else: |
- shell_dir = os.path.join(workspace, options.outdir, |
- "%s.release" % options.arch) |
- |
- results = Results() |
- for path in args: |
- path = os.path.abspath(path) |
- |
- if not os.path.exists(path): # pragma: no cover |
- results.errors.append("Benchmark file %s does not exist." % path) |
- continue |
- |
- with open(path) as f: |
- suite = json.loads(f.read()) |
- |
- # If no name is given, default to the file name without .json. |
- suite.setdefault("name", os.path.splitext(os.path.basename(path))[0]) |
- |
- for runnable in FlattenRunnables(BuildGraphs(suite, options.arch)): |
- print ">>> Running suite: %s" % "/".join(runnable.graphs) |
- runnable.ChangeCWD(path) |
- |
- def Runner(): |
- """Output generator that reruns several times.""" |
- for i in xrange(0, max(1, runnable.run_count)): |
- # TODO(machenbach): Make timeout configurable in the suite definition. |
- # Allow timeout per arch like with run_count per arch. |
- output = commands.Execute(runnable.GetCommand(shell_dir), timeout=60) |
- print ">>> Stdout (#%d):" % (i + 1) |
- print output.stdout |
- if output.stderr: # pragma: no cover |
- # Print stderr for debugging. |
- print ">>> Stderr (#%d):" % (i + 1) |
- print output.stderr |
- yield output.stdout |
- |
- # Let runnable iterate over all runs and handle output. |
- results += runnable.Run(Runner) |
- |
- if options.json_test_results: |
- results.WriteToFile(options.json_test_results) |
- else: # pragma: no cover |
- print results |
- |
- return min(1, len(results.errors)) |
- |
-if __name__ == "__main__": # pragma: no cover |
- sys.exit(Main(sys.argv[1:])) |