| Index: chrome/test/data/v8_benchmark_v6/base.js
|
| diff --git a/chrome/test/data/v8_benchmark_v6/base.js b/chrome/test/data/v8_benchmark_v6/base.js
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..ffabf24ddafc91b308c328bcce3b000d3f789870
|
| --- /dev/null
|
| +++ b/chrome/test/data/v8_benchmark_v6/base.js
|
| @@ -0,0 +1,284 @@
|
| +// Copyright 2008 the V8 project authors. All rights reserved.
|
| +// Redistribution and use in source and binary forms, with or without
|
| +// modification, are permitted provided that the following conditions are
|
| +// met:
|
| +//
|
| +// * Redistributions of source code must retain the above copyright
|
| +// notice, this list of conditions and the following disclaimer.
|
| +// * Redistributions in binary form must reproduce the above
|
| +// copyright notice, this list of conditions and the following
|
| +// disclaimer in the documentation and/or other materials provided
|
| +// with the distribution.
|
| +// * Neither the name of Google Inc. nor the names of its
|
| +// contributors may be used to endorse or promote products derived
|
| +// from this software without specific prior written permission.
|
| +//
|
| +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
| +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
| +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
| +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
| +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
| +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
| +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
| +
|
| +
|
| +// Simple framework for running the benchmark suites and
|
| +// computing a score based on the timing measurements.
|
| +
|
| +
|
| +// A benchmark has a name (string) and a function that will be run to
|
| +// do the performance measurement. The optional setup and tearDown
|
| +// arguments are functions that will be invoked before and after
|
| +// running the benchmark, but the running time of these functions will
|
| +// not be accounted for in the benchmark score.
|
| +function Benchmark(name, run, setup, tearDown) {
|
| + this.name = name;
|
| + this.run = run;
|
| + this.Setup = setup ? setup : function() { };
|
| + this.TearDown = tearDown ? tearDown : function() { };
|
| +}
|
| +
|
| +
|
| +// Benchmark results hold the benchmark and the measured time used to
|
| +// run the benchmark. The benchmark score is computed later once a
|
| +// full benchmark suite has run to completion.
|
| +function BenchmarkResult(benchmark, time) {
|
| + this.benchmark = benchmark;
|
| + this.time = time;
|
| +}
|
| +
|
| +
|
| +// Automatically convert results to numbers. Used by the geometric
|
| +// mean computation.
|
| +BenchmarkResult.prototype.valueOf = function() {
|
| + return this.time;
|
| +}
|
| +
|
| +
|
| +// Suites of benchmarks consist of a name and the set of benchmarks in
|
| +// addition to the reference timing that the final score will be based
|
| +// on. This way, all scores are relative to a reference run and higher
|
| +// scores implies better performance.
|
| +function BenchmarkSuite(name, reference, benchmarks) {
|
| + this.name = name;
|
| + this.reference = reference;
|
| + this.benchmarks = benchmarks;
|
| + BenchmarkSuite.suites.push(this);
|
| +}
|
| +
|
| +
|
| +// Keep track of all declared benchmark suites.
|
| +BenchmarkSuite.suites = [];
|
| +
|
| +
|
| +// Scores are not comparable across versions. Bump the version if
|
| +// you're making changes that will affect that scores, e.g. if you add
|
| +// a new benchmark or change an existing one.
|
| +BenchmarkSuite.version = '6';
|
| +
|
| +
|
| +// To make the benchmark results predictable, we replace Math.random
|
| +// with a 100% deterministic alternative.
|
| +Math.random = (function() {
|
| + var seed = 49734321;
|
| + return function() {
|
| + // Robert Jenkins' 32 bit integer hash function.
|
| + seed = ((seed + 0x7ed55d16) + (seed << 12)) & 0xffffffff;
|
| + seed = ((seed ^ 0xc761c23c) ^ (seed >>> 19)) & 0xffffffff;
|
| + seed = ((seed + 0x165667b1) + (seed << 5)) & 0xffffffff;
|
| + seed = ((seed + 0xd3a2646c) ^ (seed << 9)) & 0xffffffff;
|
| + seed = ((seed + 0xfd7046c5) + (seed << 3)) & 0xffffffff;
|
| + seed = ((seed ^ 0xb55a4f09) ^ (seed >>> 16)) & 0xffffffff;
|
| + return (seed & 0xfffffff) / 0x10000000;
|
| + };
|
| +})();
|
| +
|
| +
|
| +// Runs all registered benchmark suites and optionally yields between
|
| +// each individual benchmark to avoid running for too long in the
|
| +// context of browsers. Once done, the final score is reported to the
|
| +// runner.
|
| +BenchmarkSuite.RunSuites = function(runner) {
|
| + var continuation = null;
|
| + var suites = BenchmarkSuite.suites;
|
| + var length = suites.length;
|
| + BenchmarkSuite.scores = [];
|
| + var index = 0;
|
| + function RunStep() {
|
| + while (continuation || index < length) {
|
| + if (continuation) {
|
| + continuation = continuation();
|
| + } else {
|
| + var suite = suites[index++];
|
| + if (runner.NotifyStart) runner.NotifyStart(suite.name);
|
| + continuation = suite.RunStep(runner);
|
| + }
|
| + if (continuation && typeof window != 'undefined' && window.setTimeout) {
|
| + window.setTimeout(RunStep, 25);
|
| + return;
|
| + }
|
| + }
|
| + if (runner.NotifyScore) {
|
| + var score = BenchmarkSuite.GeometricMean(BenchmarkSuite.scores);
|
| + var formatted = BenchmarkSuite.FormatScore(100 * score);
|
| + runner.NotifyScore(formatted);
|
| + }
|
| + }
|
| + RunStep();
|
| +}
|
| +
|
| +
|
| +// Counts the total number of registered benchmarks. Useful for
|
| +// showing progress as a percentage.
|
| +BenchmarkSuite.CountBenchmarks = function() {
|
| + var result = 0;
|
| + var suites = BenchmarkSuite.suites;
|
| + for (var i = 0; i < suites.length; i++) {
|
| + result += suites[i].benchmarks.length;
|
| + }
|
| + return result;
|
| +}
|
| +
|
| +
|
| +// Computes the geometric mean of a set of numbers.
|
| +BenchmarkSuite.GeometricMean = function(numbers) {
|
| + var log = 0;
|
| + for (var i = 0; i < numbers.length; i++) {
|
| + log += Math.log(numbers[i]);
|
| + }
|
| + return Math.pow(Math.E, log / numbers.length);
|
| +}
|
| +
|
| +
|
| +// Converts a score value to a string with at least three significant
|
| +// digits.
|
| +BenchmarkSuite.FormatScore = function(value) {
|
| + if (value > 100) {
|
| + return value.toFixed(0);
|
| + } else {
|
| + return value.toPrecision(3);
|
| + }
|
| +}
|
| +
|
| +// Notifies the runner that we're done running a single benchmark in
|
| +// the benchmark suite. This can be useful to report progress.
|
| +BenchmarkSuite.prototype.NotifyStep = function(result) {
|
| + this.results.push(result);
|
| + if (this.runner.NotifyStep) this.runner.NotifyStep(result.benchmark.name);
|
| +}
|
| +
|
| +
|
| +// Notifies the runner that we're done with running a suite and that
|
| +// we have a result which can be reported to the user if needed.
|
| +BenchmarkSuite.prototype.NotifyResult = function() {
|
| + var mean = BenchmarkSuite.GeometricMean(this.results);
|
| + var score = this.reference / mean;
|
| + BenchmarkSuite.scores.push(score);
|
| + if (this.runner.NotifyResult) {
|
| + var formatted = BenchmarkSuite.FormatScore(100 * score);
|
| + this.runner.NotifyResult(this.name, formatted);
|
| + }
|
| +}
|
| +
|
| +
|
| +// Notifies the runner that running a benchmark resulted in an error.
|
| +BenchmarkSuite.prototype.NotifyError = function(error) {
|
| + if (this.runner.NotifyError) {
|
| + this.runner.NotifyError(this.name, error);
|
| + }
|
| + if (this.runner.NotifyStep) {
|
| + this.runner.NotifyStep(this.name);
|
| + }
|
| +}
|
| +
|
| +
|
| +// Runs a single benchmark for at least a second and computes the
|
| +// average time it takes to run a single iteration.
|
| +BenchmarkSuite.prototype.RunSingleBenchmark = function(benchmark, data) {
|
| + function Measure(data) {
|
| + var elapsed = 0;
|
| + var start = new Date();
|
| + for (var n = 0; elapsed < 1000; n++) {
|
| + benchmark.run();
|
| + elapsed = new Date() - start;
|
| + }
|
| + if (data != null) {
|
| + data.runs += n;
|
| + data.elapsed += elapsed;
|
| + }
|
| + }
|
| +
|
| + if (data == null) {
|
| + // Measure the benchmark once for warm up and throw the result
|
| + // away. Return a fresh data object.
|
| + Measure(null);
|
| + return { runs: 0, elapsed: 0 };
|
| + } else {
|
| + Measure(data);
|
| + // If we've run too few iterations, we continue for another second.
|
| + if (data.runs < 32) return data;
|
| + var usec = (data.elapsed * 1000) / data.runs;
|
| + this.NotifyStep(new BenchmarkResult(benchmark, usec));
|
| + return null;
|
| + }
|
| +}
|
| +
|
| +
|
| +// This function starts running a suite, but stops between each
|
| +// individual benchmark in the suite and returns a continuation
|
| +// function which can be invoked to run the next benchmark. Once the
|
| +// last benchmark has been executed, null is returned.
|
| +BenchmarkSuite.prototype.RunStep = function(runner) {
|
| + this.results = [];
|
| + this.runner = runner;
|
| + var length = this.benchmarks.length;
|
| + var index = 0;
|
| + var suite = this;
|
| + var data;
|
| +
|
| + // Run the setup, the actual benchmark, and the tear down in three
|
| + // separate steps to allow the framework to yield between any of the
|
| + // steps.
|
| +
|
| + function RunNextSetup() {
|
| + if (index < length) {
|
| + try {
|
| + suite.benchmarks[index].Setup();
|
| + } catch (e) {
|
| + suite.NotifyError(e);
|
| + return null;
|
| + }
|
| + return RunNextBenchmark;
|
| + }
|
| + suite.NotifyResult();
|
| + return null;
|
| + }
|
| +
|
| + function RunNextBenchmark() {
|
| + try {
|
| + data = suite.RunSingleBenchmark(suite.benchmarks[index], data);
|
| + } catch (e) {
|
| + suite.NotifyError(e);
|
| + return null;
|
| + }
|
| + // If data is null, we're done with this benchmark.
|
| + return (data == null) ? RunNextTearDown : RunNextBenchmark();
|
| + }
|
| +
|
| + function RunNextTearDown() {
|
| + try {
|
| + suite.benchmarks[index++].TearDown();
|
| + } catch (e) {
|
| + suite.NotifyError(e);
|
| + return null;
|
| + }
|
| + return RunNextSetup;
|
| + }
|
| +
|
| + // Start out running the setup.
|
| + return RunNextSetup();
|
| +}
|
|
|