| Index: tools/VisualBench/VisualLightweightBenchModule.cpp
|
| diff --git a/tools/VisualBench/VisualLightweightBenchModule.cpp b/tools/VisualBench/VisualLightweightBenchModule.cpp
|
| deleted file mode 100644
|
| index 4aeeaff09a0c58df1b146d3bfefae6439c1402db..0000000000000000000000000000000000000000
|
| --- a/tools/VisualBench/VisualLightweightBenchModule.cpp
|
| +++ /dev/null
|
| @@ -1,152 +0,0 @@
|
| -/*
|
| - * Copyright 2015 Google Inc.
|
| - *
|
| - * Use of this source code is governed by a BSD-style license that can be
|
| - * found in the LICENSE file.
|
| - */
|
| -
|
| -#include "VisualLightweightBenchModule.h"
|
| -
|
| -#include "ProcStats.h"
|
| -#include "SkApplication.h"
|
| -#include "SkCanvas.h"
|
| -#include "SkCommandLineFlags.h"
|
| -#include "SkGraphics.h"
|
| -#include "SkGr.h"
|
| -#include "SkOSFile.h"
|
| -#include "SkStream.h"
|
| -#include "Stats.h"
|
| -#include "gl/GrGLInterface.h"
|
| -
|
| -// Between samples we reset context
|
| -// Between frames we swap buffers
|
| -DEFINE_bool2(verbose, v, false, "enable verbose output from the test driver.");
|
| -DEFINE_string(outResultsFile, "", "If given, write results here as JSON.");
|
| -DEFINE_string(key, "",
|
| - "Space-separated key/value pairs to add to JSON identifying this builder.");
|
| -DEFINE_string(properties, "",
|
| - "Space-separated key/value pairs to add to JSON identifying this run.");
|
| -DEFINE_int32(samples, 10, "Number of times to time each skp.");
|
| -
|
| -static SkString humanize(double ms) {
|
| - if (FLAGS_verbose) {
|
| - return SkStringPrintf("%llu", (uint64_t)(ms*1e6));
|
| - }
|
| - return HumanizeMs(ms);
|
| -}
|
| -
|
| -#define HUMANIZE(time) humanize(time).c_str()
|
| -
|
| -VisualLightweightBenchModule::VisualLightweightBenchModule(VisualBench* owner)
|
| - : INHERITED(owner)
|
| - , fCurrentSample(0)
|
| - , fResults(new ResultsWriter) {
|
| - // Print header
|
| - SkDebugf("curr/maxrss\tloops\tmin\tmedian\tmean\tmax\tstddev\t%-*s\tconfig\tbench\n",
|
| - FLAGS_samples, "samples");
|
| -
|
| - // setup json logging if required
|
| - if (!FLAGS_outResultsFile.isEmpty()) {
|
| - fResults.reset(new NanoJSONResultsWriter(FLAGS_outResultsFile[0]));
|
| - }
|
| -
|
| - if (1 == FLAGS_key.count() % 2) {
|
| - SkDebugf("ERROR: --key must be passed with an even number of arguments.\n");
|
| - } else {
|
| - for (int i = 1; i < FLAGS_key.count(); i += 2) {
|
| - fResults->key(FLAGS_key[i - 1], FLAGS_key[i]);
|
| - }
|
| - }
|
| -
|
| - if (1 == FLAGS_properties.count() % 2) {
|
| - SkDebugf("ERROR: --properties must be passed with an even number of arguments.\n");
|
| - } else {
|
| - for (int i = 1; i < FLAGS_properties.count(); i += 2) {
|
| - fResults->property(FLAGS_properties[i - 1], FLAGS_properties[i]);
|
| - }
|
| - }
|
| -
|
| - // seed an initial record
|
| - fRecords.push_back();
|
| -}
|
| -
|
| -void VisualLightweightBenchModule::renderFrame(SkCanvas* canvas, Benchmark* benchmark, int loops) {
|
| - benchmark->draw(loops, canvas);
|
| - canvas->flush();
|
| -}
|
| -
|
| -void VisualLightweightBenchModule::printStats(Benchmark* benchmark, int loops) {
|
| - const SkTArray<double>& measurements = fRecords.back().fMeasurements;
|
| - const char* shortName = benchmark->getUniqueName();
|
| -
|
| - // update log
|
| - // Note: We currently log only the minimum. It would be interesting to log more information
|
| - SkString configName;
|
| - if (FLAGS_cpu) {
|
| - configName.append("cpu");
|
| - } else if (FLAGS_nvpr) {
|
| - if (FLAGS_offscreen) {
|
| - configName.appendf("nvpr_%d", FLAGS_msaa);
|
| - } else {
|
| - configName.appendf("nvpr_msaa_%d", FLAGS_msaa);
|
| - }
|
| - } else if (FLAGS_msaa > 0) {
|
| - if (FLAGS_offscreen) {
|
| - configName.appendf("offscreen_msaa_%d", FLAGS_msaa);
|
| - } else {
|
| - configName.appendf("msaa_%d", FLAGS_msaa);
|
| - }
|
| - } else {
|
| - if (FLAGS_offscreen) {
|
| - configName.append("offscreen");
|
| - } else {
|
| - configName.append("gpu");
|
| - }
|
| - }
|
| - // Log bench name
|
| - fResults->bench(shortName, benchmark->getSize().fX, benchmark->getSize().fY);
|
| -
|
| - fResults->config(configName.c_str());
|
| - fResults->configOption("name", shortName);
|
| - SkASSERT(measurements.count());
|
| - Stats stats(measurements);
|
| - fResults->metric("min_ms", stats.min);
|
| -
|
| - // Print output
|
| - if (FLAGS_verbose) {
|
| - for (int i = 0; i < measurements.count(); i++) {
|
| - SkDebugf("%s ", HUMANIZE(measurements[i]));
|
| - }
|
| - SkDebugf("%s\n", shortName);
|
| - } else {
|
| - const double stdDevPercent = 100 * sqrt(stats.var) / stats.mean;
|
| - SkDebugf("%4d/%-4dMB\t%d\t%s\t%s\t%s\t%s\t%.0f%%\t%s\t%s\t%s\n",
|
| - sk_tools::getCurrResidentSetSizeMB(),
|
| - sk_tools::getMaxResidentSetSizeMB(),
|
| - loops,
|
| - HUMANIZE(stats.min),
|
| - HUMANIZE(stats.median),
|
| - HUMANIZE(stats.mean),
|
| - HUMANIZE(stats.max),
|
| - stdDevPercent,
|
| - stats.plot.c_str(),
|
| - configName.c_str(),
|
| - shortName);
|
| - }
|
| -}
|
| -
|
| -bool VisualLightweightBenchModule::timingFinished(Benchmark* benchmark, int loops,
|
| - double measurement) {
|
| - fRecords.back().fMeasurements.push_back(measurement);
|
| - if (++fCurrentSample > FLAGS_samples) {
|
| - this->printStats(benchmark, loops);
|
| - fRecords.push_back();
|
| - fCurrentSample = 0;
|
| - return true;
|
| - }
|
| - return false;
|
| -}
|
| -
|
| -bool VisualLightweightBenchModule::onHandleChar(SkUnichar c) {
|
| - return true;
|
| -}
|
|
|