Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(23)

Unified Diff: content/test/test_launcher.cc

Issue 8036044: Add (not yet working) content_browsertests target. (Closed) Base URL: svn://svn.chromium.org/chrome/trunk/src/
Patch Set: '' Created 9 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « content/test/test_launcher.h ('k') | no next file » | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: content/test/test_launcher.cc
===================================================================
--- content/test/test_launcher.cc (revision 0)
+++ content/test/test_launcher.cc (revision 0)
@@ -0,0 +1,604 @@
+// Copyright (c) 2011 The Chromium Authors. All rights reserved.
+// Use of this source code is governed by a BSD-style license that can be
+// found in the LICENSE file.
+
+#include "content/test/test_launcher.h"
+
+#include <string>
+#include <vector>
+
+#include "base/command_line.h"
+#include "base/environment.h"
+#include "base/file_util.h"
+#include "base/hash_tables.h"
+#include "base/logging.h"
+#include "base/mac/scoped_nsautorelease_pool.h"
+#include "base/memory/linked_ptr.h"
+#include "base/memory/scoped_ptr.h"
+#include "base/process_util.h"
+#include "base/scoped_temp_dir.h"
+#include "base/string_number_conversions.h"
+#include "base/string_util.h"
+#include "base/test/test_suite.h"
+#include "base/test/test_timeouts.h"
+#include "base/time.h"
+#include "base/utf_string_conversions.h"
+#include "content/test/browser_test.h"
+#include "net/base/escape.h"
+#include "testing/gtest/include/gtest/gtest.h"
+
+namespace test_launcher {
+
+namespace {
+
+// The environment variable name for the total number of test shards.
+static const char kTestTotalShards[] = "GTEST_TOTAL_SHARDS";
+// The environment variable name for the test shard index.
+static const char kTestShardIndex[] = "GTEST_SHARD_INDEX";
+
+// The default output file for XML output.
+static const FilePath::CharType kDefaultOutputFile[] = FILE_PATH_LITERAL(
+ "test_detail.xml");
+
+// Name of the empty test below.
+static const char kEmptyTestName[] = "InProcessBrowserTest.Empty";
+
+// An empty test (it starts up and shuts down the browser as part of its
+// setup and teardown) used to prefetch all of the browser code into memory.
+IN_PROC_BROWSER_TEST_F(InProcessBrowserTest, Empty) {
+}
+
+// Parses the environment variable var as an Int32. If it is unset, returns
+// default_val. If it is set, unsets it then converts it to Int32 before
+// returning it. If unsetting or converting to an Int32 fails, print an
+// error and exit with failure.
+int32 Int32FromEnvOrDie(const char* const var, int32 default_val) {
+ scoped_ptr<base::Environment> env(base::Environment::Create());
+ std::string str_val;
+ int32 result;
+ if (!env->GetVar(var, &str_val))
+ return default_val;
+ if (!env->UnSetVar(var)) {
+ LOG(ERROR) << "Invalid environment: we could not unset " << var << ".\n";
+ exit(EXIT_FAILURE);
+ }
+ if (!base::StringToInt(str_val, &result)) {
+ LOG(ERROR) << "Invalid environment: " << var << " is not an integer.\n";
+ exit(EXIT_FAILURE);
+ }
+ return result;
+}
+
+// Checks whether sharding is enabled by examining the relevant
+// environment variable values. If the variables are present,
+// but inconsistent (i.e., shard_index >= total_shards), prints
+// an error and exits.
+bool ShouldShard(int32* total_shards, int32* shard_index) {
+ *total_shards = Int32FromEnvOrDie(kTestTotalShards, -1);
+ *shard_index = Int32FromEnvOrDie(kTestShardIndex, -1);
+
+ if (*total_shards == -1 && *shard_index == -1) {
+ return false;
+ } else if (*total_shards == -1 && *shard_index != -1) {
+ LOG(ERROR) << "Invalid environment variables: you have "
+ << kTestShardIndex << " = " << *shard_index
+ << ", but have left " << kTestTotalShards << " unset.\n";
+ exit(EXIT_FAILURE);
+ } else if (*total_shards != -1 && *shard_index == -1) {
+ LOG(ERROR) << "Invalid environment variables: you have "
+ << kTestTotalShards << " = " << *total_shards
+ << ", but have left " << kTestShardIndex << " unset.\n";
+ exit(EXIT_FAILURE);
+ } else if (*shard_index < 0 || *shard_index >= *total_shards) {
+ LOG(ERROR) << "Invalid environment variables: we require 0 <= "
+ << kTestShardIndex << " < " << kTestTotalShards
+ << ", but you have " << kTestShardIndex << "=" << *shard_index
+ << ", " << kTestTotalShards << "=" << *total_shards << ".\n";
+ exit(EXIT_FAILURE);
+ }
+
+ return *total_shards > 1;
+}
+
+// Given the total number of shards, the shard index, and the test id, returns
+// true iff the test should be run on this shard. The test id is some arbitrary
+// but unique non-negative integer assigned by this launcher to each test
+// method. Assumes that 0 <= shard_index < total_shards, which is first
+// verified in ShouldShard().
+bool ShouldRunTestOnShard(int total_shards, int shard_index, int test_id) {
+ return (test_id % total_shards) == shard_index;
+}
+
+// A helper class to output results.
+// Note: as currently XML is the only supported format by gtest, we don't
+// check output format (e.g. "xml:" prefix) here and output an XML file
+// unconditionally.
+// Note: we don't output per-test-case or total summary info like
+// total failed_test_count, disabled_test_count, elapsed_time and so on.
+// Only each test (testcase element in the XML) will have the correct
+// failed/disabled/elapsed_time information. Each test won't include
+// detailed failure messages either.
+class ResultsPrinter {
+ public:
+ explicit ResultsPrinter(const CommandLine& command_line);
+ ~ResultsPrinter();
+ void OnTestCaseStart(const char* name, int test_count) const;
+ void OnTestCaseEnd() const;
+
+ void OnTestEnd(const char* name, const char* case_name, bool run,
+ bool failed, bool failure_ignored, double elapsed_time) const;
+ private:
+ FILE* out_;
+
+ DISALLOW_COPY_AND_ASSIGN(ResultsPrinter);
+};
+
+ResultsPrinter::ResultsPrinter(const CommandLine& command_line) : out_(NULL) {
+ if (!command_line.HasSwitch(kGTestOutputFlag))
+ return;
+ std::string flag = command_line.GetSwitchValueASCII(kGTestOutputFlag);
+ size_t colon_pos = flag.find(':');
+ FilePath path;
+ if (colon_pos != std::string::npos) {
+ FilePath flag_path = command_line.GetSwitchValuePath(kGTestOutputFlag);
+ FilePath::StringType path_string = flag_path.value();
+ path = FilePath(path_string.substr(colon_pos + 1));
+ // If the given path ends with '/', consider it is a directory.
+ // Note: This does NOT check that a directory (or file) actually exists
+ // (the behavior is same as what gtest does).
+ if (file_util::EndsWithSeparator(path)) {
+ FilePath executable = command_line.GetProgram().BaseName();
+ path = path.Append(executable.ReplaceExtension(
+ FilePath::StringType(FILE_PATH_LITERAL("xml"))));
+ }
+ }
+ if (path.value().empty())
+ path = FilePath(kDefaultOutputFile);
+ FilePath dir_name = path.DirName();
+ if (!file_util::DirectoryExists(dir_name)) {
+ LOG(WARNING) << "The output directory does not exist. "
+ << "Creating the directory: " << dir_name.value();
+ // Create the directory if necessary (because the gtest does the same).
+ file_util::CreateDirectory(dir_name);
+ }
+ out_ = file_util::OpenFile(path, "w");
+ if (!out_) {
+ LOG(ERROR) << "Cannot open output file: "
+ << path.value() << ".";
+ return;
+ }
+ fprintf(out_, "<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n");
+ fprintf(out_, "<testsuites name=\"AllTests\" tests=\"\" failures=\"\""
+ " disabled=\"\" errors=\"\" time=\"\">\n");
+}
+
+ResultsPrinter::~ResultsPrinter() {
+ if (!out_)
+ return;
+ fprintf(out_, "</testsuites>\n");
+ fclose(out_);
+}
+
+void ResultsPrinter::OnTestCaseStart(const char* name, int test_count) const {
+ if (!out_)
+ return;
+ fprintf(out_, " <testsuite name=\"%s\" tests=\"%d\" failures=\"\""
+ " disabled=\"\" errors=\"\" time=\"\">\n", name, test_count);
+}
+
+void ResultsPrinter::OnTestCaseEnd() const {
+ if (!out_)
+ return;
+ fprintf(out_, " </testsuite>\n");
+}
+
+void ResultsPrinter::OnTestEnd(const char* name,
+ const char* case_name,
+ bool run,
+ bool failed,
+ bool failure_ignored,
+ double elapsed_time) const {
+ if (!out_)
+ return;
+ fprintf(out_, " <testcase name=\"%s\" status=\"%s\" time=\"%.3f\""
+ " classname=\"%s\"",
+ name, run ? "run" : "notrun", elapsed_time / 1000.0, case_name);
+ if (!failed) {
+ fprintf(out_, " />\n");
+ return;
+ }
+ fprintf(out_, ">\n");
+ fprintf(out_, " <failure message=\"\" type=\"\"%s></failure>\n",
+ failure_ignored ? " ignored=\"true\"" : "");
+ fprintf(out_, " </testcase>\n");
+}
+
+class TestCasePrinterHelper {
+ public:
+ TestCasePrinterHelper(const ResultsPrinter& printer,
+ const char* name,
+ int total_test_count)
+ : printer_(printer) {
+ printer_.OnTestCaseStart(name, total_test_count);
+ }
+ ~TestCasePrinterHelper() {
+ printer_.OnTestCaseEnd();
+ }
+ private:
+ const ResultsPrinter& printer_;
+
+ DISALLOW_COPY_AND_ASSIGN(TestCasePrinterHelper);
+};
+
+// For a basic pattern matching for gtest_filter options. (Copied from
+// gtest.cc, see the comment below and http://crbug.com/44497)
+bool PatternMatchesString(const char* pattern, const char* str) {
+ switch (*pattern) {
+ case '\0':
+ case ':': // Either ':' or '\0' marks the end of the pattern.
+ return *str == '\0';
+ case '?': // Matches any single character.
+ return *str != '\0' && PatternMatchesString(pattern + 1, str + 1);
+ case '*': // Matches any string (possibly empty) of characters.
+ return (*str != '\0' && PatternMatchesString(pattern, str + 1)) ||
+ PatternMatchesString(pattern + 1, str);
+ default: // Non-special character. Matches itself.
+ return *pattern == *str &&
+ PatternMatchesString(pattern + 1, str + 1);
+ }
+}
+
+// TODO(phajdan.jr): Avoid duplicating gtest code. (http://crbug.com/44497)
+// For basic pattern matching for gtest_filter options. (Copied from
+// gtest.cc)
+bool MatchesFilter(const std::string& name, const std::string& filter) {
+ const char *cur_pattern = filter.c_str();
+ for (;;) {
+ if (PatternMatchesString(cur_pattern, name.c_str())) {
+ return true;
+ }
+
+ // Finds the next pattern in the filter.
+ cur_pattern = strchr(cur_pattern, ':');
+
+ // Returns if no more pattern can be found.
+ if (cur_pattern == NULL) {
+ return false;
+ }
+
+ // Skips the pattern separater (the ':' character).
+ cur_pattern++;
+ }
+}
+
+// A multiplier for slow tests. We generally avoid multiplying
+// test timeouts by any constants. Here it is used as last resort
+// to implement the SLOW_ test prefix.
+static const int kSlowTestTimeoutMultiplier = 5;
+
+int GetTestTerminationTimeout(const std::string& test_name,
+ int default_timeout_ms) {
+ int timeout_ms = default_timeout_ms;
+
+ // Make it possible for selected tests to request a longer timeout.
+ // Generally tests should really avoid doing too much, and splitting
+ // a test instead of using SLOW prefix is strongly preferred.
+ if (test_name.find("SLOW_") != std::string::npos)
+ timeout_ms *= kSlowTestTimeoutMultiplier;
+
+ return timeout_ms;
+}
+
+// Runs test specified by |test_name| in a child process,
+// and returns the exit code.
+int RunTest(TestLauncherDelegate* launcher_delegate,
+ const std::string& test_name,
+ int default_timeout_ms) {
+ // Some of the below method calls will leak objects if there is no
+ // autorelease pool in place.
+ base::mac::ScopedNSAutoreleasePool pool;
+
+ const CommandLine* cmd_line = CommandLine::ForCurrentProcess();
+ CommandLine new_cmd_line(cmd_line->GetProgram());
+ CommandLine::SwitchMap switches = cmd_line->GetSwitches();
+
+ // Strip out gtest_output flag because otherwise we would overwrite results
+ // of the previous test. We will generate the final output file later
+ // in RunTests().
+ switches.erase(kGTestOutputFlag);
+
+ // Strip out gtest_repeat flag because we can only run one test in the child
+ // process (restarting the browser in the same process is illegal after it
+ // has been shut down and will actually crash).
+ switches.erase(kGTestRepeatFlag);
+
+ for (CommandLine::SwitchMap::const_iterator iter = switches.begin();
+ iter != switches.end(); ++iter) {
+ new_cmd_line.AppendSwitchNative((*iter).first, (*iter).second);
+ }
+
+ // Always enable disabled tests. This method is not called with disabled
+ // tests unless this flag was specified to the browser test executable.
+ new_cmd_line.AppendSwitch("gtest_also_run_disabled_tests");
+ new_cmd_line.AppendSwitchASCII("gtest_filter", test_name);
+ new_cmd_line.AppendSwitch(kSingleProcessTestsFlag);
+
+ // Do not let the child ignore failures. We need to propagate the
+ // failure status back to the parent.
+ new_cmd_line.AppendSwitch(base::TestSuite::kStrictFailureHandling);
+
+ if (!launcher_delegate->AdjustChildProcessCommandLine(&new_cmd_line))
+ return -1;
+
+ const char* browser_wrapper = getenv("BROWSER_WRAPPER");
+ if (browser_wrapper) {
+#if defined(OS_WIN)
+ new_cmd_line.PrependWrapper(ASCIIToWide(browser_wrapper));
+#elif defined(OS_POSIX)
+ new_cmd_line.PrependWrapper(browser_wrapper);
+#endif
+ VLOG(1) << "BROWSER_WRAPPER was set, prefixing command_line with "
+ << browser_wrapper;
+ }
+
+ base::ProcessHandle process_handle;
+ base::LaunchOptions options;
+
+#if defined(OS_POSIX)
+ // On POSIX, we launch the test in a new process group with pgid equal to
+ // its pid. Any child processes that the test may create will inherit the
+ // same pgid. This way, if the test is abruptly terminated, we can clean up
+ // any orphaned child processes it may have left behind.
+ options.new_process_group = true;
+#endif
+
+ if (!base::LaunchProcess(new_cmd_line, options, &process_handle))
+ return -1;
+
+ int timeout_ms = GetTestTerminationTimeout(test_name,
+ default_timeout_ms);
+
+ int exit_code = 0;
+ if (!base::WaitForExitCodeWithTimeout(process_handle, &exit_code,
+ timeout_ms)) {
+ LOG(ERROR) << "Test timeout (" << timeout_ms
+ << " ms) exceeded for " << test_name;
+
+ exit_code = -1; // Set a non-zero exit code to signal a failure.
+
+ // Ensure that the process terminates.
+ base::KillProcess(process_handle, -1, true);
+ }
+
+#if defined(OS_POSIX)
+ if (exit_code != 0) {
+ // On POSIX, in case the test does not exit cleanly, either due to a crash
+ // or due to it timing out, we need to clean up any child processes that
+ // it might have created. On Windows, child processes are automatically
+ // cleaned up using JobObjects.
+ base::KillProcessGroup(process_handle);
+ }
+#endif
+
+ base::CloseProcessHandle(process_handle);
+
+ return exit_code;
+}
+
+bool RunTests(TestLauncherDelegate* launcher_delegate,
+ bool should_shard,
+ int total_shards,
+ int shard_index) {
+ const CommandLine* command_line = CommandLine::ForCurrentProcess();
+
+ DCHECK(!command_line->HasSwitch(kGTestListTestsFlag));
+
+ testing::UnitTest* const unit_test = testing::UnitTest::GetInstance();
+
+ std::string filter = command_line->GetSwitchValueASCII(kGTestFilterFlag);
+
+ // Split --gtest_filter at '-', if there is one, to separate into
+ // positive filter and negative filter portions.
+ std::string positive_filter = filter;
+ std::string negative_filter = "";
+ size_t dash_pos = filter.find('-');
+ if (dash_pos != std::string::npos) {
+ positive_filter = filter.substr(0, dash_pos); // Everything up to the dash.
+ negative_filter = filter.substr(dash_pos + 1); // Everything after the dash.
+ }
+
+ int num_runnable_tests = 0;
+ int test_run_count = 0;
+ int ignored_failure_count = 0;
+ std::vector<std::string> failed_tests;
+
+ ResultsPrinter printer(*command_line);
+ for (int i = 0; i < unit_test->total_test_case_count(); ++i) {
+ const testing::TestCase* test_case = unit_test->GetTestCase(i);
+ TestCasePrinterHelper helper(printer, test_case->name(),
+ test_case->total_test_count());
+ for (int j = 0; j < test_case->total_test_count(); ++j) {
+ const testing::TestInfo* test_info = test_case->GetTestInfo(j);
+ std::string test_name = test_info->test_case_name();
+ test_name.append(".");
+ test_name.append(test_info->name());
+
+ // Skip our special test so it's not run twice. That confuses
+ // the log parser.
+ if (test_name == kEmptyTestName)
+ continue;
+
+ // Skip disabled tests.
+ if (test_name.find("DISABLED") != std::string::npos &&
+ !command_line->HasSwitch(kGTestRunDisabledTestsFlag)) {
+ printer.OnTestEnd(test_info->name(), test_case->name(),
+ false, false, false, 0);
+ continue;
+ }
+
+ // Skip the test that doesn't match the filter string (if given).
+ if ((!positive_filter.empty() &&
+ !MatchesFilter(test_name, positive_filter)) ||
+ MatchesFilter(test_name, negative_filter)) {
+ printer.OnTestEnd(test_info->name(), test_case->name(),
+ false, false, false, 0);
+ continue;
+ }
+
+ // Decide if this test should be run.
+ bool should_run = true;
+ if (should_shard) {
+ should_run = ShouldRunTestOnShard(total_shards, shard_index,
+ num_runnable_tests);
+ }
+ num_runnable_tests += 1;
+ // If sharding is enabled and the test should not be run, skip it.
+ if (!should_run) {
+ continue;
+ }
+
+ base::Time start_time = base::Time::Now();
+ ++test_run_count;
+ int exit_code = RunTest(launcher_delegate,
+ test_name,
+ TestTimeouts::action_max_timeout_ms());
+ if (exit_code == 0) {
+ // Test passed.
+ printer.OnTestEnd(test_info->name(), test_case->name(), true, false,
+ false,
+ (base::Time::Now() - start_time).InMillisecondsF());
+ } else {
+ failed_tests.push_back(test_name);
+
+ bool ignore_failure = false;
+
+ // -1 exit code means a crash or hang. Never ignore those failures,
+ // they are serious and should always be visible.
+ if (exit_code != -1)
+ ignore_failure = base::TestSuite::ShouldIgnoreFailure(*test_info);
+
+ printer.OnTestEnd(test_info->name(), test_case->name(), true, true,
+ ignore_failure,
+ (base::Time::Now() - start_time).InMillisecondsF());
+ if (ignore_failure)
+ ++ignored_failure_count;
+ }
+ }
+ }
+
+ printf("%d test%s run\n", test_run_count, test_run_count > 1 ? "s" : "");
+ printf("%d test%s failed (%d ignored)\n",
+ static_cast<int>(failed_tests.size()),
+ failed_tests.size() > 1 ? "s" : "", ignored_failure_count);
+ if (failed_tests.size() - ignored_failure_count == 0)
+ return true;
+
+ printf("Failing tests:\n");
+ for (std::vector<std::string>::const_iterator iter = failed_tests.begin();
+ iter != failed_tests.end(); ++iter) {
+ printf("%s\n", iter->c_str());
+ }
+
+ return false;
+}
+
+void PrintUsage() {
+ fprintf(stdout,
+ "Runs tests using the gtest framework, each test being run in its own\n"
+ "process. Any gtest flags can be specified.\n"
+ " --single_process\n"
+ " Runs the tests and the launcher in the same process. Useful for \n"
+ " debugging a specific test in a debugger.\n"
+ " --single-process\n"
+ " Same as above, and also runs Chrome in single-process mode.\n"
+ " --help\n"
+ " Shows this message.\n"
+ " --gtest_help\n"
+ " Shows the gtest help message.\n");
+}
+
+} // namespace
+
+const char kGTestFilterFlag[] = "gtest_filter";
+const char kGTestHelpFlag[] = "gtest_help";
+const char kGTestListTestsFlag[] = "gtest_list_tests";
+const char kGTestRepeatFlag[] = "gtest_repeat";
+const char kGTestRunDisabledTestsFlag[] = "gtest_also_run_disabled_tests";
+const char kGTestOutputFlag[] = "gtest_output";
+
+const char kSingleProcessTestsFlag[] = "single_process";
+const char kSingleProcessTestsAndChromeFlag[] = "single-process";
+// The following is kept for historical reasons (so people that are used to
+// using it don't get surprised).
+const char kChildProcessFlag[] = "child";
+
+const char kHelpFlag[] = "help";
+
+TestLauncherDelegate::~TestLauncherDelegate() {
+}
+
+int LaunchTests(TestLauncherDelegate* launcher_delegate,
+ int argc,
+ char** argv) {
+ launcher_delegate->EarlyInitialize();
+
+ CommandLine::Init(argc, argv);
+ const CommandLine* command_line = CommandLine::ForCurrentProcess();
+
+ if (command_line->HasSwitch(kHelpFlag)) {
+ PrintUsage();
+ return 0;
+ }
+
+ int return_code = 0;
+ if (launcher_delegate->Run(argc, argv, &return_code))
+ return return_code;
+
+ int32 total_shards;
+ int32 shard_index;
+ bool should_shard = ShouldShard(&total_shards, &shard_index);
+
+ fprintf(stdout,
+ "Starting tests...\n"
+ "IMPORTANT DEBUGGING NOTE: each test is run inside its own process.\n"
+ "For debugging a test inside a debugger, use the\n"
+ "--gtest_filter=<your_test_name> flag along with either\n"
+ "--single_process (to run all tests in one launcher/browser process) or\n"
+ "--single-process (to do the above, and also run Chrome in single-\n"
+ "process mode).\n");
+
+ testing::InitGoogleTest(&argc, argv);
+ TestTimeouts::Initialize();
+
+ // Make sure the entire browser code is loaded into memory. Reading it
+ // from disk may be slow on a busy bot, and can easily exceed the default
+ // timeout causing flaky test failures. Use an empty test that only starts
+ // and closes a browser with a long timeout to avoid those problems.
+ RunTest(launcher_delegate,
+ kEmptyTestName,
+ TestTimeouts::large_test_timeout_ms());
+
+ int cycles = 1;
+ if (command_line->HasSwitch(kGTestRepeatFlag)) {
+ base::StringToInt(command_line->GetSwitchValueASCII(kGTestRepeatFlag),
+ &cycles);
+ }
+
+ int exit_code = 0;
+ while (cycles != 0) {
+ if (!RunTests(launcher_delegate,
+ should_shard,
+ total_shards,
+ shard_index)) {
+ exit_code = 1;
+ break;
+ }
+
+ // Special value "-1" means "repeat indefinitely".
+ if (cycles != -1)
+ cycles--;
+ }
+ return exit_code;
+}
+
+} // namespcae test_launcher
Property changes on: content\test\test_launcher.cc
___________________________________________________________________
Added: svn:eol-style
+ LF
« no previous file with comments | « content/test/test_launcher.h ('k') | no next file » | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698