Index: tools/heapcheck/chrome_tests.py |
diff --git a/tools/heapcheck/chrome_tests.py b/tools/heapcheck/chrome_tests.py |
deleted file mode 100755 |
index aad6b19a89b41ac4f39cef595e2c08e643079a25..0000000000000000000000000000000000000000 |
--- a/tools/heapcheck/chrome_tests.py |
+++ /dev/null |
@@ -1,519 +0,0 @@ |
-#!/usr/bin/env python |
-# Copyright (c) 2012 The Chromium Authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
- |
-''' Runs various chrome tests through heapcheck_test.py. |
- |
-Most of this code is copied from ../valgrind/chrome_tests.py. |
-TODO(glider): put common functions to a standalone module. |
-''' |
- |
-import glob |
-import logging |
-import multiprocessing |
-import optparse |
-import os |
-import stat |
-import sys |
- |
-import logging_utils |
-import path_utils |
- |
-import common |
-import heapcheck_test |
- |
-class TestNotFound(Exception): pass |
- |
-def Dir2IsNewer(dir1, dir2): |
- if dir2 is None or not os.path.isdir(dir2): |
- return False |
- if dir1 is None or not os.path.isdir(dir1): |
- return True |
- return os.stat(dir2)[stat.ST_MTIME] > os.stat(dir1)[stat.ST_MTIME] |
- |
-def FindNewestDir(dirs): |
- newest_dir = None |
- for dir in dirs: |
- if Dir2IsNewer(newest_dir, dir): |
- newest_dir = dir |
- return newest_dir |
- |
-def File2IsNewer(file1, file2): |
- if file2 is None or not os.path.isfile(file2): |
- return False |
- if file1 is None or not os.path.isfile(file1): |
- return True |
- return os.stat(file2)[stat.ST_MTIME] > os.stat(file1)[stat.ST_MTIME] |
- |
-def FindDirContainingNewestFile(dirs, file): |
- """Searches for the directory containing the newest copy of |file|. |
- |
- Args: |
- dirs: A list of paths to the directories to search among. |
- file: A string containing the file name to search. |
- |
- Returns: |
- The string representing the the directory containing the newest copy of |
- |file|. |
- |
- Raises: |
- IOError: |file| was not found. |
- """ |
- newest_dir = None |
- newest_file = None |
- for dir in dirs: |
- the_file = os.path.join(dir, file) |
- if File2IsNewer(newest_file, the_file): |
- newest_dir = dir |
- newest_file = the_file |
- if newest_dir is None: |
- raise IOError("cannot find file %s anywhere, have you built it?" % file) |
- return newest_dir |
- |
-class ChromeTests(object): |
- '''This class is derived from the chrome_tests.py file in ../purify/. |
- ''' |
- |
- def __init__(self, options, args, test): |
- # The known list of tests. |
- # Recognise the original abbreviations as well as full executable names. |
- self._test_list = { |
- "app_list": self.TestAppList, "app_list_unittests": self.TestAppList, |
- "ash": self.TestAsh, "ash_unittests": self.TestAsh, |
- "aura": self.TestAura, "aura_unittests": self.TestAura, |
- "base": self.TestBase, "base_unittests": self.TestBase, |
- "browser": self.TestBrowser, "browser_tests": self.TestBrowser, |
- "chromeos": self.TestChromeOS, "chromeos_unittests": self.TestChromeOS, |
- "components": self.TestComponents, |
- "components_unittests": self.TestComponents, |
- "compositor": self.TestCompositor, |
- "compositor_unittests": self.TestCompositor, |
- "content": self.TestContent, "content_unittests": self.TestContent, |
- "content_browsertests": self.TestContentBrowser, |
- "courgette": self.TestCourgette, |
- "courgette_unittests": self.TestCourgette, |
- "crypto": self.TestCrypto, "crypto_unittests": self.TestCrypto, |
- "device": self.TestDevice, "device_unittests": self.TestDevice, |
- "events": self.TestEvents, "events_unittests": self.TestEvents, |
- "gpu": self.TestGPU, "gpu_unittests": self.TestGPU, |
- "ipc": self.TestIpc, "ipc_tests": self.TestIpc, |
- "jingle": self.TestJingle, "jingle_unittests": self.TestJingle, |
- "layout": self.TestLayout, "layout_tests": self.TestLayout, |
- "media": self.TestMedia, "media_unittests": self.TestMedia, |
- "message_center": self.TestMessageCenter, |
- "message_center_unittests" : self.TestMessageCenter, |
- "net": self.TestNet, "net_unittests": self.TestNet, |
- "ppapi": self.TestPPAPI, "ppapi_unittests": self.TestPPAPI, |
- "printing": self.TestPrinting, "printing_unittests": self.TestPrinting, |
- "remoting": self.TestRemoting, "remoting_unittests": self.TestRemoting, |
- "sql": self.TestSql, "sql_unittests": self.TestSql, |
- "startup": self.TestStartup, "startup_tests": self.TestStartup, |
- "sync": self.TestSync, "sync_unit_tests": self.TestSync, |
- "ui_unit": self.TestUIUnit, "ui_unittests": self.TestUIUnit, |
- "unit": self.TestUnit, "unit_tests": self.TestUnit, |
- "url": self.TestURL, "url_unittests": self.TestURL, |
- "views": self.TestViews, "views_unittests": self.TestViews, |
- } |
- |
- if test not in self._test_list: |
- raise TestNotFound("Unknown test: %s" % test) |
- |
- self._options = options |
- self._args = args |
- self._test = test |
- |
- script_dir = path_utils.ScriptDir() |
- |
- # Compute the top of the tree (the "source dir") from the script dir (where |
- # this script lives). We assume that the script dir is in tools/heapcheck/ |
- # relative to the top of the tree. |
- self._source_dir = os.path.dirname(os.path.dirname(script_dir)) |
- |
- # Since this path is used for string matching, make sure it's always |
- # an absolute Unix-style path. |
- self._source_dir = os.path.abspath(self._source_dir).replace('\\', '/') |
- |
- heapcheck_test_script = os.path.join(script_dir, "heapcheck_test.py") |
- self._command_preamble = [heapcheck_test_script] |
- |
- def _DefaultCommand(self, module, exe=None, heapcheck_test_args=None): |
- '''Generates the default command array that most tests will use. |
- |
- Args: |
- module: The module name (corresponds to the dir in src/ where the test |
- data resides). |
- exe: The executable name. |
- heapcheck_test_args: additional arguments to append to the command line. |
- Returns: |
- A string with the command to run the test. |
- ''' |
- if not self._options.build_dir: |
- dirs = [ |
- os.path.join(self._source_dir, "xcodebuild", "Debug"), |
- os.path.join(self._source_dir, "out", "Debug"), |
- ] |
- if exe: |
- self._options.build_dir = FindDirContainingNewestFile(dirs, exe) |
- else: |
- self._options.build_dir = FindNewestDir(dirs) |
- |
- cmd = list(self._command_preamble) |
- |
- if heapcheck_test_args != None: |
- for arg in heapcheck_test_args: |
- cmd.append(arg) |
- if exe: |
- cmd.append(os.path.join(self._options.build_dir, exe)) |
- # Heapcheck runs tests slowly, so slow tests hurt more; show elapased time |
- # so we can find the slowpokes. |
- cmd.append("--gtest_print_time") |
- if self._options.gtest_repeat: |
- cmd.append("--gtest_repeat=%s" % self._options.gtest_repeat) |
- return cmd |
- |
- def Suppressions(self): |
- '''Builds the list of available suppressions files.''' |
- ret = [] |
- directory = path_utils.ScriptDir() |
- suppression_file = os.path.join(directory, "suppressions.txt") |
- if os.path.exists(suppression_file): |
- ret.append(suppression_file) |
- suppression_file = os.path.join(directory, "suppressions_linux.txt") |
- if os.path.exists(suppression_file): |
- ret.append(suppression_file) |
- return ret |
- |
- def Run(self): |
- '''Runs the test specified by command-line argument --test.''' |
- logging.info("running test %s" % (self._test)) |
- return self._test_list[self._test]() |
- |
- def _ReadGtestFilterFile(self, name, cmd): |
- '''Reads files which contain lists of tests to filter out with |
- --gtest_filter and appends the command-line option to |cmd|. |
- |
- Args: |
- name: the test executable name. |
- cmd: the test running command line to be modified. |
- ''' |
- filters = [] |
- directory = path_utils.ScriptDir() |
- gtest_filter_files = [ |
- os.path.join(directory, name + ".gtest-heapcheck.txt"), |
- # TODO(glider): Linux vs. CrOS? |
- ] |
- logging.info("Reading gtest exclude filter files:") |
- for filename in gtest_filter_files: |
- # strip the leading absolute path (may be very long on the bot) |
- # and the following / or \. |
- readable_filename = filename.replace(self._source_dir, "")[1:] |
- if not os.path.exists(filename): |
- logging.info(" \"%s\" - not found" % readable_filename) |
- continue |
- logging.info(" \"%s\" - OK" % readable_filename) |
- f = open(filename, 'r') |
- for line in f.readlines(): |
- if line.startswith("#") or line.startswith("//") or line.isspace(): |
- continue |
- line = line.rstrip() |
- filters.append(line) |
- gtest_filter = self._options.gtest_filter |
- if len(filters): |
- if gtest_filter: |
- gtest_filter += ":" |
- if gtest_filter.find("-") < 0: |
- gtest_filter += "-" |
- else: |
- gtest_filter = "-" |
- gtest_filter += ":".join(filters) |
- if gtest_filter: |
- cmd.append("--gtest_filter=%s" % gtest_filter) |
- |
- def SimpleTest(self, module, name, heapcheck_test_args=None, cmd_args=None): |
- '''Builds the command line and runs the specified test. |
- |
- Args: |
- module: The module name (corresponds to the dir in src/ where the test |
- data resides). |
- name: The executable name. |
- heapcheck_test_args: Additional command line args for heap checker. |
- cmd_args: Additional command line args for the test. |
- ''' |
- cmd = self._DefaultCommand(module, name, heapcheck_test_args) |
- supp = self.Suppressions() |
- self._ReadGtestFilterFile(name, cmd) |
- if cmd_args: |
- cmd.extend(["--"]) |
- cmd.extend(cmd_args) |
- |
- # Sets LD_LIBRARY_PATH to the build folder so external libraries can be |
- # loaded. |
- if (os.getenv("LD_LIBRARY_PATH")): |
- os.putenv("LD_LIBRARY_PATH", "%s:%s" % (os.getenv("LD_LIBRARY_PATH"), |
- self._options.build_dir)) |
- else: |
- os.putenv("LD_LIBRARY_PATH", self._options.build_dir) |
- return heapcheck_test.RunTool(cmd, supp, module) |
- |
- # TODO(glider): it's an overkill to define a method for each simple test. |
- def TestAppList(self): |
- return self.SimpleTest("app_list", "app_list_unittests") |
- |
- def TestAsh(self): |
- return self.SimpleTest("ash", "ash_unittests") |
- |
- def TestAura(self): |
- return self.SimpleTest("aura", "aura_unittests") |
- |
- def TestBase(self): |
- return self.SimpleTest("base", "base_unittests") |
- |
- def TestBrowser(self): |
- return self.SimpleTest("chrome", "browser_tests") |
- |
- def TestChromeOS(self): |
- return self.SimpleTest("chromeos", "chromeos_unittests") |
- |
- def TestComponents(self): |
- return self.SimpleTest("components", "components_unittests") |
- |
- def TestCompositor(self): |
- return self.SimpleTest("compositor", "compositor_unittests") |
- |
- def TestContent(self): |
- return self.SimpleTest("content", "content_unittests") |
- |
- def TestContentBrowser(self): |
- return self.SimpleTest("content", "content_browsertests") |
- |
- def TestCourgette(self): |
- return self.SimpleTest("courgette", "courgette_unittests") |
- |
- def TestCrypto(self): |
- return self.SimpleTest("crypto", "crypto_unittests") |
- |
- def TestDevice(self): |
- return self.SimpleTest("device", "device_unittests") |
- |
- def TestEvents(self): |
- return self.SimpleTest("events", "events_unittests") |
- |
- def TestGPU(self): |
- return self.SimpleTest("gpu", "gpu_unittests") |
- |
- def TestIpc(self): |
- return self.SimpleTest("ipc", "ipc_tests") |
- |
- def TestJingle(self): |
- return self.SimpleTest("chrome", "jingle_unittests") |
- |
- def TestMedia(self): |
- return self.SimpleTest("chrome", "media_unittests") |
- |
- def TestMessageCenter(self): |
- return self.SimpleTest("message_center", "message_center_unittests") |
- |
- def TestNet(self): |
- return self.SimpleTest("net", "net_unittests") |
- |
- def TestPPAPI(self): |
- return self.SimpleTest("chrome", "ppapi_unittests") |
- |
- def TestPrinting(self): |
- return self.SimpleTest("chrome", "printing_unittests") |
- |
- def TestRemoting(self): |
- return self.SimpleTest("chrome", "remoting_unittests") |
- |
- def TestSync(self): |
- return self.SimpleTest("chrome", "sync_unit_tests") |
- |
- def TestStartup(self): |
- # We don't need the performance results, we're just looking for pointer |
- # errors, so set number of iterations down to the minimum. |
- os.putenv("STARTUP_TESTS_NUMCYCLES", "1") |
- logging.info("export STARTUP_TESTS_NUMCYCLES=1"); |
- return self.SimpleTest("chrome", "startup_tests") |
- |
- def TestUIUnit(self): |
- return self.SimpleTest("chrome", "ui_unittests") |
- |
- def TestUnit(self): |
- return self.SimpleTest("chrome", "unit_tests") |
- |
- def TestURL(self): |
- return self.SimpleTest("chrome", "url_unittests") |
- |
- def TestSql(self): |
- return self.SimpleTest("chrome", "sql_unittests") |
- |
- def TestViews(self): |
- return self.SimpleTest("views", "views_unittests") |
- |
- def TestLayoutChunk(self, chunk_num, chunk_size): |
- '''Runs tests [chunk_num*chunk_size .. (chunk_num+1)*chunk_size). |
- |
- Wrap around to beginning of list at end. If chunk_size is zero, run all |
- tests in the list once. If a text file is given as argument, it is used as |
- the list of tests. |
- ''' |
- # Build the ginormous commandline in 'cmd'. |
- # It's going to be roughly |
- # python heapcheck_test.py ... python run_webkit_tests.py ... |
- # but we'll use the --indirect flag to heapcheck_test.py |
- # to avoid heapchecking python. |
- # Start by building the heapcheck_test.py commandline. |
- cmd = self._DefaultCommand("webkit") |
- |
- # Now build script_cmd, the run_webkits_tests.py commandline |
- # Store each chunk in its own directory so that we can find the data later |
- chunk_dir = os.path.join("layout", "chunk_%05d" % chunk_num) |
- out_dir = os.path.join(path_utils.ScriptDir(), "latest") |
- out_dir = os.path.join(out_dir, chunk_dir) |
- if os.path.exists(out_dir): |
- old_files = glob.glob(os.path.join(out_dir, "*.txt")) |
- for f in old_files: |
- os.remove(f) |
- else: |
- os.makedirs(out_dir) |
- |
- script = os.path.join(self._source_dir, "webkit", "tools", "layout_tests", |
- "run_webkit_tests.py") |
- # While Heapcheck is not memory bound like Valgrind for running layout tests |
- # in parallel, it is still CPU bound. Many machines have hyper-threading |
- # turned on, so the real number of cores is actually half. |
- jobs = max(1, int(multiprocessing.cpu_count() * 0.5)) |
- script_cmd = ["python", script, "-v", |
- "--run-singly", # run a separate DumpRenderTree for each test |
- "--fully-parallel", |
- "--child-processes=%d" % jobs, |
- "--time-out-ms=200000", |
- "--no-retry-failures", # retrying takes too much time |
- # http://crbug.com/176908: Don't launch a browser when done. |
- "--no-show-results", |
- "--nocheck-sys-deps"] |
- |
- # Pass build mode to run_webkit_tests.py. We aren't passed it directly, |
- # so parse it out of build_dir. run_webkit_tests.py can only handle |
- # the two values "Release" and "Debug". |
- # TODO(Hercules): unify how all our scripts pass around build mode |
- # (--mode / --target / --build-dir / --debug) |
- if self._options.build_dir.endswith("Debug"): |
- script_cmd.append("--debug"); |
- if (chunk_size > 0): |
- script_cmd.append("--run-chunk=%d:%d" % (chunk_num, chunk_size)) |
- if len(self._args): |
- # if the arg is a txt file, then treat it as a list of tests |
- if os.path.isfile(self._args[0]) and self._args[0][-4:] == ".txt": |
- script_cmd.append("--test-list=%s" % self._args[0]) |
- else: |
- script_cmd.extend(self._args) |
- self._ReadGtestFilterFile("layout", script_cmd) |
- |
- # Now run script_cmd with the wrapper in cmd |
- cmd.extend(["--"]) |
- cmd.extend(script_cmd) |
- supp = self.Suppressions() |
- return heapcheck_test.RunTool(cmd, supp, "layout") |
- |
- def TestLayout(self): |
- '''Runs the layout tests.''' |
- # A "chunk file" is maintained in the local directory so that each test |
- # runs a slice of the layout tests of size chunk_size that increments with |
- # each run. Since tests can be added and removed from the layout tests at |
- # any time, this is not going to give exact coverage, but it will allow us |
- # to continuously run small slices of the layout tests under purify rather |
- # than having to run all of them in one shot. |
- chunk_size = self._options.num_tests |
- if (chunk_size == 0): |
- return self.TestLayoutChunk(0, 0) |
- chunk_num = 0 |
- chunk_file = os.path.join("heapcheck_layout_chunk.txt") |
- |
- logging.info("Reading state from " + chunk_file) |
- try: |
- f = open(chunk_file) |
- if f: |
- str = f.read() |
- if len(str): |
- chunk_num = int(str) |
- # This should be enough so that we have a couple of complete runs |
- # of test data stored in the archive (although note that when we loop |
- # that we almost guaranteed won't be at the end of the test list) |
- if chunk_num > 10000: |
- chunk_num = 0 |
- f.close() |
- except IOError, (errno, strerror): |
- logging.error("error reading from file %s (%d, %s)" % (chunk_file, |
- errno, strerror)) |
- ret = self.TestLayoutChunk(chunk_num, chunk_size) |
- |
- # Wait until after the test runs to completion to write out the new chunk |
- # number. This way, if the bot is killed, we'll start running again from |
- # the current chunk rather than skipping it. |
- logging.info("Saving state to " + chunk_file) |
- try: |
- f = open(chunk_file, "w") |
- chunk_num += 1 |
- f.write("%d" % chunk_num) |
- f.close() |
- except IOError, (errno, strerror): |
- logging.error("error writing to file %s (%d, %s)" % (chunk_file, errno, |
- strerror)) |
- |
- # Since we're running small chunks of the layout tests, it's important to |
- # mark the ones that have errors in them. These won't be visible in the |
- # summary list for long, but will be useful for someone reviewing this bot. |
- return ret |
- |
- |
-def main(): |
- if not sys.platform.startswith('linux'): |
- logging.error("Heap checking works only on Linux at the moment.") |
- return 1 |
- parser = optparse.OptionParser("usage: %prog -b <dir> -t <test> " |
- "[-t <test> ...]") |
- parser.add_option("-b", "--build-dir", |
- help="the location of the output of the compiler output") |
- parser.add_option("--target", help="Debug or Release") |
- parser.add_option("-t", "--test", action="append", help="which test to run") |
- parser.add_option("--gtest_filter", |
- help="additional arguments to --gtest_filter") |
- parser.add_option("--gtest_repeat", help="argument for --gtest_repeat") |
- parser.add_option("-v", "--verbose", action="store_true", default=False, |
- help="verbose output - enable debug log messages") |
- # My machine can do about 120 layout tests/hour in release mode. |
- # Let's do 30 minutes worth per run. |
- # The CPU is mostly idle, so perhaps we can raise this when |
- # we figure out how to run them more efficiently. |
- parser.add_option("-n", "--num_tests", default=60, type="int", |
- help="for layout tests: # of subtests per run. 0 for all.") |
- |
- options, args = parser.parse_args() |
- |
- # Bake target into build_dir. |
- if options.target and options.build_dir: |
- assert (options.target != |
- os.path.basename(os.path.dirname(options.build_dir))) |
- options.build_dir = os.path.join(os.path.abspath(options.build_dir), |
- options.target) |
- |
- if options.verbose: |
- logging_utils.config_root(logging.DEBUG) |
- else: |
- logging_utils.config_root() |
- |
- if not options.test or not len(options.test): |
- parser.error("--test not specified") |
- |
- for t in options.test: |
- tests = ChromeTests(options, args, t) |
- ret = tests.Run() |
- if ret: |
- return ret |
- return 0 |
- |
- |
-if __name__ == "__main__": |
- sys.exit(main()) |