Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(185)

Unified Diff: run_test_cases.py

Issue 12459014: Implement clustering support in run_test_cases.py. (Closed) Base URL: svn://svn.chromium.org/chrome/trunk/tools/swarm_client
Patch Set: Created 7 years, 9 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « no previous file | tests/gtest_fake/expected.xml » ('j') | tests/gtest_fake/gtest_fake_pass.py » ('J')
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: run_test_cases.py
diff --git a/run_test_cases.py b/run_test_cases.py
index e438884ba4d3c78dcc1515b638f6aaf5d1cbd4be..e058f7e470eeabce204a21791248b3c63c69b4b8 100755
--- a/run_test_cases.py
+++ b/run_test_cases.py
@@ -663,7 +663,7 @@ class Runner(object):
# conflict with --gtest_filter.
self.env = setup_gtest_env()
- def map(self, priority, test_case, try_count):
+ def map(self, priority, test_cases, try_count):
"""Traces a single test case and returns its output.
try_count is 0 based, the original try is 0.
@@ -671,55 +671,70 @@ class Runner(object):
if self.decider.should_stop():
return []
+ cmd = self.cmd + ['--gtest_filter=%s' % ':'.join(test_cases)]
+ if '--gtest_print_time' not in cmd:
+ cmd.append('--gtest_print_time')
start = time.time()
output, returncode = call_with_timeout(
- self.cmd + ['--gtest_filter=%s' % test_case],
+ cmd,
self.timeout,
cwd=self.cwd_dir,
stderr=subprocess.STDOUT,
env=self.env)
duration = time.time() - start
- data = {
- 'test_case': test_case,
- 'returncode': returncode,
- 'duration': duration,
- # It needs to be valid utf-8 otherwise it can't be store.
- 'output': output.decode('ascii', 'ignore').encode('utf-8'),
- }
- if '[ RUN ]' not in output:
- # Can't find gtest marker, mark it as invalid.
- returncode = returncode or 1
+
+ # It needs to be valid utf-8 otherwise it can't be store.
csharp 2013/03/11 13:33:22 nit: store -> stored
M-A Ruel 2013/03/11 17:30:53 Done.
+ # TODO(maruel): Be more intelligent than decoding to ascii.
+ utf8_output = output.decode('ascii', 'ignore').encode('utf-8')
+
+ if len(test_cases) > 1:
+ data = process_output(utf8_output, test_cases, duration, returncode)
+ else:
+ data = [
+ {
+ 'test_case': test_cases[0],
+ 'returncode': returncode,
+ 'duration': duration,
+ 'output': utf8_output,
+ }
+ ]
+ if '[ RUN ]' not in output:
+ # Can't find gtest marker, mark it as invalid.
+ returncode = returncode or 1
+
self.decider.got_result(not bool(returncode))
if sys.platform == 'win32':
output = output.replace('\r\n', '\n')
need_to_retry = returncode and try_count < self.retries
- if try_count:
- line = '%s (%.2fs) - retry #%d' % (test_case, duration, try_count)
- else:
- line = '%s (%.2fs)' % (test_case, duration)
- if self.verbose or returncode or try_count > 0:
- # Print output in one of three cases:
- # --verbose was specified.
- # The test failed.
- # The wasn't the first attempt (this is needed so the test parser can
- # detect that a test has been successfully retried).
- line += '\n' + output
- self.progress.update_item(line, True, need_to_retry)
-
- if need_to_retry:
- if try_count + 1 < self.retries:
- # The test failed and needs to be retried normally.
- # Leave a buffer of ~40 test cases before retrying.
- priority += 40
- self.add_task(priority, self.map, priority, test_case, try_count + 1)
+ for i in data:
+ if try_count:
+ line = '%s (%.2fs) - retry #%d' % (
+ i['test_case'], i['duration'] or 0, try_count)
else:
- # This test only has one retry left, so the final retry should be
- # done serially.
- self.add_serial_task(priority, self.map, priority, test_case,
- try_count + 1)
-
- return [data]
+ line = '%s (%.2fs)' % (i['test_case'], i['duration'] or 0)
+ if self.verbose or i['returncode'] or try_count > 0:
+ # Print output in one of three cases:
+ # --verbose was specified.
+ # The test failed.
+ # The wasn't the first attempt (this is needed so the test parser can
+ # detect that a test has been successfully retried).
+ line += '\n' + output
+ self.progress.update_item(line, True, need_to_retry)
+
+ if need_to_retry:
+ if try_count + 1 < self.retries:
+ # The test failed and needs to be retried normally.
+ # Leave a buffer of ~40 test cases before retrying.
+ priority += 40
+ self.add_task(
+ priority, self.map, priority, [i['test_case']], try_count + 1)
+ else:
+ # This test only has one retry left, so the final retry should be
+ # done serially.
+ self.add_serial_task(
+ priority, self.map, priority, [i['test_case']], try_count + 1)
+ return data
def get_test_cases(cmd, cwd, whitelist, blacklist, index, shards, seed):
@@ -942,7 +957,8 @@ def run_test_cases(
- cwd: working directory.
- test_cases: list of preprocessed test cases to run.
- jobs: number of parallel execution threads to do.
- - timeout: individual test case timeout.
+ - timeout: individual test case timeout. Modulated when used with
+ clustering.
- clusters: number of test cases to lump together in a single execution. 0
means the default automatic value which depends on len(test_cases) and
jobs. Capped to len(test_cases) / jobs.
@@ -991,8 +1007,10 @@ def run_test_cases(
function = runner.map
logging.debug('Adding tests to ThreadPool')
progress.use_cr_only = not no_cr
- for i, test_case in enumerate(test_cases):
- pool.add_task(i, function, i, test_case, 0)
+ # Cluster the test cases right away.
+ for i in xrange((len(test_cases) + clusters - 1) / clusters):
+ cluster = test_cases[i*clusters : (i+1)*clusters]
+ pool.add_task(i, function, i, cluster, 0)
logging.debug('All tests added to the ThreadPool')
results = pool.join()
@@ -1012,14 +1030,14 @@ def run_test_cases(
duration = time.time() - pool.tasks.progress.start
cleaned = {}
- for item in results:
- if item:
- cleaned.setdefault(item[0]['test_case'], []).extend(item)
+ for i in results:
csharp 2013/03/11 13:33:22 nit: i -> test_case?
M-A Ruel 2013/03/11 17:30:53 Clarified.
+ for item in i:
+ cleaned.setdefault(item['test_case'], []).append(item)
results = cleaned
# Total time taken to run each test case.
test_case_duration = dict(
- (test_case, sum(i.get('duration', 0) for i in item))
+ (test_case, sum((i.get('duration') or 0) for i in item))
for test_case, item in results.iteritems())
# Classify the results
« no previous file with comments | « no previous file | tests/gtest_fake/expected.xml » ('j') | tests/gtest_fake/gtest_fake_pass.py » ('J')

Powered by Google App Engine
This is Rietveld 408576698