Index: tools/testrunner/local/testsuite.py |
diff --git a/tools/testrunner/local/testsuite.py b/tools/testrunner/local/testsuite.py |
index b0372e7f7399a8842628e1387855c663f74812fa..473e8b1efed2e4b860df31707b436b784f7b3654 100644 |
--- a/tools/testrunner/local/testsuite.py |
+++ b/tools/testrunner/local/testsuite.py |
@@ -66,10 +66,7 @@ class TestSuite(object): |
# Used in the status file and for stdout printing. |
def CommonTestName(self, testcase): |
- if utils.IsWindows(): |
- return testcase.path.replace("\\", "/") |
- else: |
- return testcase.path |
+ return testcase.path |
def ListTests(self, context): |
raise NotImplementedError |
@@ -87,36 +84,32 @@ class TestSuite(object): |
def ReadTestCases(self, context): |
self.tests = self.ListTests(context) |
- @staticmethod |
- def _FilterFlaky(flaky, mode): |
- return (mode == "run" and not flaky) or (mode == "skip" and flaky) |
- |
- def FilterTestCasesByStatus(self, warn_unused_rules, flaky_tests="dontcare"): |
+ def FilterTestCasesByStatus(self, warn_unused_rules): |
filtered = [] |
used_rules = set() |
for t in self.tests: |
- flaky = False |
testname = self.CommonTestName(t) |
+ if utils.IsWindows(): |
+ testname = testname.replace("\\", "/") |
if testname in self.rules: |
used_rules.add(testname) |
- # Even for skipped tests, as the TestCase object stays around and |
- # PrintReport() uses it. |
- t.outcomes = self.rules[testname] |
- if statusfile.DoSkip(t.outcomes): |
+ outcomes = self.rules[testname] |
+ t.outcomes = outcomes # Even for skipped tests, as the TestCase |
+ # object stays around and PrintReport() uses it. |
+ if statusfile.DoSkip(outcomes): |
continue # Don't add skipped tests to |filtered|. |
- flaky = statusfile.IsFlaky(t.outcomes) |
- skip = False |
- for rule in self.wildcards: |
- assert rule[-1] == '*' |
- if testname.startswith(rule[:-1]): |
- used_rules.add(rule) |
- t.outcomes = self.wildcards[rule] |
- if statusfile.DoSkip(t.outcomes): |
- skip = True |
- break # "for rule in self.wildcards" |
- flaky = flaky or statusfile.IsFlaky(t.outcomes) |
- if skip or self._FilterFlaky(flaky, flaky_tests): |
- continue # "for t in self.tests" |
+ if len(self.wildcards) != 0: |
+ skip = False |
+ for rule in self.wildcards: |
+ assert rule[-1] == '*' |
+ if testname.startswith(rule[:-1]): |
+ used_rules.add(rule) |
+ outcomes = self.wildcards[rule] |
+ t.outcomes = outcomes |
+ if statusfile.DoSkip(outcomes): |
+ skip = True |
+ break # "for rule in self.wildcards" |
+ if skip: continue # "for t in self.tests" |
filtered.append(t) |
self.tests = filtered |