Index: tools/auto_bisect/bisect_perf_regression.py |
diff --git a/tools/auto_bisect/bisect_perf_regression.py b/tools/auto_bisect/bisect_perf_regression.py |
index 01b4fabc46dcaf47ad6d3d380b40a05d19c3b22d..e80fd43999e03584150e54fe4cd943c9b84485fb 100755 |
--- a/tools/auto_bisect/bisect_perf_regression.py |
+++ b/tools/auto_bisect/bisect_perf_regression.py |
@@ -50,7 +50,6 @@ |
os.path.dirname(__file__), os.path.pardir, 'telemetry')) |
from bisect_results import BisectResults |
-from bisect_results import ConfidenceScore |
import bisect_utils |
import builder |
import math_utils |
@@ -170,9 +169,6 @@ |
# The percentage at which confidence is considered high. |
HIGH_CONFIDENCE = 95 |
-# The confidence percentage we require to consider the initial range a |
-# regression based on the test results of the inital good and bad revisions. |
-REGRESSION_CONFIDENCE = 95 |
# Patch template to add a new file, DEPS.sha under src folder. |
# This file contains SHA1 value of the DEPS changes made while bisecting |
@@ -2475,19 +2471,6 @@ |
return results |
print message, "Therefore we continue to bisect." |
- # Check how likely it is that the good and bad results are different |
- # beyond chance-induced variation. |
- if not self.opts.debug_ignore_regression_confidence: |
- regression_confidence = ConfidenceScore(known_bad_value['values'], |
- known_good_value['values']) |
- if regression_confidence < REGRESSION_CONFIDENCE: |
- results.error = ('We could not reproduce the regression with this ' |
- 'test/metric/platform combination with enough ' |
- 'confidence. There\'s still a chance that this is ' |
- 'actually a regression, but you may need to bisect ' |
- 'a different platform.') |
- return results |
- |
# Can just mark the good and bad revisions explicitly here since we |
# already know the results. |
bad_revision_data = revision_data[revision_list[0]] |
@@ -2985,7 +2968,6 @@ |
self.debug_ignore_build = None |
self.debug_ignore_sync = None |
self.debug_ignore_perf_test = None |
- self.debug_ignore_regression_confidence = None |
self.debug_fake_first_test_mean = 0 |
self.gs_bucket = None |
self.target_arch = 'ia32' |
@@ -3153,10 +3135,6 @@ |
group.add_option('--debug_ignore_perf_test', |
action='store_true', |
help='DEBUG: Don\'t perform performance tests.') |
- group.add_option('--debug_ignore_regression_confidence', |
- action='store_true', |
- help='DEBUG: Don\'t score the confidence of the initial ' |
- 'good and bad revisions\' test results.') |
group.add_option('--debug_fake_first_test_mean', |
type='int', |
default='0', |