| Index: tools/run-bisect-perf-regression.py
|
| diff --git a/tools/run-bisect-perf-regression.py b/tools/run-bisect-perf-regression.py
|
| index a9c7e54938110f89ad0c83c67535865c9a19aa09..dda359b9b40c698cf142b2c93d72c25e8f623622 100755
|
| --- a/tools/run-bisect-perf-regression.py
|
| +++ b/tools/run-bisect-perf-regression.py
|
| @@ -37,7 +37,6 @@ WEBKIT_RUN_TEST_CONFIG_PATH = os.path.join(
|
| BISECT_SCRIPT_DIR = os.path.join(SCRIPT_DIR, 'auto_bisect')
|
|
|
| PERF_BENCHMARKS_PATH = 'tools/perf/benchmarks'
|
| -PERF_MEASUREMENTS_PATH = 'tools/perf/measurements'
|
| BUILDBOT_BUILDERNAME = 'BUILDBOT_BUILDERNAME'
|
| BENCHMARKS_JSON_FILE = 'benchmarks.json'
|
|
|
| @@ -635,8 +634,7 @@ def _GetAffectedBenchmarkModuleNames():
|
| all_affected_files = _GetModifiedFilesFromPatch()
|
| modified_benchmarks = []
|
| for affected_file in all_affected_files:
|
| - if (affected_file.startswith(PERF_BENCHMARKS_PATH) or
|
| - affected_file.startswith(PERF_MEASUREMENTS_PATH)):
|
| + if affected_file.startswith(PERF_BENCHMARKS_PATH):
|
| benchmark = os.path.basename(os.path.splitext(affected_file)[0])
|
| modified_benchmarks.append(benchmark)
|
| return modified_benchmarks
|
|
|