| Index: tools/run-bisect-perf-regression.py
 | 
| diff --git a/tools/run-bisect-perf-regression.py b/tools/run-bisect-perf-regression.py
 | 
| index a9c7e54938110f89ad0c83c67535865c9a19aa09..dda359b9b40c698cf142b2c93d72c25e8f623622 100755
 | 
| --- a/tools/run-bisect-perf-regression.py
 | 
| +++ b/tools/run-bisect-perf-regression.py
 | 
| @@ -37,7 +37,6 @@ WEBKIT_RUN_TEST_CONFIG_PATH = os.path.join(
 | 
|  BISECT_SCRIPT_DIR = os.path.join(SCRIPT_DIR, 'auto_bisect')
 | 
|  
 | 
|  PERF_BENCHMARKS_PATH = 'tools/perf/benchmarks'
 | 
| -PERF_MEASUREMENTS_PATH = 'tools/perf/measurements'
 | 
|  BUILDBOT_BUILDERNAME = 'BUILDBOT_BUILDERNAME'
 | 
|  BENCHMARKS_JSON_FILE = 'benchmarks.json'
 | 
|  
 | 
| @@ -635,8 +634,7 @@ def _GetAffectedBenchmarkModuleNames():
 | 
|    all_affected_files = _GetModifiedFilesFromPatch()
 | 
|    modified_benchmarks = []
 | 
|    for affected_file in all_affected_files:
 | 
| -    if (affected_file.startswith(PERF_BENCHMARKS_PATH) or
 | 
| -        affected_file.startswith(PERF_MEASUREMENTS_PATH)):
 | 
| +    if affected_file.startswith(PERF_BENCHMARKS_PATH):
 | 
|        benchmark = os.path.basename(os.path.splitext(affected_file)[0])
 | 
|        modified_benchmarks.append(benchmark)
 | 
|    return modified_benchmarks
 | 
| 
 |