Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(14)

Unified Diff: tools/perf/core/perf_data_generator.py

Issue 2842623002: Remove whitelist from generate_perf_data (Closed)
Patch Set: Created 3 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « no previous file | tools/perf/core/perf_data_generator_unittest.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: tools/perf/core/perf_data_generator.py
diff --git a/tools/perf/core/perf_data_generator.py b/tools/perf/core/perf_data_generator.py
index 033dc4c9ab61cedbb01ccd1222dfb43d87874dc6..c84a03a3c74e5a85a78b1632e835ee4f1ec2dd12 100755
--- a/tools/perf/core/perf_data_generator.py
+++ b/tools/perf/core/perf_data_generator.py
@@ -122,15 +122,13 @@ def add_builder(waterfall, name, additional_compile_targets=None):
return waterfall
def add_tester(waterfall, name, perf_id, platform, target_bits=64,
- num_host_shards=1, num_device_shards=1, swarming=None,
- use_whitelist=False):
+ num_host_shards=1, num_device_shards=1, swarming=None):
del perf_id # this will be needed
waterfall['testers'][name] = {
'platform': platform,
'num_device_shards': num_device_shards,
'num_host_shards': num_host_shards,
'target_bits': target_bits,
- 'use_whitelist': use_whitelist
}
if swarming:
@@ -600,7 +598,7 @@ def generate_cplusplus_isolate_script_test(dimension):
def generate_telemetry_tests(tester_config, benchmarks, benchmark_sharding_map,
- use_whitelist, benchmark_ref_build_blacklist):
+ benchmark_ref_build_blacklist):
isolated_scripts = []
# First determine the browser that you need based on the tester
browser_name = ''
@@ -622,12 +620,9 @@ def generate_telemetry_tests(tester_config, benchmarks, benchmark_sharding_map,
device_affinity = None
if benchmark_sharding_map:
sharding_map = benchmark_sharding_map.get(str(num_shards), None)
- if not sharding_map and not use_whitelist:
+ if not sharding_map:
raise Exception('Invalid number of shards, generate new sharding map')
- if use_whitelist:
- device_affinity = current_shard
- else:
- device_affinity = sharding_map.get(benchmark.Name(), None)
+ device_affinity = sharding_map.get(benchmark.Name(), None)
else:
# No sharding map was provided, default to legacy device
# affinity algorithm
@@ -656,14 +651,6 @@ def generate_telemetry_tests(tester_config, benchmarks, benchmark_sharding_map,
return isolated_scripts
-BENCHMARK_NAME_WHITELIST = set([
- u'smoothness.top_25_smooth',
- u'sunspider',
- u'system_health.webview_startup',
- u'page_cycler_v2.intl_hi_ru',
- u'dromaeo.cssqueryjquery',
-])
-
# List of benchmarks that are to never be run on a waterfall.
BENCHMARK_NAME_BLACKLIST = [
'multipage_skpicture_printer',
@@ -676,19 +663,22 @@ BENCHMARK_NAME_BLACKLIST = [
'skpicture_printer_ct',
]
+
# Overrides the default 2 hour timeout for swarming tasks.
BENCHMARK_SWARMING_TIMEOUTS = {
'loading.mobile': 14400, # 4 hours
'system_health.memory_mobile': 10800, # 4 hours
}
+
# List of benchmarks that are to never be run with reference builds.
BENCHMARK_REF_BUILD_BLACKLIST = [
'power.idle_platform',
]
-def current_benchmarks(use_whitelist):
+
+def current_benchmarks():
benchmarks_dir = os.path.join(src_dir(), 'tools', 'perf', 'benchmarks')
top_level_dir = os.path.dirname(benchmarks_dir)
@@ -702,10 +692,6 @@ def current_benchmarks(use_whitelist):
all_benchmarks.remove(benchmark)
break
- if use_whitelist:
- all_benchmarks = (
- bench for bench in all_benchmarks
- if bench.Name() in BENCHMARK_NAME_WHITELIST)
return sorted(all_benchmarks, key=lambda b: b.Name())
@@ -763,8 +749,7 @@ def shard_benchmarks(num_shards, all_benchmarks):
def generate_all_tests(waterfall):
tests = {}
- all_benchmarks = current_benchmarks(False)
- whitelist_benchmarks = current_benchmarks(True)
+ all_benchmarks = current_benchmarks()
# Get benchmark sharding according to common sharding configurations
# Currently we only have bots sharded 5 directions and 1 direction
benchmark_sharding_map = {}
@@ -774,10 +759,7 @@ def generate_all_tests(waterfall):
benchmark_sharding_map['21'] = shard_benchmarks(21, all_benchmarks)
for name, config in waterfall['testers'].iteritems():
- use_whitelist = config['use_whitelist']
benchmark_list = all_benchmarks
- if use_whitelist:
- benchmark_list = whitelist_benchmarks
if config.get('swarming', False):
# Our current configuration only ever has one set of swarming dimensions
# Make sure this still holds true
@@ -786,7 +768,7 @@ def generate_all_tests(waterfall):
# Generate benchmarks
sharding_map = benchmark_sharding_map
isolated_scripts = generate_telemetry_tests(
- config, benchmark_list, sharding_map, use_whitelist,
+ config, benchmark_list, sharding_map,
BENCHMARK_REF_BUILD_BLACKLIST)
# Generate swarmed non-telemetry tests if present
if config['swarming_dimensions'][0].get('perf_tests', False):
@@ -891,7 +873,7 @@ def get_all_waterfall_benchmarks_metadata():
def get_all_benchmarks_metadata(metadata):
- benchmark_list = current_benchmarks(False)
+ benchmark_list = current_benchmarks()
for benchmark in benchmark_list:
emails = decorators.GetEmails(benchmark)
« no previous file with comments | « no previous file | tools/perf/core/perf_data_generator_unittest.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698