Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(104)

Unified Diff: telemetry/telemetry/benchmark.py

Issue 2903023002: [Telemetry] Do not have benchmarks validate story name in expectations during runs. (Closed)
Patch Set: [Telemetry] Do not have benchmarks validate story name in expectations during runs. Created 3 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: telemetry/telemetry/benchmark.py
diff --git a/telemetry/telemetry/benchmark.py b/telemetry/telemetry/benchmark.py
index 3c36f720fbab4b39ee0f50ac7c19585db1c69a6c..65422bde013f9969342ddf19233524eec59bef47 100644
--- a/telemetry/telemetry/benchmark.py
+++ b/telemetry/telemetry/benchmark.py
@@ -267,10 +267,11 @@ class Benchmark(command_line.Command):
raise NotImplementedError('This test has no "page_set" attribute.')
return self.page_set() # pylint: disable=not-callable
- def ValidateExpectations(self, story_set):
+ def GetBrokenExpectations(self, story_set):
self.InitializeExpectations()
if self._expectations:
- self._expectations.ValidateAgainstStorySet(story_set)
+ return self._expectations.GetBrokenExpectations(story_set)
+ return []
# TODO(rnephew): Rename InitializeExpectations to GetExpectations
def InitializeExpectations(self):
« no previous file with comments | « no previous file | telemetry/telemetry/internal/story_runner.py » ('j') | telemetry/telemetry/internal/story_runner.py » ('J')

Powered by Google App Engine
This is Rietveld 408576698