| Index: tools/telemetry/telemetry/page/page_runner_unittest.py
|
| diff --git a/tools/telemetry/telemetry/page/page_runner_unittest.py b/tools/telemetry/telemetry/page/page_runner_unittest.py
|
| index 8d739e78bfec01355dc4d38a68afc4ca1e5d8027..51d540a6bce475d29f17f7989ca39230413adc43 100644
|
| --- a/tools/telemetry/telemetry/page/page_runner_unittest.py
|
| +++ b/tools/telemetry/telemetry/page/page_runner_unittest.py
|
| @@ -17,13 +17,13 @@
|
| from telemetry.core import user_agent
|
| from telemetry.core import util
|
| from telemetry.page import page as page_module
|
| +from telemetry.page import page_runner
|
| from telemetry.page import page_set
|
| from telemetry.page import page_test
|
| from telemetry.page import test_expectations
|
| from telemetry.results import results_options
|
| from telemetry.unittest_util import options_for_unittests
|
| from telemetry.unittest_util import system_stub
|
| -from telemetry.user_story import user_story_runner
|
| from telemetry.util import exception_formatter as exception_formatter_module
|
| from telemetry.value import scalar
|
| from telemetry.value import string
|
| @@ -44,11 +44,11 @@
|
| pass
|
|
|
|
|
| -def SetUpUserStoryRunnerArguments(options):
|
| +def SetUpPageRunnerArguments(options):
|
| parser = options.CreateParser()
|
| - user_story_runner.AddCommandLineArgs(parser)
|
| + page_runner.AddCommandLineArgs(parser)
|
| options.MergeDefaultValues(parser.get_default_values())
|
| - user_story_runner.ProcessCommandLineArgs(parser, options)
|
| + page_runner.ProcessCommandLineArgs(parser, options)
|
|
|
| class EmptyMetadataForTest(benchmark.BenchmarkMetadata):
|
| def __init__(self):
|
| @@ -83,25 +83,23 @@
|
| pass
|
|
|
|
|
| -# TODO: remove test cases that use real browsers and replace with a
|
| -# user_story_runner or shared_page_state unittest that tests the same logic.
|
| class PageRunnerTests(unittest.TestCase):
|
| # TODO(nduca): Move the basic "test failed, test succeeded" tests from
|
| # page_test_unittest to here.
|
|
|
| def setUp(self):
|
| - self._user_story_runner_logging_stub = None
|
| + self._page_runner_logging_stub = None
|
|
|
| def SuppressExceptionFormatting(self):
|
| - user_story_runner.exception_formatter = FakeExceptionFormatterModule
|
| - self._user_story_runner_logging_stub = system_stub.Override(
|
| - user_story_runner, ['logging'])
|
| + page_runner.exception_formatter = FakeExceptionFormatterModule
|
| + self._page_runner_logging_stub = system_stub.Override(
|
| + page_runner, ['logging'])
|
|
|
| def RestoreExceptionFormatter(self):
|
| - user_story_runner.exception_formatter = exception_formatter_module
|
| - if self._user_story_runner_logging_stub:
|
| - self._user_story_runner_logging_stub.Restore()
|
| - self._user_story_runner_logging_stub = None
|
| + page_runner.exception_formatter = exception_formatter_module
|
| + if self._page_runner_logging_stub:
|
| + self._page_runner_logging_stub.Restore()
|
| + self._page_runner_logging_stub = None
|
|
|
| def tearDown(self):
|
| self.RestoreExceptionFormatter()
|
| @@ -120,9 +118,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Test(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Test(), ps, expectations, options, results)
|
| self.assertEquals(0, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(1, len(results.failures))
|
|
|
| @@ -152,9 +150,9 @@
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| test = Test()
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| self.assertEquals(2, test.run_count)
|
| self.assertEquals(1, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(1, len(results.failures))
|
| @@ -182,9 +180,9 @@
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| test = Test()
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| self.assertEquals(2, test.run_count)
|
| self.assertEquals(1, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(1, len(results.failures))
|
| @@ -213,9 +211,9 @@
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| test = Test()
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| self.assertEquals(2, test.run_count)
|
| self.assertEquals(1, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(1, len(results.failures))
|
| @@ -231,9 +229,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(DummyTest(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(DummyTest(), ps, expectations, options, results)
|
| self.assertEquals(1, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
|
|
| @@ -263,27 +261,27 @@
|
|
|
| options.page_repeat = 1
|
| options.pageset_repeat = 1
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| self.assertEquals(0, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| self.assertEquals(0, len(results.all_page_specific_values))
|
|
|
| options.page_repeat = 1
|
| options.pageset_repeat = 2
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| self.assertEquals(2, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| self.assertEquals(2, len(results.all_page_specific_values))
|
|
|
| options.page_repeat = 2
|
| options.pageset_repeat = 1
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| self.assertEquals(2, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| self.assertEquals(2, len(results.all_page_specific_values))
|
| @@ -292,9 +290,9 @@
|
| options.suppress_gtest_report = True
|
| options.page_repeat = 1
|
| options.pageset_repeat = 1
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| self.assertEquals(0, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| self.assertEquals(0, len(results.all_page_specific_values))
|
| @@ -323,14 +321,14 @@
|
| options.results_label = None
|
| options.page_repeat = 1
|
| options.pageset_repeat = 2
|
| - SetUpUserStoryRunnerArguments(options)
|
| + SetUpPageRunnerArguments(options)
|
|
|
| output = StringIO.StringIO()
|
| real_stdout = sys.stdout
|
| sys.stdout = output
|
| try:
|
| results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| results.PrintSummary()
|
| contents = output.getvalue()
|
| self.assertEquals(4, len(GetSuccessfulPageRuns(results)))
|
| @@ -386,9 +384,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| + SetUpPageRunnerArguments(options)
|
| results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| finally:
|
| os.remove(f.name)
|
|
|
| @@ -417,13 +415,13 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
|
|
| self.assertTrue(hasattr(test, 'hasRun') and test.hasRun)
|
|
|
| - # Ensure that user_story_runner forces exactly 1 tab before running a page.
|
| + # Ensure that page_runner forces exactly 1 tab before running a page.
|
| @decorators.Enabled('has tabs')
|
| def testOneTab(self):
|
| ps = page_set.PageSet()
|
| @@ -448,12 +446,12 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| -
|
| - # Ensure that user_story_runner allows the test to customize the browser
|
| - # before it launches.
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| +
|
| + # Ensure that page_runner allows the test to customize the browser before it
|
| + # launches.
|
| def testBrowserBeforeLaunch(self):
|
| ps = page_set.PageSet()
|
| expectations = test_expectations.TestExpectations()
|
| @@ -482,9 +480,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
|
|
| def testRunPageWithStartupUrl(self):
|
| ps = page_set.PageSet()
|
| @@ -514,13 +512,13 @@
|
| if not browser_finder.FindBrowser(options):
|
| return
|
| test = Measurement()
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| self.assertEquals('about:blank', options.browser_options.startup_url)
|
| self.assertTrue(test.browser_restarted)
|
|
|
| - # Ensure that user_story_runner calls cleanUp when a page run fails.
|
| + # Ensure that page_runner calls cleanUp when a page run fails.
|
| def testCleanUpPage(self):
|
| ps = page_set.PageSet()
|
| expectations = test_expectations.TestExpectations()
|
| @@ -544,9 +542,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| assert test.did_call_clean_up
|
|
|
| # Ensure skipping the test if page cannot be run on the browser
|
| @@ -582,9 +580,9 @@
|
| options = options_for_unittests.GetCopy()
|
| options.output_formats = ['none']
|
| options.suppress_gtest_report = True
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(test, ps, expectations, options, results)
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(test, ps, expectations, options, results)
|
| self.assertFalse(test.will_navigate_to_page_called)
|
| self.assertEquals(0, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| @@ -617,9 +615,9 @@
|
| if not max_failures is None:
|
| options.max_failures = max_failures
|
| expected_max_failures = max_failures
|
| - SetUpUserStoryRunnerArguments(options)
|
| - results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Test(max_failures=2),
|
| + SetUpPageRunnerArguments(options)
|
| + results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| + page_runner.Run(Test(max_failures=2),
|
| ps, expectations, options, results)
|
| self.assertEquals(0, len(GetSuccessfulPageRuns(results)))
|
| # Runs up to max_failures+1 failing tests before stopping, since
|
| @@ -656,9 +654,9 @@
|
| options.output_dir = tempfile.mkdtemp()
|
| options.profiler = 'trace'
|
| try:
|
| - SetUpUserStoryRunnerArguments(options)
|
| + SetUpPageRunnerArguments(options)
|
| results = results_options.CreateResults(EmptyMetadataForTest(), options)
|
| - user_story_runner.Run(Measurement(), ps, expectations, options, results)
|
| + page_runner.Run(Measurement(), ps, expectations, options, results)
|
| self.assertEquals(1, len(GetSuccessfulPageRuns(results)))
|
| self.assertEquals(0, len(results.failures))
|
| self.assertEquals(0, len(results.all_page_specific_values))
|
|
|