Index: tools/telemetry/telemetry/page/page_run_end_to_end_tests.py |
diff --git a/tools/telemetry/telemetry/page/page_run_end_to_end_tests.py b/tools/telemetry/telemetry/page/page_run_end_to_end_tests.py |
deleted file mode 100644 |
index 0aad2c45d3484c534b964415a5d60d8d5f4a4dc5..0000000000000000000000000000000000000000 |
--- a/tools/telemetry/telemetry/page/page_run_end_to_end_tests.py |
+++ /dev/null |
@@ -1,485 +0,0 @@ |
-# Copyright 2014 The Chromium Authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
- |
-import os |
-import tempfile |
-import unittest |
-import shutil |
-import sys |
-import tempfile |
- |
-from telemetry import benchmark |
-from telemetry import decorators |
-from telemetry.core import browser_finder |
-from telemetry.core import exceptions |
-from telemetry.core import user_agent |
-from telemetry.core import util |
-from telemetry.page import page as page_module |
-from telemetry.page import page_set |
-from telemetry.page import page_test |
-from telemetry.page import test_expectations |
-from telemetry.results import results_options |
-from telemetry.unittest_util import options_for_unittests |
-from telemetry.unittest_util import system_stub |
-from telemetry.user_story import user_story_runner |
-from telemetry.util import exception_formatter as exception_formatter_module |
-from telemetry.value import scalar |
-from telemetry.value import string |
- |
- |
-# pylint: disable=bad-super-call |
- |
-SIMPLE_CREDENTIALS_STRING = """ |
-{ |
- "test": { |
- "username": "example", |
- "password": "asdf" |
- } |
-} |
-""" |
-class DummyTest(page_test.PageTest): |
- def ValidateAndMeasurePage(self, *_): |
- pass |
- |
- |
-def SetUpUserStoryRunnerArguments(options): |
- parser = options.CreateParser() |
- user_story_runner.AddCommandLineArgs(parser) |
- options.MergeDefaultValues(parser.get_default_values()) |
- user_story_runner.ProcessCommandLineArgs(parser, options) |
- |
-class EmptyMetadataForTest(benchmark.BenchmarkMetadata): |
- def __init__(self): |
- super(EmptyMetadataForTest, self).__init__('') |
- |
-class StubCredentialsBackend(object): |
- def __init__(self, login_return_value): |
- self.did_get_login = False |
- self.did_get_login_no_longer_needed = False |
- self.login_return_value = login_return_value |
- |
- @property |
- def credentials_type(self): |
- return 'test' |
- |
- def LoginNeeded(self, *_): |
- self.did_get_login = True |
- return self.login_return_value |
- |
- def LoginNoLongerNeeded(self, _): |
- self.did_get_login_no_longer_needed = True |
- |
- |
-def GetSuccessfulPageRuns(results): |
- return [run for run in results.all_page_runs if run.ok or run.skipped] |
- |
- |
-class FakeExceptionFormatterModule(object): |
- @staticmethod |
- def PrintFormattedException( |
- exception_class=None, exception=None, tb=None, msg=None): |
- pass |
- |
- |
-# TODO: remove test cases that use real browsers and replace with a |
-# user_story_runner or shared_page_state unittest that tests the same logic. |
-class PageRunEndToEndTests(unittest.TestCase): |
- # TODO(nduca): Move the basic "test failed, test succeeded" tests from |
- # page_test_unittest to here. |
- |
- def setUp(self): |
- self._user_story_runner_logging_stub = None |
- |
- def SuppressExceptionFormatting(self): |
- user_story_runner.exception_formatter = FakeExceptionFormatterModule |
- self._user_story_runner_logging_stub = system_stub.Override( |
- user_story_runner, ['logging']) |
- |
- def RestoreExceptionFormatter(self): |
- user_story_runner.exception_formatter = exception_formatter_module |
- if self._user_story_runner_logging_stub: |
- self._user_story_runner_logging_stub.Restore() |
- self._user_story_runner_logging_stub = None |
- |
- def tearDown(self): |
- self.RestoreExceptionFormatter() |
- |
- def testRaiseBrowserGoneExceptionFromRestartBrowserBeforeEachPage(self): |
- self.SuppressExceptionFormatting() |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- ps.pages.append(page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir())) |
- ps.pages.append(page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir())) |
- |
- class Test(page_test.PageTest): |
- def __init__(self, *args): |
- super(Test, self).__init__(*args) |
- self.run_count = 0 |
- |
- def RestartBrowserBeforeEachPage(self): |
- old_run_count = self.run_count |
- self.run_count += 1 |
- if old_run_count == 0: |
- raise exceptions.BrowserGoneException(None) |
- return self._needs_browser_restart_after_each_page |
- |
- def ValidateAndMeasurePage(self, page, tab, results): |
- pass |
- |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- test = Test() |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- self.assertEquals(2, test.run_count) |
- self.assertEquals(1, len(GetSuccessfulPageRuns(results))) |
- self.assertEquals(1, len(results.failures)) |
- |
- def testHandlingOfCrashedTabWithExpectedFailure(self): |
- self.SuppressExceptionFormatting() |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- expectations.Fail('chrome://crash') |
- page1 = page_module.Page('chrome://crash', ps) |
- ps.pages.append(page1) |
- |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(DummyTest(), ps, expectations, options, results) |
- self.assertEquals(1, len(GetSuccessfulPageRuns(results))) |
- self.assertEquals(0, len(results.failures)) |
- |
- def testCredentialsWhenLoginFails(self): |
- self.SuppressExceptionFormatting() |
- credentials_backend = StubCredentialsBackend(login_return_value=False) |
- did_run = self.runCredentialsTest(credentials_backend) |
- assert credentials_backend.did_get_login == True |
- assert credentials_backend.did_get_login_no_longer_needed == False |
- assert did_run == False |
- |
- def testCredentialsWhenLoginSucceeds(self): |
- credentials_backend = StubCredentialsBackend(login_return_value=True) |
- did_run = self.runCredentialsTest(credentials_backend) |
- assert credentials_backend.did_get_login == True |
- assert credentials_backend.did_get_login_no_longer_needed == True |
- assert did_run |
- |
- def runCredentialsTest(self, credentials_backend): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- did_run = [False] |
- |
- try: |
- with tempfile.NamedTemporaryFile(delete=False) as f: |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir(), |
- credentials_path=f.name) |
- page.credentials = "test" |
- ps.pages.append(page) |
- |
- f.write(SIMPLE_CREDENTIALS_STRING) |
- |
- class TestThatInstallsCredentialsBackend(page_test.PageTest): |
- def __init__(self, credentials_backend): |
- super(TestThatInstallsCredentialsBackend, self).__init__() |
- self._credentials_backend = credentials_backend |
- |
- def DidStartBrowser(self, browser): |
- browser.credentials.AddBackend(self._credentials_backend) |
- |
- def ValidateAndMeasurePage(self, *_): |
- did_run[0] = True |
- |
- test = TestThatInstallsCredentialsBackend(credentials_backend) |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- finally: |
- os.remove(f.name) |
- |
- return did_run[0] |
- |
- def testUserAgent(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir()) |
- ps.pages.append(page) |
- ps.user_agent_type = 'tablet' |
- |
- class TestUserAgent(page_test.PageTest): |
- def ValidateAndMeasurePage(self, _1, tab, _2): |
- actual_user_agent = tab.EvaluateJavaScript('window.navigator.userAgent') |
- expected_user_agent = user_agent.UA_TYPE_MAPPING['tablet'] |
- assert actual_user_agent.strip() == expected_user_agent |
- |
- # This is so we can check later that the test actually made it into this |
- # function. Previously it was timing out before even getting here, which |
- # should fail, but since it skipped all the asserts, it slipped by. |
- self.hasRun = True # pylint: disable=W0201 |
- |
- test = TestUserAgent() |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- |
- self.assertTrue(hasattr(test, 'hasRun') and test.hasRun) |
- |
- # Ensure that user_story_runner forces exactly 1 tab before running a page. |
- @decorators.Enabled('has tabs') |
- def testOneTab(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir()) |
- ps.pages.append(page) |
- |
- class TestOneTab(page_test.PageTest): |
- def __init__(self): |
- super(TestOneTab, self).__init__() |
- self._browser = None |
- |
- def DidStartBrowser(self, browser): |
- self._browser = browser |
- self._browser.tabs.New() |
- |
- def ValidateAndMeasurePage(self, *_): |
- assert len(self._browser.tabs) == 1 |
- |
- test = TestOneTab() |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- |
- # Ensure that user_story_runner allows the test to customize the browser |
- # before it launches. |
- def testBrowserBeforeLaunch(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir()) |
- ps.pages.append(page) |
- |
- class TestBeforeLaunch(page_test.PageTest): |
- def __init__(self): |
- super(TestBeforeLaunch, self).__init__() |
- self._did_call_will_start = False |
- self._did_call_did_start = False |
- |
- def WillStartBrowser(self, platform): |
- self._did_call_will_start = True |
- # TODO(simonjam): Test that the profile is available. |
- |
- def DidStartBrowser(self, browser): |
- assert self._did_call_will_start |
- self._did_call_did_start = True |
- |
- def ValidateAndMeasurePage(self, *_): |
- assert self._did_call_did_start |
- |
- test = TestBeforeLaunch() |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- |
- def testRunPageWithStartupUrl(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- expectations = test_expectations.TestExpectations() |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir()) |
- page.startup_url = 'about:blank' |
- ps.pages.append(page) |
- |
- class Measurement(page_test.PageTest): |
- def __init__(self): |
- super(Measurement, self).__init__() |
- self.browser_restarted = False |
- |
- def CustomizeBrowserOptionsForSinglePage(self, ps, options): |
- self.browser_restarted = True |
- super(Measurement, self).CustomizeBrowserOptionsForSinglePage(ps, |
- options) |
- def ValidateAndMeasurePage(self, page, tab, results): |
- pass |
- |
- options = options_for_unittests.GetCopy() |
- options.page_repeat = 2 |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- if not browser_finder.FindBrowser(options): |
- return |
- test = Measurement() |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- self.assertEquals('about:blank', options.browser_options.startup_url) |
- self.assertTrue(test.browser_restarted) |
- |
- # Ensure that user_story_runner calls cleanUp when a page run fails. |
- def testCleanUpPage(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- page = page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir()) |
- ps.pages.append(page) |
- |
- class Test(page_test.PageTest): |
- def __init__(self): |
- super(Test, self).__init__() |
- self.did_call_clean_up = False |
- |
- def ValidateAndMeasurePage(self, *_): |
- raise exceptions.IntentionalException |
- |
- def CleanUpAfterPage(self, page, tab): |
- self.did_call_clean_up = True |
- |
- |
- test = Test() |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- assert test.did_call_clean_up |
- |
- # Ensure skipping the test if page cannot be run on the browser |
- def testPageCannotRunOnBrowser(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- |
- class PageThatCannotRunOnBrowser(page_module.Page): |
- |
- def __init__(self): |
- super(PageThatCannotRunOnBrowser, self).__init__( |
- url='file://blank.html', page_set=ps, |
- base_dir=util.GetUnittestDataDir()) |
- |
- def CanRunOnBrowser(self, _): |
- return False |
- |
- def ValidateAndMeasurePage(self, _): |
- pass |
- |
- class Test(page_test.PageTest): |
- def __init__(self, *args, **kwargs): |
- super(Test, self).__init__(*args, **kwargs) |
- self.will_navigate_to_page_called = False |
- |
- def ValidateAndMeasurePage(self, *args): |
- pass |
- |
- def WillNavigateToPage(self, _1, _2): |
- self.will_navigate_to_page_called = True |
- |
- test = Test() |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(test, ps, expectations, options, results) |
- self.assertFalse(test.will_navigate_to_page_called) |
- self.assertEquals(0, len(GetSuccessfulPageRuns(results))) |
- self.assertEquals(0, len(results.failures)) |
- |
- def _testMaxFailuresOptionIsRespectedAndOverridable(self, max_failures=None): |
- self.SuppressExceptionFormatting() |
- class TestPage(page_module.Page): |
- def __init__(self, *args, **kwargs): |
- super(TestPage, self).__init__(*args, **kwargs) |
- self.was_run = False |
- |
- def RunNavigateSteps(self, action_runner): # pylint: disable=W0613 |
- self.was_run = True |
- raise Exception('Test exception') |
- |
- class Test(page_test.PageTest): |
- def ValidateAndMeasurePage(self, *args): |
- pass |
- |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- for ii in range(5): |
- ps.pages.append(TestPage( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir())) |
- |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- expected_max_failures = 2 |
- if not max_failures is None: |
- options.max_failures = max_failures |
- expected_max_failures = max_failures |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(Test(max_failures=2), |
- ps, expectations, options, results) |
- self.assertEquals(0, len(GetSuccessfulPageRuns(results))) |
- # Runs up to max_failures+1 failing tests before stopping, since |
- # every tests after max_failures failures have been encountered |
- # may all be passing. |
- self.assertEquals(expected_max_failures + 1, len(results.failures)) |
- for ii in range(len(ps.pages)): |
- if ii <= expected_max_failures: |
- self.assertTrue(ps.pages[ii].was_run) |
- else: |
- self.assertFalse(ps.pages[ii].was_run) |
- |
- def testMaxFailuresOptionIsRespected(self): |
- self._testMaxFailuresOptionIsRespectedAndOverridable() |
- |
- def testMaxFailuresOptionIsOverridable(self): |
- self._testMaxFailuresOptionIsRespectedAndOverridable(1) |
- |
- def testRunPageWithProfilingFlag(self): |
- ps = page_set.PageSet() |
- expectations = test_expectations.TestExpectations() |
- ps.pages.append(page_module.Page( |
- 'file://blank.html', ps, base_dir=util.GetUnittestDataDir())) |
- |
- class Measurement(page_test.PageTest): |
- def ValidateAndMeasurePage(self, page, tab, results): |
- pass |
- |
- options = options_for_unittests.GetCopy() |
- options.output_formats = ['none'] |
- options.suppress_gtest_report = True |
- options.reset_results = None |
- options.upload_results = None |
- options.results_label = None |
- options.output_dir = tempfile.mkdtemp() |
- options.profiler = 'trace' |
- try: |
- SetUpUserStoryRunnerArguments(options) |
- results = results_options.CreateResults(EmptyMetadataForTest(), options) |
- user_story_runner.Run(Measurement(), ps, expectations, options, results) |
- self.assertEquals(1, len(GetSuccessfulPageRuns(results))) |
- self.assertEquals(0, len(results.failures)) |
- self.assertEquals(0, len(results.all_page_specific_values)) |
- self.assertTrue(os.path.isfile( |
- os.path.join(options.output_dir, 'blank_html.json'))) |
- finally: |
- shutil.rmtree(options.output_dir) |