Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(474)

Unified Diff: test_backup.py

Issue 9146003: Revert change 3084. (Closed) Base URL: http://dart.googlecode.com/svn/branches/bleeding_edge/dart/tools/
Patch Set: Created 8 years, 11 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « no previous file | test_wrapper.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: test_backup.py
===================================================================
--- test_backup.py (revision 3085)
+++ test_backup.py (working copy)
@@ -1,1536 +0,0 @@
-#!/usr/bin/env python
-#
-# Copyright (c) 2011, the Dart project authors. Please see the AUTHORS file
-# for details. All rights reserved. Use of this source code is governed by a
-# BSD-style license that can be found in the LICENSE file.
-#
-
-"""Test driver for the Dart project used by continuous build and developers."""
-
-
-import imp
-import optparse
-import os
-import Queue
-import re
-import sys
-import threading
-import time
-import urllib
-
-import testing
-from testing import test_runner
-import utils
-
-
-TIMEOUT_SECS = 60
-ARCH_GUESS = utils.GuessArchitecture()
-OS_GUESS = utils.GuessOS()
-BUILT_IN_TESTS = ['dartc', 'vm', 'standalone', 'corelib', 'language', 'co19',
- 'samples', 'isolate', 'stub-generator', 'client']
-
-# Patterns for matching test options in .dart files.
-VM_OPTIONS_PATTERN = re.compile(r'// VMOptions=(.*)')
-DART_OPTIONS_PATTERN = re.compile(r'// DartOptions=(.*)')
-ISOLATE_STUB_PATTERN = re.compile(r'// IsolateStubs=(.*)')
-
-# ---------------------------------------------
-# --- P r o g r e s s I n d i c a t o r s ---
-# ---------------------------------------------
-
-
-class Error(Exception):
- pass
-
-
-class ProgressIndicator(object):
- """Base class for displaying the progress of the test run."""
-
- def __init__(self, cases, context, start_time):
- self.abort = False
- self.terminate = False
- self.cases = cases
- self.start_time = start_time
- self.queue = Queue.Queue(len(cases))
- self.batch_queues = {}
- self.context = context
-
- # Extract batchable cases.
- found_cmds = {}
- for case in cases:
- cmd = case.case.GetCommand()[0]
- if not utils.IsWindows():
- # Diagnostic check for executable (if an absolute pathname)
- if not cmd in found_cmds:
- if os.path.isabs(cmd) and not os.path.isfile(cmd):
- msg = "Can't find command %s\n" % cmd
- msg += '(Did you build first? '
- msg += 'Are you running in the correct directory?)'
- raise Exception(msg)
- else:
- found_cmds[cmd] = 1
-
- if case.case.IsBatchable():
- if not cmd in self.batch_queues:
- self.batch_queues[cmd] = Queue.Queue(len(cases))
- self.batch_queues[cmd].put(case)
- else:
- self.queue.put_nowait(case)
-
- self.succeeded = 0
- self.remaining = len(cases)
- self.total = len(cases)
- self.failed = []
- self.crashed = 0
- self.lock = threading.Lock()
-
- def PrintFailureHeader(self, test):
- if test.IsNegative():
- negative_marker = '[negative] '
- else:
- negative_marker = ''
- print '=== %(label)s %(negative)s===' % {
- 'label': test.GetLabel(),
- 'negative': negative_marker
- }
- print 'Path: %s' % '/'.join(test.path)
-
- def Run(self, tasks):
- """Starts tests and keeps running until queues are drained."""
- self.Starting()
-
- # Scale the number of tasks to the nubmer of CPUs on the machine
- if tasks == testing.USE_DEFAULT_CPUS:
- tasks = testing.HOST_CPUS
-
- # TODO(zundel): Refactor BatchSingle method and TestRunner to
- # share code and simplify this method.
-
- # Start the non-batchable items first - there are some long running
- # jobs we don't want to wait on at the end.
- threads = []
- # Spawn N-1 threads and then use this thread as the last one.
- # That way -j1 avoids threading altogether which is a nice fallback
- # in case of threading problems.
- for unused_i in xrange(tasks - 1):
- thread = threading.Thread(target=self.RunSingle, args=[])
- threads.append(thread)
- thread.start()
-
- # Next, crank up the batchable tasks. Note that this will start
- # 'tasks' more threads, but the assumption is that if batching is
- # enabled that almost all tests are batchable.
- for (cmd, queue) in self.batch_queues.items():
- if not queue.empty():
- batch_tester = None
- try:
- batch_tester = test_runner.BatchRunner(queue, tasks, self,
- [cmd, '-batch'])
- except:
- print 'Aborting batch test for ' + cmd + '. Problem on startup.'
- if batch_tester: batch_tester.Shutdown()
- raise
-
- try:
- batch_tester.WaitForCompletion()
- except:
- print 'Aborting batch cmd ' + cmd + 'while waiting for completion.'
- if batch_tester: batch_tester.Shutdown()
- raise
-
- try:
- self.RunSingle()
- if self.abort:
- raise Error('Aborted')
- # Wait for the remaining non-batched threads.
- for thread in threads:
- # Use a timeout so that signals (ctrl-c) will be processed.
- thread.join(timeout=10000000)
- if self.abort:
- raise Error('Aborted')
- except:
- # If there's an exception we schedule an interruption for any
- # remaining threads.
- self.terminate = True
- # ...and then reraise the exception to bail out
- raise
-
- self.Done()
- return not self.failed
-
- def RunSingle(self):
- while not self.terminate:
- try:
- test = self.queue.get_nowait()
- except Queue.Empty:
- return
- case = test.case
- with self.lock:
- self.AboutToRun(case)
- try:
- start = time.time()
- output = case.Run()
- case.duration = (time.time() - start)
- except KeyboardInterrupt:
- self.abort = True
- self.terminate = True
- raise
- except IOError:
- self.abort = True
- self.terminate = True
- raise
- if self.terminate:
- return
- with self.lock:
- if output.UnexpectedOutput():
- self.failed.append(output)
- if output.HasCrashed():
- self.crashed += 1
- else:
- self.succeeded += 1
- self.remaining -= 1
- self.HasRun(output)
-
-
-def EscapeCommand(command):
- parts = []
- for part in command:
- if ' ' in part:
- # Escape spaces. We may need to escape more characters for this
- # to work properly.
- parts.append('"%s"' % part)
- else:
- parts.append(part)
- return ' '.join(parts)
-
-
-class SimpleProgressIndicator(ProgressIndicator):
- """Base class for printing output of each test separately."""
-
- def Starting(self):
- """Called at the beginning before any tests are run."""
- print 'Running %i tests' % len(self.cases)
-
- def Done(self):
- """Called when all tests are complete."""
- print
- for failed in self.failed:
- self.PrintFailureHeader(failed.test)
- if failed.output.stderr:
- print '--- stderr ---'
- print failed.output.stderr.strip()
- if failed.output.stdout:
- print '--- stdout ---'
- print failed.output.stdout.strip()
- print 'Command: %s' % EscapeCommand(failed.command)
- if failed.HasCrashed():
- print '--- CRASHED ---'
- if failed.HasTimedOut():
- print '--- TIMEOUT ---'
- if not self.failed:
- print '==='
- print '=== All tests succeeded'
- print '==='
- else:
- print
- print '==='
- if len(self.failed) == 1:
- print '=== 1 test failed'
- else:
- print '=== %i tests failed' % len(self.failed)
- if self.crashed > 0:
- if self.crashed == 1:
- print '=== 1 test CRASHED'
- else:
- print '=== %i tests CRASHED' % self.crashed
- print '==='
-
-
-class VerboseProgressIndicator(SimpleProgressIndicator):
- """Print verbose information about each test that is run."""
-
- def AboutToRun(self, case):
- """Called before each test case is run."""
- print 'Starting %s...' % case.GetLabel()
- sys.stdout.flush()
-
- def HasRun(self, output):
- """Called after each test case is run."""
- if output.UnexpectedOutput():
- if output.HasCrashed():
- outcome = 'CRASH'
- else:
- outcome = 'FAIL'
- else:
- outcome = 'PASS'
- print 'Done running %s: %s' % (output.test.GetLabel(), outcome)
-
-
-class OneLineProgressIndicator(SimpleProgressIndicator):
- """Results of each test is printed like a report, on a line by itself."""
-
- def AboutToRun(self, case):
- """Called before each test case is run."""
- pass
-
- def HasRun(self, output):
- """Called after each test case is run."""
- if output.UnexpectedOutput():
- if output.HasCrashed():
- outcome = 'CRASH'
- else:
- outcome = 'FAIL'
- else:
- outcome = 'pass'
- print 'Done %s: %s' % (output.test.GetLabel(), outcome)
-
-
-class StatusFileProgressIndicator(SimpleProgressIndicator):
-
- def AboutToRun(self, case):
- """Called before each test case is run."""
- pass
-
- def HasRun(self, output):
- """Called after each test case is run."""
- actual_outcome = output.GetOutcome()
- expected_outcomes = set(output.test.outcomes)
- if not actual_outcome in expected_outcomes:
- expected_outcomes.discard(testing.PASS)
- if expected_outcomes:
- print 'Incorrect status for %s: %s' % (output.test.GetLabel(),
- ', '.join(expected_outcomes))
- else:
- print 'Update status for %s: %s' % (output.test.GetLabel(),
- actual_outcome)
-
-
-class OneLineProgressIndicatorForBuildBot(OneLineProgressIndicator):
-
- def HasRun(self, output):
- """Called after each test case is run."""
- super(OneLineProgressIndicatorForBuildBot, self).HasRun(output)
- percent = (((self.total - self.remaining) * 100) // self.total)
- print '@@@STEP_CLEAR@@@'
- print '@@@STEP_TEXT@ %3d%% +%d -%d @@@' % (
- percent, self.succeeded, len(self.failed))
-
-
-class CompactProgressIndicator(ProgressIndicator):
- """Continuously updates a single line w/ a summary of progress of the run."""
-
- def __init__(self, cases, context, start_time, templates):
- super(CompactProgressIndicator, self).__init__(cases, context, start_time)
- self.templates = templates
- self.last_status_length = 0
-
- def Starting(self):
- """Called at the beginning before any tests are run."""
- pass
-
- def Done(self):
- """Called when all tests are complete."""
- self._PrintProgress('Done')
-
- def AboutToRun(self, case):
- """Called before each test case is run."""
- self._PrintProgress(case.GetLabel())
-
- def HasRun(self, output):
- """Called after each test case is run."""
- if output.UnexpectedOutput():
- self.ClearLine(self.last_status_length)
- self.PrintFailureHeader(output.test)
- stdout = output.output.stdout.strip()
- if stdout:
- print self.templates['stdout'] % stdout
- stderr = output.output.stderr.strip()
- if stderr:
- print self.templates['stderr'] % stderr
- print 'Command: %s' % EscapeCommand(output.command)
- if output.HasCrashed():
- print '--- CRASHED ---'
- if output.HasTimedOut():
- print '--- TIMEOUT ---'
-
- def _Truncate(self, buf, length):
- """Truncate a line if it exceeds length, substituting an ellipsis..."""
- if length and (len(buf) > (length - 3)):
- return buf[:(length-3)] + '...'
- else:
- return buf
-
- def _PrintProgress(self, name):
- """Refresh the display."""
- self.ClearLine(self.last_status_length)
- elapsed = time.time() - self.start_time
- status = self.templates['status_line'] % {
- 'passed': self.succeeded,
- 'percent': (((self.total - self.remaining) * 100) // self.total),
- 'failed': len(self.failed),
- 'test': name,
- 'mins': int(elapsed) / 60,
- 'secs': int(elapsed) % 60
- }
- status = self._Truncate(status, 78)
- self.last_status_length = len(status)
- print status,
- sys.stdout.flush()
-
- def ClearLine(self, last_line_length):
- """Erase the current line w/ a linefeed and overwriting with spaces."""
- print ('\r' + (' ' * last_line_length) + '\r'),
-
-
-class MonochromeProgressIndicator(CompactProgressIndicator):
- """A CompactProgressIndicator with no color."""
-
- def __init__(self, cases, context, start_time):
- templates = {
- 'status_line': '[%(mins)02i:%(secs)02i|%%%(percent) '
- '4d|+%(passed) 4d|-%(failed) 4d]: %(test)s',
- 'stdout': '%s',
- 'stderr': '%s',
- 'clear': lambda last_line_len: self.ClearLine(last_line_len),
- 'max_length': 78
- }
- super(MonochromeProgressIndicator, self).__init__(cases,
- context,
- start_time,
- templates)
-
-
-class ColorProgressIndicator(CompactProgressIndicator):
- """A CompactProgressIndicator with pretty colors."""
-
- def __init__(self, cases, context, start_time):
- templates = {
- 'status_line': ('[%(mins)02i:%(secs)02i|%%%(percent) 4d|'
- '\033[32m+%(passed) 4d'
- '\033[0m|\033[31m-%(failed) 4d\033[0m]: %(test)s'),
- 'stdout': '%s',
- 'stderr': '%s',
- 'clear': lambda last_line_len: self.ClearLine(last_line_len),
- 'max_length': 78
- }
- super(ColorProgressIndicator, self).__init__(cases,
- context,
- start_time,
- templates)
-
-
-PROGRESS_INDICATORS = {
- 'verbose': VerboseProgressIndicator,
- 'mono': MonochromeProgressIndicator,
- 'color': ColorProgressIndicator,
- 'line': OneLineProgressIndicator,
- 'buildbot': OneLineProgressIndicatorForBuildBot,
- 'status': StatusFileProgressIndicator,
-}
-
-
-# -------------------------
-# --- F r a m e w o r k ---
-# -------------------------
-
-
-class TestCase(object):
- """A single test case, like running 'dart' on a single .dart file."""
-
- def __init__(self, context, path):
- self.path = path
- self.context = context
- self.duration = None
- self.arch = []
- self.component = []
-
- def IsBatchable(self):
- if self.context.use_batch:
- if self.component and 'dartc' in self.component:
- return True
- return False
-
- def IsNegative(self):
- return False
-
- def CompareTime(self, other):
- return cmp(other.duration, self.duration)
-
- def DidFail(self, output):
- if output.failed is None:
- output.failed = self.IsFailureOutput(output)
- return output.failed
-
- def IsFailureOutput(self, output):
- return output.exit_code != 0
-
- def RunCommand(self, command, cwd=None, cleanup=True):
- full_command = self.context.processor(command)
- try:
- output = test_runner.Execute(full_command, self.context,
- self.context.timeout, cwd)
- except OSError as e:
- raise utils.ToolError('%s: %s' % (full_command[0], e.strerror))
- test_output = test_runner.TestOutput(self, full_command, output)
- if cleanup: self.Cleanup()
- return test_output
-
- def BeforeRun(self):
- pass
-
- def AfterRun(self):
- pass
-
- def Run(self):
- self.BeforeRun()
- cmd = self.GetCommand()
- try:
- result = self.RunCommand(cmd)
- finally:
- self.AfterRun()
- return result
-
- def Cleanup(self):
- return
-
-
-class TestConfiguration(object):
- """Test configurations give test.py the list of tests, e.g. listing a dir."""
-
- def __init__(self, context, root, flags=[]):
- self.context = context
- self.root = root
- self.flags = flags
-
- def Contains(self, path, filename):
- """Returns True if the given path regexp matches the passed filename."""
-
- if len(path) > len(filename):
- return False
- for i in xrange(len(path)):
- try:
- if not path[i].match(filename[i]):
- return False
- except:
- print 'Invalid regexp %s in .status file. ' % '/'.join(path)
- print 'Try escaping special characters with \\'
- raise
-
- return True
-
- def GetTestStatus(self, sections, defs):
- pass
-
-
-class TestSuite(object):
-
- def __init__(self, name):
- self.name = name
-
- def GetName(self):
- return self.name
-
-
-class TestRepository(TestSuite):
- """A collection of test configurations."""
-
- def __init__(self, path):
- normalized_path = os.path.abspath(path)
- super(TestRepository, self).__init__(os.path.basename(normalized_path))
- self.path = normalized_path
- self.is_loaded = False
- self.config = None
-
- def GetConfiguration(self, context):
- """Retrieve a TestConfiguration subclass for this set of tests."""
- if self.is_loaded:
- return self.config
- self.is_loaded = True
- filename = None
- try:
- (filename, pathname, description) = imp.find_module(
- 'testcfg', [self.path])
- module = imp.load_module('testcfg', filename, pathname, description)
- self.config = module.GetConfiguration(context, self.path)
- finally:
- if filename:
- filename.close()
- return self.config
-
- def ListTests(self, current_path, path, context, mode, arch, component):
- return self.GetConfiguration(context).ListTests(current_path,
- path,
- mode,
- arch,
- component)
-
- def GetTestStatus(self, context, sections, defs):
- self.GetConfiguration(context).GetTestStatus(sections, defs)
-
-
-class LiteralTestSuite(TestSuite):
- """Represents one set of tests."""
-
- def __init__(self, tests):
- super(LiteralTestSuite, self).__init__('root')
- self.tests = tests
-
- def ListTests(self, current_path, path, context, mode, arch, component):
- name = path[0]
- result = []
- for test in self.tests:
- test_name = test.GetName()
- if name.match(test_name):
- full_path = current_path + [test_name]
- result += test.ListTests(full_path, path, context, mode, arch, component)
- return result
-
- def GetTestStatus(self, context, sections, defs):
- for test in self.tests:
- test.GetTestStatus(context, sections, defs)
-
-
-class Context(object):
- """A way to send global context for the test run to each test case."""
-
- def __init__(self, workspace, verbose, os_name, timeout,
- processor, suppress_dialogs, executable, flags,
- keep_temporary_files, use_batch, checked):
- self.workspace = workspace
- self.verbose = verbose
- self.os = os_name
- self.timeout = timeout
- self.processor = processor
- self.suppress_dialogs = suppress_dialogs
- self.executable = executable
- self.flags = flags
- self.keep_temporary_files = keep_temporary_files
- self.use_batch = use_batch == 'true'
- self.checked = checked
-
- def GetBuildRoot(self, mode, arch):
- """The top level directory containing compiler, runtime, tools..."""
- result = utils.GetBuildRoot(self.os, mode, arch)
- return result
-
- def GetBuildConf(self, mode, arch):
- result = utils.GetBuildConf(mode, arch)
- return result
-
- def GetExecutable(self, mode, arch, path):
- """Returns the name of the executable used to run the test."""
- if self.executable is not None:
- return self.executable
- if utils.IsWindows() and not path.endswith('.exe'):
- return path + '.exe'
- else:
- return path
-
- def GetD8(self, mode, arch):
- d8 = os.path.join(self.GetBuildRoot(mode, arch), 'd8')
- return self.GetExecutable(mode, arch, d8)
-
- def GetDart(self, mode, arch, component):
- dart = utils.GetDartRunner(mode, arch, component)
- return [self.GetExecutable(mode, arch, dart)]
-
- def GetDartC(self, mode, arch):
- """Returns the path to the Dart --> JS compiler."""
- dartc = os.path.abspath(os.path.join(
- self.GetBuildRoot(mode, arch), 'compiler', 'bin', 'dartc'))
- if utils.IsWindows(): dartc += '.exe'
- command = [dartc]
-
- # Add the flags from the context to the command line.
- command += self.flags
- return command
-
- def GetRunTests(self, mode, arch):
- path = os.path.join(self.GetBuildRoot(mode, arch), 'run_vm_tests')
- return [self.GetExecutable(mode, arch, path)]
-
-
-def RunTestCases(cases_to_run, progress, tasks, context, start_time):
- """Chooses a progress indicator and then starts the tests."""
- progress = PROGRESS_INDICATORS[progress](cases_to_run, context, start_time)
- return progress.Run(tasks)
-
-
-# -------------------------------------------
-# --- T e s t C o n f i g u r a t i o n ---
-# -------------------------------------------
-
-
-class Expression(object):
- pass
-
-
-class Constant(Expression):
-
- def __init__(self, value):
- super(Constant, self).__init__()
- self.value = value
-
- def Evaluate(self, unused_env, unused_defs):
- return self.value
-
-
-class Variable(Expression):
-
- def __init__(self, name):
- super(Variable, self).__init__()
- self.name = name
-
- def GetOutcomes(self, env, unused_defs):
- if self.name in env:
- return ListSet([env[self.name]])
- else: return Nothing()
-
- def Evaluate(self, env, defs):
- return env[self.name]
-
-
-class Outcome(Expression):
-
- def __init__(self, name):
- super(Outcome, self).__init__()
- self.name = name
-
- def GetOutcomes(self, env, defs):
- if self.name in defs:
- return defs[self.name].GetOutcomes(env, defs)
- else:
- return ListSet([self.name])
-
-
-class Set(object):
- """An abstract set class used to hold Rules."""
- pass
-
-
-class ListSet(Set):
- """A set that uses lists for storage."""
-
- def __init__(self, elms):
- super(ListSet, self).__init__()
- self.elms = elms
-
- def __str__(self):
- return 'ListSet%s' % str(self.elms)
-
- def Intersect(self, that):
- if not isinstance(that, ListSet):
- return that.Intersect(self)
- return ListSet([x for x in self.elms if x in that.elms])
-
- def Union(self, that):
- if not isinstance(that, ListSet):
- return that.Union(self)
- return ListSet(self.elms +
- [x for x in that.elms if x not in self.elms])
-
- def IsEmpty(self):
- return not self.elms
-
-
-class Everything(Set):
- """A set that represents all possible values."""
-
- def Intersect(self, that):
- return that
-
- def Union(self, unused_that):
- return self
-
- def IsEmpty(self):
- return False
-
-
-class Nothing(Set):
-
- def Intersect(self, unused_that):
- return self
-
- def Union(self, that):
- return that
-
- def IsEmpty(self):
- return True
-
-
-class Operation(Expression):
- """A conditional expression. e.g. ($arch == ia32)."""
-
- def __init__(self, left, op, right):
- super(Operation, self).__init__()
- self.left = left
- self.op = op
- self.right = right
-
- def Evaluate(self, env, defs):
- """Evaluates expression in the .status file. e.g. ($arch == ia32)."""
- if self.op == '||' or self.op == ',':
- return self.left.Evaluate(env, defs) or self.right.Evaluate(env, defs)
- elif self.op == 'if':
- return False
- elif self.op == '==':
- outcomes = self.left.GetOutcomes(env, defs)
- inter = outcomes.Intersect(self.right.GetOutcomes(env, defs))
- return not inter.IsEmpty()
- else:
- assert self.op == '&&'
- return self.left.Evaluate(env, defs) and self.right.Evaluate(env, defs)
-
- def GetOutcomes(self, env, defs):
- if self.op == '||' or self.op == ',':
- outcomes = self.left.GetOutcomes(env, defs)
- return outcomes.Union(self.right.GetOutcomes(env, defs))
- elif self.op == 'if':
- if self.right.Evaluate(env, defs):
- return self.left.GetOutcomes(env, defs)
- else: return Nothing()
- else:
- assert self.op == '&&'
- outcomes = self.left.GetOutcomes(env, defs)
- return outcomes.Intersect(self.right.GetOutcomes(env, defs))
-
-
-def IsAlpha(buf):
- """Returns True if the entire string is alphanumeric."""
- for char in buf:
- if not (char.isalpha() or char.isdigit() or char == '_'):
- return False
- return True
-
-
-class Tokenizer(object):
- """Tokenizer that chops expressions into variables, parens and operators."""
-
- def __init__(self, expr):
- self.index = 0
- self.expr = expr
- self.length = len(expr)
- self.tokens = None
-
- def Current(self, length=1):
- if not self.HasMore(length): return ''
- return self.expr[self.index:self.index+length]
-
- def HasMore(self, length=1):
- return self.index < self.length + (length - 1)
-
- def Advance(self, count=1):
- self.index += count
-
- def AddToken(self, token):
- self.tokens.append(token)
-
- def SkipSpaces(self):
- while self.HasMore() and self.Current().isspace():
- self.Advance()
-
- def Tokenize(self):
- """Lexical analysis of an expression in a .status file.
-
- Example:
- [ $mode == debug && ($component == chromium || $component == dartc) ]
-
- Args:
- None.
-
- Returns:
- A list of tokens on success, None on failure.
- """
-
- self.tokens = []
- while self.HasMore():
- self.SkipSpaces()
- if not self.HasMore():
- return None
- if self.Current() == '(':
- self.AddToken('(')
- self.Advance()
- elif self.Current() == ')':
- self.AddToken(')')
- self.Advance()
- elif self.Current() == '$':
- self.AddToken('$')
- self.Advance()
- elif self.Current() == ',':
- self.AddToken(',')
- self.Advance()
- elif IsAlpha(self.Current()):
- buf = ''
- while self.HasMore() and IsAlpha(self.Current()):
- buf += self.Current()
- self.Advance()
- self.AddToken(buf)
- elif self.Current(2) == '&&':
- self.AddToken('&&')
- self.Advance(2)
- elif self.Current(2) == '||':
- self.AddToken('||')
- self.Advance(2)
- elif self.Current(2) == '==':
- self.AddToken('==')
- self.Advance(2)
- else:
- return None
- return self.tokens
-
-
-class Scanner(object):
- """A simple scanner that can serve out tokens from a given list."""
-
- def __init__(self, tokens):
- self.tokens = tokens
- self.length = len(tokens)
- self.index = 0
-
- def HasMore(self):
- return self.index < self.length
-
- def Current(self):
- return self.tokens[self.index]
-
- def Advance(self):
- self.index += 1
-
-
-def ParseAtomicExpression(scan):
- """Parse an single (non recursive) expression in a .status file."""
-
- if scan.Current() == 'true':
- scan.Advance()
- return Constant(True)
- elif scan.Current() == 'false':
- scan.Advance()
- return Constant(False)
- elif IsAlpha(scan.Current()):
- name = scan.Current()
- scan.Advance()
- return Outcome(name.lower())
- elif scan.Current() == '$':
- scan.Advance()
- if not IsAlpha(scan.Current()):
- return None
- name = scan.Current()
- scan.Advance()
- return Variable(name.lower())
- elif scan.Current() == '(':
- scan.Advance()
- result = ParseLogicalExpression(scan)
- if (not result) or (scan.Current() != ')'):
- return None
- scan.Advance()
- return result
- else:
- return None
-
-
-def ParseOperatorExpression(scan):
- """Parse an expression that has operators."""
- left = ParseAtomicExpression(scan)
- if not left: return None
- while scan.HasMore() and (scan.Current() in ['==']):
- op = scan.Current()
- scan.Advance()
- right = ParseOperatorExpression(scan)
- if not right:
- return None
- left = Operation(left, op, right)
- return left
-
-
-def ParseConditionalExpression(scan):
- left = ParseOperatorExpression(scan)
- if not left: return None
- while scan.HasMore() and (scan.Current() == 'if'):
- scan.Advance()
- right = ParseOperatorExpression(scan)
- if not right:
- return None
- left = Operation(left, 'if', right)
- return left
-
-
-def ParseLogicalExpression(scan):
- """Parse a binary expression separated by boolean operators."""
- left = ParseConditionalExpression(scan)
- if not left: return None
- while scan.HasMore() and (scan.Current() in ['&&', '||', ',']):
- op = scan.Current()
- scan.Advance()
- right = ParseConditionalExpression(scan)
- if not right:
- return None
- left = Operation(left, op, right)
- return left
-
-
-def ParseCondition(expr):
- """Parses a boolean expression into an Expression object."""
- tokens = Tokenizer(expr).Tokenize()
- if not tokens:
- print 'Malformed expression: "%s"' % expr
- return None
- scan = Scanner(tokens)
- ast = ParseLogicalExpression(scan)
- if not ast:
- print 'Malformed expression: "%s"' % expr
- return None
- if scan.HasMore():
- print 'Malformed expression: "%s"' % expr
- return None
- return ast
-
-
-class ClassifiedTest(object):
-
- def __init__(self, case, outcomes):
- self.case = case
- self.outcomes = outcomes
-
-
-class Configuration(object):
- """The parsed contents of a configuration file."""
-
- def __init__(self, sections, defs):
- self.sections = sections
- self.defs = defs
-
- def ClassifyTests(self, cases, env):
- """Matches a test case with the test prefixes requested on the cmdline.
-
- This 'wraps' each TestCase object with some meta information
- about the test.
-
- Args:
- cases: list of TestCase objects to classify.
- env: dictionary containing values for 'mode',
- 'system', 'component', 'arch' and 'checked'.
-
- Returns:
- A triplet of (result, rules, expected_outcomes).
- """
- sections = [s for s in self.sections
- if s.condition.Evaluate(env, self.defs)]
- all_rules = reduce(list.__add__, [s.rules for s in sections], [])
- unused_rules = set(all_rules)
- result = []
- all_outcomes = set([])
- for case in cases:
- matches = [r for r in all_rules if r.Contains(case.path)]
- outcomes = set([])
- for rule in matches:
- outcomes = outcomes.union(rule.GetOutcomes(env, self.defs))
- unused_rules.discard(rule)
- if not outcomes:
- outcomes = [testing.PASS]
- case.outcomes = outcomes
- all_outcomes = all_outcomes.union(outcomes)
- result.append(ClassifiedTest(case, outcomes))
- return (result, list(unused_rules), all_outcomes)
-
-
-class Section(object):
- """A section of the configuration file.
-
- Sections are enabled or disabled prior to running the tests,
- based on their conditions.
- """
-
- def __init__(self, condition):
- self.condition = condition
- self.rules = []
-
- def AddRule(self, rule):
- self.rules.append(rule)
-
-
-class Rule(object):
- """A single rule that specifies the expected outcome for a single test."""
-
- def __init__(self, raw_path, path, value):
- self.raw_path = raw_path
- self.path = path
- self.value = value
-
- def GetOutcomes(self, env, defs):
- outcomes = self.value.GetOutcomes(env, defs)
- assert isinstance(outcomes, ListSet)
- return outcomes.elms
-
- def Contains(self, path):
- """Returns True if the specified path matches this rule (regexp)."""
- if len(self.path) > len(path):
- return False
- for i in xrange(len(self.path)):
- try:
- if not self.path[i].match(path[i]):
- return False
- except:
- print 'Invalid regexp %s in .status file. ' % '/'.join(path)
- print 'Try escaping special characters with \\'
- raise
- return True
-
-
-HEADER_PATTERN = re.compile(r'\[([^]]+)\]')
-RULE_PATTERN = re.compile(r'\s*([^: ]*)\s*:(.*)')
-DEF_PATTERN = re.compile(r'^def\s*(\w+)\s*=(.*)$')
-PREFIX_PATTERN = re.compile(r'^\s*prefix\s+([\w\_\.\-\/]+)$')
-
-
-def ReadConfigurationInto(path, sections, defs):
- """Parses a .status file into specified sections and defs arguments."""
- current_section = Section(Constant(True))
- sections.append(current_section)
- prefix = []
- for line in utils.ReadLinesFrom(path):
- header_match = HEADER_PATTERN.match(line)
- if header_match:
- condition_str = header_match.group(1).strip()
- condition = ParseCondition(condition_str)
- new_section = Section(condition)
- sections.append(new_section)
- current_section = new_section
- continue
- rule_match = RULE_PATTERN.match(line)
- if rule_match:
- path = prefix + _SplitPath(rule_match.group(1).strip())
- value_str = rule_match.group(2).strip()
- value = ParseCondition(value_str)
- if not value:
- return False
- current_section.AddRule(Rule(rule_match.group(1), path, value))
- continue
- def_match = DEF_PATTERN.match(line)
- if def_match:
- name = def_match.group(1).lower()
- value = ParseCondition(def_match.group(2).strip())
- if not value:
- return False
- defs[name] = value
- continue
- prefix_match = PREFIX_PATTERN.match(line)
- if prefix_match:
- prefix = _SplitPath(prefix_match.group(1).strip())
- continue
- print 'Malformed line: "%s".' % line
- return False
- return True
-
-
-# ---------------
-# --- M a i n ---
-# ---------------
-
-
-def BuildOptions():
- """Configures the Python optparse library with the cmdline for test.py."""
- result = optparse.OptionParser()
- result.add_option(
- '-m', '--mode',
- help='The test modes in which to run (comma-separated)',
- metavar='[all,debug,release]',
- default='debug')
- result.add_option(
- '-v', '--verbose',
- help='Verbose output',
- default=False,
- action='store_true')
- result.add_option(
- '-p', '--progress',
- help='The style of progress indicator (verbose, line, color, mono)',
- choices=PROGRESS_INDICATORS.keys(),
- default=None)
- result.add_option(
- '--report',
- help='Print a summary of the tests to be run',
- default=False,
- action='store_true')
- result.add_option(
- '--list',
- help='List all the tests, but don\'t run them',
- default=False,
- action='store_true')
- result.add_option(
- '-s', '--suite',
- help='A test suite',
- default=[],
- action='append')
- result.add_option(
- '-t', '--timeout',
- help='Timeout in seconds',
- default=None,
- type='int')
- result.add_option(
- '--checked',
- help='Run tests in checked mode',
- default=False,
- action='store_true')
- result.add_option(
- '--flag',
- help='Pass this additional flag to the VM or the program running the test',
- default=[],
- action='append')
- result.add_option(
- '--arch',
- help='The architecture to run tests for',
- metavar='[all,ia32,x64,simarm,arm]',
- default=ARCH_GUESS)
- result.add_option(
- '--os',
- help='The OS to run tests on',
- default=OS_GUESS)
- result.add_option(
- '--valgrind',
- help='Run tests through valgrind',
- default=False,
- action='store_true')
- result.add_option(
- '-j', '--tasks',
- help='The number of parallel tasks to run',
- metavar=testing.HOST_CPUS,
- default=testing.USE_DEFAULT_CPUS,
- type='int')
- result.add_option(
- '--time',
- help='Print timing information after running',
- default=False,
- action='store_true')
- result.add_option(
- '--executable',
- help='The executable with which to run the tests',
- default=None)
- result.add_option(
- '--keep_temporary_files',
- help='Do not delete temporary files after running the tests',
- default=False,
- action='store_true')
- result.add_option(
- '--batch',
- help='Run multiple tests for dartc component in a single vm',
- choices=['true', 'false'],
- default='true',
- type='choice')
- result.add_option(
- '--optimize',
- help='Invoke dart compiler with --optimize flag',
- default=False,
- action='store_true')
- result.add_option(
- '-c', '--component',
- help='The component to test against '
- '(most, vm, dartc, frog, frogsh, leg, chromium, dartium, webdriver)',
- metavar='[most,vm,dartc,chromium,dartium]',
- default='vm')
- return result
-
-
-def ProcessOptions(options):
- """Process command line options."""
- if options.arch == 'all':
- options.arch = 'ia32,x64,simarm'
- if options.mode == 'all':
- options.mode = 'debug,release'
- if options.component == 'most':
- options.component = 'vm,dartc'
-
- # By default we run with a higher timeout setting in when running on
- # a simulated architecture and in debug mode.
- if not options.timeout:
- options.timeout = TIMEOUT_SECS
- if 'dartc' in options.component:
- options.timeout *= 4
- elif 'chromium' in options.component:
- options.timeout *= 4
- elif 'dartium' in options.component:
- options.timeout *= 4
- elif 'debug' in options.mode:
- options.timeout *= 2
- options.mode = options.mode.split(',')
- options.arch = options.arch.split(',')
- options.component = options.component.split(',')
- for mode in options.mode:
- if not mode in ['debug', 'release']:
- print 'Unknown mode %s' % mode
- return False
- for arch in options.arch:
- if not arch in ['ia32', 'x64', 'simarm', 'arm']:
- print 'Unknown arch %s' % arch
- return False
- for component in options.component:
- if not component in ['vm', 'dartc', 'frog', 'frogsh', 'leg',
- 'chromium', 'dartium', 'frogium', 'webdriver']:
- print 'Unknown component %s' % component
- return False
- options.flags = []
- options.flags.append('--ignore-unrecognized-flags')
- if options.checked:
- options.flags.append('--enable_asserts')
- options.flags.append('--enable_type_checks')
- if options.optimize:
- options.flags.append('--optimize')
- for flag in options.flag:
- options.flags.append(flag)
- if options.verbose:
- print 'Flags on the command line:'
- for x in options.flags:
- print x
- # If the user hasn't specified the progress indicator, we pick
- # a good one depending on the setting of the verbose option.
- if not options.progress:
- if options.verbose: options.progress = 'verbose'
- else: options.progress = 'mono'
- # Options for future use. Such as Windows runner support.
- options.suppress_dialogs = True
- options.special_command = None
- return True
-
-
-REPORT_TEMPLATE = """\
-Total: %(total)i tests
- * %(skipped)4d tests will be skipped
- * %(nocrash)4d tests are expected to be flaky but not crash
- * %(pass)4d tests are expected to pass
- * %(fail_ok)4d tests are expected to fail that we won't fix
- * %(fail)4d tests are expected to fail that we should fix
- * %(crash)4d tests are expected to crash that we should fix
- * %(batched)4d tests are running in batch mode\
-"""
-
-
-def PrintReport(cases):
- """Print a breakdown of which tests are marked pass/skip/fail."""
-
- def IsFlaky(o):
- return ((testing.PASS in o) and (testing.FAIL in o)
- and (not testing.CRASH in o) and (not testing.OKAY in o))
-
- def IsFailOk(o):
- return (len(o) == 2) and (testing.FAIL in o) and (testing.OKAY in o)
-
- unskipped = [c for c in cases if not testing.SKIP in c.outcomes]
- print REPORT_TEMPLATE % {
- 'total': len(cases),
- 'skipped': len(cases) - len(unskipped),
- 'nocrash': len([t for t in unskipped if IsFlaky(t.outcomes)]),
- 'pass': len([t for t in unskipped
- if list(t.outcomes) == [testing.PASS]]),
- 'fail_ok': len([t for t in unskipped
- if IsFailOk(t.outcomes)]),
- 'fail': len([t for t in unskipped
- if list(t.outcomes) == [testing.FAIL]]),
- 'crash': len([t for t in unskipped
- if list(t.outcomes) == [testing.CRASH]]),
- 'batched': len([t for t in unskipped if t.case.IsBatchable()])
- }
-
-
-def PrintTests(cases):
- """Print a table of the tests to be run (--list cmdline option)."""
- has_errors = False
- for case in cases:
- try:
- case.case.GetCommand()
- except:
- # Python can throw an exception while parsing the .dart file.
- # We don't want to end the program.
- # TODO(zundel): something better... its a bit of a hack.
- sys.stderr.write(case.case.filename + '\n')
- has_errors = True
- if has_errors:
- raise Exception('Errors in above files')
- for case in [c for c in cases if not testing.SKIP in c.outcomes]:
- print '%s\t%s\t%s\t%s' %('/'.join(case.case.path),
- ','.join(case.outcomes),
- case.case.IsNegative(),
- '\t'.join(case.case.GetCommand()[1:]))
-
-
-class Pattern(object):
- """Convenience class to hold a compiled re pattern."""
-
- def __init__(self, pattern):
- self.pattern = pattern
- self.compiled = None
-
- def match(self, buf):
- if not self.compiled:
- pattern = '^%s$' % self.pattern.replace('*', '.*')
- self.compiled = re.compile(pattern)
- return self.compiled.match(buf)
-
- def __str__(self):
- return self.pattern
-
-
-def _SplitPath(s):
- """Split a path into directories - opposite of os.path.join()?"""
- stripped = [c.strip() for c in s.split('/')]
- return [Pattern(s) for s in stripped if s]
-
-
-def GetSpecialCommandProcessor(value):
- if (not value) or (value.find('@') == -1):
-
- def ExpandCommand(args):
- return args
-
- return ExpandCommand
- else:
- pos = value.find('@')
- prefix = urllib.unquote(value[:pos]).split()
- suffix = urllib.unquote(value[pos+1:]).split()
-
- def ExpandCommand(args):
- return prefix + args + suffix
-
- return ExpandCommand
-
-
-def GetSuites(test_root):
- def IsSuite(path):
- return os.path.isdir(path) and os.path.exists(
- os.path.join(path, 'testcfg.py'))
- return [f for f in os.listdir(test_root) if IsSuite(
- os.path.join(test_root, f))]
-
-
-def FormatTime(d):
- millis = round(d * 1000) % 1000
- return time.strftime('%M:%S.', time.gmtime(d)) + ('%03i' % millis)
-
-
-def Main():
- """Main loop."""
- script_start_time = time.time();
- utils.ConfigureJava()
- parser = BuildOptions()
- (options, args) = parser.parse_args()
- if not ProcessOptions(options):
- parser.print_help()
- return 1
-
- client = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '..'))
- repositories = []
- for component in os.listdir(client) + ['.']:
- test_path = os.path.join(client, component, 'tests')
- if os.path.exists(test_path) and os.path.isdir(test_path):
- suites = GetSuites(test_path)
- repositories += [TestRepository(os.path.join(test_path, name))
- for name in suites]
- repositories += [TestRepository(a) for a in options.suite]
-
- root = LiteralTestSuite(repositories)
- if args:
- paths = []
- for arg in args:
- path = _SplitPath(arg)
- paths.append(path)
- else:
- paths = [_SplitPath(t) for t in BUILT_IN_TESTS]
-
- # Check for --valgrind option. If enabled, we overwrite the special
- # command flag with a command that uses the tools/valgrind.py script.
- if options.valgrind:
- run_valgrind = os.path.join(client, 'runtime', 'tools', 'valgrind.py')
- options.special_command = 'python -u ' + run_valgrind + ' @'
-
- context = Context(client,
- options.verbose,
- options.os,
- options.timeout,
- GetSpecialCommandProcessor(options.special_command),
- options.suppress_dialogs,
- options.executable,
- options.flags,
- options.keep_temporary_files,
- options.batch,
- options.checked)
-
- # Get status for tests
- sections = []
- defs = {}
- root.GetTestStatus(context, sections, defs)
- config = Configuration(sections, defs)
-
- # List the tests
- all_cases = []
- all_unused = []
- globally_unused_rules = None
- for path in paths:
- for mode in options.mode:
- for arch in options.arch:
- for component in options.component:
- env = {
- 'mode': mode,
- 'system': utils.GuessOS(),
- 'arch': arch,
- 'component': component,
- 'checked': options.checked,
- 'unchecked': not options.checked,
- }
- test_list = root.ListTests([], path, context, mode, arch, component)
- (cases, unused_rules, unused_outcomes) = config.ClassifyTests(
- test_list, env)
- if globally_unused_rules is None:
- globally_unused_rules = set(unused_rules)
- else:
- globally_unused_rules = (
- globally_unused_rules.intersection(unused_rules))
- all_cases += cases
- all_unused.append(unused_rules)
-
- if options.report:
- PrintReport(all_cases)
-
- if options.list:
- PrintTests(all_cases)
- return 0
-
- result = None
-
- def DoSkip(case):
- return testing.SKIP in case.outcomes or testing.SLOW in case.outcomes
-
- cases_to_run = [c for c in all_cases if not DoSkip(c)]
- # Creating test cases may generate temporary files. Make sure
- # Skipped tests clean up these files.
- for c in all_cases:
- if DoSkip(c): c.case.Cleanup()
-
- if cases_to_run:
- try:
- start = time.time()
- if RunTestCases(cases_to_run, options.progress, options.tasks,
- context, script_start_time):
- result = 0
- else:
- result = 1
- duration = time.time() - start
- except KeyboardInterrupt:
- print 'Exiting on KeyboardInterrupt'
- return 1
- else:
- print 'No tests to run.'
- return 0
-
- if options.time:
- print
- print '--- Total time: %s ---' % FormatTime(duration)
- timed_tests = [t.case for t in cases_to_run if not t.case.duration is None]
- timed_tests.sort(lambda a, b: a.CompareTime(b))
- index = 1
- for entry in timed_tests[:20]:
- t = FormatTime(entry.duration)
- print '%4i (%s) %s' % (index, t, entry.GetLabel())
- index += 1
-
- return result
-
-
-if __name__ == '__main__':
- sys.exit(Main())
« no previous file with comments | « no previous file | test_wrapper.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698