| Index: tools/perf/generate_legacy_perf_dashboard_json_unittest.py | 
| diff --git a/tools/perf/generate_legacy_perf_dashboard_json_unittest.py b/tools/perf/generate_legacy_perf_dashboard_json_unittest.py | 
| new file mode 100755 | 
| index 0000000000000000000000000000000000000000..aa65bff059f6c27ed81efb45a9d5039d95ec79ae | 
| --- /dev/null | 
| +++ b/tools/perf/generate_legacy_perf_dashboard_json_unittest.py | 
| @@ -0,0 +1,101 @@ | 
| +#!/usr/bin/env python | 
| +# Copyright 2016 The Chromium Authors. All rights reserved. | 
| +# Use of this source code is governed by a BSD-style license that can be | 
| +# found in the LICENSE file. | 
| + | 
| + | 
| +import json | 
| +import os | 
| +import unittest | 
| + | 
| +import generate_legacy_perf_dashboard_json | 
| + | 
| +class LegacyResultsProcessorUnittest(unittest.TestCase): | 
| +  def setUp(self): | 
| +    """Set up for all test method of each test method below.""" | 
| +    super(LegacyResultsProcessorUnittest, self).setUp() | 
| +    self.data_directory = os.path.join(os.path.dirname( | 
| +      os.path.abspath(__file__)), 'testdata') | 
| + | 
| +  def _ConstructDefaultProcessor(self): | 
| +    """Creates a LegacyResultsProcessor instance. | 
| + | 
| +    Returns: | 
| +      An instance of LegacyResultsProcessor class | 
| +    """ | 
| +    return generate_legacy_perf_dashboard_json.LegacyResultsProcessor() | 
| + | 
| +  def _ProcessLog(self, log_processor, logfile):  # pylint: disable=R0201 | 
| +    """Reads in a input log file and processes it. | 
| + | 
| +    This changes the state of the log processor object; the output is stored | 
| +    in the object and can be gotten using the PerformanceLogs() method. | 
| + | 
| +    Args: | 
| +      log_processor: An PerformanceLogProcessor instance. | 
| +      logfile: File name of an input performance results log file. | 
| +    """ | 
| +    for line in open(os.path.join(self.data_directory, logfile)): | 
| +      log_processor.ProcessLine(line) | 
| + | 
| +  def _CheckFileExistsWithData(self, logs, graph): | 
| +    """Asserts that |graph| exists in the |logs| dict and is non-empty.""" | 
| +    self.assertTrue(graph in logs, 'File %s was not output.' % graph) | 
| +    self.assertTrue(logs[graph], 'File %s did not contain data.' % graph) | 
| + | 
| +  def _ConstructParseAndCheckLogfiles(self, inputfiles, graphs): | 
| +    """Uses a log processor to process the given input files. | 
| + | 
| +    Args: | 
| +      inputfiles: A list of input performance results log file names. | 
| +      logfiles: List of expected output ".dat" file names. | 
| + | 
| +    Returns: | 
| +      A dictionary mapping output file name to output file lines. | 
| +    """ | 
| +    parser = self._ConstructDefaultProcessor() | 
| +    for inputfile in inputfiles: | 
| +      self._ProcessLog(parser, inputfile) | 
| + | 
| +    logs = json.loads(parser.GenerateGraphJson()) | 
| +    for graph in graphs: | 
| +      self._CheckFileExistsWithData(logs, graph) | 
| + | 
| +    return logs | 
| + | 
| +  def _ConstructParseAndCheckJSON( | 
| +      self, inputfiles, logfiles, graphs): | 
| +    """Processes input with a log processor and checks against expectations. | 
| + | 
| +    Args: | 
| +      inputfiles: A list of input performance result log file names. | 
| +      logfiles: A list of expected output ".dat" file names. | 
| +      subdir: Subdirectory containing expected output files. | 
| +      log_processor_class: A log processor class. | 
| +    """ | 
| +    logs = self._ConstructParseAndCheckLogfiles(inputfiles, graphs) | 
| +    index = 0 | 
| +    for filename in logfiles: | 
| +      graph_name = graphs[index] | 
| +      actual = logs[graph_name] | 
| +      path = os.path.join(self.data_directory, filename) | 
| +      expected = json.load(open(path)) | 
| +      self.assertEqual(expected, actual, 'JSON data in %s did not match ' | 
| +          'expectations.' % filename) | 
| + | 
| +      index += 1 | 
| + | 
| + | 
| +  def testSummary(self): | 
| +    graphs = ['commit_charge', | 
| +        'ws_final_total', 'vm_final_browser', 'vm_final_total', | 
| +        'ws_final_browser', 'processes', 'artificial_graph'] | 
| +    # Tests the output of "summary" files, which contain per-graph data. | 
| +    input_files = ['graphing_processor.log'] | 
| +    output_files = ['%s-summary.dat' % graph for graph in graphs] | 
| + | 
| +    self._ConstructParseAndCheckJSON(input_files, output_files, graphs) | 
| + | 
| + | 
| +if __name__ == '__main__': | 
| +  unittest.main() | 
|  |