| Index: tools/valgrind/tsan_analyze.py
|
| diff --git a/tools/valgrind/tsan_analyze.py b/tools/valgrind/tsan_analyze.py
|
| deleted file mode 100755
|
| index 2c744e230196b476ec7c2f6b4bfa760d58e8d508..0000000000000000000000000000000000000000
|
| --- a/tools/valgrind/tsan_analyze.py
|
| +++ /dev/null
|
| @@ -1,271 +0,0 @@
|
| -#!/usr/bin/env python
|
| -# Copyright (c) 2011 The Chromium Authors. All rights reserved.
|
| -# Use of this source code is governed by a BSD-style license that can be
|
| -# found in the LICENSE file.
|
| -
|
| -# tsan_analyze.py
|
| -
|
| -''' Given a ThreadSanitizer output file, parses errors and uniques them.'''
|
| -
|
| -import gdb_helper
|
| -
|
| -from collections import defaultdict
|
| -import hashlib
|
| -import logging
|
| -import optparse
|
| -import os
|
| -import re
|
| -import subprocess
|
| -import sys
|
| -import time
|
| -
|
| -import common
|
| -
|
| -# Global symbol table (ugh)
|
| -TheAddressTable = None
|
| -
|
| -class _StackTraceLine(object):
|
| - def __init__(self, line, address, binary):
|
| - self.raw_line_ = line
|
| - self.address = address
|
| - self.binary = binary
|
| - def __str__(self):
|
| - global TheAddressTable
|
| - file, line = TheAddressTable.GetFileLine(self.binary, self.address)
|
| - if (file is None) or (line is None):
|
| - return self.raw_line_
|
| - else:
|
| - return self.raw_line_.replace(self.binary, '%s:%s' % (file, line))
|
| -
|
| -class TsanAnalyzer(object):
|
| - ''' Given a set of ThreadSanitizer output files, parse all the errors out of
|
| - them, unique them and output the results.'''
|
| -
|
| - LOAD_LIB_RE = re.compile('--[0-9]+-- ([^(:]*) \((0x[0-9a-f]+)\)')
|
| - TSAN_LINE_RE = re.compile('==[0-9]+==\s*[#0-9]+\s*'
|
| - '([0-9A-Fa-fx]+):'
|
| - '(?:[^ ]* )*'
|
| - '([^ :\n]+)'
|
| - '')
|
| - THREAD_CREATION_STR = ("INFO: T.* "
|
| - "(has been created by T.* at this point|is program's main thread)")
|
| -
|
| - SANITY_TEST_SUPPRESSION = ("ThreadSanitizer sanity test "
|
| - "(ToolsSanityTest.DataRace)")
|
| - TSAN_RACE_DESCRIPTION = "Possible data race"
|
| - TSAN_WARNING_DESCRIPTION = ("Unlocking a non-locked lock"
|
| - "|accessing an invalid lock"
|
| - "|which did not acquire this lock")
|
| - RACE_VERIFIER_LINE = "Confirmed a race|unexpected race"
|
| - TSAN_ASSERTION = "Assertion failed: "
|
| -
|
| - def __init__(self, use_gdb=False):
|
| - '''Reads in a set of files.'''
|
| -
|
| - self._use_gdb = use_gdb
|
| - self._cur_testcase = None
|
| -
|
| - def ReadLine(self):
|
| - self.line_ = self.cur_fd_.readline()
|
| - self.stack_trace_line_ = None
|
| - if not self._use_gdb:
|
| - return
|
| - global TheAddressTable
|
| - match = TsanAnalyzer.LOAD_LIB_RE.match(self.line_)
|
| - if match:
|
| - binary, ip = match.groups()
|
| - TheAddressTable.AddBinaryAt(binary, ip)
|
| - return
|
| - match = TsanAnalyzer.TSAN_LINE_RE.match(self.line_)
|
| - if match:
|
| - address, binary_name = match.groups()
|
| - stack_trace_line = _StackTraceLine(self.line_, address, binary_name)
|
| - TheAddressTable.Add(stack_trace_line.binary, stack_trace_line.address)
|
| - self.stack_trace_line_ = stack_trace_line
|
| -
|
| - def ReadSection(self):
|
| - """ Example of a section:
|
| - ==4528== WARNING: Possible data race: {{{
|
| - ==4528== T20 (L{}):
|
| - ==4528== #0 MyTest::Foo1
|
| - ==4528== #1 MyThread::ThreadBody
|
| - ==4528== Concurrent write happened at this point:
|
| - ==4528== T19 (L{}):
|
| - ==4528== #0 MyTest::Foo2
|
| - ==4528== #1 MyThread::ThreadBody
|
| - ==4528== }}}
|
| - ------- suppression -------
|
| - {
|
| - <Put your suppression name here>
|
| - ThreadSanitizer:Race
|
| - fun:MyTest::Foo1
|
| - fun:MyThread::ThreadBody
|
| - }
|
| - ------- end suppression -------
|
| - """
|
| - result = [self.line_]
|
| - if re.search("{{{", self.line_):
|
| - while not re.search('}}}', self.line_):
|
| - self.ReadLine()
|
| - if self.stack_trace_line_ is None:
|
| - result.append(self.line_)
|
| - else:
|
| - result.append(self.stack_trace_line_)
|
| - self.ReadLine()
|
| - if re.match('-+ suppression -+', self.line_):
|
| - # We need to calculate the suppression hash and prepend a line like
|
| - # "Suppression (error hash=#0123456789ABCDEF#):" so the buildbot can
|
| - # extract the suppression snippet.
|
| - supp = ""
|
| - while not re.match('-+ end suppression -+', self.line_):
|
| - self.ReadLine()
|
| - supp += self.line_
|
| - self.ReadLine()
|
| - if self._cur_testcase:
|
| - result.append("The report came from the `%s` test.\n" % \
|
| - self._cur_testcase)
|
| - result.append("Suppression (error hash=#%016X#):\n" % \
|
| - (int(hashlib.md5(supp).hexdigest()[:16], 16)))
|
| - result.append(" For more info on using suppressions see "
|
| - "http://dev.chromium.org/developers/how-tos/using-valgrind/threadsanitizer#TOC-Suppressing-data-races\n")
|
| - result.append(supp)
|
| - else:
|
| - self.ReadLine()
|
| -
|
| - return result
|
| -
|
| - def ReadTillTheEnd(self):
|
| - result = [self.line_]
|
| - while self.line_:
|
| - self.ReadLine()
|
| - result.append(self.line_)
|
| - return result
|
| -
|
| - def ParseReportFile(self, filename):
|
| - '''Parses a report file and returns a list of ThreadSanitizer reports.
|
| -
|
| -
|
| - Args:
|
| - filename: report filename.
|
| - Returns:
|
| - list of (list of (str iff self._use_gdb, _StackTraceLine otherwise)).
|
| - '''
|
| - ret = []
|
| - self.cur_fd_ = open(filename, 'r')
|
| -
|
| - while True:
|
| - # Read ThreadSanitizer reports.
|
| - self.ReadLine()
|
| - if not self.line_:
|
| - break
|
| -
|
| - while True:
|
| - tmp = []
|
| - while re.search(TsanAnalyzer.RACE_VERIFIER_LINE, self.line_):
|
| - tmp.append(self.line_)
|
| - self.ReadLine()
|
| - while re.search(TsanAnalyzer.THREAD_CREATION_STR, self.line_):
|
| - tmp.extend(self.ReadSection())
|
| - if re.search(TsanAnalyzer.TSAN_RACE_DESCRIPTION, self.line_):
|
| - tmp.extend(self.ReadSection())
|
| - ret.append(tmp) # includes RaceVerifier and thread creation stacks
|
| - elif (re.search(TsanAnalyzer.TSAN_WARNING_DESCRIPTION, self.line_) and
|
| - not common.IsWindows()): # workaround for http://crbug.com/53198
|
| - tmp.extend(self.ReadSection())
|
| - ret.append(tmp)
|
| - else:
|
| - break
|
| -
|
| - tmp = []
|
| - if re.search(TsanAnalyzer.TSAN_ASSERTION, self.line_):
|
| - tmp.extend(self.ReadTillTheEnd())
|
| - ret.append(tmp)
|
| - break
|
| -
|
| - match = re.search("used_suppression:\s+([0-9]+)\s(.*)", self.line_)
|
| - if match:
|
| - count, supp_name = match.groups()
|
| - count = int(count)
|
| - self.used_suppressions[supp_name] += count
|
| - self.cur_fd_.close()
|
| - return ret
|
| -
|
| - def GetReports(self, files):
|
| - '''Extracts reports from a set of files.
|
| -
|
| - Reads a set of files and returns a list of all discovered
|
| - ThreadSanitizer race reports. As a side effect, populates
|
| - self.used_suppressions with appropriate info.
|
| - '''
|
| -
|
| - global TheAddressTable
|
| - if self._use_gdb:
|
| - TheAddressTable = gdb_helper.AddressTable()
|
| - else:
|
| - TheAddressTable = None
|
| - reports = []
|
| - self.used_suppressions = defaultdict(int)
|
| - for file in files:
|
| - reports.extend(self.ParseReportFile(file))
|
| - if self._use_gdb:
|
| - TheAddressTable.ResolveAll()
|
| - # Make each line of each report a string.
|
| - reports = map(lambda(x): map(str, x), reports)
|
| - return [''.join(report_lines) for report_lines in reports]
|
| -
|
| - def Report(self, files, testcase, check_sanity=False):
|
| - '''Reads in a set of files and prints ThreadSanitizer report.
|
| -
|
| - Args:
|
| - files: A list of filenames.
|
| - check_sanity: if true, search for SANITY_TEST_SUPPRESSIONS
|
| - '''
|
| -
|
| - # We set up _cur_testcase class-wide variable to avoid passing it through
|
| - # about 5 functions.
|
| - self._cur_testcase = testcase
|
| - reports = self.GetReports(files)
|
| - self._cur_testcase = None # just in case, shouldn't be used anymore
|
| -
|
| - common.PrintUsedSuppressionsList(self.used_suppressions)
|
| -
|
| -
|
| - retcode = 0
|
| - if reports:
|
| - sys.stdout.flush()
|
| - sys.stderr.flush()
|
| - logging.info("FAIL! Found %i report(s)" % len(reports))
|
| - for report in reports:
|
| - logging.info('\n' + report)
|
| - sys.stdout.flush()
|
| - retcode = -1
|
| -
|
| - # Report tool's insanity even if there were errors.
|
| - if (check_sanity and
|
| - TsanAnalyzer.SANITY_TEST_SUPPRESSION not in self.used_suppressions):
|
| - logging.error("FAIL! Sanity check failed!")
|
| - retcode = -3
|
| -
|
| - if retcode != 0:
|
| - return retcode
|
| -
|
| - logging.info("PASS: No reports found")
|
| - return 0
|
| -
|
| -
|
| -def main():
|
| - '''For testing only. The TsanAnalyzer class should be imported instead.'''
|
| - parser = optparse.OptionParser("usage: %prog <files to analyze>")
|
| -
|
| - (options, args) = parser.parse_args()
|
| - if not args:
|
| - parser.error("no filename specified")
|
| - filenames = args
|
| -
|
| - logging.getLogger().setLevel(logging.INFO)
|
| - analyzer = TsanAnalyzer(use_gdb=True)
|
| - return analyzer.Report(filenames, None)
|
| -
|
| -
|
| -if __name__ == '__main__':
|
| - sys.exit(main())
|
|
|