| Index: third_party/lcov/bin/lcov
|
| ===================================================================
|
| --- third_party/lcov/bin/lcov (revision 0)
|
| +++ third_party/lcov/bin/lcov (revision 0)
|
| @@ -0,0 +1,2699 @@
|
| +#!/usr/bin/perl -w
|
| +#
|
| +# Copyright (c) International Business Machines Corp., 2002,2007
|
| +#
|
| +# This program is free software; you can redistribute it and/or modify
|
| +# it under the terms of the GNU General Public License as published by
|
| +# the Free Software Foundation; either version 2 of the License, or (at
|
| +# your option) any later version.
|
| +#
|
| +# This program is distributed in the hope that it will be useful, but
|
| +# WITHOUT ANY WARRANTY; without even the implied warranty of
|
| +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
| +# General Public License for more details.
|
| +#
|
| +# You should have received a copy of the GNU General Public License
|
| +# along with this program; if not, write to the Free Software
|
| +# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
| +#
|
| +#
|
| +# lcov
|
| +#
|
| +# This is a wrapper script which provides a single interface for accessing
|
| +# LCOV coverage data.
|
| +#
|
| +#
|
| +# History:
|
| +# 2002-08-29 created by Peter Oberparleiter <Peter.Oberparleiter@de.ibm.com>
|
| +# IBM Lab Boeblingen
|
| +# 2002-09-05 / Peter Oberparleiter: implemented --kernel-directory +
|
| +# multiple directories
|
| +# 2002-10-16 / Peter Oberparleiter: implemented --add-tracefile option
|
| +# 2002-10-17 / Peter Oberparleiter: implemented --extract option
|
| +# 2002-11-04 / Peter Oberparleiter: implemented --list option
|
| +# 2003-03-07 / Paul Larson: Changed to make it work with the latest gcov
|
| +# kernel patch. This will break it with older gcov-kernel
|
| +# patches unless you change the value of $gcovmod in this script
|
| +# 2003-04-07 / Peter Oberparleiter: fixed bug which resulted in an error
|
| +# when trying to combine .info files containing data without
|
| +# a test name
|
| +# 2003-04-10 / Peter Oberparleiter: extended Paul's change so that LCOV
|
| +# works both with the new and the old gcov-kernel patch
|
| +# 2003-04-10 / Peter Oberparleiter: added $gcov_dir constant in anticipation
|
| +# of a possible move of the gcov kernel directory to another
|
| +# file system in a future version of the gcov-kernel patch
|
| +# 2003-04-15 / Paul Larson: make info write to STDERR, not STDOUT
|
| +# 2003-04-15 / Paul Larson: added --remove option
|
| +# 2003-04-30 / Peter Oberparleiter: renamed --reset to --zerocounters
|
| +# to remove naming ambiguity with --remove
|
| +# 2003-04-30 / Peter Oberparleiter: adjusted help text to include --remove
|
| +# 2003-06-27 / Peter Oberparleiter: implemented --diff
|
| +# 2003-07-03 / Peter Oberparleiter: added line checksum support, added
|
| +# --no-checksum
|
| +# 2003-12-11 / Laurent Deniel: added --follow option
|
| +# 2004-03-29 / Peter Oberparleiter: modified --diff option to better cope with
|
| +# ambiguous patch file entries, modified --capture option to use
|
| +# modprobe before insmod (needed for 2.6)
|
| +# 2004-03-30 / Peter Oberparleiter: added --path option
|
| +# 2004-08-09 / Peter Oberparleiter: added configuration file support
|
| +# 2008-08-13 / Peter Oberparleiter: added function coverage support
|
| +#
|
| +
|
| +use strict;
|
| +use File::Basename;
|
| +use Getopt::Long;
|
| +
|
| +
|
| +# Global constants
|
| +our $lcov_version = "LCOV version 1.7";
|
| +our $lcov_url = "http://ltp.sourceforge.net/coverage/lcov.php";
|
| +our $tool_name = basename($0);
|
| +
|
| +# Names of the GCOV kernel module
|
| +our @gcovmod = ("gcov-prof", "gcov-proc");
|
| +
|
| +# Directory containing gcov kernel files
|
| +our $gcov_dir = "/proc/gcov";
|
| +
|
| +# The location of the insmod tool
|
| +our $insmod_tool = "/sbin/insmod";
|
| +
|
| +# The location of the modprobe tool
|
| +our $modprobe_tool = "/sbin/modprobe";
|
| +
|
| +# The location of the rmmod tool
|
| +our $rmmod_tool = "/sbin/rmmod";
|
| +
|
| +# Where to create temporary directories
|
| +our $tmp_dir = "/tmp";
|
| +
|
| +# How to prefix a temporary directory name
|
| +our $tmp_prefix = "tmpdir";
|
| +
|
| +
|
| +# Prototypes
|
| +sub print_usage(*);
|
| +sub check_options();
|
| +sub userspace_reset();
|
| +sub userspace_capture();
|
| +sub kernel_reset();
|
| +sub kernel_capture();
|
| +sub add_traces();
|
| +sub read_info_file($);
|
| +sub get_info_entry($);
|
| +sub set_info_entry($$$$$$$;$$$$);
|
| +sub add_counts($$);
|
| +sub merge_checksums($$$);
|
| +sub combine_info_entries($$$);
|
| +sub combine_info_files($$);
|
| +sub write_info_file(*$);
|
| +sub extract();
|
| +sub remove();
|
| +sub list();
|
| +sub get_common_filename($$);
|
| +sub read_diff($);
|
| +sub diff();
|
| +sub system_no_output($@);
|
| +sub read_config($);
|
| +sub apply_config($);
|
| +sub info(@);
|
| +sub unload_module($);
|
| +sub check_and_load_kernel_module();
|
| +sub create_temp_dir();
|
| +sub transform_pattern($);
|
| +sub warn_handler($);
|
| +sub die_handler($);
|
| +
|
| +
|
| +# Global variables & initialization
|
| +our @directory; # Specifies where to get coverage data from
|
| +our @kernel_directory; # If set, captures only from specified kernel subdirs
|
| +our @add_tracefile; # If set, reads in and combines all files in list
|
| +our $list; # If set, list contents of tracefile
|
| +our $extract; # If set, extracts parts of tracefile
|
| +our $remove; # If set, removes parts of tracefile
|
| +our $diff; # If set, modifies tracefile according to diff
|
| +our $reset; # If set, reset all coverage data to zero
|
| +our $capture; # If set, capture data
|
| +our $output_filename; # Name for file to write coverage data to
|
| +our $test_name = ""; # Test case name
|
| +our $quiet = ""; # If set, suppress information messages
|
| +our $help; # Help option flag
|
| +our $version; # Version option flag
|
| +our $convert_filenames; # If set, convert filenames when applying diff
|
| +our $strip; # If set, strip leading directories when applying diff
|
| +our $need_unload; # If set, unload gcov kernel module
|
| +our $temp_dir_name; # Name of temporary directory
|
| +our $cwd = `pwd`; # Current working directory
|
| +our $to_file; # If set, indicates that output is written to a file
|
| +our $follow; # If set, indicates that find shall follow links
|
| +our $diff_path = ""; # Path removed from tracefile when applying diff
|
| +our $base_directory; # Base directory (cwd of gcc during compilation)
|
| +our $checksum; # If set, calculate a checksum for each line
|
| +our $no_checksum; # If set, don't calculate a checksum for each line
|
| +our $compat_libtool; # If set, indicates that libtool mode is to be enabled
|
| +our $no_compat_libtool; # If set, indicates that libtool mode is to be disabled
|
| +our $gcov_tool;
|
| +our $ignore_errors;
|
| +our $initial;
|
| +our $no_recursion = 0;
|
| +our $maxdepth;
|
| +our $config; # Configuration file contents
|
| +chomp($cwd);
|
| +our $tool_dir = dirname($0); # Directory where genhtml tool is installed
|
| +
|
| +
|
| +#
|
| +# Code entry point
|
| +#
|
| +
|
| +$SIG{__WARN__} = \&warn_handler;
|
| +$SIG{__DIE__} = \&die_handler;
|
| +
|
| +# Add current working directory if $tool_dir is not already an absolute path
|
| +if (! ($tool_dir =~ /^\/(.*)$/))
|
| +{
|
| + $tool_dir = "$cwd/$tool_dir";
|
| +}
|
| +
|
| +# Read configuration file if available
|
| +if (-r $ENV{"HOME"}."/.lcovrc")
|
| +{
|
| + $config = read_config($ENV{"HOME"}."/.lcovrc");
|
| +}
|
| +elsif (-r "/etc/lcovrc")
|
| +{
|
| + $config = read_config("/etc/lcovrc");
|
| +}
|
| +
|
| +if ($config)
|
| +{
|
| + # Copy configuration file values to variables
|
| + apply_config({
|
| + "lcov_gcov_dir" => \$gcov_dir,
|
| + "lcov_insmod_tool" => \$insmod_tool,
|
| + "lcov_modprobe_tool" => \$modprobe_tool,
|
| + "lcov_rmmod_tool" => \$rmmod_tool,
|
| + "lcov_tmp_dir" => \$tmp_dir});
|
| +}
|
| +
|
| +# Parse command line options
|
| +if (!GetOptions("directory|d|di=s" => \@directory,
|
| + "add-tracefile=s" => \@add_tracefile,
|
| + "list=s" => \$list,
|
| + "kernel-directory=s" => \@kernel_directory,
|
| + "extract=s" => \$extract,
|
| + "remove=s" => \$remove,
|
| + "diff=s" => \$diff,
|
| + "convert-filenames" => \$convert_filenames,
|
| + "strip=i" => \$strip,
|
| + "capture|c" => \$capture,
|
| + "output-file=s" => \$output_filename,
|
| + "test-name=s" => \$test_name,
|
| + "zerocounters" => \$reset,
|
| + "quiet" => \$quiet,
|
| + "help|?" => \$help,
|
| + "version" => \$version,
|
| + "follow" => \$follow,
|
| + "path=s" => \$diff_path,
|
| + "base-directory=s" => \$base_directory,
|
| + "checksum" => \$checksum,
|
| + "no-checksum" => \$no_checksum,
|
| + "compat-libtool" => \$compat_libtool,
|
| + "no-compat-libtool" => \$no_compat_libtool,
|
| + "gcov-tool=s" => \$gcov_tool,
|
| + "ignore-errors=s" => \$ignore_errors,
|
| + "initial|i" => \$initial,
|
| + "no-recursion" => \$no_recursion
|
| + ))
|
| +{
|
| + print(STDERR "Use $tool_name --help to get usage information\n");
|
| + exit(1);
|
| +}
|
| +else
|
| +{
|
| + # Merge options
|
| + if (defined($no_checksum))
|
| + {
|
| + $checksum = ($no_checksum ? 0 : 1);
|
| + $no_checksum = undef;
|
| + }
|
| +
|
| + if (defined($no_compat_libtool))
|
| + {
|
| + $compat_libtool = ($no_compat_libtool ? 0 : 1);
|
| + $no_compat_libtool = undef;
|
| + }
|
| +}
|
| +
|
| +# Check for help option
|
| +if ($help)
|
| +{
|
| + print_usage(*STDOUT);
|
| + exit(0);
|
| +}
|
| +
|
| +# Check for version option
|
| +if ($version)
|
| +{
|
| + print("$tool_name: $lcov_version\n");
|
| + exit(0);
|
| +}
|
| +
|
| +# Normalize --path text
|
| +$diff_path =~ s/\/$//;
|
| +
|
| +if ($follow)
|
| +{
|
| + $follow = "-follow";
|
| +}
|
| +else
|
| +{
|
| + $follow = "";
|
| +}
|
| +
|
| +if ($no_recursion)
|
| +{
|
| + $maxdepth = "-maxdepth 1";
|
| +}
|
| +else
|
| +{
|
| + $maxdepth = "";
|
| +}
|
| +
|
| +# Check for valid options
|
| +check_options();
|
| +
|
| +# Only --extract, --remove and --diff allow unnamed parameters
|
| +if (@ARGV && !($extract || $remove || $diff))
|
| +{
|
| + die("Extra parameter found\n".
|
| + "Use $tool_name --help to get usage information\n");
|
| +}
|
| +
|
| +# Check for output filename
|
| +$to_file = ($output_filename && ($output_filename ne "-"));
|
| +
|
| +if ($capture)
|
| +{
|
| + if (!$to_file)
|
| + {
|
| + # Option that tells geninfo to write to stdout
|
| + $output_filename = "-";
|
| + }
|
| +}
|
| +else
|
| +{
|
| + if ($initial)
|
| + {
|
| + die("Option --initial is only valid when capturing data (-c)\n".
|
| + "Use $tool_name --help to get usage information\n");
|
| + }
|
| +}
|
| +
|
| +# Check for requested functionality
|
| +if ($reset)
|
| +{
|
| + # Differentiate between user space and kernel reset
|
| + if (@directory)
|
| + {
|
| + userspace_reset();
|
| + }
|
| + else
|
| + {
|
| + kernel_reset();
|
| + }
|
| +}
|
| +elsif ($capture)
|
| +{
|
| + # Differentiate between user space and kernel
|
| + if (@directory)
|
| + {
|
| + userspace_capture();
|
| + }
|
| + else
|
| + {
|
| + kernel_capture();
|
| + }
|
| +}
|
| +elsif (@add_tracefile)
|
| +{
|
| + add_traces();
|
| +}
|
| +elsif ($remove)
|
| +{
|
| + remove();
|
| +}
|
| +elsif ($extract)
|
| +{
|
| + extract();
|
| +}
|
| +elsif ($list)
|
| +{
|
| + list();
|
| +}
|
| +elsif ($diff)
|
| +{
|
| + if (scalar(@ARGV) != 1)
|
| + {
|
| + die("ERROR: option --diff requires one additional argument!\n".
|
| + "Use $tool_name --help to get usage information\n");
|
| + }
|
| + diff();
|
| +}
|
| +
|
| +info("Done.\n");
|
| +exit(0);
|
| +
|
| +#
|
| +# print_usage(handle)
|
| +#
|
| +# Print usage information.
|
| +#
|
| +
|
| +sub print_usage(*)
|
| +{
|
| + local *HANDLE = $_[0];
|
| +
|
| + print(HANDLE <<END_OF_USAGE);
|
| +Usage: $tool_name [OPTIONS]
|
| +
|
| +Use lcov to collect coverage data from either the currently running Linux
|
| +kernel or from a user space application. Specify the --directory option to
|
| +get coverage data for a user space program.
|
| +
|
| +Misc:
|
| + -h, --help Print this help, then exit
|
| + -v, --version Print version number, then exit
|
| + -q, --quiet Do not print progress messages
|
| +
|
| +Operation:
|
| + -z, --zerocounters Reset all execution counts to zero
|
| + -c, --capture Capture coverage data
|
| + -a, --add-tracefile FILE Add contents of tracefiles
|
| + -e, --extract FILE PATTERN Extract files matching PATTERN from FILE
|
| + -r, --remove FILE PATTERN Remove files matching PATTERN from FILE
|
| + -l, --list FILE List contents of tracefile FILE
|
| + --diff FILE DIFF Transform tracefile FILE according to DIFF
|
| +
|
| +Options:
|
| + -i, --initial Capture initial zero coverage data
|
| + -t, --test-name NAME Specify test name to be stored with data
|
| + -o, --output-file FILENAME Write data to FILENAME instead of stdout
|
| + -d, --directory DIR Use .da files in DIR instead of kernel
|
| + -f, --follow Follow links when searching .da files
|
| + -k, --kernel-directory KDIR Capture kernel coverage data only from KDIR
|
| + -b, --base-directory DIR Use DIR as base directory for relative paths
|
| + --convert-filenames Convert filenames when applying diff
|
| + --strip DEPTH Strip initial DEPTH directory levels in diff
|
| + --path PATH Strip PATH from tracefile when applying diff
|
| + --(no-)checksum Enable (disable) line checksumming
|
| + --(no-)compat-libtool Enable (disable) libtool compatibility mode
|
| + --gcov-tool TOOL Specify gcov tool location
|
| + --ignore-errors ERRORS Continue after ERRORS (gcov, source)
|
| + --no-recursion Exlude subdirectories from processing
|
| +
|
| +For more information see: $lcov_url
|
| +END_OF_USAGE
|
| + ;
|
| +}
|
| +
|
| +
|
| +#
|
| +# check_options()
|
| +#
|
| +# Check for valid combination of command line options. Die on error.
|
| +#
|
| +
|
| +sub check_options()
|
| +{
|
| + my $i = 0;
|
| +
|
| + # Count occurrence of mutually exclusive options
|
| + $reset && $i++;
|
| + $capture && $i++;
|
| + @add_tracefile && $i++;
|
| + $extract && $i++;
|
| + $remove && $i++;
|
| + $list && $i++;
|
| + $diff && $i++;
|
| +
|
| + if ($i == 0)
|
| + {
|
| + die("Need one of the options -z, -c, -a, -e, -r, -l or ".
|
| + "--diff\n".
|
| + "Use $tool_name --help to get usage information\n");
|
| + }
|
| + elsif ($i > 1)
|
| + {
|
| + die("ERROR: only one of -z, -c, -a, -e, -r, -l or ".
|
| + "--diff allowed!\n".
|
| + "Use $tool_name --help to get usage information\n");
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# userspace_reset()
|
| +#
|
| +# Reset coverage data found in DIRECTORY by deleting all contained .da files.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub userspace_reset()
|
| +{
|
| + my $current_dir;
|
| + my @file_list;
|
| +
|
| + foreach $current_dir (@directory)
|
| + {
|
| + info("Deleting all .da files in $current_dir".
|
| + ($no_recursion?"\n":" and subdirectories\n"));
|
| + @file_list = `find "$current_dir" $maxdepth $follow -name \\*\\.da -o -name \\*\\.gcda -type f 2>/dev/null`;
|
| + chomp(@file_list);
|
| + foreach (@file_list)
|
| + {
|
| + unlink($_) or die("ERROR: cannot remove file $_!\n");
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# userspace_capture()
|
| +#
|
| +# Capture coverage data found in DIRECTORY and write it to OUTPUT_FILENAME
|
| +# if specified, otherwise to STDOUT.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub userspace_capture()
|
| +{
|
| + my @param;
|
| + my $file_list = join(" ", @directory);
|
| +
|
| + info("Capturing coverage data from $file_list\n");
|
| + @param = ("$tool_dir/geninfo", @directory);
|
| + if ($output_filename)
|
| + {
|
| + @param = (@param, "--output-filename", $output_filename);
|
| + }
|
| + if ($test_name)
|
| + {
|
| + @param = (@param, "--test-name", $test_name);
|
| + }
|
| + if ($follow)
|
| + {
|
| + @param = (@param, "--follow");
|
| + }
|
| + if ($quiet)
|
| + {
|
| + @param = (@param, "--quiet");
|
| + }
|
| + if (defined($checksum))
|
| + {
|
| + if ($checksum)
|
| + {
|
| + @param = (@param, "--checksum");
|
| + }
|
| + else
|
| + {
|
| + @param = (@param, "--no-checksum");
|
| + }
|
| + }
|
| + if ($base_directory)
|
| + {
|
| + @param = (@param, "--base-directory", $base_directory);
|
| + }
|
| + if ($no_compat_libtool)
|
| + {
|
| + @param = (@param, "--no-compat-libtool");
|
| + }
|
| + elsif ($compat_libtool)
|
| + {
|
| + @param = (@param, "--compat-libtool");
|
| + }
|
| + if ($gcov_tool)
|
| + {
|
| + @param = (@param, "--gcov-tool", $gcov_tool);
|
| + }
|
| + if ($ignore_errors)
|
| + {
|
| + @param = (@param, "--ignore-errors", $ignore_errors);
|
| + }
|
| + if ($initial)
|
| + {
|
| + @param = (@param, "--initial");
|
| + }
|
| + if ($no_recursion)
|
| + {
|
| + @param = (@param, "--no-recursion");
|
| + }
|
| +
|
| + system(@param);
|
| + exit($? >> 8);
|
| +}
|
| +
|
| +
|
| +#
|
| +# kernel_reset()
|
| +#
|
| +# Reset kernel coverage.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub kernel_reset()
|
| +{
|
| + local *HANDLE;
|
| + check_and_load_kernel_module();
|
| +
|
| + info("Resetting kernel execution counters\n");
|
| + open(HANDLE, ">$gcov_dir/vmlinux") or
|
| + die("ERROR: cannot write to $gcov_dir/vmlinux!\n");
|
| + print(HANDLE "0");
|
| + close(HANDLE);
|
| +
|
| + # Unload module if we loaded it in the first place
|
| + if ($need_unload)
|
| + {
|
| + unload_module($need_unload);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# kernel_capture()
|
| +#
|
| +# Capture kernel coverage data and write it to OUTPUT_FILENAME if specified,
|
| +# otherwise stdout.
|
| +#
|
| +
|
| +sub kernel_capture()
|
| +{
|
| + my @param;
|
| +
|
| + check_and_load_kernel_module();
|
| +
|
| + # Make sure the temporary directory is removed upon script termination
|
| + END
|
| + {
|
| + if ($temp_dir_name)
|
| + {
|
| + stat($temp_dir_name);
|
| + if (-r _)
|
| + {
|
| + info("Removing temporary directory ".
|
| + "$temp_dir_name\n");
|
| +
|
| + # Remove temporary directory
|
| + system("rm", "-rf", $temp_dir_name)
|
| + and warn("WARNING: cannot remove ".
|
| + "temporary directory ".
|
| + "$temp_dir_name!\n");
|
| + }
|
| + }
|
| + }
|
| +
|
| + # Get temporary directory
|
| + $temp_dir_name = create_temp_dir();
|
| +
|
| + info("Copying kernel data to temporary directory $temp_dir_name\n");
|
| +
|
| + if (!@kernel_directory)
|
| + {
|
| + # Copy files from gcov kernel directory
|
| + system("cp", "-dr", $gcov_dir, $temp_dir_name)
|
| + and die("ERROR: cannot copy files from $gcov_dir!\n");
|
| + }
|
| + else
|
| + {
|
| + # Prefix list of kernel sub-directories with the gcov kernel
|
| + # directory
|
| + @kernel_directory = map("$gcov_dir/$_", @kernel_directory);
|
| +
|
| + # Copy files from gcov kernel directory
|
| + system("cp", "-dr", @kernel_directory, $temp_dir_name)
|
| + and die("ERROR: cannot copy files from ".
|
| + join(" ", @kernel_directory)."!\n");
|
| + }
|
| +
|
| + # Make directories writable
|
| + system("find", $temp_dir_name, "-type", "d", "-exec", "chmod", "u+w",
|
| + "{}", ";")
|
| + and die("ERROR: cannot modify access rights for ".
|
| + "$temp_dir_name!\n");
|
| +
|
| + # Make files writable
|
| + system("find", $temp_dir_name, "-type", "f", "-exec", "chmod", "u+w",
|
| + "{}", ";")
|
| + and die("ERROR: cannot modify access rights for ".
|
| + "$temp_dir_name!\n");
|
| +
|
| + # Capture data
|
| + info("Capturing coverage data from $temp_dir_name\n");
|
| + @param = ("$tool_dir/geninfo", $temp_dir_name);
|
| + if ($output_filename)
|
| + {
|
| + @param = (@param, "--output-filename", $output_filename);
|
| + }
|
| + if ($test_name)
|
| + {
|
| + @param = (@param, "--test-name", $test_name);
|
| + }
|
| + if ($follow)
|
| + {
|
| + @param = (@param, "--follow");
|
| + }
|
| + if ($quiet)
|
| + {
|
| + @param = (@param, "--quiet");
|
| + }
|
| + if (defined($checksum))
|
| + {
|
| + if ($checksum)
|
| + {
|
| + @param = (@param, "--checksum");
|
| + }
|
| + else
|
| + {
|
| + @param = (@param, "--no-checksum");
|
| + }
|
| + }
|
| + if ($base_directory)
|
| + {
|
| + @param = (@param, "--base-directory", $base_directory);
|
| + }
|
| + if ($no_compat_libtool)
|
| + {
|
| + @param = (@param, "--no-compat-libtool");
|
| + }
|
| + elsif ($compat_libtool)
|
| + {
|
| + @param = (@param, "--compat-libtool");
|
| + }
|
| + if ($gcov_tool)
|
| + {
|
| + @param = (@param, "--gcov-tool", $gcov_tool);
|
| + }
|
| + if ($ignore_errors)
|
| + {
|
| + @param = (@param, "--ignore-errors", $ignore_errors);
|
| + }
|
| + if ($initial)
|
| + {
|
| + @param = (@param, "--initial");
|
| + }
|
| + system(@param) and exit($? >> 8);
|
| +
|
| +
|
| + # Unload module if we loaded it in the first place
|
| + if ($need_unload)
|
| + {
|
| + unload_module($need_unload);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# info(printf_parameter)
|
| +#
|
| +# Use printf to write PRINTF_PARAMETER to stdout only when the $quiet flag
|
| +# is not set.
|
| +#
|
| +
|
| +sub info(@)
|
| +{
|
| + if (!$quiet)
|
| + {
|
| + # Print info string
|
| + if ($to_file)
|
| + {
|
| + print(@_)
|
| + }
|
| + else
|
| + {
|
| + # Don't interfer with the .info output to STDOUT
|
| + printf(STDERR @_);
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# Check if the gcov kernel module is loaded. If it is, exit, if not, try
|
| +# to load it.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub check_and_load_kernel_module()
|
| +{
|
| + my $module_name;
|
| +
|
| + # Is it loaded already?
|
| + stat("$gcov_dir");
|
| + if (-r _) { return(); }
|
| +
|
| + info("Loading required gcov kernel module.\n");
|
| +
|
| + # Do we have access to the insmod tool?
|
| + stat($insmod_tool);
|
| + if (!-x _)
|
| + {
|
| + die("ERROR: need insmod tool ($insmod_tool) to access kernel ".
|
| + "coverage data!\n");
|
| + }
|
| + # Do we have access to the modprobe tool?
|
| + stat($modprobe_tool);
|
| + if (!-x _)
|
| + {
|
| + die("ERROR: need modprobe tool ($modprobe_tool) to access ".
|
| + "kernel coverage data!\n");
|
| + }
|
| +
|
| + # Try some possibilities of where the gcov kernel module may be found
|
| + foreach $module_name (@gcovmod)
|
| + {
|
| + # Try to load module from system wide module directory
|
| + # /lib/modules
|
| + if (system_no_output(3, $modprobe_tool, $module_name) == 0)
|
| + {
|
| + # Succeeded
|
| + $need_unload = $module_name;
|
| + return();
|
| + }
|
| +
|
| + # Try to load linux 2.5/2.6 module from tool directory
|
| + if (system_no_output(3, $insmod_tool,
|
| + "$tool_dir/$module_name.ko") == 0)
|
| + {
|
| + # Succeeded
|
| + $need_unload = $module_name;
|
| + return();
|
| + }
|
| +
|
| + # Try to load linux 2.4 module from tool directory
|
| + if (system_no_output(3, $insmod_tool,
|
| + "$tool_dir/$module_name.o") == 0)
|
| + {
|
| + # Succeeded
|
| + $need_unload = $module_name;
|
| + return();
|
| + }
|
| + }
|
| +
|
| + # Hm, loading failed - maybe we aren't root?
|
| + if ($> != 0)
|
| + {
|
| + die("ERROR: need root access to load kernel module!\n");
|
| + }
|
| +
|
| + die("ERROR: cannot load required gcov kernel module!\n");
|
| +}
|
| +
|
| +
|
| +#
|
| +# unload_module()
|
| +#
|
| +# Unload the gcov kernel module.
|
| +#
|
| +
|
| +sub unload_module($)
|
| +{
|
| + my $module = $_[0];
|
| +
|
| + info("Unloading kernel module $module\n");
|
| +
|
| + # Do we have access to the rmmod tool?
|
| + stat($rmmod_tool);
|
| + if (!-x _)
|
| + {
|
| + warn("WARNING: cannot execute rmmod tool at $rmmod_tool - ".
|
| + "gcov module still loaded!\n");
|
| + }
|
| +
|
| + # Unload gcov kernel module
|
| + system_no_output(1, $rmmod_tool, $module)
|
| + and warn("WARNING: cannot unload gcov kernel module ".
|
| + "$module!\n");
|
| +}
|
| +
|
| +
|
| +#
|
| +# create_temp_dir()
|
| +#
|
| +# Create a temporary directory and return its path.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub create_temp_dir()
|
| +{
|
| + my $dirname;
|
| + my $number = sprintf("%d", rand(1000));
|
| +
|
| + # Endless loops are evil
|
| + while ($number++ < 1000)
|
| + {
|
| + $dirname = "$tmp_dir/$tmp_prefix$number";
|
| + stat($dirname);
|
| + if (-e _) { next; }
|
| +
|
| + mkdir($dirname)
|
| + or die("ERROR: cannot create temporary directory ".
|
| + "$dirname!\n");
|
| +
|
| + return($dirname);
|
| + }
|
| +
|
| + die("ERROR: cannot create temporary directory in $tmp_dir!\n");
|
| +}
|
| +
|
| +
|
| +#
|
| +# read_info_file(info_filename)
|
| +#
|
| +# Read in the contents of the .info file specified by INFO_FILENAME. Data will
|
| +# be returned as a reference to a hash containing the following mappings:
|
| +#
|
| +# %result: for each filename found in file -> \%data
|
| +#
|
| +# %data: "test" -> \%testdata
|
| +# "sum" -> \%sumcount
|
| +# "func" -> \%funcdata
|
| +# "found" -> $lines_found (number of instrumented lines found in file)
|
| +# "hit" -> $lines_hit (number of executed lines in file)
|
| +# "check" -> \%checkdata
|
| +# "testfnc" -> \%testfncdata
|
| +# "sumfnc" -> \%sumfnccount
|
| +#
|
| +# %testdata : name of test affecting this file -> \%testcount
|
| +# %testfncdata: name of test affecting this file -> \%testfnccount
|
| +#
|
| +# %testcount : line number -> execution count for a single test
|
| +# %testfnccount: function name -> execution count for a single test
|
| +# %sumcount : line number -> execution count for all tests
|
| +# %sumfnccount : function name -> execution count for all tests
|
| +# %funcdata : function name -> line number
|
| +# %checkdata : line number -> checksum of source code line
|
| +#
|
| +# Note that .info file sections referring to the same file and test name
|
| +# will automatically be combined by adding all execution counts.
|
| +#
|
| +# Note that if INFO_FILENAME ends with ".gz", it is assumed that the file
|
| +# is compressed using GZIP. If available, GUNZIP will be used to decompress
|
| +# this file.
|
| +#
|
| +# Die on error.
|
| +#
|
| +
|
| +sub read_info_file($)
|
| +{
|
| + my $tracefile = $_[0]; # Name of tracefile
|
| + my %result; # Resulting hash: file -> data
|
| + my $data; # Data handle for current entry
|
| + my $testdata; # " "
|
| + my $testcount; # " "
|
| + my $sumcount; # " "
|
| + my $funcdata; # " "
|
| + my $checkdata; # " "
|
| + my $testfncdata;
|
| + my $testfnccount;
|
| + my $sumfnccount;
|
| + my $line; # Current line read from .info file
|
| + my $testname; # Current test name
|
| + my $filename; # Current filename
|
| + my $hitcount; # Count for lines hit
|
| + my $count; # Execution count of current line
|
| + my $negative; # If set, warn about negative counts
|
| + my $changed_testname; # If set, warn about changed testname
|
| + my $line_checksum; # Checksum of current line
|
| + local *INFO_HANDLE; # Filehandle for .info file
|
| +
|
| + info("Reading tracefile $tracefile\n");
|
| +
|
| + # Check if file exists and is readable
|
| + stat($_[0]);
|
| + if (!(-r _))
|
| + {
|
| + die("ERROR: cannot read file $_[0]!\n");
|
| + }
|
| +
|
| + # Check if this is really a plain file
|
| + if (!(-f _))
|
| + {
|
| + die("ERROR: not a plain file: $_[0]!\n");
|
| + }
|
| +
|
| + # Check for .gz extension
|
| + if ($_[0] =~ /\.gz$/)
|
| + {
|
| + # Check for availability of GZIP tool
|
| + system_no_output(1, "gunzip" ,"-h")
|
| + and die("ERROR: gunzip command not available!\n");
|
| +
|
| + # Check integrity of compressed file
|
| + system_no_output(1, "gunzip", "-t", $_[0])
|
| + and die("ERROR: integrity check failed for ".
|
| + "compressed file $_[0]!\n");
|
| +
|
| + # Open compressed file
|
| + open(INFO_HANDLE, "gunzip -c $_[0]|")
|
| + or die("ERROR: cannot start gunzip to decompress ".
|
| + "file $_[0]!\n");
|
| + }
|
| + else
|
| + {
|
| + # Open decompressed file
|
| + open(INFO_HANDLE, $_[0])
|
| + or die("ERROR: cannot read file $_[0]!\n");
|
| + }
|
| +
|
| + $testname = "";
|
| + while (<INFO_HANDLE>)
|
| + {
|
| + chomp($_);
|
| + $line = $_;
|
| +
|
| + # Switch statement
|
| + foreach ($line)
|
| + {
|
| + /^TN:([^,]*)/ && do
|
| + {
|
| + # Test name information found
|
| + $testname = defined($1) ? $1 : "";
|
| + if ($testname =~ s/\W/_/g)
|
| + {
|
| + $changed_testname = 1;
|
| + }
|
| + last;
|
| + };
|
| +
|
| + /^[SK]F:(.*)/ && do
|
| + {
|
| + # Filename information found
|
| + # Retrieve data for new entry
|
| + $filename = $1;
|
| +
|
| + $data = $result{$filename};
|
| + ($testdata, $sumcount, $funcdata, $checkdata,
|
| + $testfncdata, $sumfnccount) =
|
| + get_info_entry($data);
|
| +
|
| + if (defined($testname))
|
| + {
|
| + $testcount = $testdata->{$testname};
|
| + $testfnccount = $testfncdata->{$testname};
|
| + }
|
| + else
|
| + {
|
| + $testcount = {};
|
| + $testfnccount = {};
|
| + }
|
| + last;
|
| + };
|
| +
|
| + /^DA:(\d+),(-?\d+)(,[^,\s]+)?/ && do
|
| + {
|
| + # Fix negative counts
|
| + $count = $2 < 0 ? 0 : $2;
|
| + if ($2 < 0)
|
| + {
|
| + $negative = 1;
|
| + }
|
| + # Execution count found, add to structure
|
| + # Add summary counts
|
| + $sumcount->{$1} += $count;
|
| +
|
| + # Add test-specific counts
|
| + if (defined($testname))
|
| + {
|
| + $testcount->{$1} += $count;
|
| + }
|
| +
|
| + # Store line checksum if available
|
| + if (defined($3))
|
| + {
|
| + $line_checksum = substr($3, 1);
|
| +
|
| + # Does it match a previous definition
|
| + if (defined($checkdata->{$1}) &&
|
| + ($checkdata->{$1} ne
|
| + $line_checksum))
|
| + {
|
| + die("ERROR: checksum mismatch ".
|
| + "at $filename:$1\n");
|
| + }
|
| +
|
| + $checkdata->{$1} = $line_checksum;
|
| + }
|
| + last;
|
| + };
|
| +
|
| + /^FN:(\d+),([^,]+)/ && do
|
| + {
|
| + # Function data found, add to structure
|
| + $funcdata->{$2} = $1;
|
| +
|
| + # Also initialize function call data
|
| + if (!defined($sumfnccount->{$2})) {
|
| + $sumfnccount->{$2} = 0;
|
| + }
|
| + if (defined($testname))
|
| + {
|
| + if (!defined($testfnccount->{$2})) {
|
| + $testfnccount->{$2} = 0;
|
| + }
|
| + }
|
| + last;
|
| + };
|
| +
|
| + /^FNDA:(\d+),([^,]+)/ && do
|
| + {
|
| + # Function call count found, add to structure
|
| + # Add summary counts
|
| + $sumfnccount->{$2} += $1;
|
| +
|
| + # Add test-specific counts
|
| + if (defined($testname))
|
| + {
|
| + $testfnccount->{$2} += $1;
|
| + }
|
| + last;
|
| + };
|
| + /^end_of_record/ && do
|
| + {
|
| + # Found end of section marker
|
| + if ($filename)
|
| + {
|
| + # Store current section data
|
| + if (defined($testname))
|
| + {
|
| + $testdata->{$testname} =
|
| + $testcount;
|
| + $testfncdata->{$testname} =
|
| + $testfnccount;
|
| + }
|
| +
|
| + set_info_entry($data, $testdata,
|
| + $sumcount, $funcdata,
|
| + $checkdata, $testfncdata,
|
| + $sumfnccount);
|
| + $result{$filename} = $data;
|
| + last;
|
| + }
|
| + };
|
| +
|
| + # default
|
| + last;
|
| + }
|
| + }
|
| + close(INFO_HANDLE);
|
| +
|
| + # Calculate hit and found values for lines and functions of each file
|
| + foreach $filename (keys(%result))
|
| + {
|
| + $data = $result{$filename};
|
| +
|
| + ($testdata, $sumcount, undef, undef, $testfncdata,
|
| + $sumfnccount) = get_info_entry($data);
|
| +
|
| + # Filter out empty files
|
| + if (scalar(keys(%{$sumcount})) == 0)
|
| + {
|
| + delete($result{$filename});
|
| + next;
|
| + }
|
| + # Filter out empty test cases
|
| + foreach $testname (keys(%{$testdata}))
|
| + {
|
| + if (!defined($testdata->{$testname}) ||
|
| + scalar(keys(%{$testdata->{$testname}})) == 0)
|
| + {
|
| + delete($testdata->{$testname});
|
| + delete($testfncdata->{$testname});
|
| + }
|
| + }
|
| +
|
| + $data->{"found"} = scalar(keys(%{$sumcount}));
|
| + $hitcount = 0;
|
| +
|
| + foreach (keys(%{$sumcount}))
|
| + {
|
| + if ($sumcount->{$_} > 0) { $hitcount++; }
|
| + }
|
| +
|
| + $data->{"hit"} = $hitcount;
|
| +
|
| + # Get found/hit values for function call data
|
| + $data->{"f_found"} = scalar(keys(%{$sumfnccount}));
|
| + $hitcount = 0;
|
| +
|
| + foreach (keys(%{$sumfnccount})) {
|
| + if ($sumfnccount->{$_} > 0) {
|
| + $hitcount++;
|
| + }
|
| + }
|
| + $data->{"f_hit"} = $hitcount;
|
| + }
|
| +
|
| + if (scalar(keys(%result)) == 0)
|
| + {
|
| + die("ERROR: no valid records found in tracefile $tracefile\n");
|
| + }
|
| + if ($negative)
|
| + {
|
| + warn("WARNING: negative counts found in tracefile ".
|
| + "$tracefile\n");
|
| + }
|
| + if ($changed_testname)
|
| + {
|
| + warn("WARNING: invalid characters removed from testname in ".
|
| + "tracefile $tracefile\n");
|
| + }
|
| +
|
| + return(\%result);
|
| +}
|
| +
|
| +
|
| +#
|
| +# get_info_entry(hash_ref)
|
| +#
|
| +# Retrieve data from an entry of the structure generated by read_info_file().
|
| +# Return a list of references to hashes:
|
| +# (test data hash ref, sum count hash ref, funcdata hash ref, checkdata hash
|
| +# ref, testfncdata hash ref, sumfnccount hash ref, lines found, lines hit,
|
| +# functions found, functions hit)
|
| +#
|
| +
|
| +sub get_info_entry($)
|
| +{
|
| + my $testdata_ref = $_[0]->{"test"};
|
| + my $sumcount_ref = $_[0]->{"sum"};
|
| + my $funcdata_ref = $_[0]->{"func"};
|
| + my $checkdata_ref = $_[0]->{"check"};
|
| + my $testfncdata = $_[0]->{"testfnc"};
|
| + my $sumfnccount = $_[0]->{"sumfnc"};
|
| + my $lines_found = $_[0]->{"found"};
|
| + my $lines_hit = $_[0]->{"hit"};
|
| + my $f_found = $_[0]->{"f_found"};
|
| + my $f_hit = $_[0]->{"f_hit"};
|
| +
|
| + return ($testdata_ref, $sumcount_ref, $funcdata_ref, $checkdata_ref,
|
| + $testfncdata, $sumfnccount, $lines_found, $lines_hit,
|
| + $f_found, $f_hit);
|
| +}
|
| +
|
| +
|
| +#
|
| +# set_info_entry(hash_ref, testdata_ref, sumcount_ref, funcdata_ref,
|
| +# checkdata_ref, testfncdata_ref, sumfcncount_ref[,lines_found,
|
| +# lines_hit, f_found, f_hit])
|
| +#
|
| +# Update the hash referenced by HASH_REF with the provided data references.
|
| +#
|
| +
|
| +sub set_info_entry($$$$$$$;$$$$)
|
| +{
|
| + my $data_ref = $_[0];
|
| +
|
| + $data_ref->{"test"} = $_[1];
|
| + $data_ref->{"sum"} = $_[2];
|
| + $data_ref->{"func"} = $_[3];
|
| + $data_ref->{"check"} = $_[4];
|
| + $data_ref->{"testfnc"} = $_[5];
|
| + $data_ref->{"sumfnc"} = $_[6];
|
| +
|
| + if (defined($_[7])) { $data_ref->{"found"} = $_[7]; }
|
| + if (defined($_[8])) { $data_ref->{"hit"} = $_[8]; }
|
| + if (defined($_[9])) { $data_ref->{"f_found"} = $_[9]; }
|
| + if (defined($_[10])) { $data_ref->{"f_hit"} = $_[10]; }
|
| +}
|
| +
|
| +
|
| +#
|
| +# add_counts(data1_ref, data2_ref)
|
| +#
|
| +# DATA1_REF and DATA2_REF are references to hashes containing a mapping
|
| +#
|
| +# line number -> execution count
|
| +#
|
| +# Return a list (RESULT_REF, LINES_FOUND, LINES_HIT) where RESULT_REF
|
| +# is a reference to a hash containing the combined mapping in which
|
| +# execution counts are added.
|
| +#
|
| +
|
| +sub add_counts($$)
|
| +{
|
| + my %data1 = %{$_[0]}; # Hash 1
|
| + my %data2 = %{$_[1]}; # Hash 2
|
| + my %result; # Resulting hash
|
| + my $line; # Current line iteration scalar
|
| + my $data1_count; # Count of line in hash1
|
| + my $data2_count; # Count of line in hash2
|
| + my $found = 0; # Total number of lines found
|
| + my $hit = 0; # Number of lines with a count > 0
|
| +
|
| + foreach $line (keys(%data1))
|
| + {
|
| + $data1_count = $data1{$line};
|
| + $data2_count = $data2{$line};
|
| +
|
| + # Add counts if present in both hashes
|
| + if (defined($data2_count)) { $data1_count += $data2_count; }
|
| +
|
| + # Store sum in %result
|
| + $result{$line} = $data1_count;
|
| +
|
| + $found++;
|
| + if ($data1_count > 0) { $hit++; }
|
| + }
|
| +
|
| + # Add lines unique to data2
|
| + foreach $line (keys(%data2))
|
| + {
|
| + # Skip lines already in data1
|
| + if (defined($data1{$line})) { next; }
|
| +
|
| + # Copy count from data2
|
| + $result{$line} = $data2{$line};
|
| +
|
| + $found++;
|
| + if ($result{$line} > 0) { $hit++; }
|
| + }
|
| +
|
| + return (\%result, $found, $hit);
|
| +}
|
| +
|
| +
|
| +#
|
| +# merge_checksums(ref1, ref2, filename)
|
| +#
|
| +# REF1 and REF2 are references to hashes containing a mapping
|
| +#
|
| +# line number -> checksum
|
| +#
|
| +# Merge checksum lists defined in REF1 and REF2 and return reference to
|
| +# resulting hash. Die if a checksum for a line is defined in both hashes
|
| +# but does not match.
|
| +#
|
| +
|
| +sub merge_checksums($$$)
|
| +{
|
| + my $ref1 = $_[0];
|
| + my $ref2 = $_[1];
|
| + my $filename = $_[2];
|
| + my %result;
|
| + my $line;
|
| +
|
| + foreach $line (keys(%{$ref1}))
|
| + {
|
| + if (defined($ref2->{$line}) &&
|
| + ($ref1->{$line} ne $ref2->{$line}))
|
| + {
|
| + die("ERROR: checksum mismatch at $filename:$line\n");
|
| + }
|
| + $result{$line} = $ref1->{$line};
|
| + }
|
| +
|
| + foreach $line (keys(%{$ref2}))
|
| + {
|
| + $result{$line} = $ref2->{$line};
|
| + }
|
| +
|
| + return \%result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# merge_func_data(funcdata1, funcdata2, filename)
|
| +#
|
| +
|
| +sub merge_func_data($$$)
|
| +{
|
| + my ($funcdata1, $funcdata2, $filename) = @_;
|
| + my %result;
|
| + my $func;
|
| +
|
| + %result = %{$funcdata1};
|
| +
|
| + foreach $func (keys(%{$funcdata2})) {
|
| + my $line1 = $result{$func};
|
| + my $line2 = $funcdata2->{$func};
|
| +
|
| + if (defined($line1) && ($line1 != $line2)) {
|
| + warn("WARNING: function data mismatch at ".
|
| + "$filename:$line2\n");
|
| + next;
|
| + }
|
| + $result{$func} = $line2;
|
| + }
|
| +
|
| + return \%result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# add_fnccount(fnccount1, fnccount2)
|
| +#
|
| +# Add function call count data. Return list (fnccount_added, f_found, f_hit)
|
| +#
|
| +
|
| +sub add_fnccount($$)
|
| +{
|
| + my ($fnccount1, $fnccount2) = @_;
|
| + my %result;
|
| + my $f_found;
|
| + my $f_hit;
|
| + my $function;
|
| +
|
| + %result = %{$fnccount1};
|
| + foreach $function (keys(%{$fnccount2})) {
|
| + $result{$function} += $fnccount2->{$function};
|
| + }
|
| + $f_found = scalar(keys(%result));
|
| + $f_hit = 0;
|
| + foreach $function (keys(%result)) {
|
| + if ($result{$function} > 0) {
|
| + $f_hit++;
|
| + }
|
| + }
|
| +
|
| + return (\%result, $f_found, $f_hit);
|
| +}
|
| +
|
| +#
|
| +# add_testfncdata(testfncdata1, testfncdata2)
|
| +#
|
| +# Add function call count data for several tests. Return reference to
|
| +# added_testfncdata.
|
| +#
|
| +
|
| +sub add_testfncdata($$)
|
| +{
|
| + my ($testfncdata1, $testfncdata2) = @_;
|
| + my %result;
|
| + my $testname;
|
| +
|
| + foreach $testname (keys(%{$testfncdata1})) {
|
| + if (defined($testfncdata2->{$testname})) {
|
| + my $fnccount;
|
| +
|
| + # Function call count data for this testname exists
|
| + # in both data sets: merge
|
| + ($fnccount) = add_fnccount(
|
| + $testfncdata1->{$testname},
|
| + $testfncdata2->{$testname});
|
| + $result{$testname} = $fnccount;
|
| + next;
|
| + }
|
| + # Function call count data for this testname is unique to
|
| + # data set 1: copy
|
| + $result{$testname} = $testfncdata1->{$testname};
|
| + }
|
| +
|
| + # Add count data for testnames unique to data set 2
|
| + foreach $testname (keys(%{$testfncdata2})) {
|
| + if (!defined($result{$testname})) {
|
| + $result{$testname} = $testfncdata2->{$testname};
|
| + }
|
| + }
|
| + return \%result;
|
| +}
|
| +
|
| +#
|
| +# combine_info_entries(entry_ref1, entry_ref2, filename)
|
| +#
|
| +# Combine .info data entry hashes referenced by ENTRY_REF1 and ENTRY_REF2.
|
| +# Return reference to resulting hash.
|
| +#
|
| +
|
| +sub combine_info_entries($$$)
|
| +{
|
| + my $entry1 = $_[0]; # Reference to hash containing first entry
|
| + my $testdata1;
|
| + my $sumcount1;
|
| + my $funcdata1;
|
| + my $checkdata1;
|
| + my $testfncdata1;
|
| + my $sumfnccount1;
|
| +
|
| + my $entry2 = $_[1]; # Reference to hash containing second entry
|
| + my $testdata2;
|
| + my $sumcount2;
|
| + my $funcdata2;
|
| + my $checkdata2;
|
| + my $testfncdata2;
|
| + my $sumfnccount2;
|
| +
|
| + my %result; # Hash containing combined entry
|
| + my %result_testdata;
|
| + my $result_sumcount = {};
|
| + my $result_funcdata;
|
| + my $result_testfncdata;
|
| + my $result_sumfnccount;
|
| + my $lines_found;
|
| + my $lines_hit;
|
| + my $f_found;
|
| + my $f_hit;
|
| +
|
| + my $testname;
|
| + my $filename = $_[2];
|
| +
|
| + # Retrieve data
|
| + ($testdata1, $sumcount1, $funcdata1, $checkdata1, $testfncdata1,
|
| + $sumfnccount1) = get_info_entry($entry1);
|
| + ($testdata2, $sumcount2, $funcdata2, $checkdata2, $testfncdata2,
|
| + $sumfnccount2) = get_info_entry($entry2);
|
| +
|
| + # Merge checksums
|
| + $checkdata1 = merge_checksums($checkdata1, $checkdata2, $filename);
|
| +
|
| + # Combine funcdata
|
| + $result_funcdata = merge_func_data($funcdata1, $funcdata2, $filename);
|
| +
|
| + # Combine function call count data
|
| + $result_testfncdata = add_testfncdata($testfncdata1, $testfncdata2);
|
| + ($result_sumfnccount, $f_found, $f_hit) =
|
| + add_fnccount($sumfnccount1, $sumfnccount2);
|
| +
|
| + # Combine testdata
|
| + foreach $testname (keys(%{$testdata1}))
|
| + {
|
| + if (defined($testdata2->{$testname}))
|
| + {
|
| + # testname is present in both entries, requires
|
| + # combination
|
| + ($result_testdata{$testname}) =
|
| + add_counts($testdata1->{$testname},
|
| + $testdata2->{$testname});
|
| + }
|
| + else
|
| + {
|
| + # testname only present in entry1, add to result
|
| + $result_testdata{$testname} = $testdata1->{$testname};
|
| + }
|
| +
|
| + # update sum count hash
|
| + ($result_sumcount, $lines_found, $lines_hit) =
|
| + add_counts($result_sumcount,
|
| + $result_testdata{$testname});
|
| + }
|
| +
|
| + foreach $testname (keys(%{$testdata2}))
|
| + {
|
| + # Skip testnames already covered by previous iteration
|
| + if (defined($testdata1->{$testname})) { next; }
|
| +
|
| + # testname only present in entry2, add to result hash
|
| + $result_testdata{$testname} = $testdata2->{$testname};
|
| +
|
| + # update sum count hash
|
| + ($result_sumcount, $lines_found, $lines_hit) =
|
| + add_counts($result_sumcount,
|
| + $result_testdata{$testname});
|
| + }
|
| +
|
| + # Calculate resulting sumcount
|
| +
|
| + # Store result
|
| + set_info_entry(\%result, \%result_testdata, $result_sumcount,
|
| + $result_funcdata, $checkdata1, $result_testfncdata,
|
| + $result_sumfnccount, $lines_found, $lines_hit,
|
| + $f_found, $f_hit);
|
| +
|
| + return(\%result);
|
| +}
|
| +
|
| +
|
| +#
|
| +# combine_info_files(info_ref1, info_ref2)
|
| +#
|
| +# Combine .info data in hashes referenced by INFO_REF1 and INFO_REF2. Return
|
| +# reference to resulting hash.
|
| +#
|
| +
|
| +sub combine_info_files($$)
|
| +{
|
| + my %hash1 = %{$_[0]};
|
| + my %hash2 = %{$_[1]};
|
| + my $filename;
|
| +
|
| + foreach $filename (keys(%hash2))
|
| + {
|
| + if ($hash1{$filename})
|
| + {
|
| + # Entry already exists in hash1, combine them
|
| + $hash1{$filename} =
|
| + combine_info_entries($hash1{$filename},
|
| + $hash2{$filename},
|
| + $filename);
|
| + }
|
| + else
|
| + {
|
| + # Entry is unique in both hashes, simply add to
|
| + # resulting hash
|
| + $hash1{$filename} = $hash2{$filename};
|
| + }
|
| + }
|
| +
|
| + return(\%hash1);
|
| +}
|
| +
|
| +
|
| +#
|
| +# add_traces()
|
| +#
|
| +
|
| +sub add_traces()
|
| +{
|
| + my $total_trace;
|
| + my $current_trace;
|
| + my $tracefile;
|
| + local *INFO_HANDLE;
|
| +
|
| + info("Combining tracefiles.\n");
|
| +
|
| + foreach $tracefile (@add_tracefile)
|
| + {
|
| + $current_trace = read_info_file($tracefile);
|
| + if ($total_trace)
|
| + {
|
| + $total_trace = combine_info_files($total_trace,
|
| + $current_trace);
|
| + }
|
| + else
|
| + {
|
| + $total_trace = $current_trace;
|
| + }
|
| + }
|
| +
|
| + # Write combined data
|
| + if ($to_file)
|
| + {
|
| + info("Writing data to $output_filename\n");
|
| + open(INFO_HANDLE, ">$output_filename")
|
| + or die("ERROR: cannot write to $output_filename!\n");
|
| + write_info_file(*INFO_HANDLE, $total_trace);
|
| + close(*INFO_HANDLE);
|
| + }
|
| + else
|
| + {
|
| + write_info_file(*STDOUT, $total_trace);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# write_info_file(filehandle, data)
|
| +#
|
| +
|
| +sub write_info_file(*$)
|
| +{
|
| + local *INFO_HANDLE = $_[0];
|
| + my %data = %{$_[1]};
|
| + my $source_file;
|
| + my $entry;
|
| + my $testdata;
|
| + my $sumcount;
|
| + my $funcdata;
|
| + my $checkdata;
|
| + my $testfncdata;
|
| + my $sumfnccount;
|
| + my $testname;
|
| + my $line;
|
| + my $func;
|
| + my $testcount;
|
| + my $testfnccount;
|
| + my $found;
|
| + my $hit;
|
| + my $f_found;
|
| + my $f_hit;
|
| +
|
| + foreach $source_file (keys(%data))
|
| + {
|
| + $entry = $data{$source_file};
|
| + ($testdata, $sumcount, $funcdata, $checkdata, $testfncdata,
|
| + $sumfnccount) = get_info_entry($entry);
|
| + foreach $testname (keys(%{$testdata}))
|
| + {
|
| + $testcount = $testdata->{$testname};
|
| + $testfnccount = $testfncdata->{$testname};
|
| + $found = 0;
|
| + $hit = 0;
|
| +
|
| + print(INFO_HANDLE "TN:$testname\n");
|
| + print(INFO_HANDLE "SF:$source_file\n");
|
| +
|
| + # Write function related data
|
| + foreach $func (
|
| + sort({$funcdata->{$a} <=> $funcdata->{$b}}
|
| + keys(%{$funcdata})))
|
| + {
|
| + print(INFO_HANDLE "FN:".$funcdata->{$func}.
|
| + ",$func\n");
|
| + }
|
| + foreach $func (keys(%{$testfnccount})) {
|
| + print(INFO_HANDLE "FNDA:".
|
| + $testfnccount->{$func}.
|
| + ",$func\n");
|
| + }
|
| + ($f_found, $f_hit) =
|
| + get_func_found_and_hit($testfnccount);
|
| + print(INFO_HANDLE "FNF:$f_found\n");
|
| + print(INFO_HANDLE "FNH:$f_hit\n");
|
| +
|
| + # Write line related data
|
| + foreach $line (sort({$a <=> $b} keys(%{$testcount})))
|
| + {
|
| + print(INFO_HANDLE "DA:$line,".
|
| + $testcount->{$line}.
|
| + (defined($checkdata->{$line}) &&
|
| + $checksum ?
|
| + ",".$checkdata->{$line} : "")."\n");
|
| + $found++;
|
| + if ($testcount->{$line} > 0)
|
| + {
|
| + $hit++;
|
| + }
|
| +
|
| + }
|
| + print(INFO_HANDLE "LF:$found\n");
|
| + print(INFO_HANDLE "LH:$hit\n");
|
| + print(INFO_HANDLE "end_of_record\n");
|
| + }
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# transform_pattern(pattern)
|
| +#
|
| +# Transform shell wildcard expression to equivalent PERL regular expression.
|
| +# Return transformed pattern.
|
| +#
|
| +
|
| +sub transform_pattern($)
|
| +{
|
| + my $pattern = $_[0];
|
| +
|
| + # Escape special chars
|
| +
|
| + $pattern =~ s/\\/\\\\/g;
|
| + $pattern =~ s/\//\\\//g;
|
| + $pattern =~ s/\^/\\\^/g;
|
| + $pattern =~ s/\$/\\\$/g;
|
| + $pattern =~ s/\(/\\\(/g;
|
| + $pattern =~ s/\)/\\\)/g;
|
| + $pattern =~ s/\[/\\\[/g;
|
| + $pattern =~ s/\]/\\\]/g;
|
| + $pattern =~ s/\{/\\\{/g;
|
| + $pattern =~ s/\}/\\\}/g;
|
| + $pattern =~ s/\./\\\./g;
|
| + $pattern =~ s/\,/\\\,/g;
|
| + $pattern =~ s/\|/\\\|/g;
|
| + $pattern =~ s/\+/\\\+/g;
|
| + $pattern =~ s/\!/\\\!/g;
|
| +
|
| + # Transform ? => (.) and * => (.*)
|
| +
|
| + $pattern =~ s/\*/\(\.\*\)/g;
|
| + $pattern =~ s/\?/\(\.\)/g;
|
| +
|
| + return $pattern;
|
| +}
|
| +
|
| +
|
| +#
|
| +# extract()
|
| +#
|
| +
|
| +sub extract()
|
| +{
|
| + my $data = read_info_file($extract);
|
| + my $filename;
|
| + my $keep;
|
| + my $pattern;
|
| + my @pattern_list;
|
| + my $extracted = 0;
|
| + local *INFO_HANDLE;
|
| +
|
| + # Need perlreg expressions instead of shell pattern
|
| + @pattern_list = map({ transform_pattern($_); } @ARGV);
|
| +
|
| + # Filter out files which do not match any pattern
|
| + foreach $filename (sort(keys(%{$data})))
|
| + {
|
| + $keep = 0;
|
| +
|
| + foreach $pattern (@pattern_list)
|
| + {
|
| + $keep ||= ($filename =~ (/^$pattern$/));
|
| + }
|
| +
|
| +
|
| + if (!$keep)
|
| + {
|
| + delete($data->{$filename});
|
| + }
|
| + else
|
| + {
|
| + info("Extracting $filename\n"),
|
| + $extracted++;
|
| + }
|
| + }
|
| +
|
| + # Write extracted data
|
| + if ($to_file)
|
| + {
|
| + info("Extracted $extracted files\n");
|
| + info("Writing data to $output_filename\n");
|
| + open(INFO_HANDLE, ">$output_filename")
|
| + or die("ERROR: cannot write to $output_filename!\n");
|
| + write_info_file(*INFO_HANDLE, $data);
|
| + close(*INFO_HANDLE);
|
| + }
|
| + else
|
| + {
|
| + write_info_file(*STDOUT, $data);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# remove()
|
| +#
|
| +
|
| +sub remove()
|
| +{
|
| + my $data = read_info_file($remove);
|
| + my $filename;
|
| + my $match_found;
|
| + my $pattern;
|
| + my @pattern_list;
|
| + my $removed = 0;
|
| + local *INFO_HANDLE;
|
| +
|
| + # Need perlreg expressions instead of shell pattern
|
| + @pattern_list = map({ transform_pattern($_); } @ARGV);
|
| +
|
| + # Filter out files that match the pattern
|
| + foreach $filename (sort(keys(%{$data})))
|
| + {
|
| + $match_found = 0;
|
| +
|
| + foreach $pattern (@pattern_list)
|
| + {
|
| + $match_found ||= ($filename =~ (/$pattern$/));
|
| + }
|
| +
|
| +
|
| + if ($match_found)
|
| + {
|
| + delete($data->{$filename});
|
| + info("Removing $filename\n"),
|
| + $removed++;
|
| + }
|
| + }
|
| +
|
| + # Write data
|
| + if ($to_file)
|
| + {
|
| + info("Deleted $removed files\n");
|
| + info("Writing data to $output_filename\n");
|
| + open(INFO_HANDLE, ">$output_filename")
|
| + or die("ERROR: cannot write to $output_filename!\n");
|
| + write_info_file(*INFO_HANDLE, $data);
|
| + close(*INFO_HANDLE);
|
| + }
|
| + else
|
| + {
|
| + write_info_file(*STDOUT, $data);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# list()
|
| +#
|
| +
|
| +sub list()
|
| +{
|
| + my $data = read_info_file($list);
|
| + my $filename;
|
| + my $found;
|
| + my $hit;
|
| + my $entry;
|
| +
|
| + info("Listing contents of $list:\n");
|
| +
|
| + # List all files
|
| + foreach $filename (sort(keys(%{$data})))
|
| + {
|
| + $entry = $data->{$filename};
|
| + (undef, undef, undef, undef, undef, undef, $found, $hit) =
|
| + get_info_entry($entry);
|
| + printf("$filename: $hit of $found lines hit\n");
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# get_common_filename(filename1, filename2)
|
| +#
|
| +# Check for filename components which are common to FILENAME1 and FILENAME2.
|
| +# Upon success, return
|
| +#
|
| +# (common, path1, path2)
|
| +#
|
| +# or 'undef' in case there are no such parts.
|
| +#
|
| +
|
| +sub get_common_filename($$)
|
| +{
|
| + my @list1 = split("/", $_[0]);
|
| + my @list2 = split("/", $_[1]);
|
| + my @result;
|
| +
|
| + # Work in reverse order, i.e. beginning with the filename itself
|
| + while (@list1 && @list2 && ($list1[$#list1] eq $list2[$#list2]))
|
| + {
|
| + unshift(@result, pop(@list1));
|
| + pop(@list2);
|
| + }
|
| +
|
| + # Did we find any similarities?
|
| + if (scalar(@result) > 0)
|
| + {
|
| + return (join("/", @result), join("/", @list1),
|
| + join("/", @list2));
|
| + }
|
| + else
|
| + {
|
| + return undef;
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# strip_directories($path, $depth)
|
| +#
|
| +# Remove DEPTH leading directory levels from PATH.
|
| +#
|
| +
|
| +sub strip_directories($$)
|
| +{
|
| + my $filename = $_[0];
|
| + my $depth = $_[1];
|
| + my $i;
|
| +
|
| + if (!defined($depth) || ($depth < 1))
|
| + {
|
| + return $filename;
|
| + }
|
| + for ($i = 0; $i < $depth; $i++)
|
| + {
|
| + $filename =~ s/^[^\/]*\/+(.*)$/$1/;
|
| + }
|
| + return $filename;
|
| +}
|
| +
|
| +
|
| +#
|
| +# read_diff(filename)
|
| +#
|
| +# Read diff output from FILENAME to memory. The diff file has to follow the
|
| +# format generated by 'diff -u'. Returns a list of hash references:
|
| +#
|
| +# (mapping, path mapping)
|
| +#
|
| +# mapping: filename -> reference to line hash
|
| +# line hash: line number in new file -> corresponding line number in old file
|
| +#
|
| +# path mapping: filename -> old filename
|
| +#
|
| +# Die in case of error.
|
| +#
|
| +
|
| +sub read_diff($)
|
| +{
|
| + my $diff_file = $_[0]; # Name of diff file
|
| + my %diff; # Resulting mapping filename -> line hash
|
| + my %paths; # Resulting mapping old path -> new path
|
| + my $mapping; # Reference to current line hash
|
| + my $line; # Contents of current line
|
| + my $num_old; # Current line number in old file
|
| + my $num_new; # Current line number in new file
|
| + my $file_old; # Name of old file in diff section
|
| + my $file_new; # Name of new file in diff section
|
| + my $filename; # Name of common filename of diff section
|
| + my $in_block = 0; # Non-zero while we are inside a diff block
|
| + local *HANDLE; # File handle for reading the diff file
|
| +
|
| + info("Reading diff $diff_file\n");
|
| +
|
| + # Check if file exists and is readable
|
| + stat($diff_file);
|
| + if (!(-r _))
|
| + {
|
| + die("ERROR: cannot read file $diff_file!\n");
|
| + }
|
| +
|
| + # Check if this is really a plain file
|
| + if (!(-f _))
|
| + {
|
| + die("ERROR: not a plain file: $diff_file!\n");
|
| + }
|
| +
|
| + # Check for .gz extension
|
| + if ($diff_file =~ /\.gz$/)
|
| + {
|
| + # Check for availability of GZIP tool
|
| + system_no_output(1, "gunzip", "-h")
|
| + and die("ERROR: gunzip command not available!\n");
|
| +
|
| + # Check integrity of compressed file
|
| + system_no_output(1, "gunzip", "-t", $diff_file)
|
| + and die("ERROR: integrity check failed for ".
|
| + "compressed file $diff_file!\n");
|
| +
|
| + # Open compressed file
|
| + open(HANDLE, "gunzip -c $diff_file|")
|
| + or die("ERROR: cannot start gunzip to decompress ".
|
| + "file $_[0]!\n");
|
| + }
|
| + else
|
| + {
|
| + # Open decompressed file
|
| + open(HANDLE, $diff_file)
|
| + or die("ERROR: cannot read file $_[0]!\n");
|
| + }
|
| +
|
| + # Parse diff file line by line
|
| + while (<HANDLE>)
|
| + {
|
| + chomp($_);
|
| + $line = $_;
|
| +
|
| + foreach ($line)
|
| + {
|
| + # Filename of old file:
|
| + # --- <filename> <date>
|
| + /^--- (\S+)/ && do
|
| + {
|
| + $file_old = strip_directories($1, $strip);
|
| + last;
|
| + };
|
| + # Filename of new file:
|
| + # +++ <filename> <date>
|
| + /^\+\+\+ (\S+)/ && do
|
| + {
|
| + # Add last file to resulting hash
|
| + if ($filename)
|
| + {
|
| + my %new_hash;
|
| + $diff{$filename} = $mapping;
|
| + $mapping = \%new_hash;
|
| + }
|
| + $file_new = strip_directories($1, $strip);
|
| + $filename = $file_old;
|
| + $paths{$filename} = $file_new;
|
| + $num_old = 1;
|
| + $num_new = 1;
|
| + last;
|
| + };
|
| + # Start of diff block:
|
| + # @@ -old_start,old_num, +new_start,new_num @@
|
| + /^\@\@\s+-(\d+),(\d+)\s+\+(\d+),(\d+)\s+\@\@$/ && do
|
| + {
|
| + $in_block = 1;
|
| + while ($num_old < $1)
|
| + {
|
| + $mapping->{$num_new} = $num_old;
|
| + $num_old++;
|
| + $num_new++;
|
| + }
|
| + last;
|
| + };
|
| + # Unchanged line
|
| + # <line starts with blank>
|
| + /^ / && do
|
| + {
|
| + if ($in_block == 0)
|
| + {
|
| + last;
|
| + }
|
| + $mapping->{$num_new} = $num_old;
|
| + $num_old++;
|
| + $num_new++;
|
| + last;
|
| + };
|
| + # Line as seen in old file
|
| + # <line starts with '-'>
|
| + /^-/ && do
|
| + {
|
| + if ($in_block == 0)
|
| + {
|
| + last;
|
| + }
|
| + $num_old++;
|
| + last;
|
| + };
|
| + # Line as seen in new file
|
| + # <line starts with '+'>
|
| + /^\+/ && do
|
| + {
|
| + if ($in_block == 0)
|
| + {
|
| + last;
|
| + }
|
| + $num_new++;
|
| + last;
|
| + };
|
| + # Empty line
|
| + /^$/ && do
|
| + {
|
| + if ($in_block == 0)
|
| + {
|
| + last;
|
| + }
|
| + $mapping->{$num_new} = $num_old;
|
| + $num_old++;
|
| + $num_new++;
|
| + last;
|
| + };
|
| + }
|
| + }
|
| +
|
| + close(HANDLE);
|
| +
|
| + # Add final diff file section to resulting hash
|
| + if ($filename)
|
| + {
|
| + $diff{$filename} = $mapping;
|
| + }
|
| +
|
| + if (!%diff)
|
| + {
|
| + die("ERROR: no valid diff data found in $diff_file!\n".
|
| + "Make sure to use 'diff -u' when generating the diff ".
|
| + "file.\n");
|
| + }
|
| + return (\%diff, \%paths);
|
| +}
|
| +
|
| +
|
| +#
|
| +# apply_diff($count_data, $line_hash)
|
| +#
|
| +# Transform count data using a mapping of lines:
|
| +#
|
| +# $count_data: reference to hash: line number -> data
|
| +# $line_hash: reference to hash: line number new -> line number old
|
| +#
|
| +# Return a reference to transformed count data.
|
| +#
|
| +
|
| +sub apply_diff($$)
|
| +{
|
| + my $count_data = $_[0]; # Reference to data hash: line -> hash
|
| + my $line_hash = $_[1]; # Reference to line hash: new line -> old line
|
| + my %result; # Resulting hash
|
| + my $last_new = 0; # Last new line number found in line hash
|
| + my $last_old = 0; # Last old line number found in line hash
|
| +
|
| + # Iterate all new line numbers found in the diff
|
| + foreach (sort({$a <=> $b} keys(%{$line_hash})))
|
| + {
|
| + $last_new = $_;
|
| + $last_old = $line_hash->{$last_new};
|
| +
|
| + # Is there data associated with the corresponding old line?
|
| + if (defined($count_data->{$line_hash->{$_}}))
|
| + {
|
| + # Copy data to new hash with a new line number
|
| + $result{$_} = $count_data->{$line_hash->{$_}};
|
| + }
|
| + }
|
| + # Transform all other lines which come after the last diff entry
|
| + foreach (sort({$a <=> $b} keys(%{$count_data})))
|
| + {
|
| + if ($_ <= $last_old)
|
| + {
|
| + # Skip lines which were covered by line hash
|
| + next;
|
| + }
|
| + # Copy data to new hash with an offset
|
| + $result{$_ + ($last_new - $last_old)} = $count_data->{$_};
|
| + }
|
| +
|
| + return \%result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# get_hash_max(hash_ref)
|
| +#
|
| +# Return the highest integer key from hash.
|
| +#
|
| +
|
| +sub get_hash_max($)
|
| +{
|
| + my ($hash) = @_;
|
| + my $max;
|
| +
|
| + foreach (keys(%{$hash})) {
|
| + if (!defined($max)) {
|
| + $max = $_;
|
| + } elsif ($hash->{$_} > $max) {
|
| + $max = $_;
|
| + }
|
| + }
|
| + return $max;
|
| +}
|
| +
|
| +sub get_hash_reverse($)
|
| +{
|
| + my ($hash) = @_;
|
| + my %result;
|
| +
|
| + foreach (keys(%{$hash})) {
|
| + $result{$hash->{$_}} = $_;
|
| + }
|
| +
|
| + return \%result;
|
| +}
|
| +
|
| +#
|
| +# apply_diff_to_funcdata(funcdata, line_hash)
|
| +#
|
| +
|
| +sub apply_diff_to_funcdata($$)
|
| +{
|
| + my ($funcdata, $linedata) = @_;
|
| + my $last_new = get_hash_max($linedata);
|
| + my $last_old = $linedata->{$last_new};
|
| + my $func;
|
| + my %result;
|
| + my $line_diff = get_hash_reverse($linedata);
|
| +
|
| + foreach $func (keys(%{$funcdata})) {
|
| + my $line = $funcdata->{$func};
|
| +
|
| + if (defined($line_diff->{$line})) {
|
| + $result{$func} = $line_diff->{$line};
|
| + } elsif ($line > $last_old) {
|
| + $result{$func} = $line + $last_new - $last_old;
|
| + }
|
| + }
|
| +
|
| + return \%result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# get_line_hash($filename, $diff_data, $path_data)
|
| +#
|
| +# Find line hash in DIFF_DATA which matches FILENAME. On success, return list
|
| +# line hash. or undef in case of no match. Die if more than one line hashes in
|
| +# DIFF_DATA match.
|
| +#
|
| +
|
| +sub get_line_hash($$$)
|
| +{
|
| + my $filename = $_[0];
|
| + my $diff_data = $_[1];
|
| + my $path_data = $_[2];
|
| + my $conversion;
|
| + my $old_path;
|
| + my $new_path;
|
| + my $diff_name;
|
| + my $common;
|
| + my $old_depth;
|
| + my $new_depth;
|
| +
|
| + foreach (keys(%{$diff_data}))
|
| + {
|
| + # Try to match diff filename with filename
|
| + if ($filename =~ /^\Q$diff_path\E\/$_$/)
|
| + {
|
| + if ($diff_name)
|
| + {
|
| + # Two files match, choose the more specific one
|
| + # (the one with more path components)
|
| + $old_depth = ($diff_name =~ tr/\///);
|
| + $new_depth = (tr/\///);
|
| + if ($old_depth == $new_depth)
|
| + {
|
| + die("ERROR: diff file contains ".
|
| + "ambiguous entries for ".
|
| + "$filename\n");
|
| + }
|
| + elsif ($new_depth > $old_depth)
|
| + {
|
| + $diff_name = $_;
|
| + }
|
| + }
|
| + else
|
| + {
|
| + $diff_name = $_;
|
| + }
|
| + };
|
| + }
|
| + if ($diff_name)
|
| + {
|
| + # Get converted path
|
| + if ($filename =~ /^(.*)$diff_name$/)
|
| + {
|
| + ($common, $old_path, $new_path) =
|
| + get_common_filename($filename,
|
| + $1.$path_data->{$diff_name});
|
| + }
|
| + return ($diff_data->{$diff_name}, $old_path, $new_path);
|
| + }
|
| + else
|
| + {
|
| + return undef;
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# convert_paths(trace_data, path_conversion_data)
|
| +#
|
| +# Rename all paths in TRACE_DATA which show up in PATH_CONVERSION_DATA.
|
| +#
|
| +
|
| +sub convert_paths($$)
|
| +{
|
| + my $trace_data = $_[0];
|
| + my $path_conversion_data = $_[1];
|
| + my $filename;
|
| + my $new_path;
|
| +
|
| + if (scalar(keys(%{$path_conversion_data})) == 0)
|
| + {
|
| + info("No path conversion data available.\n");
|
| + return;
|
| + }
|
| +
|
| + # Expand path conversion list
|
| + foreach $filename (keys(%{$path_conversion_data}))
|
| + {
|
| + $new_path = $path_conversion_data->{$filename};
|
| + while (($filename =~ s/^(.*)\/[^\/]+$/$1/) &&
|
| + ($new_path =~ s/^(.*)\/[^\/]+$/$1/) &&
|
| + ($filename ne $new_path))
|
| + {
|
| + $path_conversion_data->{$filename} = $new_path;
|
| + }
|
| + }
|
| +
|
| + # Adjust paths
|
| + FILENAME: foreach $filename (keys(%{$trace_data}))
|
| + {
|
| + # Find a path in our conversion table that matches, starting
|
| + # with the longest path
|
| + foreach (sort({length($b) <=> length($a)}
|
| + keys(%{$path_conversion_data})))
|
| + {
|
| + # Is this path a prefix of our filename?
|
| + if (!($filename =~ /^$_(.*)$/))
|
| + {
|
| + next;
|
| + }
|
| + $new_path = $path_conversion_data->{$_}.$1;
|
| +
|
| + # Make sure not to overwrite an existing entry under
|
| + # that path name
|
| + if ($trace_data->{$new_path})
|
| + {
|
| + # Need to combine entries
|
| + $trace_data->{$new_path} =
|
| + combine_info_entries(
|
| + $trace_data->{$filename},
|
| + $trace_data->{$new_path},
|
| + $filename);
|
| + }
|
| + else
|
| + {
|
| + # Simply rename entry
|
| + $trace_data->{$new_path} =
|
| + $trace_data->{$filename};
|
| + }
|
| + delete($trace_data->{$filename});
|
| + next FILENAME;
|
| + }
|
| + info("No conversion available for filename $filename\n");
|
| + }
|
| +}
|
| +
|
| +#
|
| +# sub adjust_fncdata(funcdata, testfncdata, sumfnccount)
|
| +#
|
| +# Remove function call count data from testfncdata and sumfnccount which
|
| +# is no longer present in funcdata.
|
| +#
|
| +
|
| +sub adjust_fncdata($$$)
|
| +{
|
| + my ($funcdata, $testfncdata, $sumfnccount) = @_;
|
| + my $testname;
|
| + my $func;
|
| + my $f_found;
|
| + my $f_hit;
|
| +
|
| + # Remove count data in testfncdata for functions which are no longer
|
| + # in funcdata
|
| + foreach $testname (%{$testfncdata}) {
|
| + my $fnccount = $testfncdata->{$testname};
|
| +
|
| + foreach $func (%{$fnccount}) {
|
| + if (!defined($funcdata->{$func})) {
|
| + delete($fnccount->{$func});
|
| + }
|
| + }
|
| + }
|
| + # Remove count data in sumfnccount for functions which are no longer
|
| + # in funcdata
|
| + foreach $func (%{$sumfnccount}) {
|
| + if (!defined($funcdata->{$func})) {
|
| + delete($sumfnccount->{$func});
|
| + }
|
| + }
|
| +}
|
| +
|
| +#
|
| +# get_func_found_and_hit(sumfnccount)
|
| +#
|
| +# Return (f_found, f_hit) for sumfnccount
|
| +#
|
| +
|
| +sub get_func_found_and_hit($)
|
| +{
|
| + my ($sumfnccount) = @_;
|
| + my $function;
|
| + my $f_found;
|
| + my $f_hit;
|
| +
|
| + $f_found = scalar(keys(%{$sumfnccount}));
|
| + $f_hit = 0;
|
| + foreach $function (keys(%{$sumfnccount})) {
|
| + if ($sumfnccount->{$function} > 0) {
|
| + $f_hit++;
|
| + }
|
| + }
|
| + return ($f_found, $f_hit);
|
| +}
|
| +
|
| +#
|
| +# diff()
|
| +#
|
| +
|
| +sub diff()
|
| +{
|
| + my $trace_data = read_info_file($diff);
|
| + my $diff_data;
|
| + my $path_data;
|
| + my $old_path;
|
| + my $new_path;
|
| + my %path_conversion_data;
|
| + my $filename;
|
| + my $line_hash;
|
| + my $new_name;
|
| + my $entry;
|
| + my $testdata;
|
| + my $testname;
|
| + my $sumcount;
|
| + my $funcdata;
|
| + my $checkdata;
|
| + my $testfncdata;
|
| + my $sumfnccount;
|
| + my $found;
|
| + my $hit;
|
| + my $f_found;
|
| + my $f_hit;
|
| + my $converted = 0;
|
| + my $unchanged = 0;
|
| + local *INFO_HANDLE;
|
| +
|
| + ($diff_data, $path_data) = read_diff($ARGV[0]);
|
| +
|
| + foreach $filename (sort(keys(%{$trace_data})))
|
| + {
|
| + # Find a diff section corresponding to this file
|
| + ($line_hash, $old_path, $new_path) =
|
| + get_line_hash($filename, $diff_data, $path_data);
|
| + if (!$line_hash)
|
| + {
|
| + # There's no diff section for this file
|
| + $unchanged++;
|
| + next;
|
| + }
|
| + $converted++;
|
| + if ($old_path && $new_path && ($old_path ne $new_path))
|
| + {
|
| + $path_conversion_data{$old_path} = $new_path;
|
| + }
|
| + # Check for deleted files
|
| + if (scalar(keys(%{$line_hash})) == 0)
|
| + {
|
| + info("Removing $filename\n");
|
| + delete($trace_data->{$filename});
|
| + next;
|
| + }
|
| + info("Converting $filename\n");
|
| + $entry = $trace_data->{$filename};
|
| + ($testdata, $sumcount, $funcdata, $checkdata, $testfncdata,
|
| + $sumfnccount) = get_info_entry($entry);
|
| + # Convert test data
|
| + foreach $testname (keys(%{$testdata}))
|
| + {
|
| + $testdata->{$testname} =
|
| + apply_diff($testdata->{$testname}, $line_hash);
|
| + # Remove empty sets of test data
|
| + if (scalar(keys(%{$testdata->{$testname}})) == 0)
|
| + {
|
| + delete($testdata->{$testname});
|
| + delete($testfncdata->{$testname});
|
| + }
|
| + }
|
| + # Rename test data to indicate conversion
|
| + foreach $testname (keys(%{$testdata}))
|
| + {
|
| + # Skip testnames which already contain an extension
|
| + if ($testname =~ /,[^,]+$/)
|
| + {
|
| + next;
|
| + }
|
| + # Check for name conflict
|
| + if (defined($testdata->{$testname.",diff"}))
|
| + {
|
| + # Add counts
|
| + ($testdata->{$testname}) = add_counts(
|
| + $testdata->{$testname},
|
| + $testdata->{$testname.",diff"});
|
| + delete($testdata->{$testname.",diff"});
|
| + # Add function call counts
|
| + ($testfncdata->{$testname}) = add_fnccount(
|
| + $testfncdata->{$testname},
|
| + $testfncdata->{$testname.",diff"});
|
| + delete($testfncdata->{$testname.",diff"});
|
| + }
|
| + # Move test data to new testname
|
| + $testdata->{$testname.",diff"} = $testdata->{$testname};
|
| + delete($testdata->{$testname});
|
| + # Move function call count data to new testname
|
| + $testfncdata->{$testname.",diff"} =
|
| + $testfncdata->{$testname};
|
| + delete($testfncdata->{$testname});
|
| + }
|
| + # Convert summary of test data
|
| + $sumcount = apply_diff($sumcount, $line_hash);
|
| + # Convert function data
|
| + $funcdata = apply_diff_to_funcdata($funcdata, $line_hash);
|
| + # Convert checksum data
|
| + $checkdata = apply_diff($checkdata, $line_hash);
|
| + # Convert function call count data
|
| + adjust_fncdata($funcdata, $testfncdata, $sumfnccount);
|
| + ($f_found, $f_hit) = get_func_found_and_hit($sumfnccount);
|
| + # Update found/hit numbers
|
| + $found = 0;
|
| + $hit = 0;
|
| + foreach (keys(%{$sumcount}))
|
| + {
|
| + $found++;
|
| + if ($sumcount->{$_} > 0)
|
| + {
|
| + $hit++;
|
| + }
|
| + }
|
| + if ($found > 0)
|
| + {
|
| + # Store converted entry
|
| + set_info_entry($entry, $testdata, $sumcount, $funcdata,
|
| + $checkdata, $testfncdata, $sumfnccount,
|
| + $found, $hit, $f_found, $f_hit);
|
| + }
|
| + else
|
| + {
|
| + # Remove empty data set
|
| + delete($trace_data->{$filename});
|
| + }
|
| + }
|
| +
|
| + # Convert filenames as well if requested
|
| + if ($convert_filenames)
|
| + {
|
| + convert_paths($trace_data, \%path_conversion_data);
|
| + }
|
| +
|
| + info("$converted entr".($converted != 1 ? "ies" : "y")." converted, ".
|
| + "$unchanged entr".($unchanged != 1 ? "ies" : "y")." left ".
|
| + "unchanged.\n");
|
| +
|
| + # Write data
|
| + if ($to_file)
|
| + {
|
| + info("Writing data to $output_filename\n");
|
| + open(INFO_HANDLE, ">$output_filename")
|
| + or die("ERROR: cannot write to $output_filename!\n");
|
| + write_info_file(*INFO_HANDLE, $trace_data);
|
| + close(*INFO_HANDLE);
|
| + }
|
| + else
|
| + {
|
| + write_info_file(*STDOUT, $trace_data);
|
| + }
|
| +}
|
| +
|
| +
|
| +#
|
| +# system_no_output(mode, parameters)
|
| +#
|
| +# Call an external program using PARAMETERS while suppressing depending on
|
| +# the value of MODE:
|
| +#
|
| +# MODE & 1: suppress STDOUT
|
| +# MODE & 2: suppress STDERR
|
| +#
|
| +# Return 0 on success, non-zero otherwise.
|
| +#
|
| +
|
| +sub system_no_output($@)
|
| +{
|
| + my $mode = shift;
|
| + my $result;
|
| + local *OLD_STDERR;
|
| + local *OLD_STDOUT;
|
| +
|
| + # Save old stdout and stderr handles
|
| + ($mode & 1) && open(OLD_STDOUT, ">>&STDOUT");
|
| + ($mode & 2) && open(OLD_STDERR, ">>&STDERR");
|
| +
|
| + # Redirect to /dev/null
|
| + ($mode & 1) && open(STDOUT, ">/dev/null");
|
| + ($mode & 2) && open(STDERR, ">/dev/null");
|
| +
|
| + system(@_);
|
| + $result = $?;
|
| +
|
| + # Close redirected handles
|
| + ($mode & 1) && close(STDOUT);
|
| + ($mode & 2) && close(STDERR);
|
| +
|
| + # Restore old handles
|
| + ($mode & 1) && open(STDOUT, ">>&OLD_STDOUT");
|
| + ($mode & 2) && open(STDERR, ">>&OLD_STDERR");
|
| +
|
| + return $result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# read_config(filename)
|
| +#
|
| +# Read configuration file FILENAME and return a reference to a hash containing
|
| +# all valid key=value pairs found.
|
| +#
|
| +
|
| +sub read_config($)
|
| +{
|
| + my $filename = $_[0];
|
| + my %result;
|
| + my $key;
|
| + my $value;
|
| + local *HANDLE;
|
| +
|
| + if (!open(HANDLE, "<$filename"))
|
| + {
|
| + warn("WARNING: cannot read configuration file $filename\n");
|
| + return undef;
|
| + }
|
| + while (<HANDLE>)
|
| + {
|
| + chomp;
|
| + # Skip comments
|
| + s/#.*//;
|
| + # Remove leading blanks
|
| + s/^\s+//;
|
| + # Remove trailing blanks
|
| + s/\s+$//;
|
| + next unless length;
|
| + ($key, $value) = split(/\s*=\s*/, $_, 2);
|
| + if (defined($key) && defined($value))
|
| + {
|
| + $result{$key} = $value;
|
| + }
|
| + else
|
| + {
|
| + warn("WARNING: malformed statement in line $. ".
|
| + "of configuration file $filename\n");
|
| + }
|
| + }
|
| + close(HANDLE);
|
| + return \%result;
|
| +}
|
| +
|
| +
|
| +#
|
| +# apply_config(REF)
|
| +#
|
| +# REF is a reference to a hash containing the following mapping:
|
| +#
|
| +# key_string => var_ref
|
| +#
|
| +# where KEY_STRING is a keyword and VAR_REF is a reference to an associated
|
| +# variable. If the global configuration hash CONFIG contains a value for
|
| +# keyword KEY_STRING, VAR_REF will be assigned the value for that keyword.
|
| +#
|
| +
|
| +sub apply_config($)
|
| +{
|
| + my $ref = $_[0];
|
| +
|
| + foreach (keys(%{$ref}))
|
| + {
|
| + if (defined($config->{$_}))
|
| + {
|
| + ${$ref->{$_}} = $config->{$_};
|
| + }
|
| + }
|
| +}
|
| +
|
| +sub warn_handler($)
|
| +{
|
| + my ($msg) = @_;
|
| +
|
| + warn("$tool_name: $msg");
|
| +}
|
| +
|
| +sub die_handler($)
|
| +{
|
| + my ($msg) = @_;
|
| +
|
| + die("$tool_name: $msg");
|
| +}
|
|
|
| Property changes on: third_party/lcov/bin/lcov
|
| ___________________________________________________________________
|
| Name: svn:executable
|
| + *
|
|
|
|
|