#!/usr/bin/perl -w
#
# Copyright (c) International Business Machines Corp., 2002,2012
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation; either version 2 of the License, or (at
# your option) any later version.
#
# This program is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
#
#
# lcov
#
# This is a wrapper script which provides a single interface for accessing
# LCOV coverage data.
#
#
# History:
# 2002-08-29 created by Peter Oberparleiter <[email protected]>
# IBM Lab Boeblingen
# 2002-09-05 / Peter Oberparleiter: implemented --kernel-directory +
# multiple directories
# 2002-10-16 / Peter Oberparleiter: implemented --add-tracefile option
# 2002-10-17 / Peter Oberparleiter: implemented --extract option
# 2002-11-04 / Peter Oberparleiter: implemented --list option
# 2003-03-07 / Paul Larson: Changed to make it work with the latest gcov
# kernel patch. This will break it with older gcov-kernel
# patches unless you change the value of $gcovmod in this script
# 2003-04-07 / Peter Oberparleiter: fixed bug which resulted in an error
# when trying to combine .info files containing data without
# a test name
# 2003-04-10 / Peter Oberparleiter: extended Paul's change so that LCOV
# works both with the new and the old gcov-kernel patch
# 2003-04-10 / Peter Oberparleiter: added $gcov_dir constant in anticipation
# of a possible move of the gcov kernel directory to another
# file system in a future version of the gcov-kernel patch
# 2003-04-15 / Paul Larson: make info write to STDERR, not STDOUT
# 2003-04-15 / Paul Larson: added --remove option
# 2003-04-30 / Peter Oberparleiter: renamed --reset to --zerocounters
# to remove naming ambiguity with --remove
# 2003-04-30 / Peter Oberparleiter: adjusted help text to include --remove
# 2003-06-27 / Peter Oberparleiter: implemented --diff
# 2003-07-03 / Peter Oberparleiter: added line checksum support, added
# --no-checksum
# 2003-12-11 / Laurent Deniel: added --follow option
# 2004-03-29 / Peter Oberparleiter: modified --diff option to better cope with
# ambiguous patch file entries, modified --capture option to use
# modprobe before insmod (needed for 2.6)
# 2004-03-30 / Peter Oberparleiter: added --path option
# 2004-08-09 / Peter Oberparleiter: added configuration file support
# 2008-08-13 / Peter Oberparleiter: added function coverage support
#
use strict;
use File::Basename;
use File::Path;
use File::Find;
use File::Temp qw /tempdir/;
use File::Spec::Functions qw /abs2rel canonpath catdir catfile catpath
file_name_is_absolute rootdir splitdir splitpath/;
use Getopt::Long;
use Cwd qw /abs_path getcwd/;
# Global constants
our $lcov_version = 'LCOV version 1.10';
our $lcov_url = "http://ltp.sourceforge.net/coverage/lcov.php";
our $tool_name = basename($0);
# Directory containing gcov kernel files
our $gcov_dir;
# Where to create temporary directories
our $tmp_dir;
# Internal constants
our $GKV_PROC = 0; # gcov-kernel data in /proc via external patch
our $GKV_SYS = 1; # gcov-kernel data in /sys via vanilla 2.6.31+
our @GKV_NAME = ( "external", "upstream" );
our $pkg_gkv_file = ".gcov_kernel_version";
our $pkg_build_file = ".build_directory";
our $BR_BLOCK = 0;
our $BR_BRANCH = 1;
our $BR_TAKEN = 2;
our $BR_VEC_ENTRIES = 3;
our $BR_VEC_WIDTH = 32;
# Branch data combination types
our $BR_SUB = 0;
our $BR_ADD = 1;
# Prototypes
sub print_usage(*);
sub check_options();
sub userspace_reset();
sub userspace_capture();
sub kernel_reset();
sub kernel_capture();
sub kernel_capture_initial();
sub package_capture();
sub add_traces();
sub read_info_file($);
sub get_info_entry($);
sub set_info_entry($$$$$$$$$;$$$$$$);
sub add_counts($$);
sub merge_checksums($$$);
sub combine_info_entries($$$);
sub combine_info_files($$);
sub write_info_file(*$);
sub extract();
sub remove();
sub list();
sub get_common_filename($$);
sub read_diff($);
sub diff();
sub system_no_output($@);
sub read_config($);
sub apply_config($);
sub info(@);
sub create_temp_dir();
sub transform_pattern($);
sub warn_handler($);
sub die_handler($);
sub abort_handler($);
sub temp_cleanup();
sub setup_gkv();
sub get_overall_line($$$$);
sub print_overall_rate($$$$$$$$$);
sub lcov_geninfo(@);
sub create_package($$$;$);
sub get_func_found_and_hit($);
sub br_ivec_get($$);
sub summary();
sub rate($$;$$$);
# Global variables & initialization
our @directory; # Specifies where to get coverage data from
our @kernel_directory; # If set, captures only from specified kernel subdirs
our @add_tracefile; # If set, reads in and combines all files in list
our $list; # If set, list contents of tracefile
our $extract; # If set, extracts parts of tracefile
our $remove; # If set, removes parts of tracefile
our $diff; # If set, modifies tracefile according to diff
our $reset; # If set, reset all coverage data to zero
our $capture; # If set, capture data
our $output_filename; # Name for file to write coverage data to
our $test_name = ""; # Test case name
our $quiet = ""; # If set, suppress information messages
our $help; # Help option flag
our $version; # Version option flag
our $convert_filenames; # If set, convert filenames when applying diff
our $strip; # If set, strip leading directories when applying diff
our $temp_dir_name; # Name of temporary directory
our $cwd = `pwd`; # Current working directory
our $to_file; # If set, indicates that output is written to a file
our $follow; # If set, indicates that find shall follow links
our $diff_path = ""; # Path removed from tracefile when applying diff
our $base_directory; # Base directory (cwd of gcc during compilation)
our $checksum; # If set, calculate a checksum for each line
our $no_checksum; # If set, don't calculate a checksum for each line
our $compat_libtool; # If set, indicates that libtool mode is to be enabled
our $no_compat_libtool; # If set, indicates that libtool mode is to be disabled
our $gcov_tool;
our @opt_ignore_errors;
our $initial;
our $no_recursion = 0;
our $to_package;
our $from_package;
our $maxdepth;
our $no_markers;
our $config; # Configuration file contents
chomp($cwd);
our $tool_dir = dirname($0); # Directory where genhtml tool is installed
our @temp_dirs;
our $gcov_gkv; # gcov kernel support version found on machine
our $opt_derive_func_data;
our $opt_debug;
our $opt_list_full_path;
our $opt_no_list_full_path;
our $opt_list_width = 80;
our $opt_list_truncate_max = 20;
our $opt_external;
our $opt_no_external;
our $opt_config_file;
our %opt_rc;
our @opt_summary;
our $opt_compat;
our $ln_overall_found;
our $ln_overall_hit;
our $fn_overall_found;
our $fn_overall_hit;
our $br_overall_found;
our $br_overall_hit;
our $func_coverage = 1;
our $br_coverage = 0;
#
# Code entry point
#
$SIG{__WARN__} = \&warn_handler;
$SIG{__DIE__} = \&die_handler;
$SIG{'INT'} = \&abort_handler;
$SIG{'QUIT'} = \&abort_handler;
# Prettify version string
$lcov_version =~ s/\$\s*Revision\s*:?\s*(\S+)\s*\$/$1/;
# Add current working directory if $tool_dir is not already an absolute path
if (! ($tool_dir =~ /^\/(.*)$/))
{
$tool_dir = "$cwd/$tool_dir";
}
# Check command line for a configuration file name
Getopt::Long::Configure("pass_through", "no_auto_abbrev");
GetOptions("config-file=s" => \$opt_config_file,
"rc=s%" => \%opt_rc);
Getopt::Long::Configure("default");
# Read configuration file if available
if (defined($opt_config_file)) {
$config = read_config($opt_config_file);
} elsif (defined($ENV{"HOME"}) && (-r $ENV{"HOME"}."/.lcovrc"))
{
$config = read_config($ENV{"HOME"}."/.lcovrc");
}
elsif (-r "/etc/lcovrc")
{
$config = read_config("/etc/lcovrc");
}
if ($config || %opt_rc)
{
# Copy configuration file and --rc values to variables
apply_config({
"lcov_gcov_dir" => \$gcov_dir,
"lcov_tmp_dir" => \$tmp_dir,
"lcov_list_full_path" => \$opt_list_full_path,
"lcov_list_width" => \$opt_list_width,
"lcov_list_truncate_max"=> \$opt_list_truncate_max,
"lcov_branch_coverage" => \$br_coverage,
"lcov_function_coverage"=> \$func_coverage,
});
}
# Parse command line options
if (!GetOptions("directory|d|di=s" => \@directory,
"add-tracefile|a=s" => \@add_tracefile,
"list|l=s" => \$list,
"kernel-directory|k=s" => \@kernel_directory,
"extract|e=s" => \$extract,
"remove|r=s" => \$remove,
"diff=s" => \$diff,
"convert-filenames" => \$convert_filenames,
"strip=i" => \$strip,
"capture|c" => \$capture,
"output-file|o=s" => \$output_filename,
"test-name|t=s" => \$test_name,
"zerocounters|z" => \$reset,
"quiet|q" => \$quiet,
"help|h|?" => \$help,
"version|v" => \$version,
"follow|f" => \$follow,
"path=s" => \$diff_path,
"base-directory|b=s" => \$base_directory,
"checksum" => \$checksum,
"no-checksum" => \$no_checksum,
"compat-libtool" => \$compat_libtool,
"no-compat-libtool" => \$no_compat_libtool,
"gcov-tool=s" => \$gcov_tool,
"ignore-errors=s" => \@opt_ignore_errors,
"initial|i" => \$initial,
"no-recursion" => \$no_recursion,
"to-package=s" => \$to_package,
"from-package=s" => \$from_package,
"no-markers" => \$no_markers,
"derive-func-data" => \$opt_derive_func_data,
"debug" => \$opt_debug,
"list-full-path" => \$opt_list_full_path,
"no-list-full-path" => \$opt_no_list_full_path,
"external" => \$opt_external,
"no-external" => \$opt_no_external,
"summary=s" => \@opt_summary,
"compat=s" => \$opt_compat,
"config-file=s" => \$opt_config_file,
"rc=s%" => \%opt_rc,
))
{
print(STDERR "Use $tool_name --help to get usage information\n");
exit(1);
}
else
{
# Merge options
if (defined($no_checksum))
{
$checksum = ($no_checksum ? 0 : 1);
$no_checksum = undef;
}
if (defined($no_compat_libtool))
{
$compat_libtool = ($no_compat_libtool ? 0 : 1);
$no_compat_libtool = undef;
}
if (defined($opt_no_list_full_path))
{
$opt_list_full_path = ($opt_no_list_full_path ? 0 : 1);
$opt_no_list_full_path = undef;
}
if (defined($opt_no_external)) {
$opt_external = 0;
$opt_no_external = undef;
}
}
# Check for help option
if ($help)
{
print_usage(*STDOUT);
exit(0);
}
# Check for version option
if ($version)
{
print("$tool_name: $lcov_version\n");
exit(0);
}
# Check list width option
if ($opt_list_width <= 40) {
die("ERROR: lcov_list_width parameter out of range (needs to be ".
"larger than 40)\n");
}
# Normalize --path text
$diff_path =~ s/\/$//;
if ($follow)
{
$follow = "-follow";
}
else
{
$follow = "";
}
if ($no_recursion)
{
$maxdepth = "-maxdepth 1";
}
else
{
$maxdepth = "";
}
# Check for valid options
check_options();
# Only --extract, --remove and --diff allow unnamed parameters
if (@ARGV && !($extract || $remove || $diff || @opt_summary))
{
die("Extra parameter found: '".join(" ", @ARGV)."'\n".
"Use $tool_name --help to get usage information\n");
}
# Check for output filename
$to_file = ($output_filename && ($output_filename ne "-"));
if ($capture)
{
if (!$to_file)
{
# Option that tells geninfo to write to stdout
$output_filename = "-";
}
}
# Determine kernel directory for gcov data
if (!$from_package && !@directory && ($capture || $reset)) {
($gcov_gkv, $gcov_dir) = setup_gkv();
}
# Check for requested functionality
if ($reset)
{
# Differentiate between user space and kernel reset
if (@directory)
{
userspace_reset();
}
else
{
kernel_reset();
}
}
elsif ($capture)
{
# Capture source can be user space, kernel or package
if ($from_package) {
package_capture();
} elsif (@directory) {
userspace_capture();
} else {
if ($initial) {
if (defined($to_package)) {
die("ERROR: --initial cannot be used together ".
"with --to-package\n");
}
kernel_capture_initial();
} else {
kernel_capture();
}
}
}
elsif (@add_tracefile)
{
($ln_overall_found, $ln_overall_hit,
$fn_overall_found, $fn_overall_hit,
$br_overall_found, $br_overall_hit) = add_traces();
}
elsif ($remove)
{
($ln_overall_found, $ln_overall_hit,
$fn_overall_found, $fn_overall_hit,
$br_overall_found, $br_overall_hit) = remove();
}
elsif ($extract)
{
($ln_overall_found, $ln_overall_hit,
$fn_overall_found, $fn_overall_hit,
$br_overall_found, $br_overall_hit) = extract();
}
elsif ($list)
{
list();
}
elsif ($diff)
{
if (scalar(@ARGV) != 1)
{
die("ERROR: option --diff requires one additional argument!\n".
"Use $tool_name --help to get usage information\n");
}
($ln_overall_found, $ln_overall_hit,
$fn_overall_found, $fn_overall_hit,
$br_overall_found, $br_overall_hit) = diff();
}
elsif (@opt_summary)
{
($ln_overall_found, $ln_overall_hit,
$fn_overall_found, $fn_overall_hit,
$br_overall_found, $br_overall_hit) = summary();
}
temp_cleanup();
if (defined($ln_overall_found)) {
print_overall_rate(1, $ln_overall_found, $ln_overall_hit,
1, $fn_overall_found, $fn_overall_hit,
1, $br_overall_found, $br_overall_hit);
} else {
info("Done.\n") if (!$list && !$capture);
}
exit(0);
#
# print_usage(handle)
#
# Print usage information.
#
sub print_usage(*)
{
local *HANDLE = $_[0];
print(HANDLE <<END_OF_USAGE);
Usage: $tool_name [OPTIONS]
Use lcov to collect coverage data from either the currently running Linux
kernel or from a user space application. Specify the --directory option to
get coverage data for a user space program.
Misc:
-h, --help Print this help, then exit
-v, --version Print version number, then exit
-q, --quiet Do not print progress messages
Operation:
-z, --zerocounters Reset all execution counts to zero
-c, --capture Capture coverage data
-a, --add-tracefile FILE Add contents of tracefiles
-e, --extract FILE PATTERN Extract files matching PATTERN from FILE
-r, --remove FILE PATTERN Remove files matching PATTERN from FILE
-l, --list FILE List contents of tracefile FILE
--diff FILE DIFF Transform tracefile FILE according to DIFF
--summary FILE Show summary coverage data for tracefiles
Options:
-i, --initial Capture initial zero coverage data
-t, --test-name NAME Specify test name to be stored with data
-o, --output-file FILENAME Write data to FILENAME instead of stdout
-d, --directory DIR Use .da files in DIR instead of kernel
-f, --follow Follow links when searching .da files
-k, --kernel-directory KDIR Capture kernel coverage data only from KDIR
-b, --base-directory DIR Use DIR as base directory for relative paths
--convert-filenames Convert filenames when applying diff
--strip DEPTH Strip initial DEPTH directory levels in diff
--path PATH Strip PATH from tracefile when applying diff
--(no-)checksum Enable (disable) line checksumming
--(no-)compat-libtool Enable (disable) libtool compatibility mode
--gcov-tool TOOL Specify gcov tool location
--ignore-errors ERRORS Continue after ERRORS (gcov, source, graph)
--no-recursion Exclude subdirectories from processing
--to-package FILENAME Store unprocessed coverage data in FILENAME
--from-package FILENAME Capture from unprocessed data in FILENAME
--no-markers Ignore exclusion markers in source code
--derive-func-data Generate function data from line data
--list-full-path Print full path during a list operation
--(no-)external Include (ignore) data for external files
--config-file FILENAME Specify configuration file location
--rc SETTING=VALUE Override configuration file setting
--compat MODE=on|off|auto Set compat MODE (libtool, hammer, split_crc)
For more information see: $lcov_url
END_OF_USAGE
;
}
#
# check_options()
#
# Check for valid combination of command line options. Die on error.
#
sub check_options()
{
my $i = 0;
# Count occurrence of mutually exclusive options
$reset && $i++;
$capture && $i++;
@add_tracefile && $i++;
$extract && $i++;
$remove && $i++;
$list && $i++;
$diff && $i++;
@opt_summary && $i++;
if ($i == 0)
{
die("Need one of options -z, -c, -a, -e, -r, -l, ".
"--diff or --summary\n".
"Use $tool_name --help to get usage information\n");
}
elsif ($i > 1)
{
die("ERROR: only one of -z, -c, -a, -e, -r, -l, ".
"--diff or --summary allowed!\n".
"Use $tool_name --help to get usage information\n");
}
}
#
# userspace_reset()
#
# Reset coverage data found in DIRECTORY by deleting all contained .da files.
#
# Die on error.
#
sub userspace_reset()
{
my $current_dir;
my @file_list;
foreach $current_dir (@directory)
{
info("Deleting all .da files in $current_dir".
($no_recursion?"\n":" and subdirectories\n"));
@file_list = `find "$current_dir" $maxdepth $follow -name \\*\\.da -o -name \\*\\.gcda -type f 2>/dev/null`;
chomp(@file_list);
foreach (@file_list)
{
unlink($_) or die("ERROR: cannot remove file $_!\n");
}
}
}
#
# userspace_capture()
#
# Capture coverage data found in DIRECTORY and write it to a package (if
# TO_PACKAGE specified) or to OUTPUT_FILENAME or STDOUT.
#
# Die on error.
#
sub userspace_capture()
{
my $dir;
my $build;
if (!defined($to_package)) {
lcov_geninfo(@directory);
return;
}
if (scalar(@directory) != 1) {
die("ERROR: -d may be specified only once with --to-package\n");
}
$dir = $directory[0];
if (defined($base_directory)) {
$build = $base_directory;
} else {
$build = $dir;
}
create_package($to_package, $dir, $build);
}
#
# kernel_reset()
#
# Reset kernel coverage.
#
# Die on error.
#
sub kernel_reset()
{
local *HANDLE;
my $reset_file;
info("Resetting kernel execution counters\n");
if (-e "$gcov_dir/vmlinux") {
$reset_file = "$gcov_dir/vmlinux";
} elsif (-e "$gcov_dir/reset") {
$reset_file = "$gcov_dir/reset";
} else {
die("ERROR: no reset control found in $gcov_dir\n");
}
open(HANDLE, ">", $reset_file) or
die("ERROR: cannot write to $reset_file!\n");
print(HANDLE "0");
close(HANDLE);
}
#
# lcov_copy_single(from, to)
#
# Copy single regular file FROM to TO without checking its size. This is
# required to work with special files generated by the kernel
# seq_file-interface.
#
#
sub lcov_copy_single($$)
{
my ($from, $to) = @_;
my $content;
local $/;
local *HANDLE;
open(HANDLE, "<", $from) or die("ERROR: cannot read $from: $!\n");
$content = <HANDLE>;
close(HANDLE);
open(HANDLE, ">", $to) or die("ERROR: cannot write $from: $!\n");
if (defined($content)) {
print(HANDLE $content);
}
close(HANDLE);
}
#
# lcov_find(dir, function, data[, extension, ...)])
#
# Search DIR for files and directories whose name matches PATTERN and run
# FUNCTION for each match. If not pattern is specified, match all names.
#
# FUNCTION has the following prototype:
# function(dir, relative_name, data)
#
# Where:
# dir: the base directory for this search
# relative_name: the name relative to the base directory of this entry
# data: the DATA variable passed to lcov_find
#
sub lcov_find($$$;@)
{
my ($dir, $fn, $data, @pattern) = @_;
my $result;
my $_fn = sub {
my $filename = $File::Find::name;
if (defined($result)) {
return;
}
$filename = abs2rel($filename, $dir);
foreach (@pattern) {
if ($filename =~ /$_/) {
goto ok;
}
}
return;
ok:
$result = &$fn($dir, $filename, $data);
};
if (scalar(@pattern) == 0) {
@pattern = ".*";
}
find( { wanted => $_fn, no_chdir => 1 }, $dir);
return $result;
}
#
# lcov_copy_fn(from, rel, to)
#
# Copy directories, files and links from/rel to to/rel.
#
sub lcov_copy_fn($$$)
{
my ($from, $rel, $to) = @_;
my $absfrom = canonpath(catfile($from, $rel));
my $absto = canonpath(catfile($to, $rel));
if (-d) {
if (! -d $absto) {
mkpath($absto) or
die("ERROR: cannot create directory $absto\n");
chmod(0700, $absto);
}
} elsif (-l) {
# Copy symbolic link
my $link = readlink($absfrom);
if (!defined($link)) {
die("ERROR: cannot read link $absfrom: $!\n");
}
symlink($link, $absto) or
die("ERROR: cannot create link $absto: $!\n");
} else {
lcov_copy_single($absfrom, $absto);
chmod(0600, $absto);
}
return undef;
}
#
# lcov_copy(from, to, subdirs)
#
# Copy all specified SUBDIRS and files from directory FROM to directory TO. For
# regular files, copy file contents without checking its size. This is required
# to work with seq_file-generated files.
#
sub lcov_copy($$;@)
{
my ($from, $to, @subdirs) = @_;
my @pattern;
foreach (@subdirs) {
push(@pattern, "^$_");
}
lcov_find($from, \&lcov_copy_fn, $to, @pattern);
}
#
# lcov_geninfo(directory)
#
# Call geninfo for the specified directory and with the parameters specified
# at the command line.
#
sub lcov_geninfo(@)
{
my (@dir) = @_;
my @param;
# Capture data
info("Capturing coverage data from ".join(" ", @dir)."\n");
@param = ("$tool_dir/geninfo", @dir);
if ($output_filename)
{
@param = (@param, "--output-filename", $output_filename);
}
if ($test_name)
{
@param = (@param, "--test-name", $test_name);
}
if ($follow)
{
@param = (@param, "--follow");
}
if ($quiet)
{
@param = (@param, "--quiet");
}
if (defined($checksum))
{
if ($checksum)
{
@param = (@param, "--checksum");
}
else
{
@param = (@param, "--no-checksum");
}
}
if ($base_directory)
{
@param = (@param, "--base-directory", $base_directory);
}
if ($no_compat_libtool)
{
@param = (@param, "--no-compat-libtool");
}
elsif ($compat_libtool)
{
@param = (@param, "--compat-libtool");
}
if ($gcov_tool)
{
@param = (@param, "--gcov-tool", $gcov_tool);
}
foreach (@opt_ignore_errors) {
@param = (@param, "--ignore-errors", $_);
}
if ($no_recursion) {
@param = (@param, "--no-recursion");
}
if ($initial)
{
@param = (@param, "--initial");
}
if ($no_markers)
{
@param = (@param, "--no-markers");
}
if ($opt_derive_func_data)
{
@param = (@param, "--derive-func-data");
}
if ($opt_debug)
{
@param = (@param, "--debug");
}
if (defined($opt_external) && $opt_external)
{
@param = (@param, "--external");
}
if (defined($opt_external) && !$opt_external)
{
@param = (@param, "--no-external");
}
if (defined($opt_compat)) {
@param = (@param, "--compat", $opt_compat);
}
if (%opt_rc) {
foreach my $key (keys(%opt_rc)) {
@param = (@param, "--rc", "$key=".$opt_rc{$key});
}
}
system(@param) and exit($? >> 8);
}
#
# read_file(filename)
#
# Return the contents of the file defined by filename.
#
sub read_file($)
{
my ($filename) = @_;
my $content;
local $\;
local *HANDLE;
open(HANDLE, "<", $filename) || return undef;
$content = <HANDLE>;
close(HANDLE);
return $content;
}
#
# get_package(package_file)
#
# Unpack unprocessed coverage data files from package_file to a temporary
# directory and return directory name, build directory and gcov kernel version
# as found in package.
#
sub get_package($)
{
my ($file) = @_;
my $dir = create_temp_dir();
my $gkv;
my $build;
my $cwd = getcwd();
my $count;
local *HANDLE;
info("Reading package $file:\n");
info(" data directory .......: $dir\n");
$file = abs_path($file);
chdir($dir);
open(HANDLE, "-|", "tar xvfz '$file' 2>/dev/null")
or die("ERROR: could not process package $file\n");
while (<HANDLE>) {
if (/\.da$/ || /\.gcda$/) {
$count++;
}
}
close(HANDLE);
$build = read_file("$dir/$pkg_build_file");
if (defined($build)) {
info(" build directory ......: $build\n");
}
$gkv = read_file("$dir/$pkg_gkv_file");
if (defined($gkv)) {
$gkv = int($gkv);
if ($gkv != $GKV_PROC && $gkv != $GKV_SYS) {
die("ERROR: unsupported gcov kernel version found ".
"($gkv)\n");
}
info(" content type .........: kernel data\n");
info(" gcov kernel version ..: %s\n", $GKV_NAME[$gkv]);
} else {
info(" content type .........: application data\n");
}
info(" data files ...........: $count\n");
chdir($cwd);
return ($dir, $build, $gkv);
}
#
# write_file(filename, $content)
#
# Create a file named filename and write the specified content to it.
#
sub write_file($$)
{
my ($filename, $content) = @_;
local *HANDLE;
open(HANDLE, ">", $filename) || return 0;
print(HANDLE $content);
close(HANDLE) || return 0;
return 1;
}
# count_package_data(filename)
#
# Count the number of coverage data files in the specified package file.
#
sub count_package_data($)
{
my ($filename) = @_;
local *HANDLE;
my $count = 0;
open(HANDLE, "-|", "tar tfz '$filename'") or return undef;
while (<HANDLE>) {
if (/\.da$/ || /\.gcda$/) {
$count++;
}
}
close(HANDLE);
return $count;
}
#
# create_package(package_file, source_directory, build_directory[,
# kernel_gcov_version])
#
# Store unprocessed coverage data files from source_directory to package_file.
#
sub create_package($$$;$)
{
my ($file, $dir, $build, $gkv) = @_;
my $cwd = getcwd();
# Print information about the package
info("Creating package $file:\n");
info(" data directory .......: $dir\n");
# Handle build directory
if (defined($build)) {
info(" build directory ......: $build\n");
write_file("$dir/$pkg_build_file", $build)
or die("ERROR: could not write to ".
"$dir/$pkg_build_file\n");
}
# Handle gcov kernel version data
if (defined($gkv)) {
info(" content type .........: kernel data\n");
info(" gcov kernel version ..: %s\n", $GKV_NAME[$gkv]);
write_file("$dir/$pkg_gkv_file", $gkv)
or die("ERROR: could not write to ".
"$dir/$pkg_gkv_file\n");
} else {
info(" content type .........: application data\n");
}
# Create package
$file = abs_path($file);
chdir($dir);
system("tar cfz $file .")
and die("ERROR: could not create package $file\n");
# Remove temporary files
unlink("$dir/$pkg_build_file");
unlink("$dir/$pkg_gkv_file");
# Show number of data files
if (!$quiet) {
my $count = count_package_data($file);
if (defined($count)) {
info(" data files ...........: $count\n");
}
}
chdir($cwd);
}
sub find_link_fn($$$)
{
my ($from, $rel, $filename) = @_;
my $absfile = catfile($from, $rel, $filename);
if (-l $absfile) {
return $absfile;
}
return undef;
}
#
# get_base(dir)
#
# Return (BASE, OBJ), where
# - BASE: is the path to the kernel base directory relative to dir
# - OBJ: is the absolute path to the kernel build directory
#
sub get_base($)
{
my ($dir) = @_;
my $marker = "kernel/gcov/base.gcno";
my $markerfile;
my $sys;
my $obj;
my $link;
$markerfile = lcov_find($dir, \&find_link_fn, $marker);
if (!defined($markerfile)) {
return (undef, undef);
}
# sys base is parent of parent of markerfile.
$sys = abs2rel(dirname(dirname(dirname($markerfile))), $dir);
# obj base is parent of parent of markerfile link target.
$link = readlink($markerfile);
if (!defined($link)) {
die("ERROR: could not read $markerfile\n");
}
$obj = dirname(dirname(dirname($link)));
return ($sys, $obj);
}
#
# apply_base_dir(data_dir, base_dir, build_dir, @directories)
#
# Make entries in @directories relative to data_dir.
#
sub apply_base_dir($$$@)
{
my ($data, $base, $build, @dirs) = @_;
my $dir;
my @result;
foreach $dir (@dirs) {
# Is directory path relative to data directory?
if (-d catdir($data, $dir)) {
push(@result, $dir);
next;
}
# Relative to the auto-detected base-directory?
if (defined($base)) {
if (-d catdir($data, $base, $dir)) {
push(@result, catdir($base, $dir));
next;
}
}
# Relative to the specified base-directory?
if (defined($base_directory)) {
if (file_name_is_absolute($base_directory)) {
$base = abs2rel($base_directory, rootdir());
} else {
$base = $base_directory;
}
if (-d catdir($data, $base, $dir)) {
push(@result, catdir($base, $dir));
next;
}
}
# Relative to the build directory?
if (defined($build)) {
if (file_name_is_absolute($build)) {
$base = abs2rel($build, rootdir());
} else {
$base = $build;
}
if (-d catdir($data, $base, $dir)) {
push(@result, catdir($base, $dir));
next;
}
}
die("ERROR: subdirectory $dir not found\n".
"Please use -b to specify the correct directory\n");
}
return @result;
}
#
# copy_gcov_dir(dir, [@subdirectories])
#
# Create a temporary directory and copy all or, if specified, only some
# subdirectories from dir to that directory. Return the name of the temporary
# directory.
#
sub copy_gcov_dir($;@)
{
my ($data, @dirs) = @_;
my $tempdir = create_temp_dir();
info("Copying data to temporary directory $tempdir\n");
lcov_copy($data, $tempdir, @dirs);
return $tempdir;
}
#
# kernel_capture_initial
#
# Capture initial kernel coverage data, i.e. create a coverage data file from
# static graph files which contains zero coverage data for all instrumented
# lines.
#
sub kernel_capture_initial()
{
my $build;
my $source;
my @params;
if (defined($base_directory)) {
$build = $base_directory;
$source = "specified";
} else {
(undef, $build) = get_base($gcov_dir);
if (!defined($build)) {
die("ERROR: could not auto-detect build directory.\n".
"Please use -b to specify the build directory\n");
}
$source = "auto-detected";
}
info("Using $build as kernel build directory ($source)\n");
# Build directory needs to be passed to geninfo
$base_directory = $build;
if (@kernel_directory) {
foreach my $dir (@kernel_directory) {
push(@params, "$build/$dir");
}
} else {
push(@params, $build);
}
lcov_geninfo(@params);
}
#
# kernel_capture_from_dir(directory, gcov_kernel_version, build)
#
# Perform the actual kernel coverage capturing from the specified directory
# assuming that the data was copied from the specified gcov kernel version.
#
sub kernel_capture_from_dir($$$)
{
my ($dir, $gkv, $build) = @_;
# Create package or coverage file
if (defined($to_package)) {
create_package($to_package, $dir, $build, $gkv);
} else {
# Build directory needs to be passed to geninfo
$base_directory = $build;
lcov_geninfo($dir);
}
}
#
# adjust_kernel_dir(dir, build)
#
# Adjust directories specified with -k so that they point to the directory
# relative to DIR. Return the build directory if specified or the auto-
# detected build-directory.
#
sub adjust_kernel_dir($$)
{
my ($dir, $build) = @_;
my ($sys_base, $build_auto) = get_base($dir);
if (!defined($build)) {
$build = $build_auto;
}
if (!defined($build)) {
die("ERROR: could not auto-detect build directory.\n".
"Please use -b to specify the build directory\n");
}
# Make @kernel_directory relative to sysfs base
if (@kernel_directory) {
@kernel_directory = apply_base_dir($dir, $sys_base, $build,
@kernel_directory);
}
return $build;
}
sub kernel_capture()
{
my $data_dir;
my $build = $base_directory;
if ($gcov_gkv == $GKV_SYS) {
$build = adjust_kernel_dir($gcov_dir, $build);
}
$data_dir = copy_gcov_dir($gcov_dir, @kernel_directory);
kernel_capture_from_dir($data_dir, $gcov_gkv, $build);
}
#
# package_capture()
#
# Capture coverage data from a package of unprocessed coverage data files
# as generated by lcov --to-package.
#
sub package_capture()
{
my $dir;
my $build;
my $gkv;
($dir, $build, $gkv) = get_package($from_package);
# Check for build directory
if (defined($base_directory)) {
if (defined($build)) {
info("Using build directory specified by -b.\n");
}
$build = $base_directory;
}
# Do the actual capture
if (defined($gkv)) {
if ($gkv == $GKV_SYS) {
$build = adjust_kernel_dir($dir, $build);
}
if (@kernel_directory) {
$dir = copy_gcov_dir($dir, @kernel_directory);
}
kernel_capture_from_dir($dir, $gkv, $build);
} else {
# Build directory needs to be passed to geninfo
$base_directory = $build;
lcov_geninfo($dir);
}
}
#
# info(printf_parameter)
#
# Use printf to write PRINTF_PARAMETER to stdout only when the $quiet flag
# is not set.
#
sub info(@)
{
if (!$quiet)
{
# Print info string
if ($to_file)
{
printf(@_)
}
else
{
# Don't interfere with the .info output to STDOUT
printf(STDERR @_);
}
}
}
#
# create_temp_dir()
#
# Create a temporary directory and return its path.
#
# Die on error.
#
sub create_temp_dir()
{
my $dir;
if (defined($tmp_dir)) {
$dir = tempdir(DIR => $tmp_dir, CLEANUP => 1);
} else {
$dir = tempdir(CLEANUP => 1);
}
if (!defined($dir)) {
die("ERROR: cannot create temporary directory\n");
}
push(@temp_dirs, $dir);
return $dir;
}
#
# br_taken_to_num(taken)
#
# Convert a branch taken value .info format to number format.
#
sub br_taken_to_num($)
{
my ($taken) = @_;
return 0 if ($taken eq '-');
return $taken + 1;
}
#
# br_num_to_taken(taken)
#
# Convert a branch taken value in number format to .info format.
#
sub br_num_to_taken($)
{
my ($taken) = @_;
return '-' if ($taken == 0);
return $taken - 1;
}
#
# br_taken_add(taken1, taken2)
#
# Return the result of taken1 + taken2 for 'branch taken' values.
#
sub br_taken_add($$)
{
my ($t1, $t2) = @_;
return $t1 if (!defined($t2));
return $t2 if (!defined($t1));
return $t1 if ($t2 eq '-');
return $t2 if ($t1 eq '-');
return $t1 + $t2;
}
#
# br_taken_sub(taken1, taken2)
#
# Return the result of taken1 - taken2 for 'branch taken' values. Return 0
# if the result would become negative.
#
sub br_taken_sub($$)
{
my ($t1, $t2) = @_;
return $t1 if (!defined($t2));
return undef if (!defined($t1));
return $t1 if ($t1 eq '-');
return $t1 if ($t2 eq '-');
return 0 if $t2 > $t1;
return $t1 - $t2;
}
#
#
# br_ivec_len(vector)
#
# Return the number of entries in the branch coverage vector.
#
sub br_ivec_len($)
{
my ($vec) = @_;
return 0 if (!defined($vec));
return (length($vec) * 8 / $BR_VEC_WIDTH) / $BR_VEC_ENTRIES;
}
#
# br_ivec_push(vector, block, branch, taken)
#
# Add an entry to the branch coverage vector. If an entry with the same
# branch ID already exists, add the corresponding taken values.
#
sub br_ivec_push($$$$)
{
my ($vec, $block, $branch, $taken) = @_;
my $offset;
my $num = br_ivec_len($vec);
my $i;
$vec = "" if (!defined($vec));
# Check if branch already exists in vector
for ($i = 0; $i < $num; $i++) {
my ($v_block, $v_branch, $v_taken) = br_ivec_get($vec, $i);
next if ($v_block != $block || $v_branch != $branch);
# Add taken counts
$taken = br_taken_add($taken, $v_taken);
last;
}
$offset = $i * $BR_VEC_ENTRIES;
$taken = br_taken_to_num($taken);
# Add to vector
vec($vec, $offset + $BR_BLOCK, $BR_VEC_WIDTH) = $block;
vec($vec, $offset + $BR_BRANCH, $BR_VEC_WIDTH) = $branch;
vec($vec, $offset + $BR_TAKEN, $BR_VEC_WIDTH) = $taken;
return $vec;
}
#
# br_ivec_get(vector, number)
#
# Return an entry from the branch coverage vector.
#
sub br_ivec_get($$)
{
my ($vec, $num) = @_;
my $block;
my $branch;
my $taken;
my $offset = $num * $BR_VEC_ENTRIES;
# Retrieve data from vector
$block = vec($vec, $offset + $BR_BLOCK, $BR_VEC_WIDTH);
$branch = vec($vec, $offset + $BR_BRANCH, $BR_VEC_WIDTH);
$taken = vec($vec, $offset + $BR_TAKEN, $BR_VEC_WIDTH);
# Decode taken value from an integer
$taken = br_num_to_taken($taken);
return ($block, $branch, $taken);
}
#
# get_br_found_and_hit(brcount)
#
# Return (br_found, br_hit) for brcount
#
sub get_br_found_and_hit($)
{
my ($brcount) = @_;
my $line;
my $br_found = 0;
my $br_hit = 0;
foreach $line (keys(%{$brcount})) {
my $brdata = $brcount->{$line};
my $i;
my $num = br_ivec_len($brdata);
for ($i = 0; $i < $num; $i++) {
my $taken;
(undef, undef, $taken) = br_ivec_get($brdata, $i);
$br_found++;
$br_hit++ if ($taken ne "-" && $taken > 0);
}
}
return ($br_found, $br_hit);
}
#
# read_info_file(info_filename)
#
# Read in the contents of the .info file specified by INFO_FILENAME. Data will
# be returned as a reference to a hash containing the following mappings:
#
# %result: for each filename found in file -> \%data
#
# %data: "test" -> \%testdata
# "sum" -> \%sumcount
# "func" -> \%funcdata
# "found" -> $lines_found (number of instrumented lines found in file)
# "hit" -> $lines_hit (number of executed lines in file)
# "check" -> \%checkdata
# "testfnc" -> \%testfncdata
# "sumfnc" -> \%sumfnccount
# "testbr" -> \%testbrdata
# "sumbr" -> \%sumbrcount
#
# %testdata : name of test affecting this file -> \%testcount
# %testfncdata: name of test affecting this file -> \%testfnccount
# %testbrdata: name of test affecting this file -> \%testbrcount
#
# %testcount : line number -> execution count for a single test
# %testfnccount: function name -> execution count for a single test
# %testbrcount : line number -> branch coverage data for a single test
# %sumcount : line number -> execution count for all tests
# %sumfnccount : function name -> execution count for all tests
# %sumbrcount : line number -> branch coverage data for all tests
# %funcdata : function name -> line number
# %checkdata : line number -> checksum of source code line
# $brdata : vector of items: block, branch, taken
#
# Note that .info file sections referring to the same file and test name
# will automatically be combined by adding all execution counts.
#
# Note that if INFO_FILENAME ends with ".gz", it is assumed that the file
# is compressed using GZIP. If available, GUNZIP will be used to decompress
# this file.
#
# Die on error.
#
sub read_info_file($)
{
my $tracefile = $_[0]; # Name of tracefile
my %result; # Resulting hash: file -> data
my $data; # Data handle for current entry
my $testdata; # " "
my $testcount; # " "
my $sumcount; # " "
my $funcdata; # " "
my $checkdata; # " "
my $testfncdata;
my $testfnccount;
my $sumfnccount;
my $testbrdata;
my $testbrcount;
my $sumbrcount;
my $line; # Current line read from .info file
my $testname; # Current test name
my $filename; # Current filename
my $hitcount; # Count for lines hit
my $count; # Execution count of current line
my $negative; # If set, warn about negative counts
my $changed_testname; # If set, warn about changed testname
my $line_checksum; # Checksum of current line
local *INFO_HANDLE; # Filehandle for .info file
info("Reading tracefile $tracefile\n");
# Check if file exists and is readable
stat($_[0]);
if (!(-r _))
{
die("ERROR: cannot read file $_[0]!\n");
}
# Check if this is really a plain file
if (!(-f _))
{
die("ERROR: not a plain file: $_[0]!\n");
}
# Check for .gz extension
if ($_[0] =~ /\.gz$/)
{
# Check for availability of GZIP tool
system_no_output(1, "gunzip" ,"-h")
and die("ERROR: gunzip command not available!\n");
# Check integrity of compressed file
system_no_output(1, "gunzip", "-t", $_[0])
and die("ERROR: integrity check failed for ".
"compressed file $_[0]!\n");
# Open compressed file
open(INFO_HANDLE, "-|", "gunzip -c '$_[0]'")
or die("ERROR: cannot start gunzip to decompress ".
"file $_[0]!\n");
}
else
{
# Open decompressed file
open(INFO_HANDLE, "<", $_[0])
or die("ERROR: cannot read file $_[0]!\n");
}
$testname = "";
while (<INFO_HANDLE>)
{
chomp($_);
$line = $_;
# Switch statement
foreach ($line)
{
/^TN:([^,]*)(,diff)?/ && do
{
# Test name information found
$testname = defined($1) ? $1 : "";
if ($testname =~ s/\W/_/g)
{
$changed_testname = 1;
}
$testname .= $2 if (defined($2));
last;
};
/^[SK]F:(.*)/ && do
{
# Filename information found
# Retrieve data for new entry
$filename = $1;
$data = $result{$filename};
($testdata, $sumcount, $funcdata, $checkdata,
$testfncdata, $sumfnccount, $testbrdata,
$sumbrcount) =
get_info_entry($data);
if (defined($testname))
{
$testcount = $testdata->{$testname};
$testfnccount = $testfncdata->{$testname};
$testbrcount = $testbrdata->{$testname};
}
else
{
$testcount = {};
$testfnccount = {};
$testbrcount = {};
}
last;
};
/^DA:(\d+),(-?\d+)(,[^,\s]+)?/ && do
{
# Fix negative counts
$count = $2 < 0 ? 0 : $2;
if ($2 < 0)
{
$negative = 1;
}
# Execution count found, add to structure
# Add summary counts
$sumcount->{$1} += $count;
# Add test-specific counts
if (defined($testname))
{
$testcount->{$1} += $count;
}
# Store line checksum if available
if (defined($3))
{
$line_checksum = substr($3, 1);
# Does it match a previous definition
if (defined($checkdata->{$1}) &&
($checkdata->{$1} ne
$line_checksum))
{
die("ERROR: checksum mismatch ".
"at $filename:$1\n");
}
$checkdata->{$1} = $line_checksum;
}
last;
};
/^FN:(\d+),([^,]+)/ && do
{
last if (!$func_coverage);
# Function data found, add to structure
$funcdata->{$2} = $1;
# Also initialize function call data
if (!defined($sumfnccount->{$2})) {
$sumfnccount->{$2} = 0;
}
if (defined($testname))
{
if (!defined($testfnccount->{$2})) {
$testfnccount->{$2} = 0;
}
}
last;
};
/^FNDA:(\d+),([^,]+)/ && do
{
last if (!$func_coverage);
# Function call count found, add to structure
# Add summary counts
$sumfnccount->{$2} += $1;
# Add test-specific counts
if (defined($testname))
{
$testfnccount->{$2} += $1;
}
last;
};
/^BRDA:(\d+),(\d+),(\d+),(\d+|-)/ && do {
# Branch coverage data found
my ($line, $block, $branch, $taken) =
($1, $2, $3, $4);
last if (!$br_coverage);
$sumbrcount->{$line} =
br_ivec_push($sumbrcount->{$line},
$block, $branch, $taken);
# Add test-specific counts
if (defined($testname)) {
$testbrcount->{$line} =
br_ivec_push(
$testbrcount->{$line},
$block, $branch,
$taken);
}
last;
};
/^end_of_record/ && do
{
# Found end of section marker
if ($filename)
{
# Store current section data
if (defined($testname))
{
$testdata->{$testname} =
$testcount;
$testfncdata->{$testname} =
$testfnccount;
$testbrdata->{$testname} =
$testbrcount;
}
set_info_entry($data, $testdata,
$sumcount, $funcdata,
$checkdata, $testfncdata,
$sumfnccount,
$testbrdata,
$sumbrcount);
$result{$filename} = $data;
last;
}
};
# default
last;
}
}
close(INFO_HANDLE);
# Calculate hit and found values for lines and functions of each file
foreach $filename (keys(%result))
{
$data = $result{$filename};
($testdata, $sumcount, undef, undef, $testfncdata,
$sumfnccount, $testbrdata, $sumbrcount) =
get_info_entry($data);
# Filter out empty files
if (scalar(keys(%{$sumcount})) == 0)
{
delete($result{$filename});
next;
}
# Filter out empty test cases
foreach $testname (keys(%{$testdata}))
{
if (!defined($testdata->{$testname}) ||
scalar(keys(%{$testdata->{$testname}})) == 0)
{
delete($testdata->{$testname});
delete($testfncdata->{$testname});
}
}
$data->{"found"} = scalar(keys(%{$sumcount}));
$hitcount = 0;
foreach (keys(%{$sumcount}))
{
if ($sumcount->{$_} > 0) { $hitcount++; }
}
$data->{"hit"} = $hitcount;
# Get found/hit values for function call data
$data->{"f_found"} = scalar(keys(%{$sumfnccount}));
$hitcount = 0;
foreach (keys(%{$sumfnccount})) {
if ($sumfnccount->{$_} > 0) {
$hitcount++;
}
}
$data->{"f_hit"} = $hitcount;
# Get found/hit values for branch data
{
my ($br_found, $br_hit) = get_br_found_and_hit($sumbrcount);
$data->{"b_found"} = $br_found;
$data->{"b_hit"} = $br_hit;
}
}
if (scalar(keys(%result)) == 0)
{
die("ERROR: no valid records found in tracefile $tracefile\n");
}
if ($negative)
{
warn("WARNING: negative counts found in tracefile ".
"$tracefile\n");
}
if ($changed_testname)
{
warn("WARNING: invalid characters removed from testname in ".
"tracefile $tracefile\n");
}
return(\%result);
}
#
# get_info_entry(hash_ref)
#
# Retrieve data from an entry of the structure generated by read_info_file().
# Return a list of references to hashes:
# (test data hash ref, sum count hash ref, funcdata hash ref, checkdata hash
# ref, testfncdata hash ref, sumfnccount hash ref, testbrdata hash ref,
# sumbrcount hash ref, lines found, lines hit, functions found,
# functions hit, branches found, branches hit)
#
sub get_info_entry($)
{
my $testdata_ref = $_[0]->{"test"};
my $sumcount_ref = $_[0]->{"sum"};
my $funcdata_ref = $_[0]->{"func"};
my $checkdata_ref = $_[0]->{"check"};
my $testfncdata = $_[0]->{"testfnc"};
my $sumfnccount = $_[0]->{"sumfnc"};
my $testbrdata = $_[0]->{"testbr"};
my $sumbrcount = $_[0]->{"sumbr"};
my $lines_found = $_[0]->{"found"};
my $lines_hit = $_[0]->{"hit"};
my $f_found = $_[0]->{"f_found"};
my $f_hit = $_[0]->{"f_hit"};
my $br_found = $_[0]->{"b_found"};
my $br_hit = $_[0]->{"b_hit"};
return ($testdata_ref, $sumcount_ref, $funcdata_ref, $checkdata_ref,
$testfncdata, $sumfnccount, $testbrdata, $sumbrcount,
$lines_found, $lines_hit, $f_found, $f_hit,
$br_found, $br_hit);
}
#
# set_info_entry(hash_ref, testdata_ref, sumcount_ref, funcdata_ref,
# checkdata_ref, testfncdata_ref, sumfcncount_ref,
# testbrdata_ref, sumbrcount_ref[,lines_found,
# lines_hit, f_found, f_hit, $b_found, $b_hit])
#
# Update the hash referenced by HASH_REF with the provided data references.
#
sub set_info_entry($$$$$$$$$;$$$$$$)
{
my $data_ref = $_[0];
$data_ref->{"test"} = $_[1];
$data_ref->{"sum"} = $_[2];
$data_ref->{"func"} = $_[3];
$data_ref->{"check"} = $_[4];
$data_ref->{"testfnc"} = $_[5];
$data_ref->{"sumfnc"} = $_[6];
$data_ref->{"testbr"} = $_[7];
$data_ref->{"sumbr"} = $_[8];
if (defined($_[9])) { $data_ref->{"found"} = $_[9]; }
if (defined($_[10])) { $data_ref->{"hit"} = $_[10]; }
if (defined($_[11])) { $data_ref->{"f_found"} = $_[11]; }
if (defined($_[12])) { $data_ref->{"f_hit"} = $_[12]; }
if (defined($_[13])) { $data_ref->{"b_found"} = $_[13]; }
if (defined($_[14])) { $data_ref->{"b_hit"} = $_[14]; }
}
#
# add_counts(data1_ref, data2_ref)
#
# DATA1_REF and DATA2_REF are references to hashes containing a mapping
#
# line number -> execution count
#
# Return a list (RESULT_REF, LINES_FOUND, LINES_HIT) where RESULT_REF
# is a reference to a hash containing the combined mapping in which
# execution counts are added.
#
sub add_counts($$)
{
my %data1 = %{$_[0]}; # Hash 1
my %data2 = %{$_[1]}; # Hash 2
my %result; # Resulting hash
my $line; # Current line iteration scalar
my $data1_count; # Count of line in hash1
my $data2_count; # Count of line in hash2
my $found = 0; # Total number of lines found
my $hit = 0; # Number of lines with a count > 0
foreach $line (keys(%data1))
{
$data1_count = $data1{$line};
$data2_count = $data2{$line};
# Add counts if present in both hashes
if (defined($data2_count)) { $data1_count += $data2_count; }
# Store sum in %result
$result{$line} = $data1_count;
$found++;
if ($data1_count > 0) { $hit++; }
}
# Add lines unique to data2
foreach $line (keys(%data2))
{
# Skip lines already in data1
if (defined($data1{$line})) { next; }
# Copy count from data2
$result{$line} = $data2{$line};
$found++;
if ($result{$line} > 0) { $hit++; }
}
return (\%result, $found, $hit);
}
#
# merge_checksums(ref1, ref2, filename)
#
# REF1 and REF2 are references to hashes containing a mapping
#
# line number -> checksum
#
# Merge checksum lists defined in REF1 and REF2 and return reference to
# resulting hash. Die if a checksum for a line is defined in both hashes
# but does not match.
#
sub merge_checksums($$$)
{
my $ref1 = $_[0];
my $ref2 = $_[1];
my $filename = $_[2];
my %result;
my $line;
foreach $line (keys(%{$ref1}))
{
if (defined($ref2->{$line}) &&
($ref1->{$line} ne $ref2->{$line}))
{
die("ERROR: checksum mismatch at $filename:$line\n");
}
$result{$line} = $ref1->{$line};
}
foreach $line (keys(%{$ref2}))
{
$result{$line} = $ref2->{$line};
}
return \%result;
}
#
# merge_func_data(funcdata1, funcdata2, filename)
#
sub merge_func_data($$$)
{
my ($funcdata1, $funcdata2, $filename) = @_;
my %result;
my $func;
if (defined($funcdata1)) {
%result = %{$funcdata1};
}
foreach $func (keys(%{$funcdata2})) {
my $line1 = $result{$func};
my $line2 = $funcdata2->{$func};
if (defined($line1) && ($line1 != $line2)) {
warn("WARNING: function data mismatch at ".
"$filename:$line2\n");
next;
}
$result{$func} = $line2;
}
return \%result;
}
#
# add_fnccount(fnccount1, fnccount2)
#
# Add function call count data. Return list (fnccount_added, f_found, f_hit)
#
sub add_fnccount($$)
{
my ($fnccount1, $fnccount2) = @_;
my %result;
my $f_found;
my $f_hit;
my $function;
if (defined($fnccount1)) {
%result = %{$fnccount1};
}
foreach $function (keys(%{$fnccount2})) {
$result{$function} += $fnccount2->{$function};
}
$f_found = scalar(keys(%result));
$f_hit = 0;
foreach $function (keys(%result)) {
if ($result{$function} > 0) {
$f_hit++;
}
}
return (\%result, $f_found, $f_hit);
}
#
# add_testfncdata(testfncdata1, testfncdata2)
#
# Add function call count data for several tests. Return reference to
# added_testfncdata.
#
sub add_testfncdata($$)
{
my ($testfncdata1, $testfncdata2) = @_;
my %result;
my $testname;
foreach $testname (keys(%{$testfncdata1})) {
if (defined($testfncdata2->{$testname})) {
my $fnccount;
# Function call count data for this testname exists
# in both data sets: merge
($fnccount) = add_fnccount(
$testfncdata1->{$testname},
$testfncdata2->{$testname});
$result{$testname} = $fnccount;
next;
}
# Function call count data for this testname is unique to
# data set 1: copy
$result{$testname} = $testfncdata1->{$testname};
}
# Add count data for testnames unique to data set 2
foreach $testname (keys(%{$testfncdata2})) {
if (!defined($result{$testname})) {
$result{$testname} = $testfncdata2->{$testname};
}
}
return \%result;
}
#
# brcount_to_db(brcount)
#
# Convert brcount data to the following format:
#
# db: line number -> block hash
# block hash: block number -> branch hash
# branch hash: branch number -> taken value
#
sub brcount_to_db($)
{
my ($brcount) = @_;
my $line;
my $db;
# Add branches from first count to database
foreach $line (keys(%{$brcount})) {
my $brdata = $brcount->{$line};
my $i;
my $num = br_ivec_len($brdata);
for ($i = 0; $i < $num; $i++) {
my ($block, $branch, $taken) = br_ivec_get($brdata, $i);
$db->{$line}->{$block}->{$branch} = $taken;
}
}
return $db;
}
#
# db_to_brcount(db)
#
# Convert branch coverage data back to brcount format.
#
sub db_to_brcount($)
{
my ($db) = @_;
my $line;
my $brcount = {};
my $br_found = 0;
my $br_hit = 0;
# Convert database back to brcount format
foreach $line (sort({$a <=> $b} keys(%{$db}))) {
my $ldata = $db->{$line};
my $brdata;
my $block;
foreach $block (sort({$a <=> $b} keys(%{$ldata}))) {
my $bdata = $ldata->{$block};
my $branch;
foreach $branch (sort({$a <=> $b} keys(%{$bdata}))) {
my $taken = $bdata->{$branch};
$br_found++;
$br_hit++ if ($taken ne "-" && $taken > 0);
$brdata = br_ivec_push($brdata, $block,
$branch, $taken);
}
}
$brcount->{$line} = $brdata;
}
return ($brcount, $br_found, $br_hit);
}
# combine_brcount(brcount1, brcount2, type)
#
# If add is BR_ADD, add branch coverage data and return list (brcount_added,
# br_found, br_hit). If add is BR_SUB, subtract the taken values of brcount2
# from brcount1 and return (brcount_sub, br_found, br_hit).
#
sub combine_brcount($$$)
{
my ($brcount1, $brcount2, $type) = @_;
my $line;
my $block;
my $branch;
my $taken;
my $db;
my $br_found = 0;
my $br_hit = 0;
my $result;
# Convert branches from first count to database
$db = brcount_to_db($brcount1);
# Combine values from database and second count
foreach $line (keys(%{$brcount2})) {
my $brdata = $brcount2->{$line};
my $num = br_ivec_len($brdata);
my $i;
for ($i = 0; $i < $num; $i++) {
($block, $branch, $taken) = br_ivec_get($brdata, $i);
my $new_taken = $db->{$line}->{$block}->{$branch};
if ($type == $BR_ADD) {
$new_taken = br_taken_add($new_taken, $taken);
} elsif ($type == $BR_SUB) {
$new_taken = br_taken_sub($new_taken, $taken);
}
$db->{$line}->{$block}->{$branch} = $new_taken
if (defined($new_taken));
}
}
# Convert database back to brcount format
($result, $br_found, $br_hit) = db_to_brcount($db);
return ($result, $br_found, $br_hit);
}
#
# add_testbrdata(testbrdata1, testbrdata2)
#
# Add branch coverage data for several tests. Return reference to
# added_testbrdata.
#
sub add_testbrdata($$)
{
my ($testbrdata1, $testbrdata2) = @_;
my %result;
my $testname;
foreach $testname (keys(%{$testbrdata1})) {
if (defined($testbrdata2->{$testname})) {
my $brcount;
# Branch coverage data for this testname exists
# in both data sets: add
($brcount) = combine_brcount(
$testbrdata1->{$testname},
$testbrdata2->{$testname}, $BR_ADD);
$result{$testname} = $brcount;
next;
}
# Branch coverage data for this testname is unique to
# data set 1: copy
$result{$testname} = $testbrdata1->{$testname};
}
# Add count data for testnames unique to data set 2
foreach $testname (keys(%{$testbrdata2})) {
if (!defined($result{$testname})) {
$result{$testname} = $testbrdata2->{$testname};
}
}
return \%result;
}
#
# combine_info_entries(entry_ref1, entry_ref2, filename)
#
# Combine .info data entry hashes referenced by ENTRY_REF1 and ENTRY_REF2.
# Return reference to resulting hash.
#
sub combine_info_entries($$$)
{
my $entry1 = $_[0]; # Reference to hash containing first entry
my $testdata1;
my $sumcount1;
my $funcdata1;
my $checkdata1;
my $testfncdata1;
my $sumfnccount1;
my $testbrdata1;
my $sumbrcount1;
my $entry2 = $_[1]; # Reference to hash containing second entry
my $testdata2;
my $sumcount2;
my $funcdata2;
my $checkdata2;
my $testfncdata2;
my $sumfnccount2;
my $testbrdata2;
my $sumbrcount2;
my %result; # Hash containing combined entry
my %result_testdata;
my $result_sumcount = {};
my $result_funcdata;
my $result_testfncdata;
my $result_sumfnccount;
my $result_testbrdata;
my $result_sumbrcount;
my $lines_found;
my $lines_hit;
my $f_found;
my $f_hit;
my $br_found;
my $br_hit;
my $testname;
my $filename = $_[2];
# Retrieve data
($testdata1, $sumcount1, $funcdata1, $checkdata1, $testfncdata1,
$sumfnccount1, $testbrdata1, $sumbrcount1) = get_info_entry($entry1);
($testdata2, $sumcount2, $funcdata2, $checkdata2, $testfncdata2,
$sumfnccount2, $testbrdata2, $sumbrcount2) = get_info_entry($entry2);
# Merge checksums
$checkdata1 = merge_checksums($checkdata1, $checkdata2, $filename);
# Combine funcdata
$result_funcdata = merge_func_data($funcdata1, $funcdata2, $filename);
# Combine function call count data
$result_testfncdata = add_testfncdata($testfncdata1, $testfncdata2);
($result_sumfnccount, $f_found, $f_hit) =
add_fnccount($sumfnccount1, $sumfnccount2);
# Combine branch coverage data
$result_testbrdata = add_testbrdata($testbrdata1, $testbrdata2);
($result_sumbrcount, $br_found, $br_hit) =
combine_brcount($sumbrcount1, $sumbrcount2, $BR_ADD);
# Combine testdata
foreach $testname (keys(%{$testdata1}))
{
if (defined($testdata2->{$testname}))
{
# testname is present in both entries, requires
# combination
($result_testdata{$testname}) =
add_counts($testdata1->{$testname},
$testdata2->{$testname});
}
else
{
# testname only present in entry1, add to result
$result_testdata{$testname} = $testdata1->{$testname};
}
# update sum count hash
($result_sumcount, $lines_found, $lines_hit) =
add_counts($result_sumcount,
$result_testdata{$testname});
}
foreach $testname (keys(%{$testdata2}))
{
# Skip testnames already covered by previous iteration
if (defined($testdata1->{$testname})) { next; }
# testname only present in entry2, add to result hash
$result_testdata{$testname} = $testdata2->{$testname};
# update sum count hash
($result_sumcount, $lines_found, $lines_hit) =
add_counts($result_sumcount,
$result_testdata{$testname});
}
# Calculate resulting sumcount
# Store result
set_info_entry(\%result, \%result_testdata, $result_sumcount,
$result_funcdata, $checkdata1, $result_testfncdata,
$result_sumfnccount, $result_testbrdata,
$result_sumbrcount, $lines_found, $lines_hit,
$f_found, $f_hit, $br_found, $br_hit);
return(\%result);
}
#
# combine_info_files(info_ref1, info_ref2)
#
# Combine .info data in hashes referenced by INFO_REF1 and INFO_REF2. Return
# reference to resulting hash.
#
sub combine_info_files($$)
{
my %hash1 = %{$_[0]};
my %hash2 = %{$_[1]};
my $filename;
foreach $filename (keys(%hash2))
{
if ($hash1{$filename})
{
# Entry already exists in hash1, combine them
$hash1{$filename} =
combine_info_entries($hash1{$filename},
$hash2{$filename},
$filename);
}
else
{
# Entry is unique in both hashes, simply add to
# resulting hash
$hash1{$filename} = $hash2{$filename};
}
}
return(\%hash1);
}
#
# add_traces()
#
sub add_traces()
{
my $total_trace;
my $current_trace;
my $tracefile;
my @result;
local *INFO_HANDLE;
info("Combining tracefiles.\n");
foreach $tracefile (@add_tracefile)
{
$current_trace = read_info_file($tracefile);
if ($total_trace)
{
$total_trace = combine_info_files($total_trace,
$current_trace);
}
else
{
$total_trace = $current_trace;
}
}
# Write combined data
if ($to_file)
{
info("Writing data to $output_filename\n");
open(INFO_HANDLE, ">", $output_filename)
or die("ERROR: cannot write to $output_filename!\n");
@result = write_info_file(*INFO_HANDLE, $total_trace);
close(*INFO_HANDLE);
}
else
{
@result = write_info_file(*STDOUT, $total_trace);
}
return @result;
}
#
# write_info_file(filehandle, data)
#
sub write_info_file(*$)
{
local *INFO_HANDLE = $_[0];
my %data = %{$_[1]};
my $source_file;
my $entry;
my $testdata;
my $sumcount;
my $funcdata;
my $checkdata;
my $testfncdata;
my $sumfnccount;
my $testbrdata;
my $sumbrcount;
my $testname;
my $line;
my $func;
my $testcount;
my $testfnccount;
my $testbrcount;
my $found;
my $hit;
my $f_found;
my $f_hit;
my $br_found;
my $br_hit;
my $ln_total_found = 0;
my $ln_total_hit = 0;
my $fn_total_found = 0;
my $fn_total_hit = 0;
my $br_total_found = 0;
my $br_total_hit = 0;
foreach $source_file (sort(keys(%data)))
{
$entry = $data{$source_file};
($testdata, $sumcount, $funcdata, $checkdata, $testfncdata,
$sumfnccount, $testbrdata, $sumbrcount, $found, $hit,
$f_found, $f_hit, $br_found, $br_hit) =
get_info_entry($entry);
# Add to totals
$ln_total_found += $found;
$ln_total_hit += $hit;
$fn_total_found += $f_found;
$fn_total_hit += $f_hit;
$br_total_found += $br_found;
$br_total_hit += $br_hit;
foreach $testname (sort(keys(%{$testdata})))
{
$testcount = $testdata->{$testname};
$testfnccount = $testfncdata->{$testname};
$testbrcount = $testbrdata->{$testname};
$found = 0;
$hit = 0;
print(INFO_HANDLE "TN:$testname\n");
print(INFO_HANDLE "SF:$source_file\n");
# Write function related data
foreach $func (
sort({$funcdata->{$a} <=> $funcdata->{$b}}
keys(%{$funcdata})))
{
print(INFO_HANDLE "FN:".$funcdata->{$func}.
",$func\n");
}
foreach $func (keys(%{$testfnccount})) {
print(INFO_HANDLE "FNDA:".
$testfnccount->{$func}.
",$func\n");
}
($f_found, $f_hit) =
get_func_found_and_hit($testfnccount);
print(INFO_HANDLE "FNF:$f_found\n");
print(INFO_HANDLE "FNH:$f_hit\n");
# Write branch related data
$br_found = 0;
$br_hit = 0;
foreach $line (sort({$a <=> $b}
keys(%{$testbrcount}))) {
my $brdata = $testbrcount->{$line};
my $num = br_ivec_len($brdata);
my $i;
for ($i = 0; $i < $num; $i++) {
my ($block, $branch, $taken) =
br_ivec_get($brdata, $i);
print(INFO_HANDLE "BRDA:$line,$block,".
"$branch,$taken\n");
$br_found++;
$br_hit++ if ($taken ne '-' &&
$taken > 0);
}
}
if ($br_found > 0) {
print(INFO_HANDLE "BRF:$br_found\n");
print(INFO_HANDLE "BRH:$br_hit\n");
}
# Write line related data
foreach $line (sort({$a <=> $b} keys(%{$testcount})))
{
print(INFO_HANDLE "DA:$line,".
$testcount->{$line}.
(defined($checkdata->{$line}) &&
$checksum ?
",".$checkdata->{$line} : "")."\n");
$found++;
if ($testcount->{$line} > 0)
{
$hit++;
}
}
print(INFO_HANDLE "LF:$found\n");
print(INFO_HANDLE "LH:$hit\n");
print(INFO_HANDLE "end_of_record\n");
}
}
return ($ln_total_found, $ln_total_hit, $fn_total_found, $fn_total_hit,
$br_total_found, $br_total_hit);
}
#
# transform_pattern(pattern)
#
# Transform shell wildcard expression to equivalent Perl regular expression.
# Return transformed pattern.
#
sub transform_pattern($)
{
my $pattern = $_[0];
# Escape special chars
$pattern =~ s/\\/\\\\/g;
$pattern =~ s/\//\\\//g;
$pattern =~ s/\^/\\\^/g;
$pattern =~ s/\$/\\\$/g;
$pattern =~ s/\(/\\\(/g;
$pattern =~ s/\)/\\\)/g;
$pattern =~ s/\[/\\\[/g;
$pattern =~ s/\]/\\\]/g;
$pattern =~ s/\{/\\\{/g;
$pattern =~ s/\}/\\\}/g;
$pattern =~ s/\./\\\./g;
$pattern =~ s/\,/\\\,/g;
$pattern =~ s/\|/\\\|/g;
$pattern =~ s/\+/\\\+/g;
$pattern =~ s/\!/\\\!/g;
# Transform ? => (.) and * => (.*)
$pattern =~ s/\*/\(\.\*\)/g;
$pattern =~ s/\?/\(\.\)/g;
return $pattern;
}
#
# extract()
#
sub extract()
{
my $data = read_info_file($extract);
my $filename;
my $keep;
my $pattern;
my @pattern_list;
my $extracted = 0;
my @result;
local *INFO_HANDLE;
# Need perlreg expressions instead of shell pattern
@pattern_list = map({ transform_pattern($_); } @ARGV);
# Filter out files which do not match any pattern
foreach $filename (sort(keys(%{$data})))
{
$keep = 0;
foreach $pattern (@pattern_list)
{
$keep ||= ($filename =~ (/^$pattern$/));
}
if (!$keep)
{
delete($data->{$filename});
}
else
{
info("Extracting $filename\n"),
$extracted++;
}
}
# Write extracted data
if ($to_file)
{
info("Extracted $extracted files\n");
info("Writing data to $output_filename\n");
open(INFO_HANDLE, ">", $output_filename)
or die("ERROR: cannot write to $output_filename!\n");
@result = write_info_file(*INFO_HANDLE, $data);
close(*INFO_HANDLE);
}
else
{
@result = write_info_file(*STDOUT, $data);
}
return @result;
}
#
# remove()
#
sub remove()
{
my $data = read_info_file($remove);
my $filename;
my $match_found;
my $pattern;
my @pattern_list;
my $removed = 0;
my @result;
local *INFO_HANDLE;
# Need perlreg expressions instead of shell pattern
@pattern_list = map({ transform_pattern($_); } @ARGV);
# Filter out files that match the pattern
foreach $filename (sort(keys(%{$data})))
{
$match_found = 0;
foreach $pattern (@pattern_list)
{
$match_found ||= ($filename =~ (/$pattern$/));
}
if ($match_found)
{
delete($data->{$filename});
info("Removing $filename\n"),
$removed++;
}
}
# Write data
if ($to_file)
{
info("Deleted $removed files\n");
info("Writing data to $output_filename\n");
open(INFO_HANDLE, ">", $output_filename)
or die("ERROR: cannot write to $output_filename!\n");
@result = write_info_file(*INFO_HANDLE, $data);
close(*INFO_HANDLE);
}
else
{
@result = write_info_file(*STDOUT, $data);
}
return @result;
}
# get_prefix(max_width, max_percentage_too_long, path_list)
#
# Return a path prefix that satisfies the following requirements:
# - is shared by more paths in path_list than any other prefix
# - the percentage of paths which would exceed the given max_width length
# after applying the prefix does not exceed max_percentage_too_long
#
# If multiple prefixes satisfy all requirements, the longest prefix is
# returned. Return an empty string if no prefix could be found.
sub get_prefix($$@)
{
my ($max_width, $max_long, @path_list) = @_;
my $path;
my $ENTRY_NUM = 0;
my $ENTRY_LONG = 1;
my %prefix;
# Build prefix hash
foreach $path (@path_list) {
my ($v, $d, $f) = splitpath($path);
my @dirs = splitdir($d);
my $p_len = length($path);
my $i;
# Remove trailing '/'
pop(@dirs) if ($dirs[scalar(@dirs) - 1] eq '');
for ($i = 0; $i < scalar(@dirs); $i++) {
my $subpath = catpath($v, catdir(@dirs[0..$i]), '');
my $entry = $prefix{$subpath};
$entry = [ 0, 0 ] if (!defined($entry));
$entry->[$ENTRY_NUM]++;
if (($p_len - length($subpath) - 1) > $max_width) {
$entry->[$ENTRY_LONG]++;
}
$prefix{$subpath} = $entry;
}
}
# Find suitable prefix (sort descending by two keys: 1. number of
# entries covered by a prefix, 2. length of prefix)
foreach $path (sort {($prefix{$a}->[$ENTRY_NUM] ==
$prefix{$b}->[$ENTRY_NUM]) ?
length($b) <=> length($a) :
$prefix{$b}->[$ENTRY_NUM] <=>
$prefix{$a}->[$ENTRY_NUM]}
keys(%prefix)) {
my ($num, $long) = @{$prefix{$path}};
# Check for additional requirement: number of filenames
# that would be too long may not exceed a certain percentage
if ($long <= $num * $max_long / 100) {
return $path;
}
}
return "";
}
#
# shorten_filename(filename, width)
#
# Truncate filename if it is longer than width characters.
#
sub shorten_filename($$)
{
my ($filename, $width) = @_;
my $l = length($filename);
my $s;
my $e;
return $filename if ($l <= $width);
$e = int(($width - 3) / 2);
$s = $width - 3 - $e;
return substr($filename, 0, $s).'...'.substr($filename, $l - $e);
}
sub shorten_number($$)
{
my ($number, $width) = @_;
my $result = sprintf("%*d", $width, $number);
return $result if (length($result) <= $width);
$number = $number / 1000;
return $result if (length($result) <= $width);
$result = sprintf("%*dk", $width - 1, $number);
return $result if (length($result) <= $width);
$number = $number / 1000;
$result = sprintf("%*dM", $width - 1, $number);
return $result if (length($result) <= $width);
return '#';
}
sub shorten_rate($$$)
{
my ($hit, $found, $width) = @_;
my $result = rate($hit, $found, "%", 1, $width);
return $result if (length($result) <= $width);
$result = rate($hit, $found, "%", 0, $width);
return $result if (length($result) <= $width);
return "#";
}
#
# list()
#
sub list()
{
my $data = read_info_file($list);
my $filename;
my $found;
my $hit;
my $entry;
my $fn_found;
my $fn_hit;
my $br_found;
my $br_hit;
my $total_found = 0;
my $total_hit = 0;
my $fn_total_found = 0;
my $fn_total_hit = 0;
my $br_total_found = 0;
my $br_total_hit = 0;
my $prefix;
my $strlen = length("Filename");
my $format;
my $heading1;
my $heading2;
my @footer;
my $barlen;
my $rate;
my $fnrate;
my $brrate;
my $lastpath;
my $F_LN_NUM = 0;
my $F_LN_RATE = 1;
my $F_FN_NUM = 2;
my $F_FN_RATE = 3;
my $F_BR_NUM = 4;
my $F_BR_RATE = 5;
my @fwidth_narrow = (5, 5, 3, 5, 4, 5);
my @fwidth_wide = (6, 5, 5, 5, 6, 5);
my @fwidth = @fwidth_wide;
my $w;
my $max_width = $opt_list_width;
my $max_long = $opt_list_truncate_max;
my $fwidth_narrow_length;
my $fwidth_wide_length;
my $got_prefix = 0;
my $root_prefix = 0;
# Calculate total width of narrow fields
$fwidth_narrow_length = 0;
foreach $w (@fwidth_narrow) {
$fwidth_narrow_length += $w + 1;
}
# Calculate total width of wide fields
$fwidth_wide_length = 0;
foreach $w (@fwidth_wide) {
$fwidth_wide_length += $w + 1;
}
# Get common file path prefix
$prefix = get_prefix($max_width - $fwidth_narrow_length, $max_long,
keys(%{$data}));
$root_prefix = 1 if ($prefix eq rootdir());
$got_prefix = 1 if (length($prefix) > 0);
$prefix =~ s/\/$//;
# Get longest filename length
foreach $filename (keys(%{$data})) {
if (!$opt_list_full_path) {
if (!$got_prefix || !$root_prefix &&
!($filename =~ s/^\Q$prefix\/\E//)) {
my ($v, $d, $f) = splitpath($filename);
$filename = $f;
}
}
# Determine maximum length of entries
if (length($filename) > $strlen) {
$strlen = length($filename)
}
}
if (!$opt_list_full_path) {
my $blanks;
$w = $fwidth_wide_length;
# Check if all columns fit into max_width characters
if ($strlen + $fwidth_wide_length > $max_width) {
# Use narrow fields
@fwidth = @fwidth_narrow;
$w = $fwidth_narrow_length;
if (($strlen + $fwidth_narrow_length) > $max_width) {
# Truncate filenames at max width
$strlen = $max_width - $fwidth_narrow_length;
}
}
# Add some blanks between filename and fields if possible
$blanks = int($strlen * 0.5);
$blanks = 4 if ($blanks < 4);
$blanks = 8 if ($blanks > 8);
if (($strlen + $w + $blanks) < $max_width) {
$strlen += $blanks;
} else {
$strlen = $max_width - $w;
}
}
# Filename
$w = $strlen;
$format = "%-${w}s|";
$heading1 = sprintf("%*s|", $w, "");
$heading2 = sprintf("%-*s|", $w, "Filename");
$barlen = $w + 1;
# Line coverage rate
$w = $fwidth[$F_LN_RATE];
$format .= "%${w}s ";
$heading1 .= sprintf("%-*s |", $w + $fwidth[$F_LN_NUM],
"Lines");
$heading2 .= sprintf("%-*s ", $w, "Rate");
$barlen += $w + 1;
# Number of lines
$w = $fwidth[$F_LN_NUM];
$format .= "%${w}s|";
$heading2 .= sprintf("%*s|", $w, "Num");
$barlen += $w + 1;
# Function coverage rate
$w = $fwidth[$F_FN_RATE];
$format .= "%${w}s ";
$heading1 .= sprintf("%-*s|", $w + $fwidth[$F_FN_NUM] + 1,
"Functions");
$heading2 .= sprintf("%-*s ", $w, "Rate");
$barlen += $w + 1;
# Number of functions
$w = $fwidth[$F_FN_NUM];
$format .= "%${w}s|";
$heading2 .= sprintf("%*s|", $w, "Num");
$barlen += $w + 1;
# Branch coverage rate
$w = $fwidth[$F_BR_RATE];
$format .= "%${w}s ";
$heading1 .= sprintf("%-*s", $w + $fwidth[$F_BR_NUM] + 1,
"Branches");
$heading2 .= sprintf("%-*s ", $w, "Rate");
$barlen += $w + 1;
# Number of branches
$w = $fwidth[$F_BR_NUM];
$format .= "%${w}s";
$heading2 .= sprintf("%*s", $w, "Num");
$barlen += $w;
# Line end
$format .= "\n";
$heading1 .= "\n";
$heading2 .= "\n";
# Print heading
print($heading1);
print($heading2);
print(("="x$barlen)."\n");
# Print per file information
foreach $filename (sort(keys(%{$data})))
{
my @file_data;
my $print_filename = $filename;
$entry = $data->{$filename};
if (!$opt_list_full_path) {
my $p;
$print_filename = $filename;
if (!$got_prefix || !$root_prefix &&
!($print_filename =~ s/^\Q$prefix\/\E//)) {
my ($v, $d, $f) = splitpath($filename);
$p = catpath($v, $d, "");
$p =~ s/\/$//;
$print_filename = $f;
} else {
$p = $prefix;
}
if (!defined($lastpath) || $lastpath ne $p) {
print("\n") if (defined($lastpath));
$lastpath = $p;
print("[$lastpath/]\n") if (!$root_prefix);
}
$print_filename = shorten_filename($print_filename,
$strlen);
}
(undef, undef, undef, undef, undef, undef, undef, undef,
$found, $hit, $fn_found, $fn_hit, $br_found, $br_hit) =
get_info_entry($entry);
# Assume zero count if there is no function data for this file
if (!defined($fn_found) || !defined($fn_hit)) {
$fn_found = 0;
$fn_hit = 0;
}
# Assume zero count if there is no branch data for this file
if (!defined($br_found) || !defined($br_hit)) {
$br_found = 0;
$br_hit = 0;
}
# Add line coverage totals
$total_found += $found;
$total_hit += $hit;
# Add function coverage totals
$fn_total_found += $fn_found;
$fn_total_hit += $fn_hit;
# Add branch coverage totals
$br_total_found += $br_found;
$br_total_hit += $br_hit;
# Determine line coverage rate for this file
$rate = shorten_rate($hit, $found, $fwidth[$F_LN_RATE]);
# Determine function coverage rate for this file
$fnrate = shorten_rate($fn_hit, $fn_found, $fwidth[$F_FN_RATE]);
# Determine branch coverage rate for this file
$brrate = shorten_rate($br_hit, $br_found, $fwidth[$F_BR_RATE]);
# Assemble line parameters
push(@file_data, $print_filename);
push(@file_data, $rate);
push(@file_data, shorten_number($found, $fwidth[$F_LN_NUM]));
push(@file_data, $fnrate);
push(@file_data, shorten_number($fn_found, $fwidth[$F_FN_NUM]));
push(@file_data, $brrate);
push(@file_data, shorten_number($br_found, $fwidth[$F_BR_NUM]));
# Print assembled line
printf($format, @file_data);
}
# Determine total line coverage rate
$rate = shorten_rate($total_hit, $total_found, $fwidth[$F_LN_RATE]);
# Determine total function coverage rate
$fnrate = shorten_rate($fn_total_hit, $fn_total_found,
$fwidth[$F_FN_RATE]);
# Determine total branch coverage rate
$brrate = shorten_rate($br_total_hit, $br_total_found,
$fwidth[$F_BR_RATE]);
# Print separator
print(("="x$barlen)."\n");
# Assemble line parameters
push(@footer, sprintf("%*s", $strlen, "Total:"));
push(@footer, $rate);
push(@footer, shorten_number($total_found, $fwidth[$F_LN_NUM]));
push(@footer, $fnrate);
push(@footer, shorten_number($fn_total_found, $fwidth[$F_FN_NUM]));
push(@footer, $brrate);
push(@footer, shorten_number($br_total_found, $fwidth[$F_BR_NUM]));
# Print assembled line
printf($format, @footer);
}
#
# get_common_filename(filename1, filename2)
#
# Check for filename components which are common to FILENAME1 and FILENAME2.
# Upon success, return
#
# (common, path1, path2)
#
# or 'undef' in case there are no such parts.
#
sub get_common_filename($$)
{
my @list1 = split("/", $_[0]);
my @list2 = split("/", $_[1]);
my @result;
# Work in reverse order, i.e. beginning with the filename itself
while (@list1 && @list2 && ($list1[$#list1] eq $list2[$#list2]))
{
unshift(@result, pop(@list1));
pop(@list2);
}
# Did we find any similarities?
if (scalar(@result) > 0)
{
return (join("/", @result), join("/", @list1),
join("/", @list2));
}
else
{
return undef;
}
}
#
# strip_directories($path, $depth)
#
# Remove DEPTH leading directory levels from PATH.
#
sub strip_directories($$)
{
my $filename = $_[0];
my $depth = $_[1];
my $i;
if (!defined($depth) || ($depth < 1))
{
return $filename;
}
for ($i = 0; $i < $depth; $i++)
{
$filename =~ s/^[^\/]*\/+(.*)$/$1/;
}
return $filename;
}
#
# read_diff(filename)
#
# Read diff output from FILENAME to memory. The diff file has to follow the
# format generated by 'diff -u'. Returns a list of hash references:
#
# (mapping, path mapping)
#
# mapping: filename -> reference to line hash
# line hash: line number in new file -> corresponding line number in old file
#
# path mapping: filename -> old filename
#
# Die in case of error.
#
sub read_diff($)
{
my $diff_file = $_[0]; # Name of diff file
my %diff; # Resulting mapping filename -> line hash
my %paths; # Resulting mapping old path -> new path
my $mapping; # Reference to current line hash
my $line; # Contents of current line
my $num_old; # Current line number in old file
my $num_new; # Current line number in new file
my $file_old; # Name of old file in diff section
my $file_new; # Name of new file in diff section
my $filename; # Name of common filename of diff section
my $in_block = 0; # Non-zero while we are inside a diff block
local *HANDLE; # File handle for reading the diff file
info("Reading diff $diff_file\n");
# Check if file exists and is readable
stat($diff_file);
if (!(-r _))
{
die("ERROR: cannot read file $diff_file!\n");
}
# Check if this is really a plain file
if (!(-f _))
{
die("ERROR: not a plain file: $diff_file!\n");
}
# Check for .gz extension
if ($diff_file =~ /\.gz$/)
{
# Check for availability of GZIP tool
system_no_output(1, "gunzip", "-h")
and die("ERROR: gunzip command not available!\n");
# Check integrity of compressed file
system_no_output(1, "gunzip", "-t", $diff_file)
and die("ERROR: integrity check failed for ".
"compressed file $diff_file!\n");
# Open compressed file
open(HANDLE, "-|", "gunzip -c '$diff_file'")
or die("ERROR: cannot start gunzip to decompress ".
"file $_[0]!\n");
}
else
{
# Open decompressed file
open(HANDLE, "<", $diff_file)
or die("ERROR: cannot read file $_[0]!\n");
}
# Parse diff file line by line
while (<HANDLE>)
{
chomp($_);
$line = $_;
foreach ($line)
{
# Filename of old file:
# --- <filename> <date>
/^--- (\S+)/ && do
{
$file_old = strip_directories($1, $strip);
last;
};
# Filename of new file:
# +++ <filename> <date>
/^\+\+\+ (\S+)/ && do
{
# Add last file to resulting hash
if ($filename)
{
my %new_hash;
$diff{$filename} = $mapping;
$mapping = \%new_hash;
}
$file_new = strip_directories($1, $strip);
$filename = $file_old;
$paths{$filename} = $file_new;
$num_old = 1;
$num_new = 1;
last;
};
# Start of diff block:
# @@ -old_start,old_num, +new_start,new_num @@
/^\@\@\s+-(\d+),(\d+)\s+\+(\d+),(\d+)\s+\@\@$/ && do
{
$in_block = 1;
while ($num_old < $1)
{
$mapping->{$num_new} = $num_old;
$num_old++;
$num_new++;
}
last;
};
# Unchanged line
# <line starts with blank>
/^ / && do
{
if ($in_block == 0)
{
last;
}
$mapping->{$num_new} = $num_old;
$num_old++;
$num_new++;
last;
};
# Line as seen in old file
# <line starts with '-'>
/^-/ && do
{
if ($in_block == 0)
{
last;
}
$num_old++;
last;
};
# Line as seen in new file
# <line starts with '+'>
/^\+/ && do
{
if ($in_block == 0)
{
last;
}
$num_new++;
last;
};
# Empty line
/^$/ && do
{
if ($in_block == 0)
{
last;
}
$mapping->{$num_new} = $num_old;
$num_old++;
$num_new++;
last;
};
}
}
close(HANDLE);
# Add final diff file section to resulting hash
if ($filename)
{
$diff{$filename} = $mapping;
}
if (!%diff)
{
die("ERROR: no valid diff data found in $diff_file!\n".
"Make sure to use 'diff -u' when generating the diff ".
"file.\n");
}
return (\%diff, \%paths);
}
#
# apply_diff($count_data, $line_hash)
#
# Transform count data using a mapping of lines:
#
# $count_data: reference to hash: line number -> data
# $line_hash: reference to hash: line number new -> line number old
#
# Return a reference to transformed count data.
#
sub apply_diff($$)
{
my $count_data = $_[0]; # Reference to data hash: line -> hash
my $line_hash = $_[1]; # Reference to line hash: new line -> old line
my %result; # Resulting hash
my $last_new = 0; # Last new line number found in line hash
my $last_old = 0; # Last old line number found in line hash
# Iterate all new line numbers found in the diff
foreach (sort({$a <=> $b} keys(%{$line_hash})))
{
$last_new = $_;
$last_old = $line_hash->{$last_new};
# Is there data associated with the corresponding old line?
if (defined($count_data->{$line_hash->{$_}}))
{
# Copy data to new hash with a new line number
$result{$_} = $count_data->{$line_hash->{$_}};
}
}
# Transform all other lines which come after the last diff entry
foreach (sort({$a <=> $b} keys(%{$count_data})))
{
if ($_ <= $last_old)
{
# Skip lines which were covered by line hash
next;
}
# Copy data to new hash with an offset
$result{$_ + ($last_new - $last_old)} = $count_data->{$_};
}
return \%result;
}
#
# apply_diff_to_brcount(brcount, linedata)
#
# Adjust line numbers of branch coverage data according to linedata.
#
sub apply_diff_to_brcount($$)
{
my ($brcount, $linedata) = @_;
my $db;
# Convert brcount to db format
$db = brcount_to_db($brcount);
# Apply diff to db format
$db = apply_diff($db, $linedata);
# Convert db format back to brcount format
($brcount) = db_to_brcount($db);
return $brcount;
}
#
# get_hash_max(hash_ref)
#
# Return the highest integer key from hash.
#
sub get_hash_max($)
{
my ($hash) = @_;
my $max;
foreach (keys(%{$hash})) {
if (!defined($max)) {
$max = $_;
} elsif ($hash->{$_} > $max) {
$max = $_;
}
}
return $max;
}
sub get_hash_reverse($)
{
my ($hash) = @_;
my %result;
foreach (keys(%{$hash})) {
$result{$hash->{$_}} = $_;
}
return \%result;
}
#
# apply_diff_to_funcdata(funcdata, line_hash)
#
sub apply_diff_to_funcdata($$)
{
my ($funcdata, $linedata) = @_;
my $last_new = get_hash_max($linedata);
my $last_old = $linedata->{$last_new};
my $func;
my %result;
my $line_diff = get_hash_reverse($linedata);
foreach $func (keys(%{$funcdata})) {
my $line = $funcdata->{$func};
if (defined($line_diff->{$line})) {
$result{$func} = $line_diff->{$line};
} elsif ($line > $last_old) {
$result{$func} = $line + $last_new - $last_old;
}
}
return \%result;
}
#
# get_line_hash($filename, $diff_data, $path_data)
#
# Find line hash in DIFF_DATA which matches FILENAME. On success, return list
# line hash. or undef in case of no match. Die if more than one line hashes in
# DIFF_DATA match.
#
sub get_line_hash($$$)
{
my $filename = $_[0];
my $diff_data = $_[1];
my $path_data = $_[2];
my $conversion;
my $old_path;
my $new_path;
my $diff_name;
my $common;
my $old_depth;
my $new_depth;
# Remove trailing slash from diff path
$diff_path =~ s/\/$//;
foreach (keys(%{$diff_data}))
{
my $sep = "";
$sep = '/' if (!/^\//);
# Try to match diff filename with filename
if ($filename =~ /^\Q$diff_path$sep$_\E$/)
{
if ($diff_name)
{
# Two files match, choose the more specific one
# (the one with more path components)
$old_depth = ($diff_name =~ tr/\///);
$new_depth = (tr/\///);
if ($old_depth == $new_depth)
{
die("ERROR: diff file contains ".
"ambiguous entries for ".
"$filename\n");
}
elsif ($new_depth > $old_depth)
{
$diff_name = $_;
}
}
else
{
$diff_name = $_;
}
};
}
if ($diff_name)
{
# Get converted path
if ($filename =~ /^(.*)$diff_name$/)
{
($common, $old_path, $new_path) =
get_common_filename($filename,
$1.$path_data->{$diff_name});
}
return ($diff_data->{$diff_name}, $old_path, $new_path);
}
else
{
return undef;
}
}
#
# convert_paths(trace_data, path_conversion_data)
#
# Rename all paths in TRACE_DATA which show up in PATH_CONVERSION_DATA.
#
sub convert_paths($$)
{
my $trace_data = $_[0];
my $path_conversion_data = $_[1];
my $filename;
my $new_path;
if (scalar(keys(%{$path_conversion_data})) == 0)
{
info("No path conversion data available.\n");
return;
}
# Expand path conversion list
foreach $filename (keys(%{$path_conversion_data}))
{
$new_path = $path_conversion_data->{$filename};
while (($filename =~ s/^(.*)\/[^\/]+$/$1/) &&
($new_path =~ s/^(.*)\/[^\/]+$/$1/) &&
($filename ne $new_path))
{
$path_conversion_data->{$filename} = $new_path;
}
}
# Adjust paths
FILENAME: foreach $filename (keys(%{$trace_data}))
{
# Find a path in our conversion table that matches, starting
# with the longest path
foreach (sort({length($b) <=> length($a)}
keys(%{$path_conversion_data})))
{
# Is this path a prefix of our filename?
if (!($filename =~ /^$_(.*)$/))
{
next;
}
$new_path = $path_conversion_data->{$_}.$1;
# Make sure not to overwrite an existing entry under
# that path name
if ($trace_data->{$new_path})
{
# Need to combine entries
$trace_data->{$new_path} =
combine_info_entries(
$trace_data->{$filename},
$trace_data->{$new_path},
$filename);
}
else
{
# Simply rename entry
$trace_data->{$new_path} =
$trace_data->{$filename};
}
delete($trace_data->{$filename});
next FILENAME;
}
info("No conversion available for filename $filename\n");
}
}
#
# sub adjust_fncdata(funcdata, testfncdata, sumfnccount)
#
# Remove function call count data from testfncdata and sumfnccount which
# is no longer present in funcdata.
#
sub adjust_fncdata($$$)
{
my ($funcdata, $testfncdata, $sumfnccount) = @_;
my $testname;
my $func;
my $f_found;
my $f_hit;
# Remove count data in testfncdata for functions which are no longer
# in funcdata
foreach $testname (%{$testfncdata}) {
my $fnccount = $testfncdata->{$testname};
foreach $func (%{$fnccount}) {
if (!defined($funcdata->{$func})) {
delete($fnccount->{$func});
}
}
}
# Remove count data in sumfnccount for functions which are no longer
# in funcdata
foreach $func (%{$sumfnccount}) {
if (!defined($funcdata->{$func})) {
delete($sumfnccount->{$func});
}
}
}
#
# get_func_found_and_hit(sumfnccount)
#
# Return (f_found, f_hit) for sumfnccount
#
sub get_func_found_and_hit($)
{
my ($sumfnccount) = @_;
my $function;
my $f_found;
my $f_hit;
$f_found = scalar(keys(%{$sumfnccount}));
$f_hit = 0;
foreach $function (keys(%{$sumfnccount})) {
if ($sumfnccount->{$function} > 0) {
$f_hit++;
}
}
return ($f_found, $f_hit);
}
#
# diff()
#
sub diff()
{
my $trace_data = read_info_file($diff);
my $diff_data;
my $path_data;
my $old_path;
my $new_path;
my %path_conversion_data;
my $filename;
my $line_hash;
my $new_name;
my $entry;
my $testdata;
my $testname;
my $sumcount;
my $funcdata;
my $checkdata;
my $testfncdata;
my $sumfnccount;
my $testbrdata;
my $sumbrcount;
my $found;
my $hit;
my $f_found;
my $f_hit;
my $br_found;
my $br_hit;
my $converted = 0;
my $unchanged = 0;
my @result;
local *INFO_HANDLE;
($diff_data, $path_data) = read_diff($ARGV[0]);
foreach $filename (sort(keys(%{$trace_data})))
{
# Find a diff section corresponding to this file
($line_hash, $old_path, $new_path) =
get_line_hash($filename, $diff_data, $path_data);
if (!$line_hash)
{
# There's no diff section for this file
$unchanged++;
next;
}
$converted++;
if ($old_path && $new_path && ($old_path ne $new_path))
{
$path_conversion_data{$old_path} = $new_path;
}
# Check for deleted files
if (scalar(keys(%{$line_hash})) == 0)
{
info("Removing $filename\n");
delete($trace_data->{$filename});
next;
}
info("Converting $filename\n");
$entry = $trace_data->{$filename};
($testdata, $sumcount, $funcdata, $checkdata, $testfncdata,
$sumfnccount, $testbrdata, $sumbrcount) =
get_info_entry($entry);
# Convert test data
foreach $testname (keys(%{$testdata}))
{
# Adjust line numbers of line coverage data
$testdata->{$testname} =
apply_diff($testdata->{$testname}, $line_hash);
# Adjust line numbers of branch coverage data
$testbrdata->{$testname} =
apply_diff_to_brcount($testbrdata->{$testname},
$line_hash);
# Remove empty sets of test data
if (scalar(keys(%{$testdata->{$testname}})) == 0)
{
delete($testdata->{$testname});
delete($testfncdata->{$testname});
delete($testbrdata->{$testname});
}
}
# Rename test data to indicate conversion
foreach $testname (keys(%{$testdata}))
{
# Skip testnames which already contain an extension
if ($testname =~ /,[^,]+$/)
{
next;
}
# Check for name conflict
if (defined($testdata->{$testname.",diff"}))
{
# Add counts
($testdata->{$testname}) = add_counts(
$testdata->{$testname},
$testdata->{$testname.",diff"});
delete($testdata->{$testname.",diff"});
# Add function call counts
($testfncdata->{$testname}) = add_fnccount(
$testfncdata->{$testname},
$testfncdata->{$testname.",diff"});
delete($testfncdata->{$testname.",diff"});
# Add branch counts
($testbrdata->{$testname}) = combine_brcount(
$testbrdata->{$testname},
$testbrdata->{$testname.",diff"},
$BR_ADD);
delete($testbrdata->{$testname.",diff"});
}
# Move test data to new testname
$testdata->{$testname.",diff"} = $testdata->{$testname};
delete($testdata->{$testname});
# Move function call count data to new testname
$testfncdata->{$testname.",diff"} =
$testfncdata->{$testname};
delete($testfncdata->{$testname});
# Move branch count data to new testname
$testbrdata->{$testname.",diff"} =
$testbrdata->{$testname};
delete($testbrdata->{$testname});
}
# Convert summary of test data
$sumcount = apply_diff($sumcount, $line_hash);
# Convert function data
$funcdata = apply_diff_to_funcdata($funcdata, $line_hash);
# Convert branch coverage data
$sumbrcount = apply_diff_to_brcount($sumbrcount, $line_hash);
# Update found/hit numbers
# Convert checksum data
$checkdata = apply_diff($checkdata, $line_hash);
# Convert function call count data
adjust_fncdata($funcdata, $testfncdata, $sumfnccount);
($f_found, $f_hit) = get_func_found_and_hit($sumfnccount);
($br_found, $br_hit) = get_br_found_and_hit($sumbrcount);
# Update found/hit numbers
$found = 0;
$hit = 0;
foreach (keys(%{$sumcount}))
{
$found++;
if ($sumcount->{$_} > 0)
{
$hit++;
}
}
if ($found > 0)
{
# Store converted entry
set_info_entry($entry, $testdata, $sumcount, $funcdata,
$checkdata, $testfncdata, $sumfnccount,
$testbrdata, $sumbrcount, $found, $hit,
$f_found, $f_hit, $br_found, $br_hit);
}
else
{
# Remove empty data set
delete($trace_data->{$filename});
}
}
# Convert filenames as well if requested
if ($convert_filenames)
{
convert_paths($trace_data, \%path_conversion_data);
}
info("$converted entr".($converted != 1 ? "ies" : "y")." converted, ".
"$unchanged entr".($unchanged != 1 ? "ies" : "y")." left ".
"unchanged.\n");
# Write data
if ($to_file)
{
info("Writing data to $output_filename\n");
open(INFO_HANDLE, ">", $output_filename)
or die("ERROR: cannot write to $output_filename!\n");
@result = write_info_file(*INFO_HANDLE, $trace_data);
close(*INFO_HANDLE);
}
else
{
@result = write_info_file(*STDOUT, $trace_data);
}
return @result;
}
#
# summary()
#
sub summary()
{
my $filename;
my $current;
my $total;
my $ln_total_found;
my $ln_total_hit;
my $fn_total_found;
my $fn_total_hit;
my $br_total_found;
my $br_total_hit;
# Read and combine trace files
foreach $filename (@opt_summary) {
$current = read_info_file($filename);
if (!defined($total)) {
$total = $current;
} else {
$total = combine_info_files($total, $current);
}
}
# Calculate coverage data
foreach $filename (keys(%{$total}))
{
my $entry = $total->{$filename};
my $ln_found;
my $ln_hit;
my $fn_found;
my $fn_hit;
my $br_found;
my $br_hit;
(undef, undef, undef, undef, undef, undef, undef, undef,
$ln_found, $ln_hit, $fn_found, $fn_hit, $br_found,
$br_hit) = get_info_entry($entry);
# Add to totals
$ln_total_found += $ln_found;
$ln_total_hit += $ln_hit;
$fn_total_found += $fn_found;
$fn_total_hit += $fn_hit;
$br_total_found += $br_found;
$br_total_hit += $br_hit;
}
return ($ln_total_found, $ln_total_hit, $fn_total_found, $fn_total_hit,
$br_total_found, $br_total_hit);
}
#
# system_no_output(mode, parameters)
#
# Call an external program using PARAMETERS while suppressing depending on
# the value of MODE:
#
# MODE & 1: suppress STDOUT
# MODE & 2: suppress STDERR
#
# Return 0 on success, non-zero otherwise.
#
sub system_no_output($@)
{
my $mode = shift;
my $result;
local *OLD_STDERR;
local *OLD_STDOUT;
# Save old stdout and stderr handles
($mode & 1) && open(OLD_STDOUT, ">>&", "STDOUT");
($mode & 2) && open(OLD_STDERR, ">>&", "STDERR");
# Redirect to /dev/null
($mode & 1) && open(STDOUT, ">", "/dev/null");
($mode & 2) && open(STDERR, ">", "/dev/null");
system(@_);
$result = $?;
# Close redirected handles
($mode & 1) && close(STDOUT);
($mode & 2) && close(STDERR);
# Restore old handles
($mode & 1) && open(STDOUT, ">>&", "OLD_STDOUT");
($mode & 2) && open(STDERR, ">>&", "OLD_STDERR");
return $result;
}
#
# read_config(filename)
#
# Read configuration file FILENAME and return a reference to a hash containing
# all valid key=value pairs found.
#
sub read_config($)
{
my $filename = $_[0];
my %result;
my $key;
my $value;
local *HANDLE;
if (!open(HANDLE, "<", $filename))
{
warn("WARNING: cannot read configuration file $filename\n");
return undef;
}
while (<HANDLE>)
{
chomp;
# Skip comments
s/#.*//;
# Remove leading blanks
s/^\s+//;
# Remove trailing blanks
s/\s+$//;
next unless length;
($key, $value) = split(/\s*=\s*/, $_, 2);
if (defined($key) && defined($value))
{
$result{$key} = $value;
}
else
{
warn("WARNING: malformed statement in line $. ".
"of configuration file $filename\n");
}
}
close(HANDLE);
return \%result;
}
#
# apply_config(REF)
#
# REF is a reference to a hash containing the following mapping:
#
# key_string => var_ref
#
# where KEY_STRING is a keyword and VAR_REF is a reference to an associated
# variable. If the global configuration hashes CONFIG or OPT_RC contain a value
# for keyword KEY_STRING, VAR_REF will be assigned the value for that keyword.
#
sub apply_config($)
{
my $ref = $_[0];
foreach (keys(%{$ref}))
{
if (defined($opt_rc{$_})) {
${$ref->{$_}} = $opt_rc{$_};
} elsif (defined($config->{$_})) {
${$ref->{$_}} = $config->{$_};
}
}
}
sub warn_handler($)
{
my ($msg) = @_;
temp_cleanup();
warn("$tool_name: $msg");
}
sub die_handler($)
{
my ($msg) = @_;
temp_cleanup();
die("$tool_name: $msg");
}
sub abort_handler($)
{
temp_cleanup();
exit(1);
}
sub temp_cleanup()
{
if (@temp_dirs) {
info("Removing temporary directories.\n");
foreach (@temp_dirs) {
rmtree($_);
}
@temp_dirs = ();
}
}
sub setup_gkv_sys()
{
system_no_output(3, "mount", "-t", "debugfs", "nodev",
"/sys/kernel/debug");
}
sub setup_gkv_proc()
{
if (system_no_output(3, "modprobe", "gcov_proc")) {
system_no_output(3, "modprobe", "gcov_prof");
}
}
sub check_gkv_sys($)
{
my ($dir) = @_;
if (-e "$dir/reset") {
return 1;
}
return 0;
}
sub check_gkv_proc($)
{
my ($dir) = @_;
if (-e "$dir/vmlinux") {
return 1;
}
return 0;
}
sub setup_gkv()
{
my $dir;
my $sys_dir = "/sys/kernel/debug/gcov";
my $proc_dir = "/proc/gcov";
my @todo;
if (!defined($gcov_dir)) {
info("Auto-detecting gcov kernel support.\n");
@todo = ( "cs", "cp", "ss", "cs", "sp", "cp" );
} elsif ($gcov_dir =~ /proc/) {
info("Checking gcov kernel support at $gcov_dir ".
"(user-specified).\n");
@todo = ( "cp", "sp", "cp", "cs", "ss", "cs");
} else {
info("Checking gcov kernel support at $gcov_dir ".
"(user-specified).\n");
@todo = ( "cs", "ss", "cs", "cp", "sp", "cp", );
}
foreach (@todo) {
if ($_ eq "cs") {
# Check /sys
$dir = defined($gcov_dir) ? $gcov_dir : $sys_dir;
if (check_gkv_sys($dir)) {
info("Found ".$GKV_NAME[$GKV_SYS]." gcov ".
"kernel support at $dir\n");
return ($GKV_SYS, $dir);
}
} elsif ($_ eq "cp") {
# Check /proc
$dir = defined($gcov_dir) ? $gcov_dir : $proc_dir;
if (check_gkv_proc($dir)) {
info("Found ".$GKV_NAME[$GKV_PROC]." gcov ".
"kernel support at $dir\n");
return ($GKV_PROC, $dir);
}
} elsif ($_ eq "ss") {
# Setup /sys
setup_gkv_sys();
} elsif ($_ eq "sp") {
# Setup /proc
setup_gkv_proc();
}
}
if (defined($gcov_dir)) {
die("ERROR: could not find gcov kernel data at $gcov_dir\n");
} else {
die("ERROR: no gcov kernel data found\n");
}
}
#
# get_overall_line(found, hit, name_singular, name_plural)
#
# Return a string containing overall information for the specified
# found/hit data.
#
sub get_overall_line($$$$)
{
my ($found, $hit, $name_sn, $name_pl) = @_;
my $name;
return "no data found" if (!defined($found) || $found == 0);
$name = ($found == 1) ? $name_sn : $name_pl;
return rate($hit, $found, "% ($hit of $found $name)");
}
#
# print_overall_rate(ln_do, ln_found, ln_hit, fn_do, fn_found, fn_hit, br_do
# br_found, br_hit)
#
# Print overall coverage rates for the specified coverage types.
#
sub print_overall_rate($$$$$$$$$)
{
my ($ln_do, $ln_found, $ln_hit, $fn_do, $fn_found, $fn_hit,
$br_do, $br_found, $br_hit) = @_;
info("Summary coverage rate:\n");
info(" lines......: %s\n",
get_overall_line($ln_found, $ln_hit, "line", "lines"))
if ($ln_do);
info(" functions..: %s\n",
get_overall_line($fn_found, $fn_hit, "function", "functions"))
if ($fn_do);
info(" branches...: %s\n",
get_overall_line($br_found, $br_hit, "branch", "branches"))
if ($br_do);
}
#
# rate(hit, found[, suffix, precision, width])
#
# Return the coverage rate [0..100] for HIT and FOUND values. 0 is only
# returned when HIT is 0. 100 is only returned when HIT equals FOUND.
# PRECISION specifies the precision of the result. SUFFIX defines a
# string that is appended to the result if FOUND is non-zero. Spaces
# are added to the start of the resulting string until it is at least WIDTH
# characters wide.
#
sub rate($$;$$$)
{
my ($hit, $found, $suffix, $precision, $width) = @_;
my $rate;
# Assign defaults if necessary
$precision = 1 if (!defined($precision));
$suffix = "" if (!defined($suffix));
$width = 0 if (!defined($width));
return sprintf("%*s", $width, "-") if (!defined($found) || $found == 0);
$rate = sprintf("%.*f", $precision, $hit * 100 / $found);
# Adjust rates if necessary
if ($rate == 0 && $hit > 0) {
$rate = sprintf("%.*f", $precision, 1 / 10 ** $precision);
} elsif ($rate == 100 && $hit != $found) {
$rate = sprintf("%.*f", $precision, 100 - 1 / 10 ** $precision);
}
return sprintf("%*s", $width, $rate.$suffix);
}