# -*- coding: utf-8 -*-
# Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
# See https://llvm.org/LICENSE.txt for license information.
# SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
import json
import libear
import libscanbuild.report as sut
import unittest
import os
import os.path
def run_bug_parse(content):
with libear.TemporaryDirectory() as tmpdir:
file_name = os.path.join(tmpdir, "test.html")
with open(file_name, "w") as handle:
handle.writelines(content)
for bug in sut.parse_bug_html(file_name):
return bug
def run_crash_parse(content, preproc):
with libear.TemporaryDirectory() as tmpdir:
file_name = os.path.join(tmpdir, preproc + ".info.txt")
with open(file_name, "w") as handle:
handle.writelines(content)
return sut.parse_crash(file_name)
class ParseFileTest(unittest.TestCase):
def test_parse_bug(self):
content = [
"some header\n",
"<!-- BUGDESC Division by zero -->\n",
"<!-- BUGTYPE Division by zero -->\n",
"<!-- BUGCATEGORY Logic error -->\n",
"<!-- BUGFILE xx -->\n",
"<!-- BUGLINE 5 -->\n",
"<!-- BUGCOLUMN 22 -->\n",
"<!-- BUGPATHLENGTH 4 -->\n",
"<!-- BUGMETAEND -->\n",
"<!-- REPORTHEADER -->\n",
"some tails\n",
]
result = run_bug_parse(content)
self.assertEqual(result["bug_category"], "Logic error")
self.assertEqual(result["bug_path_length"], 4)
self.assertEqual(result["bug_line"], 5)
self.assertEqual(result["bug_description"], "Division by zero")
self.assertEqual(result["bug_type"], "Division by zero")
self.assertEqual(result["bug_file"], "xx")
def test_parse_bug_empty(self):
content = []
result = run_bug_parse(content)
self.assertEqual(result["bug_category"], "Other")
self.assertEqual(result["bug_path_length"], 1)
self.assertEqual(result["bug_line"], 0)
def test_parse_crash(self):
content = [
"/some/path/file.c\n",
"Some very serious Error\n",
"bla\n",
"bla-bla\n",
]
result = run_crash_parse(content, "file.i")
self.assertEqual(result["source"], content[0].rstrip())
self.assertEqual(result["problem"], content[1].rstrip())
self.assertEqual(os.path.basename(result["file"]), "file.i")
self.assertEqual(os.path.basename(result["info"]), "file.i.info.txt")
self.assertEqual(os.path.basename(result["stderr"]), "file.i.stderr.txt")
def test_parse_real_crash(self):
import libscanbuild.analyze as sut2
import re
with libear.TemporaryDirectory() as tmpdir:
filename = os.path.join(tmpdir, "test.c")
with open(filename, "w") as handle:
handle.write("int main() { return 0")
# produce failure report
opts = {
"clang": "clang",
"directory": os.getcwd(),
"flags": [],
"file": filename,
"output_dir": tmpdir,
"language": "c",
"error_type": "other_error",
"error_output": "some output",
"exit_code": 13,
}
sut2.report_failure(opts)
# find the info file
pp_file = None
for root, _, files in os.walk(tmpdir):
keys = [os.path.join(root, name) for name in files]
for key in keys:
if re.match(r"^(.*/)+clang(.*)\.i$", key):
pp_file = key
self.assertIsNot(pp_file, None)
# read the failure report back
result = sut.parse_crash(pp_file + ".info.txt")
self.assertEqual(result["source"], filename)
self.assertEqual(result["problem"], "Other Error")
self.assertEqual(result["file"], pp_file)
self.assertEqual(result["info"], pp_file + ".info.txt")
self.assertEqual(result["stderr"], pp_file + ".stderr.txt")
class ReportMethodTest(unittest.TestCase):
def test_chop(self):
self.assertEqual("file", sut.chop("/prefix", "/prefix/file"))
self.assertEqual("file", sut.chop("/prefix/", "/prefix/file"))
self.assertEqual("lib/file", sut.chop("/prefix/", "/prefix/lib/file"))
self.assertEqual("/prefix/file", sut.chop("", "/prefix/file"))
def test_chop_when_cwd(self):
self.assertEqual("../src/file", sut.chop("/cwd", "/src/file"))
self.assertEqual("../src/file", sut.chop("/prefix/cwd", "/prefix/src/file"))
class GetPrefixFromCompilationDatabaseTest(unittest.TestCase):
def test_with_different_filenames(self):
self.assertEqual(sut.commonprefix(["/tmp/a.c", "/tmp/b.c"]), "/tmp")
def test_with_different_dirnames(self):
self.assertEqual(sut.commonprefix(["/tmp/abs/a.c", "/tmp/ack/b.c"]), "/tmp")
def test_no_common_prefix(self):
self.assertEqual(sut.commonprefix(["/tmp/abs/a.c", "/usr/ack/b.c"]), "/")
def test_with_single_file(self):
self.assertEqual(sut.commonprefix(["/tmp/a.c"]), "/tmp")
def test_empty(self):
self.assertEqual(sut.commonprefix([]), "")
class MergeSarifTest(unittest.TestCase):
def test_merging_sarif(self):
sarif1 = {
"$schema": "https://raw.githubusercontent.com/oasis-tcs/sarif-spec/master/Schemata/sarif-schema-2.1.0.json",
"runs": [
{
"artifacts": [
{
"length": 100,
"location": {
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py"
},
"mimeType": "text/plain",
"roles": ["resultFile"],
}
],
"columnKind": "unicodeCodePoints",
"results": [
{
"codeFlows": [
{
"threadFlows": [
{
"locations": [
{
"importance": "important",
"location": {
"message": {
"text": "test message 1"
},
"physicalLocation": {
"artifactLocation": {
"index": 0,
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py",
},
"region": {
"endColumn": 5,
"startColumn": 1,
"startLine": 2,
},
},
},
}
]
}
]
}
]
},
{
"codeFlows": [
{
"threadFlows": [
{
"locations": [
{
"importance": "important",
"location": {
"message": {
"text": "test message 2"
},
"physicalLocation": {
"artifactLocation": {
"index": 0,
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py",
},
"region": {
"endColumn": 23,
"startColumn": 9,
"startLine": 10,
},
},
},
}
]
}
]
}
]
},
],
"tool": {
"driver": {
"fullName": "clang static analyzer",
"language": "en-US",
"name": "clang",
"rules": [
{
"fullDescription": {
"text": "test rule for merge sarif test"
},
"helpUrl": "//clang/tools/scan-build-py/tests/unit/test_report.py",
"id": "testId",
"name": "testName",
}
],
"version": "test clang",
}
},
}
],
"version": "2.1.0",
}
sarif2 = {
"$schema": "https://raw.githubusercontent.com/oasis-tcs/sarif-spec/master/Schemata/sarif-schema-2.1.0.json",
"runs": [
{
"artifacts": [
{
"length": 1523,
"location": {
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py"
},
"mimeType": "text/plain",
"roles": ["resultFile"],
}
],
"columnKind": "unicodeCodePoints",
"results": [
{
"codeFlows": [
{
"threadFlows": [
{
"locations": [
{
"importance": "important",
"location": {
"message": {
"text": "test message 3"
},
"physicalLocation": {
"artifactLocation": {
"index": 0,
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py",
},
"region": {
"endColumn": 99,
"startColumn": 99,
"startLine": 17,
},
},
},
}
]
}
]
}
]
},
{
"codeFlows": [
{
"threadFlows": [
{
"locations": [
{
"importance": "important",
"location": {
"message": {
"text": "test message 4"
},
"physicalLocation": {
"artifactLocation": {
"index": 0,
"uri": "//clang/tools/scan-build-py/tests/unit/test_report.py",
},
"region": {
"endColumn": 305,
"startColumn": 304,
"startLine": 1,
},
},
},
}
]
}
]
}
]
},
],
"tool": {
"driver": {
"fullName": "clang static analyzer",
"language": "en-US",
"name": "clang",
"rules": [
{
"fullDescription": {
"text": "test rule for merge sarif test"
},
"helpUrl": "//clang/tools/scan-build-py/tests/unit/test_report.py",
"id": "testId",
"name": "testName",
}
],
"version": "test clang",
}
},
}
],
"version": "2.1.0",
}
contents = [sarif1, sarif2]
with libear.TemporaryDirectory() as tmpdir:
for idx, content in enumerate(contents):
file_name = os.path.join(tmpdir, "results-{}.sarif".format(idx))
with open(file_name, "w") as handle:
json.dump(content, handle)
sut.merge_sarif_files(tmpdir, sort_files=True)
self.assertIn("results-merged.sarif", os.listdir(tmpdir))
with open(os.path.join(tmpdir, "results-merged.sarif")) as f:
merged = json.load(f)
self.assertEqual(len(merged["runs"]), 2)
self.assertEqual(len(merged["runs"][0]["results"]), 2)
self.assertEqual(len(merged["runs"][1]["results"]), 2)
expected = sarif1
for run in sarif2["runs"]:
expected["runs"].append(run)
self.assertEqual(merged, expected)
def test_merge_updates_embedded_link(self):
sarif1 = {
"runs": [
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 1-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 1-2 [link](sarif:/runs/1/results/0)"
}
}
],
}
]
}
]
},
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 2-1 [link](sarif:/runs/0/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 2-2 [link](sarif:/runs/0/results/0)"
}
}
],
}
]
}
]
},
]
}
sarif2 = {
"runs": [
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 3-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 3-2 [link](sarif:/runs/1/results/0)"
}
}
],
}
]
}
],
},
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 4-1 [link](sarif:/runs/0/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 4-2 [link](sarif:/runs/0/results/0)"
}
}
],
}
]
}
]
},
]
}
sarif3 = {
"runs": [
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 5-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 5-2 [link](sarif:/runs/1/results/0)"
}
}
],
}
]
}
],
},
{
"results": [
{
"codeFlows": [
{
"message": {
"text": "test message 6-1 [link](sarif:/runs/0/results/0)"
},
"threadFlows": [
{
"message": {
"text": "test message 6-2 [link](sarif:/runs/0/results/0)"
}
}
],
}
]
}
]
},
]
}
contents = [sarif1, sarif2, sarif3]
with libear.TemporaryDirectory() as tmpdir:
for idx, content in enumerate(contents):
file_name = os.path.join(tmpdir, "results-{}.sarif".format(idx))
with open(file_name, "w") as handle:
json.dump(content, handle)
sut.merge_sarif_files(tmpdir, sort_files=True)
self.assertIn("results-merged.sarif", os.listdir(tmpdir))
with open(os.path.join(tmpdir, "results-merged.sarif")) as f:
merged = json.load(f)
self.assertEqual(len(merged["runs"]), 6)
code_flows = [
merged["runs"][x]["results"][0]["codeFlows"][0]["message"]["text"]
for x in range(6)
]
thread_flows = [
merged["runs"][x]["results"][0]["codeFlows"][0]["threadFlows"][0][
"message"
]["text"]
for x in range(6)
]
# The run index should be updated for the second and third sets of runs
self.assertEqual(
code_flows,
[
"test message 1-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)",
"test message 2-1 [link](sarif:/runs/0/results/0)",
"test message 3-1 [link](sarif:/runs/3/results/0) [link2](sarif:/runs/3/results/0)",
"test message 4-1 [link](sarif:/runs/2/results/0)",
"test message 5-1 [link](sarif:/runs/5/results/0) [link2](sarif:/runs/5/results/0)",
"test message 6-1 [link](sarif:/runs/4/results/0)",
],
)
self.assertEquals(
thread_flows,
[
"test message 1-2 [link](sarif:/runs/1/results/0)",
"test message 2-2 [link](sarif:/runs/0/results/0)",
"test message 3-2 [link](sarif:/runs/3/results/0)",
"test message 4-2 [link](sarif:/runs/2/results/0)",
"test message 5-2 [link](sarif:/runs/5/results/0)",
"test message 6-2 [link](sarif:/runs/4/results/0)",
],
)
def test_overflow_run_count(self):
sarif1 = {
"runs": [
{
"results": [
{"message": {"text": "run 1-0 [link](sarif:/runs/1/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-1 [link](sarif:/runs/2/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-2 [link](sarif:/runs/3/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-3 [link](sarif:/runs/4/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-4 [link](sarif:/runs/5/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-5 [link](sarif:/runs/6/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-6 [link](sarif:/runs/7/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-7 [link](sarif:/runs/8/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-8 [link](sarif:/runs/9/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 1-9 [link](sarif:/runs/0/results/0)"}}
]
},
]
}
sarif2 = {
"runs": [
{
"results": [
{
"message": {
"text": "run 2-0 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/2/results/0)"
}
}
]
},
{
"results": [
{"message": {"text": "run 2-1 [link](sarif:/runs/2/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-2 [link](sarif:/runs/3/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-3 [link](sarif:/runs/4/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-4 [link](sarif:/runs/5/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-5 [link](sarif:/runs/6/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-6 [link](sarif:/runs/7/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-7 [link](sarif:/runs/8/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-8 [link](sarif:/runs/9/results/0)"}}
]
},
{
"results": [
{"message": {"text": "run 2-9 [link](sarif:/runs/0/results/0)"}}
]
},
]
}
contents = [sarif1, sarif2]
with libear.TemporaryDirectory() as tmpdir:
for idx, content in enumerate(contents):
file_name = os.path.join(tmpdir, "results-{}.sarif".format(idx))
with open(file_name, "w") as handle:
json.dump(content, handle)
sut.merge_sarif_files(tmpdir, sort_files=True)
self.assertIn("results-merged.sarif", os.listdir(tmpdir))
with open(os.path.join(tmpdir, "results-merged.sarif")) as f:
merged = json.load(f)
self.assertEqual(len(merged["runs"]), 20)
messages = [
merged["runs"][x]["results"][0]["message"]["text"]
for x in range(20)
]
self.assertEqual(
messages,
[
"run 1-0 [link](sarif:/runs/1/results/0)",
"run 1-1 [link](sarif:/runs/2/results/0)",
"run 1-2 [link](sarif:/runs/3/results/0)",
"run 1-3 [link](sarif:/runs/4/results/0)",
"run 1-4 [link](sarif:/runs/5/results/0)",
"run 1-5 [link](sarif:/runs/6/results/0)",
"run 1-6 [link](sarif:/runs/7/results/0)",
"run 1-7 [link](sarif:/runs/8/results/0)",
"run 1-8 [link](sarif:/runs/9/results/0)",
"run 1-9 [link](sarif:/runs/0/results/0)",
"run 2-0 [link](sarif:/runs/11/results/0) [link2](sarif:/runs/12/results/0)",
"run 2-1 [link](sarif:/runs/12/results/0)",
"run 2-2 [link](sarif:/runs/13/results/0)",
"run 2-3 [link](sarif:/runs/14/results/0)",
"run 2-4 [link](sarif:/runs/15/results/0)",
"run 2-5 [link](sarif:/runs/16/results/0)",
"run 2-6 [link](sarif:/runs/17/results/0)",
"run 2-7 [link](sarif:/runs/18/results/0)",
"run 2-8 [link](sarif:/runs/19/results/0)",
"run 2-9 [link](sarif:/runs/10/results/0)",
],
)