| # -*- coding: utf-8 -*- |
| # Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| # See https://llvm.org/LICENSE.txt for license information. |
| # SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| |
| import json |
| import libear |
| import libscanbuild.report as sut |
| import unittest |
| import os |
| import os.path |
| |
| |
| def run_bug_parse(content): |
| with libear.TemporaryDirectory() as tmpdir: |
| file_name = os.path.join(tmpdir, 'test.html') |
| with open(file_name, 'w') as handle: |
| handle.writelines(content) |
| for bug in sut.parse_bug_html(file_name): |
| return bug |
| |
| |
| def run_crash_parse(content, preproc): |
| with libear.TemporaryDirectory() as tmpdir: |
| file_name = os.path.join(tmpdir, preproc + '.info.txt') |
| with open(file_name, 'w') as handle: |
| handle.writelines(content) |
| return sut.parse_crash(file_name) |
| |
| |
| class ParseFileTest(unittest.TestCase): |
| |
| def test_parse_bug(self): |
| content = [ |
| "some header\n", |
| "<!-- BUGDESC Division by zero -->\n", |
| "<!-- BUGTYPE Division by zero -->\n", |
| "<!-- BUGCATEGORY Logic error -->\n", |
| "<!-- BUGFILE xx -->\n", |
| "<!-- BUGLINE 5 -->\n", |
| "<!-- BUGCOLUMN 22 -->\n", |
| "<!-- BUGPATHLENGTH 4 -->\n", |
| "<!-- BUGMETAEND -->\n", |
| "<!-- REPORTHEADER -->\n", |
| "some tails\n"] |
| result = run_bug_parse(content) |
| self.assertEqual(result['bug_category'], 'Logic error') |
| self.assertEqual(result['bug_path_length'], 4) |
| self.assertEqual(result['bug_line'], 5) |
| self.assertEqual(result['bug_description'], 'Division by zero') |
| self.assertEqual(result['bug_type'], 'Division by zero') |
| self.assertEqual(result['bug_file'], 'xx') |
| |
| def test_parse_bug_empty(self): |
| content = [] |
| result = run_bug_parse(content) |
| self.assertEqual(result['bug_category'], 'Other') |
| self.assertEqual(result['bug_path_length'], 1) |
| self.assertEqual(result['bug_line'], 0) |
| |
| def test_parse_crash(self): |
| content = [ |
| "/some/path/file.c\n", |
| "Some very serious Error\n", |
| "bla\n", |
| "bla-bla\n"] |
| result = run_crash_parse(content, 'file.i') |
| self.assertEqual(result['source'], content[0].rstrip()) |
| self.assertEqual(result['problem'], content[1].rstrip()) |
| self.assertEqual(os.path.basename(result['file']), |
| 'file.i') |
| self.assertEqual(os.path.basename(result['info']), |
| 'file.i.info.txt') |
| self.assertEqual(os.path.basename(result['stderr']), |
| 'file.i.stderr.txt') |
| |
| def test_parse_real_crash(self): |
| import libscanbuild.analyze as sut2 |
| import re |
| with libear.TemporaryDirectory() as tmpdir: |
| filename = os.path.join(tmpdir, 'test.c') |
| with open(filename, 'w') as handle: |
| handle.write('int main() { return 0') |
| # produce failure report |
| opts = { |
| 'clang': 'clang', |
| 'directory': os.getcwd(), |
| 'flags': [], |
| 'file': filename, |
| 'output_dir': tmpdir, |
| 'language': 'c', |
| 'error_type': 'other_error', |
| 'error_output': 'some output', |
| 'exit_code': 13 |
| } |
| sut2.report_failure(opts) |
| # find the info file |
| pp_file = None |
| for root, _, files in os.walk(tmpdir): |
| keys = [os.path.join(root, name) for name in files] |
| for key in keys: |
| if re.match(r'^(.*/)+clang(.*)\.i$', key): |
| pp_file = key |
| self.assertIsNot(pp_file, None) |
| # read the failure report back |
| result = sut.parse_crash(pp_file + '.info.txt') |
| self.assertEqual(result['source'], filename) |
| self.assertEqual(result['problem'], 'Other Error') |
| self.assertEqual(result['file'], pp_file) |
| self.assertEqual(result['info'], pp_file + '.info.txt') |
| self.assertEqual(result['stderr'], pp_file + '.stderr.txt') |
| |
| |
| class ReportMethodTest(unittest.TestCase): |
| |
| def test_chop(self): |
| self.assertEqual('file', sut.chop('/prefix', '/prefix/file')) |
| self.assertEqual('file', sut.chop('/prefix/', '/prefix/file')) |
| self.assertEqual('lib/file', sut.chop('/prefix/', '/prefix/lib/file')) |
| self.assertEqual('/prefix/file', sut.chop('', '/prefix/file')) |
| |
| def test_chop_when_cwd(self): |
| self.assertEqual('../src/file', sut.chop('/cwd', '/src/file')) |
| self.assertEqual('../src/file', sut.chop('/prefix/cwd', |
| '/prefix/src/file')) |
| |
| |
| class GetPrefixFromCompilationDatabaseTest(unittest.TestCase): |
| |
| def test_with_different_filenames(self): |
| self.assertEqual( |
| sut.commonprefix(['/tmp/a.c', '/tmp/b.c']), '/tmp') |
| |
| def test_with_different_dirnames(self): |
| self.assertEqual( |
| sut.commonprefix(['/tmp/abs/a.c', '/tmp/ack/b.c']), '/tmp') |
| |
| def test_no_common_prefix(self): |
| self.assertEqual( |
| sut.commonprefix(['/tmp/abs/a.c', '/usr/ack/b.c']), '/') |
| |
| def test_with_single_file(self): |
| self.assertEqual( |
| sut.commonprefix(['/tmp/a.c']), '/tmp') |
| |
| def test_empty(self): |
| self.assertEqual( |
| sut.commonprefix([]), '') |
| |
| class MergeSarifTest(unittest.TestCase): |
| |
| def test_merging_sarif(self): |
| sarif1 = { |
| '$schema': 'https://raw.githubusercontent.com/oasis-tcs/sarif-spec/master/Schemata/sarif-schema-2.1.0.json', |
| 'runs': [ |
| { |
| 'artifacts': [ |
| { |
| 'length': 100, |
| 'location': { |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'mimeType': 'text/plain', |
| 'roles': [ |
| 'resultFile' |
| ] |
| } |
| ], |
| 'columnKind': 'unicodeCodePoints', |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'threadFlows': [ |
| { |
| 'locations': [ |
| { |
| 'importance': 'important', |
| 'location': { |
| 'message': { |
| 'text': 'test message 1' |
| }, |
| 'physicalLocation': { |
| 'artifactLocation': { |
| 'index': 0, |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'region': { |
| 'endColumn': 5, |
| 'startColumn': 1, |
| 'startLine': 2 |
| } |
| } |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| }, |
| { |
| 'codeFlows': [ |
| { |
| 'threadFlows': [ |
| { |
| 'locations': [ |
| { |
| 'importance': 'important', |
| 'location': { |
| 'message': { |
| 'text': 'test message 2' |
| }, |
| 'physicalLocation': { |
| 'artifactLocation': { |
| 'index': 0, |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'region': { |
| 'endColumn': 23, |
| 'startColumn': 9, |
| 'startLine': 10 |
| } |
| } |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| ], |
| 'tool': { |
| 'driver': { |
| 'fullName': 'clang static analyzer', |
| 'language': 'en-US', |
| 'name': 'clang', |
| 'rules': [ |
| { |
| 'fullDescription': { |
| 'text': 'test rule for merge sarif test' |
| }, |
| 'helpUrl': '//clang/tools/scan-build-py/tests/unit/test_report.py', |
| 'id': 'testId', |
| 'name': 'testName' |
| } |
| ], |
| 'version': 'test clang' |
| } |
| } |
| } |
| ], |
| 'version': '2.1.0' |
| } |
| sarif2 = { |
| '$schema': 'https://raw.githubusercontent.com/oasis-tcs/sarif-spec/master/Schemata/sarif-schema-2.1.0.json', |
| 'runs': [ |
| { |
| 'artifacts': [ |
| { |
| 'length': 1523, |
| 'location': { |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'mimeType': 'text/plain', |
| 'roles': [ |
| 'resultFile' |
| ] |
| } |
| ], |
| 'columnKind': 'unicodeCodePoints', |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'threadFlows': [ |
| { |
| 'locations': [ |
| { |
| 'importance': 'important', |
| 'location': { |
| 'message': { |
| 'text': 'test message 3' |
| }, |
| 'physicalLocation': { |
| 'artifactLocation': { |
| 'index': 0, |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'region': { |
| 'endColumn': 99, |
| 'startColumn': 99, |
| 'startLine': 17 |
| } |
| } |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| }, |
| { |
| 'codeFlows': [ |
| { |
| 'threadFlows': [ |
| { |
| 'locations': [ |
| { |
| 'importance': 'important', |
| 'location': { |
| 'message': { |
| 'text': 'test message 4' |
| }, |
| 'physicalLocation': { |
| 'artifactLocation': { |
| 'index': 0, |
| 'uri': '//clang/tools/scan-build-py/tests/unit/test_report.py' |
| }, |
| 'region': { |
| 'endColumn': 305, |
| 'startColumn': 304, |
| 'startLine': 1 |
| } |
| } |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| ], |
| 'tool': { |
| 'driver': { |
| 'fullName': 'clang static analyzer', |
| 'language': 'en-US', |
| 'name': 'clang', |
| 'rules': [ |
| { |
| 'fullDescription': { |
| 'text': 'test rule for merge sarif test' |
| }, |
| 'helpUrl': '//clang/tools/scan-build-py/tests/unit/test_report.py', |
| 'id': 'testId', |
| 'name': 'testName' |
| } |
| ], |
| 'version': 'test clang' |
| } |
| } |
| } |
| ], |
| 'version': '2.1.0' |
| } |
| |
| contents = [sarif1, sarif2] |
| with libear.TemporaryDirectory() as tmpdir: |
| for idx, content in enumerate(contents): |
| file_name = os.path.join(tmpdir, 'results-{}.sarif'.format(idx)) |
| with open(file_name, 'w') as handle: |
| json.dump(content, handle) |
| |
| sut.merge_sarif_files(tmpdir, sort_files=True) |
| |
| self.assertIn('results-merged.sarif', os.listdir(tmpdir)) |
| with open(os.path.join(tmpdir, 'results-merged.sarif')) as f: |
| merged = json.load(f) |
| self.assertEqual(len(merged['runs']), 2) |
| self.assertEqual(len(merged['runs'][0]['results']), 2) |
| self.assertEqual(len(merged['runs'][1]['results']), 2) |
| |
| expected = sarif1 |
| for run in sarif2['runs']: |
| expected['runs'].append(run) |
| |
| self.assertEqual(merged, expected) |
| |
| def test_merge_updates_embedded_link(self): |
| sarif1 = { |
| 'runs': [ |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 1-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 1-2 [link](sarif:/runs/1/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| }, |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 2-1 [link](sarif:/runs/0/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 2-2 [link](sarif:/runs/0/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| sarif2 = { |
| 'runs': [ |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 3-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 3-2 [link](sarif:/runs/1/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ], |
| }, |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 4-1 [link](sarif:/runs/0/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 4-2 [link](sarif:/runs/0/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| sarif3 = { |
| 'runs': [ |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 5-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 5-2 [link](sarif:/runs/1/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ], |
| }, |
| { |
| 'results': [ |
| { |
| 'codeFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 6-1 [link](sarif:/runs/0/results/0)' |
| }, |
| 'threadFlows': [ |
| { |
| 'message': { |
| 'text': 'test message 6-2 [link](sarif:/runs/0/results/0)' |
| } |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| ] |
| } |
| |
| contents = [sarif1, sarif2, sarif3] |
| |
| with libear.TemporaryDirectory() as tmpdir: |
| for idx, content in enumerate(contents): |
| file_name = os.path.join(tmpdir, 'results-{}.sarif'.format(idx)) |
| with open(file_name, 'w') as handle: |
| json.dump(content, handle) |
| |
| sut.merge_sarif_files(tmpdir, sort_files=True) |
| |
| self.assertIn('results-merged.sarif', os.listdir(tmpdir)) |
| with open(os.path.join(tmpdir, 'results-merged.sarif')) as f: |
| merged = json.load(f) |
| self.assertEqual(len(merged['runs']), 6) |
| |
| code_flows = [merged['runs'][x]['results'][0]['codeFlows'][0]['message']['text'] for x in range(6)] |
| thread_flows = [merged['runs'][x]['results'][0]['codeFlows'][0]['threadFlows'][0]['message']['text'] for x in range(6)] |
| |
| # The run index should be updated for the second and third sets of runs |
| self.assertEqual(code_flows, |
| [ |
| 'test message 1-1 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/1/results/0)', |
| 'test message 2-1 [link](sarif:/runs/0/results/0)', |
| 'test message 3-1 [link](sarif:/runs/3/results/0) [link2](sarif:/runs/3/results/0)', |
| 'test message 4-1 [link](sarif:/runs/2/results/0)', |
| 'test message 5-1 [link](sarif:/runs/5/results/0) [link2](sarif:/runs/5/results/0)', |
| 'test message 6-1 [link](sarif:/runs/4/results/0)' |
| ]) |
| self.assertEquals(thread_flows, |
| [ |
| 'test message 1-2 [link](sarif:/runs/1/results/0)', |
| 'test message 2-2 [link](sarif:/runs/0/results/0)', |
| 'test message 3-2 [link](sarif:/runs/3/results/0)', |
| 'test message 4-2 [link](sarif:/runs/2/results/0)', |
| 'test message 5-2 [link](sarif:/runs/5/results/0)', |
| 'test message 6-2 [link](sarif:/runs/4/results/0)' |
| ]) |
| |
| def test_overflow_run_count(self): |
| sarif1 = { |
| 'runs': [ |
| {'results': [{ |
| 'message': {'text': 'run 1-0 [link](sarif:/runs/1/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-1 [link](sarif:/runs/2/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-2 [link](sarif:/runs/3/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-3 [link](sarif:/runs/4/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-4 [link](sarif:/runs/5/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-5 [link](sarif:/runs/6/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-6 [link](sarif:/runs/7/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-7 [link](sarif:/runs/8/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-8 [link](sarif:/runs/9/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 1-9 [link](sarif:/runs/0/results/0)'} |
| }]} |
| ] |
| } |
| sarif2 = { |
| 'runs': [ |
| {'results': [{ |
| 'message': {'text': 'run 2-0 [link](sarif:/runs/1/results/0) [link2](sarif:/runs/2/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-1 [link](sarif:/runs/2/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-2 [link](sarif:/runs/3/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-3 [link](sarif:/runs/4/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-4 [link](sarif:/runs/5/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-5 [link](sarif:/runs/6/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-6 [link](sarif:/runs/7/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-7 [link](sarif:/runs/8/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-8 [link](sarif:/runs/9/results/0)'} |
| }]}, |
| {'results': [{ |
| 'message': {'text': 'run 2-9 [link](sarif:/runs/0/results/0)'} |
| }]} |
| ] |
| } |
| |
| contents = [sarif1, sarif2] |
| with libear.TemporaryDirectory() as tmpdir: |
| for idx, content in enumerate(contents): |
| file_name = os.path.join(tmpdir, 'results-{}.sarif'.format(idx)) |
| with open(file_name, 'w') as handle: |
| json.dump(content, handle) |
| |
| sut.merge_sarif_files(tmpdir, sort_files=True) |
| |
| self.assertIn('results-merged.sarif', os.listdir(tmpdir)) |
| with open(os.path.join(tmpdir, 'results-merged.sarif')) as f: |
| merged = json.load(f) |
| self.assertEqual(len(merged['runs']), 20) |
| |
| messages = [merged['runs'][x]['results'][0]['message']['text'] for x in range(20)] |
| self.assertEqual(messages, |
| [ |
| 'run 1-0 [link](sarif:/runs/1/results/0)', |
| 'run 1-1 [link](sarif:/runs/2/results/0)', |
| 'run 1-2 [link](sarif:/runs/3/results/0)', |
| 'run 1-3 [link](sarif:/runs/4/results/0)', |
| 'run 1-4 [link](sarif:/runs/5/results/0)', |
| 'run 1-5 [link](sarif:/runs/6/results/0)', |
| 'run 1-6 [link](sarif:/runs/7/results/0)', |
| 'run 1-7 [link](sarif:/runs/8/results/0)', |
| 'run 1-8 [link](sarif:/runs/9/results/0)', |
| 'run 1-9 [link](sarif:/runs/0/results/0)', |
| 'run 2-0 [link](sarif:/runs/11/results/0) [link2](sarif:/runs/12/results/0)', |
| 'run 2-1 [link](sarif:/runs/12/results/0)', |
| 'run 2-2 [link](sarif:/runs/13/results/0)', |
| 'run 2-3 [link](sarif:/runs/14/results/0)', |
| 'run 2-4 [link](sarif:/runs/15/results/0)', |
| 'run 2-5 [link](sarif:/runs/16/results/0)', |
| 'run 2-6 [link](sarif:/runs/17/results/0)', |
| 'run 2-7 [link](sarif:/runs/18/results/0)', |
| 'run 2-8 [link](sarif:/runs/19/results/0)', |
| 'run 2-9 [link](sarif:/runs/10/results/0)' |
| ]) |