mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 08:31:55 +00:00
ed01ad3a2f
This patch addresses misleading error messages reported by kunit_tool in two cases. First, in the case of TAP output having an incorrect header format or missing a header, the parser used to output an error message of 'no tests run!'. Now the parser outputs an error message of 'could not parse test results!'. As an example: Before: $ ./tools/testing/kunit/kunit.py parse /dev/null [ERROR] no tests run! ... After: $ ./tools/testing/kunit/kunit.py parse /dev/null [ERROR] could not parse test results! ... Second, in the case of TAP output with the correct header but no tests, the parser used to output an error message of 'could not parse test results!'. Now the parser outputs an error message of 'no tests run!'. As an example: Before: $ echo -e 'TAP version 14\n1..0' | ./tools/testing/kunit/kunit.py parse [ERROR] could not parse test results! After: $ echo -e 'TAP version 14\n1..0' | ./tools/testing/kunit/kunit.py parse [ERROR] no tests run! Additionally, this patch also corrects the tests in kunit_tool_test.py and adds a test to check the error in the case of TAP output with the correct header but no tests. Signed-off-by: Rae Moar <rmoar@google.com> Reviewed-by: David Gow <davidgow@google.com> Reviewed-by: Daniel Latypov <dlatypov@google.com> Reviewed-by: Brendan Higgins <brendanhiggins@google.com> Signed-off-by: Shuah Khan <skhan@linuxfoundation.org>
433 lines
12 KiB
Python
433 lines
12 KiB
Python
# SPDX-License-Identifier: GPL-2.0
|
|
#
|
|
# Parses test results from a kernel dmesg log.
|
|
#
|
|
# Copyright (C) 2019, Google LLC.
|
|
# Author: Felix Guo <felixguoxiuping@gmail.com>
|
|
# Author: Brendan Higgins <brendanhiggins@google.com>
|
|
|
|
import re
|
|
|
|
from collections import namedtuple
|
|
from datetime import datetime
|
|
from enum import Enum, auto
|
|
from functools import reduce
|
|
from typing import Iterable, Iterator, List, Optional, Tuple
|
|
|
|
TestResult = namedtuple('TestResult', ['status','suites','log'])
|
|
|
|
class TestSuite(object):
|
|
def __init__(self) -> None:
|
|
self.status = TestStatus.SUCCESS
|
|
self.name = ''
|
|
self.cases = [] # type: List[TestCase]
|
|
|
|
def __str__(self) -> str:
|
|
return 'TestSuite(' + str(self.status) + ',' + self.name + ',' + str(self.cases) + ')'
|
|
|
|
def __repr__(self) -> str:
|
|
return str(self)
|
|
|
|
class TestCase(object):
|
|
def __init__(self) -> None:
|
|
self.status = TestStatus.SUCCESS
|
|
self.name = ''
|
|
self.log = [] # type: List[str]
|
|
|
|
def __str__(self) -> str:
|
|
return 'TestCase(' + str(self.status) + ',' + self.name + ',' + str(self.log) + ')'
|
|
|
|
def __repr__(self) -> str:
|
|
return str(self)
|
|
|
|
class TestStatus(Enum):
|
|
SUCCESS = auto()
|
|
FAILURE = auto()
|
|
SKIPPED = auto()
|
|
TEST_CRASHED = auto()
|
|
NO_TESTS = auto()
|
|
FAILURE_TO_PARSE_TESTS = auto()
|
|
|
|
class LineStream:
|
|
"""Provides a peek()/pop() interface over an iterator of (line#, text)."""
|
|
_lines: Iterator[Tuple[int, str]]
|
|
_next: Tuple[int, str]
|
|
_done: bool
|
|
|
|
def __init__(self, lines: Iterator[Tuple[int, str]]):
|
|
self._lines = lines
|
|
self._done = False
|
|
self._next = (0, '')
|
|
self._get_next()
|
|
|
|
def _get_next(self) -> None:
|
|
try:
|
|
self._next = next(self._lines)
|
|
except StopIteration:
|
|
self._done = True
|
|
|
|
def peek(self) -> str:
|
|
return self._next[1]
|
|
|
|
def pop(self) -> str:
|
|
n = self._next
|
|
self._get_next()
|
|
return n[1]
|
|
|
|
def __bool__(self) -> bool:
|
|
return not self._done
|
|
|
|
# Only used by kunit_tool_test.py.
|
|
def __iter__(self) -> Iterator[str]:
|
|
while bool(self):
|
|
yield self.pop()
|
|
|
|
def line_number(self) -> int:
|
|
return self._next[0]
|
|
|
|
kunit_start_re = re.compile(r'TAP version [0-9]+$')
|
|
kunit_end_re = re.compile('(List of all partitions:|'
|
|
'Kernel panic - not syncing: VFS:|reboot: System halted)')
|
|
|
|
def extract_tap_lines(kernel_output: Iterable[str]) -> LineStream:
|
|
def isolate_kunit_output(kernel_output: Iterable[str]) -> Iterator[Tuple[int, str]]:
|
|
line_num = 0
|
|
started = False
|
|
for line in kernel_output:
|
|
line_num += 1
|
|
line = line.rstrip() # line always has a trailing \n
|
|
if kunit_start_re.search(line):
|
|
prefix_len = len(line.split('TAP version')[0])
|
|
started = True
|
|
yield line_num, line[prefix_len:]
|
|
elif kunit_end_re.search(line):
|
|
break
|
|
elif started:
|
|
yield line_num, line[prefix_len:]
|
|
return LineStream(lines=isolate_kunit_output(kernel_output))
|
|
|
|
def raw_output(kernel_output) -> None:
|
|
for line in kernel_output:
|
|
print(line.rstrip())
|
|
|
|
DIVIDER = '=' * 60
|
|
|
|
RESET = '\033[0;0m'
|
|
|
|
def red(text) -> str:
|
|
return '\033[1;31m' + text + RESET
|
|
|
|
def yellow(text) -> str:
|
|
return '\033[1;33m' + text + RESET
|
|
|
|
def green(text) -> str:
|
|
return '\033[1;32m' + text + RESET
|
|
|
|
def print_with_timestamp(message) -> None:
|
|
print('[%s] %s' % (datetime.now().strftime('%H:%M:%S'), message))
|
|
|
|
def format_suite_divider(message) -> str:
|
|
return '======== ' + message + ' ========'
|
|
|
|
def print_suite_divider(message) -> None:
|
|
print_with_timestamp(DIVIDER)
|
|
print_with_timestamp(format_suite_divider(message))
|
|
|
|
def print_log(log) -> None:
|
|
for m in log:
|
|
print_with_timestamp(m)
|
|
|
|
TAP_ENTRIES = re.compile(r'^(TAP|[\s]*ok|[\s]*not ok|[\s]*[0-9]+\.\.[0-9]+|[\s]*#).*$')
|
|
|
|
def consume_non_diagnostic(lines: LineStream) -> None:
|
|
while lines and not TAP_ENTRIES.match(lines.peek()):
|
|
lines.pop()
|
|
|
|
def save_non_diagnostic(lines: LineStream, test_case: TestCase) -> None:
|
|
while lines and not TAP_ENTRIES.match(lines.peek()):
|
|
test_case.log.append(lines.peek())
|
|
lines.pop()
|
|
|
|
OkNotOkResult = namedtuple('OkNotOkResult', ['is_ok','description', 'text'])
|
|
|
|
OK_NOT_OK_SKIP = re.compile(r'^[\s]*(ok|not ok) [0-9]+ - (.*) # SKIP(.*)$')
|
|
|
|
OK_NOT_OK_SUBTEST = re.compile(r'^[\s]+(ok|not ok) [0-9]+ - (.*)$')
|
|
|
|
OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) ([0-9]+) - (.*)$')
|
|
|
|
def parse_ok_not_ok_test_case(lines: LineStream, test_case: TestCase) -> bool:
|
|
save_non_diagnostic(lines, test_case)
|
|
if not lines:
|
|
test_case.status = TestStatus.TEST_CRASHED
|
|
return True
|
|
line = lines.peek()
|
|
match = OK_NOT_OK_SUBTEST.match(line)
|
|
while not match and lines:
|
|
line = lines.pop()
|
|
match = OK_NOT_OK_SUBTEST.match(line)
|
|
if match:
|
|
test_case.log.append(lines.pop())
|
|
test_case.name = match.group(2)
|
|
skip_match = OK_NOT_OK_SKIP.match(line)
|
|
if skip_match:
|
|
test_case.status = TestStatus.SKIPPED
|
|
return True
|
|
if test_case.status == TestStatus.TEST_CRASHED:
|
|
return True
|
|
if match.group(1) == 'ok':
|
|
test_case.status = TestStatus.SUCCESS
|
|
else:
|
|
test_case.status = TestStatus.FAILURE
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
SUBTEST_DIAGNOSTIC = re.compile(r'^[\s]+# (.*)$')
|
|
DIAGNOSTIC_CRASH_MESSAGE = re.compile(r'^[\s]+# .*?: kunit test case crashed!$')
|
|
|
|
def parse_diagnostic(lines: LineStream, test_case: TestCase) -> bool:
|
|
save_non_diagnostic(lines, test_case)
|
|
if not lines:
|
|
return False
|
|
line = lines.peek()
|
|
match = SUBTEST_DIAGNOSTIC.match(line)
|
|
if match:
|
|
test_case.log.append(lines.pop())
|
|
crash_match = DIAGNOSTIC_CRASH_MESSAGE.match(line)
|
|
if crash_match:
|
|
test_case.status = TestStatus.TEST_CRASHED
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def parse_test_case(lines: LineStream) -> Optional[TestCase]:
|
|
test_case = TestCase()
|
|
save_non_diagnostic(lines, test_case)
|
|
while parse_diagnostic(lines, test_case):
|
|
pass
|
|
if parse_ok_not_ok_test_case(lines, test_case):
|
|
return test_case
|
|
else:
|
|
return None
|
|
|
|
SUBTEST_HEADER = re.compile(r'^[\s]+# Subtest: (.*)$')
|
|
|
|
def parse_subtest_header(lines: LineStream) -> Optional[str]:
|
|
consume_non_diagnostic(lines)
|
|
if not lines:
|
|
return None
|
|
match = SUBTEST_HEADER.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return match.group(1)
|
|
else:
|
|
return None
|
|
|
|
SUBTEST_PLAN = re.compile(r'[\s]+[0-9]+\.\.([0-9]+)')
|
|
|
|
def parse_subtest_plan(lines: LineStream) -> Optional[int]:
|
|
consume_non_diagnostic(lines)
|
|
match = SUBTEST_PLAN.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return int(match.group(1))
|
|
else:
|
|
return None
|
|
|
|
def max_status(left: TestStatus, right: TestStatus) -> TestStatus:
|
|
if left == right:
|
|
return left
|
|
elif left == TestStatus.TEST_CRASHED or right == TestStatus.TEST_CRASHED:
|
|
return TestStatus.TEST_CRASHED
|
|
elif left == TestStatus.FAILURE or right == TestStatus.FAILURE:
|
|
return TestStatus.FAILURE
|
|
elif left == TestStatus.SKIPPED:
|
|
return right
|
|
else:
|
|
return left
|
|
|
|
def parse_ok_not_ok_test_suite(lines: LineStream,
|
|
test_suite: TestSuite,
|
|
expected_suite_index: int) -> bool:
|
|
consume_non_diagnostic(lines)
|
|
if not lines:
|
|
test_suite.status = TestStatus.TEST_CRASHED
|
|
return False
|
|
line = lines.peek()
|
|
match = OK_NOT_OK_MODULE.match(line)
|
|
if match:
|
|
lines.pop()
|
|
if match.group(1) == 'ok':
|
|
test_suite.status = TestStatus.SUCCESS
|
|
else:
|
|
test_suite.status = TestStatus.FAILURE
|
|
skip_match = OK_NOT_OK_SKIP.match(line)
|
|
if skip_match:
|
|
test_suite.status = TestStatus.SKIPPED
|
|
suite_index = int(match.group(2))
|
|
if suite_index != expected_suite_index:
|
|
print_with_timestamp(
|
|
red('[ERROR] ') + 'expected_suite_index ' +
|
|
str(expected_suite_index) + ', but got ' +
|
|
str(suite_index))
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def bubble_up_errors(status_list: Iterable[TestStatus]) -> TestStatus:
|
|
return reduce(max_status, status_list, TestStatus.SKIPPED)
|
|
|
|
def bubble_up_test_case_errors(test_suite: TestSuite) -> TestStatus:
|
|
max_test_case_status = bubble_up_errors(x.status for x in test_suite.cases)
|
|
return max_status(max_test_case_status, test_suite.status)
|
|
|
|
def parse_test_suite(lines: LineStream, expected_suite_index: int) -> Optional[TestSuite]:
|
|
if not lines:
|
|
return None
|
|
consume_non_diagnostic(lines)
|
|
test_suite = TestSuite()
|
|
test_suite.status = TestStatus.SUCCESS
|
|
name = parse_subtest_header(lines)
|
|
if not name:
|
|
return None
|
|
test_suite.name = name
|
|
expected_test_case_num = parse_subtest_plan(lines)
|
|
if expected_test_case_num is None:
|
|
return None
|
|
while expected_test_case_num > 0:
|
|
test_case = parse_test_case(lines)
|
|
if not test_case:
|
|
break
|
|
test_suite.cases.append(test_case)
|
|
expected_test_case_num -= 1
|
|
if parse_ok_not_ok_test_suite(lines, test_suite, expected_suite_index):
|
|
test_suite.status = bubble_up_test_case_errors(test_suite)
|
|
return test_suite
|
|
elif not lines:
|
|
print_with_timestamp(red('[ERROR] ') + 'ran out of lines before end token')
|
|
return test_suite
|
|
else:
|
|
print(f'failed to parse end of suite "{name}", at line {lines.line_number()}: {lines.peek()}')
|
|
return None
|
|
|
|
TAP_HEADER = re.compile(r'^TAP version 14$')
|
|
|
|
def parse_tap_header(lines: LineStream) -> bool:
|
|
consume_non_diagnostic(lines)
|
|
if TAP_HEADER.match(lines.peek()):
|
|
lines.pop()
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
TEST_PLAN = re.compile(r'[0-9]+\.\.([0-9]+)')
|
|
|
|
def parse_test_plan(lines: LineStream) -> Optional[int]:
|
|
consume_non_diagnostic(lines)
|
|
match = TEST_PLAN.match(lines.peek())
|
|
if match:
|
|
lines.pop()
|
|
return int(match.group(1))
|
|
else:
|
|
return None
|
|
|
|
def bubble_up_suite_errors(test_suites: Iterable[TestSuite]) -> TestStatus:
|
|
return bubble_up_errors(x.status for x in test_suites)
|
|
|
|
def parse_test_result(lines: LineStream) -> TestResult:
|
|
consume_non_diagnostic(lines)
|
|
if not lines or not parse_tap_header(lines):
|
|
return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines)
|
|
expected_test_suite_num = parse_test_plan(lines)
|
|
if expected_test_suite_num == 0:
|
|
return TestResult(TestStatus.NO_TESTS, [], lines)
|
|
elif expected_test_suite_num is None:
|
|
return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines)
|
|
test_suites = []
|
|
for i in range(1, expected_test_suite_num + 1):
|
|
test_suite = parse_test_suite(lines, i)
|
|
if test_suite:
|
|
test_suites.append(test_suite)
|
|
else:
|
|
print_with_timestamp(
|
|
red('[ERROR] ') + ' expected ' +
|
|
str(expected_test_suite_num) +
|
|
' test suites, but got ' + str(i - 2))
|
|
break
|
|
test_suite = parse_test_suite(lines, -1)
|
|
if test_suite:
|
|
print_with_timestamp(red('[ERROR] ') +
|
|
'got unexpected test suite: ' + test_suite.name)
|
|
if test_suites:
|
|
return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines)
|
|
else:
|
|
return TestResult(TestStatus.NO_TESTS, [], lines)
|
|
|
|
class TestCounts:
|
|
passed: int
|
|
failed: int
|
|
crashed: int
|
|
skipped: int
|
|
|
|
def __init__(self):
|
|
self.passed = 0
|
|
self.failed = 0
|
|
self.crashed = 0
|
|
self.skipped = 0
|
|
|
|
def total(self) -> int:
|
|
return self.passed + self.failed + self.crashed + self.skipped
|
|
|
|
def print_and_count_results(test_result: TestResult) -> TestCounts:
|
|
counts = TestCounts()
|
|
for test_suite in test_result.suites:
|
|
if test_suite.status == TestStatus.SUCCESS:
|
|
print_suite_divider(green('[PASSED] ') + test_suite.name)
|
|
elif test_suite.status == TestStatus.SKIPPED:
|
|
print_suite_divider(yellow('[SKIPPED] ') + test_suite.name)
|
|
elif test_suite.status == TestStatus.TEST_CRASHED:
|
|
print_suite_divider(red('[CRASHED] ' + test_suite.name))
|
|
else:
|
|
print_suite_divider(red('[FAILED] ') + test_suite.name)
|
|
for test_case in test_suite.cases:
|
|
if test_case.status == TestStatus.SUCCESS:
|
|
counts.passed += 1
|
|
print_with_timestamp(green('[PASSED] ') + test_case.name)
|
|
elif test_case.status == TestStatus.SKIPPED:
|
|
counts.skipped += 1
|
|
print_with_timestamp(yellow('[SKIPPED] ') + test_case.name)
|
|
elif test_case.status == TestStatus.TEST_CRASHED:
|
|
counts.crashed += 1
|
|
print_with_timestamp(red('[CRASHED] ' + test_case.name))
|
|
print_log(map(yellow, test_case.log))
|
|
print_with_timestamp('')
|
|
else:
|
|
counts.failed += 1
|
|
print_with_timestamp(red('[FAILED] ') + test_case.name)
|
|
print_log(map(yellow, test_case.log))
|
|
print_with_timestamp('')
|
|
return counts
|
|
|
|
def parse_run_tests(kernel_output: Iterable[str]) -> TestResult:
|
|
counts = TestCounts()
|
|
lines = extract_tap_lines(kernel_output)
|
|
test_result = parse_test_result(lines)
|
|
if test_result.status == TestStatus.NO_TESTS:
|
|
print(red('[ERROR] ') + yellow('no tests run!'))
|
|
elif test_result.status == TestStatus.FAILURE_TO_PARSE_TESTS:
|
|
print(red('[ERROR] ') + yellow('could not parse test results!'))
|
|
else:
|
|
counts = print_and_count_results(test_result)
|
|
print_with_timestamp(DIVIDER)
|
|
if test_result.status == TestStatus.SUCCESS:
|
|
fmt = green
|
|
elif test_result.status == TestStatus.SKIPPED:
|
|
fmt = yellow
|
|
else:
|
|
fmt =red
|
|
print_with_timestamp(
|
|
fmt('Testing complete. %d tests run. %d failed. %d crashed. %d skipped.' %
|
|
(counts.total(), counts.failed, counts.crashed, counts.skipped)))
|
|
return test_result
|