Message ID | 20211028064154.2301049-2-davidgow@google.com (mailing list archive) |
---|---|
State | New |
Delegated to: | Brendan Higgins |
Headers | show |
Series | [v3,1/4] kunit: tool: Do not error on tests without test plans | expand |
On Wed, Oct 27, 2021 at 11:42 PM David Gow <davidgow@google.com> wrote: > > It's possible for a test to have a subtest header, but zero valid > subtests. We used to error on this if the test plan had no subtests > listed, but it's possible to have subtests without a test plan (indeed, > this is how parameterised tests work). > > Tests with 0 subtests now have the result NO_TESTS, and will report an > error (which does not halt test execution, but is printed in a scary red > colour and is noted in the results summary). Tested by tweaking ext4 tests (and running with patch 3) [15:04:33] =============== ext4_inode_test (1 subtest) ================ [15:04:33] ============== inode_test_xtimestamp_decoding ============== [15:04:33] [ERROR] Test inode_test_xtimestamp_decoding: 0 tests run! [15:04:33] ====== [NO TESTS RUN] inode_test_xtimestamp_decoding ======= [15:04:33] ================ [SKIPPED] ext4_inode_test ================= [15:04:33] ============================================================ [15:04:33] Testing complete. Passed: 0, Failed: 0, Crashed: 0, Skipped: 1, Errors: 1 [15:04:33] Elapsed time: 48.581s total, 0.000s configuring, 45.486s building, 2.992s running It's maybe a bit confusing to have ERROR, NO TESTS RUN, and SKIPPED all printed for the same thing. An alternative would be to drop the error, giving [15:04:33] =============== ext4_inode_test (1 subtest) ================ [15:04:33] ============== inode_test_xtimestamp_decoding ============== [15:04:33] ====== [NO TESTS RUN] inode_test_xtimestamp_decoding ======= [15:04:33] ================ [SKIPPED] ext4_inode_test ================= [15:04:33] ============================================================ But looking at it, I think I prefer the more explicit ERROR being there. > > Signed-off-by: David Gow <davidgow@google.com> Reviewed-by: Daniel Latypov <dlatypov@google.com> A few optional nits below. > --- > > Changes since v2: > https://lore.kernel.org/linux-kselftest/20211027013702.2039566-2-davidgow@google.com/ > - Report NO_TESTS as '[NO TESTS RUN]' in yellow, instead of '[FAILED]' > in red, particularly since it doesn't get counted as a failure. > > tools/testing/kunit/kunit_parser.py | 16 +++++++++++----- > tools/testing/kunit/kunit_tool_test.py | 9 +++++++++ > .../test_is_test_passed-no_tests_no_plan.log | 7 +++++++ > 3 files changed, 27 insertions(+), 5 deletions(-) > create mode 100644 tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log > > diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py > index 50ded55c168c..68c847e8ca58 100644 > --- a/tools/testing/kunit/kunit_parser.py > +++ b/tools/testing/kunit/kunit_parser.py > @@ -360,9 +360,6 @@ def parse_test_plan(lines: LineStream, test: Test) -> bool: > test.log.append(lines.pop()) > expected_count = int(match.group(1)) > test.expected_count = expected_count > - if expected_count == 0: > - test.status = TestStatus.NO_TESTS > - test.add_error('0 tests run!') > return True > > TEST_RESULT = re.compile(r'^(ok|not ok) ([0-9]+) (- )?([^#]*)( # .*)?$') > @@ -589,6 +586,8 @@ def format_test_result(test: Test) -> str: > return (green('[PASSED] ') + test.name) > elif test.status == TestStatus.SKIPPED: > return (yellow('[SKIPPED] ') + test.name) > + elif test.status == TestStatus.NO_TESTS: > + return (yellow('[NO TESTS RUN] ') + test.name) > elif test.status == TestStatus.TEST_CRASHED: > print_log(test.log) > return (red('[CRASHED] ') + test.name) > @@ -731,6 +730,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: > # test plan > test.name = "main" > parse_test_plan(lines, test) > + parent_test = True > else: > # If KTAP/TAP header is not found, test must be subtest > # header or test result line so parse attempt to parser > @@ -744,7 +744,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: > expected_count = test.expected_count > subtests = [] > test_num = 1 > - while expected_count is None or test_num <= expected_count: > + while parent_test and (expected_count is None or test_num <= expected_count): > # Loop to parse any subtests. > # Break after parsing expected number of tests or > # if expected number of tests is unknown break when test > @@ -779,9 +779,15 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: > parse_test_result(lines, test, expected_num) > else: > test.add_error('missing subtest result line!') > + > + # Check for there being no tests > + if parent_test and len(subtests) == 0: > + test.status = TestStatus.NO_TESTS > + test.add_error('0 tests run!') > + > # Add statuses to TestCounts attribute in Test object > bubble_up_test_results(test) > - if parent_test: > + if parent_test and not main: > # If test has subtests and is not the main test object, print > # footer. > print_test_footer(test) > diff --git a/tools/testing/kunit/kunit_tool_test.py b/tools/testing/kunit/kunit_tool_test.py > index bc8793145713..c59fe0777387 100755 > --- a/tools/testing/kunit/kunit_tool_test.py > +++ b/tools/testing/kunit/kunit_tool_test.py > @@ -208,6 +208,15 @@ class KUnitParserTest(unittest.TestCase): > self.assertEqual( > kunit_parser.TestStatus.NO_TESTS, > result.status) I'd prefer we split these test cases out. Perhaps: def test_no_tests_empty_plan(self): ... def test_no_tests_no_plan(self): ... # this new test > + no_plan_log = test_data_path('test_is_test_passed-no_tests_no_plan.log') > + with open(no_plan_log) as file: > + result = kunit_parser.parse_run_tests( > + kunit_parser.extract_tap_lines(file.readlines())) > + self.assertEqual(0, len(result.test.subtests[0].subtests[0].subtests)) > + self.assertEqual( > + kunit_parser.TestStatus.NO_TESTS, > + result.test.subtests[0].subtests[0].status) optional: self.assertEqual(1, result.test.counts.errors) > + > > def test_no_kunit_output(self): > crash_log = test_data_path('test_insufficient_memory.log') > diff --git a/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log > new file mode 100644 > index 000000000000..dd873c981108 > --- /dev/null > +++ b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log > @@ -0,0 +1,7 @@ > +TAP version 14 > +1..1 > + # Subtest: suite > + 1..1 > + # Subtest: case > + ok 1 - case # SKIP > +ok 1 - suite > -- > 2.33.0.1079.g6e70778dc9-goog >
diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py index 50ded55c168c..68c847e8ca58 100644 --- a/tools/testing/kunit/kunit_parser.py +++ b/tools/testing/kunit/kunit_parser.py @@ -360,9 +360,6 @@ def parse_test_plan(lines: LineStream, test: Test) -> bool: test.log.append(lines.pop()) expected_count = int(match.group(1)) test.expected_count = expected_count - if expected_count == 0: - test.status = TestStatus.NO_TESTS - test.add_error('0 tests run!') return True TEST_RESULT = re.compile(r'^(ok|not ok) ([0-9]+) (- )?([^#]*)( # .*)?$') @@ -589,6 +586,8 @@ def format_test_result(test: Test) -> str: return (green('[PASSED] ') + test.name) elif test.status == TestStatus.SKIPPED: return (yellow('[SKIPPED] ') + test.name) + elif test.status == TestStatus.NO_TESTS: + return (yellow('[NO TESTS RUN] ') + test.name) elif test.status == TestStatus.TEST_CRASHED: print_log(test.log) return (red('[CRASHED] ') + test.name) @@ -731,6 +730,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: # test plan test.name = "main" parse_test_plan(lines, test) + parent_test = True else: # If KTAP/TAP header is not found, test must be subtest # header or test result line so parse attempt to parser @@ -744,7 +744,7 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: expected_count = test.expected_count subtests = [] test_num = 1 - while expected_count is None or test_num <= expected_count: + while parent_test and (expected_count is None or test_num <= expected_count): # Loop to parse any subtests. # Break after parsing expected number of tests or # if expected number of tests is unknown break when test @@ -779,9 +779,15 @@ def parse_test(lines: LineStream, expected_num: int, log: List[str]) -> Test: parse_test_result(lines, test, expected_num) else: test.add_error('missing subtest result line!') + + # Check for there being no tests + if parent_test and len(subtests) == 0: + test.status = TestStatus.NO_TESTS + test.add_error('0 tests run!') + # Add statuses to TestCounts attribute in Test object bubble_up_test_results(test) - if parent_test: + if parent_test and not main: # If test has subtests and is not the main test object, print # footer. print_test_footer(test) diff --git a/tools/testing/kunit/kunit_tool_test.py b/tools/testing/kunit/kunit_tool_test.py index bc8793145713..c59fe0777387 100755 --- a/tools/testing/kunit/kunit_tool_test.py +++ b/tools/testing/kunit/kunit_tool_test.py @@ -208,6 +208,15 @@ class KUnitParserTest(unittest.TestCase): self.assertEqual( kunit_parser.TestStatus.NO_TESTS, result.status) + no_plan_log = test_data_path('test_is_test_passed-no_tests_no_plan.log') + with open(no_plan_log) as file: + result = kunit_parser.parse_run_tests( + kunit_parser.extract_tap_lines(file.readlines())) + self.assertEqual(0, len(result.test.subtests[0].subtests[0].subtests)) + self.assertEqual( + kunit_parser.TestStatus.NO_TESTS, + result.test.subtests[0].subtests[0].status) + def test_no_kunit_output(self): crash_log = test_data_path('test_insufficient_memory.log') diff --git a/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log new file mode 100644 index 000000000000..dd873c981108 --- /dev/null +++ b/tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log @@ -0,0 +1,7 @@ +TAP version 14 +1..1 + # Subtest: suite + 1..1 + # Subtest: case + ok 1 - case # SKIP +ok 1 - suite
It's possible for a test to have a subtest header, but zero valid subtests. We used to error on this if the test plan had no subtests listed, but it's possible to have subtests without a test plan (indeed, this is how parameterised tests work). Tests with 0 subtests now have the result NO_TESTS, and will report an error (which does not halt test execution, but is printed in a scary red colour and is noted in the results summary). Signed-off-by: David Gow <davidgow@google.com> --- Changes since v2: https://lore.kernel.org/linux-kselftest/20211027013702.2039566-2-davidgow@google.com/ - Report NO_TESTS as '[NO TESTS RUN]' in yellow, instead of '[FAILED]' in red, particularly since it doesn't get counted as a failure. tools/testing/kunit/kunit_parser.py | 16 +++++++++++----- tools/testing/kunit/kunit_tool_test.py | 9 +++++++++ .../test_is_test_passed-no_tests_no_plan.log | 7 +++++++ 3 files changed, 27 insertions(+), 5 deletions(-) create mode 100644 tools/testing/kunit/test_data/test_is_test_passed-no_tests_no_plan.log