Message ID | 20200130230812.142642-6-brendanhiggins@google.com (mailing list archive) |
---|---|
State | Superseded, archived |
Headers | show |
Series | kunit: create a centralized executor to dispatch all KUnit tests | expand |
On 1/30/20 5:08 PM, Brendan Higgins wrote: > TAP 14 allows an optional test plan to be emitted before the start of > the start of testing[1]; this is valuable because it makes it possible > for a test harness to detect whether the number of tests run matches the > number of tests expected to be run, ensuring that no tests silently > failed. > > Link[1]: https://github.com/isaacs/testanything.github.io/blob/tap14/tap-version-14-specification.md#the-plan > Signed-off-by: Brendan Higgins <brendanhiggins@google.com> > Reviewed-by: Stephen Boyd <sboyd@kernel.org> > --- > lib/kunit/executor.c | 17 +++++ > lib/kunit/test.c | 11 --- > tools/testing/kunit/kunit_parser.py | 74 ++++++++++++++++--- > .../test_is_test_passed-all_passed.log | 1 + > .../test_data/test_is_test_passed-crash.log | 1 + > .../test_data/test_is_test_passed-failure.log | 1 + > 6 files changed, 82 insertions(+), 23 deletions(-) > > diff --git a/lib/kunit/executor.c b/lib/kunit/executor.c > index b75a46c560847..7fd16feff157e 100644 > --- a/lib/kunit/executor.c > +++ b/lib/kunit/executor.c > @@ -11,11 +11,28 @@ extern struct kunit_suite * const * const __kunit_suites_end[]; > > #if IS_BUILTIN(CONFIG_KUNIT) > > +static void kunit_print_tap_header(void) > +{ > + struct kunit_suite * const * const *suites, * const *subsuite; > + int num_of_suites = 0; > + > + for (suites = __kunit_suites_start; > + suites < __kunit_suites_end; > + suites++) > + for (subsuite = *suites; *subsuite != NULL; subsuite++) > + num_of_suites++; > + > + pr_info("TAP version 14\n"); > + pr_info("1..%d\n", num_of_suites); > +} > + Subsuites are not in Linux 5.5, as far as I can tell, so it is hard to review this. But I think this observation will still be valid even though I am not up to speed on subsuites. num_of_suites is all test suites declared (as built in). In Linux 5.5, each kunit_suite contains one or more test cases. Each test case leads to either "ok" or "not ok". So instead of counting number of tests suites, the number of test cases needs to be counted. Ahhh, I think my lack of visibility of what tree this series is against is about to lead me astray! Are the test cases in the subsuite implemented as what TAP 14 calls "subtests"? If so, then the above code that counts suites instead of test cases would be correct. I'm spinning my wheels now, since I don't know what I am reviewing against, so I'm going to stop reviewing, other than maybe a quick scan. -Frank > int kunit_run_all_tests(void) > { > struct kunit_suite * const * const *suites, * const *subsuite; > bool has_test_failed = false; > > + kunit_print_tap_header(); > + > for (suites = __kunit_suites_start; > suites < __kunit_suites_end; > suites++) { > diff --git a/lib/kunit/test.c b/lib/kunit/test.c > index 9242f932896c7..da56b94261b43 100644 > --- a/lib/kunit/test.c > +++ b/lib/kunit/test.c > @@ -18,16 +18,6 @@ static void kunit_set_failure(struct kunit *test) > WRITE_ONCE(test->success, false); > } > > -static void kunit_print_tap_version(void) > -{ > - static bool kunit_has_printed_tap_version; > - > - if (!kunit_has_printed_tap_version) { > - pr_info("TAP version 14\n"); > - kunit_has_printed_tap_version = true; > - } > -} > - > static size_t kunit_test_cases_len(struct kunit_case *test_cases) > { > struct kunit_case *test_case; > @@ -41,7 +31,6 @@ static size_t kunit_test_cases_len(struct kunit_case *test_cases) > > static void kunit_print_subtest_start(struct kunit_suite *suite) > { > - kunit_print_tap_version(); > pr_info("\t# Subtest: %s\n", suite->name); > pr_info("\t1..%zd\n", kunit_test_cases_len(suite->test_cases)); > } > diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py > index 4ffbae0f67325..78b3bdd03b1e4 100644 > --- a/tools/testing/kunit/kunit_parser.py > +++ b/tools/testing/kunit/kunit_parser.py > @@ -45,6 +45,7 @@ class TestStatus(Enum): > FAILURE = auto() > TEST_CRASHED = auto() > NO_TESTS = auto() > + FAILURE_TO_PARSE_TESTS = auto() > > kunit_start_re = re.compile(r'^TAP version [0-9]+$') > kunit_end_re = re.compile('List of all partitions:') > @@ -106,7 +107,7 @@ OkNotOkResult = namedtuple('OkNotOkResult', ['is_ok','description', 'text']) > > OK_NOT_OK_SUBTEST = re.compile(r'^\t(ok|not ok) [0-9]+ - (.*)$') > > -OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) [0-9]+ - (.*)$') > +OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) ([0-9]+) - (.*)$') > > def parse_ok_not_ok_test_case(lines: List[str], > test_case: TestCase, > @@ -196,7 +197,9 @@ def max_status(left: TestStatus, right: TestStatus) -> TestStatus: > else: > return TestStatus.SUCCESS > > -def parse_ok_not_ok_test_suite(lines: List[str], test_suite: TestSuite) -> bool: > +def parse_ok_not_ok_test_suite(lines: List[str], > + test_suite: TestSuite, > + expected_suite_index: int) -> bool: > consume_non_diagnositic(lines) > if not lines: > test_suite.status = TestStatus.TEST_CRASHED > @@ -209,6 +212,12 @@ def parse_ok_not_ok_test_suite(lines: List[str], test_suite: TestSuite) -> bool: > test_suite.status = TestStatus.SUCCESS > else: > test_suite.status = TestStatus.FAILURE > + suite_index = int(match.group(2)) > + if suite_index != expected_suite_index: > + print_with_timestamp( > + red('[ERROR] ') + 'expected_suite_index ' + > + str(expected_suite_index) + ', but got ' + > + str(suite_index)) > return True > else: > return False > @@ -221,7 +230,7 @@ def bubble_up_test_case_errors(test_suite: TestSuite) -> TestStatus: > max_test_case_status = bubble_up_errors(lambda x: x.status, test_suite.cases) > return max_status(max_test_case_status, test_suite.status) > > -def parse_test_suite(lines: List[str]) -> TestSuite: > +def parse_test_suite(lines: List[str], expected_suite_index: int) -> TestSuite: > if not lines: > return None > consume_non_diagnositic(lines) > @@ -240,7 +249,7 @@ def parse_test_suite(lines: List[str]) -> TestSuite: > test_suite.cases.append(test_case) > test_case = parse_test_case(lines, expected_test_case_num > 0) > expected_test_case_num -= 1 > - if parse_ok_not_ok_test_suite(lines, test_suite): > + if parse_ok_not_ok_test_suite(lines, test_suite, expected_suite_index): > test_suite.status = bubble_up_test_case_errors(test_suite) > return test_suite > elif not lines: > @@ -260,6 +269,17 @@ def parse_tap_header(lines: List[str]) -> bool: > else: > return False > > +TEST_PLAN = re.compile(r'[0-9]+\.\.([0-9]+)') > + > +def parse_test_plan(lines: List[str]) -> int: > + consume_non_diagnositic(lines) > + match = TEST_PLAN.match(lines[0]) > + if match: > + lines.pop(0) > + return int(match.group(1)) > + else: > + return None > + > def bubble_up_suite_errors(test_suite_list: List[TestSuite]) -> TestStatus: > return bubble_up_errors(lambda x: x.status, test_suite_list) > > @@ -268,19 +288,34 @@ def parse_test_result(lines: List[str]) -> TestResult: > return TestResult(TestStatus.NO_TESTS, [], lines) > consume_non_diagnositic(lines) > if not parse_tap_header(lines): > - return None > + return TestResult(TestStatus.NO_TESTS, [], lines) > + expected_test_suite_num = parse_test_plan(lines) > + if not expected_test_suite_num: > + return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines) > test_suites = [] > - test_suite = parse_test_suite(lines) > - while test_suite: > - test_suites.append(test_suite) > - test_suite = parse_test_suite(lines) > - return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines) > + for i in range(1, expected_test_suite_num + 1): > + test_suite = parse_test_suite(lines, i) > + if test_suite: > + test_suites.append(test_suite) > + else: > + print_with_timestamp( > + red('[ERROR] ') + ' expected ' + > + str(expected_test_suite_num) + > + ' test suites, but got ' + str(i - 2)) > + break > + test_suite = parse_test_suite(lines, -1) > + if test_suite: > + print_with_timestamp(red('[ERROR] ') + > + 'got unexpected test suite: ' + test_suite.name) > + if test_suites: > + return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines) > + else: > + return TestResult(TestStatus.NO_TESTS, [], lines) > > -def parse_run_tests(kernel_output) -> TestResult: > +def print_and_count_results(test_result: TestResult) -> None: > total_tests = 0 > failed_tests = 0 > crashed_tests = 0 > - test_result = parse_test_result(list(isolate_kunit_output(kernel_output))) > for test_suite in test_result.suites: > if test_suite.status == TestStatus.SUCCESS: > print_suite_divider(green('[PASSED] ') + test_suite.name) > @@ -302,6 +337,21 @@ def parse_run_tests(kernel_output) -> TestResult: > print_with_timestamp(red('[FAILED] ') + test_case.name) > print_log(map(yellow, test_case.log)) > print_with_timestamp('') > + return total_tests, failed_tests, crashed_tests > + > +def parse_run_tests(kernel_output) -> TestResult: > + total_tests = 0 > + failed_tests = 0 > + crashed_tests = 0 > + test_result = parse_test_result(list(isolate_kunit_output(kernel_output))) > + if test_result.status == TestStatus.NO_TESTS: > + print(red('[ERROR] ') + yellow('no tests run!')) > + elif test_result.status == TestStatus.FAILURE_TO_PARSE_TESTS: > + print(red('[ERROR] ') + yellow('could not parse test results!')) > + else: > + (total_tests, > + failed_tests, > + crashed_tests) = print_and_count_results(test_result) > print_with_timestamp(DIVIDER) > fmt = green if test_result.status == TestStatus.SUCCESS else red > print_with_timestamp( > diff --git a/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log b/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log > index 62ebc0288355c..bc0dc8fe35b76 100644 > --- a/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log > +++ b/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log > @@ -1,4 +1,5 @@ > TAP version 14 > +1..2 > # Subtest: sysctl_test > 1..8 > # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed > diff --git a/tools/testing/kunit/test_data/test_is_test_passed-crash.log b/tools/testing/kunit/test_data/test_is_test_passed-crash.log > index 0b249870c8be4..4d97f6708c4a5 100644 > --- a/tools/testing/kunit/test_data/test_is_test_passed-crash.log > +++ b/tools/testing/kunit/test_data/test_is_test_passed-crash.log > @@ -1,6 +1,7 @@ > printk: console [tty0] enabled > printk: console [mc-1] enabled > TAP version 14 > +1..2 > # Subtest: sysctl_test > 1..8 > # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed > diff --git a/tools/testing/kunit/test_data/test_is_test_passed-failure.log b/tools/testing/kunit/test_data/test_is_test_passed-failure.log > index 9e89d32d5667a..7a416497e3bec 100644 > --- a/tools/testing/kunit/test_data/test_is_test_passed-failure.log > +++ b/tools/testing/kunit/test_data/test_is_test_passed-failure.log > @@ -1,4 +1,5 @@ > TAP version 14 > +1..2 > # Subtest: sysctl_test > 1..8 > # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed >
On Tue, Feb 4, 2020 at 3:01 PM Frank Rowand <frowand.list@gmail.com> wrote: > > On 1/30/20 5:08 PM, Brendan Higgins wrote: > > TAP 14 allows an optional test plan to be emitted before the start of > > the start of testing[1]; this is valuable because it makes it possible > > for a test harness to detect whether the number of tests run matches the > > number of tests expected to be run, ensuring that no tests silently > > failed. > > > > Link[1]: https://github.com/isaacs/testanything.github.io/blob/tap14/tap-version-14-specification.md#the-plan > > Signed-off-by: Brendan Higgins <brendanhiggins@google.com> > > Reviewed-by: Stephen Boyd <sboyd@kernel.org> > > --- > > lib/kunit/executor.c | 17 +++++ > > lib/kunit/test.c | 11 --- > > tools/testing/kunit/kunit_parser.py | 74 ++++++++++++++++--- > > .../test_is_test_passed-all_passed.log | 1 + > > .../test_data/test_is_test_passed-crash.log | 1 + > > .../test_data/test_is_test_passed-failure.log | 1 + > > 6 files changed, 82 insertions(+), 23 deletions(-) > > > > diff --git a/lib/kunit/executor.c b/lib/kunit/executor.c > > index b75a46c560847..7fd16feff157e 100644 > > --- a/lib/kunit/executor.c > > +++ b/lib/kunit/executor.c > > @@ -11,11 +11,28 @@ extern struct kunit_suite * const * const __kunit_suites_end[]; > > > > #if IS_BUILTIN(CONFIG_KUNIT) > > > > +static void kunit_print_tap_header(void) > > +{ > > + struct kunit_suite * const * const *suites, * const *subsuite; > > + int num_of_suites = 0; > > + > > + for (suites = __kunit_suites_start; > > + suites < __kunit_suites_end; > > + suites++) > > + for (subsuite = *suites; *subsuite != NULL; subsuite++) > > + num_of_suites++; > > + > > + pr_info("TAP version 14\n"); > > + pr_info("1..%d\n", num_of_suites); > > +} > > + > > Subsuites are not in Linux 5.5, as far as I can tell, so it is hard to review > this. > > But I think this observation will still be valid even though I am not up > to speed on subsuites. "substitutes" are just an artifact from having to group together all suites that occur in a module. It should be mostly immaterial to most users, but we have to have all the suites that occur in a module grouped together in order to support KUnit tests as loadable modules. I just call them subsuites here because I didn't know what else to call them. Still, the relationship between them should be mostly ignored by users as you seem to recognize below. Also, module support for KUnit was just recently accepted into torvalds/master: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=08a3ef8f6b0b1341c670caba35f782c9a452d488 You should be able to apply this change to torvalds/master now. > num_of_suites is all test suites declared (as built in). In Linux 5.5, > each kunit_suite contains one or more test cases. Each test case leads > to either "ok" or "not ok". So instead of counting number of tests > suites, the number of test cases needs to be counted. Nope, but I think you understand correctly below... > Ahhh, I think my lack of visibility of what tree this series is against > is about to lead me astray! Are the test cases in the subsuite > implemented as what TAP 14 calls "subtests"? If so, then the above > code that counts suites instead of test cases would be correct. Yep! You got it! Test cases are "subtests" in TAP 14. So there is a ok/not ok line for each suite, and nested as subtests above that are indented ok/not ok lines for each test case. KUnit already does this today. We are just moving some of the printing logic out so we can count the suites and print a test plan. Technically, what we print today (before this change) is legal TAP 14. The test plan is not required. However, the test plan is very helpful as it makes it possible to know if some tests just never ran for some reason. > I'm spinning my wheels now, since I don't know what I am > reviewing against, so I'm going to stop reviewing, other > than maybe a quick scan. I just applied this patch series against torvalds/master which I pulled this morning, and it applied cleanly. I will specify in any subsequent revisions that this is based on linux-kselftest/kunit since that's where these patches will be applied, but they should apply against any rc of 5.6 just fine once they are cut. Cheers!
diff --git a/lib/kunit/executor.c b/lib/kunit/executor.c index b75a46c560847..7fd16feff157e 100644 --- a/lib/kunit/executor.c +++ b/lib/kunit/executor.c @@ -11,11 +11,28 @@ extern struct kunit_suite * const * const __kunit_suites_end[]; #if IS_BUILTIN(CONFIG_KUNIT) +static void kunit_print_tap_header(void) +{ + struct kunit_suite * const * const *suites, * const *subsuite; + int num_of_suites = 0; + + for (suites = __kunit_suites_start; + suites < __kunit_suites_end; + suites++) + for (subsuite = *suites; *subsuite != NULL; subsuite++) + num_of_suites++; + + pr_info("TAP version 14\n"); + pr_info("1..%d\n", num_of_suites); +} + int kunit_run_all_tests(void) { struct kunit_suite * const * const *suites, * const *subsuite; bool has_test_failed = false; + kunit_print_tap_header(); + for (suites = __kunit_suites_start; suites < __kunit_suites_end; suites++) { diff --git a/lib/kunit/test.c b/lib/kunit/test.c index 9242f932896c7..da56b94261b43 100644 --- a/lib/kunit/test.c +++ b/lib/kunit/test.c @@ -18,16 +18,6 @@ static void kunit_set_failure(struct kunit *test) WRITE_ONCE(test->success, false); } -static void kunit_print_tap_version(void) -{ - static bool kunit_has_printed_tap_version; - - if (!kunit_has_printed_tap_version) { - pr_info("TAP version 14\n"); - kunit_has_printed_tap_version = true; - } -} - static size_t kunit_test_cases_len(struct kunit_case *test_cases) { struct kunit_case *test_case; @@ -41,7 +31,6 @@ static size_t kunit_test_cases_len(struct kunit_case *test_cases) static void kunit_print_subtest_start(struct kunit_suite *suite) { - kunit_print_tap_version(); pr_info("\t# Subtest: %s\n", suite->name); pr_info("\t1..%zd\n", kunit_test_cases_len(suite->test_cases)); } diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py index 4ffbae0f67325..78b3bdd03b1e4 100644 --- a/tools/testing/kunit/kunit_parser.py +++ b/tools/testing/kunit/kunit_parser.py @@ -45,6 +45,7 @@ class TestStatus(Enum): FAILURE = auto() TEST_CRASHED = auto() NO_TESTS = auto() + FAILURE_TO_PARSE_TESTS = auto() kunit_start_re = re.compile(r'^TAP version [0-9]+$') kunit_end_re = re.compile('List of all partitions:') @@ -106,7 +107,7 @@ OkNotOkResult = namedtuple('OkNotOkResult', ['is_ok','description', 'text']) OK_NOT_OK_SUBTEST = re.compile(r'^\t(ok|not ok) [0-9]+ - (.*)$') -OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) [0-9]+ - (.*)$') +OK_NOT_OK_MODULE = re.compile(r'^(ok|not ok) ([0-9]+) - (.*)$') def parse_ok_not_ok_test_case(lines: List[str], test_case: TestCase, @@ -196,7 +197,9 @@ def max_status(left: TestStatus, right: TestStatus) -> TestStatus: else: return TestStatus.SUCCESS -def parse_ok_not_ok_test_suite(lines: List[str], test_suite: TestSuite) -> bool: +def parse_ok_not_ok_test_suite(lines: List[str], + test_suite: TestSuite, + expected_suite_index: int) -> bool: consume_non_diagnositic(lines) if not lines: test_suite.status = TestStatus.TEST_CRASHED @@ -209,6 +212,12 @@ def parse_ok_not_ok_test_suite(lines: List[str], test_suite: TestSuite) -> bool: test_suite.status = TestStatus.SUCCESS else: test_suite.status = TestStatus.FAILURE + suite_index = int(match.group(2)) + if suite_index != expected_suite_index: + print_with_timestamp( + red('[ERROR] ') + 'expected_suite_index ' + + str(expected_suite_index) + ', but got ' + + str(suite_index)) return True else: return False @@ -221,7 +230,7 @@ def bubble_up_test_case_errors(test_suite: TestSuite) -> TestStatus: max_test_case_status = bubble_up_errors(lambda x: x.status, test_suite.cases) return max_status(max_test_case_status, test_suite.status) -def parse_test_suite(lines: List[str]) -> TestSuite: +def parse_test_suite(lines: List[str], expected_suite_index: int) -> TestSuite: if not lines: return None consume_non_diagnositic(lines) @@ -240,7 +249,7 @@ def parse_test_suite(lines: List[str]) -> TestSuite: test_suite.cases.append(test_case) test_case = parse_test_case(lines, expected_test_case_num > 0) expected_test_case_num -= 1 - if parse_ok_not_ok_test_suite(lines, test_suite): + if parse_ok_not_ok_test_suite(lines, test_suite, expected_suite_index): test_suite.status = bubble_up_test_case_errors(test_suite) return test_suite elif not lines: @@ -260,6 +269,17 @@ def parse_tap_header(lines: List[str]) -> bool: else: return False +TEST_PLAN = re.compile(r'[0-9]+\.\.([0-9]+)') + +def parse_test_plan(lines: List[str]) -> int: + consume_non_diagnositic(lines) + match = TEST_PLAN.match(lines[0]) + if match: + lines.pop(0) + return int(match.group(1)) + else: + return None + def bubble_up_suite_errors(test_suite_list: List[TestSuite]) -> TestStatus: return bubble_up_errors(lambda x: x.status, test_suite_list) @@ -268,19 +288,34 @@ def parse_test_result(lines: List[str]) -> TestResult: return TestResult(TestStatus.NO_TESTS, [], lines) consume_non_diagnositic(lines) if not parse_tap_header(lines): - return None + return TestResult(TestStatus.NO_TESTS, [], lines) + expected_test_suite_num = parse_test_plan(lines) + if not expected_test_suite_num: + return TestResult(TestStatus.FAILURE_TO_PARSE_TESTS, [], lines) test_suites = [] - test_suite = parse_test_suite(lines) - while test_suite: - test_suites.append(test_suite) - test_suite = parse_test_suite(lines) - return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines) + for i in range(1, expected_test_suite_num + 1): + test_suite = parse_test_suite(lines, i) + if test_suite: + test_suites.append(test_suite) + else: + print_with_timestamp( + red('[ERROR] ') + ' expected ' + + str(expected_test_suite_num) + + ' test suites, but got ' + str(i - 2)) + break + test_suite = parse_test_suite(lines, -1) + if test_suite: + print_with_timestamp(red('[ERROR] ') + + 'got unexpected test suite: ' + test_suite.name) + if test_suites: + return TestResult(bubble_up_suite_errors(test_suites), test_suites, lines) + else: + return TestResult(TestStatus.NO_TESTS, [], lines) -def parse_run_tests(kernel_output) -> TestResult: +def print_and_count_results(test_result: TestResult) -> None: total_tests = 0 failed_tests = 0 crashed_tests = 0 - test_result = parse_test_result(list(isolate_kunit_output(kernel_output))) for test_suite in test_result.suites: if test_suite.status == TestStatus.SUCCESS: print_suite_divider(green('[PASSED] ') + test_suite.name) @@ -302,6 +337,21 @@ def parse_run_tests(kernel_output) -> TestResult: print_with_timestamp(red('[FAILED] ') + test_case.name) print_log(map(yellow, test_case.log)) print_with_timestamp('') + return total_tests, failed_tests, crashed_tests + +def parse_run_tests(kernel_output) -> TestResult: + total_tests = 0 + failed_tests = 0 + crashed_tests = 0 + test_result = parse_test_result(list(isolate_kunit_output(kernel_output))) + if test_result.status == TestStatus.NO_TESTS: + print(red('[ERROR] ') + yellow('no tests run!')) + elif test_result.status == TestStatus.FAILURE_TO_PARSE_TESTS: + print(red('[ERROR] ') + yellow('could not parse test results!')) + else: + (total_tests, + failed_tests, + crashed_tests) = print_and_count_results(test_result) print_with_timestamp(DIVIDER) fmt = green if test_result.status == TestStatus.SUCCESS else red print_with_timestamp( diff --git a/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log b/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log index 62ebc0288355c..bc0dc8fe35b76 100644 --- a/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log +++ b/tools/testing/kunit/test_data/test_is_test_passed-all_passed.log @@ -1,4 +1,5 @@ TAP version 14 +1..2 # Subtest: sysctl_test 1..8 # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed diff --git a/tools/testing/kunit/test_data/test_is_test_passed-crash.log b/tools/testing/kunit/test_data/test_is_test_passed-crash.log index 0b249870c8be4..4d97f6708c4a5 100644 --- a/tools/testing/kunit/test_data/test_is_test_passed-crash.log +++ b/tools/testing/kunit/test_data/test_is_test_passed-crash.log @@ -1,6 +1,7 @@ printk: console [tty0] enabled printk: console [mc-1] enabled TAP version 14 +1..2 # Subtest: sysctl_test 1..8 # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed diff --git a/tools/testing/kunit/test_data/test_is_test_passed-failure.log b/tools/testing/kunit/test_data/test_is_test_passed-failure.log index 9e89d32d5667a..7a416497e3bec 100644 --- a/tools/testing/kunit/test_data/test_is_test_passed-failure.log +++ b/tools/testing/kunit/test_data/test_is_test_passed-failure.log @@ -1,4 +1,5 @@ TAP version 14 +1..2 # Subtest: sysctl_test 1..8 # sysctl_test_dointvec_null_tbl_data: sysctl_test_dointvec_null_tbl_data passed