2022-11-29 00:55:14

by Daniel Latypov

[permalink] [raw]
Subject: [PATCH] kunit: tool: don't include KTAP headers and the like in the test log

We print the "test log" on failure.
This is meant to be all the kernel output that happened during the test.

But we also include the special KTAP lines in it, which are often
redundant.

E.g. we include the "not ok" line in the log, right before we print
that the test case failed...
[13:51:48] Expected 2 + 1 == 2, but
[13:51:48] 2 + 1 == 3 (0x3)
[13:51:48] not ok 1 example_simple_test
[13:51:48] [FAILED] example_simple_test

More full example after this patch:
[13:51:48] =================== example (4 subtests) ===================
[13:51:48] # example_simple_test: initializing
[13:51:48] # example_simple_test: EXPECTATION FAILED at lib/kunit/kunit-example-test.c:29
[13:51:48] Expected 2 + 1 == 2, but
[13:51:48] 2 + 1 == 3 (0x3)
[13:51:48] [FAILED] example_simple_test

Signed-off-by: Daniel Latypov <[email protected]>
---
tools/testing/kunit/kunit_parser.py | 8 ++++----
tools/testing/kunit/kunit_tool_test.py | 17 +++++++++++++++++
2 files changed, 21 insertions(+), 4 deletions(-)

diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py
index 4cc2f8b7ecd0..99b8f058db40 100644
--- a/tools/testing/kunit/kunit_parser.py
+++ b/tools/testing/kunit/kunit_parser.py
@@ -295,7 +295,7 @@ def parse_ktap_header(lines: LineStream, test: Test) -> bool:
check_version(version_num, TAP_VERSIONS, 'TAP', test)
else:
return False
- test.log.append(lines.pop())
+ lines.pop()
return True

TEST_HEADER = re.compile(r'^# Subtest: (.*)$')
@@ -318,8 +318,8 @@ def parse_test_header(lines: LineStream, test: Test) -> bool:
match = TEST_HEADER.match(lines.peek())
if not match:
return False
- test.log.append(lines.pop())
test.name = match.group(1)
+ lines.pop()
return True

TEST_PLAN = re.compile(r'1\.\.([0-9]+)')
@@ -345,9 +345,9 @@ def parse_test_plan(lines: LineStream, test: Test) -> bool:
if not match:
test.expected_count = None
return False
- test.log.append(lines.pop())
expected_count = int(match.group(1))
test.expected_count = expected_count
+ lines.pop()
return True

TEST_RESULT = re.compile(r'^(ok|not ok) ([0-9]+) (- )?([^#]*)( # .*)?$')
@@ -409,7 +409,7 @@ def parse_test_result(lines: LineStream, test: Test,
# Check if line matches test result line format
if not match:
return False
- test.log.append(lines.pop())
+ lines.pop()

# Set name of test object
if skip_match:
diff --git a/tools/testing/kunit/kunit_tool_test.py b/tools/testing/kunit/kunit_tool_test.py
index d7f669cbf2a8..1ef921ac4331 100755
--- a/tools/testing/kunit/kunit_tool_test.py
+++ b/tools/testing/kunit/kunit_tool_test.py
@@ -84,6 +84,10 @@ class KUnitParserTest(unittest.TestCase):
self.print_mock = mock.patch('kunit_printer.Printer.print').start()
self.addCleanup(mock.patch.stopall)

+ def noPrintCallContains(self, substr: str):
+ for call in self.print_mock.mock_calls:
+ self.assertNotIn(substr, call.args[0])
+
def assertContains(self, needle: str, haystack: kunit_parser.LineStream):
# Clone the iterator so we can print the contents on failure.
copy, backup = itertools.tee(haystack)
@@ -327,6 +331,19 @@ class KUnitParserTest(unittest.TestCase):
result = kunit_parser.parse_run_tests(file.readlines())
self.print_mock.assert_any_call(StrContains('suite (1 subtest)'))

+ def test_show_test_output_on_failure(self):
+ output = """
+ KTAP version 1
+ 1..1
+ Test output.
+ not ok 1 test1
+ """
+ result = kunit_parser.parse_run_tests(output.splitlines())
+ self.assertEqual(kunit_parser.TestStatus.FAILURE, result.status)
+
+ self.print_mock.assert_any_call(StrContains('Test output.'))
+ self.noPrintCallContains('not ok 1 test1')
+
def line_stream_from_strs(strs: Iterable[str]) -> kunit_parser.LineStream:
return kunit_parser.LineStream(enumerate(strs, start=1))


base-commit: 11300092f6f4dc4103ac4bd950d62f94effc736a
--
2.38.1.584.g0f3c55d4c2-goog


2022-11-29 08:35:00

by David Gow

[permalink] [raw]
Subject: Re: [PATCH] kunit: tool: don't include KTAP headers and the like in the test log

On Tue, Nov 29, 2022 at 8:12 AM Daniel Latypov <[email protected]> wrote:
>
> We print the "test log" on failure.
> This is meant to be all the kernel output that happened during the test.
>
> But we also include the special KTAP lines in it, which are often
> redundant.
>
> E.g. we include the "not ok" line in the log, right before we print
> that the test case failed...
> [13:51:48] Expected 2 + 1 == 2, but
> [13:51:48] 2 + 1 == 3 (0x3)
> [13:51:48] not ok 1 example_simple_test
> [13:51:48] [FAILED] example_simple_test
>
> More full example after this patch:
> [13:51:48] =================== example (4 subtests) ===================
> [13:51:48] # example_simple_test: initializing
> [13:51:48] # example_simple_test: EXPECTATION FAILED at lib/kunit/kunit-example-test.c:29
> [13:51:48] Expected 2 + 1 == 2, but
> [13:51:48] 2 + 1 == 3 (0x3)
> [13:51:48] [FAILED] example_simple_test
>
> Signed-off-by: Daniel Latypov <[email protected]>
> ---

I totally agree we should skip these from the log. (Unless
--raw_output is enabled, but that obviously doesn't apply.)

Going forward, I think we should also probably disable
kunit.stats_enabled when running via kunit.py, too (again, unless
--raw_output is used.)

In any case, this looks good and works well here.

Reviewed-by: David Gow <[email protected]>

Cheers,
-- David

> tools/testing/kunit/kunit_parser.py | 8 ++++----
> tools/testing/kunit/kunit_tool_test.py | 17 +++++++++++++++++
> 2 files changed, 21 insertions(+), 4 deletions(-)
>
> diff --git a/tools/testing/kunit/kunit_parser.py b/tools/testing/kunit/kunit_parser.py
> index 4cc2f8b7ecd0..99b8f058db40 100644
> --- a/tools/testing/kunit/kunit_parser.py
> +++ b/tools/testing/kunit/kunit_parser.py
> @@ -295,7 +295,7 @@ def parse_ktap_header(lines: LineStream, test: Test) -> bool:
> check_version(version_num, TAP_VERSIONS, 'TAP', test)
> else:
> return False
> - test.log.append(lines.pop())
> + lines.pop()
> return True
>
> TEST_HEADER = re.compile(r'^# Subtest: (.*)$')
> @@ -318,8 +318,8 @@ def parse_test_header(lines: LineStream, test: Test) -> bool:
> match = TEST_HEADER.match(lines.peek())
> if not match:
> return False
> - test.log.append(lines.pop())
> test.name = match.group(1)
> + lines.pop()
> return True
>
> TEST_PLAN = re.compile(r'1\.\.([0-9]+)')
> @@ -345,9 +345,9 @@ def parse_test_plan(lines: LineStream, test: Test) -> bool:
> if not match:
> test.expected_count = None
> return False
> - test.log.append(lines.pop())
> expected_count = int(match.group(1))
> test.expected_count = expected_count
> + lines.pop()
> return True
>
> TEST_RESULT = re.compile(r'^(ok|not ok) ([0-9]+) (- )?([^#]*)( # .*)?$')
> @@ -409,7 +409,7 @@ def parse_test_result(lines: LineStream, test: Test,
> # Check if line matches test result line format
> if not match:
> return False
> - test.log.append(lines.pop())
> + lines.pop()
>
> # Set name of test object
> if skip_match:
> diff --git a/tools/testing/kunit/kunit_tool_test.py b/tools/testing/kunit/kunit_tool_test.py
> index d7f669cbf2a8..1ef921ac4331 100755
> --- a/tools/testing/kunit/kunit_tool_test.py
> +++ b/tools/testing/kunit/kunit_tool_test.py
> @@ -84,6 +84,10 @@ class KUnitParserTest(unittest.TestCase):
> self.print_mock = mock.patch('kunit_printer.Printer.print').start()
> self.addCleanup(mock.patch.stopall)
>
> + def noPrintCallContains(self, substr: str):
> + for call in self.print_mock.mock_calls:
> + self.assertNotIn(substr, call.args[0])
> +
> def assertContains(self, needle: str, haystack: kunit_parser.LineStream):
> # Clone the iterator so we can print the contents on failure.
> copy, backup = itertools.tee(haystack)
> @@ -327,6 +331,19 @@ class KUnitParserTest(unittest.TestCase):
> result = kunit_parser.parse_run_tests(file.readlines())
> self.print_mock.assert_any_call(StrContains('suite (1 subtest)'))
>
> + def test_show_test_output_on_failure(self):
> + output = """
> + KTAP version 1
> + 1..1
> + Test output.
> + not ok 1 test1
> + """
> + result = kunit_parser.parse_run_tests(output.splitlines())
> + self.assertEqual(kunit_parser.TestStatus.FAILURE, result.status)
> +
> + self.print_mock.assert_any_call(StrContains('Test output.'))
> + self.noPrintCallContains('not ok 1 test1')
> +
> def line_stream_from_strs(strs: Iterable[str]) -> kunit_parser.LineStream:
> return kunit_parser.LineStream(enumerate(strs, start=1))
>
>
> base-commit: 11300092f6f4dc4103ac4bd950d62f94effc736a
> --
> 2.38.1.584.g0f3c55d4c2-goog
>


Attachments:
smime.p7s (3.91 kB)
S/MIME Cryptographic Signature

2022-11-29 18:05:09

by Daniel Latypov

[permalink] [raw]
Subject: Re: [PATCH] kunit: tool: don't include KTAP headers and the like in the test log

On Tue, Nov 29, 2022 at 12:31 AM David Gow <[email protected]> wrote:
>
> On Tue, Nov 29, 2022 at 8:12 AM Daniel Latypov <[email protected]> wrote:
> >
> > We print the "test log" on failure.
> > This is meant to be all the kernel output that happened during the test.
> >
> > But we also include the special KTAP lines in it, which are often
> > redundant.
> >
> > E.g. we include the "not ok" line in the log, right before we print
> > that the test case failed...
> > [13:51:48] Expected 2 + 1 == 2, but
> > [13:51:48] 2 + 1 == 3 (0x3)
> > [13:51:48] not ok 1 example_simple_test
> > [13:51:48] [FAILED] example_simple_test
> >
> > More full example after this patch:
> > [13:51:48] =================== example (4 subtests) ===================
> > [13:51:48] # example_simple_test: initializing
> > [13:51:48] # example_simple_test: EXPECTATION FAILED at lib/kunit/kunit-example-test.c:29
> > [13:51:48] Expected 2 + 1 == 2, but
> > [13:51:48] 2 + 1 == 3 (0x3)
> > [13:51:48] [FAILED] example_simple_test
> >
> > Signed-off-by: Daniel Latypov <[email protected]>
> > ---
>
> I totally agree we should skip these from the log. (Unless
> --raw_output is enabled, but that obviously doesn't apply.)
>
> Going forward, I think we should also probably disable
> kunit.stats_enabled when running via kunit.py, too (again, unless
> --raw_output is used.)

I considered including that as a patch 2/2 here.
But changing the behavior like that felt a bit iffy.

We've basically been telling people that looking at .kunit/test.log is
logically equivalent to running with kunit.py run --raw_output.
That would no longer be true after such a change.
So I'm torn between that and automatically filtering them out in the
parser side.

Cons of tweaking args based on --raw_output
* now more magic, harder to explain (see above)
* people might find test counts useful when looking at test.log

Cons of filtering out test counts in the parser
* risks false positives: filtering out other lines besides test counts
* when there's missing output, this is less debuggable
* 99% of users are *not* going to dig into the python code
* but IMO users are fairly likely to notice the extra
kunit.stats_enabled=0 cmdline arg

And overall, the benefit of hiding these is very small and cosmetic in nature.
So that means making a tradeoff to do so feels more iffy.

The hiding done in this patch seemed fine since there was no tradeoff,
we just needed to stop including lines we've already recognized as
KTAP directives.

>
> In any case, this looks good and works well here.
>
> Reviewed-by: David Gow <[email protected]>
>
> Cheers,
> -- David

2022-11-29 22:04:09

by Rae Moar

[permalink] [raw]
Subject: Re: [PATCH] kunit: tool: don't include KTAP headers and the like in the test log

On Tue, Nov 29, 2022 at 12:09 PM Daniel Latypov <[email protected]> wrote:
>
> On Tue, Nov 29, 2022 at 12:31 AM David Gow <[email protected]> wrote:
> >
> > On Tue, Nov 29, 2022 at 8:12 AM Daniel Latypov <[email protected]> wrote:
> > >
> > > We print the "test log" on failure.
> > > This is meant to be all the kernel output that happened during the test.
> > >
> > > But we also include the special KTAP lines in it, which are often
> > > redundant.
> > >
> > > E.g. we include the "not ok" line in the log, right before we print
> > > that the test case failed...
> > > [13:51:48] Expected 2 + 1 == 2, but
> > > [13:51:48] 2 + 1 == 3 (0x3)
> > > [13:51:48] not ok 1 example_simple_test
> > > [13:51:48] [FAILED] example_simple_test
> > >
> > > More full example after this patch:
> > > [13:51:48] =================== example (4 subtests) ===================
> > > [13:51:48] # example_simple_test: initializing
> > > [13:51:48] # example_simple_test: EXPECTATION FAILED at lib/kunit/kunit-example-test.c:29
> > > [13:51:48] Expected 2 + 1 == 2, but
> > > [13:51:48] 2 + 1 == 3 (0x3)
> > > [13:51:48] [FAILED] example_simple_test
> > >
> > > Signed-off-by: Daniel Latypov <[email protected]>
> > > ---
> >
> > I totally agree we should skip these from the log. (Unless
> > --raw_output is enabled, but that obviously doesn't apply.)
> >
> > Going forward, I think we should also probably disable
> > kunit.stats_enabled when running via kunit.py, too (again, unless
> > --raw_output is used.)
>
> I considered including that as a patch 2/2 here.
> But changing the behavior like that felt a bit iffy.
>
> We've basically been telling people that looking at .kunit/test.log is
> logically equivalent to running with kunit.py run --raw_output.
> That would no longer be true after such a change.
> So I'm torn between that and automatically filtering them out in the
> parser side.
>
> Cons of tweaking args based on --raw_output
> * now more magic, harder to explain (see above)
> * people might find test counts useful when looking at test.log
>
> Cons of filtering out test counts in the parser
> * risks false positives: filtering out other lines besides test counts
> * when there's missing output, this is less debuggable
> * 99% of users are *not* going to dig into the python code
> * but IMO users are fairly likely to notice the extra
> kunit.stats_enabled=0 cmdline arg
>
> And overall, the benefit of hiding these is very small and cosmetic in nature.
> So that means making a tradeoff to do so feels more iffy.
>
> The hiding done in this patch seemed fine since there was no tradeoff,
> we just needed to stop including lines we've already recognized as
> KTAP directives.
>

I tested and reviewed the changes and this all looks good to me. Nice
to no longer print the KTAP version line or test result lines in the
test logs.

To comment on the above discussion: of the options discussed, I would
prefer to either tweak the stats_enabled arg based on the raw_output
flag or simply leave the output as is. I personally think that
filtering out the test counts using the parser would be messy and does
not match the precedent set by this patch to not include only
recognized KTAP directives in the test log (with the exception of the
"# Subtest" line, which I am planning to propose for KTAP v2).

I agree with the cons to tweaking the arg based on the raw_output flag
Daniel mentioned above. Additionally, I think including the test
counts line in the log could actually be quite useful in the instance
of combining module results. Since the parser does not include a test
summary line with counts for failed subtests, this would save the time
counting up the number of failed/skipped/passed subtests for a
specific module. This would, in general, also apply to failing tests
with a large number of subtests within them.

That being said I would be ok with tweaking the stats_enabled arg
based on the raw_output flag as well. It does look cleaner without the
counts printed in the log, especially when there are only a few number
of tests or the test summary line of the parser repeats the same
information as the test counts line.

Reviewed-by: Rae Moar <[email protected]>

Thanks!
Rae

> >
> > In any case, this looks good and works well here.
> >
> > Reviewed-by: David Gow <[email protected]>
> >
> > Cheers,
> > -- David