mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-12-02 01:19:49 +00:00
10463136fd
When a single test is run we don't need to buffer the test output. This has the unfortunate side effect of suppressing test output, in particular the binman output directory normally printed with the -X option. This is a huge problem since it blocks debugging of tests. We don't actually know how many tests will be run when we set up the suite, so as a work-around, assume that test_name being specified indicates that there is likely only one. Signed-off-by: Simon Glass <sjg@chromium.org>
247 lines
9.5 KiB
Python
247 lines
9.5 KiB
Python
# SPDX-License-Identifier: GPL-2.0+
|
|
#
|
|
# Copyright (c) 2016 Google, Inc
|
|
#
|
|
|
|
from contextlib import contextmanager
|
|
import doctest
|
|
import glob
|
|
import multiprocessing
|
|
import os
|
|
import sys
|
|
import unittest
|
|
|
|
from patman import command
|
|
|
|
from io import StringIO
|
|
|
|
buffer_outputs = True
|
|
use_concurrent = True
|
|
try:
|
|
from concurrencytest.concurrencytest import ConcurrentTestSuite
|
|
from concurrencytest.concurrencytest import fork_for_tests
|
|
except:
|
|
use_concurrent = False
|
|
|
|
|
|
def run_test_coverage(prog, filter_fname, exclude_list, build_dir, required=None,
|
|
extra_args=None):
|
|
"""Run tests and check that we get 100% coverage
|
|
|
|
Args:
|
|
prog: Program to run (with be passed a '-t' argument to run tests
|
|
filter_fname: Normally all *.py files in the program's directory will
|
|
be included. If this is not None, then it is used to filter the
|
|
list so that only filenames that don't contain filter_fname are
|
|
included.
|
|
exclude_list: List of file patterns to exclude from the coverage
|
|
calculation
|
|
build_dir: Build directory, used to locate libfdt.py
|
|
required: List of modules which must be in the coverage report
|
|
extra_args (str): Extra arguments to pass to the tool before the -t/test
|
|
arg
|
|
|
|
Raises:
|
|
ValueError if the code coverage is not 100%
|
|
"""
|
|
# This uses the build output from sandbox_spl to get _libfdt.so
|
|
path = os.path.dirname(prog)
|
|
if filter_fname:
|
|
glob_list = glob.glob(os.path.join(path, '*.py'))
|
|
glob_list = [fname for fname in glob_list if filter_fname in fname]
|
|
else:
|
|
glob_list = []
|
|
glob_list += exclude_list
|
|
glob_list += ['*libfdt.py', '*site-packages*', '*dist-packages*']
|
|
glob_list += ['*concurrencytest*']
|
|
test_cmd = 'test' if 'binman' in prog or 'patman' in prog else '-t'
|
|
prefix = ''
|
|
if build_dir:
|
|
prefix = 'PYTHONPATH=$PYTHONPATH:%s/sandbox_spl/tools ' % build_dir
|
|
cmd = ('%spython3-coverage run '
|
|
'--omit "%s" %s %s %s -P1' % (prefix, ','.join(glob_list),
|
|
prog, extra_args or '', test_cmd))
|
|
os.system(cmd)
|
|
stdout = command.output('python3-coverage', 'report')
|
|
lines = stdout.splitlines()
|
|
if required:
|
|
# Convert '/path/to/name.py' just the module name 'name'
|
|
test_set = set([os.path.splitext(os.path.basename(line.split()[0]))[0]
|
|
for line in lines if '/etype/' in line])
|
|
missing_list = required
|
|
missing_list.discard('__init__')
|
|
missing_list.difference_update(test_set)
|
|
if missing_list:
|
|
print('Missing tests for %s' % (', '.join(missing_list)))
|
|
print(stdout)
|
|
ok = False
|
|
|
|
coverage = lines[-1].split(' ')[-1]
|
|
ok = True
|
|
print(coverage)
|
|
if coverage != '100%':
|
|
print(stdout)
|
|
print("To get a report in 'htmlcov/index.html', type: python3-coverage html")
|
|
print('Coverage error: %s, but should be 100%%' % coverage)
|
|
ok = False
|
|
if not ok:
|
|
raise ValueError('Test coverage failure')
|
|
|
|
|
|
# Use this to suppress stdout/stderr output:
|
|
# with capture_sys_output() as (stdout, stderr)
|
|
# ...do something...
|
|
@contextmanager
|
|
def capture_sys_output():
|
|
capture_out, capture_err = StringIO(), StringIO()
|
|
old_out, old_err = sys.stdout, sys.stderr
|
|
try:
|
|
sys.stdout, sys.stderr = capture_out, capture_err
|
|
yield capture_out, capture_err
|
|
finally:
|
|
sys.stdout, sys.stderr = old_out, old_err
|
|
|
|
|
|
class FullTextTestResult(unittest.TextTestResult):
|
|
"""A test result class that can print extended text results to a stream
|
|
|
|
This is meant to be used by a TestRunner as a result class. Like
|
|
TextTestResult, this prints out the names of tests as they are run,
|
|
errors as they occur, and a summary of the results at the end of the
|
|
test run. Beyond those, this prints information about skipped tests,
|
|
expected failures and unexpected successes.
|
|
|
|
Args:
|
|
stream: A file-like object to write results to
|
|
descriptions (bool): True to print descriptions with test names
|
|
verbosity (int): Detail of printed output per test as they run
|
|
Test stdout and stderr always get printed when buffering
|
|
them is disabled by the test runner. In addition to that,
|
|
0: Print nothing
|
|
1: Print a dot per test
|
|
2: Print test names
|
|
3: Print test names, and buffered outputs for failing tests
|
|
"""
|
|
def __init__(self, stream, descriptions, verbosity):
|
|
self.verbosity = verbosity
|
|
super().__init__(stream, descriptions, verbosity)
|
|
|
|
def printErrors(self):
|
|
"Called by TestRunner after test run to summarize the tests"
|
|
# The parent class doesn't keep unexpected successes in the same
|
|
# format as the rest. Adapt it to what printErrorList expects.
|
|
unexpected_successes = [
|
|
(test, 'Test was expected to fail, but succeeded.\n')
|
|
for test in self.unexpectedSuccesses
|
|
]
|
|
|
|
super().printErrors() # FAIL and ERROR
|
|
self.printErrorList('SKIP', self.skipped)
|
|
self.printErrorList('XFAIL', self.expectedFailures)
|
|
self.printErrorList('XPASS', unexpected_successes)
|
|
|
|
def addError(self, test, err):
|
|
"""Called when an error has occurred."""
|
|
super().addError(test, err)
|
|
self._mirrorOutput &= self.verbosity >= 3
|
|
|
|
def addFailure(self, test, err):
|
|
"""Called when a test has failed."""
|
|
super().addFailure(test, err)
|
|
self._mirrorOutput &= self.verbosity >= 3
|
|
|
|
def addSubTest(self, test, subtest, err):
|
|
"""Called at the end of a subtest."""
|
|
super().addSubTest(test, subtest, err)
|
|
self._mirrorOutput &= self.verbosity >= 3
|
|
|
|
def addSuccess(self, test):
|
|
"""Called when a test has completed successfully"""
|
|
super().addSuccess(test)
|
|
# Don't print stdout/stderr for successful tests
|
|
self._mirrorOutput = False
|
|
|
|
def addSkip(self, test, reason):
|
|
"""Called when a test is skipped."""
|
|
# Add empty line to keep spacing consistent with other results
|
|
if not reason.endswith('\n'):
|
|
reason += '\n'
|
|
super().addSkip(test, reason)
|
|
self._mirrorOutput &= self.verbosity >= 3
|
|
|
|
def addExpectedFailure(self, test, err):
|
|
"""Called when an expected failure/error occurred."""
|
|
super().addExpectedFailure(test, err)
|
|
self._mirrorOutput &= self.verbosity >= 3
|
|
|
|
|
|
def run_test_suites(toolname, debug, verbosity, test_preserve_dirs, processes,
|
|
test_name, toolpath, class_and_module_list):
|
|
"""Run a series of test suites and collect the results
|
|
|
|
Args:
|
|
toolname: Name of the tool that ran the tests
|
|
debug: True to enable debugging, which shows a full stack trace on error
|
|
verbosity: Verbosity level to use (0-4)
|
|
test_preserve_dirs: True to preserve the input directory used by tests
|
|
so that it can be examined afterwards (only useful for debugging
|
|
tests). If a single test is selected (in args[0]) it also preserves
|
|
the output directory for this test. Both directories are displayed
|
|
on the command line.
|
|
processes: Number of processes to use to run tests (None=same as #CPUs)
|
|
test_name: Name of test to run, or None for all
|
|
toolpath: List of paths to use for tools
|
|
class_and_module_list: List of test classes (type class) and module
|
|
names (type str) to run
|
|
"""
|
|
sys.argv = [sys.argv[0]]
|
|
if debug:
|
|
sys.argv.append('-D')
|
|
if verbosity:
|
|
sys.argv.append('-v%d' % verbosity)
|
|
if toolpath:
|
|
for path in toolpath:
|
|
sys.argv += ['--toolpath', path]
|
|
|
|
suite = unittest.TestSuite()
|
|
loader = unittest.TestLoader()
|
|
runner = unittest.TextTestRunner(
|
|
stream=sys.stdout,
|
|
verbosity=(1 if verbosity is None else verbosity),
|
|
buffer=False if test_name else buffer_outputs,
|
|
resultclass=FullTextTestResult,
|
|
)
|
|
|
|
if use_concurrent and processes != 1:
|
|
suite = ConcurrentTestSuite(suite,
|
|
fork_for_tests(processes or multiprocessing.cpu_count(),
|
|
buffer=False if test_name else buffer_outputs))
|
|
|
|
for module in class_and_module_list:
|
|
if isinstance(module, str) and (not test_name or test_name == module):
|
|
suite.addTests(doctest.DocTestSuite(module))
|
|
|
|
for module in class_and_module_list:
|
|
if isinstance(module, str):
|
|
continue
|
|
# Test the test module about our arguments, if it is interested
|
|
if hasattr(module, 'setup_test_args'):
|
|
setup_test_args = getattr(module, 'setup_test_args')
|
|
setup_test_args(preserve_indir=test_preserve_dirs,
|
|
preserve_outdirs=test_preserve_dirs and test_name is not None,
|
|
toolpath=toolpath, verbosity=verbosity)
|
|
if test_name:
|
|
# Since Python v3.5 If an ImportError or AttributeError occurs
|
|
# while traversing a name then a synthetic test that raises that
|
|
# error when run will be returned. Check that the requested test
|
|
# exists, otherwise these errors are included in the results.
|
|
if test_name in loader.getTestCaseNames(module):
|
|
suite.addTests(loader.loadTestsFromName(test_name, module))
|
|
else:
|
|
suite.addTests(loader.loadTestsFromTestCase(module))
|
|
|
|
print(f" Running {toolname} tests ".center(70, "="))
|
|
result = runner.run(suite)
|
|
print()
|
|
|
|
return result
|