diff options
Diffstat (limited to 'meta/lib/oeqa/runtime/cases/ptest.py')
-rw-r--r-- | meta/lib/oeqa/runtime/cases/ptest.py | 140 |
1 files changed, 89 insertions, 51 deletions
diff --git a/meta/lib/oeqa/runtime/cases/ptest.py b/meta/lib/oeqa/runtime/cases/ptest.py index ec8c038a56..fbaeb84d00 100644 --- a/meta/lib/oeqa/runtime/cases/ptest.py +++ b/meta/lib/oeqa/runtime/cases/ptest.py @@ -1,71 +1,65 @@ +# +# Copyright OpenEmbedded Contributors +# +# SPDX-License-Identifier: MIT +# + +import os +import unittest +import pprint +import datetime + from oeqa.runtime.case import OERuntimeTestCase from oeqa.core.decorator.depends import OETestDepends -from oeqa.core.decorator.oeid import OETestID from oeqa.core.decorator.data import skipIfNotFeature -from oeqa.utils.logparser import Lparser, Result +from oeqa.runtime.decorator.package import OEHasPackage +from oeqa.utils.logparser import PtestParser class PtestRunnerTest(OERuntimeTestCase): - # a ptest log parser - def parse_ptest(self, logfile): - parser = Lparser(test_0_pass_regex="^PASS:(.+)", - test_0_fail_regex="^FAIL:(.+)", - test_0_skip_regex="^SKIP:(.+)", - section_0_begin_regex="^BEGIN: .*/(.+)/ptest", - section_0_end_regex="^END: .*/(.+)/ptest") - parser.init() - result = Result() - - with open(logfile, errors='replace') as f: - for line in f: - result_tuple = parser.parse_line(line) - if not result_tuple: - continue - result_tuple = line_type, category, status, name = parser.parse_line(line) - - if line_type == 'section' and status == 'begin': - current_section = name - continue - - if line_type == 'section' and status == 'end': - current_section = None - continue - - if line_type == 'test' and status == 'pass': - result.store(current_section, name, status) - continue - - if line_type == 'test' and status == 'fail': - result.store(current_section, name, status) - continue - - if line_type == 'test' and status == 'skip': - result.store(current_section, name, status) - continue - - result.sort_tests() - return result - - @OETestID(1600) @skipIfNotFeature('ptest', 'Test requires ptest to be in DISTRO_FEATURES') - @skipIfNotFeature('ptest-pkgs', 'Test requires ptest-pkgs to be in IMAGE_FEATURES') @OETestDepends(['ssh.SSHTest.test_ssh']) - def test_ptestrunner(self): - import datetime + @OEHasPackage(['ptest-runner']) + @unittest.expectedFailure + def test_ptestrunner_expectfail(self): + if not self.td.get('PTEST_EXPECT_FAILURE'): + self.skipTest('Cannot run ptests with @expectedFailure as ptests are required to pass') + self.do_ptestrunner() + + @skipIfNotFeature('ptest', 'Test requires ptest to be in DISTRO_FEATURES') + @OETestDepends(['ssh.SSHTest.test_ssh']) + @OEHasPackage(['ptest-runner']) + def test_ptestrunner_expectsuccess(self): + if self.td.get('PTEST_EXPECT_FAILURE'): + self.skipTest('Cannot run ptests without @expectedFailure as ptests are expected to fail') + self.do_ptestrunner() + + def do_ptestrunner(self): + status, output = self.target.run('which ptest-runner', 0) + if status != 0: + self.skipTest("No -ptest packages are installed in the image") test_log_dir = self.td.get('TEST_LOG_DIR', '') # The TEST_LOG_DIR maybe NULL when testimage is added after # testdata.json is generated. if not test_log_dir: test_log_dir = os.path.join(self.td.get('WORKDIR', ''), 'testimage') + # Make the test output path absolute, otherwise the output content will be + # created relative to current directory + if not os.path.isabs(test_log_dir): + test_log_dir = os.path.join(self.td.get('TOPDIR', ''), test_log_dir) # Don't use self.td.get('DATETIME'), it's from testdata.json, not # up-to-date, and may cause "File exists" when re-reun. - datetime = datetime.datetime.now().strftime('%Y%m%d%H%M%S') + timestamp = datetime.datetime.now().strftime('%Y%m%d%H%M%S') ptest_log_dir_link = os.path.join(test_log_dir, 'ptest_log') - ptest_log_dir = '%s.%s' % (ptest_log_dir_link, datetime) + ptest_log_dir = '%s.%s' % (ptest_log_dir_link, timestamp) ptest_runner_log = os.path.join(ptest_log_dir, 'ptest-runner.log') - status, output = self.target.run('ptest-runner', 0) + libdir = self.td.get('libdir', '') + ptest_dirs = [ '/usr/lib' ] + if not libdir in ptest_dirs: + ptest_dirs.append(libdir) + status, output = self.target.run('ptest-runner -t 450 -d \"{}\"'.format(' '.join(ptest_dirs)), 0) os.makedirs(ptest_log_dir) with open(ptest_runner_log, 'w') as f: f.write(output) @@ -73,10 +67,54 @@ class PtestRunnerTest(OERuntimeTestCase): # status != 0 is OK since some ptest tests may fail self.assertTrue(status != 127, msg="Cannot execute ptest-runner!") + if not hasattr(self.tc, "extraresults"): + self.tc.extraresults = {} + extras = self.tc.extraresults + extras['ptestresult.rawlogs'] = {'log': output} + # Parse and save results - parse_result = self.parse_ptest(ptest_runner_log) - parse_result.log_as_files(ptest_log_dir, test_status = ['pass','fail', 'skip']) + parser = PtestParser() + results, sections = parser.parse(ptest_runner_log) + parser.results_as_files(ptest_log_dir) if os.path.exists(ptest_log_dir_link): # Remove the old link to create a new one os.remove(ptest_log_dir_link) os.symlink(os.path.basename(ptest_log_dir), ptest_log_dir_link) + + extras['ptestresult.sections'] = sections + + zerolength = [] + trans = str.maketrans("()", "__") + for section in results: + for test in results[section]: + result = results[section][test] + testname = "ptestresult." + (section or "No-section") + "." + "_".join(test.translate(trans).split()) + extras[testname] = {'status': result} + if not results[section]: + zerolength.append(section) + + failed_tests = {} + + for section in sections: + if 'exitcode' in sections[section].keys() or 'timeout' in sections[section].keys(): + failed_tests[section] = sections[section]["log"] + + for section in results: + failed_testcases = [ "_".join(test.translate(trans).split()) for test in results[section] if results[section][test] == 'FAILED' ] + if failed_testcases: + failed_tests[section] = failed_testcases + + failmsg = "" + status, output = self.target.run('dmesg | grep "Killed process"', 0) + if output: + failmsg = "ERROR: Processes were killed by the OOM Killer:\n%s\n" % output + + if failed_tests: + failmsg = failmsg + "\nFailed ptests:\n%s\n" % pprint.pformat(failed_tests) + + if zerolength: + failmsg = failmsg + "\nptests which had no test results:\n%s" % pprint.pformat(zerolength) + + if failmsg: + self.logger.warning("There were failing ptests.") + self.fail(failmsg) |