diff options
Diffstat (limited to 'scripts')
-rwxr-xr-x | scripts/lib/resulttool/manualexecution.py | 137 | ||||
-rwxr-xr-x | scripts/resulttool | 8 |
2 files changed, 145 insertions, 0 deletions
diff --git a/scripts/lib/resulttool/manualexecution.py b/scripts/lib/resulttool/manualexecution.py new file mode 100755 index 0000000000..64ec581a9f --- /dev/null +++ b/scripts/lib/resulttool/manualexecution.py @@ -0,0 +1,137 @@ +# test case management tool - manual execution from testopia test cases +# +# Copyright (c) 2018, Intel Corporation. +# +# This program is free software; you can redistribute it and/or modify it +# under the terms and conditions of the GNU General Public License, +# version 2, as published by the Free Software Foundation. +# +# This program is distributed in the hope it will be useful, but WITHOUT +# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or +# FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for +# more details. +# +import argparse +import json +import os +import sys +import datetime +import re +from oeqa.core.runner import OETestResultJSONHelper +from resulttool.resultsutils import load_json_file + +class ManualTestRunner(object): + def __init__(self): + self.jdata = '' + self.test_module = '' + self.test_suite = '' + self.test_cases = '' + self.configuration = '' + self.starttime = '' + self.result_id = '' + self.write_dir = '' + + def _get_testcases(self, file): + self.jdata = load_json_file(file) + self.test_cases = [] + self.test_module = self.jdata[0]['test']['@alias'].split('.', 2)[0] + self.test_suite = self.jdata[0]['test']['@alias'].split('.', 2)[1] + for i in self.jdata: + self.test_cases.append(i['test']['@alias'].split('.', 2)[2]) + + def _get_input(self, config): + while True: + output = input('{} = '.format(config)) + if re.match('^[a-zA-Z0-9_]+$', output): + break + print('Only alphanumeric and underscore are allowed. Please try again') + return output + + def _create_config(self): + self.configuration = {} + while True: + try: + conf_total = int(input('\nPlease provide how many configuration you want to save \n')) + break + except ValueError: + print('Invalid input. Please provide input as a number not character.') + for i in range(conf_total): + print('---------------------------------------------') + print('This is configuration #%s ' % (i + 1) + '. Please provide configuration name and its value') + print('---------------------------------------------') + name_conf = self._get_input('Configuration Name') + value_conf = self._get_input('Configuration Value') + print('---------------------------------------------\n') + self.configuration[name_conf.upper()] = value_conf + current_datetime = datetime.datetime.now() + self.starttime = current_datetime.strftime('%Y%m%d%H%M%S') + self.configuration['STARTTIME'] = self.starttime + self.configuration['TEST_TYPE'] = self.test_module + + def _create_result_id(self): + self.result_id = 'manual_' + self.test_module + '_' + self.starttime + + def _execute_test_steps(self, test_id): + test_result = {} + testcase_id = self.test_module + '.' + self.test_suite + '.' + self.test_cases[test_id] + total_steps = len(self.jdata[test_id]['test']['execution'].keys()) + print('------------------------------------------------------------------------') + print('Executing test case:' + '' '' + self.test_cases[test_id]) + print('------------------------------------------------------------------------') + print('You have total ' + str(total_steps) + ' test steps to be executed.') + print('------------------------------------------------------------------------\n') + for step in sorted((self.jdata[test_id]['test']['execution']).keys()): + print('Step %s: ' % step + self.jdata[test_id]['test']['execution']['%s' % step]['action']) + print('Expected output: ' + self.jdata[test_id]['test']['execution']['%s' % step]['expected_results']) + done = input('\nPlease press ENTER when you are done to proceed to next step.\n') + while True: + done = input('\nPlease provide test results: (P)assed/(F)ailed/(B)locked/(S)kipped? \n') + done = done.lower() + result_types = {'p':'PASSED', + 'f':'FAILED', + 'b':'BLOCKED', + 's':'SKIPPED'} + if done in result_types: + for r in result_types: + if done == r: + res = result_types[r] + if res == 'FAILED': + log_input = input('\nPlease enter the error and the description of the log: (Ex:log:211 Error Bitbake)\n') + test_result.update({testcase_id: {'status': '%s' % res, 'log': '%s' % log_input}}) + else: + test_result.update({testcase_id: {'status': '%s' % res}}) + break + print('Invalid input!') + return test_result + + def _create_write_dir(self): + basepath = os.environ['BUILDDIR'] + self.write_dir = basepath + '/tmp/log/manual/' + + def run_test(self, file): + self._get_testcases(file) + self._create_config() + self._create_result_id() + self._create_write_dir() + test_results = {} + print('\nTotal number of test cases in this test suite: ' + '%s\n' % len(self.jdata)) + for i in range(0, len(self.jdata)): + test_result = self._execute_test_steps(i) + test_results.update(test_result) + return self.configuration, self.result_id, self.write_dir, test_results + +def manualexecution(args, logger): + testrunner = ManualTestRunner() + get_configuration, get_result_id, get_write_dir, get_test_results = testrunner.run_test(args.file) + resultjsonhelper = OETestResultJSONHelper() + resultjsonhelper.dump_testresult_file(get_write_dir, get_configuration, get_result_id, + get_test_results) + return 0 + +def register_commands(subparsers): + """Register subcommands from this plugin""" + parser_build = subparsers.add_parser('manualexecution', help='helper script for results populating during manual test execution.', + description='helper script for results populating during manual test execution. You can find manual test case JSON file in meta/lib/oeqa/manual/', + group='manualexecution') + parser_build.set_defaults(func=manualexecution) + parser_build.add_argument('file', help='specify path to manual test case JSON file.Note: Please use \"\" to encapsulate the file path.')
\ No newline at end of file diff --git a/scripts/resulttool b/scripts/resulttool index ebb5fc81c9..13430e192a 100755 --- a/scripts/resulttool +++ b/scripts/resulttool @@ -17,6 +17,11 @@ # To perform regression file analysis, execute the below # $ resulttool regression-file <base_result_file> <target_result_file> # +# To execute manual test cases, execute the below +# $ resulttool manualexecution <manualjsonfile> +# +# By default testresults.json for manualexecution store in <build>/tmp/log/manual/ +# # Copyright (c) 2019, Intel Corporation. # # This program is free software; you can redistribute it and/or modify it @@ -42,6 +47,7 @@ import resulttool.merge import resulttool.store import resulttool.regression import resulttool.report +import resulttool.manualexecution logger = scriptutils.logger_create('resulttool') def _validate_user_input_arguments(args): @@ -58,6 +64,8 @@ def main(): parser.add_argument('-q', '--quiet', help='print only errors', action='store_true') subparsers = parser.add_subparsers(dest="subparser_name", title='subcommands', metavar='<subcommand>') subparsers.required = True + subparsers.add_subparser_group('manualexecution', 'manual testcases', 300) + resulttool.manualexecution.register_commands(subparsers) subparsers.add_subparser_group('setup', 'setup', 200) resulttool.merge.register_commands(subparsers) resulttool.store.register_commands(subparsers) |