-
Notifications
You must be signed in to change notification settings - Fork 1k
[test runner] add a script to run pytests #1861
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 5 commits
772608b
5faa155
ea82cc1
17a9c9a
58f717d
b7a45b3
dd8ac04
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,250 @@ | ||
| #!/bin/bash | ||
|
|
||
| function show_help_and_exit() | ||
| { | ||
| echo "Usage ${SCRIPT} [options]" | ||
| echo " options with (*) must be provided" | ||
| echo " -h -? : get this help" | ||
| echo " -c : specify test cases to execute (default: none, executed all matched)" | ||
| echo " -d : specify DUT name (*)" | ||
| echo " -e : specify extra parameter(s) (default: none)" | ||
| echo " -f : specify testbed file (default testbed.csv)" | ||
| echo " -i : specify inventory name" | ||
| echo " -k : specify file log level: error|warning|info|debug (default debug)" | ||
| echo " -l : specify cli log level: error|warning|info|debug (default warning)" | ||
| echo " -m : specify test method group|individual|debug (default group)" | ||
| echo " -o : omit the file logs" | ||
| echo " -n : specify testbed name (*)" | ||
| echo " -r : retain individual file log for suceeded tests (default: remove)" | ||
| echo " -s : specify list of scripts to skip (default: none)" | ||
| echo " -t : specify toplogy: t0|t1|any|combo like t0,any (*)" | ||
| echo " -u : bypass util group" | ||
|
|
||
| exit $1 | ||
| } | ||
|
|
||
| function validate_parameters() | ||
| { | ||
| RET=0 | ||
|
|
||
| if [[ -z ${DUT_NAME} ]]; then | ||
| echo "DUT name (-d) is not set.." | ||
| RET=1 | ||
| fi | ||
|
|
||
| if [[ -z ${TESTBED_NAME} ]]; then | ||
| echo "Testbed name (-n) is not set.." | ||
| RET=2 | ||
| fi | ||
|
|
||
| if [[ -z ${TOPOLOGY} ]]; then | ||
| echo "TOPOLOGY (-t) is not set.." | ||
| RET=3 | ||
| fi | ||
|
|
||
| if [[ ${RET} != 0 ]]; then | ||
| show_help_and_exit ${RET} | ||
| fi | ||
| } | ||
|
|
||
| function setup_environment() | ||
| { | ||
| SCRIPT=$0 | ||
| FULL_PATH=$(realpath ${SCRIPT}) | ||
| SCRIPT_PATH=$(dirname ${FULL_PATH}) | ||
| BASE_PATH=$(dirname ${SCRIPT_PATH}) | ||
|
|
||
| BYPASS_UTIL="False" | ||
| CLI_LOG_LEVEL='warning' | ||
| EXTRA_PARAMETERS="" | ||
| FILE_LOG_LEVEL='debug' | ||
| INVENTORY="${BASE_PATH}/ansible/lab,${BASE_PATH}/ansible/veos" | ||
| OMIT_FILE_LOG="False" | ||
| RETAIN_SUCCESS_LOG="False" | ||
| SKIP_SCRIPTS="" | ||
| SKIP_FOLDERS="ptftests acstests saitests" | ||
| TESTBED_FILE="${BASE_PATH}/ansible/testbed.csv" | ||
| TEST_CASES="" | ||
| TEST_METHOD='group' | ||
|
|
||
| export ANSIBLE_CONFIG=${BASE_PATH}/ansible | ||
| export ANSIBLE_LIBRARY=${BASE_PATH}/ansible/library/ | ||
| } | ||
|
|
||
| function setup_test_options() | ||
| { | ||
| PYTEST_COMMON_OPTS="--inventory ${INVENTORY} \ | ||
| --host-pattern ${DUT_NAME} \ | ||
| --testbed ${TESTBED_NAME} \ | ||
| --testbed_file ${TESTBED_FILE} \ | ||
| --log-cli-level ${CLI_LOG_LEVEL} \ | ||
| --log-file-level ${FILE_LOG_LEVEL} \ | ||
| --allow_recover \ | ||
| --showlocals \ | ||
| --assert plain \ | ||
| -rav" | ||
|
|
||
| for skip in ${SKIP_SCRIPTS} ${SKIP_FOLDERS}; do | ||
| PYTEST_COMMON_OPTS="${PYTEST_COMMON_OPTS} --ignore=${skip}" | ||
| done | ||
|
|
||
| if [[ -d logs ]]; then | ||
| rm -rf logs | ||
| fi | ||
|
|
||
| if [[ x"${OMIT_FILE_LOG}" == x"True" ]]; then | ||
| UTIL_LOGGING_OPTIONS="" | ||
| TEST_LOGGING_OPTIONS="" | ||
| else | ||
| mkdir logs | ||
|
|
||
| UTIL_LOGGING_OPTIONS="--junit-xml=logs/util.xml --log-file=logs/util.log" | ||
| TEST_LOGGING_OPTIONS="--junit-xml=logs/tr.xml --log-file=logs/test.log" | ||
daall marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| fi | ||
| UTIL_TOPOLOGY_OPTIONS="--topology util" | ||
| TEST_TOPOLOGY_OPTIONS="--topology ${TOPOLOGY}" | ||
| } | ||
|
|
||
| function run_debug_tests() | ||
| { | ||
| echo "=== Show test settings ===" | ||
| echo "SCRIPT: ${SCRIPT}" | ||
| echo "FULL_PATH: ${FULL_PATH}" | ||
| echo "SCRIPT_PATH: ${SCRIPT_PATH}" | ||
| echo "BASE_PATH: ${BASE_PATH}" | ||
|
|
||
| echo "ANSIBLE_CONFIG: ${ANSIBLE_CONFIG}" | ||
| echo "ANSIBLE_LIBRARY: ${ANSIBLE_LIBRARY}" | ||
| echo "BYPASS_UTIL: ${BYPASS_UTIL}" | ||
| echo "CLI_LOG_LEVEL: ${CLI_LOG_LEVEL}" | ||
| echo "EXTRA_PARAMETERS: ${EXTRA_PARAMETERS}" | ||
| echo "FILE_LOG_LEVEL: ${FILE_LOG_LEVEL}" | ||
| echo "INVENTORY: ${INVENTORY}" | ||
| echo "OMIT_FILE_LOG: ${OMIT_FILE_LOG}" | ||
| echo "RETAIN_SUCCESS_LOG: ${RETAIN_SUCCESS_LOG}" | ||
| echo "SKIP_SCRIPTS: ${SKIP_SCRIPTS}" | ||
| echo "SKIP_FOLDERS: ${SKIP_FOLDERS}" | ||
| echo "TEST_CASES: ${TEST_CASES}" | ||
| echo "TEST_METHOD: ${TEST_METHOD}" | ||
| echo "TESTBED_FILE: ${TESTBED_FILE}" | ||
| echo "TEST_LOGGING_OPTIONS: ${TEST_LOGGING_OPTIONS}" | ||
| echo "TEST_TOPOLOGY_OPTIONS: ${TEST_TOPOLOGY_OPTIONS}" | ||
| echo "UTIL_LOGGING_OPTIONS: ${UTIL_LOGGING_OPTIONS}" | ||
| echo "UTIL_TOPOLOGY_OPTIONS: ${UTIL_TOPOLOGY_OPTIONS}" | ||
|
|
||
| echo "PYTEST_COMMON_OPTS: ${PYTEST_COMMON_OPTS}" | ||
| } | ||
|
|
||
| function run_group_tests() | ||
| { | ||
| if [[ x"${BYPASS_UTIL}" == x"False" ]]; then | ||
| echo "=== Runing pre-test cases ===" | ||
| py.test ${PYTEST_COMMON_OPTS} ${UTIL_LOGGING_OPTIONS} ${UTIL_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS} | ||
|
||
| fi | ||
| echo "=== Running tests in groups ===" | ||
| py.test ${PYTEST_COMMON_OPTS} ${TEST_LOGGING_OPTIONS} ${TEST_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS} ${TEST_CASES} | ||
|
||
| } | ||
|
|
||
| function run_individual_tests() | ||
| { | ||
| if [[ x"${BYPASS_UTIL}" == x"False" ]]; then | ||
| echo "=== Runing pre-test cases ===" | ||
| py.test ${PYTEST_COMMON_OPTS} ${UTIL_LOGGING_OPTIONS} ${UTIL_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS} | ||
|
||
| fi | ||
|
|
||
| sleep 120 | ||
|
|
||
| SKIP_SCRIPTS="${SKIP_SCRIPTS} test_announce_routes.py test_nbr_health.py" | ||
|
|
||
| ignores=$(python -c "print '|'.join('''$SKIP_FOLDERS'''.split())") | ||
|
|
||
| all_scripts=$(find ./ -name 'test_*.py' | sed s:^./:: | grep -vE "^(${SKIP_FOLDERS})") | ||
| test_scripts=$(python -c "print '\n'.join(set('''$all_scripts'''.split()) - set('''$SKIP_SCRIPTS'''.split()))" | sort) | ||
|
|
||
| EXIT_CODE=0 | ||
|
|
||
| echo "=== Running tests individually ===" | ||
| for test_script in ${test_scripts}; do | ||
| if [[ x"${OMIT_FILE_LOG}" != x"True" ]]; then | ||
| test_dir=$(dirname ${test_script}) | ||
| script_name=$(basename ${test_script}) | ||
| test_name=${script_name%.py} | ||
| if [[ ${test_dir} != "." ]]; then | ||
| mkdir -p logs/${test_dir} | ||
| fi | ||
| TEST_LOGGING_OPTIONS="--log-file logs/${test_dir}/${test_name}.log --junitxml=results/${test_dir}/${test_name}.xml" | ||
| fi | ||
|
|
||
| py.test ${PYTEST_COMMON_OPTS} ${TEST_LOGGING_OPTIONS} ${TEST_TOPOLOGY_OPTIONS} ${test_script} -k "not test_restart_syncd" ${EXTRA_PARAMETERS} | ||
| ret_code=$? | ||
|
|
||
| # If test passed, no need to keep its log. | ||
daall marked this conversation as resolved.
Show resolved
Hide resolved
|
||
| if [ ${ret_code} -eq 0 ]; then | ||
| if [[ x"${OMIT_FILE_LOG}" != x"True" && x"${RETAIN_SUCCESS_LOG}" == x"False" ]]; then | ||
| rm -f logs/${test_dir}/${test_name}.log | ||
| fi | ||
| else | ||
| EXIT_CODE=1 | ||
| fi | ||
|
|
||
| done | ||
|
|
||
| return ${EXIT_CODE} | ||
| } | ||
|
|
||
| setup_environment | ||
|
|
||
| while getopts "h?c:d:e:f:i:k:l:m:n:ors:t:u" opt; do | ||
| case ${opt} in | ||
| h|\? ) | ||
| show_help_and_exit 0 | ||
| ;; | ||
| c ) | ||
| TEST_CASES=${OPTARG} | ||
| ;; | ||
| d ) | ||
| DUT_NAME=${OPTARG} | ||
| ;; | ||
| e ) | ||
| EXTRA_PARAMETERS=${OPTARG} | ||
| ;; | ||
| f ) | ||
| TESTBED_FILE=${OPTARG} | ||
| ;; | ||
| i ) | ||
| INVENTORY=${OPTARG} | ||
| ;; | ||
| k ) | ||
| FILE_LOG_LEVEL=${OPTARG} | ||
| ;; | ||
| l ) | ||
| CLI_LOG_LEVEL=${OPTARG} | ||
| ;; | ||
| m ) | ||
| TEST_METHOD=${OPTARG} | ||
| ;; | ||
| n ) | ||
| TESTBED_NAME=${OPTARG} | ||
| ;; | ||
| o ) | ||
| OMIT_FILE_LOG="True" | ||
| ;; | ||
| r ) | ||
| RETAIN_SUCCESS_LOG="True" | ||
| ;; | ||
| s ) | ||
| SKIP_SCRIPTS=${OPTARG} | ||
| ;; | ||
| t ) | ||
| TOPOLOGY=${OPTARG} | ||
| ;; | ||
| u ) | ||
| BYPASS_UTIL="True" | ||
| ;; | ||
| esac | ||
| done | ||
|
|
||
| validate_parameters | ||
| setup_test_options | ||
|
|
||
| run_${TEST_METHOD}_tests | ||
Uh oh!
There was an error while loading. Please reload this page.