Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
236 changes: 236 additions & 0 deletions tests/run_tests.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,236 @@
#!/bin/bash

function show_help_and_exit()
{
echo "Usage ${SCRIPT} [options]"
echo " options with (*) must be provided"
echo " -h -? : get this help"
echo " -c : specify test cases to execute (default: none, executed all matched)"
echo " -d : specify DUT name (*)"
echo " -e : specify extra parameter(s) (default: none)"
echo " -f : specify testbed file (default testbed.csv)"
echo " -i : specify inventory name"
echo " -k : specify file log level: error|warning|info|debug (default debug)"
echo " -l : specify cli log level: error|warning|info|debug (default warning)"
echo " -m : specify test method group|individual|dump (default group)"
Comment thread
daall marked this conversation as resolved.
Outdated
echo " -o : omit the file logs"
echo " -n : specify testbed name (*)"
echo " -s : specify list of scripts to skip (default: none)"
echo " -t : specify toplogy: t0|t1|any|combo like t0,any (*)"
echo " -u : bypass util group"

exit $1
}
Comment thread
daall marked this conversation as resolved.

function validate_parameters()
{
RET=0

if [[ -z ${DUT_NAME} ]]; then
echo "DUT name (-d) is not set.."
RET=1
fi

if [[ -z ${TESTBED_NAME} ]]; then
echo "Testbed name (-n) is not set.."
RET=2
fi

if [[ -z ${TOPOLOGY} ]]; then
echo "TOPOLOGY (-t) is not set.."
RET=3
fi

if [[ ${RET} != 0 ]]; then
show_help_and_exit ${RET}
fi
}

function setup_environment()
{
SCRIPT=$0
FULL_PATH=$(realpath ${SCRIPT})
SCRIPT_PATH=$(dirname ${FULL_PATH})
BASE_PATH=$(dirname ${SCRIPT_PATH})
CLI_LOG_LEVEL='warning'
FILE_LOG_LEVEL='debug'
SKIP_FOLDERS="ptftests acstests saitests"
EXTRA_PARAMETERS=""
TEST_CASES=""
OMIT_FILE_LOG="False"
BYPASS_UTIL="False"

TEST_METHOD='group'

TESTBED_FILE="${BASE_PATH}/ansible/testbed.csv"
INVENTORY="${BASE_PATH}/ansible/lab,${BASE_PATH}/ansible/veos"

export ANSIBLE_CONFIG=${BASE_PATH}/ansible
export ANSIBLE_LIBRARY=${BASE_PATH}/ansible/library/
}

function setup_test_options()
{
PYTEST_COMMON_OPTS="--inventory ${INVENTORY} \
--host-pattern ${DUT_NAME} \
--testbed ${TESTBED_NAME} \
--testbed_file ${TESTBED_FILE} \
--log-cli-level ${CLI_LOG_LEVEL} \
--log-file-level ${FILE_LOG_LEVEL} \
--allow_recover \
--showlocals \
--assert plain \
-rav"

for skip in ${SKIP_SCRIPTS} ${SKIP_FOLDERS}; do
PYTEST_COMMON_OPTS="${PYTEST_COMMON_OPTS} --ignore=${skip}"
done

if [[ x"${OMIT_FILE_LOG}" == x"True" ]]; then
UTIL_LOGGING_OPTIONS=""
TEST_LOGGING_OPTIONS=""
else
UTIL_LOGGING_OPTIONS="--junit-xml=logs/util.xml --log-file=logs/util.log"
TEST_LOGGING_OPTIONS="--junit-xml=logs/tr.xml --log-file=logs/test.log"
Comment thread
daall marked this conversation as resolved.
fi
UTIL_TOPOLOGY_OPTIONS="--topology util"
TEST_TOPOLOGY_OPTIONS="--topology ${TOPOLOGY}"
}

function run_dump_tests()
{
echo "SCRIPT: ${SCRIPT}"
echo "FULL_PATH: ${FULL_PATH}"
echo "SCRIPT_PATH: ${SCRIPT_PATH}"
echo "BASE_PATH: ${BASE_PATH}"

echo "ANSIBLE_CONFIG: ${ANSIBLE_CONFIG}"
echo "ANSIBLE_LIBRARY: ${ANSIBLE_LIBRARY}"
echo "BYPASS_UTIL: ${BYPASS_UTIL}"
echo "CLI_LOG_LEVEL: ${CLI_LOG_LEVEL}"
echo "EXTRA_PARAMETERS: ${EXTRA_PARAMETERS}"
echo "FILE_LOG_LEVEL: ${FILE_LOG_LEVEL}"
echo "INVENTORY: ${INVENTORY}"
echo "OMIT_FILE_LOG: ${OMIT_FILE_LOG}"
echo "SKIP_SCRIPTS: ${SKIP_SCRIPTS}"
echo "SKIP_FOLDERS: ${SKIP_FOLDERS}"
echo "TEST_CASES: ${TEST_CASES}"
echo "TEST_METHOD: ${TEST_METHOD}"
echo "TESTBED_FILE: ${TESTBED_FILE}"
echo "TEST_LOGGING_OPTIONS: ${TEST_LOGGING_OPTIONS}"
echo "TEST_TOPOLOGY_OPTIONS: ${TEST_TOPOLOGY_OPTIONS}"
echo "UTIL_LOGGING_OPTIONS: ${UTIL_LOGGING_OPTIONS}"
echo "UTIL_TOPOLOGY_OPTIONS: ${UTIL_TOPOLOGY_OPTIONS}"

echo "PYTEST_COMMON_OPTS: ${PYTEST_COMMON_OPTS}"
}

function run_group_tests()
{
echo "=== running tests in groups ==="
if [[ x"${BYPASS_UTIL}" == x"False" ]]; then
py.test ${PYTEST_COMMON_OPTS} ${UTIL_LOGGING_OPTIONS} ${UTIL_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS}
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This test_restart_syncd case in tests/platform_tests/test_sequential_restart.py has been marked as skip. It's unnecessary to skip it again here.

Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It would be better to add sleep 120 here to wait a while for the DUT to learn all the IP routes.

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! Changed both.

fi
py.test ${PYTEST_COMMON_OPTS} ${TEST_LOGGING_OPTIONS} ${TEST_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS} ${TEST_CASES}
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same for test_restart_syncd here.

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Changed.

}

function run_individual_tests()
{
echo "=== running tests individually ==="
if [[ x"${BYPASS_UTIL}" == x"False" ]]; then
py.test ${PYTEST_COMMON_OPTS} ${UTIL_LOGGING_OPTIONS} ${UTIL_TOPOLOGY_OPTIONS} -k "not test_restart_syncd" ${EXTRA_PARAMETERS}
Copy link
Copy Markdown
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same for test_restart_syncd here.

Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Changed.

fi

sleep 120

SKIP_SCRIPTS="${SKIP_SCRIPTS} test_announce_routes.py test_nbr_health.py"

ignores=$(python -c "print '|'.join('''$SKIP_FOLDERS'''.split())")

all_scripts=$(find ./ -name 'test_*.py' | sed s:^./:: | grep -vE "^(${SKIP_FOLDERS})")
test_scripts=$(python -c "print '\n'.join(set('''$all_scripts'''.split()) - set('''$SKIP_SCRIPTS'''.split()))" | sort)

EXIT_CODE=0

for test_script in $test_scripts; do

tdir=$(dirname $test_script)
ts=$(basename $test_script)
tn=${ts%.py}
Comment thread
daall marked this conversation as resolved.
Outdated
if [[ $tdir != "." ]]; then
mkdir -p logs/$tdir
fi
py.test $PYTEST_COMMON_OPTS --log-file logs/$tdir/$tn.log --junitxml=results/$tdir/$tn.xml ${TEST_TOPOLOGY_OPTIONS} $test_script -k "not test_restart_syncd" ${EXTRA_PARAMETERS}

ret_code=$?

# If test passed, no need to keep its log.
Comment thread
daall marked this conversation as resolved.
if [ $ret_code -eq 0 ]; then
rm logs/$tdir/$tn.log
else
EXIT_CODE=1
fi

done

return $EXIT_CODE
}

setup_environment

while getopts "h?c:d:e:f:i:k:l:m:n:os:t:u" opt; do
case ${opt} in
h|\? )
show_help_and_exit 0
;;
c )
TEST_CASES=${OPTARG}
;;
d )
DUT_NAME=${OPTARG}
;;
e )
EXTRA_PARAMETERS=${OPTARG}
;;
f )
TESTBED_FILE=${OPTARG}
;;
i )
INVENTORY=${OPTARG}
;;
k )
FILE_LOG_LEVEL=${OPTARG}
;;
l )
CLI_LOG_LEVEL=${OPTARG}
;;
m )
TEST_METHOD=${OPTARG}
;;
n )
TESTBED_NAME=${OPTARG}
;;
o )
OMIT_FILE_LOG="True"
;;
s )
SKIP_SCRIPTS=${OPTARG}
;;
t )
TOPOLOGY=${OPTARG}
;;
u )
BYPASS_UTIL="True"
;;
esac
done

validate_parameters
setup_test_options

if [[ x"${OMIT_FILE_LOG}" != x"True" ]]; then
rm -rf logs
mkdir logs
fi

run_${TEST_METHOD}_tests